{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9779065555957986, "eval_steps": 500, "global_step": 27000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 3.621876131836291e-05, "grad_norm": 19.014660562399033, "learning_rate": 2.4125452352231605e-08, "loss": 1.4414, "step": 1 }, { "epoch": 7.243752263672582e-05, "grad_norm": 23.290969601631677, "learning_rate": 4.825090470446321e-08, "loss": 1.456, "step": 2 }, { "epoch": 0.00010865628395508874, "grad_norm": 22.0312635218802, "learning_rate": 7.237635705669482e-08, "loss": 1.4467, "step": 3 }, { "epoch": 0.00014487504527345164, "grad_norm": 11.092012391546637, "learning_rate": 9.650180940892642e-08, "loss": 0.957, "step": 4 }, { "epoch": 0.00018109380659181456, "grad_norm": 19.842911950384583, "learning_rate": 1.2062726176115803e-07, "loss": 1.4686, "step": 5 }, { "epoch": 0.00021731256791017748, "grad_norm": 20.609251598946003, "learning_rate": 1.4475271411338965e-07, "loss": 1.3825, "step": 6 }, { "epoch": 0.0002535313292285404, "grad_norm": 17.890624007179202, "learning_rate": 1.6887816646562124e-07, "loss": 1.4625, "step": 7 }, { "epoch": 0.0002897500905469033, "grad_norm": 23.21993258764649, "learning_rate": 1.9300361881785284e-07, "loss": 1.3342, "step": 8 }, { "epoch": 0.0003259688518652662, "grad_norm": 17.90402616635559, "learning_rate": 2.1712907117008446e-07, "loss": 1.356, "step": 9 }, { "epoch": 0.0003621876131836291, "grad_norm": 19.046550895055116, "learning_rate": 2.4125452352231605e-07, "loss": 1.3826, "step": 10 }, { "epoch": 0.00039840637450199205, "grad_norm": 16.85449554710326, "learning_rate": 2.653799758745477e-07, "loss": 1.4111, "step": 11 }, { "epoch": 0.00043462513582035497, "grad_norm": 18.38324937734976, "learning_rate": 2.895054282267793e-07, "loss": 1.3782, "step": 12 }, { "epoch": 0.00047084389713871784, "grad_norm": 16.289395255296604, "learning_rate": 3.1363088057901086e-07, "loss": 1.3803, "step": 13 }, { "epoch": 0.0005070626584570808, "grad_norm": 15.48879136235226, "learning_rate": 3.377563329312425e-07, "loss": 1.3129, "step": 14 }, { "epoch": 0.0005432814197754436, "grad_norm": 13.310446639043898, "learning_rate": 3.618817852834741e-07, "loss": 1.3071, "step": 15 }, { "epoch": 0.0005795001810938066, "grad_norm": 12.002236818416927, "learning_rate": 3.860072376357057e-07, "loss": 1.3302, "step": 16 }, { "epoch": 0.0006157189424121695, "grad_norm": 11.7833908587119, "learning_rate": 4.101326899879373e-07, "loss": 0.9856, "step": 17 }, { "epoch": 0.0006519377037305324, "grad_norm": 12.534500333149413, "learning_rate": 4.342581423401689e-07, "loss": 1.0769, "step": 18 }, { "epoch": 0.0006881564650488953, "grad_norm": 10.700094019396548, "learning_rate": 4.583835946924005e-07, "loss": 1.3646, "step": 19 }, { "epoch": 0.0007243752263672582, "grad_norm": 10.629737644842686, "learning_rate": 4.825090470446321e-07, "loss": 1.3394, "step": 20 }, { "epoch": 0.0007605939876856212, "grad_norm": 6.670076605815865, "learning_rate": 5.066344993968638e-07, "loss": 1.2915, "step": 21 }, { "epoch": 0.0007968127490039841, "grad_norm": 6.0791512189315435, "learning_rate": 5.307599517490953e-07, "loss": 1.2461, "step": 22 }, { "epoch": 0.000833031510322347, "grad_norm": 6.205537725567378, "learning_rate": 5.548854041013269e-07, "loss": 1.2912, "step": 23 }, { "epoch": 0.0008692502716407099, "grad_norm": 6.713383984742424, "learning_rate": 5.790108564535586e-07, "loss": 1.2675, "step": 24 }, { "epoch": 0.0009054690329590728, "grad_norm": 9.265962339664304, "learning_rate": 6.031363088057902e-07, "loss": 0.8881, "step": 25 }, { "epoch": 0.0009416877942774357, "grad_norm": 6.259259107929034, "learning_rate": 6.272617611580217e-07, "loss": 1.2741, "step": 26 }, { "epoch": 0.0009779065555957987, "grad_norm": 9.70371690041291, "learning_rate": 6.513872135102534e-07, "loss": 0.9293, "step": 27 }, { "epoch": 0.0010141253169141616, "grad_norm": 5.795366846787815, "learning_rate": 6.75512665862485e-07, "loss": 1.2597, "step": 28 }, { "epoch": 0.0010503440782325246, "grad_norm": 6.650884567940559, "learning_rate": 6.996381182147166e-07, "loss": 0.9457, "step": 29 }, { "epoch": 0.0010865628395508873, "grad_norm": 6.266041996271789, "learning_rate": 7.237635705669482e-07, "loss": 1.2819, "step": 30 }, { "epoch": 0.0011227816008692502, "grad_norm": 7.776973797953041, "learning_rate": 7.478890229191799e-07, "loss": 1.2543, "step": 31 }, { "epoch": 0.001159000362187613, "grad_norm": 6.265748419937368, "learning_rate": 7.720144752714113e-07, "loss": 1.2165, "step": 32 }, { "epoch": 0.001195219123505976, "grad_norm": 3.3199360836060285, "learning_rate": 7.96139927623643e-07, "loss": 0.831, "step": 33 }, { "epoch": 0.001231437884824339, "grad_norm": 6.4723354253182865, "learning_rate": 8.202653799758746e-07, "loss": 1.1934, "step": 34 }, { "epoch": 0.0012676566461427019, "grad_norm": 6.680464613476801, "learning_rate": 8.443908323281063e-07, "loss": 1.2184, "step": 35 }, { "epoch": 0.0013038754074610648, "grad_norm": 5.969786459124488, "learning_rate": 8.685162846803378e-07, "loss": 1.1663, "step": 36 }, { "epoch": 0.0013400941687794277, "grad_norm": 3.349581328794547, "learning_rate": 8.926417370325695e-07, "loss": 0.8502, "step": 37 }, { "epoch": 0.0013763129300977906, "grad_norm": 3.6080372254037942, "learning_rate": 9.16767189384801e-07, "loss": 0.8641, "step": 38 }, { "epoch": 0.0014125316914161536, "grad_norm": 5.3234802092106115, "learning_rate": 9.408926417370326e-07, "loss": 1.2508, "step": 39 }, { "epoch": 0.0014487504527345165, "grad_norm": 4.799968017407003, "learning_rate": 9.650180940892642e-07, "loss": 1.2043, "step": 40 }, { "epoch": 0.0014849692140528794, "grad_norm": 3.4510812558602417, "learning_rate": 9.891435464414958e-07, "loss": 0.9178, "step": 41 }, { "epoch": 0.0015211879753712423, "grad_norm": 4.6388889494849606, "learning_rate": 1.0132689987937276e-06, "loss": 1.2443, "step": 42 }, { "epoch": 0.0015574067366896053, "grad_norm": 4.020239323714708, "learning_rate": 1.0373944511459591e-06, "loss": 1.2079, "step": 43 }, { "epoch": 0.0015936254980079682, "grad_norm": 4.351221478726818, "learning_rate": 1.0615199034981907e-06, "loss": 1.1629, "step": 44 }, { "epoch": 0.0016298442593263311, "grad_norm": 4.182154146865418, "learning_rate": 1.0856453558504223e-06, "loss": 1.2015, "step": 45 }, { "epoch": 0.001666063020644694, "grad_norm": 4.536643087637395, "learning_rate": 1.1097708082026538e-06, "loss": 1.1778, "step": 46 }, { "epoch": 0.001702281781963057, "grad_norm": 4.388129668956599, "learning_rate": 1.1338962605548854e-06, "loss": 1.1412, "step": 47 }, { "epoch": 0.0017385005432814199, "grad_norm": 4.149159147782238, "learning_rate": 1.1580217129071172e-06, "loss": 1.1416, "step": 48 }, { "epoch": 0.0017747193045997826, "grad_norm": 4.0879016334760925, "learning_rate": 1.1821471652593488e-06, "loss": 1.0844, "step": 49 }, { "epoch": 0.0018109380659181455, "grad_norm": 4.968719233411626, "learning_rate": 1.2062726176115803e-06, "loss": 1.1792, "step": 50 }, { "epoch": 0.0018471568272365084, "grad_norm": 3.9464527031062673, "learning_rate": 1.2303980699638119e-06, "loss": 1.1536, "step": 51 }, { "epoch": 0.0018833755885548714, "grad_norm": 4.063558605624582, "learning_rate": 1.2545235223160435e-06, "loss": 1.1916, "step": 52 }, { "epoch": 0.0019195943498732343, "grad_norm": 3.957620706741628, "learning_rate": 1.278648974668275e-06, "loss": 1.153, "step": 53 }, { "epoch": 0.0019558131111915974, "grad_norm": 3.5583691740725008, "learning_rate": 1.3027744270205068e-06, "loss": 1.1512, "step": 54 }, { "epoch": 0.00199203187250996, "grad_norm": 3.4436034853345996, "learning_rate": 1.3268998793727384e-06, "loss": 1.0683, "step": 55 }, { "epoch": 0.0020282506338283233, "grad_norm": 3.2893023686596083, "learning_rate": 1.35102533172497e-06, "loss": 1.0582, "step": 56 }, { "epoch": 0.002064469395146686, "grad_norm": 2.4657443161385904, "learning_rate": 1.3751507840772015e-06, "loss": 0.8861, "step": 57 }, { "epoch": 0.002100688156465049, "grad_norm": 3.498870054136221, "learning_rate": 1.3992762364294333e-06, "loss": 1.1641, "step": 58 }, { "epoch": 0.002136906917783412, "grad_norm": 2.156137624823657, "learning_rate": 1.4234016887816649e-06, "loss": 0.8394, "step": 59 }, { "epoch": 0.0021731256791017745, "grad_norm": 3.437669918889141, "learning_rate": 1.4475271411338964e-06, "loss": 1.1493, "step": 60 }, { "epoch": 0.0022093444404201377, "grad_norm": 4.027102691658461, "learning_rate": 1.4716525934861282e-06, "loss": 1.1505, "step": 61 }, { "epoch": 0.0022455632017385004, "grad_norm": 3.2171426111306882, "learning_rate": 1.4957780458383598e-06, "loss": 1.1828, "step": 62 }, { "epoch": 0.0022817819630568635, "grad_norm": 3.2376751717930916, "learning_rate": 1.5199034981905911e-06, "loss": 1.1085, "step": 63 }, { "epoch": 0.002318000724375226, "grad_norm": 3.239743205272591, "learning_rate": 1.5440289505428227e-06, "loss": 1.0974, "step": 64 }, { "epoch": 0.0023542194856935894, "grad_norm": 3.57687022767256, "learning_rate": 1.5681544028950543e-06, "loss": 1.0451, "step": 65 }, { "epoch": 0.002390438247011952, "grad_norm": 3.248075217870574, "learning_rate": 1.592279855247286e-06, "loss": 1.129, "step": 66 }, { "epoch": 0.002426657008330315, "grad_norm": 3.7854264421744435, "learning_rate": 1.6164053075995176e-06, "loss": 1.1121, "step": 67 }, { "epoch": 0.002462875769648678, "grad_norm": 3.0659641554613413, "learning_rate": 1.6405307599517492e-06, "loss": 1.1363, "step": 68 }, { "epoch": 0.002499094530967041, "grad_norm": 4.258675208753407, "learning_rate": 1.6646562123039808e-06, "loss": 0.8653, "step": 69 }, { "epoch": 0.0025353132922854038, "grad_norm": 3.3060829201506547, "learning_rate": 1.6887816646562125e-06, "loss": 1.0668, "step": 70 }, { "epoch": 0.002571532053603767, "grad_norm": 3.289065526385512, "learning_rate": 1.712907117008444e-06, "loss": 1.1116, "step": 71 }, { "epoch": 0.0026077508149221296, "grad_norm": 3.2400138646582666, "learning_rate": 1.7370325693606757e-06, "loss": 1.0377, "step": 72 }, { "epoch": 0.0026439695762404927, "grad_norm": 3.4607359093366337, "learning_rate": 1.7611580217129072e-06, "loss": 1.0445, "step": 73 }, { "epoch": 0.0026801883375588554, "grad_norm": 3.1100575929472956, "learning_rate": 1.785283474065139e-06, "loss": 1.0864, "step": 74 }, { "epoch": 0.0027164070988772186, "grad_norm": 2.9829541364066765, "learning_rate": 1.8094089264173706e-06, "loss": 1.0951, "step": 75 }, { "epoch": 0.0027526258601955813, "grad_norm": 3.2148546193195204, "learning_rate": 1.833534378769602e-06, "loss": 1.0565, "step": 76 }, { "epoch": 0.0027888446215139444, "grad_norm": 2.9262967336769337, "learning_rate": 1.8576598311218335e-06, "loss": 1.0342, "step": 77 }, { "epoch": 0.002825063382832307, "grad_norm": 3.1338162847461715, "learning_rate": 1.8817852834740653e-06, "loss": 1.0416, "step": 78 }, { "epoch": 0.00286128214415067, "grad_norm": 3.3775316871311505, "learning_rate": 1.9059107358262969e-06, "loss": 1.0854, "step": 79 }, { "epoch": 0.002897500905469033, "grad_norm": 2.022703567785184, "learning_rate": 1.9300361881785284e-06, "loss": 0.8149, "step": 80 }, { "epoch": 0.0029337196667873957, "grad_norm": 3.496087007143135, "learning_rate": 1.95416164053076e-06, "loss": 1.0573, "step": 81 }, { "epoch": 0.002969938428105759, "grad_norm": 2.0768172560986393, "learning_rate": 1.9782870928829916e-06, "loss": 0.811, "step": 82 }, { "epoch": 0.0030061571894241215, "grad_norm": 3.180136501281112, "learning_rate": 2.0024125452352233e-06, "loss": 1.0812, "step": 83 }, { "epoch": 0.0030423759507424847, "grad_norm": 3.390976984238708, "learning_rate": 2.026537997587455e-06, "loss": 1.0415, "step": 84 }, { "epoch": 0.0030785947120608474, "grad_norm": 2.922818934102044, "learning_rate": 2.0506634499396865e-06, "loss": 1.0533, "step": 85 }, { "epoch": 0.0031148134733792105, "grad_norm": 3.020281719400221, "learning_rate": 2.0747889022919183e-06, "loss": 1.0025, "step": 86 }, { "epoch": 0.0031510322346975732, "grad_norm": 3.3838892670033633, "learning_rate": 2.09891435464415e-06, "loss": 1.0359, "step": 87 }, { "epoch": 0.0031872509960159364, "grad_norm": 1.9989472374930943, "learning_rate": 2.1230398069963814e-06, "loss": 0.8612, "step": 88 }, { "epoch": 0.003223469757334299, "grad_norm": 3.2143204217883725, "learning_rate": 2.147165259348613e-06, "loss": 0.9965, "step": 89 }, { "epoch": 0.0032596885186526622, "grad_norm": 2.824657627685881, "learning_rate": 2.1712907117008445e-06, "loss": 1.0514, "step": 90 }, { "epoch": 0.003295907279971025, "grad_norm": 1.811127763783504, "learning_rate": 2.195416164053076e-06, "loss": 0.8449, "step": 91 }, { "epoch": 0.003332126041289388, "grad_norm": 3.1101954853656073, "learning_rate": 2.2195416164053077e-06, "loss": 0.9979, "step": 92 }, { "epoch": 0.0033683448026077508, "grad_norm": 2.1364119443376905, "learning_rate": 2.2436670687575394e-06, "loss": 0.9028, "step": 93 }, { "epoch": 0.003404563563926114, "grad_norm": 2.987703140287437, "learning_rate": 2.267792521109771e-06, "loss": 0.9801, "step": 94 }, { "epoch": 0.0034407823252444766, "grad_norm": 3.339244873607276, "learning_rate": 2.2919179734620026e-06, "loss": 1.0066, "step": 95 }, { "epoch": 0.0034770010865628398, "grad_norm": 3.510510169839411, "learning_rate": 2.3160434258142344e-06, "loss": 1.074, "step": 96 }, { "epoch": 0.0035132198478812025, "grad_norm": 3.002704010213538, "learning_rate": 2.3401688781664657e-06, "loss": 1.0904, "step": 97 }, { "epoch": 0.003549438609199565, "grad_norm": 2.985438202414044, "learning_rate": 2.3642943305186975e-06, "loss": 1.0144, "step": 98 }, { "epoch": 0.0035856573705179283, "grad_norm": 3.2118537365962423, "learning_rate": 2.3884197828709293e-06, "loss": 1.0081, "step": 99 }, { "epoch": 0.003621876131836291, "grad_norm": 2.971949103481469, "learning_rate": 2.4125452352231606e-06, "loss": 1.0382, "step": 100 }, { "epoch": 0.003658094893154654, "grad_norm": 2.932996742464904, "learning_rate": 2.4366706875753924e-06, "loss": 1.0378, "step": 101 }, { "epoch": 0.003694313654473017, "grad_norm": 3.1657653229577654, "learning_rate": 2.4607961399276238e-06, "loss": 0.9654, "step": 102 }, { "epoch": 0.00373053241579138, "grad_norm": 3.621715801628598, "learning_rate": 2.4849215922798556e-06, "loss": 0.874, "step": 103 }, { "epoch": 0.0037667511771097427, "grad_norm": 2.872378514682348, "learning_rate": 2.509047044632087e-06, "loss": 1.0677, "step": 104 }, { "epoch": 0.003802969938428106, "grad_norm": 3.0933863493538953, "learning_rate": 2.5331724969843187e-06, "loss": 1.0108, "step": 105 }, { "epoch": 0.0038391886997464686, "grad_norm": 1.9941667236073675, "learning_rate": 2.55729794933655e-06, "loss": 0.8624, "step": 106 }, { "epoch": 0.0038754074610648317, "grad_norm": 2.842298519782687, "learning_rate": 2.581423401688782e-06, "loss": 0.9594, "step": 107 }, { "epoch": 0.003911626222383195, "grad_norm": 3.0338429735929546, "learning_rate": 2.6055488540410136e-06, "loss": 0.9161, "step": 108 }, { "epoch": 0.0039478449837015575, "grad_norm": 1.8806225707074342, "learning_rate": 2.629674306393245e-06, "loss": 0.8253, "step": 109 }, { "epoch": 0.00398406374501992, "grad_norm": 2.877756559499516, "learning_rate": 2.6537997587454767e-06, "loss": 0.9445, "step": 110 }, { "epoch": 0.004020282506338283, "grad_norm": 2.7644288606806438, "learning_rate": 2.677925211097708e-06, "loss": 1.0271, "step": 111 }, { "epoch": 0.0040565012676566465, "grad_norm": 2.8576234732047645, "learning_rate": 2.70205066344994e-06, "loss": 1.0004, "step": 112 }, { "epoch": 0.004092720028975009, "grad_norm": 1.8922324809792588, "learning_rate": 2.7261761158021717e-06, "loss": 0.9035, "step": 113 }, { "epoch": 0.004128938790293372, "grad_norm": 2.9316295595614656, "learning_rate": 2.750301568154403e-06, "loss": 0.9605, "step": 114 }, { "epoch": 0.004165157551611735, "grad_norm": 1.8807656891436901, "learning_rate": 2.774427020506635e-06, "loss": 0.8121, "step": 115 }, { "epoch": 0.004201376312930098, "grad_norm": 2.8918575597401643, "learning_rate": 2.7985524728588666e-06, "loss": 1.0518, "step": 116 }, { "epoch": 0.004237595074248461, "grad_norm": 2.9225689171104525, "learning_rate": 2.822677925211098e-06, "loss": 1.0114, "step": 117 }, { "epoch": 0.004273813835566824, "grad_norm": 3.217073003687038, "learning_rate": 2.8468033775633297e-06, "loss": 1.0218, "step": 118 }, { "epoch": 0.004310032596885186, "grad_norm": 3.0024410786910045, "learning_rate": 2.8709288299155615e-06, "loss": 1.0561, "step": 119 }, { "epoch": 0.004346251358203549, "grad_norm": 1.7187645749671636, "learning_rate": 2.895054282267793e-06, "loss": 0.8679, "step": 120 }, { "epoch": 0.004382470119521913, "grad_norm": 2.9641119263609657, "learning_rate": 2.9191797346200246e-06, "loss": 1.0114, "step": 121 }, { "epoch": 0.004418688880840275, "grad_norm": 1.6934636751038767, "learning_rate": 2.9433051869722564e-06, "loss": 0.8032, "step": 122 }, { "epoch": 0.004454907642158638, "grad_norm": 1.5373882326431874, "learning_rate": 2.9674306393244878e-06, "loss": 0.8415, "step": 123 }, { "epoch": 0.004491126403477001, "grad_norm": 3.3465889617047573, "learning_rate": 2.9915560916767195e-06, "loss": 1.049, "step": 124 }, { "epoch": 0.004527345164795364, "grad_norm": 2.958017217621786, "learning_rate": 3.0156815440289505e-06, "loss": 1.0921, "step": 125 }, { "epoch": 0.004563563926113727, "grad_norm": 3.1033970958433077, "learning_rate": 3.0398069963811823e-06, "loss": 0.9221, "step": 126 }, { "epoch": 0.00459978268743209, "grad_norm": 2.742224727178221, "learning_rate": 3.0639324487334136e-06, "loss": 0.828, "step": 127 }, { "epoch": 0.004636001448750452, "grad_norm": 1.706445263724748, "learning_rate": 3.0880579010856454e-06, "loss": 0.836, "step": 128 }, { "epoch": 0.004672220210068816, "grad_norm": 2.607788294502132, "learning_rate": 3.112183353437877e-06, "loss": 0.9798, "step": 129 }, { "epoch": 0.004708438971387179, "grad_norm": 2.9467528031661834, "learning_rate": 3.1363088057901085e-06, "loss": 1.0009, "step": 130 }, { "epoch": 0.004744657732705541, "grad_norm": 2.8383446164227966, "learning_rate": 3.1604342581423403e-06, "loss": 1.0389, "step": 131 }, { "epoch": 0.004780876494023904, "grad_norm": 3.6248099795227375, "learning_rate": 3.184559710494572e-06, "loss": 0.9673, "step": 132 }, { "epoch": 0.004817095255342268, "grad_norm": 3.101477941326075, "learning_rate": 3.2086851628468034e-06, "loss": 1.0414, "step": 133 }, { "epoch": 0.00485331401666063, "grad_norm": 2.7090379012471004, "learning_rate": 3.2328106151990352e-06, "loss": 0.9998, "step": 134 }, { "epoch": 0.004889532777978993, "grad_norm": 2.8916726662429735, "learning_rate": 3.2569360675512666e-06, "loss": 1.0091, "step": 135 }, { "epoch": 0.004925751539297356, "grad_norm": 3.2045970790536655, "learning_rate": 3.2810615199034984e-06, "loss": 0.9439, "step": 136 }, { "epoch": 0.0049619703006157185, "grad_norm": 2.7612334287646116, "learning_rate": 3.30518697225573e-06, "loss": 0.948, "step": 137 }, { "epoch": 0.004998189061934082, "grad_norm": 3.1271236904500035, "learning_rate": 3.3293124246079615e-06, "loss": 1.0155, "step": 138 }, { "epoch": 0.005034407823252445, "grad_norm": 2.7062979691727325, "learning_rate": 3.3534378769601933e-06, "loss": 0.9446, "step": 139 }, { "epoch": 0.0050706265845708075, "grad_norm": 2.152259711651698, "learning_rate": 3.377563329312425e-06, "loss": 0.804, "step": 140 }, { "epoch": 0.00510684534588917, "grad_norm": 2.9541059307980353, "learning_rate": 3.4016887816646564e-06, "loss": 0.9555, "step": 141 }, { "epoch": 0.005143064107207534, "grad_norm": 1.980707730808823, "learning_rate": 3.425814234016888e-06, "loss": 0.7899, "step": 142 }, { "epoch": 0.0051792828685258965, "grad_norm": 3.7843454317281267, "learning_rate": 3.44993968636912e-06, "loss": 1.0345, "step": 143 }, { "epoch": 0.005215501629844259, "grad_norm": 2.900358175487933, "learning_rate": 3.4740651387213513e-06, "loss": 0.9507, "step": 144 }, { "epoch": 0.005251720391162622, "grad_norm": 1.6448769727715489, "learning_rate": 3.498190591073583e-06, "loss": 0.7584, "step": 145 }, { "epoch": 0.0052879391524809855, "grad_norm": 3.2351297474641294, "learning_rate": 3.5223160434258145e-06, "loss": 0.9394, "step": 146 }, { "epoch": 0.005324157913799348, "grad_norm": 1.9083733278982093, "learning_rate": 3.5464414957780463e-06, "loss": 0.8075, "step": 147 }, { "epoch": 0.005360376675117711, "grad_norm": 3.107236030627011, "learning_rate": 3.570566948130278e-06, "loss": 0.9421, "step": 148 }, { "epoch": 0.005396595436436074, "grad_norm": 2.767834330499927, "learning_rate": 3.5946924004825094e-06, "loss": 0.9673, "step": 149 }, { "epoch": 0.005432814197754437, "grad_norm": 3.0084663175916657, "learning_rate": 3.618817852834741e-06, "loss": 0.9654, "step": 150 }, { "epoch": 0.0054690329590728, "grad_norm": 2.902551977253228, "learning_rate": 3.642943305186973e-06, "loss": 1.0387, "step": 151 }, { "epoch": 0.005505251720391163, "grad_norm": 2.8061580442175207, "learning_rate": 3.667068757539204e-06, "loss": 0.9273, "step": 152 }, { "epoch": 0.005541470481709525, "grad_norm": 3.8274185073693516, "learning_rate": 3.6911942098914357e-06, "loss": 0.9877, "step": 153 }, { "epoch": 0.005577689243027889, "grad_norm": 2.791411690075429, "learning_rate": 3.715319662243667e-06, "loss": 0.9745, "step": 154 }, { "epoch": 0.005613908004346252, "grad_norm": 2.292485513366739, "learning_rate": 3.739445114595899e-06, "loss": 0.7237, "step": 155 }, { "epoch": 0.005650126765664614, "grad_norm": 2.8168093912631877, "learning_rate": 3.7635705669481306e-06, "loss": 1.0052, "step": 156 }, { "epoch": 0.005686345526982977, "grad_norm": 1.8827093507184105, "learning_rate": 3.787696019300362e-06, "loss": 0.7904, "step": 157 }, { "epoch": 0.00572256428830134, "grad_norm": 3.741449016423006, "learning_rate": 3.8118214716525937e-06, "loss": 0.89, "step": 158 }, { "epoch": 0.005758783049619703, "grad_norm": 3.257193424216172, "learning_rate": 3.835946924004825e-06, "loss": 0.9865, "step": 159 }, { "epoch": 0.005795001810938066, "grad_norm": 3.414529118195751, "learning_rate": 3.860072376357057e-06, "loss": 0.9887, "step": 160 }, { "epoch": 0.005831220572256429, "grad_norm": 2.976120342745276, "learning_rate": 3.884197828709289e-06, "loss": 1.0042, "step": 161 }, { "epoch": 0.005867439333574791, "grad_norm": 3.2018832940108313, "learning_rate": 3.90832328106152e-06, "loss": 1.0151, "step": 162 }, { "epoch": 0.005903658094893155, "grad_norm": 1.6922733631391222, "learning_rate": 3.932448733413752e-06, "loss": 0.7779, "step": 163 }, { "epoch": 0.005939876856211518, "grad_norm": 2.892659216445377, "learning_rate": 3.956574185765983e-06, "loss": 0.9615, "step": 164 }, { "epoch": 0.00597609561752988, "grad_norm": 4.390243730926252, "learning_rate": 3.980699638118215e-06, "loss": 0.9962, "step": 165 }, { "epoch": 0.006012314378848243, "grad_norm": 2.8166717716293985, "learning_rate": 4.004825090470447e-06, "loss": 1.017, "step": 166 }, { "epoch": 0.006048533140166607, "grad_norm": 3.1146183445194895, "learning_rate": 4.0289505428226785e-06, "loss": 1.0001, "step": 167 }, { "epoch": 0.006084751901484969, "grad_norm": 1.5699248903440275, "learning_rate": 4.05307599517491e-06, "loss": 0.7658, "step": 168 }, { "epoch": 0.006120970662803332, "grad_norm": 2.6404534769499706, "learning_rate": 4.077201447527141e-06, "loss": 0.9398, "step": 169 }, { "epoch": 0.006157189424121695, "grad_norm": 2.313008207259899, "learning_rate": 4.101326899879373e-06, "loss": 0.8202, "step": 170 }, { "epoch": 0.006193408185440058, "grad_norm": 3.1577114177760186, "learning_rate": 4.125452352231605e-06, "loss": 1.0097, "step": 171 }, { "epoch": 0.006229626946758421, "grad_norm": 2.977730536880632, "learning_rate": 4.1495778045838365e-06, "loss": 0.8916, "step": 172 }, { "epoch": 0.006265845708076784, "grad_norm": 1.4602077571325789, "learning_rate": 4.173703256936068e-06, "loss": 0.7837, "step": 173 }, { "epoch": 0.0063020644693951465, "grad_norm": 2.7964194209120143, "learning_rate": 4.1978287092883e-06, "loss": 1.0286, "step": 174 }, { "epoch": 0.006338283230713509, "grad_norm": 1.747675872370386, "learning_rate": 4.221954161640531e-06, "loss": 0.7882, "step": 175 }, { "epoch": 0.006374501992031873, "grad_norm": 2.914133393189789, "learning_rate": 4.246079613992763e-06, "loss": 0.9716, "step": 176 }, { "epoch": 0.0064107207533502355, "grad_norm": 1.5618975472930656, "learning_rate": 4.2702050663449946e-06, "loss": 0.7807, "step": 177 }, { "epoch": 0.006446939514668598, "grad_norm": 3.064898218492085, "learning_rate": 4.294330518697226e-06, "loss": 0.9463, "step": 178 }, { "epoch": 0.006483158275986961, "grad_norm": 1.5348080840552207, "learning_rate": 4.318455971049458e-06, "loss": 0.6901, "step": 179 }, { "epoch": 0.0065193770373053244, "grad_norm": 1.5577046875383134, "learning_rate": 4.342581423401689e-06, "loss": 0.724, "step": 180 }, { "epoch": 0.006555595798623687, "grad_norm": 2.91261379752279, "learning_rate": 4.366706875753921e-06, "loss": 0.9748, "step": 181 }, { "epoch": 0.00659181455994205, "grad_norm": 2.924859740896595, "learning_rate": 4.390832328106152e-06, "loss": 0.9276, "step": 182 }, { "epoch": 0.0066280333212604126, "grad_norm": 3.0936683538047847, "learning_rate": 4.4149577804583836e-06, "loss": 0.9905, "step": 183 }, { "epoch": 0.006664252082578776, "grad_norm": 2.840673427777487, "learning_rate": 4.439083232810615e-06, "loss": 0.9657, "step": 184 }, { "epoch": 0.006700470843897139, "grad_norm": 2.528904904367244, "learning_rate": 4.463208685162847e-06, "loss": 0.9099, "step": 185 }, { "epoch": 0.0067366896052155015, "grad_norm": 2.9307616591608463, "learning_rate": 4.487334137515079e-06, "loss": 0.9701, "step": 186 }, { "epoch": 0.006772908366533864, "grad_norm": 2.565295548992646, "learning_rate": 4.511459589867311e-06, "loss": 0.8491, "step": 187 }, { "epoch": 0.006809127127852228, "grad_norm": 1.7615697177449077, "learning_rate": 4.535585042219542e-06, "loss": 0.7634, "step": 188 }, { "epoch": 0.0068453458891705905, "grad_norm": 1.5583265140895237, "learning_rate": 4.559710494571773e-06, "loss": 0.8014, "step": 189 }, { "epoch": 0.006881564650488953, "grad_norm": 1.883304583333267, "learning_rate": 4.583835946924005e-06, "loss": 0.8438, "step": 190 }, { "epoch": 0.006917783411807316, "grad_norm": 3.5992416550023596, "learning_rate": 4.607961399276237e-06, "loss": 0.936, "step": 191 }, { "epoch": 0.0069540021731256795, "grad_norm": 3.0466698598883544, "learning_rate": 4.632086851628469e-06, "loss": 0.8269, "step": 192 }, { "epoch": 0.006990220934444042, "grad_norm": 2.851886443458553, "learning_rate": 4.6562123039807e-06, "loss": 1.036, "step": 193 }, { "epoch": 0.007026439695762405, "grad_norm": 3.114612490965109, "learning_rate": 4.6803377563329314e-06, "loss": 0.9055, "step": 194 }, { "epoch": 0.007062658457080768, "grad_norm": 1.9197652262689715, "learning_rate": 4.704463208685163e-06, "loss": 0.8023, "step": 195 }, { "epoch": 0.00709887721839913, "grad_norm": 2.6203655846029044, "learning_rate": 4.728588661037395e-06, "loss": 1.0032, "step": 196 }, { "epoch": 0.007135095979717494, "grad_norm": 1.5930288990393118, "learning_rate": 4.752714113389627e-06, "loss": 0.7346, "step": 197 }, { "epoch": 0.007171314741035857, "grad_norm": 2.622047918639642, "learning_rate": 4.7768395657418586e-06, "loss": 0.8972, "step": 198 }, { "epoch": 0.007207533502354219, "grad_norm": 3.4688877903144846, "learning_rate": 4.8009650180940895e-06, "loss": 0.9383, "step": 199 }, { "epoch": 0.007243752263672582, "grad_norm": 3.218436253635674, "learning_rate": 4.825090470446321e-06, "loss": 0.9651, "step": 200 }, { "epoch": 0.007279971024990946, "grad_norm": 2.6602316018074106, "learning_rate": 4.849215922798553e-06, "loss": 0.9783, "step": 201 }, { "epoch": 0.007316189786309308, "grad_norm": 2.7317248989979275, "learning_rate": 4.873341375150785e-06, "loss": 0.9389, "step": 202 }, { "epoch": 0.007352408547627671, "grad_norm": 2.806079189819918, "learning_rate": 4.897466827503017e-06, "loss": 0.9283, "step": 203 }, { "epoch": 0.007388627308946034, "grad_norm": 2.2399445477641704, "learning_rate": 4.9215922798552475e-06, "loss": 0.7867, "step": 204 }, { "epoch": 0.007424846070264397, "grad_norm": 2.246150224017164, "learning_rate": 4.945717732207479e-06, "loss": 0.7955, "step": 205 }, { "epoch": 0.00746106483158276, "grad_norm": 2.7409373446851464, "learning_rate": 4.969843184559711e-06, "loss": 0.9328, "step": 206 }, { "epoch": 0.007497283592901123, "grad_norm": 1.493791798586078, "learning_rate": 4.993968636911943e-06, "loss": 0.8194, "step": 207 }, { "epoch": 0.007533502354219485, "grad_norm": 1.563031756545117, "learning_rate": 5.018094089264174e-06, "loss": 0.7701, "step": 208 }, { "epoch": 0.007569721115537849, "grad_norm": 3.02662618685287, "learning_rate": 5.042219541616406e-06, "loss": 0.9053, "step": 209 }, { "epoch": 0.007605939876856212, "grad_norm": 3.1210454285357936, "learning_rate": 5.066344993968637e-06, "loss": 0.9182, "step": 210 }, { "epoch": 0.007642158638174574, "grad_norm": 2.6856098930594485, "learning_rate": 5.090470446320868e-06, "loss": 0.9107, "step": 211 }, { "epoch": 0.007678377399492937, "grad_norm": 2.833224493745852, "learning_rate": 5.1145958986731e-06, "loss": 0.8809, "step": 212 }, { "epoch": 0.0077145961608113, "grad_norm": 1.682302172060152, "learning_rate": 5.138721351025332e-06, "loss": 0.7815, "step": 213 }, { "epoch": 0.007750814922129663, "grad_norm": 2.956931761324029, "learning_rate": 5.162846803377564e-06, "loss": 0.9702, "step": 214 }, { "epoch": 0.007787033683448026, "grad_norm": 2.736748205010636, "learning_rate": 5.1869722557297954e-06, "loss": 0.93, "step": 215 }, { "epoch": 0.00782325244476639, "grad_norm": 2.8301426318749816, "learning_rate": 5.211097708082027e-06, "loss": 0.9631, "step": 216 }, { "epoch": 0.007859471206084752, "grad_norm": 1.7304826049971214, "learning_rate": 5.235223160434258e-06, "loss": 0.7419, "step": 217 }, { "epoch": 0.007895689967403115, "grad_norm": 1.4424455655845208, "learning_rate": 5.25934861278649e-06, "loss": 0.6933, "step": 218 }, { "epoch": 0.007931908728721477, "grad_norm": 1.7172806841841133, "learning_rate": 5.283474065138722e-06, "loss": 0.7625, "step": 219 }, { "epoch": 0.00796812749003984, "grad_norm": 2.981645024517646, "learning_rate": 5.3075995174909535e-06, "loss": 0.9417, "step": 220 }, { "epoch": 0.008004346251358204, "grad_norm": 2.7923943202308643, "learning_rate": 5.331724969843185e-06, "loss": 0.9277, "step": 221 }, { "epoch": 0.008040565012676566, "grad_norm": 2.8592860750014046, "learning_rate": 5.355850422195416e-06, "loss": 0.9367, "step": 222 }, { "epoch": 0.00807678377399493, "grad_norm": 2.6702711020665246, "learning_rate": 5.379975874547648e-06, "loss": 0.9309, "step": 223 }, { "epoch": 0.008113002535313293, "grad_norm": 3.1771747140823168, "learning_rate": 5.40410132689988e-06, "loss": 1.0059, "step": 224 }, { "epoch": 0.008149221296631655, "grad_norm": 2.75291316205735, "learning_rate": 5.4282267792521115e-06, "loss": 0.9525, "step": 225 }, { "epoch": 0.008185440057950018, "grad_norm": 2.602167024913252, "learning_rate": 5.452352231604343e-06, "loss": 1.0143, "step": 226 }, { "epoch": 0.00822165881926838, "grad_norm": 2.5167154768185176, "learning_rate": 5.476477683956575e-06, "loss": 0.9276, "step": 227 }, { "epoch": 0.008257877580586744, "grad_norm": 1.6982123628424655, "learning_rate": 5.500603136308806e-06, "loss": 0.7316, "step": 228 }, { "epoch": 0.008294096341905107, "grad_norm": 2.820369986690574, "learning_rate": 5.524728588661038e-06, "loss": 1.0414, "step": 229 }, { "epoch": 0.00833031510322347, "grad_norm": 1.69505682293959, "learning_rate": 5.54885404101327e-06, "loss": 0.7336, "step": 230 }, { "epoch": 0.008366533864541833, "grad_norm": 2.589627678374588, "learning_rate": 5.572979493365501e-06, "loss": 0.992, "step": 231 }, { "epoch": 0.008402752625860196, "grad_norm": 2.9720298919265504, "learning_rate": 5.597104945717733e-06, "loss": 0.9135, "step": 232 }, { "epoch": 0.008438971387178558, "grad_norm": 1.6912906213649406, "learning_rate": 5.621230398069965e-06, "loss": 0.7292, "step": 233 }, { "epoch": 0.008475190148496922, "grad_norm": 3.676877047972471, "learning_rate": 5.645355850422196e-06, "loss": 0.9372, "step": 234 }, { "epoch": 0.008511408909815284, "grad_norm": 2.716010349377737, "learning_rate": 5.669481302774428e-06, "loss": 0.9505, "step": 235 }, { "epoch": 0.008547627671133647, "grad_norm": 2.9667574620595443, "learning_rate": 5.6936067551266594e-06, "loss": 0.9666, "step": 236 }, { "epoch": 0.00858384643245201, "grad_norm": 2.8618235500931886, "learning_rate": 5.717732207478891e-06, "loss": 0.9146, "step": 237 }, { "epoch": 0.008620065193770373, "grad_norm": 2.756613536948661, "learning_rate": 5.741857659831123e-06, "loss": 0.942, "step": 238 }, { "epoch": 0.008656283955088736, "grad_norm": 2.7740474755483104, "learning_rate": 5.765983112183354e-06, "loss": 0.9168, "step": 239 }, { "epoch": 0.008692502716407098, "grad_norm": 2.982849650004277, "learning_rate": 5.790108564535586e-06, "loss": 0.9184, "step": 240 }, { "epoch": 0.008728721477725462, "grad_norm": 1.8460029209718842, "learning_rate": 5.8142340168878175e-06, "loss": 0.7339, "step": 241 }, { "epoch": 0.008764940239043825, "grad_norm": 1.6110738768262682, "learning_rate": 5.838359469240049e-06, "loss": 0.7524, "step": 242 }, { "epoch": 0.008801159000362187, "grad_norm": 1.6114887081670672, "learning_rate": 5.862484921592281e-06, "loss": 0.8052, "step": 243 }, { "epoch": 0.00883737776168055, "grad_norm": 2.947593822369202, "learning_rate": 5.886610373944513e-06, "loss": 0.9908, "step": 244 }, { "epoch": 0.008873596522998914, "grad_norm": 5.403944849833545, "learning_rate": 5.910735826296744e-06, "loss": 0.9737, "step": 245 }, { "epoch": 0.008909815284317276, "grad_norm": 1.6226798119037318, "learning_rate": 5.9348612786489755e-06, "loss": 0.7385, "step": 246 }, { "epoch": 0.00894603404563564, "grad_norm": 3.0195877198223013, "learning_rate": 5.958986731001207e-06, "loss": 0.9981, "step": 247 }, { "epoch": 0.008982252806954001, "grad_norm": 3.178915670572658, "learning_rate": 5.983112183353439e-06, "loss": 0.9616, "step": 248 }, { "epoch": 0.009018471568272365, "grad_norm": 2.7319727126960758, "learning_rate": 6.007237635705669e-06, "loss": 0.8955, "step": 249 }, { "epoch": 0.009054690329590729, "grad_norm": 2.6767887585481858, "learning_rate": 6.031363088057901e-06, "loss": 0.9618, "step": 250 }, { "epoch": 0.00909090909090909, "grad_norm": 3.137776508818226, "learning_rate": 6.055488540410133e-06, "loss": 0.9432, "step": 251 }, { "epoch": 0.009127127852227454, "grad_norm": 2.6749861475661603, "learning_rate": 6.0796139927623645e-06, "loss": 0.9569, "step": 252 }, { "epoch": 0.009163346613545816, "grad_norm": 2.799118798888511, "learning_rate": 6.103739445114596e-06, "loss": 0.901, "step": 253 }, { "epoch": 0.00919956537486418, "grad_norm": 2.9986912410770548, "learning_rate": 6.127864897466827e-06, "loss": 0.9541, "step": 254 }, { "epoch": 0.009235784136182543, "grad_norm": 2.6183811967075004, "learning_rate": 6.151990349819059e-06, "loss": 0.9709, "step": 255 }, { "epoch": 0.009272002897500905, "grad_norm": 2.5582838205636698, "learning_rate": 6.176115802171291e-06, "loss": 0.8563, "step": 256 }, { "epoch": 0.009308221658819268, "grad_norm": 2.495870095144383, "learning_rate": 6.2002412545235226e-06, "loss": 0.9283, "step": 257 }, { "epoch": 0.009344440420137632, "grad_norm": 2.7915793904615884, "learning_rate": 6.224366706875754e-06, "loss": 0.8618, "step": 258 }, { "epoch": 0.009380659181455994, "grad_norm": 2.03368836616756, "learning_rate": 6.248492159227985e-06, "loss": 0.7739, "step": 259 }, { "epoch": 0.009416877942774357, "grad_norm": 2.796347176705374, "learning_rate": 6.272617611580217e-06, "loss": 0.9753, "step": 260 }, { "epoch": 0.00945309670409272, "grad_norm": 2.452140413101782, "learning_rate": 6.296743063932449e-06, "loss": 0.9228, "step": 261 }, { "epoch": 0.009489315465411083, "grad_norm": 2.6525056224973738, "learning_rate": 6.320868516284681e-06, "loss": 0.9694, "step": 262 }, { "epoch": 0.009525534226729446, "grad_norm": 1.80930453127838, "learning_rate": 6.344993968636912e-06, "loss": 0.7989, "step": 263 }, { "epoch": 0.009561752988047808, "grad_norm": 2.5977638279716886, "learning_rate": 6.369119420989144e-06, "loss": 0.893, "step": 264 }, { "epoch": 0.009597971749366172, "grad_norm": 2.8604786025123663, "learning_rate": 6.393244873341375e-06, "loss": 0.9667, "step": 265 }, { "epoch": 0.009634190510684535, "grad_norm": 2.762282041596976, "learning_rate": 6.417370325693607e-06, "loss": 0.9627, "step": 266 }, { "epoch": 0.009670409272002897, "grad_norm": 2.69744121413657, "learning_rate": 6.441495778045839e-06, "loss": 0.9544, "step": 267 }, { "epoch": 0.00970662803332126, "grad_norm": 2.8166781940990835, "learning_rate": 6.4656212303980705e-06, "loss": 0.9705, "step": 268 }, { "epoch": 0.009742846794639623, "grad_norm": 3.1308806544179846, "learning_rate": 6.489746682750302e-06, "loss": 0.9336, "step": 269 }, { "epoch": 0.009779065555957986, "grad_norm": 2.6816274999643244, "learning_rate": 6.513872135102533e-06, "loss": 0.9795, "step": 270 }, { "epoch": 0.00981528431727635, "grad_norm": 1.6203679117261847, "learning_rate": 6.537997587454765e-06, "loss": 0.7616, "step": 271 }, { "epoch": 0.009851503078594712, "grad_norm": 1.803784831546618, "learning_rate": 6.562123039806997e-06, "loss": 0.7726, "step": 272 }, { "epoch": 0.009887721839913075, "grad_norm": 1.9417775940570934, "learning_rate": 6.5862484921592285e-06, "loss": 0.7446, "step": 273 }, { "epoch": 0.009923940601231437, "grad_norm": 2.9183921557844665, "learning_rate": 6.61037394451146e-06, "loss": 0.9096, "step": 274 }, { "epoch": 0.0099601593625498, "grad_norm": 3.3101238869477583, "learning_rate": 6.634499396863692e-06, "loss": 0.9952, "step": 275 }, { "epoch": 0.009996378123868164, "grad_norm": 2.5102442362965824, "learning_rate": 6.658624849215923e-06, "loss": 0.8136, "step": 276 }, { "epoch": 0.010032596885186526, "grad_norm": 2.7507128929392275, "learning_rate": 6.682750301568155e-06, "loss": 0.9379, "step": 277 }, { "epoch": 0.01006881564650489, "grad_norm": 2.8344156159305616, "learning_rate": 6.7068757539203866e-06, "loss": 0.8949, "step": 278 }, { "epoch": 0.010105034407823253, "grad_norm": 2.7224708231281176, "learning_rate": 6.731001206272618e-06, "loss": 0.9557, "step": 279 }, { "epoch": 0.010141253169141615, "grad_norm": 2.875313627717326, "learning_rate": 6.75512665862485e-06, "loss": 0.8618, "step": 280 }, { "epoch": 0.010177471930459979, "grad_norm": 2.7468003195591284, "learning_rate": 6.779252110977081e-06, "loss": 0.9823, "step": 281 }, { "epoch": 0.01021369069177834, "grad_norm": 2.6478937386974, "learning_rate": 6.803377563329313e-06, "loss": 0.8608, "step": 282 }, { "epoch": 0.010249909453096704, "grad_norm": 2.2489548060653592, "learning_rate": 6.827503015681545e-06, "loss": 0.7896, "step": 283 }, { "epoch": 0.010286128214415068, "grad_norm": 2.5859896922534853, "learning_rate": 6.851628468033776e-06, "loss": 0.8769, "step": 284 }, { "epoch": 0.01032234697573343, "grad_norm": 2.9353747060392483, "learning_rate": 6.875753920386008e-06, "loss": 0.9435, "step": 285 }, { "epoch": 0.010358565737051793, "grad_norm": 3.2206221941049775, "learning_rate": 6.89987937273824e-06, "loss": 1.0452, "step": 286 }, { "epoch": 0.010394784498370157, "grad_norm": 2.8412622742024847, "learning_rate": 6.924004825090471e-06, "loss": 0.9208, "step": 287 }, { "epoch": 0.010431003259688518, "grad_norm": 2.4606582907066303, "learning_rate": 6.948130277442703e-06, "loss": 0.9174, "step": 288 }, { "epoch": 0.010467222021006882, "grad_norm": 2.6551840550712105, "learning_rate": 6.9722557297949344e-06, "loss": 0.8586, "step": 289 }, { "epoch": 0.010503440782325244, "grad_norm": 1.5748978901313622, "learning_rate": 6.996381182147166e-06, "loss": 0.814, "step": 290 }, { "epoch": 0.010539659543643607, "grad_norm": 2.802456640644846, "learning_rate": 7.020506634499398e-06, "loss": 0.9387, "step": 291 }, { "epoch": 0.010575878304961971, "grad_norm": 2.549630321675551, "learning_rate": 7.044632086851629e-06, "loss": 0.91, "step": 292 }, { "epoch": 0.010612097066280333, "grad_norm": 3.00973309716552, "learning_rate": 7.068757539203861e-06, "loss": 0.8602, "step": 293 }, { "epoch": 0.010648315827598696, "grad_norm": 3.157483877120456, "learning_rate": 7.0928829915560925e-06, "loss": 0.9756, "step": 294 }, { "epoch": 0.010684534588917058, "grad_norm": 2.5351953312428646, "learning_rate": 7.117008443908324e-06, "loss": 0.9962, "step": 295 }, { "epoch": 0.010720753350235422, "grad_norm": 2.4732983615704325, "learning_rate": 7.141133896260556e-06, "loss": 0.9597, "step": 296 }, { "epoch": 0.010756972111553785, "grad_norm": 2.428840517177305, "learning_rate": 7.165259348612788e-06, "loss": 0.8584, "step": 297 }, { "epoch": 0.010793190872872147, "grad_norm": 1.7104478470429003, "learning_rate": 7.189384800965019e-06, "loss": 0.8121, "step": 298 }, { "epoch": 0.01082940963419051, "grad_norm": 2.7290835980641943, "learning_rate": 7.2135102533172506e-06, "loss": 1.0114, "step": 299 }, { "epoch": 0.010865628395508874, "grad_norm": 3.061831226117732, "learning_rate": 7.237635705669482e-06, "loss": 0.8591, "step": 300 }, { "epoch": 0.010901847156827236, "grad_norm": 2.801653179609577, "learning_rate": 7.261761158021714e-06, "loss": 0.9521, "step": 301 }, { "epoch": 0.0109380659181456, "grad_norm": 2.6534621531089613, "learning_rate": 7.285886610373946e-06, "loss": 0.8828, "step": 302 }, { "epoch": 0.010974284679463962, "grad_norm": 2.7403466982670057, "learning_rate": 7.310012062726177e-06, "loss": 0.9726, "step": 303 }, { "epoch": 0.011010503440782325, "grad_norm": 3.4009923429938524, "learning_rate": 7.334137515078408e-06, "loss": 0.909, "step": 304 }, { "epoch": 0.011046722202100689, "grad_norm": 1.648033382244578, "learning_rate": 7.3582629674306395e-06, "loss": 0.7374, "step": 305 }, { "epoch": 0.01108294096341905, "grad_norm": 2.4919362614193177, "learning_rate": 7.382388419782871e-06, "loss": 0.8402, "step": 306 }, { "epoch": 0.011119159724737414, "grad_norm": 3.0601737568490615, "learning_rate": 7.406513872135102e-06, "loss": 0.9419, "step": 307 }, { "epoch": 0.011155378486055778, "grad_norm": 2.5240681807895786, "learning_rate": 7.430639324487334e-06, "loss": 0.9508, "step": 308 }, { "epoch": 0.01119159724737414, "grad_norm": 2.6511443885782398, "learning_rate": 7.454764776839566e-06, "loss": 0.9138, "step": 309 }, { "epoch": 0.011227816008692503, "grad_norm": 2.522016876807927, "learning_rate": 7.478890229191798e-06, "loss": 0.9469, "step": 310 }, { "epoch": 0.011264034770010865, "grad_norm": 3.0513571856174315, "learning_rate": 7.503015681544029e-06, "loss": 0.908, "step": 311 }, { "epoch": 0.011300253531329229, "grad_norm": 2.695062570455965, "learning_rate": 7.527141133896261e-06, "loss": 0.8931, "step": 312 }, { "epoch": 0.011336472292647592, "grad_norm": 2.5385441397682404, "learning_rate": 7.551266586248492e-06, "loss": 0.9111, "step": 313 }, { "epoch": 0.011372691053965954, "grad_norm": 2.726352158879487, "learning_rate": 7.575392038600724e-06, "loss": 0.947, "step": 314 }, { "epoch": 0.011408909815284318, "grad_norm": 2.5542670846016566, "learning_rate": 7.599517490952956e-06, "loss": 0.9088, "step": 315 }, { "epoch": 0.01144512857660268, "grad_norm": 2.776319481234845, "learning_rate": 7.6236429433051874e-06, "loss": 0.9039, "step": 316 }, { "epoch": 0.011481347337921043, "grad_norm": 2.769512905495907, "learning_rate": 7.647768395657418e-06, "loss": 0.9547, "step": 317 }, { "epoch": 0.011517566099239407, "grad_norm": 2.6374923249444886, "learning_rate": 7.67189384800965e-06, "loss": 0.9046, "step": 318 }, { "epoch": 0.011553784860557768, "grad_norm": 2.6701096731692244, "learning_rate": 7.696019300361882e-06, "loss": 0.9718, "step": 319 }, { "epoch": 0.011590003621876132, "grad_norm": 2.471188935476657, "learning_rate": 7.720144752714114e-06, "loss": 0.9269, "step": 320 }, { "epoch": 0.011626222383194496, "grad_norm": 2.81618852474819, "learning_rate": 7.744270205066345e-06, "loss": 0.8029, "step": 321 }, { "epoch": 0.011662441144512857, "grad_norm": 2.8035964783012814, "learning_rate": 7.768395657418577e-06, "loss": 0.9904, "step": 322 }, { "epoch": 0.011698659905831221, "grad_norm": 1.7317966185065696, "learning_rate": 7.792521109770809e-06, "loss": 0.7844, "step": 323 }, { "epoch": 0.011734878667149583, "grad_norm": 2.799958926350901, "learning_rate": 7.81664656212304e-06, "loss": 0.9111, "step": 324 }, { "epoch": 0.011771097428467946, "grad_norm": 2.675905238710729, "learning_rate": 7.840772014475273e-06, "loss": 0.9069, "step": 325 }, { "epoch": 0.01180731618978631, "grad_norm": 2.9191977684321775, "learning_rate": 7.864897466827504e-06, "loss": 0.9603, "step": 326 }, { "epoch": 0.011843534951104672, "grad_norm": 2.6761731438556513, "learning_rate": 7.889022919179734e-06, "loss": 0.9376, "step": 327 }, { "epoch": 0.011879753712423035, "grad_norm": 2.706445177772939, "learning_rate": 7.913148371531966e-06, "loss": 0.9347, "step": 328 }, { "epoch": 0.011915972473741399, "grad_norm": 2.727684653576181, "learning_rate": 7.937273823884198e-06, "loss": 0.8303, "step": 329 }, { "epoch": 0.01195219123505976, "grad_norm": 2.6480226388720562, "learning_rate": 7.96139927623643e-06, "loss": 0.8826, "step": 330 }, { "epoch": 0.011988409996378124, "grad_norm": 2.5704094338038317, "learning_rate": 7.985524728588662e-06, "loss": 0.9017, "step": 331 }, { "epoch": 0.012024628757696486, "grad_norm": 2.469514364808141, "learning_rate": 8.009650180940893e-06, "loss": 0.9001, "step": 332 }, { "epoch": 0.01206084751901485, "grad_norm": 1.6678284430007635, "learning_rate": 8.033775633293125e-06, "loss": 0.7332, "step": 333 }, { "epoch": 0.012097066280333213, "grad_norm": 2.6317676938751196, "learning_rate": 8.057901085645357e-06, "loss": 0.8941, "step": 334 }, { "epoch": 0.012133285041651575, "grad_norm": 2.381558383724858, "learning_rate": 8.082026537997589e-06, "loss": 0.8759, "step": 335 }, { "epoch": 0.012169503802969939, "grad_norm": 1.5856316884846908, "learning_rate": 8.10615199034982e-06, "loss": 0.771, "step": 336 }, { "epoch": 0.0122057225642883, "grad_norm": 2.7638849452617333, "learning_rate": 8.130277442702052e-06, "loss": 0.936, "step": 337 }, { "epoch": 0.012241941325606664, "grad_norm": 2.802048451030444, "learning_rate": 8.154402895054282e-06, "loss": 0.9017, "step": 338 }, { "epoch": 0.012278160086925028, "grad_norm": 2.8021908036909458, "learning_rate": 8.178528347406514e-06, "loss": 0.9132, "step": 339 }, { "epoch": 0.01231437884824339, "grad_norm": 2.4020765822801784, "learning_rate": 8.202653799758746e-06, "loss": 0.9241, "step": 340 }, { "epoch": 0.012350597609561753, "grad_norm": 2.3931747689081972, "learning_rate": 8.226779252110978e-06, "loss": 0.8573, "step": 341 }, { "epoch": 0.012386816370880117, "grad_norm": 2.662481041346019, "learning_rate": 8.25090470446321e-06, "loss": 0.9309, "step": 342 }, { "epoch": 0.012423035132198479, "grad_norm": 2.9110198717241693, "learning_rate": 8.275030156815441e-06, "loss": 0.9734, "step": 343 }, { "epoch": 0.012459253893516842, "grad_norm": 2.5197534330121165, "learning_rate": 8.299155609167673e-06, "loss": 0.8305, "step": 344 }, { "epoch": 0.012495472654835204, "grad_norm": 2.562135638523902, "learning_rate": 8.323281061519905e-06, "loss": 1.021, "step": 345 }, { "epoch": 0.012531691416153568, "grad_norm": 1.6094417467760629, "learning_rate": 8.347406513872137e-06, "loss": 0.7636, "step": 346 }, { "epoch": 0.012567910177471931, "grad_norm": 2.895125545594386, "learning_rate": 8.371531966224368e-06, "loss": 0.9248, "step": 347 }, { "epoch": 0.012604128938790293, "grad_norm": 2.7169848390081817, "learning_rate": 8.3956574185766e-06, "loss": 0.879, "step": 348 }, { "epoch": 0.012640347700108657, "grad_norm": 1.5105651622756047, "learning_rate": 8.41978287092883e-06, "loss": 0.7586, "step": 349 }, { "epoch": 0.012676566461427018, "grad_norm": 2.479263481418906, "learning_rate": 8.443908323281062e-06, "loss": 0.8835, "step": 350 }, { "epoch": 0.012712785222745382, "grad_norm": 2.618946739604344, "learning_rate": 8.468033775633294e-06, "loss": 0.8461, "step": 351 }, { "epoch": 0.012749003984063745, "grad_norm": 1.8141738651805133, "learning_rate": 8.492159227985526e-06, "loss": 0.785, "step": 352 }, { "epoch": 0.012785222745382107, "grad_norm": 2.714656905735969, "learning_rate": 8.516284680337757e-06, "loss": 0.9056, "step": 353 }, { "epoch": 0.012821441506700471, "grad_norm": 2.5570013463515067, "learning_rate": 8.540410132689989e-06, "loss": 0.913, "step": 354 }, { "epoch": 0.012857660268018834, "grad_norm": 2.6890323506448914, "learning_rate": 8.564535585042221e-06, "loss": 0.9308, "step": 355 }, { "epoch": 0.012893879029337196, "grad_norm": 1.7003031287827206, "learning_rate": 8.588661037394453e-06, "loss": 0.7782, "step": 356 }, { "epoch": 0.01293009779065556, "grad_norm": 1.9386644974946157, "learning_rate": 8.612786489746684e-06, "loss": 0.7069, "step": 357 }, { "epoch": 0.012966316551973922, "grad_norm": 2.959619497996779, "learning_rate": 8.636911942098916e-06, "loss": 0.9336, "step": 358 }, { "epoch": 0.013002535313292285, "grad_norm": 1.7858786737810102, "learning_rate": 8.661037394451148e-06, "loss": 0.7623, "step": 359 }, { "epoch": 0.013038754074610649, "grad_norm": 3.0789173505457628, "learning_rate": 8.685162846803378e-06, "loss": 0.8741, "step": 360 }, { "epoch": 0.01307497283592901, "grad_norm": 2.7397696059446432, "learning_rate": 8.70928829915561e-06, "loss": 0.9418, "step": 361 }, { "epoch": 0.013111191597247374, "grad_norm": 2.753407536997807, "learning_rate": 8.733413751507842e-06, "loss": 0.7734, "step": 362 }, { "epoch": 0.013147410358565738, "grad_norm": 2.6959519857776004, "learning_rate": 8.757539203860072e-06, "loss": 0.885, "step": 363 }, { "epoch": 0.0131836291198841, "grad_norm": 2.833564227644287, "learning_rate": 8.781664656212304e-06, "loss": 0.9884, "step": 364 }, { "epoch": 0.013219847881202463, "grad_norm": 2.8237431338067607, "learning_rate": 8.805790108564535e-06, "loss": 0.8682, "step": 365 }, { "epoch": 0.013256066642520825, "grad_norm": 2.919888632084361, "learning_rate": 8.829915560916767e-06, "loss": 0.8886, "step": 366 }, { "epoch": 0.013292285403839189, "grad_norm": 1.6517563355163993, "learning_rate": 8.854041013268999e-06, "loss": 0.765, "step": 367 }, { "epoch": 0.013328504165157552, "grad_norm": 1.762712776805557, "learning_rate": 8.87816646562123e-06, "loss": 0.7599, "step": 368 }, { "epoch": 0.013364722926475914, "grad_norm": 2.6472009512161354, "learning_rate": 8.902291917973462e-06, "loss": 0.8819, "step": 369 }, { "epoch": 0.013400941687794278, "grad_norm": 2.12765950730323, "learning_rate": 8.926417370325694e-06, "loss": 0.7388, "step": 370 }, { "epoch": 0.01343716044911264, "grad_norm": 2.765290951355494, "learning_rate": 8.950542822677926e-06, "loss": 0.9041, "step": 371 }, { "epoch": 0.013473379210431003, "grad_norm": 3.0767753566187745, "learning_rate": 8.974668275030158e-06, "loss": 0.9598, "step": 372 }, { "epoch": 0.013509597971749367, "grad_norm": 1.5995469380252239, "learning_rate": 8.99879372738239e-06, "loss": 0.7501, "step": 373 }, { "epoch": 0.013545816733067729, "grad_norm": 1.655284483979649, "learning_rate": 9.022919179734621e-06, "loss": 0.7592, "step": 374 }, { "epoch": 0.013582035494386092, "grad_norm": 2.487345155612144, "learning_rate": 9.047044632086851e-06, "loss": 0.8673, "step": 375 }, { "epoch": 0.013618254255704456, "grad_norm": 2.6899982833457985, "learning_rate": 9.071170084439083e-06, "loss": 0.8551, "step": 376 }, { "epoch": 0.013654473017022817, "grad_norm": 2.5697818138687234, "learning_rate": 9.095295536791315e-06, "loss": 0.7993, "step": 377 }, { "epoch": 0.013690691778341181, "grad_norm": 2.2923861645689234, "learning_rate": 9.119420989143547e-06, "loss": 0.8434, "step": 378 }, { "epoch": 0.013726910539659543, "grad_norm": 2.563015077741197, "learning_rate": 9.143546441495779e-06, "loss": 0.8419, "step": 379 }, { "epoch": 0.013763129300977906, "grad_norm": 2.699916478913867, "learning_rate": 9.16767189384801e-06, "loss": 0.873, "step": 380 }, { "epoch": 0.01379934806229627, "grad_norm": 2.5659014874626793, "learning_rate": 9.191797346200242e-06, "loss": 0.9838, "step": 381 }, { "epoch": 0.013835566823614632, "grad_norm": 9.019815433034891, "learning_rate": 9.215922798552474e-06, "loss": 0.9272, "step": 382 }, { "epoch": 0.013871785584932995, "grad_norm": 2.5878074834171754, "learning_rate": 9.240048250904706e-06, "loss": 0.845, "step": 383 }, { "epoch": 0.013908004346251359, "grad_norm": 2.822511887943449, "learning_rate": 9.264173703256937e-06, "loss": 0.9827, "step": 384 }, { "epoch": 0.013944223107569721, "grad_norm": 2.0444935982760755, "learning_rate": 9.28829915560917e-06, "loss": 0.7529, "step": 385 }, { "epoch": 0.013980441868888084, "grad_norm": 2.416536122460168, "learning_rate": 9.3124246079614e-06, "loss": 0.9158, "step": 386 }, { "epoch": 0.014016660630206446, "grad_norm": 2.4518547148051364, "learning_rate": 9.336550060313631e-06, "loss": 0.9633, "step": 387 }, { "epoch": 0.01405287939152481, "grad_norm": 2.5017243905373445, "learning_rate": 9.360675512665863e-06, "loss": 0.9047, "step": 388 }, { "epoch": 0.014089098152843173, "grad_norm": 1.8395647674683373, "learning_rate": 9.384800965018095e-06, "loss": 0.7918, "step": 389 }, { "epoch": 0.014125316914161535, "grad_norm": 2.647093806179767, "learning_rate": 9.408926417370326e-06, "loss": 0.8996, "step": 390 }, { "epoch": 0.014161535675479899, "grad_norm": 1.6584897427571226, "learning_rate": 9.433051869722558e-06, "loss": 0.7336, "step": 391 }, { "epoch": 0.01419775443679826, "grad_norm": 1.8708353685476458, "learning_rate": 9.45717732207479e-06, "loss": 0.7773, "step": 392 }, { "epoch": 0.014233973198116624, "grad_norm": 2.5973035300431055, "learning_rate": 9.481302774427022e-06, "loss": 0.8189, "step": 393 }, { "epoch": 0.014270191959434988, "grad_norm": 3.0838756771633498, "learning_rate": 9.505428226779254e-06, "loss": 0.9571, "step": 394 }, { "epoch": 0.01430641072075335, "grad_norm": 2.4481353660380587, "learning_rate": 9.529553679131485e-06, "loss": 0.9713, "step": 395 }, { "epoch": 0.014342629482071713, "grad_norm": 1.6833594401448022, "learning_rate": 9.553679131483717e-06, "loss": 0.8075, "step": 396 }, { "epoch": 0.014378848243390077, "grad_norm": 3.027237827539918, "learning_rate": 9.577804583835947e-06, "loss": 0.9029, "step": 397 }, { "epoch": 0.014415067004708439, "grad_norm": 2.955286193904834, "learning_rate": 9.601930036188179e-06, "loss": 0.8023, "step": 398 }, { "epoch": 0.014451285766026802, "grad_norm": 2.712775747724547, "learning_rate": 9.62605548854041e-06, "loss": 1.0042, "step": 399 }, { "epoch": 0.014487504527345164, "grad_norm": 2.34527338746335, "learning_rate": 9.650180940892643e-06, "loss": 0.9675, "step": 400 }, { "epoch": 0.014523723288663528, "grad_norm": 1.8312762306974437, "learning_rate": 9.674306393244874e-06, "loss": 0.7412, "step": 401 }, { "epoch": 0.014559942049981891, "grad_norm": 2.607320993593564, "learning_rate": 9.698431845597106e-06, "loss": 0.895, "step": 402 }, { "epoch": 0.014596160811300253, "grad_norm": 2.790301186782916, "learning_rate": 9.722557297949338e-06, "loss": 0.9722, "step": 403 }, { "epoch": 0.014632379572618617, "grad_norm": 2.567955524908397, "learning_rate": 9.74668275030157e-06, "loss": 0.9365, "step": 404 }, { "epoch": 0.01466859833393698, "grad_norm": 2.7594287124020043, "learning_rate": 9.770808202653801e-06, "loss": 0.9192, "step": 405 }, { "epoch": 0.014704817095255342, "grad_norm": 2.886062475203022, "learning_rate": 9.794933655006033e-06, "loss": 0.9496, "step": 406 }, { "epoch": 0.014741035856573706, "grad_norm": 2.7283682349883973, "learning_rate": 9.819059107358265e-06, "loss": 0.8808, "step": 407 }, { "epoch": 0.014777254617892067, "grad_norm": 2.921103508222125, "learning_rate": 9.843184559710495e-06, "loss": 0.9464, "step": 408 }, { "epoch": 0.014813473379210431, "grad_norm": 1.7214427412783804, "learning_rate": 9.867310012062727e-06, "loss": 0.7713, "step": 409 }, { "epoch": 0.014849692140528795, "grad_norm": 2.739629612134116, "learning_rate": 9.891435464414959e-06, "loss": 0.8709, "step": 410 }, { "epoch": 0.014885910901847156, "grad_norm": 2.3432692813612146, "learning_rate": 9.91556091676719e-06, "loss": 0.8715, "step": 411 }, { "epoch": 0.01492212966316552, "grad_norm": 2.6270787999605356, "learning_rate": 9.939686369119422e-06, "loss": 0.9198, "step": 412 }, { "epoch": 0.014958348424483882, "grad_norm": 2.4629209881089964, "learning_rate": 9.963811821471654e-06, "loss": 0.9355, "step": 413 }, { "epoch": 0.014994567185802245, "grad_norm": 2.8083017327420166, "learning_rate": 9.987937273823886e-06, "loss": 0.8728, "step": 414 }, { "epoch": 0.015030785947120609, "grad_norm": 1.8860116140898846, "learning_rate": 1.0012062726176118e-05, "loss": 0.8056, "step": 415 }, { "epoch": 0.01506700470843897, "grad_norm": 2.4975066506809886, "learning_rate": 1.0036188178528348e-05, "loss": 0.8464, "step": 416 }, { "epoch": 0.015103223469757334, "grad_norm": 1.3589461639484088, "learning_rate": 1.0060313630880581e-05, "loss": 0.7561, "step": 417 }, { "epoch": 0.015139442231075698, "grad_norm": 2.8602049175582613, "learning_rate": 1.0084439083232811e-05, "loss": 0.9309, "step": 418 }, { "epoch": 0.01517566099239406, "grad_norm": 2.741712849719844, "learning_rate": 1.0108564535585043e-05, "loss": 0.9554, "step": 419 }, { "epoch": 0.015211879753712423, "grad_norm": 1.6355676658249385, "learning_rate": 1.0132689987937275e-05, "loss": 0.7596, "step": 420 }, { "epoch": 0.015248098515030785, "grad_norm": 1.9350154644734134, "learning_rate": 1.0156815440289507e-05, "loss": 0.7702, "step": 421 }, { "epoch": 0.015284317276349149, "grad_norm": 1.647919908898695, "learning_rate": 1.0180940892641737e-05, "loss": 0.7631, "step": 422 }, { "epoch": 0.015320536037667512, "grad_norm": 2.7403965369794068, "learning_rate": 1.020506634499397e-05, "loss": 0.9632, "step": 423 }, { "epoch": 0.015356754798985874, "grad_norm": 2.7441893549491985, "learning_rate": 1.02291917973462e-05, "loss": 1.0115, "step": 424 }, { "epoch": 0.015392973560304238, "grad_norm": 2.4323280619009187, "learning_rate": 1.0253317249698434e-05, "loss": 0.9378, "step": 425 }, { "epoch": 0.0154291923216226, "grad_norm": 2.012135976283392, "learning_rate": 1.0277442702050664e-05, "loss": 0.779, "step": 426 }, { "epoch": 0.015465411082940963, "grad_norm": 2.6899280724339527, "learning_rate": 1.0301568154402897e-05, "loss": 0.9932, "step": 427 }, { "epoch": 0.015501629844259327, "grad_norm": 1.81938241150859, "learning_rate": 1.0325693606755127e-05, "loss": 0.7528, "step": 428 }, { "epoch": 0.015537848605577689, "grad_norm": 2.558856088715358, "learning_rate": 1.034981905910736e-05, "loss": 0.9637, "step": 429 }, { "epoch": 0.015574067366896052, "grad_norm": 2.768423286562547, "learning_rate": 1.0373944511459591e-05, "loss": 0.9831, "step": 430 }, { "epoch": 0.015610286128214416, "grad_norm": 2.6846140236624616, "learning_rate": 1.0398069963811823e-05, "loss": 1.0073, "step": 431 }, { "epoch": 0.01564650488953278, "grad_norm": 2.272669360902275, "learning_rate": 1.0422195416164054e-05, "loss": 0.9112, "step": 432 }, { "epoch": 0.01568272365085114, "grad_norm": 2.5069617870291365, "learning_rate": 1.0446320868516286e-05, "loss": 0.8309, "step": 433 }, { "epoch": 0.015718942412169503, "grad_norm": 2.9119823128748017, "learning_rate": 1.0470446320868516e-05, "loss": 0.8799, "step": 434 }, { "epoch": 0.015755161173487867, "grad_norm": 2.720861976625139, "learning_rate": 1.049457177322075e-05, "loss": 0.9253, "step": 435 }, { "epoch": 0.01579137993480623, "grad_norm": 2.417821586949762, "learning_rate": 1.051869722557298e-05, "loss": 0.9597, "step": 436 }, { "epoch": 0.015827598696124594, "grad_norm": 2.66607201485157, "learning_rate": 1.0542822677925213e-05, "loss": 0.9428, "step": 437 }, { "epoch": 0.015863817457442954, "grad_norm": 1.7098621424832912, "learning_rate": 1.0566948130277443e-05, "loss": 0.7591, "step": 438 }, { "epoch": 0.015900036218761317, "grad_norm": 2.8344521586673395, "learning_rate": 1.0591073582629677e-05, "loss": 0.9145, "step": 439 }, { "epoch": 0.01593625498007968, "grad_norm": 2.5617290505926054, "learning_rate": 1.0615199034981907e-05, "loss": 0.9342, "step": 440 }, { "epoch": 0.015972473741398045, "grad_norm": 2.8022421108815108, "learning_rate": 1.0639324487334139e-05, "loss": 0.9611, "step": 441 }, { "epoch": 0.016008692502716408, "grad_norm": 2.587514264072041, "learning_rate": 1.066344993968637e-05, "loss": 0.9377, "step": 442 }, { "epoch": 0.01604491126403477, "grad_norm": 2.4972389948937814, "learning_rate": 1.06875753920386e-05, "loss": 0.8929, "step": 443 }, { "epoch": 0.016081130025353132, "grad_norm": 2.4571944609540948, "learning_rate": 1.0711700844390832e-05, "loss": 0.7326, "step": 444 }, { "epoch": 0.016117348786671495, "grad_norm": 2.493332330843036, "learning_rate": 1.0735826296743064e-05, "loss": 0.9353, "step": 445 }, { "epoch": 0.01615356754798986, "grad_norm": 2.8362611685760357, "learning_rate": 1.0759951749095296e-05, "loss": 0.9602, "step": 446 }, { "epoch": 0.016189786309308223, "grad_norm": 2.9151614727344173, "learning_rate": 1.0784077201447528e-05, "loss": 0.9106, "step": 447 }, { "epoch": 0.016226005070626586, "grad_norm": 2.43448562566588, "learning_rate": 1.080820265379976e-05, "loss": 0.9215, "step": 448 }, { "epoch": 0.016262223831944946, "grad_norm": 2.426488879094696, "learning_rate": 1.083232810615199e-05, "loss": 0.9003, "step": 449 }, { "epoch": 0.01629844259326331, "grad_norm": 2.6190031358331227, "learning_rate": 1.0856453558504223e-05, "loss": 0.9634, "step": 450 }, { "epoch": 0.016334661354581673, "grad_norm": 2.709567397127114, "learning_rate": 1.0880579010856453e-05, "loss": 0.9282, "step": 451 }, { "epoch": 0.016370880115900037, "grad_norm": 2.4136287460982193, "learning_rate": 1.0904704463208687e-05, "loss": 0.7746, "step": 452 }, { "epoch": 0.0164070988772184, "grad_norm": 2.893716416874696, "learning_rate": 1.0928829915560917e-05, "loss": 0.9287, "step": 453 }, { "epoch": 0.01644331763853676, "grad_norm": 2.488282145285932, "learning_rate": 1.095295536791315e-05, "loss": 0.9752, "step": 454 }, { "epoch": 0.016479536399855124, "grad_norm": 2.2992853813484238, "learning_rate": 1.097708082026538e-05, "loss": 0.8639, "step": 455 }, { "epoch": 0.016515755161173488, "grad_norm": 2.269073978652994, "learning_rate": 1.1001206272617612e-05, "loss": 0.9275, "step": 456 }, { "epoch": 0.01655197392249185, "grad_norm": 1.8152219328731978, "learning_rate": 1.1025331724969844e-05, "loss": 0.8226, "step": 457 }, { "epoch": 0.016588192683810215, "grad_norm": 2.5126560947932215, "learning_rate": 1.1049457177322076e-05, "loss": 0.916, "step": 458 }, { "epoch": 0.016624411445128575, "grad_norm": 2.359584298545121, "learning_rate": 1.1073582629674306e-05, "loss": 0.8656, "step": 459 }, { "epoch": 0.01666063020644694, "grad_norm": 2.266568893269498, "learning_rate": 1.109770808202654e-05, "loss": 0.9035, "step": 460 }, { "epoch": 0.016696848967765302, "grad_norm": 2.3509111104222122, "learning_rate": 1.112183353437877e-05, "loss": 0.88, "step": 461 }, { "epoch": 0.016733067729083666, "grad_norm": 2.337520040971858, "learning_rate": 1.1145958986731003e-05, "loss": 0.9184, "step": 462 }, { "epoch": 0.01676928649040203, "grad_norm": 2.488074392885902, "learning_rate": 1.1170084439083233e-05, "loss": 0.9188, "step": 463 }, { "epoch": 0.016805505251720393, "grad_norm": 1.8211378065407224, "learning_rate": 1.1194209891435466e-05, "loss": 0.7616, "step": 464 }, { "epoch": 0.016841724013038753, "grad_norm": 2.415444419891792, "learning_rate": 1.1218335343787696e-05, "loss": 0.9654, "step": 465 }, { "epoch": 0.016877942774357117, "grad_norm": 2.0473757576738563, "learning_rate": 1.124246079613993e-05, "loss": 0.7901, "step": 466 }, { "epoch": 0.01691416153567548, "grad_norm": 2.6406535423587165, "learning_rate": 1.126658624849216e-05, "loss": 0.9337, "step": 467 }, { "epoch": 0.016950380296993844, "grad_norm": 2.456496515594968, "learning_rate": 1.1290711700844392e-05, "loss": 0.8553, "step": 468 }, { "epoch": 0.016986599058312207, "grad_norm": 2.3595086842993696, "learning_rate": 1.1314837153196624e-05, "loss": 0.9777, "step": 469 }, { "epoch": 0.017022817819630567, "grad_norm": 2.371666083477102, "learning_rate": 1.1338962605548855e-05, "loss": 0.9209, "step": 470 }, { "epoch": 0.01705903658094893, "grad_norm": 1.7581233675572745, "learning_rate": 1.1363088057901085e-05, "loss": 0.71, "step": 471 }, { "epoch": 0.017095255342267295, "grad_norm": 1.5849607886903685, "learning_rate": 1.1387213510253319e-05, "loss": 0.7786, "step": 472 }, { "epoch": 0.017131474103585658, "grad_norm": 2.37955863350736, "learning_rate": 1.1411338962605549e-05, "loss": 0.9679, "step": 473 }, { "epoch": 0.01716769286490402, "grad_norm": 2.4780457549515207, "learning_rate": 1.1435464414957782e-05, "loss": 0.9404, "step": 474 }, { "epoch": 0.017203911626222382, "grad_norm": 3.0400475815293073, "learning_rate": 1.1459589867310012e-05, "loss": 0.9139, "step": 475 }, { "epoch": 0.017240130387540745, "grad_norm": 4.118572834470768, "learning_rate": 1.1483715319662246e-05, "loss": 0.9268, "step": 476 }, { "epoch": 0.01727634914885911, "grad_norm": 2.1580176480887645, "learning_rate": 1.1507840772014476e-05, "loss": 0.8149, "step": 477 }, { "epoch": 0.017312567910177473, "grad_norm": 2.3765726889944268, "learning_rate": 1.1531966224366708e-05, "loss": 0.8551, "step": 478 }, { "epoch": 0.017348786671495836, "grad_norm": 2.4907747318198568, "learning_rate": 1.155609167671894e-05, "loss": 0.9607, "step": 479 }, { "epoch": 0.017385005432814196, "grad_norm": 2.6557923890475146, "learning_rate": 1.1580217129071171e-05, "loss": 0.9418, "step": 480 }, { "epoch": 0.01742122419413256, "grad_norm": 2.234913295980479, "learning_rate": 1.1604342581423401e-05, "loss": 0.9054, "step": 481 }, { "epoch": 0.017457442955450923, "grad_norm": 2.3419746923186877, "learning_rate": 1.1628468033775635e-05, "loss": 0.8773, "step": 482 }, { "epoch": 0.017493661716769287, "grad_norm": 2.3837300769551444, "learning_rate": 1.1652593486127865e-05, "loss": 0.9107, "step": 483 }, { "epoch": 0.01752988047808765, "grad_norm": 2.45996712823553, "learning_rate": 1.1676718938480099e-05, "loss": 0.9393, "step": 484 }, { "epoch": 0.01756609923940601, "grad_norm": 2.526890072792949, "learning_rate": 1.1700844390832329e-05, "loss": 0.9188, "step": 485 }, { "epoch": 0.017602318000724374, "grad_norm": 1.8163522224569255, "learning_rate": 1.1724969843184562e-05, "loss": 0.7725, "step": 486 }, { "epoch": 0.017638536762042738, "grad_norm": 2.723248974112392, "learning_rate": 1.1749095295536792e-05, "loss": 0.9162, "step": 487 }, { "epoch": 0.0176747555233611, "grad_norm": 2.6914214976727293, "learning_rate": 1.1773220747889026e-05, "loss": 0.8543, "step": 488 }, { "epoch": 0.017710974284679465, "grad_norm": 2.751261058934199, "learning_rate": 1.1797346200241256e-05, "loss": 0.958, "step": 489 }, { "epoch": 0.01774719304599783, "grad_norm": 1.549284741865264, "learning_rate": 1.1821471652593488e-05, "loss": 0.716, "step": 490 }, { "epoch": 0.01778341180731619, "grad_norm": 1.5490463373405599, "learning_rate": 1.184559710494572e-05, "loss": 0.7148, "step": 491 }, { "epoch": 0.017819630568634552, "grad_norm": 2.8701326175779833, "learning_rate": 1.1869722557297951e-05, "loss": 0.9704, "step": 492 }, { "epoch": 0.017855849329952916, "grad_norm": 1.7951866003719028, "learning_rate": 1.1893848009650181e-05, "loss": 0.8173, "step": 493 }, { "epoch": 0.01789206809127128, "grad_norm": 2.4400293450027295, "learning_rate": 1.1917973462002415e-05, "loss": 0.8414, "step": 494 }, { "epoch": 0.017928286852589643, "grad_norm": 2.536711649154908, "learning_rate": 1.1942098914354645e-05, "loss": 0.9066, "step": 495 }, { "epoch": 0.017964505613908003, "grad_norm": 1.8133941792421113, "learning_rate": 1.1966224366706878e-05, "loss": 0.7508, "step": 496 }, { "epoch": 0.018000724375226367, "grad_norm": 1.6079923107463838, "learning_rate": 1.1990349819059108e-05, "loss": 0.7678, "step": 497 }, { "epoch": 0.01803694313654473, "grad_norm": 2.684430437548066, "learning_rate": 1.2014475271411338e-05, "loss": 0.8498, "step": 498 }, { "epoch": 0.018073161897863094, "grad_norm": 2.5495957033618666, "learning_rate": 1.2038600723763572e-05, "loss": 0.9825, "step": 499 }, { "epoch": 0.018109380659181457, "grad_norm": 1.614582747925888, "learning_rate": 1.2062726176115802e-05, "loss": 0.7012, "step": 500 }, { "epoch": 0.018145599420499817, "grad_norm": 2.565546916698094, "learning_rate": 1.2086851628468035e-05, "loss": 0.9105, "step": 501 }, { "epoch": 0.01818181818181818, "grad_norm": 2.612135332268653, "learning_rate": 1.2110977080820265e-05, "loss": 0.9381, "step": 502 }, { "epoch": 0.018218036943136545, "grad_norm": 1.4660279052651009, "learning_rate": 1.2135102533172497e-05, "loss": 0.7059, "step": 503 }, { "epoch": 0.018254255704454908, "grad_norm": 2.737458310833167, "learning_rate": 1.2159227985524729e-05, "loss": 0.9485, "step": 504 }, { "epoch": 0.01829047446577327, "grad_norm": 2.6110351886086542, "learning_rate": 1.218335343787696e-05, "loss": 1.0208, "step": 505 }, { "epoch": 0.018326693227091632, "grad_norm": 2.414625077604018, "learning_rate": 1.2207478890229193e-05, "loss": 0.9144, "step": 506 }, { "epoch": 0.018362911988409995, "grad_norm": 2.335595708739423, "learning_rate": 1.2231604342581424e-05, "loss": 0.9614, "step": 507 }, { "epoch": 0.01839913074972836, "grad_norm": 2.6067722554287127, "learning_rate": 1.2255729794933654e-05, "loss": 0.914, "step": 508 }, { "epoch": 0.018435349511046722, "grad_norm": 3.166643982835302, "learning_rate": 1.2279855247285888e-05, "loss": 0.9657, "step": 509 }, { "epoch": 0.018471568272365086, "grad_norm": 2.7286583351819216, "learning_rate": 1.2303980699638118e-05, "loss": 0.8758, "step": 510 }, { "epoch": 0.01850778703368345, "grad_norm": 3.1595478627487985, "learning_rate": 1.2328106151990351e-05, "loss": 0.8831, "step": 511 }, { "epoch": 0.01854400579500181, "grad_norm": 2.4123803484601574, "learning_rate": 1.2352231604342582e-05, "loss": 0.8556, "step": 512 }, { "epoch": 0.018580224556320173, "grad_norm": 2.4355656119132942, "learning_rate": 1.2376357056694815e-05, "loss": 0.883, "step": 513 }, { "epoch": 0.018616443317638537, "grad_norm": 2.417947873840434, "learning_rate": 1.2400482509047045e-05, "loss": 0.8907, "step": 514 }, { "epoch": 0.0186526620789569, "grad_norm": 1.8743950168305425, "learning_rate": 1.2424607961399277e-05, "loss": 0.6945, "step": 515 }, { "epoch": 0.018688880840275264, "grad_norm": 2.6104780609614178, "learning_rate": 1.2448733413751509e-05, "loss": 0.9086, "step": 516 }, { "epoch": 0.018725099601593624, "grad_norm": 2.6326598624064785, "learning_rate": 1.247285886610374e-05, "loss": 0.9537, "step": 517 }, { "epoch": 0.018761318362911988, "grad_norm": 2.75704182211756, "learning_rate": 1.249698431845597e-05, "loss": 0.8538, "step": 518 }, { "epoch": 0.01879753712423035, "grad_norm": 2.425943596485934, "learning_rate": 1.2521109770808204e-05, "loss": 0.9104, "step": 519 }, { "epoch": 0.018833755885548715, "grad_norm": 1.8272643115629008, "learning_rate": 1.2545235223160434e-05, "loss": 0.7841, "step": 520 }, { "epoch": 0.01886997464686708, "grad_norm": 2.4388216635343527, "learning_rate": 1.2569360675512668e-05, "loss": 0.9079, "step": 521 }, { "epoch": 0.01890619340818544, "grad_norm": 2.51585013703078, "learning_rate": 1.2593486127864898e-05, "loss": 0.9244, "step": 522 }, { "epoch": 0.018942412169503802, "grad_norm": 2.529393330505965, "learning_rate": 1.2617611580217131e-05, "loss": 0.8745, "step": 523 }, { "epoch": 0.018978630930822166, "grad_norm": 2.6989227303738463, "learning_rate": 1.2641737032569361e-05, "loss": 0.8987, "step": 524 }, { "epoch": 0.01901484969214053, "grad_norm": 2.7503148299680777, "learning_rate": 1.2665862484921593e-05, "loss": 0.8909, "step": 525 }, { "epoch": 0.019051068453458893, "grad_norm": 2.4763600961665033, "learning_rate": 1.2689987937273825e-05, "loss": 0.9086, "step": 526 }, { "epoch": 0.019087287214777253, "grad_norm": 1.6834710037963652, "learning_rate": 1.2714113389626057e-05, "loss": 0.7343, "step": 527 }, { "epoch": 0.019123505976095617, "grad_norm": 1.4853244854815364, "learning_rate": 1.2738238841978288e-05, "loss": 0.7635, "step": 528 }, { "epoch": 0.01915972473741398, "grad_norm": 2.963274805739315, "learning_rate": 1.276236429433052e-05, "loss": 0.8945, "step": 529 }, { "epoch": 0.019195943498732344, "grad_norm": 2.6633024865554535, "learning_rate": 1.278648974668275e-05, "loss": 0.9053, "step": 530 }, { "epoch": 0.019232162260050707, "grad_norm": 1.6363867667551617, "learning_rate": 1.2810615199034984e-05, "loss": 0.7426, "step": 531 }, { "epoch": 0.01926838102136907, "grad_norm": 2.57329663503181, "learning_rate": 1.2834740651387214e-05, "loss": 0.8687, "step": 532 }, { "epoch": 0.01930459978268743, "grad_norm": 2.2547745180950893, "learning_rate": 1.2858866103739447e-05, "loss": 0.8771, "step": 533 }, { "epoch": 0.019340818544005794, "grad_norm": 2.8118219803421765, "learning_rate": 1.2882991556091677e-05, "loss": 0.9019, "step": 534 }, { "epoch": 0.019377037305324158, "grad_norm": 2.313396855048405, "learning_rate": 1.290711700844391e-05, "loss": 0.987, "step": 535 }, { "epoch": 0.01941325606664252, "grad_norm": 2.636270354962437, "learning_rate": 1.2931242460796141e-05, "loss": 0.9579, "step": 536 }, { "epoch": 0.019449474827960885, "grad_norm": 2.521958491977086, "learning_rate": 1.2955367913148373e-05, "loss": 0.9073, "step": 537 }, { "epoch": 0.019485693589279245, "grad_norm": 2.527701698909077, "learning_rate": 1.2979493365500604e-05, "loss": 0.8633, "step": 538 }, { "epoch": 0.01952191235059761, "grad_norm": 2.782401267203926, "learning_rate": 1.3003618817852836e-05, "loss": 0.9505, "step": 539 }, { "epoch": 0.019558131111915972, "grad_norm": 2.6245598764658755, "learning_rate": 1.3027744270205066e-05, "loss": 0.9279, "step": 540 }, { "epoch": 0.019594349873234336, "grad_norm": 2.550529403287826, "learning_rate": 1.30518697225573e-05, "loss": 0.9006, "step": 541 }, { "epoch": 0.0196305686345527, "grad_norm": 2.369639503743001, "learning_rate": 1.307599517490953e-05, "loss": 0.8341, "step": 542 }, { "epoch": 0.01966678739587106, "grad_norm": 2.721408831885408, "learning_rate": 1.3100120627261763e-05, "loss": 0.9274, "step": 543 }, { "epoch": 0.019703006157189423, "grad_norm": 3.1154180107880785, "learning_rate": 1.3124246079613993e-05, "loss": 0.8975, "step": 544 }, { "epoch": 0.019739224918507787, "grad_norm": 2.2798989159933543, "learning_rate": 1.3148371531966227e-05, "loss": 0.8029, "step": 545 }, { "epoch": 0.01977544367982615, "grad_norm": 1.6169905411582535, "learning_rate": 1.3172496984318457e-05, "loss": 0.7421, "step": 546 }, { "epoch": 0.019811662441144514, "grad_norm": 1.902017811713085, "learning_rate": 1.319662243667069e-05, "loss": 0.7556, "step": 547 }, { "epoch": 0.019847881202462874, "grad_norm": 2.5737140230171947, "learning_rate": 1.322074788902292e-05, "loss": 0.9131, "step": 548 }, { "epoch": 0.019884099963781238, "grad_norm": 2.571172245789799, "learning_rate": 1.3244873341375152e-05, "loss": 0.9234, "step": 549 }, { "epoch": 0.0199203187250996, "grad_norm": 2.5245857254188446, "learning_rate": 1.3268998793727384e-05, "loss": 0.9547, "step": 550 }, { "epoch": 0.019956537486417965, "grad_norm": 2.4837485459270163, "learning_rate": 1.3293124246079616e-05, "loss": 0.939, "step": 551 }, { "epoch": 0.01999275624773633, "grad_norm": 2.5852239062118714, "learning_rate": 1.3317249698431846e-05, "loss": 0.8499, "step": 552 }, { "epoch": 0.020028975009054692, "grad_norm": 2.798151801012701, "learning_rate": 1.3341375150784078e-05, "loss": 0.9293, "step": 553 }, { "epoch": 0.020065193770373052, "grad_norm": 2.6599996453456174, "learning_rate": 1.336550060313631e-05, "loss": 0.9143, "step": 554 }, { "epoch": 0.020101412531691416, "grad_norm": 3.018233905137923, "learning_rate": 1.338962605548854e-05, "loss": 0.9169, "step": 555 }, { "epoch": 0.02013763129300978, "grad_norm": 2.293807123253535, "learning_rate": 1.3413751507840773e-05, "loss": 0.7187, "step": 556 }, { "epoch": 0.020173850054328143, "grad_norm": 2.451174799092893, "learning_rate": 1.3437876960193003e-05, "loss": 0.8985, "step": 557 }, { "epoch": 0.020210068815646506, "grad_norm": 2.4790314632835435, "learning_rate": 1.3462002412545237e-05, "loss": 0.8952, "step": 558 }, { "epoch": 0.020246287576964866, "grad_norm": 2.5659185320620366, "learning_rate": 1.3486127864897467e-05, "loss": 0.9792, "step": 559 }, { "epoch": 0.02028250633828323, "grad_norm": 2.3290703889372057, "learning_rate": 1.35102533172497e-05, "loss": 0.937, "step": 560 }, { "epoch": 0.020318725099601594, "grad_norm": 2.3517295241050484, "learning_rate": 1.353437876960193e-05, "loss": 0.9047, "step": 561 }, { "epoch": 0.020354943860919957, "grad_norm": 2.514056166741056, "learning_rate": 1.3558504221954162e-05, "loss": 0.9083, "step": 562 }, { "epoch": 0.02039116262223832, "grad_norm": 2.1390200679860283, "learning_rate": 1.3582629674306394e-05, "loss": 0.7906, "step": 563 }, { "epoch": 0.02042738138355668, "grad_norm": 2.266531521209291, "learning_rate": 1.3606755126658626e-05, "loss": 0.8697, "step": 564 }, { "epoch": 0.020463600144875044, "grad_norm": 2.5081083384677423, "learning_rate": 1.3630880579010857e-05, "loss": 0.9531, "step": 565 }, { "epoch": 0.020499818906193408, "grad_norm": 2.486074458664708, "learning_rate": 1.365500603136309e-05, "loss": 0.9501, "step": 566 }, { "epoch": 0.02053603766751177, "grad_norm": 2.4420411726276314, "learning_rate": 1.367913148371532e-05, "loss": 0.9368, "step": 567 }, { "epoch": 0.020572256428830135, "grad_norm": 2.4310673033977754, "learning_rate": 1.3703256936067553e-05, "loss": 0.857, "step": 568 }, { "epoch": 0.020608475190148495, "grad_norm": 1.776866101636466, "learning_rate": 1.3727382388419783e-05, "loss": 0.7545, "step": 569 }, { "epoch": 0.02064469395146686, "grad_norm": 2.578080806931393, "learning_rate": 1.3751507840772016e-05, "loss": 0.9598, "step": 570 }, { "epoch": 0.020680912712785222, "grad_norm": 1.514693888645448, "learning_rate": 1.3775633293124246e-05, "loss": 0.8065, "step": 571 }, { "epoch": 0.020717131474103586, "grad_norm": 2.447676667494471, "learning_rate": 1.379975874547648e-05, "loss": 0.9625, "step": 572 }, { "epoch": 0.02075335023542195, "grad_norm": 2.8749507298601453, "learning_rate": 1.382388419782871e-05, "loss": 0.9325, "step": 573 }, { "epoch": 0.020789568996740313, "grad_norm": 2.4226872082048203, "learning_rate": 1.3848009650180942e-05, "loss": 0.898, "step": 574 }, { "epoch": 0.020825787758058673, "grad_norm": 2.512986413525613, "learning_rate": 1.3872135102533174e-05, "loss": 0.9867, "step": 575 }, { "epoch": 0.020862006519377037, "grad_norm": 2.683695428912645, "learning_rate": 1.3896260554885405e-05, "loss": 0.8495, "step": 576 }, { "epoch": 0.0208982252806954, "grad_norm": 2.425146941398224, "learning_rate": 1.3920386007237635e-05, "loss": 0.8814, "step": 577 }, { "epoch": 0.020934444042013764, "grad_norm": 2.437930270879491, "learning_rate": 1.3944511459589869e-05, "loss": 0.8828, "step": 578 }, { "epoch": 0.020970662803332128, "grad_norm": 1.850977546202167, "learning_rate": 1.3968636911942099e-05, "loss": 0.8034, "step": 579 }, { "epoch": 0.021006881564650488, "grad_norm": 2.3929431979303115, "learning_rate": 1.3992762364294332e-05, "loss": 0.9843, "step": 580 }, { "epoch": 0.02104310032596885, "grad_norm": 1.5372061229417409, "learning_rate": 1.4016887816646563e-05, "loss": 0.829, "step": 581 }, { "epoch": 0.021079319087287215, "grad_norm": 2.693176501222768, "learning_rate": 1.4041013268998796e-05, "loss": 1.005, "step": 582 }, { "epoch": 0.02111553784860558, "grad_norm": 2.742446061249837, "learning_rate": 1.4065138721351026e-05, "loss": 0.9508, "step": 583 }, { "epoch": 0.021151756609923942, "grad_norm": 2.584252087715769, "learning_rate": 1.4089264173703258e-05, "loss": 0.9263, "step": 584 }, { "epoch": 0.021187975371242302, "grad_norm": 3.029588452884719, "learning_rate": 1.411338962605549e-05, "loss": 0.8983, "step": 585 }, { "epoch": 0.021224194132560666, "grad_norm": 1.8461344892268172, "learning_rate": 1.4137515078407721e-05, "loss": 0.776, "step": 586 }, { "epoch": 0.02126041289387903, "grad_norm": 1.5118662710746524, "learning_rate": 1.4161640530759953e-05, "loss": 0.7704, "step": 587 }, { "epoch": 0.021296631655197393, "grad_norm": 1.571888991301567, "learning_rate": 1.4185765983112185e-05, "loss": 0.7911, "step": 588 }, { "epoch": 0.021332850416515756, "grad_norm": 2.9523220534888908, "learning_rate": 1.4209891435464415e-05, "loss": 0.9426, "step": 589 }, { "epoch": 0.021369069177834116, "grad_norm": 1.479536848448547, "learning_rate": 1.4234016887816649e-05, "loss": 0.758, "step": 590 }, { "epoch": 0.02140528793915248, "grad_norm": 2.9094818255421786, "learning_rate": 1.4258142340168879e-05, "loss": 0.9328, "step": 591 }, { "epoch": 0.021441506700470844, "grad_norm": 2.91726711370921, "learning_rate": 1.4282267792521112e-05, "loss": 0.958, "step": 592 }, { "epoch": 0.021477725461789207, "grad_norm": 2.485141689829909, "learning_rate": 1.4306393244873342e-05, "loss": 0.9976, "step": 593 }, { "epoch": 0.02151394422310757, "grad_norm": 2.708479407234843, "learning_rate": 1.4330518697225576e-05, "loss": 0.9455, "step": 594 }, { "epoch": 0.021550162984425934, "grad_norm": 2.9125378278805223, "learning_rate": 1.4354644149577806e-05, "loss": 0.909, "step": 595 }, { "epoch": 0.021586381745744294, "grad_norm": 2.4936860998948998, "learning_rate": 1.4378769601930038e-05, "loss": 0.9055, "step": 596 }, { "epoch": 0.021622600507062658, "grad_norm": 2.586659189573727, "learning_rate": 1.440289505428227e-05, "loss": 0.9722, "step": 597 }, { "epoch": 0.02165881926838102, "grad_norm": 2.130736559996397, "learning_rate": 1.4427020506634501e-05, "loss": 0.7882, "step": 598 }, { "epoch": 0.021695038029699385, "grad_norm": 2.5227205273616553, "learning_rate": 1.4451145958986731e-05, "loss": 0.8975, "step": 599 }, { "epoch": 0.02173125679101775, "grad_norm": 2.4809724161218742, "learning_rate": 1.4475271411338965e-05, "loss": 0.9453, "step": 600 }, { "epoch": 0.02176747555233611, "grad_norm": 2.471411662978827, "learning_rate": 1.4499396863691195e-05, "loss": 0.973, "step": 601 }, { "epoch": 0.021803694313654472, "grad_norm": 1.4422175365018945, "learning_rate": 1.4523522316043428e-05, "loss": 0.7356, "step": 602 }, { "epoch": 0.021839913074972836, "grad_norm": 2.5030446169420437, "learning_rate": 1.4547647768395658e-05, "loss": 0.9054, "step": 603 }, { "epoch": 0.0218761318362912, "grad_norm": 1.5404701338662734, "learning_rate": 1.4571773220747892e-05, "loss": 0.8091, "step": 604 }, { "epoch": 0.021912350597609563, "grad_norm": 2.447388794304293, "learning_rate": 1.4595898673100122e-05, "loss": 0.8903, "step": 605 }, { "epoch": 0.021948569358927923, "grad_norm": 2.4213752800159156, "learning_rate": 1.4620024125452354e-05, "loss": 0.9813, "step": 606 }, { "epoch": 0.021984788120246287, "grad_norm": 2.5997019865189754, "learning_rate": 1.4644149577804585e-05, "loss": 0.9991, "step": 607 }, { "epoch": 0.02202100688156465, "grad_norm": 2.4592496987866457, "learning_rate": 1.4668275030156816e-05, "loss": 0.9087, "step": 608 }, { "epoch": 0.022057225642883014, "grad_norm": 2.320265565583746, "learning_rate": 1.4692400482509049e-05, "loss": 0.9335, "step": 609 }, { "epoch": 0.022093444404201378, "grad_norm": 2.4759520130244885, "learning_rate": 1.4716525934861279e-05, "loss": 0.9547, "step": 610 }, { "epoch": 0.022129663165519738, "grad_norm": 2.3592572245980885, "learning_rate": 1.4740651387213511e-05, "loss": 0.8927, "step": 611 }, { "epoch": 0.0221658819268381, "grad_norm": 2.485722776019304, "learning_rate": 1.4764776839565743e-05, "loss": 0.9806, "step": 612 }, { "epoch": 0.022202100688156465, "grad_norm": 2.3314594273804237, "learning_rate": 1.4788902291917974e-05, "loss": 0.9078, "step": 613 }, { "epoch": 0.02223831944947483, "grad_norm": 2.0034387033306733, "learning_rate": 1.4813027744270205e-05, "loss": 0.7576, "step": 614 }, { "epoch": 0.022274538210793192, "grad_norm": 2.3920561536389244, "learning_rate": 1.4837153196622438e-05, "loss": 0.9114, "step": 615 }, { "epoch": 0.022310756972111555, "grad_norm": 2.4694516868771768, "learning_rate": 1.4861278648974668e-05, "loss": 0.9464, "step": 616 }, { "epoch": 0.022346975733429916, "grad_norm": 2.5151542999180396, "learning_rate": 1.4885404101326902e-05, "loss": 0.8675, "step": 617 }, { "epoch": 0.02238319449474828, "grad_norm": 2.754637384418598, "learning_rate": 1.4909529553679132e-05, "loss": 0.9523, "step": 618 }, { "epoch": 0.022419413256066643, "grad_norm": 2.100858160847516, "learning_rate": 1.4933655006031365e-05, "loss": 0.8745, "step": 619 }, { "epoch": 0.022455632017385006, "grad_norm": 2.4397739782748915, "learning_rate": 1.4957780458383595e-05, "loss": 0.8503, "step": 620 }, { "epoch": 0.02249185077870337, "grad_norm": 2.2825427523935633, "learning_rate": 1.4981905910735827e-05, "loss": 0.9337, "step": 621 }, { "epoch": 0.02252806954002173, "grad_norm": 2.1925965950434465, "learning_rate": 1.5006031363088059e-05, "loss": 0.9024, "step": 622 }, { "epoch": 0.022564288301340094, "grad_norm": 1.7716452150709718, "learning_rate": 1.503015681544029e-05, "loss": 0.7835, "step": 623 }, { "epoch": 0.022600507062658457, "grad_norm": 2.3422909566184393, "learning_rate": 1.5054282267792522e-05, "loss": 0.9262, "step": 624 }, { "epoch": 0.02263672582397682, "grad_norm": 2.2748485001224186, "learning_rate": 1.5078407720144754e-05, "loss": 0.9587, "step": 625 }, { "epoch": 0.022672944585295184, "grad_norm": 2.3645341751825684, "learning_rate": 1.5102533172496984e-05, "loss": 0.9435, "step": 626 }, { "epoch": 0.022709163346613544, "grad_norm": 1.6866858019864068, "learning_rate": 1.5126658624849218e-05, "loss": 0.7556, "step": 627 }, { "epoch": 0.022745382107931908, "grad_norm": 2.518032586368243, "learning_rate": 1.5150784077201448e-05, "loss": 0.9106, "step": 628 }, { "epoch": 0.02278160086925027, "grad_norm": 1.4347500872653467, "learning_rate": 1.5174909529553681e-05, "loss": 0.7967, "step": 629 }, { "epoch": 0.022817819630568635, "grad_norm": 2.386971583608565, "learning_rate": 1.5199034981905911e-05, "loss": 0.9133, "step": 630 }, { "epoch": 0.022854038391887, "grad_norm": 2.4494059352741444, "learning_rate": 1.5223160434258145e-05, "loss": 0.9838, "step": 631 }, { "epoch": 0.02289025715320536, "grad_norm": 2.451590537986135, "learning_rate": 1.5247285886610375e-05, "loss": 0.9096, "step": 632 }, { "epoch": 0.022926475914523722, "grad_norm": 1.5555144786684094, "learning_rate": 1.5271411338962607e-05, "loss": 0.7071, "step": 633 }, { "epoch": 0.022962694675842086, "grad_norm": 2.5848613816398824, "learning_rate": 1.5295536791314837e-05, "loss": 0.8924, "step": 634 }, { "epoch": 0.02299891343716045, "grad_norm": 2.538571407069289, "learning_rate": 1.531966224366707e-05, "loss": 0.9721, "step": 635 }, { "epoch": 0.023035132198478813, "grad_norm": 2.730286956432724, "learning_rate": 1.53437876960193e-05, "loss": 0.8858, "step": 636 }, { "epoch": 0.023071350959797177, "grad_norm": 1.6569256793893383, "learning_rate": 1.5367913148371534e-05, "loss": 0.733, "step": 637 }, { "epoch": 0.023107569721115537, "grad_norm": 2.714358902592965, "learning_rate": 1.5392038600723764e-05, "loss": 0.9449, "step": 638 }, { "epoch": 0.0231437884824339, "grad_norm": 2.498237358030736, "learning_rate": 1.5416164053075997e-05, "loss": 0.8342, "step": 639 }, { "epoch": 0.023180007243752264, "grad_norm": 3.2090032232175636, "learning_rate": 1.5440289505428227e-05, "loss": 0.8627, "step": 640 }, { "epoch": 0.023216226005070627, "grad_norm": 3.382873637019162, "learning_rate": 1.546441495778046e-05, "loss": 0.8794, "step": 641 }, { "epoch": 0.02325244476638899, "grad_norm": 1.413499280157102, "learning_rate": 1.548854041013269e-05, "loss": 0.715, "step": 642 }, { "epoch": 0.02328866352770735, "grad_norm": 3.55212687603653, "learning_rate": 1.5512665862484924e-05, "loss": 0.9533, "step": 643 }, { "epoch": 0.023324882289025715, "grad_norm": 2.744775254793462, "learning_rate": 1.5536791314837155e-05, "loss": 0.9567, "step": 644 }, { "epoch": 0.02336110105034408, "grad_norm": 3.2092229112508717, "learning_rate": 1.5560916767189388e-05, "loss": 0.938, "step": 645 }, { "epoch": 0.023397319811662442, "grad_norm": 3.3862643877570764, "learning_rate": 1.5585042219541618e-05, "loss": 0.906, "step": 646 }, { "epoch": 0.023433538572980805, "grad_norm": 2.8601859980236792, "learning_rate": 1.560916767189385e-05, "loss": 0.9535, "step": 647 }, { "epoch": 0.023469757334299166, "grad_norm": 2.327913985194222, "learning_rate": 1.563329312424608e-05, "loss": 0.9573, "step": 648 }, { "epoch": 0.02350597609561753, "grad_norm": 2.411295993825065, "learning_rate": 1.5657418576598312e-05, "loss": 0.8623, "step": 649 }, { "epoch": 0.023542194856935893, "grad_norm": 2.518718236410565, "learning_rate": 1.5681544028950545e-05, "loss": 0.876, "step": 650 }, { "epoch": 0.023578413618254256, "grad_norm": 2.8460016925257854, "learning_rate": 1.5705669481302775e-05, "loss": 0.9062, "step": 651 }, { "epoch": 0.02361463237957262, "grad_norm": 2.547736455174824, "learning_rate": 1.572979493365501e-05, "loss": 0.9216, "step": 652 }, { "epoch": 0.02365085114089098, "grad_norm": 2.80702022674545, "learning_rate": 1.575392038600724e-05, "loss": 0.968, "step": 653 }, { "epoch": 0.023687069902209344, "grad_norm": 2.6268349035952725, "learning_rate": 1.577804583835947e-05, "loss": 0.9517, "step": 654 }, { "epoch": 0.023723288663527707, "grad_norm": 2.794899963730202, "learning_rate": 1.5802171290711702e-05, "loss": 0.8453, "step": 655 }, { "epoch": 0.02375950742484607, "grad_norm": 2.9359856212300257, "learning_rate": 1.5826296743063932e-05, "loss": 0.8983, "step": 656 }, { "epoch": 0.023795726186164434, "grad_norm": 3.6942470357323725, "learning_rate": 1.5850422195416166e-05, "loss": 0.9858, "step": 657 }, { "epoch": 0.023831944947482798, "grad_norm": 2.426495528294829, "learning_rate": 1.5874547647768396e-05, "loss": 0.9152, "step": 658 }, { "epoch": 0.023868163708801158, "grad_norm": 2.712418552903054, "learning_rate": 1.589867310012063e-05, "loss": 0.895, "step": 659 }, { "epoch": 0.02390438247011952, "grad_norm": 2.694967142729318, "learning_rate": 1.592279855247286e-05, "loss": 0.9654, "step": 660 }, { "epoch": 0.023940601231437885, "grad_norm": 1.6692168091399642, "learning_rate": 1.5946924004825093e-05, "loss": 0.7328, "step": 661 }, { "epoch": 0.02397681999275625, "grad_norm": 2.684705215345033, "learning_rate": 1.5971049457177323e-05, "loss": 0.9441, "step": 662 }, { "epoch": 0.024013038754074612, "grad_norm": 2.2769936444598424, "learning_rate": 1.5995174909529557e-05, "loss": 0.9077, "step": 663 }, { "epoch": 0.024049257515392972, "grad_norm": 2.5027364328041624, "learning_rate": 1.6019300361881787e-05, "loss": 0.9234, "step": 664 }, { "epoch": 0.024085476276711336, "grad_norm": 2.4423826843774177, "learning_rate": 1.6043425814234017e-05, "loss": 0.8101, "step": 665 }, { "epoch": 0.0241216950380297, "grad_norm": 2.4073376318511595, "learning_rate": 1.606755126658625e-05, "loss": 0.9011, "step": 666 }, { "epoch": 0.024157913799348063, "grad_norm": 2.298099857830303, "learning_rate": 1.609167671893848e-05, "loss": 0.9197, "step": 667 }, { "epoch": 0.024194132560666427, "grad_norm": 1.8166845883899594, "learning_rate": 1.6115802171290714e-05, "loss": 0.7561, "step": 668 }, { "epoch": 0.024230351321984787, "grad_norm": 2.775604339145624, "learning_rate": 1.6139927623642944e-05, "loss": 0.8866, "step": 669 }, { "epoch": 0.02426657008330315, "grad_norm": 2.3449468529047848, "learning_rate": 1.6164053075995177e-05, "loss": 0.9433, "step": 670 }, { "epoch": 0.024302788844621514, "grad_norm": 2.4424157916393097, "learning_rate": 1.6188178528347407e-05, "loss": 0.9566, "step": 671 }, { "epoch": 0.024339007605939877, "grad_norm": 2.4084637880604465, "learning_rate": 1.621230398069964e-05, "loss": 0.9135, "step": 672 }, { "epoch": 0.02437522636725824, "grad_norm": 2.6763616308363245, "learning_rate": 1.623642943305187e-05, "loss": 0.8417, "step": 673 }, { "epoch": 0.0244114451285766, "grad_norm": 2.6728763042582884, "learning_rate": 1.6260554885404105e-05, "loss": 0.9218, "step": 674 }, { "epoch": 0.024447663889894965, "grad_norm": 2.307704531598269, "learning_rate": 1.6284680337756335e-05, "loss": 0.8457, "step": 675 }, { "epoch": 0.02448388265121333, "grad_norm": 1.6680475605541645, "learning_rate": 1.6308805790108565e-05, "loss": 0.8939, "step": 676 }, { "epoch": 0.024520101412531692, "grad_norm": 2.479433761395102, "learning_rate": 1.6332931242460798e-05, "loss": 0.8709, "step": 677 }, { "epoch": 0.024556320173850055, "grad_norm": 1.4254154383119857, "learning_rate": 1.6357056694813028e-05, "loss": 0.817, "step": 678 }, { "epoch": 0.024592538935168416, "grad_norm": 2.3072982855085473, "learning_rate": 1.638118214716526e-05, "loss": 0.9586, "step": 679 }, { "epoch": 0.02462875769648678, "grad_norm": 2.268042509447891, "learning_rate": 1.6405307599517492e-05, "loss": 0.9302, "step": 680 }, { "epoch": 0.024664976457805143, "grad_norm": 2.323571039555325, "learning_rate": 1.6429433051869722e-05, "loss": 0.9078, "step": 681 }, { "epoch": 0.024701195219123506, "grad_norm": 2.2653860821207914, "learning_rate": 1.6453558504221955e-05, "loss": 0.9329, "step": 682 }, { "epoch": 0.02473741398044187, "grad_norm": 1.5496353897945403, "learning_rate": 1.6477683956574185e-05, "loss": 0.7161, "step": 683 }, { "epoch": 0.024773632741760233, "grad_norm": 1.5152750876399939, "learning_rate": 1.650180940892642e-05, "loss": 0.7597, "step": 684 }, { "epoch": 0.024809851503078593, "grad_norm": 2.69523297897458, "learning_rate": 1.652593486127865e-05, "loss": 1.0184, "step": 685 }, { "epoch": 0.024846070264396957, "grad_norm": 2.498429690355378, "learning_rate": 1.6550060313630883e-05, "loss": 1.0387, "step": 686 }, { "epoch": 0.02488228902571532, "grad_norm": 2.4865610263815783, "learning_rate": 1.6574185765983113e-05, "loss": 0.8198, "step": 687 }, { "epoch": 0.024918507787033684, "grad_norm": 2.357053151090565, "learning_rate": 1.6598311218335346e-05, "loss": 0.9495, "step": 688 }, { "epoch": 0.024954726548352048, "grad_norm": 2.4066992640646023, "learning_rate": 1.6622436670687576e-05, "loss": 0.8745, "step": 689 }, { "epoch": 0.024990945309670408, "grad_norm": 2.684626956479487, "learning_rate": 1.664656212303981e-05, "loss": 0.9428, "step": 690 }, { "epoch": 0.02502716407098877, "grad_norm": 2.476736861327514, "learning_rate": 1.667068757539204e-05, "loss": 0.9527, "step": 691 }, { "epoch": 0.025063382832307135, "grad_norm": 2.719974521586004, "learning_rate": 1.6694813027744273e-05, "loss": 0.9871, "step": 692 }, { "epoch": 0.0250996015936255, "grad_norm": 2.4214369008456123, "learning_rate": 1.6718938480096503e-05, "loss": 0.8943, "step": 693 }, { "epoch": 0.025135820354943862, "grad_norm": 2.4164297339540517, "learning_rate": 1.6743063932448737e-05, "loss": 0.9499, "step": 694 }, { "epoch": 0.025172039116262222, "grad_norm": 2.3669272474868848, "learning_rate": 1.6767189384800967e-05, "loss": 0.8765, "step": 695 }, { "epoch": 0.025208257877580586, "grad_norm": 2.403559812712755, "learning_rate": 1.67913148371532e-05, "loss": 0.9135, "step": 696 }, { "epoch": 0.02524447663889895, "grad_norm": 2.4362215699855896, "learning_rate": 1.681544028950543e-05, "loss": 0.9386, "step": 697 }, { "epoch": 0.025280695400217313, "grad_norm": 2.5558195230839345, "learning_rate": 1.683956574185766e-05, "loss": 0.8842, "step": 698 }, { "epoch": 0.025316914161535677, "grad_norm": 2.3883592879848208, "learning_rate": 1.6863691194209894e-05, "loss": 0.8549, "step": 699 }, { "epoch": 0.025353132922854037, "grad_norm": 2.4762530402378973, "learning_rate": 1.6887816646562124e-05, "loss": 0.7846, "step": 700 }, { "epoch": 0.0253893516841724, "grad_norm": 2.3939592509032397, "learning_rate": 1.6911942098914354e-05, "loss": 0.9405, "step": 701 }, { "epoch": 0.025425570445490764, "grad_norm": 2.8809906027695904, "learning_rate": 1.6936067551266588e-05, "loss": 0.9928, "step": 702 }, { "epoch": 0.025461789206809127, "grad_norm": 1.572534263359608, "learning_rate": 1.6960193003618818e-05, "loss": 0.7969, "step": 703 }, { "epoch": 0.02549800796812749, "grad_norm": 2.4558722729751516, "learning_rate": 1.698431845597105e-05, "loss": 0.9101, "step": 704 }, { "epoch": 0.025534226729445855, "grad_norm": 2.6109163677243856, "learning_rate": 1.700844390832328e-05, "loss": 1.0028, "step": 705 }, { "epoch": 0.025570445490764215, "grad_norm": 2.371482525018766, "learning_rate": 1.7032569360675515e-05, "loss": 0.9386, "step": 706 }, { "epoch": 0.025606664252082578, "grad_norm": 2.2632702495029773, "learning_rate": 1.7056694813027745e-05, "loss": 0.9348, "step": 707 }, { "epoch": 0.025642883013400942, "grad_norm": 2.845075707006544, "learning_rate": 1.7080820265379978e-05, "loss": 0.8831, "step": 708 }, { "epoch": 0.025679101774719305, "grad_norm": 2.4798111592102234, "learning_rate": 1.710494571773221e-05, "loss": 0.872, "step": 709 }, { "epoch": 0.02571532053603767, "grad_norm": 2.19693267842866, "learning_rate": 1.7129071170084442e-05, "loss": 0.9123, "step": 710 }, { "epoch": 0.02575153929735603, "grad_norm": 2.1467727748543024, "learning_rate": 1.7153196622436672e-05, "loss": 0.9117, "step": 711 }, { "epoch": 0.025787758058674393, "grad_norm": 2.2715985670439447, "learning_rate": 1.7177322074788905e-05, "loss": 0.9085, "step": 712 }, { "epoch": 0.025823976819992756, "grad_norm": 2.390066979299699, "learning_rate": 1.7201447527141135e-05, "loss": 0.93, "step": 713 }, { "epoch": 0.02586019558131112, "grad_norm": 2.707342537270767, "learning_rate": 1.722557297949337e-05, "loss": 0.9219, "step": 714 }, { "epoch": 0.025896414342629483, "grad_norm": 2.403461286307809, "learning_rate": 1.72496984318456e-05, "loss": 0.9193, "step": 715 }, { "epoch": 0.025932633103947843, "grad_norm": 2.1149565610260312, "learning_rate": 1.7273823884197833e-05, "loss": 0.8672, "step": 716 }, { "epoch": 0.025968851865266207, "grad_norm": 2.601260415339888, "learning_rate": 1.7297949336550063e-05, "loss": 0.7863, "step": 717 }, { "epoch": 0.02600507062658457, "grad_norm": 2.0605710697109547, "learning_rate": 1.7322074788902296e-05, "loss": 0.8102, "step": 718 }, { "epoch": 0.026041289387902934, "grad_norm": 2.7105530815311836, "learning_rate": 1.7346200241254526e-05, "loss": 0.8089, "step": 719 }, { "epoch": 0.026077508149221298, "grad_norm": 1.81092420458233, "learning_rate": 1.7370325693606756e-05, "loss": 0.8569, "step": 720 }, { "epoch": 0.026113726910539658, "grad_norm": 2.5865986719260348, "learning_rate": 1.739445114595899e-05, "loss": 1.0058, "step": 721 }, { "epoch": 0.02614994567185802, "grad_norm": 2.4307674632717786, "learning_rate": 1.741857659831122e-05, "loss": 1.0029, "step": 722 }, { "epoch": 0.026186164433176385, "grad_norm": 1.838729521003956, "learning_rate": 1.744270205066345e-05, "loss": 0.8055, "step": 723 }, { "epoch": 0.02622238319449475, "grad_norm": 2.355430159877384, "learning_rate": 1.7466827503015683e-05, "loss": 0.9975, "step": 724 }, { "epoch": 0.026258601955813112, "grad_norm": 1.5513496298698721, "learning_rate": 1.7490952955367913e-05, "loss": 0.7423, "step": 725 }, { "epoch": 0.026294820717131476, "grad_norm": 2.305078054730961, "learning_rate": 1.7515078407720144e-05, "loss": 0.8725, "step": 726 }, { "epoch": 0.026331039478449836, "grad_norm": 2.1591138152482805, "learning_rate": 1.7539203860072377e-05, "loss": 0.9211, "step": 727 }, { "epoch": 0.0263672582397682, "grad_norm": 2.3741793657797623, "learning_rate": 1.7563329312424607e-05, "loss": 0.9573, "step": 728 }, { "epoch": 0.026403477001086563, "grad_norm": 2.360229318567965, "learning_rate": 1.758745476477684e-05, "loss": 0.9195, "step": 729 }, { "epoch": 0.026439695762404927, "grad_norm": 2.5045135954669497, "learning_rate": 1.761158021712907e-05, "loss": 1.0276, "step": 730 }, { "epoch": 0.02647591452372329, "grad_norm": 1.7303827299167276, "learning_rate": 1.7635705669481304e-05, "loss": 0.7767, "step": 731 }, { "epoch": 0.02651213328504165, "grad_norm": 1.6887576461915992, "learning_rate": 1.7659831121833534e-05, "loss": 0.7857, "step": 732 }, { "epoch": 0.026548352046360014, "grad_norm": 2.983982791892724, "learning_rate": 1.7683956574185768e-05, "loss": 0.9913, "step": 733 }, { "epoch": 0.026584570807678377, "grad_norm": 2.6704153457398663, "learning_rate": 1.7708082026537998e-05, "loss": 0.9153, "step": 734 }, { "epoch": 0.02662078956899674, "grad_norm": 1.8510899760693296, "learning_rate": 1.773220747889023e-05, "loss": 0.8024, "step": 735 }, { "epoch": 0.026657008330315105, "grad_norm": 2.334012857367091, "learning_rate": 1.775633293124246e-05, "loss": 0.909, "step": 736 }, { "epoch": 0.026693227091633465, "grad_norm": 1.8181182130840676, "learning_rate": 1.7780458383594695e-05, "loss": 0.8197, "step": 737 }, { "epoch": 0.026729445852951828, "grad_norm": 1.5670313118659234, "learning_rate": 1.7804583835946925e-05, "loss": 0.7808, "step": 738 }, { "epoch": 0.026765664614270192, "grad_norm": 2.5843424184975636, "learning_rate": 1.782870928829916e-05, "loss": 0.9903, "step": 739 }, { "epoch": 0.026801883375588555, "grad_norm": 3.0168250814006115, "learning_rate": 1.785283474065139e-05, "loss": 0.94, "step": 740 }, { "epoch": 0.02683810213690692, "grad_norm": 2.2899665056459004, "learning_rate": 1.7876960193003622e-05, "loss": 0.9183, "step": 741 }, { "epoch": 0.02687432089822528, "grad_norm": 1.553304775151456, "learning_rate": 1.7901085645355852e-05, "loss": 0.755, "step": 742 }, { "epoch": 0.026910539659543643, "grad_norm": 2.397599869255058, "learning_rate": 1.7925211097708085e-05, "loss": 0.9964, "step": 743 }, { "epoch": 0.026946758420862006, "grad_norm": 2.3304648482123467, "learning_rate": 1.7949336550060316e-05, "loss": 0.9182, "step": 744 }, { "epoch": 0.02698297718218037, "grad_norm": 2.2474303008995133, "learning_rate": 1.7973462002412546e-05, "loss": 0.9075, "step": 745 }, { "epoch": 0.027019195943498733, "grad_norm": 1.5934430626216436, "learning_rate": 1.799758745476478e-05, "loss": 0.7837, "step": 746 }, { "epoch": 0.027055414704817097, "grad_norm": 2.4790756263298843, "learning_rate": 1.802171290711701e-05, "loss": 0.8322, "step": 747 }, { "epoch": 0.027091633466135457, "grad_norm": 1.5599031217545238, "learning_rate": 1.8045838359469243e-05, "loss": 0.7468, "step": 748 }, { "epoch": 0.02712785222745382, "grad_norm": 2.4052923498366545, "learning_rate": 1.8069963811821473e-05, "loss": 0.9505, "step": 749 }, { "epoch": 0.027164070988772184, "grad_norm": 2.115039548769046, "learning_rate": 1.8094089264173703e-05, "loss": 0.8288, "step": 750 }, { "epoch": 0.027200289750090548, "grad_norm": 2.2733982613706902, "learning_rate": 1.8118214716525936e-05, "loss": 0.9082, "step": 751 }, { "epoch": 0.02723650851140891, "grad_norm": 2.0912823191116012, "learning_rate": 1.8142340168878166e-05, "loss": 0.8519, "step": 752 }, { "epoch": 0.02727272727272727, "grad_norm": 1.7698420365885916, "learning_rate": 1.81664656212304e-05, "loss": 0.7802, "step": 753 }, { "epoch": 0.027308946034045635, "grad_norm": 1.4899279645321093, "learning_rate": 1.819059107358263e-05, "loss": 0.7684, "step": 754 }, { "epoch": 0.027345164795364, "grad_norm": 1.600663447918029, "learning_rate": 1.8214716525934863e-05, "loss": 0.7993, "step": 755 }, { "epoch": 0.027381383556682362, "grad_norm": 3.437483531739061, "learning_rate": 1.8238841978287094e-05, "loss": 1.0017, "step": 756 }, { "epoch": 0.027417602318000726, "grad_norm": 2.339807860365087, "learning_rate": 1.8262967430639327e-05, "loss": 0.8641, "step": 757 }, { "epoch": 0.027453821079319086, "grad_norm": 1.7160831172398678, "learning_rate": 1.8287092882991557e-05, "loss": 0.7787, "step": 758 }, { "epoch": 0.02749003984063745, "grad_norm": 2.6608652116808997, "learning_rate": 1.831121833534379e-05, "loss": 1.0315, "step": 759 }, { "epoch": 0.027526258601955813, "grad_norm": 2.440276603278392, "learning_rate": 1.833534378769602e-05, "loss": 0.878, "step": 760 }, { "epoch": 0.027562477363274177, "grad_norm": 1.6750331176534803, "learning_rate": 1.8359469240048254e-05, "loss": 0.7555, "step": 761 }, { "epoch": 0.02759869612459254, "grad_norm": 2.0932147477051917, "learning_rate": 1.8383594692400484e-05, "loss": 0.8799, "step": 762 }, { "epoch": 0.0276349148859109, "grad_norm": 1.9072206901668884, "learning_rate": 1.8407720144752718e-05, "loss": 0.766, "step": 763 }, { "epoch": 0.027671133647229264, "grad_norm": 1.5513406933235838, "learning_rate": 1.8431845597104948e-05, "loss": 0.7591, "step": 764 }, { "epoch": 0.027707352408547627, "grad_norm": 1.3967311139308456, "learning_rate": 1.845597104945718e-05, "loss": 0.7701, "step": 765 }, { "epoch": 0.02774357116986599, "grad_norm": 2.410070660460948, "learning_rate": 1.848009650180941e-05, "loss": 0.9443, "step": 766 }, { "epoch": 0.027779789931184354, "grad_norm": 2.677132021922613, "learning_rate": 1.850422195416164e-05, "loss": 0.8949, "step": 767 }, { "epoch": 0.027816008692502718, "grad_norm": 2.3490179273541196, "learning_rate": 1.8528347406513875e-05, "loss": 0.9597, "step": 768 }, { "epoch": 0.027852227453821078, "grad_norm": 2.325046566591718, "learning_rate": 1.8552472858866105e-05, "loss": 0.8978, "step": 769 }, { "epoch": 0.027888446215139442, "grad_norm": 1.8902932664919008, "learning_rate": 1.857659831121834e-05, "loss": 0.7782, "step": 770 }, { "epoch": 0.027924664976457805, "grad_norm": 2.222722869605412, "learning_rate": 1.860072376357057e-05, "loss": 0.8704, "step": 771 }, { "epoch": 0.02796088373777617, "grad_norm": 2.6867903268657813, "learning_rate": 1.86248492159228e-05, "loss": 0.8169, "step": 772 }, { "epoch": 0.027997102499094532, "grad_norm": 2.6701841804561086, "learning_rate": 1.8648974668275032e-05, "loss": 0.9112, "step": 773 }, { "epoch": 0.028033321260412893, "grad_norm": 1.546423001988422, "learning_rate": 1.8673100120627262e-05, "loss": 0.7395, "step": 774 }, { "epoch": 0.028069540021731256, "grad_norm": 2.3238121959392344, "learning_rate": 1.8697225572979492e-05, "loss": 1.0044, "step": 775 }, { "epoch": 0.02810575878304962, "grad_norm": 1.4962944090029655, "learning_rate": 1.8721351025331726e-05, "loss": 0.7899, "step": 776 }, { "epoch": 0.028141977544367983, "grad_norm": 2.0386925102223077, "learning_rate": 1.8745476477683956e-05, "loss": 0.8891, "step": 777 }, { "epoch": 0.028178196305686347, "grad_norm": 2.3016916426549536, "learning_rate": 1.876960193003619e-05, "loss": 0.946, "step": 778 }, { "epoch": 0.028214415067004707, "grad_norm": 2.4277091573570155, "learning_rate": 1.879372738238842e-05, "loss": 0.9347, "step": 779 }, { "epoch": 0.02825063382832307, "grad_norm": 1.4257938885781822, "learning_rate": 1.8817852834740653e-05, "loss": 0.7922, "step": 780 }, { "epoch": 0.028286852589641434, "grad_norm": 2.0515454385284784, "learning_rate": 1.8841978287092883e-05, "loss": 0.8009, "step": 781 }, { "epoch": 0.028323071350959798, "grad_norm": 2.3206381248202628, "learning_rate": 1.8866103739445116e-05, "loss": 1.0292, "step": 782 }, { "epoch": 0.02835929011227816, "grad_norm": 2.147141653614382, "learning_rate": 1.8890229191797347e-05, "loss": 0.9191, "step": 783 }, { "epoch": 0.02839550887359652, "grad_norm": 2.385311203329923, "learning_rate": 1.891435464414958e-05, "loss": 0.9437, "step": 784 }, { "epoch": 0.028431727634914885, "grad_norm": 2.280584310052067, "learning_rate": 1.893848009650181e-05, "loss": 0.9336, "step": 785 }, { "epoch": 0.02846794639623325, "grad_norm": 2.0880514534366865, "learning_rate": 1.8962605548854044e-05, "loss": 0.9146, "step": 786 }, { "epoch": 0.028504165157551612, "grad_norm": 2.44404252196686, "learning_rate": 1.8986731001206274e-05, "loss": 0.8949, "step": 787 }, { "epoch": 0.028540383918869976, "grad_norm": 2.344205619713199, "learning_rate": 1.9010856453558507e-05, "loss": 0.9643, "step": 788 }, { "epoch": 0.02857660268018834, "grad_norm": 1.5568734488330924, "learning_rate": 1.9034981905910737e-05, "loss": 0.7827, "step": 789 }, { "epoch": 0.0286128214415067, "grad_norm": 2.5643301627088153, "learning_rate": 1.905910735826297e-05, "loss": 0.9982, "step": 790 }, { "epoch": 0.028649040202825063, "grad_norm": 2.5281910986682217, "learning_rate": 1.90832328106152e-05, "loss": 0.9653, "step": 791 }, { "epoch": 0.028685258964143426, "grad_norm": 2.6913656486360997, "learning_rate": 1.9107358262967434e-05, "loss": 0.8555, "step": 792 }, { "epoch": 0.02872147772546179, "grad_norm": 1.6098008897242324, "learning_rate": 1.9131483715319664e-05, "loss": 0.8021, "step": 793 }, { "epoch": 0.028757696486780154, "grad_norm": 2.483763607536942, "learning_rate": 1.9155609167671894e-05, "loss": 0.8129, "step": 794 }, { "epoch": 0.028793915248098514, "grad_norm": 2.2256219650907747, "learning_rate": 1.9179734620024128e-05, "loss": 0.9252, "step": 795 }, { "epoch": 0.028830134009416877, "grad_norm": 2.341279888275439, "learning_rate": 1.9203860072376358e-05, "loss": 0.9636, "step": 796 }, { "epoch": 0.02886635277073524, "grad_norm": 2.0868959796498743, "learning_rate": 1.9227985524728588e-05, "loss": 0.9205, "step": 797 }, { "epoch": 0.028902571532053604, "grad_norm": 1.3587616033762424, "learning_rate": 1.925211097708082e-05, "loss": 0.7587, "step": 798 }, { "epoch": 0.028938790293371968, "grad_norm": 2.315895456859759, "learning_rate": 1.927623642943305e-05, "loss": 0.9474, "step": 799 }, { "epoch": 0.028975009054690328, "grad_norm": 2.2467081025867026, "learning_rate": 1.9300361881785285e-05, "loss": 0.9305, "step": 800 }, { "epoch": 0.02901122781600869, "grad_norm": 2.1678633552570212, "learning_rate": 1.9324487334137515e-05, "loss": 0.9549, "step": 801 }, { "epoch": 0.029047446577327055, "grad_norm": 2.292316797363366, "learning_rate": 1.934861278648975e-05, "loss": 0.9313, "step": 802 }, { "epoch": 0.02908366533864542, "grad_norm": 2.243129165304892, "learning_rate": 1.937273823884198e-05, "loss": 0.852, "step": 803 }, { "epoch": 0.029119884099963782, "grad_norm": 2.155047245367928, "learning_rate": 1.9396863691194212e-05, "loss": 0.8793, "step": 804 }, { "epoch": 0.029156102861282143, "grad_norm": 2.414428442467967, "learning_rate": 1.9420989143546442e-05, "loss": 0.7892, "step": 805 }, { "epoch": 0.029192321622600506, "grad_norm": 2.3657471331736524, "learning_rate": 1.9445114595898676e-05, "loss": 0.9332, "step": 806 }, { "epoch": 0.02922854038391887, "grad_norm": 1.5610973296105575, "learning_rate": 1.9469240048250906e-05, "loss": 0.7863, "step": 807 }, { "epoch": 0.029264759145237233, "grad_norm": 2.305398310946121, "learning_rate": 1.949336550060314e-05, "loss": 0.9036, "step": 808 }, { "epoch": 0.029300977906555597, "grad_norm": 2.132507896330459, "learning_rate": 1.951749095295537e-05, "loss": 0.9337, "step": 809 }, { "epoch": 0.02933719666787396, "grad_norm": 2.161724543934217, "learning_rate": 1.9541616405307603e-05, "loss": 0.9098, "step": 810 }, { "epoch": 0.02937341542919232, "grad_norm": 2.430386038083704, "learning_rate": 1.9565741857659833e-05, "loss": 1.0178, "step": 811 }, { "epoch": 0.029409634190510684, "grad_norm": 2.4959941897886306, "learning_rate": 1.9589867310012066e-05, "loss": 1.0154, "step": 812 }, { "epoch": 0.029445852951829048, "grad_norm": 2.1379026571193456, "learning_rate": 1.9613992762364297e-05, "loss": 0.9534, "step": 813 }, { "epoch": 0.02948207171314741, "grad_norm": 1.798567799962293, "learning_rate": 1.963811821471653e-05, "loss": 0.7763, "step": 814 }, { "epoch": 0.029518290474465775, "grad_norm": 1.2926023258614694, "learning_rate": 1.966224366706876e-05, "loss": 0.7707, "step": 815 }, { "epoch": 0.029554509235784135, "grad_norm": 1.6828633996294122, "learning_rate": 1.968636911942099e-05, "loss": 0.7936, "step": 816 }, { "epoch": 0.0295907279971025, "grad_norm": 2.1921935919515914, "learning_rate": 1.9710494571773224e-05, "loss": 0.9723, "step": 817 }, { "epoch": 0.029626946758420862, "grad_norm": 1.4997871248726762, "learning_rate": 1.9734620024125454e-05, "loss": 0.7608, "step": 818 }, { "epoch": 0.029663165519739226, "grad_norm": 1.8214471542583306, "learning_rate": 1.9758745476477684e-05, "loss": 0.7532, "step": 819 }, { "epoch": 0.02969938428105759, "grad_norm": 2.1044138493306868, "learning_rate": 1.9782870928829917e-05, "loss": 0.8823, "step": 820 }, { "epoch": 0.02973560304237595, "grad_norm": 2.376565804501154, "learning_rate": 1.9806996381182147e-05, "loss": 0.9124, "step": 821 }, { "epoch": 0.029771821803694313, "grad_norm": 1.5187865988621214, "learning_rate": 1.983112183353438e-05, "loss": 0.76, "step": 822 }, { "epoch": 0.029808040565012676, "grad_norm": 1.5079146020790517, "learning_rate": 1.985524728588661e-05, "loss": 0.7992, "step": 823 }, { "epoch": 0.02984425932633104, "grad_norm": 2.4590769652613687, "learning_rate": 1.9879372738238844e-05, "loss": 0.9363, "step": 824 }, { "epoch": 0.029880478087649404, "grad_norm": 2.3470180265982212, "learning_rate": 1.9903498190591075e-05, "loss": 0.9459, "step": 825 }, { "epoch": 0.029916696848967764, "grad_norm": 2.379146923925811, "learning_rate": 1.9927623642943308e-05, "loss": 0.9092, "step": 826 }, { "epoch": 0.029952915610286127, "grad_norm": 2.2006241253117365, "learning_rate": 1.9951749095295538e-05, "loss": 0.9154, "step": 827 }, { "epoch": 0.02998913437160449, "grad_norm": 2.2028992398567047, "learning_rate": 1.997587454764777e-05, "loss": 0.9141, "step": 828 }, { "epoch": 0.030025353132922854, "grad_norm": 2.272505242145743, "learning_rate": 2e-05, "loss": 0.8486, "step": 829 }, { "epoch": 0.030061571894241218, "grad_norm": 2.438207593940214, "learning_rate": 1.9999999931195607e-05, "loss": 0.8692, "step": 830 }, { "epoch": 0.03009779065555958, "grad_norm": 2.0792567921730427, "learning_rate": 1.999999972478243e-05, "loss": 0.7857, "step": 831 }, { "epoch": 0.03013400941687794, "grad_norm": 2.345391464810135, "learning_rate": 1.9999999380760465e-05, "loss": 0.6653, "step": 832 }, { "epoch": 0.030170228178196305, "grad_norm": 2.531588846650412, "learning_rate": 1.9999998899129725e-05, "loss": 0.9564, "step": 833 }, { "epoch": 0.03020644693951467, "grad_norm": 2.2660945504418653, "learning_rate": 1.9999998279890212e-05, "loss": 0.8664, "step": 834 }, { "epoch": 0.030242665700833032, "grad_norm": 2.249573402229777, "learning_rate": 1.999999752304194e-05, "loss": 0.7581, "step": 835 }, { "epoch": 0.030278884462151396, "grad_norm": 1.872480153575787, "learning_rate": 1.9999996628584913e-05, "loss": 0.8539, "step": 836 }, { "epoch": 0.030315103223469756, "grad_norm": 3.302377776301614, "learning_rate": 1.9999995596519143e-05, "loss": 0.874, "step": 837 }, { "epoch": 0.03035132198478812, "grad_norm": 1.7288946746368865, "learning_rate": 1.9999994426844648e-05, "loss": 0.8096, "step": 838 }, { "epoch": 0.030387540746106483, "grad_norm": 2.422047336661037, "learning_rate": 1.9999993119561444e-05, "loss": 1.013, "step": 839 }, { "epoch": 0.030423759507424847, "grad_norm": 2.2893708805762616, "learning_rate": 1.999999167466955e-05, "loss": 0.9364, "step": 840 }, { "epoch": 0.03045997826874321, "grad_norm": 1.5908853874396733, "learning_rate": 1.999999009216898e-05, "loss": 0.7464, "step": 841 }, { "epoch": 0.03049619703006157, "grad_norm": 1.9385085845531618, "learning_rate": 1.999998837205976e-05, "loss": 0.7751, "step": 842 }, { "epoch": 0.030532415791379934, "grad_norm": 2.4252020932324205, "learning_rate": 1.9999986514341912e-05, "loss": 1.0135, "step": 843 }, { "epoch": 0.030568634552698298, "grad_norm": 2.399314915584654, "learning_rate": 1.999998451901547e-05, "loss": 0.9121, "step": 844 }, { "epoch": 0.03060485331401666, "grad_norm": 2.156770736591773, "learning_rate": 1.9999982386080445e-05, "loss": 0.955, "step": 845 }, { "epoch": 0.030641072075335025, "grad_norm": 2.4376716278248383, "learning_rate": 1.999998011553688e-05, "loss": 0.9542, "step": 846 }, { "epoch": 0.030677290836653385, "grad_norm": 2.42760090602996, "learning_rate": 1.9999977707384802e-05, "loss": 0.8281, "step": 847 }, { "epoch": 0.03071350959797175, "grad_norm": 1.807669007247927, "learning_rate": 1.9999975161624244e-05, "loss": 0.7984, "step": 848 }, { "epoch": 0.030749728359290112, "grad_norm": 2.556893774020548, "learning_rate": 1.999997247825524e-05, "loss": 0.8936, "step": 849 }, { "epoch": 0.030785947120608476, "grad_norm": 2.4743923492526805, "learning_rate": 1.9999969657277828e-05, "loss": 0.9268, "step": 850 }, { "epoch": 0.03082216588192684, "grad_norm": 1.6520741877763963, "learning_rate": 1.999996669869205e-05, "loss": 0.792, "step": 851 }, { "epoch": 0.0308583846432452, "grad_norm": 1.5324742233748854, "learning_rate": 1.999996360249794e-05, "loss": 0.7696, "step": 852 }, { "epoch": 0.030894603404563563, "grad_norm": 2.7245375155471874, "learning_rate": 1.9999960368695545e-05, "loss": 0.957, "step": 853 }, { "epoch": 0.030930822165881926, "grad_norm": 2.4207099650373927, "learning_rate": 1.9999956997284908e-05, "loss": 0.924, "step": 854 }, { "epoch": 0.03096704092720029, "grad_norm": 2.1980036856538616, "learning_rate": 1.9999953488266074e-05, "loss": 0.9498, "step": 855 }, { "epoch": 0.031003259688518654, "grad_norm": 2.307444262069043, "learning_rate": 1.9999949841639098e-05, "loss": 1.0077, "step": 856 }, { "epoch": 0.031039478449837017, "grad_norm": 2.302516959347432, "learning_rate": 1.9999946057404023e-05, "loss": 0.8786, "step": 857 }, { "epoch": 0.031075697211155377, "grad_norm": 2.4021810678123248, "learning_rate": 1.9999942135560904e-05, "loss": 0.9893, "step": 858 }, { "epoch": 0.03111191597247374, "grad_norm": 1.8845755266273012, "learning_rate": 1.9999938076109795e-05, "loss": 0.781, "step": 859 }, { "epoch": 0.031148134733792104, "grad_norm": 2.2460736298696533, "learning_rate": 1.9999933879050747e-05, "loss": 0.9288, "step": 860 }, { "epoch": 0.031184353495110468, "grad_norm": 2.424254403757248, "learning_rate": 1.999992954438383e-05, "loss": 0.9023, "step": 861 }, { "epoch": 0.03122057225642883, "grad_norm": 2.370875027983531, "learning_rate": 1.999992507210909e-05, "loss": 0.8764, "step": 862 }, { "epoch": 0.03125679101774719, "grad_norm": 1.5577135487498999, "learning_rate": 1.9999920462226594e-05, "loss": 0.7389, "step": 863 }, { "epoch": 0.03129300977906556, "grad_norm": 2.2658464487964043, "learning_rate": 1.999991571473641e-05, "loss": 0.8891, "step": 864 }, { "epoch": 0.03132922854038392, "grad_norm": 2.3968517764879986, "learning_rate": 1.9999910829638596e-05, "loss": 0.915, "step": 865 }, { "epoch": 0.03136544730170228, "grad_norm": 2.4546205675407866, "learning_rate": 1.9999905806933223e-05, "loss": 0.9015, "step": 866 }, { "epoch": 0.031401666063020646, "grad_norm": 2.231322751341778, "learning_rate": 1.999990064662036e-05, "loss": 0.9103, "step": 867 }, { "epoch": 0.031437884824339006, "grad_norm": 2.176570947634809, "learning_rate": 1.9999895348700077e-05, "loss": 0.9962, "step": 868 }, { "epoch": 0.03147410358565737, "grad_norm": 1.9015033464509206, "learning_rate": 1.9999889913172446e-05, "loss": 0.7375, "step": 869 }, { "epoch": 0.03151032234697573, "grad_norm": 2.1638585693452193, "learning_rate": 1.9999884340037542e-05, "loss": 0.9942, "step": 870 }, { "epoch": 0.03154654110829409, "grad_norm": 2.1869143102151223, "learning_rate": 1.9999878629295445e-05, "loss": 1.0124, "step": 871 }, { "epoch": 0.03158275986961246, "grad_norm": 2.2981177117924902, "learning_rate": 1.9999872780946235e-05, "loss": 0.9923, "step": 872 }, { "epoch": 0.03161897863093082, "grad_norm": 2.590974057636368, "learning_rate": 1.9999866794989984e-05, "loss": 0.9198, "step": 873 }, { "epoch": 0.03165519739224919, "grad_norm": 2.572446018807414, "learning_rate": 1.999986067142678e-05, "loss": 0.9307, "step": 874 }, { "epoch": 0.03169141615356755, "grad_norm": 2.3882751384902736, "learning_rate": 1.9999854410256707e-05, "loss": 0.9547, "step": 875 }, { "epoch": 0.03172763491488591, "grad_norm": 2.0672651968300655, "learning_rate": 1.999984801147985e-05, "loss": 0.8936, "step": 876 }, { "epoch": 0.031763853676204275, "grad_norm": 2.395083975046441, "learning_rate": 1.99998414750963e-05, "loss": 0.883, "step": 877 }, { "epoch": 0.031800072437522635, "grad_norm": 2.1441197608534477, "learning_rate": 1.999983480110614e-05, "loss": 0.9934, "step": 878 }, { "epoch": 0.031836291198841, "grad_norm": 2.4693903394380525, "learning_rate": 1.999982798950947e-05, "loss": 0.9014, "step": 879 }, { "epoch": 0.03187250996015936, "grad_norm": 1.6512841527936601, "learning_rate": 1.9999821040306383e-05, "loss": 0.8144, "step": 880 }, { "epoch": 0.03190872872147772, "grad_norm": 4.523904107623979, "learning_rate": 1.9999813953496966e-05, "loss": 0.9116, "step": 881 }, { "epoch": 0.03194494748279609, "grad_norm": 2.1031459184161783, "learning_rate": 1.9999806729081327e-05, "loss": 1.0084, "step": 882 }, { "epoch": 0.03198116624411445, "grad_norm": 1.6624151715080235, "learning_rate": 1.9999799367059558e-05, "loss": 0.7991, "step": 883 }, { "epoch": 0.032017385005432816, "grad_norm": 2.196780224908659, "learning_rate": 1.9999791867431764e-05, "loss": 0.9468, "step": 884 }, { "epoch": 0.032053603766751176, "grad_norm": 1.9603209799205032, "learning_rate": 1.9999784230198047e-05, "loss": 0.8999, "step": 885 }, { "epoch": 0.03208982252806954, "grad_norm": 1.4930895758868374, "learning_rate": 1.999977645535851e-05, "loss": 0.8141, "step": 886 }, { "epoch": 0.032126041289387904, "grad_norm": 2.186425517361328, "learning_rate": 1.999976854291327e-05, "loss": 0.9309, "step": 887 }, { "epoch": 0.032162260050706264, "grad_norm": 1.3332762741304258, "learning_rate": 1.9999760492862422e-05, "loss": 0.715, "step": 888 }, { "epoch": 0.03219847881202463, "grad_norm": 2.512012654339069, "learning_rate": 1.9999752305206083e-05, "loss": 0.9485, "step": 889 }, { "epoch": 0.03223469757334299, "grad_norm": 2.3878493963345733, "learning_rate": 1.999974397994437e-05, "loss": 0.9245, "step": 890 }, { "epoch": 0.03227091633466136, "grad_norm": 2.21325115354403, "learning_rate": 1.999973551707739e-05, "loss": 0.9511, "step": 891 }, { "epoch": 0.03230713509597972, "grad_norm": 2.0322446056996637, "learning_rate": 1.999972691660526e-05, "loss": 0.8976, "step": 892 }, { "epoch": 0.03234335385729808, "grad_norm": 2.1624728744106534, "learning_rate": 1.9999718178528108e-05, "loss": 0.8929, "step": 893 }, { "epoch": 0.032379572618616445, "grad_norm": 2.079444385080172, "learning_rate": 1.999970930284604e-05, "loss": 0.7389, "step": 894 }, { "epoch": 0.032415791379934805, "grad_norm": 2.540130678113748, "learning_rate": 1.9999700289559193e-05, "loss": 0.8884, "step": 895 }, { "epoch": 0.03245201014125317, "grad_norm": 2.369822219330297, "learning_rate": 1.999969113866768e-05, "loss": 0.9465, "step": 896 }, { "epoch": 0.03248822890257153, "grad_norm": 1.5374593897061466, "learning_rate": 1.9999681850171627e-05, "loss": 0.758, "step": 897 }, { "epoch": 0.03252444766388989, "grad_norm": 1.9549917276767956, "learning_rate": 1.999967242407117e-05, "loss": 0.7954, "step": 898 }, { "epoch": 0.03256066642520826, "grad_norm": 1.8157811915339, "learning_rate": 1.9999662860366432e-05, "loss": 0.7309, "step": 899 }, { "epoch": 0.03259688518652662, "grad_norm": 2.3798260186971008, "learning_rate": 1.9999653159057547e-05, "loss": 0.9328, "step": 900 }, { "epoch": 0.03263310394784499, "grad_norm": 2.3571349430002977, "learning_rate": 1.999964332014465e-05, "loss": 0.8866, "step": 901 }, { "epoch": 0.03266932270916335, "grad_norm": 1.6167867671957503, "learning_rate": 1.999963334362787e-05, "loss": 0.7518, "step": 902 }, { "epoch": 0.03270554147048171, "grad_norm": 1.6968185312088608, "learning_rate": 1.9999623229507352e-05, "loss": 0.7895, "step": 903 }, { "epoch": 0.032741760231800074, "grad_norm": 2.303491590858691, "learning_rate": 1.999961297778323e-05, "loss": 0.8818, "step": 904 }, { "epoch": 0.032777978993118434, "grad_norm": 2.1623216505869514, "learning_rate": 1.9999602588455648e-05, "loss": 0.8256, "step": 905 }, { "epoch": 0.0328141977544368, "grad_norm": 2.5862358474344216, "learning_rate": 1.999959206152475e-05, "loss": 0.924, "step": 906 }, { "epoch": 0.03285041651575516, "grad_norm": 2.2862402537509916, "learning_rate": 1.9999581396990673e-05, "loss": 0.9476, "step": 907 }, { "epoch": 0.03288663527707352, "grad_norm": 2.1375495421791846, "learning_rate": 1.9999570594853575e-05, "loss": 0.9209, "step": 908 }, { "epoch": 0.03292285403839189, "grad_norm": 2.179066644798942, "learning_rate": 1.99995596551136e-05, "loss": 0.9157, "step": 909 }, { "epoch": 0.03295907279971025, "grad_norm": 2.538266288922225, "learning_rate": 1.9999548577770892e-05, "loss": 0.8857, "step": 910 }, { "epoch": 0.032995291561028615, "grad_norm": 1.8391447106008891, "learning_rate": 1.9999537362825613e-05, "loss": 0.7304, "step": 911 }, { "epoch": 0.033031510322346976, "grad_norm": 2.198787569019998, "learning_rate": 1.9999526010277916e-05, "loss": 0.9226, "step": 912 }, { "epoch": 0.033067729083665336, "grad_norm": 2.2130967134296347, "learning_rate": 1.9999514520127953e-05, "loss": 0.9648, "step": 913 }, { "epoch": 0.0331039478449837, "grad_norm": 1.6391484612691722, "learning_rate": 1.999950289237588e-05, "loss": 0.7674, "step": 914 }, { "epoch": 0.03314016660630206, "grad_norm": 2.2426232635574674, "learning_rate": 1.9999491127021863e-05, "loss": 0.9883, "step": 915 }, { "epoch": 0.03317638536762043, "grad_norm": 2.3266203117361286, "learning_rate": 1.9999479224066066e-05, "loss": 1.0032, "step": 916 }, { "epoch": 0.03321260412893879, "grad_norm": 1.4798460160382225, "learning_rate": 1.999946718350864e-05, "loss": 0.7705, "step": 917 }, { "epoch": 0.03324882289025715, "grad_norm": 2.357485244897964, "learning_rate": 1.9999455005349766e-05, "loss": 0.9396, "step": 918 }, { "epoch": 0.03328504165157552, "grad_norm": 2.262326084876567, "learning_rate": 1.9999442689589602e-05, "loss": 0.9381, "step": 919 }, { "epoch": 0.03332126041289388, "grad_norm": 2.2667797730224892, "learning_rate": 1.9999430236228323e-05, "loss": 0.8481, "step": 920 }, { "epoch": 0.033357479174212244, "grad_norm": 2.0445158481564394, "learning_rate": 1.9999417645266098e-05, "loss": 0.9366, "step": 921 }, { "epoch": 0.033393697935530604, "grad_norm": 2.2046486121167317, "learning_rate": 1.9999404916703096e-05, "loss": 0.9386, "step": 922 }, { "epoch": 0.033429916696848964, "grad_norm": 1.5513278905852432, "learning_rate": 1.99993920505395e-05, "loss": 0.8203, "step": 923 }, { "epoch": 0.03346613545816733, "grad_norm": 2.3736248662522583, "learning_rate": 1.999937904677548e-05, "loss": 0.9517, "step": 924 }, { "epoch": 0.03350235421948569, "grad_norm": 1.5996164366428833, "learning_rate": 1.9999365905411218e-05, "loss": 0.7529, "step": 925 }, { "epoch": 0.03353857298080406, "grad_norm": 2.2989112657471407, "learning_rate": 1.9999352626446898e-05, "loss": 0.9509, "step": 926 }, { "epoch": 0.03357479174212242, "grad_norm": 2.2548501783465302, "learning_rate": 1.99993392098827e-05, "loss": 0.9776, "step": 927 }, { "epoch": 0.033611010503440786, "grad_norm": 2.262560959384653, "learning_rate": 1.9999325655718805e-05, "loss": 0.9163, "step": 928 }, { "epoch": 0.033647229264759146, "grad_norm": 2.3177958641597023, "learning_rate": 1.99993119639554e-05, "loss": 0.9916, "step": 929 }, { "epoch": 0.033683448026077506, "grad_norm": 2.2200428795647036, "learning_rate": 1.9999298134592683e-05, "loss": 0.8821, "step": 930 }, { "epoch": 0.03371966678739587, "grad_norm": 2.2903025851696763, "learning_rate": 1.9999284167630835e-05, "loss": 0.839, "step": 931 }, { "epoch": 0.03375588554871423, "grad_norm": 2.058364606587067, "learning_rate": 1.9999270063070048e-05, "loss": 0.952, "step": 932 }, { "epoch": 0.0337921043100326, "grad_norm": 1.9452528475228987, "learning_rate": 1.9999255820910523e-05, "loss": 0.9147, "step": 933 }, { "epoch": 0.03382832307135096, "grad_norm": 2.2929295157954, "learning_rate": 1.999924144115245e-05, "loss": 0.9023, "step": 934 }, { "epoch": 0.03386454183266932, "grad_norm": 2.149275574648793, "learning_rate": 1.9999226923796027e-05, "loss": 0.931, "step": 935 }, { "epoch": 0.03390076059398769, "grad_norm": 2.0330663098155886, "learning_rate": 1.999921226884146e-05, "loss": 0.8933, "step": 936 }, { "epoch": 0.03393697935530605, "grad_norm": 2.230532489941463, "learning_rate": 1.9999197476288942e-05, "loss": 0.9291, "step": 937 }, { "epoch": 0.033973198116624415, "grad_norm": 1.5779707235403662, "learning_rate": 1.9999182546138683e-05, "loss": 0.7928, "step": 938 }, { "epoch": 0.034009416877942775, "grad_norm": 2.3436918187867257, "learning_rate": 1.9999167478390883e-05, "loss": 0.9332, "step": 939 }, { "epoch": 0.034045635639261135, "grad_norm": 2.7000398008185753, "learning_rate": 1.9999152273045755e-05, "loss": 0.916, "step": 940 }, { "epoch": 0.0340818544005795, "grad_norm": 2.1680659332487937, "learning_rate": 1.9999136930103506e-05, "loss": 0.9043, "step": 941 }, { "epoch": 0.03411807316189786, "grad_norm": 2.1617294105153877, "learning_rate": 1.9999121449564347e-05, "loss": 0.9555, "step": 942 }, { "epoch": 0.03415429192321623, "grad_norm": 1.7539228070036208, "learning_rate": 1.999910583142849e-05, "loss": 0.7842, "step": 943 }, { "epoch": 0.03419051068453459, "grad_norm": 2.1811631379416556, "learning_rate": 1.9999090075696152e-05, "loss": 0.8972, "step": 944 }, { "epoch": 0.03422672944585295, "grad_norm": 1.5911668071688598, "learning_rate": 1.9999074182367547e-05, "loss": 0.7317, "step": 945 }, { "epoch": 0.034262948207171316, "grad_norm": 1.618186379015769, "learning_rate": 1.9999058151442892e-05, "loss": 0.839, "step": 946 }, { "epoch": 0.034299166968489676, "grad_norm": 2.2331782517274896, "learning_rate": 1.9999041982922418e-05, "loss": 0.8947, "step": 947 }, { "epoch": 0.03433538572980804, "grad_norm": 2.478072695763409, "learning_rate": 1.9999025676806333e-05, "loss": 0.9826, "step": 948 }, { "epoch": 0.034371604491126403, "grad_norm": 2.3115072455280257, "learning_rate": 1.9999009233094876e-05, "loss": 0.9704, "step": 949 }, { "epoch": 0.034407823252444764, "grad_norm": 2.267031417248269, "learning_rate": 1.9998992651788262e-05, "loss": 0.9414, "step": 950 }, { "epoch": 0.03444404201376313, "grad_norm": 2.09635882860854, "learning_rate": 1.9998975932886724e-05, "loss": 0.9574, "step": 951 }, { "epoch": 0.03448026077508149, "grad_norm": 1.977213631685314, "learning_rate": 1.999895907639049e-05, "loss": 0.9756, "step": 952 }, { "epoch": 0.03451647953639986, "grad_norm": 2.1574938320583392, "learning_rate": 1.9998942082299795e-05, "loss": 0.8837, "step": 953 }, { "epoch": 0.03455269829771822, "grad_norm": 2.2925598228297313, "learning_rate": 1.999892495061487e-05, "loss": 0.9305, "step": 954 }, { "epoch": 0.03458891705903658, "grad_norm": 1.9740556682017356, "learning_rate": 1.9998907681335954e-05, "loss": 0.8335, "step": 955 }, { "epoch": 0.034625135820354945, "grad_norm": 2.0703124694104464, "learning_rate": 1.999889027446328e-05, "loss": 0.9212, "step": 956 }, { "epoch": 0.034661354581673305, "grad_norm": 2.2030640794062024, "learning_rate": 1.999887272999709e-05, "loss": 0.8477, "step": 957 }, { "epoch": 0.03469757334299167, "grad_norm": 1.479426359649428, "learning_rate": 1.9998855047937626e-05, "loss": 0.7628, "step": 958 }, { "epoch": 0.03473379210431003, "grad_norm": 2.812656133344559, "learning_rate": 1.9998837228285134e-05, "loss": 0.8918, "step": 959 }, { "epoch": 0.03477001086562839, "grad_norm": 2.2897725533799007, "learning_rate": 1.999881927103985e-05, "loss": 0.9189, "step": 960 }, { "epoch": 0.03480622962694676, "grad_norm": 2.436647073364825, "learning_rate": 1.999880117620203e-05, "loss": 0.8731, "step": 961 }, { "epoch": 0.03484244838826512, "grad_norm": 2.4441508407098835, "learning_rate": 1.9998782943771923e-05, "loss": 0.9045, "step": 962 }, { "epoch": 0.03487866714958349, "grad_norm": 2.5588450563205067, "learning_rate": 1.999876457374977e-05, "loss": 0.8498, "step": 963 }, { "epoch": 0.03491488591090185, "grad_norm": 2.2780255921316583, "learning_rate": 1.9998746066135836e-05, "loss": 0.8979, "step": 964 }, { "epoch": 0.03495110467222021, "grad_norm": 2.098894480159181, "learning_rate": 1.9998727420930373e-05, "loss": 0.8852, "step": 965 }, { "epoch": 0.034987323433538574, "grad_norm": 2.5356821475736337, "learning_rate": 1.999870863813363e-05, "loss": 1.0026, "step": 966 }, { "epoch": 0.035023542194856934, "grad_norm": 2.248662161786726, "learning_rate": 1.9998689717745872e-05, "loss": 0.9649, "step": 967 }, { "epoch": 0.0350597609561753, "grad_norm": 2.0420608226200176, "learning_rate": 1.999867065976736e-05, "loss": 0.8898, "step": 968 }, { "epoch": 0.03509597971749366, "grad_norm": 2.348295029679659, "learning_rate": 1.999865146419835e-05, "loss": 0.9055, "step": 969 }, { "epoch": 0.03513219847881202, "grad_norm": 1.8412347161436988, "learning_rate": 1.999863213103911e-05, "loss": 0.792, "step": 970 }, { "epoch": 0.03516841724013039, "grad_norm": 2.154922245003337, "learning_rate": 1.9998612660289912e-05, "loss": 1.0089, "step": 971 }, { "epoch": 0.03520463600144875, "grad_norm": 2.292647457845589, "learning_rate": 1.9998593051951014e-05, "loss": 0.9823, "step": 972 }, { "epoch": 0.035240854762767115, "grad_norm": 1.8483493074178974, "learning_rate": 1.9998573306022692e-05, "loss": 0.8489, "step": 973 }, { "epoch": 0.035277073524085475, "grad_norm": 2.2619869817612956, "learning_rate": 1.9998553422505215e-05, "loss": 0.8599, "step": 974 }, { "epoch": 0.03531329228540384, "grad_norm": 2.068181363443285, "learning_rate": 1.9998533401398856e-05, "loss": 0.9797, "step": 975 }, { "epoch": 0.0353495110467222, "grad_norm": 1.783537582105397, "learning_rate": 1.9998513242703895e-05, "loss": 0.856, "step": 976 }, { "epoch": 0.03538572980804056, "grad_norm": 2.1225594107461974, "learning_rate": 1.99984929464206e-05, "loss": 0.9183, "step": 977 }, { "epoch": 0.03542194856935893, "grad_norm": 2.136597885687692, "learning_rate": 1.9998472512549264e-05, "loss": 0.7831, "step": 978 }, { "epoch": 0.03545816733067729, "grad_norm": 2.112998206597536, "learning_rate": 1.9998451941090157e-05, "loss": 0.9744, "step": 979 }, { "epoch": 0.03549438609199566, "grad_norm": 2.2220537512136, "learning_rate": 1.9998431232043568e-05, "loss": 0.8797, "step": 980 }, { "epoch": 0.03553060485331402, "grad_norm": 2.067081917843843, "learning_rate": 1.9998410385409778e-05, "loss": 0.8834, "step": 981 }, { "epoch": 0.03556682361463238, "grad_norm": 2.0188881386891557, "learning_rate": 1.9998389401189076e-05, "loss": 0.8866, "step": 982 }, { "epoch": 0.035603042375950744, "grad_norm": 2.194645495240217, "learning_rate": 1.999836827938175e-05, "loss": 0.8848, "step": 983 }, { "epoch": 0.035639261137269104, "grad_norm": 1.8555609198314928, "learning_rate": 1.9998347019988094e-05, "loss": 0.916, "step": 984 }, { "epoch": 0.03567547989858747, "grad_norm": 2.1281044517552963, "learning_rate": 1.9998325623008396e-05, "loss": 0.943, "step": 985 }, { "epoch": 0.03571169865990583, "grad_norm": 2.1003347900231177, "learning_rate": 1.999830408844295e-05, "loss": 0.8722, "step": 986 }, { "epoch": 0.03574791742122419, "grad_norm": 2.0309137121005745, "learning_rate": 1.9998282416292057e-05, "loss": 0.9784, "step": 987 }, { "epoch": 0.03578413618254256, "grad_norm": 2.060240851181595, "learning_rate": 1.9998260606556013e-05, "loss": 0.9152, "step": 988 }, { "epoch": 0.03582035494386092, "grad_norm": 2.1550477155566226, "learning_rate": 1.9998238659235113e-05, "loss": 0.9137, "step": 989 }, { "epoch": 0.035856573705179286, "grad_norm": 1.9297644275513155, "learning_rate": 1.9998216574329668e-05, "loss": 0.8715, "step": 990 }, { "epoch": 0.035892792466497646, "grad_norm": 2.3616805180449876, "learning_rate": 1.9998194351839978e-05, "loss": 0.9655, "step": 991 }, { "epoch": 0.035929011227816006, "grad_norm": 1.975365868312018, "learning_rate": 1.999817199176635e-05, "loss": 0.8476, "step": 992 }, { "epoch": 0.03596522998913437, "grad_norm": 2.1803965013617495, "learning_rate": 1.9998149494109085e-05, "loss": 0.9199, "step": 993 }, { "epoch": 0.03600144875045273, "grad_norm": 2.0838346828649903, "learning_rate": 1.99981268588685e-05, "loss": 0.8642, "step": 994 }, { "epoch": 0.0360376675117711, "grad_norm": 2.0245839410886677, "learning_rate": 1.999810408604491e-05, "loss": 0.8126, "step": 995 }, { "epoch": 0.03607388627308946, "grad_norm": 2.5817741577689324, "learning_rate": 1.9998081175638616e-05, "loss": 0.9245, "step": 996 }, { "epoch": 0.03611010503440782, "grad_norm": 2.1194016326630365, "learning_rate": 1.999805812764994e-05, "loss": 0.8996, "step": 997 }, { "epoch": 0.03614632379572619, "grad_norm": 2.307588061303464, "learning_rate": 1.9998034942079202e-05, "loss": 0.988, "step": 998 }, { "epoch": 0.03618254255704455, "grad_norm": 2.023958792031874, "learning_rate": 1.9998011618926714e-05, "loss": 0.7927, "step": 999 }, { "epoch": 0.036218761318362915, "grad_norm": 2.6295940227859966, "learning_rate": 1.9997988158192806e-05, "loss": 0.9101, "step": 1000 }, { "epoch": 0.036254980079681275, "grad_norm": 2.218066451059634, "learning_rate": 1.9997964559877796e-05, "loss": 0.9688, "step": 1001 }, { "epoch": 0.036291198840999635, "grad_norm": 2.3329063032572672, "learning_rate": 1.9997940823982006e-05, "loss": 0.9457, "step": 1002 }, { "epoch": 0.036327417602318, "grad_norm": 1.8669682916747592, "learning_rate": 1.9997916950505765e-05, "loss": 0.7796, "step": 1003 }, { "epoch": 0.03636363636363636, "grad_norm": 2.390956903929096, "learning_rate": 1.9997892939449404e-05, "loss": 1.013, "step": 1004 }, { "epoch": 0.03639985512495473, "grad_norm": 2.2466515487000476, "learning_rate": 1.999786879081325e-05, "loss": 0.8854, "step": 1005 }, { "epoch": 0.03643607388627309, "grad_norm": 2.854112078441673, "learning_rate": 1.9997844504597636e-05, "loss": 0.9015, "step": 1006 }, { "epoch": 0.03647229264759145, "grad_norm": 1.6964216625191249, "learning_rate": 1.9997820080802896e-05, "loss": 0.785, "step": 1007 }, { "epoch": 0.036508511408909816, "grad_norm": 2.4346174609671807, "learning_rate": 1.9997795519429372e-05, "loss": 0.954, "step": 1008 }, { "epoch": 0.036544730170228176, "grad_norm": 2.0912652715933686, "learning_rate": 1.999777082047739e-05, "loss": 0.8968, "step": 1009 }, { "epoch": 0.03658094893154654, "grad_norm": 2.10618790639704, "learning_rate": 1.99977459839473e-05, "loss": 0.9411, "step": 1010 }, { "epoch": 0.0366171676928649, "grad_norm": 1.9359721064072262, "learning_rate": 1.999772100983944e-05, "loss": 0.952, "step": 1011 }, { "epoch": 0.036653386454183264, "grad_norm": 2.1800015378826845, "learning_rate": 1.9997695898154153e-05, "loss": 0.9514, "step": 1012 }, { "epoch": 0.03668960521550163, "grad_norm": 2.0452148479029435, "learning_rate": 1.999767064889179e-05, "loss": 0.8738, "step": 1013 }, { "epoch": 0.03672582397681999, "grad_norm": 1.7204857030173937, "learning_rate": 1.999764526205269e-05, "loss": 0.8445, "step": 1014 }, { "epoch": 0.03676204273813836, "grad_norm": 2.16258021313404, "learning_rate": 1.9997619737637207e-05, "loss": 0.8663, "step": 1015 }, { "epoch": 0.03679826149945672, "grad_norm": 1.4869520779660836, "learning_rate": 1.9997594075645694e-05, "loss": 0.7919, "step": 1016 }, { "epoch": 0.036834480260775085, "grad_norm": 1.5983569509278641, "learning_rate": 1.99975682760785e-05, "loss": 0.7852, "step": 1017 }, { "epoch": 0.036870699022093445, "grad_norm": 2.5153717008266794, "learning_rate": 1.9997542338935983e-05, "loss": 0.9099, "step": 1018 }, { "epoch": 0.036906917783411805, "grad_norm": 2.531972946724539, "learning_rate": 1.9997516264218494e-05, "loss": 0.9888, "step": 1019 }, { "epoch": 0.03694313654473017, "grad_norm": 2.5838974024753867, "learning_rate": 1.99974900519264e-05, "loss": 0.9852, "step": 1020 }, { "epoch": 0.03697935530604853, "grad_norm": 2.0994510555459414, "learning_rate": 1.999746370206006e-05, "loss": 0.9404, "step": 1021 }, { "epoch": 0.0370155740673669, "grad_norm": 1.5805315023273983, "learning_rate": 1.9997437214619837e-05, "loss": 0.8531, "step": 1022 }, { "epoch": 0.03705179282868526, "grad_norm": 2.1712994430361108, "learning_rate": 1.999741058960609e-05, "loss": 0.9571, "step": 1023 }, { "epoch": 0.03708801159000362, "grad_norm": 1.4756572004100195, "learning_rate": 1.9997383827019188e-05, "loss": 0.8236, "step": 1024 }, { "epoch": 0.037124230351321987, "grad_norm": 1.5917580114195666, "learning_rate": 1.9997356926859503e-05, "loss": 0.8484, "step": 1025 }, { "epoch": 0.03716044911264035, "grad_norm": 2.345673737989024, "learning_rate": 1.9997329889127397e-05, "loss": 0.9524, "step": 1026 }, { "epoch": 0.037196667873958714, "grad_norm": 2.1389383480986144, "learning_rate": 1.999730271382325e-05, "loss": 0.9779, "step": 1027 }, { "epoch": 0.037232886635277074, "grad_norm": 2.163473637885307, "learning_rate": 1.9997275400947437e-05, "loss": 0.9492, "step": 1028 }, { "epoch": 0.037269105396595434, "grad_norm": 2.5680593322593808, "learning_rate": 1.9997247950500327e-05, "loss": 0.8964, "step": 1029 }, { "epoch": 0.0373053241579138, "grad_norm": 1.83526561203841, "learning_rate": 1.9997220362482302e-05, "loss": 0.825, "step": 1030 }, { "epoch": 0.03734154291923216, "grad_norm": 2.3843290474539516, "learning_rate": 1.999719263689374e-05, "loss": 1.0514, "step": 1031 }, { "epoch": 0.03737776168055053, "grad_norm": 2.711904661047591, "learning_rate": 1.999716477373502e-05, "loss": 0.9574, "step": 1032 }, { "epoch": 0.03741398044186889, "grad_norm": 2.150806400666096, "learning_rate": 1.999713677300653e-05, "loss": 0.8889, "step": 1033 }, { "epoch": 0.03745019920318725, "grad_norm": 1.6792250803458708, "learning_rate": 1.9997108634708656e-05, "loss": 0.7782, "step": 1034 }, { "epoch": 0.037486417964505615, "grad_norm": 2.315530878540291, "learning_rate": 1.9997080358841782e-05, "loss": 1.0103, "step": 1035 }, { "epoch": 0.037522636725823975, "grad_norm": 1.5418249590544646, "learning_rate": 1.9997051945406295e-05, "loss": 0.8086, "step": 1036 }, { "epoch": 0.03755885548714234, "grad_norm": 2.434754062884018, "learning_rate": 1.9997023394402594e-05, "loss": 0.9259, "step": 1037 }, { "epoch": 0.0375950742484607, "grad_norm": 2.6480680534347596, "learning_rate": 1.9996994705831063e-05, "loss": 0.8925, "step": 1038 }, { "epoch": 0.03763129300977906, "grad_norm": 2.1155164799967983, "learning_rate": 1.99969658796921e-05, "loss": 0.8753, "step": 1039 }, { "epoch": 0.03766751177109743, "grad_norm": 2.168119602281873, "learning_rate": 1.9996936915986108e-05, "loss": 0.942, "step": 1040 }, { "epoch": 0.03770373053241579, "grad_norm": 2.3656272498892363, "learning_rate": 1.9996907814713474e-05, "loss": 0.9366, "step": 1041 }, { "epoch": 0.03773994929373416, "grad_norm": 2.3454827833273626, "learning_rate": 1.999687857587461e-05, "loss": 0.9207, "step": 1042 }, { "epoch": 0.03777616805505252, "grad_norm": 2.155679250589175, "learning_rate": 1.999684919946991e-05, "loss": 0.8782, "step": 1043 }, { "epoch": 0.03781238681637088, "grad_norm": 2.196089474712215, "learning_rate": 1.999681968549978e-05, "loss": 0.8999, "step": 1044 }, { "epoch": 0.037848605577689244, "grad_norm": 2.158262217476482, "learning_rate": 1.999679003396463e-05, "loss": 0.8344, "step": 1045 }, { "epoch": 0.037884824339007604, "grad_norm": 2.1557382515378456, "learning_rate": 1.999676024486486e-05, "loss": 0.9482, "step": 1046 }, { "epoch": 0.03792104310032597, "grad_norm": 2.0488270631392846, "learning_rate": 1.999673031820089e-05, "loss": 0.8738, "step": 1047 }, { "epoch": 0.03795726186164433, "grad_norm": 2.289735631625591, "learning_rate": 1.9996700253973123e-05, "loss": 0.8857, "step": 1048 }, { "epoch": 0.03799348062296269, "grad_norm": 2.371612177926798, "learning_rate": 1.999667005218198e-05, "loss": 0.9604, "step": 1049 }, { "epoch": 0.03802969938428106, "grad_norm": 2.2297538573690106, "learning_rate": 1.9996639712827872e-05, "loss": 0.9466, "step": 1050 }, { "epoch": 0.03806591814559942, "grad_norm": 2.214330940152788, "learning_rate": 1.9996609235911218e-05, "loss": 0.8913, "step": 1051 }, { "epoch": 0.038102136906917786, "grad_norm": 2.1285886895151025, "learning_rate": 1.9996578621432438e-05, "loss": 0.8332, "step": 1052 }, { "epoch": 0.038138355668236146, "grad_norm": 2.348951077855331, "learning_rate": 1.999654786939195e-05, "loss": 0.9652, "step": 1053 }, { "epoch": 0.038174574429554506, "grad_norm": 2.1768407656102178, "learning_rate": 1.999651697979018e-05, "loss": 0.935, "step": 1054 }, { "epoch": 0.03821079319087287, "grad_norm": 2.0353271802483786, "learning_rate": 1.9996485952627554e-05, "loss": 0.9235, "step": 1055 }, { "epoch": 0.03824701195219123, "grad_norm": 2.615795366363672, "learning_rate": 1.99964547879045e-05, "loss": 0.7807, "step": 1056 }, { "epoch": 0.0382832307135096, "grad_norm": 2.5077124097345753, "learning_rate": 1.9996423485621438e-05, "loss": 0.848, "step": 1057 }, { "epoch": 0.03831944947482796, "grad_norm": 2.2432215130647872, "learning_rate": 1.999639204577881e-05, "loss": 0.9645, "step": 1058 }, { "epoch": 0.03835566823614633, "grad_norm": 2.0173482274615995, "learning_rate": 1.9996360468377044e-05, "loss": 0.9283, "step": 1059 }, { "epoch": 0.03839188699746469, "grad_norm": 1.4866087889918556, "learning_rate": 1.9996328753416572e-05, "loss": 0.7729, "step": 1060 }, { "epoch": 0.03842810575878305, "grad_norm": 2.280610146668861, "learning_rate": 1.9996296900897835e-05, "loss": 0.9738, "step": 1061 }, { "epoch": 0.038464324520101414, "grad_norm": 2.1949315741807225, "learning_rate": 1.9996264910821266e-05, "loss": 0.9278, "step": 1062 }, { "epoch": 0.038500543281419775, "grad_norm": 1.9558768665315542, "learning_rate": 1.999623278318731e-05, "loss": 0.9134, "step": 1063 }, { "epoch": 0.03853676204273814, "grad_norm": 2.079781639209709, "learning_rate": 1.999620051799641e-05, "loss": 0.9646, "step": 1064 }, { "epoch": 0.0385729808040565, "grad_norm": 1.4249296909727902, "learning_rate": 1.9996168115249003e-05, "loss": 0.7821, "step": 1065 }, { "epoch": 0.03860919956537486, "grad_norm": 2.131625098865191, "learning_rate": 1.9996135574945543e-05, "loss": 0.8108, "step": 1066 }, { "epoch": 0.03864541832669323, "grad_norm": 2.2561425464805267, "learning_rate": 1.999610289708647e-05, "loss": 0.9423, "step": 1067 }, { "epoch": 0.03868163708801159, "grad_norm": 2.098871049809449, "learning_rate": 1.999607008167224e-05, "loss": 0.9415, "step": 1068 }, { "epoch": 0.038717855849329956, "grad_norm": 2.7088732569429577, "learning_rate": 1.9996037128703306e-05, "loss": 0.9206, "step": 1069 }, { "epoch": 0.038754074610648316, "grad_norm": 2.106621760944015, "learning_rate": 1.9996004038180112e-05, "loss": 0.9479, "step": 1070 }, { "epoch": 0.038790293371966676, "grad_norm": 2.163716097419576, "learning_rate": 1.9995970810103125e-05, "loss": 0.9347, "step": 1071 }, { "epoch": 0.03882651213328504, "grad_norm": 2.3055548377463615, "learning_rate": 1.999593744447279e-05, "loss": 0.8371, "step": 1072 }, { "epoch": 0.0388627308946034, "grad_norm": 2.0195381193948045, "learning_rate": 1.9995903941289575e-05, "loss": 0.8944, "step": 1073 }, { "epoch": 0.03889894965592177, "grad_norm": 2.3147503722324765, "learning_rate": 1.999587030055394e-05, "loss": 0.9043, "step": 1074 }, { "epoch": 0.03893516841724013, "grad_norm": 2.1415831247171813, "learning_rate": 1.9995836522266347e-05, "loss": 0.951, "step": 1075 }, { "epoch": 0.03897138717855849, "grad_norm": 2.036700433517757, "learning_rate": 1.999580260642726e-05, "loss": 0.8847, "step": 1076 }, { "epoch": 0.03900760593987686, "grad_norm": 2.081504835745472, "learning_rate": 1.9995768553037144e-05, "loss": 0.7933, "step": 1077 }, { "epoch": 0.03904382470119522, "grad_norm": 1.8133141481372783, "learning_rate": 1.999573436209647e-05, "loss": 0.9418, "step": 1078 }, { "epoch": 0.039080043462513585, "grad_norm": 2.026968726529817, "learning_rate": 1.999570003360571e-05, "loss": 0.9545, "step": 1079 }, { "epoch": 0.039116262223831945, "grad_norm": 2.024930342532885, "learning_rate": 1.9995665567565336e-05, "loss": 0.9346, "step": 1080 }, { "epoch": 0.039152480985150305, "grad_norm": 2.1457341230630185, "learning_rate": 1.9995630963975818e-05, "loss": 0.854, "step": 1081 }, { "epoch": 0.03918869974646867, "grad_norm": 1.812403538106059, "learning_rate": 1.9995596222837636e-05, "loss": 0.9074, "step": 1082 }, { "epoch": 0.03922491850778703, "grad_norm": 2.126288207609022, "learning_rate": 1.9995561344151264e-05, "loss": 0.8809, "step": 1083 }, { "epoch": 0.0392611372691054, "grad_norm": 2.1187632070821802, "learning_rate": 1.9995526327917187e-05, "loss": 0.8606, "step": 1084 }, { "epoch": 0.03929735603042376, "grad_norm": 2.0640672379005456, "learning_rate": 1.9995491174135887e-05, "loss": 0.9821, "step": 1085 }, { "epoch": 0.03933357479174212, "grad_norm": 4.693184136156419, "learning_rate": 1.9995455882807847e-05, "loss": 0.8896, "step": 1086 }, { "epoch": 0.039369793553060486, "grad_norm": 2.0123882910944535, "learning_rate": 1.9995420453933545e-05, "loss": 0.8623, "step": 1087 }, { "epoch": 0.03940601231437885, "grad_norm": 2.073722754909395, "learning_rate": 1.999538488751348e-05, "loss": 0.8847, "step": 1088 }, { "epoch": 0.039442231075697214, "grad_norm": 2.1340396236423036, "learning_rate": 1.9995349183548135e-05, "loss": 0.9211, "step": 1089 }, { "epoch": 0.039478449837015574, "grad_norm": 2.0322176939052663, "learning_rate": 1.9995313342038e-05, "loss": 0.8697, "step": 1090 }, { "epoch": 0.039514668598333934, "grad_norm": 2.18659302607991, "learning_rate": 1.9995277362983574e-05, "loss": 0.9896, "step": 1091 }, { "epoch": 0.0395508873596523, "grad_norm": 2.005257387097949, "learning_rate": 1.9995241246385347e-05, "loss": 0.927, "step": 1092 }, { "epoch": 0.03958710612097066, "grad_norm": 1.7935312062603739, "learning_rate": 1.999520499224382e-05, "loss": 0.8179, "step": 1093 }, { "epoch": 0.03962332488228903, "grad_norm": 2.2817467808335548, "learning_rate": 1.9995168600559486e-05, "loss": 0.9628, "step": 1094 }, { "epoch": 0.03965954364360739, "grad_norm": 2.2507679468163624, "learning_rate": 1.999513207133285e-05, "loss": 0.9043, "step": 1095 }, { "epoch": 0.03969576240492575, "grad_norm": 1.8984087821790883, "learning_rate": 1.9995095404564417e-05, "loss": 0.8663, "step": 1096 }, { "epoch": 0.039731981166244115, "grad_norm": 2.2307853474037977, "learning_rate": 1.999505860025469e-05, "loss": 0.8226, "step": 1097 }, { "epoch": 0.039768199927562475, "grad_norm": 2.497990778337168, "learning_rate": 1.9995021658404173e-05, "loss": 0.9725, "step": 1098 }, { "epoch": 0.03980441868888084, "grad_norm": 2.11857450104412, "learning_rate": 1.9994984579013373e-05, "loss": 0.947, "step": 1099 }, { "epoch": 0.0398406374501992, "grad_norm": 2.06469436073347, "learning_rate": 1.9994947362082806e-05, "loss": 0.8606, "step": 1100 }, { "epoch": 0.03987685621151757, "grad_norm": 2.125925979047536, "learning_rate": 1.9994910007612978e-05, "loss": 0.8756, "step": 1101 }, { "epoch": 0.03991307497283593, "grad_norm": 2.2301747901668083, "learning_rate": 1.999487251560441e-05, "loss": 0.9088, "step": 1102 }, { "epoch": 0.03994929373415429, "grad_norm": 1.543442704044643, "learning_rate": 1.999483488605761e-05, "loss": 0.7758, "step": 1103 }, { "epoch": 0.03998551249547266, "grad_norm": 2.2222047943973284, "learning_rate": 1.9994797118973098e-05, "loss": 0.9623, "step": 1104 }, { "epoch": 0.04002173125679102, "grad_norm": 2.1730659499112077, "learning_rate": 1.99947592143514e-05, "loss": 0.9528, "step": 1105 }, { "epoch": 0.040057950018109384, "grad_norm": 2.2718104180151695, "learning_rate": 1.999472117219303e-05, "loss": 0.879, "step": 1106 }, { "epoch": 0.040094168779427744, "grad_norm": 2.010411624869183, "learning_rate": 1.9994682992498517e-05, "loss": 0.818, "step": 1107 }, { "epoch": 0.040130387540746104, "grad_norm": 2.0082208110636754, "learning_rate": 1.9994644675268384e-05, "loss": 0.9416, "step": 1108 }, { "epoch": 0.04016660630206447, "grad_norm": 2.174662799086458, "learning_rate": 1.9994606220503157e-05, "loss": 0.9726, "step": 1109 }, { "epoch": 0.04020282506338283, "grad_norm": 2.2267838401534714, "learning_rate": 1.9994567628203367e-05, "loss": 0.918, "step": 1110 }, { "epoch": 0.0402390438247012, "grad_norm": 2.185340742083907, "learning_rate": 1.9994528898369544e-05, "loss": 0.9345, "step": 1111 }, { "epoch": 0.04027526258601956, "grad_norm": 2.176352294142416, "learning_rate": 1.999449003100222e-05, "loss": 0.8731, "step": 1112 }, { "epoch": 0.04031148134733792, "grad_norm": 1.9131178948156122, "learning_rate": 1.9994451026101933e-05, "loss": 0.9202, "step": 1113 }, { "epoch": 0.040347700108656286, "grad_norm": 2.264689199616226, "learning_rate": 1.999441188366922e-05, "loss": 0.9099, "step": 1114 }, { "epoch": 0.040383918869974646, "grad_norm": 2.20049298436226, "learning_rate": 1.9994372603704613e-05, "loss": 0.9297, "step": 1115 }, { "epoch": 0.04042013763129301, "grad_norm": 2.0247008161523468, "learning_rate": 1.999433318620866e-05, "loss": 0.8942, "step": 1116 }, { "epoch": 0.04045635639261137, "grad_norm": 2.416023772464173, "learning_rate": 1.99942936311819e-05, "loss": 0.9967, "step": 1117 }, { "epoch": 0.04049257515392973, "grad_norm": 2.0022825900399552, "learning_rate": 1.9994253938624877e-05, "loss": 0.9662, "step": 1118 }, { "epoch": 0.0405287939152481, "grad_norm": 1.868642137637649, "learning_rate": 1.999421410853814e-05, "loss": 0.9064, "step": 1119 }, { "epoch": 0.04056501267656646, "grad_norm": 2.0139524077578246, "learning_rate": 1.9994174140922235e-05, "loss": 0.9333, "step": 1120 }, { "epoch": 0.04060123143788483, "grad_norm": 1.663431066551771, "learning_rate": 1.999413403577771e-05, "loss": 0.8735, "step": 1121 }, { "epoch": 0.04063745019920319, "grad_norm": 1.6014525373582569, "learning_rate": 1.999409379310512e-05, "loss": 0.8547, "step": 1122 }, { "epoch": 0.04067366896052155, "grad_norm": 1.6299534495088202, "learning_rate": 1.9994053412905018e-05, "loss": 0.7996, "step": 1123 }, { "epoch": 0.040709887721839914, "grad_norm": 2.1588226853104966, "learning_rate": 1.999401289517796e-05, "loss": 0.985, "step": 1124 }, { "epoch": 0.040746106483158274, "grad_norm": 2.439446979939499, "learning_rate": 1.99939722399245e-05, "loss": 0.9314, "step": 1125 }, { "epoch": 0.04078232524447664, "grad_norm": 2.2421565086934443, "learning_rate": 1.9993931447145206e-05, "loss": 0.9182, "step": 1126 }, { "epoch": 0.040818544005795, "grad_norm": 1.924897497950085, "learning_rate": 1.9993890516840634e-05, "loss": 0.9732, "step": 1127 }, { "epoch": 0.04085476276711336, "grad_norm": 2.1238004720886177, "learning_rate": 1.999384944901134e-05, "loss": 0.9056, "step": 1128 }, { "epoch": 0.04089098152843173, "grad_norm": 2.246583997878549, "learning_rate": 1.99938082436579e-05, "loss": 0.8482, "step": 1129 }, { "epoch": 0.04092720028975009, "grad_norm": 1.9148287523912648, "learning_rate": 1.999376690078088e-05, "loss": 0.8508, "step": 1130 }, { "epoch": 0.040963419051068456, "grad_norm": 2.2752159853865566, "learning_rate": 1.9993725420380842e-05, "loss": 0.9557, "step": 1131 }, { "epoch": 0.040999637812386816, "grad_norm": 1.9952723629020526, "learning_rate": 1.9993683802458365e-05, "loss": 0.9017, "step": 1132 }, { "epoch": 0.041035856573705176, "grad_norm": 1.939156558593511, "learning_rate": 1.9993642047014014e-05, "loss": 0.8081, "step": 1133 }, { "epoch": 0.04107207533502354, "grad_norm": 2.1988874692180236, "learning_rate": 1.999360015404837e-05, "loss": 0.8402, "step": 1134 }, { "epoch": 0.0411082940963419, "grad_norm": 2.191551615063681, "learning_rate": 1.9993558123562005e-05, "loss": 0.9343, "step": 1135 }, { "epoch": 0.04114451285766027, "grad_norm": 2.307857602607854, "learning_rate": 1.9993515955555498e-05, "loss": 0.8267, "step": 1136 }, { "epoch": 0.04118073161897863, "grad_norm": 1.5616004027567416, "learning_rate": 1.9993473650029432e-05, "loss": 0.8368, "step": 1137 }, { "epoch": 0.04121695038029699, "grad_norm": 2.043436564863708, "learning_rate": 1.999343120698439e-05, "loss": 0.9296, "step": 1138 }, { "epoch": 0.04125316914161536, "grad_norm": 2.002719498927487, "learning_rate": 1.9993388626420947e-05, "loss": 0.9208, "step": 1139 }, { "epoch": 0.04128938790293372, "grad_norm": 1.834536961760295, "learning_rate": 1.99933459083397e-05, "loss": 0.8546, "step": 1140 }, { "epoch": 0.041325606664252085, "grad_norm": 2.403427180549669, "learning_rate": 1.9993303052741234e-05, "loss": 0.9482, "step": 1141 }, { "epoch": 0.041361825425570445, "grad_norm": 2.1498515001566663, "learning_rate": 1.9993260059626133e-05, "loss": 0.8639, "step": 1142 }, { "epoch": 0.041398044186888805, "grad_norm": 1.9915287387212357, "learning_rate": 1.9993216928994997e-05, "loss": 0.9184, "step": 1143 }, { "epoch": 0.04143426294820717, "grad_norm": 2.075702471524983, "learning_rate": 1.9993173660848414e-05, "loss": 0.9193, "step": 1144 }, { "epoch": 0.04147048170952553, "grad_norm": 1.381543665733601, "learning_rate": 1.999313025518698e-05, "loss": 0.8088, "step": 1145 }, { "epoch": 0.0415067004708439, "grad_norm": 2.2524776079571054, "learning_rate": 1.9993086712011293e-05, "loss": 0.959, "step": 1146 }, { "epoch": 0.04154291923216226, "grad_norm": 2.21205788117443, "learning_rate": 1.999304303132195e-05, "loss": 0.8788, "step": 1147 }, { "epoch": 0.041579137993480626, "grad_norm": 2.0302185429156014, "learning_rate": 1.9992999213119555e-05, "loss": 0.9382, "step": 1148 }, { "epoch": 0.041615356754798986, "grad_norm": 2.039802000625902, "learning_rate": 1.999295525740471e-05, "loss": 0.8695, "step": 1149 }, { "epoch": 0.041651575516117346, "grad_norm": 1.7632570065248083, "learning_rate": 1.999291116417802e-05, "loss": 0.8386, "step": 1150 }, { "epoch": 0.041687794277435714, "grad_norm": 1.988488797295451, "learning_rate": 1.9992866933440096e-05, "loss": 0.8214, "step": 1151 }, { "epoch": 0.041724013038754074, "grad_norm": 2.0254344273302967, "learning_rate": 1.999282256519154e-05, "loss": 0.8621, "step": 1152 }, { "epoch": 0.04176023180007244, "grad_norm": 2.336687331273792, "learning_rate": 1.9992778059432964e-05, "loss": 0.9826, "step": 1153 }, { "epoch": 0.0417964505613908, "grad_norm": 1.3933091445201917, "learning_rate": 1.9992733416164983e-05, "loss": 0.8113, "step": 1154 }, { "epoch": 0.04183266932270916, "grad_norm": 2.027336072815341, "learning_rate": 1.999268863538821e-05, "loss": 0.9317, "step": 1155 }, { "epoch": 0.04186888808402753, "grad_norm": 2.1108152417653714, "learning_rate": 1.999264371710326e-05, "loss": 0.9137, "step": 1156 }, { "epoch": 0.04190510684534589, "grad_norm": 1.9367206567321722, "learning_rate": 1.999259866131075e-05, "loss": 0.8452, "step": 1157 }, { "epoch": 0.041941325606664255, "grad_norm": 2.1442017154396398, "learning_rate": 1.9992553468011304e-05, "loss": 0.9804, "step": 1158 }, { "epoch": 0.041977544367982615, "grad_norm": 1.9862129089863643, "learning_rate": 1.999250813720554e-05, "loss": 0.9226, "step": 1159 }, { "epoch": 0.042013763129300975, "grad_norm": 1.9099102056826127, "learning_rate": 1.9992462668894087e-05, "loss": 0.8814, "step": 1160 }, { "epoch": 0.04204998189061934, "grad_norm": 1.9624068655705613, "learning_rate": 1.999241706307757e-05, "loss": 0.8978, "step": 1161 }, { "epoch": 0.0420862006519377, "grad_norm": 1.4404461297212108, "learning_rate": 1.999237131975661e-05, "loss": 0.8391, "step": 1162 }, { "epoch": 0.04212241941325607, "grad_norm": 2.398097803246462, "learning_rate": 1.999232543893184e-05, "loss": 0.863, "step": 1163 }, { "epoch": 0.04215863817457443, "grad_norm": 1.9627747117164298, "learning_rate": 1.9992279420603894e-05, "loss": 0.9462, "step": 1164 }, { "epoch": 0.04219485693589279, "grad_norm": 2.432794507033023, "learning_rate": 1.99922332647734e-05, "loss": 0.9105, "step": 1165 }, { "epoch": 0.04223107569721116, "grad_norm": 2.238094613314524, "learning_rate": 1.9992186971441003e-05, "loss": 0.8661, "step": 1166 }, { "epoch": 0.04226729445852952, "grad_norm": 1.9788314778101754, "learning_rate": 1.9992140540607327e-05, "loss": 0.8759, "step": 1167 }, { "epoch": 0.042303513219847884, "grad_norm": 2.0356046435281443, "learning_rate": 1.999209397227302e-05, "loss": 0.8484, "step": 1168 }, { "epoch": 0.042339731981166244, "grad_norm": 1.9799621016735858, "learning_rate": 1.9992047266438723e-05, "loss": 0.9279, "step": 1169 }, { "epoch": 0.042375950742484604, "grad_norm": 2.141059806965017, "learning_rate": 1.9992000423105067e-05, "loss": 0.9529, "step": 1170 }, { "epoch": 0.04241216950380297, "grad_norm": 1.957267577546028, "learning_rate": 1.9991953442272713e-05, "loss": 0.9, "step": 1171 }, { "epoch": 0.04244838826512133, "grad_norm": 2.153500697739636, "learning_rate": 1.99919063239423e-05, "loss": 0.8851, "step": 1172 }, { "epoch": 0.0424846070264397, "grad_norm": 1.9824585201841516, "learning_rate": 1.999185906811447e-05, "loss": 0.9979, "step": 1173 }, { "epoch": 0.04252082578775806, "grad_norm": 1.9674814883249498, "learning_rate": 1.9991811674789886e-05, "loss": 0.833, "step": 1174 }, { "epoch": 0.04255704454907642, "grad_norm": 2.2317047652781445, "learning_rate": 1.999176414396919e-05, "loss": 0.9317, "step": 1175 }, { "epoch": 0.042593263310394786, "grad_norm": 1.823137506096192, "learning_rate": 1.9991716475653043e-05, "loss": 0.9307, "step": 1176 }, { "epoch": 0.042629482071713146, "grad_norm": 2.217801541108792, "learning_rate": 1.9991668669842097e-05, "loss": 0.9816, "step": 1177 }, { "epoch": 0.04266570083303151, "grad_norm": 2.1687153415631433, "learning_rate": 1.999162072653701e-05, "loss": 0.8992, "step": 1178 }, { "epoch": 0.04270191959434987, "grad_norm": 2.1696872404063465, "learning_rate": 1.999157264573844e-05, "loss": 0.9409, "step": 1179 }, { "epoch": 0.04273813835566823, "grad_norm": 1.884322425637106, "learning_rate": 1.9991524427447054e-05, "loss": 0.9312, "step": 1180 }, { "epoch": 0.0427743571169866, "grad_norm": 1.4493869538243347, "learning_rate": 1.9991476071663515e-05, "loss": 0.7924, "step": 1181 }, { "epoch": 0.04281057587830496, "grad_norm": 1.8827106707487264, "learning_rate": 1.999142757838848e-05, "loss": 0.8604, "step": 1182 }, { "epoch": 0.04284679463962333, "grad_norm": 2.571111278132244, "learning_rate": 1.9991378947622622e-05, "loss": 0.9647, "step": 1183 }, { "epoch": 0.04288301340094169, "grad_norm": 1.800845271232339, "learning_rate": 1.9991330179366618e-05, "loss": 0.9915, "step": 1184 }, { "epoch": 0.04291923216226005, "grad_norm": 1.8909055485923556, "learning_rate": 1.9991281273621126e-05, "loss": 0.9811, "step": 1185 }, { "epoch": 0.042955450923578414, "grad_norm": 2.046342211644149, "learning_rate": 1.9991232230386826e-05, "loss": 0.985, "step": 1186 }, { "epoch": 0.042991669684896774, "grad_norm": 1.4256219505455028, "learning_rate": 1.999118304966439e-05, "loss": 0.7605, "step": 1187 }, { "epoch": 0.04302788844621514, "grad_norm": 1.8942634816053736, "learning_rate": 1.9991133731454498e-05, "loss": 0.8866, "step": 1188 }, { "epoch": 0.0430641072075335, "grad_norm": 2.200949704213438, "learning_rate": 1.9991084275757826e-05, "loss": 0.9629, "step": 1189 }, { "epoch": 0.04310032596885187, "grad_norm": 2.17781041102392, "learning_rate": 1.9991034682575058e-05, "loss": 0.85, "step": 1190 }, { "epoch": 0.04313654473017023, "grad_norm": 1.996347168663758, "learning_rate": 1.999098495190687e-05, "loss": 0.8648, "step": 1191 }, { "epoch": 0.04317276349148859, "grad_norm": 1.8490847434678288, "learning_rate": 1.999093508375395e-05, "loss": 0.888, "step": 1192 }, { "epoch": 0.043208982252806956, "grad_norm": 1.9304239859923567, "learning_rate": 1.9990885078116988e-05, "loss": 0.9428, "step": 1193 }, { "epoch": 0.043245201014125316, "grad_norm": 1.9082925662316834, "learning_rate": 1.9990834934996666e-05, "loss": 0.9429, "step": 1194 }, { "epoch": 0.04328141977544368, "grad_norm": 1.9442858438059714, "learning_rate": 1.999078465439368e-05, "loss": 0.8515, "step": 1195 }, { "epoch": 0.04331763853676204, "grad_norm": 1.9183082318969868, "learning_rate": 1.9990734236308712e-05, "loss": 0.8904, "step": 1196 }, { "epoch": 0.0433538572980804, "grad_norm": 1.4733683596572271, "learning_rate": 1.9990683680742465e-05, "loss": 0.8191, "step": 1197 }, { "epoch": 0.04339007605939877, "grad_norm": 1.8279025409116234, "learning_rate": 1.9990632987695635e-05, "loss": 0.8621, "step": 1198 }, { "epoch": 0.04342629482071713, "grad_norm": 1.9112006831068973, "learning_rate": 1.9990582157168914e-05, "loss": 0.7873, "step": 1199 }, { "epoch": 0.0434625135820355, "grad_norm": 2.6221250915488232, "learning_rate": 1.9990531189163005e-05, "loss": 0.9475, "step": 1200 }, { "epoch": 0.04349873234335386, "grad_norm": 2.0103693918037737, "learning_rate": 1.9990480083678608e-05, "loss": 0.9115, "step": 1201 }, { "epoch": 0.04353495110467222, "grad_norm": 2.0812176283045747, "learning_rate": 1.999042884071643e-05, "loss": 0.9029, "step": 1202 }, { "epoch": 0.043571169865990585, "grad_norm": 1.9580226203187288, "learning_rate": 1.9990377460277166e-05, "loss": 0.908, "step": 1203 }, { "epoch": 0.043607388627308945, "grad_norm": 1.9886250695149377, "learning_rate": 1.9990325942361534e-05, "loss": 0.9519, "step": 1204 }, { "epoch": 0.04364360738862731, "grad_norm": 1.9055152809964984, "learning_rate": 1.9990274286970236e-05, "loss": 0.8932, "step": 1205 }, { "epoch": 0.04367982614994567, "grad_norm": 1.955827188426974, "learning_rate": 1.999022249410399e-05, "loss": 0.9135, "step": 1206 }, { "epoch": 0.04371604491126403, "grad_norm": 1.647141191537758, "learning_rate": 1.99901705637635e-05, "loss": 0.8492, "step": 1207 }, { "epoch": 0.0437522636725824, "grad_norm": 2.1454612690422414, "learning_rate": 1.999011849594949e-05, "loss": 0.9592, "step": 1208 }, { "epoch": 0.04378848243390076, "grad_norm": 1.5173626926541992, "learning_rate": 1.999006629066267e-05, "loss": 0.8013, "step": 1209 }, { "epoch": 0.043824701195219126, "grad_norm": 1.8770528800567203, "learning_rate": 1.9990013947903755e-05, "loss": 0.8653, "step": 1210 }, { "epoch": 0.043860919956537486, "grad_norm": 2.023203856555014, "learning_rate": 1.9989961467673473e-05, "loss": 0.8202, "step": 1211 }, { "epoch": 0.043897138717855846, "grad_norm": 1.9246034328275277, "learning_rate": 1.9989908849972543e-05, "loss": 0.8712, "step": 1212 }, { "epoch": 0.04393335747917421, "grad_norm": 1.9607210780775401, "learning_rate": 1.998985609480169e-05, "loss": 0.8967, "step": 1213 }, { "epoch": 0.043969576240492574, "grad_norm": 2.06810499037213, "learning_rate": 1.998980320216164e-05, "loss": 0.9745, "step": 1214 }, { "epoch": 0.04400579500181094, "grad_norm": 2.1542549838215077, "learning_rate": 1.998975017205312e-05, "loss": 0.942, "step": 1215 }, { "epoch": 0.0440420137631293, "grad_norm": 2.026138568528676, "learning_rate": 1.998969700447686e-05, "loss": 0.8315, "step": 1216 }, { "epoch": 0.04407823252444766, "grad_norm": 1.8969848259298594, "learning_rate": 1.9989643699433587e-05, "loss": 0.8868, "step": 1217 }, { "epoch": 0.04411445128576603, "grad_norm": 2.5251027272827766, "learning_rate": 1.9989590256924045e-05, "loss": 0.9517, "step": 1218 }, { "epoch": 0.04415067004708439, "grad_norm": 2.00409236214733, "learning_rate": 1.998953667694896e-05, "loss": 0.9477, "step": 1219 }, { "epoch": 0.044186888808402755, "grad_norm": 1.7952766082841871, "learning_rate": 1.9989482959509073e-05, "loss": 0.8076, "step": 1220 }, { "epoch": 0.044223107569721115, "grad_norm": 1.9128274007814208, "learning_rate": 1.9989429104605125e-05, "loss": 0.7941, "step": 1221 }, { "epoch": 0.044259326331039475, "grad_norm": 2.002508981341677, "learning_rate": 1.9989375112237853e-05, "loss": 0.8404, "step": 1222 }, { "epoch": 0.04429554509235784, "grad_norm": 1.916338592008342, "learning_rate": 1.9989320982408003e-05, "loss": 0.8712, "step": 1223 }, { "epoch": 0.0443317638536762, "grad_norm": 1.4437858920195292, "learning_rate": 1.9989266715116316e-05, "loss": 0.7781, "step": 1224 }, { "epoch": 0.04436798261499457, "grad_norm": 2.264481755908576, "learning_rate": 1.998921231036355e-05, "loss": 0.9267, "step": 1225 }, { "epoch": 0.04440420137631293, "grad_norm": 2.0658601984049025, "learning_rate": 1.9989157768150437e-05, "loss": 0.9499, "step": 1226 }, { "epoch": 0.04444042013763129, "grad_norm": 1.6133675902852818, "learning_rate": 1.9989103088477738e-05, "loss": 0.8056, "step": 1227 }, { "epoch": 0.04447663889894966, "grad_norm": 2.1783369214363533, "learning_rate": 1.9989048271346204e-05, "loss": 0.9333, "step": 1228 }, { "epoch": 0.04451285766026802, "grad_norm": 2.957688882081352, "learning_rate": 1.998899331675659e-05, "loss": 0.9528, "step": 1229 }, { "epoch": 0.044549076421586384, "grad_norm": 1.5443354238340663, "learning_rate": 1.9988938224709647e-05, "loss": 0.7934, "step": 1230 }, { "epoch": 0.044585295182904744, "grad_norm": 2.001499221331559, "learning_rate": 1.998888299520614e-05, "loss": 0.8958, "step": 1231 }, { "epoch": 0.04462151394422311, "grad_norm": 2.260161487983359, "learning_rate": 1.9988827628246825e-05, "loss": 0.8856, "step": 1232 }, { "epoch": 0.04465773270554147, "grad_norm": 2.008707205254828, "learning_rate": 1.9988772123832464e-05, "loss": 0.929, "step": 1233 }, { "epoch": 0.04469395146685983, "grad_norm": 2.128200294966346, "learning_rate": 1.9988716481963825e-05, "loss": 0.9393, "step": 1234 }, { "epoch": 0.0447301702281782, "grad_norm": 2.019405602728279, "learning_rate": 1.9988660702641668e-05, "loss": 0.9456, "step": 1235 }, { "epoch": 0.04476638898949656, "grad_norm": 1.5982080459211285, "learning_rate": 1.998860478586676e-05, "loss": 0.8504, "step": 1236 }, { "epoch": 0.044802607750814925, "grad_norm": 2.1948333095069485, "learning_rate": 1.998854873163988e-05, "loss": 0.8259, "step": 1237 }, { "epoch": 0.044838826512133285, "grad_norm": 2.2098645278999296, "learning_rate": 1.9988492539961788e-05, "loss": 0.8994, "step": 1238 }, { "epoch": 0.044875045273451646, "grad_norm": 2.2147617560385866, "learning_rate": 1.998843621083326e-05, "loss": 0.8818, "step": 1239 }, { "epoch": 0.04491126403477001, "grad_norm": 1.9566683861120013, "learning_rate": 1.998837974425508e-05, "loss": 0.8692, "step": 1240 }, { "epoch": 0.04494748279608837, "grad_norm": 1.693850050743348, "learning_rate": 1.9988323140228015e-05, "loss": 0.7973, "step": 1241 }, { "epoch": 0.04498370155740674, "grad_norm": 2.192146594399342, "learning_rate": 1.9988266398752845e-05, "loss": 0.8936, "step": 1242 }, { "epoch": 0.0450199203187251, "grad_norm": 1.7328783602308673, "learning_rate": 1.9988209519830355e-05, "loss": 0.807, "step": 1243 }, { "epoch": 0.04505613908004346, "grad_norm": 2.172891382431169, "learning_rate": 1.9988152503461325e-05, "loss": 0.8367, "step": 1244 }, { "epoch": 0.04509235784136183, "grad_norm": 2.2058841872470607, "learning_rate": 1.9988095349646543e-05, "loss": 0.9387, "step": 1245 }, { "epoch": 0.04512857660268019, "grad_norm": 1.9004878639678238, "learning_rate": 1.9988038058386787e-05, "loss": 0.8641, "step": 1246 }, { "epoch": 0.045164795363998554, "grad_norm": 2.1197033649153436, "learning_rate": 1.9987980629682854e-05, "loss": 0.866, "step": 1247 }, { "epoch": 0.045201014125316914, "grad_norm": 2.006689217580402, "learning_rate": 1.9987923063535533e-05, "loss": 0.8987, "step": 1248 }, { "epoch": 0.045237232886635274, "grad_norm": 2.198635420332182, "learning_rate": 1.9987865359945614e-05, "loss": 0.8892, "step": 1249 }, { "epoch": 0.04527345164795364, "grad_norm": 1.9150790415106387, "learning_rate": 1.998780751891389e-05, "loss": 0.8492, "step": 1250 }, { "epoch": 0.045309670409272, "grad_norm": 1.4062552511594002, "learning_rate": 1.9987749540441158e-05, "loss": 0.764, "step": 1251 }, { "epoch": 0.04534588917059037, "grad_norm": 1.594289231179919, "learning_rate": 1.9987691424528214e-05, "loss": 0.8636, "step": 1252 }, { "epoch": 0.04538210793190873, "grad_norm": 1.3410886537162434, "learning_rate": 1.9987633171175864e-05, "loss": 0.7538, "step": 1253 }, { "epoch": 0.04541832669322709, "grad_norm": 1.6915847059774736, "learning_rate": 1.9987574780384906e-05, "loss": 0.7772, "step": 1254 }, { "epoch": 0.045454545454545456, "grad_norm": 2.1086781692990404, "learning_rate": 1.9987516252156143e-05, "loss": 0.8978, "step": 1255 }, { "epoch": 0.045490764215863816, "grad_norm": 1.9816002162435555, "learning_rate": 1.998745758649038e-05, "loss": 0.8917, "step": 1256 }, { "epoch": 0.04552698297718218, "grad_norm": 2.2037292017201398, "learning_rate": 1.9987398783388422e-05, "loss": 0.8586, "step": 1257 }, { "epoch": 0.04556320173850054, "grad_norm": 1.799565725250515, "learning_rate": 1.9987339842851086e-05, "loss": 0.9486, "step": 1258 }, { "epoch": 0.0455994204998189, "grad_norm": 1.7245527784148318, "learning_rate": 1.9987280764879174e-05, "loss": 0.8212, "step": 1259 }, { "epoch": 0.04563563926113727, "grad_norm": 2.0037298115544973, "learning_rate": 1.9987221549473507e-05, "loss": 0.9539, "step": 1260 }, { "epoch": 0.04567185802245563, "grad_norm": 2.2169041477413387, "learning_rate": 1.9987162196634894e-05, "loss": 0.9586, "step": 1261 }, { "epoch": 0.045708076783774, "grad_norm": 1.9004758083584399, "learning_rate": 1.9987102706364152e-05, "loss": 0.9114, "step": 1262 }, { "epoch": 0.04574429554509236, "grad_norm": 1.2820407875576254, "learning_rate": 1.99870430786621e-05, "loss": 0.7408, "step": 1263 }, { "epoch": 0.04578051430641072, "grad_norm": 2.060508997058295, "learning_rate": 1.9986983313529564e-05, "loss": 0.8302, "step": 1264 }, { "epoch": 0.045816733067729085, "grad_norm": 2.0291237780237665, "learning_rate": 1.998692341096736e-05, "loss": 0.8217, "step": 1265 }, { "epoch": 0.045852951829047445, "grad_norm": 2.1738837921335916, "learning_rate": 1.9986863370976316e-05, "loss": 0.9147, "step": 1266 }, { "epoch": 0.04588917059036581, "grad_norm": 1.794529136527556, "learning_rate": 1.998680319355726e-05, "loss": 0.8573, "step": 1267 }, { "epoch": 0.04592538935168417, "grad_norm": 1.9426711085710173, "learning_rate": 1.998674287871101e-05, "loss": 0.974, "step": 1268 }, { "epoch": 0.04596160811300253, "grad_norm": 1.9587903958368476, "learning_rate": 1.9986682426438406e-05, "loss": 0.9183, "step": 1269 }, { "epoch": 0.0459978268743209, "grad_norm": 1.8476225502598314, "learning_rate": 1.9986621836740277e-05, "loss": 0.7418, "step": 1270 }, { "epoch": 0.04603404563563926, "grad_norm": 1.9260917417141794, "learning_rate": 1.9986561109617457e-05, "loss": 0.8626, "step": 1271 }, { "epoch": 0.046070264396957626, "grad_norm": 2.1057042483829673, "learning_rate": 1.9986500245070778e-05, "loss": 0.9301, "step": 1272 }, { "epoch": 0.046106483158275986, "grad_norm": 2.038942030201656, "learning_rate": 1.9986439243101084e-05, "loss": 0.9086, "step": 1273 }, { "epoch": 0.04614270191959435, "grad_norm": 1.9303916291285945, "learning_rate": 1.9986378103709212e-05, "loss": 0.9743, "step": 1274 }, { "epoch": 0.04617892068091271, "grad_norm": 1.7776825278057424, "learning_rate": 1.9986316826896e-05, "loss": 0.8961, "step": 1275 }, { "epoch": 0.046215139442231074, "grad_norm": 2.161110056926695, "learning_rate": 1.9986255412662294e-05, "loss": 0.886, "step": 1276 }, { "epoch": 0.04625135820354944, "grad_norm": 1.790634022580961, "learning_rate": 1.998619386100894e-05, "loss": 0.9153, "step": 1277 }, { "epoch": 0.0462875769648678, "grad_norm": 1.4999033617701325, "learning_rate": 1.998613217193678e-05, "loss": 0.7856, "step": 1278 }, { "epoch": 0.04632379572618617, "grad_norm": 2.046026569988136, "learning_rate": 1.998607034544667e-05, "loss": 0.9252, "step": 1279 }, { "epoch": 0.04636001448750453, "grad_norm": 1.9762808821519435, "learning_rate": 1.998600838153946e-05, "loss": 0.9044, "step": 1280 }, { "epoch": 0.04639623324882289, "grad_norm": 1.6553168132185034, "learning_rate": 1.9985946280215996e-05, "loss": 0.7924, "step": 1281 }, { "epoch": 0.046432452010141255, "grad_norm": 2.0707283106281333, "learning_rate": 1.998588404147714e-05, "loss": 0.9148, "step": 1282 }, { "epoch": 0.046468670771459615, "grad_norm": 2.020354309814766, "learning_rate": 1.998582166532374e-05, "loss": 0.9168, "step": 1283 }, { "epoch": 0.04650488953277798, "grad_norm": 2.0351150017949577, "learning_rate": 1.9985759151756668e-05, "loss": 0.8999, "step": 1284 }, { "epoch": 0.04654110829409634, "grad_norm": 1.931842785662013, "learning_rate": 1.998569650077677e-05, "loss": 0.8731, "step": 1285 }, { "epoch": 0.0465773270554147, "grad_norm": 1.8708141611776417, "learning_rate": 1.9985633712384915e-05, "loss": 0.9034, "step": 1286 }, { "epoch": 0.04661354581673307, "grad_norm": 1.6558979334017987, "learning_rate": 1.9985570786581966e-05, "loss": 0.8588, "step": 1287 }, { "epoch": 0.04664976457805143, "grad_norm": 2.1882948214822715, "learning_rate": 1.9985507723368786e-05, "loss": 0.9347, "step": 1288 }, { "epoch": 0.046685983339369796, "grad_norm": 1.9530468872662683, "learning_rate": 1.998544452274625e-05, "loss": 0.9215, "step": 1289 }, { "epoch": 0.04672220210068816, "grad_norm": 1.9874177877124022, "learning_rate": 1.998538118471522e-05, "loss": 0.883, "step": 1290 }, { "epoch": 0.04675842086200652, "grad_norm": 2.009081037162602, "learning_rate": 1.9985317709276574e-05, "loss": 0.8854, "step": 1291 }, { "epoch": 0.046794639623324884, "grad_norm": 2.0621285501417055, "learning_rate": 1.998525409643118e-05, "loss": 0.9164, "step": 1292 }, { "epoch": 0.046830858384643244, "grad_norm": 2.203668236549104, "learning_rate": 1.9985190346179918e-05, "loss": 0.8768, "step": 1293 }, { "epoch": 0.04686707714596161, "grad_norm": 1.3536128595615857, "learning_rate": 1.998512645852366e-05, "loss": 0.726, "step": 1294 }, { "epoch": 0.04690329590727997, "grad_norm": 2.10067842812821, "learning_rate": 1.9985062433463292e-05, "loss": 0.9661, "step": 1295 }, { "epoch": 0.04693951466859833, "grad_norm": 1.958630030599935, "learning_rate": 1.9984998270999688e-05, "loss": 0.9571, "step": 1296 }, { "epoch": 0.0469757334299167, "grad_norm": 1.6670490885377323, "learning_rate": 1.9984933971133736e-05, "loss": 0.9246, "step": 1297 }, { "epoch": 0.04701195219123506, "grad_norm": 1.915239369344842, "learning_rate": 1.998486953386632e-05, "loss": 0.8915, "step": 1298 }, { "epoch": 0.047048170952553425, "grad_norm": 2.112789154930895, "learning_rate": 1.9984804959198327e-05, "loss": 0.8984, "step": 1299 }, { "epoch": 0.047084389713871785, "grad_norm": 1.737431427480254, "learning_rate": 1.998474024713064e-05, "loss": 0.8348, "step": 1300 }, { "epoch": 0.047120608475190146, "grad_norm": 1.9867390054817, "learning_rate": 1.9984675397664154e-05, "loss": 0.9334, "step": 1301 }, { "epoch": 0.04715682723650851, "grad_norm": 1.3233880531846685, "learning_rate": 1.9984610410799764e-05, "loss": 0.7984, "step": 1302 }, { "epoch": 0.04719304599782687, "grad_norm": 2.19259184284768, "learning_rate": 1.9984545286538362e-05, "loss": 0.8619, "step": 1303 }, { "epoch": 0.04722926475914524, "grad_norm": 1.9155288966607988, "learning_rate": 1.9984480024880844e-05, "loss": 0.9278, "step": 1304 }, { "epoch": 0.0472654835204636, "grad_norm": 2.063765444665992, "learning_rate": 1.9984414625828107e-05, "loss": 0.9798, "step": 1305 }, { "epoch": 0.04730170228178196, "grad_norm": 1.6143075303015564, "learning_rate": 1.9984349089381046e-05, "loss": 0.817, "step": 1306 }, { "epoch": 0.04733792104310033, "grad_norm": 1.9905820642059824, "learning_rate": 1.9984283415540575e-05, "loss": 0.8652, "step": 1307 }, { "epoch": 0.04737413980441869, "grad_norm": 1.9919015174246104, "learning_rate": 1.998421760430759e-05, "loss": 1.0027, "step": 1308 }, { "epoch": 0.047410358565737054, "grad_norm": 2.1553879617979264, "learning_rate": 1.9984151655682997e-05, "loss": 0.8134, "step": 1309 }, { "epoch": 0.047446577327055414, "grad_norm": 1.7261986845428914, "learning_rate": 1.9984085569667705e-05, "loss": 0.7958, "step": 1310 }, { "epoch": 0.047482796088373774, "grad_norm": 1.9681787021460075, "learning_rate": 1.9984019346262623e-05, "loss": 0.9043, "step": 1311 }, { "epoch": 0.04751901484969214, "grad_norm": 1.6328406814887872, "learning_rate": 1.998395298546866e-05, "loss": 0.8087, "step": 1312 }, { "epoch": 0.0475552336110105, "grad_norm": 2.1506547756955214, "learning_rate": 1.9983886487286733e-05, "loss": 0.99, "step": 1313 }, { "epoch": 0.04759145237232887, "grad_norm": 1.959615772039013, "learning_rate": 1.9983819851717753e-05, "loss": 0.9474, "step": 1314 }, { "epoch": 0.04762767113364723, "grad_norm": 1.9020136149203186, "learning_rate": 1.998375307876264e-05, "loss": 0.8886, "step": 1315 }, { "epoch": 0.047663889894965596, "grad_norm": 2.017233225007247, "learning_rate": 1.9983686168422315e-05, "loss": 0.9251, "step": 1316 }, { "epoch": 0.047700108656283956, "grad_norm": 2.071745946207053, "learning_rate": 1.9983619120697694e-05, "loss": 0.9401, "step": 1317 }, { "epoch": 0.047736327417602316, "grad_norm": 2.0570851051272583, "learning_rate": 1.9983551935589705e-05, "loss": 0.9068, "step": 1318 }, { "epoch": 0.04777254617892068, "grad_norm": 1.9695272614356587, "learning_rate": 1.9983484613099265e-05, "loss": 0.9781, "step": 1319 }, { "epoch": 0.04780876494023904, "grad_norm": 1.4532555942774068, "learning_rate": 1.9983417153227304e-05, "loss": 0.8329, "step": 1320 }, { "epoch": 0.04784498370155741, "grad_norm": 1.775011103300568, "learning_rate": 1.9983349555974754e-05, "loss": 0.8396, "step": 1321 }, { "epoch": 0.04788120246287577, "grad_norm": 1.9801532373651503, "learning_rate": 1.998328182134254e-05, "loss": 0.8816, "step": 1322 }, { "epoch": 0.04791742122419413, "grad_norm": 1.887860020113778, "learning_rate": 1.9983213949331594e-05, "loss": 0.8487, "step": 1323 }, { "epoch": 0.0479536399855125, "grad_norm": 1.5247921349590359, "learning_rate": 1.9983145939942858e-05, "loss": 0.7549, "step": 1324 }, { "epoch": 0.04798985874683086, "grad_norm": 1.8945827300166045, "learning_rate": 1.9983077793177253e-05, "loss": 0.8925, "step": 1325 }, { "epoch": 0.048026077508149224, "grad_norm": 1.8573292766568166, "learning_rate": 1.9983009509035735e-05, "loss": 0.8108, "step": 1326 }, { "epoch": 0.048062296269467585, "grad_norm": 2.221867912619188, "learning_rate": 1.998294108751923e-05, "loss": 0.8766, "step": 1327 }, { "epoch": 0.048098515030785945, "grad_norm": 1.3536292851242095, "learning_rate": 1.998287252862868e-05, "loss": 0.7474, "step": 1328 }, { "epoch": 0.04813473379210431, "grad_norm": 1.979741470832651, "learning_rate": 1.998280383236504e-05, "loss": 0.9488, "step": 1329 }, { "epoch": 0.04817095255342267, "grad_norm": 2.050458881835807, "learning_rate": 1.9982734998729244e-05, "loss": 0.9797, "step": 1330 }, { "epoch": 0.04820717131474104, "grad_norm": 1.8901871379098543, "learning_rate": 1.998266602772224e-05, "loss": 0.8652, "step": 1331 }, { "epoch": 0.0482433900760594, "grad_norm": 1.8919554750045369, "learning_rate": 1.9982596919344983e-05, "loss": 0.8924, "step": 1332 }, { "epoch": 0.04827960883737776, "grad_norm": 1.6716857154665419, "learning_rate": 1.998252767359842e-05, "loss": 0.8224, "step": 1333 }, { "epoch": 0.048315827598696126, "grad_norm": 2.0082015781214837, "learning_rate": 1.9982458290483504e-05, "loss": 0.8181, "step": 1334 }, { "epoch": 0.048352046360014486, "grad_norm": 2.014483771942628, "learning_rate": 1.9982388770001193e-05, "loss": 0.8859, "step": 1335 }, { "epoch": 0.04838826512133285, "grad_norm": 1.9011222277950974, "learning_rate": 1.998231911215244e-05, "loss": 0.9458, "step": 1336 }, { "epoch": 0.04842448388265121, "grad_norm": 2.0670460357760048, "learning_rate": 1.9982249316938205e-05, "loss": 0.9374, "step": 1337 }, { "epoch": 0.04846070264396957, "grad_norm": 2.390023666996869, "learning_rate": 1.9982179384359447e-05, "loss": 0.9192, "step": 1338 }, { "epoch": 0.04849692140528794, "grad_norm": 1.823199095801657, "learning_rate": 1.998210931441713e-05, "loss": 0.8879, "step": 1339 }, { "epoch": 0.0485331401666063, "grad_norm": 1.5478140648953902, "learning_rate": 1.9982039107112218e-05, "loss": 0.7773, "step": 1340 }, { "epoch": 0.04856935892792467, "grad_norm": 2.1397298660697124, "learning_rate": 1.9981968762445678e-05, "loss": 0.9365, "step": 1341 }, { "epoch": 0.04860557768924303, "grad_norm": 1.924520909037786, "learning_rate": 1.9981898280418477e-05, "loss": 0.9293, "step": 1342 }, { "epoch": 0.04864179645056139, "grad_norm": 1.9376525107471385, "learning_rate": 1.9981827661031586e-05, "loss": 0.8381, "step": 1343 }, { "epoch": 0.048678015211879755, "grad_norm": 1.5491945235012028, "learning_rate": 1.9981756904285973e-05, "loss": 0.7726, "step": 1344 }, { "epoch": 0.048714233973198115, "grad_norm": 2.0134077555679477, "learning_rate": 1.9981686010182614e-05, "loss": 0.8999, "step": 1345 }, { "epoch": 0.04875045273451648, "grad_norm": 2.0472509242787185, "learning_rate": 1.9981614978722485e-05, "loss": 0.8795, "step": 1346 }, { "epoch": 0.04878667149583484, "grad_norm": 2.004322046044903, "learning_rate": 1.9981543809906564e-05, "loss": 0.8198, "step": 1347 }, { "epoch": 0.0488228902571532, "grad_norm": 1.8385778048496506, "learning_rate": 1.9981472503735832e-05, "loss": 0.8934, "step": 1348 }, { "epoch": 0.04885910901847157, "grad_norm": 1.9241498513534703, "learning_rate": 1.9981401060211267e-05, "loss": 0.8491, "step": 1349 }, { "epoch": 0.04889532777978993, "grad_norm": 1.7927919094914027, "learning_rate": 1.9981329479333848e-05, "loss": 0.8201, "step": 1350 }, { "epoch": 0.048931546541108296, "grad_norm": 1.8334632941714246, "learning_rate": 1.998125776110457e-05, "loss": 0.8844, "step": 1351 }, { "epoch": 0.04896776530242666, "grad_norm": 1.9860705562200909, "learning_rate": 1.998118590552441e-05, "loss": 0.9236, "step": 1352 }, { "epoch": 0.04900398406374502, "grad_norm": 2.1174621843729873, "learning_rate": 1.9981113912594368e-05, "loss": 0.8842, "step": 1353 }, { "epoch": 0.049040202825063384, "grad_norm": 1.7615247916239571, "learning_rate": 1.9981041782315425e-05, "loss": 0.7946, "step": 1354 }, { "epoch": 0.049076421586381744, "grad_norm": 2.195503435706581, "learning_rate": 1.9980969514688575e-05, "loss": 0.7991, "step": 1355 }, { "epoch": 0.04911264034770011, "grad_norm": 1.9290323380782357, "learning_rate": 1.9980897109714818e-05, "loss": 0.8432, "step": 1356 }, { "epoch": 0.04914885910901847, "grad_norm": 2.1204859668948255, "learning_rate": 1.9980824567395146e-05, "loss": 0.8929, "step": 1357 }, { "epoch": 0.04918507787033683, "grad_norm": 1.8135211324889566, "learning_rate": 1.9980751887730556e-05, "loss": 0.9029, "step": 1358 }, { "epoch": 0.0492212966316552, "grad_norm": 2.018646326739368, "learning_rate": 1.9980679070722053e-05, "loss": 0.8035, "step": 1359 }, { "epoch": 0.04925751539297356, "grad_norm": 1.8981864266713606, "learning_rate": 1.9980606116370634e-05, "loss": 0.9847, "step": 1360 }, { "epoch": 0.049293734154291925, "grad_norm": 1.9228709284328833, "learning_rate": 1.9980533024677305e-05, "loss": 0.9443, "step": 1361 }, { "epoch": 0.049329952915610285, "grad_norm": 1.9247295939807734, "learning_rate": 1.9980459795643076e-05, "loss": 0.8267, "step": 1362 }, { "epoch": 0.04936617167692865, "grad_norm": 2.161153873387523, "learning_rate": 1.9980386429268946e-05, "loss": 0.9496, "step": 1363 }, { "epoch": 0.04940239043824701, "grad_norm": 1.8318323482889467, "learning_rate": 1.9980312925555932e-05, "loss": 0.9075, "step": 1364 }, { "epoch": 0.04943860919956537, "grad_norm": 2.116884574122328, "learning_rate": 1.998023928450504e-05, "loss": 0.9565, "step": 1365 }, { "epoch": 0.04947482796088374, "grad_norm": 1.8225741753818891, "learning_rate": 1.998016550611729e-05, "loss": 0.8845, "step": 1366 }, { "epoch": 0.0495110467222021, "grad_norm": 2.275722106233166, "learning_rate": 1.998009159039369e-05, "loss": 0.8899, "step": 1367 }, { "epoch": 0.04954726548352047, "grad_norm": 1.86932100424174, "learning_rate": 1.998001753733526e-05, "loss": 0.9077, "step": 1368 }, { "epoch": 0.04958348424483883, "grad_norm": 2.648699826214941, "learning_rate": 1.997994334694302e-05, "loss": 0.8304, "step": 1369 }, { "epoch": 0.04961970300615719, "grad_norm": 1.8683651986849859, "learning_rate": 1.9979869019217992e-05, "loss": 0.9071, "step": 1370 }, { "epoch": 0.049655921767475554, "grad_norm": 1.7925746729824388, "learning_rate": 1.9979794554161197e-05, "loss": 0.875, "step": 1371 }, { "epoch": 0.049692140528793914, "grad_norm": 1.9373562513370053, "learning_rate": 1.997971995177366e-05, "loss": 0.8853, "step": 1372 }, { "epoch": 0.04972835929011228, "grad_norm": 1.9005885604308725, "learning_rate": 1.9979645212056407e-05, "loss": 0.9727, "step": 1373 }, { "epoch": 0.04976457805143064, "grad_norm": 1.4744964325833354, "learning_rate": 1.9979570335010464e-05, "loss": 0.7795, "step": 1374 }, { "epoch": 0.049800796812749, "grad_norm": 1.369425148648175, "learning_rate": 1.997949532063687e-05, "loss": 0.8149, "step": 1375 }, { "epoch": 0.04983701557406737, "grad_norm": 2.1157811304459586, "learning_rate": 1.997942016893665e-05, "loss": 0.8862, "step": 1376 }, { "epoch": 0.04987323433538573, "grad_norm": 2.0997021038071586, "learning_rate": 1.9979344879910837e-05, "loss": 0.9108, "step": 1377 }, { "epoch": 0.049909453096704096, "grad_norm": 2.017370545675621, "learning_rate": 1.997926945356047e-05, "loss": 0.958, "step": 1378 }, { "epoch": 0.049945671858022456, "grad_norm": 2.3067910438479657, "learning_rate": 1.9979193889886588e-05, "loss": 0.9704, "step": 1379 }, { "epoch": 0.049981890619340816, "grad_norm": 2.1125576345775308, "learning_rate": 1.9979118188890234e-05, "loss": 0.9279, "step": 1380 }, { "epoch": 0.05001810938065918, "grad_norm": 1.880600856531422, "learning_rate": 1.9979042350572437e-05, "loss": 0.8924, "step": 1381 }, { "epoch": 0.05005432814197754, "grad_norm": 2.5329098145641735, "learning_rate": 1.9978966374934255e-05, "loss": 0.9026, "step": 1382 }, { "epoch": 0.05009054690329591, "grad_norm": 2.088188431165157, "learning_rate": 1.9978890261976725e-05, "loss": 0.9366, "step": 1383 }, { "epoch": 0.05012676566461427, "grad_norm": 1.5590840293950323, "learning_rate": 1.9978814011700897e-05, "loss": 0.7848, "step": 1384 }, { "epoch": 0.05016298442593263, "grad_norm": 2.024949748650513, "learning_rate": 1.997873762410782e-05, "loss": 0.8557, "step": 1385 }, { "epoch": 0.050199203187251, "grad_norm": 2.2095580363769263, "learning_rate": 1.9978661099198545e-05, "loss": 0.8815, "step": 1386 }, { "epoch": 0.05023542194856936, "grad_norm": 1.9035309801709428, "learning_rate": 1.9978584436974128e-05, "loss": 0.8849, "step": 1387 }, { "epoch": 0.050271640709887724, "grad_norm": 2.2550256947366645, "learning_rate": 1.9978507637435615e-05, "loss": 0.898, "step": 1388 }, { "epoch": 0.050307859471206084, "grad_norm": 2.014653897890577, "learning_rate": 1.9978430700584075e-05, "loss": 1.0365, "step": 1389 }, { "epoch": 0.050344078232524445, "grad_norm": 1.4319609435583092, "learning_rate": 1.997835362642056e-05, "loss": 0.7519, "step": 1390 }, { "epoch": 0.05038029699384281, "grad_norm": 1.9870963163117366, "learning_rate": 1.9978276414946126e-05, "loss": 0.8997, "step": 1391 }, { "epoch": 0.05041651575516117, "grad_norm": 2.074661356727778, "learning_rate": 1.9978199066161847e-05, "loss": 0.8677, "step": 1392 }, { "epoch": 0.05045273451647954, "grad_norm": 2.3862706683109556, "learning_rate": 1.9978121580068778e-05, "loss": 0.9489, "step": 1393 }, { "epoch": 0.0504889532777979, "grad_norm": 1.9051463438158358, "learning_rate": 1.997804395666799e-05, "loss": 0.9229, "step": 1394 }, { "epoch": 0.05052517203911626, "grad_norm": 1.5633823272855416, "learning_rate": 1.997796619596055e-05, "loss": 0.7757, "step": 1395 }, { "epoch": 0.050561390800434626, "grad_norm": 2.0952451246354165, "learning_rate": 1.9977888297947527e-05, "loss": 0.8344, "step": 1396 }, { "epoch": 0.050597609561752986, "grad_norm": 2.008051895089615, "learning_rate": 1.9977810262629993e-05, "loss": 0.838, "step": 1397 }, { "epoch": 0.05063382832307135, "grad_norm": 2.2111207057141944, "learning_rate": 1.9977732090009022e-05, "loss": 0.9339, "step": 1398 }, { "epoch": 0.05067004708438971, "grad_norm": 1.8384004173692066, "learning_rate": 1.9977653780085694e-05, "loss": 0.9176, "step": 1399 }, { "epoch": 0.05070626584570807, "grad_norm": 1.8165476097967945, "learning_rate": 1.997757533286108e-05, "loss": 0.8436, "step": 1400 }, { "epoch": 0.05074248460702644, "grad_norm": 2.205041831016036, "learning_rate": 1.9977496748336264e-05, "loss": 0.9133, "step": 1401 }, { "epoch": 0.0507787033683448, "grad_norm": 1.755241174359981, "learning_rate": 1.9977418026512323e-05, "loss": 0.8847, "step": 1402 }, { "epoch": 0.05081492212966317, "grad_norm": 1.886667782033873, "learning_rate": 1.9977339167390345e-05, "loss": 0.8503, "step": 1403 }, { "epoch": 0.05085114089098153, "grad_norm": 2.0689261194025077, "learning_rate": 1.997726017097141e-05, "loss": 0.9385, "step": 1404 }, { "epoch": 0.050887359652299895, "grad_norm": 1.7773717741800488, "learning_rate": 1.9977181037256613e-05, "loss": 0.8667, "step": 1405 }, { "epoch": 0.050923578413618255, "grad_norm": 1.5513736259469002, "learning_rate": 1.9977101766247036e-05, "loss": 0.8109, "step": 1406 }, { "epoch": 0.050959797174936615, "grad_norm": 1.859508810117088, "learning_rate": 1.997702235794377e-05, "loss": 0.8809, "step": 1407 }, { "epoch": 0.05099601593625498, "grad_norm": 1.8027795039982835, "learning_rate": 1.9976942812347914e-05, "loss": 0.9291, "step": 1408 }, { "epoch": 0.05103223469757334, "grad_norm": 1.8067256639416427, "learning_rate": 1.9976863129460557e-05, "loss": 0.8738, "step": 1409 }, { "epoch": 0.05106845345889171, "grad_norm": 1.998691987930766, "learning_rate": 1.9976783309282793e-05, "loss": 0.946, "step": 1410 }, { "epoch": 0.05110467222021007, "grad_norm": 1.9045894769694536, "learning_rate": 1.9976703351815727e-05, "loss": 0.9497, "step": 1411 }, { "epoch": 0.05114089098152843, "grad_norm": 1.9667289991135217, "learning_rate": 1.997662325706046e-05, "loss": 0.8877, "step": 1412 }, { "epoch": 0.051177109742846796, "grad_norm": 1.8530863354205105, "learning_rate": 1.9976543025018086e-05, "loss": 0.8474, "step": 1413 }, { "epoch": 0.051213328504165156, "grad_norm": 1.889490202752293, "learning_rate": 1.9976462655689716e-05, "loss": 0.8773, "step": 1414 }, { "epoch": 0.051249547265483524, "grad_norm": 1.4469466533591342, "learning_rate": 1.9976382149076455e-05, "loss": 0.8088, "step": 1415 }, { "epoch": 0.051285766026801884, "grad_norm": 1.4127580790273309, "learning_rate": 1.9976301505179404e-05, "loss": 0.8141, "step": 1416 }, { "epoch": 0.051321984788120244, "grad_norm": 1.901256377797587, "learning_rate": 1.9976220723999685e-05, "loss": 0.875, "step": 1417 }, { "epoch": 0.05135820354943861, "grad_norm": 1.403405276867515, "learning_rate": 1.99761398055384e-05, "loss": 0.8069, "step": 1418 }, { "epoch": 0.05139442231075697, "grad_norm": 2.2296803595943837, "learning_rate": 1.9976058749796667e-05, "loss": 0.8518, "step": 1419 }, { "epoch": 0.05143064107207534, "grad_norm": 2.2240225794190778, "learning_rate": 1.9975977556775597e-05, "loss": 0.9207, "step": 1420 }, { "epoch": 0.0514668598333937, "grad_norm": 2.065125419985173, "learning_rate": 1.9975896226476314e-05, "loss": 0.8658, "step": 1421 }, { "epoch": 0.05150307859471206, "grad_norm": 1.9093854700588102, "learning_rate": 1.997581475889993e-05, "loss": 0.8737, "step": 1422 }, { "epoch": 0.051539297356030425, "grad_norm": 1.582825787761256, "learning_rate": 1.997573315404757e-05, "loss": 0.7855, "step": 1423 }, { "epoch": 0.051575516117348785, "grad_norm": 1.5047694870004678, "learning_rate": 1.9975651411920357e-05, "loss": 0.8208, "step": 1424 }, { "epoch": 0.05161173487866715, "grad_norm": 1.9695541559814378, "learning_rate": 1.9975569532519413e-05, "loss": 0.954, "step": 1425 }, { "epoch": 0.05164795363998551, "grad_norm": 1.86406447985241, "learning_rate": 1.997548751584587e-05, "loss": 0.857, "step": 1426 }, { "epoch": 0.05168417240130387, "grad_norm": 1.875711965492811, "learning_rate": 1.9975405361900853e-05, "loss": 0.9097, "step": 1427 }, { "epoch": 0.05172039116262224, "grad_norm": 2.0335343161700776, "learning_rate": 1.9975323070685494e-05, "loss": 0.8812, "step": 1428 }, { "epoch": 0.0517566099239406, "grad_norm": 2.046643608059926, "learning_rate": 1.997524064220092e-05, "loss": 0.9364, "step": 1429 }, { "epoch": 0.05179282868525897, "grad_norm": 2.0679087477794638, "learning_rate": 1.9975158076448274e-05, "loss": 0.8259, "step": 1430 }, { "epoch": 0.05182904744657733, "grad_norm": 1.8854503810794225, "learning_rate": 1.9975075373428688e-05, "loss": 0.8419, "step": 1431 }, { "epoch": 0.05186526620789569, "grad_norm": 1.873432850573436, "learning_rate": 1.99749925331433e-05, "loss": 0.906, "step": 1432 }, { "epoch": 0.051901484969214054, "grad_norm": 1.7955039882053805, "learning_rate": 1.9974909555593246e-05, "loss": 0.8725, "step": 1433 }, { "epoch": 0.051937703730532414, "grad_norm": 1.9505274704199316, "learning_rate": 1.9974826440779674e-05, "loss": 0.9481, "step": 1434 }, { "epoch": 0.05197392249185078, "grad_norm": 2.163129741492113, "learning_rate": 1.9974743188703726e-05, "loss": 0.8958, "step": 1435 }, { "epoch": 0.05201014125316914, "grad_norm": 1.8141333028166615, "learning_rate": 1.9974659799366546e-05, "loss": 0.8728, "step": 1436 }, { "epoch": 0.0520463600144875, "grad_norm": 1.9420203642233524, "learning_rate": 1.9974576272769286e-05, "loss": 0.9402, "step": 1437 }, { "epoch": 0.05208257877580587, "grad_norm": 1.9377297382509198, "learning_rate": 1.9974492608913088e-05, "loss": 0.9704, "step": 1438 }, { "epoch": 0.05211879753712423, "grad_norm": 1.9477021369930045, "learning_rate": 1.997440880779911e-05, "loss": 0.8789, "step": 1439 }, { "epoch": 0.052155016298442596, "grad_norm": 1.886421561282097, "learning_rate": 1.9974324869428504e-05, "loss": 0.9381, "step": 1440 }, { "epoch": 0.052191235059760956, "grad_norm": 2.0734173480163482, "learning_rate": 1.997424079380242e-05, "loss": 0.9823, "step": 1441 }, { "epoch": 0.052227453821079316, "grad_norm": 1.9637032416232478, "learning_rate": 1.9974156580922018e-05, "loss": 0.9311, "step": 1442 }, { "epoch": 0.05226367258239768, "grad_norm": 1.6787294716336145, "learning_rate": 1.997407223078846e-05, "loss": 0.9038, "step": 1443 }, { "epoch": 0.05229989134371604, "grad_norm": 1.8522650234976525, "learning_rate": 1.9973987743402905e-05, "loss": 0.8496, "step": 1444 }, { "epoch": 0.05233611010503441, "grad_norm": 2.010898785933539, "learning_rate": 1.9973903118766515e-05, "loss": 0.9226, "step": 1445 }, { "epoch": 0.05237232886635277, "grad_norm": 1.8124132115040144, "learning_rate": 1.997381835688045e-05, "loss": 0.852, "step": 1446 }, { "epoch": 0.05240854762767114, "grad_norm": 1.9910466050104099, "learning_rate": 1.9973733457745883e-05, "loss": 0.8633, "step": 1447 }, { "epoch": 0.0524447663889895, "grad_norm": 2.076373993876166, "learning_rate": 1.997364842136398e-05, "loss": 0.9075, "step": 1448 }, { "epoch": 0.05248098515030786, "grad_norm": 2.0537573173825603, "learning_rate": 1.997356324773591e-05, "loss": 0.9735, "step": 1449 }, { "epoch": 0.052517203911626224, "grad_norm": 1.992068373166419, "learning_rate": 1.997347793686285e-05, "loss": 0.9149, "step": 1450 }, { "epoch": 0.052553422672944584, "grad_norm": 1.9664830522842325, "learning_rate": 1.9973392488745965e-05, "loss": 0.9015, "step": 1451 }, { "epoch": 0.05258964143426295, "grad_norm": 1.9373533795238422, "learning_rate": 1.997330690338644e-05, "loss": 0.8725, "step": 1452 }, { "epoch": 0.05262586019558131, "grad_norm": 1.9004587943385485, "learning_rate": 1.9973221180785445e-05, "loss": 0.8544, "step": 1453 }, { "epoch": 0.05266207895689967, "grad_norm": 1.9112012849796487, "learning_rate": 1.9973135320944164e-05, "loss": 0.7943, "step": 1454 }, { "epoch": 0.05269829771821804, "grad_norm": 2.103957775262704, "learning_rate": 1.9973049323863783e-05, "loss": 0.8392, "step": 1455 }, { "epoch": 0.0527345164795364, "grad_norm": 2.0637153292371275, "learning_rate": 1.9972963189545474e-05, "loss": 0.9307, "step": 1456 }, { "epoch": 0.052770735240854766, "grad_norm": 2.1462446552708836, "learning_rate": 1.9972876917990433e-05, "loss": 0.8716, "step": 1457 }, { "epoch": 0.052806954002173126, "grad_norm": 1.7545060733592617, "learning_rate": 1.997279050919984e-05, "loss": 0.9572, "step": 1458 }, { "epoch": 0.052843172763491486, "grad_norm": 1.3544338829755966, "learning_rate": 1.9972703963174887e-05, "loss": 0.799, "step": 1459 }, { "epoch": 0.05287939152480985, "grad_norm": 1.7170775274166852, "learning_rate": 1.9972617279916765e-05, "loss": 0.8545, "step": 1460 }, { "epoch": 0.05291561028612821, "grad_norm": 1.820167339250892, "learning_rate": 1.9972530459426663e-05, "loss": 0.9267, "step": 1461 }, { "epoch": 0.05295182904744658, "grad_norm": 1.7980615595769658, "learning_rate": 1.9972443501705787e-05, "loss": 0.8977, "step": 1462 }, { "epoch": 0.05298804780876494, "grad_norm": 1.8966199421775276, "learning_rate": 1.997235640675532e-05, "loss": 0.9109, "step": 1463 }, { "epoch": 0.0530242665700833, "grad_norm": 1.432855373335421, "learning_rate": 1.997226917457647e-05, "loss": 0.8152, "step": 1464 }, { "epoch": 0.05306048533140167, "grad_norm": 1.8105787111876253, "learning_rate": 1.997218180517043e-05, "loss": 0.8896, "step": 1465 }, { "epoch": 0.05309670409272003, "grad_norm": 1.8848699471652433, "learning_rate": 1.9972094298538407e-05, "loss": 0.9357, "step": 1466 }, { "epoch": 0.053132922854038395, "grad_norm": 1.8615890962433965, "learning_rate": 1.9972006654681605e-05, "loss": 0.8844, "step": 1467 }, { "epoch": 0.053169141615356755, "grad_norm": 1.3246754229230417, "learning_rate": 1.9971918873601227e-05, "loss": 0.788, "step": 1468 }, { "epoch": 0.053205360376675115, "grad_norm": 1.4395593076075217, "learning_rate": 1.997183095529849e-05, "loss": 0.7926, "step": 1469 }, { "epoch": 0.05324157913799348, "grad_norm": 1.493428430331403, "learning_rate": 1.9971742899774588e-05, "loss": 0.7764, "step": 1470 }, { "epoch": 0.05327779789931184, "grad_norm": 1.963646444304361, "learning_rate": 1.997165470703075e-05, "loss": 0.8252, "step": 1471 }, { "epoch": 0.05331401666063021, "grad_norm": 1.7867634443656888, "learning_rate": 1.9971566377068176e-05, "loss": 0.9253, "step": 1472 }, { "epoch": 0.05335023542194857, "grad_norm": 1.9056111480788718, "learning_rate": 1.9971477909888088e-05, "loss": 0.9105, "step": 1473 }, { "epoch": 0.05338645418326693, "grad_norm": 1.9180679159691139, "learning_rate": 1.9971389305491706e-05, "loss": 0.9278, "step": 1474 }, { "epoch": 0.053422672944585296, "grad_norm": 1.869213888073843, "learning_rate": 1.9971300563880238e-05, "loss": 0.8723, "step": 1475 }, { "epoch": 0.053458891705903656, "grad_norm": 1.796830221841402, "learning_rate": 1.9971211685054918e-05, "loss": 0.9444, "step": 1476 }, { "epoch": 0.05349511046722202, "grad_norm": 1.9291463906004982, "learning_rate": 1.9971122669016962e-05, "loss": 0.9635, "step": 1477 }, { "epoch": 0.053531329228540384, "grad_norm": 1.7892238229788437, "learning_rate": 1.9971033515767597e-05, "loss": 0.8818, "step": 1478 }, { "epoch": 0.053567547989858744, "grad_norm": 1.979381296583254, "learning_rate": 1.997094422530805e-05, "loss": 0.8774, "step": 1479 }, { "epoch": 0.05360376675117711, "grad_norm": 1.7172849393209406, "learning_rate": 1.997085479763955e-05, "loss": 0.8351, "step": 1480 }, { "epoch": 0.05363998551249547, "grad_norm": 1.45530134455428, "learning_rate": 1.9970765232763325e-05, "loss": 0.8277, "step": 1481 }, { "epoch": 0.05367620427381384, "grad_norm": 2.1584202721906367, "learning_rate": 1.997067553068061e-05, "loss": 0.926, "step": 1482 }, { "epoch": 0.0537124230351322, "grad_norm": 1.9313581933266806, "learning_rate": 1.997058569139264e-05, "loss": 0.8669, "step": 1483 }, { "epoch": 0.05374864179645056, "grad_norm": 2.165581237397584, "learning_rate": 1.997049571490065e-05, "loss": 0.9931, "step": 1484 }, { "epoch": 0.053784860557768925, "grad_norm": 1.5159639296475536, "learning_rate": 1.9970405601205874e-05, "loss": 0.7562, "step": 1485 }, { "epoch": 0.053821079319087285, "grad_norm": 1.5699270480220922, "learning_rate": 1.997031535030956e-05, "loss": 0.8016, "step": 1486 }, { "epoch": 0.05385729808040565, "grad_norm": 1.8142910711480371, "learning_rate": 1.9970224962212944e-05, "loss": 0.7791, "step": 1487 }, { "epoch": 0.05389351684172401, "grad_norm": 1.882873357090046, "learning_rate": 1.9970134436917274e-05, "loss": 0.9697, "step": 1488 }, { "epoch": 0.05392973560304238, "grad_norm": 1.5436562284701993, "learning_rate": 1.9970043774423793e-05, "loss": 0.7571, "step": 1489 }, { "epoch": 0.05396595436436074, "grad_norm": 1.9642793629747979, "learning_rate": 1.9969952974733747e-05, "loss": 0.8801, "step": 1490 }, { "epoch": 0.0540021731256791, "grad_norm": 1.9148941199169065, "learning_rate": 1.996986203784839e-05, "loss": 0.8544, "step": 1491 }, { "epoch": 0.05403839188699747, "grad_norm": 1.7194780260035558, "learning_rate": 1.996977096376897e-05, "loss": 0.9271, "step": 1492 }, { "epoch": 0.05407461064831583, "grad_norm": 1.8962250344620168, "learning_rate": 1.9969679752496744e-05, "loss": 0.8212, "step": 1493 }, { "epoch": 0.054110829409634194, "grad_norm": 1.9817216735904901, "learning_rate": 1.9969588404032964e-05, "loss": 0.9235, "step": 1494 }, { "epoch": 0.054147048170952554, "grad_norm": 1.7965073261175974, "learning_rate": 1.9969496918378883e-05, "loss": 0.9314, "step": 1495 }, { "epoch": 0.054183266932270914, "grad_norm": 1.870680693744626, "learning_rate": 1.9969405295535768e-05, "loss": 0.8293, "step": 1496 }, { "epoch": 0.05421948569358928, "grad_norm": 1.52805064933198, "learning_rate": 1.9969313535504876e-05, "loss": 0.7939, "step": 1497 }, { "epoch": 0.05425570445490764, "grad_norm": 1.3915162428416725, "learning_rate": 1.996922163828747e-05, "loss": 0.8443, "step": 1498 }, { "epoch": 0.05429192321622601, "grad_norm": 2.0194215670965905, "learning_rate": 1.9969129603884812e-05, "loss": 0.9208, "step": 1499 }, { "epoch": 0.05432814197754437, "grad_norm": 1.9188771796325015, "learning_rate": 1.9969037432298175e-05, "loss": 0.8537, "step": 1500 }, { "epoch": 0.05436436073886273, "grad_norm": 1.8433889953300284, "learning_rate": 1.9968945123528823e-05, "loss": 0.8658, "step": 1501 }, { "epoch": 0.054400579500181095, "grad_norm": 1.5409109232027243, "learning_rate": 1.9968852677578026e-05, "loss": 0.806, "step": 1502 }, { "epoch": 0.054436798261499456, "grad_norm": 1.94298231310342, "learning_rate": 1.9968760094447055e-05, "loss": 0.9146, "step": 1503 }, { "epoch": 0.05447301702281782, "grad_norm": 1.7298578525192674, "learning_rate": 1.9968667374137184e-05, "loss": 0.8304, "step": 1504 }, { "epoch": 0.05450923578413618, "grad_norm": 1.6184092731849216, "learning_rate": 1.9968574516649694e-05, "loss": 0.8408, "step": 1505 }, { "epoch": 0.05454545454545454, "grad_norm": 2.0104669750339497, "learning_rate": 1.9968481521985858e-05, "loss": 0.9906, "step": 1506 }, { "epoch": 0.05458167330677291, "grad_norm": 1.8842509466310668, "learning_rate": 1.996838839014696e-05, "loss": 0.9002, "step": 1507 }, { "epoch": 0.05461789206809127, "grad_norm": 1.883265772151176, "learning_rate": 1.996829512113428e-05, "loss": 0.9233, "step": 1508 }, { "epoch": 0.05465411082940964, "grad_norm": 2.15907928495609, "learning_rate": 1.9968201714949095e-05, "loss": 0.9153, "step": 1509 }, { "epoch": 0.054690329590728, "grad_norm": 2.24406396703376, "learning_rate": 1.9968108171592698e-05, "loss": 0.9134, "step": 1510 }, { "epoch": 0.05472654835204636, "grad_norm": 1.9537010764273857, "learning_rate": 1.9968014491066373e-05, "loss": 0.9004, "step": 1511 }, { "epoch": 0.054762767113364724, "grad_norm": 2.000865693043831, "learning_rate": 1.996792067337141e-05, "loss": 0.9159, "step": 1512 }, { "epoch": 0.054798985874683084, "grad_norm": 1.764732587884135, "learning_rate": 1.99678267185091e-05, "loss": 0.8186, "step": 1513 }, { "epoch": 0.05483520463600145, "grad_norm": 1.639446953611231, "learning_rate": 1.9967732626480736e-05, "loss": 0.7969, "step": 1514 }, { "epoch": 0.05487142339731981, "grad_norm": 2.1387463944224696, "learning_rate": 1.9967638397287613e-05, "loss": 0.9868, "step": 1515 }, { "epoch": 0.05490764215863817, "grad_norm": 1.5660470609561759, "learning_rate": 1.9967544030931026e-05, "loss": 0.8076, "step": 1516 }, { "epoch": 0.05494386091995654, "grad_norm": 1.2296116520814184, "learning_rate": 1.9967449527412277e-05, "loss": 0.7967, "step": 1517 }, { "epoch": 0.0549800796812749, "grad_norm": 1.8736339042992036, "learning_rate": 1.9967354886732662e-05, "loss": 0.85, "step": 1518 }, { "epoch": 0.055016298442593266, "grad_norm": 1.9338996173266874, "learning_rate": 1.996726010889349e-05, "loss": 0.8835, "step": 1519 }, { "epoch": 0.055052517203911626, "grad_norm": 1.94583080943449, "learning_rate": 1.9967165193896055e-05, "loss": 0.9508, "step": 1520 }, { "epoch": 0.055088735965229986, "grad_norm": 1.4354352438107776, "learning_rate": 1.9967070141741672e-05, "loss": 0.7787, "step": 1521 }, { "epoch": 0.05512495472654835, "grad_norm": 1.9585129274202637, "learning_rate": 1.9966974952431645e-05, "loss": 0.8918, "step": 1522 }, { "epoch": 0.05516117348786671, "grad_norm": 1.8887340432855682, "learning_rate": 1.9966879625967288e-05, "loss": 0.9056, "step": 1523 }, { "epoch": 0.05519739224918508, "grad_norm": 1.99394467126158, "learning_rate": 1.9966784162349907e-05, "loss": 1.0161, "step": 1524 }, { "epoch": 0.05523361101050344, "grad_norm": 1.8421690636602868, "learning_rate": 1.996668856158082e-05, "loss": 0.8307, "step": 1525 }, { "epoch": 0.0552698297718218, "grad_norm": 1.4377174588728563, "learning_rate": 1.996659282366134e-05, "loss": 0.8489, "step": 1526 }, { "epoch": 0.05530604853314017, "grad_norm": 1.3882995753924956, "learning_rate": 1.9966496948592786e-05, "loss": 0.7471, "step": 1527 }, { "epoch": 0.05534226729445853, "grad_norm": 2.2394686756429185, "learning_rate": 1.9966400936376474e-05, "loss": 0.8796, "step": 1528 }, { "epoch": 0.055378486055776895, "grad_norm": 1.6749923873352952, "learning_rate": 1.996630478701373e-05, "loss": 0.7573, "step": 1529 }, { "epoch": 0.055414704817095255, "grad_norm": 2.0660645629689496, "learning_rate": 1.9966208500505874e-05, "loss": 0.8338, "step": 1530 }, { "epoch": 0.055450923578413615, "grad_norm": 1.9037506185677449, "learning_rate": 1.9966112076854235e-05, "loss": 1.0197, "step": 1531 }, { "epoch": 0.05548714233973198, "grad_norm": 2.039821419580871, "learning_rate": 1.996601551606013e-05, "loss": 0.7728, "step": 1532 }, { "epoch": 0.05552336110105034, "grad_norm": 1.6561795503100272, "learning_rate": 1.99659188181249e-05, "loss": 0.8189, "step": 1533 }, { "epoch": 0.05555957986236871, "grad_norm": 2.4381771614227294, "learning_rate": 1.996582198304987e-05, "loss": 0.9178, "step": 1534 }, { "epoch": 0.05559579862368707, "grad_norm": 1.7170836195100438, "learning_rate": 1.9965725010836373e-05, "loss": 0.8575, "step": 1535 }, { "epoch": 0.055632017385005436, "grad_norm": 2.0605893352165454, "learning_rate": 1.9965627901485742e-05, "loss": 0.9218, "step": 1536 }, { "epoch": 0.055668236146323796, "grad_norm": 2.056519098123852, "learning_rate": 1.9965530654999315e-05, "loss": 0.9118, "step": 1537 }, { "epoch": 0.055704454907642156, "grad_norm": 1.7800430466839359, "learning_rate": 1.9965433271378427e-05, "loss": 0.8396, "step": 1538 }, { "epoch": 0.05574067366896052, "grad_norm": 1.7023673252969176, "learning_rate": 1.9965335750624425e-05, "loss": 0.7892, "step": 1539 }, { "epoch": 0.055776892430278883, "grad_norm": 1.2729118056884592, "learning_rate": 1.9965238092738643e-05, "loss": 0.8246, "step": 1540 }, { "epoch": 0.05581311119159725, "grad_norm": 2.1435761574471566, "learning_rate": 1.996514029772243e-05, "loss": 0.9922, "step": 1541 }, { "epoch": 0.05584932995291561, "grad_norm": 1.8829111295010732, "learning_rate": 1.9965042365577134e-05, "loss": 0.9096, "step": 1542 }, { "epoch": 0.05588554871423397, "grad_norm": 1.7998505357272794, "learning_rate": 1.9964944296304095e-05, "loss": 0.8698, "step": 1543 }, { "epoch": 0.05592176747555234, "grad_norm": 1.718561772855279, "learning_rate": 1.9964846089904668e-05, "loss": 0.8262, "step": 1544 }, { "epoch": 0.0559579862368707, "grad_norm": 1.849992589516876, "learning_rate": 1.9964747746380202e-05, "loss": 0.8749, "step": 1545 }, { "epoch": 0.055994204998189065, "grad_norm": 2.0208059656036776, "learning_rate": 1.996464926573205e-05, "loss": 0.933, "step": 1546 }, { "epoch": 0.056030423759507425, "grad_norm": 1.843736713167581, "learning_rate": 1.9964550647961572e-05, "loss": 0.8349, "step": 1547 }, { "epoch": 0.056066642520825785, "grad_norm": 2.082878732951732, "learning_rate": 1.996445189307012e-05, "loss": 0.8606, "step": 1548 }, { "epoch": 0.05610286128214415, "grad_norm": 2.2212996943876875, "learning_rate": 1.996435300105905e-05, "loss": 0.9184, "step": 1549 }, { "epoch": 0.05613908004346251, "grad_norm": 3.4621812208903595, "learning_rate": 1.9964253971929733e-05, "loss": 0.9672, "step": 1550 }, { "epoch": 0.05617529880478088, "grad_norm": 1.7894832541097117, "learning_rate": 1.9964154805683528e-05, "loss": 0.8579, "step": 1551 }, { "epoch": 0.05621151756609924, "grad_norm": 1.9439512223780813, "learning_rate": 1.996405550232179e-05, "loss": 0.8443, "step": 1552 }, { "epoch": 0.0562477363274176, "grad_norm": 1.723812384612423, "learning_rate": 1.9963956061845897e-05, "loss": 0.7929, "step": 1553 }, { "epoch": 0.05628395508873597, "grad_norm": 1.7902609754696588, "learning_rate": 1.9963856484257213e-05, "loss": 0.8298, "step": 1554 }, { "epoch": 0.05632017385005433, "grad_norm": 1.7721464440317758, "learning_rate": 1.996375676955711e-05, "loss": 0.8632, "step": 1555 }, { "epoch": 0.056356392611372694, "grad_norm": 1.8497309659892942, "learning_rate": 1.9963656917746957e-05, "loss": 0.9363, "step": 1556 }, { "epoch": 0.056392611372691054, "grad_norm": 1.9728382404339466, "learning_rate": 1.9963556928828133e-05, "loss": 0.9099, "step": 1557 }, { "epoch": 0.056428830134009414, "grad_norm": 1.843678828659985, "learning_rate": 1.9963456802802007e-05, "loss": 0.8747, "step": 1558 }, { "epoch": 0.05646504889532778, "grad_norm": 1.5654372442203497, "learning_rate": 1.9963356539669965e-05, "loss": 0.7941, "step": 1559 }, { "epoch": 0.05650126765664614, "grad_norm": 2.0586452921037455, "learning_rate": 1.996325613943338e-05, "loss": 0.8238, "step": 1560 }, { "epoch": 0.05653748641796451, "grad_norm": 1.7571935994668058, "learning_rate": 1.9963155602093637e-05, "loss": 0.8684, "step": 1561 }, { "epoch": 0.05657370517928287, "grad_norm": 1.7154375989067638, "learning_rate": 1.9963054927652116e-05, "loss": 0.8273, "step": 1562 }, { "epoch": 0.05660992394060123, "grad_norm": 2.269653343291368, "learning_rate": 1.9962954116110205e-05, "loss": 0.8979, "step": 1563 }, { "epoch": 0.056646142701919595, "grad_norm": 1.8972863987285482, "learning_rate": 1.9962853167469293e-05, "loss": 0.8854, "step": 1564 }, { "epoch": 0.056682361463237955, "grad_norm": 1.8759971639364807, "learning_rate": 1.9962752081730768e-05, "loss": 0.8058, "step": 1565 }, { "epoch": 0.05671858022455632, "grad_norm": 1.458226236742193, "learning_rate": 1.996265085889602e-05, "loss": 0.7191, "step": 1566 }, { "epoch": 0.05675479898587468, "grad_norm": 1.90021626220472, "learning_rate": 1.9962549498966443e-05, "loss": 0.901, "step": 1567 }, { "epoch": 0.05679101774719304, "grad_norm": 1.978472635756745, "learning_rate": 1.9962448001943427e-05, "loss": 0.8605, "step": 1568 }, { "epoch": 0.05682723650851141, "grad_norm": 1.8823710028824867, "learning_rate": 1.9962346367828377e-05, "loss": 0.9376, "step": 1569 }, { "epoch": 0.05686345526982977, "grad_norm": 1.9436621329480737, "learning_rate": 1.9962244596622683e-05, "loss": 0.9176, "step": 1570 }, { "epoch": 0.05689967403114814, "grad_norm": 2.0954570234694647, "learning_rate": 1.9962142688327753e-05, "loss": 0.9587, "step": 1571 }, { "epoch": 0.0569358927924665, "grad_norm": 1.835135960619006, "learning_rate": 1.9962040642944986e-05, "loss": 0.8521, "step": 1572 }, { "epoch": 0.05697211155378486, "grad_norm": 1.786031095466876, "learning_rate": 1.9961938460475784e-05, "loss": 0.8519, "step": 1573 }, { "epoch": 0.057008330315103224, "grad_norm": 2.0067912716315246, "learning_rate": 1.9961836140921557e-05, "loss": 0.9159, "step": 1574 }, { "epoch": 0.057044549076421584, "grad_norm": 1.9885908571523325, "learning_rate": 1.9961733684283714e-05, "loss": 0.9103, "step": 1575 }, { "epoch": 0.05708076783773995, "grad_norm": 1.9486783392731335, "learning_rate": 1.996163109056366e-05, "loss": 0.8525, "step": 1576 }, { "epoch": 0.05711698659905831, "grad_norm": 2.1111397707730335, "learning_rate": 1.9961528359762806e-05, "loss": 0.8513, "step": 1577 }, { "epoch": 0.05715320536037668, "grad_norm": 1.6658015374943622, "learning_rate": 1.9961425491882573e-05, "loss": 0.7837, "step": 1578 }, { "epoch": 0.05718942412169504, "grad_norm": 2.0081157251263164, "learning_rate": 1.9961322486924372e-05, "loss": 0.8937, "step": 1579 }, { "epoch": 0.0572256428830134, "grad_norm": 1.8559009920018434, "learning_rate": 1.996121934488962e-05, "loss": 0.7851, "step": 1580 }, { "epoch": 0.057261861644331766, "grad_norm": 1.8011153768631087, "learning_rate": 1.9961116065779737e-05, "loss": 0.8383, "step": 1581 }, { "epoch": 0.057298080405650126, "grad_norm": 1.889835736753294, "learning_rate": 1.9961012649596144e-05, "loss": 0.839, "step": 1582 }, { "epoch": 0.05733429916696849, "grad_norm": 1.8590674266150353, "learning_rate": 1.9960909096340266e-05, "loss": 0.9471, "step": 1583 }, { "epoch": 0.05737051792828685, "grad_norm": 1.8024768433583025, "learning_rate": 1.9960805406013524e-05, "loss": 0.8817, "step": 1584 }, { "epoch": 0.05740673668960521, "grad_norm": 1.82291923466002, "learning_rate": 1.996070157861735e-05, "loss": 0.8196, "step": 1585 }, { "epoch": 0.05744295545092358, "grad_norm": 1.8889046035467225, "learning_rate": 1.9960597614153166e-05, "loss": 0.8696, "step": 1586 }, { "epoch": 0.05747917421224194, "grad_norm": 1.5617038471694968, "learning_rate": 1.9960493512622412e-05, "loss": 0.8972, "step": 1587 }, { "epoch": 0.05751539297356031, "grad_norm": 2.0129646979808404, "learning_rate": 1.996038927402651e-05, "loss": 0.9642, "step": 1588 }, { "epoch": 0.05755161173487867, "grad_norm": 2.1312052780784083, "learning_rate": 1.9960284898366904e-05, "loss": 0.8836, "step": 1589 }, { "epoch": 0.05758783049619703, "grad_norm": 1.947764970657622, "learning_rate": 1.9960180385645025e-05, "loss": 0.9425, "step": 1590 }, { "epoch": 0.057624049257515395, "grad_norm": 1.7319703680306586, "learning_rate": 1.9960075735862313e-05, "loss": 0.892, "step": 1591 }, { "epoch": 0.057660268018833755, "grad_norm": 1.9370363637540597, "learning_rate": 1.9959970949020203e-05, "loss": 0.7977, "step": 1592 }, { "epoch": 0.05769648678015212, "grad_norm": 2.009138226241218, "learning_rate": 1.9959866025120143e-05, "loss": 0.9167, "step": 1593 }, { "epoch": 0.05773270554147048, "grad_norm": 2.0155670385011235, "learning_rate": 1.9959760964163576e-05, "loss": 0.9065, "step": 1594 }, { "epoch": 0.05776892430278884, "grad_norm": 1.865914652134611, "learning_rate": 1.9959655766151948e-05, "loss": 0.7655, "step": 1595 }, { "epoch": 0.05780514306410721, "grad_norm": 1.7694560060229636, "learning_rate": 1.9959550431086703e-05, "loss": 0.8734, "step": 1596 }, { "epoch": 0.05784136182542557, "grad_norm": 1.8811742216037215, "learning_rate": 1.995944495896929e-05, "loss": 0.8455, "step": 1597 }, { "epoch": 0.057877580586743936, "grad_norm": 2.2611061742432152, "learning_rate": 1.995933934980117e-05, "loss": 0.9444, "step": 1598 }, { "epoch": 0.057913799348062296, "grad_norm": 1.77049657141625, "learning_rate": 1.9959233603583786e-05, "loss": 0.902, "step": 1599 }, { "epoch": 0.057950018109380656, "grad_norm": 1.7984835929498713, "learning_rate": 1.9959127720318598e-05, "loss": 0.9215, "step": 1600 }, { "epoch": 0.05798623687069902, "grad_norm": 1.8026359568436752, "learning_rate": 1.995902170000706e-05, "loss": 0.8808, "step": 1601 }, { "epoch": 0.05802245563201738, "grad_norm": 1.765324929994326, "learning_rate": 1.9958915542650632e-05, "loss": 0.8782, "step": 1602 }, { "epoch": 0.05805867439333575, "grad_norm": 2.2474112770071897, "learning_rate": 1.9958809248250775e-05, "loss": 0.8195, "step": 1603 }, { "epoch": 0.05809489315465411, "grad_norm": 1.8925524435279113, "learning_rate": 1.9958702816808956e-05, "loss": 0.8641, "step": 1604 }, { "epoch": 0.05813111191597247, "grad_norm": 1.9919846693267913, "learning_rate": 1.9958596248326633e-05, "loss": 0.8832, "step": 1605 }, { "epoch": 0.05816733067729084, "grad_norm": 1.8895065224755887, "learning_rate": 1.9958489542805278e-05, "loss": 0.9602, "step": 1606 }, { "epoch": 0.0582035494386092, "grad_norm": 1.869957917604994, "learning_rate": 1.995838270024635e-05, "loss": 0.7961, "step": 1607 }, { "epoch": 0.058239768199927565, "grad_norm": 1.9081216441380178, "learning_rate": 1.9958275720651335e-05, "loss": 0.9406, "step": 1608 }, { "epoch": 0.058275986961245925, "grad_norm": 1.9547139856661873, "learning_rate": 1.995816860402169e-05, "loss": 0.8401, "step": 1609 }, { "epoch": 0.058312205722564285, "grad_norm": 2.028546138455331, "learning_rate": 1.9958061350358896e-05, "loss": 0.8767, "step": 1610 }, { "epoch": 0.05834842448388265, "grad_norm": 2.009705829008854, "learning_rate": 1.9957953959664425e-05, "loss": 0.9222, "step": 1611 }, { "epoch": 0.05838464324520101, "grad_norm": 1.5547618064962385, "learning_rate": 1.9957846431939765e-05, "loss": 0.7983, "step": 1612 }, { "epoch": 0.05842086200651938, "grad_norm": 1.9653951125357763, "learning_rate": 1.9957738767186385e-05, "loss": 0.9168, "step": 1613 }, { "epoch": 0.05845708076783774, "grad_norm": 1.7133437376198892, "learning_rate": 1.9957630965405767e-05, "loss": 0.8678, "step": 1614 }, { "epoch": 0.0584932995291561, "grad_norm": 1.9501086823730882, "learning_rate": 1.99575230265994e-05, "loss": 0.8746, "step": 1615 }, { "epoch": 0.058529518290474467, "grad_norm": 1.68128106657338, "learning_rate": 1.995741495076877e-05, "loss": 0.808, "step": 1616 }, { "epoch": 0.05856573705179283, "grad_norm": 1.468564176471716, "learning_rate": 1.9957306737915353e-05, "loss": 0.7765, "step": 1617 }, { "epoch": 0.058601955813111194, "grad_norm": 1.5697745226325144, "learning_rate": 1.995719838804065e-05, "loss": 0.7897, "step": 1618 }, { "epoch": 0.058638174574429554, "grad_norm": 1.8651897924397818, "learning_rate": 1.995708990114615e-05, "loss": 0.8913, "step": 1619 }, { "epoch": 0.05867439333574792, "grad_norm": 1.873493439679944, "learning_rate": 1.9956981277233342e-05, "loss": 0.8763, "step": 1620 }, { "epoch": 0.05871061209706628, "grad_norm": 1.5736715841032667, "learning_rate": 1.995687251630372e-05, "loss": 0.8235, "step": 1621 }, { "epoch": 0.05874683085838464, "grad_norm": 2.2400941580381994, "learning_rate": 1.995676361835879e-05, "loss": 0.8491, "step": 1622 }, { "epoch": 0.05878304961970301, "grad_norm": 2.0004384726202784, "learning_rate": 1.9956654583400037e-05, "loss": 0.9658, "step": 1623 }, { "epoch": 0.05881926838102137, "grad_norm": 1.9098201116606266, "learning_rate": 1.9956545411428972e-05, "loss": 0.8931, "step": 1624 }, { "epoch": 0.058855487142339735, "grad_norm": 1.8374488869275973, "learning_rate": 1.9956436102447093e-05, "loss": 0.8835, "step": 1625 }, { "epoch": 0.058891705903658095, "grad_norm": 1.750687042827042, "learning_rate": 1.9956326656455903e-05, "loss": 0.9511, "step": 1626 }, { "epoch": 0.058927924664976455, "grad_norm": 1.880450882725766, "learning_rate": 1.995621707345691e-05, "loss": 0.9034, "step": 1627 }, { "epoch": 0.05896414342629482, "grad_norm": 1.5728769380080243, "learning_rate": 1.9956107353451622e-05, "loss": 0.8632, "step": 1628 }, { "epoch": 0.05900036218761318, "grad_norm": 2.009472624511837, "learning_rate": 1.9955997496441553e-05, "loss": 0.8284, "step": 1629 }, { "epoch": 0.05903658094893155, "grad_norm": 1.9477502856791125, "learning_rate": 1.9955887502428205e-05, "loss": 0.83, "step": 1630 }, { "epoch": 0.05907279971024991, "grad_norm": 1.2881429786502723, "learning_rate": 1.9955777371413102e-05, "loss": 0.7884, "step": 1631 }, { "epoch": 0.05910901847156827, "grad_norm": 1.507928240366036, "learning_rate": 1.9955667103397748e-05, "loss": 0.7973, "step": 1632 }, { "epoch": 0.05914523723288664, "grad_norm": 1.863566860565658, "learning_rate": 1.995555669838367e-05, "loss": 0.9055, "step": 1633 }, { "epoch": 0.059181455994205, "grad_norm": 2.087778311439372, "learning_rate": 1.9955446156372387e-05, "loss": 0.8534, "step": 1634 }, { "epoch": 0.059217674755523364, "grad_norm": 1.3513553026795664, "learning_rate": 1.9955335477365415e-05, "loss": 0.8438, "step": 1635 }, { "epoch": 0.059253893516841724, "grad_norm": 1.3022202573274038, "learning_rate": 1.9955224661364278e-05, "loss": 0.7862, "step": 1636 }, { "epoch": 0.059290112278160084, "grad_norm": 1.2297768278498038, "learning_rate": 1.9955113708370503e-05, "loss": 0.7174, "step": 1637 }, { "epoch": 0.05932633103947845, "grad_norm": 1.3789713252546192, "learning_rate": 1.9955002618385615e-05, "loss": 0.7609, "step": 1638 }, { "epoch": 0.05936254980079681, "grad_norm": 2.103797881263764, "learning_rate": 1.9954891391411146e-05, "loss": 0.8825, "step": 1639 }, { "epoch": 0.05939876856211518, "grad_norm": 1.839406170005743, "learning_rate": 1.9954780027448624e-05, "loss": 0.8155, "step": 1640 }, { "epoch": 0.05943498732343354, "grad_norm": 1.9338039823356663, "learning_rate": 1.995466852649958e-05, "loss": 0.9091, "step": 1641 }, { "epoch": 0.0594712060847519, "grad_norm": 1.9144232779719306, "learning_rate": 1.9954556888565554e-05, "loss": 0.86, "step": 1642 }, { "epoch": 0.059507424846070266, "grad_norm": 1.9877477147370248, "learning_rate": 1.9954445113648074e-05, "loss": 0.8765, "step": 1643 }, { "epoch": 0.059543643607388626, "grad_norm": 2.3662454852118042, "learning_rate": 1.9954333201748682e-05, "loss": 0.8137, "step": 1644 }, { "epoch": 0.05957986236870699, "grad_norm": 1.923520103233395, "learning_rate": 1.9954221152868922e-05, "loss": 0.944, "step": 1645 }, { "epoch": 0.05961608113002535, "grad_norm": 1.9893201987699944, "learning_rate": 1.995410896701033e-05, "loss": 0.9091, "step": 1646 }, { "epoch": 0.05965229989134371, "grad_norm": 1.6329881556352934, "learning_rate": 1.9953996644174453e-05, "loss": 0.7561, "step": 1647 }, { "epoch": 0.05968851865266208, "grad_norm": 1.479448725751299, "learning_rate": 1.9953884184362838e-05, "loss": 0.7815, "step": 1648 }, { "epoch": 0.05972473741398044, "grad_norm": 2.207106708823279, "learning_rate": 1.9953771587577027e-05, "loss": 0.9092, "step": 1649 }, { "epoch": 0.05976095617529881, "grad_norm": 1.834735539810779, "learning_rate": 1.9953658853818572e-05, "loss": 0.8429, "step": 1650 }, { "epoch": 0.05979717493661717, "grad_norm": 1.477497020181283, "learning_rate": 1.995354598308903e-05, "loss": 0.7754, "step": 1651 }, { "epoch": 0.05983339369793553, "grad_norm": 1.9336844297781886, "learning_rate": 1.9953432975389947e-05, "loss": 0.8845, "step": 1652 }, { "epoch": 0.059869612459253894, "grad_norm": 2.0676874324661476, "learning_rate": 1.995331983072288e-05, "loss": 0.9149, "step": 1653 }, { "epoch": 0.059905831220572255, "grad_norm": 1.8009434319776918, "learning_rate": 1.9953206549089386e-05, "loss": 0.8222, "step": 1654 }, { "epoch": 0.05994204998189062, "grad_norm": 1.8360334422020983, "learning_rate": 1.9953093130491024e-05, "loss": 0.9497, "step": 1655 }, { "epoch": 0.05997826874320898, "grad_norm": 2.0120135563508414, "learning_rate": 1.9952979574929355e-05, "loss": 0.8882, "step": 1656 }, { "epoch": 0.06001448750452734, "grad_norm": 1.7984344588037757, "learning_rate": 1.9952865882405942e-05, "loss": 0.8299, "step": 1657 }, { "epoch": 0.06005070626584571, "grad_norm": 1.9420954710868839, "learning_rate": 1.995275205292235e-05, "loss": 0.8721, "step": 1658 }, { "epoch": 0.06008692502716407, "grad_norm": 1.849221112266292, "learning_rate": 1.9952638086480146e-05, "loss": 0.8468, "step": 1659 }, { "epoch": 0.060123143788482436, "grad_norm": 1.8811947294841724, "learning_rate": 1.995252398308089e-05, "loss": 0.8437, "step": 1660 }, { "epoch": 0.060159362549800796, "grad_norm": 1.9202119689603272, "learning_rate": 1.9952409742726163e-05, "loss": 0.8155, "step": 1661 }, { "epoch": 0.06019558131111916, "grad_norm": 1.876707204810686, "learning_rate": 1.9952295365417537e-05, "loss": 0.926, "step": 1662 }, { "epoch": 0.06023180007243752, "grad_norm": 1.8676036496663209, "learning_rate": 1.995218085115658e-05, "loss": 0.9249, "step": 1663 }, { "epoch": 0.06026801883375588, "grad_norm": 1.8870526964388208, "learning_rate": 1.9952066199944867e-05, "loss": 0.8583, "step": 1664 }, { "epoch": 0.06030423759507425, "grad_norm": 1.9601798009008786, "learning_rate": 1.9951951411783977e-05, "loss": 0.906, "step": 1665 }, { "epoch": 0.06034045635639261, "grad_norm": 1.9503834365782544, "learning_rate": 1.9951836486675495e-05, "loss": 0.8698, "step": 1666 }, { "epoch": 0.06037667511771098, "grad_norm": 1.4323233461218043, "learning_rate": 1.9951721424621e-05, "loss": 0.7986, "step": 1667 }, { "epoch": 0.06041289387902934, "grad_norm": 1.249757141031397, "learning_rate": 1.9951606225622066e-05, "loss": 0.7747, "step": 1668 }, { "epoch": 0.0604491126403477, "grad_norm": 2.4646145828192774, "learning_rate": 1.995149088968029e-05, "loss": 0.8404, "step": 1669 }, { "epoch": 0.060485331401666065, "grad_norm": 1.9854192877015315, "learning_rate": 1.9951375416797258e-05, "loss": 0.8769, "step": 1670 }, { "epoch": 0.060521550162984425, "grad_norm": 1.3753294358771464, "learning_rate": 1.9951259806974558e-05, "loss": 0.7498, "step": 1671 }, { "epoch": 0.06055776892430279, "grad_norm": 2.1508715331766317, "learning_rate": 1.9951144060213773e-05, "loss": 0.8769, "step": 1672 }, { "epoch": 0.06059398768562115, "grad_norm": 2.1975047273877952, "learning_rate": 1.9951028176516505e-05, "loss": 0.9667, "step": 1673 }, { "epoch": 0.06063020644693951, "grad_norm": 1.6760865866428072, "learning_rate": 1.9950912155884344e-05, "loss": 0.784, "step": 1674 }, { "epoch": 0.06066642520825788, "grad_norm": 2.0361841520062205, "learning_rate": 1.9950795998318888e-05, "loss": 0.9411, "step": 1675 }, { "epoch": 0.06070264396957624, "grad_norm": 2.0306594540381453, "learning_rate": 1.9950679703821737e-05, "loss": 0.9285, "step": 1676 }, { "epoch": 0.060738862730894606, "grad_norm": 1.3497760533232992, "learning_rate": 1.995056327239449e-05, "loss": 0.7487, "step": 1677 }, { "epoch": 0.060775081492212966, "grad_norm": 3.5945854091937686, "learning_rate": 1.995044670403875e-05, "loss": 0.9095, "step": 1678 }, { "epoch": 0.06081130025353133, "grad_norm": 1.990713063811788, "learning_rate": 1.995032999875612e-05, "loss": 0.9082, "step": 1679 }, { "epoch": 0.060847519014849694, "grad_norm": 1.4517602183541314, "learning_rate": 1.9950213156548202e-05, "loss": 0.7722, "step": 1680 }, { "epoch": 0.060883737776168054, "grad_norm": 1.8868962109669638, "learning_rate": 1.995009617741661e-05, "loss": 0.8704, "step": 1681 }, { "epoch": 0.06091995653748642, "grad_norm": 2.146723649743842, "learning_rate": 1.9949979061362953e-05, "loss": 0.9256, "step": 1682 }, { "epoch": 0.06095617529880478, "grad_norm": 1.9549953488900693, "learning_rate": 1.994986180838884e-05, "loss": 0.9033, "step": 1683 }, { "epoch": 0.06099239406012314, "grad_norm": 1.8122638700673825, "learning_rate": 1.9949744418495886e-05, "loss": 0.7896, "step": 1684 }, { "epoch": 0.06102861282144151, "grad_norm": 2.017624997594337, "learning_rate": 1.994962689168571e-05, "loss": 0.9353, "step": 1685 }, { "epoch": 0.06106483158275987, "grad_norm": 1.8272152903247236, "learning_rate": 1.994950922795992e-05, "loss": 0.8516, "step": 1686 }, { "epoch": 0.061101050344078235, "grad_norm": 1.9954321197718374, "learning_rate": 1.9949391427320144e-05, "loss": 0.9842, "step": 1687 }, { "epoch": 0.061137269105396595, "grad_norm": 1.7964687302769393, "learning_rate": 1.9949273489767998e-05, "loss": 0.9154, "step": 1688 }, { "epoch": 0.061173487866714955, "grad_norm": 1.979832291144308, "learning_rate": 1.9949155415305105e-05, "loss": 0.874, "step": 1689 }, { "epoch": 0.06120970662803332, "grad_norm": 1.466450840966722, "learning_rate": 1.9949037203933095e-05, "loss": 0.8295, "step": 1690 }, { "epoch": 0.06124592538935168, "grad_norm": 1.9210358790150943, "learning_rate": 1.994891885565359e-05, "loss": 0.8828, "step": 1691 }, { "epoch": 0.06128214415067005, "grad_norm": 1.9321245487945833, "learning_rate": 1.9948800370468218e-05, "loss": 0.8697, "step": 1692 }, { "epoch": 0.06131836291198841, "grad_norm": 2.013443054163139, "learning_rate": 1.9948681748378612e-05, "loss": 0.9262, "step": 1693 }, { "epoch": 0.06135458167330677, "grad_norm": 2.1263839031805296, "learning_rate": 1.9948562989386406e-05, "loss": 0.9359, "step": 1694 }, { "epoch": 0.06139080043462514, "grad_norm": 2.2855016691841885, "learning_rate": 1.9948444093493227e-05, "loss": 0.8348, "step": 1695 }, { "epoch": 0.0614270191959435, "grad_norm": 2.015897456133464, "learning_rate": 1.994832506070072e-05, "loss": 0.8757, "step": 1696 }, { "epoch": 0.061463237957261864, "grad_norm": 1.9227096511062192, "learning_rate": 1.9948205891010515e-05, "loss": 0.9095, "step": 1697 }, { "epoch": 0.061499456718580224, "grad_norm": 1.8229839492483668, "learning_rate": 1.9948086584424258e-05, "loss": 0.8826, "step": 1698 }, { "epoch": 0.061535675479898584, "grad_norm": 1.8411427809884662, "learning_rate": 1.9947967140943588e-05, "loss": 0.8156, "step": 1699 }, { "epoch": 0.06157189424121695, "grad_norm": 1.8812223591231332, "learning_rate": 1.994784756057015e-05, "loss": 0.8208, "step": 1700 }, { "epoch": 0.06160811300253531, "grad_norm": 2.022241096944684, "learning_rate": 1.9947727843305585e-05, "loss": 0.8541, "step": 1701 }, { "epoch": 0.06164433176385368, "grad_norm": 1.6785652492432643, "learning_rate": 1.9947607989151548e-05, "loss": 0.7975, "step": 1702 }, { "epoch": 0.06168055052517204, "grad_norm": 1.9032271427467884, "learning_rate": 1.9947487998109682e-05, "loss": 0.7811, "step": 1703 }, { "epoch": 0.0617167692864904, "grad_norm": 1.9931189304383794, "learning_rate": 1.994736787018164e-05, "loss": 0.8677, "step": 1704 }, { "epoch": 0.061752988047808766, "grad_norm": 2.137583205558383, "learning_rate": 1.994724760536908e-05, "loss": 0.9316, "step": 1705 }, { "epoch": 0.061789206809127126, "grad_norm": 1.8645633559026988, "learning_rate": 1.994712720367365e-05, "loss": 0.9142, "step": 1706 }, { "epoch": 0.06182542557044549, "grad_norm": 1.5238911882389752, "learning_rate": 1.994700666509701e-05, "loss": 0.8121, "step": 1707 }, { "epoch": 0.06186164433176385, "grad_norm": 2.1214442464074272, "learning_rate": 1.9946885989640816e-05, "loss": 0.9329, "step": 1708 }, { "epoch": 0.06189786309308222, "grad_norm": 1.9081034404912347, "learning_rate": 1.994676517730673e-05, "loss": 0.8337, "step": 1709 }, { "epoch": 0.06193408185440058, "grad_norm": 1.847432242921877, "learning_rate": 1.9946644228096418e-05, "loss": 0.8912, "step": 1710 }, { "epoch": 0.06197030061571894, "grad_norm": 1.8313200837124721, "learning_rate": 1.9946523142011544e-05, "loss": 0.8634, "step": 1711 }, { "epoch": 0.06200651937703731, "grad_norm": 1.734944310270195, "learning_rate": 1.9946401919053765e-05, "loss": 0.8472, "step": 1712 }, { "epoch": 0.06204273813835567, "grad_norm": 1.9601929203729702, "learning_rate": 1.9946280559224763e-05, "loss": 0.8281, "step": 1713 }, { "epoch": 0.062078956899674034, "grad_norm": 2.3798064508032875, "learning_rate": 1.9946159062526197e-05, "loss": 0.9902, "step": 1714 }, { "epoch": 0.062115175660992394, "grad_norm": 3.201384340070061, "learning_rate": 1.9946037428959744e-05, "loss": 0.9416, "step": 1715 }, { "epoch": 0.062151394422310755, "grad_norm": 2.0133327296584205, "learning_rate": 1.9945915658527072e-05, "loss": 0.8473, "step": 1716 }, { "epoch": 0.06218761318362912, "grad_norm": 1.3107427197193662, "learning_rate": 1.9945793751229867e-05, "loss": 0.7913, "step": 1717 }, { "epoch": 0.06222383194494748, "grad_norm": 2.1460540104017176, "learning_rate": 1.9945671707069797e-05, "loss": 0.8853, "step": 1718 }, { "epoch": 0.06226005070626585, "grad_norm": 1.9750861036719607, "learning_rate": 1.994554952604855e-05, "loss": 0.9069, "step": 1719 }, { "epoch": 0.06229626946758421, "grad_norm": 1.3079807806589594, "learning_rate": 1.9945427208167797e-05, "loss": 0.7823, "step": 1720 }, { "epoch": 0.06233248822890257, "grad_norm": 1.7185457845223797, "learning_rate": 1.994530475342923e-05, "loss": 0.8552, "step": 1721 }, { "epoch": 0.062368706990220936, "grad_norm": 1.887221569483969, "learning_rate": 1.9945182161834532e-05, "loss": 0.8733, "step": 1722 }, { "epoch": 0.062404925751539296, "grad_norm": 1.9996853726240533, "learning_rate": 1.9945059433385388e-05, "loss": 0.9244, "step": 1723 }, { "epoch": 0.06244114451285766, "grad_norm": 1.977066618445406, "learning_rate": 1.9944936568083488e-05, "loss": 0.8351, "step": 1724 }, { "epoch": 0.06247736327417602, "grad_norm": 2.03122555461259, "learning_rate": 1.994481356593052e-05, "loss": 0.9508, "step": 1725 }, { "epoch": 0.06251358203549438, "grad_norm": 1.8538714610973905, "learning_rate": 1.9944690426928178e-05, "loss": 0.8341, "step": 1726 }, { "epoch": 0.06254980079681274, "grad_norm": 1.992295058469695, "learning_rate": 1.9944567151078164e-05, "loss": 0.9383, "step": 1727 }, { "epoch": 0.06258601955813112, "grad_norm": 1.9564814518624623, "learning_rate": 1.9944443738382164e-05, "loss": 0.9021, "step": 1728 }, { "epoch": 0.06262223831944948, "grad_norm": 2.1605939766524505, "learning_rate": 1.994432018884188e-05, "loss": 0.8722, "step": 1729 }, { "epoch": 0.06265845708076784, "grad_norm": 2.056331852251942, "learning_rate": 1.994419650245901e-05, "loss": 0.9615, "step": 1730 }, { "epoch": 0.0626946758420862, "grad_norm": 2.38703780687304, "learning_rate": 1.994407267923526e-05, "loss": 0.9247, "step": 1731 }, { "epoch": 0.06273089460340456, "grad_norm": 1.9658414694589879, "learning_rate": 1.9943948719172333e-05, "loss": 0.905, "step": 1732 }, { "epoch": 0.06276711336472293, "grad_norm": 2.0409717409079553, "learning_rate": 1.9943824622271934e-05, "loss": 0.921, "step": 1733 }, { "epoch": 0.06280333212604129, "grad_norm": 1.5856401092788681, "learning_rate": 1.9943700388535775e-05, "loss": 0.8118, "step": 1734 }, { "epoch": 0.06283955088735965, "grad_norm": 1.9059066697789007, "learning_rate": 1.9943576017965556e-05, "loss": 0.9033, "step": 1735 }, { "epoch": 0.06287576964867801, "grad_norm": 1.8162199776003292, "learning_rate": 1.9943451510562996e-05, "loss": 0.9075, "step": 1736 }, { "epoch": 0.06291198840999637, "grad_norm": 1.9996415930967657, "learning_rate": 1.9943326866329805e-05, "loss": 0.9167, "step": 1737 }, { "epoch": 0.06294820717131475, "grad_norm": 1.7153827052744264, "learning_rate": 1.9943202085267698e-05, "loss": 0.8113, "step": 1738 }, { "epoch": 0.0629844259326331, "grad_norm": 1.3909706012091163, "learning_rate": 1.9943077167378395e-05, "loss": 0.7803, "step": 1739 }, { "epoch": 0.06302064469395147, "grad_norm": 2.1381842471006536, "learning_rate": 1.9942952112663616e-05, "loss": 0.9609, "step": 1740 }, { "epoch": 0.06305686345526983, "grad_norm": 1.7956985768992277, "learning_rate": 1.9942826921125077e-05, "loss": 0.8278, "step": 1741 }, { "epoch": 0.06309308221658819, "grad_norm": 1.757954835957024, "learning_rate": 1.9942701592764503e-05, "loss": 0.9073, "step": 1742 }, { "epoch": 0.06312930097790656, "grad_norm": 1.9998153726102479, "learning_rate": 1.9942576127583618e-05, "loss": 0.975, "step": 1743 }, { "epoch": 0.06316551973922492, "grad_norm": 1.9164440381363557, "learning_rate": 1.9942450525584148e-05, "loss": 0.8148, "step": 1744 }, { "epoch": 0.06320173850054328, "grad_norm": 1.7982581174676662, "learning_rate": 1.9942324786767826e-05, "loss": 0.9254, "step": 1745 }, { "epoch": 0.06323795726186164, "grad_norm": 1.3735832645983508, "learning_rate": 1.9942198911136375e-05, "loss": 0.8031, "step": 1746 }, { "epoch": 0.06327417602318, "grad_norm": 1.269886333985568, "learning_rate": 1.9942072898691535e-05, "loss": 0.8011, "step": 1747 }, { "epoch": 0.06331039478449837, "grad_norm": 2.1017167272310573, "learning_rate": 1.994194674943503e-05, "loss": 0.8719, "step": 1748 }, { "epoch": 0.06334661354581674, "grad_norm": 1.951810772285233, "learning_rate": 1.9941820463368612e-05, "loss": 0.83, "step": 1749 }, { "epoch": 0.0633828323071351, "grad_norm": 2.0180718832359323, "learning_rate": 1.9941694040494002e-05, "loss": 0.8896, "step": 1750 }, { "epoch": 0.06341905106845346, "grad_norm": 1.81827730184141, "learning_rate": 1.9941567480812945e-05, "loss": 0.8477, "step": 1751 }, { "epoch": 0.06345526982977182, "grad_norm": 1.5289929326639373, "learning_rate": 1.9941440784327188e-05, "loss": 0.7891, "step": 1752 }, { "epoch": 0.06349148859109019, "grad_norm": 1.9950236608063001, "learning_rate": 1.994131395103847e-05, "loss": 0.8716, "step": 1753 }, { "epoch": 0.06352770735240855, "grad_norm": 1.309778090165105, "learning_rate": 1.9941186980948534e-05, "loss": 0.8011, "step": 1754 }, { "epoch": 0.06356392611372691, "grad_norm": 1.3340151421613877, "learning_rate": 1.9941059874059135e-05, "loss": 0.7976, "step": 1755 }, { "epoch": 0.06360014487504527, "grad_norm": 1.882978677341235, "learning_rate": 1.994093263037201e-05, "loss": 0.9231, "step": 1756 }, { "epoch": 0.06363636363636363, "grad_norm": 1.847271458799492, "learning_rate": 1.9940805249888923e-05, "loss": 0.9049, "step": 1757 }, { "epoch": 0.063672582397682, "grad_norm": 1.3450629349315746, "learning_rate": 1.994067773261162e-05, "loss": 0.813, "step": 1758 }, { "epoch": 0.06370880115900036, "grad_norm": 1.7656277597456436, "learning_rate": 1.9940550078541855e-05, "loss": 0.9085, "step": 1759 }, { "epoch": 0.06374501992031872, "grad_norm": 1.808859829248941, "learning_rate": 1.9940422287681388e-05, "loss": 0.8952, "step": 1760 }, { "epoch": 0.06378123868163708, "grad_norm": 2.0339712859381502, "learning_rate": 1.9940294360031974e-05, "loss": 0.9134, "step": 1761 }, { "epoch": 0.06381745744295544, "grad_norm": 2.0450420663119453, "learning_rate": 1.994016629559538e-05, "loss": 0.9841, "step": 1762 }, { "epoch": 0.06385367620427382, "grad_norm": 1.9864143101572107, "learning_rate": 1.9940038094373358e-05, "loss": 0.897, "step": 1763 }, { "epoch": 0.06388989496559218, "grad_norm": 3.117214585188979, "learning_rate": 1.993990975636768e-05, "loss": 0.8532, "step": 1764 }, { "epoch": 0.06392611372691054, "grad_norm": 1.807235489933198, "learning_rate": 1.9939781281580108e-05, "loss": 0.9105, "step": 1765 }, { "epoch": 0.0639623324882289, "grad_norm": 1.9479710105762933, "learning_rate": 1.9939652670012416e-05, "loss": 0.9458, "step": 1766 }, { "epoch": 0.06399855124954726, "grad_norm": 1.8364946747494793, "learning_rate": 1.9939523921666363e-05, "loss": 0.9277, "step": 1767 }, { "epoch": 0.06403477001086563, "grad_norm": 1.8210103821461112, "learning_rate": 1.9939395036543733e-05, "loss": 0.8899, "step": 1768 }, { "epoch": 0.06407098877218399, "grad_norm": 2.0404221926851194, "learning_rate": 1.993926601464629e-05, "loss": 0.9634, "step": 1769 }, { "epoch": 0.06410720753350235, "grad_norm": 1.8803987434155214, "learning_rate": 1.9939136855975815e-05, "loss": 0.8651, "step": 1770 }, { "epoch": 0.06414342629482071, "grad_norm": 2.5112308707655067, "learning_rate": 1.9939007560534084e-05, "loss": 0.9277, "step": 1771 }, { "epoch": 0.06417964505613909, "grad_norm": 1.9103344656687509, "learning_rate": 1.9938878128322875e-05, "loss": 0.8336, "step": 1772 }, { "epoch": 0.06421586381745745, "grad_norm": 2.162934470632219, "learning_rate": 1.993874855934397e-05, "loss": 0.9186, "step": 1773 }, { "epoch": 0.06425208257877581, "grad_norm": 1.8537183355300957, "learning_rate": 1.9938618853599155e-05, "loss": 0.8454, "step": 1774 }, { "epoch": 0.06428830134009417, "grad_norm": 1.9639477364720346, "learning_rate": 1.9938489011090208e-05, "loss": 0.86, "step": 1775 }, { "epoch": 0.06432452010141253, "grad_norm": 1.332743980075663, "learning_rate": 1.9938359031818918e-05, "loss": 0.8278, "step": 1776 }, { "epoch": 0.0643607388627309, "grad_norm": 1.891740587055998, "learning_rate": 1.993822891578708e-05, "loss": 0.8687, "step": 1777 }, { "epoch": 0.06439695762404926, "grad_norm": 1.7454198559867329, "learning_rate": 1.9938098662996475e-05, "loss": 0.7821, "step": 1778 }, { "epoch": 0.06443317638536762, "grad_norm": 1.3330104057580225, "learning_rate": 1.9937968273448904e-05, "loss": 0.7775, "step": 1779 }, { "epoch": 0.06446939514668598, "grad_norm": 2.0253942778360354, "learning_rate": 1.9937837747146155e-05, "loss": 0.863, "step": 1780 }, { "epoch": 0.06450561390800434, "grad_norm": 1.8676305896303054, "learning_rate": 1.993770708409003e-05, "loss": 0.8431, "step": 1781 }, { "epoch": 0.06454183266932272, "grad_norm": 1.939661626566411, "learning_rate": 1.9937576284282317e-05, "loss": 0.9406, "step": 1782 }, { "epoch": 0.06457805143064108, "grad_norm": 1.7570034348019237, "learning_rate": 1.9937445347724826e-05, "loss": 0.8931, "step": 1783 }, { "epoch": 0.06461427019195944, "grad_norm": 1.8657951271020894, "learning_rate": 1.9937314274419357e-05, "loss": 0.8582, "step": 1784 }, { "epoch": 0.0646504889532778, "grad_norm": 1.90986428291742, "learning_rate": 1.993718306436771e-05, "loss": 0.8591, "step": 1785 }, { "epoch": 0.06468670771459616, "grad_norm": 1.4219647347448563, "learning_rate": 1.9937051717571688e-05, "loss": 0.7957, "step": 1786 }, { "epoch": 0.06472292647591453, "grad_norm": 1.6948967887620254, "learning_rate": 1.993692023403311e-05, "loss": 0.9241, "step": 1787 }, { "epoch": 0.06475914523723289, "grad_norm": 1.6902244317644104, "learning_rate": 1.9936788613753774e-05, "loss": 0.884, "step": 1788 }, { "epoch": 0.06479536399855125, "grad_norm": 1.3056889523510486, "learning_rate": 1.9936656856735498e-05, "loss": 0.7834, "step": 1789 }, { "epoch": 0.06483158275986961, "grad_norm": 1.7690888508662164, "learning_rate": 1.993652496298009e-05, "loss": 0.8908, "step": 1790 }, { "epoch": 0.06486780152118797, "grad_norm": 1.9801056244942943, "learning_rate": 1.9936392932489372e-05, "loss": 0.9366, "step": 1791 }, { "epoch": 0.06490402028250634, "grad_norm": 1.8207724843443962, "learning_rate": 1.9936260765265152e-05, "loss": 0.8849, "step": 1792 }, { "epoch": 0.0649402390438247, "grad_norm": 1.2594598459104143, "learning_rate": 1.9936128461309258e-05, "loss": 0.7656, "step": 1793 }, { "epoch": 0.06497645780514306, "grad_norm": 1.2286821991657932, "learning_rate": 1.99359960206235e-05, "loss": 0.8267, "step": 1794 }, { "epoch": 0.06501267656646142, "grad_norm": 2.378070860007102, "learning_rate": 1.9935863443209707e-05, "loss": 0.8699, "step": 1795 }, { "epoch": 0.06504889532777978, "grad_norm": 1.9100783669599803, "learning_rate": 1.9935730729069703e-05, "loss": 0.9037, "step": 1796 }, { "epoch": 0.06508511408909816, "grad_norm": 1.8840657946041657, "learning_rate": 1.9935597878205316e-05, "loss": 0.8459, "step": 1797 }, { "epoch": 0.06512133285041652, "grad_norm": 1.994399312695122, "learning_rate": 1.9935464890618373e-05, "loss": 0.9437, "step": 1798 }, { "epoch": 0.06515755161173488, "grad_norm": 1.7803935053895523, "learning_rate": 1.99353317663107e-05, "loss": 0.8356, "step": 1799 }, { "epoch": 0.06519377037305324, "grad_norm": 2.3280734978170057, "learning_rate": 1.993519850528413e-05, "loss": 0.9497, "step": 1800 }, { "epoch": 0.0652299891343716, "grad_norm": 1.7735102949452637, "learning_rate": 1.9935065107540505e-05, "loss": 0.8797, "step": 1801 }, { "epoch": 0.06526620789568997, "grad_norm": 1.8426193593376663, "learning_rate": 1.9934931573081648e-05, "loss": 0.9419, "step": 1802 }, { "epoch": 0.06530242665700833, "grad_norm": 1.8444795619859116, "learning_rate": 1.993479790190941e-05, "loss": 0.8692, "step": 1803 }, { "epoch": 0.0653386454183267, "grad_norm": 1.9191135478346701, "learning_rate": 1.9934664094025616e-05, "loss": 0.8367, "step": 1804 }, { "epoch": 0.06537486417964505, "grad_norm": 2.2569916860005197, "learning_rate": 1.993453014943212e-05, "loss": 0.8838, "step": 1805 }, { "epoch": 0.06541108294096341, "grad_norm": 1.773260763159065, "learning_rate": 1.993439606813076e-05, "loss": 0.9126, "step": 1806 }, { "epoch": 0.06544730170228179, "grad_norm": 1.7330404379361888, "learning_rate": 1.9934261850123374e-05, "loss": 0.9317, "step": 1807 }, { "epoch": 0.06548352046360015, "grad_norm": 1.9700806752861675, "learning_rate": 1.9934127495411824e-05, "loss": 0.9323, "step": 1808 }, { "epoch": 0.06551973922491851, "grad_norm": 1.7912031531047787, "learning_rate": 1.9933993003997948e-05, "loss": 0.8163, "step": 1809 }, { "epoch": 0.06555595798623687, "grad_norm": 1.757312373963779, "learning_rate": 1.9933858375883597e-05, "loss": 0.8545, "step": 1810 }, { "epoch": 0.06559217674755523, "grad_norm": 1.8023989604071906, "learning_rate": 1.993372361107063e-05, "loss": 0.8724, "step": 1811 }, { "epoch": 0.0656283955088736, "grad_norm": 1.9527428100489224, "learning_rate": 1.9933588709560895e-05, "loss": 0.8591, "step": 1812 }, { "epoch": 0.06566461427019196, "grad_norm": 1.491645371390761, "learning_rate": 1.9933453671356254e-05, "loss": 0.8694, "step": 1813 }, { "epoch": 0.06570083303151032, "grad_norm": 1.897281235741993, "learning_rate": 1.9933318496458557e-05, "loss": 0.8321, "step": 1814 }, { "epoch": 0.06573705179282868, "grad_norm": 1.9229580029755844, "learning_rate": 1.9933183184869675e-05, "loss": 0.8986, "step": 1815 }, { "epoch": 0.06577327055414704, "grad_norm": 1.8800675774473983, "learning_rate": 1.993304773659146e-05, "loss": 0.8971, "step": 1816 }, { "epoch": 0.06580948931546542, "grad_norm": 1.8121829217999958, "learning_rate": 1.9932912151625782e-05, "loss": 0.8568, "step": 1817 }, { "epoch": 0.06584570807678378, "grad_norm": 1.9449751659125007, "learning_rate": 1.9932776429974505e-05, "loss": 0.9235, "step": 1818 }, { "epoch": 0.06588192683810214, "grad_norm": 1.9973757512592931, "learning_rate": 1.9932640571639496e-05, "loss": 0.9433, "step": 1819 }, { "epoch": 0.0659181455994205, "grad_norm": 1.9442938757446517, "learning_rate": 1.9932504576622626e-05, "loss": 0.8567, "step": 1820 }, { "epoch": 0.06595436436073886, "grad_norm": 1.9303047702573402, "learning_rate": 1.9932368444925765e-05, "loss": 0.8945, "step": 1821 }, { "epoch": 0.06599058312205723, "grad_norm": 1.4720768908262203, "learning_rate": 1.9932232176550787e-05, "loss": 0.7701, "step": 1822 }, { "epoch": 0.06602680188337559, "grad_norm": 1.9624501422073461, "learning_rate": 1.993209577149957e-05, "loss": 0.8676, "step": 1823 }, { "epoch": 0.06606302064469395, "grad_norm": 1.7078121266365733, "learning_rate": 1.9931959229773983e-05, "loss": 0.8932, "step": 1824 }, { "epoch": 0.06609923940601231, "grad_norm": 1.8777010030064447, "learning_rate": 1.9931822551375912e-05, "loss": 0.8833, "step": 1825 }, { "epoch": 0.06613545816733067, "grad_norm": 1.8427245118924451, "learning_rate": 1.9931685736307235e-05, "loss": 0.8951, "step": 1826 }, { "epoch": 0.06617167692864905, "grad_norm": 1.9264975399549995, "learning_rate": 1.9931548784569836e-05, "loss": 0.9591, "step": 1827 }, { "epoch": 0.0662078956899674, "grad_norm": 1.3865755474838803, "learning_rate": 1.9931411696165602e-05, "loss": 0.8755, "step": 1828 }, { "epoch": 0.06624411445128577, "grad_norm": 1.8103240101259201, "learning_rate": 1.9931274471096414e-05, "loss": 0.9015, "step": 1829 }, { "epoch": 0.06628033321260413, "grad_norm": 1.462221758985554, "learning_rate": 1.9931137109364162e-05, "loss": 0.7634, "step": 1830 }, { "epoch": 0.06631655197392249, "grad_norm": 1.8902859777922345, "learning_rate": 1.9930999610970737e-05, "loss": 0.9025, "step": 1831 }, { "epoch": 0.06635277073524086, "grad_norm": 1.584306626518664, "learning_rate": 1.9930861975918035e-05, "loss": 0.8159, "step": 1832 }, { "epoch": 0.06638898949655922, "grad_norm": 1.8032961607556963, "learning_rate": 1.9930724204207942e-05, "loss": 0.8596, "step": 1833 }, { "epoch": 0.06642520825787758, "grad_norm": 1.4396866228261085, "learning_rate": 1.993058629584236e-05, "loss": 0.797, "step": 1834 }, { "epoch": 0.06646142701919594, "grad_norm": 1.9643681748842379, "learning_rate": 1.9930448250823186e-05, "loss": 0.8731, "step": 1835 }, { "epoch": 0.0664976457805143, "grad_norm": 1.4693827269668693, "learning_rate": 1.9930310069152314e-05, "loss": 0.7857, "step": 1836 }, { "epoch": 0.06653386454183267, "grad_norm": 1.848366056643258, "learning_rate": 1.9930171750831653e-05, "loss": 0.9136, "step": 1837 }, { "epoch": 0.06657008330315103, "grad_norm": 2.1564304485349486, "learning_rate": 1.9930033295863104e-05, "loss": 0.9044, "step": 1838 }, { "epoch": 0.0666063020644694, "grad_norm": 1.8908829256938047, "learning_rate": 1.992989470424857e-05, "loss": 0.8081, "step": 1839 }, { "epoch": 0.06664252082578775, "grad_norm": 1.876912829098543, "learning_rate": 1.992975597598996e-05, "loss": 0.9833, "step": 1840 }, { "epoch": 0.06667873958710611, "grad_norm": 1.7243550119498041, "learning_rate": 1.9929617111089184e-05, "loss": 0.8847, "step": 1841 }, { "epoch": 0.06671495834842449, "grad_norm": 2.16172130930604, "learning_rate": 1.992947810954815e-05, "loss": 0.9089, "step": 1842 }, { "epoch": 0.06675117710974285, "grad_norm": 1.887855501280336, "learning_rate": 1.992933897136877e-05, "loss": 0.8639, "step": 1843 }, { "epoch": 0.06678739587106121, "grad_norm": 2.0614507374126383, "learning_rate": 1.9929199696552966e-05, "loss": 0.8236, "step": 1844 }, { "epoch": 0.06682361463237957, "grad_norm": 2.08081030249804, "learning_rate": 1.992906028510265e-05, "loss": 0.9452, "step": 1845 }, { "epoch": 0.06685983339369793, "grad_norm": 2.063777961148747, "learning_rate": 1.9928920737019735e-05, "loss": 0.8485, "step": 1846 }, { "epoch": 0.0668960521550163, "grad_norm": 1.7627717569529031, "learning_rate": 1.992878105230615e-05, "loss": 0.881, "step": 1847 }, { "epoch": 0.06693227091633466, "grad_norm": 1.7747150831334277, "learning_rate": 1.992864123096381e-05, "loss": 0.8485, "step": 1848 }, { "epoch": 0.06696848967765302, "grad_norm": 1.813549856821644, "learning_rate": 1.9928501272994646e-05, "loss": 0.8539, "step": 1849 }, { "epoch": 0.06700470843897138, "grad_norm": 1.9491407487739183, "learning_rate": 1.9928361178400582e-05, "loss": 0.9045, "step": 1850 }, { "epoch": 0.06704092720028974, "grad_norm": 1.7822079634022554, "learning_rate": 1.992822094718354e-05, "loss": 0.9272, "step": 1851 }, { "epoch": 0.06707714596160812, "grad_norm": 1.9943068264140498, "learning_rate": 1.992808057934546e-05, "loss": 0.8482, "step": 1852 }, { "epoch": 0.06711336472292648, "grad_norm": 1.775486158773917, "learning_rate": 1.9927940074888262e-05, "loss": 0.9549, "step": 1853 }, { "epoch": 0.06714958348424484, "grad_norm": 1.8087232390943702, "learning_rate": 1.9927799433813887e-05, "loss": 0.8579, "step": 1854 }, { "epoch": 0.0671858022455632, "grad_norm": 1.7500108601188449, "learning_rate": 1.9927658656124268e-05, "loss": 0.8081, "step": 1855 }, { "epoch": 0.06722202100688157, "grad_norm": 1.7256947047566842, "learning_rate": 1.9927517741821343e-05, "loss": 0.9112, "step": 1856 }, { "epoch": 0.06725823976819993, "grad_norm": 1.5295708825406509, "learning_rate": 1.992737669090705e-05, "loss": 0.7735, "step": 1857 }, { "epoch": 0.06729445852951829, "grad_norm": 1.9928214867141012, "learning_rate": 1.9927235503383333e-05, "loss": 0.7363, "step": 1858 }, { "epoch": 0.06733067729083665, "grad_norm": 1.8678108834662868, "learning_rate": 1.9927094179252132e-05, "loss": 0.9628, "step": 1859 }, { "epoch": 0.06736689605215501, "grad_norm": 2.088483808236078, "learning_rate": 1.9926952718515392e-05, "loss": 0.8456, "step": 1860 }, { "epoch": 0.06740311481347339, "grad_norm": 2.1036558821733222, "learning_rate": 1.9926811121175057e-05, "loss": 0.8895, "step": 1861 }, { "epoch": 0.06743933357479175, "grad_norm": 2.115548391448331, "learning_rate": 1.9926669387233084e-05, "loss": 0.8689, "step": 1862 }, { "epoch": 0.0674755523361101, "grad_norm": 1.8655784609116095, "learning_rate": 1.9926527516691413e-05, "loss": 0.9404, "step": 1863 }, { "epoch": 0.06751177109742847, "grad_norm": 1.8331979156101486, "learning_rate": 1.9926385509552004e-05, "loss": 0.8542, "step": 1864 }, { "epoch": 0.06754798985874683, "grad_norm": 1.7897920453692804, "learning_rate": 1.9926243365816806e-05, "loss": 0.8904, "step": 1865 }, { "epoch": 0.0675842086200652, "grad_norm": 2.045253342413354, "learning_rate": 1.992610108548778e-05, "loss": 0.8291, "step": 1866 }, { "epoch": 0.06762042738138356, "grad_norm": 1.7930240061604132, "learning_rate": 1.992595866856688e-05, "loss": 0.8483, "step": 1867 }, { "epoch": 0.06765664614270192, "grad_norm": 1.803026032644912, "learning_rate": 1.992581611505607e-05, "loss": 0.8259, "step": 1868 }, { "epoch": 0.06769286490402028, "grad_norm": 1.8706648231435818, "learning_rate": 1.9925673424957305e-05, "loss": 0.8939, "step": 1869 }, { "epoch": 0.06772908366533864, "grad_norm": 1.9478614004772756, "learning_rate": 1.992553059827255e-05, "loss": 0.9638, "step": 1870 }, { "epoch": 0.06776530242665701, "grad_norm": 1.8353158202613014, "learning_rate": 1.9925387635003777e-05, "loss": 0.8599, "step": 1871 }, { "epoch": 0.06780152118797537, "grad_norm": 1.4444416247989176, "learning_rate": 1.9925244535152947e-05, "loss": 0.8218, "step": 1872 }, { "epoch": 0.06783773994929373, "grad_norm": 1.747465213775945, "learning_rate": 1.992510129872203e-05, "loss": 0.8312, "step": 1873 }, { "epoch": 0.0678739587106121, "grad_norm": 1.6451627171364216, "learning_rate": 1.9924957925713003e-05, "loss": 0.8172, "step": 1874 }, { "epoch": 0.06791017747193046, "grad_norm": 1.7409371702642067, "learning_rate": 1.992481441612783e-05, "loss": 0.9039, "step": 1875 }, { "epoch": 0.06794639623324883, "grad_norm": 1.768997386584136, "learning_rate": 1.992467076996849e-05, "loss": 0.8474, "step": 1876 }, { "epoch": 0.06798261499456719, "grad_norm": 1.5631793226273203, "learning_rate": 1.992452698723696e-05, "loss": 0.8285, "step": 1877 }, { "epoch": 0.06801883375588555, "grad_norm": 1.8063466157614783, "learning_rate": 1.992438306793522e-05, "loss": 0.8836, "step": 1878 }, { "epoch": 0.06805505251720391, "grad_norm": 1.438230667265361, "learning_rate": 1.9924239012065247e-05, "loss": 0.8323, "step": 1879 }, { "epoch": 0.06809127127852227, "grad_norm": 1.4777962891865668, "learning_rate": 1.9924094819629024e-05, "loss": 0.7703, "step": 1880 }, { "epoch": 0.06812749003984064, "grad_norm": 1.4962471988795865, "learning_rate": 1.992395049062854e-05, "loss": 0.8329, "step": 1881 }, { "epoch": 0.068163708801159, "grad_norm": 1.9343429038939342, "learning_rate": 1.9923806025065775e-05, "loss": 0.8621, "step": 1882 }, { "epoch": 0.06819992756247736, "grad_norm": 1.9780977604332906, "learning_rate": 1.992366142294272e-05, "loss": 0.9181, "step": 1883 }, { "epoch": 0.06823614632379572, "grad_norm": 1.8662108646394806, "learning_rate": 1.992351668426136e-05, "loss": 0.8383, "step": 1884 }, { "epoch": 0.06827236508511408, "grad_norm": 1.8825237994633688, "learning_rate": 1.9923371809023696e-05, "loss": 0.8117, "step": 1885 }, { "epoch": 0.06830858384643246, "grad_norm": 1.9328594075249173, "learning_rate": 1.9923226797231713e-05, "loss": 0.7932, "step": 1886 }, { "epoch": 0.06834480260775082, "grad_norm": 1.5627388989831892, "learning_rate": 1.9923081648887415e-05, "loss": 0.8627, "step": 1887 }, { "epoch": 0.06838102136906918, "grad_norm": 2.20895369868476, "learning_rate": 1.992293636399279e-05, "loss": 0.9474, "step": 1888 }, { "epoch": 0.06841724013038754, "grad_norm": 1.3653414169937341, "learning_rate": 1.9922790942549842e-05, "loss": 0.7923, "step": 1889 }, { "epoch": 0.0684534588917059, "grad_norm": 1.3373626904704063, "learning_rate": 1.9922645384560572e-05, "loss": 0.8008, "step": 1890 }, { "epoch": 0.06848967765302427, "grad_norm": 1.747343119884078, "learning_rate": 1.9922499690026983e-05, "loss": 0.8541, "step": 1891 }, { "epoch": 0.06852589641434263, "grad_norm": 1.8344399929311723, "learning_rate": 1.992235385895108e-05, "loss": 0.8739, "step": 1892 }, { "epoch": 0.06856211517566099, "grad_norm": 1.7755157040557075, "learning_rate": 1.9922207891334867e-05, "loss": 0.8587, "step": 1893 }, { "epoch": 0.06859833393697935, "grad_norm": 1.7356994400435155, "learning_rate": 1.992206178718036e-05, "loss": 0.8043, "step": 1894 }, { "epoch": 0.06863455269829771, "grad_norm": 1.7590359986261184, "learning_rate": 1.9921915546489558e-05, "loss": 0.831, "step": 1895 }, { "epoch": 0.06867077145961609, "grad_norm": 1.8614895721579452, "learning_rate": 1.9921769169264486e-05, "loss": 0.8514, "step": 1896 }, { "epoch": 0.06870699022093445, "grad_norm": 1.7836576490447695, "learning_rate": 1.992162265550715e-05, "loss": 0.8446, "step": 1897 }, { "epoch": 0.06874320898225281, "grad_norm": 1.648220563236581, "learning_rate": 1.9921476005219566e-05, "loss": 0.8512, "step": 1898 }, { "epoch": 0.06877942774357117, "grad_norm": 1.7870133327253994, "learning_rate": 1.9921329218403757e-05, "loss": 0.8379, "step": 1899 }, { "epoch": 0.06881564650488953, "grad_norm": 1.8009498247321942, "learning_rate": 1.9921182295061737e-05, "loss": 0.9609, "step": 1900 }, { "epoch": 0.0688518652662079, "grad_norm": 1.8742383671648628, "learning_rate": 1.9921035235195537e-05, "loss": 0.848, "step": 1901 }, { "epoch": 0.06888808402752626, "grad_norm": 1.7806353522887182, "learning_rate": 1.9920888038807172e-05, "loss": 0.8787, "step": 1902 }, { "epoch": 0.06892430278884462, "grad_norm": 1.8711442771333113, "learning_rate": 1.9920740705898668e-05, "loss": 0.8998, "step": 1903 }, { "epoch": 0.06896052155016298, "grad_norm": 1.4622682333193482, "learning_rate": 1.9920593236472057e-05, "loss": 0.8181, "step": 1904 }, { "epoch": 0.06899674031148134, "grad_norm": 1.9430699444422401, "learning_rate": 1.9920445630529367e-05, "loss": 0.8378, "step": 1905 }, { "epoch": 0.06903295907279972, "grad_norm": 1.6701455262088165, "learning_rate": 1.9920297888072623e-05, "loss": 0.8545, "step": 1906 }, { "epoch": 0.06906917783411808, "grad_norm": 1.6298909235684467, "learning_rate": 1.992015000910387e-05, "loss": 0.8022, "step": 1907 }, { "epoch": 0.06910539659543644, "grad_norm": 1.2627998889211027, "learning_rate": 1.9920001993625133e-05, "loss": 0.8088, "step": 1908 }, { "epoch": 0.0691416153567548, "grad_norm": 2.3197567688472893, "learning_rate": 1.9919853841638452e-05, "loss": 0.927, "step": 1909 }, { "epoch": 0.06917783411807316, "grad_norm": 1.3391788114445777, "learning_rate": 1.9919705553145866e-05, "loss": 0.8231, "step": 1910 }, { "epoch": 0.06921405287939153, "grad_norm": 1.7272932762973887, "learning_rate": 1.9919557128149418e-05, "loss": 0.9214, "step": 1911 }, { "epoch": 0.06925027164070989, "grad_norm": 2.0149479808375643, "learning_rate": 1.9919408566651143e-05, "loss": 0.8814, "step": 1912 }, { "epoch": 0.06928649040202825, "grad_norm": 1.6974221213373897, "learning_rate": 1.9919259868653093e-05, "loss": 0.9043, "step": 1913 }, { "epoch": 0.06932270916334661, "grad_norm": 1.8687569663944976, "learning_rate": 1.9919111034157312e-05, "loss": 0.9608, "step": 1914 }, { "epoch": 0.06935892792466497, "grad_norm": 1.9183617711659082, "learning_rate": 1.9918962063165845e-05, "loss": 0.9278, "step": 1915 }, { "epoch": 0.06939514668598334, "grad_norm": 1.237087433441703, "learning_rate": 1.9918812955680748e-05, "loss": 0.8224, "step": 1916 }, { "epoch": 0.0694313654473017, "grad_norm": 1.9297397631544675, "learning_rate": 1.9918663711704067e-05, "loss": 0.8587, "step": 1917 }, { "epoch": 0.06946758420862006, "grad_norm": 1.6785785296617475, "learning_rate": 1.991851433123786e-05, "loss": 0.8683, "step": 1918 }, { "epoch": 0.06950380296993842, "grad_norm": 1.836508455167708, "learning_rate": 1.9918364814284177e-05, "loss": 0.9127, "step": 1919 }, { "epoch": 0.06954002173125678, "grad_norm": 1.7875587231987042, "learning_rate": 1.991821516084508e-05, "loss": 0.8584, "step": 1920 }, { "epoch": 0.06957624049257516, "grad_norm": 1.9106075498734298, "learning_rate": 1.9918065370922632e-05, "loss": 0.8902, "step": 1921 }, { "epoch": 0.06961245925389352, "grad_norm": 1.9009634587561253, "learning_rate": 1.991791544451888e-05, "loss": 0.8941, "step": 1922 }, { "epoch": 0.06964867801521188, "grad_norm": 1.6750908931687276, "learning_rate": 1.9917765381635907e-05, "loss": 0.8356, "step": 1923 }, { "epoch": 0.06968489677653024, "grad_norm": 1.7972593134931127, "learning_rate": 1.9917615182275766e-05, "loss": 0.8469, "step": 1924 }, { "epoch": 0.0697211155378486, "grad_norm": 2.087996941367372, "learning_rate": 1.991746484644052e-05, "loss": 0.931, "step": 1925 }, { "epoch": 0.06975733429916697, "grad_norm": 1.9615269189256925, "learning_rate": 1.991731437413225e-05, "loss": 0.8378, "step": 1926 }, { "epoch": 0.06979355306048533, "grad_norm": 1.456967767912596, "learning_rate": 1.9917163765353016e-05, "loss": 0.7553, "step": 1927 }, { "epoch": 0.0698297718218037, "grad_norm": 1.8259872690005863, "learning_rate": 1.9917013020104898e-05, "loss": 0.8312, "step": 1928 }, { "epoch": 0.06986599058312205, "grad_norm": 1.8895112631045197, "learning_rate": 1.9916862138389962e-05, "loss": 0.9044, "step": 1929 }, { "epoch": 0.06990220934444041, "grad_norm": 1.9366905719312186, "learning_rate": 1.991671112021029e-05, "loss": 0.8611, "step": 1930 }, { "epoch": 0.06993842810575879, "grad_norm": 1.6542009665871082, "learning_rate": 1.9916559965567964e-05, "loss": 0.7967, "step": 1931 }, { "epoch": 0.06997464686707715, "grad_norm": 1.3370442579571589, "learning_rate": 1.9916408674465058e-05, "loss": 0.8314, "step": 1932 }, { "epoch": 0.07001086562839551, "grad_norm": 1.3376295313532807, "learning_rate": 1.9916257246903653e-05, "loss": 0.7817, "step": 1933 }, { "epoch": 0.07004708438971387, "grad_norm": 1.8996111093411616, "learning_rate": 1.9916105682885837e-05, "loss": 0.9415, "step": 1934 }, { "epoch": 0.07008330315103223, "grad_norm": 1.768708794765918, "learning_rate": 1.9915953982413693e-05, "loss": 0.8493, "step": 1935 }, { "epoch": 0.0701195219123506, "grad_norm": 1.332512283794729, "learning_rate": 1.991580214548931e-05, "loss": 0.7757, "step": 1936 }, { "epoch": 0.07015574067366896, "grad_norm": 1.8859254835863195, "learning_rate": 1.9915650172114777e-05, "loss": 0.8083, "step": 1937 }, { "epoch": 0.07019195943498732, "grad_norm": 1.4904691475613654, "learning_rate": 1.9915498062292182e-05, "loss": 0.8111, "step": 1938 }, { "epoch": 0.07022817819630568, "grad_norm": 1.4534429436593297, "learning_rate": 1.9915345816023622e-05, "loss": 0.8193, "step": 1939 }, { "epoch": 0.07026439695762404, "grad_norm": 1.9696945127301522, "learning_rate": 1.9915193433311196e-05, "loss": 0.8666, "step": 1940 }, { "epoch": 0.07030061571894242, "grad_norm": 1.8132244935330415, "learning_rate": 1.9915040914156992e-05, "loss": 0.9071, "step": 1941 }, { "epoch": 0.07033683448026078, "grad_norm": 1.7375360391673318, "learning_rate": 1.991488825856311e-05, "loss": 0.8875, "step": 1942 }, { "epoch": 0.07037305324157914, "grad_norm": 1.5097966132370395, "learning_rate": 1.991473546653166e-05, "loss": 0.7638, "step": 1943 }, { "epoch": 0.0704092720028975, "grad_norm": 1.7804009009519228, "learning_rate": 1.9914582538064735e-05, "loss": 0.9006, "step": 1944 }, { "epoch": 0.07044549076421587, "grad_norm": 1.7828658141105669, "learning_rate": 1.9914429473164442e-05, "loss": 0.8844, "step": 1945 }, { "epoch": 0.07048170952553423, "grad_norm": 1.738758504908115, "learning_rate": 1.9914276271832888e-05, "loss": 0.9137, "step": 1946 }, { "epoch": 0.07051792828685259, "grad_norm": 1.7117934306892946, "learning_rate": 1.9914122934072182e-05, "loss": 0.8976, "step": 1947 }, { "epoch": 0.07055414704817095, "grad_norm": 2.334848957256358, "learning_rate": 1.991396945988443e-05, "loss": 0.8754, "step": 1948 }, { "epoch": 0.07059036580948931, "grad_norm": 2.1296235407281143, "learning_rate": 1.991381584927175e-05, "loss": 0.9843, "step": 1949 }, { "epoch": 0.07062658457080769, "grad_norm": 1.9980820981796275, "learning_rate": 1.9913662102236255e-05, "loss": 0.8896, "step": 1950 }, { "epoch": 0.07066280333212605, "grad_norm": 1.8310044935340317, "learning_rate": 1.9913508218780055e-05, "loss": 0.7971, "step": 1951 }, { "epoch": 0.0706990220934444, "grad_norm": 1.7441963932310107, "learning_rate": 1.9913354198905273e-05, "loss": 0.8863, "step": 1952 }, { "epoch": 0.07073524085476277, "grad_norm": 1.4664194638043562, "learning_rate": 1.9913200042614025e-05, "loss": 0.8018, "step": 1953 }, { "epoch": 0.07077145961608113, "grad_norm": 2.034550765277701, "learning_rate": 1.9913045749908436e-05, "loss": 0.9066, "step": 1954 }, { "epoch": 0.0708076783773995, "grad_norm": 1.9422088245930562, "learning_rate": 1.9912891320790628e-05, "loss": 0.9127, "step": 1955 }, { "epoch": 0.07084389713871786, "grad_norm": 1.7103702451794849, "learning_rate": 1.991273675526272e-05, "loss": 0.8907, "step": 1956 }, { "epoch": 0.07088011590003622, "grad_norm": 2.0294809891244303, "learning_rate": 1.9912582053326847e-05, "loss": 0.8486, "step": 1957 }, { "epoch": 0.07091633466135458, "grad_norm": 2.2509726369165706, "learning_rate": 1.9912427214985137e-05, "loss": 0.8496, "step": 1958 }, { "epoch": 0.07095255342267294, "grad_norm": 1.674062936641963, "learning_rate": 1.9912272240239715e-05, "loss": 0.9173, "step": 1959 }, { "epoch": 0.07098877218399131, "grad_norm": 1.4682169407269774, "learning_rate": 1.9912117129092722e-05, "loss": 0.8186, "step": 1960 }, { "epoch": 0.07102499094530967, "grad_norm": 1.3662212158754485, "learning_rate": 1.9911961881546286e-05, "loss": 0.7126, "step": 1961 }, { "epoch": 0.07106120970662803, "grad_norm": 1.786198204474494, "learning_rate": 1.991180649760254e-05, "loss": 0.8777, "step": 1962 }, { "epoch": 0.0710974284679464, "grad_norm": 1.4003097512988472, "learning_rate": 1.9911650977263634e-05, "loss": 0.7853, "step": 1963 }, { "epoch": 0.07113364722926475, "grad_norm": 2.017195611792638, "learning_rate": 1.9911495320531698e-05, "loss": 0.9028, "step": 1964 }, { "epoch": 0.07116986599058313, "grad_norm": 2.122227625626912, "learning_rate": 1.9911339527408877e-05, "loss": 0.9153, "step": 1965 }, { "epoch": 0.07120608475190149, "grad_norm": 1.8104026134098805, "learning_rate": 1.9911183597897315e-05, "loss": 0.8866, "step": 1966 }, { "epoch": 0.07124230351321985, "grad_norm": 1.7902726501212076, "learning_rate": 1.991102753199916e-05, "loss": 0.8481, "step": 1967 }, { "epoch": 0.07127852227453821, "grad_norm": 1.7932436655414765, "learning_rate": 1.991087132971656e-05, "loss": 0.8352, "step": 1968 }, { "epoch": 0.07131474103585657, "grad_norm": 1.5440491562086562, "learning_rate": 1.991071499105166e-05, "loss": 0.8239, "step": 1969 }, { "epoch": 0.07135095979717494, "grad_norm": 2.1918860002140512, "learning_rate": 1.991055851600661e-05, "loss": 0.8546, "step": 1970 }, { "epoch": 0.0713871785584933, "grad_norm": 1.5411453072388257, "learning_rate": 1.991040190458357e-05, "loss": 0.789, "step": 1971 }, { "epoch": 0.07142339731981166, "grad_norm": 1.6995286516427663, "learning_rate": 1.9910245156784688e-05, "loss": 0.8405, "step": 1972 }, { "epoch": 0.07145961608113002, "grad_norm": 1.994973928617598, "learning_rate": 1.9910088272612128e-05, "loss": 0.9802, "step": 1973 }, { "epoch": 0.07149583484244838, "grad_norm": 1.7994374559281956, "learning_rate": 1.9909931252068046e-05, "loss": 0.8206, "step": 1974 }, { "epoch": 0.07153205360376676, "grad_norm": 2.049138950156838, "learning_rate": 1.9909774095154603e-05, "loss": 0.8825, "step": 1975 }, { "epoch": 0.07156827236508512, "grad_norm": 1.913942159548161, "learning_rate": 1.990961680187396e-05, "loss": 0.9215, "step": 1976 }, { "epoch": 0.07160449112640348, "grad_norm": 1.8164335946106485, "learning_rate": 1.990945937222828e-05, "loss": 0.8513, "step": 1977 }, { "epoch": 0.07164070988772184, "grad_norm": 1.9087018051508262, "learning_rate": 1.9909301806219737e-05, "loss": 0.893, "step": 1978 }, { "epoch": 0.0716769286490402, "grad_norm": 1.8725085345046382, "learning_rate": 1.9909144103850487e-05, "loss": 0.8754, "step": 1979 }, { "epoch": 0.07171314741035857, "grad_norm": 1.7803272180600054, "learning_rate": 1.9908986265122712e-05, "loss": 0.8126, "step": 1980 }, { "epoch": 0.07174936617167693, "grad_norm": 1.4695185672556468, "learning_rate": 1.9908828290038577e-05, "loss": 0.8406, "step": 1981 }, { "epoch": 0.07178558493299529, "grad_norm": 1.833935237262813, "learning_rate": 1.9908670178600264e-05, "loss": 0.8032, "step": 1982 }, { "epoch": 0.07182180369431365, "grad_norm": 1.8882361977615834, "learning_rate": 1.9908511930809935e-05, "loss": 0.9233, "step": 1983 }, { "epoch": 0.07185802245563201, "grad_norm": 2.046697390472766, "learning_rate": 1.9908353546669777e-05, "loss": 0.9405, "step": 1984 }, { "epoch": 0.07189424121695039, "grad_norm": 1.787246404928753, "learning_rate": 1.990819502618197e-05, "loss": 0.97, "step": 1985 }, { "epoch": 0.07193045997826875, "grad_norm": 1.8206196881867278, "learning_rate": 1.9908036369348695e-05, "loss": 0.9153, "step": 1986 }, { "epoch": 0.0719666787395871, "grad_norm": 1.7667991107038428, "learning_rate": 1.9907877576172133e-05, "loss": 0.837, "step": 1987 }, { "epoch": 0.07200289750090547, "grad_norm": 1.9635538167012627, "learning_rate": 1.9907718646654464e-05, "loss": 0.8752, "step": 1988 }, { "epoch": 0.07203911626222383, "grad_norm": 2.030254763929722, "learning_rate": 1.9907559580797885e-05, "loss": 0.9486, "step": 1989 }, { "epoch": 0.0720753350235422, "grad_norm": 1.363366983632892, "learning_rate": 1.9907400378604583e-05, "loss": 0.7832, "step": 1990 }, { "epoch": 0.07211155378486056, "grad_norm": 1.81345279906053, "learning_rate": 1.9907241040076744e-05, "loss": 0.9463, "step": 1991 }, { "epoch": 0.07214777254617892, "grad_norm": 1.8239019157579084, "learning_rate": 1.9907081565216564e-05, "loss": 0.8442, "step": 1992 }, { "epoch": 0.07218399130749728, "grad_norm": 1.8507404434001682, "learning_rate": 1.9906921954026234e-05, "loss": 0.8585, "step": 1993 }, { "epoch": 0.07222021006881564, "grad_norm": 1.738429956826085, "learning_rate": 1.9906762206507954e-05, "loss": 0.858, "step": 1994 }, { "epoch": 0.07225642883013401, "grad_norm": 1.8919658201854757, "learning_rate": 1.9906602322663922e-05, "loss": 0.8564, "step": 1995 }, { "epoch": 0.07229264759145237, "grad_norm": 1.4718430338036896, "learning_rate": 1.9906442302496338e-05, "loss": 0.8265, "step": 1996 }, { "epoch": 0.07232886635277073, "grad_norm": 1.3436265858612035, "learning_rate": 1.99062821460074e-05, "loss": 0.7605, "step": 1997 }, { "epoch": 0.0723650851140891, "grad_norm": 2.0600398474882637, "learning_rate": 1.990612185319932e-05, "loss": 0.8789, "step": 1998 }, { "epoch": 0.07240130387540746, "grad_norm": 1.4994334561001976, "learning_rate": 1.9905961424074294e-05, "loss": 0.7653, "step": 1999 }, { "epoch": 0.07243752263672583, "grad_norm": 1.9499380322117894, "learning_rate": 1.990580085863454e-05, "loss": 0.8364, "step": 2000 }, { "epoch": 0.07247374139804419, "grad_norm": 1.9189637520561775, "learning_rate": 1.990564015688226e-05, "loss": 0.913, "step": 2001 }, { "epoch": 0.07250996015936255, "grad_norm": 1.7958049765408248, "learning_rate": 1.9905479318819664e-05, "loss": 0.8463, "step": 2002 }, { "epoch": 0.07254617892068091, "grad_norm": 1.9251778713557435, "learning_rate": 1.990531834444897e-05, "loss": 0.9322, "step": 2003 }, { "epoch": 0.07258239768199927, "grad_norm": 1.712988686446004, "learning_rate": 1.9905157233772396e-05, "loss": 0.8654, "step": 2004 }, { "epoch": 0.07261861644331764, "grad_norm": 1.8345716042089164, "learning_rate": 1.9904995986792153e-05, "loss": 0.8497, "step": 2005 }, { "epoch": 0.072654835204636, "grad_norm": 1.7952898393434238, "learning_rate": 1.9904834603510463e-05, "loss": 0.8813, "step": 2006 }, { "epoch": 0.07269105396595436, "grad_norm": 1.6569961158340583, "learning_rate": 1.9904673083929546e-05, "loss": 0.8999, "step": 2007 }, { "epoch": 0.07272727272727272, "grad_norm": 1.709121949244265, "learning_rate": 1.990451142805162e-05, "loss": 0.8966, "step": 2008 }, { "epoch": 0.07276349148859108, "grad_norm": 1.8743236593482664, "learning_rate": 1.990434963587892e-05, "loss": 0.8568, "step": 2009 }, { "epoch": 0.07279971024990946, "grad_norm": 1.804482656883739, "learning_rate": 1.990418770741366e-05, "loss": 0.8639, "step": 2010 }, { "epoch": 0.07283592901122782, "grad_norm": 1.74124297274625, "learning_rate": 1.990402564265808e-05, "loss": 0.861, "step": 2011 }, { "epoch": 0.07287214777254618, "grad_norm": 1.988642699058151, "learning_rate": 1.99038634416144e-05, "loss": 0.8868, "step": 2012 }, { "epoch": 0.07290836653386454, "grad_norm": 1.922632839458721, "learning_rate": 1.990370110428486e-05, "loss": 0.8698, "step": 2013 }, { "epoch": 0.0729445852951829, "grad_norm": 1.869510795347459, "learning_rate": 1.9903538630671687e-05, "loss": 0.9281, "step": 2014 }, { "epoch": 0.07298080405650127, "grad_norm": 1.962058121370446, "learning_rate": 1.9903376020777126e-05, "loss": 0.8888, "step": 2015 }, { "epoch": 0.07301702281781963, "grad_norm": 1.8637488476207915, "learning_rate": 1.9903213274603404e-05, "loss": 0.9109, "step": 2016 }, { "epoch": 0.07305324157913799, "grad_norm": 1.7308784994327049, "learning_rate": 1.990305039215277e-05, "loss": 0.901, "step": 2017 }, { "epoch": 0.07308946034045635, "grad_norm": 1.8833515416554973, "learning_rate": 1.9902887373427458e-05, "loss": 0.9552, "step": 2018 }, { "epoch": 0.07312567910177471, "grad_norm": 1.6019155514423746, "learning_rate": 1.9902724218429716e-05, "loss": 0.8439, "step": 2019 }, { "epoch": 0.07316189786309309, "grad_norm": 1.7626310595692352, "learning_rate": 1.990256092716179e-05, "loss": 0.8587, "step": 2020 }, { "epoch": 0.07319811662441145, "grad_norm": 1.2699253309683063, "learning_rate": 1.9902397499625917e-05, "loss": 0.8343, "step": 2021 }, { "epoch": 0.0732343353857298, "grad_norm": 1.9599649327593323, "learning_rate": 1.990223393582436e-05, "loss": 0.9119, "step": 2022 }, { "epoch": 0.07327055414704817, "grad_norm": 1.6841675484197804, "learning_rate": 1.9902070235759358e-05, "loss": 0.8905, "step": 2023 }, { "epoch": 0.07330677290836653, "grad_norm": 1.6904388021998007, "learning_rate": 1.9901906399433172e-05, "loss": 0.8795, "step": 2024 }, { "epoch": 0.0733429916696849, "grad_norm": 1.8454683558194185, "learning_rate": 1.9901742426848055e-05, "loss": 0.9152, "step": 2025 }, { "epoch": 0.07337921043100326, "grad_norm": 1.3086911228412477, "learning_rate": 1.9901578318006258e-05, "loss": 0.8546, "step": 2026 }, { "epoch": 0.07341542919232162, "grad_norm": 1.7491981684243592, "learning_rate": 1.9901414072910045e-05, "loss": 0.9232, "step": 2027 }, { "epoch": 0.07345164795363998, "grad_norm": 1.7405772486836484, "learning_rate": 1.9901249691561672e-05, "loss": 0.9151, "step": 2028 }, { "epoch": 0.07348786671495836, "grad_norm": 1.4947785201675603, "learning_rate": 1.9901085173963407e-05, "loss": 0.7787, "step": 2029 }, { "epoch": 0.07352408547627672, "grad_norm": 1.6436441689542338, "learning_rate": 1.9900920520117507e-05, "loss": 0.803, "step": 2030 }, { "epoch": 0.07356030423759508, "grad_norm": 1.6978188020395073, "learning_rate": 1.9900755730026243e-05, "loss": 0.8099, "step": 2031 }, { "epoch": 0.07359652299891344, "grad_norm": 2.044115747453119, "learning_rate": 1.990059080369188e-05, "loss": 0.9386, "step": 2032 }, { "epoch": 0.0736327417602318, "grad_norm": 1.8440941846979897, "learning_rate": 1.990042574111669e-05, "loss": 0.8855, "step": 2033 }, { "epoch": 0.07366896052155017, "grad_norm": 1.6356549208407478, "learning_rate": 1.990026054230294e-05, "loss": 0.8551, "step": 2034 }, { "epoch": 0.07370517928286853, "grad_norm": 1.7430966660750056, "learning_rate": 1.9900095207252905e-05, "loss": 0.8279, "step": 2035 }, { "epoch": 0.07374139804418689, "grad_norm": 1.3693306544625974, "learning_rate": 1.9899929735968864e-05, "loss": 0.7956, "step": 2036 }, { "epoch": 0.07377761680550525, "grad_norm": 1.6843494718338696, "learning_rate": 1.989976412845309e-05, "loss": 0.7689, "step": 2037 }, { "epoch": 0.07381383556682361, "grad_norm": 1.74658176037128, "learning_rate": 1.989959838470786e-05, "loss": 0.8072, "step": 2038 }, { "epoch": 0.07385005432814198, "grad_norm": 1.2862951077444271, "learning_rate": 1.989943250473546e-05, "loss": 0.8372, "step": 2039 }, { "epoch": 0.07388627308946034, "grad_norm": 1.7142293292049384, "learning_rate": 1.989926648853817e-05, "loss": 0.8605, "step": 2040 }, { "epoch": 0.0739224918507787, "grad_norm": 1.740409070054331, "learning_rate": 1.9899100336118276e-05, "loss": 0.8925, "step": 2041 }, { "epoch": 0.07395871061209706, "grad_norm": 1.7926656938545324, "learning_rate": 1.9898934047478063e-05, "loss": 0.8643, "step": 2042 }, { "epoch": 0.07399492937341542, "grad_norm": 1.777123634625627, "learning_rate": 1.9898767622619818e-05, "loss": 0.8527, "step": 2043 }, { "epoch": 0.0740311481347338, "grad_norm": 1.6848674876930503, "learning_rate": 1.9898601061545833e-05, "loss": 0.9447, "step": 2044 }, { "epoch": 0.07406736689605216, "grad_norm": 1.9466919871902644, "learning_rate": 1.98984343642584e-05, "loss": 0.9141, "step": 2045 }, { "epoch": 0.07410358565737052, "grad_norm": 1.8576951753451163, "learning_rate": 1.9898267530759815e-05, "loss": 0.9821, "step": 2046 }, { "epoch": 0.07413980441868888, "grad_norm": 1.8333914004327259, "learning_rate": 1.989810056105237e-05, "loss": 0.9008, "step": 2047 }, { "epoch": 0.07417602318000724, "grad_norm": 1.7162653650344792, "learning_rate": 1.989793345513836e-05, "loss": 0.8669, "step": 2048 }, { "epoch": 0.07421224194132561, "grad_norm": 1.4329206202867253, "learning_rate": 1.9897766213020094e-05, "loss": 0.7988, "step": 2049 }, { "epoch": 0.07424846070264397, "grad_norm": 1.9203557410000431, "learning_rate": 1.9897598834699865e-05, "loss": 0.8646, "step": 2050 }, { "epoch": 0.07428467946396233, "grad_norm": 2.1823322374689287, "learning_rate": 1.989743132017998e-05, "loss": 0.7876, "step": 2051 }, { "epoch": 0.0743208982252807, "grad_norm": 1.976493118033381, "learning_rate": 1.9897263669462745e-05, "loss": 0.9013, "step": 2052 }, { "epoch": 0.07435711698659905, "grad_norm": 1.5454203917931253, "learning_rate": 1.989709588255046e-05, "loss": 0.7961, "step": 2053 }, { "epoch": 0.07439333574791743, "grad_norm": 1.6281257927759039, "learning_rate": 1.9896927959445444e-05, "loss": 0.7986, "step": 2054 }, { "epoch": 0.07442955450923579, "grad_norm": 1.8377726513966761, "learning_rate": 1.9896759900150004e-05, "loss": 0.7694, "step": 2055 }, { "epoch": 0.07446577327055415, "grad_norm": 1.8222020721801477, "learning_rate": 1.989659170466645e-05, "loss": 0.8559, "step": 2056 }, { "epoch": 0.07450199203187251, "grad_norm": 1.3880600710773932, "learning_rate": 1.98964233729971e-05, "loss": 0.8003, "step": 2057 }, { "epoch": 0.07453821079319087, "grad_norm": 1.6896173761158686, "learning_rate": 1.9896254905144265e-05, "loss": 0.8258, "step": 2058 }, { "epoch": 0.07457442955450924, "grad_norm": 1.4271521646689287, "learning_rate": 1.9896086301110268e-05, "loss": 0.7636, "step": 2059 }, { "epoch": 0.0746106483158276, "grad_norm": 1.4514503985054914, "learning_rate": 1.989591756089743e-05, "loss": 0.8383, "step": 2060 }, { "epoch": 0.07464686707714596, "grad_norm": 1.7453387012581127, "learning_rate": 1.9895748684508068e-05, "loss": 0.927, "step": 2061 }, { "epoch": 0.07468308583846432, "grad_norm": 2.0965496319614583, "learning_rate": 1.9895579671944514e-05, "loss": 0.8699, "step": 2062 }, { "epoch": 0.07471930459978268, "grad_norm": 1.794875102839881, "learning_rate": 1.9895410523209084e-05, "loss": 0.8504, "step": 2063 }, { "epoch": 0.07475552336110106, "grad_norm": 1.760709370853673, "learning_rate": 1.989524123830411e-05, "loss": 0.8366, "step": 2064 }, { "epoch": 0.07479174212241942, "grad_norm": 1.975943379822513, "learning_rate": 1.9895071817231925e-05, "loss": 0.9412, "step": 2065 }, { "epoch": 0.07482796088373778, "grad_norm": 1.6990349878501123, "learning_rate": 1.9894902259994856e-05, "loss": 0.85, "step": 2066 }, { "epoch": 0.07486417964505614, "grad_norm": 1.6129288149890422, "learning_rate": 1.9894732566595238e-05, "loss": 0.8055, "step": 2067 }, { "epoch": 0.0749003984063745, "grad_norm": 2.1408159532355575, "learning_rate": 1.9894562737035406e-05, "loss": 0.8703, "step": 2068 }, { "epoch": 0.07493661716769287, "grad_norm": 2.0915677259434906, "learning_rate": 1.9894392771317694e-05, "loss": 0.9275, "step": 2069 }, { "epoch": 0.07497283592901123, "grad_norm": 1.3270211746044749, "learning_rate": 1.9894222669444447e-05, "loss": 0.8229, "step": 2070 }, { "epoch": 0.07500905469032959, "grad_norm": 2.0591170769563756, "learning_rate": 1.9894052431418e-05, "loss": 0.9952, "step": 2071 }, { "epoch": 0.07504527345164795, "grad_norm": 1.6612990810242263, "learning_rate": 1.9893882057240698e-05, "loss": 0.8439, "step": 2072 }, { "epoch": 0.07508149221296631, "grad_norm": 2.0097624738870574, "learning_rate": 1.9893711546914885e-05, "loss": 0.8493, "step": 2073 }, { "epoch": 0.07511771097428468, "grad_norm": 1.9045918074877768, "learning_rate": 1.989354090044291e-05, "loss": 0.875, "step": 2074 }, { "epoch": 0.07515392973560305, "grad_norm": 1.965453148184106, "learning_rate": 1.9893370117827114e-05, "loss": 0.9171, "step": 2075 }, { "epoch": 0.0751901484969214, "grad_norm": 1.7390402454861345, "learning_rate": 1.9893199199069858e-05, "loss": 0.84, "step": 2076 }, { "epoch": 0.07522636725823977, "grad_norm": 1.500109711840657, "learning_rate": 1.9893028144173482e-05, "loss": 0.7635, "step": 2077 }, { "epoch": 0.07526258601955813, "grad_norm": 1.8738169116491246, "learning_rate": 1.989285695314035e-05, "loss": 0.8666, "step": 2078 }, { "epoch": 0.0752988047808765, "grad_norm": 1.9945208404663846, "learning_rate": 1.9892685625972814e-05, "loss": 0.868, "step": 2079 }, { "epoch": 0.07533502354219486, "grad_norm": 1.793924401108952, "learning_rate": 1.989251416267323e-05, "loss": 0.8798, "step": 2080 }, { "epoch": 0.07537124230351322, "grad_norm": 1.9099963107560265, "learning_rate": 1.989234256324396e-05, "loss": 0.9087, "step": 2081 }, { "epoch": 0.07540746106483158, "grad_norm": 1.7088073602349771, "learning_rate": 1.9892170827687366e-05, "loss": 0.7965, "step": 2082 }, { "epoch": 0.07544367982614994, "grad_norm": 1.414040760726413, "learning_rate": 1.9891998956005806e-05, "loss": 0.7907, "step": 2083 }, { "epoch": 0.07547989858746831, "grad_norm": 1.7117771435574893, "learning_rate": 1.9891826948201647e-05, "loss": 0.8156, "step": 2084 }, { "epoch": 0.07551611734878667, "grad_norm": 2.0663386804389763, "learning_rate": 1.989165480427726e-05, "loss": 0.945, "step": 2085 }, { "epoch": 0.07555233611010503, "grad_norm": 1.2685439639871634, "learning_rate": 1.989148252423501e-05, "loss": 0.787, "step": 2086 }, { "epoch": 0.0755885548714234, "grad_norm": 1.3265577499822163, "learning_rate": 1.9891310108077272e-05, "loss": 0.8084, "step": 2087 }, { "epoch": 0.07562477363274175, "grad_norm": 2.2129900574124655, "learning_rate": 1.9891137555806416e-05, "loss": 0.9532, "step": 2088 }, { "epoch": 0.07566099239406013, "grad_norm": 1.4179955307051189, "learning_rate": 1.9890964867424813e-05, "loss": 0.8114, "step": 2089 }, { "epoch": 0.07569721115537849, "grad_norm": 1.9011398310769743, "learning_rate": 1.9890792042934844e-05, "loss": 0.9046, "step": 2090 }, { "epoch": 0.07573342991669685, "grad_norm": 2.191626418710214, "learning_rate": 1.989061908233888e-05, "loss": 0.8659, "step": 2091 }, { "epoch": 0.07576964867801521, "grad_norm": 2.258769509824179, "learning_rate": 1.9890445985639316e-05, "loss": 0.9243, "step": 2092 }, { "epoch": 0.07580586743933357, "grad_norm": 2.686822579222545, "learning_rate": 1.989027275283852e-05, "loss": 0.8012, "step": 2093 }, { "epoch": 0.07584208620065194, "grad_norm": 1.866224192078699, "learning_rate": 1.989009938393888e-05, "loss": 0.8847, "step": 2094 }, { "epoch": 0.0758783049619703, "grad_norm": 1.6556020498418778, "learning_rate": 1.9889925878942785e-05, "loss": 0.8185, "step": 2095 }, { "epoch": 0.07591452372328866, "grad_norm": 2.0358988584041042, "learning_rate": 1.9889752237852617e-05, "loss": 0.9087, "step": 2096 }, { "epoch": 0.07595074248460702, "grad_norm": 1.3735904101063858, "learning_rate": 1.988957846067077e-05, "loss": 0.7515, "step": 2097 }, { "epoch": 0.07598696124592538, "grad_norm": 1.8252005487716942, "learning_rate": 1.9889404547399634e-05, "loss": 0.8803, "step": 2098 }, { "epoch": 0.07602318000724376, "grad_norm": 2.010080967103873, "learning_rate": 1.9889230498041598e-05, "loss": 0.8684, "step": 2099 }, { "epoch": 0.07605939876856212, "grad_norm": 1.340360553625584, "learning_rate": 1.9889056312599066e-05, "loss": 0.7468, "step": 2100 }, { "epoch": 0.07609561752988048, "grad_norm": 2.000881682455496, "learning_rate": 1.988888199107443e-05, "loss": 0.9034, "step": 2101 }, { "epoch": 0.07613183629119884, "grad_norm": 1.8866730007208805, "learning_rate": 1.9888707533470085e-05, "loss": 0.9588, "step": 2102 }, { "epoch": 0.0761680550525172, "grad_norm": 2.247842532497813, "learning_rate": 1.9888532939788437e-05, "loss": 0.9339, "step": 2103 }, { "epoch": 0.07620427381383557, "grad_norm": 1.91460587807268, "learning_rate": 1.9888358210031885e-05, "loss": 0.9362, "step": 2104 }, { "epoch": 0.07624049257515393, "grad_norm": 1.749915761708317, "learning_rate": 1.9888183344202837e-05, "loss": 0.9026, "step": 2105 }, { "epoch": 0.07627671133647229, "grad_norm": 1.5059804259311702, "learning_rate": 1.98880083423037e-05, "loss": 0.8679, "step": 2106 }, { "epoch": 0.07631293009779065, "grad_norm": 1.8621109690178237, "learning_rate": 1.9887833204336876e-05, "loss": 0.9134, "step": 2107 }, { "epoch": 0.07634914885910901, "grad_norm": 1.708922678242496, "learning_rate": 1.9887657930304782e-05, "loss": 0.8549, "step": 2108 }, { "epoch": 0.07638536762042739, "grad_norm": 1.3391949113755997, "learning_rate": 1.9887482520209824e-05, "loss": 0.8325, "step": 2109 }, { "epoch": 0.07642158638174575, "grad_norm": 1.5212732324341047, "learning_rate": 1.988730697405442e-05, "loss": 0.8025, "step": 2110 }, { "epoch": 0.0764578051430641, "grad_norm": 1.9795085979697942, "learning_rate": 1.9887131291840983e-05, "loss": 0.8428, "step": 2111 }, { "epoch": 0.07649402390438247, "grad_norm": 2.0183548184729707, "learning_rate": 1.9886955473571935e-05, "loss": 0.8696, "step": 2112 }, { "epoch": 0.07653024266570083, "grad_norm": 1.8094664717317448, "learning_rate": 1.988677951924969e-05, "loss": 0.8067, "step": 2113 }, { "epoch": 0.0765664614270192, "grad_norm": 1.6237139323764433, "learning_rate": 1.9886603428876673e-05, "loss": 0.8132, "step": 2114 }, { "epoch": 0.07660268018833756, "grad_norm": 2.138412077465679, "learning_rate": 1.98864272024553e-05, "loss": 0.9223, "step": 2115 }, { "epoch": 0.07663889894965592, "grad_norm": 1.867369521343816, "learning_rate": 1.9886250839988008e-05, "loss": 0.8391, "step": 2116 }, { "epoch": 0.07667511771097428, "grad_norm": 1.7893266607771043, "learning_rate": 1.9886074341477218e-05, "loss": 0.8086, "step": 2117 }, { "epoch": 0.07671133647229265, "grad_norm": 1.7690165804927758, "learning_rate": 1.9885897706925353e-05, "loss": 0.9117, "step": 2118 }, { "epoch": 0.07674755523361101, "grad_norm": 1.9150652069065695, "learning_rate": 1.9885720936334855e-05, "loss": 0.8767, "step": 2119 }, { "epoch": 0.07678377399492937, "grad_norm": 1.619497271131646, "learning_rate": 1.9885544029708147e-05, "loss": 0.8055, "step": 2120 }, { "epoch": 0.07681999275624773, "grad_norm": 1.752959416384423, "learning_rate": 1.988536698704767e-05, "loss": 0.925, "step": 2121 }, { "epoch": 0.0768562115175661, "grad_norm": 1.9239488631571153, "learning_rate": 1.9885189808355855e-05, "loss": 0.9638, "step": 2122 }, { "epoch": 0.07689243027888447, "grad_norm": 1.7177805019404724, "learning_rate": 1.988501249363514e-05, "loss": 0.8865, "step": 2123 }, { "epoch": 0.07692864904020283, "grad_norm": 1.8814464140217282, "learning_rate": 1.988483504288797e-05, "loss": 0.8813, "step": 2124 }, { "epoch": 0.07696486780152119, "grad_norm": 2.123982406746071, "learning_rate": 1.9884657456116785e-05, "loss": 0.8588, "step": 2125 }, { "epoch": 0.07700108656283955, "grad_norm": 1.7805396010264638, "learning_rate": 1.9884479733324027e-05, "loss": 0.9539, "step": 2126 }, { "epoch": 0.07703730532415791, "grad_norm": 1.7759235262149025, "learning_rate": 1.988430187451214e-05, "loss": 0.875, "step": 2127 }, { "epoch": 0.07707352408547628, "grad_norm": 1.7479555937306954, "learning_rate": 1.9884123879683578e-05, "loss": 0.9011, "step": 2128 }, { "epoch": 0.07710974284679464, "grad_norm": 1.7790652709535977, "learning_rate": 1.9883945748840786e-05, "loss": 0.7302, "step": 2129 }, { "epoch": 0.077145961608113, "grad_norm": 1.9513358065422697, "learning_rate": 1.9883767481986217e-05, "loss": 0.9131, "step": 2130 }, { "epoch": 0.07718218036943136, "grad_norm": 1.8118103910547916, "learning_rate": 1.9883589079122317e-05, "loss": 0.7912, "step": 2131 }, { "epoch": 0.07721839913074972, "grad_norm": 1.804702486248463, "learning_rate": 1.9883410540251553e-05, "loss": 0.866, "step": 2132 }, { "epoch": 0.0772546178920681, "grad_norm": 1.3861132175289554, "learning_rate": 1.9883231865376372e-05, "loss": 0.7477, "step": 2133 }, { "epoch": 0.07729083665338646, "grad_norm": 2.025563286950859, "learning_rate": 1.9883053054499236e-05, "loss": 0.8705, "step": 2134 }, { "epoch": 0.07732705541470482, "grad_norm": 2.3752765055366907, "learning_rate": 1.9882874107622608e-05, "loss": 0.9759, "step": 2135 }, { "epoch": 0.07736327417602318, "grad_norm": 1.249715796214323, "learning_rate": 1.988269502474895e-05, "loss": 0.8097, "step": 2136 }, { "epoch": 0.07739949293734154, "grad_norm": 1.250952725161017, "learning_rate": 1.988251580588072e-05, "loss": 0.7826, "step": 2137 }, { "epoch": 0.07743571169865991, "grad_norm": 2.1074935911489536, "learning_rate": 1.9882336451020392e-05, "loss": 0.9327, "step": 2138 }, { "epoch": 0.07747193045997827, "grad_norm": 1.677200011351866, "learning_rate": 1.988215696017043e-05, "loss": 0.9607, "step": 2139 }, { "epoch": 0.07750814922129663, "grad_norm": 1.9629584305472334, "learning_rate": 1.9881977333333307e-05, "loss": 0.9805, "step": 2140 }, { "epoch": 0.07754436798261499, "grad_norm": 1.872627028267912, "learning_rate": 1.988179757051149e-05, "loss": 0.787, "step": 2141 }, { "epoch": 0.07758058674393335, "grad_norm": 1.5912502678444285, "learning_rate": 1.9881617671707457e-05, "loss": 0.7751, "step": 2142 }, { "epoch": 0.07761680550525173, "grad_norm": 1.432086137967996, "learning_rate": 1.988143763692368e-05, "loss": 0.8541, "step": 2143 }, { "epoch": 0.07765302426657009, "grad_norm": 1.8838772434223539, "learning_rate": 1.9881257466162642e-05, "loss": 0.8938, "step": 2144 }, { "epoch": 0.07768924302788845, "grad_norm": 1.8376227111616619, "learning_rate": 1.988107715942682e-05, "loss": 0.8918, "step": 2145 }, { "epoch": 0.0777254617892068, "grad_norm": 1.8888467492741388, "learning_rate": 1.988089671671869e-05, "loss": 0.8391, "step": 2146 }, { "epoch": 0.07776168055052517, "grad_norm": 1.709348908919155, "learning_rate": 1.988071613804074e-05, "loss": 0.8037, "step": 2147 }, { "epoch": 0.07779789931184354, "grad_norm": 1.417628515506964, "learning_rate": 1.9880535423395456e-05, "loss": 0.7601, "step": 2148 }, { "epoch": 0.0778341180731619, "grad_norm": 1.8453144148575218, "learning_rate": 1.988035457278532e-05, "loss": 0.8945, "step": 2149 }, { "epoch": 0.07787033683448026, "grad_norm": 1.9598484928889086, "learning_rate": 1.9880173586212827e-05, "loss": 0.8806, "step": 2150 }, { "epoch": 0.07790655559579862, "grad_norm": 1.8434390810083314, "learning_rate": 1.987999246368046e-05, "loss": 0.9394, "step": 2151 }, { "epoch": 0.07794277435711698, "grad_norm": 2.1214353749954302, "learning_rate": 1.987981120519072e-05, "loss": 0.8875, "step": 2152 }, { "epoch": 0.07797899311843536, "grad_norm": 1.35795714130066, "learning_rate": 1.987962981074609e-05, "loss": 0.8472, "step": 2153 }, { "epoch": 0.07801521187975372, "grad_norm": 1.5973450490525902, "learning_rate": 1.9879448280349077e-05, "loss": 0.9132, "step": 2154 }, { "epoch": 0.07805143064107208, "grad_norm": 1.6171770969404102, "learning_rate": 1.9879266614002174e-05, "loss": 0.7881, "step": 2155 }, { "epoch": 0.07808764940239044, "grad_norm": 1.651873057083188, "learning_rate": 1.9879084811707882e-05, "loss": 0.7956, "step": 2156 }, { "epoch": 0.0781238681637088, "grad_norm": 1.7276594293645662, "learning_rate": 1.9878902873468702e-05, "loss": 0.8257, "step": 2157 }, { "epoch": 0.07816008692502717, "grad_norm": 1.2379169778194579, "learning_rate": 1.9878720799287138e-05, "loss": 0.7615, "step": 2158 }, { "epoch": 0.07819630568634553, "grad_norm": 1.792439789441629, "learning_rate": 1.9878538589165696e-05, "loss": 0.923, "step": 2159 }, { "epoch": 0.07823252444766389, "grad_norm": 1.6651459371548636, "learning_rate": 1.9878356243106884e-05, "loss": 0.9377, "step": 2160 }, { "epoch": 0.07826874320898225, "grad_norm": 1.3899518185816984, "learning_rate": 1.987817376111321e-05, "loss": 0.8208, "step": 2161 }, { "epoch": 0.07830496197030061, "grad_norm": 1.4151641167331426, "learning_rate": 1.9877991143187186e-05, "loss": 0.8015, "step": 2162 }, { "epoch": 0.07834118073161898, "grad_norm": 2.220114508317056, "learning_rate": 1.987780838933132e-05, "loss": 0.8959, "step": 2163 }, { "epoch": 0.07837739949293734, "grad_norm": 1.9042630236624658, "learning_rate": 1.9877625499548136e-05, "loss": 0.9459, "step": 2164 }, { "epoch": 0.0784136182542557, "grad_norm": 1.7279997988076834, "learning_rate": 1.9877442473840143e-05, "loss": 0.8221, "step": 2165 }, { "epoch": 0.07844983701557406, "grad_norm": 1.6087150586349188, "learning_rate": 1.9877259312209864e-05, "loss": 0.7905, "step": 2166 }, { "epoch": 0.07848605577689242, "grad_norm": 1.6645297113205746, "learning_rate": 1.9877076014659817e-05, "loss": 0.783, "step": 2167 }, { "epoch": 0.0785222745382108, "grad_norm": 1.5271320967721487, "learning_rate": 1.9876892581192525e-05, "loss": 0.7974, "step": 2168 }, { "epoch": 0.07855849329952916, "grad_norm": 1.8517810766961391, "learning_rate": 1.9876709011810513e-05, "loss": 0.9128, "step": 2169 }, { "epoch": 0.07859471206084752, "grad_norm": 1.9685985813494875, "learning_rate": 1.9876525306516308e-05, "loss": 0.8866, "step": 2170 }, { "epoch": 0.07863093082216588, "grad_norm": 1.5226957804983325, "learning_rate": 1.9876341465312437e-05, "loss": 0.8555, "step": 2171 }, { "epoch": 0.07866714958348424, "grad_norm": 1.8723577717470354, "learning_rate": 1.9876157488201426e-05, "loss": 0.9396, "step": 2172 }, { "epoch": 0.07870336834480261, "grad_norm": 1.7956398056347995, "learning_rate": 1.987597337518581e-05, "loss": 0.9301, "step": 2173 }, { "epoch": 0.07873958710612097, "grad_norm": 1.9614431576556213, "learning_rate": 1.9875789126268127e-05, "loss": 0.8777, "step": 2174 }, { "epoch": 0.07877580586743933, "grad_norm": 1.8781510121992615, "learning_rate": 1.98756047414509e-05, "loss": 0.9348, "step": 2175 }, { "epoch": 0.0788120246287577, "grad_norm": 1.7343654557128718, "learning_rate": 1.9875420220736682e-05, "loss": 0.9448, "step": 2176 }, { "epoch": 0.07884824339007605, "grad_norm": 1.6414480960237992, "learning_rate": 1.9875235564128003e-05, "loss": 0.8695, "step": 2177 }, { "epoch": 0.07888446215139443, "grad_norm": 1.7532159130642664, "learning_rate": 1.9875050771627403e-05, "loss": 0.9483, "step": 2178 }, { "epoch": 0.07892068091271279, "grad_norm": 1.8799646929459006, "learning_rate": 1.987486584323743e-05, "loss": 0.8765, "step": 2179 }, { "epoch": 0.07895689967403115, "grad_norm": 2.2511615453744893, "learning_rate": 1.9874680778960626e-05, "loss": 0.8862, "step": 2180 }, { "epoch": 0.07899311843534951, "grad_norm": 2.00098115167422, "learning_rate": 1.9874495578799538e-05, "loss": 0.9021, "step": 2181 }, { "epoch": 0.07902933719666787, "grad_norm": 1.847340990393732, "learning_rate": 1.987431024275671e-05, "loss": 0.9266, "step": 2182 }, { "epoch": 0.07906555595798624, "grad_norm": 1.9230109020896358, "learning_rate": 1.9874124770834703e-05, "loss": 0.8458, "step": 2183 }, { "epoch": 0.0791017747193046, "grad_norm": 1.7501170170353106, "learning_rate": 1.9873939163036058e-05, "loss": 0.9264, "step": 2184 }, { "epoch": 0.07913799348062296, "grad_norm": 1.885902315015407, "learning_rate": 1.9873753419363336e-05, "loss": 0.9454, "step": 2185 }, { "epoch": 0.07917421224194132, "grad_norm": 1.7332564260847034, "learning_rate": 1.9873567539819093e-05, "loss": 0.9315, "step": 2186 }, { "epoch": 0.07921043100325968, "grad_norm": 1.8003490233379107, "learning_rate": 1.987338152440588e-05, "loss": 0.8764, "step": 2187 }, { "epoch": 0.07924664976457806, "grad_norm": 1.5375996199274593, "learning_rate": 1.9873195373126264e-05, "loss": 0.7929, "step": 2188 }, { "epoch": 0.07928286852589642, "grad_norm": 1.8785048559653177, "learning_rate": 1.9873009085982805e-05, "loss": 0.7929, "step": 2189 }, { "epoch": 0.07931908728721478, "grad_norm": 1.9033035700351826, "learning_rate": 1.9872822662978066e-05, "loss": 0.9548, "step": 2190 }, { "epoch": 0.07935530604853314, "grad_norm": 1.7300216407194111, "learning_rate": 1.9872636104114612e-05, "loss": 0.796, "step": 2191 }, { "epoch": 0.0793915248098515, "grad_norm": 1.602344780697949, "learning_rate": 1.9872449409395012e-05, "loss": 0.7966, "step": 2192 }, { "epoch": 0.07942774357116987, "grad_norm": 1.92428986483402, "learning_rate": 1.987226257882183e-05, "loss": 0.923, "step": 2193 }, { "epoch": 0.07946396233248823, "grad_norm": 2.206201664122271, "learning_rate": 1.9872075612397642e-05, "loss": 0.871, "step": 2194 }, { "epoch": 0.07950018109380659, "grad_norm": 1.7455139122807606, "learning_rate": 1.9871888510125022e-05, "loss": 0.7907, "step": 2195 }, { "epoch": 0.07953639985512495, "grad_norm": 1.446671353595925, "learning_rate": 1.987170127200654e-05, "loss": 0.8331, "step": 2196 }, { "epoch": 0.07957261861644331, "grad_norm": 1.976411561140768, "learning_rate": 1.987151389804477e-05, "loss": 0.8612, "step": 2197 }, { "epoch": 0.07960883737776168, "grad_norm": 1.7738556699707984, "learning_rate": 1.98713263882423e-05, "loss": 0.9226, "step": 2198 }, { "epoch": 0.07964505613908004, "grad_norm": 2.1025897319834623, "learning_rate": 1.9871138742601707e-05, "loss": 0.8908, "step": 2199 }, { "epoch": 0.0796812749003984, "grad_norm": 1.5969141843359076, "learning_rate": 1.9870950961125564e-05, "loss": 0.7474, "step": 2200 }, { "epoch": 0.07971749366171677, "grad_norm": 1.9995324642609646, "learning_rate": 1.9870763043816467e-05, "loss": 0.9149, "step": 2201 }, { "epoch": 0.07975371242303514, "grad_norm": 1.8303680424374025, "learning_rate": 1.9870574990676997e-05, "loss": 0.9277, "step": 2202 }, { "epoch": 0.0797899311843535, "grad_norm": 1.773316774166236, "learning_rate": 1.9870386801709746e-05, "loss": 0.861, "step": 2203 }, { "epoch": 0.07982614994567186, "grad_norm": 2.4769179814084534, "learning_rate": 1.9870198476917296e-05, "loss": 0.8055, "step": 2204 }, { "epoch": 0.07986236870699022, "grad_norm": 1.8307993214872644, "learning_rate": 1.987001001630224e-05, "loss": 0.8574, "step": 2205 }, { "epoch": 0.07989858746830858, "grad_norm": 1.6705569392624842, "learning_rate": 1.9869821419867176e-05, "loss": 0.9172, "step": 2206 }, { "epoch": 0.07993480622962695, "grad_norm": 1.7892273041974813, "learning_rate": 1.9869632687614693e-05, "loss": 0.8365, "step": 2207 }, { "epoch": 0.07997102499094531, "grad_norm": 1.9763116771639224, "learning_rate": 1.98694438195474e-05, "loss": 0.8892, "step": 2208 }, { "epoch": 0.08000724375226367, "grad_norm": 1.9420823372583453, "learning_rate": 1.9869254815667884e-05, "loss": 0.9001, "step": 2209 }, { "epoch": 0.08004346251358203, "grad_norm": 1.7995775869140047, "learning_rate": 1.9869065675978747e-05, "loss": 0.8927, "step": 2210 }, { "epoch": 0.0800796812749004, "grad_norm": 2.032932173661069, "learning_rate": 1.9868876400482598e-05, "loss": 0.946, "step": 2211 }, { "epoch": 0.08011590003621877, "grad_norm": 1.9323526503064226, "learning_rate": 1.9868686989182037e-05, "loss": 0.9332, "step": 2212 }, { "epoch": 0.08015211879753713, "grad_norm": 1.89002038414877, "learning_rate": 1.986849744207967e-05, "loss": 0.9711, "step": 2213 }, { "epoch": 0.08018833755885549, "grad_norm": 1.931945058401707, "learning_rate": 1.986830775917811e-05, "loss": 0.8712, "step": 2214 }, { "epoch": 0.08022455632017385, "grad_norm": 1.8617903444720596, "learning_rate": 1.9868117940479963e-05, "loss": 0.8665, "step": 2215 }, { "epoch": 0.08026077508149221, "grad_norm": 1.673904367897686, "learning_rate": 1.986792798598784e-05, "loss": 0.8234, "step": 2216 }, { "epoch": 0.08029699384281058, "grad_norm": 1.7131041868866792, "learning_rate": 1.986773789570436e-05, "loss": 0.8566, "step": 2217 }, { "epoch": 0.08033321260412894, "grad_norm": 1.7959896601730292, "learning_rate": 1.9867547669632133e-05, "loss": 0.8005, "step": 2218 }, { "epoch": 0.0803694313654473, "grad_norm": 1.7797519644113395, "learning_rate": 1.9867357307773783e-05, "loss": 0.947, "step": 2219 }, { "epoch": 0.08040565012676566, "grad_norm": 1.8898562038653113, "learning_rate": 1.9867166810131925e-05, "loss": 0.8816, "step": 2220 }, { "epoch": 0.08044186888808402, "grad_norm": 1.704484898645704, "learning_rate": 1.9866976176709185e-05, "loss": 0.8896, "step": 2221 }, { "epoch": 0.0804780876494024, "grad_norm": 1.6573913185481979, "learning_rate": 1.9866785407508178e-05, "loss": 0.8422, "step": 2222 }, { "epoch": 0.08051430641072076, "grad_norm": 1.6959424031804917, "learning_rate": 1.9866594502531536e-05, "loss": 0.8269, "step": 2223 }, { "epoch": 0.08055052517203912, "grad_norm": 1.6367489793758114, "learning_rate": 1.9866403461781883e-05, "loss": 0.8108, "step": 2224 }, { "epoch": 0.08058674393335748, "grad_norm": 1.7330456506335903, "learning_rate": 1.986621228526185e-05, "loss": 0.8862, "step": 2225 }, { "epoch": 0.08062296269467584, "grad_norm": 1.6677893591278394, "learning_rate": 1.986602097297407e-05, "loss": 0.8679, "step": 2226 }, { "epoch": 0.08065918145599421, "grad_norm": 1.6657516003634758, "learning_rate": 1.986582952492117e-05, "loss": 0.8134, "step": 2227 }, { "epoch": 0.08069540021731257, "grad_norm": 1.9299435078921388, "learning_rate": 1.986563794110578e-05, "loss": 0.8716, "step": 2228 }, { "epoch": 0.08073161897863093, "grad_norm": 2.0896697212559374, "learning_rate": 1.9865446221530555e-05, "loss": 0.8948, "step": 2229 }, { "epoch": 0.08076783773994929, "grad_norm": 1.8484549993571826, "learning_rate": 1.9865254366198112e-05, "loss": 0.8974, "step": 2230 }, { "epoch": 0.08080405650126765, "grad_norm": 1.7235602625634425, "learning_rate": 1.9865062375111105e-05, "loss": 0.8753, "step": 2231 }, { "epoch": 0.08084027526258603, "grad_norm": 1.665550046132505, "learning_rate": 1.986487024827217e-05, "loss": 0.889, "step": 2232 }, { "epoch": 0.08087649402390439, "grad_norm": 1.6280951681854507, "learning_rate": 1.9864677985683954e-05, "loss": 0.9043, "step": 2233 }, { "epoch": 0.08091271278522275, "grad_norm": 1.8503937070344556, "learning_rate": 1.9864485587349097e-05, "loss": 0.9224, "step": 2234 }, { "epoch": 0.0809489315465411, "grad_norm": 1.751230362956316, "learning_rate": 1.9864293053270258e-05, "loss": 0.8964, "step": 2235 }, { "epoch": 0.08098515030785947, "grad_norm": 1.758189635615967, "learning_rate": 1.9864100383450072e-05, "loss": 0.8988, "step": 2236 }, { "epoch": 0.08102136906917784, "grad_norm": 1.590238847189799, "learning_rate": 1.98639075778912e-05, "loss": 0.8603, "step": 2237 }, { "epoch": 0.0810575878304962, "grad_norm": 1.8436736054676843, "learning_rate": 1.986371463659629e-05, "loss": 0.8318, "step": 2238 }, { "epoch": 0.08109380659181456, "grad_norm": 1.7939228452044975, "learning_rate": 1.9863521559568008e-05, "loss": 0.8957, "step": 2239 }, { "epoch": 0.08113002535313292, "grad_norm": 1.5978461813710771, "learning_rate": 1.9863328346808994e-05, "loss": 0.8188, "step": 2240 }, { "epoch": 0.08116624411445128, "grad_norm": 1.4060603980886204, "learning_rate": 1.9863134998321918e-05, "loss": 0.7473, "step": 2241 }, { "epoch": 0.08120246287576965, "grad_norm": 1.4693564004283217, "learning_rate": 1.9862941514109437e-05, "loss": 0.7537, "step": 2242 }, { "epoch": 0.08123868163708801, "grad_norm": 1.3636486955657088, "learning_rate": 1.9862747894174218e-05, "loss": 0.7521, "step": 2243 }, { "epoch": 0.08127490039840637, "grad_norm": 1.9941908244391233, "learning_rate": 1.9862554138518916e-05, "loss": 0.9166, "step": 2244 }, { "epoch": 0.08131111915972473, "grad_norm": 1.7836950319317533, "learning_rate": 1.9862360247146205e-05, "loss": 0.8336, "step": 2245 }, { "epoch": 0.0813473379210431, "grad_norm": 1.695655328832298, "learning_rate": 1.9862166220058754e-05, "loss": 0.8618, "step": 2246 }, { "epoch": 0.08138355668236147, "grad_norm": 1.642422964158945, "learning_rate": 1.9861972057259228e-05, "loss": 0.8297, "step": 2247 }, { "epoch": 0.08141977544367983, "grad_norm": 1.7886914551449038, "learning_rate": 1.9861777758750302e-05, "loss": 0.8882, "step": 2248 }, { "epoch": 0.08145599420499819, "grad_norm": 1.706762424160071, "learning_rate": 1.986158332453465e-05, "loss": 0.8497, "step": 2249 }, { "epoch": 0.08149221296631655, "grad_norm": 2.2670117655842317, "learning_rate": 1.9861388754614944e-05, "loss": 0.9195, "step": 2250 }, { "epoch": 0.08152843172763491, "grad_norm": 1.8577359191761058, "learning_rate": 1.9861194048993865e-05, "loss": 0.85, "step": 2251 }, { "epoch": 0.08156465048895328, "grad_norm": 1.824465886025131, "learning_rate": 1.986099920767409e-05, "loss": 0.9276, "step": 2252 }, { "epoch": 0.08160086925027164, "grad_norm": 2.7538295437566385, "learning_rate": 1.98608042306583e-05, "loss": 0.9108, "step": 2253 }, { "epoch": 0.08163708801159, "grad_norm": 1.7435495073325498, "learning_rate": 1.986060911794918e-05, "loss": 0.8748, "step": 2254 }, { "epoch": 0.08167330677290836, "grad_norm": 1.8965257704957874, "learning_rate": 1.9860413869549417e-05, "loss": 0.8445, "step": 2255 }, { "epoch": 0.08170952553422672, "grad_norm": 1.7327107389735021, "learning_rate": 1.9860218485461693e-05, "loss": 0.9478, "step": 2256 }, { "epoch": 0.0817457442955451, "grad_norm": 1.5969465726799454, "learning_rate": 1.98600229656887e-05, "loss": 0.8887, "step": 2257 }, { "epoch": 0.08178196305686346, "grad_norm": 1.8502718253771628, "learning_rate": 1.9859827310233125e-05, "loss": 0.7949, "step": 2258 }, { "epoch": 0.08181818181818182, "grad_norm": 2.0772953851298186, "learning_rate": 1.9859631519097664e-05, "loss": 0.8506, "step": 2259 }, { "epoch": 0.08185440057950018, "grad_norm": 1.9091104011939075, "learning_rate": 1.9859435592285007e-05, "loss": 0.8795, "step": 2260 }, { "epoch": 0.08189061934081854, "grad_norm": 1.6690965776994653, "learning_rate": 1.985923952979786e-05, "loss": 0.9055, "step": 2261 }, { "epoch": 0.08192683810213691, "grad_norm": 2.7031377396014618, "learning_rate": 1.9859043331638906e-05, "loss": 0.7574, "step": 2262 }, { "epoch": 0.08196305686345527, "grad_norm": 1.8191635922486642, "learning_rate": 1.9858846997810856e-05, "loss": 0.8115, "step": 2263 }, { "epoch": 0.08199927562477363, "grad_norm": 2.194167055524648, "learning_rate": 1.985865052831641e-05, "loss": 0.9846, "step": 2264 }, { "epoch": 0.08203549438609199, "grad_norm": 1.738732206591563, "learning_rate": 1.985845392315827e-05, "loss": 0.9409, "step": 2265 }, { "epoch": 0.08207171314741035, "grad_norm": 1.726128007841145, "learning_rate": 1.985825718233914e-05, "loss": 0.8782, "step": 2266 }, { "epoch": 0.08210793190872873, "grad_norm": 2.11341782726138, "learning_rate": 1.985806030586173e-05, "loss": 0.8938, "step": 2267 }, { "epoch": 0.08214415067004709, "grad_norm": 2.8149534708245922, "learning_rate": 1.985786329372875e-05, "loss": 0.7358, "step": 2268 }, { "epoch": 0.08218036943136545, "grad_norm": 1.6980860113509086, "learning_rate": 1.9857666145942908e-05, "loss": 0.8413, "step": 2269 }, { "epoch": 0.0822165881926838, "grad_norm": 1.6918332604877693, "learning_rate": 1.9857468862506914e-05, "loss": 0.9019, "step": 2270 }, { "epoch": 0.08225280695400217, "grad_norm": 1.7790081750957054, "learning_rate": 1.985727144342349e-05, "loss": 0.8195, "step": 2271 }, { "epoch": 0.08228902571532054, "grad_norm": 2.2576437263371103, "learning_rate": 1.985707388869535e-05, "loss": 0.8492, "step": 2272 }, { "epoch": 0.0823252444766389, "grad_norm": 1.7116997753886682, "learning_rate": 1.9856876198325213e-05, "loss": 0.8439, "step": 2273 }, { "epoch": 0.08236146323795726, "grad_norm": 1.8691447785111124, "learning_rate": 1.9856678372315797e-05, "loss": 0.8494, "step": 2274 }, { "epoch": 0.08239768199927562, "grad_norm": 1.7971119275496998, "learning_rate": 1.9856480410669827e-05, "loss": 0.8073, "step": 2275 }, { "epoch": 0.08243390076059398, "grad_norm": 1.2894225166892128, "learning_rate": 1.9856282313390026e-05, "loss": 0.8273, "step": 2276 }, { "epoch": 0.08247011952191236, "grad_norm": 1.7690049793061198, "learning_rate": 1.9856084080479115e-05, "loss": 0.8727, "step": 2277 }, { "epoch": 0.08250633828323072, "grad_norm": 2.1238227626668995, "learning_rate": 1.985588571193983e-05, "loss": 0.9305, "step": 2278 }, { "epoch": 0.08254255704454908, "grad_norm": 1.4823543238860821, "learning_rate": 1.98556872077749e-05, "loss": 0.7813, "step": 2279 }, { "epoch": 0.08257877580586744, "grad_norm": 1.4162716770509647, "learning_rate": 1.985548856798705e-05, "loss": 0.8117, "step": 2280 }, { "epoch": 0.0826149945671858, "grad_norm": 1.8357296683993298, "learning_rate": 1.985528979257902e-05, "loss": 0.9427, "step": 2281 }, { "epoch": 0.08265121332850417, "grad_norm": 2.1640219460829293, "learning_rate": 1.9855090881553546e-05, "loss": 0.895, "step": 2282 }, { "epoch": 0.08268743208982253, "grad_norm": 2.0178246564441324, "learning_rate": 1.9854891834913357e-05, "loss": 0.8639, "step": 2283 }, { "epoch": 0.08272365085114089, "grad_norm": 1.618898220720023, "learning_rate": 1.98546926526612e-05, "loss": 0.8168, "step": 2284 }, { "epoch": 0.08275986961245925, "grad_norm": 1.3699422552655678, "learning_rate": 1.985449333479981e-05, "loss": 0.8174, "step": 2285 }, { "epoch": 0.08279608837377761, "grad_norm": 1.9029990317007814, "learning_rate": 1.9854293881331935e-05, "loss": 0.9248, "step": 2286 }, { "epoch": 0.08283230713509598, "grad_norm": 1.2457431606264582, "learning_rate": 1.9854094292260316e-05, "loss": 0.8385, "step": 2287 }, { "epoch": 0.08286852589641434, "grad_norm": 1.7782973278053893, "learning_rate": 1.98538945675877e-05, "loss": 0.9356, "step": 2288 }, { "epoch": 0.0829047446577327, "grad_norm": 1.8800457990266055, "learning_rate": 1.985369470731684e-05, "loss": 0.9775, "step": 2289 }, { "epoch": 0.08294096341905106, "grad_norm": 1.7958378929399668, "learning_rate": 1.985349471145048e-05, "loss": 0.9517, "step": 2290 }, { "epoch": 0.08297718218036944, "grad_norm": 1.5727638876739474, "learning_rate": 1.9853294579991376e-05, "loss": 0.8855, "step": 2291 }, { "epoch": 0.0830134009416878, "grad_norm": 1.8626917860316337, "learning_rate": 1.985309431294228e-05, "loss": 0.8131, "step": 2292 }, { "epoch": 0.08304961970300616, "grad_norm": 1.8533571203650896, "learning_rate": 1.985289391030595e-05, "loss": 0.9162, "step": 2293 }, { "epoch": 0.08308583846432452, "grad_norm": 1.9738444083225557, "learning_rate": 1.985269337208514e-05, "loss": 0.8385, "step": 2294 }, { "epoch": 0.08312205722564288, "grad_norm": 1.736615174740935, "learning_rate": 1.9852492698282613e-05, "loss": 0.8148, "step": 2295 }, { "epoch": 0.08315827598696125, "grad_norm": 1.3456412300788492, "learning_rate": 1.985229188890113e-05, "loss": 0.7299, "step": 2296 }, { "epoch": 0.08319449474827961, "grad_norm": 1.848039030946587, "learning_rate": 1.9852090943943452e-05, "loss": 0.8498, "step": 2297 }, { "epoch": 0.08323071350959797, "grad_norm": 1.8380943797530416, "learning_rate": 1.9851889863412347e-05, "loss": 0.8888, "step": 2298 }, { "epoch": 0.08326693227091633, "grad_norm": 1.924070567621712, "learning_rate": 1.985168864731058e-05, "loss": 0.9062, "step": 2299 }, { "epoch": 0.08330315103223469, "grad_norm": 2.0346052243889416, "learning_rate": 1.985148729564092e-05, "loss": 0.8941, "step": 2300 }, { "epoch": 0.08333936979355307, "grad_norm": 1.9585387672981356, "learning_rate": 1.985128580840614e-05, "loss": 0.8754, "step": 2301 }, { "epoch": 0.08337558855487143, "grad_norm": 1.821937175212444, "learning_rate": 1.9851084185609012e-05, "loss": 0.8846, "step": 2302 }, { "epoch": 0.08341180731618979, "grad_norm": 1.886907627294931, "learning_rate": 1.985088242725231e-05, "loss": 0.8884, "step": 2303 }, { "epoch": 0.08344802607750815, "grad_norm": 1.9139060832358412, "learning_rate": 1.985068053333881e-05, "loss": 0.8339, "step": 2304 }, { "epoch": 0.08348424483882651, "grad_norm": 1.9090767639983555, "learning_rate": 1.9850478503871286e-05, "loss": 0.8958, "step": 2305 }, { "epoch": 0.08352046360014488, "grad_norm": 1.7036660312785545, "learning_rate": 1.9850276338852525e-05, "loss": 0.8036, "step": 2306 }, { "epoch": 0.08355668236146324, "grad_norm": 1.6725674035459552, "learning_rate": 1.9850074038285307e-05, "loss": 0.8091, "step": 2307 }, { "epoch": 0.0835929011227816, "grad_norm": 1.714294431420675, "learning_rate": 1.9849871602172417e-05, "loss": 0.8825, "step": 2308 }, { "epoch": 0.08362911988409996, "grad_norm": 1.8474222623803938, "learning_rate": 1.9849669030516635e-05, "loss": 1.0087, "step": 2309 }, { "epoch": 0.08366533864541832, "grad_norm": 1.3189126591956177, "learning_rate": 1.9849466323320757e-05, "loss": 0.8492, "step": 2310 }, { "epoch": 0.0837015574067367, "grad_norm": 1.9446882487238453, "learning_rate": 1.9849263480587563e-05, "loss": 0.9281, "step": 2311 }, { "epoch": 0.08373777616805506, "grad_norm": 1.8869750399697525, "learning_rate": 1.9849060502319852e-05, "loss": 0.9237, "step": 2312 }, { "epoch": 0.08377399492937342, "grad_norm": 1.409194547028952, "learning_rate": 1.9848857388520414e-05, "loss": 0.8377, "step": 2313 }, { "epoch": 0.08381021369069178, "grad_norm": 1.343427560484002, "learning_rate": 1.9848654139192046e-05, "loss": 0.7918, "step": 2314 }, { "epoch": 0.08384643245201014, "grad_norm": 1.746603212982641, "learning_rate": 1.984845075433754e-05, "loss": 0.8576, "step": 2315 }, { "epoch": 0.08388265121332851, "grad_norm": 1.6986242967273482, "learning_rate": 1.98482472339597e-05, "loss": 0.7267, "step": 2316 }, { "epoch": 0.08391886997464687, "grad_norm": 2.009420306842935, "learning_rate": 1.9848043578061326e-05, "loss": 0.8561, "step": 2317 }, { "epoch": 0.08395508873596523, "grad_norm": 1.3870860319428027, "learning_rate": 1.9847839786645214e-05, "loss": 0.7954, "step": 2318 }, { "epoch": 0.08399130749728359, "grad_norm": 1.796749840399471, "learning_rate": 1.9847635859714178e-05, "loss": 0.8092, "step": 2319 }, { "epoch": 0.08402752625860195, "grad_norm": 1.6454685251769925, "learning_rate": 1.9847431797271016e-05, "loss": 0.93, "step": 2320 }, { "epoch": 0.08406374501992032, "grad_norm": 1.836600978155173, "learning_rate": 1.9847227599318544e-05, "loss": 0.9425, "step": 2321 }, { "epoch": 0.08409996378123868, "grad_norm": 1.4256196580398555, "learning_rate": 1.9847023265859567e-05, "loss": 0.7538, "step": 2322 }, { "epoch": 0.08413618254255704, "grad_norm": 1.5553422099219636, "learning_rate": 1.9846818796896895e-05, "loss": 0.7974, "step": 2323 }, { "epoch": 0.0841724013038754, "grad_norm": 1.8080242096486139, "learning_rate": 1.9846614192433345e-05, "loss": 0.8877, "step": 2324 }, { "epoch": 0.08420862006519376, "grad_norm": 1.9268299752443092, "learning_rate": 1.9846409452471734e-05, "loss": 0.9166, "step": 2325 }, { "epoch": 0.08424483882651214, "grad_norm": 1.4363334347444179, "learning_rate": 1.9846204577014874e-05, "loss": 0.7703, "step": 2326 }, { "epoch": 0.0842810575878305, "grad_norm": 1.6585522588534498, "learning_rate": 1.9845999566065586e-05, "loss": 0.8734, "step": 2327 }, { "epoch": 0.08431727634914886, "grad_norm": 1.7184624041181586, "learning_rate": 1.9845794419626698e-05, "loss": 0.8464, "step": 2328 }, { "epoch": 0.08435349511046722, "grad_norm": 1.7816805904684132, "learning_rate": 1.9845589137701023e-05, "loss": 0.9147, "step": 2329 }, { "epoch": 0.08438971387178558, "grad_norm": 1.8147060969381736, "learning_rate": 1.9845383720291392e-05, "loss": 0.8189, "step": 2330 }, { "epoch": 0.08442593263310395, "grad_norm": 1.402674999304301, "learning_rate": 1.9845178167400633e-05, "loss": 0.7786, "step": 2331 }, { "epoch": 0.08446215139442231, "grad_norm": 1.936577009041012, "learning_rate": 1.984497247903157e-05, "loss": 0.8706, "step": 2332 }, { "epoch": 0.08449837015574067, "grad_norm": 1.6725938468207384, "learning_rate": 1.9844766655187032e-05, "loss": 0.9232, "step": 2333 }, { "epoch": 0.08453458891705903, "grad_norm": 1.7667923429432275, "learning_rate": 1.9844560695869855e-05, "loss": 0.8108, "step": 2334 }, { "epoch": 0.0845708076783774, "grad_norm": 1.5943276571600633, "learning_rate": 1.9844354601082877e-05, "loss": 0.8906, "step": 2335 }, { "epoch": 0.08460702643969577, "grad_norm": 1.4488428851368524, "learning_rate": 1.9844148370828927e-05, "loss": 0.7692, "step": 2336 }, { "epoch": 0.08464324520101413, "grad_norm": 1.7194951187091818, "learning_rate": 1.9843942005110844e-05, "loss": 0.9202, "step": 2337 }, { "epoch": 0.08467946396233249, "grad_norm": 1.7086147235547648, "learning_rate": 1.9843735503931468e-05, "loss": 0.8681, "step": 2338 }, { "epoch": 0.08471568272365085, "grad_norm": 1.692673713126779, "learning_rate": 1.9843528867293647e-05, "loss": 0.7633, "step": 2339 }, { "epoch": 0.08475190148496921, "grad_norm": 1.942516012596781, "learning_rate": 1.984332209520022e-05, "loss": 0.8799, "step": 2340 }, { "epoch": 0.08478812024628758, "grad_norm": 2.0282424311225116, "learning_rate": 1.9843115187654028e-05, "loss": 0.8837, "step": 2341 }, { "epoch": 0.08482433900760594, "grad_norm": 1.710356924108065, "learning_rate": 1.984290814465792e-05, "loss": 0.9063, "step": 2342 }, { "epoch": 0.0848605577689243, "grad_norm": 1.823035067505134, "learning_rate": 1.984270096621475e-05, "loss": 0.9277, "step": 2343 }, { "epoch": 0.08489677653024266, "grad_norm": 1.3026999903513434, "learning_rate": 1.9842493652327367e-05, "loss": 0.7951, "step": 2344 }, { "epoch": 0.08493299529156102, "grad_norm": 1.8783187962742403, "learning_rate": 1.9842286202998625e-05, "loss": 0.7488, "step": 2345 }, { "epoch": 0.0849692140528794, "grad_norm": 2.0487076075685007, "learning_rate": 1.9842078618231375e-05, "loss": 0.9295, "step": 2346 }, { "epoch": 0.08500543281419776, "grad_norm": 1.6455390859286512, "learning_rate": 1.9841870898028476e-05, "loss": 0.8142, "step": 2347 }, { "epoch": 0.08504165157551612, "grad_norm": 1.6155900113949009, "learning_rate": 1.984166304239278e-05, "loss": 0.8562, "step": 2348 }, { "epoch": 0.08507787033683448, "grad_norm": 2.283325617060333, "learning_rate": 1.9841455051327157e-05, "loss": 0.8619, "step": 2349 }, { "epoch": 0.08511408909815284, "grad_norm": 1.8291377788528564, "learning_rate": 1.9841246924834467e-05, "loss": 0.9605, "step": 2350 }, { "epoch": 0.08515030785947121, "grad_norm": 1.8270311149718477, "learning_rate": 1.984103866291757e-05, "loss": 0.9184, "step": 2351 }, { "epoch": 0.08518652662078957, "grad_norm": 1.841434489629639, "learning_rate": 1.9840830265579334e-05, "loss": 0.8961, "step": 2352 }, { "epoch": 0.08522274538210793, "grad_norm": 2.1513906920019625, "learning_rate": 1.9840621732822628e-05, "loss": 0.8844, "step": 2353 }, { "epoch": 0.08525896414342629, "grad_norm": 1.468935433720868, "learning_rate": 1.9840413064650318e-05, "loss": 0.7804, "step": 2354 }, { "epoch": 0.08529518290474465, "grad_norm": 1.6095509305642681, "learning_rate": 1.9840204261065276e-05, "loss": 0.8338, "step": 2355 }, { "epoch": 0.08533140166606303, "grad_norm": 1.7318415352483616, "learning_rate": 1.983999532207038e-05, "loss": 0.8743, "step": 2356 }, { "epoch": 0.08536762042738139, "grad_norm": 1.4391554614622941, "learning_rate": 1.98397862476685e-05, "loss": 0.7711, "step": 2357 }, { "epoch": 0.08540383918869975, "grad_norm": 1.586138454164642, "learning_rate": 1.9839577037862518e-05, "loss": 0.8592, "step": 2358 }, { "epoch": 0.0854400579500181, "grad_norm": 1.573104699660845, "learning_rate": 1.983936769265531e-05, "loss": 0.8542, "step": 2359 }, { "epoch": 0.08547627671133647, "grad_norm": 1.2854795905653233, "learning_rate": 1.9839158212049755e-05, "loss": 0.7334, "step": 2360 }, { "epoch": 0.08551249547265484, "grad_norm": 1.9615037355133367, "learning_rate": 1.983894859604874e-05, "loss": 0.9003, "step": 2361 }, { "epoch": 0.0855487142339732, "grad_norm": 1.7441048484968993, "learning_rate": 1.9838738844655144e-05, "loss": 0.8359, "step": 2362 }, { "epoch": 0.08558493299529156, "grad_norm": 1.3460449182978997, "learning_rate": 1.983852895787186e-05, "loss": 0.7859, "step": 2363 }, { "epoch": 0.08562115175660992, "grad_norm": 1.7611621762015282, "learning_rate": 1.983831893570177e-05, "loss": 0.8794, "step": 2364 }, { "epoch": 0.08565737051792828, "grad_norm": 1.913521991045232, "learning_rate": 1.9838108778147767e-05, "loss": 0.7931, "step": 2365 }, { "epoch": 0.08569358927924665, "grad_norm": 1.6476211402122307, "learning_rate": 1.9837898485212744e-05, "loss": 0.8296, "step": 2366 }, { "epoch": 0.08572980804056501, "grad_norm": 1.235439153721027, "learning_rate": 1.9837688056899593e-05, "loss": 0.8263, "step": 2367 }, { "epoch": 0.08576602680188337, "grad_norm": 1.8974015113214364, "learning_rate": 1.9837477493211207e-05, "loss": 0.9253, "step": 2368 }, { "epoch": 0.08580224556320173, "grad_norm": 1.668065056600137, "learning_rate": 1.9837266794150493e-05, "loss": 0.9483, "step": 2369 }, { "epoch": 0.0858384643245201, "grad_norm": 1.929280368427691, "learning_rate": 1.9837055959720337e-05, "loss": 0.937, "step": 2370 }, { "epoch": 0.08587468308583847, "grad_norm": 1.936386655669728, "learning_rate": 1.983684498992365e-05, "loss": 0.9543, "step": 2371 }, { "epoch": 0.08591090184715683, "grad_norm": 1.766323820584104, "learning_rate": 1.9836633884763338e-05, "loss": 0.8922, "step": 2372 }, { "epoch": 0.08594712060847519, "grad_norm": 1.714256676543886, "learning_rate": 1.9836422644242295e-05, "loss": 0.9299, "step": 2373 }, { "epoch": 0.08598333936979355, "grad_norm": 1.9981936124017479, "learning_rate": 1.983621126836343e-05, "loss": 0.8705, "step": 2374 }, { "epoch": 0.08601955813111192, "grad_norm": 1.7943538435257103, "learning_rate": 1.9835999757129665e-05, "loss": 0.8492, "step": 2375 }, { "epoch": 0.08605577689243028, "grad_norm": 1.3081908152783641, "learning_rate": 1.9835788110543893e-05, "loss": 0.8613, "step": 2376 }, { "epoch": 0.08609199565374864, "grad_norm": 1.3652463819149065, "learning_rate": 1.9835576328609035e-05, "loss": 0.8094, "step": 2377 }, { "epoch": 0.086128214415067, "grad_norm": 1.8202745603528754, "learning_rate": 1.9835364411328006e-05, "loss": 0.8277, "step": 2378 }, { "epoch": 0.08616443317638536, "grad_norm": 1.282446776714616, "learning_rate": 1.9835152358703723e-05, "loss": 0.829, "step": 2379 }, { "epoch": 0.08620065193770374, "grad_norm": 1.6025004267466505, "learning_rate": 1.98349401707391e-05, "loss": 0.8396, "step": 2380 }, { "epoch": 0.0862368706990221, "grad_norm": 1.8324247956278077, "learning_rate": 1.9834727847437055e-05, "loss": 0.8738, "step": 2381 }, { "epoch": 0.08627308946034046, "grad_norm": 1.297231630230889, "learning_rate": 1.9834515388800516e-05, "loss": 0.7957, "step": 2382 }, { "epoch": 0.08630930822165882, "grad_norm": 1.1504375836975795, "learning_rate": 1.9834302794832404e-05, "loss": 0.7617, "step": 2383 }, { "epoch": 0.08634552698297718, "grad_norm": 1.9881600762711458, "learning_rate": 1.9834090065535644e-05, "loss": 0.8243, "step": 2384 }, { "epoch": 0.08638174574429555, "grad_norm": 1.6470699405975084, "learning_rate": 1.9833877200913166e-05, "loss": 0.8828, "step": 2385 }, { "epoch": 0.08641796450561391, "grad_norm": 1.7727853042107526, "learning_rate": 1.9833664200967892e-05, "loss": 0.9821, "step": 2386 }, { "epoch": 0.08645418326693227, "grad_norm": 1.4284273072689424, "learning_rate": 1.9833451065702762e-05, "loss": 0.8064, "step": 2387 }, { "epoch": 0.08649040202825063, "grad_norm": 1.772359194607127, "learning_rate": 1.9833237795120703e-05, "loss": 0.9364, "step": 2388 }, { "epoch": 0.08652662078956899, "grad_norm": 1.8529438025096867, "learning_rate": 1.9833024389224653e-05, "loss": 0.9193, "step": 2389 }, { "epoch": 0.08656283955088737, "grad_norm": 1.713931803558457, "learning_rate": 1.9832810848017547e-05, "loss": 0.8294, "step": 2390 }, { "epoch": 0.08659905831220573, "grad_norm": 1.3587899586975627, "learning_rate": 1.983259717150232e-05, "loss": 0.8007, "step": 2391 }, { "epoch": 0.08663527707352409, "grad_norm": 1.9569302454788955, "learning_rate": 1.9832383359681923e-05, "loss": 0.8472, "step": 2392 }, { "epoch": 0.08667149583484245, "grad_norm": 1.749924494187337, "learning_rate": 1.9832169412559284e-05, "loss": 0.7852, "step": 2393 }, { "epoch": 0.0867077145961608, "grad_norm": 2.0323663614885903, "learning_rate": 1.9831955330137362e-05, "loss": 0.9493, "step": 2394 }, { "epoch": 0.08674393335747918, "grad_norm": 2.025299221025305, "learning_rate": 1.983174111241909e-05, "loss": 0.8748, "step": 2395 }, { "epoch": 0.08678015211879754, "grad_norm": 1.6789853158601775, "learning_rate": 1.9831526759407425e-05, "loss": 0.8508, "step": 2396 }, { "epoch": 0.0868163708801159, "grad_norm": 1.7796960355739857, "learning_rate": 1.9831312271105313e-05, "loss": 0.9519, "step": 2397 }, { "epoch": 0.08685258964143426, "grad_norm": 1.5359290612930976, "learning_rate": 1.9831097647515704e-05, "loss": 0.9094, "step": 2398 }, { "epoch": 0.08688880840275262, "grad_norm": 1.8544100130023424, "learning_rate": 1.9830882888641555e-05, "loss": 0.9578, "step": 2399 }, { "epoch": 0.086925027164071, "grad_norm": 1.720579132046527, "learning_rate": 1.9830667994485818e-05, "loss": 0.8797, "step": 2400 }, { "epoch": 0.08696124592538935, "grad_norm": 1.9290764706122894, "learning_rate": 1.983045296505145e-05, "loss": 0.8987, "step": 2401 }, { "epoch": 0.08699746468670772, "grad_norm": 1.8508996546611902, "learning_rate": 1.983023780034141e-05, "loss": 0.838, "step": 2402 }, { "epoch": 0.08703368344802608, "grad_norm": 1.7125829542126838, "learning_rate": 1.9830022500358664e-05, "loss": 0.8417, "step": 2403 }, { "epoch": 0.08706990220934444, "grad_norm": 1.750822929624973, "learning_rate": 1.982980706510617e-05, "loss": 0.7506, "step": 2404 }, { "epoch": 0.08710612097066281, "grad_norm": 1.8306055886426909, "learning_rate": 1.9829591494586894e-05, "loss": 0.9209, "step": 2405 }, { "epoch": 0.08714233973198117, "grad_norm": 1.5574433412069966, "learning_rate": 1.98293757888038e-05, "loss": 0.8975, "step": 2406 }, { "epoch": 0.08717855849329953, "grad_norm": 1.7380901944251612, "learning_rate": 1.982915994775986e-05, "loss": 0.8716, "step": 2407 }, { "epoch": 0.08721477725461789, "grad_norm": 1.6498165949394812, "learning_rate": 1.9828943971458042e-05, "loss": 0.8638, "step": 2408 }, { "epoch": 0.08725099601593625, "grad_norm": 1.74019498598538, "learning_rate": 1.982872785990132e-05, "loss": 0.9597, "step": 2409 }, { "epoch": 0.08728721477725462, "grad_norm": 1.715320147335542, "learning_rate": 1.982851161309266e-05, "loss": 0.8752, "step": 2410 }, { "epoch": 0.08732343353857298, "grad_norm": 1.913068133549314, "learning_rate": 1.9828295231035054e-05, "loss": 0.8821, "step": 2411 }, { "epoch": 0.08735965229989134, "grad_norm": 1.705898324693412, "learning_rate": 1.9828078713731463e-05, "loss": 0.9213, "step": 2412 }, { "epoch": 0.0873958710612097, "grad_norm": 1.7942254144228718, "learning_rate": 1.9827862061184876e-05, "loss": 0.9688, "step": 2413 }, { "epoch": 0.08743208982252806, "grad_norm": 1.54931825175748, "learning_rate": 1.982764527339827e-05, "loss": 0.8372, "step": 2414 }, { "epoch": 0.08746830858384644, "grad_norm": 1.770294298615449, "learning_rate": 1.982742835037463e-05, "loss": 0.857, "step": 2415 }, { "epoch": 0.0875045273451648, "grad_norm": 1.7564618855910528, "learning_rate": 1.9827211292116943e-05, "loss": 0.909, "step": 2416 }, { "epoch": 0.08754074610648316, "grad_norm": 1.633294624123137, "learning_rate": 1.982699409862819e-05, "loss": 0.8694, "step": 2417 }, { "epoch": 0.08757696486780152, "grad_norm": 1.7500723968585132, "learning_rate": 1.9826776769911365e-05, "loss": 0.8731, "step": 2418 }, { "epoch": 0.08761318362911988, "grad_norm": 1.6990558354338423, "learning_rate": 1.9826559305969458e-05, "loss": 0.8936, "step": 2419 }, { "epoch": 0.08764940239043825, "grad_norm": 1.6921085569084815, "learning_rate": 1.982634170680546e-05, "loss": 0.8822, "step": 2420 }, { "epoch": 0.08768562115175661, "grad_norm": 1.7622523076051053, "learning_rate": 1.9826123972422364e-05, "loss": 0.9382, "step": 2421 }, { "epoch": 0.08772183991307497, "grad_norm": 1.9583862053213417, "learning_rate": 1.982590610282317e-05, "loss": 0.9375, "step": 2422 }, { "epoch": 0.08775805867439333, "grad_norm": 1.704227091475131, "learning_rate": 1.9825688098010873e-05, "loss": 0.8713, "step": 2423 }, { "epoch": 0.08779427743571169, "grad_norm": 1.3901137333591946, "learning_rate": 1.9825469957988477e-05, "loss": 0.7928, "step": 2424 }, { "epoch": 0.08783049619703007, "grad_norm": 1.6634660679512023, "learning_rate": 1.982525168275898e-05, "loss": 0.8123, "step": 2425 }, { "epoch": 0.08786671495834843, "grad_norm": 1.6999281038392573, "learning_rate": 1.9825033272325386e-05, "loss": 0.9235, "step": 2426 }, { "epoch": 0.08790293371966679, "grad_norm": 1.687685100902862, "learning_rate": 1.98248147266907e-05, "loss": 0.8766, "step": 2427 }, { "epoch": 0.08793915248098515, "grad_norm": 1.7891044757533034, "learning_rate": 1.9824596045857932e-05, "loss": 0.889, "step": 2428 }, { "epoch": 0.08797537124230351, "grad_norm": 1.7488009685668413, "learning_rate": 1.9824377229830088e-05, "loss": 0.8917, "step": 2429 }, { "epoch": 0.08801159000362188, "grad_norm": 1.3866100201912464, "learning_rate": 1.982415827861018e-05, "loss": 0.8533, "step": 2430 }, { "epoch": 0.08804780876494024, "grad_norm": 1.7102091754480329, "learning_rate": 1.9823939192201224e-05, "loss": 0.8802, "step": 2431 }, { "epoch": 0.0880840275262586, "grad_norm": 1.8051183916895113, "learning_rate": 1.9823719970606235e-05, "loss": 0.8588, "step": 2432 }, { "epoch": 0.08812024628757696, "grad_norm": 2.006656117051799, "learning_rate": 1.9823500613828223e-05, "loss": 0.779, "step": 2433 }, { "epoch": 0.08815646504889532, "grad_norm": 1.4523753235549162, "learning_rate": 1.9823281121870215e-05, "loss": 0.7642, "step": 2434 }, { "epoch": 0.0881926838102137, "grad_norm": 1.8466581749835598, "learning_rate": 1.9823061494735224e-05, "loss": 0.9359, "step": 2435 }, { "epoch": 0.08822890257153206, "grad_norm": 1.7907417300742152, "learning_rate": 1.9822841732426277e-05, "loss": 0.7992, "step": 2436 }, { "epoch": 0.08826512133285042, "grad_norm": 2.0085962302193745, "learning_rate": 1.9822621834946397e-05, "loss": 0.982, "step": 2437 }, { "epoch": 0.08830134009416878, "grad_norm": 1.6479946757510957, "learning_rate": 1.9822401802298608e-05, "loss": 0.8986, "step": 2438 }, { "epoch": 0.08833755885548714, "grad_norm": 1.7549831423262285, "learning_rate": 1.982218163448594e-05, "loss": 0.8245, "step": 2439 }, { "epoch": 0.08837377761680551, "grad_norm": 1.5474499954718837, "learning_rate": 1.9821961331511424e-05, "loss": 0.8819, "step": 2440 }, { "epoch": 0.08840999637812387, "grad_norm": 1.6651250615661874, "learning_rate": 1.9821740893378084e-05, "loss": 0.8184, "step": 2441 }, { "epoch": 0.08844621513944223, "grad_norm": 1.990603421635502, "learning_rate": 1.9821520320088968e-05, "loss": 0.7572, "step": 2442 }, { "epoch": 0.08848243390076059, "grad_norm": 1.4226969502056448, "learning_rate": 1.9821299611647096e-05, "loss": 0.8145, "step": 2443 }, { "epoch": 0.08851865266207895, "grad_norm": 1.892690920937754, "learning_rate": 1.9821078768055516e-05, "loss": 0.8377, "step": 2444 }, { "epoch": 0.08855487142339732, "grad_norm": 1.9201798949616096, "learning_rate": 1.9820857789317255e-05, "loss": 0.9243, "step": 2445 }, { "epoch": 0.08859109018471568, "grad_norm": 1.450152521112188, "learning_rate": 1.9820636675435368e-05, "loss": 0.8184, "step": 2446 }, { "epoch": 0.08862730894603404, "grad_norm": 1.9016741939236756, "learning_rate": 1.982041542641289e-05, "loss": 0.8975, "step": 2447 }, { "epoch": 0.0886635277073524, "grad_norm": 1.688885504123767, "learning_rate": 1.9820194042252867e-05, "loss": 0.8911, "step": 2448 }, { "epoch": 0.08869974646867076, "grad_norm": 1.309151607785952, "learning_rate": 1.9819972522958345e-05, "loss": 0.7914, "step": 2449 }, { "epoch": 0.08873596522998914, "grad_norm": 1.7354011013364106, "learning_rate": 1.981975086853237e-05, "loss": 0.842, "step": 2450 }, { "epoch": 0.0887721839913075, "grad_norm": 1.2029937224130132, "learning_rate": 1.9819529078978e-05, "loss": 0.7511, "step": 2451 }, { "epoch": 0.08880840275262586, "grad_norm": 1.7533347796063916, "learning_rate": 1.9819307154298277e-05, "loss": 0.9016, "step": 2452 }, { "epoch": 0.08884462151394422, "grad_norm": 1.4235100246098353, "learning_rate": 1.981908509449626e-05, "loss": 0.7687, "step": 2453 }, { "epoch": 0.08888084027526258, "grad_norm": 1.9233409380903614, "learning_rate": 1.9818862899575008e-05, "loss": 0.8989, "step": 2454 }, { "epoch": 0.08891705903658095, "grad_norm": 1.6743447397882332, "learning_rate": 1.981864056953757e-05, "loss": 0.8572, "step": 2455 }, { "epoch": 0.08895327779789931, "grad_norm": 1.8371814085042963, "learning_rate": 1.9818418104387015e-05, "loss": 0.869, "step": 2456 }, { "epoch": 0.08898949655921767, "grad_norm": 1.7205818124845906, "learning_rate": 1.9818195504126396e-05, "loss": 0.793, "step": 2457 }, { "epoch": 0.08902571532053603, "grad_norm": 1.5212260695440658, "learning_rate": 1.981797276875878e-05, "loss": 0.817, "step": 2458 }, { "epoch": 0.0890619340818544, "grad_norm": 1.7608087445637002, "learning_rate": 1.9817749898287232e-05, "loss": 0.9634, "step": 2459 }, { "epoch": 0.08909815284317277, "grad_norm": 1.7090223188275193, "learning_rate": 1.9817526892714822e-05, "loss": 0.8259, "step": 2460 }, { "epoch": 0.08913437160449113, "grad_norm": 1.6667908455860554, "learning_rate": 1.981730375204461e-05, "loss": 0.8292, "step": 2461 }, { "epoch": 0.08917059036580949, "grad_norm": 1.62807136404281, "learning_rate": 1.9817080476279676e-05, "loss": 0.854, "step": 2462 }, { "epoch": 0.08920680912712785, "grad_norm": 1.5220208660637369, "learning_rate": 1.981685706542309e-05, "loss": 0.7877, "step": 2463 }, { "epoch": 0.08924302788844622, "grad_norm": 1.8013059116576715, "learning_rate": 1.9816633519477923e-05, "loss": 0.8219, "step": 2464 }, { "epoch": 0.08927924664976458, "grad_norm": 1.9754707106713527, "learning_rate": 1.9816409838447255e-05, "loss": 0.7993, "step": 2465 }, { "epoch": 0.08931546541108294, "grad_norm": 1.4872216316071067, "learning_rate": 1.981618602233416e-05, "loss": 0.7475, "step": 2466 }, { "epoch": 0.0893516841724013, "grad_norm": 1.795437930600086, "learning_rate": 1.981596207114172e-05, "loss": 0.9333, "step": 2467 }, { "epoch": 0.08938790293371966, "grad_norm": 1.9752732018543364, "learning_rate": 1.9815737984873016e-05, "loss": 0.892, "step": 2468 }, { "epoch": 0.08942412169503804, "grad_norm": 1.7531644021160535, "learning_rate": 1.9815513763531134e-05, "loss": 0.8779, "step": 2469 }, { "epoch": 0.0894603404563564, "grad_norm": 1.8566851774488748, "learning_rate": 1.9815289407119156e-05, "loss": 0.8985, "step": 2470 }, { "epoch": 0.08949655921767476, "grad_norm": 1.8025583561509426, "learning_rate": 1.9815064915640174e-05, "loss": 0.8654, "step": 2471 }, { "epoch": 0.08953277797899312, "grad_norm": 1.5607015595144438, "learning_rate": 1.9814840289097273e-05, "loss": 0.8904, "step": 2472 }, { "epoch": 0.08956899674031148, "grad_norm": 1.3213465450175141, "learning_rate": 1.981461552749355e-05, "loss": 0.7902, "step": 2473 }, { "epoch": 0.08960521550162985, "grad_norm": 1.751279948258937, "learning_rate": 1.9814390630832084e-05, "loss": 0.8678, "step": 2474 }, { "epoch": 0.08964143426294821, "grad_norm": 1.7218517881722393, "learning_rate": 1.9814165599115987e-05, "loss": 0.8458, "step": 2475 }, { "epoch": 0.08967765302426657, "grad_norm": 1.7727242266355365, "learning_rate": 1.9813940432348343e-05, "loss": 0.968, "step": 2476 }, { "epoch": 0.08971387178558493, "grad_norm": 1.7388424149717165, "learning_rate": 1.981371513053226e-05, "loss": 0.8243, "step": 2477 }, { "epoch": 0.08975009054690329, "grad_norm": 2.099107462307881, "learning_rate": 1.981348969367083e-05, "loss": 0.8861, "step": 2478 }, { "epoch": 0.08978630930822167, "grad_norm": 1.6646158189452342, "learning_rate": 1.9813264121767163e-05, "loss": 0.8873, "step": 2479 }, { "epoch": 0.08982252806954003, "grad_norm": 1.7700408060834174, "learning_rate": 1.9813038414824356e-05, "loss": 0.9003, "step": 2480 }, { "epoch": 0.08985874683085839, "grad_norm": 1.6689963067904134, "learning_rate": 1.9812812572845516e-05, "loss": 0.8094, "step": 2481 }, { "epoch": 0.08989496559217675, "grad_norm": 1.6774119488088997, "learning_rate": 1.9812586595833755e-05, "loss": 0.8269, "step": 2482 }, { "epoch": 0.0899311843534951, "grad_norm": 1.6780942835030057, "learning_rate": 1.9812360483792183e-05, "loss": 0.8339, "step": 2483 }, { "epoch": 0.08996740311481348, "grad_norm": 1.86605082113969, "learning_rate": 1.9812134236723905e-05, "loss": 0.8499, "step": 2484 }, { "epoch": 0.09000362187613184, "grad_norm": 1.7347674258780457, "learning_rate": 1.981190785463204e-05, "loss": 0.8884, "step": 2485 }, { "epoch": 0.0900398406374502, "grad_norm": 1.9505073286301189, "learning_rate": 1.9811681337519702e-05, "loss": 0.9538, "step": 2486 }, { "epoch": 0.09007605939876856, "grad_norm": 1.579164306771806, "learning_rate": 1.9811454685390007e-05, "loss": 0.8054, "step": 2487 }, { "epoch": 0.09011227816008692, "grad_norm": 1.7495315821055295, "learning_rate": 1.9811227898246072e-05, "loss": 0.8778, "step": 2488 }, { "epoch": 0.0901484969214053, "grad_norm": 1.3187326619824336, "learning_rate": 1.9811000976091025e-05, "loss": 0.7222, "step": 2489 }, { "epoch": 0.09018471568272365, "grad_norm": 2.3727387656043457, "learning_rate": 1.981077391892798e-05, "loss": 0.8671, "step": 2490 }, { "epoch": 0.09022093444404201, "grad_norm": 1.9309767092230379, "learning_rate": 1.9810546726760064e-05, "loss": 0.8907, "step": 2491 }, { "epoch": 0.09025715320536037, "grad_norm": 1.6277483522763918, "learning_rate": 1.9810319399590406e-05, "loss": 0.9174, "step": 2492 }, { "epoch": 0.09029337196667873, "grad_norm": 1.7372671909156243, "learning_rate": 1.9810091937422134e-05, "loss": 0.7971, "step": 2493 }, { "epoch": 0.09032959072799711, "grad_norm": 2.151761628786657, "learning_rate": 1.980986434025838e-05, "loss": 0.8804, "step": 2494 }, { "epoch": 0.09036580948931547, "grad_norm": 1.8806060588741096, "learning_rate": 1.980963660810227e-05, "loss": 0.8612, "step": 2495 }, { "epoch": 0.09040202825063383, "grad_norm": 1.9181738295139918, "learning_rate": 1.980940874095694e-05, "loss": 0.9095, "step": 2496 }, { "epoch": 0.09043824701195219, "grad_norm": 1.7382934430584118, "learning_rate": 1.9809180738825527e-05, "loss": 0.8246, "step": 2497 }, { "epoch": 0.09047446577327055, "grad_norm": 1.7531884876356008, "learning_rate": 1.980895260171117e-05, "loss": 0.8458, "step": 2498 }, { "epoch": 0.09051068453458892, "grad_norm": 1.6891557128507293, "learning_rate": 1.9808724329617e-05, "loss": 0.879, "step": 2499 }, { "epoch": 0.09054690329590728, "grad_norm": 1.75941742956823, "learning_rate": 1.980849592254617e-05, "loss": 0.7662, "step": 2500 }, { "epoch": 0.09058312205722564, "grad_norm": 1.5898378422460993, "learning_rate": 1.980826738050182e-05, "loss": 0.779, "step": 2501 }, { "epoch": 0.090619340818544, "grad_norm": 1.430753284250089, "learning_rate": 1.9808038703487088e-05, "loss": 0.7817, "step": 2502 }, { "epoch": 0.09065555957986236, "grad_norm": 1.7749922612854334, "learning_rate": 1.9807809891505126e-05, "loss": 0.8618, "step": 2503 }, { "epoch": 0.09069177834118074, "grad_norm": 1.7083537368687483, "learning_rate": 1.9807580944559083e-05, "loss": 0.7468, "step": 2504 }, { "epoch": 0.0907279971024991, "grad_norm": 1.6743266338126124, "learning_rate": 1.9807351862652108e-05, "loss": 0.8657, "step": 2505 }, { "epoch": 0.09076421586381746, "grad_norm": 1.8907763128970392, "learning_rate": 1.9807122645787354e-05, "loss": 0.882, "step": 2506 }, { "epoch": 0.09080043462513582, "grad_norm": 1.6170984775808714, "learning_rate": 1.9806893293967974e-05, "loss": 0.9525, "step": 2507 }, { "epoch": 0.09083665338645418, "grad_norm": 1.6466199382393458, "learning_rate": 1.9806663807197127e-05, "loss": 0.8488, "step": 2508 }, { "epoch": 0.09087287214777255, "grad_norm": 1.623926135719275, "learning_rate": 1.9806434185477966e-05, "loss": 0.8721, "step": 2509 }, { "epoch": 0.09090909090909091, "grad_norm": 1.8265462603797524, "learning_rate": 1.9806204428813656e-05, "loss": 0.8584, "step": 2510 }, { "epoch": 0.09094530967040927, "grad_norm": 1.4949426514087878, "learning_rate": 1.980597453720736e-05, "loss": 0.7057, "step": 2511 }, { "epoch": 0.09098152843172763, "grad_norm": 1.8184062468227138, "learning_rate": 1.9805744510662235e-05, "loss": 0.9136, "step": 2512 }, { "epoch": 0.09101774719304599, "grad_norm": 1.358974439092119, "learning_rate": 1.980551434918145e-05, "loss": 0.8027, "step": 2513 }, { "epoch": 0.09105396595436437, "grad_norm": 2.14864933833541, "learning_rate": 1.980528405276817e-05, "loss": 0.8512, "step": 2514 }, { "epoch": 0.09109018471568273, "grad_norm": 1.8950507228518452, "learning_rate": 1.9805053621425565e-05, "loss": 0.8807, "step": 2515 }, { "epoch": 0.09112640347700109, "grad_norm": 1.871153790699237, "learning_rate": 1.9804823055156807e-05, "loss": 0.8955, "step": 2516 }, { "epoch": 0.09116262223831945, "grad_norm": 1.58431749098424, "learning_rate": 1.9804592353965072e-05, "loss": 0.8771, "step": 2517 }, { "epoch": 0.0911988409996378, "grad_norm": 1.5251801612095088, "learning_rate": 1.980436151785353e-05, "loss": 0.8379, "step": 2518 }, { "epoch": 0.09123505976095618, "grad_norm": 1.595894830620527, "learning_rate": 1.9804130546825356e-05, "loss": 0.9036, "step": 2519 }, { "epoch": 0.09127127852227454, "grad_norm": 1.9237906127990365, "learning_rate": 1.9803899440883735e-05, "loss": 0.7616, "step": 2520 }, { "epoch": 0.0913074972835929, "grad_norm": 1.3234597748634953, "learning_rate": 1.980366820003184e-05, "loss": 0.7718, "step": 2521 }, { "epoch": 0.09134371604491126, "grad_norm": 1.720156100956871, "learning_rate": 1.9803436824272856e-05, "loss": 0.8719, "step": 2522 }, { "epoch": 0.09137993480622962, "grad_norm": 1.8276814998732631, "learning_rate": 1.980320531360997e-05, "loss": 0.8324, "step": 2523 }, { "epoch": 0.091416153567548, "grad_norm": 1.3209567624850829, "learning_rate": 1.9802973668046364e-05, "loss": 0.8467, "step": 2524 }, { "epoch": 0.09145237232886635, "grad_norm": 1.949346624166642, "learning_rate": 1.9802741887585224e-05, "loss": 0.872, "step": 2525 }, { "epoch": 0.09148859109018471, "grad_norm": 1.9623908208436132, "learning_rate": 1.9802509972229743e-05, "loss": 0.8865, "step": 2526 }, { "epoch": 0.09152480985150308, "grad_norm": 1.436066570898205, "learning_rate": 1.9802277921983114e-05, "loss": 0.7617, "step": 2527 }, { "epoch": 0.09156102861282144, "grad_norm": 1.6000392124374043, "learning_rate": 1.9802045736848527e-05, "loss": 0.9526, "step": 2528 }, { "epoch": 0.09159724737413981, "grad_norm": 2.0820929498458676, "learning_rate": 1.980181341682918e-05, "loss": 0.9699, "step": 2529 }, { "epoch": 0.09163346613545817, "grad_norm": 1.7672662375521457, "learning_rate": 1.9801580961928263e-05, "loss": 0.9457, "step": 2530 }, { "epoch": 0.09166968489677653, "grad_norm": 1.8137454776917858, "learning_rate": 1.980134837214898e-05, "loss": 0.8909, "step": 2531 }, { "epoch": 0.09170590365809489, "grad_norm": 1.7578785142158841, "learning_rate": 1.9801115647494537e-05, "loss": 0.8277, "step": 2532 }, { "epoch": 0.09174212241941325, "grad_norm": 1.7482599288420375, "learning_rate": 1.9800882787968126e-05, "loss": 0.8987, "step": 2533 }, { "epoch": 0.09177834118073162, "grad_norm": 1.6421089704376846, "learning_rate": 1.9800649793572956e-05, "loss": 0.8461, "step": 2534 }, { "epoch": 0.09181455994204998, "grad_norm": 1.624460634025156, "learning_rate": 1.9800416664312235e-05, "loss": 0.8038, "step": 2535 }, { "epoch": 0.09185077870336834, "grad_norm": 1.6436496815963888, "learning_rate": 1.9800183400189167e-05, "loss": 0.8467, "step": 2536 }, { "epoch": 0.0918869974646867, "grad_norm": 1.7128264079132174, "learning_rate": 1.9799950001206968e-05, "loss": 0.8803, "step": 2537 }, { "epoch": 0.09192321622600506, "grad_norm": 1.7146452406512227, "learning_rate": 1.979971646736884e-05, "loss": 0.9606, "step": 2538 }, { "epoch": 0.09195943498732344, "grad_norm": 1.8129390850630305, "learning_rate": 1.9799482798678008e-05, "loss": 0.9147, "step": 2539 }, { "epoch": 0.0919956537486418, "grad_norm": 1.5348709297208047, "learning_rate": 1.979924899513768e-05, "loss": 0.7822, "step": 2540 }, { "epoch": 0.09203187250996016, "grad_norm": 1.542993576115661, "learning_rate": 1.9799015056751078e-05, "loss": 0.8645, "step": 2541 }, { "epoch": 0.09206809127127852, "grad_norm": 1.7532362104100296, "learning_rate": 1.9798780983521417e-05, "loss": 0.9497, "step": 2542 }, { "epoch": 0.09210431003259688, "grad_norm": 1.4126336702647266, "learning_rate": 1.9798546775451917e-05, "loss": 0.8076, "step": 2543 }, { "epoch": 0.09214052879391525, "grad_norm": 1.741317368152019, "learning_rate": 1.9798312432545807e-05, "loss": 0.9083, "step": 2544 }, { "epoch": 0.09217674755523361, "grad_norm": 1.682161646442672, "learning_rate": 1.9798077954806306e-05, "loss": 0.8923, "step": 2545 }, { "epoch": 0.09221296631655197, "grad_norm": 1.7735748227345989, "learning_rate": 1.9797843342236642e-05, "loss": 0.8866, "step": 2546 }, { "epoch": 0.09224918507787033, "grad_norm": 1.8184671994865258, "learning_rate": 1.9797608594840048e-05, "loss": 0.8239, "step": 2547 }, { "epoch": 0.0922854038391887, "grad_norm": 2.0138574749635896, "learning_rate": 1.9797373712619746e-05, "loss": 0.9286, "step": 2548 }, { "epoch": 0.09232162260050707, "grad_norm": 1.804193420813798, "learning_rate": 1.979713869557898e-05, "loss": 0.9342, "step": 2549 }, { "epoch": 0.09235784136182543, "grad_norm": 1.8449304042976806, "learning_rate": 1.9796903543720974e-05, "loss": 0.9292, "step": 2550 }, { "epoch": 0.09239406012314379, "grad_norm": 1.7924022361332266, "learning_rate": 1.9796668257048965e-05, "loss": 0.9042, "step": 2551 }, { "epoch": 0.09243027888446215, "grad_norm": 1.4164323367895295, "learning_rate": 1.979643283556619e-05, "loss": 0.7844, "step": 2552 }, { "epoch": 0.09246649764578052, "grad_norm": 1.5131056390988664, "learning_rate": 1.97961972792759e-05, "loss": 0.8524, "step": 2553 }, { "epoch": 0.09250271640709888, "grad_norm": 1.5080185060618487, "learning_rate": 1.979596158818132e-05, "loss": 0.7931, "step": 2554 }, { "epoch": 0.09253893516841724, "grad_norm": 1.7623613109790892, "learning_rate": 1.9795725762285704e-05, "loss": 0.9375, "step": 2555 }, { "epoch": 0.0925751539297356, "grad_norm": 1.8620427452027712, "learning_rate": 1.9795489801592293e-05, "loss": 0.8824, "step": 2556 }, { "epoch": 0.09261137269105396, "grad_norm": 1.8401396041755989, "learning_rate": 1.979525370610434e-05, "loss": 0.9067, "step": 2557 }, { "epoch": 0.09264759145237234, "grad_norm": 1.6771483191541996, "learning_rate": 1.9795017475825084e-05, "loss": 0.881, "step": 2558 }, { "epoch": 0.0926838102136907, "grad_norm": 1.3380773735927716, "learning_rate": 1.9794781110757785e-05, "loss": 0.6987, "step": 2559 }, { "epoch": 0.09272002897500906, "grad_norm": 1.9479058350595089, "learning_rate": 1.9794544610905686e-05, "loss": 0.836, "step": 2560 }, { "epoch": 0.09275624773632742, "grad_norm": 1.7973679643780862, "learning_rate": 1.9794307976272052e-05, "loss": 0.9671, "step": 2561 }, { "epoch": 0.09279246649764578, "grad_norm": 1.387800298104646, "learning_rate": 1.979407120686013e-05, "loss": 0.7508, "step": 2562 }, { "epoch": 0.09282868525896415, "grad_norm": 1.502432251320594, "learning_rate": 1.9793834302673186e-05, "loss": 0.8095, "step": 2563 }, { "epoch": 0.09286490402028251, "grad_norm": 1.3182956222731528, "learning_rate": 1.9793597263714474e-05, "loss": 0.7813, "step": 2564 }, { "epoch": 0.09290112278160087, "grad_norm": 1.6484590671357833, "learning_rate": 1.979336008998726e-05, "loss": 0.7663, "step": 2565 }, { "epoch": 0.09293734154291923, "grad_norm": 1.718168700058977, "learning_rate": 1.9793122781494802e-05, "loss": 0.8098, "step": 2566 }, { "epoch": 0.09297356030423759, "grad_norm": 1.5866947239634974, "learning_rate": 1.9792885338240375e-05, "loss": 0.7597, "step": 2567 }, { "epoch": 0.09300977906555596, "grad_norm": 1.7578936441554183, "learning_rate": 1.9792647760227238e-05, "loss": 0.8468, "step": 2568 }, { "epoch": 0.09304599782687432, "grad_norm": 1.5804417778755193, "learning_rate": 1.9792410047458664e-05, "loss": 0.717, "step": 2569 }, { "epoch": 0.09308221658819268, "grad_norm": 1.7366106470726206, "learning_rate": 1.9792172199937926e-05, "loss": 0.8897, "step": 2570 }, { "epoch": 0.09311843534951104, "grad_norm": 1.4094495780874927, "learning_rate": 1.9791934217668286e-05, "loss": 0.77, "step": 2571 }, { "epoch": 0.0931546541108294, "grad_norm": 1.6876391816608258, "learning_rate": 1.9791696100653035e-05, "loss": 0.7766, "step": 2572 }, { "epoch": 0.09319087287214778, "grad_norm": 1.7820446433668202, "learning_rate": 1.979145784889544e-05, "loss": 0.8578, "step": 2573 }, { "epoch": 0.09322709163346614, "grad_norm": 1.7984364913710575, "learning_rate": 1.9791219462398782e-05, "loss": 0.8407, "step": 2574 }, { "epoch": 0.0932633103947845, "grad_norm": 1.6387552301532902, "learning_rate": 1.9790980941166337e-05, "loss": 0.7968, "step": 2575 }, { "epoch": 0.09329952915610286, "grad_norm": 1.68980297810136, "learning_rate": 1.9790742285201396e-05, "loss": 0.7813, "step": 2576 }, { "epoch": 0.09333574791742122, "grad_norm": 1.8224265200286356, "learning_rate": 1.9790503494507235e-05, "loss": 0.874, "step": 2577 }, { "epoch": 0.09337196667873959, "grad_norm": 2.2816067408109597, "learning_rate": 1.9790264569087147e-05, "loss": 0.8571, "step": 2578 }, { "epoch": 0.09340818544005795, "grad_norm": 1.4493496204675493, "learning_rate": 1.9790025508944412e-05, "loss": 0.7763, "step": 2579 }, { "epoch": 0.09344440420137631, "grad_norm": 1.7941313466987254, "learning_rate": 1.9789786314082327e-05, "loss": 0.8413, "step": 2580 }, { "epoch": 0.09348062296269467, "grad_norm": 1.702550814238935, "learning_rate": 1.978954698450418e-05, "loss": 0.8588, "step": 2581 }, { "epoch": 0.09351684172401303, "grad_norm": 2.2294854029511706, "learning_rate": 1.978930752021326e-05, "loss": 0.7871, "step": 2582 }, { "epoch": 0.09355306048533141, "grad_norm": 1.6744016668460915, "learning_rate": 1.9789067921212874e-05, "loss": 0.9208, "step": 2583 }, { "epoch": 0.09358927924664977, "grad_norm": 1.7479868404372645, "learning_rate": 1.9788828187506307e-05, "loss": 0.854, "step": 2584 }, { "epoch": 0.09362549800796813, "grad_norm": 1.7450273839709933, "learning_rate": 1.9788588319096864e-05, "loss": 0.8608, "step": 2585 }, { "epoch": 0.09366171676928649, "grad_norm": 1.6497784850590336, "learning_rate": 1.9788348315987843e-05, "loss": 0.8601, "step": 2586 }, { "epoch": 0.09369793553060485, "grad_norm": 1.546351690470789, "learning_rate": 1.9788108178182552e-05, "loss": 0.8845, "step": 2587 }, { "epoch": 0.09373415429192322, "grad_norm": 1.6922140637159386, "learning_rate": 1.978786790568429e-05, "loss": 0.7215, "step": 2588 }, { "epoch": 0.09377037305324158, "grad_norm": 1.8219720781874058, "learning_rate": 1.9787627498496366e-05, "loss": 0.9115, "step": 2589 }, { "epoch": 0.09380659181455994, "grad_norm": 1.6651426915113157, "learning_rate": 1.9787386956622085e-05, "loss": 0.8675, "step": 2590 }, { "epoch": 0.0938428105758783, "grad_norm": 1.6229997836349597, "learning_rate": 1.978714628006476e-05, "loss": 0.8347, "step": 2591 }, { "epoch": 0.09387902933719666, "grad_norm": 1.8446445598318795, "learning_rate": 1.9786905468827707e-05, "loss": 0.848, "step": 2592 }, { "epoch": 0.09391524809851504, "grad_norm": 1.7766203340797777, "learning_rate": 1.978666452291423e-05, "loss": 0.8848, "step": 2593 }, { "epoch": 0.0939514668598334, "grad_norm": 1.6103093781969373, "learning_rate": 1.9786423442327653e-05, "loss": 0.86, "step": 2594 }, { "epoch": 0.09398768562115176, "grad_norm": 1.8539335716234102, "learning_rate": 1.978618222707129e-05, "loss": 0.873, "step": 2595 }, { "epoch": 0.09402390438247012, "grad_norm": 1.6350144569497136, "learning_rate": 1.9785940877148457e-05, "loss": 0.8331, "step": 2596 }, { "epoch": 0.09406012314378848, "grad_norm": 1.631245182567367, "learning_rate": 1.9785699392562485e-05, "loss": 0.7879, "step": 2597 }, { "epoch": 0.09409634190510685, "grad_norm": 1.6812245355950155, "learning_rate": 1.9785457773316684e-05, "loss": 0.9406, "step": 2598 }, { "epoch": 0.09413256066642521, "grad_norm": 1.7923650951243508, "learning_rate": 1.978521601941439e-05, "loss": 0.9654, "step": 2599 }, { "epoch": 0.09416877942774357, "grad_norm": 1.7055078993885286, "learning_rate": 1.9784974130858926e-05, "loss": 0.8304, "step": 2600 }, { "epoch": 0.09420499818906193, "grad_norm": 1.5806961094737297, "learning_rate": 1.978473210765362e-05, "loss": 0.7936, "step": 2601 }, { "epoch": 0.09424121695038029, "grad_norm": 1.677714071273071, "learning_rate": 1.9784489949801798e-05, "loss": 0.8514, "step": 2602 }, { "epoch": 0.09427743571169866, "grad_norm": 1.8321478548336303, "learning_rate": 1.9784247657306804e-05, "loss": 0.9044, "step": 2603 }, { "epoch": 0.09431365447301703, "grad_norm": 1.740539134491342, "learning_rate": 1.978400523017196e-05, "loss": 0.7666, "step": 2604 }, { "epoch": 0.09434987323433539, "grad_norm": 1.7351400399943022, "learning_rate": 1.978376266840061e-05, "loss": 0.9138, "step": 2605 }, { "epoch": 0.09438609199565375, "grad_norm": 1.6452435267084435, "learning_rate": 1.9783519971996087e-05, "loss": 0.8166, "step": 2606 }, { "epoch": 0.0944223107569721, "grad_norm": 1.6395120978684155, "learning_rate": 1.9783277140961734e-05, "loss": 0.8839, "step": 2607 }, { "epoch": 0.09445852951829048, "grad_norm": 1.3779640126312487, "learning_rate": 1.9783034175300888e-05, "loss": 0.8543, "step": 2608 }, { "epoch": 0.09449474827960884, "grad_norm": 1.97090152085868, "learning_rate": 1.97827910750169e-05, "loss": 0.9202, "step": 2609 }, { "epoch": 0.0945309670409272, "grad_norm": 2.044386767280927, "learning_rate": 1.9782547840113106e-05, "loss": 0.9157, "step": 2610 }, { "epoch": 0.09456718580224556, "grad_norm": 1.810703124759699, "learning_rate": 1.978230447059286e-05, "loss": 0.9068, "step": 2611 }, { "epoch": 0.09460340456356392, "grad_norm": 1.7659461564861225, "learning_rate": 1.978206096645951e-05, "loss": 0.8178, "step": 2612 }, { "epoch": 0.0946396233248823, "grad_norm": 1.8803037573427446, "learning_rate": 1.9781817327716408e-05, "loss": 0.9098, "step": 2613 }, { "epoch": 0.09467584208620065, "grad_norm": 1.2563587021258722, "learning_rate": 1.9781573554366903e-05, "loss": 0.7684, "step": 2614 }, { "epoch": 0.09471206084751901, "grad_norm": 1.572499985139586, "learning_rate": 1.9781329646414348e-05, "loss": 0.8556, "step": 2615 }, { "epoch": 0.09474827960883737, "grad_norm": 1.698839110134779, "learning_rate": 1.9781085603862106e-05, "loss": 0.888, "step": 2616 }, { "epoch": 0.09478449837015573, "grad_norm": 1.770995667791197, "learning_rate": 1.978084142671353e-05, "loss": 0.865, "step": 2617 }, { "epoch": 0.09482071713147411, "grad_norm": 11.79283105997912, "learning_rate": 1.9780597114971983e-05, "loss": 0.9103, "step": 2618 }, { "epoch": 0.09485693589279247, "grad_norm": 2.0362492589791144, "learning_rate": 1.9780352668640822e-05, "loss": 0.8408, "step": 2619 }, { "epoch": 0.09489315465411083, "grad_norm": 1.608518272896863, "learning_rate": 1.978010808772342e-05, "loss": 0.8517, "step": 2620 }, { "epoch": 0.09492937341542919, "grad_norm": 1.7198787537384157, "learning_rate": 1.9779863372223132e-05, "loss": 0.877, "step": 2621 }, { "epoch": 0.09496559217674755, "grad_norm": 1.7652157186766158, "learning_rate": 1.9779618522143335e-05, "loss": 0.8769, "step": 2622 }, { "epoch": 0.09500181093806592, "grad_norm": 2.188276296199379, "learning_rate": 1.977937353748739e-05, "loss": 0.9126, "step": 2623 }, { "epoch": 0.09503802969938428, "grad_norm": 1.8477072738957405, "learning_rate": 1.9779128418258678e-05, "loss": 0.8464, "step": 2624 }, { "epoch": 0.09507424846070264, "grad_norm": 1.741089602151103, "learning_rate": 1.977888316446056e-05, "loss": 0.8677, "step": 2625 }, { "epoch": 0.095110467222021, "grad_norm": 1.7955425235443443, "learning_rate": 1.977863777609642e-05, "loss": 0.8958, "step": 2626 }, { "epoch": 0.09514668598333936, "grad_norm": 1.6657769813765582, "learning_rate": 1.9778392253169635e-05, "loss": 0.9121, "step": 2627 }, { "epoch": 0.09518290474465774, "grad_norm": 2.7426636982360106, "learning_rate": 1.9778146595683573e-05, "loss": 0.8422, "step": 2628 }, { "epoch": 0.0952191235059761, "grad_norm": 2.3106793873903806, "learning_rate": 1.9777900803641627e-05, "loss": 0.8635, "step": 2629 }, { "epoch": 0.09525534226729446, "grad_norm": 2.0175026826887956, "learning_rate": 1.9777654877047173e-05, "loss": 0.8695, "step": 2630 }, { "epoch": 0.09529156102861282, "grad_norm": 2.161108500240041, "learning_rate": 1.97774088159036e-05, "loss": 0.8179, "step": 2631 }, { "epoch": 0.09532777978993119, "grad_norm": 1.7305535926428424, "learning_rate": 1.977716262021429e-05, "loss": 0.8341, "step": 2632 }, { "epoch": 0.09536399855124955, "grad_norm": 1.927126126630436, "learning_rate": 1.9776916289982625e-05, "loss": 0.9176, "step": 2633 }, { "epoch": 0.09540021731256791, "grad_norm": 1.9494856917278627, "learning_rate": 1.9776669825212005e-05, "loss": 0.9034, "step": 2634 }, { "epoch": 0.09543643607388627, "grad_norm": 1.9558349234528885, "learning_rate": 1.9776423225905817e-05, "loss": 0.8099, "step": 2635 }, { "epoch": 0.09547265483520463, "grad_norm": 1.8315322100720741, "learning_rate": 1.977617649206746e-05, "loss": 0.938, "step": 2636 }, { "epoch": 0.095508873596523, "grad_norm": 1.9726884877054036, "learning_rate": 1.9775929623700318e-05, "loss": 0.8703, "step": 2637 }, { "epoch": 0.09554509235784137, "grad_norm": 1.748669294012749, "learning_rate": 1.9775682620807796e-05, "loss": 0.8489, "step": 2638 }, { "epoch": 0.09558131111915973, "grad_norm": 1.881524909050554, "learning_rate": 1.9775435483393293e-05, "loss": 0.8734, "step": 2639 }, { "epoch": 0.09561752988047809, "grad_norm": 1.5303646156641786, "learning_rate": 1.9775188211460207e-05, "loss": 0.8414, "step": 2640 }, { "epoch": 0.09565374864179645, "grad_norm": 1.7310134388888794, "learning_rate": 1.9774940805011943e-05, "loss": 0.7464, "step": 2641 }, { "epoch": 0.09568996740311482, "grad_norm": 2.49111483440697, "learning_rate": 1.9774693264051903e-05, "loss": 0.8197, "step": 2642 }, { "epoch": 0.09572618616443318, "grad_norm": 1.7852054167639944, "learning_rate": 1.9774445588583498e-05, "loss": 0.8981, "step": 2643 }, { "epoch": 0.09576240492575154, "grad_norm": 2.356748495821535, "learning_rate": 1.977419777861013e-05, "loss": 0.833, "step": 2644 }, { "epoch": 0.0957986236870699, "grad_norm": 1.8755639728108922, "learning_rate": 1.9773949834135212e-05, "loss": 0.8692, "step": 2645 }, { "epoch": 0.09583484244838826, "grad_norm": 1.5631378058212506, "learning_rate": 1.9773701755162157e-05, "loss": 0.8591, "step": 2646 }, { "epoch": 0.09587106120970663, "grad_norm": 1.5873015309252307, "learning_rate": 1.977345354169438e-05, "loss": 0.8387, "step": 2647 }, { "epoch": 0.095907279971025, "grad_norm": 1.9485731202337273, "learning_rate": 1.9773205193735293e-05, "loss": 0.8976, "step": 2648 }, { "epoch": 0.09594349873234335, "grad_norm": 1.8598316707713969, "learning_rate": 1.9772956711288313e-05, "loss": 0.9397, "step": 2649 }, { "epoch": 0.09597971749366171, "grad_norm": 1.7265450899618617, "learning_rate": 1.9772708094356865e-05, "loss": 0.8955, "step": 2650 }, { "epoch": 0.09601593625498007, "grad_norm": 4.372201218476162, "learning_rate": 1.9772459342944364e-05, "loss": 0.8779, "step": 2651 }, { "epoch": 0.09605215501629845, "grad_norm": 1.387716277572843, "learning_rate": 1.9772210457054236e-05, "loss": 0.8471, "step": 2652 }, { "epoch": 0.09608837377761681, "grad_norm": 1.7461906092616157, "learning_rate": 1.9771961436689906e-05, "loss": 0.855, "step": 2653 }, { "epoch": 0.09612459253893517, "grad_norm": 1.8050870886495294, "learning_rate": 1.97717122818548e-05, "loss": 0.8936, "step": 2654 }, { "epoch": 0.09616081130025353, "grad_norm": 2.2499747076202437, "learning_rate": 1.977146299255235e-05, "loss": 0.9425, "step": 2655 }, { "epoch": 0.09619703006157189, "grad_norm": 2.69081929946053, "learning_rate": 1.977121356878598e-05, "loss": 0.8946, "step": 2656 }, { "epoch": 0.09623324882289026, "grad_norm": 1.6192727408904524, "learning_rate": 1.9770964010559122e-05, "loss": 0.9608, "step": 2657 }, { "epoch": 0.09626946758420862, "grad_norm": 1.8141660278392857, "learning_rate": 1.9770714317875218e-05, "loss": 0.8105, "step": 2658 }, { "epoch": 0.09630568634552698, "grad_norm": 1.4548450195244789, "learning_rate": 1.97704644907377e-05, "loss": 0.8285, "step": 2659 }, { "epoch": 0.09634190510684534, "grad_norm": 1.8706721904090322, "learning_rate": 1.9770214529150004e-05, "loss": 0.9459, "step": 2660 }, { "epoch": 0.0963781238681637, "grad_norm": 2.273126031900549, "learning_rate": 1.9769964433115573e-05, "loss": 0.9523, "step": 2661 }, { "epoch": 0.09641434262948208, "grad_norm": 1.9689878497936695, "learning_rate": 1.9769714202637847e-05, "loss": 0.8724, "step": 2662 }, { "epoch": 0.09645056139080044, "grad_norm": 1.593409014143951, "learning_rate": 1.9769463837720267e-05, "loss": 0.7906, "step": 2663 }, { "epoch": 0.0964867801521188, "grad_norm": 1.8670088750398939, "learning_rate": 1.976921333836628e-05, "loss": 0.8425, "step": 2664 }, { "epoch": 0.09652299891343716, "grad_norm": 1.559457636317794, "learning_rate": 1.9768962704579335e-05, "loss": 0.8556, "step": 2665 }, { "epoch": 0.09655921767475552, "grad_norm": 1.5878123572604186, "learning_rate": 1.9768711936362878e-05, "loss": 0.8311, "step": 2666 }, { "epoch": 0.09659543643607389, "grad_norm": 1.611819324184528, "learning_rate": 1.9768461033720364e-05, "loss": 0.8234, "step": 2667 }, { "epoch": 0.09663165519739225, "grad_norm": 1.9099920003284077, "learning_rate": 1.9768209996655242e-05, "loss": 0.8188, "step": 2668 }, { "epoch": 0.09666787395871061, "grad_norm": 1.9180733658020632, "learning_rate": 1.9767958825170965e-05, "loss": 0.8505, "step": 2669 }, { "epoch": 0.09670409272002897, "grad_norm": 1.794094723733688, "learning_rate": 1.9767707519270993e-05, "loss": 0.8657, "step": 2670 }, { "epoch": 0.09674031148134733, "grad_norm": 1.8596216867596491, "learning_rate": 1.9767456078958785e-05, "loss": 0.8589, "step": 2671 }, { "epoch": 0.0967765302426657, "grad_norm": 1.8327540179773683, "learning_rate": 1.9767204504237796e-05, "loss": 0.8465, "step": 2672 }, { "epoch": 0.09681274900398407, "grad_norm": 2.6606022694382117, "learning_rate": 1.9766952795111492e-05, "loss": 0.8609, "step": 2673 }, { "epoch": 0.09684896776530243, "grad_norm": 1.8016361081750858, "learning_rate": 1.9766700951583334e-05, "loss": 0.8878, "step": 2674 }, { "epoch": 0.09688518652662079, "grad_norm": 1.7076462526017422, "learning_rate": 1.9766448973656793e-05, "loss": 0.8387, "step": 2675 }, { "epoch": 0.09692140528793915, "grad_norm": 1.5893432021217035, "learning_rate": 1.976619686133533e-05, "loss": 0.8534, "step": 2676 }, { "epoch": 0.09695762404925752, "grad_norm": 1.3574973027109647, "learning_rate": 1.9765944614622414e-05, "loss": 0.7827, "step": 2677 }, { "epoch": 0.09699384281057588, "grad_norm": 1.6559510815983658, "learning_rate": 1.9765692233521523e-05, "loss": 0.8548, "step": 2678 }, { "epoch": 0.09703006157189424, "grad_norm": 1.7802062856341736, "learning_rate": 1.9765439718036123e-05, "loss": 0.9515, "step": 2679 }, { "epoch": 0.0970662803332126, "grad_norm": 1.8379069022533143, "learning_rate": 1.9765187068169696e-05, "loss": 0.8633, "step": 2680 }, { "epoch": 0.09710249909453096, "grad_norm": 1.9256081512993057, "learning_rate": 1.9764934283925712e-05, "loss": 0.856, "step": 2681 }, { "epoch": 0.09713871785584934, "grad_norm": 1.677864868986619, "learning_rate": 1.9764681365307653e-05, "loss": 0.8565, "step": 2682 }, { "epoch": 0.0971749366171677, "grad_norm": 1.9594572700221506, "learning_rate": 1.9764428312318995e-05, "loss": 1.006, "step": 2683 }, { "epoch": 0.09721115537848606, "grad_norm": 1.5595891326743305, "learning_rate": 1.9764175124963226e-05, "loss": 0.9962, "step": 2684 }, { "epoch": 0.09724737413980442, "grad_norm": 1.4480681837336518, "learning_rate": 1.9763921803243827e-05, "loss": 0.8099, "step": 2685 }, { "epoch": 0.09728359290112278, "grad_norm": 1.7127529619632857, "learning_rate": 1.9763668347164288e-05, "loss": 0.8632, "step": 2686 }, { "epoch": 0.09731981166244115, "grad_norm": 1.6365132447510453, "learning_rate": 1.976341475672809e-05, "loss": 0.884, "step": 2687 }, { "epoch": 0.09735603042375951, "grad_norm": 1.4294659311449114, "learning_rate": 1.976316103193873e-05, "loss": 0.8296, "step": 2688 }, { "epoch": 0.09739224918507787, "grad_norm": 1.7369584290116307, "learning_rate": 1.976290717279969e-05, "loss": 0.8765, "step": 2689 }, { "epoch": 0.09742846794639623, "grad_norm": 1.4079653100807115, "learning_rate": 1.9762653179314474e-05, "loss": 0.804, "step": 2690 }, { "epoch": 0.09746468670771459, "grad_norm": 1.7330157263128096, "learning_rate": 1.976239905148657e-05, "loss": 0.8874, "step": 2691 }, { "epoch": 0.09750090546903296, "grad_norm": 1.6039882810113195, "learning_rate": 1.9762144789319475e-05, "loss": 0.8911, "step": 2692 }, { "epoch": 0.09753712423035132, "grad_norm": 1.879683823395619, "learning_rate": 1.976189039281669e-05, "loss": 0.8476, "step": 2693 }, { "epoch": 0.09757334299166968, "grad_norm": 1.4663532813487774, "learning_rate": 1.976163586198172e-05, "loss": 0.7662, "step": 2694 }, { "epoch": 0.09760956175298804, "grad_norm": 1.8380207234630588, "learning_rate": 1.9761381196818058e-05, "loss": 0.8647, "step": 2695 }, { "epoch": 0.0976457805143064, "grad_norm": 1.7819347278522162, "learning_rate": 1.9761126397329217e-05, "loss": 0.8704, "step": 2696 }, { "epoch": 0.09768199927562478, "grad_norm": 1.7256292943251665, "learning_rate": 1.9760871463518702e-05, "loss": 0.8772, "step": 2697 }, { "epoch": 0.09771821803694314, "grad_norm": 1.6830930452079191, "learning_rate": 1.976061639539001e-05, "loss": 0.7758, "step": 2698 }, { "epoch": 0.0977544367982615, "grad_norm": 1.926047812720714, "learning_rate": 1.976036119294667e-05, "loss": 0.8852, "step": 2699 }, { "epoch": 0.09779065555957986, "grad_norm": 1.9589653147324526, "learning_rate": 1.9760105856192177e-05, "loss": 0.8481, "step": 2700 }, { "epoch": 0.09782687432089822, "grad_norm": 2.205458443536841, "learning_rate": 1.9759850385130055e-05, "loss": 0.9142, "step": 2701 }, { "epoch": 0.09786309308221659, "grad_norm": 1.5771803991427356, "learning_rate": 1.9759594779763817e-05, "loss": 0.7646, "step": 2702 }, { "epoch": 0.09789931184353495, "grad_norm": 1.6131104293425316, "learning_rate": 1.9759339040096977e-05, "loss": 0.8526, "step": 2703 }, { "epoch": 0.09793553060485331, "grad_norm": 1.6458716134682723, "learning_rate": 1.9759083166133057e-05, "loss": 0.8981, "step": 2704 }, { "epoch": 0.09797174936617167, "grad_norm": 1.8535831859161855, "learning_rate": 1.9758827157875576e-05, "loss": 0.8953, "step": 2705 }, { "epoch": 0.09800796812749003, "grad_norm": 1.2927076231441161, "learning_rate": 1.9758571015328062e-05, "loss": 0.7994, "step": 2706 }, { "epoch": 0.09804418688880841, "grad_norm": 1.7567078054372687, "learning_rate": 1.9758314738494032e-05, "loss": 0.8986, "step": 2707 }, { "epoch": 0.09808040565012677, "grad_norm": 1.7203558485856423, "learning_rate": 1.975805832737702e-05, "loss": 0.8752, "step": 2708 }, { "epoch": 0.09811662441144513, "grad_norm": 1.5904548838841355, "learning_rate": 1.975780178198055e-05, "loss": 0.8929, "step": 2709 }, { "epoch": 0.09815284317276349, "grad_norm": 1.535909331496111, "learning_rate": 1.9757545102308154e-05, "loss": 0.9013, "step": 2710 }, { "epoch": 0.09818906193408185, "grad_norm": 1.312795939885876, "learning_rate": 1.9757288288363363e-05, "loss": 0.7819, "step": 2711 }, { "epoch": 0.09822528069540022, "grad_norm": 1.7169246571781793, "learning_rate": 1.975703134014971e-05, "loss": 0.808, "step": 2712 }, { "epoch": 0.09826149945671858, "grad_norm": 1.6459260199700343, "learning_rate": 1.9756774257670735e-05, "loss": 0.7658, "step": 2713 }, { "epoch": 0.09829771821803694, "grad_norm": 1.4133436041240908, "learning_rate": 1.9756517040929973e-05, "loss": 0.7743, "step": 2714 }, { "epoch": 0.0983339369793553, "grad_norm": 1.650840835561062, "learning_rate": 1.9756259689930965e-05, "loss": 0.7966, "step": 2715 }, { "epoch": 0.09837015574067366, "grad_norm": 1.920467244921172, "learning_rate": 1.9756002204677245e-05, "loss": 0.7795, "step": 2716 }, { "epoch": 0.09840637450199204, "grad_norm": 1.496469066986427, "learning_rate": 1.975574458517237e-05, "loss": 0.8199, "step": 2717 }, { "epoch": 0.0984425932633104, "grad_norm": 1.8545889183878943, "learning_rate": 1.9755486831419876e-05, "loss": 0.8893, "step": 2718 }, { "epoch": 0.09847881202462876, "grad_norm": 1.7030108242960706, "learning_rate": 1.9755228943423308e-05, "loss": 0.8615, "step": 2719 }, { "epoch": 0.09851503078594712, "grad_norm": 1.5810163221128217, "learning_rate": 1.9754970921186217e-05, "loss": 0.8183, "step": 2720 }, { "epoch": 0.09855124954726549, "grad_norm": 1.3853362412434373, "learning_rate": 1.9754712764712158e-05, "loss": 0.8079, "step": 2721 }, { "epoch": 0.09858746830858385, "grad_norm": 1.7558839528927632, "learning_rate": 1.9754454474004682e-05, "loss": 0.937, "step": 2722 }, { "epoch": 0.09862368706990221, "grad_norm": 1.7174710446194226, "learning_rate": 1.975419604906734e-05, "loss": 0.8741, "step": 2723 }, { "epoch": 0.09865990583122057, "grad_norm": 1.7660369730002732, "learning_rate": 1.9753937489903686e-05, "loss": 0.9118, "step": 2724 }, { "epoch": 0.09869612459253893, "grad_norm": 1.714649579388553, "learning_rate": 1.975367879651728e-05, "loss": 0.8274, "step": 2725 }, { "epoch": 0.0987323433538573, "grad_norm": 1.5386913270730342, "learning_rate": 1.975341996891169e-05, "loss": 0.8507, "step": 2726 }, { "epoch": 0.09876856211517566, "grad_norm": 1.334448766590499, "learning_rate": 1.9753161007090468e-05, "loss": 0.8097, "step": 2727 }, { "epoch": 0.09880478087649402, "grad_norm": 1.6923487522816296, "learning_rate": 1.975290191105718e-05, "loss": 0.8595, "step": 2728 }, { "epoch": 0.09884099963781239, "grad_norm": 1.6898311732243456, "learning_rate": 1.9752642680815396e-05, "loss": 0.9414, "step": 2729 }, { "epoch": 0.09887721839913075, "grad_norm": 1.7529829755668809, "learning_rate": 1.9752383316368674e-05, "loss": 0.9011, "step": 2730 }, { "epoch": 0.09891343716044912, "grad_norm": 1.6717587245633212, "learning_rate": 1.975212381772059e-05, "loss": 0.8812, "step": 2731 }, { "epoch": 0.09894965592176748, "grad_norm": 1.7871637096379083, "learning_rate": 1.9751864184874713e-05, "loss": 0.9574, "step": 2732 }, { "epoch": 0.09898587468308584, "grad_norm": 1.6543190007717437, "learning_rate": 1.9751604417834617e-05, "loss": 0.862, "step": 2733 }, { "epoch": 0.0990220934444042, "grad_norm": 1.743728371113457, "learning_rate": 1.9751344516603875e-05, "loss": 0.9416, "step": 2734 }, { "epoch": 0.09905831220572256, "grad_norm": 1.5927021588363202, "learning_rate": 1.9751084481186066e-05, "loss": 0.7827, "step": 2735 }, { "epoch": 0.09909453096704093, "grad_norm": 1.6012591307280732, "learning_rate": 1.9750824311584766e-05, "loss": 0.8409, "step": 2736 }, { "epoch": 0.0991307497283593, "grad_norm": 1.7055697830127452, "learning_rate": 1.9750564007803554e-05, "loss": 0.9375, "step": 2737 }, { "epoch": 0.09916696848967765, "grad_norm": 1.7302222948259365, "learning_rate": 1.9750303569846014e-05, "loss": 0.8812, "step": 2738 }, { "epoch": 0.09920318725099601, "grad_norm": 1.7743653357391427, "learning_rate": 1.975004299771573e-05, "loss": 0.9184, "step": 2739 }, { "epoch": 0.09923940601231437, "grad_norm": 1.7411091443664781, "learning_rate": 1.9749782291416287e-05, "loss": 0.858, "step": 2740 }, { "epoch": 0.09927562477363275, "grad_norm": 1.4106025496029528, "learning_rate": 1.9749521450951277e-05, "loss": 0.7932, "step": 2741 }, { "epoch": 0.09931184353495111, "grad_norm": 1.7669325624426437, "learning_rate": 1.974926047632428e-05, "loss": 0.8865, "step": 2742 }, { "epoch": 0.09934806229626947, "grad_norm": 1.4185006150560415, "learning_rate": 1.9748999367538896e-05, "loss": 0.7941, "step": 2743 }, { "epoch": 0.09938428105758783, "grad_norm": 1.849277414181873, "learning_rate": 1.974873812459871e-05, "loss": 0.9545, "step": 2744 }, { "epoch": 0.09942049981890619, "grad_norm": 1.2156932513550824, "learning_rate": 1.9748476747507325e-05, "loss": 0.7695, "step": 2745 }, { "epoch": 0.09945671858022456, "grad_norm": 1.3385756820099635, "learning_rate": 1.974821523626833e-05, "loss": 0.7903, "step": 2746 }, { "epoch": 0.09949293734154292, "grad_norm": 1.6716115676593473, "learning_rate": 1.974795359088533e-05, "loss": 0.8459, "step": 2747 }, { "epoch": 0.09952915610286128, "grad_norm": 1.7880870750336124, "learning_rate": 1.9747691811361926e-05, "loss": 0.9123, "step": 2748 }, { "epoch": 0.09956537486417964, "grad_norm": 1.6098370860881108, "learning_rate": 1.9747429897701712e-05, "loss": 0.8483, "step": 2749 }, { "epoch": 0.099601593625498, "grad_norm": 1.71242844708825, "learning_rate": 1.9747167849908305e-05, "loss": 0.8621, "step": 2750 }, { "epoch": 0.09963781238681638, "grad_norm": 1.4617361036539218, "learning_rate": 1.97469056679853e-05, "loss": 0.825, "step": 2751 }, { "epoch": 0.09967403114813474, "grad_norm": 1.578205805305504, "learning_rate": 1.9746643351936307e-05, "loss": 0.8431, "step": 2752 }, { "epoch": 0.0997102499094531, "grad_norm": 2.118101394310167, "learning_rate": 1.974638090176494e-05, "loss": 0.9046, "step": 2753 }, { "epoch": 0.09974646867077146, "grad_norm": 1.5835756046072347, "learning_rate": 1.9746118317474806e-05, "loss": 0.8186, "step": 2754 }, { "epoch": 0.09978268743208982, "grad_norm": 1.5849050933909186, "learning_rate": 1.974585559906952e-05, "loss": 0.8221, "step": 2755 }, { "epoch": 0.09981890619340819, "grad_norm": 1.6135366043760886, "learning_rate": 1.97455927465527e-05, "loss": 0.9108, "step": 2756 }, { "epoch": 0.09985512495472655, "grad_norm": 1.6124729505384534, "learning_rate": 1.974532975992796e-05, "loss": 0.8868, "step": 2757 }, { "epoch": 0.09989134371604491, "grad_norm": 1.4724615111212644, "learning_rate": 1.9745066639198918e-05, "loss": 0.7589, "step": 2758 }, { "epoch": 0.09992756247736327, "grad_norm": 1.4691354502300502, "learning_rate": 1.97448033843692e-05, "loss": 0.8081, "step": 2759 }, { "epoch": 0.09996378123868163, "grad_norm": 1.7575315971572902, "learning_rate": 1.9744539995442417e-05, "loss": 0.8803, "step": 2760 }, { "epoch": 0.1, "grad_norm": 1.3798261713172375, "learning_rate": 1.974427647242221e-05, "loss": 0.7827, "step": 2761 }, { "epoch": 0.10003621876131837, "grad_norm": 1.6314074115148025, "learning_rate": 1.9744012815312192e-05, "loss": 0.788, "step": 2762 }, { "epoch": 0.10007243752263673, "grad_norm": 1.726073626041985, "learning_rate": 1.9743749024115993e-05, "loss": 0.8017, "step": 2763 }, { "epoch": 0.10010865628395509, "grad_norm": 1.7410627538470165, "learning_rate": 1.9743485098837253e-05, "loss": 0.869, "step": 2764 }, { "epoch": 0.10014487504527345, "grad_norm": 1.8527005358510575, "learning_rate": 1.974322103947959e-05, "loss": 0.791, "step": 2765 }, { "epoch": 0.10018109380659182, "grad_norm": 1.6143453335445312, "learning_rate": 1.974295684604665e-05, "loss": 0.88, "step": 2766 }, { "epoch": 0.10021731256791018, "grad_norm": 1.7076633666591186, "learning_rate": 1.9742692518542062e-05, "loss": 0.8149, "step": 2767 }, { "epoch": 0.10025353132922854, "grad_norm": 1.8229295189722305, "learning_rate": 1.974242805696946e-05, "loss": 0.8098, "step": 2768 }, { "epoch": 0.1002897500905469, "grad_norm": 1.8709272836209043, "learning_rate": 1.974216346133249e-05, "loss": 0.9168, "step": 2769 }, { "epoch": 0.10032596885186526, "grad_norm": 1.6429046834679497, "learning_rate": 1.974189873163479e-05, "loss": 0.8135, "step": 2770 }, { "epoch": 0.10036218761318363, "grad_norm": 1.5907178252139358, "learning_rate": 1.9741633867880005e-05, "loss": 0.8697, "step": 2771 }, { "epoch": 0.100398406374502, "grad_norm": 1.4825397927990516, "learning_rate": 1.974136887007178e-05, "loss": 0.7934, "step": 2772 }, { "epoch": 0.10043462513582035, "grad_norm": 1.7729660743578215, "learning_rate": 1.9741103738213755e-05, "loss": 0.9016, "step": 2773 }, { "epoch": 0.10047084389713871, "grad_norm": 1.7786666212407702, "learning_rate": 1.9740838472309586e-05, "loss": 0.8655, "step": 2774 }, { "epoch": 0.10050706265845707, "grad_norm": 1.7165369980662757, "learning_rate": 1.974057307236292e-05, "loss": 0.9518, "step": 2775 }, { "epoch": 0.10054328141977545, "grad_norm": 1.6452832665884902, "learning_rate": 1.974030753837741e-05, "loss": 0.7716, "step": 2776 }, { "epoch": 0.10057950018109381, "grad_norm": 2.311332446557344, "learning_rate": 1.974004187035671e-05, "loss": 0.83, "step": 2777 }, { "epoch": 0.10061571894241217, "grad_norm": 1.6256620807499509, "learning_rate": 1.9739776068304474e-05, "loss": 0.8547, "step": 2778 }, { "epoch": 0.10065193770373053, "grad_norm": 1.8007670234399555, "learning_rate": 1.973951013222436e-05, "loss": 0.8678, "step": 2779 }, { "epoch": 0.10068815646504889, "grad_norm": 1.5502853916013082, "learning_rate": 1.9739244062120037e-05, "loss": 0.8786, "step": 2780 }, { "epoch": 0.10072437522636726, "grad_norm": 2.321015495514207, "learning_rate": 1.973897785799515e-05, "loss": 0.8322, "step": 2781 }, { "epoch": 0.10076059398768562, "grad_norm": 1.5769131629648534, "learning_rate": 1.9738711519853374e-05, "loss": 0.8277, "step": 2782 }, { "epoch": 0.10079681274900398, "grad_norm": 1.7736318645594482, "learning_rate": 1.9738445047698364e-05, "loss": 0.8831, "step": 2783 }, { "epoch": 0.10083303151032234, "grad_norm": 1.3027949716759821, "learning_rate": 1.97381784415338e-05, "loss": 0.7688, "step": 2784 }, { "epoch": 0.1008692502716407, "grad_norm": 1.604945786777139, "learning_rate": 1.973791170136334e-05, "loss": 0.8448, "step": 2785 }, { "epoch": 0.10090546903295908, "grad_norm": 1.7919833315596694, "learning_rate": 1.9737644827190664e-05, "loss": 0.9337, "step": 2786 }, { "epoch": 0.10094168779427744, "grad_norm": 1.673994321484239, "learning_rate": 1.9737377819019432e-05, "loss": 0.8561, "step": 2787 }, { "epoch": 0.1009779065555958, "grad_norm": 1.6528874195008474, "learning_rate": 1.9737110676853327e-05, "loss": 0.8176, "step": 2788 }, { "epoch": 0.10101412531691416, "grad_norm": 1.57355860743155, "learning_rate": 1.9736843400696025e-05, "loss": 0.8136, "step": 2789 }, { "epoch": 0.10105034407823252, "grad_norm": 1.7436962234329862, "learning_rate": 1.9736575990551203e-05, "loss": 0.8517, "step": 2790 }, { "epoch": 0.10108656283955089, "grad_norm": 1.8657576071753303, "learning_rate": 1.973630844642254e-05, "loss": 0.934, "step": 2791 }, { "epoch": 0.10112278160086925, "grad_norm": 1.9700716725342202, "learning_rate": 1.9736040768313715e-05, "loss": 0.9644, "step": 2792 }, { "epoch": 0.10115900036218761, "grad_norm": 1.65591662831686, "learning_rate": 1.9735772956228415e-05, "loss": 0.8538, "step": 2793 }, { "epoch": 0.10119521912350597, "grad_norm": 1.5232733356838777, "learning_rate": 1.973550501017033e-05, "loss": 0.8379, "step": 2794 }, { "epoch": 0.10123143788482433, "grad_norm": 1.2385607540857784, "learning_rate": 1.9735236930143134e-05, "loss": 0.7949, "step": 2795 }, { "epoch": 0.1012676566461427, "grad_norm": 1.8332815174254422, "learning_rate": 1.9734968716150527e-05, "loss": 0.7873, "step": 2796 }, { "epoch": 0.10130387540746107, "grad_norm": 1.603517617144259, "learning_rate": 1.97347003681962e-05, "loss": 0.8902, "step": 2797 }, { "epoch": 0.10134009416877943, "grad_norm": 1.605589563344172, "learning_rate": 1.9734431886283834e-05, "loss": 0.8498, "step": 2798 }, { "epoch": 0.10137631293009779, "grad_norm": 1.983546711175022, "learning_rate": 1.9734163270417136e-05, "loss": 0.9018, "step": 2799 }, { "epoch": 0.10141253169141615, "grad_norm": 2.065068192106517, "learning_rate": 1.97338945205998e-05, "loss": 0.8097, "step": 2800 }, { "epoch": 0.10144875045273452, "grad_norm": 1.5586300716613626, "learning_rate": 1.9733625636835523e-05, "loss": 0.8095, "step": 2801 }, { "epoch": 0.10148496921405288, "grad_norm": 1.469591912726614, "learning_rate": 1.9733356619128002e-05, "loss": 0.8168, "step": 2802 }, { "epoch": 0.10152118797537124, "grad_norm": 1.5648689094033545, "learning_rate": 1.973308746748094e-05, "loss": 0.7799, "step": 2803 }, { "epoch": 0.1015574067366896, "grad_norm": 1.5417708225334867, "learning_rate": 1.9732818181898046e-05, "loss": 0.8084, "step": 2804 }, { "epoch": 0.10159362549800798, "grad_norm": 1.6169696168122558, "learning_rate": 1.9732548762383018e-05, "loss": 0.8928, "step": 2805 }, { "epoch": 0.10162984425932634, "grad_norm": 1.723671566362729, "learning_rate": 1.9732279208939572e-05, "loss": 0.8508, "step": 2806 }, { "epoch": 0.1016660630206447, "grad_norm": 1.6676981000618074, "learning_rate": 1.9732009521571408e-05, "loss": 0.8772, "step": 2807 }, { "epoch": 0.10170228178196306, "grad_norm": 1.69582310032237, "learning_rate": 1.9731739700282246e-05, "loss": 0.8753, "step": 2808 }, { "epoch": 0.10173850054328142, "grad_norm": 1.7694473055902966, "learning_rate": 1.9731469745075792e-05, "loss": 0.7487, "step": 2809 }, { "epoch": 0.10177471930459979, "grad_norm": 1.7130618493723537, "learning_rate": 1.973119965595576e-05, "loss": 0.9033, "step": 2810 }, { "epoch": 0.10181093806591815, "grad_norm": 1.7052210456228263, "learning_rate": 1.9730929432925875e-05, "loss": 0.8585, "step": 2811 }, { "epoch": 0.10184715682723651, "grad_norm": 2.461254517254188, "learning_rate": 1.9730659075989852e-05, "loss": 0.8432, "step": 2812 }, { "epoch": 0.10188337558855487, "grad_norm": 1.4504074663976732, "learning_rate": 1.9730388585151408e-05, "loss": 0.8213, "step": 2813 }, { "epoch": 0.10191959434987323, "grad_norm": 1.7560307055297288, "learning_rate": 1.9730117960414266e-05, "loss": 0.8951, "step": 2814 }, { "epoch": 0.1019558131111916, "grad_norm": 1.7036542376391908, "learning_rate": 1.9729847201782154e-05, "loss": 0.8227, "step": 2815 }, { "epoch": 0.10199203187250996, "grad_norm": 1.9151961492201168, "learning_rate": 1.9729576309258793e-05, "loss": 0.9004, "step": 2816 }, { "epoch": 0.10202825063382832, "grad_norm": 1.8930358628432917, "learning_rate": 1.972930528284791e-05, "loss": 0.9395, "step": 2817 }, { "epoch": 0.10206446939514668, "grad_norm": 1.5218091082273406, "learning_rate": 1.972903412255324e-05, "loss": 0.7804, "step": 2818 }, { "epoch": 0.10210068815646504, "grad_norm": 2.0304365987019217, "learning_rate": 1.9728762828378513e-05, "loss": 0.9012, "step": 2819 }, { "epoch": 0.10213690691778342, "grad_norm": 1.4387220913168106, "learning_rate": 1.972849140032746e-05, "loss": 0.8054, "step": 2820 }, { "epoch": 0.10217312567910178, "grad_norm": 1.8145733568661457, "learning_rate": 1.972821983840382e-05, "loss": 0.9299, "step": 2821 }, { "epoch": 0.10220934444042014, "grad_norm": 1.6892538559543455, "learning_rate": 1.972794814261132e-05, "loss": 0.8327, "step": 2822 }, { "epoch": 0.1022455632017385, "grad_norm": 1.795942435625369, "learning_rate": 1.9727676312953716e-05, "loss": 0.9218, "step": 2823 }, { "epoch": 0.10228178196305686, "grad_norm": 1.8018023560349692, "learning_rate": 1.972740434943473e-05, "loss": 0.914, "step": 2824 }, { "epoch": 0.10231800072437523, "grad_norm": 1.8693770176434383, "learning_rate": 1.9727132252058113e-05, "loss": 0.9071, "step": 2825 }, { "epoch": 0.10235421948569359, "grad_norm": 1.7636345503481095, "learning_rate": 1.9726860020827615e-05, "loss": 0.9177, "step": 2826 }, { "epoch": 0.10239043824701195, "grad_norm": 1.8597475147547362, "learning_rate": 1.972658765574697e-05, "loss": 0.881, "step": 2827 }, { "epoch": 0.10242665700833031, "grad_norm": 1.7434224084206758, "learning_rate": 1.9726315156819936e-05, "loss": 0.9054, "step": 2828 }, { "epoch": 0.10246287576964867, "grad_norm": 1.9928543685680076, "learning_rate": 1.972604252405026e-05, "loss": 0.8558, "step": 2829 }, { "epoch": 0.10249909453096705, "grad_norm": 1.6969843105634477, "learning_rate": 1.972576975744169e-05, "loss": 0.8808, "step": 2830 }, { "epoch": 0.10253531329228541, "grad_norm": 1.3693375294523353, "learning_rate": 1.9725496856997987e-05, "loss": 0.7834, "step": 2831 }, { "epoch": 0.10257153205360377, "grad_norm": 1.8440750187163022, "learning_rate": 1.9725223822722897e-05, "loss": 0.9388, "step": 2832 }, { "epoch": 0.10260775081492213, "grad_norm": 1.6633663355510722, "learning_rate": 1.9724950654620182e-05, "loss": 0.8526, "step": 2833 }, { "epoch": 0.10264396957624049, "grad_norm": 1.2638373926152704, "learning_rate": 1.97246773526936e-05, "loss": 0.8502, "step": 2834 }, { "epoch": 0.10268018833755886, "grad_norm": 1.9239553380432384, "learning_rate": 1.9724403916946917e-05, "loss": 0.8878, "step": 2835 }, { "epoch": 0.10271640709887722, "grad_norm": 1.6341011575258775, "learning_rate": 1.972413034738389e-05, "loss": 0.7984, "step": 2836 }, { "epoch": 0.10275262586019558, "grad_norm": 1.240883723864413, "learning_rate": 1.9723856644008283e-05, "loss": 0.7824, "step": 2837 }, { "epoch": 0.10278884462151394, "grad_norm": 1.6334339640648248, "learning_rate": 1.9723582806823866e-05, "loss": 0.8369, "step": 2838 }, { "epoch": 0.1028250633828323, "grad_norm": 2.4898859402975386, "learning_rate": 1.9723308835834406e-05, "loss": 0.8323, "step": 2839 }, { "epoch": 0.10286128214415068, "grad_norm": 1.7831980860828014, "learning_rate": 1.9723034731043674e-05, "loss": 0.7518, "step": 2840 }, { "epoch": 0.10289750090546904, "grad_norm": 1.8611519100679776, "learning_rate": 1.972276049245544e-05, "loss": 0.8138, "step": 2841 }, { "epoch": 0.1029337196667874, "grad_norm": 1.8014652521233565, "learning_rate": 1.9722486120073478e-05, "loss": 0.8605, "step": 2842 }, { "epoch": 0.10296993842810576, "grad_norm": 1.727278244962489, "learning_rate": 1.9722211613901564e-05, "loss": 0.8751, "step": 2843 }, { "epoch": 0.10300615718942412, "grad_norm": 1.6573526000704306, "learning_rate": 1.9721936973943474e-05, "loss": 0.9214, "step": 2844 }, { "epoch": 0.10304237595074249, "grad_norm": 1.2794980189482128, "learning_rate": 1.972166220020299e-05, "loss": 0.7577, "step": 2845 }, { "epoch": 0.10307859471206085, "grad_norm": 1.438633248062822, "learning_rate": 1.972138729268389e-05, "loss": 0.7896, "step": 2846 }, { "epoch": 0.10311481347337921, "grad_norm": 1.1355961266380012, "learning_rate": 1.9721112251389963e-05, "loss": 0.7546, "step": 2847 }, { "epoch": 0.10315103223469757, "grad_norm": 2.0908747206713, "learning_rate": 1.9720837076324984e-05, "loss": 0.8507, "step": 2848 }, { "epoch": 0.10318725099601593, "grad_norm": 1.5080101236978583, "learning_rate": 1.972056176749275e-05, "loss": 0.8702, "step": 2849 }, { "epoch": 0.1032234697573343, "grad_norm": 1.7704033434177742, "learning_rate": 1.9720286324897048e-05, "loss": 0.8277, "step": 2850 }, { "epoch": 0.10325968851865266, "grad_norm": 1.7079071508181898, "learning_rate": 1.972001074854166e-05, "loss": 0.8526, "step": 2851 }, { "epoch": 0.10329590727997102, "grad_norm": 1.833571128478122, "learning_rate": 1.9719735038430382e-05, "loss": 0.8753, "step": 2852 }, { "epoch": 0.10333212604128938, "grad_norm": 1.859910258818779, "learning_rate": 1.9719459194567013e-05, "loss": 0.8885, "step": 2853 }, { "epoch": 0.10336834480260775, "grad_norm": 1.8347667065479778, "learning_rate": 1.9719183216955346e-05, "loss": 0.8396, "step": 2854 }, { "epoch": 0.10340456356392612, "grad_norm": 1.2950092938131204, "learning_rate": 1.971890710559917e-05, "loss": 0.8213, "step": 2855 }, { "epoch": 0.10344078232524448, "grad_norm": 1.656080074185453, "learning_rate": 1.9718630860502303e-05, "loss": 0.8703, "step": 2856 }, { "epoch": 0.10347700108656284, "grad_norm": 1.686864258246391, "learning_rate": 1.971835448166853e-05, "loss": 0.8076, "step": 2857 }, { "epoch": 0.1035132198478812, "grad_norm": 1.773687311466653, "learning_rate": 1.971807796910166e-05, "loss": 0.9013, "step": 2858 }, { "epoch": 0.10354943860919956, "grad_norm": 1.7957650104469842, "learning_rate": 1.97178013228055e-05, "loss": 0.9193, "step": 2859 }, { "epoch": 0.10358565737051793, "grad_norm": 1.2418376486925071, "learning_rate": 1.9717524542783854e-05, "loss": 0.7617, "step": 2860 }, { "epoch": 0.1036218761318363, "grad_norm": 1.7406979659417392, "learning_rate": 1.971724762904053e-05, "loss": 0.8528, "step": 2861 }, { "epoch": 0.10365809489315465, "grad_norm": 1.5269654786065687, "learning_rate": 1.9716970581579343e-05, "loss": 0.8512, "step": 2862 }, { "epoch": 0.10369431365447301, "grad_norm": 1.756128869681772, "learning_rate": 1.97166934004041e-05, "loss": 0.864, "step": 2863 }, { "epoch": 0.10373053241579137, "grad_norm": 1.6705169487485585, "learning_rate": 1.9716416085518622e-05, "loss": 0.8118, "step": 2864 }, { "epoch": 0.10376675117710975, "grad_norm": 1.7242092670720905, "learning_rate": 1.971613863692672e-05, "loss": 0.8962, "step": 2865 }, { "epoch": 0.10380296993842811, "grad_norm": 1.5784834664723941, "learning_rate": 1.971586105463221e-05, "loss": 0.8724, "step": 2866 }, { "epoch": 0.10383918869974647, "grad_norm": 2.111625967975878, "learning_rate": 1.971558333863892e-05, "loss": 0.8383, "step": 2867 }, { "epoch": 0.10387540746106483, "grad_norm": 1.6612514866100374, "learning_rate": 1.971530548895066e-05, "loss": 0.8798, "step": 2868 }, { "epoch": 0.10391162622238319, "grad_norm": 1.691460953239725, "learning_rate": 1.9715027505571266e-05, "loss": 0.881, "step": 2869 }, { "epoch": 0.10394784498370156, "grad_norm": 1.597970194563227, "learning_rate": 1.971474938850455e-05, "loss": 0.7694, "step": 2870 }, { "epoch": 0.10398406374501992, "grad_norm": 1.7946747639838672, "learning_rate": 1.971447113775435e-05, "loss": 0.9529, "step": 2871 }, { "epoch": 0.10402028250633828, "grad_norm": 1.6354252630638555, "learning_rate": 1.9714192753324496e-05, "loss": 0.8859, "step": 2872 }, { "epoch": 0.10405650126765664, "grad_norm": 1.6433388324049334, "learning_rate": 1.9713914235218806e-05, "loss": 0.8059, "step": 2873 }, { "epoch": 0.104092720028975, "grad_norm": 1.329528532600383, "learning_rate": 1.9713635583441125e-05, "loss": 0.749, "step": 2874 }, { "epoch": 0.10412893879029338, "grad_norm": 1.3010634543616317, "learning_rate": 1.9713356797995282e-05, "loss": 0.8202, "step": 2875 }, { "epoch": 0.10416515755161174, "grad_norm": 1.6862185721193965, "learning_rate": 1.9713077878885116e-05, "loss": 0.8433, "step": 2876 }, { "epoch": 0.1042013763129301, "grad_norm": 1.812118507595777, "learning_rate": 1.9712798826114462e-05, "loss": 0.8313, "step": 2877 }, { "epoch": 0.10423759507424846, "grad_norm": 1.6915040769764693, "learning_rate": 1.9712519639687166e-05, "loss": 0.912, "step": 2878 }, { "epoch": 0.10427381383556682, "grad_norm": 1.8022807987895106, "learning_rate": 1.9712240319607063e-05, "loss": 0.9258, "step": 2879 }, { "epoch": 0.10431003259688519, "grad_norm": 1.4760606060353105, "learning_rate": 1.9711960865877998e-05, "loss": 0.7856, "step": 2880 }, { "epoch": 0.10434625135820355, "grad_norm": 1.6006691709130128, "learning_rate": 1.9711681278503817e-05, "loss": 0.7966, "step": 2881 }, { "epoch": 0.10438247011952191, "grad_norm": 1.6809910894387667, "learning_rate": 1.9711401557488366e-05, "loss": 0.8407, "step": 2882 }, { "epoch": 0.10441868888084027, "grad_norm": 1.6217284350423466, "learning_rate": 1.9711121702835504e-05, "loss": 0.8284, "step": 2883 }, { "epoch": 0.10445490764215863, "grad_norm": 1.8859986783212768, "learning_rate": 1.971084171454907e-05, "loss": 0.9558, "step": 2884 }, { "epoch": 0.104491126403477, "grad_norm": 2.019992827270539, "learning_rate": 1.971056159263292e-05, "loss": 0.8748, "step": 2885 }, { "epoch": 0.10452734516479537, "grad_norm": 1.5315845454609214, "learning_rate": 1.971028133709091e-05, "loss": 0.8823, "step": 2886 }, { "epoch": 0.10456356392611373, "grad_norm": 1.5526703215528899, "learning_rate": 1.97100009479269e-05, "loss": 0.8201, "step": 2887 }, { "epoch": 0.10459978268743209, "grad_norm": 1.6881499363407313, "learning_rate": 1.9709720425144745e-05, "loss": 0.8517, "step": 2888 }, { "epoch": 0.10463600144875045, "grad_norm": 1.7680849391974134, "learning_rate": 1.97094397687483e-05, "loss": 0.8098, "step": 2889 }, { "epoch": 0.10467222021006882, "grad_norm": 1.8362031521251634, "learning_rate": 1.9709158978741437e-05, "loss": 0.9241, "step": 2890 }, { "epoch": 0.10470843897138718, "grad_norm": 1.6763975181800177, "learning_rate": 1.9708878055128015e-05, "loss": 0.9062, "step": 2891 }, { "epoch": 0.10474465773270554, "grad_norm": 1.6563681468120142, "learning_rate": 1.97085969979119e-05, "loss": 0.8564, "step": 2892 }, { "epoch": 0.1047808764940239, "grad_norm": 1.70855602236705, "learning_rate": 1.970831580709696e-05, "loss": 0.8641, "step": 2893 }, { "epoch": 0.10481709525534227, "grad_norm": 1.7630186318046162, "learning_rate": 1.9708034482687063e-05, "loss": 0.8834, "step": 2894 }, { "epoch": 0.10485331401666063, "grad_norm": 1.5692675675915138, "learning_rate": 1.9707753024686077e-05, "loss": 0.8432, "step": 2895 }, { "epoch": 0.104889532777979, "grad_norm": 1.6265320043975635, "learning_rate": 1.9707471433097888e-05, "loss": 0.8439, "step": 2896 }, { "epoch": 0.10492575153929735, "grad_norm": 1.7073641306255158, "learning_rate": 1.9707189707926356e-05, "loss": 0.8039, "step": 2897 }, { "epoch": 0.10496197030061571, "grad_norm": 1.636104315912256, "learning_rate": 1.9706907849175366e-05, "loss": 0.9434, "step": 2898 }, { "epoch": 0.10499818906193409, "grad_norm": 1.7042520194322526, "learning_rate": 1.9706625856848792e-05, "loss": 0.8128, "step": 2899 }, { "epoch": 0.10503440782325245, "grad_norm": 1.783470994398527, "learning_rate": 1.9706343730950523e-05, "loss": 0.8186, "step": 2900 }, { "epoch": 0.10507062658457081, "grad_norm": 1.7336685335585136, "learning_rate": 1.970606147148443e-05, "loss": 0.8833, "step": 2901 }, { "epoch": 0.10510684534588917, "grad_norm": 1.4082259446310978, "learning_rate": 1.9705779078454407e-05, "loss": 0.7546, "step": 2902 }, { "epoch": 0.10514306410720753, "grad_norm": 1.6134436874263618, "learning_rate": 1.9705496551864332e-05, "loss": 0.8159, "step": 2903 }, { "epoch": 0.1051792828685259, "grad_norm": 1.7664106243691826, "learning_rate": 1.97052138917181e-05, "loss": 0.8756, "step": 2904 }, { "epoch": 0.10521550162984426, "grad_norm": 1.7766644374368294, "learning_rate": 1.9704931098019594e-05, "loss": 0.9462, "step": 2905 }, { "epoch": 0.10525172039116262, "grad_norm": 1.618276678481997, "learning_rate": 1.970464817077271e-05, "loss": 0.8525, "step": 2906 }, { "epoch": 0.10528793915248098, "grad_norm": 1.4347665685283966, "learning_rate": 1.970436510998134e-05, "loss": 0.7757, "step": 2907 }, { "epoch": 0.10532415791379934, "grad_norm": 1.423039877152076, "learning_rate": 1.970408191564938e-05, "loss": 0.775, "step": 2908 }, { "epoch": 0.10536037667511772, "grad_norm": 1.3262969999194794, "learning_rate": 1.9703798587780723e-05, "loss": 0.7735, "step": 2909 }, { "epoch": 0.10539659543643608, "grad_norm": 1.8477249616625322, "learning_rate": 1.9703515126379274e-05, "loss": 0.7893, "step": 2910 }, { "epoch": 0.10543281419775444, "grad_norm": 1.29418466243638, "learning_rate": 1.9703231531448928e-05, "loss": 0.8609, "step": 2911 }, { "epoch": 0.1054690329590728, "grad_norm": 1.6560853232982649, "learning_rate": 1.9702947802993596e-05, "loss": 0.9533, "step": 2912 }, { "epoch": 0.10550525172039116, "grad_norm": 1.497517887752557, "learning_rate": 1.9702663941017167e-05, "loss": 0.794, "step": 2913 }, { "epoch": 0.10554147048170953, "grad_norm": 1.6650635323166703, "learning_rate": 1.9702379945523565e-05, "loss": 0.8686, "step": 2914 }, { "epoch": 0.10557768924302789, "grad_norm": 2.019592114533016, "learning_rate": 1.9702095816516687e-05, "loss": 0.8721, "step": 2915 }, { "epoch": 0.10561390800434625, "grad_norm": 1.563842996167447, "learning_rate": 1.9701811554000444e-05, "loss": 0.804, "step": 2916 }, { "epoch": 0.10565012676566461, "grad_norm": 1.5265185207290788, "learning_rate": 1.970152715797875e-05, "loss": 0.8226, "step": 2917 }, { "epoch": 0.10568634552698297, "grad_norm": 1.8105775301744411, "learning_rate": 1.9701242628455518e-05, "loss": 0.9512, "step": 2918 }, { "epoch": 0.10572256428830135, "grad_norm": 1.6808003708901762, "learning_rate": 1.9700957965434664e-05, "loss": 0.8066, "step": 2919 }, { "epoch": 0.1057587830496197, "grad_norm": 1.292284662229962, "learning_rate": 1.9700673168920103e-05, "loss": 0.7996, "step": 2920 }, { "epoch": 0.10579500181093807, "grad_norm": 1.68245678987891, "learning_rate": 1.970038823891576e-05, "loss": 0.8711, "step": 2921 }, { "epoch": 0.10583122057225643, "grad_norm": 1.8170242119765252, "learning_rate": 1.9700103175425544e-05, "loss": 0.8886, "step": 2922 }, { "epoch": 0.10586743933357479, "grad_norm": 1.6031752421405066, "learning_rate": 1.969981797845339e-05, "loss": 0.8449, "step": 2923 }, { "epoch": 0.10590365809489316, "grad_norm": 1.5911878895287306, "learning_rate": 1.9699532648003214e-05, "loss": 0.8799, "step": 2924 }, { "epoch": 0.10593987685621152, "grad_norm": 1.760011806055591, "learning_rate": 1.969924718407895e-05, "loss": 0.9239, "step": 2925 }, { "epoch": 0.10597609561752988, "grad_norm": 1.8605639899614146, "learning_rate": 1.969896158668452e-05, "loss": 0.8298, "step": 2926 }, { "epoch": 0.10601231437884824, "grad_norm": 1.5146424172267643, "learning_rate": 1.9698675855823858e-05, "loss": 0.8513, "step": 2927 }, { "epoch": 0.1060485331401666, "grad_norm": 1.2388889966917227, "learning_rate": 1.969838999150089e-05, "loss": 0.7473, "step": 2928 }, { "epoch": 0.10608475190148497, "grad_norm": 1.7426873439449047, "learning_rate": 1.9698103993719554e-05, "loss": 0.8803, "step": 2929 }, { "epoch": 0.10612097066280334, "grad_norm": 1.8137328337087715, "learning_rate": 1.9697817862483793e-05, "loss": 0.8645, "step": 2930 }, { "epoch": 0.1061571894241217, "grad_norm": 1.659809565529217, "learning_rate": 1.969753159779753e-05, "loss": 0.8414, "step": 2931 }, { "epoch": 0.10619340818544006, "grad_norm": 1.7248354738753797, "learning_rate": 1.9697245199664715e-05, "loss": 0.8729, "step": 2932 }, { "epoch": 0.10622962694675842, "grad_norm": 1.652631992648235, "learning_rate": 1.969695866808928e-05, "loss": 0.8444, "step": 2933 }, { "epoch": 0.10626584570807679, "grad_norm": 1.6655697488656749, "learning_rate": 1.969667200307518e-05, "loss": 0.8542, "step": 2934 }, { "epoch": 0.10630206446939515, "grad_norm": 1.9548049835039591, "learning_rate": 1.969638520462635e-05, "loss": 0.8893, "step": 2935 }, { "epoch": 0.10633828323071351, "grad_norm": 1.5525688383045138, "learning_rate": 1.969609827274674e-05, "loss": 0.8166, "step": 2936 }, { "epoch": 0.10637450199203187, "grad_norm": 1.4097930976873474, "learning_rate": 1.96958112074403e-05, "loss": 0.8329, "step": 2937 }, { "epoch": 0.10641072075335023, "grad_norm": 1.4170476431350036, "learning_rate": 1.9695524008710972e-05, "loss": 0.8568, "step": 2938 }, { "epoch": 0.1064469395146686, "grad_norm": 2.052368447220881, "learning_rate": 1.9695236676562717e-05, "loss": 0.9296, "step": 2939 }, { "epoch": 0.10648315827598696, "grad_norm": 1.8371829404464517, "learning_rate": 1.969494921099949e-05, "loss": 0.8338, "step": 2940 }, { "epoch": 0.10651937703730532, "grad_norm": 1.5768391840405653, "learning_rate": 1.969466161202524e-05, "loss": 0.8857, "step": 2941 }, { "epoch": 0.10655559579862368, "grad_norm": 1.3523009534771204, "learning_rate": 1.9694373879643928e-05, "loss": 0.7621, "step": 2942 }, { "epoch": 0.10659181455994204, "grad_norm": 1.6949274827891199, "learning_rate": 1.9694086013859516e-05, "loss": 0.864, "step": 2943 }, { "epoch": 0.10662803332126042, "grad_norm": 1.852134063079223, "learning_rate": 1.969379801467596e-05, "loss": 0.8781, "step": 2944 }, { "epoch": 0.10666425208257878, "grad_norm": 1.5637749143638953, "learning_rate": 1.9693509882097226e-05, "loss": 0.8303, "step": 2945 }, { "epoch": 0.10670047084389714, "grad_norm": 1.7194423808990558, "learning_rate": 1.9693221616127277e-05, "loss": 0.8873, "step": 2946 }, { "epoch": 0.1067366896052155, "grad_norm": 1.555628692933692, "learning_rate": 1.9692933216770083e-05, "loss": 0.8432, "step": 2947 }, { "epoch": 0.10677290836653386, "grad_norm": 1.431362934913301, "learning_rate": 1.969264468402961e-05, "loss": 0.8112, "step": 2948 }, { "epoch": 0.10680912712785223, "grad_norm": 1.285609305667427, "learning_rate": 1.969235601790983e-05, "loss": 0.8488, "step": 2949 }, { "epoch": 0.10684534588917059, "grad_norm": 1.890984111957499, "learning_rate": 1.9692067218414717e-05, "loss": 0.9335, "step": 2950 }, { "epoch": 0.10688156465048895, "grad_norm": 1.7294545549189566, "learning_rate": 1.969177828554824e-05, "loss": 0.7937, "step": 2951 }, { "epoch": 0.10691778341180731, "grad_norm": 1.3107534862537793, "learning_rate": 1.969148921931438e-05, "loss": 0.7934, "step": 2952 }, { "epoch": 0.10695400217312567, "grad_norm": 1.7208924119043114, "learning_rate": 1.9691200019717113e-05, "loss": 0.793, "step": 2953 }, { "epoch": 0.10699022093444405, "grad_norm": 1.6755579975331916, "learning_rate": 1.9690910686760414e-05, "loss": 0.8249, "step": 2954 }, { "epoch": 0.10702643969576241, "grad_norm": 1.547414518803855, "learning_rate": 1.969062122044827e-05, "loss": 0.8284, "step": 2955 }, { "epoch": 0.10706265845708077, "grad_norm": 1.4163964851145892, "learning_rate": 1.9690331620784665e-05, "loss": 0.8126, "step": 2956 }, { "epoch": 0.10709887721839913, "grad_norm": 1.3498961207278748, "learning_rate": 1.9690041887773584e-05, "loss": 0.7991, "step": 2957 }, { "epoch": 0.10713509597971749, "grad_norm": 1.4199924396528805, "learning_rate": 1.9689752021419008e-05, "loss": 0.7455, "step": 2958 }, { "epoch": 0.10717131474103586, "grad_norm": 1.5901244628017865, "learning_rate": 1.9689462021724933e-05, "loss": 0.8697, "step": 2959 }, { "epoch": 0.10720753350235422, "grad_norm": 1.7819216522253185, "learning_rate": 1.9689171888695345e-05, "loss": 0.8514, "step": 2960 }, { "epoch": 0.10724375226367258, "grad_norm": 1.6420432658966888, "learning_rate": 1.968888162233424e-05, "loss": 0.8248, "step": 2961 }, { "epoch": 0.10727997102499094, "grad_norm": 1.731353972511318, "learning_rate": 1.9688591222645607e-05, "loss": 0.7523, "step": 2962 }, { "epoch": 0.1073161897863093, "grad_norm": 1.563639585250706, "learning_rate": 1.9688300689633452e-05, "loss": 0.7578, "step": 2963 }, { "epoch": 0.10735240854762768, "grad_norm": 1.6237297629838197, "learning_rate": 1.968801002330176e-05, "loss": 0.8721, "step": 2964 }, { "epoch": 0.10738862730894604, "grad_norm": 1.605738766789287, "learning_rate": 1.968771922365454e-05, "loss": 0.831, "step": 2965 }, { "epoch": 0.1074248460702644, "grad_norm": 1.526077365633993, "learning_rate": 1.9687428290695792e-05, "loss": 0.7389, "step": 2966 }, { "epoch": 0.10746106483158276, "grad_norm": 1.5386900694742969, "learning_rate": 1.9687137224429515e-05, "loss": 0.7803, "step": 2967 }, { "epoch": 0.10749728359290112, "grad_norm": 1.71121575649028, "learning_rate": 1.968684602485972e-05, "loss": 0.8475, "step": 2968 }, { "epoch": 0.10753350235421949, "grad_norm": 1.3175527375644758, "learning_rate": 1.9686554691990417e-05, "loss": 0.8264, "step": 2969 }, { "epoch": 0.10756972111553785, "grad_norm": 1.7206618015033959, "learning_rate": 1.9686263225825604e-05, "loss": 0.9048, "step": 2970 }, { "epoch": 0.10760593987685621, "grad_norm": 1.3028215607649936, "learning_rate": 1.96859716263693e-05, "loss": 0.8129, "step": 2971 }, { "epoch": 0.10764215863817457, "grad_norm": 1.699977792917389, "learning_rate": 1.9685679893625512e-05, "loss": 0.8711, "step": 2972 }, { "epoch": 0.10767837739949293, "grad_norm": 1.897857473406803, "learning_rate": 1.9685388027598265e-05, "loss": 0.8641, "step": 2973 }, { "epoch": 0.1077145961608113, "grad_norm": 1.587513596770301, "learning_rate": 1.9685096028291562e-05, "loss": 0.7888, "step": 2974 }, { "epoch": 0.10775081492212966, "grad_norm": 1.611953142497356, "learning_rate": 1.9684803895709434e-05, "loss": 0.8749, "step": 2975 }, { "epoch": 0.10778703368344802, "grad_norm": 1.7693676220269432, "learning_rate": 1.968451162985589e-05, "loss": 0.8754, "step": 2976 }, { "epoch": 0.10782325244476638, "grad_norm": 1.688360536779765, "learning_rate": 1.9684219230734955e-05, "loss": 0.9577, "step": 2977 }, { "epoch": 0.10785947120608476, "grad_norm": 1.6262416388689258, "learning_rate": 1.9683926698350656e-05, "loss": 0.8343, "step": 2978 }, { "epoch": 0.10789568996740312, "grad_norm": 1.810763571403803, "learning_rate": 1.968363403270702e-05, "loss": 0.9096, "step": 2979 }, { "epoch": 0.10793190872872148, "grad_norm": 1.726680846495134, "learning_rate": 1.9683341233808067e-05, "loss": 0.8502, "step": 2980 }, { "epoch": 0.10796812749003984, "grad_norm": 1.787240568671459, "learning_rate": 1.968304830165783e-05, "loss": 0.9028, "step": 2981 }, { "epoch": 0.1080043462513582, "grad_norm": 1.6548765792648252, "learning_rate": 1.968275523626034e-05, "loss": 0.9136, "step": 2982 }, { "epoch": 0.10804056501267657, "grad_norm": 1.9872584232128294, "learning_rate": 1.9682462037619633e-05, "loss": 0.8885, "step": 2983 }, { "epoch": 0.10807678377399493, "grad_norm": 1.45562637708469, "learning_rate": 1.968216870573974e-05, "loss": 0.8218, "step": 2984 }, { "epoch": 0.1081130025353133, "grad_norm": 1.665795316138517, "learning_rate": 1.9681875240624696e-05, "loss": 0.79, "step": 2985 }, { "epoch": 0.10814922129663165, "grad_norm": 1.9465002001877547, "learning_rate": 1.968158164227854e-05, "loss": 0.8639, "step": 2986 }, { "epoch": 0.10818544005795001, "grad_norm": 1.5472633892930157, "learning_rate": 1.9681287910705317e-05, "loss": 0.7584, "step": 2987 }, { "epoch": 0.10822165881926839, "grad_norm": 1.6010830103291673, "learning_rate": 1.9680994045909064e-05, "loss": 0.8867, "step": 2988 }, { "epoch": 0.10825787758058675, "grad_norm": 1.7951672524310858, "learning_rate": 1.9680700047893827e-05, "loss": 0.8627, "step": 2989 }, { "epoch": 0.10829409634190511, "grad_norm": 1.4193417983985943, "learning_rate": 1.9680405916663652e-05, "loss": 0.8063, "step": 2990 }, { "epoch": 0.10833031510322347, "grad_norm": 1.6347534026551556, "learning_rate": 1.9680111652222584e-05, "loss": 0.8666, "step": 2991 }, { "epoch": 0.10836653386454183, "grad_norm": 1.7673451907702002, "learning_rate": 1.9679817254574676e-05, "loss": 0.9019, "step": 2992 }, { "epoch": 0.1084027526258602, "grad_norm": 1.2625420171045934, "learning_rate": 1.9679522723723975e-05, "loss": 0.7561, "step": 2993 }, { "epoch": 0.10843897138717856, "grad_norm": 1.8271588376685581, "learning_rate": 1.9679228059674534e-05, "loss": 0.8933, "step": 2994 }, { "epoch": 0.10847519014849692, "grad_norm": 1.187154352603298, "learning_rate": 1.9678933262430417e-05, "loss": 0.8143, "step": 2995 }, { "epoch": 0.10851140890981528, "grad_norm": 1.7204566415310938, "learning_rate": 1.9678638331995667e-05, "loss": 0.9034, "step": 2996 }, { "epoch": 0.10854762767113364, "grad_norm": 1.7709978512233853, "learning_rate": 1.9678343268374355e-05, "loss": 0.8578, "step": 2997 }, { "epoch": 0.10858384643245202, "grad_norm": 1.6775117855272716, "learning_rate": 1.967804807157053e-05, "loss": 0.8631, "step": 2998 }, { "epoch": 0.10862006519377038, "grad_norm": 1.9506290280154757, "learning_rate": 1.967775274158826e-05, "loss": 0.871, "step": 2999 }, { "epoch": 0.10865628395508874, "grad_norm": 1.8435739901955015, "learning_rate": 1.9677457278431614e-05, "loss": 0.9349, "step": 3000 }, { "epoch": 0.1086925027164071, "grad_norm": 1.5991924497103096, "learning_rate": 1.9677161682104647e-05, "loss": 0.7889, "step": 3001 }, { "epoch": 0.10872872147772546, "grad_norm": 1.7387443782762138, "learning_rate": 1.9676865952611436e-05, "loss": 0.9328, "step": 3002 }, { "epoch": 0.10876494023904383, "grad_norm": 1.4107630744814734, "learning_rate": 1.9676570089956043e-05, "loss": 0.7913, "step": 3003 }, { "epoch": 0.10880115900036219, "grad_norm": 1.4586032612389097, "learning_rate": 1.9676274094142544e-05, "loss": 0.7945, "step": 3004 }, { "epoch": 0.10883737776168055, "grad_norm": 1.3277869818204442, "learning_rate": 1.9675977965175015e-05, "loss": 0.7592, "step": 3005 }, { "epoch": 0.10887359652299891, "grad_norm": 1.8565835203991514, "learning_rate": 1.9675681703057523e-05, "loss": 0.843, "step": 3006 }, { "epoch": 0.10890981528431727, "grad_norm": 1.5937684457777361, "learning_rate": 1.9675385307794152e-05, "loss": 0.8336, "step": 3007 }, { "epoch": 0.10894603404563565, "grad_norm": 1.6819382846051232, "learning_rate": 1.9675088779388972e-05, "loss": 0.878, "step": 3008 }, { "epoch": 0.108982252806954, "grad_norm": 1.5468370656939687, "learning_rate": 1.9674792117846073e-05, "loss": 0.8494, "step": 3009 }, { "epoch": 0.10901847156827237, "grad_norm": 1.7577439315985754, "learning_rate": 1.9674495323169537e-05, "loss": 0.8965, "step": 3010 }, { "epoch": 0.10905469032959073, "grad_norm": 1.535637811545997, "learning_rate": 1.9674198395363436e-05, "loss": 0.7383, "step": 3011 }, { "epoch": 0.10909090909090909, "grad_norm": 1.813246908113307, "learning_rate": 1.967390133443187e-05, "loss": 0.9505, "step": 3012 }, { "epoch": 0.10912712785222746, "grad_norm": 1.9104615262032962, "learning_rate": 1.9673604140378925e-05, "loss": 0.8382, "step": 3013 }, { "epoch": 0.10916334661354582, "grad_norm": 1.8556339874852754, "learning_rate": 1.9673306813208682e-05, "loss": 0.8542, "step": 3014 }, { "epoch": 0.10919956537486418, "grad_norm": 1.8451898547396142, "learning_rate": 1.967300935292524e-05, "loss": 0.9388, "step": 3015 }, { "epoch": 0.10923578413618254, "grad_norm": 1.5365289588364488, "learning_rate": 1.9672711759532684e-05, "loss": 0.8202, "step": 3016 }, { "epoch": 0.1092720028975009, "grad_norm": 1.4480749669223743, "learning_rate": 1.967241403303512e-05, "loss": 0.7405, "step": 3017 }, { "epoch": 0.10930822165881927, "grad_norm": 1.2584975955215922, "learning_rate": 1.967211617343664e-05, "loss": 0.7875, "step": 3018 }, { "epoch": 0.10934444042013763, "grad_norm": 2.738981832789638, "learning_rate": 1.9671818180741343e-05, "loss": 0.8698, "step": 3019 }, { "epoch": 0.109380659181456, "grad_norm": 1.620531525840098, "learning_rate": 1.9671520054953327e-05, "loss": 0.7547, "step": 3020 }, { "epoch": 0.10941687794277435, "grad_norm": 1.8527495928691704, "learning_rate": 1.9671221796076702e-05, "loss": 0.8985, "step": 3021 }, { "epoch": 0.10945309670409271, "grad_norm": 1.5585950442717493, "learning_rate": 1.967092340411556e-05, "loss": 0.8936, "step": 3022 }, { "epoch": 0.10948931546541109, "grad_norm": 1.3874653161539692, "learning_rate": 1.9670624879074018e-05, "loss": 0.8452, "step": 3023 }, { "epoch": 0.10952553422672945, "grad_norm": 1.4268545110481425, "learning_rate": 1.967032622095618e-05, "loss": 0.8294, "step": 3024 }, { "epoch": 0.10956175298804781, "grad_norm": 1.7712193556855738, "learning_rate": 1.9670027429766157e-05, "loss": 0.8296, "step": 3025 }, { "epoch": 0.10959797174936617, "grad_norm": 1.2389598820665844, "learning_rate": 1.9669728505508063e-05, "loss": 0.7788, "step": 3026 }, { "epoch": 0.10963419051068453, "grad_norm": 1.7897887567391324, "learning_rate": 1.9669429448186e-05, "loss": 0.9617, "step": 3027 }, { "epoch": 0.1096704092720029, "grad_norm": 1.6671335387233992, "learning_rate": 1.9669130257804096e-05, "loss": 0.7956, "step": 3028 }, { "epoch": 0.10970662803332126, "grad_norm": 1.716682259375876, "learning_rate": 1.966883093436646e-05, "loss": 0.9049, "step": 3029 }, { "epoch": 0.10974284679463962, "grad_norm": 1.2333036629454894, "learning_rate": 1.9668531477877223e-05, "loss": 0.7806, "step": 3030 }, { "epoch": 0.10977906555595798, "grad_norm": 1.6126742444874624, "learning_rate": 1.966823188834049e-05, "loss": 0.9154, "step": 3031 }, { "epoch": 0.10981528431727634, "grad_norm": 1.6641117299100312, "learning_rate": 1.9667932165760397e-05, "loss": 0.8339, "step": 3032 }, { "epoch": 0.10985150307859472, "grad_norm": 1.6776221718533013, "learning_rate": 1.9667632310141055e-05, "loss": 0.8983, "step": 3033 }, { "epoch": 0.10988772183991308, "grad_norm": 1.6637944645890568, "learning_rate": 1.9667332321486603e-05, "loss": 0.8567, "step": 3034 }, { "epoch": 0.10992394060123144, "grad_norm": 1.5053018256431228, "learning_rate": 1.966703219980116e-05, "loss": 0.8269, "step": 3035 }, { "epoch": 0.1099601593625498, "grad_norm": 1.7432248166393494, "learning_rate": 1.9666731945088863e-05, "loss": 0.8521, "step": 3036 }, { "epoch": 0.10999637812386816, "grad_norm": 1.7878092557473928, "learning_rate": 1.966643155735384e-05, "loss": 0.8519, "step": 3037 }, { "epoch": 0.11003259688518653, "grad_norm": 3.19588836479966, "learning_rate": 1.9666131036600225e-05, "loss": 0.7464, "step": 3038 }, { "epoch": 0.11006881564650489, "grad_norm": 1.764447843110069, "learning_rate": 1.9665830382832152e-05, "loss": 0.8857, "step": 3039 }, { "epoch": 0.11010503440782325, "grad_norm": 1.6028058627895394, "learning_rate": 1.9665529596053764e-05, "loss": 0.8039, "step": 3040 }, { "epoch": 0.11014125316914161, "grad_norm": 1.6985038952716112, "learning_rate": 1.966522867626919e-05, "loss": 0.8139, "step": 3041 }, { "epoch": 0.11017747193045997, "grad_norm": 1.5970622744643124, "learning_rate": 1.966492762348258e-05, "loss": 0.8248, "step": 3042 }, { "epoch": 0.11021369069177835, "grad_norm": 1.5445263786798664, "learning_rate": 1.9664626437698074e-05, "loss": 0.8712, "step": 3043 }, { "epoch": 0.1102499094530967, "grad_norm": 1.9226857506675514, "learning_rate": 1.9664325118919813e-05, "loss": 0.8787, "step": 3044 }, { "epoch": 0.11028612821441507, "grad_norm": 1.86897077157389, "learning_rate": 1.966402366715195e-05, "loss": 0.8402, "step": 3045 }, { "epoch": 0.11032234697573343, "grad_norm": 1.2820530535786663, "learning_rate": 1.966372208239863e-05, "loss": 0.7908, "step": 3046 }, { "epoch": 0.11035856573705179, "grad_norm": 1.843776395576459, "learning_rate": 1.9663420364663998e-05, "loss": 0.926, "step": 3047 }, { "epoch": 0.11039478449837016, "grad_norm": 1.6542309271628266, "learning_rate": 1.9663118513952217e-05, "loss": 0.8517, "step": 3048 }, { "epoch": 0.11043100325968852, "grad_norm": 1.735804250274444, "learning_rate": 1.966281653026743e-05, "loss": 0.7922, "step": 3049 }, { "epoch": 0.11046722202100688, "grad_norm": 1.971313676876025, "learning_rate": 1.9662514413613796e-05, "loss": 0.9053, "step": 3050 }, { "epoch": 0.11050344078232524, "grad_norm": 1.584542981484171, "learning_rate": 1.9662212163995478e-05, "loss": 0.8238, "step": 3051 }, { "epoch": 0.1105396595436436, "grad_norm": 1.4342930757478343, "learning_rate": 1.966190978141663e-05, "loss": 0.8271, "step": 3052 }, { "epoch": 0.11057587830496197, "grad_norm": 1.6471728885419075, "learning_rate": 1.9661607265881414e-05, "loss": 0.9133, "step": 3053 }, { "epoch": 0.11061209706628033, "grad_norm": 1.4942702112537853, "learning_rate": 1.966130461739399e-05, "loss": 0.8807, "step": 3054 }, { "epoch": 0.1106483158275987, "grad_norm": 1.553942103904867, "learning_rate": 1.9661001835958526e-05, "loss": 0.8369, "step": 3055 }, { "epoch": 0.11068453458891706, "grad_norm": 1.348443986009208, "learning_rate": 1.966069892157919e-05, "loss": 0.8309, "step": 3056 }, { "epoch": 0.11072075335023542, "grad_norm": 1.6884217570529554, "learning_rate": 1.9660395874260146e-05, "loss": 0.8943, "step": 3057 }, { "epoch": 0.11075697211155379, "grad_norm": 1.370767476236133, "learning_rate": 1.9660092694005566e-05, "loss": 0.7848, "step": 3058 }, { "epoch": 0.11079319087287215, "grad_norm": 1.7625903724436638, "learning_rate": 1.9659789380819624e-05, "loss": 0.9279, "step": 3059 }, { "epoch": 0.11082940963419051, "grad_norm": 1.8122962352063485, "learning_rate": 1.965948593470649e-05, "loss": 0.9447, "step": 3060 }, { "epoch": 0.11086562839550887, "grad_norm": 1.2663458137741352, "learning_rate": 1.9659182355670345e-05, "loss": 0.774, "step": 3061 }, { "epoch": 0.11090184715682723, "grad_norm": 1.8184186042784614, "learning_rate": 1.9658878643715362e-05, "loss": 0.8845, "step": 3062 }, { "epoch": 0.1109380659181456, "grad_norm": 1.9849301635434604, "learning_rate": 1.965857479884572e-05, "loss": 0.8317, "step": 3063 }, { "epoch": 0.11097428467946396, "grad_norm": 1.5239392380515984, "learning_rate": 1.9658270821065603e-05, "loss": 0.8068, "step": 3064 }, { "epoch": 0.11101050344078232, "grad_norm": 1.6835864259423354, "learning_rate": 1.9657966710379195e-05, "loss": 0.8674, "step": 3065 }, { "epoch": 0.11104672220210068, "grad_norm": 1.2596000297332273, "learning_rate": 1.9657662466790678e-05, "loss": 0.832, "step": 3066 }, { "epoch": 0.11108294096341906, "grad_norm": 1.4103669247116148, "learning_rate": 1.965735809030424e-05, "loss": 0.8057, "step": 3067 }, { "epoch": 0.11111915972473742, "grad_norm": 1.970934477157053, "learning_rate": 1.965705358092407e-05, "loss": 0.8569, "step": 3068 }, { "epoch": 0.11115537848605578, "grad_norm": 1.5937838983203172, "learning_rate": 1.9656748938654352e-05, "loss": 0.814, "step": 3069 }, { "epoch": 0.11119159724737414, "grad_norm": 1.7180793530740628, "learning_rate": 1.9656444163499288e-05, "loss": 0.8499, "step": 3070 }, { "epoch": 0.1112278160086925, "grad_norm": 1.879330541562494, "learning_rate": 1.9656139255463066e-05, "loss": 0.9177, "step": 3071 }, { "epoch": 0.11126403477001087, "grad_norm": 1.6477451927195672, "learning_rate": 1.9655834214549883e-05, "loss": 0.8896, "step": 3072 }, { "epoch": 0.11130025353132923, "grad_norm": 1.6384779823361642, "learning_rate": 1.9655529040763936e-05, "loss": 0.9135, "step": 3073 }, { "epoch": 0.11133647229264759, "grad_norm": 1.5586523209946086, "learning_rate": 1.965522373410943e-05, "loss": 0.895, "step": 3074 }, { "epoch": 0.11137269105396595, "grad_norm": 1.543828623232856, "learning_rate": 1.9654918294590554e-05, "loss": 0.7931, "step": 3075 }, { "epoch": 0.11140890981528431, "grad_norm": 1.585113219574392, "learning_rate": 1.9654612722211525e-05, "loss": 0.8028, "step": 3076 }, { "epoch": 0.11144512857660269, "grad_norm": 1.6052828522185763, "learning_rate": 1.965430701697654e-05, "loss": 0.8841, "step": 3077 }, { "epoch": 0.11148134733792105, "grad_norm": 1.8126657391942962, "learning_rate": 1.965400117888981e-05, "loss": 0.8467, "step": 3078 }, { "epoch": 0.1115175660992394, "grad_norm": 1.4030306938829966, "learning_rate": 1.9653695207955534e-05, "loss": 0.8663, "step": 3079 }, { "epoch": 0.11155378486055777, "grad_norm": 1.794815591647874, "learning_rate": 1.9653389104177935e-05, "loss": 0.8615, "step": 3080 }, { "epoch": 0.11159000362187613, "grad_norm": 1.699241714760827, "learning_rate": 1.965308286756122e-05, "loss": 0.8321, "step": 3081 }, { "epoch": 0.1116262223831945, "grad_norm": 1.1965954368797396, "learning_rate": 1.9652776498109597e-05, "loss": 0.7999, "step": 3082 }, { "epoch": 0.11166244114451286, "grad_norm": 1.7706161103999987, "learning_rate": 1.9652469995827292e-05, "loss": 0.8532, "step": 3083 }, { "epoch": 0.11169865990583122, "grad_norm": 1.6769044864157812, "learning_rate": 1.965216336071852e-05, "loss": 0.8366, "step": 3084 }, { "epoch": 0.11173487866714958, "grad_norm": 1.6542150483955211, "learning_rate": 1.9651856592787497e-05, "loss": 0.8856, "step": 3085 }, { "epoch": 0.11177109742846794, "grad_norm": 1.6859954179709167, "learning_rate": 1.9651549692038444e-05, "loss": 0.8726, "step": 3086 }, { "epoch": 0.11180731618978632, "grad_norm": 1.6256322295648895, "learning_rate": 1.965124265847559e-05, "loss": 0.8493, "step": 3087 }, { "epoch": 0.11184353495110468, "grad_norm": 1.900134909382707, "learning_rate": 1.9650935492103158e-05, "loss": 0.9557, "step": 3088 }, { "epoch": 0.11187975371242304, "grad_norm": 1.4740177887853443, "learning_rate": 1.9650628192925372e-05, "loss": 0.774, "step": 3089 }, { "epoch": 0.1119159724737414, "grad_norm": 1.3661280818258872, "learning_rate": 1.9650320760946463e-05, "loss": 0.7777, "step": 3090 }, { "epoch": 0.11195219123505976, "grad_norm": 1.6966769423385197, "learning_rate": 1.9650013196170658e-05, "loss": 0.859, "step": 3091 }, { "epoch": 0.11198840999637813, "grad_norm": 1.6647002750853233, "learning_rate": 1.9649705498602193e-05, "loss": 0.8895, "step": 3092 }, { "epoch": 0.11202462875769649, "grad_norm": 1.342754868812484, "learning_rate": 1.9649397668245306e-05, "loss": 0.706, "step": 3093 }, { "epoch": 0.11206084751901485, "grad_norm": 1.586702593284928, "learning_rate": 1.9649089705104223e-05, "loss": 0.9201, "step": 3094 }, { "epoch": 0.11209706628033321, "grad_norm": 1.5753312132094146, "learning_rate": 1.9648781609183187e-05, "loss": 0.7924, "step": 3095 }, { "epoch": 0.11213328504165157, "grad_norm": 1.616320687569649, "learning_rate": 1.964847338048644e-05, "loss": 0.8385, "step": 3096 }, { "epoch": 0.11216950380296994, "grad_norm": 1.9509452195589954, "learning_rate": 1.964816501901822e-05, "loss": 0.8654, "step": 3097 }, { "epoch": 0.1122057225642883, "grad_norm": 1.6945088905646177, "learning_rate": 1.9647856524782774e-05, "loss": 0.9309, "step": 3098 }, { "epoch": 0.11224194132560666, "grad_norm": 1.2626226168960908, "learning_rate": 1.9647547897784346e-05, "loss": 0.7829, "step": 3099 }, { "epoch": 0.11227816008692502, "grad_norm": 1.5919143457889622, "learning_rate": 1.964723913802718e-05, "loss": 0.8618, "step": 3100 }, { "epoch": 0.11231437884824338, "grad_norm": 1.263020810187887, "learning_rate": 1.9646930245515524e-05, "loss": 0.8484, "step": 3101 }, { "epoch": 0.11235059760956176, "grad_norm": 1.8450495879848048, "learning_rate": 1.9646621220253635e-05, "loss": 0.8697, "step": 3102 }, { "epoch": 0.11238681637088012, "grad_norm": 1.6595266663032158, "learning_rate": 1.964631206224576e-05, "loss": 0.82, "step": 3103 }, { "epoch": 0.11242303513219848, "grad_norm": 1.7977440743377204, "learning_rate": 1.964600277149615e-05, "loss": 0.8911, "step": 3104 }, { "epoch": 0.11245925389351684, "grad_norm": 1.7008595694556319, "learning_rate": 1.9645693348009075e-05, "loss": 0.9448, "step": 3105 }, { "epoch": 0.1124954726548352, "grad_norm": 1.800243510474025, "learning_rate": 1.9645383791788778e-05, "loss": 0.8896, "step": 3106 }, { "epoch": 0.11253169141615357, "grad_norm": 1.927939501654226, "learning_rate": 1.964507410283953e-05, "loss": 0.9461, "step": 3107 }, { "epoch": 0.11256791017747193, "grad_norm": 1.528561735737214, "learning_rate": 1.9644764281165584e-05, "loss": 0.7811, "step": 3108 }, { "epoch": 0.1126041289387903, "grad_norm": 1.801125409758179, "learning_rate": 1.964445432677121e-05, "loss": 0.9176, "step": 3109 }, { "epoch": 0.11264034770010865, "grad_norm": 1.8761437345066634, "learning_rate": 1.9644144239660665e-05, "loss": 0.7999, "step": 3110 }, { "epoch": 0.11267656646142701, "grad_norm": 1.731256413361827, "learning_rate": 1.9643834019838226e-05, "loss": 0.9035, "step": 3111 }, { "epoch": 0.11271278522274539, "grad_norm": 1.6119007122060556, "learning_rate": 1.9643523667308157e-05, "loss": 0.9117, "step": 3112 }, { "epoch": 0.11274900398406375, "grad_norm": 1.712133057378506, "learning_rate": 1.9643213182074724e-05, "loss": 0.8345, "step": 3113 }, { "epoch": 0.11278522274538211, "grad_norm": 1.9588451602261834, "learning_rate": 1.964290256414221e-05, "loss": 0.9612, "step": 3114 }, { "epoch": 0.11282144150670047, "grad_norm": 1.7158389581699574, "learning_rate": 1.9642591813514877e-05, "loss": 0.8329, "step": 3115 }, { "epoch": 0.11285766026801883, "grad_norm": 1.767440969720395, "learning_rate": 1.9642280930197012e-05, "loss": 0.8981, "step": 3116 }, { "epoch": 0.1128938790293372, "grad_norm": 1.6782030565606252, "learning_rate": 1.964196991419289e-05, "loss": 0.8904, "step": 3117 }, { "epoch": 0.11293009779065556, "grad_norm": 1.6523391586041518, "learning_rate": 1.964165876550679e-05, "loss": 0.7981, "step": 3118 }, { "epoch": 0.11296631655197392, "grad_norm": 1.8264313467580864, "learning_rate": 1.964134748414299e-05, "loss": 0.9254, "step": 3119 }, { "epoch": 0.11300253531329228, "grad_norm": 1.6671654053111515, "learning_rate": 1.964103607010578e-05, "loss": 0.8768, "step": 3120 }, { "epoch": 0.11303875407461064, "grad_norm": 1.5488801239558212, "learning_rate": 1.964072452339944e-05, "loss": 0.7856, "step": 3121 }, { "epoch": 0.11307497283592902, "grad_norm": 1.7657997184154188, "learning_rate": 1.964041284402826e-05, "loss": 0.9162, "step": 3122 }, { "epoch": 0.11311119159724738, "grad_norm": 1.366497264229264, "learning_rate": 1.964010103199653e-05, "loss": 0.846, "step": 3123 }, { "epoch": 0.11314741035856574, "grad_norm": 1.3747446305380866, "learning_rate": 1.963978908730854e-05, "loss": 0.8173, "step": 3124 }, { "epoch": 0.1131836291198841, "grad_norm": 1.7148836376704086, "learning_rate": 1.963947700996858e-05, "loss": 0.7908, "step": 3125 }, { "epoch": 0.11321984788120246, "grad_norm": 1.7773329451515507, "learning_rate": 1.9639164799980946e-05, "loss": 0.7369, "step": 3126 }, { "epoch": 0.11325606664252083, "grad_norm": 1.6031880177647266, "learning_rate": 1.9638852457349936e-05, "loss": 0.9128, "step": 3127 }, { "epoch": 0.11329228540383919, "grad_norm": 1.7387099648192015, "learning_rate": 1.9638539982079844e-05, "loss": 0.934, "step": 3128 }, { "epoch": 0.11332850416515755, "grad_norm": 1.7670870525573121, "learning_rate": 1.9638227374174977e-05, "loss": 0.8485, "step": 3129 }, { "epoch": 0.11336472292647591, "grad_norm": 1.6739197086130384, "learning_rate": 1.963791463363963e-05, "loss": 0.8409, "step": 3130 }, { "epoch": 0.11340094168779427, "grad_norm": 1.5864115626624633, "learning_rate": 1.963760176047811e-05, "loss": 0.8427, "step": 3131 }, { "epoch": 0.11343716044911265, "grad_norm": 1.7169545983261632, "learning_rate": 1.9637288754694717e-05, "loss": 0.9108, "step": 3132 }, { "epoch": 0.113473379210431, "grad_norm": 1.4365684063328223, "learning_rate": 1.9636975616293766e-05, "loss": 0.8562, "step": 3133 }, { "epoch": 0.11350959797174937, "grad_norm": 1.6887430705483166, "learning_rate": 1.9636662345279568e-05, "loss": 0.9804, "step": 3134 }, { "epoch": 0.11354581673306773, "grad_norm": 1.6682904511694872, "learning_rate": 1.9636348941656422e-05, "loss": 0.9266, "step": 3135 }, { "epoch": 0.11358203549438609, "grad_norm": 1.6605402530411604, "learning_rate": 1.963603540542865e-05, "loss": 0.8462, "step": 3136 }, { "epoch": 0.11361825425570446, "grad_norm": 1.4541174200511429, "learning_rate": 1.9635721736600562e-05, "loss": 0.8011, "step": 3137 }, { "epoch": 0.11365447301702282, "grad_norm": 1.3005916867943272, "learning_rate": 1.9635407935176477e-05, "loss": 0.8239, "step": 3138 }, { "epoch": 0.11369069177834118, "grad_norm": 1.3590898596690344, "learning_rate": 1.9635094001160714e-05, "loss": 0.8401, "step": 3139 }, { "epoch": 0.11372691053965954, "grad_norm": 1.8823947700311157, "learning_rate": 1.963477993455759e-05, "loss": 0.8555, "step": 3140 }, { "epoch": 0.1137631293009779, "grad_norm": 1.2920828008629173, "learning_rate": 1.9634465735371427e-05, "loss": 0.8126, "step": 3141 }, { "epoch": 0.11379934806229627, "grad_norm": 1.647658291603368, "learning_rate": 1.9634151403606553e-05, "loss": 0.8564, "step": 3142 }, { "epoch": 0.11383556682361463, "grad_norm": 1.1939579343276823, "learning_rate": 1.9633836939267287e-05, "loss": 0.8095, "step": 3143 }, { "epoch": 0.113871785584933, "grad_norm": 1.6393520480398798, "learning_rate": 1.9633522342357962e-05, "loss": 0.8944, "step": 3144 }, { "epoch": 0.11390800434625135, "grad_norm": 1.553182241938764, "learning_rate": 1.9633207612882903e-05, "loss": 0.8064, "step": 3145 }, { "epoch": 0.11394422310756971, "grad_norm": 1.6181586592957573, "learning_rate": 1.9632892750846445e-05, "loss": 0.8969, "step": 3146 }, { "epoch": 0.11398044186888809, "grad_norm": 1.488677854333269, "learning_rate": 1.963257775625292e-05, "loss": 0.8864, "step": 3147 }, { "epoch": 0.11401666063020645, "grad_norm": 2.8928169315149557, "learning_rate": 1.9632262629106655e-05, "loss": 0.7663, "step": 3148 }, { "epoch": 0.11405287939152481, "grad_norm": 1.7140848135734617, "learning_rate": 1.9631947369411993e-05, "loss": 0.8545, "step": 3149 }, { "epoch": 0.11408909815284317, "grad_norm": 1.6536298891235, "learning_rate": 1.9631631977173276e-05, "loss": 0.8465, "step": 3150 }, { "epoch": 0.11412531691416154, "grad_norm": 1.9300734439512062, "learning_rate": 1.963131645239484e-05, "loss": 0.8736, "step": 3151 }, { "epoch": 0.1141615356754799, "grad_norm": 2.000080920981457, "learning_rate": 1.9631000795081024e-05, "loss": 0.8709, "step": 3152 }, { "epoch": 0.11419775443679826, "grad_norm": 1.7686834188421834, "learning_rate": 1.9630685005236175e-05, "loss": 0.8393, "step": 3153 }, { "epoch": 0.11423397319811662, "grad_norm": 1.6851823562895427, "learning_rate": 1.9630369082864643e-05, "loss": 0.8884, "step": 3154 }, { "epoch": 0.11427019195943498, "grad_norm": 1.7315750073495997, "learning_rate": 1.9630053027970764e-05, "loss": 0.8668, "step": 3155 }, { "epoch": 0.11430641072075336, "grad_norm": 1.5432440958728935, "learning_rate": 1.9629736840558896e-05, "loss": 0.801, "step": 3156 }, { "epoch": 0.11434262948207172, "grad_norm": 1.7716801649784137, "learning_rate": 1.9629420520633392e-05, "loss": 0.8577, "step": 3157 }, { "epoch": 0.11437884824339008, "grad_norm": 1.6319076389926217, "learning_rate": 1.9629104068198593e-05, "loss": 0.86, "step": 3158 }, { "epoch": 0.11441506700470844, "grad_norm": 1.3925808087947713, "learning_rate": 1.9628787483258868e-05, "loss": 0.8006, "step": 3159 }, { "epoch": 0.1144512857660268, "grad_norm": 1.8852261193200486, "learning_rate": 1.9628470765818562e-05, "loss": 0.9162, "step": 3160 }, { "epoch": 0.11448750452734517, "grad_norm": 1.2177633007487274, "learning_rate": 1.962815391588204e-05, "loss": 0.7812, "step": 3161 }, { "epoch": 0.11452372328866353, "grad_norm": 1.408940871395057, "learning_rate": 1.9627836933453656e-05, "loss": 0.8149, "step": 3162 }, { "epoch": 0.11455994204998189, "grad_norm": 1.6457273602823332, "learning_rate": 1.962751981853778e-05, "loss": 0.9007, "step": 3163 }, { "epoch": 0.11459616081130025, "grad_norm": 1.6080518637500412, "learning_rate": 1.962720257113877e-05, "loss": 0.8493, "step": 3164 }, { "epoch": 0.11463237957261861, "grad_norm": 1.5844683419331973, "learning_rate": 1.9626885191260997e-05, "loss": 0.8083, "step": 3165 }, { "epoch": 0.11466859833393699, "grad_norm": 1.4131306981904432, "learning_rate": 1.9626567678908822e-05, "loss": 0.8535, "step": 3166 }, { "epoch": 0.11470481709525535, "grad_norm": 1.6630669232646222, "learning_rate": 1.9626250034086617e-05, "loss": 0.9087, "step": 3167 }, { "epoch": 0.1147410358565737, "grad_norm": 1.5873816555415212, "learning_rate": 1.9625932256798755e-05, "loss": 0.8435, "step": 3168 }, { "epoch": 0.11477725461789207, "grad_norm": 1.615899989880564, "learning_rate": 1.96256143470496e-05, "loss": 0.8131, "step": 3169 }, { "epoch": 0.11481347337921043, "grad_norm": 1.6961958535104413, "learning_rate": 1.962529630484354e-05, "loss": 0.8814, "step": 3170 }, { "epoch": 0.1148496921405288, "grad_norm": 1.5459726861604943, "learning_rate": 1.9624978130184948e-05, "loss": 0.8377, "step": 3171 }, { "epoch": 0.11488591090184716, "grad_norm": 1.584167782646243, "learning_rate": 1.9624659823078195e-05, "loss": 0.8262, "step": 3172 }, { "epoch": 0.11492212966316552, "grad_norm": 1.2571126070923981, "learning_rate": 1.962434138352767e-05, "loss": 0.7812, "step": 3173 }, { "epoch": 0.11495834842448388, "grad_norm": 2.5109318962372904, "learning_rate": 1.9624022811537748e-05, "loss": 0.8292, "step": 3174 }, { "epoch": 0.11499456718580224, "grad_norm": 1.7906260623146486, "learning_rate": 1.9623704107112816e-05, "loss": 0.8955, "step": 3175 }, { "epoch": 0.11503078594712061, "grad_norm": 1.710387421280313, "learning_rate": 1.9623385270257258e-05, "loss": 0.8705, "step": 3176 }, { "epoch": 0.11506700470843897, "grad_norm": 1.5258113009162853, "learning_rate": 1.962306630097546e-05, "loss": 0.819, "step": 3177 }, { "epoch": 0.11510322346975733, "grad_norm": 1.6794049147275858, "learning_rate": 1.9622747199271822e-05, "loss": 0.8236, "step": 3178 }, { "epoch": 0.1151394422310757, "grad_norm": 1.773277867900767, "learning_rate": 1.9622427965150724e-05, "loss": 0.9295, "step": 3179 }, { "epoch": 0.11517566099239405, "grad_norm": 2.2311008112314994, "learning_rate": 1.962210859861656e-05, "loss": 0.8479, "step": 3180 }, { "epoch": 0.11521187975371243, "grad_norm": 1.6847231632491932, "learning_rate": 1.962178909967373e-05, "loss": 0.8468, "step": 3181 }, { "epoch": 0.11524809851503079, "grad_norm": 1.529857503505595, "learning_rate": 1.9621469468326624e-05, "loss": 0.8194, "step": 3182 }, { "epoch": 0.11528431727634915, "grad_norm": 1.517654034104247, "learning_rate": 1.9621149704579646e-05, "loss": 0.8285, "step": 3183 }, { "epoch": 0.11532053603766751, "grad_norm": 1.538464610700007, "learning_rate": 1.9620829808437193e-05, "loss": 0.8742, "step": 3184 }, { "epoch": 0.11535675479898587, "grad_norm": 1.7764865956744902, "learning_rate": 1.9620509779903673e-05, "loss": 0.844, "step": 3185 }, { "epoch": 0.11539297356030424, "grad_norm": 1.3928700952901434, "learning_rate": 1.962018961898348e-05, "loss": 0.7885, "step": 3186 }, { "epoch": 0.1154291923216226, "grad_norm": 1.2676383484147138, "learning_rate": 1.9619869325681028e-05, "loss": 0.8131, "step": 3187 }, { "epoch": 0.11546541108294096, "grad_norm": 1.2077577441200944, "learning_rate": 1.961954890000072e-05, "loss": 0.7955, "step": 3188 }, { "epoch": 0.11550162984425932, "grad_norm": 1.5651565046863363, "learning_rate": 1.9619228341946964e-05, "loss": 0.8422, "step": 3189 }, { "epoch": 0.11553784860557768, "grad_norm": 1.7843725103404353, "learning_rate": 1.9618907651524176e-05, "loss": 0.885, "step": 3190 }, { "epoch": 0.11557406736689606, "grad_norm": 1.5860557817407839, "learning_rate": 1.961858682873677e-05, "loss": 0.891, "step": 3191 }, { "epoch": 0.11561028612821442, "grad_norm": 1.781472731769155, "learning_rate": 1.9618265873589154e-05, "loss": 0.8643, "step": 3192 }, { "epoch": 0.11564650488953278, "grad_norm": 1.4442290973609926, "learning_rate": 1.9617944786085746e-05, "loss": 0.7737, "step": 3193 }, { "epoch": 0.11568272365085114, "grad_norm": 1.6397784876183767, "learning_rate": 1.961762356623097e-05, "loss": 0.8693, "step": 3194 }, { "epoch": 0.1157189424121695, "grad_norm": 1.5185681669826767, "learning_rate": 1.9617302214029242e-05, "loss": 0.8168, "step": 3195 }, { "epoch": 0.11575516117348787, "grad_norm": 1.519518013727353, "learning_rate": 1.961698072948499e-05, "loss": 0.8357, "step": 3196 }, { "epoch": 0.11579137993480623, "grad_norm": 1.4664777968699474, "learning_rate": 1.9616659112602625e-05, "loss": 0.8323, "step": 3197 }, { "epoch": 0.11582759869612459, "grad_norm": 1.9666434060522109, "learning_rate": 1.9616337363386586e-05, "loss": 0.8914, "step": 3198 }, { "epoch": 0.11586381745744295, "grad_norm": 1.6848750471830698, "learning_rate": 1.9616015481841293e-05, "loss": 0.8868, "step": 3199 }, { "epoch": 0.11590003621876131, "grad_norm": 1.647500824007326, "learning_rate": 1.961569346797118e-05, "loss": 0.7631, "step": 3200 }, { "epoch": 0.11593625498007969, "grad_norm": 1.6740517628229097, "learning_rate": 1.961537132178067e-05, "loss": 0.8202, "step": 3201 }, { "epoch": 0.11597247374139805, "grad_norm": 1.6308333069322622, "learning_rate": 1.9615049043274207e-05, "loss": 0.8467, "step": 3202 }, { "epoch": 0.1160086925027164, "grad_norm": 1.4198038497416543, "learning_rate": 1.9614726632456217e-05, "loss": 0.8768, "step": 3203 }, { "epoch": 0.11604491126403477, "grad_norm": 1.8897569452719065, "learning_rate": 1.961440408933114e-05, "loss": 0.9534, "step": 3204 }, { "epoch": 0.11608113002535313, "grad_norm": 1.5766970079922915, "learning_rate": 1.961408141390342e-05, "loss": 0.791, "step": 3205 }, { "epoch": 0.1161173487866715, "grad_norm": 1.723505548204163, "learning_rate": 1.9613758606177492e-05, "loss": 0.8744, "step": 3206 }, { "epoch": 0.11615356754798986, "grad_norm": 1.4724972813658885, "learning_rate": 1.9613435666157793e-05, "loss": 0.8669, "step": 3207 }, { "epoch": 0.11618978630930822, "grad_norm": 1.7336203537851282, "learning_rate": 1.9613112593848775e-05, "loss": 0.8967, "step": 3208 }, { "epoch": 0.11622600507062658, "grad_norm": 1.6923570856603176, "learning_rate": 1.961278938925488e-05, "loss": 0.8433, "step": 3209 }, { "epoch": 0.11626222383194494, "grad_norm": 1.6339092934348376, "learning_rate": 1.9612466052380557e-05, "loss": 0.8346, "step": 3210 }, { "epoch": 0.11629844259326332, "grad_norm": 1.6160328723742279, "learning_rate": 1.9612142583230255e-05, "loss": 0.7864, "step": 3211 }, { "epoch": 0.11633466135458168, "grad_norm": 1.634890586880662, "learning_rate": 1.9611818981808425e-05, "loss": 0.7606, "step": 3212 }, { "epoch": 0.11637088011590004, "grad_norm": 1.7871617293896944, "learning_rate": 1.961149524811952e-05, "loss": 0.8917, "step": 3213 }, { "epoch": 0.1164070988772184, "grad_norm": 1.5916945044240982, "learning_rate": 1.9611171382167998e-05, "loss": 0.8313, "step": 3214 }, { "epoch": 0.11644331763853676, "grad_norm": 1.6519268199768378, "learning_rate": 1.9610847383958308e-05, "loss": 0.8609, "step": 3215 }, { "epoch": 0.11647953639985513, "grad_norm": 1.7676951037735917, "learning_rate": 1.9610523253494915e-05, "loss": 0.9302, "step": 3216 }, { "epoch": 0.11651575516117349, "grad_norm": 1.24245791319386, "learning_rate": 1.9610198990782277e-05, "loss": 0.7603, "step": 3217 }, { "epoch": 0.11655197392249185, "grad_norm": 1.7294860443402462, "learning_rate": 1.9609874595824858e-05, "loss": 0.8924, "step": 3218 }, { "epoch": 0.11658819268381021, "grad_norm": 1.7547748980407276, "learning_rate": 1.960955006862712e-05, "loss": 0.887, "step": 3219 }, { "epoch": 0.11662441144512857, "grad_norm": 1.766453956649833, "learning_rate": 1.9609225409193526e-05, "loss": 0.8875, "step": 3220 }, { "epoch": 0.11666063020644694, "grad_norm": 1.539798063992025, "learning_rate": 1.960890061752855e-05, "loss": 0.9219, "step": 3221 }, { "epoch": 0.1166968489677653, "grad_norm": 1.2379904266918695, "learning_rate": 1.960857569363666e-05, "loss": 0.7825, "step": 3222 }, { "epoch": 0.11673306772908366, "grad_norm": 1.57269773054303, "learning_rate": 1.9608250637522322e-05, "loss": 0.8561, "step": 3223 }, { "epoch": 0.11676928649040202, "grad_norm": 1.7000731129035782, "learning_rate": 1.9607925449190016e-05, "loss": 0.8386, "step": 3224 }, { "epoch": 0.11680550525172038, "grad_norm": 1.2776206495465152, "learning_rate": 1.960760012864421e-05, "loss": 0.7789, "step": 3225 }, { "epoch": 0.11684172401303876, "grad_norm": 1.3467343118231054, "learning_rate": 1.9607274675889388e-05, "loss": 0.7505, "step": 3226 }, { "epoch": 0.11687794277435712, "grad_norm": 1.6320157594068894, "learning_rate": 1.9606949090930025e-05, "loss": 0.8229, "step": 3227 }, { "epoch": 0.11691416153567548, "grad_norm": 1.6538566399919568, "learning_rate": 1.96066233737706e-05, "loss": 0.8933, "step": 3228 }, { "epoch": 0.11695038029699384, "grad_norm": 1.4354871769902957, "learning_rate": 1.9606297524415595e-05, "loss": 0.8041, "step": 3229 }, { "epoch": 0.1169865990583122, "grad_norm": 1.6389175715070432, "learning_rate": 1.9605971542869497e-05, "loss": 0.881, "step": 3230 }, { "epoch": 0.11702281781963057, "grad_norm": 1.6952364647422276, "learning_rate": 1.960564542913679e-05, "loss": 0.9004, "step": 3231 }, { "epoch": 0.11705903658094893, "grad_norm": 3.6907653491908805, "learning_rate": 1.9605319183221956e-05, "loss": 0.7868, "step": 3232 }, { "epoch": 0.1170952553422673, "grad_norm": 1.5870406467129134, "learning_rate": 1.9604992805129493e-05, "loss": 0.8741, "step": 3233 }, { "epoch": 0.11713147410358565, "grad_norm": 1.6632302535782935, "learning_rate": 1.960466629486389e-05, "loss": 0.8416, "step": 3234 }, { "epoch": 0.11716769286490401, "grad_norm": 1.5196532430931462, "learning_rate": 1.9604339652429646e-05, "loss": 0.8776, "step": 3235 }, { "epoch": 0.11720391162622239, "grad_norm": 1.735001983380352, "learning_rate": 1.960401287783124e-05, "loss": 0.9314, "step": 3236 }, { "epoch": 0.11724013038754075, "grad_norm": 1.6045410248364125, "learning_rate": 1.9603685971073183e-05, "loss": 0.7907, "step": 3237 }, { "epoch": 0.11727634914885911, "grad_norm": 1.4459673351757356, "learning_rate": 1.9603358932159966e-05, "loss": 0.9287, "step": 3238 }, { "epoch": 0.11731256791017747, "grad_norm": 1.552258195376914, "learning_rate": 1.9603031761096095e-05, "loss": 0.8333, "step": 3239 }, { "epoch": 0.11734878667149584, "grad_norm": 1.922527816819869, "learning_rate": 1.9602704457886068e-05, "loss": 0.9277, "step": 3240 }, { "epoch": 0.1173850054328142, "grad_norm": 1.517319953098621, "learning_rate": 1.9602377022534384e-05, "loss": 0.8725, "step": 3241 }, { "epoch": 0.11742122419413256, "grad_norm": 1.6176131501467197, "learning_rate": 1.960204945504556e-05, "loss": 0.829, "step": 3242 }, { "epoch": 0.11745744295545092, "grad_norm": 1.5672271170635015, "learning_rate": 1.9601721755424098e-05, "loss": 0.812, "step": 3243 }, { "epoch": 0.11749366171676928, "grad_norm": 1.661779281256222, "learning_rate": 1.960139392367451e-05, "loss": 0.9108, "step": 3244 }, { "epoch": 0.11752988047808766, "grad_norm": 1.8209783161142692, "learning_rate": 1.9601065959801302e-05, "loss": 0.8549, "step": 3245 }, { "epoch": 0.11756609923940602, "grad_norm": 1.6637274557065116, "learning_rate": 1.9600737863808992e-05, "loss": 0.8923, "step": 3246 }, { "epoch": 0.11760231800072438, "grad_norm": 1.7018218610819136, "learning_rate": 1.960040963570209e-05, "loss": 0.8593, "step": 3247 }, { "epoch": 0.11763853676204274, "grad_norm": 1.7278356141251763, "learning_rate": 1.960008127548512e-05, "loss": 0.8597, "step": 3248 }, { "epoch": 0.1176747555233611, "grad_norm": 1.5588519472878977, "learning_rate": 1.9599752783162594e-05, "loss": 0.7831, "step": 3249 }, { "epoch": 0.11771097428467947, "grad_norm": 1.6022155256128616, "learning_rate": 1.9599424158739033e-05, "loss": 0.8652, "step": 3250 }, { "epoch": 0.11774719304599783, "grad_norm": 1.6051515094082176, "learning_rate": 1.9599095402218965e-05, "loss": 0.8971, "step": 3251 }, { "epoch": 0.11778341180731619, "grad_norm": 1.6673954641812387, "learning_rate": 1.9598766513606904e-05, "loss": 0.8815, "step": 3252 }, { "epoch": 0.11781963056863455, "grad_norm": 1.5502474079796331, "learning_rate": 1.9598437492907388e-05, "loss": 0.8822, "step": 3253 }, { "epoch": 0.11785584932995291, "grad_norm": 1.6359861345722246, "learning_rate": 1.9598108340124935e-05, "loss": 0.8888, "step": 3254 }, { "epoch": 0.11789206809127128, "grad_norm": 1.5996696009028575, "learning_rate": 1.959777905526408e-05, "loss": 0.7629, "step": 3255 }, { "epoch": 0.11792828685258964, "grad_norm": 1.7973442449704158, "learning_rate": 1.9597449638329346e-05, "loss": 0.8591, "step": 3256 }, { "epoch": 0.117964505613908, "grad_norm": 1.665962335778527, "learning_rate": 1.9597120089325273e-05, "loss": 0.8131, "step": 3257 }, { "epoch": 0.11800072437522637, "grad_norm": 1.2241561306989903, "learning_rate": 1.95967904082564e-05, "loss": 0.7678, "step": 3258 }, { "epoch": 0.11803694313654473, "grad_norm": 1.6731940685454063, "learning_rate": 1.9596460595127255e-05, "loss": 0.8044, "step": 3259 }, { "epoch": 0.1180731618978631, "grad_norm": 1.6031648441525737, "learning_rate": 1.9596130649942383e-05, "loss": 0.8941, "step": 3260 }, { "epoch": 0.11810938065918146, "grad_norm": 1.8628914363328988, "learning_rate": 1.9595800572706318e-05, "loss": 0.8949, "step": 3261 }, { "epoch": 0.11814559942049982, "grad_norm": 1.39650415976721, "learning_rate": 1.9595470363423606e-05, "loss": 0.7744, "step": 3262 }, { "epoch": 0.11818181818181818, "grad_norm": 1.7318125011096888, "learning_rate": 1.9595140022098792e-05, "loss": 0.8809, "step": 3263 }, { "epoch": 0.11821803694313654, "grad_norm": 1.3091903733820844, "learning_rate": 1.959480954873642e-05, "loss": 0.7997, "step": 3264 }, { "epoch": 0.11825425570445491, "grad_norm": 1.4650475357084238, "learning_rate": 1.9594478943341034e-05, "loss": 0.8486, "step": 3265 }, { "epoch": 0.11829047446577327, "grad_norm": 1.8100739759599909, "learning_rate": 1.9594148205917193e-05, "loss": 0.8655, "step": 3266 }, { "epoch": 0.11832669322709163, "grad_norm": 1.596472174315738, "learning_rate": 1.9593817336469437e-05, "loss": 0.8653, "step": 3267 }, { "epoch": 0.11836291198841, "grad_norm": 1.299481911602988, "learning_rate": 1.959348633500233e-05, "loss": 0.735, "step": 3268 }, { "epoch": 0.11839913074972835, "grad_norm": 1.6310078182162444, "learning_rate": 1.9593155201520418e-05, "loss": 0.7987, "step": 3269 }, { "epoch": 0.11843534951104673, "grad_norm": 1.610740288522507, "learning_rate": 1.9592823936028262e-05, "loss": 0.903, "step": 3270 }, { "epoch": 0.11847156827236509, "grad_norm": 1.5349156322168813, "learning_rate": 1.959249253853042e-05, "loss": 0.8243, "step": 3271 }, { "epoch": 0.11850778703368345, "grad_norm": 1.7781644304551298, "learning_rate": 1.9592161009031453e-05, "loss": 0.9526, "step": 3272 }, { "epoch": 0.11854400579500181, "grad_norm": 1.3179946652794539, "learning_rate": 1.959182934753592e-05, "loss": 0.7818, "step": 3273 }, { "epoch": 0.11858022455632017, "grad_norm": 2.7355773606849314, "learning_rate": 1.959149755404839e-05, "loss": 0.8235, "step": 3274 }, { "epoch": 0.11861644331763854, "grad_norm": 1.8677082981445414, "learning_rate": 1.959116562857342e-05, "loss": 0.9192, "step": 3275 }, { "epoch": 0.1186526620789569, "grad_norm": 1.6274714417065144, "learning_rate": 1.959083357111559e-05, "loss": 0.8459, "step": 3276 }, { "epoch": 0.11868888084027526, "grad_norm": 1.7646408925320012, "learning_rate": 1.959050138167946e-05, "loss": 0.7898, "step": 3277 }, { "epoch": 0.11872509960159362, "grad_norm": 1.7518759814685614, "learning_rate": 1.9590169060269602e-05, "loss": 0.9359, "step": 3278 }, { "epoch": 0.11876131836291198, "grad_norm": 1.5876822029846145, "learning_rate": 1.9589836606890594e-05, "loss": 0.7671, "step": 3279 }, { "epoch": 0.11879753712423036, "grad_norm": 1.7785416375806424, "learning_rate": 1.9589504021547007e-05, "loss": 0.9803, "step": 3280 }, { "epoch": 0.11883375588554872, "grad_norm": 1.7580437868334677, "learning_rate": 1.958917130424342e-05, "loss": 0.8344, "step": 3281 }, { "epoch": 0.11886997464686708, "grad_norm": 1.390872592228709, "learning_rate": 1.958883845498441e-05, "loss": 0.7908, "step": 3282 }, { "epoch": 0.11890619340818544, "grad_norm": 1.8769735360995703, "learning_rate": 1.9588505473774554e-05, "loss": 0.8409, "step": 3283 }, { "epoch": 0.1189424121695038, "grad_norm": 1.7499983140392354, "learning_rate": 1.958817236061844e-05, "loss": 0.9053, "step": 3284 }, { "epoch": 0.11897863093082217, "grad_norm": 2.104366726244406, "learning_rate": 1.958783911552065e-05, "loss": 0.8988, "step": 3285 }, { "epoch": 0.11901484969214053, "grad_norm": 1.8128145372484188, "learning_rate": 1.9587505738485765e-05, "loss": 0.8247, "step": 3286 }, { "epoch": 0.11905106845345889, "grad_norm": 1.5350792309803027, "learning_rate": 1.958717222951838e-05, "loss": 0.8304, "step": 3287 }, { "epoch": 0.11908728721477725, "grad_norm": 1.7644289436482163, "learning_rate": 1.9586838588623084e-05, "loss": 0.8226, "step": 3288 }, { "epoch": 0.11912350597609561, "grad_norm": 2.2381986399746148, "learning_rate": 1.9586504815804463e-05, "loss": 0.7574, "step": 3289 }, { "epoch": 0.11915972473741399, "grad_norm": 1.2626553347138114, "learning_rate": 1.958617091106711e-05, "loss": 0.7527, "step": 3290 }, { "epoch": 0.11919594349873235, "grad_norm": 1.684648829302719, "learning_rate": 1.9585836874415625e-05, "loss": 0.8015, "step": 3291 }, { "epoch": 0.1192321622600507, "grad_norm": 1.6646270409972068, "learning_rate": 1.9585502705854598e-05, "loss": 0.796, "step": 3292 }, { "epoch": 0.11926838102136907, "grad_norm": 1.6829058410079194, "learning_rate": 1.9585168405388636e-05, "loss": 0.8456, "step": 3293 }, { "epoch": 0.11930459978268743, "grad_norm": 1.6381065930435215, "learning_rate": 1.958483397302233e-05, "loss": 0.8855, "step": 3294 }, { "epoch": 0.1193408185440058, "grad_norm": 1.7448438538153075, "learning_rate": 1.9584499408760288e-05, "loss": 0.8331, "step": 3295 }, { "epoch": 0.11937703730532416, "grad_norm": 1.4778531025063881, "learning_rate": 1.9584164712607114e-05, "loss": 0.8551, "step": 3296 }, { "epoch": 0.11941325606664252, "grad_norm": 1.4156113258685858, "learning_rate": 1.9583829884567413e-05, "loss": 0.8285, "step": 3297 }, { "epoch": 0.11944947482796088, "grad_norm": 1.6176928321545418, "learning_rate": 1.9583494924645787e-05, "loss": 0.9519, "step": 3298 }, { "epoch": 0.11948569358927924, "grad_norm": 1.326600927830377, "learning_rate": 1.9583159832846853e-05, "loss": 0.8291, "step": 3299 }, { "epoch": 0.11952191235059761, "grad_norm": 1.6205802230500659, "learning_rate": 1.958282460917522e-05, "loss": 0.8653, "step": 3300 }, { "epoch": 0.11955813111191597, "grad_norm": 1.744690462398935, "learning_rate": 1.9582489253635503e-05, "loss": 0.8222, "step": 3301 }, { "epoch": 0.11959434987323433, "grad_norm": 1.3497222512984215, "learning_rate": 1.9582153766232304e-05, "loss": 0.8264, "step": 3302 }, { "epoch": 0.1196305686345527, "grad_norm": 1.7926212027970836, "learning_rate": 1.958181814697026e-05, "loss": 0.8856, "step": 3303 }, { "epoch": 0.11966678739587105, "grad_norm": 1.6991457824825202, "learning_rate": 1.9581482395853973e-05, "loss": 0.8765, "step": 3304 }, { "epoch": 0.11970300615718943, "grad_norm": 1.5316661213350917, "learning_rate": 1.9581146512888072e-05, "loss": 0.8162, "step": 3305 }, { "epoch": 0.11973922491850779, "grad_norm": 1.617609338751698, "learning_rate": 1.9580810498077177e-05, "loss": 0.8225, "step": 3306 }, { "epoch": 0.11977544367982615, "grad_norm": 2.0196492199410865, "learning_rate": 1.9580474351425905e-05, "loss": 0.8633, "step": 3307 }, { "epoch": 0.11981166244114451, "grad_norm": 1.790478422996171, "learning_rate": 1.9580138072938895e-05, "loss": 0.8942, "step": 3308 }, { "epoch": 0.11984788120246287, "grad_norm": 1.6338625114425624, "learning_rate": 1.9579801662620764e-05, "loss": 0.9121, "step": 3309 }, { "epoch": 0.11988409996378124, "grad_norm": 1.5638989670784278, "learning_rate": 1.9579465120476144e-05, "loss": 0.8071, "step": 3310 }, { "epoch": 0.1199203187250996, "grad_norm": 1.369590356950664, "learning_rate": 1.957912844650967e-05, "loss": 0.7264, "step": 3311 }, { "epoch": 0.11995653748641796, "grad_norm": 1.7374576771421955, "learning_rate": 1.957879164072597e-05, "loss": 0.8182, "step": 3312 }, { "epoch": 0.11999275624773632, "grad_norm": 1.6480095452463634, "learning_rate": 1.957845470312968e-05, "loss": 0.9004, "step": 3313 }, { "epoch": 0.12002897500905468, "grad_norm": 1.7321047315207951, "learning_rate": 1.9578117633725434e-05, "loss": 0.885, "step": 3314 }, { "epoch": 0.12006519377037306, "grad_norm": 1.7396080065070123, "learning_rate": 1.957778043251788e-05, "loss": 0.9291, "step": 3315 }, { "epoch": 0.12010141253169142, "grad_norm": 1.5772647856086055, "learning_rate": 1.9577443099511644e-05, "loss": 0.8769, "step": 3316 }, { "epoch": 0.12013763129300978, "grad_norm": 1.651856653865397, "learning_rate": 1.957710563471138e-05, "loss": 0.8372, "step": 3317 }, { "epoch": 0.12017385005432814, "grad_norm": 1.1769723422628853, "learning_rate": 1.957676803812173e-05, "loss": 0.7361, "step": 3318 }, { "epoch": 0.1202100688156465, "grad_norm": 1.6204430342710072, "learning_rate": 1.9576430309747328e-05, "loss": 0.8892, "step": 3319 }, { "epoch": 0.12024628757696487, "grad_norm": 1.3510750379204894, "learning_rate": 1.9576092449592835e-05, "loss": 0.7797, "step": 3320 }, { "epoch": 0.12028250633828323, "grad_norm": 1.6592251780955085, "learning_rate": 1.9575754457662897e-05, "loss": 0.9543, "step": 3321 }, { "epoch": 0.12031872509960159, "grad_norm": 1.547989861359881, "learning_rate": 1.9575416333962163e-05, "loss": 0.8684, "step": 3322 }, { "epoch": 0.12035494386091995, "grad_norm": 1.202250752967933, "learning_rate": 1.9575078078495287e-05, "loss": 0.7993, "step": 3323 }, { "epoch": 0.12039116262223833, "grad_norm": 1.7611914075935817, "learning_rate": 1.9574739691266923e-05, "loss": 0.9127, "step": 3324 }, { "epoch": 0.12042738138355669, "grad_norm": 2.014041683527441, "learning_rate": 1.9574401172281727e-05, "loss": 0.8482, "step": 3325 }, { "epoch": 0.12046360014487505, "grad_norm": 1.6404901591020404, "learning_rate": 1.9574062521544358e-05, "loss": 0.8649, "step": 3326 }, { "epoch": 0.1204998189061934, "grad_norm": 1.683456079492212, "learning_rate": 1.9573723739059477e-05, "loss": 0.8926, "step": 3327 }, { "epoch": 0.12053603766751177, "grad_norm": 1.661260446335503, "learning_rate": 1.9573384824831744e-05, "loss": 0.8802, "step": 3328 }, { "epoch": 0.12057225642883014, "grad_norm": 1.7399516823293826, "learning_rate": 1.9573045778865823e-05, "loss": 0.8707, "step": 3329 }, { "epoch": 0.1206084751901485, "grad_norm": 1.6178159661639986, "learning_rate": 1.9572706601166384e-05, "loss": 0.9198, "step": 3330 }, { "epoch": 0.12064469395146686, "grad_norm": 1.8482343457872046, "learning_rate": 1.957236729173809e-05, "loss": 0.9424, "step": 3331 }, { "epoch": 0.12068091271278522, "grad_norm": 1.1645660937200255, "learning_rate": 1.9572027850585606e-05, "loss": 0.7617, "step": 3332 }, { "epoch": 0.12071713147410358, "grad_norm": 1.4938153163655945, "learning_rate": 1.9571688277713614e-05, "loss": 0.7849, "step": 3333 }, { "epoch": 0.12075335023542196, "grad_norm": 1.3365837450858065, "learning_rate": 1.957134857312678e-05, "loss": 0.8282, "step": 3334 }, { "epoch": 0.12078956899674032, "grad_norm": 1.5606336706752852, "learning_rate": 1.9571008736829778e-05, "loss": 0.7811, "step": 3335 }, { "epoch": 0.12082578775805868, "grad_norm": 1.6021128569496441, "learning_rate": 1.9570668768827284e-05, "loss": 0.8747, "step": 3336 }, { "epoch": 0.12086200651937704, "grad_norm": 1.491651440462434, "learning_rate": 1.957032866912398e-05, "loss": 0.7749, "step": 3337 }, { "epoch": 0.1208982252806954, "grad_norm": 1.8895581403740112, "learning_rate": 1.9569988437724548e-05, "loss": 0.9159, "step": 3338 }, { "epoch": 0.12093444404201377, "grad_norm": 1.2988331160055078, "learning_rate": 1.9569648074633663e-05, "loss": 0.7473, "step": 3339 }, { "epoch": 0.12097066280333213, "grad_norm": 1.7412686367819277, "learning_rate": 1.956930757985601e-05, "loss": 0.8579, "step": 3340 }, { "epoch": 0.12100688156465049, "grad_norm": 1.6087409501906775, "learning_rate": 1.956896695339628e-05, "loss": 0.8485, "step": 3341 }, { "epoch": 0.12104310032596885, "grad_norm": 1.6414640943211118, "learning_rate": 1.9568626195259152e-05, "loss": 0.8129, "step": 3342 }, { "epoch": 0.12107931908728721, "grad_norm": 1.6165116148560745, "learning_rate": 1.9568285305449323e-05, "loss": 0.8262, "step": 3343 }, { "epoch": 0.12111553784860558, "grad_norm": 1.747169820170365, "learning_rate": 1.956794428397148e-05, "loss": 0.884, "step": 3344 }, { "epoch": 0.12115175660992394, "grad_norm": 1.7161333919869144, "learning_rate": 1.9567603130830317e-05, "loss": 0.8518, "step": 3345 }, { "epoch": 0.1211879753712423, "grad_norm": 1.543027025643336, "learning_rate": 1.956726184603053e-05, "loss": 0.9135, "step": 3346 }, { "epoch": 0.12122419413256066, "grad_norm": 1.6660288243369032, "learning_rate": 1.956692042957681e-05, "loss": 0.8213, "step": 3347 }, { "epoch": 0.12126041289387902, "grad_norm": 1.663071945364528, "learning_rate": 1.9566578881473862e-05, "loss": 0.8025, "step": 3348 }, { "epoch": 0.1212966316551974, "grad_norm": 1.6924984447248121, "learning_rate": 1.9566237201726382e-05, "loss": 0.8346, "step": 3349 }, { "epoch": 0.12133285041651576, "grad_norm": 1.4563392460882767, "learning_rate": 1.956589539033907e-05, "loss": 0.8095, "step": 3350 }, { "epoch": 0.12136906917783412, "grad_norm": 1.533437368394105, "learning_rate": 1.956555344731663e-05, "loss": 0.8387, "step": 3351 }, { "epoch": 0.12140528793915248, "grad_norm": 1.8902634913454086, "learning_rate": 1.9565211372663774e-05, "loss": 0.93, "step": 3352 }, { "epoch": 0.12144150670047084, "grad_norm": 1.8035309541185902, "learning_rate": 1.9564869166385204e-05, "loss": 0.8264, "step": 3353 }, { "epoch": 0.12147772546178921, "grad_norm": 1.7755637373001651, "learning_rate": 1.9564526828485633e-05, "loss": 0.9917, "step": 3354 }, { "epoch": 0.12151394422310757, "grad_norm": 1.222008666085869, "learning_rate": 1.9564184358969762e-05, "loss": 0.8427, "step": 3355 }, { "epoch": 0.12155016298442593, "grad_norm": 1.1620582889900566, "learning_rate": 1.9563841757842315e-05, "loss": 0.7909, "step": 3356 }, { "epoch": 0.12158638174574429, "grad_norm": 1.8741428025136135, "learning_rate": 1.9563499025108e-05, "loss": 0.856, "step": 3357 }, { "epoch": 0.12162260050706265, "grad_norm": 1.6762695101374896, "learning_rate": 1.9563156160771538e-05, "loss": 0.9483, "step": 3358 }, { "epoch": 0.12165881926838103, "grad_norm": 1.722277060012423, "learning_rate": 1.956281316483764e-05, "loss": 0.8406, "step": 3359 }, { "epoch": 0.12169503802969939, "grad_norm": 1.8336760405964116, "learning_rate": 1.9562470037311035e-05, "loss": 0.8014, "step": 3360 }, { "epoch": 0.12173125679101775, "grad_norm": 1.5873341377035601, "learning_rate": 1.9562126778196434e-05, "loss": 0.8595, "step": 3361 }, { "epoch": 0.12176747555233611, "grad_norm": 1.1465397766356653, "learning_rate": 1.9561783387498573e-05, "loss": 0.7808, "step": 3362 }, { "epoch": 0.12180369431365447, "grad_norm": 1.7434860167254396, "learning_rate": 1.9561439865222165e-05, "loss": 0.8159, "step": 3363 }, { "epoch": 0.12183991307497284, "grad_norm": 1.7096888795779808, "learning_rate": 1.9561096211371946e-05, "loss": 0.8798, "step": 3364 }, { "epoch": 0.1218761318362912, "grad_norm": 1.8988048370760182, "learning_rate": 1.956075242595264e-05, "loss": 0.8541, "step": 3365 }, { "epoch": 0.12191235059760956, "grad_norm": 1.3842300768591833, "learning_rate": 1.956040850896898e-05, "loss": 0.8193, "step": 3366 }, { "epoch": 0.12194856935892792, "grad_norm": 2.067474889349061, "learning_rate": 1.9560064460425705e-05, "loss": 0.9199, "step": 3367 }, { "epoch": 0.12198478812024628, "grad_norm": 1.3629449475496447, "learning_rate": 1.9559720280327536e-05, "loss": 0.7981, "step": 3368 }, { "epoch": 0.12202100688156466, "grad_norm": 1.51271570831005, "learning_rate": 1.9559375968679217e-05, "loss": 0.8324, "step": 3369 }, { "epoch": 0.12205722564288302, "grad_norm": 1.7633818538970911, "learning_rate": 1.9559031525485486e-05, "loss": 0.8099, "step": 3370 }, { "epoch": 0.12209344440420138, "grad_norm": 1.579008708339988, "learning_rate": 1.9558686950751083e-05, "loss": 0.8296, "step": 3371 }, { "epoch": 0.12212966316551974, "grad_norm": 1.6540910847289574, "learning_rate": 1.955834224448075e-05, "loss": 0.8632, "step": 3372 }, { "epoch": 0.1221658819268381, "grad_norm": 1.6884572925427763, "learning_rate": 1.9557997406679224e-05, "loss": 0.8154, "step": 3373 }, { "epoch": 0.12220210068815647, "grad_norm": 1.2397345810558849, "learning_rate": 1.9557652437351255e-05, "loss": 0.8442, "step": 3374 }, { "epoch": 0.12223831944947483, "grad_norm": 1.6269021192251176, "learning_rate": 1.9557307336501595e-05, "loss": 0.8987, "step": 3375 }, { "epoch": 0.12227453821079319, "grad_norm": 1.2460221224037287, "learning_rate": 1.9556962104134985e-05, "loss": 0.8058, "step": 3376 }, { "epoch": 0.12231075697211155, "grad_norm": 1.4996661737381916, "learning_rate": 1.9556616740256183e-05, "loss": 0.8488, "step": 3377 }, { "epoch": 0.12234697573342991, "grad_norm": 1.8799938377193142, "learning_rate": 1.9556271244869932e-05, "loss": 0.8384, "step": 3378 }, { "epoch": 0.12238319449474828, "grad_norm": 1.301786653674229, "learning_rate": 1.9555925617980997e-05, "loss": 0.8061, "step": 3379 }, { "epoch": 0.12241941325606664, "grad_norm": 1.6557452295377186, "learning_rate": 1.9555579859594126e-05, "loss": 0.8215, "step": 3380 }, { "epoch": 0.122455632017385, "grad_norm": 1.2520571487299954, "learning_rate": 1.9555233969714078e-05, "loss": 0.7647, "step": 3381 }, { "epoch": 0.12249185077870337, "grad_norm": 1.614723035834043, "learning_rate": 1.955488794834562e-05, "loss": 0.7826, "step": 3382 }, { "epoch": 0.12252806954002173, "grad_norm": 1.2222068591260764, "learning_rate": 1.9554541795493502e-05, "loss": 0.7905, "step": 3383 }, { "epoch": 0.1225642883013401, "grad_norm": 1.2314077779708898, "learning_rate": 1.9554195511162496e-05, "loss": 0.7803, "step": 3384 }, { "epoch": 0.12260050706265846, "grad_norm": 1.8261786843239212, "learning_rate": 1.9553849095357366e-05, "loss": 0.8307, "step": 3385 }, { "epoch": 0.12263672582397682, "grad_norm": 1.6668706245100324, "learning_rate": 1.9553502548082878e-05, "loss": 0.7931, "step": 3386 }, { "epoch": 0.12267294458529518, "grad_norm": 1.5189086483138956, "learning_rate": 1.95531558693438e-05, "loss": 0.8997, "step": 3387 }, { "epoch": 0.12270916334661354, "grad_norm": 1.3608205555759219, "learning_rate": 1.95528090591449e-05, "loss": 0.7908, "step": 3388 }, { "epoch": 0.12274538210793191, "grad_norm": 1.7957085306261746, "learning_rate": 1.9552462117490956e-05, "loss": 0.8468, "step": 3389 }, { "epoch": 0.12278160086925027, "grad_norm": 2.2060185615651586, "learning_rate": 1.9552115044386742e-05, "loss": 0.9037, "step": 3390 }, { "epoch": 0.12281781963056863, "grad_norm": 1.6446181447637611, "learning_rate": 1.9551767839837027e-05, "loss": 0.8727, "step": 3391 }, { "epoch": 0.122854038391887, "grad_norm": 1.22771582273493, "learning_rate": 1.95514205038466e-05, "loss": 0.8256, "step": 3392 }, { "epoch": 0.12289025715320535, "grad_norm": 1.8986951668841097, "learning_rate": 1.955107303642023e-05, "loss": 0.905, "step": 3393 }, { "epoch": 0.12292647591452373, "grad_norm": 1.6571390761670837, "learning_rate": 1.9550725437562703e-05, "loss": 0.8746, "step": 3394 }, { "epoch": 0.12296269467584209, "grad_norm": 1.633390122551778, "learning_rate": 1.95503777072788e-05, "loss": 0.8668, "step": 3395 }, { "epoch": 0.12299891343716045, "grad_norm": 1.659377252470345, "learning_rate": 1.9550029845573313e-05, "loss": 0.881, "step": 3396 }, { "epoch": 0.12303513219847881, "grad_norm": 1.667254658010485, "learning_rate": 1.9549681852451022e-05, "loss": 0.8763, "step": 3397 }, { "epoch": 0.12307135095979717, "grad_norm": 1.521480025308486, "learning_rate": 1.9549333727916716e-05, "loss": 0.9012, "step": 3398 }, { "epoch": 0.12310756972111554, "grad_norm": 1.3123101193055453, "learning_rate": 1.9548985471975186e-05, "loss": 0.7851, "step": 3399 }, { "epoch": 0.1231437884824339, "grad_norm": 1.7082947631199585, "learning_rate": 1.9548637084631228e-05, "loss": 0.822, "step": 3400 }, { "epoch": 0.12318000724375226, "grad_norm": 1.21851022599557, "learning_rate": 1.954828856588963e-05, "loss": 0.8349, "step": 3401 }, { "epoch": 0.12321622600507062, "grad_norm": 1.340694099085739, "learning_rate": 1.9547939915755196e-05, "loss": 0.8255, "step": 3402 }, { "epoch": 0.12325244476638898, "grad_norm": 1.54870300100234, "learning_rate": 1.954759113423272e-05, "loss": 0.8459, "step": 3403 }, { "epoch": 0.12328866352770736, "grad_norm": 1.5091472243832058, "learning_rate": 1.9547242221326995e-05, "loss": 0.8618, "step": 3404 }, { "epoch": 0.12332488228902572, "grad_norm": 1.5657081552653767, "learning_rate": 1.9546893177042834e-05, "loss": 0.8689, "step": 3405 }, { "epoch": 0.12336110105034408, "grad_norm": 1.5626748940814883, "learning_rate": 1.954654400138503e-05, "loss": 0.8034, "step": 3406 }, { "epoch": 0.12339731981166244, "grad_norm": 1.6395638337837886, "learning_rate": 1.9546194694358392e-05, "loss": 0.7961, "step": 3407 }, { "epoch": 0.1234335385729808, "grad_norm": 1.550904302735601, "learning_rate": 1.954584525596773e-05, "loss": 0.7975, "step": 3408 }, { "epoch": 0.12346975733429917, "grad_norm": 1.6693850651360032, "learning_rate": 1.954549568621785e-05, "loss": 0.8744, "step": 3409 }, { "epoch": 0.12350597609561753, "grad_norm": 1.6588420041555165, "learning_rate": 1.954514598511356e-05, "loss": 0.895, "step": 3410 }, { "epoch": 0.12354219485693589, "grad_norm": 1.284536725440801, "learning_rate": 1.9544796152659676e-05, "loss": 0.7872, "step": 3411 }, { "epoch": 0.12357841361825425, "grad_norm": 1.6477487801055128, "learning_rate": 1.954444618886101e-05, "loss": 0.8493, "step": 3412 }, { "epoch": 0.12361463237957263, "grad_norm": 1.2643431421123783, "learning_rate": 1.9544096093722373e-05, "loss": 0.7632, "step": 3413 }, { "epoch": 0.12365085114089099, "grad_norm": 1.3788438623234363, "learning_rate": 1.954374586724859e-05, "loss": 0.7288, "step": 3414 }, { "epoch": 0.12368706990220935, "grad_norm": 1.5553424835687386, "learning_rate": 1.954339550944448e-05, "loss": 0.8468, "step": 3415 }, { "epoch": 0.1237232886635277, "grad_norm": 1.7120364069810943, "learning_rate": 1.9543045020314864e-05, "loss": 0.8466, "step": 3416 }, { "epoch": 0.12375950742484607, "grad_norm": 1.646638714806928, "learning_rate": 1.9542694399864558e-05, "loss": 0.8809, "step": 3417 }, { "epoch": 0.12379572618616444, "grad_norm": 1.6315061508776485, "learning_rate": 1.9542343648098395e-05, "loss": 0.8258, "step": 3418 }, { "epoch": 0.1238319449474828, "grad_norm": 1.351729740113964, "learning_rate": 1.95419927650212e-05, "loss": 0.8423, "step": 3419 }, { "epoch": 0.12386816370880116, "grad_norm": 1.543872117995461, "learning_rate": 1.9541641750637797e-05, "loss": 0.8356, "step": 3420 }, { "epoch": 0.12390438247011952, "grad_norm": 1.3733668030065647, "learning_rate": 1.9541290604953023e-05, "loss": 0.8385, "step": 3421 }, { "epoch": 0.12394060123143788, "grad_norm": 1.7076091399227338, "learning_rate": 1.9540939327971707e-05, "loss": 0.891, "step": 3422 }, { "epoch": 0.12397681999275625, "grad_norm": 1.645244168631403, "learning_rate": 1.954058791969868e-05, "loss": 0.8184, "step": 3423 }, { "epoch": 0.12401303875407461, "grad_norm": 1.7158663281613924, "learning_rate": 1.9540236380138776e-05, "loss": 0.8437, "step": 3424 }, { "epoch": 0.12404925751539297, "grad_norm": 1.4928009573760428, "learning_rate": 1.9539884709296844e-05, "loss": 0.8402, "step": 3425 }, { "epoch": 0.12408547627671133, "grad_norm": 1.854038828434744, "learning_rate": 1.9539532907177713e-05, "loss": 0.8026, "step": 3426 }, { "epoch": 0.1241216950380297, "grad_norm": 1.8039296724443503, "learning_rate": 1.9539180973786223e-05, "loss": 0.8972, "step": 3427 }, { "epoch": 0.12415791379934807, "grad_norm": 1.3133298770722708, "learning_rate": 1.953882890912723e-05, "loss": 0.8016, "step": 3428 }, { "epoch": 0.12419413256066643, "grad_norm": 1.6555227952680023, "learning_rate": 1.9538476713205564e-05, "loss": 0.9325, "step": 3429 }, { "epoch": 0.12423035132198479, "grad_norm": 1.677734120760844, "learning_rate": 1.9538124386026078e-05, "loss": 0.8432, "step": 3430 }, { "epoch": 0.12426657008330315, "grad_norm": 1.5765140787114007, "learning_rate": 1.953777192759362e-05, "loss": 0.9235, "step": 3431 }, { "epoch": 0.12430278884462151, "grad_norm": 1.5753378558241855, "learning_rate": 1.9537419337913036e-05, "loss": 0.8503, "step": 3432 }, { "epoch": 0.12433900760593988, "grad_norm": 1.8677214573280105, "learning_rate": 1.9537066616989183e-05, "loss": 0.9213, "step": 3433 }, { "epoch": 0.12437522636725824, "grad_norm": 1.622444898471652, "learning_rate": 1.9536713764826917e-05, "loss": 0.9664, "step": 3434 }, { "epoch": 0.1244114451285766, "grad_norm": 1.6202674241909585, "learning_rate": 1.953636078143109e-05, "loss": 0.795, "step": 3435 }, { "epoch": 0.12444766388989496, "grad_norm": 1.6992789011207212, "learning_rate": 1.9536007666806555e-05, "loss": 0.9182, "step": 3436 }, { "epoch": 0.12448388265121332, "grad_norm": 1.386134718254954, "learning_rate": 1.9535654420958177e-05, "loss": 0.8317, "step": 3437 }, { "epoch": 0.1245201014125317, "grad_norm": 1.885633452652246, "learning_rate": 1.9535301043890817e-05, "loss": 0.8276, "step": 3438 }, { "epoch": 0.12455632017385006, "grad_norm": 1.6084886536024594, "learning_rate": 1.9534947535609334e-05, "loss": 0.7867, "step": 3439 }, { "epoch": 0.12459253893516842, "grad_norm": 1.2862167384420027, "learning_rate": 1.9534593896118596e-05, "loss": 0.7827, "step": 3440 }, { "epoch": 0.12462875769648678, "grad_norm": 1.7820174170278704, "learning_rate": 1.9534240125423468e-05, "loss": 0.8386, "step": 3441 }, { "epoch": 0.12466497645780514, "grad_norm": 1.6218370376675426, "learning_rate": 1.9533886223528818e-05, "loss": 0.7884, "step": 3442 }, { "epoch": 0.12470119521912351, "grad_norm": 1.464524516447355, "learning_rate": 1.9533532190439518e-05, "loss": 0.8086, "step": 3443 }, { "epoch": 0.12473741398044187, "grad_norm": 1.4988435481980233, "learning_rate": 1.9533178026160438e-05, "loss": 0.78, "step": 3444 }, { "epoch": 0.12477363274176023, "grad_norm": 1.6774933101141176, "learning_rate": 1.9532823730696448e-05, "loss": 0.9021, "step": 3445 }, { "epoch": 0.12480985150307859, "grad_norm": 1.6468424515158882, "learning_rate": 1.9532469304052432e-05, "loss": 0.864, "step": 3446 }, { "epoch": 0.12484607026439695, "grad_norm": 1.712047264892507, "learning_rate": 1.953211474623326e-05, "loss": 0.874, "step": 3447 }, { "epoch": 0.12488228902571533, "grad_norm": 1.5300970395778972, "learning_rate": 1.953176005724381e-05, "loss": 0.9265, "step": 3448 }, { "epoch": 0.12491850778703369, "grad_norm": 1.7113976676253457, "learning_rate": 1.953140523708897e-05, "loss": 0.8795, "step": 3449 }, { "epoch": 0.12495472654835205, "grad_norm": 1.8765808037874467, "learning_rate": 1.9531050285773622e-05, "loss": 0.8029, "step": 3450 }, { "epoch": 0.1249909453096704, "grad_norm": 1.2217888613693024, "learning_rate": 1.9530695203302645e-05, "loss": 0.7658, "step": 3451 }, { "epoch": 0.12502716407098877, "grad_norm": 1.7137563685918624, "learning_rate": 1.9530339989680927e-05, "loss": 0.8662, "step": 3452 }, { "epoch": 0.12506338283230714, "grad_norm": 1.6228917273815768, "learning_rate": 1.9529984644913356e-05, "loss": 0.8235, "step": 3453 }, { "epoch": 0.1250996015936255, "grad_norm": 1.3921613862176614, "learning_rate": 1.952962916900482e-05, "loss": 0.8247, "step": 3454 }, { "epoch": 0.12513582035494386, "grad_norm": 1.6281081263753265, "learning_rate": 1.952927356196022e-05, "loss": 0.9271, "step": 3455 }, { "epoch": 0.12517203911626223, "grad_norm": 1.5999802350475625, "learning_rate": 1.9528917823784436e-05, "loss": 0.8635, "step": 3456 }, { "epoch": 0.12520825787758058, "grad_norm": 1.6454203302237667, "learning_rate": 1.9528561954482375e-05, "loss": 0.8924, "step": 3457 }, { "epoch": 0.12524447663889895, "grad_norm": 1.4990824262214355, "learning_rate": 1.952820595405893e-05, "loss": 0.8077, "step": 3458 }, { "epoch": 0.1252806954002173, "grad_norm": 1.7325696183332229, "learning_rate": 1.952784982251899e-05, "loss": 0.8138, "step": 3459 }, { "epoch": 0.12531691416153568, "grad_norm": 1.2002629366959838, "learning_rate": 1.9527493559867473e-05, "loss": 0.7758, "step": 3460 }, { "epoch": 0.12535313292285405, "grad_norm": 1.7095391533782858, "learning_rate": 1.952713716610927e-05, "loss": 0.8332, "step": 3461 }, { "epoch": 0.1253893516841724, "grad_norm": 2.0871751842372452, "learning_rate": 1.9526780641249287e-05, "loss": 0.8208, "step": 3462 }, { "epoch": 0.12542557044549077, "grad_norm": 1.6694237016159958, "learning_rate": 1.9526423985292434e-05, "loss": 0.7735, "step": 3463 }, { "epoch": 0.12546178920680912, "grad_norm": 1.5871163840799856, "learning_rate": 1.9526067198243615e-05, "loss": 0.8764, "step": 3464 }, { "epoch": 0.1254980079681275, "grad_norm": 2.0458851324481833, "learning_rate": 1.952571028010774e-05, "loss": 0.8231, "step": 3465 }, { "epoch": 0.12553422672944586, "grad_norm": 2.5695826253629184, "learning_rate": 1.9525353230889722e-05, "loss": 0.8444, "step": 3466 }, { "epoch": 0.1255704454907642, "grad_norm": 1.7072138164931692, "learning_rate": 1.9524996050594477e-05, "loss": 0.8771, "step": 3467 }, { "epoch": 0.12560666425208258, "grad_norm": 1.391016059810878, "learning_rate": 1.9524638739226914e-05, "loss": 0.8125, "step": 3468 }, { "epoch": 0.12564288301340093, "grad_norm": 1.1877521471496646, "learning_rate": 1.9524281296791952e-05, "loss": 0.8192, "step": 3469 }, { "epoch": 0.1256791017747193, "grad_norm": 1.6899508652619366, "learning_rate": 1.9523923723294513e-05, "loss": 0.8297, "step": 3470 }, { "epoch": 0.12571532053603768, "grad_norm": 1.7710857405479326, "learning_rate": 1.9523566018739514e-05, "loss": 0.832, "step": 3471 }, { "epoch": 0.12575153929735602, "grad_norm": 1.7060641150386024, "learning_rate": 1.952320818313188e-05, "loss": 0.8282, "step": 3472 }, { "epoch": 0.1257877580586744, "grad_norm": 1.246290287076405, "learning_rate": 1.952285021647653e-05, "loss": 0.8658, "step": 3473 }, { "epoch": 0.12582397681999274, "grad_norm": 1.574824499435921, "learning_rate": 1.9522492118778396e-05, "loss": 0.8479, "step": 3474 }, { "epoch": 0.12586019558131112, "grad_norm": 1.6863119349476785, "learning_rate": 1.95221338900424e-05, "loss": 0.8279, "step": 3475 }, { "epoch": 0.1258964143426295, "grad_norm": 1.232895439645541, "learning_rate": 1.952177553027348e-05, "loss": 0.7914, "step": 3476 }, { "epoch": 0.12593263310394784, "grad_norm": 1.5527937511871415, "learning_rate": 1.952141703947656e-05, "loss": 0.7706, "step": 3477 }, { "epoch": 0.1259688518652662, "grad_norm": 1.5498298716286674, "learning_rate": 1.9521058417656575e-05, "loss": 0.8722, "step": 3478 }, { "epoch": 0.12600507062658456, "grad_norm": 1.4903966439120548, "learning_rate": 1.9520699664818462e-05, "loss": 0.8626, "step": 3479 }, { "epoch": 0.12604128938790293, "grad_norm": 1.6923785795675037, "learning_rate": 1.9520340780967152e-05, "loss": 0.8431, "step": 3480 }, { "epoch": 0.1260775081492213, "grad_norm": 1.6313287596306052, "learning_rate": 1.9519981766107593e-05, "loss": 0.7863, "step": 3481 }, { "epoch": 0.12611372691053965, "grad_norm": 1.7767467187705177, "learning_rate": 1.9519622620244716e-05, "loss": 0.925, "step": 3482 }, { "epoch": 0.12614994567185803, "grad_norm": 1.6268897553704262, "learning_rate": 1.9519263343383467e-05, "loss": 0.8634, "step": 3483 }, { "epoch": 0.12618616443317637, "grad_norm": 1.644406814579158, "learning_rate": 1.951890393552879e-05, "loss": 0.8284, "step": 3484 }, { "epoch": 0.12622238319449475, "grad_norm": 1.6665954888811068, "learning_rate": 1.951854439668563e-05, "loss": 0.9331, "step": 3485 }, { "epoch": 0.12625860195581312, "grad_norm": 1.5839750075621848, "learning_rate": 1.951818472685894e-05, "loss": 0.8714, "step": 3486 }, { "epoch": 0.12629482071713147, "grad_norm": 1.7273437040200437, "learning_rate": 1.951782492605366e-05, "loss": 0.877, "step": 3487 }, { "epoch": 0.12633103947844984, "grad_norm": 1.6698498135907534, "learning_rate": 1.951746499427475e-05, "loss": 0.8832, "step": 3488 }, { "epoch": 0.1263672582397682, "grad_norm": 1.385496015547217, "learning_rate": 1.9517104931527154e-05, "loss": 0.7254, "step": 3489 }, { "epoch": 0.12640347700108656, "grad_norm": 1.8591943941637714, "learning_rate": 1.9516744737815837e-05, "loss": 0.8578, "step": 3490 }, { "epoch": 0.12643969576240494, "grad_norm": 1.182424455852704, "learning_rate": 1.9516384413145746e-05, "loss": 0.7301, "step": 3491 }, { "epoch": 0.12647591452372328, "grad_norm": 1.6669294348840547, "learning_rate": 1.9516023957521846e-05, "loss": 0.7931, "step": 3492 }, { "epoch": 0.12651213328504166, "grad_norm": 1.752183586450965, "learning_rate": 1.9515663370949093e-05, "loss": 0.8264, "step": 3493 }, { "epoch": 0.12654835204636, "grad_norm": 1.7495758234938985, "learning_rate": 1.9515302653432452e-05, "loss": 0.8126, "step": 3494 }, { "epoch": 0.12658457080767838, "grad_norm": 1.370358874700903, "learning_rate": 1.9514941804976885e-05, "loss": 0.8054, "step": 3495 }, { "epoch": 0.12662078956899675, "grad_norm": 1.6388058118905577, "learning_rate": 1.951458082558736e-05, "loss": 0.7865, "step": 3496 }, { "epoch": 0.1266570083303151, "grad_norm": 1.62871684830299, "learning_rate": 1.951421971526884e-05, "loss": 0.7898, "step": 3497 }, { "epoch": 0.12669322709163347, "grad_norm": 1.5770925948942103, "learning_rate": 1.95138584740263e-05, "loss": 0.8632, "step": 3498 }, { "epoch": 0.12672944585295182, "grad_norm": 1.7874838056364268, "learning_rate": 1.9513497101864705e-05, "loss": 0.8291, "step": 3499 }, { "epoch": 0.1267656646142702, "grad_norm": 1.8368179959329798, "learning_rate": 1.951313559878903e-05, "loss": 0.9422, "step": 3500 }, { "epoch": 0.12680188337558856, "grad_norm": 1.6147694662436838, "learning_rate": 1.9512773964804252e-05, "loss": 0.7949, "step": 3501 }, { "epoch": 0.1268381021369069, "grad_norm": 1.6148332976587545, "learning_rate": 1.9512412199915343e-05, "loss": 0.8323, "step": 3502 }, { "epoch": 0.12687432089822528, "grad_norm": 1.4557349546677971, "learning_rate": 1.9512050304127287e-05, "loss": 0.8203, "step": 3503 }, { "epoch": 0.12691053965954363, "grad_norm": 1.5789989433373368, "learning_rate": 1.951168827744506e-05, "loss": 0.8526, "step": 3504 }, { "epoch": 0.126946758420862, "grad_norm": 1.621785443742221, "learning_rate": 1.951132611987365e-05, "loss": 0.9062, "step": 3505 }, { "epoch": 0.12698297718218038, "grad_norm": 1.742201221309393, "learning_rate": 1.9510963831418026e-05, "loss": 0.9156, "step": 3506 }, { "epoch": 0.12701919594349872, "grad_norm": 2.055395200276893, "learning_rate": 1.9510601412083184e-05, "loss": 0.9216, "step": 3507 }, { "epoch": 0.1270554147048171, "grad_norm": 1.726924769771256, "learning_rate": 1.9510238861874112e-05, "loss": 0.8739, "step": 3508 }, { "epoch": 0.12709163346613545, "grad_norm": 1.655492954833298, "learning_rate": 1.95098761807958e-05, "loss": 0.7697, "step": 3509 }, { "epoch": 0.12712785222745382, "grad_norm": 1.6806341409097716, "learning_rate": 1.9509513368853232e-05, "loss": 0.8448, "step": 3510 }, { "epoch": 0.1271640709887722, "grad_norm": 1.671338086503101, "learning_rate": 1.9509150426051405e-05, "loss": 0.8118, "step": 3511 }, { "epoch": 0.12720028975009054, "grad_norm": 1.5644599673899013, "learning_rate": 1.950878735239531e-05, "loss": 0.8651, "step": 3512 }, { "epoch": 0.1272365085114089, "grad_norm": 1.6754670771176285, "learning_rate": 1.950842414788995e-05, "loss": 0.9092, "step": 3513 }, { "epoch": 0.12727272727272726, "grad_norm": 1.7405597904838577, "learning_rate": 1.9508060812540317e-05, "loss": 0.9361, "step": 3514 }, { "epoch": 0.12730894603404563, "grad_norm": 2.099984505431597, "learning_rate": 1.9507697346351414e-05, "loss": 0.8202, "step": 3515 }, { "epoch": 0.127345164795364, "grad_norm": 1.6192456718668649, "learning_rate": 1.950733374932824e-05, "loss": 0.7994, "step": 3516 }, { "epoch": 0.12738138355668235, "grad_norm": 1.723962887290249, "learning_rate": 1.9506970021475798e-05, "loss": 0.9101, "step": 3517 }, { "epoch": 0.12741760231800073, "grad_norm": 1.8263234121514185, "learning_rate": 1.95066061627991e-05, "loss": 0.7955, "step": 3518 }, { "epoch": 0.12745382107931907, "grad_norm": 1.7140448692792083, "learning_rate": 1.9506242173303142e-05, "loss": 0.8466, "step": 3519 }, { "epoch": 0.12749003984063745, "grad_norm": 1.6904965083921966, "learning_rate": 1.9505878052992945e-05, "loss": 0.9195, "step": 3520 }, { "epoch": 0.12752625860195582, "grad_norm": 1.4210974171767725, "learning_rate": 1.950551380187351e-05, "loss": 0.8152, "step": 3521 }, { "epoch": 0.12756247736327417, "grad_norm": 1.7657362810023485, "learning_rate": 1.9505149419949853e-05, "loss": 0.7859, "step": 3522 }, { "epoch": 0.12759869612459254, "grad_norm": 1.380801564028927, "learning_rate": 1.9504784907226984e-05, "loss": 0.8288, "step": 3523 }, { "epoch": 0.1276349148859109, "grad_norm": 1.2693662497693932, "learning_rate": 1.950442026370993e-05, "loss": 0.8083, "step": 3524 }, { "epoch": 0.12767113364722926, "grad_norm": 1.7582307466321732, "learning_rate": 1.9504055489403696e-05, "loss": 0.8308, "step": 3525 }, { "epoch": 0.12770735240854764, "grad_norm": 1.5773938611050011, "learning_rate": 1.9503690584313312e-05, "loss": 0.9067, "step": 3526 }, { "epoch": 0.12774357116986598, "grad_norm": 1.6577739205037578, "learning_rate": 1.9503325548443793e-05, "loss": 0.7903, "step": 3527 }, { "epoch": 0.12777978993118436, "grad_norm": 1.3785216319190292, "learning_rate": 1.9502960381800165e-05, "loss": 0.7923, "step": 3528 }, { "epoch": 0.1278160086925027, "grad_norm": 1.6413615900633878, "learning_rate": 1.950259508438745e-05, "loss": 0.8238, "step": 3529 }, { "epoch": 0.12785222745382108, "grad_norm": 1.8698008237329389, "learning_rate": 1.950222965621068e-05, "loss": 0.804, "step": 3530 }, { "epoch": 0.12788844621513945, "grad_norm": 1.8905048883920228, "learning_rate": 1.950186409727488e-05, "loss": 0.8839, "step": 3531 }, { "epoch": 0.1279246649764578, "grad_norm": 1.6279967476728971, "learning_rate": 1.950149840758508e-05, "loss": 0.8414, "step": 3532 }, { "epoch": 0.12796088373777617, "grad_norm": 1.655698704736818, "learning_rate": 1.950113258714631e-05, "loss": 0.8529, "step": 3533 }, { "epoch": 0.12799710249909452, "grad_norm": 1.8193281227799882, "learning_rate": 1.9500766635963614e-05, "loss": 0.9391, "step": 3534 }, { "epoch": 0.1280333212604129, "grad_norm": 1.8632191221807737, "learning_rate": 1.9500400554042015e-05, "loss": 0.9354, "step": 3535 }, { "epoch": 0.12806954002173127, "grad_norm": 1.7490417505069136, "learning_rate": 1.950003434138656e-05, "loss": 0.8626, "step": 3536 }, { "epoch": 0.1281057587830496, "grad_norm": 1.711291186254693, "learning_rate": 1.9499667998002284e-05, "loss": 0.8152, "step": 3537 }, { "epoch": 0.12814197754436799, "grad_norm": 1.762395696986291, "learning_rate": 1.9499301523894226e-05, "loss": 0.8623, "step": 3538 }, { "epoch": 0.12817819630568636, "grad_norm": 1.5100321886005983, "learning_rate": 1.9498934919067435e-05, "loss": 0.782, "step": 3539 }, { "epoch": 0.1282144150670047, "grad_norm": 1.8561325933594093, "learning_rate": 1.9498568183526956e-05, "loss": 0.8885, "step": 3540 }, { "epoch": 0.12825063382832308, "grad_norm": 1.7679886230619455, "learning_rate": 1.949820131727783e-05, "loss": 0.8354, "step": 3541 }, { "epoch": 0.12828685258964143, "grad_norm": 1.7884235823731478, "learning_rate": 1.9497834320325107e-05, "loss": 0.8979, "step": 3542 }, { "epoch": 0.1283230713509598, "grad_norm": 1.5584518994288974, "learning_rate": 1.9497467192673836e-05, "loss": 0.8731, "step": 3543 }, { "epoch": 0.12835929011227817, "grad_norm": 1.6054060661534568, "learning_rate": 1.9497099934329075e-05, "loss": 0.85, "step": 3544 }, { "epoch": 0.12839550887359652, "grad_norm": 1.2066062686213312, "learning_rate": 1.9496732545295874e-05, "loss": 0.7568, "step": 3545 }, { "epoch": 0.1284317276349149, "grad_norm": 2.0660658864330133, "learning_rate": 1.949636502557929e-05, "loss": 0.8152, "step": 3546 }, { "epoch": 0.12846794639623324, "grad_norm": 1.6130727536232792, "learning_rate": 1.9495997375184377e-05, "loss": 0.8585, "step": 3547 }, { "epoch": 0.12850416515755161, "grad_norm": 1.6790805618323543, "learning_rate": 1.9495629594116193e-05, "loss": 0.9198, "step": 3548 }, { "epoch": 0.12854038391887, "grad_norm": 1.6195056098973029, "learning_rate": 1.9495261682379808e-05, "loss": 0.8723, "step": 3549 }, { "epoch": 0.12857660268018833, "grad_norm": 1.429265013595327, "learning_rate": 1.9494893639980276e-05, "loss": 0.8544, "step": 3550 }, { "epoch": 0.1286128214415067, "grad_norm": 1.5764819235255443, "learning_rate": 1.9494525466922663e-05, "loss": 0.763, "step": 3551 }, { "epoch": 0.12864904020282505, "grad_norm": 1.5909383652723497, "learning_rate": 1.949415716321204e-05, "loss": 0.9131, "step": 3552 }, { "epoch": 0.12868525896414343, "grad_norm": 1.5490546818449926, "learning_rate": 1.949378872885347e-05, "loss": 0.9134, "step": 3553 }, { "epoch": 0.1287214777254618, "grad_norm": 1.5188140675219086, "learning_rate": 1.949342016385203e-05, "loss": 0.7764, "step": 3554 }, { "epoch": 0.12875769648678015, "grad_norm": 1.668378348612181, "learning_rate": 1.949305146821278e-05, "loss": 0.9384, "step": 3555 }, { "epoch": 0.12879391524809852, "grad_norm": 1.3647946848945376, "learning_rate": 1.9492682641940803e-05, "loss": 0.7728, "step": 3556 }, { "epoch": 0.12883013400941687, "grad_norm": 1.1809049212051506, "learning_rate": 1.9492313685041174e-05, "loss": 0.7209, "step": 3557 }, { "epoch": 0.12886635277073524, "grad_norm": 1.805272110877973, "learning_rate": 1.9491944597518965e-05, "loss": 0.8639, "step": 3558 }, { "epoch": 0.12890257153205362, "grad_norm": 1.7689455233465647, "learning_rate": 1.949157537937926e-05, "loss": 0.9652, "step": 3559 }, { "epoch": 0.12893879029337196, "grad_norm": 1.6143063320430653, "learning_rate": 1.9491206030627136e-05, "loss": 0.8815, "step": 3560 }, { "epoch": 0.12897500905469034, "grad_norm": 1.6420732225659598, "learning_rate": 1.949083655126768e-05, "loss": 0.7689, "step": 3561 }, { "epoch": 0.12901122781600868, "grad_norm": 1.4209202032290187, "learning_rate": 1.9490466941305972e-05, "loss": 0.833, "step": 3562 }, { "epoch": 0.12904744657732706, "grad_norm": 1.8375799299553748, "learning_rate": 1.94900972007471e-05, "loss": 0.8727, "step": 3563 }, { "epoch": 0.12908366533864543, "grad_norm": 1.5007908097478486, "learning_rate": 1.9489727329596147e-05, "loss": 0.8196, "step": 3564 }, { "epoch": 0.12911988409996378, "grad_norm": 1.6795813826374415, "learning_rate": 1.9489357327858216e-05, "loss": 0.761, "step": 3565 }, { "epoch": 0.12915610286128215, "grad_norm": 1.251476656809259, "learning_rate": 1.9488987195538386e-05, "loss": 0.8202, "step": 3566 }, { "epoch": 0.1291923216226005, "grad_norm": 1.6706429447254976, "learning_rate": 1.9488616932641756e-05, "loss": 0.8664, "step": 3567 }, { "epoch": 0.12922854038391887, "grad_norm": 1.3876807419611294, "learning_rate": 1.9488246539173415e-05, "loss": 0.8661, "step": 3568 }, { "epoch": 0.12926475914523725, "grad_norm": 1.768336416895578, "learning_rate": 1.948787601513847e-05, "loss": 0.8589, "step": 3569 }, { "epoch": 0.1293009779065556, "grad_norm": 1.4178938024854044, "learning_rate": 1.9487505360542013e-05, "loss": 0.7839, "step": 3570 }, { "epoch": 0.12933719666787397, "grad_norm": 1.4967478711590332, "learning_rate": 1.948713457538914e-05, "loss": 0.8682, "step": 3571 }, { "epoch": 0.1293734154291923, "grad_norm": 1.570397165967181, "learning_rate": 1.948676365968497e-05, "loss": 0.8513, "step": 3572 }, { "epoch": 0.1294096341905107, "grad_norm": 1.6087174739009684, "learning_rate": 1.9486392613434584e-05, "loss": 0.8277, "step": 3573 }, { "epoch": 0.12944585295182906, "grad_norm": 1.6937308327533458, "learning_rate": 1.948602143664311e-05, "loss": 0.9231, "step": 3574 }, { "epoch": 0.1294820717131474, "grad_norm": 1.6076529222098592, "learning_rate": 1.9485650129315645e-05, "loss": 0.8326, "step": 3575 }, { "epoch": 0.12951829047446578, "grad_norm": 1.485136866806742, "learning_rate": 1.9485278691457292e-05, "loss": 0.8088, "step": 3576 }, { "epoch": 0.12955450923578413, "grad_norm": 1.5848575118771568, "learning_rate": 1.9484907123073175e-05, "loss": 0.8877, "step": 3577 }, { "epoch": 0.1295907279971025, "grad_norm": 1.8675785852211069, "learning_rate": 1.9484535424168403e-05, "loss": 0.8564, "step": 3578 }, { "epoch": 0.12962694675842087, "grad_norm": 1.7118387635508898, "learning_rate": 1.9484163594748088e-05, "loss": 0.9202, "step": 3579 }, { "epoch": 0.12966316551973922, "grad_norm": 1.4931703792535709, "learning_rate": 1.9483791634817347e-05, "loss": 0.8762, "step": 3580 }, { "epoch": 0.1296993842810576, "grad_norm": 1.6566042836059047, "learning_rate": 1.94834195443813e-05, "loss": 0.8827, "step": 3581 }, { "epoch": 0.12973560304237594, "grad_norm": 1.8824618733263572, "learning_rate": 1.948304732344507e-05, "loss": 0.9454, "step": 3582 }, { "epoch": 0.12977182180369431, "grad_norm": 1.6190417493073206, "learning_rate": 1.9482674972013775e-05, "loss": 0.8592, "step": 3583 }, { "epoch": 0.1298080405650127, "grad_norm": 1.7677743328641387, "learning_rate": 1.948230249009254e-05, "loss": 0.8712, "step": 3584 }, { "epoch": 0.12984425932633104, "grad_norm": 1.5980764596569499, "learning_rate": 1.948192987768649e-05, "loss": 0.8559, "step": 3585 }, { "epoch": 0.1298804780876494, "grad_norm": 1.603644426263025, "learning_rate": 1.9481557134800756e-05, "loss": 0.7464, "step": 3586 }, { "epoch": 0.12991669684896776, "grad_norm": 1.828073891104607, "learning_rate": 1.9481184261440462e-05, "loss": 0.882, "step": 3587 }, { "epoch": 0.12995291561028613, "grad_norm": 1.3895950086369284, "learning_rate": 1.9480811257610744e-05, "loss": 0.7605, "step": 3588 }, { "epoch": 0.1299891343716045, "grad_norm": 1.6534990145369188, "learning_rate": 1.948043812331673e-05, "loss": 0.8489, "step": 3589 }, { "epoch": 0.13002535313292285, "grad_norm": 1.6666082622805145, "learning_rate": 1.9480064858563558e-05, "loss": 0.9023, "step": 3590 }, { "epoch": 0.13006157189424122, "grad_norm": 1.5344292120129703, "learning_rate": 1.947969146335636e-05, "loss": 0.8186, "step": 3591 }, { "epoch": 0.13009779065555957, "grad_norm": 1.504190040416239, "learning_rate": 1.9479317937700284e-05, "loss": 0.8659, "step": 3592 }, { "epoch": 0.13013400941687794, "grad_norm": 1.5634786556975668, "learning_rate": 1.947894428160046e-05, "loss": 0.8604, "step": 3593 }, { "epoch": 0.13017022817819632, "grad_norm": 1.6579029476541367, "learning_rate": 1.9478570495062038e-05, "loss": 0.8979, "step": 3594 }, { "epoch": 0.13020644693951466, "grad_norm": 1.3078908733586208, "learning_rate": 1.9478196578090152e-05, "loss": 0.7582, "step": 3595 }, { "epoch": 0.13024266570083304, "grad_norm": 1.7091216288354574, "learning_rate": 1.947782253068996e-05, "loss": 0.9059, "step": 3596 }, { "epoch": 0.13027888446215138, "grad_norm": 1.6068924597520873, "learning_rate": 1.9477448352866596e-05, "loss": 0.8344, "step": 3597 }, { "epoch": 0.13031510322346976, "grad_norm": 1.5309274890947666, "learning_rate": 1.9477074044625215e-05, "loss": 0.8007, "step": 3598 }, { "epoch": 0.13035132198478813, "grad_norm": 1.4018003710519165, "learning_rate": 1.947669960597097e-05, "loss": 0.8197, "step": 3599 }, { "epoch": 0.13038754074610648, "grad_norm": 1.620196131874729, "learning_rate": 1.947632503690901e-05, "loss": 0.9203, "step": 3600 }, { "epoch": 0.13042375950742485, "grad_norm": 1.5595553713317505, "learning_rate": 1.947595033744449e-05, "loss": 0.8365, "step": 3601 }, { "epoch": 0.1304599782687432, "grad_norm": 1.532054992074004, "learning_rate": 1.947557550758257e-05, "loss": 0.808, "step": 3602 }, { "epoch": 0.13049619703006157, "grad_norm": 1.8551222427465397, "learning_rate": 1.9475200547328407e-05, "loss": 0.9225, "step": 3603 }, { "epoch": 0.13053241579137995, "grad_norm": 1.847001635216659, "learning_rate": 1.947482545668715e-05, "loss": 0.8761, "step": 3604 }, { "epoch": 0.1305686345526983, "grad_norm": 1.4759693827800875, "learning_rate": 1.9474450235663978e-05, "loss": 0.8051, "step": 3605 }, { "epoch": 0.13060485331401667, "grad_norm": 1.104457728626673, "learning_rate": 1.9474074884264043e-05, "loss": 0.7528, "step": 3606 }, { "epoch": 0.130641072075335, "grad_norm": 1.661167753279346, "learning_rate": 1.9473699402492513e-05, "loss": 0.8185, "step": 3607 }, { "epoch": 0.1306772908366534, "grad_norm": 1.6213522681960297, "learning_rate": 1.9473323790354556e-05, "loss": 0.8404, "step": 3608 }, { "epoch": 0.13071350959797176, "grad_norm": 1.2682414634512005, "learning_rate": 1.9472948047855338e-05, "loss": 0.744, "step": 3609 }, { "epoch": 0.1307497283592901, "grad_norm": 1.6600832889456283, "learning_rate": 1.947257217500003e-05, "loss": 0.8736, "step": 3610 }, { "epoch": 0.13078594712060848, "grad_norm": 1.614958668672076, "learning_rate": 1.9472196171793808e-05, "loss": 0.8801, "step": 3611 }, { "epoch": 0.13082216588192683, "grad_norm": 1.5575401394721007, "learning_rate": 1.9471820038241844e-05, "loss": 0.8859, "step": 3612 }, { "epoch": 0.1308583846432452, "grad_norm": 1.6193108250625403, "learning_rate": 1.947144377434931e-05, "loss": 0.8912, "step": 3613 }, { "epoch": 0.13089460340456358, "grad_norm": 1.715546474848477, "learning_rate": 1.947106738012139e-05, "loss": 0.9083, "step": 3614 }, { "epoch": 0.13093082216588192, "grad_norm": 1.6965566884104835, "learning_rate": 1.9470690855563263e-05, "loss": 0.8241, "step": 3615 }, { "epoch": 0.1309670409272003, "grad_norm": 1.4249659088201978, "learning_rate": 1.9470314200680104e-05, "loss": 0.807, "step": 3616 }, { "epoch": 0.13100325968851864, "grad_norm": 1.4500149335328567, "learning_rate": 1.94699374154771e-05, "loss": 0.8414, "step": 3617 }, { "epoch": 0.13103947844983702, "grad_norm": 1.6513346088518734, "learning_rate": 1.946956049995944e-05, "loss": 0.8939, "step": 3618 }, { "epoch": 0.1310756972111554, "grad_norm": 1.8159717152850907, "learning_rate": 1.9469183454132303e-05, "loss": 0.8549, "step": 3619 }, { "epoch": 0.13111191597247374, "grad_norm": 1.4809703329503294, "learning_rate": 1.9468806278000884e-05, "loss": 0.9011, "step": 3620 }, { "epoch": 0.1311481347337921, "grad_norm": 1.4606622332049692, "learning_rate": 1.946842897157037e-05, "loss": 0.8694, "step": 3621 }, { "epoch": 0.13118435349511046, "grad_norm": 1.8995241879750195, "learning_rate": 1.9468051534845954e-05, "loss": 0.8276, "step": 3622 }, { "epoch": 0.13122057225642883, "grad_norm": 1.655957860868273, "learning_rate": 1.9467673967832828e-05, "loss": 0.803, "step": 3623 }, { "epoch": 0.1312567910177472, "grad_norm": 1.1957653289818493, "learning_rate": 1.9467296270536192e-05, "loss": 0.8406, "step": 3624 }, { "epoch": 0.13129300977906555, "grad_norm": 1.618479728412982, "learning_rate": 1.9466918442961237e-05, "loss": 0.847, "step": 3625 }, { "epoch": 0.13132922854038392, "grad_norm": 1.5195335728938923, "learning_rate": 1.9466540485113167e-05, "loss": 0.839, "step": 3626 }, { "epoch": 0.13136544730170227, "grad_norm": 1.1876810943164615, "learning_rate": 1.9466162396997183e-05, "loss": 0.8182, "step": 3627 }, { "epoch": 0.13140166606302064, "grad_norm": 1.8693641600681938, "learning_rate": 1.9465784178618485e-05, "loss": 0.9274, "step": 3628 }, { "epoch": 0.13143788482433902, "grad_norm": 1.6414853837239498, "learning_rate": 1.946540582998228e-05, "loss": 0.7397, "step": 3629 }, { "epoch": 0.13147410358565736, "grad_norm": 1.6842686280720491, "learning_rate": 1.9465027351093776e-05, "loss": 0.9262, "step": 3630 }, { "epoch": 0.13151032234697574, "grad_norm": 1.6095815648686718, "learning_rate": 1.946464874195818e-05, "loss": 0.8873, "step": 3631 }, { "epoch": 0.13154654110829408, "grad_norm": 1.7651537835020432, "learning_rate": 1.9464270002580697e-05, "loss": 0.847, "step": 3632 }, { "epoch": 0.13158275986961246, "grad_norm": 1.757274373654758, "learning_rate": 1.9463891132966545e-05, "loss": 0.8886, "step": 3633 }, { "epoch": 0.13161897863093083, "grad_norm": 1.6097993175606509, "learning_rate": 1.9463512133120935e-05, "loss": 0.8543, "step": 3634 }, { "epoch": 0.13165519739224918, "grad_norm": 1.5941576062480225, "learning_rate": 1.946313300304908e-05, "loss": 0.9091, "step": 3635 }, { "epoch": 0.13169141615356755, "grad_norm": 1.64930456305078, "learning_rate": 1.94627537427562e-05, "loss": 0.8695, "step": 3636 }, { "epoch": 0.1317276349148859, "grad_norm": 1.7700159280674532, "learning_rate": 1.946237435224752e-05, "loss": 0.8195, "step": 3637 }, { "epoch": 0.13176385367620427, "grad_norm": 1.57190161098532, "learning_rate": 1.9461994831528244e-05, "loss": 0.8951, "step": 3638 }, { "epoch": 0.13180007243752265, "grad_norm": 1.669394234625384, "learning_rate": 1.9461615180603614e-05, "loss": 0.8358, "step": 3639 }, { "epoch": 0.131836291198841, "grad_norm": 1.7252641193182219, "learning_rate": 1.9461235399478842e-05, "loss": 0.8424, "step": 3640 }, { "epoch": 0.13187250996015937, "grad_norm": 1.5096267640037668, "learning_rate": 1.946085548815916e-05, "loss": 0.9087, "step": 3641 }, { "epoch": 0.1319087287214777, "grad_norm": 1.5467584570899668, "learning_rate": 1.946047544664979e-05, "loss": 0.8589, "step": 3642 }, { "epoch": 0.1319449474827961, "grad_norm": 1.4738995359345295, "learning_rate": 1.9460095274955968e-05, "loss": 0.7995, "step": 3643 }, { "epoch": 0.13198116624411446, "grad_norm": 1.2139479732254113, "learning_rate": 1.9459714973082923e-05, "loss": 0.8112, "step": 3644 }, { "epoch": 0.1320173850054328, "grad_norm": 1.6049520642651338, "learning_rate": 1.9459334541035883e-05, "loss": 0.9132, "step": 3645 }, { "epoch": 0.13205360376675118, "grad_norm": 1.7015963768192046, "learning_rate": 1.9458953978820092e-05, "loss": 0.8158, "step": 3646 }, { "epoch": 0.13208982252806953, "grad_norm": 3.1827622471584354, "learning_rate": 1.9458573286440783e-05, "loss": 0.7891, "step": 3647 }, { "epoch": 0.1321260412893879, "grad_norm": 1.2138873266059387, "learning_rate": 1.9458192463903195e-05, "loss": 0.7775, "step": 3648 }, { "epoch": 0.13216226005070628, "grad_norm": 1.6116747038365773, "learning_rate": 1.9457811511212565e-05, "loss": 0.9554, "step": 3649 }, { "epoch": 0.13219847881202462, "grad_norm": 1.770869635228577, "learning_rate": 1.945743042837414e-05, "loss": 0.9003, "step": 3650 }, { "epoch": 0.132234697573343, "grad_norm": 1.3735605956921106, "learning_rate": 1.945704921539316e-05, "loss": 0.8258, "step": 3651 }, { "epoch": 0.13227091633466134, "grad_norm": 1.6704159323674135, "learning_rate": 1.945666787227488e-05, "loss": 0.8341, "step": 3652 }, { "epoch": 0.13230713509597972, "grad_norm": 1.3852352552281118, "learning_rate": 1.9456286399024537e-05, "loss": 0.713, "step": 3653 }, { "epoch": 0.1323433538572981, "grad_norm": 3.1208414016325636, "learning_rate": 1.945590479564738e-05, "loss": 0.8888, "step": 3654 }, { "epoch": 0.13237957261861644, "grad_norm": 1.7759880030309783, "learning_rate": 1.9455523062148667e-05, "loss": 0.8655, "step": 3655 }, { "epoch": 0.1324157913799348, "grad_norm": 1.6963401876221946, "learning_rate": 1.945514119853365e-05, "loss": 0.7886, "step": 3656 }, { "epoch": 0.13245201014125316, "grad_norm": 1.7209985048672978, "learning_rate": 1.945475920480758e-05, "loss": 0.9189, "step": 3657 }, { "epoch": 0.13248822890257153, "grad_norm": 1.70928431031497, "learning_rate": 1.9454377080975718e-05, "loss": 0.8693, "step": 3658 }, { "epoch": 0.1325244476638899, "grad_norm": 1.5717330646706817, "learning_rate": 1.945399482704332e-05, "loss": 0.8486, "step": 3659 }, { "epoch": 0.13256066642520825, "grad_norm": 1.4610995593874387, "learning_rate": 1.9453612443015642e-05, "loss": 0.7637, "step": 3660 }, { "epoch": 0.13259688518652663, "grad_norm": 1.2719635731103691, "learning_rate": 1.9453229928897955e-05, "loss": 0.7885, "step": 3661 }, { "epoch": 0.13263310394784497, "grad_norm": 1.3501553505888566, "learning_rate": 1.9452847284695515e-05, "loss": 0.8158, "step": 3662 }, { "epoch": 0.13266932270916335, "grad_norm": 1.3543156811887216, "learning_rate": 1.9452464510413593e-05, "loss": 0.7632, "step": 3663 }, { "epoch": 0.13270554147048172, "grad_norm": 1.851608676173648, "learning_rate": 1.945208160605745e-05, "loss": 0.8973, "step": 3664 }, { "epoch": 0.13274176023180007, "grad_norm": 1.556229795019327, "learning_rate": 1.945169857163236e-05, "loss": 0.8479, "step": 3665 }, { "epoch": 0.13277797899311844, "grad_norm": 1.5276888106341717, "learning_rate": 1.9451315407143593e-05, "loss": 0.7907, "step": 3666 }, { "epoch": 0.13281419775443679, "grad_norm": 1.6535088735217764, "learning_rate": 1.9450932112596423e-05, "loss": 0.8417, "step": 3667 }, { "epoch": 0.13285041651575516, "grad_norm": 1.606807385023225, "learning_rate": 1.9450548687996122e-05, "loss": 0.8635, "step": 3668 }, { "epoch": 0.13288663527707353, "grad_norm": 2.456750874980454, "learning_rate": 1.9450165133347963e-05, "loss": 0.7511, "step": 3669 }, { "epoch": 0.13292285403839188, "grad_norm": 1.4506826397543757, "learning_rate": 1.9449781448657234e-05, "loss": 0.772, "step": 3670 }, { "epoch": 0.13295907279971025, "grad_norm": 1.629131465089076, "learning_rate": 1.9449397633929208e-05, "loss": 0.8707, "step": 3671 }, { "epoch": 0.1329952915610286, "grad_norm": 1.5852557053347505, "learning_rate": 1.9449013689169166e-05, "loss": 0.8213, "step": 3672 }, { "epoch": 0.13303151032234697, "grad_norm": 1.608798370482599, "learning_rate": 1.9448629614382394e-05, "loss": 0.8267, "step": 3673 }, { "epoch": 0.13306772908366535, "grad_norm": 1.4503273765721163, "learning_rate": 1.9448245409574173e-05, "loss": 0.8634, "step": 3674 }, { "epoch": 0.1331039478449837, "grad_norm": 1.3324555821657962, "learning_rate": 1.9447861074749798e-05, "loss": 0.7589, "step": 3675 }, { "epoch": 0.13314016660630207, "grad_norm": 1.7521145383714787, "learning_rate": 1.944747660991455e-05, "loss": 0.8901, "step": 3676 }, { "epoch": 0.13317638536762041, "grad_norm": 1.6447167167373407, "learning_rate": 1.9447092015073724e-05, "loss": 0.8289, "step": 3677 }, { "epoch": 0.1332126041289388, "grad_norm": 1.6448839414249752, "learning_rate": 1.9446707290232608e-05, "loss": 0.8359, "step": 3678 }, { "epoch": 0.13324882289025716, "grad_norm": 1.5844254261848045, "learning_rate": 1.9446322435396504e-05, "loss": 0.8605, "step": 3679 }, { "epoch": 0.1332850416515755, "grad_norm": 1.695200297272749, "learning_rate": 1.9445937450570697e-05, "loss": 0.8093, "step": 3680 }, { "epoch": 0.13332126041289388, "grad_norm": 1.8344117713716197, "learning_rate": 1.9445552335760497e-05, "loss": 0.8743, "step": 3681 }, { "epoch": 0.13335747917421223, "grad_norm": 1.645061324243425, "learning_rate": 1.944516709097119e-05, "loss": 0.8776, "step": 3682 }, { "epoch": 0.1333936979355306, "grad_norm": 1.7215148873413404, "learning_rate": 1.944478171620809e-05, "loss": 0.7154, "step": 3683 }, { "epoch": 0.13342991669684898, "grad_norm": 1.939815908924331, "learning_rate": 1.9444396211476495e-05, "loss": 0.7794, "step": 3684 }, { "epoch": 0.13346613545816732, "grad_norm": 1.8651761660056225, "learning_rate": 1.9444010576781708e-05, "loss": 0.8953, "step": 3685 }, { "epoch": 0.1335023542194857, "grad_norm": 1.5321308313348296, "learning_rate": 1.9443624812129037e-05, "loss": 0.8828, "step": 3686 }, { "epoch": 0.13353857298080404, "grad_norm": 1.7100871326582767, "learning_rate": 1.944323891752379e-05, "loss": 0.8511, "step": 3687 }, { "epoch": 0.13357479174212242, "grad_norm": 1.6532889569909297, "learning_rate": 1.944285289297128e-05, "loss": 0.8669, "step": 3688 }, { "epoch": 0.1336110105034408, "grad_norm": 1.5665128272084188, "learning_rate": 1.9442466738476814e-05, "loss": 0.7774, "step": 3689 }, { "epoch": 0.13364722926475914, "grad_norm": 1.6725858721210005, "learning_rate": 1.944208045404571e-05, "loss": 0.7884, "step": 3690 }, { "epoch": 0.1336834480260775, "grad_norm": 1.6001338535576388, "learning_rate": 1.9441694039683284e-05, "loss": 0.8379, "step": 3691 }, { "epoch": 0.13371966678739586, "grad_norm": 1.4165467016973878, "learning_rate": 1.944130749539485e-05, "loss": 0.8232, "step": 3692 }, { "epoch": 0.13375588554871423, "grad_norm": 1.713182751746102, "learning_rate": 1.9440920821185734e-05, "loss": 0.8586, "step": 3693 }, { "epoch": 0.1337921043100326, "grad_norm": 1.8373476963599362, "learning_rate": 1.9440534017061247e-05, "loss": 0.9339, "step": 3694 }, { "epoch": 0.13382832307135095, "grad_norm": 1.7884215222827624, "learning_rate": 1.9440147083026717e-05, "loss": 0.8449, "step": 3695 }, { "epoch": 0.13386454183266933, "grad_norm": 1.217130892427988, "learning_rate": 1.9439760019087468e-05, "loss": 0.7809, "step": 3696 }, { "epoch": 0.13390076059398767, "grad_norm": 1.3684086564817879, "learning_rate": 1.9439372825248828e-05, "loss": 0.8509, "step": 3697 }, { "epoch": 0.13393697935530605, "grad_norm": 1.7169119337408627, "learning_rate": 1.9438985501516123e-05, "loss": 0.9252, "step": 3698 }, { "epoch": 0.13397319811662442, "grad_norm": 1.826049444613092, "learning_rate": 1.9438598047894684e-05, "loss": 0.7556, "step": 3699 }, { "epoch": 0.13400941687794277, "grad_norm": 1.30245761240584, "learning_rate": 1.9438210464389842e-05, "loss": 0.8109, "step": 3700 }, { "epoch": 0.13404563563926114, "grad_norm": 1.296797638404029, "learning_rate": 1.9437822751006932e-05, "loss": 0.7822, "step": 3701 }, { "epoch": 0.1340818544005795, "grad_norm": 1.495493643998586, "learning_rate": 1.9437434907751285e-05, "loss": 0.8245, "step": 3702 }, { "epoch": 0.13411807316189786, "grad_norm": 1.7195906896816469, "learning_rate": 1.9437046934628246e-05, "loss": 0.9654, "step": 3703 }, { "epoch": 0.13415429192321623, "grad_norm": 1.5221192047967897, "learning_rate": 1.9436658831643146e-05, "loss": 0.803, "step": 3704 }, { "epoch": 0.13419051068453458, "grad_norm": 1.731923856404036, "learning_rate": 1.943627059880133e-05, "loss": 0.9783, "step": 3705 }, { "epoch": 0.13422672944585295, "grad_norm": 1.5092391099039308, "learning_rate": 1.943588223610814e-05, "loss": 0.7834, "step": 3706 }, { "epoch": 0.1342629482071713, "grad_norm": 1.6420860284811452, "learning_rate": 1.9435493743568918e-05, "loss": 0.8573, "step": 3707 }, { "epoch": 0.13429916696848967, "grad_norm": 1.569230971656738, "learning_rate": 1.9435105121189012e-05, "loss": 0.8169, "step": 3708 }, { "epoch": 0.13433538572980805, "grad_norm": 1.655734591541681, "learning_rate": 1.9434716368973765e-05, "loss": 0.8939, "step": 3709 }, { "epoch": 0.1343716044911264, "grad_norm": 1.3596676536805732, "learning_rate": 1.9434327486928537e-05, "loss": 0.794, "step": 3710 }, { "epoch": 0.13440782325244477, "grad_norm": 1.199638441397306, "learning_rate": 1.943393847505867e-05, "loss": 0.8135, "step": 3711 }, { "epoch": 0.13444404201376314, "grad_norm": 1.5991774920353425, "learning_rate": 1.943354933336952e-05, "loss": 0.8987, "step": 3712 }, { "epoch": 0.1344802607750815, "grad_norm": 1.740206270815122, "learning_rate": 1.9433160061866442e-05, "loss": 0.8482, "step": 3713 }, { "epoch": 0.13451647953639986, "grad_norm": 1.3554152745607964, "learning_rate": 1.943277066055479e-05, "loss": 0.822, "step": 3714 }, { "epoch": 0.1345526982977182, "grad_norm": 1.4084557751051203, "learning_rate": 1.943238112943993e-05, "loss": 0.7686, "step": 3715 }, { "epoch": 0.13458891705903658, "grad_norm": 1.3926956894469247, "learning_rate": 1.9431991468527217e-05, "loss": 0.8002, "step": 3716 }, { "epoch": 0.13462513582035496, "grad_norm": 1.5354415625997762, "learning_rate": 1.9431601677822014e-05, "loss": 0.8292, "step": 3717 }, { "epoch": 0.1346613545816733, "grad_norm": 1.3380269257003035, "learning_rate": 1.9431211757329683e-05, "loss": 0.832, "step": 3718 }, { "epoch": 0.13469757334299168, "grad_norm": 1.3556652792345463, "learning_rate": 1.943082170705559e-05, "loss": 0.8019, "step": 3719 }, { "epoch": 0.13473379210431002, "grad_norm": 1.6667218729402775, "learning_rate": 1.9430431527005107e-05, "loss": 0.8488, "step": 3720 }, { "epoch": 0.1347700108656284, "grad_norm": 1.764508435512173, "learning_rate": 1.94300412171836e-05, "loss": 0.8418, "step": 3721 }, { "epoch": 0.13480622962694677, "grad_norm": 1.8007962739280499, "learning_rate": 1.9429650777596437e-05, "loss": 0.8131, "step": 3722 }, { "epoch": 0.13484244838826512, "grad_norm": 1.1460278407706792, "learning_rate": 1.9429260208248994e-05, "loss": 0.6883, "step": 3723 }, { "epoch": 0.1348786671495835, "grad_norm": 1.7812285693870464, "learning_rate": 1.9428869509146648e-05, "loss": 0.8564, "step": 3724 }, { "epoch": 0.13491488591090184, "grad_norm": 1.6854002923095384, "learning_rate": 1.9428478680294768e-05, "loss": 0.7859, "step": 3725 }, { "epoch": 0.1349511046722202, "grad_norm": 1.6146273781296068, "learning_rate": 1.942808772169874e-05, "loss": 0.8215, "step": 3726 }, { "epoch": 0.1349873234335386, "grad_norm": 1.6097785041106891, "learning_rate": 1.9427696633363942e-05, "loss": 0.841, "step": 3727 }, { "epoch": 0.13502354219485693, "grad_norm": 1.3400712959115488, "learning_rate": 1.942730541529575e-05, "loss": 0.7989, "step": 3728 }, { "epoch": 0.1350597609561753, "grad_norm": 1.6968659708492397, "learning_rate": 1.9426914067499554e-05, "loss": 0.9139, "step": 3729 }, { "epoch": 0.13509597971749365, "grad_norm": 1.5340153195130737, "learning_rate": 1.942652258998074e-05, "loss": 0.861, "step": 3730 }, { "epoch": 0.13513219847881203, "grad_norm": 1.2914447321072267, "learning_rate": 1.942613098274469e-05, "loss": 0.7328, "step": 3731 }, { "epoch": 0.1351684172401304, "grad_norm": 1.6282846101660853, "learning_rate": 1.9425739245796794e-05, "loss": 0.8516, "step": 3732 }, { "epoch": 0.13520463600144875, "grad_norm": 1.726788262240189, "learning_rate": 1.9425347379142447e-05, "loss": 0.8529, "step": 3733 }, { "epoch": 0.13524085476276712, "grad_norm": 1.6932179679038737, "learning_rate": 1.9424955382787035e-05, "loss": 0.8686, "step": 3734 }, { "epoch": 0.13527707352408547, "grad_norm": 1.603365663446315, "learning_rate": 1.9424563256735956e-05, "loss": 0.8782, "step": 3735 }, { "epoch": 0.13531329228540384, "grad_norm": 1.6330357449153226, "learning_rate": 1.9424171000994607e-05, "loss": 0.8961, "step": 3736 }, { "epoch": 0.13534951104672222, "grad_norm": 1.4748626036822583, "learning_rate": 1.9423778615568386e-05, "loss": 0.7656, "step": 3737 }, { "epoch": 0.13538572980804056, "grad_norm": 1.5517294900949048, "learning_rate": 1.9423386100462688e-05, "loss": 0.8767, "step": 3738 }, { "epoch": 0.13542194856935894, "grad_norm": 1.589599540240568, "learning_rate": 1.9422993455682917e-05, "loss": 0.7838, "step": 3739 }, { "epoch": 0.13545816733067728, "grad_norm": 1.5108815161600875, "learning_rate": 1.9422600681234477e-05, "loss": 0.7968, "step": 3740 }, { "epoch": 0.13549438609199566, "grad_norm": 1.6295495786851377, "learning_rate": 1.9422207777122772e-05, "loss": 0.8836, "step": 3741 }, { "epoch": 0.13553060485331403, "grad_norm": 1.4685746808624687, "learning_rate": 1.9421814743353207e-05, "loss": 0.8608, "step": 3742 }, { "epoch": 0.13556682361463238, "grad_norm": 1.456686465528475, "learning_rate": 1.9421421579931198e-05, "loss": 0.8477, "step": 3743 }, { "epoch": 0.13560304237595075, "grad_norm": 1.6762798963493533, "learning_rate": 1.9421028286862144e-05, "loss": 0.9686, "step": 3744 }, { "epoch": 0.1356392611372691, "grad_norm": 1.673419072257197, "learning_rate": 1.942063486415146e-05, "loss": 0.8458, "step": 3745 }, { "epoch": 0.13567547989858747, "grad_norm": 1.4607067376923157, "learning_rate": 1.942024131180457e-05, "loss": 0.8132, "step": 3746 }, { "epoch": 0.13571169865990584, "grad_norm": 1.4338776407394171, "learning_rate": 1.941984762982688e-05, "loss": 0.7982, "step": 3747 }, { "epoch": 0.1357479174212242, "grad_norm": 1.6913032415025486, "learning_rate": 1.941945381822381e-05, "loss": 0.7667, "step": 3748 }, { "epoch": 0.13578413618254256, "grad_norm": 1.537061001017678, "learning_rate": 1.9419059877000776e-05, "loss": 0.8218, "step": 3749 }, { "epoch": 0.1358203549438609, "grad_norm": 2.7289069691394148, "learning_rate": 1.9418665806163203e-05, "loss": 0.8485, "step": 3750 }, { "epoch": 0.13585657370517928, "grad_norm": 1.447817125918225, "learning_rate": 1.9418271605716514e-05, "loss": 0.7617, "step": 3751 }, { "epoch": 0.13589279246649766, "grad_norm": 1.429492725787092, "learning_rate": 1.941787727566613e-05, "loss": 0.922, "step": 3752 }, { "epoch": 0.135929011227816, "grad_norm": 1.7671615967988612, "learning_rate": 1.9417482816017483e-05, "loss": 0.7806, "step": 3753 }, { "epoch": 0.13596522998913438, "grad_norm": 1.3966754790073181, "learning_rate": 1.9417088226775993e-05, "loss": 0.8254, "step": 3754 }, { "epoch": 0.13600144875045272, "grad_norm": 2.125416784926516, "learning_rate": 1.94166935079471e-05, "loss": 0.8822, "step": 3755 }, { "epoch": 0.1360376675117711, "grad_norm": 1.304028016367945, "learning_rate": 1.9416298659536225e-05, "loss": 0.7953, "step": 3756 }, { "epoch": 0.13607388627308947, "grad_norm": 1.8549135481734236, "learning_rate": 1.9415903681548808e-05, "loss": 0.9115, "step": 3757 }, { "epoch": 0.13611010503440782, "grad_norm": 1.7096745736250631, "learning_rate": 1.9415508573990284e-05, "loss": 0.8265, "step": 3758 }, { "epoch": 0.1361463237957262, "grad_norm": 1.674302464580368, "learning_rate": 1.9415113336866086e-05, "loss": 0.8712, "step": 3759 }, { "epoch": 0.13618254255704454, "grad_norm": 1.4578981549928283, "learning_rate": 1.9414717970181657e-05, "loss": 0.9069, "step": 3760 }, { "epoch": 0.1362187613183629, "grad_norm": 1.5989873605965779, "learning_rate": 1.9414322473942437e-05, "loss": 0.8181, "step": 3761 }, { "epoch": 0.1362549800796813, "grad_norm": 1.3671278689782707, "learning_rate": 1.941392684815387e-05, "loss": 0.7912, "step": 3762 }, { "epoch": 0.13629119884099963, "grad_norm": 1.6845968902260422, "learning_rate": 1.9413531092821395e-05, "loss": 0.9339, "step": 3763 }, { "epoch": 0.136327417602318, "grad_norm": 1.6569461484114048, "learning_rate": 1.941313520795046e-05, "loss": 0.8968, "step": 3764 }, { "epoch": 0.13636363636363635, "grad_norm": 1.5393161316593413, "learning_rate": 1.941273919354651e-05, "loss": 0.8699, "step": 3765 }, { "epoch": 0.13639985512495473, "grad_norm": 1.7019410480301047, "learning_rate": 1.9412343049615007e-05, "loss": 0.9238, "step": 3766 }, { "epoch": 0.1364360738862731, "grad_norm": 1.5258596434331, "learning_rate": 1.9411946776161388e-05, "loss": 0.8222, "step": 3767 }, { "epoch": 0.13647229264759145, "grad_norm": 1.5714939379798176, "learning_rate": 1.941155037319111e-05, "loss": 0.8409, "step": 3768 }, { "epoch": 0.13650851140890982, "grad_norm": 1.4481381728155998, "learning_rate": 1.9411153840709632e-05, "loss": 0.905, "step": 3769 }, { "epoch": 0.13654473017022817, "grad_norm": 1.2467553482478164, "learning_rate": 1.9410757178722406e-05, "loss": 0.8109, "step": 3770 }, { "epoch": 0.13658094893154654, "grad_norm": 1.8352983254902318, "learning_rate": 1.9410360387234895e-05, "loss": 0.9033, "step": 3771 }, { "epoch": 0.13661716769286492, "grad_norm": 1.773081369814555, "learning_rate": 1.940996346625255e-05, "loss": 0.8572, "step": 3772 }, { "epoch": 0.13665338645418326, "grad_norm": 1.420111068076072, "learning_rate": 1.9409566415780845e-05, "loss": 0.7998, "step": 3773 }, { "epoch": 0.13668960521550164, "grad_norm": 1.6249658209617457, "learning_rate": 1.9409169235825237e-05, "loss": 0.8333, "step": 3774 }, { "epoch": 0.13672582397681998, "grad_norm": 1.4867003146154012, "learning_rate": 1.9408771926391193e-05, "loss": 0.8633, "step": 3775 }, { "epoch": 0.13676204273813836, "grad_norm": 1.545418961968427, "learning_rate": 1.940837448748418e-05, "loss": 0.8306, "step": 3776 }, { "epoch": 0.13679826149945673, "grad_norm": 1.6657490366060996, "learning_rate": 1.9407976919109666e-05, "loss": 0.8908, "step": 3777 }, { "epoch": 0.13683448026077508, "grad_norm": 1.723137215317469, "learning_rate": 1.9407579221273123e-05, "loss": 0.9049, "step": 3778 }, { "epoch": 0.13687069902209345, "grad_norm": 1.578107884521962, "learning_rate": 1.9407181393980026e-05, "loss": 0.8031, "step": 3779 }, { "epoch": 0.1369069177834118, "grad_norm": 1.2540374139303623, "learning_rate": 1.9406783437235845e-05, "loss": 0.7904, "step": 3780 }, { "epoch": 0.13694313654473017, "grad_norm": 1.45521000130102, "learning_rate": 1.940638535104606e-05, "loss": 0.8032, "step": 3781 }, { "epoch": 0.13697935530604854, "grad_norm": 1.5345184090658426, "learning_rate": 1.9405987135416145e-05, "loss": 0.8621, "step": 3782 }, { "epoch": 0.1370155740673669, "grad_norm": 1.3522066456231747, "learning_rate": 1.9405588790351583e-05, "loss": 0.755, "step": 3783 }, { "epoch": 0.13705179282868526, "grad_norm": 1.6117716184701412, "learning_rate": 1.9405190315857855e-05, "loss": 0.7912, "step": 3784 }, { "epoch": 0.1370880115900036, "grad_norm": 1.7206198063790032, "learning_rate": 1.9404791711940445e-05, "loss": 0.8973, "step": 3785 }, { "epoch": 0.13712423035132199, "grad_norm": 1.620234209718286, "learning_rate": 1.9404392978604836e-05, "loss": 0.7696, "step": 3786 }, { "epoch": 0.13716044911264036, "grad_norm": 1.2131094908811524, "learning_rate": 1.9403994115856515e-05, "loss": 0.8012, "step": 3787 }, { "epoch": 0.1371966678739587, "grad_norm": 1.7552188961171549, "learning_rate": 1.9403595123700974e-05, "loss": 0.8886, "step": 3788 }, { "epoch": 0.13723288663527708, "grad_norm": 1.5314079392977218, "learning_rate": 1.94031960021437e-05, "loss": 0.8118, "step": 3789 }, { "epoch": 0.13726910539659543, "grad_norm": 1.6060585971076307, "learning_rate": 1.940279675119019e-05, "loss": 0.9311, "step": 3790 }, { "epoch": 0.1373053241579138, "grad_norm": 1.5798509998021253, "learning_rate": 1.9402397370845933e-05, "loss": 0.8714, "step": 3791 }, { "epoch": 0.13734154291923217, "grad_norm": 1.6027012021775817, "learning_rate": 1.9401997861116425e-05, "loss": 0.8286, "step": 3792 }, { "epoch": 0.13737776168055052, "grad_norm": 1.589221747144919, "learning_rate": 1.9401598222007168e-05, "loss": 0.8786, "step": 3793 }, { "epoch": 0.1374139804418689, "grad_norm": 1.4979353439523586, "learning_rate": 1.9401198453523654e-05, "loss": 0.7468, "step": 3794 }, { "epoch": 0.13745019920318724, "grad_norm": 1.6141606022941881, "learning_rate": 1.9400798555671395e-05, "loss": 0.7955, "step": 3795 }, { "epoch": 0.13748641796450561, "grad_norm": 1.5681469091678577, "learning_rate": 1.9400398528455883e-05, "loss": 0.8091, "step": 3796 }, { "epoch": 0.137522636725824, "grad_norm": 1.5623543856241897, "learning_rate": 1.9399998371882628e-05, "loss": 0.811, "step": 3797 }, { "epoch": 0.13755885548714233, "grad_norm": 1.6732694829791894, "learning_rate": 1.9399598085957135e-05, "loss": 0.81, "step": 3798 }, { "epoch": 0.1375950742484607, "grad_norm": 1.5245307533131984, "learning_rate": 1.9399197670684913e-05, "loss": 0.8624, "step": 3799 }, { "epoch": 0.13763129300977905, "grad_norm": 1.5719239497478723, "learning_rate": 1.9398797126071472e-05, "loss": 0.8655, "step": 3800 }, { "epoch": 0.13766751177109743, "grad_norm": 1.4997954648237672, "learning_rate": 1.9398396452122328e-05, "loss": 0.7897, "step": 3801 }, { "epoch": 0.1377037305324158, "grad_norm": 1.494906890638046, "learning_rate": 1.9397995648842988e-05, "loss": 0.819, "step": 3802 }, { "epoch": 0.13773994929373415, "grad_norm": 1.5599083119154913, "learning_rate": 1.939759471623897e-05, "loss": 0.7993, "step": 3803 }, { "epoch": 0.13777616805505252, "grad_norm": 1.3613229800917934, "learning_rate": 1.939719365431579e-05, "loss": 0.799, "step": 3804 }, { "epoch": 0.13781238681637087, "grad_norm": 1.5642193823228105, "learning_rate": 1.9396792463078972e-05, "loss": 0.904, "step": 3805 }, { "epoch": 0.13784860557768924, "grad_norm": 1.637073576255206, "learning_rate": 1.939639114253403e-05, "loss": 0.8305, "step": 3806 }, { "epoch": 0.13788482433900762, "grad_norm": 1.3724516912116023, "learning_rate": 1.9395989692686487e-05, "loss": 0.8299, "step": 3807 }, { "epoch": 0.13792104310032596, "grad_norm": 1.771894856371958, "learning_rate": 1.9395588113541875e-05, "loss": 0.8377, "step": 3808 }, { "epoch": 0.13795726186164434, "grad_norm": 1.4714689692436878, "learning_rate": 1.9395186405105713e-05, "loss": 0.7755, "step": 3809 }, { "epoch": 0.13799348062296268, "grad_norm": 1.7138464880233981, "learning_rate": 1.9394784567383528e-05, "loss": 0.8006, "step": 3810 }, { "epoch": 0.13802969938428106, "grad_norm": 1.4464776703681468, "learning_rate": 1.9394382600380855e-05, "loss": 0.855, "step": 3811 }, { "epoch": 0.13806591814559943, "grad_norm": 1.7109618054583957, "learning_rate": 1.9393980504103225e-05, "loss": 0.8689, "step": 3812 }, { "epoch": 0.13810213690691778, "grad_norm": 1.6074025812300559, "learning_rate": 1.9393578278556163e-05, "loss": 0.841, "step": 3813 }, { "epoch": 0.13813835566823615, "grad_norm": 1.6182096851296754, "learning_rate": 1.9393175923745215e-05, "loss": 0.8982, "step": 3814 }, { "epoch": 0.1381745744295545, "grad_norm": 1.575947310062392, "learning_rate": 1.9392773439675912e-05, "loss": 0.7937, "step": 3815 }, { "epoch": 0.13821079319087287, "grad_norm": 1.5530881166875796, "learning_rate": 1.9392370826353792e-05, "loss": 0.8672, "step": 3816 }, { "epoch": 0.13824701195219125, "grad_norm": 1.6739996941235336, "learning_rate": 1.93919680837844e-05, "loss": 0.9513, "step": 3817 }, { "epoch": 0.1382832307135096, "grad_norm": 1.454555677222042, "learning_rate": 1.9391565211973268e-05, "loss": 0.7425, "step": 3818 }, { "epoch": 0.13831944947482797, "grad_norm": 1.6675234999905961, "learning_rate": 1.939116221092595e-05, "loss": 0.8805, "step": 3819 }, { "epoch": 0.1383556682361463, "grad_norm": 1.5907651728938066, "learning_rate": 1.939075908064799e-05, "loss": 0.7895, "step": 3820 }, { "epoch": 0.13839188699746469, "grad_norm": 1.5355295420231008, "learning_rate": 1.9390355821144934e-05, "loss": 0.9071, "step": 3821 }, { "epoch": 0.13842810575878306, "grad_norm": 1.606190836963492, "learning_rate": 1.9389952432422328e-05, "loss": 0.8992, "step": 3822 }, { "epoch": 0.1384643245201014, "grad_norm": 1.748505791036977, "learning_rate": 1.9389548914485724e-05, "loss": 0.7907, "step": 3823 }, { "epoch": 0.13850054328141978, "grad_norm": 1.6049542328448707, "learning_rate": 1.938914526734068e-05, "loss": 0.8056, "step": 3824 }, { "epoch": 0.13853676204273813, "grad_norm": 1.5056624722914638, "learning_rate": 1.9388741490992746e-05, "loss": 0.8213, "step": 3825 }, { "epoch": 0.1385729808040565, "grad_norm": 1.6291985872657657, "learning_rate": 1.9388337585447482e-05, "loss": 0.8036, "step": 3826 }, { "epoch": 0.13860919956537487, "grad_norm": 1.6758218134570153, "learning_rate": 1.938793355071044e-05, "loss": 0.8489, "step": 3827 }, { "epoch": 0.13864541832669322, "grad_norm": 1.2732850782360792, "learning_rate": 1.9387529386787187e-05, "loss": 0.809, "step": 3828 }, { "epoch": 0.1386816370880116, "grad_norm": 1.6289785904223864, "learning_rate": 1.938712509368328e-05, "loss": 0.88, "step": 3829 }, { "epoch": 0.13871785584932994, "grad_norm": 1.267793695705306, "learning_rate": 1.9386720671404277e-05, "loss": 0.7756, "step": 3830 }, { "epoch": 0.13875407461064831, "grad_norm": 1.548866701126744, "learning_rate": 1.9386316119955757e-05, "loss": 0.8609, "step": 3831 }, { "epoch": 0.1387902933719667, "grad_norm": 1.4865866942448653, "learning_rate": 1.9385911439343278e-05, "loss": 0.8872, "step": 3832 }, { "epoch": 0.13882651213328503, "grad_norm": 1.6387129910326996, "learning_rate": 1.938550662957241e-05, "loss": 0.8185, "step": 3833 }, { "epoch": 0.1388627308946034, "grad_norm": 1.1888650190121088, "learning_rate": 1.938510169064872e-05, "loss": 0.7836, "step": 3834 }, { "epoch": 0.13889894965592176, "grad_norm": 1.3137223015608719, "learning_rate": 1.938469662257779e-05, "loss": 0.8333, "step": 3835 }, { "epoch": 0.13893516841724013, "grad_norm": 1.5196344088078524, "learning_rate": 1.9384291425365186e-05, "loss": 0.7991, "step": 3836 }, { "epoch": 0.1389713871785585, "grad_norm": 1.4863031014906274, "learning_rate": 1.9383886099016485e-05, "loss": 0.8534, "step": 3837 }, { "epoch": 0.13900760593987685, "grad_norm": 1.5678121365012039, "learning_rate": 1.938348064353727e-05, "loss": 0.7673, "step": 3838 }, { "epoch": 0.13904382470119522, "grad_norm": 1.2669422087996962, "learning_rate": 1.9383075058933113e-05, "loss": 0.7792, "step": 3839 }, { "epoch": 0.13908004346251357, "grad_norm": 1.8319197073777769, "learning_rate": 1.93826693452096e-05, "loss": 0.8626, "step": 3840 }, { "epoch": 0.13911626222383194, "grad_norm": 1.705312461524475, "learning_rate": 1.938226350237231e-05, "loss": 0.7807, "step": 3841 }, { "epoch": 0.13915248098515032, "grad_norm": 1.5076577800223483, "learning_rate": 1.938185753042683e-05, "loss": 0.8759, "step": 3842 }, { "epoch": 0.13918869974646866, "grad_norm": 1.5669024316595086, "learning_rate": 1.938145142937875e-05, "loss": 0.8012, "step": 3843 }, { "epoch": 0.13922491850778704, "grad_norm": 1.5742358545173964, "learning_rate": 1.9381045199233655e-05, "loss": 0.8253, "step": 3844 }, { "epoch": 0.13926113726910538, "grad_norm": 1.1945621443134566, "learning_rate": 1.9380638839997135e-05, "loss": 0.7604, "step": 3845 }, { "epoch": 0.13929735603042376, "grad_norm": 1.5633935936113916, "learning_rate": 1.938023235167478e-05, "loss": 0.8016, "step": 3846 }, { "epoch": 0.13933357479174213, "grad_norm": 1.6547305738869151, "learning_rate": 1.9379825734272186e-05, "loss": 0.9008, "step": 3847 }, { "epoch": 0.13936979355306048, "grad_norm": 1.5528958038771103, "learning_rate": 1.9379418987794948e-05, "loss": 0.9031, "step": 3848 }, { "epoch": 0.13940601231437885, "grad_norm": 1.8572519467318327, "learning_rate": 1.937901211224867e-05, "loss": 0.9155, "step": 3849 }, { "epoch": 0.1394422310756972, "grad_norm": 1.2390721285705661, "learning_rate": 1.9378605107638932e-05, "loss": 0.8259, "step": 3850 }, { "epoch": 0.13947844983701557, "grad_norm": 1.4917330524210046, "learning_rate": 1.9378197973971357e-05, "loss": 0.8953, "step": 3851 }, { "epoch": 0.13951466859833395, "grad_norm": 1.6605674245689468, "learning_rate": 1.937779071125153e-05, "loss": 0.8674, "step": 3852 }, { "epoch": 0.1395508873596523, "grad_norm": 1.6127555362377695, "learning_rate": 1.9377383319485068e-05, "loss": 0.8881, "step": 3853 }, { "epoch": 0.13958710612097067, "grad_norm": 1.580777889194217, "learning_rate": 1.937697579867757e-05, "loss": 0.8735, "step": 3854 }, { "epoch": 0.139623324882289, "grad_norm": 1.5000088202197315, "learning_rate": 1.9376568148834645e-05, "loss": 0.8458, "step": 3855 }, { "epoch": 0.1396595436436074, "grad_norm": 1.4523560254397483, "learning_rate": 1.9376160369961904e-05, "loss": 0.8187, "step": 3856 }, { "epoch": 0.13969576240492576, "grad_norm": 1.4988493429973213, "learning_rate": 1.9375752462064957e-05, "loss": 0.8746, "step": 3857 }, { "epoch": 0.1397319811662441, "grad_norm": 1.9166963400478771, "learning_rate": 1.9375344425149422e-05, "loss": 0.8349, "step": 3858 }, { "epoch": 0.13976819992756248, "grad_norm": 1.6479393705850847, "learning_rate": 1.9374936259220903e-05, "loss": 0.9261, "step": 3859 }, { "epoch": 0.13980441868888083, "grad_norm": 1.6234198662687775, "learning_rate": 1.9374527964285027e-05, "loss": 0.8953, "step": 3860 }, { "epoch": 0.1398406374501992, "grad_norm": 1.2489623128625031, "learning_rate": 1.9374119540347413e-05, "loss": 0.7737, "step": 3861 }, { "epoch": 0.13987685621151758, "grad_norm": 1.5509479082432003, "learning_rate": 1.937371098741367e-05, "loss": 0.8131, "step": 3862 }, { "epoch": 0.13991307497283592, "grad_norm": 1.5524146212625811, "learning_rate": 1.937330230548943e-05, "loss": 0.8154, "step": 3863 }, { "epoch": 0.1399492937341543, "grad_norm": 1.449290899706835, "learning_rate": 1.937289349458032e-05, "loss": 0.8318, "step": 3864 }, { "epoch": 0.13998551249547264, "grad_norm": 1.348246129917961, "learning_rate": 1.9372484554691954e-05, "loss": 0.8235, "step": 3865 }, { "epoch": 0.14002173125679102, "grad_norm": 1.0372212325741865, "learning_rate": 1.9372075485829966e-05, "loss": 0.8044, "step": 3866 }, { "epoch": 0.1400579500181094, "grad_norm": 1.5395180547195793, "learning_rate": 1.937166628799998e-05, "loss": 0.8942, "step": 3867 }, { "epoch": 0.14009416877942774, "grad_norm": 1.5917285284335716, "learning_rate": 1.9371256961207637e-05, "loss": 0.8213, "step": 3868 }, { "epoch": 0.1401303875407461, "grad_norm": 1.3841998068473773, "learning_rate": 1.9370847505458562e-05, "loss": 0.8534, "step": 3869 }, { "epoch": 0.14016660630206446, "grad_norm": 1.3023893261819794, "learning_rate": 1.937043792075839e-05, "loss": 0.7973, "step": 3870 }, { "epoch": 0.14020282506338283, "grad_norm": 1.5919953251801795, "learning_rate": 1.937002820711276e-05, "loss": 0.8507, "step": 3871 }, { "epoch": 0.1402390438247012, "grad_norm": 1.5644891237077136, "learning_rate": 1.9369618364527308e-05, "loss": 0.7911, "step": 3872 }, { "epoch": 0.14027526258601955, "grad_norm": 1.7070112892834461, "learning_rate": 1.9369208393007678e-05, "loss": 0.8226, "step": 3873 }, { "epoch": 0.14031148134733792, "grad_norm": 1.4920382599821254, "learning_rate": 1.9368798292559506e-05, "loss": 0.7759, "step": 3874 }, { "epoch": 0.14034770010865627, "grad_norm": 1.5500903440206986, "learning_rate": 1.9368388063188437e-05, "loss": 0.8194, "step": 3875 }, { "epoch": 0.14038391886997464, "grad_norm": 1.670437011681284, "learning_rate": 1.9367977704900113e-05, "loss": 0.838, "step": 3876 }, { "epoch": 0.14042013763129302, "grad_norm": 1.744113356963417, "learning_rate": 1.936756721770019e-05, "loss": 0.8521, "step": 3877 }, { "epoch": 0.14045635639261136, "grad_norm": 1.2666801064405702, "learning_rate": 1.9367156601594308e-05, "loss": 0.7903, "step": 3878 }, { "epoch": 0.14049257515392974, "grad_norm": 1.590417308588209, "learning_rate": 1.936674585658812e-05, "loss": 0.764, "step": 3879 }, { "epoch": 0.14052879391524808, "grad_norm": 1.474676116402452, "learning_rate": 1.936633498268728e-05, "loss": 0.7605, "step": 3880 }, { "epoch": 0.14056501267656646, "grad_norm": 1.723430881072517, "learning_rate": 1.936592397989744e-05, "loss": 0.9339, "step": 3881 }, { "epoch": 0.14060123143788483, "grad_norm": 1.5431432866719854, "learning_rate": 1.9365512848224257e-05, "loss": 0.8786, "step": 3882 }, { "epoch": 0.14063745019920318, "grad_norm": 1.549685618760882, "learning_rate": 1.9365101587673388e-05, "loss": 0.8258, "step": 3883 }, { "epoch": 0.14067366896052155, "grad_norm": 1.308904262922296, "learning_rate": 1.936469019825049e-05, "loss": 0.7989, "step": 3884 }, { "epoch": 0.14070988772183993, "grad_norm": 1.5755036697293505, "learning_rate": 1.936427867996123e-05, "loss": 0.899, "step": 3885 }, { "epoch": 0.14074610648315827, "grad_norm": 1.106575089952253, "learning_rate": 1.9363867032811267e-05, "loss": 0.7325, "step": 3886 }, { "epoch": 0.14078232524447665, "grad_norm": 1.4666766055155989, "learning_rate": 1.9363455256806263e-05, "loss": 0.9191, "step": 3887 }, { "epoch": 0.140818544005795, "grad_norm": 1.5108809872801587, "learning_rate": 1.9363043351951888e-05, "loss": 0.8008, "step": 3888 }, { "epoch": 0.14085476276711337, "grad_norm": 1.6494710922545466, "learning_rate": 1.9362631318253807e-05, "loss": 0.8853, "step": 3889 }, { "epoch": 0.14089098152843174, "grad_norm": 1.8341765947938358, "learning_rate": 1.9362219155717698e-05, "loss": 0.8723, "step": 3890 }, { "epoch": 0.1409272002897501, "grad_norm": 1.4769996698852372, "learning_rate": 1.9361806864349223e-05, "loss": 0.8333, "step": 3891 }, { "epoch": 0.14096341905106846, "grad_norm": 1.5588217869631047, "learning_rate": 1.9361394444154063e-05, "loss": 0.8094, "step": 3892 }, { "epoch": 0.1409996378123868, "grad_norm": 1.5612863917782347, "learning_rate": 1.9360981895137888e-05, "loss": 0.8637, "step": 3893 }, { "epoch": 0.14103585657370518, "grad_norm": 1.5762423732359028, "learning_rate": 1.9360569217306375e-05, "loss": 0.8363, "step": 3894 }, { "epoch": 0.14107207533502356, "grad_norm": 1.522946112196317, "learning_rate": 1.9360156410665206e-05, "loss": 0.8751, "step": 3895 }, { "epoch": 0.1411082940963419, "grad_norm": 1.6306156318939358, "learning_rate": 1.9359743475220057e-05, "loss": 0.8681, "step": 3896 }, { "epoch": 0.14114451285766028, "grad_norm": 1.24832603319488, "learning_rate": 1.935933041097662e-05, "loss": 0.8281, "step": 3897 }, { "epoch": 0.14118073161897862, "grad_norm": 1.5420499909180418, "learning_rate": 1.935891721794057e-05, "loss": 0.8217, "step": 3898 }, { "epoch": 0.141216950380297, "grad_norm": 1.6472703057996732, "learning_rate": 1.9358503896117594e-05, "loss": 0.8483, "step": 3899 }, { "epoch": 0.14125316914161537, "grad_norm": 1.7946246325727446, "learning_rate": 1.9358090445513383e-05, "loss": 0.8493, "step": 3900 }, { "epoch": 0.14128938790293372, "grad_norm": 1.7034233514383286, "learning_rate": 1.9357676866133625e-05, "loss": 0.8572, "step": 3901 }, { "epoch": 0.1413256066642521, "grad_norm": 1.5209321414244894, "learning_rate": 1.935726315798401e-05, "loss": 0.8174, "step": 3902 }, { "epoch": 0.14136182542557044, "grad_norm": 1.4426747951815355, "learning_rate": 1.935684932107023e-05, "loss": 0.7866, "step": 3903 }, { "epoch": 0.1413980441868888, "grad_norm": 1.55163475422531, "learning_rate": 1.9356435355397988e-05, "loss": 0.9142, "step": 3904 }, { "epoch": 0.14143426294820718, "grad_norm": 1.577676396533626, "learning_rate": 1.9356021260972965e-05, "loss": 0.8809, "step": 3905 }, { "epoch": 0.14147048170952553, "grad_norm": 1.7313811219823017, "learning_rate": 1.9355607037800875e-05, "loss": 0.8413, "step": 3906 }, { "epoch": 0.1415067004708439, "grad_norm": 1.5250346509314623, "learning_rate": 1.9355192685887408e-05, "loss": 0.7614, "step": 3907 }, { "epoch": 0.14154291923216225, "grad_norm": 1.5630062940013039, "learning_rate": 1.9354778205238275e-05, "loss": 0.8593, "step": 3908 }, { "epoch": 0.14157913799348062, "grad_norm": 1.606904350948942, "learning_rate": 1.9354363595859168e-05, "loss": 0.7951, "step": 3909 }, { "epoch": 0.141615356754799, "grad_norm": 1.5620739051863197, "learning_rate": 1.93539488577558e-05, "loss": 0.8733, "step": 3910 }, { "epoch": 0.14165157551611735, "grad_norm": 1.7911311618548242, "learning_rate": 1.935353399093388e-05, "loss": 0.833, "step": 3911 }, { "epoch": 0.14168779427743572, "grad_norm": 1.4111494948546124, "learning_rate": 1.935311899539911e-05, "loss": 0.7599, "step": 3912 }, { "epoch": 0.14172401303875407, "grad_norm": 1.2104842194186014, "learning_rate": 1.9352703871157208e-05, "loss": 0.7835, "step": 3913 }, { "epoch": 0.14176023180007244, "grad_norm": 1.5988303536092705, "learning_rate": 1.9352288618213883e-05, "loss": 0.8864, "step": 3914 }, { "epoch": 0.1417964505613908, "grad_norm": 1.7715438340711525, "learning_rate": 1.935187323657484e-05, "loss": 0.8868, "step": 3915 }, { "epoch": 0.14183266932270916, "grad_norm": 1.7484130838068306, "learning_rate": 1.9351457726245814e-05, "loss": 0.8611, "step": 3916 }, { "epoch": 0.14186888808402753, "grad_norm": 1.77922768555143, "learning_rate": 1.9351042087232508e-05, "loss": 0.802, "step": 3917 }, { "epoch": 0.14190510684534588, "grad_norm": 1.4954305765116302, "learning_rate": 1.9350626319540647e-05, "loss": 0.8095, "step": 3918 }, { "epoch": 0.14194132560666425, "grad_norm": 1.6505082387605172, "learning_rate": 1.9350210423175953e-05, "loss": 0.9639, "step": 3919 }, { "epoch": 0.14197754436798263, "grad_norm": 1.6122704270648973, "learning_rate": 1.9349794398144146e-05, "loss": 0.8666, "step": 3920 }, { "epoch": 0.14201376312930097, "grad_norm": 1.7006330403194303, "learning_rate": 1.9349378244450955e-05, "loss": 0.8762, "step": 3921 }, { "epoch": 0.14204998189061935, "grad_norm": 1.6839405596542867, "learning_rate": 1.9348961962102103e-05, "loss": 0.9039, "step": 3922 }, { "epoch": 0.1420862006519377, "grad_norm": 1.7108851897286093, "learning_rate": 1.9348545551103318e-05, "loss": 0.9798, "step": 3923 }, { "epoch": 0.14212241941325607, "grad_norm": 1.6186475911940919, "learning_rate": 1.9348129011460333e-05, "loss": 0.9274, "step": 3924 }, { "epoch": 0.14215863817457444, "grad_norm": 1.2434619798403024, "learning_rate": 1.934771234317888e-05, "loss": 0.8459, "step": 3925 }, { "epoch": 0.1421948569358928, "grad_norm": 1.5714850398338898, "learning_rate": 1.934729554626469e-05, "loss": 0.906, "step": 3926 }, { "epoch": 0.14223107569721116, "grad_norm": 1.5434208412839299, "learning_rate": 1.93468786207235e-05, "loss": 0.8611, "step": 3927 }, { "epoch": 0.1422672944585295, "grad_norm": 1.7122985943293905, "learning_rate": 1.934646156656105e-05, "loss": 0.9, "step": 3928 }, { "epoch": 0.14230351321984788, "grad_norm": 1.3113887021574284, "learning_rate": 1.9346044383783074e-05, "loss": 0.815, "step": 3929 }, { "epoch": 0.14233973198116626, "grad_norm": 1.3324565821037124, "learning_rate": 1.9345627072395314e-05, "loss": 0.8481, "step": 3930 }, { "epoch": 0.1423759507424846, "grad_norm": 1.5800589205467193, "learning_rate": 1.9345209632403516e-05, "loss": 0.8104, "step": 3931 }, { "epoch": 0.14241216950380298, "grad_norm": 1.3214434138312896, "learning_rate": 1.934479206381342e-05, "loss": 0.8161, "step": 3932 }, { "epoch": 0.14244838826512132, "grad_norm": 1.7427884821898108, "learning_rate": 1.9344374366630777e-05, "loss": 0.882, "step": 3933 }, { "epoch": 0.1424846070264397, "grad_norm": 1.592853020247486, "learning_rate": 1.9343956540861328e-05, "loss": 0.8371, "step": 3934 }, { "epoch": 0.14252082578775807, "grad_norm": 1.1483180412132208, "learning_rate": 1.934353858651083e-05, "loss": 0.7357, "step": 3935 }, { "epoch": 0.14255704454907642, "grad_norm": 1.5094420578574927, "learning_rate": 1.934312050358503e-05, "loss": 0.8725, "step": 3936 }, { "epoch": 0.1425932633103948, "grad_norm": 1.7489634439135464, "learning_rate": 1.9342702292089682e-05, "loss": 0.8092, "step": 3937 }, { "epoch": 0.14262948207171314, "grad_norm": 1.7078246950340605, "learning_rate": 1.934228395203054e-05, "loss": 0.8323, "step": 3938 }, { "epoch": 0.1426657008330315, "grad_norm": 1.7884709533823406, "learning_rate": 1.9341865483413364e-05, "loss": 0.8833, "step": 3939 }, { "epoch": 0.14270191959434989, "grad_norm": 1.4108244865987978, "learning_rate": 1.934144688624391e-05, "loss": 0.7863, "step": 3940 }, { "epoch": 0.14273813835566823, "grad_norm": 1.6029075986842267, "learning_rate": 1.9341028160527943e-05, "loss": 0.8555, "step": 3941 }, { "epoch": 0.1427743571169866, "grad_norm": 1.6013466584906322, "learning_rate": 1.9340609306271214e-05, "loss": 0.7266, "step": 3942 }, { "epoch": 0.14281057587830495, "grad_norm": 1.7206841910508026, "learning_rate": 1.9340190323479495e-05, "loss": 0.931, "step": 3943 }, { "epoch": 0.14284679463962333, "grad_norm": 1.5347934359688327, "learning_rate": 1.9339771212158553e-05, "loss": 0.8694, "step": 3944 }, { "epoch": 0.1428830134009417, "grad_norm": 1.689918491458367, "learning_rate": 1.9339351972314147e-05, "loss": 0.8097, "step": 3945 }, { "epoch": 0.14291923216226005, "grad_norm": 1.597963061171044, "learning_rate": 1.933893260395206e-05, "loss": 0.7976, "step": 3946 }, { "epoch": 0.14295545092357842, "grad_norm": 1.7683627468369385, "learning_rate": 1.9338513107078046e-05, "loss": 0.8742, "step": 3947 }, { "epoch": 0.14299166968489677, "grad_norm": 1.969781536778181, "learning_rate": 1.9338093481697894e-05, "loss": 0.7857, "step": 3948 }, { "epoch": 0.14302788844621514, "grad_norm": 1.6469943273878875, "learning_rate": 1.9337673727817362e-05, "loss": 0.8574, "step": 3949 }, { "epoch": 0.14306410720753351, "grad_norm": 1.705304562279485, "learning_rate": 1.933725384544224e-05, "loss": 0.7861, "step": 3950 }, { "epoch": 0.14310032596885186, "grad_norm": 1.657314855665885, "learning_rate": 1.9336833834578298e-05, "loss": 0.8485, "step": 3951 }, { "epoch": 0.14313654473017023, "grad_norm": 1.6319917482313284, "learning_rate": 1.9336413695231323e-05, "loss": 0.9002, "step": 3952 }, { "epoch": 0.14317276349148858, "grad_norm": 1.710088025810792, "learning_rate": 1.933599342740709e-05, "loss": 0.9393, "step": 3953 }, { "epoch": 0.14320898225280695, "grad_norm": 1.2546638650289192, "learning_rate": 1.9335573031111382e-05, "loss": 0.8195, "step": 3954 }, { "epoch": 0.14324520101412533, "grad_norm": 1.311580046932485, "learning_rate": 1.9335152506349987e-05, "loss": 0.8454, "step": 3955 }, { "epoch": 0.14328141977544367, "grad_norm": 1.6016888673287302, "learning_rate": 1.933473185312869e-05, "loss": 0.83, "step": 3956 }, { "epoch": 0.14331763853676205, "grad_norm": 1.6358259111300366, "learning_rate": 1.9334311071453283e-05, "loss": 0.9673, "step": 3957 }, { "epoch": 0.1433538572980804, "grad_norm": 1.7159305400321672, "learning_rate": 1.933389016132955e-05, "loss": 0.9044, "step": 3958 }, { "epoch": 0.14339007605939877, "grad_norm": 1.8760655594009792, "learning_rate": 1.933346912276329e-05, "loss": 0.812, "step": 3959 }, { "epoch": 0.14342629482071714, "grad_norm": 1.6614950817751695, "learning_rate": 1.9333047955760293e-05, "loss": 0.942, "step": 3960 }, { "epoch": 0.1434625135820355, "grad_norm": 1.5233583545911267, "learning_rate": 1.9332626660326357e-05, "loss": 0.863, "step": 3961 }, { "epoch": 0.14349873234335386, "grad_norm": 1.7126842524523433, "learning_rate": 1.9332205236467277e-05, "loss": 0.9157, "step": 3962 }, { "epoch": 0.1435349511046722, "grad_norm": 1.4446352597905796, "learning_rate": 1.933178368418885e-05, "loss": 0.8357, "step": 3963 }, { "epoch": 0.14357116986599058, "grad_norm": 1.5864855542814649, "learning_rate": 1.9331362003496883e-05, "loss": 0.8237, "step": 3964 }, { "epoch": 0.14360738862730896, "grad_norm": 1.4802533234547024, "learning_rate": 1.9330940194397176e-05, "loss": 0.7881, "step": 3965 }, { "epoch": 0.1436436073886273, "grad_norm": 1.7999077041460885, "learning_rate": 1.933051825689553e-05, "loss": 0.8574, "step": 3966 }, { "epoch": 0.14367982614994568, "grad_norm": 1.6961018794500566, "learning_rate": 1.933009619099776e-05, "loss": 0.855, "step": 3967 }, { "epoch": 0.14371604491126402, "grad_norm": 1.635617987344316, "learning_rate": 1.932967399670966e-05, "loss": 0.782, "step": 3968 }, { "epoch": 0.1437522636725824, "grad_norm": 1.5417611466649377, "learning_rate": 1.9329251674037053e-05, "loss": 0.7861, "step": 3969 }, { "epoch": 0.14378848243390077, "grad_norm": 1.5849118823752877, "learning_rate": 1.9328829222985746e-05, "loss": 0.8422, "step": 3970 }, { "epoch": 0.14382470119521912, "grad_norm": 1.4997089050446697, "learning_rate": 1.932840664356155e-05, "loss": 0.7489, "step": 3971 }, { "epoch": 0.1438609199565375, "grad_norm": 1.6958746686678037, "learning_rate": 1.9327983935770286e-05, "loss": 0.8735, "step": 3972 }, { "epoch": 0.14389713871785584, "grad_norm": 1.0794924556658685, "learning_rate": 1.932756109961776e-05, "loss": 0.7877, "step": 3973 }, { "epoch": 0.1439333574791742, "grad_norm": 1.5958916483888628, "learning_rate": 1.9327138135109803e-05, "loss": 0.8225, "step": 3974 }, { "epoch": 0.14396957624049259, "grad_norm": 1.5630215483691055, "learning_rate": 1.9326715042252226e-05, "loss": 0.8762, "step": 3975 }, { "epoch": 0.14400579500181093, "grad_norm": 1.489980384213427, "learning_rate": 1.932629182105086e-05, "loss": 0.765, "step": 3976 }, { "epoch": 0.1440420137631293, "grad_norm": 1.5460440301442284, "learning_rate": 1.932586847151152e-05, "loss": 0.843, "step": 3977 }, { "epoch": 0.14407823252444765, "grad_norm": 1.5200426747551172, "learning_rate": 1.9325444993640032e-05, "loss": 0.8117, "step": 3978 }, { "epoch": 0.14411445128576603, "grad_norm": 1.5163605210446558, "learning_rate": 1.932502138744223e-05, "loss": 0.8504, "step": 3979 }, { "epoch": 0.1441506700470844, "grad_norm": 1.4551784192793458, "learning_rate": 1.9324597652923943e-05, "loss": 0.8385, "step": 3980 }, { "epoch": 0.14418688880840275, "grad_norm": 1.449577097018143, "learning_rate": 1.9324173790090995e-05, "loss": 0.8015, "step": 3981 }, { "epoch": 0.14422310756972112, "grad_norm": 1.407184371416688, "learning_rate": 1.9323749798949225e-05, "loss": 0.7732, "step": 3982 }, { "epoch": 0.14425932633103947, "grad_norm": 1.5988339208502047, "learning_rate": 1.9323325679504464e-05, "loss": 0.832, "step": 3983 }, { "epoch": 0.14429554509235784, "grad_norm": 1.5611200986678793, "learning_rate": 1.932290143176255e-05, "loss": 0.8362, "step": 3984 }, { "epoch": 0.14433176385367621, "grad_norm": 1.3453641638987062, "learning_rate": 1.9322477055729317e-05, "loss": 0.8271, "step": 3985 }, { "epoch": 0.14436798261499456, "grad_norm": 1.5078473093284097, "learning_rate": 1.9322052551410617e-05, "loss": 0.8415, "step": 3986 }, { "epoch": 0.14440420137631294, "grad_norm": 1.8995151718411853, "learning_rate": 1.9321627918812273e-05, "loss": 0.8102, "step": 3987 }, { "epoch": 0.14444042013763128, "grad_norm": 1.3710872087567134, "learning_rate": 1.9321203157940144e-05, "loss": 0.7447, "step": 3988 }, { "epoch": 0.14447663889894966, "grad_norm": 1.477857837087439, "learning_rate": 1.9320778268800068e-05, "loss": 0.8495, "step": 3989 }, { "epoch": 0.14451285766026803, "grad_norm": 1.5708530394802545, "learning_rate": 1.932035325139789e-05, "loss": 0.8524, "step": 3990 }, { "epoch": 0.14454907642158638, "grad_norm": 1.3646871482389193, "learning_rate": 1.9319928105739464e-05, "loss": 0.7922, "step": 3991 }, { "epoch": 0.14458529518290475, "grad_norm": 1.7939122061459853, "learning_rate": 1.931950283183064e-05, "loss": 0.796, "step": 3992 }, { "epoch": 0.1446215139442231, "grad_norm": 1.5943091611893023, "learning_rate": 1.931907742967727e-05, "loss": 0.8622, "step": 3993 }, { "epoch": 0.14465773270554147, "grad_norm": 1.6175231884653236, "learning_rate": 1.9318651899285198e-05, "loss": 0.8479, "step": 3994 }, { "epoch": 0.14469395146685984, "grad_norm": 1.3847311682106505, "learning_rate": 1.9318226240660297e-05, "loss": 0.7904, "step": 3995 }, { "epoch": 0.1447301702281782, "grad_norm": 1.6627946431653442, "learning_rate": 1.9317800453808408e-05, "loss": 0.909, "step": 3996 }, { "epoch": 0.14476638898949656, "grad_norm": 1.2123994536774443, "learning_rate": 1.9317374538735402e-05, "loss": 0.8703, "step": 3997 }, { "epoch": 0.1448026077508149, "grad_norm": 1.5430076521786018, "learning_rate": 1.931694849544713e-05, "loss": 0.85, "step": 3998 }, { "epoch": 0.14483882651213328, "grad_norm": 1.5944879228330269, "learning_rate": 1.9316522323949466e-05, "loss": 0.8154, "step": 3999 }, { "epoch": 0.14487504527345166, "grad_norm": 1.4935419647295474, "learning_rate": 1.9316096024248265e-05, "loss": 0.8488, "step": 4000 }, { "epoch": 0.14491126403477, "grad_norm": 1.188511145130904, "learning_rate": 1.9315669596349397e-05, "loss": 0.8327, "step": 4001 }, { "epoch": 0.14494748279608838, "grad_norm": 1.5026958920752558, "learning_rate": 1.9315243040258733e-05, "loss": 0.8153, "step": 4002 }, { "epoch": 0.14498370155740672, "grad_norm": 1.6346561162769482, "learning_rate": 1.9314816355982138e-05, "loss": 0.837, "step": 4003 }, { "epoch": 0.1450199203187251, "grad_norm": 1.6185178769185917, "learning_rate": 1.9314389543525487e-05, "loss": 0.7889, "step": 4004 }, { "epoch": 0.14505613908004347, "grad_norm": 1.5125154903365916, "learning_rate": 1.9313962602894648e-05, "loss": 0.8085, "step": 4005 }, { "epoch": 0.14509235784136182, "grad_norm": 1.2124194718386614, "learning_rate": 1.93135355340955e-05, "loss": 0.7823, "step": 4006 }, { "epoch": 0.1451285766026802, "grad_norm": 1.76099706615039, "learning_rate": 1.931310833713392e-05, "loss": 0.9108, "step": 4007 }, { "epoch": 0.14516479536399854, "grad_norm": 1.373474998780671, "learning_rate": 1.931268101201579e-05, "loss": 0.8291, "step": 4008 }, { "epoch": 0.1452010141253169, "grad_norm": 1.2665857778784755, "learning_rate": 1.9312253558746984e-05, "loss": 0.8591, "step": 4009 }, { "epoch": 0.1452372328866353, "grad_norm": 1.5446831727357972, "learning_rate": 1.9311825977333383e-05, "loss": 0.8094, "step": 4010 }, { "epoch": 0.14527345164795363, "grad_norm": 1.3659888236764433, "learning_rate": 1.9311398267780878e-05, "loss": 0.7367, "step": 4011 }, { "epoch": 0.145309670409272, "grad_norm": 1.523783291184301, "learning_rate": 1.931097043009535e-05, "loss": 0.9174, "step": 4012 }, { "epoch": 0.14534588917059035, "grad_norm": 1.6835492719010288, "learning_rate": 1.9310542464282688e-05, "loss": 0.7862, "step": 4013 }, { "epoch": 0.14538210793190873, "grad_norm": 1.4024478646575862, "learning_rate": 1.931011437034878e-05, "loss": 0.8362, "step": 4014 }, { "epoch": 0.1454183266932271, "grad_norm": 1.995402694967865, "learning_rate": 1.930968614829952e-05, "loss": 0.865, "step": 4015 }, { "epoch": 0.14545454545454545, "grad_norm": 1.8631214057288807, "learning_rate": 1.9309257798140793e-05, "loss": 0.868, "step": 4016 }, { "epoch": 0.14549076421586382, "grad_norm": 1.6402261998626269, "learning_rate": 1.9308829319878503e-05, "loss": 0.8283, "step": 4017 }, { "epoch": 0.14552698297718217, "grad_norm": 1.7460128159508737, "learning_rate": 1.9308400713518542e-05, "loss": 0.8852, "step": 4018 }, { "epoch": 0.14556320173850054, "grad_norm": 1.5631139884051428, "learning_rate": 1.930797197906681e-05, "loss": 0.8786, "step": 4019 }, { "epoch": 0.14559942049981892, "grad_norm": 1.5249028625756125, "learning_rate": 1.93075431165292e-05, "loss": 0.7615, "step": 4020 }, { "epoch": 0.14563563926113726, "grad_norm": 1.5620210110100887, "learning_rate": 1.9307114125911622e-05, "loss": 0.8434, "step": 4021 }, { "epoch": 0.14567185802245564, "grad_norm": 2.2059614711485267, "learning_rate": 1.9306685007219975e-05, "loss": 0.8356, "step": 4022 }, { "epoch": 0.14570807678377398, "grad_norm": 1.7483770815149016, "learning_rate": 1.9306255760460165e-05, "loss": 0.8713, "step": 4023 }, { "epoch": 0.14574429554509236, "grad_norm": 1.522138447645665, "learning_rate": 1.9305826385638097e-05, "loss": 0.7569, "step": 4024 }, { "epoch": 0.14578051430641073, "grad_norm": 1.5423932028173268, "learning_rate": 1.9305396882759685e-05, "loss": 0.8228, "step": 4025 }, { "epoch": 0.14581673306772908, "grad_norm": 1.5433468162336543, "learning_rate": 1.930496725183083e-05, "loss": 0.8209, "step": 4026 }, { "epoch": 0.14585295182904745, "grad_norm": 1.8417215655515602, "learning_rate": 1.9304537492857453e-05, "loss": 0.8715, "step": 4027 }, { "epoch": 0.1458891705903658, "grad_norm": 1.6610390750512014, "learning_rate": 1.9304107605845465e-05, "loss": 0.7829, "step": 4028 }, { "epoch": 0.14592538935168417, "grad_norm": 1.7920284548215548, "learning_rate": 1.9303677590800775e-05, "loss": 0.8223, "step": 4029 }, { "epoch": 0.14596160811300254, "grad_norm": 1.4916448186766387, "learning_rate": 1.930324744772931e-05, "loss": 0.7695, "step": 4030 }, { "epoch": 0.1459978268743209, "grad_norm": 1.3154354944486812, "learning_rate": 1.930281717663699e-05, "loss": 0.7515, "step": 4031 }, { "epoch": 0.14603404563563926, "grad_norm": 1.7448020259824837, "learning_rate": 1.930238677752973e-05, "loss": 0.8261, "step": 4032 }, { "epoch": 0.1460702643969576, "grad_norm": 1.512089500257214, "learning_rate": 1.9301956250413448e-05, "loss": 0.824, "step": 4033 }, { "epoch": 0.14610648315827598, "grad_norm": 1.5512790475516824, "learning_rate": 1.930152559529408e-05, "loss": 0.8937, "step": 4034 }, { "epoch": 0.14614270191959436, "grad_norm": 1.192636102945844, "learning_rate": 1.9301094812177546e-05, "loss": 0.8212, "step": 4035 }, { "epoch": 0.1461789206809127, "grad_norm": 1.5764518268186984, "learning_rate": 1.9300663901069775e-05, "loss": 0.8821, "step": 4036 }, { "epoch": 0.14621513944223108, "grad_norm": 1.7750043039538361, "learning_rate": 1.9300232861976695e-05, "loss": 0.8241, "step": 4037 }, { "epoch": 0.14625135820354943, "grad_norm": 1.5919763311785888, "learning_rate": 1.9299801694904238e-05, "loss": 0.8976, "step": 4038 }, { "epoch": 0.1462875769648678, "grad_norm": 1.6545785418686583, "learning_rate": 1.929937039985834e-05, "loss": 0.7408, "step": 4039 }, { "epoch": 0.14632379572618617, "grad_norm": 1.6594781545412023, "learning_rate": 1.9298938976844934e-05, "loss": 0.8915, "step": 4040 }, { "epoch": 0.14636001448750452, "grad_norm": 1.6804076857580652, "learning_rate": 1.9298507425869955e-05, "loss": 0.8214, "step": 4041 }, { "epoch": 0.1463962332488229, "grad_norm": 1.6550712056811003, "learning_rate": 1.9298075746939348e-05, "loss": 0.8955, "step": 4042 }, { "epoch": 0.14643245201014124, "grad_norm": 1.643219562891379, "learning_rate": 1.9297643940059046e-05, "loss": 0.8122, "step": 4043 }, { "epoch": 0.1464686707714596, "grad_norm": 1.6348642025557547, "learning_rate": 1.9297212005234994e-05, "loss": 0.8327, "step": 4044 }, { "epoch": 0.146504889532778, "grad_norm": 1.8885952244602384, "learning_rate": 1.9296779942473134e-05, "loss": 0.8605, "step": 4045 }, { "epoch": 0.14654110829409633, "grad_norm": 1.6169457024487177, "learning_rate": 1.9296347751779415e-05, "loss": 0.842, "step": 4046 }, { "epoch": 0.1465773270554147, "grad_norm": 1.5974473875016564, "learning_rate": 1.929591543315978e-05, "loss": 0.8851, "step": 4047 }, { "epoch": 0.14661354581673305, "grad_norm": 1.233819534207177, "learning_rate": 1.9295482986620185e-05, "loss": 0.7831, "step": 4048 }, { "epoch": 0.14664976457805143, "grad_norm": 1.5776008382274955, "learning_rate": 1.9295050412166572e-05, "loss": 0.8469, "step": 4049 }, { "epoch": 0.1466859833393698, "grad_norm": 1.426374342768687, "learning_rate": 1.9294617709804902e-05, "loss": 0.7245, "step": 4050 }, { "epoch": 0.14672220210068815, "grad_norm": 1.3484073694947167, "learning_rate": 1.929418487954112e-05, "loss": 0.8132, "step": 4051 }, { "epoch": 0.14675842086200652, "grad_norm": 1.6800278130372672, "learning_rate": 1.929375192138119e-05, "loss": 0.897, "step": 4052 }, { "epoch": 0.14679463962332487, "grad_norm": 1.5274961278471433, "learning_rate": 1.9293318835331068e-05, "loss": 0.8929, "step": 4053 }, { "epoch": 0.14683085838464324, "grad_norm": 1.6743337363889508, "learning_rate": 1.9292885621396714e-05, "loss": 0.7458, "step": 4054 }, { "epoch": 0.14686707714596162, "grad_norm": 1.7982491919119143, "learning_rate": 1.929245227958409e-05, "loss": 0.8528, "step": 4055 }, { "epoch": 0.14690329590727996, "grad_norm": 1.324342722429458, "learning_rate": 1.9292018809899152e-05, "loss": 0.8105, "step": 4056 }, { "epoch": 0.14693951466859834, "grad_norm": 1.5133469811285936, "learning_rate": 1.9291585212347876e-05, "loss": 0.837, "step": 4057 }, { "epoch": 0.1469757334299167, "grad_norm": 1.1985276166160763, "learning_rate": 1.929115148693622e-05, "loss": 0.833, "step": 4058 }, { "epoch": 0.14701195219123506, "grad_norm": 1.4771554421159547, "learning_rate": 1.9290717633670154e-05, "loss": 0.8118, "step": 4059 }, { "epoch": 0.14704817095255343, "grad_norm": 1.1774670323061902, "learning_rate": 1.9290283652555654e-05, "loss": 0.8057, "step": 4060 }, { "epoch": 0.14708438971387178, "grad_norm": 1.6562198105345884, "learning_rate": 1.9289849543598685e-05, "loss": 0.8677, "step": 4061 }, { "epoch": 0.14712060847519015, "grad_norm": 1.5498307483695775, "learning_rate": 1.9289415306805223e-05, "loss": 0.8017, "step": 4062 }, { "epoch": 0.14715682723650853, "grad_norm": 1.6281213350080876, "learning_rate": 1.928898094218125e-05, "loss": 0.9174, "step": 4063 }, { "epoch": 0.14719304599782687, "grad_norm": 1.6285971415717808, "learning_rate": 1.928854644973273e-05, "loss": 0.9202, "step": 4064 }, { "epoch": 0.14722926475914525, "grad_norm": 1.589797511353928, "learning_rate": 1.9288111829465653e-05, "loss": 0.85, "step": 4065 }, { "epoch": 0.1472654835204636, "grad_norm": 1.6380658285049028, "learning_rate": 1.928767708138599e-05, "loss": 0.7745, "step": 4066 }, { "epoch": 0.14730170228178197, "grad_norm": 1.5024001630664578, "learning_rate": 1.9287242205499736e-05, "loss": 0.8545, "step": 4067 }, { "epoch": 0.14733792104310034, "grad_norm": 1.5549537248640863, "learning_rate": 1.9286807201812866e-05, "loss": 0.8311, "step": 4068 }, { "epoch": 0.14737413980441869, "grad_norm": 1.561103600942568, "learning_rate": 1.928637207033137e-05, "loss": 0.7707, "step": 4069 }, { "epoch": 0.14741035856573706, "grad_norm": 1.578322840791142, "learning_rate": 1.9285936811061233e-05, "loss": 0.7813, "step": 4070 }, { "epoch": 0.1474465773270554, "grad_norm": 1.1985524488211925, "learning_rate": 1.928550142400845e-05, "loss": 0.8027, "step": 4071 }, { "epoch": 0.14748279608837378, "grad_norm": 1.5960834171410694, "learning_rate": 1.9285065909179006e-05, "loss": 0.8643, "step": 4072 }, { "epoch": 0.14751901484969215, "grad_norm": 1.764140722620163, "learning_rate": 1.9284630266578897e-05, "loss": 0.8615, "step": 4073 }, { "epoch": 0.1475552336110105, "grad_norm": 1.4196519489357478, "learning_rate": 1.9284194496214118e-05, "loss": 0.7797, "step": 4074 }, { "epoch": 0.14759145237232887, "grad_norm": 1.6846123643132207, "learning_rate": 1.9283758598090664e-05, "loss": 0.861, "step": 4075 }, { "epoch": 0.14762767113364722, "grad_norm": 1.7611215873462907, "learning_rate": 1.9283322572214533e-05, "loss": 0.9027, "step": 4076 }, { "epoch": 0.1476638898949656, "grad_norm": 1.5839575554756204, "learning_rate": 1.9282886418591725e-05, "loss": 0.7848, "step": 4077 }, { "epoch": 0.14770010865628397, "grad_norm": 1.6318777125802466, "learning_rate": 1.928245013722825e-05, "loss": 0.8427, "step": 4078 }, { "epoch": 0.14773632741760231, "grad_norm": 1.4515249187400638, "learning_rate": 1.9282013728130096e-05, "loss": 0.8499, "step": 4079 }, { "epoch": 0.1477725461789207, "grad_norm": 1.1394918587745848, "learning_rate": 1.928157719130328e-05, "loss": 0.8303, "step": 4080 }, { "epoch": 0.14780876494023903, "grad_norm": 1.6999581460235824, "learning_rate": 1.9281140526753808e-05, "loss": 0.836, "step": 4081 }, { "epoch": 0.1478449837015574, "grad_norm": 1.3424247951022226, "learning_rate": 1.9280703734487688e-05, "loss": 0.8341, "step": 4082 }, { "epoch": 0.14788120246287578, "grad_norm": 1.6197764456259705, "learning_rate": 1.928026681451093e-05, "loss": 0.8527, "step": 4083 }, { "epoch": 0.14791742122419413, "grad_norm": 1.6604600662032376, "learning_rate": 1.9279829766829546e-05, "loss": 0.8496, "step": 4084 }, { "epoch": 0.1479536399855125, "grad_norm": 1.1502941661190973, "learning_rate": 1.927939259144955e-05, "loss": 0.7896, "step": 4085 }, { "epoch": 0.14798985874683085, "grad_norm": 1.9092037540879916, "learning_rate": 1.9278955288376958e-05, "loss": 0.8985, "step": 4086 }, { "epoch": 0.14802607750814922, "grad_norm": 1.422917258311004, "learning_rate": 1.9278517857617788e-05, "loss": 0.7846, "step": 4087 }, { "epoch": 0.1480622962694676, "grad_norm": 1.5320445731272827, "learning_rate": 1.927808029917806e-05, "loss": 0.8406, "step": 4088 }, { "epoch": 0.14809851503078594, "grad_norm": 1.664851462235795, "learning_rate": 1.9277642613063794e-05, "loss": 0.8403, "step": 4089 }, { "epoch": 0.14813473379210432, "grad_norm": 1.8178868996227364, "learning_rate": 1.9277204799281014e-05, "loss": 0.8582, "step": 4090 }, { "epoch": 0.14817095255342266, "grad_norm": 1.6048992230422525, "learning_rate": 1.9276766857835746e-05, "loss": 0.858, "step": 4091 }, { "epoch": 0.14820717131474104, "grad_norm": 1.7704385805980831, "learning_rate": 1.927632878873401e-05, "loss": 0.8761, "step": 4092 }, { "epoch": 0.1482433900760594, "grad_norm": 1.488597859975267, "learning_rate": 1.9275890591981842e-05, "loss": 0.7807, "step": 4093 }, { "epoch": 0.14827960883737776, "grad_norm": 1.2082912223247895, "learning_rate": 1.927545226758527e-05, "loss": 0.7614, "step": 4094 }, { "epoch": 0.14831582759869613, "grad_norm": 1.5618948491775013, "learning_rate": 1.927501381555032e-05, "loss": 0.8978, "step": 4095 }, { "epoch": 0.14835204636001448, "grad_norm": 1.5973240814837688, "learning_rate": 1.9274575235883038e-05, "loss": 0.8592, "step": 4096 }, { "epoch": 0.14838826512133285, "grad_norm": 1.9053736454564023, "learning_rate": 1.9274136528589444e-05, "loss": 0.93, "step": 4097 }, { "epoch": 0.14842448388265123, "grad_norm": 1.2505812150996958, "learning_rate": 1.927369769367559e-05, "loss": 0.8358, "step": 4098 }, { "epoch": 0.14846070264396957, "grad_norm": 1.4088996270531646, "learning_rate": 1.92732587311475e-05, "loss": 0.7901, "step": 4099 }, { "epoch": 0.14849692140528795, "grad_norm": 1.6051255216584295, "learning_rate": 1.9272819641011226e-05, "loss": 0.8261, "step": 4100 }, { "epoch": 0.1485331401666063, "grad_norm": 1.3735448589111081, "learning_rate": 1.92723804232728e-05, "loss": 0.7783, "step": 4101 }, { "epoch": 0.14856935892792467, "grad_norm": 1.8440582858181154, "learning_rate": 1.927194107793828e-05, "loss": 0.9015, "step": 4102 }, { "epoch": 0.14860557768924304, "grad_norm": 1.471959941077437, "learning_rate": 1.92715016050137e-05, "loss": 0.7496, "step": 4103 }, { "epoch": 0.1486417964505614, "grad_norm": 1.4345424972704464, "learning_rate": 1.927106200450511e-05, "loss": 0.8123, "step": 4104 }, { "epoch": 0.14867801521187976, "grad_norm": 1.4848417301238503, "learning_rate": 1.9270622276418562e-05, "loss": 0.764, "step": 4105 }, { "epoch": 0.1487142339731981, "grad_norm": 1.2702327725796394, "learning_rate": 1.9270182420760104e-05, "loss": 0.7617, "step": 4106 }, { "epoch": 0.14875045273451648, "grad_norm": 1.0302529058399315, "learning_rate": 1.926974243753579e-05, "loss": 0.7879, "step": 4107 }, { "epoch": 0.14878667149583485, "grad_norm": 1.7530115836306706, "learning_rate": 1.9269302326751678e-05, "loss": 0.899, "step": 4108 }, { "epoch": 0.1488228902571532, "grad_norm": 1.658293539060294, "learning_rate": 1.926886208841382e-05, "loss": 0.8822, "step": 4109 }, { "epoch": 0.14885910901847157, "grad_norm": 1.2932826399290165, "learning_rate": 1.9268421722528273e-05, "loss": 0.7996, "step": 4110 }, { "epoch": 0.14889532777978992, "grad_norm": 1.363036243671441, "learning_rate": 1.9267981229101105e-05, "loss": 0.7541, "step": 4111 }, { "epoch": 0.1489315465411083, "grad_norm": 1.5655764405682466, "learning_rate": 1.9267540608138368e-05, "loss": 0.8646, "step": 4112 }, { "epoch": 0.14896776530242667, "grad_norm": 1.6590561834826916, "learning_rate": 1.9267099859646126e-05, "loss": 0.8383, "step": 4113 }, { "epoch": 0.14900398406374502, "grad_norm": 1.8143522421361944, "learning_rate": 1.9266658983630452e-05, "loss": 0.8249, "step": 4114 }, { "epoch": 0.1490402028250634, "grad_norm": 1.497962738854975, "learning_rate": 1.9266217980097407e-05, "loss": 0.7915, "step": 4115 }, { "epoch": 0.14907642158638174, "grad_norm": 1.1943294337041757, "learning_rate": 1.926577684905306e-05, "loss": 0.7331, "step": 4116 }, { "epoch": 0.1491126403477001, "grad_norm": 1.2682785900301201, "learning_rate": 1.926533559050348e-05, "loss": 0.7769, "step": 4117 }, { "epoch": 0.14914885910901848, "grad_norm": 1.6475458954238138, "learning_rate": 1.9264894204454742e-05, "loss": 0.8557, "step": 4118 }, { "epoch": 0.14918507787033683, "grad_norm": 1.7588639553164993, "learning_rate": 1.9264452690912924e-05, "loss": 0.8929, "step": 4119 }, { "epoch": 0.1492212966316552, "grad_norm": 1.740161788274312, "learning_rate": 1.9264011049884092e-05, "loss": 0.8881, "step": 4120 }, { "epoch": 0.14925751539297355, "grad_norm": 1.7259225824593076, "learning_rate": 1.9263569281374325e-05, "loss": 0.881, "step": 4121 }, { "epoch": 0.14929373415429192, "grad_norm": 1.7196263084238188, "learning_rate": 1.9263127385389707e-05, "loss": 0.8775, "step": 4122 }, { "epoch": 0.1493299529156103, "grad_norm": 1.6581197443600935, "learning_rate": 1.926268536193632e-05, "loss": 0.916, "step": 4123 }, { "epoch": 0.14936617167692864, "grad_norm": 1.5117896997327733, "learning_rate": 1.926224321102024e-05, "loss": 0.7825, "step": 4124 }, { "epoch": 0.14940239043824702, "grad_norm": 1.4871202405848758, "learning_rate": 1.9261800932647557e-05, "loss": 0.8725, "step": 4125 }, { "epoch": 0.14943860919956536, "grad_norm": 1.587300964140817, "learning_rate": 1.9261358526824356e-05, "loss": 0.7974, "step": 4126 }, { "epoch": 0.14947482796088374, "grad_norm": 1.584093504772962, "learning_rate": 1.9260915993556723e-05, "loss": 0.7912, "step": 4127 }, { "epoch": 0.1495110467222021, "grad_norm": 1.5277667356237146, "learning_rate": 1.9260473332850747e-05, "loss": 0.8771, "step": 4128 }, { "epoch": 0.14954726548352046, "grad_norm": 1.4770216467640107, "learning_rate": 1.9260030544712523e-05, "loss": 0.8557, "step": 4129 }, { "epoch": 0.14958348424483883, "grad_norm": 1.5975259807053015, "learning_rate": 1.9259587629148137e-05, "loss": 0.8645, "step": 4130 }, { "epoch": 0.14961970300615718, "grad_norm": 1.5705943447753954, "learning_rate": 1.9259144586163696e-05, "loss": 0.7213, "step": 4131 }, { "epoch": 0.14965592176747555, "grad_norm": 1.5325081215848615, "learning_rate": 1.9258701415765287e-05, "loss": 0.8376, "step": 4132 }, { "epoch": 0.14969214052879393, "grad_norm": 1.6658609884213393, "learning_rate": 1.925825811795901e-05, "loss": 0.9009, "step": 4133 }, { "epoch": 0.14972835929011227, "grad_norm": 1.1683670655720904, "learning_rate": 1.9257814692750967e-05, "loss": 0.7554, "step": 4134 }, { "epoch": 0.14976457805143065, "grad_norm": 1.1384509407588914, "learning_rate": 1.925737114014726e-05, "loss": 0.8055, "step": 4135 }, { "epoch": 0.149800796812749, "grad_norm": 1.5165181907734215, "learning_rate": 1.9256927460153992e-05, "loss": 0.8305, "step": 4136 }, { "epoch": 0.14983701557406737, "grad_norm": 1.344716766525358, "learning_rate": 1.925648365277727e-05, "loss": 0.8317, "step": 4137 }, { "epoch": 0.14987323433538574, "grad_norm": 1.5944596859770181, "learning_rate": 1.9256039718023195e-05, "loss": 0.8549, "step": 4138 }, { "epoch": 0.1499094530967041, "grad_norm": 1.5471555125563317, "learning_rate": 1.925559565589788e-05, "loss": 0.8759, "step": 4139 }, { "epoch": 0.14994567185802246, "grad_norm": 1.239328779168106, "learning_rate": 1.925515146640744e-05, "loss": 0.818, "step": 4140 }, { "epoch": 0.1499818906193408, "grad_norm": 1.2632361873144309, "learning_rate": 1.925470714955798e-05, "loss": 0.7822, "step": 4141 }, { "epoch": 0.15001810938065918, "grad_norm": 1.756773287560967, "learning_rate": 1.925426270535562e-05, "loss": 0.8372, "step": 4142 }, { "epoch": 0.15005432814197756, "grad_norm": 1.688516577736832, "learning_rate": 1.9253818133806473e-05, "loss": 0.8633, "step": 4143 }, { "epoch": 0.1500905469032959, "grad_norm": 1.849626913662247, "learning_rate": 1.9253373434916657e-05, "loss": 0.8789, "step": 4144 }, { "epoch": 0.15012676566461428, "grad_norm": 1.2678740159544406, "learning_rate": 1.9252928608692296e-05, "loss": 0.8396, "step": 4145 }, { "epoch": 0.15016298442593262, "grad_norm": 1.9232063489239875, "learning_rate": 1.9252483655139503e-05, "loss": 0.85, "step": 4146 }, { "epoch": 0.150199203187251, "grad_norm": 1.28584807113473, "learning_rate": 1.9252038574264403e-05, "loss": 0.7419, "step": 4147 }, { "epoch": 0.15023542194856937, "grad_norm": 1.7428584335772754, "learning_rate": 1.9251593366073126e-05, "loss": 0.9153, "step": 4148 }, { "epoch": 0.15027164070988772, "grad_norm": 1.5209783433640756, "learning_rate": 1.9251148030571795e-05, "loss": 0.7951, "step": 4149 }, { "epoch": 0.1503078594712061, "grad_norm": 1.5422832819433858, "learning_rate": 1.9250702567766538e-05, "loss": 0.8881, "step": 4150 }, { "epoch": 0.15034407823252444, "grad_norm": 1.9344778775388243, "learning_rate": 1.9250256977663487e-05, "loss": 0.9199, "step": 4151 }, { "epoch": 0.1503802969938428, "grad_norm": 1.2180796500557733, "learning_rate": 1.924981126026877e-05, "loss": 0.8139, "step": 4152 }, { "epoch": 0.15041651575516118, "grad_norm": 1.7434389255638878, "learning_rate": 1.924936541558852e-05, "loss": 0.8743, "step": 4153 }, { "epoch": 0.15045273451647953, "grad_norm": 1.6470273175910521, "learning_rate": 1.9248919443628878e-05, "loss": 0.9029, "step": 4154 }, { "epoch": 0.1504889532777979, "grad_norm": 1.594551795171082, "learning_rate": 1.9248473344395976e-05, "loss": 0.895, "step": 4155 }, { "epoch": 0.15052517203911625, "grad_norm": 1.3318282581689191, "learning_rate": 1.9248027117895954e-05, "loss": 0.7763, "step": 4156 }, { "epoch": 0.15056139080043462, "grad_norm": 1.4776269709448542, "learning_rate": 1.9247580764134955e-05, "loss": 0.8627, "step": 4157 }, { "epoch": 0.150597609561753, "grad_norm": 1.5337339794294562, "learning_rate": 1.9247134283119116e-05, "loss": 0.8705, "step": 4158 }, { "epoch": 0.15063382832307134, "grad_norm": 1.575082625906832, "learning_rate": 1.924668767485459e-05, "loss": 0.8295, "step": 4159 }, { "epoch": 0.15067004708438972, "grad_norm": 1.5214704285204035, "learning_rate": 1.9246240939347512e-05, "loss": 0.8725, "step": 4160 }, { "epoch": 0.15070626584570807, "grad_norm": 1.1757786016497775, "learning_rate": 1.9245794076604035e-05, "loss": 0.7708, "step": 4161 }, { "epoch": 0.15074248460702644, "grad_norm": 1.5961737096440902, "learning_rate": 1.9245347086630307e-05, "loss": 0.9233, "step": 4162 }, { "epoch": 0.1507787033683448, "grad_norm": 1.531048165370842, "learning_rate": 1.924489996943248e-05, "loss": 0.8568, "step": 4163 }, { "epoch": 0.15081492212966316, "grad_norm": 1.3665717536714155, "learning_rate": 1.9244452725016707e-05, "loss": 0.8023, "step": 4164 }, { "epoch": 0.15085114089098153, "grad_norm": 1.5348386836461034, "learning_rate": 1.924400535338914e-05, "loss": 0.8569, "step": 4165 }, { "epoch": 0.15088735965229988, "grad_norm": 1.6284917503010503, "learning_rate": 1.9243557854555937e-05, "loss": 0.8881, "step": 4166 }, { "epoch": 0.15092357841361825, "grad_norm": 1.5660122223605621, "learning_rate": 1.9243110228523254e-05, "loss": 0.8035, "step": 4167 }, { "epoch": 0.15095979717493663, "grad_norm": 1.4139280538503942, "learning_rate": 1.9242662475297255e-05, "loss": 0.8079, "step": 4168 }, { "epoch": 0.15099601593625497, "grad_norm": 1.2748522760523353, "learning_rate": 1.92422145948841e-05, "loss": 0.7419, "step": 4169 }, { "epoch": 0.15103223469757335, "grad_norm": 1.706391236377424, "learning_rate": 1.924176658728995e-05, "loss": 0.8296, "step": 4170 }, { "epoch": 0.1510684534588917, "grad_norm": 1.4785075855162533, "learning_rate": 1.9241318452520973e-05, "loss": 0.8157, "step": 4171 }, { "epoch": 0.15110467222021007, "grad_norm": 1.5121139593801378, "learning_rate": 1.924087019058333e-05, "loss": 0.8495, "step": 4172 }, { "epoch": 0.15114089098152844, "grad_norm": 1.5759536327358556, "learning_rate": 1.9240421801483195e-05, "loss": 0.8414, "step": 4173 }, { "epoch": 0.1511771097428468, "grad_norm": 1.443235446703203, "learning_rate": 1.923997328522674e-05, "loss": 0.7767, "step": 4174 }, { "epoch": 0.15121332850416516, "grad_norm": 1.41433301730182, "learning_rate": 1.9239524641820127e-05, "loss": 0.81, "step": 4175 }, { "epoch": 0.1512495472654835, "grad_norm": 1.6680591456019247, "learning_rate": 1.923907587126954e-05, "loss": 0.8167, "step": 4176 }, { "epoch": 0.15128576602680188, "grad_norm": 1.4491630164968734, "learning_rate": 1.9238626973581157e-05, "loss": 0.7434, "step": 4177 }, { "epoch": 0.15132198478812026, "grad_norm": 1.6306518308414026, "learning_rate": 1.9238177948761142e-05, "loss": 0.9336, "step": 4178 }, { "epoch": 0.1513582035494386, "grad_norm": 1.2439416043312685, "learning_rate": 1.923772879681568e-05, "loss": 0.7736, "step": 4179 }, { "epoch": 0.15139442231075698, "grad_norm": 1.1791117032225813, "learning_rate": 1.9237279517750958e-05, "loss": 0.8095, "step": 4180 }, { "epoch": 0.15143064107207532, "grad_norm": 1.184300363212651, "learning_rate": 1.923683011157315e-05, "loss": 0.7847, "step": 4181 }, { "epoch": 0.1514668598333937, "grad_norm": 1.681720074937797, "learning_rate": 1.9236380578288444e-05, "loss": 0.8381, "step": 4182 }, { "epoch": 0.15150307859471207, "grad_norm": 1.6631372010443983, "learning_rate": 1.9235930917903024e-05, "loss": 0.8473, "step": 4183 }, { "epoch": 0.15153929735603042, "grad_norm": 1.739151036252898, "learning_rate": 1.923548113042308e-05, "loss": 0.7947, "step": 4184 }, { "epoch": 0.1515755161173488, "grad_norm": 1.4856451190888897, "learning_rate": 1.9235031215854803e-05, "loss": 0.8605, "step": 4185 }, { "epoch": 0.15161173487866714, "grad_norm": 1.736772206891305, "learning_rate": 1.923458117420438e-05, "loss": 0.8582, "step": 4186 }, { "epoch": 0.1516479536399855, "grad_norm": 1.6476734795354597, "learning_rate": 1.9234131005478003e-05, "loss": 0.8779, "step": 4187 }, { "epoch": 0.15168417240130389, "grad_norm": 1.6149964774336205, "learning_rate": 1.9233680709681873e-05, "loss": 0.8891, "step": 4188 }, { "epoch": 0.15172039116262223, "grad_norm": 1.8755236053516935, "learning_rate": 1.9233230286822184e-05, "loss": 0.8643, "step": 4189 }, { "epoch": 0.1517566099239406, "grad_norm": 1.6842631648806037, "learning_rate": 1.923277973690513e-05, "loss": 0.931, "step": 4190 }, { "epoch": 0.15179282868525895, "grad_norm": 1.5919896950981278, "learning_rate": 1.9232329059936917e-05, "loss": 0.922, "step": 4191 }, { "epoch": 0.15182904744657733, "grad_norm": 1.643220389237415, "learning_rate": 1.9231878255923742e-05, "loss": 0.8778, "step": 4192 }, { "epoch": 0.1518652662078957, "grad_norm": 1.5552637994552152, "learning_rate": 1.9231427324871814e-05, "loss": 0.8108, "step": 4193 }, { "epoch": 0.15190148496921405, "grad_norm": 1.5001972725036834, "learning_rate": 1.923097626678733e-05, "loss": 0.8502, "step": 4194 }, { "epoch": 0.15193770373053242, "grad_norm": 1.6381254182994764, "learning_rate": 1.92305250816765e-05, "loss": 0.8542, "step": 4195 }, { "epoch": 0.15197392249185077, "grad_norm": 1.6151337972829418, "learning_rate": 1.923007376954554e-05, "loss": 0.8404, "step": 4196 }, { "epoch": 0.15201014125316914, "grad_norm": 1.4314017464093558, "learning_rate": 1.922962233040065e-05, "loss": 0.7652, "step": 4197 }, { "epoch": 0.1520463600144875, "grad_norm": 1.59384441797285, "learning_rate": 1.922917076424805e-05, "loss": 0.8192, "step": 4198 }, { "epoch": 0.15208257877580586, "grad_norm": 1.6040563833167265, "learning_rate": 1.922871907109395e-05, "loss": 0.8964, "step": 4199 }, { "epoch": 0.15211879753712423, "grad_norm": 1.170487733413635, "learning_rate": 1.9228267250944566e-05, "loss": 0.7225, "step": 4200 }, { "epoch": 0.15215501629844258, "grad_norm": 1.626569118724479, "learning_rate": 1.9227815303806117e-05, "loss": 0.8029, "step": 4201 }, { "epoch": 0.15219123505976095, "grad_norm": 1.6244021505744903, "learning_rate": 1.922736322968482e-05, "loss": 0.8662, "step": 4202 }, { "epoch": 0.15222745382107933, "grad_norm": 1.1584657169154902, "learning_rate": 1.9226911028586895e-05, "loss": 0.7081, "step": 4203 }, { "epoch": 0.15226367258239767, "grad_norm": 1.235229905314357, "learning_rate": 1.922645870051857e-05, "loss": 0.8089, "step": 4204 }, { "epoch": 0.15229989134371605, "grad_norm": 1.4312483652180492, "learning_rate": 1.9226006245486067e-05, "loss": 0.8127, "step": 4205 }, { "epoch": 0.1523361101050344, "grad_norm": 1.543959806351308, "learning_rate": 1.922555366349561e-05, "loss": 0.8421, "step": 4206 }, { "epoch": 0.15237232886635277, "grad_norm": 1.2180454559102594, "learning_rate": 1.922510095455343e-05, "loss": 0.8012, "step": 4207 }, { "epoch": 0.15240854762767114, "grad_norm": 1.5920613285682317, "learning_rate": 1.9224648118665752e-05, "loss": 0.8263, "step": 4208 }, { "epoch": 0.1524447663889895, "grad_norm": 1.637398692356929, "learning_rate": 1.922419515583881e-05, "loss": 0.8388, "step": 4209 }, { "epoch": 0.15248098515030786, "grad_norm": 1.7002265117703397, "learning_rate": 1.9223742066078838e-05, "loss": 0.8176, "step": 4210 }, { "epoch": 0.1525172039116262, "grad_norm": 1.5271598200282974, "learning_rate": 1.9223288849392073e-05, "loss": 0.8476, "step": 4211 }, { "epoch": 0.15255342267294458, "grad_norm": 1.4778807409526358, "learning_rate": 1.9222835505784746e-05, "loss": 0.827, "step": 4212 }, { "epoch": 0.15258964143426296, "grad_norm": 1.6275985177259962, "learning_rate": 1.9222382035263105e-05, "loss": 0.9101, "step": 4213 }, { "epoch": 0.1526258601955813, "grad_norm": 1.6436307020282848, "learning_rate": 1.9221928437833373e-05, "loss": 0.7487, "step": 4214 }, { "epoch": 0.15266207895689968, "grad_norm": 1.5797485837813843, "learning_rate": 1.9221474713501813e-05, "loss": 0.8919, "step": 4215 }, { "epoch": 0.15269829771821802, "grad_norm": 1.2893831862018965, "learning_rate": 1.922102086227465e-05, "loss": 0.6961, "step": 4216 }, { "epoch": 0.1527345164795364, "grad_norm": 1.3519334284095177, "learning_rate": 1.9220566884158144e-05, "loss": 0.8134, "step": 4217 }, { "epoch": 0.15277073524085477, "grad_norm": 1.5721945865054743, "learning_rate": 1.9220112779158532e-05, "loss": 0.7918, "step": 4218 }, { "epoch": 0.15280695400217312, "grad_norm": 1.1151317663639833, "learning_rate": 1.921965854728207e-05, "loss": 0.7816, "step": 4219 }, { "epoch": 0.1528431727634915, "grad_norm": 1.6238270570312394, "learning_rate": 1.9219204188535e-05, "loss": 0.8633, "step": 4220 }, { "epoch": 0.15287939152480984, "grad_norm": 1.752466834546257, "learning_rate": 1.9218749702923583e-05, "loss": 0.9319, "step": 4221 }, { "epoch": 0.1529156102861282, "grad_norm": 1.5759848645623082, "learning_rate": 1.921829509045407e-05, "loss": 0.8489, "step": 4222 }, { "epoch": 0.15295182904744659, "grad_norm": 1.7852377299071545, "learning_rate": 1.9217840351132718e-05, "loss": 0.8466, "step": 4223 }, { "epoch": 0.15298804780876493, "grad_norm": 1.5407810141374938, "learning_rate": 1.921738548496578e-05, "loss": 0.8202, "step": 4224 }, { "epoch": 0.1530242665700833, "grad_norm": 1.7047429099373954, "learning_rate": 1.9216930491959522e-05, "loss": 0.9289, "step": 4225 }, { "epoch": 0.15306048533140165, "grad_norm": 1.5114267403587909, "learning_rate": 1.9216475372120198e-05, "loss": 0.8157, "step": 4226 }, { "epoch": 0.15309670409272003, "grad_norm": 1.7239625901681566, "learning_rate": 1.9216020125454075e-05, "loss": 0.8598, "step": 4227 }, { "epoch": 0.1531329228540384, "grad_norm": 2.062542976307336, "learning_rate": 1.921556475196742e-05, "loss": 0.8191, "step": 4228 }, { "epoch": 0.15316914161535675, "grad_norm": 1.7110810894567847, "learning_rate": 1.92151092516665e-05, "loss": 0.8611, "step": 4229 }, { "epoch": 0.15320536037667512, "grad_norm": 1.5540104762508808, "learning_rate": 1.921465362455757e-05, "loss": 0.8965, "step": 4230 }, { "epoch": 0.1532415791379935, "grad_norm": 1.3144667966668009, "learning_rate": 1.9214197870646916e-05, "loss": 0.7975, "step": 4231 }, { "epoch": 0.15327779789931184, "grad_norm": 1.6049587920245103, "learning_rate": 1.92137419899408e-05, "loss": 0.9084, "step": 4232 }, { "epoch": 0.15331401666063021, "grad_norm": 1.4322718696237544, "learning_rate": 1.92132859824455e-05, "loss": 0.8088, "step": 4233 }, { "epoch": 0.15335023542194856, "grad_norm": 1.3152022306702233, "learning_rate": 1.921282984816729e-05, "loss": 0.766, "step": 4234 }, { "epoch": 0.15338645418326693, "grad_norm": 1.650842579913284, "learning_rate": 1.9212373587112443e-05, "loss": 0.7819, "step": 4235 }, { "epoch": 0.1534226729445853, "grad_norm": 1.5925961237291033, "learning_rate": 1.9211917199287243e-05, "loss": 0.8334, "step": 4236 }, { "epoch": 0.15345889170590366, "grad_norm": 1.5143740711089262, "learning_rate": 1.921146068469797e-05, "loss": 0.8511, "step": 4237 }, { "epoch": 0.15349511046722203, "grad_norm": 2.019334235625542, "learning_rate": 1.92110040433509e-05, "loss": 0.8812, "step": 4238 }, { "epoch": 0.15353132922854038, "grad_norm": 1.6876979707340036, "learning_rate": 1.921054727525232e-05, "loss": 0.8741, "step": 4239 }, { "epoch": 0.15356754798985875, "grad_norm": 1.6635813398015653, "learning_rate": 1.921009038040852e-05, "loss": 0.9231, "step": 4240 }, { "epoch": 0.15360376675117712, "grad_norm": 1.6225097934755386, "learning_rate": 1.9209633358825783e-05, "loss": 0.8663, "step": 4241 }, { "epoch": 0.15363998551249547, "grad_norm": 1.764499964165194, "learning_rate": 1.9209176210510398e-05, "loss": 0.8867, "step": 4242 }, { "epoch": 0.15367620427381384, "grad_norm": 1.3175870367354412, "learning_rate": 1.920871893546866e-05, "loss": 0.833, "step": 4243 }, { "epoch": 0.1537124230351322, "grad_norm": 1.83735495492779, "learning_rate": 1.9208261533706855e-05, "loss": 0.833, "step": 4244 }, { "epoch": 0.15374864179645056, "grad_norm": 1.835594522621279, "learning_rate": 1.920780400523128e-05, "loss": 0.8801, "step": 4245 }, { "epoch": 0.15378486055776894, "grad_norm": 1.530365095819946, "learning_rate": 1.920734635004823e-05, "loss": 0.8998, "step": 4246 }, { "epoch": 0.15382107931908728, "grad_norm": 1.7394166774259356, "learning_rate": 1.9206888568164003e-05, "loss": 0.827, "step": 4247 }, { "epoch": 0.15385729808040566, "grad_norm": 1.7028543077812128, "learning_rate": 1.9206430659584903e-05, "loss": 0.8445, "step": 4248 }, { "epoch": 0.153893516841724, "grad_norm": 1.6772125540632379, "learning_rate": 1.9205972624317227e-05, "loss": 0.9362, "step": 4249 }, { "epoch": 0.15392973560304238, "grad_norm": 1.2248019832900745, "learning_rate": 1.920551446236728e-05, "loss": 0.8305, "step": 4250 }, { "epoch": 0.15396595436436075, "grad_norm": 1.6000106876359061, "learning_rate": 1.920505617374136e-05, "loss": 0.9406, "step": 4251 }, { "epoch": 0.1540021731256791, "grad_norm": 1.5905172809931996, "learning_rate": 1.9204597758445784e-05, "loss": 0.7732, "step": 4252 }, { "epoch": 0.15403839188699747, "grad_norm": 1.4082826533723765, "learning_rate": 1.9204139216486854e-05, "loss": 0.8225, "step": 4253 }, { "epoch": 0.15407461064831582, "grad_norm": 1.2704775070480174, "learning_rate": 1.9203680547870877e-05, "loss": 0.7897, "step": 4254 }, { "epoch": 0.1541108294096342, "grad_norm": 1.2801829328998517, "learning_rate": 1.9203221752604173e-05, "loss": 0.794, "step": 4255 }, { "epoch": 0.15414704817095257, "grad_norm": 1.1998699599751224, "learning_rate": 1.920276283069305e-05, "loss": 0.8559, "step": 4256 }, { "epoch": 0.1541832669322709, "grad_norm": 1.6345712142797812, "learning_rate": 1.9202303782143822e-05, "loss": 0.8597, "step": 4257 }, { "epoch": 0.1542194856935893, "grad_norm": 1.3095136743283688, "learning_rate": 1.920184460696281e-05, "loss": 0.8015, "step": 4258 }, { "epoch": 0.15425570445490763, "grad_norm": 1.498711802119722, "learning_rate": 1.920138530515633e-05, "loss": 0.8458, "step": 4259 }, { "epoch": 0.154291923216226, "grad_norm": 1.5846563600103316, "learning_rate": 1.9200925876730705e-05, "loss": 0.7984, "step": 4260 }, { "epoch": 0.15432814197754438, "grad_norm": 1.6828722392417081, "learning_rate": 1.920046632169225e-05, "loss": 0.8639, "step": 4261 }, { "epoch": 0.15436436073886273, "grad_norm": 2.1807708884901347, "learning_rate": 1.9200006640047303e-05, "loss": 0.8265, "step": 4262 }, { "epoch": 0.1544005795001811, "grad_norm": 1.6495843272076898, "learning_rate": 1.9199546831802177e-05, "loss": 0.8706, "step": 4263 }, { "epoch": 0.15443679826149945, "grad_norm": 1.6572530920797137, "learning_rate": 1.91990868969632e-05, "loss": 0.9068, "step": 4264 }, { "epoch": 0.15447301702281782, "grad_norm": 1.5531434117176675, "learning_rate": 1.919862683553671e-05, "loss": 0.8083, "step": 4265 }, { "epoch": 0.1545092357841362, "grad_norm": 1.5909111448973705, "learning_rate": 1.9198166647529026e-05, "loss": 0.8558, "step": 4266 }, { "epoch": 0.15454545454545454, "grad_norm": 1.9001935276411677, "learning_rate": 1.9197706332946487e-05, "loss": 0.8028, "step": 4267 }, { "epoch": 0.15458167330677292, "grad_norm": 1.6568724468039573, "learning_rate": 1.919724589179543e-05, "loss": 0.8629, "step": 4268 }, { "epoch": 0.15461789206809126, "grad_norm": 1.7601271876040885, "learning_rate": 1.9196785324082186e-05, "loss": 0.8569, "step": 4269 }, { "epoch": 0.15465411082940964, "grad_norm": 1.7570510995438229, "learning_rate": 1.91963246298131e-05, "loss": 0.8477, "step": 4270 }, { "epoch": 0.154690329590728, "grad_norm": 1.6397620929089478, "learning_rate": 1.9195863808994503e-05, "loss": 0.9408, "step": 4271 }, { "epoch": 0.15472654835204636, "grad_norm": 1.199678302292047, "learning_rate": 1.919540286163274e-05, "loss": 0.8234, "step": 4272 }, { "epoch": 0.15476276711336473, "grad_norm": 1.572495950686437, "learning_rate": 1.9194941787734153e-05, "loss": 0.8922, "step": 4273 }, { "epoch": 0.15479898587468308, "grad_norm": 2.0571229591890323, "learning_rate": 1.9194480587305086e-05, "loss": 0.923, "step": 4274 }, { "epoch": 0.15483520463600145, "grad_norm": 1.6100945900961747, "learning_rate": 1.9194019260351892e-05, "loss": 0.8988, "step": 4275 }, { "epoch": 0.15487142339731982, "grad_norm": 1.5603780977459356, "learning_rate": 1.919355780688091e-05, "loss": 0.9159, "step": 4276 }, { "epoch": 0.15490764215863817, "grad_norm": 1.8504242310543495, "learning_rate": 1.9193096226898497e-05, "loss": 0.7925, "step": 4277 }, { "epoch": 0.15494386091995654, "grad_norm": 1.5116511571297326, "learning_rate": 1.9192634520411e-05, "loss": 0.8847, "step": 4278 }, { "epoch": 0.1549800796812749, "grad_norm": 1.149372054536037, "learning_rate": 1.9192172687424776e-05, "loss": 0.716, "step": 4279 }, { "epoch": 0.15501629844259326, "grad_norm": 1.1532634139459963, "learning_rate": 1.919171072794618e-05, "loss": 0.7399, "step": 4280 }, { "epoch": 0.15505251720391164, "grad_norm": 2.2878234930937613, "learning_rate": 1.9191248641981566e-05, "loss": 0.8758, "step": 4281 }, { "epoch": 0.15508873596522998, "grad_norm": 1.7641195350107008, "learning_rate": 1.9190786429537293e-05, "loss": 0.8735, "step": 4282 }, { "epoch": 0.15512495472654836, "grad_norm": 1.656080416103743, "learning_rate": 1.919032409061973e-05, "loss": 0.9102, "step": 4283 }, { "epoch": 0.1551611734878667, "grad_norm": 1.6679644595767729, "learning_rate": 1.9189861625235226e-05, "loss": 0.8164, "step": 4284 }, { "epoch": 0.15519739224918508, "grad_norm": 1.2140098321186765, "learning_rate": 1.9189399033390154e-05, "loss": 0.7668, "step": 4285 }, { "epoch": 0.15523361101050345, "grad_norm": 1.5204992263443717, "learning_rate": 1.9188936315090873e-05, "loss": 0.8393, "step": 4286 }, { "epoch": 0.1552698297718218, "grad_norm": 1.619914563587726, "learning_rate": 1.9188473470343755e-05, "loss": 0.8646, "step": 4287 }, { "epoch": 0.15530604853314017, "grad_norm": 1.6596021580252989, "learning_rate": 1.918801049915517e-05, "loss": 0.8172, "step": 4288 }, { "epoch": 0.15534226729445852, "grad_norm": 1.5753486449883773, "learning_rate": 1.9187547401531487e-05, "loss": 0.9085, "step": 4289 }, { "epoch": 0.1553784860557769, "grad_norm": 1.5379431055749067, "learning_rate": 1.9187084177479082e-05, "loss": 0.8175, "step": 4290 }, { "epoch": 0.15541470481709527, "grad_norm": 1.7459750679184332, "learning_rate": 1.918662082700432e-05, "loss": 0.9214, "step": 4291 }, { "epoch": 0.1554509235784136, "grad_norm": 1.5970632634818496, "learning_rate": 1.9186157350113586e-05, "loss": 0.8855, "step": 4292 }, { "epoch": 0.155487142339732, "grad_norm": 1.735576796418636, "learning_rate": 1.9185693746813258e-05, "loss": 0.8413, "step": 4293 }, { "epoch": 0.15552336110105033, "grad_norm": 1.8427455598777425, "learning_rate": 1.918523001710971e-05, "loss": 0.8274, "step": 4294 }, { "epoch": 0.1555595798623687, "grad_norm": 1.9141081911589097, "learning_rate": 1.918476616100933e-05, "loss": 0.8146, "step": 4295 }, { "epoch": 0.15559579862368708, "grad_norm": 1.5333821392638682, "learning_rate": 1.918430217851849e-05, "loss": 0.8709, "step": 4296 }, { "epoch": 0.15563201738500543, "grad_norm": 1.4444429768663818, "learning_rate": 1.9183838069643586e-05, "loss": 0.9244, "step": 4297 }, { "epoch": 0.1556682361463238, "grad_norm": 1.559160114104784, "learning_rate": 1.9183373834391e-05, "loss": 0.8859, "step": 4298 }, { "epoch": 0.15570445490764215, "grad_norm": 1.6649261302069318, "learning_rate": 1.9182909472767123e-05, "loss": 0.9048, "step": 4299 }, { "epoch": 0.15574067366896052, "grad_norm": 1.3415902345117736, "learning_rate": 1.9182444984778338e-05, "loss": 0.7864, "step": 4300 }, { "epoch": 0.1557768924302789, "grad_norm": 1.6855805814694678, "learning_rate": 1.9181980370431047e-05, "loss": 0.9288, "step": 4301 }, { "epoch": 0.15581311119159724, "grad_norm": 1.584079935928369, "learning_rate": 1.9181515629731632e-05, "loss": 0.9366, "step": 4302 }, { "epoch": 0.15584932995291562, "grad_norm": 1.6929713093835541, "learning_rate": 1.91810507626865e-05, "loss": 0.8772, "step": 4303 }, { "epoch": 0.15588554871423396, "grad_norm": 1.4954366763068976, "learning_rate": 1.918058576930204e-05, "loss": 0.7883, "step": 4304 }, { "epoch": 0.15592176747555234, "grad_norm": 1.6395578762670957, "learning_rate": 1.9180120649584655e-05, "loss": 0.8547, "step": 4305 }, { "epoch": 0.1559579862368707, "grad_norm": 1.6789295693274011, "learning_rate": 1.9179655403540742e-05, "loss": 0.861, "step": 4306 }, { "epoch": 0.15599420499818906, "grad_norm": 1.2641442475343772, "learning_rate": 1.9179190031176705e-05, "loss": 0.7722, "step": 4307 }, { "epoch": 0.15603042375950743, "grad_norm": 1.4210324325827983, "learning_rate": 1.9178724532498947e-05, "loss": 0.82, "step": 4308 }, { "epoch": 0.15606664252082578, "grad_norm": 1.0612899360930796, "learning_rate": 1.9178258907513872e-05, "loss": 0.7863, "step": 4309 }, { "epoch": 0.15610286128214415, "grad_norm": 1.4447255439298308, "learning_rate": 1.9177793156227894e-05, "loss": 0.8058, "step": 4310 }, { "epoch": 0.15613908004346252, "grad_norm": 1.0663181637423529, "learning_rate": 1.9177327278647416e-05, "loss": 0.7856, "step": 4311 }, { "epoch": 0.15617529880478087, "grad_norm": 1.9226599055801479, "learning_rate": 1.917686127477885e-05, "loss": 0.783, "step": 4312 }, { "epoch": 0.15621151756609924, "grad_norm": 1.627110376349024, "learning_rate": 1.9176395144628614e-05, "loss": 0.9182, "step": 4313 }, { "epoch": 0.1562477363274176, "grad_norm": 1.667024578590999, "learning_rate": 1.9175928888203113e-05, "loss": 0.8317, "step": 4314 }, { "epoch": 0.15628395508873597, "grad_norm": 1.412411458599856, "learning_rate": 1.9175462505508767e-05, "loss": 0.8479, "step": 4315 }, { "epoch": 0.15632017385005434, "grad_norm": 1.5250933475925719, "learning_rate": 1.9174995996551997e-05, "loss": 0.8665, "step": 4316 }, { "epoch": 0.15635639261137269, "grad_norm": 1.6680582500450647, "learning_rate": 1.9174529361339217e-05, "loss": 0.8612, "step": 4317 }, { "epoch": 0.15639261137269106, "grad_norm": 1.4375495131144282, "learning_rate": 1.9174062599876854e-05, "loss": 0.8261, "step": 4318 }, { "epoch": 0.1564288301340094, "grad_norm": 1.6439976290825815, "learning_rate": 1.917359571217133e-05, "loss": 0.9077, "step": 4319 }, { "epoch": 0.15646504889532778, "grad_norm": 1.5071550372558868, "learning_rate": 1.9173128698229066e-05, "loss": 0.812, "step": 4320 }, { "epoch": 0.15650126765664615, "grad_norm": 1.768040999509431, "learning_rate": 1.9172661558056494e-05, "loss": 0.8573, "step": 4321 }, { "epoch": 0.1565374864179645, "grad_norm": 1.5489944937491844, "learning_rate": 1.9172194291660037e-05, "loss": 0.7878, "step": 4322 }, { "epoch": 0.15657370517928287, "grad_norm": 1.6576543254788985, "learning_rate": 1.9171726899046126e-05, "loss": 0.7856, "step": 4323 }, { "epoch": 0.15660992394060122, "grad_norm": 1.3772035384580525, "learning_rate": 1.9171259380221196e-05, "loss": 0.9042, "step": 4324 }, { "epoch": 0.1566461427019196, "grad_norm": 1.6733044970396562, "learning_rate": 1.9170791735191677e-05, "loss": 0.7799, "step": 4325 }, { "epoch": 0.15668236146323797, "grad_norm": 1.6724015614381686, "learning_rate": 1.9170323963964007e-05, "loss": 0.7983, "step": 4326 }, { "epoch": 0.15671858022455631, "grad_norm": 1.5622823849772103, "learning_rate": 1.916985606654462e-05, "loss": 0.825, "step": 4327 }, { "epoch": 0.1567547989858747, "grad_norm": 1.8976613565676983, "learning_rate": 1.9169388042939956e-05, "loss": 0.818, "step": 4328 }, { "epoch": 0.15679101774719303, "grad_norm": 1.5024129411885179, "learning_rate": 1.9168919893156457e-05, "loss": 0.8336, "step": 4329 }, { "epoch": 0.1568272365085114, "grad_norm": 1.4890457427572001, "learning_rate": 1.9168451617200566e-05, "loss": 0.799, "step": 4330 }, { "epoch": 0.15686345526982978, "grad_norm": 1.628979253619541, "learning_rate": 1.916798321507872e-05, "loss": 0.7715, "step": 4331 }, { "epoch": 0.15689967403114813, "grad_norm": 1.5200395081907903, "learning_rate": 1.916751468679737e-05, "loss": 0.7423, "step": 4332 }, { "epoch": 0.1569358927924665, "grad_norm": 1.20635464457496, "learning_rate": 1.9167046032362964e-05, "loss": 0.8352, "step": 4333 }, { "epoch": 0.15697211155378485, "grad_norm": 1.6311303861517639, "learning_rate": 1.9166577251781947e-05, "loss": 0.8804, "step": 4334 }, { "epoch": 0.15700833031510322, "grad_norm": 1.51559621867155, "learning_rate": 1.9166108345060777e-05, "loss": 0.8634, "step": 4335 }, { "epoch": 0.1570445490764216, "grad_norm": 1.5865109741616483, "learning_rate": 1.91656393122059e-05, "loss": 0.7874, "step": 4336 }, { "epoch": 0.15708076783773994, "grad_norm": 1.5048112623272452, "learning_rate": 1.9165170153223773e-05, "loss": 0.8997, "step": 4337 }, { "epoch": 0.15711698659905832, "grad_norm": 1.2164161191263096, "learning_rate": 1.9164700868120852e-05, "loss": 0.7543, "step": 4338 }, { "epoch": 0.15715320536037666, "grad_norm": 1.658416655668064, "learning_rate": 1.9164231456903592e-05, "loss": 0.7983, "step": 4339 }, { "epoch": 0.15718942412169504, "grad_norm": 1.1973903829812955, "learning_rate": 1.9163761919578457e-05, "loss": 0.7763, "step": 4340 }, { "epoch": 0.1572256428830134, "grad_norm": 1.544777156885375, "learning_rate": 1.9163292256151902e-05, "loss": 0.8064, "step": 4341 }, { "epoch": 0.15726186164433176, "grad_norm": 1.5559178861888072, "learning_rate": 1.91628224666304e-05, "loss": 0.8679, "step": 4342 }, { "epoch": 0.15729808040565013, "grad_norm": 1.470790612200997, "learning_rate": 1.9162352551020408e-05, "loss": 0.8457, "step": 4343 }, { "epoch": 0.15733429916696848, "grad_norm": 1.4241556129805284, "learning_rate": 1.916188250932839e-05, "loss": 0.807, "step": 4344 }, { "epoch": 0.15737051792828685, "grad_norm": 1.6030871060328402, "learning_rate": 1.9161412341560824e-05, "loss": 0.8515, "step": 4345 }, { "epoch": 0.15740673668960523, "grad_norm": 1.6435309293347817, "learning_rate": 1.916094204772417e-05, "loss": 0.8072, "step": 4346 }, { "epoch": 0.15744295545092357, "grad_norm": 1.7123196301526762, "learning_rate": 1.9160471627824904e-05, "loss": 0.9167, "step": 4347 }, { "epoch": 0.15747917421224195, "grad_norm": 1.6098685925447864, "learning_rate": 1.91600010818695e-05, "loss": 0.8524, "step": 4348 }, { "epoch": 0.1575153929735603, "grad_norm": 1.5849704859663962, "learning_rate": 1.9159530409864434e-05, "loss": 0.8494, "step": 4349 }, { "epoch": 0.15755161173487867, "grad_norm": 1.5672554493195325, "learning_rate": 1.915905961181618e-05, "loss": 0.8119, "step": 4350 }, { "epoch": 0.15758783049619704, "grad_norm": 1.590993799171811, "learning_rate": 1.9158588687731217e-05, "loss": 0.7803, "step": 4351 }, { "epoch": 0.1576240492575154, "grad_norm": 1.6097779903663598, "learning_rate": 1.915811763761603e-05, "loss": 0.8662, "step": 4352 }, { "epoch": 0.15766026801883376, "grad_norm": 1.614287404402413, "learning_rate": 1.9157646461477096e-05, "loss": 0.7921, "step": 4353 }, { "epoch": 0.1576964867801521, "grad_norm": 1.5686142834245569, "learning_rate": 1.91571751593209e-05, "loss": 0.9372, "step": 4354 }, { "epoch": 0.15773270554147048, "grad_norm": 1.7565139382121815, "learning_rate": 1.9156703731153925e-05, "loss": 0.862, "step": 4355 }, { "epoch": 0.15776892430278885, "grad_norm": 1.5986214755616965, "learning_rate": 1.915623217698266e-05, "loss": 0.8689, "step": 4356 }, { "epoch": 0.1578051430641072, "grad_norm": 1.5030070130680462, "learning_rate": 1.9155760496813597e-05, "loss": 0.7244, "step": 4357 }, { "epoch": 0.15784136182542557, "grad_norm": 1.6338260541559932, "learning_rate": 1.9155288690653226e-05, "loss": 0.825, "step": 4358 }, { "epoch": 0.15787758058674392, "grad_norm": 1.6274365454508757, "learning_rate": 1.915481675850804e-05, "loss": 0.8914, "step": 4359 }, { "epoch": 0.1579137993480623, "grad_norm": 1.5799635148890936, "learning_rate": 1.9154344700384525e-05, "loss": 0.9259, "step": 4360 }, { "epoch": 0.15795001810938067, "grad_norm": 1.197134504617577, "learning_rate": 1.9153872516289187e-05, "loss": 0.8258, "step": 4361 }, { "epoch": 0.15798623687069902, "grad_norm": 1.6584166624069439, "learning_rate": 1.915340020622852e-05, "loss": 0.8944, "step": 4362 }, { "epoch": 0.1580224556320174, "grad_norm": 1.6112561712672129, "learning_rate": 1.915292777020902e-05, "loss": 0.8409, "step": 4363 }, { "epoch": 0.15805867439333574, "grad_norm": 1.6349626078361863, "learning_rate": 1.9152455208237196e-05, "loss": 0.865, "step": 4364 }, { "epoch": 0.1580948931546541, "grad_norm": 1.34388516432095, "learning_rate": 1.9151982520319542e-05, "loss": 0.7899, "step": 4365 }, { "epoch": 0.15813111191597248, "grad_norm": 1.7972250017045233, "learning_rate": 1.915150970646257e-05, "loss": 0.8201, "step": 4366 }, { "epoch": 0.15816733067729083, "grad_norm": 1.6199678521200642, "learning_rate": 1.9151036766672782e-05, "loss": 0.8266, "step": 4367 }, { "epoch": 0.1582035494386092, "grad_norm": 1.8657960615216285, "learning_rate": 1.915056370095669e-05, "loss": 0.8196, "step": 4368 }, { "epoch": 0.15823976819992755, "grad_norm": 1.2363002605032636, "learning_rate": 1.9150090509320797e-05, "loss": 0.7904, "step": 4369 }, { "epoch": 0.15827598696124592, "grad_norm": 1.481549197120944, "learning_rate": 1.914961719177162e-05, "loss": 0.8308, "step": 4370 }, { "epoch": 0.1583122057225643, "grad_norm": 1.8239911854383533, "learning_rate": 1.914914374831567e-05, "loss": 0.8565, "step": 4371 }, { "epoch": 0.15834842448388264, "grad_norm": 1.560624727018819, "learning_rate": 1.9148670178959464e-05, "loss": 0.8488, "step": 4372 }, { "epoch": 0.15838464324520102, "grad_norm": 1.4701326658301466, "learning_rate": 1.9148196483709518e-05, "loss": 0.841, "step": 4373 }, { "epoch": 0.15842086200651936, "grad_norm": 1.2476643454041836, "learning_rate": 1.9147722662572353e-05, "loss": 0.7645, "step": 4374 }, { "epoch": 0.15845708076783774, "grad_norm": 1.574306777830953, "learning_rate": 1.914724871555448e-05, "loss": 0.8595, "step": 4375 }, { "epoch": 0.1584932995291561, "grad_norm": 1.4955715495667965, "learning_rate": 1.914677464266243e-05, "loss": 0.8362, "step": 4376 }, { "epoch": 0.15852951829047446, "grad_norm": 1.5585350226330483, "learning_rate": 1.9146300443902724e-05, "loss": 0.7669, "step": 4377 }, { "epoch": 0.15856573705179283, "grad_norm": 1.3033374216486489, "learning_rate": 1.914582611928189e-05, "loss": 0.7757, "step": 4378 }, { "epoch": 0.15860195581311118, "grad_norm": 1.6902355631362833, "learning_rate": 1.914535166880645e-05, "loss": 0.8158, "step": 4379 }, { "epoch": 0.15863817457442955, "grad_norm": 1.5974543643279353, "learning_rate": 1.9144877092482934e-05, "loss": 0.8795, "step": 4380 }, { "epoch": 0.15867439333574793, "grad_norm": 1.623029906738492, "learning_rate": 1.9144402390317872e-05, "loss": 0.8929, "step": 4381 }, { "epoch": 0.15871061209706627, "grad_norm": 1.6376597571792662, "learning_rate": 1.9143927562317803e-05, "loss": 0.8319, "step": 4382 }, { "epoch": 0.15874683085838465, "grad_norm": 1.5700571921812574, "learning_rate": 1.914345260848925e-05, "loss": 0.76, "step": 4383 }, { "epoch": 0.158783049619703, "grad_norm": 1.7435929719061902, "learning_rate": 1.9142977528838763e-05, "loss": 0.7666, "step": 4384 }, { "epoch": 0.15881926838102137, "grad_norm": 1.224290686911764, "learning_rate": 1.9142502323372867e-05, "loss": 0.7813, "step": 4385 }, { "epoch": 0.15885548714233974, "grad_norm": 1.3371806920719869, "learning_rate": 1.9142026992098105e-05, "loss": 0.7972, "step": 4386 }, { "epoch": 0.1588917059036581, "grad_norm": 1.6297942926632036, "learning_rate": 1.914155153502102e-05, "loss": 0.924, "step": 4387 }, { "epoch": 0.15892792466497646, "grad_norm": 1.6105602064466829, "learning_rate": 1.9141075952148154e-05, "loss": 0.9033, "step": 4388 }, { "epoch": 0.1589641434262948, "grad_norm": 1.6602357071206153, "learning_rate": 1.9140600243486052e-05, "loss": 0.8342, "step": 4389 }, { "epoch": 0.15900036218761318, "grad_norm": 1.601874869401869, "learning_rate": 1.9140124409041255e-05, "loss": 0.8078, "step": 4390 }, { "epoch": 0.15903658094893156, "grad_norm": 1.8252348520518085, "learning_rate": 1.913964844882032e-05, "loss": 0.8347, "step": 4391 }, { "epoch": 0.1590727997102499, "grad_norm": 1.4916037321355888, "learning_rate": 1.9139172362829786e-05, "loss": 0.8534, "step": 4392 }, { "epoch": 0.15910901847156828, "grad_norm": 1.592030906440405, "learning_rate": 1.9138696151076213e-05, "loss": 0.8127, "step": 4393 }, { "epoch": 0.15914523723288662, "grad_norm": 1.687090307411928, "learning_rate": 1.913821981356615e-05, "loss": 0.8522, "step": 4394 }, { "epoch": 0.159181455994205, "grad_norm": 1.8336087066566995, "learning_rate": 1.9137743350306155e-05, "loss": 0.8227, "step": 4395 }, { "epoch": 0.15921767475552337, "grad_norm": 1.6630079617367064, "learning_rate": 1.913726676130278e-05, "loss": 0.8081, "step": 4396 }, { "epoch": 0.15925389351684172, "grad_norm": 1.6734522661257534, "learning_rate": 1.913679004656259e-05, "loss": 0.7778, "step": 4397 }, { "epoch": 0.1592901122781601, "grad_norm": 1.4903825196291076, "learning_rate": 1.9136313206092133e-05, "loss": 0.8237, "step": 4398 }, { "epoch": 0.15932633103947844, "grad_norm": 1.639807896133519, "learning_rate": 1.913583623989798e-05, "loss": 0.8995, "step": 4399 }, { "epoch": 0.1593625498007968, "grad_norm": 1.5025719854609945, "learning_rate": 1.91353591479867e-05, "loss": 0.8215, "step": 4400 }, { "epoch": 0.15939876856211518, "grad_norm": 1.5398157069054044, "learning_rate": 1.9134881930364845e-05, "loss": 0.6942, "step": 4401 }, { "epoch": 0.15943498732343353, "grad_norm": 1.5606850567680897, "learning_rate": 1.913440458703899e-05, "loss": 0.8448, "step": 4402 }, { "epoch": 0.1594712060847519, "grad_norm": 1.7359743063524415, "learning_rate": 1.91339271180157e-05, "loss": 0.9388, "step": 4403 }, { "epoch": 0.15950742484607028, "grad_norm": 1.5707092685384834, "learning_rate": 1.913344952330155e-05, "loss": 0.7343, "step": 4404 }, { "epoch": 0.15954364360738862, "grad_norm": 1.4117013353615246, "learning_rate": 1.9132971802903105e-05, "loss": 0.7429, "step": 4405 }, { "epoch": 0.159579862368707, "grad_norm": 1.5106047251524073, "learning_rate": 1.9132493956826948e-05, "loss": 0.7824, "step": 4406 }, { "epoch": 0.15961608113002534, "grad_norm": 1.962512417316636, "learning_rate": 1.9132015985079646e-05, "loss": 0.8289, "step": 4407 }, { "epoch": 0.15965229989134372, "grad_norm": 1.605328822941265, "learning_rate": 1.913153788766778e-05, "loss": 0.8098, "step": 4408 }, { "epoch": 0.1596885186526621, "grad_norm": 1.5626244030543095, "learning_rate": 1.9131059664597932e-05, "loss": 0.8515, "step": 4409 }, { "epoch": 0.15972473741398044, "grad_norm": 1.3094181498475428, "learning_rate": 1.9130581315876676e-05, "loss": 0.7667, "step": 4410 }, { "epoch": 0.1597609561752988, "grad_norm": 1.5993019770464982, "learning_rate": 1.91301028415106e-05, "loss": 0.8271, "step": 4411 }, { "epoch": 0.15979717493661716, "grad_norm": 1.2403198213222577, "learning_rate": 1.9129624241506288e-05, "loss": 0.7607, "step": 4412 }, { "epoch": 0.15983339369793553, "grad_norm": 1.1596346015145254, "learning_rate": 1.9129145515870325e-05, "loss": 0.7902, "step": 4413 }, { "epoch": 0.1598696124592539, "grad_norm": 1.692643321084841, "learning_rate": 1.9128666664609293e-05, "loss": 0.8634, "step": 4414 }, { "epoch": 0.15990583122057225, "grad_norm": 1.6515828493176505, "learning_rate": 1.9128187687729793e-05, "loss": 0.8255, "step": 4415 }, { "epoch": 0.15994204998189063, "grad_norm": 1.6818581705126725, "learning_rate": 1.9127708585238406e-05, "loss": 0.8713, "step": 4416 }, { "epoch": 0.15997826874320897, "grad_norm": 1.603856235250188, "learning_rate": 1.9127229357141726e-05, "loss": 0.8481, "step": 4417 }, { "epoch": 0.16001448750452735, "grad_norm": 1.6505215660608996, "learning_rate": 1.9126750003446355e-05, "loss": 0.8595, "step": 4418 }, { "epoch": 0.16005070626584572, "grad_norm": 1.3152840061679532, "learning_rate": 1.9126270524158883e-05, "loss": 0.7928, "step": 4419 }, { "epoch": 0.16008692502716407, "grad_norm": 1.553485543513968, "learning_rate": 1.9125790919285906e-05, "loss": 0.9238, "step": 4420 }, { "epoch": 0.16012314378848244, "grad_norm": 1.4965288361247904, "learning_rate": 1.9125311188834034e-05, "loss": 0.8178, "step": 4421 }, { "epoch": 0.1601593625498008, "grad_norm": 1.5207618453395646, "learning_rate": 1.912483133280986e-05, "loss": 0.8528, "step": 4422 }, { "epoch": 0.16019558131111916, "grad_norm": 1.7445425828428094, "learning_rate": 1.9124351351219984e-05, "loss": 0.8924, "step": 4423 }, { "epoch": 0.16023180007243754, "grad_norm": 1.6496516576628364, "learning_rate": 1.9123871244071018e-05, "loss": 0.8954, "step": 4424 }, { "epoch": 0.16026801883375588, "grad_norm": 1.6991655626413458, "learning_rate": 1.912339101136957e-05, "loss": 0.8795, "step": 4425 }, { "epoch": 0.16030423759507426, "grad_norm": 1.631501596598342, "learning_rate": 1.9122910653122244e-05, "loss": 0.7876, "step": 4426 }, { "epoch": 0.1603404563563926, "grad_norm": 1.1775273701401596, "learning_rate": 1.912243016933565e-05, "loss": 0.8536, "step": 4427 }, { "epoch": 0.16037667511771098, "grad_norm": 1.7091255696416479, "learning_rate": 1.91219495600164e-05, "loss": 0.7876, "step": 4428 }, { "epoch": 0.16041289387902935, "grad_norm": 1.4770638734852035, "learning_rate": 1.9121468825171114e-05, "loss": 0.7924, "step": 4429 }, { "epoch": 0.1604491126403477, "grad_norm": 1.804871930569556, "learning_rate": 1.9120987964806397e-05, "loss": 0.8432, "step": 4430 }, { "epoch": 0.16048533140166607, "grad_norm": 1.529645442518422, "learning_rate": 1.9120506978928873e-05, "loss": 0.7831, "step": 4431 }, { "epoch": 0.16052155016298442, "grad_norm": 1.4956038472225655, "learning_rate": 1.912002586754516e-05, "loss": 0.8739, "step": 4432 }, { "epoch": 0.1605577689243028, "grad_norm": 1.4527528642655887, "learning_rate": 1.9119544630661877e-05, "loss": 0.8177, "step": 4433 }, { "epoch": 0.16059398768562116, "grad_norm": 1.562364581915635, "learning_rate": 1.9119063268285645e-05, "loss": 0.8906, "step": 4434 }, { "epoch": 0.1606302064469395, "grad_norm": 1.2270260499829222, "learning_rate": 1.911858178042309e-05, "loss": 0.8059, "step": 4435 }, { "epoch": 0.16066642520825788, "grad_norm": 1.5489217229609233, "learning_rate": 1.911810016708084e-05, "loss": 0.8853, "step": 4436 }, { "epoch": 0.16070264396957623, "grad_norm": 1.5743274781326595, "learning_rate": 1.911761842826552e-05, "loss": 0.822, "step": 4437 }, { "epoch": 0.1607388627308946, "grad_norm": 1.5694451386083652, "learning_rate": 1.9117136563983754e-05, "loss": 0.9278, "step": 4438 }, { "epoch": 0.16077508149221298, "grad_norm": 1.6981431088712269, "learning_rate": 1.9116654574242185e-05, "loss": 0.8538, "step": 4439 }, { "epoch": 0.16081130025353133, "grad_norm": 1.294911004523569, "learning_rate": 1.9116172459047434e-05, "loss": 0.7874, "step": 4440 }, { "epoch": 0.1608475190148497, "grad_norm": 1.545554508982316, "learning_rate": 1.911569021840614e-05, "loss": 0.7992, "step": 4441 }, { "epoch": 0.16088373777616805, "grad_norm": 1.7282354655838974, "learning_rate": 1.911520785232494e-05, "loss": 0.8475, "step": 4442 }, { "epoch": 0.16091995653748642, "grad_norm": 1.4971101883705533, "learning_rate": 1.911472536081047e-05, "loss": 0.8713, "step": 4443 }, { "epoch": 0.1609561752988048, "grad_norm": 1.4954904492742687, "learning_rate": 1.9114242743869368e-05, "loss": 0.821, "step": 4444 }, { "epoch": 0.16099239406012314, "grad_norm": 1.586967872294202, "learning_rate": 1.911376000150828e-05, "loss": 0.8372, "step": 4445 }, { "epoch": 0.1610286128214415, "grad_norm": 3.308734215250903, "learning_rate": 1.9113277133733847e-05, "loss": 0.8649, "step": 4446 }, { "epoch": 0.16106483158275986, "grad_norm": 1.5811363090760902, "learning_rate": 1.911279414055271e-05, "loss": 0.8761, "step": 4447 }, { "epoch": 0.16110105034407823, "grad_norm": 1.3662584492872807, "learning_rate": 1.911231102197152e-05, "loss": 0.7824, "step": 4448 }, { "epoch": 0.1611372691053966, "grad_norm": 1.50470598928435, "learning_rate": 1.9111827777996926e-05, "loss": 0.8739, "step": 4449 }, { "epoch": 0.16117348786671495, "grad_norm": 1.4767357671590868, "learning_rate": 1.9111344408635573e-05, "loss": 0.8913, "step": 4450 }, { "epoch": 0.16120970662803333, "grad_norm": 1.2503387558393924, "learning_rate": 1.9110860913894114e-05, "loss": 0.8067, "step": 4451 }, { "epoch": 0.16124592538935167, "grad_norm": 1.4873160607898102, "learning_rate": 1.9110377293779205e-05, "loss": 0.8328, "step": 4452 }, { "epoch": 0.16128214415067005, "grad_norm": 1.5772782541000223, "learning_rate": 1.91098935482975e-05, "loss": 0.7852, "step": 4453 }, { "epoch": 0.16131836291198842, "grad_norm": 1.4739757433330374, "learning_rate": 1.9109409677455652e-05, "loss": 0.8296, "step": 4454 }, { "epoch": 0.16135458167330677, "grad_norm": 1.5316402116924295, "learning_rate": 1.9108925681260327e-05, "loss": 0.9122, "step": 4455 }, { "epoch": 0.16139080043462514, "grad_norm": 1.69011593865016, "learning_rate": 1.910844155971818e-05, "loss": 0.8701, "step": 4456 }, { "epoch": 0.1614270191959435, "grad_norm": 1.3291330501966356, "learning_rate": 1.910795731283587e-05, "loss": 0.8035, "step": 4457 }, { "epoch": 0.16146323795726186, "grad_norm": 1.3913198714670416, "learning_rate": 1.9107472940620065e-05, "loss": 0.8582, "step": 4458 }, { "epoch": 0.16149945671858024, "grad_norm": 1.3536796449801631, "learning_rate": 1.910698844307743e-05, "loss": 0.8321, "step": 4459 }, { "epoch": 0.16153567547989858, "grad_norm": 1.5189164641481976, "learning_rate": 1.9106503820214636e-05, "loss": 0.7756, "step": 4460 }, { "epoch": 0.16157189424121696, "grad_norm": 1.5786371037543803, "learning_rate": 1.9106019072038345e-05, "loss": 0.8763, "step": 4461 }, { "epoch": 0.1616081130025353, "grad_norm": 1.923614748628756, "learning_rate": 1.910553419855523e-05, "loss": 0.7507, "step": 4462 }, { "epoch": 0.16164433176385368, "grad_norm": 1.412180795881562, "learning_rate": 1.9105049199771963e-05, "loss": 0.8379, "step": 4463 }, { "epoch": 0.16168055052517205, "grad_norm": 1.2374017913699111, "learning_rate": 1.910456407569522e-05, "loss": 0.7849, "step": 4464 }, { "epoch": 0.1617167692864904, "grad_norm": 1.7440416365638198, "learning_rate": 1.9104078826331674e-05, "loss": 0.9604, "step": 4465 }, { "epoch": 0.16175298804780877, "grad_norm": 2.010768673441204, "learning_rate": 1.9103593451688005e-05, "loss": 0.8077, "step": 4466 }, { "epoch": 0.16178920680912712, "grad_norm": 1.5756285332421103, "learning_rate": 1.910310795177089e-05, "loss": 0.8551, "step": 4467 }, { "epoch": 0.1618254255704455, "grad_norm": 1.1992734693102414, "learning_rate": 1.9102622326587013e-05, "loss": 0.7351, "step": 4468 }, { "epoch": 0.16186164433176387, "grad_norm": 1.6860311979307565, "learning_rate": 1.9102136576143052e-05, "loss": 0.8666, "step": 4469 }, { "epoch": 0.1618978630930822, "grad_norm": 1.5177694608778372, "learning_rate": 1.9101650700445697e-05, "loss": 0.8458, "step": 4470 }, { "epoch": 0.16193408185440059, "grad_norm": 1.744706888802146, "learning_rate": 1.9101164699501628e-05, "loss": 0.8559, "step": 4471 }, { "epoch": 0.16197030061571893, "grad_norm": 1.1641278789089016, "learning_rate": 1.9100678573317538e-05, "loss": 0.8133, "step": 4472 }, { "epoch": 0.1620065193770373, "grad_norm": 1.5334313338213859, "learning_rate": 1.910019232190011e-05, "loss": 0.8344, "step": 4473 }, { "epoch": 0.16204273813835568, "grad_norm": 1.4709315204964823, "learning_rate": 1.9099705945256043e-05, "loss": 0.8623, "step": 4474 }, { "epoch": 0.16207895689967403, "grad_norm": 1.3306637488638742, "learning_rate": 1.9099219443392028e-05, "loss": 0.7842, "step": 4475 }, { "epoch": 0.1621151756609924, "grad_norm": 1.241854821656767, "learning_rate": 1.9098732816314757e-05, "loss": 0.8717, "step": 4476 }, { "epoch": 0.16215139442231075, "grad_norm": 1.6844827645813647, "learning_rate": 1.9098246064030926e-05, "loss": 0.8594, "step": 4477 }, { "epoch": 0.16218761318362912, "grad_norm": 1.4647400412404894, "learning_rate": 1.9097759186547236e-05, "loss": 0.8062, "step": 4478 }, { "epoch": 0.1622238319449475, "grad_norm": 1.5618618937713733, "learning_rate": 1.9097272183870386e-05, "loss": 0.9237, "step": 4479 }, { "epoch": 0.16226005070626584, "grad_norm": 1.6156012638495578, "learning_rate": 1.9096785056007074e-05, "loss": 0.8775, "step": 4480 }, { "epoch": 0.16229626946758421, "grad_norm": 1.429348316502341, "learning_rate": 1.909629780296401e-05, "loss": 0.9692, "step": 4481 }, { "epoch": 0.16233248822890256, "grad_norm": 1.5924812285078234, "learning_rate": 1.9095810424747895e-05, "loss": 0.7795, "step": 4482 }, { "epoch": 0.16236870699022093, "grad_norm": 1.1190273369458754, "learning_rate": 1.9095322921365435e-05, "loss": 0.7536, "step": 4483 }, { "epoch": 0.1624049257515393, "grad_norm": 1.1673985085641603, "learning_rate": 1.909483529282334e-05, "loss": 0.7124, "step": 4484 }, { "epoch": 0.16244114451285765, "grad_norm": 1.26464022063933, "learning_rate": 1.909434753912832e-05, "loss": 0.7903, "step": 4485 }, { "epoch": 0.16247736327417603, "grad_norm": 1.6802621695612805, "learning_rate": 1.9093859660287087e-05, "loss": 0.9078, "step": 4486 }, { "epoch": 0.16251358203549437, "grad_norm": 1.5929382058901378, "learning_rate": 1.9093371656306355e-05, "loss": 0.7984, "step": 4487 }, { "epoch": 0.16254980079681275, "grad_norm": 1.5000090623621574, "learning_rate": 1.9092883527192837e-05, "loss": 0.7794, "step": 4488 }, { "epoch": 0.16258601955813112, "grad_norm": 1.690172937333203, "learning_rate": 1.9092395272953257e-05, "loss": 0.8353, "step": 4489 }, { "epoch": 0.16262223831944947, "grad_norm": 1.66636031298088, "learning_rate": 1.9091906893594324e-05, "loss": 0.8582, "step": 4490 }, { "epoch": 0.16265845708076784, "grad_norm": 1.7014484717165927, "learning_rate": 1.9091418389122766e-05, "loss": 0.8121, "step": 4491 }, { "epoch": 0.1626946758420862, "grad_norm": 1.5777243610809468, "learning_rate": 1.90909297595453e-05, "loss": 0.8548, "step": 4492 }, { "epoch": 0.16273089460340456, "grad_norm": 1.5738649420314965, "learning_rate": 1.9090441004868655e-05, "loss": 0.7862, "step": 4493 }, { "epoch": 0.16276711336472294, "grad_norm": 1.4997109436383553, "learning_rate": 1.908995212509955e-05, "loss": 0.8183, "step": 4494 }, { "epoch": 0.16280333212604128, "grad_norm": 1.4282759429902738, "learning_rate": 1.908946312024472e-05, "loss": 0.8103, "step": 4495 }, { "epoch": 0.16283955088735966, "grad_norm": 1.5455903091405474, "learning_rate": 1.9088973990310894e-05, "loss": 0.8616, "step": 4496 }, { "epoch": 0.162875769648678, "grad_norm": 1.3294898996183555, "learning_rate": 1.9088484735304793e-05, "loss": 0.8441, "step": 4497 }, { "epoch": 0.16291198840999638, "grad_norm": 1.5548895466168366, "learning_rate": 1.9087995355233157e-05, "loss": 0.8214, "step": 4498 }, { "epoch": 0.16294820717131475, "grad_norm": 1.3615003483544699, "learning_rate": 1.9087505850102724e-05, "loss": 0.7685, "step": 4499 }, { "epoch": 0.1629844259326331, "grad_norm": 2.4115292735750202, "learning_rate": 1.9087016219920222e-05, "loss": 0.8198, "step": 4500 }, { "epoch": 0.16302064469395147, "grad_norm": 1.3597258671309984, "learning_rate": 1.9086526464692394e-05, "loss": 0.8172, "step": 4501 }, { "epoch": 0.16305686345526982, "grad_norm": 1.1963271201171919, "learning_rate": 1.908603658442598e-05, "loss": 0.8042, "step": 4502 }, { "epoch": 0.1630930822165882, "grad_norm": 1.6444184543228284, "learning_rate": 1.9085546579127713e-05, "loss": 0.826, "step": 4503 }, { "epoch": 0.16312930097790657, "grad_norm": 1.484146025841052, "learning_rate": 1.9085056448804345e-05, "loss": 0.8485, "step": 4504 }, { "epoch": 0.1631655197392249, "grad_norm": 1.5113623153287759, "learning_rate": 1.9084566193462617e-05, "loss": 0.8397, "step": 4505 }, { "epoch": 0.1632017385005433, "grad_norm": 1.376924295157422, "learning_rate": 1.9084075813109275e-05, "loss": 0.7769, "step": 4506 }, { "epoch": 0.16323795726186163, "grad_norm": 1.3804241274134452, "learning_rate": 1.9083585307751068e-05, "loss": 0.7465, "step": 4507 }, { "epoch": 0.16327417602318, "grad_norm": 1.2013400168510036, "learning_rate": 1.908309467739475e-05, "loss": 0.7577, "step": 4508 }, { "epoch": 0.16331039478449838, "grad_norm": 1.6643435622545197, "learning_rate": 1.9082603922047063e-05, "loss": 0.7663, "step": 4509 }, { "epoch": 0.16334661354581673, "grad_norm": 1.6081199567043365, "learning_rate": 1.9082113041714767e-05, "loss": 0.791, "step": 4510 }, { "epoch": 0.1633828323071351, "grad_norm": 1.660378918020236, "learning_rate": 1.9081622036404616e-05, "loss": 0.8334, "step": 4511 }, { "epoch": 0.16341905106845345, "grad_norm": 1.631129580516305, "learning_rate": 1.9081130906123366e-05, "loss": 0.7911, "step": 4512 }, { "epoch": 0.16345526982977182, "grad_norm": 1.5914365214276402, "learning_rate": 1.9080639650877774e-05, "loss": 0.8886, "step": 4513 }, { "epoch": 0.1634914885910902, "grad_norm": 1.739641812240064, "learning_rate": 1.90801482706746e-05, "loss": 0.8091, "step": 4514 }, { "epoch": 0.16352770735240854, "grad_norm": 1.8042012732136634, "learning_rate": 1.907965676552061e-05, "loss": 0.8097, "step": 4515 }, { "epoch": 0.16356392611372692, "grad_norm": 2.1939179544840974, "learning_rate": 1.9079165135422567e-05, "loss": 0.8244, "step": 4516 }, { "epoch": 0.16360014487504526, "grad_norm": 1.7771540592958681, "learning_rate": 1.907867338038723e-05, "loss": 0.8382, "step": 4517 }, { "epoch": 0.16363636363636364, "grad_norm": 1.4615012741090654, "learning_rate": 1.907818150042137e-05, "loss": 0.7419, "step": 4518 }, { "epoch": 0.163672582397682, "grad_norm": 1.6284900346219902, "learning_rate": 1.9077689495531755e-05, "loss": 0.8224, "step": 4519 }, { "epoch": 0.16370880115900036, "grad_norm": 1.458061909678054, "learning_rate": 1.9077197365725163e-05, "loss": 0.7406, "step": 4520 }, { "epoch": 0.16374501992031873, "grad_norm": 1.7835289468409343, "learning_rate": 1.907670511100835e-05, "loss": 0.9316, "step": 4521 }, { "epoch": 0.16378123868163708, "grad_norm": 1.335493818406584, "learning_rate": 1.9076212731388106e-05, "loss": 0.8279, "step": 4522 }, { "epoch": 0.16381745744295545, "grad_norm": 1.6625406885680278, "learning_rate": 1.90757202268712e-05, "loss": 0.7582, "step": 4523 }, { "epoch": 0.16385367620427382, "grad_norm": 1.94804656688687, "learning_rate": 1.9075227597464404e-05, "loss": 0.8915, "step": 4524 }, { "epoch": 0.16388989496559217, "grad_norm": 1.576169805749908, "learning_rate": 1.9074734843174504e-05, "loss": 0.9072, "step": 4525 }, { "epoch": 0.16392611372691054, "grad_norm": 1.4879459831994066, "learning_rate": 1.907424196400828e-05, "loss": 0.808, "step": 4526 }, { "epoch": 0.1639623324882289, "grad_norm": 1.3721869286770763, "learning_rate": 1.9073748959972514e-05, "loss": 0.874, "step": 4527 }, { "epoch": 0.16399855124954726, "grad_norm": 1.5773678311082198, "learning_rate": 1.9073255831073985e-05, "loss": 0.8834, "step": 4528 }, { "epoch": 0.16403477001086564, "grad_norm": 1.5789400783672993, "learning_rate": 1.9072762577319487e-05, "loss": 0.8167, "step": 4529 }, { "epoch": 0.16407098877218398, "grad_norm": 1.5093284445250623, "learning_rate": 1.9072269198715805e-05, "loss": 0.8034, "step": 4530 }, { "epoch": 0.16410720753350236, "grad_norm": 1.5557690541290394, "learning_rate": 1.9071775695269726e-05, "loss": 0.7965, "step": 4531 }, { "epoch": 0.1641434262948207, "grad_norm": 1.6001991924237455, "learning_rate": 1.907128206698804e-05, "loss": 0.8228, "step": 4532 }, { "epoch": 0.16417964505613908, "grad_norm": 1.7439201275406886, "learning_rate": 1.9070788313877546e-05, "loss": 0.8695, "step": 4533 }, { "epoch": 0.16421586381745745, "grad_norm": 1.545638413770186, "learning_rate": 1.9070294435945032e-05, "loss": 0.8636, "step": 4534 }, { "epoch": 0.1642520825787758, "grad_norm": 1.5611618534681453, "learning_rate": 1.9069800433197296e-05, "loss": 0.8604, "step": 4535 }, { "epoch": 0.16428830134009417, "grad_norm": 1.8142443609709202, "learning_rate": 1.906930630564114e-05, "loss": 0.7902, "step": 4536 }, { "epoch": 0.16432452010141252, "grad_norm": 1.5600085977294853, "learning_rate": 1.9068812053283355e-05, "loss": 0.8409, "step": 4537 }, { "epoch": 0.1643607388627309, "grad_norm": 1.6870704187981713, "learning_rate": 1.9068317676130753e-05, "loss": 0.8532, "step": 4538 }, { "epoch": 0.16439695762404927, "grad_norm": 1.511547595815779, "learning_rate": 1.9067823174190127e-05, "loss": 0.8358, "step": 4539 }, { "epoch": 0.1644331763853676, "grad_norm": 1.462362250031598, "learning_rate": 1.906732854746829e-05, "loss": 0.8638, "step": 4540 }, { "epoch": 0.164469395146686, "grad_norm": 1.474464660768785, "learning_rate": 1.906683379597204e-05, "loss": 0.88, "step": 4541 }, { "epoch": 0.16450561390800433, "grad_norm": 1.5298658660774107, "learning_rate": 1.90663389197082e-05, "loss": 0.8295, "step": 4542 }, { "epoch": 0.1645418326693227, "grad_norm": 1.7809125210395362, "learning_rate": 1.9065843918683563e-05, "loss": 0.8539, "step": 4543 }, { "epoch": 0.16457805143064108, "grad_norm": 1.3923687889555685, "learning_rate": 1.9065348792904946e-05, "loss": 0.7891, "step": 4544 }, { "epoch": 0.16461427019195943, "grad_norm": 1.6564321327832114, "learning_rate": 1.9064853542379166e-05, "loss": 0.8405, "step": 4545 }, { "epoch": 0.1646504889532778, "grad_norm": 1.5122692380507166, "learning_rate": 1.9064358167113036e-05, "loss": 0.7829, "step": 4546 }, { "epoch": 0.16468670771459615, "grad_norm": 1.6771779764216699, "learning_rate": 1.9063862667113376e-05, "loss": 0.8743, "step": 4547 }, { "epoch": 0.16472292647591452, "grad_norm": 1.608408288195936, "learning_rate": 1.9063367042387e-05, "loss": 0.7913, "step": 4548 }, { "epoch": 0.1647591452372329, "grad_norm": 1.6283263934235688, "learning_rate": 1.9062871292940728e-05, "loss": 0.8867, "step": 4549 }, { "epoch": 0.16479536399855124, "grad_norm": 1.632691798129355, "learning_rate": 1.9062375418781384e-05, "loss": 0.8306, "step": 4550 }, { "epoch": 0.16483158275986962, "grad_norm": 1.133396296274811, "learning_rate": 1.906187941991579e-05, "loss": 0.7607, "step": 4551 }, { "epoch": 0.16486780152118796, "grad_norm": 1.0875583296135416, "learning_rate": 1.9061383296350777e-05, "loss": 0.7151, "step": 4552 }, { "epoch": 0.16490402028250634, "grad_norm": 1.8584477111891242, "learning_rate": 1.9060887048093166e-05, "loss": 0.9171, "step": 4553 }, { "epoch": 0.1649402390438247, "grad_norm": 1.2871619201702864, "learning_rate": 1.9060390675149787e-05, "loss": 0.7441, "step": 4554 }, { "epoch": 0.16497645780514306, "grad_norm": 1.4630461796003027, "learning_rate": 1.905989417752747e-05, "loss": 0.864, "step": 4555 }, { "epoch": 0.16501267656646143, "grad_norm": 1.6871387019857664, "learning_rate": 1.905939755523305e-05, "loss": 0.8594, "step": 4556 }, { "epoch": 0.16504889532777978, "grad_norm": 1.2967960237748688, "learning_rate": 1.9058900808273358e-05, "loss": 0.8108, "step": 4557 }, { "epoch": 0.16508511408909815, "grad_norm": 1.66861282371488, "learning_rate": 1.9058403936655235e-05, "loss": 0.906, "step": 4558 }, { "epoch": 0.16512133285041652, "grad_norm": 1.3570056249629152, "learning_rate": 1.9057906940385508e-05, "loss": 0.7914, "step": 4559 }, { "epoch": 0.16515755161173487, "grad_norm": 1.565030109432752, "learning_rate": 1.905740981947103e-05, "loss": 0.7393, "step": 4560 }, { "epoch": 0.16519377037305324, "grad_norm": 1.7086119894581269, "learning_rate": 1.905691257391863e-05, "loss": 0.823, "step": 4561 }, { "epoch": 0.1652299891343716, "grad_norm": 1.5055115665470047, "learning_rate": 1.905641520373515e-05, "loss": 0.8483, "step": 4562 }, { "epoch": 0.16526620789568996, "grad_norm": 1.4335668181922168, "learning_rate": 1.905591770892745e-05, "loss": 0.8309, "step": 4563 }, { "epoch": 0.16530242665700834, "grad_norm": 1.499675463546475, "learning_rate": 1.905542008950236e-05, "loss": 0.8256, "step": 4564 }, { "epoch": 0.16533864541832669, "grad_norm": 1.3591457514394985, "learning_rate": 1.905492234546673e-05, "loss": 0.7573, "step": 4565 }, { "epoch": 0.16537486417964506, "grad_norm": 1.6034310262808094, "learning_rate": 1.9054424476827417e-05, "loss": 0.805, "step": 4566 }, { "epoch": 0.1654110829409634, "grad_norm": 1.247035623047777, "learning_rate": 1.905392648359126e-05, "loss": 0.7956, "step": 4567 }, { "epoch": 0.16544730170228178, "grad_norm": 1.4980178939622146, "learning_rate": 1.9053428365765126e-05, "loss": 0.7796, "step": 4568 }, { "epoch": 0.16548352046360015, "grad_norm": 1.2239677095730437, "learning_rate": 1.9052930123355858e-05, "loss": 0.8205, "step": 4569 }, { "epoch": 0.1655197392249185, "grad_norm": 1.2861934553793917, "learning_rate": 1.905243175637032e-05, "loss": 0.7964, "step": 4570 }, { "epoch": 0.16555595798623687, "grad_norm": 1.5579134118311486, "learning_rate": 1.9051933264815368e-05, "loss": 0.8559, "step": 4571 }, { "epoch": 0.16559217674755522, "grad_norm": 1.4523304861713755, "learning_rate": 1.9051434648697857e-05, "loss": 0.773, "step": 4572 }, { "epoch": 0.1656283955088736, "grad_norm": 1.5009918480468936, "learning_rate": 1.9050935908024656e-05, "loss": 0.7514, "step": 4573 }, { "epoch": 0.16566461427019197, "grad_norm": 1.72567883054201, "learning_rate": 1.9050437042802622e-05, "loss": 0.7976, "step": 4574 }, { "epoch": 0.16570083303151031, "grad_norm": 7.351230009791583, "learning_rate": 1.904993805303862e-05, "loss": 0.8598, "step": 4575 }, { "epoch": 0.1657370517928287, "grad_norm": 1.8346922526998921, "learning_rate": 1.9049438938739522e-05, "loss": 0.8294, "step": 4576 }, { "epoch": 0.16577327055414706, "grad_norm": 1.5479839484944464, "learning_rate": 1.904893969991219e-05, "loss": 0.8935, "step": 4577 }, { "epoch": 0.1658094893154654, "grad_norm": 1.3987218902957108, "learning_rate": 1.9048440336563495e-05, "loss": 0.84, "step": 4578 }, { "epoch": 0.16584570807678378, "grad_norm": 1.7127742967066861, "learning_rate": 1.9047940848700315e-05, "loss": 0.9029, "step": 4579 }, { "epoch": 0.16588192683810213, "grad_norm": 1.8825309273786999, "learning_rate": 1.9047441236329516e-05, "loss": 0.865, "step": 4580 }, { "epoch": 0.1659181455994205, "grad_norm": 1.8874759852542695, "learning_rate": 1.904694149945798e-05, "loss": 0.7906, "step": 4581 }, { "epoch": 0.16595436436073888, "grad_norm": 1.50716731246789, "learning_rate": 1.9046441638092572e-05, "loss": 0.7227, "step": 4582 }, { "epoch": 0.16599058312205722, "grad_norm": 1.5205736556402247, "learning_rate": 1.9045941652240183e-05, "loss": 0.8846, "step": 4583 }, { "epoch": 0.1660268018833756, "grad_norm": 1.503336857475081, "learning_rate": 1.9045441541907686e-05, "loss": 0.8882, "step": 4584 }, { "epoch": 0.16606302064469394, "grad_norm": 1.247249396429446, "learning_rate": 1.9044941307101966e-05, "loss": 0.8163, "step": 4585 }, { "epoch": 0.16609923940601232, "grad_norm": 1.5385998839063066, "learning_rate": 1.9044440947829908e-05, "loss": 0.7813, "step": 4586 }, { "epoch": 0.1661354581673307, "grad_norm": 1.7139933118574595, "learning_rate": 1.9043940464098397e-05, "loss": 0.828, "step": 4587 }, { "epoch": 0.16617167692864904, "grad_norm": 1.6391266838431047, "learning_rate": 1.9043439855914313e-05, "loss": 0.8135, "step": 4588 }, { "epoch": 0.1662078956899674, "grad_norm": 1.618478118357685, "learning_rate": 1.9042939123284555e-05, "loss": 0.8401, "step": 4589 }, { "epoch": 0.16624411445128576, "grad_norm": 1.8810899995173438, "learning_rate": 1.9042438266216004e-05, "loss": 0.8718, "step": 4590 }, { "epoch": 0.16628033321260413, "grad_norm": 2.319828169138248, "learning_rate": 1.9041937284715562e-05, "loss": 0.7714, "step": 4591 }, { "epoch": 0.1663165519739225, "grad_norm": 1.6141371707870993, "learning_rate": 1.9041436178790117e-05, "loss": 0.9369, "step": 4592 }, { "epoch": 0.16635277073524085, "grad_norm": 1.656371406837863, "learning_rate": 1.9040934948446564e-05, "loss": 0.853, "step": 4593 }, { "epoch": 0.16638898949655923, "grad_norm": 1.480260315970431, "learning_rate": 1.90404335936918e-05, "loss": 0.8637, "step": 4594 }, { "epoch": 0.16642520825787757, "grad_norm": 1.580052030769909, "learning_rate": 1.903993211453273e-05, "loss": 0.8352, "step": 4595 }, { "epoch": 0.16646142701919595, "grad_norm": 1.6680200482906817, "learning_rate": 1.9039430510976252e-05, "loss": 0.8532, "step": 4596 }, { "epoch": 0.16649764578051432, "grad_norm": 1.6780633365190063, "learning_rate": 1.9038928783029266e-05, "loss": 0.8831, "step": 4597 }, { "epoch": 0.16653386454183267, "grad_norm": 1.455505672148789, "learning_rate": 1.9038426930698676e-05, "loss": 0.8142, "step": 4598 }, { "epoch": 0.16657008330315104, "grad_norm": 1.6194044411308073, "learning_rate": 1.903792495399139e-05, "loss": 0.8194, "step": 4599 }, { "epoch": 0.16660630206446939, "grad_norm": 1.4679398610478596, "learning_rate": 1.9037422852914316e-05, "loss": 0.7835, "step": 4600 }, { "epoch": 0.16664252082578776, "grad_norm": 1.5633427814668452, "learning_rate": 1.9036920627474362e-05, "loss": 0.8673, "step": 4601 }, { "epoch": 0.16667873958710613, "grad_norm": 1.4434477807407902, "learning_rate": 1.903641827767844e-05, "loss": 0.7623, "step": 4602 }, { "epoch": 0.16671495834842448, "grad_norm": 1.6067881256071554, "learning_rate": 1.903591580353346e-05, "loss": 0.8505, "step": 4603 }, { "epoch": 0.16675117710974285, "grad_norm": 1.4820492166728156, "learning_rate": 1.9035413205046344e-05, "loss": 0.7651, "step": 4604 }, { "epoch": 0.1667873958710612, "grad_norm": 1.200242700409478, "learning_rate": 1.9034910482224e-05, "loss": 0.7815, "step": 4605 }, { "epoch": 0.16682361463237957, "grad_norm": 1.580211323757481, "learning_rate": 1.9034407635073348e-05, "loss": 0.856, "step": 4606 }, { "epoch": 0.16685983339369795, "grad_norm": 1.4102545896118655, "learning_rate": 1.903390466360131e-05, "loss": 0.758, "step": 4607 }, { "epoch": 0.1668960521550163, "grad_norm": 1.6186822465576942, "learning_rate": 1.9033401567814807e-05, "loss": 0.8843, "step": 4608 }, { "epoch": 0.16693227091633467, "grad_norm": 1.704320360144747, "learning_rate": 1.903289834772076e-05, "loss": 0.791, "step": 4609 }, { "epoch": 0.16696848967765301, "grad_norm": 1.146530989149315, "learning_rate": 1.9032395003326096e-05, "loss": 0.8173, "step": 4610 }, { "epoch": 0.1670047084389714, "grad_norm": 1.6522417110247678, "learning_rate": 1.9031891534637735e-05, "loss": 0.8059, "step": 4611 }, { "epoch": 0.16704092720028976, "grad_norm": 1.8868718460315048, "learning_rate": 1.9031387941662614e-05, "loss": 0.8254, "step": 4612 }, { "epoch": 0.1670771459616081, "grad_norm": 1.5012742636458432, "learning_rate": 1.9030884224407658e-05, "loss": 0.8434, "step": 4613 }, { "epoch": 0.16711336472292648, "grad_norm": 1.5895230722024825, "learning_rate": 1.9030380382879806e-05, "loss": 0.8766, "step": 4614 }, { "epoch": 0.16714958348424483, "grad_norm": 1.652282163586013, "learning_rate": 1.902987641708598e-05, "loss": 0.8905, "step": 4615 }, { "epoch": 0.1671858022455632, "grad_norm": 1.4960953411163553, "learning_rate": 1.902937232703312e-05, "loss": 0.7776, "step": 4616 }, { "epoch": 0.16722202100688158, "grad_norm": 1.5522463769423474, "learning_rate": 1.902886811272816e-05, "loss": 0.798, "step": 4617 }, { "epoch": 0.16725823976819992, "grad_norm": 1.5992453478812894, "learning_rate": 1.9028363774178045e-05, "loss": 0.8625, "step": 4618 }, { "epoch": 0.1672944585295183, "grad_norm": 1.446430442797705, "learning_rate": 1.902785931138971e-05, "loss": 0.7868, "step": 4619 }, { "epoch": 0.16733067729083664, "grad_norm": 1.7566377253356849, "learning_rate": 1.9027354724370102e-05, "loss": 0.9062, "step": 4620 }, { "epoch": 0.16736689605215502, "grad_norm": 1.4063420407983287, "learning_rate": 1.902685001312616e-05, "loss": 0.795, "step": 4621 }, { "epoch": 0.1674031148134734, "grad_norm": 1.5968636001452483, "learning_rate": 1.9026345177664826e-05, "loss": 0.8081, "step": 4622 }, { "epoch": 0.16743933357479174, "grad_norm": 1.4863449414037544, "learning_rate": 1.9025840217993056e-05, "loss": 0.8084, "step": 4623 }, { "epoch": 0.1674755523361101, "grad_norm": 1.5198717508063264, "learning_rate": 1.9025335134117792e-05, "loss": 0.8272, "step": 4624 }, { "epoch": 0.16751177109742846, "grad_norm": 1.3687253606180392, "learning_rate": 1.9024829926045984e-05, "loss": 0.8037, "step": 4625 }, { "epoch": 0.16754798985874683, "grad_norm": 1.1560904364174913, "learning_rate": 1.9024324593784586e-05, "loss": 0.785, "step": 4626 }, { "epoch": 0.1675842086200652, "grad_norm": 1.6435608333896592, "learning_rate": 1.9023819137340552e-05, "loss": 0.8888, "step": 4627 }, { "epoch": 0.16762042738138355, "grad_norm": 1.7004321972431338, "learning_rate": 1.902331355672084e-05, "loss": 0.9959, "step": 4628 }, { "epoch": 0.16765664614270193, "grad_norm": 1.4773545272873811, "learning_rate": 1.902280785193241e-05, "loss": 0.7357, "step": 4629 }, { "epoch": 0.16769286490402027, "grad_norm": 1.513688862612983, "learning_rate": 1.9022302022982207e-05, "loss": 0.8705, "step": 4630 }, { "epoch": 0.16772908366533865, "grad_norm": 1.4192766720246308, "learning_rate": 1.9021796069877207e-05, "loss": 0.7969, "step": 4631 }, { "epoch": 0.16776530242665702, "grad_norm": 1.5843634921940724, "learning_rate": 1.902128999262436e-05, "loss": 0.925, "step": 4632 }, { "epoch": 0.16780152118797537, "grad_norm": 1.5770716054997678, "learning_rate": 1.9020783791230636e-05, "loss": 0.8644, "step": 4633 }, { "epoch": 0.16783773994929374, "grad_norm": 1.6299558215191625, "learning_rate": 1.9020277465703006e-05, "loss": 0.8132, "step": 4634 }, { "epoch": 0.1678739587106121, "grad_norm": 1.4146468067725964, "learning_rate": 1.901977101604843e-05, "loss": 0.8774, "step": 4635 }, { "epoch": 0.16791017747193046, "grad_norm": 1.4777960758603343, "learning_rate": 1.901926444227388e-05, "loss": 0.8669, "step": 4636 }, { "epoch": 0.16794639623324883, "grad_norm": 1.5857489443251054, "learning_rate": 1.9018757744386324e-05, "loss": 0.8524, "step": 4637 }, { "epoch": 0.16798261499456718, "grad_norm": 1.5345945865662538, "learning_rate": 1.9018250922392737e-05, "loss": 0.8477, "step": 4638 }, { "epoch": 0.16801883375588555, "grad_norm": 1.5516945722211897, "learning_rate": 1.9017743976300092e-05, "loss": 0.7985, "step": 4639 }, { "epoch": 0.1680550525172039, "grad_norm": 1.797352805828646, "learning_rate": 1.901723690611537e-05, "loss": 0.8134, "step": 4640 }, { "epoch": 0.16809127127852228, "grad_norm": 1.634048892723414, "learning_rate": 1.9016729711845543e-05, "loss": 0.8336, "step": 4641 }, { "epoch": 0.16812749003984065, "grad_norm": 1.8247683357585374, "learning_rate": 1.901622239349759e-05, "loss": 0.8473, "step": 4642 }, { "epoch": 0.168163708801159, "grad_norm": 1.2055360927668042, "learning_rate": 1.90157149510785e-05, "loss": 0.7871, "step": 4643 }, { "epoch": 0.16819992756247737, "grad_norm": 1.487627507853544, "learning_rate": 1.901520738459525e-05, "loss": 0.7867, "step": 4644 }, { "epoch": 0.16823614632379572, "grad_norm": 1.2530118384043731, "learning_rate": 1.901469969405482e-05, "loss": 0.8, "step": 4645 }, { "epoch": 0.1682723650851141, "grad_norm": 1.6057337611232507, "learning_rate": 1.9014191879464203e-05, "loss": 0.7998, "step": 4646 }, { "epoch": 0.16830858384643246, "grad_norm": 1.1597534569299994, "learning_rate": 1.9013683940830385e-05, "loss": 0.7329, "step": 4647 }, { "epoch": 0.1683448026077508, "grad_norm": 1.5603782702375961, "learning_rate": 1.9013175878160355e-05, "loss": 0.9341, "step": 4648 }, { "epoch": 0.16838102136906918, "grad_norm": 1.5313947339165441, "learning_rate": 1.901266769146111e-05, "loss": 0.8648, "step": 4649 }, { "epoch": 0.16841724013038753, "grad_norm": 1.667364340406488, "learning_rate": 1.9012159380739635e-05, "loss": 0.9187, "step": 4650 }, { "epoch": 0.1684534588917059, "grad_norm": 1.4257329222248667, "learning_rate": 1.901165094600293e-05, "loss": 0.7662, "step": 4651 }, { "epoch": 0.16848967765302428, "grad_norm": 1.4648801853099103, "learning_rate": 1.9011142387257985e-05, "loss": 0.8285, "step": 4652 }, { "epoch": 0.16852589641434262, "grad_norm": 1.5248080220460254, "learning_rate": 1.9010633704511808e-05, "loss": 0.8363, "step": 4653 }, { "epoch": 0.168562115175661, "grad_norm": 1.7647120343450526, "learning_rate": 1.901012489777139e-05, "loss": 0.8912, "step": 4654 }, { "epoch": 0.16859833393697934, "grad_norm": 1.552509781650966, "learning_rate": 1.9009615967043736e-05, "loss": 0.857, "step": 4655 }, { "epoch": 0.16863455269829772, "grad_norm": 1.571997277241847, "learning_rate": 1.9009106912335854e-05, "loss": 0.883, "step": 4656 }, { "epoch": 0.1686707714596161, "grad_norm": 1.604600202717246, "learning_rate": 1.9008597733654743e-05, "loss": 0.8787, "step": 4657 }, { "epoch": 0.16870699022093444, "grad_norm": 1.6822189373400267, "learning_rate": 1.900808843100741e-05, "loss": 0.9156, "step": 4658 }, { "epoch": 0.1687432089822528, "grad_norm": 1.4543178338593172, "learning_rate": 1.9007579004400868e-05, "loss": 0.7625, "step": 4659 }, { "epoch": 0.16877942774357116, "grad_norm": 1.551332202926573, "learning_rate": 1.900706945384212e-05, "loss": 0.8529, "step": 4660 }, { "epoch": 0.16881564650488953, "grad_norm": 1.531267848446035, "learning_rate": 1.9006559779338186e-05, "loss": 0.8576, "step": 4661 }, { "epoch": 0.1688518652662079, "grad_norm": 1.5747908802035133, "learning_rate": 1.9006049980896074e-05, "loss": 0.8426, "step": 4662 }, { "epoch": 0.16888808402752625, "grad_norm": 1.6095094613042031, "learning_rate": 1.90055400585228e-05, "loss": 0.8856, "step": 4663 }, { "epoch": 0.16892430278884463, "grad_norm": 1.3981991176276833, "learning_rate": 1.9005030012225383e-05, "loss": 0.7898, "step": 4664 }, { "epoch": 0.16896052155016297, "grad_norm": 1.4891057208834646, "learning_rate": 1.900451984201084e-05, "loss": 0.7698, "step": 4665 }, { "epoch": 0.16899674031148135, "grad_norm": 1.392569568154314, "learning_rate": 1.9004009547886193e-05, "loss": 0.8097, "step": 4666 }, { "epoch": 0.16903295907279972, "grad_norm": 1.6183918123311505, "learning_rate": 1.900349912985846e-05, "loss": 0.8651, "step": 4667 }, { "epoch": 0.16906917783411807, "grad_norm": 1.5848433755932299, "learning_rate": 1.900298858793467e-05, "loss": 0.8757, "step": 4668 }, { "epoch": 0.16910539659543644, "grad_norm": 1.39558167058583, "learning_rate": 1.9002477922121846e-05, "loss": 0.824, "step": 4669 }, { "epoch": 0.1691416153567548, "grad_norm": 1.52161867310484, "learning_rate": 1.9001967132427017e-05, "loss": 0.8319, "step": 4670 }, { "epoch": 0.16917783411807316, "grad_norm": 1.6368335237524627, "learning_rate": 1.9001456218857207e-05, "loss": 0.8508, "step": 4671 }, { "epoch": 0.16921405287939154, "grad_norm": 1.5641934641304327, "learning_rate": 1.900094518141945e-05, "loss": 0.8645, "step": 4672 }, { "epoch": 0.16925027164070988, "grad_norm": 1.6508761825922669, "learning_rate": 1.9000434020120785e-05, "loss": 0.9281, "step": 4673 }, { "epoch": 0.16928649040202826, "grad_norm": 1.6155459612009009, "learning_rate": 1.8999922734968236e-05, "loss": 0.8085, "step": 4674 }, { "epoch": 0.1693227091633466, "grad_norm": 1.3824978505133718, "learning_rate": 1.899941132596884e-05, "loss": 0.7165, "step": 4675 }, { "epoch": 0.16935892792466498, "grad_norm": 1.6214818466927727, "learning_rate": 1.8998899793129637e-05, "loss": 0.8731, "step": 4676 }, { "epoch": 0.16939514668598335, "grad_norm": 1.2663977182944135, "learning_rate": 1.8998388136457667e-05, "loss": 0.8268, "step": 4677 }, { "epoch": 0.1694313654473017, "grad_norm": 1.5247621493185828, "learning_rate": 1.899787635595997e-05, "loss": 0.7829, "step": 4678 }, { "epoch": 0.16946758420862007, "grad_norm": 1.3023568080571835, "learning_rate": 1.8997364451643585e-05, "loss": 0.8142, "step": 4679 }, { "epoch": 0.16950380296993842, "grad_norm": 1.4334638794013754, "learning_rate": 1.8996852423515562e-05, "loss": 0.8188, "step": 4680 }, { "epoch": 0.1695400217312568, "grad_norm": 1.5409964879374287, "learning_rate": 1.8996340271582943e-05, "loss": 0.8556, "step": 4681 }, { "epoch": 0.16957624049257516, "grad_norm": 1.5707309459545749, "learning_rate": 1.899582799585278e-05, "loss": 0.8234, "step": 4682 }, { "epoch": 0.1696124592538935, "grad_norm": 1.8024406328110247, "learning_rate": 1.8995315596332115e-05, "loss": 0.8152, "step": 4683 }, { "epoch": 0.16964867801521188, "grad_norm": 1.84806465558348, "learning_rate": 1.8994803073028004e-05, "loss": 0.8585, "step": 4684 }, { "epoch": 0.16968489677653023, "grad_norm": 1.4949731402332593, "learning_rate": 1.89942904259475e-05, "loss": 0.7951, "step": 4685 }, { "epoch": 0.1697211155378486, "grad_norm": 1.45727666112506, "learning_rate": 1.8993777655097658e-05, "loss": 0.8145, "step": 4686 }, { "epoch": 0.16975733429916698, "grad_norm": 1.0552600308861702, "learning_rate": 1.8993264760485535e-05, "loss": 0.7388, "step": 4687 }, { "epoch": 0.16979355306048532, "grad_norm": 1.833570569756962, "learning_rate": 1.8992751742118184e-05, "loss": 0.8497, "step": 4688 }, { "epoch": 0.1698297718218037, "grad_norm": 1.1540697292023148, "learning_rate": 1.8992238600002668e-05, "loss": 0.8252, "step": 4689 }, { "epoch": 0.16986599058312205, "grad_norm": 1.5895938489431964, "learning_rate": 1.8991725334146046e-05, "loss": 0.8323, "step": 4690 }, { "epoch": 0.16990220934444042, "grad_norm": 1.1350078982762275, "learning_rate": 1.899121194455538e-05, "loss": 0.8058, "step": 4691 }, { "epoch": 0.1699384281057588, "grad_norm": 1.6766488734711142, "learning_rate": 1.8990698431237744e-05, "loss": 0.8885, "step": 4692 }, { "epoch": 0.16997464686707714, "grad_norm": 1.1776721366170375, "learning_rate": 1.899018479420019e-05, "loss": 0.7626, "step": 4693 }, { "epoch": 0.1700108656283955, "grad_norm": 1.1724838245904166, "learning_rate": 1.89896710334498e-05, "loss": 0.7649, "step": 4694 }, { "epoch": 0.17004708438971386, "grad_norm": 1.6443219459428686, "learning_rate": 1.8989157148993634e-05, "loss": 0.9114, "step": 4695 }, { "epoch": 0.17008330315103223, "grad_norm": 1.133668977372785, "learning_rate": 1.898864314083877e-05, "loss": 0.7621, "step": 4696 }, { "epoch": 0.1701195219123506, "grad_norm": 1.6237601877159793, "learning_rate": 1.8988129008992276e-05, "loss": 0.8692, "step": 4697 }, { "epoch": 0.17015574067366895, "grad_norm": 1.4720443724312682, "learning_rate": 1.8987614753461234e-05, "loss": 0.8132, "step": 4698 }, { "epoch": 0.17019195943498733, "grad_norm": 1.541045240593875, "learning_rate": 1.898710037425271e-05, "loss": 0.8653, "step": 4699 }, { "epoch": 0.17022817819630567, "grad_norm": 1.4916582592024858, "learning_rate": 1.8986585871373792e-05, "loss": 0.8294, "step": 4700 }, { "epoch": 0.17026439695762405, "grad_norm": 1.665481192095601, "learning_rate": 1.898607124483155e-05, "loss": 0.8315, "step": 4701 }, { "epoch": 0.17030061571894242, "grad_norm": 1.6085483698861076, "learning_rate": 1.8985556494633078e-05, "loss": 0.8425, "step": 4702 }, { "epoch": 0.17033683448026077, "grad_norm": 1.460027398475346, "learning_rate": 1.898504162078545e-05, "loss": 0.7662, "step": 4703 }, { "epoch": 0.17037305324157914, "grad_norm": 1.5717920905254232, "learning_rate": 1.898452662329576e-05, "loss": 0.8465, "step": 4704 }, { "epoch": 0.1704092720028975, "grad_norm": 1.4260132013056956, "learning_rate": 1.8984011502171082e-05, "loss": 0.8099, "step": 4705 }, { "epoch": 0.17044549076421586, "grad_norm": 1.6021907853348978, "learning_rate": 1.8983496257418515e-05, "loss": 0.7486, "step": 4706 }, { "epoch": 0.17048170952553424, "grad_norm": 1.5938738536393808, "learning_rate": 1.8982980889045147e-05, "loss": 0.8738, "step": 4707 }, { "epoch": 0.17051792828685258, "grad_norm": 1.5580348970455566, "learning_rate": 1.8982465397058065e-05, "loss": 0.7934, "step": 4708 }, { "epoch": 0.17055414704817096, "grad_norm": 1.3973084645826646, "learning_rate": 1.8981949781464367e-05, "loss": 0.8735, "step": 4709 }, { "epoch": 0.1705903658094893, "grad_norm": 1.400041859317101, "learning_rate": 1.8981434042271153e-05, "loss": 0.8446, "step": 4710 }, { "epoch": 0.17062658457080768, "grad_norm": 1.4542658772573223, "learning_rate": 1.8980918179485508e-05, "loss": 0.8481, "step": 4711 }, { "epoch": 0.17066280333212605, "grad_norm": 1.4073083391856125, "learning_rate": 1.898040219311454e-05, "loss": 0.7863, "step": 4712 }, { "epoch": 0.1706990220934444, "grad_norm": 1.6667279592465514, "learning_rate": 1.897988608316535e-05, "loss": 0.7864, "step": 4713 }, { "epoch": 0.17073524085476277, "grad_norm": 1.2729625037260588, "learning_rate": 1.8979369849645036e-05, "loss": 0.8196, "step": 4714 }, { "epoch": 0.17077145961608112, "grad_norm": 1.5523998415949245, "learning_rate": 1.89788534925607e-05, "loss": 0.8354, "step": 4715 }, { "epoch": 0.1708076783773995, "grad_norm": 1.4296646676559401, "learning_rate": 1.897833701191945e-05, "loss": 0.8099, "step": 4716 }, { "epoch": 0.17084389713871787, "grad_norm": 1.2668601581280325, "learning_rate": 1.8977820407728397e-05, "loss": 0.8623, "step": 4717 }, { "epoch": 0.1708801159000362, "grad_norm": 1.3361087180970617, "learning_rate": 1.8977303679994645e-05, "loss": 0.8654, "step": 4718 }, { "epoch": 0.17091633466135459, "grad_norm": 1.131382649217619, "learning_rate": 1.8976786828725305e-05, "loss": 0.8236, "step": 4719 }, { "epoch": 0.17095255342267293, "grad_norm": 1.6694735800904568, "learning_rate": 1.8976269853927492e-05, "loss": 0.8172, "step": 4720 }, { "epoch": 0.1709887721839913, "grad_norm": 1.5623217278347759, "learning_rate": 1.897575275560832e-05, "loss": 0.9253, "step": 4721 }, { "epoch": 0.17102499094530968, "grad_norm": 1.52071537715465, "learning_rate": 1.89752355337749e-05, "loss": 0.9422, "step": 4722 }, { "epoch": 0.17106120970662803, "grad_norm": 1.570145598124982, "learning_rate": 1.8974718188434357e-05, "loss": 0.7994, "step": 4723 }, { "epoch": 0.1710974284679464, "grad_norm": 1.5283747879654725, "learning_rate": 1.8974200719593803e-05, "loss": 0.9028, "step": 4724 }, { "epoch": 0.17113364722926475, "grad_norm": 1.9834008209800205, "learning_rate": 1.8973683127260362e-05, "loss": 0.8858, "step": 4725 }, { "epoch": 0.17116986599058312, "grad_norm": 1.444249715739798, "learning_rate": 1.8973165411441156e-05, "loss": 0.9104, "step": 4726 }, { "epoch": 0.1712060847519015, "grad_norm": 1.626895629901859, "learning_rate": 1.8972647572143312e-05, "loss": 0.8861, "step": 4727 }, { "epoch": 0.17124230351321984, "grad_norm": 1.6556653703731885, "learning_rate": 1.897212960937395e-05, "loss": 0.8553, "step": 4728 }, { "epoch": 0.17127852227453821, "grad_norm": 1.6388662198004413, "learning_rate": 1.8971611523140203e-05, "loss": 0.8069, "step": 4729 }, { "epoch": 0.17131474103585656, "grad_norm": 1.4044736147787644, "learning_rate": 1.8971093313449197e-05, "loss": 0.8192, "step": 4730 }, { "epoch": 0.17135095979717493, "grad_norm": 1.4877920736041632, "learning_rate": 1.8970574980308062e-05, "loss": 0.8366, "step": 4731 }, { "epoch": 0.1713871785584933, "grad_norm": 1.5217342232778126, "learning_rate": 1.8970056523723937e-05, "loss": 0.7906, "step": 4732 }, { "epoch": 0.17142339731981165, "grad_norm": 1.6150900275783135, "learning_rate": 1.8969537943703953e-05, "loss": 0.8185, "step": 4733 }, { "epoch": 0.17145961608113003, "grad_norm": 1.3425343489634232, "learning_rate": 1.8969019240255244e-05, "loss": 0.8762, "step": 4734 }, { "epoch": 0.17149583484244837, "grad_norm": 1.3003322186779231, "learning_rate": 1.8968500413384946e-05, "loss": 0.8121, "step": 4735 }, { "epoch": 0.17153205360376675, "grad_norm": 1.2880792489286672, "learning_rate": 1.8967981463100207e-05, "loss": 0.796, "step": 4736 }, { "epoch": 0.17156827236508512, "grad_norm": 1.4452950950167398, "learning_rate": 1.896746238940816e-05, "loss": 0.8552, "step": 4737 }, { "epoch": 0.17160449112640347, "grad_norm": 1.6176495744244503, "learning_rate": 1.896694319231595e-05, "loss": 0.8849, "step": 4738 }, { "epoch": 0.17164070988772184, "grad_norm": 1.657198837713741, "learning_rate": 1.896642387183073e-05, "loss": 0.8819, "step": 4739 }, { "epoch": 0.1716769286490402, "grad_norm": 1.587317471213933, "learning_rate": 1.896590442795963e-05, "loss": 0.8489, "step": 4740 }, { "epoch": 0.17171314741035856, "grad_norm": 1.4404541537259836, "learning_rate": 1.896538486070981e-05, "loss": 0.8308, "step": 4741 }, { "epoch": 0.17174936617167694, "grad_norm": 1.784058452427212, "learning_rate": 1.896486517008842e-05, "loss": 0.8687, "step": 4742 }, { "epoch": 0.17178558493299528, "grad_norm": 1.575064481633851, "learning_rate": 1.8964345356102606e-05, "loss": 0.7677, "step": 4743 }, { "epoch": 0.17182180369431366, "grad_norm": 1.6630633090009124, "learning_rate": 1.8963825418759524e-05, "loss": 0.8256, "step": 4744 }, { "epoch": 0.171858022455632, "grad_norm": 1.239268480229437, "learning_rate": 1.8963305358066325e-05, "loss": 0.7488, "step": 4745 }, { "epoch": 0.17189424121695038, "grad_norm": 1.4719798436011844, "learning_rate": 1.8962785174030166e-05, "loss": 0.7814, "step": 4746 }, { "epoch": 0.17193045997826875, "grad_norm": 1.5138313234087446, "learning_rate": 1.8962264866658214e-05, "loss": 0.8464, "step": 4747 }, { "epoch": 0.1719666787395871, "grad_norm": 1.6325129777252283, "learning_rate": 1.8961744435957618e-05, "loss": 0.8131, "step": 4748 }, { "epoch": 0.17200289750090547, "grad_norm": 1.5079703371914934, "learning_rate": 1.896122388193555e-05, "loss": 0.8236, "step": 4749 }, { "epoch": 0.17203911626222385, "grad_norm": 1.6906003122148938, "learning_rate": 1.896070320459916e-05, "loss": 0.8703, "step": 4750 }, { "epoch": 0.1720753350235422, "grad_norm": 1.4342927047396254, "learning_rate": 1.8960182403955626e-05, "loss": 0.8475, "step": 4751 }, { "epoch": 0.17211155378486057, "grad_norm": 1.5660037774608166, "learning_rate": 1.8959661480012105e-05, "loss": 0.7651, "step": 4752 }, { "epoch": 0.1721477725461789, "grad_norm": 1.6819139894185604, "learning_rate": 1.8959140432775772e-05, "loss": 0.8163, "step": 4753 }, { "epoch": 0.17218399130749729, "grad_norm": 1.5778729082816472, "learning_rate": 1.895861926225379e-05, "loss": 0.8417, "step": 4754 }, { "epoch": 0.17222021006881566, "grad_norm": 1.5584033458970428, "learning_rate": 1.895809796845334e-05, "loss": 0.8493, "step": 4755 }, { "epoch": 0.172256428830134, "grad_norm": 1.4153201756592444, "learning_rate": 1.895757655138159e-05, "loss": 0.8258, "step": 4756 }, { "epoch": 0.17229264759145238, "grad_norm": 1.522353931817716, "learning_rate": 1.8957055011045718e-05, "loss": 0.7757, "step": 4757 }, { "epoch": 0.17232886635277073, "grad_norm": 1.7035099095487247, "learning_rate": 1.8956533347452895e-05, "loss": 0.7967, "step": 4758 }, { "epoch": 0.1723650851140891, "grad_norm": 1.4265701489071139, "learning_rate": 1.8956011560610304e-05, "loss": 0.8346, "step": 4759 }, { "epoch": 0.17240130387540747, "grad_norm": 1.5860194075860485, "learning_rate": 1.8955489650525124e-05, "loss": 0.7705, "step": 4760 }, { "epoch": 0.17243752263672582, "grad_norm": 1.4215377617942644, "learning_rate": 1.895496761720454e-05, "loss": 0.7666, "step": 4761 }, { "epoch": 0.1724737413980442, "grad_norm": 1.442364848831499, "learning_rate": 1.895444546065573e-05, "loss": 0.7536, "step": 4762 }, { "epoch": 0.17250996015936254, "grad_norm": 1.271475822482959, "learning_rate": 1.8953923180885883e-05, "loss": 0.7809, "step": 4763 }, { "epoch": 0.17254617892068091, "grad_norm": 1.4169641415734107, "learning_rate": 1.8953400777902188e-05, "loss": 0.7655, "step": 4764 }, { "epoch": 0.1725823976819993, "grad_norm": 1.6392093886276384, "learning_rate": 1.8952878251711828e-05, "loss": 0.8395, "step": 4765 }, { "epoch": 0.17261861644331764, "grad_norm": 1.5544556805526908, "learning_rate": 1.8952355602322e-05, "loss": 0.8123, "step": 4766 }, { "epoch": 0.172654835204636, "grad_norm": 1.2672270088704174, "learning_rate": 1.8951832829739887e-05, "loss": 0.8715, "step": 4767 }, { "epoch": 0.17269105396595436, "grad_norm": 1.488434543374421, "learning_rate": 1.8951309933972694e-05, "loss": 0.8568, "step": 4768 }, { "epoch": 0.17272727272727273, "grad_norm": 1.4603995598832087, "learning_rate": 1.895078691502761e-05, "loss": 0.8715, "step": 4769 }, { "epoch": 0.1727634914885911, "grad_norm": 1.365003860114286, "learning_rate": 1.895026377291183e-05, "loss": 0.8804, "step": 4770 }, { "epoch": 0.17279971024990945, "grad_norm": 1.4896709950840932, "learning_rate": 1.894974050763256e-05, "loss": 0.753, "step": 4771 }, { "epoch": 0.17283592901122782, "grad_norm": 1.2555064227081452, "learning_rate": 1.8949217119196995e-05, "loss": 0.7625, "step": 4772 }, { "epoch": 0.17287214777254617, "grad_norm": 1.296856234095855, "learning_rate": 1.8948693607612337e-05, "loss": 0.8409, "step": 4773 }, { "epoch": 0.17290836653386454, "grad_norm": 1.7427434141246392, "learning_rate": 1.8948169972885795e-05, "loss": 0.8791, "step": 4774 }, { "epoch": 0.17294458529518292, "grad_norm": 1.252214509551847, "learning_rate": 1.8947646215024573e-05, "loss": 0.7712, "step": 4775 }, { "epoch": 0.17298080405650126, "grad_norm": 1.6255689430792235, "learning_rate": 1.894712233403587e-05, "loss": 0.9453, "step": 4776 }, { "epoch": 0.17301702281781964, "grad_norm": 1.6869428330750862, "learning_rate": 1.894659832992691e-05, "loss": 0.8858, "step": 4777 }, { "epoch": 0.17305324157913798, "grad_norm": 1.1250230525328024, "learning_rate": 1.8946074202704888e-05, "loss": 0.7851, "step": 4778 }, { "epoch": 0.17308946034045636, "grad_norm": 1.5662966031953949, "learning_rate": 1.894554995237703e-05, "loss": 0.7836, "step": 4779 }, { "epoch": 0.17312567910177473, "grad_norm": 1.5824515537914936, "learning_rate": 1.8945025578950544e-05, "loss": 0.8915, "step": 4780 }, { "epoch": 0.17316189786309308, "grad_norm": 1.5415332967803788, "learning_rate": 1.8944501082432644e-05, "loss": 0.8155, "step": 4781 }, { "epoch": 0.17319811662441145, "grad_norm": 1.4366859548973385, "learning_rate": 1.894397646283055e-05, "loss": 0.8121, "step": 4782 }, { "epoch": 0.1732343353857298, "grad_norm": 1.5972882694617947, "learning_rate": 1.8943451720151486e-05, "loss": 0.9089, "step": 4783 }, { "epoch": 0.17327055414704817, "grad_norm": 1.390832972177649, "learning_rate": 1.894292685440266e-05, "loss": 0.7662, "step": 4784 }, { "epoch": 0.17330677290836655, "grad_norm": 1.1109655085422623, "learning_rate": 1.894240186559131e-05, "loss": 0.79, "step": 4785 }, { "epoch": 0.1733429916696849, "grad_norm": 1.1404113324964713, "learning_rate": 1.8941876753724648e-05, "loss": 0.8267, "step": 4786 }, { "epoch": 0.17337921043100327, "grad_norm": 1.4229303592130826, "learning_rate": 1.8941351518809903e-05, "loss": 0.8538, "step": 4787 }, { "epoch": 0.1734154291923216, "grad_norm": 1.3510441712986998, "learning_rate": 1.894082616085431e-05, "loss": 0.7694, "step": 4788 }, { "epoch": 0.17345164795364, "grad_norm": 1.489386623627755, "learning_rate": 1.8940300679865092e-05, "loss": 0.8455, "step": 4789 }, { "epoch": 0.17348786671495836, "grad_norm": 1.6918774184205125, "learning_rate": 1.893977507584948e-05, "loss": 0.8807, "step": 4790 }, { "epoch": 0.1735240854762767, "grad_norm": 1.617672007999919, "learning_rate": 1.8939249348814704e-05, "loss": 0.8287, "step": 4791 }, { "epoch": 0.17356030423759508, "grad_norm": 1.6155627238668857, "learning_rate": 1.8938723498768007e-05, "loss": 0.9253, "step": 4792 }, { "epoch": 0.17359652299891343, "grad_norm": 1.4198532238752608, "learning_rate": 1.893819752571662e-05, "loss": 0.826, "step": 4793 }, { "epoch": 0.1736327417602318, "grad_norm": 1.5375060135638479, "learning_rate": 1.893767142966778e-05, "loss": 0.7938, "step": 4794 }, { "epoch": 0.17366896052155018, "grad_norm": 1.6042544167856145, "learning_rate": 1.8937145210628727e-05, "loss": 0.821, "step": 4795 }, { "epoch": 0.17370517928286852, "grad_norm": 1.5765800467536422, "learning_rate": 1.8936618868606707e-05, "loss": 0.8626, "step": 4796 }, { "epoch": 0.1737413980441869, "grad_norm": 1.5989213041176342, "learning_rate": 1.8936092403608954e-05, "loss": 0.8933, "step": 4797 }, { "epoch": 0.17377761680550524, "grad_norm": 1.5338002238856647, "learning_rate": 1.893556581564272e-05, "loss": 0.8476, "step": 4798 }, { "epoch": 0.17381383556682362, "grad_norm": 1.558602085664874, "learning_rate": 1.893503910471525e-05, "loss": 0.7717, "step": 4799 }, { "epoch": 0.173850054328142, "grad_norm": 1.4678612021008315, "learning_rate": 1.8934512270833787e-05, "loss": 0.8063, "step": 4800 }, { "epoch": 0.17388627308946034, "grad_norm": 1.5484783016866854, "learning_rate": 1.8933985314005587e-05, "loss": 0.8954, "step": 4801 }, { "epoch": 0.1739224918507787, "grad_norm": 1.5184336071254982, "learning_rate": 1.89334582342379e-05, "loss": 0.8576, "step": 4802 }, { "epoch": 0.17395871061209706, "grad_norm": 1.5748492083571763, "learning_rate": 1.893293103153798e-05, "loss": 0.7803, "step": 4803 }, { "epoch": 0.17399492937341543, "grad_norm": 1.6285206969656394, "learning_rate": 1.8932403705913077e-05, "loss": 0.8668, "step": 4804 }, { "epoch": 0.1740311481347338, "grad_norm": 1.5745609753910634, "learning_rate": 1.8931876257370448e-05, "loss": 0.8388, "step": 4805 }, { "epoch": 0.17406736689605215, "grad_norm": 1.2809875444986325, "learning_rate": 1.8931348685917354e-05, "loss": 0.827, "step": 4806 }, { "epoch": 0.17410358565737052, "grad_norm": 1.3514932943931908, "learning_rate": 1.8930820991561057e-05, "loss": 0.7675, "step": 4807 }, { "epoch": 0.17413980441868887, "grad_norm": 1.8367177025805381, "learning_rate": 1.8930293174308815e-05, "loss": 0.8833, "step": 4808 }, { "epoch": 0.17417602318000724, "grad_norm": 1.5937054454896118, "learning_rate": 1.8929765234167893e-05, "loss": 0.8452, "step": 4809 }, { "epoch": 0.17421224194132562, "grad_norm": 1.2397008914301753, "learning_rate": 1.8929237171145554e-05, "loss": 0.7823, "step": 4810 }, { "epoch": 0.17424846070264396, "grad_norm": 1.4571957367511466, "learning_rate": 1.8928708985249065e-05, "loss": 0.7579, "step": 4811 }, { "epoch": 0.17428467946396234, "grad_norm": 1.4995134956699625, "learning_rate": 1.8928180676485697e-05, "loss": 0.843, "step": 4812 }, { "epoch": 0.17432089822528068, "grad_norm": 1.2794452444699647, "learning_rate": 1.8927652244862718e-05, "loss": 0.8327, "step": 4813 }, { "epoch": 0.17435711698659906, "grad_norm": 1.4437716312150373, "learning_rate": 1.8927123690387397e-05, "loss": 0.816, "step": 4814 }, { "epoch": 0.17439333574791743, "grad_norm": 1.4357243914714939, "learning_rate": 1.892659501306701e-05, "loss": 0.7893, "step": 4815 }, { "epoch": 0.17442955450923578, "grad_norm": 1.1483025123309238, "learning_rate": 1.8926066212908835e-05, "loss": 0.749, "step": 4816 }, { "epoch": 0.17446577327055415, "grad_norm": 1.1586786228293398, "learning_rate": 1.892553728992014e-05, "loss": 0.8199, "step": 4817 }, { "epoch": 0.1745019920318725, "grad_norm": 1.4759027026207987, "learning_rate": 1.8925008244108215e-05, "loss": 0.8575, "step": 4818 }, { "epoch": 0.17453821079319087, "grad_norm": 1.1202075305194024, "learning_rate": 1.892447907548033e-05, "loss": 0.769, "step": 4819 }, { "epoch": 0.17457442955450925, "grad_norm": 1.5841631453323792, "learning_rate": 1.8923949784043777e-05, "loss": 0.8615, "step": 4820 }, { "epoch": 0.1746106483158276, "grad_norm": 1.6809126211868766, "learning_rate": 1.892342036980583e-05, "loss": 0.9304, "step": 4821 }, { "epoch": 0.17464686707714597, "grad_norm": 1.601756916806413, "learning_rate": 1.892289083277378e-05, "loss": 0.8637, "step": 4822 }, { "epoch": 0.1746830858384643, "grad_norm": 1.4722921138818708, "learning_rate": 1.892236117295491e-05, "loss": 0.8007, "step": 4823 }, { "epoch": 0.1747193045997827, "grad_norm": 1.6956214579310755, "learning_rate": 1.892183139035651e-05, "loss": 0.8783, "step": 4824 }, { "epoch": 0.17475552336110106, "grad_norm": 1.477671623586658, "learning_rate": 1.892130148498587e-05, "loss": 0.8233, "step": 4825 }, { "epoch": 0.1747917421224194, "grad_norm": 1.5632283599290764, "learning_rate": 1.892077145685028e-05, "loss": 0.8677, "step": 4826 }, { "epoch": 0.17482796088373778, "grad_norm": 1.5561863788082246, "learning_rate": 1.8920241305957043e-05, "loss": 0.8379, "step": 4827 }, { "epoch": 0.17486417964505613, "grad_norm": 1.5228712801759432, "learning_rate": 1.8919711032313444e-05, "loss": 0.7228, "step": 4828 }, { "epoch": 0.1749003984063745, "grad_norm": 1.5474863361555016, "learning_rate": 1.8919180635926782e-05, "loss": 0.7332, "step": 4829 }, { "epoch": 0.17493661716769288, "grad_norm": 1.3109923684296692, "learning_rate": 1.891865011680436e-05, "loss": 0.751, "step": 4830 }, { "epoch": 0.17497283592901122, "grad_norm": 1.5394287962395852, "learning_rate": 1.8918119474953477e-05, "loss": 0.8593, "step": 4831 }, { "epoch": 0.1750090546903296, "grad_norm": 1.6330189996397075, "learning_rate": 1.8917588710381434e-05, "loss": 0.8888, "step": 4832 }, { "epoch": 0.17504527345164794, "grad_norm": 1.533181417518068, "learning_rate": 1.8917057823095533e-05, "loss": 0.89, "step": 4833 }, { "epoch": 0.17508149221296632, "grad_norm": 1.5231877592986283, "learning_rate": 1.891652681310308e-05, "loss": 0.834, "step": 4834 }, { "epoch": 0.1751177109742847, "grad_norm": 1.6217735324478777, "learning_rate": 1.8915995680411385e-05, "loss": 0.8357, "step": 4835 }, { "epoch": 0.17515392973560304, "grad_norm": 1.5687108887018149, "learning_rate": 1.8915464425027753e-05, "loss": 0.838, "step": 4836 }, { "epoch": 0.1751901484969214, "grad_norm": 1.3360217305589246, "learning_rate": 1.89149330469595e-05, "loss": 0.8533, "step": 4837 }, { "epoch": 0.17522636725823976, "grad_norm": 1.5407264544526076, "learning_rate": 1.8914401546213935e-05, "loss": 0.7431, "step": 4838 }, { "epoch": 0.17526258601955813, "grad_norm": 1.413217718996013, "learning_rate": 1.8913869922798368e-05, "loss": 0.7707, "step": 4839 }, { "epoch": 0.1752988047808765, "grad_norm": 1.6179581248083246, "learning_rate": 1.891333817672012e-05, "loss": 0.864, "step": 4840 }, { "epoch": 0.17533502354219485, "grad_norm": 1.351076672468725, "learning_rate": 1.891280630798651e-05, "loss": 0.7704, "step": 4841 }, { "epoch": 0.17537124230351323, "grad_norm": 1.5803207993850497, "learning_rate": 1.891227431660485e-05, "loss": 0.8737, "step": 4842 }, { "epoch": 0.17540746106483157, "grad_norm": 1.458240212010509, "learning_rate": 1.8911742202582463e-05, "loss": 0.8373, "step": 4843 }, { "epoch": 0.17544367982614995, "grad_norm": 1.273929960010791, "learning_rate": 1.8911209965926677e-05, "loss": 0.8131, "step": 4844 }, { "epoch": 0.17547989858746832, "grad_norm": 1.4487047292790427, "learning_rate": 1.891067760664481e-05, "loss": 0.7977, "step": 4845 }, { "epoch": 0.17551611734878667, "grad_norm": 1.444039029264168, "learning_rate": 1.891014512474419e-05, "loss": 0.7303, "step": 4846 }, { "epoch": 0.17555233611010504, "grad_norm": 1.4675032290166459, "learning_rate": 1.890961252023214e-05, "loss": 0.7642, "step": 4847 }, { "epoch": 0.17558855487142339, "grad_norm": 1.2232033341372512, "learning_rate": 1.8909079793115998e-05, "loss": 0.8184, "step": 4848 }, { "epoch": 0.17562477363274176, "grad_norm": 1.6067905229034969, "learning_rate": 1.8908546943403087e-05, "loss": 0.7933, "step": 4849 }, { "epoch": 0.17566099239406013, "grad_norm": 1.2975961523203636, "learning_rate": 1.8908013971100744e-05, "loss": 0.7856, "step": 4850 }, { "epoch": 0.17569721115537848, "grad_norm": 1.1433134658060604, "learning_rate": 1.89074808762163e-05, "loss": 0.7549, "step": 4851 }, { "epoch": 0.17573342991669685, "grad_norm": 1.5954135389211002, "learning_rate": 1.8906947658757095e-05, "loss": 0.8206, "step": 4852 }, { "epoch": 0.1757696486780152, "grad_norm": 1.5610455890601616, "learning_rate": 1.8906414318730462e-05, "loss": 0.8819, "step": 4853 }, { "epoch": 0.17580586743933357, "grad_norm": 1.6841343710915468, "learning_rate": 1.8905880856143742e-05, "loss": 0.8769, "step": 4854 }, { "epoch": 0.17584208620065195, "grad_norm": 1.2955322569348655, "learning_rate": 1.8905347271004277e-05, "loss": 0.7927, "step": 4855 }, { "epoch": 0.1758783049619703, "grad_norm": 1.2986942155902372, "learning_rate": 1.890481356331941e-05, "loss": 0.8765, "step": 4856 }, { "epoch": 0.17591452372328867, "grad_norm": 1.726699926119956, "learning_rate": 1.890427973309648e-05, "loss": 0.8766, "step": 4857 }, { "epoch": 0.17595074248460701, "grad_norm": 1.8366845437243278, "learning_rate": 1.8903745780342838e-05, "loss": 0.8515, "step": 4858 }, { "epoch": 0.1759869612459254, "grad_norm": 1.8396351750713174, "learning_rate": 1.8903211705065833e-05, "loss": 0.9055, "step": 4859 }, { "epoch": 0.17602318000724376, "grad_norm": 1.452733781921643, "learning_rate": 1.890267750727281e-05, "loss": 0.8347, "step": 4860 }, { "epoch": 0.1760593987685621, "grad_norm": 1.6127023616589127, "learning_rate": 1.8902143186971123e-05, "loss": 0.7879, "step": 4861 }, { "epoch": 0.17609561752988048, "grad_norm": 1.4365109968910952, "learning_rate": 1.890160874416812e-05, "loss": 0.8072, "step": 4862 }, { "epoch": 0.17613183629119883, "grad_norm": 1.419203975871584, "learning_rate": 1.8901074178871163e-05, "loss": 0.8198, "step": 4863 }, { "epoch": 0.1761680550525172, "grad_norm": 1.4944548382445906, "learning_rate": 1.8900539491087605e-05, "loss": 0.8031, "step": 4864 }, { "epoch": 0.17620427381383558, "grad_norm": 1.5027296498300957, "learning_rate": 1.8900004680824798e-05, "loss": 0.7316, "step": 4865 }, { "epoch": 0.17624049257515392, "grad_norm": 1.4953106842807282, "learning_rate": 1.889946974809011e-05, "loss": 0.8793, "step": 4866 }, { "epoch": 0.1762767113364723, "grad_norm": 1.3875972342680114, "learning_rate": 1.8898934692890897e-05, "loss": 0.8184, "step": 4867 }, { "epoch": 0.17631293009779064, "grad_norm": 1.5271387962722924, "learning_rate": 1.8898399515234525e-05, "loss": 0.7846, "step": 4868 }, { "epoch": 0.17634914885910902, "grad_norm": 1.5853288133095067, "learning_rate": 1.8897864215128356e-05, "loss": 0.7931, "step": 4869 }, { "epoch": 0.1763853676204274, "grad_norm": 1.584974939939047, "learning_rate": 1.889732879257976e-05, "loss": 0.8066, "step": 4870 }, { "epoch": 0.17642158638174574, "grad_norm": 1.534365935891255, "learning_rate": 1.88967932475961e-05, "loss": 0.8776, "step": 4871 }, { "epoch": 0.1764578051430641, "grad_norm": 1.619920271391359, "learning_rate": 1.8896257580184744e-05, "loss": 0.8767, "step": 4872 }, { "epoch": 0.17649402390438246, "grad_norm": 1.3999616140026923, "learning_rate": 1.889572179035307e-05, "loss": 0.7944, "step": 4873 }, { "epoch": 0.17653024266570083, "grad_norm": 1.408782300214934, "learning_rate": 1.8895185878108445e-05, "loss": 0.7514, "step": 4874 }, { "epoch": 0.1765664614270192, "grad_norm": 1.5250672383464805, "learning_rate": 1.889464984345825e-05, "loss": 0.8087, "step": 4875 }, { "epoch": 0.17660268018833755, "grad_norm": 1.4464578600870852, "learning_rate": 1.8894113686409854e-05, "loss": 0.8648, "step": 4876 }, { "epoch": 0.17663889894965593, "grad_norm": 1.6206799174256403, "learning_rate": 1.8893577406970642e-05, "loss": 0.8473, "step": 4877 }, { "epoch": 0.17667511771097427, "grad_norm": 1.3383874096160162, "learning_rate": 1.889304100514799e-05, "loss": 0.8042, "step": 4878 }, { "epoch": 0.17671133647229265, "grad_norm": 1.7050358416371263, "learning_rate": 1.8892504480949275e-05, "loss": 0.7745, "step": 4879 }, { "epoch": 0.17674755523361102, "grad_norm": 1.4420048793857818, "learning_rate": 1.889196783438189e-05, "loss": 0.7248, "step": 4880 }, { "epoch": 0.17678377399492937, "grad_norm": 1.571238164272964, "learning_rate": 1.8891431065453216e-05, "loss": 0.8776, "step": 4881 }, { "epoch": 0.17681999275624774, "grad_norm": 1.7267503226292198, "learning_rate": 1.889089417417063e-05, "loss": 0.8851, "step": 4882 }, { "epoch": 0.1768562115175661, "grad_norm": 1.573250839954312, "learning_rate": 1.8890357160541533e-05, "loss": 0.7539, "step": 4883 }, { "epoch": 0.17689243027888446, "grad_norm": 1.6354537094582102, "learning_rate": 1.8889820024573308e-05, "loss": 0.8105, "step": 4884 }, { "epoch": 0.17692864904020283, "grad_norm": 1.5511779325852553, "learning_rate": 1.888928276627335e-05, "loss": 0.769, "step": 4885 }, { "epoch": 0.17696486780152118, "grad_norm": 1.4318088602433527, "learning_rate": 1.888874538564905e-05, "loss": 0.8084, "step": 4886 }, { "epoch": 0.17700108656283955, "grad_norm": 1.4393408434242774, "learning_rate": 1.8888207882707802e-05, "loss": 0.8508, "step": 4887 }, { "epoch": 0.1770373053241579, "grad_norm": 1.074274183273556, "learning_rate": 1.8887670257457006e-05, "loss": 0.6893, "step": 4888 }, { "epoch": 0.17707352408547627, "grad_norm": 1.5275875992033177, "learning_rate": 1.8887132509904054e-05, "loss": 0.9087, "step": 4889 }, { "epoch": 0.17710974284679465, "grad_norm": 1.502166229428373, "learning_rate": 1.8886594640056354e-05, "loss": 0.8046, "step": 4890 }, { "epoch": 0.177145961608113, "grad_norm": 1.4349631605186912, "learning_rate": 1.88860566479213e-05, "loss": 0.8464, "step": 4891 }, { "epoch": 0.17718218036943137, "grad_norm": 1.4236549242989596, "learning_rate": 1.88855185335063e-05, "loss": 0.8471, "step": 4892 }, { "epoch": 0.17721839913074972, "grad_norm": 1.5351676521897022, "learning_rate": 1.888498029681876e-05, "loss": 0.8813, "step": 4893 }, { "epoch": 0.1772546178920681, "grad_norm": 1.5858389825006385, "learning_rate": 1.888444193786608e-05, "loss": 0.9106, "step": 4894 }, { "epoch": 0.17729083665338646, "grad_norm": 1.3842521335519655, "learning_rate": 1.888390345665567e-05, "loss": 0.8177, "step": 4895 }, { "epoch": 0.1773270554147048, "grad_norm": 1.1160734617936396, "learning_rate": 1.888336485319495e-05, "loss": 0.7755, "step": 4896 }, { "epoch": 0.17736327417602318, "grad_norm": 1.5650076484645623, "learning_rate": 1.888282612749132e-05, "loss": 0.8163, "step": 4897 }, { "epoch": 0.17739949293734153, "grad_norm": 1.2616712386675346, "learning_rate": 1.8882287279552197e-05, "loss": 0.7394, "step": 4898 }, { "epoch": 0.1774357116986599, "grad_norm": 1.4460754648924905, "learning_rate": 1.8881748309384998e-05, "loss": 0.707, "step": 4899 }, { "epoch": 0.17747193045997828, "grad_norm": 1.521235756644802, "learning_rate": 1.888120921699714e-05, "loss": 0.8967, "step": 4900 }, { "epoch": 0.17750814922129662, "grad_norm": 1.4086280388686545, "learning_rate": 1.8880670002396034e-05, "loss": 0.8269, "step": 4901 }, { "epoch": 0.177544367982615, "grad_norm": 1.4671692496171145, "learning_rate": 1.8880130665589106e-05, "loss": 0.8275, "step": 4902 }, { "epoch": 0.17758058674393334, "grad_norm": 1.2845352651468993, "learning_rate": 1.887959120658378e-05, "loss": 0.7598, "step": 4903 }, { "epoch": 0.17761680550525172, "grad_norm": 1.1520777221495475, "learning_rate": 1.8879051625387477e-05, "loss": 0.8589, "step": 4904 }, { "epoch": 0.1776530242665701, "grad_norm": 1.579506626914581, "learning_rate": 1.887851192200762e-05, "loss": 0.7874, "step": 4905 }, { "epoch": 0.17768924302788844, "grad_norm": 1.6974549336097344, "learning_rate": 1.887797209645164e-05, "loss": 0.9801, "step": 4906 }, { "epoch": 0.1777254617892068, "grad_norm": 1.5691543324474586, "learning_rate": 1.8877432148726962e-05, "loss": 0.8629, "step": 4907 }, { "epoch": 0.17776168055052516, "grad_norm": 1.5528016705595034, "learning_rate": 1.8876892078841017e-05, "loss": 0.8666, "step": 4908 }, { "epoch": 0.17779789931184353, "grad_norm": 1.636685127513837, "learning_rate": 1.8876351886801236e-05, "loss": 0.8783, "step": 4909 }, { "epoch": 0.1778341180731619, "grad_norm": 1.4254923122741274, "learning_rate": 1.887581157261505e-05, "loss": 0.8329, "step": 4910 }, { "epoch": 0.17787033683448025, "grad_norm": 1.3244558559178166, "learning_rate": 1.8875271136289904e-05, "loss": 0.84, "step": 4911 }, { "epoch": 0.17790655559579863, "grad_norm": 1.6876733125767636, "learning_rate": 1.8874730577833228e-05, "loss": 0.8582, "step": 4912 }, { "epoch": 0.17794277435711697, "grad_norm": 1.625928636365812, "learning_rate": 1.8874189897252463e-05, "loss": 0.8625, "step": 4913 }, { "epoch": 0.17797899311843535, "grad_norm": 1.446594782938991, "learning_rate": 1.8873649094555042e-05, "loss": 0.8462, "step": 4914 }, { "epoch": 0.17801521187975372, "grad_norm": 1.429520332746298, "learning_rate": 1.8873108169748414e-05, "loss": 0.8568, "step": 4915 }, { "epoch": 0.17805143064107207, "grad_norm": 1.586655210070795, "learning_rate": 1.8872567122840024e-05, "loss": 0.8264, "step": 4916 }, { "epoch": 0.17808764940239044, "grad_norm": 1.4843712430203728, "learning_rate": 1.887202595383731e-05, "loss": 0.8083, "step": 4917 }, { "epoch": 0.1781238681637088, "grad_norm": 1.5339782820605607, "learning_rate": 1.8871484662747726e-05, "loss": 0.8457, "step": 4918 }, { "epoch": 0.17816008692502716, "grad_norm": 1.471104480366968, "learning_rate": 1.8870943249578715e-05, "loss": 0.9037, "step": 4919 }, { "epoch": 0.17819630568634554, "grad_norm": 1.5764138551236124, "learning_rate": 1.8870401714337736e-05, "loss": 0.8737, "step": 4920 }, { "epoch": 0.17823252444766388, "grad_norm": 1.6752786271613533, "learning_rate": 1.8869860057032234e-05, "loss": 0.7868, "step": 4921 }, { "epoch": 0.17826874320898226, "grad_norm": 1.5195225424872536, "learning_rate": 1.886931827766966e-05, "loss": 0.85, "step": 4922 }, { "epoch": 0.17830496197030063, "grad_norm": 1.3655332571897514, "learning_rate": 1.8868776376257476e-05, "loss": 0.8742, "step": 4923 }, { "epoch": 0.17834118073161898, "grad_norm": 1.3832301318045719, "learning_rate": 1.8868234352803134e-05, "loss": 0.8804, "step": 4924 }, { "epoch": 0.17837739949293735, "grad_norm": 1.0996128493206705, "learning_rate": 1.8867692207314096e-05, "loss": 0.8223, "step": 4925 }, { "epoch": 0.1784136182542557, "grad_norm": 1.4622852632020442, "learning_rate": 1.8867149939797822e-05, "loss": 0.8774, "step": 4926 }, { "epoch": 0.17844983701557407, "grad_norm": 1.340398204538796, "learning_rate": 1.8866607550261773e-05, "loss": 0.8036, "step": 4927 }, { "epoch": 0.17848605577689244, "grad_norm": 1.4836648836102206, "learning_rate": 1.8866065038713414e-05, "loss": 0.7742, "step": 4928 }, { "epoch": 0.1785222745382108, "grad_norm": 1.45308112390994, "learning_rate": 1.8865522405160207e-05, "loss": 0.795, "step": 4929 }, { "epoch": 0.17855849329952916, "grad_norm": 1.4141601935777117, "learning_rate": 1.8864979649609624e-05, "loss": 0.8072, "step": 4930 }, { "epoch": 0.1785947120608475, "grad_norm": 1.1913942120483278, "learning_rate": 1.886443677206913e-05, "loss": 0.7878, "step": 4931 }, { "epoch": 0.17863093082216588, "grad_norm": 1.4943414839827482, "learning_rate": 1.88638937725462e-05, "loss": 0.8219, "step": 4932 }, { "epoch": 0.17866714958348426, "grad_norm": 1.1042010721445152, "learning_rate": 1.88633506510483e-05, "loss": 0.7866, "step": 4933 }, { "epoch": 0.1787033683448026, "grad_norm": 1.6163753193990653, "learning_rate": 1.8862807407582907e-05, "loss": 0.7903, "step": 4934 }, { "epoch": 0.17873958710612098, "grad_norm": 1.5777985588093608, "learning_rate": 1.88622640421575e-05, "loss": 0.839, "step": 4935 }, { "epoch": 0.17877580586743932, "grad_norm": 1.6139667707860936, "learning_rate": 1.886172055477955e-05, "loss": 0.8639, "step": 4936 }, { "epoch": 0.1788120246287577, "grad_norm": 1.2492218781971323, "learning_rate": 1.8861176945456542e-05, "loss": 0.8201, "step": 4937 }, { "epoch": 0.17884824339007607, "grad_norm": 1.5794757524703482, "learning_rate": 1.8860633214195947e-05, "loss": 0.7859, "step": 4938 }, { "epoch": 0.17888446215139442, "grad_norm": 1.7679243654907526, "learning_rate": 1.8860089361005255e-05, "loss": 0.8587, "step": 4939 }, { "epoch": 0.1789206809127128, "grad_norm": 1.5062181441833289, "learning_rate": 1.8859545385891952e-05, "loss": 0.7519, "step": 4940 }, { "epoch": 0.17895689967403114, "grad_norm": 1.801319951969818, "learning_rate": 1.885900128886352e-05, "loss": 0.8796, "step": 4941 }, { "epoch": 0.1789931184353495, "grad_norm": 1.5093310842384424, "learning_rate": 1.8858457069927443e-05, "loss": 0.8222, "step": 4942 }, { "epoch": 0.1790293371966679, "grad_norm": 1.1385734262098857, "learning_rate": 1.8857912729091213e-05, "loss": 0.794, "step": 4943 }, { "epoch": 0.17906555595798623, "grad_norm": 1.5324310211299603, "learning_rate": 1.885736826636232e-05, "loss": 0.787, "step": 4944 }, { "epoch": 0.1791017747193046, "grad_norm": 1.6118013288225508, "learning_rate": 1.885682368174826e-05, "loss": 0.7883, "step": 4945 }, { "epoch": 0.17913799348062295, "grad_norm": 1.5936959435140428, "learning_rate": 1.885627897525652e-05, "loss": 0.7992, "step": 4946 }, { "epoch": 0.17917421224194133, "grad_norm": 1.4886796692088877, "learning_rate": 1.88557341468946e-05, "loss": 0.8171, "step": 4947 }, { "epoch": 0.1792104310032597, "grad_norm": 1.4811069090076836, "learning_rate": 1.8855189196670004e-05, "loss": 0.8658, "step": 4948 }, { "epoch": 0.17924664976457805, "grad_norm": 1.5766975809512866, "learning_rate": 1.8854644124590216e-05, "loss": 0.848, "step": 4949 }, { "epoch": 0.17928286852589642, "grad_norm": 1.484340314710835, "learning_rate": 1.8854098930662744e-05, "loss": 0.8108, "step": 4950 }, { "epoch": 0.17931908728721477, "grad_norm": 1.4613111242607046, "learning_rate": 1.8853553614895098e-05, "loss": 0.8543, "step": 4951 }, { "epoch": 0.17935530604853314, "grad_norm": 1.4402801836952535, "learning_rate": 1.8853008177294766e-05, "loss": 0.8122, "step": 4952 }, { "epoch": 0.17939152480985152, "grad_norm": 1.5642414490051912, "learning_rate": 1.885246261786927e-05, "loss": 0.7989, "step": 4953 }, { "epoch": 0.17942774357116986, "grad_norm": 1.5336992615186176, "learning_rate": 1.8851916936626104e-05, "loss": 0.887, "step": 4954 }, { "epoch": 0.17946396233248824, "grad_norm": 1.2605354049380646, "learning_rate": 1.8851371133572787e-05, "loss": 0.7607, "step": 4955 }, { "epoch": 0.17950018109380658, "grad_norm": 1.6487020284826084, "learning_rate": 1.8850825208716824e-05, "loss": 0.887, "step": 4956 }, { "epoch": 0.17953639985512496, "grad_norm": 1.6386779954407236, "learning_rate": 1.885027916206573e-05, "loss": 0.8862, "step": 4957 }, { "epoch": 0.17957261861644333, "grad_norm": 1.6853287435240774, "learning_rate": 1.884973299362702e-05, "loss": 0.913, "step": 4958 }, { "epoch": 0.17960883737776168, "grad_norm": 1.7296733190546167, "learning_rate": 1.8849186703408205e-05, "loss": 0.7558, "step": 4959 }, { "epoch": 0.17964505613908005, "grad_norm": 1.5599701607105414, "learning_rate": 1.8848640291416807e-05, "loss": 0.859, "step": 4960 }, { "epoch": 0.1796812749003984, "grad_norm": 1.5026478766546898, "learning_rate": 1.884809375766034e-05, "loss": 0.8766, "step": 4961 }, { "epoch": 0.17971749366171677, "grad_norm": 1.4395441656683305, "learning_rate": 1.884754710214633e-05, "loss": 0.7431, "step": 4962 }, { "epoch": 0.17975371242303514, "grad_norm": 1.4028175617484882, "learning_rate": 1.88470003248823e-05, "loss": 0.8673, "step": 4963 }, { "epoch": 0.1797899311843535, "grad_norm": 1.4498358695268987, "learning_rate": 1.884645342587577e-05, "loss": 0.8702, "step": 4964 }, { "epoch": 0.17982614994567186, "grad_norm": 1.2559384803864249, "learning_rate": 1.884590640513427e-05, "loss": 0.8008, "step": 4965 }, { "epoch": 0.1798623687069902, "grad_norm": 1.6108956085940977, "learning_rate": 1.8845359262665322e-05, "loss": 0.8612, "step": 4966 }, { "epoch": 0.17989858746830859, "grad_norm": 1.3652554673774404, "learning_rate": 1.884481199847646e-05, "loss": 0.8213, "step": 4967 }, { "epoch": 0.17993480622962696, "grad_norm": 1.7618676670344064, "learning_rate": 1.8844264612575213e-05, "loss": 0.8536, "step": 4968 }, { "epoch": 0.1799710249909453, "grad_norm": 1.9899323171518573, "learning_rate": 1.8843717104969114e-05, "loss": 0.8368, "step": 4969 }, { "epoch": 0.18000724375226368, "grad_norm": 1.6409220721722415, "learning_rate": 1.8843169475665698e-05, "loss": 0.8049, "step": 4970 }, { "epoch": 0.18004346251358203, "grad_norm": 1.4427160118400202, "learning_rate": 1.8842621724672498e-05, "loss": 0.808, "step": 4971 }, { "epoch": 0.1800796812749004, "grad_norm": 1.2166600919691066, "learning_rate": 1.884207385199705e-05, "loss": 0.7856, "step": 4972 }, { "epoch": 0.18011590003621877, "grad_norm": 1.4810146807308238, "learning_rate": 1.8841525857646905e-05, "loss": 0.7918, "step": 4973 }, { "epoch": 0.18015211879753712, "grad_norm": 1.4551567504428924, "learning_rate": 1.884097774162959e-05, "loss": 0.8366, "step": 4974 }, { "epoch": 0.1801883375588555, "grad_norm": 1.397658575692009, "learning_rate": 1.884042950395265e-05, "loss": 0.7397, "step": 4975 }, { "epoch": 0.18022455632017384, "grad_norm": 1.556265819973757, "learning_rate": 1.883988114462364e-05, "loss": 0.8646, "step": 4976 }, { "epoch": 0.1802607750814922, "grad_norm": 1.4524085857864133, "learning_rate": 1.8839332663650092e-05, "loss": 0.8258, "step": 4977 }, { "epoch": 0.1802969938428106, "grad_norm": 2.0971291588421823, "learning_rate": 1.883878406103956e-05, "loss": 0.856, "step": 4978 }, { "epoch": 0.18033321260412893, "grad_norm": 1.5039691738489742, "learning_rate": 1.8838235336799593e-05, "loss": 0.8992, "step": 4979 }, { "epoch": 0.1803694313654473, "grad_norm": 1.5680089175397003, "learning_rate": 1.8837686490937744e-05, "loss": 0.8465, "step": 4980 }, { "epoch": 0.18040565012676565, "grad_norm": 1.4335237246881256, "learning_rate": 1.8837137523461558e-05, "loss": 0.7475, "step": 4981 }, { "epoch": 0.18044186888808403, "grad_norm": 1.3979514955081713, "learning_rate": 1.8836588434378603e-05, "loss": 0.7939, "step": 4982 }, { "epoch": 0.1804780876494024, "grad_norm": 1.5081258161679338, "learning_rate": 1.883603922369642e-05, "loss": 0.8783, "step": 4983 }, { "epoch": 0.18051430641072075, "grad_norm": 1.388426549785894, "learning_rate": 1.883548989142258e-05, "loss": 0.7716, "step": 4984 }, { "epoch": 0.18055052517203912, "grad_norm": 1.378532966332823, "learning_rate": 1.8834940437564628e-05, "loss": 0.7994, "step": 4985 }, { "epoch": 0.18058674393335747, "grad_norm": 1.514559503990506, "learning_rate": 1.8834390862130134e-05, "loss": 0.8308, "step": 4986 }, { "epoch": 0.18062296269467584, "grad_norm": 1.537038006513194, "learning_rate": 1.8833841165126662e-05, "loss": 0.7926, "step": 4987 }, { "epoch": 0.18065918145599422, "grad_norm": 1.6303078001360232, "learning_rate": 1.8833291346561772e-05, "loss": 0.8262, "step": 4988 }, { "epoch": 0.18069540021731256, "grad_norm": 1.3051267412037, "learning_rate": 1.8832741406443033e-05, "loss": 0.7817, "step": 4989 }, { "epoch": 0.18073161897863094, "grad_norm": 1.5883240794380165, "learning_rate": 1.883219134477801e-05, "loss": 0.825, "step": 4990 }, { "epoch": 0.18076783773994928, "grad_norm": 1.5220856133583678, "learning_rate": 1.8831641161574272e-05, "loss": 0.7569, "step": 4991 }, { "epoch": 0.18080405650126766, "grad_norm": 1.6991971968237571, "learning_rate": 1.8831090856839393e-05, "loss": 0.8309, "step": 4992 }, { "epoch": 0.18084027526258603, "grad_norm": 1.5905861878390786, "learning_rate": 1.8830540430580947e-05, "loss": 0.7659, "step": 4993 }, { "epoch": 0.18087649402390438, "grad_norm": 1.4000411252570988, "learning_rate": 1.88299898828065e-05, "loss": 0.8372, "step": 4994 }, { "epoch": 0.18091271278522275, "grad_norm": 1.3962700199125864, "learning_rate": 1.8829439213523636e-05, "loss": 0.774, "step": 4995 }, { "epoch": 0.1809489315465411, "grad_norm": 1.4971946532787799, "learning_rate": 1.8828888422739933e-05, "loss": 0.8398, "step": 4996 }, { "epoch": 0.18098515030785947, "grad_norm": 1.641799710561957, "learning_rate": 1.882833751046296e-05, "loss": 0.8241, "step": 4997 }, { "epoch": 0.18102136906917785, "grad_norm": 1.470498533063823, "learning_rate": 1.882778647670031e-05, "loss": 0.7883, "step": 4998 }, { "epoch": 0.1810575878304962, "grad_norm": 1.7067245867003467, "learning_rate": 1.8827235321459565e-05, "loss": 0.8096, "step": 4999 }, { "epoch": 0.18109380659181457, "grad_norm": 1.6041643825944103, "learning_rate": 1.8826684044748303e-05, "loss": 0.8918, "step": 5000 }, { "epoch": 0.1811300253531329, "grad_norm": 1.1998239142189049, "learning_rate": 1.882613264657411e-05, "loss": 0.7843, "step": 5001 }, { "epoch": 0.18116624411445129, "grad_norm": 1.4544105406468975, "learning_rate": 1.882558112694458e-05, "loss": 0.8401, "step": 5002 }, { "epoch": 0.18120246287576966, "grad_norm": 1.5167945160603067, "learning_rate": 1.88250294858673e-05, "loss": 0.8406, "step": 5003 }, { "epoch": 0.181238681637088, "grad_norm": 1.3293601435497893, "learning_rate": 1.8824477723349853e-05, "loss": 0.8006, "step": 5004 }, { "epoch": 0.18127490039840638, "grad_norm": 1.845986935830397, "learning_rate": 1.8823925839399845e-05, "loss": 0.894, "step": 5005 }, { "epoch": 0.18131111915972473, "grad_norm": 1.4754097626348583, "learning_rate": 1.8823373834024862e-05, "loss": 0.7458, "step": 5006 }, { "epoch": 0.1813473379210431, "grad_norm": 1.697038209470811, "learning_rate": 1.8822821707232503e-05, "loss": 0.8221, "step": 5007 }, { "epoch": 0.18138355668236147, "grad_norm": 1.5368130045553368, "learning_rate": 1.8822269459030362e-05, "loss": 0.8541, "step": 5008 }, { "epoch": 0.18141977544367982, "grad_norm": 1.5951679317417666, "learning_rate": 1.8821717089426045e-05, "loss": 0.9058, "step": 5009 }, { "epoch": 0.1814559942049982, "grad_norm": 1.2197251057406142, "learning_rate": 1.8821164598427148e-05, "loss": 0.8516, "step": 5010 }, { "epoch": 0.18149221296631654, "grad_norm": 1.476702368131481, "learning_rate": 1.8820611986041277e-05, "loss": 0.8516, "step": 5011 }, { "epoch": 0.18152843172763491, "grad_norm": 1.8490056479583374, "learning_rate": 1.882005925227603e-05, "loss": 0.8756, "step": 5012 }, { "epoch": 0.1815646504889533, "grad_norm": 1.533009226498612, "learning_rate": 1.881950639713902e-05, "loss": 0.9093, "step": 5013 }, { "epoch": 0.18160086925027163, "grad_norm": 1.5270114833612578, "learning_rate": 1.8818953420637855e-05, "loss": 0.8884, "step": 5014 }, { "epoch": 0.18163708801159, "grad_norm": 1.5230359501922603, "learning_rate": 1.881840032278014e-05, "loss": 0.8936, "step": 5015 }, { "epoch": 0.18167330677290836, "grad_norm": 1.4564647583227959, "learning_rate": 1.8817847103573484e-05, "loss": 0.8492, "step": 5016 }, { "epoch": 0.18170952553422673, "grad_norm": 1.4863195431595628, "learning_rate": 1.881729376302551e-05, "loss": 0.7705, "step": 5017 }, { "epoch": 0.1817457442955451, "grad_norm": 1.6001416439823513, "learning_rate": 1.8816740301143823e-05, "loss": 0.8587, "step": 5018 }, { "epoch": 0.18178196305686345, "grad_norm": 1.5807060189466449, "learning_rate": 1.8816186717936045e-05, "loss": 0.7974, "step": 5019 }, { "epoch": 0.18181818181818182, "grad_norm": 1.435687910664532, "learning_rate": 1.881563301340979e-05, "loss": 0.8501, "step": 5020 }, { "epoch": 0.18185440057950017, "grad_norm": 1.4275433833988092, "learning_rate": 1.881507918757268e-05, "loss": 0.8079, "step": 5021 }, { "epoch": 0.18189061934081854, "grad_norm": 1.5627664648772894, "learning_rate": 1.8814525240432327e-05, "loss": 0.8435, "step": 5022 }, { "epoch": 0.18192683810213692, "grad_norm": 1.6348967146412845, "learning_rate": 1.881397117199637e-05, "loss": 0.955, "step": 5023 }, { "epoch": 0.18196305686345526, "grad_norm": 1.423705725648154, "learning_rate": 1.8813416982272422e-05, "loss": 0.8349, "step": 5024 }, { "epoch": 0.18199927562477364, "grad_norm": 1.596775886134169, "learning_rate": 1.8812862671268112e-05, "loss": 0.8062, "step": 5025 }, { "epoch": 0.18203549438609198, "grad_norm": 1.9585999009804598, "learning_rate": 1.881230823899107e-05, "loss": 0.9557, "step": 5026 }, { "epoch": 0.18207171314741036, "grad_norm": 1.5267037608160736, "learning_rate": 1.881175368544892e-05, "loss": 0.8481, "step": 5027 }, { "epoch": 0.18210793190872873, "grad_norm": 1.652590428870426, "learning_rate": 1.88111990106493e-05, "loss": 0.8603, "step": 5028 }, { "epoch": 0.18214415067004708, "grad_norm": 1.3743286004057784, "learning_rate": 1.881064421459984e-05, "loss": 0.7967, "step": 5029 }, { "epoch": 0.18218036943136545, "grad_norm": 1.2900469148578808, "learning_rate": 1.881008929730817e-05, "loss": 0.8328, "step": 5030 }, { "epoch": 0.1822165881926838, "grad_norm": 1.585616363465031, "learning_rate": 1.880953425878193e-05, "loss": 0.9442, "step": 5031 }, { "epoch": 0.18225280695400217, "grad_norm": 1.4835016410592838, "learning_rate": 1.8808979099028765e-05, "loss": 0.8621, "step": 5032 }, { "epoch": 0.18228902571532055, "grad_norm": 1.3979688350568313, "learning_rate": 1.88084238180563e-05, "loss": 0.7758, "step": 5033 }, { "epoch": 0.1823252444766389, "grad_norm": 1.626673190063577, "learning_rate": 1.8807868415872187e-05, "loss": 0.8495, "step": 5034 }, { "epoch": 0.18236146323795727, "grad_norm": 1.475668163899232, "learning_rate": 1.8807312892484062e-05, "loss": 0.8439, "step": 5035 }, { "epoch": 0.1823976819992756, "grad_norm": 1.5968392190358505, "learning_rate": 1.8806757247899577e-05, "loss": 0.8682, "step": 5036 }, { "epoch": 0.182433900760594, "grad_norm": 1.6409647215203707, "learning_rate": 1.880620148212637e-05, "loss": 0.8585, "step": 5037 }, { "epoch": 0.18247011952191236, "grad_norm": 1.5998722694120928, "learning_rate": 1.8805645595172096e-05, "loss": 0.812, "step": 5038 }, { "epoch": 0.1825063382832307, "grad_norm": 1.503969638281248, "learning_rate": 1.8805089587044397e-05, "loss": 0.7732, "step": 5039 }, { "epoch": 0.18254255704454908, "grad_norm": 1.5824803644999037, "learning_rate": 1.8804533457750933e-05, "loss": 0.881, "step": 5040 }, { "epoch": 0.18257877580586743, "grad_norm": 1.7078775609223789, "learning_rate": 1.8803977207299352e-05, "loss": 0.8024, "step": 5041 }, { "epoch": 0.1826149945671858, "grad_norm": 1.4638085604857296, "learning_rate": 1.8803420835697307e-05, "loss": 0.8242, "step": 5042 }, { "epoch": 0.18265121332850418, "grad_norm": 1.5528979893019568, "learning_rate": 1.8802864342952457e-05, "loss": 0.8065, "step": 5043 }, { "epoch": 0.18268743208982252, "grad_norm": 1.4700840119295044, "learning_rate": 1.8802307729072456e-05, "loss": 0.7768, "step": 5044 }, { "epoch": 0.1827236508511409, "grad_norm": 1.388830554456507, "learning_rate": 1.8801750994064968e-05, "loss": 0.852, "step": 5045 }, { "epoch": 0.18275986961245924, "grad_norm": 1.6604687295678395, "learning_rate": 1.8801194137937654e-05, "loss": 0.8282, "step": 5046 }, { "epoch": 0.18279608837377762, "grad_norm": 1.5689974546226844, "learning_rate": 1.8800637160698175e-05, "loss": 0.8658, "step": 5047 }, { "epoch": 0.182832307135096, "grad_norm": 1.5331107093512009, "learning_rate": 1.8800080062354197e-05, "loss": 0.7652, "step": 5048 }, { "epoch": 0.18286852589641434, "grad_norm": 1.6297883954365404, "learning_rate": 1.879952284291338e-05, "loss": 0.92, "step": 5049 }, { "epoch": 0.1829047446577327, "grad_norm": 1.4439492910694864, "learning_rate": 1.8798965502383398e-05, "loss": 0.8098, "step": 5050 }, { "epoch": 0.18294096341905106, "grad_norm": 1.6364267436998632, "learning_rate": 1.8798408040771922e-05, "loss": 0.861, "step": 5051 }, { "epoch": 0.18297718218036943, "grad_norm": 1.496953191984657, "learning_rate": 1.879785045808662e-05, "loss": 0.7858, "step": 5052 }, { "epoch": 0.1830134009416878, "grad_norm": 1.40826580950387, "learning_rate": 1.8797292754335164e-05, "loss": 0.841, "step": 5053 }, { "epoch": 0.18304961970300615, "grad_norm": 1.5986781757608817, "learning_rate": 1.8796734929525227e-05, "loss": 0.8167, "step": 5054 }, { "epoch": 0.18308583846432452, "grad_norm": 1.7104651787677818, "learning_rate": 1.8796176983664494e-05, "loss": 0.79, "step": 5055 }, { "epoch": 0.18312205722564287, "grad_norm": 1.6618676794333116, "learning_rate": 1.8795618916760634e-05, "loss": 0.8164, "step": 5056 }, { "epoch": 0.18315827598696124, "grad_norm": 1.8861084754671757, "learning_rate": 1.879506072882133e-05, "loss": 0.8781, "step": 5057 }, { "epoch": 0.18319449474827962, "grad_norm": 1.674875732783426, "learning_rate": 1.879450241985426e-05, "loss": 0.8994, "step": 5058 }, { "epoch": 0.18323071350959796, "grad_norm": 1.7057734794296442, "learning_rate": 1.879394398986711e-05, "loss": 0.8307, "step": 5059 }, { "epoch": 0.18326693227091634, "grad_norm": 1.5748811799975808, "learning_rate": 1.8793385438867564e-05, "loss": 0.8445, "step": 5060 }, { "epoch": 0.18330315103223468, "grad_norm": 1.501869549676901, "learning_rate": 1.879282676686331e-05, "loss": 0.8362, "step": 5061 }, { "epoch": 0.18333936979355306, "grad_norm": 1.5239421491590337, "learning_rate": 1.879226797386203e-05, "loss": 0.853, "step": 5062 }, { "epoch": 0.18337558855487143, "grad_norm": 1.635202308494562, "learning_rate": 1.879170905987142e-05, "loss": 0.8046, "step": 5063 }, { "epoch": 0.18341180731618978, "grad_norm": 1.2783086104473016, "learning_rate": 1.8791150024899164e-05, "loss": 0.8154, "step": 5064 }, { "epoch": 0.18344802607750815, "grad_norm": 1.5164648388643889, "learning_rate": 1.8790590868952963e-05, "loss": 0.8418, "step": 5065 }, { "epoch": 0.1834842448388265, "grad_norm": 1.245480505911805, "learning_rate": 1.879003159204051e-05, "loss": 0.8243, "step": 5066 }, { "epoch": 0.18352046360014487, "grad_norm": 1.6153077047785693, "learning_rate": 1.8789472194169493e-05, "loss": 0.8791, "step": 5067 }, { "epoch": 0.18355668236146325, "grad_norm": 1.6116071702181984, "learning_rate": 1.8788912675347617e-05, "loss": 0.8271, "step": 5068 }, { "epoch": 0.1835929011227816, "grad_norm": 1.5515171077999361, "learning_rate": 1.8788353035582583e-05, "loss": 0.7844, "step": 5069 }, { "epoch": 0.18362911988409997, "grad_norm": 1.451919969955529, "learning_rate": 1.8787793274882084e-05, "loss": 0.8189, "step": 5070 }, { "epoch": 0.1836653386454183, "grad_norm": 1.4532713457472823, "learning_rate": 1.8787233393253832e-05, "loss": 0.8719, "step": 5071 }, { "epoch": 0.1837015574067367, "grad_norm": 1.4602472866522878, "learning_rate": 1.8786673390705528e-05, "loss": 0.7789, "step": 5072 }, { "epoch": 0.18373777616805506, "grad_norm": 1.4371369113238543, "learning_rate": 1.8786113267244873e-05, "loss": 0.8367, "step": 5073 }, { "epoch": 0.1837739949293734, "grad_norm": 1.5398983791343779, "learning_rate": 1.8785553022879583e-05, "loss": 0.8676, "step": 5074 }, { "epoch": 0.18381021369069178, "grad_norm": 1.4576410201413117, "learning_rate": 1.878499265761736e-05, "loss": 0.8172, "step": 5075 }, { "epoch": 0.18384643245201013, "grad_norm": 1.3007461579632613, "learning_rate": 1.8784432171465923e-05, "loss": 0.7578, "step": 5076 }, { "epoch": 0.1838826512133285, "grad_norm": 1.2468344070225355, "learning_rate": 1.878387156443298e-05, "loss": 0.7752, "step": 5077 }, { "epoch": 0.18391886997464688, "grad_norm": 1.5827138534865732, "learning_rate": 1.8783310836526246e-05, "loss": 0.891, "step": 5078 }, { "epoch": 0.18395508873596522, "grad_norm": 1.497904254734131, "learning_rate": 1.8782749987753437e-05, "loss": 0.7956, "step": 5079 }, { "epoch": 0.1839913074972836, "grad_norm": 1.656064679072594, "learning_rate": 1.878218901812227e-05, "loss": 0.8174, "step": 5080 }, { "epoch": 0.18402752625860194, "grad_norm": 1.4484506702647346, "learning_rate": 1.8781627927640465e-05, "loss": 0.8211, "step": 5081 }, { "epoch": 0.18406374501992032, "grad_norm": 1.4734449336348845, "learning_rate": 1.8781066716315742e-05, "loss": 0.8707, "step": 5082 }, { "epoch": 0.1840999637812387, "grad_norm": 1.4826736801755886, "learning_rate": 1.878050538415583e-05, "loss": 0.7848, "step": 5083 }, { "epoch": 0.18413618254255704, "grad_norm": 1.619327245662958, "learning_rate": 1.8779943931168446e-05, "loss": 0.803, "step": 5084 }, { "epoch": 0.1841724013038754, "grad_norm": 1.4210340553137968, "learning_rate": 1.877938235736132e-05, "loss": 0.8044, "step": 5085 }, { "epoch": 0.18420862006519376, "grad_norm": 1.530958259778151, "learning_rate": 1.8778820662742178e-05, "loss": 0.8161, "step": 5086 }, { "epoch": 0.18424483882651213, "grad_norm": 1.5696402303989345, "learning_rate": 1.8778258847318748e-05, "loss": 0.8342, "step": 5087 }, { "epoch": 0.1842810575878305, "grad_norm": 1.4541749796785526, "learning_rate": 1.8777696911098762e-05, "loss": 0.8041, "step": 5088 }, { "epoch": 0.18431727634914885, "grad_norm": 1.5019124550143452, "learning_rate": 1.8777134854089958e-05, "loss": 0.9014, "step": 5089 }, { "epoch": 0.18435349511046722, "grad_norm": 1.4910567415628073, "learning_rate": 1.8776572676300066e-05, "loss": 0.9039, "step": 5090 }, { "epoch": 0.1843897138717856, "grad_norm": 1.7410016460940205, "learning_rate": 1.8776010377736818e-05, "loss": 0.8844, "step": 5091 }, { "epoch": 0.18442593263310395, "grad_norm": 1.3751768606676575, "learning_rate": 1.877544795840796e-05, "loss": 0.7912, "step": 5092 }, { "epoch": 0.18446215139442232, "grad_norm": 1.4681092688345747, "learning_rate": 1.877488541832123e-05, "loss": 0.8178, "step": 5093 }, { "epoch": 0.18449837015574067, "grad_norm": 1.5229423610854522, "learning_rate": 1.877432275748436e-05, "loss": 0.8234, "step": 5094 }, { "epoch": 0.18453458891705904, "grad_norm": 1.8273138671466571, "learning_rate": 1.8773759975905098e-05, "loss": 0.8028, "step": 5095 }, { "epoch": 0.1845708076783774, "grad_norm": 1.3824708274111943, "learning_rate": 1.8773197073591195e-05, "loss": 0.8255, "step": 5096 }, { "epoch": 0.18460702643969576, "grad_norm": 1.5898144072559344, "learning_rate": 1.8772634050550388e-05, "loss": 0.905, "step": 5097 }, { "epoch": 0.18464324520101413, "grad_norm": 1.5588493711187708, "learning_rate": 1.8772070906790428e-05, "loss": 0.8188, "step": 5098 }, { "epoch": 0.18467946396233248, "grad_norm": 1.4708826156908008, "learning_rate": 1.8771507642319067e-05, "loss": 0.7124, "step": 5099 }, { "epoch": 0.18471568272365085, "grad_norm": 1.4132721349212425, "learning_rate": 1.877094425714405e-05, "loss": 0.8571, "step": 5100 }, { "epoch": 0.18475190148496923, "grad_norm": 1.5161469331406676, "learning_rate": 1.877038075127314e-05, "loss": 0.793, "step": 5101 }, { "epoch": 0.18478812024628757, "grad_norm": 1.294578421831969, "learning_rate": 1.876981712471408e-05, "loss": 0.697, "step": 5102 }, { "epoch": 0.18482433900760595, "grad_norm": 1.563323974311509, "learning_rate": 1.876925337747463e-05, "loss": 0.7731, "step": 5103 }, { "epoch": 0.1848605577689243, "grad_norm": 1.2612324584313153, "learning_rate": 1.8768689509562548e-05, "loss": 0.7752, "step": 5104 }, { "epoch": 0.18489677653024267, "grad_norm": 1.657729939545114, "learning_rate": 1.8768125520985594e-05, "loss": 0.8803, "step": 5105 }, { "epoch": 0.18493299529156104, "grad_norm": 1.5209818880694461, "learning_rate": 1.8767561411751528e-05, "loss": 0.8725, "step": 5106 }, { "epoch": 0.1849692140528794, "grad_norm": 1.5955822710942127, "learning_rate": 1.8766997181868115e-05, "loss": 0.8406, "step": 5107 }, { "epoch": 0.18500543281419776, "grad_norm": 1.5378950333574484, "learning_rate": 1.8766432831343117e-05, "loss": 0.8068, "step": 5108 }, { "epoch": 0.1850416515755161, "grad_norm": 1.2522054120856958, "learning_rate": 1.87658683601843e-05, "loss": 0.6989, "step": 5109 }, { "epoch": 0.18507787033683448, "grad_norm": 1.5685559205430506, "learning_rate": 1.876530376839943e-05, "loss": 0.8759, "step": 5110 }, { "epoch": 0.18511408909815286, "grad_norm": 1.5051005086563498, "learning_rate": 1.876473905599628e-05, "loss": 0.8252, "step": 5111 }, { "epoch": 0.1851503078594712, "grad_norm": 1.475671697531643, "learning_rate": 1.8764174222982617e-05, "loss": 0.7593, "step": 5112 }, { "epoch": 0.18518652662078958, "grad_norm": 1.4127018399261169, "learning_rate": 1.8763609269366217e-05, "loss": 0.8116, "step": 5113 }, { "epoch": 0.18522274538210792, "grad_norm": 1.511980613187001, "learning_rate": 1.876304419515486e-05, "loss": 0.7756, "step": 5114 }, { "epoch": 0.1852589641434263, "grad_norm": 1.3332319666788215, "learning_rate": 1.8762479000356307e-05, "loss": 0.7788, "step": 5115 }, { "epoch": 0.18529518290474467, "grad_norm": 1.7933088189960988, "learning_rate": 1.8761913684978346e-05, "loss": 0.7912, "step": 5116 }, { "epoch": 0.18533140166606302, "grad_norm": 1.6754283666862324, "learning_rate": 1.876134824902875e-05, "loss": 0.8027, "step": 5117 }, { "epoch": 0.1853676204273814, "grad_norm": 1.5483012337443545, "learning_rate": 1.876078269251531e-05, "loss": 0.9232, "step": 5118 }, { "epoch": 0.18540383918869974, "grad_norm": 1.5120631948025294, "learning_rate": 1.87602170154458e-05, "loss": 0.9032, "step": 5119 }, { "epoch": 0.1854400579500181, "grad_norm": 1.4239772789149892, "learning_rate": 1.8759651217828002e-05, "loss": 0.8391, "step": 5120 }, { "epoch": 0.18547627671133649, "grad_norm": 1.648084548479814, "learning_rate": 1.8759085299669713e-05, "loss": 0.8659, "step": 5121 }, { "epoch": 0.18551249547265483, "grad_norm": 1.3901209909067245, "learning_rate": 1.875851926097871e-05, "loss": 0.7123, "step": 5122 }, { "epoch": 0.1855487142339732, "grad_norm": 1.5433261741044686, "learning_rate": 1.8757953101762786e-05, "loss": 0.8433, "step": 5123 }, { "epoch": 0.18558493299529155, "grad_norm": 1.587909227991869, "learning_rate": 1.8757386822029733e-05, "loss": 0.7781, "step": 5124 }, { "epoch": 0.18562115175660993, "grad_norm": 1.3586966520963688, "learning_rate": 1.875682042178734e-05, "loss": 0.8439, "step": 5125 }, { "epoch": 0.1856573705179283, "grad_norm": 1.4773387502114645, "learning_rate": 1.875625390104341e-05, "loss": 0.8579, "step": 5126 }, { "epoch": 0.18569358927924665, "grad_norm": 1.4682700864491902, "learning_rate": 1.875568725980573e-05, "loss": 0.8035, "step": 5127 }, { "epoch": 0.18572980804056502, "grad_norm": 1.0287606409644823, "learning_rate": 1.87551204980821e-05, "loss": 0.7894, "step": 5128 }, { "epoch": 0.18576602680188337, "grad_norm": 1.3617303170682973, "learning_rate": 1.8754553615880312e-05, "loss": 0.7972, "step": 5129 }, { "epoch": 0.18580224556320174, "grad_norm": 1.6784179303078302, "learning_rate": 1.8753986613208183e-05, "loss": 0.8314, "step": 5130 }, { "epoch": 0.18583846432452011, "grad_norm": 1.4666653941182375, "learning_rate": 1.87534194900735e-05, "loss": 0.6933, "step": 5131 }, { "epoch": 0.18587468308583846, "grad_norm": 1.5407116862785004, "learning_rate": 1.8752852246484077e-05, "loss": 0.8918, "step": 5132 }, { "epoch": 0.18591090184715683, "grad_norm": 1.420768058913817, "learning_rate": 1.8752284882447713e-05, "loss": 0.8154, "step": 5133 }, { "epoch": 0.18594712060847518, "grad_norm": 1.3071619901720304, "learning_rate": 1.8751717397972222e-05, "loss": 0.8296, "step": 5134 }, { "epoch": 0.18598333936979355, "grad_norm": 1.5937756391407694, "learning_rate": 1.875114979306541e-05, "loss": 0.8606, "step": 5135 }, { "epoch": 0.18601955813111193, "grad_norm": 1.5935753762891238, "learning_rate": 1.8750582067735084e-05, "loss": 0.8847, "step": 5136 }, { "epoch": 0.18605577689243027, "grad_norm": 1.2560707923993804, "learning_rate": 1.875001422198906e-05, "loss": 0.7629, "step": 5137 }, { "epoch": 0.18609199565374865, "grad_norm": 1.4161387781323684, "learning_rate": 1.8749446255835154e-05, "loss": 0.8531, "step": 5138 }, { "epoch": 0.186128214415067, "grad_norm": 1.621497763886827, "learning_rate": 1.874887816928118e-05, "loss": 0.762, "step": 5139 }, { "epoch": 0.18616443317638537, "grad_norm": 1.6015790240668146, "learning_rate": 1.874830996233495e-05, "loss": 0.8271, "step": 5140 }, { "epoch": 0.18620065193770374, "grad_norm": 1.7525160692008146, "learning_rate": 1.8747741635004296e-05, "loss": 0.8464, "step": 5141 }, { "epoch": 0.1862368706990221, "grad_norm": 1.4324428397652975, "learning_rate": 1.8747173187297025e-05, "loss": 0.7262, "step": 5142 }, { "epoch": 0.18627308946034046, "grad_norm": 1.5873168451757265, "learning_rate": 1.8746604619220968e-05, "loss": 0.8822, "step": 5143 }, { "epoch": 0.1863093082216588, "grad_norm": 1.26544995333645, "learning_rate": 1.8746035930783942e-05, "loss": 0.7847, "step": 5144 }, { "epoch": 0.18634552698297718, "grad_norm": 1.4447709821798622, "learning_rate": 1.874546712199378e-05, "loss": 0.8188, "step": 5145 }, { "epoch": 0.18638174574429556, "grad_norm": 1.7512671255619419, "learning_rate": 1.8744898192858305e-05, "loss": 0.7716, "step": 5146 }, { "epoch": 0.1864179645056139, "grad_norm": 1.5756971257009655, "learning_rate": 1.8744329143385346e-05, "loss": 0.794, "step": 5147 }, { "epoch": 0.18645418326693228, "grad_norm": 1.356789715737942, "learning_rate": 1.8743759973582735e-05, "loss": 0.7651, "step": 5148 }, { "epoch": 0.18649040202825062, "grad_norm": 1.5032224562094727, "learning_rate": 1.8743190683458308e-05, "loss": 0.785, "step": 5149 }, { "epoch": 0.186526620789569, "grad_norm": 1.666693718263514, "learning_rate": 1.874262127301989e-05, "loss": 0.7864, "step": 5150 }, { "epoch": 0.18656283955088737, "grad_norm": 1.393458831580628, "learning_rate": 1.874205174227532e-05, "loss": 0.821, "step": 5151 }, { "epoch": 0.18659905831220572, "grad_norm": 1.5876648943034737, "learning_rate": 1.8741482091232442e-05, "loss": 0.8028, "step": 5152 }, { "epoch": 0.1866352770735241, "grad_norm": 1.3831984796543, "learning_rate": 1.8740912319899084e-05, "loss": 0.7924, "step": 5153 }, { "epoch": 0.18667149583484244, "grad_norm": 1.4284047472041914, "learning_rate": 1.8740342428283096e-05, "loss": 0.7794, "step": 5154 }, { "epoch": 0.1867077145961608, "grad_norm": 1.3447759860488877, "learning_rate": 1.8739772416392313e-05, "loss": 0.8615, "step": 5155 }, { "epoch": 0.18674393335747919, "grad_norm": 1.5659268744429258, "learning_rate": 1.8739202284234584e-05, "loss": 0.9226, "step": 5156 }, { "epoch": 0.18678015211879753, "grad_norm": 16.988709163473544, "learning_rate": 1.8738632031817754e-05, "loss": 0.8529, "step": 5157 }, { "epoch": 0.1868163708801159, "grad_norm": 1.266745459699472, "learning_rate": 1.873806165914967e-05, "loss": 0.7868, "step": 5158 }, { "epoch": 0.18685258964143425, "grad_norm": 1.6413524995860815, "learning_rate": 1.8737491166238175e-05, "loss": 0.8082, "step": 5159 }, { "epoch": 0.18688880840275263, "grad_norm": 2.311287632217278, "learning_rate": 1.8736920553091127e-05, "loss": 0.7873, "step": 5160 }, { "epoch": 0.186925027164071, "grad_norm": 5.692592071859553, "learning_rate": 1.873634981971637e-05, "loss": 0.9254, "step": 5161 }, { "epoch": 0.18696124592538935, "grad_norm": 1.5547660671203025, "learning_rate": 1.8735778966121772e-05, "loss": 0.8281, "step": 5162 }, { "epoch": 0.18699746468670772, "grad_norm": 1.4401996273652906, "learning_rate": 1.8735207992315173e-05, "loss": 0.8458, "step": 5163 }, { "epoch": 0.18703368344802607, "grad_norm": 1.8165186139699718, "learning_rate": 1.8734636898304438e-05, "loss": 0.8614, "step": 5164 }, { "epoch": 0.18706990220934444, "grad_norm": 1.7530609991519202, "learning_rate": 1.8734065684097427e-05, "loss": 0.764, "step": 5165 }, { "epoch": 0.18710612097066281, "grad_norm": 1.6484276310232073, "learning_rate": 1.8733494349701994e-05, "loss": 0.8335, "step": 5166 }, { "epoch": 0.18714233973198116, "grad_norm": 1.7331449530178076, "learning_rate": 1.8732922895126006e-05, "loss": 0.8481, "step": 5167 }, { "epoch": 0.18717855849329953, "grad_norm": 1.5357285917772605, "learning_rate": 1.8732351320377327e-05, "loss": 0.9027, "step": 5168 }, { "epoch": 0.18721477725461788, "grad_norm": 1.2898151981710297, "learning_rate": 1.873177962546382e-05, "loss": 0.7675, "step": 5169 }, { "epoch": 0.18725099601593626, "grad_norm": 1.4415018675395421, "learning_rate": 1.873120781039335e-05, "loss": 0.8243, "step": 5170 }, { "epoch": 0.18728721477725463, "grad_norm": 1.4239746837251654, "learning_rate": 1.8730635875173796e-05, "loss": 0.8276, "step": 5171 }, { "epoch": 0.18732343353857298, "grad_norm": 1.4936274850353615, "learning_rate": 1.8730063819813016e-05, "loss": 0.7938, "step": 5172 }, { "epoch": 0.18735965229989135, "grad_norm": 1.7682821652574454, "learning_rate": 1.8729491644318887e-05, "loss": 0.8344, "step": 5173 }, { "epoch": 0.1873958710612097, "grad_norm": 1.3568736917973945, "learning_rate": 1.8728919348699285e-05, "loss": 0.7753, "step": 5174 }, { "epoch": 0.18743208982252807, "grad_norm": 1.470092745562937, "learning_rate": 1.872834693296208e-05, "loss": 0.8313, "step": 5175 }, { "epoch": 0.18746830858384644, "grad_norm": 1.5131337624463173, "learning_rate": 1.872777439711515e-05, "loss": 0.8523, "step": 5176 }, { "epoch": 0.1875045273451648, "grad_norm": 1.4742850837581667, "learning_rate": 1.8727201741166377e-05, "loss": 0.7651, "step": 5177 }, { "epoch": 0.18754074610648316, "grad_norm": 1.7065372079470955, "learning_rate": 1.8726628965123643e-05, "loss": 0.7693, "step": 5178 }, { "epoch": 0.1875769648678015, "grad_norm": 1.6867851635171804, "learning_rate": 1.8726056068994825e-05, "loss": 0.8799, "step": 5179 }, { "epoch": 0.18761318362911988, "grad_norm": 1.5355722261760731, "learning_rate": 1.8725483052787807e-05, "loss": 0.8907, "step": 5180 }, { "epoch": 0.18764940239043826, "grad_norm": 1.6987973995241439, "learning_rate": 1.8724909916510478e-05, "loss": 0.9383, "step": 5181 }, { "epoch": 0.1876856211517566, "grad_norm": 1.5216796086208133, "learning_rate": 1.8724336660170718e-05, "loss": 0.7558, "step": 5182 }, { "epoch": 0.18772183991307498, "grad_norm": 1.574742428877072, "learning_rate": 1.872376328377642e-05, "loss": 0.884, "step": 5183 }, { "epoch": 0.18775805867439332, "grad_norm": 1.6692311671022901, "learning_rate": 1.8723189787335475e-05, "loss": 0.7817, "step": 5184 }, { "epoch": 0.1877942774357117, "grad_norm": 1.4615070124162903, "learning_rate": 1.8722616170855774e-05, "loss": 0.76, "step": 5185 }, { "epoch": 0.18783049619703007, "grad_norm": 1.6435392212437523, "learning_rate": 1.872204243434521e-05, "loss": 0.8121, "step": 5186 }, { "epoch": 0.18786671495834842, "grad_norm": 1.382935783176259, "learning_rate": 1.872146857781168e-05, "loss": 0.7732, "step": 5187 }, { "epoch": 0.1879029337196668, "grad_norm": 1.2756863316238856, "learning_rate": 1.8720894601263077e-05, "loss": 0.8422, "step": 5188 }, { "epoch": 0.18793915248098514, "grad_norm": 1.8106495993730338, "learning_rate": 1.87203205047073e-05, "loss": 0.8888, "step": 5189 }, { "epoch": 0.1879753712423035, "grad_norm": 1.6789646844559951, "learning_rate": 1.8719746288152252e-05, "loss": 0.8645, "step": 5190 }, { "epoch": 0.1880115900036219, "grad_norm": 1.2821381495573545, "learning_rate": 1.8719171951605835e-05, "loss": 0.7629, "step": 5191 }, { "epoch": 0.18804780876494023, "grad_norm": 2.01672963989538, "learning_rate": 1.871859749507595e-05, "loss": 0.9259, "step": 5192 }, { "epoch": 0.1880840275262586, "grad_norm": 1.603241403924017, "learning_rate": 1.87180229185705e-05, "loss": 0.8418, "step": 5193 }, { "epoch": 0.18812024628757695, "grad_norm": 1.3849568764566926, "learning_rate": 1.8717448222097395e-05, "loss": 0.8541, "step": 5194 }, { "epoch": 0.18815646504889533, "grad_norm": 1.63009338664286, "learning_rate": 1.8716873405664546e-05, "loss": 0.8985, "step": 5195 }, { "epoch": 0.1881926838102137, "grad_norm": 1.5964284290491069, "learning_rate": 1.8716298469279858e-05, "loss": 0.8355, "step": 5196 }, { "epoch": 0.18822890257153205, "grad_norm": 1.3922471303721098, "learning_rate": 1.8715723412951243e-05, "loss": 0.763, "step": 5197 }, { "epoch": 0.18826512133285042, "grad_norm": 1.5088048508197658, "learning_rate": 1.8715148236686617e-05, "loss": 0.7883, "step": 5198 }, { "epoch": 0.18830134009416877, "grad_norm": 1.557101993800723, "learning_rate": 1.8714572940493894e-05, "loss": 0.7403, "step": 5199 }, { "epoch": 0.18833755885548714, "grad_norm": 1.5050373981994047, "learning_rate": 1.8713997524380987e-05, "loss": 0.8502, "step": 5200 }, { "epoch": 0.18837377761680552, "grad_norm": 1.8353852760439318, "learning_rate": 1.8713421988355817e-05, "loss": 0.9274, "step": 5201 }, { "epoch": 0.18840999637812386, "grad_norm": 1.2810384920877358, "learning_rate": 1.8712846332426307e-05, "loss": 0.7723, "step": 5202 }, { "epoch": 0.18844621513944224, "grad_norm": 1.517858858543316, "learning_rate": 1.8712270556600374e-05, "loss": 0.8644, "step": 5203 }, { "epoch": 0.18848243390076058, "grad_norm": 1.5418987046521775, "learning_rate": 1.8711694660885947e-05, "loss": 0.8303, "step": 5204 }, { "epoch": 0.18851865266207896, "grad_norm": 1.4894364790757872, "learning_rate": 1.8711118645290943e-05, "loss": 0.8654, "step": 5205 }, { "epoch": 0.18855487142339733, "grad_norm": 1.6134285513200988, "learning_rate": 1.871054250982329e-05, "loss": 0.8704, "step": 5206 }, { "epoch": 0.18859109018471568, "grad_norm": 1.1903951890639128, "learning_rate": 1.8709966254490923e-05, "loss": 0.8394, "step": 5207 }, { "epoch": 0.18862730894603405, "grad_norm": 1.5403905284843058, "learning_rate": 1.8709389879301766e-05, "loss": 0.8192, "step": 5208 }, { "epoch": 0.1886635277073524, "grad_norm": 1.6556544965238393, "learning_rate": 1.870881338426375e-05, "loss": 0.797, "step": 5209 }, { "epoch": 0.18869974646867077, "grad_norm": 1.3966950504405986, "learning_rate": 1.870823676938481e-05, "loss": 0.8152, "step": 5210 }, { "epoch": 0.18873596522998914, "grad_norm": 1.4872666583894478, "learning_rate": 1.870766003467288e-05, "loss": 0.7388, "step": 5211 }, { "epoch": 0.1887721839913075, "grad_norm": 1.5602543229012218, "learning_rate": 1.8707083180135898e-05, "loss": 0.8228, "step": 5212 }, { "epoch": 0.18880840275262586, "grad_norm": 1.6248502000321565, "learning_rate": 1.87065062057818e-05, "loss": 0.9188, "step": 5213 }, { "epoch": 0.1888446215139442, "grad_norm": 1.5189393552918642, "learning_rate": 1.8705929111618525e-05, "loss": 0.7893, "step": 5214 }, { "epoch": 0.18888084027526258, "grad_norm": 1.436697480623541, "learning_rate": 1.8705351897654016e-05, "loss": 0.7368, "step": 5215 }, { "epoch": 0.18891705903658096, "grad_norm": 1.6086422666502196, "learning_rate": 1.8704774563896216e-05, "loss": 0.7872, "step": 5216 }, { "epoch": 0.1889532777978993, "grad_norm": 1.431987073906823, "learning_rate": 1.8704197110353073e-05, "loss": 0.7952, "step": 5217 }, { "epoch": 0.18898949655921768, "grad_norm": 1.2871708727711915, "learning_rate": 1.8703619537032526e-05, "loss": 0.8429, "step": 5218 }, { "epoch": 0.18902571532053603, "grad_norm": 1.514051082856403, "learning_rate": 1.8703041843942526e-05, "loss": 0.7991, "step": 5219 }, { "epoch": 0.1890619340818544, "grad_norm": 1.6254032854184408, "learning_rate": 1.8702464031091026e-05, "loss": 0.7972, "step": 5220 }, { "epoch": 0.18909815284317277, "grad_norm": 1.6184840349290002, "learning_rate": 1.870188609848597e-05, "loss": 0.8839, "step": 5221 }, { "epoch": 0.18913437160449112, "grad_norm": 1.4448885527865472, "learning_rate": 1.870130804613532e-05, "loss": 0.8308, "step": 5222 }, { "epoch": 0.1891705903658095, "grad_norm": 1.3798285910339902, "learning_rate": 1.870072987404702e-05, "loss": 0.7595, "step": 5223 }, { "epoch": 0.18920680912712784, "grad_norm": 1.2551688510243721, "learning_rate": 1.8700151582229035e-05, "loss": 0.8211, "step": 5224 }, { "epoch": 0.1892430278884462, "grad_norm": 1.4033072224791598, "learning_rate": 1.8699573170689322e-05, "loss": 0.8362, "step": 5225 }, { "epoch": 0.1892792466497646, "grad_norm": 1.2541600586769337, "learning_rate": 1.8698994639435833e-05, "loss": 0.8381, "step": 5226 }, { "epoch": 0.18931546541108293, "grad_norm": 1.66782170123242, "learning_rate": 1.8698415988476538e-05, "loss": 0.8622, "step": 5227 }, { "epoch": 0.1893516841724013, "grad_norm": 1.4373339992540375, "learning_rate": 1.8697837217819398e-05, "loss": 0.8606, "step": 5228 }, { "epoch": 0.18938790293371965, "grad_norm": 1.6087147135966156, "learning_rate": 1.869725832747237e-05, "loss": 0.8853, "step": 5229 }, { "epoch": 0.18942412169503803, "grad_norm": 1.5097484390247877, "learning_rate": 1.8696679317443428e-05, "loss": 0.7709, "step": 5230 }, { "epoch": 0.1894603404563564, "grad_norm": 1.453239917056612, "learning_rate": 1.8696100187740535e-05, "loss": 0.8255, "step": 5231 }, { "epoch": 0.18949655921767475, "grad_norm": 1.6702228721038053, "learning_rate": 1.8695520938371667e-05, "loss": 0.7836, "step": 5232 }, { "epoch": 0.18953277797899312, "grad_norm": 1.7214895414963471, "learning_rate": 1.8694941569344785e-05, "loss": 0.8287, "step": 5233 }, { "epoch": 0.18956899674031147, "grad_norm": 1.3030054832256646, "learning_rate": 1.869436208066787e-05, "loss": 0.7563, "step": 5234 }, { "epoch": 0.18960521550162984, "grad_norm": 1.7901000425083873, "learning_rate": 1.869378247234889e-05, "loss": 0.9475, "step": 5235 }, { "epoch": 0.18964143426294822, "grad_norm": 1.6254680869491562, "learning_rate": 1.869320274439583e-05, "loss": 0.8466, "step": 5236 }, { "epoch": 0.18967765302426656, "grad_norm": 1.4530416068635812, "learning_rate": 1.8692622896816653e-05, "loss": 0.7915, "step": 5237 }, { "epoch": 0.18971387178558494, "grad_norm": 1.578153432929856, "learning_rate": 1.8692042929619352e-05, "loss": 0.784, "step": 5238 }, { "epoch": 0.18975009054690328, "grad_norm": 1.606640348218787, "learning_rate": 1.8691462842811906e-05, "loss": 0.8769, "step": 5239 }, { "epoch": 0.18978630930822166, "grad_norm": 1.4487162153149065, "learning_rate": 1.869088263640229e-05, "loss": 0.7968, "step": 5240 }, { "epoch": 0.18982252806954003, "grad_norm": 1.6005781346911192, "learning_rate": 1.8690302310398495e-05, "loss": 0.7876, "step": 5241 }, { "epoch": 0.18985874683085838, "grad_norm": 2.031071295579985, "learning_rate": 1.8689721864808502e-05, "loss": 0.7909, "step": 5242 }, { "epoch": 0.18989496559217675, "grad_norm": 1.3613048459803272, "learning_rate": 1.86891412996403e-05, "loss": 0.8389, "step": 5243 }, { "epoch": 0.1899311843534951, "grad_norm": 1.4276230465100366, "learning_rate": 1.8688560614901884e-05, "loss": 0.7552, "step": 5244 }, { "epoch": 0.18996740311481347, "grad_norm": 1.3563068201957305, "learning_rate": 1.8687979810601234e-05, "loss": 0.7841, "step": 5245 }, { "epoch": 0.19000362187613185, "grad_norm": 1.394129660561398, "learning_rate": 1.868739888674635e-05, "loss": 0.7307, "step": 5246 }, { "epoch": 0.1900398406374502, "grad_norm": 1.2029985786796709, "learning_rate": 1.8686817843345225e-05, "loss": 0.8068, "step": 5247 }, { "epoch": 0.19007605939876857, "grad_norm": 1.5871933242392815, "learning_rate": 1.8686236680405848e-05, "loss": 0.8902, "step": 5248 }, { "epoch": 0.1901122781600869, "grad_norm": 1.4962102966629902, "learning_rate": 1.8685655397936228e-05, "loss": 0.8709, "step": 5249 }, { "epoch": 0.19014849692140529, "grad_norm": 1.3071959304333953, "learning_rate": 1.8685073995944356e-05, "loss": 0.7648, "step": 5250 }, { "epoch": 0.19018471568272366, "grad_norm": 1.6154694899234392, "learning_rate": 1.8684492474438233e-05, "loss": 0.8344, "step": 5251 }, { "epoch": 0.190220934444042, "grad_norm": 1.5510379677263642, "learning_rate": 1.8683910833425868e-05, "loss": 0.8664, "step": 5252 }, { "epoch": 0.19025715320536038, "grad_norm": 1.4914974869146798, "learning_rate": 1.8683329072915253e-05, "loss": 0.8047, "step": 5253 }, { "epoch": 0.19029337196667873, "grad_norm": 1.3440550745283262, "learning_rate": 1.86827471929144e-05, "loss": 0.7869, "step": 5254 }, { "epoch": 0.1903295907279971, "grad_norm": 1.5233433718823584, "learning_rate": 1.868216519343132e-05, "loss": 0.7789, "step": 5255 }, { "epoch": 0.19036580948931547, "grad_norm": 1.4815531875572747, "learning_rate": 1.8681583074474018e-05, "loss": 0.8193, "step": 5256 }, { "epoch": 0.19040202825063382, "grad_norm": 1.2628794178722071, "learning_rate": 1.8681000836050504e-05, "loss": 0.8074, "step": 5257 }, { "epoch": 0.1904382470119522, "grad_norm": 1.5920438726967263, "learning_rate": 1.868041847816879e-05, "loss": 0.8987, "step": 5258 }, { "epoch": 0.19047446577327054, "grad_norm": 1.3974235111132518, "learning_rate": 1.8679836000836894e-05, "loss": 0.7412, "step": 5259 }, { "epoch": 0.19051068453458891, "grad_norm": 1.5263290510337353, "learning_rate": 1.867925340406282e-05, "loss": 0.8899, "step": 5260 }, { "epoch": 0.1905469032959073, "grad_norm": 1.470911209487626, "learning_rate": 1.86786706878546e-05, "loss": 0.763, "step": 5261 }, { "epoch": 0.19058312205722563, "grad_norm": 1.4728863989091, "learning_rate": 1.8678087852220242e-05, "loss": 0.7616, "step": 5262 }, { "epoch": 0.190619340818544, "grad_norm": 1.4322341219584065, "learning_rate": 1.867750489716777e-05, "loss": 0.8004, "step": 5263 }, { "epoch": 0.19065555957986238, "grad_norm": 1.6118846429330789, "learning_rate": 1.867692182270521e-05, "loss": 0.818, "step": 5264 }, { "epoch": 0.19069177834118073, "grad_norm": 1.421943383223047, "learning_rate": 1.8676338628840577e-05, "loss": 0.7924, "step": 5265 }, { "epoch": 0.1907279971024991, "grad_norm": 1.581662369331544, "learning_rate": 1.8675755315581903e-05, "loss": 0.8949, "step": 5266 }, { "epoch": 0.19076421586381745, "grad_norm": 1.3914299628214637, "learning_rate": 1.867517188293721e-05, "loss": 0.7458, "step": 5267 }, { "epoch": 0.19080043462513582, "grad_norm": 1.3609814144198145, "learning_rate": 1.8674588330914535e-05, "loss": 0.7871, "step": 5268 }, { "epoch": 0.1908366533864542, "grad_norm": 1.5060968310782417, "learning_rate": 1.8674004659521895e-05, "loss": 0.847, "step": 5269 }, { "epoch": 0.19087287214777254, "grad_norm": 1.6614888665752963, "learning_rate": 1.8673420868767333e-05, "loss": 0.8647, "step": 5270 }, { "epoch": 0.19090909090909092, "grad_norm": 1.3760924896235762, "learning_rate": 1.867283695865888e-05, "loss": 0.8141, "step": 5271 }, { "epoch": 0.19094530967040926, "grad_norm": 1.524380499637094, "learning_rate": 1.867225292920457e-05, "loss": 0.8415, "step": 5272 }, { "epoch": 0.19098152843172764, "grad_norm": 1.4300109752695684, "learning_rate": 1.8671668780412436e-05, "loss": 0.7031, "step": 5273 }, { "epoch": 0.191017747193046, "grad_norm": 1.2745152215435995, "learning_rate": 1.8671084512290522e-05, "loss": 0.8264, "step": 5274 }, { "epoch": 0.19105396595436436, "grad_norm": 1.1198204156854776, "learning_rate": 1.8670500124846866e-05, "loss": 0.7347, "step": 5275 }, { "epoch": 0.19109018471568273, "grad_norm": 1.5653093580902517, "learning_rate": 1.866991561808951e-05, "loss": 0.866, "step": 5276 }, { "epoch": 0.19112640347700108, "grad_norm": 1.419545639826712, "learning_rate": 1.86693309920265e-05, "loss": 0.796, "step": 5277 }, { "epoch": 0.19116262223831945, "grad_norm": 1.5842201478419395, "learning_rate": 1.8668746246665873e-05, "loss": 0.898, "step": 5278 }, { "epoch": 0.19119884099963783, "grad_norm": 1.1434725958761516, "learning_rate": 1.8668161382015683e-05, "loss": 0.7881, "step": 5279 }, { "epoch": 0.19123505976095617, "grad_norm": 1.5019726187702183, "learning_rate": 1.8667576398083972e-05, "loss": 0.808, "step": 5280 }, { "epoch": 0.19127127852227455, "grad_norm": 1.5203821146219991, "learning_rate": 1.8666991294878798e-05, "loss": 0.8481, "step": 5281 }, { "epoch": 0.1913074972835929, "grad_norm": 1.6488307268049422, "learning_rate": 1.8666406072408206e-05, "loss": 0.8217, "step": 5282 }, { "epoch": 0.19134371604491127, "grad_norm": 1.5640268387954372, "learning_rate": 1.8665820730680254e-05, "loss": 0.8544, "step": 5283 }, { "epoch": 0.19137993480622964, "grad_norm": 1.4594857281341647, "learning_rate": 1.8665235269702993e-05, "loss": 0.8231, "step": 5284 }, { "epoch": 0.191416153567548, "grad_norm": 1.5577119589168995, "learning_rate": 1.866464968948448e-05, "loss": 0.8012, "step": 5285 }, { "epoch": 0.19145237232886636, "grad_norm": 1.5832335977659424, "learning_rate": 1.8664063990032775e-05, "loss": 0.7602, "step": 5286 }, { "epoch": 0.1914885910901847, "grad_norm": 1.498662579620016, "learning_rate": 1.866347817135594e-05, "loss": 0.8296, "step": 5287 }, { "epoch": 0.19152480985150308, "grad_norm": 1.4548361717908007, "learning_rate": 1.8662892233462028e-05, "loss": 0.8143, "step": 5288 }, { "epoch": 0.19156102861282145, "grad_norm": 1.4862038225787748, "learning_rate": 1.8662306176359103e-05, "loss": 0.862, "step": 5289 }, { "epoch": 0.1915972473741398, "grad_norm": 1.4867498411113877, "learning_rate": 1.8661720000055243e-05, "loss": 0.8133, "step": 5290 }, { "epoch": 0.19163346613545817, "grad_norm": 1.5171964354404186, "learning_rate": 1.86611337045585e-05, "loss": 0.7896, "step": 5291 }, { "epoch": 0.19166968489677652, "grad_norm": 1.2696730920983073, "learning_rate": 1.8660547289876945e-05, "loss": 0.72, "step": 5292 }, { "epoch": 0.1917059036580949, "grad_norm": 1.6832172510506465, "learning_rate": 1.8659960756018652e-05, "loss": 0.797, "step": 5293 }, { "epoch": 0.19174212241941327, "grad_norm": 1.652985861196855, "learning_rate": 1.8659374102991693e-05, "loss": 0.8834, "step": 5294 }, { "epoch": 0.19177834118073162, "grad_norm": 1.50260687962248, "learning_rate": 1.8658787330804133e-05, "loss": 0.809, "step": 5295 }, { "epoch": 0.19181455994205, "grad_norm": 1.6644144290525358, "learning_rate": 1.865820043946405e-05, "loss": 0.8404, "step": 5296 }, { "epoch": 0.19185077870336834, "grad_norm": 1.6424275010546108, "learning_rate": 1.8657613428979526e-05, "loss": 0.9155, "step": 5297 }, { "epoch": 0.1918869974646867, "grad_norm": 1.6029653313523446, "learning_rate": 1.865702629935863e-05, "loss": 0.8754, "step": 5298 }, { "epoch": 0.19192321622600508, "grad_norm": 1.6059360438988106, "learning_rate": 1.865643905060945e-05, "loss": 0.9072, "step": 5299 }, { "epoch": 0.19195943498732343, "grad_norm": 1.5932206713538397, "learning_rate": 1.865585168274006e-05, "loss": 0.7999, "step": 5300 }, { "epoch": 0.1919956537486418, "grad_norm": 8.064463848458884, "learning_rate": 1.8655264195758542e-05, "loss": 0.8418, "step": 5301 }, { "epoch": 0.19203187250996015, "grad_norm": 1.4814583797293444, "learning_rate": 1.8654676589672988e-05, "loss": 0.8421, "step": 5302 }, { "epoch": 0.19206809127127852, "grad_norm": 1.2552914673016549, "learning_rate": 1.8654088864491474e-05, "loss": 0.8025, "step": 5303 }, { "epoch": 0.1921043100325969, "grad_norm": 1.2457042235398463, "learning_rate": 1.8653501020222098e-05, "loss": 0.7619, "step": 5304 }, { "epoch": 0.19214052879391524, "grad_norm": 1.4071576772027004, "learning_rate": 1.8652913056872944e-05, "loss": 0.8492, "step": 5305 }, { "epoch": 0.19217674755523362, "grad_norm": 1.5213116081292068, "learning_rate": 1.86523249744521e-05, "loss": 0.7966, "step": 5306 }, { "epoch": 0.19221296631655196, "grad_norm": 1.2327303279374748, "learning_rate": 1.8651736772967663e-05, "loss": 0.7741, "step": 5307 }, { "epoch": 0.19224918507787034, "grad_norm": 1.6777664206906016, "learning_rate": 1.865114845242773e-05, "loss": 0.7829, "step": 5308 }, { "epoch": 0.1922854038391887, "grad_norm": 1.2838740923468048, "learning_rate": 1.8650560012840387e-05, "loss": 0.8306, "step": 5309 }, { "epoch": 0.19232162260050706, "grad_norm": 1.2539470729001954, "learning_rate": 1.864997145421374e-05, "loss": 0.8162, "step": 5310 }, { "epoch": 0.19235784136182543, "grad_norm": 1.6655723556748152, "learning_rate": 1.8649382776555885e-05, "loss": 0.9157, "step": 5311 }, { "epoch": 0.19239406012314378, "grad_norm": 1.5638684038481705, "learning_rate": 1.8648793979874924e-05, "loss": 0.8597, "step": 5312 }, { "epoch": 0.19243027888446215, "grad_norm": 1.6293812740656546, "learning_rate": 1.8648205064178957e-05, "loss": 0.8283, "step": 5313 }, { "epoch": 0.19246649764578053, "grad_norm": 1.470141893931583, "learning_rate": 1.864761602947609e-05, "loss": 0.8497, "step": 5314 }, { "epoch": 0.19250271640709887, "grad_norm": 1.6882995283204463, "learning_rate": 1.8647026875774426e-05, "loss": 0.8948, "step": 5315 }, { "epoch": 0.19253893516841725, "grad_norm": 1.44583910292547, "learning_rate": 1.8646437603082078e-05, "loss": 0.7559, "step": 5316 }, { "epoch": 0.1925751539297356, "grad_norm": 1.541936020555985, "learning_rate": 1.864584821140715e-05, "loss": 0.9515, "step": 5317 }, { "epoch": 0.19261137269105397, "grad_norm": 1.2582558022306647, "learning_rate": 1.864525870075775e-05, "loss": 0.7945, "step": 5318 }, { "epoch": 0.19264759145237234, "grad_norm": 1.5669976828588252, "learning_rate": 1.8644669071141995e-05, "loss": 0.8699, "step": 5319 }, { "epoch": 0.1926838102136907, "grad_norm": 1.2633268046492412, "learning_rate": 1.8644079322568004e-05, "loss": 0.8387, "step": 5320 }, { "epoch": 0.19272002897500906, "grad_norm": 1.5560827368354109, "learning_rate": 1.864348945504388e-05, "loss": 0.7684, "step": 5321 }, { "epoch": 0.1927562477363274, "grad_norm": 1.4765623885960764, "learning_rate": 1.8642899468577747e-05, "loss": 0.7625, "step": 5322 }, { "epoch": 0.19279246649764578, "grad_norm": 1.7602671608181135, "learning_rate": 1.8642309363177723e-05, "loss": 0.8428, "step": 5323 }, { "epoch": 0.19282868525896416, "grad_norm": 1.3220917036722761, "learning_rate": 1.864171913885193e-05, "loss": 0.7647, "step": 5324 }, { "epoch": 0.1928649040202825, "grad_norm": 1.5254334420558002, "learning_rate": 1.8641128795608486e-05, "loss": 0.8805, "step": 5325 }, { "epoch": 0.19290112278160088, "grad_norm": 1.6138781096942798, "learning_rate": 1.864053833345552e-05, "loss": 0.8667, "step": 5326 }, { "epoch": 0.19293734154291922, "grad_norm": 1.4956216783233611, "learning_rate": 1.8639947752401152e-05, "loss": 0.8405, "step": 5327 }, { "epoch": 0.1929735603042376, "grad_norm": 1.3316816202513169, "learning_rate": 1.8639357052453514e-05, "loss": 0.8109, "step": 5328 }, { "epoch": 0.19300977906555597, "grad_norm": 1.1889951907677412, "learning_rate": 1.8638766233620725e-05, "loss": 0.7606, "step": 5329 }, { "epoch": 0.19304599782687432, "grad_norm": 1.6258242721951166, "learning_rate": 1.863817529591093e-05, "loss": 0.8662, "step": 5330 }, { "epoch": 0.1930822165881927, "grad_norm": 1.2381656073115985, "learning_rate": 1.8637584239332248e-05, "loss": 0.7494, "step": 5331 }, { "epoch": 0.19311843534951104, "grad_norm": 1.5711747976119146, "learning_rate": 1.8636993063892822e-05, "loss": 0.7637, "step": 5332 }, { "epoch": 0.1931546541108294, "grad_norm": 1.7022149253023915, "learning_rate": 1.8636401769600776e-05, "loss": 0.8668, "step": 5333 }, { "epoch": 0.19319087287214778, "grad_norm": 1.3963826169106397, "learning_rate": 1.8635810356464254e-05, "loss": 0.8326, "step": 5334 }, { "epoch": 0.19322709163346613, "grad_norm": 1.294902169986037, "learning_rate": 1.8635218824491398e-05, "loss": 0.8227, "step": 5335 }, { "epoch": 0.1932633103947845, "grad_norm": 1.4765103276688354, "learning_rate": 1.863462717369034e-05, "loss": 0.8509, "step": 5336 }, { "epoch": 0.19329952915610285, "grad_norm": 1.684860296306996, "learning_rate": 1.8634035404069225e-05, "loss": 0.815, "step": 5337 }, { "epoch": 0.19333574791742122, "grad_norm": 1.2988635159184592, "learning_rate": 1.8633443515636197e-05, "loss": 0.7899, "step": 5338 }, { "epoch": 0.1933719666787396, "grad_norm": 1.316981352708203, "learning_rate": 1.8632851508399402e-05, "loss": 0.8376, "step": 5339 }, { "epoch": 0.19340818544005794, "grad_norm": 1.6088337474201075, "learning_rate": 1.8632259382366985e-05, "loss": 0.8257, "step": 5340 }, { "epoch": 0.19344440420137632, "grad_norm": 1.4774104898841234, "learning_rate": 1.8631667137547088e-05, "loss": 0.7634, "step": 5341 }, { "epoch": 0.19348062296269466, "grad_norm": 1.4747280881477378, "learning_rate": 1.8631074773947872e-05, "loss": 0.8301, "step": 5342 }, { "epoch": 0.19351684172401304, "grad_norm": 1.4306502776083976, "learning_rate": 1.8630482291577483e-05, "loss": 0.7429, "step": 5343 }, { "epoch": 0.1935530604853314, "grad_norm": 1.5361529213764722, "learning_rate": 1.8629889690444073e-05, "loss": 0.8938, "step": 5344 }, { "epoch": 0.19358927924664976, "grad_norm": 1.5410040835636112, "learning_rate": 1.86292969705558e-05, "loss": 0.7659, "step": 5345 }, { "epoch": 0.19362549800796813, "grad_norm": 1.451641214800816, "learning_rate": 1.8628704131920815e-05, "loss": 0.8052, "step": 5346 }, { "epoch": 0.19366171676928648, "grad_norm": 1.485321195092605, "learning_rate": 1.8628111174547282e-05, "loss": 0.9126, "step": 5347 }, { "epoch": 0.19369793553060485, "grad_norm": 1.370605408274748, "learning_rate": 1.8627518098443355e-05, "loss": 0.8338, "step": 5348 }, { "epoch": 0.19373415429192323, "grad_norm": 1.4878276361445881, "learning_rate": 1.86269249036172e-05, "loss": 0.8461, "step": 5349 }, { "epoch": 0.19377037305324157, "grad_norm": 1.2609303656805213, "learning_rate": 1.8626331590076978e-05, "loss": 0.7816, "step": 5350 }, { "epoch": 0.19380659181455995, "grad_norm": 1.5392369768077903, "learning_rate": 1.8625738157830855e-05, "loss": 0.8636, "step": 5351 }, { "epoch": 0.1938428105758783, "grad_norm": 1.5095737638864681, "learning_rate": 1.8625144606886993e-05, "loss": 0.8593, "step": 5352 }, { "epoch": 0.19387902933719667, "grad_norm": 1.3543183332723219, "learning_rate": 1.862455093725356e-05, "loss": 0.797, "step": 5353 }, { "epoch": 0.19391524809851504, "grad_norm": 1.5092774770360842, "learning_rate": 1.8623957148938735e-05, "loss": 0.737, "step": 5354 }, { "epoch": 0.1939514668598334, "grad_norm": 1.4947304434123414, "learning_rate": 1.862336324195068e-05, "loss": 0.8792, "step": 5355 }, { "epoch": 0.19398768562115176, "grad_norm": 1.5042197334244334, "learning_rate": 1.8622769216297567e-05, "loss": 0.8621, "step": 5356 }, { "epoch": 0.1940239043824701, "grad_norm": 1.576130709293731, "learning_rate": 1.8622175071987576e-05, "loss": 0.7326, "step": 5357 }, { "epoch": 0.19406012314378848, "grad_norm": 1.4561512158414884, "learning_rate": 1.8621580809028876e-05, "loss": 0.804, "step": 5358 }, { "epoch": 0.19409634190510686, "grad_norm": 1.3596173034863794, "learning_rate": 1.8620986427429652e-05, "loss": 0.7714, "step": 5359 }, { "epoch": 0.1941325606664252, "grad_norm": 1.523831491233151, "learning_rate": 1.8620391927198077e-05, "loss": 0.8051, "step": 5360 }, { "epoch": 0.19416877942774358, "grad_norm": 1.578585424746751, "learning_rate": 1.8619797308342337e-05, "loss": 0.8651, "step": 5361 }, { "epoch": 0.19420499818906192, "grad_norm": 1.6001700966076406, "learning_rate": 1.8619202570870613e-05, "loss": 0.8952, "step": 5362 }, { "epoch": 0.1942412169503803, "grad_norm": 1.009240314600709, "learning_rate": 1.8618607714791084e-05, "loss": 0.792, "step": 5363 }, { "epoch": 0.19427743571169867, "grad_norm": 1.2872611160440128, "learning_rate": 1.861801274011194e-05, "loss": 0.7987, "step": 5364 }, { "epoch": 0.19431365447301702, "grad_norm": 1.5816418615964283, "learning_rate": 1.861741764684137e-05, "loss": 0.8342, "step": 5365 }, { "epoch": 0.1943498732343354, "grad_norm": 1.4108564812452946, "learning_rate": 1.8616822434987564e-05, "loss": 0.7721, "step": 5366 }, { "epoch": 0.19438609199565374, "grad_norm": 1.4580476816734473, "learning_rate": 1.8616227104558703e-05, "loss": 0.7712, "step": 5367 }, { "epoch": 0.1944223107569721, "grad_norm": 1.5031090431694811, "learning_rate": 1.8615631655562993e-05, "loss": 0.7622, "step": 5368 }, { "epoch": 0.19445852951829048, "grad_norm": 1.326876440519107, "learning_rate": 1.861503608800862e-05, "loss": 0.7833, "step": 5369 }, { "epoch": 0.19449474827960883, "grad_norm": 1.4694692304037806, "learning_rate": 1.8614440401903776e-05, "loss": 0.8527, "step": 5370 }, { "epoch": 0.1945309670409272, "grad_norm": 1.393192949373931, "learning_rate": 1.8613844597256668e-05, "loss": 0.7801, "step": 5371 }, { "epoch": 0.19456718580224555, "grad_norm": 1.4262586003569957, "learning_rate": 1.861324867407549e-05, "loss": 0.7734, "step": 5372 }, { "epoch": 0.19460340456356393, "grad_norm": 1.258947389484302, "learning_rate": 1.8612652632368436e-05, "loss": 0.8151, "step": 5373 }, { "epoch": 0.1946396233248823, "grad_norm": 1.5794887769814872, "learning_rate": 1.8612056472143717e-05, "loss": 0.837, "step": 5374 }, { "epoch": 0.19467584208620065, "grad_norm": 1.1591860057257708, "learning_rate": 1.8611460193409535e-05, "loss": 0.8264, "step": 5375 }, { "epoch": 0.19471206084751902, "grad_norm": 1.4854823659326193, "learning_rate": 1.861086379617409e-05, "loss": 0.7619, "step": 5376 }, { "epoch": 0.19474827960883737, "grad_norm": 1.568296190131578, "learning_rate": 1.86102672804456e-05, "loss": 0.819, "step": 5377 }, { "epoch": 0.19478449837015574, "grad_norm": 1.399140309316805, "learning_rate": 1.860967064623226e-05, "loss": 0.7946, "step": 5378 }, { "epoch": 0.1948207171314741, "grad_norm": 1.7259175122899657, "learning_rate": 1.860907389354229e-05, "loss": 0.9076, "step": 5379 }, { "epoch": 0.19485693589279246, "grad_norm": 1.7886031604354995, "learning_rate": 1.86084770223839e-05, "loss": 0.8821, "step": 5380 }, { "epoch": 0.19489315465411083, "grad_norm": 1.6239564271636953, "learning_rate": 1.8607880032765298e-05, "loss": 0.8963, "step": 5381 }, { "epoch": 0.19492937341542918, "grad_norm": 1.5677383536546237, "learning_rate": 1.8607282924694706e-05, "loss": 0.8318, "step": 5382 }, { "epoch": 0.19496559217674755, "grad_norm": 1.5440888383375784, "learning_rate": 1.860668569818034e-05, "loss": 0.8666, "step": 5383 }, { "epoch": 0.19500181093806593, "grad_norm": 1.5120692019397037, "learning_rate": 1.860608835323041e-05, "loss": 0.9022, "step": 5384 }, { "epoch": 0.19503802969938427, "grad_norm": 1.4628120421629125, "learning_rate": 1.8605490889853145e-05, "loss": 0.8503, "step": 5385 }, { "epoch": 0.19507424846070265, "grad_norm": 1.6235846834587644, "learning_rate": 1.8604893308056765e-05, "loss": 0.8696, "step": 5386 }, { "epoch": 0.195110467222021, "grad_norm": 1.4806227314570557, "learning_rate": 1.8604295607849495e-05, "loss": 0.7725, "step": 5387 }, { "epoch": 0.19514668598333937, "grad_norm": 1.59104981619831, "learning_rate": 1.8603697789239555e-05, "loss": 0.8927, "step": 5388 }, { "epoch": 0.19518290474465774, "grad_norm": 1.334491077949175, "learning_rate": 1.860309985223517e-05, "loss": 0.8191, "step": 5389 }, { "epoch": 0.1952191235059761, "grad_norm": 1.499811363031489, "learning_rate": 1.8602501796844578e-05, "loss": 0.808, "step": 5390 }, { "epoch": 0.19525534226729446, "grad_norm": 1.330280237484339, "learning_rate": 1.8601903623075996e-05, "loss": 0.8905, "step": 5391 }, { "epoch": 0.1952915610286128, "grad_norm": 1.5513190248666915, "learning_rate": 1.8601305330937664e-05, "loss": 0.8054, "step": 5392 }, { "epoch": 0.19532777978993118, "grad_norm": 1.560182228024917, "learning_rate": 1.8600706920437815e-05, "loss": 0.8095, "step": 5393 }, { "epoch": 0.19536399855124956, "grad_norm": 1.1693333609716516, "learning_rate": 1.860010839158468e-05, "loss": 0.7509, "step": 5394 }, { "epoch": 0.1954002173125679, "grad_norm": 1.7727998772004838, "learning_rate": 1.8599509744386498e-05, "loss": 0.7617, "step": 5395 }, { "epoch": 0.19543643607388628, "grad_norm": 1.2253316832111931, "learning_rate": 1.8598910978851504e-05, "loss": 0.8139, "step": 5396 }, { "epoch": 0.19547265483520462, "grad_norm": 1.491538457121246, "learning_rate": 1.859831209498794e-05, "loss": 0.872, "step": 5397 }, { "epoch": 0.195508873596523, "grad_norm": 1.4990355755193394, "learning_rate": 1.8597713092804045e-05, "loss": 0.8502, "step": 5398 }, { "epoch": 0.19554509235784137, "grad_norm": 1.443330514126323, "learning_rate": 1.8597113972308064e-05, "loss": 0.8505, "step": 5399 }, { "epoch": 0.19558131111915972, "grad_norm": 1.5225956155595535, "learning_rate": 1.8596514733508244e-05, "loss": 0.8671, "step": 5400 }, { "epoch": 0.1956175298804781, "grad_norm": 1.4717455585321029, "learning_rate": 1.8595915376412823e-05, "loss": 0.7593, "step": 5401 }, { "epoch": 0.19565374864179644, "grad_norm": 1.654917751345476, "learning_rate": 1.8595315901030057e-05, "loss": 0.9227, "step": 5402 }, { "epoch": 0.1956899674031148, "grad_norm": 1.4992487641759682, "learning_rate": 1.859471630736819e-05, "loss": 0.8165, "step": 5403 }, { "epoch": 0.19572618616443319, "grad_norm": 1.4560207533128162, "learning_rate": 1.859411659543547e-05, "loss": 0.7902, "step": 5404 }, { "epoch": 0.19576240492575153, "grad_norm": 1.3530587344484388, "learning_rate": 1.8593516765240164e-05, "loss": 0.7879, "step": 5405 }, { "epoch": 0.1957986236870699, "grad_norm": 1.486738350108088, "learning_rate": 1.859291681679051e-05, "loss": 0.7628, "step": 5406 }, { "epoch": 0.19583484244838825, "grad_norm": 1.6195261263961822, "learning_rate": 1.859231675009477e-05, "loss": 0.8849, "step": 5407 }, { "epoch": 0.19587106120970663, "grad_norm": 1.5677190283371443, "learning_rate": 1.8591716565161203e-05, "loss": 0.8187, "step": 5408 }, { "epoch": 0.195907279971025, "grad_norm": 1.7070135534665511, "learning_rate": 1.8591116261998068e-05, "loss": 0.8157, "step": 5409 }, { "epoch": 0.19594349873234335, "grad_norm": 1.4249695538051925, "learning_rate": 1.8590515840613622e-05, "loss": 0.8412, "step": 5410 }, { "epoch": 0.19597971749366172, "grad_norm": 1.2402532335784973, "learning_rate": 1.858991530101613e-05, "loss": 0.757, "step": 5411 }, { "epoch": 0.19601593625498007, "grad_norm": 1.5115967018677634, "learning_rate": 1.8589314643213855e-05, "loss": 0.8845, "step": 5412 }, { "epoch": 0.19605215501629844, "grad_norm": 1.5236633469605483, "learning_rate": 1.858871386721507e-05, "loss": 0.7918, "step": 5413 }, { "epoch": 0.19608837377761681, "grad_norm": 1.3642531328660334, "learning_rate": 1.8588112973028026e-05, "loss": 0.9009, "step": 5414 }, { "epoch": 0.19612459253893516, "grad_norm": 1.421813664515116, "learning_rate": 1.8587511960661006e-05, "loss": 0.7967, "step": 5415 }, { "epoch": 0.19616081130025353, "grad_norm": 1.5009367905896438, "learning_rate": 1.8586910830122277e-05, "loss": 0.8711, "step": 5416 }, { "epoch": 0.19619703006157188, "grad_norm": 1.5880184931491244, "learning_rate": 1.858630958142011e-05, "loss": 0.8006, "step": 5417 }, { "epoch": 0.19623324882289025, "grad_norm": 1.4573511949770077, "learning_rate": 1.8585708214562776e-05, "loss": 0.7718, "step": 5418 }, { "epoch": 0.19626946758420863, "grad_norm": 1.4086772012242288, "learning_rate": 1.8585106729558553e-05, "loss": 0.8988, "step": 5419 }, { "epoch": 0.19630568634552698, "grad_norm": 1.2517465089065742, "learning_rate": 1.858450512641572e-05, "loss": 0.749, "step": 5420 }, { "epoch": 0.19634190510684535, "grad_norm": 1.5164750869933876, "learning_rate": 1.8583903405142552e-05, "loss": 0.8368, "step": 5421 }, { "epoch": 0.1963781238681637, "grad_norm": 1.8430688340274115, "learning_rate": 1.858330156574733e-05, "loss": 0.8438, "step": 5422 }, { "epoch": 0.19641434262948207, "grad_norm": 1.4955405676718247, "learning_rate": 1.858269960823834e-05, "loss": 0.8707, "step": 5423 }, { "epoch": 0.19645056139080044, "grad_norm": 1.5825394508828938, "learning_rate": 1.8582097532623856e-05, "loss": 0.7042, "step": 5424 }, { "epoch": 0.1964867801521188, "grad_norm": 1.6828445174731934, "learning_rate": 1.858149533891217e-05, "loss": 0.8195, "step": 5425 }, { "epoch": 0.19652299891343716, "grad_norm": 1.29589436946538, "learning_rate": 1.8580893027111574e-05, "loss": 0.871, "step": 5426 }, { "epoch": 0.1965592176747555, "grad_norm": 1.5035576738427636, "learning_rate": 1.8580290597230342e-05, "loss": 0.8951, "step": 5427 }, { "epoch": 0.19659543643607388, "grad_norm": 1.4781017221740969, "learning_rate": 1.8579688049276777e-05, "loss": 0.8592, "step": 5428 }, { "epoch": 0.19663165519739226, "grad_norm": 1.935791823884562, "learning_rate": 1.8579085383259165e-05, "loss": 0.8909, "step": 5429 }, { "epoch": 0.1966678739587106, "grad_norm": 1.4645092709458047, "learning_rate": 1.8578482599185804e-05, "loss": 0.8579, "step": 5430 }, { "epoch": 0.19670409272002898, "grad_norm": 1.4746227314325202, "learning_rate": 1.8577879697064978e-05, "loss": 0.8574, "step": 5431 }, { "epoch": 0.19674031148134732, "grad_norm": 1.350711808369526, "learning_rate": 1.857727667690499e-05, "loss": 0.8027, "step": 5432 }, { "epoch": 0.1967765302426657, "grad_norm": 1.556411340979185, "learning_rate": 1.8576673538714144e-05, "loss": 0.8948, "step": 5433 }, { "epoch": 0.19681274900398407, "grad_norm": 1.1043315430915834, "learning_rate": 1.8576070282500732e-05, "loss": 0.7688, "step": 5434 }, { "epoch": 0.19684896776530242, "grad_norm": 1.438860665217634, "learning_rate": 1.8575466908273056e-05, "loss": 0.8179, "step": 5435 }, { "epoch": 0.1968851865266208, "grad_norm": 1.4824857823665516, "learning_rate": 1.857486341603942e-05, "loss": 0.8455, "step": 5436 }, { "epoch": 0.19692140528793917, "grad_norm": 1.5563713999204056, "learning_rate": 1.857425980580813e-05, "loss": 0.8862, "step": 5437 }, { "epoch": 0.1969576240492575, "grad_norm": 1.4714933883719041, "learning_rate": 1.857365607758749e-05, "loss": 0.7599, "step": 5438 }, { "epoch": 0.1969938428105759, "grad_norm": 1.4427709373657989, "learning_rate": 1.857305223138581e-05, "loss": 0.9128, "step": 5439 }, { "epoch": 0.19703006157189423, "grad_norm": 1.6365637762768737, "learning_rate": 1.8572448267211397e-05, "loss": 0.8287, "step": 5440 }, { "epoch": 0.1970662803332126, "grad_norm": 1.4116506603130639, "learning_rate": 1.8571844185072563e-05, "loss": 0.8493, "step": 5441 }, { "epoch": 0.19710249909453098, "grad_norm": 1.334495714520905, "learning_rate": 1.8571239984977625e-05, "loss": 0.8102, "step": 5442 }, { "epoch": 0.19713871785584933, "grad_norm": 1.3607354830722773, "learning_rate": 1.8570635666934888e-05, "loss": 0.8284, "step": 5443 }, { "epoch": 0.1971749366171677, "grad_norm": 1.558711406024254, "learning_rate": 1.8570031230952676e-05, "loss": 0.7338, "step": 5444 }, { "epoch": 0.19721115537848605, "grad_norm": 1.3466219678740368, "learning_rate": 1.8569426677039306e-05, "loss": 0.853, "step": 5445 }, { "epoch": 0.19724737413980442, "grad_norm": 1.6215851073564704, "learning_rate": 1.8568822005203094e-05, "loss": 0.8475, "step": 5446 }, { "epoch": 0.1972835929011228, "grad_norm": 1.09414867283593, "learning_rate": 1.856821721545236e-05, "loss": 0.7459, "step": 5447 }, { "epoch": 0.19731981166244114, "grad_norm": 1.3298627872334745, "learning_rate": 1.856761230779543e-05, "loss": 0.8047, "step": 5448 }, { "epoch": 0.19735603042375952, "grad_norm": 1.4381335426366555, "learning_rate": 1.8567007282240627e-05, "loss": 0.8622, "step": 5449 }, { "epoch": 0.19739224918507786, "grad_norm": 1.402997273725422, "learning_rate": 1.8566402138796277e-05, "loss": 0.8852, "step": 5450 }, { "epoch": 0.19742846794639624, "grad_norm": 1.2973423594696254, "learning_rate": 1.85657968774707e-05, "loss": 0.7676, "step": 5451 }, { "epoch": 0.1974646867077146, "grad_norm": 1.516379187788081, "learning_rate": 1.8565191498272236e-05, "loss": 0.7687, "step": 5452 }, { "epoch": 0.19750090546903296, "grad_norm": 2.166622164655901, "learning_rate": 1.8564586001209214e-05, "loss": 0.7768, "step": 5453 }, { "epoch": 0.19753712423035133, "grad_norm": 1.5002580702579633, "learning_rate": 1.856398038628996e-05, "loss": 0.8428, "step": 5454 }, { "epoch": 0.19757334299166968, "grad_norm": 1.4259647995887237, "learning_rate": 1.8563374653522814e-05, "loss": 0.8348, "step": 5455 }, { "epoch": 0.19760956175298805, "grad_norm": 1.254032557207494, "learning_rate": 1.8562768802916104e-05, "loss": 0.8414, "step": 5456 }, { "epoch": 0.19764578051430642, "grad_norm": 1.3701983068077415, "learning_rate": 1.8562162834478173e-05, "loss": 0.8354, "step": 5457 }, { "epoch": 0.19768199927562477, "grad_norm": 1.368024591583912, "learning_rate": 1.8561556748217357e-05, "loss": 0.8216, "step": 5458 }, { "epoch": 0.19771821803694314, "grad_norm": 1.5773085793366604, "learning_rate": 1.8560950544142e-05, "loss": 0.8245, "step": 5459 }, { "epoch": 0.1977544367982615, "grad_norm": 1.205094387152995, "learning_rate": 1.856034422226044e-05, "loss": 0.7559, "step": 5460 }, { "epoch": 0.19779065555957986, "grad_norm": 1.4919003552378522, "learning_rate": 1.8559737782581017e-05, "loss": 0.7819, "step": 5461 }, { "epoch": 0.19782687432089824, "grad_norm": 1.544452061694982, "learning_rate": 1.8559131225112085e-05, "loss": 0.8309, "step": 5462 }, { "epoch": 0.19786309308221658, "grad_norm": 1.4297105091619808, "learning_rate": 1.8558524549861986e-05, "loss": 0.7669, "step": 5463 }, { "epoch": 0.19789931184353496, "grad_norm": 1.7801617949901742, "learning_rate": 1.855791775683907e-05, "loss": 0.8722, "step": 5464 }, { "epoch": 0.1979355306048533, "grad_norm": 1.5316284153662516, "learning_rate": 1.8557310846051687e-05, "loss": 0.8326, "step": 5465 }, { "epoch": 0.19797174936617168, "grad_norm": 1.6975366970753027, "learning_rate": 1.8556703817508187e-05, "loss": 0.9345, "step": 5466 }, { "epoch": 0.19800796812749005, "grad_norm": 1.4242851095648432, "learning_rate": 1.855609667121692e-05, "loss": 0.7014, "step": 5467 }, { "epoch": 0.1980441868888084, "grad_norm": 1.5835554527616424, "learning_rate": 1.855548940718625e-05, "loss": 0.8463, "step": 5468 }, { "epoch": 0.19808040565012677, "grad_norm": 1.4618035282961994, "learning_rate": 1.8554882025424526e-05, "loss": 0.8268, "step": 5469 }, { "epoch": 0.19811662441144512, "grad_norm": 1.1698547721765542, "learning_rate": 1.855427452594011e-05, "loss": 0.7816, "step": 5470 }, { "epoch": 0.1981528431727635, "grad_norm": 2.1121408978369782, "learning_rate": 1.855366690874136e-05, "loss": 0.8872, "step": 5471 }, { "epoch": 0.19818906193408187, "grad_norm": 1.448236831351605, "learning_rate": 1.8553059173836634e-05, "loss": 0.8333, "step": 5472 }, { "epoch": 0.1982252806954002, "grad_norm": 1.7133237515147113, "learning_rate": 1.8552451321234303e-05, "loss": 0.8563, "step": 5473 }, { "epoch": 0.1982614994567186, "grad_norm": 1.5158898402389085, "learning_rate": 1.8551843350942723e-05, "loss": 0.9074, "step": 5474 }, { "epoch": 0.19829771821803693, "grad_norm": 1.4835496656493106, "learning_rate": 1.8551235262970264e-05, "loss": 0.8819, "step": 5475 }, { "epoch": 0.1983339369793553, "grad_norm": 1.4581023081507045, "learning_rate": 1.8550627057325298e-05, "loss": 0.8141, "step": 5476 }, { "epoch": 0.19837015574067368, "grad_norm": 1.4986984304437716, "learning_rate": 1.8550018734016187e-05, "loss": 0.8099, "step": 5477 }, { "epoch": 0.19840637450199203, "grad_norm": 1.4678687383863658, "learning_rate": 1.8549410293051305e-05, "loss": 0.8039, "step": 5478 }, { "epoch": 0.1984425932633104, "grad_norm": 1.712689370497285, "learning_rate": 1.854880173443903e-05, "loss": 0.887, "step": 5479 }, { "epoch": 0.19847881202462875, "grad_norm": 1.364637412860256, "learning_rate": 1.8548193058187725e-05, "loss": 0.7638, "step": 5480 }, { "epoch": 0.19851503078594712, "grad_norm": 1.6292360804320103, "learning_rate": 1.8547584264305777e-05, "loss": 0.8341, "step": 5481 }, { "epoch": 0.1985512495472655, "grad_norm": 1.6210974291104772, "learning_rate": 1.8546975352801557e-05, "loss": 0.874, "step": 5482 }, { "epoch": 0.19858746830858384, "grad_norm": 1.3941442923792868, "learning_rate": 1.8546366323683445e-05, "loss": 0.796, "step": 5483 }, { "epoch": 0.19862368706990222, "grad_norm": 1.4766069341210504, "learning_rate": 1.8545757176959826e-05, "loss": 0.8196, "step": 5484 }, { "epoch": 0.19865990583122056, "grad_norm": 1.5197519262256614, "learning_rate": 1.854514791263908e-05, "loss": 0.7724, "step": 5485 }, { "epoch": 0.19869612459253894, "grad_norm": 1.4353229477807565, "learning_rate": 1.8544538530729587e-05, "loss": 0.7911, "step": 5486 }, { "epoch": 0.1987323433538573, "grad_norm": 1.6168448037856482, "learning_rate": 1.8543929031239737e-05, "loss": 0.8419, "step": 5487 }, { "epoch": 0.19876856211517566, "grad_norm": 1.3609434549631683, "learning_rate": 1.8543319414177916e-05, "loss": 0.7909, "step": 5488 }, { "epoch": 0.19880478087649403, "grad_norm": 1.4705919499823692, "learning_rate": 1.8542709679552513e-05, "loss": 0.8649, "step": 5489 }, { "epoch": 0.19884099963781238, "grad_norm": 1.6580958269898145, "learning_rate": 1.854209982737192e-05, "loss": 0.7433, "step": 5490 }, { "epoch": 0.19887721839913075, "grad_norm": 1.4776406111428657, "learning_rate": 1.8541489857644527e-05, "loss": 0.803, "step": 5491 }, { "epoch": 0.19891343716044912, "grad_norm": 1.5041103162660614, "learning_rate": 1.854087977037873e-05, "loss": 0.9117, "step": 5492 }, { "epoch": 0.19894965592176747, "grad_norm": 1.5260852198168549, "learning_rate": 1.8540269565582918e-05, "loss": 0.8032, "step": 5493 }, { "epoch": 0.19898587468308584, "grad_norm": 1.6346919135426263, "learning_rate": 1.8539659243265497e-05, "loss": 0.832, "step": 5494 }, { "epoch": 0.1990220934444042, "grad_norm": 1.5917023806227812, "learning_rate": 1.8539048803434858e-05, "loss": 0.9119, "step": 5495 }, { "epoch": 0.19905831220572257, "grad_norm": 1.4202621910052609, "learning_rate": 1.8538438246099405e-05, "loss": 0.8024, "step": 5496 }, { "epoch": 0.19909453096704094, "grad_norm": 1.4858043715310367, "learning_rate": 1.8537827571267542e-05, "loss": 0.8243, "step": 5497 }, { "epoch": 0.19913074972835929, "grad_norm": 1.56819468710754, "learning_rate": 1.8537216778947666e-05, "loss": 0.9793, "step": 5498 }, { "epoch": 0.19916696848967766, "grad_norm": 1.0714918704925445, "learning_rate": 1.8536605869148186e-05, "loss": 0.7826, "step": 5499 }, { "epoch": 0.199203187250996, "grad_norm": 1.5520464602886517, "learning_rate": 1.8535994841877512e-05, "loss": 0.9032, "step": 5500 }, { "epoch": 0.19923940601231438, "grad_norm": 1.4800450409408012, "learning_rate": 1.8535383697144046e-05, "loss": 0.7314, "step": 5501 }, { "epoch": 0.19927562477363275, "grad_norm": 1.0817696944014228, "learning_rate": 1.85347724349562e-05, "loss": 0.829, "step": 5502 }, { "epoch": 0.1993118435349511, "grad_norm": 1.4420018564696033, "learning_rate": 1.8534161055322388e-05, "loss": 0.8065, "step": 5503 }, { "epoch": 0.19934806229626947, "grad_norm": 1.455002499221633, "learning_rate": 1.8533549558251018e-05, "loss": 0.8366, "step": 5504 }, { "epoch": 0.19938428105758782, "grad_norm": 1.0498419499992178, "learning_rate": 1.853293794375051e-05, "loss": 0.7286, "step": 5505 }, { "epoch": 0.1994204998189062, "grad_norm": 1.4971060913448855, "learning_rate": 1.853232621182928e-05, "loss": 0.8166, "step": 5506 }, { "epoch": 0.19945671858022457, "grad_norm": 1.471671036692952, "learning_rate": 1.853171436249574e-05, "loss": 0.8124, "step": 5507 }, { "epoch": 0.19949293734154291, "grad_norm": 1.5484918635820961, "learning_rate": 1.8531102395758318e-05, "loss": 0.8091, "step": 5508 }, { "epoch": 0.1995291561028613, "grad_norm": 1.146228305208168, "learning_rate": 1.8530490311625428e-05, "loss": 0.7531, "step": 5509 }, { "epoch": 0.19956537486417963, "grad_norm": 1.6411664228339233, "learning_rate": 1.8529878110105498e-05, "loss": 0.7823, "step": 5510 }, { "epoch": 0.199601593625498, "grad_norm": 1.5769564638723765, "learning_rate": 1.8529265791206947e-05, "loss": 0.9012, "step": 5511 }, { "epoch": 0.19963781238681638, "grad_norm": 1.173729389373311, "learning_rate": 1.8528653354938205e-05, "loss": 0.7711, "step": 5512 }, { "epoch": 0.19967403114813473, "grad_norm": 1.5224730856351443, "learning_rate": 1.8528040801307702e-05, "loss": 0.856, "step": 5513 }, { "epoch": 0.1997102499094531, "grad_norm": 1.3988005403974937, "learning_rate": 1.852742813032386e-05, "loss": 0.7574, "step": 5514 }, { "epoch": 0.19974646867077145, "grad_norm": 1.3961338350190828, "learning_rate": 1.8526815341995117e-05, "loss": 0.8541, "step": 5515 }, { "epoch": 0.19978268743208982, "grad_norm": 1.4917580688155896, "learning_rate": 1.85262024363299e-05, "loss": 0.8706, "step": 5516 }, { "epoch": 0.1998189061934082, "grad_norm": 1.269860956220486, "learning_rate": 1.852558941333665e-05, "loss": 0.8025, "step": 5517 }, { "epoch": 0.19985512495472654, "grad_norm": 1.5285291382835446, "learning_rate": 1.8524976273023795e-05, "loss": 0.9061, "step": 5518 }, { "epoch": 0.19989134371604492, "grad_norm": 1.6423807388439553, "learning_rate": 1.8524363015399774e-05, "loss": 0.8362, "step": 5519 }, { "epoch": 0.19992756247736326, "grad_norm": 1.3861319549587996, "learning_rate": 1.852374964047303e-05, "loss": 0.8099, "step": 5520 }, { "epoch": 0.19996378123868164, "grad_norm": 1.6268650159298639, "learning_rate": 1.8523136148252003e-05, "loss": 0.7821, "step": 5521 }, { "epoch": 0.2, "grad_norm": 1.6633466257548541, "learning_rate": 1.8522522538745133e-05, "loss": 0.8442, "step": 5522 }, { "epoch": 0.20003621876131836, "grad_norm": 1.6938434247788519, "learning_rate": 1.8521908811960862e-05, "loss": 0.8116, "step": 5523 }, { "epoch": 0.20007243752263673, "grad_norm": 1.2472911134590638, "learning_rate": 1.852129496790764e-05, "loss": 0.8108, "step": 5524 }, { "epoch": 0.20010865628395508, "grad_norm": 1.3801276798944804, "learning_rate": 1.8520681006593913e-05, "loss": 0.787, "step": 5525 }, { "epoch": 0.20014487504527345, "grad_norm": 1.3742513277845212, "learning_rate": 1.8520066928028124e-05, "loss": 0.7313, "step": 5526 }, { "epoch": 0.20018109380659183, "grad_norm": 1.5067137979586085, "learning_rate": 1.8519452732218732e-05, "loss": 0.871, "step": 5527 }, { "epoch": 0.20021731256791017, "grad_norm": 1.4438117103114196, "learning_rate": 1.851883841917418e-05, "loss": 0.8443, "step": 5528 }, { "epoch": 0.20025353132922855, "grad_norm": 1.2392429219114847, "learning_rate": 1.8518223988902934e-05, "loss": 0.7816, "step": 5529 }, { "epoch": 0.2002897500905469, "grad_norm": 1.519879180550718, "learning_rate": 1.8517609441413434e-05, "loss": 0.8335, "step": 5530 }, { "epoch": 0.20032596885186527, "grad_norm": 1.4098144570093243, "learning_rate": 1.8516994776714145e-05, "loss": 0.8019, "step": 5531 }, { "epoch": 0.20036218761318364, "grad_norm": 1.460522741331081, "learning_rate": 1.8516379994813526e-05, "loss": 0.8539, "step": 5532 }, { "epoch": 0.200398406374502, "grad_norm": 1.584166192980342, "learning_rate": 1.8515765095720033e-05, "loss": 0.8851, "step": 5533 }, { "epoch": 0.20043462513582036, "grad_norm": 1.4708268796164117, "learning_rate": 1.851515007944213e-05, "loss": 0.7601, "step": 5534 }, { "epoch": 0.2004708438971387, "grad_norm": 1.45010432975837, "learning_rate": 1.851453494598828e-05, "loss": 0.8195, "step": 5535 }, { "epoch": 0.20050706265845708, "grad_norm": 1.5023864671582314, "learning_rate": 1.8513919695366952e-05, "loss": 0.8383, "step": 5536 }, { "epoch": 0.20054328141977545, "grad_norm": 1.4998025739406273, "learning_rate": 1.85133043275866e-05, "loss": 0.7998, "step": 5537 }, { "epoch": 0.2005795001810938, "grad_norm": 1.3190828742622702, "learning_rate": 1.8512688842655705e-05, "loss": 0.8208, "step": 5538 }, { "epoch": 0.20061571894241217, "grad_norm": 1.4848417301238503, "learning_rate": 1.8512073240582732e-05, "loss": 0.8494, "step": 5539 }, { "epoch": 0.20065193770373052, "grad_norm": 1.5974297945937541, "learning_rate": 1.851145752137615e-05, "loss": 0.7865, "step": 5540 }, { "epoch": 0.2006881564650489, "grad_norm": 1.2178051447507043, "learning_rate": 1.8510841685044436e-05, "loss": 0.7934, "step": 5541 }, { "epoch": 0.20072437522636727, "grad_norm": 1.2716686585203225, "learning_rate": 1.8510225731596058e-05, "loss": 0.7806, "step": 5542 }, { "epoch": 0.20076059398768561, "grad_norm": 1.585154124982828, "learning_rate": 1.85096096610395e-05, "loss": 0.878, "step": 5543 }, { "epoch": 0.200796812749004, "grad_norm": 1.5491557243837148, "learning_rate": 1.8508993473383234e-05, "loss": 0.8603, "step": 5544 }, { "epoch": 0.20083303151032234, "grad_norm": 1.5745027051739942, "learning_rate": 1.8508377168635743e-05, "loss": 0.7993, "step": 5545 }, { "epoch": 0.2008692502716407, "grad_norm": 1.6278446074589323, "learning_rate": 1.8507760746805507e-05, "loss": 0.834, "step": 5546 }, { "epoch": 0.20090546903295908, "grad_norm": 1.6281033739641586, "learning_rate": 1.8507144207901003e-05, "loss": 0.8534, "step": 5547 }, { "epoch": 0.20094168779427743, "grad_norm": 2.0213174852412537, "learning_rate": 1.850652755193072e-05, "loss": 0.82, "step": 5548 }, { "epoch": 0.2009779065555958, "grad_norm": 2.0108847769025724, "learning_rate": 1.8505910778903146e-05, "loss": 0.8288, "step": 5549 }, { "epoch": 0.20101412531691415, "grad_norm": 1.6226523180286299, "learning_rate": 1.8505293888826764e-05, "loss": 0.8893, "step": 5550 }, { "epoch": 0.20105034407823252, "grad_norm": 1.7120204028598625, "learning_rate": 1.8504676881710066e-05, "loss": 0.8765, "step": 5551 }, { "epoch": 0.2010865628395509, "grad_norm": 1.4030189475939856, "learning_rate": 1.850405975756154e-05, "loss": 0.8868, "step": 5552 }, { "epoch": 0.20112278160086924, "grad_norm": 1.5300361739059318, "learning_rate": 1.8503442516389684e-05, "loss": 0.8905, "step": 5553 }, { "epoch": 0.20115900036218762, "grad_norm": 1.5990464089694614, "learning_rate": 1.8502825158202983e-05, "loss": 0.933, "step": 5554 }, { "epoch": 0.20119521912350596, "grad_norm": 1.6745826279257345, "learning_rate": 1.850220768300994e-05, "loss": 0.7468, "step": 5555 }, { "epoch": 0.20123143788482434, "grad_norm": 1.4271841711131197, "learning_rate": 1.8501590090819044e-05, "loss": 0.7892, "step": 5556 }, { "epoch": 0.2012676566461427, "grad_norm": 1.7456419720867786, "learning_rate": 1.8500972381638797e-05, "loss": 0.8266, "step": 5557 }, { "epoch": 0.20130387540746106, "grad_norm": 1.1896801034244118, "learning_rate": 1.8500354555477706e-05, "loss": 0.7371, "step": 5558 }, { "epoch": 0.20134009416877943, "grad_norm": 1.485641463114883, "learning_rate": 1.8499736612344265e-05, "loss": 0.7842, "step": 5559 }, { "epoch": 0.20137631293009778, "grad_norm": 1.4736736082139537, "learning_rate": 1.849911855224698e-05, "loss": 0.8865, "step": 5560 }, { "epoch": 0.20141253169141615, "grad_norm": 1.556933603990798, "learning_rate": 1.849850037519435e-05, "loss": 0.8268, "step": 5561 }, { "epoch": 0.20144875045273453, "grad_norm": 1.6314307051536383, "learning_rate": 1.8497882081194896e-05, "loss": 0.8283, "step": 5562 }, { "epoch": 0.20148496921405287, "grad_norm": 1.5395372476657416, "learning_rate": 1.8497263670257112e-05, "loss": 0.7783, "step": 5563 }, { "epoch": 0.20152118797537125, "grad_norm": 1.5695812387324184, "learning_rate": 1.8496645142389516e-05, "loss": 0.7913, "step": 5564 }, { "epoch": 0.2015574067366896, "grad_norm": 1.776268407372592, "learning_rate": 1.849602649760062e-05, "loss": 0.882, "step": 5565 }, { "epoch": 0.20159362549800797, "grad_norm": 1.6303758646851523, "learning_rate": 1.849540773589893e-05, "loss": 0.802, "step": 5566 }, { "epoch": 0.20162984425932634, "grad_norm": 1.5032364387849348, "learning_rate": 1.8494788857292968e-05, "loss": 0.7926, "step": 5567 }, { "epoch": 0.2016660630206447, "grad_norm": 1.535943577430359, "learning_rate": 1.849416986179124e-05, "loss": 0.828, "step": 5568 }, { "epoch": 0.20170228178196306, "grad_norm": 1.4622573621461086, "learning_rate": 1.8493550749402278e-05, "loss": 0.7256, "step": 5569 }, { "epoch": 0.2017385005432814, "grad_norm": 1.278256291137028, "learning_rate": 1.8492931520134593e-05, "loss": 0.8535, "step": 5570 }, { "epoch": 0.20177471930459978, "grad_norm": 1.4817230854375274, "learning_rate": 1.8492312173996707e-05, "loss": 0.8081, "step": 5571 }, { "epoch": 0.20181093806591816, "grad_norm": 1.4401412403503653, "learning_rate": 1.849169271099714e-05, "loss": 0.8942, "step": 5572 }, { "epoch": 0.2018471568272365, "grad_norm": 1.425391808977417, "learning_rate": 1.8491073131144427e-05, "loss": 0.851, "step": 5573 }, { "epoch": 0.20188337558855488, "grad_norm": 1.174649491973343, "learning_rate": 1.8490453434447085e-05, "loss": 0.752, "step": 5574 }, { "epoch": 0.20191959434987322, "grad_norm": 1.3294393857587683, "learning_rate": 1.8489833620913644e-05, "loss": 0.8391, "step": 5575 }, { "epoch": 0.2019558131111916, "grad_norm": 1.562405920772677, "learning_rate": 1.8489213690552627e-05, "loss": 0.8378, "step": 5576 }, { "epoch": 0.20199203187250997, "grad_norm": 1.5090573093998871, "learning_rate": 1.8488593643372574e-05, "loss": 0.8422, "step": 5577 }, { "epoch": 0.20202825063382832, "grad_norm": 1.55627075465373, "learning_rate": 1.8487973479382013e-05, "loss": 0.8166, "step": 5578 }, { "epoch": 0.2020644693951467, "grad_norm": 1.5721656341033061, "learning_rate": 1.848735319858948e-05, "loss": 0.8429, "step": 5579 }, { "epoch": 0.20210068815646504, "grad_norm": 1.6230707717678705, "learning_rate": 1.8486732801003508e-05, "loss": 0.8322, "step": 5580 }, { "epoch": 0.2021369069177834, "grad_norm": 1.3869437485152694, "learning_rate": 1.8486112286632634e-05, "loss": 0.8084, "step": 5581 }, { "epoch": 0.20217312567910178, "grad_norm": 1.4209050719116114, "learning_rate": 1.84854916554854e-05, "loss": 0.788, "step": 5582 }, { "epoch": 0.20220934444042013, "grad_norm": 1.662426623539274, "learning_rate": 1.848487090757035e-05, "loss": 0.7835, "step": 5583 }, { "epoch": 0.2022455632017385, "grad_norm": 1.392736417246018, "learning_rate": 1.8484250042896014e-05, "loss": 0.8242, "step": 5584 }, { "epoch": 0.20228178196305685, "grad_norm": 1.2883631604580676, "learning_rate": 1.8483629061470946e-05, "loss": 0.7528, "step": 5585 }, { "epoch": 0.20231800072437522, "grad_norm": 1.5267119393564257, "learning_rate": 1.8483007963303684e-05, "loss": 0.8716, "step": 5586 }, { "epoch": 0.2023542194856936, "grad_norm": 1.519913649800245, "learning_rate": 1.8482386748402784e-05, "loss": 0.7974, "step": 5587 }, { "epoch": 0.20239043824701194, "grad_norm": 1.3265269707189233, "learning_rate": 1.8481765416776786e-05, "loss": 0.8145, "step": 5588 }, { "epoch": 0.20242665700833032, "grad_norm": 1.3364754487447175, "learning_rate": 1.8481143968434242e-05, "loss": 0.796, "step": 5589 }, { "epoch": 0.20246287576964866, "grad_norm": 1.5464460616293139, "learning_rate": 1.8480522403383704e-05, "loss": 0.8423, "step": 5590 }, { "epoch": 0.20249909453096704, "grad_norm": 1.2492859481018537, "learning_rate": 1.847990072163373e-05, "loss": 0.8387, "step": 5591 }, { "epoch": 0.2025353132922854, "grad_norm": 1.3690705832443246, "learning_rate": 1.847927892319287e-05, "loss": 0.7774, "step": 5592 }, { "epoch": 0.20257153205360376, "grad_norm": 1.4240289384282836, "learning_rate": 1.847865700806968e-05, "loss": 0.7758, "step": 5593 }, { "epoch": 0.20260775081492213, "grad_norm": 1.1072111213575193, "learning_rate": 1.8478034976272717e-05, "loss": 0.7479, "step": 5594 }, { "epoch": 0.20264396957624048, "grad_norm": 1.5558343328385225, "learning_rate": 1.8477412827810547e-05, "loss": 0.8011, "step": 5595 }, { "epoch": 0.20268018833755885, "grad_norm": 1.5391943174397391, "learning_rate": 1.847679056269173e-05, "loss": 0.8428, "step": 5596 }, { "epoch": 0.20271640709887723, "grad_norm": 1.55591499196045, "learning_rate": 1.847616818092482e-05, "loss": 0.7788, "step": 5597 }, { "epoch": 0.20275262586019557, "grad_norm": 1.1732923862814881, "learning_rate": 1.8475545682518393e-05, "loss": 0.7565, "step": 5598 }, { "epoch": 0.20278884462151395, "grad_norm": 1.5314835521125811, "learning_rate": 1.847492306748101e-05, "loss": 0.7266, "step": 5599 }, { "epoch": 0.2028250633828323, "grad_norm": 1.372234381508176, "learning_rate": 1.8474300335821235e-05, "loss": 0.7984, "step": 5600 }, { "epoch": 0.20286128214415067, "grad_norm": 1.090973737833145, "learning_rate": 1.8473677487547644e-05, "loss": 0.7822, "step": 5601 }, { "epoch": 0.20289750090546904, "grad_norm": 1.4440253931741682, "learning_rate": 1.8473054522668804e-05, "loss": 0.8866, "step": 5602 }, { "epoch": 0.2029337196667874, "grad_norm": 1.5324022485340176, "learning_rate": 1.8472431441193293e-05, "loss": 0.8321, "step": 5603 }, { "epoch": 0.20296993842810576, "grad_norm": 1.4210585408937224, "learning_rate": 1.8471808243129678e-05, "loss": 0.885, "step": 5604 }, { "epoch": 0.2030061571894241, "grad_norm": 1.539056860845662, "learning_rate": 1.8471184928486536e-05, "loss": 0.9105, "step": 5605 }, { "epoch": 0.20304237595074248, "grad_norm": 1.691566248279466, "learning_rate": 1.8470561497272446e-05, "loss": 0.8643, "step": 5606 }, { "epoch": 0.20307859471206086, "grad_norm": 1.397250041852778, "learning_rate": 1.8469937949495987e-05, "loss": 0.8444, "step": 5607 }, { "epoch": 0.2031148134733792, "grad_norm": 1.1281273432319945, "learning_rate": 1.8469314285165742e-05, "loss": 0.8303, "step": 5608 }, { "epoch": 0.20315103223469758, "grad_norm": 1.4213148227960941, "learning_rate": 1.846869050429029e-05, "loss": 0.7942, "step": 5609 }, { "epoch": 0.20318725099601595, "grad_norm": 1.870303662119088, "learning_rate": 1.846806660687821e-05, "loss": 0.8299, "step": 5610 }, { "epoch": 0.2032234697573343, "grad_norm": 1.5788015813170966, "learning_rate": 1.84674425929381e-05, "loss": 0.8227, "step": 5611 }, { "epoch": 0.20325968851865267, "grad_norm": 1.3197762194453089, "learning_rate": 1.8466818462478536e-05, "loss": 0.7776, "step": 5612 }, { "epoch": 0.20329590727997102, "grad_norm": 1.6938490660049321, "learning_rate": 1.8466194215508112e-05, "loss": 0.8322, "step": 5613 }, { "epoch": 0.2033321260412894, "grad_norm": 1.3005759260437046, "learning_rate": 1.8465569852035415e-05, "loss": 0.7212, "step": 5614 }, { "epoch": 0.20336834480260776, "grad_norm": 1.5404762972975634, "learning_rate": 1.8464945372069037e-05, "loss": 0.846, "step": 5615 }, { "epoch": 0.2034045635639261, "grad_norm": 1.441312509036851, "learning_rate": 1.8464320775617577e-05, "loss": 0.7325, "step": 5616 }, { "epoch": 0.20344078232524448, "grad_norm": 1.5499272222188494, "learning_rate": 1.8463696062689623e-05, "loss": 0.8324, "step": 5617 }, { "epoch": 0.20347700108656283, "grad_norm": 1.3391163939777484, "learning_rate": 1.8463071233293775e-05, "loss": 0.8454, "step": 5618 }, { "epoch": 0.2035132198478812, "grad_norm": 1.4141629873908226, "learning_rate": 1.8462446287438632e-05, "loss": 0.8393, "step": 5619 }, { "epoch": 0.20354943860919958, "grad_norm": 1.5878472240142574, "learning_rate": 1.846182122513279e-05, "loss": 0.8688, "step": 5620 }, { "epoch": 0.20358565737051793, "grad_norm": 1.1428091088266803, "learning_rate": 1.8461196046384853e-05, "loss": 0.7494, "step": 5621 }, { "epoch": 0.2036218761318363, "grad_norm": 1.4961504676466166, "learning_rate": 1.8460570751203423e-05, "loss": 0.856, "step": 5622 }, { "epoch": 0.20365809489315465, "grad_norm": 1.6028566841893561, "learning_rate": 1.845994533959711e-05, "loss": 0.9279, "step": 5623 }, { "epoch": 0.20369431365447302, "grad_norm": 1.26744714625609, "learning_rate": 1.8459319811574508e-05, "loss": 0.7884, "step": 5624 }, { "epoch": 0.2037305324157914, "grad_norm": 1.5257749061237733, "learning_rate": 1.845869416714424e-05, "loss": 0.8368, "step": 5625 }, { "epoch": 0.20376675117710974, "grad_norm": 1.489798754021099, "learning_rate": 1.8458068406314903e-05, "loss": 0.7616, "step": 5626 }, { "epoch": 0.2038029699384281, "grad_norm": 1.2821657234627928, "learning_rate": 1.8457442529095114e-05, "loss": 0.7469, "step": 5627 }, { "epoch": 0.20383918869974646, "grad_norm": 1.5604185438325215, "learning_rate": 1.8456816535493484e-05, "loss": 0.7955, "step": 5628 }, { "epoch": 0.20387540746106483, "grad_norm": 1.2584585316493826, "learning_rate": 1.845619042551863e-05, "loss": 0.8105, "step": 5629 }, { "epoch": 0.2039116262223832, "grad_norm": 1.41596530101406, "learning_rate": 1.8455564199179162e-05, "loss": 0.8515, "step": 5630 }, { "epoch": 0.20394784498370155, "grad_norm": 1.4828322876553564, "learning_rate": 1.8454937856483703e-05, "loss": 0.8222, "step": 5631 }, { "epoch": 0.20398406374501993, "grad_norm": 1.3953712902467013, "learning_rate": 1.8454311397440873e-05, "loss": 0.8167, "step": 5632 }, { "epoch": 0.20402028250633827, "grad_norm": 1.577439832193608, "learning_rate": 1.8453684822059284e-05, "loss": 0.8253, "step": 5633 }, { "epoch": 0.20405650126765665, "grad_norm": 1.5522465545375221, "learning_rate": 1.8453058130347565e-05, "loss": 0.8127, "step": 5634 }, { "epoch": 0.20409272002897502, "grad_norm": 1.4311024687401794, "learning_rate": 1.845243132231434e-05, "loss": 0.7821, "step": 5635 }, { "epoch": 0.20412893879029337, "grad_norm": 1.138047319196951, "learning_rate": 1.845180439796823e-05, "loss": 0.8002, "step": 5636 }, { "epoch": 0.20416515755161174, "grad_norm": 1.8807806501619735, "learning_rate": 1.8451177357317868e-05, "loss": 0.8812, "step": 5637 }, { "epoch": 0.2042013763129301, "grad_norm": 1.5286481511771606, "learning_rate": 1.845055020037188e-05, "loss": 0.7609, "step": 5638 }, { "epoch": 0.20423759507424846, "grad_norm": 1.4666865684491526, "learning_rate": 1.8449922927138897e-05, "loss": 0.8298, "step": 5639 }, { "epoch": 0.20427381383556684, "grad_norm": 1.473896988206021, "learning_rate": 1.844929553762755e-05, "loss": 0.8067, "step": 5640 }, { "epoch": 0.20431003259688518, "grad_norm": 1.4528659293513866, "learning_rate": 1.8448668031846468e-05, "loss": 0.8473, "step": 5641 }, { "epoch": 0.20434625135820356, "grad_norm": 1.5452759441896473, "learning_rate": 1.8448040409804294e-05, "loss": 0.8583, "step": 5642 }, { "epoch": 0.2043824701195219, "grad_norm": 1.540995867255627, "learning_rate": 1.844741267150966e-05, "loss": 0.7707, "step": 5643 }, { "epoch": 0.20441868888084028, "grad_norm": 1.3714249345409093, "learning_rate": 1.8446784816971203e-05, "loss": 0.8279, "step": 5644 }, { "epoch": 0.20445490764215865, "grad_norm": 1.4924534091394661, "learning_rate": 1.8446156846197565e-05, "loss": 0.8685, "step": 5645 }, { "epoch": 0.204491126403477, "grad_norm": 1.3850799770031925, "learning_rate": 1.8445528759197387e-05, "loss": 0.8131, "step": 5646 }, { "epoch": 0.20452734516479537, "grad_norm": 1.3900617285152548, "learning_rate": 1.8444900555979318e-05, "loss": 0.8406, "step": 5647 }, { "epoch": 0.20456356392611372, "grad_norm": 1.4239910468149626, "learning_rate": 1.844427223655199e-05, "loss": 0.8179, "step": 5648 }, { "epoch": 0.2045997826874321, "grad_norm": 1.404463000810562, "learning_rate": 1.8443643800924063e-05, "loss": 0.8034, "step": 5649 }, { "epoch": 0.20463600144875047, "grad_norm": 1.2178776362051222, "learning_rate": 1.8443015249104174e-05, "loss": 0.7848, "step": 5650 }, { "epoch": 0.2046722202100688, "grad_norm": 1.5352440486193784, "learning_rate": 1.8442386581100977e-05, "loss": 0.8904, "step": 5651 }, { "epoch": 0.20470843897138719, "grad_norm": 1.5118987028078221, "learning_rate": 1.844175779692312e-05, "loss": 0.8579, "step": 5652 }, { "epoch": 0.20474465773270553, "grad_norm": 1.5424955294249063, "learning_rate": 1.8441128896579264e-05, "loss": 0.8931, "step": 5653 }, { "epoch": 0.2047808764940239, "grad_norm": 1.4237601636174464, "learning_rate": 1.8440499880078052e-05, "loss": 0.8655, "step": 5654 }, { "epoch": 0.20481709525534228, "grad_norm": 2.1167070052059795, "learning_rate": 1.8439870747428148e-05, "loss": 0.8499, "step": 5655 }, { "epoch": 0.20485331401666063, "grad_norm": 1.2674492470529886, "learning_rate": 1.8439241498638206e-05, "loss": 0.842, "step": 5656 }, { "epoch": 0.204889532777979, "grad_norm": 1.3957588725429797, "learning_rate": 1.8438612133716883e-05, "loss": 0.7337, "step": 5657 }, { "epoch": 0.20492575153929735, "grad_norm": 1.384396955243897, "learning_rate": 1.8437982652672848e-05, "loss": 0.7788, "step": 5658 }, { "epoch": 0.20496197030061572, "grad_norm": 1.4113156487309118, "learning_rate": 1.843735305551475e-05, "loss": 0.7687, "step": 5659 }, { "epoch": 0.2049981890619341, "grad_norm": 1.5122563369788158, "learning_rate": 1.8436723342251267e-05, "loss": 0.8159, "step": 5660 }, { "epoch": 0.20503440782325244, "grad_norm": 1.4037051088088182, "learning_rate": 1.8436093512891052e-05, "loss": 0.877, "step": 5661 }, { "epoch": 0.20507062658457081, "grad_norm": 1.5044603138389527, "learning_rate": 1.8435463567442783e-05, "loss": 0.8787, "step": 5662 }, { "epoch": 0.20510684534588916, "grad_norm": 1.5239592221253786, "learning_rate": 1.8434833505915123e-05, "loss": 0.8495, "step": 5663 }, { "epoch": 0.20514306410720753, "grad_norm": 1.6444791742538152, "learning_rate": 1.843420332831674e-05, "loss": 0.8434, "step": 5664 }, { "epoch": 0.2051792828685259, "grad_norm": 1.367688353622706, "learning_rate": 1.8433573034656308e-05, "loss": 0.8081, "step": 5665 }, { "epoch": 0.20521550162984425, "grad_norm": 1.6030745869698098, "learning_rate": 1.84329426249425e-05, "loss": 0.8395, "step": 5666 }, { "epoch": 0.20525172039116263, "grad_norm": 1.5380199236554397, "learning_rate": 1.8432312099183993e-05, "loss": 0.8224, "step": 5667 }, { "epoch": 0.20528793915248097, "grad_norm": 1.4515001665221015, "learning_rate": 1.8431681457389463e-05, "loss": 0.8555, "step": 5668 }, { "epoch": 0.20532415791379935, "grad_norm": 1.4709180653313179, "learning_rate": 1.8431050699567585e-05, "loss": 0.8111, "step": 5669 }, { "epoch": 0.20536037667511772, "grad_norm": 1.421777079296788, "learning_rate": 1.8430419825727044e-05, "loss": 0.815, "step": 5670 }, { "epoch": 0.20539659543643607, "grad_norm": 1.367650482940674, "learning_rate": 1.8429788835876515e-05, "loss": 0.7677, "step": 5671 }, { "epoch": 0.20543281419775444, "grad_norm": 1.5186331400584037, "learning_rate": 1.842915773002469e-05, "loss": 0.8204, "step": 5672 }, { "epoch": 0.2054690329590728, "grad_norm": 1.4727165692975979, "learning_rate": 1.8428526508180245e-05, "loss": 0.8091, "step": 5673 }, { "epoch": 0.20550525172039116, "grad_norm": 1.3863106429297543, "learning_rate": 1.8427895170351868e-05, "loss": 0.7728, "step": 5674 }, { "epoch": 0.20554147048170954, "grad_norm": 1.4989572106558877, "learning_rate": 1.8427263716548248e-05, "loss": 0.8311, "step": 5675 }, { "epoch": 0.20557768924302788, "grad_norm": 1.3242174875765385, "learning_rate": 1.8426632146778075e-05, "loss": 0.7961, "step": 5676 }, { "epoch": 0.20561390800434626, "grad_norm": 1.709691817860911, "learning_rate": 1.8426000461050045e-05, "loss": 0.9174, "step": 5677 }, { "epoch": 0.2056501267656646, "grad_norm": 1.4217913807395406, "learning_rate": 1.842536865937284e-05, "loss": 0.8024, "step": 5678 }, { "epoch": 0.20568634552698298, "grad_norm": 1.5525982374214888, "learning_rate": 1.8424736741755156e-05, "loss": 0.8553, "step": 5679 }, { "epoch": 0.20572256428830135, "grad_norm": 1.4513904899746017, "learning_rate": 1.8424104708205694e-05, "loss": 0.8368, "step": 5680 }, { "epoch": 0.2057587830496197, "grad_norm": 1.1696133281346848, "learning_rate": 1.842347255873315e-05, "loss": 0.7581, "step": 5681 }, { "epoch": 0.20579500181093807, "grad_norm": 0.9839214659582464, "learning_rate": 1.842284029334622e-05, "loss": 0.7788, "step": 5682 }, { "epoch": 0.20583122057225642, "grad_norm": 1.576194703930641, "learning_rate": 1.8422207912053612e-05, "loss": 0.769, "step": 5683 }, { "epoch": 0.2058674393335748, "grad_norm": 1.525372497753716, "learning_rate": 1.842157541486402e-05, "loss": 0.8115, "step": 5684 }, { "epoch": 0.20590365809489317, "grad_norm": 1.2220171106804154, "learning_rate": 1.842094280178615e-05, "loss": 0.707, "step": 5685 }, { "epoch": 0.2059398768562115, "grad_norm": 1.5376514038984082, "learning_rate": 1.8420310072828707e-05, "loss": 0.8567, "step": 5686 }, { "epoch": 0.2059760956175299, "grad_norm": 1.726604506469241, "learning_rate": 1.84196772280004e-05, "loss": 0.8169, "step": 5687 }, { "epoch": 0.20601231437884823, "grad_norm": 1.5640021428982043, "learning_rate": 1.841904426730994e-05, "loss": 0.8513, "step": 5688 }, { "epoch": 0.2060485331401666, "grad_norm": 1.3127510519521526, "learning_rate": 1.8418411190766026e-05, "loss": 0.8637, "step": 5689 }, { "epoch": 0.20608475190148498, "grad_norm": 1.4958942270325517, "learning_rate": 1.841777799837738e-05, "loss": 0.845, "step": 5690 }, { "epoch": 0.20612097066280333, "grad_norm": 1.1806184885696527, "learning_rate": 1.8417144690152712e-05, "loss": 0.8044, "step": 5691 }, { "epoch": 0.2061571894241217, "grad_norm": 1.5250031117012766, "learning_rate": 1.8416511266100736e-05, "loss": 0.8292, "step": 5692 }, { "epoch": 0.20619340818544005, "grad_norm": 1.4238808087261776, "learning_rate": 1.841587772623017e-05, "loss": 0.7692, "step": 5693 }, { "epoch": 0.20622962694675842, "grad_norm": 1.4164506322973798, "learning_rate": 1.8415244070549733e-05, "loss": 0.8601, "step": 5694 }, { "epoch": 0.2062658457080768, "grad_norm": 1.5920043898541751, "learning_rate": 1.8414610299068142e-05, "loss": 0.7689, "step": 5695 }, { "epoch": 0.20630206446939514, "grad_norm": 1.5001327976163292, "learning_rate": 1.841397641179412e-05, "loss": 0.8237, "step": 5696 }, { "epoch": 0.20633828323071352, "grad_norm": 1.3552070661896105, "learning_rate": 1.8413342408736393e-05, "loss": 0.7696, "step": 5697 }, { "epoch": 0.20637450199203186, "grad_norm": 1.5754373752279396, "learning_rate": 1.8412708289903676e-05, "loss": 0.8719, "step": 5698 }, { "epoch": 0.20641072075335024, "grad_norm": 1.496063222150369, "learning_rate": 1.8412074055304704e-05, "loss": 0.7955, "step": 5699 }, { "epoch": 0.2064469395146686, "grad_norm": 1.4612092438578705, "learning_rate": 1.84114397049482e-05, "loss": 0.7942, "step": 5700 }, { "epoch": 0.20648315827598696, "grad_norm": 1.499384873148193, "learning_rate": 1.8410805238842898e-05, "loss": 0.8127, "step": 5701 }, { "epoch": 0.20651937703730533, "grad_norm": 1.3561098598501728, "learning_rate": 1.841017065699752e-05, "loss": 0.7827, "step": 5702 }, { "epoch": 0.20655559579862368, "grad_norm": 1.5846944856461493, "learning_rate": 1.840953595942081e-05, "loss": 0.8402, "step": 5703 }, { "epoch": 0.20659181455994205, "grad_norm": 1.5831924992018165, "learning_rate": 1.8408901146121493e-05, "loss": 0.9095, "step": 5704 }, { "epoch": 0.20662803332126042, "grad_norm": 1.392043445098918, "learning_rate": 1.8408266217108305e-05, "loss": 0.8342, "step": 5705 }, { "epoch": 0.20666425208257877, "grad_norm": 1.59863167065374, "learning_rate": 1.840763117238999e-05, "loss": 0.762, "step": 5706 }, { "epoch": 0.20670047084389714, "grad_norm": 1.370910246434474, "learning_rate": 1.840699601197528e-05, "loss": 0.7719, "step": 5707 }, { "epoch": 0.2067366896052155, "grad_norm": 1.8676431012255252, "learning_rate": 1.840636073587292e-05, "loss": 0.7712, "step": 5708 }, { "epoch": 0.20677290836653386, "grad_norm": 1.4604709385130772, "learning_rate": 1.840572534409165e-05, "loss": 0.7777, "step": 5709 }, { "epoch": 0.20680912712785224, "grad_norm": 1.515386614289994, "learning_rate": 1.840508983664021e-05, "loss": 0.7853, "step": 5710 }, { "epoch": 0.20684534588917058, "grad_norm": 1.3677369212682888, "learning_rate": 1.840445421352735e-05, "loss": 0.8016, "step": 5711 }, { "epoch": 0.20688156465048896, "grad_norm": 1.430735750322663, "learning_rate": 1.8403818474761815e-05, "loss": 0.841, "step": 5712 }, { "epoch": 0.2069177834118073, "grad_norm": 1.3714638434942112, "learning_rate": 1.8403182620352352e-05, "loss": 0.7082, "step": 5713 }, { "epoch": 0.20695400217312568, "grad_norm": 1.506141826286465, "learning_rate": 1.8402546650307717e-05, "loss": 0.8166, "step": 5714 }, { "epoch": 0.20699022093444405, "grad_norm": 1.5094637797291413, "learning_rate": 1.8401910564636653e-05, "loss": 0.82, "step": 5715 }, { "epoch": 0.2070264396957624, "grad_norm": 1.39205505811996, "learning_rate": 1.8401274363347917e-05, "loss": 0.785, "step": 5716 }, { "epoch": 0.20706265845708077, "grad_norm": 1.4497974145299075, "learning_rate": 1.8400638046450266e-05, "loss": 0.7985, "step": 5717 }, { "epoch": 0.20709887721839912, "grad_norm": 1.6812742079973768, "learning_rate": 1.840000161395245e-05, "loss": 0.8179, "step": 5718 }, { "epoch": 0.2071350959797175, "grad_norm": 1.550328286436991, "learning_rate": 1.839936506586324e-05, "loss": 0.8066, "step": 5719 }, { "epoch": 0.20717131474103587, "grad_norm": 1.5741737858619456, "learning_rate": 1.8398728402191376e-05, "loss": 0.8467, "step": 5720 }, { "epoch": 0.2072075335023542, "grad_norm": 1.4173184340660143, "learning_rate": 1.8398091622945636e-05, "loss": 0.7962, "step": 5721 }, { "epoch": 0.2072437522636726, "grad_norm": 1.5141437926972041, "learning_rate": 1.8397454728134775e-05, "loss": 0.7743, "step": 5722 }, { "epoch": 0.20727997102499093, "grad_norm": 1.6319410757076267, "learning_rate": 1.8396817717767557e-05, "loss": 0.8148, "step": 5723 }, { "epoch": 0.2073161897863093, "grad_norm": 1.6363134235059507, "learning_rate": 1.839618059185275e-05, "loss": 0.8769, "step": 5724 }, { "epoch": 0.20735240854762768, "grad_norm": 1.3290450927095363, "learning_rate": 1.8395543350399117e-05, "loss": 0.7771, "step": 5725 }, { "epoch": 0.20738862730894603, "grad_norm": 1.5336071004592524, "learning_rate": 1.8394905993415436e-05, "loss": 0.8346, "step": 5726 }, { "epoch": 0.2074248460702644, "grad_norm": 1.4790501352756893, "learning_rate": 1.8394268520910467e-05, "loss": 0.8476, "step": 5727 }, { "epoch": 0.20746106483158275, "grad_norm": 1.3237750644974278, "learning_rate": 1.8393630932892992e-05, "loss": 0.8226, "step": 5728 }, { "epoch": 0.20749728359290112, "grad_norm": 1.9218380544569365, "learning_rate": 1.8392993229371774e-05, "loss": 0.934, "step": 5729 }, { "epoch": 0.2075335023542195, "grad_norm": 1.5756762161294948, "learning_rate": 1.83923554103556e-05, "loss": 0.8448, "step": 5730 }, { "epoch": 0.20756972111553784, "grad_norm": 1.3885090533973348, "learning_rate": 1.8391717475853242e-05, "loss": 0.8762, "step": 5731 }, { "epoch": 0.20760593987685622, "grad_norm": 1.5055993674518342, "learning_rate": 1.8391079425873475e-05, "loss": 0.8883, "step": 5732 }, { "epoch": 0.20764215863817456, "grad_norm": 1.6659622116739088, "learning_rate": 1.8390441260425082e-05, "loss": 0.9254, "step": 5733 }, { "epoch": 0.20767837739949294, "grad_norm": 1.6004473630304263, "learning_rate": 1.8389802979516846e-05, "loss": 0.9148, "step": 5734 }, { "epoch": 0.2077145961608113, "grad_norm": 1.5768438464370924, "learning_rate": 1.838916458315755e-05, "loss": 0.8383, "step": 5735 }, { "epoch": 0.20775081492212966, "grad_norm": 1.558632839401068, "learning_rate": 1.8388526071355976e-05, "loss": 0.8443, "step": 5736 }, { "epoch": 0.20778703368344803, "grad_norm": 1.5711543094756488, "learning_rate": 1.8387887444120914e-05, "loss": 0.8027, "step": 5737 }, { "epoch": 0.20782325244476638, "grad_norm": 1.4986385944920608, "learning_rate": 1.838724870146115e-05, "loss": 0.7512, "step": 5738 }, { "epoch": 0.20785947120608475, "grad_norm": 1.3324978633766977, "learning_rate": 1.8386609843385473e-05, "loss": 0.8554, "step": 5739 }, { "epoch": 0.20789568996740312, "grad_norm": 1.4775328304726, "learning_rate": 1.8385970869902675e-05, "loss": 0.7789, "step": 5740 }, { "epoch": 0.20793190872872147, "grad_norm": 1.1486227995408989, "learning_rate": 1.838533178102155e-05, "loss": 0.7465, "step": 5741 }, { "epoch": 0.20796812749003984, "grad_norm": 1.5227259001012372, "learning_rate": 1.8384692576750892e-05, "loss": 0.7416, "step": 5742 }, { "epoch": 0.2080043462513582, "grad_norm": 1.16030120134744, "learning_rate": 1.8384053257099498e-05, "loss": 0.7397, "step": 5743 }, { "epoch": 0.20804056501267656, "grad_norm": 1.5557033745778848, "learning_rate": 1.838341382207616e-05, "loss": 0.8138, "step": 5744 }, { "epoch": 0.20807678377399494, "grad_norm": 1.6865001095923866, "learning_rate": 1.838277427168969e-05, "loss": 0.812, "step": 5745 }, { "epoch": 0.20811300253531329, "grad_norm": 1.762952560716288, "learning_rate": 1.8382134605948872e-05, "loss": 0.8099, "step": 5746 }, { "epoch": 0.20814922129663166, "grad_norm": 1.3156932621262227, "learning_rate": 1.838149482486252e-05, "loss": 0.8637, "step": 5747 }, { "epoch": 0.20818544005795, "grad_norm": 1.4354394989338586, "learning_rate": 1.8380854928439434e-05, "loss": 0.7916, "step": 5748 }, { "epoch": 0.20822165881926838, "grad_norm": 1.4299882431772415, "learning_rate": 1.838021491668842e-05, "loss": 0.7546, "step": 5749 }, { "epoch": 0.20825787758058675, "grad_norm": 1.273565425345623, "learning_rate": 1.8379574789618284e-05, "loss": 0.7961, "step": 5750 }, { "epoch": 0.2082940963419051, "grad_norm": 1.4648242752052194, "learning_rate": 1.837893454723784e-05, "loss": 0.8159, "step": 5751 }, { "epoch": 0.20833031510322347, "grad_norm": 1.4370787964906673, "learning_rate": 1.837829418955589e-05, "loss": 0.7876, "step": 5752 }, { "epoch": 0.20836653386454182, "grad_norm": 1.5333223430180936, "learning_rate": 1.837765371658125e-05, "loss": 0.7947, "step": 5753 }, { "epoch": 0.2084027526258602, "grad_norm": 1.5697593901019675, "learning_rate": 1.8377013128322736e-05, "loss": 0.787, "step": 5754 }, { "epoch": 0.20843897138717857, "grad_norm": 1.4681482578861549, "learning_rate": 1.837637242478916e-05, "loss": 0.8393, "step": 5755 }, { "epoch": 0.20847519014849691, "grad_norm": 1.528091526883794, "learning_rate": 1.837573160598934e-05, "loss": 0.891, "step": 5756 }, { "epoch": 0.2085114089098153, "grad_norm": 1.6055286714179813, "learning_rate": 1.8375090671932093e-05, "loss": 0.8392, "step": 5757 }, { "epoch": 0.20854762767113363, "grad_norm": 1.2385562231259184, "learning_rate": 1.837444962262624e-05, "loss": 0.7611, "step": 5758 }, { "epoch": 0.208583846432452, "grad_norm": 1.485141667214668, "learning_rate": 1.83738084580806e-05, "loss": 0.8471, "step": 5759 }, { "epoch": 0.20862006519377038, "grad_norm": 1.4537486262255532, "learning_rate": 1.8373167178304e-05, "loss": 0.8842, "step": 5760 }, { "epoch": 0.20865628395508873, "grad_norm": 1.2170505293739244, "learning_rate": 1.837252578330526e-05, "loss": 0.8396, "step": 5761 }, { "epoch": 0.2086925027164071, "grad_norm": 1.4372687438833132, "learning_rate": 1.8371884273093213e-05, "loss": 0.905, "step": 5762 }, { "epoch": 0.20872872147772545, "grad_norm": 1.2757970357198314, "learning_rate": 1.8371242647676678e-05, "loss": 0.883, "step": 5763 }, { "epoch": 0.20876494023904382, "grad_norm": 1.4886616091898504, "learning_rate": 1.837060090706449e-05, "loss": 0.7955, "step": 5764 }, { "epoch": 0.2088011590003622, "grad_norm": 1.6480097023496183, "learning_rate": 1.8369959051265476e-05, "loss": 0.856, "step": 5765 }, { "epoch": 0.20883737776168054, "grad_norm": 1.4590563226549862, "learning_rate": 1.836931708028847e-05, "loss": 0.8219, "step": 5766 }, { "epoch": 0.20887359652299892, "grad_norm": 1.3599701459318958, "learning_rate": 1.8368674994142314e-05, "loss": 0.7589, "step": 5767 }, { "epoch": 0.20890981528431726, "grad_norm": 1.5710408184765607, "learning_rate": 1.836803279283583e-05, "loss": 0.8602, "step": 5768 }, { "epoch": 0.20894603404563564, "grad_norm": 1.5419586388173894, "learning_rate": 1.8367390476377864e-05, "loss": 0.8186, "step": 5769 }, { "epoch": 0.208982252806954, "grad_norm": 1.5520549907326011, "learning_rate": 1.8366748044777254e-05, "loss": 0.775, "step": 5770 }, { "epoch": 0.20901847156827236, "grad_norm": 1.10512104132706, "learning_rate": 1.8366105498042837e-05, "loss": 0.7986, "step": 5771 }, { "epoch": 0.20905469032959073, "grad_norm": 1.1995375163275475, "learning_rate": 1.836546283618346e-05, "loss": 0.7752, "step": 5772 }, { "epoch": 0.20909090909090908, "grad_norm": 1.4624101176875213, "learning_rate": 1.8364820059207964e-05, "loss": 0.7963, "step": 5773 }, { "epoch": 0.20912712785222745, "grad_norm": 1.425039457676273, "learning_rate": 1.836417716712519e-05, "loss": 0.8205, "step": 5774 }, { "epoch": 0.20916334661354583, "grad_norm": 1.1044698976675777, "learning_rate": 1.8363534159943996e-05, "loss": 0.7145, "step": 5775 }, { "epoch": 0.20919956537486417, "grad_norm": 4.721823746338669, "learning_rate": 1.8362891037673215e-05, "loss": 0.781, "step": 5776 }, { "epoch": 0.20923578413618255, "grad_norm": 1.4061921352826101, "learning_rate": 1.836224780032171e-05, "loss": 0.8681, "step": 5777 }, { "epoch": 0.2092720028975009, "grad_norm": 1.2736528202592623, "learning_rate": 1.8361604447898325e-05, "loss": 0.8215, "step": 5778 }, { "epoch": 0.20930822165881927, "grad_norm": 1.581817536577334, "learning_rate": 1.836096098041192e-05, "loss": 0.8489, "step": 5779 }, { "epoch": 0.20934444042013764, "grad_norm": 1.644065675502191, "learning_rate": 1.8360317397871342e-05, "loss": 0.9141, "step": 5780 }, { "epoch": 0.20938065918145599, "grad_norm": 1.5221620306642085, "learning_rate": 1.835967370028545e-05, "loss": 0.7828, "step": 5781 }, { "epoch": 0.20941687794277436, "grad_norm": 3.2549158478624696, "learning_rate": 1.8359029887663103e-05, "loss": 0.8083, "step": 5782 }, { "epoch": 0.20945309670409273, "grad_norm": 1.1440584746037346, "learning_rate": 1.835838596001316e-05, "loss": 0.7665, "step": 5783 }, { "epoch": 0.20948931546541108, "grad_norm": 1.7262053810798994, "learning_rate": 1.835774191734448e-05, "loss": 0.7824, "step": 5784 }, { "epoch": 0.20952553422672945, "grad_norm": 1.661963973089967, "learning_rate": 1.835709775966593e-05, "loss": 0.9301, "step": 5785 }, { "epoch": 0.2095617529880478, "grad_norm": 1.5051121453440868, "learning_rate": 1.835645348698637e-05, "loss": 0.7791, "step": 5786 }, { "epoch": 0.20959797174936617, "grad_norm": 1.515734754599747, "learning_rate": 1.835580909931467e-05, "loss": 0.8989, "step": 5787 }, { "epoch": 0.20963419051068455, "grad_norm": 1.4356298695142482, "learning_rate": 1.8355164596659693e-05, "loss": 0.8231, "step": 5788 }, { "epoch": 0.2096704092720029, "grad_norm": 1.3896051547225574, "learning_rate": 1.835451997903031e-05, "loss": 0.7395, "step": 5789 }, { "epoch": 0.20970662803332127, "grad_norm": 1.6251004879599402, "learning_rate": 1.835387524643539e-05, "loss": 0.8041, "step": 5790 }, { "epoch": 0.20974284679463961, "grad_norm": 1.545669462798097, "learning_rate": 1.8353230398883805e-05, "loss": 0.9067, "step": 5791 }, { "epoch": 0.209779065555958, "grad_norm": 1.460775147847975, "learning_rate": 1.8352585436384432e-05, "loss": 0.8044, "step": 5792 }, { "epoch": 0.20981528431727636, "grad_norm": 1.368963425334546, "learning_rate": 1.8351940358946145e-05, "loss": 0.8033, "step": 5793 }, { "epoch": 0.2098515030785947, "grad_norm": 1.1204194696290006, "learning_rate": 1.835129516657782e-05, "loss": 0.7692, "step": 5794 }, { "epoch": 0.20988772183991308, "grad_norm": 1.402291192661099, "learning_rate": 1.8350649859288336e-05, "loss": 0.6811, "step": 5795 }, { "epoch": 0.20992394060123143, "grad_norm": 1.323484955227338, "learning_rate": 1.835000443708657e-05, "loss": 0.8219, "step": 5796 }, { "epoch": 0.2099601593625498, "grad_norm": 1.613293193976129, "learning_rate": 1.834935889998141e-05, "loss": 0.7997, "step": 5797 }, { "epoch": 0.20999637812386818, "grad_norm": 1.7265995763880018, "learning_rate": 1.834871324798173e-05, "loss": 0.8132, "step": 5798 }, { "epoch": 0.21003259688518652, "grad_norm": 1.5075931401860865, "learning_rate": 1.8348067481096423e-05, "loss": 0.8319, "step": 5799 }, { "epoch": 0.2100688156465049, "grad_norm": 1.8179133489976165, "learning_rate": 1.8347421599334368e-05, "loss": 0.8358, "step": 5800 }, { "epoch": 0.21010503440782324, "grad_norm": 1.5796432863419057, "learning_rate": 1.8346775602704464e-05, "loss": 0.9202, "step": 5801 }, { "epoch": 0.21014125316914162, "grad_norm": 1.8099302972398463, "learning_rate": 1.834612949121559e-05, "loss": 0.8723, "step": 5802 }, { "epoch": 0.21017747193046, "grad_norm": 1.60501731317309, "learning_rate": 1.834548326487664e-05, "loss": 0.8621, "step": 5803 }, { "epoch": 0.21021369069177834, "grad_norm": 1.6878347870521169, "learning_rate": 1.8344836923696506e-05, "loss": 0.9204, "step": 5804 }, { "epoch": 0.2102499094530967, "grad_norm": 1.516609022388669, "learning_rate": 1.8344190467684087e-05, "loss": 0.7987, "step": 5805 }, { "epoch": 0.21028612821441506, "grad_norm": 1.4986203617845002, "learning_rate": 1.8343543896848275e-05, "loss": 0.8063, "step": 5806 }, { "epoch": 0.21032234697573343, "grad_norm": 1.4377302524428983, "learning_rate": 1.8342897211197964e-05, "loss": 0.7467, "step": 5807 }, { "epoch": 0.2103585657370518, "grad_norm": 1.559028740464031, "learning_rate": 1.834225041074206e-05, "loss": 0.837, "step": 5808 }, { "epoch": 0.21039478449837015, "grad_norm": 1.4136547691573067, "learning_rate": 1.834160349548946e-05, "loss": 0.7924, "step": 5809 }, { "epoch": 0.21043100325968853, "grad_norm": 1.4995294164310704, "learning_rate": 1.8340956465449066e-05, "loss": 0.7774, "step": 5810 }, { "epoch": 0.21046722202100687, "grad_norm": 1.540291285376477, "learning_rate": 1.834030932062978e-05, "loss": 0.8335, "step": 5811 }, { "epoch": 0.21050344078232525, "grad_norm": 1.509250140801152, "learning_rate": 1.833966206104051e-05, "loss": 0.8366, "step": 5812 }, { "epoch": 0.21053965954364362, "grad_norm": 1.395839951953806, "learning_rate": 1.8339014686690166e-05, "loss": 0.7897, "step": 5813 }, { "epoch": 0.21057587830496197, "grad_norm": 1.365354975706835, "learning_rate": 1.833836719758765e-05, "loss": 0.8204, "step": 5814 }, { "epoch": 0.21061209706628034, "grad_norm": 1.079224932207806, "learning_rate": 1.8337719593741874e-05, "loss": 0.8127, "step": 5815 }, { "epoch": 0.2106483158275987, "grad_norm": 1.134733963514822, "learning_rate": 1.833707187516175e-05, "loss": 0.7918, "step": 5816 }, { "epoch": 0.21068453458891706, "grad_norm": 1.3696298990229037, "learning_rate": 1.8336424041856193e-05, "loss": 0.7886, "step": 5817 }, { "epoch": 0.21072075335023543, "grad_norm": 1.575010550176011, "learning_rate": 1.8335776093834116e-05, "loss": 0.8485, "step": 5818 }, { "epoch": 0.21075697211155378, "grad_norm": 1.4988413659774258, "learning_rate": 1.8335128031104436e-05, "loss": 0.7906, "step": 5819 }, { "epoch": 0.21079319087287215, "grad_norm": 1.4419398295840602, "learning_rate": 1.8334479853676065e-05, "loss": 0.8056, "step": 5820 }, { "epoch": 0.2108294096341905, "grad_norm": 1.2590453215450987, "learning_rate": 1.8333831561557933e-05, "loss": 0.759, "step": 5821 }, { "epoch": 0.21086562839550888, "grad_norm": 1.522517562924053, "learning_rate": 1.8333183154758957e-05, "loss": 0.8587, "step": 5822 }, { "epoch": 0.21090184715682725, "grad_norm": 1.4108066874967677, "learning_rate": 1.8332534633288056e-05, "loss": 0.8426, "step": 5823 }, { "epoch": 0.2109380659181456, "grad_norm": 1.1941357776877577, "learning_rate": 1.8331885997154154e-05, "loss": 0.859, "step": 5824 }, { "epoch": 0.21097428467946397, "grad_norm": 1.4343109032650165, "learning_rate": 1.8331237246366183e-05, "loss": 0.8923, "step": 5825 }, { "epoch": 0.21101050344078232, "grad_norm": 1.3545276383489218, "learning_rate": 1.833058838093307e-05, "loss": 0.8034, "step": 5826 }, { "epoch": 0.2110467222021007, "grad_norm": 1.285861584849052, "learning_rate": 1.832993940086374e-05, "loss": 0.8441, "step": 5827 }, { "epoch": 0.21108294096341906, "grad_norm": 1.2209145311877103, "learning_rate": 1.832929030616712e-05, "loss": 0.7864, "step": 5828 }, { "epoch": 0.2111191597247374, "grad_norm": 1.62031064364475, "learning_rate": 1.8328641096852147e-05, "loss": 0.8691, "step": 5829 }, { "epoch": 0.21115537848605578, "grad_norm": 1.4398170797824363, "learning_rate": 1.832799177292776e-05, "loss": 0.779, "step": 5830 }, { "epoch": 0.21119159724737413, "grad_norm": 1.5239019810689043, "learning_rate": 1.8327342334402883e-05, "loss": 0.8189, "step": 5831 }, { "epoch": 0.2112278160086925, "grad_norm": 1.5224777781085468, "learning_rate": 1.832669278128646e-05, "loss": 0.8617, "step": 5832 }, { "epoch": 0.21126403477001088, "grad_norm": 1.5394569474845305, "learning_rate": 1.832604311358743e-05, "loss": 0.8427, "step": 5833 }, { "epoch": 0.21130025353132922, "grad_norm": 1.616616308309332, "learning_rate": 1.8325393331314728e-05, "loss": 0.8312, "step": 5834 }, { "epoch": 0.2113364722926476, "grad_norm": 1.5478570775556537, "learning_rate": 1.8324743434477302e-05, "loss": 0.7591, "step": 5835 }, { "epoch": 0.21137269105396594, "grad_norm": 1.4591645081069715, "learning_rate": 1.8324093423084087e-05, "loss": 0.8148, "step": 5836 }, { "epoch": 0.21140890981528432, "grad_norm": 1.3831508971348354, "learning_rate": 1.8323443297144036e-05, "loss": 0.837, "step": 5837 }, { "epoch": 0.2114451285766027, "grad_norm": 1.497120126690513, "learning_rate": 1.832279305666609e-05, "loss": 0.8139, "step": 5838 }, { "epoch": 0.21148134733792104, "grad_norm": 1.6085369650334755, "learning_rate": 1.83221427016592e-05, "loss": 0.9159, "step": 5839 }, { "epoch": 0.2115175660992394, "grad_norm": 1.4766277155022487, "learning_rate": 1.8321492232132312e-05, "loss": 0.8161, "step": 5840 }, { "epoch": 0.21155378486055776, "grad_norm": 1.7501736878305498, "learning_rate": 1.832084164809438e-05, "loss": 0.8961, "step": 5841 }, { "epoch": 0.21159000362187613, "grad_norm": 1.725547228790812, "learning_rate": 1.8320190949554356e-05, "loss": 0.8073, "step": 5842 }, { "epoch": 0.2116262223831945, "grad_norm": 1.92570545558068, "learning_rate": 1.831954013652119e-05, "loss": 0.8942, "step": 5843 }, { "epoch": 0.21166244114451285, "grad_norm": 1.5718230827149267, "learning_rate": 1.8318889209003847e-05, "loss": 0.7975, "step": 5844 }, { "epoch": 0.21169865990583123, "grad_norm": 1.6238772475002645, "learning_rate": 1.8318238167011276e-05, "loss": 0.817, "step": 5845 }, { "epoch": 0.21173487866714957, "grad_norm": 1.0762200744984987, "learning_rate": 1.831758701055244e-05, "loss": 0.7495, "step": 5846 }, { "epoch": 0.21177109742846795, "grad_norm": 1.4731035887503883, "learning_rate": 1.8316935739636298e-05, "loss": 0.8208, "step": 5847 }, { "epoch": 0.21180731618978632, "grad_norm": 1.4071301017542768, "learning_rate": 1.831628435427181e-05, "loss": 0.7763, "step": 5848 }, { "epoch": 0.21184353495110467, "grad_norm": 1.226737142843328, "learning_rate": 1.8315632854467944e-05, "loss": 0.7442, "step": 5849 }, { "epoch": 0.21187975371242304, "grad_norm": 2.2326406536883043, "learning_rate": 1.8314981240233663e-05, "loss": 0.725, "step": 5850 }, { "epoch": 0.2119159724737414, "grad_norm": 1.443922487912094, "learning_rate": 1.8314329511577933e-05, "loss": 0.7723, "step": 5851 }, { "epoch": 0.21195219123505976, "grad_norm": 1.341730906046176, "learning_rate": 1.8313677668509723e-05, "loss": 0.7997, "step": 5852 }, { "epoch": 0.21198840999637814, "grad_norm": 1.5414537682050846, "learning_rate": 1.8313025711038007e-05, "loss": 0.8091, "step": 5853 }, { "epoch": 0.21202462875769648, "grad_norm": 1.2289862292370743, "learning_rate": 1.831237363917175e-05, "loss": 0.7869, "step": 5854 }, { "epoch": 0.21206084751901486, "grad_norm": 1.3767540867319832, "learning_rate": 1.831172145291993e-05, "loss": 0.8139, "step": 5855 }, { "epoch": 0.2120970662803332, "grad_norm": 1.3522769587515688, "learning_rate": 1.8311069152291513e-05, "loss": 0.809, "step": 5856 }, { "epoch": 0.21213328504165158, "grad_norm": 1.7473787789192383, "learning_rate": 1.8310416737295486e-05, "loss": 0.7846, "step": 5857 }, { "epoch": 0.21216950380296995, "grad_norm": 1.5901667567883664, "learning_rate": 1.8309764207940826e-05, "loss": 0.7951, "step": 5858 }, { "epoch": 0.2122057225642883, "grad_norm": 1.5956302530465207, "learning_rate": 1.8309111564236504e-05, "loss": 0.7993, "step": 5859 }, { "epoch": 0.21224194132560667, "grad_norm": 1.152589231343652, "learning_rate": 1.8308458806191507e-05, "loss": 0.751, "step": 5860 }, { "epoch": 0.21227816008692502, "grad_norm": 1.0456304370430023, "learning_rate": 1.8307805933814818e-05, "loss": 0.7293, "step": 5861 }, { "epoch": 0.2123143788482434, "grad_norm": 1.5229809994086074, "learning_rate": 1.8307152947115418e-05, "loss": 0.8458, "step": 5862 }, { "epoch": 0.21235059760956176, "grad_norm": 1.7535300934692422, "learning_rate": 1.8306499846102294e-05, "loss": 0.7928, "step": 5863 }, { "epoch": 0.2123868163708801, "grad_norm": 1.5062810910941593, "learning_rate": 1.8305846630784436e-05, "loss": 0.8597, "step": 5864 }, { "epoch": 0.21242303513219848, "grad_norm": 1.5569153714353348, "learning_rate": 1.8305193301170828e-05, "loss": 0.8102, "step": 5865 }, { "epoch": 0.21245925389351683, "grad_norm": 1.6169240645982905, "learning_rate": 1.8304539857270463e-05, "loss": 0.7988, "step": 5866 }, { "epoch": 0.2124954726548352, "grad_norm": 1.5363400404996892, "learning_rate": 1.830388629909233e-05, "loss": 0.92, "step": 5867 }, { "epoch": 0.21253169141615358, "grad_norm": 1.4871989955906693, "learning_rate": 1.8303232626645425e-05, "loss": 0.8698, "step": 5868 }, { "epoch": 0.21256791017747192, "grad_norm": 1.4013325477598786, "learning_rate": 1.8302578839938748e-05, "loss": 0.818, "step": 5869 }, { "epoch": 0.2126041289387903, "grad_norm": 1.7374935036819095, "learning_rate": 1.830192493898129e-05, "loss": 0.8641, "step": 5870 }, { "epoch": 0.21264034770010865, "grad_norm": 1.3319633472172758, "learning_rate": 1.830127092378205e-05, "loss": 0.8029, "step": 5871 }, { "epoch": 0.21267656646142702, "grad_norm": 1.75802683477215, "learning_rate": 1.8300616794350025e-05, "loss": 0.8533, "step": 5872 }, { "epoch": 0.2127127852227454, "grad_norm": 1.38651009588275, "learning_rate": 1.829996255069422e-05, "loss": 0.771, "step": 5873 }, { "epoch": 0.21274900398406374, "grad_norm": 1.5192150890892189, "learning_rate": 1.829930819282364e-05, "loss": 0.8659, "step": 5874 }, { "epoch": 0.2127852227453821, "grad_norm": 1.5131848640842773, "learning_rate": 1.8298653720747283e-05, "loss": 0.9044, "step": 5875 }, { "epoch": 0.21282144150670046, "grad_norm": 1.1241111564752089, "learning_rate": 1.8297999134474162e-05, "loss": 0.7674, "step": 5876 }, { "epoch": 0.21285766026801883, "grad_norm": 1.5014807766945137, "learning_rate": 1.829734443401328e-05, "loss": 0.7979, "step": 5877 }, { "epoch": 0.2128938790293372, "grad_norm": 1.5798141473530727, "learning_rate": 1.829668961937365e-05, "loss": 0.8528, "step": 5878 }, { "epoch": 0.21293009779065555, "grad_norm": 1.4373383256190129, "learning_rate": 1.8296034690564276e-05, "loss": 0.8801, "step": 5879 }, { "epoch": 0.21296631655197393, "grad_norm": 1.5281001776452003, "learning_rate": 1.829537964759418e-05, "loss": 0.781, "step": 5880 }, { "epoch": 0.21300253531329227, "grad_norm": 1.5071162805452931, "learning_rate": 1.829472449047237e-05, "loss": 0.846, "step": 5881 }, { "epoch": 0.21303875407461065, "grad_norm": 1.2660124370247328, "learning_rate": 1.829406921920786e-05, "loss": 0.7771, "step": 5882 }, { "epoch": 0.21307497283592902, "grad_norm": 1.5277577867082042, "learning_rate": 1.829341383380967e-05, "loss": 0.844, "step": 5883 }, { "epoch": 0.21311119159724737, "grad_norm": 1.170221412683708, "learning_rate": 1.8292758334286822e-05, "loss": 0.7562, "step": 5884 }, { "epoch": 0.21314741035856574, "grad_norm": 1.3707548762646222, "learning_rate": 1.8292102720648333e-05, "loss": 0.81, "step": 5885 }, { "epoch": 0.2131836291198841, "grad_norm": 1.7093475827767775, "learning_rate": 1.829144699290322e-05, "loss": 0.9205, "step": 5886 }, { "epoch": 0.21321984788120246, "grad_norm": 1.5268765787508012, "learning_rate": 1.8290791151060513e-05, "loss": 0.8638, "step": 5887 }, { "epoch": 0.21325606664252084, "grad_norm": 1.3872873591173656, "learning_rate": 1.8290135195129237e-05, "loss": 0.7518, "step": 5888 }, { "epoch": 0.21329228540383918, "grad_norm": 1.4467488381663929, "learning_rate": 1.8289479125118412e-05, "loss": 0.9333, "step": 5889 }, { "epoch": 0.21332850416515756, "grad_norm": 1.8103343423256297, "learning_rate": 1.8288822941037074e-05, "loss": 0.8218, "step": 5890 }, { "epoch": 0.2133647229264759, "grad_norm": 1.311830027485476, "learning_rate": 1.828816664289425e-05, "loss": 0.7549, "step": 5891 }, { "epoch": 0.21340094168779428, "grad_norm": 1.620684693903555, "learning_rate": 1.828751023069897e-05, "loss": 0.8221, "step": 5892 }, { "epoch": 0.21343716044911265, "grad_norm": 1.5261208682669438, "learning_rate": 1.8286853704460264e-05, "loss": 0.8145, "step": 5893 }, { "epoch": 0.213473379210431, "grad_norm": 1.3942615437889878, "learning_rate": 1.828619706418717e-05, "loss": 0.7697, "step": 5894 }, { "epoch": 0.21350959797174937, "grad_norm": 1.4726195925291201, "learning_rate": 1.8285540309888724e-05, "loss": 0.7683, "step": 5895 }, { "epoch": 0.21354581673306772, "grad_norm": 1.4441710707509656, "learning_rate": 1.8284883441573968e-05, "loss": 0.8078, "step": 5896 }, { "epoch": 0.2135820354943861, "grad_norm": 1.3123587392355387, "learning_rate": 1.8284226459251928e-05, "loss": 0.7632, "step": 5897 }, { "epoch": 0.21361825425570447, "grad_norm": 1.5835828533498146, "learning_rate": 1.8283569362931657e-05, "loss": 0.8393, "step": 5898 }, { "epoch": 0.2136544730170228, "grad_norm": 1.3917915000458427, "learning_rate": 1.828291215262219e-05, "loss": 0.8038, "step": 5899 }, { "epoch": 0.21369069177834119, "grad_norm": 1.5538581518254155, "learning_rate": 1.8282254828332578e-05, "loss": 0.8793, "step": 5900 }, { "epoch": 0.21372691053965953, "grad_norm": 1.5250861630907728, "learning_rate": 1.8281597390071862e-05, "loss": 0.8534, "step": 5901 }, { "epoch": 0.2137631293009779, "grad_norm": 1.3355640464874332, "learning_rate": 1.8280939837849087e-05, "loss": 0.7572, "step": 5902 }, { "epoch": 0.21379934806229628, "grad_norm": 1.4650723128816805, "learning_rate": 1.8280282171673303e-05, "loss": 0.7889, "step": 5903 }, { "epoch": 0.21383556682361463, "grad_norm": 1.4823705779939378, "learning_rate": 1.827962439155356e-05, "loss": 0.8318, "step": 5904 }, { "epoch": 0.213871785584933, "grad_norm": 1.472802165692474, "learning_rate": 1.8278966497498912e-05, "loss": 0.831, "step": 5905 }, { "epoch": 0.21390800434625135, "grad_norm": 1.3745338350078526, "learning_rate": 1.8278308489518412e-05, "loss": 0.9023, "step": 5906 }, { "epoch": 0.21394422310756972, "grad_norm": 1.5382162904754784, "learning_rate": 1.827765036762111e-05, "loss": 0.8442, "step": 5907 }, { "epoch": 0.2139804418688881, "grad_norm": 1.4665858412806934, "learning_rate": 1.827699213181607e-05, "loss": 0.8624, "step": 5908 }, { "epoch": 0.21401666063020644, "grad_norm": 1.4283270062530087, "learning_rate": 1.827633378211234e-05, "loss": 0.8104, "step": 5909 }, { "epoch": 0.21405287939152481, "grad_norm": 1.461329280752884, "learning_rate": 1.8275675318518988e-05, "loss": 0.8564, "step": 5910 }, { "epoch": 0.21408909815284316, "grad_norm": 1.31887753935441, "learning_rate": 1.8275016741045074e-05, "loss": 0.7321, "step": 5911 }, { "epoch": 0.21412531691416153, "grad_norm": 1.397556249699064, "learning_rate": 1.8274358049699656e-05, "loss": 0.8099, "step": 5912 }, { "epoch": 0.2141615356754799, "grad_norm": 1.4713845945935264, "learning_rate": 1.82736992444918e-05, "loss": 0.8282, "step": 5913 }, { "epoch": 0.21419775443679825, "grad_norm": 1.689977108176787, "learning_rate": 1.8273040325430575e-05, "loss": 0.8221, "step": 5914 }, { "epoch": 0.21423397319811663, "grad_norm": 1.4486923302752974, "learning_rate": 1.8272381292525044e-05, "loss": 0.8548, "step": 5915 }, { "epoch": 0.21427019195943497, "grad_norm": 2.473355125487427, "learning_rate": 1.8271722145784277e-05, "loss": 0.8385, "step": 5916 }, { "epoch": 0.21430641072075335, "grad_norm": 1.6448723853753107, "learning_rate": 1.8271062885217346e-05, "loss": 0.7591, "step": 5917 }, { "epoch": 0.21434262948207172, "grad_norm": 1.65484852245998, "learning_rate": 1.827040351083332e-05, "loss": 0.769, "step": 5918 }, { "epoch": 0.21437884824339007, "grad_norm": 1.6608715221905208, "learning_rate": 1.8269744022641278e-05, "loss": 0.7736, "step": 5919 }, { "epoch": 0.21441506700470844, "grad_norm": 1.5307625778171496, "learning_rate": 1.826908442065029e-05, "loss": 0.8201, "step": 5920 }, { "epoch": 0.2144512857660268, "grad_norm": 1.3888165764726297, "learning_rate": 1.8268424704869436e-05, "loss": 0.8061, "step": 5921 }, { "epoch": 0.21448750452734516, "grad_norm": 1.45364705444712, "learning_rate": 1.826776487530779e-05, "loss": 0.8579, "step": 5922 }, { "epoch": 0.21452372328866354, "grad_norm": 1.3138136229151645, "learning_rate": 1.826710493197444e-05, "loss": 0.8298, "step": 5923 }, { "epoch": 0.21455994204998188, "grad_norm": 1.5125146219831458, "learning_rate": 1.8266444874878456e-05, "loss": 0.8401, "step": 5924 }, { "epoch": 0.21459616081130026, "grad_norm": 1.4807407109225046, "learning_rate": 1.8265784704028934e-05, "loss": 0.8282, "step": 5925 }, { "epoch": 0.2146323795726186, "grad_norm": 1.5162151333172103, "learning_rate": 1.8265124419434945e-05, "loss": 0.8122, "step": 5926 }, { "epoch": 0.21466859833393698, "grad_norm": 1.278731868276745, "learning_rate": 1.826446402110558e-05, "loss": 0.8004, "step": 5927 }, { "epoch": 0.21470481709525535, "grad_norm": 1.415984281653421, "learning_rate": 1.8263803509049936e-05, "loss": 0.8143, "step": 5928 }, { "epoch": 0.2147410358565737, "grad_norm": 1.600648852002665, "learning_rate": 1.826314288327709e-05, "loss": 0.7499, "step": 5929 }, { "epoch": 0.21477725461789207, "grad_norm": 1.6613672125851031, "learning_rate": 1.8262482143796136e-05, "loss": 0.8598, "step": 5930 }, { "epoch": 0.21481347337921042, "grad_norm": 1.584102737905687, "learning_rate": 1.826182129061617e-05, "loss": 0.822, "step": 5931 }, { "epoch": 0.2148496921405288, "grad_norm": 1.5118563291257043, "learning_rate": 1.8261160323746283e-05, "loss": 0.7968, "step": 5932 }, { "epoch": 0.21488591090184717, "grad_norm": 1.5810914391716482, "learning_rate": 1.826049924319557e-05, "loss": 0.7699, "step": 5933 }, { "epoch": 0.2149221296631655, "grad_norm": 1.5778280571317742, "learning_rate": 1.8259838048973132e-05, "loss": 0.7875, "step": 5934 }, { "epoch": 0.21495834842448389, "grad_norm": 1.3550131489704742, "learning_rate": 1.8259176741088062e-05, "loss": 0.8282, "step": 5935 }, { "epoch": 0.21499456718580223, "grad_norm": 1.3004332963411513, "learning_rate": 1.8258515319549463e-05, "loss": 0.8306, "step": 5936 }, { "epoch": 0.2150307859471206, "grad_norm": 1.45875956374893, "learning_rate": 1.8257853784366436e-05, "loss": 0.8372, "step": 5937 }, { "epoch": 0.21506700470843898, "grad_norm": 1.418003119953275, "learning_rate": 1.8257192135548087e-05, "loss": 0.8148, "step": 5938 }, { "epoch": 0.21510322346975733, "grad_norm": 1.363682144156969, "learning_rate": 1.8256530373103518e-05, "loss": 0.7951, "step": 5939 }, { "epoch": 0.2151394422310757, "grad_norm": 1.5258702209225472, "learning_rate": 1.8255868497041835e-05, "loss": 0.8106, "step": 5940 }, { "epoch": 0.21517566099239405, "grad_norm": 1.414570324690008, "learning_rate": 1.825520650737215e-05, "loss": 0.8334, "step": 5941 }, { "epoch": 0.21521187975371242, "grad_norm": 1.3635792365579802, "learning_rate": 1.825454440410357e-05, "loss": 0.8237, "step": 5942 }, { "epoch": 0.2152480985150308, "grad_norm": 1.450146143172838, "learning_rate": 1.82538821872452e-05, "loss": 0.8376, "step": 5943 }, { "epoch": 0.21528431727634914, "grad_norm": 1.381515363974406, "learning_rate": 1.8253219856806164e-05, "loss": 0.8318, "step": 5944 }, { "epoch": 0.21532053603766751, "grad_norm": 1.5586337261283796, "learning_rate": 1.8252557412795572e-05, "loss": 0.8446, "step": 5945 }, { "epoch": 0.21535675479898586, "grad_norm": 1.1556827706962634, "learning_rate": 1.8251894855222537e-05, "loss": 0.8177, "step": 5946 }, { "epoch": 0.21539297356030424, "grad_norm": 1.3981695437913395, "learning_rate": 1.8251232184096177e-05, "loss": 0.6982, "step": 5947 }, { "epoch": 0.2154291923216226, "grad_norm": 1.541686669801985, "learning_rate": 1.8250569399425613e-05, "loss": 0.9077, "step": 5948 }, { "epoch": 0.21546541108294096, "grad_norm": 1.6745928149319935, "learning_rate": 1.8249906501219962e-05, "loss": 0.746, "step": 5949 }, { "epoch": 0.21550162984425933, "grad_norm": 1.219470466473068, "learning_rate": 1.8249243489488352e-05, "loss": 0.7496, "step": 5950 }, { "epoch": 0.21553784860557768, "grad_norm": 1.5488076202651595, "learning_rate": 1.82485803642399e-05, "loss": 0.7923, "step": 5951 }, { "epoch": 0.21557406736689605, "grad_norm": 1.1207396360458508, "learning_rate": 1.8247917125483735e-05, "loss": 0.7864, "step": 5952 }, { "epoch": 0.21561028612821442, "grad_norm": 1.2503708695296316, "learning_rate": 1.8247253773228987e-05, "loss": 0.7901, "step": 5953 }, { "epoch": 0.21564650488953277, "grad_norm": 1.4605767408971424, "learning_rate": 1.8246590307484776e-05, "loss": 0.7509, "step": 5954 }, { "epoch": 0.21568272365085114, "grad_norm": 1.3257882472085716, "learning_rate": 1.8245926728260238e-05, "loss": 0.8231, "step": 5955 }, { "epoch": 0.21571894241216952, "grad_norm": 1.1990921251388553, "learning_rate": 1.8245263035564503e-05, "loss": 0.7845, "step": 5956 }, { "epoch": 0.21575516117348786, "grad_norm": 1.3458414384722386, "learning_rate": 1.8244599229406702e-05, "loss": 0.7317, "step": 5957 }, { "epoch": 0.21579137993480624, "grad_norm": 1.5174312471380116, "learning_rate": 1.8243935309795974e-05, "loss": 0.9291, "step": 5958 }, { "epoch": 0.21582759869612458, "grad_norm": 1.6180170599053676, "learning_rate": 1.824327127674145e-05, "loss": 0.7976, "step": 5959 }, { "epoch": 0.21586381745744296, "grad_norm": 1.5197529790367161, "learning_rate": 1.824260713025227e-05, "loss": 0.8766, "step": 5960 }, { "epoch": 0.21590003621876133, "grad_norm": 1.536988916487898, "learning_rate": 1.8241942870337575e-05, "loss": 0.805, "step": 5961 }, { "epoch": 0.21593625498007968, "grad_norm": 1.5002468869562053, "learning_rate": 1.8241278497006503e-05, "loss": 0.8469, "step": 5962 }, { "epoch": 0.21597247374139805, "grad_norm": 1.541316612206843, "learning_rate": 1.82406140102682e-05, "loss": 0.8774, "step": 5963 }, { "epoch": 0.2160086925027164, "grad_norm": 1.5962466670588138, "learning_rate": 1.8239949410131803e-05, "loss": 0.849, "step": 5964 }, { "epoch": 0.21604491126403477, "grad_norm": 1.4268841449283254, "learning_rate": 1.8239284696606465e-05, "loss": 0.7715, "step": 5965 }, { "epoch": 0.21608113002535315, "grad_norm": 1.4165541558753125, "learning_rate": 1.8238619869701328e-05, "loss": 0.8274, "step": 5966 }, { "epoch": 0.2161173487866715, "grad_norm": 1.4002006460421605, "learning_rate": 1.8237954929425544e-05, "loss": 0.7559, "step": 5967 }, { "epoch": 0.21615356754798987, "grad_norm": 1.5102168971897374, "learning_rate": 1.823728987578826e-05, "loss": 0.8163, "step": 5968 }, { "epoch": 0.2161897863093082, "grad_norm": 1.4731314297865725, "learning_rate": 1.823662470879863e-05, "loss": 0.7441, "step": 5969 }, { "epoch": 0.2162260050706266, "grad_norm": 1.5118561467859586, "learning_rate": 1.8235959428465806e-05, "loss": 0.8394, "step": 5970 }, { "epoch": 0.21626222383194496, "grad_norm": 1.4821852770383936, "learning_rate": 1.8235294034798948e-05, "loss": 0.7509, "step": 5971 }, { "epoch": 0.2162984425932633, "grad_norm": 1.6130088005740046, "learning_rate": 1.8234628527807202e-05, "loss": 0.869, "step": 5972 }, { "epoch": 0.21633466135458168, "grad_norm": 1.5025237063683474, "learning_rate": 1.8233962907499737e-05, "loss": 0.7926, "step": 5973 }, { "epoch": 0.21637088011590003, "grad_norm": 1.4840961093916365, "learning_rate": 1.8233297173885702e-05, "loss": 0.7606, "step": 5974 }, { "epoch": 0.2164070988772184, "grad_norm": 1.4510181635675914, "learning_rate": 1.8232631326974268e-05, "loss": 0.8213, "step": 5975 }, { "epoch": 0.21644331763853678, "grad_norm": 1.0869850281081797, "learning_rate": 1.8231965366774592e-05, "loss": 0.7468, "step": 5976 }, { "epoch": 0.21647953639985512, "grad_norm": 1.5089740269439693, "learning_rate": 1.8231299293295837e-05, "loss": 0.7942, "step": 5977 }, { "epoch": 0.2165157551611735, "grad_norm": 1.5136731424633996, "learning_rate": 1.823063310654717e-05, "loss": 0.8611, "step": 5978 }, { "epoch": 0.21655197392249184, "grad_norm": 1.2859904812238325, "learning_rate": 1.8229966806537766e-05, "loss": 0.7923, "step": 5979 }, { "epoch": 0.21658819268381022, "grad_norm": 1.4127700494768909, "learning_rate": 1.8229300393276783e-05, "loss": 0.7649, "step": 5980 }, { "epoch": 0.2166244114451286, "grad_norm": 1.4586305010394618, "learning_rate": 1.8228633866773396e-05, "loss": 0.8439, "step": 5981 }, { "epoch": 0.21666063020644694, "grad_norm": 1.2687334399727408, "learning_rate": 1.822796722703678e-05, "loss": 0.8774, "step": 5982 }, { "epoch": 0.2166968489677653, "grad_norm": 1.3447135839480064, "learning_rate": 1.8227300474076098e-05, "loss": 0.8393, "step": 5983 }, { "epoch": 0.21673306772908366, "grad_norm": 1.4972130482830155, "learning_rate": 1.822663360790054e-05, "loss": 0.8678, "step": 5984 }, { "epoch": 0.21676928649040203, "grad_norm": 1.6181223535313483, "learning_rate": 1.822596662851927e-05, "loss": 0.8947, "step": 5985 }, { "epoch": 0.2168055052517204, "grad_norm": 1.4592831993738629, "learning_rate": 1.8225299535941473e-05, "loss": 0.8491, "step": 5986 }, { "epoch": 0.21684172401303875, "grad_norm": 1.6399849233428596, "learning_rate": 1.8224632330176325e-05, "loss": 0.8755, "step": 5987 }, { "epoch": 0.21687794277435712, "grad_norm": 1.2696318828605673, "learning_rate": 1.822396501123301e-05, "loss": 0.8382, "step": 5988 }, { "epoch": 0.21691416153567547, "grad_norm": 1.4594315809925675, "learning_rate": 1.8223297579120713e-05, "loss": 0.7604, "step": 5989 }, { "epoch": 0.21695038029699384, "grad_norm": 1.349781885176664, "learning_rate": 1.8222630033848615e-05, "loss": 0.794, "step": 5990 }, { "epoch": 0.21698659905831222, "grad_norm": 1.738095486819707, "learning_rate": 1.8221962375425902e-05, "loss": 0.8277, "step": 5991 }, { "epoch": 0.21702281781963056, "grad_norm": 1.3950714633075714, "learning_rate": 1.822129460386176e-05, "loss": 0.8357, "step": 5992 }, { "epoch": 0.21705903658094894, "grad_norm": 1.573658521654559, "learning_rate": 1.8220626719165383e-05, "loss": 0.8016, "step": 5993 }, { "epoch": 0.21709525534226728, "grad_norm": 1.3536802469741982, "learning_rate": 1.8219958721345957e-05, "loss": 0.8422, "step": 5994 }, { "epoch": 0.21713147410358566, "grad_norm": 1.5539402591690954, "learning_rate": 1.821929061041268e-05, "loss": 0.8323, "step": 5995 }, { "epoch": 0.21716769286490403, "grad_norm": 1.2167716646051208, "learning_rate": 1.821862238637474e-05, "loss": 0.8624, "step": 5996 }, { "epoch": 0.21720391162622238, "grad_norm": 1.7022136975557773, "learning_rate": 1.8217954049241334e-05, "loss": 0.8323, "step": 5997 }, { "epoch": 0.21724013038754075, "grad_norm": 1.215653501455905, "learning_rate": 1.8217285599021658e-05, "loss": 0.8366, "step": 5998 }, { "epoch": 0.2172763491488591, "grad_norm": 1.5665478226190792, "learning_rate": 1.8216617035724913e-05, "loss": 0.8153, "step": 5999 }, { "epoch": 0.21731256791017747, "grad_norm": 1.5533360904835753, "learning_rate": 1.82159483593603e-05, "loss": 0.7896, "step": 6000 }, { "epoch": 0.21734878667149585, "grad_norm": 1.4529268173848882, "learning_rate": 1.8215279569937016e-05, "loss": 0.846, "step": 6001 }, { "epoch": 0.2173850054328142, "grad_norm": 1.7593437243708772, "learning_rate": 1.8214610667464264e-05, "loss": 0.8456, "step": 6002 }, { "epoch": 0.21742122419413257, "grad_norm": 1.3281850017639827, "learning_rate": 1.8213941651951255e-05, "loss": 0.7985, "step": 6003 }, { "epoch": 0.2174574429554509, "grad_norm": 1.673294912758557, "learning_rate": 1.8213272523407192e-05, "loss": 0.7821, "step": 6004 }, { "epoch": 0.2174936617167693, "grad_norm": 1.4687611916805552, "learning_rate": 1.8212603281841283e-05, "loss": 0.7843, "step": 6005 }, { "epoch": 0.21752988047808766, "grad_norm": 1.4759517666837751, "learning_rate": 1.8211933927262734e-05, "loss": 0.7931, "step": 6006 }, { "epoch": 0.217566099239406, "grad_norm": 1.559208003162964, "learning_rate": 1.8211264459680757e-05, "loss": 0.825, "step": 6007 }, { "epoch": 0.21760231800072438, "grad_norm": 1.5547168210903233, "learning_rate": 1.821059487910457e-05, "loss": 0.8214, "step": 6008 }, { "epoch": 0.21763853676204273, "grad_norm": 1.6537306716262306, "learning_rate": 1.820992518554338e-05, "loss": 0.8462, "step": 6009 }, { "epoch": 0.2176747555233611, "grad_norm": 1.4880249348393537, "learning_rate": 1.820925537900641e-05, "loss": 0.8455, "step": 6010 }, { "epoch": 0.21771097428467948, "grad_norm": 1.483317754081911, "learning_rate": 1.8208585459502868e-05, "loss": 0.7508, "step": 6011 }, { "epoch": 0.21774719304599782, "grad_norm": 1.4431062251819606, "learning_rate": 1.8207915427041977e-05, "loss": 0.8134, "step": 6012 }, { "epoch": 0.2177834118073162, "grad_norm": 1.5119630409214777, "learning_rate": 1.820724528163296e-05, "loss": 0.8659, "step": 6013 }, { "epoch": 0.21781963056863454, "grad_norm": 1.360815684470572, "learning_rate": 1.8206575023285035e-05, "loss": 0.7962, "step": 6014 }, { "epoch": 0.21785584932995292, "grad_norm": 1.5265858128152687, "learning_rate": 1.820590465200743e-05, "loss": 0.8755, "step": 6015 }, { "epoch": 0.2178920680912713, "grad_norm": 1.6461899370805961, "learning_rate": 1.8205234167809363e-05, "loss": 0.8399, "step": 6016 }, { "epoch": 0.21792828685258964, "grad_norm": 1.621923599061698, "learning_rate": 1.8204563570700063e-05, "loss": 0.8493, "step": 6017 }, { "epoch": 0.217964505613908, "grad_norm": 1.5067499436608223, "learning_rate": 1.8203892860688763e-05, "loss": 0.8661, "step": 6018 }, { "epoch": 0.21800072437522636, "grad_norm": 1.5929013800353835, "learning_rate": 1.820322203778469e-05, "loss": 0.7757, "step": 6019 }, { "epoch": 0.21803694313654473, "grad_norm": 1.320183901036557, "learning_rate": 1.8202551101997068e-05, "loss": 0.7925, "step": 6020 }, { "epoch": 0.2180731618978631, "grad_norm": 1.2472148576156181, "learning_rate": 1.820188005333514e-05, "loss": 0.7724, "step": 6021 }, { "epoch": 0.21810938065918145, "grad_norm": 1.4867410637568688, "learning_rate": 1.8201208891808132e-05, "loss": 0.8287, "step": 6022 }, { "epoch": 0.21814559942049982, "grad_norm": 1.5374862228334536, "learning_rate": 1.8200537617425288e-05, "loss": 0.8101, "step": 6023 }, { "epoch": 0.21818181818181817, "grad_norm": 1.3153549368150659, "learning_rate": 1.819986623019584e-05, "loss": 0.7885, "step": 6024 }, { "epoch": 0.21821803694313655, "grad_norm": 1.2354621461708015, "learning_rate": 1.8199194730129025e-05, "loss": 0.791, "step": 6025 }, { "epoch": 0.21825425570445492, "grad_norm": 1.5878656744316386, "learning_rate": 1.8198523117234085e-05, "loss": 0.8919, "step": 6026 }, { "epoch": 0.21829047446577327, "grad_norm": 1.4996717697382942, "learning_rate": 1.8197851391520265e-05, "loss": 0.8301, "step": 6027 }, { "epoch": 0.21832669322709164, "grad_norm": 1.1329822439314063, "learning_rate": 1.8197179552996805e-05, "loss": 0.7838, "step": 6028 }, { "epoch": 0.21836291198840999, "grad_norm": 1.2529021542185712, "learning_rate": 1.8196507601672954e-05, "loss": 0.7425, "step": 6029 }, { "epoch": 0.21839913074972836, "grad_norm": 1.1882960733750023, "learning_rate": 1.819583553755796e-05, "loss": 0.7396, "step": 6030 }, { "epoch": 0.21843534951104673, "grad_norm": 1.9291164010629995, "learning_rate": 1.819516336066106e-05, "loss": 0.8301, "step": 6031 }, { "epoch": 0.21847156827236508, "grad_norm": 1.196520252659736, "learning_rate": 1.8194491070991517e-05, "loss": 0.8385, "step": 6032 }, { "epoch": 0.21850778703368345, "grad_norm": 1.464446188064959, "learning_rate": 1.8193818668558573e-05, "loss": 0.8209, "step": 6033 }, { "epoch": 0.2185440057950018, "grad_norm": 1.4719894478072428, "learning_rate": 1.8193146153371483e-05, "loss": 0.7371, "step": 6034 }, { "epoch": 0.21858022455632017, "grad_norm": 1.4134984142685387, "learning_rate": 1.8192473525439506e-05, "loss": 0.7421, "step": 6035 }, { "epoch": 0.21861644331763855, "grad_norm": 1.5301135070280105, "learning_rate": 1.8191800784771893e-05, "loss": 0.8839, "step": 6036 }, { "epoch": 0.2186526620789569, "grad_norm": 1.6326600635592454, "learning_rate": 1.8191127931377903e-05, "loss": 0.8778, "step": 6037 }, { "epoch": 0.21868888084027527, "grad_norm": 1.579877138027414, "learning_rate": 1.81904549652668e-05, "loss": 0.8481, "step": 6038 }, { "epoch": 0.21872509960159361, "grad_norm": 1.4236122950220096, "learning_rate": 1.8189781886447834e-05, "loss": 0.8389, "step": 6039 }, { "epoch": 0.218761318362912, "grad_norm": 1.2912616271728437, "learning_rate": 1.818910869493028e-05, "loss": 0.787, "step": 6040 }, { "epoch": 0.21879753712423036, "grad_norm": 1.4339982223569552, "learning_rate": 1.8188435390723386e-05, "loss": 0.8498, "step": 6041 }, { "epoch": 0.2188337558855487, "grad_norm": 1.4423209534166876, "learning_rate": 1.818776197383643e-05, "loss": 0.8599, "step": 6042 }, { "epoch": 0.21886997464686708, "grad_norm": 1.5670536819454601, "learning_rate": 1.8187088444278675e-05, "loss": 0.7547, "step": 6043 }, { "epoch": 0.21890619340818543, "grad_norm": 1.349059377132593, "learning_rate": 1.818641480205939e-05, "loss": 0.7793, "step": 6044 }, { "epoch": 0.2189424121695038, "grad_norm": 1.470885546010689, "learning_rate": 1.8185741047187843e-05, "loss": 0.8652, "step": 6045 }, { "epoch": 0.21897863093082218, "grad_norm": 1.5382102710182597, "learning_rate": 1.818506717967331e-05, "loss": 0.8437, "step": 6046 }, { "epoch": 0.21901484969214052, "grad_norm": 1.430868095847299, "learning_rate": 1.8184393199525054e-05, "loss": 0.8094, "step": 6047 }, { "epoch": 0.2190510684534589, "grad_norm": 1.5021221415512436, "learning_rate": 1.818371910675236e-05, "loss": 0.8092, "step": 6048 }, { "epoch": 0.21908728721477724, "grad_norm": 1.9941986404333971, "learning_rate": 1.8183044901364502e-05, "loss": 0.9409, "step": 6049 }, { "epoch": 0.21912350597609562, "grad_norm": 1.566342114683796, "learning_rate": 1.8182370583370753e-05, "loss": 0.7782, "step": 6050 }, { "epoch": 0.219159724737414, "grad_norm": 1.6568977151074145, "learning_rate": 1.8181696152780395e-05, "loss": 0.7922, "step": 6051 }, { "epoch": 0.21919594349873234, "grad_norm": 1.5984252848582614, "learning_rate": 1.8181021609602712e-05, "loss": 0.8452, "step": 6052 }, { "epoch": 0.2192321622600507, "grad_norm": 1.411835990423475, "learning_rate": 1.818034695384698e-05, "loss": 0.7446, "step": 6053 }, { "epoch": 0.21926838102136906, "grad_norm": 1.3654489935185474, "learning_rate": 1.8179672185522487e-05, "loss": 0.7779, "step": 6054 }, { "epoch": 0.21930459978268743, "grad_norm": 1.3219115915177015, "learning_rate": 1.8178997304638515e-05, "loss": 0.7973, "step": 6055 }, { "epoch": 0.2193408185440058, "grad_norm": 1.4557073493549824, "learning_rate": 1.817832231120436e-05, "loss": 0.8668, "step": 6056 }, { "epoch": 0.21937703730532415, "grad_norm": 1.6285062145474583, "learning_rate": 1.8177647205229295e-05, "loss": 0.9215, "step": 6057 }, { "epoch": 0.21941325606664253, "grad_norm": 1.3644856021793503, "learning_rate": 1.8176971986722625e-05, "loss": 0.8709, "step": 6058 }, { "epoch": 0.21944947482796087, "grad_norm": 1.4336805258897296, "learning_rate": 1.8176296655693635e-05, "loss": 0.8149, "step": 6059 }, { "epoch": 0.21948569358927925, "grad_norm": 1.1973241765751554, "learning_rate": 1.8175621212151616e-05, "loss": 0.7786, "step": 6060 }, { "epoch": 0.21952191235059762, "grad_norm": 1.2831230549565602, "learning_rate": 1.817494565610587e-05, "loss": 0.813, "step": 6061 }, { "epoch": 0.21955813111191597, "grad_norm": 1.4292539026337794, "learning_rate": 1.8174269987565686e-05, "loss": 0.8459, "step": 6062 }, { "epoch": 0.21959434987323434, "grad_norm": 1.1801869690159223, "learning_rate": 1.8173594206540363e-05, "loss": 0.7785, "step": 6063 }, { "epoch": 0.2196305686345527, "grad_norm": 1.3529481585874414, "learning_rate": 1.8172918313039206e-05, "loss": 0.7458, "step": 6064 }, { "epoch": 0.21966678739587106, "grad_norm": 1.448579554062889, "learning_rate": 1.8172242307071507e-05, "loss": 0.8578, "step": 6065 }, { "epoch": 0.21970300615718943, "grad_norm": 1.5112573895498556, "learning_rate": 1.8171566188646575e-05, "loss": 0.8339, "step": 6066 }, { "epoch": 0.21973922491850778, "grad_norm": 1.5418171858820902, "learning_rate": 1.817088995777371e-05, "loss": 0.8935, "step": 6067 }, { "epoch": 0.21977544367982615, "grad_norm": 1.570935664197685, "learning_rate": 1.8170213614462223e-05, "loss": 0.8105, "step": 6068 }, { "epoch": 0.2198116624411445, "grad_norm": 1.4974791153090183, "learning_rate": 1.8169537158721415e-05, "loss": 0.7956, "step": 6069 }, { "epoch": 0.21984788120246287, "grad_norm": 1.1709501800983622, "learning_rate": 1.81688605905606e-05, "loss": 0.7833, "step": 6070 }, { "epoch": 0.21988409996378125, "grad_norm": 1.487418514161556, "learning_rate": 1.8168183909989085e-05, "loss": 0.9041, "step": 6071 }, { "epoch": 0.2199203187250996, "grad_norm": 1.350139519316637, "learning_rate": 1.816750711701618e-05, "loss": 0.8348, "step": 6072 }, { "epoch": 0.21995653748641797, "grad_norm": 1.2890473036159316, "learning_rate": 1.81668302116512e-05, "loss": 0.8361, "step": 6073 }, { "epoch": 0.21999275624773632, "grad_norm": 1.5883301212336938, "learning_rate": 1.8166153193903464e-05, "loss": 0.8313, "step": 6074 }, { "epoch": 0.2200289750090547, "grad_norm": 1.4838721327019297, "learning_rate": 1.816547606378228e-05, "loss": 0.8038, "step": 6075 }, { "epoch": 0.22006519377037306, "grad_norm": 2.2599263962418794, "learning_rate": 1.8164798821296977e-05, "loss": 0.8377, "step": 6076 }, { "epoch": 0.2201014125316914, "grad_norm": 1.2228897103399001, "learning_rate": 1.816412146645686e-05, "loss": 0.813, "step": 6077 }, { "epoch": 0.22013763129300978, "grad_norm": 1.6503583114891613, "learning_rate": 1.8163443999271267e-05, "loss": 0.9444, "step": 6078 }, { "epoch": 0.22017385005432813, "grad_norm": 1.4639109044485032, "learning_rate": 1.8162766419749503e-05, "loss": 0.8178, "step": 6079 }, { "epoch": 0.2202100688156465, "grad_norm": 1.1290577661825891, "learning_rate": 1.8162088727900905e-05, "loss": 0.751, "step": 6080 }, { "epoch": 0.22024628757696488, "grad_norm": 1.5324945458609538, "learning_rate": 1.8161410923734795e-05, "loss": 0.8633, "step": 6081 }, { "epoch": 0.22028250633828322, "grad_norm": 1.5230482606766556, "learning_rate": 1.81607330072605e-05, "loss": 0.8795, "step": 6082 }, { "epoch": 0.2203187250996016, "grad_norm": 1.3110465621096907, "learning_rate": 1.8160054978487345e-05, "loss": 0.824, "step": 6083 }, { "epoch": 0.22035494386091994, "grad_norm": 1.4296549750364926, "learning_rate": 1.8159376837424665e-05, "loss": 0.8265, "step": 6084 }, { "epoch": 0.22039116262223832, "grad_norm": 1.4843009045578972, "learning_rate": 1.8158698584081788e-05, "loss": 0.8494, "step": 6085 }, { "epoch": 0.2204273813835567, "grad_norm": 1.5505682861480485, "learning_rate": 1.815802021846805e-05, "loss": 0.791, "step": 6086 }, { "epoch": 0.22046360014487504, "grad_norm": 1.4809074399841067, "learning_rate": 1.815734174059279e-05, "loss": 0.8538, "step": 6087 }, { "epoch": 0.2204998189061934, "grad_norm": 1.4791488299391329, "learning_rate": 1.8156663150465338e-05, "loss": 0.75, "step": 6088 }, { "epoch": 0.22053603766751176, "grad_norm": 1.4849090017187299, "learning_rate": 1.8155984448095032e-05, "loss": 0.8948, "step": 6089 }, { "epoch": 0.22057225642883013, "grad_norm": 1.4759862244499136, "learning_rate": 1.8155305633491217e-05, "loss": 0.7958, "step": 6090 }, { "epoch": 0.2206084751901485, "grad_norm": 1.5056569052834277, "learning_rate": 1.8154626706663225e-05, "loss": 0.816, "step": 6091 }, { "epoch": 0.22064469395146685, "grad_norm": 1.438632698934317, "learning_rate": 1.815394766762041e-05, "loss": 0.7918, "step": 6092 }, { "epoch": 0.22068091271278523, "grad_norm": 1.4328105776237923, "learning_rate": 1.8153268516372105e-05, "loss": 0.7603, "step": 6093 }, { "epoch": 0.22071713147410357, "grad_norm": 1.5285215751070047, "learning_rate": 1.8152589252927666e-05, "loss": 0.8481, "step": 6094 }, { "epoch": 0.22075335023542195, "grad_norm": 1.4939191965819507, "learning_rate": 1.815190987729643e-05, "loss": 0.8476, "step": 6095 }, { "epoch": 0.22078956899674032, "grad_norm": 1.583347051230618, "learning_rate": 1.8151230389487757e-05, "loss": 0.8382, "step": 6096 }, { "epoch": 0.22082578775805867, "grad_norm": 1.5071907775257185, "learning_rate": 1.815055078951099e-05, "loss": 0.7854, "step": 6097 }, { "epoch": 0.22086200651937704, "grad_norm": 1.4357368110034021, "learning_rate": 1.8149871077375482e-05, "loss": 0.8163, "step": 6098 }, { "epoch": 0.2208982252806954, "grad_norm": 1.5081037748979778, "learning_rate": 1.8149191253090586e-05, "loss": 0.8066, "step": 6099 }, { "epoch": 0.22093444404201376, "grad_norm": 1.4203715674861919, "learning_rate": 1.8148511316665657e-05, "loss": 0.8966, "step": 6100 }, { "epoch": 0.22097066280333214, "grad_norm": 1.3352057951679213, "learning_rate": 1.8147831268110052e-05, "loss": 0.7759, "step": 6101 }, { "epoch": 0.22100688156465048, "grad_norm": 1.4940020203970905, "learning_rate": 1.8147151107433133e-05, "loss": 0.7814, "step": 6102 }, { "epoch": 0.22104310032596886, "grad_norm": 1.4549388719656755, "learning_rate": 1.8146470834644255e-05, "loss": 0.8563, "step": 6103 }, { "epoch": 0.2210793190872872, "grad_norm": 1.5936465446001524, "learning_rate": 1.8145790449752778e-05, "loss": 0.8584, "step": 6104 }, { "epoch": 0.22111553784860558, "grad_norm": 1.198269177461632, "learning_rate": 1.8145109952768068e-05, "loss": 0.7624, "step": 6105 }, { "epoch": 0.22115175660992395, "grad_norm": 1.5240125115039729, "learning_rate": 1.8144429343699486e-05, "loss": 0.8515, "step": 6106 }, { "epoch": 0.2211879753712423, "grad_norm": 1.374743329322834, "learning_rate": 1.8143748622556404e-05, "loss": 0.7762, "step": 6107 }, { "epoch": 0.22122419413256067, "grad_norm": 1.7170357933925067, "learning_rate": 1.8143067789348183e-05, "loss": 0.806, "step": 6108 }, { "epoch": 0.22126041289387902, "grad_norm": 1.4826943325565407, "learning_rate": 1.8142386844084195e-05, "loss": 0.9159, "step": 6109 }, { "epoch": 0.2212966316551974, "grad_norm": 1.5591433363992842, "learning_rate": 1.814170578677381e-05, "loss": 0.7732, "step": 6110 }, { "epoch": 0.22133285041651576, "grad_norm": 1.5730485542426702, "learning_rate": 1.81410246174264e-05, "loss": 0.7961, "step": 6111 }, { "epoch": 0.2213690691778341, "grad_norm": 1.0209620074767225, "learning_rate": 1.8140343336051334e-05, "loss": 0.7371, "step": 6112 }, { "epoch": 0.22140528793915248, "grad_norm": 1.1348810918397785, "learning_rate": 1.8139661942657995e-05, "loss": 0.8013, "step": 6113 }, { "epoch": 0.22144150670047083, "grad_norm": 1.4302858079652445, "learning_rate": 1.8138980437255754e-05, "loss": 0.7763, "step": 6114 }, { "epoch": 0.2214777254617892, "grad_norm": 1.6073968347705727, "learning_rate": 1.813829881985399e-05, "loss": 0.7571, "step": 6115 }, { "epoch": 0.22151394422310758, "grad_norm": 1.289563904315975, "learning_rate": 1.8137617090462082e-05, "loss": 0.7463, "step": 6116 }, { "epoch": 0.22155016298442592, "grad_norm": 1.2371804612323962, "learning_rate": 1.8136935249089416e-05, "loss": 0.7711, "step": 6117 }, { "epoch": 0.2215863817457443, "grad_norm": 2.083321090403982, "learning_rate": 1.813625329574537e-05, "loss": 0.8866, "step": 6118 }, { "epoch": 0.22162260050706264, "grad_norm": 1.1232584552382021, "learning_rate": 1.813557123043933e-05, "loss": 0.7788, "step": 6119 }, { "epoch": 0.22165881926838102, "grad_norm": 1.66872850017592, "learning_rate": 1.8134889053180677e-05, "loss": 0.8622, "step": 6120 }, { "epoch": 0.2216950380296994, "grad_norm": 1.1397354506262962, "learning_rate": 1.813420676397881e-05, "loss": 0.8189, "step": 6121 }, { "epoch": 0.22173125679101774, "grad_norm": 1.4722723877979964, "learning_rate": 1.8133524362843105e-05, "loss": 0.7487, "step": 6122 }, { "epoch": 0.2217674755523361, "grad_norm": 1.5259438072135176, "learning_rate": 1.8132841849782957e-05, "loss": 0.7827, "step": 6123 }, { "epoch": 0.22180369431365446, "grad_norm": 1.430776143810096, "learning_rate": 1.8132159224807765e-05, "loss": 0.7759, "step": 6124 }, { "epoch": 0.22183991307497283, "grad_norm": 1.488687576185119, "learning_rate": 1.8131476487926914e-05, "loss": 0.7831, "step": 6125 }, { "epoch": 0.2218761318362912, "grad_norm": 1.40873580941971, "learning_rate": 1.8130793639149798e-05, "loss": 0.8091, "step": 6126 }, { "epoch": 0.22191235059760955, "grad_norm": 1.7744830952108117, "learning_rate": 1.813011067848582e-05, "loss": 0.8573, "step": 6127 }, { "epoch": 0.22194856935892793, "grad_norm": 1.59518361126719, "learning_rate": 1.812942760594438e-05, "loss": 0.8818, "step": 6128 }, { "epoch": 0.2219847881202463, "grad_norm": 1.6666537792234062, "learning_rate": 1.812874442153487e-05, "loss": 0.8122, "step": 6129 }, { "epoch": 0.22202100688156465, "grad_norm": 1.384571161406428, "learning_rate": 1.8128061125266686e-05, "loss": 0.7914, "step": 6130 }, { "epoch": 0.22205722564288302, "grad_norm": 1.2311523574571297, "learning_rate": 1.812737771714925e-05, "loss": 0.8552, "step": 6131 }, { "epoch": 0.22209344440420137, "grad_norm": 1.4827869472120865, "learning_rate": 1.8126694197191947e-05, "loss": 0.8823, "step": 6132 }, { "epoch": 0.22212966316551974, "grad_norm": 1.456022118294351, "learning_rate": 1.81260105654042e-05, "loss": 0.8493, "step": 6133 }, { "epoch": 0.22216588192683812, "grad_norm": 1.5031879827722965, "learning_rate": 1.81253268217954e-05, "loss": 0.7727, "step": 6134 }, { "epoch": 0.22220210068815646, "grad_norm": 1.712460249592621, "learning_rate": 1.8124642966374964e-05, "loss": 0.8485, "step": 6135 }, { "epoch": 0.22223831944947484, "grad_norm": 1.3890354724031067, "learning_rate": 1.8123958999152305e-05, "loss": 0.8144, "step": 6136 }, { "epoch": 0.22227453821079318, "grad_norm": 1.4463460390715177, "learning_rate": 1.812327492013683e-05, "loss": 0.8526, "step": 6137 }, { "epoch": 0.22231075697211156, "grad_norm": 1.5371207002952045, "learning_rate": 1.8122590729337957e-05, "loss": 0.77, "step": 6138 }, { "epoch": 0.22234697573342993, "grad_norm": 1.3410962377254252, "learning_rate": 1.8121906426765096e-05, "loss": 0.8038, "step": 6139 }, { "epoch": 0.22238319449474828, "grad_norm": 1.9477040728421766, "learning_rate": 1.8121222012427666e-05, "loss": 0.7873, "step": 6140 }, { "epoch": 0.22241941325606665, "grad_norm": 1.5341290466187154, "learning_rate": 1.8120537486335085e-05, "loss": 0.8558, "step": 6141 }, { "epoch": 0.222455632017385, "grad_norm": 1.5215142593582933, "learning_rate": 1.8119852848496774e-05, "loss": 0.825, "step": 6142 }, { "epoch": 0.22249185077870337, "grad_norm": 1.5445857851948466, "learning_rate": 1.8119168098922152e-05, "loss": 0.814, "step": 6143 }, { "epoch": 0.22252806954002174, "grad_norm": 1.4858996479235376, "learning_rate": 1.8118483237620643e-05, "loss": 0.8854, "step": 6144 }, { "epoch": 0.2225642883013401, "grad_norm": 1.606757943479336, "learning_rate": 1.811779826460167e-05, "loss": 0.8273, "step": 6145 }, { "epoch": 0.22260050706265846, "grad_norm": 1.3550162274565216, "learning_rate": 1.811711317987466e-05, "loss": 0.8491, "step": 6146 }, { "epoch": 0.2226367258239768, "grad_norm": 1.5270240752813624, "learning_rate": 1.8116427983449045e-05, "loss": 0.8508, "step": 6147 }, { "epoch": 0.22267294458529518, "grad_norm": 1.4246750158820185, "learning_rate": 1.8115742675334244e-05, "loss": 0.7564, "step": 6148 }, { "epoch": 0.22270916334661356, "grad_norm": 1.6580827610917301, "learning_rate": 1.8115057255539694e-05, "loss": 0.8728, "step": 6149 }, { "epoch": 0.2227453821079319, "grad_norm": 1.2887141325043139, "learning_rate": 1.8114371724074827e-05, "loss": 0.7782, "step": 6150 }, { "epoch": 0.22278160086925028, "grad_norm": 1.2379500200117557, "learning_rate": 1.8113686080949076e-05, "loss": 0.8053, "step": 6151 }, { "epoch": 0.22281781963056863, "grad_norm": 1.5168750292860924, "learning_rate": 1.8113000326171874e-05, "loss": 0.8036, "step": 6152 }, { "epoch": 0.222854038391887, "grad_norm": 1.4451191308856204, "learning_rate": 1.811231445975266e-05, "loss": 0.7407, "step": 6153 }, { "epoch": 0.22289025715320537, "grad_norm": 1.5571848876306529, "learning_rate": 1.8111628481700865e-05, "loss": 0.7857, "step": 6154 }, { "epoch": 0.22292647591452372, "grad_norm": 1.5299831600729208, "learning_rate": 1.811094239202594e-05, "loss": 0.7945, "step": 6155 }, { "epoch": 0.2229626946758421, "grad_norm": 1.6406434955008513, "learning_rate": 1.811025619073732e-05, "loss": 0.8339, "step": 6156 }, { "epoch": 0.22299891343716044, "grad_norm": 1.6700406093660292, "learning_rate": 1.8109569877844447e-05, "loss": 0.8268, "step": 6157 }, { "epoch": 0.2230351321984788, "grad_norm": 1.6386696721131495, "learning_rate": 1.810888345335677e-05, "loss": 0.8945, "step": 6158 }, { "epoch": 0.2230713509597972, "grad_norm": 1.5524289830250606, "learning_rate": 1.810819691728373e-05, "loss": 0.7595, "step": 6159 }, { "epoch": 0.22310756972111553, "grad_norm": 1.3916955307136778, "learning_rate": 1.8107510269634775e-05, "loss": 0.7883, "step": 6160 }, { "epoch": 0.2231437884824339, "grad_norm": 1.5934146627364707, "learning_rate": 1.8106823510419354e-05, "loss": 0.8012, "step": 6161 }, { "epoch": 0.22318000724375225, "grad_norm": 1.585195902379686, "learning_rate": 1.810613663964692e-05, "loss": 0.8592, "step": 6162 }, { "epoch": 0.22321622600507063, "grad_norm": 1.4632952172650957, "learning_rate": 1.8105449657326924e-05, "loss": 0.8193, "step": 6163 }, { "epoch": 0.223252444766389, "grad_norm": 1.6945122618818305, "learning_rate": 1.8104762563468816e-05, "loss": 0.8449, "step": 6164 }, { "epoch": 0.22328866352770735, "grad_norm": 1.525696563812076, "learning_rate": 1.8104075358082057e-05, "loss": 0.8428, "step": 6165 }, { "epoch": 0.22332488228902572, "grad_norm": 1.7378318162752446, "learning_rate": 1.81033880411761e-05, "loss": 0.8344, "step": 6166 }, { "epoch": 0.22336110105034407, "grad_norm": 1.5545609982221067, "learning_rate": 1.81027006127604e-05, "loss": 0.8968, "step": 6167 }, { "epoch": 0.22339731981166244, "grad_norm": 1.5426799528042967, "learning_rate": 1.810201307284442e-05, "loss": 0.7267, "step": 6168 }, { "epoch": 0.22343353857298082, "grad_norm": 1.4623479307620135, "learning_rate": 1.8101325421437622e-05, "loss": 0.8658, "step": 6169 }, { "epoch": 0.22346975733429916, "grad_norm": 1.512851823353013, "learning_rate": 1.810063765854947e-05, "loss": 0.8776, "step": 6170 }, { "epoch": 0.22350597609561754, "grad_norm": 1.445148555375479, "learning_rate": 1.8099949784189425e-05, "loss": 0.849, "step": 6171 }, { "epoch": 0.22354219485693588, "grad_norm": 1.5861891490493762, "learning_rate": 1.8099261798366952e-05, "loss": 0.7762, "step": 6172 }, { "epoch": 0.22357841361825426, "grad_norm": 1.5243342587649185, "learning_rate": 1.809857370109152e-05, "loss": 0.7717, "step": 6173 }, { "epoch": 0.22361463237957263, "grad_norm": 1.2919072160996312, "learning_rate": 1.80978854923726e-05, "loss": 0.7233, "step": 6174 }, { "epoch": 0.22365085114089098, "grad_norm": 1.281752864936332, "learning_rate": 1.809719717221966e-05, "loss": 0.7652, "step": 6175 }, { "epoch": 0.22368706990220935, "grad_norm": 1.5192796699921685, "learning_rate": 1.809650874064217e-05, "loss": 0.8875, "step": 6176 }, { "epoch": 0.2237232886635277, "grad_norm": 1.4856153388124487, "learning_rate": 1.8095820197649607e-05, "loss": 0.8503, "step": 6177 }, { "epoch": 0.22375950742484607, "grad_norm": 1.465754431248536, "learning_rate": 1.8095131543251442e-05, "loss": 0.7708, "step": 6178 }, { "epoch": 0.22379572618616445, "grad_norm": 1.599934777731647, "learning_rate": 1.809444277745716e-05, "loss": 0.8063, "step": 6179 }, { "epoch": 0.2238319449474828, "grad_norm": 1.4931978629797964, "learning_rate": 1.8093753900276226e-05, "loss": 0.7546, "step": 6180 }, { "epoch": 0.22386816370880117, "grad_norm": 1.4755055609523215, "learning_rate": 1.809306491171813e-05, "loss": 0.8514, "step": 6181 }, { "epoch": 0.2239043824701195, "grad_norm": 1.5009059827363878, "learning_rate": 1.809237581179235e-05, "loss": 0.7247, "step": 6182 }, { "epoch": 0.22394060123143789, "grad_norm": 1.4539005724751484, "learning_rate": 1.809168660050837e-05, "loss": 0.8002, "step": 6183 }, { "epoch": 0.22397681999275626, "grad_norm": 1.590448833763984, "learning_rate": 1.8090997277875665e-05, "loss": 0.8034, "step": 6184 }, { "epoch": 0.2240130387540746, "grad_norm": 1.5089930701986602, "learning_rate": 1.8090307843903738e-05, "loss": 0.8365, "step": 6185 }, { "epoch": 0.22404925751539298, "grad_norm": 1.5572570025584676, "learning_rate": 1.8089618298602058e-05, "loss": 0.8307, "step": 6186 }, { "epoch": 0.22408547627671133, "grad_norm": 1.4153879602976414, "learning_rate": 1.8088928641980127e-05, "loss": 0.8266, "step": 6187 }, { "epoch": 0.2241216950380297, "grad_norm": 1.6376107227587375, "learning_rate": 1.8088238874047428e-05, "loss": 0.8847, "step": 6188 }, { "epoch": 0.22415791379934807, "grad_norm": 1.4915373225787516, "learning_rate": 1.8087548994813455e-05, "loss": 0.7903, "step": 6189 }, { "epoch": 0.22419413256066642, "grad_norm": 2.0223312891840446, "learning_rate": 1.8086859004287704e-05, "loss": 0.8076, "step": 6190 }, { "epoch": 0.2242303513219848, "grad_norm": 1.4598003681898108, "learning_rate": 1.8086168902479665e-05, "loss": 0.733, "step": 6191 }, { "epoch": 0.22426657008330314, "grad_norm": 1.4222724065747305, "learning_rate": 1.8085478689398836e-05, "loss": 0.8686, "step": 6192 }, { "epoch": 0.22430278884462151, "grad_norm": 1.7182783780234827, "learning_rate": 1.8084788365054718e-05, "loss": 0.8576, "step": 6193 }, { "epoch": 0.2243390076059399, "grad_norm": 1.55707007877477, "learning_rate": 1.808409792945681e-05, "loss": 0.9507, "step": 6194 }, { "epoch": 0.22437522636725823, "grad_norm": 1.282905574960856, "learning_rate": 1.8083407382614604e-05, "loss": 0.8348, "step": 6195 }, { "epoch": 0.2244114451285766, "grad_norm": 1.4429976444934798, "learning_rate": 1.8082716724537617e-05, "loss": 0.8444, "step": 6196 }, { "epoch": 0.22444766388989495, "grad_norm": 1.5173479811497743, "learning_rate": 1.8082025955235344e-05, "loss": 0.8434, "step": 6197 }, { "epoch": 0.22448388265121333, "grad_norm": 1.2801783369622126, "learning_rate": 1.808133507471729e-05, "loss": 0.7948, "step": 6198 }, { "epoch": 0.2245201014125317, "grad_norm": 1.4125341393048794, "learning_rate": 1.8080644082992967e-05, "loss": 0.8111, "step": 6199 }, { "epoch": 0.22455632017385005, "grad_norm": 1.3536636491265959, "learning_rate": 1.807995298007188e-05, "loss": 0.7432, "step": 6200 }, { "epoch": 0.22459253893516842, "grad_norm": 1.5462389972691577, "learning_rate": 1.8079261765963537e-05, "loss": 0.8203, "step": 6201 }, { "epoch": 0.22462875769648677, "grad_norm": 1.4840271399037155, "learning_rate": 1.8078570440677457e-05, "loss": 0.8448, "step": 6202 }, { "epoch": 0.22466497645780514, "grad_norm": 2.4961102758227676, "learning_rate": 1.8077879004223148e-05, "loss": 0.868, "step": 6203 }, { "epoch": 0.22470119521912352, "grad_norm": 1.7099045167391276, "learning_rate": 1.8077187456610127e-05, "loss": 0.8448, "step": 6204 }, { "epoch": 0.22473741398044186, "grad_norm": 1.4769562115620605, "learning_rate": 1.807649579784791e-05, "loss": 0.7845, "step": 6205 }, { "epoch": 0.22477363274176024, "grad_norm": 1.2919460001479892, "learning_rate": 1.8075804027946014e-05, "loss": 0.783, "step": 6206 }, { "epoch": 0.22480985150307858, "grad_norm": 1.4294045355889051, "learning_rate": 1.8075112146913957e-05, "loss": 0.8201, "step": 6207 }, { "epoch": 0.22484607026439696, "grad_norm": 1.4187458932602888, "learning_rate": 1.8074420154761263e-05, "loss": 0.794, "step": 6208 }, { "epoch": 0.22488228902571533, "grad_norm": 1.5215504868213932, "learning_rate": 1.807372805149745e-05, "loss": 0.8956, "step": 6209 }, { "epoch": 0.22491850778703368, "grad_norm": 1.5184816717029932, "learning_rate": 1.8073035837132047e-05, "loss": 0.8401, "step": 6210 }, { "epoch": 0.22495472654835205, "grad_norm": 1.4786560408912963, "learning_rate": 1.8072343511674576e-05, "loss": 0.9042, "step": 6211 }, { "epoch": 0.2249909453096704, "grad_norm": 1.4387800986709027, "learning_rate": 1.8071651075134568e-05, "loss": 0.8102, "step": 6212 }, { "epoch": 0.22502716407098877, "grad_norm": 1.4369603652410823, "learning_rate": 1.8070958527521547e-05, "loss": 0.7962, "step": 6213 }, { "epoch": 0.22506338283230715, "grad_norm": 1.6867524650317405, "learning_rate": 1.8070265868845048e-05, "loss": 0.7808, "step": 6214 }, { "epoch": 0.2250996015936255, "grad_norm": 1.3675013066236312, "learning_rate": 1.8069573099114595e-05, "loss": 0.8946, "step": 6215 }, { "epoch": 0.22513582035494387, "grad_norm": 1.4911148087770205, "learning_rate": 1.8068880218339728e-05, "loss": 0.8181, "step": 6216 }, { "epoch": 0.2251720391162622, "grad_norm": 1.581779861492468, "learning_rate": 1.806818722652998e-05, "loss": 0.862, "step": 6217 }, { "epoch": 0.2252082578775806, "grad_norm": 1.5317061593868029, "learning_rate": 1.8067494123694884e-05, "loss": 0.872, "step": 6218 }, { "epoch": 0.22524447663889896, "grad_norm": 1.1991750739497287, "learning_rate": 1.806680090984398e-05, "loss": 0.7466, "step": 6219 }, { "epoch": 0.2252806954002173, "grad_norm": 1.3453917943430103, "learning_rate": 1.806610758498681e-05, "loss": 0.8124, "step": 6220 }, { "epoch": 0.22531691416153568, "grad_norm": 1.298234766036047, "learning_rate": 1.806541414913291e-05, "loss": 0.8164, "step": 6221 }, { "epoch": 0.22535313292285403, "grad_norm": 2.1009283066105424, "learning_rate": 1.8064720602291825e-05, "loss": 0.8536, "step": 6222 }, { "epoch": 0.2253893516841724, "grad_norm": 1.5678610027681108, "learning_rate": 1.80640269444731e-05, "loss": 0.8368, "step": 6223 }, { "epoch": 0.22542557044549077, "grad_norm": 1.357302270738317, "learning_rate": 1.8063333175686275e-05, "loss": 0.7633, "step": 6224 }, { "epoch": 0.22546178920680912, "grad_norm": 1.4532777221471993, "learning_rate": 1.8062639295940903e-05, "loss": 0.8574, "step": 6225 }, { "epoch": 0.2254980079681275, "grad_norm": 1.5043295227486155, "learning_rate": 1.8061945305246527e-05, "loss": 0.7998, "step": 6226 }, { "epoch": 0.22553422672944584, "grad_norm": 1.5721628031073251, "learning_rate": 1.8061251203612702e-05, "loss": 0.8137, "step": 6227 }, { "epoch": 0.22557044549076422, "grad_norm": 1.859892465357952, "learning_rate": 1.806055699104898e-05, "loss": 0.8145, "step": 6228 }, { "epoch": 0.2256066642520826, "grad_norm": 1.2271659984736885, "learning_rate": 1.8059862667564906e-05, "loss": 0.7856, "step": 6229 }, { "epoch": 0.22564288301340094, "grad_norm": 1.5851352863702235, "learning_rate": 1.8059168233170045e-05, "loss": 0.9061, "step": 6230 }, { "epoch": 0.2256791017747193, "grad_norm": 1.54733926736983, "learning_rate": 1.805847368787394e-05, "loss": 0.8805, "step": 6231 }, { "epoch": 0.22571532053603766, "grad_norm": 1.6194546916051809, "learning_rate": 1.8057779031686162e-05, "loss": 0.9061, "step": 6232 }, { "epoch": 0.22575153929735603, "grad_norm": 1.4694117262731674, "learning_rate": 1.8057084264616264e-05, "loss": 0.8055, "step": 6233 }, { "epoch": 0.2257877580586744, "grad_norm": 1.4441386985095266, "learning_rate": 1.8056389386673805e-05, "loss": 0.7719, "step": 6234 }, { "epoch": 0.22582397681999275, "grad_norm": 1.3532078406453438, "learning_rate": 1.8055694397868354e-05, "loss": 0.8284, "step": 6235 }, { "epoch": 0.22586019558131112, "grad_norm": 1.438334915790658, "learning_rate": 1.8054999298209465e-05, "loss": 0.8175, "step": 6236 }, { "epoch": 0.22589641434262947, "grad_norm": 1.420028100113475, "learning_rate": 1.8054304087706706e-05, "loss": 0.8372, "step": 6237 }, { "epoch": 0.22593263310394784, "grad_norm": 2.0511361387267573, "learning_rate": 1.805360876636965e-05, "loss": 0.7802, "step": 6238 }, { "epoch": 0.22596885186526622, "grad_norm": 1.4955071358730883, "learning_rate": 1.805291333420786e-05, "loss": 0.8069, "step": 6239 }, { "epoch": 0.22600507062658456, "grad_norm": 1.4141809969722299, "learning_rate": 1.8052217791230908e-05, "loss": 0.8778, "step": 6240 }, { "epoch": 0.22604128938790294, "grad_norm": 1.2008370943794775, "learning_rate": 1.805152213744836e-05, "loss": 0.7538, "step": 6241 }, { "epoch": 0.22607750814922128, "grad_norm": 1.619422877642642, "learning_rate": 1.8050826372869793e-05, "loss": 0.8276, "step": 6242 }, { "epoch": 0.22611372691053966, "grad_norm": 1.583215762204876, "learning_rate": 1.8050130497504782e-05, "loss": 0.7564, "step": 6243 }, { "epoch": 0.22614994567185803, "grad_norm": 1.5093485859520341, "learning_rate": 1.8049434511362904e-05, "loss": 0.8419, "step": 6244 }, { "epoch": 0.22618616443317638, "grad_norm": 1.39504477643639, "learning_rate": 1.8048738414453732e-05, "loss": 0.7965, "step": 6245 }, { "epoch": 0.22622238319449475, "grad_norm": 1.148760074467302, "learning_rate": 1.804804220678685e-05, "loss": 0.7718, "step": 6246 }, { "epoch": 0.2262586019558131, "grad_norm": 1.5163595039575715, "learning_rate": 1.8047345888371833e-05, "loss": 0.8296, "step": 6247 }, { "epoch": 0.22629482071713147, "grad_norm": 1.532919606035502, "learning_rate": 1.8046649459218267e-05, "loss": 0.8929, "step": 6248 }, { "epoch": 0.22633103947844985, "grad_norm": 1.4923911479971281, "learning_rate": 1.8045952919335732e-05, "loss": 0.8953, "step": 6249 }, { "epoch": 0.2263672582397682, "grad_norm": 1.5528727613716, "learning_rate": 1.8045256268733818e-05, "loss": 0.7951, "step": 6250 }, { "epoch": 0.22640347700108657, "grad_norm": 1.4949073195714493, "learning_rate": 1.8044559507422105e-05, "loss": 0.85, "step": 6251 }, { "epoch": 0.2264396957624049, "grad_norm": 1.2128114113917057, "learning_rate": 1.8043862635410186e-05, "loss": 0.7545, "step": 6252 }, { "epoch": 0.2264759145237233, "grad_norm": 1.607301109570124, "learning_rate": 1.804316565270765e-05, "loss": 0.7754, "step": 6253 }, { "epoch": 0.22651213328504166, "grad_norm": 1.5105961036700095, "learning_rate": 1.8042468559324086e-05, "loss": 0.8175, "step": 6254 }, { "epoch": 0.22654835204636, "grad_norm": 1.474450633365707, "learning_rate": 1.804177135526909e-05, "loss": 0.7913, "step": 6255 }, { "epoch": 0.22658457080767838, "grad_norm": 1.4600492463467434, "learning_rate": 1.8041074040552253e-05, "loss": 0.8207, "step": 6256 }, { "epoch": 0.22662078956899673, "grad_norm": 1.3072970911876551, "learning_rate": 1.804037661518317e-05, "loss": 0.8048, "step": 6257 }, { "epoch": 0.2266570083303151, "grad_norm": 1.546222004671703, "learning_rate": 1.8039679079171443e-05, "loss": 0.8457, "step": 6258 }, { "epoch": 0.22669322709163348, "grad_norm": 4.722880545086649, "learning_rate": 1.8038981432526664e-05, "loss": 0.849, "step": 6259 }, { "epoch": 0.22672944585295182, "grad_norm": 1.5874464791016671, "learning_rate": 1.803828367525844e-05, "loss": 0.8001, "step": 6260 }, { "epoch": 0.2267656646142702, "grad_norm": 1.4266377750374737, "learning_rate": 1.8037585807376366e-05, "loss": 0.8034, "step": 6261 }, { "epoch": 0.22680188337558854, "grad_norm": 1.6828305734037536, "learning_rate": 1.803688782889005e-05, "loss": 0.8895, "step": 6262 }, { "epoch": 0.22683810213690692, "grad_norm": 1.5249816862896437, "learning_rate": 1.80361897398091e-05, "loss": 0.7541, "step": 6263 }, { "epoch": 0.2268743208982253, "grad_norm": 1.637351155159207, "learning_rate": 1.8035491540143114e-05, "loss": 0.8515, "step": 6264 }, { "epoch": 0.22691053965954364, "grad_norm": 1.5297754410996227, "learning_rate": 1.8034793229901706e-05, "loss": 0.7966, "step": 6265 }, { "epoch": 0.226946758420862, "grad_norm": 1.671916488774399, "learning_rate": 1.803409480909448e-05, "loss": 0.844, "step": 6266 }, { "epoch": 0.22698297718218036, "grad_norm": 1.5315705007753926, "learning_rate": 1.8033396277731053e-05, "loss": 0.8524, "step": 6267 }, { "epoch": 0.22701919594349873, "grad_norm": 1.2096532654848389, "learning_rate": 1.8032697635821033e-05, "loss": 0.7655, "step": 6268 }, { "epoch": 0.2270554147048171, "grad_norm": 1.475893603930504, "learning_rate": 1.8031998883374036e-05, "loss": 0.8584, "step": 6269 }, { "epoch": 0.22709163346613545, "grad_norm": 1.4329795028200725, "learning_rate": 1.8031300020399675e-05, "loss": 0.79, "step": 6270 }, { "epoch": 0.22712785222745382, "grad_norm": 1.2475576813203644, "learning_rate": 1.803060104690757e-05, "loss": 0.7801, "step": 6271 }, { "epoch": 0.22716407098877217, "grad_norm": 1.5512145647751643, "learning_rate": 1.802990196290734e-05, "loss": 0.8409, "step": 6272 }, { "epoch": 0.22720028975009054, "grad_norm": 1.5641110745184081, "learning_rate": 1.80292027684086e-05, "loss": 0.866, "step": 6273 }, { "epoch": 0.22723650851140892, "grad_norm": 1.3900928389896563, "learning_rate": 1.8028503463420975e-05, "loss": 0.7559, "step": 6274 }, { "epoch": 0.22727272727272727, "grad_norm": 1.2645625404234937, "learning_rate": 1.802780404795409e-05, "loss": 0.743, "step": 6275 }, { "epoch": 0.22730894603404564, "grad_norm": 1.617017658732013, "learning_rate": 1.802710452201757e-05, "loss": 0.7532, "step": 6276 }, { "epoch": 0.22734516479536399, "grad_norm": 1.4270166133652937, "learning_rate": 1.8026404885621036e-05, "loss": 0.8162, "step": 6277 }, { "epoch": 0.22738138355668236, "grad_norm": 1.5782167530769664, "learning_rate": 1.8025705138774117e-05, "loss": 0.8401, "step": 6278 }, { "epoch": 0.22741760231800073, "grad_norm": 1.3727400899586517, "learning_rate": 1.8025005281486444e-05, "loss": 0.859, "step": 6279 }, { "epoch": 0.22745382107931908, "grad_norm": 1.3052490776080126, "learning_rate": 1.8024305313767648e-05, "loss": 0.8062, "step": 6280 }, { "epoch": 0.22749003984063745, "grad_norm": 1.5198185558393749, "learning_rate": 1.8023605235627356e-05, "loss": 0.8648, "step": 6281 }, { "epoch": 0.2275262586019558, "grad_norm": 1.4296085096123468, "learning_rate": 1.802290504707521e-05, "loss": 0.7933, "step": 6282 }, { "epoch": 0.22756247736327417, "grad_norm": 1.5021054918565664, "learning_rate": 1.802220474812084e-05, "loss": 0.8615, "step": 6283 }, { "epoch": 0.22759869612459255, "grad_norm": 1.5247663558826863, "learning_rate": 1.8021504338773884e-05, "loss": 0.7851, "step": 6284 }, { "epoch": 0.2276349148859109, "grad_norm": 1.5698035800935723, "learning_rate": 1.802080381904398e-05, "loss": 0.8177, "step": 6285 }, { "epoch": 0.22767113364722927, "grad_norm": 1.4379247829491126, "learning_rate": 1.8020103188940767e-05, "loss": 0.811, "step": 6286 }, { "epoch": 0.22770735240854761, "grad_norm": 1.4938001640992047, "learning_rate": 1.8019402448473887e-05, "loss": 0.7905, "step": 6287 }, { "epoch": 0.227743571169866, "grad_norm": 1.3763621908708594, "learning_rate": 1.8018701597652987e-05, "loss": 0.7524, "step": 6288 }, { "epoch": 0.22777978993118436, "grad_norm": 1.6914600689719308, "learning_rate": 1.80180006364877e-05, "loss": 0.8413, "step": 6289 }, { "epoch": 0.2278160086925027, "grad_norm": 1.4545399893226552, "learning_rate": 1.801729956498768e-05, "loss": 0.8088, "step": 6290 }, { "epoch": 0.22785222745382108, "grad_norm": 1.5452753812769544, "learning_rate": 1.8016598383162578e-05, "loss": 0.8693, "step": 6291 }, { "epoch": 0.22788844621513943, "grad_norm": 1.3014446076355353, "learning_rate": 1.8015897091022035e-05, "loss": 0.7861, "step": 6292 }, { "epoch": 0.2279246649764578, "grad_norm": 1.3233845211913027, "learning_rate": 1.801519568857571e-05, "loss": 0.7835, "step": 6293 }, { "epoch": 0.22796088373777618, "grad_norm": 1.6327202179945395, "learning_rate": 1.8014494175833244e-05, "loss": 0.796, "step": 6294 }, { "epoch": 0.22799710249909452, "grad_norm": 1.4828824507631366, "learning_rate": 1.8013792552804298e-05, "loss": 0.8578, "step": 6295 }, { "epoch": 0.2280333212604129, "grad_norm": 1.6287563469976098, "learning_rate": 1.8013090819498523e-05, "loss": 0.8695, "step": 6296 }, { "epoch": 0.22806954002173124, "grad_norm": 1.3419519873394008, "learning_rate": 1.801238897592558e-05, "loss": 0.7626, "step": 6297 }, { "epoch": 0.22810575878304962, "grad_norm": 1.2840404551962332, "learning_rate": 1.8011687022095122e-05, "loss": 0.8284, "step": 6298 }, { "epoch": 0.228141977544368, "grad_norm": 1.4924948460683223, "learning_rate": 1.8010984958016812e-05, "loss": 0.8262, "step": 6299 }, { "epoch": 0.22817819630568634, "grad_norm": 1.2983101623150843, "learning_rate": 1.8010282783700307e-05, "loss": 0.7855, "step": 6300 }, { "epoch": 0.2282144150670047, "grad_norm": 1.5882904375207423, "learning_rate": 1.8009580499155276e-05, "loss": 0.8225, "step": 6301 }, { "epoch": 0.22825063382832309, "grad_norm": 1.8453797011611086, "learning_rate": 1.8008878104391382e-05, "loss": 0.8199, "step": 6302 }, { "epoch": 0.22828685258964143, "grad_norm": 1.359846738238702, "learning_rate": 1.800817559941828e-05, "loss": 0.7519, "step": 6303 }, { "epoch": 0.2283230713509598, "grad_norm": 1.295648182347233, "learning_rate": 1.800747298424565e-05, "loss": 0.718, "step": 6304 }, { "epoch": 0.22835929011227815, "grad_norm": 1.524834513186257, "learning_rate": 1.800677025888316e-05, "loss": 0.6583, "step": 6305 }, { "epoch": 0.22839550887359653, "grad_norm": 1.4992034642472485, "learning_rate": 1.800606742334047e-05, "loss": 0.9047, "step": 6306 }, { "epoch": 0.2284317276349149, "grad_norm": 1.4569407261092742, "learning_rate": 1.800536447762726e-05, "loss": 0.7902, "step": 6307 }, { "epoch": 0.22846794639623325, "grad_norm": 1.3214296208381169, "learning_rate": 1.80046614217532e-05, "loss": 0.8161, "step": 6308 }, { "epoch": 0.22850416515755162, "grad_norm": 1.479686714200352, "learning_rate": 1.8003958255727965e-05, "loss": 0.7343, "step": 6309 }, { "epoch": 0.22854038391886997, "grad_norm": 1.2786076170398821, "learning_rate": 1.800325497956123e-05, "loss": 0.7694, "step": 6310 }, { "epoch": 0.22857660268018834, "grad_norm": 1.4297551806121407, "learning_rate": 1.8002551593262675e-05, "loss": 0.7979, "step": 6311 }, { "epoch": 0.22861282144150671, "grad_norm": 1.3577414513799797, "learning_rate": 1.800184809684198e-05, "loss": 0.8037, "step": 6312 }, { "epoch": 0.22864904020282506, "grad_norm": 1.5039767292153932, "learning_rate": 1.8001144490308823e-05, "loss": 0.7505, "step": 6313 }, { "epoch": 0.22868525896414343, "grad_norm": 1.4211338718267545, "learning_rate": 1.8000440773672888e-05, "loss": 0.8026, "step": 6314 }, { "epoch": 0.22872147772546178, "grad_norm": 1.583055037794787, "learning_rate": 1.7999736946943856e-05, "loss": 0.8575, "step": 6315 }, { "epoch": 0.22875769648678015, "grad_norm": 1.072751934655678, "learning_rate": 1.7999033010131417e-05, "loss": 0.8065, "step": 6316 }, { "epoch": 0.22879391524809853, "grad_norm": 1.5203416252090802, "learning_rate": 1.7998328963245254e-05, "loss": 0.7389, "step": 6317 }, { "epoch": 0.22883013400941687, "grad_norm": 1.5754386917223553, "learning_rate": 1.799762480629506e-05, "loss": 0.7902, "step": 6318 }, { "epoch": 0.22886635277073525, "grad_norm": 1.5769759423374796, "learning_rate": 1.7996920539290518e-05, "loss": 0.7546, "step": 6319 }, { "epoch": 0.2289025715320536, "grad_norm": 1.3148182115557971, "learning_rate": 1.7996216162241324e-05, "loss": 0.8043, "step": 6320 }, { "epoch": 0.22893879029337197, "grad_norm": 1.2204412649623084, "learning_rate": 1.7995511675157166e-05, "loss": 0.8041, "step": 6321 }, { "epoch": 0.22897500905469034, "grad_norm": 1.8355967895416925, "learning_rate": 1.7994807078047745e-05, "loss": 0.8125, "step": 6322 }, { "epoch": 0.2290112278160087, "grad_norm": 1.1840745059983537, "learning_rate": 1.799410237092275e-05, "loss": 0.8186, "step": 6323 }, { "epoch": 0.22904744657732706, "grad_norm": 1.5918709029787868, "learning_rate": 1.799339755379189e-05, "loss": 0.8426, "step": 6324 }, { "epoch": 0.2290836653386454, "grad_norm": 1.169693797436549, "learning_rate": 1.799269262666485e-05, "loss": 0.7672, "step": 6325 }, { "epoch": 0.22911988409996378, "grad_norm": 1.4450368244579799, "learning_rate": 1.7991987589551338e-05, "loss": 0.8024, "step": 6326 }, { "epoch": 0.22915610286128216, "grad_norm": 1.4156767141142819, "learning_rate": 1.7991282442461052e-05, "loss": 0.8402, "step": 6327 }, { "epoch": 0.2291923216226005, "grad_norm": 1.0421911124764387, "learning_rate": 1.7990577185403702e-05, "loss": 0.7206, "step": 6328 }, { "epoch": 0.22922854038391888, "grad_norm": 1.083263224671421, "learning_rate": 1.7989871818388986e-05, "loss": 0.7672, "step": 6329 }, { "epoch": 0.22926475914523722, "grad_norm": 1.4582897611194119, "learning_rate": 1.7989166341426615e-05, "loss": 0.8483, "step": 6330 }, { "epoch": 0.2293009779065556, "grad_norm": 1.514735066895925, "learning_rate": 1.7988460754526293e-05, "loss": 0.8377, "step": 6331 }, { "epoch": 0.22933719666787397, "grad_norm": 1.344328533985214, "learning_rate": 1.7987755057697734e-05, "loss": 0.8366, "step": 6332 }, { "epoch": 0.22937341542919232, "grad_norm": 1.4777586219768724, "learning_rate": 1.7987049250950644e-05, "loss": 0.8225, "step": 6333 }, { "epoch": 0.2294096341905107, "grad_norm": 1.6054379629631292, "learning_rate": 1.798634333429474e-05, "loss": 0.8398, "step": 6334 }, { "epoch": 0.22944585295182904, "grad_norm": 1.3810912879050552, "learning_rate": 1.7985637307739735e-05, "loss": 0.8089, "step": 6335 }, { "epoch": 0.2294820717131474, "grad_norm": 1.4666366972241014, "learning_rate": 1.7984931171295344e-05, "loss": 0.8839, "step": 6336 }, { "epoch": 0.22951829047446579, "grad_norm": 1.4728591820864128, "learning_rate": 1.7984224924971287e-05, "loss": 0.8191, "step": 6337 }, { "epoch": 0.22955450923578413, "grad_norm": 1.5350136824162688, "learning_rate": 1.7983518568777275e-05, "loss": 0.7544, "step": 6338 }, { "epoch": 0.2295907279971025, "grad_norm": 1.5664926110335982, "learning_rate": 1.798281210272303e-05, "loss": 0.9006, "step": 6339 }, { "epoch": 0.22962694675842085, "grad_norm": 1.416130905391573, "learning_rate": 1.798210552681828e-05, "loss": 0.8024, "step": 6340 }, { "epoch": 0.22966316551973923, "grad_norm": 1.1732034476952238, "learning_rate": 1.7981398841072747e-05, "loss": 0.7731, "step": 6341 }, { "epoch": 0.2296993842810576, "grad_norm": 1.1123450761613625, "learning_rate": 1.7980692045496146e-05, "loss": 0.7544, "step": 6342 }, { "epoch": 0.22973560304237595, "grad_norm": 1.6208109351145374, "learning_rate": 1.7979985140098212e-05, "loss": 0.7782, "step": 6343 }, { "epoch": 0.22977182180369432, "grad_norm": 1.3339654563690495, "learning_rate": 1.7979278124888675e-05, "loss": 0.7786, "step": 6344 }, { "epoch": 0.22980804056501267, "grad_norm": 1.568443433853359, "learning_rate": 1.7978570999877256e-05, "loss": 0.8251, "step": 6345 }, { "epoch": 0.22984425932633104, "grad_norm": 1.410672745070941, "learning_rate": 1.7977863765073692e-05, "loss": 0.8162, "step": 6346 }, { "epoch": 0.22988047808764941, "grad_norm": 1.513091894832104, "learning_rate": 1.797715642048771e-05, "loss": 0.8443, "step": 6347 }, { "epoch": 0.22991669684896776, "grad_norm": 1.4679601293444455, "learning_rate": 1.7976448966129047e-05, "loss": 0.7709, "step": 6348 }, { "epoch": 0.22995291561028613, "grad_norm": 1.4157894585302746, "learning_rate": 1.7975741402007438e-05, "loss": 0.8028, "step": 6349 }, { "epoch": 0.22998913437160448, "grad_norm": 1.4955322504257642, "learning_rate": 1.7975033728132622e-05, "loss": 0.8527, "step": 6350 }, { "epoch": 0.23002535313292286, "grad_norm": 1.4145405383491405, "learning_rate": 1.7974325944514327e-05, "loss": 0.9258, "step": 6351 }, { "epoch": 0.23006157189424123, "grad_norm": 1.62463186981778, "learning_rate": 1.7973618051162307e-05, "loss": 0.8954, "step": 6352 }, { "epoch": 0.23009779065555958, "grad_norm": 1.5351539810785122, "learning_rate": 1.7972910048086294e-05, "loss": 0.8653, "step": 6353 }, { "epoch": 0.23013400941687795, "grad_norm": 1.3657451848724749, "learning_rate": 1.7972201935296033e-05, "loss": 0.7147, "step": 6354 }, { "epoch": 0.2301702281781963, "grad_norm": 1.6549347590101753, "learning_rate": 1.7971493712801268e-05, "loss": 0.8158, "step": 6355 }, { "epoch": 0.23020644693951467, "grad_norm": 1.505544245434891, "learning_rate": 1.7970785380611747e-05, "loss": 0.8483, "step": 6356 }, { "epoch": 0.23024266570083304, "grad_norm": 1.4700466618523813, "learning_rate": 1.7970076938737214e-05, "loss": 0.8025, "step": 6357 }, { "epoch": 0.2302788844621514, "grad_norm": 1.4874968213660473, "learning_rate": 1.7969368387187417e-05, "loss": 0.7835, "step": 6358 }, { "epoch": 0.23031510322346976, "grad_norm": 1.517442148516365, "learning_rate": 1.7968659725972113e-05, "loss": 0.8071, "step": 6359 }, { "epoch": 0.2303513219847881, "grad_norm": 1.4463602540488254, "learning_rate": 1.7967950955101047e-05, "loss": 0.7946, "step": 6360 }, { "epoch": 0.23038754074610648, "grad_norm": 1.5658882930502291, "learning_rate": 1.7967242074583976e-05, "loss": 0.9042, "step": 6361 }, { "epoch": 0.23042375950742486, "grad_norm": 1.4802920870191891, "learning_rate": 1.796653308443065e-05, "loss": 0.8398, "step": 6362 }, { "epoch": 0.2304599782687432, "grad_norm": 1.4883702550641267, "learning_rate": 1.7965823984650834e-05, "loss": 0.8355, "step": 6363 }, { "epoch": 0.23049619703006158, "grad_norm": 1.4232063999730955, "learning_rate": 1.7965114775254274e-05, "loss": 0.9116, "step": 6364 }, { "epoch": 0.23053241579137992, "grad_norm": 1.5539673384477206, "learning_rate": 1.796440545625074e-05, "loss": 0.8822, "step": 6365 }, { "epoch": 0.2305686345526983, "grad_norm": 1.4930573806425025, "learning_rate": 1.796369602764999e-05, "loss": 0.7756, "step": 6366 }, { "epoch": 0.23060485331401667, "grad_norm": 1.583303633341264, "learning_rate": 1.796298648946178e-05, "loss": 0.9164, "step": 6367 }, { "epoch": 0.23064107207533502, "grad_norm": 1.5409337240775112, "learning_rate": 1.7962276841695884e-05, "loss": 0.8097, "step": 6368 }, { "epoch": 0.2306772908366534, "grad_norm": 1.5540758887568933, "learning_rate": 1.7961567084362063e-05, "loss": 0.824, "step": 6369 }, { "epoch": 0.23071350959797174, "grad_norm": 1.3780211358170127, "learning_rate": 1.7960857217470077e-05, "loss": 0.871, "step": 6370 }, { "epoch": 0.2307497283592901, "grad_norm": 1.463195579995664, "learning_rate": 1.7960147241029706e-05, "loss": 0.8356, "step": 6371 }, { "epoch": 0.2307859471206085, "grad_norm": 1.4749347706692744, "learning_rate": 1.7959437155050713e-05, "loss": 0.8577, "step": 6372 }, { "epoch": 0.23082216588192683, "grad_norm": 1.499386972587873, "learning_rate": 1.795872695954287e-05, "loss": 0.8109, "step": 6373 }, { "epoch": 0.2308583846432452, "grad_norm": 1.6976973820718466, "learning_rate": 1.7958016654515952e-05, "loss": 0.8096, "step": 6374 }, { "epoch": 0.23089460340456355, "grad_norm": 1.3679692131664771, "learning_rate": 1.7957306239979733e-05, "loss": 0.8166, "step": 6375 }, { "epoch": 0.23093082216588193, "grad_norm": 1.473726394745768, "learning_rate": 1.7956595715943987e-05, "loss": 0.8608, "step": 6376 }, { "epoch": 0.2309670409272003, "grad_norm": 1.4378965714488954, "learning_rate": 1.7955885082418495e-05, "loss": 0.7976, "step": 6377 }, { "epoch": 0.23100325968851865, "grad_norm": 1.4031739234891096, "learning_rate": 1.795517433941303e-05, "loss": 0.7685, "step": 6378 }, { "epoch": 0.23103947844983702, "grad_norm": 1.4025362116470335, "learning_rate": 1.795446348693738e-05, "loss": 0.7882, "step": 6379 }, { "epoch": 0.23107569721115537, "grad_norm": 1.4797389488782333, "learning_rate": 1.795375252500132e-05, "loss": 0.8379, "step": 6380 }, { "epoch": 0.23111191597247374, "grad_norm": 1.405767703797193, "learning_rate": 1.7953041453614635e-05, "loss": 0.7926, "step": 6381 }, { "epoch": 0.23114813473379212, "grad_norm": 1.2703076940511717, "learning_rate": 1.7952330272787116e-05, "loss": 0.8501, "step": 6382 }, { "epoch": 0.23118435349511046, "grad_norm": 1.5643872546021953, "learning_rate": 1.7951618982528544e-05, "loss": 0.8046, "step": 6383 }, { "epoch": 0.23122057225642884, "grad_norm": 1.64245804156095, "learning_rate": 1.7950907582848708e-05, "loss": 0.8368, "step": 6384 }, { "epoch": 0.23125679101774718, "grad_norm": 1.530533683863712, "learning_rate": 1.795019607375739e-05, "loss": 0.8513, "step": 6385 }, { "epoch": 0.23129300977906556, "grad_norm": 1.3352886981312178, "learning_rate": 1.7949484455264398e-05, "loss": 0.7754, "step": 6386 }, { "epoch": 0.23132922854038393, "grad_norm": 1.4614422654159993, "learning_rate": 1.794877272737951e-05, "loss": 0.8128, "step": 6387 }, { "epoch": 0.23136544730170228, "grad_norm": 1.139207852048157, "learning_rate": 1.7948060890112527e-05, "loss": 0.7525, "step": 6388 }, { "epoch": 0.23140166606302065, "grad_norm": 1.4265758769169639, "learning_rate": 1.794734894347324e-05, "loss": 0.8389, "step": 6389 }, { "epoch": 0.231437884824339, "grad_norm": 1.132603358362992, "learning_rate": 1.7946636887471453e-05, "loss": 0.769, "step": 6390 }, { "epoch": 0.23147410358565737, "grad_norm": 1.5700494963518692, "learning_rate": 1.7945924722116956e-05, "loss": 0.7925, "step": 6391 }, { "epoch": 0.23151032234697574, "grad_norm": 1.2681710809518276, "learning_rate": 1.7945212447419552e-05, "loss": 0.8119, "step": 6392 }, { "epoch": 0.2315465411082941, "grad_norm": 1.5774354917046598, "learning_rate": 1.7944500063389046e-05, "loss": 0.7872, "step": 6393 }, { "epoch": 0.23158275986961246, "grad_norm": 1.4574736480902954, "learning_rate": 1.7943787570035237e-05, "loss": 0.8697, "step": 6394 }, { "epoch": 0.2316189786309308, "grad_norm": 1.1258679245113687, "learning_rate": 1.7943074967367936e-05, "loss": 0.7892, "step": 6395 }, { "epoch": 0.23165519739224918, "grad_norm": 1.5944918490732631, "learning_rate": 1.7942362255396936e-05, "loss": 0.8744, "step": 6396 }, { "epoch": 0.23169141615356756, "grad_norm": 1.5459028269259032, "learning_rate": 1.7941649434132058e-05, "loss": 0.8429, "step": 6397 }, { "epoch": 0.2317276349148859, "grad_norm": 1.4667609549440506, "learning_rate": 1.7940936503583105e-05, "loss": 0.8404, "step": 6398 }, { "epoch": 0.23176385367620428, "grad_norm": 1.4474136023347555, "learning_rate": 1.7940223463759884e-05, "loss": 0.7567, "step": 6399 }, { "epoch": 0.23180007243752263, "grad_norm": 1.3452018249075595, "learning_rate": 1.793951031467222e-05, "loss": 0.811, "step": 6400 }, { "epoch": 0.231836291198841, "grad_norm": 1.3291958343474544, "learning_rate": 1.7938797056329907e-05, "loss": 0.759, "step": 6401 }, { "epoch": 0.23187250996015937, "grad_norm": 1.3225466056017887, "learning_rate": 1.7938083688742778e-05, "loss": 0.8296, "step": 6402 }, { "epoch": 0.23190872872147772, "grad_norm": 1.559896817124767, "learning_rate": 1.793737021192064e-05, "loss": 0.8306, "step": 6403 }, { "epoch": 0.2319449474827961, "grad_norm": 1.412003432055041, "learning_rate": 1.7936656625873316e-05, "loss": 0.7983, "step": 6404 }, { "epoch": 0.23198116624411444, "grad_norm": 1.4697162817915455, "learning_rate": 1.793594293061062e-05, "loss": 0.8311, "step": 6405 }, { "epoch": 0.2320173850054328, "grad_norm": 1.408364182916498, "learning_rate": 1.7935229126142376e-05, "loss": 0.8163, "step": 6406 }, { "epoch": 0.2320536037667512, "grad_norm": 1.4218209041533378, "learning_rate": 1.7934515212478408e-05, "loss": 0.7929, "step": 6407 }, { "epoch": 0.23208982252806953, "grad_norm": 1.6858552339049446, "learning_rate": 1.793380118962854e-05, "loss": 0.7768, "step": 6408 }, { "epoch": 0.2321260412893879, "grad_norm": 1.7524663562545058, "learning_rate": 1.7933087057602595e-05, "loss": 0.8527, "step": 6409 }, { "epoch": 0.23216226005070625, "grad_norm": 1.2007325024007731, "learning_rate": 1.79323728164104e-05, "loss": 0.8024, "step": 6410 }, { "epoch": 0.23219847881202463, "grad_norm": 1.4678403892399432, "learning_rate": 1.7931658466061788e-05, "loss": 0.8447, "step": 6411 }, { "epoch": 0.232234697573343, "grad_norm": 1.8123855544411827, "learning_rate": 1.7930944006566584e-05, "loss": 0.8538, "step": 6412 }, { "epoch": 0.23227091633466135, "grad_norm": 1.5158128880025405, "learning_rate": 1.7930229437934624e-05, "loss": 0.7531, "step": 6413 }, { "epoch": 0.23230713509597972, "grad_norm": 1.48914404450163, "learning_rate": 1.7929514760175736e-05, "loss": 0.87, "step": 6414 }, { "epoch": 0.23234335385729807, "grad_norm": 1.3085250308903122, "learning_rate": 1.7928799973299756e-05, "loss": 0.8428, "step": 6415 }, { "epoch": 0.23237957261861644, "grad_norm": 1.2499408305122701, "learning_rate": 1.7928085077316528e-05, "loss": 0.7849, "step": 6416 }, { "epoch": 0.23241579137993482, "grad_norm": 1.414093545446444, "learning_rate": 1.792737007223588e-05, "loss": 0.7903, "step": 6417 }, { "epoch": 0.23245201014125316, "grad_norm": 1.4505072993418158, "learning_rate": 1.792665495806765e-05, "loss": 0.8763, "step": 6418 }, { "epoch": 0.23248822890257154, "grad_norm": 1.4705282698814555, "learning_rate": 1.792593973482169e-05, "loss": 0.859, "step": 6419 }, { "epoch": 0.23252444766388988, "grad_norm": 1.7224785524958373, "learning_rate": 1.7925224402507832e-05, "loss": 0.8645, "step": 6420 }, { "epoch": 0.23256066642520826, "grad_norm": 1.4254393539974886, "learning_rate": 1.792450896113592e-05, "loss": 0.808, "step": 6421 }, { "epoch": 0.23259688518652663, "grad_norm": 1.5014242771530146, "learning_rate": 1.7923793410715804e-05, "loss": 0.8143, "step": 6422 }, { "epoch": 0.23263310394784498, "grad_norm": 1.5835105337777875, "learning_rate": 1.7923077751257328e-05, "loss": 0.8712, "step": 6423 }, { "epoch": 0.23266932270916335, "grad_norm": 1.566293908606963, "learning_rate": 1.792236198277034e-05, "loss": 0.8594, "step": 6424 }, { "epoch": 0.2327055414704817, "grad_norm": 1.605861813513914, "learning_rate": 1.7921646105264688e-05, "loss": 0.7342, "step": 6425 }, { "epoch": 0.23274176023180007, "grad_norm": 1.5779931702647336, "learning_rate": 1.792093011875023e-05, "loss": 0.7931, "step": 6426 }, { "epoch": 0.23277797899311845, "grad_norm": 1.3800558290858915, "learning_rate": 1.7920214023236807e-05, "loss": 0.8243, "step": 6427 }, { "epoch": 0.2328141977544368, "grad_norm": 1.7465235274031747, "learning_rate": 1.7919497818734287e-05, "loss": 0.8019, "step": 6428 }, { "epoch": 0.23285041651575517, "grad_norm": 1.595915165123618, "learning_rate": 1.791878150525251e-05, "loss": 0.8638, "step": 6429 }, { "epoch": 0.2328866352770735, "grad_norm": 1.7237597203340465, "learning_rate": 1.7918065082801348e-05, "loss": 0.8461, "step": 6430 }, { "epoch": 0.23292285403839189, "grad_norm": 1.1165363389908103, "learning_rate": 1.7917348551390652e-05, "loss": 0.7396, "step": 6431 }, { "epoch": 0.23295907279971026, "grad_norm": 1.3865688001125458, "learning_rate": 1.7916631911030283e-05, "loss": 0.8362, "step": 6432 }, { "epoch": 0.2329952915610286, "grad_norm": 1.5496674520765428, "learning_rate": 1.79159151617301e-05, "loss": 0.7912, "step": 6433 }, { "epoch": 0.23303151032234698, "grad_norm": 1.5599594299607151, "learning_rate": 1.791519830349997e-05, "loss": 0.7508, "step": 6434 }, { "epoch": 0.23306772908366533, "grad_norm": 1.452571239758767, "learning_rate": 1.7914481336349757e-05, "loss": 0.8663, "step": 6435 }, { "epoch": 0.2331039478449837, "grad_norm": 1.20904086068454, "learning_rate": 1.7913764260289328e-05, "loss": 0.8893, "step": 6436 }, { "epoch": 0.23314016660630207, "grad_norm": 1.5497876949376286, "learning_rate": 1.791304707532855e-05, "loss": 0.865, "step": 6437 }, { "epoch": 0.23317638536762042, "grad_norm": 1.1049527346741665, "learning_rate": 1.7912329781477287e-05, "loss": 0.7604, "step": 6438 }, { "epoch": 0.2332126041289388, "grad_norm": 1.5486622254898936, "learning_rate": 1.7911612378745418e-05, "loss": 0.8693, "step": 6439 }, { "epoch": 0.23324882289025714, "grad_norm": 1.6222677885739665, "learning_rate": 1.791089486714281e-05, "loss": 0.8333, "step": 6440 }, { "epoch": 0.23328504165157551, "grad_norm": 1.744283590192288, "learning_rate": 1.7910177246679336e-05, "loss": 0.8156, "step": 6441 }, { "epoch": 0.2333212604128939, "grad_norm": 1.5108867137208215, "learning_rate": 1.790945951736487e-05, "loss": 0.8599, "step": 6442 }, { "epoch": 0.23335747917421223, "grad_norm": 1.5230175240264834, "learning_rate": 1.7908741679209295e-05, "loss": 0.8629, "step": 6443 }, { "epoch": 0.2333936979355306, "grad_norm": 1.4207798306950368, "learning_rate": 1.7908023732222485e-05, "loss": 0.8109, "step": 6444 }, { "epoch": 0.23342991669684895, "grad_norm": 1.4813033119498709, "learning_rate": 1.790730567641432e-05, "loss": 0.8011, "step": 6445 }, { "epoch": 0.23346613545816733, "grad_norm": 1.4309112955257504, "learning_rate": 1.790658751179468e-05, "loss": 0.7804, "step": 6446 }, { "epoch": 0.2335023542194857, "grad_norm": 1.3071853381840566, "learning_rate": 1.7905869238373448e-05, "loss": 0.8223, "step": 6447 }, { "epoch": 0.23353857298080405, "grad_norm": 1.4424622945768548, "learning_rate": 1.790515085616051e-05, "loss": 0.8257, "step": 6448 }, { "epoch": 0.23357479174212242, "grad_norm": 1.09122586811493, "learning_rate": 1.7904432365165748e-05, "loss": 0.7416, "step": 6449 }, { "epoch": 0.23361101050344077, "grad_norm": 1.4063430516928037, "learning_rate": 1.7903713765399054e-05, "loss": 0.8632, "step": 6450 }, { "epoch": 0.23364722926475914, "grad_norm": 1.440903696851072, "learning_rate": 1.790299505687031e-05, "loss": 0.8241, "step": 6451 }, { "epoch": 0.23368344802607752, "grad_norm": 1.5100566983157102, "learning_rate": 1.7902276239589413e-05, "loss": 0.8478, "step": 6452 }, { "epoch": 0.23371966678739586, "grad_norm": 1.1375242114847246, "learning_rate": 1.790155731356625e-05, "loss": 0.7364, "step": 6453 }, { "epoch": 0.23375588554871424, "grad_norm": 1.4046667776393273, "learning_rate": 1.790083827881072e-05, "loss": 0.8532, "step": 6454 }, { "epoch": 0.23379210431003258, "grad_norm": 1.4749429679104722, "learning_rate": 1.7900119135332707e-05, "loss": 0.8295, "step": 6455 }, { "epoch": 0.23382832307135096, "grad_norm": 1.3358711399082566, "learning_rate": 1.7899399883142115e-05, "loss": 0.8299, "step": 6456 }, { "epoch": 0.23386454183266933, "grad_norm": 1.2157331569008085, "learning_rate": 1.789868052224884e-05, "loss": 0.7546, "step": 6457 }, { "epoch": 0.23390076059398768, "grad_norm": 1.4427030655928843, "learning_rate": 1.789796105266278e-05, "loss": 0.8144, "step": 6458 }, { "epoch": 0.23393697935530605, "grad_norm": 1.4866434995838052, "learning_rate": 1.7897241474393837e-05, "loss": 0.8512, "step": 6459 }, { "epoch": 0.2339731981166244, "grad_norm": 1.3411349561552655, "learning_rate": 1.7896521787451912e-05, "loss": 0.6888, "step": 6460 }, { "epoch": 0.23400941687794277, "grad_norm": 1.4863153192667768, "learning_rate": 1.789580199184691e-05, "loss": 0.8629, "step": 6461 }, { "epoch": 0.23404563563926115, "grad_norm": 1.3627190475890003, "learning_rate": 1.789508208758873e-05, "loss": 0.7973, "step": 6462 }, { "epoch": 0.2340818544005795, "grad_norm": 1.360861679192142, "learning_rate": 1.7894362074687288e-05, "loss": 0.742, "step": 6463 }, { "epoch": 0.23411807316189787, "grad_norm": 1.2292975438484055, "learning_rate": 1.7893641953152484e-05, "loss": 0.8318, "step": 6464 }, { "epoch": 0.2341542919232162, "grad_norm": 1.5457576776787672, "learning_rate": 1.7892921722994235e-05, "loss": 0.8802, "step": 6465 }, { "epoch": 0.2341905106845346, "grad_norm": 1.1485828852033015, "learning_rate": 1.7892201384222443e-05, "loss": 0.8021, "step": 6466 }, { "epoch": 0.23422672944585296, "grad_norm": 1.5488597065711096, "learning_rate": 1.7891480936847027e-05, "loss": 0.8904, "step": 6467 }, { "epoch": 0.2342629482071713, "grad_norm": 1.3664727209007625, "learning_rate": 1.78907603808779e-05, "loss": 0.7819, "step": 6468 }, { "epoch": 0.23429916696848968, "grad_norm": 1.4195570759163882, "learning_rate": 1.789003971632498e-05, "loss": 0.8691, "step": 6469 }, { "epoch": 0.23433538572980803, "grad_norm": 1.3954680979732195, "learning_rate": 1.7889318943198174e-05, "loss": 0.802, "step": 6470 }, { "epoch": 0.2343716044911264, "grad_norm": 1.4893384705260666, "learning_rate": 1.788859806150741e-05, "loss": 0.8141, "step": 6471 }, { "epoch": 0.23440782325244477, "grad_norm": 1.1910706063414505, "learning_rate": 1.7887877071262605e-05, "loss": 0.7894, "step": 6472 }, { "epoch": 0.23444404201376312, "grad_norm": 1.3135949131140443, "learning_rate": 1.7887155972473682e-05, "loss": 0.7878, "step": 6473 }, { "epoch": 0.2344802607750815, "grad_norm": 1.4211312316326237, "learning_rate": 1.7886434765150562e-05, "loss": 0.8985, "step": 6474 }, { "epoch": 0.23451647953639987, "grad_norm": 1.4195648063163866, "learning_rate": 1.788571344930317e-05, "loss": 0.704, "step": 6475 }, { "epoch": 0.23455269829771822, "grad_norm": 1.411883181082397, "learning_rate": 1.788499202494143e-05, "loss": 0.7739, "step": 6476 }, { "epoch": 0.2345889170590366, "grad_norm": 1.463875142610908, "learning_rate": 1.7884270492075275e-05, "loss": 0.7591, "step": 6477 }, { "epoch": 0.23462513582035494, "grad_norm": 1.5563076834683587, "learning_rate": 1.788354885071463e-05, "loss": 0.8383, "step": 6478 }, { "epoch": 0.2346613545816733, "grad_norm": 1.2623231967175192, "learning_rate": 1.788282710086942e-05, "loss": 0.8382, "step": 6479 }, { "epoch": 0.23469757334299168, "grad_norm": 1.6914161099285268, "learning_rate": 1.788210524254959e-05, "loss": 0.7791, "step": 6480 }, { "epoch": 0.23473379210431003, "grad_norm": 1.353257626691899, "learning_rate": 1.7881383275765062e-05, "loss": 0.748, "step": 6481 }, { "epoch": 0.2347700108656284, "grad_norm": 1.0532440625556727, "learning_rate": 1.7880661200525777e-05, "loss": 0.7374, "step": 6482 }, { "epoch": 0.23480622962694675, "grad_norm": 1.43352770003561, "learning_rate": 1.7879939016841666e-05, "loss": 0.8317, "step": 6483 }, { "epoch": 0.23484244838826512, "grad_norm": 1.4838071991983375, "learning_rate": 1.787921672472267e-05, "loss": 0.7844, "step": 6484 }, { "epoch": 0.2348786671495835, "grad_norm": 1.4147198648102253, "learning_rate": 1.7878494324178732e-05, "loss": 0.7879, "step": 6485 }, { "epoch": 0.23491488591090184, "grad_norm": 2.3694300525863494, "learning_rate": 1.787777181521979e-05, "loss": 0.7048, "step": 6486 }, { "epoch": 0.23495110467222022, "grad_norm": 1.1768287957812014, "learning_rate": 1.7877049197855783e-05, "loss": 0.7713, "step": 6487 }, { "epoch": 0.23498732343353856, "grad_norm": 1.381156408555742, "learning_rate": 1.787632647209666e-05, "loss": 0.7668, "step": 6488 }, { "epoch": 0.23502354219485694, "grad_norm": 1.370097211853929, "learning_rate": 1.787560363795236e-05, "loss": 0.8405, "step": 6489 }, { "epoch": 0.2350597609561753, "grad_norm": 1.5192906892891547, "learning_rate": 1.7874880695432836e-05, "loss": 0.8292, "step": 6490 }, { "epoch": 0.23509597971749366, "grad_norm": 1.4978032007193365, "learning_rate": 1.7874157644548035e-05, "loss": 0.7858, "step": 6491 }, { "epoch": 0.23513219847881203, "grad_norm": 1.5108226681942996, "learning_rate": 1.7873434485307907e-05, "loss": 0.8305, "step": 6492 }, { "epoch": 0.23516841724013038, "grad_norm": 1.6576491071788253, "learning_rate": 1.7872711217722402e-05, "loss": 0.9287, "step": 6493 }, { "epoch": 0.23520463600144875, "grad_norm": 1.2853277240674137, "learning_rate": 1.7871987841801467e-05, "loss": 0.694, "step": 6494 }, { "epoch": 0.23524085476276713, "grad_norm": 1.5140219572758764, "learning_rate": 1.787126435755507e-05, "loss": 0.7994, "step": 6495 }, { "epoch": 0.23527707352408547, "grad_norm": 1.3860797271887557, "learning_rate": 1.7870540764993155e-05, "loss": 0.7586, "step": 6496 }, { "epoch": 0.23531329228540385, "grad_norm": 1.4283249542899534, "learning_rate": 1.7869817064125686e-05, "loss": 0.8947, "step": 6497 }, { "epoch": 0.2353495110467222, "grad_norm": 1.4569038522061193, "learning_rate": 1.7869093254962618e-05, "loss": 0.8921, "step": 6498 }, { "epoch": 0.23538572980804057, "grad_norm": 1.300076526472146, "learning_rate": 1.7868369337513913e-05, "loss": 0.8352, "step": 6499 }, { "epoch": 0.23542194856935894, "grad_norm": 1.3997697145338743, "learning_rate": 1.7867645311789533e-05, "loss": 0.7946, "step": 6500 }, { "epoch": 0.2354581673306773, "grad_norm": 1.236505090483693, "learning_rate": 1.7866921177799437e-05, "loss": 0.7598, "step": 6501 }, { "epoch": 0.23549438609199566, "grad_norm": 1.325547694868222, "learning_rate": 1.7866196935553593e-05, "loss": 0.7898, "step": 6502 }, { "epoch": 0.235530604853314, "grad_norm": 1.4844577634476033, "learning_rate": 1.786547258506197e-05, "loss": 0.8214, "step": 6503 }, { "epoch": 0.23556682361463238, "grad_norm": 1.531690669148196, "learning_rate": 1.7864748126334534e-05, "loss": 0.7932, "step": 6504 }, { "epoch": 0.23560304237595076, "grad_norm": 2.4269164987179868, "learning_rate": 1.7864023559381253e-05, "loss": 0.7691, "step": 6505 }, { "epoch": 0.2356392611372691, "grad_norm": 1.2916681448449885, "learning_rate": 1.7863298884212093e-05, "loss": 0.7883, "step": 6506 }, { "epoch": 0.23567547989858748, "grad_norm": 1.4455605835257646, "learning_rate": 1.786257410083704e-05, "loss": 0.8525, "step": 6507 }, { "epoch": 0.23571169865990582, "grad_norm": 2.730304238270986, "learning_rate": 1.7861849209266052e-05, "loss": 0.8604, "step": 6508 }, { "epoch": 0.2357479174212242, "grad_norm": 1.63756544500166, "learning_rate": 1.786112420950911e-05, "loss": 0.9211, "step": 6509 }, { "epoch": 0.23578413618254257, "grad_norm": 1.3816759589770113, "learning_rate": 1.7860399101576194e-05, "loss": 0.7866, "step": 6510 }, { "epoch": 0.23582035494386092, "grad_norm": 1.1192203404087286, "learning_rate": 1.7859673885477276e-05, "loss": 0.7864, "step": 6511 }, { "epoch": 0.2358565737051793, "grad_norm": 1.5130125295593921, "learning_rate": 1.7858948561222344e-05, "loss": 0.8019, "step": 6512 }, { "epoch": 0.23589279246649764, "grad_norm": 1.1547928922769093, "learning_rate": 1.7858223128821376e-05, "loss": 0.7321, "step": 6513 }, { "epoch": 0.235929011227816, "grad_norm": 1.3768220426513262, "learning_rate": 1.785749758828435e-05, "loss": 0.7974, "step": 6514 }, { "epoch": 0.23596522998913438, "grad_norm": 1.3835070786451005, "learning_rate": 1.785677193962125e-05, "loss": 0.7286, "step": 6515 }, { "epoch": 0.23600144875045273, "grad_norm": 1.154879274676162, "learning_rate": 1.785604618284207e-05, "loss": 0.7289, "step": 6516 }, { "epoch": 0.2360376675117711, "grad_norm": 1.4466005984030075, "learning_rate": 1.7855320317956785e-05, "loss": 0.8723, "step": 6517 }, { "epoch": 0.23607388627308945, "grad_norm": 1.4901718648996578, "learning_rate": 1.7854594344975398e-05, "loss": 0.8609, "step": 6518 }, { "epoch": 0.23611010503440782, "grad_norm": 1.0487690432009233, "learning_rate": 1.7853868263907885e-05, "loss": 0.7456, "step": 6519 }, { "epoch": 0.2361463237957262, "grad_norm": 1.4216184705626485, "learning_rate": 1.7853142074764246e-05, "loss": 0.7634, "step": 6520 }, { "epoch": 0.23618254255704454, "grad_norm": 1.3795150164082082, "learning_rate": 1.7852415777554473e-05, "loss": 0.7242, "step": 6521 }, { "epoch": 0.23621876131836292, "grad_norm": 1.5380145392422069, "learning_rate": 1.7851689372288555e-05, "loss": 0.8022, "step": 6522 }, { "epoch": 0.23625498007968126, "grad_norm": 1.1971668413460161, "learning_rate": 1.7850962858976496e-05, "loss": 0.7802, "step": 6523 }, { "epoch": 0.23629119884099964, "grad_norm": 1.216371960139803, "learning_rate": 1.785023623762829e-05, "loss": 0.7436, "step": 6524 }, { "epoch": 0.236327417602318, "grad_norm": 1.4993881949821668, "learning_rate": 1.784950950825393e-05, "loss": 0.7871, "step": 6525 }, { "epoch": 0.23636363636363636, "grad_norm": 1.5087689203936314, "learning_rate": 1.7848782670863428e-05, "loss": 0.8562, "step": 6526 }, { "epoch": 0.23639985512495473, "grad_norm": 1.5149681970011089, "learning_rate": 1.784805572546678e-05, "loss": 0.7817, "step": 6527 }, { "epoch": 0.23643607388627308, "grad_norm": 1.7366159166255566, "learning_rate": 1.7847328672073983e-05, "loss": 0.769, "step": 6528 }, { "epoch": 0.23647229264759145, "grad_norm": 1.5374432641024471, "learning_rate": 1.7846601510695055e-05, "loss": 0.8103, "step": 6529 }, { "epoch": 0.23650851140890983, "grad_norm": 1.5503708118509636, "learning_rate": 1.7845874241339992e-05, "loss": 0.8263, "step": 6530 }, { "epoch": 0.23654473017022817, "grad_norm": 1.5600248923041884, "learning_rate": 1.7845146864018806e-05, "loss": 0.819, "step": 6531 }, { "epoch": 0.23658094893154655, "grad_norm": 1.5765416587230858, "learning_rate": 1.7844419378741506e-05, "loss": 0.857, "step": 6532 }, { "epoch": 0.2366171676928649, "grad_norm": 1.4031143495255307, "learning_rate": 1.7843691785518103e-05, "loss": 0.7561, "step": 6533 }, { "epoch": 0.23665338645418327, "grad_norm": 1.553329273439296, "learning_rate": 1.784296408435861e-05, "loss": 0.823, "step": 6534 }, { "epoch": 0.23668960521550164, "grad_norm": 1.900647944782603, "learning_rate": 1.7842236275273038e-05, "loss": 0.861, "step": 6535 }, { "epoch": 0.23672582397682, "grad_norm": 1.5211457088600977, "learning_rate": 1.7841508358271403e-05, "loss": 0.9607, "step": 6536 }, { "epoch": 0.23676204273813836, "grad_norm": 1.5307745134466875, "learning_rate": 1.7840780333363724e-05, "loss": 0.8204, "step": 6537 }, { "epoch": 0.2367982614994567, "grad_norm": 1.590034079816251, "learning_rate": 1.784005220056002e-05, "loss": 0.8343, "step": 6538 }, { "epoch": 0.23683448026077508, "grad_norm": 1.514316424068616, "learning_rate": 1.7839323959870306e-05, "loss": 0.8175, "step": 6539 }, { "epoch": 0.23687069902209346, "grad_norm": 1.4068407963450777, "learning_rate": 1.783859561130461e-05, "loss": 0.7854, "step": 6540 }, { "epoch": 0.2369069177834118, "grad_norm": 1.2124515133442033, "learning_rate": 1.7837867154872948e-05, "loss": 0.7762, "step": 6541 }, { "epoch": 0.23694313654473018, "grad_norm": 1.4043439975254794, "learning_rate": 1.7837138590585347e-05, "loss": 0.8081, "step": 6542 }, { "epoch": 0.23697935530604852, "grad_norm": 1.0533020338238719, "learning_rate": 1.783640991845183e-05, "loss": 0.7574, "step": 6543 }, { "epoch": 0.2370155740673669, "grad_norm": 1.3274199507688538, "learning_rate": 1.7835681138482434e-05, "loss": 0.7828, "step": 6544 }, { "epoch": 0.23705179282868527, "grad_norm": 1.3507172988868539, "learning_rate": 1.7834952250687173e-05, "loss": 0.7902, "step": 6545 }, { "epoch": 0.23708801159000362, "grad_norm": 1.6669336855097652, "learning_rate": 1.7834223255076087e-05, "loss": 0.7356, "step": 6546 }, { "epoch": 0.237124230351322, "grad_norm": 1.6216017168522026, "learning_rate": 1.7833494151659208e-05, "loss": 0.8117, "step": 6547 }, { "epoch": 0.23716044911264034, "grad_norm": 1.4045908173289383, "learning_rate": 1.7832764940446564e-05, "loss": 0.7867, "step": 6548 }, { "epoch": 0.2371966678739587, "grad_norm": 1.6015537633890184, "learning_rate": 1.7832035621448192e-05, "loss": 0.8579, "step": 6549 }, { "epoch": 0.23723288663527708, "grad_norm": 1.6937855060047657, "learning_rate": 1.783130619467413e-05, "loss": 0.8254, "step": 6550 }, { "epoch": 0.23726910539659543, "grad_norm": 1.4686631204902312, "learning_rate": 1.7830576660134412e-05, "loss": 0.8141, "step": 6551 }, { "epoch": 0.2373053241579138, "grad_norm": 1.2027108490585996, "learning_rate": 1.7829847017839078e-05, "loss": 0.7569, "step": 6552 }, { "epoch": 0.23734154291923215, "grad_norm": 1.812069559741736, "learning_rate": 1.7829117267798172e-05, "loss": 0.7673, "step": 6553 }, { "epoch": 0.23737776168055053, "grad_norm": 1.1685807967664312, "learning_rate": 1.782838741002173e-05, "loss": 0.8277, "step": 6554 }, { "epoch": 0.2374139804418689, "grad_norm": 1.1877636602567914, "learning_rate": 1.78276574445198e-05, "loss": 0.8087, "step": 6555 }, { "epoch": 0.23745019920318725, "grad_norm": 1.9186428452914515, "learning_rate": 1.7826927371302426e-05, "loss": 0.7587, "step": 6556 }, { "epoch": 0.23748641796450562, "grad_norm": 1.45576102213959, "learning_rate": 1.782619719037965e-05, "loss": 0.822, "step": 6557 }, { "epoch": 0.23752263672582397, "grad_norm": 1.4797294539754682, "learning_rate": 1.782546690176153e-05, "loss": 0.7341, "step": 6558 }, { "epoch": 0.23755885548714234, "grad_norm": 1.4804328886434486, "learning_rate": 1.7824736505458106e-05, "loss": 0.8914, "step": 6559 }, { "epoch": 0.2375950742484607, "grad_norm": 1.9402883196912397, "learning_rate": 1.782400600147943e-05, "loss": 0.8781, "step": 6560 }, { "epoch": 0.23763129300977906, "grad_norm": 1.4046029810412755, "learning_rate": 1.782327538983556e-05, "loss": 0.8331, "step": 6561 }, { "epoch": 0.23766751177109743, "grad_norm": 1.4830179159623342, "learning_rate": 1.7822544670536544e-05, "loss": 0.8176, "step": 6562 }, { "epoch": 0.23770373053241578, "grad_norm": 1.4511276417430403, "learning_rate": 1.7821813843592447e-05, "loss": 0.8122, "step": 6563 }, { "epoch": 0.23773994929373415, "grad_norm": 1.596579749219886, "learning_rate": 1.782108290901331e-05, "loss": 0.8803, "step": 6564 }, { "epoch": 0.23777616805505253, "grad_norm": 1.935641679068365, "learning_rate": 1.7820351866809204e-05, "loss": 0.8779, "step": 6565 }, { "epoch": 0.23781238681637087, "grad_norm": 1.125867556095278, "learning_rate": 1.7819620716990184e-05, "loss": 0.8626, "step": 6566 }, { "epoch": 0.23784860557768925, "grad_norm": 1.53037128914642, "learning_rate": 1.7818889459566314e-05, "loss": 0.8326, "step": 6567 }, { "epoch": 0.2378848243390076, "grad_norm": 1.5361710390256385, "learning_rate": 1.781815809454765e-05, "loss": 0.8525, "step": 6568 }, { "epoch": 0.23792104310032597, "grad_norm": 1.448728628899898, "learning_rate": 1.7817426621944266e-05, "loss": 0.8432, "step": 6569 }, { "epoch": 0.23795726186164434, "grad_norm": 1.3550438063655956, "learning_rate": 1.781669504176622e-05, "loss": 0.8122, "step": 6570 }, { "epoch": 0.2379934806229627, "grad_norm": 1.1566771819820347, "learning_rate": 1.7815963354023586e-05, "loss": 0.7655, "step": 6571 }, { "epoch": 0.23802969938428106, "grad_norm": 1.4551954588130196, "learning_rate": 1.7815231558726422e-05, "loss": 0.7728, "step": 6572 }, { "epoch": 0.2380659181455994, "grad_norm": 1.270782583474003, "learning_rate": 1.781449965588481e-05, "loss": 0.7726, "step": 6573 }, { "epoch": 0.23810213690691778, "grad_norm": 1.7519624214393514, "learning_rate": 1.7813767645508812e-05, "loss": 0.8839, "step": 6574 }, { "epoch": 0.23813835566823616, "grad_norm": 1.48444224318232, "learning_rate": 1.7813035527608508e-05, "loss": 0.8293, "step": 6575 }, { "epoch": 0.2381745744295545, "grad_norm": 1.3654829878096968, "learning_rate": 1.7812303302193966e-05, "loss": 0.8071, "step": 6576 }, { "epoch": 0.23821079319087288, "grad_norm": 1.2471344202452117, "learning_rate": 1.7811570969275274e-05, "loss": 0.7546, "step": 6577 }, { "epoch": 0.23824701195219122, "grad_norm": 1.4509680147210648, "learning_rate": 1.7810838528862496e-05, "loss": 0.8641, "step": 6578 }, { "epoch": 0.2382832307135096, "grad_norm": 1.0848422155052788, "learning_rate": 1.7810105980965717e-05, "loss": 0.8216, "step": 6579 }, { "epoch": 0.23831944947482797, "grad_norm": 1.1776065871751478, "learning_rate": 1.780937332559502e-05, "loss": 0.7546, "step": 6580 }, { "epoch": 0.23835566823614632, "grad_norm": 1.5905064436529575, "learning_rate": 1.780864056276048e-05, "loss": 0.8224, "step": 6581 }, { "epoch": 0.2383918869974647, "grad_norm": 1.443078441534835, "learning_rate": 1.7807907692472188e-05, "loss": 0.7701, "step": 6582 }, { "epoch": 0.23842810575878304, "grad_norm": 1.3595679409300696, "learning_rate": 1.7807174714740225e-05, "loss": 0.7822, "step": 6583 }, { "epoch": 0.2384643245201014, "grad_norm": 1.3575434767505574, "learning_rate": 1.7806441629574675e-05, "loss": 0.8096, "step": 6584 }, { "epoch": 0.23850054328141979, "grad_norm": 1.3072864427652526, "learning_rate": 1.7805708436985632e-05, "loss": 0.7922, "step": 6585 }, { "epoch": 0.23853676204273813, "grad_norm": 1.6184202531270164, "learning_rate": 1.780497513698318e-05, "loss": 0.8093, "step": 6586 }, { "epoch": 0.2385729808040565, "grad_norm": 1.5186748996757287, "learning_rate": 1.7804241729577413e-05, "loss": 0.8108, "step": 6587 }, { "epoch": 0.23860919956537485, "grad_norm": 1.5415939423176905, "learning_rate": 1.7803508214778423e-05, "loss": 0.8405, "step": 6588 }, { "epoch": 0.23864541832669323, "grad_norm": 1.3484497636655086, "learning_rate": 1.7802774592596302e-05, "loss": 0.7409, "step": 6589 }, { "epoch": 0.2386816370880116, "grad_norm": 1.1812256158153396, "learning_rate": 1.7802040863041148e-05, "loss": 0.7246, "step": 6590 }, { "epoch": 0.23871785584932995, "grad_norm": 1.6131512356294417, "learning_rate": 1.7801307026123056e-05, "loss": 0.9287, "step": 6591 }, { "epoch": 0.23875407461064832, "grad_norm": 1.2557944951050724, "learning_rate": 1.7800573081852124e-05, "loss": 0.8182, "step": 6592 }, { "epoch": 0.23879029337196667, "grad_norm": 1.6919025294347414, "learning_rate": 1.779983903023845e-05, "loss": 0.7604, "step": 6593 }, { "epoch": 0.23882651213328504, "grad_norm": 1.4524231453976655, "learning_rate": 1.779910487129214e-05, "loss": 0.8463, "step": 6594 }, { "epoch": 0.23886273089460341, "grad_norm": 1.4611937832307884, "learning_rate": 1.7798370605023295e-05, "loss": 0.804, "step": 6595 }, { "epoch": 0.23889894965592176, "grad_norm": 1.4664512289249043, "learning_rate": 1.7797636231442018e-05, "loss": 0.8399, "step": 6596 }, { "epoch": 0.23893516841724013, "grad_norm": 1.4805657466979805, "learning_rate": 1.779690175055841e-05, "loss": 0.8368, "step": 6597 }, { "epoch": 0.23897138717855848, "grad_norm": 1.3284988311966845, "learning_rate": 1.779616716238259e-05, "loss": 0.8065, "step": 6598 }, { "epoch": 0.23900760593987685, "grad_norm": 1.5778856184431833, "learning_rate": 1.7795432466924656e-05, "loss": 0.8485, "step": 6599 }, { "epoch": 0.23904382470119523, "grad_norm": 1.4460783012179075, "learning_rate": 1.7794697664194722e-05, "loss": 0.7522, "step": 6600 }, { "epoch": 0.23908004346251358, "grad_norm": 1.5172380854407055, "learning_rate": 1.77939627542029e-05, "loss": 0.7956, "step": 6601 }, { "epoch": 0.23911626222383195, "grad_norm": 1.4460196382552526, "learning_rate": 1.77932277369593e-05, "loss": 0.7809, "step": 6602 }, { "epoch": 0.2391524809851503, "grad_norm": 1.552913058094393, "learning_rate": 1.7792492612474038e-05, "loss": 0.8585, "step": 6603 }, { "epoch": 0.23918869974646867, "grad_norm": 1.428403379303601, "learning_rate": 1.7791757380757234e-05, "loss": 0.8432, "step": 6604 }, { "epoch": 0.23922491850778704, "grad_norm": 1.5623238780181568, "learning_rate": 1.7791022041819e-05, "loss": 0.8346, "step": 6605 }, { "epoch": 0.2392611372691054, "grad_norm": 1.6162675484599294, "learning_rate": 1.7790286595669455e-05, "loss": 0.8529, "step": 6606 }, { "epoch": 0.23929735603042376, "grad_norm": 1.4018161406758674, "learning_rate": 1.7789551042318724e-05, "loss": 0.8301, "step": 6607 }, { "epoch": 0.2393335747917421, "grad_norm": 1.3865878560520695, "learning_rate": 1.778881538177693e-05, "loss": 0.7953, "step": 6608 }, { "epoch": 0.23936979355306048, "grad_norm": 1.2750767271157968, "learning_rate": 1.7788079614054186e-05, "loss": 0.7043, "step": 6609 }, { "epoch": 0.23940601231437886, "grad_norm": 1.4587953858743492, "learning_rate": 1.7787343739160624e-05, "loss": 0.8235, "step": 6610 }, { "epoch": 0.2394422310756972, "grad_norm": 1.5142628703481262, "learning_rate": 1.7786607757106372e-05, "loss": 0.8145, "step": 6611 }, { "epoch": 0.23947844983701558, "grad_norm": 1.5482239079806905, "learning_rate": 1.7785871667901555e-05, "loss": 0.7915, "step": 6612 }, { "epoch": 0.23951466859833392, "grad_norm": 1.59748940191201, "learning_rate": 1.7785135471556302e-05, "loss": 0.8427, "step": 6613 }, { "epoch": 0.2395508873596523, "grad_norm": 1.5140694846640728, "learning_rate": 1.778439916808074e-05, "loss": 0.8006, "step": 6614 }, { "epoch": 0.23958710612097067, "grad_norm": 1.4986343176963561, "learning_rate": 1.778366275748501e-05, "loss": 0.7574, "step": 6615 }, { "epoch": 0.23962332488228902, "grad_norm": 1.3733116175483744, "learning_rate": 1.7782926239779242e-05, "loss": 0.8223, "step": 6616 }, { "epoch": 0.2396595436436074, "grad_norm": 1.7944844089501308, "learning_rate": 1.778218961497357e-05, "loss": 0.8464, "step": 6617 }, { "epoch": 0.23969576240492574, "grad_norm": 1.4167466381380873, "learning_rate": 1.778145288307813e-05, "loss": 0.8011, "step": 6618 }, { "epoch": 0.2397319811662441, "grad_norm": 1.397659922041102, "learning_rate": 1.7780716044103062e-05, "loss": 0.7688, "step": 6619 }, { "epoch": 0.2397681999275625, "grad_norm": 1.3902941086442906, "learning_rate": 1.7779979098058506e-05, "loss": 0.7893, "step": 6620 }, { "epoch": 0.23980441868888083, "grad_norm": 1.2716851636235242, "learning_rate": 1.7779242044954596e-05, "loss": 0.8093, "step": 6621 }, { "epoch": 0.2398406374501992, "grad_norm": 1.4702067601422895, "learning_rate": 1.7778504884801483e-05, "loss": 0.8106, "step": 6622 }, { "epoch": 0.23987685621151755, "grad_norm": 1.5743341072355264, "learning_rate": 1.7777767617609307e-05, "loss": 0.744, "step": 6623 }, { "epoch": 0.23991307497283593, "grad_norm": 1.5527369085327207, "learning_rate": 1.7777030243388217e-05, "loss": 0.9125, "step": 6624 }, { "epoch": 0.2399492937341543, "grad_norm": 1.2520463123246242, "learning_rate": 1.7776292762148356e-05, "loss": 0.8415, "step": 6625 }, { "epoch": 0.23998551249547265, "grad_norm": 1.3795092712212988, "learning_rate": 1.7775555173899876e-05, "loss": 0.778, "step": 6626 }, { "epoch": 0.24002173125679102, "grad_norm": 1.558587821193882, "learning_rate": 1.777481747865292e-05, "loss": 0.8102, "step": 6627 }, { "epoch": 0.24005795001810937, "grad_norm": 1.7372868640450694, "learning_rate": 1.7774079676417644e-05, "loss": 0.8075, "step": 6628 }, { "epoch": 0.24009416877942774, "grad_norm": 1.4796056609497636, "learning_rate": 1.7773341767204205e-05, "loss": 0.8309, "step": 6629 }, { "epoch": 0.24013038754074612, "grad_norm": 1.5719006570898664, "learning_rate": 1.777260375102275e-05, "loss": 0.8079, "step": 6630 }, { "epoch": 0.24016660630206446, "grad_norm": 1.4230707462195031, "learning_rate": 1.777186562788344e-05, "loss": 0.8182, "step": 6631 }, { "epoch": 0.24020282506338284, "grad_norm": 1.2072695547315757, "learning_rate": 1.7771127397796427e-05, "loss": 0.7337, "step": 6632 }, { "epoch": 0.24023904382470118, "grad_norm": 3.108198373530122, "learning_rate": 1.7770389060771875e-05, "loss": 0.8011, "step": 6633 }, { "epoch": 0.24027526258601956, "grad_norm": 1.5135730009703545, "learning_rate": 1.776965061681994e-05, "loss": 0.8654, "step": 6634 }, { "epoch": 0.24031148134733793, "grad_norm": 1.4265417017574817, "learning_rate": 1.7768912065950786e-05, "loss": 0.8547, "step": 6635 }, { "epoch": 0.24034770010865628, "grad_norm": 1.1928640749713526, "learning_rate": 1.776817340817458e-05, "loss": 0.7723, "step": 6636 }, { "epoch": 0.24038391886997465, "grad_norm": 1.178595012320841, "learning_rate": 1.7767434643501476e-05, "loss": 0.7456, "step": 6637 }, { "epoch": 0.240420137631293, "grad_norm": 1.1775111414176214, "learning_rate": 1.776669577194165e-05, "loss": 0.7414, "step": 6638 }, { "epoch": 0.24045635639261137, "grad_norm": 1.5518386482468403, "learning_rate": 1.7765956793505265e-05, "loss": 0.8185, "step": 6639 }, { "epoch": 0.24049257515392974, "grad_norm": 1.5058472913660486, "learning_rate": 1.776521770820249e-05, "loss": 0.8922, "step": 6640 }, { "epoch": 0.2405287939152481, "grad_norm": 1.2913866812411994, "learning_rate": 1.7764478516043495e-05, "loss": 0.8168, "step": 6641 }, { "epoch": 0.24056501267656646, "grad_norm": 1.437368486377448, "learning_rate": 1.7763739217038455e-05, "loss": 0.8225, "step": 6642 }, { "epoch": 0.2406012314378848, "grad_norm": 1.450584944849187, "learning_rate": 1.7762999811197538e-05, "loss": 0.7832, "step": 6643 }, { "epoch": 0.24063745019920318, "grad_norm": 1.4987688277034739, "learning_rate": 1.776226029853093e-05, "loss": 0.7908, "step": 6644 }, { "epoch": 0.24067366896052156, "grad_norm": 1.2001843115369917, "learning_rate": 1.7761520679048794e-05, "loss": 0.7431, "step": 6645 }, { "epoch": 0.2407098877218399, "grad_norm": 1.4128474908941084, "learning_rate": 1.7760780952761314e-05, "loss": 0.7826, "step": 6646 }, { "epoch": 0.24074610648315828, "grad_norm": 1.6449646457971743, "learning_rate": 1.776004111967867e-05, "loss": 0.7115, "step": 6647 }, { "epoch": 0.24078232524447665, "grad_norm": 1.076079477451733, "learning_rate": 1.775930117981104e-05, "loss": 0.7557, "step": 6648 }, { "epoch": 0.240818544005795, "grad_norm": 1.4982618478562455, "learning_rate": 1.775856113316861e-05, "loss": 0.8713, "step": 6649 }, { "epoch": 0.24085476276711337, "grad_norm": 1.2884028083320713, "learning_rate": 1.7757820979761563e-05, "loss": 0.8686, "step": 6650 }, { "epoch": 0.24089098152843172, "grad_norm": 1.42777383778485, "learning_rate": 1.7757080719600078e-05, "loss": 0.807, "step": 6651 }, { "epoch": 0.2409272002897501, "grad_norm": 1.417572772745733, "learning_rate": 1.775634035269435e-05, "loss": 0.8678, "step": 6652 }, { "epoch": 0.24096341905106847, "grad_norm": 1.5177629786597262, "learning_rate": 1.775559987905456e-05, "loss": 0.8847, "step": 6653 }, { "epoch": 0.2409996378123868, "grad_norm": 1.3667292930616788, "learning_rate": 1.7754859298690905e-05, "loss": 0.7645, "step": 6654 }, { "epoch": 0.2410358565737052, "grad_norm": 1.6508780724546759, "learning_rate": 1.7754118611613573e-05, "loss": 0.8431, "step": 6655 }, { "epoch": 0.24107207533502353, "grad_norm": 1.602660690006541, "learning_rate": 1.7753377817832752e-05, "loss": 0.818, "step": 6656 }, { "epoch": 0.2411082940963419, "grad_norm": 1.256569481401941, "learning_rate": 1.7752636917358642e-05, "loss": 0.7464, "step": 6657 }, { "epoch": 0.24114451285766028, "grad_norm": 1.4074894515920324, "learning_rate": 1.7751895910201434e-05, "loss": 0.7539, "step": 6658 }, { "epoch": 0.24118073161897863, "grad_norm": 1.388128677392921, "learning_rate": 1.7751154796371325e-05, "loss": 0.7563, "step": 6659 }, { "epoch": 0.241216950380297, "grad_norm": 1.1641793094422945, "learning_rate": 1.7750413575878518e-05, "loss": 0.8436, "step": 6660 }, { "epoch": 0.24125316914161535, "grad_norm": 1.4122621533905448, "learning_rate": 1.7749672248733208e-05, "loss": 0.897, "step": 6661 }, { "epoch": 0.24128938790293372, "grad_norm": 1.5200719202450914, "learning_rate": 1.7748930814945602e-05, "loss": 0.8137, "step": 6662 }, { "epoch": 0.2413256066642521, "grad_norm": 1.5674857477155424, "learning_rate": 1.7748189274525896e-05, "loss": 0.8093, "step": 6663 }, { "epoch": 0.24136182542557044, "grad_norm": 1.2932640747547761, "learning_rate": 1.77474476274843e-05, "loss": 0.7945, "step": 6664 }, { "epoch": 0.24139804418688882, "grad_norm": 1.429668323461196, "learning_rate": 1.7746705873831014e-05, "loss": 0.7736, "step": 6665 }, { "epoch": 0.24143426294820716, "grad_norm": 1.5488511573262331, "learning_rate": 1.7745964013576248e-05, "loss": 0.8894, "step": 6666 }, { "epoch": 0.24147048170952554, "grad_norm": 1.582414793687512, "learning_rate": 1.7745222046730216e-05, "loss": 0.8892, "step": 6667 }, { "epoch": 0.2415067004708439, "grad_norm": 1.6222969324042729, "learning_rate": 1.774447997330312e-05, "loss": 0.8127, "step": 6668 }, { "epoch": 0.24154291923216226, "grad_norm": 1.5795740919787025, "learning_rate": 1.7743737793305175e-05, "loss": 0.8015, "step": 6669 }, { "epoch": 0.24157913799348063, "grad_norm": 1.4943905702355746, "learning_rate": 1.774299550674659e-05, "loss": 0.8489, "step": 6670 }, { "epoch": 0.24161535675479898, "grad_norm": 1.3990260739017384, "learning_rate": 1.7742253113637587e-05, "loss": 0.8621, "step": 6671 }, { "epoch": 0.24165157551611735, "grad_norm": 1.4530518623790547, "learning_rate": 1.774151061398838e-05, "loss": 0.7716, "step": 6672 }, { "epoch": 0.24168779427743572, "grad_norm": 1.3276567881113537, "learning_rate": 1.774076800780918e-05, "loss": 0.8099, "step": 6673 }, { "epoch": 0.24172401303875407, "grad_norm": 1.4511519876095444, "learning_rate": 1.7740025295110214e-05, "loss": 0.784, "step": 6674 }, { "epoch": 0.24176023180007244, "grad_norm": 1.1388388760928316, "learning_rate": 1.77392824759017e-05, "loss": 0.7393, "step": 6675 }, { "epoch": 0.2417964505613908, "grad_norm": 1.381904567995685, "learning_rate": 1.7738539550193856e-05, "loss": 0.7838, "step": 6676 }, { "epoch": 0.24183266932270917, "grad_norm": 1.3219470273481382, "learning_rate": 1.773779651799691e-05, "loss": 0.7897, "step": 6677 }, { "epoch": 0.24186888808402754, "grad_norm": 1.4758362333738893, "learning_rate": 1.7737053379321084e-05, "loss": 0.8035, "step": 6678 }, { "epoch": 0.24190510684534589, "grad_norm": 1.402809732089283, "learning_rate": 1.7736310134176604e-05, "loss": 0.7959, "step": 6679 }, { "epoch": 0.24194132560666426, "grad_norm": 1.359820577200914, "learning_rate": 1.77355667825737e-05, "loss": 0.7753, "step": 6680 }, { "epoch": 0.2419775443679826, "grad_norm": 1.4847071727717138, "learning_rate": 1.77348233245226e-05, "loss": 0.7427, "step": 6681 }, { "epoch": 0.24201376312930098, "grad_norm": 1.6178438431533337, "learning_rate": 1.7734079760033532e-05, "loss": 0.8343, "step": 6682 }, { "epoch": 0.24204998189061935, "grad_norm": 1.4576851806566564, "learning_rate": 1.7733336089116736e-05, "loss": 0.7681, "step": 6683 }, { "epoch": 0.2420862006519377, "grad_norm": 1.091969472516771, "learning_rate": 1.7732592311782438e-05, "loss": 0.7557, "step": 6684 }, { "epoch": 0.24212241941325607, "grad_norm": 1.4776992395869104, "learning_rate": 1.7731848428040873e-05, "loss": 0.87, "step": 6685 }, { "epoch": 0.24215863817457442, "grad_norm": 1.5352259218504345, "learning_rate": 1.7731104437902283e-05, "loss": 0.8184, "step": 6686 }, { "epoch": 0.2421948569358928, "grad_norm": 1.096344983200657, "learning_rate": 1.77303603413769e-05, "loss": 0.8244, "step": 6687 }, { "epoch": 0.24223107569721117, "grad_norm": 1.1003841680714597, "learning_rate": 1.772961613847497e-05, "loss": 0.8083, "step": 6688 }, { "epoch": 0.24226729445852951, "grad_norm": 1.4041891998479044, "learning_rate": 1.7728871829206725e-05, "loss": 0.7608, "step": 6689 }, { "epoch": 0.2423035132198479, "grad_norm": 1.3604505712403718, "learning_rate": 1.7728127413582418e-05, "loss": 0.8523, "step": 6690 }, { "epoch": 0.24233973198116623, "grad_norm": 1.0768413691251897, "learning_rate": 1.7727382891612284e-05, "loss": 0.7323, "step": 6691 }, { "epoch": 0.2423759507424846, "grad_norm": 1.597753812100816, "learning_rate": 1.772663826330657e-05, "loss": 0.8755, "step": 6692 }, { "epoch": 0.24241216950380298, "grad_norm": 1.5471493713690456, "learning_rate": 1.7725893528675526e-05, "loss": 0.8801, "step": 6693 }, { "epoch": 0.24244838826512133, "grad_norm": 1.600363403453792, "learning_rate": 1.77251486877294e-05, "loss": 0.8019, "step": 6694 }, { "epoch": 0.2424846070264397, "grad_norm": 1.3382111402130477, "learning_rate": 1.772440374047844e-05, "loss": 0.7864, "step": 6695 }, { "epoch": 0.24252082578775805, "grad_norm": 1.3175630987795186, "learning_rate": 1.7723658686932898e-05, "loss": 0.8434, "step": 6696 }, { "epoch": 0.24255704454907642, "grad_norm": 1.5230631661846277, "learning_rate": 1.772291352710302e-05, "loss": 0.8088, "step": 6697 }, { "epoch": 0.2425932633103948, "grad_norm": 1.4086024290971597, "learning_rate": 1.772216826099907e-05, "loss": 0.8328, "step": 6698 }, { "epoch": 0.24262948207171314, "grad_norm": 1.528336271988778, "learning_rate": 1.77214228886313e-05, "loss": 0.7526, "step": 6699 }, { "epoch": 0.24266570083303152, "grad_norm": 1.5154625372971204, "learning_rate": 1.772067741000997e-05, "loss": 0.8873, "step": 6700 }, { "epoch": 0.24270191959434986, "grad_norm": 1.135016157855781, "learning_rate": 1.7719931825145326e-05, "loss": 0.7755, "step": 6701 }, { "epoch": 0.24273813835566824, "grad_norm": 1.5801521420677682, "learning_rate": 1.7719186134047643e-05, "loss": 0.9151, "step": 6702 }, { "epoch": 0.2427743571169866, "grad_norm": 1.2661893499387085, "learning_rate": 1.7718440336727173e-05, "loss": 0.7932, "step": 6703 }, { "epoch": 0.24281057587830496, "grad_norm": 1.41103760289211, "learning_rate": 1.7717694433194186e-05, "loss": 0.7222, "step": 6704 }, { "epoch": 0.24284679463962333, "grad_norm": 1.4561160987373503, "learning_rate": 1.771694842345894e-05, "loss": 0.7754, "step": 6705 }, { "epoch": 0.24288301340094168, "grad_norm": 1.5316481522713479, "learning_rate": 1.7716202307531704e-05, "loss": 0.7914, "step": 6706 }, { "epoch": 0.24291923216226005, "grad_norm": 1.3845822773816119, "learning_rate": 1.771545608542274e-05, "loss": 0.8198, "step": 6707 }, { "epoch": 0.24295545092357843, "grad_norm": 1.3982393641202908, "learning_rate": 1.7714709757142324e-05, "loss": 0.7611, "step": 6708 }, { "epoch": 0.24299166968489677, "grad_norm": 1.2793611066287023, "learning_rate": 1.7713963322700724e-05, "loss": 0.7738, "step": 6709 }, { "epoch": 0.24302788844621515, "grad_norm": 1.6121126981367757, "learning_rate": 1.771321678210821e-05, "loss": 0.8214, "step": 6710 }, { "epoch": 0.2430641072075335, "grad_norm": 1.3641529548122322, "learning_rate": 1.7712470135375053e-05, "loss": 0.7954, "step": 6711 }, { "epoch": 0.24310032596885187, "grad_norm": 1.4441719181280728, "learning_rate": 1.7711723382511536e-05, "loss": 0.8212, "step": 6712 }, { "epoch": 0.24313654473017024, "grad_norm": 1.6076956373308426, "learning_rate": 1.7710976523527925e-05, "loss": 0.8869, "step": 6713 }, { "epoch": 0.24317276349148859, "grad_norm": 1.512891434699565, "learning_rate": 1.7710229558434502e-05, "loss": 0.8124, "step": 6714 }, { "epoch": 0.24320898225280696, "grad_norm": 1.5903723749390806, "learning_rate": 1.7709482487241545e-05, "loss": 0.8977, "step": 6715 }, { "epoch": 0.2432452010141253, "grad_norm": 1.1708873504864616, "learning_rate": 1.7708735309959335e-05, "loss": 0.7442, "step": 6716 }, { "epoch": 0.24328141977544368, "grad_norm": 1.2305682370104347, "learning_rate": 1.7707988026598154e-05, "loss": 0.7213, "step": 6717 }, { "epoch": 0.24331763853676205, "grad_norm": 1.6868662195906465, "learning_rate": 1.7707240637168285e-05, "loss": 0.8463, "step": 6718 }, { "epoch": 0.2433538572980804, "grad_norm": 1.4656822462329755, "learning_rate": 1.770649314168001e-05, "loss": 0.7939, "step": 6719 }, { "epoch": 0.24339007605939877, "grad_norm": 1.299697358524422, "learning_rate": 1.7705745540143622e-05, "loss": 0.8072, "step": 6720 }, { "epoch": 0.24342629482071712, "grad_norm": 1.1753646631010572, "learning_rate": 1.7704997832569405e-05, "loss": 0.8371, "step": 6721 }, { "epoch": 0.2434625135820355, "grad_norm": 1.0901817753379262, "learning_rate": 1.7704250018967645e-05, "loss": 0.7932, "step": 6722 }, { "epoch": 0.24349873234335387, "grad_norm": 1.4835876961896421, "learning_rate": 1.7703502099348636e-05, "loss": 0.796, "step": 6723 }, { "epoch": 0.24353495110467221, "grad_norm": 1.6562585110715593, "learning_rate": 1.7702754073722668e-05, "loss": 0.8026, "step": 6724 }, { "epoch": 0.2435711698659906, "grad_norm": 1.5923411743454665, "learning_rate": 1.7702005942100037e-05, "loss": 0.7955, "step": 6725 }, { "epoch": 0.24360738862730894, "grad_norm": 1.3682899914912656, "learning_rate": 1.7701257704491036e-05, "loss": 0.8134, "step": 6726 }, { "epoch": 0.2436436073886273, "grad_norm": 1.4598621717746456, "learning_rate": 1.7700509360905962e-05, "loss": 0.8306, "step": 6727 }, { "epoch": 0.24367982614994568, "grad_norm": 1.3814152761381715, "learning_rate": 1.7699760911355115e-05, "loss": 0.8206, "step": 6728 }, { "epoch": 0.24371604491126403, "grad_norm": 1.4217909379360694, "learning_rate": 1.7699012355848787e-05, "loss": 0.7738, "step": 6729 }, { "epoch": 0.2437522636725824, "grad_norm": 1.468475569795555, "learning_rate": 1.769826369439729e-05, "loss": 0.8458, "step": 6730 }, { "epoch": 0.24378848243390075, "grad_norm": 1.2268670400298372, "learning_rate": 1.7697514927010916e-05, "loss": 0.7929, "step": 6731 }, { "epoch": 0.24382470119521912, "grad_norm": 1.8028755008760555, "learning_rate": 1.7696766053699978e-05, "loss": 0.7769, "step": 6732 }, { "epoch": 0.2438609199565375, "grad_norm": 1.4279844910367323, "learning_rate": 1.7696017074474773e-05, "loss": 0.7826, "step": 6733 }, { "epoch": 0.24389713871785584, "grad_norm": 1.4381095731451579, "learning_rate": 1.769526798934561e-05, "loss": 0.7752, "step": 6734 }, { "epoch": 0.24393335747917422, "grad_norm": 1.3513617111461178, "learning_rate": 1.7694518798322797e-05, "loss": 0.8338, "step": 6735 }, { "epoch": 0.24396957624049256, "grad_norm": 1.3965339238373116, "learning_rate": 1.7693769501416645e-05, "loss": 0.862, "step": 6736 }, { "epoch": 0.24400579500181094, "grad_norm": 1.0820280158514448, "learning_rate": 1.7693020098637464e-05, "loss": 0.7558, "step": 6737 }, { "epoch": 0.2440420137631293, "grad_norm": 1.5290941741069066, "learning_rate": 1.7692270589995573e-05, "loss": 0.7504, "step": 6738 }, { "epoch": 0.24407823252444766, "grad_norm": 1.3828929336237215, "learning_rate": 1.7691520975501272e-05, "loss": 0.8256, "step": 6739 }, { "epoch": 0.24411445128576603, "grad_norm": 1.2310925353891278, "learning_rate": 1.7690771255164888e-05, "loss": 0.7719, "step": 6740 }, { "epoch": 0.24415067004708438, "grad_norm": 1.2215367049806587, "learning_rate": 1.7690021428996732e-05, "loss": 0.8484, "step": 6741 }, { "epoch": 0.24418688880840275, "grad_norm": 1.5040083293206996, "learning_rate": 1.7689271497007127e-05, "loss": 0.9037, "step": 6742 }, { "epoch": 0.24422310756972113, "grad_norm": 1.2696183710282474, "learning_rate": 1.7688521459206386e-05, "loss": 0.8331, "step": 6743 }, { "epoch": 0.24425932633103947, "grad_norm": 1.4298249019718248, "learning_rate": 1.7687771315604833e-05, "loss": 0.7579, "step": 6744 }, { "epoch": 0.24429554509235785, "grad_norm": 1.4431019522491801, "learning_rate": 1.7687021066212795e-05, "loss": 0.8604, "step": 6745 }, { "epoch": 0.2443317638536762, "grad_norm": 1.4005463984372606, "learning_rate": 1.7686270711040593e-05, "loss": 0.8176, "step": 6746 }, { "epoch": 0.24436798261499457, "grad_norm": 1.4233875168526129, "learning_rate": 1.768552025009855e-05, "loss": 0.8196, "step": 6747 }, { "epoch": 0.24440420137631294, "grad_norm": 1.1932831962716803, "learning_rate": 1.7684769683396996e-05, "loss": 0.7497, "step": 6748 }, { "epoch": 0.2444404201376313, "grad_norm": 1.4693866970447726, "learning_rate": 1.7684019010946262e-05, "loss": 0.8105, "step": 6749 }, { "epoch": 0.24447663889894966, "grad_norm": 1.4904605123946302, "learning_rate": 1.768326823275667e-05, "loss": 0.8401, "step": 6750 }, { "epoch": 0.244512857660268, "grad_norm": 1.361611688380728, "learning_rate": 1.7682517348838555e-05, "loss": 0.7279, "step": 6751 }, { "epoch": 0.24454907642158638, "grad_norm": 1.4686015471298024, "learning_rate": 1.7681766359202256e-05, "loss": 0.8209, "step": 6752 }, { "epoch": 0.24458529518290476, "grad_norm": 1.473160275484843, "learning_rate": 1.76810152638581e-05, "loss": 0.8362, "step": 6753 }, { "epoch": 0.2446215139442231, "grad_norm": 1.490376028275481, "learning_rate": 1.768026406281642e-05, "loss": 0.8553, "step": 6754 }, { "epoch": 0.24465773270554148, "grad_norm": 1.4474434376570984, "learning_rate": 1.7679512756087564e-05, "loss": 0.7644, "step": 6755 }, { "epoch": 0.24469395146685982, "grad_norm": 1.4245940811668876, "learning_rate": 1.7678761343681864e-05, "loss": 0.8289, "step": 6756 }, { "epoch": 0.2447301702281782, "grad_norm": 1.4578329787599613, "learning_rate": 1.7678009825609658e-05, "loss": 0.7343, "step": 6757 }, { "epoch": 0.24476638898949657, "grad_norm": 1.5264002759181823, "learning_rate": 1.7677258201881297e-05, "loss": 0.8013, "step": 6758 }, { "epoch": 0.24480260775081492, "grad_norm": 1.1979774814805024, "learning_rate": 1.767650647250711e-05, "loss": 0.8171, "step": 6759 }, { "epoch": 0.2448388265121333, "grad_norm": 1.435169908762783, "learning_rate": 1.7675754637497452e-05, "loss": 0.7812, "step": 6760 }, { "epoch": 0.24487504527345164, "grad_norm": 1.4719856687061852, "learning_rate": 1.7675002696862664e-05, "loss": 0.8311, "step": 6761 }, { "epoch": 0.24491126403477, "grad_norm": 1.173990763799937, "learning_rate": 1.7674250650613096e-05, "loss": 0.8058, "step": 6762 }, { "epoch": 0.24494748279608838, "grad_norm": 1.4233483099971913, "learning_rate": 1.7673498498759095e-05, "loss": 0.8342, "step": 6763 }, { "epoch": 0.24498370155740673, "grad_norm": 1.447707117878178, "learning_rate": 1.7672746241311013e-05, "loss": 0.8506, "step": 6764 }, { "epoch": 0.2450199203187251, "grad_norm": 1.3224262663995443, "learning_rate": 1.7671993878279204e-05, "loss": 0.7456, "step": 6765 }, { "epoch": 0.24505613908004345, "grad_norm": 1.5191782845409756, "learning_rate": 1.7671241409674014e-05, "loss": 0.8446, "step": 6766 }, { "epoch": 0.24509235784136182, "grad_norm": 1.6106252142825, "learning_rate": 1.7670488835505805e-05, "loss": 0.8962, "step": 6767 }, { "epoch": 0.2451285766026802, "grad_norm": 1.4288799501421257, "learning_rate": 1.7669736155784924e-05, "loss": 0.8551, "step": 6768 }, { "epoch": 0.24516479536399854, "grad_norm": 1.5703320169896622, "learning_rate": 1.766898337052174e-05, "loss": 0.8118, "step": 6769 }, { "epoch": 0.24520101412531692, "grad_norm": 1.4101888971168808, "learning_rate": 1.7668230479726603e-05, "loss": 0.7875, "step": 6770 }, { "epoch": 0.24523723288663526, "grad_norm": 1.337862557095889, "learning_rate": 1.7667477483409875e-05, "loss": 0.8226, "step": 6771 }, { "epoch": 0.24527345164795364, "grad_norm": 1.4475074702260924, "learning_rate": 1.7666724381581923e-05, "loss": 0.8433, "step": 6772 }, { "epoch": 0.245309670409272, "grad_norm": 1.4123501704249748, "learning_rate": 1.7665971174253106e-05, "loss": 0.8278, "step": 6773 }, { "epoch": 0.24534588917059036, "grad_norm": 1.4451846427037969, "learning_rate": 1.766521786143379e-05, "loss": 0.7927, "step": 6774 }, { "epoch": 0.24538210793190873, "grad_norm": 1.4625028432406804, "learning_rate": 1.7664464443134338e-05, "loss": 0.8154, "step": 6775 }, { "epoch": 0.24541832669322708, "grad_norm": 1.8509726922909586, "learning_rate": 1.7663710919365128e-05, "loss": 0.8452, "step": 6776 }, { "epoch": 0.24545454545454545, "grad_norm": 1.4350997095360585, "learning_rate": 1.7662957290136515e-05, "loss": 0.7091, "step": 6777 }, { "epoch": 0.24549076421586383, "grad_norm": 1.4581579806852891, "learning_rate": 1.766220355545888e-05, "loss": 0.7667, "step": 6778 }, { "epoch": 0.24552698297718217, "grad_norm": 1.8362960627616625, "learning_rate": 1.7661449715342588e-05, "loss": 0.8236, "step": 6779 }, { "epoch": 0.24556320173850055, "grad_norm": 1.6211313524107738, "learning_rate": 1.7660695769798018e-05, "loss": 0.7919, "step": 6780 }, { "epoch": 0.2455994204998189, "grad_norm": 1.5259821212954887, "learning_rate": 1.7659941718835543e-05, "loss": 0.8081, "step": 6781 }, { "epoch": 0.24563563926113727, "grad_norm": 1.525619428621345, "learning_rate": 1.765918756246554e-05, "loss": 0.7943, "step": 6782 }, { "epoch": 0.24567185802245564, "grad_norm": 1.1833294669409111, "learning_rate": 1.7658433300698384e-05, "loss": 0.8055, "step": 6783 }, { "epoch": 0.245708076783774, "grad_norm": 1.1841783289734054, "learning_rate": 1.765767893354446e-05, "loss": 0.769, "step": 6784 }, { "epoch": 0.24574429554509236, "grad_norm": 1.4572595316490016, "learning_rate": 1.7656924461014144e-05, "loss": 0.867, "step": 6785 }, { "epoch": 0.2457805143064107, "grad_norm": 1.1532194816625774, "learning_rate": 1.765616988311782e-05, "loss": 0.8541, "step": 6786 }, { "epoch": 0.24581673306772908, "grad_norm": 1.3738989783816624, "learning_rate": 1.765541519986587e-05, "loss": 0.7216, "step": 6787 }, { "epoch": 0.24585295182904746, "grad_norm": 1.4788277056049208, "learning_rate": 1.7654660411268678e-05, "loss": 0.8258, "step": 6788 }, { "epoch": 0.2458891705903658, "grad_norm": 1.633393884579714, "learning_rate": 1.7653905517336632e-05, "loss": 0.8415, "step": 6789 }, { "epoch": 0.24592538935168418, "grad_norm": 1.4869773220183315, "learning_rate": 1.7653150518080125e-05, "loss": 0.8221, "step": 6790 }, { "epoch": 0.24596160811300252, "grad_norm": 1.5363543242202422, "learning_rate": 1.7652395413509542e-05, "loss": 0.8986, "step": 6791 }, { "epoch": 0.2459978268743209, "grad_norm": 1.2798217376969931, "learning_rate": 1.7651640203635274e-05, "loss": 0.7485, "step": 6792 }, { "epoch": 0.24603404563563927, "grad_norm": 1.4253754730757406, "learning_rate": 1.7650884888467712e-05, "loss": 0.8553, "step": 6793 }, { "epoch": 0.24607026439695762, "grad_norm": 1.357406646313765, "learning_rate": 1.7650129468017252e-05, "loss": 0.8355, "step": 6794 }, { "epoch": 0.246106483158276, "grad_norm": 1.554847102862655, "learning_rate": 1.7649373942294288e-05, "loss": 0.7667, "step": 6795 }, { "epoch": 0.24614270191959434, "grad_norm": 1.2115019589002993, "learning_rate": 1.7648618311309218e-05, "loss": 0.7648, "step": 6796 }, { "epoch": 0.2461789206809127, "grad_norm": 1.3703001250955538, "learning_rate": 1.764786257507244e-05, "loss": 0.8028, "step": 6797 }, { "epoch": 0.24621513944223108, "grad_norm": 1.465491746627873, "learning_rate": 1.764710673359435e-05, "loss": 0.8364, "step": 6798 }, { "epoch": 0.24625135820354943, "grad_norm": 1.4555621097505123, "learning_rate": 1.7646350786885352e-05, "loss": 0.8585, "step": 6799 }, { "epoch": 0.2462875769648678, "grad_norm": 1.6773696964924023, "learning_rate": 1.7645594734955854e-05, "loss": 0.8455, "step": 6800 }, { "epoch": 0.24632379572618615, "grad_norm": 1.3664539562750309, "learning_rate": 1.764483857781625e-05, "loss": 0.8164, "step": 6801 }, { "epoch": 0.24636001448750453, "grad_norm": 1.5083301068360593, "learning_rate": 1.764408231547695e-05, "loss": 0.8128, "step": 6802 }, { "epoch": 0.2463962332488229, "grad_norm": 1.502234287509428, "learning_rate": 1.764332594794836e-05, "loss": 0.8428, "step": 6803 }, { "epoch": 0.24643245201014125, "grad_norm": 1.5585450979577062, "learning_rate": 1.764256947524089e-05, "loss": 0.8109, "step": 6804 }, { "epoch": 0.24646867077145962, "grad_norm": 1.0485841907638682, "learning_rate": 1.764181289736495e-05, "loss": 0.7403, "step": 6805 }, { "epoch": 0.24650488953277797, "grad_norm": 1.530810763898175, "learning_rate": 1.7641056214330946e-05, "loss": 0.8117, "step": 6806 }, { "epoch": 0.24654110829409634, "grad_norm": 1.526903130983229, "learning_rate": 1.76402994261493e-05, "loss": 0.7726, "step": 6807 }, { "epoch": 0.2465773270554147, "grad_norm": 1.3915268017948492, "learning_rate": 1.7639542532830417e-05, "loss": 0.7811, "step": 6808 }, { "epoch": 0.24661354581673306, "grad_norm": 1.5931043035643315, "learning_rate": 1.7638785534384718e-05, "loss": 0.7758, "step": 6809 }, { "epoch": 0.24664976457805143, "grad_norm": 1.5544351296246703, "learning_rate": 1.7638028430822617e-05, "loss": 0.7979, "step": 6810 }, { "epoch": 0.24668598333936978, "grad_norm": 1.4124533259436385, "learning_rate": 1.763727122215453e-05, "loss": 0.8211, "step": 6811 }, { "epoch": 0.24672220210068815, "grad_norm": 1.3339575063867561, "learning_rate": 1.7636513908390888e-05, "loss": 0.8454, "step": 6812 }, { "epoch": 0.24675842086200653, "grad_norm": 1.610443311567913, "learning_rate": 1.76357564895421e-05, "loss": 0.8038, "step": 6813 }, { "epoch": 0.24679463962332487, "grad_norm": 1.372701552551424, "learning_rate": 1.763499896561859e-05, "loss": 0.8311, "step": 6814 }, { "epoch": 0.24683085838464325, "grad_norm": 1.695810950084264, "learning_rate": 1.7634241336630792e-05, "loss": 0.8542, "step": 6815 }, { "epoch": 0.2468670771459616, "grad_norm": 1.6077243860954804, "learning_rate": 1.7633483602589122e-05, "loss": 0.8241, "step": 6816 }, { "epoch": 0.24690329590727997, "grad_norm": 1.6067440647964104, "learning_rate": 1.763272576350401e-05, "loss": 0.8869, "step": 6817 }, { "epoch": 0.24693951466859834, "grad_norm": 1.6562450134454192, "learning_rate": 1.7631967819385883e-05, "loss": 0.8917, "step": 6818 }, { "epoch": 0.2469757334299167, "grad_norm": 1.1546909773732943, "learning_rate": 1.7631209770245178e-05, "loss": 0.8185, "step": 6819 }, { "epoch": 0.24701195219123506, "grad_norm": 1.447890096115891, "learning_rate": 1.7630451616092317e-05, "loss": 0.7413, "step": 6820 }, { "epoch": 0.24704817095255344, "grad_norm": 1.5024754790322006, "learning_rate": 1.762969335693774e-05, "loss": 0.8002, "step": 6821 }, { "epoch": 0.24708438971387178, "grad_norm": 1.3290504954475584, "learning_rate": 1.7628934992791874e-05, "loss": 0.7851, "step": 6822 }, { "epoch": 0.24712060847519016, "grad_norm": 1.557755489846221, "learning_rate": 1.7628176523665163e-05, "loss": 0.8667, "step": 6823 }, { "epoch": 0.2471568272365085, "grad_norm": 1.8541820441547756, "learning_rate": 1.7627417949568037e-05, "loss": 0.8431, "step": 6824 }, { "epoch": 0.24719304599782688, "grad_norm": 1.6469835766753422, "learning_rate": 1.7626659270510942e-05, "loss": 0.8347, "step": 6825 }, { "epoch": 0.24722926475914525, "grad_norm": 1.4453249125334273, "learning_rate": 1.7625900486504314e-05, "loss": 0.7024, "step": 6826 }, { "epoch": 0.2472654835204636, "grad_norm": 1.5794272473239, "learning_rate": 1.762514159755859e-05, "loss": 0.8593, "step": 6827 }, { "epoch": 0.24730170228178197, "grad_norm": 1.4332709411461073, "learning_rate": 1.762438260368422e-05, "loss": 0.7739, "step": 6828 }, { "epoch": 0.24733792104310032, "grad_norm": 1.511935673164881, "learning_rate": 1.762362350489165e-05, "loss": 0.8144, "step": 6829 }, { "epoch": 0.2473741398044187, "grad_norm": 1.9504790570063848, "learning_rate": 1.7622864301191315e-05, "loss": 0.7972, "step": 6830 }, { "epoch": 0.24741035856573707, "grad_norm": 1.6221202955175547, "learning_rate": 1.7622104992593675e-05, "loss": 0.8796, "step": 6831 }, { "epoch": 0.2474465773270554, "grad_norm": 1.4986799697784634, "learning_rate": 1.762134557910917e-05, "loss": 0.7774, "step": 6832 }, { "epoch": 0.24748279608837379, "grad_norm": 1.5949576514699948, "learning_rate": 1.762058606074825e-05, "loss": 0.8479, "step": 6833 }, { "epoch": 0.24751901484969213, "grad_norm": 1.527008111226695, "learning_rate": 1.7619826437521375e-05, "loss": 0.7644, "step": 6834 }, { "epoch": 0.2475552336110105, "grad_norm": 1.191518753243146, "learning_rate": 1.761906670943899e-05, "loss": 0.7512, "step": 6835 }, { "epoch": 0.24759145237232888, "grad_norm": 1.4228463706273309, "learning_rate": 1.7618306876511554e-05, "loss": 0.8242, "step": 6836 }, { "epoch": 0.24762767113364723, "grad_norm": 1.445733341601122, "learning_rate": 1.761754693874952e-05, "loss": 0.7902, "step": 6837 }, { "epoch": 0.2476638898949656, "grad_norm": 1.1664857116468266, "learning_rate": 1.7616786896163347e-05, "loss": 0.8215, "step": 6838 }, { "epoch": 0.24770010865628395, "grad_norm": 1.5633581225050055, "learning_rate": 1.7616026748763495e-05, "loss": 0.8142, "step": 6839 }, { "epoch": 0.24773632741760232, "grad_norm": 1.5488144030931241, "learning_rate": 1.7615266496560425e-05, "loss": 0.7724, "step": 6840 }, { "epoch": 0.2477725461789207, "grad_norm": 1.209002379031463, "learning_rate": 1.761450613956459e-05, "loss": 0.8323, "step": 6841 }, { "epoch": 0.24780876494023904, "grad_norm": 1.6284325686233512, "learning_rate": 1.7613745677786464e-05, "loss": 0.7959, "step": 6842 }, { "epoch": 0.24784498370155741, "grad_norm": 1.5064962700190867, "learning_rate": 1.7612985111236508e-05, "loss": 0.7822, "step": 6843 }, { "epoch": 0.24788120246287576, "grad_norm": 1.5675080436518414, "learning_rate": 1.7612224439925185e-05, "loss": 0.8174, "step": 6844 }, { "epoch": 0.24791742122419413, "grad_norm": 1.1074818078984536, "learning_rate": 1.7611463663862968e-05, "loss": 0.7954, "step": 6845 }, { "epoch": 0.2479536399855125, "grad_norm": 1.5744468598238053, "learning_rate": 1.7610702783060323e-05, "loss": 0.8838, "step": 6846 }, { "epoch": 0.24798985874683085, "grad_norm": 1.4805529037050098, "learning_rate": 1.7609941797527718e-05, "loss": 0.7798, "step": 6847 }, { "epoch": 0.24802607750814923, "grad_norm": 1.1667401432648499, "learning_rate": 1.7609180707275627e-05, "loss": 0.7701, "step": 6848 }, { "epoch": 0.24806229626946757, "grad_norm": 1.6080396005265516, "learning_rate": 1.7608419512314527e-05, "loss": 0.8078, "step": 6849 }, { "epoch": 0.24809851503078595, "grad_norm": 1.7023614650538108, "learning_rate": 1.7607658212654885e-05, "loss": 0.7506, "step": 6850 }, { "epoch": 0.24813473379210432, "grad_norm": 1.2958589956034978, "learning_rate": 1.7606896808307182e-05, "loss": 0.8218, "step": 6851 }, { "epoch": 0.24817095255342267, "grad_norm": 1.5961189755366318, "learning_rate": 1.7606135299281897e-05, "loss": 0.821, "step": 6852 }, { "epoch": 0.24820717131474104, "grad_norm": 1.4767082838721477, "learning_rate": 1.7605373685589503e-05, "loss": 0.8118, "step": 6853 }, { "epoch": 0.2482433900760594, "grad_norm": 1.6661582977364904, "learning_rate": 1.760461196724049e-05, "loss": 0.8406, "step": 6854 }, { "epoch": 0.24827960883737776, "grad_norm": 1.6063562177167519, "learning_rate": 1.7603850144245333e-05, "loss": 0.9004, "step": 6855 }, { "epoch": 0.24831582759869614, "grad_norm": 1.559269912586386, "learning_rate": 1.7603088216614512e-05, "loss": 0.8594, "step": 6856 }, { "epoch": 0.24835204636001448, "grad_norm": 1.116375178165769, "learning_rate": 1.7602326184358524e-05, "loss": 0.7951, "step": 6857 }, { "epoch": 0.24838826512133286, "grad_norm": 1.522291182571918, "learning_rate": 1.760156404748784e-05, "loss": 0.8543, "step": 6858 }, { "epoch": 0.2484244838826512, "grad_norm": 1.0980201737070991, "learning_rate": 1.760080180601296e-05, "loss": 0.7562, "step": 6859 }, { "epoch": 0.24846070264396958, "grad_norm": 1.315298473919379, "learning_rate": 1.7600039459944363e-05, "loss": 0.8107, "step": 6860 }, { "epoch": 0.24849692140528795, "grad_norm": 1.3626794402160542, "learning_rate": 1.759927700929255e-05, "loss": 0.8254, "step": 6861 }, { "epoch": 0.2485331401666063, "grad_norm": 1.6354683353579895, "learning_rate": 1.7598514454068013e-05, "loss": 0.7915, "step": 6862 }, { "epoch": 0.24856935892792467, "grad_norm": 1.5038149310539004, "learning_rate": 1.759775179428123e-05, "loss": 0.8337, "step": 6863 }, { "epoch": 0.24860557768924302, "grad_norm": 1.5492411900875094, "learning_rate": 1.7596989029942714e-05, "loss": 0.8381, "step": 6864 }, { "epoch": 0.2486417964505614, "grad_norm": 1.3543078204934802, "learning_rate": 1.759622616106295e-05, "loss": 0.7784, "step": 6865 }, { "epoch": 0.24867801521187977, "grad_norm": 1.6378759598433683, "learning_rate": 1.759546318765244e-05, "loss": 0.8018, "step": 6866 }, { "epoch": 0.2487142339731981, "grad_norm": 1.439937176961758, "learning_rate": 1.7594700109721686e-05, "loss": 0.7533, "step": 6867 }, { "epoch": 0.2487504527345165, "grad_norm": 1.5240279765087772, "learning_rate": 1.759393692728118e-05, "loss": 0.8986, "step": 6868 }, { "epoch": 0.24878667149583483, "grad_norm": 1.262487970515577, "learning_rate": 1.7593173640341434e-05, "loss": 0.7804, "step": 6869 }, { "epoch": 0.2488228902571532, "grad_norm": 1.200232497850517, "learning_rate": 1.7592410248912943e-05, "loss": 0.7726, "step": 6870 }, { "epoch": 0.24885910901847158, "grad_norm": 1.2290964220589027, "learning_rate": 1.7591646753006214e-05, "loss": 0.7836, "step": 6871 }, { "epoch": 0.24889532777978993, "grad_norm": 1.2055555476784645, "learning_rate": 1.759088315263176e-05, "loss": 0.7809, "step": 6872 }, { "epoch": 0.2489315465411083, "grad_norm": 1.3913864330848942, "learning_rate": 1.7590119447800083e-05, "loss": 0.8305, "step": 6873 }, { "epoch": 0.24896776530242665, "grad_norm": 1.4094042159644977, "learning_rate": 1.758935563852169e-05, "loss": 0.8133, "step": 6874 }, { "epoch": 0.24900398406374502, "grad_norm": 1.0890102564045725, "learning_rate": 1.7588591724807098e-05, "loss": 0.8038, "step": 6875 }, { "epoch": 0.2490402028250634, "grad_norm": 1.3441292368304711, "learning_rate": 1.7587827706666813e-05, "loss": 0.7942, "step": 6876 }, { "epoch": 0.24907642158638174, "grad_norm": 1.4367077136795627, "learning_rate": 1.7587063584111354e-05, "loss": 0.7833, "step": 6877 }, { "epoch": 0.24911264034770011, "grad_norm": 1.1093068309696223, "learning_rate": 1.7586299357151234e-05, "loss": 0.7012, "step": 6878 }, { "epoch": 0.24914885910901846, "grad_norm": 1.555966787823022, "learning_rate": 1.7585535025796966e-05, "loss": 0.8695, "step": 6879 }, { "epoch": 0.24918507787033684, "grad_norm": 1.3927835397242962, "learning_rate": 1.758477059005907e-05, "loss": 0.7982, "step": 6880 }, { "epoch": 0.2492212966316552, "grad_norm": 1.4862314912158474, "learning_rate": 1.7584006049948072e-05, "loss": 0.7716, "step": 6881 }, { "epoch": 0.24925751539297356, "grad_norm": 1.4437141034176126, "learning_rate": 1.7583241405474482e-05, "loss": 0.8387, "step": 6882 }, { "epoch": 0.24929373415429193, "grad_norm": 1.562017259531309, "learning_rate": 1.758247665664883e-05, "loss": 0.8051, "step": 6883 }, { "epoch": 0.24932995291561028, "grad_norm": 1.495879246296884, "learning_rate": 1.7581711803481635e-05, "loss": 0.7982, "step": 6884 }, { "epoch": 0.24936617167692865, "grad_norm": 1.6198194229000704, "learning_rate": 1.7580946845983425e-05, "loss": 0.7702, "step": 6885 }, { "epoch": 0.24940239043824702, "grad_norm": 1.612398850702487, "learning_rate": 1.758018178416472e-05, "loss": 0.7713, "step": 6886 }, { "epoch": 0.24943860919956537, "grad_norm": 1.4848751323544387, "learning_rate": 1.757941661803606e-05, "loss": 0.8567, "step": 6887 }, { "epoch": 0.24947482796088374, "grad_norm": 1.4904928420719386, "learning_rate": 1.7578651347607964e-05, "loss": 0.7774, "step": 6888 }, { "epoch": 0.2495110467222021, "grad_norm": 1.423394329408805, "learning_rate": 1.7577885972890966e-05, "loss": 0.8187, "step": 6889 }, { "epoch": 0.24954726548352046, "grad_norm": 1.4629559425928886, "learning_rate": 1.7577120493895597e-05, "loss": 0.8669, "step": 6890 }, { "epoch": 0.24958348424483884, "grad_norm": 1.720802497619367, "learning_rate": 1.7576354910632397e-05, "loss": 0.778, "step": 6891 }, { "epoch": 0.24961970300615718, "grad_norm": 1.356672403922806, "learning_rate": 1.757558922311189e-05, "loss": 0.7787, "step": 6892 }, { "epoch": 0.24965592176747556, "grad_norm": 1.5148010152461755, "learning_rate": 1.7574823431344626e-05, "loss": 0.7828, "step": 6893 }, { "epoch": 0.2496921405287939, "grad_norm": 1.3829311344206072, "learning_rate": 1.7574057535341134e-05, "loss": 0.8223, "step": 6894 }, { "epoch": 0.24972835929011228, "grad_norm": 1.8197235916848788, "learning_rate": 1.7573291535111955e-05, "loss": 0.8895, "step": 6895 }, { "epoch": 0.24976457805143065, "grad_norm": 1.2123359285296789, "learning_rate": 1.7572525430667627e-05, "loss": 0.7785, "step": 6896 }, { "epoch": 0.249800796812749, "grad_norm": 1.2625065315382733, "learning_rate": 1.7571759222018697e-05, "loss": 0.7764, "step": 6897 }, { "epoch": 0.24983701557406737, "grad_norm": 1.1440794738374338, "learning_rate": 1.7570992909175706e-05, "loss": 0.7683, "step": 6898 }, { "epoch": 0.24987323433538572, "grad_norm": 1.4344711204498255, "learning_rate": 1.75702264921492e-05, "loss": 0.7547, "step": 6899 }, { "epoch": 0.2499094530967041, "grad_norm": 1.0121901871190129, "learning_rate": 1.7569459970949727e-05, "loss": 0.8333, "step": 6900 }, { "epoch": 0.24994567185802247, "grad_norm": 1.499707539304893, "learning_rate": 1.7568693345587834e-05, "loss": 0.7889, "step": 6901 }, { "epoch": 0.2499818906193408, "grad_norm": 1.5338529369957037, "learning_rate": 1.756792661607407e-05, "loss": 0.9038, "step": 6902 }, { "epoch": 0.25001810938065916, "grad_norm": 1.4161709458569631, "learning_rate": 1.7567159782418985e-05, "loss": 0.8067, "step": 6903 }, { "epoch": 0.25005432814197753, "grad_norm": 1.2341344039066173, "learning_rate": 1.7566392844633133e-05, "loss": 0.7819, "step": 6904 }, { "epoch": 0.2500905469032959, "grad_norm": 1.1908623030647405, "learning_rate": 1.7565625802727068e-05, "loss": 0.7834, "step": 6905 }, { "epoch": 0.2501267656646143, "grad_norm": 1.0617100989700772, "learning_rate": 1.7564858656711343e-05, "loss": 0.8159, "step": 6906 }, { "epoch": 0.25016298442593266, "grad_norm": 1.5069017531879112, "learning_rate": 1.756409140659651e-05, "loss": 0.8724, "step": 6907 }, { "epoch": 0.250199203187251, "grad_norm": 1.5863911979242624, "learning_rate": 1.7563324052393142e-05, "loss": 0.7915, "step": 6908 }, { "epoch": 0.25023542194856935, "grad_norm": 1.2644147014852742, "learning_rate": 1.7562556594111786e-05, "loss": 0.8164, "step": 6909 }, { "epoch": 0.2502716407098877, "grad_norm": 1.4619544883036317, "learning_rate": 1.756178903176301e-05, "loss": 0.826, "step": 6910 }, { "epoch": 0.2503078594712061, "grad_norm": 1.593657453505303, "learning_rate": 1.7561021365357366e-05, "loss": 0.9339, "step": 6911 }, { "epoch": 0.25034407823252447, "grad_norm": 1.4299030545337348, "learning_rate": 1.7560253594905425e-05, "loss": 0.8215, "step": 6912 }, { "epoch": 0.2503802969938428, "grad_norm": 1.7272216091166506, "learning_rate": 1.7559485720417753e-05, "loss": 0.8498, "step": 6913 }, { "epoch": 0.25041651575516116, "grad_norm": 1.4973475715212663, "learning_rate": 1.755871774190492e-05, "loss": 0.8968, "step": 6914 }, { "epoch": 0.25045273451647954, "grad_norm": 1.3860606660271295, "learning_rate": 1.755794965937748e-05, "loss": 0.8342, "step": 6915 }, { "epoch": 0.2504889532777979, "grad_norm": 1.2972980186018424, "learning_rate": 1.755718147284602e-05, "loss": 0.7357, "step": 6916 }, { "epoch": 0.2505251720391163, "grad_norm": 1.4488551224738477, "learning_rate": 1.75564131823211e-05, "loss": 0.7907, "step": 6917 }, { "epoch": 0.2505613908004346, "grad_norm": 1.400004932328885, "learning_rate": 1.755564478781329e-05, "loss": 0.8083, "step": 6918 }, { "epoch": 0.250597609561753, "grad_norm": 1.5171634826860982, "learning_rate": 1.7554876289333175e-05, "loss": 0.7817, "step": 6919 }, { "epoch": 0.25063382832307135, "grad_norm": 1.4515050644182579, "learning_rate": 1.7554107686891324e-05, "loss": 0.7239, "step": 6920 }, { "epoch": 0.2506700470843897, "grad_norm": 1.4135043619449623, "learning_rate": 1.755333898049831e-05, "loss": 0.7719, "step": 6921 }, { "epoch": 0.2507062658457081, "grad_norm": 1.1709793038437464, "learning_rate": 1.7552570170164718e-05, "loss": 0.7929, "step": 6922 }, { "epoch": 0.2507424846070264, "grad_norm": 1.6446642810538687, "learning_rate": 1.755180125590112e-05, "loss": 0.8624, "step": 6923 }, { "epoch": 0.2507787033683448, "grad_norm": 1.5257449324729728, "learning_rate": 1.7551032237718105e-05, "loss": 0.7736, "step": 6924 }, { "epoch": 0.25081492212966316, "grad_norm": 1.441537206278201, "learning_rate": 1.755026311562625e-05, "loss": 0.738, "step": 6925 }, { "epoch": 0.25085114089098154, "grad_norm": 1.3967594916101587, "learning_rate": 1.754949388963614e-05, "loss": 0.7615, "step": 6926 }, { "epoch": 0.2508873596522999, "grad_norm": 1.432650500055208, "learning_rate": 1.754872455975836e-05, "loss": 0.8666, "step": 6927 }, { "epoch": 0.25092357841361823, "grad_norm": 3.187269004347602, "learning_rate": 1.75479551260035e-05, "loss": 0.8383, "step": 6928 }, { "epoch": 0.2509597971749366, "grad_norm": 1.5083825795279657, "learning_rate": 1.754718558838214e-05, "loss": 0.7842, "step": 6929 }, { "epoch": 0.250996015936255, "grad_norm": 1.456940288874783, "learning_rate": 1.754641594690488e-05, "loss": 0.7653, "step": 6930 }, { "epoch": 0.25103223469757335, "grad_norm": 1.4469677815545399, "learning_rate": 1.7545646201582304e-05, "loss": 0.8012, "step": 6931 }, { "epoch": 0.2510684534588917, "grad_norm": 1.409302436494726, "learning_rate": 1.7544876352425005e-05, "loss": 0.8198, "step": 6932 }, { "epoch": 0.25110467222021005, "grad_norm": 1.4557422300126799, "learning_rate": 1.754410639944358e-05, "loss": 0.8785, "step": 6933 }, { "epoch": 0.2511408909815284, "grad_norm": 1.5538121004096812, "learning_rate": 1.754333634264862e-05, "loss": 0.9138, "step": 6934 }, { "epoch": 0.2511771097428468, "grad_norm": 1.455946785902733, "learning_rate": 1.7542566182050725e-05, "loss": 0.7718, "step": 6935 }, { "epoch": 0.25121332850416517, "grad_norm": 1.4535752080993414, "learning_rate": 1.754179591766049e-05, "loss": 0.8827, "step": 6936 }, { "epoch": 0.25124954726548354, "grad_norm": 1.5825260042165867, "learning_rate": 1.7541025549488517e-05, "loss": 0.8533, "step": 6937 }, { "epoch": 0.25128576602680186, "grad_norm": 1.9682456952528489, "learning_rate": 1.7540255077545408e-05, "loss": 0.7971, "step": 6938 }, { "epoch": 0.25132198478812023, "grad_norm": 1.4138747039196742, "learning_rate": 1.7539484501841763e-05, "loss": 0.7392, "step": 6939 }, { "epoch": 0.2513582035494386, "grad_norm": 1.4253015004994214, "learning_rate": 1.7538713822388186e-05, "loss": 0.7968, "step": 6940 }, { "epoch": 0.251394422310757, "grad_norm": 1.580376636943514, "learning_rate": 1.7537943039195282e-05, "loss": 0.7894, "step": 6941 }, { "epoch": 0.25143064107207536, "grad_norm": 1.5369321061690793, "learning_rate": 1.753717215227366e-05, "loss": 0.9033, "step": 6942 }, { "epoch": 0.2514668598333937, "grad_norm": 1.4580698121488558, "learning_rate": 1.7536401161633927e-05, "loss": 0.7844, "step": 6943 }, { "epoch": 0.25150307859471205, "grad_norm": 1.4002836324874162, "learning_rate": 1.7535630067286692e-05, "loss": 0.8316, "step": 6944 }, { "epoch": 0.2515392973560304, "grad_norm": 1.5882302891051419, "learning_rate": 1.7534858869242563e-05, "loss": 0.8465, "step": 6945 }, { "epoch": 0.2515755161173488, "grad_norm": 1.361821832163914, "learning_rate": 1.7534087567512156e-05, "loss": 0.7615, "step": 6946 }, { "epoch": 0.25161173487866717, "grad_norm": 1.5799717248426752, "learning_rate": 1.7533316162106088e-05, "loss": 0.7858, "step": 6947 }, { "epoch": 0.2516479536399855, "grad_norm": 1.4385970014094351, "learning_rate": 1.7532544653034964e-05, "loss": 0.8186, "step": 6948 }, { "epoch": 0.25168417240130386, "grad_norm": 1.3551112714987283, "learning_rate": 1.753177304030941e-05, "loss": 0.7968, "step": 6949 }, { "epoch": 0.25172039116262224, "grad_norm": 1.562480769039064, "learning_rate": 1.753100132394004e-05, "loss": 0.9784, "step": 6950 }, { "epoch": 0.2517566099239406, "grad_norm": 1.4043095678304967, "learning_rate": 1.7530229503937474e-05, "loss": 0.8295, "step": 6951 }, { "epoch": 0.251792828685259, "grad_norm": 1.4579234996315096, "learning_rate": 1.7529457580312336e-05, "loss": 0.8332, "step": 6952 }, { "epoch": 0.2518290474465773, "grad_norm": 1.4285220538248422, "learning_rate": 1.7528685553075244e-05, "loss": 0.8077, "step": 6953 }, { "epoch": 0.2518652662078957, "grad_norm": 1.3493476067280763, "learning_rate": 1.7527913422236828e-05, "loss": 0.7316, "step": 6954 }, { "epoch": 0.25190148496921405, "grad_norm": 1.643283110975993, "learning_rate": 1.7527141187807705e-05, "loss": 0.8626, "step": 6955 }, { "epoch": 0.2519377037305324, "grad_norm": 1.4802565938472214, "learning_rate": 1.7526368849798504e-05, "loss": 0.8114, "step": 6956 }, { "epoch": 0.2519739224918508, "grad_norm": 1.711991984568225, "learning_rate": 1.7525596408219858e-05, "loss": 0.7369, "step": 6957 }, { "epoch": 0.2520101412531691, "grad_norm": 1.3834946284821874, "learning_rate": 1.7524823863082387e-05, "loss": 0.8065, "step": 6958 }, { "epoch": 0.2520463600144875, "grad_norm": 1.4608684985037452, "learning_rate": 1.752405121439673e-05, "loss": 0.8254, "step": 6959 }, { "epoch": 0.25208257877580587, "grad_norm": 1.3672968670834145, "learning_rate": 1.7523278462173518e-05, "loss": 0.8306, "step": 6960 }, { "epoch": 0.25211879753712424, "grad_norm": 1.2808393132391578, "learning_rate": 1.7522505606423387e-05, "loss": 0.7666, "step": 6961 }, { "epoch": 0.2521550162984426, "grad_norm": 1.6416746335443428, "learning_rate": 1.7521732647156963e-05, "loss": 0.8016, "step": 6962 }, { "epoch": 0.25219123505976093, "grad_norm": 1.560115187548235, "learning_rate": 1.7520959584384895e-05, "loss": 0.8054, "step": 6963 }, { "epoch": 0.2522274538210793, "grad_norm": 1.5939821174765265, "learning_rate": 1.7520186418117808e-05, "loss": 0.8098, "step": 6964 }, { "epoch": 0.2522636725823977, "grad_norm": 1.3123425534908904, "learning_rate": 1.751941314836635e-05, "loss": 0.7872, "step": 6965 }, { "epoch": 0.25229989134371605, "grad_norm": 1.5141817944458897, "learning_rate": 1.7518639775141163e-05, "loss": 0.7744, "step": 6966 }, { "epoch": 0.25233611010503443, "grad_norm": 1.4086588612834856, "learning_rate": 1.7517866298452886e-05, "loss": 0.8434, "step": 6967 }, { "epoch": 0.25237232886635275, "grad_norm": 1.4660268195894028, "learning_rate": 1.751709271831216e-05, "loss": 0.972, "step": 6968 }, { "epoch": 0.2524085476276711, "grad_norm": 1.438600370384206, "learning_rate": 1.7516319034729634e-05, "loss": 0.7842, "step": 6969 }, { "epoch": 0.2524447663889895, "grad_norm": 1.38257004476556, "learning_rate": 1.7515545247715956e-05, "loss": 0.8259, "step": 6970 }, { "epoch": 0.25248098515030787, "grad_norm": 1.3354080661102712, "learning_rate": 1.751477135728177e-05, "loss": 0.6745, "step": 6971 }, { "epoch": 0.25251720391162624, "grad_norm": 1.5467335029389417, "learning_rate": 1.7513997363437728e-05, "loss": 0.738, "step": 6972 }, { "epoch": 0.25255342267294456, "grad_norm": 1.3965590103585532, "learning_rate": 1.7513223266194476e-05, "loss": 0.8035, "step": 6973 }, { "epoch": 0.25258964143426293, "grad_norm": 1.5544714400144295, "learning_rate": 1.7512449065562674e-05, "loss": 0.8588, "step": 6974 }, { "epoch": 0.2526258601955813, "grad_norm": 1.5824078017555143, "learning_rate": 1.7511674761552973e-05, "loss": 0.8146, "step": 6975 }, { "epoch": 0.2526620789568997, "grad_norm": 1.3340276443032204, "learning_rate": 1.7510900354176024e-05, "loss": 0.764, "step": 6976 }, { "epoch": 0.25269829771821806, "grad_norm": 1.6405850076342392, "learning_rate": 1.7510125843442485e-05, "loss": 0.9056, "step": 6977 }, { "epoch": 0.2527345164795364, "grad_norm": 1.5033098016514923, "learning_rate": 1.750935122936302e-05, "loss": 0.8367, "step": 6978 }, { "epoch": 0.25277073524085475, "grad_norm": 1.2540572341441936, "learning_rate": 1.750857651194828e-05, "loss": 0.7864, "step": 6979 }, { "epoch": 0.2528069540021731, "grad_norm": 1.4872340419119217, "learning_rate": 1.7507801691208932e-05, "loss": 0.9058, "step": 6980 }, { "epoch": 0.2528431727634915, "grad_norm": 1.343780513073225, "learning_rate": 1.7507026767155634e-05, "loss": 0.8264, "step": 6981 }, { "epoch": 0.25287939152480987, "grad_norm": 1.3711895043201188, "learning_rate": 1.7506251739799052e-05, "loss": 0.8055, "step": 6982 }, { "epoch": 0.2529156102861282, "grad_norm": 1.1570669520337427, "learning_rate": 1.750547660914985e-05, "loss": 0.7995, "step": 6983 }, { "epoch": 0.25295182904744656, "grad_norm": 1.4954222200413148, "learning_rate": 1.7504701375218697e-05, "loss": 0.8633, "step": 6984 }, { "epoch": 0.25298804780876494, "grad_norm": 1.4730994394979446, "learning_rate": 1.750392603801626e-05, "loss": 0.8455, "step": 6985 }, { "epoch": 0.2530242665700833, "grad_norm": 1.354809453671147, "learning_rate": 1.7503150597553205e-05, "loss": 0.8028, "step": 6986 }, { "epoch": 0.2530604853314017, "grad_norm": 1.2084975798433975, "learning_rate": 1.7502375053840205e-05, "loss": 0.8736, "step": 6987 }, { "epoch": 0.25309670409272, "grad_norm": 1.4359914869655335, "learning_rate": 1.750159940688793e-05, "loss": 0.7273, "step": 6988 }, { "epoch": 0.2531329228540384, "grad_norm": 1.6184723807244439, "learning_rate": 1.7500823656707056e-05, "loss": 0.7927, "step": 6989 }, { "epoch": 0.25316914161535675, "grad_norm": 1.5608645606349811, "learning_rate": 1.750004780330826e-05, "loss": 0.8918, "step": 6990 }, { "epoch": 0.2532053603766751, "grad_norm": 1.618835290434827, "learning_rate": 1.7499271846702216e-05, "loss": 0.868, "step": 6991 }, { "epoch": 0.2532415791379935, "grad_norm": 1.4454297340375943, "learning_rate": 1.74984957868996e-05, "loss": 0.8653, "step": 6992 }, { "epoch": 0.2532777978993118, "grad_norm": 1.5515509768247318, "learning_rate": 1.7497719623911095e-05, "loss": 0.7519, "step": 6993 }, { "epoch": 0.2533140166606302, "grad_norm": 1.584690350594098, "learning_rate": 1.749694335774738e-05, "loss": 0.9318, "step": 6994 }, { "epoch": 0.25335023542194857, "grad_norm": 1.4028888303183662, "learning_rate": 1.7496166988419135e-05, "loss": 0.8025, "step": 6995 }, { "epoch": 0.25338645418326694, "grad_norm": 1.4697526897756847, "learning_rate": 1.749539051593705e-05, "loss": 0.8289, "step": 6996 }, { "epoch": 0.2534226729445853, "grad_norm": 1.1789349160304055, "learning_rate": 1.74946139403118e-05, "loss": 0.8165, "step": 6997 }, { "epoch": 0.25345889170590363, "grad_norm": 1.4720554232872445, "learning_rate": 1.7493837261554077e-05, "loss": 0.7961, "step": 6998 }, { "epoch": 0.253495110467222, "grad_norm": 1.3766933155470384, "learning_rate": 1.7493060479674573e-05, "loss": 0.8507, "step": 6999 }, { "epoch": 0.2535313292285404, "grad_norm": 1.4409963960675314, "learning_rate": 1.749228359468397e-05, "loss": 0.8636, "step": 7000 }, { "epoch": 0.25356754798985875, "grad_norm": 1.4675109384899163, "learning_rate": 1.7491506606592962e-05, "loss": 0.866, "step": 7001 }, { "epoch": 0.25360376675117713, "grad_norm": 1.204589794268721, "learning_rate": 1.749072951541224e-05, "loss": 0.7938, "step": 7002 }, { "epoch": 0.25363998551249545, "grad_norm": 1.5134181997556075, "learning_rate": 1.74899523211525e-05, "loss": 0.8846, "step": 7003 }, { "epoch": 0.2536762042738138, "grad_norm": 1.3966860085420212, "learning_rate": 1.7489175023824435e-05, "loss": 0.8755, "step": 7004 }, { "epoch": 0.2537124230351322, "grad_norm": 1.4648748241943323, "learning_rate": 1.7488397623438738e-05, "loss": 0.8101, "step": 7005 }, { "epoch": 0.25374864179645057, "grad_norm": 1.3505912933739903, "learning_rate": 1.748762012000611e-05, "loss": 0.8136, "step": 7006 }, { "epoch": 0.25378486055776894, "grad_norm": 1.2911690333717067, "learning_rate": 1.7486842513537253e-05, "loss": 0.8288, "step": 7007 }, { "epoch": 0.25382107931908726, "grad_norm": 1.5697725824433422, "learning_rate": 1.748606480404286e-05, "loss": 0.8002, "step": 7008 }, { "epoch": 0.25385729808040564, "grad_norm": 1.3724129557823253, "learning_rate": 1.748528699153364e-05, "loss": 0.7425, "step": 7009 }, { "epoch": 0.253893516841724, "grad_norm": 1.4071012781350878, "learning_rate": 1.74845090760203e-05, "loss": 0.8338, "step": 7010 }, { "epoch": 0.2539297356030424, "grad_norm": 1.4429708017149807, "learning_rate": 1.748373105751353e-05, "loss": 0.8566, "step": 7011 }, { "epoch": 0.25396595436436076, "grad_norm": 1.3412379967323724, "learning_rate": 1.7482952936024045e-05, "loss": 0.7951, "step": 7012 }, { "epoch": 0.2540021731256791, "grad_norm": 1.3057874647641705, "learning_rate": 1.748217471156256e-05, "loss": 0.8627, "step": 7013 }, { "epoch": 0.25403839188699745, "grad_norm": 1.3907063284944698, "learning_rate": 1.748139638413977e-05, "loss": 0.7447, "step": 7014 }, { "epoch": 0.2540746106483158, "grad_norm": 1.2321532277058054, "learning_rate": 1.7480617953766393e-05, "loss": 0.8271, "step": 7015 }, { "epoch": 0.2541108294096342, "grad_norm": 1.4343290944171105, "learning_rate": 1.747983942045314e-05, "loss": 0.791, "step": 7016 }, { "epoch": 0.25414704817095257, "grad_norm": 1.4253363679484066, "learning_rate": 1.7479060784210723e-05, "loss": 0.8045, "step": 7017 }, { "epoch": 0.2541832669322709, "grad_norm": 1.559581893897006, "learning_rate": 1.747828204504986e-05, "loss": 0.7776, "step": 7018 }, { "epoch": 0.25421948569358926, "grad_norm": 1.4584938894739854, "learning_rate": 1.7477503202981266e-05, "loss": 0.8272, "step": 7019 }, { "epoch": 0.25425570445490764, "grad_norm": 1.4352329468830318, "learning_rate": 1.7476724258015655e-05, "loss": 0.8471, "step": 7020 }, { "epoch": 0.254291923216226, "grad_norm": 1.453027170343297, "learning_rate": 1.747594521016375e-05, "loss": 0.7778, "step": 7021 }, { "epoch": 0.2543281419775444, "grad_norm": 1.8421179611890586, "learning_rate": 1.747516605943627e-05, "loss": 0.8649, "step": 7022 }, { "epoch": 0.2543643607388627, "grad_norm": 1.5502216220568192, "learning_rate": 1.7474386805843936e-05, "loss": 0.8335, "step": 7023 }, { "epoch": 0.2544005795001811, "grad_norm": 1.193592747190873, "learning_rate": 1.7473607449397475e-05, "loss": 0.7703, "step": 7024 }, { "epoch": 0.25443679826149945, "grad_norm": 1.0945500704578766, "learning_rate": 1.7472827990107607e-05, "loss": 0.7964, "step": 7025 }, { "epoch": 0.2544730170228178, "grad_norm": 1.5501704322266923, "learning_rate": 1.7472048427985056e-05, "loss": 0.7938, "step": 7026 }, { "epoch": 0.2545092357841362, "grad_norm": 1.172425980581397, "learning_rate": 1.747126876304056e-05, "loss": 0.7986, "step": 7027 }, { "epoch": 0.2545454545454545, "grad_norm": 1.3404861229598806, "learning_rate": 1.7470488995284833e-05, "loss": 0.7383, "step": 7028 }, { "epoch": 0.2545816733067729, "grad_norm": 1.5288716988224198, "learning_rate": 1.746970912472862e-05, "loss": 0.8448, "step": 7029 }, { "epoch": 0.25461789206809127, "grad_norm": 1.4984198283803243, "learning_rate": 1.7468929151382644e-05, "loss": 0.7481, "step": 7030 }, { "epoch": 0.25465411082940964, "grad_norm": 1.37126774512574, "learning_rate": 1.7468149075257638e-05, "loss": 0.8263, "step": 7031 }, { "epoch": 0.254690329590728, "grad_norm": 1.6832768823571773, "learning_rate": 1.746736889636434e-05, "loss": 0.714, "step": 7032 }, { "epoch": 0.25472654835204633, "grad_norm": 1.5330782323131376, "learning_rate": 1.7466588614713484e-05, "loss": 0.7384, "step": 7033 }, { "epoch": 0.2547627671133647, "grad_norm": 1.44928081461809, "learning_rate": 1.746580823031581e-05, "loss": 0.8349, "step": 7034 }, { "epoch": 0.2547989858746831, "grad_norm": 1.588250559309998, "learning_rate": 1.7465027743182053e-05, "loss": 0.8725, "step": 7035 }, { "epoch": 0.25483520463600146, "grad_norm": 1.1818613559490345, "learning_rate": 1.7464247153322955e-05, "loss": 0.828, "step": 7036 }, { "epoch": 0.25487142339731983, "grad_norm": 1.2558095322917846, "learning_rate": 1.7463466460749257e-05, "loss": 0.7327, "step": 7037 }, { "epoch": 0.25490764215863815, "grad_norm": 1.7881727379856913, "learning_rate": 1.7462685665471703e-05, "loss": 0.8085, "step": 7038 }, { "epoch": 0.2549438609199565, "grad_norm": 1.4316301254366235, "learning_rate": 1.746190476750104e-05, "loss": 0.7755, "step": 7039 }, { "epoch": 0.2549800796812749, "grad_norm": 1.4455762145732611, "learning_rate": 1.7461123766848005e-05, "loss": 0.8487, "step": 7040 }, { "epoch": 0.25501629844259327, "grad_norm": 1.2028711971655979, "learning_rate": 1.7460342663523352e-05, "loss": 0.8352, "step": 7041 }, { "epoch": 0.25505251720391164, "grad_norm": 1.3010736479053249, "learning_rate": 1.7459561457537835e-05, "loss": 0.8103, "step": 7042 }, { "epoch": 0.25508873596522996, "grad_norm": 1.3503095839897554, "learning_rate": 1.7458780148902193e-05, "loss": 0.787, "step": 7043 }, { "epoch": 0.25512495472654834, "grad_norm": 1.56564053586763, "learning_rate": 1.7457998737627183e-05, "loss": 0.8344, "step": 7044 }, { "epoch": 0.2551611734878667, "grad_norm": 1.4452501979493522, "learning_rate": 1.7457217223723556e-05, "loss": 0.8226, "step": 7045 }, { "epoch": 0.2551973922491851, "grad_norm": 1.4762608160655097, "learning_rate": 1.7456435607202067e-05, "loss": 0.8835, "step": 7046 }, { "epoch": 0.25523361101050346, "grad_norm": 1.4613579996578867, "learning_rate": 1.7455653888073473e-05, "loss": 0.8609, "step": 7047 }, { "epoch": 0.2552698297718218, "grad_norm": 1.566378291158801, "learning_rate": 1.7454872066348533e-05, "loss": 0.8138, "step": 7048 }, { "epoch": 0.25530604853314015, "grad_norm": 1.4262629508493878, "learning_rate": 1.7454090142038e-05, "loss": 0.832, "step": 7049 }, { "epoch": 0.2553422672944585, "grad_norm": 1.4000119211975104, "learning_rate": 1.7453308115152638e-05, "loss": 0.7548, "step": 7050 }, { "epoch": 0.2553784860557769, "grad_norm": 1.5159460875484776, "learning_rate": 1.745252598570321e-05, "loss": 0.875, "step": 7051 }, { "epoch": 0.2554147048170953, "grad_norm": 1.402545036544369, "learning_rate": 1.7451743753700472e-05, "loss": 0.8574, "step": 7052 }, { "epoch": 0.2554509235784136, "grad_norm": 1.3379163033046866, "learning_rate": 1.7450961419155192e-05, "loss": 0.8083, "step": 7053 }, { "epoch": 0.25548714233973197, "grad_norm": 1.3976609995192488, "learning_rate": 1.745017898207814e-05, "loss": 0.8471, "step": 7054 }, { "epoch": 0.25552336110105034, "grad_norm": 1.2812595266222049, "learning_rate": 1.7449396442480076e-05, "loss": 0.7934, "step": 7055 }, { "epoch": 0.2555595798623687, "grad_norm": 1.5313609433368551, "learning_rate": 1.7448613800371772e-05, "loss": 0.7501, "step": 7056 }, { "epoch": 0.2555957986236871, "grad_norm": 1.136779155551813, "learning_rate": 1.7447831055763995e-05, "loss": 0.8226, "step": 7057 }, { "epoch": 0.2556320173850054, "grad_norm": 1.641371075971859, "learning_rate": 1.744704820866752e-05, "loss": 0.866, "step": 7058 }, { "epoch": 0.2556682361463238, "grad_norm": 1.150741805074237, "learning_rate": 1.744626525909312e-05, "loss": 0.7894, "step": 7059 }, { "epoch": 0.25570445490764215, "grad_norm": 1.47964382024079, "learning_rate": 1.7445482207051564e-05, "loss": 0.8508, "step": 7060 }, { "epoch": 0.2557406736689605, "grad_norm": 1.498007592925854, "learning_rate": 1.7444699052553634e-05, "loss": 0.8318, "step": 7061 }, { "epoch": 0.2557768924302789, "grad_norm": 1.4271320130376866, "learning_rate": 1.7443915795610104e-05, "loss": 0.8695, "step": 7062 }, { "epoch": 0.2558131111915972, "grad_norm": 1.5545593764835404, "learning_rate": 1.7443132436231746e-05, "loss": 0.8226, "step": 7063 }, { "epoch": 0.2558493299529156, "grad_norm": 1.5732739431726983, "learning_rate": 1.744234897442935e-05, "loss": 0.829, "step": 7064 }, { "epoch": 0.25588554871423397, "grad_norm": 1.464926171209271, "learning_rate": 1.7441565410213693e-05, "loss": 0.9084, "step": 7065 }, { "epoch": 0.25592176747555234, "grad_norm": 1.441251208849927, "learning_rate": 1.7440781743595556e-05, "loss": 0.8471, "step": 7066 }, { "epoch": 0.2559579862368707, "grad_norm": 1.3095184510317963, "learning_rate": 1.7439997974585722e-05, "loss": 0.791, "step": 7067 }, { "epoch": 0.25599420499818903, "grad_norm": 1.632148057638897, "learning_rate": 1.7439214103194985e-05, "loss": 0.7801, "step": 7068 }, { "epoch": 0.2560304237595074, "grad_norm": 1.602173408435543, "learning_rate": 1.743843012943412e-05, "loss": 0.8497, "step": 7069 }, { "epoch": 0.2560666425208258, "grad_norm": 1.968095307257914, "learning_rate": 1.743764605331392e-05, "loss": 0.7671, "step": 7070 }, { "epoch": 0.25610286128214416, "grad_norm": 1.6230280584720438, "learning_rate": 1.7436861874845178e-05, "loss": 0.7807, "step": 7071 }, { "epoch": 0.25613908004346253, "grad_norm": 1.0601020971646535, "learning_rate": 1.7436077594038678e-05, "loss": 0.7656, "step": 7072 }, { "epoch": 0.25617529880478085, "grad_norm": 1.555338535146499, "learning_rate": 1.743529321090522e-05, "loss": 0.8411, "step": 7073 }, { "epoch": 0.2562115175660992, "grad_norm": 1.497370641883468, "learning_rate": 1.7434508725455594e-05, "loss": 0.853, "step": 7074 }, { "epoch": 0.2562477363274176, "grad_norm": 1.0792777243204608, "learning_rate": 1.7433724137700596e-05, "loss": 0.7879, "step": 7075 }, { "epoch": 0.25628395508873597, "grad_norm": 1.5480284725137878, "learning_rate": 1.7432939447651018e-05, "loss": 0.8578, "step": 7076 }, { "epoch": 0.25632017385005434, "grad_norm": 1.3546090422859314, "learning_rate": 1.7432154655317668e-05, "loss": 0.7589, "step": 7077 }, { "epoch": 0.2563563926113727, "grad_norm": 1.3589164451009652, "learning_rate": 1.7431369760711338e-05, "loss": 0.7515, "step": 7078 }, { "epoch": 0.25639261137269104, "grad_norm": 1.5369110967142352, "learning_rate": 1.7430584763842824e-05, "loss": 0.8757, "step": 7079 }, { "epoch": 0.2564288301340094, "grad_norm": 1.4555424794172738, "learning_rate": 1.742979966472294e-05, "loss": 0.783, "step": 7080 }, { "epoch": 0.2564650488953278, "grad_norm": 1.4866410313296536, "learning_rate": 1.7429014463362483e-05, "loss": 0.8179, "step": 7081 }, { "epoch": 0.25650126765664616, "grad_norm": 1.2132035573907238, "learning_rate": 1.742822915977226e-05, "loss": 0.8046, "step": 7082 }, { "epoch": 0.25653748641796453, "grad_norm": 1.5802880667562236, "learning_rate": 1.7427443753963077e-05, "loss": 0.8273, "step": 7083 }, { "epoch": 0.25657370517928285, "grad_norm": 1.526480938973804, "learning_rate": 1.7426658245945742e-05, "loss": 0.8352, "step": 7084 }, { "epoch": 0.2566099239406012, "grad_norm": 1.4790260613017234, "learning_rate": 1.742587263573106e-05, "loss": 0.8085, "step": 7085 }, { "epoch": 0.2566461427019196, "grad_norm": 1.2106771165616295, "learning_rate": 1.742508692332985e-05, "loss": 0.8504, "step": 7086 }, { "epoch": 0.256682361463238, "grad_norm": 1.490235743640756, "learning_rate": 1.742430110875292e-05, "loss": 0.7892, "step": 7087 }, { "epoch": 0.25671858022455635, "grad_norm": 1.6197496069622697, "learning_rate": 1.7423515192011078e-05, "loss": 0.8635, "step": 7088 }, { "epoch": 0.25675479898587467, "grad_norm": 1.655403021294746, "learning_rate": 1.7422729173115148e-05, "loss": 0.8361, "step": 7089 }, { "epoch": 0.25679101774719304, "grad_norm": 1.4047341879467996, "learning_rate": 1.7421943052075942e-05, "loss": 0.876, "step": 7090 }, { "epoch": 0.2568272365085114, "grad_norm": 1.499209619199093, "learning_rate": 1.742115682890428e-05, "loss": 0.8008, "step": 7091 }, { "epoch": 0.2568634552698298, "grad_norm": 1.1985581979282753, "learning_rate": 1.7420370503610975e-05, "loss": 0.7708, "step": 7092 }, { "epoch": 0.25689967403114816, "grad_norm": 1.3893310783638677, "learning_rate": 1.7419584076206854e-05, "loss": 0.8093, "step": 7093 }, { "epoch": 0.2569358927924665, "grad_norm": 1.638746880507246, "learning_rate": 1.7418797546702737e-05, "loss": 0.7649, "step": 7094 }, { "epoch": 0.25697211155378485, "grad_norm": 1.501481068841371, "learning_rate": 1.741801091510945e-05, "loss": 0.8763, "step": 7095 }, { "epoch": 0.25700833031510323, "grad_norm": 1.4575102634870107, "learning_rate": 1.7417224181437807e-05, "loss": 0.8042, "step": 7096 }, { "epoch": 0.2570445490764216, "grad_norm": 1.5116410826906563, "learning_rate": 1.7416437345698648e-05, "loss": 0.9024, "step": 7097 }, { "epoch": 0.25708076783774, "grad_norm": 1.2924114285145705, "learning_rate": 1.741565040790279e-05, "loss": 0.8059, "step": 7098 }, { "epoch": 0.2571169865990583, "grad_norm": 1.555077927892922, "learning_rate": 1.741486336806107e-05, "loss": 0.7796, "step": 7099 }, { "epoch": 0.25715320536037667, "grad_norm": 1.1579956973386785, "learning_rate": 1.7414076226184313e-05, "loss": 0.7884, "step": 7100 }, { "epoch": 0.25718942412169504, "grad_norm": 1.3731389164936998, "learning_rate": 1.7413288982283355e-05, "loss": 0.818, "step": 7101 }, { "epoch": 0.2572256428830134, "grad_norm": 1.163370176987187, "learning_rate": 1.7412501636369025e-05, "loss": 0.7844, "step": 7102 }, { "epoch": 0.2572618616443318, "grad_norm": 1.5344022520212928, "learning_rate": 1.741171418845216e-05, "loss": 0.8385, "step": 7103 }, { "epoch": 0.2572980804056501, "grad_norm": 1.469994610520907, "learning_rate": 1.7410926638543594e-05, "loss": 0.8423, "step": 7104 }, { "epoch": 0.2573342991669685, "grad_norm": 1.6055981544449514, "learning_rate": 1.7410138986654167e-05, "loss": 0.8506, "step": 7105 }, { "epoch": 0.25737051792828686, "grad_norm": 1.4782147271421908, "learning_rate": 1.7409351232794716e-05, "loss": 0.8242, "step": 7106 }, { "epoch": 0.25740673668960523, "grad_norm": 1.5248213962417843, "learning_rate": 1.7408563376976082e-05, "loss": 0.8088, "step": 7107 }, { "epoch": 0.2574429554509236, "grad_norm": 1.471221030943871, "learning_rate": 1.7407775419209106e-05, "loss": 0.8213, "step": 7108 }, { "epoch": 0.2574791742122419, "grad_norm": 1.5070152954732285, "learning_rate": 1.740698735950463e-05, "loss": 0.7864, "step": 7109 }, { "epoch": 0.2575153929735603, "grad_norm": 1.374284165079488, "learning_rate": 1.74061991978735e-05, "loss": 0.7564, "step": 7110 }, { "epoch": 0.25755161173487867, "grad_norm": 1.482092050442367, "learning_rate": 1.7405410934326562e-05, "loss": 0.7949, "step": 7111 }, { "epoch": 0.25758783049619705, "grad_norm": 1.728412126147293, "learning_rate": 1.7404622568874662e-05, "loss": 0.8056, "step": 7112 }, { "epoch": 0.2576240492575154, "grad_norm": 1.4533756857669682, "learning_rate": 1.7403834101528653e-05, "loss": 0.7363, "step": 7113 }, { "epoch": 0.25766026801883374, "grad_norm": 1.4269998620558821, "learning_rate": 1.7403045532299376e-05, "loss": 0.85, "step": 7114 }, { "epoch": 0.2576964867801521, "grad_norm": 1.2639964079789565, "learning_rate": 1.7402256861197687e-05, "loss": 0.7944, "step": 7115 }, { "epoch": 0.2577327055414705, "grad_norm": 1.529828249668037, "learning_rate": 1.7401468088234445e-05, "loss": 0.746, "step": 7116 }, { "epoch": 0.25776892430278886, "grad_norm": 1.3550951948655008, "learning_rate": 1.7400679213420492e-05, "loss": 0.7526, "step": 7117 }, { "epoch": 0.25780514306410723, "grad_norm": 1.3966995940392168, "learning_rate": 1.7399890236766693e-05, "loss": 0.7445, "step": 7118 }, { "epoch": 0.25784136182542555, "grad_norm": 1.3955738527169432, "learning_rate": 1.7399101158283905e-05, "loss": 0.7529, "step": 7119 }, { "epoch": 0.2578775805867439, "grad_norm": 1.715136234783158, "learning_rate": 1.739831197798298e-05, "loss": 0.8956, "step": 7120 }, { "epoch": 0.2579137993480623, "grad_norm": 1.3922197401068448, "learning_rate": 1.7397522695874786e-05, "loss": 0.7875, "step": 7121 }, { "epoch": 0.2579500181093807, "grad_norm": 1.5577296117790793, "learning_rate": 1.7396733311970175e-05, "loss": 0.8577, "step": 7122 }, { "epoch": 0.25798623687069905, "grad_norm": 1.220172450682811, "learning_rate": 1.7395943826280018e-05, "loss": 0.762, "step": 7123 }, { "epoch": 0.25802245563201737, "grad_norm": 1.481364857796772, "learning_rate": 1.739515423881517e-05, "loss": 0.8263, "step": 7124 }, { "epoch": 0.25805867439333574, "grad_norm": 1.7128570209717118, "learning_rate": 1.7394364549586505e-05, "loss": 0.7669, "step": 7125 }, { "epoch": 0.2580948931546541, "grad_norm": 1.1316820750082617, "learning_rate": 1.7393574758604887e-05, "loss": 0.7933, "step": 7126 }, { "epoch": 0.2581311119159725, "grad_norm": 1.773127890650215, "learning_rate": 1.7392784865881182e-05, "loss": 0.9197, "step": 7127 }, { "epoch": 0.25816733067729086, "grad_norm": 1.512296433187939, "learning_rate": 1.7391994871426262e-05, "loss": 0.8082, "step": 7128 }, { "epoch": 0.2582035494386092, "grad_norm": 1.7198999112832343, "learning_rate": 1.7391204775251e-05, "loss": 0.8115, "step": 7129 }, { "epoch": 0.25823976819992756, "grad_norm": 1.2957283540294438, "learning_rate": 1.7390414577366263e-05, "loss": 0.7586, "step": 7130 }, { "epoch": 0.25827598696124593, "grad_norm": 1.4731622605715733, "learning_rate": 1.738962427778293e-05, "loss": 0.8303, "step": 7131 }, { "epoch": 0.2583122057225643, "grad_norm": 1.3376113683690232, "learning_rate": 1.738883387651187e-05, "loss": 0.8262, "step": 7132 }, { "epoch": 0.2583484244838827, "grad_norm": 1.1722533574855913, "learning_rate": 1.7388043373563967e-05, "loss": 0.7794, "step": 7133 }, { "epoch": 0.258384643245201, "grad_norm": 1.0203247469649708, "learning_rate": 1.7387252768950093e-05, "loss": 0.7463, "step": 7134 }, { "epoch": 0.25842086200651937, "grad_norm": 1.5866701531399043, "learning_rate": 1.7386462062681132e-05, "loss": 0.8181, "step": 7135 }, { "epoch": 0.25845708076783774, "grad_norm": 1.8567316555867428, "learning_rate": 1.7385671254767964e-05, "loss": 0.8733, "step": 7136 }, { "epoch": 0.2584932995291561, "grad_norm": 1.51998135316267, "learning_rate": 1.7384880345221466e-05, "loss": 0.8688, "step": 7137 }, { "epoch": 0.2585295182904745, "grad_norm": 1.3605060684766623, "learning_rate": 1.7384089334052527e-05, "loss": 0.782, "step": 7138 }, { "epoch": 0.2585657370517928, "grad_norm": 1.4911432931642774, "learning_rate": 1.738329822127203e-05, "loss": 0.7828, "step": 7139 }, { "epoch": 0.2586019558131112, "grad_norm": 1.3563312078466405, "learning_rate": 1.7382507006890865e-05, "loss": 0.7328, "step": 7140 }, { "epoch": 0.25863817457442956, "grad_norm": 1.5043823888004504, "learning_rate": 1.7381715690919914e-05, "loss": 0.8981, "step": 7141 }, { "epoch": 0.25867439333574793, "grad_norm": 1.5347658660764207, "learning_rate": 1.738092427337007e-05, "loss": 0.8043, "step": 7142 }, { "epoch": 0.2587106120970663, "grad_norm": 1.6488676385679402, "learning_rate": 1.7380132754252222e-05, "loss": 0.8873, "step": 7143 }, { "epoch": 0.2587468308583846, "grad_norm": 1.3984903899638912, "learning_rate": 1.7379341133577262e-05, "loss": 0.6573, "step": 7144 }, { "epoch": 0.258783049619703, "grad_norm": 1.2623243521337357, "learning_rate": 1.7378549411356083e-05, "loss": 0.8511, "step": 7145 }, { "epoch": 0.2588192683810214, "grad_norm": 1.490660097900582, "learning_rate": 1.737775758759958e-05, "loss": 0.8513, "step": 7146 }, { "epoch": 0.25885548714233975, "grad_norm": 1.434220901846713, "learning_rate": 1.7376965662318654e-05, "loss": 0.8172, "step": 7147 }, { "epoch": 0.2588917059036581, "grad_norm": 1.4599044288907332, "learning_rate": 1.7376173635524196e-05, "loss": 0.811, "step": 7148 }, { "epoch": 0.25892792466497644, "grad_norm": 1.487939188914034, "learning_rate": 1.7375381507227108e-05, "loss": 0.8819, "step": 7149 }, { "epoch": 0.2589641434262948, "grad_norm": 1.4075690910842162, "learning_rate": 1.737458927743829e-05, "loss": 0.7904, "step": 7150 }, { "epoch": 0.2590003621876132, "grad_norm": 1.4399906950722843, "learning_rate": 1.7373796946168645e-05, "loss": 0.8103, "step": 7151 }, { "epoch": 0.25903658094893156, "grad_norm": 1.201055459258219, "learning_rate": 1.7373004513429074e-05, "loss": 0.7637, "step": 7152 }, { "epoch": 0.25907279971024993, "grad_norm": 1.3888942336602157, "learning_rate": 1.7372211979230483e-05, "loss": 0.8213, "step": 7153 }, { "epoch": 0.25910901847156825, "grad_norm": 1.5642069249015291, "learning_rate": 1.7371419343583777e-05, "loss": 0.8421, "step": 7154 }, { "epoch": 0.2591452372328866, "grad_norm": 1.4395344800487415, "learning_rate": 1.7370626606499863e-05, "loss": 0.828, "step": 7155 }, { "epoch": 0.259181455994205, "grad_norm": 1.3764872168555948, "learning_rate": 1.7369833767989647e-05, "loss": 0.8912, "step": 7156 }, { "epoch": 0.2592176747555234, "grad_norm": 1.6092382866002897, "learning_rate": 1.7369040828064046e-05, "loss": 0.7894, "step": 7157 }, { "epoch": 0.25925389351684175, "grad_norm": 1.4125556379148354, "learning_rate": 1.736824778673397e-05, "loss": 0.807, "step": 7158 }, { "epoch": 0.25929011227816007, "grad_norm": 1.6354070599136123, "learning_rate": 1.7367454644010327e-05, "loss": 0.9216, "step": 7159 }, { "epoch": 0.25932633103947844, "grad_norm": 1.412716031512619, "learning_rate": 1.736666139990404e-05, "loss": 0.7968, "step": 7160 }, { "epoch": 0.2593625498007968, "grad_norm": 1.3291181427358, "learning_rate": 1.7365868054426012e-05, "loss": 0.789, "step": 7161 }, { "epoch": 0.2593987685621152, "grad_norm": 1.5050217423866534, "learning_rate": 1.736507460758717e-05, "loss": 0.773, "step": 7162 }, { "epoch": 0.25943498732343356, "grad_norm": 1.390664031118567, "learning_rate": 1.7364281059398433e-05, "loss": 0.8306, "step": 7163 }, { "epoch": 0.2594712060847519, "grad_norm": 1.5158108168428555, "learning_rate": 1.7363487409870714e-05, "loss": 0.8486, "step": 7164 }, { "epoch": 0.25950742484607026, "grad_norm": 1.1718551830761952, "learning_rate": 1.736269365901494e-05, "loss": 0.8782, "step": 7165 }, { "epoch": 0.25954364360738863, "grad_norm": 1.3371424519960804, "learning_rate": 1.7361899806842028e-05, "loss": 0.7828, "step": 7166 }, { "epoch": 0.259579862368707, "grad_norm": 1.5417134284861973, "learning_rate": 1.7361105853362912e-05, "loss": 0.833, "step": 7167 }, { "epoch": 0.2596160811300254, "grad_norm": 1.3767184499278886, "learning_rate": 1.7360311798588508e-05, "loss": 0.7627, "step": 7168 }, { "epoch": 0.2596522998913437, "grad_norm": 1.5510794871686973, "learning_rate": 1.735951764252975e-05, "loss": 0.8505, "step": 7169 }, { "epoch": 0.25968851865266207, "grad_norm": 1.3577896244363077, "learning_rate": 1.7358723385197557e-05, "loss": 0.758, "step": 7170 }, { "epoch": 0.25972473741398044, "grad_norm": 1.2501870526675762, "learning_rate": 1.7357929026602867e-05, "loss": 0.8363, "step": 7171 }, { "epoch": 0.2597609561752988, "grad_norm": 1.1550331986985696, "learning_rate": 1.7357134566756606e-05, "loss": 0.8179, "step": 7172 }, { "epoch": 0.2597971749366172, "grad_norm": 1.745664665445539, "learning_rate": 1.735634000566971e-05, "loss": 0.9353, "step": 7173 }, { "epoch": 0.2598333936979355, "grad_norm": 1.5173418801333183, "learning_rate": 1.7355545343353116e-05, "loss": 0.7993, "step": 7174 }, { "epoch": 0.2598696124592539, "grad_norm": 1.5708838936799914, "learning_rate": 1.735475057981775e-05, "loss": 0.8901, "step": 7175 }, { "epoch": 0.25990583122057226, "grad_norm": 1.441792174755115, "learning_rate": 1.7353955715074556e-05, "loss": 0.8766, "step": 7176 }, { "epoch": 0.25994204998189063, "grad_norm": 1.3430993402498719, "learning_rate": 1.7353160749134467e-05, "loss": 0.7907, "step": 7177 }, { "epoch": 0.259978268743209, "grad_norm": 1.4834675311332886, "learning_rate": 1.7352365682008426e-05, "loss": 0.8035, "step": 7178 }, { "epoch": 0.2600144875045273, "grad_norm": 1.2048284301030723, "learning_rate": 1.7351570513707376e-05, "loss": 0.7946, "step": 7179 }, { "epoch": 0.2600507062658457, "grad_norm": 1.4121995312783266, "learning_rate": 1.7350775244242253e-05, "loss": 0.8536, "step": 7180 }, { "epoch": 0.2600869250271641, "grad_norm": 1.5856165138283684, "learning_rate": 1.7349979873624002e-05, "loss": 0.8695, "step": 7181 }, { "epoch": 0.26012314378848245, "grad_norm": 1.641908985605918, "learning_rate": 1.7349184401863573e-05, "loss": 0.8385, "step": 7182 }, { "epoch": 0.2601593625498008, "grad_norm": 1.5116262567795349, "learning_rate": 1.734838882897191e-05, "loss": 0.8349, "step": 7183 }, { "epoch": 0.26019558131111914, "grad_norm": 1.5155763765774497, "learning_rate": 1.7347593154959955e-05, "loss": 0.7948, "step": 7184 }, { "epoch": 0.2602318000724375, "grad_norm": 1.1912440334168894, "learning_rate": 1.7346797379838667e-05, "loss": 0.7545, "step": 7185 }, { "epoch": 0.2602680188337559, "grad_norm": 1.3540921992373416, "learning_rate": 1.734600150361899e-05, "loss": 0.8189, "step": 7186 }, { "epoch": 0.26030423759507426, "grad_norm": 1.729979205797269, "learning_rate": 1.7345205526311875e-05, "loss": 0.852, "step": 7187 }, { "epoch": 0.26034045635639264, "grad_norm": 1.4601648653938597, "learning_rate": 1.7344409447928282e-05, "loss": 0.8316, "step": 7188 }, { "epoch": 0.26037667511771095, "grad_norm": 1.1800492102017357, "learning_rate": 1.734361326847916e-05, "loss": 0.7899, "step": 7189 }, { "epoch": 0.26041289387902933, "grad_norm": 1.473191057936541, "learning_rate": 1.7342816987975465e-05, "loss": 0.7785, "step": 7190 }, { "epoch": 0.2604491126403477, "grad_norm": 1.364245519807957, "learning_rate": 1.734202060642816e-05, "loss": 0.7995, "step": 7191 }, { "epoch": 0.2604853314016661, "grad_norm": 1.5832041042747453, "learning_rate": 1.7341224123848196e-05, "loss": 0.7617, "step": 7192 }, { "epoch": 0.26052155016298445, "grad_norm": 1.4754178249951495, "learning_rate": 1.734042754024654e-05, "loss": 0.8623, "step": 7193 }, { "epoch": 0.26055776892430277, "grad_norm": 1.4160089466332109, "learning_rate": 1.733963085563415e-05, "loss": 0.7968, "step": 7194 }, { "epoch": 0.26059398768562114, "grad_norm": 1.247020155933117, "learning_rate": 1.733883407002199e-05, "loss": 0.8434, "step": 7195 }, { "epoch": 0.2606302064469395, "grad_norm": 1.5810046974464618, "learning_rate": 1.7338037183421028e-05, "loss": 0.7651, "step": 7196 }, { "epoch": 0.2606664252082579, "grad_norm": 1.3381191232564256, "learning_rate": 1.7337240195842224e-05, "loss": 0.7602, "step": 7197 }, { "epoch": 0.26070264396957626, "grad_norm": 1.6757984838510984, "learning_rate": 1.7336443107296547e-05, "loss": 0.811, "step": 7198 }, { "epoch": 0.2607388627308946, "grad_norm": 1.4126948393413903, "learning_rate": 1.7335645917794968e-05, "loss": 0.7053, "step": 7199 }, { "epoch": 0.26077508149221296, "grad_norm": 1.38033358452446, "learning_rate": 1.7334848627348454e-05, "loss": 0.8199, "step": 7200 }, { "epoch": 0.26081130025353133, "grad_norm": 1.1750747229803817, "learning_rate": 1.733405123596798e-05, "loss": 0.8729, "step": 7201 }, { "epoch": 0.2608475190148497, "grad_norm": 1.5824459968673856, "learning_rate": 1.7333253743664516e-05, "loss": 0.7726, "step": 7202 }, { "epoch": 0.2608837377761681, "grad_norm": 1.492285952390912, "learning_rate": 1.7332456150449037e-05, "loss": 0.77, "step": 7203 }, { "epoch": 0.2609199565374864, "grad_norm": 1.4688080005150743, "learning_rate": 1.7331658456332516e-05, "loss": 0.7882, "step": 7204 }, { "epoch": 0.26095617529880477, "grad_norm": 1.2778356186383113, "learning_rate": 1.7330860661325935e-05, "loss": 0.8302, "step": 7205 }, { "epoch": 0.26099239406012315, "grad_norm": 1.6003585738896489, "learning_rate": 1.733006276544027e-05, "loss": 0.8556, "step": 7206 }, { "epoch": 0.2610286128214415, "grad_norm": 1.5419010802225397, "learning_rate": 1.73292647686865e-05, "loss": 0.7408, "step": 7207 }, { "epoch": 0.2610648315827599, "grad_norm": 1.5084091399525759, "learning_rate": 1.7328466671075607e-05, "loss": 0.8158, "step": 7208 }, { "epoch": 0.2611010503440782, "grad_norm": 1.2571343348920856, "learning_rate": 1.732766847261857e-05, "loss": 0.7419, "step": 7209 }, { "epoch": 0.2611372691053966, "grad_norm": 1.374942567693339, "learning_rate": 1.732687017332638e-05, "loss": 0.8339, "step": 7210 }, { "epoch": 0.26117348786671496, "grad_norm": 1.6133118862078815, "learning_rate": 1.732607177321002e-05, "loss": 0.9535, "step": 7211 }, { "epoch": 0.26120970662803333, "grad_norm": 1.5361693311838132, "learning_rate": 1.7325273272280473e-05, "loss": 0.8029, "step": 7212 }, { "epoch": 0.2612459253893517, "grad_norm": 1.4300994474983009, "learning_rate": 1.732447467054873e-05, "loss": 0.8457, "step": 7213 }, { "epoch": 0.26128214415067, "grad_norm": 1.4025383059881158, "learning_rate": 1.7323675968025776e-05, "loss": 0.785, "step": 7214 }, { "epoch": 0.2613183629119884, "grad_norm": 1.1137552576099512, "learning_rate": 1.7322877164722606e-05, "loss": 0.7546, "step": 7215 }, { "epoch": 0.2613545816733068, "grad_norm": 1.1895455777046868, "learning_rate": 1.7322078260650216e-05, "loss": 0.8067, "step": 7216 }, { "epoch": 0.26139080043462515, "grad_norm": 1.4358329280346067, "learning_rate": 1.7321279255819595e-05, "loss": 0.7979, "step": 7217 }, { "epoch": 0.2614270191959435, "grad_norm": 1.4585662307068963, "learning_rate": 1.7320480150241734e-05, "loss": 0.7881, "step": 7218 }, { "epoch": 0.26146323795726184, "grad_norm": 1.1740306436169676, "learning_rate": 1.7319680943927636e-05, "loss": 0.7406, "step": 7219 }, { "epoch": 0.2614994567185802, "grad_norm": 1.5047956041617885, "learning_rate": 1.7318881636888297e-05, "loss": 0.7884, "step": 7220 }, { "epoch": 0.2615356754798986, "grad_norm": 1.517576957761102, "learning_rate": 1.7318082229134718e-05, "loss": 0.798, "step": 7221 }, { "epoch": 0.26157189424121696, "grad_norm": 1.2766347216565421, "learning_rate": 1.7317282720677893e-05, "loss": 0.7764, "step": 7222 }, { "epoch": 0.26160811300253534, "grad_norm": 1.4272846454238537, "learning_rate": 1.731648311152883e-05, "loss": 0.7713, "step": 7223 }, { "epoch": 0.26164433176385365, "grad_norm": 1.367795592701522, "learning_rate": 1.7315683401698532e-05, "loss": 0.8703, "step": 7224 }, { "epoch": 0.26168055052517203, "grad_norm": 1.2217014821179735, "learning_rate": 1.7314883591198005e-05, "loss": 0.7484, "step": 7225 }, { "epoch": 0.2617167692864904, "grad_norm": 1.296661255928574, "learning_rate": 1.7314083680038244e-05, "loss": 0.7258, "step": 7226 }, { "epoch": 0.2617529880478088, "grad_norm": 1.6431488945987665, "learning_rate": 1.731328366823027e-05, "loss": 0.7675, "step": 7227 }, { "epoch": 0.26178920680912715, "grad_norm": 1.999874411864331, "learning_rate": 1.7312483555785087e-05, "loss": 0.8516, "step": 7228 }, { "epoch": 0.26182542557044547, "grad_norm": 1.0982220095724975, "learning_rate": 1.7311683342713707e-05, "loss": 0.8393, "step": 7229 }, { "epoch": 0.26186164433176384, "grad_norm": 1.5763637734344154, "learning_rate": 1.7310883029027135e-05, "loss": 0.766, "step": 7230 }, { "epoch": 0.2618978630930822, "grad_norm": 1.3192055240330922, "learning_rate": 1.731008261473639e-05, "loss": 0.7473, "step": 7231 }, { "epoch": 0.2619340818544006, "grad_norm": 1.8415576136207192, "learning_rate": 1.7309282099852484e-05, "loss": 0.7976, "step": 7232 }, { "epoch": 0.26197030061571897, "grad_norm": 1.5663177163178137, "learning_rate": 1.7308481484386436e-05, "loss": 0.8682, "step": 7233 }, { "epoch": 0.2620065193770373, "grad_norm": 1.4125559191136978, "learning_rate": 1.730768076834926e-05, "loss": 0.8267, "step": 7234 }, { "epoch": 0.26204273813835566, "grad_norm": 1.8249234083226296, "learning_rate": 1.7306879951751977e-05, "loss": 0.8011, "step": 7235 }, { "epoch": 0.26207895689967403, "grad_norm": 1.0558364486569185, "learning_rate": 1.73060790346056e-05, "loss": 0.7259, "step": 7236 }, { "epoch": 0.2621151756609924, "grad_norm": 1.5467201712939287, "learning_rate": 1.7305278016921162e-05, "loss": 0.8045, "step": 7237 }, { "epoch": 0.2621513944223108, "grad_norm": 1.1394509049403934, "learning_rate": 1.730447689870968e-05, "loss": 0.8337, "step": 7238 }, { "epoch": 0.2621876131836291, "grad_norm": 1.3564374376264772, "learning_rate": 1.7303675679982173e-05, "loss": 0.7801, "step": 7239 }, { "epoch": 0.26222383194494747, "grad_norm": 1.5052904044917923, "learning_rate": 1.7302874360749674e-05, "loss": 0.7799, "step": 7240 }, { "epoch": 0.26226005070626585, "grad_norm": 1.4267012032016912, "learning_rate": 1.7302072941023204e-05, "loss": 0.7613, "step": 7241 }, { "epoch": 0.2622962694675842, "grad_norm": 1.4657516234668835, "learning_rate": 1.7301271420813795e-05, "loss": 0.8218, "step": 7242 }, { "epoch": 0.2623324882289026, "grad_norm": 1.4874435530298584, "learning_rate": 1.730046980013248e-05, "loss": 0.8252, "step": 7243 }, { "epoch": 0.2623687069902209, "grad_norm": 1.4187880413799225, "learning_rate": 1.7299668078990283e-05, "loss": 0.821, "step": 7244 }, { "epoch": 0.2624049257515393, "grad_norm": 1.454606648681313, "learning_rate": 1.7298866257398237e-05, "loss": 0.8186, "step": 7245 }, { "epoch": 0.26244114451285766, "grad_norm": 1.4472695400787403, "learning_rate": 1.7298064335367382e-05, "loss": 0.7428, "step": 7246 }, { "epoch": 0.26247736327417603, "grad_norm": 1.5518788559975576, "learning_rate": 1.7297262312908747e-05, "loss": 0.7995, "step": 7247 }, { "epoch": 0.2625135820354944, "grad_norm": 1.4695056153555508, "learning_rate": 1.729646019003337e-05, "loss": 0.7611, "step": 7248 }, { "epoch": 0.2625498007968127, "grad_norm": 1.4214980462840683, "learning_rate": 1.729565796675229e-05, "loss": 0.7894, "step": 7249 }, { "epoch": 0.2625860195581311, "grad_norm": 1.4631819619558768, "learning_rate": 1.7294855643076548e-05, "loss": 0.7913, "step": 7250 }, { "epoch": 0.2626222383194495, "grad_norm": 1.4001152208025642, "learning_rate": 1.729405321901718e-05, "loss": 0.8021, "step": 7251 }, { "epoch": 0.26265845708076785, "grad_norm": 1.5395444947902917, "learning_rate": 1.729325069458523e-05, "loss": 0.7936, "step": 7252 }, { "epoch": 0.2626946758420862, "grad_norm": 1.561315241109846, "learning_rate": 1.7292448069791742e-05, "loss": 0.9051, "step": 7253 }, { "epoch": 0.26273089460340454, "grad_norm": 1.2146101054020828, "learning_rate": 1.7291645344647764e-05, "loss": 0.7465, "step": 7254 }, { "epoch": 0.2627671133647229, "grad_norm": 1.344893413988213, "learning_rate": 1.7290842519164334e-05, "loss": 0.7329, "step": 7255 }, { "epoch": 0.2628033321260413, "grad_norm": 1.5279831828286026, "learning_rate": 1.729003959335251e-05, "loss": 0.8452, "step": 7256 }, { "epoch": 0.26283955088735966, "grad_norm": 1.4086636565332955, "learning_rate": 1.7289236567223337e-05, "loss": 0.7628, "step": 7257 }, { "epoch": 0.26287576964867804, "grad_norm": 1.2078048254882066, "learning_rate": 1.7288433440787857e-05, "loss": 0.817, "step": 7258 }, { "epoch": 0.26291198840999636, "grad_norm": 1.5441898342931226, "learning_rate": 1.728763021405713e-05, "loss": 0.8691, "step": 7259 }, { "epoch": 0.26294820717131473, "grad_norm": 1.3897585349390214, "learning_rate": 1.7286826887042213e-05, "loss": 0.8793, "step": 7260 }, { "epoch": 0.2629844259326331, "grad_norm": 1.437714085238051, "learning_rate": 1.7286023459754152e-05, "loss": 0.8161, "step": 7261 }, { "epoch": 0.2630206446939515, "grad_norm": 1.7170650415951232, "learning_rate": 1.7285219932204004e-05, "loss": 0.8039, "step": 7262 }, { "epoch": 0.26305686345526985, "grad_norm": 1.4521863287245196, "learning_rate": 1.7284416304402832e-05, "loss": 0.8419, "step": 7263 }, { "epoch": 0.26309308221658817, "grad_norm": 1.3891158981848786, "learning_rate": 1.728361257636169e-05, "loss": 0.7594, "step": 7264 }, { "epoch": 0.26312930097790654, "grad_norm": 1.2048450229838656, "learning_rate": 1.7282808748091637e-05, "loss": 0.7442, "step": 7265 }, { "epoch": 0.2631655197392249, "grad_norm": 1.5287519071018116, "learning_rate": 1.7282004819603736e-05, "loss": 0.8338, "step": 7266 }, { "epoch": 0.2632017385005433, "grad_norm": 1.2911864754138744, "learning_rate": 1.7281200790909056e-05, "loss": 0.7901, "step": 7267 }, { "epoch": 0.26323795726186167, "grad_norm": 1.4655658838964918, "learning_rate": 1.728039666201865e-05, "loss": 0.8288, "step": 7268 }, { "epoch": 0.26327417602318, "grad_norm": 1.5792150663066993, "learning_rate": 1.727959243294359e-05, "loss": 0.832, "step": 7269 }, { "epoch": 0.26331039478449836, "grad_norm": 1.4397939591639521, "learning_rate": 1.7278788103694944e-05, "loss": 0.8206, "step": 7270 }, { "epoch": 0.26334661354581673, "grad_norm": 1.4944389737272987, "learning_rate": 1.7277983674283776e-05, "loss": 0.8984, "step": 7271 }, { "epoch": 0.2633828323071351, "grad_norm": 1.2670430801376547, "learning_rate": 1.7277179144721156e-05, "loss": 0.786, "step": 7272 }, { "epoch": 0.2634190510684535, "grad_norm": 1.381437977615358, "learning_rate": 1.7276374515018158e-05, "loss": 0.7821, "step": 7273 }, { "epoch": 0.2634552698297718, "grad_norm": 1.458840698868692, "learning_rate": 1.7275569785185854e-05, "loss": 0.766, "step": 7274 }, { "epoch": 0.2634914885910902, "grad_norm": 1.3548966947209764, "learning_rate": 1.7274764955235315e-05, "loss": 0.7874, "step": 7275 }, { "epoch": 0.26352770735240855, "grad_norm": 1.4219386574000166, "learning_rate": 1.727396002517762e-05, "loss": 0.7685, "step": 7276 }, { "epoch": 0.2635639261137269, "grad_norm": 1.0662849526710962, "learning_rate": 1.7273154995023844e-05, "loss": 0.7463, "step": 7277 }, { "epoch": 0.2636001448750453, "grad_norm": 1.4344756671022894, "learning_rate": 1.7272349864785062e-05, "loss": 0.8184, "step": 7278 }, { "epoch": 0.2636363636363636, "grad_norm": 1.3982226190691036, "learning_rate": 1.7271544634472356e-05, "loss": 0.7938, "step": 7279 }, { "epoch": 0.263672582397682, "grad_norm": 1.2459102811043363, "learning_rate": 1.727073930409681e-05, "loss": 0.7944, "step": 7280 }, { "epoch": 0.26370880115900036, "grad_norm": 1.4929454398918036, "learning_rate": 1.72699338736695e-05, "loss": 0.9175, "step": 7281 }, { "epoch": 0.26374501992031874, "grad_norm": 1.6128356834388842, "learning_rate": 1.726912834320151e-05, "loss": 0.8062, "step": 7282 }, { "epoch": 0.2637812386816371, "grad_norm": 1.4240780403050208, "learning_rate": 1.7268322712703932e-05, "loss": 0.8094, "step": 7283 }, { "epoch": 0.2638174574429554, "grad_norm": 1.4712575042390623, "learning_rate": 1.7267516982187843e-05, "loss": 0.7841, "step": 7284 }, { "epoch": 0.2638536762042738, "grad_norm": 1.3653696208756396, "learning_rate": 1.7266711151664334e-05, "loss": 0.7792, "step": 7285 }, { "epoch": 0.2638898949655922, "grad_norm": 1.3353524530420908, "learning_rate": 1.7265905221144497e-05, "loss": 0.8198, "step": 7286 }, { "epoch": 0.26392611372691055, "grad_norm": 1.353569097136998, "learning_rate": 1.7265099190639417e-05, "loss": 0.7678, "step": 7287 }, { "epoch": 0.2639623324882289, "grad_norm": 1.5939331159860082, "learning_rate": 1.7264293060160194e-05, "loss": 0.7561, "step": 7288 }, { "epoch": 0.26399855124954724, "grad_norm": 1.573879857419059, "learning_rate": 1.726348682971791e-05, "loss": 0.8282, "step": 7289 }, { "epoch": 0.2640347700108656, "grad_norm": 1.3737982054948505, "learning_rate": 1.7262680499323664e-05, "loss": 0.7551, "step": 7290 }, { "epoch": 0.264070988772184, "grad_norm": 1.5365635109856015, "learning_rate": 1.7261874068988557e-05, "loss": 0.7693, "step": 7291 }, { "epoch": 0.26410720753350236, "grad_norm": 1.450843185947432, "learning_rate": 1.726106753872368e-05, "loss": 0.8267, "step": 7292 }, { "epoch": 0.26414342629482074, "grad_norm": 1.2784117706204818, "learning_rate": 1.7260260908540135e-05, "loss": 0.7939, "step": 7293 }, { "epoch": 0.26417964505613906, "grad_norm": 1.3457692300314419, "learning_rate": 1.725945417844902e-05, "loss": 0.7771, "step": 7294 }, { "epoch": 0.26421586381745743, "grad_norm": 1.3388651911124518, "learning_rate": 1.7258647348461436e-05, "loss": 0.7848, "step": 7295 }, { "epoch": 0.2642520825787758, "grad_norm": 1.504533612075807, "learning_rate": 1.7257840418588486e-05, "loss": 0.768, "step": 7296 }, { "epoch": 0.2642883013400942, "grad_norm": 1.3637529016021028, "learning_rate": 1.7257033388841277e-05, "loss": 0.8432, "step": 7297 }, { "epoch": 0.26432452010141255, "grad_norm": 1.562744008323355, "learning_rate": 1.7256226259230907e-05, "loss": 0.8574, "step": 7298 }, { "epoch": 0.26436073886273087, "grad_norm": 1.5391332735935845, "learning_rate": 1.725541902976849e-05, "loss": 0.7414, "step": 7299 }, { "epoch": 0.26439695762404924, "grad_norm": 1.3207451813959397, "learning_rate": 1.7254611700465137e-05, "loss": 0.7716, "step": 7300 }, { "epoch": 0.2644331763853676, "grad_norm": 1.5154939550594178, "learning_rate": 1.7253804271331946e-05, "loss": 0.8546, "step": 7301 }, { "epoch": 0.264469395146686, "grad_norm": 1.5075812421858306, "learning_rate": 1.725299674238004e-05, "loss": 0.7843, "step": 7302 }, { "epoch": 0.26450561390800437, "grad_norm": 1.2334697556282934, "learning_rate": 1.7252189113620522e-05, "loss": 0.83, "step": 7303 }, { "epoch": 0.2645418326693227, "grad_norm": 1.4707231304940847, "learning_rate": 1.725138138506451e-05, "loss": 0.8338, "step": 7304 }, { "epoch": 0.26457805143064106, "grad_norm": 1.011373264949875, "learning_rate": 1.7250573556723118e-05, "loss": 0.7868, "step": 7305 }, { "epoch": 0.26461427019195943, "grad_norm": 1.0688440750096055, "learning_rate": 1.7249765628607463e-05, "loss": 0.7913, "step": 7306 }, { "epoch": 0.2646504889532778, "grad_norm": 1.2415514493162905, "learning_rate": 1.7248957600728664e-05, "loss": 0.6923, "step": 7307 }, { "epoch": 0.2646867077145962, "grad_norm": 1.5015102702146284, "learning_rate": 1.724814947309784e-05, "loss": 0.8063, "step": 7308 }, { "epoch": 0.2647229264759145, "grad_norm": 1.410552118535148, "learning_rate": 1.7247341245726107e-05, "loss": 0.766, "step": 7309 }, { "epoch": 0.2647591452372329, "grad_norm": 1.2443157705115184, "learning_rate": 1.724653291862459e-05, "loss": 0.8651, "step": 7310 }, { "epoch": 0.26479536399855125, "grad_norm": 1.1886478142398262, "learning_rate": 1.7245724491804418e-05, "loss": 0.798, "step": 7311 }, { "epoch": 0.2648315827598696, "grad_norm": 1.0816923889951082, "learning_rate": 1.7244915965276706e-05, "loss": 0.8053, "step": 7312 }, { "epoch": 0.264867801521188, "grad_norm": 1.0365210627278318, "learning_rate": 1.7244107339052587e-05, "loss": 0.7479, "step": 7313 }, { "epoch": 0.2649040202825063, "grad_norm": 1.3880326581138747, "learning_rate": 1.7243298613143182e-05, "loss": 0.7736, "step": 7314 }, { "epoch": 0.2649402390438247, "grad_norm": 1.4012400586086526, "learning_rate": 1.724248978755963e-05, "loss": 0.8388, "step": 7315 }, { "epoch": 0.26497645780514306, "grad_norm": 1.6418598568671412, "learning_rate": 1.724168086231305e-05, "loss": 0.8114, "step": 7316 }, { "epoch": 0.26501267656646144, "grad_norm": 1.3857279333896373, "learning_rate": 1.7240871837414577e-05, "loss": 0.7837, "step": 7317 }, { "epoch": 0.2650488953277798, "grad_norm": 1.4078571725292568, "learning_rate": 1.7240062712875346e-05, "loss": 0.7879, "step": 7318 }, { "epoch": 0.26508511408909813, "grad_norm": 1.05860431635517, "learning_rate": 1.723925348870649e-05, "loss": 0.8108, "step": 7319 }, { "epoch": 0.2651213328504165, "grad_norm": 1.3763137489523396, "learning_rate": 1.723844416491915e-05, "loss": 0.7685, "step": 7320 }, { "epoch": 0.2651575516117349, "grad_norm": 1.6641711906970307, "learning_rate": 1.7237634741524455e-05, "loss": 0.7805, "step": 7321 }, { "epoch": 0.26519377037305325, "grad_norm": 1.4193297539446772, "learning_rate": 1.7236825218533547e-05, "loss": 0.7872, "step": 7322 }, { "epoch": 0.2652299891343716, "grad_norm": 1.527405118075248, "learning_rate": 1.7236015595957562e-05, "loss": 0.8075, "step": 7323 }, { "epoch": 0.26526620789568994, "grad_norm": 1.0849946872065508, "learning_rate": 1.7235205873807646e-05, "loss": 0.7695, "step": 7324 }, { "epoch": 0.2653024266570083, "grad_norm": 1.3516984206965983, "learning_rate": 1.7234396052094943e-05, "loss": 0.7143, "step": 7325 }, { "epoch": 0.2653386454183267, "grad_norm": 1.405162268079267, "learning_rate": 1.7233586130830587e-05, "loss": 0.7728, "step": 7326 }, { "epoch": 0.26537486417964506, "grad_norm": 1.5237613168772233, "learning_rate": 1.7232776110025734e-05, "loss": 0.8719, "step": 7327 }, { "epoch": 0.26541108294096344, "grad_norm": 1.1991968694595008, "learning_rate": 1.723196598969153e-05, "loss": 0.7734, "step": 7328 }, { "epoch": 0.26544730170228176, "grad_norm": 2.2343876411507213, "learning_rate": 1.7231155769839113e-05, "loss": 0.8518, "step": 7329 }, { "epoch": 0.26548352046360013, "grad_norm": 1.8208940021264879, "learning_rate": 1.723034545047964e-05, "loss": 0.8568, "step": 7330 }, { "epoch": 0.2655197392249185, "grad_norm": 1.4398106677031737, "learning_rate": 1.7229535031624264e-05, "loss": 0.8116, "step": 7331 }, { "epoch": 0.2655559579862369, "grad_norm": 1.7165016966453577, "learning_rate": 1.7228724513284133e-05, "loss": 0.859, "step": 7332 }, { "epoch": 0.26559217674755525, "grad_norm": 1.4346655235609846, "learning_rate": 1.7227913895470395e-05, "loss": 0.7531, "step": 7333 }, { "epoch": 0.26562839550887357, "grad_norm": 1.2875270809698092, "learning_rate": 1.722710317819422e-05, "loss": 0.7953, "step": 7334 }, { "epoch": 0.26566461427019195, "grad_norm": 1.1796082942153523, "learning_rate": 1.722629236146675e-05, "loss": 0.777, "step": 7335 }, { "epoch": 0.2657008330315103, "grad_norm": 1.4993789238916169, "learning_rate": 1.7225481445299143e-05, "loss": 0.8258, "step": 7336 }, { "epoch": 0.2657370517928287, "grad_norm": 1.3591301490870231, "learning_rate": 1.7224670429702568e-05, "loss": 0.7186, "step": 7337 }, { "epoch": 0.26577327055414707, "grad_norm": 1.559266201076711, "learning_rate": 1.7223859314688177e-05, "loss": 0.7817, "step": 7338 }, { "epoch": 0.2658094893154654, "grad_norm": 1.0438893501003692, "learning_rate": 1.7223048100267138e-05, "loss": 0.7371, "step": 7339 }, { "epoch": 0.26584570807678376, "grad_norm": 1.6570197523168704, "learning_rate": 1.7222236786450607e-05, "loss": 0.7816, "step": 7340 }, { "epoch": 0.26588192683810213, "grad_norm": 1.5204884258987035, "learning_rate": 1.7221425373249757e-05, "loss": 0.7739, "step": 7341 }, { "epoch": 0.2659181455994205, "grad_norm": 1.1015716658669126, "learning_rate": 1.7220613860675742e-05, "loss": 0.7288, "step": 7342 }, { "epoch": 0.2659543643607389, "grad_norm": 1.8691371391526965, "learning_rate": 1.7219802248739742e-05, "loss": 0.78, "step": 7343 }, { "epoch": 0.2659905831220572, "grad_norm": 1.4767993807633926, "learning_rate": 1.7218990537452913e-05, "loss": 0.842, "step": 7344 }, { "epoch": 0.2660268018833756, "grad_norm": 1.4820902017361932, "learning_rate": 1.7218178726826433e-05, "loss": 0.8402, "step": 7345 }, { "epoch": 0.26606302064469395, "grad_norm": 1.6019624978167788, "learning_rate": 1.7217366816871472e-05, "loss": 0.8067, "step": 7346 }, { "epoch": 0.2660992394060123, "grad_norm": 1.1948068359897077, "learning_rate": 1.72165548075992e-05, "loss": 0.778, "step": 7347 }, { "epoch": 0.2661354581673307, "grad_norm": 1.0416049472336109, "learning_rate": 1.7215742699020797e-05, "loss": 0.72, "step": 7348 }, { "epoch": 0.266171676928649, "grad_norm": 1.5958630331125143, "learning_rate": 1.7214930491147432e-05, "loss": 0.8289, "step": 7349 }, { "epoch": 0.2662078956899674, "grad_norm": 1.6135845961338335, "learning_rate": 1.7214118183990285e-05, "loss": 0.8668, "step": 7350 }, { "epoch": 0.26624411445128576, "grad_norm": 1.516369660057415, "learning_rate": 1.7213305777560532e-05, "loss": 0.7688, "step": 7351 }, { "epoch": 0.26628033321260414, "grad_norm": 1.7027043547131282, "learning_rate": 1.7212493271869352e-05, "loss": 0.9167, "step": 7352 }, { "epoch": 0.2663165519739225, "grad_norm": 1.4540268716786184, "learning_rate": 1.7211680666927927e-05, "loss": 0.7935, "step": 7353 }, { "epoch": 0.26635277073524083, "grad_norm": 1.523372943439584, "learning_rate": 1.721086796274744e-05, "loss": 0.7526, "step": 7354 }, { "epoch": 0.2663889894965592, "grad_norm": 1.1527370834270292, "learning_rate": 1.7210055159339078e-05, "loss": 0.7575, "step": 7355 }, { "epoch": 0.2664252082578776, "grad_norm": 1.2187407524571117, "learning_rate": 1.7209242256714015e-05, "loss": 0.763, "step": 7356 }, { "epoch": 0.26646142701919595, "grad_norm": 1.5132724499532741, "learning_rate": 1.720842925488345e-05, "loss": 0.8283, "step": 7357 }, { "epoch": 0.2664976457805143, "grad_norm": 0.9833365938127561, "learning_rate": 1.720761615385856e-05, "loss": 0.6931, "step": 7358 }, { "epoch": 0.26653386454183264, "grad_norm": 1.5247388783254276, "learning_rate": 1.720680295365054e-05, "loss": 0.7981, "step": 7359 }, { "epoch": 0.266570083303151, "grad_norm": 1.1045635762213972, "learning_rate": 1.7205989654270577e-05, "loss": 0.8201, "step": 7360 }, { "epoch": 0.2666063020644694, "grad_norm": 1.2336799647812944, "learning_rate": 1.720517625572987e-05, "loss": 0.7892, "step": 7361 }, { "epoch": 0.26664252082578777, "grad_norm": 1.3315500266906102, "learning_rate": 1.72043627580396e-05, "loss": 0.7631, "step": 7362 }, { "epoch": 0.26667873958710614, "grad_norm": 1.4950850587329148, "learning_rate": 1.7203549161210973e-05, "loss": 0.8215, "step": 7363 }, { "epoch": 0.26671495834842446, "grad_norm": 1.2771117315189233, "learning_rate": 1.7202735465255177e-05, "loss": 0.7419, "step": 7364 }, { "epoch": 0.26675117710974283, "grad_norm": 1.7028147519860348, "learning_rate": 1.7201921670183414e-05, "loss": 0.8235, "step": 7365 }, { "epoch": 0.2667873958710612, "grad_norm": 1.3804711884788474, "learning_rate": 1.720110777600688e-05, "loss": 0.7885, "step": 7366 }, { "epoch": 0.2668236146323796, "grad_norm": 1.1117272723466738, "learning_rate": 1.7200293782736776e-05, "loss": 0.704, "step": 7367 }, { "epoch": 0.26685983339369795, "grad_norm": 1.5274669434116415, "learning_rate": 1.7199479690384303e-05, "loss": 0.8554, "step": 7368 }, { "epoch": 0.2668960521550163, "grad_norm": 1.0853072323257342, "learning_rate": 1.7198665498960664e-05, "loss": 0.735, "step": 7369 }, { "epoch": 0.26693227091633465, "grad_norm": 1.4086357866013959, "learning_rate": 1.719785120847706e-05, "loss": 0.7283, "step": 7370 }, { "epoch": 0.266968489677653, "grad_norm": 1.1661776275570188, "learning_rate": 1.71970368189447e-05, "loss": 0.7867, "step": 7371 }, { "epoch": 0.2670047084389714, "grad_norm": 1.5395260514566764, "learning_rate": 1.719622233037479e-05, "loss": 0.808, "step": 7372 }, { "epoch": 0.26704092720028977, "grad_norm": 1.2957772457534895, "learning_rate": 1.7195407742778537e-05, "loss": 0.8181, "step": 7373 }, { "epoch": 0.2670771459616081, "grad_norm": 1.40907440003704, "learning_rate": 1.719459305616715e-05, "loss": 0.8016, "step": 7374 }, { "epoch": 0.26711336472292646, "grad_norm": 1.2243646820900034, "learning_rate": 1.7193778270551845e-05, "loss": 0.778, "step": 7375 }, { "epoch": 0.26714958348424483, "grad_norm": 1.3970803226313384, "learning_rate": 1.7192963385943825e-05, "loss": 0.8073, "step": 7376 }, { "epoch": 0.2671858022455632, "grad_norm": 1.4247007881177849, "learning_rate": 1.7192148402354313e-05, "loss": 0.7602, "step": 7377 }, { "epoch": 0.2672220210068816, "grad_norm": 1.4773302630369176, "learning_rate": 1.7191333319794514e-05, "loss": 0.8023, "step": 7378 }, { "epoch": 0.2672582397681999, "grad_norm": 1.6974301801657339, "learning_rate": 1.7190518138275652e-05, "loss": 0.891, "step": 7379 }, { "epoch": 0.2672944585295183, "grad_norm": 1.224103632911967, "learning_rate": 1.7189702857808946e-05, "loss": 0.8469, "step": 7380 }, { "epoch": 0.26733067729083665, "grad_norm": 1.1238261403065282, "learning_rate": 1.718888747840561e-05, "loss": 0.7824, "step": 7381 }, { "epoch": 0.267366896052155, "grad_norm": 1.5021765038072157, "learning_rate": 1.7188072000076862e-05, "loss": 0.7497, "step": 7382 }, { "epoch": 0.2674031148134734, "grad_norm": 1.1615172519546397, "learning_rate": 1.7187256422833928e-05, "loss": 0.7385, "step": 7383 }, { "epoch": 0.2674393335747917, "grad_norm": 1.2379315736014531, "learning_rate": 1.7186440746688032e-05, "loss": 0.8225, "step": 7384 }, { "epoch": 0.2674755523361101, "grad_norm": 1.5009014133163452, "learning_rate": 1.7185624971650395e-05, "loss": 0.8661, "step": 7385 }, { "epoch": 0.26751177109742846, "grad_norm": 1.587330150341343, "learning_rate": 1.7184809097732247e-05, "loss": 0.8027, "step": 7386 }, { "epoch": 0.26754798985874684, "grad_norm": 1.4023711791583626, "learning_rate": 1.718399312494481e-05, "loss": 0.8086, "step": 7387 }, { "epoch": 0.2675842086200652, "grad_norm": 1.1336046154269166, "learning_rate": 1.7183177053299317e-05, "loss": 0.7424, "step": 7388 }, { "epoch": 0.26762042738138353, "grad_norm": 1.9002095652140918, "learning_rate": 1.7182360882806998e-05, "loss": 0.7972, "step": 7389 }, { "epoch": 0.2676566461427019, "grad_norm": 1.5788408558443756, "learning_rate": 1.7181544613479077e-05, "loss": 0.811, "step": 7390 }, { "epoch": 0.2676928649040203, "grad_norm": 1.351664664012447, "learning_rate": 1.7180728245326796e-05, "loss": 0.7164, "step": 7391 }, { "epoch": 0.26772908366533865, "grad_norm": 1.5444711529255912, "learning_rate": 1.7179911778361386e-05, "loss": 0.8712, "step": 7392 }, { "epoch": 0.267765302426657, "grad_norm": 1.2260695374383128, "learning_rate": 1.7179095212594078e-05, "loss": 0.8498, "step": 7393 }, { "epoch": 0.26780152118797534, "grad_norm": 1.4074947662810966, "learning_rate": 1.717827854803611e-05, "loss": 0.796, "step": 7394 }, { "epoch": 0.2678377399492937, "grad_norm": 1.3895924463740599, "learning_rate": 1.7177461784698725e-05, "loss": 0.8208, "step": 7395 }, { "epoch": 0.2678739587106121, "grad_norm": 1.1260304449469283, "learning_rate": 1.7176644922593156e-05, "loss": 0.7327, "step": 7396 }, { "epoch": 0.26791017747193047, "grad_norm": 1.46548858309673, "learning_rate": 1.717582796173065e-05, "loss": 0.8609, "step": 7397 }, { "epoch": 0.26794639623324884, "grad_norm": 1.601560485652494, "learning_rate": 1.717501090212245e-05, "loss": 0.8124, "step": 7398 }, { "epoch": 0.26798261499456716, "grad_norm": 1.4555227048520567, "learning_rate": 1.717419374377979e-05, "loss": 0.8336, "step": 7399 }, { "epoch": 0.26801883375588553, "grad_norm": 1.2642493667744747, "learning_rate": 1.717337648671392e-05, "loss": 0.7765, "step": 7400 }, { "epoch": 0.2680550525172039, "grad_norm": 1.494730854638872, "learning_rate": 1.7172559130936088e-05, "loss": 0.8485, "step": 7401 }, { "epoch": 0.2680912712785223, "grad_norm": 1.055148756306923, "learning_rate": 1.717174167645754e-05, "loss": 0.7308, "step": 7402 }, { "epoch": 0.26812749003984065, "grad_norm": 1.5066060875496292, "learning_rate": 1.7170924123289522e-05, "loss": 0.8198, "step": 7403 }, { "epoch": 0.268163708801159, "grad_norm": 1.3129193998812312, "learning_rate": 1.7170106471443292e-05, "loss": 0.7563, "step": 7404 }, { "epoch": 0.26819992756247735, "grad_norm": 1.6859042870588374, "learning_rate": 1.7169288720930093e-05, "loss": 0.8599, "step": 7405 }, { "epoch": 0.2682361463237957, "grad_norm": 1.6318138321955062, "learning_rate": 1.7168470871761184e-05, "loss": 0.8455, "step": 7406 }, { "epoch": 0.2682723650851141, "grad_norm": 1.3672142421286864, "learning_rate": 1.7167652923947815e-05, "loss": 0.8301, "step": 7407 }, { "epoch": 0.26830858384643247, "grad_norm": 1.4121694116769827, "learning_rate": 1.7166834877501247e-05, "loss": 0.7735, "step": 7408 }, { "epoch": 0.2683448026077508, "grad_norm": 1.3999859865053494, "learning_rate": 1.7166016732432732e-05, "loss": 0.8687, "step": 7409 }, { "epoch": 0.26838102136906916, "grad_norm": 1.5205838975141301, "learning_rate": 1.7165198488753526e-05, "loss": 0.832, "step": 7410 }, { "epoch": 0.26841724013038754, "grad_norm": 1.331523221565086, "learning_rate": 1.7164380146474897e-05, "loss": 0.8337, "step": 7411 }, { "epoch": 0.2684534588917059, "grad_norm": 1.0636886476826553, "learning_rate": 1.7163561705608104e-05, "loss": 0.7991, "step": 7412 }, { "epoch": 0.2684896776530243, "grad_norm": 1.5097269643001174, "learning_rate": 1.7162743166164403e-05, "loss": 0.8432, "step": 7413 }, { "epoch": 0.2685258964143426, "grad_norm": 1.3896747956554119, "learning_rate": 1.7161924528155066e-05, "loss": 0.7776, "step": 7414 }, { "epoch": 0.268562115175661, "grad_norm": 1.6403431082558921, "learning_rate": 1.716110579159135e-05, "loss": 0.8689, "step": 7415 }, { "epoch": 0.26859833393697935, "grad_norm": 1.5462502508772304, "learning_rate": 1.716028695648453e-05, "loss": 0.8553, "step": 7416 }, { "epoch": 0.2686345526982977, "grad_norm": 1.4147842742545587, "learning_rate": 1.715946802284587e-05, "loss": 0.8265, "step": 7417 }, { "epoch": 0.2686707714596161, "grad_norm": 1.6372891962086895, "learning_rate": 1.715864899068664e-05, "loss": 0.8001, "step": 7418 }, { "epoch": 0.2687069902209344, "grad_norm": 1.3688364929148975, "learning_rate": 1.7157829860018104e-05, "loss": 0.7621, "step": 7419 }, { "epoch": 0.2687432089822528, "grad_norm": 1.3885489662922434, "learning_rate": 1.7157010630851544e-05, "loss": 0.7867, "step": 7420 }, { "epoch": 0.26877942774357116, "grad_norm": 2.302110095977929, "learning_rate": 1.7156191303198227e-05, "loss": 0.7859, "step": 7421 }, { "epoch": 0.26881564650488954, "grad_norm": 1.6185245365645622, "learning_rate": 1.7155371877069433e-05, "loss": 0.8485, "step": 7422 }, { "epoch": 0.2688518652662079, "grad_norm": 1.5357719404920525, "learning_rate": 1.7154552352476432e-05, "loss": 0.7807, "step": 7423 }, { "epoch": 0.2688880840275263, "grad_norm": 1.4148691770076591, "learning_rate": 1.7153732729430506e-05, "loss": 0.8264, "step": 7424 }, { "epoch": 0.2689243027888446, "grad_norm": 1.4747816111542957, "learning_rate": 1.715291300794293e-05, "loss": 0.7893, "step": 7425 }, { "epoch": 0.268960521550163, "grad_norm": 1.3677264935843867, "learning_rate": 1.715209318802499e-05, "loss": 0.8263, "step": 7426 }, { "epoch": 0.26899674031148135, "grad_norm": 1.4673901880536655, "learning_rate": 1.7151273269687954e-05, "loss": 0.8483, "step": 7427 }, { "epoch": 0.2690329590727997, "grad_norm": 1.4626597150937408, "learning_rate": 1.7150453252943126e-05, "loss": 0.7697, "step": 7428 }, { "epoch": 0.2690691778341181, "grad_norm": 1.3407375618556674, "learning_rate": 1.714963313780177e-05, "loss": 0.8179, "step": 7429 }, { "epoch": 0.2691053965954364, "grad_norm": 1.4398920329500806, "learning_rate": 1.7148812924275183e-05, "loss": 0.8095, "step": 7430 }, { "epoch": 0.2691416153567548, "grad_norm": 1.4165089412206553, "learning_rate": 1.7147992612374647e-05, "loss": 0.8734, "step": 7431 }, { "epoch": 0.26917783411807317, "grad_norm": 1.3539289253568043, "learning_rate": 1.7147172202111454e-05, "loss": 0.83, "step": 7432 }, { "epoch": 0.26921405287939154, "grad_norm": 1.5252953705981538, "learning_rate": 1.714635169349689e-05, "loss": 0.7464, "step": 7433 }, { "epoch": 0.2692502716407099, "grad_norm": 1.386264605771989, "learning_rate": 1.7145531086542246e-05, "loss": 0.8264, "step": 7434 }, { "epoch": 0.26928649040202823, "grad_norm": 1.4905464188138022, "learning_rate": 1.714471038125882e-05, "loss": 0.8401, "step": 7435 }, { "epoch": 0.2693227091633466, "grad_norm": 1.3579540142725388, "learning_rate": 1.7143889577657896e-05, "loss": 0.8565, "step": 7436 }, { "epoch": 0.269358927924665, "grad_norm": 1.4979031641835265, "learning_rate": 1.7143068675750777e-05, "loss": 0.8484, "step": 7437 }, { "epoch": 0.26939514668598336, "grad_norm": 1.2560153116485149, "learning_rate": 1.7142247675548754e-05, "loss": 0.8641, "step": 7438 }, { "epoch": 0.26943136544730173, "grad_norm": 1.436138631211499, "learning_rate": 1.7141426577063133e-05, "loss": 0.7883, "step": 7439 }, { "epoch": 0.26946758420862005, "grad_norm": 1.3743070148468892, "learning_rate": 1.7140605380305203e-05, "loss": 0.7944, "step": 7440 }, { "epoch": 0.2695038029699384, "grad_norm": 1.5605406325386098, "learning_rate": 1.7139784085286268e-05, "loss": 0.7814, "step": 7441 }, { "epoch": 0.2695400217312568, "grad_norm": 1.298029738108462, "learning_rate": 1.7138962692017633e-05, "loss": 0.8409, "step": 7442 }, { "epoch": 0.26957624049257517, "grad_norm": 1.5192147234175069, "learning_rate": 1.7138141200510595e-05, "loss": 0.7868, "step": 7443 }, { "epoch": 0.26961245925389354, "grad_norm": 1.2697682965408128, "learning_rate": 1.7137319610776466e-05, "loss": 0.78, "step": 7444 }, { "epoch": 0.26964867801521186, "grad_norm": 1.4050526309623583, "learning_rate": 1.7136497922826546e-05, "loss": 0.7682, "step": 7445 }, { "epoch": 0.26968489677653024, "grad_norm": 1.2692000129374004, "learning_rate": 1.7135676136672144e-05, "loss": 0.7967, "step": 7446 }, { "epoch": 0.2697211155378486, "grad_norm": 1.435337426353086, "learning_rate": 1.7134854252324567e-05, "loss": 0.8107, "step": 7447 }, { "epoch": 0.269757334299167, "grad_norm": 1.3322101756711175, "learning_rate": 1.713403226979513e-05, "loss": 0.8657, "step": 7448 }, { "epoch": 0.26979355306048536, "grad_norm": 1.4867000120859426, "learning_rate": 1.7133210189095136e-05, "loss": 0.8612, "step": 7449 }, { "epoch": 0.2698297718218037, "grad_norm": 1.4117934096416638, "learning_rate": 1.7132388010235908e-05, "loss": 0.845, "step": 7450 }, { "epoch": 0.26986599058312205, "grad_norm": 1.3318035729611806, "learning_rate": 1.713156573322875e-05, "loss": 0.8005, "step": 7451 }, { "epoch": 0.2699022093444404, "grad_norm": 1.7765976209532406, "learning_rate": 1.713074335808498e-05, "loss": 0.8718, "step": 7452 }, { "epoch": 0.2699384281057588, "grad_norm": 1.0912175463772817, "learning_rate": 1.7129920884815917e-05, "loss": 0.7665, "step": 7453 }, { "epoch": 0.2699746468670772, "grad_norm": 1.3693159864498716, "learning_rate": 1.712909831343288e-05, "loss": 0.8225, "step": 7454 }, { "epoch": 0.2700108656283955, "grad_norm": 1.2246674972754297, "learning_rate": 1.712827564394718e-05, "loss": 0.8082, "step": 7455 }, { "epoch": 0.27004708438971387, "grad_norm": 1.3938439315631743, "learning_rate": 1.712745287637015e-05, "loss": 0.803, "step": 7456 }, { "epoch": 0.27008330315103224, "grad_norm": 1.503818202847329, "learning_rate": 1.7126630010713105e-05, "loss": 0.8596, "step": 7457 }, { "epoch": 0.2701195219123506, "grad_norm": 1.490583577620177, "learning_rate": 1.7125807046987364e-05, "loss": 0.8569, "step": 7458 }, { "epoch": 0.270155740673669, "grad_norm": 1.1921925281606736, "learning_rate": 1.712498398520426e-05, "loss": 0.7421, "step": 7459 }, { "epoch": 0.2701919594349873, "grad_norm": 1.5506465832975596, "learning_rate": 1.712416082537512e-05, "loss": 0.8272, "step": 7460 }, { "epoch": 0.2702281781963057, "grad_norm": 1.1473664858680197, "learning_rate": 1.712333756751126e-05, "loss": 0.8118, "step": 7461 }, { "epoch": 0.27026439695762405, "grad_norm": 1.5492434660305323, "learning_rate": 1.712251421162402e-05, "loss": 0.8471, "step": 7462 }, { "epoch": 0.2703006157189424, "grad_norm": 1.1202993061972657, "learning_rate": 1.7121690757724722e-05, "loss": 0.7823, "step": 7463 }, { "epoch": 0.2703368344802608, "grad_norm": 1.4286894187167158, "learning_rate": 1.7120867205824704e-05, "loss": 0.7146, "step": 7464 }, { "epoch": 0.2703730532415791, "grad_norm": 1.065423370099154, "learning_rate": 1.71200435559353e-05, "loss": 0.7827, "step": 7465 }, { "epoch": 0.2704092720028975, "grad_norm": 1.39087930507873, "learning_rate": 1.7119219808067838e-05, "loss": 0.8254, "step": 7466 }, { "epoch": 0.27044549076421587, "grad_norm": 1.450732476519244, "learning_rate": 1.7118395962233655e-05, "loss": 0.8473, "step": 7467 }, { "epoch": 0.27048170952553424, "grad_norm": 1.081153233392701, "learning_rate": 1.711757201844409e-05, "loss": 0.8242, "step": 7468 }, { "epoch": 0.2705179282868526, "grad_norm": 1.7539528061187861, "learning_rate": 1.7116747976710476e-05, "loss": 0.8531, "step": 7469 }, { "epoch": 0.27055414704817093, "grad_norm": 1.3604811855156147, "learning_rate": 1.7115923837044158e-05, "loss": 0.7446, "step": 7470 }, { "epoch": 0.2705903658094893, "grad_norm": 1.4250497352344944, "learning_rate": 1.711509959945648e-05, "loss": 0.7485, "step": 7471 }, { "epoch": 0.2706265845708077, "grad_norm": 1.486919970734279, "learning_rate": 1.7114275263958778e-05, "loss": 0.8724, "step": 7472 }, { "epoch": 0.27066280333212606, "grad_norm": 1.501502307486486, "learning_rate": 1.7113450830562396e-05, "loss": 0.7937, "step": 7473 }, { "epoch": 0.27069902209344443, "grad_norm": 1.102877074294381, "learning_rate": 1.711262629927868e-05, "loss": 0.7816, "step": 7474 }, { "epoch": 0.27073524085476275, "grad_norm": 1.4776092381525345, "learning_rate": 1.7111801670118977e-05, "loss": 0.8195, "step": 7475 }, { "epoch": 0.2707714596160811, "grad_norm": 1.5026763413718258, "learning_rate": 1.711097694309463e-05, "loss": 0.8521, "step": 7476 }, { "epoch": 0.2708076783773995, "grad_norm": 1.4960907563809556, "learning_rate": 1.7110152118216997e-05, "loss": 0.8264, "step": 7477 }, { "epoch": 0.27084389713871787, "grad_norm": 1.4561782582358676, "learning_rate": 1.710932719549742e-05, "loss": 0.8246, "step": 7478 }, { "epoch": 0.27088011590003624, "grad_norm": 1.5050484834064437, "learning_rate": 1.710850217494726e-05, "loss": 0.8371, "step": 7479 }, { "epoch": 0.27091633466135456, "grad_norm": 1.5385209839768739, "learning_rate": 1.7107677056577856e-05, "loss": 0.8204, "step": 7480 }, { "epoch": 0.27095255342267294, "grad_norm": 1.4206713675263138, "learning_rate": 1.7106851840400574e-05, "loss": 0.7715, "step": 7481 }, { "epoch": 0.2709887721839913, "grad_norm": 1.5450741503550058, "learning_rate": 1.710602652642677e-05, "loss": 0.8571, "step": 7482 }, { "epoch": 0.2710249909453097, "grad_norm": 1.471398593786056, "learning_rate": 1.710520111466779e-05, "loss": 0.8504, "step": 7483 }, { "epoch": 0.27106120970662806, "grad_norm": 1.4790310547063312, "learning_rate": 1.7104375605135e-05, "loss": 0.8841, "step": 7484 }, { "epoch": 0.2710974284679464, "grad_norm": 1.4591569938830153, "learning_rate": 1.7103549997839764e-05, "loss": 0.7578, "step": 7485 }, { "epoch": 0.27113364722926475, "grad_norm": 1.3535032423149325, "learning_rate": 1.7102724292793435e-05, "loss": 0.734, "step": 7486 }, { "epoch": 0.2711698659905831, "grad_norm": 1.496561960704943, "learning_rate": 1.7101898490007382e-05, "loss": 0.7799, "step": 7487 }, { "epoch": 0.2712060847519015, "grad_norm": 1.6585914260517358, "learning_rate": 1.710107258949296e-05, "loss": 0.8957, "step": 7488 }, { "epoch": 0.2712423035132199, "grad_norm": 1.3817657095650802, "learning_rate": 1.7100246591261545e-05, "loss": 0.7987, "step": 7489 }, { "epoch": 0.2712785222745382, "grad_norm": 1.516269033681495, "learning_rate": 1.7099420495324497e-05, "loss": 0.7461, "step": 7490 }, { "epoch": 0.27131474103585657, "grad_norm": 1.1323640232607493, "learning_rate": 1.7098594301693183e-05, "loss": 0.8217, "step": 7491 }, { "epoch": 0.27135095979717494, "grad_norm": 1.0866960253694562, "learning_rate": 1.7097768010378976e-05, "loss": 0.7453, "step": 7492 }, { "epoch": 0.2713871785584933, "grad_norm": 1.2997146796682837, "learning_rate": 1.709694162139324e-05, "loss": 0.786, "step": 7493 }, { "epoch": 0.2714233973198117, "grad_norm": 1.5527340043265423, "learning_rate": 1.7096115134747358e-05, "loss": 0.7622, "step": 7494 }, { "epoch": 0.27145961608113, "grad_norm": 1.1488036693896861, "learning_rate": 1.7095288550452695e-05, "loss": 0.7517, "step": 7495 }, { "epoch": 0.2714958348424484, "grad_norm": 1.5215467396149274, "learning_rate": 1.7094461868520625e-05, "loss": 0.7887, "step": 7496 }, { "epoch": 0.27153205360376675, "grad_norm": 1.4774117046161668, "learning_rate": 1.7093635088962526e-05, "loss": 0.7824, "step": 7497 }, { "epoch": 0.27156827236508513, "grad_norm": 1.4717448637160577, "learning_rate": 1.7092808211789776e-05, "loss": 0.7647, "step": 7498 }, { "epoch": 0.2716044911264035, "grad_norm": 1.4075300973368448, "learning_rate": 1.7091981237013754e-05, "loss": 0.7855, "step": 7499 }, { "epoch": 0.2716407098877218, "grad_norm": 1.1846094069983737, "learning_rate": 1.7091154164645835e-05, "loss": 0.7832, "step": 7500 }, { "epoch": 0.2716769286490402, "grad_norm": 1.5833884610494324, "learning_rate": 1.7090326994697406e-05, "loss": 0.8601, "step": 7501 }, { "epoch": 0.27171314741035857, "grad_norm": 1.2473054784461988, "learning_rate": 1.708949972717985e-05, "loss": 0.7618, "step": 7502 }, { "epoch": 0.27174936617167694, "grad_norm": 1.2007585970779187, "learning_rate": 1.7088672362104543e-05, "loss": 0.8132, "step": 7503 }, { "epoch": 0.2717855849329953, "grad_norm": 1.4160089295327631, "learning_rate": 1.708784489948288e-05, "loss": 0.8609, "step": 7504 }, { "epoch": 0.27182180369431364, "grad_norm": 1.5012431084882052, "learning_rate": 1.708701733932624e-05, "loss": 0.8547, "step": 7505 }, { "epoch": 0.271858022455632, "grad_norm": 1.463730297342865, "learning_rate": 1.7086189681646016e-05, "loss": 0.8169, "step": 7506 }, { "epoch": 0.2718942412169504, "grad_norm": 1.492736931204348, "learning_rate": 1.7085361926453596e-05, "loss": 0.7961, "step": 7507 }, { "epoch": 0.27193045997826876, "grad_norm": 1.3983938014539334, "learning_rate": 1.708453407376037e-05, "loss": 0.7747, "step": 7508 }, { "epoch": 0.27196667873958713, "grad_norm": 1.4306741238914018, "learning_rate": 1.708370612357773e-05, "loss": 0.8394, "step": 7509 }, { "epoch": 0.27200289750090545, "grad_norm": 1.421287081026282, "learning_rate": 1.7082878075917068e-05, "loss": 0.7488, "step": 7510 }, { "epoch": 0.2720391162622238, "grad_norm": 1.3832463658729088, "learning_rate": 1.7082049930789782e-05, "loss": 0.7562, "step": 7511 }, { "epoch": 0.2720753350235422, "grad_norm": 1.6768305867687137, "learning_rate": 1.7081221688207262e-05, "loss": 0.8702, "step": 7512 }, { "epoch": 0.27211155378486057, "grad_norm": 1.548201023929462, "learning_rate": 1.7080393348180914e-05, "loss": 0.8284, "step": 7513 }, { "epoch": 0.27214777254617895, "grad_norm": 1.4662956163541558, "learning_rate": 1.7079564910722134e-05, "loss": 0.8246, "step": 7514 }, { "epoch": 0.27218399130749726, "grad_norm": 1.145618884156261, "learning_rate": 1.7078736375842317e-05, "loss": 0.7821, "step": 7515 }, { "epoch": 0.27222021006881564, "grad_norm": 1.5847996649072882, "learning_rate": 1.7077907743552867e-05, "loss": 0.9045, "step": 7516 }, { "epoch": 0.272256428830134, "grad_norm": 1.5535789885678568, "learning_rate": 1.7077079013865186e-05, "loss": 0.8154, "step": 7517 }, { "epoch": 0.2722926475914524, "grad_norm": 1.4787462324254077, "learning_rate": 1.7076250186790685e-05, "loss": 0.7932, "step": 7518 }, { "epoch": 0.27232886635277076, "grad_norm": 1.722669497588205, "learning_rate": 1.7075421262340757e-05, "loss": 0.8329, "step": 7519 }, { "epoch": 0.2723650851140891, "grad_norm": 1.4458410482264912, "learning_rate": 1.707459224052682e-05, "loss": 0.8206, "step": 7520 }, { "epoch": 0.27240130387540745, "grad_norm": 1.3698087557178737, "learning_rate": 1.7073763121360273e-05, "loss": 0.7548, "step": 7521 }, { "epoch": 0.2724375226367258, "grad_norm": 1.8286618988948824, "learning_rate": 1.7072933904852535e-05, "loss": 0.8267, "step": 7522 }, { "epoch": 0.2724737413980442, "grad_norm": 1.580192974334874, "learning_rate": 1.707210459101501e-05, "loss": 0.866, "step": 7523 }, { "epoch": 0.2725099601593626, "grad_norm": 1.6143598114669695, "learning_rate": 1.707127517985911e-05, "loss": 0.7352, "step": 7524 }, { "epoch": 0.2725461789206809, "grad_norm": 1.3323937717024945, "learning_rate": 1.7070445671396252e-05, "loss": 0.8033, "step": 7525 }, { "epoch": 0.27258239768199927, "grad_norm": 1.7225088069383347, "learning_rate": 1.7069616065637847e-05, "loss": 0.8391, "step": 7526 }, { "epoch": 0.27261861644331764, "grad_norm": 1.4467208251316124, "learning_rate": 1.7068786362595316e-05, "loss": 0.844, "step": 7527 }, { "epoch": 0.272654835204636, "grad_norm": 1.5308077876750052, "learning_rate": 1.7067956562280074e-05, "loss": 0.8252, "step": 7528 }, { "epoch": 0.2726910539659544, "grad_norm": 1.309437141929726, "learning_rate": 1.7067126664703534e-05, "loss": 0.7505, "step": 7529 }, { "epoch": 0.2727272727272727, "grad_norm": 1.5967383382681817, "learning_rate": 1.7066296669877123e-05, "loss": 0.8379, "step": 7530 }, { "epoch": 0.2727634914885911, "grad_norm": 1.4115639532842663, "learning_rate": 1.7065466577812264e-05, "loss": 0.7825, "step": 7531 }, { "epoch": 0.27279971024990946, "grad_norm": 1.0956786114125205, "learning_rate": 1.7064636388520373e-05, "loss": 0.7872, "step": 7532 }, { "epoch": 0.27283592901122783, "grad_norm": 1.4260098757403534, "learning_rate": 1.706380610201288e-05, "loss": 0.7998, "step": 7533 }, { "epoch": 0.2728721477725462, "grad_norm": 1.4490663841640563, "learning_rate": 1.706297571830121e-05, "loss": 0.7376, "step": 7534 }, { "epoch": 0.2729083665338645, "grad_norm": 1.5179340210346723, "learning_rate": 1.7062145237396786e-05, "loss": 0.8147, "step": 7535 }, { "epoch": 0.2729445852951829, "grad_norm": 1.6256018519009263, "learning_rate": 1.7061314659311036e-05, "loss": 0.8301, "step": 7536 }, { "epoch": 0.27298080405650127, "grad_norm": 1.2075066728713277, "learning_rate": 1.7060483984055394e-05, "loss": 0.7672, "step": 7537 }, { "epoch": 0.27301702281781964, "grad_norm": 1.416112888547368, "learning_rate": 1.705965321164129e-05, "loss": 0.8483, "step": 7538 }, { "epoch": 0.273053241579138, "grad_norm": 1.5354368290737648, "learning_rate": 1.705882234208015e-05, "loss": 0.7491, "step": 7539 }, { "epoch": 0.27308946034045634, "grad_norm": 1.5358790377786333, "learning_rate": 1.7057991375383417e-05, "loss": 0.7814, "step": 7540 }, { "epoch": 0.2731256791017747, "grad_norm": 1.6415525357597378, "learning_rate": 1.7057160311562517e-05, "loss": 0.7995, "step": 7541 }, { "epoch": 0.2731618978630931, "grad_norm": 1.3869719495781319, "learning_rate": 1.7056329150628893e-05, "loss": 0.7567, "step": 7542 }, { "epoch": 0.27319811662441146, "grad_norm": 1.5258818298311836, "learning_rate": 1.705549789259398e-05, "loss": 0.8596, "step": 7543 }, { "epoch": 0.27323433538572983, "grad_norm": 1.2369636343725363, "learning_rate": 1.7054666537469213e-05, "loss": 0.7942, "step": 7544 }, { "epoch": 0.27327055414704815, "grad_norm": 1.464382381261864, "learning_rate": 1.7053835085266038e-05, "loss": 0.8623, "step": 7545 }, { "epoch": 0.2733067729083665, "grad_norm": 1.4303177339507243, "learning_rate": 1.7053003535995895e-05, "loss": 0.7869, "step": 7546 }, { "epoch": 0.2733429916696849, "grad_norm": 1.2476609819765294, "learning_rate": 1.7052171889670224e-05, "loss": 0.818, "step": 7547 }, { "epoch": 0.27337921043100327, "grad_norm": 1.4919006973282007, "learning_rate": 1.7051340146300473e-05, "loss": 0.7827, "step": 7548 }, { "epoch": 0.27341542919232165, "grad_norm": 1.5118687066165613, "learning_rate": 1.7050508305898086e-05, "loss": 0.7718, "step": 7549 }, { "epoch": 0.27345164795363996, "grad_norm": 1.5733679609264748, "learning_rate": 1.704967636847451e-05, "loss": 0.7781, "step": 7550 }, { "epoch": 0.27348786671495834, "grad_norm": 1.3184390262984624, "learning_rate": 1.7048844334041187e-05, "loss": 0.6853, "step": 7551 }, { "epoch": 0.2735240854762767, "grad_norm": 0.996466508942212, "learning_rate": 1.704801220260958e-05, "loss": 0.754, "step": 7552 }, { "epoch": 0.2735603042375951, "grad_norm": 1.3516826521106173, "learning_rate": 1.7047179974191127e-05, "loss": 0.7676, "step": 7553 }, { "epoch": 0.27359652299891346, "grad_norm": 1.5795186328902884, "learning_rate": 1.7046347648797288e-05, "loss": 0.8377, "step": 7554 }, { "epoch": 0.2736327417602318, "grad_norm": 1.6583143091388521, "learning_rate": 1.7045515226439512e-05, "loss": 0.8548, "step": 7555 }, { "epoch": 0.27366896052155015, "grad_norm": 1.236578179097699, "learning_rate": 1.7044682707129258e-05, "loss": 0.8577, "step": 7556 }, { "epoch": 0.2737051792828685, "grad_norm": 1.3243747847121687, "learning_rate": 1.704385009087798e-05, "loss": 0.757, "step": 7557 }, { "epoch": 0.2737413980441869, "grad_norm": 1.6008774978853966, "learning_rate": 1.7043017377697137e-05, "loss": 0.7051, "step": 7558 }, { "epoch": 0.2737776168055053, "grad_norm": 1.4465880320012574, "learning_rate": 1.704218456759818e-05, "loss": 0.8196, "step": 7559 }, { "epoch": 0.2738138355668236, "grad_norm": 1.5710003642749013, "learning_rate": 1.704135166059258e-05, "loss": 0.8232, "step": 7560 }, { "epoch": 0.27385005432814197, "grad_norm": 1.2623329172684477, "learning_rate": 1.7040518656691793e-05, "loss": 0.7332, "step": 7561 }, { "epoch": 0.27388627308946034, "grad_norm": 1.373689516020726, "learning_rate": 1.7039685555907284e-05, "loss": 0.8767, "step": 7562 }, { "epoch": 0.2739224918507787, "grad_norm": 1.435439750670785, "learning_rate": 1.7038852358250514e-05, "loss": 0.7175, "step": 7563 }, { "epoch": 0.2739587106120971, "grad_norm": 1.4338300429664794, "learning_rate": 1.703801906373295e-05, "loss": 0.8478, "step": 7564 }, { "epoch": 0.2739949293734154, "grad_norm": 1.839240546265464, "learning_rate": 1.7037185672366065e-05, "loss": 0.875, "step": 7565 }, { "epoch": 0.2740311481347338, "grad_norm": 1.463316460092638, "learning_rate": 1.7036352184161318e-05, "loss": 0.8055, "step": 7566 }, { "epoch": 0.27406736689605216, "grad_norm": 1.7699271279254756, "learning_rate": 1.7035518599130184e-05, "loss": 0.6862, "step": 7567 }, { "epoch": 0.27410358565737053, "grad_norm": 1.4353475646408285, "learning_rate": 1.7034684917284126e-05, "loss": 0.8496, "step": 7568 }, { "epoch": 0.2741398044186889, "grad_norm": 1.4256438672490888, "learning_rate": 1.703385113863463e-05, "loss": 0.7935, "step": 7569 }, { "epoch": 0.2741760231800072, "grad_norm": 1.4924825194447762, "learning_rate": 1.7033017263193158e-05, "loss": 0.8828, "step": 7570 }, { "epoch": 0.2742122419413256, "grad_norm": 1.3764935991491405, "learning_rate": 1.7032183290971192e-05, "loss": 0.7936, "step": 7571 }, { "epoch": 0.27424846070264397, "grad_norm": 1.4539903158823142, "learning_rate": 1.70313492219802e-05, "loss": 0.8376, "step": 7572 }, { "epoch": 0.27428467946396234, "grad_norm": 1.4237803816149361, "learning_rate": 1.7030515056231668e-05, "loss": 0.7944, "step": 7573 }, { "epoch": 0.2743208982252807, "grad_norm": 1.2225015452595716, "learning_rate": 1.702968079373707e-05, "loss": 0.7165, "step": 7574 }, { "epoch": 0.27435711698659904, "grad_norm": 1.6508740760983038, "learning_rate": 1.702884643450789e-05, "loss": 0.8152, "step": 7575 }, { "epoch": 0.2743933357479174, "grad_norm": 1.5694242035992723, "learning_rate": 1.7028011978555602e-05, "loss": 0.9002, "step": 7576 }, { "epoch": 0.2744295545092358, "grad_norm": 1.5057753008813715, "learning_rate": 1.7027177425891695e-05, "loss": 0.8553, "step": 7577 }, { "epoch": 0.27446577327055416, "grad_norm": 1.6594716848521984, "learning_rate": 1.7026342776527654e-05, "loss": 0.8019, "step": 7578 }, { "epoch": 0.27450199203187253, "grad_norm": 1.4560027644924294, "learning_rate": 1.702550803047496e-05, "loss": 0.7983, "step": 7579 }, { "epoch": 0.27453821079319085, "grad_norm": 1.6572117689615464, "learning_rate": 1.702467318774511e-05, "loss": 0.8383, "step": 7580 }, { "epoch": 0.2745744295545092, "grad_norm": 1.4899010329193685, "learning_rate": 1.7023838248349574e-05, "loss": 0.7951, "step": 7581 }, { "epoch": 0.2746106483158276, "grad_norm": 1.3591426508292477, "learning_rate": 1.702300321229986e-05, "loss": 0.8521, "step": 7582 }, { "epoch": 0.274646867077146, "grad_norm": 1.1544878753424432, "learning_rate": 1.7022168079607446e-05, "loss": 0.8115, "step": 7583 }, { "epoch": 0.27468308583846435, "grad_norm": 1.4495656896012894, "learning_rate": 1.7021332850283833e-05, "loss": 0.7555, "step": 7584 }, { "epoch": 0.27471930459978267, "grad_norm": 1.3765140888651497, "learning_rate": 1.7020497524340505e-05, "loss": 0.8296, "step": 7585 }, { "epoch": 0.27475552336110104, "grad_norm": 1.5131966244732589, "learning_rate": 1.7019662101788967e-05, "loss": 0.7743, "step": 7586 }, { "epoch": 0.2747917421224194, "grad_norm": 1.5829772687409216, "learning_rate": 1.7018826582640708e-05, "loss": 0.8115, "step": 7587 }, { "epoch": 0.2748279608837378, "grad_norm": 1.4046405125079904, "learning_rate": 1.701799096690723e-05, "loss": 0.7771, "step": 7588 }, { "epoch": 0.27486417964505616, "grad_norm": 1.2499032761732525, "learning_rate": 1.701715525460003e-05, "loss": 0.8023, "step": 7589 }, { "epoch": 0.2749003984063745, "grad_norm": 1.409164215156278, "learning_rate": 1.7016319445730608e-05, "loss": 0.8035, "step": 7590 }, { "epoch": 0.27493661716769285, "grad_norm": 1.37641783448293, "learning_rate": 1.7015483540310462e-05, "loss": 0.8695, "step": 7591 }, { "epoch": 0.27497283592901123, "grad_norm": 1.4818032095801825, "learning_rate": 1.70146475383511e-05, "loss": 0.8432, "step": 7592 }, { "epoch": 0.2750090546903296, "grad_norm": 1.4523761227012664, "learning_rate": 1.7013811439864026e-05, "loss": 0.8473, "step": 7593 }, { "epoch": 0.275045273451648, "grad_norm": 1.4951570419590876, "learning_rate": 1.701297524486074e-05, "loss": 0.8572, "step": 7594 }, { "epoch": 0.2750814922129663, "grad_norm": 1.3771830623809267, "learning_rate": 1.7012138953352753e-05, "loss": 0.8372, "step": 7595 }, { "epoch": 0.27511771097428467, "grad_norm": 1.4781219507593235, "learning_rate": 1.7011302565351578e-05, "loss": 0.9047, "step": 7596 }, { "epoch": 0.27515392973560304, "grad_norm": 1.4376242774388661, "learning_rate": 1.7010466080868712e-05, "loss": 0.7687, "step": 7597 }, { "epoch": 0.2751901484969214, "grad_norm": 1.1726180648856819, "learning_rate": 1.7009629499915677e-05, "loss": 0.8431, "step": 7598 }, { "epoch": 0.2752263672582398, "grad_norm": 1.5494793200025296, "learning_rate": 1.7008792822503978e-05, "loss": 0.907, "step": 7599 }, { "epoch": 0.2752625860195581, "grad_norm": 1.4394203451453729, "learning_rate": 1.7007956048645136e-05, "loss": 0.7733, "step": 7600 }, { "epoch": 0.2752988047808765, "grad_norm": 1.113299603739663, "learning_rate": 1.7007119178350658e-05, "loss": 0.755, "step": 7601 }, { "epoch": 0.27533502354219486, "grad_norm": 1.549683665289868, "learning_rate": 1.700628221163206e-05, "loss": 0.7654, "step": 7602 }, { "epoch": 0.27537124230351323, "grad_norm": 1.631551748626479, "learning_rate": 1.700544514850087e-05, "loss": 0.8565, "step": 7603 }, { "epoch": 0.2754074610648316, "grad_norm": 1.4805613529060457, "learning_rate": 1.7004607988968594e-05, "loss": 0.7519, "step": 7604 }, { "epoch": 0.2754436798261499, "grad_norm": 1.3895675899714972, "learning_rate": 1.7003770733046757e-05, "loss": 0.796, "step": 7605 }, { "epoch": 0.2754798985874683, "grad_norm": 1.0209090086477401, "learning_rate": 1.7002933380746883e-05, "loss": 0.6934, "step": 7606 }, { "epoch": 0.27551611734878667, "grad_norm": 1.1303484144572185, "learning_rate": 1.700209593208049e-05, "loss": 0.7809, "step": 7607 }, { "epoch": 0.27555233611010505, "grad_norm": 1.6020824007206036, "learning_rate": 1.7001258387059106e-05, "loss": 0.9091, "step": 7608 }, { "epoch": 0.2755885548714234, "grad_norm": 1.324525350337262, "learning_rate": 1.7000420745694256e-05, "loss": 0.7359, "step": 7609 }, { "epoch": 0.27562477363274174, "grad_norm": 1.4014367036490982, "learning_rate": 1.699958300799746e-05, "loss": 0.7955, "step": 7610 }, { "epoch": 0.2756609923940601, "grad_norm": 1.1235572887571643, "learning_rate": 1.699874517398026e-05, "loss": 0.769, "step": 7611 }, { "epoch": 0.2756972111553785, "grad_norm": 1.4409467098998223, "learning_rate": 1.699790724365417e-05, "loss": 0.8029, "step": 7612 }, { "epoch": 0.27573342991669686, "grad_norm": 1.4351076063863568, "learning_rate": 1.699706921703073e-05, "loss": 0.8474, "step": 7613 }, { "epoch": 0.27576964867801523, "grad_norm": 1.3834103716936426, "learning_rate": 1.699623109412147e-05, "loss": 0.7429, "step": 7614 }, { "epoch": 0.27580586743933355, "grad_norm": 1.537329870014592, "learning_rate": 1.699539287493792e-05, "loss": 0.746, "step": 7615 }, { "epoch": 0.2758420862006519, "grad_norm": 1.4535039848470506, "learning_rate": 1.699455455949162e-05, "loss": 0.7767, "step": 7616 }, { "epoch": 0.2758783049619703, "grad_norm": 1.474030291302675, "learning_rate": 1.69937161477941e-05, "loss": 0.8218, "step": 7617 }, { "epoch": 0.2759145237232887, "grad_norm": 1.661064128588297, "learning_rate": 1.6992877639856903e-05, "loss": 0.8076, "step": 7618 }, { "epoch": 0.27595074248460705, "grad_norm": 1.4249326558261868, "learning_rate": 1.699203903569157e-05, "loss": 0.8447, "step": 7619 }, { "epoch": 0.27598696124592537, "grad_norm": 1.4558266212283493, "learning_rate": 1.6991200335309628e-05, "loss": 0.7556, "step": 7620 }, { "epoch": 0.27602318000724374, "grad_norm": 1.4259526633075679, "learning_rate": 1.699036153872263e-05, "loss": 0.8335, "step": 7621 }, { "epoch": 0.2760593987685621, "grad_norm": 2.0975405614090143, "learning_rate": 1.6989522645942112e-05, "loss": 0.8103, "step": 7622 }, { "epoch": 0.2760956175298805, "grad_norm": 1.3761392799045067, "learning_rate": 1.6988683656979624e-05, "loss": 0.7535, "step": 7623 }, { "epoch": 0.27613183629119886, "grad_norm": 1.2549171235258725, "learning_rate": 1.698784457184671e-05, "loss": 0.8101, "step": 7624 }, { "epoch": 0.2761680550525172, "grad_norm": 1.462484970688012, "learning_rate": 1.698700539055491e-05, "loss": 0.8959, "step": 7625 }, { "epoch": 0.27620427381383555, "grad_norm": 1.414295469437818, "learning_rate": 1.698616611311578e-05, "loss": 0.8288, "step": 7626 }, { "epoch": 0.27624049257515393, "grad_norm": 1.606314278682025, "learning_rate": 1.6985326739540867e-05, "loss": 0.7921, "step": 7627 }, { "epoch": 0.2762767113364723, "grad_norm": 1.3816081772591742, "learning_rate": 1.6984487269841716e-05, "loss": 0.7597, "step": 7628 }, { "epoch": 0.2763129300977907, "grad_norm": 1.1364753047068237, "learning_rate": 1.6983647704029886e-05, "loss": 0.859, "step": 7629 }, { "epoch": 0.276349148859109, "grad_norm": 1.3345082009945353, "learning_rate": 1.6982808042116926e-05, "loss": 0.7869, "step": 7630 }, { "epoch": 0.27638536762042737, "grad_norm": 1.2802727934423261, "learning_rate": 1.698196828411439e-05, "loss": 0.8087, "step": 7631 }, { "epoch": 0.27642158638174574, "grad_norm": 1.4589252172679128, "learning_rate": 1.698112843003384e-05, "loss": 0.8419, "step": 7632 }, { "epoch": 0.2764578051430641, "grad_norm": 1.464519099817288, "learning_rate": 1.6980288479886822e-05, "loss": 0.7776, "step": 7633 }, { "epoch": 0.2764940239043825, "grad_norm": 1.4252349852176047, "learning_rate": 1.6979448433684908e-05, "loss": 0.7756, "step": 7634 }, { "epoch": 0.2765302426657008, "grad_norm": 1.4819997794691526, "learning_rate": 1.6978608291439648e-05, "loss": 0.8573, "step": 7635 }, { "epoch": 0.2765664614270192, "grad_norm": 1.286613342815556, "learning_rate": 1.69777680531626e-05, "loss": 0.7937, "step": 7636 }, { "epoch": 0.27660268018833756, "grad_norm": 1.2404834058004899, "learning_rate": 1.697692771886534e-05, "loss": 0.8064, "step": 7637 }, { "epoch": 0.27663889894965593, "grad_norm": 1.520184088603121, "learning_rate": 1.6976087288559425e-05, "loss": 0.7466, "step": 7638 }, { "epoch": 0.2766751177109743, "grad_norm": 1.512839867597061, "learning_rate": 1.6975246762256414e-05, "loss": 0.8533, "step": 7639 }, { "epoch": 0.2767113364722926, "grad_norm": 1.432531516121024, "learning_rate": 1.697440613996788e-05, "loss": 0.8108, "step": 7640 }, { "epoch": 0.276747555233611, "grad_norm": 1.4878073122931235, "learning_rate": 1.6973565421705392e-05, "loss": 0.7847, "step": 7641 }, { "epoch": 0.27678377399492937, "grad_norm": 1.638914221074045, "learning_rate": 1.6972724607480513e-05, "loss": 0.857, "step": 7642 }, { "epoch": 0.27681999275624775, "grad_norm": 1.3834848547397054, "learning_rate": 1.697188369730482e-05, "loss": 0.7726, "step": 7643 }, { "epoch": 0.2768562115175661, "grad_norm": 1.2767958800377277, "learning_rate": 1.697104269118988e-05, "loss": 0.8066, "step": 7644 }, { "epoch": 0.27689243027888444, "grad_norm": 1.1784245384863323, "learning_rate": 1.6970201589147266e-05, "loss": 0.8156, "step": 7645 }, { "epoch": 0.2769286490402028, "grad_norm": 1.421629973667923, "learning_rate": 1.6969360391188554e-05, "loss": 0.8232, "step": 7646 }, { "epoch": 0.2769648678015212, "grad_norm": 1.4930968010612011, "learning_rate": 1.696851909732532e-05, "loss": 0.7818, "step": 7647 }, { "epoch": 0.27700108656283956, "grad_norm": 1.357785650949199, "learning_rate": 1.696767770756914e-05, "loss": 0.783, "step": 7648 }, { "epoch": 0.27703730532415793, "grad_norm": 1.2991466769747846, "learning_rate": 1.696683622193159e-05, "loss": 0.7595, "step": 7649 }, { "epoch": 0.27707352408547625, "grad_norm": 1.4756605456391843, "learning_rate": 1.6965994640424254e-05, "loss": 0.7484, "step": 7650 }, { "epoch": 0.2771097428467946, "grad_norm": 1.691951353321697, "learning_rate": 1.6965152963058714e-05, "loss": 0.8402, "step": 7651 }, { "epoch": 0.277145961608113, "grad_norm": 1.3890792930411804, "learning_rate": 1.6964311189846544e-05, "loss": 0.849, "step": 7652 }, { "epoch": 0.2771821803694314, "grad_norm": 1.3270967851008937, "learning_rate": 1.6963469320799336e-05, "loss": 0.8083, "step": 7653 }, { "epoch": 0.27721839913074975, "grad_norm": 1.567983973124627, "learning_rate": 1.696262735592867e-05, "loss": 0.8036, "step": 7654 }, { "epoch": 0.27725461789206807, "grad_norm": 1.5594199412067784, "learning_rate": 1.6961785295246134e-05, "loss": 0.8041, "step": 7655 }, { "epoch": 0.27729083665338644, "grad_norm": 1.1974332717939091, "learning_rate": 1.6960943138763312e-05, "loss": 0.7804, "step": 7656 }, { "epoch": 0.2773270554147048, "grad_norm": 1.453705979153881, "learning_rate": 1.6960100886491798e-05, "loss": 0.7233, "step": 7657 }, { "epoch": 0.2773632741760232, "grad_norm": 1.4225103485273307, "learning_rate": 1.6959258538443185e-05, "loss": 0.8008, "step": 7658 }, { "epoch": 0.27739949293734156, "grad_norm": 1.1637065767549546, "learning_rate": 1.6958416094629054e-05, "loss": 0.8236, "step": 7659 }, { "epoch": 0.2774357116986599, "grad_norm": 1.4138848307731682, "learning_rate": 1.6957573555061003e-05, "loss": 0.8177, "step": 7660 }, { "epoch": 0.27747193045997826, "grad_norm": 1.1296803668410806, "learning_rate": 1.695673091975063e-05, "loss": 0.8435, "step": 7661 }, { "epoch": 0.27750814922129663, "grad_norm": 1.3249004397178605, "learning_rate": 1.6955888188709527e-05, "loss": 0.8221, "step": 7662 }, { "epoch": 0.277544367982615, "grad_norm": 1.5824419689385807, "learning_rate": 1.6955045361949287e-05, "loss": 0.7603, "step": 7663 }, { "epoch": 0.2775805867439334, "grad_norm": 1.364960320980242, "learning_rate": 1.6954202439481517e-05, "loss": 0.726, "step": 7664 }, { "epoch": 0.2776168055052517, "grad_norm": 1.4514779606233974, "learning_rate": 1.6953359421317804e-05, "loss": 0.8067, "step": 7665 }, { "epoch": 0.27765302426657007, "grad_norm": 1.5708505049292059, "learning_rate": 1.695251630746976e-05, "loss": 0.8425, "step": 7666 }, { "epoch": 0.27768924302788844, "grad_norm": 1.1962335533770811, "learning_rate": 1.695167309794898e-05, "loss": 0.7217, "step": 7667 }, { "epoch": 0.2777254617892068, "grad_norm": 1.3932594154912668, "learning_rate": 1.6950829792767073e-05, "loss": 0.8029, "step": 7668 }, { "epoch": 0.2777616805505252, "grad_norm": 1.5131018549755373, "learning_rate": 1.694998639193564e-05, "loss": 0.8305, "step": 7669 }, { "epoch": 0.2777978993118435, "grad_norm": 1.1759867265834159, "learning_rate": 1.6949142895466283e-05, "loss": 0.7891, "step": 7670 }, { "epoch": 0.2778341180731619, "grad_norm": 1.5715277652311566, "learning_rate": 1.6948299303370616e-05, "loss": 0.8437, "step": 7671 }, { "epoch": 0.27787033683448026, "grad_norm": 1.3969931155976272, "learning_rate": 1.694745561566025e-05, "loss": 0.7817, "step": 7672 }, { "epoch": 0.27790655559579863, "grad_norm": 1.1964807374337247, "learning_rate": 1.6946611832346787e-05, "loss": 0.6892, "step": 7673 }, { "epoch": 0.277942774357117, "grad_norm": 1.5782872655982558, "learning_rate": 1.694576795344184e-05, "loss": 0.8243, "step": 7674 }, { "epoch": 0.2779789931184353, "grad_norm": 1.386691811796921, "learning_rate": 1.694492397895702e-05, "loss": 0.7606, "step": 7675 }, { "epoch": 0.2780152118797537, "grad_norm": 1.1192310580278553, "learning_rate": 1.6944079908903952e-05, "loss": 0.7348, "step": 7676 }, { "epoch": 0.2780514306410721, "grad_norm": 1.4938319398490045, "learning_rate": 1.6943235743294237e-05, "loss": 0.793, "step": 7677 }, { "epoch": 0.27808764940239045, "grad_norm": 1.490761348748259, "learning_rate": 1.6942391482139496e-05, "loss": 0.7822, "step": 7678 }, { "epoch": 0.2781238681637088, "grad_norm": 1.1525408515126814, "learning_rate": 1.6941547125451355e-05, "loss": 0.7018, "step": 7679 }, { "epoch": 0.27816008692502714, "grad_norm": 1.390932664563917, "learning_rate": 1.694070267324142e-05, "loss": 0.7837, "step": 7680 }, { "epoch": 0.2781963056863455, "grad_norm": 1.4268788646022412, "learning_rate": 1.693985812552132e-05, "loss": 0.8238, "step": 7681 }, { "epoch": 0.2782325244476639, "grad_norm": 1.4763270901926027, "learning_rate": 1.693901348230267e-05, "loss": 0.7868, "step": 7682 }, { "epoch": 0.27826874320898226, "grad_norm": 1.075307154447125, "learning_rate": 1.6938168743597105e-05, "loss": 0.813, "step": 7683 }, { "epoch": 0.27830496197030063, "grad_norm": 1.1280438535677781, "learning_rate": 1.693732390941624e-05, "loss": 0.8653, "step": 7684 }, { "epoch": 0.27834118073161895, "grad_norm": 1.7174726377900695, "learning_rate": 1.6936478979771697e-05, "loss": 0.8025, "step": 7685 }, { "epoch": 0.2783773994929373, "grad_norm": 1.443954294838012, "learning_rate": 1.6935633954675112e-05, "loss": 0.8492, "step": 7686 }, { "epoch": 0.2784136182542557, "grad_norm": 1.890890240166735, "learning_rate": 1.693478883413811e-05, "loss": 0.7173, "step": 7687 }, { "epoch": 0.2784498370155741, "grad_norm": 1.261660136604873, "learning_rate": 1.693394361817232e-05, "loss": 0.821, "step": 7688 }, { "epoch": 0.27848605577689245, "grad_norm": 1.3638373650692734, "learning_rate": 1.693309830678937e-05, "loss": 0.7796, "step": 7689 }, { "epoch": 0.27852227453821077, "grad_norm": 1.4515775303529557, "learning_rate": 1.6932252900000898e-05, "loss": 0.7594, "step": 7690 }, { "epoch": 0.27855849329952914, "grad_norm": 1.4136730337441241, "learning_rate": 1.6931407397818532e-05, "loss": 0.7155, "step": 7691 }, { "epoch": 0.2785947120608475, "grad_norm": 1.50743272240703, "learning_rate": 1.6930561800253916e-05, "loss": 0.7462, "step": 7692 }, { "epoch": 0.2786309308221659, "grad_norm": 1.351168229086314, "learning_rate": 1.6929716107318673e-05, "loss": 0.7703, "step": 7693 }, { "epoch": 0.27866714958348426, "grad_norm": 1.7307986310733305, "learning_rate": 1.6928870319024454e-05, "loss": 0.822, "step": 7694 }, { "epoch": 0.2787033683448026, "grad_norm": 1.3428617885306875, "learning_rate": 1.6928024435382887e-05, "loss": 0.7157, "step": 7695 }, { "epoch": 0.27873958710612096, "grad_norm": 1.531149759952806, "learning_rate": 1.6927178456405617e-05, "loss": 0.825, "step": 7696 }, { "epoch": 0.27877580586743933, "grad_norm": 1.5032034383598711, "learning_rate": 1.6926332382104286e-05, "loss": 0.8108, "step": 7697 }, { "epoch": 0.2788120246287577, "grad_norm": 1.0683334701204876, "learning_rate": 1.692548621249053e-05, "loss": 0.7657, "step": 7698 }, { "epoch": 0.2788482433900761, "grad_norm": 1.3564713216750164, "learning_rate": 1.6924639947576006e-05, "loss": 0.753, "step": 7699 }, { "epoch": 0.2788844621513944, "grad_norm": 1.408133704793093, "learning_rate": 1.6923793587372347e-05, "loss": 0.819, "step": 7700 }, { "epoch": 0.27892068091271277, "grad_norm": 1.3659306852938269, "learning_rate": 1.6922947131891208e-05, "loss": 0.7987, "step": 7701 }, { "epoch": 0.27895689967403114, "grad_norm": 1.2881320134613539, "learning_rate": 1.6922100581144228e-05, "loss": 0.7752, "step": 7702 }, { "epoch": 0.2789931184353495, "grad_norm": 1.4255862076489012, "learning_rate": 1.6921253935143068e-05, "loss": 0.7341, "step": 7703 }, { "epoch": 0.2790293371966679, "grad_norm": 1.6553440011436875, "learning_rate": 1.6920407193899368e-05, "loss": 0.837, "step": 7704 }, { "epoch": 0.2790655559579862, "grad_norm": 1.4865573388006665, "learning_rate": 1.6919560357424787e-05, "loss": 0.8292, "step": 7705 }, { "epoch": 0.2791017747193046, "grad_norm": 1.2710577985159586, "learning_rate": 1.6918713425730974e-05, "loss": 0.8113, "step": 7706 }, { "epoch": 0.27913799348062296, "grad_norm": 1.5235446525450405, "learning_rate": 1.6917866398829583e-05, "loss": 0.8358, "step": 7707 }, { "epoch": 0.27917421224194133, "grad_norm": 1.485443841182647, "learning_rate": 1.6917019276732274e-05, "loss": 0.8375, "step": 7708 }, { "epoch": 0.2792104310032597, "grad_norm": 1.5256193431575906, "learning_rate": 1.69161720594507e-05, "loss": 0.851, "step": 7709 }, { "epoch": 0.279246649764578, "grad_norm": 2.154452969316302, "learning_rate": 1.6915324746996526e-05, "loss": 0.8571, "step": 7710 }, { "epoch": 0.2792828685258964, "grad_norm": 1.5909500424825513, "learning_rate": 1.69144773393814e-05, "loss": 0.8508, "step": 7711 }, { "epoch": 0.2793190872872148, "grad_norm": 1.4167962888596013, "learning_rate": 1.6913629836616996e-05, "loss": 0.8331, "step": 7712 }, { "epoch": 0.27935530604853315, "grad_norm": 1.5782535838157146, "learning_rate": 1.691278223871497e-05, "loss": 0.8871, "step": 7713 }, { "epoch": 0.2793915248098515, "grad_norm": 1.4549274712216145, "learning_rate": 1.6911934545686986e-05, "loss": 0.767, "step": 7714 }, { "epoch": 0.27942774357116984, "grad_norm": 1.4608619002298926, "learning_rate": 1.691108675754471e-05, "loss": 0.8453, "step": 7715 }, { "epoch": 0.2794639623324882, "grad_norm": 1.3851614394469929, "learning_rate": 1.6910238874299805e-05, "loss": 0.884, "step": 7716 }, { "epoch": 0.2795001810938066, "grad_norm": 1.4768680686491396, "learning_rate": 1.690939089596394e-05, "loss": 0.7933, "step": 7717 }, { "epoch": 0.27953639985512496, "grad_norm": 1.4118774877917226, "learning_rate": 1.690854282254879e-05, "loss": 0.7598, "step": 7718 }, { "epoch": 0.27957261861644334, "grad_norm": 1.4853211361529466, "learning_rate": 1.690769465406602e-05, "loss": 0.7772, "step": 7719 }, { "epoch": 0.27960883737776165, "grad_norm": 1.2161813773615213, "learning_rate": 1.6906846390527296e-05, "loss": 0.8107, "step": 7720 }, { "epoch": 0.27964505613908003, "grad_norm": 1.3915781059295975, "learning_rate": 1.69059980319443e-05, "loss": 0.8033, "step": 7721 }, { "epoch": 0.2796812749003984, "grad_norm": 1.663096641241764, "learning_rate": 1.6905149578328705e-05, "loss": 0.8149, "step": 7722 }, { "epoch": 0.2797174936617168, "grad_norm": 1.332535815951938, "learning_rate": 1.6904301029692184e-05, "loss": 0.7815, "step": 7723 }, { "epoch": 0.27975371242303515, "grad_norm": 1.631962044213187, "learning_rate": 1.6903452386046413e-05, "loss": 0.899, "step": 7724 }, { "epoch": 0.27978993118435347, "grad_norm": 1.6305112364369234, "learning_rate": 1.690260364740307e-05, "loss": 0.8534, "step": 7725 }, { "epoch": 0.27982614994567184, "grad_norm": 1.3748891593408794, "learning_rate": 1.6901754813773838e-05, "loss": 0.7879, "step": 7726 }, { "epoch": 0.2798623687069902, "grad_norm": 1.6437701799691458, "learning_rate": 1.6900905885170392e-05, "loss": 0.8597, "step": 7727 }, { "epoch": 0.2798985874683086, "grad_norm": 1.4952607699467808, "learning_rate": 1.6900056861604422e-05, "loss": 0.8761, "step": 7728 }, { "epoch": 0.27993480622962696, "grad_norm": 1.4269215061037768, "learning_rate": 1.6899207743087603e-05, "loss": 0.8279, "step": 7729 }, { "epoch": 0.2799710249909453, "grad_norm": 1.6014666569339713, "learning_rate": 1.6898358529631625e-05, "loss": 0.883, "step": 7730 }, { "epoch": 0.28000724375226366, "grad_norm": 1.4360753131272754, "learning_rate": 1.689750922124817e-05, "loss": 0.8308, "step": 7731 }, { "epoch": 0.28004346251358203, "grad_norm": 1.121243018130386, "learning_rate": 1.6896659817948934e-05, "loss": 0.8481, "step": 7732 }, { "epoch": 0.2800796812749004, "grad_norm": 1.3185637449875327, "learning_rate": 1.6895810319745592e-05, "loss": 0.8028, "step": 7733 }, { "epoch": 0.2801159000362188, "grad_norm": 1.4682376416854765, "learning_rate": 1.6894960726649844e-05, "loss": 0.7731, "step": 7734 }, { "epoch": 0.2801521187975371, "grad_norm": 1.275656632322681, "learning_rate": 1.6894111038673378e-05, "loss": 0.8385, "step": 7735 }, { "epoch": 0.28018833755885547, "grad_norm": 1.3518634294717644, "learning_rate": 1.6893261255827887e-05, "loss": 0.7956, "step": 7736 }, { "epoch": 0.28022455632017385, "grad_norm": 1.46253969060193, "learning_rate": 1.6892411378125063e-05, "loss": 0.922, "step": 7737 }, { "epoch": 0.2802607750814922, "grad_norm": 1.464075061017506, "learning_rate": 1.68915614055766e-05, "loss": 0.8044, "step": 7738 }, { "epoch": 0.2802969938428106, "grad_norm": 1.0936871960822199, "learning_rate": 1.68907113381942e-05, "loss": 0.7457, "step": 7739 }, { "epoch": 0.2803332126041289, "grad_norm": 1.1660697611345368, "learning_rate": 1.6889861175989556e-05, "loss": 0.8407, "step": 7740 }, { "epoch": 0.2803694313654473, "grad_norm": 1.5358093434059075, "learning_rate": 1.6889010918974367e-05, "loss": 0.886, "step": 7741 }, { "epoch": 0.28040565012676566, "grad_norm": 1.4475343264308687, "learning_rate": 1.6888160567160335e-05, "loss": 0.7874, "step": 7742 }, { "epoch": 0.28044186888808403, "grad_norm": 1.4837098137009883, "learning_rate": 1.6887310120559162e-05, "loss": 0.7913, "step": 7743 }, { "epoch": 0.2804780876494024, "grad_norm": 1.6490627093023125, "learning_rate": 1.688645957918255e-05, "loss": 0.8287, "step": 7744 }, { "epoch": 0.2805143064107207, "grad_norm": 1.258632623926362, "learning_rate": 1.6885608943042202e-05, "loss": 0.7187, "step": 7745 }, { "epoch": 0.2805505251720391, "grad_norm": 1.6074321794005788, "learning_rate": 1.6884758212149827e-05, "loss": 0.8359, "step": 7746 }, { "epoch": 0.2805867439333575, "grad_norm": 1.1036232725951418, "learning_rate": 1.6883907386517128e-05, "loss": 0.7846, "step": 7747 }, { "epoch": 0.28062296269467585, "grad_norm": 1.478672181814334, "learning_rate": 1.6883056466155815e-05, "loss": 0.8464, "step": 7748 }, { "epoch": 0.2806591814559942, "grad_norm": 1.3944432774478748, "learning_rate": 1.6882205451077597e-05, "loss": 0.8197, "step": 7749 }, { "epoch": 0.28069540021731254, "grad_norm": 1.5467400050170823, "learning_rate": 1.6881354341294184e-05, "loss": 0.7843, "step": 7750 }, { "epoch": 0.2807316189786309, "grad_norm": 1.419027345810374, "learning_rate": 1.688050313681729e-05, "loss": 0.7515, "step": 7751 }, { "epoch": 0.2807678377399493, "grad_norm": 1.4940968662910128, "learning_rate": 1.6879651837658623e-05, "loss": 0.8034, "step": 7752 }, { "epoch": 0.28080405650126766, "grad_norm": 1.3954153070295365, "learning_rate": 1.6878800443829904e-05, "loss": 0.8466, "step": 7753 }, { "epoch": 0.28084027526258604, "grad_norm": 1.3698493005016368, "learning_rate": 1.687794895534285e-05, "loss": 0.7503, "step": 7754 }, { "epoch": 0.28087649402390436, "grad_norm": 1.3843521171975635, "learning_rate": 1.6877097372209174e-05, "loss": 0.764, "step": 7755 }, { "epoch": 0.28091271278522273, "grad_norm": 1.242580433440215, "learning_rate": 1.6876245694440595e-05, "loss": 0.7995, "step": 7756 }, { "epoch": 0.2809489315465411, "grad_norm": 1.0052085134083169, "learning_rate": 1.687539392204883e-05, "loss": 0.7858, "step": 7757 }, { "epoch": 0.2809851503078595, "grad_norm": 1.3654536315407524, "learning_rate": 1.687454205504561e-05, "loss": 0.7965, "step": 7758 }, { "epoch": 0.28102136906917785, "grad_norm": 1.408513298212595, "learning_rate": 1.6873690093442645e-05, "loss": 0.7407, "step": 7759 }, { "epoch": 0.28105758783049617, "grad_norm": 1.3341227986019955, "learning_rate": 1.6872838037251665e-05, "loss": 0.7609, "step": 7760 }, { "epoch": 0.28109380659181454, "grad_norm": 1.5564629119460112, "learning_rate": 1.6871985886484394e-05, "loss": 0.7992, "step": 7761 }, { "epoch": 0.2811300253531329, "grad_norm": 1.4247018765228083, "learning_rate": 1.6871133641152566e-05, "loss": 0.7323, "step": 7762 }, { "epoch": 0.2811662441144513, "grad_norm": 1.3660807789554674, "learning_rate": 1.6870281301267893e-05, "loss": 0.7682, "step": 7763 }, { "epoch": 0.28120246287576967, "grad_norm": 1.4850952709829084, "learning_rate": 1.686942886684212e-05, "loss": 0.7859, "step": 7764 }, { "epoch": 0.281238681637088, "grad_norm": 1.3771462839337245, "learning_rate": 1.6868576337886963e-05, "loss": 0.7854, "step": 7765 }, { "epoch": 0.28127490039840636, "grad_norm": 1.4778664044160306, "learning_rate": 1.686772371441416e-05, "loss": 0.7616, "step": 7766 }, { "epoch": 0.28131111915972473, "grad_norm": 1.3314843817125959, "learning_rate": 1.6866870996435448e-05, "loss": 0.7274, "step": 7767 }, { "epoch": 0.2813473379210431, "grad_norm": 1.1821606065343553, "learning_rate": 1.686601818396256e-05, "loss": 0.7859, "step": 7768 }, { "epoch": 0.2813835566823615, "grad_norm": 1.4104728871592593, "learning_rate": 1.686516527700722e-05, "loss": 0.7083, "step": 7769 }, { "epoch": 0.28141977544367985, "grad_norm": 1.4446048603894523, "learning_rate": 1.686431227558118e-05, "loss": 0.7785, "step": 7770 }, { "epoch": 0.28145599420499817, "grad_norm": 1.5170791338299152, "learning_rate": 1.686345917969617e-05, "loss": 0.8335, "step": 7771 }, { "epoch": 0.28149221296631655, "grad_norm": 1.5484335145312, "learning_rate": 1.6862605989363933e-05, "loss": 0.8856, "step": 7772 }, { "epoch": 0.2815284317276349, "grad_norm": 1.4430895477006898, "learning_rate": 1.6861752704596206e-05, "loss": 0.8489, "step": 7773 }, { "epoch": 0.2815646504889533, "grad_norm": 1.3307458627055293, "learning_rate": 1.6860899325404732e-05, "loss": 0.7945, "step": 7774 }, { "epoch": 0.28160086925027167, "grad_norm": 1.4531698726319382, "learning_rate": 1.6860045851801256e-05, "loss": 0.8124, "step": 7775 }, { "epoch": 0.28163708801159, "grad_norm": 1.4272879053778398, "learning_rate": 1.685919228379752e-05, "loss": 0.856, "step": 7776 }, { "epoch": 0.28167330677290836, "grad_norm": 1.9172600920528255, "learning_rate": 1.685833862140527e-05, "loss": 0.8527, "step": 7777 }, { "epoch": 0.28170952553422673, "grad_norm": 1.3872992495989167, "learning_rate": 1.685748486463626e-05, "loss": 0.8454, "step": 7778 }, { "epoch": 0.2817457442955451, "grad_norm": 1.4380369932670658, "learning_rate": 1.6856631013502226e-05, "loss": 0.7917, "step": 7779 }, { "epoch": 0.2817819630568635, "grad_norm": 1.2671731353222955, "learning_rate": 1.6855777068014925e-05, "loss": 0.7815, "step": 7780 }, { "epoch": 0.2818181818181818, "grad_norm": 1.4027656389426362, "learning_rate": 1.6854923028186112e-05, "loss": 0.7667, "step": 7781 }, { "epoch": 0.2818544005795002, "grad_norm": 1.351472676605431, "learning_rate": 1.6854068894027532e-05, "loss": 0.7866, "step": 7782 }, { "epoch": 0.28189061934081855, "grad_norm": 1.5024462691443157, "learning_rate": 1.6853214665550942e-05, "loss": 0.6947, "step": 7783 }, { "epoch": 0.2819268381021369, "grad_norm": 1.095244082214923, "learning_rate": 1.6852360342768096e-05, "loss": 0.7759, "step": 7784 }, { "epoch": 0.2819630568634553, "grad_norm": 1.3783481943738327, "learning_rate": 1.685150592569075e-05, "loss": 0.7962, "step": 7785 }, { "epoch": 0.2819992756247736, "grad_norm": 1.4897670463056314, "learning_rate": 1.6850651414330666e-05, "loss": 0.7901, "step": 7786 }, { "epoch": 0.282035494386092, "grad_norm": 1.610039133007621, "learning_rate": 1.6849796808699595e-05, "loss": 0.7015, "step": 7787 }, { "epoch": 0.28207171314741036, "grad_norm": 1.513116947096116, "learning_rate": 1.6848942108809302e-05, "loss": 0.8529, "step": 7788 }, { "epoch": 0.28210793190872874, "grad_norm": 1.4606862406722163, "learning_rate": 1.6848087314671547e-05, "loss": 0.8701, "step": 7789 }, { "epoch": 0.2821441506700471, "grad_norm": 1.4779915126342624, "learning_rate": 1.6847232426298097e-05, "loss": 0.7983, "step": 7790 }, { "epoch": 0.28218036943136543, "grad_norm": 1.4355017187840438, "learning_rate": 1.684637744370071e-05, "loss": 0.8088, "step": 7791 }, { "epoch": 0.2822165881926838, "grad_norm": 1.3650442074192828, "learning_rate": 1.6845522366891152e-05, "loss": 0.8664, "step": 7792 }, { "epoch": 0.2822528069540022, "grad_norm": 1.4165332394034535, "learning_rate": 1.684466719588119e-05, "loss": 0.8024, "step": 7793 }, { "epoch": 0.28228902571532055, "grad_norm": 1.2948034722869737, "learning_rate": 1.6843811930682597e-05, "loss": 0.8032, "step": 7794 }, { "epoch": 0.2823252444766389, "grad_norm": 1.3381050362683506, "learning_rate": 1.684295657130714e-05, "loss": 0.8325, "step": 7795 }, { "epoch": 0.28236146323795724, "grad_norm": 1.1921629902243815, "learning_rate": 1.6842101117766584e-05, "loss": 0.7713, "step": 7796 }, { "epoch": 0.2823976819992756, "grad_norm": 1.6958070222721646, "learning_rate": 1.6841245570072703e-05, "loss": 0.8318, "step": 7797 }, { "epoch": 0.282433900760594, "grad_norm": 1.0117716664942793, "learning_rate": 1.6840389928237276e-05, "loss": 0.7284, "step": 7798 }, { "epoch": 0.28247011952191237, "grad_norm": 1.4008930048693498, "learning_rate": 1.683953419227207e-05, "loss": 0.7733, "step": 7799 }, { "epoch": 0.28250633828323074, "grad_norm": 1.3558078152497872, "learning_rate": 1.6838678362188864e-05, "loss": 0.7445, "step": 7800 }, { "epoch": 0.28254255704454906, "grad_norm": 1.5079086376830055, "learning_rate": 1.683782243799944e-05, "loss": 0.8409, "step": 7801 }, { "epoch": 0.28257877580586743, "grad_norm": 1.4373347560610714, "learning_rate": 1.6836966419715564e-05, "loss": 0.7588, "step": 7802 }, { "epoch": 0.2826149945671858, "grad_norm": 1.3652153136596736, "learning_rate": 1.6836110307349026e-05, "loss": 0.7848, "step": 7803 }, { "epoch": 0.2826512133285042, "grad_norm": 1.5749798702474958, "learning_rate": 1.68352541009116e-05, "loss": 0.8199, "step": 7804 }, { "epoch": 0.28268743208982255, "grad_norm": 1.5207985910492838, "learning_rate": 1.6834397800415077e-05, "loss": 0.8188, "step": 7805 }, { "epoch": 0.2827236508511409, "grad_norm": 1.239302936480936, "learning_rate": 1.6833541405871235e-05, "loss": 0.7758, "step": 7806 }, { "epoch": 0.28275986961245925, "grad_norm": 1.1795399694765385, "learning_rate": 1.6832684917291855e-05, "loss": 0.8488, "step": 7807 }, { "epoch": 0.2827960883737776, "grad_norm": 1.5779315551376576, "learning_rate": 1.683182833468873e-05, "loss": 0.8098, "step": 7808 }, { "epoch": 0.282832307135096, "grad_norm": 1.4502677289304984, "learning_rate": 1.683097165807364e-05, "loss": 0.7895, "step": 7809 }, { "epoch": 0.28286852589641437, "grad_norm": 1.5143951662960802, "learning_rate": 1.683011488745838e-05, "loss": 0.8577, "step": 7810 }, { "epoch": 0.2829047446577327, "grad_norm": 1.5433583552367545, "learning_rate": 1.6829258022854738e-05, "loss": 0.8079, "step": 7811 }, { "epoch": 0.28294096341905106, "grad_norm": 1.3536311522772988, "learning_rate": 1.6828401064274502e-05, "loss": 0.7044, "step": 7812 }, { "epoch": 0.28297718218036944, "grad_norm": 1.4099405479040141, "learning_rate": 1.682754401172947e-05, "loss": 0.8721, "step": 7813 }, { "epoch": 0.2830134009416878, "grad_norm": 1.5856874543653434, "learning_rate": 1.6826686865231434e-05, "loss": 0.7797, "step": 7814 }, { "epoch": 0.2830496197030062, "grad_norm": 1.6243696113348802, "learning_rate": 1.6825829624792187e-05, "loss": 0.7511, "step": 7815 }, { "epoch": 0.2830858384643245, "grad_norm": 1.5187025862914312, "learning_rate": 1.6824972290423527e-05, "loss": 0.8531, "step": 7816 }, { "epoch": 0.2831220572256429, "grad_norm": 1.5228375653539399, "learning_rate": 1.6824114862137254e-05, "loss": 0.7955, "step": 7817 }, { "epoch": 0.28315827598696125, "grad_norm": 1.379408277595015, "learning_rate": 1.682325733994516e-05, "loss": 0.7302, "step": 7818 }, { "epoch": 0.2831944947482796, "grad_norm": 1.4208259708238489, "learning_rate": 1.6822399723859053e-05, "loss": 0.7785, "step": 7819 }, { "epoch": 0.283230713509598, "grad_norm": 1.089406356195765, "learning_rate": 1.6821542013890727e-05, "loss": 0.7727, "step": 7820 }, { "epoch": 0.2832669322709163, "grad_norm": 1.5595064812999773, "learning_rate": 1.6820684210051993e-05, "loss": 0.8505, "step": 7821 }, { "epoch": 0.2833031510322347, "grad_norm": 1.3000671471393093, "learning_rate": 1.6819826312354647e-05, "loss": 0.7863, "step": 7822 }, { "epoch": 0.28333936979355306, "grad_norm": 1.2226896384284163, "learning_rate": 1.6818968320810502e-05, "loss": 0.786, "step": 7823 }, { "epoch": 0.28337558855487144, "grad_norm": 1.4342015703337887, "learning_rate": 1.6818110235431362e-05, "loss": 0.7632, "step": 7824 }, { "epoch": 0.2834118073161898, "grad_norm": 1.202792885654291, "learning_rate": 1.681725205622903e-05, "loss": 0.7953, "step": 7825 }, { "epoch": 0.28344802607750813, "grad_norm": 1.296000200417734, "learning_rate": 1.6816393783215325e-05, "loss": 0.8089, "step": 7826 }, { "epoch": 0.2834842448388265, "grad_norm": 1.455142615956498, "learning_rate": 1.681553541640205e-05, "loss": 0.7844, "step": 7827 }, { "epoch": 0.2835204636001449, "grad_norm": 1.3205673237088316, "learning_rate": 1.681467695580102e-05, "loss": 0.7541, "step": 7828 }, { "epoch": 0.28355668236146325, "grad_norm": 1.430943916732953, "learning_rate": 1.6813818401424042e-05, "loss": 0.7342, "step": 7829 }, { "epoch": 0.2835929011227816, "grad_norm": 1.1829555383412327, "learning_rate": 1.681295975328294e-05, "loss": 0.7279, "step": 7830 }, { "epoch": 0.28362911988409995, "grad_norm": 1.5361750755146735, "learning_rate": 1.681210101138952e-05, "loss": 0.8157, "step": 7831 }, { "epoch": 0.2836653386454183, "grad_norm": 1.5133002040229429, "learning_rate": 1.681124217575561e-05, "loss": 0.8027, "step": 7832 }, { "epoch": 0.2837015574067367, "grad_norm": 1.4685845801863378, "learning_rate": 1.6810383246393022e-05, "loss": 0.7729, "step": 7833 }, { "epoch": 0.28373777616805507, "grad_norm": 1.258627931352981, "learning_rate": 1.6809524223313577e-05, "loss": 0.8076, "step": 7834 }, { "epoch": 0.28377399492937344, "grad_norm": 1.8465221707014756, "learning_rate": 1.6808665106529096e-05, "loss": 0.786, "step": 7835 }, { "epoch": 0.28381021369069176, "grad_norm": 1.536491812863868, "learning_rate": 1.6807805896051397e-05, "loss": 0.8303, "step": 7836 }, { "epoch": 0.28384643245201013, "grad_norm": 1.4531999627640806, "learning_rate": 1.6806946591892308e-05, "loss": 0.798, "step": 7837 }, { "epoch": 0.2838826512133285, "grad_norm": 1.4080769664901827, "learning_rate": 1.6806087194063657e-05, "loss": 0.8815, "step": 7838 }, { "epoch": 0.2839188699746469, "grad_norm": 1.3775472583416066, "learning_rate": 1.6805227702577262e-05, "loss": 0.7753, "step": 7839 }, { "epoch": 0.28395508873596526, "grad_norm": 1.5878054172509557, "learning_rate": 1.6804368117444954e-05, "loss": 0.791, "step": 7840 }, { "epoch": 0.2839913074972836, "grad_norm": 1.43777334423595, "learning_rate": 1.6803508438678564e-05, "loss": 0.8108, "step": 7841 }, { "epoch": 0.28402752625860195, "grad_norm": 1.428015269685154, "learning_rate": 1.6802648666289915e-05, "loss": 0.8028, "step": 7842 }, { "epoch": 0.2840637450199203, "grad_norm": 1.4797822296570806, "learning_rate": 1.6801788800290846e-05, "loss": 0.8079, "step": 7843 }, { "epoch": 0.2840999637812387, "grad_norm": 1.4893025450945485, "learning_rate": 1.680092884069319e-05, "loss": 0.8412, "step": 7844 }, { "epoch": 0.28413618254255707, "grad_norm": 1.3203995474217156, "learning_rate": 1.6800068787508775e-05, "loss": 0.7171, "step": 7845 }, { "epoch": 0.2841724013038754, "grad_norm": 1.1211663293895726, "learning_rate": 1.679920864074944e-05, "loss": 0.7835, "step": 7846 }, { "epoch": 0.28420862006519376, "grad_norm": 1.4712785225949876, "learning_rate": 1.6798348400427012e-05, "loss": 0.7998, "step": 7847 }, { "epoch": 0.28424483882651214, "grad_norm": 1.4079785994685916, "learning_rate": 1.6797488066553344e-05, "loss": 0.8161, "step": 7848 }, { "epoch": 0.2842810575878305, "grad_norm": 1.093151073513251, "learning_rate": 1.6796627639140266e-05, "loss": 0.7446, "step": 7849 }, { "epoch": 0.2843172763491489, "grad_norm": 1.3798947067180072, "learning_rate": 1.6795767118199618e-05, "loss": 0.8391, "step": 7850 }, { "epoch": 0.2843534951104672, "grad_norm": 1.5290246467867197, "learning_rate": 1.6794906503743244e-05, "loss": 0.8202, "step": 7851 }, { "epoch": 0.2843897138717856, "grad_norm": 1.336231102525432, "learning_rate": 1.6794045795782987e-05, "loss": 0.8097, "step": 7852 }, { "epoch": 0.28442593263310395, "grad_norm": 1.1737223274424056, "learning_rate": 1.679318499433069e-05, "loss": 0.8376, "step": 7853 }, { "epoch": 0.2844621513944223, "grad_norm": 1.4116602601588937, "learning_rate": 1.6792324099398196e-05, "loss": 0.7887, "step": 7854 }, { "epoch": 0.2844983701557407, "grad_norm": 1.395152475860563, "learning_rate": 1.6791463110997358e-05, "loss": 0.7873, "step": 7855 }, { "epoch": 0.284534588917059, "grad_norm": 1.5402006549515674, "learning_rate": 1.6790602029140014e-05, "loss": 0.8471, "step": 7856 }, { "epoch": 0.2845708076783774, "grad_norm": 1.352554441701992, "learning_rate": 1.6789740853838026e-05, "loss": 0.8625, "step": 7857 }, { "epoch": 0.28460702643969576, "grad_norm": 1.2645246258261604, "learning_rate": 1.678887958510323e-05, "loss": 0.769, "step": 7858 }, { "epoch": 0.28464324520101414, "grad_norm": 1.4486859651935782, "learning_rate": 1.6788018222947496e-05, "loss": 0.7194, "step": 7859 }, { "epoch": 0.2846794639623325, "grad_norm": 1.5618171935961138, "learning_rate": 1.678715676738266e-05, "loss": 0.7846, "step": 7860 }, { "epoch": 0.28471568272365083, "grad_norm": 1.476118915523436, "learning_rate": 1.6786295218420587e-05, "loss": 0.7953, "step": 7861 }, { "epoch": 0.2847519014849692, "grad_norm": 1.4648076205566423, "learning_rate": 1.6785433576073125e-05, "loss": 0.7998, "step": 7862 }, { "epoch": 0.2847881202462876, "grad_norm": 1.3578105582727082, "learning_rate": 1.6784571840352137e-05, "loss": 0.7859, "step": 7863 }, { "epoch": 0.28482433900760595, "grad_norm": 1.7296915310999652, "learning_rate": 1.678371001126948e-05, "loss": 0.7709, "step": 7864 }, { "epoch": 0.2848605577689243, "grad_norm": 1.3859375284382023, "learning_rate": 1.6782848088837008e-05, "loss": 0.7896, "step": 7865 }, { "epoch": 0.28489677653024265, "grad_norm": 1.1519570686358365, "learning_rate": 1.6781986073066588e-05, "loss": 0.7949, "step": 7866 }, { "epoch": 0.284932995291561, "grad_norm": 1.45439010969845, "learning_rate": 1.678112396397008e-05, "loss": 0.8512, "step": 7867 }, { "epoch": 0.2849692140528794, "grad_norm": 1.5438199532254306, "learning_rate": 1.6780261761559353e-05, "loss": 0.839, "step": 7868 }, { "epoch": 0.28500543281419777, "grad_norm": 1.6041791892698245, "learning_rate": 1.6779399465846262e-05, "loss": 0.8424, "step": 7869 }, { "epoch": 0.28504165157551614, "grad_norm": 1.4381233268497617, "learning_rate": 1.677853707684268e-05, "loss": 0.777, "step": 7870 }, { "epoch": 0.28507787033683446, "grad_norm": 1.4395942955202374, "learning_rate": 1.677767459456047e-05, "loss": 0.7607, "step": 7871 }, { "epoch": 0.28511408909815283, "grad_norm": 1.3272023740900072, "learning_rate": 1.6776812019011503e-05, "loss": 0.8053, "step": 7872 }, { "epoch": 0.2851503078594712, "grad_norm": 1.4243052949364101, "learning_rate": 1.6775949350207646e-05, "loss": 0.8022, "step": 7873 }, { "epoch": 0.2851865266207896, "grad_norm": 1.4299023973530323, "learning_rate": 1.6775086588160773e-05, "loss": 0.8476, "step": 7874 }, { "epoch": 0.28522274538210796, "grad_norm": 1.4620588411846682, "learning_rate": 1.6774223732882758e-05, "loss": 0.7595, "step": 7875 }, { "epoch": 0.2852589641434263, "grad_norm": 1.3468160774592015, "learning_rate": 1.6773360784385467e-05, "loss": 0.7986, "step": 7876 }, { "epoch": 0.28529518290474465, "grad_norm": 1.3108745135553945, "learning_rate": 1.6772497742680783e-05, "loss": 0.7837, "step": 7877 }, { "epoch": 0.285331401666063, "grad_norm": 1.1676937663865932, "learning_rate": 1.6771634607780584e-05, "loss": 0.7183, "step": 7878 }, { "epoch": 0.2853676204273814, "grad_norm": 1.120891922135668, "learning_rate": 1.6770771379696735e-05, "loss": 0.8003, "step": 7879 }, { "epoch": 0.28540383918869977, "grad_norm": 1.376049431711673, "learning_rate": 1.6769908058441125e-05, "loss": 0.754, "step": 7880 }, { "epoch": 0.2854400579500181, "grad_norm": 1.1710464589386784, "learning_rate": 1.676904464402563e-05, "loss": 0.7677, "step": 7881 }, { "epoch": 0.28547627671133646, "grad_norm": 1.3908586051371823, "learning_rate": 1.676818113646214e-05, "loss": 0.7987, "step": 7882 }, { "epoch": 0.28551249547265484, "grad_norm": 1.4312188579781926, "learning_rate": 1.6767317535762523e-05, "loss": 0.8432, "step": 7883 }, { "epoch": 0.2855487142339732, "grad_norm": 1.3527314279971958, "learning_rate": 1.6766453841938672e-05, "loss": 0.7658, "step": 7884 }, { "epoch": 0.2855849329952916, "grad_norm": 1.513482756336794, "learning_rate": 1.676559005500247e-05, "loss": 0.8535, "step": 7885 }, { "epoch": 0.2856211517566099, "grad_norm": 1.5620281770726892, "learning_rate": 1.676472617496581e-05, "loss": 0.7773, "step": 7886 }, { "epoch": 0.2856573705179283, "grad_norm": 1.3007684523007323, "learning_rate": 1.676386220184057e-05, "loss": 0.8271, "step": 7887 }, { "epoch": 0.28569358927924665, "grad_norm": 1.3613556540290346, "learning_rate": 1.676299813563864e-05, "loss": 0.7786, "step": 7888 }, { "epoch": 0.285729808040565, "grad_norm": 1.348489015469289, "learning_rate": 1.6762133976371914e-05, "loss": 0.7797, "step": 7889 }, { "epoch": 0.2857660268018834, "grad_norm": 1.6015420377401082, "learning_rate": 1.6761269724052285e-05, "loss": 0.9106, "step": 7890 }, { "epoch": 0.2858022455632017, "grad_norm": 1.4076695310859777, "learning_rate": 1.6760405378691646e-05, "loss": 0.8143, "step": 7891 }, { "epoch": 0.2858384643245201, "grad_norm": 1.4450943569571604, "learning_rate": 1.675954094030188e-05, "loss": 0.8115, "step": 7892 }, { "epoch": 0.28587468308583847, "grad_norm": 1.1923452520976516, "learning_rate": 1.67586764088949e-05, "loss": 0.8138, "step": 7893 }, { "epoch": 0.28591090184715684, "grad_norm": 1.2253028535098345, "learning_rate": 1.675781178448259e-05, "loss": 0.7679, "step": 7894 }, { "epoch": 0.2859471206084752, "grad_norm": 1.1111776324616964, "learning_rate": 1.6756947067076855e-05, "loss": 0.8081, "step": 7895 }, { "epoch": 0.28598333936979353, "grad_norm": 1.4786529716584993, "learning_rate": 1.675608225668959e-05, "loss": 0.8243, "step": 7896 }, { "epoch": 0.2860195581311119, "grad_norm": 1.5885081275281223, "learning_rate": 1.6755217353332694e-05, "loss": 0.8262, "step": 7897 }, { "epoch": 0.2860557768924303, "grad_norm": 1.749603056166988, "learning_rate": 1.675435235701807e-05, "loss": 0.8432, "step": 7898 }, { "epoch": 0.28609199565374865, "grad_norm": 1.4489291408124751, "learning_rate": 1.6753487267757624e-05, "loss": 0.8404, "step": 7899 }, { "epoch": 0.28612821441506703, "grad_norm": 1.5040958037375758, "learning_rate": 1.675262208556326e-05, "loss": 0.7791, "step": 7900 }, { "epoch": 0.28616443317638535, "grad_norm": 1.4631596720189386, "learning_rate": 1.6751756810446888e-05, "loss": 0.8048, "step": 7901 }, { "epoch": 0.2862006519377037, "grad_norm": 1.213893625252687, "learning_rate": 1.6750891442420404e-05, "loss": 0.798, "step": 7902 }, { "epoch": 0.2862368706990221, "grad_norm": 1.2215567265066354, "learning_rate": 1.6750025981495723e-05, "loss": 0.7969, "step": 7903 }, { "epoch": 0.28627308946034047, "grad_norm": 1.4313326773751465, "learning_rate": 1.6749160427684753e-05, "loss": 0.7726, "step": 7904 }, { "epoch": 0.28630930822165884, "grad_norm": 1.4750296492748642, "learning_rate": 1.6748294780999404e-05, "loss": 0.7972, "step": 7905 }, { "epoch": 0.28634552698297716, "grad_norm": 1.4359833527384716, "learning_rate": 1.674742904145159e-05, "loss": 0.7899, "step": 7906 }, { "epoch": 0.28638174574429553, "grad_norm": 1.4205051677526608, "learning_rate": 1.6746563209053226e-05, "loss": 0.7558, "step": 7907 }, { "epoch": 0.2864179645056139, "grad_norm": 1.3522060576519384, "learning_rate": 1.6745697283816224e-05, "loss": 0.799, "step": 7908 }, { "epoch": 0.2864541832669323, "grad_norm": 1.3503137556967164, "learning_rate": 1.67448312657525e-05, "loss": 0.7597, "step": 7909 }, { "epoch": 0.28649040202825066, "grad_norm": 1.4725587531784825, "learning_rate": 1.6743965154873966e-05, "loss": 0.8076, "step": 7910 }, { "epoch": 0.286526620789569, "grad_norm": 1.365177630369449, "learning_rate": 1.6743098951192555e-05, "loss": 0.809, "step": 7911 }, { "epoch": 0.28656283955088735, "grad_norm": 1.369872543098277, "learning_rate": 1.674223265472017e-05, "loss": 0.9155, "step": 7912 }, { "epoch": 0.2865990583122057, "grad_norm": 1.960347926657242, "learning_rate": 1.674136626546874e-05, "loss": 0.7391, "step": 7913 }, { "epoch": 0.2866352770735241, "grad_norm": 1.4488792908705934, "learning_rate": 1.674049978345019e-05, "loss": 0.836, "step": 7914 }, { "epoch": 0.28667149583484247, "grad_norm": 1.273243691353832, "learning_rate": 1.6739633208676438e-05, "loss": 0.7916, "step": 7915 }, { "epoch": 0.2867077145961608, "grad_norm": 1.3740230874974146, "learning_rate": 1.673876654115941e-05, "loss": 0.8299, "step": 7916 }, { "epoch": 0.28674393335747916, "grad_norm": 1.0683772577692972, "learning_rate": 1.6737899780911036e-05, "loss": 0.8535, "step": 7917 }, { "epoch": 0.28678015211879754, "grad_norm": 1.4046247720237839, "learning_rate": 1.673703292794324e-05, "loss": 0.8068, "step": 7918 }, { "epoch": 0.2868163708801159, "grad_norm": 1.402841251288553, "learning_rate": 1.6736165982267953e-05, "loss": 0.7856, "step": 7919 }, { "epoch": 0.2868525896414343, "grad_norm": 1.3397102859464431, "learning_rate": 1.6735298943897097e-05, "loss": 0.8169, "step": 7920 }, { "epoch": 0.2868888084027526, "grad_norm": 1.5045211798112708, "learning_rate": 1.6734431812842613e-05, "loss": 0.8127, "step": 7921 }, { "epoch": 0.286925027164071, "grad_norm": 1.2192760278551784, "learning_rate": 1.6733564589116433e-05, "loss": 0.786, "step": 7922 }, { "epoch": 0.28696124592538935, "grad_norm": 1.5028855166720485, "learning_rate": 1.6732697272730482e-05, "loss": 0.7933, "step": 7923 }, { "epoch": 0.2869974646867077, "grad_norm": 1.4561283385891168, "learning_rate": 1.6731829863696705e-05, "loss": 0.7736, "step": 7924 }, { "epoch": 0.2870336834480261, "grad_norm": 1.2558109924309386, "learning_rate": 1.6730962362027034e-05, "loss": 0.7976, "step": 7925 }, { "epoch": 0.2870699022093444, "grad_norm": 1.1450259392524464, "learning_rate": 1.6730094767733402e-05, "loss": 0.7579, "step": 7926 }, { "epoch": 0.2871061209706628, "grad_norm": 1.4313700746887585, "learning_rate": 1.6729227080827756e-05, "loss": 0.7808, "step": 7927 }, { "epoch": 0.28714233973198117, "grad_norm": 1.476233875954145, "learning_rate": 1.6728359301322032e-05, "loss": 0.7636, "step": 7928 }, { "epoch": 0.28717855849329954, "grad_norm": 1.4553350528999345, "learning_rate": 1.6727491429228173e-05, "loss": 0.7801, "step": 7929 }, { "epoch": 0.2872147772546179, "grad_norm": 1.2200499268985678, "learning_rate": 1.6726623464558116e-05, "loss": 0.7542, "step": 7930 }, { "epoch": 0.28725099601593623, "grad_norm": 1.3810247884682547, "learning_rate": 1.6725755407323814e-05, "loss": 0.8565, "step": 7931 }, { "epoch": 0.2872872147772546, "grad_norm": 1.5468397016062108, "learning_rate": 1.6724887257537207e-05, "loss": 0.8341, "step": 7932 }, { "epoch": 0.287323433538573, "grad_norm": 1.4187073924897031, "learning_rate": 1.6724019015210245e-05, "loss": 0.8104, "step": 7933 }, { "epoch": 0.28735965229989135, "grad_norm": 1.3799485253725463, "learning_rate": 1.672315068035487e-05, "loss": 0.833, "step": 7934 }, { "epoch": 0.28739587106120973, "grad_norm": 1.192697129776515, "learning_rate": 1.6722282252983034e-05, "loss": 0.7382, "step": 7935 }, { "epoch": 0.28743208982252805, "grad_norm": 1.5839733706655323, "learning_rate": 1.6721413733106685e-05, "loss": 0.7962, "step": 7936 }, { "epoch": 0.2874683085838464, "grad_norm": 1.4504201213869177, "learning_rate": 1.6720545120737784e-05, "loss": 0.8046, "step": 7937 }, { "epoch": 0.2875045273451648, "grad_norm": 1.687803397896599, "learning_rate": 1.671967641588827e-05, "loss": 0.8272, "step": 7938 }, { "epoch": 0.28754074610648317, "grad_norm": 1.1339975553685155, "learning_rate": 1.671880761857011e-05, "loss": 0.758, "step": 7939 }, { "epoch": 0.28757696486780154, "grad_norm": 1.3897160743503967, "learning_rate": 1.6717938728795247e-05, "loss": 0.8844, "step": 7940 }, { "epoch": 0.28761318362911986, "grad_norm": 1.4385050798633645, "learning_rate": 1.671706974657565e-05, "loss": 0.7485, "step": 7941 }, { "epoch": 0.28764940239043824, "grad_norm": 1.4786156679722093, "learning_rate": 1.671620067192327e-05, "loss": 0.8892, "step": 7942 }, { "epoch": 0.2876856211517566, "grad_norm": 1.406905747878997, "learning_rate": 1.6715331504850065e-05, "loss": 0.7838, "step": 7943 }, { "epoch": 0.287721839913075, "grad_norm": 1.1437364052315688, "learning_rate": 1.6714462245368e-05, "loss": 0.8378, "step": 7944 }, { "epoch": 0.28775805867439336, "grad_norm": 1.338163378680514, "learning_rate": 1.6713592893489032e-05, "loss": 0.8136, "step": 7945 }, { "epoch": 0.2877942774357117, "grad_norm": 1.4984079081179784, "learning_rate": 1.671272344922513e-05, "loss": 0.8028, "step": 7946 }, { "epoch": 0.28783049619703005, "grad_norm": 1.4617884906256993, "learning_rate": 1.6711853912588256e-05, "loss": 0.7607, "step": 7947 }, { "epoch": 0.2878667149583484, "grad_norm": 1.316259602378532, "learning_rate": 1.671098428359037e-05, "loss": 0.8109, "step": 7948 }, { "epoch": 0.2879029337196668, "grad_norm": 1.3752788781076257, "learning_rate": 1.6710114562243448e-05, "loss": 0.8474, "step": 7949 }, { "epoch": 0.28793915248098517, "grad_norm": 1.4415833033651049, "learning_rate": 1.6709244748559448e-05, "loss": 0.7811, "step": 7950 }, { "epoch": 0.2879753712423035, "grad_norm": 1.4648473230954144, "learning_rate": 1.670837484255035e-05, "loss": 0.8565, "step": 7951 }, { "epoch": 0.28801159000362186, "grad_norm": 1.3891878256028316, "learning_rate": 1.6707504844228116e-05, "loss": 0.7233, "step": 7952 }, { "epoch": 0.28804780876494024, "grad_norm": 1.399279221562395, "learning_rate": 1.6706634753604725e-05, "loss": 0.7572, "step": 7953 }, { "epoch": 0.2880840275262586, "grad_norm": 1.333060501091366, "learning_rate": 1.6705764570692148e-05, "loss": 0.8649, "step": 7954 }, { "epoch": 0.288120246287577, "grad_norm": 1.371055160968418, "learning_rate": 1.6704894295502355e-05, "loss": 0.7741, "step": 7955 }, { "epoch": 0.2881564650488953, "grad_norm": 1.230576247390177, "learning_rate": 1.6704023928047323e-05, "loss": 0.8302, "step": 7956 }, { "epoch": 0.2881926838102137, "grad_norm": 1.4209202505842853, "learning_rate": 1.6703153468339036e-05, "loss": 0.7344, "step": 7957 }, { "epoch": 0.28822890257153205, "grad_norm": 1.403110971678038, "learning_rate": 1.6702282916389464e-05, "loss": 0.7752, "step": 7958 }, { "epoch": 0.2882651213328504, "grad_norm": 1.5042117279231983, "learning_rate": 1.670141227221059e-05, "loss": 0.7676, "step": 7959 }, { "epoch": 0.2883013400941688, "grad_norm": 1.6084650809882497, "learning_rate": 1.6700541535814397e-05, "loss": 0.8196, "step": 7960 }, { "epoch": 0.2883375588554871, "grad_norm": 1.1508955412120891, "learning_rate": 1.6699670707212863e-05, "loss": 0.8312, "step": 7961 }, { "epoch": 0.2883737776168055, "grad_norm": 1.4251206905824978, "learning_rate": 1.6698799786417973e-05, "loss": 0.7821, "step": 7962 }, { "epoch": 0.28840999637812387, "grad_norm": 1.1705917272563162, "learning_rate": 1.6697928773441713e-05, "loss": 0.8357, "step": 7963 }, { "epoch": 0.28844621513944224, "grad_norm": 1.4249678309839686, "learning_rate": 1.669705766829607e-05, "loss": 0.765, "step": 7964 }, { "epoch": 0.2884824339007606, "grad_norm": 1.6281728946135707, "learning_rate": 1.6696186470993026e-05, "loss": 0.8002, "step": 7965 }, { "epoch": 0.28851865266207893, "grad_norm": 1.02667694154275, "learning_rate": 1.6695315181544572e-05, "loss": 0.7674, "step": 7966 }, { "epoch": 0.2885548714233973, "grad_norm": 1.5202552454567722, "learning_rate": 1.6694443799962698e-05, "loss": 0.8696, "step": 7967 }, { "epoch": 0.2885910901847157, "grad_norm": 1.3899061435859636, "learning_rate": 1.6693572326259397e-05, "loss": 0.8473, "step": 7968 }, { "epoch": 0.28862730894603406, "grad_norm": 1.4370384544350823, "learning_rate": 1.6692700760446658e-05, "loss": 0.7945, "step": 7969 }, { "epoch": 0.28866352770735243, "grad_norm": 1.1463826115962206, "learning_rate": 1.6691829102536477e-05, "loss": 0.7528, "step": 7970 }, { "epoch": 0.28869974646867075, "grad_norm": 1.5296513474257323, "learning_rate": 1.6690957352540848e-05, "loss": 0.7381, "step": 7971 }, { "epoch": 0.2887359652299891, "grad_norm": 1.4992029244142755, "learning_rate": 1.669008551047176e-05, "loss": 0.7479, "step": 7972 }, { "epoch": 0.2887721839913075, "grad_norm": 1.52668134723061, "learning_rate": 1.6689213576341226e-05, "loss": 0.8522, "step": 7973 }, { "epoch": 0.28880840275262587, "grad_norm": 1.467955453009524, "learning_rate": 1.668834155016123e-05, "loss": 0.7903, "step": 7974 }, { "epoch": 0.28884462151394424, "grad_norm": 1.4585271882457989, "learning_rate": 1.668746943194378e-05, "loss": 0.7747, "step": 7975 }, { "epoch": 0.28888084027526256, "grad_norm": 1.5023558092947527, "learning_rate": 1.6686597221700874e-05, "loss": 0.8342, "step": 7976 }, { "epoch": 0.28891705903658094, "grad_norm": 1.8310921667713724, "learning_rate": 1.6685724919444514e-05, "loss": 0.7722, "step": 7977 }, { "epoch": 0.2889532777978993, "grad_norm": 1.3171213037514193, "learning_rate": 1.6684852525186704e-05, "loss": 0.7573, "step": 7978 }, { "epoch": 0.2889894965592177, "grad_norm": 1.5266066623886703, "learning_rate": 1.668398003893945e-05, "loss": 0.8077, "step": 7979 }, { "epoch": 0.28902571532053606, "grad_norm": 1.5783427626433209, "learning_rate": 1.668310746071476e-05, "loss": 0.7668, "step": 7980 }, { "epoch": 0.2890619340818544, "grad_norm": 1.3296474720278895, "learning_rate": 1.668223479052464e-05, "loss": 0.7614, "step": 7981 }, { "epoch": 0.28909815284317275, "grad_norm": 1.3976940591036293, "learning_rate": 1.668136202838109e-05, "loss": 0.8389, "step": 7982 }, { "epoch": 0.2891343716044911, "grad_norm": 1.227268422836484, "learning_rate": 1.6680489174296135e-05, "loss": 0.804, "step": 7983 }, { "epoch": 0.2891705903658095, "grad_norm": 1.3073154269360436, "learning_rate": 1.6679616228281775e-05, "loss": 0.802, "step": 7984 }, { "epoch": 0.2892068091271279, "grad_norm": 1.3337590427318937, "learning_rate": 1.6678743190350032e-05, "loss": 0.8117, "step": 7985 }, { "epoch": 0.2892430278884462, "grad_norm": 1.5019534052735701, "learning_rate": 1.6677870060512908e-05, "loss": 0.8154, "step": 7986 }, { "epoch": 0.28927924664976457, "grad_norm": 1.5075517512962655, "learning_rate": 1.6676996838782425e-05, "loss": 0.7822, "step": 7987 }, { "epoch": 0.28931546541108294, "grad_norm": 1.313437510906104, "learning_rate": 1.6676123525170603e-05, "loss": 0.7571, "step": 7988 }, { "epoch": 0.2893516841724013, "grad_norm": 1.464038513749435, "learning_rate": 1.667525011968945e-05, "loss": 0.7258, "step": 7989 }, { "epoch": 0.2893879029337197, "grad_norm": 1.3781071595250365, "learning_rate": 1.6674376622350995e-05, "loss": 0.7905, "step": 7990 }, { "epoch": 0.289424121695038, "grad_norm": 1.3899248294649902, "learning_rate": 1.6673503033167253e-05, "loss": 0.8007, "step": 7991 }, { "epoch": 0.2894603404563564, "grad_norm": 1.50050925635224, "learning_rate": 1.6672629352150237e-05, "loss": 0.7689, "step": 7992 }, { "epoch": 0.28949655921767475, "grad_norm": 1.5495478700317913, "learning_rate": 1.6671755579311984e-05, "loss": 0.8567, "step": 7993 }, { "epoch": 0.2895327779789931, "grad_norm": 1.190850989916528, "learning_rate": 1.6670881714664515e-05, "loss": 0.7518, "step": 7994 }, { "epoch": 0.2895689967403115, "grad_norm": 1.5823317271855337, "learning_rate": 1.667000775821985e-05, "loss": 0.8603, "step": 7995 }, { "epoch": 0.2896052155016298, "grad_norm": 1.471236473555479, "learning_rate": 1.6669133709990016e-05, "loss": 0.8022, "step": 7996 }, { "epoch": 0.2896414342629482, "grad_norm": 1.182274086075959, "learning_rate": 1.6668259569987043e-05, "loss": 0.7405, "step": 7997 }, { "epoch": 0.28967765302426657, "grad_norm": 1.0846360803554773, "learning_rate": 1.6667385338222954e-05, "loss": 0.8229, "step": 7998 }, { "epoch": 0.28971387178558494, "grad_norm": 1.4832357514545098, "learning_rate": 1.6666511014709788e-05, "loss": 0.8093, "step": 7999 }, { "epoch": 0.2897500905469033, "grad_norm": 1.4217853976407466, "learning_rate": 1.6665636599459576e-05, "loss": 0.8348, "step": 8000 }, { "epoch": 0.28978630930822163, "grad_norm": 1.3352745130169856, "learning_rate": 1.6664762092484346e-05, "loss": 0.7885, "step": 8001 }, { "epoch": 0.28982252806954, "grad_norm": 1.4165889703929126, "learning_rate": 1.666388749379613e-05, "loss": 0.7722, "step": 8002 }, { "epoch": 0.2898587468308584, "grad_norm": 1.3161699757936183, "learning_rate": 1.6663012803406972e-05, "loss": 0.7934, "step": 8003 }, { "epoch": 0.28989496559217676, "grad_norm": 1.2140305191323726, "learning_rate": 1.66621380213289e-05, "loss": 0.7593, "step": 8004 }, { "epoch": 0.28993118435349513, "grad_norm": 1.5456337162925873, "learning_rate": 1.6661263147573954e-05, "loss": 0.7993, "step": 8005 }, { "epoch": 0.28996740311481345, "grad_norm": 1.5113418199131177, "learning_rate": 1.6660388182154176e-05, "loss": 0.8145, "step": 8006 }, { "epoch": 0.2900036218761318, "grad_norm": 1.4517517336827575, "learning_rate": 1.66595131250816e-05, "loss": 0.7556, "step": 8007 }, { "epoch": 0.2900398406374502, "grad_norm": 1.5805930181648784, "learning_rate": 1.6658637976368276e-05, "loss": 0.8137, "step": 8008 }, { "epoch": 0.29007605939876857, "grad_norm": 0.9999693916151486, "learning_rate": 1.6657762736026245e-05, "loss": 0.6828, "step": 8009 }, { "epoch": 0.29011227816008694, "grad_norm": 1.5074887743495553, "learning_rate": 1.6656887404067544e-05, "loss": 0.7644, "step": 8010 }, { "epoch": 0.29014849692140526, "grad_norm": 1.4773665886590783, "learning_rate": 1.6656011980504226e-05, "loss": 0.8054, "step": 8011 }, { "epoch": 0.29018471568272364, "grad_norm": 1.5335769483089952, "learning_rate": 1.6655136465348332e-05, "loss": 0.792, "step": 8012 }, { "epoch": 0.290220934444042, "grad_norm": 1.146810551320691, "learning_rate": 1.6654260858611913e-05, "loss": 0.7652, "step": 8013 }, { "epoch": 0.2902571532053604, "grad_norm": 1.523644633761356, "learning_rate": 1.665338516030702e-05, "loss": 0.8305, "step": 8014 }, { "epoch": 0.29029337196667876, "grad_norm": 1.1496624210350768, "learning_rate": 1.66525093704457e-05, "loss": 0.7301, "step": 8015 }, { "epoch": 0.2903295907279971, "grad_norm": 1.5576249113459402, "learning_rate": 1.6651633489040003e-05, "loss": 0.7799, "step": 8016 }, { "epoch": 0.29036580948931545, "grad_norm": 1.4556345707731058, "learning_rate": 1.6650757516101988e-05, "loss": 0.823, "step": 8017 }, { "epoch": 0.2904020282506338, "grad_norm": 1.0624997802078495, "learning_rate": 1.6649881451643706e-05, "loss": 0.7763, "step": 8018 }, { "epoch": 0.2904382470119522, "grad_norm": 1.3077396366827794, "learning_rate": 1.6649005295677206e-05, "loss": 0.7853, "step": 8019 }, { "epoch": 0.2904744657732706, "grad_norm": 1.4691876472123035, "learning_rate": 1.664812904821456e-05, "loss": 0.8778, "step": 8020 }, { "epoch": 0.2905106845345889, "grad_norm": 1.7508604284559675, "learning_rate": 1.664725270926781e-05, "loss": 0.7689, "step": 8021 }, { "epoch": 0.29054690329590727, "grad_norm": 1.1555735519911658, "learning_rate": 1.6646376278849024e-05, "loss": 0.8359, "step": 8022 }, { "epoch": 0.29058312205722564, "grad_norm": 1.4424625780162013, "learning_rate": 1.664549975697026e-05, "loss": 0.7562, "step": 8023 }, { "epoch": 0.290619340818544, "grad_norm": 1.1985751665207245, "learning_rate": 1.664462314364358e-05, "loss": 0.798, "step": 8024 }, { "epoch": 0.2906555595798624, "grad_norm": 1.5510472434684235, "learning_rate": 1.664374643888105e-05, "loss": 0.7821, "step": 8025 }, { "epoch": 0.2906917783411807, "grad_norm": 1.4161668354722763, "learning_rate": 1.664286964269473e-05, "loss": 0.7457, "step": 8026 }, { "epoch": 0.2907279971024991, "grad_norm": 1.486109362830959, "learning_rate": 1.6641992755096685e-05, "loss": 0.7875, "step": 8027 }, { "epoch": 0.29076421586381745, "grad_norm": 1.3025621632071787, "learning_rate": 1.6641115776098983e-05, "loss": 0.7648, "step": 8028 }, { "epoch": 0.29080043462513583, "grad_norm": 1.3700465077381803, "learning_rate": 1.6640238705713693e-05, "loss": 0.822, "step": 8029 }, { "epoch": 0.2908366533864542, "grad_norm": 1.4674103009114223, "learning_rate": 1.6639361543952885e-05, "loss": 0.7486, "step": 8030 }, { "epoch": 0.2908728721477725, "grad_norm": 1.5546857843437507, "learning_rate": 1.663848429082863e-05, "loss": 0.7239, "step": 8031 }, { "epoch": 0.2909090909090909, "grad_norm": 1.5072085401350537, "learning_rate": 1.6637606946352996e-05, "loss": 0.7669, "step": 8032 }, { "epoch": 0.29094530967040927, "grad_norm": 1.6485744553324393, "learning_rate": 1.6636729510538056e-05, "loss": 0.8052, "step": 8033 }, { "epoch": 0.29098152843172764, "grad_norm": 1.1717296635884853, "learning_rate": 1.6635851983395887e-05, "loss": 0.8112, "step": 8034 }, { "epoch": 0.291017747193046, "grad_norm": 1.086233697235636, "learning_rate": 1.6634974364938566e-05, "loss": 0.7613, "step": 8035 }, { "epoch": 0.29105396595436434, "grad_norm": 1.537899086516892, "learning_rate": 1.6634096655178168e-05, "loss": 0.74, "step": 8036 }, { "epoch": 0.2910901847156827, "grad_norm": 1.449674649003802, "learning_rate": 1.6633218854126768e-05, "loss": 0.6945, "step": 8037 }, { "epoch": 0.2911264034770011, "grad_norm": 1.4912160962008698, "learning_rate": 1.6632340961796455e-05, "loss": 0.7786, "step": 8038 }, { "epoch": 0.29116262223831946, "grad_norm": 1.4996396714787203, "learning_rate": 1.6631462978199297e-05, "loss": 0.7488, "step": 8039 }, { "epoch": 0.29119884099963783, "grad_norm": 1.4637648805644552, "learning_rate": 1.663058490334738e-05, "loss": 0.8723, "step": 8040 }, { "epoch": 0.29123505976095615, "grad_norm": 1.5180270047870443, "learning_rate": 1.6629706737252792e-05, "loss": 0.8274, "step": 8041 }, { "epoch": 0.2912712785222745, "grad_norm": 1.4416961236300492, "learning_rate": 1.6628828479927613e-05, "loss": 0.7296, "step": 8042 }, { "epoch": 0.2913074972835929, "grad_norm": 1.1313227196761797, "learning_rate": 1.6627950131383928e-05, "loss": 0.8175, "step": 8043 }, { "epoch": 0.29134371604491127, "grad_norm": 1.7856038138212085, "learning_rate": 1.6627071691633826e-05, "loss": 0.8521, "step": 8044 }, { "epoch": 0.29137993480622965, "grad_norm": 1.462566842840064, "learning_rate": 1.6626193160689395e-05, "loss": 0.7966, "step": 8045 }, { "epoch": 0.29141615356754796, "grad_norm": 1.4304451797789388, "learning_rate": 1.6625314538562725e-05, "loss": 0.8126, "step": 8046 }, { "epoch": 0.29145237232886634, "grad_norm": 1.4456849262748477, "learning_rate": 1.6624435825265903e-05, "loss": 0.8114, "step": 8047 }, { "epoch": 0.2914885910901847, "grad_norm": 1.4873827640130601, "learning_rate": 1.6623557020811024e-05, "loss": 0.7718, "step": 8048 }, { "epoch": 0.2915248098515031, "grad_norm": 1.5091849255288883, "learning_rate": 1.662267812521018e-05, "loss": 0.8286, "step": 8049 }, { "epoch": 0.29156102861282146, "grad_norm": 1.4907176551522547, "learning_rate": 1.6621799138475468e-05, "loss": 0.8504, "step": 8050 }, { "epoch": 0.2915972473741398, "grad_norm": 1.2344546261249212, "learning_rate": 1.662092006061898e-05, "loss": 0.7802, "step": 8051 }, { "epoch": 0.29163346613545815, "grad_norm": 1.383010574610656, "learning_rate": 1.6620040891652816e-05, "loss": 0.7943, "step": 8052 }, { "epoch": 0.2916696848967765, "grad_norm": 1.363764193326704, "learning_rate": 1.6619161631589072e-05, "loss": 0.7985, "step": 8053 }, { "epoch": 0.2917059036580949, "grad_norm": 1.3276325027943237, "learning_rate": 1.661828228043985e-05, "loss": 0.7787, "step": 8054 }, { "epoch": 0.2917421224194133, "grad_norm": 1.3436575742750476, "learning_rate": 1.6617402838217242e-05, "loss": 0.7905, "step": 8055 }, { "epoch": 0.2917783411807316, "grad_norm": 1.5493421363076596, "learning_rate": 1.661652330493336e-05, "loss": 0.9163, "step": 8056 }, { "epoch": 0.29181455994204997, "grad_norm": 1.3008785232911075, "learning_rate": 1.6615643680600306e-05, "loss": 0.7861, "step": 8057 }, { "epoch": 0.29185077870336834, "grad_norm": 1.3398480317962114, "learning_rate": 1.6614763965230175e-05, "loss": 0.8418, "step": 8058 }, { "epoch": 0.2918869974646867, "grad_norm": 1.5147611058410884, "learning_rate": 1.6613884158835086e-05, "loss": 0.8513, "step": 8059 }, { "epoch": 0.2919232162260051, "grad_norm": 1.8925520034686083, "learning_rate": 1.6613004261427137e-05, "loss": 0.7582, "step": 8060 }, { "epoch": 0.2919594349873234, "grad_norm": 1.3377690287833806, "learning_rate": 1.661212427301844e-05, "loss": 0.7991, "step": 8061 }, { "epoch": 0.2919956537486418, "grad_norm": 1.3804978142816113, "learning_rate": 1.66112441936211e-05, "loss": 0.775, "step": 8062 }, { "epoch": 0.29203187250996016, "grad_norm": 1.483348313188547, "learning_rate": 1.6610364023247234e-05, "loss": 0.8339, "step": 8063 }, { "epoch": 0.29206809127127853, "grad_norm": 1.405400485829543, "learning_rate": 1.660948376190895e-05, "loss": 0.876, "step": 8064 }, { "epoch": 0.2921043100325969, "grad_norm": 1.317336706241191, "learning_rate": 1.660860340961836e-05, "loss": 0.7904, "step": 8065 }, { "epoch": 0.2921405287939152, "grad_norm": 1.6361397331303633, "learning_rate": 1.6607722966387584e-05, "loss": 0.8759, "step": 8066 }, { "epoch": 0.2921767475552336, "grad_norm": 1.4406748029702008, "learning_rate": 1.660684243222873e-05, "loss": 0.8356, "step": 8067 }, { "epoch": 0.29221296631655197, "grad_norm": 1.4857198708769939, "learning_rate": 1.660596180715392e-05, "loss": 0.8085, "step": 8068 }, { "epoch": 0.29224918507787034, "grad_norm": 1.373432459350033, "learning_rate": 1.6605081091175273e-05, "loss": 0.7707, "step": 8069 }, { "epoch": 0.2922854038391887, "grad_norm": 1.375556701025568, "learning_rate": 1.6604200284304904e-05, "loss": 0.8015, "step": 8070 }, { "epoch": 0.29232162260050704, "grad_norm": 1.461322603620608, "learning_rate": 1.6603319386554937e-05, "loss": 0.7896, "step": 8071 }, { "epoch": 0.2923578413618254, "grad_norm": 1.3500712830604444, "learning_rate": 1.660243839793749e-05, "loss": 0.7578, "step": 8072 }, { "epoch": 0.2923940601231438, "grad_norm": 1.3833515600248814, "learning_rate": 1.6601557318464696e-05, "loss": 0.8386, "step": 8073 }, { "epoch": 0.29243027888446216, "grad_norm": 1.197891049779561, "learning_rate": 1.6600676148148667e-05, "loss": 0.7844, "step": 8074 }, { "epoch": 0.29246649764578053, "grad_norm": 1.408477913795316, "learning_rate": 1.6599794887001535e-05, "loss": 0.7585, "step": 8075 }, { "epoch": 0.29250271640709885, "grad_norm": 1.4667630280603596, "learning_rate": 1.659891353503543e-05, "loss": 0.8135, "step": 8076 }, { "epoch": 0.2925389351684172, "grad_norm": 1.3737980502516127, "learning_rate": 1.659803209226247e-05, "loss": 0.8386, "step": 8077 }, { "epoch": 0.2925751539297356, "grad_norm": 1.4660472897454788, "learning_rate": 1.6597150558694798e-05, "loss": 0.8121, "step": 8078 }, { "epoch": 0.292611372691054, "grad_norm": 1.369313653577705, "learning_rate": 1.6596268934344536e-05, "loss": 0.7635, "step": 8079 }, { "epoch": 0.29264759145237235, "grad_norm": 1.4879853293568672, "learning_rate": 1.6595387219223814e-05, "loss": 0.777, "step": 8080 }, { "epoch": 0.29268381021369066, "grad_norm": 1.598403713842501, "learning_rate": 1.6594505413344772e-05, "loss": 0.8839, "step": 8081 }, { "epoch": 0.29272002897500904, "grad_norm": 1.144327789311569, "learning_rate": 1.659362351671954e-05, "loss": 0.8271, "step": 8082 }, { "epoch": 0.2927562477363274, "grad_norm": 1.3474127535672453, "learning_rate": 1.6592741529360254e-05, "loss": 0.7419, "step": 8083 }, { "epoch": 0.2927924664976458, "grad_norm": 1.1676385439253296, "learning_rate": 1.6591859451279053e-05, "loss": 0.8156, "step": 8084 }, { "epoch": 0.29282868525896416, "grad_norm": 1.3389513829564383, "learning_rate": 1.659097728248807e-05, "loss": 0.768, "step": 8085 }, { "epoch": 0.2928649040202825, "grad_norm": 1.4735379501357748, "learning_rate": 1.6590095022999454e-05, "loss": 0.8539, "step": 8086 }, { "epoch": 0.29290112278160085, "grad_norm": 1.5156452371779963, "learning_rate": 1.6589212672825334e-05, "loss": 0.824, "step": 8087 }, { "epoch": 0.2929373415429192, "grad_norm": 1.3582535496071628, "learning_rate": 1.6588330231977862e-05, "loss": 0.8459, "step": 8088 }, { "epoch": 0.2929735603042376, "grad_norm": 1.411413880178401, "learning_rate": 1.6587447700469178e-05, "loss": 0.8193, "step": 8089 }, { "epoch": 0.293009779065556, "grad_norm": 1.2930735256781836, "learning_rate": 1.6586565078311424e-05, "loss": 0.7569, "step": 8090 }, { "epoch": 0.2930459978268743, "grad_norm": 1.4033901307398533, "learning_rate": 1.6585682365516743e-05, "loss": 0.794, "step": 8091 }, { "epoch": 0.29308221658819267, "grad_norm": 1.5256087840739028, "learning_rate": 1.6584799562097293e-05, "loss": 0.7776, "step": 8092 }, { "epoch": 0.29311843534951104, "grad_norm": 1.492574011341008, "learning_rate": 1.658391666806521e-05, "loss": 0.8425, "step": 8093 }, { "epoch": 0.2931546541108294, "grad_norm": 1.3181794739649062, "learning_rate": 1.658303368343265e-05, "loss": 0.7905, "step": 8094 }, { "epoch": 0.2931908728721478, "grad_norm": 1.3356197834058057, "learning_rate": 1.658215060821176e-05, "loss": 0.8075, "step": 8095 }, { "epoch": 0.2932270916334661, "grad_norm": 1.5293477160795363, "learning_rate": 1.6581267442414697e-05, "loss": 0.9766, "step": 8096 }, { "epoch": 0.2932633103947845, "grad_norm": 1.5605760053606137, "learning_rate": 1.658038418605361e-05, "loss": 0.8149, "step": 8097 }, { "epoch": 0.29329952915610286, "grad_norm": 1.0844342578340185, "learning_rate": 1.6579500839140656e-05, "loss": 0.7367, "step": 8098 }, { "epoch": 0.29333574791742123, "grad_norm": 1.4477380175948935, "learning_rate": 1.6578617401687988e-05, "loss": 0.7659, "step": 8099 }, { "epoch": 0.2933719666787396, "grad_norm": 1.4755094055182532, "learning_rate": 1.6577733873707762e-05, "loss": 0.859, "step": 8100 }, { "epoch": 0.2934081854400579, "grad_norm": 1.4695600092100236, "learning_rate": 1.6576850255212142e-05, "loss": 0.8056, "step": 8101 }, { "epoch": 0.2934444042013763, "grad_norm": 1.4668636998879878, "learning_rate": 1.657596654621328e-05, "loss": 0.8299, "step": 8102 }, { "epoch": 0.29348062296269467, "grad_norm": 1.4844313510438008, "learning_rate": 1.6575082746723342e-05, "loss": 0.8232, "step": 8103 }, { "epoch": 0.29351684172401304, "grad_norm": 1.5217072564801557, "learning_rate": 1.657419885675449e-05, "loss": 0.813, "step": 8104 }, { "epoch": 0.2935530604853314, "grad_norm": 1.5365863307612324, "learning_rate": 1.6573314876318885e-05, "loss": 0.7879, "step": 8105 }, { "epoch": 0.29358927924664974, "grad_norm": 1.5928512474714787, "learning_rate": 1.6572430805428687e-05, "loss": 0.7557, "step": 8106 }, { "epoch": 0.2936254980079681, "grad_norm": 1.409783357628684, "learning_rate": 1.657154664409607e-05, "loss": 0.8504, "step": 8107 }, { "epoch": 0.2936617167692865, "grad_norm": 1.592845429805482, "learning_rate": 1.6570662392333196e-05, "loss": 0.8223, "step": 8108 }, { "epoch": 0.29369793553060486, "grad_norm": 1.4086563899350752, "learning_rate": 1.6569778050152233e-05, "loss": 0.7235, "step": 8109 }, { "epoch": 0.29373415429192323, "grad_norm": 1.367755913421309, "learning_rate": 1.6568893617565353e-05, "loss": 0.7887, "step": 8110 }, { "epoch": 0.2937703730532416, "grad_norm": 1.4529213852407374, "learning_rate": 1.6568009094584723e-05, "loss": 0.8331, "step": 8111 }, { "epoch": 0.2938065918145599, "grad_norm": 1.3445745934942743, "learning_rate": 1.6567124481222516e-05, "loss": 0.7749, "step": 8112 }, { "epoch": 0.2938428105758783, "grad_norm": 1.123382017258752, "learning_rate": 1.656623977749091e-05, "loss": 0.753, "step": 8113 }, { "epoch": 0.2938790293371967, "grad_norm": 1.3593138874227444, "learning_rate": 1.6565354983402074e-05, "loss": 0.8248, "step": 8114 }, { "epoch": 0.29391524809851505, "grad_norm": 1.1995893500439823, "learning_rate": 1.656447009896818e-05, "loss": 0.7645, "step": 8115 }, { "epoch": 0.2939514668598334, "grad_norm": 1.9987151541132269, "learning_rate": 1.6563585124201414e-05, "loss": 0.8438, "step": 8116 }, { "epoch": 0.29398768562115174, "grad_norm": 1.4033146780454413, "learning_rate": 1.6562700059113948e-05, "loss": 0.7778, "step": 8117 }, { "epoch": 0.2940239043824701, "grad_norm": 1.540658123673917, "learning_rate": 1.656181490371796e-05, "loss": 0.8077, "step": 8118 }, { "epoch": 0.2940601231437885, "grad_norm": 1.6321358465055016, "learning_rate": 1.6560929658025635e-05, "loss": 0.8374, "step": 8119 }, { "epoch": 0.29409634190510686, "grad_norm": 1.4194650708865908, "learning_rate": 1.6560044322049154e-05, "loss": 0.7869, "step": 8120 }, { "epoch": 0.29413256066642524, "grad_norm": 1.4461749210174122, "learning_rate": 1.65591588958007e-05, "loss": 0.7825, "step": 8121 }, { "epoch": 0.29416877942774355, "grad_norm": 1.467310258139685, "learning_rate": 1.6558273379292456e-05, "loss": 0.7802, "step": 8122 }, { "epoch": 0.29420499818906193, "grad_norm": 1.5527823278831845, "learning_rate": 1.6557387772536606e-05, "loss": 0.809, "step": 8123 }, { "epoch": 0.2942412169503803, "grad_norm": 1.4600888390795785, "learning_rate": 1.655650207554534e-05, "loss": 0.8674, "step": 8124 }, { "epoch": 0.2942774357116987, "grad_norm": 1.2988780607218287, "learning_rate": 1.6555616288330844e-05, "loss": 0.7663, "step": 8125 }, { "epoch": 0.29431365447301705, "grad_norm": 1.472645782312078, "learning_rate": 1.6554730410905303e-05, "loss": 0.8173, "step": 8126 }, { "epoch": 0.29434987323433537, "grad_norm": 1.4365363434886713, "learning_rate": 1.655384444328092e-05, "loss": 0.8302, "step": 8127 }, { "epoch": 0.29438609199565374, "grad_norm": 1.3594245072645332, "learning_rate": 1.6552958385469872e-05, "loss": 0.7374, "step": 8128 }, { "epoch": 0.2944223107569721, "grad_norm": 1.0881534592837436, "learning_rate": 1.6552072237484363e-05, "loss": 0.809, "step": 8129 }, { "epoch": 0.2944585295182905, "grad_norm": 1.1232033717436338, "learning_rate": 1.655118599933658e-05, "loss": 0.7659, "step": 8130 }, { "epoch": 0.29449474827960886, "grad_norm": 1.4563691709785427, "learning_rate": 1.6550299671038722e-05, "loss": 0.7313, "step": 8131 }, { "epoch": 0.2945309670409272, "grad_norm": 1.4965056778720336, "learning_rate": 1.654941325260299e-05, "loss": 0.7508, "step": 8132 }, { "epoch": 0.29456718580224556, "grad_norm": 1.4012305109981054, "learning_rate": 1.654852674404157e-05, "loss": 0.7452, "step": 8133 }, { "epoch": 0.29460340456356393, "grad_norm": 1.2124032501819555, "learning_rate": 1.654764014536667e-05, "loss": 0.7856, "step": 8134 }, { "epoch": 0.2946396233248823, "grad_norm": 1.2052211792972818, "learning_rate": 1.6546753456590488e-05, "loss": 0.7586, "step": 8135 }, { "epoch": 0.2946758420862007, "grad_norm": 1.5600422447912015, "learning_rate": 1.6545866677725227e-05, "loss": 0.7988, "step": 8136 }, { "epoch": 0.294712060847519, "grad_norm": 1.571780337200176, "learning_rate": 1.6544979808783092e-05, "loss": 0.7462, "step": 8137 }, { "epoch": 0.29474827960883737, "grad_norm": 1.5575927194329762, "learning_rate": 1.6544092849776278e-05, "loss": 0.8424, "step": 8138 }, { "epoch": 0.29478449837015575, "grad_norm": 1.5249215746866138, "learning_rate": 1.6543205800716997e-05, "loss": 0.7797, "step": 8139 }, { "epoch": 0.2948207171314741, "grad_norm": 1.3932425403992923, "learning_rate": 1.6542318661617457e-05, "loss": 0.7559, "step": 8140 }, { "epoch": 0.2948569358927925, "grad_norm": 1.243003045880239, "learning_rate": 1.654143143248986e-05, "loss": 0.7726, "step": 8141 }, { "epoch": 0.2948931546541108, "grad_norm": 1.4415299768250192, "learning_rate": 1.6540544113346424e-05, "loss": 0.8217, "step": 8142 }, { "epoch": 0.2949293734154292, "grad_norm": 1.5008588805566943, "learning_rate": 1.6539656704199353e-05, "loss": 0.8055, "step": 8143 }, { "epoch": 0.29496559217674756, "grad_norm": 1.5209054800507114, "learning_rate": 1.6538769205060856e-05, "loss": 0.7034, "step": 8144 }, { "epoch": 0.29500181093806593, "grad_norm": 1.4308193539128813, "learning_rate": 1.6537881615943153e-05, "loss": 0.8359, "step": 8145 }, { "epoch": 0.2950380296993843, "grad_norm": 1.4115922318592036, "learning_rate": 1.6536993936858455e-05, "loss": 0.8659, "step": 8146 }, { "epoch": 0.2950742484607026, "grad_norm": 1.4342073011447856, "learning_rate": 1.6536106167818974e-05, "loss": 0.7594, "step": 8147 }, { "epoch": 0.295110467222021, "grad_norm": 1.4330066640349373, "learning_rate": 1.6535218308836927e-05, "loss": 0.7722, "step": 8148 }, { "epoch": 0.2951466859833394, "grad_norm": 1.0918125883346737, "learning_rate": 1.6534330359924537e-05, "loss": 0.8129, "step": 8149 }, { "epoch": 0.29518290474465775, "grad_norm": 1.4764053624417222, "learning_rate": 1.6533442321094017e-05, "loss": 0.7879, "step": 8150 }, { "epoch": 0.2952191235059761, "grad_norm": 1.326622199153309, "learning_rate": 1.6532554192357593e-05, "loss": 0.7615, "step": 8151 }, { "epoch": 0.29525534226729444, "grad_norm": 1.2311409849122625, "learning_rate": 1.653166597372748e-05, "loss": 0.7898, "step": 8152 }, { "epoch": 0.2952915610286128, "grad_norm": 1.3562816839758216, "learning_rate": 1.6530777665215907e-05, "loss": 0.7474, "step": 8153 }, { "epoch": 0.2953277797899312, "grad_norm": 1.291338916003667, "learning_rate": 1.6529889266835093e-05, "loss": 0.7954, "step": 8154 }, { "epoch": 0.29536399855124956, "grad_norm": 1.5238170138836271, "learning_rate": 1.6529000778597266e-05, "loss": 0.9054, "step": 8155 }, { "epoch": 0.29540021731256794, "grad_norm": 1.3204241922638462, "learning_rate": 1.6528112200514652e-05, "loss": 0.784, "step": 8156 }, { "epoch": 0.29543643607388625, "grad_norm": 1.4510816074960848, "learning_rate": 1.652722353259948e-05, "loss": 0.8542, "step": 8157 }, { "epoch": 0.29547265483520463, "grad_norm": 1.3433533398593056, "learning_rate": 1.6526334774863973e-05, "loss": 0.7656, "step": 8158 }, { "epoch": 0.295508873596523, "grad_norm": 1.462296274761438, "learning_rate": 1.6525445927320367e-05, "loss": 0.8399, "step": 8159 }, { "epoch": 0.2955450923578414, "grad_norm": 1.4549492951780276, "learning_rate": 1.6524556989980888e-05, "loss": 0.8256, "step": 8160 }, { "epoch": 0.29558131111915975, "grad_norm": 1.3926435437827078, "learning_rate": 1.6523667962857776e-05, "loss": 0.7719, "step": 8161 }, { "epoch": 0.29561752988047807, "grad_norm": 1.4353210505150493, "learning_rate": 1.652277884596326e-05, "loss": 0.7719, "step": 8162 }, { "epoch": 0.29565374864179644, "grad_norm": 1.3193376860194053, "learning_rate": 1.652188963930957e-05, "loss": 0.7669, "step": 8163 }, { "epoch": 0.2956899674031148, "grad_norm": 1.4476325217400672, "learning_rate": 1.6521000342908956e-05, "loss": 0.8334, "step": 8164 }, { "epoch": 0.2957261861644332, "grad_norm": 1.071427111654834, "learning_rate": 1.6520110956773643e-05, "loss": 0.8297, "step": 8165 }, { "epoch": 0.29576240492575157, "grad_norm": 1.5043289147974057, "learning_rate": 1.6519221480915875e-05, "loss": 0.8471, "step": 8166 }, { "epoch": 0.2957986236870699, "grad_norm": 1.3650230515665844, "learning_rate": 1.6518331915347888e-05, "loss": 0.9114, "step": 8167 }, { "epoch": 0.29583484244838826, "grad_norm": 1.5032139770142265, "learning_rate": 1.651744226008193e-05, "loss": 0.7683, "step": 8168 }, { "epoch": 0.29587106120970663, "grad_norm": 1.5045327466984415, "learning_rate": 1.651655251513024e-05, "loss": 0.822, "step": 8169 }, { "epoch": 0.295907279971025, "grad_norm": 1.4775384006637433, "learning_rate": 1.651566268050506e-05, "loss": 0.833, "step": 8170 }, { "epoch": 0.2959434987323434, "grad_norm": 1.4324544691991625, "learning_rate": 1.651477275621864e-05, "loss": 0.8878, "step": 8171 }, { "epoch": 0.2959797174936617, "grad_norm": 1.319881522641944, "learning_rate": 1.6513882742283217e-05, "loss": 0.8608, "step": 8172 }, { "epoch": 0.29601593625498007, "grad_norm": 1.4439738698679487, "learning_rate": 1.6512992638711046e-05, "loss": 0.8224, "step": 8173 }, { "epoch": 0.29605215501629845, "grad_norm": 1.3996099852662562, "learning_rate": 1.6512102445514376e-05, "loss": 0.8133, "step": 8174 }, { "epoch": 0.2960883737776168, "grad_norm": 1.5054920802463945, "learning_rate": 1.651121216270545e-05, "loss": 0.8005, "step": 8175 }, { "epoch": 0.2961245925389352, "grad_norm": 1.155228430078058, "learning_rate": 1.6510321790296527e-05, "loss": 0.7579, "step": 8176 }, { "epoch": 0.2961608113002535, "grad_norm": 1.5354834467544667, "learning_rate": 1.6509431328299853e-05, "loss": 0.8158, "step": 8177 }, { "epoch": 0.2961970300615719, "grad_norm": 1.459893016447284, "learning_rate": 1.6508540776727687e-05, "loss": 0.7986, "step": 8178 }, { "epoch": 0.29623324882289026, "grad_norm": 1.00704256461818, "learning_rate": 1.650765013559228e-05, "loss": 0.7315, "step": 8179 }, { "epoch": 0.29626946758420863, "grad_norm": 1.3069525917905418, "learning_rate": 1.6506759404905886e-05, "loss": 0.8037, "step": 8180 }, { "epoch": 0.296305686345527, "grad_norm": 1.4487411188776214, "learning_rate": 1.6505868584680768e-05, "loss": 0.8459, "step": 8181 }, { "epoch": 0.2963419051068453, "grad_norm": 1.7246813170482398, "learning_rate": 1.6504977674929184e-05, "loss": 0.771, "step": 8182 }, { "epoch": 0.2963781238681637, "grad_norm": 1.0492950244121204, "learning_rate": 1.6504086675663388e-05, "loss": 0.7581, "step": 8183 }, { "epoch": 0.2964143426294821, "grad_norm": 1.6698951183651733, "learning_rate": 1.6503195586895646e-05, "loss": 0.7944, "step": 8184 }, { "epoch": 0.29645056139080045, "grad_norm": 1.0957765097786187, "learning_rate": 1.6502304408638215e-05, "loss": 0.7794, "step": 8185 }, { "epoch": 0.2964867801521188, "grad_norm": 1.3407626997481639, "learning_rate": 1.6501413140903365e-05, "loss": 0.7686, "step": 8186 }, { "epoch": 0.29652299891343714, "grad_norm": 1.2006350133091923, "learning_rate": 1.650052178370336e-05, "loss": 0.7915, "step": 8187 }, { "epoch": 0.2965592176747555, "grad_norm": 1.4478537874038748, "learning_rate": 1.649963033705046e-05, "loss": 0.842, "step": 8188 }, { "epoch": 0.2965954364360739, "grad_norm": 1.3900244044069314, "learning_rate": 1.6498738800956936e-05, "loss": 0.8488, "step": 8189 }, { "epoch": 0.29663165519739226, "grad_norm": 1.2092643869564303, "learning_rate": 1.6497847175435057e-05, "loss": 0.6947, "step": 8190 }, { "epoch": 0.29666787395871064, "grad_norm": 1.4216791642957527, "learning_rate": 1.6496955460497087e-05, "loss": 0.8538, "step": 8191 }, { "epoch": 0.29670409272002896, "grad_norm": 1.3889784172793302, "learning_rate": 1.6496063656155304e-05, "loss": 0.8143, "step": 8192 }, { "epoch": 0.29674031148134733, "grad_norm": 1.558360632245312, "learning_rate": 1.649517176242198e-05, "loss": 0.8222, "step": 8193 }, { "epoch": 0.2967765302426657, "grad_norm": 1.648998781892137, "learning_rate": 1.6494279779309385e-05, "loss": 0.893, "step": 8194 }, { "epoch": 0.2968127490039841, "grad_norm": 1.5302793701095827, "learning_rate": 1.649338770682979e-05, "loss": 0.8019, "step": 8195 }, { "epoch": 0.29684896776530245, "grad_norm": 1.4753184830387311, "learning_rate": 1.6492495544995474e-05, "loss": 0.8158, "step": 8196 }, { "epoch": 0.29688518652662077, "grad_norm": 1.5886482716315433, "learning_rate": 1.649160329381872e-05, "loss": 0.8695, "step": 8197 }, { "epoch": 0.29692140528793914, "grad_norm": 1.7654708630452711, "learning_rate": 1.6490710953311796e-05, "loss": 0.8964, "step": 8198 }, { "epoch": 0.2969576240492575, "grad_norm": 1.5990574010875476, "learning_rate": 1.6489818523486987e-05, "loss": 0.7959, "step": 8199 }, { "epoch": 0.2969938428105759, "grad_norm": 1.4019626264563725, "learning_rate": 1.6488926004356573e-05, "loss": 0.8649, "step": 8200 }, { "epoch": 0.29703006157189427, "grad_norm": 1.43187630692347, "learning_rate": 1.6488033395932838e-05, "loss": 0.7989, "step": 8201 }, { "epoch": 0.2970662803332126, "grad_norm": 1.443752232127714, "learning_rate": 1.648714069822806e-05, "loss": 0.8252, "step": 8202 }, { "epoch": 0.29710249909453096, "grad_norm": 1.4052651559507485, "learning_rate": 1.6486247911254525e-05, "loss": 0.8271, "step": 8203 }, { "epoch": 0.29713871785584933, "grad_norm": 1.3970861878807013, "learning_rate": 1.648535503502452e-05, "loss": 0.7985, "step": 8204 }, { "epoch": 0.2971749366171677, "grad_norm": 1.5779496510977218, "learning_rate": 1.648446206955033e-05, "loss": 0.7292, "step": 8205 }, { "epoch": 0.2972111553784861, "grad_norm": 1.0544599167225348, "learning_rate": 1.6483569014844247e-05, "loss": 0.7375, "step": 8206 }, { "epoch": 0.2972473741398044, "grad_norm": 1.4448961096304926, "learning_rate": 1.6482675870918557e-05, "loss": 0.7341, "step": 8207 }, { "epoch": 0.2972835929011228, "grad_norm": 1.4266818338519336, "learning_rate": 1.648178263778555e-05, "loss": 0.8172, "step": 8208 }, { "epoch": 0.29731981166244115, "grad_norm": 1.3186454836899955, "learning_rate": 1.648088931545752e-05, "loss": 0.7802, "step": 8209 }, { "epoch": 0.2973560304237595, "grad_norm": 1.1198316875826455, "learning_rate": 1.6479995903946756e-05, "loss": 0.7112, "step": 8210 }, { "epoch": 0.2973922491850779, "grad_norm": 1.620901845580693, "learning_rate": 1.6479102403265557e-05, "loss": 0.8222, "step": 8211 }, { "epoch": 0.2974284679463962, "grad_norm": 1.44885248635289, "learning_rate": 1.6478208813426217e-05, "loss": 0.8112, "step": 8212 }, { "epoch": 0.2974646867077146, "grad_norm": 1.403256395619156, "learning_rate": 1.6477315134441025e-05, "loss": 0.8071, "step": 8213 }, { "epoch": 0.29750090546903296, "grad_norm": 1.50067062337571, "learning_rate": 1.647642136632229e-05, "loss": 0.8429, "step": 8214 }, { "epoch": 0.29753712423035134, "grad_norm": 1.2590718017477087, "learning_rate": 1.6475527509082304e-05, "loss": 0.7871, "step": 8215 }, { "epoch": 0.2975733429916697, "grad_norm": 1.5327409399945757, "learning_rate": 1.6474633562733372e-05, "loss": 0.7438, "step": 8216 }, { "epoch": 0.297609561752988, "grad_norm": 1.4599216417618743, "learning_rate": 1.6473739527287792e-05, "loss": 0.7863, "step": 8217 }, { "epoch": 0.2976457805143064, "grad_norm": 1.6573671776949963, "learning_rate": 1.647284540275787e-05, "loss": 0.9161, "step": 8218 }, { "epoch": 0.2976819992756248, "grad_norm": 1.7147709563843856, "learning_rate": 1.6471951189155905e-05, "loss": 0.7605, "step": 8219 }, { "epoch": 0.29771821803694315, "grad_norm": 1.4810911458078366, "learning_rate": 1.6471056886494205e-05, "loss": 0.8486, "step": 8220 }, { "epoch": 0.2977544367982615, "grad_norm": 1.352542699540409, "learning_rate": 1.6470162494785077e-05, "loss": 0.7594, "step": 8221 }, { "epoch": 0.29779065555957984, "grad_norm": 1.5117145926088145, "learning_rate": 1.646926801404083e-05, "loss": 0.7686, "step": 8222 }, { "epoch": 0.2978268743208982, "grad_norm": 1.4825626703479895, "learning_rate": 1.6468373444273767e-05, "loss": 0.769, "step": 8223 }, { "epoch": 0.2978630930822166, "grad_norm": 1.6482296943756123, "learning_rate": 1.64674787854962e-05, "loss": 0.8162, "step": 8224 }, { "epoch": 0.29789931184353496, "grad_norm": 1.2998958279120603, "learning_rate": 1.646658403772045e-05, "loss": 0.7658, "step": 8225 }, { "epoch": 0.29793553060485334, "grad_norm": 1.224682284615329, "learning_rate": 1.6465689200958818e-05, "loss": 0.7549, "step": 8226 }, { "epoch": 0.29797174936617166, "grad_norm": 1.0781112214938684, "learning_rate": 1.646479427522362e-05, "loss": 0.7841, "step": 8227 }, { "epoch": 0.29800796812749003, "grad_norm": 1.2058836836955105, "learning_rate": 1.646389926052717e-05, "loss": 0.7794, "step": 8228 }, { "epoch": 0.2980441868888084, "grad_norm": 1.094939509556666, "learning_rate": 1.646300415688179e-05, "loss": 0.7605, "step": 8229 }, { "epoch": 0.2980804056501268, "grad_norm": 1.5805789922260851, "learning_rate": 1.646210896429979e-05, "loss": 0.7898, "step": 8230 }, { "epoch": 0.29811662441144515, "grad_norm": 1.0363778027821935, "learning_rate": 1.64612136827935e-05, "loss": 0.7693, "step": 8231 }, { "epoch": 0.29815284317276347, "grad_norm": 1.376242717311724, "learning_rate": 1.646031831237523e-05, "loss": 0.8073, "step": 8232 }, { "epoch": 0.29818906193408184, "grad_norm": 1.3545610417919127, "learning_rate": 1.64594228530573e-05, "loss": 0.8235, "step": 8233 }, { "epoch": 0.2982252806954002, "grad_norm": 1.4335220797135282, "learning_rate": 1.6458527304852036e-05, "loss": 0.758, "step": 8234 }, { "epoch": 0.2982614994567186, "grad_norm": 1.5660863725720058, "learning_rate": 1.6457631667771764e-05, "loss": 0.8874, "step": 8235 }, { "epoch": 0.29829771821803697, "grad_norm": 1.4869475101741134, "learning_rate": 1.6456735941828802e-05, "loss": 0.8287, "step": 8236 }, { "epoch": 0.2983339369793553, "grad_norm": 1.374409792764571, "learning_rate": 1.6455840127035483e-05, "loss": 0.7422, "step": 8237 }, { "epoch": 0.29837015574067366, "grad_norm": 1.4280650796934573, "learning_rate": 1.645494422340413e-05, "loss": 0.7955, "step": 8238 }, { "epoch": 0.29840637450199203, "grad_norm": 1.4332072818079273, "learning_rate": 1.6454048230947074e-05, "loss": 0.7939, "step": 8239 }, { "epoch": 0.2984425932633104, "grad_norm": 1.3304961936694017, "learning_rate": 1.6453152149676643e-05, "loss": 0.7892, "step": 8240 }, { "epoch": 0.2984788120246288, "grad_norm": 1.2117193814555158, "learning_rate": 1.645225597960517e-05, "loss": 0.7858, "step": 8241 }, { "epoch": 0.2985150307859471, "grad_norm": 1.0931157033354157, "learning_rate": 1.6451359720744984e-05, "loss": 0.7595, "step": 8242 }, { "epoch": 0.2985512495472655, "grad_norm": 1.5267175551766943, "learning_rate": 1.6450463373108418e-05, "loss": 0.8823, "step": 8243 }, { "epoch": 0.29858746830858385, "grad_norm": 1.6537120848372726, "learning_rate": 1.644956693670781e-05, "loss": 0.8227, "step": 8244 }, { "epoch": 0.2986236870699022, "grad_norm": 1.5339900408883753, "learning_rate": 1.6448670411555493e-05, "loss": 0.8224, "step": 8245 }, { "epoch": 0.2986599058312206, "grad_norm": 1.4428933840609655, "learning_rate": 1.6447773797663806e-05, "loss": 0.7515, "step": 8246 }, { "epoch": 0.2986961245925389, "grad_norm": 1.4933563124874483, "learning_rate": 1.6446877095045083e-05, "loss": 0.7835, "step": 8247 }, { "epoch": 0.2987323433538573, "grad_norm": 1.3192451334269963, "learning_rate": 1.644598030371167e-05, "loss": 0.7443, "step": 8248 }, { "epoch": 0.29876856211517566, "grad_norm": 3.719777429122227, "learning_rate": 1.6445083423675902e-05, "loss": 0.8823, "step": 8249 }, { "epoch": 0.29880478087649404, "grad_norm": 1.409523405848874, "learning_rate": 1.6444186454950123e-05, "loss": 0.7972, "step": 8250 }, { "epoch": 0.2988409996378124, "grad_norm": 1.470984159141476, "learning_rate": 1.6443289397546676e-05, "loss": 0.7986, "step": 8251 }, { "epoch": 0.29887721839913073, "grad_norm": 1.39988654782711, "learning_rate": 1.6442392251477904e-05, "loss": 0.7475, "step": 8252 }, { "epoch": 0.2989134371604491, "grad_norm": 1.413826143549321, "learning_rate": 1.6441495016756156e-05, "loss": 0.8508, "step": 8253 }, { "epoch": 0.2989496559217675, "grad_norm": 1.4889275294034388, "learning_rate": 1.6440597693393776e-05, "loss": 0.7725, "step": 8254 }, { "epoch": 0.29898587468308585, "grad_norm": 1.633509653645998, "learning_rate": 1.6439700281403113e-05, "loss": 0.7744, "step": 8255 }, { "epoch": 0.2990220934444042, "grad_norm": 1.304341925750976, "learning_rate": 1.6438802780796517e-05, "loss": 0.7546, "step": 8256 }, { "epoch": 0.29905831220572254, "grad_norm": 1.1496259524714174, "learning_rate": 1.6437905191586338e-05, "loss": 0.7823, "step": 8257 }, { "epoch": 0.2990945309670409, "grad_norm": 1.301372058114446, "learning_rate": 1.6437007513784918e-05, "loss": 0.819, "step": 8258 }, { "epoch": 0.2991307497283593, "grad_norm": 1.3155921476049874, "learning_rate": 1.643610974740463e-05, "loss": 0.7821, "step": 8259 }, { "epoch": 0.29916696848967766, "grad_norm": 1.3273101459704046, "learning_rate": 1.643521189245781e-05, "loss": 0.7478, "step": 8260 }, { "epoch": 0.29920318725099604, "grad_norm": 1.584465134989281, "learning_rate": 1.643431394895682e-05, "loss": 0.754, "step": 8261 }, { "epoch": 0.29923940601231436, "grad_norm": 1.3978449735636764, "learning_rate": 1.6433415916914018e-05, "loss": 0.8426, "step": 8262 }, { "epoch": 0.29927562477363273, "grad_norm": 1.1185740277380318, "learning_rate": 1.643251779634176e-05, "loss": 0.7568, "step": 8263 }, { "epoch": 0.2993118435349511, "grad_norm": 1.4916752377844107, "learning_rate": 1.6431619587252407e-05, "loss": 0.7411, "step": 8264 }, { "epoch": 0.2993480622962695, "grad_norm": 1.8841553683415306, "learning_rate": 1.6430721289658315e-05, "loss": 0.8463, "step": 8265 }, { "epoch": 0.29938428105758785, "grad_norm": 1.3995807479221922, "learning_rate": 1.6429822903571848e-05, "loss": 0.7326, "step": 8266 }, { "epoch": 0.29942049981890617, "grad_norm": 1.639598535547422, "learning_rate": 1.6428924429005366e-05, "loss": 0.7947, "step": 8267 }, { "epoch": 0.29945671858022455, "grad_norm": 1.3766644357832727, "learning_rate": 1.6428025865971236e-05, "loss": 0.7393, "step": 8268 }, { "epoch": 0.2994929373415429, "grad_norm": 1.458684745595108, "learning_rate": 1.6427127214481825e-05, "loss": 0.7907, "step": 8269 }, { "epoch": 0.2995291561028613, "grad_norm": 1.3154784719572572, "learning_rate": 1.6426228474549492e-05, "loss": 0.7563, "step": 8270 }, { "epoch": 0.29956537486417967, "grad_norm": 1.3561428836963112, "learning_rate": 1.6425329646186612e-05, "loss": 0.7776, "step": 8271 }, { "epoch": 0.299601593625498, "grad_norm": 1.544013552089636, "learning_rate": 1.6424430729405546e-05, "loss": 0.7203, "step": 8272 }, { "epoch": 0.29963781238681636, "grad_norm": 1.310065090662386, "learning_rate": 1.642353172421867e-05, "loss": 0.7608, "step": 8273 }, { "epoch": 0.29967403114813473, "grad_norm": 1.5451498888162356, "learning_rate": 1.6422632630638355e-05, "loss": 0.7933, "step": 8274 }, { "epoch": 0.2997102499094531, "grad_norm": 1.3323592686825834, "learning_rate": 1.642173344867697e-05, "loss": 0.7819, "step": 8275 }, { "epoch": 0.2997464686707715, "grad_norm": 1.3780200152729103, "learning_rate": 1.642083417834689e-05, "loss": 0.8341, "step": 8276 }, { "epoch": 0.2997826874320898, "grad_norm": 1.3309250690302494, "learning_rate": 1.6419934819660493e-05, "loss": 0.8151, "step": 8277 }, { "epoch": 0.2998189061934082, "grad_norm": 1.3293552626028762, "learning_rate": 1.6419035372630148e-05, "loss": 0.8251, "step": 8278 }, { "epoch": 0.29985512495472655, "grad_norm": 1.4035838403323295, "learning_rate": 1.641813583726824e-05, "loss": 0.8466, "step": 8279 }, { "epoch": 0.2998913437160449, "grad_norm": 1.0686620709745989, "learning_rate": 1.6417236213587138e-05, "loss": 0.7625, "step": 8280 }, { "epoch": 0.2999275624773633, "grad_norm": 1.415696529740199, "learning_rate": 1.641633650159923e-05, "loss": 0.842, "step": 8281 }, { "epoch": 0.2999637812386816, "grad_norm": 1.3234405594106935, "learning_rate": 1.6415436701316894e-05, "loss": 0.789, "step": 8282 }, { "epoch": 0.3, "grad_norm": 1.1751277565875502, "learning_rate": 1.6414536812752513e-05, "loss": 0.7479, "step": 8283 }, { "epoch": 0.30003621876131836, "grad_norm": 1.5578506134877121, "learning_rate": 1.6413636835918466e-05, "loss": 0.8067, "step": 8284 }, { "epoch": 0.30007243752263674, "grad_norm": 1.1354448449653534, "learning_rate": 1.641273677082714e-05, "loss": 0.7943, "step": 8285 }, { "epoch": 0.3001086562839551, "grad_norm": 1.4249361518779617, "learning_rate": 1.6411836617490925e-05, "loss": 0.7518, "step": 8286 }, { "epoch": 0.30014487504527343, "grad_norm": 1.5257495941300332, "learning_rate": 1.64109363759222e-05, "loss": 0.7684, "step": 8287 }, { "epoch": 0.3001810938065918, "grad_norm": 1.0786393567661736, "learning_rate": 1.6410036046133364e-05, "loss": 0.784, "step": 8288 }, { "epoch": 0.3002173125679102, "grad_norm": 1.4369754611699537, "learning_rate": 1.6409135628136795e-05, "loss": 0.8173, "step": 8289 }, { "epoch": 0.30025353132922855, "grad_norm": 1.556131049765478, "learning_rate": 1.640823512194489e-05, "loss": 0.8081, "step": 8290 }, { "epoch": 0.3002897500905469, "grad_norm": 1.4305617819930139, "learning_rate": 1.6407334527570037e-05, "loss": 0.7797, "step": 8291 }, { "epoch": 0.30032596885186524, "grad_norm": 1.504964730956046, "learning_rate": 1.6406433845024632e-05, "loss": 0.7537, "step": 8292 }, { "epoch": 0.3003621876131836, "grad_norm": 1.4890759554031894, "learning_rate": 1.6405533074321067e-05, "loss": 0.7836, "step": 8293 }, { "epoch": 0.300398406374502, "grad_norm": 1.2095574332699888, "learning_rate": 1.640463221547174e-05, "loss": 0.7759, "step": 8294 }, { "epoch": 0.30043462513582037, "grad_norm": 1.3916922014399622, "learning_rate": 1.6403731268489046e-05, "loss": 0.7669, "step": 8295 }, { "epoch": 0.30047084389713874, "grad_norm": 1.3859317305809278, "learning_rate": 1.640283023338538e-05, "loss": 0.735, "step": 8296 }, { "epoch": 0.30050706265845706, "grad_norm": 1.491332731163397, "learning_rate": 1.640192911017315e-05, "loss": 0.8678, "step": 8297 }, { "epoch": 0.30054328141977543, "grad_norm": 1.1048440685213257, "learning_rate": 1.6401027898864745e-05, "loss": 0.7587, "step": 8298 }, { "epoch": 0.3005795001810938, "grad_norm": 1.402266534124589, "learning_rate": 1.6400126599472576e-05, "loss": 0.8061, "step": 8299 }, { "epoch": 0.3006157189424122, "grad_norm": 1.553454582361777, "learning_rate": 1.639922521200904e-05, "loss": 0.8098, "step": 8300 }, { "epoch": 0.30065193770373055, "grad_norm": 1.1087961306875, "learning_rate": 1.6398323736486536e-05, "loss": 0.7713, "step": 8301 }, { "epoch": 0.3006881564650489, "grad_norm": 1.2984468689269502, "learning_rate": 1.6397422172917484e-05, "loss": 0.8317, "step": 8302 }, { "epoch": 0.30072437522636725, "grad_norm": 1.2389674624390348, "learning_rate": 1.6396520521314277e-05, "loss": 0.8157, "step": 8303 }, { "epoch": 0.3007605939876856, "grad_norm": 1.562381924697796, "learning_rate": 1.6395618781689328e-05, "loss": 0.7613, "step": 8304 }, { "epoch": 0.300796812749004, "grad_norm": 1.3682344338224066, "learning_rate": 1.6394716954055043e-05, "loss": 0.837, "step": 8305 }, { "epoch": 0.30083303151032237, "grad_norm": 1.3009375436781911, "learning_rate": 1.6393815038423837e-05, "loss": 0.8335, "step": 8306 }, { "epoch": 0.3008692502716407, "grad_norm": 1.330977906806452, "learning_rate": 1.6392913034808117e-05, "loss": 0.761, "step": 8307 }, { "epoch": 0.30090546903295906, "grad_norm": 1.4116393605689506, "learning_rate": 1.6392010943220293e-05, "loss": 0.8616, "step": 8308 }, { "epoch": 0.30094168779427743, "grad_norm": 1.495531300129507, "learning_rate": 1.6391108763672787e-05, "loss": 0.7484, "step": 8309 }, { "epoch": 0.3009779065555958, "grad_norm": 1.4049470653714018, "learning_rate": 1.6390206496178007e-05, "loss": 0.7438, "step": 8310 }, { "epoch": 0.3010141253169142, "grad_norm": 1.0703456069586066, "learning_rate": 1.638930414074837e-05, "loss": 0.7257, "step": 8311 }, { "epoch": 0.3010503440782325, "grad_norm": 1.2484561999071402, "learning_rate": 1.6388401697396293e-05, "loss": 0.8159, "step": 8312 }, { "epoch": 0.3010865628395509, "grad_norm": 1.5507266577905363, "learning_rate": 1.63874991661342e-05, "loss": 0.7954, "step": 8313 }, { "epoch": 0.30112278160086925, "grad_norm": 1.2037608641352249, "learning_rate": 1.63865965469745e-05, "loss": 0.7419, "step": 8314 }, { "epoch": 0.3011590003621876, "grad_norm": 1.5213582858283767, "learning_rate": 1.6385693839929623e-05, "loss": 0.8321, "step": 8315 }, { "epoch": 0.301195219123506, "grad_norm": 1.4960860502428366, "learning_rate": 1.6384791045011986e-05, "loss": 0.7377, "step": 8316 }, { "epoch": 0.3012314378848243, "grad_norm": 1.4712661321573126, "learning_rate": 1.6383888162234014e-05, "loss": 0.7983, "step": 8317 }, { "epoch": 0.3012676566461427, "grad_norm": 1.4286090373832099, "learning_rate": 1.6382985191608134e-05, "loss": 0.9, "step": 8318 }, { "epoch": 0.30130387540746106, "grad_norm": 1.4276012209836717, "learning_rate": 1.6382082133146767e-05, "loss": 0.7797, "step": 8319 }, { "epoch": 0.30134009416877944, "grad_norm": 1.3774023878757091, "learning_rate": 1.6381178986862344e-05, "loss": 0.8682, "step": 8320 }, { "epoch": 0.3013763129300978, "grad_norm": 1.085043221193908, "learning_rate": 1.638027575276729e-05, "loss": 0.7739, "step": 8321 }, { "epoch": 0.30141253169141613, "grad_norm": 1.7703266377233833, "learning_rate": 1.6379372430874037e-05, "loss": 0.7726, "step": 8322 }, { "epoch": 0.3014487504527345, "grad_norm": 1.5906863744563942, "learning_rate": 1.637846902119501e-05, "loss": 0.8543, "step": 8323 }, { "epoch": 0.3014849692140529, "grad_norm": 1.470073715348702, "learning_rate": 1.6377565523742648e-05, "loss": 0.7662, "step": 8324 }, { "epoch": 0.30152118797537125, "grad_norm": 1.3884302316679749, "learning_rate": 1.637666193852938e-05, "loss": 0.8714, "step": 8325 }, { "epoch": 0.3015574067366896, "grad_norm": 1.607237220792097, "learning_rate": 1.637575826556764e-05, "loss": 0.7953, "step": 8326 }, { "epoch": 0.30159362549800794, "grad_norm": 1.4595703278649654, "learning_rate": 1.6374854504869864e-05, "loss": 0.8394, "step": 8327 }, { "epoch": 0.3016298442593263, "grad_norm": 1.447350985182226, "learning_rate": 1.637395065644849e-05, "loss": 0.7777, "step": 8328 }, { "epoch": 0.3016660630206447, "grad_norm": 1.4177358412392127, "learning_rate": 1.6373046720315954e-05, "loss": 0.8257, "step": 8329 }, { "epoch": 0.30170228178196307, "grad_norm": 1.3125842555568006, "learning_rate": 1.6372142696484693e-05, "loss": 0.8685, "step": 8330 }, { "epoch": 0.30173850054328144, "grad_norm": 1.3623732060801603, "learning_rate": 1.637123858496715e-05, "loss": 0.7818, "step": 8331 }, { "epoch": 0.30177471930459976, "grad_norm": 1.5242006640043138, "learning_rate": 1.6370334385775766e-05, "loss": 0.8534, "step": 8332 }, { "epoch": 0.30181093806591813, "grad_norm": 1.279366408343927, "learning_rate": 1.6369430098922987e-05, "loss": 0.7995, "step": 8333 }, { "epoch": 0.3018471568272365, "grad_norm": 1.2521645026068184, "learning_rate": 1.6368525724421248e-05, "loss": 0.7228, "step": 8334 }, { "epoch": 0.3018833755885549, "grad_norm": 1.2641917969187029, "learning_rate": 1.6367621262283003e-05, "loss": 0.7151, "step": 8335 }, { "epoch": 0.30191959434987325, "grad_norm": 1.3648965698021147, "learning_rate": 1.636671671252069e-05, "loss": 0.8586, "step": 8336 }, { "epoch": 0.3019558131111916, "grad_norm": 1.1907087426134124, "learning_rate": 1.6365812075146765e-05, "loss": 0.7288, "step": 8337 }, { "epoch": 0.30199203187250995, "grad_norm": 1.5219201258484263, "learning_rate": 1.6364907350173672e-05, "loss": 0.8535, "step": 8338 }, { "epoch": 0.3020282506338283, "grad_norm": 1.3257055965726285, "learning_rate": 1.6364002537613858e-05, "loss": 0.6597, "step": 8339 }, { "epoch": 0.3020644693951467, "grad_norm": 1.3506659982580633, "learning_rate": 1.6363097637479782e-05, "loss": 0.7608, "step": 8340 }, { "epoch": 0.30210068815646507, "grad_norm": 1.5091496388571546, "learning_rate": 1.636219264978389e-05, "loss": 0.791, "step": 8341 }, { "epoch": 0.3021369069177834, "grad_norm": 1.395332286012014, "learning_rate": 1.6361287574538633e-05, "loss": 0.8099, "step": 8342 }, { "epoch": 0.30217312567910176, "grad_norm": 1.3314172071960266, "learning_rate": 1.6360382411756472e-05, "loss": 0.7602, "step": 8343 }, { "epoch": 0.30220934444042014, "grad_norm": 1.400731163703847, "learning_rate": 1.635947716144986e-05, "loss": 0.747, "step": 8344 }, { "epoch": 0.3022455632017385, "grad_norm": 1.3565036645783364, "learning_rate": 1.6358571823631257e-05, "loss": 0.786, "step": 8345 }, { "epoch": 0.3022817819630569, "grad_norm": 1.357730940069183, "learning_rate": 1.6357666398313115e-05, "loss": 0.8113, "step": 8346 }, { "epoch": 0.3023180007243752, "grad_norm": 1.2114110314491053, "learning_rate": 1.6356760885507897e-05, "loss": 0.7972, "step": 8347 }, { "epoch": 0.3023542194856936, "grad_norm": 1.2785733470130303, "learning_rate": 1.6355855285228063e-05, "loss": 0.6725, "step": 8348 }, { "epoch": 0.30239043824701195, "grad_norm": 1.5721478755763298, "learning_rate": 1.635494959748608e-05, "loss": 0.7922, "step": 8349 }, { "epoch": 0.3024266570083303, "grad_norm": 1.4979702023476942, "learning_rate": 1.6354043822294402e-05, "loss": 0.8337, "step": 8350 }, { "epoch": 0.3024628757696487, "grad_norm": 1.3888684199401082, "learning_rate": 1.6353137959665502e-05, "loss": 0.8403, "step": 8351 }, { "epoch": 0.302499094530967, "grad_norm": 1.3070509897692102, "learning_rate": 1.635223200961184e-05, "loss": 0.7237, "step": 8352 }, { "epoch": 0.3025353132922854, "grad_norm": 1.2185591866784207, "learning_rate": 1.6351325972145882e-05, "loss": 0.7911, "step": 8353 }, { "epoch": 0.30257153205360376, "grad_norm": 1.4735225145787583, "learning_rate": 1.63504198472801e-05, "loss": 0.7947, "step": 8354 }, { "epoch": 0.30260775081492214, "grad_norm": 1.4525450656731425, "learning_rate": 1.6349513635026955e-05, "loss": 0.7998, "step": 8355 }, { "epoch": 0.3026439695762405, "grad_norm": 1.405950271930412, "learning_rate": 1.6348607335398932e-05, "loss": 0.8279, "step": 8356 }, { "epoch": 0.30268018833755883, "grad_norm": 1.455260215477017, "learning_rate": 1.6347700948408486e-05, "loss": 0.8536, "step": 8357 }, { "epoch": 0.3027164070988772, "grad_norm": 1.2534691941686877, "learning_rate": 1.6346794474068102e-05, "loss": 0.8094, "step": 8358 }, { "epoch": 0.3027526258601956, "grad_norm": 1.1986430531538739, "learning_rate": 1.634588791239025e-05, "loss": 0.8427, "step": 8359 }, { "epoch": 0.30278884462151395, "grad_norm": 1.3206005360661353, "learning_rate": 1.6344981263387404e-05, "loss": 0.8031, "step": 8360 }, { "epoch": 0.3028250633828323, "grad_norm": 1.5125479949674014, "learning_rate": 1.634407452707204e-05, "loss": 0.7729, "step": 8361 }, { "epoch": 0.30286128214415065, "grad_norm": 1.5395655620723208, "learning_rate": 1.6343167703456634e-05, "loss": 0.7915, "step": 8362 }, { "epoch": 0.302897500905469, "grad_norm": 1.379865336459345, "learning_rate": 1.6342260792553667e-05, "loss": 0.7581, "step": 8363 }, { "epoch": 0.3029337196667874, "grad_norm": 1.460356128752323, "learning_rate": 1.634135379437562e-05, "loss": 0.797, "step": 8364 }, { "epoch": 0.30296993842810577, "grad_norm": 1.5925121864944725, "learning_rate": 1.6340446708934975e-05, "loss": 0.9042, "step": 8365 }, { "epoch": 0.30300615718942414, "grad_norm": 1.3894140028358983, "learning_rate": 1.6339539536244207e-05, "loss": 0.7629, "step": 8366 }, { "epoch": 0.30304237595074246, "grad_norm": 1.4919491166790413, "learning_rate": 1.633863227631581e-05, "loss": 0.7918, "step": 8367 }, { "epoch": 0.30307859471206083, "grad_norm": 1.1660510441977254, "learning_rate": 1.633772492916226e-05, "loss": 0.8101, "step": 8368 }, { "epoch": 0.3031148134733792, "grad_norm": 1.4025191732626434, "learning_rate": 1.633681749479605e-05, "loss": 0.8624, "step": 8369 }, { "epoch": 0.3031510322346976, "grad_norm": 1.0582691571787386, "learning_rate": 1.633590997322966e-05, "loss": 0.7817, "step": 8370 }, { "epoch": 0.30318725099601596, "grad_norm": 1.38840937782577, "learning_rate": 1.633500236447558e-05, "loss": 0.6893, "step": 8371 }, { "epoch": 0.3032234697573343, "grad_norm": 1.3819751399797602, "learning_rate": 1.6334094668546304e-05, "loss": 0.82, "step": 8372 }, { "epoch": 0.30325968851865265, "grad_norm": 1.4313737007629834, "learning_rate": 1.6333186885454317e-05, "loss": 0.7548, "step": 8373 }, { "epoch": 0.303295907279971, "grad_norm": 1.2608539404711483, "learning_rate": 1.6332279015212114e-05, "loss": 0.7009, "step": 8374 }, { "epoch": 0.3033321260412894, "grad_norm": 1.393861941259908, "learning_rate": 1.633137105783219e-05, "loss": 0.7584, "step": 8375 }, { "epoch": 0.30336834480260777, "grad_norm": 1.4373731421129712, "learning_rate": 1.6330463013327034e-05, "loss": 0.754, "step": 8376 }, { "epoch": 0.3034045635639261, "grad_norm": 1.325869821496406, "learning_rate": 1.632955488170915e-05, "loss": 0.7454, "step": 8377 }, { "epoch": 0.30344078232524446, "grad_norm": 1.3977180207165503, "learning_rate": 1.6328646662991025e-05, "loss": 0.7981, "step": 8378 }, { "epoch": 0.30347700108656284, "grad_norm": 1.4780387453440333, "learning_rate": 1.632773835718516e-05, "loss": 0.8068, "step": 8379 }, { "epoch": 0.3035132198478812, "grad_norm": 1.1368203554530512, "learning_rate": 1.632682996430406e-05, "loss": 0.7874, "step": 8380 }, { "epoch": 0.3035494386091996, "grad_norm": 1.1309530781123525, "learning_rate": 1.6325921484360215e-05, "loss": 0.7106, "step": 8381 }, { "epoch": 0.3035856573705179, "grad_norm": 1.5030675023020383, "learning_rate": 1.6325012917366135e-05, "loss": 0.7599, "step": 8382 }, { "epoch": 0.3036218761318363, "grad_norm": 1.5233490942710768, "learning_rate": 1.632410426333432e-05, "loss": 0.7837, "step": 8383 }, { "epoch": 0.30365809489315465, "grad_norm": 1.536080529886816, "learning_rate": 1.6323195522277272e-05, "loss": 0.8427, "step": 8384 }, { "epoch": 0.303694313654473, "grad_norm": 1.2388308986910144, "learning_rate": 1.6322286694207495e-05, "loss": 0.7861, "step": 8385 }, { "epoch": 0.3037305324157914, "grad_norm": 1.3338832635385822, "learning_rate": 1.6321377779137503e-05, "loss": 0.7678, "step": 8386 }, { "epoch": 0.3037667511771097, "grad_norm": 1.3721327276644162, "learning_rate": 1.6320468777079796e-05, "loss": 0.8036, "step": 8387 }, { "epoch": 0.3038029699384281, "grad_norm": 1.6368181476136532, "learning_rate": 1.6319559688046886e-05, "loss": 0.7765, "step": 8388 }, { "epoch": 0.30383918869974647, "grad_norm": 1.3988842703367488, "learning_rate": 1.6318650512051278e-05, "loss": 0.7754, "step": 8389 }, { "epoch": 0.30387540746106484, "grad_norm": 1.3250748320244485, "learning_rate": 1.631774124910549e-05, "loss": 0.7772, "step": 8390 }, { "epoch": 0.3039116262223832, "grad_norm": 1.2368106917208985, "learning_rate": 1.6316831899222033e-05, "loss": 0.7273, "step": 8391 }, { "epoch": 0.30394784498370153, "grad_norm": 1.4451286082944919, "learning_rate": 1.6315922462413417e-05, "loss": 0.8084, "step": 8392 }, { "epoch": 0.3039840637450199, "grad_norm": 1.408520066007633, "learning_rate": 1.6315012938692153e-05, "loss": 0.7879, "step": 8393 }, { "epoch": 0.3040202825063383, "grad_norm": 1.5397458520016771, "learning_rate": 1.631410332807077e-05, "loss": 0.7959, "step": 8394 }, { "epoch": 0.30405650126765665, "grad_norm": 1.404449250939958, "learning_rate": 1.631319363056177e-05, "loss": 0.8431, "step": 8395 }, { "epoch": 0.304092720028975, "grad_norm": 1.417873855573417, "learning_rate": 1.6312283846177684e-05, "loss": 0.8458, "step": 8396 }, { "epoch": 0.30412893879029335, "grad_norm": 1.815531989914413, "learning_rate": 1.6311373974931026e-05, "loss": 0.7438, "step": 8397 }, { "epoch": 0.3041651575516117, "grad_norm": 1.266117499035902, "learning_rate": 1.6310464016834316e-05, "loss": 0.7852, "step": 8398 }, { "epoch": 0.3042013763129301, "grad_norm": 1.0666325659996823, "learning_rate": 1.6309553971900073e-05, "loss": 0.7789, "step": 8399 }, { "epoch": 0.30423759507424847, "grad_norm": 1.4021977711250995, "learning_rate": 1.630864384014083e-05, "loss": 0.8454, "step": 8400 }, { "epoch": 0.30427381383556684, "grad_norm": 1.7787632733544774, "learning_rate": 1.6307733621569094e-05, "loss": 0.8379, "step": 8401 }, { "epoch": 0.30431003259688516, "grad_norm": 1.4383381759519718, "learning_rate": 1.630682331619741e-05, "loss": 0.8479, "step": 8402 }, { "epoch": 0.30434625135820353, "grad_norm": 1.2942193594015443, "learning_rate": 1.630591292403829e-05, "loss": 0.7586, "step": 8403 }, { "epoch": 0.3043824701195219, "grad_norm": 1.4114173212961159, "learning_rate": 1.630500244510427e-05, "loss": 0.8044, "step": 8404 }, { "epoch": 0.3044186888808403, "grad_norm": 1.485170020973217, "learning_rate": 1.6304091879407875e-05, "loss": 0.862, "step": 8405 }, { "epoch": 0.30445490764215866, "grad_norm": 1.445979115400702, "learning_rate": 1.6303181226961636e-05, "loss": 0.8438, "step": 8406 }, { "epoch": 0.304491126403477, "grad_norm": 1.3145493136350213, "learning_rate": 1.6302270487778088e-05, "loss": 0.7597, "step": 8407 }, { "epoch": 0.30452734516479535, "grad_norm": 1.2628885908225491, "learning_rate": 1.630135966186976e-05, "loss": 0.8246, "step": 8408 }, { "epoch": 0.3045635639261137, "grad_norm": 1.5390165051011464, "learning_rate": 1.6300448749249182e-05, "loss": 0.816, "step": 8409 }, { "epoch": 0.3045997826874321, "grad_norm": 1.3310174982301017, "learning_rate": 1.6299537749928897e-05, "loss": 0.7768, "step": 8410 }, { "epoch": 0.30463600144875047, "grad_norm": 1.5313863644896344, "learning_rate": 1.6298626663921434e-05, "loss": 0.7697, "step": 8411 }, { "epoch": 0.3046722202100688, "grad_norm": 1.5261189459631352, "learning_rate": 1.629771549123934e-05, "loss": 0.8309, "step": 8412 }, { "epoch": 0.30470843897138716, "grad_norm": 1.4319013192991528, "learning_rate": 1.629680423189514e-05, "loss": 0.7612, "step": 8413 }, { "epoch": 0.30474465773270554, "grad_norm": 1.4385942469559543, "learning_rate": 1.6295892885901384e-05, "loss": 0.8175, "step": 8414 }, { "epoch": 0.3047808764940239, "grad_norm": 1.577287611120589, "learning_rate": 1.629498145327061e-05, "loss": 0.8013, "step": 8415 }, { "epoch": 0.3048170952553423, "grad_norm": 1.8735701721991398, "learning_rate": 1.629406993401536e-05, "loss": 0.8706, "step": 8416 }, { "epoch": 0.3048533140166606, "grad_norm": 1.3881341641616096, "learning_rate": 1.6293158328148177e-05, "loss": 0.7846, "step": 8417 }, { "epoch": 0.304889532777979, "grad_norm": 1.0625437837495606, "learning_rate": 1.6292246635681607e-05, "loss": 0.7535, "step": 8418 }, { "epoch": 0.30492575153929735, "grad_norm": 1.552256972618896, "learning_rate": 1.6291334856628195e-05, "loss": 0.8606, "step": 8419 }, { "epoch": 0.3049619703006157, "grad_norm": 1.22788914638509, "learning_rate": 1.6290422991000484e-05, "loss": 0.7756, "step": 8420 }, { "epoch": 0.3049981890619341, "grad_norm": 1.1052250015253038, "learning_rate": 1.6289511038811027e-05, "loss": 0.7562, "step": 8421 }, { "epoch": 0.3050344078232524, "grad_norm": 1.362920934149314, "learning_rate": 1.6288599000072373e-05, "loss": 0.7494, "step": 8422 }, { "epoch": 0.3050706265845708, "grad_norm": 1.3403971060431483, "learning_rate": 1.6287686874797066e-05, "loss": 0.8226, "step": 8423 }, { "epoch": 0.30510684534588917, "grad_norm": 1.3723106642112417, "learning_rate": 1.6286774662997665e-05, "loss": 0.8115, "step": 8424 }, { "epoch": 0.30514306410720754, "grad_norm": 1.6517038656335803, "learning_rate": 1.628586236468672e-05, "loss": 0.8163, "step": 8425 }, { "epoch": 0.3051792828685259, "grad_norm": 1.4089603431805446, "learning_rate": 1.628494997987679e-05, "loss": 0.8575, "step": 8426 }, { "epoch": 0.30521550162984423, "grad_norm": 1.2894505552481788, "learning_rate": 1.6284037508580425e-05, "loss": 0.8153, "step": 8427 }, { "epoch": 0.3052517203911626, "grad_norm": 1.4439089893966892, "learning_rate": 1.628312495081018e-05, "loss": 0.7523, "step": 8428 }, { "epoch": 0.305287939152481, "grad_norm": 1.4814146896194726, "learning_rate": 1.6282212306578616e-05, "loss": 0.7742, "step": 8429 }, { "epoch": 0.30532415791379935, "grad_norm": 1.381324390311182, "learning_rate": 1.628129957589829e-05, "loss": 0.7909, "step": 8430 }, { "epoch": 0.30536037667511773, "grad_norm": 1.3910698400203865, "learning_rate": 1.6280386758781763e-05, "loss": 0.8633, "step": 8431 }, { "epoch": 0.30539659543643605, "grad_norm": 1.5004203761159614, "learning_rate": 1.6279473855241596e-05, "loss": 0.7729, "step": 8432 }, { "epoch": 0.3054328141977544, "grad_norm": 1.1255706207748082, "learning_rate": 1.627856086529035e-05, "loss": 0.6792, "step": 8433 }, { "epoch": 0.3054690329590728, "grad_norm": 1.1191732987646865, "learning_rate": 1.627764778894059e-05, "loss": 0.7935, "step": 8434 }, { "epoch": 0.30550525172039117, "grad_norm": 1.4185539724607372, "learning_rate": 1.627673462620488e-05, "loss": 0.7685, "step": 8435 }, { "epoch": 0.30554147048170954, "grad_norm": 1.421884245894836, "learning_rate": 1.627582137709579e-05, "loss": 0.7454, "step": 8436 }, { "epoch": 0.30557768924302786, "grad_norm": 1.3867601208828662, "learning_rate": 1.6274908041625877e-05, "loss": 0.7843, "step": 8437 }, { "epoch": 0.30561390800434624, "grad_norm": 1.4041361020390546, "learning_rate": 1.627399461980772e-05, "loss": 0.7311, "step": 8438 }, { "epoch": 0.3056501267656646, "grad_norm": 1.4388970519925623, "learning_rate": 1.6273081111653884e-05, "loss": 0.7855, "step": 8439 }, { "epoch": 0.305686345526983, "grad_norm": 1.3632108607333118, "learning_rate": 1.6272167517176938e-05, "loss": 0.7693, "step": 8440 }, { "epoch": 0.30572256428830136, "grad_norm": 1.384194789843514, "learning_rate": 1.627125383638946e-05, "loss": 0.7475, "step": 8441 }, { "epoch": 0.3057587830496197, "grad_norm": 1.2895753172764999, "learning_rate": 1.6270340069304014e-05, "loss": 0.7336, "step": 8442 }, { "epoch": 0.30579500181093805, "grad_norm": 1.3884129162060197, "learning_rate": 1.626942621593318e-05, "loss": 0.7994, "step": 8443 }, { "epoch": 0.3058312205722564, "grad_norm": 1.355833147521183, "learning_rate": 1.6268512276289535e-05, "loss": 0.8173, "step": 8444 }, { "epoch": 0.3058674393335748, "grad_norm": 1.085398004201378, "learning_rate": 1.626759825038565e-05, "loss": 0.7705, "step": 8445 }, { "epoch": 0.30590365809489317, "grad_norm": 1.3707268956358085, "learning_rate": 1.6266684138234108e-05, "loss": 0.7622, "step": 8446 }, { "epoch": 0.3059398768562115, "grad_norm": 1.3002373847422766, "learning_rate": 1.6265769939847485e-05, "loss": 0.7822, "step": 8447 }, { "epoch": 0.30597609561752986, "grad_norm": 1.4411239308775887, "learning_rate": 1.626485565523836e-05, "loss": 0.7866, "step": 8448 }, { "epoch": 0.30601231437884824, "grad_norm": 1.1053868167463161, "learning_rate": 1.6263941284419322e-05, "loss": 0.7836, "step": 8449 }, { "epoch": 0.3060485331401666, "grad_norm": 1.5215969607183466, "learning_rate": 1.6263026827402943e-05, "loss": 0.7823, "step": 8450 }, { "epoch": 0.306084751901485, "grad_norm": 1.3369010523406162, "learning_rate": 1.6262112284201815e-05, "loss": 0.7304, "step": 8451 }, { "epoch": 0.3061209706628033, "grad_norm": 1.0358208790451293, "learning_rate": 1.6261197654828515e-05, "loss": 0.8403, "step": 8452 }, { "epoch": 0.3061571894241217, "grad_norm": 1.3193000383384894, "learning_rate": 1.6260282939295637e-05, "loss": 0.7809, "step": 8453 }, { "epoch": 0.30619340818544005, "grad_norm": 1.0899240361549534, "learning_rate": 1.6259368137615767e-05, "loss": 0.7433, "step": 8454 }, { "epoch": 0.3062296269467584, "grad_norm": 1.372068918302106, "learning_rate": 1.6258453249801487e-05, "loss": 0.8221, "step": 8455 }, { "epoch": 0.3062658457080768, "grad_norm": 1.407855465812307, "learning_rate": 1.6257538275865396e-05, "loss": 0.7348, "step": 8456 }, { "epoch": 0.3063020644693952, "grad_norm": 1.2123303186513572, "learning_rate": 1.6256623215820074e-05, "loss": 0.7399, "step": 8457 }, { "epoch": 0.3063382832307135, "grad_norm": 1.4003554104323799, "learning_rate": 1.6255708069678122e-05, "loss": 0.8529, "step": 8458 }, { "epoch": 0.30637450199203187, "grad_norm": 1.4453503436211803, "learning_rate": 1.625479283745213e-05, "loss": 0.7747, "step": 8459 }, { "epoch": 0.30641072075335024, "grad_norm": 1.327491745323713, "learning_rate": 1.6253877519154693e-05, "loss": 0.8143, "step": 8460 }, { "epoch": 0.3064469395146686, "grad_norm": 1.2433223338467856, "learning_rate": 1.6252962114798407e-05, "loss": 0.8048, "step": 8461 }, { "epoch": 0.306483158275987, "grad_norm": 1.5208529091145613, "learning_rate": 1.625204662439587e-05, "loss": 0.8256, "step": 8462 }, { "epoch": 0.3065193770373053, "grad_norm": 1.4620211937230219, "learning_rate": 1.6251131047959675e-05, "loss": 0.8434, "step": 8463 }, { "epoch": 0.3065555957986237, "grad_norm": 1.1739136400210564, "learning_rate": 1.625021538550242e-05, "loss": 0.7858, "step": 8464 }, { "epoch": 0.30659181455994206, "grad_norm": 1.3123998532802756, "learning_rate": 1.6249299637036716e-05, "loss": 0.7422, "step": 8465 }, { "epoch": 0.30662803332126043, "grad_norm": 1.3274637623823013, "learning_rate": 1.6248383802575154e-05, "loss": 0.7263, "step": 8466 }, { "epoch": 0.3066642520825788, "grad_norm": 1.4716904360727352, "learning_rate": 1.624746788213034e-05, "loss": 0.831, "step": 8467 }, { "epoch": 0.3067004708438971, "grad_norm": 1.1187686986615943, "learning_rate": 1.6246551875714883e-05, "loss": 0.7578, "step": 8468 }, { "epoch": 0.3067366896052155, "grad_norm": 1.413511214886402, "learning_rate": 1.624563578334138e-05, "loss": 0.8215, "step": 8469 }, { "epoch": 0.30677290836653387, "grad_norm": 0.9375610349699977, "learning_rate": 1.624471960502244e-05, "loss": 0.7464, "step": 8470 }, { "epoch": 0.30680912712785224, "grad_norm": 1.2954137451950378, "learning_rate": 1.6243803340770676e-05, "loss": 0.7157, "step": 8471 }, { "epoch": 0.3068453458891706, "grad_norm": 1.420545928199292, "learning_rate": 1.6242886990598688e-05, "loss": 0.8251, "step": 8472 }, { "epoch": 0.30688156465048894, "grad_norm": 1.367001450643577, "learning_rate": 1.624197055451909e-05, "loss": 0.7718, "step": 8473 }, { "epoch": 0.3069177834118073, "grad_norm": 1.3077301958721403, "learning_rate": 1.6241054032544493e-05, "loss": 0.7333, "step": 8474 }, { "epoch": 0.3069540021731257, "grad_norm": 1.4432609999482142, "learning_rate": 1.624013742468751e-05, "loss": 0.7844, "step": 8475 }, { "epoch": 0.30699022093444406, "grad_norm": 1.3545681087116137, "learning_rate": 1.623922073096075e-05, "loss": 0.8196, "step": 8476 }, { "epoch": 0.30702643969576243, "grad_norm": 1.1456895357980919, "learning_rate": 1.623830395137683e-05, "loss": 0.7608, "step": 8477 }, { "epoch": 0.30706265845708075, "grad_norm": 1.073188707757586, "learning_rate": 1.6237387085948365e-05, "loss": 0.7488, "step": 8478 }, { "epoch": 0.3070988772183991, "grad_norm": 1.4882337351095178, "learning_rate": 1.6236470134687977e-05, "loss": 0.7438, "step": 8479 }, { "epoch": 0.3071350959797175, "grad_norm": 1.4230243627676973, "learning_rate": 1.6235553097608278e-05, "loss": 0.8777, "step": 8480 }, { "epoch": 0.3071713147410359, "grad_norm": 1.349141118345922, "learning_rate": 1.623463597472189e-05, "loss": 0.7765, "step": 8481 }, { "epoch": 0.30720753350235425, "grad_norm": 1.4687441333694167, "learning_rate": 1.6233718766041435e-05, "loss": 0.8032, "step": 8482 }, { "epoch": 0.30724375226367256, "grad_norm": 1.1292814760154462, "learning_rate": 1.623280147157953e-05, "loss": 0.7137, "step": 8483 }, { "epoch": 0.30727997102499094, "grad_norm": 1.1899542418465825, "learning_rate": 1.6231884091348803e-05, "loss": 0.7515, "step": 8484 }, { "epoch": 0.3073161897863093, "grad_norm": 1.383368286817841, "learning_rate": 1.623096662536187e-05, "loss": 0.7966, "step": 8485 }, { "epoch": 0.3073524085476277, "grad_norm": 1.4659373308473196, "learning_rate": 1.6230049073631364e-05, "loss": 0.7465, "step": 8486 }, { "epoch": 0.30738862730894606, "grad_norm": 1.3319767034666523, "learning_rate": 1.622913143616991e-05, "loss": 0.7609, "step": 8487 }, { "epoch": 0.3074248460702644, "grad_norm": 1.1008751292727754, "learning_rate": 1.622821371299013e-05, "loss": 0.6919, "step": 8488 }, { "epoch": 0.30746106483158275, "grad_norm": 1.3554958002760789, "learning_rate": 1.622729590410466e-05, "loss": 0.7509, "step": 8489 }, { "epoch": 0.3074972835929011, "grad_norm": 1.3051994156044435, "learning_rate": 1.6226378009526126e-05, "loss": 0.829, "step": 8490 }, { "epoch": 0.3075335023542195, "grad_norm": 1.419533737502703, "learning_rate": 1.622546002926716e-05, "loss": 0.8121, "step": 8491 }, { "epoch": 0.3075697211155379, "grad_norm": 1.2937706129432065, "learning_rate": 1.6224541963340392e-05, "loss": 0.7577, "step": 8492 }, { "epoch": 0.3076059398768562, "grad_norm": 1.386771372683166, "learning_rate": 1.6223623811758458e-05, "loss": 0.783, "step": 8493 }, { "epoch": 0.30764215863817457, "grad_norm": 1.355915046951291, "learning_rate": 1.622270557453399e-05, "loss": 0.7584, "step": 8494 }, { "epoch": 0.30767837739949294, "grad_norm": 1.0115899625734481, "learning_rate": 1.622178725167963e-05, "loss": 0.7599, "step": 8495 }, { "epoch": 0.3077145961608113, "grad_norm": 1.3664662754458146, "learning_rate": 1.6220868843208006e-05, "loss": 0.7241, "step": 8496 }, { "epoch": 0.3077508149221297, "grad_norm": 1.4402489914138195, "learning_rate": 1.6219950349131765e-05, "loss": 0.8048, "step": 8497 }, { "epoch": 0.307787033683448, "grad_norm": 1.1097171983522183, "learning_rate": 1.6219031769463537e-05, "loss": 0.8066, "step": 8498 }, { "epoch": 0.3078232524447664, "grad_norm": 1.5403238007882987, "learning_rate": 1.621811310421597e-05, "loss": 0.7745, "step": 8499 }, { "epoch": 0.30785947120608476, "grad_norm": 1.56006526975442, "learning_rate": 1.6217194353401703e-05, "loss": 0.7823, "step": 8500 }, { "epoch": 0.30789568996740313, "grad_norm": 1.2918083165682646, "learning_rate": 1.621627551703338e-05, "loss": 0.8306, "step": 8501 }, { "epoch": 0.3079319087287215, "grad_norm": 1.166048937455273, "learning_rate": 1.621535659512364e-05, "loss": 0.7446, "step": 8502 }, { "epoch": 0.3079681274900398, "grad_norm": 1.3717717236115667, "learning_rate": 1.6214437587685136e-05, "loss": 0.7835, "step": 8503 }, { "epoch": 0.3080043462513582, "grad_norm": 1.44343239634372, "learning_rate": 1.621351849473051e-05, "loss": 0.8107, "step": 8504 }, { "epoch": 0.30804056501267657, "grad_norm": 1.3637840858881989, "learning_rate": 1.6212599316272408e-05, "loss": 0.7547, "step": 8505 }, { "epoch": 0.30807678377399494, "grad_norm": 1.3064265118496678, "learning_rate": 1.6211680052323483e-05, "loss": 0.7829, "step": 8506 }, { "epoch": 0.3081130025353133, "grad_norm": 1.325419215189938, "learning_rate": 1.621076070289638e-05, "loss": 0.7409, "step": 8507 }, { "epoch": 0.30814922129663164, "grad_norm": 1.2189420514750002, "learning_rate": 1.6209841268003753e-05, "loss": 0.7181, "step": 8508 }, { "epoch": 0.30818544005795, "grad_norm": 1.3768920932257238, "learning_rate": 1.6208921747658257e-05, "loss": 0.8443, "step": 8509 }, { "epoch": 0.3082216588192684, "grad_norm": 1.3676525149397358, "learning_rate": 1.6208002141872533e-05, "loss": 0.6783, "step": 8510 }, { "epoch": 0.30825787758058676, "grad_norm": 1.3255969420982066, "learning_rate": 1.620708245065925e-05, "loss": 0.7346, "step": 8511 }, { "epoch": 0.30829409634190513, "grad_norm": 1.5210582297967763, "learning_rate": 1.6206162674031062e-05, "loss": 0.7713, "step": 8512 }, { "epoch": 0.30833031510322345, "grad_norm": 1.3893526946120756, "learning_rate": 1.620524281200062e-05, "loss": 0.7608, "step": 8513 }, { "epoch": 0.3083665338645418, "grad_norm": 1.0808768770136774, "learning_rate": 1.6204322864580583e-05, "loss": 0.6822, "step": 8514 }, { "epoch": 0.3084027526258602, "grad_norm": 1.454252839145386, "learning_rate": 1.6203402831783613e-05, "loss": 0.7714, "step": 8515 }, { "epoch": 0.3084389713871786, "grad_norm": 1.4526980731733445, "learning_rate": 1.6202482713622368e-05, "loss": 0.7695, "step": 8516 }, { "epoch": 0.30847519014849695, "grad_norm": 1.0622482360355001, "learning_rate": 1.6201562510109513e-05, "loss": 0.7675, "step": 8517 }, { "epoch": 0.30851140890981527, "grad_norm": 1.0606638159393535, "learning_rate": 1.6200642221257705e-05, "loss": 0.8119, "step": 8518 }, { "epoch": 0.30854762767113364, "grad_norm": 1.4087060197042465, "learning_rate": 1.6199721847079616e-05, "loss": 0.7947, "step": 8519 }, { "epoch": 0.308583846432452, "grad_norm": 1.3272902516096698, "learning_rate": 1.6198801387587905e-05, "loss": 0.8525, "step": 8520 }, { "epoch": 0.3086200651937704, "grad_norm": 1.4209746009365765, "learning_rate": 1.6197880842795243e-05, "loss": 0.8423, "step": 8521 }, { "epoch": 0.30865628395508876, "grad_norm": 1.5133911112381622, "learning_rate": 1.619696021271429e-05, "loss": 0.8448, "step": 8522 }, { "epoch": 0.3086925027164071, "grad_norm": 1.608284526420613, "learning_rate": 1.619603949735772e-05, "loss": 0.72, "step": 8523 }, { "epoch": 0.30872872147772545, "grad_norm": 1.452221597933449, "learning_rate": 1.6195118696738202e-05, "loss": 0.7485, "step": 8524 }, { "epoch": 0.30876494023904383, "grad_norm": 1.2653795477271343, "learning_rate": 1.6194197810868408e-05, "loss": 0.8617, "step": 8525 }, { "epoch": 0.3088011590003622, "grad_norm": 1.3255671876834985, "learning_rate": 1.619327683976101e-05, "loss": 0.783, "step": 8526 }, { "epoch": 0.3088373777616806, "grad_norm": 1.3237874874666304, "learning_rate": 1.6192355783428682e-05, "loss": 0.8255, "step": 8527 }, { "epoch": 0.3088735965229989, "grad_norm": 1.3907489132054904, "learning_rate": 1.6191434641884097e-05, "loss": 0.7611, "step": 8528 }, { "epoch": 0.30890981528431727, "grad_norm": 1.231811383566519, "learning_rate": 1.619051341513993e-05, "loss": 0.7471, "step": 8529 }, { "epoch": 0.30894603404563564, "grad_norm": 1.4865761197861356, "learning_rate": 1.618959210320886e-05, "loss": 0.8056, "step": 8530 }, { "epoch": 0.308982252806954, "grad_norm": 1.4158395552514853, "learning_rate": 1.6188670706103562e-05, "loss": 0.879, "step": 8531 }, { "epoch": 0.3090184715682724, "grad_norm": 1.5152397809149774, "learning_rate": 1.6187749223836718e-05, "loss": 0.7538, "step": 8532 }, { "epoch": 0.3090546903295907, "grad_norm": 1.460149944153013, "learning_rate": 1.6186827656421006e-05, "loss": 0.8269, "step": 8533 }, { "epoch": 0.3090909090909091, "grad_norm": 1.6990839904754216, "learning_rate": 1.6185906003869114e-05, "loss": 0.7398, "step": 8534 }, { "epoch": 0.30912712785222746, "grad_norm": 1.35127756527575, "learning_rate": 1.618498426619372e-05, "loss": 0.8241, "step": 8535 }, { "epoch": 0.30916334661354583, "grad_norm": 1.3742071651124403, "learning_rate": 1.6184062443407503e-05, "loss": 0.8119, "step": 8536 }, { "epoch": 0.3091995653748642, "grad_norm": 1.1745551766217497, "learning_rate": 1.6183140535523154e-05, "loss": 0.8345, "step": 8537 }, { "epoch": 0.3092357841361825, "grad_norm": 1.1086696685710327, "learning_rate": 1.618221854255336e-05, "loss": 0.8052, "step": 8538 }, { "epoch": 0.3092720028975009, "grad_norm": 1.5065295809749777, "learning_rate": 1.6181296464510804e-05, "loss": 0.7995, "step": 8539 }, { "epoch": 0.30930822165881927, "grad_norm": 1.1076454160051568, "learning_rate": 1.618037430140818e-05, "loss": 0.753, "step": 8540 }, { "epoch": 0.30934444042013765, "grad_norm": 1.3856080509830269, "learning_rate": 1.6179452053258174e-05, "loss": 0.7597, "step": 8541 }, { "epoch": 0.309380659181456, "grad_norm": 1.40267507084618, "learning_rate": 1.6178529720073483e-05, "loss": 0.876, "step": 8542 }, { "epoch": 0.30941687794277434, "grad_norm": 1.1776744517795348, "learning_rate": 1.6177607301866786e-05, "loss": 0.751, "step": 8543 }, { "epoch": 0.3094530967040927, "grad_norm": 1.3461533874190645, "learning_rate": 1.6176684798650788e-05, "loss": 0.7517, "step": 8544 }, { "epoch": 0.3094893154654111, "grad_norm": 1.3321100180932373, "learning_rate": 1.617576221043818e-05, "loss": 0.8279, "step": 8545 }, { "epoch": 0.30952553422672946, "grad_norm": 1.4061222005098681, "learning_rate": 1.6174839537241658e-05, "loss": 0.8268, "step": 8546 }, { "epoch": 0.30956175298804783, "grad_norm": 1.4967004726021502, "learning_rate": 1.6173916779073917e-05, "loss": 0.7412, "step": 8547 }, { "epoch": 0.30959797174936615, "grad_norm": 1.4194497362445233, "learning_rate": 1.617299393594766e-05, "loss": 0.7908, "step": 8548 }, { "epoch": 0.3096341905106845, "grad_norm": 1.308501794772608, "learning_rate": 1.6172071007875575e-05, "loss": 0.773, "step": 8549 }, { "epoch": 0.3096704092720029, "grad_norm": 1.3152572111485987, "learning_rate": 1.6171147994870375e-05, "loss": 0.7964, "step": 8550 }, { "epoch": 0.3097066280333213, "grad_norm": 1.055446352667436, "learning_rate": 1.6170224896944755e-05, "loss": 0.7407, "step": 8551 }, { "epoch": 0.30974284679463965, "grad_norm": 1.3053623016058764, "learning_rate": 1.6169301714111418e-05, "loss": 0.7838, "step": 8552 }, { "epoch": 0.30977906555595797, "grad_norm": 1.3268951401662157, "learning_rate": 1.616837844638307e-05, "loss": 0.7158, "step": 8553 }, { "epoch": 0.30981528431727634, "grad_norm": 1.081800768612542, "learning_rate": 1.6167455093772413e-05, "loss": 0.7467, "step": 8554 }, { "epoch": 0.3098515030785947, "grad_norm": 1.249949462807795, "learning_rate": 1.6166531656292156e-05, "loss": 0.7829, "step": 8555 }, { "epoch": 0.3098877218399131, "grad_norm": 1.453280375233998, "learning_rate": 1.6165608133955006e-05, "loss": 0.8633, "step": 8556 }, { "epoch": 0.30992394060123146, "grad_norm": 1.5739959294061725, "learning_rate": 1.616468452677367e-05, "loss": 0.8652, "step": 8557 }, { "epoch": 0.3099601593625498, "grad_norm": 1.4153927505156312, "learning_rate": 1.6163760834760856e-05, "loss": 0.8352, "step": 8558 }, { "epoch": 0.30999637812386815, "grad_norm": 1.0980973932208362, "learning_rate": 1.616283705792928e-05, "loss": 0.7848, "step": 8559 }, { "epoch": 0.31003259688518653, "grad_norm": 1.363092669721672, "learning_rate": 1.6161913196291646e-05, "loss": 0.7337, "step": 8560 }, { "epoch": 0.3100688156465049, "grad_norm": 1.2394954898373434, "learning_rate": 1.6160989249860675e-05, "loss": 0.7516, "step": 8561 }, { "epoch": 0.3101050344078233, "grad_norm": 1.6044336797038616, "learning_rate": 1.6160065218649083e-05, "loss": 0.7981, "step": 8562 }, { "epoch": 0.3101412531691416, "grad_norm": 1.4347994487004525, "learning_rate": 1.6159141102669576e-05, "loss": 0.7647, "step": 8563 }, { "epoch": 0.31017747193045997, "grad_norm": 1.5643658988724474, "learning_rate": 1.6158216901934873e-05, "loss": 0.8526, "step": 8564 }, { "epoch": 0.31021369069177834, "grad_norm": 1.4071064235276547, "learning_rate": 1.6157292616457697e-05, "loss": 0.7421, "step": 8565 }, { "epoch": 0.3102499094530967, "grad_norm": 1.1225736852646087, "learning_rate": 1.6156368246250766e-05, "loss": 0.779, "step": 8566 }, { "epoch": 0.3102861282144151, "grad_norm": 1.679388140460893, "learning_rate": 1.6155443791326798e-05, "loss": 0.81, "step": 8567 }, { "epoch": 0.3103223469757334, "grad_norm": 1.4927179082903688, "learning_rate": 1.615451925169851e-05, "loss": 0.811, "step": 8568 }, { "epoch": 0.3103585657370518, "grad_norm": 1.3939595613396765, "learning_rate": 1.6153594627378636e-05, "loss": 0.7795, "step": 8569 }, { "epoch": 0.31039478449837016, "grad_norm": 1.4536571262219495, "learning_rate": 1.615266991837989e-05, "loss": 0.8593, "step": 8570 }, { "epoch": 0.31043100325968853, "grad_norm": 1.4408269088809236, "learning_rate": 1.6151745124715003e-05, "loss": 0.8471, "step": 8571 }, { "epoch": 0.3104672220210069, "grad_norm": 1.416271712725087, "learning_rate": 1.6150820246396693e-05, "loss": 0.799, "step": 8572 }, { "epoch": 0.3105034407823252, "grad_norm": 1.4353086883789985, "learning_rate": 1.6149895283437694e-05, "loss": 0.7372, "step": 8573 }, { "epoch": 0.3105396595436436, "grad_norm": 1.357025769298952, "learning_rate": 1.6148970235850732e-05, "loss": 0.7927, "step": 8574 }, { "epoch": 0.31057587830496197, "grad_norm": 1.152496810826119, "learning_rate": 1.6148045103648537e-05, "loss": 0.7928, "step": 8575 }, { "epoch": 0.31061209706628035, "grad_norm": 1.5773607400486445, "learning_rate": 1.6147119886843837e-05, "loss": 0.7856, "step": 8576 }, { "epoch": 0.3106483158275987, "grad_norm": 1.4688704509753128, "learning_rate": 1.6146194585449368e-05, "loss": 0.8358, "step": 8577 }, { "epoch": 0.31068453458891704, "grad_norm": 1.352672177866862, "learning_rate": 1.6145269199477856e-05, "loss": 0.7965, "step": 8578 }, { "epoch": 0.3107207533502354, "grad_norm": 1.0970826637597177, "learning_rate": 1.6144343728942048e-05, "loss": 0.7447, "step": 8579 }, { "epoch": 0.3107569721115538, "grad_norm": 1.4685720477959678, "learning_rate": 1.614341817385467e-05, "loss": 0.7977, "step": 8580 }, { "epoch": 0.31079319087287216, "grad_norm": 1.3024741567747369, "learning_rate": 1.6142492534228455e-05, "loss": 0.8705, "step": 8581 }, { "epoch": 0.31082940963419053, "grad_norm": 1.1655698793653777, "learning_rate": 1.614156681007615e-05, "loss": 0.7988, "step": 8582 }, { "epoch": 0.31086562839550885, "grad_norm": 1.354748867964751, "learning_rate": 1.6140641001410487e-05, "loss": 0.8053, "step": 8583 }, { "epoch": 0.3109018471568272, "grad_norm": 1.4876257179882526, "learning_rate": 1.613971510824421e-05, "loss": 0.7796, "step": 8584 }, { "epoch": 0.3109380659181456, "grad_norm": 1.5953218793733404, "learning_rate": 1.6138789130590056e-05, "loss": 0.7308, "step": 8585 }, { "epoch": 0.310974284679464, "grad_norm": 1.07253670891972, "learning_rate": 1.613786306846077e-05, "loss": 0.7554, "step": 8586 }, { "epoch": 0.31101050344078235, "grad_norm": 1.433888035512329, "learning_rate": 1.6136936921869097e-05, "loss": 0.8083, "step": 8587 }, { "epoch": 0.31104672220210067, "grad_norm": 1.0555235682565143, "learning_rate": 1.6136010690827776e-05, "loss": 0.7636, "step": 8588 }, { "epoch": 0.31108294096341904, "grad_norm": 1.605542299576789, "learning_rate": 1.613508437534956e-05, "loss": 0.7899, "step": 8589 }, { "epoch": 0.3111191597247374, "grad_norm": 1.5690719812613054, "learning_rate": 1.6134157975447192e-05, "loss": 0.7866, "step": 8590 }, { "epoch": 0.3111553784860558, "grad_norm": 1.3913414615024264, "learning_rate": 1.6133231491133418e-05, "loss": 0.8928, "step": 8591 }, { "epoch": 0.31119159724737416, "grad_norm": 1.3385388641631133, "learning_rate": 1.613230492242099e-05, "loss": 0.7744, "step": 8592 }, { "epoch": 0.3112278160086925, "grad_norm": 1.5980456560867349, "learning_rate": 1.6131378269322656e-05, "loss": 0.8733, "step": 8593 }, { "epoch": 0.31126403477001086, "grad_norm": 1.10914662601005, "learning_rate": 1.613045153185117e-05, "loss": 0.7802, "step": 8594 }, { "epoch": 0.31130025353132923, "grad_norm": 1.379551011383706, "learning_rate": 1.612952471001929e-05, "loss": 0.7648, "step": 8595 }, { "epoch": 0.3113364722926476, "grad_norm": 1.0332277330310178, "learning_rate": 1.612859780383976e-05, "loss": 0.8196, "step": 8596 }, { "epoch": 0.311372691053966, "grad_norm": 1.4473207646006423, "learning_rate": 1.612767081332534e-05, "loss": 0.8379, "step": 8597 }, { "epoch": 0.3114089098152843, "grad_norm": 1.4781956572107378, "learning_rate": 1.6126743738488782e-05, "loss": 0.8864, "step": 8598 }, { "epoch": 0.31144512857660267, "grad_norm": 1.187223571186899, "learning_rate": 1.6125816579342853e-05, "loss": 0.7336, "step": 8599 }, { "epoch": 0.31148134733792104, "grad_norm": 1.4428667820383574, "learning_rate": 1.6124889335900302e-05, "loss": 0.7744, "step": 8600 }, { "epoch": 0.3115175660992394, "grad_norm": 1.3022766853461045, "learning_rate": 1.612396200817389e-05, "loss": 0.8529, "step": 8601 }, { "epoch": 0.3115537848605578, "grad_norm": 1.3854170647031407, "learning_rate": 1.6123034596176382e-05, "loss": 0.7447, "step": 8602 }, { "epoch": 0.3115900036218761, "grad_norm": 1.3807626701172016, "learning_rate": 1.6122107099920535e-05, "loss": 0.7296, "step": 8603 }, { "epoch": 0.3116262223831945, "grad_norm": 1.1142497686648245, "learning_rate": 1.612117951941912e-05, "loss": 0.7852, "step": 8604 }, { "epoch": 0.31166244114451286, "grad_norm": 1.4877707708522512, "learning_rate": 1.6120251854684892e-05, "loss": 0.806, "step": 8605 }, { "epoch": 0.31169865990583123, "grad_norm": 1.0765353070963006, "learning_rate": 1.6119324105730622e-05, "loss": 0.7457, "step": 8606 }, { "epoch": 0.3117348786671496, "grad_norm": 1.310600985662103, "learning_rate": 1.6118396272569077e-05, "loss": 0.7644, "step": 8607 }, { "epoch": 0.3117710974284679, "grad_norm": 1.4469940191950952, "learning_rate": 1.6117468355213023e-05, "loss": 0.8042, "step": 8608 }, { "epoch": 0.3118073161897863, "grad_norm": 1.4562554812635156, "learning_rate": 1.6116540353675226e-05, "loss": 0.8542, "step": 8609 }, { "epoch": 0.3118435349511047, "grad_norm": 1.3533801978480846, "learning_rate": 1.6115612267968463e-05, "loss": 0.8247, "step": 8610 }, { "epoch": 0.31187975371242305, "grad_norm": 1.3796517063187361, "learning_rate": 1.61146840981055e-05, "loss": 0.8367, "step": 8611 }, { "epoch": 0.3119159724737414, "grad_norm": 1.2016473792325706, "learning_rate": 1.611375584409911e-05, "loss": 0.7558, "step": 8612 }, { "epoch": 0.31195219123505974, "grad_norm": 1.4897592694635429, "learning_rate": 1.6112827505962066e-05, "loss": 0.8621, "step": 8613 }, { "epoch": 0.3119884099963781, "grad_norm": 1.318850761955076, "learning_rate": 1.6111899083707148e-05, "loss": 0.8091, "step": 8614 }, { "epoch": 0.3120246287576965, "grad_norm": 1.3175321335997567, "learning_rate": 1.6110970577347127e-05, "loss": 0.7923, "step": 8615 }, { "epoch": 0.31206084751901486, "grad_norm": 1.3022026216383291, "learning_rate": 1.6110041986894782e-05, "loss": 0.7158, "step": 8616 }, { "epoch": 0.31209706628033324, "grad_norm": 1.3144719900446753, "learning_rate": 1.6109113312362888e-05, "loss": 0.7513, "step": 8617 }, { "epoch": 0.31213328504165155, "grad_norm": 1.4764211665233746, "learning_rate": 1.610818455376423e-05, "loss": 0.8175, "step": 8618 }, { "epoch": 0.3121695038029699, "grad_norm": 1.5550596628001072, "learning_rate": 1.6107255711111584e-05, "loss": 0.7608, "step": 8619 }, { "epoch": 0.3122057225642883, "grad_norm": 1.511585398516932, "learning_rate": 1.6106326784417734e-05, "loss": 0.755, "step": 8620 }, { "epoch": 0.3122419413256067, "grad_norm": 1.487828372274937, "learning_rate": 1.6105397773695458e-05, "loss": 0.7837, "step": 8621 }, { "epoch": 0.31227816008692505, "grad_norm": 1.388693228556101, "learning_rate": 1.610446867895755e-05, "loss": 0.7794, "step": 8622 }, { "epoch": 0.31231437884824337, "grad_norm": 1.4335956521502355, "learning_rate": 1.6103539500216783e-05, "loss": 0.7309, "step": 8623 }, { "epoch": 0.31235059760956174, "grad_norm": 1.4956943835493453, "learning_rate": 1.6102610237485956e-05, "loss": 0.8199, "step": 8624 }, { "epoch": 0.3123868163708801, "grad_norm": 1.116786240851131, "learning_rate": 1.6101680890777846e-05, "loss": 0.754, "step": 8625 }, { "epoch": 0.3124230351321985, "grad_norm": 1.2032225526895255, "learning_rate": 1.6100751460105244e-05, "loss": 0.6797, "step": 8626 }, { "epoch": 0.31245925389351686, "grad_norm": 1.5017571439231787, "learning_rate": 1.6099821945480946e-05, "loss": 0.8063, "step": 8627 }, { "epoch": 0.3124954726548352, "grad_norm": 1.275103454915561, "learning_rate": 1.6098892346917737e-05, "loss": 0.7723, "step": 8628 }, { "epoch": 0.31253169141615356, "grad_norm": 1.180424305810056, "learning_rate": 1.6097962664428405e-05, "loss": 0.7745, "step": 8629 }, { "epoch": 0.31256791017747193, "grad_norm": 1.5299252282482676, "learning_rate": 1.6097032898025756e-05, "loss": 0.7382, "step": 8630 }, { "epoch": 0.3126041289387903, "grad_norm": 1.0574727547029386, "learning_rate": 1.6096103047722573e-05, "loss": 0.7663, "step": 8631 }, { "epoch": 0.3126403477001087, "grad_norm": 1.6474924802445732, "learning_rate": 1.6095173113531655e-05, "loss": 0.7813, "step": 8632 }, { "epoch": 0.312676566461427, "grad_norm": 1.388087736401735, "learning_rate": 1.60942430954658e-05, "loss": 0.7776, "step": 8633 }, { "epoch": 0.31271278522274537, "grad_norm": 1.4637346786675274, "learning_rate": 1.6093312993537806e-05, "loss": 0.7832, "step": 8634 }, { "epoch": 0.31274900398406374, "grad_norm": 1.1336836449624177, "learning_rate": 1.609238280776047e-05, "loss": 0.7922, "step": 8635 }, { "epoch": 0.3127852227453821, "grad_norm": 1.0366129344231851, "learning_rate": 1.6091452538146593e-05, "loss": 0.7498, "step": 8636 }, { "epoch": 0.3128214415067005, "grad_norm": 1.2755133744938125, "learning_rate": 1.609052218470898e-05, "loss": 0.8522, "step": 8637 }, { "epoch": 0.3128576602680188, "grad_norm": 1.4252709929055078, "learning_rate": 1.6089591747460428e-05, "loss": 0.7764, "step": 8638 }, { "epoch": 0.3128938790293372, "grad_norm": 1.0581295680059717, "learning_rate": 1.608866122641374e-05, "loss": 0.772, "step": 8639 }, { "epoch": 0.31293009779065556, "grad_norm": 1.5976316827876482, "learning_rate": 1.608773062158173e-05, "loss": 0.9182, "step": 8640 }, { "epoch": 0.31296631655197393, "grad_norm": 1.5060627177848933, "learning_rate": 1.6086799932977193e-05, "loss": 0.8081, "step": 8641 }, { "epoch": 0.3130025353132923, "grad_norm": 1.2665137529449988, "learning_rate": 1.608586916061294e-05, "loss": 0.7235, "step": 8642 }, { "epoch": 0.3130387540746106, "grad_norm": 1.331078080652537, "learning_rate": 1.608493830450178e-05, "loss": 0.739, "step": 8643 }, { "epoch": 0.313074972835929, "grad_norm": 1.5521541666785819, "learning_rate": 1.6084007364656523e-05, "loss": 0.8034, "step": 8644 }, { "epoch": 0.3131111915972474, "grad_norm": 1.3472755703015677, "learning_rate": 1.608307634108998e-05, "loss": 0.7982, "step": 8645 }, { "epoch": 0.31314741035856575, "grad_norm": 1.288134722910107, "learning_rate": 1.608214523381496e-05, "loss": 0.7381, "step": 8646 }, { "epoch": 0.3131836291198841, "grad_norm": 1.4420833399438633, "learning_rate": 1.608121404284428e-05, "loss": 0.7994, "step": 8647 }, { "epoch": 0.31321984788120244, "grad_norm": 1.3182094757197493, "learning_rate": 1.6080282768190748e-05, "loss": 0.7868, "step": 8648 }, { "epoch": 0.3132560666425208, "grad_norm": 1.1680969168756772, "learning_rate": 1.6079351409867185e-05, "loss": 0.8615, "step": 8649 }, { "epoch": 0.3132922854038392, "grad_norm": 1.3377910838025793, "learning_rate": 1.6078419967886402e-05, "loss": 0.8805, "step": 8650 }, { "epoch": 0.31332850416515756, "grad_norm": 1.5482250400833766, "learning_rate": 1.6077488442261218e-05, "loss": 0.8898, "step": 8651 }, { "epoch": 0.31336472292647594, "grad_norm": 1.3436411533517698, "learning_rate": 1.607655683300446e-05, "loss": 0.8136, "step": 8652 }, { "epoch": 0.31340094168779425, "grad_norm": 1.372861716753487, "learning_rate": 1.6075625140128934e-05, "loss": 0.7967, "step": 8653 }, { "epoch": 0.31343716044911263, "grad_norm": 1.5216042489526767, "learning_rate": 1.607469336364747e-05, "loss": 0.7689, "step": 8654 }, { "epoch": 0.313473379210431, "grad_norm": 1.3577238858392497, "learning_rate": 1.607376150357289e-05, "loss": 0.7329, "step": 8655 }, { "epoch": 0.3135095979717494, "grad_norm": 1.3681533644242296, "learning_rate": 1.607282955991801e-05, "loss": 0.7854, "step": 8656 }, { "epoch": 0.31354581673306775, "grad_norm": 1.3505360924438234, "learning_rate": 1.6071897532695662e-05, "loss": 0.773, "step": 8657 }, { "epoch": 0.31358203549438607, "grad_norm": 1.4717231829809156, "learning_rate": 1.6070965421918668e-05, "loss": 0.7629, "step": 8658 }, { "epoch": 0.31361825425570444, "grad_norm": 1.3509571464503338, "learning_rate": 1.6070033227599858e-05, "loss": 0.8027, "step": 8659 }, { "epoch": 0.3136544730170228, "grad_norm": 1.5469407190465314, "learning_rate": 1.6069100949752052e-05, "loss": 0.838, "step": 8660 }, { "epoch": 0.3136906917783412, "grad_norm": 1.4071654161541312, "learning_rate": 1.6068168588388088e-05, "loss": 0.7485, "step": 8661 }, { "epoch": 0.31372691053965956, "grad_norm": 1.4332850641016674, "learning_rate": 1.606723614352079e-05, "loss": 0.8076, "step": 8662 }, { "epoch": 0.3137631293009779, "grad_norm": 1.4139072939950248, "learning_rate": 1.6066303615162994e-05, "loss": 0.8033, "step": 8663 }, { "epoch": 0.31379934806229626, "grad_norm": 1.4106551566405463, "learning_rate": 1.606537100332753e-05, "loss": 0.7715, "step": 8664 }, { "epoch": 0.31383556682361463, "grad_norm": 1.1669844317876799, "learning_rate": 1.6064438308027227e-05, "loss": 0.7518, "step": 8665 }, { "epoch": 0.313871785584933, "grad_norm": 1.5666138185640661, "learning_rate": 1.606350552927493e-05, "loss": 0.8497, "step": 8666 }, { "epoch": 0.3139080043462514, "grad_norm": 1.4018088824305288, "learning_rate": 1.606257266708347e-05, "loss": 0.8344, "step": 8667 }, { "epoch": 0.3139442231075697, "grad_norm": 1.4077543795779985, "learning_rate": 1.6061639721465675e-05, "loss": 0.8027, "step": 8668 }, { "epoch": 0.31398044186888807, "grad_norm": 1.3863218276718428, "learning_rate": 1.6060706692434398e-05, "loss": 0.7047, "step": 8669 }, { "epoch": 0.31401666063020645, "grad_norm": 1.2861482068800933, "learning_rate": 1.6059773580002468e-05, "loss": 0.7687, "step": 8670 }, { "epoch": 0.3140528793915248, "grad_norm": 1.4490905445384503, "learning_rate": 1.6058840384182732e-05, "loss": 0.7904, "step": 8671 }, { "epoch": 0.3140890981528432, "grad_norm": 1.300590161730438, "learning_rate": 1.6057907104988026e-05, "loss": 0.7649, "step": 8672 }, { "epoch": 0.3141253169141615, "grad_norm": 1.5752789261745226, "learning_rate": 1.6056973742431194e-05, "loss": 0.8288, "step": 8673 }, { "epoch": 0.3141615356754799, "grad_norm": 1.181219805903649, "learning_rate": 1.6056040296525086e-05, "loss": 0.7437, "step": 8674 }, { "epoch": 0.31419775443679826, "grad_norm": 1.3972625758250055, "learning_rate": 1.6055106767282537e-05, "loss": 0.7025, "step": 8675 }, { "epoch": 0.31423397319811663, "grad_norm": 1.4525910429625557, "learning_rate": 1.60541731547164e-05, "loss": 0.8289, "step": 8676 }, { "epoch": 0.314270191959435, "grad_norm": 2.1686720386091287, "learning_rate": 1.6053239458839524e-05, "loss": 0.752, "step": 8677 }, { "epoch": 0.3143064107207533, "grad_norm": 1.3259858521845322, "learning_rate": 1.6052305679664752e-05, "loss": 0.812, "step": 8678 }, { "epoch": 0.3143426294820717, "grad_norm": 1.3637951603662095, "learning_rate": 1.6051371817204934e-05, "loss": 0.821, "step": 8679 }, { "epoch": 0.3143788482433901, "grad_norm": 1.3962030020597345, "learning_rate": 1.6050437871472927e-05, "loss": 0.7967, "step": 8680 }, { "epoch": 0.31441506700470845, "grad_norm": 1.3271461343640012, "learning_rate": 1.6049503842481576e-05, "loss": 0.6435, "step": 8681 }, { "epoch": 0.3144512857660268, "grad_norm": 1.4385361060245208, "learning_rate": 1.6048569730243735e-05, "loss": 0.835, "step": 8682 }, { "epoch": 0.31448750452734514, "grad_norm": 1.5293543331026613, "learning_rate": 1.604763553477226e-05, "loss": 0.7457, "step": 8683 }, { "epoch": 0.3145237232886635, "grad_norm": 1.6361121439689896, "learning_rate": 1.6046701256080008e-05, "loss": 0.835, "step": 8684 }, { "epoch": 0.3145599420499819, "grad_norm": 1.445756210057541, "learning_rate": 1.6045766894179836e-05, "loss": 0.7969, "step": 8685 }, { "epoch": 0.31459616081130026, "grad_norm": 1.2731459617021499, "learning_rate": 1.6044832449084594e-05, "loss": 0.7634, "step": 8686 }, { "epoch": 0.31463237957261864, "grad_norm": 1.3527126118283233, "learning_rate": 1.6043897920807148e-05, "loss": 0.7562, "step": 8687 }, { "epoch": 0.31466859833393696, "grad_norm": 1.5141130616590455, "learning_rate": 1.6042963309360357e-05, "loss": 0.7882, "step": 8688 }, { "epoch": 0.31470481709525533, "grad_norm": 1.4526109542485213, "learning_rate": 1.604202861475708e-05, "loss": 0.7548, "step": 8689 }, { "epoch": 0.3147410358565737, "grad_norm": 1.2335953906327997, "learning_rate": 1.604109383701018e-05, "loss": 0.7794, "step": 8690 }, { "epoch": 0.3147772546178921, "grad_norm": 1.4685259759782394, "learning_rate": 1.604015897613252e-05, "loss": 0.8169, "step": 8691 }, { "epoch": 0.31481347337921045, "grad_norm": 1.2178342842920555, "learning_rate": 1.6039224032136965e-05, "loss": 0.8033, "step": 8692 }, { "epoch": 0.31484969214052877, "grad_norm": 1.1071163303987244, "learning_rate": 1.6038289005036385e-05, "loss": 0.672, "step": 8693 }, { "epoch": 0.31488591090184714, "grad_norm": 1.4276191272955832, "learning_rate": 1.603735389484364e-05, "loss": 0.7972, "step": 8694 }, { "epoch": 0.3149221296631655, "grad_norm": 1.3602570400274512, "learning_rate": 1.60364187015716e-05, "loss": 0.7916, "step": 8695 }, { "epoch": 0.3149583484244839, "grad_norm": 1.4218026263293972, "learning_rate": 1.6035483425233134e-05, "loss": 0.8149, "step": 8696 }, { "epoch": 0.31499456718580227, "grad_norm": 1.5250000300954596, "learning_rate": 1.6034548065841113e-05, "loss": 0.8264, "step": 8697 }, { "epoch": 0.3150307859471206, "grad_norm": 1.3562060042740698, "learning_rate": 1.603361262340841e-05, "loss": 0.7106, "step": 8698 }, { "epoch": 0.31506700470843896, "grad_norm": 1.2901005214111685, "learning_rate": 1.6032677097947895e-05, "loss": 0.7819, "step": 8699 }, { "epoch": 0.31510322346975733, "grad_norm": 1.257965355153538, "learning_rate": 1.6031741489472442e-05, "loss": 0.717, "step": 8700 }, { "epoch": 0.3151394422310757, "grad_norm": 1.4508198956861564, "learning_rate": 1.6030805797994923e-05, "loss": 0.8025, "step": 8701 }, { "epoch": 0.3151756609923941, "grad_norm": 1.4856794071987296, "learning_rate": 1.6029870023528222e-05, "loss": 0.8049, "step": 8702 }, { "epoch": 0.3152118797537124, "grad_norm": 1.3144797192075433, "learning_rate": 1.602893416608521e-05, "loss": 0.7529, "step": 8703 }, { "epoch": 0.3152480985150308, "grad_norm": 1.5384148769741572, "learning_rate": 1.6027998225678763e-05, "loss": 0.8302, "step": 8704 }, { "epoch": 0.31528431727634915, "grad_norm": 1.3254635791616998, "learning_rate": 1.6027062202321765e-05, "loss": 0.8262, "step": 8705 }, { "epoch": 0.3153205360376675, "grad_norm": 1.3163849640718546, "learning_rate": 1.60261260960271e-05, "loss": 0.8009, "step": 8706 }, { "epoch": 0.3153567547989859, "grad_norm": 1.4348916511220846, "learning_rate": 1.6025189906807642e-05, "loss": 0.8379, "step": 8707 }, { "epoch": 0.3153929735603042, "grad_norm": 1.5116239057157765, "learning_rate": 1.6024253634676275e-05, "loss": 0.8222, "step": 8708 }, { "epoch": 0.3154291923216226, "grad_norm": 1.4634338759343422, "learning_rate": 1.6023317279645887e-05, "loss": 0.8566, "step": 8709 }, { "epoch": 0.31546541108294096, "grad_norm": 1.3712370884516965, "learning_rate": 1.602238084172936e-05, "loss": 0.7108, "step": 8710 }, { "epoch": 0.31550162984425933, "grad_norm": 1.408409379994098, "learning_rate": 1.602144432093958e-05, "loss": 0.7913, "step": 8711 }, { "epoch": 0.3155378486055777, "grad_norm": 1.3920093108833154, "learning_rate": 1.6020507717289434e-05, "loss": 0.8413, "step": 8712 }, { "epoch": 0.315574067366896, "grad_norm": 1.3113426877844667, "learning_rate": 1.6019571030791815e-05, "loss": 0.7155, "step": 8713 }, { "epoch": 0.3156102861282144, "grad_norm": 1.5569911643465404, "learning_rate": 1.601863426145961e-05, "loss": 0.7782, "step": 8714 }, { "epoch": 0.3156465048895328, "grad_norm": 1.608271551541207, "learning_rate": 1.6017697409305703e-05, "loss": 0.7811, "step": 8715 }, { "epoch": 0.31568272365085115, "grad_norm": 1.0887186892137928, "learning_rate": 1.6016760474342997e-05, "loss": 0.7651, "step": 8716 }, { "epoch": 0.3157189424121695, "grad_norm": 1.4045483771160583, "learning_rate": 1.601582345658438e-05, "loss": 0.8364, "step": 8717 }, { "epoch": 0.31575516117348784, "grad_norm": 1.7628636837165803, "learning_rate": 1.6014886356042744e-05, "loss": 0.7787, "step": 8718 }, { "epoch": 0.3157913799348062, "grad_norm": 1.2609513743690055, "learning_rate": 1.6013949172730987e-05, "loss": 0.753, "step": 8719 }, { "epoch": 0.3158275986961246, "grad_norm": 1.036495788060603, "learning_rate": 1.6013011906662008e-05, "loss": 0.8264, "step": 8720 }, { "epoch": 0.31586381745744296, "grad_norm": 1.4433167042186799, "learning_rate": 1.6012074557848697e-05, "loss": 0.7947, "step": 8721 }, { "epoch": 0.31590003621876134, "grad_norm": 1.4350993178885327, "learning_rate": 1.601113712630396e-05, "loss": 0.847, "step": 8722 }, { "epoch": 0.31593625498007966, "grad_norm": 1.47293039573029, "learning_rate": 1.6010199612040693e-05, "loss": 0.8232, "step": 8723 }, { "epoch": 0.31597247374139803, "grad_norm": 1.266224620129682, "learning_rate": 1.60092620150718e-05, "loss": 0.7681, "step": 8724 }, { "epoch": 0.3160086925027164, "grad_norm": 2.6852780183247207, "learning_rate": 1.6008324335410183e-05, "loss": 0.7751, "step": 8725 }, { "epoch": 0.3160449112640348, "grad_norm": 1.321916381941216, "learning_rate": 1.6007386573068743e-05, "loss": 0.7544, "step": 8726 }, { "epoch": 0.31608113002535315, "grad_norm": 1.388924009769827, "learning_rate": 1.600644872806038e-05, "loss": 0.7714, "step": 8727 }, { "epoch": 0.31611734878667147, "grad_norm": 1.3599275895669949, "learning_rate": 1.600551080039801e-05, "loss": 0.8255, "step": 8728 }, { "epoch": 0.31615356754798984, "grad_norm": 1.4790448122201827, "learning_rate": 1.6004572790094535e-05, "loss": 0.7828, "step": 8729 }, { "epoch": 0.3161897863093082, "grad_norm": 1.426977635446437, "learning_rate": 1.600363469716286e-05, "loss": 0.8003, "step": 8730 }, { "epoch": 0.3162260050706266, "grad_norm": 1.3540240574137505, "learning_rate": 1.6002696521615897e-05, "loss": 0.7868, "step": 8731 }, { "epoch": 0.31626222383194497, "grad_norm": 1.1702671935489712, "learning_rate": 1.6001758263466554e-05, "loss": 0.7093, "step": 8732 }, { "epoch": 0.3162984425932633, "grad_norm": 1.2683168596508956, "learning_rate": 1.6000819922727745e-05, "loss": 0.8214, "step": 8733 }, { "epoch": 0.31633466135458166, "grad_norm": 1.3046433186474191, "learning_rate": 1.599988149941238e-05, "loss": 0.7642, "step": 8734 }, { "epoch": 0.31637088011590003, "grad_norm": 1.3965891465318643, "learning_rate": 1.599894299353338e-05, "loss": 0.7697, "step": 8735 }, { "epoch": 0.3164070988772184, "grad_norm": 1.324772162476143, "learning_rate": 1.5998004405103644e-05, "loss": 0.7933, "step": 8736 }, { "epoch": 0.3164433176385368, "grad_norm": 1.7131001073607663, "learning_rate": 1.5997065734136104e-05, "loss": 0.7673, "step": 8737 }, { "epoch": 0.3164795363998551, "grad_norm": 1.2078640870807893, "learning_rate": 1.5996126980643665e-05, "loss": 0.7841, "step": 8738 }, { "epoch": 0.3165157551611735, "grad_norm": 1.478821121965846, "learning_rate": 1.599518814463925e-05, "loss": 0.7673, "step": 8739 }, { "epoch": 0.31655197392249185, "grad_norm": 1.1198786040360107, "learning_rate": 1.599424922613578e-05, "loss": 0.8095, "step": 8740 }, { "epoch": 0.3165881926838102, "grad_norm": 2.033072562015309, "learning_rate": 1.5993310225146178e-05, "loss": 0.7718, "step": 8741 }, { "epoch": 0.3166244114451286, "grad_norm": 1.4506269705194863, "learning_rate": 1.5992371141683357e-05, "loss": 0.8099, "step": 8742 }, { "epoch": 0.3166606302064469, "grad_norm": 1.4904888305807014, "learning_rate": 1.5991431975760243e-05, "loss": 0.8382, "step": 8743 }, { "epoch": 0.3166968489677653, "grad_norm": 1.3680652307705061, "learning_rate": 1.599049272738976e-05, "loss": 0.82, "step": 8744 }, { "epoch": 0.31673306772908366, "grad_norm": 1.3981058697165487, "learning_rate": 1.5989553396584836e-05, "loss": 0.8452, "step": 8745 }, { "epoch": 0.31676928649040204, "grad_norm": 1.4703987789480277, "learning_rate": 1.5988613983358394e-05, "loss": 0.7846, "step": 8746 }, { "epoch": 0.3168055052517204, "grad_norm": 1.4882549649017518, "learning_rate": 1.5987674487723358e-05, "loss": 0.8807, "step": 8747 }, { "epoch": 0.31684172401303873, "grad_norm": 1.750505889420952, "learning_rate": 1.5986734909692664e-05, "loss": 0.8349, "step": 8748 }, { "epoch": 0.3168779427743571, "grad_norm": 1.310479236950545, "learning_rate": 1.5985795249279237e-05, "loss": 0.8329, "step": 8749 }, { "epoch": 0.3169141615356755, "grad_norm": 1.3823838242993298, "learning_rate": 1.5984855506496006e-05, "loss": 0.8278, "step": 8750 }, { "epoch": 0.31695038029699385, "grad_norm": 1.3296343025118078, "learning_rate": 1.5983915681355907e-05, "loss": 0.7091, "step": 8751 }, { "epoch": 0.3169865990583122, "grad_norm": 1.4269808615112107, "learning_rate": 1.5982975773871866e-05, "loss": 0.7803, "step": 8752 }, { "epoch": 0.31702281781963054, "grad_norm": 1.3610549581537459, "learning_rate": 1.5982035784056823e-05, "loss": 0.8107, "step": 8753 }, { "epoch": 0.3170590365809489, "grad_norm": 1.3477321105835405, "learning_rate": 1.5981095711923713e-05, "loss": 0.7492, "step": 8754 }, { "epoch": 0.3170952553422673, "grad_norm": 1.2641085646951034, "learning_rate": 1.598015555748547e-05, "loss": 0.7186, "step": 8755 }, { "epoch": 0.31713147410358566, "grad_norm": 1.0731286710403212, "learning_rate": 1.597921532075503e-05, "loss": 0.7956, "step": 8756 }, { "epoch": 0.31716769286490404, "grad_norm": 1.448667947251285, "learning_rate": 1.5978275001745336e-05, "loss": 0.7875, "step": 8757 }, { "epoch": 0.31720391162622236, "grad_norm": 1.0730846383108898, "learning_rate": 1.5977334600469324e-05, "loss": 0.7428, "step": 8758 }, { "epoch": 0.31724013038754073, "grad_norm": 1.6019922191702376, "learning_rate": 1.5976394116939934e-05, "loss": 0.7872, "step": 8759 }, { "epoch": 0.3172763491488591, "grad_norm": 1.4098715943959856, "learning_rate": 1.5975453551170113e-05, "loss": 0.7808, "step": 8760 }, { "epoch": 0.3173125679101775, "grad_norm": 1.4382435058881535, "learning_rate": 1.5974512903172797e-05, "loss": 0.8248, "step": 8761 }, { "epoch": 0.31734878667149585, "grad_norm": 1.6844295178192423, "learning_rate": 1.5973572172960935e-05, "loss": 0.8012, "step": 8762 }, { "epoch": 0.31738500543281417, "grad_norm": 1.3073497004343348, "learning_rate": 1.5972631360547473e-05, "loss": 0.8166, "step": 8763 }, { "epoch": 0.31742122419413255, "grad_norm": 1.5106601495543914, "learning_rate": 1.5971690465945354e-05, "loss": 0.786, "step": 8764 }, { "epoch": 0.3174574429554509, "grad_norm": 1.464492600632134, "learning_rate": 1.5970749489167528e-05, "loss": 0.7985, "step": 8765 }, { "epoch": 0.3174936617167693, "grad_norm": 1.3670793292986507, "learning_rate": 1.596980843022694e-05, "loss": 0.7785, "step": 8766 }, { "epoch": 0.31752988047808767, "grad_norm": 1.330443763330282, "learning_rate": 1.5968867289136545e-05, "loss": 0.7687, "step": 8767 }, { "epoch": 0.317566099239406, "grad_norm": 1.6247832435690472, "learning_rate": 1.5967926065909288e-05, "loss": 0.7954, "step": 8768 }, { "epoch": 0.31760231800072436, "grad_norm": 1.1456923843227274, "learning_rate": 1.596698476055813e-05, "loss": 0.7204, "step": 8769 }, { "epoch": 0.31763853676204273, "grad_norm": 1.3320469442824556, "learning_rate": 1.5966043373096018e-05, "loss": 0.7964, "step": 8770 }, { "epoch": 0.3176747555233611, "grad_norm": 1.3039870537667566, "learning_rate": 1.5965101903535903e-05, "loss": 0.772, "step": 8771 }, { "epoch": 0.3177109742846795, "grad_norm": 1.4140667118357533, "learning_rate": 1.5964160351890743e-05, "loss": 0.7495, "step": 8772 }, { "epoch": 0.3177471930459978, "grad_norm": 1.5995539840777409, "learning_rate": 1.5963218718173502e-05, "loss": 0.8379, "step": 8773 }, { "epoch": 0.3177834118073162, "grad_norm": 1.4203218368369213, "learning_rate": 1.596227700239713e-05, "loss": 0.7308, "step": 8774 }, { "epoch": 0.31781963056863455, "grad_norm": 1.4562783398909709, "learning_rate": 1.5961335204574586e-05, "loss": 0.7814, "step": 8775 }, { "epoch": 0.3178558493299529, "grad_norm": 1.3721478789861579, "learning_rate": 1.596039332471883e-05, "loss": 0.7508, "step": 8776 }, { "epoch": 0.3178920680912713, "grad_norm": 1.521991653128012, "learning_rate": 1.5959451362842827e-05, "loss": 0.7723, "step": 8777 }, { "epoch": 0.3179282868525896, "grad_norm": 1.3984968110245677, "learning_rate": 1.5958509318959537e-05, "loss": 0.753, "step": 8778 }, { "epoch": 0.317964505613908, "grad_norm": 1.5018648349820392, "learning_rate": 1.5957567193081924e-05, "loss": 0.8781, "step": 8779 }, { "epoch": 0.31800072437522636, "grad_norm": 1.3709422826650937, "learning_rate": 1.595662498522295e-05, "loss": 0.7297, "step": 8780 }, { "epoch": 0.31803694313654474, "grad_norm": 1.4526766590429796, "learning_rate": 1.595568269539558e-05, "loss": 0.8207, "step": 8781 }, { "epoch": 0.3180731618978631, "grad_norm": 1.5324055000109016, "learning_rate": 1.5954740323612784e-05, "loss": 0.8377, "step": 8782 }, { "epoch": 0.31810938065918143, "grad_norm": 1.7245064787879856, "learning_rate": 1.595379786988753e-05, "loss": 0.7795, "step": 8783 }, { "epoch": 0.3181455994204998, "grad_norm": 1.4313123659742923, "learning_rate": 1.5952855334232785e-05, "loss": 0.7994, "step": 8784 }, { "epoch": 0.3181818181818182, "grad_norm": 1.5135451909857995, "learning_rate": 1.595191271666152e-05, "loss": 0.7177, "step": 8785 }, { "epoch": 0.31821803694313655, "grad_norm": 1.4904343882034072, "learning_rate": 1.595097001718671e-05, "loss": 0.8075, "step": 8786 }, { "epoch": 0.3182542557044549, "grad_norm": 1.3469767105921975, "learning_rate": 1.595002723582132e-05, "loss": 0.7977, "step": 8787 }, { "epoch": 0.31829047446577324, "grad_norm": 1.4947071185829328, "learning_rate": 1.594908437257833e-05, "loss": 0.7952, "step": 8788 }, { "epoch": 0.3183266932270916, "grad_norm": 1.236043827121108, "learning_rate": 1.5948141427470708e-05, "loss": 0.7698, "step": 8789 }, { "epoch": 0.31836291198841, "grad_norm": 1.4479423065854151, "learning_rate": 1.5947198400511438e-05, "loss": 0.7861, "step": 8790 }, { "epoch": 0.31839913074972837, "grad_norm": 1.4494108448062246, "learning_rate": 1.594625529171349e-05, "loss": 0.7356, "step": 8791 }, { "epoch": 0.31843534951104674, "grad_norm": 1.099263209891452, "learning_rate": 1.594531210108984e-05, "loss": 0.7739, "step": 8792 }, { "epoch": 0.31847156827236506, "grad_norm": 1.0639546058712805, "learning_rate": 1.5944368828653476e-05, "loss": 0.8078, "step": 8793 }, { "epoch": 0.31850778703368343, "grad_norm": 1.501285224653367, "learning_rate": 1.5943425474417376e-05, "loss": 0.8271, "step": 8794 }, { "epoch": 0.3185440057950018, "grad_norm": 1.5432622195873476, "learning_rate": 1.5942482038394515e-05, "loss": 0.8125, "step": 8795 }, { "epoch": 0.3185802245563202, "grad_norm": 1.4791691713888382, "learning_rate": 1.594153852059788e-05, "loss": 0.8439, "step": 8796 }, { "epoch": 0.31861644331763855, "grad_norm": 1.105274676353251, "learning_rate": 1.5940594921040457e-05, "loss": 0.7328, "step": 8797 }, { "epoch": 0.31865266207895687, "grad_norm": 1.0821168975703381, "learning_rate": 1.5939651239735225e-05, "loss": 0.7487, "step": 8798 }, { "epoch": 0.31868888084027525, "grad_norm": 1.5413875789690927, "learning_rate": 1.5938707476695173e-05, "loss": 0.7876, "step": 8799 }, { "epoch": 0.3187250996015936, "grad_norm": 1.5061765984797972, "learning_rate": 1.5937763631933288e-05, "loss": 0.7251, "step": 8800 }, { "epoch": 0.318761318362912, "grad_norm": 1.391038063803529, "learning_rate": 1.593681970546256e-05, "loss": 0.869, "step": 8801 }, { "epoch": 0.31879753712423037, "grad_norm": 1.560791682734437, "learning_rate": 1.5935875697295973e-05, "loss": 0.7894, "step": 8802 }, { "epoch": 0.31883375588554874, "grad_norm": 1.2501479202904129, "learning_rate": 1.5934931607446524e-05, "loss": 0.7086, "step": 8803 }, { "epoch": 0.31886997464686706, "grad_norm": 1.5531638613104681, "learning_rate": 1.5933987435927202e-05, "loss": 0.8688, "step": 8804 }, { "epoch": 0.31890619340818543, "grad_norm": 1.3738693863998976, "learning_rate": 1.5933043182750996e-05, "loss": 0.6874, "step": 8805 }, { "epoch": 0.3189424121695038, "grad_norm": 1.2561459905785057, "learning_rate": 1.5932098847930905e-05, "loss": 0.7615, "step": 8806 }, { "epoch": 0.3189786309308222, "grad_norm": 1.4015052474862209, "learning_rate": 1.5931154431479917e-05, "loss": 0.8475, "step": 8807 }, { "epoch": 0.31901484969214056, "grad_norm": 1.5027945261281195, "learning_rate": 1.5930209933411036e-05, "loss": 0.8101, "step": 8808 }, { "epoch": 0.3190510684534589, "grad_norm": 1.3374686515445486, "learning_rate": 1.5929265353737253e-05, "loss": 0.8561, "step": 8809 }, { "epoch": 0.31908728721477725, "grad_norm": 1.3589713906032161, "learning_rate": 1.5928320692471572e-05, "loss": 0.7588, "step": 8810 }, { "epoch": 0.3191235059760956, "grad_norm": 1.3252414759984443, "learning_rate": 1.5927375949626988e-05, "loss": 0.8024, "step": 8811 }, { "epoch": 0.319159724737414, "grad_norm": 1.2609226679727301, "learning_rate": 1.5926431125216503e-05, "loss": 0.7303, "step": 8812 }, { "epoch": 0.31919594349873237, "grad_norm": 1.3758741534252064, "learning_rate": 1.5925486219253117e-05, "loss": 0.7958, "step": 8813 }, { "epoch": 0.3192321622600507, "grad_norm": 1.3343143704250853, "learning_rate": 1.5924541231749836e-05, "loss": 0.7836, "step": 8814 }, { "epoch": 0.31926838102136906, "grad_norm": 1.3752985617950462, "learning_rate": 1.592359616271966e-05, "loss": 0.82, "step": 8815 }, { "epoch": 0.31930459978268744, "grad_norm": 1.415384422237489, "learning_rate": 1.5922651012175596e-05, "loss": 0.8508, "step": 8816 }, { "epoch": 0.3193408185440058, "grad_norm": 1.2995540640950762, "learning_rate": 1.592170578013065e-05, "loss": 0.8043, "step": 8817 }, { "epoch": 0.3193770373053242, "grad_norm": 1.4994866752738094, "learning_rate": 1.5920760466597832e-05, "loss": 0.7971, "step": 8818 }, { "epoch": 0.3194132560666425, "grad_norm": 1.3469632163553968, "learning_rate": 1.5919815071590145e-05, "loss": 0.6966, "step": 8819 }, { "epoch": 0.3194494748279609, "grad_norm": 1.479046352176624, "learning_rate": 1.59188695951206e-05, "loss": 0.8092, "step": 8820 }, { "epoch": 0.31948569358927925, "grad_norm": 1.3034850517776615, "learning_rate": 1.591792403720221e-05, "loss": 0.77, "step": 8821 }, { "epoch": 0.3195219123505976, "grad_norm": 1.4386728040229724, "learning_rate": 1.5916978397847987e-05, "loss": 0.8284, "step": 8822 }, { "epoch": 0.319558131111916, "grad_norm": 1.354440215034346, "learning_rate": 1.591603267707094e-05, "loss": 0.8037, "step": 8823 }, { "epoch": 0.3195943498732343, "grad_norm": 1.4118774825146496, "learning_rate": 1.5915086874884086e-05, "loss": 0.8827, "step": 8824 }, { "epoch": 0.3196305686345527, "grad_norm": 1.354769940903687, "learning_rate": 1.5914140991300436e-05, "loss": 0.8174, "step": 8825 }, { "epoch": 0.31966678739587107, "grad_norm": 1.346768900811827, "learning_rate": 1.5913195026333016e-05, "loss": 0.7643, "step": 8826 }, { "epoch": 0.31970300615718944, "grad_norm": 1.3053945214657838, "learning_rate": 1.591224897999483e-05, "loss": 0.8249, "step": 8827 }, { "epoch": 0.3197392249185078, "grad_norm": 1.4989697288310213, "learning_rate": 1.5911302852298907e-05, "loss": 0.8195, "step": 8828 }, { "epoch": 0.31977544367982613, "grad_norm": 1.432365496134228, "learning_rate": 1.5910356643258263e-05, "loss": 0.848, "step": 8829 }, { "epoch": 0.3198116624411445, "grad_norm": 1.0379158485805509, "learning_rate": 1.5909410352885914e-05, "loss": 0.7505, "step": 8830 }, { "epoch": 0.3198478812024629, "grad_norm": 1.2628787653124578, "learning_rate": 1.590846398119489e-05, "loss": 0.7148, "step": 8831 }, { "epoch": 0.31988409996378125, "grad_norm": 1.3258133556022387, "learning_rate": 1.590751752819821e-05, "loss": 0.7114, "step": 8832 }, { "epoch": 0.31992031872509963, "grad_norm": 1.50464354437999, "learning_rate": 1.5906570993908896e-05, "loss": 0.8237, "step": 8833 }, { "epoch": 0.31995653748641795, "grad_norm": 1.341442856875115, "learning_rate": 1.5905624378339976e-05, "loss": 0.8092, "step": 8834 }, { "epoch": 0.3199927562477363, "grad_norm": 1.4666681931656602, "learning_rate": 1.5904677681504475e-05, "loss": 0.8079, "step": 8835 }, { "epoch": 0.3200289750090547, "grad_norm": 1.5476714995189917, "learning_rate": 1.590373090341542e-05, "loss": 0.8243, "step": 8836 }, { "epoch": 0.32006519377037307, "grad_norm": 1.3769934377788788, "learning_rate": 1.590278404408584e-05, "loss": 0.7972, "step": 8837 }, { "epoch": 0.32010141253169144, "grad_norm": 1.5643080704839039, "learning_rate": 1.5901837103528767e-05, "loss": 0.8435, "step": 8838 }, { "epoch": 0.32013763129300976, "grad_norm": 1.1941755564141112, "learning_rate": 1.590089008175723e-05, "loss": 0.7278, "step": 8839 }, { "epoch": 0.32017385005432814, "grad_norm": 1.514511343225887, "learning_rate": 1.589994297878426e-05, "loss": 0.8349, "step": 8840 }, { "epoch": 0.3202100688156465, "grad_norm": 1.3843908125510351, "learning_rate": 1.589899579462289e-05, "loss": 0.8023, "step": 8841 }, { "epoch": 0.3202462875769649, "grad_norm": 1.0933463691891563, "learning_rate": 1.5898048529286153e-05, "loss": 0.7871, "step": 8842 }, { "epoch": 0.32028250633828326, "grad_norm": 1.3277175811351936, "learning_rate": 1.5897101182787088e-05, "loss": 0.7733, "step": 8843 }, { "epoch": 0.3203187250996016, "grad_norm": 1.4095578694068551, "learning_rate": 1.589615375513873e-05, "loss": 0.8186, "step": 8844 }, { "epoch": 0.32035494386091995, "grad_norm": 1.3890302459900516, "learning_rate": 1.5895206246354115e-05, "loss": 0.8418, "step": 8845 }, { "epoch": 0.3203911626222383, "grad_norm": 1.4638194050284592, "learning_rate": 1.589425865644628e-05, "loss": 0.7683, "step": 8846 }, { "epoch": 0.3204273813835567, "grad_norm": 1.3616153520155803, "learning_rate": 1.5893310985428264e-05, "loss": 0.7372, "step": 8847 }, { "epoch": 0.32046360014487507, "grad_norm": 1.1257656197377823, "learning_rate": 1.5892363233313117e-05, "loss": 0.7629, "step": 8848 }, { "epoch": 0.3204998189061934, "grad_norm": 1.4727041365732807, "learning_rate": 1.589141540011387e-05, "loss": 0.767, "step": 8849 }, { "epoch": 0.32053603766751176, "grad_norm": 1.4571813890812502, "learning_rate": 1.5890467485843573e-05, "loss": 0.801, "step": 8850 }, { "epoch": 0.32057225642883014, "grad_norm": 1.234716049518752, "learning_rate": 1.5889519490515265e-05, "loss": 0.7417, "step": 8851 }, { "epoch": 0.3206084751901485, "grad_norm": 1.3745120575330192, "learning_rate": 1.5888571414141997e-05, "loss": 0.8369, "step": 8852 }, { "epoch": 0.3206446939514669, "grad_norm": 1.4436084405952774, "learning_rate": 1.588762325673681e-05, "loss": 0.8761, "step": 8853 }, { "epoch": 0.3206809127127852, "grad_norm": 1.5563124522543164, "learning_rate": 1.5886675018312754e-05, "loss": 0.7799, "step": 8854 }, { "epoch": 0.3207171314741036, "grad_norm": 1.3820456746665637, "learning_rate": 1.5885726698882877e-05, "loss": 0.8549, "step": 8855 }, { "epoch": 0.32075335023542195, "grad_norm": 1.3701358122317653, "learning_rate": 1.5884778298460232e-05, "loss": 0.8045, "step": 8856 }, { "epoch": 0.3207895689967403, "grad_norm": 1.3464124295419468, "learning_rate": 1.5883829817057862e-05, "loss": 0.8296, "step": 8857 }, { "epoch": 0.3208257877580587, "grad_norm": 1.2460419665013467, "learning_rate": 1.588288125468883e-05, "loss": 0.8353, "step": 8858 }, { "epoch": 0.320862006519377, "grad_norm": 1.4089338454304166, "learning_rate": 1.5881932611366178e-05, "loss": 0.8129, "step": 8859 }, { "epoch": 0.3208982252806954, "grad_norm": 1.6443629779035476, "learning_rate": 1.5880983887102965e-05, "loss": 0.8444, "step": 8860 }, { "epoch": 0.32093444404201377, "grad_norm": 1.4718095789177381, "learning_rate": 1.5880035081912246e-05, "loss": 0.868, "step": 8861 }, { "epoch": 0.32097066280333214, "grad_norm": 1.398970276023446, "learning_rate": 1.587908619580708e-05, "loss": 0.7498, "step": 8862 }, { "epoch": 0.3210068815646505, "grad_norm": 1.3803170986025526, "learning_rate": 1.587813722880052e-05, "loss": 0.7758, "step": 8863 }, { "epoch": 0.32104310032596883, "grad_norm": 1.4876954823258854, "learning_rate": 1.5877188180905626e-05, "loss": 0.7256, "step": 8864 }, { "epoch": 0.3210793190872872, "grad_norm": 1.4318870407827893, "learning_rate": 1.5876239052135463e-05, "loss": 0.868, "step": 8865 }, { "epoch": 0.3211155378486056, "grad_norm": 1.2168059411866872, "learning_rate": 1.5875289842503083e-05, "loss": 0.8131, "step": 8866 }, { "epoch": 0.32115175660992396, "grad_norm": 1.1059567697789976, "learning_rate": 1.5874340552021554e-05, "loss": 0.7771, "step": 8867 }, { "epoch": 0.32118797537124233, "grad_norm": 1.2125374811448266, "learning_rate": 1.5873391180703937e-05, "loss": 0.7461, "step": 8868 }, { "epoch": 0.32122419413256065, "grad_norm": 1.7106732813300787, "learning_rate": 1.5872441728563293e-05, "loss": 0.8058, "step": 8869 }, { "epoch": 0.321260412893879, "grad_norm": 1.474143288070325, "learning_rate": 1.5871492195612698e-05, "loss": 0.7886, "step": 8870 }, { "epoch": 0.3212966316551974, "grad_norm": 1.3968906625193889, "learning_rate": 1.5870542581865205e-05, "loss": 0.7142, "step": 8871 }, { "epoch": 0.32133285041651577, "grad_norm": 1.0470473459881173, "learning_rate": 1.586959288733389e-05, "loss": 0.7691, "step": 8872 }, { "epoch": 0.32136906917783414, "grad_norm": 1.3392294322643499, "learning_rate": 1.5868643112031824e-05, "loss": 0.839, "step": 8873 }, { "epoch": 0.32140528793915246, "grad_norm": 1.436865448596026, "learning_rate": 1.5867693255972064e-05, "loss": 0.7853, "step": 8874 }, { "epoch": 0.32144150670047084, "grad_norm": 1.1026257015478917, "learning_rate": 1.5866743319167696e-05, "loss": 0.7697, "step": 8875 }, { "epoch": 0.3214777254617892, "grad_norm": 1.3711835147132585, "learning_rate": 1.586579330163178e-05, "loss": 0.7452, "step": 8876 }, { "epoch": 0.3215139442231076, "grad_norm": 1.3709232857421396, "learning_rate": 1.5864843203377397e-05, "loss": 0.6834, "step": 8877 }, { "epoch": 0.32155016298442596, "grad_norm": 1.5682854070833545, "learning_rate": 1.5863893024417615e-05, "loss": 0.783, "step": 8878 }, { "epoch": 0.3215863817457443, "grad_norm": 1.3231315524313363, "learning_rate": 1.5862942764765518e-05, "loss": 0.7644, "step": 8879 }, { "epoch": 0.32162260050706265, "grad_norm": 1.5630771000374082, "learning_rate": 1.5861992424434173e-05, "loss": 0.7792, "step": 8880 }, { "epoch": 0.321658819268381, "grad_norm": 1.331080286027538, "learning_rate": 1.586104200343666e-05, "loss": 0.7507, "step": 8881 }, { "epoch": 0.3216950380296994, "grad_norm": 1.1422906889520168, "learning_rate": 1.586009150178606e-05, "loss": 0.786, "step": 8882 }, { "epoch": 0.3217312567910178, "grad_norm": 1.4357321872604218, "learning_rate": 1.5859140919495452e-05, "loss": 0.7538, "step": 8883 }, { "epoch": 0.3217674755523361, "grad_norm": 1.5802710501417419, "learning_rate": 1.5858190256577917e-05, "loss": 0.802, "step": 8884 }, { "epoch": 0.32180369431365446, "grad_norm": 1.4830804638479396, "learning_rate": 1.585723951304654e-05, "loss": 0.8565, "step": 8885 }, { "epoch": 0.32183991307497284, "grad_norm": 1.093348876593349, "learning_rate": 1.5856288688914396e-05, "loss": 0.705, "step": 8886 }, { "epoch": 0.3218761318362912, "grad_norm": 1.3969223789419714, "learning_rate": 1.5855337784194576e-05, "loss": 0.9036, "step": 8887 }, { "epoch": 0.3219123505976096, "grad_norm": 1.438246375142348, "learning_rate": 1.5854386798900165e-05, "loss": 0.8378, "step": 8888 }, { "epoch": 0.3219485693589279, "grad_norm": 1.6908194496431954, "learning_rate": 1.5853435733044246e-05, "loss": 0.8629, "step": 8889 }, { "epoch": 0.3219847881202463, "grad_norm": 1.0850511009347965, "learning_rate": 1.5852484586639905e-05, "loss": 0.7534, "step": 8890 }, { "epoch": 0.32202100688156465, "grad_norm": 1.5532972934877862, "learning_rate": 1.5851533359700238e-05, "loss": 0.8007, "step": 8891 }, { "epoch": 0.322057225642883, "grad_norm": 1.4237664854315202, "learning_rate": 1.585058205223833e-05, "loss": 0.8322, "step": 8892 }, { "epoch": 0.3220934444042014, "grad_norm": 1.5664361965142208, "learning_rate": 1.5849630664267266e-05, "loss": 0.8073, "step": 8893 }, { "epoch": 0.3221296631655197, "grad_norm": 1.6091674907012354, "learning_rate": 1.584867919580015e-05, "loss": 0.7731, "step": 8894 }, { "epoch": 0.3221658819268381, "grad_norm": 1.3772840744707386, "learning_rate": 1.584772764685007e-05, "loss": 0.8121, "step": 8895 }, { "epoch": 0.32220210068815647, "grad_norm": 1.5181273641092825, "learning_rate": 1.5846776017430115e-05, "loss": 0.7835, "step": 8896 }, { "epoch": 0.32223831944947484, "grad_norm": 1.5496840333124564, "learning_rate": 1.584582430755339e-05, "loss": 0.8306, "step": 8897 }, { "epoch": 0.3222745382107932, "grad_norm": 1.4922384557961281, "learning_rate": 1.584487251723298e-05, "loss": 0.7926, "step": 8898 }, { "epoch": 0.32231075697211153, "grad_norm": 1.4834236785770345, "learning_rate": 1.584392064648199e-05, "loss": 0.7904, "step": 8899 }, { "epoch": 0.3223469757334299, "grad_norm": 1.4284743408410752, "learning_rate": 1.584296869531352e-05, "loss": 0.7837, "step": 8900 }, { "epoch": 0.3223831944947483, "grad_norm": 1.3915362754364289, "learning_rate": 1.5842016663740664e-05, "loss": 0.8605, "step": 8901 }, { "epoch": 0.32241941325606666, "grad_norm": 1.470559856002033, "learning_rate": 1.5841064551776528e-05, "loss": 0.852, "step": 8902 }, { "epoch": 0.32245563201738503, "grad_norm": 1.3166314881798333, "learning_rate": 1.5840112359434207e-05, "loss": 0.7418, "step": 8903 }, { "epoch": 0.32249185077870335, "grad_norm": 1.0257097644530728, "learning_rate": 1.5839160086726812e-05, "loss": 0.7638, "step": 8904 }, { "epoch": 0.3225280695400217, "grad_norm": 1.2143968874958493, "learning_rate": 1.5838207733667443e-05, "loss": 0.8182, "step": 8905 }, { "epoch": 0.3225642883013401, "grad_norm": 1.3068995803628978, "learning_rate": 1.5837255300269204e-05, "loss": 0.7865, "step": 8906 }, { "epoch": 0.32260050706265847, "grad_norm": 1.1531176150739357, "learning_rate": 1.5836302786545205e-05, "loss": 0.7921, "step": 8907 }, { "epoch": 0.32263672582397684, "grad_norm": 1.7683156682421817, "learning_rate": 1.5835350192508553e-05, "loss": 0.8623, "step": 8908 }, { "epoch": 0.32267294458529516, "grad_norm": 1.2096686224675062, "learning_rate": 1.5834397518172352e-05, "loss": 0.8483, "step": 8909 }, { "epoch": 0.32270916334661354, "grad_norm": 1.6038222234709263, "learning_rate": 1.5833444763549716e-05, "loss": 0.877, "step": 8910 }, { "epoch": 0.3227453821079319, "grad_norm": 1.3250437265392023, "learning_rate": 1.583249192865375e-05, "loss": 0.7578, "step": 8911 }, { "epoch": 0.3227816008692503, "grad_norm": 1.371958412413072, "learning_rate": 1.5831539013497574e-05, "loss": 0.7551, "step": 8912 }, { "epoch": 0.32281781963056866, "grad_norm": 1.13759940033716, "learning_rate": 1.58305860180943e-05, "loss": 0.7588, "step": 8913 }, { "epoch": 0.322854038391887, "grad_norm": 1.4244874845595492, "learning_rate": 1.5829632942457032e-05, "loss": 0.8297, "step": 8914 }, { "epoch": 0.32289025715320535, "grad_norm": 1.1889115744119312, "learning_rate": 1.5828679786598897e-05, "loss": 0.8057, "step": 8915 }, { "epoch": 0.3229264759145237, "grad_norm": 1.3464613731307564, "learning_rate": 1.5827726550533007e-05, "loss": 0.894, "step": 8916 }, { "epoch": 0.3229626946758421, "grad_norm": 1.0837639132330037, "learning_rate": 1.5826773234272476e-05, "loss": 0.7338, "step": 8917 }, { "epoch": 0.3229989134371605, "grad_norm": 1.289555767357347, "learning_rate": 1.582581983783043e-05, "loss": 0.8702, "step": 8918 }, { "epoch": 0.3230351321984788, "grad_norm": 1.355413455000183, "learning_rate": 1.582486636121998e-05, "loss": 0.8245, "step": 8919 }, { "epoch": 0.32307135095979717, "grad_norm": 1.3665283861017654, "learning_rate": 1.582391280445425e-05, "loss": 0.7458, "step": 8920 }, { "epoch": 0.32310756972111554, "grad_norm": 1.5305172160664353, "learning_rate": 1.5822959167546365e-05, "loss": 0.7909, "step": 8921 }, { "epoch": 0.3231437884824339, "grad_norm": 1.4090360633056198, "learning_rate": 1.5822005450509447e-05, "loss": 0.7527, "step": 8922 }, { "epoch": 0.3231800072437523, "grad_norm": 1.4478654308689232, "learning_rate": 1.5821051653356617e-05, "loss": 0.7792, "step": 8923 }, { "epoch": 0.3232162260050706, "grad_norm": 1.1503050069266485, "learning_rate": 1.5820097776101e-05, "loss": 0.8202, "step": 8924 }, { "epoch": 0.323252444766389, "grad_norm": 1.344222494348593, "learning_rate": 1.5819143818755725e-05, "loss": 0.7663, "step": 8925 }, { "epoch": 0.32328866352770735, "grad_norm": 1.1058369046564829, "learning_rate": 1.5818189781333914e-05, "loss": 0.7865, "step": 8926 }, { "epoch": 0.32332488228902573, "grad_norm": 1.3943058661239753, "learning_rate": 1.5817235663848704e-05, "loss": 0.78, "step": 8927 }, { "epoch": 0.3233611010503441, "grad_norm": 1.4443766934445985, "learning_rate": 1.5816281466313218e-05, "loss": 0.7971, "step": 8928 }, { "epoch": 0.3233973198116624, "grad_norm": 1.4941721482784933, "learning_rate": 1.581532718874059e-05, "loss": 0.8485, "step": 8929 }, { "epoch": 0.3234335385729808, "grad_norm": 1.3932404247316312, "learning_rate": 1.5814372831143946e-05, "loss": 0.7653, "step": 8930 }, { "epoch": 0.32346975733429917, "grad_norm": 1.3997835126163873, "learning_rate": 1.581341839353643e-05, "loss": 0.8328, "step": 8931 }, { "epoch": 0.32350597609561754, "grad_norm": 1.1613676823792711, "learning_rate": 1.581246387593116e-05, "loss": 0.7571, "step": 8932 }, { "epoch": 0.3235421948569359, "grad_norm": 1.2488851135109003, "learning_rate": 1.5811509278341285e-05, "loss": 0.7114, "step": 8933 }, { "epoch": 0.32357841361825423, "grad_norm": 1.6118257436218955, "learning_rate": 1.5810554600779936e-05, "loss": 0.8444, "step": 8934 }, { "epoch": 0.3236146323795726, "grad_norm": 1.3767537053761703, "learning_rate": 1.5809599843260248e-05, "loss": 0.8346, "step": 8935 }, { "epoch": 0.323650851140891, "grad_norm": 1.372989617397846, "learning_rate": 1.580864500579536e-05, "loss": 0.7045, "step": 8936 }, { "epoch": 0.32368706990220936, "grad_norm": 1.3515715746905685, "learning_rate": 1.580769008839842e-05, "loss": 0.8021, "step": 8937 }, { "epoch": 0.32372328866352773, "grad_norm": 1.2538717924713432, "learning_rate": 1.5806735091082555e-05, "loss": 0.7374, "step": 8938 }, { "epoch": 0.32375950742484605, "grad_norm": 1.273014931689153, "learning_rate": 1.5805780013860914e-05, "loss": 0.7843, "step": 8939 }, { "epoch": 0.3237957261861644, "grad_norm": 1.3409258577963323, "learning_rate": 1.580482485674664e-05, "loss": 0.8025, "step": 8940 }, { "epoch": 0.3238319449474828, "grad_norm": 1.1989540444189004, "learning_rate": 1.5803869619752874e-05, "loss": 0.8059, "step": 8941 }, { "epoch": 0.32386816370880117, "grad_norm": 1.4468416540225073, "learning_rate": 1.5802914302892764e-05, "loss": 0.8753, "step": 8942 }, { "epoch": 0.32390438247011955, "grad_norm": 1.4690182534727703, "learning_rate": 1.580195890617945e-05, "loss": 0.767, "step": 8943 }, { "epoch": 0.32394060123143786, "grad_norm": 1.0696943690202012, "learning_rate": 1.580100342962609e-05, "loss": 0.795, "step": 8944 }, { "epoch": 0.32397681999275624, "grad_norm": 1.2951415488388653, "learning_rate": 1.580004787324582e-05, "loss": 0.7769, "step": 8945 }, { "epoch": 0.3240130387540746, "grad_norm": 1.4429966589547558, "learning_rate": 1.57990922370518e-05, "loss": 0.8821, "step": 8946 }, { "epoch": 0.324049257515393, "grad_norm": 1.4722252387046149, "learning_rate": 1.579813652105717e-05, "loss": 0.8663, "step": 8947 }, { "epoch": 0.32408547627671136, "grad_norm": 1.5388638239878365, "learning_rate": 1.5797180725275092e-05, "loss": 0.8192, "step": 8948 }, { "epoch": 0.3241216950380297, "grad_norm": 1.4577634418815109, "learning_rate": 1.5796224849718708e-05, "loss": 0.7839, "step": 8949 }, { "epoch": 0.32415791379934805, "grad_norm": 1.5728697688224837, "learning_rate": 1.579526889440118e-05, "loss": 0.8127, "step": 8950 }, { "epoch": 0.3241941325606664, "grad_norm": 1.4045009497914391, "learning_rate": 1.579431285933566e-05, "loss": 0.8381, "step": 8951 }, { "epoch": 0.3242303513219848, "grad_norm": 1.4371424786269258, "learning_rate": 1.5793356744535303e-05, "loss": 0.6681, "step": 8952 }, { "epoch": 0.3242665700833032, "grad_norm": 1.3857915021362506, "learning_rate": 1.5792400550013267e-05, "loss": 0.8452, "step": 8953 }, { "epoch": 0.3243027888446215, "grad_norm": 1.3685117772538193, "learning_rate": 1.579144427578271e-05, "loss": 0.8095, "step": 8954 }, { "epoch": 0.32433900760593987, "grad_norm": 1.1808243236322504, "learning_rate": 1.579048792185679e-05, "loss": 0.8591, "step": 8955 }, { "epoch": 0.32437522636725824, "grad_norm": 1.389300786762429, "learning_rate": 1.578953148824867e-05, "loss": 0.7938, "step": 8956 }, { "epoch": 0.3244114451285766, "grad_norm": 1.6828832587558213, "learning_rate": 1.5788574974971506e-05, "loss": 0.8492, "step": 8957 }, { "epoch": 0.324447663889895, "grad_norm": 1.3535256736349335, "learning_rate": 1.5787618382038467e-05, "loss": 0.7508, "step": 8958 }, { "epoch": 0.3244838826512133, "grad_norm": 1.3711817657444783, "learning_rate": 1.578666170946271e-05, "loss": 0.76, "step": 8959 }, { "epoch": 0.3245201014125317, "grad_norm": 1.2956825887244625, "learning_rate": 1.578570495725741e-05, "loss": 0.7565, "step": 8960 }, { "epoch": 0.32455632017385005, "grad_norm": 1.6079585991636243, "learning_rate": 1.578474812543572e-05, "loss": 0.7844, "step": 8961 }, { "epoch": 0.32459253893516843, "grad_norm": 1.3076940860030921, "learning_rate": 1.5783791214010818e-05, "loss": 0.7405, "step": 8962 }, { "epoch": 0.3246287576964868, "grad_norm": 1.444290148800426, "learning_rate": 1.5782834222995864e-05, "loss": 0.9016, "step": 8963 }, { "epoch": 0.3246649764578051, "grad_norm": 1.4277851047452979, "learning_rate": 1.578187715240403e-05, "loss": 0.913, "step": 8964 }, { "epoch": 0.3247011952191235, "grad_norm": 1.4126533473588059, "learning_rate": 1.5780920002248484e-05, "loss": 0.7763, "step": 8965 }, { "epoch": 0.32473741398044187, "grad_norm": 1.334922714634987, "learning_rate": 1.5779962772542404e-05, "loss": 0.7976, "step": 8966 }, { "epoch": 0.32477363274176024, "grad_norm": 1.4955961642472502, "learning_rate": 1.5779005463298954e-05, "loss": 0.891, "step": 8967 }, { "epoch": 0.3248098515030786, "grad_norm": 1.5299775862184226, "learning_rate": 1.577804807453131e-05, "loss": 0.8259, "step": 8968 }, { "epoch": 0.32484607026439694, "grad_norm": 1.5508276730658659, "learning_rate": 1.577709060625265e-05, "loss": 0.8241, "step": 8969 }, { "epoch": 0.3248822890257153, "grad_norm": 1.4714179967577645, "learning_rate": 1.5776133058476143e-05, "loss": 0.8618, "step": 8970 }, { "epoch": 0.3249185077870337, "grad_norm": 1.4195404583333655, "learning_rate": 1.5775175431214976e-05, "loss": 0.7697, "step": 8971 }, { "epoch": 0.32495472654835206, "grad_norm": 1.4386702819512958, "learning_rate": 1.577421772448232e-05, "loss": 0.781, "step": 8972 }, { "epoch": 0.32499094530967043, "grad_norm": 1.4511780642432668, "learning_rate": 1.5773259938291347e-05, "loss": 0.8022, "step": 8973 }, { "epoch": 0.32502716407098875, "grad_norm": 1.4567789736329253, "learning_rate": 1.577230207265525e-05, "loss": 0.8794, "step": 8974 }, { "epoch": 0.3250633828323071, "grad_norm": 1.208019345245033, "learning_rate": 1.5771344127587204e-05, "loss": 0.7619, "step": 8975 }, { "epoch": 0.3250996015936255, "grad_norm": 1.3121908680990026, "learning_rate": 1.5770386103100392e-05, "loss": 0.7244, "step": 8976 }, { "epoch": 0.32513582035494387, "grad_norm": 1.2683607280595928, "learning_rate": 1.5769427999208e-05, "loss": 0.708, "step": 8977 }, { "epoch": 0.32517203911626225, "grad_norm": 1.4312976085353257, "learning_rate": 1.57684698159232e-05, "loss": 0.8106, "step": 8978 }, { "epoch": 0.32520825787758056, "grad_norm": 1.3744712776210883, "learning_rate": 1.5767511553259195e-05, "loss": 0.7224, "step": 8979 }, { "epoch": 0.32524447663889894, "grad_norm": 1.3826934020969535, "learning_rate": 1.576655321122916e-05, "loss": 0.7417, "step": 8980 }, { "epoch": 0.3252806954002173, "grad_norm": 1.3798061546635603, "learning_rate": 1.5765594789846284e-05, "loss": 0.7778, "step": 8981 }, { "epoch": 0.3253169141615357, "grad_norm": 1.5148362252991379, "learning_rate": 1.5764636289123763e-05, "loss": 0.7782, "step": 8982 }, { "epoch": 0.32535313292285406, "grad_norm": 1.140339682502654, "learning_rate": 1.5763677709074777e-05, "loss": 0.7761, "step": 8983 }, { "epoch": 0.3253893516841724, "grad_norm": 1.6411132426603812, "learning_rate": 1.5762719049712526e-05, "loss": 0.8357, "step": 8984 }, { "epoch": 0.32542557044549075, "grad_norm": 1.4254772953133568, "learning_rate": 1.5761760311050193e-05, "loss": 0.8057, "step": 8985 }, { "epoch": 0.3254617892068091, "grad_norm": 1.2788900441918074, "learning_rate": 1.5760801493100975e-05, "loss": 0.7654, "step": 8986 }, { "epoch": 0.3254980079681275, "grad_norm": 1.2704353612404728, "learning_rate": 1.575984259587807e-05, "loss": 0.7084, "step": 8987 }, { "epoch": 0.3255342267294459, "grad_norm": 1.4971106238266154, "learning_rate": 1.5758883619394665e-05, "loss": 0.7591, "step": 8988 }, { "epoch": 0.3255704454907642, "grad_norm": 1.4522120032799284, "learning_rate": 1.5757924563663966e-05, "loss": 0.7906, "step": 8989 }, { "epoch": 0.32560666425208257, "grad_norm": 1.155406460254183, "learning_rate": 1.5756965428699163e-05, "loss": 0.7163, "step": 8990 }, { "epoch": 0.32564288301340094, "grad_norm": 1.4573533389934603, "learning_rate": 1.5756006214513458e-05, "loss": 0.8671, "step": 8991 }, { "epoch": 0.3256791017747193, "grad_norm": 1.280436067559754, "learning_rate": 1.5755046921120046e-05, "loss": 0.7769, "step": 8992 }, { "epoch": 0.3257153205360377, "grad_norm": 1.4717092327551748, "learning_rate": 1.5754087548532136e-05, "loss": 0.7791, "step": 8993 }, { "epoch": 0.325751539297356, "grad_norm": 1.5301005436810908, "learning_rate": 1.5753128096762926e-05, "loss": 0.8066, "step": 8994 }, { "epoch": 0.3257877580586744, "grad_norm": 1.3705195232498266, "learning_rate": 1.5752168565825616e-05, "loss": 0.8079, "step": 8995 }, { "epoch": 0.32582397681999276, "grad_norm": 1.466996541256133, "learning_rate": 1.5751208955733415e-05, "loss": 0.7912, "step": 8996 }, { "epoch": 0.32586019558131113, "grad_norm": 1.526790061210915, "learning_rate": 1.5750249266499524e-05, "loss": 0.8392, "step": 8997 }, { "epoch": 0.3258964143426295, "grad_norm": 1.4439071116366446, "learning_rate": 1.5749289498137147e-05, "loss": 0.7266, "step": 8998 }, { "epoch": 0.3259326331039478, "grad_norm": 1.2637866099301411, "learning_rate": 1.57483296506595e-05, "loss": 0.8119, "step": 8999 }, { "epoch": 0.3259688518652662, "grad_norm": 1.4156028626092085, "learning_rate": 1.5747369724079785e-05, "loss": 0.8341, "step": 9000 }, { "epoch": 0.32600507062658457, "grad_norm": 1.324712408092145, "learning_rate": 1.574640971841121e-05, "loss": 0.7308, "step": 9001 }, { "epoch": 0.32604128938790294, "grad_norm": 1.40741299116352, "learning_rate": 1.574544963366699e-05, "loss": 0.8066, "step": 9002 }, { "epoch": 0.3260775081492213, "grad_norm": 1.0833586586950668, "learning_rate": 1.5744489469860334e-05, "loss": 0.7735, "step": 9003 }, { "epoch": 0.32611372691053964, "grad_norm": 1.2303372452869328, "learning_rate": 1.5743529227004456e-05, "loss": 0.737, "step": 9004 }, { "epoch": 0.326149945671858, "grad_norm": 1.5290767272499386, "learning_rate": 1.574256890511257e-05, "loss": 0.7603, "step": 9005 }, { "epoch": 0.3261861644331764, "grad_norm": 1.3538483572643478, "learning_rate": 1.5741608504197887e-05, "loss": 0.7442, "step": 9006 }, { "epoch": 0.32622238319449476, "grad_norm": 1.4457163826355945, "learning_rate": 1.5740648024273627e-05, "loss": 0.7068, "step": 9007 }, { "epoch": 0.32625860195581313, "grad_norm": 1.1127191975668034, "learning_rate": 1.5739687465353005e-05, "loss": 0.743, "step": 9008 }, { "epoch": 0.32629482071713145, "grad_norm": 1.5925760953823365, "learning_rate": 1.573872682744924e-05, "loss": 0.8298, "step": 9009 }, { "epoch": 0.3263310394784498, "grad_norm": 1.428882475155137, "learning_rate": 1.5737766110575554e-05, "loss": 0.747, "step": 9010 }, { "epoch": 0.3263672582397682, "grad_norm": 1.1388618355157358, "learning_rate": 1.5736805314745166e-05, "loss": 0.8008, "step": 9011 }, { "epoch": 0.3264034770010866, "grad_norm": 1.2628074737794306, "learning_rate": 1.5735844439971295e-05, "loss": 0.7846, "step": 9012 }, { "epoch": 0.32643969576240495, "grad_norm": 1.38828558585714, "learning_rate": 1.5734883486267162e-05, "loss": 0.7894, "step": 9013 }, { "epoch": 0.32647591452372327, "grad_norm": 1.535334033232722, "learning_rate": 1.5733922453645993e-05, "loss": 0.7988, "step": 9014 }, { "epoch": 0.32651213328504164, "grad_norm": 1.1833235784232836, "learning_rate": 1.5732961342121015e-05, "loss": 0.811, "step": 9015 }, { "epoch": 0.32654835204636, "grad_norm": 1.5537366822973218, "learning_rate": 1.573200015170545e-05, "loss": 0.7966, "step": 9016 }, { "epoch": 0.3265845708076784, "grad_norm": 1.521488379447476, "learning_rate": 1.5731038882412527e-05, "loss": 0.8096, "step": 9017 }, { "epoch": 0.32662078956899676, "grad_norm": 1.3299989193595085, "learning_rate": 1.573007753425547e-05, "loss": 0.7663, "step": 9018 }, { "epoch": 0.3266570083303151, "grad_norm": 1.0380051790926177, "learning_rate": 1.5729116107247516e-05, "loss": 0.7068, "step": 9019 }, { "epoch": 0.32669322709163345, "grad_norm": 1.3917525487270952, "learning_rate": 1.572815460140189e-05, "loss": 0.7274, "step": 9020 }, { "epoch": 0.3267294458529518, "grad_norm": 1.3356449869384792, "learning_rate": 1.5727193016731823e-05, "loss": 0.7563, "step": 9021 }, { "epoch": 0.3267656646142702, "grad_norm": 1.4036976600850095, "learning_rate": 1.5726231353250544e-05, "loss": 0.7979, "step": 9022 }, { "epoch": 0.3268018833755886, "grad_norm": 1.3653960316443967, "learning_rate": 1.5725269610971294e-05, "loss": 0.8729, "step": 9023 }, { "epoch": 0.3268381021369069, "grad_norm": 1.3602292114684762, "learning_rate": 1.5724307789907304e-05, "loss": 0.7161, "step": 9024 }, { "epoch": 0.32687432089822527, "grad_norm": 1.4021184469316277, "learning_rate": 1.5723345890071805e-05, "loss": 0.807, "step": 9025 }, { "epoch": 0.32691053965954364, "grad_norm": 1.147399640576633, "learning_rate": 1.572238391147804e-05, "loss": 0.7796, "step": 9026 }, { "epoch": 0.326946758420862, "grad_norm": 1.093399481507978, "learning_rate": 1.5721421854139248e-05, "loss": 0.7404, "step": 9027 }, { "epoch": 0.3269829771821804, "grad_norm": 1.5407535272070765, "learning_rate": 1.5720459718068658e-05, "loss": 0.8549, "step": 9028 }, { "epoch": 0.3270191959434987, "grad_norm": 1.621470769449717, "learning_rate": 1.571949750327952e-05, "loss": 0.7945, "step": 9029 }, { "epoch": 0.3270554147048171, "grad_norm": 1.4201717589722747, "learning_rate": 1.5718535209785068e-05, "loss": 0.7641, "step": 9030 }, { "epoch": 0.32709163346613546, "grad_norm": 1.393842590548399, "learning_rate": 1.5717572837598545e-05, "loss": 0.8236, "step": 9031 }, { "epoch": 0.32712785222745383, "grad_norm": 1.1881278375369455, "learning_rate": 1.57166103867332e-05, "loss": 0.7598, "step": 9032 }, { "epoch": 0.3271640709887722, "grad_norm": 1.4576258136586537, "learning_rate": 1.5715647857202273e-05, "loss": 0.8423, "step": 9033 }, { "epoch": 0.3272002897500905, "grad_norm": 1.4775226652618452, "learning_rate": 1.5714685249019006e-05, "loss": 0.7879, "step": 9034 }, { "epoch": 0.3272365085114089, "grad_norm": 1.5791805255848133, "learning_rate": 1.571372256219665e-05, "loss": 0.8328, "step": 9035 }, { "epoch": 0.32727272727272727, "grad_norm": 1.6386895730082507, "learning_rate": 1.5712759796748454e-05, "loss": 0.7377, "step": 9036 }, { "epoch": 0.32730894603404564, "grad_norm": 1.3552197913440018, "learning_rate": 1.5711796952687663e-05, "loss": 0.8046, "step": 9037 }, { "epoch": 0.327345164795364, "grad_norm": 1.3162978641308456, "learning_rate": 1.5710834030027523e-05, "loss": 0.7607, "step": 9038 }, { "epoch": 0.32738138355668234, "grad_norm": 1.1936051865533037, "learning_rate": 1.570987102878129e-05, "loss": 0.7128, "step": 9039 }, { "epoch": 0.3274176023180007, "grad_norm": 1.4163964062110384, "learning_rate": 1.5708907948962216e-05, "loss": 0.8303, "step": 9040 }, { "epoch": 0.3274538210793191, "grad_norm": 1.3860962303228872, "learning_rate": 1.570794479058355e-05, "loss": 0.8379, "step": 9041 }, { "epoch": 0.32749003984063746, "grad_norm": 1.1142035122335854, "learning_rate": 1.570698155365855e-05, "loss": 0.7938, "step": 9042 }, { "epoch": 0.32752625860195583, "grad_norm": 1.58961065678426, "learning_rate": 1.570601823820047e-05, "loss": 0.7673, "step": 9043 }, { "epoch": 0.32756247736327415, "grad_norm": 1.4789783668265053, "learning_rate": 1.5705054844222564e-05, "loss": 0.8026, "step": 9044 }, { "epoch": 0.3275986961245925, "grad_norm": 1.3345700551907451, "learning_rate": 1.570409137173809e-05, "loss": 0.7575, "step": 9045 }, { "epoch": 0.3276349148859109, "grad_norm": 1.3909347449069485, "learning_rate": 1.5703127820760312e-05, "loss": 0.8019, "step": 9046 }, { "epoch": 0.3276711336472293, "grad_norm": 1.4447102554925417, "learning_rate": 1.5702164191302478e-05, "loss": 0.8014, "step": 9047 }, { "epoch": 0.32770735240854765, "grad_norm": 1.4121792480519306, "learning_rate": 1.5701200483377855e-05, "loss": 0.8346, "step": 9048 }, { "epoch": 0.32774357116986597, "grad_norm": 1.3720877948339572, "learning_rate": 1.5700236696999703e-05, "loss": 0.7637, "step": 9049 }, { "epoch": 0.32777978993118434, "grad_norm": 1.433199706220357, "learning_rate": 1.569927283218129e-05, "loss": 0.7894, "step": 9050 }, { "epoch": 0.3278160086925027, "grad_norm": 1.4695564748240768, "learning_rate": 1.569830888893587e-05, "loss": 0.7817, "step": 9051 }, { "epoch": 0.3278522274538211, "grad_norm": 1.4361947180786439, "learning_rate": 1.5697344867276716e-05, "loss": 0.8742, "step": 9052 }, { "epoch": 0.32788844621513946, "grad_norm": 1.4125877186187974, "learning_rate": 1.569638076721709e-05, "loss": 0.7587, "step": 9053 }, { "epoch": 0.3279246649764578, "grad_norm": 1.4113465737922164, "learning_rate": 1.569541658877026e-05, "loss": 0.8433, "step": 9054 }, { "epoch": 0.32796088373777615, "grad_norm": 1.4299939809414928, "learning_rate": 1.569445233194949e-05, "loss": 0.8591, "step": 9055 }, { "epoch": 0.32799710249909453, "grad_norm": 1.2782311943480016, "learning_rate": 1.5693487996768056e-05, "loss": 0.8361, "step": 9056 }, { "epoch": 0.3280333212604129, "grad_norm": 1.7634486217389314, "learning_rate": 1.5692523583239222e-05, "loss": 0.7127, "step": 9057 }, { "epoch": 0.3280695400217313, "grad_norm": 1.4126384207061677, "learning_rate": 1.569155909137626e-05, "loss": 0.8147, "step": 9058 }, { "epoch": 0.3281057587830496, "grad_norm": 1.475937253028014, "learning_rate": 1.569059452119245e-05, "loss": 0.8856, "step": 9059 }, { "epoch": 0.32814197754436797, "grad_norm": 1.4698879824990485, "learning_rate": 1.5689629872701053e-05, "loss": 0.8521, "step": 9060 }, { "epoch": 0.32817819630568634, "grad_norm": 1.2738384390147017, "learning_rate": 1.5688665145915352e-05, "loss": 0.8013, "step": 9061 }, { "epoch": 0.3282144150670047, "grad_norm": 1.2894149262387455, "learning_rate": 1.5687700340848618e-05, "loss": 0.7927, "step": 9062 }, { "epoch": 0.3282506338283231, "grad_norm": 1.4043813064617794, "learning_rate": 1.5686735457514136e-05, "loss": 0.8004, "step": 9063 }, { "epoch": 0.3282868525896414, "grad_norm": 1.3431893637391605, "learning_rate": 1.5685770495925173e-05, "loss": 0.7981, "step": 9064 }, { "epoch": 0.3283230713509598, "grad_norm": 1.404028783364097, "learning_rate": 1.568480545609501e-05, "loss": 0.8199, "step": 9065 }, { "epoch": 0.32835929011227816, "grad_norm": 1.241434803049771, "learning_rate": 1.5683840338036934e-05, "loss": 0.8143, "step": 9066 }, { "epoch": 0.32839550887359653, "grad_norm": 1.4078766302792631, "learning_rate": 1.5682875141764218e-05, "loss": 0.8123, "step": 9067 }, { "epoch": 0.3284317276349149, "grad_norm": 1.3560334233217417, "learning_rate": 1.5681909867290144e-05, "loss": 0.7291, "step": 9068 }, { "epoch": 0.3284679463962332, "grad_norm": 1.2205092345750008, "learning_rate": 1.5680944514628e-05, "loss": 0.8196, "step": 9069 }, { "epoch": 0.3285041651575516, "grad_norm": 1.0773884670309661, "learning_rate": 1.567997908379107e-05, "loss": 0.739, "step": 9070 }, { "epoch": 0.32854038391886997, "grad_norm": 1.3074001129582815, "learning_rate": 1.5679013574792634e-05, "loss": 0.763, "step": 9071 }, { "epoch": 0.32857660268018835, "grad_norm": 1.3928409725607938, "learning_rate": 1.567804798764598e-05, "loss": 0.8641, "step": 9072 }, { "epoch": 0.3286128214415067, "grad_norm": 1.3458201955412639, "learning_rate": 1.56770823223644e-05, "loss": 0.7969, "step": 9073 }, { "epoch": 0.32864904020282504, "grad_norm": 1.3703617386143645, "learning_rate": 1.5676116578961176e-05, "loss": 0.7968, "step": 9074 }, { "epoch": 0.3286852589641434, "grad_norm": 1.4599827921070478, "learning_rate": 1.5675150757449603e-05, "loss": 0.8373, "step": 9075 }, { "epoch": 0.3287214777254618, "grad_norm": 1.2805893011525757, "learning_rate": 1.5674184857842967e-05, "loss": 0.7382, "step": 9076 }, { "epoch": 0.32875769648678016, "grad_norm": 1.2405953219969363, "learning_rate": 1.567321888015456e-05, "loss": 0.803, "step": 9077 }, { "epoch": 0.32879391524809853, "grad_norm": 1.3128712078089837, "learning_rate": 1.5672252824397683e-05, "loss": 0.7716, "step": 9078 }, { "epoch": 0.32883013400941685, "grad_norm": 1.3351498367619827, "learning_rate": 1.5671286690585615e-05, "loss": 0.845, "step": 9079 }, { "epoch": 0.3288663527707352, "grad_norm": 1.3600397464621081, "learning_rate": 1.567032047873166e-05, "loss": 0.7813, "step": 9080 }, { "epoch": 0.3289025715320536, "grad_norm": 1.4808254969612595, "learning_rate": 1.5669354188849116e-05, "loss": 0.6772, "step": 9081 }, { "epoch": 0.328938790293372, "grad_norm": 1.2232142804487685, "learning_rate": 1.566838782095128e-05, "loss": 0.7699, "step": 9082 }, { "epoch": 0.32897500905469035, "grad_norm": 1.4372438469487332, "learning_rate": 1.5667421375051443e-05, "loss": 0.8394, "step": 9083 }, { "epoch": 0.32901122781600867, "grad_norm": 1.4147394934683477, "learning_rate": 1.5666454851162908e-05, "loss": 0.8339, "step": 9084 }, { "epoch": 0.32904744657732704, "grad_norm": 1.2959548633195057, "learning_rate": 1.5665488249298976e-05, "loss": 0.7416, "step": 9085 }, { "epoch": 0.3290836653386454, "grad_norm": 1.211230710317226, "learning_rate": 1.5664521569472945e-05, "loss": 0.7432, "step": 9086 }, { "epoch": 0.3291198840999638, "grad_norm": 1.645859270454802, "learning_rate": 1.5663554811698122e-05, "loss": 0.7972, "step": 9087 }, { "epoch": 0.32915610286128216, "grad_norm": 1.3709625197641155, "learning_rate": 1.5662587975987807e-05, "loss": 0.7893, "step": 9088 }, { "epoch": 0.3291923216226005, "grad_norm": 1.3949268538600632, "learning_rate": 1.566162106235531e-05, "loss": 0.8026, "step": 9089 }, { "epoch": 0.32922854038391886, "grad_norm": 1.2758536119139943, "learning_rate": 1.5660654070813928e-05, "loss": 0.7478, "step": 9090 }, { "epoch": 0.32926475914523723, "grad_norm": 1.341145025333297, "learning_rate": 1.5659687001376974e-05, "loss": 0.7287, "step": 9091 }, { "epoch": 0.3293009779065556, "grad_norm": 1.4843720963098332, "learning_rate": 1.5658719854057752e-05, "loss": 0.8257, "step": 9092 }, { "epoch": 0.329337196667874, "grad_norm": 1.157060854451447, "learning_rate": 1.5657752628869576e-05, "loss": 0.7273, "step": 9093 }, { "epoch": 0.3293734154291923, "grad_norm": 1.2098748804382062, "learning_rate": 1.565678532582575e-05, "loss": 0.7133, "step": 9094 }, { "epoch": 0.32940963419051067, "grad_norm": 1.3723728565871998, "learning_rate": 1.5655817944939588e-05, "loss": 0.8082, "step": 9095 }, { "epoch": 0.32944585295182904, "grad_norm": 1.8379222267182518, "learning_rate": 1.56548504862244e-05, "loss": 0.7964, "step": 9096 }, { "epoch": 0.3294820717131474, "grad_norm": 1.3760233479242319, "learning_rate": 1.5653882949693503e-05, "loss": 0.7842, "step": 9097 }, { "epoch": 0.3295182904744658, "grad_norm": 1.3511576790462287, "learning_rate": 1.5652915335360207e-05, "loss": 0.8003, "step": 9098 }, { "epoch": 0.3295545092357841, "grad_norm": 1.457584828203254, "learning_rate": 1.565194764323783e-05, "loss": 0.8132, "step": 9099 }, { "epoch": 0.3295907279971025, "grad_norm": 1.414690765944196, "learning_rate": 1.5650979873339686e-05, "loss": 0.7759, "step": 9100 }, { "epoch": 0.32962694675842086, "grad_norm": 1.3667589888700777, "learning_rate": 1.5650012025679095e-05, "loss": 0.7898, "step": 9101 }, { "epoch": 0.32966316551973923, "grad_norm": 1.3008260948309462, "learning_rate": 1.5649044100269374e-05, "loss": 0.8039, "step": 9102 }, { "epoch": 0.3296993842810576, "grad_norm": 1.1142979246069191, "learning_rate": 1.564807609712384e-05, "loss": 0.7549, "step": 9103 }, { "epoch": 0.3297356030423759, "grad_norm": 1.4683237928507418, "learning_rate": 1.564710801625582e-05, "loss": 0.7809, "step": 9104 }, { "epoch": 0.3297718218036943, "grad_norm": 1.47860384669077, "learning_rate": 1.564613985767863e-05, "loss": 0.8637, "step": 9105 }, { "epoch": 0.3298080405650127, "grad_norm": 1.736241564829737, "learning_rate": 1.5645171621405594e-05, "loss": 0.8042, "step": 9106 }, { "epoch": 0.32984425932633105, "grad_norm": 1.3843210923213136, "learning_rate": 1.5644203307450036e-05, "loss": 0.7507, "step": 9107 }, { "epoch": 0.3298804780876494, "grad_norm": 1.439560003725315, "learning_rate": 1.564323491582528e-05, "loss": 0.7934, "step": 9108 }, { "epoch": 0.32991669684896774, "grad_norm": 1.1219531447123237, "learning_rate": 1.5642266446544655e-05, "loss": 0.7622, "step": 9109 }, { "epoch": 0.3299529156102861, "grad_norm": 1.1685641350161815, "learning_rate": 1.5641297899621486e-05, "loss": 0.7912, "step": 9110 }, { "epoch": 0.3299891343716045, "grad_norm": 1.0641296627189936, "learning_rate": 1.56403292750691e-05, "loss": 0.7821, "step": 9111 }, { "epoch": 0.33002535313292286, "grad_norm": 1.439699188174212, "learning_rate": 1.563936057290083e-05, "loss": 0.8165, "step": 9112 }, { "epoch": 0.33006157189424123, "grad_norm": 1.479109128456405, "learning_rate": 1.5638391793129998e-05, "loss": 0.7644, "step": 9113 }, { "epoch": 0.33009779065555955, "grad_norm": 1.4256926691939802, "learning_rate": 1.5637422935769942e-05, "loss": 0.7755, "step": 9114 }, { "epoch": 0.3301340094168779, "grad_norm": 1.0143463552306033, "learning_rate": 1.5636454000833995e-05, "loss": 0.7546, "step": 9115 }, { "epoch": 0.3301702281781963, "grad_norm": 1.1379886569806779, "learning_rate": 1.5635484988335488e-05, "loss": 0.7855, "step": 9116 }, { "epoch": 0.3302064469395147, "grad_norm": 1.5141724896905098, "learning_rate": 1.5634515898287753e-05, "loss": 0.7675, "step": 9117 }, { "epoch": 0.33024266570083305, "grad_norm": 1.4120081908827464, "learning_rate": 1.5633546730704134e-05, "loss": 0.7261, "step": 9118 }, { "epoch": 0.33027888446215137, "grad_norm": 1.2842477064285907, "learning_rate": 1.5632577485597958e-05, "loss": 0.8312, "step": 9119 }, { "epoch": 0.33031510322346974, "grad_norm": 1.5317899484283723, "learning_rate": 1.563160816298257e-05, "loss": 0.8459, "step": 9120 }, { "epoch": 0.3303513219847881, "grad_norm": 1.4877617422394334, "learning_rate": 1.56306387628713e-05, "loss": 0.8287, "step": 9121 }, { "epoch": 0.3303875407461065, "grad_norm": 1.2790311923309081, "learning_rate": 1.5629669285277497e-05, "loss": 0.7181, "step": 9122 }, { "epoch": 0.33042375950742486, "grad_norm": 1.3548101369660586, "learning_rate": 1.5628699730214495e-05, "loss": 0.8376, "step": 9123 }, { "epoch": 0.3304599782687432, "grad_norm": 1.1932276428299788, "learning_rate": 1.562773009769564e-05, "loss": 0.7996, "step": 9124 }, { "epoch": 0.33049619703006156, "grad_norm": 1.2548436188421657, "learning_rate": 1.5626760387734273e-05, "loss": 0.7339, "step": 9125 }, { "epoch": 0.33053241579137993, "grad_norm": 1.2095449241813048, "learning_rate": 1.562579060034374e-05, "loss": 0.757, "step": 9126 }, { "epoch": 0.3305686345526983, "grad_norm": 1.1441097420146373, "learning_rate": 1.562482073553739e-05, "loss": 0.8682, "step": 9127 }, { "epoch": 0.3306048533140167, "grad_norm": 1.5570680547183813, "learning_rate": 1.5623850793328556e-05, "loss": 0.7533, "step": 9128 }, { "epoch": 0.330641072075335, "grad_norm": 1.4052150487533124, "learning_rate": 1.56228807737306e-05, "loss": 0.7763, "step": 9129 }, { "epoch": 0.33067729083665337, "grad_norm": 1.344480059766937, "learning_rate": 1.562191067675686e-05, "loss": 0.7946, "step": 9130 }, { "epoch": 0.33071350959797174, "grad_norm": 1.3924683407876628, "learning_rate": 1.562094050242069e-05, "loss": 0.8599, "step": 9131 }, { "epoch": 0.3307497283592901, "grad_norm": 1.4783817593482977, "learning_rate": 1.561997025073544e-05, "loss": 0.8231, "step": 9132 }, { "epoch": 0.3307859471206085, "grad_norm": 1.184155072584466, "learning_rate": 1.561899992171446e-05, "loss": 0.789, "step": 9133 }, { "epoch": 0.3308221658819268, "grad_norm": 1.434953439759656, "learning_rate": 1.5618029515371106e-05, "loss": 0.7567, "step": 9134 }, { "epoch": 0.3308583846432452, "grad_norm": 1.430342196038957, "learning_rate": 1.561705903171873e-05, "loss": 0.7354, "step": 9135 }, { "epoch": 0.33089460340456356, "grad_norm": 1.419571763181151, "learning_rate": 1.561608847077068e-05, "loss": 0.7774, "step": 9136 }, { "epoch": 0.33093082216588193, "grad_norm": 1.753759673265927, "learning_rate": 1.5615117832540324e-05, "loss": 0.7264, "step": 9137 }, { "epoch": 0.3309670409272003, "grad_norm": 1.0881969869762196, "learning_rate": 1.5614147117041012e-05, "loss": 0.7521, "step": 9138 }, { "epoch": 0.3310032596885186, "grad_norm": 1.49213416131541, "learning_rate": 1.5613176324286102e-05, "loss": 0.8003, "step": 9139 }, { "epoch": 0.331039478449837, "grad_norm": 1.3996464880029333, "learning_rate": 1.5612205454288957e-05, "loss": 0.8168, "step": 9140 }, { "epoch": 0.3310756972111554, "grad_norm": 1.4427815533785806, "learning_rate": 1.5611234507062928e-05, "loss": 0.7972, "step": 9141 }, { "epoch": 0.33111191597247375, "grad_norm": 1.108910235118236, "learning_rate": 1.5610263482621383e-05, "loss": 0.7701, "step": 9142 }, { "epoch": 0.3311481347337921, "grad_norm": 1.42034845490264, "learning_rate": 1.5609292380977684e-05, "loss": 0.7959, "step": 9143 }, { "epoch": 0.33118435349511044, "grad_norm": 1.4107898480179428, "learning_rate": 1.5608321202145194e-05, "loss": 0.7885, "step": 9144 }, { "epoch": 0.3312205722564288, "grad_norm": 1.4118742707524183, "learning_rate": 1.5607349946137275e-05, "loss": 0.7995, "step": 9145 }, { "epoch": 0.3312567910177472, "grad_norm": 1.443556993374527, "learning_rate": 1.5606378612967295e-05, "loss": 0.7148, "step": 9146 }, { "epoch": 0.33129300977906556, "grad_norm": 1.3456881339590492, "learning_rate": 1.5605407202648618e-05, "loss": 0.749, "step": 9147 }, { "epoch": 0.33132922854038394, "grad_norm": 1.2906286176072006, "learning_rate": 1.5604435715194612e-05, "loss": 0.7809, "step": 9148 }, { "epoch": 0.3313654473017023, "grad_norm": 1.4662519400186216, "learning_rate": 1.5603464150618645e-05, "loss": 0.8989, "step": 9149 }, { "epoch": 0.33140166606302063, "grad_norm": 1.4385057538314905, "learning_rate": 1.560249250893409e-05, "loss": 0.8125, "step": 9150 }, { "epoch": 0.331437884824339, "grad_norm": 1.4507475498293154, "learning_rate": 1.5601520790154313e-05, "loss": 0.7711, "step": 9151 }, { "epoch": 0.3314741035856574, "grad_norm": 1.2009010089980063, "learning_rate": 1.560054899429269e-05, "loss": 0.8109, "step": 9152 }, { "epoch": 0.33151032234697575, "grad_norm": 1.3648903062541076, "learning_rate": 1.5599577121362592e-05, "loss": 0.7666, "step": 9153 }, { "epoch": 0.3315465411082941, "grad_norm": 1.528266067273462, "learning_rate": 1.559860517137739e-05, "loss": 0.7799, "step": 9154 }, { "epoch": 0.33158275986961244, "grad_norm": 1.1420028893165208, "learning_rate": 1.5597633144350465e-05, "loss": 0.8082, "step": 9155 }, { "epoch": 0.3316189786309308, "grad_norm": 1.168810447493361, "learning_rate": 1.559666104029519e-05, "loss": 0.7818, "step": 9156 }, { "epoch": 0.3316551973922492, "grad_norm": 1.3493049032473736, "learning_rate": 1.5595688859224937e-05, "loss": 0.758, "step": 9157 }, { "epoch": 0.33169141615356756, "grad_norm": 1.4360909877535124, "learning_rate": 1.559471660115309e-05, "loss": 0.8635, "step": 9158 }, { "epoch": 0.33172763491488594, "grad_norm": 1.0931371092434035, "learning_rate": 1.559374426609303e-05, "loss": 0.8021, "step": 9159 }, { "epoch": 0.33176385367620426, "grad_norm": 1.4515563447840243, "learning_rate": 1.5592771854058128e-05, "loss": 0.8137, "step": 9160 }, { "epoch": 0.33180007243752263, "grad_norm": 1.180019539957564, "learning_rate": 1.5591799365061774e-05, "loss": 0.752, "step": 9161 }, { "epoch": 0.331836291198841, "grad_norm": 1.2223920364324132, "learning_rate": 1.5590826799117348e-05, "loss": 0.7436, "step": 9162 }, { "epoch": 0.3318725099601594, "grad_norm": 1.2568411189715707, "learning_rate": 1.5589854156238235e-05, "loss": 0.7489, "step": 9163 }, { "epoch": 0.33190872872147775, "grad_norm": 1.535717119160047, "learning_rate": 1.5588881436437813e-05, "loss": 0.8105, "step": 9164 }, { "epoch": 0.33194494748279607, "grad_norm": 1.4137665428345578, "learning_rate": 1.5587908639729472e-05, "loss": 0.7558, "step": 9165 }, { "epoch": 0.33198116624411445, "grad_norm": 1.4264151679374595, "learning_rate": 1.5586935766126597e-05, "loss": 0.7934, "step": 9166 }, { "epoch": 0.3320173850054328, "grad_norm": 1.1000461241423753, "learning_rate": 1.5585962815642583e-05, "loss": 0.7942, "step": 9167 }, { "epoch": 0.3320536037667512, "grad_norm": 1.5355054176768193, "learning_rate": 1.5584989788290806e-05, "loss": 0.7811, "step": 9168 }, { "epoch": 0.33208982252806957, "grad_norm": 1.3214730024059838, "learning_rate": 1.5584016684084664e-05, "loss": 0.8294, "step": 9169 }, { "epoch": 0.3321260412893879, "grad_norm": 1.4613243613338667, "learning_rate": 1.5583043503037547e-05, "loss": 0.737, "step": 9170 }, { "epoch": 0.33216226005070626, "grad_norm": 1.1285098729452463, "learning_rate": 1.5582070245162847e-05, "loss": 0.7931, "step": 9171 }, { "epoch": 0.33219847881202463, "grad_norm": 1.4052937574421895, "learning_rate": 1.558109691047395e-05, "loss": 0.8117, "step": 9172 }, { "epoch": 0.332234697573343, "grad_norm": 1.1720165942500527, "learning_rate": 1.558012349898426e-05, "loss": 0.8385, "step": 9173 }, { "epoch": 0.3322709163346614, "grad_norm": 1.3474579264348254, "learning_rate": 1.557915001070717e-05, "loss": 0.7411, "step": 9174 }, { "epoch": 0.3323071350959797, "grad_norm": 1.4092303264980734, "learning_rate": 1.557817644565607e-05, "loss": 0.8421, "step": 9175 }, { "epoch": 0.3323433538572981, "grad_norm": 1.2250271476990164, "learning_rate": 1.557720280384436e-05, "loss": 0.7333, "step": 9176 }, { "epoch": 0.33237957261861645, "grad_norm": 1.0876568274643046, "learning_rate": 1.5576229085285438e-05, "loss": 0.733, "step": 9177 }, { "epoch": 0.3324157913799348, "grad_norm": 1.435266473204707, "learning_rate": 1.5575255289992707e-05, "loss": 0.8019, "step": 9178 }, { "epoch": 0.3324520101412532, "grad_norm": 1.1026411656497892, "learning_rate": 1.557428141797956e-05, "loss": 0.7394, "step": 9179 }, { "epoch": 0.3324882289025715, "grad_norm": 1.372565350537953, "learning_rate": 1.557330746925941e-05, "loss": 0.7666, "step": 9180 }, { "epoch": 0.3325244476638899, "grad_norm": 1.3710432138394655, "learning_rate": 1.5572333443845647e-05, "loss": 0.8282, "step": 9181 }, { "epoch": 0.33256066642520826, "grad_norm": 1.2394618267742106, "learning_rate": 1.557135934175168e-05, "loss": 0.6946, "step": 9182 }, { "epoch": 0.33259688518652664, "grad_norm": 1.5065011068422978, "learning_rate": 1.5570385162990916e-05, "loss": 0.7946, "step": 9183 }, { "epoch": 0.332633103947845, "grad_norm": 1.3914863043453616, "learning_rate": 1.5569410907576753e-05, "loss": 0.8107, "step": 9184 }, { "epoch": 0.33266932270916333, "grad_norm": 1.3437409747175484, "learning_rate": 1.5568436575522604e-05, "loss": 0.7715, "step": 9185 }, { "epoch": 0.3327055414704817, "grad_norm": 1.40469373048158, "learning_rate": 1.5567462166841874e-05, "loss": 0.716, "step": 9186 }, { "epoch": 0.3327417602318001, "grad_norm": 1.3155370992690016, "learning_rate": 1.5566487681547977e-05, "loss": 0.7501, "step": 9187 }, { "epoch": 0.33277797899311845, "grad_norm": 1.3632628358404117, "learning_rate": 1.556551311965431e-05, "loss": 0.7838, "step": 9188 }, { "epoch": 0.3328141977544368, "grad_norm": 1.5174186259420872, "learning_rate": 1.55645384811743e-05, "loss": 0.764, "step": 9189 }, { "epoch": 0.33285041651575514, "grad_norm": 1.466781487873098, "learning_rate": 1.556356376612135e-05, "loss": 0.7547, "step": 9190 }, { "epoch": 0.3328866352770735, "grad_norm": 1.3786292324668739, "learning_rate": 1.556258897450887e-05, "loss": 0.7901, "step": 9191 }, { "epoch": 0.3329228540383919, "grad_norm": 1.0119637908057244, "learning_rate": 1.556161410635028e-05, "loss": 0.7221, "step": 9192 }, { "epoch": 0.33295907279971027, "grad_norm": 1.374754455029152, "learning_rate": 1.5560639161658997e-05, "loss": 0.8286, "step": 9193 }, { "epoch": 0.33299529156102864, "grad_norm": 1.0459944412058837, "learning_rate": 1.5559664140448428e-05, "loss": 0.7426, "step": 9194 }, { "epoch": 0.33303151032234696, "grad_norm": 1.365523785843818, "learning_rate": 1.5558689042732e-05, "loss": 0.7419, "step": 9195 }, { "epoch": 0.33306772908366533, "grad_norm": 1.188771287429811, "learning_rate": 1.5557713868523123e-05, "loss": 0.7623, "step": 9196 }, { "epoch": 0.3331039478449837, "grad_norm": 1.0316944146839293, "learning_rate": 1.5556738617835217e-05, "loss": 0.7053, "step": 9197 }, { "epoch": 0.3331401666063021, "grad_norm": 1.3124878856548312, "learning_rate": 1.555576329068171e-05, "loss": 0.7363, "step": 9198 }, { "epoch": 0.33317638536762045, "grad_norm": 1.509226566449074, "learning_rate": 1.5554787887076016e-05, "loss": 0.8698, "step": 9199 }, { "epoch": 0.33321260412893877, "grad_norm": 1.1653320205168682, "learning_rate": 1.555381240703156e-05, "loss": 0.7614, "step": 9200 }, { "epoch": 0.33324882289025715, "grad_norm": 1.5718025971037413, "learning_rate": 1.5552836850561764e-05, "loss": 0.7585, "step": 9201 }, { "epoch": 0.3332850416515755, "grad_norm": 1.429664735404376, "learning_rate": 1.5551861217680057e-05, "loss": 0.7937, "step": 9202 }, { "epoch": 0.3333212604128939, "grad_norm": 1.5388711783779436, "learning_rate": 1.5550885508399857e-05, "loss": 0.7892, "step": 9203 }, { "epoch": 0.33335747917421227, "grad_norm": 1.3624888798058656, "learning_rate": 1.55499097227346e-05, "loss": 0.7682, "step": 9204 }, { "epoch": 0.3333936979355306, "grad_norm": 1.6211105805408077, "learning_rate": 1.5548933860697706e-05, "loss": 0.7304, "step": 9205 }, { "epoch": 0.33342991669684896, "grad_norm": 1.2885861557049483, "learning_rate": 1.5547957922302606e-05, "loss": 0.7031, "step": 9206 }, { "epoch": 0.33346613545816733, "grad_norm": 1.4021440658681852, "learning_rate": 1.554698190756273e-05, "loss": 0.7895, "step": 9207 }, { "epoch": 0.3335023542194857, "grad_norm": 1.2826517817108891, "learning_rate": 1.554600581649151e-05, "loss": 0.7952, "step": 9208 }, { "epoch": 0.3335385729808041, "grad_norm": 1.3559124698535379, "learning_rate": 1.5545029649102378e-05, "loss": 0.7473, "step": 9209 }, { "epoch": 0.3335747917421224, "grad_norm": 1.3237891238735235, "learning_rate": 1.5544053405408763e-05, "loss": 0.742, "step": 9210 }, { "epoch": 0.3336110105034408, "grad_norm": 1.4911981605826792, "learning_rate": 1.5543077085424104e-05, "loss": 0.8258, "step": 9211 }, { "epoch": 0.33364722926475915, "grad_norm": 1.4848402561558898, "learning_rate": 1.5542100689161833e-05, "loss": 0.779, "step": 9212 }, { "epoch": 0.3336834480260775, "grad_norm": 1.658273106958441, "learning_rate": 1.554112421663539e-05, "loss": 0.7991, "step": 9213 }, { "epoch": 0.3337196667873959, "grad_norm": 1.1191244114309644, "learning_rate": 1.5540147667858204e-05, "loss": 0.8503, "step": 9214 }, { "epoch": 0.3337558855487142, "grad_norm": 1.6627323404170726, "learning_rate": 1.553917104284372e-05, "loss": 0.7135, "step": 9215 }, { "epoch": 0.3337921043100326, "grad_norm": 1.4618147954575782, "learning_rate": 1.5538194341605372e-05, "loss": 0.8897, "step": 9216 }, { "epoch": 0.33382832307135096, "grad_norm": 1.5205799017093147, "learning_rate": 1.553721756415661e-05, "loss": 0.8049, "step": 9217 }, { "epoch": 0.33386454183266934, "grad_norm": 1.4510262417526822, "learning_rate": 1.5536240710510865e-05, "loss": 0.7517, "step": 9218 }, { "epoch": 0.3339007605939877, "grad_norm": 1.0155061449943932, "learning_rate": 1.5535263780681582e-05, "loss": 0.7183, "step": 9219 }, { "epoch": 0.33393697935530603, "grad_norm": 1.4981514112163996, "learning_rate": 1.5534286774682205e-05, "loss": 0.802, "step": 9220 }, { "epoch": 0.3339731981166244, "grad_norm": 1.3545643932291194, "learning_rate": 1.5533309692526183e-05, "loss": 0.7487, "step": 9221 }, { "epoch": 0.3340094168779428, "grad_norm": 1.4215827067686009, "learning_rate": 1.5532332534226957e-05, "loss": 0.8234, "step": 9222 }, { "epoch": 0.33404563563926115, "grad_norm": 1.3811444563391964, "learning_rate": 1.5531355299797973e-05, "loss": 0.7075, "step": 9223 }, { "epoch": 0.3340818544005795, "grad_norm": 1.4537294794345705, "learning_rate": 1.553037798925268e-05, "loss": 0.821, "step": 9224 }, { "epoch": 0.33411807316189784, "grad_norm": 1.4524687597802146, "learning_rate": 1.552940060260453e-05, "loss": 0.7549, "step": 9225 }, { "epoch": 0.3341542919232162, "grad_norm": 1.458795887671534, "learning_rate": 1.552842313986696e-05, "loss": 0.7895, "step": 9226 }, { "epoch": 0.3341905106845346, "grad_norm": 1.3721466083966727, "learning_rate": 1.5527445601053437e-05, "loss": 0.8504, "step": 9227 }, { "epoch": 0.33422672944585297, "grad_norm": 1.0954859279467255, "learning_rate": 1.5526467986177405e-05, "loss": 0.7494, "step": 9228 }, { "epoch": 0.33426294820717134, "grad_norm": 1.2665863077091384, "learning_rate": 1.5525490295252316e-05, "loss": 0.821, "step": 9229 }, { "epoch": 0.33429916696848966, "grad_norm": 1.3489376503246788, "learning_rate": 1.5524512528291627e-05, "loss": 0.7665, "step": 9230 }, { "epoch": 0.33433538572980803, "grad_norm": 1.414831271951208, "learning_rate": 1.5523534685308792e-05, "loss": 0.8374, "step": 9231 }, { "epoch": 0.3343716044911264, "grad_norm": 1.6992904034119043, "learning_rate": 1.5522556766317266e-05, "loss": 0.9256, "step": 9232 }, { "epoch": 0.3344078232524448, "grad_norm": 1.464051804406783, "learning_rate": 1.55215787713305e-05, "loss": 0.7727, "step": 9233 }, { "epoch": 0.33444404201376315, "grad_norm": 1.5799711566077443, "learning_rate": 1.5520600700361964e-05, "loss": 0.8131, "step": 9234 }, { "epoch": 0.3344802607750815, "grad_norm": 1.6565406722502805, "learning_rate": 1.5519622553425114e-05, "loss": 0.7444, "step": 9235 }, { "epoch": 0.33451647953639985, "grad_norm": 1.1291961453420807, "learning_rate": 1.5518644330533403e-05, "loss": 0.7579, "step": 9236 }, { "epoch": 0.3345526982977182, "grad_norm": 1.1711933039712905, "learning_rate": 1.55176660317003e-05, "loss": 0.7534, "step": 9237 }, { "epoch": 0.3345889170590366, "grad_norm": 1.2323626407286874, "learning_rate": 1.5516687656939263e-05, "loss": 0.7066, "step": 9238 }, { "epoch": 0.33462513582035497, "grad_norm": 1.6908983376233935, "learning_rate": 1.5515709206263757e-05, "loss": 0.7245, "step": 9239 }, { "epoch": 0.3346613545816733, "grad_norm": 1.4076815365368314, "learning_rate": 1.5514730679687244e-05, "loss": 0.7719, "step": 9240 }, { "epoch": 0.33469757334299166, "grad_norm": 1.5492032848650055, "learning_rate": 1.5513752077223193e-05, "loss": 0.7217, "step": 9241 }, { "epoch": 0.33473379210431004, "grad_norm": 1.1927614173327221, "learning_rate": 1.551277339888507e-05, "loss": 0.8372, "step": 9242 }, { "epoch": 0.3347700108656284, "grad_norm": 1.425804029570157, "learning_rate": 1.5511794644686337e-05, "loss": 0.7721, "step": 9243 }, { "epoch": 0.3348062296269468, "grad_norm": 1.2244378974000751, "learning_rate": 1.5510815814640468e-05, "loss": 0.7788, "step": 9244 }, { "epoch": 0.3348424483882651, "grad_norm": 1.2754767924237416, "learning_rate": 1.550983690876093e-05, "loss": 0.8581, "step": 9245 }, { "epoch": 0.3348786671495835, "grad_norm": 1.4212091112839387, "learning_rate": 1.55088579270612e-05, "loss": 0.7641, "step": 9246 }, { "epoch": 0.33491488591090185, "grad_norm": 1.3013355732203264, "learning_rate": 1.5507878869554738e-05, "loss": 0.7904, "step": 9247 }, { "epoch": 0.3349511046722202, "grad_norm": 1.4092338535740008, "learning_rate": 1.5506899736255027e-05, "loss": 0.8033, "step": 9248 }, { "epoch": 0.3349873234335386, "grad_norm": 1.1336913137023485, "learning_rate": 1.5505920527175534e-05, "loss": 0.7563, "step": 9249 }, { "epoch": 0.3350235421948569, "grad_norm": 1.5512825758508337, "learning_rate": 1.550494124232974e-05, "loss": 0.8438, "step": 9250 }, { "epoch": 0.3350597609561753, "grad_norm": 0.9827793612179958, "learning_rate": 1.550396188173112e-05, "loss": 0.7653, "step": 9251 }, { "epoch": 0.33509597971749366, "grad_norm": 1.5756039021532888, "learning_rate": 1.5502982445393144e-05, "loss": 0.8463, "step": 9252 }, { "epoch": 0.33513219847881204, "grad_norm": 1.2820083014034218, "learning_rate": 1.550200293332929e-05, "loss": 0.8102, "step": 9253 }, { "epoch": 0.3351684172401304, "grad_norm": 1.5282939872058814, "learning_rate": 1.5501023345553046e-05, "loss": 0.8135, "step": 9254 }, { "epoch": 0.33520463600144873, "grad_norm": 1.5485387293640933, "learning_rate": 1.5500043682077887e-05, "loss": 0.7979, "step": 9255 }, { "epoch": 0.3352408547627671, "grad_norm": 1.3543544993589578, "learning_rate": 1.5499063942917295e-05, "loss": 0.6953, "step": 9256 }, { "epoch": 0.3352770735240855, "grad_norm": 1.4921056839847198, "learning_rate": 1.549808412808475e-05, "loss": 0.7676, "step": 9257 }, { "epoch": 0.33531329228540385, "grad_norm": 1.3749950792213732, "learning_rate": 1.5497104237593736e-05, "loss": 0.8315, "step": 9258 }, { "epoch": 0.3353495110467222, "grad_norm": 1.3541515749921789, "learning_rate": 1.549612427145774e-05, "loss": 0.7395, "step": 9259 }, { "epoch": 0.33538572980804054, "grad_norm": 1.2967362010027235, "learning_rate": 1.5495144229690243e-05, "loss": 0.7925, "step": 9260 }, { "epoch": 0.3354219485693589, "grad_norm": 1.3505361634720792, "learning_rate": 1.5494164112304734e-05, "loss": 0.7278, "step": 9261 }, { "epoch": 0.3354581673306773, "grad_norm": 1.4166832652212336, "learning_rate": 1.5493183919314696e-05, "loss": 0.805, "step": 9262 }, { "epoch": 0.33549438609199567, "grad_norm": 1.7790208513709604, "learning_rate": 1.5492203650733623e-05, "loss": 0.7521, "step": 9263 }, { "epoch": 0.33553060485331404, "grad_norm": 1.4486937651631018, "learning_rate": 1.5491223306575e-05, "loss": 0.8058, "step": 9264 }, { "epoch": 0.33556682361463236, "grad_norm": 1.1024858894663359, "learning_rate": 1.5490242886852323e-05, "loss": 0.7023, "step": 9265 }, { "epoch": 0.33560304237595073, "grad_norm": 1.4988690747236266, "learning_rate": 1.5489262391579076e-05, "loss": 0.8173, "step": 9266 }, { "epoch": 0.3356392611372691, "grad_norm": 1.580981823689518, "learning_rate": 1.548828182076876e-05, "loss": 0.7852, "step": 9267 }, { "epoch": 0.3356754798985875, "grad_norm": 1.321583727694086, "learning_rate": 1.5487301174434858e-05, "loss": 0.7727, "step": 9268 }, { "epoch": 0.33571169865990586, "grad_norm": 1.32762950321438, "learning_rate": 1.5486320452590872e-05, "loss": 0.8061, "step": 9269 }, { "epoch": 0.3357479174212242, "grad_norm": 1.3563877722425608, "learning_rate": 1.5485339655250294e-05, "loss": 0.7964, "step": 9270 }, { "epoch": 0.33578413618254255, "grad_norm": 1.344747304929549, "learning_rate": 1.5484358782426626e-05, "loss": 0.7728, "step": 9271 }, { "epoch": 0.3358203549438609, "grad_norm": 1.6047264968197166, "learning_rate": 1.5483377834133362e-05, "loss": 0.8914, "step": 9272 }, { "epoch": 0.3358565737051793, "grad_norm": 1.4534494232708985, "learning_rate": 1.5482396810384002e-05, "loss": 0.7973, "step": 9273 }, { "epoch": 0.33589279246649767, "grad_norm": 1.9030224786418164, "learning_rate": 1.5481415711192042e-05, "loss": 0.8444, "step": 9274 }, { "epoch": 0.335929011227816, "grad_norm": 1.271864943236114, "learning_rate": 1.5480434536570986e-05, "loss": 0.7715, "step": 9275 }, { "epoch": 0.33596522998913436, "grad_norm": 1.407274317543275, "learning_rate": 1.5479453286534334e-05, "loss": 0.7685, "step": 9276 }, { "epoch": 0.33600144875045274, "grad_norm": 1.3345141503826659, "learning_rate": 1.547847196109559e-05, "loss": 0.7819, "step": 9277 }, { "epoch": 0.3360376675117711, "grad_norm": 1.2665601833329907, "learning_rate": 1.547749056026826e-05, "loss": 0.786, "step": 9278 }, { "epoch": 0.3360738862730895, "grad_norm": 1.3282131972199438, "learning_rate": 1.5476509084065846e-05, "loss": 0.8083, "step": 9279 }, { "epoch": 0.3361101050344078, "grad_norm": 1.5160603475621055, "learning_rate": 1.5475527532501857e-05, "loss": 0.851, "step": 9280 }, { "epoch": 0.3361463237957262, "grad_norm": 1.3884289773204668, "learning_rate": 1.5474545905589797e-05, "loss": 0.7669, "step": 9281 }, { "epoch": 0.33618254255704455, "grad_norm": 1.5233316690414032, "learning_rate": 1.5473564203343173e-05, "loss": 0.8198, "step": 9282 }, { "epoch": 0.3362187613183629, "grad_norm": 1.4912497060116767, "learning_rate": 1.54725824257755e-05, "loss": 0.8021, "step": 9283 }, { "epoch": 0.3362549800796813, "grad_norm": 1.148743273189319, "learning_rate": 1.5471600572900278e-05, "loss": 0.8106, "step": 9284 }, { "epoch": 0.3362911988409996, "grad_norm": 1.7599365879550806, "learning_rate": 1.5470618644731027e-05, "loss": 0.7742, "step": 9285 }, { "epoch": 0.336327417602318, "grad_norm": 1.3788050838674082, "learning_rate": 1.5469636641281262e-05, "loss": 0.7636, "step": 9286 }, { "epoch": 0.33636363636363636, "grad_norm": 1.01076586362733, "learning_rate": 1.546865456256448e-05, "loss": 0.7293, "step": 9287 }, { "epoch": 0.33639985512495474, "grad_norm": 1.4754227643328914, "learning_rate": 1.5467672408594214e-05, "loss": 0.7171, "step": 9288 }, { "epoch": 0.3364360738862731, "grad_norm": 1.5153393046786003, "learning_rate": 1.546669017938397e-05, "loss": 0.8067, "step": 9289 }, { "epoch": 0.33647229264759143, "grad_norm": 1.4541031854747275, "learning_rate": 1.5465707874947266e-05, "loss": 0.8215, "step": 9290 }, { "epoch": 0.3365085114089098, "grad_norm": 1.2858504852250916, "learning_rate": 1.546472549529762e-05, "loss": 0.8117, "step": 9291 }, { "epoch": 0.3365447301702282, "grad_norm": 1.5189355317456084, "learning_rate": 1.5463743040448547e-05, "loss": 0.6994, "step": 9292 }, { "epoch": 0.33658094893154655, "grad_norm": 1.3686927268214877, "learning_rate": 1.5462760510413568e-05, "loss": 0.818, "step": 9293 }, { "epoch": 0.3366171676928649, "grad_norm": 1.1939362747557658, "learning_rate": 1.5461777905206206e-05, "loss": 0.6959, "step": 9294 }, { "epoch": 0.33665338645418325, "grad_norm": 1.3349210388539072, "learning_rate": 1.5460795224839982e-05, "loss": 0.7579, "step": 9295 }, { "epoch": 0.3366896052155016, "grad_norm": 1.2907075276267115, "learning_rate": 1.5459812469328417e-05, "loss": 0.7916, "step": 9296 }, { "epoch": 0.33672582397682, "grad_norm": 1.3588083342610757, "learning_rate": 1.5458829638685032e-05, "loss": 0.7837, "step": 9297 }, { "epoch": 0.33676204273813837, "grad_norm": 1.3873233370445106, "learning_rate": 1.5457846732923358e-05, "loss": 0.8443, "step": 9298 }, { "epoch": 0.33679826149945674, "grad_norm": 1.3431793022594751, "learning_rate": 1.5456863752056918e-05, "loss": 0.7545, "step": 9299 }, { "epoch": 0.33683448026077506, "grad_norm": 1.418705173658124, "learning_rate": 1.5455880696099236e-05, "loss": 0.7689, "step": 9300 }, { "epoch": 0.33687069902209343, "grad_norm": 1.4280483524217409, "learning_rate": 1.5454897565063843e-05, "loss": 0.8134, "step": 9301 }, { "epoch": 0.3369069177834118, "grad_norm": 2.8381618506667112, "learning_rate": 1.5453914358964265e-05, "loss": 0.7324, "step": 9302 }, { "epoch": 0.3369431365447302, "grad_norm": 1.3157368784846268, "learning_rate": 1.545293107781403e-05, "loss": 0.7911, "step": 9303 }, { "epoch": 0.33697935530604856, "grad_norm": 1.4217051244745826, "learning_rate": 1.5451947721626676e-05, "loss": 0.746, "step": 9304 }, { "epoch": 0.3370155740673669, "grad_norm": 1.6789972550448948, "learning_rate": 1.5450964290415732e-05, "loss": 0.7239, "step": 9305 }, { "epoch": 0.33705179282868525, "grad_norm": 1.346573068161662, "learning_rate": 1.544998078419473e-05, "loss": 0.887, "step": 9306 }, { "epoch": 0.3370880115900036, "grad_norm": 1.330067123389993, "learning_rate": 1.5448997202977198e-05, "loss": 0.7277, "step": 9307 }, { "epoch": 0.337124230351322, "grad_norm": 1.3921797130294158, "learning_rate": 1.5448013546776682e-05, "loss": 0.787, "step": 9308 }, { "epoch": 0.33716044911264037, "grad_norm": 1.097434252323551, "learning_rate": 1.5447029815606713e-05, "loss": 0.7418, "step": 9309 }, { "epoch": 0.3371966678739587, "grad_norm": 1.2754204149257602, "learning_rate": 1.5446046009480823e-05, "loss": 0.7327, "step": 9310 }, { "epoch": 0.33723288663527706, "grad_norm": 1.2896574274468602, "learning_rate": 1.5445062128412555e-05, "loss": 0.7468, "step": 9311 }, { "epoch": 0.33726910539659544, "grad_norm": 1.4770969755657375, "learning_rate": 1.544407817241545e-05, "loss": 0.8631, "step": 9312 }, { "epoch": 0.3373053241579138, "grad_norm": 1.4487177903225186, "learning_rate": 1.5443094141503045e-05, "loss": 0.8325, "step": 9313 }, { "epoch": 0.3373415429192322, "grad_norm": 1.3842836363042208, "learning_rate": 1.5442110035688883e-05, "loss": 0.8671, "step": 9314 }, { "epoch": 0.3373777616805505, "grad_norm": 1.1833425293008792, "learning_rate": 1.5441125854986504e-05, "loss": 0.7157, "step": 9315 }, { "epoch": 0.3374139804418689, "grad_norm": 1.16015453771985, "learning_rate": 1.5440141599409452e-05, "loss": 0.7566, "step": 9316 }, { "epoch": 0.33745019920318725, "grad_norm": 1.5260293175378852, "learning_rate": 1.5439157268971276e-05, "loss": 0.8087, "step": 9317 }, { "epoch": 0.3374864179645056, "grad_norm": 1.1132841072024942, "learning_rate": 1.543817286368551e-05, "loss": 0.8097, "step": 9318 }, { "epoch": 0.337522636725824, "grad_norm": 1.0990095448416681, "learning_rate": 1.543718838356571e-05, "loss": 0.7728, "step": 9319 }, { "epoch": 0.3375588554871423, "grad_norm": 1.1484458070125325, "learning_rate": 1.543620382862542e-05, "loss": 0.7228, "step": 9320 }, { "epoch": 0.3375950742484607, "grad_norm": 1.123369216787788, "learning_rate": 1.5435219198878192e-05, "loss": 0.7723, "step": 9321 }, { "epoch": 0.33763129300977907, "grad_norm": 1.365194745777917, "learning_rate": 1.5434234494337568e-05, "loss": 0.7589, "step": 9322 }, { "epoch": 0.33766751177109744, "grad_norm": 1.2176128642847375, "learning_rate": 1.5433249715017103e-05, "loss": 0.7323, "step": 9323 }, { "epoch": 0.3377037305324158, "grad_norm": 1.5180432185049044, "learning_rate": 1.5432264860930345e-05, "loss": 0.7826, "step": 9324 }, { "epoch": 0.33773994929373413, "grad_norm": 1.489349653820342, "learning_rate": 1.5431279932090856e-05, "loss": 0.7895, "step": 9325 }, { "epoch": 0.3377761680550525, "grad_norm": 1.3977235309765355, "learning_rate": 1.5430294928512178e-05, "loss": 0.7707, "step": 9326 }, { "epoch": 0.3378123868163709, "grad_norm": 1.4228848909651985, "learning_rate": 1.542930985020787e-05, "loss": 0.7657, "step": 9327 }, { "epoch": 0.33784860557768925, "grad_norm": 1.0743151360044294, "learning_rate": 1.542832469719149e-05, "loss": 0.7249, "step": 9328 }, { "epoch": 0.33788482433900763, "grad_norm": 1.7899416799741235, "learning_rate": 1.5427339469476594e-05, "loss": 0.8009, "step": 9329 }, { "epoch": 0.33792104310032595, "grad_norm": 1.456091274617958, "learning_rate": 1.5426354167076733e-05, "loss": 0.7278, "step": 9330 }, { "epoch": 0.3379572618616443, "grad_norm": 1.3634694537895624, "learning_rate": 1.5425368790005474e-05, "loss": 0.7471, "step": 9331 }, { "epoch": 0.3379934806229627, "grad_norm": 1.3409135109558752, "learning_rate": 1.5424383338276375e-05, "loss": 0.7065, "step": 9332 }, { "epoch": 0.33802969938428107, "grad_norm": 1.256653506444496, "learning_rate": 1.5423397811902988e-05, "loss": 0.7219, "step": 9333 }, { "epoch": 0.33806591814559944, "grad_norm": 1.269551812335532, "learning_rate": 1.542241221089889e-05, "loss": 0.7357, "step": 9334 }, { "epoch": 0.33810213690691776, "grad_norm": 1.2690429349916166, "learning_rate": 1.5421426535277628e-05, "loss": 0.6926, "step": 9335 }, { "epoch": 0.33813835566823613, "grad_norm": 1.0419538768739105, "learning_rate": 1.542044078505278e-05, "loss": 0.7399, "step": 9336 }, { "epoch": 0.3381745744295545, "grad_norm": 1.0872696879972192, "learning_rate": 1.54194549602379e-05, "loss": 0.7464, "step": 9337 }, { "epoch": 0.3382107931908729, "grad_norm": 1.3593941613880267, "learning_rate": 1.5418469060846558e-05, "loss": 0.7706, "step": 9338 }, { "epoch": 0.33824701195219126, "grad_norm": 1.4844288477416416, "learning_rate": 1.5417483086892322e-05, "loss": 0.7369, "step": 9339 }, { "epoch": 0.3382832307135096, "grad_norm": 1.1151512043687164, "learning_rate": 1.541649703838876e-05, "loss": 0.7611, "step": 9340 }, { "epoch": 0.33831944947482795, "grad_norm": 1.5246907336854976, "learning_rate": 1.5415510915349433e-05, "loss": 0.7922, "step": 9341 }, { "epoch": 0.3383556682361463, "grad_norm": 1.455156389714228, "learning_rate": 1.5414524717787925e-05, "loss": 0.7717, "step": 9342 }, { "epoch": 0.3383918869974647, "grad_norm": 1.667232773174161, "learning_rate": 1.5413538445717795e-05, "loss": 0.8748, "step": 9343 }, { "epoch": 0.33842810575878307, "grad_norm": 1.5505065963018023, "learning_rate": 1.5412552099152617e-05, "loss": 0.7183, "step": 9344 }, { "epoch": 0.3384643245201014, "grad_norm": 1.3395310278860852, "learning_rate": 1.5411565678105972e-05, "loss": 0.7487, "step": 9345 }, { "epoch": 0.33850054328141976, "grad_norm": 1.0724263381563355, "learning_rate": 1.5410579182591427e-05, "loss": 0.7753, "step": 9346 }, { "epoch": 0.33853676204273814, "grad_norm": 1.121112027501104, "learning_rate": 1.5409592612622556e-05, "loss": 0.6895, "step": 9347 }, { "epoch": 0.3385729808040565, "grad_norm": 1.4322471660867968, "learning_rate": 1.5408605968212937e-05, "loss": 0.7962, "step": 9348 }, { "epoch": 0.3386091995653749, "grad_norm": 1.4527785275266447, "learning_rate": 1.540761924937615e-05, "loss": 0.8507, "step": 9349 }, { "epoch": 0.3386454183266932, "grad_norm": 1.4678803599759684, "learning_rate": 1.540663245612577e-05, "loss": 0.7997, "step": 9350 }, { "epoch": 0.3386816370880116, "grad_norm": 1.5846225503045492, "learning_rate": 1.5405645588475375e-05, "loss": 0.7963, "step": 9351 }, { "epoch": 0.33871785584932995, "grad_norm": 1.377720960979525, "learning_rate": 1.5404658646438547e-05, "loss": 0.8071, "step": 9352 }, { "epoch": 0.3387540746106483, "grad_norm": 1.3270962640332855, "learning_rate": 1.540367163002887e-05, "loss": 0.8233, "step": 9353 }, { "epoch": 0.3387902933719667, "grad_norm": 1.6106035210091685, "learning_rate": 1.5402684539259918e-05, "loss": 0.7606, "step": 9354 }, { "epoch": 0.338826512133285, "grad_norm": 1.2206888699652896, "learning_rate": 1.5401697374145285e-05, "loss": 0.8631, "step": 9355 }, { "epoch": 0.3388627308946034, "grad_norm": 1.5473373942979916, "learning_rate": 1.5400710134698548e-05, "loss": 0.7979, "step": 9356 }, { "epoch": 0.33889894965592177, "grad_norm": 1.1136641475519116, "learning_rate": 1.5399722820933295e-05, "loss": 0.7775, "step": 9357 }, { "epoch": 0.33893516841724014, "grad_norm": 1.3801161514094356, "learning_rate": 1.5398735432863107e-05, "loss": 0.7088, "step": 9358 }, { "epoch": 0.3389713871785585, "grad_norm": 1.3863326609782642, "learning_rate": 1.539774797050158e-05, "loss": 0.7841, "step": 9359 }, { "epoch": 0.33900760593987683, "grad_norm": 1.1616568998804924, "learning_rate": 1.5396760433862295e-05, "loss": 0.7316, "step": 9360 }, { "epoch": 0.3390438247011952, "grad_norm": 1.3653380677952165, "learning_rate": 1.5395772822958844e-05, "loss": 0.83, "step": 9361 }, { "epoch": 0.3390800434625136, "grad_norm": 1.3312382665920757, "learning_rate": 1.5394785137804823e-05, "loss": 0.7758, "step": 9362 }, { "epoch": 0.33911626222383195, "grad_norm": 1.3658865542872236, "learning_rate": 1.5393797378413816e-05, "loss": 0.7827, "step": 9363 }, { "epoch": 0.33915248098515033, "grad_norm": 1.2513186506685658, "learning_rate": 1.5392809544799415e-05, "loss": 0.7551, "step": 9364 }, { "epoch": 0.33918869974646865, "grad_norm": 2.5174328123583773, "learning_rate": 1.539182163697522e-05, "loss": 0.8269, "step": 9365 }, { "epoch": 0.339224918507787, "grad_norm": 1.314920134855711, "learning_rate": 1.5390833654954818e-05, "loss": 0.8002, "step": 9366 }, { "epoch": 0.3392611372691054, "grad_norm": 1.3745751889686058, "learning_rate": 1.538984559875181e-05, "loss": 0.811, "step": 9367 }, { "epoch": 0.33929735603042377, "grad_norm": 1.457221792791271, "learning_rate": 1.538885746837979e-05, "loss": 0.7208, "step": 9368 }, { "epoch": 0.33933357479174214, "grad_norm": 1.2921431482803538, "learning_rate": 1.5387869263852356e-05, "loss": 0.7569, "step": 9369 }, { "epoch": 0.33936979355306046, "grad_norm": 1.3239887236093073, "learning_rate": 1.5386880985183108e-05, "loss": 0.7966, "step": 9370 }, { "epoch": 0.33940601231437884, "grad_norm": 1.272479813597489, "learning_rate": 1.5385892632385643e-05, "loss": 0.761, "step": 9371 }, { "epoch": 0.3394422310756972, "grad_norm": 1.2706021604464843, "learning_rate": 1.5384904205473564e-05, "loss": 0.7892, "step": 9372 }, { "epoch": 0.3394784498370156, "grad_norm": 1.4550130675960464, "learning_rate": 1.538391570446047e-05, "loss": 0.7774, "step": 9373 }, { "epoch": 0.33951466859833396, "grad_norm": 1.4787764741187606, "learning_rate": 1.5382927129359968e-05, "loss": 0.8182, "step": 9374 }, { "epoch": 0.3395508873596523, "grad_norm": 1.108806218639806, "learning_rate": 1.5381938480185658e-05, "loss": 0.8232, "step": 9375 }, { "epoch": 0.33958710612097065, "grad_norm": 1.136061029926366, "learning_rate": 1.5380949756951144e-05, "loss": 0.7055, "step": 9376 }, { "epoch": 0.339623324882289, "grad_norm": 1.613858653044476, "learning_rate": 1.5379960959670036e-05, "loss": 0.8181, "step": 9377 }, { "epoch": 0.3396595436436074, "grad_norm": 1.1936885948156783, "learning_rate": 1.5378972088355934e-05, "loss": 0.831, "step": 9378 }, { "epoch": 0.33969576240492577, "grad_norm": 1.3623590520040303, "learning_rate": 1.5377983143022453e-05, "loss": 0.7104, "step": 9379 }, { "epoch": 0.3397319811662441, "grad_norm": 1.3609704145596653, "learning_rate": 1.5376994123683194e-05, "loss": 0.7566, "step": 9380 }, { "epoch": 0.33976819992756246, "grad_norm": 1.5440533966316459, "learning_rate": 1.5376005030351776e-05, "loss": 0.7967, "step": 9381 }, { "epoch": 0.33980441868888084, "grad_norm": 1.455299827867952, "learning_rate": 1.53750158630418e-05, "loss": 0.7798, "step": 9382 }, { "epoch": 0.3398406374501992, "grad_norm": 1.2970934168316395, "learning_rate": 1.5374026621766886e-05, "loss": 0.7498, "step": 9383 }, { "epoch": 0.3398768562115176, "grad_norm": 1.439144545286209, "learning_rate": 1.5373037306540646e-05, "loss": 0.8504, "step": 9384 }, { "epoch": 0.3399130749728359, "grad_norm": 1.3328513553609078, "learning_rate": 1.537204791737669e-05, "loss": 0.742, "step": 9385 }, { "epoch": 0.3399492937341543, "grad_norm": 1.2552472316451297, "learning_rate": 1.5371058454288633e-05, "loss": 0.7551, "step": 9386 }, { "epoch": 0.33998551249547265, "grad_norm": 1.4910799492324553, "learning_rate": 1.537006891729009e-05, "loss": 0.6902, "step": 9387 }, { "epoch": 0.340021731256791, "grad_norm": 1.367138459824738, "learning_rate": 1.5369079306394684e-05, "loss": 0.7574, "step": 9388 }, { "epoch": 0.3400579500181094, "grad_norm": 1.054268874594335, "learning_rate": 1.5368089621616026e-05, "loss": 0.7187, "step": 9389 }, { "epoch": 0.3400941687794277, "grad_norm": 1.2548224905930385, "learning_rate": 1.5367099862967742e-05, "loss": 0.6819, "step": 9390 }, { "epoch": 0.3401303875407461, "grad_norm": 1.2950069768481351, "learning_rate": 1.5366110030463444e-05, "loss": 0.7834, "step": 9391 }, { "epoch": 0.34016660630206447, "grad_norm": 1.4087852426940166, "learning_rate": 1.536512012411676e-05, "loss": 0.7908, "step": 9392 }, { "epoch": 0.34020282506338284, "grad_norm": 1.4199778107666956, "learning_rate": 1.536413014394131e-05, "loss": 0.8027, "step": 9393 }, { "epoch": 0.3402390438247012, "grad_norm": 1.2582168099595656, "learning_rate": 1.536314008995071e-05, "loss": 0.763, "step": 9394 }, { "epoch": 0.34027526258601953, "grad_norm": 1.1429339237486482, "learning_rate": 1.5362149962158593e-05, "loss": 0.7527, "step": 9395 }, { "epoch": 0.3403114813473379, "grad_norm": 1.7189384309745577, "learning_rate": 1.5361159760578584e-05, "loss": 0.7773, "step": 9396 }, { "epoch": 0.3403477001086563, "grad_norm": 1.4617996890954477, "learning_rate": 1.5360169485224302e-05, "loss": 0.7825, "step": 9397 }, { "epoch": 0.34038391886997466, "grad_norm": 1.2250095585577403, "learning_rate": 1.535917913610938e-05, "loss": 0.7671, "step": 9398 }, { "epoch": 0.34042013763129303, "grad_norm": 1.2170794925710173, "learning_rate": 1.5358188713247443e-05, "loss": 0.8321, "step": 9399 }, { "epoch": 0.34045635639261135, "grad_norm": 1.4540258481390764, "learning_rate": 1.5357198216652123e-05, "loss": 0.7971, "step": 9400 }, { "epoch": 0.3404925751539297, "grad_norm": 1.3396072123294218, "learning_rate": 1.5356207646337045e-05, "loss": 0.7744, "step": 9401 }, { "epoch": 0.3405287939152481, "grad_norm": 1.297626473689779, "learning_rate": 1.5355217002315844e-05, "loss": 0.7389, "step": 9402 }, { "epoch": 0.34056501267656647, "grad_norm": 1.191983471833402, "learning_rate": 1.5354226284602153e-05, "loss": 0.8009, "step": 9403 }, { "epoch": 0.34060123143788484, "grad_norm": 1.2497515606605984, "learning_rate": 1.5353235493209606e-05, "loss": 0.7779, "step": 9404 }, { "epoch": 0.34063745019920316, "grad_norm": 1.3916068127181749, "learning_rate": 1.535224462815183e-05, "loss": 0.8172, "step": 9405 }, { "epoch": 0.34067366896052154, "grad_norm": 1.2643272507537928, "learning_rate": 1.535125368944247e-05, "loss": 0.761, "step": 9406 }, { "epoch": 0.3407098877218399, "grad_norm": 1.138013824298819, "learning_rate": 1.5350262677095153e-05, "loss": 0.7697, "step": 9407 }, { "epoch": 0.3407461064831583, "grad_norm": 1.384588004209213, "learning_rate": 1.5349271591123524e-05, "loss": 0.694, "step": 9408 }, { "epoch": 0.34078232524447666, "grad_norm": 1.4357939565907691, "learning_rate": 1.5348280431541215e-05, "loss": 0.7523, "step": 9409 }, { "epoch": 0.340818544005795, "grad_norm": 1.4977060810668161, "learning_rate": 1.5347289198361866e-05, "loss": 0.8073, "step": 9410 }, { "epoch": 0.34085476276711335, "grad_norm": 1.5226074372155072, "learning_rate": 1.5346297891599126e-05, "loss": 0.7536, "step": 9411 }, { "epoch": 0.3408909815284317, "grad_norm": 1.2920874303195946, "learning_rate": 1.5345306511266624e-05, "loss": 0.7918, "step": 9412 }, { "epoch": 0.3409272002897501, "grad_norm": 1.2994569468047843, "learning_rate": 1.5344315057378008e-05, "loss": 0.7989, "step": 9413 }, { "epoch": 0.3409634190510685, "grad_norm": 1.4161361432582842, "learning_rate": 1.5343323529946918e-05, "loss": 0.7165, "step": 9414 }, { "epoch": 0.3409996378123868, "grad_norm": 1.3663205770163773, "learning_rate": 1.5342331928987005e-05, "loss": 0.7974, "step": 9415 }, { "epoch": 0.34103585657370517, "grad_norm": 1.2668847243260917, "learning_rate": 1.534134025451191e-05, "loss": 0.8631, "step": 9416 }, { "epoch": 0.34107207533502354, "grad_norm": 1.4775946928747425, "learning_rate": 1.534034850653528e-05, "loss": 0.7799, "step": 9417 }, { "epoch": 0.3411082940963419, "grad_norm": 1.3536279332953374, "learning_rate": 1.533935668507076e-05, "loss": 0.76, "step": 9418 }, { "epoch": 0.3411445128576603, "grad_norm": 1.2272009232089034, "learning_rate": 1.533836479013201e-05, "loss": 0.7607, "step": 9419 }, { "epoch": 0.3411807316189786, "grad_norm": 1.413384630735538, "learning_rate": 1.5337372821732656e-05, "loss": 0.8603, "step": 9420 }, { "epoch": 0.341216950380297, "grad_norm": 1.5284422987252486, "learning_rate": 1.5336380779886372e-05, "loss": 0.8065, "step": 9421 }, { "epoch": 0.34125316914161535, "grad_norm": 1.5366362851885649, "learning_rate": 1.5335388664606793e-05, "loss": 0.7595, "step": 9422 }, { "epoch": 0.3412893879029337, "grad_norm": 1.1072259137523, "learning_rate": 1.5334396475907585e-05, "loss": 0.7743, "step": 9423 }, { "epoch": 0.3413256066642521, "grad_norm": 1.330138458586449, "learning_rate": 1.533340421380239e-05, "loss": 0.7913, "step": 9424 }, { "epoch": 0.3413618254255704, "grad_norm": 1.4684172999187421, "learning_rate": 1.5332411878304864e-05, "loss": 0.7508, "step": 9425 }, { "epoch": 0.3413980441868888, "grad_norm": 1.069800666521973, "learning_rate": 1.533141946942867e-05, "loss": 0.746, "step": 9426 }, { "epoch": 0.34143426294820717, "grad_norm": 1.3564615241799107, "learning_rate": 1.5330426987187456e-05, "loss": 0.7848, "step": 9427 }, { "epoch": 0.34147048170952554, "grad_norm": 1.592884712378522, "learning_rate": 1.5329434431594883e-05, "loss": 0.7633, "step": 9428 }, { "epoch": 0.3415067004708439, "grad_norm": 1.591726913479778, "learning_rate": 1.5328441802664612e-05, "loss": 0.8159, "step": 9429 }, { "epoch": 0.34154291923216223, "grad_norm": 1.4699035537960683, "learning_rate": 1.53274491004103e-05, "loss": 0.8765, "step": 9430 }, { "epoch": 0.3415791379934806, "grad_norm": 1.4548387865069534, "learning_rate": 1.5326456324845605e-05, "loss": 0.8097, "step": 9431 }, { "epoch": 0.341615356754799, "grad_norm": 1.3939511751594422, "learning_rate": 1.532546347598419e-05, "loss": 0.7798, "step": 9432 }, { "epoch": 0.34165157551611736, "grad_norm": 1.1879099495898662, "learning_rate": 1.532447055383972e-05, "loss": 0.7594, "step": 9433 }, { "epoch": 0.34168779427743573, "grad_norm": 1.2604704378581826, "learning_rate": 1.5323477558425858e-05, "loss": 0.7737, "step": 9434 }, { "epoch": 0.34172401303875405, "grad_norm": 1.3258901810951622, "learning_rate": 1.5322484489756267e-05, "loss": 0.743, "step": 9435 }, { "epoch": 0.3417602318000724, "grad_norm": 1.0866699515501048, "learning_rate": 1.5321491347844608e-05, "loss": 0.7689, "step": 9436 }, { "epoch": 0.3417964505613908, "grad_norm": 1.152050955396398, "learning_rate": 1.5320498132704558e-05, "loss": 0.6934, "step": 9437 }, { "epoch": 0.34183266932270917, "grad_norm": 1.3084745463124479, "learning_rate": 1.5319504844349773e-05, "loss": 0.6969, "step": 9438 }, { "epoch": 0.34186888808402754, "grad_norm": 1.3547545758529287, "learning_rate": 1.531851148279393e-05, "loss": 0.8263, "step": 9439 }, { "epoch": 0.34190510684534586, "grad_norm": 1.3043283215102464, "learning_rate": 1.5317518048050698e-05, "loss": 0.7159, "step": 9440 }, { "epoch": 0.34194132560666424, "grad_norm": 1.4323134344065194, "learning_rate": 1.531652454013374e-05, "loss": 0.7525, "step": 9441 }, { "epoch": 0.3419775443679826, "grad_norm": 1.28431956069742, "learning_rate": 1.531553095905674e-05, "loss": 0.7902, "step": 9442 }, { "epoch": 0.342013763129301, "grad_norm": 1.5806163017498196, "learning_rate": 1.5314537304833356e-05, "loss": 0.7875, "step": 9443 }, { "epoch": 0.34204998189061936, "grad_norm": 1.4604302414324655, "learning_rate": 1.5313543577477274e-05, "loss": 0.8261, "step": 9444 }, { "epoch": 0.3420862006519377, "grad_norm": 1.2204536333708909, "learning_rate": 1.531254977700216e-05, "loss": 0.81, "step": 9445 }, { "epoch": 0.34212241941325605, "grad_norm": 1.305866331355229, "learning_rate": 1.5311555903421697e-05, "loss": 0.7258, "step": 9446 }, { "epoch": 0.3421586381745744, "grad_norm": 1.4609720167376583, "learning_rate": 1.5310561956749554e-05, "loss": 0.8074, "step": 9447 }, { "epoch": 0.3421948569358928, "grad_norm": 1.4793898411718691, "learning_rate": 1.5309567936999413e-05, "loss": 0.8259, "step": 9448 }, { "epoch": 0.3422310756972112, "grad_norm": 1.3576080421015693, "learning_rate": 1.5308573844184955e-05, "loss": 0.7472, "step": 9449 }, { "epoch": 0.3422672944585295, "grad_norm": 1.284821482022187, "learning_rate": 1.5307579678319852e-05, "loss": 0.8228, "step": 9450 }, { "epoch": 0.34230351321984787, "grad_norm": 1.3842113445500281, "learning_rate": 1.5306585439417792e-05, "loss": 0.7847, "step": 9451 }, { "epoch": 0.34233973198116624, "grad_norm": 1.4980971975714779, "learning_rate": 1.5305591127492455e-05, "loss": 0.8156, "step": 9452 }, { "epoch": 0.3423759507424846, "grad_norm": 1.2834301545219935, "learning_rate": 1.5304596742557516e-05, "loss": 0.7553, "step": 9453 }, { "epoch": 0.342412169503803, "grad_norm": 1.4679477636350649, "learning_rate": 1.530360228462667e-05, "loss": 0.8552, "step": 9454 }, { "epoch": 0.3424483882651213, "grad_norm": 1.426216278596933, "learning_rate": 1.5302607753713597e-05, "loss": 0.768, "step": 9455 }, { "epoch": 0.3424846070264397, "grad_norm": 1.3095147955304145, "learning_rate": 1.5301613149831978e-05, "loss": 0.7181, "step": 9456 }, { "epoch": 0.34252082578775805, "grad_norm": 1.505265674865553, "learning_rate": 1.5300618472995507e-05, "loss": 0.8233, "step": 9457 }, { "epoch": 0.34255704454907643, "grad_norm": 1.3937569974608168, "learning_rate": 1.5299623723217867e-05, "loss": 0.7421, "step": 9458 }, { "epoch": 0.3425932633103948, "grad_norm": 1.5620464906581468, "learning_rate": 1.5298628900512747e-05, "loss": 0.828, "step": 9459 }, { "epoch": 0.3426294820717131, "grad_norm": 1.3966868087123978, "learning_rate": 1.5297634004893838e-05, "loss": 0.7704, "step": 9460 }, { "epoch": 0.3426657008330315, "grad_norm": 1.6509134389971394, "learning_rate": 1.529663903637483e-05, "loss": 0.8437, "step": 9461 }, { "epoch": 0.34270191959434987, "grad_norm": 1.5144560782143266, "learning_rate": 1.5295643994969417e-05, "loss": 0.8352, "step": 9462 }, { "epoch": 0.34273813835566824, "grad_norm": 1.558746182844089, "learning_rate": 1.5294648880691284e-05, "loss": 0.8406, "step": 9463 }, { "epoch": 0.3427743571169866, "grad_norm": 1.3861676652366355, "learning_rate": 1.529365369355414e-05, "loss": 0.7418, "step": 9464 }, { "epoch": 0.34281057587830494, "grad_norm": 1.7470585186700434, "learning_rate": 1.529265843357166e-05, "loss": 0.8102, "step": 9465 }, { "epoch": 0.3428467946396233, "grad_norm": 1.087751243985682, "learning_rate": 1.5291663100757554e-05, "loss": 0.7553, "step": 9466 }, { "epoch": 0.3428830134009417, "grad_norm": 1.3711519955764402, "learning_rate": 1.5290667695125512e-05, "loss": 0.786, "step": 9467 }, { "epoch": 0.34291923216226006, "grad_norm": 1.2686442758532537, "learning_rate": 1.5289672216689236e-05, "loss": 0.827, "step": 9468 }, { "epoch": 0.34295545092357843, "grad_norm": 1.4021610012094674, "learning_rate": 1.528867666546242e-05, "loss": 0.7542, "step": 9469 }, { "epoch": 0.34299166968489675, "grad_norm": 1.3768853569939923, "learning_rate": 1.5287681041458767e-05, "loss": 0.6992, "step": 9470 }, { "epoch": 0.3430278884462151, "grad_norm": 1.4663657038514069, "learning_rate": 1.5286685344691977e-05, "loss": 0.7911, "step": 9471 }, { "epoch": 0.3430641072075335, "grad_norm": 2.0065562028903785, "learning_rate": 1.528568957517575e-05, "loss": 0.8089, "step": 9472 }, { "epoch": 0.34310032596885187, "grad_norm": 1.8496942530877871, "learning_rate": 1.5284693732923793e-05, "loss": 0.7836, "step": 9473 }, { "epoch": 0.34313654473017025, "grad_norm": 1.3490541229580002, "learning_rate": 1.5283697817949806e-05, "loss": 0.7462, "step": 9474 }, { "epoch": 0.34317276349148856, "grad_norm": 1.4579270947975032, "learning_rate": 1.5282701830267493e-05, "loss": 0.7733, "step": 9475 }, { "epoch": 0.34320898225280694, "grad_norm": 1.3259620251169821, "learning_rate": 1.528170576989056e-05, "loss": 0.8096, "step": 9476 }, { "epoch": 0.3432452010141253, "grad_norm": 1.211916368155319, "learning_rate": 1.5280709636832718e-05, "loss": 0.8439, "step": 9477 }, { "epoch": 0.3432814197754437, "grad_norm": 1.3370180535398006, "learning_rate": 1.527971343110767e-05, "loss": 0.7721, "step": 9478 }, { "epoch": 0.34331763853676206, "grad_norm": 1.3839535706268393, "learning_rate": 1.5278717152729124e-05, "loss": 0.7601, "step": 9479 }, { "epoch": 0.3433538572980804, "grad_norm": 1.3057739686608874, "learning_rate": 1.5277720801710795e-05, "loss": 0.7718, "step": 9480 }, { "epoch": 0.34339007605939875, "grad_norm": 1.2560559581654664, "learning_rate": 1.5276724378066384e-05, "loss": 0.7702, "step": 9481 }, { "epoch": 0.3434262948207171, "grad_norm": 1.2780727425595744, "learning_rate": 1.5275727881809617e-05, "loss": 0.7354, "step": 9482 }, { "epoch": 0.3434625135820355, "grad_norm": 1.3001613359168587, "learning_rate": 1.5274731312954198e-05, "loss": 0.7957, "step": 9483 }, { "epoch": 0.3434987323433539, "grad_norm": 1.4194119763114261, "learning_rate": 1.5273734671513838e-05, "loss": 0.7599, "step": 9484 }, { "epoch": 0.3435349511046722, "grad_norm": 1.58851992886684, "learning_rate": 1.5272737957502254e-05, "loss": 0.903, "step": 9485 }, { "epoch": 0.34357116986599057, "grad_norm": 1.6299394605699333, "learning_rate": 1.5271741170933167e-05, "loss": 0.7613, "step": 9486 }, { "epoch": 0.34360738862730894, "grad_norm": 1.4192386620634625, "learning_rate": 1.527074431182028e-05, "loss": 0.8151, "step": 9487 }, { "epoch": 0.3436436073886273, "grad_norm": 1.3113476876205028, "learning_rate": 1.526974738017733e-05, "loss": 0.8178, "step": 9488 }, { "epoch": 0.3436798261499457, "grad_norm": 1.339651777316406, "learning_rate": 1.5268750376018022e-05, "loss": 0.8223, "step": 9489 }, { "epoch": 0.343716044911264, "grad_norm": 1.3018483289998288, "learning_rate": 1.526775329935608e-05, "loss": 0.8567, "step": 9490 }, { "epoch": 0.3437522636725824, "grad_norm": 1.3911638782922846, "learning_rate": 1.5266756150205224e-05, "loss": 0.7538, "step": 9491 }, { "epoch": 0.34378848243390076, "grad_norm": 1.092740562300696, "learning_rate": 1.5265758928579175e-05, "loss": 0.7395, "step": 9492 }, { "epoch": 0.34382470119521913, "grad_norm": 1.4029885353131801, "learning_rate": 1.5264761634491656e-05, "loss": 0.7623, "step": 9493 }, { "epoch": 0.3438609199565375, "grad_norm": 1.3007952486035304, "learning_rate": 1.5263764267956392e-05, "loss": 0.7755, "step": 9494 }, { "epoch": 0.3438971387178559, "grad_norm": 1.3049173880673917, "learning_rate": 1.526276682898711e-05, "loss": 0.7555, "step": 9495 }, { "epoch": 0.3439333574791742, "grad_norm": 1.1945524877140998, "learning_rate": 1.5261769317597523e-05, "loss": 0.744, "step": 9496 }, { "epoch": 0.34396957624049257, "grad_norm": 1.4110886721983773, "learning_rate": 1.5260771733801375e-05, "loss": 0.7519, "step": 9497 }, { "epoch": 0.34400579500181094, "grad_norm": 1.3441867936288967, "learning_rate": 1.525977407761238e-05, "loss": 0.7253, "step": 9498 }, { "epoch": 0.3440420137631293, "grad_norm": 1.3976506931435415, "learning_rate": 1.5258776349044276e-05, "loss": 0.7841, "step": 9499 }, { "epoch": 0.3440782325244477, "grad_norm": 1.4493259045863034, "learning_rate": 1.5257778548110786e-05, "loss": 0.8206, "step": 9500 }, { "epoch": 0.344114451285766, "grad_norm": 1.1318035237831836, "learning_rate": 1.5256780674825641e-05, "loss": 0.731, "step": 9501 }, { "epoch": 0.3441506700470844, "grad_norm": 1.0624660279320415, "learning_rate": 1.525578272920258e-05, "loss": 0.7308, "step": 9502 }, { "epoch": 0.34418688880840276, "grad_norm": 1.4097621988303497, "learning_rate": 1.5254784711255329e-05, "loss": 0.8367, "step": 9503 }, { "epoch": 0.34422310756972113, "grad_norm": 1.4100145020398605, "learning_rate": 1.525378662099762e-05, "loss": 0.7558, "step": 9504 }, { "epoch": 0.3442593263310395, "grad_norm": 1.3419300541452202, "learning_rate": 1.5252788458443197e-05, "loss": 0.7752, "step": 9505 }, { "epoch": 0.3442955450923578, "grad_norm": 1.5396050015044447, "learning_rate": 1.5251790223605784e-05, "loss": 0.8659, "step": 9506 }, { "epoch": 0.3443317638536762, "grad_norm": 1.3346795135671967, "learning_rate": 1.5250791916499126e-05, "loss": 0.8221, "step": 9507 }, { "epoch": 0.34436798261499457, "grad_norm": 1.1183051517960878, "learning_rate": 1.5249793537136957e-05, "loss": 0.7521, "step": 9508 }, { "epoch": 0.34440420137631295, "grad_norm": 1.1513850176945657, "learning_rate": 1.5248795085533013e-05, "loss": 0.7625, "step": 9509 }, { "epoch": 0.3444404201376313, "grad_norm": 1.5867846706164832, "learning_rate": 1.5247796561701038e-05, "loss": 0.7769, "step": 9510 }, { "epoch": 0.34447663889894964, "grad_norm": 1.4059733569042656, "learning_rate": 1.5246797965654773e-05, "loss": 0.8171, "step": 9511 }, { "epoch": 0.344512857660268, "grad_norm": 1.2792018611218723, "learning_rate": 1.5245799297407957e-05, "loss": 0.6183, "step": 9512 }, { "epoch": 0.3445490764215864, "grad_norm": 1.326741125888305, "learning_rate": 1.5244800556974334e-05, "loss": 0.7849, "step": 9513 }, { "epoch": 0.34458529518290476, "grad_norm": 1.356111806127802, "learning_rate": 1.5243801744367645e-05, "loss": 0.7388, "step": 9514 }, { "epoch": 0.34462151394422313, "grad_norm": 1.4235736792251426, "learning_rate": 1.5242802859601637e-05, "loss": 0.7965, "step": 9515 }, { "epoch": 0.34465773270554145, "grad_norm": 1.5121327161157687, "learning_rate": 1.5241803902690056e-05, "loss": 0.7625, "step": 9516 }, { "epoch": 0.3446939514668598, "grad_norm": 1.3401426310034603, "learning_rate": 1.5240804873646649e-05, "loss": 0.7575, "step": 9517 }, { "epoch": 0.3447301702281782, "grad_norm": 1.7687336469372825, "learning_rate": 1.5239805772485157e-05, "loss": 0.7924, "step": 9518 }, { "epoch": 0.3447663889894966, "grad_norm": 1.1110102925163574, "learning_rate": 1.5238806599219337e-05, "loss": 0.6893, "step": 9519 }, { "epoch": 0.34480260775081495, "grad_norm": 1.1452993511602136, "learning_rate": 1.5237807353862933e-05, "loss": 0.7701, "step": 9520 }, { "epoch": 0.34483882651213327, "grad_norm": 1.3441672574841212, "learning_rate": 1.5236808036429698e-05, "loss": 0.8013, "step": 9521 }, { "epoch": 0.34487504527345164, "grad_norm": 1.4756806732607275, "learning_rate": 1.5235808646933383e-05, "loss": 0.8151, "step": 9522 }, { "epoch": 0.34491126403477, "grad_norm": 1.2991364807670658, "learning_rate": 1.5234809185387743e-05, "loss": 0.8072, "step": 9523 }, { "epoch": 0.3449474827960884, "grad_norm": 1.2606581298676238, "learning_rate": 1.5233809651806528e-05, "loss": 0.6967, "step": 9524 }, { "epoch": 0.34498370155740676, "grad_norm": 1.403374141969433, "learning_rate": 1.523281004620349e-05, "loss": 0.8094, "step": 9525 }, { "epoch": 0.3450199203187251, "grad_norm": 1.3639741983705385, "learning_rate": 1.523181036859239e-05, "loss": 0.7795, "step": 9526 }, { "epoch": 0.34505613908004346, "grad_norm": 1.2222129677852787, "learning_rate": 1.5230810618986982e-05, "loss": 0.733, "step": 9527 }, { "epoch": 0.34509235784136183, "grad_norm": 1.4346901198932192, "learning_rate": 1.5229810797401026e-05, "loss": 0.8177, "step": 9528 }, { "epoch": 0.3451285766026802, "grad_norm": 1.4728533640720105, "learning_rate": 1.5228810903848276e-05, "loss": 0.7849, "step": 9529 }, { "epoch": 0.3451647953639986, "grad_norm": 1.4293668939871849, "learning_rate": 1.5227810938342493e-05, "loss": 0.7125, "step": 9530 }, { "epoch": 0.3452010141253169, "grad_norm": 1.274931028898804, "learning_rate": 1.522681090089744e-05, "loss": 0.765, "step": 9531 }, { "epoch": 0.34523723288663527, "grad_norm": 1.4529973212416472, "learning_rate": 1.5225810791526873e-05, "loss": 0.8172, "step": 9532 }, { "epoch": 0.34527345164795364, "grad_norm": 1.4300085160737117, "learning_rate": 1.5224810610244563e-05, "loss": 0.7535, "step": 9533 }, { "epoch": 0.345309670409272, "grad_norm": 1.3222910936345436, "learning_rate": 1.5223810357064261e-05, "loss": 0.7454, "step": 9534 }, { "epoch": 0.3453458891705904, "grad_norm": 1.2907499012478336, "learning_rate": 1.5222810031999746e-05, "loss": 0.7427, "step": 9535 }, { "epoch": 0.3453821079319087, "grad_norm": 1.4227170330796548, "learning_rate": 1.522180963506477e-05, "loss": 0.7444, "step": 9536 }, { "epoch": 0.3454183266932271, "grad_norm": 1.4657561258269851, "learning_rate": 1.5220809166273104e-05, "loss": 0.8884, "step": 9537 }, { "epoch": 0.34545454545454546, "grad_norm": 1.2668218185869795, "learning_rate": 1.5219808625638521e-05, "loss": 0.8106, "step": 9538 }, { "epoch": 0.34549076421586383, "grad_norm": 1.3113742652641387, "learning_rate": 1.5218808013174784e-05, "loss": 0.7411, "step": 9539 }, { "epoch": 0.3455269829771822, "grad_norm": 1.3041172808776393, "learning_rate": 1.5217807328895663e-05, "loss": 0.7835, "step": 9540 }, { "epoch": 0.3455632017385005, "grad_norm": 1.521911295540245, "learning_rate": 1.5216806572814929e-05, "loss": 0.7434, "step": 9541 }, { "epoch": 0.3455994204998189, "grad_norm": 1.8154016495663876, "learning_rate": 1.521580574494635e-05, "loss": 0.7261, "step": 9542 }, { "epoch": 0.3456356392611373, "grad_norm": 1.3449329780166843, "learning_rate": 1.52148048453037e-05, "loss": 0.7569, "step": 9543 }, { "epoch": 0.34567185802245565, "grad_norm": 1.365599510954228, "learning_rate": 1.521380387390076e-05, "loss": 0.8044, "step": 9544 }, { "epoch": 0.345708076783774, "grad_norm": 1.3008738311549453, "learning_rate": 1.5212802830751292e-05, "loss": 0.7952, "step": 9545 }, { "epoch": 0.34574429554509234, "grad_norm": 1.4987095581870797, "learning_rate": 1.5211801715869077e-05, "loss": 0.8098, "step": 9546 }, { "epoch": 0.3457805143064107, "grad_norm": 1.2591113841460144, "learning_rate": 1.5210800529267893e-05, "loss": 0.7982, "step": 9547 }, { "epoch": 0.3458167330677291, "grad_norm": 1.4932475219514723, "learning_rate": 1.5209799270961512e-05, "loss": 0.8137, "step": 9548 }, { "epoch": 0.34585295182904746, "grad_norm": 1.4295865229087215, "learning_rate": 1.5208797940963718e-05, "loss": 0.8181, "step": 9549 }, { "epoch": 0.34588917059036584, "grad_norm": 0.9712666359653287, "learning_rate": 1.5207796539288286e-05, "loss": 0.695, "step": 9550 }, { "epoch": 0.34592538935168415, "grad_norm": 1.3206784471983521, "learning_rate": 1.5206795065948999e-05, "loss": 0.803, "step": 9551 }, { "epoch": 0.34596160811300253, "grad_norm": 1.42717921165238, "learning_rate": 1.5205793520959634e-05, "loss": 0.8014, "step": 9552 }, { "epoch": 0.3459978268743209, "grad_norm": 1.440961342665826, "learning_rate": 1.520479190433398e-05, "loss": 0.7456, "step": 9553 }, { "epoch": 0.3460340456356393, "grad_norm": 1.4360105712819962, "learning_rate": 1.5203790216085814e-05, "loss": 0.856, "step": 9554 }, { "epoch": 0.34607026439695765, "grad_norm": 1.4085645145803516, "learning_rate": 1.520278845622892e-05, "loss": 0.7006, "step": 9555 }, { "epoch": 0.34610648315827597, "grad_norm": 1.4318553007692698, "learning_rate": 1.5201786624777088e-05, "loss": 0.7815, "step": 9556 }, { "epoch": 0.34614270191959434, "grad_norm": 1.4142811988742938, "learning_rate": 1.5200784721744103e-05, "loss": 0.8324, "step": 9557 }, { "epoch": 0.3461789206809127, "grad_norm": 1.386924375144653, "learning_rate": 1.5199782747143745e-05, "loss": 0.8142, "step": 9558 }, { "epoch": 0.3462151394422311, "grad_norm": 1.8449597681445093, "learning_rate": 1.5198780700989813e-05, "loss": 0.7908, "step": 9559 }, { "epoch": 0.34625135820354946, "grad_norm": 1.5551785647189205, "learning_rate": 1.5197778583296088e-05, "loss": 0.8697, "step": 9560 }, { "epoch": 0.3462875769648678, "grad_norm": 1.3792586997341891, "learning_rate": 1.519677639407636e-05, "loss": 0.7789, "step": 9561 }, { "epoch": 0.34632379572618616, "grad_norm": 1.0868206765392352, "learning_rate": 1.5195774133344428e-05, "loss": 0.6839, "step": 9562 }, { "epoch": 0.34636001448750453, "grad_norm": 1.4518098686612688, "learning_rate": 1.5194771801114074e-05, "loss": 0.6858, "step": 9563 }, { "epoch": 0.3463962332488229, "grad_norm": 1.4529301560169503, "learning_rate": 1.5193769397399099e-05, "loss": 0.7851, "step": 9564 }, { "epoch": 0.3464324520101413, "grad_norm": 1.2205517736066949, "learning_rate": 1.519276692221329e-05, "loss": 0.7323, "step": 9565 }, { "epoch": 0.3464686707714596, "grad_norm": 1.3974541857894205, "learning_rate": 1.5191764375570449e-05, "loss": 0.8407, "step": 9566 }, { "epoch": 0.34650488953277797, "grad_norm": 1.207772046135902, "learning_rate": 1.5190761757484365e-05, "loss": 0.8157, "step": 9567 }, { "epoch": 0.34654110829409634, "grad_norm": 1.3942352900479993, "learning_rate": 1.5189759067968838e-05, "loss": 0.8385, "step": 9568 }, { "epoch": 0.3465773270554147, "grad_norm": 1.0872553889128431, "learning_rate": 1.518875630703767e-05, "loss": 0.8222, "step": 9569 }, { "epoch": 0.3466135458167331, "grad_norm": 1.4719513754355684, "learning_rate": 1.5187753474704652e-05, "loss": 0.7247, "step": 9570 }, { "epoch": 0.3466497645780514, "grad_norm": 1.3659333239452032, "learning_rate": 1.518675057098359e-05, "loss": 0.7604, "step": 9571 }, { "epoch": 0.3466859833393698, "grad_norm": 1.2269069989751897, "learning_rate": 1.5185747595888281e-05, "loss": 0.7621, "step": 9572 }, { "epoch": 0.34672220210068816, "grad_norm": 1.324718262975978, "learning_rate": 1.5184744549432533e-05, "loss": 0.8148, "step": 9573 }, { "epoch": 0.34675842086200653, "grad_norm": 1.486770275258471, "learning_rate": 1.5183741431630138e-05, "loss": 0.721, "step": 9574 }, { "epoch": 0.3467946396233249, "grad_norm": 1.48745479690365, "learning_rate": 1.5182738242494913e-05, "loss": 0.725, "step": 9575 }, { "epoch": 0.3468308583846432, "grad_norm": 1.438444484241659, "learning_rate": 1.5181734982040648e-05, "loss": 0.8356, "step": 9576 }, { "epoch": 0.3468670771459616, "grad_norm": 1.266828837184117, "learning_rate": 1.5180731650281161e-05, "loss": 0.7911, "step": 9577 }, { "epoch": 0.34690329590728, "grad_norm": 1.3119700304172366, "learning_rate": 1.5179728247230255e-05, "loss": 0.7747, "step": 9578 }, { "epoch": 0.34693951466859835, "grad_norm": 1.3935412063427735, "learning_rate": 1.5178724772901733e-05, "loss": 0.7973, "step": 9579 }, { "epoch": 0.3469757334299167, "grad_norm": 1.1909463997528742, "learning_rate": 1.5177721227309412e-05, "loss": 0.74, "step": 9580 }, { "epoch": 0.34701195219123504, "grad_norm": 1.3626995632126657, "learning_rate": 1.5176717610467096e-05, "loss": 0.8202, "step": 9581 }, { "epoch": 0.3470481709525534, "grad_norm": 1.4560584433473263, "learning_rate": 1.5175713922388598e-05, "loss": 0.7483, "step": 9582 }, { "epoch": 0.3470843897138718, "grad_norm": 1.541483566299821, "learning_rate": 1.5174710163087727e-05, "loss": 0.8397, "step": 9583 }, { "epoch": 0.34712060847519016, "grad_norm": 1.236341768644479, "learning_rate": 1.5173706332578298e-05, "loss": 0.6869, "step": 9584 }, { "epoch": 0.34715682723650854, "grad_norm": 1.1169270104634448, "learning_rate": 1.5172702430874122e-05, "loss": 0.7908, "step": 9585 }, { "epoch": 0.34719304599782685, "grad_norm": 1.9169913013551425, "learning_rate": 1.5171698457989017e-05, "loss": 0.8118, "step": 9586 }, { "epoch": 0.34722926475914523, "grad_norm": 1.3833316142439454, "learning_rate": 1.5170694413936798e-05, "loss": 0.7985, "step": 9587 }, { "epoch": 0.3472654835204636, "grad_norm": 1.4932474084400067, "learning_rate": 1.5169690298731277e-05, "loss": 0.9062, "step": 9588 }, { "epoch": 0.347301702281782, "grad_norm": 1.4516076862130793, "learning_rate": 1.5168686112386279e-05, "loss": 0.7689, "step": 9589 }, { "epoch": 0.34733792104310035, "grad_norm": 1.315838010549185, "learning_rate": 1.5167681854915617e-05, "loss": 0.8032, "step": 9590 }, { "epoch": 0.34737413980441867, "grad_norm": 1.3568074029404316, "learning_rate": 1.5166677526333113e-05, "loss": 0.7597, "step": 9591 }, { "epoch": 0.34741035856573704, "grad_norm": 1.3292009427196279, "learning_rate": 1.5165673126652585e-05, "loss": 0.8216, "step": 9592 }, { "epoch": 0.3474465773270554, "grad_norm": 1.0915894526715815, "learning_rate": 1.5164668655887855e-05, "loss": 0.7553, "step": 9593 }, { "epoch": 0.3474827960883738, "grad_norm": 1.5113896614250049, "learning_rate": 1.5163664114052748e-05, "loss": 0.7878, "step": 9594 }, { "epoch": 0.34751901484969216, "grad_norm": 1.1055622007623427, "learning_rate": 1.5162659501161083e-05, "loss": 0.7795, "step": 9595 }, { "epoch": 0.3475552336110105, "grad_norm": 1.4205342261875702, "learning_rate": 1.516165481722669e-05, "loss": 0.7816, "step": 9596 }, { "epoch": 0.34759145237232886, "grad_norm": 1.527444345912735, "learning_rate": 1.5160650062263392e-05, "loss": 0.8132, "step": 9597 }, { "epoch": 0.34762767113364723, "grad_norm": 1.3496912974158228, "learning_rate": 1.515964523628501e-05, "loss": 0.8673, "step": 9598 }, { "epoch": 0.3476638898949656, "grad_norm": 1.7023295942455867, "learning_rate": 1.5158640339305376e-05, "loss": 0.8407, "step": 9599 }, { "epoch": 0.347700108656284, "grad_norm": 1.3246329012980533, "learning_rate": 1.5157635371338324e-05, "loss": 0.7466, "step": 9600 }, { "epoch": 0.3477363274176023, "grad_norm": 1.0028752943200554, "learning_rate": 1.5156630332397675e-05, "loss": 0.7075, "step": 9601 }, { "epoch": 0.34777254617892067, "grad_norm": 1.206367921114741, "learning_rate": 1.515562522249726e-05, "loss": 0.7838, "step": 9602 }, { "epoch": 0.34780876494023905, "grad_norm": 1.187261928467924, "learning_rate": 1.5154620041650914e-05, "loss": 0.8089, "step": 9603 }, { "epoch": 0.3478449837015574, "grad_norm": 1.5297920599476353, "learning_rate": 1.5153614789872464e-05, "loss": 0.8115, "step": 9604 }, { "epoch": 0.3478812024628758, "grad_norm": 1.1368104749799606, "learning_rate": 1.5152609467175748e-05, "loss": 0.8152, "step": 9605 }, { "epoch": 0.3479174212241941, "grad_norm": 1.4093306225001063, "learning_rate": 1.5151604073574598e-05, "loss": 0.7451, "step": 9606 }, { "epoch": 0.3479536399855125, "grad_norm": 1.5586470210023202, "learning_rate": 1.5150598609082849e-05, "loss": 0.7612, "step": 9607 }, { "epoch": 0.34798985874683086, "grad_norm": 1.256884555424656, "learning_rate": 1.5149593073714339e-05, "loss": 0.8244, "step": 9608 }, { "epoch": 0.34802607750814923, "grad_norm": 1.5580172225363833, "learning_rate": 1.5148587467482903e-05, "loss": 0.8371, "step": 9609 }, { "epoch": 0.3480622962694676, "grad_norm": 1.6066171293351543, "learning_rate": 1.5147581790402374e-05, "loss": 0.7494, "step": 9610 }, { "epoch": 0.3480985150307859, "grad_norm": 1.568436275131316, "learning_rate": 1.5146576042486604e-05, "loss": 0.7811, "step": 9611 }, { "epoch": 0.3481347337921043, "grad_norm": 1.2905576545772717, "learning_rate": 1.5145570223749424e-05, "loss": 0.7894, "step": 9612 }, { "epoch": 0.3481709525534227, "grad_norm": 1.3087542798376288, "learning_rate": 1.5144564334204675e-05, "loss": 0.8129, "step": 9613 }, { "epoch": 0.34820717131474105, "grad_norm": 1.4787257699017218, "learning_rate": 1.5143558373866198e-05, "loss": 0.8161, "step": 9614 }, { "epoch": 0.3482433900760594, "grad_norm": 1.6523087364797755, "learning_rate": 1.5142552342747843e-05, "loss": 0.7976, "step": 9615 }, { "epoch": 0.34827960883737774, "grad_norm": 1.4185513082613972, "learning_rate": 1.5141546240863446e-05, "loss": 0.825, "step": 9616 }, { "epoch": 0.3483158275986961, "grad_norm": 1.5079463679425684, "learning_rate": 1.5140540068226857e-05, "loss": 0.7845, "step": 9617 }, { "epoch": 0.3483520463600145, "grad_norm": 1.4839133548636776, "learning_rate": 1.5139533824851917e-05, "loss": 0.8079, "step": 9618 }, { "epoch": 0.34838826512133286, "grad_norm": 1.435566109249466, "learning_rate": 1.5138527510752476e-05, "loss": 0.7585, "step": 9619 }, { "epoch": 0.34842448388265124, "grad_norm": 1.3480904536193576, "learning_rate": 1.5137521125942384e-05, "loss": 0.7164, "step": 9620 }, { "epoch": 0.34846070264396956, "grad_norm": 1.5206622037371804, "learning_rate": 1.5136514670435485e-05, "loss": 0.8384, "step": 9621 }, { "epoch": 0.34849692140528793, "grad_norm": 1.4077071772858507, "learning_rate": 1.5135508144245635e-05, "loss": 0.769, "step": 9622 }, { "epoch": 0.3485331401666063, "grad_norm": 1.32988419078207, "learning_rate": 1.5134501547386675e-05, "loss": 0.761, "step": 9623 }, { "epoch": 0.3485693589279247, "grad_norm": 1.4678751402014876, "learning_rate": 1.5133494879872464e-05, "loss": 0.7492, "step": 9624 }, { "epoch": 0.34860557768924305, "grad_norm": 1.3706016431754404, "learning_rate": 1.5132488141716852e-05, "loss": 0.8175, "step": 9625 }, { "epoch": 0.34864179645056137, "grad_norm": 1.537899850154432, "learning_rate": 1.5131481332933695e-05, "loss": 0.7891, "step": 9626 }, { "epoch": 0.34867801521187974, "grad_norm": 1.1491901563277587, "learning_rate": 1.5130474453536845e-05, "loss": 0.7465, "step": 9627 }, { "epoch": 0.3487142339731981, "grad_norm": 1.3921108735983794, "learning_rate": 1.5129467503540158e-05, "loss": 0.8879, "step": 9628 }, { "epoch": 0.3487504527345165, "grad_norm": 1.1221892714625872, "learning_rate": 1.512846048295749e-05, "loss": 0.8271, "step": 9629 }, { "epoch": 0.34878667149583487, "grad_norm": 1.2057996676929919, "learning_rate": 1.51274533918027e-05, "loss": 0.7471, "step": 9630 }, { "epoch": 0.3488228902571532, "grad_norm": 1.3273131483770402, "learning_rate": 1.5126446230089648e-05, "loss": 0.6785, "step": 9631 }, { "epoch": 0.34885910901847156, "grad_norm": 1.0927089085864254, "learning_rate": 1.5125438997832191e-05, "loss": 0.715, "step": 9632 }, { "epoch": 0.34889532777978993, "grad_norm": 1.2040264238359277, "learning_rate": 1.5124431695044193e-05, "loss": 0.7909, "step": 9633 }, { "epoch": 0.3489315465411083, "grad_norm": 1.3955383637796641, "learning_rate": 1.5123424321739508e-05, "loss": 0.7666, "step": 9634 }, { "epoch": 0.3489677653024267, "grad_norm": 1.3496871110193973, "learning_rate": 1.5122416877932003e-05, "loss": 0.743, "step": 9635 }, { "epoch": 0.349003984063745, "grad_norm": 1.3259242280236965, "learning_rate": 1.5121409363635539e-05, "loss": 0.7421, "step": 9636 }, { "epoch": 0.3490402028250634, "grad_norm": 1.1825876158927846, "learning_rate": 1.5120401778863984e-05, "loss": 0.7951, "step": 9637 }, { "epoch": 0.34907642158638175, "grad_norm": 1.07591463582631, "learning_rate": 1.5119394123631201e-05, "loss": 0.7342, "step": 9638 }, { "epoch": 0.3491126403477001, "grad_norm": 1.4345392557499772, "learning_rate": 1.5118386397951057e-05, "loss": 0.7636, "step": 9639 }, { "epoch": 0.3491488591090185, "grad_norm": 1.3557588984045126, "learning_rate": 1.5117378601837418e-05, "loss": 0.7513, "step": 9640 }, { "epoch": 0.3491850778703368, "grad_norm": 1.264203065340668, "learning_rate": 1.5116370735304153e-05, "loss": 0.8056, "step": 9641 }, { "epoch": 0.3492212966316552, "grad_norm": 1.494892495287527, "learning_rate": 1.5115362798365131e-05, "loss": 0.8634, "step": 9642 }, { "epoch": 0.34925751539297356, "grad_norm": 1.4634901506604383, "learning_rate": 1.5114354791034225e-05, "loss": 0.781, "step": 9643 }, { "epoch": 0.34929373415429193, "grad_norm": 1.1307509804056883, "learning_rate": 1.51133467133253e-05, "loss": 0.8067, "step": 9644 }, { "epoch": 0.3493299529156103, "grad_norm": 1.1990041665061688, "learning_rate": 1.5112338565252231e-05, "loss": 0.6799, "step": 9645 }, { "epoch": 0.3493661716769286, "grad_norm": 1.3789189003102058, "learning_rate": 1.5111330346828891e-05, "loss": 0.7695, "step": 9646 }, { "epoch": 0.349402390438247, "grad_norm": 1.3310708977656074, "learning_rate": 1.5110322058069157e-05, "loss": 0.7808, "step": 9647 }, { "epoch": 0.3494386091995654, "grad_norm": 1.4312433703864529, "learning_rate": 1.5109313698986899e-05, "loss": 0.8179, "step": 9648 }, { "epoch": 0.34947482796088375, "grad_norm": 1.4290808063974125, "learning_rate": 1.5108305269595997e-05, "loss": 0.7967, "step": 9649 }, { "epoch": 0.3495110467222021, "grad_norm": 1.3096833567859847, "learning_rate": 1.5107296769910322e-05, "loss": 0.8037, "step": 9650 }, { "epoch": 0.34954726548352044, "grad_norm": 1.5421815638732295, "learning_rate": 1.5106288199943762e-05, "loss": 0.823, "step": 9651 }, { "epoch": 0.3495834842448388, "grad_norm": 1.3894060148618297, "learning_rate": 1.5105279559710185e-05, "loss": 0.7849, "step": 9652 }, { "epoch": 0.3496197030061572, "grad_norm": 1.419051283878355, "learning_rate": 1.5104270849223476e-05, "loss": 0.7338, "step": 9653 }, { "epoch": 0.34965592176747556, "grad_norm": 1.1644896736146997, "learning_rate": 1.5103262068497516e-05, "loss": 0.7442, "step": 9654 }, { "epoch": 0.34969214052879394, "grad_norm": 1.3457784327126865, "learning_rate": 1.5102253217546185e-05, "loss": 0.7455, "step": 9655 }, { "epoch": 0.34972835929011226, "grad_norm": 1.4814572091778508, "learning_rate": 1.5101244296383368e-05, "loss": 0.7603, "step": 9656 }, { "epoch": 0.34976457805143063, "grad_norm": 1.3632227535322372, "learning_rate": 1.5100235305022945e-05, "loss": 0.8864, "step": 9657 }, { "epoch": 0.349800796812749, "grad_norm": 1.2041402379454809, "learning_rate": 1.5099226243478805e-05, "loss": 0.7721, "step": 9658 }, { "epoch": 0.3498370155740674, "grad_norm": 1.3147713458796813, "learning_rate": 1.5098217111764827e-05, "loss": 0.7082, "step": 9659 }, { "epoch": 0.34987323433538575, "grad_norm": 1.4097539996383674, "learning_rate": 1.5097207909894907e-05, "loss": 0.8082, "step": 9660 }, { "epoch": 0.34990945309670407, "grad_norm": 1.4827228122575697, "learning_rate": 1.5096198637882924e-05, "loss": 0.8882, "step": 9661 }, { "epoch": 0.34994567185802244, "grad_norm": 1.058066407870294, "learning_rate": 1.5095189295742772e-05, "loss": 0.6913, "step": 9662 }, { "epoch": 0.3499818906193408, "grad_norm": 1.3337992432172658, "learning_rate": 1.5094179883488335e-05, "loss": 0.7994, "step": 9663 }, { "epoch": 0.3500181093806592, "grad_norm": 1.4918884232272356, "learning_rate": 1.5093170401133512e-05, "loss": 0.8422, "step": 9664 }, { "epoch": 0.35005432814197757, "grad_norm": 1.6038868041237022, "learning_rate": 1.5092160848692183e-05, "loss": 0.8978, "step": 9665 }, { "epoch": 0.3500905469032959, "grad_norm": 1.3894227797010015, "learning_rate": 1.5091151226178253e-05, "loss": 0.8531, "step": 9666 }, { "epoch": 0.35012676566461426, "grad_norm": 1.0922464875784719, "learning_rate": 1.5090141533605605e-05, "loss": 0.8268, "step": 9667 }, { "epoch": 0.35016298442593263, "grad_norm": 1.4604389422433606, "learning_rate": 1.5089131770988134e-05, "loss": 0.7857, "step": 9668 }, { "epoch": 0.350199203187251, "grad_norm": 1.1430672459597235, "learning_rate": 1.5088121938339741e-05, "loss": 0.7125, "step": 9669 }, { "epoch": 0.3502354219485694, "grad_norm": 1.4296536838982747, "learning_rate": 1.5087112035674318e-05, "loss": 0.8315, "step": 9670 }, { "epoch": 0.3502716407098877, "grad_norm": 1.5787566196911846, "learning_rate": 1.5086102063005768e-05, "loss": 0.7953, "step": 9671 }, { "epoch": 0.3503078594712061, "grad_norm": 1.3644169791042817, "learning_rate": 1.508509202034798e-05, "loss": 0.8477, "step": 9672 }, { "epoch": 0.35034407823252445, "grad_norm": 1.411155811373188, "learning_rate": 1.508408190771486e-05, "loss": 0.6381, "step": 9673 }, { "epoch": 0.3503802969938428, "grad_norm": 1.2777382404801882, "learning_rate": 1.5083071725120304e-05, "loss": 0.7381, "step": 9674 }, { "epoch": 0.3504165157551612, "grad_norm": 1.3429605755881737, "learning_rate": 1.5082061472578215e-05, "loss": 0.7206, "step": 9675 }, { "epoch": 0.3504527345164795, "grad_norm": 1.5524283779375332, "learning_rate": 1.5081051150102495e-05, "loss": 0.879, "step": 9676 }, { "epoch": 0.3504889532777979, "grad_norm": 1.592636438719207, "learning_rate": 1.5080040757707045e-05, "loss": 0.7893, "step": 9677 }, { "epoch": 0.35052517203911626, "grad_norm": 1.288900771254846, "learning_rate": 1.5079030295405774e-05, "loss": 0.8199, "step": 9678 }, { "epoch": 0.35056139080043464, "grad_norm": 1.4411090477234143, "learning_rate": 1.5078019763212582e-05, "loss": 0.8522, "step": 9679 }, { "epoch": 0.350597609561753, "grad_norm": 1.0778602477702612, "learning_rate": 1.5077009161141375e-05, "loss": 0.7366, "step": 9680 }, { "epoch": 0.35063382832307133, "grad_norm": 1.393763461710833, "learning_rate": 1.5075998489206063e-05, "loss": 0.8426, "step": 9681 }, { "epoch": 0.3506700470843897, "grad_norm": 1.4297096297630718, "learning_rate": 1.5074987747420551e-05, "loss": 0.7277, "step": 9682 }, { "epoch": 0.3507062658457081, "grad_norm": 1.3488869926458766, "learning_rate": 1.5073976935798747e-05, "loss": 0.7572, "step": 9683 }, { "epoch": 0.35074248460702645, "grad_norm": 1.4123472617496384, "learning_rate": 1.5072966054354561e-05, "loss": 0.7602, "step": 9684 }, { "epoch": 0.3507787033683448, "grad_norm": 1.4126854799029445, "learning_rate": 1.5071955103101905e-05, "loss": 0.7382, "step": 9685 }, { "epoch": 0.35081492212966314, "grad_norm": 1.4515821895968066, "learning_rate": 1.5070944082054693e-05, "loss": 0.8522, "step": 9686 }, { "epoch": 0.3508511408909815, "grad_norm": 1.3605384212245497, "learning_rate": 1.5069932991226832e-05, "loss": 0.7997, "step": 9687 }, { "epoch": 0.3508873596522999, "grad_norm": 1.3369437075798012, "learning_rate": 1.5068921830632238e-05, "loss": 0.8038, "step": 9688 }, { "epoch": 0.35092357841361826, "grad_norm": 1.5696161775981288, "learning_rate": 1.5067910600284827e-05, "loss": 0.9017, "step": 9689 }, { "epoch": 0.35095979717493664, "grad_norm": 1.3630158895719453, "learning_rate": 1.5066899300198512e-05, "loss": 0.6932, "step": 9690 }, { "epoch": 0.35099601593625496, "grad_norm": 1.298843454686384, "learning_rate": 1.5065887930387214e-05, "loss": 0.727, "step": 9691 }, { "epoch": 0.35103223469757333, "grad_norm": 1.4535116199211664, "learning_rate": 1.5064876490864842e-05, "loss": 0.7651, "step": 9692 }, { "epoch": 0.3510684534588917, "grad_norm": 1.470758976578735, "learning_rate": 1.5063864981645322e-05, "loss": 0.7983, "step": 9693 }, { "epoch": 0.3511046722202101, "grad_norm": 1.0510123254271861, "learning_rate": 1.506285340274257e-05, "loss": 0.7287, "step": 9694 }, { "epoch": 0.35114089098152845, "grad_norm": 1.5822590322521495, "learning_rate": 1.5061841754170505e-05, "loss": 0.7506, "step": 9695 }, { "epoch": 0.35117710974284677, "grad_norm": 1.3832548216655665, "learning_rate": 1.506083003594305e-05, "loss": 0.7638, "step": 9696 }, { "epoch": 0.35121332850416515, "grad_norm": 1.3781269521656179, "learning_rate": 1.505981824807413e-05, "loss": 0.7901, "step": 9697 }, { "epoch": 0.3512495472654835, "grad_norm": 1.508171404485476, "learning_rate": 1.505880639057766e-05, "loss": 0.7698, "step": 9698 }, { "epoch": 0.3512857660268019, "grad_norm": 1.1539435574089585, "learning_rate": 1.5057794463467572e-05, "loss": 0.7032, "step": 9699 }, { "epoch": 0.35132198478812027, "grad_norm": 0.9843317359521245, "learning_rate": 1.5056782466757787e-05, "loss": 0.7252, "step": 9700 }, { "epoch": 0.3513582035494386, "grad_norm": 1.3959152182241183, "learning_rate": 1.5055770400462233e-05, "loss": 0.6998, "step": 9701 }, { "epoch": 0.35139442231075696, "grad_norm": 1.0775847034622414, "learning_rate": 1.5054758264594838e-05, "loss": 0.7531, "step": 9702 }, { "epoch": 0.35143064107207533, "grad_norm": 1.3910575639968632, "learning_rate": 1.5053746059169522e-05, "loss": 0.7373, "step": 9703 }, { "epoch": 0.3514668598333937, "grad_norm": 1.2226966951767704, "learning_rate": 1.5052733784200224e-05, "loss": 0.712, "step": 9704 }, { "epoch": 0.3515030785947121, "grad_norm": 1.4044492437284957, "learning_rate": 1.5051721439700869e-05, "loss": 0.7418, "step": 9705 }, { "epoch": 0.3515392973560304, "grad_norm": 1.2749434507968287, "learning_rate": 1.5050709025685387e-05, "loss": 0.7328, "step": 9706 }, { "epoch": 0.3515755161173488, "grad_norm": 1.5874113119594255, "learning_rate": 1.504969654216771e-05, "loss": 0.8165, "step": 9707 }, { "epoch": 0.35161173487866715, "grad_norm": 1.2375201369455522, "learning_rate": 1.5048683989161771e-05, "loss": 0.7491, "step": 9708 }, { "epoch": 0.3516479536399855, "grad_norm": 1.2612720597838418, "learning_rate": 1.5047671366681509e-05, "loss": 0.7401, "step": 9709 }, { "epoch": 0.3516841724013039, "grad_norm": 1.3852146144453836, "learning_rate": 1.504665867474085e-05, "loss": 0.7837, "step": 9710 }, { "epoch": 0.3517203911626222, "grad_norm": 1.3616566815677964, "learning_rate": 1.5045645913353735e-05, "loss": 0.7769, "step": 9711 }, { "epoch": 0.3517566099239406, "grad_norm": 1.1912105485559301, "learning_rate": 1.5044633082534098e-05, "loss": 0.6773, "step": 9712 }, { "epoch": 0.35179282868525896, "grad_norm": 1.3816990343085147, "learning_rate": 1.5043620182295877e-05, "loss": 0.7827, "step": 9713 }, { "epoch": 0.35182904744657734, "grad_norm": 1.259865462751313, "learning_rate": 1.5042607212653008e-05, "loss": 0.7787, "step": 9714 }, { "epoch": 0.3518652662078957, "grad_norm": 1.3370707345561932, "learning_rate": 1.5041594173619439e-05, "loss": 0.7687, "step": 9715 }, { "epoch": 0.35190148496921403, "grad_norm": 1.4030319593216507, "learning_rate": 1.5040581065209098e-05, "loss": 0.7937, "step": 9716 }, { "epoch": 0.3519377037305324, "grad_norm": 1.2330740684194146, "learning_rate": 1.5039567887435938e-05, "loss": 0.7474, "step": 9717 }, { "epoch": 0.3519739224918508, "grad_norm": 1.3509107476351296, "learning_rate": 1.5038554640313892e-05, "loss": 0.7174, "step": 9718 }, { "epoch": 0.35201014125316915, "grad_norm": 1.3283516662379364, "learning_rate": 1.5037541323856906e-05, "loss": 0.7963, "step": 9719 }, { "epoch": 0.3520463600144875, "grad_norm": 1.0661223615217041, "learning_rate": 1.5036527938078926e-05, "loss": 0.7075, "step": 9720 }, { "epoch": 0.35208257877580584, "grad_norm": 1.802587922699167, "learning_rate": 1.5035514482993898e-05, "loss": 0.774, "step": 9721 }, { "epoch": 0.3521187975371242, "grad_norm": 1.3721426961682417, "learning_rate": 1.5034500958615765e-05, "loss": 0.7636, "step": 9722 }, { "epoch": 0.3521550162984426, "grad_norm": 1.4837287638007415, "learning_rate": 1.5033487364958474e-05, "loss": 0.8075, "step": 9723 }, { "epoch": 0.35219123505976097, "grad_norm": 1.1790155504172712, "learning_rate": 1.5032473702035977e-05, "loss": 0.8057, "step": 9724 }, { "epoch": 0.35222745382107934, "grad_norm": 1.429208513053072, "learning_rate": 1.5031459969862216e-05, "loss": 0.5926, "step": 9725 }, { "epoch": 0.35226367258239766, "grad_norm": 1.3490911225708206, "learning_rate": 1.503044616845115e-05, "loss": 0.72, "step": 9726 }, { "epoch": 0.35229989134371603, "grad_norm": 0.9906943675228436, "learning_rate": 1.502943229781672e-05, "loss": 0.7534, "step": 9727 }, { "epoch": 0.3523361101050344, "grad_norm": 1.3799117696764949, "learning_rate": 1.5028418357972882e-05, "loss": 0.8163, "step": 9728 }, { "epoch": 0.3523723288663528, "grad_norm": 1.3295912068416402, "learning_rate": 1.5027404348933593e-05, "loss": 0.8023, "step": 9729 }, { "epoch": 0.35240854762767115, "grad_norm": 1.4226478471408321, "learning_rate": 1.5026390270712799e-05, "loss": 0.8825, "step": 9730 }, { "epoch": 0.35244476638898947, "grad_norm": 1.2781285512951805, "learning_rate": 1.502537612332446e-05, "loss": 0.7455, "step": 9731 }, { "epoch": 0.35248098515030785, "grad_norm": 1.5791248515622804, "learning_rate": 1.5024361906782529e-05, "loss": 0.7631, "step": 9732 }, { "epoch": 0.3525172039116262, "grad_norm": 1.3390752792722314, "learning_rate": 1.5023347621100969e-05, "loss": 0.7382, "step": 9733 }, { "epoch": 0.3525534226729446, "grad_norm": 1.489707347999925, "learning_rate": 1.5022333266293725e-05, "loss": 0.774, "step": 9734 }, { "epoch": 0.35258964143426297, "grad_norm": 1.477342936728913, "learning_rate": 1.5021318842374763e-05, "loss": 0.7955, "step": 9735 }, { "epoch": 0.3526258601955813, "grad_norm": 1.3904538719023405, "learning_rate": 1.5020304349358043e-05, "loss": 0.7689, "step": 9736 }, { "epoch": 0.35266207895689966, "grad_norm": 1.3255153043371828, "learning_rate": 1.5019289787257526e-05, "loss": 0.7908, "step": 9737 }, { "epoch": 0.35269829771821803, "grad_norm": 1.442525594433004, "learning_rate": 1.5018275156087167e-05, "loss": 0.7858, "step": 9738 }, { "epoch": 0.3527345164795364, "grad_norm": 1.3546701409262774, "learning_rate": 1.5017260455860933e-05, "loss": 0.7916, "step": 9739 }, { "epoch": 0.3527707352408548, "grad_norm": 1.5390502074762502, "learning_rate": 1.501624568659279e-05, "loss": 0.8384, "step": 9740 }, { "epoch": 0.3528069540021731, "grad_norm": 1.1458550650033166, "learning_rate": 1.5015230848296695e-05, "loss": 0.8184, "step": 9741 }, { "epoch": 0.3528431727634915, "grad_norm": 1.3661929150497647, "learning_rate": 1.5014215940986621e-05, "loss": 0.6753, "step": 9742 }, { "epoch": 0.35287939152480985, "grad_norm": 2.485762459114857, "learning_rate": 1.5013200964676524e-05, "loss": 0.7622, "step": 9743 }, { "epoch": 0.3529156102861282, "grad_norm": 1.5200081618656134, "learning_rate": 1.5012185919380382e-05, "loss": 0.7352, "step": 9744 }, { "epoch": 0.3529518290474466, "grad_norm": 1.532461754938522, "learning_rate": 1.5011170805112154e-05, "loss": 0.7782, "step": 9745 }, { "epoch": 0.3529880478087649, "grad_norm": 1.6855996898275671, "learning_rate": 1.5010155621885814e-05, "loss": 0.8346, "step": 9746 }, { "epoch": 0.3530242665700833, "grad_norm": 1.3437515517991, "learning_rate": 1.5009140369715332e-05, "loss": 0.7814, "step": 9747 }, { "epoch": 0.35306048533140166, "grad_norm": 1.3936154052242162, "learning_rate": 1.5008125048614675e-05, "loss": 0.8745, "step": 9748 }, { "epoch": 0.35309670409272004, "grad_norm": 1.4026413710634613, "learning_rate": 1.5007109658597819e-05, "loss": 0.8398, "step": 9749 }, { "epoch": 0.3531329228540384, "grad_norm": 1.370399420198771, "learning_rate": 1.500609419967873e-05, "loss": 0.7829, "step": 9750 }, { "epoch": 0.35316914161535673, "grad_norm": 1.3246116688223901, "learning_rate": 1.5005078671871388e-05, "loss": 0.7776, "step": 9751 }, { "epoch": 0.3532053603766751, "grad_norm": 1.4720351904999915, "learning_rate": 1.5004063075189768e-05, "loss": 0.8643, "step": 9752 }, { "epoch": 0.3532415791379935, "grad_norm": 1.4248095872466648, "learning_rate": 1.5003047409647841e-05, "loss": 0.8408, "step": 9753 }, { "epoch": 0.35327779789931185, "grad_norm": 1.2313218663694763, "learning_rate": 1.5002031675259585e-05, "loss": 0.7647, "step": 9754 }, { "epoch": 0.3533140166606302, "grad_norm": 1.3739633438590424, "learning_rate": 1.500101587203898e-05, "loss": 0.8501, "step": 9755 }, { "epoch": 0.35335023542194854, "grad_norm": 1.1690260367098988, "learning_rate": 1.5000000000000002e-05, "loss": 0.8091, "step": 9756 }, { "epoch": 0.3533864541832669, "grad_norm": 1.1773842654751279, "learning_rate": 1.499898405915663e-05, "loss": 0.7817, "step": 9757 }, { "epoch": 0.3534226729445853, "grad_norm": 1.461664760842362, "learning_rate": 1.4997968049522844e-05, "loss": 0.797, "step": 9758 }, { "epoch": 0.35345889170590367, "grad_norm": 1.052711090226712, "learning_rate": 1.4996951971112628e-05, "loss": 0.7774, "step": 9759 }, { "epoch": 0.35349511046722204, "grad_norm": 1.3547224572400176, "learning_rate": 1.4995935823939962e-05, "loss": 0.7416, "step": 9760 }, { "epoch": 0.35353132922854036, "grad_norm": 1.2645186147738496, "learning_rate": 1.4994919608018828e-05, "loss": 0.7372, "step": 9761 }, { "epoch": 0.35356754798985873, "grad_norm": 1.3451387754280766, "learning_rate": 1.4993903323363212e-05, "loss": 0.7612, "step": 9762 }, { "epoch": 0.3536037667511771, "grad_norm": 1.306053578573252, "learning_rate": 1.4992886969987096e-05, "loss": 0.7753, "step": 9763 }, { "epoch": 0.3536399855124955, "grad_norm": 1.4406736425940416, "learning_rate": 1.4991870547904472e-05, "loss": 0.8286, "step": 9764 }, { "epoch": 0.35367620427381385, "grad_norm": 1.5637451360552161, "learning_rate": 1.4990854057129321e-05, "loss": 0.7582, "step": 9765 }, { "epoch": 0.3537124230351322, "grad_norm": 1.5001426035428522, "learning_rate": 1.4989837497675632e-05, "loss": 0.8324, "step": 9766 }, { "epoch": 0.35374864179645055, "grad_norm": 1.1248192683906972, "learning_rate": 1.4988820869557397e-05, "loss": 0.8183, "step": 9767 }, { "epoch": 0.3537848605577689, "grad_norm": 1.299073536712191, "learning_rate": 1.49878041727886e-05, "loss": 0.7621, "step": 9768 }, { "epoch": 0.3538210793190873, "grad_norm": 1.2171956406492945, "learning_rate": 1.4986787407383237e-05, "loss": 0.8042, "step": 9769 }, { "epoch": 0.35385729808040567, "grad_norm": 1.2596578714581315, "learning_rate": 1.4985770573355297e-05, "loss": 0.7103, "step": 9770 }, { "epoch": 0.353893516841724, "grad_norm": 1.4203755285003077, "learning_rate": 1.4984753670718774e-05, "loss": 0.7802, "step": 9771 }, { "epoch": 0.35392973560304236, "grad_norm": 1.4125572169792344, "learning_rate": 1.4983736699487659e-05, "loss": 0.7981, "step": 9772 }, { "epoch": 0.35396595436436074, "grad_norm": 1.2276563251731603, "learning_rate": 1.4982719659675948e-05, "loss": 0.8067, "step": 9773 }, { "epoch": 0.3540021731256791, "grad_norm": 1.371364832975596, "learning_rate": 1.4981702551297636e-05, "loss": 0.8239, "step": 9774 }, { "epoch": 0.3540383918869975, "grad_norm": 1.3834712579549382, "learning_rate": 1.4980685374366722e-05, "loss": 0.8123, "step": 9775 }, { "epoch": 0.3540746106483158, "grad_norm": 1.3884371285467019, "learning_rate": 1.49796681288972e-05, "loss": 0.766, "step": 9776 }, { "epoch": 0.3541108294096342, "grad_norm": 1.5004880931840476, "learning_rate": 1.4978650814903066e-05, "loss": 0.8861, "step": 9777 }, { "epoch": 0.35414704817095255, "grad_norm": 1.3309964778091894, "learning_rate": 1.4977633432398326e-05, "loss": 0.7683, "step": 9778 }, { "epoch": 0.3541832669322709, "grad_norm": 1.5446478773125263, "learning_rate": 1.4976615981396973e-05, "loss": 0.8713, "step": 9779 }, { "epoch": 0.3542194856935893, "grad_norm": 1.4403277344910836, "learning_rate": 1.4975598461913015e-05, "loss": 0.7479, "step": 9780 }, { "epoch": 0.3542557044549076, "grad_norm": 1.4193131783202548, "learning_rate": 1.4974580873960447e-05, "loss": 0.8212, "step": 9781 }, { "epoch": 0.354291923216226, "grad_norm": 1.3277210259334666, "learning_rate": 1.4973563217553278e-05, "loss": 0.7741, "step": 9782 }, { "epoch": 0.35432814197754436, "grad_norm": 1.5280060807666942, "learning_rate": 1.4972545492705505e-05, "loss": 0.7514, "step": 9783 }, { "epoch": 0.35436436073886274, "grad_norm": 1.3148485398637368, "learning_rate": 1.497152769943114e-05, "loss": 0.8119, "step": 9784 }, { "epoch": 0.3544005795001811, "grad_norm": 1.6836182809356908, "learning_rate": 1.4970509837744185e-05, "loss": 0.7233, "step": 9785 }, { "epoch": 0.35443679826149943, "grad_norm": 1.5521292328981495, "learning_rate": 1.4969491907658644e-05, "loss": 0.8401, "step": 9786 }, { "epoch": 0.3544730170228178, "grad_norm": 1.3881306780733775, "learning_rate": 1.4968473909188532e-05, "loss": 0.7658, "step": 9787 }, { "epoch": 0.3545092357841362, "grad_norm": 1.196478556980954, "learning_rate": 1.4967455842347851e-05, "loss": 0.7646, "step": 9788 }, { "epoch": 0.35454545454545455, "grad_norm": 1.136449431274658, "learning_rate": 1.4966437707150617e-05, "loss": 0.7502, "step": 9789 }, { "epoch": 0.3545816733067729, "grad_norm": 1.4065724294626003, "learning_rate": 1.496541950361083e-05, "loss": 0.7708, "step": 9790 }, { "epoch": 0.35461789206809124, "grad_norm": 1.503504060558048, "learning_rate": 1.4964401231742512e-05, "loss": 0.8588, "step": 9791 }, { "epoch": 0.3546541108294096, "grad_norm": 1.1554146742402327, "learning_rate": 1.4963382891559667e-05, "loss": 0.7507, "step": 9792 }, { "epoch": 0.354690329590728, "grad_norm": 1.094442843143417, "learning_rate": 1.4962364483076318e-05, "loss": 0.764, "step": 9793 }, { "epoch": 0.35472654835204637, "grad_norm": 1.1468269261959323, "learning_rate": 1.4961346006306468e-05, "loss": 0.7313, "step": 9794 }, { "epoch": 0.35476276711336474, "grad_norm": 1.4033863314905641, "learning_rate": 1.4960327461264141e-05, "loss": 0.8196, "step": 9795 }, { "epoch": 0.35479898587468306, "grad_norm": 1.043867648271661, "learning_rate": 1.4959308847963348e-05, "loss": 0.8027, "step": 9796 }, { "epoch": 0.35483520463600143, "grad_norm": 1.5116585810710392, "learning_rate": 1.4958290166418106e-05, "loss": 0.7961, "step": 9797 }, { "epoch": 0.3548714233973198, "grad_norm": 1.5900276497151438, "learning_rate": 1.4957271416642438e-05, "loss": 0.8061, "step": 9798 }, { "epoch": 0.3549076421586382, "grad_norm": 1.2387923876466032, "learning_rate": 1.4956252598650356e-05, "loss": 0.7248, "step": 9799 }, { "epoch": 0.35494386091995656, "grad_norm": 1.5215129843483985, "learning_rate": 1.4955233712455887e-05, "loss": 0.7857, "step": 9800 }, { "epoch": 0.3549800796812749, "grad_norm": 1.5021043132143748, "learning_rate": 1.4954214758073044e-05, "loss": 0.8798, "step": 9801 }, { "epoch": 0.35501629844259325, "grad_norm": 1.4358740972109438, "learning_rate": 1.4953195735515858e-05, "loss": 0.7881, "step": 9802 }, { "epoch": 0.3550525172039116, "grad_norm": 1.3120608435925711, "learning_rate": 1.4952176644798342e-05, "loss": 0.7733, "step": 9803 }, { "epoch": 0.35508873596523, "grad_norm": 1.5197326862509413, "learning_rate": 1.4951157485934527e-05, "loss": 0.8147, "step": 9804 }, { "epoch": 0.35512495472654837, "grad_norm": 1.4634933923303184, "learning_rate": 1.4950138258938431e-05, "loss": 0.8402, "step": 9805 }, { "epoch": 0.3551611734878667, "grad_norm": 1.5479140885777865, "learning_rate": 1.4949118963824086e-05, "loss": 0.7208, "step": 9806 }, { "epoch": 0.35519739224918506, "grad_norm": 1.389155388205446, "learning_rate": 1.4948099600605515e-05, "loss": 0.817, "step": 9807 }, { "epoch": 0.35523361101050344, "grad_norm": 1.3666376309216748, "learning_rate": 1.4947080169296744e-05, "loss": 0.7715, "step": 9808 }, { "epoch": 0.3552698297718218, "grad_norm": 1.3325135418543084, "learning_rate": 1.4946060669911803e-05, "loss": 0.7695, "step": 9809 }, { "epoch": 0.3553060485331402, "grad_norm": 1.539923906102877, "learning_rate": 1.494504110246472e-05, "loss": 0.8642, "step": 9810 }, { "epoch": 0.3553422672944585, "grad_norm": 1.3264065395826299, "learning_rate": 1.4944021466969532e-05, "loss": 0.7539, "step": 9811 }, { "epoch": 0.3553784860557769, "grad_norm": 1.2974715806799302, "learning_rate": 1.4943001763440262e-05, "loss": 0.7392, "step": 9812 }, { "epoch": 0.35541470481709525, "grad_norm": 1.3317582754060984, "learning_rate": 1.4941981991890941e-05, "loss": 0.7215, "step": 9813 }, { "epoch": 0.3554509235784136, "grad_norm": 1.5514852573725157, "learning_rate": 1.4940962152335607e-05, "loss": 0.7473, "step": 9814 }, { "epoch": 0.355487142339732, "grad_norm": 1.3906229124964358, "learning_rate": 1.4939942244788294e-05, "loss": 0.831, "step": 9815 }, { "epoch": 0.3555233611010503, "grad_norm": 1.3099956212955652, "learning_rate": 1.4938922269263035e-05, "loss": 0.7149, "step": 9816 }, { "epoch": 0.3555595798623687, "grad_norm": 1.2324130348248974, "learning_rate": 1.4937902225773864e-05, "loss": 0.7653, "step": 9817 }, { "epoch": 0.35559579862368706, "grad_norm": 1.326687279442573, "learning_rate": 1.4936882114334821e-05, "loss": 0.7573, "step": 9818 }, { "epoch": 0.35563201738500544, "grad_norm": 1.4211983773804904, "learning_rate": 1.4935861934959941e-05, "loss": 0.7474, "step": 9819 }, { "epoch": 0.3556682361463238, "grad_norm": 1.31685767230034, "learning_rate": 1.4934841687663266e-05, "loss": 0.7701, "step": 9820 }, { "epoch": 0.35570445490764213, "grad_norm": 1.290525491967895, "learning_rate": 1.493382137245883e-05, "loss": 0.7776, "step": 9821 }, { "epoch": 0.3557406736689605, "grad_norm": 1.427161189087966, "learning_rate": 1.493280098936068e-05, "loss": 0.7943, "step": 9822 }, { "epoch": 0.3557768924302789, "grad_norm": 1.3353980652602933, "learning_rate": 1.4931780538382852e-05, "loss": 0.6988, "step": 9823 }, { "epoch": 0.35581311119159725, "grad_norm": 1.301388139311319, "learning_rate": 1.4930760019539392e-05, "loss": 0.654, "step": 9824 }, { "epoch": 0.3558493299529156, "grad_norm": 1.1297287312511182, "learning_rate": 1.492973943284434e-05, "loss": 0.8185, "step": 9825 }, { "epoch": 0.35588554871423395, "grad_norm": 1.409641688451716, "learning_rate": 1.4928718778311742e-05, "loss": 0.8068, "step": 9826 }, { "epoch": 0.3559217674755523, "grad_norm": 1.3301998088367366, "learning_rate": 1.4927698055955643e-05, "loss": 0.7489, "step": 9827 }, { "epoch": 0.3559579862368707, "grad_norm": 1.3699803545728424, "learning_rate": 1.4926677265790088e-05, "loss": 0.7438, "step": 9828 }, { "epoch": 0.35599420499818907, "grad_norm": 1.3496943742398602, "learning_rate": 1.4925656407829125e-05, "loss": 0.8022, "step": 9829 }, { "epoch": 0.35603042375950744, "grad_norm": 1.466054599039596, "learning_rate": 1.4924635482086803e-05, "loss": 0.7464, "step": 9830 }, { "epoch": 0.35606664252082576, "grad_norm": 1.5116567574340631, "learning_rate": 1.492361448857717e-05, "loss": 0.7512, "step": 9831 }, { "epoch": 0.35610286128214413, "grad_norm": 1.392287091504486, "learning_rate": 1.4922593427314274e-05, "loss": 0.827, "step": 9832 }, { "epoch": 0.3561390800434625, "grad_norm": 1.309185618023342, "learning_rate": 1.4921572298312168e-05, "loss": 0.79, "step": 9833 }, { "epoch": 0.3561752988047809, "grad_norm": 1.3943956346579474, "learning_rate": 1.4920551101584902e-05, "loss": 0.7383, "step": 9834 }, { "epoch": 0.35621151756609926, "grad_norm": 1.4454171477914761, "learning_rate": 1.4919529837146529e-05, "loss": 0.8293, "step": 9835 }, { "epoch": 0.3562477363274176, "grad_norm": 1.4008957844219694, "learning_rate": 1.4918508505011104e-05, "loss": 0.8178, "step": 9836 }, { "epoch": 0.35628395508873595, "grad_norm": 1.3714027307960042, "learning_rate": 1.4917487105192681e-05, "loss": 0.7986, "step": 9837 }, { "epoch": 0.3563201738500543, "grad_norm": 1.3692692241868432, "learning_rate": 1.491646563770531e-05, "loss": 0.7682, "step": 9838 }, { "epoch": 0.3563563926113727, "grad_norm": 1.4361075714126068, "learning_rate": 1.4915444102563057e-05, "loss": 0.8183, "step": 9839 }, { "epoch": 0.35639261137269107, "grad_norm": 1.316675158691585, "learning_rate": 1.4914422499779972e-05, "loss": 0.749, "step": 9840 }, { "epoch": 0.35642883013400944, "grad_norm": 1.1235432271717245, "learning_rate": 1.4913400829370113e-05, "loss": 0.7721, "step": 9841 }, { "epoch": 0.35646504889532776, "grad_norm": 1.5135018351629677, "learning_rate": 1.4912379091347547e-05, "loss": 0.857, "step": 9842 }, { "epoch": 0.35650126765664614, "grad_norm": 1.4141945066008288, "learning_rate": 1.4911357285726322e-05, "loss": 0.856, "step": 9843 }, { "epoch": 0.3565374864179645, "grad_norm": 1.3794637983366915, "learning_rate": 1.491033541252051e-05, "loss": 0.7411, "step": 9844 }, { "epoch": 0.3565737051792829, "grad_norm": 1.2892753931010463, "learning_rate": 1.4909313471744167e-05, "loss": 0.7637, "step": 9845 }, { "epoch": 0.35660992394060126, "grad_norm": 1.2782956289264396, "learning_rate": 1.4908291463411356e-05, "loss": 0.7378, "step": 9846 }, { "epoch": 0.3566461427019196, "grad_norm": 1.3404830857921117, "learning_rate": 1.4907269387536141e-05, "loss": 0.8075, "step": 9847 }, { "epoch": 0.35668236146323795, "grad_norm": 1.3816720784494847, "learning_rate": 1.4906247244132588e-05, "loss": 0.8096, "step": 9848 }, { "epoch": 0.3567185802245563, "grad_norm": 1.267393290470082, "learning_rate": 1.4905225033214763e-05, "loss": 0.7005, "step": 9849 }, { "epoch": 0.3567547989858747, "grad_norm": 1.105327309818422, "learning_rate": 1.4904202754796729e-05, "loss": 0.838, "step": 9850 }, { "epoch": 0.3567910177471931, "grad_norm": 1.4161142821308743, "learning_rate": 1.4903180408892558e-05, "loss": 0.7638, "step": 9851 }, { "epoch": 0.3568272365085114, "grad_norm": 1.300517521218915, "learning_rate": 1.4902157995516315e-05, "loss": 0.6949, "step": 9852 }, { "epoch": 0.35686345526982977, "grad_norm": 3.6498677607400216, "learning_rate": 1.4901135514682072e-05, "loss": 0.7982, "step": 9853 }, { "epoch": 0.35689967403114814, "grad_norm": 1.4204229745715946, "learning_rate": 1.4900112966403897e-05, "loss": 0.8159, "step": 9854 }, { "epoch": 0.3569358927924665, "grad_norm": 1.438504929013587, "learning_rate": 1.4899090350695859e-05, "loss": 0.8112, "step": 9855 }, { "epoch": 0.3569721115537849, "grad_norm": 1.4757078714944178, "learning_rate": 1.4898067667572038e-05, "loss": 0.8568, "step": 9856 }, { "epoch": 0.3570083303151032, "grad_norm": 1.418603730617766, "learning_rate": 1.4897044917046499e-05, "loss": 0.8485, "step": 9857 }, { "epoch": 0.3570445490764216, "grad_norm": 1.4637364239403126, "learning_rate": 1.4896022099133322e-05, "loss": 0.8004, "step": 9858 }, { "epoch": 0.35708076783773995, "grad_norm": 1.1921141810599667, "learning_rate": 1.4894999213846576e-05, "loss": 0.7856, "step": 9859 }, { "epoch": 0.35711698659905833, "grad_norm": 1.4834549838470807, "learning_rate": 1.4893976261200342e-05, "loss": 0.7868, "step": 9860 }, { "epoch": 0.3571532053603767, "grad_norm": 1.4721888374633478, "learning_rate": 1.4892953241208692e-05, "loss": 0.851, "step": 9861 }, { "epoch": 0.357189424121695, "grad_norm": 1.316853008103604, "learning_rate": 1.4891930153885708e-05, "loss": 0.7444, "step": 9862 }, { "epoch": 0.3572256428830134, "grad_norm": 1.1832304507516567, "learning_rate": 1.4890906999245464e-05, "loss": 0.7507, "step": 9863 }, { "epoch": 0.35726186164433177, "grad_norm": 1.3645506877594737, "learning_rate": 1.4889883777302044e-05, "loss": 0.8953, "step": 9864 }, { "epoch": 0.35729808040565014, "grad_norm": 1.469334643250686, "learning_rate": 1.4888860488069528e-05, "loss": 0.8439, "step": 9865 }, { "epoch": 0.3573342991669685, "grad_norm": 1.299107349469005, "learning_rate": 1.4887837131561994e-05, "loss": 0.7896, "step": 9866 }, { "epoch": 0.35737051792828683, "grad_norm": 1.259481190944403, "learning_rate": 1.4886813707793528e-05, "loss": 0.791, "step": 9867 }, { "epoch": 0.3574067366896052, "grad_norm": 1.3409893689195103, "learning_rate": 1.4885790216778208e-05, "loss": 0.7558, "step": 9868 }, { "epoch": 0.3574429554509236, "grad_norm": 12.274877338709445, "learning_rate": 1.4884766658530126e-05, "loss": 0.894, "step": 9869 }, { "epoch": 0.35747917421224196, "grad_norm": 1.1734710558280543, "learning_rate": 1.4883743033063357e-05, "loss": 0.7672, "step": 9870 }, { "epoch": 0.35751539297356033, "grad_norm": 1.3957415839426126, "learning_rate": 1.4882719340392e-05, "loss": 0.8415, "step": 9871 }, { "epoch": 0.35755161173487865, "grad_norm": 1.3776961182035432, "learning_rate": 1.4881695580530129e-05, "loss": 0.8618, "step": 9872 }, { "epoch": 0.357587830496197, "grad_norm": 1.2859599327474953, "learning_rate": 1.488067175349184e-05, "loss": 0.8045, "step": 9873 }, { "epoch": 0.3576240492575154, "grad_norm": 1.282378518011869, "learning_rate": 1.4879647859291217e-05, "loss": 0.7066, "step": 9874 }, { "epoch": 0.35766026801883377, "grad_norm": 1.3283355560011418, "learning_rate": 1.4878623897942353e-05, "loss": 0.8247, "step": 9875 }, { "epoch": 0.35769648678015215, "grad_norm": 1.4411179000843548, "learning_rate": 1.4877599869459337e-05, "loss": 0.7524, "step": 9876 }, { "epoch": 0.35773270554147046, "grad_norm": 1.3960480575256793, "learning_rate": 1.487657577385626e-05, "loss": 0.8232, "step": 9877 }, { "epoch": 0.35776892430278884, "grad_norm": 1.4131579417012763, "learning_rate": 1.4875551611147219e-05, "loss": 0.7958, "step": 9878 }, { "epoch": 0.3578051430641072, "grad_norm": 1.236936467096708, "learning_rate": 1.48745273813463e-05, "loss": 0.7539, "step": 9879 }, { "epoch": 0.3578413618254256, "grad_norm": 1.446485215046984, "learning_rate": 1.4873503084467603e-05, "loss": 0.8334, "step": 9880 }, { "epoch": 0.35787758058674396, "grad_norm": 1.29173461822007, "learning_rate": 1.487247872052522e-05, "loss": 0.73, "step": 9881 }, { "epoch": 0.3579137993480623, "grad_norm": 1.4275795445591453, "learning_rate": 1.4871454289533252e-05, "loss": 0.8343, "step": 9882 }, { "epoch": 0.35795001810938065, "grad_norm": 1.2420795717769886, "learning_rate": 1.4870429791505789e-05, "loss": 0.7157, "step": 9883 }, { "epoch": 0.357986236870699, "grad_norm": 1.366877388865407, "learning_rate": 1.4869405226456934e-05, "loss": 0.8596, "step": 9884 }, { "epoch": 0.3580224556320174, "grad_norm": 1.4412919518658684, "learning_rate": 1.4868380594400785e-05, "loss": 0.7614, "step": 9885 }, { "epoch": 0.3580586743933358, "grad_norm": 1.5088937303354135, "learning_rate": 1.4867355895351438e-05, "loss": 0.8211, "step": 9886 }, { "epoch": 0.3580948931546541, "grad_norm": 1.3746370414683216, "learning_rate": 1.4866331129323e-05, "loss": 0.7356, "step": 9887 }, { "epoch": 0.35813111191597247, "grad_norm": 1.2393732184399335, "learning_rate": 1.4865306296329567e-05, "loss": 0.8037, "step": 9888 }, { "epoch": 0.35816733067729084, "grad_norm": 1.1573946191989928, "learning_rate": 1.486428139638525e-05, "loss": 0.7808, "step": 9889 }, { "epoch": 0.3582035494386092, "grad_norm": 1.3545837859553949, "learning_rate": 1.4863256429504141e-05, "loss": 0.7108, "step": 9890 }, { "epoch": 0.3582397681999276, "grad_norm": 1.3686694641893355, "learning_rate": 1.4862231395700353e-05, "loss": 0.76, "step": 9891 }, { "epoch": 0.3582759869612459, "grad_norm": 1.3617893474813536, "learning_rate": 1.4861206294987986e-05, "loss": 0.7561, "step": 9892 }, { "epoch": 0.3583122057225643, "grad_norm": 1.3403638908893858, "learning_rate": 1.4860181127381155e-05, "loss": 0.7976, "step": 9893 }, { "epoch": 0.35834842448388265, "grad_norm": 1.4220668013604922, "learning_rate": 1.4859155892893956e-05, "loss": 0.7874, "step": 9894 }, { "epoch": 0.35838464324520103, "grad_norm": 1.459999453951949, "learning_rate": 1.4858130591540503e-05, "loss": 0.8753, "step": 9895 }, { "epoch": 0.3584208620065194, "grad_norm": 1.3078554753964244, "learning_rate": 1.4857105223334905e-05, "loss": 0.7558, "step": 9896 }, { "epoch": 0.3584570807678377, "grad_norm": 1.3472482360786089, "learning_rate": 1.485607978829127e-05, "loss": 0.8025, "step": 9897 }, { "epoch": 0.3584932995291561, "grad_norm": 1.4324710487973589, "learning_rate": 1.4855054286423714e-05, "loss": 0.8837, "step": 9898 }, { "epoch": 0.35852951829047447, "grad_norm": 1.3407016487706966, "learning_rate": 1.485402871774634e-05, "loss": 0.7957, "step": 9899 }, { "epoch": 0.35856573705179284, "grad_norm": 1.4453179114472403, "learning_rate": 1.4853003082273269e-05, "loss": 0.7858, "step": 9900 }, { "epoch": 0.3586019558131112, "grad_norm": 1.3274195031457017, "learning_rate": 1.485197738001861e-05, "loss": 0.8191, "step": 9901 }, { "epoch": 0.35863817457442954, "grad_norm": 1.449708733665261, "learning_rate": 1.485095161099648e-05, "loss": 0.7653, "step": 9902 }, { "epoch": 0.3586743933357479, "grad_norm": 1.6914693609185085, "learning_rate": 1.4849925775220993e-05, "loss": 0.8447, "step": 9903 }, { "epoch": 0.3587106120970663, "grad_norm": 1.1814525449639903, "learning_rate": 1.4848899872706267e-05, "loss": 0.8236, "step": 9904 }, { "epoch": 0.35874683085838466, "grad_norm": 1.1947258683536741, "learning_rate": 1.4847873903466414e-05, "loss": 0.8396, "step": 9905 }, { "epoch": 0.35878304961970303, "grad_norm": 1.37879476826027, "learning_rate": 1.484684786751556e-05, "loss": 0.7513, "step": 9906 }, { "epoch": 0.35881926838102135, "grad_norm": 1.304387986187523, "learning_rate": 1.484582176486782e-05, "loss": 0.8107, "step": 9907 }, { "epoch": 0.3588554871423397, "grad_norm": 1.49222854737551, "learning_rate": 1.4844795595537314e-05, "loss": 0.7458, "step": 9908 }, { "epoch": 0.3588917059036581, "grad_norm": 1.2291848462856323, "learning_rate": 1.4843769359538164e-05, "loss": 0.6976, "step": 9909 }, { "epoch": 0.35892792466497647, "grad_norm": 1.2208889644305942, "learning_rate": 1.484274305688449e-05, "loss": 0.7278, "step": 9910 }, { "epoch": 0.35896414342629485, "grad_norm": 1.2976754531359567, "learning_rate": 1.4841716687590419e-05, "loss": 0.8066, "step": 9911 }, { "epoch": 0.35900036218761316, "grad_norm": 1.3776407919007074, "learning_rate": 1.4840690251670071e-05, "loss": 0.7798, "step": 9912 }, { "epoch": 0.35903658094893154, "grad_norm": 1.3974413640472885, "learning_rate": 1.4839663749137572e-05, "loss": 0.6919, "step": 9913 }, { "epoch": 0.3590727997102499, "grad_norm": 1.5989346367843502, "learning_rate": 1.4838637180007048e-05, "loss": 0.7813, "step": 9914 }, { "epoch": 0.3591090184715683, "grad_norm": 1.2981646712992203, "learning_rate": 1.4837610544292622e-05, "loss": 0.7633, "step": 9915 }, { "epoch": 0.35914523723288666, "grad_norm": 1.42562329449399, "learning_rate": 1.4836583842008425e-05, "loss": 0.7813, "step": 9916 }, { "epoch": 0.359181455994205, "grad_norm": 1.5953824854565095, "learning_rate": 1.4835557073168584e-05, "loss": 0.7986, "step": 9917 }, { "epoch": 0.35921767475552335, "grad_norm": 1.5306512759337692, "learning_rate": 1.4834530237787233e-05, "loss": 0.8543, "step": 9918 }, { "epoch": 0.3592538935168417, "grad_norm": 1.5385735892219188, "learning_rate": 1.4833503335878494e-05, "loss": 0.8326, "step": 9919 }, { "epoch": 0.3592901122781601, "grad_norm": 1.371351788698101, "learning_rate": 1.4832476367456503e-05, "loss": 0.8191, "step": 9920 }, { "epoch": 0.3593263310394785, "grad_norm": 1.350543673137436, "learning_rate": 1.483144933253539e-05, "loss": 0.7163, "step": 9921 }, { "epoch": 0.3593625498007968, "grad_norm": 1.4141674371635589, "learning_rate": 1.483042223112929e-05, "loss": 0.8216, "step": 9922 }, { "epoch": 0.35939876856211517, "grad_norm": 1.5598259503138923, "learning_rate": 1.4829395063252336e-05, "loss": 0.9051, "step": 9923 }, { "epoch": 0.35943498732343354, "grad_norm": 1.2722474780617092, "learning_rate": 1.4828367828918663e-05, "loss": 0.7706, "step": 9924 }, { "epoch": 0.3594712060847519, "grad_norm": 1.520755120502382, "learning_rate": 1.4827340528142404e-05, "loss": 0.7695, "step": 9925 }, { "epoch": 0.3595074248460703, "grad_norm": 1.5735853421507258, "learning_rate": 1.48263131609377e-05, "loss": 0.7957, "step": 9926 }, { "epoch": 0.3595436436073886, "grad_norm": 1.0721516731039096, "learning_rate": 1.4825285727318685e-05, "loss": 0.8094, "step": 9927 }, { "epoch": 0.359579862368707, "grad_norm": 1.5491630990502188, "learning_rate": 1.4824258227299497e-05, "loss": 0.8169, "step": 9928 }, { "epoch": 0.35961608113002536, "grad_norm": 1.1318537817762888, "learning_rate": 1.4823230660894282e-05, "loss": 0.7277, "step": 9929 }, { "epoch": 0.35965229989134373, "grad_norm": 1.2119647810049703, "learning_rate": 1.482220302811717e-05, "loss": 0.7886, "step": 9930 }, { "epoch": 0.3596885186526621, "grad_norm": 1.2625218653614902, "learning_rate": 1.4821175328982309e-05, "loss": 0.7191, "step": 9931 }, { "epoch": 0.3597247374139804, "grad_norm": 1.431645106481639, "learning_rate": 1.4820147563503838e-05, "loss": 0.7125, "step": 9932 }, { "epoch": 0.3597609561752988, "grad_norm": 1.486099372183728, "learning_rate": 1.4819119731695903e-05, "loss": 0.7729, "step": 9933 }, { "epoch": 0.35979717493661717, "grad_norm": 1.4137662892147327, "learning_rate": 1.4818091833572643e-05, "loss": 0.7974, "step": 9934 }, { "epoch": 0.35983339369793554, "grad_norm": 1.484178074523234, "learning_rate": 1.481706386914821e-05, "loss": 0.8063, "step": 9935 }, { "epoch": 0.3598696124592539, "grad_norm": 1.0862999608283443, "learning_rate": 1.4816035838436743e-05, "loss": 0.7947, "step": 9936 }, { "epoch": 0.35990583122057224, "grad_norm": 1.3655016968926788, "learning_rate": 1.481500774145239e-05, "loss": 0.8178, "step": 9937 }, { "epoch": 0.3599420499818906, "grad_norm": 1.3465216290095856, "learning_rate": 1.4813979578209303e-05, "loss": 0.7303, "step": 9938 }, { "epoch": 0.359978268743209, "grad_norm": 1.482435875978557, "learning_rate": 1.4812951348721623e-05, "loss": 0.7787, "step": 9939 }, { "epoch": 0.36001448750452736, "grad_norm": 1.5674286048355668, "learning_rate": 1.4811923053003508e-05, "loss": 0.871, "step": 9940 }, { "epoch": 0.36005070626584573, "grad_norm": 1.256693674511676, "learning_rate": 1.48108946910691e-05, "loss": 0.8126, "step": 9941 }, { "epoch": 0.36008692502716405, "grad_norm": 1.4105710670353098, "learning_rate": 1.4809866262932558e-05, "loss": 0.8192, "step": 9942 }, { "epoch": 0.3601231437884824, "grad_norm": 1.4100091664649559, "learning_rate": 1.4808837768608024e-05, "loss": 0.7868, "step": 9943 }, { "epoch": 0.3601593625498008, "grad_norm": 1.0104585473620904, "learning_rate": 1.480780920810966e-05, "loss": 0.7692, "step": 9944 }, { "epoch": 0.3601955813111192, "grad_norm": 1.2678435559085623, "learning_rate": 1.4806780581451619e-05, "loss": 0.8053, "step": 9945 }, { "epoch": 0.36023180007243755, "grad_norm": 1.431028926919995, "learning_rate": 1.480575188864805e-05, "loss": 0.7592, "step": 9946 }, { "epoch": 0.36026801883375587, "grad_norm": 1.388987482531866, "learning_rate": 1.4804723129713115e-05, "loss": 0.8092, "step": 9947 }, { "epoch": 0.36030423759507424, "grad_norm": 1.2154724040314722, "learning_rate": 1.4803694304660965e-05, "loss": 0.749, "step": 9948 }, { "epoch": 0.3603404563563926, "grad_norm": 1.4245730453472476, "learning_rate": 1.4802665413505763e-05, "loss": 0.7843, "step": 9949 }, { "epoch": 0.360376675117711, "grad_norm": 1.4509299026465354, "learning_rate": 1.4801636456261665e-05, "loss": 0.7289, "step": 9950 }, { "epoch": 0.36041289387902936, "grad_norm": 1.19884844974356, "learning_rate": 1.480060743294283e-05, "loss": 0.744, "step": 9951 }, { "epoch": 0.3604491126403477, "grad_norm": 1.5096854864715858, "learning_rate": 1.4799578343563417e-05, "loss": 0.769, "step": 9952 }, { "epoch": 0.36048533140166605, "grad_norm": 1.3761599731851026, "learning_rate": 1.4798549188137588e-05, "loss": 0.771, "step": 9953 }, { "epoch": 0.3605215501629844, "grad_norm": 1.3988093828704995, "learning_rate": 1.4797519966679508e-05, "loss": 0.7821, "step": 9954 }, { "epoch": 0.3605577689243028, "grad_norm": 1.2740488851889722, "learning_rate": 1.4796490679203337e-05, "loss": 0.7519, "step": 9955 }, { "epoch": 0.3605939876856212, "grad_norm": 1.3288670794787671, "learning_rate": 1.479546132572324e-05, "loss": 0.8126, "step": 9956 }, { "epoch": 0.3606302064469395, "grad_norm": 1.3560415804104677, "learning_rate": 1.4794431906253381e-05, "loss": 0.8455, "step": 9957 }, { "epoch": 0.36066642520825787, "grad_norm": 1.3959663537384404, "learning_rate": 1.4793402420807927e-05, "loss": 0.8301, "step": 9958 }, { "epoch": 0.36070264396957624, "grad_norm": 1.0611270001524504, "learning_rate": 1.4792372869401042e-05, "loss": 0.7243, "step": 9959 }, { "epoch": 0.3607388627308946, "grad_norm": 1.4238057245970164, "learning_rate": 1.47913432520469e-05, "loss": 0.7245, "step": 9960 }, { "epoch": 0.360775081492213, "grad_norm": 1.3193072874288905, "learning_rate": 1.4790313568759658e-05, "loss": 0.7274, "step": 9961 }, { "epoch": 0.3608113002535313, "grad_norm": 1.4062681766500889, "learning_rate": 1.4789283819553497e-05, "loss": 0.885, "step": 9962 }, { "epoch": 0.3608475190148497, "grad_norm": 1.4278973676843012, "learning_rate": 1.4788254004442582e-05, "loss": 0.7921, "step": 9963 }, { "epoch": 0.36088373777616806, "grad_norm": 1.124822548927561, "learning_rate": 1.4787224123441081e-05, "loss": 0.7687, "step": 9964 }, { "epoch": 0.36091995653748643, "grad_norm": 1.4962556000704723, "learning_rate": 1.4786194176563174e-05, "loss": 0.853, "step": 9965 }, { "epoch": 0.3609561752988048, "grad_norm": 1.4059690843910606, "learning_rate": 1.4785164163823026e-05, "loss": 0.8448, "step": 9966 }, { "epoch": 0.3609923940601231, "grad_norm": 1.1767682994577058, "learning_rate": 1.4784134085234817e-05, "loss": 0.7592, "step": 9967 }, { "epoch": 0.3610286128214415, "grad_norm": 1.3523442172782967, "learning_rate": 1.4783103940812719e-05, "loss": 0.7587, "step": 9968 }, { "epoch": 0.36106483158275987, "grad_norm": 1.0298870022119446, "learning_rate": 1.478207373057091e-05, "loss": 0.7436, "step": 9969 }, { "epoch": 0.36110105034407824, "grad_norm": 0.9900490685700644, "learning_rate": 1.478104345452356e-05, "loss": 0.7835, "step": 9970 }, { "epoch": 0.3611372691053966, "grad_norm": 1.308037960820917, "learning_rate": 1.4780013112684855e-05, "loss": 0.7199, "step": 9971 }, { "epoch": 0.36117348786671494, "grad_norm": 1.3279793168556535, "learning_rate": 1.4778982705068968e-05, "loss": 0.735, "step": 9972 }, { "epoch": 0.3612097066280333, "grad_norm": 1.4054221257843664, "learning_rate": 1.4777952231690078e-05, "loss": 0.8684, "step": 9973 }, { "epoch": 0.3612459253893517, "grad_norm": 1.1407971356102098, "learning_rate": 1.4776921692562369e-05, "loss": 0.7177, "step": 9974 }, { "epoch": 0.36128214415067006, "grad_norm": 1.3597738479709722, "learning_rate": 1.4775891087700018e-05, "loss": 0.7341, "step": 9975 }, { "epoch": 0.36131836291198843, "grad_norm": 1.465756280861436, "learning_rate": 1.4774860417117211e-05, "loss": 0.8034, "step": 9976 }, { "epoch": 0.36135458167330675, "grad_norm": 1.3761974811201965, "learning_rate": 1.4773829680828129e-05, "loss": 0.7842, "step": 9977 }, { "epoch": 0.3613908004346251, "grad_norm": 1.1540036734766135, "learning_rate": 1.4772798878846957e-05, "loss": 0.7788, "step": 9978 }, { "epoch": 0.3614270191959435, "grad_norm": 1.3358521134433323, "learning_rate": 1.4771768011187878e-05, "loss": 0.6983, "step": 9979 }, { "epoch": 0.3614632379572619, "grad_norm": 1.1477743756565617, "learning_rate": 1.4770737077865078e-05, "loss": 0.7934, "step": 9980 }, { "epoch": 0.36149945671858025, "grad_norm": 1.4176667997997145, "learning_rate": 1.4769706078892742e-05, "loss": 0.7885, "step": 9981 }, { "epoch": 0.36153567547989857, "grad_norm": 1.4164385302190245, "learning_rate": 1.4768675014285063e-05, "loss": 0.7824, "step": 9982 }, { "epoch": 0.36157189424121694, "grad_norm": 1.3921547623359034, "learning_rate": 1.4767643884056222e-05, "loss": 0.8068, "step": 9983 }, { "epoch": 0.3616081130025353, "grad_norm": 1.235036007531443, "learning_rate": 1.4766612688220414e-05, "loss": 0.6912, "step": 9984 }, { "epoch": 0.3616443317638537, "grad_norm": 1.3921340325231966, "learning_rate": 1.4765581426791826e-05, "loss": 0.8322, "step": 9985 }, { "epoch": 0.36168055052517206, "grad_norm": 1.4285840090874644, "learning_rate": 1.4764550099784651e-05, "loss": 0.7473, "step": 9986 }, { "epoch": 0.3617167692864904, "grad_norm": 0.9775085558785154, "learning_rate": 1.4763518707213083e-05, "loss": 0.7941, "step": 9987 }, { "epoch": 0.36175298804780875, "grad_norm": 1.2457649378603906, "learning_rate": 1.4762487249091306e-05, "loss": 0.7109, "step": 9988 }, { "epoch": 0.36178920680912713, "grad_norm": 1.4505542054152647, "learning_rate": 1.4761455725433525e-05, "loss": 0.7737, "step": 9989 }, { "epoch": 0.3618254255704455, "grad_norm": 1.4360526134748997, "learning_rate": 1.4760424136253926e-05, "loss": 0.85, "step": 9990 }, { "epoch": 0.3618616443317639, "grad_norm": 1.2712861242009812, "learning_rate": 1.475939248156671e-05, "loss": 0.7904, "step": 9991 }, { "epoch": 0.3618978630930822, "grad_norm": 1.5416908767019262, "learning_rate": 1.475836076138607e-05, "loss": 0.8039, "step": 9992 }, { "epoch": 0.36193408185440057, "grad_norm": 1.2011121932764093, "learning_rate": 1.4757328975726207e-05, "loss": 0.8396, "step": 9993 }, { "epoch": 0.36197030061571894, "grad_norm": 1.3050659140045846, "learning_rate": 1.4756297124601316e-05, "loss": 0.7454, "step": 9994 }, { "epoch": 0.3620065193770373, "grad_norm": 1.3403279629282536, "learning_rate": 1.4755265208025595e-05, "loss": 0.7525, "step": 9995 }, { "epoch": 0.3620427381383557, "grad_norm": 1.4460451756299537, "learning_rate": 1.4754233226013251e-05, "loss": 0.772, "step": 9996 }, { "epoch": 0.362078956899674, "grad_norm": 1.6761981244411026, "learning_rate": 1.4753201178578476e-05, "loss": 0.7957, "step": 9997 }, { "epoch": 0.3621151756609924, "grad_norm": 1.497065697157104, "learning_rate": 1.475216906573548e-05, "loss": 0.8066, "step": 9998 }, { "epoch": 0.36215139442231076, "grad_norm": 1.3861988402625196, "learning_rate": 1.475113688749846e-05, "loss": 0.8032, "step": 9999 }, { "epoch": 0.36218761318362913, "grad_norm": 1.365943445546422, "learning_rate": 1.4750104643881623e-05, "loss": 0.8014, "step": 10000 }, { "epoch": 0.3622238319449475, "grad_norm": 1.2527482729028032, "learning_rate": 1.4749072334899168e-05, "loss": 0.8211, "step": 10001 }, { "epoch": 0.3622600507062658, "grad_norm": 1.4136262472325394, "learning_rate": 1.474803996056531e-05, "loss": 0.7565, "step": 10002 }, { "epoch": 0.3622962694675842, "grad_norm": 1.580799096261165, "learning_rate": 1.4747007520894245e-05, "loss": 0.7927, "step": 10003 }, { "epoch": 0.36233248822890257, "grad_norm": 1.543677167390595, "learning_rate": 1.474597501590019e-05, "loss": 0.7482, "step": 10004 }, { "epoch": 0.36236870699022095, "grad_norm": 1.354649786935698, "learning_rate": 1.4744942445597349e-05, "loss": 0.7733, "step": 10005 }, { "epoch": 0.3624049257515393, "grad_norm": 1.4526317974616283, "learning_rate": 1.4743909809999926e-05, "loss": 0.8646, "step": 10006 }, { "epoch": 0.36244114451285764, "grad_norm": 1.0661476585561032, "learning_rate": 1.474287710912214e-05, "loss": 0.7185, "step": 10007 }, { "epoch": 0.362477363274176, "grad_norm": 1.3768011388023775, "learning_rate": 1.4741844342978196e-05, "loss": 0.7085, "step": 10008 }, { "epoch": 0.3625135820354944, "grad_norm": 1.1863499889440967, "learning_rate": 1.4740811511582308e-05, "loss": 0.7483, "step": 10009 }, { "epoch": 0.36254980079681276, "grad_norm": 1.2558160369478943, "learning_rate": 1.4739778614948685e-05, "loss": 0.699, "step": 10010 }, { "epoch": 0.36258601955813113, "grad_norm": 1.347160700204286, "learning_rate": 1.4738745653091547e-05, "loss": 0.7885, "step": 10011 }, { "epoch": 0.36262223831944945, "grad_norm": 1.3286727547083306, "learning_rate": 1.4737712626025103e-05, "loss": 0.7215, "step": 10012 }, { "epoch": 0.3626584570807678, "grad_norm": 1.349891910431249, "learning_rate": 1.4736679533763573e-05, "loss": 0.7525, "step": 10013 }, { "epoch": 0.3626946758420862, "grad_norm": 1.4146643840876416, "learning_rate": 1.4735646376321166e-05, "loss": 0.8243, "step": 10014 }, { "epoch": 0.3627308946034046, "grad_norm": 1.3797890630625291, "learning_rate": 1.4734613153712107e-05, "loss": 0.809, "step": 10015 }, { "epoch": 0.36276711336472295, "grad_norm": 1.0248228536851087, "learning_rate": 1.4733579865950611e-05, "loss": 0.7902, "step": 10016 }, { "epoch": 0.36280333212604127, "grad_norm": 1.2576693089053983, "learning_rate": 1.4732546513050894e-05, "loss": 0.7959, "step": 10017 }, { "epoch": 0.36283955088735964, "grad_norm": 1.139047417312712, "learning_rate": 1.4731513095027183e-05, "loss": 0.766, "step": 10018 }, { "epoch": 0.362875769648678, "grad_norm": 1.4054740645905586, "learning_rate": 1.473047961189369e-05, "loss": 0.7705, "step": 10019 }, { "epoch": 0.3629119884099964, "grad_norm": 1.4124633522153363, "learning_rate": 1.4729446063664645e-05, "loss": 0.8301, "step": 10020 }, { "epoch": 0.36294820717131476, "grad_norm": 1.2961911644867483, "learning_rate": 1.4728412450354264e-05, "loss": 0.7768, "step": 10021 }, { "epoch": 0.3629844259326331, "grad_norm": 1.3916840645945612, "learning_rate": 1.4727378771976772e-05, "loss": 0.8117, "step": 10022 }, { "epoch": 0.36302064469395146, "grad_norm": 1.3979891679060872, "learning_rate": 1.4726345028546396e-05, "loss": 0.7763, "step": 10023 }, { "epoch": 0.36305686345526983, "grad_norm": 1.4881870660491654, "learning_rate": 1.472531122007736e-05, "loss": 0.8152, "step": 10024 }, { "epoch": 0.3630930822165882, "grad_norm": 1.3972951909371967, "learning_rate": 1.4724277346583889e-05, "loss": 0.8341, "step": 10025 }, { "epoch": 0.3631293009779066, "grad_norm": 1.4330344180896832, "learning_rate": 1.472324340808021e-05, "loss": 0.8157, "step": 10026 }, { "epoch": 0.3631655197392249, "grad_norm": 1.412612249072875, "learning_rate": 1.4722209404580551e-05, "loss": 0.7268, "step": 10027 }, { "epoch": 0.36320173850054327, "grad_norm": 1.4493063210367227, "learning_rate": 1.4721175336099142e-05, "loss": 0.7302, "step": 10028 }, { "epoch": 0.36323795726186164, "grad_norm": 1.3232566297805755, "learning_rate": 1.4720141202650214e-05, "loss": 0.7885, "step": 10029 }, { "epoch": 0.36327417602318, "grad_norm": 1.3155941453292173, "learning_rate": 1.4719107004247992e-05, "loss": 0.6991, "step": 10030 }, { "epoch": 0.3633103947844984, "grad_norm": 1.5083027336722965, "learning_rate": 1.4718072740906713e-05, "loss": 0.8109, "step": 10031 }, { "epoch": 0.3633466135458167, "grad_norm": 1.1376784507582485, "learning_rate": 1.4717038412640607e-05, "loss": 0.7278, "step": 10032 }, { "epoch": 0.3633828323071351, "grad_norm": 1.4420836208744234, "learning_rate": 1.471600401946391e-05, "loss": 0.7796, "step": 10033 }, { "epoch": 0.36341905106845346, "grad_norm": 1.1395926791319162, "learning_rate": 1.4714969561390851e-05, "loss": 0.7527, "step": 10034 }, { "epoch": 0.36345526982977183, "grad_norm": 1.080096438452687, "learning_rate": 1.4713935038435666e-05, "loss": 0.7363, "step": 10035 }, { "epoch": 0.3634914885910902, "grad_norm": 1.44016043716634, "learning_rate": 1.4712900450612598e-05, "loss": 0.8189, "step": 10036 }, { "epoch": 0.3635277073524085, "grad_norm": 1.4805198952488872, "learning_rate": 1.4711865797935873e-05, "loss": 0.7488, "step": 10037 }, { "epoch": 0.3635639261137269, "grad_norm": 1.2428561793541706, "learning_rate": 1.4710831080419739e-05, "loss": 0.7745, "step": 10038 }, { "epoch": 0.3636001448750453, "grad_norm": 1.1463637431191238, "learning_rate": 1.4709796298078428e-05, "loss": 0.753, "step": 10039 }, { "epoch": 0.36363636363636365, "grad_norm": 1.4225928808450794, "learning_rate": 1.4708761450926184e-05, "loss": 0.7812, "step": 10040 }, { "epoch": 0.363672582397682, "grad_norm": 1.300685149443276, "learning_rate": 1.470772653897724e-05, "loss": 0.7494, "step": 10041 }, { "epoch": 0.36370880115900034, "grad_norm": 1.373158291654913, "learning_rate": 1.4706691562245847e-05, "loss": 0.8189, "step": 10042 }, { "epoch": 0.3637450199203187, "grad_norm": 1.454440701721383, "learning_rate": 1.470565652074624e-05, "loss": 0.7789, "step": 10043 }, { "epoch": 0.3637812386816371, "grad_norm": 1.38702004880122, "learning_rate": 1.4704621414492666e-05, "loss": 0.7644, "step": 10044 }, { "epoch": 0.36381745744295546, "grad_norm": 1.3456025598152512, "learning_rate": 1.4703586243499366e-05, "loss": 0.8031, "step": 10045 }, { "epoch": 0.36385367620427383, "grad_norm": 1.8265001682102837, "learning_rate": 1.4702551007780587e-05, "loss": 0.7357, "step": 10046 }, { "epoch": 0.36388989496559215, "grad_norm": 1.4369213083584107, "learning_rate": 1.4701515707350576e-05, "loss": 0.749, "step": 10047 }, { "epoch": 0.3639261137269105, "grad_norm": 1.138273648126267, "learning_rate": 1.4700480342223574e-05, "loss": 0.8222, "step": 10048 }, { "epoch": 0.3639623324882289, "grad_norm": 1.5582865529403795, "learning_rate": 1.4699444912413835e-05, "loss": 0.9012, "step": 10049 }, { "epoch": 0.3639985512495473, "grad_norm": 1.0764621977568958, "learning_rate": 1.4698409417935603e-05, "loss": 0.7909, "step": 10050 }, { "epoch": 0.36403477001086565, "grad_norm": 1.4217579054563203, "learning_rate": 1.4697373858803133e-05, "loss": 0.7268, "step": 10051 }, { "epoch": 0.36407098877218397, "grad_norm": 2.6253369487987506, "learning_rate": 1.4696338235030668e-05, "loss": 0.8591, "step": 10052 }, { "epoch": 0.36410720753350234, "grad_norm": 1.4141145065660203, "learning_rate": 1.4695302546632462e-05, "loss": 0.8025, "step": 10053 }, { "epoch": 0.3641434262948207, "grad_norm": 1.5207148284226586, "learning_rate": 1.4694266793622766e-05, "loss": 0.7627, "step": 10054 }, { "epoch": 0.3641796450561391, "grad_norm": 1.301155701573684, "learning_rate": 1.4693230976015837e-05, "loss": 0.7963, "step": 10055 }, { "epoch": 0.36421586381745746, "grad_norm": 1.016313923584849, "learning_rate": 1.4692195093825927e-05, "loss": 0.7278, "step": 10056 }, { "epoch": 0.3642520825787758, "grad_norm": 1.2387547550365172, "learning_rate": 1.4691159147067287e-05, "loss": 0.6912, "step": 10057 }, { "epoch": 0.36428830134009416, "grad_norm": 1.299645799928857, "learning_rate": 1.4690123135754175e-05, "loss": 0.7891, "step": 10058 }, { "epoch": 0.36432452010141253, "grad_norm": 1.4747165742018096, "learning_rate": 1.4689087059900848e-05, "loss": 0.7823, "step": 10059 }, { "epoch": 0.3643607388627309, "grad_norm": 1.356256323516185, "learning_rate": 1.4688050919521563e-05, "loss": 0.7823, "step": 10060 }, { "epoch": 0.3643969576240493, "grad_norm": 1.2336659530215837, "learning_rate": 1.4687014714630578e-05, "loss": 0.7329, "step": 10061 }, { "epoch": 0.3644331763853676, "grad_norm": 1.3017732858274795, "learning_rate": 1.4685978445242153e-05, "loss": 0.7394, "step": 10062 }, { "epoch": 0.36446939514668597, "grad_norm": 1.8544896242787916, "learning_rate": 1.4684942111370545e-05, "loss": 0.7507, "step": 10063 }, { "epoch": 0.36450561390800434, "grad_norm": 1.480273276637444, "learning_rate": 1.4683905713030018e-05, "loss": 0.7647, "step": 10064 }, { "epoch": 0.3645418326693227, "grad_norm": 1.3965769557008405, "learning_rate": 1.4682869250234835e-05, "loss": 0.8253, "step": 10065 }, { "epoch": 0.3645780514306411, "grad_norm": 1.2238775953758771, "learning_rate": 1.4681832722999255e-05, "loss": 0.8075, "step": 10066 }, { "epoch": 0.3646142701919594, "grad_norm": 1.4177721914561012, "learning_rate": 1.4680796131337543e-05, "loss": 0.7178, "step": 10067 }, { "epoch": 0.3646504889532778, "grad_norm": 1.2660639569744512, "learning_rate": 1.4679759475263962e-05, "loss": 0.7699, "step": 10068 }, { "epoch": 0.36468670771459616, "grad_norm": 1.2751048441162296, "learning_rate": 1.467872275479278e-05, "loss": 0.6963, "step": 10069 }, { "epoch": 0.36472292647591453, "grad_norm": 1.2625013971084948, "learning_rate": 1.467768596993826e-05, "loss": 0.7972, "step": 10070 }, { "epoch": 0.3647591452372329, "grad_norm": 1.4239274848471006, "learning_rate": 1.4676649120714674e-05, "loss": 0.7478, "step": 10071 }, { "epoch": 0.3647953639985512, "grad_norm": 1.3826759238604935, "learning_rate": 1.4675612207136283e-05, "loss": 0.7913, "step": 10072 }, { "epoch": 0.3648315827598696, "grad_norm": 1.1291173055802053, "learning_rate": 1.4674575229217361e-05, "loss": 0.8093, "step": 10073 }, { "epoch": 0.364867801521188, "grad_norm": 0.9944129657001308, "learning_rate": 1.4673538186972177e-05, "loss": 0.7653, "step": 10074 }, { "epoch": 0.36490402028250635, "grad_norm": 1.0737450790799048, "learning_rate": 1.4672501080415002e-05, "loss": 0.7326, "step": 10075 }, { "epoch": 0.3649402390438247, "grad_norm": 1.4131543987169504, "learning_rate": 1.4671463909560107e-05, "loss": 0.8, "step": 10076 }, { "epoch": 0.36497645780514304, "grad_norm": 1.404102378953226, "learning_rate": 1.467042667442176e-05, "loss": 0.7525, "step": 10077 }, { "epoch": 0.3650126765664614, "grad_norm": 1.249501947183626, "learning_rate": 1.4669389375014243e-05, "loss": 0.7273, "step": 10078 }, { "epoch": 0.3650488953277798, "grad_norm": 1.3173911269019578, "learning_rate": 1.466835201135182e-05, "loss": 0.7346, "step": 10079 }, { "epoch": 0.36508511408909816, "grad_norm": 1.450243236250814, "learning_rate": 1.4667314583448777e-05, "loss": 0.7872, "step": 10080 }, { "epoch": 0.36512133285041654, "grad_norm": 1.0258651455133787, "learning_rate": 1.466627709131938e-05, "loss": 0.7271, "step": 10081 }, { "epoch": 0.36515755161173485, "grad_norm": 1.3528392141960095, "learning_rate": 1.4665239534977912e-05, "loss": 0.6633, "step": 10082 }, { "epoch": 0.36519377037305323, "grad_norm": 1.5411976155633598, "learning_rate": 1.4664201914438647e-05, "loss": 0.7439, "step": 10083 }, { "epoch": 0.3652299891343716, "grad_norm": 1.1636776047789952, "learning_rate": 1.4663164229715868e-05, "loss": 0.7472, "step": 10084 }, { "epoch": 0.36526620789569, "grad_norm": 1.4179718842484905, "learning_rate": 1.466212648082385e-05, "loss": 0.8262, "step": 10085 }, { "epoch": 0.36530242665700835, "grad_norm": 1.3252494234522527, "learning_rate": 1.4661088667776876e-05, "loss": 0.7631, "step": 10086 }, { "epoch": 0.36533864541832667, "grad_norm": 1.3681735672167923, "learning_rate": 1.4660050790589227e-05, "loss": 0.7864, "step": 10087 }, { "epoch": 0.36537486417964504, "grad_norm": 1.3905397212208135, "learning_rate": 1.4659012849275183e-05, "loss": 0.8162, "step": 10088 }, { "epoch": 0.3654110829409634, "grad_norm": 1.1730205556476232, "learning_rate": 1.465797484384903e-05, "loss": 0.7529, "step": 10089 }, { "epoch": 0.3654473017022818, "grad_norm": 1.3757564618735874, "learning_rate": 1.4656936774325047e-05, "loss": 0.7994, "step": 10090 }, { "epoch": 0.36548352046360016, "grad_norm": 1.3826612003173997, "learning_rate": 1.4655898640717525e-05, "loss": 0.7543, "step": 10091 }, { "epoch": 0.3655197392249185, "grad_norm": 1.3495478491829478, "learning_rate": 1.4654860443040746e-05, "loss": 0.7329, "step": 10092 }, { "epoch": 0.36555595798623686, "grad_norm": 1.2375134154276544, "learning_rate": 1.4653822181308997e-05, "loss": 0.7242, "step": 10093 }, { "epoch": 0.36559217674755523, "grad_norm": 1.3385765878844673, "learning_rate": 1.4652783855536565e-05, "loss": 0.8246, "step": 10094 }, { "epoch": 0.3656283955088736, "grad_norm": 1.207858734831274, "learning_rate": 1.465174546573774e-05, "loss": 0.6721, "step": 10095 }, { "epoch": 0.365664614270192, "grad_norm": 1.4439725105890235, "learning_rate": 1.4650707011926808e-05, "loss": 0.855, "step": 10096 }, { "epoch": 0.3657008330315103, "grad_norm": 1.4183921628422795, "learning_rate": 1.4649668494118063e-05, "loss": 0.7864, "step": 10097 }, { "epoch": 0.36573705179282867, "grad_norm": 1.3459702864653176, "learning_rate": 1.4648629912325794e-05, "loss": 0.7608, "step": 10098 }, { "epoch": 0.36577327055414705, "grad_norm": 1.3867041378863232, "learning_rate": 1.4647591266564291e-05, "loss": 0.7768, "step": 10099 }, { "epoch": 0.3658094893154654, "grad_norm": 1.3636458149924713, "learning_rate": 1.464655255684785e-05, "loss": 0.8175, "step": 10100 }, { "epoch": 0.3658457080767838, "grad_norm": 1.4702006320019434, "learning_rate": 1.4645513783190764e-05, "loss": 0.7539, "step": 10101 }, { "epoch": 0.3658819268381021, "grad_norm": 1.6374061661327168, "learning_rate": 1.4644474945607325e-05, "loss": 0.7679, "step": 10102 }, { "epoch": 0.3659181455994205, "grad_norm": 1.3538431573608112, "learning_rate": 1.4643436044111829e-05, "loss": 0.7588, "step": 10103 }, { "epoch": 0.36595436436073886, "grad_norm": 1.4594431848954788, "learning_rate": 1.4642397078718574e-05, "loss": 0.8158, "step": 10104 }, { "epoch": 0.36599058312205723, "grad_norm": 1.4139929156280218, "learning_rate": 1.4641358049441855e-05, "loss": 0.8519, "step": 10105 }, { "epoch": 0.3660268018833756, "grad_norm": 1.3530634837725617, "learning_rate": 1.4640318956295971e-05, "loss": 0.8138, "step": 10106 }, { "epoch": 0.3660630206446939, "grad_norm": 1.2827359133537561, "learning_rate": 1.4639279799295226e-05, "loss": 0.7419, "step": 10107 }, { "epoch": 0.3660992394060123, "grad_norm": 1.4277377448192232, "learning_rate": 1.4638240578453907e-05, "loss": 0.6965, "step": 10108 }, { "epoch": 0.3661354581673307, "grad_norm": 1.413855316859513, "learning_rate": 1.4637201293786328e-05, "loss": 0.7573, "step": 10109 }, { "epoch": 0.36617167692864905, "grad_norm": 1.1417469670564235, "learning_rate": 1.463616194530678e-05, "loss": 0.7294, "step": 10110 }, { "epoch": 0.3662078956899674, "grad_norm": 1.2575873153374646, "learning_rate": 1.4635122533029573e-05, "loss": 0.7918, "step": 10111 }, { "epoch": 0.36624411445128574, "grad_norm": 1.0684378391000362, "learning_rate": 1.4634083056969006e-05, "loss": 0.7618, "step": 10112 }, { "epoch": 0.3662803332126041, "grad_norm": 1.441706522754032, "learning_rate": 1.4633043517139382e-05, "loss": 0.8025, "step": 10113 }, { "epoch": 0.3663165519739225, "grad_norm": 1.379411873835017, "learning_rate": 1.4632003913555012e-05, "loss": 0.8738, "step": 10114 }, { "epoch": 0.36635277073524086, "grad_norm": 1.2902266440360592, "learning_rate": 1.4630964246230197e-05, "loss": 0.7588, "step": 10115 }, { "epoch": 0.36638898949655924, "grad_norm": 1.2535123648817303, "learning_rate": 1.4629924515179243e-05, "loss": 0.7605, "step": 10116 }, { "epoch": 0.36642520825787755, "grad_norm": 1.516644490132491, "learning_rate": 1.4628884720416461e-05, "loss": 0.7462, "step": 10117 }, { "epoch": 0.36646142701919593, "grad_norm": 1.2609300104140975, "learning_rate": 1.462784486195616e-05, "loss": 0.7335, "step": 10118 }, { "epoch": 0.3664976457805143, "grad_norm": 1.4222333556058078, "learning_rate": 1.4626804939812645e-05, "loss": 0.7425, "step": 10119 }, { "epoch": 0.3665338645418327, "grad_norm": 1.4324811462579246, "learning_rate": 1.4625764954000228e-05, "loss": 0.7956, "step": 10120 }, { "epoch": 0.36657008330315105, "grad_norm": 1.5398186010013095, "learning_rate": 1.4624724904533221e-05, "loss": 0.8201, "step": 10121 }, { "epoch": 0.36660630206446937, "grad_norm": 1.276629372850571, "learning_rate": 1.4623684791425939e-05, "loss": 0.7443, "step": 10122 }, { "epoch": 0.36664252082578774, "grad_norm": 1.1807617499784757, "learning_rate": 1.4622644614692687e-05, "loss": 0.7702, "step": 10123 }, { "epoch": 0.3666787395871061, "grad_norm": 1.5589133200323104, "learning_rate": 1.4621604374347781e-05, "loss": 0.7908, "step": 10124 }, { "epoch": 0.3667149583484245, "grad_norm": 1.3774698086162929, "learning_rate": 1.4620564070405545e-05, "loss": 0.8298, "step": 10125 }, { "epoch": 0.36675117710974287, "grad_norm": 1.4379098342974115, "learning_rate": 1.4619523702880285e-05, "loss": 0.808, "step": 10126 }, { "epoch": 0.3667873958710612, "grad_norm": 1.3652325592315453, "learning_rate": 1.461848327178632e-05, "loss": 0.7199, "step": 10127 }, { "epoch": 0.36682361463237956, "grad_norm": 1.5429937360647838, "learning_rate": 1.4617442777137965e-05, "loss": 0.8263, "step": 10128 }, { "epoch": 0.36685983339369793, "grad_norm": 1.4604450883873907, "learning_rate": 1.4616402218949541e-05, "loss": 0.7036, "step": 10129 }, { "epoch": 0.3668960521550163, "grad_norm": 1.2995113024499882, "learning_rate": 1.4615361597235366e-05, "loss": 0.7233, "step": 10130 }, { "epoch": 0.3669322709163347, "grad_norm": 1.0647682378080543, "learning_rate": 1.461432091200976e-05, "loss": 0.7817, "step": 10131 }, { "epoch": 0.366968489677653, "grad_norm": 1.0241545920423918, "learning_rate": 1.4613280163287044e-05, "loss": 0.7441, "step": 10132 }, { "epoch": 0.36700470843897137, "grad_norm": 1.4471495082415806, "learning_rate": 1.461223935108154e-05, "loss": 0.7722, "step": 10133 }, { "epoch": 0.36704092720028975, "grad_norm": 1.4718047293311656, "learning_rate": 1.4611198475407572e-05, "loss": 0.8972, "step": 10134 }, { "epoch": 0.3670771459616081, "grad_norm": 1.4151501819269483, "learning_rate": 1.4610157536279456e-05, "loss": 0.8378, "step": 10135 }, { "epoch": 0.3671133647229265, "grad_norm": 1.1716665354720386, "learning_rate": 1.4609116533711526e-05, "loss": 0.8204, "step": 10136 }, { "epoch": 0.3671495834842448, "grad_norm": 1.358364008803836, "learning_rate": 1.4608075467718098e-05, "loss": 0.7216, "step": 10137 }, { "epoch": 0.3671858022455632, "grad_norm": 1.4324401896949281, "learning_rate": 1.4607034338313506e-05, "loss": 0.7153, "step": 10138 }, { "epoch": 0.36722202100688156, "grad_norm": 1.5994372943395578, "learning_rate": 1.4605993145512073e-05, "loss": 0.7314, "step": 10139 }, { "epoch": 0.36725823976819993, "grad_norm": 1.4306972063671533, "learning_rate": 1.4604951889328127e-05, "loss": 0.8087, "step": 10140 }, { "epoch": 0.3672944585295183, "grad_norm": 1.2191048384113368, "learning_rate": 1.4603910569775997e-05, "loss": 0.7535, "step": 10141 }, { "epoch": 0.3673306772908366, "grad_norm": 1.0237173470084036, "learning_rate": 1.4602869186870012e-05, "loss": 0.7841, "step": 10142 }, { "epoch": 0.367366896052155, "grad_norm": 1.2887655784095093, "learning_rate": 1.4601827740624501e-05, "loss": 0.773, "step": 10143 }, { "epoch": 0.3674031148134734, "grad_norm": 1.3757546058362273, "learning_rate": 1.4600786231053796e-05, "loss": 0.6873, "step": 10144 }, { "epoch": 0.36743933357479175, "grad_norm": 0.9980060708789887, "learning_rate": 1.4599744658172234e-05, "loss": 0.7559, "step": 10145 }, { "epoch": 0.3674755523361101, "grad_norm": 1.4825788698758895, "learning_rate": 1.459870302199414e-05, "loss": 0.8088, "step": 10146 }, { "epoch": 0.36751177109742844, "grad_norm": 1.329654101564198, "learning_rate": 1.4597661322533855e-05, "loss": 0.789, "step": 10147 }, { "epoch": 0.3675479898587468, "grad_norm": 1.3313714502010776, "learning_rate": 1.4596619559805708e-05, "loss": 0.8219, "step": 10148 }, { "epoch": 0.3675842086200652, "grad_norm": 1.385789885180584, "learning_rate": 1.4595577733824038e-05, "loss": 0.7864, "step": 10149 }, { "epoch": 0.36762042738138356, "grad_norm": 1.280414240388478, "learning_rate": 1.4594535844603179e-05, "loss": 0.7383, "step": 10150 }, { "epoch": 0.36765664614270194, "grad_norm": 1.350569610469433, "learning_rate": 1.4593493892157473e-05, "loss": 0.8001, "step": 10151 }, { "epoch": 0.36769286490402026, "grad_norm": 1.4181789342249946, "learning_rate": 1.4592451876501253e-05, "loss": 0.7854, "step": 10152 }, { "epoch": 0.36772908366533863, "grad_norm": 1.1845485006354253, "learning_rate": 1.4591409797648858e-05, "loss": 0.7641, "step": 10153 }, { "epoch": 0.367765302426657, "grad_norm": 1.2652801595365375, "learning_rate": 1.459036765561463e-05, "loss": 0.7628, "step": 10154 }, { "epoch": 0.3678015211879754, "grad_norm": 1.3121496013824596, "learning_rate": 1.4589325450412912e-05, "loss": 0.8085, "step": 10155 }, { "epoch": 0.36783773994929375, "grad_norm": 1.126054276417233, "learning_rate": 1.4588283182058042e-05, "loss": 0.7398, "step": 10156 }, { "epoch": 0.36787395871061207, "grad_norm": 1.343895710208696, "learning_rate": 1.4587240850564364e-05, "loss": 0.6945, "step": 10157 }, { "epoch": 0.36791017747193044, "grad_norm": 1.4260345325461032, "learning_rate": 1.4586198455946222e-05, "loss": 0.7606, "step": 10158 }, { "epoch": 0.3679463962332488, "grad_norm": 1.3672089452424008, "learning_rate": 1.4585155998217956e-05, "loss": 0.7494, "step": 10159 }, { "epoch": 0.3679826149945672, "grad_norm": 1.338101657178753, "learning_rate": 1.4584113477393919e-05, "loss": 0.8119, "step": 10160 }, { "epoch": 0.36801883375588557, "grad_norm": 1.3850069622601147, "learning_rate": 1.4583070893488451e-05, "loss": 0.7852, "step": 10161 }, { "epoch": 0.3680550525172039, "grad_norm": 1.4359454231022866, "learning_rate": 1.45820282465159e-05, "loss": 0.8341, "step": 10162 }, { "epoch": 0.36809127127852226, "grad_norm": 1.3737675415642494, "learning_rate": 1.4580985536490615e-05, "loss": 0.8614, "step": 10163 }, { "epoch": 0.36812749003984063, "grad_norm": 1.281457392114243, "learning_rate": 1.4579942763426944e-05, "loss": 0.8082, "step": 10164 }, { "epoch": 0.368163708801159, "grad_norm": 1.4254020526330675, "learning_rate": 1.4578899927339236e-05, "loss": 0.7969, "step": 10165 }, { "epoch": 0.3681999275624774, "grad_norm": 1.2212144463618084, "learning_rate": 1.4577857028241839e-05, "loss": 0.7292, "step": 10166 }, { "epoch": 0.3682361463237957, "grad_norm": 1.3945691800769362, "learning_rate": 1.4576814066149112e-05, "loss": 0.8794, "step": 10167 }, { "epoch": 0.3682723650851141, "grad_norm": 1.398660310385167, "learning_rate": 1.4575771041075397e-05, "loss": 0.8155, "step": 10168 }, { "epoch": 0.36830858384643245, "grad_norm": 1.3332443809888004, "learning_rate": 1.4574727953035055e-05, "loss": 0.8022, "step": 10169 }, { "epoch": 0.3683448026077508, "grad_norm": 1.4110409822250654, "learning_rate": 1.4573684802042433e-05, "loss": 0.7635, "step": 10170 }, { "epoch": 0.3683810213690692, "grad_norm": 1.3450081592052199, "learning_rate": 1.4572641588111891e-05, "loss": 0.7906, "step": 10171 }, { "epoch": 0.3684172401303875, "grad_norm": 1.0140395075506916, "learning_rate": 1.4571598311257783e-05, "loss": 0.7579, "step": 10172 }, { "epoch": 0.3684534588917059, "grad_norm": 1.1551281021574593, "learning_rate": 1.4570554971494466e-05, "loss": 0.7814, "step": 10173 }, { "epoch": 0.36848967765302426, "grad_norm": 1.426639565039773, "learning_rate": 1.4569511568836299e-05, "loss": 0.8546, "step": 10174 }, { "epoch": 0.36852589641434264, "grad_norm": 1.3420304444930535, "learning_rate": 1.4568468103297635e-05, "loss": 0.7914, "step": 10175 }, { "epoch": 0.368562115175661, "grad_norm": 1.4543167554514473, "learning_rate": 1.4567424574892836e-05, "loss": 0.8072, "step": 10176 }, { "epoch": 0.3685983339369793, "grad_norm": 1.4071036999269095, "learning_rate": 1.4566380983636262e-05, "loss": 0.8324, "step": 10177 }, { "epoch": 0.3686345526982977, "grad_norm": 1.0109694587038898, "learning_rate": 1.456533732954227e-05, "loss": 0.7255, "step": 10178 }, { "epoch": 0.3686707714596161, "grad_norm": 1.0914736042678337, "learning_rate": 1.4564293612625229e-05, "loss": 0.773, "step": 10179 }, { "epoch": 0.36870699022093445, "grad_norm": 1.5990898723320826, "learning_rate": 1.4563249832899495e-05, "loss": 0.7949, "step": 10180 }, { "epoch": 0.3687432089822528, "grad_norm": 1.3037574216863657, "learning_rate": 1.4562205990379434e-05, "loss": 0.7362, "step": 10181 }, { "epoch": 0.3687794277435712, "grad_norm": 0.9607879690846173, "learning_rate": 1.456116208507941e-05, "loss": 0.7617, "step": 10182 }, { "epoch": 0.3688156465048895, "grad_norm": 1.2702844322219347, "learning_rate": 1.456011811701379e-05, "loss": 0.7926, "step": 10183 }, { "epoch": 0.3688518652662079, "grad_norm": 0.9783750869045581, "learning_rate": 1.4559074086196935e-05, "loss": 0.7853, "step": 10184 }, { "epoch": 0.36888808402752626, "grad_norm": 1.3434278939341382, "learning_rate": 1.4558029992643216e-05, "loss": 0.688, "step": 10185 }, { "epoch": 0.36892430278884464, "grad_norm": 1.311195414887068, "learning_rate": 1.4556985836366999e-05, "loss": 0.761, "step": 10186 }, { "epoch": 0.368960521550163, "grad_norm": 0.9992392881630912, "learning_rate": 1.4555941617382653e-05, "loss": 0.7373, "step": 10187 }, { "epoch": 0.36899674031148133, "grad_norm": 1.4590610799324986, "learning_rate": 1.4554897335704548e-05, "loss": 0.8502, "step": 10188 }, { "epoch": 0.3690329590727997, "grad_norm": 1.2606323805742878, "learning_rate": 1.4553852991347053e-05, "loss": 0.7423, "step": 10189 }, { "epoch": 0.3690691778341181, "grad_norm": 1.41104288375632, "learning_rate": 1.4552808584324538e-05, "loss": 0.7797, "step": 10190 }, { "epoch": 0.36910539659543645, "grad_norm": 1.3640505709077357, "learning_rate": 1.455176411465138e-05, "loss": 0.7794, "step": 10191 }, { "epoch": 0.3691416153567548, "grad_norm": 1.2909958407953617, "learning_rate": 1.4550719582341944e-05, "loss": 0.7459, "step": 10192 }, { "epoch": 0.36917783411807314, "grad_norm": 1.3404406199564578, "learning_rate": 1.4549674987410607e-05, "loss": 0.6961, "step": 10193 }, { "epoch": 0.3692140528793915, "grad_norm": 2.070959364403057, "learning_rate": 1.4548630329871749e-05, "loss": 0.7543, "step": 10194 }, { "epoch": 0.3692502716407099, "grad_norm": 1.3738323287530774, "learning_rate": 1.4547585609739737e-05, "loss": 0.7748, "step": 10195 }, { "epoch": 0.36928649040202827, "grad_norm": 1.6547459079638813, "learning_rate": 1.4546540827028955e-05, "loss": 0.7342, "step": 10196 }, { "epoch": 0.36932270916334664, "grad_norm": 1.3216681814052256, "learning_rate": 1.4545495981753773e-05, "loss": 0.8047, "step": 10197 }, { "epoch": 0.36935892792466496, "grad_norm": 1.086211890800241, "learning_rate": 1.4544451073928574e-05, "loss": 0.7589, "step": 10198 }, { "epoch": 0.36939514668598333, "grad_norm": 1.2647213104785364, "learning_rate": 1.4543406103567735e-05, "loss": 0.7055, "step": 10199 }, { "epoch": 0.3694313654473017, "grad_norm": 1.4968034988002696, "learning_rate": 1.4542361070685636e-05, "loss": 0.7206, "step": 10200 }, { "epoch": 0.3694675842086201, "grad_norm": 1.5455348526483639, "learning_rate": 1.4541315975296655e-05, "loss": 0.7799, "step": 10201 }, { "epoch": 0.36950380296993846, "grad_norm": 1.5235755035665108, "learning_rate": 1.4540270817415176e-05, "loss": 0.7549, "step": 10202 }, { "epoch": 0.3695400217312568, "grad_norm": 1.528268853439445, "learning_rate": 1.4539225597055584e-05, "loss": 0.7992, "step": 10203 }, { "epoch": 0.36957624049257515, "grad_norm": 1.0525888359420206, "learning_rate": 1.4538180314232257e-05, "loss": 0.8011, "step": 10204 }, { "epoch": 0.3696124592538935, "grad_norm": 1.2245583837333305, "learning_rate": 1.4537134968959582e-05, "loss": 0.725, "step": 10205 }, { "epoch": 0.3696486780152119, "grad_norm": 1.3695871010178367, "learning_rate": 1.4536089561251945e-05, "loss": 0.8026, "step": 10206 }, { "epoch": 0.36968489677653027, "grad_norm": 1.018694798410934, "learning_rate": 1.453504409112373e-05, "loss": 0.8064, "step": 10207 }, { "epoch": 0.3697211155378486, "grad_norm": 1.3931072722496902, "learning_rate": 1.4533998558589319e-05, "loss": 0.7704, "step": 10208 }, { "epoch": 0.36975733429916696, "grad_norm": 1.3606349474635726, "learning_rate": 1.453295296366311e-05, "loss": 0.7984, "step": 10209 }, { "epoch": 0.36979355306048534, "grad_norm": 1.3844942060947245, "learning_rate": 1.453190730635948e-05, "loss": 0.7827, "step": 10210 }, { "epoch": 0.3698297718218037, "grad_norm": 1.4772949975259109, "learning_rate": 1.4530861586692826e-05, "loss": 0.707, "step": 10211 }, { "epoch": 0.3698659905831221, "grad_norm": 1.4858547577248946, "learning_rate": 1.4529815804677534e-05, "loss": 0.8186, "step": 10212 }, { "epoch": 0.3699022093444404, "grad_norm": 1.5881064902859185, "learning_rate": 1.4528769960327997e-05, "loss": 0.7704, "step": 10213 }, { "epoch": 0.3699384281057588, "grad_norm": 1.488650575879331, "learning_rate": 1.4527724053658608e-05, "loss": 0.8399, "step": 10214 }, { "epoch": 0.36997464686707715, "grad_norm": 1.0580964269040491, "learning_rate": 1.4526678084683755e-05, "loss": 0.821, "step": 10215 }, { "epoch": 0.3700108656283955, "grad_norm": 1.5572959114152944, "learning_rate": 1.4525632053417836e-05, "loss": 0.8182, "step": 10216 }, { "epoch": 0.3700470843897139, "grad_norm": 1.4009702379484648, "learning_rate": 1.452458595987524e-05, "loss": 0.767, "step": 10217 }, { "epoch": 0.3700833031510322, "grad_norm": 1.4417861086459696, "learning_rate": 1.4523539804070368e-05, "loss": 0.7741, "step": 10218 }, { "epoch": 0.3701195219123506, "grad_norm": 1.372370576414989, "learning_rate": 1.4522493586017611e-05, "loss": 0.7775, "step": 10219 }, { "epoch": 0.37015574067366896, "grad_norm": 1.4281386730574614, "learning_rate": 1.4521447305731374e-05, "loss": 0.86, "step": 10220 }, { "epoch": 0.37019195943498734, "grad_norm": 1.2953903636721602, "learning_rate": 1.4520400963226044e-05, "loss": 0.7081, "step": 10221 }, { "epoch": 0.3702281781963057, "grad_norm": 1.3824755364496804, "learning_rate": 1.4519354558516029e-05, "loss": 0.8048, "step": 10222 }, { "epoch": 0.37026439695762403, "grad_norm": 1.2082866747344163, "learning_rate": 1.451830809161572e-05, "loss": 0.7449, "step": 10223 }, { "epoch": 0.3703006157189424, "grad_norm": 1.4449379132688627, "learning_rate": 1.4517261562539523e-05, "loss": 0.8193, "step": 10224 }, { "epoch": 0.3703368344802608, "grad_norm": 1.4714481460186326, "learning_rate": 1.4516214971301839e-05, "loss": 0.7792, "step": 10225 }, { "epoch": 0.37037305324157915, "grad_norm": 1.3502848186758731, "learning_rate": 1.4515168317917066e-05, "loss": 0.7287, "step": 10226 }, { "epoch": 0.3704092720028975, "grad_norm": 1.4525949122720074, "learning_rate": 1.4514121602399613e-05, "loss": 0.7847, "step": 10227 }, { "epoch": 0.37044549076421585, "grad_norm": 1.3301143221193996, "learning_rate": 1.4513074824763877e-05, "loss": 0.738, "step": 10228 }, { "epoch": 0.3704817095255342, "grad_norm": 1.2720398971152185, "learning_rate": 1.4512027985024269e-05, "loss": 0.8203, "step": 10229 }, { "epoch": 0.3705179282868526, "grad_norm": 1.3814928585959734, "learning_rate": 1.4510981083195188e-05, "loss": 0.7928, "step": 10230 }, { "epoch": 0.37055414704817097, "grad_norm": 1.4254623023682524, "learning_rate": 1.4509934119291048e-05, "loss": 0.7831, "step": 10231 }, { "epoch": 0.37059036580948934, "grad_norm": 1.4454882482692515, "learning_rate": 1.4508887093326246e-05, "loss": 0.7336, "step": 10232 }, { "epoch": 0.37062658457080766, "grad_norm": 1.269499635302766, "learning_rate": 1.45078400053152e-05, "loss": 0.664, "step": 10233 }, { "epoch": 0.37066280333212603, "grad_norm": 1.1259594216659088, "learning_rate": 1.4506792855272314e-05, "loss": 0.7715, "step": 10234 }, { "epoch": 0.3706990220934444, "grad_norm": 1.0382501609313732, "learning_rate": 1.4505745643211997e-05, "loss": 0.6948, "step": 10235 }, { "epoch": 0.3707352408547628, "grad_norm": 1.1105600733103564, "learning_rate": 1.4504698369148664e-05, "loss": 0.7413, "step": 10236 }, { "epoch": 0.37077145961608116, "grad_norm": 1.0650728152611664, "learning_rate": 1.4503651033096719e-05, "loss": 0.8067, "step": 10237 }, { "epoch": 0.3708076783773995, "grad_norm": 1.4796844313457778, "learning_rate": 1.4502603635070583e-05, "loss": 0.7375, "step": 10238 }, { "epoch": 0.37084389713871785, "grad_norm": 1.4352519523875686, "learning_rate": 1.450155617508466e-05, "loss": 0.7712, "step": 10239 }, { "epoch": 0.3708801159000362, "grad_norm": 1.4149987193293891, "learning_rate": 1.4500508653153372e-05, "loss": 0.7752, "step": 10240 }, { "epoch": 0.3709163346613546, "grad_norm": 1.4690590663901026, "learning_rate": 1.4499461069291128e-05, "loss": 0.8025, "step": 10241 }, { "epoch": 0.37095255342267297, "grad_norm": 1.5851997012316403, "learning_rate": 1.4498413423512347e-05, "loss": 0.8683, "step": 10242 }, { "epoch": 0.3709887721839913, "grad_norm": 1.36488324876806, "learning_rate": 1.4497365715831446e-05, "loss": 0.7238, "step": 10243 }, { "epoch": 0.37102499094530966, "grad_norm": 1.2764488987936933, "learning_rate": 1.4496317946262839e-05, "loss": 0.693, "step": 10244 }, { "epoch": 0.37106120970662804, "grad_norm": 1.4286587999625553, "learning_rate": 1.4495270114820947e-05, "loss": 0.8036, "step": 10245 }, { "epoch": 0.3710974284679464, "grad_norm": 1.5183040632866114, "learning_rate": 1.449422222152019e-05, "loss": 0.8175, "step": 10246 }, { "epoch": 0.3711336472292648, "grad_norm": 1.4250101538764854, "learning_rate": 1.4493174266374985e-05, "loss": 0.7577, "step": 10247 }, { "epoch": 0.3711698659905831, "grad_norm": 1.5242158820820881, "learning_rate": 1.4492126249399752e-05, "loss": 0.7829, "step": 10248 }, { "epoch": 0.3712060847519015, "grad_norm": 1.4304129962847891, "learning_rate": 1.4491078170608916e-05, "loss": 0.8109, "step": 10249 }, { "epoch": 0.37124230351321985, "grad_norm": 1.4166238392785422, "learning_rate": 1.44900300300169e-05, "loss": 0.764, "step": 10250 }, { "epoch": 0.3712785222745382, "grad_norm": 1.5420391952568162, "learning_rate": 1.4488981827638125e-05, "loss": 0.7782, "step": 10251 }, { "epoch": 0.3713147410358566, "grad_norm": 1.0347688005220186, "learning_rate": 1.4487933563487014e-05, "loss": 0.7094, "step": 10252 }, { "epoch": 0.3713509597971749, "grad_norm": 1.4964242126989566, "learning_rate": 1.4486885237577994e-05, "loss": 0.7979, "step": 10253 }, { "epoch": 0.3713871785584933, "grad_norm": 1.4595021097804495, "learning_rate": 1.4485836849925492e-05, "loss": 0.8211, "step": 10254 }, { "epoch": 0.37142339731981167, "grad_norm": 1.40718532094145, "learning_rate": 1.448478840054393e-05, "loss": 0.8004, "step": 10255 }, { "epoch": 0.37145961608113004, "grad_norm": 1.1814148038849739, "learning_rate": 1.4483739889447743e-05, "loss": 0.7952, "step": 10256 }, { "epoch": 0.3714958348424484, "grad_norm": 1.2817659802395733, "learning_rate": 1.4482691316651354e-05, "loss": 0.8233, "step": 10257 }, { "epoch": 0.37153205360376673, "grad_norm": 1.02735829518677, "learning_rate": 1.4481642682169194e-05, "loss": 0.8083, "step": 10258 }, { "epoch": 0.3715682723650851, "grad_norm": 1.3369789846190074, "learning_rate": 1.4480593986015691e-05, "loss": 0.6963, "step": 10259 }, { "epoch": 0.3716044911264035, "grad_norm": 1.3279556895775293, "learning_rate": 1.4479545228205281e-05, "loss": 0.7403, "step": 10260 }, { "epoch": 0.37164070988772185, "grad_norm": 1.133541694207642, "learning_rate": 1.4478496408752389e-05, "loss": 0.8205, "step": 10261 }, { "epoch": 0.37167692864904023, "grad_norm": 1.3598482545471011, "learning_rate": 1.4477447527671453e-05, "loss": 0.8385, "step": 10262 }, { "epoch": 0.37171314741035855, "grad_norm": 1.1957407370559272, "learning_rate": 1.4476398584976905e-05, "loss": 0.7901, "step": 10263 }, { "epoch": 0.3717493661716769, "grad_norm": 1.4883510405529523, "learning_rate": 1.4475349580683179e-05, "loss": 0.8112, "step": 10264 }, { "epoch": 0.3717855849329953, "grad_norm": 1.2736487228061997, "learning_rate": 1.4474300514804709e-05, "loss": 0.7779, "step": 10265 }, { "epoch": 0.37182180369431367, "grad_norm": 1.0704331181033786, "learning_rate": 1.4473251387355936e-05, "loss": 0.7364, "step": 10266 }, { "epoch": 0.37185802245563204, "grad_norm": 1.2868249351691856, "learning_rate": 1.447220219835129e-05, "loss": 0.73, "step": 10267 }, { "epoch": 0.37189424121695036, "grad_norm": 1.409909570204525, "learning_rate": 1.4471152947805214e-05, "loss": 0.7004, "step": 10268 }, { "epoch": 0.37193045997826873, "grad_norm": 1.3850342042173258, "learning_rate": 1.4470103635732144e-05, "loss": 0.7532, "step": 10269 }, { "epoch": 0.3719666787395871, "grad_norm": 1.5776497106273975, "learning_rate": 1.446905426214652e-05, "loss": 0.7404, "step": 10270 }, { "epoch": 0.3720028975009055, "grad_norm": 1.166221522498181, "learning_rate": 1.4468004827062782e-05, "loss": 0.7986, "step": 10271 }, { "epoch": 0.37203911626222386, "grad_norm": 1.5249617727170817, "learning_rate": 1.4466955330495374e-05, "loss": 0.7882, "step": 10272 }, { "epoch": 0.3720753350235422, "grad_norm": 1.5783685434905277, "learning_rate": 1.4465905772458732e-05, "loss": 0.7406, "step": 10273 }, { "epoch": 0.37211155378486055, "grad_norm": 1.4406815861420517, "learning_rate": 1.4464856152967305e-05, "loss": 0.7572, "step": 10274 }, { "epoch": 0.3721477725461789, "grad_norm": 1.32601388080028, "learning_rate": 1.4463806472035532e-05, "loss": 0.758, "step": 10275 }, { "epoch": 0.3721839913074973, "grad_norm": 1.3537590989987491, "learning_rate": 1.4462756729677863e-05, "loss": 0.8236, "step": 10276 }, { "epoch": 0.37222021006881567, "grad_norm": 1.28433773678188, "learning_rate": 1.4461706925908735e-05, "loss": 0.6518, "step": 10277 }, { "epoch": 0.372256428830134, "grad_norm": 1.2738617694164343, "learning_rate": 1.4460657060742607e-05, "loss": 0.8765, "step": 10278 }, { "epoch": 0.37229264759145236, "grad_norm": 1.3439239816257629, "learning_rate": 1.445960713419391e-05, "loss": 0.7531, "step": 10279 }, { "epoch": 0.37232886635277074, "grad_norm": 1.6581808735732793, "learning_rate": 1.4458557146277106e-05, "loss": 0.8191, "step": 10280 }, { "epoch": 0.3723650851140891, "grad_norm": 1.0226403123337315, "learning_rate": 1.4457507097006636e-05, "loss": 0.7691, "step": 10281 }, { "epoch": 0.3724013038754075, "grad_norm": 1.41800147799089, "learning_rate": 1.4456456986396951e-05, "loss": 0.7875, "step": 10282 }, { "epoch": 0.3724375226367258, "grad_norm": 0.9953892073283614, "learning_rate": 1.4455406814462505e-05, "loss": 0.7433, "step": 10283 }, { "epoch": 0.3724737413980442, "grad_norm": 1.485869256617586, "learning_rate": 1.4454356581217743e-05, "loss": 0.7355, "step": 10284 }, { "epoch": 0.37250996015936255, "grad_norm": 1.265642609416718, "learning_rate": 1.4453306286677124e-05, "loss": 0.7984, "step": 10285 }, { "epoch": 0.3725461789206809, "grad_norm": 1.553986853958645, "learning_rate": 1.4452255930855095e-05, "loss": 0.7451, "step": 10286 }, { "epoch": 0.3725823976819993, "grad_norm": 1.2960904625212288, "learning_rate": 1.4451205513766113e-05, "loss": 0.7559, "step": 10287 }, { "epoch": 0.3726186164433176, "grad_norm": 1.2996676535419907, "learning_rate": 1.4450155035424632e-05, "loss": 0.7702, "step": 10288 }, { "epoch": 0.372654835204636, "grad_norm": 1.4202294646174616, "learning_rate": 1.4449104495845106e-05, "loss": 0.7723, "step": 10289 }, { "epoch": 0.37269105396595437, "grad_norm": 1.5515196361172, "learning_rate": 1.4448053895041993e-05, "loss": 0.764, "step": 10290 }, { "epoch": 0.37272727272727274, "grad_norm": 1.1121266582509872, "learning_rate": 1.444700323302975e-05, "loss": 0.7404, "step": 10291 }, { "epoch": 0.3727634914885911, "grad_norm": 1.4423887863500364, "learning_rate": 1.4445952509822836e-05, "loss": 0.7975, "step": 10292 }, { "epoch": 0.37279971024990943, "grad_norm": 1.2659583734846447, "learning_rate": 1.444490172543571e-05, "loss": 0.7485, "step": 10293 }, { "epoch": 0.3728359290112278, "grad_norm": 1.504727793141175, "learning_rate": 1.444385087988283e-05, "loss": 0.7784, "step": 10294 }, { "epoch": 0.3728721477725462, "grad_norm": 1.6345464913903207, "learning_rate": 1.4442799973178656e-05, "loss": 0.8073, "step": 10295 }, { "epoch": 0.37290836653386455, "grad_norm": 1.5239998122020226, "learning_rate": 1.4441749005337651e-05, "loss": 0.8419, "step": 10296 }, { "epoch": 0.37294458529518293, "grad_norm": 1.3996196530587812, "learning_rate": 1.4440697976374276e-05, "loss": 0.7597, "step": 10297 }, { "epoch": 0.37298080405650125, "grad_norm": 1.1910578730394434, "learning_rate": 1.4439646886302998e-05, "loss": 0.7606, "step": 10298 }, { "epoch": 0.3730170228178196, "grad_norm": 1.299031208768159, "learning_rate": 1.4438595735138274e-05, "loss": 0.7395, "step": 10299 }, { "epoch": 0.373053241579138, "grad_norm": 1.3931700149498816, "learning_rate": 1.4437544522894576e-05, "loss": 0.7747, "step": 10300 }, { "epoch": 0.37308946034045637, "grad_norm": 1.5723804775487282, "learning_rate": 1.4436493249586363e-05, "loss": 0.8027, "step": 10301 }, { "epoch": 0.37312567910177474, "grad_norm": 1.3645828886492444, "learning_rate": 1.4435441915228106e-05, "loss": 0.6961, "step": 10302 }, { "epoch": 0.37316189786309306, "grad_norm": 1.3828596723991167, "learning_rate": 1.4434390519834272e-05, "loss": 0.7477, "step": 10303 }, { "epoch": 0.37319811662441144, "grad_norm": 1.3759525789725089, "learning_rate": 1.4433339063419326e-05, "loss": 0.817, "step": 10304 }, { "epoch": 0.3732343353857298, "grad_norm": 1.3053009673883817, "learning_rate": 1.4432287545997741e-05, "loss": 0.7105, "step": 10305 }, { "epoch": 0.3732705541470482, "grad_norm": 1.4390874499126565, "learning_rate": 1.4431235967583983e-05, "loss": 0.7894, "step": 10306 }, { "epoch": 0.37330677290836656, "grad_norm": 1.4303157069820585, "learning_rate": 1.4430184328192527e-05, "loss": 0.7748, "step": 10307 }, { "epoch": 0.3733429916696849, "grad_norm": 1.4727156156620334, "learning_rate": 1.442913262783784e-05, "loss": 0.7151, "step": 10308 }, { "epoch": 0.37337921043100325, "grad_norm": 1.3356260129481994, "learning_rate": 1.4428080866534397e-05, "loss": 0.8363, "step": 10309 }, { "epoch": 0.3734154291923216, "grad_norm": 1.3756053483387785, "learning_rate": 1.4427029044296669e-05, "loss": 0.7542, "step": 10310 }, { "epoch": 0.37345164795364, "grad_norm": 1.5475396103692722, "learning_rate": 1.4425977161139134e-05, "loss": 0.8373, "step": 10311 }, { "epoch": 0.37348786671495837, "grad_norm": 1.230877663487242, "learning_rate": 1.4424925217076262e-05, "loss": 0.8125, "step": 10312 }, { "epoch": 0.3735240854762767, "grad_norm": 1.4072158807776642, "learning_rate": 1.442387321212253e-05, "loss": 0.7823, "step": 10313 }, { "epoch": 0.37356030423759506, "grad_norm": 1.3714219356709874, "learning_rate": 1.4422821146292418e-05, "loss": 0.7354, "step": 10314 }, { "epoch": 0.37359652299891344, "grad_norm": 1.3261929890970139, "learning_rate": 1.44217690196004e-05, "loss": 0.778, "step": 10315 }, { "epoch": 0.3736327417602318, "grad_norm": 1.2910927491228605, "learning_rate": 1.4420716832060957e-05, "loss": 0.7032, "step": 10316 }, { "epoch": 0.3736689605215502, "grad_norm": 1.195704909189034, "learning_rate": 1.4419664583688563e-05, "loss": 0.7634, "step": 10317 }, { "epoch": 0.3737051792828685, "grad_norm": 1.3506768831096414, "learning_rate": 1.44186122744977e-05, "loss": 0.7202, "step": 10318 }, { "epoch": 0.3737413980441869, "grad_norm": 1.277332684267741, "learning_rate": 1.4417559904502852e-05, "loss": 0.7507, "step": 10319 }, { "epoch": 0.37377761680550525, "grad_norm": 1.3564679649938773, "learning_rate": 1.44165074737185e-05, "loss": 0.7361, "step": 10320 }, { "epoch": 0.3738138355668236, "grad_norm": 1.3761648451327524, "learning_rate": 1.4415454982159121e-05, "loss": 0.7054, "step": 10321 }, { "epoch": 0.373850054328142, "grad_norm": 1.1416057092727552, "learning_rate": 1.4414402429839201e-05, "loss": 0.718, "step": 10322 }, { "epoch": 0.3738862730894603, "grad_norm": 1.3571482296800685, "learning_rate": 1.4413349816773228e-05, "loss": 0.8, "step": 10323 }, { "epoch": 0.3739224918507787, "grad_norm": 1.2885338255465852, "learning_rate": 1.441229714297568e-05, "loss": 0.8234, "step": 10324 }, { "epoch": 0.37395871061209707, "grad_norm": 1.063419725297814, "learning_rate": 1.441124440846105e-05, "loss": 0.7848, "step": 10325 }, { "epoch": 0.37399492937341544, "grad_norm": 1.4459436559247223, "learning_rate": 1.4410191613243817e-05, "loss": 0.7509, "step": 10326 }, { "epoch": 0.3740311481347338, "grad_norm": 1.4344035753507873, "learning_rate": 1.4409138757338477e-05, "loss": 0.8854, "step": 10327 }, { "epoch": 0.37406736689605213, "grad_norm": 0.9900629313220529, "learning_rate": 1.4408085840759509e-05, "loss": 0.7083, "step": 10328 }, { "epoch": 0.3741035856573705, "grad_norm": 0.9779247474374213, "learning_rate": 1.4407032863521411e-05, "loss": 0.7649, "step": 10329 }, { "epoch": 0.3741398044186889, "grad_norm": 1.4424279876742276, "learning_rate": 1.4405979825638663e-05, "loss": 0.8659, "step": 10330 }, { "epoch": 0.37417602318000726, "grad_norm": 1.4724690864783783, "learning_rate": 1.4404926727125768e-05, "loss": 0.8603, "step": 10331 }, { "epoch": 0.37421224194132563, "grad_norm": 0.9978647387214801, "learning_rate": 1.4403873567997205e-05, "loss": 0.7034, "step": 10332 }, { "epoch": 0.37424846070264395, "grad_norm": 1.3851849300461974, "learning_rate": 1.4402820348267473e-05, "loss": 0.8197, "step": 10333 }, { "epoch": 0.3742846794639623, "grad_norm": 1.365974484201299, "learning_rate": 1.4401767067951066e-05, "loss": 0.8447, "step": 10334 }, { "epoch": 0.3743208982252807, "grad_norm": 1.5577351743681815, "learning_rate": 1.4400713727062475e-05, "loss": 0.8878, "step": 10335 }, { "epoch": 0.37435711698659907, "grad_norm": 1.6026869702121478, "learning_rate": 1.4399660325616198e-05, "loss": 0.7554, "step": 10336 }, { "epoch": 0.37439333574791744, "grad_norm": 1.350620629285001, "learning_rate": 1.4398606863626727e-05, "loss": 0.7309, "step": 10337 }, { "epoch": 0.37442955450923576, "grad_norm": 1.3020980823794133, "learning_rate": 1.4397553341108563e-05, "loss": 0.7345, "step": 10338 }, { "epoch": 0.37446577327055414, "grad_norm": 1.5075032813971976, "learning_rate": 1.43964997580762e-05, "loss": 0.8389, "step": 10339 }, { "epoch": 0.3745019920318725, "grad_norm": 1.3125250933155743, "learning_rate": 1.4395446114544135e-05, "loss": 0.826, "step": 10340 }, { "epoch": 0.3745382107931909, "grad_norm": 1.5151990975396075, "learning_rate": 1.4394392410526871e-05, "loss": 0.8283, "step": 10341 }, { "epoch": 0.37457442955450926, "grad_norm": 1.5303417840531266, "learning_rate": 1.4393338646038906e-05, "loss": 0.8102, "step": 10342 }, { "epoch": 0.3746106483158276, "grad_norm": 1.3822138434689097, "learning_rate": 1.4392284821094742e-05, "loss": 0.7894, "step": 10343 }, { "epoch": 0.37464686707714595, "grad_norm": 1.434509711027516, "learning_rate": 1.4391230935708878e-05, "loss": 0.8248, "step": 10344 }, { "epoch": 0.3746830858384643, "grad_norm": 1.3894468779231768, "learning_rate": 1.439017698989582e-05, "loss": 0.7068, "step": 10345 }, { "epoch": 0.3747193045997827, "grad_norm": 1.4376438317696951, "learning_rate": 1.4389122983670066e-05, "loss": 0.8097, "step": 10346 }, { "epoch": 0.3747555233611011, "grad_norm": 1.2716456803901406, "learning_rate": 1.4388068917046125e-05, "loss": 0.7677, "step": 10347 }, { "epoch": 0.3747917421224194, "grad_norm": 1.4184491671493566, "learning_rate": 1.43870147900385e-05, "loss": 0.7479, "step": 10348 }, { "epoch": 0.37482796088373777, "grad_norm": 1.3846510703908526, "learning_rate": 1.4385960602661697e-05, "loss": 0.7482, "step": 10349 }, { "epoch": 0.37486417964505614, "grad_norm": 1.3462278031170276, "learning_rate": 1.4384906354930222e-05, "loss": 0.8405, "step": 10350 }, { "epoch": 0.3749003984063745, "grad_norm": 1.201682330717079, "learning_rate": 1.4383852046858584e-05, "loss": 0.775, "step": 10351 }, { "epoch": 0.3749366171676929, "grad_norm": 1.3959690406929284, "learning_rate": 1.4382797678461287e-05, "loss": 0.779, "step": 10352 }, { "epoch": 0.3749728359290112, "grad_norm": 1.3926395433479462, "learning_rate": 1.4381743249752842e-05, "loss": 0.8209, "step": 10353 }, { "epoch": 0.3750090546903296, "grad_norm": 1.43659932618761, "learning_rate": 1.4380688760747762e-05, "loss": 0.8224, "step": 10354 }, { "epoch": 0.37504527345164795, "grad_norm": 1.0308694854295093, "learning_rate": 1.4379634211460556e-05, "loss": 0.7286, "step": 10355 }, { "epoch": 0.3750814922129663, "grad_norm": 1.1745457854955943, "learning_rate": 1.4378579601905735e-05, "loss": 0.778, "step": 10356 }, { "epoch": 0.3751177109742847, "grad_norm": 1.6628894768268043, "learning_rate": 1.4377524932097811e-05, "loss": 0.8181, "step": 10357 }, { "epoch": 0.375153929735603, "grad_norm": 1.5830409201255058, "learning_rate": 1.4376470202051298e-05, "loss": 0.8744, "step": 10358 }, { "epoch": 0.3751901484969214, "grad_norm": 1.0526560310130255, "learning_rate": 1.4375415411780706e-05, "loss": 0.782, "step": 10359 }, { "epoch": 0.37522636725823977, "grad_norm": 1.4357767453041528, "learning_rate": 1.4374360561300558e-05, "loss": 0.736, "step": 10360 }, { "epoch": 0.37526258601955814, "grad_norm": 1.600560865968103, "learning_rate": 1.4373305650625362e-05, "loss": 0.7684, "step": 10361 }, { "epoch": 0.3752988047808765, "grad_norm": 1.1811733080378202, "learning_rate": 1.4372250679769639e-05, "loss": 0.7538, "step": 10362 }, { "epoch": 0.37533502354219483, "grad_norm": 1.3719289712802352, "learning_rate": 1.4371195648747905e-05, "loss": 0.7137, "step": 10363 }, { "epoch": 0.3753712423035132, "grad_norm": 1.2278058953955913, "learning_rate": 1.4370140557574677e-05, "loss": 0.7856, "step": 10364 }, { "epoch": 0.3754074610648316, "grad_norm": 1.471319876049015, "learning_rate": 1.4369085406264476e-05, "loss": 0.8034, "step": 10365 }, { "epoch": 0.37544367982614996, "grad_norm": 1.260044141779808, "learning_rate": 1.4368030194831819e-05, "loss": 0.7683, "step": 10366 }, { "epoch": 0.37547989858746833, "grad_norm": 1.3980693685561225, "learning_rate": 1.436697492329123e-05, "loss": 0.8244, "step": 10367 }, { "epoch": 0.37551611734878665, "grad_norm": 1.306225802058267, "learning_rate": 1.436591959165723e-05, "loss": 0.6834, "step": 10368 }, { "epoch": 0.375552336110105, "grad_norm": 1.483561176014766, "learning_rate": 1.4364864199944336e-05, "loss": 0.9186, "step": 10369 }, { "epoch": 0.3755885548714234, "grad_norm": 1.3577275995350726, "learning_rate": 1.4363808748167079e-05, "loss": 0.8147, "step": 10370 }, { "epoch": 0.37562477363274177, "grad_norm": 1.0675912702142303, "learning_rate": 1.4362753236339979e-05, "loss": 0.731, "step": 10371 }, { "epoch": 0.37566099239406014, "grad_norm": 1.405110696378493, "learning_rate": 1.436169766447756e-05, "loss": 0.763, "step": 10372 }, { "epoch": 0.37569721115537846, "grad_norm": 1.44591407628293, "learning_rate": 1.4360642032594349e-05, "loss": 0.7747, "step": 10373 }, { "epoch": 0.37573342991669684, "grad_norm": 1.1535785274555292, "learning_rate": 1.4359586340704873e-05, "loss": 0.6702, "step": 10374 }, { "epoch": 0.3757696486780152, "grad_norm": 1.498213573466521, "learning_rate": 1.4358530588823656e-05, "loss": 0.7867, "step": 10375 }, { "epoch": 0.3758058674393336, "grad_norm": 1.2143575719509156, "learning_rate": 1.435747477696523e-05, "loss": 0.7084, "step": 10376 }, { "epoch": 0.37584208620065196, "grad_norm": 1.4042371019579993, "learning_rate": 1.4356418905144123e-05, "loss": 0.7248, "step": 10377 }, { "epoch": 0.3758783049619703, "grad_norm": 1.4414038994105745, "learning_rate": 1.4355362973374863e-05, "loss": 0.6452, "step": 10378 }, { "epoch": 0.37591452372328865, "grad_norm": 1.5083964753721644, "learning_rate": 1.4354306981671982e-05, "loss": 0.7593, "step": 10379 }, { "epoch": 0.375950742484607, "grad_norm": 1.0408754985805013, "learning_rate": 1.435325093005001e-05, "loss": 0.7135, "step": 10380 }, { "epoch": 0.3759869612459254, "grad_norm": 1.2780780795410114, "learning_rate": 1.435219481852348e-05, "loss": 0.7396, "step": 10381 }, { "epoch": 0.3760231800072438, "grad_norm": 1.3999985481205717, "learning_rate": 1.4351138647106923e-05, "loss": 0.6957, "step": 10382 }, { "epoch": 0.3760593987685621, "grad_norm": 1.0587562765306413, "learning_rate": 1.4350082415814878e-05, "loss": 0.7144, "step": 10383 }, { "epoch": 0.37609561752988047, "grad_norm": 1.3852329144989983, "learning_rate": 1.4349026124661878e-05, "loss": 0.7197, "step": 10384 }, { "epoch": 0.37613183629119884, "grad_norm": 1.3555627338191216, "learning_rate": 1.4347969773662453e-05, "loss": 0.8125, "step": 10385 }, { "epoch": 0.3761680550525172, "grad_norm": 1.4063008948229103, "learning_rate": 1.4346913362831147e-05, "loss": 0.7975, "step": 10386 }, { "epoch": 0.3762042738138356, "grad_norm": 1.2856488950360478, "learning_rate": 1.4345856892182492e-05, "loss": 0.8359, "step": 10387 }, { "epoch": 0.3762404925751539, "grad_norm": 1.1290800861432029, "learning_rate": 1.4344800361731028e-05, "loss": 0.744, "step": 10388 }, { "epoch": 0.3762767113364723, "grad_norm": 1.3964963112167614, "learning_rate": 1.4343743771491292e-05, "loss": 0.7484, "step": 10389 }, { "epoch": 0.37631293009779065, "grad_norm": 1.501249694296133, "learning_rate": 1.4342687121477827e-05, "loss": 0.8516, "step": 10390 }, { "epoch": 0.37634914885910903, "grad_norm": 1.3621209304369741, "learning_rate": 1.4341630411705169e-05, "loss": 0.7737, "step": 10391 }, { "epoch": 0.3763853676204274, "grad_norm": 1.009344719087584, "learning_rate": 1.4340573642187865e-05, "loss": 0.7699, "step": 10392 }, { "epoch": 0.3764215863817457, "grad_norm": 1.3144220763701364, "learning_rate": 1.433951681294045e-05, "loss": 0.7865, "step": 10393 }, { "epoch": 0.3764578051430641, "grad_norm": 1.429436622558147, "learning_rate": 1.4338459923977474e-05, "loss": 0.8076, "step": 10394 }, { "epoch": 0.37649402390438247, "grad_norm": 1.4166051199913192, "learning_rate": 1.4337402975313475e-05, "loss": 0.8038, "step": 10395 }, { "epoch": 0.37653024266570084, "grad_norm": 1.3358419332565319, "learning_rate": 1.4336345966963e-05, "loss": 0.804, "step": 10396 }, { "epoch": 0.3765664614270192, "grad_norm": 1.453290506904961, "learning_rate": 1.4335288898940595e-05, "loss": 0.7834, "step": 10397 }, { "epoch": 0.37660268018833754, "grad_norm": 1.3537678504200228, "learning_rate": 1.4334231771260807e-05, "loss": 0.7447, "step": 10398 }, { "epoch": 0.3766388989496559, "grad_norm": 1.410774804606614, "learning_rate": 1.4333174583938179e-05, "loss": 0.7776, "step": 10399 }, { "epoch": 0.3766751177109743, "grad_norm": 1.4668801686578228, "learning_rate": 1.4332117336987262e-05, "loss": 0.7776, "step": 10400 }, { "epoch": 0.37671133647229266, "grad_norm": 1.1633036220680484, "learning_rate": 1.4331060030422606e-05, "loss": 0.7995, "step": 10401 }, { "epoch": 0.37674755523361103, "grad_norm": 1.3044003246247546, "learning_rate": 1.4330002664258755e-05, "loss": 0.7335, "step": 10402 }, { "epoch": 0.37678377399492935, "grad_norm": 1.2668246048578067, "learning_rate": 1.4328945238510264e-05, "loss": 0.7151, "step": 10403 }, { "epoch": 0.3768199927562477, "grad_norm": 1.391873671045427, "learning_rate": 1.4327887753191683e-05, "loss": 0.7596, "step": 10404 }, { "epoch": 0.3768562115175661, "grad_norm": 1.4605763391837054, "learning_rate": 1.4326830208317564e-05, "loss": 0.8187, "step": 10405 }, { "epoch": 0.37689243027888447, "grad_norm": 1.104417028281472, "learning_rate": 1.4325772603902458e-05, "loss": 0.7691, "step": 10406 }, { "epoch": 0.37692864904020285, "grad_norm": 1.2472492121809202, "learning_rate": 1.4324714939960922e-05, "loss": 0.7021, "step": 10407 }, { "epoch": 0.37696486780152116, "grad_norm": 1.4656182746488895, "learning_rate": 1.4323657216507507e-05, "loss": 0.8357, "step": 10408 }, { "epoch": 0.37700108656283954, "grad_norm": 1.5745928393219424, "learning_rate": 1.432259943355677e-05, "loss": 0.7909, "step": 10409 }, { "epoch": 0.3770373053241579, "grad_norm": 1.3901677525134384, "learning_rate": 1.4321541591123268e-05, "loss": 0.7641, "step": 10410 }, { "epoch": 0.3770735240854763, "grad_norm": 1.2090696203101519, "learning_rate": 1.432048368922155e-05, "loss": 0.7651, "step": 10411 }, { "epoch": 0.37710974284679466, "grad_norm": 1.4590548302977282, "learning_rate": 1.431942572786619e-05, "loss": 0.7118, "step": 10412 }, { "epoch": 0.377145961608113, "grad_norm": 1.2617509098944182, "learning_rate": 1.431836770707173e-05, "loss": 0.7359, "step": 10413 }, { "epoch": 0.37718218036943135, "grad_norm": 1.3878305276688938, "learning_rate": 1.4317309626852739e-05, "loss": 0.8141, "step": 10414 }, { "epoch": 0.3772183991307497, "grad_norm": 1.3279202534828196, "learning_rate": 1.4316251487223771e-05, "loss": 0.8129, "step": 10415 }, { "epoch": 0.3772546178920681, "grad_norm": 1.261097148889029, "learning_rate": 1.4315193288199392e-05, "loss": 0.7289, "step": 10416 }, { "epoch": 0.3772908366533865, "grad_norm": 1.0503481001862314, "learning_rate": 1.431413502979416e-05, "loss": 0.7614, "step": 10417 }, { "epoch": 0.3773270554147048, "grad_norm": 1.4340003499819716, "learning_rate": 1.4313076712022642e-05, "loss": 0.7296, "step": 10418 }, { "epoch": 0.37736327417602317, "grad_norm": 1.3265299910458699, "learning_rate": 1.4312018334899398e-05, "loss": 0.7829, "step": 10419 }, { "epoch": 0.37739949293734154, "grad_norm": 1.1944346642491355, "learning_rate": 1.4310959898438989e-05, "loss": 0.7677, "step": 10420 }, { "epoch": 0.3774357116986599, "grad_norm": 1.299038416825305, "learning_rate": 1.430990140265599e-05, "loss": 0.6952, "step": 10421 }, { "epoch": 0.3774719304599783, "grad_norm": 1.3241217072693154, "learning_rate": 1.4308842847564956e-05, "loss": 0.7685, "step": 10422 }, { "epoch": 0.3775081492212966, "grad_norm": 1.2338387301102218, "learning_rate": 1.4307784233180462e-05, "loss": 0.8, "step": 10423 }, { "epoch": 0.377544367982615, "grad_norm": 1.3610763132195627, "learning_rate": 1.430672555951707e-05, "loss": 0.7777, "step": 10424 }, { "epoch": 0.37758058674393336, "grad_norm": 1.3492344502483704, "learning_rate": 1.4305666826589348e-05, "loss": 0.7472, "step": 10425 }, { "epoch": 0.37761680550525173, "grad_norm": 1.4363518008785456, "learning_rate": 1.430460803441187e-05, "loss": 0.7639, "step": 10426 }, { "epoch": 0.3776530242665701, "grad_norm": 1.2351825291255705, "learning_rate": 1.4303549182999202e-05, "loss": 0.8085, "step": 10427 }, { "epoch": 0.3776892430278884, "grad_norm": 1.4891938930614719, "learning_rate": 1.4302490272365916e-05, "loss": 0.794, "step": 10428 }, { "epoch": 0.3777254617892068, "grad_norm": 1.3703742013935212, "learning_rate": 1.4301431302526582e-05, "loss": 0.825, "step": 10429 }, { "epoch": 0.37776168055052517, "grad_norm": 1.2715416820637917, "learning_rate": 1.4300372273495774e-05, "loss": 0.79, "step": 10430 }, { "epoch": 0.37779789931184354, "grad_norm": 1.1213689670413414, "learning_rate": 1.4299313185288067e-05, "loss": 0.7269, "step": 10431 }, { "epoch": 0.3778341180731619, "grad_norm": 1.6046692186928158, "learning_rate": 1.4298254037918034e-05, "loss": 0.8293, "step": 10432 }, { "epoch": 0.37787033683448024, "grad_norm": 1.3139130997861987, "learning_rate": 1.4297194831400246e-05, "loss": 0.7727, "step": 10433 }, { "epoch": 0.3779065555957986, "grad_norm": 1.150543701300476, "learning_rate": 1.429613556574928e-05, "loss": 0.7428, "step": 10434 }, { "epoch": 0.377942774357117, "grad_norm": 1.4351061546696744, "learning_rate": 1.4295076240979716e-05, "loss": 0.771, "step": 10435 }, { "epoch": 0.37797899311843536, "grad_norm": 1.2984788252867967, "learning_rate": 1.4294016857106128e-05, "loss": 0.7943, "step": 10436 }, { "epoch": 0.37801521187975373, "grad_norm": 1.7837300310774784, "learning_rate": 1.4292957414143096e-05, "loss": 0.8248, "step": 10437 }, { "epoch": 0.37805143064107205, "grad_norm": 1.1322495801712111, "learning_rate": 1.4291897912105198e-05, "loss": 0.7526, "step": 10438 }, { "epoch": 0.3780876494023904, "grad_norm": 1.5484303562555748, "learning_rate": 1.429083835100701e-05, "loss": 0.7044, "step": 10439 }, { "epoch": 0.3781238681637088, "grad_norm": 1.0396496779253301, "learning_rate": 1.428977873086312e-05, "loss": 0.8102, "step": 10440 }, { "epoch": 0.3781600869250272, "grad_norm": 1.5934016230304155, "learning_rate": 1.4288719051688105e-05, "loss": 0.7691, "step": 10441 }, { "epoch": 0.37819630568634555, "grad_norm": 1.106186578835786, "learning_rate": 1.4287659313496545e-05, "loss": 0.8272, "step": 10442 }, { "epoch": 0.37823252444766386, "grad_norm": 1.3323767289537785, "learning_rate": 1.4286599516303028e-05, "loss": 0.7894, "step": 10443 }, { "epoch": 0.37826874320898224, "grad_norm": 1.263188747340934, "learning_rate": 1.4285539660122135e-05, "loss": 0.6839, "step": 10444 }, { "epoch": 0.3783049619703006, "grad_norm": 1.4107417511837004, "learning_rate": 1.428447974496845e-05, "loss": 0.7973, "step": 10445 }, { "epoch": 0.378341180731619, "grad_norm": 1.0441436400764224, "learning_rate": 1.4283419770856557e-05, "loss": 0.76, "step": 10446 }, { "epoch": 0.37837739949293736, "grad_norm": 1.5414805780578316, "learning_rate": 1.4282359737801046e-05, "loss": 0.8489, "step": 10447 }, { "epoch": 0.3784136182542557, "grad_norm": 2.09764811000985, "learning_rate": 1.4281299645816502e-05, "loss": 0.8128, "step": 10448 }, { "epoch": 0.37844983701557405, "grad_norm": 1.0227863554827503, "learning_rate": 1.4280239494917513e-05, "loss": 0.7674, "step": 10449 }, { "epoch": 0.3784860557768924, "grad_norm": 1.4362932237270565, "learning_rate": 1.4279179285118667e-05, "loss": 0.7183, "step": 10450 }, { "epoch": 0.3785222745382108, "grad_norm": 1.3154141592257598, "learning_rate": 1.4278119016434555e-05, "loss": 0.7877, "step": 10451 }, { "epoch": 0.3785584932995292, "grad_norm": 1.2907779026755788, "learning_rate": 1.4277058688879768e-05, "loss": 0.7367, "step": 10452 }, { "epoch": 0.3785947120608475, "grad_norm": 2.2783280612439345, "learning_rate": 1.427599830246889e-05, "loss": 0.7383, "step": 10453 }, { "epoch": 0.37863093082216587, "grad_norm": 1.4306953947522605, "learning_rate": 1.4274937857216524e-05, "loss": 0.8046, "step": 10454 }, { "epoch": 0.37866714958348424, "grad_norm": 1.3101793090321407, "learning_rate": 1.4273877353137253e-05, "loss": 0.7471, "step": 10455 }, { "epoch": 0.3787033683448026, "grad_norm": 1.138633045745799, "learning_rate": 1.4272816790245675e-05, "loss": 0.7676, "step": 10456 }, { "epoch": 0.378739587106121, "grad_norm": 1.3333340591438623, "learning_rate": 1.4271756168556383e-05, "loss": 0.7057, "step": 10457 }, { "epoch": 0.3787758058674393, "grad_norm": 1.6846412002464983, "learning_rate": 1.4270695488083976e-05, "loss": 0.8075, "step": 10458 }, { "epoch": 0.3788120246287577, "grad_norm": 1.8039747560527235, "learning_rate": 1.4269634748843044e-05, "loss": 0.7416, "step": 10459 }, { "epoch": 0.37884824339007606, "grad_norm": 1.4233890459510925, "learning_rate": 1.4268573950848187e-05, "loss": 0.8018, "step": 10460 }, { "epoch": 0.37888446215139443, "grad_norm": 1.0648083654011513, "learning_rate": 1.4267513094114001e-05, "loss": 0.7592, "step": 10461 }, { "epoch": 0.3789206809127128, "grad_norm": 1.4957798617342686, "learning_rate": 1.4266452178655083e-05, "loss": 0.7955, "step": 10462 }, { "epoch": 0.3789568996740311, "grad_norm": 1.4769970289806116, "learning_rate": 1.4265391204486037e-05, "loss": 0.7472, "step": 10463 }, { "epoch": 0.3789931184353495, "grad_norm": 1.0683333490285838, "learning_rate": 1.4264330171621459e-05, "loss": 0.8236, "step": 10464 }, { "epoch": 0.37902933719666787, "grad_norm": 1.5116083230487665, "learning_rate": 1.4263269080075952e-05, "loss": 0.7744, "step": 10465 }, { "epoch": 0.37906555595798624, "grad_norm": 1.3436871641460189, "learning_rate": 1.4262207929864116e-05, "loss": 0.7718, "step": 10466 }, { "epoch": 0.3791017747193046, "grad_norm": 1.472613847562251, "learning_rate": 1.4261146721000554e-05, "loss": 0.797, "step": 10467 }, { "epoch": 0.37913799348062294, "grad_norm": 1.195222077745974, "learning_rate": 1.4260085453499868e-05, "loss": 0.7441, "step": 10468 }, { "epoch": 0.3791742122419413, "grad_norm": 1.3412910525546637, "learning_rate": 1.4259024127376665e-05, "loss": 0.7193, "step": 10469 }, { "epoch": 0.3792104310032597, "grad_norm": 1.4870490814453392, "learning_rate": 1.4257962742645544e-05, "loss": 0.8235, "step": 10470 }, { "epoch": 0.37924664976457806, "grad_norm": 1.4764458595103493, "learning_rate": 1.4256901299321118e-05, "loss": 0.7969, "step": 10471 }, { "epoch": 0.37928286852589643, "grad_norm": 1.457127899001861, "learning_rate": 1.4255839797417987e-05, "loss": 0.8014, "step": 10472 }, { "epoch": 0.37931908728721475, "grad_norm": 1.1962270599957692, "learning_rate": 1.4254778236950761e-05, "loss": 0.7368, "step": 10473 }, { "epoch": 0.3793553060485331, "grad_norm": 1.2267029990832308, "learning_rate": 1.425371661793405e-05, "loss": 0.8013, "step": 10474 }, { "epoch": 0.3793915248098515, "grad_norm": 1.3431825749701174, "learning_rate": 1.425265494038246e-05, "loss": 0.7081, "step": 10475 }, { "epoch": 0.3794277435711699, "grad_norm": 1.4833272988474804, "learning_rate": 1.4251593204310601e-05, "loss": 0.8024, "step": 10476 }, { "epoch": 0.37946396233248825, "grad_norm": 1.3823488759462892, "learning_rate": 1.4250531409733084e-05, "loss": 0.7086, "step": 10477 }, { "epoch": 0.37950018109380657, "grad_norm": 1.4243965139475798, "learning_rate": 1.4249469556664518e-05, "loss": 0.7641, "step": 10478 }, { "epoch": 0.37953639985512494, "grad_norm": 1.1141292834207037, "learning_rate": 1.4248407645119522e-05, "loss": 0.7315, "step": 10479 }, { "epoch": 0.3795726186164433, "grad_norm": 1.305769107247934, "learning_rate": 1.4247345675112697e-05, "loss": 0.7408, "step": 10480 }, { "epoch": 0.3796088373777617, "grad_norm": 1.4014108591219767, "learning_rate": 1.424628364665867e-05, "loss": 0.7986, "step": 10481 }, { "epoch": 0.37964505613908006, "grad_norm": 1.1839415624219254, "learning_rate": 1.4245221559772045e-05, "loss": 0.6594, "step": 10482 }, { "epoch": 0.3796812749003984, "grad_norm": 1.2664992210965693, "learning_rate": 1.4244159414467442e-05, "loss": 0.6957, "step": 10483 }, { "epoch": 0.37971749366171675, "grad_norm": 1.4011441548404797, "learning_rate": 1.4243097210759476e-05, "loss": 0.6877, "step": 10484 }, { "epoch": 0.37975371242303513, "grad_norm": 1.3744800286171677, "learning_rate": 1.4242034948662764e-05, "loss": 0.8093, "step": 10485 }, { "epoch": 0.3797899311843535, "grad_norm": 1.0221585976245882, "learning_rate": 1.4240972628191924e-05, "loss": 0.7491, "step": 10486 }, { "epoch": 0.3798261499456719, "grad_norm": 1.0415015281532956, "learning_rate": 1.4239910249361575e-05, "loss": 0.7269, "step": 10487 }, { "epoch": 0.3798623687069902, "grad_norm": 0.9706915630692035, "learning_rate": 1.4238847812186334e-05, "loss": 0.7634, "step": 10488 }, { "epoch": 0.37989858746830857, "grad_norm": 1.3646740773261323, "learning_rate": 1.4237785316680821e-05, "loss": 0.8001, "step": 10489 }, { "epoch": 0.37993480622962694, "grad_norm": 0.9988519109048412, "learning_rate": 1.4236722762859662e-05, "loss": 0.7787, "step": 10490 }, { "epoch": 0.3799710249909453, "grad_norm": 1.3700696837713406, "learning_rate": 1.4235660150737474e-05, "loss": 0.7971, "step": 10491 }, { "epoch": 0.3800072437522637, "grad_norm": 1.4325945474698847, "learning_rate": 1.423459748032888e-05, "loss": 0.7888, "step": 10492 }, { "epoch": 0.380043462513582, "grad_norm": 1.2752995349421565, "learning_rate": 1.4233534751648504e-05, "loss": 0.7625, "step": 10493 }, { "epoch": 0.3800796812749004, "grad_norm": 1.4643456343282109, "learning_rate": 1.423247196471097e-05, "loss": 0.8026, "step": 10494 }, { "epoch": 0.38011590003621876, "grad_norm": 1.4331093328316633, "learning_rate": 1.4231409119530901e-05, "loss": 0.7867, "step": 10495 }, { "epoch": 0.38015211879753713, "grad_norm": 1.1020007490793542, "learning_rate": 1.4230346216122927e-05, "loss": 0.7987, "step": 10496 }, { "epoch": 0.3801883375588555, "grad_norm": 1.4821976013309888, "learning_rate": 1.4229283254501672e-05, "loss": 0.7746, "step": 10497 }, { "epoch": 0.3802245563201738, "grad_norm": 1.3216844997484696, "learning_rate": 1.422822023468176e-05, "loss": 0.7724, "step": 10498 }, { "epoch": 0.3802607750814922, "grad_norm": 1.389859985634129, "learning_rate": 1.4227157156677826e-05, "loss": 0.7644, "step": 10499 }, { "epoch": 0.38029699384281057, "grad_norm": 1.227614876887612, "learning_rate": 1.4226094020504495e-05, "loss": 0.7306, "step": 10500 }, { "epoch": 0.38033321260412895, "grad_norm": 1.222996494715474, "learning_rate": 1.4225030826176396e-05, "loss": 0.7543, "step": 10501 }, { "epoch": 0.3803694313654473, "grad_norm": 1.5740582725096284, "learning_rate": 1.4223967573708159e-05, "loss": 0.8354, "step": 10502 }, { "epoch": 0.38040565012676564, "grad_norm": 1.3456099835068212, "learning_rate": 1.4222904263114419e-05, "loss": 0.7461, "step": 10503 }, { "epoch": 0.380441868888084, "grad_norm": 1.0209514790264265, "learning_rate": 1.4221840894409803e-05, "loss": 0.742, "step": 10504 }, { "epoch": 0.3804780876494024, "grad_norm": 1.3837475955540788, "learning_rate": 1.422077746760895e-05, "loss": 0.7206, "step": 10505 }, { "epoch": 0.38051430641072076, "grad_norm": 1.6143955741941813, "learning_rate": 1.4219713982726488e-05, "loss": 0.7901, "step": 10506 }, { "epoch": 0.38055052517203913, "grad_norm": 1.036477301763879, "learning_rate": 1.4218650439777055e-05, "loss": 0.7629, "step": 10507 }, { "epoch": 0.38058674393335745, "grad_norm": 1.2878838005266922, "learning_rate": 1.4217586838775283e-05, "loss": 0.7214, "step": 10508 }, { "epoch": 0.3806229626946758, "grad_norm": 1.196488019007988, "learning_rate": 1.4216523179735812e-05, "loss": 0.7274, "step": 10509 }, { "epoch": 0.3806591814559942, "grad_norm": 1.3180484998597004, "learning_rate": 1.4215459462673278e-05, "loss": 0.7593, "step": 10510 }, { "epoch": 0.3806954002173126, "grad_norm": 1.1482023806159773, "learning_rate": 1.4214395687602316e-05, "loss": 0.7495, "step": 10511 }, { "epoch": 0.38073161897863095, "grad_norm": 1.4353760533326885, "learning_rate": 1.4213331854537566e-05, "loss": 0.8294, "step": 10512 }, { "epoch": 0.38076783773994927, "grad_norm": 1.3545723006491877, "learning_rate": 1.421226796349367e-05, "loss": 0.8404, "step": 10513 }, { "epoch": 0.38080405650126764, "grad_norm": 1.0555340237326092, "learning_rate": 1.4211204014485264e-05, "loss": 0.7648, "step": 10514 }, { "epoch": 0.380840275262586, "grad_norm": 1.1032236365958645, "learning_rate": 1.421014000752699e-05, "loss": 0.7188, "step": 10515 }, { "epoch": 0.3808764940239044, "grad_norm": 1.4903726751065745, "learning_rate": 1.420907594263349e-05, "loss": 0.7868, "step": 10516 }, { "epoch": 0.38091271278522276, "grad_norm": 1.5807005760795627, "learning_rate": 1.4208011819819407e-05, "loss": 0.7838, "step": 10517 }, { "epoch": 0.3809489315465411, "grad_norm": 1.4270746387149535, "learning_rate": 1.4206947639099386e-05, "loss": 0.7465, "step": 10518 }, { "epoch": 0.38098515030785945, "grad_norm": 1.395380216521995, "learning_rate": 1.4205883400488066e-05, "loss": 0.7597, "step": 10519 }, { "epoch": 0.38102136906917783, "grad_norm": 1.6794224152865145, "learning_rate": 1.4204819104000093e-05, "loss": 0.7166, "step": 10520 }, { "epoch": 0.3810575878304962, "grad_norm": 1.4680387297806252, "learning_rate": 1.420375474965012e-05, "loss": 0.8372, "step": 10521 }, { "epoch": 0.3810938065918146, "grad_norm": 1.0885623337274903, "learning_rate": 1.4202690337452784e-05, "loss": 0.7544, "step": 10522 }, { "epoch": 0.3811300253531329, "grad_norm": 1.0405991426334422, "learning_rate": 1.420162586742274e-05, "loss": 0.7326, "step": 10523 }, { "epoch": 0.38116624411445127, "grad_norm": 1.542830584170694, "learning_rate": 1.4200561339574629e-05, "loss": 0.8201, "step": 10524 }, { "epoch": 0.38120246287576964, "grad_norm": 1.311832396558115, "learning_rate": 1.4199496753923104e-05, "loss": 0.7753, "step": 10525 }, { "epoch": 0.381238681637088, "grad_norm": 1.0411727326658946, "learning_rate": 1.4198432110482812e-05, "loss": 0.7798, "step": 10526 }, { "epoch": 0.3812749003984064, "grad_norm": 1.391535058021882, "learning_rate": 1.4197367409268407e-05, "loss": 0.7853, "step": 10527 }, { "epoch": 0.38131111915972477, "grad_norm": 1.485464521899415, "learning_rate": 1.4196302650294538e-05, "loss": 0.7524, "step": 10528 }, { "epoch": 0.3813473379210431, "grad_norm": 1.0367524777569768, "learning_rate": 1.4195237833575854e-05, "loss": 0.7584, "step": 10529 }, { "epoch": 0.38138355668236146, "grad_norm": 1.5761435410459355, "learning_rate": 1.4194172959127018e-05, "loss": 0.7409, "step": 10530 }, { "epoch": 0.38141977544367983, "grad_norm": 1.430909573837949, "learning_rate": 1.419310802696267e-05, "loss": 0.717, "step": 10531 }, { "epoch": 0.3814559942049982, "grad_norm": 1.7435650660779722, "learning_rate": 1.4192043037097473e-05, "loss": 0.7229, "step": 10532 }, { "epoch": 0.3814922129663166, "grad_norm": 1.1714114238817919, "learning_rate": 1.4190977989546082e-05, "loss": 0.775, "step": 10533 }, { "epoch": 0.3815284317276349, "grad_norm": 1.052548242488854, "learning_rate": 1.4189912884323152e-05, "loss": 0.714, "step": 10534 }, { "epoch": 0.38156465048895327, "grad_norm": 1.3306013687159988, "learning_rate": 1.4188847721443337e-05, "loss": 0.7618, "step": 10535 }, { "epoch": 0.38160086925027165, "grad_norm": 1.3809444867980512, "learning_rate": 1.41877825009213e-05, "loss": 0.786, "step": 10536 }, { "epoch": 0.38163708801159, "grad_norm": 1.4467621962455677, "learning_rate": 1.418671722277169e-05, "loss": 0.7856, "step": 10537 }, { "epoch": 0.3816733067729084, "grad_norm": 1.4530605191619228, "learning_rate": 1.418565188700918e-05, "loss": 0.8154, "step": 10538 }, { "epoch": 0.3817095255342267, "grad_norm": 1.1809464867021031, "learning_rate": 1.4184586493648416e-05, "loss": 0.7517, "step": 10539 }, { "epoch": 0.3817457442955451, "grad_norm": 1.3697906045099741, "learning_rate": 1.4183521042704067e-05, "loss": 0.7826, "step": 10540 }, { "epoch": 0.38178196305686346, "grad_norm": 1.2621072706683452, "learning_rate": 1.4182455534190793e-05, "loss": 0.7604, "step": 10541 }, { "epoch": 0.38181818181818183, "grad_norm": 1.413324894436786, "learning_rate": 1.4181389968123254e-05, "loss": 0.7666, "step": 10542 }, { "epoch": 0.3818544005795002, "grad_norm": 1.3547523643326864, "learning_rate": 1.4180324344516119e-05, "loss": 0.7453, "step": 10543 }, { "epoch": 0.3818906193408185, "grad_norm": 1.3887070022303634, "learning_rate": 1.4179258663384043e-05, "loss": 0.815, "step": 10544 }, { "epoch": 0.3819268381021369, "grad_norm": 1.4466611459542207, "learning_rate": 1.4178192924741699e-05, "loss": 0.7604, "step": 10545 }, { "epoch": 0.3819630568634553, "grad_norm": 1.2951294774286526, "learning_rate": 1.4177127128603748e-05, "loss": 0.7561, "step": 10546 }, { "epoch": 0.38199927562477365, "grad_norm": 1.3254543612069147, "learning_rate": 1.4176061274984858e-05, "loss": 0.7871, "step": 10547 }, { "epoch": 0.382035494386092, "grad_norm": 1.3504511268729908, "learning_rate": 1.4174995363899693e-05, "loss": 0.8264, "step": 10548 }, { "epoch": 0.38207171314741034, "grad_norm": 1.1311605368308175, "learning_rate": 1.4173929395362926e-05, "loss": 0.7836, "step": 10549 }, { "epoch": 0.3821079319087287, "grad_norm": 1.1906333890660674, "learning_rate": 1.417286336938922e-05, "loss": 0.6722, "step": 10550 }, { "epoch": 0.3821441506700471, "grad_norm": 1.380334078410873, "learning_rate": 1.417179728599325e-05, "loss": 0.8419, "step": 10551 }, { "epoch": 0.38218036943136546, "grad_norm": 1.2870927593970565, "learning_rate": 1.4170731145189685e-05, "loss": 0.678, "step": 10552 }, { "epoch": 0.38221658819268384, "grad_norm": 1.4729955709805007, "learning_rate": 1.4169664946993192e-05, "loss": 0.8088, "step": 10553 }, { "epoch": 0.38225280695400216, "grad_norm": 1.4346337190797929, "learning_rate": 1.4168598691418444e-05, "loss": 0.8002, "step": 10554 }, { "epoch": 0.38228902571532053, "grad_norm": 1.4212414697770228, "learning_rate": 1.416753237848012e-05, "loss": 0.6652, "step": 10555 }, { "epoch": 0.3823252444766389, "grad_norm": 1.300007224693305, "learning_rate": 1.4166466008192885e-05, "loss": 0.7626, "step": 10556 }, { "epoch": 0.3823614632379573, "grad_norm": 1.3278318320002978, "learning_rate": 1.4165399580571417e-05, "loss": 0.7445, "step": 10557 }, { "epoch": 0.38239768199927565, "grad_norm": 1.4150541263934233, "learning_rate": 1.4164333095630394e-05, "loss": 0.7072, "step": 10558 }, { "epoch": 0.38243390076059397, "grad_norm": 1.0610126087226108, "learning_rate": 1.4163266553384484e-05, "loss": 0.7091, "step": 10559 }, { "epoch": 0.38247011952191234, "grad_norm": 1.4428864996228021, "learning_rate": 1.416219995384837e-05, "loss": 0.7793, "step": 10560 }, { "epoch": 0.3825063382832307, "grad_norm": 1.441309580623915, "learning_rate": 1.4161133297036729e-05, "loss": 0.7416, "step": 10561 }, { "epoch": 0.3825425570445491, "grad_norm": 1.364215179793025, "learning_rate": 1.4160066582964233e-05, "loss": 0.826, "step": 10562 }, { "epoch": 0.38257877580586747, "grad_norm": 1.3812075766730716, "learning_rate": 1.4158999811645571e-05, "loss": 0.7433, "step": 10563 }, { "epoch": 0.3826149945671858, "grad_norm": 1.4239942482528425, "learning_rate": 1.4157932983095415e-05, "loss": 0.7402, "step": 10564 }, { "epoch": 0.38265121332850416, "grad_norm": 1.347125855714339, "learning_rate": 1.4156866097328448e-05, "loss": 0.7521, "step": 10565 }, { "epoch": 0.38268743208982253, "grad_norm": 1.204263640453092, "learning_rate": 1.4155799154359349e-05, "loss": 0.7398, "step": 10566 }, { "epoch": 0.3827236508511409, "grad_norm": 1.2651194751996906, "learning_rate": 1.4154732154202803e-05, "loss": 0.8047, "step": 10567 }, { "epoch": 0.3827598696124593, "grad_norm": 1.3666125968911786, "learning_rate": 1.4153665096873492e-05, "loss": 0.7454, "step": 10568 }, { "epoch": 0.3827960883737776, "grad_norm": 1.0080396882227987, "learning_rate": 1.41525979823861e-05, "loss": 0.734, "step": 10569 }, { "epoch": 0.382832307135096, "grad_norm": 1.4617317235066112, "learning_rate": 1.4151530810755309e-05, "loss": 0.771, "step": 10570 }, { "epoch": 0.38286852589641435, "grad_norm": 0.9736011729197931, "learning_rate": 1.4150463581995811e-05, "loss": 0.777, "step": 10571 }, { "epoch": 0.3829047446577327, "grad_norm": 1.407236856982812, "learning_rate": 1.4149396296122282e-05, "loss": 0.7985, "step": 10572 }, { "epoch": 0.3829409634190511, "grad_norm": 1.3546490045597614, "learning_rate": 1.4148328953149416e-05, "loss": 0.796, "step": 10573 }, { "epoch": 0.3829771821803694, "grad_norm": 1.5694726066684765, "learning_rate": 1.41472615530919e-05, "loss": 0.8761, "step": 10574 }, { "epoch": 0.3830134009416878, "grad_norm": 1.2805479702465703, "learning_rate": 1.4146194095964419e-05, "loss": 0.8171, "step": 10575 }, { "epoch": 0.38304961970300616, "grad_norm": 1.0884001416615827, "learning_rate": 1.4145126581781665e-05, "loss": 0.7131, "step": 10576 }, { "epoch": 0.38308583846432454, "grad_norm": 1.0700717691414616, "learning_rate": 1.4144059010558327e-05, "loss": 0.7421, "step": 10577 }, { "epoch": 0.3831220572256429, "grad_norm": 1.385291323931291, "learning_rate": 1.4142991382309097e-05, "loss": 0.6909, "step": 10578 }, { "epoch": 0.3831582759869612, "grad_norm": 1.2813532935932062, "learning_rate": 1.4141923697048663e-05, "loss": 0.738, "step": 10579 }, { "epoch": 0.3831944947482796, "grad_norm": 1.3145064069299022, "learning_rate": 1.4140855954791719e-05, "loss": 0.7592, "step": 10580 }, { "epoch": 0.383230713509598, "grad_norm": 1.4722622640343146, "learning_rate": 1.4139788155552961e-05, "loss": 0.7789, "step": 10581 }, { "epoch": 0.38326693227091635, "grad_norm": 1.3067567528523185, "learning_rate": 1.413872029934708e-05, "loss": 0.7269, "step": 10582 }, { "epoch": 0.3833031510322347, "grad_norm": 1.1558791083038564, "learning_rate": 1.4137652386188769e-05, "loss": 0.8237, "step": 10583 }, { "epoch": 0.38333936979355304, "grad_norm": 1.3442954506098643, "learning_rate": 1.4136584416092727e-05, "loss": 0.795, "step": 10584 }, { "epoch": 0.3833755885548714, "grad_norm": 1.2065752594658816, "learning_rate": 1.4135516389073648e-05, "loss": 0.7878, "step": 10585 }, { "epoch": 0.3834118073161898, "grad_norm": 1.4539905573616265, "learning_rate": 1.4134448305146231e-05, "loss": 0.8848, "step": 10586 }, { "epoch": 0.38344802607750816, "grad_norm": 1.1156967461622151, "learning_rate": 1.413338016432517e-05, "loss": 0.7131, "step": 10587 }, { "epoch": 0.38348424483882654, "grad_norm": 1.5799384131983323, "learning_rate": 1.413231196662517e-05, "loss": 0.7839, "step": 10588 }, { "epoch": 0.38352046360014486, "grad_norm": 1.389787681718074, "learning_rate": 1.4131243712060922e-05, "loss": 0.7684, "step": 10589 }, { "epoch": 0.38355668236146323, "grad_norm": 1.277659713239619, "learning_rate": 1.4130175400647135e-05, "loss": 0.7083, "step": 10590 }, { "epoch": 0.3835929011227816, "grad_norm": 1.3990533228015751, "learning_rate": 1.4129107032398501e-05, "loss": 0.8155, "step": 10591 }, { "epoch": 0.3836291198841, "grad_norm": 0.9762464028129457, "learning_rate": 1.4128038607329728e-05, "loss": 0.7965, "step": 10592 }, { "epoch": 0.38366533864541835, "grad_norm": 1.2567090316464518, "learning_rate": 1.4126970125455517e-05, "loss": 0.7903, "step": 10593 }, { "epoch": 0.38370155740673667, "grad_norm": 1.4547438576800207, "learning_rate": 1.412590158679057e-05, "loss": 0.8087, "step": 10594 }, { "epoch": 0.38373777616805504, "grad_norm": 1.399443085709582, "learning_rate": 1.4124832991349592e-05, "loss": 0.784, "step": 10595 }, { "epoch": 0.3837739949293734, "grad_norm": 1.4954120350239675, "learning_rate": 1.412376433914729e-05, "loss": 0.7526, "step": 10596 }, { "epoch": 0.3838102136906918, "grad_norm": 1.3233005630454473, "learning_rate": 1.4122695630198365e-05, "loss": 0.7084, "step": 10597 }, { "epoch": 0.38384643245201017, "grad_norm": 1.3829146168336974, "learning_rate": 1.4121626864517523e-05, "loss": 0.7883, "step": 10598 }, { "epoch": 0.3838826512133285, "grad_norm": 1.5157127060043472, "learning_rate": 1.4120558042119479e-05, "loss": 0.8835, "step": 10599 }, { "epoch": 0.38391886997464686, "grad_norm": 1.2704411283170896, "learning_rate": 1.4119489163018932e-05, "loss": 0.7397, "step": 10600 }, { "epoch": 0.38395508873596523, "grad_norm": 1.272479237961347, "learning_rate": 1.4118420227230596e-05, "loss": 0.751, "step": 10601 }, { "epoch": 0.3839913074972836, "grad_norm": 1.405272856944841, "learning_rate": 1.4117351234769178e-05, "loss": 0.7442, "step": 10602 }, { "epoch": 0.384027526258602, "grad_norm": 1.3155909696400225, "learning_rate": 1.411628218564939e-05, "loss": 0.7217, "step": 10603 }, { "epoch": 0.3840637450199203, "grad_norm": 1.2698089061231264, "learning_rate": 1.4115213079885941e-05, "loss": 0.7632, "step": 10604 }, { "epoch": 0.3840999637812387, "grad_norm": 1.3272134457596925, "learning_rate": 1.4114143917493546e-05, "loss": 0.7357, "step": 10605 }, { "epoch": 0.38413618254255705, "grad_norm": 1.3267440355198552, "learning_rate": 1.4113074698486915e-05, "loss": 0.7697, "step": 10606 }, { "epoch": 0.3841724013038754, "grad_norm": 1.037849477811535, "learning_rate": 1.4112005422880762e-05, "loss": 0.7607, "step": 10607 }, { "epoch": 0.3842086200651938, "grad_norm": 1.2954741382003716, "learning_rate": 1.41109360906898e-05, "loss": 0.829, "step": 10608 }, { "epoch": 0.3842448388265121, "grad_norm": 1.2331906001597144, "learning_rate": 1.4109866701928745e-05, "loss": 0.7921, "step": 10609 }, { "epoch": 0.3842810575878305, "grad_norm": 1.0797748343020162, "learning_rate": 1.4108797256612317e-05, "loss": 0.7415, "step": 10610 }, { "epoch": 0.38431727634914886, "grad_norm": 1.4287710770356452, "learning_rate": 1.4107727754755222e-05, "loss": 0.7892, "step": 10611 }, { "epoch": 0.38435349511046724, "grad_norm": 1.160697272965598, "learning_rate": 1.410665819637219e-05, "loss": 0.7877, "step": 10612 }, { "epoch": 0.3843897138717856, "grad_norm": 1.2609119271470581, "learning_rate": 1.410558858147793e-05, "loss": 0.7865, "step": 10613 }, { "epoch": 0.38442593263310393, "grad_norm": 1.4104112793414774, "learning_rate": 1.4104518910087164e-05, "loss": 0.8152, "step": 10614 }, { "epoch": 0.3844621513944223, "grad_norm": 1.5015851278118442, "learning_rate": 1.410344918221461e-05, "loss": 0.7481, "step": 10615 }, { "epoch": 0.3844983701557407, "grad_norm": 1.31986293742988, "learning_rate": 1.4102379397874992e-05, "loss": 0.7707, "step": 10616 }, { "epoch": 0.38453458891705905, "grad_norm": 1.381242752811355, "learning_rate": 1.4101309557083027e-05, "loss": 0.7929, "step": 10617 }, { "epoch": 0.3845708076783774, "grad_norm": 1.3461030274534742, "learning_rate": 1.4100239659853439e-05, "loss": 0.7627, "step": 10618 }, { "epoch": 0.38460702643969574, "grad_norm": 1.4439473548295778, "learning_rate": 1.409916970620095e-05, "loss": 0.7925, "step": 10619 }, { "epoch": 0.3846432452010141, "grad_norm": 1.4102257066882617, "learning_rate": 1.4098099696140284e-05, "loss": 0.7878, "step": 10620 }, { "epoch": 0.3846794639623325, "grad_norm": 1.3626566989278124, "learning_rate": 1.4097029629686167e-05, "loss": 0.7778, "step": 10621 }, { "epoch": 0.38471568272365086, "grad_norm": 1.3597125156455874, "learning_rate": 1.4095959506853323e-05, "loss": 0.7996, "step": 10622 }, { "epoch": 0.38475190148496924, "grad_norm": 1.3445213136300274, "learning_rate": 1.4094889327656475e-05, "loss": 0.7463, "step": 10623 }, { "epoch": 0.38478812024628756, "grad_norm": 1.2733564749637354, "learning_rate": 1.4093819092110353e-05, "loss": 0.7859, "step": 10624 }, { "epoch": 0.38482433900760593, "grad_norm": 1.5187309696384723, "learning_rate": 1.4092748800229684e-05, "loss": 0.8955, "step": 10625 }, { "epoch": 0.3848605577689243, "grad_norm": 1.126369831334453, "learning_rate": 1.4091678452029192e-05, "loss": 0.734, "step": 10626 }, { "epoch": 0.3848967765302427, "grad_norm": 1.1025391144494854, "learning_rate": 1.409060804752361e-05, "loss": 0.7736, "step": 10627 }, { "epoch": 0.38493299529156105, "grad_norm": 1.5466969488726774, "learning_rate": 1.4089537586727668e-05, "loss": 0.81, "step": 10628 }, { "epoch": 0.38496921405287937, "grad_norm": 1.3256049320776124, "learning_rate": 1.4088467069656095e-05, "loss": 0.6949, "step": 10629 }, { "epoch": 0.38500543281419775, "grad_norm": 1.3966280723212952, "learning_rate": 1.4087396496323625e-05, "loss": 0.7713, "step": 10630 }, { "epoch": 0.3850416515755161, "grad_norm": 1.2925589558888175, "learning_rate": 1.4086325866744984e-05, "loss": 0.7744, "step": 10631 }, { "epoch": 0.3850778703368345, "grad_norm": 1.392542040064444, "learning_rate": 1.4085255180934912e-05, "loss": 0.7504, "step": 10632 }, { "epoch": 0.38511408909815287, "grad_norm": 1.5393207739983596, "learning_rate": 1.4084184438908136e-05, "loss": 0.8037, "step": 10633 }, { "epoch": 0.3851503078594712, "grad_norm": 1.3919190737247333, "learning_rate": 1.4083113640679395e-05, "loss": 0.7636, "step": 10634 }, { "epoch": 0.38518652662078956, "grad_norm": 1.3713361832803506, "learning_rate": 1.408204278626342e-05, "loss": 0.7819, "step": 10635 }, { "epoch": 0.38522274538210793, "grad_norm": 1.4721175797222736, "learning_rate": 1.4080971875674957e-05, "loss": 0.8072, "step": 10636 }, { "epoch": 0.3852589641434263, "grad_norm": 1.2982175582218396, "learning_rate": 1.4079900908928727e-05, "loss": 0.7783, "step": 10637 }, { "epoch": 0.3852951829047447, "grad_norm": 1.120113834045166, "learning_rate": 1.4078829886039478e-05, "loss": 0.7822, "step": 10638 }, { "epoch": 0.385331401666063, "grad_norm": 1.4310457657819275, "learning_rate": 1.4077758807021946e-05, "loss": 0.7716, "step": 10639 }, { "epoch": 0.3853676204273814, "grad_norm": 1.3764083948160888, "learning_rate": 1.407668767189087e-05, "loss": 0.7314, "step": 10640 }, { "epoch": 0.38540383918869975, "grad_norm": 1.4201712172965313, "learning_rate": 1.4075616480660989e-05, "loss": 0.7124, "step": 10641 }, { "epoch": 0.3854400579500181, "grad_norm": 1.6611206058724246, "learning_rate": 1.4074545233347045e-05, "loss": 0.7154, "step": 10642 }, { "epoch": 0.3854762767113365, "grad_norm": 1.3621522360103278, "learning_rate": 1.407347392996378e-05, "loss": 0.7629, "step": 10643 }, { "epoch": 0.3855124954726548, "grad_norm": 1.3499364139231609, "learning_rate": 1.4072402570525931e-05, "loss": 0.782, "step": 10644 }, { "epoch": 0.3855487142339732, "grad_norm": 1.134625377940359, "learning_rate": 1.4071331155048246e-05, "loss": 0.7296, "step": 10645 }, { "epoch": 0.38558493299529156, "grad_norm": 1.3266384542780036, "learning_rate": 1.4070259683545469e-05, "loss": 0.8004, "step": 10646 }, { "epoch": 0.38562115175660994, "grad_norm": 1.384707722710989, "learning_rate": 1.406918815603234e-05, "loss": 0.8025, "step": 10647 }, { "epoch": 0.3856573705179283, "grad_norm": 1.4462527094077289, "learning_rate": 1.4068116572523605e-05, "loss": 0.7508, "step": 10648 }, { "epoch": 0.38569358927924663, "grad_norm": 1.389304690225081, "learning_rate": 1.4067044933034013e-05, "loss": 0.7078, "step": 10649 }, { "epoch": 0.385729808040565, "grad_norm": 1.2707102912574109, "learning_rate": 1.406597323757831e-05, "loss": 0.7682, "step": 10650 }, { "epoch": 0.3857660268018834, "grad_norm": 1.2829703870388185, "learning_rate": 1.406490148617124e-05, "loss": 0.7517, "step": 10651 }, { "epoch": 0.38580224556320175, "grad_norm": 1.2943523090789832, "learning_rate": 1.4063829678827556e-05, "loss": 0.7314, "step": 10652 }, { "epoch": 0.3858384643245201, "grad_norm": 1.254472815853873, "learning_rate": 1.4062757815562005e-05, "loss": 0.7373, "step": 10653 }, { "epoch": 0.38587468308583844, "grad_norm": 1.5285882891612308, "learning_rate": 1.4061685896389337e-05, "loss": 0.7526, "step": 10654 }, { "epoch": 0.3859109018471568, "grad_norm": 1.2875226866589016, "learning_rate": 1.40606139213243e-05, "loss": 0.6816, "step": 10655 }, { "epoch": 0.3859471206084752, "grad_norm": 1.3661070082687696, "learning_rate": 1.405954189038165e-05, "loss": 0.7693, "step": 10656 }, { "epoch": 0.38598333936979357, "grad_norm": 1.4277558429476473, "learning_rate": 1.4058469803576133e-05, "loss": 0.8348, "step": 10657 }, { "epoch": 0.38601955813111194, "grad_norm": 1.3778865369478832, "learning_rate": 1.405739766092251e-05, "loss": 0.7431, "step": 10658 }, { "epoch": 0.38605577689243026, "grad_norm": 1.0302614841852513, "learning_rate": 1.4056325462435525e-05, "loss": 0.6997, "step": 10659 }, { "epoch": 0.38609199565374863, "grad_norm": 1.5796021851613187, "learning_rate": 1.405525320812994e-05, "loss": 0.7695, "step": 10660 }, { "epoch": 0.386128214415067, "grad_norm": 1.3512085702020764, "learning_rate": 1.405418089802051e-05, "loss": 0.7997, "step": 10661 }, { "epoch": 0.3861644331763854, "grad_norm": 1.3344628214691812, "learning_rate": 1.4053108532121984e-05, "loss": 0.7157, "step": 10662 }, { "epoch": 0.38620065193770375, "grad_norm": 1.4471950977144945, "learning_rate": 1.4052036110449126e-05, "loss": 0.7894, "step": 10663 }, { "epoch": 0.3862368706990221, "grad_norm": 1.225358643295336, "learning_rate": 1.4050963633016687e-05, "loss": 0.7879, "step": 10664 }, { "epoch": 0.38627308946034045, "grad_norm": 1.456128304690939, "learning_rate": 1.4049891099839434e-05, "loss": 0.7243, "step": 10665 }, { "epoch": 0.3863093082216588, "grad_norm": 1.299673526660825, "learning_rate": 1.4048818510932118e-05, "loss": 0.7836, "step": 10666 }, { "epoch": 0.3863455269829772, "grad_norm": 1.437712490402559, "learning_rate": 1.4047745866309504e-05, "loss": 0.7788, "step": 10667 }, { "epoch": 0.38638174574429557, "grad_norm": 1.2896538952843832, "learning_rate": 1.4046673165986348e-05, "loss": 0.726, "step": 10668 }, { "epoch": 0.3864179645056139, "grad_norm": 1.594075293044659, "learning_rate": 1.4045600409977414e-05, "loss": 0.7117, "step": 10669 }, { "epoch": 0.38645418326693226, "grad_norm": 1.44888026919397, "learning_rate": 1.4044527598297463e-05, "loss": 0.8569, "step": 10670 }, { "epoch": 0.38649040202825063, "grad_norm": 1.1261778829440297, "learning_rate": 1.404345473096126e-05, "loss": 0.788, "step": 10671 }, { "epoch": 0.386526620789569, "grad_norm": 1.3402626818728194, "learning_rate": 1.4042381807983567e-05, "loss": 0.7643, "step": 10672 }, { "epoch": 0.3865628395508874, "grad_norm": 1.246019080146076, "learning_rate": 1.4041308829379148e-05, "loss": 0.7822, "step": 10673 }, { "epoch": 0.3865990583122057, "grad_norm": 1.0357117107280363, "learning_rate": 1.404023579516277e-05, "loss": 0.7414, "step": 10674 }, { "epoch": 0.3866352770735241, "grad_norm": 1.2986979448449105, "learning_rate": 1.4039162705349193e-05, "loss": 0.7924, "step": 10675 }, { "epoch": 0.38667149583484245, "grad_norm": 1.403016911720975, "learning_rate": 1.4038089559953192e-05, "loss": 0.7744, "step": 10676 }, { "epoch": 0.3867077145961608, "grad_norm": 1.3654355718120015, "learning_rate": 1.4037016358989527e-05, "loss": 0.744, "step": 10677 }, { "epoch": 0.3867439333574792, "grad_norm": 1.6468482379078004, "learning_rate": 1.4035943102472973e-05, "loss": 0.7718, "step": 10678 }, { "epoch": 0.3867801521187975, "grad_norm": 1.3575717740344733, "learning_rate": 1.4034869790418295e-05, "loss": 0.7741, "step": 10679 }, { "epoch": 0.3868163708801159, "grad_norm": 1.2268576342864124, "learning_rate": 1.4033796422840262e-05, "loss": 0.7161, "step": 10680 }, { "epoch": 0.38685258964143426, "grad_norm": 1.4038865951760717, "learning_rate": 1.4032722999753647e-05, "loss": 0.793, "step": 10681 }, { "epoch": 0.38688880840275264, "grad_norm": 1.570177079524853, "learning_rate": 1.4031649521173218e-05, "loss": 0.8657, "step": 10682 }, { "epoch": 0.386925027164071, "grad_norm": 1.4316735205312336, "learning_rate": 1.4030575987113753e-05, "loss": 0.8024, "step": 10683 }, { "epoch": 0.38696124592538933, "grad_norm": 1.3253372034897586, "learning_rate": 1.4029502397590019e-05, "loss": 0.8002, "step": 10684 }, { "epoch": 0.3869974646867077, "grad_norm": 1.379217006096235, "learning_rate": 1.402842875261679e-05, "loss": 0.7336, "step": 10685 }, { "epoch": 0.3870336834480261, "grad_norm": 1.4678977093908094, "learning_rate": 1.4027355052208842e-05, "loss": 0.8181, "step": 10686 }, { "epoch": 0.38706990220934445, "grad_norm": 1.4962168909296087, "learning_rate": 1.402628129638095e-05, "loss": 0.7267, "step": 10687 }, { "epoch": 0.3871061209706628, "grad_norm": 1.1664244057648552, "learning_rate": 1.4025207485147892e-05, "loss": 0.8353, "step": 10688 }, { "epoch": 0.38714233973198114, "grad_norm": 1.7868173270064964, "learning_rate": 1.4024133618524438e-05, "loss": 0.7469, "step": 10689 }, { "epoch": 0.3871785584932995, "grad_norm": 1.5154363991694075, "learning_rate": 1.4023059696525375e-05, "loss": 0.8384, "step": 10690 }, { "epoch": 0.3872147772546179, "grad_norm": 1.3547289901922395, "learning_rate": 1.402198571916547e-05, "loss": 0.8063, "step": 10691 }, { "epoch": 0.38725099601593627, "grad_norm": 1.4368816476297273, "learning_rate": 1.402091168645951e-05, "loss": 0.8409, "step": 10692 }, { "epoch": 0.38728721477725464, "grad_norm": 1.5098541845722802, "learning_rate": 1.401983759842227e-05, "loss": 0.7294, "step": 10693 }, { "epoch": 0.38732343353857296, "grad_norm": 1.3128003233794612, "learning_rate": 1.4018763455068536e-05, "loss": 0.7827, "step": 10694 }, { "epoch": 0.38735965229989133, "grad_norm": 1.3491209011512644, "learning_rate": 1.4017689256413083e-05, "loss": 0.805, "step": 10695 }, { "epoch": 0.3873958710612097, "grad_norm": 1.0852253117085735, "learning_rate": 1.4016615002470697e-05, "loss": 0.7548, "step": 10696 }, { "epoch": 0.3874320898225281, "grad_norm": 1.3150105336033902, "learning_rate": 1.401554069325616e-05, "loss": 0.8097, "step": 10697 }, { "epoch": 0.38746830858384645, "grad_norm": 1.540451796382572, "learning_rate": 1.4014466328784251e-05, "loss": 0.8182, "step": 10698 }, { "epoch": 0.3875045273451648, "grad_norm": 1.318619309690798, "learning_rate": 1.4013391909069763e-05, "loss": 0.7323, "step": 10699 }, { "epoch": 0.38754074610648315, "grad_norm": 1.3428386214647208, "learning_rate": 1.4012317434127472e-05, "loss": 0.8039, "step": 10700 }, { "epoch": 0.3875769648678015, "grad_norm": 1.5941823905740011, "learning_rate": 1.401124290397217e-05, "loss": 0.8182, "step": 10701 }, { "epoch": 0.3876131836291199, "grad_norm": 1.4561601008347462, "learning_rate": 1.4010168318618637e-05, "loss": 0.744, "step": 10702 }, { "epoch": 0.38764940239043827, "grad_norm": 1.3795366291684852, "learning_rate": 1.4009093678081669e-05, "loss": 0.766, "step": 10703 }, { "epoch": 0.3876856211517566, "grad_norm": 1.1585517222197301, "learning_rate": 1.4008018982376045e-05, "loss": 0.6441, "step": 10704 }, { "epoch": 0.38772183991307496, "grad_norm": 1.3844573786892826, "learning_rate": 1.4006944231516564e-05, "loss": 0.8207, "step": 10705 }, { "epoch": 0.38775805867439334, "grad_norm": 1.3518977700303, "learning_rate": 1.4005869425518003e-05, "loss": 0.7441, "step": 10706 }, { "epoch": 0.3877942774357117, "grad_norm": 1.3790346793647505, "learning_rate": 1.4004794564395163e-05, "loss": 0.8687, "step": 10707 }, { "epoch": 0.3878304961970301, "grad_norm": 1.4091700595201644, "learning_rate": 1.400371964816283e-05, "loss": 0.8392, "step": 10708 }, { "epoch": 0.3878667149583484, "grad_norm": 1.3161223576620453, "learning_rate": 1.4002644676835796e-05, "loss": 0.8898, "step": 10709 }, { "epoch": 0.3879029337196668, "grad_norm": 1.2257596831250943, "learning_rate": 1.4001569650428855e-05, "loss": 0.7044, "step": 10710 }, { "epoch": 0.38793915248098515, "grad_norm": 1.0747892026844656, "learning_rate": 1.4000494568956797e-05, "loss": 0.697, "step": 10711 }, { "epoch": 0.3879753712423035, "grad_norm": 1.392507097859245, "learning_rate": 1.3999419432434422e-05, "loss": 0.7941, "step": 10712 }, { "epoch": 0.3880115900036219, "grad_norm": 1.3211821035609852, "learning_rate": 1.399834424087652e-05, "loss": 0.8183, "step": 10713 }, { "epoch": 0.3880478087649402, "grad_norm": 1.3000163759291365, "learning_rate": 1.399726899429789e-05, "loss": 0.7965, "step": 10714 }, { "epoch": 0.3880840275262586, "grad_norm": 1.3523781842564395, "learning_rate": 1.3996193692713323e-05, "loss": 0.7528, "step": 10715 }, { "epoch": 0.38812024628757696, "grad_norm": 1.295135884559182, "learning_rate": 1.3995118336137624e-05, "loss": 0.8106, "step": 10716 }, { "epoch": 0.38815646504889534, "grad_norm": 1.3644771577087533, "learning_rate": 1.399404292458558e-05, "loss": 0.8115, "step": 10717 }, { "epoch": 0.3881926838102137, "grad_norm": 1.2569741948872872, "learning_rate": 1.3992967458071997e-05, "loss": 0.6644, "step": 10718 }, { "epoch": 0.38822890257153203, "grad_norm": 1.351308602054039, "learning_rate": 1.3991891936611678e-05, "loss": 0.7748, "step": 10719 }, { "epoch": 0.3882651213328504, "grad_norm": 1.427113166958996, "learning_rate": 1.3990816360219415e-05, "loss": 0.7681, "step": 10720 }, { "epoch": 0.3883013400941688, "grad_norm": 1.3000239424495197, "learning_rate": 1.3989740728910014e-05, "loss": 0.7732, "step": 10721 }, { "epoch": 0.38833755885548715, "grad_norm": 1.2813337606870494, "learning_rate": 1.3988665042698273e-05, "loss": 0.7346, "step": 10722 }, { "epoch": 0.3883737776168055, "grad_norm": 1.2541964631828935, "learning_rate": 1.3987589301598998e-05, "loss": 0.7607, "step": 10723 }, { "epoch": 0.38840999637812385, "grad_norm": 1.40106282800224, "learning_rate": 1.3986513505626988e-05, "loss": 0.8287, "step": 10724 }, { "epoch": 0.3884462151394422, "grad_norm": 1.3918538089148813, "learning_rate": 1.3985437654797052e-05, "loss": 0.7168, "step": 10725 }, { "epoch": 0.3884824339007606, "grad_norm": 1.178186469129019, "learning_rate": 1.398436174912399e-05, "loss": 0.7678, "step": 10726 }, { "epoch": 0.38851865266207897, "grad_norm": 1.3676110190679844, "learning_rate": 1.398328578862261e-05, "loss": 0.7573, "step": 10727 }, { "epoch": 0.38855487142339734, "grad_norm": 1.0297299072537642, "learning_rate": 1.3982209773307717e-05, "loss": 0.7874, "step": 10728 }, { "epoch": 0.38859109018471566, "grad_norm": 1.537466377330672, "learning_rate": 1.3981133703194118e-05, "loss": 0.8303, "step": 10729 }, { "epoch": 0.38862730894603403, "grad_norm": 1.42520109287599, "learning_rate": 1.3980057578296624e-05, "loss": 0.7378, "step": 10730 }, { "epoch": 0.3886635277073524, "grad_norm": 1.3704773462830753, "learning_rate": 1.3978981398630038e-05, "loss": 0.7611, "step": 10731 }, { "epoch": 0.3886997464686708, "grad_norm": 1.0430225867291396, "learning_rate": 1.3977905164209174e-05, "loss": 0.7874, "step": 10732 }, { "epoch": 0.38873596522998916, "grad_norm": 1.3560462925027628, "learning_rate": 1.3976828875048836e-05, "loss": 0.806, "step": 10733 }, { "epoch": 0.3887721839913075, "grad_norm": 1.1190096910574445, "learning_rate": 1.3975752531163842e-05, "loss": 0.8444, "step": 10734 }, { "epoch": 0.38880840275262585, "grad_norm": 1.1074398905938705, "learning_rate": 1.3974676132568999e-05, "loss": 0.7418, "step": 10735 }, { "epoch": 0.3888446215139442, "grad_norm": 1.3756350833166833, "learning_rate": 1.3973599679279119e-05, "loss": 0.7429, "step": 10736 }, { "epoch": 0.3888808402752626, "grad_norm": 1.4101190403601942, "learning_rate": 1.3972523171309015e-05, "loss": 0.7602, "step": 10737 }, { "epoch": 0.38891705903658097, "grad_norm": 1.476214193079238, "learning_rate": 1.3971446608673502e-05, "loss": 0.8446, "step": 10738 }, { "epoch": 0.3889532777978993, "grad_norm": 1.7217664474982874, "learning_rate": 1.3970369991387397e-05, "loss": 0.8011, "step": 10739 }, { "epoch": 0.38898949655921766, "grad_norm": 1.231574403816724, "learning_rate": 1.3969293319465508e-05, "loss": 0.6867, "step": 10740 }, { "epoch": 0.38902571532053604, "grad_norm": 1.2331188238717803, "learning_rate": 1.3968216592922656e-05, "loss": 0.7967, "step": 10741 }, { "epoch": 0.3890619340818544, "grad_norm": 1.212153745529722, "learning_rate": 1.3967139811773658e-05, "loss": 0.7778, "step": 10742 }, { "epoch": 0.3890981528431728, "grad_norm": 1.514600359115506, "learning_rate": 1.3966062976033332e-05, "loss": 0.8106, "step": 10743 }, { "epoch": 0.3891343716044911, "grad_norm": 1.4322602985299475, "learning_rate": 1.3964986085716493e-05, "loss": 0.8512, "step": 10744 }, { "epoch": 0.3891705903658095, "grad_norm": 1.043080964329262, "learning_rate": 1.3963909140837962e-05, "loss": 0.7474, "step": 10745 }, { "epoch": 0.38920680912712785, "grad_norm": 1.5004532273353712, "learning_rate": 1.3962832141412558e-05, "loss": 0.7583, "step": 10746 }, { "epoch": 0.3892430278884462, "grad_norm": 1.3619760164682733, "learning_rate": 1.39617550874551e-05, "loss": 0.6715, "step": 10747 }, { "epoch": 0.3892792466497646, "grad_norm": 1.48387575914328, "learning_rate": 1.3960677978980413e-05, "loss": 0.7076, "step": 10748 }, { "epoch": 0.3893154654110829, "grad_norm": 1.3428936760127648, "learning_rate": 1.3959600816003314e-05, "loss": 0.782, "step": 10749 }, { "epoch": 0.3893516841724013, "grad_norm": 1.3436798545914972, "learning_rate": 1.3958523598538632e-05, "loss": 0.7337, "step": 10750 }, { "epoch": 0.38938790293371967, "grad_norm": 1.2855217263246177, "learning_rate": 1.3957446326601185e-05, "loss": 0.7332, "step": 10751 }, { "epoch": 0.38942412169503804, "grad_norm": 1.2851494525161509, "learning_rate": 1.3956369000205801e-05, "loss": 0.8027, "step": 10752 }, { "epoch": 0.3894603404563564, "grad_norm": 1.332030062451211, "learning_rate": 1.3955291619367302e-05, "loss": 0.6908, "step": 10753 }, { "epoch": 0.38949655921767473, "grad_norm": 1.107327858742219, "learning_rate": 1.3954214184100515e-05, "loss": 0.7898, "step": 10754 }, { "epoch": 0.3895327779789931, "grad_norm": 1.3886466936116608, "learning_rate": 1.3953136694420267e-05, "loss": 0.8079, "step": 10755 }, { "epoch": 0.3895689967403115, "grad_norm": 1.4300152866673526, "learning_rate": 1.3952059150341383e-05, "loss": 0.7571, "step": 10756 }, { "epoch": 0.38960521550162985, "grad_norm": 1.3621968633441943, "learning_rate": 1.3950981551878693e-05, "loss": 0.7438, "step": 10757 }, { "epoch": 0.3896414342629482, "grad_norm": 1.415950631590337, "learning_rate": 1.3949903899047023e-05, "loss": 0.7156, "step": 10758 }, { "epoch": 0.38967765302426655, "grad_norm": 1.2639992134180804, "learning_rate": 1.3948826191861211e-05, "loss": 0.7192, "step": 10759 }, { "epoch": 0.3897138717855849, "grad_norm": 1.2520457209714015, "learning_rate": 1.3947748430336075e-05, "loss": 0.7913, "step": 10760 }, { "epoch": 0.3897500905469033, "grad_norm": 1.3889026235523874, "learning_rate": 1.3946670614486455e-05, "loss": 0.7771, "step": 10761 }, { "epoch": 0.38978630930822167, "grad_norm": 1.225588340728425, "learning_rate": 1.3945592744327177e-05, "loss": 0.7414, "step": 10762 }, { "epoch": 0.38982252806954004, "grad_norm": 1.4168764303051469, "learning_rate": 1.3944514819873078e-05, "loss": 0.7869, "step": 10763 }, { "epoch": 0.38985874683085836, "grad_norm": 1.2339745340471118, "learning_rate": 1.3943436841138988e-05, "loss": 0.712, "step": 10764 }, { "epoch": 0.38989496559217673, "grad_norm": 1.3058123636299062, "learning_rate": 1.3942358808139745e-05, "loss": 0.8336, "step": 10765 }, { "epoch": 0.3899311843534951, "grad_norm": 1.3658897787272353, "learning_rate": 1.3941280720890178e-05, "loss": 0.7874, "step": 10766 }, { "epoch": 0.3899674031148135, "grad_norm": 1.4050362183295153, "learning_rate": 1.3940202579405125e-05, "loss": 0.796, "step": 10767 }, { "epoch": 0.39000362187613186, "grad_norm": 1.3971186524588535, "learning_rate": 1.3939124383699422e-05, "loss": 0.8342, "step": 10768 }, { "epoch": 0.3900398406374502, "grad_norm": 1.4219733547420919, "learning_rate": 1.3938046133787906e-05, "loss": 0.8247, "step": 10769 }, { "epoch": 0.39007605939876855, "grad_norm": 1.3135508058971443, "learning_rate": 1.3936967829685418e-05, "loss": 0.8409, "step": 10770 }, { "epoch": 0.3901122781600869, "grad_norm": 1.432617829783581, "learning_rate": 1.393588947140679e-05, "loss": 0.7679, "step": 10771 }, { "epoch": 0.3901484969214053, "grad_norm": 1.368519248846541, "learning_rate": 1.3934811058966865e-05, "loss": 0.7408, "step": 10772 }, { "epoch": 0.39018471568272367, "grad_norm": 1.3707494554764252, "learning_rate": 1.3933732592380485e-05, "loss": 0.7558, "step": 10773 }, { "epoch": 0.390220934444042, "grad_norm": 1.5288426210557444, "learning_rate": 1.3932654071662486e-05, "loss": 0.7871, "step": 10774 }, { "epoch": 0.39025715320536036, "grad_norm": 1.6569482662962993, "learning_rate": 1.3931575496827711e-05, "loss": 0.7322, "step": 10775 }, { "epoch": 0.39029337196667874, "grad_norm": 1.1044734897177635, "learning_rate": 1.3930496867891004e-05, "loss": 0.7767, "step": 10776 }, { "epoch": 0.3903295907279971, "grad_norm": 1.0190282212775001, "learning_rate": 1.3929418184867206e-05, "loss": 0.6711, "step": 10777 }, { "epoch": 0.3903658094893155, "grad_norm": 1.3975286407384928, "learning_rate": 1.3928339447771162e-05, "loss": 0.7045, "step": 10778 }, { "epoch": 0.3904020282506338, "grad_norm": 1.372208552336955, "learning_rate": 1.3927260656617713e-05, "loss": 0.8485, "step": 10779 }, { "epoch": 0.3904382470119522, "grad_norm": 1.4067310358223035, "learning_rate": 1.3926181811421705e-05, "loss": 0.805, "step": 10780 }, { "epoch": 0.39047446577327055, "grad_norm": 1.364463580375691, "learning_rate": 1.3925102912197993e-05, "loss": 0.7616, "step": 10781 }, { "epoch": 0.3905106845345889, "grad_norm": 1.441965987556382, "learning_rate": 1.3924023958961408e-05, "loss": 0.8613, "step": 10782 }, { "epoch": 0.3905469032959073, "grad_norm": 1.3528175005431533, "learning_rate": 1.3922944951726811e-05, "loss": 0.7315, "step": 10783 }, { "epoch": 0.3905831220572256, "grad_norm": 1.3889491559569345, "learning_rate": 1.3921865890509042e-05, "loss": 0.7917, "step": 10784 }, { "epoch": 0.390619340818544, "grad_norm": 1.573514656658281, "learning_rate": 1.3920786775322952e-05, "loss": 0.7282, "step": 10785 }, { "epoch": 0.39065555957986237, "grad_norm": 1.4527893736312107, "learning_rate": 1.391970760618339e-05, "loss": 0.7857, "step": 10786 }, { "epoch": 0.39069177834118074, "grad_norm": 1.2713872891524505, "learning_rate": 1.391862838310521e-05, "loss": 0.7909, "step": 10787 }, { "epoch": 0.3907279971024991, "grad_norm": 1.2876690514581661, "learning_rate": 1.391754910610326e-05, "loss": 0.8121, "step": 10788 }, { "epoch": 0.39076421586381743, "grad_norm": 1.3257775724761258, "learning_rate": 1.3916469775192386e-05, "loss": 0.7443, "step": 10789 }, { "epoch": 0.3908004346251358, "grad_norm": 1.3463024734212077, "learning_rate": 1.3915390390387455e-05, "loss": 0.7778, "step": 10790 }, { "epoch": 0.3908366533864542, "grad_norm": 1.5414860615318196, "learning_rate": 1.3914310951703306e-05, "loss": 0.7305, "step": 10791 }, { "epoch": 0.39087287214777255, "grad_norm": 1.2043073247102924, "learning_rate": 1.3913231459154803e-05, "loss": 0.7451, "step": 10792 }, { "epoch": 0.39090909090909093, "grad_norm": 1.2582773347801044, "learning_rate": 1.3912151912756796e-05, "loss": 0.7067, "step": 10793 }, { "epoch": 0.39094530967040925, "grad_norm": 1.4862804499184656, "learning_rate": 1.391107231252414e-05, "loss": 0.8033, "step": 10794 }, { "epoch": 0.3909815284317276, "grad_norm": 1.2628910492622134, "learning_rate": 1.3909992658471692e-05, "loss": 0.7743, "step": 10795 }, { "epoch": 0.391017747193046, "grad_norm": 1.4733530607557992, "learning_rate": 1.3908912950614309e-05, "loss": 0.8527, "step": 10796 }, { "epoch": 0.39105396595436437, "grad_norm": 1.1571637569659219, "learning_rate": 1.390783318896685e-05, "loss": 0.7625, "step": 10797 }, { "epoch": 0.39109018471568274, "grad_norm": 1.3415632349910185, "learning_rate": 1.3906753373544173e-05, "loss": 0.7038, "step": 10798 }, { "epoch": 0.39112640347700106, "grad_norm": 1.0982106005739245, "learning_rate": 1.3905673504361136e-05, "loss": 0.7737, "step": 10799 }, { "epoch": 0.39116262223831944, "grad_norm": 1.4380481649212615, "learning_rate": 1.39045935814326e-05, "loss": 0.753, "step": 10800 }, { "epoch": 0.3911988409996378, "grad_norm": 1.2655055164815938, "learning_rate": 1.3903513604773427e-05, "loss": 0.7514, "step": 10801 }, { "epoch": 0.3912350597609562, "grad_norm": 1.41149310063168, "learning_rate": 1.3902433574398476e-05, "loss": 0.8183, "step": 10802 }, { "epoch": 0.39127127852227456, "grad_norm": 1.3172259523677592, "learning_rate": 1.390135349032261e-05, "loss": 0.752, "step": 10803 }, { "epoch": 0.3913074972835929, "grad_norm": 1.4592798679373118, "learning_rate": 1.3900273352560689e-05, "loss": 0.7545, "step": 10804 }, { "epoch": 0.39134371604491125, "grad_norm": 1.4415335598990187, "learning_rate": 1.3899193161127583e-05, "loss": 0.8117, "step": 10805 }, { "epoch": 0.3913799348062296, "grad_norm": 1.115594889599864, "learning_rate": 1.3898112916038151e-05, "loss": 0.7344, "step": 10806 }, { "epoch": 0.391416153567548, "grad_norm": 1.4450986575111437, "learning_rate": 1.389703261730726e-05, "loss": 0.7705, "step": 10807 }, { "epoch": 0.39145237232886637, "grad_norm": 1.4165624937126502, "learning_rate": 1.3895952264949779e-05, "loss": 0.7645, "step": 10808 }, { "epoch": 0.3914885910901847, "grad_norm": 1.4054914190108612, "learning_rate": 1.3894871858980568e-05, "loss": 0.7001, "step": 10809 }, { "epoch": 0.39152480985150306, "grad_norm": 1.2655566625336128, "learning_rate": 1.38937913994145e-05, "loss": 0.6934, "step": 10810 }, { "epoch": 0.39156102861282144, "grad_norm": 1.300831541028576, "learning_rate": 1.389271088626644e-05, "loss": 0.8335, "step": 10811 }, { "epoch": 0.3915972473741398, "grad_norm": 1.3576571541935853, "learning_rate": 1.389163031955126e-05, "loss": 0.771, "step": 10812 }, { "epoch": 0.3916334661354582, "grad_norm": 1.3503721190155034, "learning_rate": 1.3890549699283824e-05, "loss": 0.7125, "step": 10813 }, { "epoch": 0.3916696848967765, "grad_norm": 1.3604710533969913, "learning_rate": 1.3889469025479009e-05, "loss": 0.8143, "step": 10814 }, { "epoch": 0.3917059036580949, "grad_norm": 1.5489700872872751, "learning_rate": 1.3888388298151679e-05, "loss": 0.8428, "step": 10815 }, { "epoch": 0.39174212241941325, "grad_norm": 1.1798571778789844, "learning_rate": 1.388730751731671e-05, "loss": 0.6954, "step": 10816 }, { "epoch": 0.3917783411807316, "grad_norm": 1.3307657346684465, "learning_rate": 1.3886226682988975e-05, "loss": 0.7795, "step": 10817 }, { "epoch": 0.39181455994205, "grad_norm": 1.4410788091786495, "learning_rate": 1.3885145795183345e-05, "loss": 0.8114, "step": 10818 }, { "epoch": 0.3918507787033683, "grad_norm": 1.3727246098927917, "learning_rate": 1.3884064853914698e-05, "loss": 0.7912, "step": 10819 }, { "epoch": 0.3918869974646867, "grad_norm": 1.4237450351724812, "learning_rate": 1.3882983859197903e-05, "loss": 0.7253, "step": 10820 }, { "epoch": 0.39192321622600507, "grad_norm": 1.32729342055456, "learning_rate": 1.388190281104784e-05, "loss": 0.7646, "step": 10821 }, { "epoch": 0.39195943498732344, "grad_norm": 1.3863463304667314, "learning_rate": 1.3880821709479383e-05, "loss": 0.7596, "step": 10822 }, { "epoch": 0.3919956537486418, "grad_norm": 1.3895104427542782, "learning_rate": 1.3879740554507408e-05, "loss": 0.7709, "step": 10823 }, { "epoch": 0.39203187250996013, "grad_norm": 1.3665720194575095, "learning_rate": 1.3878659346146795e-05, "loss": 0.7381, "step": 10824 }, { "epoch": 0.3920680912712785, "grad_norm": 1.3723957938231066, "learning_rate": 1.3877578084412422e-05, "loss": 0.7013, "step": 10825 }, { "epoch": 0.3921043100325969, "grad_norm": 1.3686640803922276, "learning_rate": 1.3876496769319166e-05, "loss": 0.6758, "step": 10826 }, { "epoch": 0.39214052879391526, "grad_norm": 1.2935683223069705, "learning_rate": 1.3875415400881907e-05, "loss": 0.7486, "step": 10827 }, { "epoch": 0.39217674755523363, "grad_norm": 1.215373971122712, "learning_rate": 1.3874333979115529e-05, "loss": 0.7661, "step": 10828 }, { "epoch": 0.39221296631655195, "grad_norm": 1.368240220906541, "learning_rate": 1.387325250403491e-05, "loss": 0.7941, "step": 10829 }, { "epoch": 0.3922491850778703, "grad_norm": 1.366063069759658, "learning_rate": 1.3872170975654935e-05, "loss": 0.8062, "step": 10830 }, { "epoch": 0.3922854038391887, "grad_norm": 1.1569198364875326, "learning_rate": 1.3871089393990482e-05, "loss": 0.8051, "step": 10831 }, { "epoch": 0.39232162260050707, "grad_norm": 1.5108184166424994, "learning_rate": 1.3870007759056441e-05, "loss": 0.7721, "step": 10832 }, { "epoch": 0.39235784136182544, "grad_norm": 1.4154662349020413, "learning_rate": 1.386892607086769e-05, "loss": 0.8527, "step": 10833 }, { "epoch": 0.39239406012314376, "grad_norm": 1.3651173067994749, "learning_rate": 1.386784432943912e-05, "loss": 0.7924, "step": 10834 }, { "epoch": 0.39243027888446214, "grad_norm": 1.3365350204048365, "learning_rate": 1.386676253478561e-05, "loss": 0.7964, "step": 10835 }, { "epoch": 0.3924664976457805, "grad_norm": 1.2242288812109787, "learning_rate": 1.386568068692205e-05, "loss": 0.7183, "step": 10836 }, { "epoch": 0.3925027164070989, "grad_norm": 1.3257724767292374, "learning_rate": 1.386459878586333e-05, "loss": 0.7432, "step": 10837 }, { "epoch": 0.39253893516841726, "grad_norm": 1.137905047284693, "learning_rate": 1.3863516831624333e-05, "loss": 0.6817, "step": 10838 }, { "epoch": 0.3925751539297356, "grad_norm": 1.0751078840357084, "learning_rate": 1.386243482421995e-05, "loss": 0.7249, "step": 10839 }, { "epoch": 0.39261137269105395, "grad_norm": 1.713918702527806, "learning_rate": 1.3861352763665067e-05, "loss": 0.9099, "step": 10840 }, { "epoch": 0.3926475914523723, "grad_norm": 1.4075451578925031, "learning_rate": 1.3860270649974582e-05, "loss": 0.7647, "step": 10841 }, { "epoch": 0.3926838102136907, "grad_norm": 1.487925702275192, "learning_rate": 1.385918848316338e-05, "loss": 0.7948, "step": 10842 }, { "epoch": 0.3927200289750091, "grad_norm": 1.4293777607135107, "learning_rate": 1.3858106263246351e-05, "loss": 0.8561, "step": 10843 }, { "epoch": 0.3927562477363274, "grad_norm": 1.4301929290514488, "learning_rate": 1.3857023990238391e-05, "loss": 0.6909, "step": 10844 }, { "epoch": 0.39279246649764576, "grad_norm": 1.4328558189349874, "learning_rate": 1.3855941664154393e-05, "loss": 0.7857, "step": 10845 }, { "epoch": 0.39282868525896414, "grad_norm": 1.4736442413796955, "learning_rate": 1.3854859285009248e-05, "loss": 0.8135, "step": 10846 }, { "epoch": 0.3928649040202825, "grad_norm": 1.1447602844787328, "learning_rate": 1.385377685281785e-05, "loss": 0.7889, "step": 10847 }, { "epoch": 0.3929011227816009, "grad_norm": 1.2684210632755444, "learning_rate": 1.3852694367595102e-05, "loss": 0.7281, "step": 10848 }, { "epoch": 0.3929373415429192, "grad_norm": 1.254194689193515, "learning_rate": 1.385161182935589e-05, "loss": 0.7221, "step": 10849 }, { "epoch": 0.3929735603042376, "grad_norm": 1.3495327545476514, "learning_rate": 1.3850529238115115e-05, "loss": 0.8079, "step": 10850 }, { "epoch": 0.39300977906555595, "grad_norm": 1.2988655061648682, "learning_rate": 1.3849446593887677e-05, "loss": 0.7287, "step": 10851 }, { "epoch": 0.3930459978268743, "grad_norm": 1.4788893494393507, "learning_rate": 1.384836389668847e-05, "loss": 0.7521, "step": 10852 }, { "epoch": 0.3930822165881927, "grad_norm": 1.3967105282213885, "learning_rate": 1.3847281146532392e-05, "loss": 0.7703, "step": 10853 }, { "epoch": 0.393118435349511, "grad_norm": 1.360097525606322, "learning_rate": 1.384619834343435e-05, "loss": 0.7716, "step": 10854 }, { "epoch": 0.3931546541108294, "grad_norm": 1.1039814482778945, "learning_rate": 1.3845115487409233e-05, "loss": 0.8223, "step": 10855 }, { "epoch": 0.39319087287214777, "grad_norm": 1.367503724993195, "learning_rate": 1.3844032578471953e-05, "loss": 0.7753, "step": 10856 }, { "epoch": 0.39322709163346614, "grad_norm": 1.4526992668982857, "learning_rate": 1.3842949616637404e-05, "loss": 0.7598, "step": 10857 }, { "epoch": 0.3932633103947845, "grad_norm": 1.572359885772475, "learning_rate": 1.3841866601920493e-05, "loss": 0.7486, "step": 10858 }, { "epoch": 0.39329952915610283, "grad_norm": 1.4878844710964132, "learning_rate": 1.3840783534336123e-05, "loss": 0.8144, "step": 10859 }, { "epoch": 0.3933357479174212, "grad_norm": 1.4062247883338066, "learning_rate": 1.3839700413899194e-05, "loss": 0.7524, "step": 10860 }, { "epoch": 0.3933719666787396, "grad_norm": 1.2984587667630945, "learning_rate": 1.3838617240624616e-05, "loss": 0.7786, "step": 10861 }, { "epoch": 0.39340818544005796, "grad_norm": 1.246173397944789, "learning_rate": 1.3837534014527292e-05, "loss": 0.7537, "step": 10862 }, { "epoch": 0.39344440420137633, "grad_norm": 1.714602022958279, "learning_rate": 1.3836450735622126e-05, "loss": 0.7928, "step": 10863 }, { "epoch": 0.39348062296269465, "grad_norm": 1.1994708760921828, "learning_rate": 1.383536740392403e-05, "loss": 0.7663, "step": 10864 }, { "epoch": 0.393516841724013, "grad_norm": 1.1548428867821634, "learning_rate": 1.3834284019447906e-05, "loss": 0.8001, "step": 10865 }, { "epoch": 0.3935530604853314, "grad_norm": 1.3620453755227293, "learning_rate": 1.3833200582208664e-05, "loss": 0.8029, "step": 10866 }, { "epoch": 0.39358927924664977, "grad_norm": 1.2856524043814628, "learning_rate": 1.3832117092221213e-05, "loss": 0.7613, "step": 10867 }, { "epoch": 0.39362549800796814, "grad_norm": 1.30424158141054, "learning_rate": 1.3831033549500468e-05, "loss": 0.7702, "step": 10868 }, { "epoch": 0.39366171676928646, "grad_norm": 1.2018298324215153, "learning_rate": 1.3829949954061332e-05, "loss": 0.7995, "step": 10869 }, { "epoch": 0.39369793553060484, "grad_norm": 1.418797315268585, "learning_rate": 1.382886630591872e-05, "loss": 0.75, "step": 10870 }, { "epoch": 0.3937341542919232, "grad_norm": 1.3166422193808769, "learning_rate": 1.3827782605087543e-05, "loss": 0.7683, "step": 10871 }, { "epoch": 0.3937703730532416, "grad_norm": 1.0619110099133988, "learning_rate": 1.3826698851582716e-05, "loss": 0.7537, "step": 10872 }, { "epoch": 0.39380659181455996, "grad_norm": 1.033273632466502, "learning_rate": 1.3825615045419148e-05, "loss": 0.7582, "step": 10873 }, { "epoch": 0.39384281057587833, "grad_norm": 1.2187525232606835, "learning_rate": 1.3824531186611756e-05, "loss": 0.6961, "step": 10874 }, { "epoch": 0.39387902933719665, "grad_norm": 2.133542746396434, "learning_rate": 1.3823447275175454e-05, "loss": 0.8083, "step": 10875 }, { "epoch": 0.393915248098515, "grad_norm": 1.4015037217567958, "learning_rate": 1.382236331112516e-05, "loss": 0.7883, "step": 10876 }, { "epoch": 0.3939514668598334, "grad_norm": 1.320800832371173, "learning_rate": 1.3821279294475785e-05, "loss": 0.8174, "step": 10877 }, { "epoch": 0.3939876856211518, "grad_norm": 1.3163257391728749, "learning_rate": 1.382019522524225e-05, "loss": 0.773, "step": 10878 }, { "epoch": 0.39402390438247015, "grad_norm": 1.4544167007399, "learning_rate": 1.3819111103439473e-05, "loss": 0.8259, "step": 10879 }, { "epoch": 0.39406012314378847, "grad_norm": 1.5366350687880637, "learning_rate": 1.381802692908237e-05, "loss": 0.8696, "step": 10880 }, { "epoch": 0.39409634190510684, "grad_norm": 1.5168683295816467, "learning_rate": 1.3816942702185864e-05, "loss": 0.8335, "step": 10881 }, { "epoch": 0.3941325606664252, "grad_norm": 1.3425447303896292, "learning_rate": 1.3815858422764872e-05, "loss": 0.8174, "step": 10882 }, { "epoch": 0.3941687794277436, "grad_norm": 1.3352002457478778, "learning_rate": 1.3814774090834315e-05, "loss": 0.734, "step": 10883 }, { "epoch": 0.39420499818906196, "grad_norm": 1.2190332281823952, "learning_rate": 1.3813689706409114e-05, "loss": 0.7625, "step": 10884 }, { "epoch": 0.3942412169503803, "grad_norm": 1.2577021443783059, "learning_rate": 1.3812605269504194e-05, "loss": 0.722, "step": 10885 }, { "epoch": 0.39427743571169865, "grad_norm": 1.3394653631248095, "learning_rate": 1.3811520780134471e-05, "loss": 0.8198, "step": 10886 }, { "epoch": 0.39431365447301703, "grad_norm": 1.3594350911155448, "learning_rate": 1.3810436238314877e-05, "loss": 0.8304, "step": 10887 }, { "epoch": 0.3943498732343354, "grad_norm": 2.699001695487531, "learning_rate": 1.3809351644060329e-05, "loss": 0.6996, "step": 10888 }, { "epoch": 0.3943860919956538, "grad_norm": 1.3935104689235631, "learning_rate": 1.3808266997385756e-05, "loss": 0.7529, "step": 10889 }, { "epoch": 0.3944223107569721, "grad_norm": 1.0848798308413763, "learning_rate": 1.3807182298306086e-05, "loss": 0.7318, "step": 10890 }, { "epoch": 0.39445852951829047, "grad_norm": 1.3980295888999736, "learning_rate": 1.380609754683624e-05, "loss": 0.7034, "step": 10891 }, { "epoch": 0.39449474827960884, "grad_norm": 0.9470746775756849, "learning_rate": 1.3805012742991147e-05, "loss": 0.7303, "step": 10892 }, { "epoch": 0.3945309670409272, "grad_norm": 1.3802573457748533, "learning_rate": 1.3803927886785732e-05, "loss": 0.8343, "step": 10893 }, { "epoch": 0.3945671858022456, "grad_norm": 1.3650207004411932, "learning_rate": 1.3802842978234933e-05, "loss": 0.7468, "step": 10894 }, { "epoch": 0.3946034045635639, "grad_norm": 1.58065047348405, "learning_rate": 1.380175801735367e-05, "loss": 0.8194, "step": 10895 }, { "epoch": 0.3946396233248823, "grad_norm": 1.1796501824120436, "learning_rate": 1.380067300415688e-05, "loss": 0.7539, "step": 10896 }, { "epoch": 0.39467584208620066, "grad_norm": 1.3731619439398866, "learning_rate": 1.3799587938659485e-05, "loss": 0.8402, "step": 10897 }, { "epoch": 0.39471206084751903, "grad_norm": 1.4350961885982494, "learning_rate": 1.379850282087642e-05, "loss": 0.8303, "step": 10898 }, { "epoch": 0.3947482796088374, "grad_norm": 1.4576575049393805, "learning_rate": 1.3797417650822626e-05, "loss": 0.7877, "step": 10899 }, { "epoch": 0.3947844983701557, "grad_norm": 1.9971531484720935, "learning_rate": 1.3796332428513024e-05, "loss": 0.8346, "step": 10900 }, { "epoch": 0.3948207171314741, "grad_norm": 1.4112021881262258, "learning_rate": 1.3795247153962553e-05, "loss": 0.8154, "step": 10901 }, { "epoch": 0.39485693589279247, "grad_norm": 1.5055081833721446, "learning_rate": 1.3794161827186143e-05, "loss": 0.8427, "step": 10902 }, { "epoch": 0.39489315465411085, "grad_norm": 1.5657799841069897, "learning_rate": 1.3793076448198737e-05, "loss": 0.7035, "step": 10903 }, { "epoch": 0.3949293734154292, "grad_norm": 1.3246842427021321, "learning_rate": 1.3791991017015262e-05, "loss": 0.7292, "step": 10904 }, { "epoch": 0.39496559217674754, "grad_norm": 1.6346045657432058, "learning_rate": 1.3790905533650661e-05, "loss": 0.7585, "step": 10905 }, { "epoch": 0.3950018109380659, "grad_norm": 1.413956207257182, "learning_rate": 1.3789819998119869e-05, "loss": 0.7588, "step": 10906 }, { "epoch": 0.3950380296993843, "grad_norm": 1.3173453224565481, "learning_rate": 1.3788734410437821e-05, "loss": 0.7279, "step": 10907 }, { "epoch": 0.39507424846070266, "grad_norm": 1.370045892542189, "learning_rate": 1.3787648770619462e-05, "loss": 0.7956, "step": 10908 }, { "epoch": 0.39511046722202103, "grad_norm": 1.445422069159488, "learning_rate": 1.3786563078679727e-05, "loss": 0.7684, "step": 10909 }, { "epoch": 0.39514668598333935, "grad_norm": 1.2290578587071233, "learning_rate": 1.3785477334633557e-05, "loss": 0.791, "step": 10910 }, { "epoch": 0.3951829047446577, "grad_norm": 1.4891168246211721, "learning_rate": 1.3784391538495892e-05, "loss": 0.7525, "step": 10911 }, { "epoch": 0.3952191235059761, "grad_norm": 1.4375040459835084, "learning_rate": 1.3783305690281675e-05, "loss": 0.8753, "step": 10912 }, { "epoch": 0.3952553422672945, "grad_norm": 1.3643619092139838, "learning_rate": 1.3782219790005846e-05, "loss": 0.7427, "step": 10913 }, { "epoch": 0.39529156102861285, "grad_norm": 1.6136796242373876, "learning_rate": 1.378113383768335e-05, "loss": 0.74, "step": 10914 }, { "epoch": 0.39532777978993117, "grad_norm": 1.177677054595822, "learning_rate": 1.3780047833329129e-05, "loss": 0.7485, "step": 10915 }, { "epoch": 0.39536399855124954, "grad_norm": 1.3401330616058884, "learning_rate": 1.3778961776958128e-05, "loss": 0.6608, "step": 10916 }, { "epoch": 0.3954002173125679, "grad_norm": 0.9596284600299674, "learning_rate": 1.3777875668585296e-05, "loss": 0.6712, "step": 10917 }, { "epoch": 0.3954364360738863, "grad_norm": 1.1262895142686766, "learning_rate": 1.3776789508225572e-05, "loss": 0.7505, "step": 10918 }, { "epoch": 0.39547265483520466, "grad_norm": 1.3223471905161983, "learning_rate": 1.3775703295893907e-05, "loss": 0.7211, "step": 10919 }, { "epoch": 0.395508873596523, "grad_norm": 1.459799385860687, "learning_rate": 1.3774617031605246e-05, "loss": 0.8211, "step": 10920 }, { "epoch": 0.39554509235784135, "grad_norm": 1.2636653179193147, "learning_rate": 1.3773530715374542e-05, "loss": 0.8074, "step": 10921 }, { "epoch": 0.39558131111915973, "grad_norm": 1.4574927003266271, "learning_rate": 1.3772444347216734e-05, "loss": 0.7379, "step": 10922 }, { "epoch": 0.3956175298804781, "grad_norm": 1.7850351542963419, "learning_rate": 1.377135792714678e-05, "loss": 0.7254, "step": 10923 }, { "epoch": 0.3956537486417965, "grad_norm": 1.2744393064841453, "learning_rate": 1.3770271455179627e-05, "loss": 0.7614, "step": 10924 }, { "epoch": 0.3956899674031148, "grad_norm": 1.1300784854875154, "learning_rate": 1.3769184931330223e-05, "loss": 0.7804, "step": 10925 }, { "epoch": 0.39572618616443317, "grad_norm": 0.98028027903715, "learning_rate": 1.3768098355613528e-05, "loss": 0.7445, "step": 10926 }, { "epoch": 0.39576240492575154, "grad_norm": 1.600240860759526, "learning_rate": 1.3767011728044483e-05, "loss": 0.7987, "step": 10927 }, { "epoch": 0.3957986236870699, "grad_norm": 1.737466484027866, "learning_rate": 1.3765925048638051e-05, "loss": 0.7075, "step": 10928 }, { "epoch": 0.3958348424483883, "grad_norm": 1.4456087073091601, "learning_rate": 1.376483831740918e-05, "loss": 0.7738, "step": 10929 }, { "epoch": 0.3958710612097066, "grad_norm": 1.3776919871613553, "learning_rate": 1.3763751534372825e-05, "loss": 0.7142, "step": 10930 }, { "epoch": 0.395907279971025, "grad_norm": 1.3630368244513162, "learning_rate": 1.3762664699543941e-05, "loss": 0.6604, "step": 10931 }, { "epoch": 0.39594349873234336, "grad_norm": 0.9770633412768698, "learning_rate": 1.3761577812937489e-05, "loss": 0.6835, "step": 10932 }, { "epoch": 0.39597971749366173, "grad_norm": 1.4910028608503854, "learning_rate": 1.3760490874568415e-05, "loss": 0.8518, "step": 10933 }, { "epoch": 0.3960159362549801, "grad_norm": 1.3180396391657707, "learning_rate": 1.3759403884451687e-05, "loss": 0.7616, "step": 10934 }, { "epoch": 0.3960521550162984, "grad_norm": 1.376275779294715, "learning_rate": 1.3758316842602256e-05, "loss": 0.8049, "step": 10935 }, { "epoch": 0.3960883737776168, "grad_norm": 1.5793558976331903, "learning_rate": 1.3757229749035081e-05, "loss": 0.7316, "step": 10936 }, { "epoch": 0.39612459253893517, "grad_norm": 2.14095708805825, "learning_rate": 1.3756142603765126e-05, "loss": 0.7643, "step": 10937 }, { "epoch": 0.39616081130025355, "grad_norm": 1.0676782078001437, "learning_rate": 1.3755055406807348e-05, "loss": 0.6806, "step": 10938 }, { "epoch": 0.3961970300615719, "grad_norm": 1.3640939857185315, "learning_rate": 1.375396815817671e-05, "loss": 0.7193, "step": 10939 }, { "epoch": 0.39623324882289024, "grad_norm": 1.4805373500541539, "learning_rate": 1.3752880857888166e-05, "loss": 0.7725, "step": 10940 }, { "epoch": 0.3962694675842086, "grad_norm": 1.749363340643557, "learning_rate": 1.375179350595669e-05, "loss": 0.8032, "step": 10941 }, { "epoch": 0.396305686345527, "grad_norm": 1.5114812947850786, "learning_rate": 1.3750706102397235e-05, "loss": 0.808, "step": 10942 }, { "epoch": 0.39634190510684536, "grad_norm": 1.3670493614709673, "learning_rate": 1.3749618647224772e-05, "loss": 0.7201, "step": 10943 }, { "epoch": 0.39637812386816373, "grad_norm": 1.4401902209870685, "learning_rate": 1.3748531140454258e-05, "loss": 0.7838, "step": 10944 }, { "epoch": 0.39641434262948205, "grad_norm": 1.0163723661371873, "learning_rate": 1.3747443582100661e-05, "loss": 0.719, "step": 10945 }, { "epoch": 0.3964505613908004, "grad_norm": 1.0463061182617983, "learning_rate": 1.3746355972178951e-05, "loss": 0.7843, "step": 10946 }, { "epoch": 0.3964867801521188, "grad_norm": 1.4158279287970383, "learning_rate": 1.3745268310704087e-05, "loss": 0.7743, "step": 10947 }, { "epoch": 0.3965229989134372, "grad_norm": 1.75394032069293, "learning_rate": 1.3744180597691044e-05, "loss": 0.7265, "step": 10948 }, { "epoch": 0.39655921767475555, "grad_norm": 1.4049894517805077, "learning_rate": 1.3743092833154785e-05, "loss": 0.7862, "step": 10949 }, { "epoch": 0.39659543643607387, "grad_norm": 1.3043064671746643, "learning_rate": 1.3742005017110281e-05, "loss": 0.7616, "step": 10950 }, { "epoch": 0.39663165519739224, "grad_norm": 1.3696656465277484, "learning_rate": 1.3740917149572497e-05, "loss": 0.7842, "step": 10951 }, { "epoch": 0.3966678739587106, "grad_norm": 1.5169817072866216, "learning_rate": 1.3739829230556408e-05, "loss": 0.8192, "step": 10952 }, { "epoch": 0.396704092720029, "grad_norm": 1.2731722346278735, "learning_rate": 1.373874126007698e-05, "loss": 0.7699, "step": 10953 }, { "epoch": 0.39674031148134736, "grad_norm": 1.2417281828943743, "learning_rate": 1.3737653238149192e-05, "loss": 0.7305, "step": 10954 }, { "epoch": 0.3967765302426657, "grad_norm": 1.1082032688284988, "learning_rate": 1.3736565164788008e-05, "loss": 0.7117, "step": 10955 }, { "epoch": 0.39681274900398406, "grad_norm": 1.4295692629465926, "learning_rate": 1.3735477040008403e-05, "loss": 0.7954, "step": 10956 }, { "epoch": 0.39684896776530243, "grad_norm": 1.3903137431971935, "learning_rate": 1.3734388863825352e-05, "loss": 0.7401, "step": 10957 }, { "epoch": 0.3968851865266208, "grad_norm": 1.6831606487085597, "learning_rate": 1.373330063625383e-05, "loss": 0.7746, "step": 10958 }, { "epoch": 0.3969214052879392, "grad_norm": 1.2374842400740904, "learning_rate": 1.3732212357308812e-05, "loss": 0.7563, "step": 10959 }, { "epoch": 0.3969576240492575, "grad_norm": 1.4770326685780104, "learning_rate": 1.373112402700527e-05, "loss": 0.8291, "step": 10960 }, { "epoch": 0.39699384281057587, "grad_norm": 1.147115536608876, "learning_rate": 1.3730035645358187e-05, "loss": 0.7312, "step": 10961 }, { "epoch": 0.39703006157189424, "grad_norm": 1.4103688736230764, "learning_rate": 1.372894721238253e-05, "loss": 0.6919, "step": 10962 }, { "epoch": 0.3970662803332126, "grad_norm": 1.3959701725146674, "learning_rate": 1.3727858728093288e-05, "loss": 0.7578, "step": 10963 }, { "epoch": 0.397102499094531, "grad_norm": 1.1706904480317468, "learning_rate": 1.3726770192505432e-05, "loss": 0.7999, "step": 10964 }, { "epoch": 0.3971387178558493, "grad_norm": 1.1083042042750217, "learning_rate": 1.3725681605633946e-05, "loss": 0.7158, "step": 10965 }, { "epoch": 0.3971749366171677, "grad_norm": 1.146265121145107, "learning_rate": 1.3724592967493804e-05, "loss": 0.7355, "step": 10966 }, { "epoch": 0.39721115537848606, "grad_norm": 1.130312439867448, "learning_rate": 1.3723504278099991e-05, "loss": 0.8184, "step": 10967 }, { "epoch": 0.39724737413980443, "grad_norm": 1.0364379380730846, "learning_rate": 1.3722415537467488e-05, "loss": 0.7691, "step": 10968 }, { "epoch": 0.3972835929011228, "grad_norm": 2.3781575980815806, "learning_rate": 1.3721326745611277e-05, "loss": 0.8096, "step": 10969 }, { "epoch": 0.3973198116624411, "grad_norm": 1.3898934793662012, "learning_rate": 1.3720237902546339e-05, "loss": 0.7541, "step": 10970 }, { "epoch": 0.3973560304237595, "grad_norm": 1.3747648978270706, "learning_rate": 1.3719149008287658e-05, "loss": 0.7074, "step": 10971 }, { "epoch": 0.3973922491850779, "grad_norm": 1.4468775703868735, "learning_rate": 1.371806006285022e-05, "loss": 0.8127, "step": 10972 }, { "epoch": 0.39742846794639625, "grad_norm": 1.0984096973943733, "learning_rate": 1.3716971066249008e-05, "loss": 0.7822, "step": 10973 }, { "epoch": 0.3974646867077146, "grad_norm": 1.3735940068313797, "learning_rate": 1.371588201849901e-05, "loss": 0.7399, "step": 10974 }, { "epoch": 0.39750090546903294, "grad_norm": 1.1011877728796715, "learning_rate": 1.3714792919615207e-05, "loss": 0.8193, "step": 10975 }, { "epoch": 0.3975371242303513, "grad_norm": 1.1632352372524413, "learning_rate": 1.371370376961259e-05, "loss": 0.7012, "step": 10976 }, { "epoch": 0.3975733429916697, "grad_norm": 1.390447992407156, "learning_rate": 1.3712614568506149e-05, "loss": 0.7855, "step": 10977 }, { "epoch": 0.39760956175298806, "grad_norm": 1.358541367317806, "learning_rate": 1.3711525316310867e-05, "loss": 0.7029, "step": 10978 }, { "epoch": 0.39764578051430644, "grad_norm": 1.4144382201586476, "learning_rate": 1.3710436013041738e-05, "loss": 0.8195, "step": 10979 }, { "epoch": 0.39768199927562475, "grad_norm": 1.3796872724627935, "learning_rate": 1.3709346658713745e-05, "loss": 0.8311, "step": 10980 }, { "epoch": 0.3977182180369431, "grad_norm": 1.349385621144792, "learning_rate": 1.3708257253341886e-05, "loss": 0.7258, "step": 10981 }, { "epoch": 0.3977544367982615, "grad_norm": 1.343736464132843, "learning_rate": 1.3707167796941146e-05, "loss": 0.7573, "step": 10982 }, { "epoch": 0.3977906555595799, "grad_norm": 1.2279756326575968, "learning_rate": 1.3706078289526521e-05, "loss": 0.8358, "step": 10983 }, { "epoch": 0.39782687432089825, "grad_norm": 1.4389126382761133, "learning_rate": 1.3704988731113003e-05, "loss": 0.8645, "step": 10984 }, { "epoch": 0.39786309308221657, "grad_norm": 1.1128457989189102, "learning_rate": 1.3703899121715585e-05, "loss": 0.7545, "step": 10985 }, { "epoch": 0.39789931184353494, "grad_norm": 1.4430044173916887, "learning_rate": 1.3702809461349258e-05, "loss": 0.8207, "step": 10986 }, { "epoch": 0.3979355306048533, "grad_norm": 1.4048676425131248, "learning_rate": 1.3701719750029021e-05, "loss": 0.7766, "step": 10987 }, { "epoch": 0.3979717493661717, "grad_norm": 1.1912348100654238, "learning_rate": 1.3700629987769868e-05, "loss": 0.8005, "step": 10988 }, { "epoch": 0.39800796812749006, "grad_norm": 1.8230503330725123, "learning_rate": 1.3699540174586793e-05, "loss": 0.8089, "step": 10989 }, { "epoch": 0.3980441868888084, "grad_norm": 1.4851136574307773, "learning_rate": 1.3698450310494798e-05, "loss": 0.8102, "step": 10990 }, { "epoch": 0.39808040565012676, "grad_norm": 1.3900942834474446, "learning_rate": 1.3697360395508871e-05, "loss": 0.8278, "step": 10991 }, { "epoch": 0.39811662441144513, "grad_norm": 1.2877117630092774, "learning_rate": 1.3696270429644023e-05, "loss": 0.6995, "step": 10992 }, { "epoch": 0.3981528431727635, "grad_norm": 1.2395223454707716, "learning_rate": 1.369518041291524e-05, "loss": 0.6829, "step": 10993 }, { "epoch": 0.3981890619340819, "grad_norm": 1.3260472699815091, "learning_rate": 1.3694090345337534e-05, "loss": 0.7782, "step": 10994 }, { "epoch": 0.3982252806954002, "grad_norm": 1.276911437811939, "learning_rate": 1.3693000226925894e-05, "loss": 0.7658, "step": 10995 }, { "epoch": 0.39826149945671857, "grad_norm": 1.0784405494561056, "learning_rate": 1.3691910057695328e-05, "loss": 0.7317, "step": 10996 }, { "epoch": 0.39829771821803694, "grad_norm": 1.3527443981267484, "learning_rate": 1.3690819837660834e-05, "loss": 0.7735, "step": 10997 }, { "epoch": 0.3983339369793553, "grad_norm": 1.3146733787442197, "learning_rate": 1.3689729566837416e-05, "loss": 0.7447, "step": 10998 }, { "epoch": 0.3983701557406737, "grad_norm": 1.3485280237224035, "learning_rate": 1.3688639245240078e-05, "loss": 0.8213, "step": 10999 }, { "epoch": 0.398406374501992, "grad_norm": 1.3938532003654462, "learning_rate": 1.3687548872883824e-05, "loss": 0.7604, "step": 11000 }, { "epoch": 0.3984425932633104, "grad_norm": 1.26476863820078, "learning_rate": 1.3686458449783657e-05, "loss": 0.7199, "step": 11001 }, { "epoch": 0.39847881202462876, "grad_norm": 1.3842341986497178, "learning_rate": 1.368536797595458e-05, "loss": 0.6832, "step": 11002 }, { "epoch": 0.39851503078594713, "grad_norm": 1.3023018938105866, "learning_rate": 1.3684277451411606e-05, "loss": 0.7897, "step": 11003 }, { "epoch": 0.3985512495472655, "grad_norm": 1.0212133396869747, "learning_rate": 1.3683186876169734e-05, "loss": 0.7762, "step": 11004 }, { "epoch": 0.3985874683085838, "grad_norm": 1.3865987101164798, "learning_rate": 1.3682096250243975e-05, "loss": 0.7384, "step": 11005 }, { "epoch": 0.3986236870699022, "grad_norm": 1.3135738152445282, "learning_rate": 1.3681005573649335e-05, "loss": 0.7526, "step": 11006 }, { "epoch": 0.3986599058312206, "grad_norm": 1.1732356235129384, "learning_rate": 1.3679914846400824e-05, "loss": 0.6443, "step": 11007 }, { "epoch": 0.39869612459253895, "grad_norm": 1.37155994005023, "learning_rate": 1.3678824068513455e-05, "loss": 0.7384, "step": 11008 }, { "epoch": 0.3987323433538573, "grad_norm": 1.4220853953798762, "learning_rate": 1.3677733240002228e-05, "loss": 0.8083, "step": 11009 }, { "epoch": 0.39876856211517564, "grad_norm": 1.521011909659804, "learning_rate": 1.3676642360882167e-05, "loss": 0.8367, "step": 11010 }, { "epoch": 0.398804780876494, "grad_norm": 1.3766314732811487, "learning_rate": 1.367555143116827e-05, "loss": 0.7765, "step": 11011 }, { "epoch": 0.3988409996378124, "grad_norm": 1.3282304721913913, "learning_rate": 1.367446045087556e-05, "loss": 0.7296, "step": 11012 }, { "epoch": 0.39887721839913076, "grad_norm": 1.4870184499581112, "learning_rate": 1.3673369420019043e-05, "loss": 0.7204, "step": 11013 }, { "epoch": 0.39891343716044914, "grad_norm": 1.8244614674508899, "learning_rate": 1.3672278338613736e-05, "loss": 0.834, "step": 11014 }, { "epoch": 0.39894965592176745, "grad_norm": 1.1950238202497279, "learning_rate": 1.3671187206674652e-05, "loss": 0.7384, "step": 11015 }, { "epoch": 0.39898587468308583, "grad_norm": 1.2793818756795445, "learning_rate": 1.3670096024216806e-05, "loss": 0.7986, "step": 11016 }, { "epoch": 0.3990220934444042, "grad_norm": 1.1162059618829816, "learning_rate": 1.3669004791255216e-05, "loss": 0.7376, "step": 11017 }, { "epoch": 0.3990583122057226, "grad_norm": 1.8251564966751401, "learning_rate": 1.3667913507804892e-05, "loss": 0.7794, "step": 11018 }, { "epoch": 0.39909453096704095, "grad_norm": 1.3132341831306806, "learning_rate": 1.3666822173880861e-05, "loss": 0.7573, "step": 11019 }, { "epoch": 0.39913074972835927, "grad_norm": 1.4867314619782748, "learning_rate": 1.366573078949813e-05, "loss": 0.7649, "step": 11020 }, { "epoch": 0.39916696848967764, "grad_norm": 1.2476882818639081, "learning_rate": 1.3664639354671724e-05, "loss": 0.8285, "step": 11021 }, { "epoch": 0.399203187250996, "grad_norm": 1.4739058736428416, "learning_rate": 1.366354786941666e-05, "loss": 0.805, "step": 11022 }, { "epoch": 0.3992394060123144, "grad_norm": 0.9897216771279952, "learning_rate": 1.366245633374796e-05, "loss": 0.734, "step": 11023 }, { "epoch": 0.39927562477363276, "grad_norm": 1.3245480389626199, "learning_rate": 1.366136474768064e-05, "loss": 0.7046, "step": 11024 }, { "epoch": 0.3993118435349511, "grad_norm": 1.4451961940931066, "learning_rate": 1.3660273111229727e-05, "loss": 0.8208, "step": 11025 }, { "epoch": 0.39934806229626946, "grad_norm": 1.514569503060437, "learning_rate": 1.365918142441024e-05, "loss": 0.7363, "step": 11026 }, { "epoch": 0.39938428105758783, "grad_norm": 1.1039114086729307, "learning_rate": 1.3658089687237197e-05, "loss": 0.725, "step": 11027 }, { "epoch": 0.3994204998189062, "grad_norm": 1.4193739666928722, "learning_rate": 1.365699789972563e-05, "loss": 0.7912, "step": 11028 }, { "epoch": 0.3994567185802246, "grad_norm": 1.3063820888215902, "learning_rate": 1.3655906061890557e-05, "loss": 0.8228, "step": 11029 }, { "epoch": 0.3994929373415429, "grad_norm": 1.3463101478208306, "learning_rate": 1.3654814173747005e-05, "loss": 0.7742, "step": 11030 }, { "epoch": 0.39952915610286127, "grad_norm": 1.4740102826748744, "learning_rate": 1.3653722235309996e-05, "loss": 0.7216, "step": 11031 }, { "epoch": 0.39956537486417965, "grad_norm": 1.3296529521667693, "learning_rate": 1.365263024659456e-05, "loss": 0.7414, "step": 11032 }, { "epoch": 0.399601593625498, "grad_norm": 1.4110173702868114, "learning_rate": 1.3651538207615724e-05, "loss": 0.8075, "step": 11033 }, { "epoch": 0.3996378123868164, "grad_norm": 1.6509049387303476, "learning_rate": 1.3650446118388509e-05, "loss": 0.8905, "step": 11034 }, { "epoch": 0.3996740311481347, "grad_norm": 1.1858647687594115, "learning_rate": 1.3649353978927952e-05, "loss": 0.7111, "step": 11035 }, { "epoch": 0.3997102499094531, "grad_norm": 1.2761158978554075, "learning_rate": 1.3648261789249077e-05, "loss": 0.7931, "step": 11036 }, { "epoch": 0.39974646867077146, "grad_norm": 1.4449914601471279, "learning_rate": 1.3647169549366917e-05, "loss": 0.771, "step": 11037 }, { "epoch": 0.39978268743208983, "grad_norm": 1.3735198404195263, "learning_rate": 1.3646077259296492e-05, "loss": 0.7836, "step": 11038 }, { "epoch": 0.3998189061934082, "grad_norm": 1.3584337675544846, "learning_rate": 1.3644984919052849e-05, "loss": 0.7625, "step": 11039 }, { "epoch": 0.3998551249547265, "grad_norm": 1.422781618555225, "learning_rate": 1.3643892528651004e-05, "loss": 0.7352, "step": 11040 }, { "epoch": 0.3998913437160449, "grad_norm": 1.354511249850902, "learning_rate": 1.3642800088105998e-05, "loss": 0.7681, "step": 11041 }, { "epoch": 0.3999275624773633, "grad_norm": 1.4600036448990872, "learning_rate": 1.3641707597432862e-05, "loss": 0.7718, "step": 11042 }, { "epoch": 0.39996378123868165, "grad_norm": 1.2941618259592045, "learning_rate": 1.3640615056646632e-05, "loss": 0.6812, "step": 11043 }, { "epoch": 0.4, "grad_norm": 1.4462318267418177, "learning_rate": 1.3639522465762336e-05, "loss": 0.7949, "step": 11044 }, { "epoch": 0.40003621876131834, "grad_norm": 1.2281403646921527, "learning_rate": 1.3638429824795014e-05, "loss": 0.7539, "step": 11045 }, { "epoch": 0.4000724375226367, "grad_norm": 1.3654269804318604, "learning_rate": 1.3637337133759703e-05, "loss": 0.7224, "step": 11046 }, { "epoch": 0.4001086562839551, "grad_norm": 1.4628580290927926, "learning_rate": 1.3636244392671434e-05, "loss": 0.8051, "step": 11047 }, { "epoch": 0.40014487504527346, "grad_norm": 1.415167671714331, "learning_rate": 1.3635151601545249e-05, "loss": 0.8357, "step": 11048 }, { "epoch": 0.40018109380659184, "grad_norm": 1.241876137670664, "learning_rate": 1.363405876039618e-05, "loss": 0.7602, "step": 11049 }, { "epoch": 0.40021731256791016, "grad_norm": 1.2076466385873112, "learning_rate": 1.3632965869239274e-05, "loss": 0.7463, "step": 11050 }, { "epoch": 0.40025353132922853, "grad_norm": 1.3609082708137912, "learning_rate": 1.3631872928089564e-05, "loss": 0.74, "step": 11051 }, { "epoch": 0.4002897500905469, "grad_norm": 1.5429794987171344, "learning_rate": 1.3630779936962092e-05, "loss": 0.6939, "step": 11052 }, { "epoch": 0.4003259688518653, "grad_norm": 1.8189127148763296, "learning_rate": 1.3629686895871896e-05, "loss": 0.7381, "step": 11053 }, { "epoch": 0.40036218761318365, "grad_norm": 1.3805418627440025, "learning_rate": 1.3628593804834021e-05, "loss": 0.754, "step": 11054 }, { "epoch": 0.40039840637450197, "grad_norm": 1.3182610789494635, "learning_rate": 1.3627500663863505e-05, "loss": 0.7515, "step": 11055 }, { "epoch": 0.40043462513582034, "grad_norm": 1.348516207825769, "learning_rate": 1.3626407472975393e-05, "loss": 0.7122, "step": 11056 }, { "epoch": 0.4004708438971387, "grad_norm": 1.4760734805434308, "learning_rate": 1.3625314232184727e-05, "loss": 0.8038, "step": 11057 }, { "epoch": 0.4005070626584571, "grad_norm": 1.415137844310321, "learning_rate": 1.3624220941506554e-05, "loss": 0.79, "step": 11058 }, { "epoch": 0.40054328141977547, "grad_norm": 1.3182563586076266, "learning_rate": 1.3623127600955914e-05, "loss": 0.6722, "step": 11059 }, { "epoch": 0.4005795001810938, "grad_norm": 1.2034493719695136, "learning_rate": 1.3622034210547856e-05, "loss": 0.7089, "step": 11060 }, { "epoch": 0.40061571894241216, "grad_norm": 1.579526926523975, "learning_rate": 1.3620940770297427e-05, "loss": 0.8179, "step": 11061 }, { "epoch": 0.40065193770373053, "grad_norm": 1.3943963586658141, "learning_rate": 1.3619847280219664e-05, "loss": 0.7859, "step": 11062 }, { "epoch": 0.4006881564650489, "grad_norm": 1.590107945040893, "learning_rate": 1.3618753740329629e-05, "loss": 0.8071, "step": 11063 }, { "epoch": 0.4007243752263673, "grad_norm": 1.287325554378934, "learning_rate": 1.361766015064236e-05, "loss": 0.763, "step": 11064 }, { "epoch": 0.4007605939876856, "grad_norm": 1.3615801357666328, "learning_rate": 1.3616566511172908e-05, "loss": 0.7385, "step": 11065 }, { "epoch": 0.400796812749004, "grad_norm": 1.0834186564103232, "learning_rate": 1.3615472821936324e-05, "loss": 0.7752, "step": 11066 }, { "epoch": 0.40083303151032235, "grad_norm": 1.4657056320129949, "learning_rate": 1.3614379082947657e-05, "loss": 0.7469, "step": 11067 }, { "epoch": 0.4008692502716407, "grad_norm": 1.4555876076736862, "learning_rate": 1.3613285294221959e-05, "loss": 0.7842, "step": 11068 }, { "epoch": 0.4009054690329591, "grad_norm": 1.2907556771309245, "learning_rate": 1.3612191455774277e-05, "loss": 0.769, "step": 11069 }, { "epoch": 0.4009416877942774, "grad_norm": 1.0163652027211654, "learning_rate": 1.3611097567619671e-05, "loss": 0.7664, "step": 11070 }, { "epoch": 0.4009779065555958, "grad_norm": 1.3097158382440564, "learning_rate": 1.3610003629773186e-05, "loss": 0.6654, "step": 11071 }, { "epoch": 0.40101412531691416, "grad_norm": 1.342048475672595, "learning_rate": 1.3608909642249882e-05, "loss": 0.7386, "step": 11072 }, { "epoch": 0.40105034407823253, "grad_norm": 1.3923741093552318, "learning_rate": 1.3607815605064807e-05, "loss": 0.7352, "step": 11073 }, { "epoch": 0.4010865628395509, "grad_norm": 1.4927379657126236, "learning_rate": 1.3606721518233021e-05, "loss": 0.7461, "step": 11074 }, { "epoch": 0.4011227816008692, "grad_norm": 1.4107651214974735, "learning_rate": 1.3605627381769577e-05, "loss": 0.7237, "step": 11075 }, { "epoch": 0.4011590003621876, "grad_norm": 1.4343311525623608, "learning_rate": 1.3604533195689532e-05, "loss": 0.704, "step": 11076 }, { "epoch": 0.401195219123506, "grad_norm": 1.1957859718987522, "learning_rate": 1.3603438960007943e-05, "loss": 0.8032, "step": 11077 }, { "epoch": 0.40123143788482435, "grad_norm": 1.465942103912965, "learning_rate": 1.3602344674739868e-05, "loss": 0.8631, "step": 11078 }, { "epoch": 0.4012676566461427, "grad_norm": 1.3683553625925158, "learning_rate": 1.3601250339900367e-05, "loss": 0.7674, "step": 11079 }, { "epoch": 0.40130387540746104, "grad_norm": 1.4478430735000736, "learning_rate": 1.3600155955504496e-05, "loss": 0.8198, "step": 11080 }, { "epoch": 0.4013400941687794, "grad_norm": 1.381635318386768, "learning_rate": 1.3599061521567317e-05, "loss": 0.8604, "step": 11081 }, { "epoch": 0.4013763129300978, "grad_norm": 1.600825014349262, "learning_rate": 1.3597967038103885e-05, "loss": 0.7685, "step": 11082 }, { "epoch": 0.40141253169141616, "grad_norm": 1.188893605081772, "learning_rate": 1.3596872505129269e-05, "loss": 0.7921, "step": 11083 }, { "epoch": 0.40144875045273454, "grad_norm": 1.507683691640256, "learning_rate": 1.3595777922658524e-05, "loss": 0.8132, "step": 11084 }, { "epoch": 0.40148496921405286, "grad_norm": 1.3849997409973522, "learning_rate": 1.3594683290706714e-05, "loss": 0.8541, "step": 11085 }, { "epoch": 0.40152118797537123, "grad_norm": 1.420435325292036, "learning_rate": 1.3593588609288909e-05, "loss": 0.7071, "step": 11086 }, { "epoch": 0.4015574067366896, "grad_norm": 1.0861932792102191, "learning_rate": 1.3592493878420161e-05, "loss": 0.82, "step": 11087 }, { "epoch": 0.401593625498008, "grad_norm": 1.3695084068343055, "learning_rate": 1.3591399098115545e-05, "loss": 0.8181, "step": 11088 }, { "epoch": 0.40162984425932635, "grad_norm": 1.4243356136508607, "learning_rate": 1.3590304268390118e-05, "loss": 0.7925, "step": 11089 }, { "epoch": 0.40166606302064467, "grad_norm": 1.181234580808498, "learning_rate": 1.3589209389258953e-05, "loss": 0.7516, "step": 11090 }, { "epoch": 0.40170228178196304, "grad_norm": 1.3454307839766515, "learning_rate": 1.3588114460737112e-05, "loss": 0.8148, "step": 11091 }, { "epoch": 0.4017385005432814, "grad_norm": 1.1981923006285389, "learning_rate": 1.3587019482839664e-05, "loss": 0.7192, "step": 11092 }, { "epoch": 0.4017747193045998, "grad_norm": 1.1778699447327923, "learning_rate": 1.3585924455581672e-05, "loss": 0.7546, "step": 11093 }, { "epoch": 0.40181093806591817, "grad_norm": 1.6665950284154973, "learning_rate": 1.3584829378978213e-05, "loss": 0.829, "step": 11094 }, { "epoch": 0.4018471568272365, "grad_norm": 1.4046032078044954, "learning_rate": 1.3583734253044349e-05, "loss": 0.7503, "step": 11095 }, { "epoch": 0.40188337558855486, "grad_norm": 1.399153391977227, "learning_rate": 1.3582639077795154e-05, "loss": 0.8126, "step": 11096 }, { "epoch": 0.40191959434987323, "grad_norm": 1.3115043120363776, "learning_rate": 1.3581543853245698e-05, "loss": 0.6772, "step": 11097 }, { "epoch": 0.4019558131111916, "grad_norm": 1.4977644762346598, "learning_rate": 1.358044857941105e-05, "loss": 0.6677, "step": 11098 }, { "epoch": 0.40199203187251, "grad_norm": 1.3163234001202564, "learning_rate": 1.3579353256306287e-05, "loss": 0.7684, "step": 11099 }, { "epoch": 0.4020282506338283, "grad_norm": 1.1739715319575972, "learning_rate": 1.3578257883946473e-05, "loss": 0.7823, "step": 11100 }, { "epoch": 0.4020644693951467, "grad_norm": 1.5238926433809228, "learning_rate": 1.3577162462346692e-05, "loss": 0.7732, "step": 11101 }, { "epoch": 0.40210068815646505, "grad_norm": 1.3269639802791362, "learning_rate": 1.3576066991522008e-05, "loss": 0.7445, "step": 11102 }, { "epoch": 0.4021369069177834, "grad_norm": 1.2919760135004439, "learning_rate": 1.3574971471487502e-05, "loss": 0.7377, "step": 11103 }, { "epoch": 0.4021731256791018, "grad_norm": 1.1190283658278029, "learning_rate": 1.3573875902258244e-05, "loss": 0.7581, "step": 11104 }, { "epoch": 0.4022093444404201, "grad_norm": 1.3962901660617615, "learning_rate": 1.3572780283849318e-05, "loss": 0.8324, "step": 11105 }, { "epoch": 0.4022455632017385, "grad_norm": 1.245989520846928, "learning_rate": 1.3571684616275792e-05, "loss": 0.7165, "step": 11106 }, { "epoch": 0.40228178196305686, "grad_norm": 1.4072374056226942, "learning_rate": 1.357058889955275e-05, "loss": 0.7985, "step": 11107 }, { "epoch": 0.40231800072437524, "grad_norm": 1.3428734426538207, "learning_rate": 1.3569493133695266e-05, "loss": 0.7266, "step": 11108 }, { "epoch": 0.4023542194856936, "grad_norm": 1.1350690067623947, "learning_rate": 1.356839731871842e-05, "loss": 0.783, "step": 11109 }, { "epoch": 0.40239043824701193, "grad_norm": 0.9989275049572688, "learning_rate": 1.3567301454637297e-05, "loss": 0.7094, "step": 11110 }, { "epoch": 0.4024266570083303, "grad_norm": 1.353633611939352, "learning_rate": 1.3566205541466966e-05, "loss": 0.7918, "step": 11111 }, { "epoch": 0.4024628757696487, "grad_norm": 1.443897584446016, "learning_rate": 1.3565109579222511e-05, "loss": 0.7961, "step": 11112 }, { "epoch": 0.40249909453096705, "grad_norm": 1.1565723281024105, "learning_rate": 1.356401356791902e-05, "loss": 0.6852, "step": 11113 }, { "epoch": 0.4025353132922854, "grad_norm": 1.2855212278888133, "learning_rate": 1.3562917507571566e-05, "loss": 0.7879, "step": 11114 }, { "epoch": 0.40257153205360374, "grad_norm": 1.2355344995929047, "learning_rate": 1.3561821398195242e-05, "loss": 0.8052, "step": 11115 }, { "epoch": 0.4026077508149221, "grad_norm": 1.0495004351657047, "learning_rate": 1.3560725239805121e-05, "loss": 0.7638, "step": 11116 }, { "epoch": 0.4026439695762405, "grad_norm": 1.233395490771207, "learning_rate": 1.3559629032416297e-05, "loss": 0.7549, "step": 11117 }, { "epoch": 0.40268018833755886, "grad_norm": 0.9876519428735445, "learning_rate": 1.3558532776043846e-05, "loss": 0.7641, "step": 11118 }, { "epoch": 0.40271640709887724, "grad_norm": 1.0935289899909604, "learning_rate": 1.3557436470702858e-05, "loss": 0.7944, "step": 11119 }, { "epoch": 0.40275262586019556, "grad_norm": 1.4685898044150452, "learning_rate": 1.355634011640842e-05, "loss": 0.7688, "step": 11120 }, { "epoch": 0.40278884462151393, "grad_norm": 0.9462722288667762, "learning_rate": 1.3555243713175617e-05, "loss": 0.7181, "step": 11121 }, { "epoch": 0.4028250633828323, "grad_norm": 1.5311045139641726, "learning_rate": 1.3554147261019535e-05, "loss": 0.7742, "step": 11122 }, { "epoch": 0.4028612821441507, "grad_norm": 1.3961188818404493, "learning_rate": 1.3553050759955262e-05, "loss": 0.813, "step": 11123 }, { "epoch": 0.40289750090546905, "grad_norm": 1.3360765117130564, "learning_rate": 1.3551954209997891e-05, "loss": 0.7736, "step": 11124 }, { "epoch": 0.40293371966678737, "grad_norm": 1.1240504709107493, "learning_rate": 1.3550857611162508e-05, "loss": 0.7452, "step": 11125 }, { "epoch": 0.40296993842810575, "grad_norm": 0.9578664352228144, "learning_rate": 1.3549760963464205e-05, "loss": 0.7361, "step": 11126 }, { "epoch": 0.4030061571894241, "grad_norm": 1.397492138496305, "learning_rate": 1.3548664266918072e-05, "loss": 0.7572, "step": 11127 }, { "epoch": 0.4030423759507425, "grad_norm": 1.6910713791550502, "learning_rate": 1.3547567521539202e-05, "loss": 0.7086, "step": 11128 }, { "epoch": 0.40307859471206087, "grad_norm": 1.5145831509064884, "learning_rate": 1.3546470727342682e-05, "loss": 0.7926, "step": 11129 }, { "epoch": 0.4031148134733792, "grad_norm": 1.5096376952122927, "learning_rate": 1.3545373884343614e-05, "loss": 0.8197, "step": 11130 }, { "epoch": 0.40315103223469756, "grad_norm": 1.6720234802749299, "learning_rate": 1.354427699255708e-05, "loss": 0.7621, "step": 11131 }, { "epoch": 0.40318725099601593, "grad_norm": 1.3004163251301293, "learning_rate": 1.3543180051998182e-05, "loss": 0.7638, "step": 11132 }, { "epoch": 0.4032234697573343, "grad_norm": 1.3594384969893594, "learning_rate": 1.3542083062682016e-05, "loss": 0.7213, "step": 11133 }, { "epoch": 0.4032596885186527, "grad_norm": 1.4288615670786216, "learning_rate": 1.3540986024623673e-05, "loss": 0.7584, "step": 11134 }, { "epoch": 0.403295907279971, "grad_norm": 1.841903024084236, "learning_rate": 1.3539888937838253e-05, "loss": 0.815, "step": 11135 }, { "epoch": 0.4033321260412894, "grad_norm": 1.3761023138398534, "learning_rate": 1.3538791802340846e-05, "loss": 0.7497, "step": 11136 }, { "epoch": 0.40336834480260775, "grad_norm": 1.0961216302214452, "learning_rate": 1.353769461814656e-05, "loss": 0.7431, "step": 11137 }, { "epoch": 0.4034045635639261, "grad_norm": 1.3572098600727025, "learning_rate": 1.3536597385270482e-05, "loss": 0.7764, "step": 11138 }, { "epoch": 0.4034407823252445, "grad_norm": 1.064040608704279, "learning_rate": 1.3535500103727722e-05, "loss": 0.7357, "step": 11139 }, { "epoch": 0.4034770010865628, "grad_norm": 1.145107095284097, "learning_rate": 1.3534402773533368e-05, "loss": 0.7777, "step": 11140 }, { "epoch": 0.4035132198478812, "grad_norm": 1.2903030776529576, "learning_rate": 1.3533305394702532e-05, "loss": 0.7005, "step": 11141 }, { "epoch": 0.40354943860919956, "grad_norm": 1.3619752882184397, "learning_rate": 1.3532207967250305e-05, "loss": 0.7432, "step": 11142 }, { "epoch": 0.40358565737051794, "grad_norm": 1.462354949030326, "learning_rate": 1.3531110491191793e-05, "loss": 0.7314, "step": 11143 }, { "epoch": 0.4036218761318363, "grad_norm": 1.512397669106781, "learning_rate": 1.35300129665421e-05, "loss": 0.8187, "step": 11144 }, { "epoch": 0.40365809489315463, "grad_norm": 1.2701898194472396, "learning_rate": 1.3528915393316327e-05, "loss": 0.7053, "step": 11145 }, { "epoch": 0.403694313654473, "grad_norm": 1.3195706408619314, "learning_rate": 1.3527817771529577e-05, "loss": 0.7824, "step": 11146 }, { "epoch": 0.4037305324157914, "grad_norm": 1.3800131040426935, "learning_rate": 1.3526720101196953e-05, "loss": 0.7921, "step": 11147 }, { "epoch": 0.40376675117710975, "grad_norm": 1.4400489135740508, "learning_rate": 1.3525622382333562e-05, "loss": 0.753, "step": 11148 }, { "epoch": 0.4038029699384281, "grad_norm": 1.6107643181427789, "learning_rate": 1.352452461495451e-05, "loss": 0.7799, "step": 11149 }, { "epoch": 0.40383918869974644, "grad_norm": 1.4606933172941003, "learning_rate": 1.3523426799074903e-05, "loss": 0.6983, "step": 11150 }, { "epoch": 0.4038754074610648, "grad_norm": 1.505572042221454, "learning_rate": 1.3522328934709847e-05, "loss": 0.7278, "step": 11151 }, { "epoch": 0.4039116262223832, "grad_norm": 1.3334710409687796, "learning_rate": 1.352123102187445e-05, "loss": 0.7604, "step": 11152 }, { "epoch": 0.40394784498370157, "grad_norm": 1.6085900735095169, "learning_rate": 1.3520133060583818e-05, "loss": 0.7727, "step": 11153 }, { "epoch": 0.40398406374501994, "grad_norm": 1.3797490094882536, "learning_rate": 1.3519035050853063e-05, "loss": 0.7841, "step": 11154 }, { "epoch": 0.40402028250633826, "grad_norm": 1.4140830466584948, "learning_rate": 1.3517936992697296e-05, "loss": 0.7664, "step": 11155 }, { "epoch": 0.40405650126765663, "grad_norm": 1.4483497705711634, "learning_rate": 1.351683888613162e-05, "loss": 0.7218, "step": 11156 }, { "epoch": 0.404092720028975, "grad_norm": 1.396128839970801, "learning_rate": 1.3515740731171157e-05, "loss": 0.8168, "step": 11157 }, { "epoch": 0.4041289387902934, "grad_norm": 1.4222791825643375, "learning_rate": 1.3514642527831008e-05, "loss": 0.8067, "step": 11158 }, { "epoch": 0.40416515755161175, "grad_norm": 1.4359317912257037, "learning_rate": 1.3513544276126292e-05, "loss": 0.7816, "step": 11159 }, { "epoch": 0.40420137631293007, "grad_norm": 1.4445273689684148, "learning_rate": 1.3512445976072114e-05, "loss": 0.7261, "step": 11160 }, { "epoch": 0.40423759507424845, "grad_norm": 1.4647486961042029, "learning_rate": 1.35113476276836e-05, "loss": 0.7845, "step": 11161 }, { "epoch": 0.4042738138355668, "grad_norm": 1.4382510322450726, "learning_rate": 1.3510249230975853e-05, "loss": 0.8208, "step": 11162 }, { "epoch": 0.4043100325968852, "grad_norm": 1.365228633841257, "learning_rate": 1.3509150785963994e-05, "loss": 0.7377, "step": 11163 }, { "epoch": 0.40434625135820357, "grad_norm": 1.2732412782760325, "learning_rate": 1.3508052292663137e-05, "loss": 0.8229, "step": 11164 }, { "epoch": 0.4043824701195219, "grad_norm": 1.2598083917618281, "learning_rate": 1.3506953751088396e-05, "loss": 0.7642, "step": 11165 }, { "epoch": 0.40441868888084026, "grad_norm": 1.559036595906772, "learning_rate": 1.3505855161254893e-05, "loss": 0.6789, "step": 11166 }, { "epoch": 0.40445490764215863, "grad_norm": 1.4237405922753048, "learning_rate": 1.350475652317774e-05, "loss": 0.8257, "step": 11167 }, { "epoch": 0.404491126403477, "grad_norm": 1.5544545847417195, "learning_rate": 1.350365783687206e-05, "loss": 0.8725, "step": 11168 }, { "epoch": 0.4045273451647954, "grad_norm": 1.1055481699677068, "learning_rate": 1.3502559102352967e-05, "loss": 0.7288, "step": 11169 }, { "epoch": 0.4045635639261137, "grad_norm": 1.121764963512095, "learning_rate": 1.3501460319635584e-05, "loss": 0.7139, "step": 11170 }, { "epoch": 0.4045997826874321, "grad_norm": 1.2996980944403695, "learning_rate": 1.350036148873503e-05, "loss": 0.6987, "step": 11171 }, { "epoch": 0.40463600144875045, "grad_norm": 1.514965603380125, "learning_rate": 1.3499262609666428e-05, "loss": 0.7248, "step": 11172 }, { "epoch": 0.4046722202100688, "grad_norm": 1.36949158318388, "learning_rate": 1.3498163682444896e-05, "loss": 0.7603, "step": 11173 }, { "epoch": 0.4047084389713872, "grad_norm": 1.5219223912456352, "learning_rate": 1.3497064707085557e-05, "loss": 0.7048, "step": 11174 }, { "epoch": 0.4047446577327055, "grad_norm": 1.3654495187071467, "learning_rate": 1.3495965683603537e-05, "loss": 0.7885, "step": 11175 }, { "epoch": 0.4047808764940239, "grad_norm": 1.293752948929241, "learning_rate": 1.3494866612013956e-05, "loss": 0.6588, "step": 11176 }, { "epoch": 0.40481709525534226, "grad_norm": 1.1255913797891028, "learning_rate": 1.3493767492331942e-05, "loss": 0.7602, "step": 11177 }, { "epoch": 0.40485331401666064, "grad_norm": 1.4867938235717975, "learning_rate": 1.3492668324572615e-05, "loss": 0.7647, "step": 11178 }, { "epoch": 0.404889532777979, "grad_norm": 1.3772505797379582, "learning_rate": 1.3491569108751106e-05, "loss": 0.803, "step": 11179 }, { "epoch": 0.40492575153929733, "grad_norm": 1.4081212865193822, "learning_rate": 1.3490469844882535e-05, "loss": 0.7374, "step": 11180 }, { "epoch": 0.4049619703006157, "grad_norm": 0.9251403622357497, "learning_rate": 1.3489370532982034e-05, "loss": 0.6984, "step": 11181 }, { "epoch": 0.4049981890619341, "grad_norm": 1.4599033944007518, "learning_rate": 1.3488271173064725e-05, "loss": 0.8178, "step": 11182 }, { "epoch": 0.40503440782325245, "grad_norm": 1.6533174306322769, "learning_rate": 1.3487171765145743e-05, "loss": 0.7134, "step": 11183 }, { "epoch": 0.4050706265845708, "grad_norm": 1.3623898674371289, "learning_rate": 1.3486072309240212e-05, "loss": 0.7977, "step": 11184 }, { "epoch": 0.40510684534588914, "grad_norm": 1.2913387142460127, "learning_rate": 1.3484972805363262e-05, "loss": 0.7517, "step": 11185 }, { "epoch": 0.4051430641072075, "grad_norm": 1.4946899033979137, "learning_rate": 1.3483873253530027e-05, "loss": 0.8326, "step": 11186 }, { "epoch": 0.4051792828685259, "grad_norm": 1.2870844938452133, "learning_rate": 1.348277365375563e-05, "loss": 0.7662, "step": 11187 }, { "epoch": 0.40521550162984427, "grad_norm": 1.3632837612489672, "learning_rate": 1.3481674006055212e-05, "loss": 0.7793, "step": 11188 }, { "epoch": 0.40525172039116264, "grad_norm": 1.2649322758291135, "learning_rate": 1.3480574310443897e-05, "loss": 0.7527, "step": 11189 }, { "epoch": 0.40528793915248096, "grad_norm": 1.5817740525644264, "learning_rate": 1.3479474566936825e-05, "loss": 0.7609, "step": 11190 }, { "epoch": 0.40532415791379933, "grad_norm": 1.3623466328005123, "learning_rate": 1.3478374775549121e-05, "loss": 0.8266, "step": 11191 }, { "epoch": 0.4053603766751177, "grad_norm": 1.400550186771255, "learning_rate": 1.3477274936295928e-05, "loss": 0.7257, "step": 11192 }, { "epoch": 0.4053965954364361, "grad_norm": 1.3944368330464187, "learning_rate": 1.3476175049192373e-05, "loss": 0.7717, "step": 11193 }, { "epoch": 0.40543281419775445, "grad_norm": 1.3413364538042702, "learning_rate": 1.3475075114253597e-05, "loss": 0.7894, "step": 11194 }, { "epoch": 0.4054690329590728, "grad_norm": 1.4552360141781986, "learning_rate": 1.3473975131494735e-05, "loss": 0.7913, "step": 11195 }, { "epoch": 0.40550525172039115, "grad_norm": 1.306324306826033, "learning_rate": 1.347287510093092e-05, "loss": 0.7451, "step": 11196 }, { "epoch": 0.4055414704817095, "grad_norm": 1.1859954653952203, "learning_rate": 1.3471775022577294e-05, "loss": 0.7738, "step": 11197 }, { "epoch": 0.4055776892430279, "grad_norm": 1.2228264667316322, "learning_rate": 1.347067489644899e-05, "loss": 0.7508, "step": 11198 }, { "epoch": 0.40561390800434627, "grad_norm": 1.4433416030277468, "learning_rate": 1.3469574722561154e-05, "loss": 0.8162, "step": 11199 }, { "epoch": 0.4056501267656646, "grad_norm": 1.5233332158121855, "learning_rate": 1.3468474500928919e-05, "loss": 0.8158, "step": 11200 }, { "epoch": 0.40568634552698296, "grad_norm": 1.403982308959205, "learning_rate": 1.3467374231567427e-05, "loss": 0.8222, "step": 11201 }, { "epoch": 0.40572256428830134, "grad_norm": 1.00905536750256, "learning_rate": 1.3466273914491818e-05, "loss": 0.68, "step": 11202 }, { "epoch": 0.4057587830496197, "grad_norm": 1.4158165383969445, "learning_rate": 1.346517354971724e-05, "loss": 0.7722, "step": 11203 }, { "epoch": 0.4057950018109381, "grad_norm": 1.083468079097341, "learning_rate": 1.3464073137258823e-05, "loss": 0.6663, "step": 11204 }, { "epoch": 0.4058312205722564, "grad_norm": 1.0949987085773456, "learning_rate": 1.3462972677131716e-05, "loss": 0.7924, "step": 11205 }, { "epoch": 0.4058674393335748, "grad_norm": 1.4546677830084094, "learning_rate": 1.3461872169351064e-05, "loss": 0.8363, "step": 11206 }, { "epoch": 0.40590365809489315, "grad_norm": 1.3856892820962525, "learning_rate": 1.3460771613932008e-05, "loss": 0.7662, "step": 11207 }, { "epoch": 0.4059398768562115, "grad_norm": 1.514379727147211, "learning_rate": 1.3459671010889696e-05, "loss": 0.734, "step": 11208 }, { "epoch": 0.4059760956175299, "grad_norm": 1.1595445708827357, "learning_rate": 1.3458570360239268e-05, "loss": 0.7993, "step": 11209 }, { "epoch": 0.4060123143788482, "grad_norm": 1.388715802676924, "learning_rate": 1.3457469661995877e-05, "loss": 0.763, "step": 11210 }, { "epoch": 0.4060485331401666, "grad_norm": 1.5578972747008104, "learning_rate": 1.345636891617466e-05, "loss": 0.7547, "step": 11211 }, { "epoch": 0.40608475190148496, "grad_norm": 1.13201483032945, "learning_rate": 1.3455268122790774e-05, "loss": 0.7078, "step": 11212 }, { "epoch": 0.40612097066280334, "grad_norm": 1.445768381726492, "learning_rate": 1.3454167281859363e-05, "loss": 0.7345, "step": 11213 }, { "epoch": 0.4061571894241217, "grad_norm": 1.3571003546823188, "learning_rate": 1.3453066393395572e-05, "loss": 0.8095, "step": 11214 }, { "epoch": 0.40619340818544003, "grad_norm": 1.1566117802436933, "learning_rate": 1.3451965457414558e-05, "loss": 0.7565, "step": 11215 }, { "epoch": 0.4062296269467584, "grad_norm": 1.669741809350115, "learning_rate": 1.345086447393146e-05, "loss": 0.7464, "step": 11216 }, { "epoch": 0.4062658457080768, "grad_norm": 1.563510710934828, "learning_rate": 1.3449763442961438e-05, "loss": 0.7708, "step": 11217 }, { "epoch": 0.40630206446939515, "grad_norm": 1.4202524873332016, "learning_rate": 1.3448662364519638e-05, "loss": 0.6974, "step": 11218 }, { "epoch": 0.4063382832307135, "grad_norm": 1.262965079106012, "learning_rate": 1.3447561238621216e-05, "loss": 0.7209, "step": 11219 }, { "epoch": 0.4063745019920319, "grad_norm": 1.4790034579065097, "learning_rate": 1.3446460065281318e-05, "loss": 0.7588, "step": 11220 }, { "epoch": 0.4064107207533502, "grad_norm": 1.359176458186915, "learning_rate": 1.3445358844515105e-05, "loss": 0.7414, "step": 11221 }, { "epoch": 0.4064469395146686, "grad_norm": 1.4295291789359932, "learning_rate": 1.3444257576337728e-05, "loss": 0.789, "step": 11222 }, { "epoch": 0.40648315827598697, "grad_norm": 1.5541763447795987, "learning_rate": 1.3443156260764336e-05, "loss": 0.693, "step": 11223 }, { "epoch": 0.40651937703730534, "grad_norm": 1.0717915144908974, "learning_rate": 1.3442054897810093e-05, "loss": 0.8158, "step": 11224 }, { "epoch": 0.4065555957986237, "grad_norm": 1.424868077651819, "learning_rate": 1.3440953487490145e-05, "loss": 0.7932, "step": 11225 }, { "epoch": 0.40659181455994203, "grad_norm": 1.3345017846980083, "learning_rate": 1.3439852029819658e-05, "loss": 0.7392, "step": 11226 }, { "epoch": 0.4066280333212604, "grad_norm": 1.0429341087455124, "learning_rate": 1.3438750524813781e-05, "loss": 0.7371, "step": 11227 }, { "epoch": 0.4066642520825788, "grad_norm": 1.4678536683145211, "learning_rate": 1.3437648972487681e-05, "loss": 0.8029, "step": 11228 }, { "epoch": 0.40670047084389715, "grad_norm": 1.3602197142304944, "learning_rate": 1.3436547372856503e-05, "loss": 0.7696, "step": 11229 }, { "epoch": 0.40673668960521553, "grad_norm": 1.3404187916120096, "learning_rate": 1.3435445725935422e-05, "loss": 0.6896, "step": 11230 }, { "epoch": 0.40677290836653385, "grad_norm": 1.0668448587198027, "learning_rate": 1.3434344031739581e-05, "loss": 0.7209, "step": 11231 }, { "epoch": 0.4068091271278522, "grad_norm": 1.3855345054904582, "learning_rate": 1.343324229028415e-05, "loss": 0.8012, "step": 11232 }, { "epoch": 0.4068453458891706, "grad_norm": 1.4609083068195232, "learning_rate": 1.3432140501584293e-05, "loss": 0.8032, "step": 11233 }, { "epoch": 0.40688156465048897, "grad_norm": 1.6550787662147428, "learning_rate": 1.3431038665655163e-05, "loss": 0.7471, "step": 11234 }, { "epoch": 0.40691778341180734, "grad_norm": 1.351393215652689, "learning_rate": 1.3429936782511927e-05, "loss": 0.7204, "step": 11235 }, { "epoch": 0.40695400217312566, "grad_norm": 1.510212486061695, "learning_rate": 1.3428834852169747e-05, "loss": 0.8022, "step": 11236 }, { "epoch": 0.40699022093444404, "grad_norm": 1.4947954918123498, "learning_rate": 1.3427732874643788e-05, "loss": 0.7295, "step": 11237 }, { "epoch": 0.4070264396957624, "grad_norm": 1.3889717188779447, "learning_rate": 1.3426630849949212e-05, "loss": 0.7695, "step": 11238 }, { "epoch": 0.4070626584570808, "grad_norm": 1.3607306126890297, "learning_rate": 1.3425528778101184e-05, "loss": 0.7414, "step": 11239 }, { "epoch": 0.40709887721839916, "grad_norm": 1.3247843736314067, "learning_rate": 1.342442665911487e-05, "loss": 0.7573, "step": 11240 }, { "epoch": 0.4071350959797175, "grad_norm": 1.0136931071809496, "learning_rate": 1.3423324493005438e-05, "loss": 0.7785, "step": 11241 }, { "epoch": 0.40717131474103585, "grad_norm": 1.6348912545154526, "learning_rate": 1.342222227978805e-05, "loss": 0.8171, "step": 11242 }, { "epoch": 0.4072075335023542, "grad_norm": 1.26195635349258, "learning_rate": 1.3421120019477877e-05, "loss": 0.7525, "step": 11243 }, { "epoch": 0.4072437522636726, "grad_norm": 1.378511192582729, "learning_rate": 1.342001771209009e-05, "loss": 0.7888, "step": 11244 }, { "epoch": 0.40727997102499097, "grad_norm": 1.4498442247332208, "learning_rate": 1.341891535763985e-05, "loss": 0.794, "step": 11245 }, { "epoch": 0.4073161897863093, "grad_norm": 0.992082376653363, "learning_rate": 1.3417812956142333e-05, "loss": 0.7776, "step": 11246 }, { "epoch": 0.40735240854762766, "grad_norm": 1.1424399402393723, "learning_rate": 1.3416710507612705e-05, "loss": 0.7517, "step": 11247 }, { "epoch": 0.40738862730894604, "grad_norm": 1.5150922829605722, "learning_rate": 1.341560801206614e-05, "loss": 0.8161, "step": 11248 }, { "epoch": 0.4074248460702644, "grad_norm": 1.3845111891609776, "learning_rate": 1.3414505469517807e-05, "loss": 0.7155, "step": 11249 }, { "epoch": 0.4074610648315828, "grad_norm": 1.0136173868510558, "learning_rate": 1.3413402879982878e-05, "loss": 0.7822, "step": 11250 }, { "epoch": 0.4074972835929011, "grad_norm": 1.4908058275748264, "learning_rate": 1.3412300243476526e-05, "loss": 0.8044, "step": 11251 }, { "epoch": 0.4075335023542195, "grad_norm": 1.3693747123067668, "learning_rate": 1.3411197560013922e-05, "loss": 0.6973, "step": 11252 }, { "epoch": 0.40756972111553785, "grad_norm": 1.2069011721051668, "learning_rate": 1.3410094829610248e-05, "loss": 0.763, "step": 11253 }, { "epoch": 0.4076059398768562, "grad_norm": 1.0900545532707846, "learning_rate": 1.3408992052280667e-05, "loss": 0.8073, "step": 11254 }, { "epoch": 0.4076421586381746, "grad_norm": 1.132204424040942, "learning_rate": 1.3407889228040366e-05, "loss": 0.7665, "step": 11255 }, { "epoch": 0.4076783773994929, "grad_norm": 1.277152493723174, "learning_rate": 1.3406786356904509e-05, "loss": 0.7845, "step": 11256 }, { "epoch": 0.4077145961608113, "grad_norm": 1.351450035482137, "learning_rate": 1.3405683438888281e-05, "loss": 0.8236, "step": 11257 }, { "epoch": 0.40775081492212967, "grad_norm": 1.3784158566790137, "learning_rate": 1.3404580474006856e-05, "loss": 0.7812, "step": 11258 }, { "epoch": 0.40778703368344804, "grad_norm": 1.5909087950919856, "learning_rate": 1.3403477462275412e-05, "loss": 0.7523, "step": 11259 }, { "epoch": 0.4078232524447664, "grad_norm": 1.4263262216701869, "learning_rate": 1.3402374403709126e-05, "loss": 0.7931, "step": 11260 }, { "epoch": 0.40785947120608473, "grad_norm": 1.5176236502988039, "learning_rate": 1.3401271298323182e-05, "loss": 0.7946, "step": 11261 }, { "epoch": 0.4078956899674031, "grad_norm": 1.1562770462477363, "learning_rate": 1.3400168146132755e-05, "loss": 0.7716, "step": 11262 }, { "epoch": 0.4079319087287215, "grad_norm": 1.1697308763754741, "learning_rate": 1.3399064947153024e-05, "loss": 0.7416, "step": 11263 }, { "epoch": 0.40796812749003986, "grad_norm": 1.5827001029021484, "learning_rate": 1.3397961701399175e-05, "loss": 0.7729, "step": 11264 }, { "epoch": 0.40800434625135823, "grad_norm": 1.767190688308071, "learning_rate": 1.3396858408886386e-05, "loss": 0.8466, "step": 11265 }, { "epoch": 0.40804056501267655, "grad_norm": 1.6283561729132825, "learning_rate": 1.3395755069629845e-05, "loss": 0.7484, "step": 11266 }, { "epoch": 0.4080767837739949, "grad_norm": 1.3439847296195595, "learning_rate": 1.3394651683644727e-05, "loss": 0.7447, "step": 11267 }, { "epoch": 0.4081130025353133, "grad_norm": 1.1557585320753125, "learning_rate": 1.3393548250946217e-05, "loss": 0.7908, "step": 11268 }, { "epoch": 0.40814922129663167, "grad_norm": 1.5267581713508769, "learning_rate": 1.3392444771549504e-05, "loss": 0.7917, "step": 11269 }, { "epoch": 0.40818544005795004, "grad_norm": 1.472912113516738, "learning_rate": 1.339134124546977e-05, "loss": 0.743, "step": 11270 }, { "epoch": 0.40822165881926836, "grad_norm": 1.276895399180419, "learning_rate": 1.33902376727222e-05, "loss": 0.7592, "step": 11271 }, { "epoch": 0.40825787758058674, "grad_norm": 1.5069253522441015, "learning_rate": 1.3389134053321981e-05, "loss": 0.7621, "step": 11272 }, { "epoch": 0.4082940963419051, "grad_norm": 1.4889599374012383, "learning_rate": 1.3388030387284299e-05, "loss": 0.8295, "step": 11273 }, { "epoch": 0.4083303151032235, "grad_norm": 1.3280048442230323, "learning_rate": 1.3386926674624341e-05, "loss": 0.7799, "step": 11274 }, { "epoch": 0.40836653386454186, "grad_norm": 1.4422221478477595, "learning_rate": 1.3385822915357298e-05, "loss": 0.7189, "step": 11275 }, { "epoch": 0.4084027526258602, "grad_norm": 1.085392217302016, "learning_rate": 1.3384719109498354e-05, "loss": 0.8012, "step": 11276 }, { "epoch": 0.40843897138717855, "grad_norm": 1.7669888177913151, "learning_rate": 1.3383615257062705e-05, "loss": 0.7743, "step": 11277 }, { "epoch": 0.4084751901484969, "grad_norm": 1.5333825850696807, "learning_rate": 1.3382511358065533e-05, "loss": 0.8267, "step": 11278 }, { "epoch": 0.4085114089098153, "grad_norm": 1.688863161196276, "learning_rate": 1.3381407412522036e-05, "loss": 0.7337, "step": 11279 }, { "epoch": 0.4085476276711337, "grad_norm": 1.4905999698876626, "learning_rate": 1.33803034204474e-05, "loss": 0.7197, "step": 11280 }, { "epoch": 0.408583846432452, "grad_norm": 1.572178694886717, "learning_rate": 1.337919938185682e-05, "loss": 0.7549, "step": 11281 }, { "epoch": 0.40862006519377037, "grad_norm": 1.2095833255339468, "learning_rate": 1.3378095296765485e-05, "loss": 0.7653, "step": 11282 }, { "epoch": 0.40865628395508874, "grad_norm": 1.1683607185289724, "learning_rate": 1.337699116518859e-05, "loss": 0.7463, "step": 11283 }, { "epoch": 0.4086925027164071, "grad_norm": 1.7702094793014824, "learning_rate": 1.3375886987141331e-05, "loss": 0.8571, "step": 11284 }, { "epoch": 0.4087287214777255, "grad_norm": 1.503349248115248, "learning_rate": 1.3374782762638901e-05, "loss": 0.7908, "step": 11285 }, { "epoch": 0.4087649402390438, "grad_norm": 1.2986844750608753, "learning_rate": 1.3373678491696495e-05, "loss": 0.8159, "step": 11286 }, { "epoch": 0.4088011590003622, "grad_norm": 1.4365263536269348, "learning_rate": 1.3372574174329306e-05, "loss": 0.8487, "step": 11287 }, { "epoch": 0.40883737776168055, "grad_norm": 1.4359764883540962, "learning_rate": 1.3371469810552538e-05, "loss": 0.7342, "step": 11288 }, { "epoch": 0.40887359652299893, "grad_norm": 1.4667049866808082, "learning_rate": 1.3370365400381377e-05, "loss": 0.7116, "step": 11289 }, { "epoch": 0.4089098152843173, "grad_norm": 1.5153949479384952, "learning_rate": 1.3369260943831032e-05, "loss": 0.724, "step": 11290 }, { "epoch": 0.4089460340456356, "grad_norm": 1.2865617184628495, "learning_rate": 1.3368156440916693e-05, "loss": 0.6998, "step": 11291 }, { "epoch": 0.408982252806954, "grad_norm": 1.4405363321427973, "learning_rate": 1.3367051891653564e-05, "loss": 0.893, "step": 11292 }, { "epoch": 0.40901847156827237, "grad_norm": 1.4093375021794903, "learning_rate": 1.3365947296056838e-05, "loss": 0.7824, "step": 11293 }, { "epoch": 0.40905469032959074, "grad_norm": 1.3154973637136202, "learning_rate": 1.3364842654141723e-05, "loss": 0.7763, "step": 11294 }, { "epoch": 0.4090909090909091, "grad_norm": 1.4337729347256267, "learning_rate": 1.3363737965923415e-05, "loss": 0.7598, "step": 11295 }, { "epoch": 0.40912712785222743, "grad_norm": 1.3887388519037944, "learning_rate": 1.336263323141712e-05, "loss": 0.7871, "step": 11296 }, { "epoch": 0.4091633466135458, "grad_norm": 1.3963921923323024, "learning_rate": 1.3361528450638036e-05, "loss": 0.8218, "step": 11297 }, { "epoch": 0.4091995653748642, "grad_norm": 1.4497552342044289, "learning_rate": 1.3360423623601361e-05, "loss": 0.7602, "step": 11298 }, { "epoch": 0.40923578413618256, "grad_norm": 1.425130643419037, "learning_rate": 1.3359318750322312e-05, "loss": 0.789, "step": 11299 }, { "epoch": 0.40927200289750093, "grad_norm": 1.3503854717281003, "learning_rate": 1.335821383081608e-05, "loss": 0.7295, "step": 11300 }, { "epoch": 0.40930822165881925, "grad_norm": 1.1218925204285333, "learning_rate": 1.3357108865097881e-05, "loss": 0.7707, "step": 11301 }, { "epoch": 0.4093444404201376, "grad_norm": 1.3763188220052531, "learning_rate": 1.3356003853182908e-05, "loss": 0.7528, "step": 11302 }, { "epoch": 0.409380659181456, "grad_norm": 1.3719938236514957, "learning_rate": 1.3354898795086374e-05, "loss": 0.7446, "step": 11303 }, { "epoch": 0.40941687794277437, "grad_norm": 1.3070087064097438, "learning_rate": 1.3353793690823489e-05, "loss": 0.7184, "step": 11304 }, { "epoch": 0.40945309670409274, "grad_norm": 0.9929543419568488, "learning_rate": 1.335268854040945e-05, "loss": 0.7819, "step": 11305 }, { "epoch": 0.40948931546541106, "grad_norm": 0.9984082050346433, "learning_rate": 1.3351583343859475e-05, "loss": 0.7367, "step": 11306 }, { "epoch": 0.40952553422672944, "grad_norm": 1.4185368605718247, "learning_rate": 1.3350478101188767e-05, "loss": 0.7626, "step": 11307 }, { "epoch": 0.4095617529880478, "grad_norm": 1.4147863208420441, "learning_rate": 1.3349372812412537e-05, "loss": 0.7184, "step": 11308 }, { "epoch": 0.4095979717493662, "grad_norm": 1.501434822092517, "learning_rate": 1.3348267477545993e-05, "loss": 0.7793, "step": 11309 }, { "epoch": 0.40963419051068456, "grad_norm": 1.4877341684177927, "learning_rate": 1.3347162096604347e-05, "loss": 0.7674, "step": 11310 }, { "epoch": 0.4096704092720029, "grad_norm": 1.3015823775464557, "learning_rate": 1.334605666960281e-05, "loss": 0.7093, "step": 11311 }, { "epoch": 0.40970662803332125, "grad_norm": 1.442236863889878, "learning_rate": 1.3344951196556592e-05, "loss": 0.7869, "step": 11312 }, { "epoch": 0.4097428467946396, "grad_norm": 1.417034538288152, "learning_rate": 1.3343845677480906e-05, "loss": 0.8466, "step": 11313 }, { "epoch": 0.409779065555958, "grad_norm": 1.4809536177648992, "learning_rate": 1.3342740112390967e-05, "loss": 0.8202, "step": 11314 }, { "epoch": 0.4098152843172764, "grad_norm": 1.4426812371373066, "learning_rate": 1.3341634501301985e-05, "loss": 0.8161, "step": 11315 }, { "epoch": 0.4098515030785947, "grad_norm": 1.3219128480412339, "learning_rate": 1.3340528844229177e-05, "loss": 0.7623, "step": 11316 }, { "epoch": 0.40988772183991307, "grad_norm": 1.3514080028211624, "learning_rate": 1.3339423141187759e-05, "loss": 0.7737, "step": 11317 }, { "epoch": 0.40992394060123144, "grad_norm": 1.2733046880749326, "learning_rate": 1.3338317392192942e-05, "loss": 0.7528, "step": 11318 }, { "epoch": 0.4099601593625498, "grad_norm": 1.5048142057965983, "learning_rate": 1.3337211597259945e-05, "loss": 0.76, "step": 11319 }, { "epoch": 0.4099963781238682, "grad_norm": 1.301795136903217, "learning_rate": 1.3336105756403982e-05, "loss": 0.7583, "step": 11320 }, { "epoch": 0.4100325968851865, "grad_norm": 1.3939973700125632, "learning_rate": 1.3334999869640274e-05, "loss": 0.7966, "step": 11321 }, { "epoch": 0.4100688156465049, "grad_norm": 1.5026277420986642, "learning_rate": 1.3333893936984042e-05, "loss": 0.7455, "step": 11322 }, { "epoch": 0.41010503440782325, "grad_norm": 1.3530874620983688, "learning_rate": 1.3332787958450492e-05, "loss": 0.7949, "step": 11323 }, { "epoch": 0.41014125316914163, "grad_norm": 1.2176838542074968, "learning_rate": 1.3331681934054857e-05, "loss": 0.7114, "step": 11324 }, { "epoch": 0.41017747193046, "grad_norm": 1.6239302912991334, "learning_rate": 1.3330575863812347e-05, "loss": 0.7347, "step": 11325 }, { "epoch": 0.4102136906917783, "grad_norm": 1.279032317317346, "learning_rate": 1.332946974773819e-05, "loss": 0.712, "step": 11326 }, { "epoch": 0.4102499094530967, "grad_norm": 1.3455899257414208, "learning_rate": 1.3328363585847602e-05, "loss": 0.7326, "step": 11327 }, { "epoch": 0.41028612821441507, "grad_norm": 1.3424852068803506, "learning_rate": 1.3327257378155807e-05, "loss": 0.7334, "step": 11328 }, { "epoch": 0.41032234697573344, "grad_norm": 1.320480678783428, "learning_rate": 1.3326151124678026e-05, "loss": 0.7922, "step": 11329 }, { "epoch": 0.4103585657370518, "grad_norm": 1.3332929411062273, "learning_rate": 1.3325044825429483e-05, "loss": 0.7506, "step": 11330 }, { "epoch": 0.41039478449837014, "grad_norm": 1.4510075461991943, "learning_rate": 1.3323938480425404e-05, "loss": 0.8128, "step": 11331 }, { "epoch": 0.4104310032596885, "grad_norm": 1.4520490813753657, "learning_rate": 1.3322832089681009e-05, "loss": 0.8056, "step": 11332 }, { "epoch": 0.4104672220210069, "grad_norm": 2.078417328233678, "learning_rate": 1.3321725653211526e-05, "loss": 0.6883, "step": 11333 }, { "epoch": 0.41050344078232526, "grad_norm": 1.1541370468378553, "learning_rate": 1.3320619171032176e-05, "loss": 0.7561, "step": 11334 }, { "epoch": 0.41053965954364363, "grad_norm": 1.4005627785692232, "learning_rate": 1.3319512643158193e-05, "loss": 0.7931, "step": 11335 }, { "epoch": 0.41057587830496195, "grad_norm": 1.3073597470131488, "learning_rate": 1.3318406069604794e-05, "loss": 0.7168, "step": 11336 }, { "epoch": 0.4106120970662803, "grad_norm": 1.3190194476710624, "learning_rate": 1.3317299450387216e-05, "loss": 0.7104, "step": 11337 }, { "epoch": 0.4106483158275987, "grad_norm": 1.405812186063676, "learning_rate": 1.331619278552068e-05, "loss": 0.7667, "step": 11338 }, { "epoch": 0.41068453458891707, "grad_norm": 1.3761998327762823, "learning_rate": 1.331508607502042e-05, "loss": 0.6777, "step": 11339 }, { "epoch": 0.41072075335023545, "grad_norm": 1.5528953102969236, "learning_rate": 1.331397931890166e-05, "loss": 0.762, "step": 11340 }, { "epoch": 0.41075697211155376, "grad_norm": 1.4563991866980506, "learning_rate": 1.3312872517179633e-05, "loss": 0.7921, "step": 11341 }, { "epoch": 0.41079319087287214, "grad_norm": 1.572928429329597, "learning_rate": 1.3311765669869572e-05, "loss": 0.8449, "step": 11342 }, { "epoch": 0.4108294096341905, "grad_norm": 1.3210176004359713, "learning_rate": 1.3310658776986702e-05, "loss": 0.7145, "step": 11343 }, { "epoch": 0.4108656283955089, "grad_norm": 1.2948934104575731, "learning_rate": 1.3309551838546261e-05, "loss": 0.7808, "step": 11344 }, { "epoch": 0.41090184715682726, "grad_norm": 1.2922521954529274, "learning_rate": 1.3308444854563477e-05, "loss": 0.7046, "step": 11345 }, { "epoch": 0.4109380659181456, "grad_norm": 1.4134543412525362, "learning_rate": 1.3307337825053586e-05, "loss": 0.7597, "step": 11346 }, { "epoch": 0.41097428467946395, "grad_norm": 1.222246753882821, "learning_rate": 1.3306230750031818e-05, "loss": 0.7813, "step": 11347 }, { "epoch": 0.4110105034407823, "grad_norm": 1.092357323926537, "learning_rate": 1.3305123629513413e-05, "loss": 0.7901, "step": 11348 }, { "epoch": 0.4110467222021007, "grad_norm": 1.302578689088541, "learning_rate": 1.33040164635136e-05, "loss": 0.7745, "step": 11349 }, { "epoch": 0.4110829409634191, "grad_norm": 1.3149122722923987, "learning_rate": 1.330290925204762e-05, "loss": 0.7584, "step": 11350 }, { "epoch": 0.4111191597247374, "grad_norm": 1.3472562334458336, "learning_rate": 1.3301801995130703e-05, "loss": 0.7642, "step": 11351 }, { "epoch": 0.41115537848605577, "grad_norm": 1.3221274223396513, "learning_rate": 1.3300694692778091e-05, "loss": 0.7103, "step": 11352 }, { "epoch": 0.41119159724737414, "grad_norm": 1.4358920156043198, "learning_rate": 1.329958734500502e-05, "loss": 0.81, "step": 11353 }, { "epoch": 0.4112278160086925, "grad_norm": 1.1572637858822354, "learning_rate": 1.3298479951826727e-05, "loss": 0.7539, "step": 11354 }, { "epoch": 0.4112640347700109, "grad_norm": 1.4471040273840554, "learning_rate": 1.3297372513258455e-05, "loss": 0.8699, "step": 11355 }, { "epoch": 0.4113002535313292, "grad_norm": 1.3679714080854437, "learning_rate": 1.3296265029315435e-05, "loss": 0.7372, "step": 11356 }, { "epoch": 0.4113364722926476, "grad_norm": 1.408469055980566, "learning_rate": 1.3295157500012917e-05, "loss": 0.8148, "step": 11357 }, { "epoch": 0.41137269105396596, "grad_norm": 1.5277281372598424, "learning_rate": 1.3294049925366133e-05, "loss": 0.7703, "step": 11358 }, { "epoch": 0.41140890981528433, "grad_norm": 1.5813741508473615, "learning_rate": 1.3292942305390329e-05, "loss": 0.7897, "step": 11359 }, { "epoch": 0.4114451285766027, "grad_norm": 1.443109649456371, "learning_rate": 1.3291834640100745e-05, "loss": 0.7427, "step": 11360 }, { "epoch": 0.411481347337921, "grad_norm": 1.2346626313303941, "learning_rate": 1.329072692951262e-05, "loss": 0.7601, "step": 11361 }, { "epoch": 0.4115175660992394, "grad_norm": 1.6034168272213598, "learning_rate": 1.3289619173641208e-05, "loss": 0.7949, "step": 11362 }, { "epoch": 0.41155378486055777, "grad_norm": 1.4648143333024488, "learning_rate": 1.3288511372501741e-05, "loss": 0.8277, "step": 11363 }, { "epoch": 0.41159000362187614, "grad_norm": 1.4022050220439561, "learning_rate": 1.3287403526109471e-05, "loss": 0.7227, "step": 11364 }, { "epoch": 0.4116262223831945, "grad_norm": 1.306676585439605, "learning_rate": 1.3286295634479636e-05, "loss": 0.7787, "step": 11365 }, { "epoch": 0.41166244114451284, "grad_norm": 1.390416147053865, "learning_rate": 1.3285187697627488e-05, "loss": 0.8107, "step": 11366 }, { "epoch": 0.4116986599058312, "grad_norm": 1.2567542915827288, "learning_rate": 1.328407971556827e-05, "loss": 0.707, "step": 11367 }, { "epoch": 0.4117348786671496, "grad_norm": 1.329853711122863, "learning_rate": 1.328297168831723e-05, "loss": 0.8001, "step": 11368 }, { "epoch": 0.41177109742846796, "grad_norm": 1.3796805978053717, "learning_rate": 1.3281863615889614e-05, "loss": 0.7856, "step": 11369 }, { "epoch": 0.41180731618978633, "grad_norm": 1.4118707476341208, "learning_rate": 1.3280755498300676e-05, "loss": 0.7263, "step": 11370 }, { "epoch": 0.41184353495110465, "grad_norm": 1.3473274043975259, "learning_rate": 1.327964733556565e-05, "loss": 0.8701, "step": 11371 }, { "epoch": 0.411879753712423, "grad_norm": 1.4198774995914065, "learning_rate": 1.3278539127699802e-05, "loss": 0.746, "step": 11372 }, { "epoch": 0.4119159724737414, "grad_norm": 1.027059136187858, "learning_rate": 1.3277430874718373e-05, "loss": 0.7302, "step": 11373 }, { "epoch": 0.4119521912350598, "grad_norm": 1.0863207135132855, "learning_rate": 1.3276322576636616e-05, "loss": 0.7439, "step": 11374 }, { "epoch": 0.41198840999637815, "grad_norm": 1.438349035117222, "learning_rate": 1.327521423346978e-05, "loss": 0.8211, "step": 11375 }, { "epoch": 0.41202462875769646, "grad_norm": 1.3413260562566827, "learning_rate": 1.3274105845233118e-05, "loss": 0.823, "step": 11376 }, { "epoch": 0.41206084751901484, "grad_norm": 1.4622822605561883, "learning_rate": 1.3272997411941885e-05, "loss": 0.7371, "step": 11377 }, { "epoch": 0.4120970662803332, "grad_norm": 1.1623113089612276, "learning_rate": 1.327188893361133e-05, "loss": 0.7312, "step": 11378 }, { "epoch": 0.4121332850416516, "grad_norm": 1.3488594571873314, "learning_rate": 1.327078041025671e-05, "loss": 0.6847, "step": 11379 }, { "epoch": 0.41216950380296996, "grad_norm": 1.2330399103154688, "learning_rate": 1.3269671841893275e-05, "loss": 0.7313, "step": 11380 }, { "epoch": 0.4122057225642883, "grad_norm": 1.4044651426751897, "learning_rate": 1.3268563228536287e-05, "loss": 0.7725, "step": 11381 }, { "epoch": 0.41224194132560665, "grad_norm": 1.3788952124588851, "learning_rate": 1.326745457020099e-05, "loss": 0.7999, "step": 11382 }, { "epoch": 0.412278160086925, "grad_norm": 1.469874074890124, "learning_rate": 1.3266345866902653e-05, "loss": 0.7936, "step": 11383 }, { "epoch": 0.4123143788482434, "grad_norm": 1.3248973249968934, "learning_rate": 1.3265237118656525e-05, "loss": 0.7072, "step": 11384 }, { "epoch": 0.4123505976095618, "grad_norm": 1.2427676198782118, "learning_rate": 1.3264128325477864e-05, "loss": 0.7818, "step": 11385 }, { "epoch": 0.4123868163708801, "grad_norm": 1.3797401278860246, "learning_rate": 1.326301948738193e-05, "loss": 0.8282, "step": 11386 }, { "epoch": 0.41242303513219847, "grad_norm": 1.2971007426396899, "learning_rate": 1.3261910604383981e-05, "loss": 0.7607, "step": 11387 }, { "epoch": 0.41245925389351684, "grad_norm": 1.329054055916991, "learning_rate": 1.3260801676499275e-05, "loss": 0.8092, "step": 11388 }, { "epoch": 0.4124954726548352, "grad_norm": 1.1496507112182388, "learning_rate": 1.3259692703743073e-05, "loss": 0.7261, "step": 11389 }, { "epoch": 0.4125316914161536, "grad_norm": 1.202252874633367, "learning_rate": 1.3258583686130637e-05, "loss": 0.7368, "step": 11390 }, { "epoch": 0.4125679101774719, "grad_norm": 1.1413174608956254, "learning_rate": 1.3257474623677224e-05, "loss": 0.7377, "step": 11391 }, { "epoch": 0.4126041289387903, "grad_norm": 1.2152473173813367, "learning_rate": 1.3256365516398098e-05, "loss": 0.7564, "step": 11392 }, { "epoch": 0.41264034770010866, "grad_norm": 1.3480073077680017, "learning_rate": 1.3255256364308525e-05, "loss": 0.7605, "step": 11393 }, { "epoch": 0.41267656646142703, "grad_norm": 1.3981357913641026, "learning_rate": 1.3254147167423759e-05, "loss": 0.6707, "step": 11394 }, { "epoch": 0.4127127852227454, "grad_norm": 1.392513856835994, "learning_rate": 1.325303792575907e-05, "loss": 0.7704, "step": 11395 }, { "epoch": 0.4127490039840637, "grad_norm": 1.4021104934750839, "learning_rate": 1.325192863932972e-05, "loss": 0.7808, "step": 11396 }, { "epoch": 0.4127852227453821, "grad_norm": 1.4131006201855862, "learning_rate": 1.325081930815098e-05, "loss": 0.8238, "step": 11397 }, { "epoch": 0.41282144150670047, "grad_norm": 1.2884247661736663, "learning_rate": 1.3249709932238105e-05, "loss": 0.7608, "step": 11398 }, { "epoch": 0.41285766026801884, "grad_norm": 1.555461351893414, "learning_rate": 1.3248600511606366e-05, "loss": 0.8022, "step": 11399 }, { "epoch": 0.4128938790293372, "grad_norm": 1.3123268016002825, "learning_rate": 1.324749104627103e-05, "loss": 0.7973, "step": 11400 }, { "epoch": 0.41293009779065554, "grad_norm": 1.368299664047519, "learning_rate": 1.3246381536247364e-05, "loss": 0.7587, "step": 11401 }, { "epoch": 0.4129663165519739, "grad_norm": 1.306133426279706, "learning_rate": 1.3245271981550637e-05, "loss": 0.8318, "step": 11402 }, { "epoch": 0.4130025353132923, "grad_norm": 1.3232715279680003, "learning_rate": 1.3244162382196111e-05, "loss": 0.7527, "step": 11403 }, { "epoch": 0.41303875407461066, "grad_norm": 1.5764762671353978, "learning_rate": 1.3243052738199067e-05, "loss": 0.7607, "step": 11404 }, { "epoch": 0.41307497283592903, "grad_norm": 1.3130477112995353, "learning_rate": 1.3241943049574762e-05, "loss": 0.719, "step": 11405 }, { "epoch": 0.41311119159724735, "grad_norm": 1.3786632807371832, "learning_rate": 1.3240833316338475e-05, "loss": 0.7916, "step": 11406 }, { "epoch": 0.4131474103585657, "grad_norm": 1.4666139714661863, "learning_rate": 1.3239723538505473e-05, "loss": 0.7681, "step": 11407 }, { "epoch": 0.4131836291198841, "grad_norm": 1.2365600163567632, "learning_rate": 1.3238613716091028e-05, "loss": 0.7123, "step": 11408 }, { "epoch": 0.4132198478812025, "grad_norm": 1.3547411327514034, "learning_rate": 1.3237503849110414e-05, "loss": 0.7804, "step": 11409 }, { "epoch": 0.41325606664252085, "grad_norm": 1.3331896703283153, "learning_rate": 1.3236393937578902e-05, "loss": 0.7148, "step": 11410 }, { "epoch": 0.41329228540383917, "grad_norm": 1.4662046684658576, "learning_rate": 1.3235283981511764e-05, "loss": 0.8562, "step": 11411 }, { "epoch": 0.41332850416515754, "grad_norm": 1.4436002731609396, "learning_rate": 1.323417398092428e-05, "loss": 0.7282, "step": 11412 }, { "epoch": 0.4133647229264759, "grad_norm": 1.6411039357258346, "learning_rate": 1.3233063935831717e-05, "loss": 0.7939, "step": 11413 }, { "epoch": 0.4134009416877943, "grad_norm": 1.115384020949296, "learning_rate": 1.323195384624935e-05, "loss": 0.8185, "step": 11414 }, { "epoch": 0.41343716044911266, "grad_norm": 1.4509276287454962, "learning_rate": 1.3230843712192463e-05, "loss": 0.821, "step": 11415 }, { "epoch": 0.413473379210431, "grad_norm": 1.2491899961647832, "learning_rate": 1.3229733533676325e-05, "loss": 0.6732, "step": 11416 }, { "epoch": 0.41350959797174935, "grad_norm": 1.4305931470825914, "learning_rate": 1.322862331071622e-05, "loss": 0.8636, "step": 11417 }, { "epoch": 0.41354581673306773, "grad_norm": 1.3542566954200925, "learning_rate": 1.3227513043327416e-05, "loss": 0.7762, "step": 11418 }, { "epoch": 0.4135820354943861, "grad_norm": 1.3570174925331333, "learning_rate": 1.32264027315252e-05, "loss": 0.803, "step": 11419 }, { "epoch": 0.4136182542557045, "grad_norm": 1.534035494394361, "learning_rate": 1.3225292375324848e-05, "loss": 0.7002, "step": 11420 }, { "epoch": 0.4136544730170228, "grad_norm": 1.6071629182380929, "learning_rate": 1.3224181974741637e-05, "loss": 0.7525, "step": 11421 }, { "epoch": 0.41369069177834117, "grad_norm": 1.3484793947978602, "learning_rate": 1.3223071529790848e-05, "loss": 0.724, "step": 11422 }, { "epoch": 0.41372691053965954, "grad_norm": 1.3515183945995068, "learning_rate": 1.3221961040487763e-05, "loss": 0.8625, "step": 11423 }, { "epoch": 0.4137631293009779, "grad_norm": 1.4246850550158963, "learning_rate": 1.3220850506847667e-05, "loss": 0.7441, "step": 11424 }, { "epoch": 0.4137993480622963, "grad_norm": 1.4965679821456448, "learning_rate": 1.3219739928885834e-05, "loss": 0.7679, "step": 11425 }, { "epoch": 0.4138355668236146, "grad_norm": 1.4117903513924166, "learning_rate": 1.3218629306617552e-05, "loss": 0.7691, "step": 11426 }, { "epoch": 0.413871785584933, "grad_norm": 1.5139026404412672, "learning_rate": 1.3217518640058103e-05, "loss": 0.7464, "step": 11427 }, { "epoch": 0.41390800434625136, "grad_norm": 1.5611991426294989, "learning_rate": 1.321640792922277e-05, "loss": 0.7841, "step": 11428 }, { "epoch": 0.41394422310756973, "grad_norm": 1.2834652962378676, "learning_rate": 1.3215297174126835e-05, "loss": 0.7981, "step": 11429 }, { "epoch": 0.4139804418688881, "grad_norm": 1.2769239009805025, "learning_rate": 1.3214186374785589e-05, "loss": 0.8067, "step": 11430 }, { "epoch": 0.4140166606302064, "grad_norm": 1.1789759132264734, "learning_rate": 1.3213075531214315e-05, "loss": 0.7787, "step": 11431 }, { "epoch": 0.4140528793915248, "grad_norm": 1.4843267959999493, "learning_rate": 1.3211964643428295e-05, "loss": 0.8073, "step": 11432 }, { "epoch": 0.41408909815284317, "grad_norm": 1.3676237881578996, "learning_rate": 1.321085371144282e-05, "loss": 0.7263, "step": 11433 }, { "epoch": 0.41412531691416155, "grad_norm": 1.352801068264938, "learning_rate": 1.3209742735273179e-05, "loss": 0.7281, "step": 11434 }, { "epoch": 0.4141615356754799, "grad_norm": 1.4573496201933165, "learning_rate": 1.3208631714934655e-05, "loss": 0.7644, "step": 11435 }, { "epoch": 0.41419775443679824, "grad_norm": 1.14501049835286, "learning_rate": 1.320752065044254e-05, "loss": 0.7427, "step": 11436 }, { "epoch": 0.4142339731981166, "grad_norm": 1.415370696296464, "learning_rate": 1.3206409541812123e-05, "loss": 0.7226, "step": 11437 }, { "epoch": 0.414270191959435, "grad_norm": 1.0636614236211381, "learning_rate": 1.3205298389058694e-05, "loss": 0.7202, "step": 11438 }, { "epoch": 0.41430641072075336, "grad_norm": 1.4462206333001169, "learning_rate": 1.320418719219754e-05, "loss": 0.7103, "step": 11439 }, { "epoch": 0.41434262948207173, "grad_norm": 1.305823763222842, "learning_rate": 1.3203075951243957e-05, "loss": 0.7123, "step": 11440 }, { "epoch": 0.41437884824339005, "grad_norm": 1.187554040576385, "learning_rate": 1.3201964666213234e-05, "loss": 0.787, "step": 11441 }, { "epoch": 0.4144150670047084, "grad_norm": 1.6142453610308671, "learning_rate": 1.3200853337120665e-05, "loss": 0.8549, "step": 11442 }, { "epoch": 0.4144512857660268, "grad_norm": 1.370766882637961, "learning_rate": 1.3199741963981538e-05, "loss": 0.8638, "step": 11443 }, { "epoch": 0.4144875045273452, "grad_norm": 1.2962396419738045, "learning_rate": 1.3198630546811155e-05, "loss": 0.7853, "step": 11444 }, { "epoch": 0.41452372328866355, "grad_norm": 1.2619914555134968, "learning_rate": 1.3197519085624801e-05, "loss": 0.7339, "step": 11445 }, { "epoch": 0.41455994204998187, "grad_norm": 1.626036304372987, "learning_rate": 1.3196407580437777e-05, "loss": 0.7459, "step": 11446 }, { "epoch": 0.41459616081130024, "grad_norm": 1.487860880579723, "learning_rate": 1.3195296031265375e-05, "loss": 0.7726, "step": 11447 }, { "epoch": 0.4146323795726186, "grad_norm": 1.3759675726180631, "learning_rate": 1.3194184438122894e-05, "loss": 0.8341, "step": 11448 }, { "epoch": 0.414668598333937, "grad_norm": 1.2719105381359002, "learning_rate": 1.3193072801025625e-05, "loss": 0.7827, "step": 11449 }, { "epoch": 0.41470481709525536, "grad_norm": 1.3324312853222249, "learning_rate": 1.319196111998887e-05, "loss": 0.781, "step": 11450 }, { "epoch": 0.4147410358565737, "grad_norm": 1.3587586112569843, "learning_rate": 1.3190849395027926e-05, "loss": 0.7597, "step": 11451 }, { "epoch": 0.41477725461789205, "grad_norm": 1.3355486252241333, "learning_rate": 1.318973762615809e-05, "loss": 0.7986, "step": 11452 }, { "epoch": 0.41481347337921043, "grad_norm": 1.3880150793670807, "learning_rate": 1.3188625813394663e-05, "loss": 0.7175, "step": 11453 }, { "epoch": 0.4148496921405288, "grad_norm": 1.066809975832782, "learning_rate": 1.318751395675294e-05, "loss": 0.756, "step": 11454 }, { "epoch": 0.4148859109018472, "grad_norm": 1.41139684008854, "learning_rate": 1.318640205624823e-05, "loss": 0.7866, "step": 11455 }, { "epoch": 0.4149221296631655, "grad_norm": 1.3735458639400828, "learning_rate": 1.318529011189582e-05, "loss": 0.7196, "step": 11456 }, { "epoch": 0.41495834842448387, "grad_norm": 1.3630911843520968, "learning_rate": 1.3184178123711025e-05, "loss": 0.7395, "step": 11457 }, { "epoch": 0.41499456718580224, "grad_norm": 1.2429080094122709, "learning_rate": 1.3183066091709138e-05, "loss": 0.7519, "step": 11458 }, { "epoch": 0.4150307859471206, "grad_norm": 1.357194243356587, "learning_rate": 1.3181954015905467e-05, "loss": 0.8618, "step": 11459 }, { "epoch": 0.415067004708439, "grad_norm": 1.133477676187252, "learning_rate": 1.3180841896315311e-05, "loss": 0.7998, "step": 11460 }, { "epoch": 0.4151032234697573, "grad_norm": 1.0592131407671752, "learning_rate": 1.3179729732953973e-05, "loss": 0.7603, "step": 11461 }, { "epoch": 0.4151394422310757, "grad_norm": 1.3468157980933955, "learning_rate": 1.3178617525836765e-05, "loss": 0.754, "step": 11462 }, { "epoch": 0.41517566099239406, "grad_norm": 1.2730730330901252, "learning_rate": 1.3177505274978986e-05, "loss": 0.6691, "step": 11463 }, { "epoch": 0.41521187975371243, "grad_norm": 1.3661175151238227, "learning_rate": 1.3176392980395941e-05, "loss": 0.8429, "step": 11464 }, { "epoch": 0.4152480985150308, "grad_norm": 1.3705666600771373, "learning_rate": 1.3175280642102939e-05, "loss": 0.7727, "step": 11465 }, { "epoch": 0.4152843172763491, "grad_norm": 1.4226568778747313, "learning_rate": 1.3174168260115286e-05, "loss": 0.8434, "step": 11466 }, { "epoch": 0.4153205360376675, "grad_norm": 1.229983812473745, "learning_rate": 1.3173055834448285e-05, "loss": 0.7897, "step": 11467 }, { "epoch": 0.41535675479898587, "grad_norm": 1.1293442004097025, "learning_rate": 1.317194336511725e-05, "loss": 0.7953, "step": 11468 }, { "epoch": 0.41539297356030425, "grad_norm": 1.3107240233102766, "learning_rate": 1.3170830852137488e-05, "loss": 0.6559, "step": 11469 }, { "epoch": 0.4154291923216226, "grad_norm": 1.2901998754642001, "learning_rate": 1.3169718295524306e-05, "loss": 0.7111, "step": 11470 }, { "epoch": 0.41546541108294094, "grad_norm": 1.3711002056263288, "learning_rate": 1.3168605695293015e-05, "loss": 0.7716, "step": 11471 }, { "epoch": 0.4155016298442593, "grad_norm": 1.13158635988956, "learning_rate": 1.3167493051458923e-05, "loss": 0.7139, "step": 11472 }, { "epoch": 0.4155378486055777, "grad_norm": 1.694910990725576, "learning_rate": 1.316638036403735e-05, "loss": 0.8205, "step": 11473 }, { "epoch": 0.41557406736689606, "grad_norm": 1.3397423772581687, "learning_rate": 1.3165267633043594e-05, "loss": 0.781, "step": 11474 }, { "epoch": 0.41561028612821443, "grad_norm": 1.3076342303110555, "learning_rate": 1.3164154858492979e-05, "loss": 0.8282, "step": 11475 }, { "epoch": 0.41564650488953275, "grad_norm": 1.2783211812977355, "learning_rate": 1.3163042040400811e-05, "loss": 0.7276, "step": 11476 }, { "epoch": 0.4156827236508511, "grad_norm": 1.3955927182977064, "learning_rate": 1.3161929178782407e-05, "loss": 0.6885, "step": 11477 }, { "epoch": 0.4157189424121695, "grad_norm": 1.2981045613965059, "learning_rate": 1.3160816273653077e-05, "loss": 0.7822, "step": 11478 }, { "epoch": 0.4157551611734879, "grad_norm": 1.3983126456717088, "learning_rate": 1.315970332502814e-05, "loss": 0.7251, "step": 11479 }, { "epoch": 0.41579137993480625, "grad_norm": 1.3890181289501036, "learning_rate": 1.3158590332922904e-05, "loss": 0.8043, "step": 11480 }, { "epoch": 0.41582759869612457, "grad_norm": 1.2843063038242544, "learning_rate": 1.3157477297352693e-05, "loss": 0.771, "step": 11481 }, { "epoch": 0.41586381745744294, "grad_norm": 1.0626411503103732, "learning_rate": 1.315636421833282e-05, "loss": 0.7295, "step": 11482 }, { "epoch": 0.4159000362187613, "grad_norm": 1.332571242179323, "learning_rate": 1.31552510958786e-05, "loss": 0.7801, "step": 11483 }, { "epoch": 0.4159362549800797, "grad_norm": 1.2998513171102626, "learning_rate": 1.3154137930005357e-05, "loss": 0.6961, "step": 11484 }, { "epoch": 0.41597247374139806, "grad_norm": 1.822686209189207, "learning_rate": 1.31530247207284e-05, "loss": 0.7504, "step": 11485 }, { "epoch": 0.4160086925027164, "grad_norm": 1.3816573752158237, "learning_rate": 1.3151911468063056e-05, "loss": 0.769, "step": 11486 }, { "epoch": 0.41604491126403476, "grad_norm": 1.3178336670667956, "learning_rate": 1.3150798172024639e-05, "loss": 0.7967, "step": 11487 }, { "epoch": 0.41608113002535313, "grad_norm": 1.2617175970278924, "learning_rate": 1.3149684832628471e-05, "loss": 0.7405, "step": 11488 }, { "epoch": 0.4161173487866715, "grad_norm": 1.0348152812548914, "learning_rate": 1.3148571449889872e-05, "loss": 0.6579, "step": 11489 }, { "epoch": 0.4161535675479899, "grad_norm": 1.259473815932076, "learning_rate": 1.3147458023824163e-05, "loss": 0.7676, "step": 11490 }, { "epoch": 0.4161897863093082, "grad_norm": 1.422592849093799, "learning_rate": 1.3146344554446664e-05, "loss": 0.749, "step": 11491 }, { "epoch": 0.41622600507062657, "grad_norm": 0.9954594264836639, "learning_rate": 1.3145231041772701e-05, "loss": 0.7226, "step": 11492 }, { "epoch": 0.41626222383194494, "grad_norm": 1.4497425894714355, "learning_rate": 1.3144117485817595e-05, "loss": 0.8039, "step": 11493 }, { "epoch": 0.4162984425932633, "grad_norm": 1.3608583367620404, "learning_rate": 1.314300388659667e-05, "loss": 0.7519, "step": 11494 }, { "epoch": 0.4163346613545817, "grad_norm": 1.351878956380783, "learning_rate": 1.3141890244125253e-05, "loss": 0.7758, "step": 11495 }, { "epoch": 0.4163708801159, "grad_norm": 1.2257873248113254, "learning_rate": 1.314077655841866e-05, "loss": 0.7509, "step": 11496 }, { "epoch": 0.4164070988772184, "grad_norm": 1.0430472504091466, "learning_rate": 1.3139662829492226e-05, "loss": 0.7746, "step": 11497 }, { "epoch": 0.41644331763853676, "grad_norm": 1.1471535580414238, "learning_rate": 1.313854905736127e-05, "loss": 0.7903, "step": 11498 }, { "epoch": 0.41647953639985513, "grad_norm": 1.443281404856692, "learning_rate": 1.3137435242041123e-05, "loss": 0.7456, "step": 11499 }, { "epoch": 0.4165157551611735, "grad_norm": 1.3754526254882502, "learning_rate": 1.3136321383547109e-05, "loss": 0.7996, "step": 11500 }, { "epoch": 0.4165519739224918, "grad_norm": 1.502895556849949, "learning_rate": 1.3135207481894557e-05, "loss": 0.8184, "step": 11501 }, { "epoch": 0.4165881926838102, "grad_norm": 1.0882548984204394, "learning_rate": 1.3134093537098797e-05, "loss": 0.7491, "step": 11502 }, { "epoch": 0.4166244114451286, "grad_norm": 1.3315177379325802, "learning_rate": 1.3132979549175153e-05, "loss": 0.7238, "step": 11503 }, { "epoch": 0.41666063020644695, "grad_norm": 1.3877825028412272, "learning_rate": 1.313186551813896e-05, "loss": 0.8148, "step": 11504 }, { "epoch": 0.4166968489677653, "grad_norm": 1.3722680570438073, "learning_rate": 1.3130751444005546e-05, "loss": 0.7472, "step": 11505 }, { "epoch": 0.41673306772908364, "grad_norm": 1.4025858408462248, "learning_rate": 1.312963732679024e-05, "loss": 0.7406, "step": 11506 }, { "epoch": 0.416769286490402, "grad_norm": 1.4300436295509327, "learning_rate": 1.3128523166508374e-05, "loss": 0.8022, "step": 11507 }, { "epoch": 0.4168055052517204, "grad_norm": 1.3695954398624095, "learning_rate": 1.312740896317528e-05, "loss": 0.8386, "step": 11508 }, { "epoch": 0.41684172401303876, "grad_norm": 1.1254116557717329, "learning_rate": 1.3126294716806292e-05, "loss": 0.7776, "step": 11509 }, { "epoch": 0.41687794277435714, "grad_norm": 1.3010260062449799, "learning_rate": 1.3125180427416741e-05, "loss": 0.7253, "step": 11510 }, { "epoch": 0.41691416153567545, "grad_norm": 1.3368136630872072, "learning_rate": 1.312406609502196e-05, "loss": 0.7011, "step": 11511 }, { "epoch": 0.41695038029699383, "grad_norm": 1.5099254094262728, "learning_rate": 1.3122951719637287e-05, "loss": 0.791, "step": 11512 }, { "epoch": 0.4169865990583122, "grad_norm": 1.4428189611940918, "learning_rate": 1.3121837301278055e-05, "loss": 0.8262, "step": 11513 }, { "epoch": 0.4170228178196306, "grad_norm": 1.374330028227076, "learning_rate": 1.3120722839959595e-05, "loss": 0.7518, "step": 11514 }, { "epoch": 0.41705903658094895, "grad_norm": 1.4938466139659647, "learning_rate": 1.3119608335697249e-05, "loss": 0.7723, "step": 11515 }, { "epoch": 0.41709525534226727, "grad_norm": 1.2789355619540106, "learning_rate": 1.3118493788506351e-05, "loss": 0.7666, "step": 11516 }, { "epoch": 0.41713147410358564, "grad_norm": 1.088446829679973, "learning_rate": 1.311737919840224e-05, "loss": 0.7164, "step": 11517 }, { "epoch": 0.417167692864904, "grad_norm": 1.367468676944958, "learning_rate": 1.3116264565400247e-05, "loss": 0.754, "step": 11518 }, { "epoch": 0.4172039116262224, "grad_norm": 1.6580384535513386, "learning_rate": 1.3115149889515722e-05, "loss": 0.7645, "step": 11519 }, { "epoch": 0.41724013038754076, "grad_norm": 1.3486344090921176, "learning_rate": 1.311403517076399e-05, "loss": 0.7192, "step": 11520 }, { "epoch": 0.4172763491488591, "grad_norm": 1.2991358951670842, "learning_rate": 1.3112920409160403e-05, "loss": 0.7223, "step": 11521 }, { "epoch": 0.41731256791017746, "grad_norm": 1.2545292272672002, "learning_rate": 1.3111805604720296e-05, "loss": 0.7445, "step": 11522 }, { "epoch": 0.41734878667149583, "grad_norm": 1.3943639556061798, "learning_rate": 1.3110690757459005e-05, "loss": 0.8122, "step": 11523 }, { "epoch": 0.4173850054328142, "grad_norm": 1.3381496480640689, "learning_rate": 1.3109575867391881e-05, "loss": 0.7601, "step": 11524 }, { "epoch": 0.4174212241941326, "grad_norm": 1.5260199702861161, "learning_rate": 1.3108460934534257e-05, "loss": 0.6802, "step": 11525 }, { "epoch": 0.4174574429554509, "grad_norm": 1.4876145599232244, "learning_rate": 1.3107345958901484e-05, "loss": 0.8901, "step": 11526 }, { "epoch": 0.41749366171676927, "grad_norm": 1.4147305958617138, "learning_rate": 1.3106230940508895e-05, "loss": 0.7622, "step": 11527 }, { "epoch": 0.41752988047808764, "grad_norm": 1.4163169776948743, "learning_rate": 1.3105115879371844e-05, "loss": 0.7401, "step": 11528 }, { "epoch": 0.417566099239406, "grad_norm": 1.248778029177368, "learning_rate": 1.3104000775505665e-05, "loss": 0.8031, "step": 11529 }, { "epoch": 0.4176023180007244, "grad_norm": 1.3011491637536847, "learning_rate": 1.3102885628925712e-05, "loss": 0.753, "step": 11530 }, { "epoch": 0.4176385367620427, "grad_norm": 1.1323285232818263, "learning_rate": 1.3101770439647325e-05, "loss": 0.6902, "step": 11531 }, { "epoch": 0.4176747555233611, "grad_norm": 1.0488412577523476, "learning_rate": 1.310065520768585e-05, "loss": 0.7522, "step": 11532 }, { "epoch": 0.41771097428467946, "grad_norm": 1.1018178419621072, "learning_rate": 1.3099539933056638e-05, "loss": 0.7783, "step": 11533 }, { "epoch": 0.41774719304599783, "grad_norm": 1.0461226392980225, "learning_rate": 1.3098424615775027e-05, "loss": 0.7684, "step": 11534 }, { "epoch": 0.4177834118073162, "grad_norm": 1.4416479202753232, "learning_rate": 1.3097309255856377e-05, "loss": 0.7841, "step": 11535 }, { "epoch": 0.4178196305686345, "grad_norm": 1.3733374842941044, "learning_rate": 1.3096193853316027e-05, "loss": 0.7426, "step": 11536 }, { "epoch": 0.4178558493299529, "grad_norm": 1.339770732146738, "learning_rate": 1.309507840816933e-05, "loss": 0.7128, "step": 11537 }, { "epoch": 0.4178920680912713, "grad_norm": 1.4663206615072086, "learning_rate": 1.3093962920431633e-05, "loss": 0.7037, "step": 11538 }, { "epoch": 0.41792828685258965, "grad_norm": 1.4586251696291392, "learning_rate": 1.3092847390118288e-05, "loss": 0.7194, "step": 11539 }, { "epoch": 0.417964505613908, "grad_norm": 1.08514869251668, "learning_rate": 1.3091731817244647e-05, "loss": 0.7707, "step": 11540 }, { "epoch": 0.41800072437522634, "grad_norm": 1.3644870129987563, "learning_rate": 1.3090616201826056e-05, "loss": 0.7209, "step": 11541 }, { "epoch": 0.4180369431365447, "grad_norm": 1.3238866419112807, "learning_rate": 1.3089500543877876e-05, "loss": 0.7256, "step": 11542 }, { "epoch": 0.4180731618978631, "grad_norm": 1.3123599329642246, "learning_rate": 1.3088384843415449e-05, "loss": 0.7836, "step": 11543 }, { "epoch": 0.41810938065918146, "grad_norm": 1.272770530002632, "learning_rate": 1.3087269100454136e-05, "loss": 0.7848, "step": 11544 }, { "epoch": 0.41814559942049984, "grad_norm": 1.4512362542895803, "learning_rate": 1.3086153315009285e-05, "loss": 0.7937, "step": 11545 }, { "epoch": 0.41818181818181815, "grad_norm": 1.3710582176994914, "learning_rate": 1.3085037487096255e-05, "loss": 0.8197, "step": 11546 }, { "epoch": 0.41821803694313653, "grad_norm": 1.5290162826595615, "learning_rate": 1.3083921616730396e-05, "loss": 0.7987, "step": 11547 }, { "epoch": 0.4182542557044549, "grad_norm": 1.3199249445068668, "learning_rate": 1.3082805703927066e-05, "loss": 0.7677, "step": 11548 }, { "epoch": 0.4182904744657733, "grad_norm": 1.3116222502011956, "learning_rate": 1.3081689748701623e-05, "loss": 0.7857, "step": 11549 }, { "epoch": 0.41832669322709165, "grad_norm": 1.341090024478275, "learning_rate": 1.308057375106942e-05, "loss": 0.7394, "step": 11550 }, { "epoch": 0.41836291198840997, "grad_norm": 1.367104476042721, "learning_rate": 1.3079457711045815e-05, "loss": 0.7704, "step": 11551 }, { "epoch": 0.41839913074972834, "grad_norm": 1.368263801217545, "learning_rate": 1.3078341628646168e-05, "loss": 0.7827, "step": 11552 }, { "epoch": 0.4184353495110467, "grad_norm": 1.450816402949227, "learning_rate": 1.3077225503885834e-05, "loss": 0.7946, "step": 11553 }, { "epoch": 0.4184715682723651, "grad_norm": 1.3704648259991465, "learning_rate": 1.3076109336780172e-05, "loss": 0.7974, "step": 11554 }, { "epoch": 0.41850778703368346, "grad_norm": 1.2024127296269496, "learning_rate": 1.3074993127344545e-05, "loss": 0.7336, "step": 11555 }, { "epoch": 0.4185440057950018, "grad_norm": 1.4705343409240557, "learning_rate": 1.3073876875594307e-05, "loss": 0.7623, "step": 11556 }, { "epoch": 0.41858022455632016, "grad_norm": 1.0546595824705838, "learning_rate": 1.3072760581544826e-05, "loss": 0.7975, "step": 11557 }, { "epoch": 0.41861644331763853, "grad_norm": 1.281428826168916, "learning_rate": 1.3071644245211456e-05, "loss": 0.751, "step": 11558 }, { "epoch": 0.4186526620789569, "grad_norm": 1.3320360452811773, "learning_rate": 1.3070527866609564e-05, "loss": 0.7501, "step": 11559 }, { "epoch": 0.4186888808402753, "grad_norm": 1.3459287351885751, "learning_rate": 1.306941144575451e-05, "loss": 0.7055, "step": 11560 }, { "epoch": 0.4187250996015936, "grad_norm": 1.265043339141218, "learning_rate": 1.3068294982661659e-05, "loss": 0.6563, "step": 11561 }, { "epoch": 0.41876131836291197, "grad_norm": 1.4418165132192315, "learning_rate": 1.3067178477346372e-05, "loss": 0.8111, "step": 11562 }, { "epoch": 0.41879753712423035, "grad_norm": 1.3587284318064032, "learning_rate": 1.3066061929824013e-05, "loss": 0.7863, "step": 11563 }, { "epoch": 0.4188337558855487, "grad_norm": 1.3564503019445726, "learning_rate": 1.306494534010995e-05, "loss": 0.8377, "step": 11564 }, { "epoch": 0.4188699746468671, "grad_norm": 1.6149120493537306, "learning_rate": 1.3063828708219543e-05, "loss": 0.8738, "step": 11565 }, { "epoch": 0.41890619340818547, "grad_norm": 1.0709598396109499, "learning_rate": 1.3062712034168166e-05, "loss": 0.775, "step": 11566 }, { "epoch": 0.4189424121695038, "grad_norm": 1.324808547801541, "learning_rate": 1.3061595317971176e-05, "loss": 0.7811, "step": 11567 }, { "epoch": 0.41897863093082216, "grad_norm": 1.268894159678419, "learning_rate": 1.3060478559643946e-05, "loss": 0.7904, "step": 11568 }, { "epoch": 0.41901484969214053, "grad_norm": 1.1780510630034475, "learning_rate": 1.305936175920184e-05, "loss": 0.6697, "step": 11569 }, { "epoch": 0.4190510684534589, "grad_norm": 1.3533879948588963, "learning_rate": 1.3058244916660228e-05, "loss": 0.7847, "step": 11570 }, { "epoch": 0.4190872872147773, "grad_norm": 1.0632324718032984, "learning_rate": 1.305712803203448e-05, "loss": 0.7759, "step": 11571 }, { "epoch": 0.4191235059760956, "grad_norm": 1.142110232790552, "learning_rate": 1.3056011105339965e-05, "loss": 0.7146, "step": 11572 }, { "epoch": 0.419159724737414, "grad_norm": 1.2853923474935973, "learning_rate": 1.3054894136592052e-05, "loss": 0.7467, "step": 11573 }, { "epoch": 0.41919594349873235, "grad_norm": 1.069992623329629, "learning_rate": 1.3053777125806108e-05, "loss": 0.7582, "step": 11574 }, { "epoch": 0.4192321622600507, "grad_norm": 1.0342990686567834, "learning_rate": 1.3052660072997511e-05, "loss": 0.7607, "step": 11575 }, { "epoch": 0.4192683810213691, "grad_norm": 1.2618163450483313, "learning_rate": 1.3051542978181629e-05, "loss": 0.7207, "step": 11576 }, { "epoch": 0.4193045997826874, "grad_norm": 1.4268173764097993, "learning_rate": 1.3050425841373833e-05, "loss": 0.8387, "step": 11577 }, { "epoch": 0.4193408185440058, "grad_norm": 1.3056612117275423, "learning_rate": 1.3049308662589498e-05, "loss": 0.7672, "step": 11578 }, { "epoch": 0.41937703730532416, "grad_norm": 1.3940755115458496, "learning_rate": 1.3048191441843993e-05, "loss": 0.7899, "step": 11579 }, { "epoch": 0.41941325606664254, "grad_norm": 1.4319635237866124, "learning_rate": 1.30470741791527e-05, "loss": 0.7589, "step": 11580 }, { "epoch": 0.4194494748279609, "grad_norm": 1.3281277362009922, "learning_rate": 1.3045956874530986e-05, "loss": 0.7538, "step": 11581 }, { "epoch": 0.41948569358927923, "grad_norm": 1.0374264010301133, "learning_rate": 1.3044839527994231e-05, "loss": 0.7217, "step": 11582 }, { "epoch": 0.4195219123505976, "grad_norm": 1.2705709363281863, "learning_rate": 1.3043722139557806e-05, "loss": 0.6774, "step": 11583 }, { "epoch": 0.419558131111916, "grad_norm": 1.2942405428658332, "learning_rate": 1.3042604709237092e-05, "loss": 0.7131, "step": 11584 }, { "epoch": 0.41959434987323435, "grad_norm": 1.6081607043027313, "learning_rate": 1.3041487237047462e-05, "loss": 0.8134, "step": 11585 }, { "epoch": 0.4196305686345527, "grad_norm": 1.1829820152103792, "learning_rate": 1.30403697230043e-05, "loss": 0.7123, "step": 11586 }, { "epoch": 0.41966678739587104, "grad_norm": 1.0143713101671115, "learning_rate": 1.3039252167122973e-05, "loss": 0.7379, "step": 11587 }, { "epoch": 0.4197030061571894, "grad_norm": 1.418980733871724, "learning_rate": 1.303813456941887e-05, "loss": 0.7269, "step": 11588 }, { "epoch": 0.4197392249185078, "grad_norm": 1.1385601139757275, "learning_rate": 1.3037016929907362e-05, "loss": 0.736, "step": 11589 }, { "epoch": 0.41977544367982617, "grad_norm": 1.3037099688237053, "learning_rate": 1.3035899248603837e-05, "loss": 0.7793, "step": 11590 }, { "epoch": 0.41981166244114454, "grad_norm": 1.258808060377662, "learning_rate": 1.3034781525523668e-05, "loss": 0.7474, "step": 11591 }, { "epoch": 0.41984788120246286, "grad_norm": 1.3383112260197092, "learning_rate": 1.3033663760682239e-05, "loss": 0.7558, "step": 11592 }, { "epoch": 0.41988409996378123, "grad_norm": 1.4382269669317465, "learning_rate": 1.3032545954094931e-05, "loss": 0.7633, "step": 11593 }, { "epoch": 0.4199203187250996, "grad_norm": 1.3297782068266375, "learning_rate": 1.3031428105777124e-05, "loss": 0.7596, "step": 11594 }, { "epoch": 0.419956537486418, "grad_norm": 1.4218895995424132, "learning_rate": 1.3030310215744206e-05, "loss": 0.8354, "step": 11595 }, { "epoch": 0.41999275624773635, "grad_norm": 1.4459072088037106, "learning_rate": 1.3029192284011554e-05, "loss": 0.852, "step": 11596 }, { "epoch": 0.4200289750090547, "grad_norm": 1.111484872171251, "learning_rate": 1.3028074310594558e-05, "loss": 0.7029, "step": 11597 }, { "epoch": 0.42006519377037305, "grad_norm": 1.4215477839387662, "learning_rate": 1.3026956295508594e-05, "loss": 0.7697, "step": 11598 }, { "epoch": 0.4201014125316914, "grad_norm": 1.0648606960429114, "learning_rate": 1.3025838238769058e-05, "loss": 0.7981, "step": 11599 }, { "epoch": 0.4201376312930098, "grad_norm": 1.0551863455067036, "learning_rate": 1.3024720140391322e-05, "loss": 0.7537, "step": 11600 }, { "epoch": 0.42017385005432817, "grad_norm": 1.317435363219093, "learning_rate": 1.3023602000390783e-05, "loss": 0.8075, "step": 11601 }, { "epoch": 0.4202100688156465, "grad_norm": 1.4027311300071803, "learning_rate": 1.3022483818782823e-05, "loss": 0.8268, "step": 11602 }, { "epoch": 0.42024628757696486, "grad_norm": 1.0636440784233236, "learning_rate": 1.302136559558283e-05, "loss": 0.7545, "step": 11603 }, { "epoch": 0.42028250633828323, "grad_norm": 1.3167122773816806, "learning_rate": 1.3020247330806193e-05, "loss": 0.7034, "step": 11604 }, { "epoch": 0.4203187250996016, "grad_norm": 1.4563378085014047, "learning_rate": 1.3019129024468295e-05, "loss": 0.8218, "step": 11605 }, { "epoch": 0.42035494386092, "grad_norm": 1.3371996473541579, "learning_rate": 1.3018010676584533e-05, "loss": 0.754, "step": 11606 }, { "epoch": 0.4203911626222383, "grad_norm": 1.4754929024067711, "learning_rate": 1.301689228717029e-05, "loss": 0.7829, "step": 11607 }, { "epoch": 0.4204273813835567, "grad_norm": 1.6260156941788453, "learning_rate": 1.301577385624096e-05, "loss": 0.7758, "step": 11608 }, { "epoch": 0.42046360014487505, "grad_norm": 1.2880884207382732, "learning_rate": 1.3014655383811929e-05, "loss": 0.7714, "step": 11609 }, { "epoch": 0.4204998189061934, "grad_norm": 1.4657282301604706, "learning_rate": 1.3013536869898593e-05, "loss": 0.8599, "step": 11610 }, { "epoch": 0.4205360376675118, "grad_norm": 1.564025663347927, "learning_rate": 1.3012418314516342e-05, "loss": 0.7873, "step": 11611 }, { "epoch": 0.4205722564288301, "grad_norm": 1.1265489767368375, "learning_rate": 1.3011299717680568e-05, "loss": 0.7092, "step": 11612 }, { "epoch": 0.4206084751901485, "grad_norm": 1.3834240565929652, "learning_rate": 1.3010181079406663e-05, "loss": 0.7628, "step": 11613 }, { "epoch": 0.42064469395146686, "grad_norm": 1.3685880843941451, "learning_rate": 1.3009062399710021e-05, "loss": 0.8396, "step": 11614 }, { "epoch": 0.42068091271278524, "grad_norm": 1.342071870112926, "learning_rate": 1.3007943678606038e-05, "loss": 0.7198, "step": 11615 }, { "epoch": 0.4207171314741036, "grad_norm": 1.3690280078930617, "learning_rate": 1.3006824916110105e-05, "loss": 0.7976, "step": 11616 }, { "epoch": 0.42075335023542193, "grad_norm": 1.1484121745384164, "learning_rate": 1.3005706112237622e-05, "loss": 0.7064, "step": 11617 }, { "epoch": 0.4207895689967403, "grad_norm": 1.4664112401927165, "learning_rate": 1.3004587267003979e-05, "loss": 0.7032, "step": 11618 }, { "epoch": 0.4208257877580587, "grad_norm": 1.4542582294924917, "learning_rate": 1.3003468380424578e-05, "loss": 0.7455, "step": 11619 }, { "epoch": 0.42086200651937705, "grad_norm": 1.3537514950333651, "learning_rate": 1.300234945251481e-05, "loss": 0.7555, "step": 11620 }, { "epoch": 0.4208982252806954, "grad_norm": 1.258965917079737, "learning_rate": 1.3001230483290076e-05, "loss": 0.868, "step": 11621 }, { "epoch": 0.42093444404201374, "grad_norm": 1.7435840731127576, "learning_rate": 1.3000111472765776e-05, "loss": 0.8545, "step": 11622 }, { "epoch": 0.4209706628033321, "grad_norm": 1.5712195132684703, "learning_rate": 1.2998992420957301e-05, "loss": 0.8323, "step": 11623 }, { "epoch": 0.4210068815646505, "grad_norm": 1.3288286410285925, "learning_rate": 1.299787332788006e-05, "loss": 0.7999, "step": 11624 }, { "epoch": 0.42104310032596887, "grad_norm": 1.5326925751866085, "learning_rate": 1.2996754193549446e-05, "loss": 0.8061, "step": 11625 }, { "epoch": 0.42107931908728724, "grad_norm": 1.2598101019862722, "learning_rate": 1.2995635017980863e-05, "loss": 0.7752, "step": 11626 }, { "epoch": 0.42111553784860556, "grad_norm": 1.0342394685554723, "learning_rate": 1.2994515801189706e-05, "loss": 0.7617, "step": 11627 }, { "epoch": 0.42115175660992393, "grad_norm": 1.377963762517611, "learning_rate": 1.2993396543191383e-05, "loss": 0.7461, "step": 11628 }, { "epoch": 0.4211879753712423, "grad_norm": 1.3870076906032522, "learning_rate": 1.2992277244001294e-05, "loss": 0.7808, "step": 11629 }, { "epoch": 0.4212241941325607, "grad_norm": 1.3076049321222092, "learning_rate": 1.299115790363484e-05, "loss": 0.7358, "step": 11630 }, { "epoch": 0.42126041289387905, "grad_norm": 1.146517856988748, "learning_rate": 1.2990038522107424e-05, "loss": 0.7235, "step": 11631 }, { "epoch": 0.4212966316551974, "grad_norm": 1.4284744731910015, "learning_rate": 1.298891909943445e-05, "loss": 0.8362, "step": 11632 }, { "epoch": 0.42133285041651575, "grad_norm": 1.3709777663425844, "learning_rate": 1.2987799635631327e-05, "loss": 0.7993, "step": 11633 }, { "epoch": 0.4213690691778341, "grad_norm": 1.271680588971643, "learning_rate": 1.2986680130713451e-05, "loss": 0.7506, "step": 11634 }, { "epoch": 0.4214052879391525, "grad_norm": 1.2778430569456949, "learning_rate": 1.2985560584696236e-05, "loss": 0.7095, "step": 11635 }, { "epoch": 0.42144150670047087, "grad_norm": 1.3735376372248826, "learning_rate": 1.298444099759508e-05, "loss": 0.7964, "step": 11636 }, { "epoch": 0.4214777254617892, "grad_norm": 1.385832071917595, "learning_rate": 1.2983321369425397e-05, "loss": 0.7857, "step": 11637 }, { "epoch": 0.42151394422310756, "grad_norm": 1.250090806613153, "learning_rate": 1.298220170020259e-05, "loss": 0.75, "step": 11638 }, { "epoch": 0.42155016298442594, "grad_norm": 1.3398584283019153, "learning_rate": 1.2981081989942066e-05, "loss": 0.6837, "step": 11639 }, { "epoch": 0.4215863817457443, "grad_norm": 1.3618369567677326, "learning_rate": 1.2979962238659236e-05, "loss": 0.8121, "step": 11640 }, { "epoch": 0.4216226005070627, "grad_norm": 1.154438194508917, "learning_rate": 1.2978842446369506e-05, "loss": 0.7387, "step": 11641 }, { "epoch": 0.421658819268381, "grad_norm": 1.4615387928630283, "learning_rate": 1.2977722613088286e-05, "loss": 0.7547, "step": 11642 }, { "epoch": 0.4216950380296994, "grad_norm": 1.0306726705054061, "learning_rate": 1.2976602738830986e-05, "loss": 0.6954, "step": 11643 }, { "epoch": 0.42173125679101775, "grad_norm": 1.3018220569414185, "learning_rate": 1.2975482823613019e-05, "loss": 0.6468, "step": 11644 }, { "epoch": 0.4217674755523361, "grad_norm": 1.2769767134743724, "learning_rate": 1.2974362867449792e-05, "loss": 0.707, "step": 11645 }, { "epoch": 0.4218036943136545, "grad_norm": 1.1823256504973043, "learning_rate": 1.297324287035672e-05, "loss": 0.7112, "step": 11646 }, { "epoch": 0.4218399130749728, "grad_norm": 1.3200628479125902, "learning_rate": 1.2972122832349212e-05, "loss": 0.7542, "step": 11647 }, { "epoch": 0.4218761318362912, "grad_norm": 1.2664353409390443, "learning_rate": 1.2971002753442683e-05, "loss": 0.7378, "step": 11648 }, { "epoch": 0.42191235059760956, "grad_norm": 1.4519738584294783, "learning_rate": 1.2969882633652546e-05, "loss": 0.7723, "step": 11649 }, { "epoch": 0.42194856935892794, "grad_norm": 1.410694743669347, "learning_rate": 1.2968762472994212e-05, "loss": 0.7653, "step": 11650 }, { "epoch": 0.4219847881202463, "grad_norm": 1.3815641344816199, "learning_rate": 1.29676422714831e-05, "loss": 0.8054, "step": 11651 }, { "epoch": 0.42202100688156463, "grad_norm": 1.108704549682457, "learning_rate": 1.2966522029134622e-05, "loss": 0.6857, "step": 11652 }, { "epoch": 0.422057225642883, "grad_norm": 1.2228632807265367, "learning_rate": 1.2965401745964197e-05, "loss": 0.8029, "step": 11653 }, { "epoch": 0.4220934444042014, "grad_norm": 1.3520113834910656, "learning_rate": 1.2964281421987235e-05, "loss": 0.8004, "step": 11654 }, { "epoch": 0.42212966316551975, "grad_norm": 1.222851889926615, "learning_rate": 1.2963161057219157e-05, "loss": 0.7365, "step": 11655 }, { "epoch": 0.4221658819268381, "grad_norm": 1.3908896710040832, "learning_rate": 1.296204065167538e-05, "loss": 0.7797, "step": 11656 }, { "epoch": 0.42220210068815645, "grad_norm": 1.3423487591652783, "learning_rate": 1.2960920205371322e-05, "loss": 0.7814, "step": 11657 }, { "epoch": 0.4222383194494748, "grad_norm": 1.2926775432948434, "learning_rate": 1.2959799718322396e-05, "loss": 0.6902, "step": 11658 }, { "epoch": 0.4222745382107932, "grad_norm": 1.4973658850342517, "learning_rate": 1.2958679190544028e-05, "loss": 0.7647, "step": 11659 }, { "epoch": 0.42231075697211157, "grad_norm": 1.4504307398448484, "learning_rate": 1.2957558622051633e-05, "loss": 0.7512, "step": 11660 }, { "epoch": 0.42234697573342994, "grad_norm": 1.2963028681302835, "learning_rate": 1.2956438012860634e-05, "loss": 0.7503, "step": 11661 }, { "epoch": 0.42238319449474826, "grad_norm": 1.4907938312041213, "learning_rate": 1.2955317362986453e-05, "loss": 0.7466, "step": 11662 }, { "epoch": 0.42241941325606663, "grad_norm": 1.3215452623900614, "learning_rate": 1.2954196672444503e-05, "loss": 0.6983, "step": 11663 }, { "epoch": 0.422455632017385, "grad_norm": 1.3665331078231273, "learning_rate": 1.2953075941250216e-05, "loss": 0.7862, "step": 11664 }, { "epoch": 0.4224918507787034, "grad_norm": 1.5064783283850782, "learning_rate": 1.2951955169419004e-05, "loss": 0.7616, "step": 11665 }, { "epoch": 0.42252806954002176, "grad_norm": 1.523729690016626, "learning_rate": 1.29508343569663e-05, "loss": 0.7715, "step": 11666 }, { "epoch": 0.4225642883013401, "grad_norm": 1.4024811520644473, "learning_rate": 1.2949713503907516e-05, "loss": 0.7885, "step": 11667 }, { "epoch": 0.42260050706265845, "grad_norm": 1.0388411808532885, "learning_rate": 1.2948592610258088e-05, "loss": 0.7503, "step": 11668 }, { "epoch": 0.4226367258239768, "grad_norm": 1.782554983945223, "learning_rate": 1.2947471676033434e-05, "loss": 0.8017, "step": 11669 }, { "epoch": 0.4226729445852952, "grad_norm": 1.5217902839471305, "learning_rate": 1.2946350701248977e-05, "loss": 0.7839, "step": 11670 }, { "epoch": 0.42270916334661357, "grad_norm": 1.139313404429554, "learning_rate": 1.2945229685920149e-05, "loss": 0.8248, "step": 11671 }, { "epoch": 0.4227453821079319, "grad_norm": 1.3430573291647958, "learning_rate": 1.294410863006237e-05, "loss": 0.7472, "step": 11672 }, { "epoch": 0.42278160086925026, "grad_norm": 1.4384041396981286, "learning_rate": 1.2942987533691071e-05, "loss": 0.7994, "step": 11673 }, { "epoch": 0.42281781963056864, "grad_norm": 1.350149570315463, "learning_rate": 1.2941866396821676e-05, "loss": 0.78, "step": 11674 }, { "epoch": 0.422854038391887, "grad_norm": 1.2201145691139017, "learning_rate": 1.2940745219469619e-05, "loss": 0.7166, "step": 11675 }, { "epoch": 0.4228902571532054, "grad_norm": 1.266214287991896, "learning_rate": 1.293962400165032e-05, "loss": 0.7742, "step": 11676 }, { "epoch": 0.4229264759145237, "grad_norm": 1.0155363051034416, "learning_rate": 1.2938502743379212e-05, "loss": 0.7125, "step": 11677 }, { "epoch": 0.4229626946758421, "grad_norm": 1.3371401902776048, "learning_rate": 1.2937381444671723e-05, "loss": 0.7911, "step": 11678 }, { "epoch": 0.42299891343716045, "grad_norm": 1.4572055011899685, "learning_rate": 1.2936260105543286e-05, "loss": 0.6683, "step": 11679 }, { "epoch": 0.4230351321984788, "grad_norm": 1.4641535137308004, "learning_rate": 1.2935138726009328e-05, "loss": 0.7052, "step": 11680 }, { "epoch": 0.4230713509597972, "grad_norm": 1.153610609130264, "learning_rate": 1.2934017306085284e-05, "loss": 0.7505, "step": 11681 }, { "epoch": 0.4231075697211155, "grad_norm": 1.167404829875186, "learning_rate": 1.2932895845786587e-05, "loss": 0.7353, "step": 11682 }, { "epoch": 0.4231437884824339, "grad_norm": 1.5218924732889267, "learning_rate": 1.2931774345128661e-05, "loss": 0.777, "step": 11683 }, { "epoch": 0.42318000724375227, "grad_norm": 1.3633764276098768, "learning_rate": 1.2930652804126944e-05, "loss": 0.7781, "step": 11684 }, { "epoch": 0.42321622600507064, "grad_norm": 0.9974170200759992, "learning_rate": 1.2929531222796872e-05, "loss": 0.7464, "step": 11685 }, { "epoch": 0.423252444766389, "grad_norm": 1.2389444303661277, "learning_rate": 1.2928409601153875e-05, "loss": 0.784, "step": 11686 }, { "epoch": 0.42328866352770733, "grad_norm": 1.1455118434539142, "learning_rate": 1.2927287939213388e-05, "loss": 0.7777, "step": 11687 }, { "epoch": 0.4233248822890257, "grad_norm": 0.958141076361194, "learning_rate": 1.2926166236990847e-05, "loss": 0.724, "step": 11688 }, { "epoch": 0.4233611010503441, "grad_norm": 1.1838598713453814, "learning_rate": 1.2925044494501688e-05, "loss": 0.7646, "step": 11689 }, { "epoch": 0.42339731981166245, "grad_norm": 1.4252557690853827, "learning_rate": 1.2923922711761347e-05, "loss": 0.796, "step": 11690 }, { "epoch": 0.4234335385729808, "grad_norm": 1.0268689232696204, "learning_rate": 1.292280088878526e-05, "loss": 0.7609, "step": 11691 }, { "epoch": 0.42346975733429915, "grad_norm": 1.2683223470482219, "learning_rate": 1.2921679025588865e-05, "loss": 0.7159, "step": 11692 }, { "epoch": 0.4235059760956175, "grad_norm": 1.319058604910991, "learning_rate": 1.29205571221876e-05, "loss": 0.7919, "step": 11693 }, { "epoch": 0.4235421948569359, "grad_norm": 1.025087814067124, "learning_rate": 1.2919435178596901e-05, "loss": 0.7216, "step": 11694 }, { "epoch": 0.42357841361825427, "grad_norm": 2.1248245427592267, "learning_rate": 1.291831319483221e-05, "loss": 0.7739, "step": 11695 }, { "epoch": 0.42361463237957264, "grad_norm": 1.3095026894845743, "learning_rate": 1.2917191170908964e-05, "loss": 0.7121, "step": 11696 }, { "epoch": 0.42365085114089096, "grad_norm": 1.3243725344165536, "learning_rate": 1.2916069106842607e-05, "loss": 0.7519, "step": 11697 }, { "epoch": 0.42368706990220933, "grad_norm": 1.26565764636744, "learning_rate": 1.2914947002648573e-05, "loss": 0.7278, "step": 11698 }, { "epoch": 0.4237232886635277, "grad_norm": 1.6516563798465071, "learning_rate": 1.2913824858342308e-05, "loss": 0.7259, "step": 11699 }, { "epoch": 0.4237595074248461, "grad_norm": 1.4132062014564686, "learning_rate": 1.2912702673939252e-05, "loss": 0.7353, "step": 11700 }, { "epoch": 0.42379572618616446, "grad_norm": 1.2543998625713664, "learning_rate": 1.291158044945485e-05, "loss": 0.6817, "step": 11701 }, { "epoch": 0.4238319449474828, "grad_norm": 1.331639936446406, "learning_rate": 1.2910458184904542e-05, "loss": 0.7276, "step": 11702 }, { "epoch": 0.42386816370880115, "grad_norm": 1.5218163000913831, "learning_rate": 1.2909335880303772e-05, "loss": 0.7312, "step": 11703 }, { "epoch": 0.4239043824701195, "grad_norm": 1.4013984490606088, "learning_rate": 1.2908213535667985e-05, "loss": 0.7509, "step": 11704 }, { "epoch": 0.4239406012314379, "grad_norm": 1.3488553955979163, "learning_rate": 1.2907091151012621e-05, "loss": 0.7692, "step": 11705 }, { "epoch": 0.42397681999275627, "grad_norm": 1.436107797740745, "learning_rate": 1.2905968726353133e-05, "loss": 0.7849, "step": 11706 }, { "epoch": 0.4240130387540746, "grad_norm": 1.3868082051235797, "learning_rate": 1.2904846261704957e-05, "loss": 0.7236, "step": 11707 }, { "epoch": 0.42404925751539296, "grad_norm": 1.1654833596363943, "learning_rate": 1.2903723757083548e-05, "loss": 0.7584, "step": 11708 }, { "epoch": 0.42408547627671134, "grad_norm": 1.372003575395511, "learning_rate": 1.2902601212504346e-05, "loss": 0.7538, "step": 11709 }, { "epoch": 0.4241216950380297, "grad_norm": 1.6065737641063382, "learning_rate": 1.2901478627982802e-05, "loss": 0.807, "step": 11710 }, { "epoch": 0.4241579137993481, "grad_norm": 1.3328405072644858, "learning_rate": 1.2900356003534364e-05, "loss": 0.7559, "step": 11711 }, { "epoch": 0.4241941325606664, "grad_norm": 1.337171657230599, "learning_rate": 1.2899233339174474e-05, "loss": 0.7645, "step": 11712 }, { "epoch": 0.4242303513219848, "grad_norm": 1.3679033530663467, "learning_rate": 1.2898110634918592e-05, "loss": 0.741, "step": 11713 }, { "epoch": 0.42426657008330315, "grad_norm": 1.3110237186792524, "learning_rate": 1.2896987890782158e-05, "loss": 0.7969, "step": 11714 }, { "epoch": 0.4243027888446215, "grad_norm": 1.3174931598913353, "learning_rate": 1.2895865106780625e-05, "loss": 0.7186, "step": 11715 }, { "epoch": 0.4243390076059399, "grad_norm": 1.3534314024787217, "learning_rate": 1.2894742282929444e-05, "loss": 0.7778, "step": 11716 }, { "epoch": 0.4243752263672582, "grad_norm": 1.286072993208907, "learning_rate": 1.2893619419244068e-05, "loss": 0.7044, "step": 11717 }, { "epoch": 0.4244114451285766, "grad_norm": 1.4610900104445477, "learning_rate": 1.2892496515739943e-05, "loss": 0.8438, "step": 11718 }, { "epoch": 0.42444766388989497, "grad_norm": 1.4394638030071503, "learning_rate": 1.2891373572432524e-05, "loss": 0.7612, "step": 11719 }, { "epoch": 0.42448388265121334, "grad_norm": 1.2588751432759455, "learning_rate": 1.2890250589337265e-05, "loss": 0.7859, "step": 11720 }, { "epoch": 0.4245201014125317, "grad_norm": 1.4796520272708917, "learning_rate": 1.288912756646962e-05, "loss": 0.7096, "step": 11721 }, { "epoch": 0.42455632017385003, "grad_norm": 1.3268986460575056, "learning_rate": 1.288800450384504e-05, "loss": 0.7537, "step": 11722 }, { "epoch": 0.4245925389351684, "grad_norm": 1.1995066989113345, "learning_rate": 1.2886881401478978e-05, "loss": 0.7629, "step": 11723 }, { "epoch": 0.4246287576964868, "grad_norm": 1.3712192108484578, "learning_rate": 1.2885758259386896e-05, "loss": 0.7851, "step": 11724 }, { "epoch": 0.42466497645780515, "grad_norm": 1.323902277981596, "learning_rate": 1.288463507758424e-05, "loss": 0.7823, "step": 11725 }, { "epoch": 0.42470119521912353, "grad_norm": 1.3954698411969084, "learning_rate": 1.2883511856086475e-05, "loss": 0.7219, "step": 11726 }, { "epoch": 0.42473741398044185, "grad_norm": 1.045032863760933, "learning_rate": 1.288238859490905e-05, "loss": 0.7572, "step": 11727 }, { "epoch": 0.4247736327417602, "grad_norm": 1.4634944410668966, "learning_rate": 1.2881265294067426e-05, "loss": 0.817, "step": 11728 }, { "epoch": 0.4248098515030786, "grad_norm": 1.323337457190363, "learning_rate": 1.288014195357706e-05, "loss": 0.686, "step": 11729 }, { "epoch": 0.42484607026439697, "grad_norm": 1.2113161745692083, "learning_rate": 1.287901857345341e-05, "loss": 0.7821, "step": 11730 }, { "epoch": 0.42488228902571534, "grad_norm": 1.2196532693959081, "learning_rate": 1.2877895153711935e-05, "loss": 0.7431, "step": 11731 }, { "epoch": 0.42491850778703366, "grad_norm": 1.4300931839617654, "learning_rate": 1.2876771694368094e-05, "loss": 0.8459, "step": 11732 }, { "epoch": 0.42495472654835204, "grad_norm": 1.0811628686095256, "learning_rate": 1.2875648195437347e-05, "loss": 0.7245, "step": 11733 }, { "epoch": 0.4249909453096704, "grad_norm": 1.3220959016349751, "learning_rate": 1.2874524656935153e-05, "loss": 0.7352, "step": 11734 }, { "epoch": 0.4250271640709888, "grad_norm": 1.4286610691756856, "learning_rate": 1.2873401078876974e-05, "loss": 0.6769, "step": 11735 }, { "epoch": 0.42506338283230716, "grad_norm": 1.3613720918229764, "learning_rate": 1.2872277461278272e-05, "loss": 0.8114, "step": 11736 }, { "epoch": 0.4250996015936255, "grad_norm": 1.452021651643181, "learning_rate": 1.2871153804154511e-05, "loss": 0.7226, "step": 11737 }, { "epoch": 0.42513582035494385, "grad_norm": 1.4066764959652032, "learning_rate": 1.2870030107521145e-05, "loss": 0.7932, "step": 11738 }, { "epoch": 0.4251720391162622, "grad_norm": 1.3470571692910862, "learning_rate": 1.2868906371393648e-05, "loss": 0.7329, "step": 11739 }, { "epoch": 0.4252082578775806, "grad_norm": 1.325815320012719, "learning_rate": 1.2867782595787474e-05, "loss": 0.6541, "step": 11740 }, { "epoch": 0.42524447663889897, "grad_norm": 1.2877614738400873, "learning_rate": 1.2866658780718096e-05, "loss": 0.7611, "step": 11741 }, { "epoch": 0.4252806954002173, "grad_norm": 1.4573591243073631, "learning_rate": 1.2865534926200974e-05, "loss": 0.7668, "step": 11742 }, { "epoch": 0.42531691416153566, "grad_norm": 1.5066694426012084, "learning_rate": 1.286441103225157e-05, "loss": 0.7466, "step": 11743 }, { "epoch": 0.42535313292285404, "grad_norm": 1.3699902362153609, "learning_rate": 1.2863287098885358e-05, "loss": 0.6889, "step": 11744 }, { "epoch": 0.4253893516841724, "grad_norm": 1.3632635508921362, "learning_rate": 1.2862163126117794e-05, "loss": 0.7864, "step": 11745 }, { "epoch": 0.4254255704454908, "grad_norm": 1.3963349095633637, "learning_rate": 1.2861039113964354e-05, "loss": 0.7895, "step": 11746 }, { "epoch": 0.4254617892068091, "grad_norm": 1.4839762528556952, "learning_rate": 1.2859915062440502e-05, "loss": 0.7153, "step": 11747 }, { "epoch": 0.4254980079681275, "grad_norm": 1.2993410302387405, "learning_rate": 1.2858790971561704e-05, "loss": 0.7481, "step": 11748 }, { "epoch": 0.42553422672944585, "grad_norm": 1.134587901834901, "learning_rate": 1.2857666841343433e-05, "loss": 0.792, "step": 11749 }, { "epoch": 0.4255704454907642, "grad_norm": 1.2362064705485716, "learning_rate": 1.285654267180115e-05, "loss": 0.7768, "step": 11750 }, { "epoch": 0.4256066642520826, "grad_norm": 0.9893922916789659, "learning_rate": 1.2855418462950335e-05, "loss": 0.7518, "step": 11751 }, { "epoch": 0.4256428830134009, "grad_norm": 1.313822851656443, "learning_rate": 1.2854294214806453e-05, "loss": 0.7464, "step": 11752 }, { "epoch": 0.4256791017747193, "grad_norm": 1.3702584103475561, "learning_rate": 1.2853169927384972e-05, "loss": 0.8213, "step": 11753 }, { "epoch": 0.42571532053603767, "grad_norm": 1.2007903957724442, "learning_rate": 1.2852045600701366e-05, "loss": 0.7566, "step": 11754 }, { "epoch": 0.42575153929735604, "grad_norm": 1.3626414528802135, "learning_rate": 1.285092123477111e-05, "loss": 0.7606, "step": 11755 }, { "epoch": 0.4257877580586744, "grad_norm": 1.3263945701546382, "learning_rate": 1.2849796829609669e-05, "loss": 0.6838, "step": 11756 }, { "epoch": 0.42582397681999273, "grad_norm": 1.2893918094520762, "learning_rate": 1.2848672385232519e-05, "loss": 0.763, "step": 11757 }, { "epoch": 0.4258601955813111, "grad_norm": 1.2393918329431313, "learning_rate": 1.2847547901655134e-05, "loss": 0.6785, "step": 11758 }, { "epoch": 0.4258964143426295, "grad_norm": 1.4655062537900958, "learning_rate": 1.2846423378892992e-05, "loss": 0.8542, "step": 11759 }, { "epoch": 0.42593263310394786, "grad_norm": 1.548726753198174, "learning_rate": 1.284529881696156e-05, "loss": 0.7309, "step": 11760 }, { "epoch": 0.42596885186526623, "grad_norm": 1.0121229341423001, "learning_rate": 1.2844174215876315e-05, "loss": 0.727, "step": 11761 }, { "epoch": 0.42600507062658455, "grad_norm": 0.9400017906584505, "learning_rate": 1.2843049575652737e-05, "loss": 0.828, "step": 11762 }, { "epoch": 0.4260412893879029, "grad_norm": 1.1445340212057695, "learning_rate": 1.2841924896306298e-05, "loss": 0.8132, "step": 11763 }, { "epoch": 0.4260775081492213, "grad_norm": 1.2740523903022016, "learning_rate": 1.2840800177852474e-05, "loss": 0.6604, "step": 11764 }, { "epoch": 0.42611372691053967, "grad_norm": 1.0915427336003243, "learning_rate": 1.2839675420306742e-05, "loss": 0.7248, "step": 11765 }, { "epoch": 0.42614994567185804, "grad_norm": 1.3542462437124094, "learning_rate": 1.2838550623684585e-05, "loss": 0.734, "step": 11766 }, { "epoch": 0.42618616443317636, "grad_norm": 1.8748055794371297, "learning_rate": 1.2837425788001475e-05, "loss": 0.7621, "step": 11767 }, { "epoch": 0.42622238319449474, "grad_norm": 1.143192205815055, "learning_rate": 1.2836300913272892e-05, "loss": 0.7522, "step": 11768 }, { "epoch": 0.4262586019558131, "grad_norm": 1.336797535696441, "learning_rate": 1.2835175999514318e-05, "loss": 0.7476, "step": 11769 }, { "epoch": 0.4262948207171315, "grad_norm": 1.0126652844762967, "learning_rate": 1.2834051046741229e-05, "loss": 0.7147, "step": 11770 }, { "epoch": 0.42633103947844986, "grad_norm": 1.3187749864541094, "learning_rate": 1.2832926054969111e-05, "loss": 0.692, "step": 11771 }, { "epoch": 0.4263672582397682, "grad_norm": 1.3603221085009907, "learning_rate": 1.2831801024213437e-05, "loss": 0.7609, "step": 11772 }, { "epoch": 0.42640347700108655, "grad_norm": 1.3037763114854501, "learning_rate": 1.2830675954489697e-05, "loss": 0.7342, "step": 11773 }, { "epoch": 0.4264396957624049, "grad_norm": 1.7765762977768855, "learning_rate": 1.2829550845813365e-05, "loss": 0.8304, "step": 11774 }, { "epoch": 0.4264759145237233, "grad_norm": 1.2900656546509373, "learning_rate": 1.2828425698199928e-05, "loss": 0.7772, "step": 11775 }, { "epoch": 0.4265121332850417, "grad_norm": 1.1245210992344836, "learning_rate": 1.2827300511664868e-05, "loss": 0.7927, "step": 11776 }, { "epoch": 0.42654835204636, "grad_norm": 1.3727880141642512, "learning_rate": 1.282617528622367e-05, "loss": 0.6941, "step": 11777 }, { "epoch": 0.42658457080767836, "grad_norm": 1.2868738154250237, "learning_rate": 1.2825050021891816e-05, "loss": 0.8052, "step": 11778 }, { "epoch": 0.42662078956899674, "grad_norm": 1.2955297960520782, "learning_rate": 1.282392471868479e-05, "loss": 0.8349, "step": 11779 }, { "epoch": 0.4266570083303151, "grad_norm": 1.2844850169929674, "learning_rate": 1.2822799376618083e-05, "loss": 0.712, "step": 11780 }, { "epoch": 0.4266932270916335, "grad_norm": 1.1274338965044004, "learning_rate": 1.2821673995707171e-05, "loss": 0.7527, "step": 11781 }, { "epoch": 0.4267294458529518, "grad_norm": 1.443572717728735, "learning_rate": 1.2820548575967549e-05, "loss": 0.7949, "step": 11782 }, { "epoch": 0.4267656646142702, "grad_norm": 1.0462342256621031, "learning_rate": 1.2819423117414698e-05, "loss": 0.7642, "step": 11783 }, { "epoch": 0.42680188337558855, "grad_norm": 1.3910328556269456, "learning_rate": 1.2818297620064109e-05, "loss": 0.7618, "step": 11784 }, { "epoch": 0.4268381021369069, "grad_norm": 1.3718259390148337, "learning_rate": 1.2817172083931266e-05, "loss": 0.7524, "step": 11785 }, { "epoch": 0.4268743208982253, "grad_norm": 1.3654480734149168, "learning_rate": 1.2816046509031663e-05, "loss": 0.8043, "step": 11786 }, { "epoch": 0.4269105396595436, "grad_norm": 1.4863657695950672, "learning_rate": 1.2814920895380785e-05, "loss": 0.7921, "step": 11787 }, { "epoch": 0.426946758420862, "grad_norm": 1.452505303857824, "learning_rate": 1.2813795242994119e-05, "loss": 0.8089, "step": 11788 }, { "epoch": 0.42698297718218037, "grad_norm": 1.0449873527172628, "learning_rate": 1.2812669551887163e-05, "loss": 0.735, "step": 11789 }, { "epoch": 0.42701919594349874, "grad_norm": 1.729239403868547, "learning_rate": 1.2811543822075396e-05, "loss": 0.7596, "step": 11790 }, { "epoch": 0.4270554147048171, "grad_norm": 1.1297826002326006, "learning_rate": 1.281041805357432e-05, "loss": 0.8154, "step": 11791 }, { "epoch": 0.42709163346613543, "grad_norm": 1.133958076610425, "learning_rate": 1.2809292246399422e-05, "loss": 0.7154, "step": 11792 }, { "epoch": 0.4271278522274538, "grad_norm": 1.3244801638834842, "learning_rate": 1.2808166400566195e-05, "loss": 0.7469, "step": 11793 }, { "epoch": 0.4271640709887722, "grad_norm": 1.0745838163760415, "learning_rate": 1.2807040516090128e-05, "loss": 0.7489, "step": 11794 }, { "epoch": 0.42720028975009056, "grad_norm": 1.2653555009153126, "learning_rate": 1.280591459298672e-05, "loss": 0.7718, "step": 11795 }, { "epoch": 0.42723650851140893, "grad_norm": 1.1009416936049554, "learning_rate": 1.2804788631271459e-05, "loss": 0.7385, "step": 11796 }, { "epoch": 0.42727272727272725, "grad_norm": 1.0727495737947639, "learning_rate": 1.2803662630959845e-05, "loss": 0.7269, "step": 11797 }, { "epoch": 0.4273089460340456, "grad_norm": 1.3710259304635326, "learning_rate": 1.2802536592067367e-05, "loss": 0.7842, "step": 11798 }, { "epoch": 0.427345164795364, "grad_norm": 1.2986768673896811, "learning_rate": 1.2801410514609522e-05, "loss": 0.7599, "step": 11799 }, { "epoch": 0.42738138355668237, "grad_norm": 1.1946624729977664, "learning_rate": 1.280028439860181e-05, "loss": 0.7373, "step": 11800 }, { "epoch": 0.42741760231800074, "grad_norm": 1.298914514981327, "learning_rate": 1.2799158244059724e-05, "loss": 0.7795, "step": 11801 }, { "epoch": 0.42745382107931906, "grad_norm": 1.360023638380918, "learning_rate": 1.279803205099876e-05, "loss": 0.7789, "step": 11802 }, { "epoch": 0.42749003984063744, "grad_norm": 1.4604201395244727, "learning_rate": 1.2796905819434415e-05, "loss": 0.7628, "step": 11803 }, { "epoch": 0.4275262586019558, "grad_norm": 1.2978713053456734, "learning_rate": 1.279577954938219e-05, "loss": 0.778, "step": 11804 }, { "epoch": 0.4275624773632742, "grad_norm": 1.3194214385851075, "learning_rate": 1.2794653240857582e-05, "loss": 0.7099, "step": 11805 }, { "epoch": 0.42759869612459256, "grad_norm": 1.6583734633771512, "learning_rate": 1.279352689387609e-05, "loss": 0.8043, "step": 11806 }, { "epoch": 0.4276349148859109, "grad_norm": 1.3588816923321976, "learning_rate": 1.279240050845321e-05, "loss": 0.778, "step": 11807 }, { "epoch": 0.42767113364722925, "grad_norm": 1.382075199400932, "learning_rate": 1.2791274084604449e-05, "loss": 0.7184, "step": 11808 }, { "epoch": 0.4277073524085476, "grad_norm": 1.3741121839590014, "learning_rate": 1.2790147622345303e-05, "loss": 0.7353, "step": 11809 }, { "epoch": 0.427743571169866, "grad_norm": 1.6130901666861308, "learning_rate": 1.2789021121691273e-05, "loss": 0.7634, "step": 11810 }, { "epoch": 0.4277797899311844, "grad_norm": 1.418620332913575, "learning_rate": 1.2787894582657864e-05, "loss": 0.7386, "step": 11811 }, { "epoch": 0.4278160086925027, "grad_norm": 1.2646650497346987, "learning_rate": 1.2786768005260574e-05, "loss": 0.7493, "step": 11812 }, { "epoch": 0.42785222745382107, "grad_norm": 1.2374147375252473, "learning_rate": 1.278564138951491e-05, "loss": 0.7535, "step": 11813 }, { "epoch": 0.42788844621513944, "grad_norm": 1.395194364146307, "learning_rate": 1.278451473543637e-05, "loss": 0.8038, "step": 11814 }, { "epoch": 0.4279246649764578, "grad_norm": 1.3072223523053772, "learning_rate": 1.278338804304046e-05, "loss": 0.7367, "step": 11815 }, { "epoch": 0.4279608837377762, "grad_norm": 1.0517102132676979, "learning_rate": 1.2782261312342687e-05, "loss": 0.7088, "step": 11816 }, { "epoch": 0.4279971024990945, "grad_norm": 1.3576846911183946, "learning_rate": 1.2781134543358556e-05, "loss": 0.7982, "step": 11817 }, { "epoch": 0.4280333212604129, "grad_norm": 1.3146043664949398, "learning_rate": 1.2780007736103563e-05, "loss": 0.7602, "step": 11818 }, { "epoch": 0.42806954002173125, "grad_norm": 1.5187254696207182, "learning_rate": 1.2778880890593227e-05, "loss": 0.6875, "step": 11819 }, { "epoch": 0.42810575878304963, "grad_norm": 1.339515901720351, "learning_rate": 1.2777754006843046e-05, "loss": 0.7575, "step": 11820 }, { "epoch": 0.428141977544368, "grad_norm": 1.3591324244483858, "learning_rate": 1.2776627084868527e-05, "loss": 0.7828, "step": 11821 }, { "epoch": 0.4281781963056863, "grad_norm": 1.441797946287722, "learning_rate": 1.2775500124685183e-05, "loss": 0.8069, "step": 11822 }, { "epoch": 0.4282144150670047, "grad_norm": 1.288841180875739, "learning_rate": 1.2774373126308517e-05, "loss": 0.7062, "step": 11823 }, { "epoch": 0.42825063382832307, "grad_norm": 1.3703730623635522, "learning_rate": 1.277324608975404e-05, "loss": 0.7597, "step": 11824 }, { "epoch": 0.42828685258964144, "grad_norm": 1.1192220019896622, "learning_rate": 1.2772119015037259e-05, "loss": 0.7502, "step": 11825 }, { "epoch": 0.4283230713509598, "grad_norm": 1.2594409716509403, "learning_rate": 1.2770991902173683e-05, "loss": 0.7405, "step": 11826 }, { "epoch": 0.42835929011227813, "grad_norm": 1.3475903329070145, "learning_rate": 1.2769864751178827e-05, "loss": 0.7546, "step": 11827 }, { "epoch": 0.4283955088735965, "grad_norm": 1.2743155861585231, "learning_rate": 1.2768737562068197e-05, "loss": 0.7872, "step": 11828 }, { "epoch": 0.4284317276349149, "grad_norm": 1.3798150803882723, "learning_rate": 1.2767610334857303e-05, "loss": 0.7012, "step": 11829 }, { "epoch": 0.42846794639623326, "grad_norm": 0.9883106310603628, "learning_rate": 1.2766483069561662e-05, "loss": 0.7394, "step": 11830 }, { "epoch": 0.42850416515755163, "grad_norm": 1.2645032283824993, "learning_rate": 1.2765355766196781e-05, "loss": 0.6957, "step": 11831 }, { "epoch": 0.42854038391886995, "grad_norm": 1.507990776527045, "learning_rate": 1.2764228424778175e-05, "loss": 0.7568, "step": 11832 }, { "epoch": 0.4285766026801883, "grad_norm": 1.3697883703520919, "learning_rate": 1.276310104532136e-05, "loss": 0.7815, "step": 11833 }, { "epoch": 0.4286128214415067, "grad_norm": 1.1233421350117665, "learning_rate": 1.2761973627841844e-05, "loss": 0.7459, "step": 11834 }, { "epoch": 0.42864904020282507, "grad_norm": 1.340177473596452, "learning_rate": 1.2760846172355144e-05, "loss": 0.7487, "step": 11835 }, { "epoch": 0.42868525896414345, "grad_norm": 1.4133184092809785, "learning_rate": 1.2759718678876775e-05, "loss": 0.7826, "step": 11836 }, { "epoch": 0.42872147772546176, "grad_norm": 1.5084456710404452, "learning_rate": 1.2758591147422253e-05, "loss": 0.842, "step": 11837 }, { "epoch": 0.42875769648678014, "grad_norm": 1.6782517950898492, "learning_rate": 1.2757463578007092e-05, "loss": 0.7995, "step": 11838 }, { "epoch": 0.4287939152480985, "grad_norm": 1.2792647696244754, "learning_rate": 1.2756335970646809e-05, "loss": 0.7596, "step": 11839 }, { "epoch": 0.4288301340094169, "grad_norm": 1.4076417777648886, "learning_rate": 1.2755208325356922e-05, "loss": 0.7506, "step": 11840 }, { "epoch": 0.42886635277073526, "grad_norm": 1.4459347493755823, "learning_rate": 1.2754080642152946e-05, "loss": 0.7517, "step": 11841 }, { "epoch": 0.4289025715320536, "grad_norm": 1.3603365651527393, "learning_rate": 1.2752952921050403e-05, "loss": 0.7256, "step": 11842 }, { "epoch": 0.42893879029337195, "grad_norm": 1.0739140815982993, "learning_rate": 1.2751825162064807e-05, "loss": 0.7016, "step": 11843 }, { "epoch": 0.4289750090546903, "grad_norm": 1.4249624906170835, "learning_rate": 1.2750697365211679e-05, "loss": 0.7074, "step": 11844 }, { "epoch": 0.4290112278160087, "grad_norm": 1.6337883362708612, "learning_rate": 1.2749569530506538e-05, "loss": 0.673, "step": 11845 }, { "epoch": 0.4290474465773271, "grad_norm": 0.999448071134225, "learning_rate": 1.2748441657964907e-05, "loss": 0.7719, "step": 11846 }, { "epoch": 0.4290836653386454, "grad_norm": 1.3080309169893811, "learning_rate": 1.2747313747602299e-05, "loss": 0.7067, "step": 11847 }, { "epoch": 0.42911988409996377, "grad_norm": 1.1360887934828527, "learning_rate": 1.2746185799434243e-05, "loss": 0.7269, "step": 11848 }, { "epoch": 0.42915610286128214, "grad_norm": 1.0657227972475005, "learning_rate": 1.2745057813476258e-05, "loss": 0.7422, "step": 11849 }, { "epoch": 0.4291923216226005, "grad_norm": 1.4866088741140828, "learning_rate": 1.2743929789743863e-05, "loss": 0.7651, "step": 11850 }, { "epoch": 0.4292285403839189, "grad_norm": 1.3352854026729504, "learning_rate": 1.2742801728252587e-05, "loss": 0.7623, "step": 11851 }, { "epoch": 0.4292647591452372, "grad_norm": 1.2761277738213228, "learning_rate": 1.2741673629017946e-05, "loss": 0.7401, "step": 11852 }, { "epoch": 0.4293009779065556, "grad_norm": 1.5293140580928093, "learning_rate": 1.2740545492055467e-05, "loss": 0.7714, "step": 11853 }, { "epoch": 0.42933719666787395, "grad_norm": 1.3484341988241335, "learning_rate": 1.2739417317380674e-05, "loss": 0.7216, "step": 11854 }, { "epoch": 0.42937341542919233, "grad_norm": 1.4353277700946183, "learning_rate": 1.2738289105009091e-05, "loss": 0.7834, "step": 11855 }, { "epoch": 0.4294096341905107, "grad_norm": 1.270199468523576, "learning_rate": 1.2737160854956245e-05, "loss": 0.6793, "step": 11856 }, { "epoch": 0.429445852951829, "grad_norm": 1.193489853116393, "learning_rate": 1.273603256723766e-05, "loss": 0.7873, "step": 11857 }, { "epoch": 0.4294820717131474, "grad_norm": 1.5472892750755427, "learning_rate": 1.2734904241868864e-05, "loss": 0.7787, "step": 11858 }, { "epoch": 0.42951829047446577, "grad_norm": 1.2466572075068623, "learning_rate": 1.2733775878865382e-05, "loss": 0.6853, "step": 11859 }, { "epoch": 0.42955450923578414, "grad_norm": 1.3733587168642, "learning_rate": 1.2732647478242741e-05, "loss": 0.8245, "step": 11860 }, { "epoch": 0.4295907279971025, "grad_norm": 1.3023445047869493, "learning_rate": 1.273151904001647e-05, "loss": 0.7137, "step": 11861 }, { "epoch": 0.42962694675842084, "grad_norm": 1.568277970311832, "learning_rate": 1.2730390564202097e-05, "loss": 0.8162, "step": 11862 }, { "epoch": 0.4296631655197392, "grad_norm": 1.5302868801712057, "learning_rate": 1.272926205081515e-05, "loss": 0.8184, "step": 11863 }, { "epoch": 0.4296993842810576, "grad_norm": 1.3098391471021071, "learning_rate": 1.272813349987116e-05, "loss": 0.7551, "step": 11864 }, { "epoch": 0.42973560304237596, "grad_norm": 1.3388306907003273, "learning_rate": 1.2727004911385657e-05, "loss": 0.6476, "step": 11865 }, { "epoch": 0.42977182180369433, "grad_norm": 1.3215716229739758, "learning_rate": 1.2725876285374166e-05, "loss": 0.7715, "step": 11866 }, { "epoch": 0.42980804056501265, "grad_norm": 1.2482238153088363, "learning_rate": 1.2724747621852227e-05, "loss": 0.7457, "step": 11867 }, { "epoch": 0.429844259326331, "grad_norm": 1.336602212941568, "learning_rate": 1.2723618920835364e-05, "loss": 0.7884, "step": 11868 }, { "epoch": 0.4298804780876494, "grad_norm": 1.0989166384283322, "learning_rate": 1.2722490182339114e-05, "loss": 0.7167, "step": 11869 }, { "epoch": 0.42991669684896777, "grad_norm": 1.5080125285671337, "learning_rate": 1.2721361406379003e-05, "loss": 0.7632, "step": 11870 }, { "epoch": 0.42995291561028615, "grad_norm": 1.3635752649159125, "learning_rate": 1.2720232592970571e-05, "loss": 0.7632, "step": 11871 }, { "epoch": 0.42998913437160446, "grad_norm": 1.1761805449535698, "learning_rate": 1.2719103742129346e-05, "loss": 0.8152, "step": 11872 }, { "epoch": 0.43002535313292284, "grad_norm": 1.4836605586295846, "learning_rate": 1.2717974853870866e-05, "loss": 0.7909, "step": 11873 }, { "epoch": 0.4300615718942412, "grad_norm": 1.4878105873639658, "learning_rate": 1.2716845928210662e-05, "loss": 0.6967, "step": 11874 }, { "epoch": 0.4300977906555596, "grad_norm": 0.9989683602994983, "learning_rate": 1.2715716965164273e-05, "loss": 0.6689, "step": 11875 }, { "epoch": 0.43013400941687796, "grad_norm": 1.3912397215378691, "learning_rate": 1.2714587964747231e-05, "loss": 0.746, "step": 11876 }, { "epoch": 0.4301702281781963, "grad_norm": 1.2893342013556421, "learning_rate": 1.2713458926975073e-05, "loss": 0.7007, "step": 11877 }, { "epoch": 0.43020644693951465, "grad_norm": 1.3867388844707342, "learning_rate": 1.2712329851863337e-05, "loss": 0.7412, "step": 11878 }, { "epoch": 0.430242665700833, "grad_norm": 0.9938213940344431, "learning_rate": 1.2711200739427558e-05, "loss": 0.7495, "step": 11879 }, { "epoch": 0.4302788844621514, "grad_norm": 1.2414826036550675, "learning_rate": 1.2710071589683275e-05, "loss": 0.7938, "step": 11880 }, { "epoch": 0.4303151032234698, "grad_norm": 1.3693791078371393, "learning_rate": 1.2708942402646025e-05, "loss": 0.7432, "step": 11881 }, { "epoch": 0.4303513219847881, "grad_norm": 1.52994492138238, "learning_rate": 1.270781317833135e-05, "loss": 0.8956, "step": 11882 }, { "epoch": 0.43038754074610647, "grad_norm": 1.2497366349768073, "learning_rate": 1.2706683916754782e-05, "loss": 0.8058, "step": 11883 }, { "epoch": 0.43042375950742484, "grad_norm": 1.331135203024682, "learning_rate": 1.270555461793187e-05, "loss": 0.7148, "step": 11884 }, { "epoch": 0.4304599782687432, "grad_norm": 1.1931252433274293, "learning_rate": 1.2704425281878145e-05, "loss": 0.7728, "step": 11885 }, { "epoch": 0.4304961970300616, "grad_norm": 1.3027529893886394, "learning_rate": 1.2703295908609153e-05, "loss": 0.7177, "step": 11886 }, { "epoch": 0.4305324157913799, "grad_norm": 1.5128693957603614, "learning_rate": 1.2702166498140436e-05, "loss": 0.7586, "step": 11887 }, { "epoch": 0.4305686345526983, "grad_norm": 1.3026028381552766, "learning_rate": 1.270103705048753e-05, "loss": 0.7555, "step": 11888 }, { "epoch": 0.43060485331401666, "grad_norm": 1.18128081086715, "learning_rate": 1.2699907565665983e-05, "loss": 0.7747, "step": 11889 }, { "epoch": 0.43064107207533503, "grad_norm": 1.213904994477876, "learning_rate": 1.2698778043691336e-05, "loss": 0.7717, "step": 11890 }, { "epoch": 0.4306772908366534, "grad_norm": 1.4129603612868644, "learning_rate": 1.2697648484579133e-05, "loss": 0.7603, "step": 11891 }, { "epoch": 0.4307135095979717, "grad_norm": 1.4059656332507597, "learning_rate": 1.2696518888344911e-05, "loss": 0.8074, "step": 11892 }, { "epoch": 0.4307497283592901, "grad_norm": 1.1903920274282038, "learning_rate": 1.2695389255004224e-05, "loss": 0.7912, "step": 11893 }, { "epoch": 0.43078594712060847, "grad_norm": 1.4454370071291365, "learning_rate": 1.2694259584572612e-05, "loss": 0.8336, "step": 11894 }, { "epoch": 0.43082216588192684, "grad_norm": 1.405847853538912, "learning_rate": 1.2693129877065622e-05, "loss": 0.7227, "step": 11895 }, { "epoch": 0.4308583846432452, "grad_norm": 1.8558377792396215, "learning_rate": 1.2692000132498796e-05, "loss": 0.7733, "step": 11896 }, { "epoch": 0.43089460340456354, "grad_norm": 1.2813778608477855, "learning_rate": 1.2690870350887683e-05, "loss": 0.7046, "step": 11897 }, { "epoch": 0.4309308221658819, "grad_norm": 1.4123606016464318, "learning_rate": 1.2689740532247832e-05, "loss": 0.7904, "step": 11898 }, { "epoch": 0.4309670409272003, "grad_norm": 2.2847011463473614, "learning_rate": 1.2688610676594783e-05, "loss": 0.8491, "step": 11899 }, { "epoch": 0.43100325968851866, "grad_norm": 1.1537046359429175, "learning_rate": 1.2687480783944096e-05, "loss": 0.738, "step": 11900 }, { "epoch": 0.43103947844983703, "grad_norm": 1.149464819722481, "learning_rate": 1.2686350854311304e-05, "loss": 0.7352, "step": 11901 }, { "epoch": 0.43107569721115535, "grad_norm": 1.460475180397394, "learning_rate": 1.268522088771197e-05, "loss": 0.7683, "step": 11902 }, { "epoch": 0.4311119159724737, "grad_norm": 1.3797895908925677, "learning_rate": 1.2684090884161636e-05, "loss": 0.7737, "step": 11903 }, { "epoch": 0.4311481347337921, "grad_norm": 1.4001301632289926, "learning_rate": 1.2682960843675852e-05, "loss": 0.7276, "step": 11904 }, { "epoch": 0.4311843534951105, "grad_norm": 1.4799530507527316, "learning_rate": 1.2681830766270168e-05, "loss": 0.7738, "step": 11905 }, { "epoch": 0.43122057225642885, "grad_norm": 1.3588321378755335, "learning_rate": 1.2680700651960138e-05, "loss": 0.7005, "step": 11906 }, { "epoch": 0.4312567910177472, "grad_norm": 1.571092314698158, "learning_rate": 1.2679570500761309e-05, "loss": 0.7438, "step": 11907 }, { "epoch": 0.43129300977906554, "grad_norm": 1.3618878954112303, "learning_rate": 1.2678440312689238e-05, "loss": 0.8012, "step": 11908 }, { "epoch": 0.4313292285403839, "grad_norm": 1.3322612516335555, "learning_rate": 1.2677310087759476e-05, "loss": 0.7636, "step": 11909 }, { "epoch": 0.4313654473017023, "grad_norm": 1.271607576135186, "learning_rate": 1.2676179825987572e-05, "loss": 0.7098, "step": 11910 }, { "epoch": 0.43140166606302066, "grad_norm": 1.448220383370002, "learning_rate": 1.2675049527389083e-05, "loss": 0.7474, "step": 11911 }, { "epoch": 0.43143788482433904, "grad_norm": 1.2651478913394307, "learning_rate": 1.267391919197956e-05, "loss": 0.7337, "step": 11912 }, { "epoch": 0.43147410358565735, "grad_norm": 1.333955000233508, "learning_rate": 1.2672788819774562e-05, "loss": 0.8229, "step": 11913 }, { "epoch": 0.4315103223469757, "grad_norm": 1.0582481970372986, "learning_rate": 1.2671658410789642e-05, "loss": 0.7459, "step": 11914 }, { "epoch": 0.4315465411082941, "grad_norm": 1.3529791363064865, "learning_rate": 1.2670527965040353e-05, "loss": 0.7664, "step": 11915 }, { "epoch": 0.4315827598696125, "grad_norm": 1.4362698959465698, "learning_rate": 1.2669397482542252e-05, "loss": 0.7442, "step": 11916 }, { "epoch": 0.43161897863093085, "grad_norm": 1.513504524824296, "learning_rate": 1.2668266963310898e-05, "loss": 0.8315, "step": 11917 }, { "epoch": 0.43165519739224917, "grad_norm": 1.3583277979459583, "learning_rate": 1.2667136407361844e-05, "loss": 0.7481, "step": 11918 }, { "epoch": 0.43169141615356754, "grad_norm": 1.2784601503709425, "learning_rate": 1.2666005814710652e-05, "loss": 0.7362, "step": 11919 }, { "epoch": 0.4317276349148859, "grad_norm": 1.4864898047315709, "learning_rate": 1.2664875185372875e-05, "loss": 0.8443, "step": 11920 }, { "epoch": 0.4317638536762043, "grad_norm": 0.9485274186923823, "learning_rate": 1.2663744519364073e-05, "loss": 0.656, "step": 11921 }, { "epoch": 0.43180007243752266, "grad_norm": 1.424547170201956, "learning_rate": 1.2662613816699812e-05, "loss": 0.7927, "step": 11922 }, { "epoch": 0.431836291198841, "grad_norm": 1.5137764347018179, "learning_rate": 1.2661483077395639e-05, "loss": 0.7707, "step": 11923 }, { "epoch": 0.43187250996015936, "grad_norm": 1.497486855325022, "learning_rate": 1.2660352301467126e-05, "loss": 0.7463, "step": 11924 }, { "epoch": 0.43190872872147773, "grad_norm": 1.3667675197498304, "learning_rate": 1.2659221488929823e-05, "loss": 0.7287, "step": 11925 }, { "epoch": 0.4319449474827961, "grad_norm": 1.0209735151126758, "learning_rate": 1.2658090639799297e-05, "loss": 0.7998, "step": 11926 }, { "epoch": 0.4319811662441145, "grad_norm": 1.5145824493013615, "learning_rate": 1.2656959754091107e-05, "loss": 0.8198, "step": 11927 }, { "epoch": 0.4320173850054328, "grad_norm": 1.3322008680766397, "learning_rate": 1.2655828831820815e-05, "loss": 0.7242, "step": 11928 }, { "epoch": 0.43205360376675117, "grad_norm": 1.3595052826592178, "learning_rate": 1.265469787300399e-05, "loss": 0.7174, "step": 11929 }, { "epoch": 0.43208982252806954, "grad_norm": 1.6632015909893598, "learning_rate": 1.2653566877656184e-05, "loss": 0.7777, "step": 11930 }, { "epoch": 0.4321260412893879, "grad_norm": 1.3117181316735524, "learning_rate": 1.2652435845792969e-05, "loss": 0.7005, "step": 11931 }, { "epoch": 0.4321622600507063, "grad_norm": 1.2473029828472852, "learning_rate": 1.2651304777429907e-05, "loss": 0.8125, "step": 11932 }, { "epoch": 0.4321984788120246, "grad_norm": 1.4564774972010186, "learning_rate": 1.2650173672582562e-05, "loss": 0.8022, "step": 11933 }, { "epoch": 0.432234697573343, "grad_norm": 1.3743382644297504, "learning_rate": 1.2649042531266495e-05, "loss": 0.7134, "step": 11934 }, { "epoch": 0.43227091633466136, "grad_norm": 1.3669873404440782, "learning_rate": 1.2647911353497277e-05, "loss": 0.7348, "step": 11935 }, { "epoch": 0.43230713509597973, "grad_norm": 1.4693381242939898, "learning_rate": 1.2646780139290473e-05, "loss": 0.7985, "step": 11936 }, { "epoch": 0.4323433538572981, "grad_norm": 1.2910828277133282, "learning_rate": 1.2645648888661648e-05, "loss": 0.6976, "step": 11937 }, { "epoch": 0.4323795726186164, "grad_norm": 1.3764744609755253, "learning_rate": 1.264451760162637e-05, "loss": 0.826, "step": 11938 }, { "epoch": 0.4324157913799348, "grad_norm": 1.4698954982555335, "learning_rate": 1.2643386278200204e-05, "loss": 0.7928, "step": 11939 }, { "epoch": 0.4324520101412532, "grad_norm": 1.233244845974775, "learning_rate": 1.2642254918398722e-05, "loss": 0.7625, "step": 11940 }, { "epoch": 0.43248822890257155, "grad_norm": 1.363500956023556, "learning_rate": 1.264112352223749e-05, "loss": 0.7592, "step": 11941 }, { "epoch": 0.4325244476638899, "grad_norm": 1.2667223204412663, "learning_rate": 1.2639992089732081e-05, "loss": 0.6737, "step": 11942 }, { "epoch": 0.43256066642520824, "grad_norm": 1.295982346624585, "learning_rate": 1.2638860620898057e-05, "loss": 0.7083, "step": 11943 }, { "epoch": 0.4325968851865266, "grad_norm": 1.276288443823741, "learning_rate": 1.2637729115750995e-05, "loss": 0.68, "step": 11944 }, { "epoch": 0.432633103947845, "grad_norm": 1.1421657582793123, "learning_rate": 1.263659757430646e-05, "loss": 0.7124, "step": 11945 }, { "epoch": 0.43266932270916336, "grad_norm": 1.0463318400960395, "learning_rate": 1.2635465996580028e-05, "loss": 0.7467, "step": 11946 }, { "epoch": 0.43270554147048174, "grad_norm": 1.4780494516897553, "learning_rate": 1.2634334382587264e-05, "loss": 0.7013, "step": 11947 }, { "epoch": 0.43274176023180005, "grad_norm": 1.6101569780498934, "learning_rate": 1.2633202732343746e-05, "loss": 0.7802, "step": 11948 }, { "epoch": 0.43277797899311843, "grad_norm": 1.3848911545946907, "learning_rate": 1.2632071045865048e-05, "loss": 0.7312, "step": 11949 }, { "epoch": 0.4328141977544368, "grad_norm": 1.3143156243289458, "learning_rate": 1.2630939323166735e-05, "loss": 0.7548, "step": 11950 }, { "epoch": 0.4328504165157552, "grad_norm": 1.4085872591813808, "learning_rate": 1.2629807564264389e-05, "loss": 0.7157, "step": 11951 }, { "epoch": 0.43288663527707355, "grad_norm": 1.1686972573112975, "learning_rate": 1.262867576917358e-05, "loss": 0.7021, "step": 11952 }, { "epoch": 0.43292285403839187, "grad_norm": 1.3895666409329432, "learning_rate": 1.2627543937909882e-05, "loss": 0.7333, "step": 11953 }, { "epoch": 0.43295907279971024, "grad_norm": 1.3064919438468223, "learning_rate": 1.262641207048887e-05, "loss": 0.7518, "step": 11954 }, { "epoch": 0.4329952915610286, "grad_norm": 1.3276287028108222, "learning_rate": 1.2625280166926119e-05, "loss": 0.8369, "step": 11955 }, { "epoch": 0.433031510322347, "grad_norm": 1.52861435959067, "learning_rate": 1.262414822723721e-05, "loss": 0.8495, "step": 11956 }, { "epoch": 0.43306772908366536, "grad_norm": 1.4540250564646937, "learning_rate": 1.262301625143771e-05, "loss": 0.7216, "step": 11957 }, { "epoch": 0.4331039478449837, "grad_norm": 1.4249391049533373, "learning_rate": 1.2621884239543208e-05, "loss": 0.7631, "step": 11958 }, { "epoch": 0.43314016660630206, "grad_norm": 1.3961584931385953, "learning_rate": 1.2620752191569267e-05, "loss": 0.7972, "step": 11959 }, { "epoch": 0.43317638536762043, "grad_norm": 1.386253958029768, "learning_rate": 1.261962010753148e-05, "loss": 0.774, "step": 11960 }, { "epoch": 0.4332126041289388, "grad_norm": 1.4461423398947726, "learning_rate": 1.2618487987445417e-05, "loss": 0.7502, "step": 11961 }, { "epoch": 0.4332488228902572, "grad_norm": 1.0234033883985638, "learning_rate": 1.2617355831326657e-05, "loss": 0.7816, "step": 11962 }, { "epoch": 0.4332850416515755, "grad_norm": 1.6531164340185345, "learning_rate": 1.2616223639190782e-05, "loss": 0.7862, "step": 11963 }, { "epoch": 0.43332126041289387, "grad_norm": 1.433042090085153, "learning_rate": 1.2615091411053369e-05, "loss": 0.8736, "step": 11964 }, { "epoch": 0.43335747917421225, "grad_norm": 1.8135224697565775, "learning_rate": 1.2613959146930004e-05, "loss": 0.7561, "step": 11965 }, { "epoch": 0.4333936979355306, "grad_norm": 1.4594469881783811, "learning_rate": 1.261282684683626e-05, "loss": 0.7043, "step": 11966 }, { "epoch": 0.433429916696849, "grad_norm": 1.3148341246681563, "learning_rate": 1.2611694510787727e-05, "loss": 0.7228, "step": 11967 }, { "epoch": 0.4334661354581673, "grad_norm": 1.3925588534226823, "learning_rate": 1.2610562138799977e-05, "loss": 0.7402, "step": 11968 }, { "epoch": 0.4335023542194857, "grad_norm": 1.3673622925603621, "learning_rate": 1.2609429730888605e-05, "loss": 0.7527, "step": 11969 }, { "epoch": 0.43353857298080406, "grad_norm": 1.382055849429995, "learning_rate": 1.2608297287069182e-05, "loss": 0.7493, "step": 11970 }, { "epoch": 0.43357479174212243, "grad_norm": 1.5412177985496311, "learning_rate": 1.2607164807357299e-05, "loss": 0.7954, "step": 11971 }, { "epoch": 0.4336110105034408, "grad_norm": 0.974760278276531, "learning_rate": 1.2606032291768537e-05, "loss": 0.6471, "step": 11972 }, { "epoch": 0.4336472292647591, "grad_norm": 1.4558506956422599, "learning_rate": 1.260489974031848e-05, "loss": 0.7237, "step": 11973 }, { "epoch": 0.4336834480260775, "grad_norm": 1.4108561096024765, "learning_rate": 1.2603767153022715e-05, "loss": 0.8367, "step": 11974 }, { "epoch": 0.4337196667873959, "grad_norm": 1.377948228939762, "learning_rate": 1.2602634529896824e-05, "loss": 0.7451, "step": 11975 }, { "epoch": 0.43375588554871425, "grad_norm": 1.2739628610591587, "learning_rate": 1.2601501870956396e-05, "loss": 0.7737, "step": 11976 }, { "epoch": 0.4337921043100326, "grad_norm": 1.4947489989623153, "learning_rate": 1.2600369176217018e-05, "loss": 0.7036, "step": 11977 }, { "epoch": 0.43382832307135094, "grad_norm": 1.7159089297325256, "learning_rate": 1.2599236445694274e-05, "loss": 0.7489, "step": 11978 }, { "epoch": 0.4338645418326693, "grad_norm": 1.9235182594896993, "learning_rate": 1.259810367940375e-05, "loss": 0.8308, "step": 11979 }, { "epoch": 0.4339007605939877, "grad_norm": 1.2983726893269736, "learning_rate": 1.2596970877361041e-05, "loss": 0.7098, "step": 11980 }, { "epoch": 0.43393697935530606, "grad_norm": 1.4189931916562228, "learning_rate": 1.2595838039581729e-05, "loss": 0.8288, "step": 11981 }, { "epoch": 0.43397319811662444, "grad_norm": 1.2744155043378254, "learning_rate": 1.2594705166081404e-05, "loss": 0.7377, "step": 11982 }, { "epoch": 0.43400941687794276, "grad_norm": 1.3229417713260059, "learning_rate": 1.2593572256875656e-05, "loss": 0.7533, "step": 11983 }, { "epoch": 0.43404563563926113, "grad_norm": 1.198736077643206, "learning_rate": 1.2592439311980078e-05, "loss": 0.8392, "step": 11984 }, { "epoch": 0.4340818544005795, "grad_norm": 1.5309328850503148, "learning_rate": 1.2591306331410254e-05, "loss": 0.7352, "step": 11985 }, { "epoch": 0.4341180731618979, "grad_norm": 1.4193705494582876, "learning_rate": 1.2590173315181776e-05, "loss": 0.7541, "step": 11986 }, { "epoch": 0.43415429192321625, "grad_norm": 1.391847980833975, "learning_rate": 1.2589040263310241e-05, "loss": 0.7674, "step": 11987 }, { "epoch": 0.43419051068453457, "grad_norm": 1.2661547435204803, "learning_rate": 1.2587907175811234e-05, "loss": 0.7042, "step": 11988 }, { "epoch": 0.43422672944585294, "grad_norm": 1.1466746930808094, "learning_rate": 1.2586774052700354e-05, "loss": 0.7762, "step": 11989 }, { "epoch": 0.4342629482071713, "grad_norm": 1.373371228397437, "learning_rate": 1.2585640893993187e-05, "loss": 0.7163, "step": 11990 }, { "epoch": 0.4342991669684897, "grad_norm": 1.0817759823641198, "learning_rate": 1.2584507699705332e-05, "loss": 0.7178, "step": 11991 }, { "epoch": 0.43433538572980807, "grad_norm": 1.2227840440587028, "learning_rate": 1.2583374469852379e-05, "loss": 0.7208, "step": 11992 }, { "epoch": 0.4343716044911264, "grad_norm": 1.4883750049967182, "learning_rate": 1.2582241204449923e-05, "loss": 0.8679, "step": 11993 }, { "epoch": 0.43440782325244476, "grad_norm": 1.1168343750360787, "learning_rate": 1.2581107903513558e-05, "loss": 0.7095, "step": 11994 }, { "epoch": 0.43444404201376313, "grad_norm": 1.4645959695426478, "learning_rate": 1.2579974567058879e-05, "loss": 0.7163, "step": 11995 }, { "epoch": 0.4344802607750815, "grad_norm": 1.359625721287227, "learning_rate": 1.2578841195101487e-05, "loss": 0.7905, "step": 11996 }, { "epoch": 0.4345164795363999, "grad_norm": 1.3197271450358152, "learning_rate": 1.2577707787656972e-05, "loss": 0.7681, "step": 11997 }, { "epoch": 0.4345526982977182, "grad_norm": 1.3479438350497845, "learning_rate": 1.2576574344740935e-05, "loss": 0.7406, "step": 11998 }, { "epoch": 0.4345889170590366, "grad_norm": 1.3998540890602775, "learning_rate": 1.2575440866368969e-05, "loss": 0.7369, "step": 11999 }, { "epoch": 0.43462513582035495, "grad_norm": 1.2656138163531647, "learning_rate": 1.2574307352556675e-05, "loss": 0.7488, "step": 12000 }, { "epoch": 0.4346613545816733, "grad_norm": 1.4116609271514606, "learning_rate": 1.2573173803319647e-05, "loss": 0.8061, "step": 12001 }, { "epoch": 0.4346975733429917, "grad_norm": 1.3805553190551518, "learning_rate": 1.2572040218673492e-05, "loss": 0.7507, "step": 12002 }, { "epoch": 0.43473379210431, "grad_norm": 1.4194849226114925, "learning_rate": 1.2570906598633798e-05, "loss": 0.8655, "step": 12003 }, { "epoch": 0.4347700108656284, "grad_norm": 1.1970048142390235, "learning_rate": 1.2569772943216174e-05, "loss": 0.7866, "step": 12004 }, { "epoch": 0.43480622962694676, "grad_norm": 1.2616885915346545, "learning_rate": 1.2568639252436212e-05, "loss": 0.6788, "step": 12005 }, { "epoch": 0.43484244838826513, "grad_norm": 1.3786908459661915, "learning_rate": 1.256750552630952e-05, "loss": 0.7418, "step": 12006 }, { "epoch": 0.4348786671495835, "grad_norm": 1.6371180125904976, "learning_rate": 1.2566371764851694e-05, "loss": 0.8499, "step": 12007 }, { "epoch": 0.4349148859109018, "grad_norm": 1.530965861407997, "learning_rate": 1.2565237968078339e-05, "loss": 0.7919, "step": 12008 }, { "epoch": 0.4349511046722202, "grad_norm": 1.3625227996248461, "learning_rate": 1.2564104136005058e-05, "loss": 0.7946, "step": 12009 }, { "epoch": 0.4349873234335386, "grad_norm": 1.578846755787332, "learning_rate": 1.2562970268647447e-05, "loss": 0.772, "step": 12010 }, { "epoch": 0.43502354219485695, "grad_norm": 1.2385403759856461, "learning_rate": 1.2561836366021116e-05, "loss": 0.7389, "step": 12011 }, { "epoch": 0.4350597609561753, "grad_norm": 1.648389741015909, "learning_rate": 1.2560702428141662e-05, "loss": 0.8213, "step": 12012 }, { "epoch": 0.43509597971749364, "grad_norm": 1.4745900323369712, "learning_rate": 1.2559568455024698e-05, "loss": 0.799, "step": 12013 }, { "epoch": 0.435132198478812, "grad_norm": 1.6292640166786567, "learning_rate": 1.2558434446685816e-05, "loss": 0.7525, "step": 12014 }, { "epoch": 0.4351684172401304, "grad_norm": 1.2221449236577429, "learning_rate": 1.2557300403140636e-05, "loss": 0.7783, "step": 12015 }, { "epoch": 0.43520463600144876, "grad_norm": 1.178015214071683, "learning_rate": 1.2556166324404747e-05, "loss": 0.7669, "step": 12016 }, { "epoch": 0.43524085476276714, "grad_norm": 1.2732667684193792, "learning_rate": 1.2555032210493768e-05, "loss": 0.674, "step": 12017 }, { "epoch": 0.43527707352408546, "grad_norm": 1.2962397016076634, "learning_rate": 1.2553898061423302e-05, "loss": 0.7109, "step": 12018 }, { "epoch": 0.43531329228540383, "grad_norm": 1.3638680166760684, "learning_rate": 1.2552763877208952e-05, "loss": 0.8165, "step": 12019 }, { "epoch": 0.4353495110467222, "grad_norm": 1.3124319361170433, "learning_rate": 1.2551629657866329e-05, "loss": 0.7095, "step": 12020 }, { "epoch": 0.4353857298080406, "grad_norm": 1.0972964722872016, "learning_rate": 1.2550495403411038e-05, "loss": 0.7662, "step": 12021 }, { "epoch": 0.43542194856935895, "grad_norm": 1.0879447669818039, "learning_rate": 1.2549361113858693e-05, "loss": 0.7875, "step": 12022 }, { "epoch": 0.43545816733067727, "grad_norm": 1.344150085826106, "learning_rate": 1.2548226789224896e-05, "loss": 0.7405, "step": 12023 }, { "epoch": 0.43549438609199564, "grad_norm": 1.275766791488352, "learning_rate": 1.2547092429525261e-05, "loss": 0.768, "step": 12024 }, { "epoch": 0.435530604853314, "grad_norm": 1.2632871857387489, "learning_rate": 1.2545958034775393e-05, "loss": 0.7747, "step": 12025 }, { "epoch": 0.4355668236146324, "grad_norm": 1.3631461658609818, "learning_rate": 1.2544823604990908e-05, "loss": 0.7083, "step": 12026 }, { "epoch": 0.43560304237595077, "grad_norm": 1.3972446559519534, "learning_rate": 1.2543689140187417e-05, "loss": 0.8011, "step": 12027 }, { "epoch": 0.4356392611372691, "grad_norm": 1.323602323334341, "learning_rate": 1.2542554640380522e-05, "loss": 0.7503, "step": 12028 }, { "epoch": 0.43567547989858746, "grad_norm": 1.021510397509404, "learning_rate": 1.2541420105585846e-05, "loss": 0.7964, "step": 12029 }, { "epoch": 0.43571169865990583, "grad_norm": 1.1060414346272796, "learning_rate": 1.2540285535818992e-05, "loss": 0.6887, "step": 12030 }, { "epoch": 0.4357479174212242, "grad_norm": 1.400557950263792, "learning_rate": 1.253915093109558e-05, "loss": 0.7737, "step": 12031 }, { "epoch": 0.4357841361825426, "grad_norm": 1.400315346550001, "learning_rate": 1.253801629143122e-05, "loss": 0.7809, "step": 12032 }, { "epoch": 0.4358203549438609, "grad_norm": 1.3882003941493792, "learning_rate": 1.2536881616841527e-05, "loss": 0.7803, "step": 12033 }, { "epoch": 0.4358565737051793, "grad_norm": 1.4026931291612232, "learning_rate": 1.253574690734211e-05, "loss": 0.774, "step": 12034 }, { "epoch": 0.43589279246649765, "grad_norm": 1.4247203832031152, "learning_rate": 1.2534612162948591e-05, "loss": 0.7464, "step": 12035 }, { "epoch": 0.435929011227816, "grad_norm": 1.4663029574382125, "learning_rate": 1.253347738367658e-05, "loss": 0.7347, "step": 12036 }, { "epoch": 0.4359652299891344, "grad_norm": 1.3623467503825375, "learning_rate": 1.2532342569541691e-05, "loss": 0.8033, "step": 12037 }, { "epoch": 0.4360014487504527, "grad_norm": 1.321748820582988, "learning_rate": 1.2531207720559548e-05, "loss": 0.8095, "step": 12038 }, { "epoch": 0.4360376675117711, "grad_norm": 1.4927615726612833, "learning_rate": 1.2530072836745762e-05, "loss": 0.6976, "step": 12039 }, { "epoch": 0.43607388627308946, "grad_norm": 1.2871893957730247, "learning_rate": 1.2528937918115948e-05, "loss": 0.7522, "step": 12040 }, { "epoch": 0.43611010503440784, "grad_norm": 1.4180676002831045, "learning_rate": 1.252780296468573e-05, "loss": 0.772, "step": 12041 }, { "epoch": 0.4361463237957262, "grad_norm": 1.2297596234396044, "learning_rate": 1.252666797647072e-05, "loss": 0.7407, "step": 12042 }, { "epoch": 0.43618254255704453, "grad_norm": 1.4889377988094359, "learning_rate": 1.2525532953486537e-05, "loss": 0.797, "step": 12043 }, { "epoch": 0.4362187613183629, "grad_norm": 1.3898923054072871, "learning_rate": 1.2524397895748804e-05, "loss": 0.7718, "step": 12044 }, { "epoch": 0.4362549800796813, "grad_norm": 1.4874923491231118, "learning_rate": 1.2523262803273135e-05, "loss": 0.8352, "step": 12045 }, { "epoch": 0.43629119884099965, "grad_norm": 1.3159347943590913, "learning_rate": 1.2522127676075155e-05, "loss": 0.6699, "step": 12046 }, { "epoch": 0.436327417602318, "grad_norm": 1.6104150838220088, "learning_rate": 1.252099251417048e-05, "loss": 0.7829, "step": 12047 }, { "epoch": 0.43636363636363634, "grad_norm": 1.5084588760503586, "learning_rate": 1.2519857317574733e-05, "loss": 0.7917, "step": 12048 }, { "epoch": 0.4363998551249547, "grad_norm": 1.4236594196339518, "learning_rate": 1.251872208630354e-05, "loss": 0.8157, "step": 12049 }, { "epoch": 0.4364360738862731, "grad_norm": 1.5256841428436227, "learning_rate": 1.2517586820372511e-05, "loss": 0.8211, "step": 12050 }, { "epoch": 0.43647229264759146, "grad_norm": 1.2918652322895428, "learning_rate": 1.251645151979728e-05, "loss": 0.76, "step": 12051 }, { "epoch": 0.43650851140890984, "grad_norm": 1.0999157795868246, "learning_rate": 1.2515316184593462e-05, "loss": 0.7393, "step": 12052 }, { "epoch": 0.43654473017022816, "grad_norm": 1.1395039119459247, "learning_rate": 1.2514180814776686e-05, "loss": 0.7625, "step": 12053 }, { "epoch": 0.43658094893154653, "grad_norm": 1.484992684292514, "learning_rate": 1.251304541036257e-05, "loss": 0.7614, "step": 12054 }, { "epoch": 0.4366171676928649, "grad_norm": 0.9371910941677961, "learning_rate": 1.2511909971366744e-05, "loss": 0.7312, "step": 12055 }, { "epoch": 0.4366533864541833, "grad_norm": 1.3053007540540782, "learning_rate": 1.2510774497804827e-05, "loss": 0.7682, "step": 12056 }, { "epoch": 0.43668960521550165, "grad_norm": 1.45924229301853, "learning_rate": 1.2509638989692447e-05, "loss": 0.723, "step": 12057 }, { "epoch": 0.43672582397681997, "grad_norm": 1.4270483304752364, "learning_rate": 1.250850344704523e-05, "loss": 0.7895, "step": 12058 }, { "epoch": 0.43676204273813835, "grad_norm": 1.4611626237414144, "learning_rate": 1.2507367869878804e-05, "loss": 0.7778, "step": 12059 }, { "epoch": 0.4367982614994567, "grad_norm": 1.2711964660693875, "learning_rate": 1.2506232258208793e-05, "loss": 0.7563, "step": 12060 }, { "epoch": 0.4368344802607751, "grad_norm": 1.3137570801302774, "learning_rate": 1.250509661205082e-05, "loss": 0.7027, "step": 12061 }, { "epoch": 0.43687069902209347, "grad_norm": 1.3229646701407427, "learning_rate": 1.2503960931420519e-05, "loss": 0.6754, "step": 12062 }, { "epoch": 0.4369069177834118, "grad_norm": 1.4575881762929102, "learning_rate": 1.2502825216333512e-05, "loss": 0.7604, "step": 12063 }, { "epoch": 0.43694313654473016, "grad_norm": 1.2458702016926828, "learning_rate": 1.2501689466805433e-05, "loss": 0.7533, "step": 12064 }, { "epoch": 0.43697935530604853, "grad_norm": 1.50998072923946, "learning_rate": 1.250055368285191e-05, "loss": 0.7914, "step": 12065 }, { "epoch": 0.4370155740673669, "grad_norm": 1.3779420041048178, "learning_rate": 1.249941786448857e-05, "loss": 0.7304, "step": 12066 }, { "epoch": 0.4370517928286853, "grad_norm": 1.298845047231015, "learning_rate": 1.2498282011731045e-05, "loss": 0.7763, "step": 12067 }, { "epoch": 0.4370880115900036, "grad_norm": 1.388392939499663, "learning_rate": 1.249714612459496e-05, "loss": 0.8071, "step": 12068 }, { "epoch": 0.437124230351322, "grad_norm": 1.284777230314436, "learning_rate": 1.2496010203095955e-05, "loss": 0.7623, "step": 12069 }, { "epoch": 0.43716044911264035, "grad_norm": 1.3995003601273683, "learning_rate": 1.2494874247249656e-05, "loss": 0.7653, "step": 12070 }, { "epoch": 0.4371966678739587, "grad_norm": 1.3632864562593752, "learning_rate": 1.2493738257071695e-05, "loss": 0.7626, "step": 12071 }, { "epoch": 0.4372328866352771, "grad_norm": 1.4038516865626034, "learning_rate": 1.2492602232577702e-05, "loss": 0.805, "step": 12072 }, { "epoch": 0.4372691053965954, "grad_norm": 1.2974954924661328, "learning_rate": 1.2491466173783313e-05, "loss": 0.8133, "step": 12073 }, { "epoch": 0.4373053241579138, "grad_norm": 1.4277852847759016, "learning_rate": 1.2490330080704163e-05, "loss": 0.7974, "step": 12074 }, { "epoch": 0.43734154291923216, "grad_norm": 1.4149605983991422, "learning_rate": 1.2489193953355879e-05, "loss": 0.7986, "step": 12075 }, { "epoch": 0.43737776168055054, "grad_norm": 1.3379955988445051, "learning_rate": 1.2488057791754102e-05, "loss": 0.6599, "step": 12076 }, { "epoch": 0.4374139804418689, "grad_norm": 1.4238986768367305, "learning_rate": 1.2486921595914461e-05, "loss": 0.7253, "step": 12077 }, { "epoch": 0.43745019920318723, "grad_norm": 1.2558720915819452, "learning_rate": 1.2485785365852596e-05, "loss": 0.8182, "step": 12078 }, { "epoch": 0.4374864179645056, "grad_norm": 1.5357670491018327, "learning_rate": 1.248464910158414e-05, "loss": 0.7974, "step": 12079 }, { "epoch": 0.437522636725824, "grad_norm": 1.5220405115563018, "learning_rate": 1.2483512803124729e-05, "loss": 0.7817, "step": 12080 }, { "epoch": 0.43755885548714235, "grad_norm": 1.2968706765914284, "learning_rate": 1.2482376470489996e-05, "loss": 0.7603, "step": 12081 }, { "epoch": 0.4375950742484607, "grad_norm": 1.1769986842242008, "learning_rate": 1.2481240103695587e-05, "loss": 0.7563, "step": 12082 }, { "epoch": 0.43763129300977904, "grad_norm": 1.355181085667294, "learning_rate": 1.248010370275713e-05, "loss": 0.7269, "step": 12083 }, { "epoch": 0.4376675117710974, "grad_norm": 0.9940000600498489, "learning_rate": 1.2478967267690267e-05, "loss": 0.7096, "step": 12084 }, { "epoch": 0.4377037305324158, "grad_norm": 1.6674087307915095, "learning_rate": 1.247783079851064e-05, "loss": 0.7842, "step": 12085 }, { "epoch": 0.43773994929373417, "grad_norm": 1.3755099742073902, "learning_rate": 1.247669429523388e-05, "loss": 0.7837, "step": 12086 }, { "epoch": 0.43777616805505254, "grad_norm": 1.1296926740670332, "learning_rate": 1.2475557757875632e-05, "loss": 0.6523, "step": 12087 }, { "epoch": 0.43781238681637086, "grad_norm": 0.9416943695117961, "learning_rate": 1.2474421186451533e-05, "loss": 0.6175, "step": 12088 }, { "epoch": 0.43784860557768923, "grad_norm": 1.378578136190943, "learning_rate": 1.2473284580977224e-05, "loss": 0.7593, "step": 12089 }, { "epoch": 0.4378848243390076, "grad_norm": 2.169386491214801, "learning_rate": 1.247214794146835e-05, "loss": 0.7873, "step": 12090 }, { "epoch": 0.437921043100326, "grad_norm": 1.6612526627799824, "learning_rate": 1.2471011267940545e-05, "loss": 0.8043, "step": 12091 }, { "epoch": 0.43795726186164435, "grad_norm": 1.0639847631130417, "learning_rate": 1.2469874560409452e-05, "loss": 0.8055, "step": 12092 }, { "epoch": 0.43799348062296267, "grad_norm": 1.510172833432053, "learning_rate": 1.2468737818890717e-05, "loss": 0.693, "step": 12093 }, { "epoch": 0.43802969938428105, "grad_norm": 1.3611099117529486, "learning_rate": 1.246760104339998e-05, "loss": 0.7673, "step": 12094 }, { "epoch": 0.4380659181455994, "grad_norm": 1.3509873423271055, "learning_rate": 1.2466464233952883e-05, "loss": 0.7944, "step": 12095 }, { "epoch": 0.4381021369069178, "grad_norm": 1.4120933372198867, "learning_rate": 1.2465327390565074e-05, "loss": 0.7267, "step": 12096 }, { "epoch": 0.43813835566823617, "grad_norm": 1.4234856349289258, "learning_rate": 1.2464190513252192e-05, "loss": 0.7856, "step": 12097 }, { "epoch": 0.4381745744295545, "grad_norm": 1.5244231972185496, "learning_rate": 1.2463053602029884e-05, "loss": 0.8131, "step": 12098 }, { "epoch": 0.43821079319087286, "grad_norm": 1.2949143448347473, "learning_rate": 1.2461916656913792e-05, "loss": 0.7093, "step": 12099 }, { "epoch": 0.43824701195219123, "grad_norm": 1.435391642812485, "learning_rate": 1.2460779677919567e-05, "loss": 0.7945, "step": 12100 }, { "epoch": 0.4382832307135096, "grad_norm": 1.5953276436411081, "learning_rate": 1.245964266506285e-05, "loss": 0.7188, "step": 12101 }, { "epoch": 0.438319449474828, "grad_norm": 1.2361981491569873, "learning_rate": 1.245850561835929e-05, "loss": 0.7084, "step": 12102 }, { "epoch": 0.4383556682361463, "grad_norm": 1.3777584737363284, "learning_rate": 1.2457368537824528e-05, "loss": 0.7803, "step": 12103 }, { "epoch": 0.4383918869974647, "grad_norm": 0.9683315463648245, "learning_rate": 1.245623142347422e-05, "loss": 0.7506, "step": 12104 }, { "epoch": 0.43842810575878305, "grad_norm": 1.4249964905892207, "learning_rate": 1.245509427532401e-05, "loss": 0.7911, "step": 12105 }, { "epoch": 0.4384643245201014, "grad_norm": 1.4825127627790715, "learning_rate": 1.2453957093389542e-05, "loss": 0.7815, "step": 12106 }, { "epoch": 0.4385005432814198, "grad_norm": 1.3683732612769917, "learning_rate": 1.2452819877686469e-05, "loss": 0.7509, "step": 12107 }, { "epoch": 0.4385367620427381, "grad_norm": 1.4476905500112813, "learning_rate": 1.245168262823044e-05, "loss": 0.8418, "step": 12108 }, { "epoch": 0.4385729808040565, "grad_norm": 1.3463146719190962, "learning_rate": 1.2450545345037104e-05, "loss": 0.7595, "step": 12109 }, { "epoch": 0.43860919956537486, "grad_norm": 1.306931916513502, "learning_rate": 1.244940802812211e-05, "loss": 0.6556, "step": 12110 }, { "epoch": 0.43864541832669324, "grad_norm": 1.0931442059369185, "learning_rate": 1.244827067750111e-05, "loss": 0.6755, "step": 12111 }, { "epoch": 0.4386816370880116, "grad_norm": 1.5102384851989892, "learning_rate": 1.2447133293189753e-05, "loss": 0.8229, "step": 12112 }, { "epoch": 0.43871785584932993, "grad_norm": 1.381423538189575, "learning_rate": 1.2445995875203693e-05, "loss": 0.7211, "step": 12113 }, { "epoch": 0.4387540746106483, "grad_norm": 1.3601212974145958, "learning_rate": 1.2444858423558581e-05, "loss": 0.7387, "step": 12114 }, { "epoch": 0.4387902933719667, "grad_norm": 1.4272355044347629, "learning_rate": 1.2443720938270067e-05, "loss": 0.7333, "step": 12115 }, { "epoch": 0.43882651213328505, "grad_norm": 1.3344627558665112, "learning_rate": 1.2442583419353808e-05, "loss": 0.745, "step": 12116 }, { "epoch": 0.4388627308946034, "grad_norm": 1.490816457616739, "learning_rate": 1.2441445866825452e-05, "loss": 0.723, "step": 12117 }, { "epoch": 0.43889894965592174, "grad_norm": 1.3772394964691312, "learning_rate": 1.244030828070066e-05, "loss": 0.7978, "step": 12118 }, { "epoch": 0.4389351684172401, "grad_norm": 1.4985543824991967, "learning_rate": 1.2439170660995077e-05, "loss": 0.7536, "step": 12119 }, { "epoch": 0.4389713871785585, "grad_norm": 1.4761475649878864, "learning_rate": 1.2438033007724367e-05, "loss": 0.8027, "step": 12120 }, { "epoch": 0.43900760593987687, "grad_norm": 1.419387862574719, "learning_rate": 1.2436895320904176e-05, "loss": 0.7567, "step": 12121 }, { "epoch": 0.43904382470119524, "grad_norm": 1.4375874310382415, "learning_rate": 1.2435757600550167e-05, "loss": 0.7037, "step": 12122 }, { "epoch": 0.43908004346251356, "grad_norm": 1.0902500028841555, "learning_rate": 1.2434619846677991e-05, "loss": 0.733, "step": 12123 }, { "epoch": 0.43911626222383193, "grad_norm": 1.3681213581900227, "learning_rate": 1.243348205930331e-05, "loss": 0.6815, "step": 12124 }, { "epoch": 0.4391524809851503, "grad_norm": 1.3458592894510875, "learning_rate": 1.2432344238441774e-05, "loss": 0.7789, "step": 12125 }, { "epoch": 0.4391886997464687, "grad_norm": 1.1209189298610718, "learning_rate": 1.2431206384109045e-05, "loss": 0.7614, "step": 12126 }, { "epoch": 0.43922491850778705, "grad_norm": 1.5110573073169844, "learning_rate": 1.2430068496320782e-05, "loss": 0.7482, "step": 12127 }, { "epoch": 0.4392611372691054, "grad_norm": 1.3416242752229621, "learning_rate": 1.242893057509264e-05, "loss": 0.7615, "step": 12128 }, { "epoch": 0.43929735603042375, "grad_norm": 1.2411284466669446, "learning_rate": 1.242779262044028e-05, "loss": 0.7445, "step": 12129 }, { "epoch": 0.4393335747917421, "grad_norm": 1.5423777498662359, "learning_rate": 1.2426654632379359e-05, "loss": 0.8382, "step": 12130 }, { "epoch": 0.4393697935530605, "grad_norm": 1.0791873112436619, "learning_rate": 1.2425516610925539e-05, "loss": 0.7021, "step": 12131 }, { "epoch": 0.43940601231437887, "grad_norm": 1.996254203453965, "learning_rate": 1.2424378556094476e-05, "loss": 0.7974, "step": 12132 }, { "epoch": 0.4394422310756972, "grad_norm": 1.476490861678724, "learning_rate": 1.2423240467901838e-05, "loss": 0.7719, "step": 12133 }, { "epoch": 0.43947844983701556, "grad_norm": 1.3937940358573833, "learning_rate": 1.2422102346363278e-05, "loss": 0.7668, "step": 12134 }, { "epoch": 0.43951466859833394, "grad_norm": 1.4406075991657454, "learning_rate": 1.2420964191494464e-05, "loss": 0.7704, "step": 12135 }, { "epoch": 0.4395508873596523, "grad_norm": 1.0625259851434126, "learning_rate": 1.2419826003311056e-05, "loss": 0.7159, "step": 12136 }, { "epoch": 0.4395871061209707, "grad_norm": 1.9324109862332906, "learning_rate": 1.2418687781828713e-05, "loss": 0.7868, "step": 12137 }, { "epoch": 0.439623324882289, "grad_norm": 1.1972978776306582, "learning_rate": 1.2417549527063106e-05, "loss": 0.7335, "step": 12138 }, { "epoch": 0.4396595436436074, "grad_norm": 1.6250329106005974, "learning_rate": 1.241641123902989e-05, "loss": 0.8145, "step": 12139 }, { "epoch": 0.43969576240492575, "grad_norm": 1.4757212716278907, "learning_rate": 1.2415272917744731e-05, "loss": 0.809, "step": 12140 }, { "epoch": 0.4397319811662441, "grad_norm": 1.4427789984678905, "learning_rate": 1.2414134563223294e-05, "loss": 0.6959, "step": 12141 }, { "epoch": 0.4397681999275625, "grad_norm": 1.043549293822376, "learning_rate": 1.241299617548125e-05, "loss": 0.804, "step": 12142 }, { "epoch": 0.4398044186888808, "grad_norm": 1.050011735614442, "learning_rate": 1.2411857754534254e-05, "loss": 0.8084, "step": 12143 }, { "epoch": 0.4398406374501992, "grad_norm": 1.2479914375314336, "learning_rate": 1.2410719300397975e-05, "loss": 0.7996, "step": 12144 }, { "epoch": 0.43987685621151756, "grad_norm": 1.3916787390473209, "learning_rate": 1.2409580813088082e-05, "loss": 0.6766, "step": 12145 }, { "epoch": 0.43991307497283594, "grad_norm": 1.3082533721067569, "learning_rate": 1.2408442292620237e-05, "loss": 0.7105, "step": 12146 }, { "epoch": 0.4399492937341543, "grad_norm": 1.4261497008418649, "learning_rate": 1.2407303739010113e-05, "loss": 0.7478, "step": 12147 }, { "epoch": 0.43998551249547263, "grad_norm": 1.375373740093939, "learning_rate": 1.2406165152273374e-05, "loss": 0.7469, "step": 12148 }, { "epoch": 0.440021731256791, "grad_norm": 1.380712461547375, "learning_rate": 1.2405026532425688e-05, "loss": 0.8392, "step": 12149 }, { "epoch": 0.4400579500181094, "grad_norm": 1.2083190216038688, "learning_rate": 1.2403887879482723e-05, "loss": 0.7002, "step": 12150 }, { "epoch": 0.44009416877942775, "grad_norm": 1.3891076693147588, "learning_rate": 1.2402749193460148e-05, "loss": 0.7425, "step": 12151 }, { "epoch": 0.4401303875407461, "grad_norm": 1.4263067100247413, "learning_rate": 1.2401610474373635e-05, "loss": 0.6982, "step": 12152 }, { "epoch": 0.44016660630206444, "grad_norm": 1.1024728553260803, "learning_rate": 1.2400471722238851e-05, "loss": 0.7854, "step": 12153 }, { "epoch": 0.4402028250633828, "grad_norm": 1.3908429189376068, "learning_rate": 1.2399332937071465e-05, "loss": 0.7329, "step": 12154 }, { "epoch": 0.4402390438247012, "grad_norm": 1.1824148778882442, "learning_rate": 1.2398194118887152e-05, "loss": 0.6944, "step": 12155 }, { "epoch": 0.44027526258601957, "grad_norm": 1.369628129710693, "learning_rate": 1.2397055267701579e-05, "loss": 0.7288, "step": 12156 }, { "epoch": 0.44031148134733794, "grad_norm": 1.6567093801408839, "learning_rate": 1.2395916383530418e-05, "loss": 0.7826, "step": 12157 }, { "epoch": 0.44034770010865626, "grad_norm": 1.307652274267147, "learning_rate": 1.2394777466389345e-05, "loss": 0.7633, "step": 12158 }, { "epoch": 0.44038391886997463, "grad_norm": 1.5660890819385223, "learning_rate": 1.2393638516294029e-05, "loss": 0.8145, "step": 12159 }, { "epoch": 0.440420137631293, "grad_norm": 1.4019433355746944, "learning_rate": 1.2392499533260145e-05, "loss": 0.697, "step": 12160 }, { "epoch": 0.4404563563926114, "grad_norm": 1.1155899204183017, "learning_rate": 1.2391360517303362e-05, "loss": 0.7671, "step": 12161 }, { "epoch": 0.44049257515392976, "grad_norm": 1.4252136145649417, "learning_rate": 1.239022146843936e-05, "loss": 0.7962, "step": 12162 }, { "epoch": 0.4405287939152481, "grad_norm": 1.508775719642281, "learning_rate": 1.2389082386683809e-05, "loss": 0.7702, "step": 12163 }, { "epoch": 0.44056501267656645, "grad_norm": 1.3558248002796471, "learning_rate": 1.2387943272052385e-05, "loss": 0.715, "step": 12164 }, { "epoch": 0.4406012314378848, "grad_norm": 1.347255010582401, "learning_rate": 1.2386804124560764e-05, "loss": 0.7795, "step": 12165 }, { "epoch": 0.4406374501992032, "grad_norm": 1.2652135889690208, "learning_rate": 1.2385664944224622e-05, "loss": 0.748, "step": 12166 }, { "epoch": 0.44067366896052157, "grad_norm": 1.4174887288661369, "learning_rate": 1.2384525731059634e-05, "loss": 0.7171, "step": 12167 }, { "epoch": 0.4407098877218399, "grad_norm": 1.042679689587954, "learning_rate": 1.2383386485081475e-05, "loss": 0.6946, "step": 12168 }, { "epoch": 0.44074610648315826, "grad_norm": 1.0604501121320773, "learning_rate": 1.2382247206305825e-05, "loss": 0.7346, "step": 12169 }, { "epoch": 0.44078232524447664, "grad_norm": 1.1882487828025954, "learning_rate": 1.238110789474836e-05, "loss": 0.788, "step": 12170 }, { "epoch": 0.440818544005795, "grad_norm": 1.6109222721726926, "learning_rate": 1.2379968550424756e-05, "loss": 0.835, "step": 12171 }, { "epoch": 0.4408547627671134, "grad_norm": 1.2059887466473753, "learning_rate": 1.2378829173350695e-05, "loss": 0.7564, "step": 12172 }, { "epoch": 0.4408909815284317, "grad_norm": 1.361685666236653, "learning_rate": 1.2377689763541853e-05, "loss": 0.8363, "step": 12173 }, { "epoch": 0.4409272002897501, "grad_norm": 1.2857681588764949, "learning_rate": 1.2376550321013914e-05, "loss": 0.7529, "step": 12174 }, { "epoch": 0.44096341905106845, "grad_norm": 1.2953627110724477, "learning_rate": 1.2375410845782551e-05, "loss": 0.7639, "step": 12175 }, { "epoch": 0.4409996378123868, "grad_norm": 2.167373636547203, "learning_rate": 1.237427133786345e-05, "loss": 0.8609, "step": 12176 }, { "epoch": 0.4410358565737052, "grad_norm": 1.2298858014785723, "learning_rate": 1.2373131797272289e-05, "loss": 0.7915, "step": 12177 }, { "epoch": 0.4410720753350235, "grad_norm": 1.4640311314124088, "learning_rate": 1.2371992224024749e-05, "loss": 0.8092, "step": 12178 }, { "epoch": 0.4411082940963419, "grad_norm": 1.4811082379795264, "learning_rate": 1.237085261813651e-05, "loss": 0.713, "step": 12179 }, { "epoch": 0.44114451285766026, "grad_norm": 1.4987301952808976, "learning_rate": 1.2369712979623259e-05, "loss": 0.8169, "step": 12180 }, { "epoch": 0.44118073161897864, "grad_norm": 1.4500393712204593, "learning_rate": 1.2368573308500673e-05, "loss": 0.7491, "step": 12181 }, { "epoch": 0.441216950380297, "grad_norm": 1.7182779487521957, "learning_rate": 1.2367433604784438e-05, "loss": 0.8234, "step": 12182 }, { "epoch": 0.44125316914161533, "grad_norm": 1.1459640677547989, "learning_rate": 1.2366293868490236e-05, "loss": 0.7605, "step": 12183 }, { "epoch": 0.4412893879029337, "grad_norm": 1.052162627175721, "learning_rate": 1.2365154099633749e-05, "loss": 0.7367, "step": 12184 }, { "epoch": 0.4413256066642521, "grad_norm": 1.377414978974628, "learning_rate": 1.2364014298230667e-05, "loss": 0.8038, "step": 12185 }, { "epoch": 0.44136182542557045, "grad_norm": 1.3938832547277604, "learning_rate": 1.2362874464296669e-05, "loss": 0.7912, "step": 12186 }, { "epoch": 0.4413980441868888, "grad_norm": 1.422759027554995, "learning_rate": 1.2361734597847442e-05, "loss": 0.8468, "step": 12187 }, { "epoch": 0.44143426294820715, "grad_norm": 1.297583611916744, "learning_rate": 1.236059469889867e-05, "loss": 0.7337, "step": 12188 }, { "epoch": 0.4414704817095255, "grad_norm": 1.3815052297232928, "learning_rate": 1.2359454767466045e-05, "loss": 0.8015, "step": 12189 }, { "epoch": 0.4415067004708439, "grad_norm": 1.3786886829742258, "learning_rate": 1.2358314803565246e-05, "loss": 0.7934, "step": 12190 }, { "epoch": 0.44154291923216227, "grad_norm": 1.4190488490053133, "learning_rate": 1.2357174807211966e-05, "loss": 0.7594, "step": 12191 }, { "epoch": 0.44157913799348064, "grad_norm": 1.143875719527111, "learning_rate": 1.2356034778421887e-05, "loss": 0.7682, "step": 12192 }, { "epoch": 0.44161535675479896, "grad_norm": 1.2248690331213454, "learning_rate": 1.2354894717210699e-05, "loss": 0.7844, "step": 12193 }, { "epoch": 0.44165157551611733, "grad_norm": 1.130606954460246, "learning_rate": 1.2353754623594092e-05, "loss": 0.7596, "step": 12194 }, { "epoch": 0.4416877942774357, "grad_norm": 1.5577000706543533, "learning_rate": 1.235261449758775e-05, "loss": 0.8312, "step": 12195 }, { "epoch": 0.4417240130387541, "grad_norm": 1.4727184183876099, "learning_rate": 1.2351474339207372e-05, "loss": 0.752, "step": 12196 }, { "epoch": 0.44176023180007246, "grad_norm": 0.9907068267279618, "learning_rate": 1.2350334148468635e-05, "loss": 0.7545, "step": 12197 }, { "epoch": 0.4417964505613908, "grad_norm": 1.2441683945595885, "learning_rate": 1.234919392538724e-05, "loss": 0.7549, "step": 12198 }, { "epoch": 0.44183266932270915, "grad_norm": 1.630652124404842, "learning_rate": 1.2348053669978866e-05, "loss": 0.7645, "step": 12199 }, { "epoch": 0.4418688880840275, "grad_norm": 1.3169170550272007, "learning_rate": 1.2346913382259216e-05, "loss": 0.7142, "step": 12200 }, { "epoch": 0.4419051068453459, "grad_norm": 1.3566408002750583, "learning_rate": 1.2345773062243971e-05, "loss": 0.7239, "step": 12201 }, { "epoch": 0.44194132560666427, "grad_norm": 1.4001878980051048, "learning_rate": 1.2344632709948831e-05, "loss": 0.7179, "step": 12202 }, { "epoch": 0.4419775443679826, "grad_norm": 1.4782450162500995, "learning_rate": 1.2343492325389483e-05, "loss": 0.7729, "step": 12203 }, { "epoch": 0.44201376312930096, "grad_norm": 1.3726162323304771, "learning_rate": 1.2342351908581622e-05, "loss": 0.7703, "step": 12204 }, { "epoch": 0.44204998189061934, "grad_norm": 1.1131480621728547, "learning_rate": 1.234121145954094e-05, "loss": 0.8217, "step": 12205 }, { "epoch": 0.4420862006519377, "grad_norm": 1.2750611989735166, "learning_rate": 1.2340070978283132e-05, "loss": 0.7948, "step": 12206 }, { "epoch": 0.4421224194132561, "grad_norm": 1.514830850700713, "learning_rate": 1.233893046482389e-05, "loss": 0.7874, "step": 12207 }, { "epoch": 0.4421586381745744, "grad_norm": 1.2384343057387486, "learning_rate": 1.2337789919178911e-05, "loss": 0.7882, "step": 12208 }, { "epoch": 0.4421948569358928, "grad_norm": 1.510612673346164, "learning_rate": 1.2336649341363889e-05, "loss": 0.7383, "step": 12209 }, { "epoch": 0.44223107569721115, "grad_norm": 1.4911949123128554, "learning_rate": 1.2335508731394516e-05, "loss": 0.7404, "step": 12210 }, { "epoch": 0.4422672944585295, "grad_norm": 1.479175454407991, "learning_rate": 1.2334368089286494e-05, "loss": 0.7645, "step": 12211 }, { "epoch": 0.4423035132198479, "grad_norm": 1.5234198031253467, "learning_rate": 1.2333227415055515e-05, "loss": 0.8516, "step": 12212 }, { "epoch": 0.4423397319811662, "grad_norm": 1.4484585165095736, "learning_rate": 1.2332086708717277e-05, "loss": 0.8324, "step": 12213 }, { "epoch": 0.4423759507424846, "grad_norm": 1.3961413248897214, "learning_rate": 1.2330945970287473e-05, "loss": 0.6744, "step": 12214 }, { "epoch": 0.44241216950380297, "grad_norm": 1.432139391751797, "learning_rate": 1.2329805199781808e-05, "loss": 0.756, "step": 12215 }, { "epoch": 0.44244838826512134, "grad_norm": 1.4086036251492091, "learning_rate": 1.2328664397215976e-05, "loss": 0.8105, "step": 12216 }, { "epoch": 0.4424846070264397, "grad_norm": 1.4194863489699043, "learning_rate": 1.2327523562605675e-05, "loss": 0.8835, "step": 12217 }, { "epoch": 0.44252082578775803, "grad_norm": 1.0697519275712768, "learning_rate": 1.2326382695966605e-05, "loss": 0.782, "step": 12218 }, { "epoch": 0.4425570445490764, "grad_norm": 1.211400504196408, "learning_rate": 1.2325241797314465e-05, "loss": 0.6839, "step": 12219 }, { "epoch": 0.4425932633103948, "grad_norm": 1.435938610423077, "learning_rate": 1.2324100866664955e-05, "loss": 0.7196, "step": 12220 }, { "epoch": 0.44262948207171315, "grad_norm": 1.2542139197004571, "learning_rate": 1.2322959904033774e-05, "loss": 0.8128, "step": 12221 }, { "epoch": 0.44266570083303153, "grad_norm": 1.4551010616466074, "learning_rate": 1.2321818909436625e-05, "loss": 0.7649, "step": 12222 }, { "epoch": 0.44270191959434985, "grad_norm": 1.3924943288815175, "learning_rate": 1.2320677882889205e-05, "loss": 0.8055, "step": 12223 }, { "epoch": 0.4427381383556682, "grad_norm": 1.3049318568744248, "learning_rate": 1.231953682440722e-05, "loss": 0.7362, "step": 12224 }, { "epoch": 0.4427743571169866, "grad_norm": 1.422414081424539, "learning_rate": 1.2318395734006372e-05, "loss": 0.7605, "step": 12225 }, { "epoch": 0.44281057587830497, "grad_norm": 1.4473993848335214, "learning_rate": 1.2317254611702358e-05, "loss": 0.8118, "step": 12226 }, { "epoch": 0.44284679463962334, "grad_norm": 1.5551530186590725, "learning_rate": 1.2316113457510887e-05, "loss": 0.7356, "step": 12227 }, { "epoch": 0.44288301340094166, "grad_norm": 1.5081994197694226, "learning_rate": 1.2314972271447657e-05, "loss": 0.8158, "step": 12228 }, { "epoch": 0.44291923216226003, "grad_norm": 1.2751144530779412, "learning_rate": 1.2313831053528378e-05, "loss": 0.7044, "step": 12229 }, { "epoch": 0.4429554509235784, "grad_norm": 1.2981921503276104, "learning_rate": 1.2312689803768749e-05, "loss": 0.8006, "step": 12230 }, { "epoch": 0.4429916696848968, "grad_norm": 1.3481976250981018, "learning_rate": 1.2311548522184476e-05, "loss": 0.678, "step": 12231 }, { "epoch": 0.44302788844621516, "grad_norm": 1.5343245602395224, "learning_rate": 1.2310407208791264e-05, "loss": 0.8234, "step": 12232 }, { "epoch": 0.4430641072075335, "grad_norm": 1.474327013632406, "learning_rate": 1.230926586360482e-05, "loss": 0.7998, "step": 12233 }, { "epoch": 0.44310032596885185, "grad_norm": 1.3947474061043874, "learning_rate": 1.2308124486640847e-05, "loss": 0.7203, "step": 12234 }, { "epoch": 0.4431365447301702, "grad_norm": 1.412843657090759, "learning_rate": 1.2306983077915052e-05, "loss": 0.7325, "step": 12235 }, { "epoch": 0.4431727634914886, "grad_norm": 1.1247619546352907, "learning_rate": 1.2305841637443146e-05, "loss": 0.7499, "step": 12236 }, { "epoch": 0.44320898225280697, "grad_norm": 1.2580330504284596, "learning_rate": 1.230470016524083e-05, "loss": 0.716, "step": 12237 }, { "epoch": 0.4432452010141253, "grad_norm": 1.401572445714018, "learning_rate": 1.2303558661323818e-05, "loss": 0.729, "step": 12238 }, { "epoch": 0.44328141977544366, "grad_norm": 1.3394315783507855, "learning_rate": 1.230241712570781e-05, "loss": 0.7302, "step": 12239 }, { "epoch": 0.44331763853676204, "grad_norm": 1.2294176979012403, "learning_rate": 1.2301275558408526e-05, "loss": 0.7906, "step": 12240 }, { "epoch": 0.4433538572980804, "grad_norm": 1.0107684146431506, "learning_rate": 1.2300133959441663e-05, "loss": 0.7677, "step": 12241 }, { "epoch": 0.4433900760593988, "grad_norm": 0.9962533058674197, "learning_rate": 1.2298992328822937e-05, "loss": 0.7917, "step": 12242 }, { "epoch": 0.4434262948207171, "grad_norm": 1.2651529071851328, "learning_rate": 1.2297850666568054e-05, "loss": 0.7439, "step": 12243 }, { "epoch": 0.4434625135820355, "grad_norm": 1.259449983575965, "learning_rate": 1.2296708972692728e-05, "loss": 0.6881, "step": 12244 }, { "epoch": 0.44349873234335385, "grad_norm": 1.2860542800550903, "learning_rate": 1.2295567247212672e-05, "loss": 0.7405, "step": 12245 }, { "epoch": 0.4435349511046722, "grad_norm": 1.5564790328331015, "learning_rate": 1.2294425490143591e-05, "loss": 0.7649, "step": 12246 }, { "epoch": 0.4435711698659906, "grad_norm": 1.312023693040012, "learning_rate": 1.22932837015012e-05, "loss": 0.7109, "step": 12247 }, { "epoch": 0.4436073886273089, "grad_norm": 1.3643945619971467, "learning_rate": 1.2292141881301208e-05, "loss": 0.7417, "step": 12248 }, { "epoch": 0.4436436073886273, "grad_norm": 1.4133770200650497, "learning_rate": 1.2291000029559332e-05, "loss": 0.7266, "step": 12249 }, { "epoch": 0.44367982614994567, "grad_norm": 1.361704552653588, "learning_rate": 1.2289858146291283e-05, "loss": 0.7957, "step": 12250 }, { "epoch": 0.44371604491126404, "grad_norm": 1.1719434924772745, "learning_rate": 1.2288716231512772e-05, "loss": 0.7084, "step": 12251 }, { "epoch": 0.4437522636725824, "grad_norm": 1.3754373499683488, "learning_rate": 1.2287574285239513e-05, "loss": 0.7206, "step": 12252 }, { "epoch": 0.4437884824339008, "grad_norm": 1.3885641820813013, "learning_rate": 1.2286432307487226e-05, "loss": 0.7317, "step": 12253 }, { "epoch": 0.4438247011952191, "grad_norm": 1.4474676920360412, "learning_rate": 1.2285290298271619e-05, "loss": 0.7626, "step": 12254 }, { "epoch": 0.4438609199565375, "grad_norm": 1.3666419187243188, "learning_rate": 1.2284148257608411e-05, "loss": 0.6869, "step": 12255 }, { "epoch": 0.44389713871785585, "grad_norm": 1.4259684322236692, "learning_rate": 1.2283006185513315e-05, "loss": 0.8384, "step": 12256 }, { "epoch": 0.44393335747917423, "grad_norm": 1.355589099317864, "learning_rate": 1.2281864082002046e-05, "loss": 0.7205, "step": 12257 }, { "epoch": 0.4439695762404926, "grad_norm": 1.3826779176124537, "learning_rate": 1.2280721947090327e-05, "loss": 0.7159, "step": 12258 }, { "epoch": 0.4440057950018109, "grad_norm": 1.3044093736422577, "learning_rate": 1.2279579780793864e-05, "loss": 0.7094, "step": 12259 }, { "epoch": 0.4440420137631293, "grad_norm": 1.6179443030580338, "learning_rate": 1.2278437583128384e-05, "loss": 0.8074, "step": 12260 }, { "epoch": 0.44407823252444767, "grad_norm": 1.0228300418942584, "learning_rate": 1.22772953541096e-05, "loss": 0.702, "step": 12261 }, { "epoch": 0.44411445128576604, "grad_norm": 1.3741558657761823, "learning_rate": 1.2276153093753231e-05, "loss": 0.7086, "step": 12262 }, { "epoch": 0.4441506700470844, "grad_norm": 1.166317894538869, "learning_rate": 1.2275010802074996e-05, "loss": 0.7534, "step": 12263 }, { "epoch": 0.44418688880840274, "grad_norm": 1.3268250383816984, "learning_rate": 1.2273868479090614e-05, "loss": 0.7199, "step": 12264 }, { "epoch": 0.4442231075697211, "grad_norm": 1.3081184356661901, "learning_rate": 1.2272726124815802e-05, "loss": 0.7152, "step": 12265 }, { "epoch": 0.4442593263310395, "grad_norm": 1.2806325826285436, "learning_rate": 1.227158373926628e-05, "loss": 0.7187, "step": 12266 }, { "epoch": 0.44429554509235786, "grad_norm": 1.2545802178306444, "learning_rate": 1.2270441322457773e-05, "loss": 0.7482, "step": 12267 }, { "epoch": 0.44433176385367623, "grad_norm": 1.3913071832993107, "learning_rate": 1.2269298874405995e-05, "loss": 0.7579, "step": 12268 }, { "epoch": 0.44436798261499455, "grad_norm": 1.4300631240749835, "learning_rate": 1.2268156395126675e-05, "loss": 0.7173, "step": 12269 }, { "epoch": 0.4444042013763129, "grad_norm": 1.1726025838492014, "learning_rate": 1.2267013884635528e-05, "loss": 0.7135, "step": 12270 }, { "epoch": 0.4444404201376313, "grad_norm": 1.0733807947914604, "learning_rate": 1.2265871342948276e-05, "loss": 0.7489, "step": 12271 }, { "epoch": 0.44447663889894967, "grad_norm": 1.2809847522274065, "learning_rate": 1.2264728770080647e-05, "loss": 0.673, "step": 12272 }, { "epoch": 0.44451285766026805, "grad_norm": 1.3363020557056344, "learning_rate": 1.2263586166048359e-05, "loss": 0.7331, "step": 12273 }, { "epoch": 0.44454907642158636, "grad_norm": 1.33450406397329, "learning_rate": 1.2262443530867136e-05, "loss": 0.7119, "step": 12274 }, { "epoch": 0.44458529518290474, "grad_norm": 1.2453501404897465, "learning_rate": 1.2261300864552701e-05, "loss": 0.6673, "step": 12275 }, { "epoch": 0.4446215139442231, "grad_norm": 1.0333793674381013, "learning_rate": 1.226015816712078e-05, "loss": 0.7434, "step": 12276 }, { "epoch": 0.4446577327055415, "grad_norm": 1.3157209641689152, "learning_rate": 1.2259015438587096e-05, "loss": 0.8131, "step": 12277 }, { "epoch": 0.44469395146685986, "grad_norm": 1.453556959235483, "learning_rate": 1.2257872678967375e-05, "loss": 0.6934, "step": 12278 }, { "epoch": 0.4447301702281782, "grad_norm": 1.4269221842377322, "learning_rate": 1.2256729888277341e-05, "loss": 0.7793, "step": 12279 }, { "epoch": 0.44476638898949655, "grad_norm": 1.4024951920781792, "learning_rate": 1.2255587066532723e-05, "loss": 0.8082, "step": 12280 }, { "epoch": 0.4448026077508149, "grad_norm": 1.5865528605277048, "learning_rate": 1.2254444213749243e-05, "loss": 0.7966, "step": 12281 }, { "epoch": 0.4448388265121333, "grad_norm": 1.3096057146024989, "learning_rate": 1.2253301329942631e-05, "loss": 0.743, "step": 12282 }, { "epoch": 0.4448750452734517, "grad_norm": 1.3584598442365412, "learning_rate": 1.2252158415128612e-05, "loss": 0.7206, "step": 12283 }, { "epoch": 0.44491126403477, "grad_norm": 1.2746719521447905, "learning_rate": 1.2251015469322915e-05, "loss": 0.7527, "step": 12284 }, { "epoch": 0.44494748279608837, "grad_norm": 1.3031401481251113, "learning_rate": 1.2249872492541268e-05, "loss": 0.6907, "step": 12285 }, { "epoch": 0.44498370155740674, "grad_norm": 1.4210933702341488, "learning_rate": 1.2248729484799395e-05, "loss": 0.7322, "step": 12286 }, { "epoch": 0.4450199203187251, "grad_norm": 1.2893363315000228, "learning_rate": 1.2247586446113034e-05, "loss": 0.742, "step": 12287 }, { "epoch": 0.4450561390800435, "grad_norm": 1.313410362007652, "learning_rate": 1.2246443376497903e-05, "loss": 0.7789, "step": 12288 }, { "epoch": 0.4450923578413618, "grad_norm": 1.393403728785511, "learning_rate": 1.2245300275969742e-05, "loss": 0.7118, "step": 12289 }, { "epoch": 0.4451285766026802, "grad_norm": 1.2898552436593784, "learning_rate": 1.2244157144544273e-05, "loss": 0.7195, "step": 12290 }, { "epoch": 0.44516479536399856, "grad_norm": 1.405454411229416, "learning_rate": 1.224301398223723e-05, "loss": 0.7904, "step": 12291 }, { "epoch": 0.44520101412531693, "grad_norm": 1.0648288643290504, "learning_rate": 1.2241870789064345e-05, "loss": 0.707, "step": 12292 }, { "epoch": 0.4452372328866353, "grad_norm": 1.35308485070675, "learning_rate": 1.2240727565041346e-05, "loss": 0.6977, "step": 12293 }, { "epoch": 0.4452734516479536, "grad_norm": 1.3790160856728217, "learning_rate": 1.223958431018397e-05, "loss": 0.7855, "step": 12294 }, { "epoch": 0.445309670409272, "grad_norm": 1.3357833039935834, "learning_rate": 1.2238441024507943e-05, "loss": 0.777, "step": 12295 }, { "epoch": 0.44534588917059037, "grad_norm": 1.6216619219766506, "learning_rate": 1.2237297708029004e-05, "loss": 0.7969, "step": 12296 }, { "epoch": 0.44538210793190874, "grad_norm": 1.5113694307347547, "learning_rate": 1.223615436076288e-05, "loss": 0.759, "step": 12297 }, { "epoch": 0.4454183266932271, "grad_norm": 1.3731308571566205, "learning_rate": 1.2235010982725308e-05, "loss": 0.7297, "step": 12298 }, { "epoch": 0.44545454545454544, "grad_norm": 1.3983606625071543, "learning_rate": 1.2233867573932022e-05, "loss": 0.7934, "step": 12299 }, { "epoch": 0.4454907642158638, "grad_norm": 1.1235342923061162, "learning_rate": 1.2232724134398755e-05, "loss": 0.8346, "step": 12300 }, { "epoch": 0.4455269829771822, "grad_norm": 1.4102960303287597, "learning_rate": 1.2231580664141241e-05, "loss": 0.7591, "step": 12301 }, { "epoch": 0.44556320173850056, "grad_norm": 1.602682077045948, "learning_rate": 1.2230437163175216e-05, "loss": 0.7279, "step": 12302 }, { "epoch": 0.44559942049981893, "grad_norm": 1.0734234677168717, "learning_rate": 1.2229293631516414e-05, "loss": 0.7311, "step": 12303 }, { "epoch": 0.44563563926113725, "grad_norm": 1.30490985705872, "learning_rate": 1.2228150069180575e-05, "loss": 0.7434, "step": 12304 }, { "epoch": 0.4456718580224556, "grad_norm": 0.9846537877387425, "learning_rate": 1.2227006476183435e-05, "loss": 0.7311, "step": 12305 }, { "epoch": 0.445708076783774, "grad_norm": 0.9739266817034609, "learning_rate": 1.2225862852540727e-05, "loss": 0.7386, "step": 12306 }, { "epoch": 0.4457442955450924, "grad_norm": 1.3427270437375496, "learning_rate": 1.222471919826819e-05, "loss": 0.7471, "step": 12307 }, { "epoch": 0.44578051430641075, "grad_norm": 1.2984822723436646, "learning_rate": 1.2223575513381563e-05, "loss": 0.6956, "step": 12308 }, { "epoch": 0.44581673306772907, "grad_norm": 1.2933182776951648, "learning_rate": 1.2222431797896586e-05, "loss": 0.8001, "step": 12309 }, { "epoch": 0.44585295182904744, "grad_norm": 1.3594179256282706, "learning_rate": 1.222128805182899e-05, "loss": 0.7557, "step": 12310 }, { "epoch": 0.4458891705903658, "grad_norm": 1.376842526803129, "learning_rate": 1.222014427519452e-05, "loss": 0.7444, "step": 12311 }, { "epoch": 0.4459253893516842, "grad_norm": 1.2837846677847988, "learning_rate": 1.2219000468008915e-05, "loss": 0.7021, "step": 12312 }, { "epoch": 0.44596160811300256, "grad_norm": 1.3819785405097653, "learning_rate": 1.2217856630287912e-05, "loss": 0.6971, "step": 12313 }, { "epoch": 0.4459978268743209, "grad_norm": 1.3661234625188055, "learning_rate": 1.2216712762047254e-05, "loss": 0.7497, "step": 12314 }, { "epoch": 0.44603404563563925, "grad_norm": 1.4981545666974643, "learning_rate": 1.2215568863302681e-05, "loss": 0.7045, "step": 12315 }, { "epoch": 0.4460702643969576, "grad_norm": 2.376795651266858, "learning_rate": 1.2214424934069933e-05, "loss": 0.775, "step": 12316 }, { "epoch": 0.446106483158276, "grad_norm": 1.201863682022107, "learning_rate": 1.2213280974364751e-05, "loss": 0.6543, "step": 12317 }, { "epoch": 0.4461427019195944, "grad_norm": 1.0393653942743724, "learning_rate": 1.221213698420288e-05, "loss": 0.7296, "step": 12318 }, { "epoch": 0.4461789206809127, "grad_norm": 1.3308626114705433, "learning_rate": 1.221099296360006e-05, "loss": 0.7837, "step": 12319 }, { "epoch": 0.44621513944223107, "grad_norm": 1.0859118835833474, "learning_rate": 1.2209848912572036e-05, "loss": 0.7437, "step": 12320 }, { "epoch": 0.44625135820354944, "grad_norm": 1.33939997118961, "learning_rate": 1.2208704831134545e-05, "loss": 0.8335, "step": 12321 }, { "epoch": 0.4462875769648678, "grad_norm": 1.238000975762129, "learning_rate": 1.2207560719303337e-05, "loss": 0.7362, "step": 12322 }, { "epoch": 0.4463237957261862, "grad_norm": 1.3128076899426082, "learning_rate": 1.2206416577094153e-05, "loss": 0.743, "step": 12323 }, { "epoch": 0.4463600144875045, "grad_norm": 1.4632624241544079, "learning_rate": 1.2205272404522739e-05, "loss": 0.7779, "step": 12324 }, { "epoch": 0.4463962332488229, "grad_norm": 1.3841445482395116, "learning_rate": 1.220412820160484e-05, "loss": 0.8343, "step": 12325 }, { "epoch": 0.44643245201014126, "grad_norm": 1.3263401407903155, "learning_rate": 1.2202983968356197e-05, "loss": 0.7739, "step": 12326 }, { "epoch": 0.44646867077145963, "grad_norm": 0.9616544718153767, "learning_rate": 1.2201839704792562e-05, "loss": 0.7597, "step": 12327 }, { "epoch": 0.446504889532778, "grad_norm": 1.3979107881114237, "learning_rate": 1.2200695410929677e-05, "loss": 0.8126, "step": 12328 }, { "epoch": 0.4465411082940963, "grad_norm": 1.395234184587661, "learning_rate": 1.219955108678329e-05, "loss": 0.7488, "step": 12329 }, { "epoch": 0.4465773270554147, "grad_norm": 1.4405473990754651, "learning_rate": 1.2198406732369147e-05, "loss": 0.7855, "step": 12330 }, { "epoch": 0.44661354581673307, "grad_norm": 1.686942010479947, "learning_rate": 1.2197262347702993e-05, "loss": 0.8251, "step": 12331 }, { "epoch": 0.44664976457805144, "grad_norm": 1.710430993482067, "learning_rate": 1.2196117932800581e-05, "loss": 0.7029, "step": 12332 }, { "epoch": 0.4466859833393698, "grad_norm": 1.4792259656634035, "learning_rate": 1.2194973487677655e-05, "loss": 0.7905, "step": 12333 }, { "epoch": 0.44672220210068814, "grad_norm": 1.485250242591065, "learning_rate": 1.2193829012349968e-05, "loss": 0.7457, "step": 12334 }, { "epoch": 0.4467584208620065, "grad_norm": 1.3191776950169707, "learning_rate": 1.2192684506833264e-05, "loss": 0.7308, "step": 12335 }, { "epoch": 0.4467946396233249, "grad_norm": 1.2867820074517062, "learning_rate": 1.2191539971143296e-05, "loss": 0.6961, "step": 12336 }, { "epoch": 0.44683085838464326, "grad_norm": 1.3136230847908472, "learning_rate": 1.2190395405295811e-05, "loss": 0.7223, "step": 12337 }, { "epoch": 0.44686707714596163, "grad_norm": 1.3536276516372765, "learning_rate": 1.218925080930656e-05, "loss": 0.6713, "step": 12338 }, { "epoch": 0.44690329590727995, "grad_norm": 1.2087679284632526, "learning_rate": 1.2188106183191298e-05, "loss": 0.7812, "step": 12339 }, { "epoch": 0.4469395146685983, "grad_norm": 1.4532937649174782, "learning_rate": 1.2186961526965768e-05, "loss": 0.8028, "step": 12340 }, { "epoch": 0.4469757334299167, "grad_norm": 1.5392898540504183, "learning_rate": 1.2185816840645727e-05, "loss": 0.8685, "step": 12341 }, { "epoch": 0.4470119521912351, "grad_norm": 1.3466923065092247, "learning_rate": 1.2184672124246928e-05, "loss": 0.7806, "step": 12342 }, { "epoch": 0.44704817095255345, "grad_norm": 1.4876281814783983, "learning_rate": 1.2183527377785116e-05, "loss": 0.7316, "step": 12343 }, { "epoch": 0.44708438971387177, "grad_norm": 1.3416936201654446, "learning_rate": 1.2182382601276053e-05, "loss": 0.8047, "step": 12344 }, { "epoch": 0.44712060847519014, "grad_norm": 1.4993095218150303, "learning_rate": 1.2181237794735486e-05, "loss": 0.8075, "step": 12345 }, { "epoch": 0.4471568272365085, "grad_norm": 1.0812126774768562, "learning_rate": 1.2180092958179171e-05, "loss": 0.7261, "step": 12346 }, { "epoch": 0.4471930459978269, "grad_norm": 1.0098732114168445, "learning_rate": 1.217894809162286e-05, "loss": 0.7381, "step": 12347 }, { "epoch": 0.44722926475914526, "grad_norm": 1.3890707888962004, "learning_rate": 1.217780319508231e-05, "loss": 0.7811, "step": 12348 }, { "epoch": 0.4472654835204636, "grad_norm": 1.3886459981275205, "learning_rate": 1.2176658268573273e-05, "loss": 0.6266, "step": 12349 }, { "epoch": 0.44730170228178195, "grad_norm": 1.3683027902338443, "learning_rate": 1.2175513312111507e-05, "loss": 0.793, "step": 12350 }, { "epoch": 0.44733792104310033, "grad_norm": 1.3794509241510342, "learning_rate": 1.2174368325712766e-05, "loss": 0.7701, "step": 12351 }, { "epoch": 0.4473741398044187, "grad_norm": 1.397914071928687, "learning_rate": 1.2173223309392808e-05, "loss": 0.7178, "step": 12352 }, { "epoch": 0.4474103585657371, "grad_norm": 1.6480057860661752, "learning_rate": 1.2172078263167387e-05, "loss": 0.7596, "step": 12353 }, { "epoch": 0.4474465773270554, "grad_norm": 1.3661078399848712, "learning_rate": 1.2170933187052257e-05, "loss": 0.7492, "step": 12354 }, { "epoch": 0.44748279608837377, "grad_norm": 1.3681485932841826, "learning_rate": 1.2169788081063181e-05, "loss": 0.717, "step": 12355 }, { "epoch": 0.44751901484969214, "grad_norm": 1.2827986180993158, "learning_rate": 1.2168642945215916e-05, "loss": 0.7537, "step": 12356 }, { "epoch": 0.4475552336110105, "grad_norm": 1.4937596608342316, "learning_rate": 1.2167497779526215e-05, "loss": 0.7436, "step": 12357 }, { "epoch": 0.4475914523723289, "grad_norm": 1.692077928431678, "learning_rate": 1.2166352584009843e-05, "loss": 0.6558, "step": 12358 }, { "epoch": 0.4476276711336472, "grad_norm": 1.3936831823399558, "learning_rate": 1.2165207358682555e-05, "loss": 0.7292, "step": 12359 }, { "epoch": 0.4476638898949656, "grad_norm": 1.3691710919025513, "learning_rate": 1.2164062103560112e-05, "loss": 0.8277, "step": 12360 }, { "epoch": 0.44770010865628396, "grad_norm": 1.3619653737249282, "learning_rate": 1.2162916818658272e-05, "loss": 0.7819, "step": 12361 }, { "epoch": 0.44773632741760233, "grad_norm": 1.0447175870159673, "learning_rate": 1.2161771503992798e-05, "loss": 0.8028, "step": 12362 }, { "epoch": 0.4477725461789207, "grad_norm": 1.239894082945017, "learning_rate": 1.2160626159579447e-05, "loss": 0.7477, "step": 12363 }, { "epoch": 0.447808764940239, "grad_norm": 1.3557458094209047, "learning_rate": 1.2159480785433978e-05, "loss": 0.6979, "step": 12364 }, { "epoch": 0.4478449837015574, "grad_norm": 1.432011393665894, "learning_rate": 1.2158335381572162e-05, "loss": 0.7365, "step": 12365 }, { "epoch": 0.44788120246287577, "grad_norm": 1.2666558325382868, "learning_rate": 1.2157189948009752e-05, "loss": 0.655, "step": 12366 }, { "epoch": 0.44791742122419415, "grad_norm": 1.333197388076398, "learning_rate": 1.2156044484762514e-05, "loss": 0.8214, "step": 12367 }, { "epoch": 0.4479536399855125, "grad_norm": 1.330610014874538, "learning_rate": 1.2154898991846207e-05, "loss": 0.7168, "step": 12368 }, { "epoch": 0.44798985874683084, "grad_norm": 1.3544819363217824, "learning_rate": 1.2153753469276599e-05, "loss": 0.756, "step": 12369 }, { "epoch": 0.4480260775081492, "grad_norm": 1.2693789310044634, "learning_rate": 1.215260791706945e-05, "loss": 0.699, "step": 12370 }, { "epoch": 0.4480622962694676, "grad_norm": 1.3755444917036919, "learning_rate": 1.2151462335240524e-05, "loss": 0.7402, "step": 12371 }, { "epoch": 0.44809851503078596, "grad_norm": 1.2735642506482254, "learning_rate": 1.2150316723805586e-05, "loss": 0.7614, "step": 12372 }, { "epoch": 0.44813473379210433, "grad_norm": 1.1411635614167757, "learning_rate": 1.2149171082780402e-05, "loss": 0.7193, "step": 12373 }, { "epoch": 0.44817095255342265, "grad_norm": 1.3435372299200292, "learning_rate": 1.2148025412180733e-05, "loss": 0.7009, "step": 12374 }, { "epoch": 0.448207171314741, "grad_norm": 1.8668876929980278, "learning_rate": 1.2146879712022348e-05, "loss": 0.7971, "step": 12375 }, { "epoch": 0.4482433900760594, "grad_norm": 1.3727182887761282, "learning_rate": 1.2145733982321012e-05, "loss": 0.747, "step": 12376 }, { "epoch": 0.4482796088373778, "grad_norm": 1.4017378360100348, "learning_rate": 1.214458822309249e-05, "loss": 0.7705, "step": 12377 }, { "epoch": 0.44831582759869615, "grad_norm": 1.554539550038636, "learning_rate": 1.2143442434352552e-05, "loss": 0.7722, "step": 12378 }, { "epoch": 0.44835204636001447, "grad_norm": 1.0482912900756323, "learning_rate": 1.2142296616116963e-05, "loss": 0.7851, "step": 12379 }, { "epoch": 0.44838826512133284, "grad_norm": 1.3099605980049809, "learning_rate": 1.2141150768401486e-05, "loss": 0.7931, "step": 12380 }, { "epoch": 0.4484244838826512, "grad_norm": 1.2963962999984793, "learning_rate": 1.2140004891221895e-05, "loss": 0.6997, "step": 12381 }, { "epoch": 0.4484607026439696, "grad_norm": 1.348012518438034, "learning_rate": 1.2138858984593956e-05, "loss": 0.8142, "step": 12382 }, { "epoch": 0.44849692140528796, "grad_norm": 1.392037439848514, "learning_rate": 1.2137713048533439e-05, "loss": 0.7426, "step": 12383 }, { "epoch": 0.4485331401666063, "grad_norm": 1.8092492704529364, "learning_rate": 1.2136567083056112e-05, "loss": 0.7184, "step": 12384 }, { "epoch": 0.44856935892792466, "grad_norm": 1.3543662065584472, "learning_rate": 1.2135421088177746e-05, "loss": 0.726, "step": 12385 }, { "epoch": 0.44860557768924303, "grad_norm": 1.4592830500337666, "learning_rate": 1.2134275063914104e-05, "loss": 0.8271, "step": 12386 }, { "epoch": 0.4486417964505614, "grad_norm": 1.4083767855276899, "learning_rate": 1.213312901028097e-05, "loss": 0.7315, "step": 12387 }, { "epoch": 0.4486780152118798, "grad_norm": 1.3775873515942307, "learning_rate": 1.2131982927294099e-05, "loss": 0.7131, "step": 12388 }, { "epoch": 0.4487142339731981, "grad_norm": 1.1651244210184208, "learning_rate": 1.2130836814969275e-05, "loss": 0.7667, "step": 12389 }, { "epoch": 0.44875045273451647, "grad_norm": 1.1841315046770107, "learning_rate": 1.2129690673322261e-05, "loss": 0.7165, "step": 12390 }, { "epoch": 0.44878667149583484, "grad_norm": 1.2681001490531125, "learning_rate": 1.212854450236883e-05, "loss": 0.746, "step": 12391 }, { "epoch": 0.4488228902571532, "grad_norm": 1.4413392224531536, "learning_rate": 1.2127398302124761e-05, "loss": 0.7471, "step": 12392 }, { "epoch": 0.4488591090184716, "grad_norm": 1.4680016575617245, "learning_rate": 1.212625207260582e-05, "loss": 0.8037, "step": 12393 }, { "epoch": 0.4488953277797899, "grad_norm": 1.4810392524843594, "learning_rate": 1.2125105813827782e-05, "loss": 0.7551, "step": 12394 }, { "epoch": 0.4489315465411083, "grad_norm": 1.3915872024330616, "learning_rate": 1.2123959525806421e-05, "loss": 0.7082, "step": 12395 }, { "epoch": 0.44896776530242666, "grad_norm": 1.1340449528655199, "learning_rate": 1.2122813208557512e-05, "loss": 0.7354, "step": 12396 }, { "epoch": 0.44900398406374503, "grad_norm": 1.268955785465339, "learning_rate": 1.2121666862096827e-05, "loss": 0.7192, "step": 12397 }, { "epoch": 0.4490402028250634, "grad_norm": 1.0870594506223292, "learning_rate": 1.2120520486440142e-05, "loss": 0.7461, "step": 12398 }, { "epoch": 0.4490764215863817, "grad_norm": 1.1786542339955512, "learning_rate": 1.211937408160323e-05, "loss": 0.7249, "step": 12399 }, { "epoch": 0.4491126403477001, "grad_norm": 1.3223717194765396, "learning_rate": 1.211822764760187e-05, "loss": 0.7257, "step": 12400 }, { "epoch": 0.4491488591090185, "grad_norm": 1.6118462798716788, "learning_rate": 1.2117081184451835e-05, "loss": 0.8277, "step": 12401 }, { "epoch": 0.44918507787033685, "grad_norm": 1.2030303278434773, "learning_rate": 1.2115934692168902e-05, "loss": 0.7679, "step": 12402 }, { "epoch": 0.4492212966316552, "grad_norm": 1.0456566417577842, "learning_rate": 1.2114788170768852e-05, "loss": 0.7473, "step": 12403 }, { "epoch": 0.44925751539297354, "grad_norm": 1.592723776664693, "learning_rate": 1.2113641620267456e-05, "loss": 0.8199, "step": 12404 }, { "epoch": 0.4492937341542919, "grad_norm": 0.9149764629182946, "learning_rate": 1.2112495040680496e-05, "loss": 0.6518, "step": 12405 }, { "epoch": 0.4493299529156103, "grad_norm": 1.4930554843842199, "learning_rate": 1.2111348432023746e-05, "loss": 0.7593, "step": 12406 }, { "epoch": 0.44936617167692866, "grad_norm": 1.3974951480647329, "learning_rate": 1.2110201794312987e-05, "loss": 0.7191, "step": 12407 }, { "epoch": 0.44940239043824703, "grad_norm": 1.357787873304893, "learning_rate": 1.2109055127563998e-05, "loss": 0.7689, "step": 12408 }, { "epoch": 0.44943860919956535, "grad_norm": 1.4657589399538717, "learning_rate": 1.2107908431792556e-05, "loss": 0.7661, "step": 12409 }, { "epoch": 0.4494748279608837, "grad_norm": 1.6250780664545468, "learning_rate": 1.2106761707014443e-05, "loss": 0.7481, "step": 12410 }, { "epoch": 0.4495110467222021, "grad_norm": 1.2747451137576833, "learning_rate": 1.2105614953245436e-05, "loss": 0.7317, "step": 12411 }, { "epoch": 0.4495472654835205, "grad_norm": 1.4249864832232864, "learning_rate": 1.210446817050132e-05, "loss": 0.7669, "step": 12412 }, { "epoch": 0.44958348424483885, "grad_norm": 1.3248416735190474, "learning_rate": 1.2103321358797868e-05, "loss": 0.7732, "step": 12413 }, { "epoch": 0.44961970300615717, "grad_norm": 1.7763491632130117, "learning_rate": 1.210217451815087e-05, "loss": 0.7453, "step": 12414 }, { "epoch": 0.44965592176747554, "grad_norm": 1.6907098164498047, "learning_rate": 1.2101027648576103e-05, "loss": 0.6657, "step": 12415 }, { "epoch": 0.4496921405287939, "grad_norm": 1.4437715563878146, "learning_rate": 1.209988075008935e-05, "loss": 0.7294, "step": 12416 }, { "epoch": 0.4497283592901123, "grad_norm": 1.3145905751580764, "learning_rate": 1.209873382270639e-05, "loss": 0.6925, "step": 12417 }, { "epoch": 0.44976457805143066, "grad_norm": 1.3213732298597467, "learning_rate": 1.2097586866443011e-05, "loss": 0.7427, "step": 12418 }, { "epoch": 0.449800796812749, "grad_norm": 1.3609940352889018, "learning_rate": 1.2096439881314991e-05, "loss": 0.6898, "step": 12419 }, { "epoch": 0.44983701557406736, "grad_norm": 1.1986670739528316, "learning_rate": 1.2095292867338122e-05, "loss": 0.7802, "step": 12420 }, { "epoch": 0.44987323433538573, "grad_norm": 1.5691083138871036, "learning_rate": 1.2094145824528176e-05, "loss": 0.7594, "step": 12421 }, { "epoch": 0.4499094530967041, "grad_norm": 1.193810706844238, "learning_rate": 1.2092998752900945e-05, "loss": 0.8071, "step": 12422 }, { "epoch": 0.4499456718580225, "grad_norm": 1.3383115816214282, "learning_rate": 1.2091851652472213e-05, "loss": 0.7692, "step": 12423 }, { "epoch": 0.4499818906193408, "grad_norm": 1.3893874413207392, "learning_rate": 1.2090704523257763e-05, "loss": 0.7429, "step": 12424 }, { "epoch": 0.45001810938065917, "grad_norm": 1.338650898119268, "learning_rate": 1.2089557365273383e-05, "loss": 0.79, "step": 12425 }, { "epoch": 0.45005432814197754, "grad_norm": 1.3694718541149598, "learning_rate": 1.2088410178534857e-05, "loss": 0.7412, "step": 12426 }, { "epoch": 0.4500905469032959, "grad_norm": 1.0980259027474846, "learning_rate": 1.2087262963057973e-05, "loss": 0.6714, "step": 12427 }, { "epoch": 0.4501267656646143, "grad_norm": 1.3511833667955706, "learning_rate": 1.2086115718858513e-05, "loss": 0.8032, "step": 12428 }, { "epoch": 0.4501629844259326, "grad_norm": 1.340545689380939, "learning_rate": 1.2084968445952272e-05, "loss": 0.7037, "step": 12429 }, { "epoch": 0.450199203187251, "grad_norm": 1.6999472721191538, "learning_rate": 1.208382114435503e-05, "loss": 0.7337, "step": 12430 }, { "epoch": 0.45023542194856936, "grad_norm": 1.2759162300277052, "learning_rate": 1.2082673814082581e-05, "loss": 0.7937, "step": 12431 }, { "epoch": 0.45027164070988773, "grad_norm": 1.3665398596463363, "learning_rate": 1.2081526455150707e-05, "loss": 0.7417, "step": 12432 }, { "epoch": 0.4503078594712061, "grad_norm": 1.3830935979149355, "learning_rate": 1.20803790675752e-05, "loss": 0.7945, "step": 12433 }, { "epoch": 0.4503440782325244, "grad_norm": 1.412676524529556, "learning_rate": 1.2079231651371854e-05, "loss": 0.7608, "step": 12434 }, { "epoch": 0.4503802969938428, "grad_norm": 1.3186142039382303, "learning_rate": 1.2078084206556448e-05, "loss": 0.8288, "step": 12435 }, { "epoch": 0.4504165157551612, "grad_norm": 1.0680440898634607, "learning_rate": 1.2076936733144781e-05, "loss": 0.7257, "step": 12436 }, { "epoch": 0.45045273451647955, "grad_norm": 1.3665775839160466, "learning_rate": 1.2075789231152637e-05, "loss": 0.7529, "step": 12437 }, { "epoch": 0.4504889532777979, "grad_norm": 1.405856625850244, "learning_rate": 1.2074641700595813e-05, "loss": 0.7754, "step": 12438 }, { "epoch": 0.45052517203911624, "grad_norm": 1.362806652250517, "learning_rate": 1.2073494141490092e-05, "loss": 0.712, "step": 12439 }, { "epoch": 0.4505613908004346, "grad_norm": 1.3967982119591127, "learning_rate": 1.2072346553851273e-05, "loss": 0.7788, "step": 12440 }, { "epoch": 0.450597609561753, "grad_norm": 1.0541159290806998, "learning_rate": 1.2071198937695141e-05, "loss": 0.7382, "step": 12441 }, { "epoch": 0.45063382832307136, "grad_norm": 1.4944981863987812, "learning_rate": 1.2070051293037493e-05, "loss": 0.7445, "step": 12442 }, { "epoch": 0.45067004708438974, "grad_norm": 1.5181646686406989, "learning_rate": 1.206890361989412e-05, "loss": 0.8294, "step": 12443 }, { "epoch": 0.45070626584570805, "grad_norm": 1.3215399346778318, "learning_rate": 1.2067755918280817e-05, "loss": 0.7475, "step": 12444 }, { "epoch": 0.45074248460702643, "grad_norm": 1.1488055803848993, "learning_rate": 1.2066608188213376e-05, "loss": 0.6902, "step": 12445 }, { "epoch": 0.4507787033683448, "grad_norm": 1.5379576989861716, "learning_rate": 1.2065460429707589e-05, "loss": 0.66, "step": 12446 }, { "epoch": 0.4508149221296632, "grad_norm": 1.3657216442705864, "learning_rate": 1.2064312642779253e-05, "loss": 0.7291, "step": 12447 }, { "epoch": 0.45085114089098155, "grad_norm": 1.6683328921181788, "learning_rate": 1.2063164827444158e-05, "loss": 0.7996, "step": 12448 }, { "epoch": 0.45088735965229987, "grad_norm": 0.9914292057609696, "learning_rate": 1.2062016983718106e-05, "loss": 0.7347, "step": 12449 }, { "epoch": 0.45092357841361824, "grad_norm": 1.4195275159166316, "learning_rate": 1.2060869111616886e-05, "loss": 0.7253, "step": 12450 }, { "epoch": 0.4509597971749366, "grad_norm": 1.3742780374985208, "learning_rate": 1.2059721211156296e-05, "loss": 0.7456, "step": 12451 }, { "epoch": 0.450996015936255, "grad_norm": 1.428153173598639, "learning_rate": 1.2058573282352132e-05, "loss": 0.7722, "step": 12452 }, { "epoch": 0.45103223469757336, "grad_norm": 1.1337135700110095, "learning_rate": 1.2057425325220193e-05, "loss": 0.7474, "step": 12453 }, { "epoch": 0.4510684534588917, "grad_norm": 1.4320975676423306, "learning_rate": 1.2056277339776275e-05, "loss": 0.6505, "step": 12454 }, { "epoch": 0.45110467222021006, "grad_norm": 1.1059154430047828, "learning_rate": 1.205512932603617e-05, "loss": 0.8297, "step": 12455 }, { "epoch": 0.45114089098152843, "grad_norm": 1.387492820789906, "learning_rate": 1.2053981284015682e-05, "loss": 0.7366, "step": 12456 }, { "epoch": 0.4511771097428468, "grad_norm": 1.2791428771768305, "learning_rate": 1.2052833213730607e-05, "loss": 0.7493, "step": 12457 }, { "epoch": 0.4512133285041652, "grad_norm": 1.2605762860082734, "learning_rate": 1.205168511519674e-05, "loss": 0.7615, "step": 12458 }, { "epoch": 0.4512495472654835, "grad_norm": 1.1625294437548388, "learning_rate": 1.2050536988429886e-05, "loss": 0.7777, "step": 12459 }, { "epoch": 0.45128576602680187, "grad_norm": 1.3902332465340603, "learning_rate": 1.2049388833445842e-05, "loss": 0.8231, "step": 12460 }, { "epoch": 0.45132198478812025, "grad_norm": 1.135912288878089, "learning_rate": 1.2048240650260402e-05, "loss": 0.7653, "step": 12461 }, { "epoch": 0.4513582035494386, "grad_norm": 1.3383312314270812, "learning_rate": 1.2047092438889378e-05, "loss": 0.7582, "step": 12462 }, { "epoch": 0.451394422310757, "grad_norm": 1.329947560437734, "learning_rate": 1.2045944199348558e-05, "loss": 0.8064, "step": 12463 }, { "epoch": 0.4514306410720753, "grad_norm": 1.1123143328921667, "learning_rate": 1.204479593165375e-05, "loss": 0.7033, "step": 12464 }, { "epoch": 0.4514668598333937, "grad_norm": 1.2141518630175085, "learning_rate": 1.2043647635820755e-05, "loss": 0.7862, "step": 12465 }, { "epoch": 0.45150307859471206, "grad_norm": 1.4304720056404057, "learning_rate": 1.2042499311865367e-05, "loss": 0.8235, "step": 12466 }, { "epoch": 0.45153929735603043, "grad_norm": 1.406048961307623, "learning_rate": 1.20413509598034e-05, "loss": 0.7744, "step": 12467 }, { "epoch": 0.4515755161173488, "grad_norm": 0.9713254843110218, "learning_rate": 1.2040202579650649e-05, "loss": 0.6949, "step": 12468 }, { "epoch": 0.4516117348786671, "grad_norm": 1.4072247795713144, "learning_rate": 1.2039054171422916e-05, "loss": 0.7097, "step": 12469 }, { "epoch": 0.4516479536399855, "grad_norm": 1.4586248337815613, "learning_rate": 1.2037905735136006e-05, "loss": 0.7625, "step": 12470 }, { "epoch": 0.4516841724013039, "grad_norm": 1.3423541069707994, "learning_rate": 1.2036757270805724e-05, "loss": 0.789, "step": 12471 }, { "epoch": 0.45172039116262225, "grad_norm": 1.2701111855409812, "learning_rate": 1.2035608778447869e-05, "loss": 0.7046, "step": 12472 }, { "epoch": 0.4517566099239406, "grad_norm": 1.3134976707946264, "learning_rate": 1.203446025807825e-05, "loss": 0.7824, "step": 12473 }, { "epoch": 0.45179282868525894, "grad_norm": 1.3833701477877072, "learning_rate": 1.2033311709712675e-05, "loss": 0.7542, "step": 12474 }, { "epoch": 0.4518290474465773, "grad_norm": 1.3296249446655504, "learning_rate": 1.2032163133366939e-05, "loss": 0.7642, "step": 12475 }, { "epoch": 0.4518652662078957, "grad_norm": 1.2546957425752918, "learning_rate": 1.2031014529056852e-05, "loss": 0.6775, "step": 12476 }, { "epoch": 0.45190148496921406, "grad_norm": 1.0932056741460532, "learning_rate": 1.2029865896798222e-05, "loss": 0.8075, "step": 12477 }, { "epoch": 0.45193770373053244, "grad_norm": 1.3897148137917907, "learning_rate": 1.2028717236606857e-05, "loss": 0.7694, "step": 12478 }, { "epoch": 0.45197392249185075, "grad_norm": 1.315343025598297, "learning_rate": 1.2027568548498556e-05, "loss": 0.7366, "step": 12479 }, { "epoch": 0.45201014125316913, "grad_norm": 1.4865157376156233, "learning_rate": 1.2026419832489133e-05, "loss": 0.7737, "step": 12480 }, { "epoch": 0.4520463600144875, "grad_norm": 1.333688039804045, "learning_rate": 1.2025271088594392e-05, "loss": 0.7767, "step": 12481 }, { "epoch": 0.4520825787758059, "grad_norm": 1.5153941096591115, "learning_rate": 1.202412231683014e-05, "loss": 0.7657, "step": 12482 }, { "epoch": 0.45211879753712425, "grad_norm": 1.637324350888112, "learning_rate": 1.2022973517212186e-05, "loss": 0.7451, "step": 12483 }, { "epoch": 0.45215501629844257, "grad_norm": 1.3591510421052002, "learning_rate": 1.202182468975634e-05, "loss": 0.7027, "step": 12484 }, { "epoch": 0.45219123505976094, "grad_norm": 1.2745512738665243, "learning_rate": 1.202067583447841e-05, "loss": 0.6667, "step": 12485 }, { "epoch": 0.4522274538210793, "grad_norm": 1.0844162121614014, "learning_rate": 1.2019526951394203e-05, "loss": 0.812, "step": 12486 }, { "epoch": 0.4522636725823977, "grad_norm": 1.6942304781763144, "learning_rate": 1.2018378040519534e-05, "loss": 0.7993, "step": 12487 }, { "epoch": 0.45229989134371607, "grad_norm": 1.3759023837257214, "learning_rate": 1.2017229101870206e-05, "loss": 0.7977, "step": 12488 }, { "epoch": 0.4523361101050344, "grad_norm": 1.439275090686808, "learning_rate": 1.2016080135462034e-05, "loss": 0.7955, "step": 12489 }, { "epoch": 0.45237232886635276, "grad_norm": 1.4936054343333118, "learning_rate": 1.201493114131083e-05, "loss": 0.7266, "step": 12490 }, { "epoch": 0.45240854762767113, "grad_norm": 1.296769490459893, "learning_rate": 1.20137821194324e-05, "loss": 0.749, "step": 12491 }, { "epoch": 0.4524447663889895, "grad_norm": 1.3802622295748852, "learning_rate": 1.201263306984256e-05, "loss": 0.7315, "step": 12492 }, { "epoch": 0.4524809851503079, "grad_norm": 1.3344544452383553, "learning_rate": 1.201148399255712e-05, "loss": 0.7615, "step": 12493 }, { "epoch": 0.4525172039116262, "grad_norm": 1.41419299192724, "learning_rate": 1.2010334887591894e-05, "loss": 0.7833, "step": 12494 }, { "epoch": 0.45255342267294457, "grad_norm": 1.3993583503063185, "learning_rate": 1.2009185754962693e-05, "loss": 0.8222, "step": 12495 }, { "epoch": 0.45258964143426295, "grad_norm": 2.3720410280591193, "learning_rate": 1.2008036594685333e-05, "loss": 0.7151, "step": 12496 }, { "epoch": 0.4526258601955813, "grad_norm": 1.302458051089063, "learning_rate": 1.2006887406775621e-05, "loss": 0.7681, "step": 12497 }, { "epoch": 0.4526620789568997, "grad_norm": 1.295144834354221, "learning_rate": 1.2005738191249379e-05, "loss": 0.7551, "step": 12498 }, { "epoch": 0.452698297718218, "grad_norm": 1.332922406656998, "learning_rate": 1.2004588948122414e-05, "loss": 0.7575, "step": 12499 }, { "epoch": 0.4527345164795364, "grad_norm": 1.3705488684375864, "learning_rate": 1.2003439677410542e-05, "loss": 0.6933, "step": 12500 }, { "epoch": 0.45277073524085476, "grad_norm": 1.3636442598799279, "learning_rate": 1.2002290379129582e-05, "loss": 0.6659, "step": 12501 }, { "epoch": 0.45280695400217313, "grad_norm": 1.3894231416600338, "learning_rate": 1.2001141053295347e-05, "loss": 0.7929, "step": 12502 }, { "epoch": 0.4528431727634915, "grad_norm": 1.7119547071490073, "learning_rate": 1.1999991699923654e-05, "loss": 0.8148, "step": 12503 }, { "epoch": 0.4528793915248098, "grad_norm": 1.4493629031706685, "learning_rate": 1.1998842319030315e-05, "loss": 0.7923, "step": 12504 }, { "epoch": 0.4529156102861282, "grad_norm": 1.3363182901312136, "learning_rate": 1.199769291063115e-05, "loss": 0.6991, "step": 12505 }, { "epoch": 0.4529518290474466, "grad_norm": 1.0844058076744885, "learning_rate": 1.1996543474741977e-05, "loss": 0.7218, "step": 12506 }, { "epoch": 0.45298804780876495, "grad_norm": 1.4488652291667397, "learning_rate": 1.1995394011378611e-05, "loss": 0.711, "step": 12507 }, { "epoch": 0.4530242665700833, "grad_norm": 1.110079861404308, "learning_rate": 1.199424452055687e-05, "loss": 0.7398, "step": 12508 }, { "epoch": 0.45306048533140164, "grad_norm": 1.0763502335558646, "learning_rate": 1.1993095002292573e-05, "loss": 0.7635, "step": 12509 }, { "epoch": 0.45309670409272, "grad_norm": 1.2130942800448723, "learning_rate": 1.1991945456601534e-05, "loss": 0.8017, "step": 12510 }, { "epoch": 0.4531329228540384, "grad_norm": 1.0630952136407097, "learning_rate": 1.1990795883499578e-05, "loss": 0.7103, "step": 12511 }, { "epoch": 0.45316914161535676, "grad_norm": 1.3604633890539017, "learning_rate": 1.198964628300252e-05, "loss": 0.8287, "step": 12512 }, { "epoch": 0.45320536037667514, "grad_norm": 1.4893335166825632, "learning_rate": 1.1988496655126182e-05, "loss": 0.7725, "step": 12513 }, { "epoch": 0.45324157913799346, "grad_norm": 1.3204798888577498, "learning_rate": 1.1987346999886384e-05, "loss": 0.7061, "step": 12514 }, { "epoch": 0.45327779789931183, "grad_norm": 1.3333007532537597, "learning_rate": 1.1986197317298943e-05, "loss": 0.7281, "step": 12515 }, { "epoch": 0.4533140166606302, "grad_norm": 1.247783806287268, "learning_rate": 1.1985047607379683e-05, "loss": 0.7348, "step": 12516 }, { "epoch": 0.4533502354219486, "grad_norm": 1.214669905885475, "learning_rate": 1.1983897870144423e-05, "loss": 0.6911, "step": 12517 }, { "epoch": 0.45338645418326695, "grad_norm": 1.6120336929850745, "learning_rate": 1.1982748105608985e-05, "loss": 0.7374, "step": 12518 }, { "epoch": 0.45342267294458527, "grad_norm": 1.4782701096973279, "learning_rate": 1.1981598313789192e-05, "loss": 0.7894, "step": 12519 }, { "epoch": 0.45345889170590364, "grad_norm": 1.4035643257251205, "learning_rate": 1.1980448494700861e-05, "loss": 0.7585, "step": 12520 }, { "epoch": 0.453495110467222, "grad_norm": 1.2288181586936557, "learning_rate": 1.1979298648359823e-05, "loss": 0.6961, "step": 12521 }, { "epoch": 0.4535313292285404, "grad_norm": 1.1496977418045347, "learning_rate": 1.1978148774781895e-05, "loss": 0.8005, "step": 12522 }, { "epoch": 0.45356754798985877, "grad_norm": 1.370379918276653, "learning_rate": 1.1976998873982902e-05, "loss": 0.6425, "step": 12523 }, { "epoch": 0.4536037667511771, "grad_norm": 1.4305757651568785, "learning_rate": 1.1975848945978665e-05, "loss": 0.7502, "step": 12524 }, { "epoch": 0.45363998551249546, "grad_norm": 1.3849311845322023, "learning_rate": 1.1974698990785015e-05, "loss": 0.7509, "step": 12525 }, { "epoch": 0.45367620427381383, "grad_norm": 1.3481152851506764, "learning_rate": 1.1973549008417767e-05, "loss": 0.7895, "step": 12526 }, { "epoch": 0.4537124230351322, "grad_norm": 1.0702835929884376, "learning_rate": 1.1972398998892753e-05, "loss": 0.7262, "step": 12527 }, { "epoch": 0.4537486417964506, "grad_norm": 1.3441082468542616, "learning_rate": 1.1971248962225797e-05, "loss": 0.6982, "step": 12528 }, { "epoch": 0.4537848605577689, "grad_norm": 1.2898883522184827, "learning_rate": 1.1970098898432721e-05, "loss": 0.7252, "step": 12529 }, { "epoch": 0.4538210793190873, "grad_norm": 1.4253641002485136, "learning_rate": 1.1968948807529353e-05, "loss": 0.7886, "step": 12530 }, { "epoch": 0.45385729808040565, "grad_norm": 1.2878063562226272, "learning_rate": 1.196779868953152e-05, "loss": 0.764, "step": 12531 }, { "epoch": 0.453893516841724, "grad_norm": 1.1258435456116955, "learning_rate": 1.1966648544455048e-05, "loss": 0.6579, "step": 12532 }, { "epoch": 0.4539297356030424, "grad_norm": 1.1810367195682903, "learning_rate": 1.1965498372315762e-05, "loss": 0.7379, "step": 12533 }, { "epoch": 0.4539659543643607, "grad_norm": 1.6875667680229556, "learning_rate": 1.1964348173129497e-05, "loss": 0.7944, "step": 12534 }, { "epoch": 0.4540021731256791, "grad_norm": 1.3079810857641128, "learning_rate": 1.1963197946912068e-05, "loss": 0.7605, "step": 12535 }, { "epoch": 0.45403839188699746, "grad_norm": 1.3309211090186541, "learning_rate": 1.1962047693679315e-05, "loss": 0.7815, "step": 12536 }, { "epoch": 0.45407461064831584, "grad_norm": 0.9907912366433245, "learning_rate": 1.1960897413447059e-05, "loss": 0.7831, "step": 12537 }, { "epoch": 0.4541108294096342, "grad_norm": 1.314521367624123, "learning_rate": 1.1959747106231134e-05, "loss": 0.7421, "step": 12538 }, { "epoch": 0.4541470481709525, "grad_norm": 1.3164343544396502, "learning_rate": 1.1958596772047367e-05, "loss": 0.7184, "step": 12539 }, { "epoch": 0.4541832669322709, "grad_norm": 1.4745490068116545, "learning_rate": 1.1957446410911586e-05, "loss": 0.7387, "step": 12540 }, { "epoch": 0.4542194856935893, "grad_norm": 1.1386591740279377, "learning_rate": 1.1956296022839622e-05, "loss": 0.8347, "step": 12541 }, { "epoch": 0.45425570445490765, "grad_norm": 1.414300461895045, "learning_rate": 1.1955145607847306e-05, "loss": 0.7876, "step": 12542 }, { "epoch": 0.454291923216226, "grad_norm": 1.3950666887633083, "learning_rate": 1.1953995165950469e-05, "loss": 0.7847, "step": 12543 }, { "epoch": 0.45432814197754434, "grad_norm": 1.246000666863715, "learning_rate": 1.195284469716494e-05, "loss": 0.6855, "step": 12544 }, { "epoch": 0.4543643607388627, "grad_norm": 1.3658490823131975, "learning_rate": 1.1951694201506556e-05, "loss": 0.8406, "step": 12545 }, { "epoch": 0.4544005795001811, "grad_norm": 1.3150173027949896, "learning_rate": 1.1950543678991142e-05, "loss": 0.736, "step": 12546 }, { "epoch": 0.45443679826149946, "grad_norm": 1.3985473208254326, "learning_rate": 1.1949393129634534e-05, "loss": 0.7354, "step": 12547 }, { "epoch": 0.45447301702281784, "grad_norm": 1.0730654378522344, "learning_rate": 1.1948242553452563e-05, "loss": 0.6998, "step": 12548 }, { "epoch": 0.45450923578413616, "grad_norm": 0.9303196168812877, "learning_rate": 1.1947091950461062e-05, "loss": 0.6725, "step": 12549 }, { "epoch": 0.45454545454545453, "grad_norm": 1.1282016009420903, "learning_rate": 1.1945941320675866e-05, "loss": 0.7496, "step": 12550 }, { "epoch": 0.4545816733067729, "grad_norm": 1.3375443156534625, "learning_rate": 1.1944790664112807e-05, "loss": 0.7026, "step": 12551 }, { "epoch": 0.4546178920680913, "grad_norm": 1.3606148702522995, "learning_rate": 1.194363998078772e-05, "loss": 0.7531, "step": 12552 }, { "epoch": 0.45465411082940965, "grad_norm": 1.6404302549582617, "learning_rate": 1.1942489270716437e-05, "loss": 0.7932, "step": 12553 }, { "epoch": 0.45469032959072797, "grad_norm": 1.3622210829043768, "learning_rate": 1.1941338533914798e-05, "loss": 0.7767, "step": 12554 }, { "epoch": 0.45472654835204634, "grad_norm": 1.076441923068936, "learning_rate": 1.1940187770398633e-05, "loss": 0.7597, "step": 12555 }, { "epoch": 0.4547627671133647, "grad_norm": 1.2917906221169801, "learning_rate": 1.1939036980183781e-05, "loss": 0.8079, "step": 12556 }, { "epoch": 0.4547989858746831, "grad_norm": 1.2570341720096632, "learning_rate": 1.1937886163286075e-05, "loss": 0.7231, "step": 12557 }, { "epoch": 0.45483520463600147, "grad_norm": 1.3581890408276922, "learning_rate": 1.1936735319721354e-05, "loss": 0.731, "step": 12558 }, { "epoch": 0.4548714233973198, "grad_norm": 1.3275821242448722, "learning_rate": 1.1935584449505451e-05, "loss": 0.6935, "step": 12559 }, { "epoch": 0.45490764215863816, "grad_norm": 1.304245914750111, "learning_rate": 1.1934433552654208e-05, "loss": 0.7498, "step": 12560 }, { "epoch": 0.45494386091995653, "grad_norm": 1.5490373804419413, "learning_rate": 1.1933282629183458e-05, "loss": 0.7036, "step": 12561 }, { "epoch": 0.4549800796812749, "grad_norm": 1.3302493804382833, "learning_rate": 1.1932131679109038e-05, "loss": 0.7253, "step": 12562 }, { "epoch": 0.4550162984425933, "grad_norm": 1.3701932237534544, "learning_rate": 1.1930980702446793e-05, "loss": 0.7754, "step": 12563 }, { "epoch": 0.4550525172039116, "grad_norm": 1.473957405213602, "learning_rate": 1.1929829699212554e-05, "loss": 0.7948, "step": 12564 }, { "epoch": 0.45508873596523, "grad_norm": 1.2788231161719643, "learning_rate": 1.1928678669422164e-05, "loss": 0.6744, "step": 12565 }, { "epoch": 0.45512495472654835, "grad_norm": 1.3344666578239686, "learning_rate": 1.1927527613091459e-05, "loss": 0.748, "step": 12566 }, { "epoch": 0.4551611734878667, "grad_norm": 1.3523833835949115, "learning_rate": 1.1926376530236283e-05, "loss": 0.6512, "step": 12567 }, { "epoch": 0.4551973922491851, "grad_norm": 1.2992916054636279, "learning_rate": 1.1925225420872471e-05, "loss": 0.7837, "step": 12568 }, { "epoch": 0.4552336110105034, "grad_norm": 1.1648194583914264, "learning_rate": 1.1924074285015867e-05, "loss": 0.7642, "step": 12569 }, { "epoch": 0.4552698297718218, "grad_norm": 1.2245475268222217, "learning_rate": 1.1922923122682308e-05, "loss": 0.6822, "step": 12570 }, { "epoch": 0.45530604853314016, "grad_norm": 1.4431452882748192, "learning_rate": 1.192177193388764e-05, "loss": 0.7618, "step": 12571 }, { "epoch": 0.45534226729445854, "grad_norm": 1.5772097560902536, "learning_rate": 1.19206207186477e-05, "loss": 0.762, "step": 12572 }, { "epoch": 0.4553784860557769, "grad_norm": 1.3973851733192693, "learning_rate": 1.1919469476978329e-05, "loss": 0.7355, "step": 12573 }, { "epoch": 0.45541470481709523, "grad_norm": 1.3273157817044954, "learning_rate": 1.1918318208895377e-05, "loss": 0.7634, "step": 12574 }, { "epoch": 0.4554509235784136, "grad_norm": 1.2037291859808423, "learning_rate": 1.1917166914414678e-05, "loss": 0.727, "step": 12575 }, { "epoch": 0.455487142339732, "grad_norm": 1.1688775263598283, "learning_rate": 1.1916015593552077e-05, "loss": 0.7905, "step": 12576 }, { "epoch": 0.45552336110105035, "grad_norm": 1.2779383312228463, "learning_rate": 1.1914864246323417e-05, "loss": 0.788, "step": 12577 }, { "epoch": 0.4555595798623687, "grad_norm": 1.2502371801425802, "learning_rate": 1.1913712872744545e-05, "loss": 0.7909, "step": 12578 }, { "epoch": 0.45559579862368704, "grad_norm": 1.3217208198112775, "learning_rate": 1.1912561472831301e-05, "loss": 0.7569, "step": 12579 }, { "epoch": 0.4556320173850054, "grad_norm": 1.038721098532382, "learning_rate": 1.1911410046599533e-05, "loss": 0.7485, "step": 12580 }, { "epoch": 0.4556682361463238, "grad_norm": 1.345178803724415, "learning_rate": 1.1910258594065079e-05, "loss": 0.7003, "step": 12581 }, { "epoch": 0.45570445490764216, "grad_norm": 1.3661818586449237, "learning_rate": 1.1909107115243788e-05, "loss": 0.7439, "step": 12582 }, { "epoch": 0.45574067366896054, "grad_norm": 1.0677806605445723, "learning_rate": 1.1907955610151509e-05, "loss": 0.7844, "step": 12583 }, { "epoch": 0.45577689243027886, "grad_norm": 1.5328703610312626, "learning_rate": 1.1906804078804081e-05, "loss": 0.7849, "step": 12584 }, { "epoch": 0.45581311119159723, "grad_norm": 1.0251552426858233, "learning_rate": 1.1905652521217359e-05, "loss": 0.7269, "step": 12585 }, { "epoch": 0.4558493299529156, "grad_norm": 1.36128652455536, "learning_rate": 1.1904500937407181e-05, "loss": 0.7576, "step": 12586 }, { "epoch": 0.455885548714234, "grad_norm": 1.3723387832262894, "learning_rate": 1.1903349327389397e-05, "loss": 0.7425, "step": 12587 }, { "epoch": 0.45592176747555235, "grad_norm": 1.4773501755984064, "learning_rate": 1.1902197691179853e-05, "loss": 0.657, "step": 12588 }, { "epoch": 0.45595798623687067, "grad_norm": 1.46227204024101, "learning_rate": 1.1901046028794396e-05, "loss": 0.7428, "step": 12589 }, { "epoch": 0.45599420499818905, "grad_norm": 1.287171298666208, "learning_rate": 1.1899894340248879e-05, "loss": 0.625, "step": 12590 }, { "epoch": 0.4560304237595074, "grad_norm": 1.3217807154265437, "learning_rate": 1.1898742625559141e-05, "loss": 0.7607, "step": 12591 }, { "epoch": 0.4560666425208258, "grad_norm": 1.3442201041569863, "learning_rate": 1.1897590884741042e-05, "loss": 0.7031, "step": 12592 }, { "epoch": 0.45610286128214417, "grad_norm": 1.428099905102404, "learning_rate": 1.1896439117810421e-05, "loss": 0.7352, "step": 12593 }, { "epoch": 0.4561390800434625, "grad_norm": 1.402075789732665, "learning_rate": 1.1895287324783134e-05, "loss": 0.7027, "step": 12594 }, { "epoch": 0.45617529880478086, "grad_norm": 1.3305429537044906, "learning_rate": 1.1894135505675025e-05, "loss": 0.7582, "step": 12595 }, { "epoch": 0.45621151756609923, "grad_norm": 2.1227148054271794, "learning_rate": 1.1892983660501948e-05, "loss": 0.8041, "step": 12596 }, { "epoch": 0.4562477363274176, "grad_norm": 1.3996018525309246, "learning_rate": 1.1891831789279752e-05, "loss": 0.7327, "step": 12597 }, { "epoch": 0.456283955088736, "grad_norm": 1.2457816404250894, "learning_rate": 1.189067989202429e-05, "loss": 0.7027, "step": 12598 }, { "epoch": 0.45632017385005436, "grad_norm": 1.3426122908563836, "learning_rate": 1.1889527968751413e-05, "loss": 0.733, "step": 12599 }, { "epoch": 0.4563563926113727, "grad_norm": 1.241883021651779, "learning_rate": 1.1888376019476966e-05, "loss": 0.7545, "step": 12600 }, { "epoch": 0.45639261137269105, "grad_norm": 1.4088918824239063, "learning_rate": 1.188722404421681e-05, "loss": 0.7587, "step": 12601 }, { "epoch": 0.4564288301340094, "grad_norm": 1.5637484486223052, "learning_rate": 1.1886072042986791e-05, "loss": 0.8604, "step": 12602 }, { "epoch": 0.4564650488953278, "grad_norm": 1.3661797978722316, "learning_rate": 1.1884920015802765e-05, "loss": 0.8113, "step": 12603 }, { "epoch": 0.45650126765664617, "grad_norm": 1.4446968777958884, "learning_rate": 1.1883767962680581e-05, "loss": 0.8, "step": 12604 }, { "epoch": 0.4565374864179645, "grad_norm": 1.5449538016607194, "learning_rate": 1.1882615883636097e-05, "loss": 0.7627, "step": 12605 }, { "epoch": 0.45657370517928286, "grad_norm": 1.1460965221855166, "learning_rate": 1.1881463778685163e-05, "loss": 0.7122, "step": 12606 }, { "epoch": 0.45660992394060124, "grad_norm": 1.6034559954336913, "learning_rate": 1.1880311647843635e-05, "loss": 0.7541, "step": 12607 }, { "epoch": 0.4566461427019196, "grad_norm": 1.340708022892514, "learning_rate": 1.1879159491127363e-05, "loss": 0.6973, "step": 12608 }, { "epoch": 0.456682361463238, "grad_norm": 1.3227319071346038, "learning_rate": 1.1878007308552208e-05, "loss": 0.7968, "step": 12609 }, { "epoch": 0.4567185802245563, "grad_norm": 0.9879783500847374, "learning_rate": 1.1876855100134024e-05, "loss": 0.7341, "step": 12610 }, { "epoch": 0.4567547989858747, "grad_norm": 1.2330696776083856, "learning_rate": 1.187570286588866e-05, "loss": 0.7567, "step": 12611 }, { "epoch": 0.45679101774719305, "grad_norm": 1.4527122106089625, "learning_rate": 1.1874550605831982e-05, "loss": 0.7496, "step": 12612 }, { "epoch": 0.4568272365085114, "grad_norm": 1.5553174946892367, "learning_rate": 1.1873398319979837e-05, "loss": 0.7696, "step": 12613 }, { "epoch": 0.4568634552698298, "grad_norm": 1.3186265660548857, "learning_rate": 1.1872246008348088e-05, "loss": 0.7024, "step": 12614 }, { "epoch": 0.4568996740311481, "grad_norm": 1.0690935420823298, "learning_rate": 1.1871093670952585e-05, "loss": 0.7341, "step": 12615 }, { "epoch": 0.4569358927924665, "grad_norm": 1.4365821031563129, "learning_rate": 1.1869941307809191e-05, "loss": 0.8123, "step": 12616 }, { "epoch": 0.45697211155378487, "grad_norm": 1.320837518177588, "learning_rate": 1.186878891893376e-05, "loss": 0.8293, "step": 12617 }, { "epoch": 0.45700833031510324, "grad_norm": 1.4674163543897092, "learning_rate": 1.1867636504342155e-05, "loss": 0.7555, "step": 12618 }, { "epoch": 0.4570445490764216, "grad_norm": 1.457050561137743, "learning_rate": 1.1866484064050228e-05, "loss": 0.7566, "step": 12619 }, { "epoch": 0.45708076783773993, "grad_norm": 1.306514927815291, "learning_rate": 1.186533159807384e-05, "loss": 0.7507, "step": 12620 }, { "epoch": 0.4571169865990583, "grad_norm": 1.4939180925243196, "learning_rate": 1.186417910642885e-05, "loss": 0.8349, "step": 12621 }, { "epoch": 0.4571532053603767, "grad_norm": 1.4883945890008008, "learning_rate": 1.1863026589131118e-05, "loss": 0.7252, "step": 12622 }, { "epoch": 0.45718942412169505, "grad_norm": 1.3071281296304156, "learning_rate": 1.1861874046196504e-05, "loss": 0.7557, "step": 12623 }, { "epoch": 0.45722564288301343, "grad_norm": 1.230753858777732, "learning_rate": 1.1860721477640866e-05, "loss": 0.7445, "step": 12624 }, { "epoch": 0.45726186164433175, "grad_norm": 1.5885539739684411, "learning_rate": 1.1859568883480067e-05, "loss": 0.8329, "step": 12625 }, { "epoch": 0.4572980804056501, "grad_norm": 1.2094602326242863, "learning_rate": 1.1858416263729965e-05, "loss": 0.7647, "step": 12626 }, { "epoch": 0.4573342991669685, "grad_norm": 1.3333806155868717, "learning_rate": 1.1857263618406423e-05, "loss": 0.7835, "step": 12627 }, { "epoch": 0.45737051792828687, "grad_norm": 1.3268512529230527, "learning_rate": 1.18561109475253e-05, "loss": 0.65, "step": 12628 }, { "epoch": 0.45740673668960524, "grad_norm": 1.4578247536668616, "learning_rate": 1.1854958251102459e-05, "loss": 0.694, "step": 12629 }, { "epoch": 0.45744295545092356, "grad_norm": 1.4797372671658373, "learning_rate": 1.1853805529153766e-05, "loss": 0.8209, "step": 12630 }, { "epoch": 0.45747917421224193, "grad_norm": 1.6103466327654241, "learning_rate": 1.1852652781695078e-05, "loss": 0.7439, "step": 12631 }, { "epoch": 0.4575153929735603, "grad_norm": 1.5842598911846173, "learning_rate": 1.1851500008742261e-05, "loss": 0.8316, "step": 12632 }, { "epoch": 0.4575516117348787, "grad_norm": 1.3845700531419964, "learning_rate": 1.1850347210311175e-05, "loss": 0.7829, "step": 12633 }, { "epoch": 0.45758783049619706, "grad_norm": 1.2708066256119122, "learning_rate": 1.1849194386417686e-05, "loss": 0.7045, "step": 12634 }, { "epoch": 0.4576240492575154, "grad_norm": 1.275446715297187, "learning_rate": 1.1848041537077657e-05, "loss": 0.6981, "step": 12635 }, { "epoch": 0.45766026801883375, "grad_norm": 1.3374671715330226, "learning_rate": 1.1846888662306954e-05, "loss": 0.7694, "step": 12636 }, { "epoch": 0.4576964867801521, "grad_norm": 1.2139375095452374, "learning_rate": 1.184573576212144e-05, "loss": 0.7739, "step": 12637 }, { "epoch": 0.4577327055414705, "grad_norm": 1.432510297879308, "learning_rate": 1.1844582836536982e-05, "loss": 0.7795, "step": 12638 }, { "epoch": 0.45776892430278887, "grad_norm": 1.3410783249953415, "learning_rate": 1.184342988556944e-05, "loss": 0.7072, "step": 12639 }, { "epoch": 0.4578051430641072, "grad_norm": 1.1023134274106927, "learning_rate": 1.1842276909234682e-05, "loss": 0.7291, "step": 12640 }, { "epoch": 0.45784136182542556, "grad_norm": 1.4629290675013014, "learning_rate": 1.184112390754858e-05, "loss": 0.7208, "step": 12641 }, { "epoch": 0.45787758058674394, "grad_norm": 1.3920819092641172, "learning_rate": 1.183997088052699e-05, "loss": 0.7288, "step": 12642 }, { "epoch": 0.4579137993480623, "grad_norm": 1.151030159344741, "learning_rate": 1.1838817828185787e-05, "loss": 0.7559, "step": 12643 }, { "epoch": 0.4579500181093807, "grad_norm": 1.1224447534766606, "learning_rate": 1.1837664750540834e-05, "loss": 0.7962, "step": 12644 }, { "epoch": 0.457986236870699, "grad_norm": 1.2526489955793867, "learning_rate": 1.1836511647608e-05, "loss": 0.7526, "step": 12645 }, { "epoch": 0.4580224556320174, "grad_norm": 1.4281570745555783, "learning_rate": 1.1835358519403153e-05, "loss": 0.7358, "step": 12646 }, { "epoch": 0.45805867439333575, "grad_norm": 1.3683277007762213, "learning_rate": 1.1834205365942158e-05, "loss": 0.7266, "step": 12647 }, { "epoch": 0.4580948931546541, "grad_norm": 1.2877733011697223, "learning_rate": 1.1833052187240886e-05, "loss": 0.6965, "step": 12648 }, { "epoch": 0.4581311119159725, "grad_norm": 1.4006912382303294, "learning_rate": 1.1831898983315206e-05, "loss": 0.7706, "step": 12649 }, { "epoch": 0.4581673306772908, "grad_norm": 1.1260932724781183, "learning_rate": 1.1830745754180985e-05, "loss": 0.7483, "step": 12650 }, { "epoch": 0.4582035494386092, "grad_norm": 1.1973202819409317, "learning_rate": 1.1829592499854094e-05, "loss": 0.7657, "step": 12651 }, { "epoch": 0.45823976819992757, "grad_norm": 1.2927539451087446, "learning_rate": 1.1828439220350405e-05, "loss": 0.7518, "step": 12652 }, { "epoch": 0.45827598696124594, "grad_norm": 1.4118254463173934, "learning_rate": 1.1827285915685783e-05, "loss": 0.7943, "step": 12653 }, { "epoch": 0.4583122057225643, "grad_norm": 0.9940614215876703, "learning_rate": 1.1826132585876102e-05, "loss": 0.7664, "step": 12654 }, { "epoch": 0.45834842448388263, "grad_norm": 1.0668382438306536, "learning_rate": 1.1824979230937233e-05, "loss": 0.7951, "step": 12655 }, { "epoch": 0.458384643245201, "grad_norm": 1.388356042778194, "learning_rate": 1.1823825850885045e-05, "loss": 0.7782, "step": 12656 }, { "epoch": 0.4584208620065194, "grad_norm": 1.5060293396103026, "learning_rate": 1.182267244573541e-05, "loss": 0.7367, "step": 12657 }, { "epoch": 0.45845708076783775, "grad_norm": 1.406495164984403, "learning_rate": 1.1821519015504203e-05, "loss": 0.7471, "step": 12658 }, { "epoch": 0.45849329952915613, "grad_norm": 1.3455138004096092, "learning_rate": 1.1820365560207292e-05, "loss": 0.818, "step": 12659 }, { "epoch": 0.45852951829047445, "grad_norm": 1.4110522276892268, "learning_rate": 1.1819212079860551e-05, "loss": 0.7714, "step": 12660 }, { "epoch": 0.4585657370517928, "grad_norm": 1.4105116945695195, "learning_rate": 1.1818058574479855e-05, "loss": 0.7552, "step": 12661 }, { "epoch": 0.4586019558131112, "grad_norm": 1.4304702158763631, "learning_rate": 1.1816905044081076e-05, "loss": 0.8209, "step": 12662 }, { "epoch": 0.45863817457442957, "grad_norm": 1.4181153902045143, "learning_rate": 1.1815751488680084e-05, "loss": 0.8261, "step": 12663 }, { "epoch": 0.45867439333574794, "grad_norm": 1.8637495522031622, "learning_rate": 1.1814597908292757e-05, "loss": 0.7595, "step": 12664 }, { "epoch": 0.45871061209706626, "grad_norm": 1.3554819096797663, "learning_rate": 1.181344430293497e-05, "loss": 0.6529, "step": 12665 }, { "epoch": 0.45874683085838464, "grad_norm": 1.3013642210442085, "learning_rate": 1.1812290672622595e-05, "loss": 0.7173, "step": 12666 }, { "epoch": 0.458783049619703, "grad_norm": 1.3170620497370742, "learning_rate": 1.181113701737151e-05, "loss": 0.7784, "step": 12667 }, { "epoch": 0.4588192683810214, "grad_norm": 1.4068839280145151, "learning_rate": 1.1809983337197586e-05, "loss": 0.7268, "step": 12668 }, { "epoch": 0.45885548714233976, "grad_norm": 1.3617920680177624, "learning_rate": 1.1808829632116703e-05, "loss": 0.7575, "step": 12669 }, { "epoch": 0.4588917059036581, "grad_norm": 1.1393695331073914, "learning_rate": 1.1807675902144731e-05, "loss": 0.8045, "step": 12670 }, { "epoch": 0.45892792466497645, "grad_norm": 1.618814112201885, "learning_rate": 1.1806522147297552e-05, "loss": 0.9267, "step": 12671 }, { "epoch": 0.4589641434262948, "grad_norm": 1.1230242725875932, "learning_rate": 1.1805368367591043e-05, "loss": 0.7668, "step": 12672 }, { "epoch": 0.4590003621876132, "grad_norm": 1.4907837063995328, "learning_rate": 1.1804214563041077e-05, "loss": 0.7502, "step": 12673 }, { "epoch": 0.45903658094893157, "grad_norm": 1.399599988023824, "learning_rate": 1.1803060733663535e-05, "loss": 0.8167, "step": 12674 }, { "epoch": 0.4590727997102499, "grad_norm": 1.4358483964506668, "learning_rate": 1.180190687947429e-05, "loss": 0.8019, "step": 12675 }, { "epoch": 0.45910901847156826, "grad_norm": 1.1151088156538544, "learning_rate": 1.1800753000489225e-05, "loss": 0.6834, "step": 12676 }, { "epoch": 0.45914523723288664, "grad_norm": 1.2014525225216155, "learning_rate": 1.1799599096724215e-05, "loss": 0.7072, "step": 12677 }, { "epoch": 0.459181455994205, "grad_norm": 0.9526222047098515, "learning_rate": 1.1798445168195141e-05, "loss": 0.7499, "step": 12678 }, { "epoch": 0.4592176747555234, "grad_norm": 1.6204460021910572, "learning_rate": 1.1797291214917882e-05, "loss": 0.7904, "step": 12679 }, { "epoch": 0.4592538935168417, "grad_norm": 1.1659378327525072, "learning_rate": 1.1796137236908315e-05, "loss": 0.7006, "step": 12680 }, { "epoch": 0.4592901122781601, "grad_norm": 1.3486739434800536, "learning_rate": 1.1794983234182323e-05, "loss": 0.649, "step": 12681 }, { "epoch": 0.45932633103947845, "grad_norm": 0.9431684876697772, "learning_rate": 1.1793829206755782e-05, "loss": 0.6572, "step": 12682 }, { "epoch": 0.4593625498007968, "grad_norm": 1.3530098154398051, "learning_rate": 1.179267515464458e-05, "loss": 0.7568, "step": 12683 }, { "epoch": 0.4593987685621152, "grad_norm": 1.359720564361075, "learning_rate": 1.1791521077864587e-05, "loss": 0.7549, "step": 12684 }, { "epoch": 0.4594349873234335, "grad_norm": 1.0727230357002695, "learning_rate": 1.1790366976431692e-05, "loss": 0.7407, "step": 12685 }, { "epoch": 0.4594712060847519, "grad_norm": 1.2742605233785582, "learning_rate": 1.1789212850361775e-05, "loss": 0.776, "step": 12686 }, { "epoch": 0.45950742484607027, "grad_norm": 1.1727455679469767, "learning_rate": 1.1788058699670719e-05, "loss": 0.7626, "step": 12687 }, { "epoch": 0.45954364360738864, "grad_norm": 1.019199045968407, "learning_rate": 1.1786904524374401e-05, "loss": 0.694, "step": 12688 }, { "epoch": 0.459579862368707, "grad_norm": 1.433447193689296, "learning_rate": 1.1785750324488708e-05, "loss": 0.7999, "step": 12689 }, { "epoch": 0.45961608113002533, "grad_norm": 1.108187821570918, "learning_rate": 1.1784596100029519e-05, "loss": 0.6747, "step": 12690 }, { "epoch": 0.4596522998913437, "grad_norm": 1.2797330927965294, "learning_rate": 1.1783441851012722e-05, "loss": 0.693, "step": 12691 }, { "epoch": 0.4596885186526621, "grad_norm": 1.3259929359291562, "learning_rate": 1.1782287577454199e-05, "loss": 0.7478, "step": 12692 }, { "epoch": 0.45972473741398046, "grad_norm": 1.4113018845379013, "learning_rate": 1.1781133279369831e-05, "loss": 0.7852, "step": 12693 }, { "epoch": 0.45976095617529883, "grad_norm": 4.245277663699763, "learning_rate": 1.1779978956775507e-05, "loss": 0.7848, "step": 12694 }, { "epoch": 0.45979717493661715, "grad_norm": 1.1758372623782842, "learning_rate": 1.1778824609687103e-05, "loss": 0.7148, "step": 12695 }, { "epoch": 0.4598333936979355, "grad_norm": 1.0850603554239013, "learning_rate": 1.1777670238120513e-05, "loss": 0.7436, "step": 12696 }, { "epoch": 0.4598696124592539, "grad_norm": 1.4094374806509224, "learning_rate": 1.1776515842091619e-05, "loss": 0.7738, "step": 12697 }, { "epoch": 0.45990583122057227, "grad_norm": 1.4361297384307263, "learning_rate": 1.1775361421616303e-05, "loss": 0.7532, "step": 12698 }, { "epoch": 0.45994204998189064, "grad_norm": 1.3258424948200564, "learning_rate": 1.1774206976710455e-05, "loss": 0.7441, "step": 12699 }, { "epoch": 0.45997826874320896, "grad_norm": 1.4752285282754696, "learning_rate": 1.1773052507389962e-05, "loss": 0.7627, "step": 12700 }, { "epoch": 0.46001448750452734, "grad_norm": 1.694526970301408, "learning_rate": 1.1771898013670708e-05, "loss": 0.7425, "step": 12701 }, { "epoch": 0.4600507062658457, "grad_norm": 1.3033960329986451, "learning_rate": 1.1770743495568579e-05, "loss": 0.7653, "step": 12702 }, { "epoch": 0.4600869250271641, "grad_norm": 1.3664627756539567, "learning_rate": 1.1769588953099464e-05, "loss": 0.6936, "step": 12703 }, { "epoch": 0.46012314378848246, "grad_norm": 1.5018461186818488, "learning_rate": 1.176843438627925e-05, "loss": 0.8, "step": 12704 }, { "epoch": 0.4601593625498008, "grad_norm": 1.4676853440295736, "learning_rate": 1.1767279795123827e-05, "loss": 0.719, "step": 12705 }, { "epoch": 0.46019558131111915, "grad_norm": 1.6336684515943054, "learning_rate": 1.1766125179649079e-05, "loss": 0.7089, "step": 12706 }, { "epoch": 0.4602318000724375, "grad_norm": 1.2469941409453944, "learning_rate": 1.1764970539870899e-05, "loss": 0.7863, "step": 12707 }, { "epoch": 0.4602680188337559, "grad_norm": 1.3851899222010395, "learning_rate": 1.176381587580517e-05, "loss": 0.7604, "step": 12708 }, { "epoch": 0.4603042375950743, "grad_norm": 1.355091925496051, "learning_rate": 1.1762661187467785e-05, "loss": 0.7651, "step": 12709 }, { "epoch": 0.4603404563563926, "grad_norm": 1.4476630711484846, "learning_rate": 1.1761506474874635e-05, "loss": 0.7709, "step": 12710 }, { "epoch": 0.46037667511771097, "grad_norm": 1.014817376272202, "learning_rate": 1.1760351738041609e-05, "loss": 0.6953, "step": 12711 }, { "epoch": 0.46041289387902934, "grad_norm": 1.0381151507664055, "learning_rate": 1.1759196976984595e-05, "loss": 0.7392, "step": 12712 }, { "epoch": 0.4604491126403477, "grad_norm": 1.4289843961809443, "learning_rate": 1.1758042191719485e-05, "loss": 0.7629, "step": 12713 }, { "epoch": 0.4604853314016661, "grad_norm": 1.328744850247399, "learning_rate": 1.175688738226217e-05, "loss": 0.7347, "step": 12714 }, { "epoch": 0.4605215501629844, "grad_norm": 1.3571833750921003, "learning_rate": 1.175573254862854e-05, "loss": 0.792, "step": 12715 }, { "epoch": 0.4605577689243028, "grad_norm": 1.2011692692565565, "learning_rate": 1.1754577690834488e-05, "loss": 0.7657, "step": 12716 }, { "epoch": 0.46059398768562115, "grad_norm": 1.4564103319659087, "learning_rate": 1.1753422808895904e-05, "loss": 0.7459, "step": 12717 }, { "epoch": 0.4606302064469395, "grad_norm": 1.142892852967379, "learning_rate": 1.1752267902828682e-05, "loss": 0.7967, "step": 12718 }, { "epoch": 0.4606664252082579, "grad_norm": 1.423385388412842, "learning_rate": 1.1751112972648717e-05, "loss": 0.7195, "step": 12719 }, { "epoch": 0.4607026439695762, "grad_norm": 1.4387566280233035, "learning_rate": 1.1749958018371895e-05, "loss": 0.7495, "step": 12720 }, { "epoch": 0.4607388627308946, "grad_norm": 1.366690178064753, "learning_rate": 1.1748803040014116e-05, "loss": 0.7947, "step": 12721 }, { "epoch": 0.46077508149221297, "grad_norm": 1.198234054134438, "learning_rate": 1.1747648037591268e-05, "loss": 0.7185, "step": 12722 }, { "epoch": 0.46081130025353134, "grad_norm": 1.2024872609923971, "learning_rate": 1.1746493011119248e-05, "loss": 0.6567, "step": 12723 }, { "epoch": 0.4608475190148497, "grad_norm": 1.4730282036784832, "learning_rate": 1.174533796061395e-05, "loss": 0.7907, "step": 12724 }, { "epoch": 0.46088373777616803, "grad_norm": 1.5123957515288882, "learning_rate": 1.1744182886091269e-05, "loss": 0.7759, "step": 12725 }, { "epoch": 0.4609199565374864, "grad_norm": 1.31594879100139, "learning_rate": 1.1743027787567099e-05, "loss": 0.7138, "step": 12726 }, { "epoch": 0.4609561752988048, "grad_norm": 1.6298902304579401, "learning_rate": 1.1741872665057334e-05, "loss": 0.7539, "step": 12727 }, { "epoch": 0.46099239406012316, "grad_norm": 1.3879156421809773, "learning_rate": 1.1740717518577869e-05, "loss": 0.737, "step": 12728 }, { "epoch": 0.46102861282144153, "grad_norm": 1.5866229873150983, "learning_rate": 1.1739562348144602e-05, "loss": 0.7317, "step": 12729 }, { "epoch": 0.46106483158275985, "grad_norm": 1.3604157316979686, "learning_rate": 1.173840715377343e-05, "loss": 0.7413, "step": 12730 }, { "epoch": 0.4611010503440782, "grad_norm": 1.1715793834792134, "learning_rate": 1.1737251935480247e-05, "loss": 0.7326, "step": 12731 }, { "epoch": 0.4611372691053966, "grad_norm": 1.040577481040652, "learning_rate": 1.1736096693280953e-05, "loss": 0.719, "step": 12732 }, { "epoch": 0.46117348786671497, "grad_norm": 1.4663153745714856, "learning_rate": 1.1734941427191442e-05, "loss": 0.8681, "step": 12733 }, { "epoch": 0.46120970662803334, "grad_norm": 1.03916087527412, "learning_rate": 1.1733786137227612e-05, "loss": 0.7727, "step": 12734 }, { "epoch": 0.46124592538935166, "grad_norm": 1.42814849530557, "learning_rate": 1.1732630823405362e-05, "loss": 0.7734, "step": 12735 }, { "epoch": 0.46128214415067004, "grad_norm": 1.4243914336281331, "learning_rate": 1.1731475485740589e-05, "loss": 0.8025, "step": 12736 }, { "epoch": 0.4613183629119884, "grad_norm": 1.4149552005166768, "learning_rate": 1.1730320124249191e-05, "loss": 0.789, "step": 12737 }, { "epoch": 0.4613545816733068, "grad_norm": 1.3493318092895503, "learning_rate": 1.172916473894707e-05, "loss": 0.7037, "step": 12738 }, { "epoch": 0.46139080043462516, "grad_norm": 1.4141645519830246, "learning_rate": 1.172800932985012e-05, "loss": 0.7917, "step": 12739 }, { "epoch": 0.4614270191959435, "grad_norm": 1.4737395417915755, "learning_rate": 1.1726853896974245e-05, "loss": 0.8117, "step": 12740 }, { "epoch": 0.46146323795726185, "grad_norm": 1.3161585798377204, "learning_rate": 1.1725698440335343e-05, "loss": 0.7799, "step": 12741 }, { "epoch": 0.4614994567185802, "grad_norm": 1.3396675832336613, "learning_rate": 1.1724542959949313e-05, "loss": 0.7367, "step": 12742 }, { "epoch": 0.4615356754798986, "grad_norm": 1.508062513654036, "learning_rate": 1.1723387455832058e-05, "loss": 0.7286, "step": 12743 }, { "epoch": 0.461571894241217, "grad_norm": 1.3189721809586792, "learning_rate": 1.1722231927999474e-05, "loss": 0.7671, "step": 12744 }, { "epoch": 0.4616081130025353, "grad_norm": 1.4768041092631305, "learning_rate": 1.1721076376467471e-05, "loss": 0.7661, "step": 12745 }, { "epoch": 0.46164433176385367, "grad_norm": 1.2967059940209913, "learning_rate": 1.171992080125194e-05, "loss": 0.7395, "step": 12746 }, { "epoch": 0.46168055052517204, "grad_norm": 1.3082809294968234, "learning_rate": 1.1718765202368791e-05, "loss": 0.7514, "step": 12747 }, { "epoch": 0.4617167692864904, "grad_norm": 1.4352701024127121, "learning_rate": 1.171760957983392e-05, "loss": 0.8071, "step": 12748 }, { "epoch": 0.4617529880478088, "grad_norm": 1.3274418504419974, "learning_rate": 1.171645393366323e-05, "loss": 0.716, "step": 12749 }, { "epoch": 0.4617892068091271, "grad_norm": 1.4790144089092583, "learning_rate": 1.1715298263872632e-05, "loss": 0.8539, "step": 12750 }, { "epoch": 0.4618254255704455, "grad_norm": 1.0481958736834305, "learning_rate": 1.1714142570478018e-05, "loss": 0.7721, "step": 12751 }, { "epoch": 0.46186164433176385, "grad_norm": 1.3580309448236139, "learning_rate": 1.1712986853495298e-05, "loss": 0.6679, "step": 12752 }, { "epoch": 0.46189786309308223, "grad_norm": 1.0058716372210077, "learning_rate": 1.1711831112940371e-05, "loss": 0.8315, "step": 12753 }, { "epoch": 0.4619340818544006, "grad_norm": 1.111190480447135, "learning_rate": 1.1710675348829148e-05, "loss": 0.7445, "step": 12754 }, { "epoch": 0.4619703006157189, "grad_norm": 1.6203052832351694, "learning_rate": 1.1709519561177524e-05, "loss": 0.7533, "step": 12755 }, { "epoch": 0.4620065193770373, "grad_norm": 1.5671974892549883, "learning_rate": 1.1708363750001415e-05, "loss": 0.7412, "step": 12756 }, { "epoch": 0.46204273813835567, "grad_norm": 1.6801215830727658, "learning_rate": 1.1707207915316714e-05, "loss": 0.8344, "step": 12757 }, { "epoch": 0.46207895689967404, "grad_norm": 1.3377101971442444, "learning_rate": 1.1706052057139335e-05, "loss": 0.6976, "step": 12758 }, { "epoch": 0.4621151756609924, "grad_norm": 1.0502660928352976, "learning_rate": 1.1704896175485179e-05, "loss": 0.7517, "step": 12759 }, { "epoch": 0.46215139442231074, "grad_norm": 1.0733939007527313, "learning_rate": 1.1703740270370154e-05, "loss": 0.744, "step": 12760 }, { "epoch": 0.4621876131836291, "grad_norm": 1.4028934262175448, "learning_rate": 1.1702584341810166e-05, "loss": 0.7801, "step": 12761 }, { "epoch": 0.4622238319449475, "grad_norm": 1.3474805731226698, "learning_rate": 1.1701428389821121e-05, "loss": 0.7517, "step": 12762 }, { "epoch": 0.46226005070626586, "grad_norm": 1.4196020498134527, "learning_rate": 1.1700272414418929e-05, "loss": 0.7202, "step": 12763 }, { "epoch": 0.46229626946758423, "grad_norm": 1.022594000224585, "learning_rate": 1.1699116415619492e-05, "loss": 0.7374, "step": 12764 }, { "epoch": 0.46233248822890255, "grad_norm": 1.4504182958801513, "learning_rate": 1.1697960393438721e-05, "loss": 0.7574, "step": 12765 }, { "epoch": 0.4623687069902209, "grad_norm": 1.287232519721808, "learning_rate": 1.1696804347892524e-05, "loss": 0.7537, "step": 12766 }, { "epoch": 0.4624049257515393, "grad_norm": 1.3399576258298027, "learning_rate": 1.1695648278996808e-05, "loss": 0.7393, "step": 12767 }, { "epoch": 0.46244114451285767, "grad_norm": 1.4528356284509643, "learning_rate": 1.169449218676748e-05, "loss": 0.7875, "step": 12768 }, { "epoch": 0.46247736327417605, "grad_norm": 1.5081168532181122, "learning_rate": 1.1693336071220452e-05, "loss": 0.7494, "step": 12769 }, { "epoch": 0.46251358203549436, "grad_norm": 1.5892472401698956, "learning_rate": 1.169217993237163e-05, "loss": 0.7493, "step": 12770 }, { "epoch": 0.46254980079681274, "grad_norm": 1.355911004092663, "learning_rate": 1.1691023770236925e-05, "loss": 0.6843, "step": 12771 }, { "epoch": 0.4625860195581311, "grad_norm": 1.1326774283360914, "learning_rate": 1.1689867584832248e-05, "loss": 0.7459, "step": 12772 }, { "epoch": 0.4626222383194495, "grad_norm": 1.0976162969794756, "learning_rate": 1.1688711376173509e-05, "loss": 0.6919, "step": 12773 }, { "epoch": 0.46265845708076786, "grad_norm": 1.5995378207527036, "learning_rate": 1.168755514427662e-05, "loss": 0.7163, "step": 12774 }, { "epoch": 0.4626946758420862, "grad_norm": 1.3917207854213116, "learning_rate": 1.1686398889157485e-05, "loss": 0.7594, "step": 12775 }, { "epoch": 0.46273089460340455, "grad_norm": 1.425649141043906, "learning_rate": 1.1685242610832022e-05, "loss": 0.7302, "step": 12776 }, { "epoch": 0.4627671133647229, "grad_norm": 1.2596450485514896, "learning_rate": 1.1684086309316136e-05, "loss": 0.6994, "step": 12777 }, { "epoch": 0.4628033321260413, "grad_norm": 1.232998206894136, "learning_rate": 1.1682929984625749e-05, "loss": 0.7584, "step": 12778 }, { "epoch": 0.4628395508873597, "grad_norm": 1.3686137185445746, "learning_rate": 1.168177363677676e-05, "loss": 0.7633, "step": 12779 }, { "epoch": 0.462875769648678, "grad_norm": 1.2807340383538006, "learning_rate": 1.1680617265785093e-05, "loss": 0.7089, "step": 12780 }, { "epoch": 0.46291198840999637, "grad_norm": 1.2898443932101324, "learning_rate": 1.1679460871666652e-05, "loss": 0.8074, "step": 12781 }, { "epoch": 0.46294820717131474, "grad_norm": 1.407223826557282, "learning_rate": 1.1678304454437358e-05, "loss": 0.8055, "step": 12782 }, { "epoch": 0.4629844259326331, "grad_norm": 1.3676717963087432, "learning_rate": 1.1677148014113117e-05, "loss": 0.7596, "step": 12783 }, { "epoch": 0.4630206446939515, "grad_norm": 1.4143067657402164, "learning_rate": 1.1675991550709846e-05, "loss": 0.7897, "step": 12784 }, { "epoch": 0.4630568634552698, "grad_norm": 1.5586966830756346, "learning_rate": 1.1674835064243459e-05, "loss": 0.7183, "step": 12785 }, { "epoch": 0.4630930822165882, "grad_norm": 1.4527799314539567, "learning_rate": 1.1673678554729871e-05, "loss": 0.7907, "step": 12786 }, { "epoch": 0.46312930097790656, "grad_norm": 1.2383973160509325, "learning_rate": 1.1672522022184993e-05, "loss": 0.7245, "step": 12787 }, { "epoch": 0.46316551973922493, "grad_norm": 1.2268231907957061, "learning_rate": 1.1671365466624744e-05, "loss": 0.7502, "step": 12788 }, { "epoch": 0.4632017385005433, "grad_norm": 1.3687989666396132, "learning_rate": 1.1670208888065038e-05, "loss": 0.8, "step": 12789 }, { "epoch": 0.4632379572618616, "grad_norm": 1.4009202530224938, "learning_rate": 1.166905228652179e-05, "loss": 0.7382, "step": 12790 }, { "epoch": 0.46327417602318, "grad_norm": 1.1758341951584492, "learning_rate": 1.1667895662010913e-05, "loss": 0.6812, "step": 12791 }, { "epoch": 0.46331039478449837, "grad_norm": 0.9687494216485084, "learning_rate": 1.166673901454833e-05, "loss": 0.7234, "step": 12792 }, { "epoch": 0.46334661354581674, "grad_norm": 1.261503447772787, "learning_rate": 1.1665582344149952e-05, "loss": 0.7008, "step": 12793 }, { "epoch": 0.4633828323071351, "grad_norm": 1.440491340740196, "learning_rate": 1.1664425650831698e-05, "loss": 0.7594, "step": 12794 }, { "epoch": 0.46341905106845344, "grad_norm": 1.4004857893521887, "learning_rate": 1.1663268934609484e-05, "loss": 0.7715, "step": 12795 }, { "epoch": 0.4634552698297718, "grad_norm": 1.6213894447984292, "learning_rate": 1.1662112195499229e-05, "loss": 0.7191, "step": 12796 }, { "epoch": 0.4634914885910902, "grad_norm": 1.3557464132452761, "learning_rate": 1.1660955433516849e-05, "loss": 0.6725, "step": 12797 }, { "epoch": 0.46352770735240856, "grad_norm": 1.2986886838763638, "learning_rate": 1.1659798648678262e-05, "loss": 0.7157, "step": 12798 }, { "epoch": 0.46356392611372693, "grad_norm": 0.9999164222970901, "learning_rate": 1.165864184099939e-05, "loss": 0.8075, "step": 12799 }, { "epoch": 0.46360014487504525, "grad_norm": 1.3017963151896792, "learning_rate": 1.1657485010496146e-05, "loss": 0.7861, "step": 12800 }, { "epoch": 0.4636363636363636, "grad_norm": 1.3796367230095277, "learning_rate": 1.1656328157184453e-05, "loss": 0.7773, "step": 12801 }, { "epoch": 0.463672582397682, "grad_norm": 1.386662604310211, "learning_rate": 1.1655171281080229e-05, "loss": 0.6866, "step": 12802 }, { "epoch": 0.46370880115900037, "grad_norm": 1.4226735081534516, "learning_rate": 1.1654014382199395e-05, "loss": 0.7628, "step": 12803 }, { "epoch": 0.46374501992031875, "grad_norm": 1.354804968250823, "learning_rate": 1.1652857460557867e-05, "loss": 0.7723, "step": 12804 }, { "epoch": 0.46378123868163706, "grad_norm": 1.2757606831146655, "learning_rate": 1.1651700516171569e-05, "loss": 0.684, "step": 12805 }, { "epoch": 0.46381745744295544, "grad_norm": 1.4259409312096805, "learning_rate": 1.1650543549056422e-05, "loss": 0.8075, "step": 12806 }, { "epoch": 0.4638536762042738, "grad_norm": 1.1382865208711248, "learning_rate": 1.1649386559228342e-05, "loss": 0.7993, "step": 12807 }, { "epoch": 0.4638898949655922, "grad_norm": 1.0715514640182098, "learning_rate": 1.1648229546703258e-05, "loss": 0.7258, "step": 12808 }, { "epoch": 0.46392611372691056, "grad_norm": 1.4972680266749434, "learning_rate": 1.1647072511497084e-05, "loss": 0.7544, "step": 12809 }, { "epoch": 0.4639623324882289, "grad_norm": 1.0263765374469174, "learning_rate": 1.1645915453625746e-05, "loss": 0.7095, "step": 12810 }, { "epoch": 0.46399855124954725, "grad_norm": 1.288685411149157, "learning_rate": 1.1644758373105165e-05, "loss": 0.746, "step": 12811 }, { "epoch": 0.4640347700108656, "grad_norm": 0.9799591194225986, "learning_rate": 1.1643601269951266e-05, "loss": 0.7914, "step": 12812 }, { "epoch": 0.464070988772184, "grad_norm": 1.573122923371858, "learning_rate": 1.1642444144179965e-05, "loss": 0.7653, "step": 12813 }, { "epoch": 0.4641072075335024, "grad_norm": 1.325085118827853, "learning_rate": 1.1641286995807193e-05, "loss": 0.7124, "step": 12814 }, { "epoch": 0.4641434262948207, "grad_norm": 1.0029738565000779, "learning_rate": 1.164012982484887e-05, "loss": 0.7905, "step": 12815 }, { "epoch": 0.46417964505613907, "grad_norm": 1.1397749849762444, "learning_rate": 1.1638972631320916e-05, "loss": 0.7302, "step": 12816 }, { "epoch": 0.46421586381745744, "grad_norm": 1.4331831149068734, "learning_rate": 1.163781541523926e-05, "loss": 0.7928, "step": 12817 }, { "epoch": 0.4642520825787758, "grad_norm": 1.2087533911950354, "learning_rate": 1.1636658176619824e-05, "loss": 0.714, "step": 12818 }, { "epoch": 0.4642883013400942, "grad_norm": 1.3942062576249004, "learning_rate": 1.1635500915478536e-05, "loss": 0.7736, "step": 12819 }, { "epoch": 0.4643245201014125, "grad_norm": 1.2830761849600674, "learning_rate": 1.1634343631831317e-05, "loss": 0.7322, "step": 12820 }, { "epoch": 0.4643607388627309, "grad_norm": 1.2583940521002779, "learning_rate": 1.1633186325694094e-05, "loss": 0.6569, "step": 12821 }, { "epoch": 0.46439695762404926, "grad_norm": 1.721637160565212, "learning_rate": 1.1632028997082792e-05, "loss": 0.7464, "step": 12822 }, { "epoch": 0.46443317638536763, "grad_norm": 1.5281734822584254, "learning_rate": 1.1630871646013335e-05, "loss": 0.6737, "step": 12823 }, { "epoch": 0.464469395146686, "grad_norm": 1.459993121580233, "learning_rate": 1.1629714272501653e-05, "loss": 0.8557, "step": 12824 }, { "epoch": 0.4645056139080043, "grad_norm": 1.2573191551921326, "learning_rate": 1.162855687656367e-05, "loss": 0.8035, "step": 12825 }, { "epoch": 0.4645418326693227, "grad_norm": 1.4091172353185757, "learning_rate": 1.1627399458215314e-05, "loss": 0.771, "step": 12826 }, { "epoch": 0.46457805143064107, "grad_norm": 1.1339216182021483, "learning_rate": 1.1626242017472511e-05, "loss": 0.7772, "step": 12827 }, { "epoch": 0.46461427019195944, "grad_norm": 1.4124294527653312, "learning_rate": 1.162508455435119e-05, "loss": 0.7625, "step": 12828 }, { "epoch": 0.4646504889532778, "grad_norm": 1.6363812482137385, "learning_rate": 1.1623927068867276e-05, "loss": 0.7438, "step": 12829 }, { "epoch": 0.46468670771459614, "grad_norm": 1.422119616070451, "learning_rate": 1.16227695610367e-05, "loss": 0.7913, "step": 12830 }, { "epoch": 0.4647229264759145, "grad_norm": 1.261691332303642, "learning_rate": 1.1621612030875386e-05, "loss": 0.6577, "step": 12831 }, { "epoch": 0.4647591452372329, "grad_norm": 2.1092999621572552, "learning_rate": 1.162045447839927e-05, "loss": 0.7797, "step": 12832 }, { "epoch": 0.46479536399855126, "grad_norm": 1.4690671892787128, "learning_rate": 1.1619296903624274e-05, "loss": 0.6875, "step": 12833 }, { "epoch": 0.46483158275986963, "grad_norm": 1.3178306324624751, "learning_rate": 1.1618139306566331e-05, "loss": 0.7101, "step": 12834 }, { "epoch": 0.46486780152118795, "grad_norm": 1.2147281436865098, "learning_rate": 1.1616981687241366e-05, "loss": 0.6865, "step": 12835 }, { "epoch": 0.4649040202825063, "grad_norm": 1.3476615076377145, "learning_rate": 1.1615824045665315e-05, "loss": 0.6872, "step": 12836 }, { "epoch": 0.4649402390438247, "grad_norm": 1.4505782626806634, "learning_rate": 1.1614666381854107e-05, "loss": 0.7867, "step": 12837 }, { "epoch": 0.4649764578051431, "grad_norm": 1.4338322786636728, "learning_rate": 1.1613508695823668e-05, "loss": 0.7533, "step": 12838 }, { "epoch": 0.46501267656646145, "grad_norm": 1.4464699350677284, "learning_rate": 1.1612350987589932e-05, "loss": 0.7756, "step": 12839 }, { "epoch": 0.46504889532777977, "grad_norm": 1.457728739281425, "learning_rate": 1.161119325716883e-05, "loss": 0.8039, "step": 12840 }, { "epoch": 0.46508511408909814, "grad_norm": 1.438395927173614, "learning_rate": 1.1610035504576293e-05, "loss": 0.6559, "step": 12841 }, { "epoch": 0.4651213328504165, "grad_norm": 1.2253241118746998, "learning_rate": 1.1608877729828252e-05, "loss": 0.7047, "step": 12842 }, { "epoch": 0.4651575516117349, "grad_norm": 1.0340953149499816, "learning_rate": 1.1607719932940643e-05, "loss": 0.7863, "step": 12843 }, { "epoch": 0.46519377037305326, "grad_norm": 1.433606968852537, "learning_rate": 1.1606562113929392e-05, "loss": 0.7947, "step": 12844 }, { "epoch": 0.4652299891343716, "grad_norm": 1.3891654049685365, "learning_rate": 1.1605404272810437e-05, "loss": 0.7367, "step": 12845 }, { "epoch": 0.46526620789568995, "grad_norm": 1.5280188235693066, "learning_rate": 1.1604246409599704e-05, "loss": 0.764, "step": 12846 }, { "epoch": 0.46530242665700833, "grad_norm": 1.3867314371683785, "learning_rate": 1.1603088524313135e-05, "loss": 0.7653, "step": 12847 }, { "epoch": 0.4653386454183267, "grad_norm": 1.2521476847352266, "learning_rate": 1.1601930616966656e-05, "loss": 0.6148, "step": 12848 }, { "epoch": 0.4653748641796451, "grad_norm": 1.2809566992197772, "learning_rate": 1.1600772687576207e-05, "loss": 0.7338, "step": 12849 }, { "epoch": 0.4654110829409634, "grad_norm": 1.0225906507259497, "learning_rate": 1.159961473615772e-05, "loss": 0.718, "step": 12850 }, { "epoch": 0.46544730170228177, "grad_norm": 1.4108941929107333, "learning_rate": 1.1598456762727127e-05, "loss": 0.7311, "step": 12851 }, { "epoch": 0.46548352046360014, "grad_norm": 1.4009903504115824, "learning_rate": 1.1597298767300365e-05, "loss": 0.7831, "step": 12852 }, { "epoch": 0.4655197392249185, "grad_norm": 1.4105293316675849, "learning_rate": 1.1596140749893366e-05, "loss": 0.7169, "step": 12853 }, { "epoch": 0.4655559579862369, "grad_norm": 1.4642065875010604, "learning_rate": 1.159498271052207e-05, "loss": 0.7786, "step": 12854 }, { "epoch": 0.4655921767475552, "grad_norm": 1.433061545119682, "learning_rate": 1.159382464920241e-05, "loss": 0.7484, "step": 12855 }, { "epoch": 0.4656283955088736, "grad_norm": 1.3876653455952248, "learning_rate": 1.1592666565950324e-05, "loss": 0.7126, "step": 12856 }, { "epoch": 0.46566461427019196, "grad_norm": 1.465933874764531, "learning_rate": 1.1591508460781743e-05, "loss": 0.7421, "step": 12857 }, { "epoch": 0.46570083303151033, "grad_norm": 1.3705044727195046, "learning_rate": 1.159035033371261e-05, "loss": 0.8161, "step": 12858 }, { "epoch": 0.4657370517928287, "grad_norm": 1.3956827423781106, "learning_rate": 1.1589192184758859e-05, "loss": 0.7503, "step": 12859 }, { "epoch": 0.465773270554147, "grad_norm": 1.1405551375935863, "learning_rate": 1.1588034013936427e-05, "loss": 0.73, "step": 12860 }, { "epoch": 0.4658094893154654, "grad_norm": 1.1852630410192444, "learning_rate": 1.1586875821261252e-05, "loss": 0.786, "step": 12861 }, { "epoch": 0.46584570807678377, "grad_norm": 1.487266886951496, "learning_rate": 1.158571760674927e-05, "loss": 0.7193, "step": 12862 }, { "epoch": 0.46588192683810215, "grad_norm": 1.3731229836979137, "learning_rate": 1.1584559370416422e-05, "loss": 0.7259, "step": 12863 }, { "epoch": 0.4659181455994205, "grad_norm": 1.473366577437886, "learning_rate": 1.158340111227864e-05, "loss": 0.85, "step": 12864 }, { "epoch": 0.46595436436073884, "grad_norm": 1.1457303164727528, "learning_rate": 1.1582242832351873e-05, "loss": 0.8084, "step": 12865 }, { "epoch": 0.4659905831220572, "grad_norm": 1.274498307523842, "learning_rate": 1.1581084530652051e-05, "loss": 0.7026, "step": 12866 }, { "epoch": 0.4660268018833756, "grad_norm": 1.2859382554651129, "learning_rate": 1.157992620719512e-05, "loss": 0.7476, "step": 12867 }, { "epoch": 0.46606302064469396, "grad_norm": 1.3906899253551392, "learning_rate": 1.1578767861997014e-05, "loss": 0.6981, "step": 12868 }, { "epoch": 0.46609923940601233, "grad_norm": 1.275762418355674, "learning_rate": 1.1577609495073673e-05, "loss": 0.6543, "step": 12869 }, { "epoch": 0.46613545816733065, "grad_norm": 1.1143015351752488, "learning_rate": 1.1576451106441043e-05, "loss": 0.7355, "step": 12870 }, { "epoch": 0.466171676928649, "grad_norm": 1.1162521939984293, "learning_rate": 1.1575292696115056e-05, "loss": 0.7809, "step": 12871 }, { "epoch": 0.4662078956899674, "grad_norm": 1.341681289414228, "learning_rate": 1.1574134264111662e-05, "loss": 0.6969, "step": 12872 }, { "epoch": 0.4662441144512858, "grad_norm": 0.9642329843515158, "learning_rate": 1.1572975810446796e-05, "loss": 0.7001, "step": 12873 }, { "epoch": 0.46628033321260415, "grad_norm": 1.3244440631143544, "learning_rate": 1.1571817335136402e-05, "loss": 0.7401, "step": 12874 }, { "epoch": 0.46631655197392247, "grad_norm": 1.4160191430672977, "learning_rate": 1.1570658838196416e-05, "loss": 0.8147, "step": 12875 }, { "epoch": 0.46635277073524084, "grad_norm": 1.4013811330068175, "learning_rate": 1.1569500319642789e-05, "loss": 0.7974, "step": 12876 }, { "epoch": 0.4663889894965592, "grad_norm": 1.3235136387112458, "learning_rate": 1.1568341779491454e-05, "loss": 0.6945, "step": 12877 }, { "epoch": 0.4664252082578776, "grad_norm": 1.3648344444652052, "learning_rate": 1.1567183217758363e-05, "loss": 0.7766, "step": 12878 }, { "epoch": 0.46646142701919596, "grad_norm": 1.3690227561202528, "learning_rate": 1.156602463445945e-05, "loss": 0.6842, "step": 12879 }, { "epoch": 0.4664976457805143, "grad_norm": 1.3746675826440258, "learning_rate": 1.1564866029610661e-05, "loss": 0.7122, "step": 12880 }, { "epoch": 0.46653386454183265, "grad_norm": 1.8388380553804016, "learning_rate": 1.1563707403227944e-05, "loss": 0.6516, "step": 12881 }, { "epoch": 0.46657008330315103, "grad_norm": 1.394244396936158, "learning_rate": 1.1562548755327236e-05, "loss": 0.7941, "step": 12882 }, { "epoch": 0.4666063020644694, "grad_norm": 1.2435094453979605, "learning_rate": 1.156139008592449e-05, "loss": 0.7091, "step": 12883 }, { "epoch": 0.4666425208257878, "grad_norm": 1.30157925997404, "learning_rate": 1.1560231395035638e-05, "loss": 0.6393, "step": 12884 }, { "epoch": 0.4666787395871061, "grad_norm": 1.3465240698425556, "learning_rate": 1.1559072682676635e-05, "loss": 0.7401, "step": 12885 }, { "epoch": 0.46671495834842447, "grad_norm": 1.0063883686090094, "learning_rate": 1.155791394886342e-05, "loss": 0.7113, "step": 12886 }, { "epoch": 0.46675117710974284, "grad_norm": 1.3545635630195263, "learning_rate": 1.1556755193611941e-05, "loss": 0.7552, "step": 12887 }, { "epoch": 0.4667873958710612, "grad_norm": 1.2602570085513194, "learning_rate": 1.1555596416938142e-05, "loss": 0.7614, "step": 12888 }, { "epoch": 0.4668236146323796, "grad_norm": 1.1287870418928543, "learning_rate": 1.1554437618857967e-05, "loss": 0.7358, "step": 12889 }, { "epoch": 0.4668598333936979, "grad_norm": 1.3069166617489218, "learning_rate": 1.155327879938737e-05, "loss": 0.74, "step": 12890 }, { "epoch": 0.4668960521550163, "grad_norm": 1.4089907547944058, "learning_rate": 1.1552119958542288e-05, "loss": 0.6477, "step": 12891 }, { "epoch": 0.46693227091633466, "grad_norm": 1.2931501197956334, "learning_rate": 1.1550961096338673e-05, "loss": 0.6968, "step": 12892 }, { "epoch": 0.46696848967765303, "grad_norm": 1.2499542846366736, "learning_rate": 1.1549802212792469e-05, "loss": 0.7241, "step": 12893 }, { "epoch": 0.4670047084389714, "grad_norm": 1.4205405698975242, "learning_rate": 1.1548643307919627e-05, "loss": 0.7196, "step": 12894 }, { "epoch": 0.4670409272002897, "grad_norm": 1.6511141065947468, "learning_rate": 1.1547484381736091e-05, "loss": 0.814, "step": 12895 }, { "epoch": 0.4670771459616081, "grad_norm": 1.1722069256309262, "learning_rate": 1.1546325434257812e-05, "loss": 0.7155, "step": 12896 }, { "epoch": 0.46711336472292647, "grad_norm": 1.280875889164739, "learning_rate": 1.1545166465500732e-05, "loss": 0.7581, "step": 12897 }, { "epoch": 0.46714958348424485, "grad_norm": 1.2714720998631497, "learning_rate": 1.1544007475480809e-05, "loss": 0.7705, "step": 12898 }, { "epoch": 0.4671858022455632, "grad_norm": 1.1736567964498092, "learning_rate": 1.1542848464213983e-05, "loss": 0.76, "step": 12899 }, { "epoch": 0.46722202100688154, "grad_norm": 0.9592542562334548, "learning_rate": 1.1541689431716208e-05, "loss": 0.7201, "step": 12900 }, { "epoch": 0.4672582397681999, "grad_norm": 1.486181281977991, "learning_rate": 1.1540530378003431e-05, "loss": 0.7797, "step": 12901 }, { "epoch": 0.4672944585295183, "grad_norm": 1.4216986182611513, "learning_rate": 1.1539371303091605e-05, "loss": 0.7879, "step": 12902 }, { "epoch": 0.46733067729083666, "grad_norm": 1.3148833438025325, "learning_rate": 1.1538212206996675e-05, "loss": 0.7892, "step": 12903 }, { "epoch": 0.46736689605215503, "grad_norm": 1.2942984700050642, "learning_rate": 1.153705308973459e-05, "loss": 0.7657, "step": 12904 }, { "epoch": 0.46740311481347335, "grad_norm": 1.4745669466192628, "learning_rate": 1.1535893951321309e-05, "loss": 0.7172, "step": 12905 }, { "epoch": 0.4674393335747917, "grad_norm": 1.4969331637452654, "learning_rate": 1.1534734791772776e-05, "loss": 0.7734, "step": 12906 }, { "epoch": 0.4674755523361101, "grad_norm": 1.0285640074128861, "learning_rate": 1.1533575611104943e-05, "loss": 0.7255, "step": 12907 }, { "epoch": 0.4675117710974285, "grad_norm": 1.2637925186230237, "learning_rate": 1.1532416409333764e-05, "loss": 0.7317, "step": 12908 }, { "epoch": 0.46754798985874685, "grad_norm": 1.329262795853597, "learning_rate": 1.1531257186475186e-05, "loss": 0.7138, "step": 12909 }, { "epoch": 0.46758420862006517, "grad_norm": 1.5722538112409792, "learning_rate": 1.1530097942545167e-05, "loss": 0.8265, "step": 12910 }, { "epoch": 0.46762042738138354, "grad_norm": 1.3107392998431788, "learning_rate": 1.1528938677559652e-05, "loss": 0.7684, "step": 12911 }, { "epoch": 0.4676566461427019, "grad_norm": 1.2869099518951566, "learning_rate": 1.1527779391534601e-05, "loss": 0.7363, "step": 12912 }, { "epoch": 0.4676928649040203, "grad_norm": 1.3635861153600326, "learning_rate": 1.152662008448596e-05, "loss": 0.7546, "step": 12913 }, { "epoch": 0.46772908366533866, "grad_norm": 1.2927001273598937, "learning_rate": 1.1525460756429687e-05, "loss": 0.7183, "step": 12914 }, { "epoch": 0.467765302426657, "grad_norm": 1.443024147294195, "learning_rate": 1.1524301407381733e-05, "loss": 0.7139, "step": 12915 }, { "epoch": 0.46780152118797536, "grad_norm": 1.2978037946649283, "learning_rate": 1.152314203735805e-05, "loss": 0.752, "step": 12916 }, { "epoch": 0.46783773994929373, "grad_norm": 1.377454229577113, "learning_rate": 1.15219826463746e-05, "loss": 0.6433, "step": 12917 }, { "epoch": 0.4678739587106121, "grad_norm": 1.1018270367679306, "learning_rate": 1.1520823234447327e-05, "loss": 0.7047, "step": 12918 }, { "epoch": 0.4679101774719305, "grad_norm": 1.3150391058990503, "learning_rate": 1.151966380159219e-05, "loss": 0.7964, "step": 12919 }, { "epoch": 0.4679463962332488, "grad_norm": 1.4504808119184296, "learning_rate": 1.1518504347825146e-05, "loss": 0.7612, "step": 12920 }, { "epoch": 0.46798261499456717, "grad_norm": 1.354354860375432, "learning_rate": 1.1517344873162148e-05, "loss": 0.793, "step": 12921 }, { "epoch": 0.46801883375588554, "grad_norm": 1.0125724132550544, "learning_rate": 1.1516185377619149e-05, "loss": 0.7515, "step": 12922 }, { "epoch": 0.4680550525172039, "grad_norm": 1.2806307174477376, "learning_rate": 1.151502586121211e-05, "loss": 0.7856, "step": 12923 }, { "epoch": 0.4680912712785223, "grad_norm": 1.147318512214169, "learning_rate": 1.1513866323956979e-05, "loss": 0.7002, "step": 12924 }, { "epoch": 0.4681274900398406, "grad_norm": 1.2231345140119294, "learning_rate": 1.1512706765869724e-05, "loss": 0.6909, "step": 12925 }, { "epoch": 0.468163708801159, "grad_norm": 1.5263034870720733, "learning_rate": 1.1511547186966287e-05, "loss": 0.7915, "step": 12926 }, { "epoch": 0.46819992756247736, "grad_norm": 1.2654135550960102, "learning_rate": 1.1510387587262637e-05, "loss": 0.6755, "step": 12927 }, { "epoch": 0.46823614632379573, "grad_norm": 1.061714887956289, "learning_rate": 1.1509227966774728e-05, "loss": 0.7361, "step": 12928 }, { "epoch": 0.4682723650851141, "grad_norm": 1.251289100824411, "learning_rate": 1.1508068325518514e-05, "loss": 0.6883, "step": 12929 }, { "epoch": 0.4683085838464324, "grad_norm": 1.2050301168978363, "learning_rate": 1.1506908663509954e-05, "loss": 0.7527, "step": 12930 }, { "epoch": 0.4683448026077508, "grad_norm": 1.493614912109634, "learning_rate": 1.1505748980765006e-05, "loss": 0.7669, "step": 12931 }, { "epoch": 0.4683810213690692, "grad_norm": 1.2207140193929902, "learning_rate": 1.1504589277299632e-05, "loss": 0.8162, "step": 12932 }, { "epoch": 0.46841724013038755, "grad_norm": 1.3493004085014548, "learning_rate": 1.1503429553129786e-05, "loss": 0.6699, "step": 12933 }, { "epoch": 0.4684534588917059, "grad_norm": 1.343640643898617, "learning_rate": 1.1502269808271429e-05, "loss": 0.7136, "step": 12934 }, { "epoch": 0.46848967765302424, "grad_norm": 1.331001883821691, "learning_rate": 1.1501110042740518e-05, "loss": 0.7281, "step": 12935 }, { "epoch": 0.4685258964143426, "grad_norm": 1.7030258215453327, "learning_rate": 1.1499950256553013e-05, "loss": 0.7566, "step": 12936 }, { "epoch": 0.468562115175661, "grad_norm": 1.43631324682552, "learning_rate": 1.1498790449724878e-05, "loss": 0.7762, "step": 12937 }, { "epoch": 0.46859833393697936, "grad_norm": 1.2475333313894372, "learning_rate": 1.1497630622272067e-05, "loss": 0.7266, "step": 12938 }, { "epoch": 0.46863455269829773, "grad_norm": 1.2860735877441014, "learning_rate": 1.1496470774210544e-05, "loss": 0.7327, "step": 12939 }, { "epoch": 0.46867077145961605, "grad_norm": 1.4101620314735817, "learning_rate": 1.1495310905556267e-05, "loss": 0.7613, "step": 12940 }, { "epoch": 0.4687069902209344, "grad_norm": 1.4358761014150403, "learning_rate": 1.14941510163252e-05, "loss": 0.7204, "step": 12941 }, { "epoch": 0.4687432089822528, "grad_norm": 1.1715157147576754, "learning_rate": 1.1492991106533297e-05, "loss": 0.7265, "step": 12942 }, { "epoch": 0.4687794277435712, "grad_norm": 1.2662719695341211, "learning_rate": 1.1491831176196531e-05, "loss": 0.8131, "step": 12943 }, { "epoch": 0.46881564650488955, "grad_norm": 1.3126410780312423, "learning_rate": 1.1490671225330855e-05, "loss": 0.7187, "step": 12944 }, { "epoch": 0.4688518652662079, "grad_norm": 1.6082626313232806, "learning_rate": 1.1489511253952232e-05, "loss": 0.8229, "step": 12945 }, { "epoch": 0.46888808402752624, "grad_norm": 1.3822419409215103, "learning_rate": 1.1488351262076625e-05, "loss": 0.7012, "step": 12946 }, { "epoch": 0.4689243027888446, "grad_norm": 1.4148555278551154, "learning_rate": 1.1487191249719998e-05, "loss": 0.7648, "step": 12947 }, { "epoch": 0.468960521550163, "grad_norm": 1.3285525366605426, "learning_rate": 1.1486031216898312e-05, "loss": 0.7255, "step": 12948 }, { "epoch": 0.46899674031148136, "grad_norm": 1.2850170661610754, "learning_rate": 1.148487116362753e-05, "loss": 0.694, "step": 12949 }, { "epoch": 0.46903295907279974, "grad_norm": 1.2599416668928642, "learning_rate": 1.148371108992362e-05, "loss": 0.7075, "step": 12950 }, { "epoch": 0.46906917783411806, "grad_norm": 1.2835280697910203, "learning_rate": 1.1482550995802537e-05, "loss": 0.7563, "step": 12951 }, { "epoch": 0.46910539659543643, "grad_norm": 1.329581806664301, "learning_rate": 1.1481390881280252e-05, "loss": 0.7659, "step": 12952 }, { "epoch": 0.4691416153567548, "grad_norm": 1.0809099976376992, "learning_rate": 1.1480230746372727e-05, "loss": 0.6876, "step": 12953 }, { "epoch": 0.4691778341180732, "grad_norm": 1.4484546554397184, "learning_rate": 1.1479070591095924e-05, "loss": 0.7472, "step": 12954 }, { "epoch": 0.46921405287939155, "grad_norm": 0.9929226654728619, "learning_rate": 1.1477910415465811e-05, "loss": 0.7114, "step": 12955 }, { "epoch": 0.46925027164070987, "grad_norm": 1.3588021490569817, "learning_rate": 1.1476750219498354e-05, "loss": 0.7044, "step": 12956 }, { "epoch": 0.46928649040202824, "grad_norm": 1.3664910511259198, "learning_rate": 1.1475590003209516e-05, "loss": 0.7901, "step": 12957 }, { "epoch": 0.4693227091633466, "grad_norm": 1.1857136834092294, "learning_rate": 1.147442976661526e-05, "loss": 0.7375, "step": 12958 }, { "epoch": 0.469358927924665, "grad_norm": 1.1600283044848727, "learning_rate": 1.1473269509731557e-05, "loss": 0.7189, "step": 12959 }, { "epoch": 0.46939514668598337, "grad_norm": 1.3416556515843712, "learning_rate": 1.147210923257437e-05, "loss": 0.7587, "step": 12960 }, { "epoch": 0.4694313654473017, "grad_norm": 1.3702627581851703, "learning_rate": 1.147094893515967e-05, "loss": 0.6921, "step": 12961 }, { "epoch": 0.46946758420862006, "grad_norm": 1.373495259334422, "learning_rate": 1.1469788617503417e-05, "loss": 0.6803, "step": 12962 }, { "epoch": 0.46950380296993843, "grad_norm": 1.0022835284917202, "learning_rate": 1.146862827962158e-05, "loss": 0.7194, "step": 12963 }, { "epoch": 0.4695400217312568, "grad_norm": 1.4755158133162736, "learning_rate": 1.1467467921530132e-05, "loss": 0.7806, "step": 12964 }, { "epoch": 0.4695762404925752, "grad_norm": 1.3508124302949351, "learning_rate": 1.1466307543245033e-05, "loss": 0.7493, "step": 12965 }, { "epoch": 0.4696124592538935, "grad_norm": 1.3468138791806614, "learning_rate": 1.1465147144782252e-05, "loss": 0.7714, "step": 12966 }, { "epoch": 0.4696486780152119, "grad_norm": 1.1891696374633791, "learning_rate": 1.146398672615776e-05, "loss": 0.8036, "step": 12967 }, { "epoch": 0.46968489677653025, "grad_norm": 1.409986011871957, "learning_rate": 1.1462826287387528e-05, "loss": 0.7147, "step": 12968 }, { "epoch": 0.4697211155378486, "grad_norm": 1.5012076739138043, "learning_rate": 1.1461665828487517e-05, "loss": 0.7999, "step": 12969 }, { "epoch": 0.469757334299167, "grad_norm": 1.5061244960046474, "learning_rate": 1.14605053494737e-05, "loss": 0.788, "step": 12970 }, { "epoch": 0.4697935530604853, "grad_norm": 1.009612834572833, "learning_rate": 1.1459344850362047e-05, "loss": 0.7582, "step": 12971 }, { "epoch": 0.4698297718218037, "grad_norm": 0.982138610013661, "learning_rate": 1.1458184331168526e-05, "loss": 0.733, "step": 12972 }, { "epoch": 0.46986599058312206, "grad_norm": 1.3616411760418714, "learning_rate": 1.1457023791909107e-05, "loss": 0.7666, "step": 12973 }, { "epoch": 0.46990220934444044, "grad_norm": 1.346193184227058, "learning_rate": 1.1455863232599764e-05, "loss": 0.7238, "step": 12974 }, { "epoch": 0.4699384281057588, "grad_norm": 1.4452771762449097, "learning_rate": 1.145470265325646e-05, "loss": 0.8258, "step": 12975 }, { "epoch": 0.46997464686707713, "grad_norm": 1.4814345448051718, "learning_rate": 1.145354205389517e-05, "loss": 0.7415, "step": 12976 }, { "epoch": 0.4700108656283955, "grad_norm": 1.6102630516883112, "learning_rate": 1.1452381434531863e-05, "loss": 0.8288, "step": 12977 }, { "epoch": 0.4700470843897139, "grad_norm": 1.4219809790038198, "learning_rate": 1.1451220795182511e-05, "loss": 0.7395, "step": 12978 }, { "epoch": 0.47008330315103225, "grad_norm": 1.5218648618403519, "learning_rate": 1.145006013586309e-05, "loss": 0.7256, "step": 12979 }, { "epoch": 0.4701195219123506, "grad_norm": 1.2799583137441775, "learning_rate": 1.1448899456589563e-05, "loss": 0.7563, "step": 12980 }, { "epoch": 0.47015574067366894, "grad_norm": 1.2363207991154739, "learning_rate": 1.1447738757377911e-05, "loss": 0.7696, "step": 12981 }, { "epoch": 0.4701919594349873, "grad_norm": 1.296769102639028, "learning_rate": 1.1446578038244096e-05, "loss": 0.7543, "step": 12982 }, { "epoch": 0.4702281781963057, "grad_norm": 1.441969044455859, "learning_rate": 1.1445417299204101e-05, "loss": 0.7882, "step": 12983 }, { "epoch": 0.47026439695762406, "grad_norm": 1.3063978417249946, "learning_rate": 1.144425654027389e-05, "loss": 0.7494, "step": 12984 }, { "epoch": 0.47030061571894244, "grad_norm": 1.659162816528261, "learning_rate": 1.1443095761469443e-05, "loss": 0.6995, "step": 12985 }, { "epoch": 0.47033683448026076, "grad_norm": 1.030773531244322, "learning_rate": 1.144193496280673e-05, "loss": 0.75, "step": 12986 }, { "epoch": 0.47037305324157913, "grad_norm": 1.5303220966346687, "learning_rate": 1.1440774144301723e-05, "loss": 0.7099, "step": 12987 }, { "epoch": 0.4704092720028975, "grad_norm": 1.2836295708251189, "learning_rate": 1.1439613305970398e-05, "loss": 0.8037, "step": 12988 }, { "epoch": 0.4704454907642159, "grad_norm": 1.4671487259235452, "learning_rate": 1.143845244782873e-05, "loss": 0.8168, "step": 12989 }, { "epoch": 0.47048170952553425, "grad_norm": 1.359620151666672, "learning_rate": 1.1437291569892693e-05, "loss": 0.7624, "step": 12990 }, { "epoch": 0.47051792828685257, "grad_norm": 0.9491022319569967, "learning_rate": 1.1436130672178258e-05, "loss": 0.6822, "step": 12991 }, { "epoch": 0.47055414704817095, "grad_norm": 1.356058040726285, "learning_rate": 1.1434969754701406e-05, "loss": 0.6945, "step": 12992 }, { "epoch": 0.4705903658094893, "grad_norm": 1.353142137727286, "learning_rate": 1.1433808817478107e-05, "loss": 0.7748, "step": 12993 }, { "epoch": 0.4706265845708077, "grad_norm": 1.7634954945182515, "learning_rate": 1.1432647860524342e-05, "loss": 0.7129, "step": 12994 }, { "epoch": 0.47066280333212607, "grad_norm": 1.297644207523604, "learning_rate": 1.1431486883856082e-05, "loss": 0.7945, "step": 12995 }, { "epoch": 0.4706990220934444, "grad_norm": 1.3797216530522411, "learning_rate": 1.1430325887489304e-05, "loss": 0.7168, "step": 12996 }, { "epoch": 0.47073524085476276, "grad_norm": 1.4095324057024678, "learning_rate": 1.1429164871439985e-05, "loss": 0.8438, "step": 12997 }, { "epoch": 0.47077145961608113, "grad_norm": 1.4093374830156242, "learning_rate": 1.1428003835724102e-05, "loss": 0.8524, "step": 12998 }, { "epoch": 0.4708076783773995, "grad_norm": 1.6130007391435428, "learning_rate": 1.1426842780357633e-05, "loss": 0.7831, "step": 12999 }, { "epoch": 0.4708438971387179, "grad_norm": 1.389405224574746, "learning_rate": 1.142568170535655e-05, "loss": 0.7783, "step": 13000 }, { "epoch": 0.4708801159000362, "grad_norm": 1.362362848769915, "learning_rate": 1.1424520610736837e-05, "loss": 0.8055, "step": 13001 }, { "epoch": 0.4709163346613546, "grad_norm": 1.527172745524487, "learning_rate": 1.1423359496514469e-05, "loss": 0.7243, "step": 13002 }, { "epoch": 0.47095255342267295, "grad_norm": 1.2907148308728524, "learning_rate": 1.1422198362705424e-05, "loss": 0.7046, "step": 13003 }, { "epoch": 0.4709887721839913, "grad_norm": 1.4675167732421517, "learning_rate": 1.1421037209325676e-05, "loss": 0.769, "step": 13004 }, { "epoch": 0.4710249909453097, "grad_norm": 1.4236238291186543, "learning_rate": 1.141987603639121e-05, "loss": 0.7492, "step": 13005 }, { "epoch": 0.471061209706628, "grad_norm": 1.2692847709538835, "learning_rate": 1.1418714843918002e-05, "loss": 0.7371, "step": 13006 }, { "epoch": 0.4710974284679464, "grad_norm": 1.3519613333527025, "learning_rate": 1.141755363192203e-05, "loss": 0.8109, "step": 13007 }, { "epoch": 0.47113364722926476, "grad_norm": 1.3499873662949091, "learning_rate": 1.1416392400419279e-05, "loss": 0.8115, "step": 13008 }, { "epoch": 0.47116986599058314, "grad_norm": 1.4292225127151332, "learning_rate": 1.1415231149425721e-05, "loss": 0.7476, "step": 13009 }, { "epoch": 0.4712060847519015, "grad_norm": 1.3316628416495935, "learning_rate": 1.141406987895734e-05, "loss": 0.7651, "step": 13010 }, { "epoch": 0.47124230351321983, "grad_norm": 2.5931590524475516, "learning_rate": 1.1412908589030112e-05, "loss": 0.7943, "step": 13011 }, { "epoch": 0.4712785222745382, "grad_norm": 1.3018355650407212, "learning_rate": 1.1411747279660021e-05, "loss": 0.7628, "step": 13012 }, { "epoch": 0.4713147410358566, "grad_norm": 1.2901530219083068, "learning_rate": 1.1410585950863047e-05, "loss": 0.7165, "step": 13013 }, { "epoch": 0.47135095979717495, "grad_norm": 1.361195986805184, "learning_rate": 1.140942460265517e-05, "loss": 0.8246, "step": 13014 }, { "epoch": 0.4713871785584933, "grad_norm": 1.3920262026815728, "learning_rate": 1.1408263235052374e-05, "loss": 0.6971, "step": 13015 }, { "epoch": 0.47142339731981164, "grad_norm": 1.4836294133749597, "learning_rate": 1.1407101848070638e-05, "loss": 0.7397, "step": 13016 }, { "epoch": 0.47145961608113, "grad_norm": 1.371068065082122, "learning_rate": 1.1405940441725947e-05, "loss": 0.7364, "step": 13017 }, { "epoch": 0.4714958348424484, "grad_norm": 1.3001315782743854, "learning_rate": 1.1404779016034275e-05, "loss": 0.7204, "step": 13018 }, { "epoch": 0.47153205360376677, "grad_norm": 1.2569008086505409, "learning_rate": 1.1403617571011614e-05, "loss": 0.6937, "step": 13019 }, { "epoch": 0.47156827236508514, "grad_norm": 1.411807661813777, "learning_rate": 1.1402456106673938e-05, "loss": 0.7715, "step": 13020 }, { "epoch": 0.47160449112640346, "grad_norm": 1.4160419093356258, "learning_rate": 1.1401294623037238e-05, "loss": 0.656, "step": 13021 }, { "epoch": 0.47164070988772183, "grad_norm": 1.386929972774527, "learning_rate": 1.1400133120117486e-05, "loss": 0.7709, "step": 13022 }, { "epoch": 0.4716769286490402, "grad_norm": 1.7461564348263634, "learning_rate": 1.1398971597930679e-05, "loss": 0.7329, "step": 13023 }, { "epoch": 0.4717131474103586, "grad_norm": 1.1901434299157385, "learning_rate": 1.1397810056492791e-05, "loss": 0.6909, "step": 13024 }, { "epoch": 0.47174936617167695, "grad_norm": 1.3479730178510747, "learning_rate": 1.1396648495819807e-05, "loss": 0.7186, "step": 13025 }, { "epoch": 0.47178558493299527, "grad_norm": 1.2730179407074504, "learning_rate": 1.1395486915927713e-05, "loss": 0.7274, "step": 13026 }, { "epoch": 0.47182180369431365, "grad_norm": 1.3382552353155976, "learning_rate": 1.1394325316832493e-05, "loss": 0.8115, "step": 13027 }, { "epoch": 0.471858022455632, "grad_norm": 1.3307496166959674, "learning_rate": 1.1393163698550133e-05, "loss": 0.7932, "step": 13028 }, { "epoch": 0.4718942412169504, "grad_norm": 1.3489780888681877, "learning_rate": 1.1392002061096614e-05, "loss": 0.7786, "step": 13029 }, { "epoch": 0.47193045997826877, "grad_norm": 1.0623771624084035, "learning_rate": 1.1390840404487926e-05, "loss": 0.7634, "step": 13030 }, { "epoch": 0.4719666787395871, "grad_norm": 1.3788046279343331, "learning_rate": 1.138967872874005e-05, "loss": 0.7868, "step": 13031 }, { "epoch": 0.47200289750090546, "grad_norm": 1.299933874214833, "learning_rate": 1.1388517033868976e-05, "loss": 0.7847, "step": 13032 }, { "epoch": 0.47203911626222383, "grad_norm": 1.3590860724329117, "learning_rate": 1.1387355319890685e-05, "loss": 0.7583, "step": 13033 }, { "epoch": 0.4720753350235422, "grad_norm": 1.569994676035765, "learning_rate": 1.1386193586821165e-05, "loss": 0.7486, "step": 13034 }, { "epoch": 0.4721115537848606, "grad_norm": 1.386315287070811, "learning_rate": 1.1385031834676405e-05, "loss": 0.7768, "step": 13035 }, { "epoch": 0.4721477725461789, "grad_norm": 0.985185716311346, "learning_rate": 1.138387006347239e-05, "loss": 0.7476, "step": 13036 }, { "epoch": 0.4721839913074973, "grad_norm": 1.480627177274302, "learning_rate": 1.1382708273225108e-05, "loss": 0.7756, "step": 13037 }, { "epoch": 0.47222021006881565, "grad_norm": 1.60440667377855, "learning_rate": 1.1381546463950542e-05, "loss": 0.7167, "step": 13038 }, { "epoch": 0.472256428830134, "grad_norm": 1.3887949089616287, "learning_rate": 1.1380384635664688e-05, "loss": 0.7759, "step": 13039 }, { "epoch": 0.4722926475914524, "grad_norm": 1.4422904354982669, "learning_rate": 1.1379222788383522e-05, "loss": 0.7446, "step": 13040 }, { "epoch": 0.4723288663527707, "grad_norm": 1.3962991231475046, "learning_rate": 1.1378060922123044e-05, "loss": 0.7304, "step": 13041 }, { "epoch": 0.4723650851140891, "grad_norm": 1.412125148790618, "learning_rate": 1.1376899036899234e-05, "loss": 0.7266, "step": 13042 }, { "epoch": 0.47240130387540746, "grad_norm": 1.3376830977267764, "learning_rate": 1.1375737132728084e-05, "loss": 0.7045, "step": 13043 }, { "epoch": 0.47243752263672584, "grad_norm": 1.2935821793791291, "learning_rate": 1.137457520962558e-05, "loss": 0.7555, "step": 13044 }, { "epoch": 0.4724737413980442, "grad_norm": 1.07305663098619, "learning_rate": 1.1373413267607716e-05, "loss": 0.7909, "step": 13045 }, { "epoch": 0.47250996015936253, "grad_norm": 1.328364100391464, "learning_rate": 1.1372251306690478e-05, "loss": 0.7071, "step": 13046 }, { "epoch": 0.4725461789206809, "grad_norm": 1.404208307829094, "learning_rate": 1.1371089326889857e-05, "loss": 0.7809, "step": 13047 }, { "epoch": 0.4725823976819993, "grad_norm": 1.4460695835569355, "learning_rate": 1.1369927328221842e-05, "loss": 0.7205, "step": 13048 }, { "epoch": 0.47261861644331765, "grad_norm": 1.3533129210014943, "learning_rate": 1.1368765310702422e-05, "loss": 0.7913, "step": 13049 }, { "epoch": 0.472654835204636, "grad_norm": 1.317862038229677, "learning_rate": 1.1367603274347591e-05, "loss": 0.7267, "step": 13050 }, { "epoch": 0.47269105396595434, "grad_norm": 1.3266795307551662, "learning_rate": 1.1366441219173335e-05, "loss": 0.7642, "step": 13051 }, { "epoch": 0.4727272727272727, "grad_norm": 1.4660330458555588, "learning_rate": 1.1365279145195649e-05, "loss": 0.7522, "step": 13052 }, { "epoch": 0.4727634914885911, "grad_norm": 1.4705575418797265, "learning_rate": 1.136411705243052e-05, "loss": 0.7069, "step": 13053 }, { "epoch": 0.47279971024990947, "grad_norm": 1.3300441591964596, "learning_rate": 1.1362954940893944e-05, "loss": 0.7121, "step": 13054 }, { "epoch": 0.47283592901122784, "grad_norm": 1.1042589080187009, "learning_rate": 1.1361792810601908e-05, "loss": 0.7767, "step": 13055 }, { "epoch": 0.47287214777254616, "grad_norm": 1.368736744216631, "learning_rate": 1.1360630661570406e-05, "loss": 0.8197, "step": 13056 }, { "epoch": 0.47290836653386453, "grad_norm": 1.3961780505566483, "learning_rate": 1.1359468493815434e-05, "loss": 0.7182, "step": 13057 }, { "epoch": 0.4729445852951829, "grad_norm": 1.165863521447915, "learning_rate": 1.1358306307352978e-05, "loss": 0.7699, "step": 13058 }, { "epoch": 0.4729808040565013, "grad_norm": 0.9741128292545813, "learning_rate": 1.1357144102199035e-05, "loss": 0.7534, "step": 13059 }, { "epoch": 0.47301702281781965, "grad_norm": 1.3936484050715126, "learning_rate": 1.1355981878369598e-05, "loss": 0.751, "step": 13060 }, { "epoch": 0.473053241579138, "grad_norm": 1.4215754662448816, "learning_rate": 1.1354819635880658e-05, "loss": 0.8325, "step": 13061 }, { "epoch": 0.47308946034045635, "grad_norm": 1.429627716594181, "learning_rate": 1.1353657374748206e-05, "loss": 0.7419, "step": 13062 }, { "epoch": 0.4731256791017747, "grad_norm": 0.9977934063138701, "learning_rate": 1.1352495094988245e-05, "loss": 0.6613, "step": 13063 }, { "epoch": 0.4731618978630931, "grad_norm": 1.6018996622326254, "learning_rate": 1.1351332796616759e-05, "loss": 0.8751, "step": 13064 }, { "epoch": 0.47319811662441147, "grad_norm": 0.9695197713382412, "learning_rate": 1.1350170479649747e-05, "loss": 0.7189, "step": 13065 }, { "epoch": 0.4732343353857298, "grad_norm": 1.0530164368184973, "learning_rate": 1.1349008144103202e-05, "loss": 0.7358, "step": 13066 }, { "epoch": 0.47327055414704816, "grad_norm": 1.1315517869067298, "learning_rate": 1.134784578999312e-05, "loss": 0.7418, "step": 13067 }, { "epoch": 0.47330677290836654, "grad_norm": 1.32677009416228, "learning_rate": 1.1346683417335497e-05, "loss": 0.7195, "step": 13068 }, { "epoch": 0.4733429916696849, "grad_norm": 1.4296615909615566, "learning_rate": 1.1345521026146327e-05, "loss": 0.7579, "step": 13069 }, { "epoch": 0.4733792104310033, "grad_norm": 1.3868031187316117, "learning_rate": 1.1344358616441605e-05, "loss": 0.6723, "step": 13070 }, { "epoch": 0.4734154291923216, "grad_norm": 1.0309156514839921, "learning_rate": 1.1343196188237327e-05, "loss": 0.7067, "step": 13071 }, { "epoch": 0.47345164795364, "grad_norm": 1.1301637389318473, "learning_rate": 1.134203374154949e-05, "loss": 0.6909, "step": 13072 }, { "epoch": 0.47348786671495835, "grad_norm": 1.5008833439905505, "learning_rate": 1.1340871276394085e-05, "loss": 0.7539, "step": 13073 }, { "epoch": 0.4735240854762767, "grad_norm": 1.0429359468938737, "learning_rate": 1.1339708792787119e-05, "loss": 0.8043, "step": 13074 }, { "epoch": 0.4735603042375951, "grad_norm": 0.9742502959214651, "learning_rate": 1.1338546290744578e-05, "loss": 0.7093, "step": 13075 }, { "epoch": 0.4735965229989134, "grad_norm": 1.3659387314600424, "learning_rate": 1.1337383770282466e-05, "loss": 0.7345, "step": 13076 }, { "epoch": 0.4736327417602318, "grad_norm": 1.3469730276249572, "learning_rate": 1.133622123141678e-05, "loss": 0.7398, "step": 13077 }, { "epoch": 0.47366896052155016, "grad_norm": 1.5062586555588409, "learning_rate": 1.1335058674163512e-05, "loss": 0.83, "step": 13078 }, { "epoch": 0.47370517928286854, "grad_norm": 1.371893753964141, "learning_rate": 1.1333896098538669e-05, "loss": 0.6856, "step": 13079 }, { "epoch": 0.4737413980441869, "grad_norm": 1.5122024306959283, "learning_rate": 1.1332733504558239e-05, "loss": 0.7924, "step": 13080 }, { "epoch": 0.47377761680550523, "grad_norm": 1.3795079277487579, "learning_rate": 1.1331570892238228e-05, "loss": 0.7176, "step": 13081 }, { "epoch": 0.4738138355668236, "grad_norm": 1.3519406850728009, "learning_rate": 1.1330408261594627e-05, "loss": 0.7662, "step": 13082 }, { "epoch": 0.473850054328142, "grad_norm": 1.2657065074630107, "learning_rate": 1.1329245612643445e-05, "loss": 0.8238, "step": 13083 }, { "epoch": 0.47388627308946035, "grad_norm": 1.3601758792803926, "learning_rate": 1.132808294540067e-05, "loss": 0.7528, "step": 13084 }, { "epoch": 0.4739224918507787, "grad_norm": 1.2311278160038124, "learning_rate": 1.1326920259882312e-05, "loss": 0.7185, "step": 13085 }, { "epoch": 0.47395871061209705, "grad_norm": 1.4741812697186059, "learning_rate": 1.1325757556104362e-05, "loss": 0.833, "step": 13086 }, { "epoch": 0.4739949293734154, "grad_norm": 1.2807175566505271, "learning_rate": 1.1324594834082823e-05, "loss": 0.6805, "step": 13087 }, { "epoch": 0.4740311481347338, "grad_norm": 1.300997589352599, "learning_rate": 1.1323432093833695e-05, "loss": 0.6836, "step": 13088 }, { "epoch": 0.47406736689605217, "grad_norm": 1.3372891584664435, "learning_rate": 1.132226933537298e-05, "loss": 0.7927, "step": 13089 }, { "epoch": 0.47410358565737054, "grad_norm": 1.2052644366492486, "learning_rate": 1.1321106558716676e-05, "loss": 0.7848, "step": 13090 }, { "epoch": 0.47413980441868886, "grad_norm": 1.5633068659779294, "learning_rate": 1.1319943763880786e-05, "loss": 0.737, "step": 13091 }, { "epoch": 0.47417602318000723, "grad_norm": 1.3484672186450892, "learning_rate": 1.131878095088131e-05, "loss": 0.7266, "step": 13092 }, { "epoch": 0.4742122419413256, "grad_norm": 1.4065761803788277, "learning_rate": 1.1317618119734249e-05, "loss": 0.8437, "step": 13093 }, { "epoch": 0.474248460702644, "grad_norm": 1.3699300166405624, "learning_rate": 1.1316455270455604e-05, "loss": 0.7448, "step": 13094 }, { "epoch": 0.47428467946396236, "grad_norm": 1.6852842194801363, "learning_rate": 1.1315292403061378e-05, "loss": 0.7499, "step": 13095 }, { "epoch": 0.4743208982252807, "grad_norm": 1.2826146034713097, "learning_rate": 1.1314129517567576e-05, "loss": 0.6661, "step": 13096 }, { "epoch": 0.47435711698659905, "grad_norm": 1.294063749830927, "learning_rate": 1.1312966613990192e-05, "loss": 0.7644, "step": 13097 }, { "epoch": 0.4743933357479174, "grad_norm": 1.4362402610266016, "learning_rate": 1.1311803692345236e-05, "loss": 0.7704, "step": 13098 }, { "epoch": 0.4744295545092358, "grad_norm": 1.3268016547376573, "learning_rate": 1.1310640752648709e-05, "loss": 0.8083, "step": 13099 }, { "epoch": 0.47446577327055417, "grad_norm": 1.2391016296585606, "learning_rate": 1.1309477794916612e-05, "loss": 0.7246, "step": 13100 }, { "epoch": 0.4745019920318725, "grad_norm": 1.419532674002471, "learning_rate": 1.1308314819164951e-05, "loss": 0.7529, "step": 13101 }, { "epoch": 0.47453821079319086, "grad_norm": 1.4935687921854994, "learning_rate": 1.130715182540973e-05, "loss": 0.7154, "step": 13102 }, { "epoch": 0.47457442955450924, "grad_norm": 1.0120796649312571, "learning_rate": 1.1305988813666951e-05, "loss": 0.7419, "step": 13103 }, { "epoch": 0.4746106483158276, "grad_norm": 1.3725281789714951, "learning_rate": 1.1304825783952615e-05, "loss": 0.7604, "step": 13104 }, { "epoch": 0.474646867077146, "grad_norm": 1.080613181740015, "learning_rate": 1.1303662736282735e-05, "loss": 0.7656, "step": 13105 }, { "epoch": 0.4746830858384643, "grad_norm": 1.4420981839813833, "learning_rate": 1.1302499670673306e-05, "loss": 0.8268, "step": 13106 }, { "epoch": 0.4747193045997827, "grad_norm": 1.276517775247041, "learning_rate": 1.1301336587140337e-05, "loss": 0.6694, "step": 13107 }, { "epoch": 0.47475552336110105, "grad_norm": 1.3150652470510815, "learning_rate": 1.1300173485699834e-05, "loss": 0.7244, "step": 13108 }, { "epoch": 0.4747917421224194, "grad_norm": 1.1265584331061576, "learning_rate": 1.1299010366367801e-05, "loss": 0.7521, "step": 13109 }, { "epoch": 0.4748279608837378, "grad_norm": 1.1483673273757542, "learning_rate": 1.1297847229160248e-05, "loss": 0.7041, "step": 13110 }, { "epoch": 0.4748641796450561, "grad_norm": 1.787360743786927, "learning_rate": 1.1296684074093171e-05, "loss": 0.8221, "step": 13111 }, { "epoch": 0.4749003984063745, "grad_norm": 0.9804053599843416, "learning_rate": 1.1295520901182586e-05, "loss": 0.7475, "step": 13112 }, { "epoch": 0.47493661716769286, "grad_norm": 1.2340261843216869, "learning_rate": 1.1294357710444494e-05, "loss": 0.725, "step": 13113 }, { "epoch": 0.47497283592901124, "grad_norm": 1.2846033079315755, "learning_rate": 1.1293194501894901e-05, "loss": 0.773, "step": 13114 }, { "epoch": 0.4750090546903296, "grad_norm": 1.2989582369461983, "learning_rate": 1.1292031275549817e-05, "loss": 0.7022, "step": 13115 }, { "epoch": 0.47504527345164793, "grad_norm": 1.0900157968169586, "learning_rate": 1.1290868031425246e-05, "loss": 0.7377, "step": 13116 }, { "epoch": 0.4750814922129663, "grad_norm": 1.284847963164551, "learning_rate": 1.12897047695372e-05, "loss": 0.6976, "step": 13117 }, { "epoch": 0.4751177109742847, "grad_norm": 1.3678736994632674, "learning_rate": 1.1288541489901679e-05, "loss": 0.7423, "step": 13118 }, { "epoch": 0.47515392973560305, "grad_norm": 1.3037010335367463, "learning_rate": 1.1287378192534697e-05, "loss": 0.7839, "step": 13119 }, { "epoch": 0.4751901484969214, "grad_norm": 1.1495136874981873, "learning_rate": 1.1286214877452259e-05, "loss": 0.7345, "step": 13120 }, { "epoch": 0.47522636725823975, "grad_norm": 1.4520043671694343, "learning_rate": 1.1285051544670375e-05, "loss": 0.8271, "step": 13121 }, { "epoch": 0.4752625860195581, "grad_norm": 1.1008844897269803, "learning_rate": 1.128388819420505e-05, "loss": 0.758, "step": 13122 }, { "epoch": 0.4752988047808765, "grad_norm": 1.3587986886408796, "learning_rate": 1.1282724826072297e-05, "loss": 0.7933, "step": 13123 }, { "epoch": 0.47533502354219487, "grad_norm": 1.077900043448595, "learning_rate": 1.1281561440288125e-05, "loss": 0.7375, "step": 13124 }, { "epoch": 0.47537124230351324, "grad_norm": 1.326942523310918, "learning_rate": 1.1280398036868538e-05, "loss": 0.7405, "step": 13125 }, { "epoch": 0.47540746106483156, "grad_norm": 1.443712036554334, "learning_rate": 1.1279234615829553e-05, "loss": 0.7481, "step": 13126 }, { "epoch": 0.47544367982614993, "grad_norm": 1.1722720155141346, "learning_rate": 1.1278071177187174e-05, "loss": 0.7452, "step": 13127 }, { "epoch": 0.4754798985874683, "grad_norm": 1.0523889070211523, "learning_rate": 1.1276907720957414e-05, "loss": 0.7194, "step": 13128 }, { "epoch": 0.4755161173487867, "grad_norm": 1.4231402635101342, "learning_rate": 1.1275744247156277e-05, "loss": 0.7357, "step": 13129 }, { "epoch": 0.47555233611010506, "grad_norm": 1.1568750005895256, "learning_rate": 1.1274580755799785e-05, "loss": 0.7276, "step": 13130 }, { "epoch": 0.4755885548714234, "grad_norm": 1.31234139097165, "learning_rate": 1.1273417246903935e-05, "loss": 0.7852, "step": 13131 }, { "epoch": 0.47562477363274175, "grad_norm": 1.1637746555836235, "learning_rate": 1.127225372048475e-05, "loss": 0.6749, "step": 13132 }, { "epoch": 0.4756609923940601, "grad_norm": 1.2258878377842655, "learning_rate": 1.1271090176558233e-05, "loss": 0.7041, "step": 13133 }, { "epoch": 0.4756972111553785, "grad_norm": 1.009584648158216, "learning_rate": 1.1269926615140397e-05, "loss": 0.7217, "step": 13134 }, { "epoch": 0.47573342991669687, "grad_norm": 1.2708529924135221, "learning_rate": 1.1268763036247259e-05, "loss": 0.7968, "step": 13135 }, { "epoch": 0.4757696486780152, "grad_norm": 1.3755064134685724, "learning_rate": 1.1267599439894822e-05, "loss": 0.7374, "step": 13136 }, { "epoch": 0.47580586743933356, "grad_norm": 0.9771044134731343, "learning_rate": 1.126643582609911e-05, "loss": 0.7241, "step": 13137 }, { "epoch": 0.47584208620065194, "grad_norm": 1.1314029673908834, "learning_rate": 1.126527219487612e-05, "loss": 0.7554, "step": 13138 }, { "epoch": 0.4758783049619703, "grad_norm": 1.325801750664133, "learning_rate": 1.126410854624188e-05, "loss": 0.7496, "step": 13139 }, { "epoch": 0.4759145237232887, "grad_norm": 1.5117905621882013, "learning_rate": 1.126294488021239e-05, "loss": 0.7995, "step": 13140 }, { "epoch": 0.475950742484607, "grad_norm": 1.3267655399209228, "learning_rate": 1.1261781196803674e-05, "loss": 0.691, "step": 13141 }, { "epoch": 0.4759869612459254, "grad_norm": 1.273984478204669, "learning_rate": 1.1260617496031733e-05, "loss": 0.7161, "step": 13142 }, { "epoch": 0.47602318000724375, "grad_norm": 0.9636996471702255, "learning_rate": 1.1259453777912595e-05, "loss": 0.653, "step": 13143 }, { "epoch": 0.4760593987685621, "grad_norm": 1.3395170482168326, "learning_rate": 1.1258290042462261e-05, "loss": 0.7098, "step": 13144 }, { "epoch": 0.4760956175298805, "grad_norm": 1.2443391410991174, "learning_rate": 1.125712628969675e-05, "loss": 0.7164, "step": 13145 }, { "epoch": 0.4761318362911988, "grad_norm": 1.4432929027566432, "learning_rate": 1.1255962519632082e-05, "loss": 0.7501, "step": 13146 }, { "epoch": 0.4761680550525172, "grad_norm": 1.4162751269152707, "learning_rate": 1.1254798732284263e-05, "loss": 0.7844, "step": 13147 }, { "epoch": 0.47620427381383557, "grad_norm": 1.2854071563513663, "learning_rate": 1.1253634927669313e-05, "loss": 0.6399, "step": 13148 }, { "epoch": 0.47624049257515394, "grad_norm": 1.269522769606216, "learning_rate": 1.1252471105803241e-05, "loss": 0.7013, "step": 13149 }, { "epoch": 0.4762767113364723, "grad_norm": 0.9647144889104997, "learning_rate": 1.125130726670207e-05, "loss": 0.71, "step": 13150 }, { "epoch": 0.47631293009779063, "grad_norm": 1.3967704759978488, "learning_rate": 1.1250143410381808e-05, "loss": 0.6876, "step": 13151 }, { "epoch": 0.476349148859109, "grad_norm": 1.2034658008929928, "learning_rate": 1.1248979536858477e-05, "loss": 0.7274, "step": 13152 }, { "epoch": 0.4763853676204274, "grad_norm": 1.4424251622495874, "learning_rate": 1.1247815646148088e-05, "loss": 0.7802, "step": 13153 }, { "epoch": 0.47642158638174575, "grad_norm": 1.11738574499154, "learning_rate": 1.1246651738266659e-05, "loss": 0.7462, "step": 13154 }, { "epoch": 0.47645780514306413, "grad_norm": 1.2215773412279138, "learning_rate": 1.1245487813230209e-05, "loss": 0.6663, "step": 13155 }, { "epoch": 0.47649402390438245, "grad_norm": 1.0730678328972088, "learning_rate": 1.124432387105475e-05, "loss": 0.725, "step": 13156 }, { "epoch": 0.4765302426657008, "grad_norm": 1.3334975590064018, "learning_rate": 1.1243159911756306e-05, "loss": 0.7262, "step": 13157 }, { "epoch": 0.4765664614270192, "grad_norm": 1.3874906131227103, "learning_rate": 1.1241995935350882e-05, "loss": 0.7741, "step": 13158 }, { "epoch": 0.47660268018833757, "grad_norm": 1.0988291643366965, "learning_rate": 1.1240831941854509e-05, "loss": 0.7689, "step": 13159 }, { "epoch": 0.47663889894965594, "grad_norm": 1.1178275986816966, "learning_rate": 1.1239667931283193e-05, "loss": 0.7128, "step": 13160 }, { "epoch": 0.47667511771097426, "grad_norm": 1.4973152841105097, "learning_rate": 1.1238503903652962e-05, "loss": 0.7944, "step": 13161 }, { "epoch": 0.47671133647229263, "grad_norm": 1.5378241609905658, "learning_rate": 1.1237339858979825e-05, "loss": 0.811, "step": 13162 }, { "epoch": 0.476747555233611, "grad_norm": 1.336353017089553, "learning_rate": 1.1236175797279807e-05, "loss": 0.7014, "step": 13163 }, { "epoch": 0.4767837739949294, "grad_norm": 1.2754368912394474, "learning_rate": 1.1235011718568922e-05, "loss": 0.7523, "step": 13164 }, { "epoch": 0.47681999275624776, "grad_norm": 1.4037354896418357, "learning_rate": 1.123384762286319e-05, "loss": 0.8202, "step": 13165 }, { "epoch": 0.4768562115175661, "grad_norm": 1.0631154290530391, "learning_rate": 1.1232683510178632e-05, "loss": 0.735, "step": 13166 }, { "epoch": 0.47689243027888445, "grad_norm": 1.1251892435761555, "learning_rate": 1.1231519380531265e-05, "loss": 0.6712, "step": 13167 }, { "epoch": 0.4769286490402028, "grad_norm": 1.3256119851008983, "learning_rate": 1.1230355233937112e-05, "loss": 0.7529, "step": 13168 }, { "epoch": 0.4769648678015212, "grad_norm": 1.2720981239483597, "learning_rate": 1.1229191070412187e-05, "loss": 0.7596, "step": 13169 }, { "epoch": 0.47700108656283957, "grad_norm": 1.0951919296300043, "learning_rate": 1.1228026889972514e-05, "loss": 0.7882, "step": 13170 }, { "epoch": 0.4770373053241579, "grad_norm": 1.2018131769398002, "learning_rate": 1.122686269263411e-05, "loss": 0.6598, "step": 13171 }, { "epoch": 0.47707352408547626, "grad_norm": 1.4374945319113186, "learning_rate": 1.1225698478413e-05, "loss": 0.8096, "step": 13172 }, { "epoch": 0.47710974284679464, "grad_norm": 1.3681694162714655, "learning_rate": 1.12245342473252e-05, "loss": 0.7778, "step": 13173 }, { "epoch": 0.477145961608113, "grad_norm": 1.0392341778369356, "learning_rate": 1.1223369999386736e-05, "loss": 0.7519, "step": 13174 }, { "epoch": 0.4771821803694314, "grad_norm": 1.3019276134041329, "learning_rate": 1.1222205734613621e-05, "loss": 0.7242, "step": 13175 }, { "epoch": 0.4772183991307497, "grad_norm": 1.4967075313948521, "learning_rate": 1.122104145302188e-05, "loss": 0.7968, "step": 13176 }, { "epoch": 0.4772546178920681, "grad_norm": 1.3146038861703366, "learning_rate": 1.1219877154627543e-05, "loss": 0.7673, "step": 13177 }, { "epoch": 0.47729083665338645, "grad_norm": 1.2803771226508376, "learning_rate": 1.121871283944662e-05, "loss": 0.6668, "step": 13178 }, { "epoch": 0.4773270554147048, "grad_norm": 1.3617696203522682, "learning_rate": 1.1217548507495137e-05, "loss": 0.7563, "step": 13179 }, { "epoch": 0.4773632741760232, "grad_norm": 1.3095880467396737, "learning_rate": 1.1216384158789117e-05, "loss": 0.7474, "step": 13180 }, { "epoch": 0.4773994929373415, "grad_norm": 1.4676257482401205, "learning_rate": 1.1215219793344585e-05, "loss": 0.7289, "step": 13181 }, { "epoch": 0.4774357116986599, "grad_norm": 1.1034728824726694, "learning_rate": 1.1214055411177558e-05, "loss": 0.7047, "step": 13182 }, { "epoch": 0.47747193045997827, "grad_norm": 1.7580884059965212, "learning_rate": 1.1212891012304064e-05, "loss": 0.8497, "step": 13183 }, { "epoch": 0.47750814922129664, "grad_norm": 1.330387056494778, "learning_rate": 1.1211726596740123e-05, "loss": 0.7502, "step": 13184 }, { "epoch": 0.477544367982615, "grad_norm": 1.559520780365587, "learning_rate": 1.1210562164501757e-05, "loss": 0.7234, "step": 13185 }, { "epoch": 0.47758058674393333, "grad_norm": 1.6954316303920642, "learning_rate": 1.1209397715604994e-05, "loss": 0.7718, "step": 13186 }, { "epoch": 0.4776168055052517, "grad_norm": 1.2450439184734037, "learning_rate": 1.1208233250065857e-05, "loss": 0.8189, "step": 13187 }, { "epoch": 0.4776530242665701, "grad_norm": 1.3711200443963447, "learning_rate": 1.1207068767900367e-05, "loss": 0.6504, "step": 13188 }, { "epoch": 0.47768924302788845, "grad_norm": 1.0979899051201716, "learning_rate": 1.120590426912455e-05, "loss": 0.6893, "step": 13189 }, { "epoch": 0.47772546178920683, "grad_norm": 1.3294282563553383, "learning_rate": 1.1204739753754433e-05, "loss": 0.7893, "step": 13190 }, { "epoch": 0.47776168055052515, "grad_norm": 1.2124973487271984, "learning_rate": 1.1203575221806035e-05, "loss": 0.8077, "step": 13191 }, { "epoch": 0.4777978993118435, "grad_norm": 1.0017350019366955, "learning_rate": 1.1202410673295387e-05, "loss": 0.7613, "step": 13192 }, { "epoch": 0.4778341180731619, "grad_norm": 1.3188095680451175, "learning_rate": 1.120124610823851e-05, "loss": 0.7107, "step": 13193 }, { "epoch": 0.47787033683448027, "grad_norm": 1.5196612266955671, "learning_rate": 1.1200081526651432e-05, "loss": 0.8376, "step": 13194 }, { "epoch": 0.47790655559579864, "grad_norm": 1.2005292693218923, "learning_rate": 1.1198916928550176e-05, "loss": 0.7484, "step": 13195 }, { "epoch": 0.47794277435711696, "grad_norm": 1.5973899568085423, "learning_rate": 1.1197752313950771e-05, "loss": 0.7724, "step": 13196 }, { "epoch": 0.47797899311843534, "grad_norm": 1.0229636961178792, "learning_rate": 1.1196587682869243e-05, "loss": 0.7493, "step": 13197 }, { "epoch": 0.4780152118797537, "grad_norm": 1.9826391054095054, "learning_rate": 1.1195423035321615e-05, "loss": 0.8323, "step": 13198 }, { "epoch": 0.4780514306410721, "grad_norm": 1.146629304839732, "learning_rate": 1.1194258371323917e-05, "loss": 0.7384, "step": 13199 }, { "epoch": 0.47808764940239046, "grad_norm": 1.3643210894055504, "learning_rate": 1.1193093690892173e-05, "loss": 0.673, "step": 13200 }, { "epoch": 0.4781238681637088, "grad_norm": 1.3924176229298462, "learning_rate": 1.1191928994042414e-05, "loss": 0.7527, "step": 13201 }, { "epoch": 0.47816008692502715, "grad_norm": 1.3871053062879897, "learning_rate": 1.1190764280790663e-05, "loss": 0.7357, "step": 13202 }, { "epoch": 0.4781963056863455, "grad_norm": 1.3085361638849526, "learning_rate": 1.118959955115295e-05, "loss": 0.7695, "step": 13203 }, { "epoch": 0.4782325244476639, "grad_norm": 1.3278796796868864, "learning_rate": 1.1188434805145299e-05, "loss": 0.7432, "step": 13204 }, { "epoch": 0.47826874320898227, "grad_norm": 1.4482697248252363, "learning_rate": 1.1187270042783744e-05, "loss": 0.7668, "step": 13205 }, { "epoch": 0.4783049619703006, "grad_norm": 1.1515861273542114, "learning_rate": 1.1186105264084309e-05, "loss": 0.7133, "step": 13206 }, { "epoch": 0.47834118073161896, "grad_norm": 1.6138276197912174, "learning_rate": 1.1184940469063022e-05, "loss": 0.7094, "step": 13207 }, { "epoch": 0.47837739949293734, "grad_norm": 1.236126448617467, "learning_rate": 1.1183775657735916e-05, "loss": 0.7138, "step": 13208 }, { "epoch": 0.4784136182542557, "grad_norm": 1.0092498537249865, "learning_rate": 1.1182610830119013e-05, "loss": 0.7052, "step": 13209 }, { "epoch": 0.4784498370155741, "grad_norm": 1.3775553766028532, "learning_rate": 1.1181445986228349e-05, "loss": 0.7851, "step": 13210 }, { "epoch": 0.4784860557768924, "grad_norm": 1.462565517078791, "learning_rate": 1.1180281126079948e-05, "loss": 0.6905, "step": 13211 }, { "epoch": 0.4785222745382108, "grad_norm": 1.3134166476967606, "learning_rate": 1.1179116249689846e-05, "loss": 0.7694, "step": 13212 }, { "epoch": 0.47855849329952915, "grad_norm": 1.3133219326566732, "learning_rate": 1.1177951357074062e-05, "loss": 0.6666, "step": 13213 }, { "epoch": 0.4785947120608475, "grad_norm": 1.4500278671645126, "learning_rate": 1.1176786448248637e-05, "loss": 0.8185, "step": 13214 }, { "epoch": 0.4786309308221659, "grad_norm": 1.3860940399165318, "learning_rate": 1.1175621523229595e-05, "loss": 0.8143, "step": 13215 }, { "epoch": 0.4786671495834842, "grad_norm": 1.6159833338689975, "learning_rate": 1.1174456582032965e-05, "loss": 0.7904, "step": 13216 }, { "epoch": 0.4787033683448026, "grad_norm": 1.2717264519226357, "learning_rate": 1.1173291624674785e-05, "loss": 0.72, "step": 13217 }, { "epoch": 0.47873958710612097, "grad_norm": 1.3745034768109843, "learning_rate": 1.117212665117108e-05, "loss": 0.783, "step": 13218 }, { "epoch": 0.47877580586743934, "grad_norm": 1.4339892935713976, "learning_rate": 1.1170961661537882e-05, "loss": 0.7491, "step": 13219 }, { "epoch": 0.4788120246287577, "grad_norm": 1.4138323849545724, "learning_rate": 1.1169796655791222e-05, "loss": 0.7408, "step": 13220 }, { "epoch": 0.47884824339007603, "grad_norm": 1.029497643788128, "learning_rate": 1.1168631633947134e-05, "loss": 0.7212, "step": 13221 }, { "epoch": 0.4788844621513944, "grad_norm": 1.3134328466181402, "learning_rate": 1.1167466596021645e-05, "loss": 0.7611, "step": 13222 }, { "epoch": 0.4789206809127128, "grad_norm": 1.682464084534176, "learning_rate": 1.116630154203079e-05, "loss": 0.6362, "step": 13223 }, { "epoch": 0.47895689967403116, "grad_norm": 1.385887674674633, "learning_rate": 1.1165136471990603e-05, "loss": 0.7838, "step": 13224 }, { "epoch": 0.47899311843534953, "grad_norm": 1.4437660578664537, "learning_rate": 1.1163971385917114e-05, "loss": 0.8, "step": 13225 }, { "epoch": 0.47902933719666785, "grad_norm": 1.4652653373583253, "learning_rate": 1.1162806283826355e-05, "loss": 0.7604, "step": 13226 }, { "epoch": 0.4790655559579862, "grad_norm": 1.3597829455987829, "learning_rate": 1.116164116573436e-05, "loss": 0.7058, "step": 13227 }, { "epoch": 0.4791017747193046, "grad_norm": 1.339834595529406, "learning_rate": 1.1160476031657164e-05, "loss": 0.7359, "step": 13228 }, { "epoch": 0.47913799348062297, "grad_norm": 1.5206858310847873, "learning_rate": 1.1159310881610795e-05, "loss": 0.7541, "step": 13229 }, { "epoch": 0.47917421224194134, "grad_norm": 1.1058497590551757, "learning_rate": 1.1158145715611293e-05, "loss": 0.725, "step": 13230 }, { "epoch": 0.47921043100325966, "grad_norm": 1.4999725120727891, "learning_rate": 1.1156980533674684e-05, "loss": 0.735, "step": 13231 }, { "epoch": 0.47924664976457804, "grad_norm": 1.41147355978191, "learning_rate": 1.115581533581701e-05, "loss": 0.7466, "step": 13232 }, { "epoch": 0.4792828685258964, "grad_norm": 1.2991172726811615, "learning_rate": 1.11546501220543e-05, "loss": 0.6799, "step": 13233 }, { "epoch": 0.4793190872872148, "grad_norm": 1.396747031553375, "learning_rate": 1.1153484892402589e-05, "loss": 0.8086, "step": 13234 }, { "epoch": 0.47935530604853316, "grad_norm": 1.3550005565906977, "learning_rate": 1.1152319646877913e-05, "loss": 0.7411, "step": 13235 }, { "epoch": 0.4793915248098515, "grad_norm": 1.4666079476969756, "learning_rate": 1.1151154385496305e-05, "loss": 0.8047, "step": 13236 }, { "epoch": 0.47942774357116985, "grad_norm": 1.4670355713074006, "learning_rate": 1.1149989108273804e-05, "loss": 0.7519, "step": 13237 }, { "epoch": 0.4794639623324882, "grad_norm": 1.3116357404470336, "learning_rate": 1.114882381522644e-05, "loss": 0.7325, "step": 13238 }, { "epoch": 0.4795001810938066, "grad_norm": 1.1317744735168094, "learning_rate": 1.1147658506370252e-05, "loss": 0.7821, "step": 13239 }, { "epoch": 0.479536399855125, "grad_norm": 1.3167728159903913, "learning_rate": 1.1146493181721271e-05, "loss": 0.7085, "step": 13240 }, { "epoch": 0.4795726186164433, "grad_norm": 1.2538593148274002, "learning_rate": 1.1145327841295543e-05, "loss": 0.731, "step": 13241 }, { "epoch": 0.47960883737776167, "grad_norm": 2.0177684451366296, "learning_rate": 1.1144162485109093e-05, "loss": 0.7708, "step": 13242 }, { "epoch": 0.47964505613908004, "grad_norm": 0.9137533101271481, "learning_rate": 1.1142997113177962e-05, "loss": 0.755, "step": 13243 }, { "epoch": 0.4796812749003984, "grad_norm": 1.5459839386950422, "learning_rate": 1.1141831725518189e-05, "loss": 0.8316, "step": 13244 }, { "epoch": 0.4797174936617168, "grad_norm": 0.9823510589894832, "learning_rate": 1.1140666322145804e-05, "loss": 0.7157, "step": 13245 }, { "epoch": 0.4797537124230351, "grad_norm": 1.4404321444215673, "learning_rate": 1.1139500903076852e-05, "loss": 0.7703, "step": 13246 }, { "epoch": 0.4797899311843535, "grad_norm": 1.3352942284392535, "learning_rate": 1.1138335468327363e-05, "loss": 0.7547, "step": 13247 }, { "epoch": 0.47982614994567185, "grad_norm": 1.3700090944824073, "learning_rate": 1.1137170017913382e-05, "loss": 0.7905, "step": 13248 }, { "epoch": 0.4798623687069902, "grad_norm": 1.4131624356635937, "learning_rate": 1.113600455185094e-05, "loss": 0.6994, "step": 13249 }, { "epoch": 0.4798985874683086, "grad_norm": 1.2593283079025022, "learning_rate": 1.1134839070156075e-05, "loss": 0.7363, "step": 13250 }, { "epoch": 0.4799348062296269, "grad_norm": 1.3659612797002703, "learning_rate": 1.1133673572844829e-05, "loss": 0.7301, "step": 13251 }, { "epoch": 0.4799710249909453, "grad_norm": 1.0613990380589013, "learning_rate": 1.113250805993324e-05, "loss": 0.6994, "step": 13252 }, { "epoch": 0.48000724375226367, "grad_norm": 1.3868985731095018, "learning_rate": 1.1131342531437344e-05, "loss": 0.7621, "step": 13253 }, { "epoch": 0.48004346251358204, "grad_norm": 1.277202526941702, "learning_rate": 1.1130176987373181e-05, "loss": 0.7324, "step": 13254 }, { "epoch": 0.4800796812749004, "grad_norm": 1.541427883512612, "learning_rate": 1.1129011427756788e-05, "loss": 0.7053, "step": 13255 }, { "epoch": 0.48011590003621873, "grad_norm": 1.3184022112350975, "learning_rate": 1.1127845852604207e-05, "loss": 0.7122, "step": 13256 }, { "epoch": 0.4801521187975371, "grad_norm": 1.0869239254976957, "learning_rate": 1.1126680261931478e-05, "loss": 0.7375, "step": 13257 }, { "epoch": 0.4801883375588555, "grad_norm": 1.2477427217953017, "learning_rate": 1.1125514655754634e-05, "loss": 0.7422, "step": 13258 }, { "epoch": 0.48022455632017386, "grad_norm": 1.3804668485059524, "learning_rate": 1.1124349034089724e-05, "loss": 0.7553, "step": 13259 }, { "epoch": 0.48026077508149223, "grad_norm": 1.4225784751836772, "learning_rate": 1.1123183396952779e-05, "loss": 0.7157, "step": 13260 }, { "epoch": 0.48029699384281055, "grad_norm": 1.3894252550967232, "learning_rate": 1.1122017744359848e-05, "loss": 0.7711, "step": 13261 }, { "epoch": 0.4803332126041289, "grad_norm": 1.0224631847691181, "learning_rate": 1.1120852076326965e-05, "loss": 0.671, "step": 13262 }, { "epoch": 0.4803694313654473, "grad_norm": 1.4110063595195907, "learning_rate": 1.1119686392870172e-05, "loss": 0.7257, "step": 13263 }, { "epoch": 0.48040565012676567, "grad_norm": 1.3525480814058122, "learning_rate": 1.111852069400551e-05, "loss": 0.6864, "step": 13264 }, { "epoch": 0.48044186888808404, "grad_norm": 1.5000047907156735, "learning_rate": 1.111735497974902e-05, "loss": 0.7139, "step": 13265 }, { "epoch": 0.48047808764940236, "grad_norm": 1.3661685797973968, "learning_rate": 1.1116189250116744e-05, "loss": 0.8063, "step": 13266 }, { "epoch": 0.48051430641072074, "grad_norm": 0.9772374347429134, "learning_rate": 1.1115023505124724e-05, "loss": 0.7414, "step": 13267 }, { "epoch": 0.4805505251720391, "grad_norm": 1.4557420700731991, "learning_rate": 1.1113857744789002e-05, "loss": 0.7429, "step": 13268 }, { "epoch": 0.4805867439333575, "grad_norm": 1.459025863711216, "learning_rate": 1.1112691969125618e-05, "loss": 0.7645, "step": 13269 }, { "epoch": 0.48062296269467586, "grad_norm": 1.1196431532031135, "learning_rate": 1.1111526178150615e-05, "loss": 0.7009, "step": 13270 }, { "epoch": 0.4806591814559942, "grad_norm": 1.345271649000722, "learning_rate": 1.1110360371880034e-05, "loss": 0.8167, "step": 13271 }, { "epoch": 0.48069540021731255, "grad_norm": 1.5463288744751806, "learning_rate": 1.1109194550329919e-05, "loss": 0.7853, "step": 13272 }, { "epoch": 0.4807316189786309, "grad_norm": 1.3604468341719143, "learning_rate": 1.1108028713516312e-05, "loss": 0.7483, "step": 13273 }, { "epoch": 0.4807678377399493, "grad_norm": 1.031919991150511, "learning_rate": 1.1106862861455255e-05, "loss": 0.7601, "step": 13274 }, { "epoch": 0.4808040565012677, "grad_norm": 1.4314218681852537, "learning_rate": 1.1105696994162796e-05, "loss": 0.8132, "step": 13275 }, { "epoch": 0.480840275262586, "grad_norm": 1.2941996630309265, "learning_rate": 1.1104531111654973e-05, "loss": 0.6953, "step": 13276 }, { "epoch": 0.48087649402390437, "grad_norm": 1.3407960689489486, "learning_rate": 1.1103365213947833e-05, "loss": 0.7504, "step": 13277 }, { "epoch": 0.48091271278522274, "grad_norm": 1.5541164982600453, "learning_rate": 1.1102199301057413e-05, "loss": 0.7996, "step": 13278 }, { "epoch": 0.4809489315465411, "grad_norm": 1.2560629991052619, "learning_rate": 1.1101033372999767e-05, "loss": 0.7493, "step": 13279 }, { "epoch": 0.4809851503078595, "grad_norm": 0.9651459394908241, "learning_rate": 1.1099867429790931e-05, "loss": 0.7166, "step": 13280 }, { "epoch": 0.4810213690691778, "grad_norm": 1.3727360898671992, "learning_rate": 1.1098701471446958e-05, "loss": 0.6433, "step": 13281 }, { "epoch": 0.4810575878304962, "grad_norm": 1.4193608908705793, "learning_rate": 1.1097535497983883e-05, "loss": 0.7675, "step": 13282 }, { "epoch": 0.48109380659181455, "grad_norm": 1.4084155799394857, "learning_rate": 1.1096369509417757e-05, "loss": 0.771, "step": 13283 }, { "epoch": 0.48113002535313293, "grad_norm": 1.2843041256347154, "learning_rate": 1.109520350576462e-05, "loss": 0.7499, "step": 13284 }, { "epoch": 0.4811662441144513, "grad_norm": 1.3934179144011365, "learning_rate": 1.1094037487040523e-05, "loss": 0.7929, "step": 13285 }, { "epoch": 0.4812024628757696, "grad_norm": 1.26043739963522, "learning_rate": 1.1092871453261509e-05, "loss": 0.7149, "step": 13286 }, { "epoch": 0.481238681637088, "grad_norm": 1.381450955306461, "learning_rate": 1.1091705404443622e-05, "loss": 0.7412, "step": 13287 }, { "epoch": 0.48127490039840637, "grad_norm": 1.0125925678193293, "learning_rate": 1.1090539340602912e-05, "loss": 0.7421, "step": 13288 }, { "epoch": 0.48131111915972474, "grad_norm": 1.420069691416345, "learning_rate": 1.108937326175542e-05, "loss": 0.7124, "step": 13289 }, { "epoch": 0.4813473379210431, "grad_norm": 1.0818180759324296, "learning_rate": 1.1088207167917198e-05, "loss": 0.7395, "step": 13290 }, { "epoch": 0.4813835566823615, "grad_norm": 1.3321694836526876, "learning_rate": 1.1087041059104287e-05, "loss": 0.7337, "step": 13291 }, { "epoch": 0.4814197754436798, "grad_norm": 1.1541227190256862, "learning_rate": 1.1085874935332736e-05, "loss": 0.7243, "step": 13292 }, { "epoch": 0.4814559942049982, "grad_norm": 1.0799913229756046, "learning_rate": 1.108470879661859e-05, "loss": 0.7329, "step": 13293 }, { "epoch": 0.48149221296631656, "grad_norm": 1.4017120621480954, "learning_rate": 1.1083542642977902e-05, "loss": 0.7018, "step": 13294 }, { "epoch": 0.48152843172763493, "grad_norm": 1.1157941667763969, "learning_rate": 1.108237647442671e-05, "loss": 0.7565, "step": 13295 }, { "epoch": 0.4815646504889533, "grad_norm": 1.3557864608325993, "learning_rate": 1.1081210290981068e-05, "loss": 0.6368, "step": 13296 }, { "epoch": 0.4816008692502716, "grad_norm": 0.9491109872395794, "learning_rate": 1.1080044092657025e-05, "loss": 0.7016, "step": 13297 }, { "epoch": 0.48163708801159, "grad_norm": 1.1679876606663795, "learning_rate": 1.1078877879470625e-05, "loss": 0.6483, "step": 13298 }, { "epoch": 0.48167330677290837, "grad_norm": 1.4123457391602285, "learning_rate": 1.107771165143792e-05, "loss": 0.7966, "step": 13299 }, { "epoch": 0.48170952553422675, "grad_norm": 1.3639707911906538, "learning_rate": 1.1076545408574951e-05, "loss": 0.7672, "step": 13300 }, { "epoch": 0.4817457442955451, "grad_norm": 1.2850599528239617, "learning_rate": 1.1075379150897774e-05, "loss": 0.7901, "step": 13301 }, { "epoch": 0.48178196305686344, "grad_norm": 1.4226239348946292, "learning_rate": 1.1074212878422432e-05, "loss": 0.8047, "step": 13302 }, { "epoch": 0.4818181818181818, "grad_norm": 1.280630959435965, "learning_rate": 1.1073046591164982e-05, "loss": 0.8066, "step": 13303 }, { "epoch": 0.4818544005795002, "grad_norm": 1.2225518023477253, "learning_rate": 1.1071880289141463e-05, "loss": 0.7255, "step": 13304 }, { "epoch": 0.48189061934081856, "grad_norm": 1.354991977402209, "learning_rate": 1.107071397236793e-05, "loss": 0.7556, "step": 13305 }, { "epoch": 0.48192683810213693, "grad_norm": 1.3983963322307271, "learning_rate": 1.1069547640860437e-05, "loss": 0.7611, "step": 13306 }, { "epoch": 0.48196305686345525, "grad_norm": 1.1639905393432866, "learning_rate": 1.1068381294635022e-05, "loss": 0.7779, "step": 13307 }, { "epoch": 0.4819992756247736, "grad_norm": 1.3410432037108941, "learning_rate": 1.1067214933707746e-05, "loss": 0.7167, "step": 13308 }, { "epoch": 0.482035494386092, "grad_norm": 1.4455929690233382, "learning_rate": 1.1066048558094652e-05, "loss": 0.7279, "step": 13309 }, { "epoch": 0.4820717131474104, "grad_norm": 1.27882145572587, "learning_rate": 1.1064882167811796e-05, "loss": 0.689, "step": 13310 }, { "epoch": 0.48210793190872875, "grad_norm": 1.4488835981882113, "learning_rate": 1.1063715762875225e-05, "loss": 0.8594, "step": 13311 }, { "epoch": 0.48214415067004707, "grad_norm": 1.2844160101833364, "learning_rate": 1.106254934330099e-05, "loss": 0.7375, "step": 13312 }, { "epoch": 0.48218036943136544, "grad_norm": 1.364788159181057, "learning_rate": 1.106138290910514e-05, "loss": 0.7635, "step": 13313 }, { "epoch": 0.4822165881926838, "grad_norm": 1.1436271125359416, "learning_rate": 1.1060216460303735e-05, "loss": 0.7315, "step": 13314 }, { "epoch": 0.4822528069540022, "grad_norm": 1.26214883105684, "learning_rate": 1.1059049996912814e-05, "loss": 0.7251, "step": 13315 }, { "epoch": 0.48228902571532056, "grad_norm": 1.4892778389006345, "learning_rate": 1.1057883518948433e-05, "loss": 0.6768, "step": 13316 }, { "epoch": 0.4823252444766389, "grad_norm": 1.350950005146056, "learning_rate": 1.105671702642665e-05, "loss": 0.7373, "step": 13317 }, { "epoch": 0.48236146323795726, "grad_norm": 1.2699286895142003, "learning_rate": 1.1055550519363511e-05, "loss": 0.7135, "step": 13318 }, { "epoch": 0.48239768199927563, "grad_norm": 1.3532161670371483, "learning_rate": 1.1054383997775068e-05, "loss": 0.8149, "step": 13319 }, { "epoch": 0.482433900760594, "grad_norm": 0.9491058324942873, "learning_rate": 1.1053217461677377e-05, "loss": 0.6836, "step": 13320 }, { "epoch": 0.4824701195219124, "grad_norm": 1.3725547722621116, "learning_rate": 1.1052050911086487e-05, "loss": 0.8046, "step": 13321 }, { "epoch": 0.4825063382832307, "grad_norm": 1.3431112197492572, "learning_rate": 1.1050884346018452e-05, "loss": 0.7764, "step": 13322 }, { "epoch": 0.48254255704454907, "grad_norm": 1.4890691356270915, "learning_rate": 1.1049717766489325e-05, "loss": 0.7776, "step": 13323 }, { "epoch": 0.48257877580586744, "grad_norm": 1.1920717882897003, "learning_rate": 1.104855117251516e-05, "loss": 0.7895, "step": 13324 }, { "epoch": 0.4826149945671858, "grad_norm": 1.0714673930675356, "learning_rate": 1.104738456411201e-05, "loss": 0.8125, "step": 13325 }, { "epoch": 0.4826512133285042, "grad_norm": 1.5348589611726873, "learning_rate": 1.1046217941295926e-05, "loss": 0.8273, "step": 13326 }, { "epoch": 0.4826874320898225, "grad_norm": 1.4690804464870597, "learning_rate": 1.1045051304082966e-05, "loss": 0.7153, "step": 13327 }, { "epoch": 0.4827236508511409, "grad_norm": 1.095967570292259, "learning_rate": 1.104388465248918e-05, "loss": 0.7805, "step": 13328 }, { "epoch": 0.48275986961245926, "grad_norm": 1.6269474395799155, "learning_rate": 1.1042717986530624e-05, "loss": 0.7846, "step": 13329 }, { "epoch": 0.48279608837377763, "grad_norm": 1.3831961323222803, "learning_rate": 1.1041551306223353e-05, "loss": 0.8034, "step": 13330 }, { "epoch": 0.482832307135096, "grad_norm": 1.3237637959235076, "learning_rate": 1.1040384611583419e-05, "loss": 0.6834, "step": 13331 }, { "epoch": 0.4828685258964143, "grad_norm": 1.3890511099548992, "learning_rate": 1.103921790262688e-05, "loss": 0.7788, "step": 13332 }, { "epoch": 0.4829047446577327, "grad_norm": 1.3663243600474027, "learning_rate": 1.103805117936979e-05, "loss": 0.7008, "step": 13333 }, { "epoch": 0.4829409634190511, "grad_norm": 1.4940247010781644, "learning_rate": 1.1036884441828203e-05, "loss": 0.7267, "step": 13334 }, { "epoch": 0.48297718218036945, "grad_norm": 1.4056248354742091, "learning_rate": 1.1035717690018177e-05, "loss": 0.7659, "step": 13335 }, { "epoch": 0.4830134009416878, "grad_norm": 1.3637053450337278, "learning_rate": 1.1034550923955762e-05, "loss": 0.6717, "step": 13336 }, { "epoch": 0.48304961970300614, "grad_norm": 1.2356110160887264, "learning_rate": 1.103338414365702e-05, "loss": 0.7702, "step": 13337 }, { "epoch": 0.4830858384643245, "grad_norm": 1.0380772170739456, "learning_rate": 1.1032217349138003e-05, "loss": 0.6631, "step": 13338 }, { "epoch": 0.4831220572256429, "grad_norm": 1.4072160091705954, "learning_rate": 1.1031050540414769e-05, "loss": 0.7839, "step": 13339 }, { "epoch": 0.48315827598696126, "grad_norm": 1.025994755705084, "learning_rate": 1.1029883717503374e-05, "loss": 0.7228, "step": 13340 }, { "epoch": 0.48319449474827963, "grad_norm": 1.332320250485897, "learning_rate": 1.1028716880419874e-05, "loss": 0.7695, "step": 13341 }, { "epoch": 0.48323071350959795, "grad_norm": 1.5990946643771042, "learning_rate": 1.1027550029180323e-05, "loss": 0.8095, "step": 13342 }, { "epoch": 0.4832669322709163, "grad_norm": 1.331262308526112, "learning_rate": 1.1026383163800782e-05, "loss": 0.7083, "step": 13343 }, { "epoch": 0.4833031510322347, "grad_norm": 1.1279010942476182, "learning_rate": 1.102521628429731e-05, "loss": 0.7087, "step": 13344 }, { "epoch": 0.4833393697935531, "grad_norm": 1.4663344097373692, "learning_rate": 1.1024049390685957e-05, "loss": 0.7054, "step": 13345 }, { "epoch": 0.48337558855487145, "grad_norm": 1.0182869805462775, "learning_rate": 1.1022882482982786e-05, "loss": 0.7425, "step": 13346 }, { "epoch": 0.48341180731618977, "grad_norm": 1.378023421507887, "learning_rate": 1.1021715561203854e-05, "loss": 0.7586, "step": 13347 }, { "epoch": 0.48344802607750814, "grad_norm": 1.0608174136655677, "learning_rate": 1.102054862536522e-05, "loss": 0.7749, "step": 13348 }, { "epoch": 0.4834842448388265, "grad_norm": 1.2821259724251262, "learning_rate": 1.1019381675482937e-05, "loss": 0.7125, "step": 13349 }, { "epoch": 0.4835204636001449, "grad_norm": 1.3250533845275412, "learning_rate": 1.1018214711573067e-05, "loss": 0.7353, "step": 13350 }, { "epoch": 0.48355668236146326, "grad_norm": 1.2630023229035676, "learning_rate": 1.1017047733651666e-05, "loss": 0.723, "step": 13351 }, { "epoch": 0.4835929011227816, "grad_norm": 1.3678976530665463, "learning_rate": 1.1015880741734796e-05, "loss": 0.7461, "step": 13352 }, { "epoch": 0.48362911988409996, "grad_norm": 1.0371223733333474, "learning_rate": 1.1014713735838516e-05, "loss": 0.7413, "step": 13353 }, { "epoch": 0.48366533864541833, "grad_norm": 1.447775568164793, "learning_rate": 1.1013546715978883e-05, "loss": 0.7027, "step": 13354 }, { "epoch": 0.4837015574067367, "grad_norm": 1.4161231322419723, "learning_rate": 1.1012379682171956e-05, "loss": 0.6784, "step": 13355 }, { "epoch": 0.4837377761680551, "grad_norm": 1.5009226147121848, "learning_rate": 1.1011212634433795e-05, "loss": 0.7058, "step": 13356 }, { "epoch": 0.4837739949293734, "grad_norm": 0.9732453870416898, "learning_rate": 1.1010045572780461e-05, "loss": 0.6841, "step": 13357 }, { "epoch": 0.48381021369069177, "grad_norm": 1.3570068081558138, "learning_rate": 1.1008878497228009e-05, "loss": 0.6605, "step": 13358 }, { "epoch": 0.48384643245201014, "grad_norm": 1.5948395499238313, "learning_rate": 1.1007711407792504e-05, "loss": 0.7303, "step": 13359 }, { "epoch": 0.4838826512133285, "grad_norm": 1.5232046536251438, "learning_rate": 1.1006544304490005e-05, "loss": 0.7896, "step": 13360 }, { "epoch": 0.4839188699746469, "grad_norm": 1.0421545682887845, "learning_rate": 1.1005377187336575e-05, "loss": 0.7623, "step": 13361 }, { "epoch": 0.4839550887359652, "grad_norm": 1.3149197509716286, "learning_rate": 1.1004210056348265e-05, "loss": 0.7753, "step": 13362 }, { "epoch": 0.4839913074972836, "grad_norm": 0.9730092285715187, "learning_rate": 1.1003042911541146e-05, "loss": 0.698, "step": 13363 }, { "epoch": 0.48402752625860196, "grad_norm": 1.6701889449152825, "learning_rate": 1.1001875752931277e-05, "loss": 0.7557, "step": 13364 }, { "epoch": 0.48406374501992033, "grad_norm": 1.4048882991196259, "learning_rate": 1.1000708580534713e-05, "loss": 0.8048, "step": 13365 }, { "epoch": 0.4840999637812387, "grad_norm": 1.2660973838811438, "learning_rate": 1.0999541394367523e-05, "loss": 0.754, "step": 13366 }, { "epoch": 0.484136182542557, "grad_norm": 1.3611328848363797, "learning_rate": 1.0998374194445764e-05, "loss": 0.7335, "step": 13367 }, { "epoch": 0.4841724013038754, "grad_norm": 1.3652580079091976, "learning_rate": 1.09972069807855e-05, "loss": 0.7223, "step": 13368 }, { "epoch": 0.4842086200651938, "grad_norm": 1.3947740692020234, "learning_rate": 1.099603975340279e-05, "loss": 0.7119, "step": 13369 }, { "epoch": 0.48424483882651215, "grad_norm": 1.3277634661529785, "learning_rate": 1.09948725123137e-05, "loss": 0.7032, "step": 13370 }, { "epoch": 0.4842810575878305, "grad_norm": 1.4516986440207096, "learning_rate": 1.0993705257534286e-05, "loss": 0.7217, "step": 13371 }, { "epoch": 0.48431727634914884, "grad_norm": 1.464367544016924, "learning_rate": 1.0992537989080618e-05, "loss": 0.6557, "step": 13372 }, { "epoch": 0.4843534951104672, "grad_norm": 1.391246352788612, "learning_rate": 1.0991370706968754e-05, "loss": 0.7542, "step": 13373 }, { "epoch": 0.4843897138717856, "grad_norm": 1.071862801672778, "learning_rate": 1.0990203411214757e-05, "loss": 0.6837, "step": 13374 }, { "epoch": 0.48442593263310396, "grad_norm": 1.380717486061478, "learning_rate": 1.0989036101834693e-05, "loss": 0.7368, "step": 13375 }, { "epoch": 0.48446215139442234, "grad_norm": 1.4708561026154388, "learning_rate": 1.0987868778844621e-05, "loss": 0.7637, "step": 13376 }, { "epoch": 0.48449837015574065, "grad_norm": 1.3015971159710935, "learning_rate": 1.098670144226061e-05, "loss": 0.7833, "step": 13377 }, { "epoch": 0.48453458891705903, "grad_norm": 1.4152960507722947, "learning_rate": 1.0985534092098715e-05, "loss": 0.8411, "step": 13378 }, { "epoch": 0.4845708076783774, "grad_norm": 1.0046068320901116, "learning_rate": 1.0984366728375012e-05, "loss": 0.7843, "step": 13379 }, { "epoch": 0.4846070264396958, "grad_norm": 1.343322312968841, "learning_rate": 1.098319935110555e-05, "loss": 0.7258, "step": 13380 }, { "epoch": 0.48464324520101415, "grad_norm": 1.3591729197530298, "learning_rate": 1.0982031960306408e-05, "loss": 0.6241, "step": 13381 }, { "epoch": 0.48467946396233247, "grad_norm": 1.3534906407041654, "learning_rate": 1.0980864555993638e-05, "loss": 0.7403, "step": 13382 }, { "epoch": 0.48471568272365084, "grad_norm": 1.2953683161242384, "learning_rate": 1.0979697138183311e-05, "loss": 0.6545, "step": 13383 }, { "epoch": 0.4847519014849692, "grad_norm": 1.325360287887732, "learning_rate": 1.0978529706891493e-05, "loss": 0.7332, "step": 13384 }, { "epoch": 0.4847881202462876, "grad_norm": 1.514058675874343, "learning_rate": 1.0977362262134243e-05, "loss": 0.6711, "step": 13385 }, { "epoch": 0.48482433900760596, "grad_norm": 1.2841652517631452, "learning_rate": 1.0976194803927634e-05, "loss": 0.6944, "step": 13386 }, { "epoch": 0.4848605577689243, "grad_norm": 1.0623013618821868, "learning_rate": 1.0975027332287721e-05, "loss": 0.7622, "step": 13387 }, { "epoch": 0.48489677653024266, "grad_norm": 1.1629948365652722, "learning_rate": 1.0973859847230582e-05, "loss": 0.7618, "step": 13388 }, { "epoch": 0.48493299529156103, "grad_norm": 1.0473890268055877, "learning_rate": 1.0972692348772271e-05, "loss": 0.7413, "step": 13389 }, { "epoch": 0.4849692140528794, "grad_norm": 1.4444173063889694, "learning_rate": 1.097152483692886e-05, "loss": 0.8535, "step": 13390 }, { "epoch": 0.4850054328141978, "grad_norm": 1.0204010814084699, "learning_rate": 1.0970357311716414e-05, "loss": 0.7666, "step": 13391 }, { "epoch": 0.4850416515755161, "grad_norm": 1.3151363202844577, "learning_rate": 1.0969189773150998e-05, "loss": 0.696, "step": 13392 }, { "epoch": 0.48507787033683447, "grad_norm": 0.9652371710881901, "learning_rate": 1.0968022221248678e-05, "loss": 0.7654, "step": 13393 }, { "epoch": 0.48511408909815285, "grad_norm": 1.40840273300868, "learning_rate": 1.0966854656025523e-05, "loss": 0.7133, "step": 13394 }, { "epoch": 0.4851503078594712, "grad_norm": 1.433006563949093, "learning_rate": 1.0965687077497597e-05, "loss": 0.763, "step": 13395 }, { "epoch": 0.4851865266207896, "grad_norm": 1.3921783657288085, "learning_rate": 1.096451948568097e-05, "loss": 0.6965, "step": 13396 }, { "epoch": 0.4852227453821079, "grad_norm": 1.2882771910365782, "learning_rate": 1.0963351880591708e-05, "loss": 0.7574, "step": 13397 }, { "epoch": 0.4852589641434263, "grad_norm": 1.5974394550924196, "learning_rate": 1.0962184262245876e-05, "loss": 0.7194, "step": 13398 }, { "epoch": 0.48529518290474466, "grad_norm": 1.384578861049286, "learning_rate": 1.0961016630659544e-05, "loss": 0.7437, "step": 13399 }, { "epoch": 0.48533140166606303, "grad_norm": 1.3594562572002307, "learning_rate": 1.0959848985848778e-05, "loss": 0.716, "step": 13400 }, { "epoch": 0.4853676204273814, "grad_norm": 1.3373282468783887, "learning_rate": 1.0958681327829647e-05, "loss": 0.7391, "step": 13401 }, { "epoch": 0.4854038391886997, "grad_norm": 1.343929727866164, "learning_rate": 1.095751365661822e-05, "loss": 0.7311, "step": 13402 }, { "epoch": 0.4854400579500181, "grad_norm": 0.982210346982392, "learning_rate": 1.0956345972230561e-05, "loss": 0.7168, "step": 13403 }, { "epoch": 0.4854762767113365, "grad_norm": 1.4983405182275085, "learning_rate": 1.0955178274682744e-05, "loss": 0.6718, "step": 13404 }, { "epoch": 0.48551249547265485, "grad_norm": 1.2847675305204798, "learning_rate": 1.0954010563990831e-05, "loss": 0.6419, "step": 13405 }, { "epoch": 0.4855487142339732, "grad_norm": 1.281708304645044, "learning_rate": 1.0952842840170896e-05, "loss": 0.8065, "step": 13406 }, { "epoch": 0.48558493299529154, "grad_norm": 2.3762198745755883, "learning_rate": 1.0951675103239009e-05, "loss": 0.8007, "step": 13407 }, { "epoch": 0.4856211517566099, "grad_norm": 1.1443597450260437, "learning_rate": 1.0950507353211233e-05, "loss": 0.7735, "step": 13408 }, { "epoch": 0.4856573705179283, "grad_norm": 1.438507977082417, "learning_rate": 1.094933959010364e-05, "loss": 0.7379, "step": 13409 }, { "epoch": 0.48569358927924666, "grad_norm": 1.560726205598141, "learning_rate": 1.0948171813932304e-05, "loss": 0.7955, "step": 13410 }, { "epoch": 0.48572980804056504, "grad_norm": 1.3844565193165226, "learning_rate": 1.0947004024713289e-05, "loss": 0.7898, "step": 13411 }, { "epoch": 0.48576602680188335, "grad_norm": 0.9718757518447447, "learning_rate": 1.0945836222462664e-05, "loss": 0.797, "step": 13412 }, { "epoch": 0.48580224556320173, "grad_norm": 1.3124777055447672, "learning_rate": 1.0944668407196503e-05, "loss": 0.7626, "step": 13413 }, { "epoch": 0.4858384643245201, "grad_norm": 1.7717714834082172, "learning_rate": 1.0943500578930874e-05, "loss": 0.7941, "step": 13414 }, { "epoch": 0.4858746830858385, "grad_norm": 1.3564389203266336, "learning_rate": 1.094233273768185e-05, "loss": 0.7781, "step": 13415 }, { "epoch": 0.48591090184715685, "grad_norm": 1.0302179486012255, "learning_rate": 1.0941164883465499e-05, "loss": 0.7591, "step": 13416 }, { "epoch": 0.48594712060847517, "grad_norm": 4.254168412158071, "learning_rate": 1.0939997016297892e-05, "loss": 0.7862, "step": 13417 }, { "epoch": 0.48598333936979354, "grad_norm": 1.094077229669913, "learning_rate": 1.0938829136195099e-05, "loss": 0.7145, "step": 13418 }, { "epoch": 0.4860195581311119, "grad_norm": 1.0135592980843595, "learning_rate": 1.0937661243173195e-05, "loss": 0.7113, "step": 13419 }, { "epoch": 0.4860557768924303, "grad_norm": 1.3457179366294592, "learning_rate": 1.0936493337248247e-05, "loss": 0.7346, "step": 13420 }, { "epoch": 0.48609199565374867, "grad_norm": 1.3188419115943462, "learning_rate": 1.0935325418436326e-05, "loss": 0.6987, "step": 13421 }, { "epoch": 0.486128214415067, "grad_norm": 1.0647871148585972, "learning_rate": 1.0934157486753507e-05, "loss": 0.7676, "step": 13422 }, { "epoch": 0.48616443317638536, "grad_norm": 1.2907241948272916, "learning_rate": 1.0932989542215861e-05, "loss": 0.7307, "step": 13423 }, { "epoch": 0.48620065193770373, "grad_norm": 1.3054723777241184, "learning_rate": 1.0931821584839457e-05, "loss": 0.7339, "step": 13424 }, { "epoch": 0.4862368706990221, "grad_norm": 1.3385572749525403, "learning_rate": 1.093065361464037e-05, "loss": 0.6744, "step": 13425 }, { "epoch": 0.4862730894603405, "grad_norm": 1.0953407674664468, "learning_rate": 1.0929485631634673e-05, "loss": 0.7692, "step": 13426 }, { "epoch": 0.4863093082216588, "grad_norm": 1.0215066620504236, "learning_rate": 1.0928317635838434e-05, "loss": 0.7692, "step": 13427 }, { "epoch": 0.48634552698297717, "grad_norm": 1.72401472188653, "learning_rate": 1.092714962726773e-05, "loss": 0.792, "step": 13428 }, { "epoch": 0.48638174574429555, "grad_norm": 1.1647421519667829, "learning_rate": 1.0925981605938635e-05, "loss": 0.672, "step": 13429 }, { "epoch": 0.4864179645056139, "grad_norm": 1.434750544974833, "learning_rate": 1.0924813571867216e-05, "loss": 0.7295, "step": 13430 }, { "epoch": 0.4864541832669323, "grad_norm": 1.2219364874905225, "learning_rate": 1.0923645525069552e-05, "loss": 0.7292, "step": 13431 }, { "epoch": 0.4864904020282506, "grad_norm": 1.459787445859708, "learning_rate": 1.0922477465561712e-05, "loss": 0.8009, "step": 13432 }, { "epoch": 0.486526620789569, "grad_norm": 1.3396068153583813, "learning_rate": 1.0921309393359774e-05, "loss": 0.7997, "step": 13433 }, { "epoch": 0.48656283955088736, "grad_norm": 1.0032682337267096, "learning_rate": 1.0920141308479805e-05, "loss": 0.7156, "step": 13434 }, { "epoch": 0.48659905831220573, "grad_norm": 1.3768817024104623, "learning_rate": 1.0918973210937888e-05, "loss": 0.8033, "step": 13435 }, { "epoch": 0.4866352770735241, "grad_norm": 1.3238817379705001, "learning_rate": 1.091780510075009e-05, "loss": 0.7245, "step": 13436 }, { "epoch": 0.4866714958348424, "grad_norm": 1.4697048388425005, "learning_rate": 1.0916636977932489e-05, "loss": 0.7891, "step": 13437 }, { "epoch": 0.4867077145961608, "grad_norm": 1.2834301135227026, "learning_rate": 1.0915468842501153e-05, "loss": 0.7784, "step": 13438 }, { "epoch": 0.4867439333574792, "grad_norm": 1.5172393585185617, "learning_rate": 1.0914300694472166e-05, "loss": 0.7952, "step": 13439 }, { "epoch": 0.48678015211879755, "grad_norm": 1.1587045092510968, "learning_rate": 1.0913132533861597e-05, "loss": 0.7868, "step": 13440 }, { "epoch": 0.4868163708801159, "grad_norm": 0.9651987413589164, "learning_rate": 1.091196436068552e-05, "loss": 0.7545, "step": 13441 }, { "epoch": 0.48685258964143424, "grad_norm": 1.1273506544871426, "learning_rate": 1.0910796174960015e-05, "loss": 0.7433, "step": 13442 }, { "epoch": 0.4868888084027526, "grad_norm": 1.0774218060742775, "learning_rate": 1.0909627976701152e-05, "loss": 0.731, "step": 13443 }, { "epoch": 0.486925027164071, "grad_norm": 1.5531608139913249, "learning_rate": 1.090845976592501e-05, "loss": 0.7496, "step": 13444 }, { "epoch": 0.48696124592538936, "grad_norm": 1.344083885224741, "learning_rate": 1.090729154264766e-05, "loss": 0.8146, "step": 13445 }, { "epoch": 0.48699746468670774, "grad_norm": 1.0459600933206892, "learning_rate": 1.0906123306885185e-05, "loss": 0.7065, "step": 13446 }, { "epoch": 0.48703368344802606, "grad_norm": 1.3408880682021544, "learning_rate": 1.0904955058653657e-05, "loss": 0.7124, "step": 13447 }, { "epoch": 0.48706990220934443, "grad_norm": 1.0120981521193095, "learning_rate": 1.0903786797969152e-05, "loss": 0.7208, "step": 13448 }, { "epoch": 0.4871061209706628, "grad_norm": 1.3090072626135265, "learning_rate": 1.0902618524847746e-05, "loss": 0.7431, "step": 13449 }, { "epoch": 0.4871423397319812, "grad_norm": 1.3999165174008226, "learning_rate": 1.0901450239305516e-05, "loss": 0.7284, "step": 13450 }, { "epoch": 0.48717855849329955, "grad_norm": 1.2967133708566032, "learning_rate": 1.0900281941358537e-05, "loss": 0.7472, "step": 13451 }, { "epoch": 0.48721477725461787, "grad_norm": 1.339770992822631, "learning_rate": 1.0899113631022889e-05, "loss": 0.7191, "step": 13452 }, { "epoch": 0.48725099601593624, "grad_norm": 1.3328763771234704, "learning_rate": 1.0897945308314647e-05, "loss": 0.7631, "step": 13453 }, { "epoch": 0.4872872147772546, "grad_norm": 1.2837226304515152, "learning_rate": 1.0896776973249889e-05, "loss": 0.6902, "step": 13454 }, { "epoch": 0.487323433538573, "grad_norm": 1.427186215580537, "learning_rate": 1.0895608625844694e-05, "loss": 0.7296, "step": 13455 }, { "epoch": 0.48735965229989137, "grad_norm": 1.6312873804077297, "learning_rate": 1.0894440266115134e-05, "loss": 0.8152, "step": 13456 }, { "epoch": 0.4873958710612097, "grad_norm": 1.2721952679206487, "learning_rate": 1.0893271894077291e-05, "loss": 0.7376, "step": 13457 }, { "epoch": 0.48743208982252806, "grad_norm": 1.2594560523727052, "learning_rate": 1.089210350974724e-05, "loss": 0.8183, "step": 13458 }, { "epoch": 0.48746830858384643, "grad_norm": 1.19082981532533, "learning_rate": 1.0890935113141066e-05, "loss": 0.731, "step": 13459 }, { "epoch": 0.4875045273451648, "grad_norm": 1.3196004465456155, "learning_rate": 1.0889766704274838e-05, "loss": 0.7451, "step": 13460 }, { "epoch": 0.4875407461064832, "grad_norm": 1.0293473695554047, "learning_rate": 1.0888598283164638e-05, "loss": 0.7828, "step": 13461 }, { "epoch": 0.4875769648678015, "grad_norm": 1.3686525710259152, "learning_rate": 1.0887429849826548e-05, "loss": 0.8029, "step": 13462 }, { "epoch": 0.4876131836291199, "grad_norm": 1.3225810570644643, "learning_rate": 1.088626140427664e-05, "loss": 0.6947, "step": 13463 }, { "epoch": 0.48764940239043825, "grad_norm": 1.3384756319818474, "learning_rate": 1.0885092946530997e-05, "loss": 0.7026, "step": 13464 }, { "epoch": 0.4876856211517566, "grad_norm": 1.3575414234457301, "learning_rate": 1.0883924476605698e-05, "loss": 0.6793, "step": 13465 }, { "epoch": 0.487721839913075, "grad_norm": 1.323054995792043, "learning_rate": 1.0882755994516822e-05, "loss": 0.722, "step": 13466 }, { "epoch": 0.4877580586743933, "grad_norm": 1.32617550008107, "learning_rate": 1.0881587500280448e-05, "loss": 0.7694, "step": 13467 }, { "epoch": 0.4877942774357117, "grad_norm": 1.424979143005749, "learning_rate": 1.0880418993912656e-05, "loss": 0.7385, "step": 13468 }, { "epoch": 0.48783049619703006, "grad_norm": 1.0154061700092796, "learning_rate": 1.0879250475429523e-05, "loss": 0.6385, "step": 13469 }, { "epoch": 0.48786671495834844, "grad_norm": 1.3616485068193627, "learning_rate": 1.0878081944847133e-05, "loss": 0.7162, "step": 13470 }, { "epoch": 0.4879029337196668, "grad_norm": 1.361609269122065, "learning_rate": 1.0876913402181562e-05, "loss": 0.7321, "step": 13471 }, { "epoch": 0.4879391524809851, "grad_norm": 1.212145955695199, "learning_rate": 1.0875744847448893e-05, "loss": 0.7212, "step": 13472 }, { "epoch": 0.4879753712423035, "grad_norm": 1.349104365860529, "learning_rate": 1.0874576280665206e-05, "loss": 0.6942, "step": 13473 }, { "epoch": 0.4880115900036219, "grad_norm": 1.3760068063004502, "learning_rate": 1.087340770184658e-05, "loss": 0.7949, "step": 13474 }, { "epoch": 0.48804780876494025, "grad_norm": 1.285635865932402, "learning_rate": 1.0872239111009095e-05, "loss": 0.7699, "step": 13475 }, { "epoch": 0.4880840275262586, "grad_norm": 1.2700791423789075, "learning_rate": 1.0871070508168836e-05, "loss": 0.6856, "step": 13476 }, { "epoch": 0.48812024628757694, "grad_norm": 1.3388869314038718, "learning_rate": 1.086990189334188e-05, "loss": 0.7539, "step": 13477 }, { "epoch": 0.4881564650488953, "grad_norm": 1.3249758128891587, "learning_rate": 1.086873326654431e-05, "loss": 0.7217, "step": 13478 }, { "epoch": 0.4881926838102137, "grad_norm": 1.3909560382810067, "learning_rate": 1.0867564627792208e-05, "loss": 0.7225, "step": 13479 }, { "epoch": 0.48822890257153206, "grad_norm": 1.3815447200527566, "learning_rate": 1.0866395977101654e-05, "loss": 0.7487, "step": 13480 }, { "epoch": 0.48826512133285044, "grad_norm": 1.2952223410479635, "learning_rate": 1.0865227314488728e-05, "loss": 0.7527, "step": 13481 }, { "epoch": 0.48830134009416876, "grad_norm": 1.385690563119372, "learning_rate": 1.0864058639969514e-05, "loss": 0.7407, "step": 13482 }, { "epoch": 0.48833755885548713, "grad_norm": 1.209383625290775, "learning_rate": 1.0862889953560096e-05, "loss": 0.7539, "step": 13483 }, { "epoch": 0.4883737776168055, "grad_norm": 1.3688602324924737, "learning_rate": 1.0861721255276554e-05, "loss": 0.7086, "step": 13484 }, { "epoch": 0.4884099963781239, "grad_norm": 1.1460516361617512, "learning_rate": 1.086055254513497e-05, "loss": 0.7443, "step": 13485 }, { "epoch": 0.48844621513944225, "grad_norm": 1.127758395566742, "learning_rate": 1.0859383823151426e-05, "loss": 0.7125, "step": 13486 }, { "epoch": 0.48848243390076057, "grad_norm": 1.448292737144213, "learning_rate": 1.0858215089342006e-05, "loss": 0.7899, "step": 13487 }, { "epoch": 0.48851865266207894, "grad_norm": 1.3878581188463308, "learning_rate": 1.0857046343722792e-05, "loss": 0.7064, "step": 13488 }, { "epoch": 0.4885548714233973, "grad_norm": 1.3553813540621884, "learning_rate": 1.0855877586309868e-05, "loss": 0.8006, "step": 13489 }, { "epoch": 0.4885910901847157, "grad_norm": 1.356943350774354, "learning_rate": 1.0854708817119318e-05, "loss": 0.7275, "step": 13490 }, { "epoch": 0.48862730894603407, "grad_norm": 1.2487566821902334, "learning_rate": 1.085354003616722e-05, "loss": 0.6776, "step": 13491 }, { "epoch": 0.4886635277073524, "grad_norm": 1.0541222867484328, "learning_rate": 1.0852371243469662e-05, "loss": 0.7466, "step": 13492 }, { "epoch": 0.48869974646867076, "grad_norm": 1.351210386378638, "learning_rate": 1.0851202439042727e-05, "loss": 0.7091, "step": 13493 }, { "epoch": 0.48873596522998913, "grad_norm": 1.493918735259268, "learning_rate": 1.0850033622902498e-05, "loss": 0.7707, "step": 13494 }, { "epoch": 0.4887721839913075, "grad_norm": 1.3542358613828513, "learning_rate": 1.0848864795065062e-05, "loss": 0.7543, "step": 13495 }, { "epoch": 0.4888084027526259, "grad_norm": 2.529398448480514, "learning_rate": 1.0847695955546498e-05, "loss": 0.7354, "step": 13496 }, { "epoch": 0.4888446215139442, "grad_norm": 1.5706889901596033, "learning_rate": 1.0846527104362896e-05, "loss": 0.8427, "step": 13497 }, { "epoch": 0.4888808402752626, "grad_norm": 1.4072985971506276, "learning_rate": 1.0845358241530333e-05, "loss": 0.7247, "step": 13498 }, { "epoch": 0.48891705903658095, "grad_norm": 1.3787622646770221, "learning_rate": 1.0844189367064901e-05, "loss": 0.741, "step": 13499 }, { "epoch": 0.4889532777978993, "grad_norm": 1.514639398780283, "learning_rate": 1.084302048098268e-05, "loss": 0.775, "step": 13500 }, { "epoch": 0.4889894965592177, "grad_norm": 1.3259994355399383, "learning_rate": 1.0841851583299755e-05, "loss": 0.6493, "step": 13501 }, { "epoch": 0.489025715320536, "grad_norm": 1.4377659080076681, "learning_rate": 1.0840682674032214e-05, "loss": 0.7178, "step": 13502 }, { "epoch": 0.4890619340818544, "grad_norm": 1.4756912279717291, "learning_rate": 1.0839513753196141e-05, "loss": 0.7599, "step": 13503 }, { "epoch": 0.48909815284317276, "grad_norm": 1.3603131952235792, "learning_rate": 1.0838344820807623e-05, "loss": 0.698, "step": 13504 }, { "epoch": 0.48913437160449114, "grad_norm": 1.2846327377760283, "learning_rate": 1.0837175876882739e-05, "loss": 0.7745, "step": 13505 }, { "epoch": 0.4891705903658095, "grad_norm": 1.5595339924327942, "learning_rate": 1.0836006921437586e-05, "loss": 0.8172, "step": 13506 }, { "epoch": 0.48920680912712783, "grad_norm": 1.4247120466960486, "learning_rate": 1.0834837954488236e-05, "loss": 0.8082, "step": 13507 }, { "epoch": 0.4892430278884462, "grad_norm": 1.2844135231078886, "learning_rate": 1.0833668976050787e-05, "loss": 0.7081, "step": 13508 }, { "epoch": 0.4892792466497646, "grad_norm": 1.2995119166372264, "learning_rate": 1.0832499986141318e-05, "loss": 0.7518, "step": 13509 }, { "epoch": 0.48931546541108295, "grad_norm": 1.3309093292418541, "learning_rate": 1.0831330984775918e-05, "loss": 0.708, "step": 13510 }, { "epoch": 0.4893516841724013, "grad_norm": 1.4393221848662394, "learning_rate": 1.0830161971970675e-05, "loss": 0.7914, "step": 13511 }, { "epoch": 0.48938790293371964, "grad_norm": 1.0359627848110828, "learning_rate": 1.0828992947741672e-05, "loss": 0.7814, "step": 13512 }, { "epoch": 0.489424121695038, "grad_norm": 1.0677506625741948, "learning_rate": 1.0827823912104998e-05, "loss": 0.761, "step": 13513 }, { "epoch": 0.4894603404563564, "grad_norm": 1.4528273090384012, "learning_rate": 1.0826654865076738e-05, "loss": 0.7313, "step": 13514 }, { "epoch": 0.48949655921767476, "grad_norm": 1.5219788292669993, "learning_rate": 1.0825485806672983e-05, "loss": 0.8035, "step": 13515 }, { "epoch": 0.48953277797899314, "grad_norm": 1.2335680016722894, "learning_rate": 1.0824316736909817e-05, "loss": 0.6602, "step": 13516 }, { "epoch": 0.48956899674031146, "grad_norm": 1.4530902328917776, "learning_rate": 1.082314765580333e-05, "loss": 0.7735, "step": 13517 }, { "epoch": 0.48960521550162983, "grad_norm": 1.2411908283900706, "learning_rate": 1.0821978563369605e-05, "loss": 0.689, "step": 13518 }, { "epoch": 0.4896414342629482, "grad_norm": 1.4531078196607194, "learning_rate": 1.0820809459624735e-05, "loss": 0.769, "step": 13519 }, { "epoch": 0.4896776530242666, "grad_norm": 1.444611887334205, "learning_rate": 1.0819640344584805e-05, "loss": 0.7584, "step": 13520 }, { "epoch": 0.48971387178558495, "grad_norm": 1.3243130731888806, "learning_rate": 1.0818471218265905e-05, "loss": 0.7509, "step": 13521 }, { "epoch": 0.48975009054690327, "grad_norm": 1.3227292421945134, "learning_rate": 1.0817302080684118e-05, "loss": 0.7856, "step": 13522 }, { "epoch": 0.48978630930822165, "grad_norm": 1.5338253872367578, "learning_rate": 1.0816132931855539e-05, "loss": 0.823, "step": 13523 }, { "epoch": 0.48982252806954, "grad_norm": 1.1469506747903027, "learning_rate": 1.0814963771796255e-05, "loss": 0.7612, "step": 13524 }, { "epoch": 0.4898587468308584, "grad_norm": 1.3338833284715486, "learning_rate": 1.081379460052235e-05, "loss": 0.7756, "step": 13525 }, { "epoch": 0.48989496559217677, "grad_norm": 1.6102423182643173, "learning_rate": 1.0812625418049918e-05, "loss": 0.7879, "step": 13526 }, { "epoch": 0.4899311843534951, "grad_norm": 1.08238805128133, "learning_rate": 1.0811456224395048e-05, "loss": 0.7203, "step": 13527 }, { "epoch": 0.48996740311481346, "grad_norm": 1.471022038122403, "learning_rate": 1.0810287019573823e-05, "loss": 0.7106, "step": 13528 }, { "epoch": 0.49000362187613183, "grad_norm": 1.2076765568400833, "learning_rate": 1.080911780360234e-05, "loss": 0.7756, "step": 13529 }, { "epoch": 0.4900398406374502, "grad_norm": 1.27677463399617, "learning_rate": 1.0807948576496685e-05, "loss": 0.7468, "step": 13530 }, { "epoch": 0.4900760593987686, "grad_norm": 1.029194406913862, "learning_rate": 1.0806779338272945e-05, "loss": 0.7332, "step": 13531 }, { "epoch": 0.4901122781600869, "grad_norm": 1.5542515833665447, "learning_rate": 1.0805610088947219e-05, "loss": 0.7757, "step": 13532 }, { "epoch": 0.4901484969214053, "grad_norm": 1.3249358413003367, "learning_rate": 1.0804440828535583e-05, "loss": 0.6584, "step": 13533 }, { "epoch": 0.49018471568272365, "grad_norm": 1.3683631549263553, "learning_rate": 1.0803271557054136e-05, "loss": 0.6902, "step": 13534 }, { "epoch": 0.490220934444042, "grad_norm": 1.348561318178486, "learning_rate": 1.080210227451897e-05, "loss": 0.7281, "step": 13535 }, { "epoch": 0.4902571532053604, "grad_norm": 1.2926455955384446, "learning_rate": 1.080093298094617e-05, "loss": 0.8164, "step": 13536 }, { "epoch": 0.4902933719666787, "grad_norm": 1.9874225703337607, "learning_rate": 1.0799763676351829e-05, "loss": 0.7651, "step": 13537 }, { "epoch": 0.4903295907279971, "grad_norm": 1.032169326610339, "learning_rate": 1.0798594360752034e-05, "loss": 0.7827, "step": 13538 }, { "epoch": 0.49036580948931546, "grad_norm": 1.2718258309877368, "learning_rate": 1.0797425034162881e-05, "loss": 0.6391, "step": 13539 }, { "epoch": 0.49040202825063384, "grad_norm": 1.3951497429418167, "learning_rate": 1.0796255696600459e-05, "loss": 0.736, "step": 13540 }, { "epoch": 0.4904382470119522, "grad_norm": 1.3547974987041134, "learning_rate": 1.0795086348080859e-05, "loss": 0.7197, "step": 13541 }, { "epoch": 0.49047446577327053, "grad_norm": 1.2305662328464013, "learning_rate": 1.0793916988620172e-05, "loss": 0.6803, "step": 13542 }, { "epoch": 0.4905106845345889, "grad_norm": 1.400585004119711, "learning_rate": 1.0792747618234489e-05, "loss": 0.7066, "step": 13543 }, { "epoch": 0.4905469032959073, "grad_norm": 1.215484340142548, "learning_rate": 1.0791578236939903e-05, "loss": 0.7413, "step": 13544 }, { "epoch": 0.49058312205722565, "grad_norm": 1.3644666068248945, "learning_rate": 1.0790408844752505e-05, "loss": 0.7334, "step": 13545 }, { "epoch": 0.490619340818544, "grad_norm": 1.4989220320855834, "learning_rate": 1.0789239441688388e-05, "loss": 0.8188, "step": 13546 }, { "epoch": 0.49065555957986234, "grad_norm": 1.0366980309174954, "learning_rate": 1.078807002776364e-05, "loss": 0.7661, "step": 13547 }, { "epoch": 0.4906917783411807, "grad_norm": 1.3640815918690772, "learning_rate": 1.078690060299436e-05, "loss": 0.7504, "step": 13548 }, { "epoch": 0.4907279971024991, "grad_norm": 1.1992926737898313, "learning_rate": 1.0785731167396632e-05, "loss": 0.6814, "step": 13549 }, { "epoch": 0.49076421586381747, "grad_norm": 1.2711180981213182, "learning_rate": 1.0784561720986552e-05, "loss": 0.7098, "step": 13550 }, { "epoch": 0.49080043462513584, "grad_norm": 1.055783306408699, "learning_rate": 1.0783392263780218e-05, "loss": 0.7943, "step": 13551 }, { "epoch": 0.49083665338645416, "grad_norm": 1.0995536719369468, "learning_rate": 1.0782222795793714e-05, "loss": 0.7551, "step": 13552 }, { "epoch": 0.49087287214777253, "grad_norm": 1.1414901019816832, "learning_rate": 1.078105331704314e-05, "loss": 0.77, "step": 13553 }, { "epoch": 0.4909090909090909, "grad_norm": 2.058315097183328, "learning_rate": 1.0779883827544584e-05, "loss": 0.685, "step": 13554 }, { "epoch": 0.4909453096704093, "grad_norm": 1.14189090551608, "learning_rate": 1.0778714327314143e-05, "loss": 0.7452, "step": 13555 }, { "epoch": 0.49098152843172765, "grad_norm": 1.3304181882565782, "learning_rate": 1.0777544816367907e-05, "loss": 0.7304, "step": 13556 }, { "epoch": 0.491017747193046, "grad_norm": 1.2810196974835526, "learning_rate": 1.0776375294721972e-05, "loss": 0.7152, "step": 13557 }, { "epoch": 0.49105396595436435, "grad_norm": 1.3397375515147523, "learning_rate": 1.077520576239243e-05, "loss": 0.6074, "step": 13558 }, { "epoch": 0.4910901847156827, "grad_norm": 1.3305548809147305, "learning_rate": 1.0774036219395376e-05, "loss": 0.8006, "step": 13559 }, { "epoch": 0.4911264034770011, "grad_norm": 1.4015916491723213, "learning_rate": 1.0772866665746904e-05, "loss": 0.815, "step": 13560 }, { "epoch": 0.49116262223831947, "grad_norm": 1.3268233126270677, "learning_rate": 1.0771697101463106e-05, "loss": 0.6878, "step": 13561 }, { "epoch": 0.4911988409996378, "grad_norm": 1.4396029644104684, "learning_rate": 1.077052752656008e-05, "loss": 0.7468, "step": 13562 }, { "epoch": 0.49123505976095616, "grad_norm": 1.046381803455192, "learning_rate": 1.0769357941053916e-05, "loss": 0.735, "step": 13563 }, { "epoch": 0.49127127852227453, "grad_norm": 1.5657294873340974, "learning_rate": 1.0768188344960712e-05, "loss": 0.7135, "step": 13564 }, { "epoch": 0.4913074972835929, "grad_norm": 1.5367437167677633, "learning_rate": 1.0767018738296561e-05, "loss": 0.8152, "step": 13565 }, { "epoch": 0.4913437160449113, "grad_norm": 1.0438833591328, "learning_rate": 1.0765849121077561e-05, "loss": 0.7518, "step": 13566 }, { "epoch": 0.4913799348062296, "grad_norm": 1.1793640965253112, "learning_rate": 1.0764679493319801e-05, "loss": 0.7872, "step": 13567 }, { "epoch": 0.491416153567548, "grad_norm": 1.3208846095589717, "learning_rate": 1.076350985503938e-05, "loss": 0.768, "step": 13568 }, { "epoch": 0.49145237232886635, "grad_norm": 1.5796724230068497, "learning_rate": 1.076234020625239e-05, "loss": 0.7716, "step": 13569 }, { "epoch": 0.4914885910901847, "grad_norm": 1.1006169934352763, "learning_rate": 1.076117054697493e-05, "loss": 0.722, "step": 13570 }, { "epoch": 0.4915248098515031, "grad_norm": 1.3421454400276924, "learning_rate": 1.0760000877223097e-05, "loss": 0.7722, "step": 13571 }, { "epoch": 0.4915610286128214, "grad_norm": 1.434500625714757, "learning_rate": 1.0758831197012985e-05, "loss": 0.8074, "step": 13572 }, { "epoch": 0.4915972473741398, "grad_norm": 1.2693214543929894, "learning_rate": 1.0757661506360686e-05, "loss": 0.696, "step": 13573 }, { "epoch": 0.49163346613545816, "grad_norm": 1.252985622260534, "learning_rate": 1.0756491805282298e-05, "loss": 0.6907, "step": 13574 }, { "epoch": 0.49166968489677654, "grad_norm": 1.0605936313433468, "learning_rate": 1.0755322093793924e-05, "loss": 0.7144, "step": 13575 }, { "epoch": 0.4917059036580949, "grad_norm": 1.465127041444104, "learning_rate": 1.0754152371911648e-05, "loss": 0.7466, "step": 13576 }, { "epoch": 0.49174212241941323, "grad_norm": 1.350980323200581, "learning_rate": 1.0752982639651575e-05, "loss": 0.7365, "step": 13577 }, { "epoch": 0.4917783411807316, "grad_norm": 1.3963599829859141, "learning_rate": 1.07518128970298e-05, "loss": 0.7062, "step": 13578 }, { "epoch": 0.49181455994205, "grad_norm": 1.2535857833909663, "learning_rate": 1.0750643144062418e-05, "loss": 0.7141, "step": 13579 }, { "epoch": 0.49185077870336835, "grad_norm": 1.3227631451356974, "learning_rate": 1.0749473380765525e-05, "loss": 0.7658, "step": 13580 }, { "epoch": 0.4918869974646867, "grad_norm": 1.2888641154192948, "learning_rate": 1.0748303607155222e-05, "loss": 0.7121, "step": 13581 }, { "epoch": 0.49192321622600504, "grad_norm": 1.2980522326411374, "learning_rate": 1.0747133823247604e-05, "loss": 0.6457, "step": 13582 }, { "epoch": 0.4919594349873234, "grad_norm": 1.2068490481413057, "learning_rate": 1.0745964029058766e-05, "loss": 0.7606, "step": 13583 }, { "epoch": 0.4919956537486418, "grad_norm": 1.3767283729666742, "learning_rate": 1.074479422460481e-05, "loss": 0.7523, "step": 13584 }, { "epoch": 0.49203187250996017, "grad_norm": 1.45246254269434, "learning_rate": 1.0743624409901829e-05, "loss": 0.745, "step": 13585 }, { "epoch": 0.49206809127127854, "grad_norm": 1.3300850947865446, "learning_rate": 1.0742454584965924e-05, "loss": 0.7882, "step": 13586 }, { "epoch": 0.49210431003259686, "grad_norm": 1.2877662246008703, "learning_rate": 1.0741284749813192e-05, "loss": 0.7409, "step": 13587 }, { "epoch": 0.49214052879391523, "grad_norm": 1.0691963749219138, "learning_rate": 1.074011490445973e-05, "loss": 0.7497, "step": 13588 }, { "epoch": 0.4921767475552336, "grad_norm": 1.4188577594774139, "learning_rate": 1.0738945048921636e-05, "loss": 0.6972, "step": 13589 }, { "epoch": 0.492212966316552, "grad_norm": 1.0109649918263925, "learning_rate": 1.0737775183215011e-05, "loss": 0.7448, "step": 13590 }, { "epoch": 0.49224918507787035, "grad_norm": 1.309855412307637, "learning_rate": 1.0736605307355948e-05, "loss": 0.7374, "step": 13591 }, { "epoch": 0.4922854038391887, "grad_norm": 1.3244350483160912, "learning_rate": 1.0735435421360549e-05, "loss": 0.7152, "step": 13592 }, { "epoch": 0.49232162260050705, "grad_norm": 1.3445960496108416, "learning_rate": 1.0734265525244916e-05, "loss": 0.7797, "step": 13593 }, { "epoch": 0.4923578413618254, "grad_norm": 1.445679241767521, "learning_rate": 1.0733095619025141e-05, "loss": 0.8091, "step": 13594 }, { "epoch": 0.4923940601231438, "grad_norm": 1.216668161194381, "learning_rate": 1.0731925702717327e-05, "loss": 0.6737, "step": 13595 }, { "epoch": 0.49243027888446217, "grad_norm": 1.4273641187094577, "learning_rate": 1.0730755776337571e-05, "loss": 0.7163, "step": 13596 }, { "epoch": 0.4924664976457805, "grad_norm": 1.3101634119778653, "learning_rate": 1.0729585839901975e-05, "loss": 0.7099, "step": 13597 }, { "epoch": 0.49250271640709886, "grad_norm": 1.3932200252524214, "learning_rate": 1.0728415893426636e-05, "loss": 0.8561, "step": 13598 }, { "epoch": 0.49253893516841724, "grad_norm": 1.2951011087762523, "learning_rate": 1.0727245936927655e-05, "loss": 0.7566, "step": 13599 }, { "epoch": 0.4925751539297356, "grad_norm": 1.314091898496311, "learning_rate": 1.072607597042113e-05, "loss": 0.6887, "step": 13600 }, { "epoch": 0.492611372691054, "grad_norm": 1.3485591538221073, "learning_rate": 1.072490599392316e-05, "loss": 0.711, "step": 13601 }, { "epoch": 0.4926475914523723, "grad_norm": 1.3788404624633834, "learning_rate": 1.072373600744985e-05, "loss": 0.7804, "step": 13602 }, { "epoch": 0.4926838102136907, "grad_norm": 1.08277886959536, "learning_rate": 1.0722566011017295e-05, "loss": 0.7532, "step": 13603 }, { "epoch": 0.49272002897500905, "grad_norm": 1.0405078379401758, "learning_rate": 1.0721396004641598e-05, "loss": 0.8153, "step": 13604 }, { "epoch": 0.4927562477363274, "grad_norm": 1.5193540916039914, "learning_rate": 1.0720225988338857e-05, "loss": 0.7441, "step": 13605 }, { "epoch": 0.4927924664976458, "grad_norm": 1.3878311725608807, "learning_rate": 1.0719055962125173e-05, "loss": 0.7811, "step": 13606 }, { "epoch": 0.4928286852589641, "grad_norm": 1.4978364475873918, "learning_rate": 1.0717885926016647e-05, "loss": 0.729, "step": 13607 }, { "epoch": 0.4928649040202825, "grad_norm": 1.2941701233101592, "learning_rate": 1.0716715880029383e-05, "loss": 0.7659, "step": 13608 }, { "epoch": 0.49290112278160086, "grad_norm": 1.3122416821090994, "learning_rate": 1.0715545824179474e-05, "loss": 0.775, "step": 13609 }, { "epoch": 0.49293734154291924, "grad_norm": 1.3169551944303415, "learning_rate": 1.071437575848303e-05, "loss": 0.7587, "step": 13610 }, { "epoch": 0.4929735603042376, "grad_norm": 1.094066969250649, "learning_rate": 1.0713205682956145e-05, "loss": 0.7643, "step": 13611 }, { "epoch": 0.49300977906555593, "grad_norm": 1.4612450550057268, "learning_rate": 1.0712035597614923e-05, "loss": 0.8484, "step": 13612 }, { "epoch": 0.4930459978268743, "grad_norm": 3.0406327199494867, "learning_rate": 1.0710865502475469e-05, "loss": 0.7542, "step": 13613 }, { "epoch": 0.4930822165881927, "grad_norm": 1.4079726899658862, "learning_rate": 1.0709695397553877e-05, "loss": 0.7462, "step": 13614 }, { "epoch": 0.49311843534951105, "grad_norm": 1.329208362769093, "learning_rate": 1.0708525282866256e-05, "loss": 0.7391, "step": 13615 }, { "epoch": 0.4931546541108294, "grad_norm": 1.397810375128644, "learning_rate": 1.0707355158428702e-05, "loss": 0.8052, "step": 13616 }, { "epoch": 0.49319087287214775, "grad_norm": 1.349312263085395, "learning_rate": 1.0706185024257321e-05, "loss": 0.7675, "step": 13617 }, { "epoch": 0.4932270916334661, "grad_norm": 1.3949969045842259, "learning_rate": 1.0705014880368212e-05, "loss": 0.7028, "step": 13618 }, { "epoch": 0.4932633103947845, "grad_norm": 1.9206170640602103, "learning_rate": 1.0703844726777482e-05, "loss": 0.7547, "step": 13619 }, { "epoch": 0.49329952915610287, "grad_norm": 1.1506119679188294, "learning_rate": 1.0702674563501227e-05, "loss": 0.6894, "step": 13620 }, { "epoch": 0.49333574791742124, "grad_norm": 1.475561545704996, "learning_rate": 1.0701504390555555e-05, "loss": 0.8249, "step": 13621 }, { "epoch": 0.49337196667873956, "grad_norm": 0.9776327587739473, "learning_rate": 1.0700334207956563e-05, "loss": 0.733, "step": 13622 }, { "epoch": 0.49340818544005793, "grad_norm": 1.3288340288773766, "learning_rate": 1.0699164015720358e-05, "loss": 0.7202, "step": 13623 }, { "epoch": 0.4934444042013763, "grad_norm": 1.330937957682982, "learning_rate": 1.0697993813863043e-05, "loss": 0.7734, "step": 13624 }, { "epoch": 0.4934806229626947, "grad_norm": 1.0499544364004916, "learning_rate": 1.0696823602400718e-05, "loss": 0.716, "step": 13625 }, { "epoch": 0.49351684172401306, "grad_norm": 1.4034136048951806, "learning_rate": 1.0695653381349491e-05, "loss": 0.7013, "step": 13626 }, { "epoch": 0.4935530604853314, "grad_norm": 1.4801068064834584, "learning_rate": 1.0694483150725458e-05, "loss": 0.7513, "step": 13627 }, { "epoch": 0.49358927924664975, "grad_norm": 1.327852841849846, "learning_rate": 1.069331291054473e-05, "loss": 0.7541, "step": 13628 }, { "epoch": 0.4936254980079681, "grad_norm": 1.0681367342283716, "learning_rate": 1.0692142660823406e-05, "loss": 0.6592, "step": 13629 }, { "epoch": 0.4936617167692865, "grad_norm": 1.3527436456299287, "learning_rate": 1.0690972401577593e-05, "loss": 0.7595, "step": 13630 }, { "epoch": 0.49369793553060487, "grad_norm": 2.049007082488352, "learning_rate": 1.068980213282339e-05, "loss": 0.7548, "step": 13631 }, { "epoch": 0.4937341542919232, "grad_norm": 1.1134495262911233, "learning_rate": 1.0688631854576903e-05, "loss": 0.775, "step": 13632 }, { "epoch": 0.49377037305324156, "grad_norm": 1.7112750902645841, "learning_rate": 1.0687461566854237e-05, "loss": 0.7546, "step": 13633 }, { "epoch": 0.49380659181455994, "grad_norm": 1.5555616238494376, "learning_rate": 1.0686291269671496e-05, "loss": 0.8052, "step": 13634 }, { "epoch": 0.4938428105758783, "grad_norm": 1.3289976815150728, "learning_rate": 1.0685120963044785e-05, "loss": 0.7476, "step": 13635 }, { "epoch": 0.4938790293371967, "grad_norm": 1.2717519967776072, "learning_rate": 1.0683950646990209e-05, "loss": 0.7622, "step": 13636 }, { "epoch": 0.49391524809851506, "grad_norm": 1.1234290650240755, "learning_rate": 1.0682780321523868e-05, "loss": 0.7559, "step": 13637 }, { "epoch": 0.4939514668598334, "grad_norm": 1.3267851143437994, "learning_rate": 1.0681609986661872e-05, "loss": 0.7333, "step": 13638 }, { "epoch": 0.49398768562115175, "grad_norm": 1.0588609803632956, "learning_rate": 1.0680439642420323e-05, "loss": 0.7167, "step": 13639 }, { "epoch": 0.4940239043824701, "grad_norm": 1.000046377100776, "learning_rate": 1.0679269288815328e-05, "loss": 0.6789, "step": 13640 }, { "epoch": 0.4940601231437885, "grad_norm": 1.558167712463558, "learning_rate": 1.0678098925862988e-05, "loss": 0.7324, "step": 13641 }, { "epoch": 0.4940963419051069, "grad_norm": 1.4495390232235916, "learning_rate": 1.0676928553579414e-05, "loss": 0.7902, "step": 13642 }, { "epoch": 0.4941325606664252, "grad_norm": 1.351939857040854, "learning_rate": 1.0675758171980706e-05, "loss": 0.7424, "step": 13643 }, { "epoch": 0.49416877942774357, "grad_norm": 1.433701036944606, "learning_rate": 1.0674587781082973e-05, "loss": 0.728, "step": 13644 }, { "epoch": 0.49420499818906194, "grad_norm": 0.9636327710360518, "learning_rate": 1.0673417380902317e-05, "loss": 0.734, "step": 13645 }, { "epoch": 0.4942412169503803, "grad_norm": 1.4344398745657223, "learning_rate": 1.0672246971454849e-05, "loss": 0.7397, "step": 13646 }, { "epoch": 0.4942774357116987, "grad_norm": 1.234099555821264, "learning_rate": 1.0671076552756669e-05, "loss": 0.6964, "step": 13647 }, { "epoch": 0.494313654473017, "grad_norm": 1.0721503795231526, "learning_rate": 1.066990612482389e-05, "loss": 0.6412, "step": 13648 }, { "epoch": 0.4943498732343354, "grad_norm": 1.3462637050669561, "learning_rate": 1.0668735687672611e-05, "loss": 0.7907, "step": 13649 }, { "epoch": 0.49438609199565375, "grad_norm": 1.3215429910677519, "learning_rate": 1.066756524131894e-05, "loss": 0.7404, "step": 13650 }, { "epoch": 0.4944223107569721, "grad_norm": 1.0596666796541692, "learning_rate": 1.066639478577899e-05, "loss": 0.682, "step": 13651 }, { "epoch": 0.4944585295182905, "grad_norm": 1.186115223713187, "learning_rate": 1.0665224321068857e-05, "loss": 0.6846, "step": 13652 }, { "epoch": 0.4944947482796088, "grad_norm": 1.3235075223598258, "learning_rate": 1.0664053847204656e-05, "loss": 0.6992, "step": 13653 }, { "epoch": 0.4945309670409272, "grad_norm": 1.3768486659691395, "learning_rate": 1.0662883364202488e-05, "loss": 0.6644, "step": 13654 }, { "epoch": 0.49456718580224557, "grad_norm": 1.2700732446008796, "learning_rate": 1.0661712872078464e-05, "loss": 0.6966, "step": 13655 }, { "epoch": 0.49460340456356394, "grad_norm": 1.5746778997047994, "learning_rate": 1.066054237084869e-05, "loss": 0.8205, "step": 13656 }, { "epoch": 0.4946396233248823, "grad_norm": 1.2049970081732932, "learning_rate": 1.0659371860529273e-05, "loss": 0.7365, "step": 13657 }, { "epoch": 0.49467584208620063, "grad_norm": 1.4026412263163177, "learning_rate": 1.0658201341136316e-05, "loss": 0.7713, "step": 13658 }, { "epoch": 0.494712060847519, "grad_norm": 1.2800244817474038, "learning_rate": 1.0657030812685934e-05, "loss": 0.7534, "step": 13659 }, { "epoch": 0.4947482796088374, "grad_norm": 1.2812556646093893, "learning_rate": 1.065586027519423e-05, "loss": 0.7264, "step": 13660 }, { "epoch": 0.49478449837015576, "grad_norm": 1.2535983194764544, "learning_rate": 1.065468972867731e-05, "loss": 0.7628, "step": 13661 }, { "epoch": 0.49482071713147413, "grad_norm": 1.0177730389185953, "learning_rate": 1.0653519173151287e-05, "loss": 0.7403, "step": 13662 }, { "epoch": 0.49485693589279245, "grad_norm": 1.9399680849435283, "learning_rate": 1.0652348608632264e-05, "loss": 0.7757, "step": 13663 }, { "epoch": 0.4948931546541108, "grad_norm": 1.3614503150241457, "learning_rate": 1.0651178035136353e-05, "loss": 0.7531, "step": 13664 }, { "epoch": 0.4949293734154292, "grad_norm": 1.1122329618037234, "learning_rate": 1.0650007452679658e-05, "loss": 0.6929, "step": 13665 }, { "epoch": 0.49496559217674757, "grad_norm": 1.4705714183259546, "learning_rate": 1.064883686127829e-05, "loss": 0.8237, "step": 13666 }, { "epoch": 0.49500181093806594, "grad_norm": 1.0343471136383087, "learning_rate": 1.0647666260948357e-05, "loss": 0.6765, "step": 13667 }, { "epoch": 0.49503802969938426, "grad_norm": 0.9870273951082587, "learning_rate": 1.0646495651705967e-05, "loss": 0.7307, "step": 13668 }, { "epoch": 0.49507424846070264, "grad_norm": 1.3256922979754064, "learning_rate": 1.064532503356723e-05, "loss": 0.7217, "step": 13669 }, { "epoch": 0.495110467222021, "grad_norm": 1.3275346551900273, "learning_rate": 1.064415440654825e-05, "loss": 0.6718, "step": 13670 }, { "epoch": 0.4951466859833394, "grad_norm": 1.0502510498039883, "learning_rate": 1.0642983770665144e-05, "loss": 0.7407, "step": 13671 }, { "epoch": 0.49518290474465776, "grad_norm": 1.424940858364477, "learning_rate": 1.0641813125934013e-05, "loss": 0.7439, "step": 13672 }, { "epoch": 0.4952191235059761, "grad_norm": 1.085365938723327, "learning_rate": 1.0640642472370972e-05, "loss": 0.7, "step": 13673 }, { "epoch": 0.49525534226729445, "grad_norm": 1.198620201058169, "learning_rate": 1.0639471809992126e-05, "loss": 0.7279, "step": 13674 }, { "epoch": 0.4952915610286128, "grad_norm": 1.1992098414994166, "learning_rate": 1.0638301138813588e-05, "loss": 0.7289, "step": 13675 }, { "epoch": 0.4953277797899312, "grad_norm": 1.421848847730351, "learning_rate": 1.0637130458851462e-05, "loss": 0.7924, "step": 13676 }, { "epoch": 0.4953639985512496, "grad_norm": 1.2597392892672237, "learning_rate": 1.0635959770121862e-05, "loss": 0.7043, "step": 13677 }, { "epoch": 0.4954002173125679, "grad_norm": 1.2698080311346074, "learning_rate": 1.0634789072640897e-05, "loss": 0.7375, "step": 13678 }, { "epoch": 0.49543643607388627, "grad_norm": 1.4279460104342132, "learning_rate": 1.0633618366424676e-05, "loss": 0.7839, "step": 13679 }, { "epoch": 0.49547265483520464, "grad_norm": 1.3503260621054343, "learning_rate": 1.0632447651489313e-05, "loss": 0.6555, "step": 13680 }, { "epoch": 0.495508873596523, "grad_norm": 1.2588962031388828, "learning_rate": 1.063127692785091e-05, "loss": 0.7249, "step": 13681 }, { "epoch": 0.4955450923578414, "grad_norm": 1.2347611619447207, "learning_rate": 1.0630106195525585e-05, "loss": 0.7212, "step": 13682 }, { "epoch": 0.4955813111191597, "grad_norm": 1.4062630877945417, "learning_rate": 1.0628935454529443e-05, "loss": 0.8605, "step": 13683 }, { "epoch": 0.4956175298804781, "grad_norm": 1.2848906822382247, "learning_rate": 1.0627764704878598e-05, "loss": 0.7056, "step": 13684 }, { "epoch": 0.49565374864179645, "grad_norm": 1.3641059338003196, "learning_rate": 1.062659394658916e-05, "loss": 0.6986, "step": 13685 }, { "epoch": 0.49568996740311483, "grad_norm": 1.3403648372450043, "learning_rate": 1.0625423179677236e-05, "loss": 0.7673, "step": 13686 }, { "epoch": 0.4957261861644332, "grad_norm": 1.292344628413191, "learning_rate": 1.0624252404158941e-05, "loss": 0.7703, "step": 13687 }, { "epoch": 0.4957624049257515, "grad_norm": 1.3865722834108898, "learning_rate": 1.0623081620050386e-05, "loss": 0.7315, "step": 13688 }, { "epoch": 0.4957986236870699, "grad_norm": 1.306770410920287, "learning_rate": 1.0621910827367676e-05, "loss": 0.7077, "step": 13689 }, { "epoch": 0.49583484244838827, "grad_norm": 1.2501529354177388, "learning_rate": 1.062074002612693e-05, "loss": 0.6975, "step": 13690 }, { "epoch": 0.49587106120970664, "grad_norm": 1.282099732069898, "learning_rate": 1.0619569216344257e-05, "loss": 0.7624, "step": 13691 }, { "epoch": 0.495907279971025, "grad_norm": 1.3796554831724515, "learning_rate": 1.0618398398035764e-05, "loss": 0.7244, "step": 13692 }, { "epoch": 0.49594349873234334, "grad_norm": 1.376461402534224, "learning_rate": 1.0617227571217567e-05, "loss": 0.7833, "step": 13693 }, { "epoch": 0.4959797174936617, "grad_norm": 1.3315600886151857, "learning_rate": 1.0616056735905775e-05, "loss": 0.7623, "step": 13694 }, { "epoch": 0.4960159362549801, "grad_norm": 1.759467004340398, "learning_rate": 1.0614885892116504e-05, "loss": 0.7526, "step": 13695 }, { "epoch": 0.49605215501629846, "grad_norm": 1.3748644862262442, "learning_rate": 1.0613715039865859e-05, "loss": 0.6925, "step": 13696 }, { "epoch": 0.49608837377761683, "grad_norm": 1.390423435969284, "learning_rate": 1.061254417916996e-05, "loss": 0.8609, "step": 13697 }, { "epoch": 0.49612459253893515, "grad_norm": 1.0922388633762725, "learning_rate": 1.061137331004491e-05, "loss": 0.7008, "step": 13698 }, { "epoch": 0.4961608113002535, "grad_norm": 1.0293461654472589, "learning_rate": 1.0610202432506831e-05, "loss": 0.7589, "step": 13699 }, { "epoch": 0.4961970300615719, "grad_norm": 1.4029706567132538, "learning_rate": 1.0609031546571827e-05, "loss": 0.731, "step": 13700 }, { "epoch": 0.49623324882289027, "grad_norm": 1.445883914395867, "learning_rate": 1.0607860652256015e-05, "loss": 0.7754, "step": 13701 }, { "epoch": 0.49626946758420865, "grad_norm": 1.444860301287537, "learning_rate": 1.0606689749575504e-05, "loss": 0.7502, "step": 13702 }, { "epoch": 0.49630568634552696, "grad_norm": 1.3387855091681686, "learning_rate": 1.0605518838546412e-05, "loss": 0.693, "step": 13703 }, { "epoch": 0.49634190510684534, "grad_norm": 1.2091110945741712, "learning_rate": 1.060434791918485e-05, "loss": 0.724, "step": 13704 }, { "epoch": 0.4963781238681637, "grad_norm": 1.5651879341084671, "learning_rate": 1.0603176991506925e-05, "loss": 0.6965, "step": 13705 }, { "epoch": 0.4964143426294821, "grad_norm": 1.370594863116712, "learning_rate": 1.060200605552876e-05, "loss": 0.6985, "step": 13706 }, { "epoch": 0.49645056139080046, "grad_norm": 1.2010484228815363, "learning_rate": 1.0600835111266457e-05, "loss": 0.6863, "step": 13707 }, { "epoch": 0.4964867801521188, "grad_norm": 1.3956886952431262, "learning_rate": 1.059966415873614e-05, "loss": 0.7862, "step": 13708 }, { "epoch": 0.49652299891343715, "grad_norm": 0.9825878355071971, "learning_rate": 1.0598493197953913e-05, "loss": 0.7009, "step": 13709 }, { "epoch": 0.4965592176747555, "grad_norm": 1.2796661317238318, "learning_rate": 1.0597322228935895e-05, "loss": 0.6931, "step": 13710 }, { "epoch": 0.4965954364360739, "grad_norm": 1.4056294900093946, "learning_rate": 1.05961512516982e-05, "loss": 0.7618, "step": 13711 }, { "epoch": 0.4966316551973923, "grad_norm": 1.3416433072950482, "learning_rate": 1.0594980266256938e-05, "loss": 0.7556, "step": 13712 }, { "epoch": 0.4966678739587106, "grad_norm": 1.3598767319293126, "learning_rate": 1.0593809272628226e-05, "loss": 0.6874, "step": 13713 }, { "epoch": 0.49670409272002897, "grad_norm": 2.0330231893384676, "learning_rate": 1.0592638270828175e-05, "loss": 0.7568, "step": 13714 }, { "epoch": 0.49674031148134734, "grad_norm": 1.099348979381216, "learning_rate": 1.0591467260872905e-05, "loss": 0.6848, "step": 13715 }, { "epoch": 0.4967765302426657, "grad_norm": 1.319519990863951, "learning_rate": 1.0590296242778521e-05, "loss": 0.8136, "step": 13716 }, { "epoch": 0.4968127490039841, "grad_norm": 1.2145861382432281, "learning_rate": 1.0589125216561144e-05, "loss": 0.7124, "step": 13717 }, { "epoch": 0.4968489677653024, "grad_norm": 1.5164208976985447, "learning_rate": 1.0587954182236886e-05, "loss": 0.7447, "step": 13718 }, { "epoch": 0.4968851865266208, "grad_norm": 1.2481950931561654, "learning_rate": 1.0586783139821863e-05, "loss": 0.6821, "step": 13719 }, { "epoch": 0.49692140528793916, "grad_norm": 1.4023412157211144, "learning_rate": 1.0585612089332186e-05, "loss": 0.7904, "step": 13720 }, { "epoch": 0.49695762404925753, "grad_norm": 2.052265158979254, "learning_rate": 1.0584441030783971e-05, "loss": 0.7252, "step": 13721 }, { "epoch": 0.4969938428105759, "grad_norm": 1.2444464645971454, "learning_rate": 1.058326996419334e-05, "loss": 0.7513, "step": 13722 }, { "epoch": 0.4970300615718942, "grad_norm": 1.3419673749231302, "learning_rate": 1.0582098889576396e-05, "loss": 0.7268, "step": 13723 }, { "epoch": 0.4970662803332126, "grad_norm": 1.309106181290831, "learning_rate": 1.0580927806949264e-05, "loss": 0.7611, "step": 13724 }, { "epoch": 0.49710249909453097, "grad_norm": 1.8215883024518094, "learning_rate": 1.0579756716328052e-05, "loss": 0.823, "step": 13725 }, { "epoch": 0.49713871785584934, "grad_norm": 1.4638104443689388, "learning_rate": 1.057858561772888e-05, "loss": 0.7612, "step": 13726 }, { "epoch": 0.4971749366171677, "grad_norm": 1.2474000597704171, "learning_rate": 1.0577414511167858e-05, "loss": 0.7203, "step": 13727 }, { "epoch": 0.49721115537848604, "grad_norm": 1.1310070793038394, "learning_rate": 1.057624339666111e-05, "loss": 0.7791, "step": 13728 }, { "epoch": 0.4972473741398044, "grad_norm": 1.262678497332302, "learning_rate": 1.0575072274224743e-05, "loss": 0.8148, "step": 13729 }, { "epoch": 0.4972835929011228, "grad_norm": 1.4225723762444236, "learning_rate": 1.0573901143874878e-05, "loss": 0.7628, "step": 13730 }, { "epoch": 0.49731981166244116, "grad_norm": 1.3236083477765743, "learning_rate": 1.0572730005627628e-05, "loss": 0.71, "step": 13731 }, { "epoch": 0.49735603042375953, "grad_norm": 1.2252542586447177, "learning_rate": 1.0571558859499106e-05, "loss": 0.7569, "step": 13732 }, { "epoch": 0.49739224918507785, "grad_norm": 1.2975372137560743, "learning_rate": 1.0570387705505438e-05, "loss": 0.7202, "step": 13733 }, { "epoch": 0.4974284679463962, "grad_norm": 1.289892687929129, "learning_rate": 1.0569216543662733e-05, "loss": 0.6869, "step": 13734 }, { "epoch": 0.4974646867077146, "grad_norm": 1.367117897714302, "learning_rate": 1.0568045373987107e-05, "loss": 0.7835, "step": 13735 }, { "epoch": 0.497500905469033, "grad_norm": 1.2823931323632067, "learning_rate": 1.0566874196494677e-05, "loss": 0.7402, "step": 13736 }, { "epoch": 0.49753712423035135, "grad_norm": 1.2999637507866129, "learning_rate": 1.0565703011201563e-05, "loss": 0.7771, "step": 13737 }, { "epoch": 0.49757334299166966, "grad_norm": 1.3684182077360287, "learning_rate": 1.0564531818123874e-05, "loss": 0.7179, "step": 13738 }, { "epoch": 0.49760956175298804, "grad_norm": 1.4509643204742946, "learning_rate": 1.0563360617277734e-05, "loss": 0.7793, "step": 13739 }, { "epoch": 0.4976457805143064, "grad_norm": 1.05748944430727, "learning_rate": 1.0562189408679255e-05, "loss": 0.7689, "step": 13740 }, { "epoch": 0.4976819992756248, "grad_norm": 1.4641894990256286, "learning_rate": 1.0561018192344557e-05, "loss": 0.7005, "step": 13741 }, { "epoch": 0.49771821803694316, "grad_norm": 1.3912379482452908, "learning_rate": 1.0559846968289755e-05, "loss": 0.816, "step": 13742 }, { "epoch": 0.4977544367982615, "grad_norm": 1.5049473304954222, "learning_rate": 1.0558675736530968e-05, "loss": 0.7753, "step": 13743 }, { "epoch": 0.49779065555957985, "grad_norm": 1.4107701360201579, "learning_rate": 1.0557504497084312e-05, "loss": 0.7392, "step": 13744 }, { "epoch": 0.4978268743208982, "grad_norm": 1.025069321860133, "learning_rate": 1.0556333249965902e-05, "loss": 0.6621, "step": 13745 }, { "epoch": 0.4978630930822166, "grad_norm": 1.3937284899877862, "learning_rate": 1.0555161995191863e-05, "loss": 0.7514, "step": 13746 }, { "epoch": 0.497899311843535, "grad_norm": 1.4686978437415865, "learning_rate": 1.0553990732778303e-05, "loss": 0.7884, "step": 13747 }, { "epoch": 0.4979355306048533, "grad_norm": 1.1037199200511227, "learning_rate": 1.0552819462741343e-05, "loss": 0.7383, "step": 13748 }, { "epoch": 0.49797174936617167, "grad_norm": 1.3792219165386645, "learning_rate": 1.0551648185097103e-05, "loss": 0.7952, "step": 13749 }, { "epoch": 0.49800796812749004, "grad_norm": 1.5161674219560335, "learning_rate": 1.0550476899861699e-05, "loss": 0.7494, "step": 13750 }, { "epoch": 0.4980441868888084, "grad_norm": 1.3290383452124916, "learning_rate": 1.054930560705125e-05, "loss": 0.6971, "step": 13751 }, { "epoch": 0.4980804056501268, "grad_norm": 1.29686421797763, "learning_rate": 1.0548134306681871e-05, "loss": 0.7341, "step": 13752 }, { "epoch": 0.4981166244114451, "grad_norm": 1.15982989229797, "learning_rate": 1.0546962998769686e-05, "loss": 0.7789, "step": 13753 }, { "epoch": 0.4981528431727635, "grad_norm": 1.3807454122865648, "learning_rate": 1.0545791683330803e-05, "loss": 0.7008, "step": 13754 }, { "epoch": 0.49818906193408186, "grad_norm": 1.0650075313007716, "learning_rate": 1.0544620360381355e-05, "loss": 0.7587, "step": 13755 }, { "epoch": 0.49822528069540023, "grad_norm": 1.3997946898384261, "learning_rate": 1.0543449029937447e-05, "loss": 0.7881, "step": 13756 }, { "epoch": 0.4982614994567186, "grad_norm": 1.4204884976642544, "learning_rate": 1.0542277692015204e-05, "loss": 0.7422, "step": 13757 }, { "epoch": 0.4982977182180369, "grad_norm": 1.3848215510181503, "learning_rate": 1.0541106346630744e-05, "loss": 0.8225, "step": 13758 }, { "epoch": 0.4983339369793553, "grad_norm": 1.4934229866407402, "learning_rate": 1.0539934993800183e-05, "loss": 0.7479, "step": 13759 }, { "epoch": 0.49837015574067367, "grad_norm": 1.004102366566452, "learning_rate": 1.0538763633539643e-05, "loss": 0.6797, "step": 13760 }, { "epoch": 0.49840637450199204, "grad_norm": 0.9776219131966002, "learning_rate": 1.0537592265865245e-05, "loss": 0.7038, "step": 13761 }, { "epoch": 0.4984425932633104, "grad_norm": 1.0625908458116518, "learning_rate": 1.0536420890793101e-05, "loss": 0.7636, "step": 13762 }, { "epoch": 0.49847881202462874, "grad_norm": 1.0691011688782495, "learning_rate": 1.0535249508339337e-05, "loss": 0.7152, "step": 13763 }, { "epoch": 0.4985150307859471, "grad_norm": 0.9514327273541967, "learning_rate": 1.0534078118520068e-05, "loss": 0.6711, "step": 13764 }, { "epoch": 0.4985512495472655, "grad_norm": 1.2580827769150562, "learning_rate": 1.0532906721351415e-05, "loss": 0.7371, "step": 13765 }, { "epoch": 0.49858746830858386, "grad_norm": 1.080813018788677, "learning_rate": 1.0531735316849498e-05, "loss": 0.738, "step": 13766 }, { "epoch": 0.49862368706990223, "grad_norm": 1.331690848077081, "learning_rate": 1.0530563905030436e-05, "loss": 0.7898, "step": 13767 }, { "epoch": 0.49865990583122055, "grad_norm": 1.4006788670068404, "learning_rate": 1.0529392485910346e-05, "loss": 0.7639, "step": 13768 }, { "epoch": 0.4986961245925389, "grad_norm": 1.3373162589025926, "learning_rate": 1.0528221059505353e-05, "loss": 0.6668, "step": 13769 }, { "epoch": 0.4987323433538573, "grad_norm": 1.3755810692433055, "learning_rate": 1.0527049625831572e-05, "loss": 0.8016, "step": 13770 }, { "epoch": 0.4987685621151757, "grad_norm": 1.6846183161875734, "learning_rate": 1.0525878184905127e-05, "loss": 0.7515, "step": 13771 }, { "epoch": 0.49880478087649405, "grad_norm": 1.3116610746671424, "learning_rate": 1.0524706736742135e-05, "loss": 0.7162, "step": 13772 }, { "epoch": 0.49884099963781237, "grad_norm": 1.5024615761559388, "learning_rate": 1.0523535281358718e-05, "loss": 0.6989, "step": 13773 }, { "epoch": 0.49887721839913074, "grad_norm": 1.1026906981147497, "learning_rate": 1.0522363818770995e-05, "loss": 0.707, "step": 13774 }, { "epoch": 0.4989134371604491, "grad_norm": 1.6113691220703408, "learning_rate": 1.052119234899509e-05, "loss": 0.7206, "step": 13775 }, { "epoch": 0.4989496559217675, "grad_norm": 1.3736600625817286, "learning_rate": 1.0520020872047115e-05, "loss": 0.8398, "step": 13776 }, { "epoch": 0.49898587468308586, "grad_norm": 1.3293969690054908, "learning_rate": 1.0518849387943198e-05, "loss": 0.7675, "step": 13777 }, { "epoch": 0.4990220934444042, "grad_norm": 1.3174650705274789, "learning_rate": 1.0517677896699456e-05, "loss": 0.7169, "step": 13778 }, { "epoch": 0.49905831220572255, "grad_norm": 1.3418306189601523, "learning_rate": 1.0516506398332014e-05, "loss": 0.7721, "step": 13779 }, { "epoch": 0.49909453096704093, "grad_norm": 1.1224516327729355, "learning_rate": 1.051533489285699e-05, "loss": 0.7932, "step": 13780 }, { "epoch": 0.4991307497283593, "grad_norm": 1.3738619655956645, "learning_rate": 1.0514163380290502e-05, "loss": 0.8114, "step": 13781 }, { "epoch": 0.4991669684896777, "grad_norm": 0.9915292952093654, "learning_rate": 1.0512991860648678e-05, "loss": 0.7804, "step": 13782 }, { "epoch": 0.499203187250996, "grad_norm": 1.4181021031287708, "learning_rate": 1.0511820333947632e-05, "loss": 0.6954, "step": 13783 }, { "epoch": 0.49923940601231437, "grad_norm": 1.4382438555606027, "learning_rate": 1.0510648800203491e-05, "loss": 0.7827, "step": 13784 }, { "epoch": 0.49927562477363274, "grad_norm": 1.0572081185434323, "learning_rate": 1.0509477259432372e-05, "loss": 0.7413, "step": 13785 }, { "epoch": 0.4993118435349511, "grad_norm": 1.4044472284962988, "learning_rate": 1.05083057116504e-05, "loss": 0.7143, "step": 13786 }, { "epoch": 0.4993480622962695, "grad_norm": 0.9903105291144094, "learning_rate": 1.0507134156873693e-05, "loss": 0.7496, "step": 13787 }, { "epoch": 0.4993842810575878, "grad_norm": 0.9499311969245701, "learning_rate": 1.0505962595118376e-05, "loss": 0.7126, "step": 13788 }, { "epoch": 0.4994204998189062, "grad_norm": 1.0359194911619771, "learning_rate": 1.0504791026400567e-05, "loss": 0.7553, "step": 13789 }, { "epoch": 0.49945671858022456, "grad_norm": 1.373673902244441, "learning_rate": 1.0503619450736392e-05, "loss": 0.6899, "step": 13790 }, { "epoch": 0.49949293734154293, "grad_norm": 1.3799346784494324, "learning_rate": 1.0502447868141969e-05, "loss": 0.7416, "step": 13791 }, { "epoch": 0.4995291561028613, "grad_norm": 1.3750965425887636, "learning_rate": 1.0501276278633423e-05, "loss": 0.8081, "step": 13792 }, { "epoch": 0.4995653748641796, "grad_norm": 1.3508951029113008, "learning_rate": 1.0500104682226876e-05, "loss": 0.7238, "step": 13793 }, { "epoch": 0.499601593625498, "grad_norm": 1.3232023620242903, "learning_rate": 1.0498933078938447e-05, "loss": 0.6945, "step": 13794 }, { "epoch": 0.49963781238681637, "grad_norm": 1.0664344866170545, "learning_rate": 1.0497761468784262e-05, "loss": 0.6976, "step": 13795 }, { "epoch": 0.49967403114813475, "grad_norm": 1.233334721273356, "learning_rate": 1.049658985178044e-05, "loss": 0.7473, "step": 13796 }, { "epoch": 0.4997102499094531, "grad_norm": 1.1035280181074925, "learning_rate": 1.0495418227943109e-05, "loss": 0.7414, "step": 13797 }, { "epoch": 0.49974646867077144, "grad_norm": 1.516075905970353, "learning_rate": 1.0494246597288382e-05, "loss": 0.799, "step": 13798 }, { "epoch": 0.4997826874320898, "grad_norm": 1.1598259665121253, "learning_rate": 1.049307495983239e-05, "loss": 0.7184, "step": 13799 }, { "epoch": 0.4998189061934082, "grad_norm": 1.3263380602448596, "learning_rate": 1.0491903315591256e-05, "loss": 0.7262, "step": 13800 }, { "epoch": 0.49985512495472656, "grad_norm": 1.5835488399127677, "learning_rate": 1.0490731664581098e-05, "loss": 0.7764, "step": 13801 }, { "epoch": 0.49989134371604493, "grad_norm": 1.4211629955540082, "learning_rate": 1.0489560006818042e-05, "loss": 0.6762, "step": 13802 }, { "epoch": 0.49992756247736325, "grad_norm": 1.30235535371047, "learning_rate": 1.048838834231821e-05, "loss": 0.785, "step": 13803 }, { "epoch": 0.4999637812386816, "grad_norm": 1.3652086037207622, "learning_rate": 1.0487216671097725e-05, "loss": 0.7405, "step": 13804 }, { "epoch": 0.5, "grad_norm": 1.7134567421349791, "learning_rate": 1.0486044993172708e-05, "loss": 0.6999, "step": 13805 }, { "epoch": 0.5000362187613183, "grad_norm": 1.3021996414372878, "learning_rate": 1.048487330855929e-05, "loss": 0.7106, "step": 13806 }, { "epoch": 0.5000724375226367, "grad_norm": 1.0906600404313547, "learning_rate": 1.0483701617273585e-05, "loss": 0.7457, "step": 13807 }, { "epoch": 0.5001086562839551, "grad_norm": 1.3889426947807542, "learning_rate": 1.0482529919331723e-05, "loss": 0.7723, "step": 13808 }, { "epoch": 0.5001448750452735, "grad_norm": 1.3644217690950218, "learning_rate": 1.0481358214749824e-05, "loss": 0.7078, "step": 13809 }, { "epoch": 0.5001810938065918, "grad_norm": 1.4883657328839441, "learning_rate": 1.0480186503544013e-05, "loss": 0.6836, "step": 13810 }, { "epoch": 0.5002173125679101, "grad_norm": 0.9748155359937305, "learning_rate": 1.0479014785730414e-05, "loss": 0.7594, "step": 13811 }, { "epoch": 0.5002535313292286, "grad_norm": 1.4009206931156726, "learning_rate": 1.047784306132515e-05, "loss": 0.7088, "step": 13812 }, { "epoch": 0.5002897500905469, "grad_norm": 1.3628063276422722, "learning_rate": 1.0476671330344349e-05, "loss": 0.738, "step": 13813 }, { "epoch": 0.5003259688518653, "grad_norm": 0.9517228116053658, "learning_rate": 1.0475499592804128e-05, "loss": 0.7203, "step": 13814 }, { "epoch": 0.5003621876131836, "grad_norm": 1.4193416259912, "learning_rate": 1.0474327848720616e-05, "loss": 0.6866, "step": 13815 }, { "epoch": 0.500398406374502, "grad_norm": 1.390057062051538, "learning_rate": 1.0473156098109935e-05, "loss": 0.7013, "step": 13816 }, { "epoch": 0.5004346251358204, "grad_norm": 1.415624507952388, "learning_rate": 1.0471984340988212e-05, "loss": 0.7411, "step": 13817 }, { "epoch": 0.5004708438971387, "grad_norm": 1.3293533507128004, "learning_rate": 1.047081257737157e-05, "loss": 0.7197, "step": 13818 }, { "epoch": 0.5005070626584571, "grad_norm": 1.4656091413422807, "learning_rate": 1.0469640807276133e-05, "loss": 0.7706, "step": 13819 }, { "epoch": 0.5005432814197754, "grad_norm": 1.0781550843768313, "learning_rate": 1.0468469030718023e-05, "loss": 0.7118, "step": 13820 }, { "epoch": 0.5005795001810938, "grad_norm": 1.2673688842224429, "learning_rate": 1.0467297247713368e-05, "loss": 0.701, "step": 13821 }, { "epoch": 0.5006157189424122, "grad_norm": 1.3639192918933338, "learning_rate": 1.0466125458278295e-05, "loss": 0.6935, "step": 13822 }, { "epoch": 0.5006519377037305, "grad_norm": 1.3075286758357214, "learning_rate": 1.0464953662428925e-05, "loss": 0.7411, "step": 13823 }, { "epoch": 0.5006881564650489, "grad_norm": 1.400252896749197, "learning_rate": 1.0463781860181385e-05, "loss": 0.6914, "step": 13824 }, { "epoch": 0.5007243752263673, "grad_norm": 1.0765219818166851, "learning_rate": 1.0462610051551798e-05, "loss": 0.7801, "step": 13825 }, { "epoch": 0.5007605939876856, "grad_norm": 1.0587120624573627, "learning_rate": 1.0461438236556292e-05, "loss": 0.7393, "step": 13826 }, { "epoch": 0.500796812749004, "grad_norm": 1.2285276573089225, "learning_rate": 1.0460266415210986e-05, "loss": 0.6809, "step": 13827 }, { "epoch": 0.5008330315103223, "grad_norm": 1.4912561960603226, "learning_rate": 1.0459094587532015e-05, "loss": 0.7202, "step": 13828 }, { "epoch": 0.5008692502716408, "grad_norm": 1.357074390504183, "learning_rate": 1.0457922753535494e-05, "loss": 0.7736, "step": 13829 }, { "epoch": 0.5009054690329591, "grad_norm": 1.3400946915965275, "learning_rate": 1.0456750913237557e-05, "loss": 0.6813, "step": 13830 }, { "epoch": 0.5009416877942774, "grad_norm": 1.2773866879540396, "learning_rate": 1.0455579066654327e-05, "loss": 0.7904, "step": 13831 }, { "epoch": 0.5009779065555958, "grad_norm": 1.5073987833010507, "learning_rate": 1.0454407213801924e-05, "loss": 0.7985, "step": 13832 }, { "epoch": 0.5010141253169141, "grad_norm": 1.373465221969348, "learning_rate": 1.0453235354696482e-05, "loss": 0.7414, "step": 13833 }, { "epoch": 0.5010503440782326, "grad_norm": 1.3594238365652627, "learning_rate": 1.0452063489354124e-05, "loss": 0.7866, "step": 13834 }, { "epoch": 0.5010865628395509, "grad_norm": 1.3394845739673717, "learning_rate": 1.0450891617790975e-05, "loss": 0.7408, "step": 13835 }, { "epoch": 0.5011227816008692, "grad_norm": 1.0353381703325069, "learning_rate": 1.0449719740023158e-05, "loss": 0.7414, "step": 13836 }, { "epoch": 0.5011590003621876, "grad_norm": 1.8573275019661148, "learning_rate": 1.0448547856066806e-05, "loss": 0.7193, "step": 13837 }, { "epoch": 0.501195219123506, "grad_norm": 0.948660288694681, "learning_rate": 1.0447375965938037e-05, "loss": 0.6754, "step": 13838 }, { "epoch": 0.5012314378848244, "grad_norm": 1.3383664607597023, "learning_rate": 1.0446204069652986e-05, "loss": 0.6677, "step": 13839 }, { "epoch": 0.5012676566461427, "grad_norm": 1.2147652119213799, "learning_rate": 1.0445032167227773e-05, "loss": 0.713, "step": 13840 }, { "epoch": 0.501303875407461, "grad_norm": 1.0891356161924313, "learning_rate": 1.0443860258678525e-05, "loss": 0.7422, "step": 13841 }, { "epoch": 0.5013400941687794, "grad_norm": 1.4932513975520565, "learning_rate": 1.0442688344021372e-05, "loss": 0.7825, "step": 13842 }, { "epoch": 0.5013763129300978, "grad_norm": 0.881598879548797, "learning_rate": 1.0441516423272437e-05, "loss": 0.7191, "step": 13843 }, { "epoch": 0.5014125316914162, "grad_norm": 1.196008660823254, "learning_rate": 1.044034449644785e-05, "loss": 0.7853, "step": 13844 }, { "epoch": 0.5014487504527345, "grad_norm": 1.402951158802435, "learning_rate": 1.0439172563563734e-05, "loss": 0.7414, "step": 13845 }, { "epoch": 0.5014849692140528, "grad_norm": 1.317204140210733, "learning_rate": 1.043800062463622e-05, "loss": 0.7357, "step": 13846 }, { "epoch": 0.5015211879753713, "grad_norm": 1.190084929688139, "learning_rate": 1.043682867968143e-05, "loss": 0.7373, "step": 13847 }, { "epoch": 0.5015574067366896, "grad_norm": 1.1273047971381733, "learning_rate": 1.0435656728715496e-05, "loss": 0.6513, "step": 13848 }, { "epoch": 0.501593625498008, "grad_norm": 1.3622192185092132, "learning_rate": 1.0434484771754538e-05, "loss": 0.7175, "step": 13849 }, { "epoch": 0.5016298442593263, "grad_norm": 1.3171137159527047, "learning_rate": 1.0433312808814694e-05, "loss": 0.7011, "step": 13850 }, { "epoch": 0.5016660630206446, "grad_norm": 1.0604698995586017, "learning_rate": 1.0432140839912081e-05, "loss": 0.8117, "step": 13851 }, { "epoch": 0.5017022817819631, "grad_norm": 1.0217558482831344, "learning_rate": 1.043096886506283e-05, "loss": 0.6665, "step": 13852 }, { "epoch": 0.5017385005432814, "grad_norm": 1.3319222483965303, "learning_rate": 1.0429796884283071e-05, "loss": 0.7514, "step": 13853 }, { "epoch": 0.5017747193045998, "grad_norm": 1.3165155571967504, "learning_rate": 1.0428624897588926e-05, "loss": 0.7034, "step": 13854 }, { "epoch": 0.5018109380659181, "grad_norm": 1.4458355202152173, "learning_rate": 1.0427452904996528e-05, "loss": 0.7789, "step": 13855 }, { "epoch": 0.5018471568272365, "grad_norm": 1.389948747452412, "learning_rate": 1.0426280906522003e-05, "loss": 0.7368, "step": 13856 }, { "epoch": 0.5018833755885549, "grad_norm": 1.3374491729002767, "learning_rate": 1.0425108902181475e-05, "loss": 0.7341, "step": 13857 }, { "epoch": 0.5019195943498732, "grad_norm": 1.0214078502546207, "learning_rate": 1.042393689199108e-05, "loss": 0.6796, "step": 13858 }, { "epoch": 0.5019558131111916, "grad_norm": 1.2845677083299203, "learning_rate": 1.0422764875966938e-05, "loss": 0.7078, "step": 13859 }, { "epoch": 0.50199203187251, "grad_norm": 1.632830043634594, "learning_rate": 1.042159285412518e-05, "loss": 0.7843, "step": 13860 }, { "epoch": 0.5020282506338283, "grad_norm": 1.3600509116889241, "learning_rate": 1.0420420826481935e-05, "loss": 0.73, "step": 13861 }, { "epoch": 0.5020644693951467, "grad_norm": 1.1355610929213498, "learning_rate": 1.041924879305333e-05, "loss": 0.7801, "step": 13862 }, { "epoch": 0.502100688156465, "grad_norm": 1.3032471417599105, "learning_rate": 1.041807675385549e-05, "loss": 0.7435, "step": 13863 }, { "epoch": 0.5021369069177835, "grad_norm": 1.2964096045763895, "learning_rate": 1.041690470890455e-05, "loss": 0.7552, "step": 13864 }, { "epoch": 0.5021731256791018, "grad_norm": 1.12348706429717, "learning_rate": 1.0415732658216634e-05, "loss": 0.7478, "step": 13865 }, { "epoch": 0.5022093444404201, "grad_norm": 1.2482182234664183, "learning_rate": 1.041456060180787e-05, "loss": 0.7168, "step": 13866 }, { "epoch": 0.5022455632017385, "grad_norm": 1.1936255149858237, "learning_rate": 1.0413388539694388e-05, "loss": 0.6895, "step": 13867 }, { "epoch": 0.5022817819630568, "grad_norm": 1.0455536108619303, "learning_rate": 1.0412216471892316e-05, "loss": 0.8176, "step": 13868 }, { "epoch": 0.5023180007243753, "grad_norm": 1.487857467832616, "learning_rate": 1.0411044398417784e-05, "loss": 0.7459, "step": 13869 }, { "epoch": 0.5023542194856936, "grad_norm": 1.1765105150880593, "learning_rate": 1.040987231928692e-05, "loss": 0.777, "step": 13870 }, { "epoch": 0.5023904382470119, "grad_norm": 1.3667330000049347, "learning_rate": 1.0408700234515855e-05, "loss": 0.6877, "step": 13871 }, { "epoch": 0.5024266570083303, "grad_norm": 1.3039719731033255, "learning_rate": 1.0407528144120713e-05, "loss": 0.6982, "step": 13872 }, { "epoch": 0.5024628757696487, "grad_norm": 1.4031462923167561, "learning_rate": 1.0406356048117628e-05, "loss": 0.7832, "step": 13873 }, { "epoch": 0.5024990945309671, "grad_norm": 1.4380904849507816, "learning_rate": 1.0405183946522723e-05, "loss": 0.6851, "step": 13874 }, { "epoch": 0.5025353132922854, "grad_norm": 1.033983512540471, "learning_rate": 1.0404011839352133e-05, "loss": 0.6938, "step": 13875 }, { "epoch": 0.5025715320536037, "grad_norm": 1.4079289162753843, "learning_rate": 1.0402839726621986e-05, "loss": 0.7097, "step": 13876 }, { "epoch": 0.5026077508149221, "grad_norm": 1.3739471834627843, "learning_rate": 1.0401667608348408e-05, "loss": 0.793, "step": 13877 }, { "epoch": 0.5026439695762405, "grad_norm": 1.4122795830765047, "learning_rate": 1.0400495484547533e-05, "loss": 0.7329, "step": 13878 }, { "epoch": 0.5026801883375589, "grad_norm": 1.3225314043657481, "learning_rate": 1.0399323355235486e-05, "loss": 0.7551, "step": 13879 }, { "epoch": 0.5027164070988772, "grad_norm": 1.340910272990824, "learning_rate": 1.03981512204284e-05, "loss": 0.7224, "step": 13880 }, { "epoch": 0.5027526258601955, "grad_norm": 1.0127993502534, "learning_rate": 1.0396979080142405e-05, "loss": 0.8026, "step": 13881 }, { "epoch": 0.502788844621514, "grad_norm": 1.372527965229702, "learning_rate": 1.0395806934393626e-05, "loss": 0.7188, "step": 13882 }, { "epoch": 0.5028250633828323, "grad_norm": 1.4856423338543645, "learning_rate": 1.0394634783198198e-05, "loss": 0.7192, "step": 13883 }, { "epoch": 0.5028612821441507, "grad_norm": 1.2605348604195346, "learning_rate": 1.0393462626572249e-05, "loss": 0.7029, "step": 13884 }, { "epoch": 0.502897500905469, "grad_norm": 1.0882494358168435, "learning_rate": 1.0392290464531906e-05, "loss": 0.671, "step": 13885 }, { "epoch": 0.5029337196667873, "grad_norm": 1.5327602062579286, "learning_rate": 1.0391118297093303e-05, "loss": 0.7466, "step": 13886 }, { "epoch": 0.5029699384281058, "grad_norm": 1.7558201731378134, "learning_rate": 1.038994612427257e-05, "loss": 0.7436, "step": 13887 }, { "epoch": 0.5030061571894241, "grad_norm": 1.3536907113077365, "learning_rate": 1.0388773946085833e-05, "loss": 0.7831, "step": 13888 }, { "epoch": 0.5030423759507425, "grad_norm": 1.0451066591859424, "learning_rate": 1.0387601762549227e-05, "loss": 0.7436, "step": 13889 }, { "epoch": 0.5030785947120608, "grad_norm": 1.0605054086759595, "learning_rate": 1.0386429573678878e-05, "loss": 0.6629, "step": 13890 }, { "epoch": 0.5031148134733792, "grad_norm": 1.409276685991465, "learning_rate": 1.038525737949092e-05, "loss": 0.7774, "step": 13891 }, { "epoch": 0.5031510322346976, "grad_norm": 1.492981831599583, "learning_rate": 1.0384085180001483e-05, "loss": 0.7941, "step": 13892 }, { "epoch": 0.5031872509960159, "grad_norm": 1.2933166758247003, "learning_rate": 1.0382912975226696e-05, "loss": 0.7525, "step": 13893 }, { "epoch": 0.5032234697573343, "grad_norm": 1.1967710744241449, "learning_rate": 1.0381740765182687e-05, "loss": 0.7289, "step": 13894 }, { "epoch": 0.5032596885186527, "grad_norm": 1.3212306079494782, "learning_rate": 1.0380568549885593e-05, "loss": 0.7425, "step": 13895 }, { "epoch": 0.503295907279971, "grad_norm": 0.9990330352000554, "learning_rate": 1.0379396329351538e-05, "loss": 0.7253, "step": 13896 }, { "epoch": 0.5033321260412894, "grad_norm": 1.4943864981609012, "learning_rate": 1.0378224103596662e-05, "loss": 0.788, "step": 13897 }, { "epoch": 0.5033683448026077, "grad_norm": 1.2701899213641585, "learning_rate": 1.0377051872637085e-05, "loss": 0.7103, "step": 13898 }, { "epoch": 0.5034045635639262, "grad_norm": 1.3579201730511643, "learning_rate": 1.0375879636488942e-05, "loss": 0.7838, "step": 13899 }, { "epoch": 0.5034407823252445, "grad_norm": 1.3421805109590366, "learning_rate": 1.0374707395168368e-05, "loss": 0.6873, "step": 13900 }, { "epoch": 0.5034770010865628, "grad_norm": 1.2594584576623589, "learning_rate": 1.037353514869149e-05, "loss": 0.6894, "step": 13901 }, { "epoch": 0.5035132198478812, "grad_norm": 1.2999981701074614, "learning_rate": 1.0372362897074441e-05, "loss": 0.7149, "step": 13902 }, { "epoch": 0.5035494386091995, "grad_norm": 1.2926692241154556, "learning_rate": 1.0371190640333351e-05, "loss": 0.7503, "step": 13903 }, { "epoch": 0.503585657370518, "grad_norm": 1.2822919986498078, "learning_rate": 1.0370018378484351e-05, "loss": 0.7599, "step": 13904 }, { "epoch": 0.5036218761318363, "grad_norm": 1.2103343821053076, "learning_rate": 1.0368846111543574e-05, "loss": 0.7906, "step": 13905 }, { "epoch": 0.5036580948931546, "grad_norm": 1.3955977395389318, "learning_rate": 1.036767383952715e-05, "loss": 0.7168, "step": 13906 }, { "epoch": 0.503694313654473, "grad_norm": 1.1717919971952806, "learning_rate": 1.036650156245121e-05, "loss": 0.7164, "step": 13907 }, { "epoch": 0.5037305324157914, "grad_norm": 1.344507140636678, "learning_rate": 1.0365329280331886e-05, "loss": 0.7317, "step": 13908 }, { "epoch": 0.5037667511771098, "grad_norm": 1.3593274134221196, "learning_rate": 1.0364156993185312e-05, "loss": 0.6849, "step": 13909 }, { "epoch": 0.5038029699384281, "grad_norm": 1.316583740312945, "learning_rate": 1.0362984701027617e-05, "loss": 0.7349, "step": 13910 }, { "epoch": 0.5038391886997464, "grad_norm": 1.4492023440098172, "learning_rate": 1.0361812403874934e-05, "loss": 0.7699, "step": 13911 }, { "epoch": 0.5038754074610649, "grad_norm": 1.2957548858885246, "learning_rate": 1.0360640101743393e-05, "loss": 0.6496, "step": 13912 }, { "epoch": 0.5039116262223832, "grad_norm": 1.1921650687136365, "learning_rate": 1.0359467794649129e-05, "loss": 0.7474, "step": 13913 }, { "epoch": 0.5039478449837016, "grad_norm": 1.3593483694800423, "learning_rate": 1.0358295482608269e-05, "loss": 0.7879, "step": 13914 }, { "epoch": 0.5039840637450199, "grad_norm": 1.2332537691563845, "learning_rate": 1.0357123165636953e-05, "loss": 0.7354, "step": 13915 }, { "epoch": 0.5040202825063382, "grad_norm": 1.1905051778362146, "learning_rate": 1.0355950843751306e-05, "loss": 0.7364, "step": 13916 }, { "epoch": 0.5040565012676567, "grad_norm": 1.2752095074885301, "learning_rate": 1.0354778516967463e-05, "loss": 0.7113, "step": 13917 }, { "epoch": 0.504092720028975, "grad_norm": 1.464236133440905, "learning_rate": 1.0353606185301554e-05, "loss": 0.7649, "step": 13918 }, { "epoch": 0.5041289387902934, "grad_norm": 1.0906497560300363, "learning_rate": 1.0352433848769717e-05, "loss": 0.7142, "step": 13919 }, { "epoch": 0.5041651575516117, "grad_norm": 1.2409678332894647, "learning_rate": 1.0351261507388077e-05, "loss": 0.7, "step": 13920 }, { "epoch": 0.50420137631293, "grad_norm": 1.3266744718175896, "learning_rate": 1.0350089161172771e-05, "loss": 0.7743, "step": 13921 }, { "epoch": 0.5042375950742485, "grad_norm": 0.951175204042096, "learning_rate": 1.0348916810139933e-05, "loss": 0.7518, "step": 13922 }, { "epoch": 0.5042738138355668, "grad_norm": 1.501992887360025, "learning_rate": 1.034774445430569e-05, "loss": 0.8356, "step": 13923 }, { "epoch": 0.5043100325968852, "grad_norm": 1.3200356655964978, "learning_rate": 1.0346572093686181e-05, "loss": 0.6666, "step": 13924 }, { "epoch": 0.5043462513582035, "grad_norm": 1.2236545512101904, "learning_rate": 1.0345399728297532e-05, "loss": 0.6894, "step": 13925 }, { "epoch": 0.5043824701195219, "grad_norm": 1.4231438869945252, "learning_rate": 1.0344227358155883e-05, "loss": 0.7267, "step": 13926 }, { "epoch": 0.5044186888808403, "grad_norm": 1.2788677005282658, "learning_rate": 1.0343054983277358e-05, "loss": 0.7353, "step": 13927 }, { "epoch": 0.5044549076421586, "grad_norm": 1.3321287071955525, "learning_rate": 1.0341882603678099e-05, "loss": 0.7647, "step": 13928 }, { "epoch": 0.504491126403477, "grad_norm": 1.3728334279964955, "learning_rate": 1.0340710219374233e-05, "loss": 0.7122, "step": 13929 }, { "epoch": 0.5045273451647954, "grad_norm": 1.363714733313542, "learning_rate": 1.0339537830381893e-05, "loss": 0.7524, "step": 13930 }, { "epoch": 0.5045635639261137, "grad_norm": 1.4107609796837006, "learning_rate": 1.033836543671722e-05, "loss": 0.6823, "step": 13931 }, { "epoch": 0.5045997826874321, "grad_norm": 1.3111707265649175, "learning_rate": 1.0337193038396338e-05, "loss": 0.6784, "step": 13932 }, { "epoch": 0.5046360014487504, "grad_norm": 1.5223839956322087, "learning_rate": 1.0336020635435384e-05, "loss": 0.7182, "step": 13933 }, { "epoch": 0.5046722202100689, "grad_norm": 1.1878874937273067, "learning_rate": 1.033484822785049e-05, "loss": 0.7371, "step": 13934 }, { "epoch": 0.5047084389713872, "grad_norm": 0.9943538374601866, "learning_rate": 1.033367581565779e-05, "loss": 0.7885, "step": 13935 }, { "epoch": 0.5047446577327055, "grad_norm": 1.129382402760634, "learning_rate": 1.0332503398873418e-05, "loss": 0.7152, "step": 13936 }, { "epoch": 0.5047808764940239, "grad_norm": 1.3258120936434612, "learning_rate": 1.033133097751351e-05, "loss": 0.7772, "step": 13937 }, { "epoch": 0.5048170952553422, "grad_norm": 1.0612799595325508, "learning_rate": 1.0330158551594191e-05, "loss": 0.7191, "step": 13938 }, { "epoch": 0.5048533140166607, "grad_norm": 1.5068780395005663, "learning_rate": 1.0328986121131604e-05, "loss": 0.7766, "step": 13939 }, { "epoch": 0.504889532777979, "grad_norm": 1.165620280041643, "learning_rate": 1.0327813686141878e-05, "loss": 0.7929, "step": 13940 }, { "epoch": 0.5049257515392973, "grad_norm": 1.1667229692015435, "learning_rate": 1.0326641246641147e-05, "loss": 0.7113, "step": 13941 }, { "epoch": 0.5049619703006157, "grad_norm": 1.5827219456096546, "learning_rate": 1.0325468802645546e-05, "loss": 0.6733, "step": 13942 }, { "epoch": 0.504998189061934, "grad_norm": 0.939940149071088, "learning_rate": 1.0324296354171209e-05, "loss": 0.7125, "step": 13943 }, { "epoch": 0.5050344078232525, "grad_norm": 1.0494706326543743, "learning_rate": 1.032312390123427e-05, "loss": 0.6874, "step": 13944 }, { "epoch": 0.5050706265845708, "grad_norm": 1.416811773170831, "learning_rate": 1.032195144385086e-05, "loss": 0.7528, "step": 13945 }, { "epoch": 0.5051068453458891, "grad_norm": 0.9759608274614288, "learning_rate": 1.0320778982037117e-05, "loss": 0.748, "step": 13946 }, { "epoch": 0.5051430641072076, "grad_norm": 1.054452471794187, "learning_rate": 1.031960651580917e-05, "loss": 0.7418, "step": 13947 }, { "epoch": 0.5051792828685259, "grad_norm": 1.1470317175783284, "learning_rate": 1.031843404518316e-05, "loss": 0.7396, "step": 13948 }, { "epoch": 0.5052155016298443, "grad_norm": 1.4650047913084838, "learning_rate": 1.0317261570175216e-05, "loss": 0.7687, "step": 13949 }, { "epoch": 0.5052517203911626, "grad_norm": 1.4317897482173392, "learning_rate": 1.0316089090801474e-05, "loss": 0.7126, "step": 13950 }, { "epoch": 0.5052879391524809, "grad_norm": 1.1740795987210162, "learning_rate": 1.031491660707807e-05, "loss": 0.7636, "step": 13951 }, { "epoch": 0.5053241579137994, "grad_norm": 1.0423698238712027, "learning_rate": 1.0313744119021134e-05, "loss": 0.7575, "step": 13952 }, { "epoch": 0.5053603766751177, "grad_norm": 1.4046575947782913, "learning_rate": 1.0312571626646805e-05, "loss": 0.7671, "step": 13953 }, { "epoch": 0.5053965954364361, "grad_norm": 1.4316458123023343, "learning_rate": 1.0311399129971215e-05, "loss": 0.7663, "step": 13954 }, { "epoch": 0.5054328141977544, "grad_norm": 1.5094865328786473, "learning_rate": 1.0310226629010503e-05, "loss": 0.739, "step": 13955 }, { "epoch": 0.5054690329590728, "grad_norm": 1.3014087495174123, "learning_rate": 1.0309054123780797e-05, "loss": 0.7261, "step": 13956 }, { "epoch": 0.5055052517203912, "grad_norm": 1.3620407983743452, "learning_rate": 1.0307881614298234e-05, "loss": 0.7014, "step": 13957 }, { "epoch": 0.5055414704817095, "grad_norm": 1.2317805004210982, "learning_rate": 1.0306709100578952e-05, "loss": 0.78, "step": 13958 }, { "epoch": 0.5055776892430279, "grad_norm": 1.0814251750538963, "learning_rate": 1.0305536582639082e-05, "loss": 0.6749, "step": 13959 }, { "epoch": 0.5056139080043462, "grad_norm": 0.9698132608131891, "learning_rate": 1.030436406049476e-05, "loss": 0.6904, "step": 13960 }, { "epoch": 0.5056501267656646, "grad_norm": 1.3006895608634068, "learning_rate": 1.030319153416212e-05, "loss": 0.7135, "step": 13961 }, { "epoch": 0.505686345526983, "grad_norm": 1.0025118707477592, "learning_rate": 1.03020190036573e-05, "loss": 0.7017, "step": 13962 }, { "epoch": 0.5057225642883013, "grad_norm": 1.321298858594986, "learning_rate": 1.0300846468996431e-05, "loss": 0.7295, "step": 13963 }, { "epoch": 0.5057587830496197, "grad_norm": 1.3270821958323153, "learning_rate": 1.0299673930195654e-05, "loss": 0.7785, "step": 13964 }, { "epoch": 0.5057950018109381, "grad_norm": 0.965090896093686, "learning_rate": 1.0298501387271094e-05, "loss": 0.71, "step": 13965 }, { "epoch": 0.5058312205722564, "grad_norm": 1.4590018959474307, "learning_rate": 1.0297328840238896e-05, "loss": 0.6699, "step": 13966 }, { "epoch": 0.5058674393335748, "grad_norm": 1.0032621483116144, "learning_rate": 1.0296156289115194e-05, "loss": 0.7483, "step": 13967 }, { "epoch": 0.5059036580948931, "grad_norm": 1.2886854819729665, "learning_rate": 1.0294983733916118e-05, "loss": 0.7201, "step": 13968 }, { "epoch": 0.5059398768562116, "grad_norm": 2.0330931444132156, "learning_rate": 1.0293811174657809e-05, "loss": 0.7467, "step": 13969 }, { "epoch": 0.5059760956175299, "grad_norm": 1.3981953842870465, "learning_rate": 1.0292638611356398e-05, "loss": 0.7029, "step": 13970 }, { "epoch": 0.5060123143788482, "grad_norm": 1.465828991482409, "learning_rate": 1.0291466044028024e-05, "loss": 0.7148, "step": 13971 }, { "epoch": 0.5060485331401666, "grad_norm": 1.2294197970191036, "learning_rate": 1.0290293472688821e-05, "loss": 0.7845, "step": 13972 }, { "epoch": 0.5060847519014849, "grad_norm": 1.3634540289883073, "learning_rate": 1.0289120897354925e-05, "loss": 0.7104, "step": 13973 }, { "epoch": 0.5061209706628034, "grad_norm": 1.3709678157344016, "learning_rate": 1.028794831804247e-05, "loss": 0.6974, "step": 13974 }, { "epoch": 0.5061571894241217, "grad_norm": 1.121121294458093, "learning_rate": 1.0286775734767596e-05, "loss": 0.7073, "step": 13975 }, { "epoch": 0.50619340818544, "grad_norm": 1.2527881506691958, "learning_rate": 1.0285603147546433e-05, "loss": 0.7221, "step": 13976 }, { "epoch": 0.5062296269467584, "grad_norm": 1.2755707121468869, "learning_rate": 1.028443055639512e-05, "loss": 0.7456, "step": 13977 }, { "epoch": 0.5062658457080768, "grad_norm": 1.3184150111856656, "learning_rate": 1.0283257961329794e-05, "loss": 0.7292, "step": 13978 }, { "epoch": 0.5063020644693952, "grad_norm": 1.3374796270886653, "learning_rate": 1.0282085362366587e-05, "loss": 0.7499, "step": 13979 }, { "epoch": 0.5063382832307135, "grad_norm": 1.044227619914048, "learning_rate": 1.0280912759521641e-05, "loss": 0.7497, "step": 13980 }, { "epoch": 0.5063745019920318, "grad_norm": 0.9332251656641197, "learning_rate": 1.0279740152811087e-05, "loss": 0.7162, "step": 13981 }, { "epoch": 0.5064107207533503, "grad_norm": 1.5157960483047688, "learning_rate": 1.0278567542251063e-05, "loss": 0.7879, "step": 13982 }, { "epoch": 0.5064469395146686, "grad_norm": 1.394371740851618, "learning_rate": 1.0277394927857705e-05, "loss": 0.7451, "step": 13983 }, { "epoch": 0.506483158275987, "grad_norm": 1.2588093269912908, "learning_rate": 1.027622230964715e-05, "loss": 0.6908, "step": 13984 }, { "epoch": 0.5065193770373053, "grad_norm": 0.987659415794878, "learning_rate": 1.027504968763553e-05, "loss": 0.7228, "step": 13985 }, { "epoch": 0.5065555957986236, "grad_norm": 1.2661633619531987, "learning_rate": 1.0273877061838986e-05, "loss": 0.6625, "step": 13986 }, { "epoch": 0.5065918145599421, "grad_norm": 1.5477664322412115, "learning_rate": 1.0272704432273654e-05, "loss": 0.7217, "step": 13987 }, { "epoch": 0.5066280333212604, "grad_norm": 1.3637265500111972, "learning_rate": 1.027153179895567e-05, "loss": 0.7434, "step": 13988 }, { "epoch": 0.5066642520825788, "grad_norm": 1.4250129465144545, "learning_rate": 1.027035916190117e-05, "loss": 0.7644, "step": 13989 }, { "epoch": 0.5067004708438971, "grad_norm": 1.0359115276089852, "learning_rate": 1.0269186521126287e-05, "loss": 0.7589, "step": 13990 }, { "epoch": 0.5067366896052155, "grad_norm": 1.390927102459947, "learning_rate": 1.0268013876647165e-05, "loss": 0.7347, "step": 13991 }, { "epoch": 0.5067729083665339, "grad_norm": 1.4668703962407401, "learning_rate": 1.0266841228479933e-05, "loss": 0.6536, "step": 13992 }, { "epoch": 0.5068091271278522, "grad_norm": 1.0115424287564259, "learning_rate": 1.0265668576640735e-05, "loss": 0.6889, "step": 13993 }, { "epoch": 0.5068453458891706, "grad_norm": 1.3171431660685968, "learning_rate": 1.02644959211457e-05, "loss": 0.6553, "step": 13994 }, { "epoch": 0.506881564650489, "grad_norm": 0.9898444431730793, "learning_rate": 1.0263323262010972e-05, "loss": 0.6916, "step": 13995 }, { "epoch": 0.5069177834118073, "grad_norm": 1.5583669119759875, "learning_rate": 1.0262150599252682e-05, "loss": 0.7255, "step": 13996 }, { "epoch": 0.5069540021731257, "grad_norm": 1.2815777901579402, "learning_rate": 1.026097793288697e-05, "loss": 0.7434, "step": 13997 }, { "epoch": 0.506990220934444, "grad_norm": 1.2388876201608137, "learning_rate": 1.0259805262929974e-05, "loss": 0.7479, "step": 13998 }, { "epoch": 0.5070264396957624, "grad_norm": 1.3808317829164911, "learning_rate": 1.0258632589397828e-05, "loss": 0.7196, "step": 13999 }, { "epoch": 0.5070626584570808, "grad_norm": 1.2497819568780453, "learning_rate": 1.025745991230667e-05, "loss": 0.7125, "step": 14000 }, { "epoch": 0.5070988772183991, "grad_norm": 1.3050564776694469, "learning_rate": 1.0256287231672637e-05, "loss": 0.7722, "step": 14001 }, { "epoch": 0.5071350959797175, "grad_norm": 1.5118105118770113, "learning_rate": 1.0255114547511868e-05, "loss": 0.7494, "step": 14002 }, { "epoch": 0.5071713147410358, "grad_norm": 1.233351640878206, "learning_rate": 1.0253941859840498e-05, "loss": 0.7454, "step": 14003 }, { "epoch": 0.5072075335023543, "grad_norm": 1.0241780550989565, "learning_rate": 1.0252769168674666e-05, "loss": 0.7285, "step": 14004 }, { "epoch": 0.5072437522636726, "grad_norm": 1.1523067749282172, "learning_rate": 1.0251596474030506e-05, "loss": 0.7369, "step": 14005 }, { "epoch": 0.5072799710249909, "grad_norm": 1.3380940442778613, "learning_rate": 1.025042377592416e-05, "loss": 0.7742, "step": 14006 }, { "epoch": 0.5073161897863093, "grad_norm": 1.6298101110683334, "learning_rate": 1.024925107437176e-05, "loss": 0.8411, "step": 14007 }, { "epoch": 0.5073524085476276, "grad_norm": 1.1078777478979032, "learning_rate": 1.0248078369389446e-05, "loss": 0.8047, "step": 14008 }, { "epoch": 0.5073886273089461, "grad_norm": 1.2108297653399511, "learning_rate": 1.0246905660993359e-05, "loss": 0.6935, "step": 14009 }, { "epoch": 0.5074248460702644, "grad_norm": 1.4044524386514026, "learning_rate": 1.0245732949199629e-05, "loss": 0.7647, "step": 14010 }, { "epoch": 0.5074610648315827, "grad_norm": 1.0862720240326642, "learning_rate": 1.0244560234024403e-05, "loss": 0.7207, "step": 14011 }, { "epoch": 0.5074972835929011, "grad_norm": 1.2972403666568104, "learning_rate": 1.024338751548381e-05, "loss": 0.7236, "step": 14012 }, { "epoch": 0.5075335023542195, "grad_norm": 1.3489582993502471, "learning_rate": 1.0242214793593993e-05, "loss": 0.7238, "step": 14013 }, { "epoch": 0.5075697211155379, "grad_norm": 1.2453374211630892, "learning_rate": 1.0241042068371084e-05, "loss": 0.7728, "step": 14014 }, { "epoch": 0.5076059398768562, "grad_norm": 1.6410596112932914, "learning_rate": 1.0239869339831225e-05, "loss": 0.7923, "step": 14015 }, { "epoch": 0.5076421586381745, "grad_norm": 1.2502964927291758, "learning_rate": 1.0238696607990554e-05, "loss": 0.7384, "step": 14016 }, { "epoch": 0.507678377399493, "grad_norm": 1.4012274775494342, "learning_rate": 1.0237523872865207e-05, "loss": 0.7704, "step": 14017 }, { "epoch": 0.5077145961608113, "grad_norm": 1.268151745662218, "learning_rate": 1.0236351134471325e-05, "loss": 0.7386, "step": 14018 }, { "epoch": 0.5077508149221297, "grad_norm": 1.2930034124274687, "learning_rate": 1.0235178392825043e-05, "loss": 0.7375, "step": 14019 }, { "epoch": 0.507787033683448, "grad_norm": 1.3509956119171023, "learning_rate": 1.0234005647942499e-05, "loss": 0.7633, "step": 14020 }, { "epoch": 0.5078232524447663, "grad_norm": 1.4709408771803538, "learning_rate": 1.0232832899839833e-05, "loss": 0.6991, "step": 14021 }, { "epoch": 0.5078594712060848, "grad_norm": 1.444331751385981, "learning_rate": 1.0231660148533183e-05, "loss": 0.8103, "step": 14022 }, { "epoch": 0.5078956899674031, "grad_norm": 1.528142800549232, "learning_rate": 1.0230487394038681e-05, "loss": 0.7677, "step": 14023 }, { "epoch": 0.5079319087287215, "grad_norm": 1.8102903014545153, "learning_rate": 1.0229314636372474e-05, "loss": 0.8158, "step": 14024 }, { "epoch": 0.5079681274900398, "grad_norm": 1.628282921723138, "learning_rate": 1.0228141875550695e-05, "loss": 0.7388, "step": 14025 }, { "epoch": 0.5080043462513582, "grad_norm": 1.3732469244766623, "learning_rate": 1.0226969111589482e-05, "loss": 0.7199, "step": 14026 }, { "epoch": 0.5080405650126766, "grad_norm": 1.3883498310673157, "learning_rate": 1.0225796344504975e-05, "loss": 0.7409, "step": 14027 }, { "epoch": 0.5080767837739949, "grad_norm": 1.273222362447646, "learning_rate": 1.0224623574313312e-05, "loss": 0.7571, "step": 14028 }, { "epoch": 0.5081130025353133, "grad_norm": 1.0441274409914767, "learning_rate": 1.0223450801030631e-05, "loss": 0.7902, "step": 14029 }, { "epoch": 0.5081492212966316, "grad_norm": 1.23502522646736, "learning_rate": 1.0222278024673072e-05, "loss": 0.6991, "step": 14030 }, { "epoch": 0.50818544005795, "grad_norm": 1.3302233728883983, "learning_rate": 1.0221105245256774e-05, "loss": 0.7466, "step": 14031 }, { "epoch": 0.5082216588192684, "grad_norm": 1.3762622939980376, "learning_rate": 1.0219932462797868e-05, "loss": 0.7664, "step": 14032 }, { "epoch": 0.5082578775805867, "grad_norm": 1.1668151729197482, "learning_rate": 1.0218759677312502e-05, "loss": 0.6876, "step": 14033 }, { "epoch": 0.5082940963419051, "grad_norm": 1.1240516572263415, "learning_rate": 1.0217586888816808e-05, "loss": 0.7107, "step": 14034 }, { "epoch": 0.5083303151032235, "grad_norm": 0.957279414576124, "learning_rate": 1.0216414097326933e-05, "loss": 0.6727, "step": 14035 }, { "epoch": 0.5083665338645418, "grad_norm": 1.3442096464666278, "learning_rate": 1.0215241302859003e-05, "loss": 0.705, "step": 14036 }, { "epoch": 0.5084027526258602, "grad_norm": 1.0451592962482974, "learning_rate": 1.021406850542917e-05, "loss": 0.7408, "step": 14037 }, { "epoch": 0.5084389713871785, "grad_norm": 1.1866608057447734, "learning_rate": 1.0212895705053559e-05, "loss": 0.7025, "step": 14038 }, { "epoch": 0.508475190148497, "grad_norm": 1.3486763079104747, "learning_rate": 1.021172290174832e-05, "loss": 0.731, "step": 14039 }, { "epoch": 0.5085114089098153, "grad_norm": 1.3195001223299112, "learning_rate": 1.0210550095529587e-05, "loss": 0.7124, "step": 14040 }, { "epoch": 0.5085476276711336, "grad_norm": 1.368963522619193, "learning_rate": 1.02093772864135e-05, "loss": 0.7036, "step": 14041 }, { "epoch": 0.508583846432452, "grad_norm": 1.4301253831187877, "learning_rate": 1.0208204474416199e-05, "loss": 0.8343, "step": 14042 }, { "epoch": 0.5086200651937703, "grad_norm": 1.3728034657003703, "learning_rate": 1.020703165955382e-05, "loss": 0.7334, "step": 14043 }, { "epoch": 0.5086562839550888, "grad_norm": 1.5020419219794754, "learning_rate": 1.0205858841842504e-05, "loss": 0.6776, "step": 14044 }, { "epoch": 0.5086925027164071, "grad_norm": 1.316309107470389, "learning_rate": 1.0204686021298387e-05, "loss": 0.7872, "step": 14045 }, { "epoch": 0.5087287214777254, "grad_norm": 1.2921247608909374, "learning_rate": 1.0203513197937614e-05, "loss": 0.7029, "step": 14046 }, { "epoch": 0.5087649402390438, "grad_norm": 1.3875019333370588, "learning_rate": 1.0202340371776318e-05, "loss": 0.7193, "step": 14047 }, { "epoch": 0.5088011590003622, "grad_norm": 1.4302776114176172, "learning_rate": 1.020116754283064e-05, "loss": 0.8039, "step": 14048 }, { "epoch": 0.5088373777616806, "grad_norm": 1.1241724256701278, "learning_rate": 1.0199994711116723e-05, "loss": 0.7677, "step": 14049 }, { "epoch": 0.5088735965229989, "grad_norm": 1.2586395131924872, "learning_rate": 1.0198821876650702e-05, "loss": 0.7332, "step": 14050 }, { "epoch": 0.5089098152843172, "grad_norm": 1.306454231099457, "learning_rate": 1.0197649039448718e-05, "loss": 0.7733, "step": 14051 }, { "epoch": 0.5089460340456357, "grad_norm": 1.3243202800706735, "learning_rate": 1.0196476199526907e-05, "loss": 0.7659, "step": 14052 }, { "epoch": 0.508982252806954, "grad_norm": 1.2749142465433918, "learning_rate": 1.0195303356901414e-05, "loss": 0.6711, "step": 14053 }, { "epoch": 0.5090184715682724, "grad_norm": 1.3652372141905418, "learning_rate": 1.0194130511588371e-05, "loss": 0.7402, "step": 14054 }, { "epoch": 0.5090546903295907, "grad_norm": 1.396174703288526, "learning_rate": 1.0192957663603927e-05, "loss": 0.7662, "step": 14055 }, { "epoch": 0.509090909090909, "grad_norm": 0.9874447076002836, "learning_rate": 1.0191784812964212e-05, "loss": 0.7537, "step": 14056 }, { "epoch": 0.5091271278522275, "grad_norm": 1.0831993146981356, "learning_rate": 1.0190611959685372e-05, "loss": 0.7473, "step": 14057 }, { "epoch": 0.5091633466135458, "grad_norm": 1.250750769220543, "learning_rate": 1.0189439103783541e-05, "loss": 0.743, "step": 14058 }, { "epoch": 0.5091995653748642, "grad_norm": 1.3423800124437937, "learning_rate": 1.0188266245274862e-05, "loss": 0.7046, "step": 14059 }, { "epoch": 0.5092357841361825, "grad_norm": 1.1712583938280556, "learning_rate": 1.0187093384175476e-05, "loss": 0.7323, "step": 14060 }, { "epoch": 0.5092720028975009, "grad_norm": 1.5108855745985093, "learning_rate": 1.018592052050152e-05, "loss": 0.7706, "step": 14061 }, { "epoch": 0.5093082216588193, "grad_norm": 1.5241263006183345, "learning_rate": 1.0184747654269136e-05, "loss": 0.7464, "step": 14062 }, { "epoch": 0.5093444404201376, "grad_norm": 1.0000955973226746, "learning_rate": 1.018357478549446e-05, "loss": 0.6911, "step": 14063 }, { "epoch": 0.509380659181456, "grad_norm": 1.2361080303712588, "learning_rate": 1.0182401914193631e-05, "loss": 0.7249, "step": 14064 }, { "epoch": 0.5094168779427743, "grad_norm": 1.1209508735464744, "learning_rate": 1.0181229040382797e-05, "loss": 0.7766, "step": 14065 }, { "epoch": 0.5094530967040927, "grad_norm": 1.3329895437476895, "learning_rate": 1.0180056164078087e-05, "loss": 0.748, "step": 14066 }, { "epoch": 0.5094893154654111, "grad_norm": 1.351855882874552, "learning_rate": 1.017888328529565e-05, "loss": 0.7959, "step": 14067 }, { "epoch": 0.5095255342267294, "grad_norm": 1.495830566275073, "learning_rate": 1.0177710404051619e-05, "loss": 0.7801, "step": 14068 }, { "epoch": 0.5095617529880478, "grad_norm": 1.4402190102858747, "learning_rate": 1.017653752036214e-05, "loss": 0.7886, "step": 14069 }, { "epoch": 0.5095979717493662, "grad_norm": 1.2515220800163993, "learning_rate": 1.0175364634243345e-05, "loss": 0.6993, "step": 14070 }, { "epoch": 0.5096341905106845, "grad_norm": 1.207440168130473, "learning_rate": 1.0174191745711381e-05, "loss": 0.682, "step": 14071 }, { "epoch": 0.5096704092720029, "grad_norm": 1.3925760919200294, "learning_rate": 1.0173018854782386e-05, "loss": 0.7042, "step": 14072 }, { "epoch": 0.5097066280333212, "grad_norm": 1.376109954021294, "learning_rate": 1.01718459614725e-05, "loss": 0.7737, "step": 14073 }, { "epoch": 0.5097428467946397, "grad_norm": 1.2983828978987366, "learning_rate": 1.017067306579786e-05, "loss": 0.7223, "step": 14074 }, { "epoch": 0.509779065555958, "grad_norm": 1.0231848805471506, "learning_rate": 1.016950016777461e-05, "loss": 0.7836, "step": 14075 }, { "epoch": 0.5098152843172763, "grad_norm": 1.4515888531492585, "learning_rate": 1.0168327267418889e-05, "loss": 0.7054, "step": 14076 }, { "epoch": 0.5098515030785947, "grad_norm": 1.287033230029898, "learning_rate": 1.0167154364746836e-05, "loss": 0.7592, "step": 14077 }, { "epoch": 0.509887721839913, "grad_norm": 1.6340501694059917, "learning_rate": 1.0165981459774595e-05, "loss": 0.7704, "step": 14078 }, { "epoch": 0.5099239406012315, "grad_norm": 1.3168240925440968, "learning_rate": 1.0164808552518299e-05, "loss": 0.7496, "step": 14079 }, { "epoch": 0.5099601593625498, "grad_norm": 1.3161683560950046, "learning_rate": 1.0163635642994094e-05, "loss": 0.6689, "step": 14080 }, { "epoch": 0.5099963781238681, "grad_norm": 1.2214461262664333, "learning_rate": 1.016246273121812e-05, "loss": 0.674, "step": 14081 }, { "epoch": 0.5100325968851865, "grad_norm": 1.4236300648581506, "learning_rate": 1.0161289817206517e-05, "loss": 0.7578, "step": 14082 }, { "epoch": 0.5100688156465049, "grad_norm": 1.3874998955039253, "learning_rate": 1.016011690097542e-05, "loss": 0.78, "step": 14083 }, { "epoch": 0.5101050344078233, "grad_norm": 1.178977827846808, "learning_rate": 1.0158943982540978e-05, "loss": 0.7433, "step": 14084 }, { "epoch": 0.5101412531691416, "grad_norm": 1.3507681762973838, "learning_rate": 1.0157771061919327e-05, "loss": 0.74, "step": 14085 }, { "epoch": 0.5101774719304599, "grad_norm": 1.3301567741981368, "learning_rate": 1.0156598139126603e-05, "loss": 0.821, "step": 14086 }, { "epoch": 0.5102136906917784, "grad_norm": 1.4010268698647188, "learning_rate": 1.0155425214178955e-05, "loss": 0.6437, "step": 14087 }, { "epoch": 0.5102499094530967, "grad_norm": 1.2909559713654473, "learning_rate": 1.015425228709252e-05, "loss": 0.7495, "step": 14088 }, { "epoch": 0.5102861282144151, "grad_norm": 1.531503500986585, "learning_rate": 1.0153079357883435e-05, "loss": 0.7112, "step": 14089 }, { "epoch": 0.5103223469757334, "grad_norm": 1.0656029560778464, "learning_rate": 1.0151906426567845e-05, "loss": 0.7832, "step": 14090 }, { "epoch": 0.5103585657370517, "grad_norm": 1.3698260255560866, "learning_rate": 1.015073349316189e-05, "loss": 0.7055, "step": 14091 }, { "epoch": 0.5103947844983702, "grad_norm": 1.0804822762305328, "learning_rate": 1.0149560557681708e-05, "loss": 0.7162, "step": 14092 }, { "epoch": 0.5104310032596885, "grad_norm": 1.3049334005945254, "learning_rate": 1.0148387620143445e-05, "loss": 0.7515, "step": 14093 }, { "epoch": 0.5104672220210069, "grad_norm": 1.371673007200873, "learning_rate": 1.0147214680563233e-05, "loss": 0.7023, "step": 14094 }, { "epoch": 0.5105034407823252, "grad_norm": 1.2212168121974463, "learning_rate": 1.014604173895722e-05, "loss": 0.7481, "step": 14095 }, { "epoch": 0.5105396595436436, "grad_norm": 1.7252048235909305, "learning_rate": 1.0144868795341544e-05, "loss": 0.7352, "step": 14096 }, { "epoch": 0.510575878304962, "grad_norm": 1.3931143732701015, "learning_rate": 1.0143695849732347e-05, "loss": 0.7587, "step": 14097 }, { "epoch": 0.5106120970662803, "grad_norm": 1.285981664704539, "learning_rate": 1.0142522902145768e-05, "loss": 0.7726, "step": 14098 }, { "epoch": 0.5106483158275987, "grad_norm": 1.0787242828617998, "learning_rate": 1.0141349952597948e-05, "loss": 0.8088, "step": 14099 }, { "epoch": 0.510684534588917, "grad_norm": 1.3960509754772945, "learning_rate": 1.0140177001105031e-05, "loss": 0.7365, "step": 14100 }, { "epoch": 0.5107207533502354, "grad_norm": 0.9843210205122814, "learning_rate": 1.0139004047683152e-05, "loss": 0.6818, "step": 14101 }, { "epoch": 0.5107569721115538, "grad_norm": 1.4704853874418944, "learning_rate": 1.0137831092348458e-05, "loss": 0.7858, "step": 14102 }, { "epoch": 0.5107931908728721, "grad_norm": 1.4269277026305562, "learning_rate": 1.0136658135117086e-05, "loss": 0.7114, "step": 14103 }, { "epoch": 0.5108294096341905, "grad_norm": 0.9666857289174987, "learning_rate": 1.0135485176005178e-05, "loss": 0.7683, "step": 14104 }, { "epoch": 0.5108656283955089, "grad_norm": 1.3149631674092388, "learning_rate": 1.0134312215028875e-05, "loss": 0.734, "step": 14105 }, { "epoch": 0.5109018471568272, "grad_norm": 1.1234498950594625, "learning_rate": 1.0133139252204315e-05, "loss": 0.7555, "step": 14106 }, { "epoch": 0.5109380659181456, "grad_norm": 1.40493376247169, "learning_rate": 1.0131966287547646e-05, "loss": 0.7108, "step": 14107 }, { "epoch": 0.5109742846794639, "grad_norm": 1.3372282644001436, "learning_rate": 1.0130793321075003e-05, "loss": 0.7793, "step": 14108 }, { "epoch": 0.5110105034407824, "grad_norm": 1.2970358399823618, "learning_rate": 1.012962035280253e-05, "loss": 0.7514, "step": 14109 }, { "epoch": 0.5110467222021007, "grad_norm": 1.005617096226157, "learning_rate": 1.0128447382746365e-05, "loss": 0.7347, "step": 14110 }, { "epoch": 0.511082940963419, "grad_norm": 1.3193281046940513, "learning_rate": 1.0127274410922653e-05, "loss": 0.7713, "step": 14111 }, { "epoch": 0.5111191597247374, "grad_norm": 1.6135711674972208, "learning_rate": 1.0126101437347533e-05, "loss": 0.7588, "step": 14112 }, { "epoch": 0.5111553784860557, "grad_norm": 1.3939217055402169, "learning_rate": 1.0124928462037149e-05, "loss": 0.7165, "step": 14113 }, { "epoch": 0.5111915972473742, "grad_norm": 1.3634745022285917, "learning_rate": 1.0123755485007637e-05, "loss": 0.7219, "step": 14114 }, { "epoch": 0.5112278160086925, "grad_norm": 1.1159724789526742, "learning_rate": 1.0122582506275141e-05, "loss": 0.7124, "step": 14115 }, { "epoch": 0.5112640347700108, "grad_norm": 1.7895174718153364, "learning_rate": 1.0121409525855798e-05, "loss": 0.7085, "step": 14116 }, { "epoch": 0.5113002535313292, "grad_norm": 1.3222686628084208, "learning_rate": 1.0120236543765759e-05, "loss": 0.7229, "step": 14117 }, { "epoch": 0.5113364722926476, "grad_norm": 1.427736699823869, "learning_rate": 1.0119063560021156e-05, "loss": 0.7633, "step": 14118 }, { "epoch": 0.511372691053966, "grad_norm": 1.3356099396393393, "learning_rate": 1.0117890574638137e-05, "loss": 0.7241, "step": 14119 }, { "epoch": 0.5114089098152843, "grad_norm": 0.9698636193975085, "learning_rate": 1.0116717587632838e-05, "loss": 0.7354, "step": 14120 }, { "epoch": 0.5114451285766026, "grad_norm": 1.046314433592052, "learning_rate": 1.01155445990214e-05, "loss": 0.7333, "step": 14121 }, { "epoch": 0.511481347337921, "grad_norm": 1.3298421050262004, "learning_rate": 1.0114371608819972e-05, "loss": 0.7237, "step": 14122 }, { "epoch": 0.5115175660992394, "grad_norm": 1.3863419336462668, "learning_rate": 1.0113198617044686e-05, "loss": 0.72, "step": 14123 }, { "epoch": 0.5115537848605578, "grad_norm": 1.3619588180435698, "learning_rate": 1.0112025623711692e-05, "loss": 0.7174, "step": 14124 }, { "epoch": 0.5115900036218761, "grad_norm": 1.2756490796029858, "learning_rate": 1.0110852628837122e-05, "loss": 0.7103, "step": 14125 }, { "epoch": 0.5116262223831944, "grad_norm": 1.3340598157659993, "learning_rate": 1.0109679632437123e-05, "loss": 0.694, "step": 14126 }, { "epoch": 0.5116624411445129, "grad_norm": 1.0261833192560084, "learning_rate": 1.0108506634527839e-05, "loss": 0.7412, "step": 14127 }, { "epoch": 0.5116986599058312, "grad_norm": 1.4096889236631585, "learning_rate": 1.0107333635125407e-05, "loss": 0.7453, "step": 14128 }, { "epoch": 0.5117348786671496, "grad_norm": 1.3576468027501567, "learning_rate": 1.0106160634245972e-05, "loss": 0.809, "step": 14129 }, { "epoch": 0.5117710974284679, "grad_norm": 1.0508044377789967, "learning_rate": 1.010498763190567e-05, "loss": 0.7491, "step": 14130 }, { "epoch": 0.5118073161897863, "grad_norm": 1.3197537289395325, "learning_rate": 1.0103814628120648e-05, "loss": 0.7211, "step": 14131 }, { "epoch": 0.5118435349511047, "grad_norm": 1.2509113007376582, "learning_rate": 1.0102641622907043e-05, "loss": 0.7005, "step": 14132 }, { "epoch": 0.511879753712423, "grad_norm": 1.3297909995286339, "learning_rate": 1.0101468616281003e-05, "loss": 0.6825, "step": 14133 }, { "epoch": 0.5119159724737414, "grad_norm": 1.3887578237469413, "learning_rate": 1.010029560825866e-05, "loss": 0.7987, "step": 14134 }, { "epoch": 0.5119521912350598, "grad_norm": 1.3156025319151665, "learning_rate": 1.0099122598856168e-05, "loss": 0.6916, "step": 14135 }, { "epoch": 0.5119884099963781, "grad_norm": 1.3691844579452424, "learning_rate": 1.0097949588089657e-05, "loss": 0.8206, "step": 14136 }, { "epoch": 0.5120246287576965, "grad_norm": 1.2906420473065126, "learning_rate": 1.0096776575975276e-05, "loss": 0.67, "step": 14137 }, { "epoch": 0.5120608475190148, "grad_norm": 1.7497924730598657, "learning_rate": 1.0095603562529163e-05, "loss": 0.6883, "step": 14138 }, { "epoch": 0.5120970662803332, "grad_norm": 1.4067211633373635, "learning_rate": 1.0094430547767462e-05, "loss": 0.7159, "step": 14139 }, { "epoch": 0.5121332850416516, "grad_norm": 1.4221714989188048, "learning_rate": 1.0093257531706313e-05, "loss": 0.7159, "step": 14140 }, { "epoch": 0.5121695038029699, "grad_norm": 1.3578920092256586, "learning_rate": 1.0092084514361857e-05, "loss": 0.6884, "step": 14141 }, { "epoch": 0.5122057225642883, "grad_norm": 1.4486461173849754, "learning_rate": 1.0090911495750243e-05, "loss": 0.6377, "step": 14142 }, { "epoch": 0.5122419413256066, "grad_norm": 1.4330259952571835, "learning_rate": 1.00897384758876e-05, "loss": 0.7857, "step": 14143 }, { "epoch": 0.5122781600869251, "grad_norm": 1.2731120031446705, "learning_rate": 1.008856545479008e-05, "loss": 0.7234, "step": 14144 }, { "epoch": 0.5123143788482434, "grad_norm": 1.3519503114236224, "learning_rate": 1.008739243247382e-05, "loss": 0.763, "step": 14145 }, { "epoch": 0.5123505976095617, "grad_norm": 1.3415603359794979, "learning_rate": 1.0086219408954966e-05, "loss": 0.7374, "step": 14146 }, { "epoch": 0.5123868163708801, "grad_norm": 1.1224439269128175, "learning_rate": 1.0085046384249654e-05, "loss": 0.7298, "step": 14147 }, { "epoch": 0.5124230351321984, "grad_norm": 1.4865312725698174, "learning_rate": 1.008387335837403e-05, "loss": 0.7797, "step": 14148 }, { "epoch": 0.5124592538935169, "grad_norm": 1.2968252550201813, "learning_rate": 1.0082700331344238e-05, "loss": 0.8052, "step": 14149 }, { "epoch": 0.5124954726548352, "grad_norm": 1.260593600570037, "learning_rate": 1.0081527303176413e-05, "loss": 0.6569, "step": 14150 }, { "epoch": 0.5125316914161536, "grad_norm": 1.4306980337334505, "learning_rate": 1.0080354273886701e-05, "loss": 0.8307, "step": 14151 }, { "epoch": 0.5125679101774719, "grad_norm": 2.1257989499406857, "learning_rate": 1.0079181243491245e-05, "loss": 0.7263, "step": 14152 }, { "epoch": 0.5126041289387903, "grad_norm": 1.523423484591057, "learning_rate": 1.0078008212006185e-05, "loss": 0.7313, "step": 14153 }, { "epoch": 0.5126403477001087, "grad_norm": 1.531068049052706, "learning_rate": 1.0076835179447664e-05, "loss": 0.7553, "step": 14154 }, { "epoch": 0.512676566461427, "grad_norm": 1.375706505383254, "learning_rate": 1.0075662145831823e-05, "loss": 0.7275, "step": 14155 }, { "epoch": 0.5127127852227454, "grad_norm": 1.3264153355197015, "learning_rate": 1.0074489111174804e-05, "loss": 0.6887, "step": 14156 }, { "epoch": 0.5127490039840638, "grad_norm": 1.3485022874630823, "learning_rate": 1.0073316075492748e-05, "loss": 0.7315, "step": 14157 }, { "epoch": 0.5127852227453821, "grad_norm": 1.4142915101627314, "learning_rate": 1.00721430388018e-05, "loss": 0.7928, "step": 14158 }, { "epoch": 0.5128214415067005, "grad_norm": 1.4012781626638697, "learning_rate": 1.00709700011181e-05, "loss": 0.6703, "step": 14159 }, { "epoch": 0.5128576602680188, "grad_norm": 0.9520498791755518, "learning_rate": 1.0069796962457792e-05, "loss": 0.6888, "step": 14160 }, { "epoch": 0.5128938790293373, "grad_norm": 1.349299878407583, "learning_rate": 1.0068623922837016e-05, "loss": 0.7645, "step": 14161 }, { "epoch": 0.5129300977906556, "grad_norm": 1.3099095945475976, "learning_rate": 1.0067450882271915e-05, "loss": 0.7517, "step": 14162 }, { "epoch": 0.5129663165519739, "grad_norm": 1.2766175940179827, "learning_rate": 1.0066277840778626e-05, "loss": 0.7036, "step": 14163 }, { "epoch": 0.5130025353132923, "grad_norm": 1.3433752972179764, "learning_rate": 1.0065104798373302e-05, "loss": 0.7266, "step": 14164 }, { "epoch": 0.5130387540746106, "grad_norm": 1.25913228075463, "learning_rate": 1.0063931755072074e-05, "loss": 0.7921, "step": 14165 }, { "epoch": 0.5130749728359291, "grad_norm": 1.4463327827516672, "learning_rate": 1.0062758710891091e-05, "loss": 0.7777, "step": 14166 }, { "epoch": 0.5131111915972474, "grad_norm": 1.2956124157531357, "learning_rate": 1.0061585665846494e-05, "loss": 0.7182, "step": 14167 }, { "epoch": 0.5131474103585657, "grad_norm": 1.4216204634189211, "learning_rate": 1.0060412619954423e-05, "loss": 0.7163, "step": 14168 }, { "epoch": 0.5131836291198841, "grad_norm": 1.3869541707139956, "learning_rate": 1.0059239573231021e-05, "loss": 0.8199, "step": 14169 }, { "epoch": 0.5132198478812025, "grad_norm": 1.6354987507061396, "learning_rate": 1.005806652569243e-05, "loss": 0.7334, "step": 14170 }, { "epoch": 0.5132560666425209, "grad_norm": 1.4114722559712614, "learning_rate": 1.0056893477354795e-05, "loss": 0.7027, "step": 14171 }, { "epoch": 0.5132922854038392, "grad_norm": 1.229107117742632, "learning_rate": 1.0055720428234253e-05, "loss": 0.6759, "step": 14172 }, { "epoch": 0.5133285041651575, "grad_norm": 1.370491635331881, "learning_rate": 1.0054547378346951e-05, "loss": 0.7433, "step": 14173 }, { "epoch": 0.513364722926476, "grad_norm": 1.4422250053120662, "learning_rate": 1.0053374327709027e-05, "loss": 0.7975, "step": 14174 }, { "epoch": 0.5134009416877943, "grad_norm": 1.4852665382142545, "learning_rate": 1.0052201276336624e-05, "loss": 0.7486, "step": 14175 }, { "epoch": 0.5134371604491127, "grad_norm": 1.3423969087675092, "learning_rate": 1.0051028224245888e-05, "loss": 0.7385, "step": 14176 }, { "epoch": 0.513473379210431, "grad_norm": 1.4202900990208358, "learning_rate": 1.004985517145296e-05, "loss": 0.8025, "step": 14177 }, { "epoch": 0.5135095979717493, "grad_norm": 1.3538726869617435, "learning_rate": 1.004868211797398e-05, "loss": 0.6523, "step": 14178 }, { "epoch": 0.5135458167330678, "grad_norm": 1.432181805096382, "learning_rate": 1.004750906382509e-05, "loss": 0.7295, "step": 14179 }, { "epoch": 0.5135820354943861, "grad_norm": 1.2413524266416953, "learning_rate": 1.0046336009022435e-05, "loss": 0.7075, "step": 14180 }, { "epoch": 0.5136182542557045, "grad_norm": 1.4049330518488272, "learning_rate": 1.0045162953582154e-05, "loss": 0.7757, "step": 14181 }, { "epoch": 0.5136544730170228, "grad_norm": 1.0009521947778968, "learning_rate": 1.0043989897520394e-05, "loss": 0.7227, "step": 14182 }, { "epoch": 0.5136906917783411, "grad_norm": 1.2430821561838254, "learning_rate": 1.0042816840853293e-05, "loss": 0.7378, "step": 14183 }, { "epoch": 0.5137269105396596, "grad_norm": 1.3769630242390933, "learning_rate": 1.0041643783596993e-05, "loss": 0.7201, "step": 14184 }, { "epoch": 0.5137631293009779, "grad_norm": 1.4213997680736, "learning_rate": 1.004047072576764e-05, "loss": 0.8147, "step": 14185 }, { "epoch": 0.5137993480622963, "grad_norm": 1.3566443432538113, "learning_rate": 1.0039297667381373e-05, "loss": 0.6708, "step": 14186 }, { "epoch": 0.5138355668236146, "grad_norm": 1.3761279914219482, "learning_rate": 1.0038124608454337e-05, "loss": 0.6591, "step": 14187 }, { "epoch": 0.513871785584933, "grad_norm": 1.3389684943471503, "learning_rate": 1.0036951549002672e-05, "loss": 0.6692, "step": 14188 }, { "epoch": 0.5139080043462514, "grad_norm": 1.2444626860670698, "learning_rate": 1.0035778489042522e-05, "loss": 0.771, "step": 14189 }, { "epoch": 0.5139442231075697, "grad_norm": 1.5185008551571164, "learning_rate": 1.0034605428590027e-05, "loss": 0.7378, "step": 14190 }, { "epoch": 0.5139804418688881, "grad_norm": 1.4134642615895912, "learning_rate": 1.0033432367661333e-05, "loss": 0.7028, "step": 14191 }, { "epoch": 0.5140166606302065, "grad_norm": 1.2891256808200822, "learning_rate": 1.0032259306272579e-05, "loss": 0.71, "step": 14192 }, { "epoch": 0.5140528793915248, "grad_norm": 1.409591749381627, "learning_rate": 1.003108624443991e-05, "loss": 0.7202, "step": 14193 }, { "epoch": 0.5140890981528432, "grad_norm": 1.3249466507446013, "learning_rate": 1.0029913182179463e-05, "loss": 0.7713, "step": 14194 }, { "epoch": 0.5141253169141615, "grad_norm": 1.0944347822214022, "learning_rate": 1.0028740119507386e-05, "loss": 0.6434, "step": 14195 }, { "epoch": 0.51416153567548, "grad_norm": 1.0120783325284686, "learning_rate": 1.0027567056439821e-05, "loss": 0.7486, "step": 14196 }, { "epoch": 0.5141977544367983, "grad_norm": 1.4385668382461154, "learning_rate": 1.002639399299291e-05, "loss": 0.7701, "step": 14197 }, { "epoch": 0.5142339731981166, "grad_norm": 1.3590492345996263, "learning_rate": 1.0025220929182793e-05, "loss": 0.7479, "step": 14198 }, { "epoch": 0.514270191959435, "grad_norm": 1.3338527733214474, "learning_rate": 1.0024047865025612e-05, "loss": 0.7233, "step": 14199 }, { "epoch": 0.5143064107207533, "grad_norm": 1.3579820862847016, "learning_rate": 1.0022874800537514e-05, "loss": 0.7177, "step": 14200 }, { "epoch": 0.5143426294820718, "grad_norm": 0.9858238452493695, "learning_rate": 1.0021701735734636e-05, "loss": 0.755, "step": 14201 }, { "epoch": 0.5143788482433901, "grad_norm": 0.9966482834799452, "learning_rate": 1.0020528670633127e-05, "loss": 0.7064, "step": 14202 }, { "epoch": 0.5144150670047084, "grad_norm": 1.3244782843332457, "learning_rate": 1.0019355605249123e-05, "loss": 0.7667, "step": 14203 }, { "epoch": 0.5144512857660268, "grad_norm": 1.2445188618107463, "learning_rate": 1.001818253959877e-05, "loss": 0.7385, "step": 14204 }, { "epoch": 0.5144875045273452, "grad_norm": 1.4245222294090756, "learning_rate": 1.0017009473698207e-05, "loss": 0.7907, "step": 14205 }, { "epoch": 0.5145237232886636, "grad_norm": 1.0836126087637628, "learning_rate": 1.0015836407563579e-05, "loss": 0.7747, "step": 14206 }, { "epoch": 0.5145599420499819, "grad_norm": 1.393712417775814, "learning_rate": 1.001466334121103e-05, "loss": 0.7101, "step": 14207 }, { "epoch": 0.5145961608113002, "grad_norm": 1.4747331055779525, "learning_rate": 1.00134902746567e-05, "loss": 0.7728, "step": 14208 }, { "epoch": 0.5146323795726186, "grad_norm": 1.3617927443895956, "learning_rate": 1.0012317207916733e-05, "loss": 0.8025, "step": 14209 }, { "epoch": 0.514668598333937, "grad_norm": 1.4484980730756898, "learning_rate": 1.0011144141007269e-05, "loss": 0.7176, "step": 14210 }, { "epoch": 0.5147048170952554, "grad_norm": 1.122764513551764, "learning_rate": 1.0009971073944453e-05, "loss": 0.7064, "step": 14211 }, { "epoch": 0.5147410358565737, "grad_norm": 0.9978706822315425, "learning_rate": 1.0008798006744425e-05, "loss": 0.7891, "step": 14212 }, { "epoch": 0.514777254617892, "grad_norm": 1.3992388456531988, "learning_rate": 1.0007624939423328e-05, "loss": 0.7953, "step": 14213 }, { "epoch": 0.5148134733792105, "grad_norm": 1.1268796470189437, "learning_rate": 1.0006451871997306e-05, "loss": 0.7817, "step": 14214 }, { "epoch": 0.5148496921405288, "grad_norm": 1.376797579373507, "learning_rate": 1.0005278804482502e-05, "loss": 0.7494, "step": 14215 }, { "epoch": 0.5148859109018472, "grad_norm": 1.5418850557728436, "learning_rate": 1.0004105736895056e-05, "loss": 0.6618, "step": 14216 }, { "epoch": 0.5149221296631655, "grad_norm": 1.3038540158000338, "learning_rate": 1.0002932669251109e-05, "loss": 0.7194, "step": 14217 }, { "epoch": 0.5149583484244838, "grad_norm": 1.4437591556572529, "learning_rate": 1.000175960156681e-05, "loss": 0.7234, "step": 14218 }, { "epoch": 0.5149945671858023, "grad_norm": 1.2807437133396025, "learning_rate": 1.0000586533858294e-05, "loss": 0.7206, "step": 14219 }, { "epoch": 0.5150307859471206, "grad_norm": 1.3145642106104436, "learning_rate": 9.999413466141711e-06, "loss": 0.8143, "step": 14220 }, { "epoch": 0.515067004708439, "grad_norm": 1.3470425715057537, "learning_rate": 9.998240398433195e-06, "loss": 0.7678, "step": 14221 }, { "epoch": 0.5151032234697573, "grad_norm": 1.2809335903795338, "learning_rate": 9.997067330748894e-06, "loss": 0.7509, "step": 14222 }, { "epoch": 0.5151394422310757, "grad_norm": 1.3751456193054685, "learning_rate": 9.995894263104951e-06, "loss": 0.7601, "step": 14223 }, { "epoch": 0.5151756609923941, "grad_norm": 2.394729954501143, "learning_rate": 9.994721195517503e-06, "loss": 0.7736, "step": 14224 }, { "epoch": 0.5152118797537124, "grad_norm": 1.6324151527574908, "learning_rate": 9.993548128002699e-06, "loss": 0.7484, "step": 14225 }, { "epoch": 0.5152480985150308, "grad_norm": 1.6726321150505308, "learning_rate": 9.992375060576675e-06, "loss": 0.7555, "step": 14226 }, { "epoch": 0.5152843172763492, "grad_norm": 1.4487456888878556, "learning_rate": 9.991201993255582e-06, "loss": 0.7661, "step": 14227 }, { "epoch": 0.5153205360376675, "grad_norm": 1.2199801096726273, "learning_rate": 9.990028926055552e-06, "loss": 0.6191, "step": 14228 }, { "epoch": 0.5153567547989859, "grad_norm": 1.5191267230181755, "learning_rate": 9.988855858992736e-06, "loss": 0.6818, "step": 14229 }, { "epoch": 0.5153929735603042, "grad_norm": 1.2467211964285982, "learning_rate": 9.987682792083272e-06, "loss": 0.6995, "step": 14230 }, { "epoch": 0.5154291923216227, "grad_norm": 1.3545203429460746, "learning_rate": 9.986509725343303e-06, "loss": 0.6918, "step": 14231 }, { "epoch": 0.515465411082941, "grad_norm": 1.3289575351594682, "learning_rate": 9.985336658788973e-06, "loss": 0.723, "step": 14232 }, { "epoch": 0.5155016298442593, "grad_norm": 1.5907912289728259, "learning_rate": 9.984163592436423e-06, "loss": 0.7302, "step": 14233 }, { "epoch": 0.5155378486055777, "grad_norm": 1.1684538170283705, "learning_rate": 9.982990526301798e-06, "loss": 0.726, "step": 14234 }, { "epoch": 0.515574067366896, "grad_norm": 1.3113675383439252, "learning_rate": 9.981817460401235e-06, "loss": 0.7266, "step": 14235 }, { "epoch": 0.5156102861282145, "grad_norm": 1.3994674912001044, "learning_rate": 9.980644394750882e-06, "loss": 0.7453, "step": 14236 }, { "epoch": 0.5156465048895328, "grad_norm": 1.5154398591233273, "learning_rate": 9.979471329366878e-06, "loss": 0.6959, "step": 14237 }, { "epoch": 0.5156827236508511, "grad_norm": 1.3759868687770984, "learning_rate": 9.978298264265367e-06, "loss": 0.7685, "step": 14238 }, { "epoch": 0.5157189424121695, "grad_norm": 1.4285671346889406, "learning_rate": 9.977125199462491e-06, "loss": 0.7578, "step": 14239 }, { "epoch": 0.5157551611734879, "grad_norm": 1.1086943074493933, "learning_rate": 9.975952134974393e-06, "loss": 0.7212, "step": 14240 }, { "epoch": 0.5157913799348063, "grad_norm": 1.0914692753497448, "learning_rate": 9.974779070817213e-06, "loss": 0.7538, "step": 14241 }, { "epoch": 0.5158275986961246, "grad_norm": 0.9218362663959414, "learning_rate": 9.973606007007096e-06, "loss": 0.7445, "step": 14242 }, { "epoch": 0.5158638174574429, "grad_norm": 1.8972444106446102, "learning_rate": 9.972432943560184e-06, "loss": 0.7372, "step": 14243 }, { "epoch": 0.5159000362187613, "grad_norm": 1.4345047554685288, "learning_rate": 9.971259880492617e-06, "loss": 0.7169, "step": 14244 }, { "epoch": 0.5159362549800797, "grad_norm": 1.3168155093417806, "learning_rate": 9.970086817820542e-06, "loss": 0.7159, "step": 14245 }, { "epoch": 0.5159724737413981, "grad_norm": 1.291187351062159, "learning_rate": 9.968913755560097e-06, "loss": 0.7153, "step": 14246 }, { "epoch": 0.5160086925027164, "grad_norm": 1.377234914366832, "learning_rate": 9.967740693727428e-06, "loss": 0.6915, "step": 14247 }, { "epoch": 0.5160449112640347, "grad_norm": 1.5556271734850862, "learning_rate": 9.966567632338672e-06, "loss": 0.8256, "step": 14248 }, { "epoch": 0.5160811300253532, "grad_norm": 1.2630398458019907, "learning_rate": 9.96539457140998e-06, "loss": 0.6993, "step": 14249 }, { "epoch": 0.5161173487866715, "grad_norm": 1.2144735379181424, "learning_rate": 9.964221510957483e-06, "loss": 0.7876, "step": 14250 }, { "epoch": 0.5161535675479899, "grad_norm": 1.3265078741178526, "learning_rate": 9.963048450997333e-06, "loss": 0.747, "step": 14251 }, { "epoch": 0.5161897863093082, "grad_norm": 1.1557336412217551, "learning_rate": 9.961875391545668e-06, "loss": 0.79, "step": 14252 }, { "epoch": 0.5162260050706265, "grad_norm": 1.2883362465414947, "learning_rate": 9.96070233261863e-06, "loss": 0.7657, "step": 14253 }, { "epoch": 0.516262223831945, "grad_norm": 1.1480418885444945, "learning_rate": 9.959529274232364e-06, "loss": 0.6228, "step": 14254 }, { "epoch": 0.5162984425932633, "grad_norm": 1.201680189056386, "learning_rate": 9.95835621640301e-06, "loss": 0.734, "step": 14255 }, { "epoch": 0.5163346613545817, "grad_norm": 1.171804077068572, "learning_rate": 9.957183159146713e-06, "loss": 0.6417, "step": 14256 }, { "epoch": 0.5163708801159, "grad_norm": 1.3169194026609672, "learning_rate": 9.956010102479611e-06, "loss": 0.7632, "step": 14257 }, { "epoch": 0.5164070988772184, "grad_norm": 1.2464736266765106, "learning_rate": 9.95483704641785e-06, "loss": 0.6758, "step": 14258 }, { "epoch": 0.5164433176385368, "grad_norm": 1.4482228746624823, "learning_rate": 9.953663990977568e-06, "loss": 0.705, "step": 14259 }, { "epoch": 0.5164795363998551, "grad_norm": 1.1349540549284403, "learning_rate": 9.952490936174914e-06, "loss": 0.7704, "step": 14260 }, { "epoch": 0.5165157551611735, "grad_norm": 1.3168051098430544, "learning_rate": 9.951317882026025e-06, "loss": 0.7496, "step": 14261 }, { "epoch": 0.5165519739224919, "grad_norm": 1.2481239344308648, "learning_rate": 9.950144828547043e-06, "loss": 0.6855, "step": 14262 }, { "epoch": 0.5165881926838102, "grad_norm": 1.2893057304715725, "learning_rate": 9.948971775754115e-06, "loss": 0.661, "step": 14263 }, { "epoch": 0.5166244114451286, "grad_norm": 1.1434664778258523, "learning_rate": 9.947798723663377e-06, "loss": 0.7583, "step": 14264 }, { "epoch": 0.5166606302064469, "grad_norm": 1.3559276328988343, "learning_rate": 9.94662567229098e-06, "loss": 0.7814, "step": 14265 }, { "epoch": 0.5166968489677654, "grad_norm": 1.3677623548416848, "learning_rate": 9.945452621653054e-06, "loss": 0.7062, "step": 14266 }, { "epoch": 0.5167330677290837, "grad_norm": 1.0096374054814776, "learning_rate": 9.944279571765752e-06, "loss": 0.7792, "step": 14267 }, { "epoch": 0.516769286490402, "grad_norm": 0.8992751572898464, "learning_rate": 9.94310652264521e-06, "loss": 0.6787, "step": 14268 }, { "epoch": 0.5168055052517204, "grad_norm": 1.366955508576527, "learning_rate": 9.941933474307575e-06, "loss": 0.7085, "step": 14269 }, { "epoch": 0.5168417240130387, "grad_norm": 1.3228907157074739, "learning_rate": 9.940760426768982e-06, "loss": 0.7533, "step": 14270 }, { "epoch": 0.5168779427743572, "grad_norm": 1.3248867161924947, "learning_rate": 9.939587380045582e-06, "loss": 0.6471, "step": 14271 }, { "epoch": 0.5169141615356755, "grad_norm": 1.3129556289363606, "learning_rate": 9.93841433415351e-06, "loss": 0.669, "step": 14272 }, { "epoch": 0.5169503802969938, "grad_norm": 1.4364172353586897, "learning_rate": 9.93724128910891e-06, "loss": 0.7135, "step": 14273 }, { "epoch": 0.5169865990583122, "grad_norm": 1.1130169335886042, "learning_rate": 9.936068244927929e-06, "loss": 0.688, "step": 14274 }, { "epoch": 0.5170228178196306, "grad_norm": 1.3026039803201155, "learning_rate": 9.934895201626701e-06, "loss": 0.7335, "step": 14275 }, { "epoch": 0.517059036580949, "grad_norm": 1.3884040531334338, "learning_rate": 9.933722159221375e-06, "loss": 0.7885, "step": 14276 }, { "epoch": 0.5170952553422673, "grad_norm": 1.3296613243240827, "learning_rate": 9.93254911772809e-06, "loss": 0.8057, "step": 14277 }, { "epoch": 0.5171314741035856, "grad_norm": 1.333738652772505, "learning_rate": 9.93137607716299e-06, "loss": 0.7518, "step": 14278 }, { "epoch": 0.517167692864904, "grad_norm": 1.4745914227611816, "learning_rate": 9.930203037542211e-06, "loss": 0.7929, "step": 14279 }, { "epoch": 0.5172039116262224, "grad_norm": 1.5777883111490147, "learning_rate": 9.929029998881903e-06, "loss": 0.7042, "step": 14280 }, { "epoch": 0.5172401303875408, "grad_norm": 1.459256448744835, "learning_rate": 9.927856961198203e-06, "loss": 0.6973, "step": 14281 }, { "epoch": 0.5172763491488591, "grad_norm": 0.9786157851224736, "learning_rate": 9.926683924507256e-06, "loss": 0.7138, "step": 14282 }, { "epoch": 0.5173125679101774, "grad_norm": 1.4093745183520965, "learning_rate": 9.9255108888252e-06, "loss": 0.8339, "step": 14283 }, { "epoch": 0.5173487866714959, "grad_norm": 1.4469830067170353, "learning_rate": 9.92433785416818e-06, "loss": 0.7429, "step": 14284 }, { "epoch": 0.5173850054328142, "grad_norm": 1.3664127402590405, "learning_rate": 9.923164820552341e-06, "loss": 0.7386, "step": 14285 }, { "epoch": 0.5174212241941326, "grad_norm": 1.4432759325378213, "learning_rate": 9.921991787993817e-06, "loss": 0.7692, "step": 14286 }, { "epoch": 0.5174574429554509, "grad_norm": 1.0934380650610853, "learning_rate": 9.920818756508758e-06, "loss": 0.6781, "step": 14287 }, { "epoch": 0.5174936617167692, "grad_norm": 1.387740812287242, "learning_rate": 9.9196457261133e-06, "loss": 0.6833, "step": 14288 }, { "epoch": 0.5175298804780877, "grad_norm": 1.251655904121733, "learning_rate": 9.91847269682359e-06, "loss": 0.7137, "step": 14289 }, { "epoch": 0.517566099239406, "grad_norm": 1.1043819133494968, "learning_rate": 9.917299668655767e-06, "loss": 0.6945, "step": 14290 }, { "epoch": 0.5176023180007244, "grad_norm": 1.1925838326250184, "learning_rate": 9.916126641625973e-06, "loss": 0.7752, "step": 14291 }, { "epoch": 0.5176385367620427, "grad_norm": 1.8837895450225577, "learning_rate": 9.914953615750349e-06, "loss": 0.7842, "step": 14292 }, { "epoch": 0.5176747555233611, "grad_norm": 1.3217522034299323, "learning_rate": 9.913780591045037e-06, "loss": 0.7133, "step": 14293 }, { "epoch": 0.5177109742846795, "grad_norm": 1.4893951677861064, "learning_rate": 9.912607567526183e-06, "loss": 0.7609, "step": 14294 }, { "epoch": 0.5177471930459978, "grad_norm": 1.3791427301459835, "learning_rate": 9.911434545209922e-06, "loss": 0.7794, "step": 14295 }, { "epoch": 0.5177834118073162, "grad_norm": 1.4570169064631384, "learning_rate": 9.910261524112404e-06, "loss": 0.7409, "step": 14296 }, { "epoch": 0.5178196305686346, "grad_norm": 1.4522537008587397, "learning_rate": 9.909088504249762e-06, "loss": 0.6781, "step": 14297 }, { "epoch": 0.5178558493299529, "grad_norm": 1.2214194683119999, "learning_rate": 9.907915485638144e-06, "loss": 0.7117, "step": 14298 }, { "epoch": 0.5178920680912713, "grad_norm": 1.1262136337567796, "learning_rate": 9.906742468293689e-06, "loss": 0.6912, "step": 14299 }, { "epoch": 0.5179282868525896, "grad_norm": 1.721729794972359, "learning_rate": 9.905569452232541e-06, "loss": 0.7159, "step": 14300 }, { "epoch": 0.517964505613908, "grad_norm": 1.356263256282889, "learning_rate": 9.904396437470839e-06, "loss": 0.7935, "step": 14301 }, { "epoch": 0.5180007243752264, "grad_norm": 1.2711476921799547, "learning_rate": 9.903223424024729e-06, "loss": 0.7272, "step": 14302 }, { "epoch": 0.5180369431365447, "grad_norm": 1.2895538903472985, "learning_rate": 9.902050411910346e-06, "loss": 0.7601, "step": 14303 }, { "epoch": 0.5180731618978631, "grad_norm": 1.417282643904398, "learning_rate": 9.900877401143834e-06, "loss": 0.7823, "step": 14304 }, { "epoch": 0.5181093806591814, "grad_norm": 1.4306755208673, "learning_rate": 9.89970439174134e-06, "loss": 0.749, "step": 14305 }, { "epoch": 0.5181455994204999, "grad_norm": 1.4076884608288958, "learning_rate": 9.898531383719e-06, "loss": 0.7939, "step": 14306 }, { "epoch": 0.5181818181818182, "grad_norm": 1.05163855062454, "learning_rate": 9.897358377092959e-06, "loss": 0.7518, "step": 14307 }, { "epoch": 0.5182180369431365, "grad_norm": 1.3745981626278272, "learning_rate": 9.896185371879354e-06, "loss": 0.7624, "step": 14308 }, { "epoch": 0.5182542557044549, "grad_norm": 1.3179249952104934, "learning_rate": 9.895012368094332e-06, "loss": 0.688, "step": 14309 }, { "epoch": 0.5182904744657733, "grad_norm": 1.3361166114533096, "learning_rate": 9.893839365754031e-06, "loss": 0.7103, "step": 14310 }, { "epoch": 0.5183266932270917, "grad_norm": 1.425943971519323, "learning_rate": 9.892666364874595e-06, "loss": 0.7594, "step": 14311 }, { "epoch": 0.51836291198841, "grad_norm": 1.091422307833968, "learning_rate": 9.891493365472163e-06, "loss": 0.7475, "step": 14312 }, { "epoch": 0.5183991307497283, "grad_norm": 1.2329670233878172, "learning_rate": 9.890320367562879e-06, "loss": 0.7001, "step": 14313 }, { "epoch": 0.5184353495110468, "grad_norm": 1.212169730524453, "learning_rate": 9.88914737116288e-06, "loss": 0.7265, "step": 14314 }, { "epoch": 0.5184715682723651, "grad_norm": 1.4337848437223424, "learning_rate": 9.887974376288312e-06, "loss": 0.7541, "step": 14315 }, { "epoch": 0.5185077870336835, "grad_norm": 1.39922960187707, "learning_rate": 9.886801382955316e-06, "loss": 0.7424, "step": 14316 }, { "epoch": 0.5185440057950018, "grad_norm": 1.3687133938758533, "learning_rate": 9.885628391180031e-06, "loss": 0.7077, "step": 14317 }, { "epoch": 0.5185802245563201, "grad_norm": 1.2469363853873028, "learning_rate": 9.884455400978601e-06, "loss": 0.7604, "step": 14318 }, { "epoch": 0.5186164433176386, "grad_norm": 1.044346236633548, "learning_rate": 9.883282412367165e-06, "loss": 0.7084, "step": 14319 }, { "epoch": 0.5186526620789569, "grad_norm": 1.3265450785361244, "learning_rate": 9.882109425361868e-06, "loss": 0.6934, "step": 14320 }, { "epoch": 0.5186888808402753, "grad_norm": 1.0288150278631634, "learning_rate": 9.880936439978846e-06, "loss": 0.7305, "step": 14321 }, { "epoch": 0.5187250996015936, "grad_norm": 1.4284810215631807, "learning_rate": 9.879763456234246e-06, "loss": 0.6669, "step": 14322 }, { "epoch": 0.518761318362912, "grad_norm": 1.3864154262307837, "learning_rate": 9.878590474144204e-06, "loss": 0.7351, "step": 14323 }, { "epoch": 0.5187975371242304, "grad_norm": 1.4987927697361116, "learning_rate": 9.877417493724864e-06, "loss": 0.6906, "step": 14324 }, { "epoch": 0.5188337558855487, "grad_norm": 1.2609947809931266, "learning_rate": 9.876244514992368e-06, "loss": 0.6956, "step": 14325 }, { "epoch": 0.5188699746468671, "grad_norm": 1.6724922553679282, "learning_rate": 9.875071537962855e-06, "loss": 0.7419, "step": 14326 }, { "epoch": 0.5189061934081854, "grad_norm": 1.3273334507545431, "learning_rate": 9.87389856265247e-06, "loss": 0.7455, "step": 14327 }, { "epoch": 0.5189424121695038, "grad_norm": 1.348013194124081, "learning_rate": 9.872725589077348e-06, "loss": 0.7708, "step": 14328 }, { "epoch": 0.5189786309308222, "grad_norm": 1.0223603089774274, "learning_rate": 9.871552617253636e-06, "loss": 0.69, "step": 14329 }, { "epoch": 0.5190148496921405, "grad_norm": 1.200597802754836, "learning_rate": 9.870379647197472e-06, "loss": 0.7387, "step": 14330 }, { "epoch": 0.5190510684534589, "grad_norm": 1.3027405817321522, "learning_rate": 9.869206678925e-06, "loss": 0.7619, "step": 14331 }, { "epoch": 0.5190872872147773, "grad_norm": 1.3220126101547343, "learning_rate": 9.868033712452356e-06, "loss": 0.7044, "step": 14332 }, { "epoch": 0.5191235059760956, "grad_norm": 1.3265237439743087, "learning_rate": 9.866860747795686e-06, "loss": 0.7647, "step": 14333 }, { "epoch": 0.519159724737414, "grad_norm": 1.253996090478056, "learning_rate": 9.865687784971128e-06, "loss": 0.651, "step": 14334 }, { "epoch": 0.5191959434987323, "grad_norm": 1.300443715758967, "learning_rate": 9.864514823994824e-06, "loss": 0.6908, "step": 14335 }, { "epoch": 0.5192321622600508, "grad_norm": 1.1289427782393215, "learning_rate": 9.863341864882918e-06, "loss": 0.7688, "step": 14336 }, { "epoch": 0.5192683810213691, "grad_norm": 1.354314311091056, "learning_rate": 9.862168907651546e-06, "loss": 0.6942, "step": 14337 }, { "epoch": 0.5193045997826874, "grad_norm": 1.3157543668094633, "learning_rate": 9.860995952316851e-06, "loss": 0.6742, "step": 14338 }, { "epoch": 0.5193408185440058, "grad_norm": 1.3913004539381766, "learning_rate": 9.859822998894972e-06, "loss": 0.8021, "step": 14339 }, { "epoch": 0.5193770373053241, "grad_norm": 1.008103117318327, "learning_rate": 9.858650047402053e-06, "loss": 0.7291, "step": 14340 }, { "epoch": 0.5194132560666426, "grad_norm": 1.2945084242629965, "learning_rate": 9.857477097854234e-06, "loss": 0.7371, "step": 14341 }, { "epoch": 0.5194494748279609, "grad_norm": 1.4285342791077098, "learning_rate": 9.856304150267656e-06, "loss": 0.7289, "step": 14342 }, { "epoch": 0.5194856935892792, "grad_norm": 0.9463478304243192, "learning_rate": 9.855131204658457e-06, "loss": 0.6693, "step": 14343 }, { "epoch": 0.5195219123505976, "grad_norm": 1.4615900385168576, "learning_rate": 9.853958261042784e-06, "loss": 0.6711, "step": 14344 }, { "epoch": 0.519558131111916, "grad_norm": 1.2734268201192938, "learning_rate": 9.852785319436769e-06, "loss": 0.7506, "step": 14345 }, { "epoch": 0.5195943498732344, "grad_norm": 1.3654853047189006, "learning_rate": 9.851612379856559e-06, "loss": 0.7493, "step": 14346 }, { "epoch": 0.5196305686345527, "grad_norm": 1.389218548048642, "learning_rate": 9.850439442318294e-06, "loss": 0.7565, "step": 14347 }, { "epoch": 0.519666787395871, "grad_norm": 1.308306357154089, "learning_rate": 9.849266506838113e-06, "loss": 0.7648, "step": 14348 }, { "epoch": 0.5197030061571895, "grad_norm": 1.2324113739581934, "learning_rate": 9.848093573432158e-06, "loss": 0.6769, "step": 14349 }, { "epoch": 0.5197392249185078, "grad_norm": 1.3643675864420604, "learning_rate": 9.846920642116567e-06, "loss": 0.7556, "step": 14350 }, { "epoch": 0.5197754436798262, "grad_norm": 1.121915389903889, "learning_rate": 9.845747712907484e-06, "loss": 0.7579, "step": 14351 }, { "epoch": 0.5198116624411445, "grad_norm": 1.3100403017721307, "learning_rate": 9.844574785821047e-06, "loss": 0.7693, "step": 14352 }, { "epoch": 0.5198478812024628, "grad_norm": 1.4609163124669635, "learning_rate": 9.8434018608734e-06, "loss": 0.8192, "step": 14353 }, { "epoch": 0.5198840999637813, "grad_norm": 1.315470042809738, "learning_rate": 9.842228938080678e-06, "loss": 0.7947, "step": 14354 }, { "epoch": 0.5199203187250996, "grad_norm": 1.319584149011196, "learning_rate": 9.841056017459024e-06, "loss": 0.7064, "step": 14355 }, { "epoch": 0.519956537486418, "grad_norm": 1.3170063189173096, "learning_rate": 9.839883099024581e-06, "loss": 0.6861, "step": 14356 }, { "epoch": 0.5199927562477363, "grad_norm": 1.241776613965361, "learning_rate": 9.838710182793488e-06, "loss": 0.6925, "step": 14357 }, { "epoch": 0.5200289750090547, "grad_norm": 1.0399611085681597, "learning_rate": 9.837537268781884e-06, "loss": 0.71, "step": 14358 }, { "epoch": 0.5200651937703731, "grad_norm": 1.337103245141892, "learning_rate": 9.836364357005908e-06, "loss": 0.6965, "step": 14359 }, { "epoch": 0.5201014125316914, "grad_norm": 1.4046460673848662, "learning_rate": 9.835191447481705e-06, "loss": 0.7303, "step": 14360 }, { "epoch": 0.5201376312930098, "grad_norm": 1.0888876322892482, "learning_rate": 9.83401854022541e-06, "loss": 0.7161, "step": 14361 }, { "epoch": 0.5201738500543281, "grad_norm": 1.3287338467100118, "learning_rate": 9.832845635253167e-06, "loss": 0.7103, "step": 14362 }, { "epoch": 0.5202100688156465, "grad_norm": 1.0219033767302927, "learning_rate": 9.831672732581115e-06, "loss": 0.7221, "step": 14363 }, { "epoch": 0.5202462875769649, "grad_norm": 1.406848746891113, "learning_rate": 9.830499832225394e-06, "loss": 0.7343, "step": 14364 }, { "epoch": 0.5202825063382832, "grad_norm": 1.242569286677165, "learning_rate": 9.829326934202143e-06, "loss": 0.7621, "step": 14365 }, { "epoch": 0.5203187250996016, "grad_norm": 1.3265887905690064, "learning_rate": 9.828154038527503e-06, "loss": 0.7528, "step": 14366 }, { "epoch": 0.52035494386092, "grad_norm": 1.4433247041990351, "learning_rate": 9.826981145217618e-06, "loss": 0.7162, "step": 14367 }, { "epoch": 0.5203911626222383, "grad_norm": 1.0795307678650898, "learning_rate": 9.825808254288622e-06, "loss": 0.7192, "step": 14368 }, { "epoch": 0.5204273813835567, "grad_norm": 1.3962351489920803, "learning_rate": 9.824635365756658e-06, "loss": 0.8008, "step": 14369 }, { "epoch": 0.520463600144875, "grad_norm": 1.326594585220519, "learning_rate": 9.823462479637864e-06, "loss": 0.7043, "step": 14370 }, { "epoch": 0.5204998189061935, "grad_norm": 1.8083355639007395, "learning_rate": 9.822289595948384e-06, "loss": 0.7253, "step": 14371 }, { "epoch": 0.5205360376675118, "grad_norm": 1.0840370717964818, "learning_rate": 9.821116714704353e-06, "loss": 0.7224, "step": 14372 }, { "epoch": 0.5205722564288301, "grad_norm": 1.4421803184425528, "learning_rate": 9.819943835921916e-06, "loss": 0.7453, "step": 14373 }, { "epoch": 0.5206084751901485, "grad_norm": 1.3514215858957463, "learning_rate": 9.818770959617207e-06, "loss": 0.7276, "step": 14374 }, { "epoch": 0.5206446939514668, "grad_norm": 1.6847217563611099, "learning_rate": 9.81759808580637e-06, "loss": 0.6974, "step": 14375 }, { "epoch": 0.5206809127127853, "grad_norm": 1.3645880974499576, "learning_rate": 9.816425214505542e-06, "loss": 0.6627, "step": 14376 }, { "epoch": 0.5207171314741036, "grad_norm": 1.4101482771545057, "learning_rate": 9.815252345730866e-06, "loss": 0.7293, "step": 14377 }, { "epoch": 0.5207533502354219, "grad_norm": 1.5091792111441087, "learning_rate": 9.814079479498481e-06, "loss": 0.8375, "step": 14378 }, { "epoch": 0.5207895689967403, "grad_norm": 1.4092892797817254, "learning_rate": 9.812906615824524e-06, "loss": 0.8136, "step": 14379 }, { "epoch": 0.5208257877580587, "grad_norm": 1.0438169738037955, "learning_rate": 9.811733754725138e-06, "loss": 0.7264, "step": 14380 }, { "epoch": 0.5208620065193771, "grad_norm": 1.2721931471426395, "learning_rate": 9.810560896216459e-06, "loss": 0.6886, "step": 14381 }, { "epoch": 0.5208982252806954, "grad_norm": 1.2128031018754781, "learning_rate": 9.809388040314632e-06, "loss": 0.8024, "step": 14382 }, { "epoch": 0.5209344440420137, "grad_norm": 1.377907710910575, "learning_rate": 9.80821518703579e-06, "loss": 0.8081, "step": 14383 }, { "epoch": 0.5209706628033322, "grad_norm": 1.261473416497117, "learning_rate": 9.807042336396077e-06, "loss": 0.7708, "step": 14384 }, { "epoch": 0.5210068815646505, "grad_norm": 1.1085654292053646, "learning_rate": 9.805869488411627e-06, "loss": 0.7158, "step": 14385 }, { "epoch": 0.5210431003259689, "grad_norm": 8.547519889209672, "learning_rate": 9.804696643098586e-06, "loss": 0.6753, "step": 14386 }, { "epoch": 0.5210793190872872, "grad_norm": 1.3399682716972041, "learning_rate": 9.803523800473094e-06, "loss": 0.8017, "step": 14387 }, { "epoch": 0.5211155378486055, "grad_norm": 1.2529381108437547, "learning_rate": 9.802350960551284e-06, "loss": 0.6811, "step": 14388 }, { "epoch": 0.521151756609924, "grad_norm": 1.2722193597691998, "learning_rate": 9.801178123349298e-06, "loss": 0.7025, "step": 14389 }, { "epoch": 0.5211879753712423, "grad_norm": 1.3411556388595287, "learning_rate": 9.800005288883277e-06, "loss": 0.7678, "step": 14390 }, { "epoch": 0.5212241941325607, "grad_norm": 1.3974326775317194, "learning_rate": 9.79883245716936e-06, "loss": 0.7389, "step": 14391 }, { "epoch": 0.521260412893879, "grad_norm": 1.2431987782326048, "learning_rate": 9.797659628223682e-06, "loss": 0.7555, "step": 14392 }, { "epoch": 0.5212966316551974, "grad_norm": 1.3200046777224652, "learning_rate": 9.796486802062388e-06, "loss": 0.7311, "step": 14393 }, { "epoch": 0.5213328504165158, "grad_norm": 1.4249699570600871, "learning_rate": 9.795313978701613e-06, "loss": 0.7838, "step": 14394 }, { "epoch": 0.5213690691778341, "grad_norm": 1.3133837557199741, "learning_rate": 9.7941411581575e-06, "loss": 0.7385, "step": 14395 }, { "epoch": 0.5214052879391525, "grad_norm": 1.2822663053351444, "learning_rate": 9.792968340446182e-06, "loss": 0.7385, "step": 14396 }, { "epoch": 0.5214415067004708, "grad_norm": 1.4840083459965014, "learning_rate": 9.791795525583803e-06, "loss": 0.7849, "step": 14397 }, { "epoch": 0.5214777254617892, "grad_norm": 1.2645127227309267, "learning_rate": 9.7906227135865e-06, "loss": 0.642, "step": 14398 }, { "epoch": 0.5215139442231076, "grad_norm": 1.289761351668945, "learning_rate": 9.789449904470413e-06, "loss": 0.7317, "step": 14399 }, { "epoch": 0.5215501629844259, "grad_norm": 1.0783352605805874, "learning_rate": 9.788277098251681e-06, "loss": 0.7612, "step": 14400 }, { "epoch": 0.5215863817457443, "grad_norm": 1.3438526200321397, "learning_rate": 9.787104294946441e-06, "loss": 0.6589, "step": 14401 }, { "epoch": 0.5216226005070627, "grad_norm": 1.2886153283595772, "learning_rate": 9.785931494570836e-06, "loss": 0.7587, "step": 14402 }, { "epoch": 0.521658819268381, "grad_norm": 1.430987333125155, "learning_rate": 9.784758697140997e-06, "loss": 0.7332, "step": 14403 }, { "epoch": 0.5216950380296994, "grad_norm": 1.370573131096898, "learning_rate": 9.783585902673072e-06, "loss": 0.6732, "step": 14404 }, { "epoch": 0.5217312567910177, "grad_norm": 1.4377239774539339, "learning_rate": 9.782413111183192e-06, "loss": 0.7155, "step": 14405 }, { "epoch": 0.5217674755523362, "grad_norm": 1.4280985356055378, "learning_rate": 9.7812403226875e-06, "loss": 0.7003, "step": 14406 }, { "epoch": 0.5218036943136545, "grad_norm": 1.3289850500640044, "learning_rate": 9.780067537202132e-06, "loss": 0.6112, "step": 14407 }, { "epoch": 0.5218399130749728, "grad_norm": 1.329179280129911, "learning_rate": 9.778894754743228e-06, "loss": 0.7408, "step": 14408 }, { "epoch": 0.5218761318362912, "grad_norm": 1.5069909260191623, "learning_rate": 9.777721975326928e-06, "loss": 0.7636, "step": 14409 }, { "epoch": 0.5219123505976095, "grad_norm": 1.567871255319231, "learning_rate": 9.776549198969369e-06, "loss": 0.7892, "step": 14410 }, { "epoch": 0.521948569358928, "grad_norm": 1.1459849768584165, "learning_rate": 9.77537642568669e-06, "loss": 0.7072, "step": 14411 }, { "epoch": 0.5219847881202463, "grad_norm": 1.01217976746528, "learning_rate": 9.774203655495025e-06, "loss": 0.7217, "step": 14412 }, { "epoch": 0.5220210068815646, "grad_norm": 1.5890967737194504, "learning_rate": 9.77303088841052e-06, "loss": 0.7918, "step": 14413 }, { "epoch": 0.522057225642883, "grad_norm": 0.9777473506236036, "learning_rate": 9.771858124449307e-06, "loss": 0.6672, "step": 14414 }, { "epoch": 0.5220934444042014, "grad_norm": 1.483700295248868, "learning_rate": 9.77068536362753e-06, "loss": 0.7769, "step": 14415 }, { "epoch": 0.5221296631655198, "grad_norm": 1.408703338202283, "learning_rate": 9.76951260596132e-06, "loss": 0.7422, "step": 14416 }, { "epoch": 0.5221658819268381, "grad_norm": 1.563005049858963, "learning_rate": 9.768339851466818e-06, "loss": 0.7827, "step": 14417 }, { "epoch": 0.5222021006881564, "grad_norm": 0.947315638490164, "learning_rate": 9.767167100160167e-06, "loss": 0.6824, "step": 14418 }, { "epoch": 0.5222383194494749, "grad_norm": 1.0282750897503052, "learning_rate": 9.7659943520575e-06, "loss": 0.7166, "step": 14419 }, { "epoch": 0.5222745382107932, "grad_norm": 1.414271453177368, "learning_rate": 9.764821607174957e-06, "loss": 0.7965, "step": 14420 }, { "epoch": 0.5223107569721116, "grad_norm": 1.2292345378430805, "learning_rate": 9.763648865528674e-06, "loss": 0.6904, "step": 14421 }, { "epoch": 0.5223469757334299, "grad_norm": 1.368904151847406, "learning_rate": 9.762476127134793e-06, "loss": 0.7657, "step": 14422 }, { "epoch": 0.5223831944947482, "grad_norm": 1.3494698152827878, "learning_rate": 9.761303392009446e-06, "loss": 0.7827, "step": 14423 }, { "epoch": 0.5224194132560667, "grad_norm": 1.3604954748167366, "learning_rate": 9.760130660168777e-06, "loss": 0.7345, "step": 14424 }, { "epoch": 0.522455632017385, "grad_norm": 1.3994389190382972, "learning_rate": 9.758957931628918e-06, "loss": 0.7105, "step": 14425 }, { "epoch": 0.5224918507787034, "grad_norm": 1.2542450267455254, "learning_rate": 9.757785206406012e-06, "loss": 0.7113, "step": 14426 }, { "epoch": 0.5225280695400217, "grad_norm": 1.342902246712659, "learning_rate": 9.756612484516192e-06, "loss": 0.7396, "step": 14427 }, { "epoch": 0.52256428830134, "grad_norm": 1.4302572798242785, "learning_rate": 9.755439765975599e-06, "loss": 0.6988, "step": 14428 }, { "epoch": 0.5226005070626585, "grad_norm": 1.4433846116602003, "learning_rate": 9.75426705080037e-06, "loss": 0.7381, "step": 14429 }, { "epoch": 0.5226367258239768, "grad_norm": 1.2599442007963448, "learning_rate": 9.753094339006641e-06, "loss": 0.7028, "step": 14430 }, { "epoch": 0.5226729445852952, "grad_norm": 1.3497254243081436, "learning_rate": 9.751921630610557e-06, "loss": 0.701, "step": 14431 }, { "epoch": 0.5227091633466135, "grad_norm": 1.045817272748374, "learning_rate": 9.750748925628244e-06, "loss": 0.6655, "step": 14432 }, { "epoch": 0.5227453821079319, "grad_norm": 1.4533693706730781, "learning_rate": 9.749576224075846e-06, "loss": 0.7519, "step": 14433 }, { "epoch": 0.5227816008692503, "grad_norm": 1.3202795061243813, "learning_rate": 9.7484035259695e-06, "loss": 0.7068, "step": 14434 }, { "epoch": 0.5228178196305686, "grad_norm": 1.0039950109597386, "learning_rate": 9.747230831325339e-06, "loss": 0.7219, "step": 14435 }, { "epoch": 0.522854038391887, "grad_norm": 1.067991059251395, "learning_rate": 9.746058140159507e-06, "loss": 0.708, "step": 14436 }, { "epoch": 0.5228902571532054, "grad_norm": 1.2658290653451463, "learning_rate": 9.744885452488137e-06, "loss": 0.7474, "step": 14437 }, { "epoch": 0.5229264759145237, "grad_norm": 1.5112460836611408, "learning_rate": 9.743712768327368e-06, "loss": 0.7874, "step": 14438 }, { "epoch": 0.5229626946758421, "grad_norm": 1.4318702371808216, "learning_rate": 9.742540087693334e-06, "loss": 0.6907, "step": 14439 }, { "epoch": 0.5229989134371604, "grad_norm": 1.362013190198861, "learning_rate": 9.741367410602177e-06, "loss": 0.7113, "step": 14440 }, { "epoch": 0.5230351321984789, "grad_norm": 0.9813625589042619, "learning_rate": 9.740194737070031e-06, "loss": 0.7103, "step": 14441 }, { "epoch": 0.5230713509597972, "grad_norm": 1.163017419867568, "learning_rate": 9.739022067113032e-06, "loss": 0.7319, "step": 14442 }, { "epoch": 0.5231075697211155, "grad_norm": 1.3984533394284129, "learning_rate": 9.737849400747324e-06, "loss": 0.7346, "step": 14443 }, { "epoch": 0.5231437884824339, "grad_norm": 1.133338963326838, "learning_rate": 9.736676737989033e-06, "loss": 0.7154, "step": 14444 }, { "epoch": 0.5231800072437522, "grad_norm": 1.4400010560656027, "learning_rate": 9.735504078854305e-06, "loss": 0.7023, "step": 14445 }, { "epoch": 0.5232162260050707, "grad_norm": 1.8546638754556846, "learning_rate": 9.734331423359272e-06, "loss": 0.7176, "step": 14446 }, { "epoch": 0.523252444766389, "grad_norm": 1.2844476072238014, "learning_rate": 9.733158771520072e-06, "loss": 0.7311, "step": 14447 }, { "epoch": 0.5232886635277073, "grad_norm": 1.33350687640708, "learning_rate": 9.731986123352842e-06, "loss": 0.807, "step": 14448 }, { "epoch": 0.5233248822890257, "grad_norm": 1.273772877135839, "learning_rate": 9.730813478873716e-06, "loss": 0.6652, "step": 14449 }, { "epoch": 0.5233611010503441, "grad_norm": 1.0175661057267602, "learning_rate": 9.729640838098835e-06, "loss": 0.7931, "step": 14450 }, { "epoch": 0.5233973198116625, "grad_norm": 1.268263842959177, "learning_rate": 9.728468201044335e-06, "loss": 0.6636, "step": 14451 }, { "epoch": 0.5234335385729808, "grad_norm": 1.4494975713789782, "learning_rate": 9.72729556772635e-06, "loss": 0.7789, "step": 14452 }, { "epoch": 0.5234697573342991, "grad_norm": 1.3266395732924217, "learning_rate": 9.726122938161017e-06, "loss": 0.6665, "step": 14453 }, { "epoch": 0.5235059760956176, "grad_norm": 1.2793765487321425, "learning_rate": 9.724950312364474e-06, "loss": 0.6785, "step": 14454 }, { "epoch": 0.5235421948569359, "grad_norm": 1.3282020483387882, "learning_rate": 9.723777690352855e-06, "loss": 0.7697, "step": 14455 }, { "epoch": 0.5235784136182543, "grad_norm": 1.026234893979321, "learning_rate": 9.7226050721423e-06, "loss": 0.6776, "step": 14456 }, { "epoch": 0.5236146323795726, "grad_norm": 1.2228280409881396, "learning_rate": 9.721432457748942e-06, "loss": 0.6732, "step": 14457 }, { "epoch": 0.5236508511408909, "grad_norm": 1.38265095322084, "learning_rate": 9.720259847188917e-06, "loss": 0.7299, "step": 14458 }, { "epoch": 0.5236870699022094, "grad_norm": 1.3450401892017678, "learning_rate": 9.719087240478364e-06, "loss": 0.7705, "step": 14459 }, { "epoch": 0.5237232886635277, "grad_norm": 1.3245420912165853, "learning_rate": 9.717914637633415e-06, "loss": 0.7144, "step": 14460 }, { "epoch": 0.5237595074248461, "grad_norm": 1.5286182417862495, "learning_rate": 9.716742038670211e-06, "loss": 0.7233, "step": 14461 }, { "epoch": 0.5237957261861644, "grad_norm": 1.503169870301675, "learning_rate": 9.715569443604882e-06, "loss": 0.7886, "step": 14462 }, { "epoch": 0.5238319449474828, "grad_norm": 1.2677658552596278, "learning_rate": 9.714396852453572e-06, "loss": 0.6954, "step": 14463 }, { "epoch": 0.5238681637088012, "grad_norm": 1.1908617925278404, "learning_rate": 9.713224265232409e-06, "loss": 0.7447, "step": 14464 }, { "epoch": 0.5239043824701195, "grad_norm": 1.290813792990802, "learning_rate": 9.712051681957534e-06, "loss": 0.766, "step": 14465 }, { "epoch": 0.5239406012314379, "grad_norm": 1.6013479786928266, "learning_rate": 9.71087910264508e-06, "loss": 0.7078, "step": 14466 }, { "epoch": 0.5239768199927562, "grad_norm": 1.3493436311580265, "learning_rate": 9.709706527311184e-06, "loss": 0.7403, "step": 14467 }, { "epoch": 0.5240130387540746, "grad_norm": 1.3359890741019107, "learning_rate": 9.708533955971981e-06, "loss": 0.6572, "step": 14468 }, { "epoch": 0.524049257515393, "grad_norm": 1.4569581974830428, "learning_rate": 9.707361388643607e-06, "loss": 0.6665, "step": 14469 }, { "epoch": 0.5240854762767113, "grad_norm": 1.319863027749323, "learning_rate": 9.706188825342196e-06, "loss": 0.6983, "step": 14470 }, { "epoch": 0.5241216950380297, "grad_norm": 1.226357713235345, "learning_rate": 9.705016266083886e-06, "loss": 0.7377, "step": 14471 }, { "epoch": 0.5241579137993481, "grad_norm": 1.3998376015047926, "learning_rate": 9.703843710884811e-06, "loss": 0.737, "step": 14472 }, { "epoch": 0.5241941325606664, "grad_norm": 1.2584149394358775, "learning_rate": 9.702671159761107e-06, "loss": 0.7201, "step": 14473 }, { "epoch": 0.5242303513219848, "grad_norm": 1.417516460218009, "learning_rate": 9.70149861272891e-06, "loss": 0.7734, "step": 14474 }, { "epoch": 0.5242665700833031, "grad_norm": 1.377011818683237, "learning_rate": 9.700326069804353e-06, "loss": 0.8002, "step": 14475 }, { "epoch": 0.5243027888446216, "grad_norm": 1.0542535177903551, "learning_rate": 9.699153531003574e-06, "loss": 0.7306, "step": 14476 }, { "epoch": 0.5243390076059399, "grad_norm": 1.0469101863879853, "learning_rate": 9.697980996342706e-06, "loss": 0.7042, "step": 14477 }, { "epoch": 0.5243752263672582, "grad_norm": 1.3415854820579507, "learning_rate": 9.696808465837886e-06, "loss": 0.7617, "step": 14478 }, { "epoch": 0.5244114451285766, "grad_norm": 1.2835516311215331, "learning_rate": 9.695635939505245e-06, "loss": 0.7177, "step": 14479 }, { "epoch": 0.5244476638898949, "grad_norm": 1.4697182689952362, "learning_rate": 9.694463417360921e-06, "loss": 0.7979, "step": 14480 }, { "epoch": 0.5244838826512134, "grad_norm": 1.490145379405107, "learning_rate": 9.693290899421053e-06, "loss": 0.717, "step": 14481 }, { "epoch": 0.5245201014125317, "grad_norm": 1.42421223648892, "learning_rate": 9.692118385701768e-06, "loss": 0.7005, "step": 14482 }, { "epoch": 0.52455632017385, "grad_norm": 1.302806601316559, "learning_rate": 9.690945876219206e-06, "loss": 0.752, "step": 14483 }, { "epoch": 0.5245925389351684, "grad_norm": 1.4043877257881243, "learning_rate": 9.6897733709895e-06, "loss": 0.7799, "step": 14484 }, { "epoch": 0.5246287576964868, "grad_norm": 1.315482109514902, "learning_rate": 9.688600870028786e-06, "loss": 0.7185, "step": 14485 }, { "epoch": 0.5246649764578052, "grad_norm": 1.4158539719224084, "learning_rate": 9.687428373353197e-06, "loss": 0.7657, "step": 14486 }, { "epoch": 0.5247011952191235, "grad_norm": 1.3646785514646036, "learning_rate": 9.68625588097887e-06, "loss": 0.7138, "step": 14487 }, { "epoch": 0.5247374139804418, "grad_norm": 1.387734358236206, "learning_rate": 9.685083392921933e-06, "loss": 0.7171, "step": 14488 }, { "epoch": 0.5247736327417603, "grad_norm": 1.3795046770798853, "learning_rate": 9.68391090919853e-06, "loss": 0.7811, "step": 14489 }, { "epoch": 0.5248098515030786, "grad_norm": 1.5421278436672736, "learning_rate": 9.682738429824787e-06, "loss": 0.689, "step": 14490 }, { "epoch": 0.524846070264397, "grad_norm": 1.3076128564297071, "learning_rate": 9.681565954816843e-06, "loss": 0.7797, "step": 14491 }, { "epoch": 0.5248822890257153, "grad_norm": 1.04462546446029, "learning_rate": 9.680393484190833e-06, "loss": 0.6767, "step": 14492 }, { "epoch": 0.5249185077870336, "grad_norm": 1.8768073431238441, "learning_rate": 9.679221017962888e-06, "loss": 0.7842, "step": 14493 }, { "epoch": 0.5249547265483521, "grad_norm": 1.4726199036825647, "learning_rate": 9.678048556149144e-06, "loss": 0.7792, "step": 14494 }, { "epoch": 0.5249909453096704, "grad_norm": 1.3044670020018339, "learning_rate": 9.676876098765734e-06, "loss": 0.7057, "step": 14495 }, { "epoch": 0.5250271640709888, "grad_norm": 1.3010571834368356, "learning_rate": 9.675703645828795e-06, "loss": 0.6964, "step": 14496 }, { "epoch": 0.5250633828323071, "grad_norm": 1.4804015433726012, "learning_rate": 9.674531197354457e-06, "loss": 0.649, "step": 14497 }, { "epoch": 0.5250996015936255, "grad_norm": 1.4274506156464066, "learning_rate": 9.673358753358857e-06, "loss": 0.7507, "step": 14498 }, { "epoch": 0.5251358203549439, "grad_norm": 1.337487025537798, "learning_rate": 9.672186313858125e-06, "loss": 0.7133, "step": 14499 }, { "epoch": 0.5251720391162622, "grad_norm": 1.3609315343763286, "learning_rate": 9.671013878868401e-06, "loss": 0.6978, "step": 14500 }, { "epoch": 0.5252082578775806, "grad_norm": 1.3518397096568109, "learning_rate": 9.66984144840581e-06, "loss": 0.6691, "step": 14501 }, { "epoch": 0.525244476638899, "grad_norm": 1.3576173051623592, "learning_rate": 9.668669022486495e-06, "loss": 0.6937, "step": 14502 }, { "epoch": 0.5252806954002173, "grad_norm": 1.5247349361660507, "learning_rate": 9.667496601126585e-06, "loss": 0.775, "step": 14503 }, { "epoch": 0.5253169141615357, "grad_norm": 1.5573979214447244, "learning_rate": 9.666324184342211e-06, "loss": 0.7298, "step": 14504 }, { "epoch": 0.525353132922854, "grad_norm": 1.4634252190222246, "learning_rate": 9.665151772149513e-06, "loss": 0.7175, "step": 14505 }, { "epoch": 0.5253893516841724, "grad_norm": 1.1222684997389165, "learning_rate": 9.663979364564619e-06, "loss": 0.7607, "step": 14506 }, { "epoch": 0.5254255704454908, "grad_norm": 1.396239767459712, "learning_rate": 9.662806961603665e-06, "loss": 0.7898, "step": 14507 }, { "epoch": 0.5254617892068091, "grad_norm": 1.006891263601181, "learning_rate": 9.661634563282782e-06, "loss": 0.7452, "step": 14508 }, { "epoch": 0.5254980079681275, "grad_norm": 1.2470252179003625, "learning_rate": 9.660462169618108e-06, "loss": 0.6471, "step": 14509 }, { "epoch": 0.5255342267294458, "grad_norm": 1.3235205325907147, "learning_rate": 9.659289780625768e-06, "loss": 0.7373, "step": 14510 }, { "epoch": 0.5255704454907643, "grad_norm": 1.2853326335846966, "learning_rate": 9.658117396321903e-06, "loss": 0.7777, "step": 14511 }, { "epoch": 0.5256066642520826, "grad_norm": 1.5093572521992684, "learning_rate": 9.656945016722645e-06, "loss": 0.6929, "step": 14512 }, { "epoch": 0.5256428830134009, "grad_norm": 1.3722565018925854, "learning_rate": 9.655772641844122e-06, "loss": 0.7419, "step": 14513 }, { "epoch": 0.5256791017747193, "grad_norm": 1.4554932622949346, "learning_rate": 9.654600271702471e-06, "loss": 0.7207, "step": 14514 }, { "epoch": 0.5257153205360376, "grad_norm": 1.5499179620816947, "learning_rate": 9.653427906313822e-06, "loss": 0.7143, "step": 14515 }, { "epoch": 0.5257515392973561, "grad_norm": 1.4370722600774428, "learning_rate": 9.652255545694313e-06, "loss": 0.7849, "step": 14516 }, { "epoch": 0.5257877580586744, "grad_norm": 1.4948988081640397, "learning_rate": 9.65108318986007e-06, "loss": 0.7714, "step": 14517 }, { "epoch": 0.5258239768199927, "grad_norm": 1.4845260079016969, "learning_rate": 9.64991083882723e-06, "loss": 0.7353, "step": 14518 }, { "epoch": 0.5258601955813111, "grad_norm": 1.5428321777937088, "learning_rate": 9.648738492611924e-06, "loss": 0.7268, "step": 14519 }, { "epoch": 0.5258964143426295, "grad_norm": 1.4135934761280615, "learning_rate": 9.647566151230288e-06, "loss": 0.7268, "step": 14520 }, { "epoch": 0.5259326331039479, "grad_norm": 0.9267731567063212, "learning_rate": 9.646393814698448e-06, "loss": 0.6978, "step": 14521 }, { "epoch": 0.5259688518652662, "grad_norm": 1.3925308216829926, "learning_rate": 9.645221483032538e-06, "loss": 0.6922, "step": 14522 }, { "epoch": 0.5260050706265845, "grad_norm": 0.973394989109661, "learning_rate": 9.644049156248697e-06, "loss": 0.6659, "step": 14523 }, { "epoch": 0.526041289387903, "grad_norm": 1.3086468657346528, "learning_rate": 9.64287683436305e-06, "loss": 0.7571, "step": 14524 }, { "epoch": 0.5260775081492213, "grad_norm": 1.3804738442009075, "learning_rate": 9.641704517391733e-06, "loss": 0.7428, "step": 14525 }, { "epoch": 0.5261137269105397, "grad_norm": 1.26493666836533, "learning_rate": 9.640532205350874e-06, "loss": 0.7531, "step": 14526 }, { "epoch": 0.526149945671858, "grad_norm": 1.6118012606404093, "learning_rate": 9.63935989825661e-06, "loss": 0.744, "step": 14527 }, { "epoch": 0.5261861644331763, "grad_norm": 1.448455214829328, "learning_rate": 9.63818759612507e-06, "loss": 0.8101, "step": 14528 }, { "epoch": 0.5262223831944948, "grad_norm": 1.3127785341717408, "learning_rate": 9.637015298972388e-06, "loss": 0.6911, "step": 14529 }, { "epoch": 0.5262586019558131, "grad_norm": 1.6137130208011343, "learning_rate": 9.635843006814691e-06, "loss": 0.7588, "step": 14530 }, { "epoch": 0.5262948207171315, "grad_norm": 1.5106388493156762, "learning_rate": 9.634670719668117e-06, "loss": 0.7277, "step": 14531 }, { "epoch": 0.5263310394784498, "grad_norm": 1.2951724067551873, "learning_rate": 9.633498437548794e-06, "loss": 0.6993, "step": 14532 }, { "epoch": 0.5263672582397682, "grad_norm": 1.5133120472127128, "learning_rate": 9.632326160472854e-06, "loss": 0.7752, "step": 14533 }, { "epoch": 0.5264034770010866, "grad_norm": 1.3418782853606896, "learning_rate": 9.631153888456431e-06, "loss": 0.7519, "step": 14534 }, { "epoch": 0.5264396957624049, "grad_norm": 1.0278705519482458, "learning_rate": 9.629981621515652e-06, "loss": 0.6482, "step": 14535 }, { "epoch": 0.5264759145237233, "grad_norm": 1.2975806282844307, "learning_rate": 9.628809359666654e-06, "loss": 0.7792, "step": 14536 }, { "epoch": 0.5265121332850417, "grad_norm": 1.2724919260486478, "learning_rate": 9.62763710292556e-06, "loss": 0.7537, "step": 14537 }, { "epoch": 0.52654835204636, "grad_norm": 1.3061598447780378, "learning_rate": 9.626464851308514e-06, "loss": 0.7355, "step": 14538 }, { "epoch": 0.5265845708076784, "grad_norm": 1.3322132019885258, "learning_rate": 9.625292604831634e-06, "loss": 0.7192, "step": 14539 }, { "epoch": 0.5266207895689967, "grad_norm": 1.272610568473101, "learning_rate": 9.62412036351106e-06, "loss": 0.6619, "step": 14540 }, { "epoch": 0.5266570083303151, "grad_norm": 1.247166876314136, "learning_rate": 9.622948127362918e-06, "loss": 0.7557, "step": 14541 }, { "epoch": 0.5266932270916335, "grad_norm": 1.368878784391822, "learning_rate": 9.621775896403343e-06, "loss": 0.6851, "step": 14542 }, { "epoch": 0.5267294458529518, "grad_norm": 1.3726438090803308, "learning_rate": 9.620603670648464e-06, "loss": 0.7269, "step": 14543 }, { "epoch": 0.5267656646142702, "grad_norm": 1.4779434753708458, "learning_rate": 9.61943145011441e-06, "loss": 0.8046, "step": 14544 }, { "epoch": 0.5268018833755885, "grad_norm": 1.3044021502837275, "learning_rate": 9.618259234817315e-06, "loss": 0.721, "step": 14545 }, { "epoch": 0.526838102136907, "grad_norm": 1.4191154527623604, "learning_rate": 9.617087024773309e-06, "loss": 0.7284, "step": 14546 }, { "epoch": 0.5268743208982253, "grad_norm": 1.5148835679123667, "learning_rate": 9.615914819998522e-06, "loss": 0.7295, "step": 14547 }, { "epoch": 0.5269105396595436, "grad_norm": 1.403604107102672, "learning_rate": 9.614742620509083e-06, "loss": 0.7679, "step": 14548 }, { "epoch": 0.526946758420862, "grad_norm": 1.461428376210487, "learning_rate": 9.613570426321125e-06, "loss": 0.7568, "step": 14549 }, { "epoch": 0.5269829771821803, "grad_norm": 1.386945360096765, "learning_rate": 9.612398237450776e-06, "loss": 0.7713, "step": 14550 }, { "epoch": 0.5270191959434988, "grad_norm": 1.4334906000452665, "learning_rate": 9.61122605391417e-06, "loss": 0.7525, "step": 14551 }, { "epoch": 0.5270554147048171, "grad_norm": 1.2165113407987902, "learning_rate": 9.610053875727435e-06, "loss": 0.7022, "step": 14552 }, { "epoch": 0.5270916334661354, "grad_norm": 1.083643190606934, "learning_rate": 9.608881702906698e-06, "loss": 0.703, "step": 14553 }, { "epoch": 0.5271278522274538, "grad_norm": 1.2700397478112995, "learning_rate": 9.607709535468097e-06, "loss": 0.7854, "step": 14554 }, { "epoch": 0.5271640709887722, "grad_norm": 1.2708547501114336, "learning_rate": 9.606537373427754e-06, "loss": 0.7391, "step": 14555 }, { "epoch": 0.5272002897500906, "grad_norm": 1.3898129496503984, "learning_rate": 9.605365216801805e-06, "loss": 0.7764, "step": 14556 }, { "epoch": 0.5272365085114089, "grad_norm": 1.3207056205159542, "learning_rate": 9.604193065606376e-06, "loss": 0.7234, "step": 14557 }, { "epoch": 0.5272727272727272, "grad_norm": 1.342627179590116, "learning_rate": 9.6030209198576e-06, "loss": 0.6758, "step": 14558 }, { "epoch": 0.5273089460340457, "grad_norm": 1.4213299007716051, "learning_rate": 9.601848779571603e-06, "loss": 0.787, "step": 14559 }, { "epoch": 0.527345164795364, "grad_norm": 1.2685381893409553, "learning_rate": 9.600676644764517e-06, "loss": 0.6882, "step": 14560 }, { "epoch": 0.5273813835566824, "grad_norm": 0.9278077357549848, "learning_rate": 9.59950451545247e-06, "loss": 0.7219, "step": 14561 }, { "epoch": 0.5274176023180007, "grad_norm": 1.316582330507531, "learning_rate": 9.598332391651596e-06, "loss": 0.781, "step": 14562 }, { "epoch": 0.527453821079319, "grad_norm": 1.2955086171345978, "learning_rate": 9.597160273378018e-06, "loss": 0.7348, "step": 14563 }, { "epoch": 0.5274900398406375, "grad_norm": 1.469915325931798, "learning_rate": 9.595988160647869e-06, "loss": 0.6959, "step": 14564 }, { "epoch": 0.5275262586019558, "grad_norm": 1.6432937034069035, "learning_rate": 9.59481605347728e-06, "loss": 0.7087, "step": 14565 }, { "epoch": 0.5275624773632742, "grad_norm": 1.4247003115726213, "learning_rate": 9.593643951882377e-06, "loss": 0.7436, "step": 14566 }, { "epoch": 0.5275986961245925, "grad_norm": 1.1962055269580876, "learning_rate": 9.59247185587929e-06, "loss": 0.6928, "step": 14567 }, { "epoch": 0.5276349148859109, "grad_norm": 1.510976975338531, "learning_rate": 9.591299765484147e-06, "loss": 0.7114, "step": 14568 }, { "epoch": 0.5276711336472293, "grad_norm": 1.4402612232036822, "learning_rate": 9.590127680713083e-06, "loss": 0.7753, "step": 14569 }, { "epoch": 0.5277073524085476, "grad_norm": 1.2559205377021307, "learning_rate": 9.588955601582218e-06, "loss": 0.7531, "step": 14570 }, { "epoch": 0.527743571169866, "grad_norm": 1.2809058474422095, "learning_rate": 9.587783528107687e-06, "loss": 0.7466, "step": 14571 }, { "epoch": 0.5277797899311844, "grad_norm": 1.1210398536446935, "learning_rate": 9.586611460305614e-06, "loss": 0.6827, "step": 14572 }, { "epoch": 0.5278160086925027, "grad_norm": 1.5423438424381, "learning_rate": 9.585439398192133e-06, "loss": 0.6796, "step": 14573 }, { "epoch": 0.5278522274538211, "grad_norm": 1.4312375543398899, "learning_rate": 9.584267341783372e-06, "loss": 0.6714, "step": 14574 }, { "epoch": 0.5278884462151394, "grad_norm": 1.4461570469570728, "learning_rate": 9.583095291095454e-06, "loss": 0.7427, "step": 14575 }, { "epoch": 0.5279246649764578, "grad_norm": 1.4220531824885396, "learning_rate": 9.581923246144514e-06, "loss": 0.6881, "step": 14576 }, { "epoch": 0.5279608837377762, "grad_norm": 1.4558171885145557, "learning_rate": 9.580751206946675e-06, "loss": 0.7133, "step": 14577 }, { "epoch": 0.5279971024990945, "grad_norm": 1.3068366225354417, "learning_rate": 9.57957917351807e-06, "loss": 0.7814, "step": 14578 }, { "epoch": 0.5280333212604129, "grad_norm": 1.3799435648735852, "learning_rate": 9.578407145874824e-06, "loss": 0.7577, "step": 14579 }, { "epoch": 0.5280695400217312, "grad_norm": 1.033146515656089, "learning_rate": 9.577235124033065e-06, "loss": 0.7213, "step": 14580 }, { "epoch": 0.5281057587830497, "grad_norm": 1.4120537165202842, "learning_rate": 9.576063108008924e-06, "loss": 0.7635, "step": 14581 }, { "epoch": 0.528141977544368, "grad_norm": 1.3109104587348508, "learning_rate": 9.574891097818527e-06, "loss": 0.7156, "step": 14582 }, { "epoch": 0.5281781963056863, "grad_norm": 1.3533012501104513, "learning_rate": 9.573719093478e-06, "loss": 0.7345, "step": 14583 }, { "epoch": 0.5282144150670047, "grad_norm": 1.27180911237349, "learning_rate": 9.572547095003473e-06, "loss": 0.6988, "step": 14584 }, { "epoch": 0.528250633828323, "grad_norm": 1.378679085923033, "learning_rate": 9.571375102411075e-06, "loss": 0.7078, "step": 14585 }, { "epoch": 0.5282868525896415, "grad_norm": 1.4149840820088182, "learning_rate": 9.570203115716932e-06, "loss": 0.775, "step": 14586 }, { "epoch": 0.5283230713509598, "grad_norm": 1.0682217201013255, "learning_rate": 9.569031134937172e-06, "loss": 0.716, "step": 14587 }, { "epoch": 0.5283592901122781, "grad_norm": 1.281802496618472, "learning_rate": 9.567859160087922e-06, "loss": 0.7301, "step": 14588 }, { "epoch": 0.5283955088735965, "grad_norm": 1.3913344037283364, "learning_rate": 9.566687191185311e-06, "loss": 0.6725, "step": 14589 }, { "epoch": 0.5284317276349149, "grad_norm": 1.341364809017094, "learning_rate": 9.565515228245462e-06, "loss": 0.7592, "step": 14590 }, { "epoch": 0.5284679463962333, "grad_norm": 1.3785734072103848, "learning_rate": 9.564343271284509e-06, "loss": 0.7364, "step": 14591 }, { "epoch": 0.5285041651575516, "grad_norm": 1.3622337281809804, "learning_rate": 9.563171320318571e-06, "loss": 0.7194, "step": 14592 }, { "epoch": 0.5285403839188699, "grad_norm": 1.498681677463245, "learning_rate": 9.561999375363784e-06, "loss": 0.8125, "step": 14593 }, { "epoch": 0.5285766026801884, "grad_norm": 1.3212255905311463, "learning_rate": 9.560827436436266e-06, "loss": 0.6894, "step": 14594 }, { "epoch": 0.5286128214415067, "grad_norm": 1.4884278495555017, "learning_rate": 9.55965550355215e-06, "loss": 0.7359, "step": 14595 }, { "epoch": 0.5286490402028251, "grad_norm": 0.9754879481214989, "learning_rate": 9.558483576727563e-06, "loss": 0.7389, "step": 14596 }, { "epoch": 0.5286852589641434, "grad_norm": 1.3478250812427117, "learning_rate": 9.557311655978628e-06, "loss": 0.6918, "step": 14597 }, { "epoch": 0.5287214777254617, "grad_norm": 1.1290140652197431, "learning_rate": 9.556139741321475e-06, "loss": 0.7583, "step": 14598 }, { "epoch": 0.5287576964867802, "grad_norm": 1.526431171420947, "learning_rate": 9.554967832772228e-06, "loss": 0.7172, "step": 14599 }, { "epoch": 0.5287939152480985, "grad_norm": 1.3561738046654037, "learning_rate": 9.553795930347016e-06, "loss": 0.7402, "step": 14600 }, { "epoch": 0.5288301340094169, "grad_norm": 1.5343119165462231, "learning_rate": 9.552624034061963e-06, "loss": 0.661, "step": 14601 }, { "epoch": 0.5288663527707352, "grad_norm": 1.1603902593210926, "learning_rate": 9.551452143933197e-06, "loss": 0.7225, "step": 14602 }, { "epoch": 0.5289025715320536, "grad_norm": 1.2106471397122247, "learning_rate": 9.550280259976844e-06, "loss": 0.7703, "step": 14603 }, { "epoch": 0.528938790293372, "grad_norm": 1.2835454513206817, "learning_rate": 9.549108382209026e-06, "loss": 0.6748, "step": 14604 }, { "epoch": 0.5289750090546903, "grad_norm": 1.0536661845647737, "learning_rate": 9.547936510645878e-06, "loss": 0.7974, "step": 14605 }, { "epoch": 0.5290112278160087, "grad_norm": 1.096141938694872, "learning_rate": 9.546764645303518e-06, "loss": 0.7596, "step": 14606 }, { "epoch": 0.529047446577327, "grad_norm": 1.1061122645345032, "learning_rate": 9.545592786198076e-06, "loss": 0.7927, "step": 14607 }, { "epoch": 0.5290836653386454, "grad_norm": 1.3503136005123448, "learning_rate": 9.544420933345675e-06, "loss": 0.7127, "step": 14608 }, { "epoch": 0.5291198840999638, "grad_norm": 1.3530303379605275, "learning_rate": 9.543249086762443e-06, "loss": 0.6812, "step": 14609 }, { "epoch": 0.5291561028612821, "grad_norm": 1.3468153942575734, "learning_rate": 9.542077246464506e-06, "loss": 0.7731, "step": 14610 }, { "epoch": 0.5291923216226005, "grad_norm": 1.3161860721920848, "learning_rate": 9.540905412467988e-06, "loss": 0.8195, "step": 14611 }, { "epoch": 0.5292285403839189, "grad_norm": 1.3272547479285202, "learning_rate": 9.539733584789014e-06, "loss": 0.7026, "step": 14612 }, { "epoch": 0.5292647591452372, "grad_norm": 1.4603025266438197, "learning_rate": 9.538561763443711e-06, "loss": 0.7585, "step": 14613 }, { "epoch": 0.5293009779065556, "grad_norm": 1.414253124539587, "learning_rate": 9.537389948448203e-06, "loss": 0.7507, "step": 14614 }, { "epoch": 0.5293371966678739, "grad_norm": 1.2854274772260808, "learning_rate": 9.536218139818615e-06, "loss": 0.6968, "step": 14615 }, { "epoch": 0.5293734154291924, "grad_norm": 1.3933735283960804, "learning_rate": 9.535046337571075e-06, "loss": 0.7454, "step": 14616 }, { "epoch": 0.5294096341905107, "grad_norm": 1.2958439518700708, "learning_rate": 9.533874541721704e-06, "loss": 0.6935, "step": 14617 }, { "epoch": 0.529445852951829, "grad_norm": 1.0877067288605162, "learning_rate": 9.532702752286632e-06, "loss": 0.6855, "step": 14618 }, { "epoch": 0.5294820717131474, "grad_norm": 1.2443909153777912, "learning_rate": 9.531530969281977e-06, "loss": 0.6779, "step": 14619 }, { "epoch": 0.5295182904744657, "grad_norm": 1.4288510392724372, "learning_rate": 9.53035919272387e-06, "loss": 0.7519, "step": 14620 }, { "epoch": 0.5295545092357842, "grad_norm": 1.4220705369494135, "learning_rate": 9.529187422628432e-06, "loss": 0.6684, "step": 14621 }, { "epoch": 0.5295907279971025, "grad_norm": 1.3173782887250345, "learning_rate": 9.52801565901179e-06, "loss": 0.7391, "step": 14622 }, { "epoch": 0.5296269467584208, "grad_norm": 1.3652899107327094, "learning_rate": 9.526843901890065e-06, "loss": 0.7334, "step": 14623 }, { "epoch": 0.5296631655197392, "grad_norm": 1.3477502907546322, "learning_rate": 9.525672151279386e-06, "loss": 0.749, "step": 14624 }, { "epoch": 0.5296993842810576, "grad_norm": 1.4202890924795246, "learning_rate": 9.524500407195873e-06, "loss": 0.7867, "step": 14625 }, { "epoch": 0.529735603042376, "grad_norm": 1.4611266920167796, "learning_rate": 9.523328669655653e-06, "loss": 0.7376, "step": 14626 }, { "epoch": 0.5297718218036943, "grad_norm": 1.250161025732249, "learning_rate": 9.522156938674851e-06, "loss": 0.7104, "step": 14627 }, { "epoch": 0.5298080405650126, "grad_norm": 1.307855647012272, "learning_rate": 9.520985214269586e-06, "loss": 0.7368, "step": 14628 }, { "epoch": 0.5298442593263311, "grad_norm": 1.255199397633695, "learning_rate": 9.519813496455989e-06, "loss": 0.7195, "step": 14629 }, { "epoch": 0.5298804780876494, "grad_norm": 1.0569566987822816, "learning_rate": 9.518641785250177e-06, "loss": 0.7062, "step": 14630 }, { "epoch": 0.5299166968489678, "grad_norm": 1.33961058892488, "learning_rate": 9.51747008066828e-06, "loss": 0.7617, "step": 14631 }, { "epoch": 0.5299529156102861, "grad_norm": 1.334989186052929, "learning_rate": 9.516298382726415e-06, "loss": 0.7302, "step": 14632 }, { "epoch": 0.5299891343716044, "grad_norm": 1.068802562383646, "learning_rate": 9.515126691440714e-06, "loss": 0.7136, "step": 14633 }, { "epoch": 0.5300253531329229, "grad_norm": 1.4268008636204135, "learning_rate": 9.513955006827292e-06, "loss": 0.8676, "step": 14634 }, { "epoch": 0.5300615718942412, "grad_norm": 1.1440023075856565, "learning_rate": 9.512783328902275e-06, "loss": 0.7209, "step": 14635 }, { "epoch": 0.5300977906555596, "grad_norm": 1.3007728645123156, "learning_rate": 9.511611657681792e-06, "loss": 0.7368, "step": 14636 }, { "epoch": 0.5301340094168779, "grad_norm": 1.307188719534117, "learning_rate": 9.510439993181958e-06, "loss": 0.8201, "step": 14637 }, { "epoch": 0.5301702281781963, "grad_norm": 1.3172497438224926, "learning_rate": 9.509268335418902e-06, "loss": 0.7246, "step": 14638 }, { "epoch": 0.5302064469395147, "grad_norm": 1.3501742680974915, "learning_rate": 9.508096684408744e-06, "loss": 0.7309, "step": 14639 }, { "epoch": 0.530242665700833, "grad_norm": 1.0327792739384023, "learning_rate": 9.506925040167609e-06, "loss": 0.7101, "step": 14640 }, { "epoch": 0.5302788844621514, "grad_norm": 1.4050023266070506, "learning_rate": 9.505753402711618e-06, "loss": 0.7188, "step": 14641 }, { "epoch": 0.5303151032234698, "grad_norm": 1.099962034280629, "learning_rate": 9.504581772056898e-06, "loss": 0.7229, "step": 14642 }, { "epoch": 0.5303513219847881, "grad_norm": 1.2902128924048983, "learning_rate": 9.503410148219565e-06, "loss": 0.6995, "step": 14643 }, { "epoch": 0.5303875407461065, "grad_norm": 2.036129292363845, "learning_rate": 9.502238531215743e-06, "loss": 0.7445, "step": 14644 }, { "epoch": 0.5304237595074248, "grad_norm": 1.385126756837125, "learning_rate": 9.50106692106156e-06, "loss": 0.6693, "step": 14645 }, { "epoch": 0.5304599782687432, "grad_norm": 1.4094530654642061, "learning_rate": 9.49989531777313e-06, "loss": 0.7802, "step": 14646 }, { "epoch": 0.5304961970300616, "grad_norm": 1.1753547225912708, "learning_rate": 9.49872372136658e-06, "loss": 0.7228, "step": 14647 }, { "epoch": 0.5305324157913799, "grad_norm": 1.7191125823258049, "learning_rate": 9.497552131858035e-06, "loss": 0.7309, "step": 14648 }, { "epoch": 0.5305686345526983, "grad_norm": 1.2369394847293917, "learning_rate": 9.496380549263613e-06, "loss": 0.6321, "step": 14649 }, { "epoch": 0.5306048533140166, "grad_norm": 1.428349129057371, "learning_rate": 9.495208973599438e-06, "loss": 0.7538, "step": 14650 }, { "epoch": 0.5306410720753351, "grad_norm": 1.3317189647607697, "learning_rate": 9.494037404881629e-06, "loss": 0.7706, "step": 14651 }, { "epoch": 0.5306772908366534, "grad_norm": 1.410799603572565, "learning_rate": 9.492865843126312e-06, "loss": 0.7286, "step": 14652 }, { "epoch": 0.5307135095979717, "grad_norm": 1.2632893642266385, "learning_rate": 9.491694288349605e-06, "loss": 0.6713, "step": 14653 }, { "epoch": 0.5307497283592901, "grad_norm": 1.207142080664928, "learning_rate": 9.490522740567633e-06, "loss": 0.6907, "step": 14654 }, { "epoch": 0.5307859471206084, "grad_norm": 1.267825604307119, "learning_rate": 9.489351199796514e-06, "loss": 0.6643, "step": 14655 }, { "epoch": 0.5308221658819269, "grad_norm": 1.2648864610469233, "learning_rate": 9.488179666052374e-06, "loss": 0.7751, "step": 14656 }, { "epoch": 0.5308583846432452, "grad_norm": 1.3088307970193116, "learning_rate": 9.487008139351327e-06, "loss": 0.7673, "step": 14657 }, { "epoch": 0.5308946034045635, "grad_norm": 1.0393507382776184, "learning_rate": 9.485836619709501e-06, "loss": 0.6874, "step": 14658 }, { "epoch": 0.5309308221658819, "grad_norm": 1.5338924923454977, "learning_rate": 9.484665107143017e-06, "loss": 0.7064, "step": 14659 }, { "epoch": 0.5309670409272003, "grad_norm": 1.3707744336597407, "learning_rate": 9.48349360166799e-06, "loss": 0.7036, "step": 14660 }, { "epoch": 0.5310032596885187, "grad_norm": 1.395269570459161, "learning_rate": 9.48232210330055e-06, "loss": 0.7319, "step": 14661 }, { "epoch": 0.531039478449837, "grad_norm": 1.3488643828621898, "learning_rate": 9.481150612056807e-06, "loss": 0.7695, "step": 14662 }, { "epoch": 0.5310756972111553, "grad_norm": 1.3904321589061466, "learning_rate": 9.47997912795289e-06, "loss": 0.7349, "step": 14663 }, { "epoch": 0.5311119159724738, "grad_norm": 1.118825377923142, "learning_rate": 9.478807651004918e-06, "loss": 0.6982, "step": 14664 }, { "epoch": 0.5311481347337921, "grad_norm": 1.4705891520702437, "learning_rate": 9.47763618122901e-06, "loss": 0.7201, "step": 14665 }, { "epoch": 0.5311843534951105, "grad_norm": 1.4554886194102474, "learning_rate": 9.476464718641287e-06, "loss": 0.7177, "step": 14666 }, { "epoch": 0.5312205722564288, "grad_norm": 1.2380340811591406, "learning_rate": 9.47529326325787e-06, "loss": 0.7956, "step": 14667 }, { "epoch": 0.5312567910177471, "grad_norm": 1.3133752883100007, "learning_rate": 9.474121815094878e-06, "loss": 0.7124, "step": 14668 }, { "epoch": 0.5312930097790656, "grad_norm": 1.6905413898693002, "learning_rate": 9.47295037416843e-06, "loss": 0.7597, "step": 14669 }, { "epoch": 0.5313292285403839, "grad_norm": 1.3080692136723377, "learning_rate": 9.471778940494652e-06, "loss": 0.6688, "step": 14670 }, { "epoch": 0.5313654473017023, "grad_norm": 1.261007512995712, "learning_rate": 9.470607514089657e-06, "loss": 0.7629, "step": 14671 }, { "epoch": 0.5314016660630206, "grad_norm": 1.4290146573314728, "learning_rate": 9.46943609496957e-06, "loss": 0.782, "step": 14672 }, { "epoch": 0.531437884824339, "grad_norm": 1.428122495089782, "learning_rate": 9.468264683150507e-06, "loss": 0.746, "step": 14673 }, { "epoch": 0.5314741035856574, "grad_norm": 1.305633092326062, "learning_rate": 9.46709327864859e-06, "loss": 0.6951, "step": 14674 }, { "epoch": 0.5315103223469757, "grad_norm": 1.5340510527038604, "learning_rate": 9.465921881479935e-06, "loss": 0.8004, "step": 14675 }, { "epoch": 0.5315465411082941, "grad_norm": 2.115488165708479, "learning_rate": 9.464750491660668e-06, "loss": 0.7502, "step": 14676 }, { "epoch": 0.5315827598696125, "grad_norm": 1.407252784659412, "learning_rate": 9.463579109206904e-06, "loss": 0.7305, "step": 14677 }, { "epoch": 0.5316189786309308, "grad_norm": 0.9154500450245121, "learning_rate": 9.46240773413476e-06, "loss": 0.7722, "step": 14678 }, { "epoch": 0.5316551973922492, "grad_norm": 1.3561050346628896, "learning_rate": 9.461236366460362e-06, "loss": 0.6945, "step": 14679 }, { "epoch": 0.5316914161535675, "grad_norm": 1.4600577204338536, "learning_rate": 9.46006500619982e-06, "loss": 0.7227, "step": 14680 }, { "epoch": 0.531727634914886, "grad_norm": 1.0696747580554895, "learning_rate": 9.458893653369261e-06, "loss": 0.716, "step": 14681 }, { "epoch": 0.5317638536762043, "grad_norm": 1.4843634134392525, "learning_rate": 9.457722307984801e-06, "loss": 0.7861, "step": 14682 }, { "epoch": 0.5318000724375226, "grad_norm": 1.3209027510113858, "learning_rate": 9.456550970062559e-06, "loss": 0.7695, "step": 14683 }, { "epoch": 0.531836291198841, "grad_norm": 1.3034043923212038, "learning_rate": 9.455379639618651e-06, "loss": 0.8056, "step": 14684 }, { "epoch": 0.5318725099601593, "grad_norm": 1.4619402593228905, "learning_rate": 9.4542083166692e-06, "loss": 0.7778, "step": 14685 }, { "epoch": 0.5319087287214778, "grad_norm": 1.0921985027402992, "learning_rate": 9.45303700123032e-06, "loss": 0.6898, "step": 14686 }, { "epoch": 0.5319449474827961, "grad_norm": 2.0837545423457824, "learning_rate": 9.451865693318134e-06, "loss": 0.6751, "step": 14687 }, { "epoch": 0.5319811662441144, "grad_norm": 1.4688452235764058, "learning_rate": 9.450694392948755e-06, "loss": 0.7232, "step": 14688 }, { "epoch": 0.5320173850054328, "grad_norm": 1.4073770390996467, "learning_rate": 9.449523100138305e-06, "loss": 0.7974, "step": 14689 }, { "epoch": 0.5320536037667511, "grad_norm": 1.2720914115309818, "learning_rate": 9.4483518149029e-06, "loss": 0.733, "step": 14690 }, { "epoch": 0.5320898225280696, "grad_norm": 1.2781040851711036, "learning_rate": 9.44718053725866e-06, "loss": 0.7433, "step": 14691 }, { "epoch": 0.5321260412893879, "grad_norm": 1.256975399258965, "learning_rate": 9.446009267221702e-06, "loss": 0.7759, "step": 14692 }, { "epoch": 0.5321622600507062, "grad_norm": 1.2871410701710235, "learning_rate": 9.44483800480814e-06, "loss": 0.765, "step": 14693 }, { "epoch": 0.5321984788120246, "grad_norm": 1.3331635017441903, "learning_rate": 9.4436667500341e-06, "loss": 0.7537, "step": 14694 }, { "epoch": 0.532234697573343, "grad_norm": 1.4311484171183253, "learning_rate": 9.44249550291569e-06, "loss": 0.746, "step": 14695 }, { "epoch": 0.5322709163346614, "grad_norm": 1.3504430042908004, "learning_rate": 9.441324263469035e-06, "loss": 0.7754, "step": 14696 }, { "epoch": 0.5323071350959797, "grad_norm": 1.0110433839212456, "learning_rate": 9.440153031710247e-06, "loss": 0.7074, "step": 14697 }, { "epoch": 0.532343353857298, "grad_norm": 1.2032130902145923, "learning_rate": 9.438981807655446e-06, "loss": 0.7771, "step": 14698 }, { "epoch": 0.5323795726186165, "grad_norm": 1.405993406409869, "learning_rate": 9.437810591320748e-06, "loss": 0.8525, "step": 14699 }, { "epoch": 0.5324157913799348, "grad_norm": 1.8161145771135523, "learning_rate": 9.436639382722268e-06, "loss": 0.7577, "step": 14700 }, { "epoch": 0.5324520101412532, "grad_norm": 1.3771640102061622, "learning_rate": 9.435468181876128e-06, "loss": 0.7122, "step": 14701 }, { "epoch": 0.5324882289025715, "grad_norm": 1.3310737699454818, "learning_rate": 9.434296988798442e-06, "loss": 0.696, "step": 14702 }, { "epoch": 0.5325244476638898, "grad_norm": 1.5008516327907135, "learning_rate": 9.433125803505325e-06, "loss": 0.792, "step": 14703 }, { "epoch": 0.5325606664252083, "grad_norm": 1.2832637914777265, "learning_rate": 9.431954626012896e-06, "loss": 0.6488, "step": 14704 }, { "epoch": 0.5325968851865266, "grad_norm": 1.3862283685822145, "learning_rate": 9.43078345633727e-06, "loss": 0.6992, "step": 14705 }, { "epoch": 0.532633103947845, "grad_norm": 1.3446420602835907, "learning_rate": 9.429612294494563e-06, "loss": 0.7604, "step": 14706 }, { "epoch": 0.5326693227091633, "grad_norm": 1.3652881508033552, "learning_rate": 9.428441140500895e-06, "loss": 0.7701, "step": 14707 }, { "epoch": 0.5327055414704817, "grad_norm": 1.386316172499092, "learning_rate": 9.427269994372376e-06, "loss": 0.7416, "step": 14708 }, { "epoch": 0.5327417602318001, "grad_norm": 1.4320463773903425, "learning_rate": 9.426098856125125e-06, "loss": 0.618, "step": 14709 }, { "epoch": 0.5327779789931184, "grad_norm": 1.526096088059635, "learning_rate": 9.42492772577526e-06, "loss": 0.7868, "step": 14710 }, { "epoch": 0.5328141977544368, "grad_norm": 1.4136549081651169, "learning_rate": 9.423756603338894e-06, "loss": 0.7556, "step": 14711 }, { "epoch": 0.5328504165157552, "grad_norm": 1.1240121175318607, "learning_rate": 9.422585488832144e-06, "loss": 0.6993, "step": 14712 }, { "epoch": 0.5328866352770735, "grad_norm": 1.4229863115439427, "learning_rate": 9.421414382271123e-06, "loss": 0.7515, "step": 14713 }, { "epoch": 0.5329228540383919, "grad_norm": 1.3363250134143883, "learning_rate": 9.420243283671952e-06, "loss": 0.7746, "step": 14714 }, { "epoch": 0.5329590727997102, "grad_norm": 1.3432581009533557, "learning_rate": 9.41907219305074e-06, "loss": 0.7376, "step": 14715 }, { "epoch": 0.5329952915610287, "grad_norm": 1.2645328100655124, "learning_rate": 9.417901110423605e-06, "loss": 0.7515, "step": 14716 }, { "epoch": 0.533031510322347, "grad_norm": 1.3865749761436243, "learning_rate": 9.416730035806664e-06, "loss": 0.6427, "step": 14717 }, { "epoch": 0.5330677290836653, "grad_norm": 1.2398702490555162, "learning_rate": 9.41555896921603e-06, "loss": 0.7489, "step": 14718 }, { "epoch": 0.5331039478449837, "grad_norm": 1.2981500368511625, "learning_rate": 9.414387910667816e-06, "loss": 0.7212, "step": 14719 }, { "epoch": 0.533140166606302, "grad_norm": 1.5880468932111587, "learning_rate": 9.41321686017814e-06, "loss": 0.7307, "step": 14720 }, { "epoch": 0.5331763853676205, "grad_norm": 1.3987193536160316, "learning_rate": 9.412045817763116e-06, "loss": 0.7539, "step": 14721 }, { "epoch": 0.5332126041289388, "grad_norm": 1.4397426408535987, "learning_rate": 9.410874783438858e-06, "loss": 0.7976, "step": 14722 }, { "epoch": 0.5332488228902571, "grad_norm": 1.3325912222861864, "learning_rate": 9.409703757221482e-06, "loss": 0.7443, "step": 14723 }, { "epoch": 0.5332850416515755, "grad_norm": 0.9895026667921992, "learning_rate": 9.408532739127098e-06, "loss": 0.7346, "step": 14724 }, { "epoch": 0.5333212604128938, "grad_norm": 1.4616818883917992, "learning_rate": 9.407361729171827e-06, "loss": 0.7345, "step": 14725 }, { "epoch": 0.5333574791742123, "grad_norm": 1.546943395720908, "learning_rate": 9.406190727371775e-06, "loss": 0.7325, "step": 14726 }, { "epoch": 0.5333936979355306, "grad_norm": 1.231728909090202, "learning_rate": 9.405019733743066e-06, "loss": 0.7289, "step": 14727 }, { "epoch": 0.5334299166968489, "grad_norm": 1.4042667516997767, "learning_rate": 9.403848748301802e-06, "loss": 0.8213, "step": 14728 }, { "epoch": 0.5334661354581673, "grad_norm": 1.0705153455403524, "learning_rate": 9.402677771064108e-06, "loss": 0.6759, "step": 14729 }, { "epoch": 0.5335023542194857, "grad_norm": 1.2949751631530246, "learning_rate": 9.40150680204609e-06, "loss": 0.7074, "step": 14730 }, { "epoch": 0.5335385729808041, "grad_norm": 1.1731756747234598, "learning_rate": 9.400335841263863e-06, "loss": 0.665, "step": 14731 }, { "epoch": 0.5335747917421224, "grad_norm": 2.290542652905985, "learning_rate": 9.399164888733544e-06, "loss": 0.7196, "step": 14732 }, { "epoch": 0.5336110105034407, "grad_norm": 1.3978981850835968, "learning_rate": 9.397993944471245e-06, "loss": 0.7538, "step": 14733 }, { "epoch": 0.5336472292647592, "grad_norm": 1.1500724468742265, "learning_rate": 9.396823008493077e-06, "loss": 0.774, "step": 14734 }, { "epoch": 0.5336834480260775, "grad_norm": 1.397651404137311, "learning_rate": 9.395652080815154e-06, "loss": 0.7719, "step": 14735 }, { "epoch": 0.5337196667873959, "grad_norm": 1.376875342884402, "learning_rate": 9.39448116145359e-06, "loss": 0.7751, "step": 14736 }, { "epoch": 0.5337558855487142, "grad_norm": 1.5372117310196411, "learning_rate": 9.393310250424497e-06, "loss": 0.767, "step": 14737 }, { "epoch": 0.5337921043100325, "grad_norm": 1.3142640259803202, "learning_rate": 9.392139347743989e-06, "loss": 0.7338, "step": 14738 }, { "epoch": 0.533828323071351, "grad_norm": 1.410937510244342, "learning_rate": 9.390968453428176e-06, "loss": 0.727, "step": 14739 }, { "epoch": 0.5338645418326693, "grad_norm": 1.439151976979479, "learning_rate": 9.389797567493172e-06, "loss": 0.8086, "step": 14740 }, { "epoch": 0.5339007605939877, "grad_norm": 1.4057692036987615, "learning_rate": 9.388626689955091e-06, "loss": 0.7443, "step": 14741 }, { "epoch": 0.533936979355306, "grad_norm": 1.3813548386226266, "learning_rate": 9.387455820830044e-06, "loss": 0.7044, "step": 14742 }, { "epoch": 0.5339731981166244, "grad_norm": 1.3933561527160396, "learning_rate": 9.386284960134143e-06, "loss": 0.6846, "step": 14743 }, { "epoch": 0.5340094168779428, "grad_norm": 1.2581499813904609, "learning_rate": 9.3851141078835e-06, "loss": 0.7194, "step": 14744 }, { "epoch": 0.5340456356392611, "grad_norm": 1.3431202995350953, "learning_rate": 9.383943264094227e-06, "loss": 0.6948, "step": 14745 }, { "epoch": 0.5340818544005795, "grad_norm": 0.9971702987442688, "learning_rate": 9.382772428782436e-06, "loss": 0.7223, "step": 14746 }, { "epoch": 0.5341180731618979, "grad_norm": 1.4045983894097198, "learning_rate": 9.38160160196424e-06, "loss": 0.7436, "step": 14747 }, { "epoch": 0.5341542919232162, "grad_norm": 1.3770053813027694, "learning_rate": 9.380430783655748e-06, "loss": 0.7571, "step": 14748 }, { "epoch": 0.5341905106845346, "grad_norm": 1.9563348638999574, "learning_rate": 9.379259973873075e-06, "loss": 0.8041, "step": 14749 }, { "epoch": 0.5342267294458529, "grad_norm": 1.3904676924359731, "learning_rate": 9.378089172632325e-06, "loss": 0.7043, "step": 14750 }, { "epoch": 0.5342629482071714, "grad_norm": 1.1139456513760457, "learning_rate": 9.376918379949618e-06, "loss": 0.7508, "step": 14751 }, { "epoch": 0.5342991669684897, "grad_norm": 1.0451563983065753, "learning_rate": 9.375747595841062e-06, "loss": 0.7355, "step": 14752 }, { "epoch": 0.534335385729808, "grad_norm": 2.653309798544508, "learning_rate": 9.374576820322765e-06, "loss": 0.72, "step": 14753 }, { "epoch": 0.5343716044911264, "grad_norm": 1.2352161069345133, "learning_rate": 9.373406053410844e-06, "loss": 0.7186, "step": 14754 }, { "epoch": 0.5344078232524447, "grad_norm": 1.2726278964022724, "learning_rate": 9.372235295121404e-06, "loss": 0.7347, "step": 14755 }, { "epoch": 0.5344440420137632, "grad_norm": 0.9818579293922669, "learning_rate": 9.37106454547056e-06, "loss": 0.6687, "step": 14756 }, { "epoch": 0.5344802607750815, "grad_norm": 1.2557846684134644, "learning_rate": 9.369893804474416e-06, "loss": 0.7081, "step": 14757 }, { "epoch": 0.5345164795363998, "grad_norm": 1.1309780568557433, "learning_rate": 9.368723072149093e-06, "loss": 0.6919, "step": 14758 }, { "epoch": 0.5345526982977182, "grad_norm": 1.4053613119637827, "learning_rate": 9.367552348510692e-06, "loss": 0.8204, "step": 14759 }, { "epoch": 0.5345889170590366, "grad_norm": 1.387943424162137, "learning_rate": 9.366381633575327e-06, "loss": 0.7476, "step": 14760 }, { "epoch": 0.534625135820355, "grad_norm": 1.422445745988058, "learning_rate": 9.365210927359105e-06, "loss": 0.6876, "step": 14761 }, { "epoch": 0.5346613545816733, "grad_norm": 1.0862260647771662, "learning_rate": 9.36404022987814e-06, "loss": 0.7347, "step": 14762 }, { "epoch": 0.5346975733429916, "grad_norm": 1.1851400898250968, "learning_rate": 9.362869541148542e-06, "loss": 0.7142, "step": 14763 }, { "epoch": 0.53473379210431, "grad_norm": 1.2732440627508959, "learning_rate": 9.361698861186417e-06, "loss": 0.7485, "step": 14764 }, { "epoch": 0.5347700108656284, "grad_norm": 1.279216597855676, "learning_rate": 9.360528190007879e-06, "loss": 0.765, "step": 14765 }, { "epoch": 0.5348062296269468, "grad_norm": 1.264697169686855, "learning_rate": 9.359357527629032e-06, "loss": 0.6863, "step": 14766 }, { "epoch": 0.5348424483882651, "grad_norm": 1.117315761051388, "learning_rate": 9.35818687406599e-06, "loss": 0.6738, "step": 14767 }, { "epoch": 0.5348786671495834, "grad_norm": 1.2371022511196723, "learning_rate": 9.357016229334859e-06, "loss": 0.6774, "step": 14768 }, { "epoch": 0.5349148859109019, "grad_norm": 1.040631389461822, "learning_rate": 9.355845593451751e-06, "loss": 0.7329, "step": 14769 }, { "epoch": 0.5349511046722202, "grad_norm": 1.3093836834148456, "learning_rate": 9.354674966432772e-06, "loss": 0.7, "step": 14770 }, { "epoch": 0.5349873234335386, "grad_norm": 1.5173585664165012, "learning_rate": 9.353504348294034e-06, "loss": 0.758, "step": 14771 }, { "epoch": 0.5350235421948569, "grad_norm": 1.2707085835644207, "learning_rate": 9.352333739051646e-06, "loss": 0.6851, "step": 14772 }, { "epoch": 0.5350597609561752, "grad_norm": 1.3294069834754871, "learning_rate": 9.351163138721711e-06, "loss": 0.6851, "step": 14773 }, { "epoch": 0.5350959797174937, "grad_norm": 1.4783078719180234, "learning_rate": 9.349992547320344e-06, "loss": 0.7888, "step": 14774 }, { "epoch": 0.535132198478812, "grad_norm": 0.9838272957096681, "learning_rate": 9.34882196486365e-06, "loss": 0.6576, "step": 14775 }, { "epoch": 0.5351684172401304, "grad_norm": 1.597187122590482, "learning_rate": 9.347651391367738e-06, "loss": 0.6843, "step": 14776 }, { "epoch": 0.5352046360014487, "grad_norm": 1.3553294665427085, "learning_rate": 9.346480826848714e-06, "loss": 0.7506, "step": 14777 }, { "epoch": 0.5352408547627671, "grad_norm": 1.0999606038035181, "learning_rate": 9.345310271322692e-06, "loss": 0.7752, "step": 14778 }, { "epoch": 0.5352770735240855, "grad_norm": 1.3011715127411345, "learning_rate": 9.344139724805774e-06, "loss": 0.7018, "step": 14779 }, { "epoch": 0.5353132922854038, "grad_norm": 1.3379717057433889, "learning_rate": 9.34296918731407e-06, "loss": 0.6949, "step": 14780 }, { "epoch": 0.5353495110467222, "grad_norm": 1.3303531447824373, "learning_rate": 9.341798658863686e-06, "loss": 0.7913, "step": 14781 }, { "epoch": 0.5353857298080406, "grad_norm": 1.3482442492198916, "learning_rate": 9.34062813947073e-06, "loss": 0.721, "step": 14782 }, { "epoch": 0.5354219485693589, "grad_norm": 1.0417208073498658, "learning_rate": 9.339457629151314e-06, "loss": 0.7075, "step": 14783 }, { "epoch": 0.5354581673306773, "grad_norm": 1.5654312426487367, "learning_rate": 9.338287127921538e-06, "loss": 0.6846, "step": 14784 }, { "epoch": 0.5354943860919956, "grad_norm": 1.1963382422789723, "learning_rate": 9.337116635797516e-06, "loss": 0.7082, "step": 14785 }, { "epoch": 0.535530604853314, "grad_norm": 1.3184888853815553, "learning_rate": 9.335946152795347e-06, "loss": 0.6663, "step": 14786 }, { "epoch": 0.5355668236146324, "grad_norm": 1.716715377852219, "learning_rate": 9.334775678931147e-06, "loss": 0.7619, "step": 14787 }, { "epoch": 0.5356030423759507, "grad_norm": 1.392610134249099, "learning_rate": 9.333605214221014e-06, "loss": 0.8142, "step": 14788 }, { "epoch": 0.5356392611372691, "grad_norm": 1.3633768176599783, "learning_rate": 9.332434758681062e-06, "loss": 0.6893, "step": 14789 }, { "epoch": 0.5356754798985874, "grad_norm": 1.0337082871498886, "learning_rate": 9.331264312327394e-06, "loss": 0.758, "step": 14790 }, { "epoch": 0.5357116986599059, "grad_norm": 1.5273613611654782, "learning_rate": 9.330093875176113e-06, "loss": 0.7621, "step": 14791 }, { "epoch": 0.5357479174212242, "grad_norm": 1.4479659608699025, "learning_rate": 9.328923447243333e-06, "loss": 0.8371, "step": 14792 }, { "epoch": 0.5357841361825425, "grad_norm": 1.3614560816926682, "learning_rate": 9.327753028545155e-06, "loss": 0.7932, "step": 14793 }, { "epoch": 0.5358203549438609, "grad_norm": 1.2497836752804983, "learning_rate": 9.326582619097686e-06, "loss": 0.7346, "step": 14794 }, { "epoch": 0.5358565737051793, "grad_norm": 1.2696103592213246, "learning_rate": 9.32541221891703e-06, "loss": 0.7321, "step": 14795 }, { "epoch": 0.5358927924664977, "grad_norm": 1.3401543205866002, "learning_rate": 9.324241828019299e-06, "loss": 0.7216, "step": 14796 }, { "epoch": 0.535929011227816, "grad_norm": 1.2696116774556891, "learning_rate": 9.32307144642059e-06, "loss": 0.7217, "step": 14797 }, { "epoch": 0.5359652299891343, "grad_norm": 1.4233930653065276, "learning_rate": 9.321901074137013e-06, "loss": 0.8089, "step": 14798 }, { "epoch": 0.5360014487504527, "grad_norm": 1.4837303307509964, "learning_rate": 9.320730711184674e-06, "loss": 0.7841, "step": 14799 }, { "epoch": 0.5360376675117711, "grad_norm": 1.1611484655818236, "learning_rate": 9.319560357579679e-06, "loss": 0.6882, "step": 14800 }, { "epoch": 0.5360738862730895, "grad_norm": 1.8336506937121804, "learning_rate": 9.31839001333813e-06, "loss": 0.8044, "step": 14801 }, { "epoch": 0.5361101050344078, "grad_norm": 1.2953678024246957, "learning_rate": 9.317219678476132e-06, "loss": 0.7497, "step": 14802 }, { "epoch": 0.5361463237957261, "grad_norm": 1.1273689346885811, "learning_rate": 9.316049353009793e-06, "loss": 0.7183, "step": 14803 }, { "epoch": 0.5361825425570446, "grad_norm": 1.1546266600249835, "learning_rate": 9.314879036955213e-06, "loss": 0.6526, "step": 14804 }, { "epoch": 0.5362187613183629, "grad_norm": 1.0446560816873316, "learning_rate": 9.313708730328503e-06, "loss": 0.7511, "step": 14805 }, { "epoch": 0.5362549800796813, "grad_norm": 1.3441792041061227, "learning_rate": 9.312538433145763e-06, "loss": 0.7391, "step": 14806 }, { "epoch": 0.5362911988409996, "grad_norm": 1.3046440885368487, "learning_rate": 9.311368145423098e-06, "loss": 0.7358, "step": 14807 }, { "epoch": 0.536327417602318, "grad_norm": 1.3253634233041545, "learning_rate": 9.310197867176612e-06, "loss": 0.7254, "step": 14808 }, { "epoch": 0.5363636363636364, "grad_norm": 1.0109593626010547, "learning_rate": 9.30902759842241e-06, "loss": 0.6813, "step": 14809 }, { "epoch": 0.5363998551249547, "grad_norm": 1.3510547508323334, "learning_rate": 9.307857339176593e-06, "loss": 0.702, "step": 14810 }, { "epoch": 0.5364360738862731, "grad_norm": 1.2466047787710746, "learning_rate": 9.306687089455271e-06, "loss": 0.7386, "step": 14811 }, { "epoch": 0.5364722926475914, "grad_norm": 1.3904185570486904, "learning_rate": 9.305516849274542e-06, "loss": 0.8122, "step": 14812 }, { "epoch": 0.5365085114089098, "grad_norm": 1.8739999398909732, "learning_rate": 9.30434661865051e-06, "loss": 0.585, "step": 14813 }, { "epoch": 0.5365447301702282, "grad_norm": 1.250807776443836, "learning_rate": 9.303176397599283e-06, "loss": 0.7195, "step": 14814 }, { "epoch": 0.5365809489315465, "grad_norm": 1.319163697334381, "learning_rate": 9.302006186136959e-06, "loss": 0.7611, "step": 14815 }, { "epoch": 0.5366171676928649, "grad_norm": 1.3579024129823043, "learning_rate": 9.300835984279644e-06, "loss": 0.7776, "step": 14816 }, { "epoch": 0.5366533864541833, "grad_norm": 1.091442908805115, "learning_rate": 9.299665792043437e-06, "loss": 0.7249, "step": 14817 }, { "epoch": 0.5366896052155016, "grad_norm": 1.5678083835832994, "learning_rate": 9.29849560944445e-06, "loss": 0.7601, "step": 14818 }, { "epoch": 0.53672582397682, "grad_norm": 1.3824492687336387, "learning_rate": 9.297325436498775e-06, "loss": 0.7663, "step": 14819 }, { "epoch": 0.5367620427381383, "grad_norm": 1.3969421690061141, "learning_rate": 9.296155273222521e-06, "loss": 0.7091, "step": 14820 }, { "epoch": 0.5367982614994568, "grad_norm": 1.1903533244108222, "learning_rate": 9.294985119631788e-06, "loss": 0.7497, "step": 14821 }, { "epoch": 0.5368344802607751, "grad_norm": 1.3937152711215057, "learning_rate": 9.293814975742679e-06, "loss": 0.7202, "step": 14822 }, { "epoch": 0.5368706990220934, "grad_norm": 1.156250330946688, "learning_rate": 9.2926448415713e-06, "loss": 0.7282, "step": 14823 }, { "epoch": 0.5369069177834118, "grad_norm": 1.2937371932308066, "learning_rate": 9.291474717133744e-06, "loss": 0.8058, "step": 14824 }, { "epoch": 0.5369431365447301, "grad_norm": 1.414748157324082, "learning_rate": 9.290304602446123e-06, "loss": 0.7165, "step": 14825 }, { "epoch": 0.5369793553060486, "grad_norm": 1.4663183724518425, "learning_rate": 9.289134497524533e-06, "loss": 0.7977, "step": 14826 }, { "epoch": 0.5370155740673669, "grad_norm": 1.253943006030261, "learning_rate": 9.287964402385077e-06, "loss": 0.7618, "step": 14827 }, { "epoch": 0.5370517928286852, "grad_norm": 0.9855033472992453, "learning_rate": 9.286794317043855e-06, "loss": 0.7687, "step": 14828 }, { "epoch": 0.5370880115900036, "grad_norm": 1.2877315390877953, "learning_rate": 9.285624241516973e-06, "loss": 0.7477, "step": 14829 }, { "epoch": 0.537124230351322, "grad_norm": 1.4025809265441913, "learning_rate": 9.284454175820526e-06, "loss": 0.7612, "step": 14830 }, { "epoch": 0.5371604491126404, "grad_norm": 1.348261897509318, "learning_rate": 9.28328411997062e-06, "loss": 0.7659, "step": 14831 }, { "epoch": 0.5371966678739587, "grad_norm": 1.4757622033501105, "learning_rate": 9.282114073983353e-06, "loss": 0.7628, "step": 14832 }, { "epoch": 0.537232886635277, "grad_norm": 1.5093463362367139, "learning_rate": 9.280944037874826e-06, "loss": 0.7995, "step": 14833 }, { "epoch": 0.5372691053965954, "grad_norm": 1.0888575729877936, "learning_rate": 9.279774011661144e-06, "loss": 0.7615, "step": 14834 }, { "epoch": 0.5373053241579138, "grad_norm": 1.6590748585977002, "learning_rate": 9.278603995358402e-06, "loss": 0.7577, "step": 14835 }, { "epoch": 0.5373415429192322, "grad_norm": 1.4687532275245594, "learning_rate": 9.277433988982705e-06, "loss": 0.7417, "step": 14836 }, { "epoch": 0.5373777616805505, "grad_norm": 1.4981790115041895, "learning_rate": 9.27626399255015e-06, "loss": 0.7069, "step": 14837 }, { "epoch": 0.5374139804418688, "grad_norm": 1.3448574713023274, "learning_rate": 9.27509400607684e-06, "loss": 0.7351, "step": 14838 }, { "epoch": 0.5374501992031873, "grad_norm": 1.6145666113238755, "learning_rate": 9.273924029578872e-06, "loss": 0.8428, "step": 14839 }, { "epoch": 0.5374864179645056, "grad_norm": 1.3268929323974878, "learning_rate": 9.272754063072348e-06, "loss": 0.6681, "step": 14840 }, { "epoch": 0.537522636725824, "grad_norm": 0.9743227908812415, "learning_rate": 9.271584106573364e-06, "loss": 0.6917, "step": 14841 }, { "epoch": 0.5375588554871423, "grad_norm": 1.1622905971570134, "learning_rate": 9.270414160098028e-06, "loss": 0.7014, "step": 14842 }, { "epoch": 0.5375950742484608, "grad_norm": 1.6192897226412433, "learning_rate": 9.26924422366243e-06, "loss": 0.7774, "step": 14843 }, { "epoch": 0.5376312930097791, "grad_norm": 1.3631540296485805, "learning_rate": 9.268074297282675e-06, "loss": 0.7247, "step": 14844 }, { "epoch": 0.5376675117710974, "grad_norm": 1.3526007765589279, "learning_rate": 9.26690438097486e-06, "loss": 0.6962, "step": 14845 }, { "epoch": 0.5377037305324158, "grad_norm": 1.418506269535981, "learning_rate": 9.265734474755085e-06, "loss": 0.7075, "step": 14846 }, { "epoch": 0.5377399492937341, "grad_norm": 1.335824900337955, "learning_rate": 9.26456457863945e-06, "loss": 0.7097, "step": 14847 }, { "epoch": 0.5377761680550526, "grad_norm": 1.0533002643305946, "learning_rate": 9.263394692644052e-06, "loss": 0.7317, "step": 14848 }, { "epoch": 0.5378123868163709, "grad_norm": 1.4555822561604606, "learning_rate": 9.262224816784992e-06, "loss": 0.6928, "step": 14849 }, { "epoch": 0.5378486055776892, "grad_norm": 1.566656961853789, "learning_rate": 9.261054951078364e-06, "loss": 0.7323, "step": 14850 }, { "epoch": 0.5378848243390076, "grad_norm": 1.2692530785302154, "learning_rate": 9.259885095540271e-06, "loss": 0.7656, "step": 14851 }, { "epoch": 0.537921043100326, "grad_norm": 1.1448416985859489, "learning_rate": 9.25871525018681e-06, "loss": 0.7006, "step": 14852 }, { "epoch": 0.5379572618616444, "grad_norm": 1.2160919799400314, "learning_rate": 9.25754541503408e-06, "loss": 0.7595, "step": 14853 }, { "epoch": 0.5379934806229627, "grad_norm": 1.2398214456286267, "learning_rate": 9.256375590098176e-06, "loss": 0.6603, "step": 14854 }, { "epoch": 0.538029699384281, "grad_norm": 1.3449350138700253, "learning_rate": 9.255205775395194e-06, "loss": 0.654, "step": 14855 }, { "epoch": 0.5380659181455995, "grad_norm": 1.4899095222455534, "learning_rate": 9.254035970941237e-06, "loss": 0.8355, "step": 14856 }, { "epoch": 0.5381021369069178, "grad_norm": 1.2717438329080692, "learning_rate": 9.2528661767524e-06, "loss": 0.6922, "step": 14857 }, { "epoch": 0.5381383556682362, "grad_norm": 1.3022147097272594, "learning_rate": 9.251696392844781e-06, "loss": 0.7544, "step": 14858 }, { "epoch": 0.5381745744295545, "grad_norm": 1.4468780673058381, "learning_rate": 9.250526619234478e-06, "loss": 0.7605, "step": 14859 }, { "epoch": 0.5382107931908728, "grad_norm": 1.5819427547899574, "learning_rate": 9.249356855937587e-06, "loss": 0.819, "step": 14860 }, { "epoch": 0.5382470119521913, "grad_norm": 1.5321277661926864, "learning_rate": 9.248187102970206e-06, "loss": 0.7122, "step": 14861 }, { "epoch": 0.5382832307135096, "grad_norm": 1.2615335050949832, "learning_rate": 9.247017360348429e-06, "loss": 0.7185, "step": 14862 }, { "epoch": 0.538319449474828, "grad_norm": 1.3346581872687837, "learning_rate": 9.245847628088358e-06, "loss": 0.7219, "step": 14863 }, { "epoch": 0.5383556682361463, "grad_norm": 1.248977145782965, "learning_rate": 9.244677906206083e-06, "loss": 0.7013, "step": 14864 }, { "epoch": 0.5383918869974647, "grad_norm": 1.043566666435407, "learning_rate": 9.243508194717703e-06, "loss": 0.742, "step": 14865 }, { "epoch": 0.5384281057587831, "grad_norm": 1.2071245683626513, "learning_rate": 9.24233849363932e-06, "loss": 0.7373, "step": 14866 }, { "epoch": 0.5384643245201014, "grad_norm": 1.3001904692523105, "learning_rate": 9.24116880298702e-06, "loss": 0.6645, "step": 14867 }, { "epoch": 0.5385005432814198, "grad_norm": 0.9959282885771883, "learning_rate": 9.239999122776906e-06, "loss": 0.75, "step": 14868 }, { "epoch": 0.5385367620427381, "grad_norm": 1.4539333948228201, "learning_rate": 9.238829453025072e-06, "loss": 0.7762, "step": 14869 }, { "epoch": 0.5385729808040565, "grad_norm": 1.3762683836463145, "learning_rate": 9.237659793747615e-06, "loss": 0.7329, "step": 14870 }, { "epoch": 0.5386091995653749, "grad_norm": 1.1992714770123118, "learning_rate": 9.236490144960625e-06, "loss": 0.7533, "step": 14871 }, { "epoch": 0.5386454183266932, "grad_norm": 1.7863167938617128, "learning_rate": 9.235320506680206e-06, "loss": 0.706, "step": 14872 }, { "epoch": 0.5386816370880116, "grad_norm": 1.4979533995668184, "learning_rate": 9.234150878922445e-06, "loss": 0.7016, "step": 14873 }, { "epoch": 0.53871785584933, "grad_norm": 1.4379083802283266, "learning_rate": 9.232981261703442e-06, "loss": 0.7259, "step": 14874 }, { "epoch": 0.5387540746106483, "grad_norm": 1.4730110322915613, "learning_rate": 9.231811655039293e-06, "loss": 0.747, "step": 14875 }, { "epoch": 0.5387902933719667, "grad_norm": 1.4667823806026232, "learning_rate": 9.230642058946086e-06, "loss": 0.722, "step": 14876 }, { "epoch": 0.538826512133285, "grad_norm": 1.174637133732495, "learning_rate": 9.229472473439925e-06, "loss": 0.7623, "step": 14877 }, { "epoch": 0.5388627308946035, "grad_norm": 1.4678762568680643, "learning_rate": 9.228302898536897e-06, "loss": 0.7144, "step": 14878 }, { "epoch": 0.5388989496559218, "grad_norm": 1.0720710117994035, "learning_rate": 9.227133334253101e-06, "loss": 0.6895, "step": 14879 }, { "epoch": 0.5389351684172401, "grad_norm": 1.0612213773226975, "learning_rate": 9.225963780604627e-06, "loss": 0.745, "step": 14880 }, { "epoch": 0.5389713871785585, "grad_norm": 1.014582330984455, "learning_rate": 9.224794237607575e-06, "loss": 0.7315, "step": 14881 }, { "epoch": 0.5390076059398768, "grad_norm": 1.363535238450705, "learning_rate": 9.223624705278033e-06, "loss": 0.776, "step": 14882 }, { "epoch": 0.5390438247011953, "grad_norm": 1.3610379718478505, "learning_rate": 9.222455183632098e-06, "loss": 0.7025, "step": 14883 }, { "epoch": 0.5390800434625136, "grad_norm": 1.343807913674406, "learning_rate": 9.221285672685862e-06, "loss": 0.7348, "step": 14884 }, { "epoch": 0.5391162622238319, "grad_norm": 1.4862889160618966, "learning_rate": 9.220116172455422e-06, "loss": 0.7445, "step": 14885 }, { "epoch": 0.5391524809851503, "grad_norm": 1.426548460732534, "learning_rate": 9.218946682956865e-06, "loss": 0.7008, "step": 14886 }, { "epoch": 0.5391886997464687, "grad_norm": 1.3328369897513237, "learning_rate": 9.21777720420629e-06, "loss": 0.7328, "step": 14887 }, { "epoch": 0.5392249185077871, "grad_norm": 0.9753952104244336, "learning_rate": 9.216607736219787e-06, "loss": 0.7242, "step": 14888 }, { "epoch": 0.5392611372691054, "grad_norm": 1.4010763152233041, "learning_rate": 9.215438279013451e-06, "loss": 0.7371, "step": 14889 }, { "epoch": 0.5392973560304237, "grad_norm": 1.4130922012875706, "learning_rate": 9.214268832603375e-06, "loss": 0.7497, "step": 14890 }, { "epoch": 0.5393335747917422, "grad_norm": 1.4247439314639267, "learning_rate": 9.213099397005647e-06, "loss": 0.7458, "step": 14891 }, { "epoch": 0.5393697935530605, "grad_norm": 1.2321774018077278, "learning_rate": 9.211929972236365e-06, "loss": 0.7134, "step": 14892 }, { "epoch": 0.5394060123143789, "grad_norm": 1.4276680546084963, "learning_rate": 9.210760558311619e-06, "loss": 0.8376, "step": 14893 }, { "epoch": 0.5394422310756972, "grad_norm": 1.5619273972356478, "learning_rate": 9.2095911552475e-06, "loss": 0.7735, "step": 14894 }, { "epoch": 0.5394784498370155, "grad_norm": 1.0010569061041552, "learning_rate": 9.2084217630601e-06, "loss": 0.7524, "step": 14895 }, { "epoch": 0.539514668598334, "grad_norm": 1.2260518385326906, "learning_rate": 9.207252381765513e-06, "loss": 0.6595, "step": 14896 }, { "epoch": 0.5395508873596523, "grad_norm": 1.3073815004447658, "learning_rate": 9.206083011379833e-06, "loss": 0.6955, "step": 14897 }, { "epoch": 0.5395871061209707, "grad_norm": 1.3204714690670085, "learning_rate": 9.204913651919144e-06, "loss": 0.7036, "step": 14898 }, { "epoch": 0.539623324882289, "grad_norm": 1.3439412500713404, "learning_rate": 9.203744303399546e-06, "loss": 0.7481, "step": 14899 }, { "epoch": 0.5396595436436074, "grad_norm": 1.4791062836903512, "learning_rate": 9.202574965837122e-06, "loss": 0.747, "step": 14900 }, { "epoch": 0.5396957624049258, "grad_norm": 1.058612964350616, "learning_rate": 9.20140563924797e-06, "loss": 0.6571, "step": 14901 }, { "epoch": 0.5397319811662441, "grad_norm": 1.426792201150603, "learning_rate": 9.200236323648176e-06, "loss": 0.739, "step": 14902 }, { "epoch": 0.5397681999275625, "grad_norm": 1.3932080429175169, "learning_rate": 9.199067019053836e-06, "loss": 0.7434, "step": 14903 }, { "epoch": 0.5398044186888808, "grad_norm": 1.3383923919232126, "learning_rate": 9.197897725481035e-06, "loss": 0.6969, "step": 14904 }, { "epoch": 0.5398406374501992, "grad_norm": 1.4263379512518806, "learning_rate": 9.196728442945867e-06, "loss": 0.7511, "step": 14905 }, { "epoch": 0.5398768562115176, "grad_norm": 1.3659116091245724, "learning_rate": 9.19555917146442e-06, "loss": 0.7195, "step": 14906 }, { "epoch": 0.5399130749728359, "grad_norm": 1.378104644877001, "learning_rate": 9.194389911052786e-06, "loss": 0.7486, "step": 14907 }, { "epoch": 0.5399492937341543, "grad_norm": 1.5271796842052698, "learning_rate": 9.193220661727056e-06, "loss": 0.7047, "step": 14908 }, { "epoch": 0.5399855124954727, "grad_norm": 1.880051647382651, "learning_rate": 9.192051423503316e-06, "loss": 0.7582, "step": 14909 }, { "epoch": 0.540021731256791, "grad_norm": 1.4505784925293135, "learning_rate": 9.190882196397663e-06, "loss": 0.6961, "step": 14910 }, { "epoch": 0.5400579500181094, "grad_norm": 1.3395481958355295, "learning_rate": 9.189712980426178e-06, "loss": 0.7788, "step": 14911 }, { "epoch": 0.5400941687794277, "grad_norm": 1.2929237893086796, "learning_rate": 9.188543775604957e-06, "loss": 0.7195, "step": 14912 }, { "epoch": 0.5401303875407462, "grad_norm": 1.4615482587438602, "learning_rate": 9.187374581950083e-06, "loss": 0.7131, "step": 14913 }, { "epoch": 0.5401666063020645, "grad_norm": 1.0901424425279507, "learning_rate": 9.186205399477654e-06, "loss": 0.7075, "step": 14914 }, { "epoch": 0.5402028250633828, "grad_norm": 1.5057953079982924, "learning_rate": 9.185036228203749e-06, "loss": 0.7801, "step": 14915 }, { "epoch": 0.5402390438247012, "grad_norm": 0.9925342987661906, "learning_rate": 9.183867068144465e-06, "loss": 0.8143, "step": 14916 }, { "epoch": 0.5402752625860195, "grad_norm": 1.444347892876822, "learning_rate": 9.182697919315884e-06, "loss": 0.7785, "step": 14917 }, { "epoch": 0.540311481347338, "grad_norm": 1.2955945301090135, "learning_rate": 9.1815287817341e-06, "loss": 0.6979, "step": 14918 }, { "epoch": 0.5403477001086563, "grad_norm": 1.0849925108173166, "learning_rate": 9.180359655415198e-06, "loss": 0.7931, "step": 14919 }, { "epoch": 0.5403839188699746, "grad_norm": 0.9541360568597087, "learning_rate": 9.179190540375268e-06, "loss": 0.6917, "step": 14920 }, { "epoch": 0.540420137631293, "grad_norm": 1.408883741791694, "learning_rate": 9.178021436630397e-06, "loss": 0.7738, "step": 14921 }, { "epoch": 0.5404563563926114, "grad_norm": 1.433534809376896, "learning_rate": 9.176852344196674e-06, "loss": 0.7638, "step": 14922 }, { "epoch": 0.5404925751539298, "grad_norm": 1.0261662544685397, "learning_rate": 9.175683263090188e-06, "loss": 0.6913, "step": 14923 }, { "epoch": 0.5405287939152481, "grad_norm": 1.4289213724363328, "learning_rate": 9.17451419332702e-06, "loss": 0.8053, "step": 14924 }, { "epoch": 0.5405650126765664, "grad_norm": 1.386589953660373, "learning_rate": 9.173345134923265e-06, "loss": 0.7703, "step": 14925 }, { "epoch": 0.5406012314378849, "grad_norm": 1.2603863997871252, "learning_rate": 9.172176087895006e-06, "loss": 0.722, "step": 14926 }, { "epoch": 0.5406374501992032, "grad_norm": 0.9533120564406297, "learning_rate": 9.171007052258331e-06, "loss": 0.7422, "step": 14927 }, { "epoch": 0.5406736689605216, "grad_norm": 1.3998394324300703, "learning_rate": 9.16983802802933e-06, "loss": 0.7047, "step": 14928 }, { "epoch": 0.5407098877218399, "grad_norm": 1.3324351229829645, "learning_rate": 9.168669015224085e-06, "loss": 0.7408, "step": 14929 }, { "epoch": 0.5407461064831582, "grad_norm": 0.9956242748170938, "learning_rate": 9.167500013858685e-06, "loss": 0.737, "step": 14930 }, { "epoch": 0.5407823252444767, "grad_norm": 1.1722280998533563, "learning_rate": 9.166331023949217e-06, "loss": 0.709, "step": 14931 }, { "epoch": 0.540818544005795, "grad_norm": 1.2711323303470732, "learning_rate": 9.165162045511766e-06, "loss": 0.6566, "step": 14932 }, { "epoch": 0.5408547627671134, "grad_norm": 1.5811182495874512, "learning_rate": 9.16399307856242e-06, "loss": 0.7425, "step": 14933 }, { "epoch": 0.5408909815284317, "grad_norm": 1.3436462575765071, "learning_rate": 9.162824123117263e-06, "loss": 0.6846, "step": 14934 }, { "epoch": 0.54092720028975, "grad_norm": 1.4374449657188424, "learning_rate": 9.16165517919238e-06, "loss": 0.724, "step": 14935 }, { "epoch": 0.5409634190510685, "grad_norm": 1.2461530243817762, "learning_rate": 9.16048624680386e-06, "loss": 0.6606, "step": 14936 }, { "epoch": 0.5409996378123868, "grad_norm": 0.9679582887642573, "learning_rate": 9.159317325967787e-06, "loss": 0.7142, "step": 14937 }, { "epoch": 0.5410358565737052, "grad_norm": 1.4605837313308982, "learning_rate": 9.158148416700246e-06, "loss": 0.8165, "step": 14938 }, { "epoch": 0.5410720753350236, "grad_norm": 1.3920630329002865, "learning_rate": 9.156979519017324e-06, "loss": 0.799, "step": 14939 }, { "epoch": 0.5411082940963419, "grad_norm": 1.5518567988163319, "learning_rate": 9.155810632935102e-06, "loss": 0.6847, "step": 14940 }, { "epoch": 0.5411445128576603, "grad_norm": 1.2888657795470169, "learning_rate": 9.15464175846967e-06, "loss": 0.7515, "step": 14941 }, { "epoch": 0.5411807316189786, "grad_norm": 1.3073745421214953, "learning_rate": 9.153472895637107e-06, "loss": 0.7216, "step": 14942 }, { "epoch": 0.541216950380297, "grad_norm": 1.2720180430158357, "learning_rate": 9.152304044453505e-06, "loss": 0.7294, "step": 14943 }, { "epoch": 0.5412531691416154, "grad_norm": 1.3369159162075213, "learning_rate": 9.151135204934941e-06, "loss": 0.7464, "step": 14944 }, { "epoch": 0.5412893879029337, "grad_norm": 1.0274145239984454, "learning_rate": 9.149966377097504e-06, "loss": 0.7417, "step": 14945 }, { "epoch": 0.5413256066642521, "grad_norm": 1.3634203536542173, "learning_rate": 9.148797560957276e-06, "loss": 0.7622, "step": 14946 }, { "epoch": 0.5413618254255704, "grad_norm": 1.0554444126885252, "learning_rate": 9.147628756530342e-06, "loss": 0.7437, "step": 14947 }, { "epoch": 0.5413980441868889, "grad_norm": 1.0348667556443027, "learning_rate": 9.146459963832783e-06, "loss": 0.6984, "step": 14948 }, { "epoch": 0.5414342629482072, "grad_norm": 1.3531680968306425, "learning_rate": 9.145291182880685e-06, "loss": 0.7196, "step": 14949 }, { "epoch": 0.5414704817095255, "grad_norm": 1.3169666040009884, "learning_rate": 9.144122413690135e-06, "loss": 0.6352, "step": 14950 }, { "epoch": 0.5415067004708439, "grad_norm": 1.5070644484759574, "learning_rate": 9.14295365627721e-06, "loss": 0.8224, "step": 14951 }, { "epoch": 0.5415429192321622, "grad_norm": 1.5185952202640374, "learning_rate": 9.141784910657997e-06, "loss": 0.8286, "step": 14952 }, { "epoch": 0.5415791379934807, "grad_norm": 1.374695016621884, "learning_rate": 9.140616176848576e-06, "loss": 0.7205, "step": 14953 }, { "epoch": 0.541615356754799, "grad_norm": 1.376234401318678, "learning_rate": 9.139447454865034e-06, "loss": 0.7534, "step": 14954 }, { "epoch": 0.5416515755161173, "grad_norm": 1.2277112865036894, "learning_rate": 9.138278744723448e-06, "loss": 0.7264, "step": 14955 }, { "epoch": 0.5416877942774357, "grad_norm": 1.1179474021031675, "learning_rate": 9.137110046439906e-06, "loss": 0.7213, "step": 14956 }, { "epoch": 0.5417240130387541, "grad_norm": 1.3670031859272511, "learning_rate": 9.135941360030487e-06, "loss": 0.7453, "step": 14957 }, { "epoch": 0.5417602318000725, "grad_norm": 1.1193765362797943, "learning_rate": 9.134772685511273e-06, "loss": 0.706, "step": 14958 }, { "epoch": 0.5417964505613908, "grad_norm": 1.3164215292171235, "learning_rate": 9.133604022898351e-06, "loss": 0.7316, "step": 14959 }, { "epoch": 0.5418326693227091, "grad_norm": 1.3769728097793685, "learning_rate": 9.132435372207795e-06, "loss": 0.7729, "step": 14960 }, { "epoch": 0.5418688880840276, "grad_norm": 1.5456468992910355, "learning_rate": 9.131266733455692e-06, "loss": 0.7296, "step": 14961 }, { "epoch": 0.5419051068453459, "grad_norm": 1.3094359462033063, "learning_rate": 9.130098106658121e-06, "loss": 0.7303, "step": 14962 }, { "epoch": 0.5419413256066643, "grad_norm": 1.3801929519504008, "learning_rate": 9.128929491831167e-06, "loss": 0.7276, "step": 14963 }, { "epoch": 0.5419775443679826, "grad_norm": 1.3008029731162347, "learning_rate": 9.127760888990907e-06, "loss": 0.7547, "step": 14964 }, { "epoch": 0.5420137631293009, "grad_norm": 1.1722009845168802, "learning_rate": 9.126592298153425e-06, "loss": 0.6659, "step": 14965 }, { "epoch": 0.5420499818906194, "grad_norm": 1.196265450750613, "learning_rate": 9.125423719334797e-06, "loss": 0.6819, "step": 14966 }, { "epoch": 0.5420862006519377, "grad_norm": 1.4009027967679553, "learning_rate": 9.124255152551112e-06, "loss": 0.7484, "step": 14967 }, { "epoch": 0.5421224194132561, "grad_norm": 1.3635360812977904, "learning_rate": 9.123086597818442e-06, "loss": 0.7882, "step": 14968 }, { "epoch": 0.5421586381745744, "grad_norm": 1.3534717938040433, "learning_rate": 9.121918055152869e-06, "loss": 0.6357, "step": 14969 }, { "epoch": 0.5421948569358928, "grad_norm": 1.4322325632876614, "learning_rate": 9.12074952457048e-06, "loss": 0.7412, "step": 14970 }, { "epoch": 0.5422310756972112, "grad_norm": 1.8833372366247265, "learning_rate": 9.119581006087348e-06, "loss": 0.7679, "step": 14971 }, { "epoch": 0.5422672944585295, "grad_norm": 1.285026269339197, "learning_rate": 9.118412499719556e-06, "loss": 0.7372, "step": 14972 }, { "epoch": 0.5423035132198479, "grad_norm": 1.4736154882627386, "learning_rate": 9.11724400548318e-06, "loss": 0.7636, "step": 14973 }, { "epoch": 0.5423397319811663, "grad_norm": 1.424274314491197, "learning_rate": 9.116075523394305e-06, "loss": 0.6659, "step": 14974 }, { "epoch": 0.5423759507424846, "grad_norm": 1.3444347134145815, "learning_rate": 9.114907053469005e-06, "loss": 0.6682, "step": 14975 }, { "epoch": 0.542412169503803, "grad_norm": 1.348175029140256, "learning_rate": 9.113738595723364e-06, "loss": 0.7508, "step": 14976 }, { "epoch": 0.5424483882651213, "grad_norm": 1.3358534304060847, "learning_rate": 9.112570150173457e-06, "loss": 0.7593, "step": 14977 }, { "epoch": 0.5424846070264397, "grad_norm": 1.5766504156875625, "learning_rate": 9.111401716835365e-06, "loss": 0.8282, "step": 14978 }, { "epoch": 0.5425208257877581, "grad_norm": 1.348128169816048, "learning_rate": 9.110233295725167e-06, "loss": 0.7381, "step": 14979 }, { "epoch": 0.5425570445490764, "grad_norm": 1.3035514005279092, "learning_rate": 9.109064886858938e-06, "loss": 0.7394, "step": 14980 }, { "epoch": 0.5425932633103948, "grad_norm": 1.3498336416579817, "learning_rate": 9.107896490252762e-06, "loss": 0.7294, "step": 14981 }, { "epoch": 0.5426294820717131, "grad_norm": 1.3956476712734398, "learning_rate": 9.10672810592271e-06, "loss": 0.7476, "step": 14982 }, { "epoch": 0.5426657008330316, "grad_norm": 1.4336606277748878, "learning_rate": 9.10555973388487e-06, "loss": 0.6892, "step": 14983 }, { "epoch": 0.5427019195943499, "grad_norm": 1.8819540795387115, "learning_rate": 9.104391374155311e-06, "loss": 0.7527, "step": 14984 }, { "epoch": 0.5427381383556682, "grad_norm": 1.4092408297169168, "learning_rate": 9.103223026750114e-06, "loss": 0.7071, "step": 14985 }, { "epoch": 0.5427743571169866, "grad_norm": 1.2939522142046276, "learning_rate": 9.102054691685354e-06, "loss": 0.6833, "step": 14986 }, { "epoch": 0.542810575878305, "grad_norm": 1.3924870715395075, "learning_rate": 9.100886368977114e-06, "loss": 0.7563, "step": 14987 }, { "epoch": 0.5428467946396234, "grad_norm": 1.5034613812117288, "learning_rate": 9.099718058641466e-06, "loss": 0.7555, "step": 14988 }, { "epoch": 0.5428830134009417, "grad_norm": 1.5029805393551754, "learning_rate": 9.098549760694488e-06, "loss": 0.7151, "step": 14989 }, { "epoch": 0.54291923216226, "grad_norm": 1.3279859063296684, "learning_rate": 9.097381475152258e-06, "loss": 0.7025, "step": 14990 }, { "epoch": 0.5429554509235784, "grad_norm": 1.311037066438729, "learning_rate": 9.096213202030851e-06, "loss": 0.6907, "step": 14991 }, { "epoch": 0.5429916696848968, "grad_norm": 1.1005033827640969, "learning_rate": 9.095044941346346e-06, "loss": 0.7312, "step": 14992 }, { "epoch": 0.5430278884462152, "grad_norm": 1.4318409502459162, "learning_rate": 9.093876693114817e-06, "loss": 0.6555, "step": 14993 }, { "epoch": 0.5430641072075335, "grad_norm": 0.9500571214564684, "learning_rate": 9.092708457352342e-06, "loss": 0.6514, "step": 14994 }, { "epoch": 0.5431003259688518, "grad_norm": 1.3526449058791146, "learning_rate": 9.091540234074994e-06, "loss": 0.696, "step": 14995 }, { "epoch": 0.5431365447301703, "grad_norm": 1.3919615363617164, "learning_rate": 9.090372023298853e-06, "loss": 0.7101, "step": 14996 }, { "epoch": 0.5431727634914886, "grad_norm": 1.3572826547377175, "learning_rate": 9.08920382503999e-06, "loss": 0.728, "step": 14997 }, { "epoch": 0.543208982252807, "grad_norm": 1.344992558690717, "learning_rate": 9.088035639314484e-06, "loss": 0.7191, "step": 14998 }, { "epoch": 0.5432452010141253, "grad_norm": 1.1061272981661956, "learning_rate": 9.086867466138407e-06, "loss": 0.7014, "step": 14999 }, { "epoch": 0.5432814197754436, "grad_norm": 1.382418178584128, "learning_rate": 9.085699305527836e-06, "loss": 0.7143, "step": 15000 }, { "epoch": 0.5433176385367621, "grad_norm": 1.8555595815536177, "learning_rate": 9.084531157498848e-06, "loss": 0.7386, "step": 15001 }, { "epoch": 0.5433538572980804, "grad_norm": 1.2051765544283954, "learning_rate": 9.083363022067515e-06, "loss": 0.7263, "step": 15002 }, { "epoch": 0.5433900760593988, "grad_norm": 1.5284294016359714, "learning_rate": 9.082194899249914e-06, "loss": 0.8285, "step": 15003 }, { "epoch": 0.5434262948207171, "grad_norm": 1.4226353689878706, "learning_rate": 9.081026789062113e-06, "loss": 0.7024, "step": 15004 }, { "epoch": 0.5434625135820355, "grad_norm": 1.4326658071629148, "learning_rate": 9.079858691520196e-06, "loss": 0.7108, "step": 15005 }, { "epoch": 0.5434987323433539, "grad_norm": 1.2023954753550368, "learning_rate": 9.078690606640229e-06, "loss": 0.7324, "step": 15006 }, { "epoch": 0.5435349511046722, "grad_norm": 1.0110508714034143, "learning_rate": 9.07752253443829e-06, "loss": 0.706, "step": 15007 }, { "epoch": 0.5435711698659906, "grad_norm": 1.3607919545639047, "learning_rate": 9.07635447493045e-06, "loss": 0.7399, "step": 15008 }, { "epoch": 0.543607388627309, "grad_norm": 1.5033709254497851, "learning_rate": 9.075186428132785e-06, "loss": 0.6866, "step": 15009 }, { "epoch": 0.5436436073886273, "grad_norm": 1.322871745082462, "learning_rate": 9.074018394061368e-06, "loss": 0.6709, "step": 15010 }, { "epoch": 0.5436798261499457, "grad_norm": 1.304806311988099, "learning_rate": 9.072850372732268e-06, "loss": 0.7248, "step": 15011 }, { "epoch": 0.543716044911264, "grad_norm": 1.4224648693092707, "learning_rate": 9.071682364161567e-06, "loss": 0.6626, "step": 15012 }, { "epoch": 0.5437522636725824, "grad_norm": 1.2980774036414051, "learning_rate": 9.070514368365329e-06, "loss": 0.7602, "step": 15013 }, { "epoch": 0.5437884824339008, "grad_norm": 1.2286216295142276, "learning_rate": 9.069346385359632e-06, "loss": 0.7292, "step": 15014 }, { "epoch": 0.5438247011952191, "grad_norm": 1.339040475116704, "learning_rate": 9.068178415160543e-06, "loss": 0.7387, "step": 15015 }, { "epoch": 0.5438609199565375, "grad_norm": 1.3574436563349759, "learning_rate": 9.067010457784142e-06, "loss": 0.7836, "step": 15016 }, { "epoch": 0.5438971387178558, "grad_norm": 1.4503163316111387, "learning_rate": 9.065842513246493e-06, "loss": 0.7983, "step": 15017 }, { "epoch": 0.5439333574791743, "grad_norm": 1.4658833222694183, "learning_rate": 9.064674581563676e-06, "loss": 0.7144, "step": 15018 }, { "epoch": 0.5439695762404926, "grad_norm": 1.288405526969818, "learning_rate": 9.063506662751755e-06, "loss": 0.7112, "step": 15019 }, { "epoch": 0.5440057950018109, "grad_norm": 1.2782017847397689, "learning_rate": 9.062338756826805e-06, "loss": 0.6925, "step": 15020 }, { "epoch": 0.5440420137631293, "grad_norm": 1.3610961345592163, "learning_rate": 9.061170863804901e-06, "loss": 0.7472, "step": 15021 }, { "epoch": 0.5440782325244476, "grad_norm": 1.2984489583004488, "learning_rate": 9.060002983702108e-06, "loss": 0.6231, "step": 15022 }, { "epoch": 0.5441144512857661, "grad_norm": 1.432611330874589, "learning_rate": 9.058835116534501e-06, "loss": 0.7226, "step": 15023 }, { "epoch": 0.5441506700470844, "grad_norm": 1.0772974832199431, "learning_rate": 9.05766726231815e-06, "loss": 0.8073, "step": 15024 }, { "epoch": 0.5441868888084027, "grad_norm": 1.233376750090899, "learning_rate": 9.056499421069125e-06, "loss": 0.7088, "step": 15025 }, { "epoch": 0.5442231075697211, "grad_norm": 1.3709666028122793, "learning_rate": 9.055331592803497e-06, "loss": 0.7722, "step": 15026 }, { "epoch": 0.5442593263310395, "grad_norm": 1.5517393678849742, "learning_rate": 9.054163777537337e-06, "loss": 0.7408, "step": 15027 }, { "epoch": 0.5442955450923579, "grad_norm": 1.3740775789711408, "learning_rate": 9.052995975286713e-06, "loss": 0.7223, "step": 15028 }, { "epoch": 0.5443317638536762, "grad_norm": 1.4011715436732441, "learning_rate": 9.051828186067699e-06, "loss": 0.7109, "step": 15029 }, { "epoch": 0.5443679826149945, "grad_norm": 1.9436191452312197, "learning_rate": 9.05066040989636e-06, "loss": 0.7453, "step": 15030 }, { "epoch": 0.544404201376313, "grad_norm": 1.0839053743266471, "learning_rate": 9.049492646788768e-06, "loss": 0.6955, "step": 15031 }, { "epoch": 0.5444404201376313, "grad_norm": 1.3864314668220008, "learning_rate": 9.048324896760993e-06, "loss": 0.7278, "step": 15032 }, { "epoch": 0.5444766388989497, "grad_norm": 1.1494725030900046, "learning_rate": 9.047157159829103e-06, "loss": 0.7373, "step": 15033 }, { "epoch": 0.544512857660268, "grad_norm": 1.3414341680668598, "learning_rate": 9.04598943600917e-06, "loss": 0.7085, "step": 15034 }, { "epoch": 0.5445490764215863, "grad_norm": 1.189128671752328, "learning_rate": 9.044821725317258e-06, "loss": 0.6461, "step": 15035 }, { "epoch": 0.5445852951829048, "grad_norm": 1.5154078625448266, "learning_rate": 9.04365402776944e-06, "loss": 0.8118, "step": 15036 }, { "epoch": 0.5446215139442231, "grad_norm": 1.4333956656953526, "learning_rate": 9.042486343381782e-06, "loss": 0.7508, "step": 15037 }, { "epoch": 0.5446577327055415, "grad_norm": 1.45934527502685, "learning_rate": 9.041318672170355e-06, "loss": 0.7392, "step": 15038 }, { "epoch": 0.5446939514668598, "grad_norm": 1.3745667563441768, "learning_rate": 9.040151014151223e-06, "loss": 0.6589, "step": 15039 }, { "epoch": 0.5447301702281782, "grad_norm": 1.4337388167926255, "learning_rate": 9.038983369340458e-06, "loss": 0.8069, "step": 15040 }, { "epoch": 0.5447663889894966, "grad_norm": 1.0013591602137433, "learning_rate": 9.037815737754124e-06, "loss": 0.7599, "step": 15041 }, { "epoch": 0.5448026077508149, "grad_norm": 1.3069915478440088, "learning_rate": 9.036648119408292e-06, "loss": 0.7261, "step": 15042 }, { "epoch": 0.5448388265121333, "grad_norm": 1.3976716123347563, "learning_rate": 9.035480514319031e-06, "loss": 0.6942, "step": 15043 }, { "epoch": 0.5448750452734517, "grad_norm": 1.7173462423827632, "learning_rate": 9.034312922502402e-06, "loss": 0.7907, "step": 15044 }, { "epoch": 0.54491126403477, "grad_norm": 1.1130426345118243, "learning_rate": 9.033145343974479e-06, "loss": 0.7584, "step": 15045 }, { "epoch": 0.5449474827960884, "grad_norm": 1.4250214315777079, "learning_rate": 9.031977778751322e-06, "loss": 0.7607, "step": 15046 }, { "epoch": 0.5449837015574067, "grad_norm": 1.3875936931365853, "learning_rate": 9.030810226849005e-06, "loss": 0.6859, "step": 15047 }, { "epoch": 0.5450199203187251, "grad_norm": 1.0351462134288567, "learning_rate": 9.029642688283588e-06, "loss": 0.72, "step": 15048 }, { "epoch": 0.5450561390800435, "grad_norm": 1.4424052821057622, "learning_rate": 9.028475163071142e-06, "loss": 0.7178, "step": 15049 }, { "epoch": 0.5450923578413618, "grad_norm": 0.9306025922574972, "learning_rate": 9.02730765122773e-06, "loss": 0.7614, "step": 15050 }, { "epoch": 0.5451285766026802, "grad_norm": 1.3541663691783234, "learning_rate": 9.02614015276942e-06, "loss": 0.6912, "step": 15051 }, { "epoch": 0.5451647953639985, "grad_norm": 1.3115396860782067, "learning_rate": 9.024972667712277e-06, "loss": 0.7178, "step": 15052 }, { "epoch": 0.545201014125317, "grad_norm": 1.2363800705336774, "learning_rate": 9.023805196072368e-06, "loss": 0.7568, "step": 15053 }, { "epoch": 0.5452372328866353, "grad_norm": 1.371138195665104, "learning_rate": 9.022637737865757e-06, "loss": 0.5875, "step": 15054 }, { "epoch": 0.5452734516479536, "grad_norm": 1.4075376800720034, "learning_rate": 9.021470293108507e-06, "loss": 0.6843, "step": 15055 }, { "epoch": 0.545309670409272, "grad_norm": 1.2421314693862024, "learning_rate": 9.02030286181669e-06, "loss": 0.8113, "step": 15056 }, { "epoch": 0.5453458891705903, "grad_norm": 1.5081801361608314, "learning_rate": 9.019135444006362e-06, "loss": 0.7372, "step": 15057 }, { "epoch": 0.5453821079319088, "grad_norm": 1.3731539991055226, "learning_rate": 9.017968039693595e-06, "loss": 0.7586, "step": 15058 }, { "epoch": 0.5454183266932271, "grad_norm": 1.1975191308451525, "learning_rate": 9.01680064889445e-06, "loss": 0.761, "step": 15059 }, { "epoch": 0.5454545454545454, "grad_norm": 1.353781748284469, "learning_rate": 9.015633271624993e-06, "loss": 0.7469, "step": 15060 }, { "epoch": 0.5454907642158638, "grad_norm": 1.4210323611458904, "learning_rate": 9.014465907901284e-06, "loss": 0.7089, "step": 15061 }, { "epoch": 0.5455269829771822, "grad_norm": 1.3439131585027742, "learning_rate": 9.013298557739391e-06, "loss": 0.7319, "step": 15062 }, { "epoch": 0.5455632017385006, "grad_norm": 1.2273184905133028, "learning_rate": 9.01213122115538e-06, "loss": 0.7175, "step": 15063 }, { "epoch": 0.5455994204998189, "grad_norm": 1.33523216785063, "learning_rate": 9.010963898165312e-06, "loss": 0.7194, "step": 15064 }, { "epoch": 0.5456356392611372, "grad_norm": 1.1090287439800581, "learning_rate": 9.009796588785245e-06, "loss": 0.7905, "step": 15065 }, { "epoch": 0.5456718580224557, "grad_norm": 1.2988241019043247, "learning_rate": 9.008629293031251e-06, "loss": 0.7096, "step": 15066 }, { "epoch": 0.545708076783774, "grad_norm": 1.2408669848305705, "learning_rate": 9.007462010919387e-06, "loss": 0.6856, "step": 15067 }, { "epoch": 0.5457442955450924, "grad_norm": 1.1587032232312926, "learning_rate": 9.00629474246572e-06, "loss": 0.7149, "step": 15068 }, { "epoch": 0.5457805143064107, "grad_norm": 1.3846748023655153, "learning_rate": 9.005127487686306e-06, "loss": 0.7823, "step": 15069 }, { "epoch": 0.545816733067729, "grad_norm": 1.3876004236947983, "learning_rate": 9.003960246597215e-06, "loss": 0.7389, "step": 15070 }, { "epoch": 0.5458529518290475, "grad_norm": 1.4374177756324398, "learning_rate": 9.002793019214506e-06, "loss": 0.7212, "step": 15071 }, { "epoch": 0.5458891705903658, "grad_norm": 1.4088181701428604, "learning_rate": 9.001625805554242e-06, "loss": 0.7264, "step": 15072 }, { "epoch": 0.5459253893516842, "grad_norm": 1.335454473537449, "learning_rate": 9.000458605632482e-06, "loss": 0.7717, "step": 15073 }, { "epoch": 0.5459616081130025, "grad_norm": 1.3658314573928902, "learning_rate": 8.999291419465289e-06, "loss": 0.6401, "step": 15074 }, { "epoch": 0.5459978268743209, "grad_norm": 1.3711000095347492, "learning_rate": 8.99812424706873e-06, "loss": 0.8242, "step": 15075 }, { "epoch": 0.5460340456356393, "grad_norm": 1.2079030737806882, "learning_rate": 8.996957088458857e-06, "loss": 0.7406, "step": 15076 }, { "epoch": 0.5460702643969576, "grad_norm": 1.3860684838000903, "learning_rate": 8.995789943651738e-06, "loss": 0.7224, "step": 15077 }, { "epoch": 0.546106483158276, "grad_norm": 1.3176680949504382, "learning_rate": 8.99462281266343e-06, "loss": 0.6614, "step": 15078 }, { "epoch": 0.5461427019195944, "grad_norm": 1.411039728176671, "learning_rate": 8.99345569551e-06, "loss": 0.7278, "step": 15079 }, { "epoch": 0.5461789206809127, "grad_norm": 1.4749909570384439, "learning_rate": 8.992288592207499e-06, "loss": 0.7776, "step": 15080 }, { "epoch": 0.5462151394422311, "grad_norm": 1.301026590367854, "learning_rate": 8.991121502771995e-06, "loss": 0.8212, "step": 15081 }, { "epoch": 0.5462513582035494, "grad_norm": 1.4776226241944839, "learning_rate": 8.989954427219545e-06, "loss": 0.7414, "step": 15082 }, { "epoch": 0.5462875769648678, "grad_norm": 1.3346067047907486, "learning_rate": 8.988787365566209e-06, "loss": 0.7245, "step": 15083 }, { "epoch": 0.5463237957261862, "grad_norm": 1.3099928742373743, "learning_rate": 8.98762031782805e-06, "loss": 0.7418, "step": 15084 }, { "epoch": 0.5463600144875045, "grad_norm": 1.0058793307843255, "learning_rate": 8.98645328402112e-06, "loss": 0.7253, "step": 15085 }, { "epoch": 0.5463962332488229, "grad_norm": 0.8950611675565153, "learning_rate": 8.985286264161487e-06, "loss": 0.7372, "step": 15086 }, { "epoch": 0.5464324520101412, "grad_norm": 1.4515795770344841, "learning_rate": 8.984119258265205e-06, "loss": 0.7854, "step": 15087 }, { "epoch": 0.5464686707714597, "grad_norm": 1.4572314503251211, "learning_rate": 8.982952266348339e-06, "loss": 0.7657, "step": 15088 }, { "epoch": 0.546504889532778, "grad_norm": 1.388311778700503, "learning_rate": 8.981785288426938e-06, "loss": 0.6972, "step": 15089 }, { "epoch": 0.5465411082940963, "grad_norm": 1.3289135747248786, "learning_rate": 8.98061832451707e-06, "loss": 0.7587, "step": 15090 }, { "epoch": 0.5465773270554147, "grad_norm": 1.3877508619104164, "learning_rate": 8.979451374634787e-06, "loss": 0.7207, "step": 15091 }, { "epoch": 0.546613545816733, "grad_norm": 1.225528495323925, "learning_rate": 8.978284438796151e-06, "loss": 0.6836, "step": 15092 }, { "epoch": 0.5466497645780515, "grad_norm": 1.5311320877414103, "learning_rate": 8.977117517017218e-06, "loss": 0.7085, "step": 15093 }, { "epoch": 0.5466859833393698, "grad_norm": 1.3210949667096525, "learning_rate": 8.975950609314045e-06, "loss": 0.8199, "step": 15094 }, { "epoch": 0.5467222021006881, "grad_norm": 1.0348756480976458, "learning_rate": 8.974783715702695e-06, "loss": 0.7481, "step": 15095 }, { "epoch": 0.5467584208620065, "grad_norm": 1.2353126067401077, "learning_rate": 8.97361683619922e-06, "loss": 0.7555, "step": 15096 }, { "epoch": 0.5467946396233249, "grad_norm": 1.1320309522400012, "learning_rate": 8.972449970819682e-06, "loss": 0.694, "step": 15097 }, { "epoch": 0.5468308583846433, "grad_norm": 1.389531146693679, "learning_rate": 8.971283119580132e-06, "loss": 0.6993, "step": 15098 }, { "epoch": 0.5468670771459616, "grad_norm": 1.3306883911771077, "learning_rate": 8.970116282496633e-06, "loss": 0.7449, "step": 15099 }, { "epoch": 0.5469032959072799, "grad_norm": 1.035554649727831, "learning_rate": 8.968949459585236e-06, "loss": 0.7313, "step": 15100 }, { "epoch": 0.5469395146685984, "grad_norm": 1.3738090779012109, "learning_rate": 8.967782650862002e-06, "loss": 0.7477, "step": 15101 }, { "epoch": 0.5469757334299167, "grad_norm": 1.2853691155232556, "learning_rate": 8.966615856342984e-06, "loss": 0.7528, "step": 15102 }, { "epoch": 0.5470119521912351, "grad_norm": 1.3556283917259324, "learning_rate": 8.965449076044243e-06, "loss": 0.7309, "step": 15103 }, { "epoch": 0.5470481709525534, "grad_norm": 1.0479219350299391, "learning_rate": 8.964282309981828e-06, "loss": 0.629, "step": 15104 }, { "epoch": 0.5470843897138717, "grad_norm": 1.3447105286251075, "learning_rate": 8.9631155581718e-06, "loss": 0.735, "step": 15105 }, { "epoch": 0.5471206084751902, "grad_norm": 1.4271047693809853, "learning_rate": 8.961948820630215e-06, "loss": 0.7892, "step": 15106 }, { "epoch": 0.5471568272365085, "grad_norm": 1.3143490059980898, "learning_rate": 8.960782097373124e-06, "loss": 0.6501, "step": 15107 }, { "epoch": 0.5471930459978269, "grad_norm": 1.3041380300560592, "learning_rate": 8.959615388416584e-06, "loss": 0.6946, "step": 15108 }, { "epoch": 0.5472292647591452, "grad_norm": 1.38072900547855, "learning_rate": 8.95844869377665e-06, "loss": 0.8015, "step": 15109 }, { "epoch": 0.5472654835204636, "grad_norm": 1.4089240538214776, "learning_rate": 8.957282013469381e-06, "loss": 0.6611, "step": 15110 }, { "epoch": 0.547301702281782, "grad_norm": 1.0925285455498697, "learning_rate": 8.956115347510825e-06, "loss": 0.7589, "step": 15111 }, { "epoch": 0.5473379210431003, "grad_norm": 0.9798267669961305, "learning_rate": 8.95494869591704e-06, "loss": 0.6841, "step": 15112 }, { "epoch": 0.5473741398044187, "grad_norm": 1.2203175851432655, "learning_rate": 8.953782058704079e-06, "loss": 0.6306, "step": 15113 }, { "epoch": 0.547410358565737, "grad_norm": 1.4992309035334876, "learning_rate": 8.952615435887994e-06, "loss": 0.7699, "step": 15114 }, { "epoch": 0.5474465773270554, "grad_norm": 1.4436122385010188, "learning_rate": 8.951448827484843e-06, "loss": 0.7185, "step": 15115 }, { "epoch": 0.5474827960883738, "grad_norm": 1.312998713816843, "learning_rate": 8.950282233510677e-06, "loss": 0.7048, "step": 15116 }, { "epoch": 0.5475190148496921, "grad_norm": 1.021822438030664, "learning_rate": 8.949115653981551e-06, "loss": 0.691, "step": 15117 }, { "epoch": 0.5475552336110106, "grad_norm": 1.1294497028271806, "learning_rate": 8.947949088913516e-06, "loss": 0.7672, "step": 15118 }, { "epoch": 0.5475914523723289, "grad_norm": 1.2374496757949247, "learning_rate": 8.946782538322628e-06, "loss": 0.7034, "step": 15119 }, { "epoch": 0.5476276711336472, "grad_norm": 1.3935670243955536, "learning_rate": 8.945616002224934e-06, "loss": 0.6671, "step": 15120 }, { "epoch": 0.5476638898949656, "grad_norm": 1.2830600295849741, "learning_rate": 8.944449480636492e-06, "loss": 0.6993, "step": 15121 }, { "epoch": 0.5477001086562839, "grad_norm": 1.2387387174503675, "learning_rate": 8.943282973573351e-06, "loss": 0.7107, "step": 15122 }, { "epoch": 0.5477363274176024, "grad_norm": 1.2515058938630428, "learning_rate": 8.942116481051569e-06, "loss": 0.6598, "step": 15123 }, { "epoch": 0.5477725461789207, "grad_norm": 0.9591425119003009, "learning_rate": 8.94095000308719e-06, "loss": 0.7264, "step": 15124 }, { "epoch": 0.547808764940239, "grad_norm": 1.5238041986612734, "learning_rate": 8.939783539696269e-06, "loss": 0.7555, "step": 15125 }, { "epoch": 0.5478449837015574, "grad_norm": 0.9569474302170967, "learning_rate": 8.938617090894861e-06, "loss": 0.7267, "step": 15126 }, { "epoch": 0.5478812024628757, "grad_norm": 1.3476406552614926, "learning_rate": 8.937450656699013e-06, "loss": 0.7592, "step": 15127 }, { "epoch": 0.5479174212241942, "grad_norm": 0.9124150507440281, "learning_rate": 8.936284237124779e-06, "loss": 0.7225, "step": 15128 }, { "epoch": 0.5479536399855125, "grad_norm": 1.3163537027698498, "learning_rate": 8.935117832188205e-06, "loss": 0.7059, "step": 15129 }, { "epoch": 0.5479898587468308, "grad_norm": 1.335329616532842, "learning_rate": 8.93395144190535e-06, "loss": 0.7026, "step": 15130 }, { "epoch": 0.5480260775081492, "grad_norm": 1.3726491820175672, "learning_rate": 8.932785066292257e-06, "loss": 0.7167, "step": 15131 }, { "epoch": 0.5480622962694676, "grad_norm": 1.4020119157561453, "learning_rate": 8.931618705364981e-06, "loss": 0.7283, "step": 15132 }, { "epoch": 0.548098515030786, "grad_norm": 1.6915379664560422, "learning_rate": 8.930452359139568e-06, "loss": 0.7437, "step": 15133 }, { "epoch": 0.5481347337921043, "grad_norm": 1.0392825642140977, "learning_rate": 8.92928602763207e-06, "loss": 0.7048, "step": 15134 }, { "epoch": 0.5481709525534226, "grad_norm": 1.5633005368359822, "learning_rate": 8.928119710858539e-06, "loss": 0.788, "step": 15135 }, { "epoch": 0.5482071713147411, "grad_norm": 1.414771589822406, "learning_rate": 8.926953408835022e-06, "loss": 0.7681, "step": 15136 }, { "epoch": 0.5482433900760594, "grad_norm": 1.1025025990591615, "learning_rate": 8.92578712157757e-06, "loss": 0.7195, "step": 15137 }, { "epoch": 0.5482796088373778, "grad_norm": 1.0784029183464106, "learning_rate": 8.924620849102229e-06, "loss": 0.6554, "step": 15138 }, { "epoch": 0.5483158275986961, "grad_norm": 1.0720780498217093, "learning_rate": 8.923454591425052e-06, "loss": 0.6541, "step": 15139 }, { "epoch": 0.5483520463600144, "grad_norm": 1.4619042148398567, "learning_rate": 8.922288348562084e-06, "loss": 0.7181, "step": 15140 }, { "epoch": 0.5483882651213329, "grad_norm": 1.8274439170623231, "learning_rate": 8.921122120529377e-06, "loss": 0.7385, "step": 15141 }, { "epoch": 0.5484244838826512, "grad_norm": 1.3848430554057767, "learning_rate": 8.919955907342977e-06, "loss": 0.7069, "step": 15142 }, { "epoch": 0.5484607026439696, "grad_norm": 1.4703622412806214, "learning_rate": 8.918789709018934e-06, "loss": 0.6686, "step": 15143 }, { "epoch": 0.5484969214052879, "grad_norm": 1.3503796515765147, "learning_rate": 8.917623525573292e-06, "loss": 0.7553, "step": 15144 }, { "epoch": 0.5485331401666063, "grad_norm": 1.2799768839132748, "learning_rate": 8.916457357022101e-06, "loss": 0.6364, "step": 15145 }, { "epoch": 0.5485693589279247, "grad_norm": 1.3120084634051241, "learning_rate": 8.915291203381413e-06, "loss": 0.7005, "step": 15146 }, { "epoch": 0.548605577689243, "grad_norm": 1.3655855596235156, "learning_rate": 8.914125064667267e-06, "loss": 0.7202, "step": 15147 }, { "epoch": 0.5486417964505614, "grad_norm": 1.1428650066913137, "learning_rate": 8.912958940895717e-06, "loss": 0.6779, "step": 15148 }, { "epoch": 0.5486780152118798, "grad_norm": 1.411911987554345, "learning_rate": 8.911792832082805e-06, "loss": 0.7209, "step": 15149 }, { "epoch": 0.5487142339731981, "grad_norm": 1.3155525662718002, "learning_rate": 8.910626738244583e-06, "loss": 0.6913, "step": 15150 }, { "epoch": 0.5487504527345165, "grad_norm": 1.2690871650423703, "learning_rate": 8.90946065939709e-06, "loss": 0.7322, "step": 15151 }, { "epoch": 0.5487866714958348, "grad_norm": 1.548579666297662, "learning_rate": 8.90829459555638e-06, "loss": 0.7445, "step": 15152 }, { "epoch": 0.5488228902571533, "grad_norm": 1.2711377194559057, "learning_rate": 8.907128546738493e-06, "loss": 0.811, "step": 15153 }, { "epoch": 0.5488591090184716, "grad_norm": 1.394611990752408, "learning_rate": 8.90596251295948e-06, "loss": 0.711, "step": 15154 }, { "epoch": 0.5488953277797899, "grad_norm": 1.1217861958127313, "learning_rate": 8.904796494235382e-06, "loss": 0.8284, "step": 15155 }, { "epoch": 0.5489315465411083, "grad_norm": 1.5008822171325324, "learning_rate": 8.903630490582246e-06, "loss": 0.7667, "step": 15156 }, { "epoch": 0.5489677653024266, "grad_norm": 1.3327347493217288, "learning_rate": 8.90246450201612e-06, "loss": 0.7149, "step": 15157 }, { "epoch": 0.5490039840637451, "grad_norm": 1.0594889829335614, "learning_rate": 8.901298528553045e-06, "loss": 0.701, "step": 15158 }, { "epoch": 0.5490402028250634, "grad_norm": 1.3566050006364077, "learning_rate": 8.90013257020907e-06, "loss": 0.7267, "step": 15159 }, { "epoch": 0.5490764215863817, "grad_norm": 1.3868173256109428, "learning_rate": 8.898966627000235e-06, "loss": 0.7139, "step": 15160 }, { "epoch": 0.5491126403477001, "grad_norm": 1.0723458637465868, "learning_rate": 8.897800698942589e-06, "loss": 0.7283, "step": 15161 }, { "epoch": 0.5491488591090185, "grad_norm": 1.010270467851051, "learning_rate": 8.896634786052172e-06, "loss": 0.7093, "step": 15162 }, { "epoch": 0.5491850778703369, "grad_norm": 1.3114109516548116, "learning_rate": 8.895468888345032e-06, "loss": 0.6864, "step": 15163 }, { "epoch": 0.5492212966316552, "grad_norm": 1.4942974718832305, "learning_rate": 8.894303005837208e-06, "loss": 0.7598, "step": 15164 }, { "epoch": 0.5492575153929735, "grad_norm": 0.9242156966223264, "learning_rate": 8.893137138544747e-06, "loss": 0.6987, "step": 15165 }, { "epoch": 0.549293734154292, "grad_norm": 1.3844185388803274, "learning_rate": 8.891971286483691e-06, "loss": 0.7821, "step": 15166 }, { "epoch": 0.5493299529156103, "grad_norm": 1.3871889123267065, "learning_rate": 8.890805449670083e-06, "loss": 0.7596, "step": 15167 }, { "epoch": 0.5493661716769287, "grad_norm": 1.369814496037454, "learning_rate": 8.889639628119971e-06, "loss": 0.7763, "step": 15168 }, { "epoch": 0.549402390438247, "grad_norm": 1.3342314040422367, "learning_rate": 8.888473821849388e-06, "loss": 0.7914, "step": 15169 }, { "epoch": 0.5494386091995653, "grad_norm": 1.3804578912391028, "learning_rate": 8.887308030874387e-06, "loss": 0.7964, "step": 15170 }, { "epoch": 0.5494748279608838, "grad_norm": 1.316315663387565, "learning_rate": 8.886142255211e-06, "loss": 0.781, "step": 15171 }, { "epoch": 0.5495110467222021, "grad_norm": 1.3598458088654646, "learning_rate": 8.884976494875277e-06, "loss": 0.7047, "step": 15172 }, { "epoch": 0.5495472654835205, "grad_norm": 1.3155777634762287, "learning_rate": 8.883810749883257e-06, "loss": 0.7234, "step": 15173 }, { "epoch": 0.5495834842448388, "grad_norm": 1.291510553469074, "learning_rate": 8.882645020250982e-06, "loss": 0.7084, "step": 15174 }, { "epoch": 0.5496197030061571, "grad_norm": 0.998189444609595, "learning_rate": 8.881479305994493e-06, "loss": 0.7049, "step": 15175 }, { "epoch": 0.5496559217674756, "grad_norm": 1.3131998344955236, "learning_rate": 8.880313607129831e-06, "loss": 0.6959, "step": 15176 }, { "epoch": 0.5496921405287939, "grad_norm": 1.4022469181643098, "learning_rate": 8.879147923673039e-06, "loss": 0.6372, "step": 15177 }, { "epoch": 0.5497283592901123, "grad_norm": 1.3748159041491126, "learning_rate": 8.877982255640155e-06, "loss": 0.6925, "step": 15178 }, { "epoch": 0.5497645780514306, "grad_norm": 1.3021228089706438, "learning_rate": 8.876816603047224e-06, "loss": 0.746, "step": 15179 }, { "epoch": 0.549800796812749, "grad_norm": 1.2851044918351424, "learning_rate": 8.87565096591028e-06, "loss": 0.7203, "step": 15180 }, { "epoch": 0.5498370155740674, "grad_norm": 1.4619876967891225, "learning_rate": 8.874485344245369e-06, "loss": 0.7121, "step": 15181 }, { "epoch": 0.5498732343353857, "grad_norm": 1.4779773964585121, "learning_rate": 8.873319738068527e-06, "loss": 0.7496, "step": 15182 }, { "epoch": 0.5499094530967041, "grad_norm": 1.233137104939439, "learning_rate": 8.872154147395797e-06, "loss": 0.6804, "step": 15183 }, { "epoch": 0.5499456718580225, "grad_norm": 1.4063563637774807, "learning_rate": 8.870988572243214e-06, "loss": 0.686, "step": 15184 }, { "epoch": 0.5499818906193408, "grad_norm": 1.0177180652529718, "learning_rate": 8.869823012626824e-06, "loss": 0.7028, "step": 15185 }, { "epoch": 0.5500181093806592, "grad_norm": 1.1977876853028595, "learning_rate": 8.86865746856266e-06, "loss": 0.7801, "step": 15186 }, { "epoch": 0.5500543281419775, "grad_norm": 1.2365754186772688, "learning_rate": 8.867491940066762e-06, "loss": 0.7237, "step": 15187 }, { "epoch": 0.550090546903296, "grad_norm": 1.596480907817237, "learning_rate": 8.866326427155173e-06, "loss": 0.7182, "step": 15188 }, { "epoch": 0.5501267656646143, "grad_norm": 1.2938065760388737, "learning_rate": 8.865160929843926e-06, "loss": 0.711, "step": 15189 }, { "epoch": 0.5501629844259326, "grad_norm": 1.33197445063155, "learning_rate": 8.863995448149064e-06, "loss": 0.7323, "step": 15190 }, { "epoch": 0.550199203187251, "grad_norm": 1.296495374519119, "learning_rate": 8.862829982086622e-06, "loss": 0.7795, "step": 15191 }, { "epoch": 0.5502354219485693, "grad_norm": 1.1952144368896676, "learning_rate": 8.861664531672639e-06, "loss": 0.7081, "step": 15192 }, { "epoch": 0.5502716407098878, "grad_norm": 1.3416152009313802, "learning_rate": 8.860499096923151e-06, "loss": 0.7954, "step": 15193 }, { "epoch": 0.5503078594712061, "grad_norm": 1.0538686926672554, "learning_rate": 8.859333677854198e-06, "loss": 0.723, "step": 15194 }, { "epoch": 0.5503440782325244, "grad_norm": 1.313055270817245, "learning_rate": 8.858168274481816e-06, "loss": 0.7396, "step": 15195 }, { "epoch": 0.5503802969938428, "grad_norm": 1.2532226207705355, "learning_rate": 8.857002886822041e-06, "loss": 0.7269, "step": 15196 }, { "epoch": 0.5504165157551612, "grad_norm": 1.299678725452388, "learning_rate": 8.85583751489091e-06, "loss": 0.6741, "step": 15197 }, { "epoch": 0.5504527345164796, "grad_norm": 1.4238151267129486, "learning_rate": 8.85467215870446e-06, "loss": 0.6606, "step": 15198 }, { "epoch": 0.5504889532777979, "grad_norm": 1.3501402470525765, "learning_rate": 8.85350681827873e-06, "loss": 0.7214, "step": 15199 }, { "epoch": 0.5505251720391162, "grad_norm": 1.2902175619558447, "learning_rate": 8.852341493629752e-06, "loss": 0.6739, "step": 15200 }, { "epoch": 0.5505613908004346, "grad_norm": 1.2107791154863312, "learning_rate": 8.851176184773563e-06, "loss": 0.6971, "step": 15201 }, { "epoch": 0.550597609561753, "grad_norm": 1.4751868778181065, "learning_rate": 8.8500108917262e-06, "loss": 0.6933, "step": 15202 }, { "epoch": 0.5506338283230714, "grad_norm": 1.3012515476551527, "learning_rate": 8.848845614503698e-06, "loss": 0.8027, "step": 15203 }, { "epoch": 0.5506700470843897, "grad_norm": 1.1500480353032354, "learning_rate": 8.84768035312209e-06, "loss": 0.765, "step": 15204 }, { "epoch": 0.550706265845708, "grad_norm": 0.9257534228893021, "learning_rate": 8.846515107597415e-06, "loss": 0.7684, "step": 15205 }, { "epoch": 0.5507424846070265, "grad_norm": 1.3104214821477096, "learning_rate": 8.845349877945703e-06, "loss": 0.6903, "step": 15206 }, { "epoch": 0.5507787033683448, "grad_norm": 2.207543085888516, "learning_rate": 8.844184664182993e-06, "loss": 0.756, "step": 15207 }, { "epoch": 0.5508149221296632, "grad_norm": 1.3335373975795652, "learning_rate": 8.843019466325317e-06, "loss": 0.6877, "step": 15208 }, { "epoch": 0.5508511408909815, "grad_norm": 1.3751000508840963, "learning_rate": 8.84185428438871e-06, "loss": 0.7313, "step": 15209 }, { "epoch": 0.5508873596522998, "grad_norm": 1.2476360442574281, "learning_rate": 8.840689118389207e-06, "loss": 0.6923, "step": 15210 }, { "epoch": 0.5509235784136183, "grad_norm": 1.2693399637337166, "learning_rate": 8.83952396834284e-06, "loss": 0.8043, "step": 15211 }, { "epoch": 0.5509597971749366, "grad_norm": 1.3069087923671503, "learning_rate": 8.838358834265642e-06, "loss": 0.7532, "step": 15212 }, { "epoch": 0.550996015936255, "grad_norm": 1.3064419569707235, "learning_rate": 8.837193716173646e-06, "loss": 0.6931, "step": 15213 }, { "epoch": 0.5510322346975733, "grad_norm": 1.2956555059041706, "learning_rate": 8.83602861408289e-06, "loss": 0.7032, "step": 15214 }, { "epoch": 0.5510684534588917, "grad_norm": 1.3661191301399913, "learning_rate": 8.834863528009398e-06, "loss": 0.6864, "step": 15215 }, { "epoch": 0.5511046722202101, "grad_norm": 1.3300222225456917, "learning_rate": 8.833698457969213e-06, "loss": 0.7436, "step": 15216 }, { "epoch": 0.5511408909815284, "grad_norm": 1.3108088889478104, "learning_rate": 8.832533403978357e-06, "loss": 0.7262, "step": 15217 }, { "epoch": 0.5511771097428468, "grad_norm": 1.3533402545377413, "learning_rate": 8.83136836605287e-06, "loss": 0.7223, "step": 15218 }, { "epoch": 0.5512133285041652, "grad_norm": 0.8881496383107087, "learning_rate": 8.830203344208781e-06, "loss": 0.6736, "step": 15219 }, { "epoch": 0.5512495472654835, "grad_norm": 0.9423825858787896, "learning_rate": 8.82903833846212e-06, "loss": 0.736, "step": 15220 }, { "epoch": 0.5512857660268019, "grad_norm": 1.1224964818658394, "learning_rate": 8.827873348828922e-06, "loss": 0.7225, "step": 15221 }, { "epoch": 0.5513219847881202, "grad_norm": 1.187723420326025, "learning_rate": 8.826708375325215e-06, "loss": 0.725, "step": 15222 }, { "epoch": 0.5513582035494387, "grad_norm": 1.21083747663135, "learning_rate": 8.825543417967035e-06, "loss": 0.735, "step": 15223 }, { "epoch": 0.551394422310757, "grad_norm": 1.2708127435074141, "learning_rate": 8.824378476770407e-06, "loss": 0.754, "step": 15224 }, { "epoch": 0.5514306410720753, "grad_norm": 1.2458996259071002, "learning_rate": 8.823213551751366e-06, "loss": 0.7535, "step": 15225 }, { "epoch": 0.5514668598333937, "grad_norm": 1.3235451699824297, "learning_rate": 8.822048642925938e-06, "loss": 0.795, "step": 15226 }, { "epoch": 0.551503078594712, "grad_norm": 1.5300526374459726, "learning_rate": 8.820883750310159e-06, "loss": 0.7219, "step": 15227 }, { "epoch": 0.5515392973560305, "grad_norm": 1.483040461940069, "learning_rate": 8.819718873920052e-06, "loss": 0.7555, "step": 15228 }, { "epoch": 0.5515755161173488, "grad_norm": 1.4163710838619246, "learning_rate": 8.818554013771653e-06, "loss": 0.7746, "step": 15229 }, { "epoch": 0.5516117348786671, "grad_norm": 1.4636831657782527, "learning_rate": 8.817389169880988e-06, "loss": 0.7194, "step": 15230 }, { "epoch": 0.5516479536399855, "grad_norm": 1.478535549981975, "learning_rate": 8.816224342264085e-06, "loss": 0.8445, "step": 15231 }, { "epoch": 0.5516841724013039, "grad_norm": 1.0978926403237803, "learning_rate": 8.81505953093698e-06, "loss": 0.6992, "step": 15232 }, { "epoch": 0.5517203911626223, "grad_norm": 1.2152283925297294, "learning_rate": 8.813894735915693e-06, "loss": 0.6815, "step": 15233 }, { "epoch": 0.5517566099239406, "grad_norm": 1.199871048527344, "learning_rate": 8.812729957216259e-06, "loss": 0.6898, "step": 15234 }, { "epoch": 0.5517928286852589, "grad_norm": 1.1816492950056359, "learning_rate": 8.811565194854701e-06, "loss": 0.7688, "step": 15235 }, { "epoch": 0.5518290474465773, "grad_norm": 1.247647233415675, "learning_rate": 8.810400448847054e-06, "loss": 0.7348, "step": 15236 }, { "epoch": 0.5518652662078957, "grad_norm": 0.9567119257244291, "learning_rate": 8.809235719209339e-06, "loss": 0.7364, "step": 15237 }, { "epoch": 0.5519014849692141, "grad_norm": 1.335590779753527, "learning_rate": 8.808071005957586e-06, "loss": 0.7438, "step": 15238 }, { "epoch": 0.5519377037305324, "grad_norm": 1.3524191696517194, "learning_rate": 8.806906309107828e-06, "loss": 0.7661, "step": 15239 }, { "epoch": 0.5519739224918507, "grad_norm": 1.4353942426647246, "learning_rate": 8.805741628676083e-06, "loss": 0.7144, "step": 15240 }, { "epoch": 0.5520101412531692, "grad_norm": 1.354553992711623, "learning_rate": 8.804576964678386e-06, "loss": 0.771, "step": 15241 }, { "epoch": 0.5520463600144875, "grad_norm": 1.4206116927740817, "learning_rate": 8.803412317130759e-06, "loss": 0.7924, "step": 15242 }, { "epoch": 0.5520825787758059, "grad_norm": 1.3921950155701805, "learning_rate": 8.80224768604923e-06, "loss": 0.723, "step": 15243 }, { "epoch": 0.5521187975371242, "grad_norm": 1.3148331850820796, "learning_rate": 8.801083071449824e-06, "loss": 0.7295, "step": 15244 }, { "epoch": 0.5521550162984425, "grad_norm": 0.9919867693842662, "learning_rate": 8.79991847334857e-06, "loss": 0.749, "step": 15245 }, { "epoch": 0.552191235059761, "grad_norm": 1.2382620169171816, "learning_rate": 8.798753891761492e-06, "loss": 0.6915, "step": 15246 }, { "epoch": 0.5522274538210793, "grad_norm": 1.2982834900228266, "learning_rate": 8.797589326704616e-06, "loss": 0.7336, "step": 15247 }, { "epoch": 0.5522636725823977, "grad_norm": 1.0795601414114635, "learning_rate": 8.796424778193966e-06, "loss": 0.6699, "step": 15248 }, { "epoch": 0.552299891343716, "grad_norm": 1.3007754549120785, "learning_rate": 8.795260246245569e-06, "loss": 0.6624, "step": 15249 }, { "epoch": 0.5523361101050344, "grad_norm": 1.3703354258445022, "learning_rate": 8.794095730875452e-06, "loss": 0.7262, "step": 15250 }, { "epoch": 0.5523723288663528, "grad_norm": 1.4555350181576188, "learning_rate": 8.792931232099633e-06, "loss": 0.7464, "step": 15251 }, { "epoch": 0.5524085476276711, "grad_norm": 1.3179365327692871, "learning_rate": 8.791766749934145e-06, "loss": 0.7025, "step": 15252 }, { "epoch": 0.5524447663889895, "grad_norm": 1.4057188534337701, "learning_rate": 8.790602284395006e-06, "loss": 0.7102, "step": 15253 }, { "epoch": 0.5524809851503079, "grad_norm": 1.0564812251437528, "learning_rate": 8.789437835498243e-06, "loss": 0.7361, "step": 15254 }, { "epoch": 0.5525172039116262, "grad_norm": 1.0158144591041727, "learning_rate": 8.788273403259879e-06, "loss": 0.7241, "step": 15255 }, { "epoch": 0.5525534226729446, "grad_norm": 1.3587452221644292, "learning_rate": 8.787108987695937e-06, "loss": 0.8005, "step": 15256 }, { "epoch": 0.5525896414342629, "grad_norm": 1.524532706942326, "learning_rate": 8.785944588822443e-06, "loss": 0.7552, "step": 15257 }, { "epoch": 0.5526258601955814, "grad_norm": 1.3386907113405535, "learning_rate": 8.784780206655419e-06, "loss": 0.7775, "step": 15258 }, { "epoch": 0.5526620789568997, "grad_norm": 1.033743391905612, "learning_rate": 8.783615841210883e-06, "loss": 0.7517, "step": 15259 }, { "epoch": 0.552698297718218, "grad_norm": 1.258635641060333, "learning_rate": 8.782451492504861e-06, "loss": 0.7131, "step": 15260 }, { "epoch": 0.5527345164795364, "grad_norm": 1.23481248922334, "learning_rate": 8.781287160553382e-06, "loss": 0.712, "step": 15261 }, { "epoch": 0.5527707352408547, "grad_norm": 1.3544393280210238, "learning_rate": 8.780122845372459e-06, "loss": 0.78, "step": 15262 }, { "epoch": 0.5528069540021732, "grad_norm": 1.340476313539762, "learning_rate": 8.77895854697812e-06, "loss": 0.7922, "step": 15263 }, { "epoch": 0.5528431727634915, "grad_norm": 1.4957370195809838, "learning_rate": 8.77779426538638e-06, "loss": 0.7544, "step": 15264 }, { "epoch": 0.5528793915248098, "grad_norm": 1.3732033093737666, "learning_rate": 8.77663000061327e-06, "loss": 0.7311, "step": 15265 }, { "epoch": 0.5529156102861282, "grad_norm": 1.4745276371615617, "learning_rate": 8.775465752674802e-06, "loss": 0.8113, "step": 15266 }, { "epoch": 0.5529518290474466, "grad_norm": 1.512781634493603, "learning_rate": 8.774301521587003e-06, "loss": 0.75, "step": 15267 }, { "epoch": 0.552988047808765, "grad_norm": 1.0314082404772225, "learning_rate": 8.773137307365892e-06, "loss": 0.7591, "step": 15268 }, { "epoch": 0.5530242665700833, "grad_norm": 1.390464268460796, "learning_rate": 8.771973110027487e-06, "loss": 0.8003, "step": 15269 }, { "epoch": 0.5530604853314016, "grad_norm": 1.2611844087254493, "learning_rate": 8.770808929587815e-06, "loss": 0.697, "step": 15270 }, { "epoch": 0.55309670409272, "grad_norm": 1.1459076847916232, "learning_rate": 8.76964476606289e-06, "loss": 0.7025, "step": 15271 }, { "epoch": 0.5531329228540384, "grad_norm": 1.2948008404491413, "learning_rate": 8.768480619468736e-06, "loss": 0.7112, "step": 15272 }, { "epoch": 0.5531691416153568, "grad_norm": 1.061020658342864, "learning_rate": 8.767316489821368e-06, "loss": 0.7412, "step": 15273 }, { "epoch": 0.5532053603766751, "grad_norm": 1.2256830344169605, "learning_rate": 8.766152377136813e-06, "loss": 0.709, "step": 15274 }, { "epoch": 0.5532415791379934, "grad_norm": 1.4115906352211278, "learning_rate": 8.764988281431083e-06, "loss": 0.7064, "step": 15275 }, { "epoch": 0.5532777978993119, "grad_norm": 1.0030005777840876, "learning_rate": 8.763824202720198e-06, "loss": 0.6982, "step": 15276 }, { "epoch": 0.5533140166606302, "grad_norm": 1.6054028499842679, "learning_rate": 8.76266014102018e-06, "loss": 0.7236, "step": 15277 }, { "epoch": 0.5533502354219486, "grad_norm": 1.2701387184473298, "learning_rate": 8.761496096347045e-06, "loss": 0.6744, "step": 15278 }, { "epoch": 0.5533864541832669, "grad_norm": 1.2090014878634892, "learning_rate": 8.76033206871681e-06, "loss": 0.6805, "step": 15279 }, { "epoch": 0.5534226729445852, "grad_norm": 1.0744916694911126, "learning_rate": 8.759168058145496e-06, "loss": 0.7064, "step": 15280 }, { "epoch": 0.5534588917059037, "grad_norm": 1.4110919557088393, "learning_rate": 8.75800406464912e-06, "loss": 0.7543, "step": 15281 }, { "epoch": 0.553495110467222, "grad_norm": 1.1969634090046777, "learning_rate": 8.756840088243701e-06, "loss": 0.6326, "step": 15282 }, { "epoch": 0.5535313292285404, "grad_norm": 1.4361357382914843, "learning_rate": 8.755676128945251e-06, "loss": 0.7356, "step": 15283 }, { "epoch": 0.5535675479898587, "grad_norm": 1.3017924247656019, "learning_rate": 8.754512186769796e-06, "loss": 0.654, "step": 15284 }, { "epoch": 0.5536037667511771, "grad_norm": 1.3497458415174923, "learning_rate": 8.753348261733343e-06, "loss": 0.6138, "step": 15285 }, { "epoch": 0.5536399855124955, "grad_norm": 0.9348859783285648, "learning_rate": 8.752184353851917e-06, "loss": 0.6852, "step": 15286 }, { "epoch": 0.5536762042738138, "grad_norm": 1.359002204727619, "learning_rate": 8.751020463141528e-06, "loss": 0.7314, "step": 15287 }, { "epoch": 0.5537124230351322, "grad_norm": 1.0223053242888516, "learning_rate": 8.749856589618197e-06, "loss": 0.6461, "step": 15288 }, { "epoch": 0.5537486417964506, "grad_norm": 1.2500525053739149, "learning_rate": 8.748692733297935e-06, "loss": 0.6594, "step": 15289 }, { "epoch": 0.5537848605577689, "grad_norm": 1.4574709591847261, "learning_rate": 8.747528894196764e-06, "loss": 0.7787, "step": 15290 }, { "epoch": 0.5538210793190873, "grad_norm": 0.9183781464543804, "learning_rate": 8.746365072330692e-06, "loss": 0.6996, "step": 15291 }, { "epoch": 0.5538572980804056, "grad_norm": 1.1063129884837528, "learning_rate": 8.745201267715739e-06, "loss": 0.7511, "step": 15292 }, { "epoch": 0.553893516841724, "grad_norm": 1.3520287546035397, "learning_rate": 8.744037480367922e-06, "loss": 0.7713, "step": 15293 }, { "epoch": 0.5539297356030424, "grad_norm": 1.523057709327861, "learning_rate": 8.74287371030325e-06, "loss": 0.7452, "step": 15294 }, { "epoch": 0.5539659543643607, "grad_norm": 1.2490986987025443, "learning_rate": 8.741709957537742e-06, "loss": 0.698, "step": 15295 }, { "epoch": 0.5540021731256791, "grad_norm": 1.4113561763662776, "learning_rate": 8.74054622208741e-06, "loss": 0.7781, "step": 15296 }, { "epoch": 0.5540383918869974, "grad_norm": 1.233912127663046, "learning_rate": 8.73938250396827e-06, "loss": 0.6666, "step": 15297 }, { "epoch": 0.5540746106483159, "grad_norm": 1.314243422497631, "learning_rate": 8.738218803196332e-06, "loss": 0.8252, "step": 15298 }, { "epoch": 0.5541108294096342, "grad_norm": 1.302158909909043, "learning_rate": 8.737055119787615e-06, "loss": 0.7007, "step": 15299 }, { "epoch": 0.5541470481709525, "grad_norm": 1.3453284075479923, "learning_rate": 8.735891453758126e-06, "loss": 0.6856, "step": 15300 }, { "epoch": 0.5541832669322709, "grad_norm": 1.417179844508581, "learning_rate": 8.734727805123883e-06, "loss": 0.8107, "step": 15301 }, { "epoch": 0.5542194856935893, "grad_norm": 1.2936730821809932, "learning_rate": 8.733564173900898e-06, "loss": 0.7643, "step": 15302 }, { "epoch": 0.5542557044549077, "grad_norm": 1.4609785941153108, "learning_rate": 8.73240056010518e-06, "loss": 0.6929, "step": 15303 }, { "epoch": 0.554291923216226, "grad_norm": 1.3505434097137012, "learning_rate": 8.731236963752746e-06, "loss": 0.7589, "step": 15304 }, { "epoch": 0.5543281419775443, "grad_norm": 1.2733788769339895, "learning_rate": 8.730073384859605e-06, "loss": 0.7154, "step": 15305 }, { "epoch": 0.5543643607388627, "grad_norm": 1.1918190030082034, "learning_rate": 8.728909823441772e-06, "loss": 0.7049, "step": 15306 }, { "epoch": 0.5544005795001811, "grad_norm": 1.293228484555381, "learning_rate": 8.727746279515255e-06, "loss": 0.688, "step": 15307 }, { "epoch": 0.5544367982614995, "grad_norm": 1.350236668250669, "learning_rate": 8.72658275309607e-06, "loss": 0.7233, "step": 15308 }, { "epoch": 0.5544730170228178, "grad_norm": 1.282412587404389, "learning_rate": 8.725419244200222e-06, "loss": 0.7257, "step": 15309 }, { "epoch": 0.5545092357841361, "grad_norm": 1.449704132642511, "learning_rate": 8.724255752843726e-06, "loss": 0.6975, "step": 15310 }, { "epoch": 0.5545454545454546, "grad_norm": 1.2136888074698957, "learning_rate": 8.723092279042591e-06, "loss": 0.7029, "step": 15311 }, { "epoch": 0.5545816733067729, "grad_norm": 1.2544959937876354, "learning_rate": 8.72192882281283e-06, "loss": 0.6814, "step": 15312 }, { "epoch": 0.5546178920680913, "grad_norm": 1.2334339758917114, "learning_rate": 8.72076538417045e-06, "loss": 0.6605, "step": 15313 }, { "epoch": 0.5546541108294096, "grad_norm": 1.3381399897264783, "learning_rate": 8.719601963131464e-06, "loss": 0.7956, "step": 15314 }, { "epoch": 0.554690329590728, "grad_norm": 1.3912382695668772, "learning_rate": 8.718438559711879e-06, "loss": 0.683, "step": 15315 }, { "epoch": 0.5547265483520464, "grad_norm": 1.544313696887705, "learning_rate": 8.717275173927704e-06, "loss": 0.7874, "step": 15316 }, { "epoch": 0.5547627671133647, "grad_norm": 1.4735309642737429, "learning_rate": 8.716111805794953e-06, "loss": 0.7394, "step": 15317 }, { "epoch": 0.5547989858746831, "grad_norm": 1.4509637524240482, "learning_rate": 8.71494845532963e-06, "loss": 0.7587, "step": 15318 }, { "epoch": 0.5548352046360014, "grad_norm": 1.2764927007755142, "learning_rate": 8.713785122547746e-06, "loss": 0.7038, "step": 15319 }, { "epoch": 0.5548714233973198, "grad_norm": 1.4346623478834468, "learning_rate": 8.712621807465307e-06, "loss": 0.706, "step": 15320 }, { "epoch": 0.5549076421586382, "grad_norm": 1.3685709262744383, "learning_rate": 8.711458510098326e-06, "loss": 0.7298, "step": 15321 }, { "epoch": 0.5549438609199565, "grad_norm": 1.4188220869143484, "learning_rate": 8.710295230462805e-06, "loss": 0.8391, "step": 15322 }, { "epoch": 0.5549800796812749, "grad_norm": 1.4018481017543314, "learning_rate": 8.709131968574756e-06, "loss": 0.8075, "step": 15323 }, { "epoch": 0.5550162984425933, "grad_norm": 1.0526220099904091, "learning_rate": 8.707968724450186e-06, "loss": 0.7055, "step": 15324 }, { "epoch": 0.5550525172039116, "grad_norm": 1.3413347783985439, "learning_rate": 8.7068054981051e-06, "loss": 0.6826, "step": 15325 }, { "epoch": 0.55508873596523, "grad_norm": 1.3713114326563827, "learning_rate": 8.705642289555511e-06, "loss": 0.6993, "step": 15326 }, { "epoch": 0.5551249547265483, "grad_norm": 1.2869879928741887, "learning_rate": 8.704479098817417e-06, "loss": 0.7251, "step": 15327 }, { "epoch": 0.5551611734878668, "grad_norm": 1.3195414421925604, "learning_rate": 8.70331592590683e-06, "loss": 0.718, "step": 15328 }, { "epoch": 0.5551973922491851, "grad_norm": 1.385006084736479, "learning_rate": 8.702152770839755e-06, "loss": 0.6404, "step": 15329 }, { "epoch": 0.5552336110105034, "grad_norm": 1.3866210607586043, "learning_rate": 8.7009896336322e-06, "loss": 0.7603, "step": 15330 }, { "epoch": 0.5552698297718218, "grad_norm": 1.2004558925965423, "learning_rate": 8.699826514300168e-06, "loss": 0.6974, "step": 15331 }, { "epoch": 0.5553060485331401, "grad_norm": 1.0827139553259275, "learning_rate": 8.698663412859667e-06, "loss": 0.7033, "step": 15332 }, { "epoch": 0.5553422672944586, "grad_norm": 1.2057265657729184, "learning_rate": 8.697500329326697e-06, "loss": 0.6787, "step": 15333 }, { "epoch": 0.5553784860557769, "grad_norm": 1.3485426765611768, "learning_rate": 8.696337263717269e-06, "loss": 0.8122, "step": 15334 }, { "epoch": 0.5554147048170952, "grad_norm": 1.431837561464254, "learning_rate": 8.695174216047388e-06, "loss": 0.7678, "step": 15335 }, { "epoch": 0.5554509235784136, "grad_norm": 1.3973213441972563, "learning_rate": 8.694011186333052e-06, "loss": 0.7265, "step": 15336 }, { "epoch": 0.555487142339732, "grad_norm": 1.4196996652708336, "learning_rate": 8.692848174590274e-06, "loss": 0.7498, "step": 15337 }, { "epoch": 0.5555233611010504, "grad_norm": 1.3248669928131407, "learning_rate": 8.69168518083505e-06, "loss": 0.6907, "step": 15338 }, { "epoch": 0.5555595798623687, "grad_norm": 1.459885233571711, "learning_rate": 8.69052220508339e-06, "loss": 0.755, "step": 15339 }, { "epoch": 0.555595798623687, "grad_norm": 1.4693219980950036, "learning_rate": 8.689359247351294e-06, "loss": 0.7656, "step": 15340 }, { "epoch": 0.5556320173850055, "grad_norm": 1.3732322904110514, "learning_rate": 8.688196307654769e-06, "loss": 0.7051, "step": 15341 }, { "epoch": 0.5556682361463238, "grad_norm": 1.4465376423312932, "learning_rate": 8.687033386009811e-06, "loss": 0.7658, "step": 15342 }, { "epoch": 0.5557044549076422, "grad_norm": 1.0695665082612742, "learning_rate": 8.685870482432429e-06, "loss": 0.6264, "step": 15343 }, { "epoch": 0.5557406736689605, "grad_norm": 1.3343220677083913, "learning_rate": 8.684707596938624e-06, "loss": 0.7113, "step": 15344 }, { "epoch": 0.5557768924302788, "grad_norm": 1.5096747598039435, "learning_rate": 8.683544729544398e-06, "loss": 0.7176, "step": 15345 }, { "epoch": 0.5558131111915973, "grad_norm": 1.0602355124780747, "learning_rate": 8.682381880265756e-06, "loss": 0.7109, "step": 15346 }, { "epoch": 0.5558493299529156, "grad_norm": 1.4057921962475106, "learning_rate": 8.681219049118691e-06, "loss": 0.7393, "step": 15347 }, { "epoch": 0.555885548714234, "grad_norm": 1.4053447723839618, "learning_rate": 8.680056236119218e-06, "loss": 0.7693, "step": 15348 }, { "epoch": 0.5559217674755523, "grad_norm": 1.7390970828387065, "learning_rate": 8.678893441283325e-06, "loss": 0.7983, "step": 15349 }, { "epoch": 0.5559579862368706, "grad_norm": 1.4153971140084718, "learning_rate": 8.677730664627023e-06, "loss": 0.7408, "step": 15350 }, { "epoch": 0.5559942049981891, "grad_norm": 1.2831544583815988, "learning_rate": 8.676567906166307e-06, "loss": 0.7312, "step": 15351 }, { "epoch": 0.5560304237595074, "grad_norm": 1.5491149462087515, "learning_rate": 8.675405165917182e-06, "loss": 0.7241, "step": 15352 }, { "epoch": 0.5560666425208258, "grad_norm": 1.5284290189747574, "learning_rate": 8.674242443895641e-06, "loss": 0.7221, "step": 15353 }, { "epoch": 0.5561028612821441, "grad_norm": 1.3477785960237678, "learning_rate": 8.673079740117692e-06, "loss": 0.7287, "step": 15354 }, { "epoch": 0.5561390800434625, "grad_norm": 1.1910217953686635, "learning_rate": 8.671917054599331e-06, "loss": 0.7278, "step": 15355 }, { "epoch": 0.5561752988047809, "grad_norm": 1.507348485812481, "learning_rate": 8.670754387356558e-06, "loss": 0.6967, "step": 15356 }, { "epoch": 0.5562115175660992, "grad_norm": 1.2703356248667488, "learning_rate": 8.669591738405374e-06, "loss": 0.6886, "step": 15357 }, { "epoch": 0.5562477363274176, "grad_norm": 1.4495310935200982, "learning_rate": 8.668429107761775e-06, "loss": 0.7472, "step": 15358 }, { "epoch": 0.556283955088736, "grad_norm": 1.2888706324577972, "learning_rate": 8.667266495441764e-06, "loss": 0.7606, "step": 15359 }, { "epoch": 0.5563201738500543, "grad_norm": 1.2756835752941424, "learning_rate": 8.666103901461333e-06, "loss": 0.7144, "step": 15360 }, { "epoch": 0.5563563926113727, "grad_norm": 1.59583949343295, "learning_rate": 8.66494132583649e-06, "loss": 0.677, "step": 15361 }, { "epoch": 0.556392611372691, "grad_norm": 1.3750826759299914, "learning_rate": 8.663778768583222e-06, "loss": 0.7233, "step": 15362 }, { "epoch": 0.5564288301340095, "grad_norm": 1.339924032792191, "learning_rate": 8.662616229717537e-06, "loss": 0.7195, "step": 15363 }, { "epoch": 0.5564650488953278, "grad_norm": 1.336319158507827, "learning_rate": 8.661453709255423e-06, "loss": 0.7291, "step": 15364 }, { "epoch": 0.5565012676566461, "grad_norm": 1.3829487973192796, "learning_rate": 8.660291207212883e-06, "loss": 0.7727, "step": 15365 }, { "epoch": 0.5565374864179645, "grad_norm": 1.5566605012646586, "learning_rate": 8.659128723605916e-06, "loss": 0.6496, "step": 15366 }, { "epoch": 0.5565737051792828, "grad_norm": 1.538887875714793, "learning_rate": 8.657966258450514e-06, "loss": 0.7608, "step": 15367 }, { "epoch": 0.5566099239406013, "grad_norm": 1.3900177170761567, "learning_rate": 8.656803811762678e-06, "loss": 0.7296, "step": 15368 }, { "epoch": 0.5566461427019196, "grad_norm": 1.3105279115251451, "learning_rate": 8.655641383558397e-06, "loss": 0.7298, "step": 15369 }, { "epoch": 0.5566823614632379, "grad_norm": 1.0806675571427893, "learning_rate": 8.654478973853678e-06, "loss": 0.7566, "step": 15370 }, { "epoch": 0.5567185802245563, "grad_norm": 1.5203695951053815, "learning_rate": 8.653316582664504e-06, "loss": 0.731, "step": 15371 }, { "epoch": 0.5567547989858747, "grad_norm": 1.4411045536656661, "learning_rate": 8.652154210006882e-06, "loss": 0.6458, "step": 15372 }, { "epoch": 0.5567910177471931, "grad_norm": 0.9773677951053168, "learning_rate": 8.650991855896799e-06, "loss": 0.8253, "step": 15373 }, { "epoch": 0.5568272365085114, "grad_norm": 1.3201211143070655, "learning_rate": 8.649829520350255e-06, "loss": 0.7188, "step": 15374 }, { "epoch": 0.5568634552698297, "grad_norm": 1.053321013193094, "learning_rate": 8.648667203383245e-06, "loss": 0.7734, "step": 15375 }, { "epoch": 0.5568996740311482, "grad_norm": 1.2715920001360552, "learning_rate": 8.647504905011759e-06, "loss": 0.6593, "step": 15376 }, { "epoch": 0.5569358927924665, "grad_norm": 1.3806030772084554, "learning_rate": 8.646342625251795e-06, "loss": 0.7892, "step": 15377 }, { "epoch": 0.5569721115537849, "grad_norm": 1.3490212852623233, "learning_rate": 8.645180364119345e-06, "loss": 0.668, "step": 15378 }, { "epoch": 0.5570083303151032, "grad_norm": 1.0614802379490509, "learning_rate": 8.644018121630406e-06, "loss": 0.74, "step": 15379 }, { "epoch": 0.5570445490764215, "grad_norm": 1.304673665224608, "learning_rate": 8.642855897800967e-06, "loss": 0.7431, "step": 15380 }, { "epoch": 0.55708076783774, "grad_norm": 1.4772122873329248, "learning_rate": 8.641693692647025e-06, "loss": 0.7726, "step": 15381 }, { "epoch": 0.5571169865990583, "grad_norm": 1.0238357896305965, "learning_rate": 8.640531506184568e-06, "loss": 0.7501, "step": 15382 }, { "epoch": 0.5571532053603767, "grad_norm": 1.314504621695811, "learning_rate": 8.639369338429596e-06, "loss": 0.6908, "step": 15383 }, { "epoch": 0.557189424121695, "grad_norm": 1.2662761795285762, "learning_rate": 8.638207189398096e-06, "loss": 0.5938, "step": 15384 }, { "epoch": 0.5572256428830134, "grad_norm": 1.1141486842679684, "learning_rate": 8.63704505910606e-06, "loss": 0.6624, "step": 15385 }, { "epoch": 0.5572618616443318, "grad_norm": 1.3403139213908908, "learning_rate": 8.635882947569484e-06, "loss": 0.7308, "step": 15386 }, { "epoch": 0.5572980804056501, "grad_norm": 1.433531262831538, "learning_rate": 8.634720854804354e-06, "loss": 0.762, "step": 15387 }, { "epoch": 0.5573342991669685, "grad_norm": 1.3892240861552736, "learning_rate": 8.633558780826668e-06, "loss": 0.6944, "step": 15388 }, { "epoch": 0.5573705179282868, "grad_norm": 1.1090516838423732, "learning_rate": 8.632396725652412e-06, "loss": 0.7537, "step": 15389 }, { "epoch": 0.5574067366896052, "grad_norm": 1.381509979004091, "learning_rate": 8.63123468929758e-06, "loss": 0.7824, "step": 15390 }, { "epoch": 0.5574429554509236, "grad_norm": 1.3720769888827433, "learning_rate": 8.63007267177816e-06, "loss": 0.7483, "step": 15391 }, { "epoch": 0.5574791742122419, "grad_norm": 1.6546415214301278, "learning_rate": 8.628910673110146e-06, "loss": 0.7202, "step": 15392 }, { "epoch": 0.5575153929735603, "grad_norm": 1.0980264639503152, "learning_rate": 8.627748693309524e-06, "loss": 0.7837, "step": 15393 }, { "epoch": 0.5575516117348787, "grad_norm": 1.1628622822716095, "learning_rate": 8.626586732392287e-06, "loss": 0.7608, "step": 15394 }, { "epoch": 0.557587830496197, "grad_norm": 1.2787567192238574, "learning_rate": 8.625424790374422e-06, "loss": 0.6924, "step": 15395 }, { "epoch": 0.5576240492575154, "grad_norm": 0.9064193165792256, "learning_rate": 8.62426286727192e-06, "loss": 0.7188, "step": 15396 }, { "epoch": 0.5576602680188337, "grad_norm": 1.5225453772383202, "learning_rate": 8.62310096310077e-06, "loss": 0.7754, "step": 15397 }, { "epoch": 0.5576964867801522, "grad_norm": 1.3649978848991582, "learning_rate": 8.62193907787696e-06, "loss": 0.7537, "step": 15398 }, { "epoch": 0.5577327055414705, "grad_norm": 1.3983659626069083, "learning_rate": 8.62077721161648e-06, "loss": 0.7137, "step": 15399 }, { "epoch": 0.5577689243027888, "grad_norm": 1.2635477092004321, "learning_rate": 8.619615364335316e-06, "loss": 0.6832, "step": 15400 }, { "epoch": 0.5578051430641072, "grad_norm": 1.3510537292448836, "learning_rate": 8.61845353604946e-06, "loss": 0.6865, "step": 15401 }, { "epoch": 0.5578413618254255, "grad_norm": 1.9581322142924407, "learning_rate": 8.617291726774896e-06, "loss": 0.7141, "step": 15402 }, { "epoch": 0.557877580586744, "grad_norm": 1.4477326673007993, "learning_rate": 8.616129936527612e-06, "loss": 0.7278, "step": 15403 }, { "epoch": 0.5579137993480623, "grad_norm": 1.2388821980293059, "learning_rate": 8.614968165323598e-06, "loss": 0.6496, "step": 15404 }, { "epoch": 0.5579500181093806, "grad_norm": 1.3069362498403692, "learning_rate": 8.613806413178836e-06, "loss": 0.6801, "step": 15405 }, { "epoch": 0.557986236870699, "grad_norm": 1.141843987447915, "learning_rate": 8.61264468010932e-06, "loss": 0.6886, "step": 15406 }, { "epoch": 0.5580224556320174, "grad_norm": 1.2842251729705962, "learning_rate": 8.611482966131027e-06, "loss": 0.7648, "step": 15407 }, { "epoch": 0.5580586743933358, "grad_norm": 1.3995897950749328, "learning_rate": 8.610321271259954e-06, "loss": 0.8276, "step": 15408 }, { "epoch": 0.5580948931546541, "grad_norm": 1.362277196791362, "learning_rate": 8.609159595512077e-06, "loss": 0.6968, "step": 15409 }, { "epoch": 0.5581311119159724, "grad_norm": 1.3873087903745793, "learning_rate": 8.607997938903389e-06, "loss": 0.7353, "step": 15410 }, { "epoch": 0.5581673306772909, "grad_norm": 1.343977711340664, "learning_rate": 8.60683630144987e-06, "loss": 0.7444, "step": 15411 }, { "epoch": 0.5582035494386092, "grad_norm": 1.362199775183674, "learning_rate": 8.60567468316751e-06, "loss": 0.694, "step": 15412 }, { "epoch": 0.5582397681999276, "grad_norm": 1.482260178805438, "learning_rate": 8.604513084072288e-06, "loss": 0.7565, "step": 15413 }, { "epoch": 0.5582759869612459, "grad_norm": 1.2963949077512793, "learning_rate": 8.603351504180198e-06, "loss": 0.6829, "step": 15414 }, { "epoch": 0.5583122057225642, "grad_norm": 1.398974511993768, "learning_rate": 8.602189943507214e-06, "loss": 0.7336, "step": 15415 }, { "epoch": 0.5583484244838827, "grad_norm": 1.422220578484788, "learning_rate": 8.601028402069324e-06, "loss": 0.6925, "step": 15416 }, { "epoch": 0.558384643245201, "grad_norm": 1.2695608869174377, "learning_rate": 8.599866879882515e-06, "loss": 0.6736, "step": 15417 }, { "epoch": 0.5584208620065194, "grad_norm": 1.446593016341506, "learning_rate": 8.598705376962765e-06, "loss": 0.7747, "step": 15418 }, { "epoch": 0.5584570807678377, "grad_norm": 1.2643464961506627, "learning_rate": 8.597543893326065e-06, "loss": 0.6945, "step": 15419 }, { "epoch": 0.558493299529156, "grad_norm": 1.3794501504403125, "learning_rate": 8.59638242898839e-06, "loss": 0.6611, "step": 15420 }, { "epoch": 0.5585295182904745, "grad_norm": 1.3259453685404754, "learning_rate": 8.595220983965727e-06, "loss": 0.6744, "step": 15421 }, { "epoch": 0.5585657370517928, "grad_norm": 1.2694951756560535, "learning_rate": 8.594059558274056e-06, "loss": 0.7113, "step": 15422 }, { "epoch": 0.5586019558131112, "grad_norm": 0.9616520223382611, "learning_rate": 8.592898151929364e-06, "loss": 0.6974, "step": 15423 }, { "epoch": 0.5586381745744295, "grad_norm": 1.236306207622286, "learning_rate": 8.591736764947628e-06, "loss": 0.7257, "step": 15424 }, { "epoch": 0.5586743933357479, "grad_norm": 1.2465880386739723, "learning_rate": 8.590575397344831e-06, "loss": 0.7453, "step": 15425 }, { "epoch": 0.5587106120970663, "grad_norm": 1.3042221460073113, "learning_rate": 8.589414049136955e-06, "loss": 0.7058, "step": 15426 }, { "epoch": 0.5587468308583846, "grad_norm": 1.400080548091648, "learning_rate": 8.58825272033998e-06, "loss": 0.7903, "step": 15427 }, { "epoch": 0.558783049619703, "grad_norm": 1.393627938931373, "learning_rate": 8.587091410969893e-06, "loss": 0.7191, "step": 15428 }, { "epoch": 0.5588192683810214, "grad_norm": 1.5396856556814185, "learning_rate": 8.585930121042666e-06, "loss": 0.7566, "step": 15429 }, { "epoch": 0.5588554871423397, "grad_norm": 1.2416155836085663, "learning_rate": 8.584768850574284e-06, "loss": 0.6569, "step": 15430 }, { "epoch": 0.5588917059036581, "grad_norm": 1.433579570845663, "learning_rate": 8.583607599580723e-06, "loss": 0.7355, "step": 15431 }, { "epoch": 0.5589279246649764, "grad_norm": 1.0277028570644424, "learning_rate": 8.582446368077971e-06, "loss": 0.6857, "step": 15432 }, { "epoch": 0.5589641434262949, "grad_norm": 1.298124476973585, "learning_rate": 8.581285156081998e-06, "loss": 0.6584, "step": 15433 }, { "epoch": 0.5590003621876132, "grad_norm": 1.0643084473674873, "learning_rate": 8.580123963608791e-06, "loss": 0.715, "step": 15434 }, { "epoch": 0.5590365809489315, "grad_norm": 1.015884051159651, "learning_rate": 8.578962790674324e-06, "loss": 0.703, "step": 15435 }, { "epoch": 0.5590727997102499, "grad_norm": 1.3779435876892736, "learning_rate": 8.577801637294578e-06, "loss": 0.7509, "step": 15436 }, { "epoch": 0.5591090184715682, "grad_norm": 1.419338518389435, "learning_rate": 8.576640503485533e-06, "loss": 0.73, "step": 15437 }, { "epoch": 0.5591452372328867, "grad_norm": 1.326515259339669, "learning_rate": 8.575479389263163e-06, "loss": 0.7034, "step": 15438 }, { "epoch": 0.559181455994205, "grad_norm": 1.39815300573003, "learning_rate": 8.57431829464345e-06, "loss": 0.7403, "step": 15439 }, { "epoch": 0.5592176747555233, "grad_norm": 1.01538500342999, "learning_rate": 8.573157219642369e-06, "loss": 0.7108, "step": 15440 }, { "epoch": 0.5592538935168417, "grad_norm": 0.9823414222028299, "learning_rate": 8.5719961642759e-06, "loss": 0.6809, "step": 15441 }, { "epoch": 0.5592901122781601, "grad_norm": 1.1247366452195748, "learning_rate": 8.570835128560016e-06, "loss": 0.6914, "step": 15442 }, { "epoch": 0.5593263310394785, "grad_norm": 1.3409185088805753, "learning_rate": 8.569674112510698e-06, "loss": 0.808, "step": 15443 }, { "epoch": 0.5593625498007968, "grad_norm": 0.9300126866724299, "learning_rate": 8.56851311614392e-06, "loss": 0.6926, "step": 15444 }, { "epoch": 0.5593987685621151, "grad_norm": 1.3330501102541719, "learning_rate": 8.567352139475661e-06, "loss": 0.7405, "step": 15445 }, { "epoch": 0.5594349873234336, "grad_norm": 1.3895851665041021, "learning_rate": 8.566191182521893e-06, "loss": 0.7608, "step": 15446 }, { "epoch": 0.5594712060847519, "grad_norm": 1.5692904488902988, "learning_rate": 8.565030245298594e-06, "loss": 0.7511, "step": 15447 }, { "epoch": 0.5595074248460703, "grad_norm": 1.4480242652043778, "learning_rate": 8.563869327821744e-06, "loss": 0.6512, "step": 15448 }, { "epoch": 0.5595436436073886, "grad_norm": 0.9747694326987394, "learning_rate": 8.562708430107309e-06, "loss": 0.6487, "step": 15449 }, { "epoch": 0.5595798623687069, "grad_norm": 1.4243993097501404, "learning_rate": 8.561547552171272e-06, "loss": 0.7246, "step": 15450 }, { "epoch": 0.5596160811300254, "grad_norm": 1.3064199398118737, "learning_rate": 8.560386694029603e-06, "loss": 0.806, "step": 15451 }, { "epoch": 0.5596522998913437, "grad_norm": 1.2628818987761117, "learning_rate": 8.55922585569828e-06, "loss": 0.6909, "step": 15452 }, { "epoch": 0.5596885186526621, "grad_norm": 1.138683552952029, "learning_rate": 8.558065037193272e-06, "loss": 0.8559, "step": 15453 }, { "epoch": 0.5597247374139804, "grad_norm": 0.9451399875548553, "learning_rate": 8.55690423853056e-06, "loss": 0.7467, "step": 15454 }, { "epoch": 0.5597609561752988, "grad_norm": 1.3209380052115234, "learning_rate": 8.55574345972611e-06, "loss": 0.6956, "step": 15455 }, { "epoch": 0.5597971749366172, "grad_norm": 1.656226709040136, "learning_rate": 8.554582700795902e-06, "loss": 0.7963, "step": 15456 }, { "epoch": 0.5598333936979355, "grad_norm": 1.3447079328245213, "learning_rate": 8.553421961755904e-06, "loss": 0.7313, "step": 15457 }, { "epoch": 0.5598696124592539, "grad_norm": 1.246055459686856, "learning_rate": 8.55226124262209e-06, "loss": 0.6715, "step": 15458 }, { "epoch": 0.5599058312205722, "grad_norm": 1.3884908756704788, "learning_rate": 8.551100543410437e-06, "loss": 0.7244, "step": 15459 }, { "epoch": 0.5599420499818906, "grad_norm": 1.3007756825920762, "learning_rate": 8.549939864136912e-06, "loss": 0.7182, "step": 15460 }, { "epoch": 0.559978268743209, "grad_norm": 1.2929366171148744, "learning_rate": 8.548779204817487e-06, "loss": 0.7892, "step": 15461 }, { "epoch": 0.5600144875045273, "grad_norm": 1.4458323587655715, "learning_rate": 8.547618565468137e-06, "loss": 0.7446, "step": 15462 }, { "epoch": 0.5600507062658457, "grad_norm": 1.4649629442676086, "learning_rate": 8.546457946104832e-06, "loss": 0.8054, "step": 15463 }, { "epoch": 0.5600869250271641, "grad_norm": 1.3456631704359214, "learning_rate": 8.545297346743541e-06, "loss": 0.733, "step": 15464 }, { "epoch": 0.5601231437884824, "grad_norm": 1.2490448625529966, "learning_rate": 8.54413676740024e-06, "loss": 0.7365, "step": 15465 }, { "epoch": 0.5601593625498008, "grad_norm": 1.3312010632114626, "learning_rate": 8.542976208090893e-06, "loss": 0.7389, "step": 15466 }, { "epoch": 0.5601955813111191, "grad_norm": 1.319018520598325, "learning_rate": 8.541815668831474e-06, "loss": 0.7634, "step": 15467 }, { "epoch": 0.5602318000724376, "grad_norm": 1.2315704197448678, "learning_rate": 8.540655149637955e-06, "loss": 0.7209, "step": 15468 }, { "epoch": 0.5602680188337559, "grad_norm": 1.3830652463936461, "learning_rate": 8.5394946505263e-06, "loss": 0.6801, "step": 15469 }, { "epoch": 0.5603042375950742, "grad_norm": 1.3593543684150118, "learning_rate": 8.538334171512485e-06, "loss": 0.742, "step": 15470 }, { "epoch": 0.5603404563563926, "grad_norm": 1.0126178604008298, "learning_rate": 8.537173712612474e-06, "loss": 0.7422, "step": 15471 }, { "epoch": 0.5603766751177109, "grad_norm": 1.6334955917524565, "learning_rate": 8.536013273842242e-06, "loss": 0.7404, "step": 15472 }, { "epoch": 0.5604128938790294, "grad_norm": 1.3484446845299731, "learning_rate": 8.534852855217748e-06, "loss": 0.6967, "step": 15473 }, { "epoch": 0.5604491126403477, "grad_norm": 1.320415513968885, "learning_rate": 8.53369245675497e-06, "loss": 0.692, "step": 15474 }, { "epoch": 0.560485331401666, "grad_norm": 1.341203822268012, "learning_rate": 8.53253207846987e-06, "loss": 0.6774, "step": 15475 }, { "epoch": 0.5605215501629844, "grad_norm": 1.3556060501610285, "learning_rate": 8.531371720378421e-06, "loss": 0.7232, "step": 15476 }, { "epoch": 0.5605577689243028, "grad_norm": 1.4633047608871217, "learning_rate": 8.530211382496584e-06, "loss": 0.7587, "step": 15477 }, { "epoch": 0.5605939876856212, "grad_norm": 1.4497337923441909, "learning_rate": 8.529051064840332e-06, "loss": 0.6865, "step": 15478 }, { "epoch": 0.5606302064469395, "grad_norm": 1.5462300938267028, "learning_rate": 8.52789076742563e-06, "loss": 0.7844, "step": 15479 }, { "epoch": 0.5606664252082578, "grad_norm": 1.1240526825450206, "learning_rate": 8.526730490268443e-06, "loss": 0.7297, "step": 15480 }, { "epoch": 0.5607026439695763, "grad_norm": 1.4459600467173046, "learning_rate": 8.525570233384741e-06, "loss": 0.6793, "step": 15481 }, { "epoch": 0.5607388627308946, "grad_norm": 1.0811187544027334, "learning_rate": 8.524409996790485e-06, "loss": 0.7286, "step": 15482 }, { "epoch": 0.560775081492213, "grad_norm": 1.2995189765445994, "learning_rate": 8.523249780501647e-06, "loss": 0.7849, "step": 15483 }, { "epoch": 0.5608113002535313, "grad_norm": 1.4211577667121957, "learning_rate": 8.522089584534189e-06, "loss": 0.7091, "step": 15484 }, { "epoch": 0.5608475190148496, "grad_norm": 1.3613644954907298, "learning_rate": 8.520929408904079e-06, "loss": 0.7415, "step": 15485 }, { "epoch": 0.5608837377761681, "grad_norm": 1.369884802292467, "learning_rate": 8.51976925362728e-06, "loss": 0.7348, "step": 15486 }, { "epoch": 0.5609199565374864, "grad_norm": 1.3566778710038923, "learning_rate": 8.518609118719753e-06, "loss": 0.7832, "step": 15487 }, { "epoch": 0.5609561752988048, "grad_norm": 1.5288989364164507, "learning_rate": 8.517449004197468e-06, "loss": 0.7845, "step": 15488 }, { "epoch": 0.5609923940601231, "grad_norm": 1.0968328636059383, "learning_rate": 8.516288910076385e-06, "loss": 0.7115, "step": 15489 }, { "epoch": 0.5610286128214415, "grad_norm": 1.2820210099610407, "learning_rate": 8.515128836372473e-06, "loss": 0.7069, "step": 15490 }, { "epoch": 0.5610648315827599, "grad_norm": 1.9506007247338328, "learning_rate": 8.513968783101693e-06, "loss": 0.7948, "step": 15491 }, { "epoch": 0.5611010503440782, "grad_norm": 1.3090082729039836, "learning_rate": 8.512808750280007e-06, "loss": 0.7261, "step": 15492 }, { "epoch": 0.5611372691053966, "grad_norm": 1.3442907825419925, "learning_rate": 8.511648737923379e-06, "loss": 0.7165, "step": 15493 }, { "epoch": 0.561173487866715, "grad_norm": 1.4684899409343808, "learning_rate": 8.510488746047773e-06, "loss": 0.7199, "step": 15494 }, { "epoch": 0.5612097066280333, "grad_norm": 1.4324405784935867, "learning_rate": 8.50932877466915e-06, "loss": 0.6945, "step": 15495 }, { "epoch": 0.5612459253893517, "grad_norm": 1.4742877066212292, "learning_rate": 8.508168823803474e-06, "loss": 0.7299, "step": 15496 }, { "epoch": 0.56128214415067, "grad_norm": 0.9650643327362718, "learning_rate": 8.507008893466706e-06, "loss": 0.7019, "step": 15497 }, { "epoch": 0.5613183629119884, "grad_norm": 1.2674876612995511, "learning_rate": 8.505848983674805e-06, "loss": 0.757, "step": 15498 }, { "epoch": 0.5613545816733068, "grad_norm": 1.4078578148628906, "learning_rate": 8.504689094443737e-06, "loss": 0.6572, "step": 15499 }, { "epoch": 0.5613908004346251, "grad_norm": 1.3761590683639222, "learning_rate": 8.503529225789461e-06, "loss": 0.7261, "step": 15500 }, { "epoch": 0.5614270191959435, "grad_norm": 1.246162373049363, "learning_rate": 8.502369377727936e-06, "loss": 0.6655, "step": 15501 }, { "epoch": 0.5614632379572618, "grad_norm": 1.0324751596090782, "learning_rate": 8.501209550275127e-06, "loss": 0.6176, "step": 15502 }, { "epoch": 0.5614994567185803, "grad_norm": 1.3827018275935772, "learning_rate": 8.50004974344699e-06, "loss": 0.7411, "step": 15503 }, { "epoch": 0.5615356754798986, "grad_norm": 1.4195898145247885, "learning_rate": 8.498889957259487e-06, "loss": 0.7304, "step": 15504 }, { "epoch": 0.5615718942412169, "grad_norm": 1.316898141558277, "learning_rate": 8.497730191728576e-06, "loss": 0.7197, "step": 15505 }, { "epoch": 0.5616081130025353, "grad_norm": 1.2736472998434079, "learning_rate": 8.49657044687022e-06, "loss": 0.7007, "step": 15506 }, { "epoch": 0.5616443317638536, "grad_norm": 1.3557751059274796, "learning_rate": 8.495410722700373e-06, "loss": 0.6994, "step": 15507 }, { "epoch": 0.5616805505251721, "grad_norm": 1.3499523367459192, "learning_rate": 8.494251019234999e-06, "loss": 0.6697, "step": 15508 }, { "epoch": 0.5617167692864904, "grad_norm": 1.285279379688529, "learning_rate": 8.493091336490051e-06, "loss": 0.7333, "step": 15509 }, { "epoch": 0.5617529880478087, "grad_norm": 1.334099448366584, "learning_rate": 8.491931674481491e-06, "loss": 0.6964, "step": 15510 }, { "epoch": 0.5617892068091271, "grad_norm": 1.4561875856412527, "learning_rate": 8.49077203322528e-06, "loss": 0.6984, "step": 15511 }, { "epoch": 0.5618254255704455, "grad_norm": 1.3554225805208353, "learning_rate": 8.489612412737366e-06, "loss": 0.7151, "step": 15512 }, { "epoch": 0.5618616443317639, "grad_norm": 1.4310769738374907, "learning_rate": 8.488452813033716e-06, "loss": 0.7934, "step": 15513 }, { "epoch": 0.5618978630930822, "grad_norm": 1.3725752474962967, "learning_rate": 8.487293234130283e-06, "loss": 0.7501, "step": 15514 }, { "epoch": 0.5619340818544005, "grad_norm": 1.3063703087736704, "learning_rate": 8.486133676043025e-06, "loss": 0.7155, "step": 15515 }, { "epoch": 0.561970300615719, "grad_norm": 1.5375669542326178, "learning_rate": 8.484974138787897e-06, "loss": 0.794, "step": 15516 }, { "epoch": 0.5620065193770373, "grad_norm": 1.3854543226963718, "learning_rate": 8.483814622380856e-06, "loss": 0.7778, "step": 15517 }, { "epoch": 0.5620427381383557, "grad_norm": 1.272309443574521, "learning_rate": 8.482655126837857e-06, "loss": 0.6955, "step": 15518 }, { "epoch": 0.562078956899674, "grad_norm": 1.188010831337428, "learning_rate": 8.481495652174859e-06, "loss": 0.6759, "step": 15519 }, { "epoch": 0.5621151756609923, "grad_norm": 1.1899091609965675, "learning_rate": 8.480336198407814e-06, "loss": 0.6599, "step": 15520 }, { "epoch": 0.5621513944223108, "grad_norm": 1.247079690137718, "learning_rate": 8.479176765552677e-06, "loss": 0.6973, "step": 15521 }, { "epoch": 0.5621876131836291, "grad_norm": 1.3461162385650556, "learning_rate": 8.478017353625406e-06, "loss": 0.7269, "step": 15522 }, { "epoch": 0.5622238319449475, "grad_norm": 1.5640685005149102, "learning_rate": 8.476857962641951e-06, "loss": 0.7505, "step": 15523 }, { "epoch": 0.5622600507062658, "grad_norm": 1.054112332302997, "learning_rate": 8.475698592618272e-06, "loss": 0.677, "step": 15524 }, { "epoch": 0.5622962694675842, "grad_norm": 1.3376691899510893, "learning_rate": 8.474539243570316e-06, "loss": 0.7019, "step": 15525 }, { "epoch": 0.5623324882289026, "grad_norm": 1.2063672784200568, "learning_rate": 8.473379915514045e-06, "loss": 0.632, "step": 15526 }, { "epoch": 0.5623687069902209, "grad_norm": 1.4632246822720614, "learning_rate": 8.472220608465404e-06, "loss": 0.7906, "step": 15527 }, { "epoch": 0.5624049257515393, "grad_norm": 1.4121741633504046, "learning_rate": 8.471061322440353e-06, "loss": 0.6528, "step": 15528 }, { "epoch": 0.5624411445128576, "grad_norm": 1.06459897468253, "learning_rate": 8.469902057454838e-06, "loss": 0.7551, "step": 15529 }, { "epoch": 0.562477363274176, "grad_norm": 1.147456453801587, "learning_rate": 8.468742813524817e-06, "loss": 0.7052, "step": 15530 }, { "epoch": 0.5625135820354944, "grad_norm": 1.215226348984871, "learning_rate": 8.467583590666242e-06, "loss": 0.6757, "step": 15531 }, { "epoch": 0.5625498007968127, "grad_norm": 1.4244291295591516, "learning_rate": 8.46642438889506e-06, "loss": 0.732, "step": 15532 }, { "epoch": 0.5625860195581311, "grad_norm": 0.9621059562767179, "learning_rate": 8.465265208227229e-06, "loss": 0.6911, "step": 15533 }, { "epoch": 0.5626222383194495, "grad_norm": 1.4239748701237858, "learning_rate": 8.464106048678696e-06, "loss": 0.7982, "step": 15534 }, { "epoch": 0.5626584570807679, "grad_norm": 1.0797808482287805, "learning_rate": 8.462946910265413e-06, "loss": 0.7391, "step": 15535 }, { "epoch": 0.5626946758420862, "grad_norm": 1.4078345140393531, "learning_rate": 8.46178779300333e-06, "loss": 0.721, "step": 15536 }, { "epoch": 0.5627308946034045, "grad_norm": 1.387327752907765, "learning_rate": 8.460628696908402e-06, "loss": 0.749, "step": 15537 }, { "epoch": 0.562767113364723, "grad_norm": 1.081860889418927, "learning_rate": 8.459469621996572e-06, "loss": 0.7678, "step": 15538 }, { "epoch": 0.5628033321260413, "grad_norm": 1.0047650973506357, "learning_rate": 8.458310568283797e-06, "loss": 0.7177, "step": 15539 }, { "epoch": 0.5628395508873597, "grad_norm": 1.3131081727468967, "learning_rate": 8.457151535786019e-06, "loss": 0.6982, "step": 15540 }, { "epoch": 0.562875769648678, "grad_norm": 1.269452902503754, "learning_rate": 8.455992524519194e-06, "loss": 0.7155, "step": 15541 }, { "epoch": 0.5629119884099963, "grad_norm": 0.9662795114863467, "learning_rate": 8.45483353449927e-06, "loss": 0.6866, "step": 15542 }, { "epoch": 0.5629482071713148, "grad_norm": 1.4238927699616462, "learning_rate": 8.453674565742191e-06, "loss": 0.7112, "step": 15543 }, { "epoch": 0.5629844259326331, "grad_norm": 0.982398097435919, "learning_rate": 8.452515618263912e-06, "loss": 0.6848, "step": 15544 }, { "epoch": 0.5630206446939515, "grad_norm": 1.2732241208976933, "learning_rate": 8.451356692080375e-06, "loss": 0.7507, "step": 15545 }, { "epoch": 0.5630568634552698, "grad_norm": 1.0137059419848138, "learning_rate": 8.450197787207533e-06, "loss": 0.744, "step": 15546 }, { "epoch": 0.5630930822165882, "grad_norm": 1.351017646446121, "learning_rate": 8.44903890366133e-06, "loss": 0.7462, "step": 15547 }, { "epoch": 0.5631293009779066, "grad_norm": 1.2733011392122984, "learning_rate": 8.447880041457715e-06, "loss": 0.6278, "step": 15548 }, { "epoch": 0.5631655197392249, "grad_norm": 1.3709999572701643, "learning_rate": 8.446721200612632e-06, "loss": 0.7147, "step": 15549 }, { "epoch": 0.5632017385005433, "grad_norm": 1.1234376325239563, "learning_rate": 8.445562381142035e-06, "loss": 0.7169, "step": 15550 }, { "epoch": 0.5632379572618617, "grad_norm": 1.3883531790890187, "learning_rate": 8.444403583061861e-06, "loss": 0.7197, "step": 15551 }, { "epoch": 0.56327417602318, "grad_norm": 1.351616888660825, "learning_rate": 8.443244806388062e-06, "loss": 0.6946, "step": 15552 }, { "epoch": 0.5633103947844984, "grad_norm": 1.3489492978679574, "learning_rate": 8.442086051136584e-06, "loss": 0.6801, "step": 15553 }, { "epoch": 0.5633466135458167, "grad_norm": 1.6152605333712486, "learning_rate": 8.440927317323368e-06, "loss": 0.749, "step": 15554 }, { "epoch": 0.5633828323071352, "grad_norm": 1.503853944389038, "learning_rate": 8.439768604964364e-06, "loss": 0.7824, "step": 15555 }, { "epoch": 0.5634190510684535, "grad_norm": 1.328866457833962, "learning_rate": 8.438609914075514e-06, "loss": 0.7284, "step": 15556 }, { "epoch": 0.5634552698297718, "grad_norm": 1.5761342522764594, "learning_rate": 8.437451244672766e-06, "loss": 0.7708, "step": 15557 }, { "epoch": 0.5634914885910902, "grad_norm": 1.2978591660390941, "learning_rate": 8.436292596772058e-06, "loss": 0.6709, "step": 15558 }, { "epoch": 0.5635277073524085, "grad_norm": 1.3557749809062165, "learning_rate": 8.43513397038934e-06, "loss": 0.7204, "step": 15559 }, { "epoch": 0.563563926113727, "grad_norm": 1.236573834432377, "learning_rate": 8.433975365540554e-06, "loss": 0.6986, "step": 15560 }, { "epoch": 0.5636001448750453, "grad_norm": 0.9233584739378035, "learning_rate": 8.43281678224164e-06, "loss": 0.688, "step": 15561 }, { "epoch": 0.5636363636363636, "grad_norm": 1.591260540012219, "learning_rate": 8.431658220508549e-06, "loss": 0.7693, "step": 15562 }, { "epoch": 0.563672582397682, "grad_norm": 1.752608739007103, "learning_rate": 8.430499680357215e-06, "loss": 0.79, "step": 15563 }, { "epoch": 0.5637088011590004, "grad_norm": 1.4194292108437263, "learning_rate": 8.429341161803587e-06, "loss": 0.7474, "step": 15564 }, { "epoch": 0.5637450199203188, "grad_norm": 1.282848756303358, "learning_rate": 8.428182664863602e-06, "loss": 0.7337, "step": 15565 }, { "epoch": 0.5637812386816371, "grad_norm": 1.3864495526531797, "learning_rate": 8.427024189553209e-06, "loss": 0.7525, "step": 15566 }, { "epoch": 0.5638174574429554, "grad_norm": 1.3186371348205361, "learning_rate": 8.42586573588834e-06, "loss": 0.7385, "step": 15567 }, { "epoch": 0.5638536762042738, "grad_norm": 1.100728476294801, "learning_rate": 8.424707303884945e-06, "loss": 0.688, "step": 15568 }, { "epoch": 0.5638898949655922, "grad_norm": 1.0776822129590735, "learning_rate": 8.42354889355896e-06, "loss": 0.7286, "step": 15569 }, { "epoch": 0.5639261137269106, "grad_norm": 1.327521301954411, "learning_rate": 8.42239050492633e-06, "loss": 0.7435, "step": 15570 }, { "epoch": 0.5639623324882289, "grad_norm": 1.2651394284924926, "learning_rate": 8.42123213800299e-06, "loss": 0.6971, "step": 15571 }, { "epoch": 0.5639985512495472, "grad_norm": 1.3880083045221063, "learning_rate": 8.420073792804882e-06, "loss": 0.7109, "step": 15572 }, { "epoch": 0.5640347700108657, "grad_norm": 1.2970976875321785, "learning_rate": 8.418915469347952e-06, "loss": 0.7082, "step": 15573 }, { "epoch": 0.564070988772184, "grad_norm": 1.0318494837021746, "learning_rate": 8.41775716764813e-06, "loss": 0.7016, "step": 15574 }, { "epoch": 0.5641072075335024, "grad_norm": 1.4667132819847848, "learning_rate": 8.416598887721362e-06, "loss": 0.7474, "step": 15575 }, { "epoch": 0.5641434262948207, "grad_norm": 1.4227992626152934, "learning_rate": 8.415440629583583e-06, "loss": 0.7529, "step": 15576 }, { "epoch": 0.564179645056139, "grad_norm": 1.3790359463092108, "learning_rate": 8.414282393250734e-06, "loss": 0.7342, "step": 15577 }, { "epoch": 0.5642158638174575, "grad_norm": 1.341312243884008, "learning_rate": 8.413124178738753e-06, "loss": 0.7148, "step": 15578 }, { "epoch": 0.5642520825787758, "grad_norm": 1.5186675198453936, "learning_rate": 8.411965986063578e-06, "loss": 0.7127, "step": 15579 }, { "epoch": 0.5642883013400942, "grad_norm": 1.0555189553026953, "learning_rate": 8.410807815241143e-06, "loss": 0.7483, "step": 15580 }, { "epoch": 0.5643245201014125, "grad_norm": 1.3379295023260662, "learning_rate": 8.409649666287393e-06, "loss": 0.7063, "step": 15581 }, { "epoch": 0.5643607388627309, "grad_norm": 1.3809473881062106, "learning_rate": 8.408491539218258e-06, "loss": 0.7726, "step": 15582 }, { "epoch": 0.5643969576240493, "grad_norm": 1.1938024586428053, "learning_rate": 8.407333434049678e-06, "loss": 0.6997, "step": 15583 }, { "epoch": 0.5644331763853676, "grad_norm": 1.1720698033782893, "learning_rate": 8.406175350797591e-06, "loss": 0.7074, "step": 15584 }, { "epoch": 0.564469395146686, "grad_norm": 1.188979042642082, "learning_rate": 8.405017289477931e-06, "loss": 0.7251, "step": 15585 }, { "epoch": 0.5645056139080044, "grad_norm": 1.3374338692956709, "learning_rate": 8.403859250106636e-06, "loss": 0.6791, "step": 15586 }, { "epoch": 0.5645418326693227, "grad_norm": 1.02219677111259, "learning_rate": 8.402701232699638e-06, "loss": 0.7041, "step": 15587 }, { "epoch": 0.5645780514306411, "grad_norm": 1.4016853521282178, "learning_rate": 8.401543237272877e-06, "loss": 0.7154, "step": 15588 }, { "epoch": 0.5646142701919594, "grad_norm": 1.076897218398166, "learning_rate": 8.400385263842283e-06, "loss": 0.665, "step": 15589 }, { "epoch": 0.5646504889532779, "grad_norm": 1.0518567669406864, "learning_rate": 8.399227312423796e-06, "loss": 0.6723, "step": 15590 }, { "epoch": 0.5646867077145962, "grad_norm": 1.3059498118306072, "learning_rate": 8.398069383033346e-06, "loss": 0.7696, "step": 15591 }, { "epoch": 0.5647229264759145, "grad_norm": 1.4250594049335383, "learning_rate": 8.396911475686867e-06, "loss": 0.7539, "step": 15592 }, { "epoch": 0.5647591452372329, "grad_norm": 1.3611170715963274, "learning_rate": 8.395753590400297e-06, "loss": 0.7326, "step": 15593 }, { "epoch": 0.5647953639985512, "grad_norm": 1.2187812247922463, "learning_rate": 8.394595727189568e-06, "loss": 0.6459, "step": 15594 }, { "epoch": 0.5648315827598697, "grad_norm": 1.162088489094346, "learning_rate": 8.393437886070612e-06, "loss": 0.7221, "step": 15595 }, { "epoch": 0.564867801521188, "grad_norm": 1.3256190451122256, "learning_rate": 8.392280067059359e-06, "loss": 0.7045, "step": 15596 }, { "epoch": 0.5649040202825063, "grad_norm": 1.3378049635090097, "learning_rate": 8.39112227017175e-06, "loss": 0.734, "step": 15597 }, { "epoch": 0.5649402390438247, "grad_norm": 1.3122154158636419, "learning_rate": 8.389964495423709e-06, "loss": 0.7399, "step": 15598 }, { "epoch": 0.564976457805143, "grad_norm": 1.4729273423875202, "learning_rate": 8.388806742831172e-06, "loss": 0.7012, "step": 15599 }, { "epoch": 0.5650126765664615, "grad_norm": 1.3013859637705192, "learning_rate": 8.387649012410071e-06, "loss": 0.7446, "step": 15600 }, { "epoch": 0.5650488953277798, "grad_norm": 1.388381829130916, "learning_rate": 8.386491304176337e-06, "loss": 0.7648, "step": 15601 }, { "epoch": 0.5650851140890981, "grad_norm": 1.4019666047722348, "learning_rate": 8.385333618145896e-06, "loss": 0.6855, "step": 15602 }, { "epoch": 0.5651213328504165, "grad_norm": 1.339072057379419, "learning_rate": 8.384175954334686e-06, "loss": 0.7022, "step": 15603 }, { "epoch": 0.5651575516117349, "grad_norm": 1.3658727004953581, "learning_rate": 8.383018312758635e-06, "loss": 0.7839, "step": 15604 }, { "epoch": 0.5651937703730533, "grad_norm": 1.2464643177019465, "learning_rate": 8.381860693433672e-06, "loss": 0.6382, "step": 15605 }, { "epoch": 0.5652299891343716, "grad_norm": 1.407743953174956, "learning_rate": 8.38070309637573e-06, "loss": 0.7468, "step": 15606 }, { "epoch": 0.5652662078956899, "grad_norm": 1.085718210595148, "learning_rate": 8.379545521600732e-06, "loss": 0.7252, "step": 15607 }, { "epoch": 0.5653024266570084, "grad_norm": 1.3386364799965111, "learning_rate": 8.378387969124615e-06, "loss": 0.714, "step": 15608 }, { "epoch": 0.5653386454183267, "grad_norm": 1.1944151763955735, "learning_rate": 8.377230438963302e-06, "loss": 0.6422, "step": 15609 }, { "epoch": 0.5653748641796451, "grad_norm": 1.5374163269035133, "learning_rate": 8.376072931132727e-06, "loss": 0.7314, "step": 15610 }, { "epoch": 0.5654110829409634, "grad_norm": 1.337236066790089, "learning_rate": 8.374915445648814e-06, "loss": 0.7191, "step": 15611 }, { "epoch": 0.5654473017022817, "grad_norm": 1.2590505514822996, "learning_rate": 8.373757982527494e-06, "loss": 0.6745, "step": 15612 }, { "epoch": 0.5654835204636002, "grad_norm": 1.2875426594949064, "learning_rate": 8.372600541784689e-06, "loss": 0.7053, "step": 15613 }, { "epoch": 0.5655197392249185, "grad_norm": 1.121589750453092, "learning_rate": 8.371443123436331e-06, "loss": 0.7057, "step": 15614 }, { "epoch": 0.5655559579862369, "grad_norm": 1.2534766025597297, "learning_rate": 8.370285727498349e-06, "loss": 0.6827, "step": 15615 }, { "epoch": 0.5655921767475552, "grad_norm": 1.3179026907673168, "learning_rate": 8.369128353986667e-06, "loss": 0.7149, "step": 15616 }, { "epoch": 0.5656283955088736, "grad_norm": 1.5173047462647802, "learning_rate": 8.367971002917213e-06, "loss": 0.7154, "step": 15617 }, { "epoch": 0.565664614270192, "grad_norm": 1.5270649589675216, "learning_rate": 8.366813674305909e-06, "loss": 0.6829, "step": 15618 }, { "epoch": 0.5657008330315103, "grad_norm": 1.4338368981408456, "learning_rate": 8.365656368168686e-06, "loss": 0.7414, "step": 15619 }, { "epoch": 0.5657370517928287, "grad_norm": 1.329457177986934, "learning_rate": 8.364499084521467e-06, "loss": 0.7036, "step": 15620 }, { "epoch": 0.5657732705541471, "grad_norm": 1.3300594147866038, "learning_rate": 8.363341823380178e-06, "loss": 0.6838, "step": 15621 }, { "epoch": 0.5658094893154654, "grad_norm": 1.2812245718247015, "learning_rate": 8.362184584760742e-06, "loss": 0.7546, "step": 15622 }, { "epoch": 0.5658457080767838, "grad_norm": 1.3244964075022678, "learning_rate": 8.361027368679085e-06, "loss": 0.7604, "step": 15623 }, { "epoch": 0.5658819268381021, "grad_norm": 0.9064423215979492, "learning_rate": 8.359870175151135e-06, "loss": 0.6895, "step": 15624 }, { "epoch": 0.5659181455994206, "grad_norm": 1.3802235972862003, "learning_rate": 8.358713004192809e-06, "loss": 0.7926, "step": 15625 }, { "epoch": 0.5659543643607389, "grad_norm": 1.3453608494221505, "learning_rate": 8.357555855820036e-06, "loss": 0.7314, "step": 15626 }, { "epoch": 0.5659905831220572, "grad_norm": 1.0344436950893812, "learning_rate": 8.356398730048737e-06, "loss": 0.6834, "step": 15627 }, { "epoch": 0.5660268018833756, "grad_norm": 1.3515846796365787, "learning_rate": 8.355241626894837e-06, "loss": 0.7809, "step": 15628 }, { "epoch": 0.5660630206446939, "grad_norm": 1.4143640805154591, "learning_rate": 8.354084546374255e-06, "loss": 0.7305, "step": 15629 }, { "epoch": 0.5660992394060124, "grad_norm": 1.3975436221492263, "learning_rate": 8.35292748850292e-06, "loss": 0.7771, "step": 15630 }, { "epoch": 0.5661354581673307, "grad_norm": 1.250823434425714, "learning_rate": 8.351770453296746e-06, "loss": 0.6503, "step": 15631 }, { "epoch": 0.566171676928649, "grad_norm": 1.3330726083790037, "learning_rate": 8.350613440771661e-06, "loss": 0.7305, "step": 15632 }, { "epoch": 0.5662078956899674, "grad_norm": 1.359649535935535, "learning_rate": 8.349456450943583e-06, "loss": 0.7363, "step": 15633 }, { "epoch": 0.5662441144512858, "grad_norm": 1.4830377277043338, "learning_rate": 8.348299483828433e-06, "loss": 0.7485, "step": 15634 }, { "epoch": 0.5662803332126042, "grad_norm": 1.346213862550411, "learning_rate": 8.347142539442136e-06, "loss": 0.7256, "step": 15635 }, { "epoch": 0.5663165519739225, "grad_norm": 0.8958026178171279, "learning_rate": 8.345985617800609e-06, "loss": 0.7298, "step": 15636 }, { "epoch": 0.5663527707352408, "grad_norm": 1.2888349382794235, "learning_rate": 8.344828718919775e-06, "loss": 0.7522, "step": 15637 }, { "epoch": 0.5663889894965592, "grad_norm": 1.2053290671271994, "learning_rate": 8.34367184281555e-06, "loss": 0.711, "step": 15638 }, { "epoch": 0.5664252082578776, "grad_norm": 1.2941450922065916, "learning_rate": 8.342514989503857e-06, "loss": 0.7236, "step": 15639 }, { "epoch": 0.566461427019196, "grad_norm": 0.9760032756420125, "learning_rate": 8.341358159000613e-06, "loss": 0.7737, "step": 15640 }, { "epoch": 0.5664976457805143, "grad_norm": 1.4378334719836978, "learning_rate": 8.340201351321741e-06, "loss": 0.7426, "step": 15641 }, { "epoch": 0.5665338645418326, "grad_norm": 1.2602598115530668, "learning_rate": 8.339044566483153e-06, "loss": 0.7505, "step": 15642 }, { "epoch": 0.5665700833031511, "grad_norm": 1.4302992163489905, "learning_rate": 8.337887804500773e-06, "loss": 0.6758, "step": 15643 }, { "epoch": 0.5666063020644694, "grad_norm": 1.294572086760628, "learning_rate": 8.336731065390517e-06, "loss": 0.7775, "step": 15644 }, { "epoch": 0.5666425208257878, "grad_norm": 1.5776782760884687, "learning_rate": 8.335574349168303e-06, "loss": 0.7019, "step": 15645 }, { "epoch": 0.5666787395871061, "grad_norm": 1.5274868407868865, "learning_rate": 8.33441765585005e-06, "loss": 0.678, "step": 15646 }, { "epoch": 0.5667149583484244, "grad_norm": 1.4875003193856948, "learning_rate": 8.33326098545167e-06, "loss": 0.688, "step": 15647 }, { "epoch": 0.5667511771097429, "grad_norm": 1.292314440213034, "learning_rate": 8.332104337989086e-06, "loss": 0.7543, "step": 15648 }, { "epoch": 0.5667873958710612, "grad_norm": 1.2395373330543444, "learning_rate": 8.330947713478212e-06, "loss": 0.7189, "step": 15649 }, { "epoch": 0.5668236146323796, "grad_norm": 1.268589039866751, "learning_rate": 8.329791111934963e-06, "loss": 0.6952, "step": 15650 }, { "epoch": 0.5668598333936979, "grad_norm": 1.303523528098251, "learning_rate": 8.328634533375258e-06, "loss": 0.7253, "step": 15651 }, { "epoch": 0.5668960521550163, "grad_norm": 1.324550885208751, "learning_rate": 8.327477977815009e-06, "loss": 0.7132, "step": 15652 }, { "epoch": 0.5669322709163347, "grad_norm": 1.1975335930525393, "learning_rate": 8.32632144527013e-06, "loss": 0.7194, "step": 15653 }, { "epoch": 0.566968489677653, "grad_norm": 1.4575112551864728, "learning_rate": 8.32516493575654e-06, "loss": 0.6645, "step": 15654 }, { "epoch": 0.5670047084389714, "grad_norm": 1.402486202846582, "learning_rate": 8.324008449290155e-06, "loss": 0.7205, "step": 15655 }, { "epoch": 0.5670409272002898, "grad_norm": 1.42760946429273, "learning_rate": 8.322851985886884e-06, "loss": 0.8359, "step": 15656 }, { "epoch": 0.5670771459616081, "grad_norm": 1.3429971224266002, "learning_rate": 8.321695545562644e-06, "loss": 0.789, "step": 15657 }, { "epoch": 0.5671133647229265, "grad_norm": 1.260672016983343, "learning_rate": 8.320539128333346e-06, "loss": 0.6538, "step": 15658 }, { "epoch": 0.5671495834842448, "grad_norm": 1.4066817111048195, "learning_rate": 8.319382734214909e-06, "loss": 0.6657, "step": 15659 }, { "epoch": 0.5671858022455633, "grad_norm": 1.3480019036283215, "learning_rate": 8.31822636322324e-06, "loss": 0.7, "step": 15660 }, { "epoch": 0.5672220210068816, "grad_norm": 1.4265521251764988, "learning_rate": 8.317070015374256e-06, "loss": 0.7749, "step": 15661 }, { "epoch": 0.5672582397681999, "grad_norm": 1.3172134215574824, "learning_rate": 8.315913690683864e-06, "loss": 0.6752, "step": 15662 }, { "epoch": 0.5672944585295183, "grad_norm": 1.459298647694604, "learning_rate": 8.314757389167983e-06, "loss": 0.8271, "step": 15663 }, { "epoch": 0.5673306772908366, "grad_norm": 1.3221220124435356, "learning_rate": 8.313601110842517e-06, "loss": 0.7123, "step": 15664 }, { "epoch": 0.5673668960521551, "grad_norm": 1.2602122217531841, "learning_rate": 8.312444855723382e-06, "loss": 0.67, "step": 15665 }, { "epoch": 0.5674031148134734, "grad_norm": 1.2629441490358884, "learning_rate": 8.311288623826493e-06, "loss": 0.7062, "step": 15666 }, { "epoch": 0.5674393335747917, "grad_norm": 1.4984665274170301, "learning_rate": 8.310132415167752e-06, "loss": 0.7192, "step": 15667 }, { "epoch": 0.5674755523361101, "grad_norm": 1.3397903390031165, "learning_rate": 8.308976229763076e-06, "loss": 0.7492, "step": 15668 }, { "epoch": 0.5675117710974285, "grad_norm": 1.288290318129822, "learning_rate": 8.307820067628371e-06, "loss": 0.6985, "step": 15669 }, { "epoch": 0.5675479898587469, "grad_norm": 1.4452523077209944, "learning_rate": 8.306663928779553e-06, "loss": 0.7704, "step": 15670 }, { "epoch": 0.5675842086200652, "grad_norm": 1.2416756740592452, "learning_rate": 8.305507813232522e-06, "loss": 0.6968, "step": 15671 }, { "epoch": 0.5676204273813835, "grad_norm": 1.2651047314244352, "learning_rate": 8.304351721003197e-06, "loss": 0.7065, "step": 15672 }, { "epoch": 0.567656646142702, "grad_norm": 1.283463550367804, "learning_rate": 8.303195652107478e-06, "loss": 0.7128, "step": 15673 }, { "epoch": 0.5676928649040203, "grad_norm": 1.3496828556069205, "learning_rate": 8.302039606561282e-06, "loss": 0.8294, "step": 15674 }, { "epoch": 0.5677290836653387, "grad_norm": 1.3823887416828107, "learning_rate": 8.30088358438051e-06, "loss": 0.7339, "step": 15675 }, { "epoch": 0.567765302426657, "grad_norm": 1.3428731368072016, "learning_rate": 8.299727585581073e-06, "loss": 0.7322, "step": 15676 }, { "epoch": 0.5678015211879753, "grad_norm": 1.4398067381640416, "learning_rate": 8.298571610178879e-06, "loss": 0.7322, "step": 15677 }, { "epoch": 0.5678377399492938, "grad_norm": 0.9955901666964803, "learning_rate": 8.297415658189834e-06, "loss": 0.7119, "step": 15678 }, { "epoch": 0.5678739587106121, "grad_norm": 1.2743371947627, "learning_rate": 8.296259729629847e-06, "loss": 0.7243, "step": 15679 }, { "epoch": 0.5679101774719305, "grad_norm": 1.35498165096197, "learning_rate": 8.295103824514821e-06, "loss": 0.7132, "step": 15680 }, { "epoch": 0.5679463962332488, "grad_norm": 1.2948738743910193, "learning_rate": 8.293947942860667e-06, "loss": 0.7502, "step": 15681 }, { "epoch": 0.5679826149945671, "grad_norm": 1.5931350547765806, "learning_rate": 8.292792084683287e-06, "loss": 0.7728, "step": 15682 }, { "epoch": 0.5680188337558856, "grad_norm": 1.3684588651781608, "learning_rate": 8.291636249998588e-06, "loss": 0.7317, "step": 15683 }, { "epoch": 0.5680550525172039, "grad_norm": 1.3631193692028947, "learning_rate": 8.290480438822476e-06, "loss": 0.7483, "step": 15684 }, { "epoch": 0.5680912712785223, "grad_norm": 1.4392103018847673, "learning_rate": 8.289324651170853e-06, "loss": 0.7598, "step": 15685 }, { "epoch": 0.5681274900398406, "grad_norm": 1.3282036197056277, "learning_rate": 8.288168887059629e-06, "loss": 0.6871, "step": 15686 }, { "epoch": 0.568163708801159, "grad_norm": 1.31674027308551, "learning_rate": 8.287013146504703e-06, "loss": 0.6981, "step": 15687 }, { "epoch": 0.5681999275624774, "grad_norm": 1.269282805268037, "learning_rate": 8.285857429521983e-06, "loss": 0.6799, "step": 15688 }, { "epoch": 0.5682361463237957, "grad_norm": 1.6570860196764763, "learning_rate": 8.28470173612737e-06, "loss": 0.7626, "step": 15689 }, { "epoch": 0.5682723650851141, "grad_norm": 1.2954717521531471, "learning_rate": 8.28354606633677e-06, "loss": 0.693, "step": 15690 }, { "epoch": 0.5683085838464325, "grad_norm": 1.4363504638885118, "learning_rate": 8.28239042016608e-06, "loss": 0.6915, "step": 15691 }, { "epoch": 0.5683448026077508, "grad_norm": 1.4182693330996003, "learning_rate": 8.281234797631212e-06, "loss": 0.6668, "step": 15692 }, { "epoch": 0.5683810213690692, "grad_norm": 1.248956773215166, "learning_rate": 8.280079198748061e-06, "loss": 0.6637, "step": 15693 }, { "epoch": 0.5684172401303875, "grad_norm": 1.332304244159785, "learning_rate": 8.278923623532532e-06, "loss": 0.724, "step": 15694 }, { "epoch": 0.568453458891706, "grad_norm": 1.4020286055860596, "learning_rate": 8.277768072000526e-06, "loss": 0.7736, "step": 15695 }, { "epoch": 0.5684896776530243, "grad_norm": 1.2661968222765017, "learning_rate": 8.276612544167947e-06, "loss": 0.6636, "step": 15696 }, { "epoch": 0.5685258964143426, "grad_norm": 1.3968320240433287, "learning_rate": 8.27545704005069e-06, "loss": 0.7157, "step": 15697 }, { "epoch": 0.568562115175661, "grad_norm": 1.262179022059499, "learning_rate": 8.274301559664662e-06, "loss": 0.7056, "step": 15698 }, { "epoch": 0.5685983339369793, "grad_norm": 1.196609556318488, "learning_rate": 8.27314610302576e-06, "loss": 0.6991, "step": 15699 }, { "epoch": 0.5686345526982978, "grad_norm": 1.337084039787432, "learning_rate": 8.271990670149885e-06, "loss": 0.7992, "step": 15700 }, { "epoch": 0.5686707714596161, "grad_norm": 1.285006092598402, "learning_rate": 8.270835261052935e-06, "loss": 0.6646, "step": 15701 }, { "epoch": 0.5687069902209344, "grad_norm": 1.3943427944268563, "learning_rate": 8.269679875750814e-06, "loss": 0.7537, "step": 15702 }, { "epoch": 0.5687432089822528, "grad_norm": 1.0267513357994147, "learning_rate": 8.268524514259416e-06, "loss": 0.6851, "step": 15703 }, { "epoch": 0.5687794277435712, "grad_norm": 0.998085676869483, "learning_rate": 8.267369176594643e-06, "loss": 0.6751, "step": 15704 }, { "epoch": 0.5688156465048896, "grad_norm": 1.5614941446921782, "learning_rate": 8.266213862772393e-06, "loss": 0.7324, "step": 15705 }, { "epoch": 0.5688518652662079, "grad_norm": 1.1880585275464919, "learning_rate": 8.265058572808565e-06, "loss": 0.7224, "step": 15706 }, { "epoch": 0.5688880840275262, "grad_norm": 1.381298957659378, "learning_rate": 8.263903306719052e-06, "loss": 0.8232, "step": 15707 }, { "epoch": 0.5689243027888446, "grad_norm": 1.3149793656936772, "learning_rate": 8.262748064519754e-06, "loss": 0.7698, "step": 15708 }, { "epoch": 0.568960521550163, "grad_norm": 1.381755203746289, "learning_rate": 8.261592846226574e-06, "loss": 0.6812, "step": 15709 }, { "epoch": 0.5689967403114814, "grad_norm": 1.3390639638391992, "learning_rate": 8.2604376518554e-06, "loss": 0.6779, "step": 15710 }, { "epoch": 0.5690329590727997, "grad_norm": 1.461946160901195, "learning_rate": 8.259282481422134e-06, "loss": 0.6995, "step": 15711 }, { "epoch": 0.569069177834118, "grad_norm": 0.9257344687287463, "learning_rate": 8.258127334942672e-06, "loss": 0.7158, "step": 15712 }, { "epoch": 0.5691053965954365, "grad_norm": 1.0136858445557375, "learning_rate": 8.256972212432908e-06, "loss": 0.7109, "step": 15713 }, { "epoch": 0.5691416153567548, "grad_norm": 1.2902597459521281, "learning_rate": 8.255817113908735e-06, "loss": 0.7295, "step": 15714 }, { "epoch": 0.5691778341180732, "grad_norm": 1.3719530700332547, "learning_rate": 8.254662039386055e-06, "loss": 0.6823, "step": 15715 }, { "epoch": 0.5692140528793915, "grad_norm": 1.2150321899395071, "learning_rate": 8.253506988880757e-06, "loss": 0.7046, "step": 15716 }, { "epoch": 0.5692502716407098, "grad_norm": 1.2664974569861969, "learning_rate": 8.252351962408738e-06, "loss": 0.6741, "step": 15717 }, { "epoch": 0.5692864904020283, "grad_norm": 1.4198626174000615, "learning_rate": 8.251196959985889e-06, "loss": 0.7776, "step": 15718 }, { "epoch": 0.5693227091633466, "grad_norm": 1.2338852408747094, "learning_rate": 8.250041981628107e-06, "loss": 0.6895, "step": 15719 }, { "epoch": 0.569358927924665, "grad_norm": 1.017628716016141, "learning_rate": 8.24888702735129e-06, "loss": 0.7758, "step": 15720 }, { "epoch": 0.5693951466859833, "grad_norm": 1.3238955619578228, "learning_rate": 8.24773209717132e-06, "loss": 0.765, "step": 15721 }, { "epoch": 0.5694313654473017, "grad_norm": 1.2711526376439544, "learning_rate": 8.246577191104101e-06, "loss": 0.6959, "step": 15722 }, { "epoch": 0.5694675842086201, "grad_norm": 0.9554790288509908, "learning_rate": 8.245422309165516e-06, "loss": 0.7251, "step": 15723 }, { "epoch": 0.5695038029699384, "grad_norm": 1.3771624896320334, "learning_rate": 8.244267451371465e-06, "loss": 0.7895, "step": 15724 }, { "epoch": 0.5695400217312568, "grad_norm": 1.2551930010783838, "learning_rate": 8.243112617737835e-06, "loss": 0.6373, "step": 15725 }, { "epoch": 0.5695762404925752, "grad_norm": 1.5331584194979209, "learning_rate": 8.241957808280522e-06, "loss": 0.7749, "step": 15726 }, { "epoch": 0.5696124592538935, "grad_norm": 1.338666544361816, "learning_rate": 8.24080302301541e-06, "loss": 0.7768, "step": 15727 }, { "epoch": 0.5696486780152119, "grad_norm": 1.139822024161388, "learning_rate": 8.239648261958394e-06, "loss": 0.7496, "step": 15728 }, { "epoch": 0.5696848967765302, "grad_norm": 1.3234179341346015, "learning_rate": 8.238493525125369e-06, "loss": 0.7269, "step": 15729 }, { "epoch": 0.5697211155378487, "grad_norm": 1.160812926604122, "learning_rate": 8.237338812532217e-06, "loss": 0.677, "step": 15730 }, { "epoch": 0.569757334299167, "grad_norm": 1.3109069718931334, "learning_rate": 8.236184124194834e-06, "loss": 0.7718, "step": 15731 }, { "epoch": 0.5697935530604853, "grad_norm": 1.6053655896085801, "learning_rate": 8.235029460129108e-06, "loss": 0.6811, "step": 15732 }, { "epoch": 0.5698297718218037, "grad_norm": 1.2964144339176136, "learning_rate": 8.233874820350926e-06, "loss": 0.6647, "step": 15733 }, { "epoch": 0.569865990583122, "grad_norm": 1.092351016837364, "learning_rate": 8.232720204876178e-06, "loss": 0.6912, "step": 15734 }, { "epoch": 0.5699022093444405, "grad_norm": 1.0776502401422319, "learning_rate": 8.231565613720755e-06, "loss": 0.7138, "step": 15735 }, { "epoch": 0.5699384281057588, "grad_norm": 1.4294032533461223, "learning_rate": 8.23041104690054e-06, "loss": 0.7736, "step": 15736 }, { "epoch": 0.5699746468670771, "grad_norm": 1.2743166290687602, "learning_rate": 8.229256504431426e-06, "loss": 0.6896, "step": 15737 }, { "epoch": 0.5700108656283955, "grad_norm": 1.358928091068422, "learning_rate": 8.228101986329297e-06, "loss": 0.7803, "step": 15738 }, { "epoch": 0.5700470843897139, "grad_norm": 1.436984485467974, "learning_rate": 8.226947492610041e-06, "loss": 0.7675, "step": 15739 }, { "epoch": 0.5700833031510323, "grad_norm": 1.332861552755687, "learning_rate": 8.22579302328955e-06, "loss": 0.6764, "step": 15740 }, { "epoch": 0.5701195219123506, "grad_norm": 1.3016879806332218, "learning_rate": 8.2246385783837e-06, "loss": 0.7209, "step": 15741 }, { "epoch": 0.5701557406736689, "grad_norm": 1.2973135869390384, "learning_rate": 8.223484157908386e-06, "loss": 0.734, "step": 15742 }, { "epoch": 0.5701919594349873, "grad_norm": 1.364779984764788, "learning_rate": 8.22232976187949e-06, "loss": 0.7375, "step": 15743 }, { "epoch": 0.5702281781963057, "grad_norm": 1.2819534012243634, "learning_rate": 8.2211753903129e-06, "loss": 0.7267, "step": 15744 }, { "epoch": 0.5702643969576241, "grad_norm": 1.3951760679228706, "learning_rate": 8.2200210432245e-06, "loss": 0.7646, "step": 15745 }, { "epoch": 0.5703006157189424, "grad_norm": 1.029211272497488, "learning_rate": 8.218866720630174e-06, "loss": 0.6509, "step": 15746 }, { "epoch": 0.5703368344802607, "grad_norm": 1.3385578169541061, "learning_rate": 8.217712422545805e-06, "loss": 0.7531, "step": 15747 }, { "epoch": 0.5703730532415792, "grad_norm": 1.1947465239200086, "learning_rate": 8.216558148987282e-06, "loss": 0.6647, "step": 15748 }, { "epoch": 0.5704092720028975, "grad_norm": 1.464679921207716, "learning_rate": 8.215403899970483e-06, "loss": 0.7472, "step": 15749 }, { "epoch": 0.5704454907642159, "grad_norm": 1.3721758486251556, "learning_rate": 8.214249675511295e-06, "loss": 0.7344, "step": 15750 }, { "epoch": 0.5704817095255342, "grad_norm": 1.479256025041924, "learning_rate": 8.213095475625602e-06, "loss": 0.7769, "step": 15751 }, { "epoch": 0.5705179282868525, "grad_norm": 1.3681757284534293, "learning_rate": 8.211941300329284e-06, "loss": 0.7139, "step": 15752 }, { "epoch": 0.570554147048171, "grad_norm": 1.089742815435124, "learning_rate": 8.210787149638227e-06, "loss": 0.7491, "step": 15753 }, { "epoch": 0.5705903658094893, "grad_norm": 1.3181578051155305, "learning_rate": 8.20963302356831e-06, "loss": 0.6701, "step": 15754 }, { "epoch": 0.5706265845708077, "grad_norm": 1.4839353261595194, "learning_rate": 8.208478922135416e-06, "loss": 0.6863, "step": 15755 }, { "epoch": 0.570662803332126, "grad_norm": 1.3499733803859817, "learning_rate": 8.207324845355425e-06, "loss": 0.7674, "step": 15756 }, { "epoch": 0.5706990220934444, "grad_norm": 1.3361069525781388, "learning_rate": 8.20617079324422e-06, "loss": 0.7826, "step": 15757 }, { "epoch": 0.5707352408547628, "grad_norm": 1.343269493027837, "learning_rate": 8.205016765817679e-06, "loss": 0.7578, "step": 15758 }, { "epoch": 0.5707714596160811, "grad_norm": 1.2335257611736192, "learning_rate": 8.203862763091686e-06, "loss": 0.689, "step": 15759 }, { "epoch": 0.5708076783773995, "grad_norm": 1.4123844397216576, "learning_rate": 8.202708785082122e-06, "loss": 0.7634, "step": 15760 }, { "epoch": 0.5708438971387179, "grad_norm": 1.1567036376763595, "learning_rate": 8.20155483180486e-06, "loss": 0.7112, "step": 15761 }, { "epoch": 0.5708801159000362, "grad_norm": 1.422148864828386, "learning_rate": 8.200400903275788e-06, "loss": 0.7501, "step": 15762 }, { "epoch": 0.5709163346613546, "grad_norm": 1.3689020414258333, "learning_rate": 8.199246999510779e-06, "loss": 0.7138, "step": 15763 }, { "epoch": 0.5709525534226729, "grad_norm": 1.4373856931829057, "learning_rate": 8.198093120525713e-06, "loss": 0.67, "step": 15764 }, { "epoch": 0.5709887721839914, "grad_norm": 1.822472944865001, "learning_rate": 8.19693926633647e-06, "loss": 0.8011, "step": 15765 }, { "epoch": 0.5710249909453097, "grad_norm": 1.3454876493287948, "learning_rate": 8.195785436958928e-06, "loss": 0.6833, "step": 15766 }, { "epoch": 0.571061209706628, "grad_norm": 1.2141435029403385, "learning_rate": 8.19463163240896e-06, "loss": 0.7204, "step": 15767 }, { "epoch": 0.5710974284679464, "grad_norm": 1.0403162109569897, "learning_rate": 8.193477852702451e-06, "loss": 0.6917, "step": 15768 }, { "epoch": 0.5711336472292647, "grad_norm": 1.3825995838991587, "learning_rate": 8.19232409785527e-06, "loss": 0.7521, "step": 15769 }, { "epoch": 0.5711698659905832, "grad_norm": 1.4195772018983435, "learning_rate": 8.1911703678833e-06, "loss": 0.7503, "step": 15770 }, { "epoch": 0.5712060847519015, "grad_norm": 1.424336627794427, "learning_rate": 8.190016662802417e-06, "loss": 0.7605, "step": 15771 }, { "epoch": 0.5712423035132198, "grad_norm": 1.327963680399491, "learning_rate": 8.188862982628492e-06, "loss": 0.7339, "step": 15772 }, { "epoch": 0.5712785222745382, "grad_norm": 1.3063102911879425, "learning_rate": 8.187709327377407e-06, "loss": 0.8045, "step": 15773 }, { "epoch": 0.5713147410358566, "grad_norm": 1.4250990757621116, "learning_rate": 8.186555697065031e-06, "loss": 0.7395, "step": 15774 }, { "epoch": 0.571350959797175, "grad_norm": 1.440531015226578, "learning_rate": 8.185402091707244e-06, "loss": 0.7187, "step": 15775 }, { "epoch": 0.5713871785584933, "grad_norm": 0.9011066550125855, "learning_rate": 8.184248511319917e-06, "loss": 0.6761, "step": 15776 }, { "epoch": 0.5714233973198116, "grad_norm": 1.2822755948037459, "learning_rate": 8.18309495591893e-06, "loss": 0.762, "step": 15777 }, { "epoch": 0.57145961608113, "grad_norm": 0.8880248696015474, "learning_rate": 8.181941425520148e-06, "loss": 0.7329, "step": 15778 }, { "epoch": 0.5714958348424484, "grad_norm": 1.6003585593409944, "learning_rate": 8.180787920139452e-06, "loss": 0.7668, "step": 15779 }, { "epoch": 0.5715320536037668, "grad_norm": 1.031397715440579, "learning_rate": 8.179634439792712e-06, "loss": 0.6973, "step": 15780 }, { "epoch": 0.5715682723650851, "grad_norm": 1.4302617408964915, "learning_rate": 8.1784809844958e-06, "loss": 0.7655, "step": 15781 }, { "epoch": 0.5716044911264034, "grad_norm": 0.9849556411278902, "learning_rate": 8.177327554264592e-06, "loss": 0.7569, "step": 15782 }, { "epoch": 0.5716407098877219, "grad_norm": 1.25670570321787, "learning_rate": 8.176174149114957e-06, "loss": 0.7704, "step": 15783 }, { "epoch": 0.5716769286490402, "grad_norm": 1.3526045304852294, "learning_rate": 8.175020769062772e-06, "loss": 0.761, "step": 15784 }, { "epoch": 0.5717131474103586, "grad_norm": 1.3036446379473066, "learning_rate": 8.1738674141239e-06, "loss": 0.746, "step": 15785 }, { "epoch": 0.5717493661716769, "grad_norm": 1.4343837411568439, "learning_rate": 8.17271408431422e-06, "loss": 0.7514, "step": 15786 }, { "epoch": 0.5717855849329953, "grad_norm": 1.4272804386627802, "learning_rate": 8.171560779649598e-06, "loss": 0.843, "step": 15787 }, { "epoch": 0.5718218036943137, "grad_norm": 1.1136066786705985, "learning_rate": 8.17040750014591e-06, "loss": 0.7101, "step": 15788 }, { "epoch": 0.571858022455632, "grad_norm": 1.2437886439363126, "learning_rate": 8.169254245819018e-06, "loss": 0.7216, "step": 15789 }, { "epoch": 0.5718942412169504, "grad_norm": 1.3350791941561524, "learning_rate": 8.168101016684796e-06, "loss": 0.7489, "step": 15790 }, { "epoch": 0.5719304599782687, "grad_norm": 1.3429955219065788, "learning_rate": 8.166947812759118e-06, "loss": 0.7, "step": 15791 }, { "epoch": 0.5719666787395871, "grad_norm": 1.3604088878670864, "learning_rate": 8.165794634057843e-06, "loss": 0.7252, "step": 15792 }, { "epoch": 0.5720028975009055, "grad_norm": 1.4098299446770237, "learning_rate": 8.164641480596852e-06, "loss": 0.784, "step": 15793 }, { "epoch": 0.5720391162622238, "grad_norm": 1.05925063054893, "learning_rate": 8.163488352392001e-06, "loss": 0.7443, "step": 15794 }, { "epoch": 0.5720753350235422, "grad_norm": 1.4864923960354595, "learning_rate": 8.162335249459169e-06, "loss": 0.7806, "step": 15795 }, { "epoch": 0.5721115537848606, "grad_norm": 1.2618098469671393, "learning_rate": 8.161182171814214e-06, "loss": 0.6453, "step": 15796 }, { "epoch": 0.5721477725461789, "grad_norm": 1.3363876058570725, "learning_rate": 8.160029119473014e-06, "loss": 0.6951, "step": 15797 }, { "epoch": 0.5721839913074973, "grad_norm": 1.264950461849617, "learning_rate": 8.158876092451424e-06, "loss": 0.6652, "step": 15798 }, { "epoch": 0.5722202100688156, "grad_norm": 1.4578012446575148, "learning_rate": 8.157723090765321e-06, "loss": 0.7059, "step": 15799 }, { "epoch": 0.5722564288301341, "grad_norm": 1.3977806817142524, "learning_rate": 8.156570114430564e-06, "loss": 0.6735, "step": 15800 }, { "epoch": 0.5722926475914524, "grad_norm": 1.3218009419584114, "learning_rate": 8.155417163463022e-06, "loss": 0.6862, "step": 15801 }, { "epoch": 0.5723288663527707, "grad_norm": 1.5224187330032573, "learning_rate": 8.154264237878564e-06, "loss": 0.6161, "step": 15802 }, { "epoch": 0.5723650851140891, "grad_norm": 1.3420491522795117, "learning_rate": 8.153111337693047e-06, "loss": 0.7147, "step": 15803 }, { "epoch": 0.5724013038754074, "grad_norm": 1.2869919562723828, "learning_rate": 8.151958462922345e-06, "loss": 0.7016, "step": 15804 }, { "epoch": 0.5724375226367259, "grad_norm": 1.2580764061241052, "learning_rate": 8.150805613582317e-06, "loss": 0.6721, "step": 15805 }, { "epoch": 0.5724737413980442, "grad_norm": 1.33232509960002, "learning_rate": 8.149652789688828e-06, "loss": 0.7095, "step": 15806 }, { "epoch": 0.5725099601593625, "grad_norm": 1.3717052612136835, "learning_rate": 8.148499991257744e-06, "loss": 0.7202, "step": 15807 }, { "epoch": 0.5725461789206809, "grad_norm": 1.208799432773827, "learning_rate": 8.147347218304927e-06, "loss": 0.6689, "step": 15808 }, { "epoch": 0.5725823976819993, "grad_norm": 1.0167390099932527, "learning_rate": 8.146194470846237e-06, "loss": 0.6949, "step": 15809 }, { "epoch": 0.5726186164433177, "grad_norm": 1.2166093030949103, "learning_rate": 8.145041748897544e-06, "loss": 0.6722, "step": 15810 }, { "epoch": 0.572654835204636, "grad_norm": 1.44991584165699, "learning_rate": 8.143889052474702e-06, "loss": 0.7036, "step": 15811 }, { "epoch": 0.5726910539659543, "grad_norm": 1.3654856764334349, "learning_rate": 8.14273638159358e-06, "loss": 0.7739, "step": 15812 }, { "epoch": 0.5727272727272728, "grad_norm": 1.4598274246457648, "learning_rate": 8.141583736270039e-06, "loss": 0.7859, "step": 15813 }, { "epoch": 0.5727634914885911, "grad_norm": 1.3253140807429078, "learning_rate": 8.140431116519935e-06, "loss": 0.7167, "step": 15814 }, { "epoch": 0.5727997102499095, "grad_norm": 1.2747389251452712, "learning_rate": 8.139278522359137e-06, "loss": 0.7047, "step": 15815 }, { "epoch": 0.5728359290112278, "grad_norm": 1.5596760753965255, "learning_rate": 8.138125953803498e-06, "loss": 0.7253, "step": 15816 }, { "epoch": 0.5728721477725461, "grad_norm": 1.439295808019068, "learning_rate": 8.136973410868884e-06, "loss": 0.6597, "step": 15817 }, { "epoch": 0.5729083665338646, "grad_norm": 1.3601147732444425, "learning_rate": 8.135820893571152e-06, "loss": 0.7103, "step": 15818 }, { "epoch": 0.5729445852951829, "grad_norm": 1.5008594526819363, "learning_rate": 8.134668401926164e-06, "loss": 0.7547, "step": 15819 }, { "epoch": 0.5729808040565013, "grad_norm": 1.3807093965105957, "learning_rate": 8.133515935949776e-06, "loss": 0.6665, "step": 15820 }, { "epoch": 0.5730170228178196, "grad_norm": 1.4991126465734939, "learning_rate": 8.132363495657849e-06, "loss": 0.7282, "step": 15821 }, { "epoch": 0.573053241579138, "grad_norm": 1.2936020910558819, "learning_rate": 8.131211081066241e-06, "loss": 0.6955, "step": 15822 }, { "epoch": 0.5730894603404564, "grad_norm": 1.2641645921722888, "learning_rate": 8.13005869219081e-06, "loss": 0.704, "step": 15823 }, { "epoch": 0.5731256791017747, "grad_norm": 1.5141326547515948, "learning_rate": 8.128906329047418e-06, "loss": 0.7732, "step": 15824 }, { "epoch": 0.5731618978630931, "grad_norm": 1.4485213691643206, "learning_rate": 8.127753991651916e-06, "loss": 0.7063, "step": 15825 }, { "epoch": 0.5731981166244114, "grad_norm": 1.143957380706921, "learning_rate": 8.126601680020165e-06, "loss": 0.6315, "step": 15826 }, { "epoch": 0.5732343353857298, "grad_norm": 1.4899816348737, "learning_rate": 8.125449394168021e-06, "loss": 0.7185, "step": 15827 }, { "epoch": 0.5732705541470482, "grad_norm": 1.1127000459106142, "learning_rate": 8.124297134111341e-06, "loss": 0.7456, "step": 15828 }, { "epoch": 0.5733067729083665, "grad_norm": 1.046772612756449, "learning_rate": 8.123144899865979e-06, "loss": 0.6878, "step": 15829 }, { "epoch": 0.5733429916696849, "grad_norm": 1.235093830197485, "learning_rate": 8.121992691447793e-06, "loss": 0.7264, "step": 15830 }, { "epoch": 0.5733792104310033, "grad_norm": 1.1601032247370673, "learning_rate": 8.120840508872639e-06, "loss": 0.7157, "step": 15831 }, { "epoch": 0.5734154291923216, "grad_norm": 1.5459537616693377, "learning_rate": 8.119688352156369e-06, "loss": 0.7389, "step": 15832 }, { "epoch": 0.57345164795364, "grad_norm": 1.2784529749004352, "learning_rate": 8.11853622131484e-06, "loss": 0.7111, "step": 15833 }, { "epoch": 0.5734878667149583, "grad_norm": 1.1365487420747238, "learning_rate": 8.117384116363905e-06, "loss": 0.6698, "step": 15834 }, { "epoch": 0.5735240854762768, "grad_norm": 1.378491320332282, "learning_rate": 8.11623203731942e-06, "loss": 0.653, "step": 15835 }, { "epoch": 0.5735603042375951, "grad_norm": 1.4450541806705628, "learning_rate": 8.115079984197238e-06, "loss": 0.7724, "step": 15836 }, { "epoch": 0.5735965229989134, "grad_norm": 1.4202466354173908, "learning_rate": 8.113927957013212e-06, "loss": 0.7806, "step": 15837 }, { "epoch": 0.5736327417602318, "grad_norm": 1.3393904416826925, "learning_rate": 8.112775955783192e-06, "loss": 0.7224, "step": 15838 }, { "epoch": 0.5736689605215501, "grad_norm": 1.4906441923721314, "learning_rate": 8.111623980523036e-06, "loss": 0.7806, "step": 15839 }, { "epoch": 0.5737051792828686, "grad_norm": 1.4082049238852177, "learning_rate": 8.11047203124859e-06, "loss": 0.6699, "step": 15840 }, { "epoch": 0.5737413980441869, "grad_norm": 1.3960831305799608, "learning_rate": 8.109320107975711e-06, "loss": 0.7263, "step": 15841 }, { "epoch": 0.5737776168055052, "grad_norm": 1.2765039656711883, "learning_rate": 8.10816821072025e-06, "loss": 0.7847, "step": 15842 }, { "epoch": 0.5738138355668236, "grad_norm": 1.5574248168205682, "learning_rate": 8.107016339498053e-06, "loss": 0.7666, "step": 15843 }, { "epoch": 0.573850054328142, "grad_norm": 1.3333351795238388, "learning_rate": 8.105864494324979e-06, "loss": 0.7392, "step": 15844 }, { "epoch": 0.5738862730894604, "grad_norm": 1.2607086355943837, "learning_rate": 8.104712675216871e-06, "loss": 0.7331, "step": 15845 }, { "epoch": 0.5739224918507787, "grad_norm": 2.192378431965757, "learning_rate": 8.103560882189582e-06, "loss": 0.7371, "step": 15846 }, { "epoch": 0.573958710612097, "grad_norm": 1.320840281461718, "learning_rate": 8.102409115258962e-06, "loss": 0.7194, "step": 15847 }, { "epoch": 0.5739949293734155, "grad_norm": 1.4330446463517763, "learning_rate": 8.10125737444086e-06, "loss": 0.6984, "step": 15848 }, { "epoch": 0.5740311481347338, "grad_norm": 2.4312256577702445, "learning_rate": 8.100105659751126e-06, "loss": 0.6978, "step": 15849 }, { "epoch": 0.5740673668960522, "grad_norm": 1.4339925525715798, "learning_rate": 8.098953971205607e-06, "loss": 0.6662, "step": 15850 }, { "epoch": 0.5741035856573705, "grad_norm": 1.36657513460298, "learning_rate": 8.09780230882015e-06, "loss": 0.6878, "step": 15851 }, { "epoch": 0.5741398044186888, "grad_norm": 1.5948647102278453, "learning_rate": 8.096650672610605e-06, "loss": 0.7647, "step": 15852 }, { "epoch": 0.5741760231800073, "grad_norm": 1.1117810390084237, "learning_rate": 8.095499062592822e-06, "loss": 0.6592, "step": 15853 }, { "epoch": 0.5742122419413256, "grad_norm": 1.4830243340210876, "learning_rate": 8.094347478782643e-06, "loss": 0.7895, "step": 15854 }, { "epoch": 0.574248460702644, "grad_norm": 1.423766142669492, "learning_rate": 8.093195921195919e-06, "loss": 0.7319, "step": 15855 }, { "epoch": 0.5742846794639623, "grad_norm": 1.2867629897683184, "learning_rate": 8.092044389848491e-06, "loss": 0.6908, "step": 15856 }, { "epoch": 0.5743208982252807, "grad_norm": 1.377677759273245, "learning_rate": 8.090892884756212e-06, "loss": 0.7737, "step": 15857 }, { "epoch": 0.5743571169865991, "grad_norm": 1.3106435798349942, "learning_rate": 8.089741405934923e-06, "loss": 0.7054, "step": 15858 }, { "epoch": 0.5743933357479174, "grad_norm": 1.3991043176615086, "learning_rate": 8.088589953400472e-06, "loss": 0.7143, "step": 15859 }, { "epoch": 0.5744295545092358, "grad_norm": 1.5166004842427652, "learning_rate": 8.0874385271687e-06, "loss": 0.7974, "step": 15860 }, { "epoch": 0.5744657732705541, "grad_norm": 1.3004241301350292, "learning_rate": 8.086287127255457e-06, "loss": 0.6937, "step": 15861 }, { "epoch": 0.5745019920318725, "grad_norm": 1.485790154074075, "learning_rate": 8.085135753676583e-06, "loss": 0.7068, "step": 15862 }, { "epoch": 0.5745382107931909, "grad_norm": 1.1011346209424704, "learning_rate": 8.083984406447925e-06, "loss": 0.6685, "step": 15863 }, { "epoch": 0.5745744295545092, "grad_norm": 1.382959923725195, "learning_rate": 8.082833085585324e-06, "loss": 0.7554, "step": 15864 }, { "epoch": 0.5746106483158276, "grad_norm": 1.3785368273221428, "learning_rate": 8.081681791104624e-06, "loss": 0.7298, "step": 15865 }, { "epoch": 0.574646867077146, "grad_norm": 1.3244431025701278, "learning_rate": 8.08053052302167e-06, "loss": 0.6965, "step": 15866 }, { "epoch": 0.5746830858384643, "grad_norm": 1.4724957697419798, "learning_rate": 8.0793792813523e-06, "loss": 0.6778, "step": 15867 }, { "epoch": 0.5747193045997827, "grad_norm": 1.487469518748896, "learning_rate": 8.078228066112363e-06, "loss": 0.764, "step": 15868 }, { "epoch": 0.574755523361101, "grad_norm": 1.3707469069478002, "learning_rate": 8.077076877317692e-06, "loss": 0.717, "step": 15869 }, { "epoch": 0.5747917421224195, "grad_norm": 1.3082649978068788, "learning_rate": 8.075925714984136e-06, "loss": 0.7277, "step": 15870 }, { "epoch": 0.5748279608837378, "grad_norm": 1.3116937279657668, "learning_rate": 8.07477457912753e-06, "loss": 0.7574, "step": 15871 }, { "epoch": 0.5748641796450561, "grad_norm": 1.3643022999647165, "learning_rate": 8.07362346976372e-06, "loss": 0.7559, "step": 15872 }, { "epoch": 0.5749003984063745, "grad_norm": 1.4447152166577306, "learning_rate": 8.072472386908543e-06, "loss": 0.7239, "step": 15873 }, { "epoch": 0.5749366171676928, "grad_norm": 1.465671709674672, "learning_rate": 8.071321330577838e-06, "loss": 0.6863, "step": 15874 }, { "epoch": 0.5749728359290113, "grad_norm": 1.5519130189842145, "learning_rate": 8.070170300787448e-06, "loss": 0.6677, "step": 15875 }, { "epoch": 0.5750090546903296, "grad_norm": 1.3373572560259137, "learning_rate": 8.069019297553208e-06, "loss": 0.7152, "step": 15876 }, { "epoch": 0.5750452734516479, "grad_norm": 1.7614083534728306, "learning_rate": 8.067868320890961e-06, "loss": 0.7316, "step": 15877 }, { "epoch": 0.5750814922129663, "grad_norm": 1.0895807568891256, "learning_rate": 8.066717370816543e-06, "loss": 0.6782, "step": 15878 }, { "epoch": 0.5751177109742847, "grad_norm": 1.384008352527934, "learning_rate": 8.065566447345794e-06, "loss": 0.6785, "step": 15879 }, { "epoch": 0.5751539297356031, "grad_norm": 1.3554034718946826, "learning_rate": 8.064415550494549e-06, "loss": 0.6814, "step": 15880 }, { "epoch": 0.5751901484969214, "grad_norm": 1.2859911634261083, "learning_rate": 8.063264680278648e-06, "loss": 0.6545, "step": 15881 }, { "epoch": 0.5752263672582397, "grad_norm": 0.9986734506678482, "learning_rate": 8.062113836713925e-06, "loss": 0.711, "step": 15882 }, { "epoch": 0.5752625860195582, "grad_norm": 1.0749780162913674, "learning_rate": 8.060963019816219e-06, "loss": 0.7021, "step": 15883 }, { "epoch": 0.5752988047808765, "grad_norm": 1.426902753189944, "learning_rate": 8.059812229601367e-06, "loss": 0.7139, "step": 15884 }, { "epoch": 0.5753350235421949, "grad_norm": 1.3018494700381682, "learning_rate": 8.058661466085203e-06, "loss": 0.6917, "step": 15885 }, { "epoch": 0.5753712423035132, "grad_norm": 1.497209926895878, "learning_rate": 8.057510729283563e-06, "loss": 0.7311, "step": 15886 }, { "epoch": 0.5754074610648315, "grad_norm": 1.4160029391016469, "learning_rate": 8.056360019212281e-06, "loss": 0.722, "step": 15887 }, { "epoch": 0.57544367982615, "grad_norm": 1.1878729630814353, "learning_rate": 8.055209335887195e-06, "loss": 0.7899, "step": 15888 }, { "epoch": 0.5754798985874683, "grad_norm": 1.2039292127856442, "learning_rate": 8.054058679324135e-06, "loss": 0.6255, "step": 15889 }, { "epoch": 0.5755161173487867, "grad_norm": 1.2645395939062767, "learning_rate": 8.05290804953894e-06, "loss": 0.6899, "step": 15890 }, { "epoch": 0.575552336110105, "grad_norm": 1.628442710905828, "learning_rate": 8.051757446547439e-06, "loss": 0.742, "step": 15891 }, { "epoch": 0.5755885548714234, "grad_norm": 1.3566567363985058, "learning_rate": 8.05060687036547e-06, "loss": 0.701, "step": 15892 }, { "epoch": 0.5756247736327418, "grad_norm": 1.3326483671557006, "learning_rate": 8.049456321008859e-06, "loss": 0.7273, "step": 15893 }, { "epoch": 0.5756609923940601, "grad_norm": 1.2374744765462427, "learning_rate": 8.048305798493446e-06, "loss": 0.6487, "step": 15894 }, { "epoch": 0.5756972111553785, "grad_norm": 1.472917415344092, "learning_rate": 8.04715530283506e-06, "loss": 0.7072, "step": 15895 }, { "epoch": 0.5757334299166968, "grad_norm": 0.9002822541309866, "learning_rate": 8.046004834049531e-06, "loss": 0.7175, "step": 15896 }, { "epoch": 0.5757696486780152, "grad_norm": 1.2724205978276126, "learning_rate": 8.044854392152696e-06, "loss": 0.6871, "step": 15897 }, { "epoch": 0.5758058674393336, "grad_norm": 1.3667790334059817, "learning_rate": 8.043703977160378e-06, "loss": 0.7178, "step": 15898 }, { "epoch": 0.5758420862006519, "grad_norm": 1.6668481635369898, "learning_rate": 8.042553589088416e-06, "loss": 0.7055, "step": 15899 }, { "epoch": 0.5758783049619703, "grad_norm": 1.0027630205738256, "learning_rate": 8.041403227952635e-06, "loss": 0.7392, "step": 15900 }, { "epoch": 0.5759145237232887, "grad_norm": 1.2915725898867736, "learning_rate": 8.040252893768868e-06, "loss": 0.6964, "step": 15901 }, { "epoch": 0.575950742484607, "grad_norm": 1.0513707239373924, "learning_rate": 8.039102586552941e-06, "loss": 0.7095, "step": 15902 }, { "epoch": 0.5759869612459254, "grad_norm": 1.0214640285866192, "learning_rate": 8.037952306320684e-06, "loss": 0.7042, "step": 15903 }, { "epoch": 0.5760231800072437, "grad_norm": 1.1599606281218442, "learning_rate": 8.036802053087932e-06, "loss": 0.7517, "step": 15904 }, { "epoch": 0.5760593987685622, "grad_norm": 1.1448631796507431, "learning_rate": 8.035651826870505e-06, "loss": 0.6932, "step": 15905 }, { "epoch": 0.5760956175298805, "grad_norm": 1.4650000043818567, "learning_rate": 8.034501627684237e-06, "loss": 0.7312, "step": 15906 }, { "epoch": 0.5761318362911988, "grad_norm": 1.3986018286752864, "learning_rate": 8.033351455544957e-06, "loss": 0.7404, "step": 15907 }, { "epoch": 0.5761680550525172, "grad_norm": 1.3317318800847686, "learning_rate": 8.032201310468484e-06, "loss": 0.6305, "step": 15908 }, { "epoch": 0.5762042738138355, "grad_norm": 1.6173077183880635, "learning_rate": 8.03105119247065e-06, "loss": 0.7538, "step": 15909 }, { "epoch": 0.576240492575154, "grad_norm": 1.3922621755127524, "learning_rate": 8.029901101567284e-06, "loss": 0.7353, "step": 15910 }, { "epoch": 0.5762767113364723, "grad_norm": 1.4045164774848724, "learning_rate": 8.02875103777421e-06, "loss": 0.6571, "step": 15911 }, { "epoch": 0.5763129300977906, "grad_norm": 1.2954751712661554, "learning_rate": 8.02760100110725e-06, "loss": 0.6874, "step": 15912 }, { "epoch": 0.576349148859109, "grad_norm": 1.287658662475522, "learning_rate": 8.026450991582237e-06, "loss": 0.7228, "step": 15913 }, { "epoch": 0.5763853676204274, "grad_norm": 1.4516695973704865, "learning_rate": 8.02530100921499e-06, "loss": 0.7482, "step": 15914 }, { "epoch": 0.5764215863817458, "grad_norm": 1.350117535298188, "learning_rate": 8.024151054021337e-06, "loss": 0.708, "step": 15915 }, { "epoch": 0.5764578051430641, "grad_norm": 1.4595223689873433, "learning_rate": 8.023001126017102e-06, "loss": 0.7187, "step": 15916 }, { "epoch": 0.5764940239043824, "grad_norm": 1.3858390012153459, "learning_rate": 8.02185122521811e-06, "loss": 0.7391, "step": 15917 }, { "epoch": 0.5765302426657009, "grad_norm": 1.376704500627057, "learning_rate": 8.020701351640182e-06, "loss": 0.7385, "step": 15918 }, { "epoch": 0.5765664614270192, "grad_norm": 1.322743831552502, "learning_rate": 8.01955150529914e-06, "loss": 0.7656, "step": 15919 }, { "epoch": 0.5766026801883376, "grad_norm": 1.2870689654917045, "learning_rate": 8.018401686210815e-06, "loss": 0.7546, "step": 15920 }, { "epoch": 0.5766388989496559, "grad_norm": 1.398719772429005, "learning_rate": 8.01725189439102e-06, "loss": 0.7901, "step": 15921 }, { "epoch": 0.5766751177109742, "grad_norm": 1.281669174511169, "learning_rate": 8.016102129855584e-06, "loss": 0.7869, "step": 15922 }, { "epoch": 0.5767113364722927, "grad_norm": 1.0314635753053152, "learning_rate": 8.014952392620322e-06, "loss": 0.6692, "step": 15923 }, { "epoch": 0.576747555233611, "grad_norm": 1.3908610337837868, "learning_rate": 8.013802682701063e-06, "loss": 0.7305, "step": 15924 }, { "epoch": 0.5767837739949294, "grad_norm": 1.4002764763307476, "learning_rate": 8.012653000113621e-06, "loss": 0.7695, "step": 15925 }, { "epoch": 0.5768199927562477, "grad_norm": 1.3317966149652625, "learning_rate": 8.011503344873821e-06, "loss": 0.6394, "step": 15926 }, { "epoch": 0.576856211517566, "grad_norm": 1.3737803307138379, "learning_rate": 8.010353716997483e-06, "loss": 0.755, "step": 15927 }, { "epoch": 0.5768924302788845, "grad_norm": 1.4582286010895689, "learning_rate": 8.009204116500426e-06, "loss": 0.7404, "step": 15928 }, { "epoch": 0.5769286490402028, "grad_norm": 1.47071701337075, "learning_rate": 8.008054543398471e-06, "loss": 0.7244, "step": 15929 }, { "epoch": 0.5769648678015212, "grad_norm": 1.2910168131247173, "learning_rate": 8.006904997707432e-06, "loss": 0.7542, "step": 15930 }, { "epoch": 0.5770010865628395, "grad_norm": 1.3940357548782571, "learning_rate": 8.005755479443136e-06, "loss": 0.7982, "step": 15931 }, { "epoch": 0.5770373053241579, "grad_norm": 1.4114301374887233, "learning_rate": 8.004605988621394e-06, "loss": 0.72, "step": 15932 }, { "epoch": 0.5770735240854763, "grad_norm": 1.3365347681564488, "learning_rate": 8.003456525258028e-06, "loss": 0.6996, "step": 15933 }, { "epoch": 0.5771097428467946, "grad_norm": 0.9056289550488267, "learning_rate": 8.002307089368853e-06, "loss": 0.6834, "step": 15934 }, { "epoch": 0.577145961608113, "grad_norm": 1.3547213586725069, "learning_rate": 8.00115768096969e-06, "loss": 0.6852, "step": 15935 }, { "epoch": 0.5771821803694314, "grad_norm": 1.5197369502464728, "learning_rate": 8.000008300076351e-06, "loss": 0.7532, "step": 15936 }, { "epoch": 0.5772183991307497, "grad_norm": 1.4516614598930144, "learning_rate": 7.998858946704657e-06, "loss": 0.7961, "step": 15937 }, { "epoch": 0.5772546178920681, "grad_norm": 1.4247939362358888, "learning_rate": 7.997709620870421e-06, "loss": 0.6843, "step": 15938 }, { "epoch": 0.5772908366533864, "grad_norm": 0.9848230879569772, "learning_rate": 7.996560322589462e-06, "loss": 0.7316, "step": 15939 }, { "epoch": 0.5773270554147049, "grad_norm": 1.0378435529881553, "learning_rate": 7.995411051877593e-06, "loss": 0.6925, "step": 15940 }, { "epoch": 0.5773632741760232, "grad_norm": 1.2939653646067, "learning_rate": 7.994261808750628e-06, "loss": 0.6868, "step": 15941 }, { "epoch": 0.5773994929373415, "grad_norm": 1.0604638938005209, "learning_rate": 7.993112593224384e-06, "loss": 0.704, "step": 15942 }, { "epoch": 0.5774357116986599, "grad_norm": 1.414025335850958, "learning_rate": 7.991963405314672e-06, "loss": 0.7624, "step": 15943 }, { "epoch": 0.5774719304599782, "grad_norm": 1.003673789671585, "learning_rate": 7.990814245037312e-06, "loss": 0.7799, "step": 15944 }, { "epoch": 0.5775081492212967, "grad_norm": 1.3015847989011968, "learning_rate": 7.98966511240811e-06, "loss": 0.7576, "step": 15945 }, { "epoch": 0.577544367982615, "grad_norm": 3.433351883598941, "learning_rate": 7.988516007442882e-06, "loss": 0.7369, "step": 15946 }, { "epoch": 0.5775805867439333, "grad_norm": 1.1893002465850653, "learning_rate": 7.987366930157443e-06, "loss": 0.7259, "step": 15947 }, { "epoch": 0.5776168055052517, "grad_norm": 1.4720825727250075, "learning_rate": 7.986217880567603e-06, "loss": 0.7124, "step": 15948 }, { "epoch": 0.5776530242665701, "grad_norm": 1.2592606001647157, "learning_rate": 7.985068858689176e-06, "loss": 0.6874, "step": 15949 }, { "epoch": 0.5776892430278885, "grad_norm": 1.343661811341121, "learning_rate": 7.98391986453797e-06, "loss": 0.738, "step": 15950 }, { "epoch": 0.5777254617892068, "grad_norm": 1.2783731732593178, "learning_rate": 7.982770898129799e-06, "loss": 0.6731, "step": 15951 }, { "epoch": 0.5777616805505251, "grad_norm": 1.2629651594836557, "learning_rate": 7.981621959480471e-06, "loss": 0.7324, "step": 15952 }, { "epoch": 0.5777978993118436, "grad_norm": 1.0455248720263366, "learning_rate": 7.980473048605802e-06, "loss": 0.6817, "step": 15953 }, { "epoch": 0.5778341180731619, "grad_norm": 1.7483355353080243, "learning_rate": 7.979324165521596e-06, "loss": 0.7219, "step": 15954 }, { "epoch": 0.5778703368344803, "grad_norm": 1.4658193689911287, "learning_rate": 7.978175310243665e-06, "loss": 0.7886, "step": 15955 }, { "epoch": 0.5779065555957986, "grad_norm": 0.9004919755074074, "learning_rate": 7.977026482787817e-06, "loss": 0.6519, "step": 15956 }, { "epoch": 0.5779427743571169, "grad_norm": 1.3996381919570806, "learning_rate": 7.975877683169863e-06, "loss": 0.7115, "step": 15957 }, { "epoch": 0.5779789931184354, "grad_norm": 1.2987331126173471, "learning_rate": 7.974728911405613e-06, "loss": 0.6857, "step": 15958 }, { "epoch": 0.5780152118797537, "grad_norm": 1.3348141661400623, "learning_rate": 7.97358016751087e-06, "loss": 0.7065, "step": 15959 }, { "epoch": 0.5780514306410721, "grad_norm": 1.4989942330364154, "learning_rate": 7.972431451501446e-06, "loss": 0.7715, "step": 15960 }, { "epoch": 0.5780876494023904, "grad_norm": 1.0858678269837416, "learning_rate": 7.971282763393147e-06, "loss": 0.7402, "step": 15961 }, { "epoch": 0.5781238681637088, "grad_norm": 1.4552517780908452, "learning_rate": 7.97013410320178e-06, "loss": 0.741, "step": 15962 }, { "epoch": 0.5781600869250272, "grad_norm": 1.490261367613388, "learning_rate": 7.96898547094315e-06, "loss": 0.7649, "step": 15963 }, { "epoch": 0.5781963056863455, "grad_norm": 1.3363828551116688, "learning_rate": 7.967836866633065e-06, "loss": 0.7505, "step": 15964 }, { "epoch": 0.5782325244476639, "grad_norm": 1.3340537854647438, "learning_rate": 7.96668829028733e-06, "loss": 0.6985, "step": 15965 }, { "epoch": 0.5782687432089822, "grad_norm": 1.0636028789423249, "learning_rate": 7.965539741921753e-06, "loss": 0.7216, "step": 15966 }, { "epoch": 0.5783049619703006, "grad_norm": 1.1402924559948, "learning_rate": 7.964391221552133e-06, "loss": 0.6831, "step": 15967 }, { "epoch": 0.578341180731619, "grad_norm": 1.3748109535825612, "learning_rate": 7.96324272919428e-06, "loss": 0.7393, "step": 15968 }, { "epoch": 0.5783773994929373, "grad_norm": 2.3962956964775786, "learning_rate": 7.962094264863997e-06, "loss": 0.6209, "step": 15969 }, { "epoch": 0.5784136182542557, "grad_norm": 1.4318450167733243, "learning_rate": 7.960945828577087e-06, "loss": 0.7647, "step": 15970 }, { "epoch": 0.5784498370155741, "grad_norm": 1.4748345386311636, "learning_rate": 7.959797420349356e-06, "loss": 0.7523, "step": 15971 }, { "epoch": 0.5784860557768924, "grad_norm": 1.2773061361811835, "learning_rate": 7.958649040196602e-06, "loss": 0.6724, "step": 15972 }, { "epoch": 0.5785222745382108, "grad_norm": 1.2694524714896498, "learning_rate": 7.957500688134635e-06, "loss": 0.6334, "step": 15973 }, { "epoch": 0.5785584932995291, "grad_norm": 1.4024093182252235, "learning_rate": 7.95635236417925e-06, "loss": 0.69, "step": 15974 }, { "epoch": 0.5785947120608476, "grad_norm": 1.3600108561406834, "learning_rate": 7.955204068346254e-06, "loss": 0.6705, "step": 15975 }, { "epoch": 0.5786309308221659, "grad_norm": 1.445471518377503, "learning_rate": 7.954055800651445e-06, "loss": 0.7308, "step": 15976 }, { "epoch": 0.5786671495834842, "grad_norm": 1.3266959685713837, "learning_rate": 7.952907561110625e-06, "loss": 0.7358, "step": 15977 }, { "epoch": 0.5787033683448026, "grad_norm": 1.3463167478801992, "learning_rate": 7.9517593497396e-06, "loss": 0.7068, "step": 15978 }, { "epoch": 0.578739587106121, "grad_norm": 1.3476275240559403, "learning_rate": 7.95061116655416e-06, "loss": 0.7226, "step": 15979 }, { "epoch": 0.5787758058674394, "grad_norm": 1.3050025770784928, "learning_rate": 7.949463011570117e-06, "loss": 0.7266, "step": 15980 }, { "epoch": 0.5788120246287577, "grad_norm": 0.9596361878738988, "learning_rate": 7.94831488480326e-06, "loss": 0.6479, "step": 15981 }, { "epoch": 0.578848243390076, "grad_norm": 1.314636675360245, "learning_rate": 7.947166786269398e-06, "loss": 0.7247, "step": 15982 }, { "epoch": 0.5788844621513944, "grad_norm": 1.336958326466583, "learning_rate": 7.94601871598432e-06, "loss": 0.7011, "step": 15983 }, { "epoch": 0.5789206809127128, "grad_norm": 1.237938593587779, "learning_rate": 7.944870673963833e-06, "loss": 0.7584, "step": 15984 }, { "epoch": 0.5789568996740312, "grad_norm": 1.352414912506193, "learning_rate": 7.943722660223729e-06, "loss": 0.7675, "step": 15985 }, { "epoch": 0.5789931184353495, "grad_norm": 1.0492210133474702, "learning_rate": 7.94257467477981e-06, "loss": 0.6317, "step": 15986 }, { "epoch": 0.5790293371966678, "grad_norm": 1.3441580167502736, "learning_rate": 7.94142671764787e-06, "loss": 0.7209, "step": 15987 }, { "epoch": 0.5790655559579863, "grad_norm": 1.005059811121876, "learning_rate": 7.940278788843706e-06, "loss": 0.6522, "step": 15988 }, { "epoch": 0.5791017747193046, "grad_norm": 1.1104576042170005, "learning_rate": 7.939130888383117e-06, "loss": 0.7033, "step": 15989 }, { "epoch": 0.579137993480623, "grad_norm": 1.442696324848184, "learning_rate": 7.937983016281895e-06, "loss": 0.7386, "step": 15990 }, { "epoch": 0.5791742122419413, "grad_norm": 1.445859534896963, "learning_rate": 7.936835172555843e-06, "loss": 0.6872, "step": 15991 }, { "epoch": 0.5792104310032596, "grad_norm": 1.2431596319894287, "learning_rate": 7.93568735722075e-06, "loss": 0.6866, "step": 15992 }, { "epoch": 0.5792466497645781, "grad_norm": 1.5049877597398809, "learning_rate": 7.934539570292415e-06, "loss": 0.7657, "step": 15993 }, { "epoch": 0.5792828685258964, "grad_norm": 1.5837537971622677, "learning_rate": 7.933391811786625e-06, "loss": 0.8309, "step": 15994 }, { "epoch": 0.5793190872872148, "grad_norm": 34.963449903178606, "learning_rate": 7.932244081719186e-06, "loss": 0.7403, "step": 15995 }, { "epoch": 0.5793553060485331, "grad_norm": 1.1277343466800511, "learning_rate": 7.931096380105881e-06, "loss": 0.7098, "step": 15996 }, { "epoch": 0.5793915248098515, "grad_norm": 1.3103089962349974, "learning_rate": 7.929948706962508e-06, "loss": 0.6936, "step": 15997 }, { "epoch": 0.5794277435711699, "grad_norm": 1.4023719568258224, "learning_rate": 7.928801062304862e-06, "loss": 0.7217, "step": 15998 }, { "epoch": 0.5794639623324882, "grad_norm": 1.3386604996791482, "learning_rate": 7.927653446148732e-06, "loss": 0.681, "step": 15999 }, { "epoch": 0.5795001810938066, "grad_norm": 1.443333978519858, "learning_rate": 7.926505858509911e-06, "loss": 0.6472, "step": 16000 }, { "epoch": 0.579536399855125, "grad_norm": 1.3335386581639745, "learning_rate": 7.92535829940419e-06, "loss": 0.6826, "step": 16001 }, { "epoch": 0.5795726186164433, "grad_norm": 1.4834504931409036, "learning_rate": 7.924210768847365e-06, "loss": 0.8344, "step": 16002 }, { "epoch": 0.5796088373777617, "grad_norm": 1.0548032838903156, "learning_rate": 7.92306326685522e-06, "loss": 0.6982, "step": 16003 }, { "epoch": 0.57964505613908, "grad_norm": 1.4112177298579802, "learning_rate": 7.921915793443553e-06, "loss": 0.7285, "step": 16004 }, { "epoch": 0.5796812749003984, "grad_norm": 1.3628514043948994, "learning_rate": 7.920768348628149e-06, "loss": 0.6904, "step": 16005 }, { "epoch": 0.5797174936617168, "grad_norm": 1.1470553929371756, "learning_rate": 7.919620932424801e-06, "loss": 0.7188, "step": 16006 }, { "epoch": 0.5797537124230351, "grad_norm": 2.3307366355939707, "learning_rate": 7.918473544849295e-06, "loss": 0.7511, "step": 16007 }, { "epoch": 0.5797899311843535, "grad_norm": 1.3690798925569765, "learning_rate": 7.917326185917422e-06, "loss": 0.7285, "step": 16008 }, { "epoch": 0.5798261499456718, "grad_norm": 1.3388763151963694, "learning_rate": 7.916178855644972e-06, "loss": 0.7123, "step": 16009 }, { "epoch": 0.5798623687069903, "grad_norm": 1.2581814393665047, "learning_rate": 7.915031554047732e-06, "loss": 0.6382, "step": 16010 }, { "epoch": 0.5798985874683086, "grad_norm": 1.5597647376742227, "learning_rate": 7.913884281141489e-06, "loss": 0.7231, "step": 16011 }, { "epoch": 0.5799348062296269, "grad_norm": 1.5569424737407762, "learning_rate": 7.912737036942032e-06, "loss": 0.6959, "step": 16012 }, { "epoch": 0.5799710249909453, "grad_norm": 1.4025907166189437, "learning_rate": 7.911589821465146e-06, "loss": 0.6871, "step": 16013 }, { "epoch": 0.5800072437522636, "grad_norm": 1.569678007587375, "learning_rate": 7.910442634726619e-06, "loss": 0.7629, "step": 16014 }, { "epoch": 0.5800434625135821, "grad_norm": 1.2900772789341513, "learning_rate": 7.90929547674224e-06, "loss": 0.7085, "step": 16015 }, { "epoch": 0.5800796812749004, "grad_norm": 1.3537024827263004, "learning_rate": 7.90814834752779e-06, "loss": 0.7247, "step": 16016 }, { "epoch": 0.5801159000362187, "grad_norm": 1.374884750944034, "learning_rate": 7.907001247099059e-06, "loss": 0.7276, "step": 16017 }, { "epoch": 0.5801521187975371, "grad_norm": 1.338185222240293, "learning_rate": 7.905854175471827e-06, "loss": 0.6676, "step": 16018 }, { "epoch": 0.5801883375588555, "grad_norm": 1.157094052932996, "learning_rate": 7.904707132661882e-06, "loss": 0.6675, "step": 16019 }, { "epoch": 0.5802245563201739, "grad_norm": 1.1000851416373372, "learning_rate": 7.90356011868501e-06, "loss": 0.6322, "step": 16020 }, { "epoch": 0.5802607750814922, "grad_norm": 1.3139017820730283, "learning_rate": 7.90241313355699e-06, "loss": 0.7325, "step": 16021 }, { "epoch": 0.5802969938428105, "grad_norm": 1.337647007947145, "learning_rate": 7.901266177293612e-06, "loss": 0.6968, "step": 16022 }, { "epoch": 0.580333212604129, "grad_norm": 1.275540790655499, "learning_rate": 7.900119249910653e-06, "loss": 0.6675, "step": 16023 }, { "epoch": 0.5803694313654473, "grad_norm": 1.3497132600861792, "learning_rate": 7.898972351423899e-06, "loss": 0.7563, "step": 16024 }, { "epoch": 0.5804056501267657, "grad_norm": 1.3606473692854322, "learning_rate": 7.897825481849131e-06, "loss": 0.7578, "step": 16025 }, { "epoch": 0.580441868888084, "grad_norm": 1.0019821356594534, "learning_rate": 7.896678641202133e-06, "loss": 0.7285, "step": 16026 }, { "epoch": 0.5804780876494023, "grad_norm": 1.2873857307129002, "learning_rate": 7.895531829498683e-06, "loss": 0.7095, "step": 16027 }, { "epoch": 0.5805143064107208, "grad_norm": 1.1795066338008338, "learning_rate": 7.894385046754566e-06, "loss": 0.7123, "step": 16028 }, { "epoch": 0.5805505251720391, "grad_norm": 1.262217502901036, "learning_rate": 7.89323829298556e-06, "loss": 0.6898, "step": 16029 }, { "epoch": 0.5805867439333575, "grad_norm": 1.2674659567696618, "learning_rate": 7.892091568207445e-06, "loss": 0.7513, "step": 16030 }, { "epoch": 0.5806229626946758, "grad_norm": 1.2424779168609243, "learning_rate": 7.890944872436005e-06, "loss": 0.6754, "step": 16031 }, { "epoch": 0.5806591814559942, "grad_norm": 1.2851350805747868, "learning_rate": 7.889798205687014e-06, "loss": 0.7606, "step": 16032 }, { "epoch": 0.5806954002173126, "grad_norm": 1.405606300664056, "learning_rate": 7.888651567976256e-06, "loss": 0.7989, "step": 16033 }, { "epoch": 0.5807316189786309, "grad_norm": 2.262293699709287, "learning_rate": 7.887504959319507e-06, "loss": 0.7361, "step": 16034 }, { "epoch": 0.5807678377399493, "grad_norm": 1.2155449870048036, "learning_rate": 7.886358379732546e-06, "loss": 0.7483, "step": 16035 }, { "epoch": 0.5808040565012677, "grad_norm": 1.3609157071629738, "learning_rate": 7.88521182923115e-06, "loss": 0.744, "step": 16036 }, { "epoch": 0.580840275262586, "grad_norm": 1.2834810191202872, "learning_rate": 7.8840653078311e-06, "loss": 0.7337, "step": 16037 }, { "epoch": 0.5808764940239044, "grad_norm": 1.3901676339348692, "learning_rate": 7.882918815548168e-06, "loss": 0.6815, "step": 16038 }, { "epoch": 0.5809127127852227, "grad_norm": 1.141370995180744, "learning_rate": 7.881772352398132e-06, "loss": 0.6821, "step": 16039 }, { "epoch": 0.5809489315465411, "grad_norm": 1.0882430222811832, "learning_rate": 7.880625918396773e-06, "loss": 0.7478, "step": 16040 }, { "epoch": 0.5809851503078595, "grad_norm": 1.4303313691959654, "learning_rate": 7.879479513559861e-06, "loss": 0.7161, "step": 16041 }, { "epoch": 0.5810213690691778, "grad_norm": 1.5067640882400608, "learning_rate": 7.878333137903177e-06, "loss": 0.7525, "step": 16042 }, { "epoch": 0.5810575878304962, "grad_norm": 1.4134639479561963, "learning_rate": 7.87718679144249e-06, "loss": 0.7415, "step": 16043 }, { "epoch": 0.5810938065918145, "grad_norm": 1.320431361124435, "learning_rate": 7.87604047419358e-06, "loss": 0.6195, "step": 16044 }, { "epoch": 0.581130025353133, "grad_norm": 1.3760504354189835, "learning_rate": 7.87489418617222e-06, "loss": 0.7184, "step": 16045 }, { "epoch": 0.5811662441144513, "grad_norm": 1.2998354669562195, "learning_rate": 7.873747927394182e-06, "loss": 0.66, "step": 16046 }, { "epoch": 0.5812024628757696, "grad_norm": 2.7172764259602564, "learning_rate": 7.872601697875242e-06, "loss": 0.729, "step": 16047 }, { "epoch": 0.581238681637088, "grad_norm": 1.3278198347396342, "learning_rate": 7.871455497631171e-06, "loss": 0.6094, "step": 16048 }, { "epoch": 0.5812749003984063, "grad_norm": 1.3302044101314698, "learning_rate": 7.870309326677742e-06, "loss": 0.6738, "step": 16049 }, { "epoch": 0.5813111191597248, "grad_norm": 1.2771403809903088, "learning_rate": 7.869163185030728e-06, "loss": 0.6389, "step": 16050 }, { "epoch": 0.5813473379210431, "grad_norm": 1.3664336871636293, "learning_rate": 7.868017072705903e-06, "loss": 0.7146, "step": 16051 }, { "epoch": 0.5813835566823614, "grad_norm": 1.2810993033269098, "learning_rate": 7.866870989719034e-06, "loss": 0.7368, "step": 16052 }, { "epoch": 0.5814197754436798, "grad_norm": 1.1310614780923889, "learning_rate": 7.865724936085897e-06, "loss": 0.6672, "step": 16053 }, { "epoch": 0.5814559942049982, "grad_norm": 1.36892871196237, "learning_rate": 7.864578911822258e-06, "loss": 0.676, "step": 16054 }, { "epoch": 0.5814922129663166, "grad_norm": 1.2047281827532612, "learning_rate": 7.863432916943892e-06, "loss": 0.6768, "step": 16055 }, { "epoch": 0.5815284317276349, "grad_norm": 1.4621773854040565, "learning_rate": 7.862286951466563e-06, "loss": 0.7837, "step": 16056 }, { "epoch": 0.5815646504889532, "grad_norm": 1.3593834980885382, "learning_rate": 7.861141015406046e-06, "loss": 0.6904, "step": 16057 }, { "epoch": 0.5816008692502717, "grad_norm": 1.3872168506031748, "learning_rate": 7.859995108778106e-06, "loss": 0.6662, "step": 16058 }, { "epoch": 0.58163708801159, "grad_norm": 1.3058147086727017, "learning_rate": 7.858849231598517e-06, "loss": 0.6348, "step": 16059 }, { "epoch": 0.5816733067729084, "grad_norm": 1.2843799382353263, "learning_rate": 7.857703383883042e-06, "loss": 0.7183, "step": 16060 }, { "epoch": 0.5817095255342267, "grad_norm": 1.4440940618295715, "learning_rate": 7.85655756564745e-06, "loss": 0.7581, "step": 16061 }, { "epoch": 0.581745744295545, "grad_norm": 1.067343393960831, "learning_rate": 7.85541177690751e-06, "loss": 0.7294, "step": 16062 }, { "epoch": 0.5817819630568635, "grad_norm": 1.3581817304633834, "learning_rate": 7.85426601767899e-06, "loss": 0.7835, "step": 16063 }, { "epoch": 0.5818181818181818, "grad_norm": 1.3784448585225202, "learning_rate": 7.853120287977654e-06, "loss": 0.784, "step": 16064 }, { "epoch": 0.5818544005795002, "grad_norm": 1.106245184879226, "learning_rate": 7.851974587819267e-06, "loss": 0.696, "step": 16065 }, { "epoch": 0.5818906193408185, "grad_norm": 1.5704602461434634, "learning_rate": 7.850828917219601e-06, "loss": 0.6954, "step": 16066 }, { "epoch": 0.5819268381021369, "grad_norm": 1.3939074174813877, "learning_rate": 7.849683276194416e-06, "loss": 0.7379, "step": 16067 }, { "epoch": 0.5819630568634553, "grad_norm": 1.438429041153283, "learning_rate": 7.848537664759477e-06, "loss": 0.7228, "step": 16068 }, { "epoch": 0.5819992756247736, "grad_norm": 1.0150578398469692, "learning_rate": 7.847392082930552e-06, "loss": 0.7213, "step": 16069 }, { "epoch": 0.582035494386092, "grad_norm": 1.2482744958546577, "learning_rate": 7.846246530723401e-06, "loss": 0.6803, "step": 16070 }, { "epoch": 0.5820717131474104, "grad_norm": 1.3767471172898478, "learning_rate": 7.845101008153793e-06, "loss": 0.689, "step": 16071 }, { "epoch": 0.5821079319087287, "grad_norm": 1.3110620804904352, "learning_rate": 7.843955515237487e-06, "loss": 0.7237, "step": 16072 }, { "epoch": 0.5821441506700471, "grad_norm": 1.1658342127375538, "learning_rate": 7.842810051990248e-06, "loss": 0.7169, "step": 16073 }, { "epoch": 0.5821803694313654, "grad_norm": 1.331514925638692, "learning_rate": 7.841664618427838e-06, "loss": 0.7699, "step": 16074 }, { "epoch": 0.5822165881926838, "grad_norm": 1.473136294920052, "learning_rate": 7.84051921456602e-06, "loss": 0.732, "step": 16075 }, { "epoch": 0.5822528069540022, "grad_norm": 1.3910873674447448, "learning_rate": 7.839373840420555e-06, "loss": 0.6945, "step": 16076 }, { "epoch": 0.5822890257153205, "grad_norm": 1.2867972801647436, "learning_rate": 7.838228496007206e-06, "loss": 0.6731, "step": 16077 }, { "epoch": 0.5823252444766389, "grad_norm": 1.4469978107892734, "learning_rate": 7.837083181341728e-06, "loss": 0.7938, "step": 16078 }, { "epoch": 0.5823614632379572, "grad_norm": 1.3952300554098962, "learning_rate": 7.83593789643989e-06, "loss": 0.7043, "step": 16079 }, { "epoch": 0.5823976819992757, "grad_norm": 1.2673132997065626, "learning_rate": 7.834792641317447e-06, "loss": 0.7617, "step": 16080 }, { "epoch": 0.582433900760594, "grad_norm": 0.972199120733735, "learning_rate": 7.833647415990157e-06, "loss": 0.7062, "step": 16081 }, { "epoch": 0.5824701195219123, "grad_norm": 1.4509865602689804, "learning_rate": 7.832502220473784e-06, "loss": 0.7147, "step": 16082 }, { "epoch": 0.5825063382832307, "grad_norm": 1.2577067043403314, "learning_rate": 7.831357054784085e-06, "loss": 0.6795, "step": 16083 }, { "epoch": 0.582542557044549, "grad_norm": 1.2712605252962523, "learning_rate": 7.83021191893682e-06, "loss": 0.7055, "step": 16084 }, { "epoch": 0.5825787758058675, "grad_norm": 1.1880746008286491, "learning_rate": 7.829066812947743e-06, "loss": 0.7106, "step": 16085 }, { "epoch": 0.5826149945671858, "grad_norm": 1.3109720204244726, "learning_rate": 7.827921736832617e-06, "loss": 0.7156, "step": 16086 }, { "epoch": 0.5826512133285041, "grad_norm": 1.3191457119968732, "learning_rate": 7.826776690607194e-06, "loss": 0.771, "step": 16087 }, { "epoch": 0.5826874320898225, "grad_norm": 1.350437571967739, "learning_rate": 7.825631674287235e-06, "loss": 0.6863, "step": 16088 }, { "epoch": 0.5827236508511409, "grad_norm": 1.206152249455844, "learning_rate": 7.824486687888493e-06, "loss": 0.6818, "step": 16089 }, { "epoch": 0.5827598696124593, "grad_norm": 0.9242482379375327, "learning_rate": 7.823341731426728e-06, "loss": 0.6995, "step": 16090 }, { "epoch": 0.5827960883737776, "grad_norm": 1.2890973758314532, "learning_rate": 7.822196804917692e-06, "loss": 0.7066, "step": 16091 }, { "epoch": 0.5828323071350959, "grad_norm": 0.9406037654150173, "learning_rate": 7.82105190837714e-06, "loss": 0.657, "step": 16092 }, { "epoch": 0.5828685258964144, "grad_norm": 1.2955164292415726, "learning_rate": 7.81990704182083e-06, "loss": 0.6794, "step": 16093 }, { "epoch": 0.5829047446577327, "grad_norm": 1.3198344011157428, "learning_rate": 7.818762205264514e-06, "loss": 0.7199, "step": 16094 }, { "epoch": 0.5829409634190511, "grad_norm": 1.1516116415010842, "learning_rate": 7.817617398723949e-06, "loss": 0.6948, "step": 16095 }, { "epoch": 0.5829771821803694, "grad_norm": 1.240711305934928, "learning_rate": 7.816472622214884e-06, "loss": 0.7447, "step": 16096 }, { "epoch": 0.5830134009416877, "grad_norm": 1.2554683599400227, "learning_rate": 7.815327875753077e-06, "loss": 0.7462, "step": 16097 }, { "epoch": 0.5830496197030062, "grad_norm": 1.0017394153743104, "learning_rate": 7.814183159354274e-06, "loss": 0.7182, "step": 16098 }, { "epoch": 0.5830858384643245, "grad_norm": 0.9974075957335417, "learning_rate": 7.813038473034233e-06, "loss": 0.7223, "step": 16099 }, { "epoch": 0.5831220572256429, "grad_norm": 1.0359231264110087, "learning_rate": 7.811893816808704e-06, "loss": 0.7622, "step": 16100 }, { "epoch": 0.5831582759869612, "grad_norm": 1.2364517706771456, "learning_rate": 7.810749190693439e-06, "loss": 0.7134, "step": 16101 }, { "epoch": 0.5831944947482796, "grad_norm": 1.3753192758512447, "learning_rate": 7.80960459470419e-06, "loss": 0.7137, "step": 16102 }, { "epoch": 0.583230713509598, "grad_norm": 1.2404558487134891, "learning_rate": 7.808460028856705e-06, "loss": 0.6345, "step": 16103 }, { "epoch": 0.5832669322709163, "grad_norm": 1.4315694866746889, "learning_rate": 7.807315493166736e-06, "loss": 0.7729, "step": 16104 }, { "epoch": 0.5833031510322347, "grad_norm": 1.3648099021120883, "learning_rate": 7.806170987650033e-06, "loss": 0.7063, "step": 16105 }, { "epoch": 0.583339369793553, "grad_norm": 1.3715113930253338, "learning_rate": 7.805026512322344e-06, "loss": 0.7851, "step": 16106 }, { "epoch": 0.5833755885548714, "grad_norm": 1.348220675209923, "learning_rate": 7.803882067199419e-06, "loss": 0.8018, "step": 16107 }, { "epoch": 0.5834118073161898, "grad_norm": 1.109324927033171, "learning_rate": 7.802737652297008e-06, "loss": 0.7932, "step": 16108 }, { "epoch": 0.5834480260775081, "grad_norm": 1.0856734573799607, "learning_rate": 7.801593267630855e-06, "loss": 0.717, "step": 16109 }, { "epoch": 0.5834842448388265, "grad_norm": 1.4148493719278887, "learning_rate": 7.800448913216712e-06, "loss": 0.7683, "step": 16110 }, { "epoch": 0.5835204636001449, "grad_norm": 1.2978112477857509, "learning_rate": 7.799304589070325e-06, "loss": 0.6705, "step": 16111 }, { "epoch": 0.5835566823614632, "grad_norm": 1.3347893753260822, "learning_rate": 7.798160295207438e-06, "loss": 0.7026, "step": 16112 }, { "epoch": 0.5835929011227816, "grad_norm": 1.3451147834963508, "learning_rate": 7.797016031643803e-06, "loss": 0.7165, "step": 16113 }, { "epoch": 0.5836291198840999, "grad_norm": 1.3233614195705352, "learning_rate": 7.795871798395162e-06, "loss": 0.6621, "step": 16114 }, { "epoch": 0.5836653386454184, "grad_norm": 1.091286561517389, "learning_rate": 7.794727595477263e-06, "loss": 0.7573, "step": 16115 }, { "epoch": 0.5837015574067367, "grad_norm": 1.2653609466899558, "learning_rate": 7.793583422905847e-06, "loss": 0.7797, "step": 16116 }, { "epoch": 0.583737776168055, "grad_norm": 1.156576084860111, "learning_rate": 7.792439280696665e-06, "loss": 0.687, "step": 16117 }, { "epoch": 0.5837739949293734, "grad_norm": 1.3552183234595234, "learning_rate": 7.79129516886546e-06, "loss": 0.7107, "step": 16118 }, { "epoch": 0.5838102136906917, "grad_norm": 1.3233429712012679, "learning_rate": 7.79015108742797e-06, "loss": 0.6526, "step": 16119 }, { "epoch": 0.5838464324520102, "grad_norm": 1.450786688989069, "learning_rate": 7.789007036399945e-06, "loss": 0.7011, "step": 16120 }, { "epoch": 0.5838826512133285, "grad_norm": 1.3146817248035998, "learning_rate": 7.787863015797123e-06, "loss": 0.7372, "step": 16121 }, { "epoch": 0.5839188699746468, "grad_norm": 1.4318147321844588, "learning_rate": 7.786719025635254e-06, "loss": 0.7675, "step": 16122 }, { "epoch": 0.5839550887359652, "grad_norm": 1.3602789478131787, "learning_rate": 7.785575065930072e-06, "loss": 0.7602, "step": 16123 }, { "epoch": 0.5839913074972836, "grad_norm": 1.5470562982761227, "learning_rate": 7.784431136697324e-06, "loss": 0.7055, "step": 16124 }, { "epoch": 0.584027526258602, "grad_norm": 1.2300737572002183, "learning_rate": 7.783287237952751e-06, "loss": 0.6455, "step": 16125 }, { "epoch": 0.5840637450199203, "grad_norm": 1.2522213837399623, "learning_rate": 7.782143369712093e-06, "loss": 0.7053, "step": 16126 }, { "epoch": 0.5840999637812386, "grad_norm": 1.2968303172905138, "learning_rate": 7.780999531991092e-06, "loss": 0.7143, "step": 16127 }, { "epoch": 0.5841361825425571, "grad_norm": 1.3489283694311003, "learning_rate": 7.779855724805483e-06, "loss": 0.7385, "step": 16128 }, { "epoch": 0.5841724013038754, "grad_norm": 1.1111204644905253, "learning_rate": 7.778711948171016e-06, "loss": 0.7076, "step": 16129 }, { "epoch": 0.5842086200651938, "grad_norm": 1.011725828199407, "learning_rate": 7.77756820210342e-06, "loss": 0.7406, "step": 16130 }, { "epoch": 0.5842448388265121, "grad_norm": 0.9695662188717307, "learning_rate": 7.77642448661844e-06, "loss": 0.6833, "step": 16131 }, { "epoch": 0.5842810575878304, "grad_norm": 1.146052263364039, "learning_rate": 7.775280801731813e-06, "loss": 0.7094, "step": 16132 }, { "epoch": 0.5843172763491489, "grad_norm": 1.3229481669552934, "learning_rate": 7.774137147459278e-06, "loss": 0.7536, "step": 16133 }, { "epoch": 0.5843534951104672, "grad_norm": 0.9903040799142012, "learning_rate": 7.77299352381657e-06, "loss": 0.662, "step": 16134 }, { "epoch": 0.5843897138717856, "grad_norm": 1.3579047901496084, "learning_rate": 7.771849930819427e-06, "loss": 0.7023, "step": 16135 }, { "epoch": 0.5844259326331039, "grad_norm": 1.1272548398918947, "learning_rate": 7.770706368483589e-06, "loss": 0.7121, "step": 16136 }, { "epoch": 0.5844621513944223, "grad_norm": 1.304684257788827, "learning_rate": 7.76956283682479e-06, "loss": 0.7046, "step": 16137 }, { "epoch": 0.5844983701557407, "grad_norm": 1.3958311774079435, "learning_rate": 7.768419335858766e-06, "loss": 0.6927, "step": 16138 }, { "epoch": 0.584534588917059, "grad_norm": 1.3978993216725726, "learning_rate": 7.76727586560125e-06, "loss": 0.687, "step": 16139 }, { "epoch": 0.5845708076783774, "grad_norm": 1.4690798670572514, "learning_rate": 7.766132426067983e-06, "loss": 0.8506, "step": 16140 }, { "epoch": 0.5846070264396958, "grad_norm": 1.2612391466764494, "learning_rate": 7.764989017274696e-06, "loss": 0.7666, "step": 16141 }, { "epoch": 0.5846432452010141, "grad_norm": 1.224170309887363, "learning_rate": 7.763845639237125e-06, "loss": 0.7359, "step": 16142 }, { "epoch": 0.5846794639623325, "grad_norm": 1.2870246616207397, "learning_rate": 7.762702291971e-06, "loss": 0.7098, "step": 16143 }, { "epoch": 0.5847156827236508, "grad_norm": 1.2923011259044832, "learning_rate": 7.761558975492059e-06, "loss": 0.6788, "step": 16144 }, { "epoch": 0.5847519014849692, "grad_norm": 1.1360954209329706, "learning_rate": 7.760415689816035e-06, "loss": 0.6749, "step": 16145 }, { "epoch": 0.5847881202462876, "grad_norm": 1.2347771873644782, "learning_rate": 7.759272434958656e-06, "loss": 0.7065, "step": 16146 }, { "epoch": 0.5848243390076059, "grad_norm": 1.2470378886453486, "learning_rate": 7.75812921093566e-06, "loss": 0.6955, "step": 16147 }, { "epoch": 0.5848605577689243, "grad_norm": 1.241628547334409, "learning_rate": 7.756986017762773e-06, "loss": 0.6866, "step": 16148 }, { "epoch": 0.5848967765302426, "grad_norm": 1.247572112336792, "learning_rate": 7.755842855455734e-06, "loss": 0.7695, "step": 16149 }, { "epoch": 0.5849329952915611, "grad_norm": 1.273088264699641, "learning_rate": 7.754699724030263e-06, "loss": 0.7671, "step": 16150 }, { "epoch": 0.5849692140528794, "grad_norm": 1.1856304046283994, "learning_rate": 7.7535566235021e-06, "loss": 0.7261, "step": 16151 }, { "epoch": 0.5850054328141977, "grad_norm": 1.326376598607316, "learning_rate": 7.752413553886971e-06, "loss": 0.7305, "step": 16152 }, { "epoch": 0.5850416515755161, "grad_norm": 1.472995885215845, "learning_rate": 7.751270515200608e-06, "loss": 0.7328, "step": 16153 }, { "epoch": 0.5850778703368344, "grad_norm": 1.1711324945462647, "learning_rate": 7.750127507458737e-06, "loss": 0.6749, "step": 16154 }, { "epoch": 0.5851140890981529, "grad_norm": 1.1597779817599851, "learning_rate": 7.748984530677088e-06, "loss": 0.728, "step": 16155 }, { "epoch": 0.5851503078594712, "grad_norm": 0.9839828741212463, "learning_rate": 7.747841584871391e-06, "loss": 0.7115, "step": 16156 }, { "epoch": 0.5851865266207895, "grad_norm": 1.2701242147829375, "learning_rate": 7.746698670057372e-06, "loss": 0.6736, "step": 16157 }, { "epoch": 0.585222745382108, "grad_norm": 1.532847725320566, "learning_rate": 7.745555786250762e-06, "loss": 0.7583, "step": 16158 }, { "epoch": 0.5852589641434263, "grad_norm": 1.2982712663355482, "learning_rate": 7.744412933467282e-06, "loss": 0.6743, "step": 16159 }, { "epoch": 0.5852951829047447, "grad_norm": 0.9690891513873885, "learning_rate": 7.743270111722662e-06, "loss": 0.711, "step": 16160 }, { "epoch": 0.585331401666063, "grad_norm": 1.3664514563047152, "learning_rate": 7.742127321032628e-06, "loss": 0.68, "step": 16161 }, { "epoch": 0.5853676204273813, "grad_norm": 1.3986414748617297, "learning_rate": 7.74098456141291e-06, "loss": 0.7153, "step": 16162 }, { "epoch": 0.5854038391886998, "grad_norm": 1.322970914307294, "learning_rate": 7.739841832879225e-06, "loss": 0.7906, "step": 16163 }, { "epoch": 0.5854400579500181, "grad_norm": 1.2438113669191462, "learning_rate": 7.738699135447305e-06, "loss": 0.6342, "step": 16164 }, { "epoch": 0.5854762767113365, "grad_norm": 1.4978028114772737, "learning_rate": 7.73755646913287e-06, "loss": 0.7504, "step": 16165 }, { "epoch": 0.5855124954726548, "grad_norm": 1.2783237505313578, "learning_rate": 7.736413833951646e-06, "loss": 0.6923, "step": 16166 }, { "epoch": 0.5855487142339731, "grad_norm": 1.3248424629529552, "learning_rate": 7.735271229919358e-06, "loss": 0.69, "step": 16167 }, { "epoch": 0.5855849329952916, "grad_norm": 1.2766903739527231, "learning_rate": 7.734128657051725e-06, "loss": 0.6651, "step": 16168 }, { "epoch": 0.5856211517566099, "grad_norm": 1.1060085350953424, "learning_rate": 7.732986115364477e-06, "loss": 0.8017, "step": 16169 }, { "epoch": 0.5856573705179283, "grad_norm": 1.0129324083457938, "learning_rate": 7.731843604873328e-06, "loss": 0.7132, "step": 16170 }, { "epoch": 0.5856935892792466, "grad_norm": 1.4498229720822857, "learning_rate": 7.730701125594007e-06, "loss": 0.7643, "step": 16171 }, { "epoch": 0.585729808040565, "grad_norm": 1.335471065558732, "learning_rate": 7.72955867754223e-06, "loss": 0.6783, "step": 16172 }, { "epoch": 0.5857660268018834, "grad_norm": 1.384520364172988, "learning_rate": 7.728416260733723e-06, "loss": 0.6944, "step": 16173 }, { "epoch": 0.5858022455632017, "grad_norm": 1.4304912293489274, "learning_rate": 7.727273875184203e-06, "loss": 0.742, "step": 16174 }, { "epoch": 0.5858384643245201, "grad_norm": 1.3742332684626664, "learning_rate": 7.72613152090939e-06, "loss": 0.7639, "step": 16175 }, { "epoch": 0.5858746830858385, "grad_norm": 1.4697494758518075, "learning_rate": 7.724989197925008e-06, "loss": 0.7466, "step": 16176 }, { "epoch": 0.5859109018471568, "grad_norm": 1.285349390045171, "learning_rate": 7.72384690624677e-06, "loss": 0.6615, "step": 16177 }, { "epoch": 0.5859471206084752, "grad_norm": 1.3769634036769454, "learning_rate": 7.722704645890402e-06, "loss": 0.6895, "step": 16178 }, { "epoch": 0.5859833393697935, "grad_norm": 1.443714782049979, "learning_rate": 7.721562416871617e-06, "loss": 0.7851, "step": 16179 }, { "epoch": 0.586019558131112, "grad_norm": 1.435661643494353, "learning_rate": 7.720420219206138e-06, "loss": 0.7141, "step": 16180 }, { "epoch": 0.5860557768924303, "grad_norm": 1.5506618895826467, "learning_rate": 7.719278052909678e-06, "loss": 0.7291, "step": 16181 }, { "epoch": 0.5860919956537486, "grad_norm": 1.21581504181022, "learning_rate": 7.718135917997956e-06, "loss": 0.7144, "step": 16182 }, { "epoch": 0.586128214415067, "grad_norm": 1.3809188434739474, "learning_rate": 7.716993814486689e-06, "loss": 0.7517, "step": 16183 }, { "epoch": 0.5861644331763853, "grad_norm": 1.3396811887144398, "learning_rate": 7.715851742391594e-06, "loss": 0.7469, "step": 16184 }, { "epoch": 0.5862006519377038, "grad_norm": 1.406689996231643, "learning_rate": 7.714709701728383e-06, "loss": 0.7006, "step": 16185 }, { "epoch": 0.5862368706990221, "grad_norm": 1.1049063838299162, "learning_rate": 7.713567692512775e-06, "loss": 0.6271, "step": 16186 }, { "epoch": 0.5862730894603404, "grad_norm": 1.670997381485717, "learning_rate": 7.712425714760488e-06, "loss": 0.7744, "step": 16187 }, { "epoch": 0.5863093082216588, "grad_norm": 1.303705834796649, "learning_rate": 7.71128376848723e-06, "loss": 0.6877, "step": 16188 }, { "epoch": 0.5863455269829771, "grad_norm": 1.3108311613418149, "learning_rate": 7.710141853708722e-06, "loss": 0.6614, "step": 16189 }, { "epoch": 0.5863817457442956, "grad_norm": 1.093830778872793, "learning_rate": 7.70899997044067e-06, "loss": 0.7202, "step": 16190 }, { "epoch": 0.5864179645056139, "grad_norm": 1.0160365724755385, "learning_rate": 7.707858118698794e-06, "loss": 0.7207, "step": 16191 }, { "epoch": 0.5864541832669322, "grad_norm": 1.4103246380927932, "learning_rate": 7.706716298498802e-06, "loss": 0.7042, "step": 16192 }, { "epoch": 0.5864904020282506, "grad_norm": 1.3201788490411805, "learning_rate": 7.705574509856412e-06, "loss": 0.7636, "step": 16193 }, { "epoch": 0.586526620789569, "grad_norm": 1.2372995348629625, "learning_rate": 7.704432752787331e-06, "loss": 0.6505, "step": 16194 }, { "epoch": 0.5865628395508874, "grad_norm": 1.1868186026259138, "learning_rate": 7.703291027307273e-06, "loss": 0.6633, "step": 16195 }, { "epoch": 0.5865990583122057, "grad_norm": 1.3214442940090414, "learning_rate": 7.702149333431947e-06, "loss": 0.7064, "step": 16196 }, { "epoch": 0.586635277073524, "grad_norm": 1.347137678966568, "learning_rate": 7.701007671177066e-06, "loss": 0.8031, "step": 16197 }, { "epoch": 0.5866714958348425, "grad_norm": 0.9480948704510929, "learning_rate": 7.69986604055834e-06, "loss": 0.6782, "step": 16198 }, { "epoch": 0.5867077145961608, "grad_norm": 1.4209051204144079, "learning_rate": 7.69872444159148e-06, "loss": 0.7201, "step": 16199 }, { "epoch": 0.5867439333574792, "grad_norm": 1.0678859246017298, "learning_rate": 7.697582874292192e-06, "loss": 0.7043, "step": 16200 }, { "epoch": 0.5867801521187975, "grad_norm": 1.4347160251071251, "learning_rate": 7.696441338676185e-06, "loss": 0.7001, "step": 16201 }, { "epoch": 0.5868163708801158, "grad_norm": 1.290685731287921, "learning_rate": 7.695299834759172e-06, "loss": 0.6973, "step": 16202 }, { "epoch": 0.5868525896414343, "grad_norm": 1.4644433326220818, "learning_rate": 7.694158362556858e-06, "loss": 0.7109, "step": 16203 }, { "epoch": 0.5868888084027526, "grad_norm": 1.36973783362481, "learning_rate": 7.69301692208495e-06, "loss": 0.7166, "step": 16204 }, { "epoch": 0.586925027164071, "grad_norm": 1.338844568324435, "learning_rate": 7.691875513359156e-06, "loss": 0.6907, "step": 16205 }, { "epoch": 0.5869612459253893, "grad_norm": 1.41112561735543, "learning_rate": 7.690734136395183e-06, "loss": 0.7642, "step": 16206 }, { "epoch": 0.5869974646867077, "grad_norm": 1.3020627779927605, "learning_rate": 7.68959279120874e-06, "loss": 0.7039, "step": 16207 }, { "epoch": 0.5870336834480261, "grad_norm": 1.2052763037121288, "learning_rate": 7.688451477815526e-06, "loss": 0.6506, "step": 16208 }, { "epoch": 0.5870699022093444, "grad_norm": 1.1823118738689622, "learning_rate": 7.687310196231255e-06, "loss": 0.6775, "step": 16209 }, { "epoch": 0.5871061209706628, "grad_norm": 1.1250170155486798, "learning_rate": 7.686168946471624e-06, "loss": 0.7526, "step": 16210 }, { "epoch": 0.5871423397319812, "grad_norm": 1.3381633995596365, "learning_rate": 7.685027728552344e-06, "loss": 0.7378, "step": 16211 }, { "epoch": 0.5871785584932995, "grad_norm": 0.979889641139302, "learning_rate": 7.683886542489117e-06, "loss": 0.6773, "step": 16212 }, { "epoch": 0.5872147772546179, "grad_norm": 1.3142900953892922, "learning_rate": 7.682745388297645e-06, "loss": 0.7441, "step": 16213 }, { "epoch": 0.5872509960159362, "grad_norm": 1.3599729428685237, "learning_rate": 7.681604265993631e-06, "loss": 0.756, "step": 16214 }, { "epoch": 0.5872872147772547, "grad_norm": 1.3201476662914684, "learning_rate": 7.680463175592782e-06, "loss": 0.695, "step": 16215 }, { "epoch": 0.587323433538573, "grad_norm": 1.5688364187028687, "learning_rate": 7.679322117110797e-06, "loss": 0.7477, "step": 16216 }, { "epoch": 0.5873596522998913, "grad_norm": 1.361787678773577, "learning_rate": 7.678181090563379e-06, "loss": 0.6629, "step": 16217 }, { "epoch": 0.5873958710612097, "grad_norm": 1.3882158076425999, "learning_rate": 7.67704009596623e-06, "loss": 0.7862, "step": 16218 }, { "epoch": 0.587432089822528, "grad_norm": 1.3788134736845334, "learning_rate": 7.675899133335047e-06, "loss": 0.6793, "step": 16219 }, { "epoch": 0.5874683085838465, "grad_norm": 0.9988002766154351, "learning_rate": 7.674758202685538e-06, "loss": 0.7639, "step": 16220 }, { "epoch": 0.5875045273451648, "grad_norm": 1.3435977493297147, "learning_rate": 7.673617304033397e-06, "loss": 0.7786, "step": 16221 }, { "epoch": 0.5875407461064832, "grad_norm": 1.241273837517917, "learning_rate": 7.672476437394328e-06, "loss": 0.7269, "step": 16222 }, { "epoch": 0.5875769648678015, "grad_norm": 1.2130288649605423, "learning_rate": 7.671335602784027e-06, "loss": 0.7446, "step": 16223 }, { "epoch": 0.5876131836291199, "grad_norm": 1.1034075979509612, "learning_rate": 7.670194800218195e-06, "loss": 0.8175, "step": 16224 }, { "epoch": 0.5876494023904383, "grad_norm": 1.5240510548659925, "learning_rate": 7.669054029712528e-06, "loss": 0.7391, "step": 16225 }, { "epoch": 0.5876856211517566, "grad_norm": 1.3245938523760237, "learning_rate": 7.667913291282725e-06, "loss": 0.7023, "step": 16226 }, { "epoch": 0.587721839913075, "grad_norm": 1.2830088379789482, "learning_rate": 7.666772584944488e-06, "loss": 0.7432, "step": 16227 }, { "epoch": 0.5877580586743933, "grad_norm": 1.182430415993786, "learning_rate": 7.665631910713507e-06, "loss": 0.6843, "step": 16228 }, { "epoch": 0.5877942774357117, "grad_norm": 1.3553951975904566, "learning_rate": 7.664491268605486e-06, "loss": 0.6811, "step": 16229 }, { "epoch": 0.5878304961970301, "grad_norm": 1.2815249550723466, "learning_rate": 7.663350658636115e-06, "loss": 0.7737, "step": 16230 }, { "epoch": 0.5878667149583484, "grad_norm": 1.4207913492754807, "learning_rate": 7.662210080821092e-06, "loss": 0.7154, "step": 16231 }, { "epoch": 0.5879029337196668, "grad_norm": 1.0853174951639122, "learning_rate": 7.661069535176111e-06, "loss": 0.7173, "step": 16232 }, { "epoch": 0.5879391524809852, "grad_norm": 1.309718583750056, "learning_rate": 7.659929021716871e-06, "loss": 0.7599, "step": 16233 }, { "epoch": 0.5879753712423035, "grad_norm": 1.3472410944704318, "learning_rate": 7.658788540459063e-06, "loss": 0.7322, "step": 16234 }, { "epoch": 0.5880115900036219, "grad_norm": 1.3832017632191982, "learning_rate": 7.657648091418381e-06, "loss": 0.7186, "step": 16235 }, { "epoch": 0.5880478087649402, "grad_norm": 1.3830103193911407, "learning_rate": 7.65650767461052e-06, "loss": 0.6439, "step": 16236 }, { "epoch": 0.5880840275262587, "grad_norm": 1.1507819423261523, "learning_rate": 7.65536729005117e-06, "loss": 0.6831, "step": 16237 }, { "epoch": 0.588120246287577, "grad_norm": 1.1180981735843807, "learning_rate": 7.65422693775603e-06, "loss": 0.7144, "step": 16238 }, { "epoch": 0.5881564650488953, "grad_norm": 1.355350164893238, "learning_rate": 7.653086617740788e-06, "loss": 0.6972, "step": 16239 }, { "epoch": 0.5881926838102137, "grad_norm": 1.530293229004925, "learning_rate": 7.651946330021136e-06, "loss": 0.7614, "step": 16240 }, { "epoch": 0.588228902571532, "grad_norm": 1.4476227930920262, "learning_rate": 7.650806074612765e-06, "loss": 0.6999, "step": 16241 }, { "epoch": 0.5882651213328505, "grad_norm": 1.2536465409862039, "learning_rate": 7.649665851531366e-06, "loss": 0.7081, "step": 16242 }, { "epoch": 0.5883013400941688, "grad_norm": 1.3545183029352512, "learning_rate": 7.648525660792631e-06, "loss": 0.7083, "step": 16243 }, { "epoch": 0.5883375588554871, "grad_norm": 0.9897414485759346, "learning_rate": 7.647385502412251e-06, "loss": 0.7179, "step": 16244 }, { "epoch": 0.5883737776168055, "grad_norm": 1.2863476326902867, "learning_rate": 7.64624537640591e-06, "loss": 0.6853, "step": 16245 }, { "epoch": 0.5884099963781239, "grad_norm": 1.421007805642903, "learning_rate": 7.645105282789303e-06, "loss": 0.7567, "step": 16246 }, { "epoch": 0.5884462151394423, "grad_norm": 1.3940023453066244, "learning_rate": 7.643965221578116e-06, "loss": 0.77, "step": 16247 }, { "epoch": 0.5884824339007606, "grad_norm": 1.2794691362955495, "learning_rate": 7.642825192788037e-06, "loss": 0.67, "step": 16248 }, { "epoch": 0.5885186526620789, "grad_norm": 1.334166301452962, "learning_rate": 7.641685196434756e-06, "loss": 0.776, "step": 16249 }, { "epoch": 0.5885548714233974, "grad_norm": 1.4315296158680813, "learning_rate": 7.640545232533957e-06, "loss": 0.7056, "step": 16250 }, { "epoch": 0.5885910901847157, "grad_norm": 1.3702859713401643, "learning_rate": 7.639405301101331e-06, "loss": 0.7463, "step": 16251 }, { "epoch": 0.5886273089460341, "grad_norm": 1.3377026300540664, "learning_rate": 7.63826540215256e-06, "loss": 0.7163, "step": 16252 }, { "epoch": 0.5886635277073524, "grad_norm": 1.0585003017932626, "learning_rate": 7.637125535703336e-06, "loss": 0.7189, "step": 16253 }, { "epoch": 0.5886997464686707, "grad_norm": 1.2881015009514638, "learning_rate": 7.635985701769338e-06, "loss": 0.6832, "step": 16254 }, { "epoch": 0.5887359652299892, "grad_norm": 1.4499219206431735, "learning_rate": 7.634845900366255e-06, "loss": 0.7573, "step": 16255 }, { "epoch": 0.5887721839913075, "grad_norm": 1.207250434764125, "learning_rate": 7.633706131509769e-06, "loss": 0.7043, "step": 16256 }, { "epoch": 0.5888084027526259, "grad_norm": 1.311048057251151, "learning_rate": 7.632566395215565e-06, "loss": 0.7183, "step": 16257 }, { "epoch": 0.5888446215139442, "grad_norm": 1.3995860680224033, "learning_rate": 7.63142669149933e-06, "loss": 0.7287, "step": 16258 }, { "epoch": 0.5888808402752626, "grad_norm": 1.31911394777397, "learning_rate": 7.630287020376743e-06, "loss": 0.8055, "step": 16259 }, { "epoch": 0.588917059036581, "grad_norm": 1.9641670100615467, "learning_rate": 7.629147381863492e-06, "loss": 0.6877, "step": 16260 }, { "epoch": 0.5889532777978993, "grad_norm": 1.2612718567239813, "learning_rate": 7.628007775975254e-06, "loss": 0.6211, "step": 16261 }, { "epoch": 0.5889894965592177, "grad_norm": 1.4086453197219448, "learning_rate": 7.626868202727714e-06, "loss": 0.742, "step": 16262 }, { "epoch": 0.589025715320536, "grad_norm": 1.301905084250711, "learning_rate": 7.6257286621365516e-06, "loss": 0.6914, "step": 16263 }, { "epoch": 0.5890619340818544, "grad_norm": 1.4234379821236354, "learning_rate": 7.6245891542174495e-06, "loss": 0.6829, "step": 16264 }, { "epoch": 0.5890981528431728, "grad_norm": 1.02607777538992, "learning_rate": 7.6234496789860875e-06, "loss": 0.6578, "step": 16265 }, { "epoch": 0.5891343716044911, "grad_norm": 1.2487779716119232, "learning_rate": 7.622310236458148e-06, "loss": 0.6251, "step": 16266 }, { "epoch": 0.5891705903658095, "grad_norm": 1.3470319960671426, "learning_rate": 7.6211708266493065e-06, "loss": 0.7376, "step": 16267 }, { "epoch": 0.5892068091271279, "grad_norm": 1.0482552058465415, "learning_rate": 7.6200314495752446e-06, "loss": 0.7793, "step": 16268 }, { "epoch": 0.5892430278884462, "grad_norm": 1.4155381772382396, "learning_rate": 7.618892105251644e-06, "loss": 0.7465, "step": 16269 }, { "epoch": 0.5892792466497646, "grad_norm": 1.3423465834043657, "learning_rate": 7.617752793694178e-06, "loss": 0.757, "step": 16270 }, { "epoch": 0.5893154654110829, "grad_norm": 1.2199184169103423, "learning_rate": 7.616613514918528e-06, "loss": 0.6551, "step": 16271 }, { "epoch": 0.5893516841724014, "grad_norm": 1.3755010659108144, "learning_rate": 7.61547426894037e-06, "loss": 0.7304, "step": 16272 }, { "epoch": 0.5893879029337197, "grad_norm": 1.328298853967448, "learning_rate": 7.6143350557753815e-06, "loss": 0.7288, "step": 16273 }, { "epoch": 0.589424121695038, "grad_norm": 1.0582690231371503, "learning_rate": 7.613195875439236e-06, "loss": 0.6825, "step": 16274 }, { "epoch": 0.5894603404563564, "grad_norm": 1.418490947508315, "learning_rate": 7.612056727947616e-06, "loss": 0.7507, "step": 16275 }, { "epoch": 0.5894965592176747, "grad_norm": 1.1426637697419877, "learning_rate": 7.610917613316192e-06, "loss": 0.6946, "step": 16276 }, { "epoch": 0.5895327779789932, "grad_norm": 1.1273403085225184, "learning_rate": 7.609778531560643e-06, "loss": 0.7063, "step": 16277 }, { "epoch": 0.5895689967403115, "grad_norm": 1.2919074172728908, "learning_rate": 7.608639482696638e-06, "loss": 0.7212, "step": 16278 }, { "epoch": 0.5896052155016298, "grad_norm": 1.0862537259579759, "learning_rate": 7.607500466739856e-06, "loss": 0.7283, "step": 16279 }, { "epoch": 0.5896414342629482, "grad_norm": 1.5168302760252237, "learning_rate": 7.606361483705973e-06, "loss": 0.6793, "step": 16280 }, { "epoch": 0.5896776530242666, "grad_norm": 1.4398882439890368, "learning_rate": 7.605222533610656e-06, "loss": 0.7011, "step": 16281 }, { "epoch": 0.589713871785585, "grad_norm": 1.2823944541156158, "learning_rate": 7.604083616469582e-06, "loss": 0.6796, "step": 16282 }, { "epoch": 0.5897500905469033, "grad_norm": 1.4270646655005668, "learning_rate": 7.602944732298423e-06, "loss": 0.7359, "step": 16283 }, { "epoch": 0.5897863093082216, "grad_norm": 1.2081757631481178, "learning_rate": 7.601805881112851e-06, "loss": 0.6888, "step": 16284 }, { "epoch": 0.58982252806954, "grad_norm": 1.055135670726581, "learning_rate": 7.600667062928535e-06, "loss": 0.6882, "step": 16285 }, { "epoch": 0.5898587468308584, "grad_norm": 1.3826650642607021, "learning_rate": 7.599528277761152e-06, "loss": 0.7216, "step": 16286 }, { "epoch": 0.5898949655921768, "grad_norm": 1.383762962402066, "learning_rate": 7.5983895256263664e-06, "loss": 0.7146, "step": 16287 }, { "epoch": 0.5899311843534951, "grad_norm": 1.3238996723344358, "learning_rate": 7.597250806539851e-06, "loss": 0.7396, "step": 16288 }, { "epoch": 0.5899674031148134, "grad_norm": 1.372512559447861, "learning_rate": 7.596112120517279e-06, "loss": 0.7527, "step": 16289 }, { "epoch": 0.5900036218761319, "grad_norm": 1.416857043235803, "learning_rate": 7.594973467574313e-06, "loss": 0.7097, "step": 16290 }, { "epoch": 0.5900398406374502, "grad_norm": 1.403276076445872, "learning_rate": 7.593834847726628e-06, "loss": 0.8027, "step": 16291 }, { "epoch": 0.5900760593987686, "grad_norm": 1.4162470588975502, "learning_rate": 7.5926962609898865e-06, "loss": 0.7116, "step": 16292 }, { "epoch": 0.5901122781600869, "grad_norm": 1.6959798604627956, "learning_rate": 7.591557707379763e-06, "loss": 0.7135, "step": 16293 }, { "epoch": 0.5901484969214053, "grad_norm": 1.251614529293772, "learning_rate": 7.590419186911919e-06, "loss": 0.7674, "step": 16294 }, { "epoch": 0.5901847156827237, "grad_norm": 1.4657460371154523, "learning_rate": 7.589280699602028e-06, "loss": 0.6894, "step": 16295 }, { "epoch": 0.590220934444042, "grad_norm": 1.4181551889536759, "learning_rate": 7.588142245465748e-06, "loss": 0.7497, "step": 16296 }, { "epoch": 0.5902571532053604, "grad_norm": 1.0385716937742402, "learning_rate": 7.587003824518754e-06, "loss": 0.6975, "step": 16297 }, { "epoch": 0.5902933719666787, "grad_norm": 1.1123372424107354, "learning_rate": 7.585865436776705e-06, "loss": 0.6541, "step": 16298 }, { "epoch": 0.5903295907279971, "grad_norm": 1.153304062120376, "learning_rate": 7.584727082255269e-06, "loss": 0.7103, "step": 16299 }, { "epoch": 0.5903658094893155, "grad_norm": 1.3571479850372798, "learning_rate": 7.583588760970112e-06, "loss": 0.7073, "step": 16300 }, { "epoch": 0.5904020282506338, "grad_norm": 1.4064118735871978, "learning_rate": 7.582450472936895e-06, "loss": 0.7027, "step": 16301 }, { "epoch": 0.5904382470119522, "grad_norm": 1.4077745319775965, "learning_rate": 7.581312218171286e-06, "loss": 0.6936, "step": 16302 }, { "epoch": 0.5904744657732706, "grad_norm": 1.377315199031904, "learning_rate": 7.580173996688944e-06, "loss": 0.693, "step": 16303 }, { "epoch": 0.5905106845345889, "grad_norm": 1.6055278419144978, "learning_rate": 7.579035808505537e-06, "loss": 0.6759, "step": 16304 }, { "epoch": 0.5905469032959073, "grad_norm": 1.3619074006316367, "learning_rate": 7.577897653636721e-06, "loss": 0.745, "step": 16305 }, { "epoch": 0.5905831220572256, "grad_norm": 1.52043312149007, "learning_rate": 7.5767595320981645e-06, "loss": 0.7237, "step": 16306 }, { "epoch": 0.5906193408185441, "grad_norm": 1.2925408590264456, "learning_rate": 7.575621443905524e-06, "loss": 0.692, "step": 16307 }, { "epoch": 0.5906555595798624, "grad_norm": 1.349764318860228, "learning_rate": 7.574483389074465e-06, "loss": 0.7474, "step": 16308 }, { "epoch": 0.5906917783411807, "grad_norm": 1.3428068441223497, "learning_rate": 7.573345367620644e-06, "loss": 0.7428, "step": 16309 }, { "epoch": 0.5907279971024991, "grad_norm": 1.3553135690988323, "learning_rate": 7.572207379559722e-06, "loss": 0.6551, "step": 16310 }, { "epoch": 0.5907642158638174, "grad_norm": 1.2550815565903521, "learning_rate": 7.5710694249073615e-06, "loss": 0.6593, "step": 16311 }, { "epoch": 0.5908004346251359, "grad_norm": 1.2748868284592632, "learning_rate": 7.5699315036792185e-06, "loss": 0.6263, "step": 16312 }, { "epoch": 0.5908366533864542, "grad_norm": 1.2031830395905228, "learning_rate": 7.568793615890955e-06, "loss": 0.7072, "step": 16313 }, { "epoch": 0.5908728721477725, "grad_norm": 1.2572341457065355, "learning_rate": 7.5676557615582255e-06, "loss": 0.6822, "step": 16314 }, { "epoch": 0.5909090909090909, "grad_norm": 1.1733884701116106, "learning_rate": 7.566517940696693e-06, "loss": 0.7411, "step": 16315 }, { "epoch": 0.5909453096704093, "grad_norm": 1.3420418164544217, "learning_rate": 7.565380153322008e-06, "loss": 0.7148, "step": 16316 }, { "epoch": 0.5909815284317277, "grad_norm": 0.9538843054945833, "learning_rate": 7.564242399449836e-06, "loss": 0.7508, "step": 16317 }, { "epoch": 0.591017747193046, "grad_norm": 1.329417533790397, "learning_rate": 7.563104679095825e-06, "loss": 0.6778, "step": 16318 }, { "epoch": 0.5910539659543643, "grad_norm": 1.5693258808623227, "learning_rate": 7.561966992275636e-06, "loss": 0.8183, "step": 16319 }, { "epoch": 0.5910901847156828, "grad_norm": 1.4386106468738071, "learning_rate": 7.560829339004923e-06, "loss": 0.8388, "step": 16320 }, { "epoch": 0.5911264034770011, "grad_norm": 1.2701299803393282, "learning_rate": 7.559691719299342e-06, "loss": 0.685, "step": 16321 }, { "epoch": 0.5911626222383195, "grad_norm": 1.2430476293704804, "learning_rate": 7.558554133174548e-06, "loss": 0.6955, "step": 16322 }, { "epoch": 0.5911988409996378, "grad_norm": 0.957222474101539, "learning_rate": 7.557416580646193e-06, "loss": 0.7322, "step": 16323 }, { "epoch": 0.5912350597609561, "grad_norm": 1.260895062344609, "learning_rate": 7.556279061729934e-06, "loss": 0.7207, "step": 16324 }, { "epoch": 0.5912712785222746, "grad_norm": 1.4297645488155704, "learning_rate": 7.55514157644142e-06, "loss": 0.7798, "step": 16325 }, { "epoch": 0.5913074972835929, "grad_norm": 1.2739357798956343, "learning_rate": 7.554004124796308e-06, "loss": 0.7517, "step": 16326 }, { "epoch": 0.5913437160449113, "grad_norm": 1.2972468394613883, "learning_rate": 7.552866706810247e-06, "loss": 0.7664, "step": 16327 }, { "epoch": 0.5913799348062296, "grad_norm": 1.2630580807302791, "learning_rate": 7.551729322498892e-06, "loss": 0.6892, "step": 16328 }, { "epoch": 0.591416153567548, "grad_norm": 1.1015858916156187, "learning_rate": 7.550591971877895e-06, "loss": 0.7497, "step": 16329 }, { "epoch": 0.5914523723288664, "grad_norm": 1.3209822482468294, "learning_rate": 7.5494546549629015e-06, "loss": 0.702, "step": 16330 }, { "epoch": 0.5914885910901847, "grad_norm": 1.0561541413121263, "learning_rate": 7.548317371769563e-06, "loss": 0.7583, "step": 16331 }, { "epoch": 0.5915248098515031, "grad_norm": 1.300900819655588, "learning_rate": 7.547180122313535e-06, "loss": 0.6513, "step": 16332 }, { "epoch": 0.5915610286128214, "grad_norm": 1.449967496811842, "learning_rate": 7.546042906610462e-06, "loss": 0.7356, "step": 16333 }, { "epoch": 0.5915972473741398, "grad_norm": 0.9009940645937068, "learning_rate": 7.544905724675997e-06, "loss": 0.6711, "step": 16334 }, { "epoch": 0.5916334661354582, "grad_norm": 1.3690019311461121, "learning_rate": 7.543768576525784e-06, "loss": 0.7735, "step": 16335 }, { "epoch": 0.5916696848967765, "grad_norm": 3.48601864955837, "learning_rate": 7.542631462175476e-06, "loss": 0.739, "step": 16336 }, { "epoch": 0.591705903658095, "grad_norm": 1.464905863398687, "learning_rate": 7.541494381640716e-06, "loss": 0.708, "step": 16337 }, { "epoch": 0.5917421224194133, "grad_norm": 1.2951459093897182, "learning_rate": 7.540357334937156e-06, "loss": 0.7324, "step": 16338 }, { "epoch": 0.5917783411807316, "grad_norm": 1.3264563456231315, "learning_rate": 7.539220322080437e-06, "loss": 0.7081, "step": 16339 }, { "epoch": 0.59181455994205, "grad_norm": 1.1143637997192404, "learning_rate": 7.538083343086212e-06, "loss": 0.6598, "step": 16340 }, { "epoch": 0.5918507787033683, "grad_norm": 1.3058780084379453, "learning_rate": 7.536946397970121e-06, "loss": 0.6738, "step": 16341 }, { "epoch": 0.5918869974646868, "grad_norm": 1.2343903190531387, "learning_rate": 7.535809486747811e-06, "loss": 0.7249, "step": 16342 }, { "epoch": 0.5919232162260051, "grad_norm": 1.071419394770809, "learning_rate": 7.534672609434931e-06, "loss": 0.6204, "step": 16343 }, { "epoch": 0.5919594349873234, "grad_norm": 1.5080747499840703, "learning_rate": 7.53353576604712e-06, "loss": 0.6539, "step": 16344 }, { "epoch": 0.5919956537486418, "grad_norm": 1.2745124018936718, "learning_rate": 7.532398956600025e-06, "loss": 0.6773, "step": 16345 }, { "epoch": 0.5920318725099601, "grad_norm": 1.0778527717883721, "learning_rate": 7.531262181109287e-06, "loss": 0.6632, "step": 16346 }, { "epoch": 0.5920680912712786, "grad_norm": 1.4962032741167823, "learning_rate": 7.530125439590552e-06, "loss": 0.7713, "step": 16347 }, { "epoch": 0.5921043100325969, "grad_norm": 1.536898466422069, "learning_rate": 7.5289887320594605e-06, "loss": 0.7941, "step": 16348 }, { "epoch": 0.5921405287939152, "grad_norm": 1.4716784598099233, "learning_rate": 7.527852058531657e-06, "loss": 0.8087, "step": 16349 }, { "epoch": 0.5921767475552336, "grad_norm": 1.2348347061368588, "learning_rate": 7.526715419022779e-06, "loss": 0.6976, "step": 16350 }, { "epoch": 0.592212966316552, "grad_norm": 1.5019613707257786, "learning_rate": 7.525578813548472e-06, "loss": 0.7182, "step": 16351 }, { "epoch": 0.5922491850778704, "grad_norm": 1.367866839158988, "learning_rate": 7.5244422421243724e-06, "loss": 0.7761, "step": 16352 }, { "epoch": 0.5922854038391887, "grad_norm": 1.071657580003734, "learning_rate": 7.523305704766124e-06, "loss": 0.712, "step": 16353 }, { "epoch": 0.592321622600507, "grad_norm": 1.338701684518032, "learning_rate": 7.522169201489367e-06, "loss": 0.643, "step": 16354 }, { "epoch": 0.5923578413618255, "grad_norm": 1.287829984637864, "learning_rate": 7.521032732309736e-06, "loss": 0.7334, "step": 16355 }, { "epoch": 0.5923940601231438, "grad_norm": 1.2936851232670323, "learning_rate": 7.519896297242874e-06, "loss": 0.6843, "step": 16356 }, { "epoch": 0.5924302788844622, "grad_norm": 1.275095225586977, "learning_rate": 7.518759896304419e-06, "loss": 0.7484, "step": 16357 }, { "epoch": 0.5924664976457805, "grad_norm": 1.3136533795432084, "learning_rate": 7.517623529510009e-06, "loss": 0.7519, "step": 16358 }, { "epoch": 0.5925027164070988, "grad_norm": 1.3631801656557898, "learning_rate": 7.516487196875277e-06, "loss": 0.7414, "step": 16359 }, { "epoch": 0.5925389351684173, "grad_norm": 1.0147906979053212, "learning_rate": 7.515350898415867e-06, "loss": 0.7044, "step": 16360 }, { "epoch": 0.5925751539297356, "grad_norm": 1.2483154061715784, "learning_rate": 7.514214634147409e-06, "loss": 0.7631, "step": 16361 }, { "epoch": 0.592611372691054, "grad_norm": 1.4017186704328797, "learning_rate": 7.513078404085542e-06, "loss": 0.7597, "step": 16362 }, { "epoch": 0.5926475914523723, "grad_norm": 1.4027538636281762, "learning_rate": 7.511942208245904e-06, "loss": 0.7386, "step": 16363 }, { "epoch": 0.5926838102136907, "grad_norm": 1.1440111639003525, "learning_rate": 7.510806046644125e-06, "loss": 0.7694, "step": 16364 }, { "epoch": 0.5927200289750091, "grad_norm": 1.3251066406623513, "learning_rate": 7.509669919295843e-06, "loss": 0.7685, "step": 16365 }, { "epoch": 0.5927562477363274, "grad_norm": 1.4433727057309813, "learning_rate": 7.50853382621669e-06, "loss": 0.7025, "step": 16366 }, { "epoch": 0.5927924664976458, "grad_norm": 1.0904427375871764, "learning_rate": 7.507397767422303e-06, "loss": 0.7944, "step": 16367 }, { "epoch": 0.5928286852589641, "grad_norm": 1.3610230347539116, "learning_rate": 7.50626174292831e-06, "loss": 0.7366, "step": 16368 }, { "epoch": 0.5928649040202825, "grad_norm": 1.4832207282616379, "learning_rate": 7.50512575275035e-06, "loss": 0.8013, "step": 16369 }, { "epoch": 0.5929011227816009, "grad_norm": 1.3092067706029757, "learning_rate": 7.503989796904049e-06, "loss": 0.6883, "step": 16370 }, { "epoch": 0.5929373415429192, "grad_norm": 1.1481108739255605, "learning_rate": 7.502853875405043e-06, "loss": 0.7264, "step": 16371 }, { "epoch": 0.5929735603042376, "grad_norm": 1.4549238353573268, "learning_rate": 7.50171798826896e-06, "loss": 0.6965, "step": 16372 }, { "epoch": 0.593009779065556, "grad_norm": 1.173955418147809, "learning_rate": 7.500582135511434e-06, "loss": 0.692, "step": 16373 }, { "epoch": 0.5930459978268743, "grad_norm": 1.4172623677124558, "learning_rate": 7.499446317148096e-06, "loss": 0.7948, "step": 16374 }, { "epoch": 0.5930822165881927, "grad_norm": 1.0030412793526144, "learning_rate": 7.49831053319457e-06, "loss": 0.6865, "step": 16375 }, { "epoch": 0.593118435349511, "grad_norm": 1.3542494250342627, "learning_rate": 7.497174783666492e-06, "loss": 0.6933, "step": 16376 }, { "epoch": 0.5931546541108295, "grad_norm": 1.25928914130055, "learning_rate": 7.496039068579486e-06, "loss": 0.6884, "step": 16377 }, { "epoch": 0.5931908728721478, "grad_norm": 0.9058008363212097, "learning_rate": 7.4949033879491865e-06, "loss": 0.6864, "step": 16378 }, { "epoch": 0.5932270916334661, "grad_norm": 1.0264393781523178, "learning_rate": 7.493767741791213e-06, "loss": 0.6815, "step": 16379 }, { "epoch": 0.5932633103947845, "grad_norm": 1.3204883135318097, "learning_rate": 7.4926321301212e-06, "loss": 0.7008, "step": 16380 }, { "epoch": 0.5932995291561028, "grad_norm": 1.3589578843611885, "learning_rate": 7.491496552954771e-06, "loss": 0.6941, "step": 16381 }, { "epoch": 0.5933357479174213, "grad_norm": 1.2583131711836122, "learning_rate": 7.490361010307556e-06, "loss": 0.6421, "step": 16382 }, { "epoch": 0.5933719666787396, "grad_norm": 1.321777182396892, "learning_rate": 7.489225502195176e-06, "loss": 0.6577, "step": 16383 }, { "epoch": 0.5934081854400579, "grad_norm": 1.2625401763731705, "learning_rate": 7.488090028633259e-06, "loss": 0.7043, "step": 16384 }, { "epoch": 0.5934444042013763, "grad_norm": 1.2379495111677592, "learning_rate": 7.486954589637432e-06, "loss": 0.6804, "step": 16385 }, { "epoch": 0.5934806229626947, "grad_norm": 1.3556229451446649, "learning_rate": 7.485819185223316e-06, "loss": 0.6482, "step": 16386 }, { "epoch": 0.5935168417240131, "grad_norm": 1.5562295059689266, "learning_rate": 7.48468381540654e-06, "loss": 0.7582, "step": 16387 }, { "epoch": 0.5935530604853314, "grad_norm": 0.935464888516219, "learning_rate": 7.483548480202722e-06, "loss": 0.6921, "step": 16388 }, { "epoch": 0.5935892792466497, "grad_norm": 1.46127162264301, "learning_rate": 7.4824131796274915e-06, "loss": 0.6884, "step": 16389 }, { "epoch": 0.5936254980079682, "grad_norm": 1.4889972421290412, "learning_rate": 7.481277913696465e-06, "loss": 0.7263, "step": 16390 }, { "epoch": 0.5936617167692865, "grad_norm": 1.3712115766971886, "learning_rate": 7.480142682425268e-06, "loss": 0.7413, "step": 16391 }, { "epoch": 0.5936979355306049, "grad_norm": 1.2699464441819257, "learning_rate": 7.479007485829523e-06, "loss": 0.6637, "step": 16392 }, { "epoch": 0.5937341542919232, "grad_norm": 1.2312043927672687, "learning_rate": 7.477872323924847e-06, "loss": 0.7247, "step": 16393 }, { "epoch": 0.5937703730532415, "grad_norm": 1.4188739822020104, "learning_rate": 7.476737196726868e-06, "loss": 0.6444, "step": 16394 }, { "epoch": 0.59380659181456, "grad_norm": 1.8350429480423003, "learning_rate": 7.4756021042511995e-06, "loss": 0.7473, "step": 16395 }, { "epoch": 0.5938428105758783, "grad_norm": 1.3265851013271182, "learning_rate": 7.474467046513467e-06, "loss": 0.7295, "step": 16396 }, { "epoch": 0.5938790293371967, "grad_norm": 1.3668420900256144, "learning_rate": 7.473332023529284e-06, "loss": 0.6943, "step": 16397 }, { "epoch": 0.593915248098515, "grad_norm": 1.0086287053553933, "learning_rate": 7.472197035314274e-06, "loss": 0.6852, "step": 16398 }, { "epoch": 0.5939514668598334, "grad_norm": 1.160688878565428, "learning_rate": 7.471062081884053e-06, "loss": 0.7633, "step": 16399 }, { "epoch": 0.5939876856211518, "grad_norm": 1.2302003734311187, "learning_rate": 7.469927163254243e-06, "loss": 0.6966, "step": 16400 }, { "epoch": 0.5940239043824701, "grad_norm": 1.4552171845520754, "learning_rate": 7.468792279440455e-06, "loss": 0.7992, "step": 16401 }, { "epoch": 0.5940601231437885, "grad_norm": 1.2394612966665437, "learning_rate": 7.4676574304583105e-06, "loss": 0.7256, "step": 16402 }, { "epoch": 0.5940963419051069, "grad_norm": 0.9666196213294771, "learning_rate": 7.466522616323424e-06, "loss": 0.6811, "step": 16403 }, { "epoch": 0.5941325606664252, "grad_norm": 1.3490387473446472, "learning_rate": 7.465387837051412e-06, "loss": 0.7357, "step": 16404 }, { "epoch": 0.5941687794277436, "grad_norm": 1.3953024323987697, "learning_rate": 7.4642530926578925e-06, "loss": 0.6755, "step": 16405 }, { "epoch": 0.5942049981890619, "grad_norm": 1.3038567365064493, "learning_rate": 7.4631183831584765e-06, "loss": 0.6908, "step": 16406 }, { "epoch": 0.5942412169503803, "grad_norm": 1.3283651962179646, "learning_rate": 7.461983708568783e-06, "loss": 0.7972, "step": 16407 }, { "epoch": 0.5942774357116987, "grad_norm": 1.3155336997322007, "learning_rate": 7.4608490689044215e-06, "loss": 0.7567, "step": 16408 }, { "epoch": 0.594313654473017, "grad_norm": 1.2429600538504642, "learning_rate": 7.459714464181009e-06, "loss": 0.7653, "step": 16409 }, { "epoch": 0.5943498732343354, "grad_norm": 1.5584439125554435, "learning_rate": 7.458579894414158e-06, "loss": 0.775, "step": 16410 }, { "epoch": 0.5943860919956537, "grad_norm": 1.3495894622437834, "learning_rate": 7.457445359619481e-06, "loss": 0.747, "step": 16411 }, { "epoch": 0.5944223107569722, "grad_norm": 1.4577986359336284, "learning_rate": 7.456310859812588e-06, "loss": 0.7266, "step": 16412 }, { "epoch": 0.5944585295182905, "grad_norm": 1.3439326688894924, "learning_rate": 7.455176395009095e-06, "loss": 0.6912, "step": 16413 }, { "epoch": 0.5944947482796088, "grad_norm": 1.4697916086193734, "learning_rate": 7.454041965224608e-06, "loss": 0.7496, "step": 16414 }, { "epoch": 0.5945309670409272, "grad_norm": 1.2706793772615281, "learning_rate": 7.452907570474741e-06, "loss": 0.6887, "step": 16415 }, { "epoch": 0.5945671858022455, "grad_norm": 1.2491929727343685, "learning_rate": 7.4517732107751075e-06, "loss": 0.6557, "step": 16416 }, { "epoch": 0.594603404563564, "grad_norm": 1.3021740194370834, "learning_rate": 7.450638886141311e-06, "loss": 0.6842, "step": 16417 }, { "epoch": 0.5946396233248823, "grad_norm": 1.2804069827661455, "learning_rate": 7.449504596588964e-06, "loss": 0.7011, "step": 16418 }, { "epoch": 0.5946758420862006, "grad_norm": 1.297609943950854, "learning_rate": 7.4483703421336736e-06, "loss": 0.6779, "step": 16419 }, { "epoch": 0.594712060847519, "grad_norm": 1.2707308149091245, "learning_rate": 7.447236122791052e-06, "loss": 0.7104, "step": 16420 }, { "epoch": 0.5947482796088374, "grad_norm": 1.4371795348916965, "learning_rate": 7.446101938576702e-06, "loss": 0.6801, "step": 16421 }, { "epoch": 0.5947844983701558, "grad_norm": 1.3030956273576435, "learning_rate": 7.444967789506234e-06, "loss": 0.756, "step": 16422 }, { "epoch": 0.5948207171314741, "grad_norm": 1.4137158495853595, "learning_rate": 7.443833675595254e-06, "loss": 0.7355, "step": 16423 }, { "epoch": 0.5948569358927924, "grad_norm": 1.3560617933234114, "learning_rate": 7.442699596859369e-06, "loss": 0.6619, "step": 16424 }, { "epoch": 0.5948931546541109, "grad_norm": 1.0235190245320978, "learning_rate": 7.441565553314185e-06, "loss": 0.7069, "step": 16425 }, { "epoch": 0.5949293734154292, "grad_norm": 1.3319186158816605, "learning_rate": 7.440431544975306e-06, "loss": 0.7552, "step": 16426 }, { "epoch": 0.5949655921767476, "grad_norm": 1.0547607123000533, "learning_rate": 7.439297571858341e-06, "loss": 0.6934, "step": 16427 }, { "epoch": 0.5950018109380659, "grad_norm": 2.3646009386068223, "learning_rate": 7.438163633978888e-06, "loss": 0.7508, "step": 16428 }, { "epoch": 0.5950380296993842, "grad_norm": 1.35442558405355, "learning_rate": 7.437029731352556e-06, "loss": 0.7251, "step": 16429 }, { "epoch": 0.5950742484607027, "grad_norm": 1.0489583083382503, "learning_rate": 7.435895863994946e-06, "loss": 0.7454, "step": 16430 }, { "epoch": 0.595110467222021, "grad_norm": 1.0002645786553948, "learning_rate": 7.434762031921663e-06, "loss": 0.701, "step": 16431 }, { "epoch": 0.5951466859833394, "grad_norm": 1.2840106512003286, "learning_rate": 7.4336282351483065e-06, "loss": 0.685, "step": 16432 }, { "epoch": 0.5951829047446577, "grad_norm": 1.439527337582289, "learning_rate": 7.432494473690484e-06, "loss": 0.6784, "step": 16433 }, { "epoch": 0.595219123505976, "grad_norm": 1.2220362829695641, "learning_rate": 7.431360747563791e-06, "loss": 0.6829, "step": 16434 }, { "epoch": 0.5952553422672945, "grad_norm": 1.0227265746151584, "learning_rate": 7.430227056783829e-06, "loss": 0.7695, "step": 16435 }, { "epoch": 0.5952915610286128, "grad_norm": 1.3054995686404982, "learning_rate": 7.429093401366205e-06, "loss": 0.7351, "step": 16436 }, { "epoch": 0.5953277797899312, "grad_norm": 1.1391488728534687, "learning_rate": 7.427959781326512e-06, "loss": 0.639, "step": 16437 }, { "epoch": 0.5953639985512496, "grad_norm": 1.3939013009982495, "learning_rate": 7.426826196680354e-06, "loss": 0.69, "step": 16438 }, { "epoch": 0.5954002173125679, "grad_norm": 1.3563650380735437, "learning_rate": 7.425692647443328e-06, "loss": 0.7856, "step": 16439 }, { "epoch": 0.5954364360738863, "grad_norm": 1.3914661843555096, "learning_rate": 7.424559133631035e-06, "loss": 0.7774, "step": 16440 }, { "epoch": 0.5954726548352046, "grad_norm": 1.2737441439699881, "learning_rate": 7.423425655259068e-06, "loss": 0.7381, "step": 16441 }, { "epoch": 0.595508873596523, "grad_norm": 1.4469055546403606, "learning_rate": 7.42229221234303e-06, "loss": 0.7636, "step": 16442 }, { "epoch": 0.5955450923578414, "grad_norm": 1.2859159278045171, "learning_rate": 7.421158804898515e-06, "loss": 0.753, "step": 16443 }, { "epoch": 0.5955813111191597, "grad_norm": 1.3772236646189, "learning_rate": 7.420025432941122e-06, "loss": 0.7104, "step": 16444 }, { "epoch": 0.5956175298804781, "grad_norm": 1.043041188474996, "learning_rate": 7.4188920964864455e-06, "loss": 0.7778, "step": 16445 }, { "epoch": 0.5956537486417964, "grad_norm": 1.4457754663056746, "learning_rate": 7.417758795550081e-06, "loss": 0.7547, "step": 16446 }, { "epoch": 0.5956899674031149, "grad_norm": 1.6512550584385688, "learning_rate": 7.4166255301476255e-06, "loss": 0.7411, "step": 16447 }, { "epoch": 0.5957261861644332, "grad_norm": 1.034450133963567, "learning_rate": 7.415492300294671e-06, "loss": 0.6734, "step": 16448 }, { "epoch": 0.5957624049257515, "grad_norm": 1.318624091239462, "learning_rate": 7.4143591060068154e-06, "loss": 0.7821, "step": 16449 }, { "epoch": 0.5957986236870699, "grad_norm": 1.2322645925825764, "learning_rate": 7.4132259472996496e-06, "loss": 0.7069, "step": 16450 }, { "epoch": 0.5958348424483882, "grad_norm": 1.1551405338681138, "learning_rate": 7.412092824188768e-06, "loss": 0.6857, "step": 16451 }, { "epoch": 0.5958710612097067, "grad_norm": 1.379619586780847, "learning_rate": 7.410959736689762e-06, "loss": 0.6572, "step": 16452 }, { "epoch": 0.595907279971025, "grad_norm": 1.2960386214796407, "learning_rate": 7.409826684818227e-06, "loss": 0.7127, "step": 16453 }, { "epoch": 0.5959434987323433, "grad_norm": 1.0379285433163512, "learning_rate": 7.40869366858975e-06, "loss": 0.6917, "step": 16454 }, { "epoch": 0.5959797174936617, "grad_norm": 1.39159671247206, "learning_rate": 7.407560688019926e-06, "loss": 0.6532, "step": 16455 }, { "epoch": 0.5960159362549801, "grad_norm": 1.7429764761554731, "learning_rate": 7.406427743124346e-06, "loss": 0.739, "step": 16456 }, { "epoch": 0.5960521550162985, "grad_norm": 1.2661341957103796, "learning_rate": 7.405294833918597e-06, "loss": 0.736, "step": 16457 }, { "epoch": 0.5960883737776168, "grad_norm": 1.3331947041940524, "learning_rate": 7.4041619604182745e-06, "loss": 0.651, "step": 16458 }, { "epoch": 0.5961245925389351, "grad_norm": 1.1675384200278403, "learning_rate": 7.403029122638961e-06, "loss": 0.7549, "step": 16459 }, { "epoch": 0.5961608113002536, "grad_norm": 1.3055082683300485, "learning_rate": 7.4018963205962515e-06, "loss": 0.6755, "step": 16460 }, { "epoch": 0.5961970300615719, "grad_norm": 1.1414828222407947, "learning_rate": 7.400763554305729e-06, "loss": 0.7163, "step": 16461 }, { "epoch": 0.5962332488228903, "grad_norm": 1.302638921143114, "learning_rate": 7.399630823782986e-06, "loss": 0.657, "step": 16462 }, { "epoch": 0.5962694675842086, "grad_norm": 1.3141730316193576, "learning_rate": 7.398498129043605e-06, "loss": 0.7241, "step": 16463 }, { "epoch": 0.5963056863455269, "grad_norm": 1.2738338106818703, "learning_rate": 7.39736547010318e-06, "loss": 0.6971, "step": 16464 }, { "epoch": 0.5963419051068454, "grad_norm": 0.9099223442361783, "learning_rate": 7.396232846977289e-06, "loss": 0.7081, "step": 16465 }, { "epoch": 0.5963781238681637, "grad_norm": 1.3029047991789129, "learning_rate": 7.395100259681523e-06, "loss": 0.7103, "step": 16466 }, { "epoch": 0.5964143426294821, "grad_norm": 1.0153243007738948, "learning_rate": 7.393967708231466e-06, "loss": 0.6654, "step": 16467 }, { "epoch": 0.5964505613908004, "grad_norm": 1.3386793276087072, "learning_rate": 7.392835192642702e-06, "loss": 0.632, "step": 16468 }, { "epoch": 0.5964867801521188, "grad_norm": 1.073048689474947, "learning_rate": 7.39170271293082e-06, "loss": 0.7525, "step": 16469 }, { "epoch": 0.5965229989134372, "grad_norm": 1.034825536412235, "learning_rate": 7.390570269111399e-06, "loss": 0.7143, "step": 16470 }, { "epoch": 0.5965592176747555, "grad_norm": 1.3960197903041918, "learning_rate": 7.389437861200024e-06, "loss": 0.7044, "step": 16471 }, { "epoch": 0.5965954364360739, "grad_norm": 1.4592762863154327, "learning_rate": 7.3883054892122774e-06, "loss": 0.7519, "step": 16472 }, { "epoch": 0.5966316551973923, "grad_norm": 1.4199243142790483, "learning_rate": 7.387173153163742e-06, "loss": 0.673, "step": 16473 }, { "epoch": 0.5966678739587106, "grad_norm": 1.4937005594857142, "learning_rate": 7.3860408530700004e-06, "loss": 0.7957, "step": 16474 }, { "epoch": 0.596704092720029, "grad_norm": 1.1568329920466376, "learning_rate": 7.384908588946634e-06, "loss": 0.7044, "step": 16475 }, { "epoch": 0.5967403114813473, "grad_norm": 1.2553181186390745, "learning_rate": 7.3837763608092215e-06, "loss": 0.6238, "step": 16476 }, { "epoch": 0.5967765302426657, "grad_norm": 1.3057657172290411, "learning_rate": 7.382644168673345e-06, "loss": 0.6516, "step": 16477 }, { "epoch": 0.5968127490039841, "grad_norm": 1.3279855458590744, "learning_rate": 7.381512012554587e-06, "loss": 0.7068, "step": 16478 }, { "epoch": 0.5968489677653024, "grad_norm": 1.3129083389907732, "learning_rate": 7.380379892468522e-06, "loss": 0.6549, "step": 16479 }, { "epoch": 0.5968851865266208, "grad_norm": 1.2825623005171192, "learning_rate": 7.379247808430734e-06, "loss": 0.6702, "step": 16480 }, { "epoch": 0.5969214052879391, "grad_norm": 1.619873754045552, "learning_rate": 7.3781157604567965e-06, "loss": 0.6999, "step": 16481 }, { "epoch": 0.5969576240492576, "grad_norm": 1.2728839062966637, "learning_rate": 7.3769837485622916e-06, "loss": 0.7149, "step": 16482 }, { "epoch": 0.5969938428105759, "grad_norm": 1.0878375418093658, "learning_rate": 7.375851772762795e-06, "loss": 0.7457, "step": 16483 }, { "epoch": 0.5970300615718942, "grad_norm": 1.487144613136576, "learning_rate": 7.3747198330738834e-06, "loss": 0.7221, "step": 16484 }, { "epoch": 0.5970662803332126, "grad_norm": 1.3993001659919648, "learning_rate": 7.3735879295111325e-06, "loss": 0.7019, "step": 16485 }, { "epoch": 0.597102499094531, "grad_norm": 1.2575757428772498, "learning_rate": 7.3724560620901194e-06, "loss": 0.6588, "step": 16486 }, { "epoch": 0.5971387178558494, "grad_norm": 1.0906813168186418, "learning_rate": 7.371324230826422e-06, "loss": 0.7036, "step": 16487 }, { "epoch": 0.5971749366171677, "grad_norm": 1.4683043902190287, "learning_rate": 7.37019243573561e-06, "loss": 0.6653, "step": 16488 }, { "epoch": 0.597211155378486, "grad_norm": 1.0844335193648438, "learning_rate": 7.369060676833264e-06, "loss": 0.7477, "step": 16489 }, { "epoch": 0.5972473741398044, "grad_norm": 1.0345683656914182, "learning_rate": 7.367928954134953e-06, "loss": 0.7025, "step": 16490 }, { "epoch": 0.5972835929011228, "grad_norm": 1.336838233836991, "learning_rate": 7.366797267656254e-06, "loss": 0.7128, "step": 16491 }, { "epoch": 0.5973198116624412, "grad_norm": 1.4667802776713235, "learning_rate": 7.365665617412736e-06, "loss": 0.7025, "step": 16492 }, { "epoch": 0.5973560304237595, "grad_norm": 1.332943856797564, "learning_rate": 7.364534003419976e-06, "loss": 0.726, "step": 16493 }, { "epoch": 0.5973922491850778, "grad_norm": 1.3924474313782795, "learning_rate": 7.363402425693542e-06, "loss": 0.7095, "step": 16494 }, { "epoch": 0.5974284679463963, "grad_norm": 1.3208532918761886, "learning_rate": 7.3622708842490095e-06, "loss": 0.6743, "step": 16495 }, { "epoch": 0.5974646867077146, "grad_norm": 1.4024385895709746, "learning_rate": 7.361139379101945e-06, "loss": 0.6404, "step": 16496 }, { "epoch": 0.597500905469033, "grad_norm": 1.034232416285626, "learning_rate": 7.3600079102679215e-06, "loss": 0.7256, "step": 16497 }, { "epoch": 0.5975371242303513, "grad_norm": 1.1010820322964816, "learning_rate": 7.35887647776251e-06, "loss": 0.7289, "step": 16498 }, { "epoch": 0.5975733429916696, "grad_norm": 1.3436382103040874, "learning_rate": 7.357745081601278e-06, "loss": 0.6868, "step": 16499 }, { "epoch": 0.5976095617529881, "grad_norm": 1.057688287397292, "learning_rate": 7.356613721799797e-06, "loss": 0.7884, "step": 16500 }, { "epoch": 0.5976457805143064, "grad_norm": 1.5382706497935337, "learning_rate": 7.355482398373631e-06, "loss": 0.7445, "step": 16501 }, { "epoch": 0.5976819992756248, "grad_norm": 3.211041077569825, "learning_rate": 7.354351111338354e-06, "loss": 0.7315, "step": 16502 }, { "epoch": 0.5977182180369431, "grad_norm": 1.4407002534367628, "learning_rate": 7.353219860709529e-06, "loss": 0.6842, "step": 16503 }, { "epoch": 0.5977544367982615, "grad_norm": 1.456209957995641, "learning_rate": 7.352088646502724e-06, "loss": 0.7246, "step": 16504 }, { "epoch": 0.5977906555595799, "grad_norm": 1.4157910977955612, "learning_rate": 7.350957468733506e-06, "loss": 0.7118, "step": 16505 }, { "epoch": 0.5978268743208982, "grad_norm": 1.2106952909533812, "learning_rate": 7.349826327417442e-06, "loss": 0.6382, "step": 16506 }, { "epoch": 0.5978630930822166, "grad_norm": 1.469949446852187, "learning_rate": 7.348695222570095e-06, "loss": 0.7091, "step": 16507 }, { "epoch": 0.597899311843535, "grad_norm": 1.3779934754128507, "learning_rate": 7.34756415420703e-06, "loss": 0.6984, "step": 16508 }, { "epoch": 0.5979355306048533, "grad_norm": 1.277124428366628, "learning_rate": 7.346433122343817e-06, "loss": 0.6621, "step": 16509 }, { "epoch": 0.5979717493661717, "grad_norm": 1.3574053907422645, "learning_rate": 7.345302126996013e-06, "loss": 0.7511, "step": 16510 }, { "epoch": 0.59800796812749, "grad_norm": 0.980359723243395, "learning_rate": 7.344171168179185e-06, "loss": 0.7274, "step": 16511 }, { "epoch": 0.5980441868888084, "grad_norm": 1.5625601160882583, "learning_rate": 7.343040245908894e-06, "loss": 0.6759, "step": 16512 }, { "epoch": 0.5980804056501268, "grad_norm": 2.1748443055195006, "learning_rate": 7.341909360200706e-06, "loss": 0.7877, "step": 16513 }, { "epoch": 0.5981166244114451, "grad_norm": 1.4428794215134666, "learning_rate": 7.340778511070179e-06, "loss": 0.7511, "step": 16514 }, { "epoch": 0.5981528431727635, "grad_norm": 1.2571327308093434, "learning_rate": 7.3396476985328795e-06, "loss": 0.7367, "step": 16515 }, { "epoch": 0.5981890619340818, "grad_norm": 1.4897300146704924, "learning_rate": 7.338516922604362e-06, "loss": 0.7455, "step": 16516 }, { "epoch": 0.5982252806954003, "grad_norm": 1.2470724782424558, "learning_rate": 7.337386183300189e-06, "loss": 0.7277, "step": 16517 }, { "epoch": 0.5982614994567186, "grad_norm": 1.2227382952306953, "learning_rate": 7.336255480635925e-06, "loss": 0.6846, "step": 16518 }, { "epoch": 0.5982977182180369, "grad_norm": 1.4762683649872947, "learning_rate": 7.335124814627124e-06, "loss": 0.7063, "step": 16519 }, { "epoch": 0.5983339369793553, "grad_norm": 1.3881075956008384, "learning_rate": 7.333994185289351e-06, "loss": 0.7427, "step": 16520 }, { "epoch": 0.5983701557406736, "grad_norm": 1.1727943620859782, "learning_rate": 7.332863592638155e-06, "loss": 0.7442, "step": 16521 }, { "epoch": 0.5984063745019921, "grad_norm": 1.35621600619357, "learning_rate": 7.331733036689104e-06, "loss": 0.7226, "step": 16522 }, { "epoch": 0.5984425932633104, "grad_norm": 1.6744573205103204, "learning_rate": 7.330602517457748e-06, "loss": 0.8344, "step": 16523 }, { "epoch": 0.5984788120246287, "grad_norm": 1.2167255663336385, "learning_rate": 7.329472034959649e-06, "loss": 0.7255, "step": 16524 }, { "epoch": 0.5985150307859471, "grad_norm": 1.2641340122635787, "learning_rate": 7.3283415892103596e-06, "loss": 0.5983, "step": 16525 }, { "epoch": 0.5985512495472655, "grad_norm": 1.370397694015299, "learning_rate": 7.32721118022544e-06, "loss": 0.6856, "step": 16526 }, { "epoch": 0.5985874683085839, "grad_norm": 1.3076226303408949, "learning_rate": 7.32608080802044e-06, "loss": 0.7304, "step": 16527 }, { "epoch": 0.5986236870699022, "grad_norm": 1.2357857676129371, "learning_rate": 7.3249504726109175e-06, "loss": 0.6996, "step": 16528 }, { "epoch": 0.5986599058312205, "grad_norm": 1.3714935217139215, "learning_rate": 7.32382017401243e-06, "loss": 0.7072, "step": 16529 }, { "epoch": 0.598696124592539, "grad_norm": 1.0618947642153023, "learning_rate": 7.322689912240526e-06, "loss": 0.7295, "step": 16530 }, { "epoch": 0.5987323433538573, "grad_norm": 1.3571509989784645, "learning_rate": 7.321559687310763e-06, "loss": 0.6993, "step": 16531 }, { "epoch": 0.5987685621151757, "grad_norm": 1.322793037649648, "learning_rate": 7.32042949923869e-06, "loss": 0.6711, "step": 16532 }, { "epoch": 0.598804780876494, "grad_norm": 1.2583728520955373, "learning_rate": 7.319299348039864e-06, "loss": 0.6626, "step": 16533 }, { "epoch": 0.5988409996378123, "grad_norm": 1.460342007282175, "learning_rate": 7.3181692337298326e-06, "loss": 0.7762, "step": 16534 }, { "epoch": 0.5988772183991308, "grad_norm": 1.5082220697287576, "learning_rate": 7.317039156324152e-06, "loss": 0.6722, "step": 16535 }, { "epoch": 0.5989134371604491, "grad_norm": 1.8317111778867463, "learning_rate": 7.315909115838367e-06, "loss": 0.6899, "step": 16536 }, { "epoch": 0.5989496559217675, "grad_norm": 1.4557258444414347, "learning_rate": 7.314779112288033e-06, "loss": 0.7452, "step": 16537 }, { "epoch": 0.5989858746830858, "grad_norm": 1.5301612564576847, "learning_rate": 7.313649145688695e-06, "loss": 0.6961, "step": 16538 }, { "epoch": 0.5990220934444042, "grad_norm": 1.20624418739045, "learning_rate": 7.312519216055907e-06, "loss": 0.7085, "step": 16539 }, { "epoch": 0.5990583122057226, "grad_norm": 1.4823214819867987, "learning_rate": 7.311389323405218e-06, "loss": 0.6792, "step": 16540 }, { "epoch": 0.5990945309670409, "grad_norm": 1.3921584818552684, "learning_rate": 7.3102594677521735e-06, "loss": 0.7492, "step": 16541 }, { "epoch": 0.5991307497283593, "grad_norm": 1.2368958052771528, "learning_rate": 7.3091296491123194e-06, "loss": 0.6267, "step": 16542 }, { "epoch": 0.5991669684896777, "grad_norm": 1.3179665305147779, "learning_rate": 7.307999867501209e-06, "loss": 0.6698, "step": 16543 }, { "epoch": 0.599203187250996, "grad_norm": 1.355837336234446, "learning_rate": 7.306870122934383e-06, "loss": 0.7045, "step": 16544 }, { "epoch": 0.5992394060123144, "grad_norm": 0.9667936654350682, "learning_rate": 7.3057404154273925e-06, "loss": 0.7524, "step": 16545 }, { "epoch": 0.5992756247736327, "grad_norm": 1.2144243138780113, "learning_rate": 7.304610744995779e-06, "loss": 0.6297, "step": 16546 }, { "epoch": 0.5993118435349511, "grad_norm": 0.9661003298456727, "learning_rate": 7.303481111655093e-06, "loss": 0.7228, "step": 16547 }, { "epoch": 0.5993480622962695, "grad_norm": 1.5101818877844737, "learning_rate": 7.302351515420874e-06, "loss": 0.6638, "step": 16548 }, { "epoch": 0.5993842810575878, "grad_norm": 1.220300824630594, "learning_rate": 7.301221956308667e-06, "loss": 0.7256, "step": 16549 }, { "epoch": 0.5994204998189062, "grad_norm": 1.372433319807364, "learning_rate": 7.300092434334021e-06, "loss": 0.7002, "step": 16550 }, { "epoch": 0.5994567185802245, "grad_norm": 1.413145238716862, "learning_rate": 7.298962949512474e-06, "loss": 0.7319, "step": 16551 }, { "epoch": 0.599492937341543, "grad_norm": 1.199009260206267, "learning_rate": 7.29783350185957e-06, "loss": 0.7045, "step": 16552 }, { "epoch": 0.5995291561028613, "grad_norm": 1.389377157361496, "learning_rate": 7.29670409139085e-06, "loss": 0.6368, "step": 16553 }, { "epoch": 0.5995653748641796, "grad_norm": 1.3890060661146038, "learning_rate": 7.295574718121859e-06, "loss": 0.6867, "step": 16554 }, { "epoch": 0.599601593625498, "grad_norm": 1.223308530156874, "learning_rate": 7.294445382068136e-06, "loss": 0.6757, "step": 16555 }, { "epoch": 0.5996378123868163, "grad_norm": 1.3583469758203113, "learning_rate": 7.293316083245221e-06, "loss": 0.7525, "step": 16556 }, { "epoch": 0.5996740311481348, "grad_norm": 1.3531125818590155, "learning_rate": 7.292186821668655e-06, "loss": 0.7332, "step": 16557 }, { "epoch": 0.5997102499094531, "grad_norm": 1.4296488001020478, "learning_rate": 7.291057597353979e-06, "loss": 0.7383, "step": 16558 }, { "epoch": 0.5997464686707714, "grad_norm": 1.2530280912294642, "learning_rate": 7.289928410316729e-06, "loss": 0.6837, "step": 16559 }, { "epoch": 0.5997826874320898, "grad_norm": 1.2718205820504784, "learning_rate": 7.288799260572446e-06, "loss": 0.7076, "step": 16560 }, { "epoch": 0.5998189061934082, "grad_norm": 1.0679326334045862, "learning_rate": 7.287670148136667e-06, "loss": 0.7194, "step": 16561 }, { "epoch": 0.5998551249547266, "grad_norm": 1.319634117269182, "learning_rate": 7.2865410730249306e-06, "loss": 0.7066, "step": 16562 }, { "epoch": 0.5998913437160449, "grad_norm": 1.046821235146397, "learning_rate": 7.285412035252774e-06, "loss": 0.7662, "step": 16563 }, { "epoch": 0.5999275624773632, "grad_norm": 1.339267960796874, "learning_rate": 7.284283034835731e-06, "loss": 0.7652, "step": 16564 }, { "epoch": 0.5999637812386817, "grad_norm": 1.3164606068988387, "learning_rate": 7.283154071789342e-06, "loss": 0.7548, "step": 16565 }, { "epoch": 0.6, "grad_norm": 1.5476428447760964, "learning_rate": 7.282025146129139e-06, "loss": 0.7615, "step": 16566 }, { "epoch": 0.6000362187613184, "grad_norm": 1.3423373197297843, "learning_rate": 7.2808962578706596e-06, "loss": 0.7231, "step": 16567 }, { "epoch": 0.6000724375226367, "grad_norm": 1.3641983468456012, "learning_rate": 7.279767407029435e-06, "loss": 0.716, "step": 16568 }, { "epoch": 0.600108656283955, "grad_norm": 1.3398410738704192, "learning_rate": 7.278638593621003e-06, "loss": 0.6676, "step": 16569 }, { "epoch": 0.6001448750452735, "grad_norm": 1.244881807406046, "learning_rate": 7.277509817660893e-06, "loss": 0.7228, "step": 16570 }, { "epoch": 0.6001810938065918, "grad_norm": 1.005236815058235, "learning_rate": 7.276381079164639e-06, "loss": 0.6605, "step": 16571 }, { "epoch": 0.6002173125679102, "grad_norm": 1.0174950543568348, "learning_rate": 7.275252378147778e-06, "loss": 0.7593, "step": 16572 }, { "epoch": 0.6002535313292285, "grad_norm": 1.2643482617890058, "learning_rate": 7.274123714625836e-06, "loss": 0.6591, "step": 16573 }, { "epoch": 0.6002897500905469, "grad_norm": 1.3143429745438049, "learning_rate": 7.272995088614348e-06, "loss": 0.7479, "step": 16574 }, { "epoch": 0.6003259688518653, "grad_norm": 1.4037314166539954, "learning_rate": 7.271866500128844e-06, "loss": 0.7246, "step": 16575 }, { "epoch": 0.6003621876131836, "grad_norm": 1.3328082721303267, "learning_rate": 7.270737949184854e-06, "loss": 0.6954, "step": 16576 }, { "epoch": 0.600398406374502, "grad_norm": 1.323696260562556, "learning_rate": 7.269609435797906e-06, "loss": 0.7442, "step": 16577 }, { "epoch": 0.6004346251358204, "grad_norm": 1.373332017074555, "learning_rate": 7.268480959983535e-06, "loss": 0.6341, "step": 16578 }, { "epoch": 0.6004708438971387, "grad_norm": 1.3304461181583889, "learning_rate": 7.267352521757263e-06, "loss": 0.7623, "step": 16579 }, { "epoch": 0.6005070626584571, "grad_norm": 1.3640262964366667, "learning_rate": 7.266224121134622e-06, "loss": 0.7429, "step": 16580 }, { "epoch": 0.6005432814197754, "grad_norm": 1.4998214498686762, "learning_rate": 7.2650957581311395e-06, "loss": 0.6959, "step": 16581 }, { "epoch": 0.6005795001810939, "grad_norm": 1.1877926635239986, "learning_rate": 7.263967432762342e-06, "loss": 0.6459, "step": 16582 }, { "epoch": 0.6006157189424122, "grad_norm": 0.8767698525635413, "learning_rate": 7.262839145043759e-06, "loss": 0.6686, "step": 16583 }, { "epoch": 0.6006519377037305, "grad_norm": 1.2922406444264865, "learning_rate": 7.261710894990912e-06, "loss": 0.7062, "step": 16584 }, { "epoch": 0.6006881564650489, "grad_norm": 1.3643529199367943, "learning_rate": 7.26058268261933e-06, "loss": 0.716, "step": 16585 }, { "epoch": 0.6007243752263672, "grad_norm": 1.357846850140685, "learning_rate": 7.2594545079445366e-06, "loss": 0.6865, "step": 16586 }, { "epoch": 0.6007605939876857, "grad_norm": 1.2975860270964044, "learning_rate": 7.2583263709820595e-06, "loss": 0.7813, "step": 16587 }, { "epoch": 0.600796812749004, "grad_norm": 1.1906987763122636, "learning_rate": 7.257198271747418e-06, "loss": 0.7319, "step": 16588 }, { "epoch": 0.6008330315103223, "grad_norm": 1.2701171359373524, "learning_rate": 7.256070210256141e-06, "loss": 0.621, "step": 16589 }, { "epoch": 0.6008692502716407, "grad_norm": 1.095066884450999, "learning_rate": 7.254942186523746e-06, "loss": 0.6809, "step": 16590 }, { "epoch": 0.600905469032959, "grad_norm": 1.2667901325958022, "learning_rate": 7.2538142005657585e-06, "loss": 0.7073, "step": 16591 }, { "epoch": 0.6009416877942775, "grad_norm": 1.320447256050771, "learning_rate": 7.252686252397703e-06, "loss": 0.7364, "step": 16592 }, { "epoch": 0.6009779065555958, "grad_norm": 1.2856459546977739, "learning_rate": 7.251558342035097e-06, "loss": 0.6723, "step": 16593 }, { "epoch": 0.6010141253169141, "grad_norm": 1.3093059139188945, "learning_rate": 7.250430469493465e-06, "loss": 0.7326, "step": 16594 }, { "epoch": 0.6010503440782325, "grad_norm": 1.218021610732752, "learning_rate": 7.249302634788324e-06, "loss": 0.6537, "step": 16595 }, { "epoch": 0.6010865628395509, "grad_norm": 1.0303833526583115, "learning_rate": 7.248174837935197e-06, "loss": 0.7722, "step": 16596 }, { "epoch": 0.6011227816008693, "grad_norm": 1.0106471804872452, "learning_rate": 7.2470470789496e-06, "loss": 0.7638, "step": 16597 }, { "epoch": 0.6011590003621876, "grad_norm": 1.2510588322581317, "learning_rate": 7.245919357847057e-06, "loss": 0.7605, "step": 16598 }, { "epoch": 0.6011952191235059, "grad_norm": 1.2946296192359157, "learning_rate": 7.2447916746430815e-06, "loss": 0.7226, "step": 16599 }, { "epoch": 0.6012314378848244, "grad_norm": 1.4361754552849164, "learning_rate": 7.243664029353195e-06, "loss": 0.6869, "step": 16600 }, { "epoch": 0.6012676566461427, "grad_norm": 1.1751226003344215, "learning_rate": 7.242536421992912e-06, "loss": 0.7002, "step": 16601 }, { "epoch": 0.6013038754074611, "grad_norm": 1.2738147967248696, "learning_rate": 7.2414088525777495e-06, "loss": 0.648, "step": 16602 }, { "epoch": 0.6013400941687794, "grad_norm": 1.445502087040925, "learning_rate": 7.240281321123228e-06, "loss": 0.7491, "step": 16603 }, { "epoch": 0.6013763129300977, "grad_norm": 1.218075587821346, "learning_rate": 7.2391538276448584e-06, "loss": 0.6824, "step": 16604 }, { "epoch": 0.6014125316914162, "grad_norm": 1.0238704392151234, "learning_rate": 7.23802637215816e-06, "loss": 0.7004, "step": 16605 }, { "epoch": 0.6014487504527345, "grad_norm": 0.9451336818680353, "learning_rate": 7.236898954678644e-06, "loss": 0.7346, "step": 16606 }, { "epoch": 0.6014849692140529, "grad_norm": 1.3256278727080797, "learning_rate": 7.235771575221828e-06, "loss": 0.7837, "step": 16607 }, { "epoch": 0.6015211879753712, "grad_norm": 1.2077416767294396, "learning_rate": 7.234644233803221e-06, "loss": 0.6723, "step": 16608 }, { "epoch": 0.6015574067366896, "grad_norm": 1.3754621956233692, "learning_rate": 7.233516930438343e-06, "loss": 0.6317, "step": 16609 }, { "epoch": 0.601593625498008, "grad_norm": 1.4235558596184537, "learning_rate": 7.232389665142699e-06, "loss": 0.7034, "step": 16610 }, { "epoch": 0.6016298442593263, "grad_norm": 1.285000253908804, "learning_rate": 7.231262437931806e-06, "loss": 0.645, "step": 16611 }, { "epoch": 0.6016660630206447, "grad_norm": 1.4999279263188205, "learning_rate": 7.230135248821178e-06, "loss": 0.8343, "step": 16612 }, { "epoch": 0.601702281781963, "grad_norm": 1.3493549035251269, "learning_rate": 7.229008097826318e-06, "loss": 0.7204, "step": 16613 }, { "epoch": 0.6017385005432814, "grad_norm": 1.0765134043166995, "learning_rate": 7.227880984962745e-06, "loss": 0.6889, "step": 16614 }, { "epoch": 0.6017747193045998, "grad_norm": 1.8644063220012985, "learning_rate": 7.226753910245963e-06, "loss": 0.8057, "step": 16615 }, { "epoch": 0.6018109380659181, "grad_norm": 1.294909282266367, "learning_rate": 7.225626873691487e-06, "loss": 0.6719, "step": 16616 }, { "epoch": 0.6018471568272366, "grad_norm": 1.2569975774959985, "learning_rate": 7.22449987531482e-06, "loss": 0.6979, "step": 16617 }, { "epoch": 0.6018833755885549, "grad_norm": 1.1055735470581864, "learning_rate": 7.223372915131476e-06, "loss": 0.7197, "step": 16618 }, { "epoch": 0.6019195943498732, "grad_norm": 1.37184993290864, "learning_rate": 7.222245993156958e-06, "loss": 0.7748, "step": 16619 }, { "epoch": 0.6019558131111916, "grad_norm": 1.266954305231547, "learning_rate": 7.221119109406778e-06, "loss": 0.7099, "step": 16620 }, { "epoch": 0.6019920318725099, "grad_norm": 1.309034170224178, "learning_rate": 7.219992263896438e-06, "loss": 0.7393, "step": 16621 }, { "epoch": 0.6020282506338284, "grad_norm": 1.3290796102915379, "learning_rate": 7.218865456641448e-06, "loss": 0.7113, "step": 16622 }, { "epoch": 0.6020644693951467, "grad_norm": 1.5263070175797544, "learning_rate": 7.217738687657316e-06, "loss": 0.7207, "step": 16623 }, { "epoch": 0.602100688156465, "grad_norm": 1.3456683500417004, "learning_rate": 7.21661195695954e-06, "loss": 0.697, "step": 16624 }, { "epoch": 0.6021369069177834, "grad_norm": 1.4220674968723568, "learning_rate": 7.215485264563633e-06, "loss": 0.6901, "step": 16625 }, { "epoch": 0.6021731256791018, "grad_norm": 1.4324545166607385, "learning_rate": 7.2143586104850936e-06, "loss": 0.7124, "step": 16626 }, { "epoch": 0.6022093444404202, "grad_norm": 1.2153282755528954, "learning_rate": 7.213231994739429e-06, "loss": 0.6772, "step": 16627 }, { "epoch": 0.6022455632017385, "grad_norm": 1.0097011366524962, "learning_rate": 7.212105417342139e-06, "loss": 0.7339, "step": 16628 }, { "epoch": 0.6022817819630568, "grad_norm": 1.3062070161304336, "learning_rate": 7.21097887830873e-06, "loss": 0.6954, "step": 16629 }, { "epoch": 0.6023180007243752, "grad_norm": 1.4458898705601373, "learning_rate": 7.2098523776547e-06, "loss": 0.776, "step": 16630 }, { "epoch": 0.6023542194856936, "grad_norm": 1.5868360811293603, "learning_rate": 7.2087259153955544e-06, "loss": 0.6183, "step": 16631 }, { "epoch": 0.602390438247012, "grad_norm": 1.2396686525260776, "learning_rate": 7.207599491546791e-06, "loss": 0.6761, "step": 16632 }, { "epoch": 0.6024266570083303, "grad_norm": 1.3698996080873707, "learning_rate": 7.206473106123913e-06, "loss": 0.6648, "step": 16633 }, { "epoch": 0.6024628757696486, "grad_norm": 1.45891853615481, "learning_rate": 7.205346759142422e-06, "loss": 0.7648, "step": 16634 }, { "epoch": 0.6024990945309671, "grad_norm": 1.3044502833570135, "learning_rate": 7.204220450617811e-06, "loss": 0.6963, "step": 16635 }, { "epoch": 0.6025353132922854, "grad_norm": 1.359727983575633, "learning_rate": 7.203094180565589e-06, "loss": 0.7186, "step": 16636 }, { "epoch": 0.6025715320536038, "grad_norm": 1.2752328836471698, "learning_rate": 7.201967949001243e-06, "loss": 0.6357, "step": 16637 }, { "epoch": 0.6026077508149221, "grad_norm": 1.549711413301867, "learning_rate": 7.20084175594028e-06, "loss": 0.6881, "step": 16638 }, { "epoch": 0.6026439695762404, "grad_norm": 1.75241792202868, "learning_rate": 7.199715601398192e-06, "loss": 0.6937, "step": 16639 }, { "epoch": 0.6026801883375589, "grad_norm": 1.3613536536779869, "learning_rate": 7.1985894853904795e-06, "loss": 0.7644, "step": 16640 }, { "epoch": 0.6027164070988772, "grad_norm": 1.2802245896751634, "learning_rate": 7.197463407932636e-06, "loss": 0.6624, "step": 16641 }, { "epoch": 0.6027526258601956, "grad_norm": 1.0642248680533517, "learning_rate": 7.196337369040158e-06, "loss": 0.7346, "step": 16642 }, { "epoch": 0.6027888446215139, "grad_norm": 1.289155769586306, "learning_rate": 7.195211368728543e-06, "loss": 0.7201, "step": 16643 }, { "epoch": 0.6028250633828323, "grad_norm": 1.4163119078482123, "learning_rate": 7.194085407013283e-06, "loss": 0.6849, "step": 16644 }, { "epoch": 0.6028612821441507, "grad_norm": 1.3068779452235197, "learning_rate": 7.1929594839098736e-06, "loss": 0.7257, "step": 16645 }, { "epoch": 0.602897500905469, "grad_norm": 1.4172144530823503, "learning_rate": 7.191833599433809e-06, "loss": 0.7244, "step": 16646 }, { "epoch": 0.6029337196667874, "grad_norm": 1.3362604979112964, "learning_rate": 7.190707753600581e-06, "loss": 0.6186, "step": 16647 }, { "epoch": 0.6029699384281058, "grad_norm": 1.3372203874432902, "learning_rate": 7.189581946425681e-06, "loss": 0.7677, "step": 16648 }, { "epoch": 0.6030061571894241, "grad_norm": 1.362510792717928, "learning_rate": 7.1884561779246055e-06, "loss": 0.7335, "step": 16649 }, { "epoch": 0.6030423759507425, "grad_norm": 1.3514106822329854, "learning_rate": 7.187330448112841e-06, "loss": 0.6713, "step": 16650 }, { "epoch": 0.6030785947120608, "grad_norm": 1.3146953423361138, "learning_rate": 7.186204757005884e-06, "loss": 0.6765, "step": 16651 }, { "epoch": 0.6031148134733793, "grad_norm": 1.0375672673936365, "learning_rate": 7.185079104619219e-06, "loss": 0.7263, "step": 16652 }, { "epoch": 0.6031510322346976, "grad_norm": 1.3506768100202173, "learning_rate": 7.18395349096834e-06, "loss": 0.7888, "step": 16653 }, { "epoch": 0.6031872509960159, "grad_norm": 1.392699033081522, "learning_rate": 7.182827916068736e-06, "loss": 0.6707, "step": 16654 }, { "epoch": 0.6032234697573343, "grad_norm": 1.422902453892864, "learning_rate": 7.181702379935894e-06, "loss": 0.7187, "step": 16655 }, { "epoch": 0.6032596885186526, "grad_norm": 1.5076675691277557, "learning_rate": 7.180576882585306e-06, "loss": 0.72, "step": 16656 }, { "epoch": 0.6032959072799711, "grad_norm": 1.0416401512650546, "learning_rate": 7.179451424032455e-06, "loss": 0.7712, "step": 16657 }, { "epoch": 0.6033321260412894, "grad_norm": 1.2338257741727567, "learning_rate": 7.178326004292832e-06, "loss": 0.7087, "step": 16658 }, { "epoch": 0.6033683448026077, "grad_norm": 1.4017411693293482, "learning_rate": 7.177200623381921e-06, "loss": 0.7555, "step": 16659 }, { "epoch": 0.6034045635639261, "grad_norm": 1.214821095424037, "learning_rate": 7.1760752813152125e-06, "loss": 0.7433, "step": 16660 }, { "epoch": 0.6034407823252445, "grad_norm": 1.4101861649401974, "learning_rate": 7.174949978108188e-06, "loss": 0.7725, "step": 16661 }, { "epoch": 0.6034770010865629, "grad_norm": 1.3141474552622232, "learning_rate": 7.173824713776334e-06, "loss": 0.7287, "step": 16662 }, { "epoch": 0.6035132198478812, "grad_norm": 1.52158464889737, "learning_rate": 7.172699488335134e-06, "loss": 0.7765, "step": 16663 }, { "epoch": 0.6035494386091995, "grad_norm": 1.1702880472414203, "learning_rate": 7.171574301800074e-06, "loss": 0.7048, "step": 16664 }, { "epoch": 0.603585657370518, "grad_norm": 1.4724597907374963, "learning_rate": 7.170449154186638e-06, "loss": 0.7834, "step": 16665 }, { "epoch": 0.6036218761318363, "grad_norm": 1.3398329398373205, "learning_rate": 7.169324045510306e-06, "loss": 0.7136, "step": 16666 }, { "epoch": 0.6036580948931547, "grad_norm": 1.211716374512391, "learning_rate": 7.168198975786565e-06, "loss": 0.7025, "step": 16667 }, { "epoch": 0.603694313654473, "grad_norm": 1.4404647111675735, "learning_rate": 7.167073945030892e-06, "loss": 0.7184, "step": 16668 }, { "epoch": 0.6037305324157913, "grad_norm": 1.3839568303613667, "learning_rate": 7.165948953258773e-06, "loss": 0.7821, "step": 16669 }, { "epoch": 0.6037667511771098, "grad_norm": 1.2949540211929345, "learning_rate": 7.164824000485685e-06, "loss": 0.6976, "step": 16670 }, { "epoch": 0.6038029699384281, "grad_norm": 1.2899110869200914, "learning_rate": 7.163699086727112e-06, "loss": 0.6547, "step": 16671 }, { "epoch": 0.6038391886997465, "grad_norm": 1.346773232163377, "learning_rate": 7.162574211998529e-06, "loss": 0.7151, "step": 16672 }, { "epoch": 0.6038754074610648, "grad_norm": 1.252079880274141, "learning_rate": 7.1614493763154166e-06, "loss": 0.693, "step": 16673 }, { "epoch": 0.6039116262223831, "grad_norm": 1.522432281733414, "learning_rate": 7.160324579693259e-06, "loss": 0.7371, "step": 16674 }, { "epoch": 0.6039478449837016, "grad_norm": 1.3436866172823736, "learning_rate": 7.159199822147529e-06, "loss": 0.7174, "step": 16675 }, { "epoch": 0.6039840637450199, "grad_norm": 1.0129020200843093, "learning_rate": 7.158075103693706e-06, "loss": 0.6508, "step": 16676 }, { "epoch": 0.6040202825063383, "grad_norm": 1.2699707994274387, "learning_rate": 7.156950424347264e-06, "loss": 0.7015, "step": 16677 }, { "epoch": 0.6040565012676566, "grad_norm": 1.3037463579831041, "learning_rate": 7.155825784123687e-06, "loss": 0.7576, "step": 16678 }, { "epoch": 0.604092720028975, "grad_norm": 1.3142412349309052, "learning_rate": 7.154701183038442e-06, "loss": 0.6937, "step": 16679 }, { "epoch": 0.6041289387902934, "grad_norm": 1.3883050214034256, "learning_rate": 7.153576621107013e-06, "loss": 0.7377, "step": 16680 }, { "epoch": 0.6041651575516117, "grad_norm": 1.0389791851638712, "learning_rate": 7.152452098344866e-06, "loss": 0.7087, "step": 16681 }, { "epoch": 0.6042013763129301, "grad_norm": 1.4829651361095255, "learning_rate": 7.151327614767484e-06, "loss": 0.6608, "step": 16682 }, { "epoch": 0.6042375950742485, "grad_norm": 0.9260921329133123, "learning_rate": 7.150203170390335e-06, "loss": 0.7454, "step": 16683 }, { "epoch": 0.6042738138355668, "grad_norm": 1.4256574511670645, "learning_rate": 7.149078765228894e-06, "loss": 0.7534, "step": 16684 }, { "epoch": 0.6043100325968852, "grad_norm": 1.3214480198035166, "learning_rate": 7.147954399298635e-06, "loss": 0.7805, "step": 16685 }, { "epoch": 0.6043462513582035, "grad_norm": 1.2687500677672494, "learning_rate": 7.1468300726150295e-06, "loss": 0.7146, "step": 16686 }, { "epoch": 0.604382470119522, "grad_norm": 1.1681806306759477, "learning_rate": 7.145705785193551e-06, "loss": 0.6935, "step": 16687 }, { "epoch": 0.6044186888808403, "grad_norm": 1.3457109287738838, "learning_rate": 7.144581537049666e-06, "loss": 0.7091, "step": 16688 }, { "epoch": 0.6044549076421586, "grad_norm": 1.4416589399144029, "learning_rate": 7.143457328198851e-06, "loss": 0.7381, "step": 16689 }, { "epoch": 0.604491126403477, "grad_norm": 1.2757356777157647, "learning_rate": 7.14233315865657e-06, "loss": 0.6657, "step": 16690 }, { "epoch": 0.6045273451647953, "grad_norm": 1.0090003716083717, "learning_rate": 7.1412090284382985e-06, "loss": 0.6561, "step": 16691 }, { "epoch": 0.6045635639261138, "grad_norm": 1.4281782224867234, "learning_rate": 7.1400849375595e-06, "loss": 0.7313, "step": 16692 }, { "epoch": 0.6045997826874321, "grad_norm": 1.2638695266011086, "learning_rate": 7.138960886035649e-06, "loss": 0.6755, "step": 16693 }, { "epoch": 0.6046360014487504, "grad_norm": 1.0155448968526586, "learning_rate": 7.137836873882207e-06, "loss": 0.7161, "step": 16694 }, { "epoch": 0.6046722202100688, "grad_norm": 1.0976107447285544, "learning_rate": 7.136712901114645e-06, "loss": 0.7114, "step": 16695 }, { "epoch": 0.6047084389713872, "grad_norm": 1.0357765646856016, "learning_rate": 7.135588967748432e-06, "loss": 0.7021, "step": 16696 }, { "epoch": 0.6047446577327056, "grad_norm": 1.4801253974958475, "learning_rate": 7.134465073799028e-06, "loss": 0.8088, "step": 16697 }, { "epoch": 0.6047808764940239, "grad_norm": 1.4486270053742503, "learning_rate": 7.133341219281907e-06, "loss": 0.7088, "step": 16698 }, { "epoch": 0.6048170952553422, "grad_norm": 1.389381753714831, "learning_rate": 7.132217404212525e-06, "loss": 0.7249, "step": 16699 }, { "epoch": 0.6048533140166606, "grad_norm": 1.0998408222630327, "learning_rate": 7.131093628606355e-06, "loss": 0.6734, "step": 16700 }, { "epoch": 0.604889532777979, "grad_norm": 1.0026433698167545, "learning_rate": 7.129969892478854e-06, "loss": 0.6753, "step": 16701 }, { "epoch": 0.6049257515392974, "grad_norm": 1.3368517921775396, "learning_rate": 7.128846195845493e-06, "loss": 0.7249, "step": 16702 }, { "epoch": 0.6049619703006157, "grad_norm": 1.3086165753551684, "learning_rate": 7.127722538721729e-06, "loss": 0.7152, "step": 16703 }, { "epoch": 0.604998189061934, "grad_norm": 1.226838475663596, "learning_rate": 7.126598921123025e-06, "loss": 0.6384, "step": 16704 }, { "epoch": 0.6050344078232525, "grad_norm": 1.3832230524212905, "learning_rate": 7.125475343064847e-06, "loss": 0.6883, "step": 16705 }, { "epoch": 0.6050706265845708, "grad_norm": 1.2520251130669304, "learning_rate": 7.1243518045626535e-06, "loss": 0.7405, "step": 16706 }, { "epoch": 0.6051068453458892, "grad_norm": 1.0663959359449238, "learning_rate": 7.123228305631906e-06, "loss": 0.737, "step": 16707 }, { "epoch": 0.6051430641072075, "grad_norm": 0.9614972189250502, "learning_rate": 7.122104846288065e-06, "loss": 0.6989, "step": 16708 }, { "epoch": 0.6051792828685258, "grad_norm": 0.9993256204413854, "learning_rate": 7.1209814265465915e-06, "loss": 0.7308, "step": 16709 }, { "epoch": 0.6052155016298443, "grad_norm": 1.0244519300116754, "learning_rate": 7.11985804642294e-06, "loss": 0.7087, "step": 16710 }, { "epoch": 0.6052517203911626, "grad_norm": 1.340600854288306, "learning_rate": 7.118734705932575e-06, "loss": 0.743, "step": 16711 }, { "epoch": 0.605287939152481, "grad_norm": 1.530435161041868, "learning_rate": 7.117611405090951e-06, "loss": 0.6575, "step": 16712 }, { "epoch": 0.6053241579137993, "grad_norm": 1.4080858538892667, "learning_rate": 7.116488143913528e-06, "loss": 0.6678, "step": 16713 }, { "epoch": 0.6053603766751177, "grad_norm": 1.0047191293325863, "learning_rate": 7.1153649224157595e-06, "loss": 0.7168, "step": 16714 }, { "epoch": 0.6053965954364361, "grad_norm": 1.4405966826798204, "learning_rate": 7.114241740613105e-06, "loss": 0.69, "step": 16715 }, { "epoch": 0.6054328141977544, "grad_norm": 1.0472401671088143, "learning_rate": 7.1131185985210214e-06, "loss": 0.7204, "step": 16716 }, { "epoch": 0.6054690329590728, "grad_norm": 1.0876209444949876, "learning_rate": 7.111995496154961e-06, "loss": 0.72, "step": 16717 }, { "epoch": 0.6055052517203912, "grad_norm": 1.2776066045923167, "learning_rate": 7.110872433530383e-06, "loss": 0.7259, "step": 16718 }, { "epoch": 0.6055414704817095, "grad_norm": 1.4382415069256873, "learning_rate": 7.109749410662734e-06, "loss": 0.7773, "step": 16719 }, { "epoch": 0.6055776892430279, "grad_norm": 1.065951089741975, "learning_rate": 7.108626427567478e-06, "loss": 0.7386, "step": 16720 }, { "epoch": 0.6056139080043462, "grad_norm": 1.3380775989277662, "learning_rate": 7.107503484260059e-06, "loss": 0.7057, "step": 16721 }, { "epoch": 0.6056501267656647, "grad_norm": 1.1752530738199116, "learning_rate": 7.106380580755936e-06, "loss": 0.6648, "step": 16722 }, { "epoch": 0.605686345526983, "grad_norm": 1.351279530919315, "learning_rate": 7.105257717070556e-06, "loss": 0.7101, "step": 16723 }, { "epoch": 0.6057225642883013, "grad_norm": 1.34377824676012, "learning_rate": 7.104134893219376e-06, "loss": 0.7189, "step": 16724 }, { "epoch": 0.6057587830496197, "grad_norm": 0.9679019596761875, "learning_rate": 7.1030121092178435e-06, "loss": 0.6832, "step": 16725 }, { "epoch": 0.605795001810938, "grad_norm": 1.3315695601557929, "learning_rate": 7.101889365081409e-06, "loss": 0.6821, "step": 16726 }, { "epoch": 0.6058312205722565, "grad_norm": 1.272145115002926, "learning_rate": 7.100766660825525e-06, "loss": 0.6675, "step": 16727 }, { "epoch": 0.6058674393335748, "grad_norm": 1.2322210051465206, "learning_rate": 7.099643996465637e-06, "loss": 0.6887, "step": 16728 }, { "epoch": 0.6059036580948931, "grad_norm": 1.4730199281306078, "learning_rate": 7.0985213720172e-06, "loss": 0.6972, "step": 16729 }, { "epoch": 0.6059398768562115, "grad_norm": 1.3817970337799979, "learning_rate": 7.097398787495654e-06, "loss": 0.7127, "step": 16730 }, { "epoch": 0.6059760956175299, "grad_norm": 1.4012971096261293, "learning_rate": 7.096276242916454e-06, "loss": 0.6994, "step": 16731 }, { "epoch": 0.6060123143788483, "grad_norm": 1.5089149743537278, "learning_rate": 7.095153738295043e-06, "loss": 0.778, "step": 16732 }, { "epoch": 0.6060485331401666, "grad_norm": 1.1149282777560774, "learning_rate": 7.094031273646871e-06, "loss": 0.7026, "step": 16733 }, { "epoch": 0.6060847519014849, "grad_norm": 1.4047816438601182, "learning_rate": 7.092908848987379e-06, "loss": 0.8069, "step": 16734 }, { "epoch": 0.6061209706628033, "grad_norm": 1.400254949277969, "learning_rate": 7.091786464332015e-06, "loss": 0.7211, "step": 16735 }, { "epoch": 0.6061571894241217, "grad_norm": 1.3266600762213836, "learning_rate": 7.090664119696229e-06, "loss": 0.6822, "step": 16736 }, { "epoch": 0.6061934081854401, "grad_norm": 1.3152424403635277, "learning_rate": 7.0895418150954575e-06, "loss": 0.6654, "step": 16737 }, { "epoch": 0.6062296269467584, "grad_norm": 1.4157002327992982, "learning_rate": 7.088419550545151e-06, "loss": 0.7054, "step": 16738 }, { "epoch": 0.6062658457080767, "grad_norm": 1.378236540456371, "learning_rate": 7.087297326060747e-06, "loss": 0.7269, "step": 16739 }, { "epoch": 0.6063020644693952, "grad_norm": 1.3056829767847271, "learning_rate": 7.086175141657694e-06, "loss": 0.7136, "step": 16740 }, { "epoch": 0.6063382832307135, "grad_norm": 1.3405430924629946, "learning_rate": 7.085052997351428e-06, "loss": 0.7323, "step": 16741 }, { "epoch": 0.6063745019920319, "grad_norm": 1.3252010745919094, "learning_rate": 7.083930893157397e-06, "loss": 0.7324, "step": 16742 }, { "epoch": 0.6064107207533502, "grad_norm": 1.4661864343506117, "learning_rate": 7.0828088290910365e-06, "loss": 0.7395, "step": 16743 }, { "epoch": 0.6064469395146685, "grad_norm": 1.2899247183071534, "learning_rate": 7.081686805167792e-06, "loss": 0.7596, "step": 16744 }, { "epoch": 0.606483158275987, "grad_norm": 1.7162948933430335, "learning_rate": 7.080564821403099e-06, "loss": 0.6756, "step": 16745 }, { "epoch": 0.6065193770373053, "grad_norm": 1.3849669607384134, "learning_rate": 7.079442877812401e-06, "loss": 0.7106, "step": 16746 }, { "epoch": 0.6065555957986237, "grad_norm": 1.21230582570523, "learning_rate": 7.078320974411136e-06, "loss": 0.7294, "step": 16747 }, { "epoch": 0.606591814559942, "grad_norm": 1.1813212680384289, "learning_rate": 7.0771991112147386e-06, "loss": 0.6505, "step": 16748 }, { "epoch": 0.6066280333212604, "grad_norm": 1.397674586867195, "learning_rate": 7.076077288238654e-06, "loss": 0.6868, "step": 16749 }, { "epoch": 0.6066642520825788, "grad_norm": 1.26216647379958, "learning_rate": 7.074955505498315e-06, "loss": 0.6238, "step": 16750 }, { "epoch": 0.6067004708438971, "grad_norm": 1.3390821497197665, "learning_rate": 7.073833763009156e-06, "loss": 0.6828, "step": 16751 }, { "epoch": 0.6067366896052155, "grad_norm": 1.3892868259113473, "learning_rate": 7.072712060786618e-06, "loss": 0.7838, "step": 16752 }, { "epoch": 0.6067729083665339, "grad_norm": 1.2979831026123405, "learning_rate": 7.071590398846129e-06, "loss": 0.7145, "step": 16753 }, { "epoch": 0.6068091271278522, "grad_norm": 0.9500900615304684, "learning_rate": 7.070468777203135e-06, "loss": 0.7153, "step": 16754 }, { "epoch": 0.6068453458891706, "grad_norm": 1.2366569008064325, "learning_rate": 7.069347195873059e-06, "loss": 0.7192, "step": 16755 }, { "epoch": 0.6068815646504889, "grad_norm": 1.4265783818531814, "learning_rate": 7.068225654871346e-06, "loss": 0.7464, "step": 16756 }, { "epoch": 0.6069177834118074, "grad_norm": 1.1471708272892012, "learning_rate": 7.06710415421342e-06, "loss": 0.7148, "step": 16757 }, { "epoch": 0.6069540021731257, "grad_norm": 1.3724005020165944, "learning_rate": 7.065982693914719e-06, "loss": 0.7521, "step": 16758 }, { "epoch": 0.606990220934444, "grad_norm": 1.399955274168697, "learning_rate": 7.064861273990675e-06, "loss": 0.7427, "step": 16759 }, { "epoch": 0.6070264396957624, "grad_norm": 1.2506715624580487, "learning_rate": 7.063739894456718e-06, "loss": 0.6634, "step": 16760 }, { "epoch": 0.6070626584570807, "grad_norm": 1.2406263640598034, "learning_rate": 7.0626185553282825e-06, "loss": 0.6952, "step": 16761 }, { "epoch": 0.6070988772183992, "grad_norm": 1.2999962414596342, "learning_rate": 7.061497256620793e-06, "loss": 0.7452, "step": 16762 }, { "epoch": 0.6071350959797175, "grad_norm": 1.004149089669095, "learning_rate": 7.060375998349686e-06, "loss": 0.717, "step": 16763 }, { "epoch": 0.6071713147410358, "grad_norm": 1.3356636823600585, "learning_rate": 7.059254780530387e-06, "loss": 0.7046, "step": 16764 }, { "epoch": 0.6072075335023542, "grad_norm": 1.4787305401072233, "learning_rate": 7.058133603178328e-06, "loss": 0.7896, "step": 16765 }, { "epoch": 0.6072437522636726, "grad_norm": 1.2395665525922077, "learning_rate": 7.057012466308932e-06, "loss": 0.7248, "step": 16766 }, { "epoch": 0.607279971024991, "grad_norm": 1.5284050622421572, "learning_rate": 7.055891369937634e-06, "loss": 0.7555, "step": 16767 }, { "epoch": 0.6073161897863093, "grad_norm": 1.2859397068329204, "learning_rate": 7.054770314079855e-06, "loss": 0.7034, "step": 16768 }, { "epoch": 0.6073524085476276, "grad_norm": 1.0558214448004652, "learning_rate": 7.053649298751026e-06, "loss": 0.6995, "step": 16769 }, { "epoch": 0.607388627308946, "grad_norm": 1.372500409223282, "learning_rate": 7.052528323966572e-06, "loss": 0.6302, "step": 16770 }, { "epoch": 0.6074248460702644, "grad_norm": 1.5832946442177078, "learning_rate": 7.051407389741916e-06, "loss": 0.7109, "step": 16771 }, { "epoch": 0.6074610648315828, "grad_norm": 1.427260647731185, "learning_rate": 7.050286496092487e-06, "loss": 0.8195, "step": 16772 }, { "epoch": 0.6074972835929011, "grad_norm": 1.3260336046792873, "learning_rate": 7.049165643033706e-06, "loss": 0.7201, "step": 16773 }, { "epoch": 0.6075335023542194, "grad_norm": 1.4177550142948634, "learning_rate": 7.048044830581001e-06, "loss": 0.6606, "step": 16774 }, { "epoch": 0.6075697211155379, "grad_norm": 1.3515430549645038, "learning_rate": 7.046924058749791e-06, "loss": 0.6487, "step": 16775 }, { "epoch": 0.6076059398768562, "grad_norm": 1.3360768421184201, "learning_rate": 7.045803327555502e-06, "loss": 0.684, "step": 16776 }, { "epoch": 0.6076421586381746, "grad_norm": 1.3505442475665013, "learning_rate": 7.0446826370135535e-06, "loss": 0.6881, "step": 16777 }, { "epoch": 0.6076783773994929, "grad_norm": 1.3619677684178124, "learning_rate": 7.043561987139369e-06, "loss": 0.7312, "step": 16778 }, { "epoch": 0.6077145961608112, "grad_norm": 1.428519141568661, "learning_rate": 7.0424413779483705e-06, "loss": 0.6483, "step": 16779 }, { "epoch": 0.6077508149221297, "grad_norm": 1.4376720229652715, "learning_rate": 7.041320809455976e-06, "loss": 0.7714, "step": 16780 }, { "epoch": 0.607787033683448, "grad_norm": 1.2984984359489424, "learning_rate": 7.040200281677609e-06, "loss": 0.6962, "step": 16781 }, { "epoch": 0.6078232524447664, "grad_norm": 1.2909188551711792, "learning_rate": 7.039079794628685e-06, "loss": 0.7652, "step": 16782 }, { "epoch": 0.6078594712060847, "grad_norm": 0.9741909349831145, "learning_rate": 7.037959348324626e-06, "loss": 0.6871, "step": 16783 }, { "epoch": 0.6078956899674031, "grad_norm": 1.471366227502819, "learning_rate": 7.036838942780846e-06, "loss": 0.7581, "step": 16784 }, { "epoch": 0.6079319087287215, "grad_norm": 1.279691226252812, "learning_rate": 7.03571857801277e-06, "loss": 0.6888, "step": 16785 }, { "epoch": 0.6079681274900398, "grad_norm": 1.332042984896789, "learning_rate": 7.034598254035807e-06, "loss": 0.7719, "step": 16786 }, { "epoch": 0.6080043462513582, "grad_norm": 1.0646058130011362, "learning_rate": 7.033477970865381e-06, "loss": 0.7119, "step": 16787 }, { "epoch": 0.6080405650126766, "grad_norm": 0.9900383757364323, "learning_rate": 7.032357728516903e-06, "loss": 0.7118, "step": 16788 }, { "epoch": 0.6080767837739949, "grad_norm": 1.029751517569186, "learning_rate": 7.0312375270057895e-06, "loss": 0.7544, "step": 16789 }, { "epoch": 0.6081130025353133, "grad_norm": 1.0175105965886346, "learning_rate": 7.030117366347458e-06, "loss": 0.7201, "step": 16790 }, { "epoch": 0.6081492212966316, "grad_norm": 1.336375858237164, "learning_rate": 7.028997246557321e-06, "loss": 0.7783, "step": 16791 }, { "epoch": 0.60818544005795, "grad_norm": 1.3088481143437554, "learning_rate": 7.0278771676507915e-06, "loss": 0.7114, "step": 16792 }, { "epoch": 0.6082216588192684, "grad_norm": 1.3537964825341953, "learning_rate": 7.026757129643284e-06, "loss": 0.7042, "step": 16793 }, { "epoch": 0.6082578775805867, "grad_norm": 1.2777178374047968, "learning_rate": 7.025637132550213e-06, "loss": 0.6504, "step": 16794 }, { "epoch": 0.6082940963419051, "grad_norm": 1.2897582690670226, "learning_rate": 7.024517176386985e-06, "loss": 0.6956, "step": 16795 }, { "epoch": 0.6083303151032234, "grad_norm": 1.4285164522737648, "learning_rate": 7.023397261169018e-06, "loss": 0.7649, "step": 16796 }, { "epoch": 0.6083665338645419, "grad_norm": 1.0168906578577073, "learning_rate": 7.022277386911718e-06, "loss": 0.6983, "step": 16797 }, { "epoch": 0.6084027526258602, "grad_norm": 1.3868994568223714, "learning_rate": 7.0211575536305e-06, "loss": 0.7114, "step": 16798 }, { "epoch": 0.6084389713871785, "grad_norm": 1.3587168465028965, "learning_rate": 7.020037761340769e-06, "loss": 0.7516, "step": 16799 }, { "epoch": 0.6084751901484969, "grad_norm": 1.7367174529509648, "learning_rate": 7.018918010057936e-06, "loss": 0.6635, "step": 16800 }, { "epoch": 0.6085114089098153, "grad_norm": 0.9134106372998705, "learning_rate": 7.017798299797414e-06, "loss": 0.684, "step": 16801 }, { "epoch": 0.6085476276711337, "grad_norm": 1.4231287477445613, "learning_rate": 7.016678630574605e-06, "loss": 0.7398, "step": 16802 }, { "epoch": 0.608583846432452, "grad_norm": 2.1629499371143193, "learning_rate": 7.015559002404921e-06, "loss": 0.7085, "step": 16803 }, { "epoch": 0.6086200651937703, "grad_norm": 1.4261372337376217, "learning_rate": 7.014439415303768e-06, "loss": 0.6789, "step": 16804 }, { "epoch": 0.6086562839550888, "grad_norm": 1.4190842599172806, "learning_rate": 7.013319869286552e-06, "loss": 0.745, "step": 16805 }, { "epoch": 0.6086925027164071, "grad_norm": 1.3853660278608455, "learning_rate": 7.012200364368677e-06, "loss": 0.7033, "step": 16806 }, { "epoch": 0.6087287214777255, "grad_norm": 1.3787534435910966, "learning_rate": 7.011080900565552e-06, "loss": 0.6683, "step": 16807 }, { "epoch": 0.6087649402390438, "grad_norm": 0.9932263326616707, "learning_rate": 7.0099614778925785e-06, "loss": 0.6774, "step": 16808 }, { "epoch": 0.6088011590003621, "grad_norm": 1.2732416716175348, "learning_rate": 7.008842096365163e-06, "loss": 0.5936, "step": 16809 }, { "epoch": 0.6088373777616806, "grad_norm": 1.3281889798377275, "learning_rate": 7.007722755998709e-06, "loss": 0.7326, "step": 16810 }, { "epoch": 0.6088735965229989, "grad_norm": 1.611924388264944, "learning_rate": 7.006603456808618e-06, "loss": 0.6642, "step": 16811 }, { "epoch": 0.6089098152843173, "grad_norm": 1.2467993204811993, "learning_rate": 7.005484198810296e-06, "loss": 0.763, "step": 16812 }, { "epoch": 0.6089460340456356, "grad_norm": 1.379347299425462, "learning_rate": 7.004364982019141e-06, "loss": 0.6778, "step": 16813 }, { "epoch": 0.608982252806954, "grad_norm": 1.0265424944586339, "learning_rate": 7.003245806450557e-06, "loss": 0.7249, "step": 16814 }, { "epoch": 0.6090184715682724, "grad_norm": 1.4474214400457928, "learning_rate": 7.002126672119942e-06, "loss": 0.7967, "step": 16815 }, { "epoch": 0.6090546903295907, "grad_norm": 1.3306940175028492, "learning_rate": 7.0010075790427e-06, "loss": 0.7433, "step": 16816 }, { "epoch": 0.6090909090909091, "grad_norm": 1.0332140581691878, "learning_rate": 6.9998885272342276e-06, "loss": 0.708, "step": 16817 }, { "epoch": 0.6091271278522274, "grad_norm": 1.2447052322755632, "learning_rate": 6.998769516709927e-06, "loss": 0.6469, "step": 16818 }, { "epoch": 0.6091633466135458, "grad_norm": 1.4044495434600528, "learning_rate": 6.997650547485193e-06, "loss": 0.7795, "step": 16819 }, { "epoch": 0.6091995653748642, "grad_norm": 1.1758745643487882, "learning_rate": 6.996531619575425e-06, "loss": 0.6757, "step": 16820 }, { "epoch": 0.6092357841361825, "grad_norm": 1.372255454689472, "learning_rate": 6.9954127329960235e-06, "loss": 0.7362, "step": 16821 }, { "epoch": 0.6092720028975009, "grad_norm": 1.40328117613351, "learning_rate": 6.99429388776238e-06, "loss": 0.7662, "step": 16822 }, { "epoch": 0.6093082216588193, "grad_norm": 1.5257297729855948, "learning_rate": 6.993175083889897e-06, "loss": 0.7774, "step": 16823 }, { "epoch": 0.6093444404201376, "grad_norm": 1.8302177362715215, "learning_rate": 6.992056321393964e-06, "loss": 0.637, "step": 16824 }, { "epoch": 0.609380659181456, "grad_norm": 1.588627533989129, "learning_rate": 6.990937600289982e-06, "loss": 0.7029, "step": 16825 }, { "epoch": 0.6094168779427743, "grad_norm": 1.1008900821471796, "learning_rate": 6.9898189205933385e-06, "loss": 0.7472, "step": 16826 }, { "epoch": 0.6094530967040928, "grad_norm": 1.3342788924106537, "learning_rate": 6.988700282319436e-06, "loss": 0.6697, "step": 16827 }, { "epoch": 0.6094893154654111, "grad_norm": 1.6343780602796651, "learning_rate": 6.987581685483659e-06, "loss": 0.7853, "step": 16828 }, { "epoch": 0.6095255342267294, "grad_norm": 0.9795397606565859, "learning_rate": 6.986463130101411e-06, "loss": 0.7217, "step": 16829 }, { "epoch": 0.6095617529880478, "grad_norm": 1.3372765252440804, "learning_rate": 6.9853446161880726e-06, "loss": 0.7034, "step": 16830 }, { "epoch": 0.6095979717493661, "grad_norm": 1.4836735147022513, "learning_rate": 6.984226143759043e-06, "loss": 0.7111, "step": 16831 }, { "epoch": 0.6096341905106846, "grad_norm": 1.4718168551659478, "learning_rate": 6.983107712829713e-06, "loss": 0.752, "step": 16832 }, { "epoch": 0.6096704092720029, "grad_norm": 1.4301849969245937, "learning_rate": 6.981989323415469e-06, "loss": 0.7393, "step": 16833 }, { "epoch": 0.6097066280333212, "grad_norm": 4.475544848785136, "learning_rate": 6.980870975531706e-06, "loss": 0.7102, "step": 16834 }, { "epoch": 0.6097428467946396, "grad_norm": 1.2600435216588088, "learning_rate": 6.979752669193811e-06, "loss": 0.6221, "step": 16835 }, { "epoch": 0.609779065555958, "grad_norm": 1.4486711037985585, "learning_rate": 6.9786344044171725e-06, "loss": 0.7773, "step": 16836 }, { "epoch": 0.6098152843172764, "grad_norm": 1.5747317313689342, "learning_rate": 6.977516181217179e-06, "loss": 0.7651, "step": 16837 }, { "epoch": 0.6098515030785947, "grad_norm": 1.3197657769431792, "learning_rate": 6.9763979996092215e-06, "loss": 0.6739, "step": 16838 }, { "epoch": 0.609887721839913, "grad_norm": 1.3590075397682735, "learning_rate": 6.97527985960868e-06, "loss": 0.7503, "step": 16839 }, { "epoch": 0.6099239406012315, "grad_norm": 1.2794535278910857, "learning_rate": 6.974161761230947e-06, "loss": 0.6775, "step": 16840 }, { "epoch": 0.6099601593625498, "grad_norm": 1.37764027541587, "learning_rate": 6.973043704491408e-06, "loss": 0.7344, "step": 16841 }, { "epoch": 0.6099963781238682, "grad_norm": 1.5054867997248007, "learning_rate": 6.9719256894054455e-06, "loss": 0.7537, "step": 16842 }, { "epoch": 0.6100325968851865, "grad_norm": 1.450134476435643, "learning_rate": 6.970807715988448e-06, "loss": 0.7429, "step": 16843 }, { "epoch": 0.6100688156465048, "grad_norm": 1.2815473445183057, "learning_rate": 6.969689784255798e-06, "loss": 0.6446, "step": 16844 }, { "epoch": 0.6101050344078233, "grad_norm": 1.4072882554709172, "learning_rate": 6.9685718942228785e-06, "loss": 0.7627, "step": 16845 }, { "epoch": 0.6101412531691416, "grad_norm": 1.3582487503772631, "learning_rate": 6.967454045905073e-06, "loss": 0.7122, "step": 16846 }, { "epoch": 0.61017747193046, "grad_norm": 1.1517494674091298, "learning_rate": 6.966336239317766e-06, "loss": 0.753, "step": 16847 }, { "epoch": 0.6102136906917783, "grad_norm": 0.987955409046139, "learning_rate": 6.965218474476336e-06, "loss": 0.6901, "step": 16848 }, { "epoch": 0.6102499094530967, "grad_norm": 1.1262038980301516, "learning_rate": 6.964100751396168e-06, "loss": 0.6881, "step": 16849 }, { "epoch": 0.6102861282144151, "grad_norm": 1.3222464288194329, "learning_rate": 6.962983070092639e-06, "loss": 0.7087, "step": 16850 }, { "epoch": 0.6103223469757334, "grad_norm": 1.3817277306559588, "learning_rate": 6.961865430581133e-06, "loss": 0.7311, "step": 16851 }, { "epoch": 0.6103585657370518, "grad_norm": 1.3916256678317167, "learning_rate": 6.960747832877028e-06, "loss": 0.6612, "step": 16852 }, { "epoch": 0.6103947844983701, "grad_norm": 1.2898640488645263, "learning_rate": 6.959630276995704e-06, "loss": 0.723, "step": 16853 }, { "epoch": 0.6104310032596885, "grad_norm": 1.4160965633280207, "learning_rate": 6.958512762952539e-06, "loss": 0.7414, "step": 16854 }, { "epoch": 0.6104672220210069, "grad_norm": 1.322665991383938, "learning_rate": 6.957395290762909e-06, "loss": 0.7, "step": 16855 }, { "epoch": 0.6105034407823252, "grad_norm": 1.113762301217092, "learning_rate": 6.9562778604421965e-06, "loss": 0.704, "step": 16856 }, { "epoch": 0.6105396595436436, "grad_norm": 1.3693088150657364, "learning_rate": 6.955160472005772e-06, "loss": 0.7242, "step": 16857 }, { "epoch": 0.610575878304962, "grad_norm": 1.375062762258548, "learning_rate": 6.954043125469017e-06, "loss": 0.7072, "step": 16858 }, { "epoch": 0.6106120970662803, "grad_norm": 1.8626890140989747, "learning_rate": 6.952925820847302e-06, "loss": 0.6315, "step": 16859 }, { "epoch": 0.6106483158275987, "grad_norm": 1.4702684778460098, "learning_rate": 6.95180855815601e-06, "loss": 0.624, "step": 16860 }, { "epoch": 0.610684534588917, "grad_norm": 2.2641864682121327, "learning_rate": 6.9506913374105065e-06, "loss": 0.7306, "step": 16861 }, { "epoch": 0.6107207533502355, "grad_norm": 1.3489982322612537, "learning_rate": 6.949574158626169e-06, "loss": 0.6422, "step": 16862 }, { "epoch": 0.6107569721115538, "grad_norm": 1.2621886900412205, "learning_rate": 6.948457021818376e-06, "loss": 0.6987, "step": 16863 }, { "epoch": 0.6107931908728721, "grad_norm": 1.536697200348662, "learning_rate": 6.947339927002491e-06, "loss": 0.713, "step": 16864 }, { "epoch": 0.6108294096341905, "grad_norm": 1.5243018779249247, "learning_rate": 6.946222874193893e-06, "loss": 0.7354, "step": 16865 }, { "epoch": 0.6108656283955088, "grad_norm": 1.4340545552952833, "learning_rate": 6.94510586340795e-06, "loss": 0.7084, "step": 16866 }, { "epoch": 0.6109018471568273, "grad_norm": 1.2891845132419402, "learning_rate": 6.943988894660039e-06, "loss": 0.6854, "step": 16867 }, { "epoch": 0.6109380659181456, "grad_norm": 1.3073197269341157, "learning_rate": 6.942871967965521e-06, "loss": 0.6882, "step": 16868 }, { "epoch": 0.6109742846794639, "grad_norm": 1.3612284480180552, "learning_rate": 6.941755083339774e-06, "loss": 0.7223, "step": 16869 }, { "epoch": 0.6110105034407823, "grad_norm": 1.3503604846930939, "learning_rate": 6.940638240798163e-06, "loss": 0.7368, "step": 16870 }, { "epoch": 0.6110467222021007, "grad_norm": 1.3013580177667656, "learning_rate": 6.939521440356056e-06, "loss": 0.7103, "step": 16871 }, { "epoch": 0.6110829409634191, "grad_norm": 1.306445915528847, "learning_rate": 6.938404682028827e-06, "loss": 0.7254, "step": 16872 }, { "epoch": 0.6111191597247374, "grad_norm": 9.044993150496538, "learning_rate": 6.937287965831838e-06, "loss": 0.6669, "step": 16873 }, { "epoch": 0.6111553784860557, "grad_norm": 1.1861543735840043, "learning_rate": 6.936171291780458e-06, "loss": 0.7086, "step": 16874 }, { "epoch": 0.6111915972473742, "grad_norm": 1.488074292038345, "learning_rate": 6.935054659890053e-06, "loss": 0.7665, "step": 16875 }, { "epoch": 0.6112278160086925, "grad_norm": 1.2091364044683428, "learning_rate": 6.93393807017599e-06, "loss": 0.7378, "step": 16876 }, { "epoch": 0.6112640347700109, "grad_norm": 1.4559545092337296, "learning_rate": 6.93282152265363e-06, "loss": 0.7222, "step": 16877 }, { "epoch": 0.6113002535313292, "grad_norm": 1.1060664885925544, "learning_rate": 6.931705017338345e-06, "loss": 0.6849, "step": 16878 }, { "epoch": 0.6113364722926475, "grad_norm": 1.3148838813967556, "learning_rate": 6.930588554245492e-06, "loss": 0.7517, "step": 16879 }, { "epoch": 0.611372691053966, "grad_norm": 1.3554001634158697, "learning_rate": 6.929472133390441e-06, "loss": 0.6772, "step": 16880 }, { "epoch": 0.6114089098152843, "grad_norm": 1.267516003866002, "learning_rate": 6.928355754788547e-06, "loss": 0.7231, "step": 16881 }, { "epoch": 0.6114451285766027, "grad_norm": 1.4675811976709559, "learning_rate": 6.9272394184551785e-06, "loss": 0.6274, "step": 16882 }, { "epoch": 0.611481347337921, "grad_norm": 1.2999439723751802, "learning_rate": 6.926123124405696e-06, "loss": 0.7428, "step": 16883 }, { "epoch": 0.6115175660992394, "grad_norm": 1.350628240533778, "learning_rate": 6.92500687265546e-06, "loss": 0.6366, "step": 16884 }, { "epoch": 0.6115537848605578, "grad_norm": 1.4313717911054438, "learning_rate": 6.923890663219832e-06, "loss": 0.7102, "step": 16885 }, { "epoch": 0.6115900036218761, "grad_norm": 1.2534717892652167, "learning_rate": 6.92277449611417e-06, "loss": 0.7117, "step": 16886 }, { "epoch": 0.6116262223831945, "grad_norm": 2.040670551898654, "learning_rate": 6.921658371353836e-06, "loss": 0.6935, "step": 16887 }, { "epoch": 0.6116624411445128, "grad_norm": 1.3696699207713086, "learning_rate": 6.920542288954186e-06, "loss": 0.7278, "step": 16888 }, { "epoch": 0.6116986599058312, "grad_norm": 1.3361035874297031, "learning_rate": 6.919426248930583e-06, "loss": 0.73, "step": 16889 }, { "epoch": 0.6117348786671496, "grad_norm": 1.2365481175647182, "learning_rate": 6.918310251298379e-06, "loss": 0.714, "step": 16890 }, { "epoch": 0.6117710974284679, "grad_norm": 1.0622373581842082, "learning_rate": 6.917194296072936e-06, "loss": 0.7073, "step": 16891 }, { "epoch": 0.6118073161897863, "grad_norm": 1.3207902749255414, "learning_rate": 6.916078383269607e-06, "loss": 0.6759, "step": 16892 }, { "epoch": 0.6118435349511047, "grad_norm": 1.39974354713825, "learning_rate": 6.914962512903748e-06, "loss": 0.7737, "step": 16893 }, { "epoch": 0.611879753712423, "grad_norm": 1.3737469891887435, "learning_rate": 6.913846684990717e-06, "loss": 0.6869, "step": 16894 }, { "epoch": 0.6119159724737414, "grad_norm": 1.0668416613020748, "learning_rate": 6.912730899545867e-06, "loss": 0.7491, "step": 16895 }, { "epoch": 0.6119521912350597, "grad_norm": 1.3131173504157674, "learning_rate": 6.911615156584554e-06, "loss": 0.6734, "step": 16896 }, { "epoch": 0.6119884099963782, "grad_norm": 1.3879085970143645, "learning_rate": 6.910499456122128e-06, "loss": 0.7465, "step": 16897 }, { "epoch": 0.6120246287576965, "grad_norm": 1.0522052374163435, "learning_rate": 6.909383798173945e-06, "loss": 0.7364, "step": 16898 }, { "epoch": 0.6120608475190148, "grad_norm": 1.327928204541922, "learning_rate": 6.908268182755356e-06, "loss": 0.6416, "step": 16899 }, { "epoch": 0.6120970662803332, "grad_norm": 1.1596029548490425, "learning_rate": 6.907152609881716e-06, "loss": 0.6481, "step": 16900 }, { "epoch": 0.6121332850416515, "grad_norm": 1.3225330113422002, "learning_rate": 6.906037079568371e-06, "loss": 0.6487, "step": 16901 }, { "epoch": 0.61216950380297, "grad_norm": 1.513588212685268, "learning_rate": 6.904921591830673e-06, "loss": 0.7527, "step": 16902 }, { "epoch": 0.6122057225642883, "grad_norm": 1.3684076259860236, "learning_rate": 6.9038061466839765e-06, "loss": 0.738, "step": 16903 }, { "epoch": 0.6122419413256066, "grad_norm": 0.9413168827525588, "learning_rate": 6.902690744143627e-06, "loss": 0.6759, "step": 16904 }, { "epoch": 0.612278160086925, "grad_norm": 1.3829830503624085, "learning_rate": 6.901575384224974e-06, "loss": 0.724, "step": 16905 }, { "epoch": 0.6123143788482434, "grad_norm": 1.0661982757189863, "learning_rate": 6.900460066943366e-06, "loss": 0.8168, "step": 16906 }, { "epoch": 0.6123505976095618, "grad_norm": 1.4291306177167504, "learning_rate": 6.8993447923141535e-06, "loss": 0.7771, "step": 16907 }, { "epoch": 0.6123868163708801, "grad_norm": 1.3307776075257098, "learning_rate": 6.898229560352678e-06, "loss": 0.6982, "step": 16908 }, { "epoch": 0.6124230351321984, "grad_norm": 1.1871310186476363, "learning_rate": 6.897114371074292e-06, "loss": 0.7498, "step": 16909 }, { "epoch": 0.6124592538935169, "grad_norm": 1.3565496551500476, "learning_rate": 6.895999224494336e-06, "loss": 0.6917, "step": 16910 }, { "epoch": 0.6124954726548352, "grad_norm": 0.936765745090086, "learning_rate": 6.89488412062816e-06, "loss": 0.7216, "step": 16911 }, { "epoch": 0.6125316914161536, "grad_norm": 1.4462669636173322, "learning_rate": 6.893769059491106e-06, "loss": 0.7614, "step": 16912 }, { "epoch": 0.6125679101774719, "grad_norm": 1.383581989977167, "learning_rate": 6.892654041098517e-06, "loss": 0.7812, "step": 16913 }, { "epoch": 0.6126041289387903, "grad_norm": 1.3272742442669003, "learning_rate": 6.8915390654657425e-06, "loss": 0.6991, "step": 16914 }, { "epoch": 0.6126403477001087, "grad_norm": 1.543149457079522, "learning_rate": 6.89042413260812e-06, "loss": 0.7951, "step": 16915 }, { "epoch": 0.612676566461427, "grad_norm": 1.0556249591989613, "learning_rate": 6.8893092425409946e-06, "loss": 0.6662, "step": 16916 }, { "epoch": 0.6127127852227454, "grad_norm": 1.3996428278077377, "learning_rate": 6.888194395279706e-06, "loss": 0.815, "step": 16917 }, { "epoch": 0.6127490039840637, "grad_norm": 1.273459579032831, "learning_rate": 6.887079590839598e-06, "loss": 0.6673, "step": 16918 }, { "epoch": 0.6127852227453822, "grad_norm": 1.3861461914769202, "learning_rate": 6.88596482923601e-06, "loss": 0.7555, "step": 16919 }, { "epoch": 0.6128214415067005, "grad_norm": 1.284292426436634, "learning_rate": 6.884850110484282e-06, "loss": 0.7435, "step": 16920 }, { "epoch": 0.6128576602680188, "grad_norm": 1.366773271479512, "learning_rate": 6.883735434599752e-06, "loss": 0.7003, "step": 16921 }, { "epoch": 0.6128938790293372, "grad_norm": 1.1684602488502551, "learning_rate": 6.882620801597763e-06, "loss": 0.7368, "step": 16922 }, { "epoch": 0.6129300977906555, "grad_norm": 1.117474280117453, "learning_rate": 6.881506211493649e-06, "loss": 0.6973, "step": 16923 }, { "epoch": 0.612966316551974, "grad_norm": 1.3427700028809058, "learning_rate": 6.88039166430275e-06, "loss": 0.6779, "step": 16924 }, { "epoch": 0.6130025353132923, "grad_norm": 1.2642583643138428, "learning_rate": 6.879277160040405e-06, "loss": 0.6918, "step": 16925 }, { "epoch": 0.6130387540746106, "grad_norm": 1.3580581737996968, "learning_rate": 6.878162698721946e-06, "loss": 0.6807, "step": 16926 }, { "epoch": 0.613074972835929, "grad_norm": 1.477511630714454, "learning_rate": 6.877048280362713e-06, "loss": 0.7475, "step": 16927 }, { "epoch": 0.6131111915972474, "grad_norm": 1.3418049180457474, "learning_rate": 6.875933904978038e-06, "loss": 0.6546, "step": 16928 }, { "epoch": 0.6131474103585658, "grad_norm": 1.2519810520217383, "learning_rate": 6.87481957258326e-06, "loss": 0.7485, "step": 16929 }, { "epoch": 0.6131836291198841, "grad_norm": 0.9857605690420432, "learning_rate": 6.873705283193708e-06, "loss": 0.6606, "step": 16930 }, { "epoch": 0.6132198478812024, "grad_norm": 1.3254340054534806, "learning_rate": 6.8725910368247206e-06, "loss": 0.7312, "step": 16931 }, { "epoch": 0.6132560666425209, "grad_norm": 1.4898350994437073, "learning_rate": 6.871476833491627e-06, "loss": 0.7197, "step": 16932 }, { "epoch": 0.6132922854038392, "grad_norm": 1.403740711727053, "learning_rate": 6.870362673209761e-06, "loss": 0.718, "step": 16933 }, { "epoch": 0.6133285041651576, "grad_norm": 2.122606416018307, "learning_rate": 6.869248555994457e-06, "loss": 0.6372, "step": 16934 }, { "epoch": 0.6133647229264759, "grad_norm": 1.0073308140704746, "learning_rate": 6.86813448186104e-06, "loss": 0.7434, "step": 16935 }, { "epoch": 0.6134009416877942, "grad_norm": 1.4373618576544311, "learning_rate": 6.867020450824847e-06, "loss": 0.7223, "step": 16936 }, { "epoch": 0.6134371604491127, "grad_norm": 0.9716038815722159, "learning_rate": 6.865906462901204e-06, "loss": 0.6709, "step": 16937 }, { "epoch": 0.613473379210431, "grad_norm": 1.315174546380754, "learning_rate": 6.864792518105444e-06, "loss": 0.7169, "step": 16938 }, { "epoch": 0.6135095979717494, "grad_norm": 1.2620600483829496, "learning_rate": 6.863678616452891e-06, "loss": 0.6032, "step": 16939 }, { "epoch": 0.6135458167330677, "grad_norm": 1.4034638215327333, "learning_rate": 6.8625647579588795e-06, "loss": 0.7851, "step": 16940 }, { "epoch": 0.6135820354943861, "grad_norm": 1.277204624088763, "learning_rate": 6.861450942638731e-06, "loss": 0.686, "step": 16941 }, { "epoch": 0.6136182542557045, "grad_norm": 1.4548453637664074, "learning_rate": 6.860337170507777e-06, "loss": 0.6552, "step": 16942 }, { "epoch": 0.6136544730170228, "grad_norm": 0.9362001885486921, "learning_rate": 6.859223441581341e-06, "loss": 0.7209, "step": 16943 }, { "epoch": 0.6136906917783412, "grad_norm": 1.2879629589775758, "learning_rate": 6.858109755874749e-06, "loss": 0.6531, "step": 16944 }, { "epoch": 0.6137269105396596, "grad_norm": 1.065228161581255, "learning_rate": 6.85699611340333e-06, "loss": 0.694, "step": 16945 }, { "epoch": 0.6137631293009779, "grad_norm": 1.414535351854422, "learning_rate": 6.855882514182405e-06, "loss": 0.7309, "step": 16946 }, { "epoch": 0.6137993480622963, "grad_norm": 1.4873039493884574, "learning_rate": 6.8547689582273e-06, "loss": 0.7468, "step": 16947 }, { "epoch": 0.6138355668236146, "grad_norm": 1.2263736568091352, "learning_rate": 6.853655445553336e-06, "loss": 0.732, "step": 16948 }, { "epoch": 0.613871785584933, "grad_norm": 1.0365475668898803, "learning_rate": 6.8525419761758405e-06, "loss": 0.6736, "step": 16949 }, { "epoch": 0.6139080043462514, "grad_norm": 1.2599229448249718, "learning_rate": 6.851428550110131e-06, "loss": 0.5982, "step": 16950 }, { "epoch": 0.6139442231075697, "grad_norm": 1.183197985858263, "learning_rate": 6.850315167371533e-06, "loss": 0.7545, "step": 16951 }, { "epoch": 0.6139804418688881, "grad_norm": 1.3947957092653434, "learning_rate": 6.849201827975362e-06, "loss": 0.6301, "step": 16952 }, { "epoch": 0.6140166606302064, "grad_norm": 1.3936501866574476, "learning_rate": 6.848088531936947e-06, "loss": 0.6986, "step": 16953 }, { "epoch": 0.6140528793915249, "grad_norm": 1.0302918816108744, "learning_rate": 6.846975279271599e-06, "loss": 0.6884, "step": 16954 }, { "epoch": 0.6140890981528432, "grad_norm": 1.4559245022576137, "learning_rate": 6.845862069994644e-06, "loss": 0.7901, "step": 16955 }, { "epoch": 0.6141253169141615, "grad_norm": 1.3230941478504574, "learning_rate": 6.844748904121399e-06, "loss": 0.6717, "step": 16956 }, { "epoch": 0.6141615356754799, "grad_norm": 1.3321497394629345, "learning_rate": 6.843635781667179e-06, "loss": 0.6861, "step": 16957 }, { "epoch": 0.6141977544367982, "grad_norm": 1.4738626433596125, "learning_rate": 6.842522702647307e-06, "loss": 0.7198, "step": 16958 }, { "epoch": 0.6142339731981167, "grad_norm": 1.1602574918232968, "learning_rate": 6.841409667077096e-06, "loss": 0.7102, "step": 16959 }, { "epoch": 0.614270191959435, "grad_norm": 1.6792294343413412, "learning_rate": 6.840296674971864e-06, "loss": 0.7311, "step": 16960 }, { "epoch": 0.6143064107207533, "grad_norm": 1.3327180847991738, "learning_rate": 6.83918372634693e-06, "loss": 0.695, "step": 16961 }, { "epoch": 0.6143426294820717, "grad_norm": 1.400189345351477, "learning_rate": 6.838070821217598e-06, "loss": 0.6742, "step": 16962 }, { "epoch": 0.6143788482433901, "grad_norm": 1.29061721910251, "learning_rate": 6.836957959599194e-06, "loss": 0.6374, "step": 16963 }, { "epoch": 0.6144150670047085, "grad_norm": 1.356696157145416, "learning_rate": 6.835845141507025e-06, "loss": 0.6829, "step": 16964 }, { "epoch": 0.6144512857660268, "grad_norm": 1.6640086904073013, "learning_rate": 6.834732366956407e-06, "loss": 0.7176, "step": 16965 }, { "epoch": 0.6144875045273451, "grad_norm": 1.413350911529708, "learning_rate": 6.833619635962656e-06, "loss": 0.6853, "step": 16966 }, { "epoch": 0.6145237232886636, "grad_norm": 1.4775498132005715, "learning_rate": 6.8325069485410775e-06, "loss": 0.7289, "step": 16967 }, { "epoch": 0.6145599420499819, "grad_norm": 1.3372475352386708, "learning_rate": 6.8313943047069905e-06, "loss": 0.7238, "step": 16968 }, { "epoch": 0.6145961608113003, "grad_norm": 1.116595262681012, "learning_rate": 6.830281704475699e-06, "loss": 0.6819, "step": 16969 }, { "epoch": 0.6146323795726186, "grad_norm": 1.3178652112729887, "learning_rate": 6.829169147862517e-06, "loss": 0.664, "step": 16970 }, { "epoch": 0.6146685983339369, "grad_norm": 1.289251000175861, "learning_rate": 6.828056634882754e-06, "loss": 0.7028, "step": 16971 }, { "epoch": 0.6147048170952554, "grad_norm": 1.3186347546661887, "learning_rate": 6.826944165551719e-06, "loss": 0.6693, "step": 16972 }, { "epoch": 0.6147410358565737, "grad_norm": 1.0390072096469913, "learning_rate": 6.82583173988472e-06, "loss": 0.7132, "step": 16973 }, { "epoch": 0.6147772546178921, "grad_norm": 1.1508839368507937, "learning_rate": 6.824719357897066e-06, "loss": 0.752, "step": 16974 }, { "epoch": 0.6148134733792104, "grad_norm": 1.0817183948074862, "learning_rate": 6.823607019604063e-06, "loss": 0.712, "step": 16975 }, { "epoch": 0.6148496921405288, "grad_norm": 1.3369488234420945, "learning_rate": 6.822494725021017e-06, "loss": 0.7558, "step": 16976 }, { "epoch": 0.6148859109018472, "grad_norm": 1.1050820700071728, "learning_rate": 6.821382474163239e-06, "loss": 0.6931, "step": 16977 }, { "epoch": 0.6149221296631655, "grad_norm": 1.3184758057220722, "learning_rate": 6.820270267046027e-06, "loss": 0.6479, "step": 16978 }, { "epoch": 0.6149583484244839, "grad_norm": 1.2849391982962926, "learning_rate": 6.819158103684693e-06, "loss": 0.6913, "step": 16979 }, { "epoch": 0.6149945671858023, "grad_norm": 1.352344464511782, "learning_rate": 6.818045984094538e-06, "loss": 0.7297, "step": 16980 }, { "epoch": 0.6150307859471206, "grad_norm": 1.0976591648435579, "learning_rate": 6.816933908290867e-06, "loss": 0.6835, "step": 16981 }, { "epoch": 0.615067004708439, "grad_norm": 1.4325094364524935, "learning_rate": 6.815821876288979e-06, "loss": 0.7373, "step": 16982 }, { "epoch": 0.6151032234697573, "grad_norm": 1.3807785895428695, "learning_rate": 6.8147098881041855e-06, "loss": 0.6564, "step": 16983 }, { "epoch": 0.6151394422310758, "grad_norm": 1.3344894614698628, "learning_rate": 6.8135979437517775e-06, "loss": 0.7319, "step": 16984 }, { "epoch": 0.6151756609923941, "grad_norm": 1.1418687506334346, "learning_rate": 6.812486043247064e-06, "loss": 0.6547, "step": 16985 }, { "epoch": 0.6152118797537124, "grad_norm": 1.3461643329730317, "learning_rate": 6.811374186605342e-06, "loss": 0.6854, "step": 16986 }, { "epoch": 0.6152480985150308, "grad_norm": 1.3398230936165276, "learning_rate": 6.810262373841914e-06, "loss": 0.7084, "step": 16987 }, { "epoch": 0.6152843172763491, "grad_norm": 0.9639696051826169, "learning_rate": 6.809150604972079e-06, "loss": 0.6907, "step": 16988 }, { "epoch": 0.6153205360376676, "grad_norm": 1.4173547371644628, "learning_rate": 6.808038880011133e-06, "loss": 0.7353, "step": 16989 }, { "epoch": 0.6153567547989859, "grad_norm": 1.3760777775322723, "learning_rate": 6.80692719897438e-06, "loss": 0.6745, "step": 16990 }, { "epoch": 0.6153929735603042, "grad_norm": 1.2391580676984806, "learning_rate": 6.805815561877111e-06, "loss": 0.704, "step": 16991 }, { "epoch": 0.6154291923216226, "grad_norm": 0.9724030321872225, "learning_rate": 6.804703968734629e-06, "loss": 0.6941, "step": 16992 }, { "epoch": 0.615465411082941, "grad_norm": 1.1254881124692862, "learning_rate": 6.803592419562228e-06, "loss": 0.7313, "step": 16993 }, { "epoch": 0.6155016298442594, "grad_norm": 1.3056115323201152, "learning_rate": 6.802480914375205e-06, "loss": 0.674, "step": 16994 }, { "epoch": 0.6155378486055777, "grad_norm": 1.454306929351454, "learning_rate": 6.801369453188851e-06, "loss": 0.6904, "step": 16995 }, { "epoch": 0.615574067366896, "grad_norm": 1.0903368110355511, "learning_rate": 6.800258036018463e-06, "loss": 0.7118, "step": 16996 }, { "epoch": 0.6156102861282144, "grad_norm": 1.3711120789001054, "learning_rate": 6.79914666287934e-06, "loss": 0.7332, "step": 16997 }, { "epoch": 0.6156465048895328, "grad_norm": 1.4795803818051754, "learning_rate": 6.798035333786769e-06, "loss": 0.7516, "step": 16998 }, { "epoch": 0.6156827236508512, "grad_norm": 1.3172415734682372, "learning_rate": 6.796924048756047e-06, "loss": 0.7027, "step": 16999 }, { "epoch": 0.6157189424121695, "grad_norm": 1.4283333303493917, "learning_rate": 6.795812807802462e-06, "loss": 0.6945, "step": 17000 }, { "epoch": 0.6157551611734878, "grad_norm": 1.3920648319027207, "learning_rate": 6.794701610941312e-06, "loss": 0.6651, "step": 17001 }, { "epoch": 0.6157913799348063, "grad_norm": 1.4633689118872324, "learning_rate": 6.793590458187881e-06, "loss": 0.6504, "step": 17002 }, { "epoch": 0.6158275986961246, "grad_norm": 1.3633940446476462, "learning_rate": 6.792479349557465e-06, "loss": 0.6777, "step": 17003 }, { "epoch": 0.615863817457443, "grad_norm": 1.0858863751637458, "learning_rate": 6.791368285065349e-06, "loss": 0.7185, "step": 17004 }, { "epoch": 0.6159000362187613, "grad_norm": 1.7372409322103726, "learning_rate": 6.790257264726827e-06, "loss": 0.7114, "step": 17005 }, { "epoch": 0.6159362549800796, "grad_norm": 1.3585107346936518, "learning_rate": 6.789146288557184e-06, "loss": 0.6883, "step": 17006 }, { "epoch": 0.6159724737413981, "grad_norm": 1.274447150549924, "learning_rate": 6.788035356571709e-06, "loss": 0.6927, "step": 17007 }, { "epoch": 0.6160086925027164, "grad_norm": 1.358394085282704, "learning_rate": 6.786924468785691e-06, "loss": 0.7806, "step": 17008 }, { "epoch": 0.6160449112640348, "grad_norm": 1.2919491085272177, "learning_rate": 6.785813625214415e-06, "loss": 0.6634, "step": 17009 }, { "epoch": 0.6160811300253531, "grad_norm": 1.309627900077057, "learning_rate": 6.784702825873168e-06, "loss": 0.7206, "step": 17010 }, { "epoch": 0.6161173487866715, "grad_norm": 1.2900264776916937, "learning_rate": 6.783592070777234e-06, "loss": 0.6568, "step": 17011 }, { "epoch": 0.6161535675479899, "grad_norm": 1.3675304078181263, "learning_rate": 6.782481359941903e-06, "loss": 0.6551, "step": 17012 }, { "epoch": 0.6161897863093082, "grad_norm": 1.3770734883181084, "learning_rate": 6.781370693382452e-06, "loss": 0.6704, "step": 17013 }, { "epoch": 0.6162260050706266, "grad_norm": 1.3147228468705794, "learning_rate": 6.7802600711141705e-06, "loss": 0.6912, "step": 17014 }, { "epoch": 0.616262223831945, "grad_norm": 1.3691861842978827, "learning_rate": 6.779149493152337e-06, "loss": 0.7335, "step": 17015 }, { "epoch": 0.6162984425932633, "grad_norm": 1.1251250278164597, "learning_rate": 6.7780389595122395e-06, "loss": 0.7286, "step": 17016 }, { "epoch": 0.6163346613545817, "grad_norm": 1.537047519432921, "learning_rate": 6.776928470209154e-06, "loss": 0.6493, "step": 17017 }, { "epoch": 0.6163708801159, "grad_norm": 0.9470696901985405, "learning_rate": 6.775818025258366e-06, "loss": 0.6885, "step": 17018 }, { "epoch": 0.6164070988772185, "grad_norm": 1.4153848505952513, "learning_rate": 6.774707624675157e-06, "loss": 0.6953, "step": 17019 }, { "epoch": 0.6164433176385368, "grad_norm": 1.247818803618454, "learning_rate": 6.773597268474802e-06, "loss": 0.6923, "step": 17020 }, { "epoch": 0.6164795363998551, "grad_norm": 1.5275111027470558, "learning_rate": 6.772486956672586e-06, "loss": 0.6771, "step": 17021 }, { "epoch": 0.6165157551611735, "grad_norm": 1.2800312424200573, "learning_rate": 6.771376689283784e-06, "loss": 0.727, "step": 17022 }, { "epoch": 0.6165519739224918, "grad_norm": 1.4235851802699875, "learning_rate": 6.770266466323677e-06, "loss": 0.6906, "step": 17023 }, { "epoch": 0.6165881926838103, "grad_norm": 1.783488082679762, "learning_rate": 6.769156287807539e-06, "loss": 0.7657, "step": 17024 }, { "epoch": 0.6166244114451286, "grad_norm": 1.2981542617477935, "learning_rate": 6.768046153750653e-06, "loss": 0.7189, "step": 17025 }, { "epoch": 0.6166606302064469, "grad_norm": 1.2939874016016384, "learning_rate": 6.7669360641682876e-06, "loss": 0.6904, "step": 17026 }, { "epoch": 0.6166968489677653, "grad_norm": 1.090458200123104, "learning_rate": 6.765826019075724e-06, "loss": 0.7421, "step": 17027 }, { "epoch": 0.6167330677290837, "grad_norm": 1.4060749713268148, "learning_rate": 6.7647160184882375e-06, "loss": 0.7606, "step": 17028 }, { "epoch": 0.6167692864904021, "grad_norm": 1.262428034133471, "learning_rate": 6.7636060624211e-06, "loss": 0.7161, "step": 17029 }, { "epoch": 0.6168055052517204, "grad_norm": 1.2382528534199428, "learning_rate": 6.7624961508895894e-06, "loss": 0.7229, "step": 17030 }, { "epoch": 0.6168417240130387, "grad_norm": 1.4007211731337261, "learning_rate": 6.761386283908973e-06, "loss": 0.7603, "step": 17031 }, { "epoch": 0.6168779427743571, "grad_norm": 1.4036275033662229, "learning_rate": 6.76027646149453e-06, "loss": 0.7309, "step": 17032 }, { "epoch": 0.6169141615356755, "grad_norm": 1.2807053038359193, "learning_rate": 6.759166683661527e-06, "loss": 0.757, "step": 17033 }, { "epoch": 0.6169503802969939, "grad_norm": 1.3026227045309862, "learning_rate": 6.758056950425241e-06, "loss": 0.658, "step": 17034 }, { "epoch": 0.6169865990583122, "grad_norm": 1.0119858201354626, "learning_rate": 6.7569472618009364e-06, "loss": 0.6573, "step": 17035 }, { "epoch": 0.6170228178196305, "grad_norm": 1.3794933979083663, "learning_rate": 6.755837617803891e-06, "loss": 0.7321, "step": 17036 }, { "epoch": 0.617059036580949, "grad_norm": 1.0136182175705055, "learning_rate": 6.754728018449368e-06, "loss": 0.7689, "step": 17037 }, { "epoch": 0.6170952553422673, "grad_norm": 0.9542456843090183, "learning_rate": 6.753618463752638e-06, "loss": 0.6624, "step": 17038 }, { "epoch": 0.6171314741035857, "grad_norm": 1.5300783277622971, "learning_rate": 6.752508953728973e-06, "loss": 0.7712, "step": 17039 }, { "epoch": 0.617167692864904, "grad_norm": 1.2953571527294363, "learning_rate": 6.751399488393637e-06, "loss": 0.6654, "step": 17040 }, { "epoch": 0.6172039116262223, "grad_norm": 1.3367834036046702, "learning_rate": 6.7502900677619e-06, "loss": 0.7318, "step": 17041 }, { "epoch": 0.6172401303875408, "grad_norm": 1.319687817574864, "learning_rate": 6.749180691849024e-06, "loss": 0.7775, "step": 17042 }, { "epoch": 0.6172763491488591, "grad_norm": 1.2869417750017236, "learning_rate": 6.748071360670281e-06, "loss": 0.7446, "step": 17043 }, { "epoch": 0.6173125679101775, "grad_norm": 1.0102353038407257, "learning_rate": 6.746962074240933e-06, "loss": 0.7072, "step": 17044 }, { "epoch": 0.6173487866714958, "grad_norm": 1.3411850638037481, "learning_rate": 6.745852832576245e-06, "loss": 0.6992, "step": 17045 }, { "epoch": 0.6173850054328142, "grad_norm": 1.0523893814151637, "learning_rate": 6.7447436356914796e-06, "loss": 0.6803, "step": 17046 }, { "epoch": 0.6174212241941326, "grad_norm": 1.3012989849600538, "learning_rate": 6.743634483601904e-06, "loss": 0.7068, "step": 17047 }, { "epoch": 0.6174574429554509, "grad_norm": 1.5404260363268902, "learning_rate": 6.74252537632278e-06, "loss": 0.7774, "step": 17048 }, { "epoch": 0.6174936617167693, "grad_norm": 1.3171621537965799, "learning_rate": 6.7414163138693655e-06, "loss": 0.6536, "step": 17049 }, { "epoch": 0.6175298804780877, "grad_norm": 1.339017064576733, "learning_rate": 6.7403072962569295e-06, "loss": 0.765, "step": 17050 }, { "epoch": 0.617566099239406, "grad_norm": 1.3093287083652614, "learning_rate": 6.739198323500727e-06, "loss": 0.6697, "step": 17051 }, { "epoch": 0.6176023180007244, "grad_norm": 1.100000300069976, "learning_rate": 6.738089395616023e-06, "loss": 0.7641, "step": 17052 }, { "epoch": 0.6176385367620427, "grad_norm": 1.3318816175065942, "learning_rate": 6.736980512618072e-06, "loss": 0.7455, "step": 17053 }, { "epoch": 0.6176747555233612, "grad_norm": 1.3425419268869863, "learning_rate": 6.73587167452214e-06, "loss": 0.7628, "step": 17054 }, { "epoch": 0.6177109742846795, "grad_norm": 1.1372736198962436, "learning_rate": 6.7347628813434784e-06, "loss": 0.7555, "step": 17055 }, { "epoch": 0.6177471930459978, "grad_norm": 1.266380346622802, "learning_rate": 6.733654133097351e-06, "loss": 0.6747, "step": 17056 }, { "epoch": 0.6177834118073162, "grad_norm": 1.1679694840937638, "learning_rate": 6.732545429799011e-06, "loss": 0.7196, "step": 17057 }, { "epoch": 0.6178196305686345, "grad_norm": 1.3477917133194826, "learning_rate": 6.731436771463718e-06, "loss": 0.7014, "step": 17058 }, { "epoch": 0.617855849329953, "grad_norm": 1.3753167906414487, "learning_rate": 6.730328158106727e-06, "loss": 0.7396, "step": 17059 }, { "epoch": 0.6178920680912713, "grad_norm": 1.4213121165934757, "learning_rate": 6.729219589743292e-06, "loss": 0.6464, "step": 17060 }, { "epoch": 0.6179282868525896, "grad_norm": 1.3841554012781292, "learning_rate": 6.728111066388673e-06, "loss": 0.6566, "step": 17061 }, { "epoch": 0.617964505613908, "grad_norm": 1.3510327727451046, "learning_rate": 6.727002588058118e-06, "loss": 0.7456, "step": 17062 }, { "epoch": 0.6180007243752264, "grad_norm": 0.9624955198369282, "learning_rate": 6.725894154766884e-06, "loss": 0.7086, "step": 17063 }, { "epoch": 0.6180369431365448, "grad_norm": 1.2805310367716989, "learning_rate": 6.724785766530222e-06, "loss": 0.7713, "step": 17064 }, { "epoch": 0.6180731618978631, "grad_norm": 0.9948026302201186, "learning_rate": 6.723677423363389e-06, "loss": 0.6809, "step": 17065 }, { "epoch": 0.6181093806591814, "grad_norm": 1.3913840237624875, "learning_rate": 6.72256912528163e-06, "loss": 0.7425, "step": 17066 }, { "epoch": 0.6181455994204998, "grad_norm": 1.374180662083834, "learning_rate": 6.7214608723002015e-06, "loss": 0.7039, "step": 17067 }, { "epoch": 0.6181818181818182, "grad_norm": 1.3613785579758544, "learning_rate": 6.720352664434351e-06, "loss": 0.7447, "step": 17068 }, { "epoch": 0.6182180369431366, "grad_norm": 1.3192102039152516, "learning_rate": 6.719244501699329e-06, "loss": 0.731, "step": 17069 }, { "epoch": 0.6182542557044549, "grad_norm": 1.4196035757728893, "learning_rate": 6.718136384110388e-06, "loss": 0.6999, "step": 17070 }, { "epoch": 0.6182904744657732, "grad_norm": 1.4180029038711337, "learning_rate": 6.717028311682771e-06, "loss": 0.6869, "step": 17071 }, { "epoch": 0.6183266932270917, "grad_norm": 1.5362815966763366, "learning_rate": 6.7159202844317325e-06, "loss": 0.7935, "step": 17072 }, { "epoch": 0.61836291198841, "grad_norm": 1.4299684942810011, "learning_rate": 6.7148123023725145e-06, "loss": 0.7822, "step": 17073 }, { "epoch": 0.6183991307497284, "grad_norm": 1.3593087728306232, "learning_rate": 6.713704365520366e-06, "loss": 0.7421, "step": 17074 }, { "epoch": 0.6184353495110467, "grad_norm": 1.342663689029719, "learning_rate": 6.712596473890533e-06, "loss": 0.6936, "step": 17075 }, { "epoch": 0.618471568272365, "grad_norm": 0.9678789619208352, "learning_rate": 6.711488627498262e-06, "loss": 0.6061, "step": 17076 }, { "epoch": 0.6185077870336835, "grad_norm": 1.3223915152862993, "learning_rate": 6.710380826358795e-06, "loss": 0.705, "step": 17077 }, { "epoch": 0.6185440057950018, "grad_norm": 1.2880387742241362, "learning_rate": 6.709273070487381e-06, "loss": 0.6958, "step": 17078 }, { "epoch": 0.6185802245563202, "grad_norm": 1.3089718251151798, "learning_rate": 6.708165359899259e-06, "loss": 0.6243, "step": 17079 }, { "epoch": 0.6186164433176385, "grad_norm": 1.4223198286569985, "learning_rate": 6.707057694609673e-06, "loss": 0.6185, "step": 17080 }, { "epoch": 0.6186526620789569, "grad_norm": 1.3952969070708134, "learning_rate": 6.70595007463387e-06, "loss": 0.7142, "step": 17081 }, { "epoch": 0.6186888808402753, "grad_norm": 1.4836482163586242, "learning_rate": 6.7048424999870855e-06, "loss": 0.7393, "step": 17082 }, { "epoch": 0.6187250996015936, "grad_norm": 1.2277509749956432, "learning_rate": 6.7037349706845656e-06, "loss": 0.6577, "step": 17083 }, { "epoch": 0.618761318362912, "grad_norm": 1.1138929338316592, "learning_rate": 6.702627486741548e-06, "loss": 0.7438, "step": 17084 }, { "epoch": 0.6187975371242304, "grad_norm": 1.3218858397436182, "learning_rate": 6.701520048173274e-06, "loss": 0.7126, "step": 17085 }, { "epoch": 0.6188337558855487, "grad_norm": 1.344116555995895, "learning_rate": 6.7004126549949814e-06, "loss": 0.7559, "step": 17086 }, { "epoch": 0.6188699746468671, "grad_norm": 1.417916112374656, "learning_rate": 6.699305307221912e-06, "loss": 0.7602, "step": 17087 }, { "epoch": 0.6189061934081854, "grad_norm": 1.2848515308733224, "learning_rate": 6.6981980048692985e-06, "loss": 0.7001, "step": 17088 }, { "epoch": 0.6189424121695039, "grad_norm": 1.4810184650904026, "learning_rate": 6.697090747952383e-06, "loss": 0.7384, "step": 17089 }, { "epoch": 0.6189786309308222, "grad_norm": 1.511293149273748, "learning_rate": 6.695983536486403e-06, "loss": 0.6777, "step": 17090 }, { "epoch": 0.6190148496921405, "grad_norm": 1.4698711736005845, "learning_rate": 6.69487637048659e-06, "loss": 0.6918, "step": 17091 }, { "epoch": 0.6190510684534589, "grad_norm": 1.0239023693466112, "learning_rate": 6.693769249968185e-06, "loss": 0.6995, "step": 17092 }, { "epoch": 0.6190872872147772, "grad_norm": 1.397089882938493, "learning_rate": 6.692662174946417e-06, "loss": 0.6953, "step": 17093 }, { "epoch": 0.6191235059760957, "grad_norm": 1.6094089690808564, "learning_rate": 6.691555145436526e-06, "loss": 0.6945, "step": 17094 }, { "epoch": 0.619159724737414, "grad_norm": 1.3248871295241265, "learning_rate": 6.690448161453742e-06, "loss": 0.6754, "step": 17095 }, { "epoch": 0.6191959434987323, "grad_norm": 1.4506287970476777, "learning_rate": 6.689341223013302e-06, "loss": 0.716, "step": 17096 }, { "epoch": 0.6192321622600507, "grad_norm": 1.3368908606355696, "learning_rate": 6.688234330130432e-06, "loss": 0.7319, "step": 17097 }, { "epoch": 0.619268381021369, "grad_norm": 1.1624916119719113, "learning_rate": 6.68712748282037e-06, "loss": 0.7594, "step": 17098 }, { "epoch": 0.6193045997826875, "grad_norm": 1.1230732106666101, "learning_rate": 6.686020681098343e-06, "loss": 0.7324, "step": 17099 }, { "epoch": 0.6193408185440058, "grad_norm": 1.260682914250624, "learning_rate": 6.684913924979584e-06, "loss": 0.7232, "step": 17100 }, { "epoch": 0.6193770373053241, "grad_norm": 1.310018661806081, "learning_rate": 6.683807214479323e-06, "loss": 0.6638, "step": 17101 }, { "epoch": 0.6194132560666425, "grad_norm": 1.4884554042696396, "learning_rate": 6.6827005496127874e-06, "loss": 0.7227, "step": 17102 }, { "epoch": 0.6194494748279609, "grad_norm": 1.3214401023455542, "learning_rate": 6.681593930395209e-06, "loss": 0.7411, "step": 17103 }, { "epoch": 0.6194856935892793, "grad_norm": 1.3698365637150287, "learning_rate": 6.680487356841812e-06, "loss": 0.7344, "step": 17104 }, { "epoch": 0.6195219123505976, "grad_norm": 1.327506610714487, "learning_rate": 6.679380828967828e-06, "loss": 0.6874, "step": 17105 }, { "epoch": 0.6195581311119159, "grad_norm": 1.3418126785535933, "learning_rate": 6.678274346788478e-06, "loss": 0.6238, "step": 17106 }, { "epoch": 0.6195943498732344, "grad_norm": 1.2951648133296583, "learning_rate": 6.677167910318996e-06, "loss": 0.7176, "step": 17107 }, { "epoch": 0.6196305686345527, "grad_norm": 1.4208536783358925, "learning_rate": 6.676061519574599e-06, "loss": 0.6998, "step": 17108 }, { "epoch": 0.6196667873958711, "grad_norm": 1.2730546848341995, "learning_rate": 6.67495517457052e-06, "loss": 0.7064, "step": 17109 }, { "epoch": 0.6197030061571894, "grad_norm": 1.043480613169353, "learning_rate": 6.673848875321977e-06, "loss": 0.6751, "step": 17110 }, { "epoch": 0.6197392249185077, "grad_norm": 1.5634783203334326, "learning_rate": 6.6727426218441945e-06, "loss": 0.7036, "step": 17111 }, { "epoch": 0.6197754436798262, "grad_norm": 1.3594092551113743, "learning_rate": 6.671636414152401e-06, "loss": 0.6902, "step": 17112 }, { "epoch": 0.6198116624411445, "grad_norm": 1.3910799936346856, "learning_rate": 6.670530252261813e-06, "loss": 0.8071, "step": 17113 }, { "epoch": 0.6198478812024629, "grad_norm": 1.3589259240337213, "learning_rate": 6.6694241361876546e-06, "loss": 0.6908, "step": 17114 }, { "epoch": 0.6198840999637812, "grad_norm": 1.3008403764638328, "learning_rate": 6.668318065945146e-06, "loss": 0.7666, "step": 17115 }, { "epoch": 0.6199203187250996, "grad_norm": 1.3815384237870316, "learning_rate": 6.66721204154951e-06, "loss": 0.7304, "step": 17116 }, { "epoch": 0.619956537486418, "grad_norm": 1.511025678966965, "learning_rate": 6.666106063015962e-06, "loss": 0.7469, "step": 17117 }, { "epoch": 0.6199927562477363, "grad_norm": 1.3284482786880512, "learning_rate": 6.665000130359726e-06, "loss": 0.6529, "step": 17118 }, { "epoch": 0.6200289750090547, "grad_norm": 1.3707643932511404, "learning_rate": 6.663894243596018e-06, "loss": 0.7228, "step": 17119 }, { "epoch": 0.6200651937703731, "grad_norm": 1.5241744520885987, "learning_rate": 6.662788402740057e-06, "loss": 0.777, "step": 17120 }, { "epoch": 0.6201014125316914, "grad_norm": 1.3823780317487135, "learning_rate": 6.661682607807061e-06, "loss": 0.75, "step": 17121 }, { "epoch": 0.6201376312930098, "grad_norm": 1.3338312295718817, "learning_rate": 6.660576858812242e-06, "loss": 0.6871, "step": 17122 }, { "epoch": 0.6201738500543281, "grad_norm": 1.4729642394201046, "learning_rate": 6.659471155770824e-06, "loss": 0.6976, "step": 17123 }, { "epoch": 0.6202100688156466, "grad_norm": 1.3487678366322207, "learning_rate": 6.658365498698015e-06, "loss": 0.6736, "step": 17124 }, { "epoch": 0.6202462875769649, "grad_norm": 1.316505989376967, "learning_rate": 6.657259887609036e-06, "loss": 0.6412, "step": 17125 }, { "epoch": 0.6202825063382832, "grad_norm": 1.4613201716251722, "learning_rate": 6.656154322519095e-06, "loss": 0.673, "step": 17126 }, { "epoch": 0.6203187250996016, "grad_norm": 1.5315449160613535, "learning_rate": 6.65504880344341e-06, "loss": 0.7626, "step": 17127 }, { "epoch": 0.6203549438609199, "grad_norm": 1.2766324878835058, "learning_rate": 6.653943330397193e-06, "loss": 0.7095, "step": 17128 }, { "epoch": 0.6203911626222384, "grad_norm": 1.360043392889409, "learning_rate": 6.652837903395655e-06, "loss": 0.8272, "step": 17129 }, { "epoch": 0.6204273813835567, "grad_norm": 0.9660738085837728, "learning_rate": 6.651732522454008e-06, "loss": 0.7262, "step": 17130 }, { "epoch": 0.620463600144875, "grad_norm": 1.2203606392112065, "learning_rate": 6.650627187587464e-06, "loss": 0.6537, "step": 17131 }, { "epoch": 0.6204998189061934, "grad_norm": 1.4280492393653137, "learning_rate": 6.649521898811234e-06, "loss": 0.7081, "step": 17132 }, { "epoch": 0.6205360376675118, "grad_norm": 1.3343552273596382, "learning_rate": 6.648416656140525e-06, "loss": 0.7206, "step": 17133 }, { "epoch": 0.6205722564288302, "grad_norm": 1.3974197282986316, "learning_rate": 6.6473114595905506e-06, "loss": 0.7568, "step": 17134 }, { "epoch": 0.6206084751901485, "grad_norm": 1.3022045125998882, "learning_rate": 6.646206309176513e-06, "loss": 0.6747, "step": 17135 }, { "epoch": 0.6206446939514668, "grad_norm": 1.129185702673879, "learning_rate": 6.6451012049136265e-06, "loss": 0.667, "step": 17136 }, { "epoch": 0.6206809127127852, "grad_norm": 1.275311674241862, "learning_rate": 6.643996146817093e-06, "loss": 0.695, "step": 17137 }, { "epoch": 0.6207171314741036, "grad_norm": 1.5603628599125796, "learning_rate": 6.642891134902123e-06, "loss": 0.6814, "step": 17138 }, { "epoch": 0.620753350235422, "grad_norm": 1.3640262677601311, "learning_rate": 6.6417861691839194e-06, "loss": 0.7421, "step": 17139 }, { "epoch": 0.6207895689967403, "grad_norm": 1.5336883700345898, "learning_rate": 6.640681249677691e-06, "loss": 0.7602, "step": 17140 }, { "epoch": 0.6208257877580586, "grad_norm": 1.4399665863950377, "learning_rate": 6.639576376398637e-06, "loss": 0.7186, "step": 17141 }, { "epoch": 0.6208620065193771, "grad_norm": 1.3449217807682212, "learning_rate": 6.638471549361967e-06, "loss": 0.7197, "step": 17142 }, { "epoch": 0.6208982252806954, "grad_norm": 1.2139943747510902, "learning_rate": 6.637366768582882e-06, "loss": 0.6686, "step": 17143 }, { "epoch": 0.6209344440420138, "grad_norm": 1.4252373480978828, "learning_rate": 6.636262034076584e-06, "loss": 0.6369, "step": 17144 }, { "epoch": 0.6209706628033321, "grad_norm": 1.482743575129243, "learning_rate": 6.635157345858278e-06, "loss": 0.6729, "step": 17145 }, { "epoch": 0.6210068815646504, "grad_norm": 1.3611035975937624, "learning_rate": 6.634052703943162e-06, "loss": 0.6515, "step": 17146 }, { "epoch": 0.6210431003259689, "grad_norm": 1.346220134474232, "learning_rate": 6.63294810834644e-06, "loss": 0.6929, "step": 17147 }, { "epoch": 0.6210793190872872, "grad_norm": 1.1231376576271357, "learning_rate": 6.631843559083309e-06, "loss": 0.6565, "step": 17148 }, { "epoch": 0.6211155378486056, "grad_norm": 1.3573155917266697, "learning_rate": 6.630739056168971e-06, "loss": 0.6481, "step": 17149 }, { "epoch": 0.6211517566099239, "grad_norm": 1.042119521551924, "learning_rate": 6.629634599618622e-06, "loss": 0.7033, "step": 17150 }, { "epoch": 0.6211879753712423, "grad_norm": 1.2899675923402139, "learning_rate": 6.628530189447463e-06, "loss": 0.7175, "step": 17151 }, { "epoch": 0.6212241941325607, "grad_norm": 1.2728414932573846, "learning_rate": 6.627425825670693e-06, "loss": 0.6739, "step": 17152 }, { "epoch": 0.621260412893879, "grad_norm": 1.3344256277273565, "learning_rate": 6.6263215083035055e-06, "loss": 0.7693, "step": 17153 }, { "epoch": 0.6212966316551974, "grad_norm": 1.2689527220448769, "learning_rate": 6.625217237361099e-06, "loss": 0.6824, "step": 17154 }, { "epoch": 0.6213328504165158, "grad_norm": 1.5702013479341188, "learning_rate": 6.624113012858669e-06, "loss": 0.646, "step": 17155 }, { "epoch": 0.6213690691778341, "grad_norm": 1.339154854608686, "learning_rate": 6.6230088348114106e-06, "loss": 0.7176, "step": 17156 }, { "epoch": 0.6214052879391525, "grad_norm": 1.0299263284694804, "learning_rate": 6.621904703234516e-06, "loss": 0.6743, "step": 17157 }, { "epoch": 0.6214415067004708, "grad_norm": 1.4555052175268623, "learning_rate": 6.620800618143184e-06, "loss": 0.744, "step": 17158 }, { "epoch": 0.6214777254617893, "grad_norm": 1.1959704898435293, "learning_rate": 6.619696579552602e-06, "loss": 0.6403, "step": 17159 }, { "epoch": 0.6215139442231076, "grad_norm": 1.6876660878958052, "learning_rate": 6.618592587477966e-06, "loss": 0.7372, "step": 17160 }, { "epoch": 0.6215501629844259, "grad_norm": 1.231659471281145, "learning_rate": 6.617488641934467e-06, "loss": 0.6206, "step": 17161 }, { "epoch": 0.6215863817457443, "grad_norm": 1.3089186300696571, "learning_rate": 6.616384742937295e-06, "loss": 0.724, "step": 17162 }, { "epoch": 0.6216226005070626, "grad_norm": 1.0097768515358445, "learning_rate": 6.615280890501645e-06, "loss": 0.7116, "step": 17163 }, { "epoch": 0.6216588192683811, "grad_norm": 1.4395615382886284, "learning_rate": 6.614177084642701e-06, "loss": 0.7311, "step": 17164 }, { "epoch": 0.6216950380296994, "grad_norm": 1.3743800084919242, "learning_rate": 6.613073325375658e-06, "loss": 0.7492, "step": 17165 }, { "epoch": 0.6217312567910177, "grad_norm": 1.2810667462253207, "learning_rate": 6.6119696127157e-06, "loss": 0.6533, "step": 17166 }, { "epoch": 0.6217674755523361, "grad_norm": 1.307981851196798, "learning_rate": 6.61086594667802e-06, "loss": 0.6715, "step": 17167 }, { "epoch": 0.6218036943136545, "grad_norm": 1.3205413838635547, "learning_rate": 6.6097623272778e-06, "loss": 0.7282, "step": 17168 }, { "epoch": 0.6218399130749729, "grad_norm": 1.3155348756246201, "learning_rate": 6.608658754530232e-06, "loss": 0.6737, "step": 17169 }, { "epoch": 0.6218761318362912, "grad_norm": 1.3184490349975446, "learning_rate": 6.607555228450496e-06, "loss": 0.674, "step": 17170 }, { "epoch": 0.6219123505976095, "grad_norm": 1.442195619976664, "learning_rate": 6.606451749053784e-06, "loss": 0.6553, "step": 17171 }, { "epoch": 0.621948569358928, "grad_norm": 1.3413899115487768, "learning_rate": 6.60534831635528e-06, "loss": 0.6962, "step": 17172 }, { "epoch": 0.6219847881202463, "grad_norm": 1.388285066624055, "learning_rate": 6.6042449303701605e-06, "loss": 0.702, "step": 17173 }, { "epoch": 0.6220210068815647, "grad_norm": 0.9752442841130328, "learning_rate": 6.603141591113615e-06, "loss": 0.7117, "step": 17174 }, { "epoch": 0.622057225642883, "grad_norm": 1.3556430056032645, "learning_rate": 6.602038298600828e-06, "loss": 0.6883, "step": 17175 }, { "epoch": 0.6220934444042013, "grad_norm": 1.2588315842625397, "learning_rate": 6.600935052846978e-06, "loss": 0.6906, "step": 17176 }, { "epoch": 0.6221296631655198, "grad_norm": 1.2773781415993275, "learning_rate": 6.599831853867251e-06, "loss": 0.7411, "step": 17177 }, { "epoch": 0.6221658819268381, "grad_norm": 1.092197458203438, "learning_rate": 6.598728701676822e-06, "loss": 0.6942, "step": 17178 }, { "epoch": 0.6222021006881565, "grad_norm": 1.2111292771790876, "learning_rate": 6.597625596290877e-06, "loss": 0.7104, "step": 17179 }, { "epoch": 0.6222383194494748, "grad_norm": 1.3405190161958176, "learning_rate": 6.596522537724592e-06, "loss": 0.7334, "step": 17180 }, { "epoch": 0.6222745382107931, "grad_norm": 1.1359473965169788, "learning_rate": 6.5954195259931495e-06, "loss": 0.6877, "step": 17181 }, { "epoch": 0.6223107569721116, "grad_norm": 1.0527994192408556, "learning_rate": 6.5943165611117244e-06, "loss": 0.7487, "step": 17182 }, { "epoch": 0.6223469757334299, "grad_norm": 1.3556434123053824, "learning_rate": 6.5932136430954965e-06, "loss": 0.7938, "step": 17183 }, { "epoch": 0.6223831944947483, "grad_norm": 1.2910876434368268, "learning_rate": 6.592110771959641e-06, "loss": 0.7891, "step": 17184 }, { "epoch": 0.6224194132560666, "grad_norm": 1.4296171178847281, "learning_rate": 6.5910079477193345e-06, "loss": 0.6651, "step": 17185 }, { "epoch": 0.622455632017385, "grad_norm": 1.3816568966318026, "learning_rate": 6.589905170389757e-06, "loss": 0.7062, "step": 17186 }, { "epoch": 0.6224918507787034, "grad_norm": 1.5132733811078871, "learning_rate": 6.58880243998608e-06, "loss": 0.8025, "step": 17187 }, { "epoch": 0.6225280695400217, "grad_norm": 1.345583261901435, "learning_rate": 6.587699756523479e-06, "loss": 0.6565, "step": 17188 }, { "epoch": 0.6225642883013401, "grad_norm": 1.3193063534992335, "learning_rate": 6.5865971200171265e-06, "loss": 0.6904, "step": 17189 }, { "epoch": 0.6226005070626585, "grad_norm": 1.2278350675759697, "learning_rate": 6.585494530482199e-06, "loss": 0.7659, "step": 17190 }, { "epoch": 0.6226367258239768, "grad_norm": 1.241221847306153, "learning_rate": 6.584391987933865e-06, "loss": 0.6575, "step": 17191 }, { "epoch": 0.6226729445852952, "grad_norm": 1.5834908545610613, "learning_rate": 6.5832894923873e-06, "loss": 0.7506, "step": 17192 }, { "epoch": 0.6227091633466135, "grad_norm": 1.42202694573572, "learning_rate": 6.582187043857671e-06, "loss": 0.6553, "step": 17193 }, { "epoch": 0.622745382107932, "grad_norm": 1.5827466289185803, "learning_rate": 6.581084642360153e-06, "loss": 0.7657, "step": 17194 }, { "epoch": 0.6227816008692503, "grad_norm": 1.2007978609123748, "learning_rate": 6.579982287909916e-06, "loss": 0.6276, "step": 17195 }, { "epoch": 0.6228178196305686, "grad_norm": 1.322890793148096, "learning_rate": 6.578879980522125e-06, "loss": 0.7319, "step": 17196 }, { "epoch": 0.622854038391887, "grad_norm": 1.3156059436624228, "learning_rate": 6.577777720211954e-06, "loss": 0.6738, "step": 17197 }, { "epoch": 0.6228902571532053, "grad_norm": 1.2484567106231648, "learning_rate": 6.576675506994567e-06, "loss": 0.7745, "step": 17198 }, { "epoch": 0.6229264759145238, "grad_norm": 1.3071433577025449, "learning_rate": 6.5755733408851355e-06, "loss": 0.6972, "step": 17199 }, { "epoch": 0.6229626946758421, "grad_norm": 1.2962198785870616, "learning_rate": 6.57447122189882e-06, "loss": 0.6632, "step": 17200 }, { "epoch": 0.6229989134371604, "grad_norm": 1.2661608051894861, "learning_rate": 6.573369150050794e-06, "loss": 0.7327, "step": 17201 }, { "epoch": 0.6230351321984788, "grad_norm": 1.0131148824028093, "learning_rate": 6.572267125356216e-06, "loss": 0.7096, "step": 17202 }, { "epoch": 0.6230713509597972, "grad_norm": 1.0509670918718865, "learning_rate": 6.571165147830257e-06, "loss": 0.7264, "step": 17203 }, { "epoch": 0.6231075697211156, "grad_norm": 1.7870788321464681, "learning_rate": 6.570063217488076e-06, "loss": 0.7517, "step": 17204 }, { "epoch": 0.6231437884824339, "grad_norm": 1.3552864004269172, "learning_rate": 6.568961334344841e-06, "loss": 0.7246, "step": 17205 }, { "epoch": 0.6231800072437522, "grad_norm": 1.4268911026581983, "learning_rate": 6.567859498415713e-06, "loss": 0.7477, "step": 17206 }, { "epoch": 0.6232162260050707, "grad_norm": 1.015735090460386, "learning_rate": 6.566757709715852e-06, "loss": 0.7525, "step": 17207 }, { "epoch": 0.623252444766389, "grad_norm": 1.720955686187383, "learning_rate": 6.565655968260424e-06, "loss": 0.6735, "step": 17208 }, { "epoch": 0.6232886635277074, "grad_norm": 1.6832157781678223, "learning_rate": 6.564554274064585e-06, "loss": 0.7254, "step": 17209 }, { "epoch": 0.6233248822890257, "grad_norm": 1.2710024649965597, "learning_rate": 6.5634526271435e-06, "loss": 0.6818, "step": 17210 }, { "epoch": 0.623361101050344, "grad_norm": 1.1412995392610839, "learning_rate": 6.562351027512326e-06, "loss": 0.7093, "step": 17211 }, { "epoch": 0.6233973198116625, "grad_norm": 1.3480302816832725, "learning_rate": 6.561249475186222e-06, "loss": 0.7549, "step": 17212 }, { "epoch": 0.6234335385729808, "grad_norm": 1.342135825580728, "learning_rate": 6.560147970180346e-06, "loss": 0.7964, "step": 17213 }, { "epoch": 0.6234697573342992, "grad_norm": 1.3324139756958562, "learning_rate": 6.559046512509859e-06, "loss": 0.6831, "step": 17214 }, { "epoch": 0.6235059760956175, "grad_norm": 1.4184034064919386, "learning_rate": 6.557945102189913e-06, "loss": 0.8228, "step": 17215 }, { "epoch": 0.6235421948569358, "grad_norm": 1.237832276813024, "learning_rate": 6.556843739235667e-06, "loss": 0.6466, "step": 17216 }, { "epoch": 0.6235784136182543, "grad_norm": 0.9904931151146484, "learning_rate": 6.555742423662278e-06, "loss": 0.6896, "step": 17217 }, { "epoch": 0.6236146323795726, "grad_norm": 1.6036371662434916, "learning_rate": 6.554641155484898e-06, "loss": 0.6988, "step": 17218 }, { "epoch": 0.623650851140891, "grad_norm": 1.3343430884362375, "learning_rate": 6.553539934718685e-06, "loss": 0.7431, "step": 17219 }, { "epoch": 0.6236870699022093, "grad_norm": 1.3348030640205226, "learning_rate": 6.5524387613787885e-06, "loss": 0.6664, "step": 17220 }, { "epoch": 0.6237232886635277, "grad_norm": 1.3670768876970432, "learning_rate": 6.551337635480366e-06, "loss": 0.7541, "step": 17221 }, { "epoch": 0.6237595074248461, "grad_norm": 1.4146192202934111, "learning_rate": 6.550236557038566e-06, "loss": 0.7278, "step": 17222 }, { "epoch": 0.6237957261861644, "grad_norm": 1.4421720176394752, "learning_rate": 6.5491355260685445e-06, "loss": 0.7514, "step": 17223 }, { "epoch": 0.6238319449474828, "grad_norm": 1.3988406276916452, "learning_rate": 6.548034542585449e-06, "loss": 0.7117, "step": 17224 }, { "epoch": 0.6238681637088012, "grad_norm": 1.3427909156355382, "learning_rate": 6.546933606604431e-06, "loss": 0.6926, "step": 17225 }, { "epoch": 0.6239043824701195, "grad_norm": 1.6797081807438075, "learning_rate": 6.545832718140641e-06, "loss": 0.7315, "step": 17226 }, { "epoch": 0.6239406012314379, "grad_norm": 1.4068947055949943, "learning_rate": 6.544731877209227e-06, "loss": 0.7543, "step": 17227 }, { "epoch": 0.6239768199927562, "grad_norm": 1.312516413526425, "learning_rate": 6.543631083825342e-06, "loss": 0.7566, "step": 17228 }, { "epoch": 0.6240130387540747, "grad_norm": 1.303935016705596, "learning_rate": 6.542530338004126e-06, "loss": 0.5655, "step": 17229 }, { "epoch": 0.624049257515393, "grad_norm": 1.3443501533655609, "learning_rate": 6.541429639760733e-06, "loss": 0.8082, "step": 17230 }, { "epoch": 0.6240854762767113, "grad_norm": 1.3609424726196084, "learning_rate": 6.5403289891103065e-06, "loss": 0.72, "step": 17231 }, { "epoch": 0.6241216950380297, "grad_norm": 1.3692839958286924, "learning_rate": 6.539228386067994e-06, "loss": 0.7186, "step": 17232 }, { "epoch": 0.624157913799348, "grad_norm": 1.3256838817842007, "learning_rate": 6.538127830648937e-06, "loss": 0.7138, "step": 17233 }, { "epoch": 0.6241941325606665, "grad_norm": 1.1490708959447542, "learning_rate": 6.537027322868287e-06, "loss": 0.6339, "step": 17234 }, { "epoch": 0.6242303513219848, "grad_norm": 1.292057876906499, "learning_rate": 6.53592686274118e-06, "loss": 0.7095, "step": 17235 }, { "epoch": 0.6242665700833031, "grad_norm": 1.606732645977249, "learning_rate": 6.534826450282764e-06, "loss": 0.6585, "step": 17236 }, { "epoch": 0.6243027888446215, "grad_norm": 0.973661225745123, "learning_rate": 6.533726085508183e-06, "loss": 0.7425, "step": 17237 }, { "epoch": 0.6243390076059399, "grad_norm": 1.2100617846449266, "learning_rate": 6.532625768432575e-06, "loss": 0.745, "step": 17238 }, { "epoch": 0.6243752263672583, "grad_norm": 1.227214357559648, "learning_rate": 6.531525499071084e-06, "loss": 0.6872, "step": 17239 }, { "epoch": 0.6244114451285766, "grad_norm": 1.0393750122824366, "learning_rate": 6.530425277438849e-06, "loss": 0.7209, "step": 17240 }, { "epoch": 0.6244476638898949, "grad_norm": 1.2930947189106632, "learning_rate": 6.529325103551012e-06, "loss": 0.6432, "step": 17241 }, { "epoch": 0.6244838826512134, "grad_norm": 1.332237821933226, "learning_rate": 6.528224977422709e-06, "loss": 0.6665, "step": 17242 }, { "epoch": 0.6245201014125317, "grad_norm": 1.4997886133754603, "learning_rate": 6.5271248990690835e-06, "loss": 0.7513, "step": 17243 }, { "epoch": 0.6245563201738501, "grad_norm": 1.4651918506273771, "learning_rate": 6.526024868505269e-06, "loss": 0.7579, "step": 17244 }, { "epoch": 0.6245925389351684, "grad_norm": 1.3449764268898396, "learning_rate": 6.524924885746407e-06, "loss": 0.7788, "step": 17245 }, { "epoch": 0.6246287576964867, "grad_norm": 1.4139829374159878, "learning_rate": 6.523824950807629e-06, "loss": 0.7977, "step": 17246 }, { "epoch": 0.6246649764578052, "grad_norm": 1.3358692952084497, "learning_rate": 6.522725063704074e-06, "loss": 0.6823, "step": 17247 }, { "epoch": 0.6247011952191235, "grad_norm": 1.426281848973123, "learning_rate": 6.52162522445088e-06, "loss": 0.783, "step": 17248 }, { "epoch": 0.6247374139804419, "grad_norm": 1.347911181732301, "learning_rate": 6.520525433063178e-06, "loss": 0.7725, "step": 17249 }, { "epoch": 0.6247736327417602, "grad_norm": 0.9766631843400564, "learning_rate": 6.519425689556104e-06, "loss": 0.7331, "step": 17250 }, { "epoch": 0.6248098515030786, "grad_norm": 1.2914729138064744, "learning_rate": 6.518325993944792e-06, "loss": 0.6671, "step": 17251 }, { "epoch": 0.624846070264397, "grad_norm": 1.3601745769661757, "learning_rate": 6.517226346244372e-06, "loss": 0.7605, "step": 17252 }, { "epoch": 0.6248822890257153, "grad_norm": 1.367900798554323, "learning_rate": 6.5161267464699775e-06, "loss": 0.7598, "step": 17253 }, { "epoch": 0.6249185077870337, "grad_norm": 0.9728942644448095, "learning_rate": 6.515027194636741e-06, "loss": 0.6754, "step": 17254 }, { "epoch": 0.624954726548352, "grad_norm": 1.3724869648669653, "learning_rate": 6.513927690759792e-06, "loss": 0.7214, "step": 17255 }, { "epoch": 0.6249909453096704, "grad_norm": 1.2928602390624586, "learning_rate": 6.51282823485426e-06, "loss": 0.6366, "step": 17256 }, { "epoch": 0.6250271640709888, "grad_norm": 1.3116010286467727, "learning_rate": 6.511728826935278e-06, "loss": 0.726, "step": 17257 }, { "epoch": 0.6250633828323071, "grad_norm": 1.4902025446860907, "learning_rate": 6.51062946701797e-06, "loss": 0.6799, "step": 17258 }, { "epoch": 0.6250996015936255, "grad_norm": 1.384779952907284, "learning_rate": 6.509530155117469e-06, "loss": 0.6941, "step": 17259 }, { "epoch": 0.6251358203549439, "grad_norm": 1.5069759078911387, "learning_rate": 6.508430891248898e-06, "loss": 0.7152, "step": 17260 }, { "epoch": 0.6251720391162622, "grad_norm": 1.2312627380215644, "learning_rate": 6.507331675427388e-06, "loss": 0.6378, "step": 17261 }, { "epoch": 0.6252082578775806, "grad_norm": 1.4236202232106698, "learning_rate": 6.50623250766806e-06, "loss": 0.7117, "step": 17262 }, { "epoch": 0.6252444766388989, "grad_norm": 1.4290716878738279, "learning_rate": 6.505133387986045e-06, "loss": 0.7427, "step": 17263 }, { "epoch": 0.6252806954002174, "grad_norm": 1.3822166161140899, "learning_rate": 6.504034316396465e-06, "loss": 0.6636, "step": 17264 }, { "epoch": 0.6253169141615357, "grad_norm": 1.3014566734253867, "learning_rate": 6.5029352929144455e-06, "loss": 0.7759, "step": 17265 }, { "epoch": 0.625353132922854, "grad_norm": 0.9530298886542198, "learning_rate": 6.501836317555106e-06, "loss": 0.7576, "step": 17266 }, { "epoch": 0.6253893516841724, "grad_norm": 1.2184548234503005, "learning_rate": 6.500737390333573e-06, "loss": 0.6853, "step": 17267 }, { "epoch": 0.6254255704454907, "grad_norm": 1.3316217027809747, "learning_rate": 6.4996385112649714e-06, "loss": 0.7031, "step": 17268 }, { "epoch": 0.6254617892068092, "grad_norm": 1.283298373497682, "learning_rate": 6.498539680364418e-06, "loss": 0.6881, "step": 17269 }, { "epoch": 0.6254980079681275, "grad_norm": 0.89022456904562, "learning_rate": 6.497440897647036e-06, "loss": 0.6718, "step": 17270 }, { "epoch": 0.6255342267294458, "grad_norm": 1.300199203032692, "learning_rate": 6.496342163127943e-06, "loss": 0.7045, "step": 17271 }, { "epoch": 0.6255704454907642, "grad_norm": 1.310011460134923, "learning_rate": 6.495243476822264e-06, "loss": 0.6256, "step": 17272 }, { "epoch": 0.6256066642520826, "grad_norm": 1.415830408019011, "learning_rate": 6.49414483874511e-06, "loss": 0.7473, "step": 17273 }, { "epoch": 0.625642883013401, "grad_norm": 0.9606879698882104, "learning_rate": 6.493046248911606e-06, "loss": 0.6455, "step": 17274 }, { "epoch": 0.6256791017747193, "grad_norm": 1.003395782988841, "learning_rate": 6.491947707336865e-06, "loss": 0.7166, "step": 17275 }, { "epoch": 0.6257153205360376, "grad_norm": 1.322648285330886, "learning_rate": 6.4908492140360105e-06, "loss": 0.7056, "step": 17276 }, { "epoch": 0.625751539297356, "grad_norm": 0.9590333694644608, "learning_rate": 6.489750769024149e-06, "loss": 0.6891, "step": 17277 }, { "epoch": 0.6257877580586744, "grad_norm": 1.0280726923506607, "learning_rate": 6.488652372316403e-06, "loss": 0.6731, "step": 17278 }, { "epoch": 0.6258239768199928, "grad_norm": 1.1546000192621013, "learning_rate": 6.487554023927887e-06, "loss": 0.7975, "step": 17279 }, { "epoch": 0.6258601955813111, "grad_norm": 1.2874020902898475, "learning_rate": 6.486455723873713e-06, "loss": 0.6929, "step": 17280 }, { "epoch": 0.6258964143426294, "grad_norm": 1.2526603886530112, "learning_rate": 6.485357472168996e-06, "loss": 0.727, "step": 17281 }, { "epoch": 0.6259326331039479, "grad_norm": 1.5198364423262603, "learning_rate": 6.484259268828847e-06, "loss": 0.7178, "step": 17282 }, { "epoch": 0.6259688518652662, "grad_norm": 1.0255832293682998, "learning_rate": 6.4831611138683815e-06, "loss": 0.7383, "step": 17283 }, { "epoch": 0.6260050706265846, "grad_norm": 1.26344192202061, "learning_rate": 6.482063007302708e-06, "loss": 0.7256, "step": 17284 }, { "epoch": 0.6260412893879029, "grad_norm": 1.266935022263256, "learning_rate": 6.480964949146939e-06, "loss": 0.6739, "step": 17285 }, { "epoch": 0.6260775081492213, "grad_norm": 1.1738053378536997, "learning_rate": 6.479866939416184e-06, "loss": 0.7669, "step": 17286 }, { "epoch": 0.6261137269105397, "grad_norm": 1.3587062371790142, "learning_rate": 6.478768978125554e-06, "loss": 0.7854, "step": 17287 }, { "epoch": 0.626149945671858, "grad_norm": 1.8601420447244934, "learning_rate": 6.4776710652901564e-06, "loss": 0.7201, "step": 17288 }, { "epoch": 0.6261861644331764, "grad_norm": 1.195215040484759, "learning_rate": 6.476573200925099e-06, "loss": 0.6409, "step": 17289 }, { "epoch": 0.6262223831944947, "grad_norm": 1.3155618436960688, "learning_rate": 6.4754753850454925e-06, "loss": 0.7816, "step": 17290 }, { "epoch": 0.6262586019558131, "grad_norm": 1.4120782384433928, "learning_rate": 6.4743776176664386e-06, "loss": 0.6633, "step": 17291 }, { "epoch": 0.6262948207171315, "grad_norm": 1.2814931871320157, "learning_rate": 6.47327989880305e-06, "loss": 0.7129, "step": 17292 }, { "epoch": 0.6263310394784498, "grad_norm": 1.4218391345814083, "learning_rate": 6.472182228470426e-06, "loss": 0.6625, "step": 17293 }, { "epoch": 0.6263672582397682, "grad_norm": 1.341925103364201, "learning_rate": 6.471084606683678e-06, "loss": 0.6848, "step": 17294 }, { "epoch": 0.6264034770010866, "grad_norm": 1.23356136106199, "learning_rate": 6.4699870334579015e-06, "loss": 0.7296, "step": 17295 }, { "epoch": 0.6264396957624049, "grad_norm": 1.3950504797579182, "learning_rate": 6.46888950880821e-06, "loss": 0.6858, "step": 17296 }, { "epoch": 0.6264759145237233, "grad_norm": 1.0629899394460913, "learning_rate": 6.467792032749698e-06, "loss": 0.7399, "step": 17297 }, { "epoch": 0.6265121332850416, "grad_norm": 1.3969078315593124, "learning_rate": 6.466694605297471e-06, "loss": 0.7126, "step": 17298 }, { "epoch": 0.6265483520463601, "grad_norm": 1.236785064492923, "learning_rate": 6.465597226466633e-06, "loss": 0.6187, "step": 17299 }, { "epoch": 0.6265845708076784, "grad_norm": 1.4694308151307072, "learning_rate": 6.4644998962722814e-06, "loss": 0.6666, "step": 17300 }, { "epoch": 0.6266207895689967, "grad_norm": 1.3358362463453617, "learning_rate": 6.46340261472952e-06, "loss": 0.7431, "step": 17301 }, { "epoch": 0.6266570083303151, "grad_norm": 1.0186117060045714, "learning_rate": 6.462305381853444e-06, "loss": 0.6956, "step": 17302 }, { "epoch": 0.6266932270916334, "grad_norm": 0.9696580379670121, "learning_rate": 6.461208197659157e-06, "loss": 0.7278, "step": 17303 }, { "epoch": 0.6267294458529519, "grad_norm": 1.1021228253294866, "learning_rate": 6.460111062161751e-06, "loss": 0.7151, "step": 17304 }, { "epoch": 0.6267656646142702, "grad_norm": 1.2909777891123277, "learning_rate": 6.459013975376331e-06, "loss": 0.6858, "step": 17305 }, { "epoch": 0.6268018833755885, "grad_norm": 1.1843879642695478, "learning_rate": 6.457916937317986e-06, "loss": 0.6891, "step": 17306 }, { "epoch": 0.6268381021369069, "grad_norm": 1.1143871612804215, "learning_rate": 6.45681994800182e-06, "loss": 0.7539, "step": 17307 }, { "epoch": 0.6268743208982253, "grad_norm": 1.3819332243569986, "learning_rate": 6.455723007442922e-06, "loss": 0.8355, "step": 17308 }, { "epoch": 0.6269105396595437, "grad_norm": 5.0658963747598404, "learning_rate": 6.45462611565639e-06, "loss": 0.7018, "step": 17309 }, { "epoch": 0.626946758420862, "grad_norm": 0.9893623203999189, "learning_rate": 6.453529272657319e-06, "loss": 0.7171, "step": 17310 }, { "epoch": 0.6269829771821803, "grad_norm": 1.4091395989027422, "learning_rate": 6.452432478460802e-06, "loss": 0.7787, "step": 17311 }, { "epoch": 0.6270191959434988, "grad_norm": 0.9430876190137928, "learning_rate": 6.451335733081931e-06, "loss": 0.6367, "step": 17312 }, { "epoch": 0.6270554147048171, "grad_norm": 1.5417169333706564, "learning_rate": 6.450239036535797e-06, "loss": 0.7682, "step": 17313 }, { "epoch": 0.6270916334661355, "grad_norm": 1.3491228720079949, "learning_rate": 6.449142388837495e-06, "loss": 0.7059, "step": 17314 }, { "epoch": 0.6271278522274538, "grad_norm": 1.0399953727209266, "learning_rate": 6.448045790002111e-06, "loss": 0.6806, "step": 17315 }, { "epoch": 0.6271640709887721, "grad_norm": 1.3952198538987632, "learning_rate": 6.446949240044742e-06, "loss": 0.706, "step": 17316 }, { "epoch": 0.6272002897500906, "grad_norm": 1.3508851401894544, "learning_rate": 6.445852738980469e-06, "loss": 0.6766, "step": 17317 }, { "epoch": 0.6272365085114089, "grad_norm": 0.9780630228026091, "learning_rate": 6.444756286824387e-06, "loss": 0.6847, "step": 17318 }, { "epoch": 0.6272727272727273, "grad_norm": 0.9612423967456357, "learning_rate": 6.443659883591584e-06, "loss": 0.6931, "step": 17319 }, { "epoch": 0.6273089460340456, "grad_norm": 1.340675633411864, "learning_rate": 6.4425635292971435e-06, "loss": 0.7281, "step": 17320 }, { "epoch": 0.627345164795364, "grad_norm": 1.0934886593540134, "learning_rate": 6.441467223956157e-06, "loss": 0.7174, "step": 17321 }, { "epoch": 0.6273813835566824, "grad_norm": 1.0432409453575675, "learning_rate": 6.440370967583705e-06, "loss": 0.6382, "step": 17322 }, { "epoch": 0.6274176023180007, "grad_norm": 1.3136945068605903, "learning_rate": 6.4392747601948805e-06, "loss": 0.7365, "step": 17323 }, { "epoch": 0.6274538210793191, "grad_norm": 1.386321149831612, "learning_rate": 6.438178601804761e-06, "loss": 0.781, "step": 17324 }, { "epoch": 0.6274900398406374, "grad_norm": 1.0326173788210837, "learning_rate": 6.4370824924284345e-06, "loss": 0.7499, "step": 17325 }, { "epoch": 0.6275262586019558, "grad_norm": 1.2293457253825772, "learning_rate": 6.435986432080983e-06, "loss": 0.6664, "step": 17326 }, { "epoch": 0.6275624773632742, "grad_norm": 1.3872351000751397, "learning_rate": 6.434890420777491e-06, "loss": 0.7333, "step": 17327 }, { "epoch": 0.6275986961245925, "grad_norm": 1.3491565107404042, "learning_rate": 6.433794458533038e-06, "loss": 0.7622, "step": 17328 }, { "epoch": 0.6276349148859109, "grad_norm": 0.9708776146872219, "learning_rate": 6.432698545362707e-06, "loss": 0.6975, "step": 17329 }, { "epoch": 0.6276711336472293, "grad_norm": 1.2673215856296518, "learning_rate": 6.43160268128158e-06, "loss": 0.6694, "step": 17330 }, { "epoch": 0.6277073524085476, "grad_norm": 1.3908716475959833, "learning_rate": 6.4305068663047335e-06, "loss": 0.7005, "step": 17331 }, { "epoch": 0.627743571169866, "grad_norm": 1.181800231321709, "learning_rate": 6.429411100447252e-06, "loss": 0.6265, "step": 17332 }, { "epoch": 0.6277797899311843, "grad_norm": 1.404264269968902, "learning_rate": 6.428315383724208e-06, "loss": 0.7342, "step": 17333 }, { "epoch": 0.6278160086925028, "grad_norm": 0.9500344442013011, "learning_rate": 6.427219716150685e-06, "loss": 0.6467, "step": 17334 }, { "epoch": 0.6278522274538211, "grad_norm": 1.4036805945381599, "learning_rate": 6.426124097741755e-06, "loss": 0.6639, "step": 17335 }, { "epoch": 0.6278884462151394, "grad_norm": 1.4117989475890904, "learning_rate": 6.425028528512501e-06, "loss": 0.7648, "step": 17336 }, { "epoch": 0.6279246649764578, "grad_norm": 1.3092109911188279, "learning_rate": 6.423933008477994e-06, "loss": 0.7242, "step": 17337 }, { "epoch": 0.6279608837377761, "grad_norm": 1.3432800994746983, "learning_rate": 6.42283753765331e-06, "loss": 0.799, "step": 17338 }, { "epoch": 0.6279971024990946, "grad_norm": 1.354431417074191, "learning_rate": 6.421742116053526e-06, "loss": 0.7578, "step": 17339 }, { "epoch": 0.6280333212604129, "grad_norm": 1.2629635762643519, "learning_rate": 6.420646743693715e-06, "loss": 0.6717, "step": 17340 }, { "epoch": 0.6280695400217312, "grad_norm": 1.2750355458447014, "learning_rate": 6.4195514205889495e-06, "loss": 0.747, "step": 17341 }, { "epoch": 0.6281057587830496, "grad_norm": 1.604607121164058, "learning_rate": 6.418456146754302e-06, "loss": 0.7256, "step": 17342 }, { "epoch": 0.628141977544368, "grad_norm": 1.4508976426936446, "learning_rate": 6.417360922204847e-06, "loss": 0.7081, "step": 17343 }, { "epoch": 0.6281781963056864, "grad_norm": 1.246878200660747, "learning_rate": 6.416265746955651e-06, "loss": 0.6655, "step": 17344 }, { "epoch": 0.6282144150670047, "grad_norm": 1.0737460759461075, "learning_rate": 6.415170621021789e-06, "loss": 0.7118, "step": 17345 }, { "epoch": 0.628250633828323, "grad_norm": 1.2925764991413404, "learning_rate": 6.414075544418328e-06, "loss": 0.6892, "step": 17346 }, { "epoch": 0.6282868525896415, "grad_norm": 1.3196507487045197, "learning_rate": 6.412980517160341e-06, "loss": 0.6682, "step": 17347 }, { "epoch": 0.6283230713509598, "grad_norm": 1.2295851630642214, "learning_rate": 6.41188553926289e-06, "loss": 0.7256, "step": 17348 }, { "epoch": 0.6283592901122782, "grad_norm": 1.6214456658830616, "learning_rate": 6.410790610741048e-06, "loss": 0.7245, "step": 17349 }, { "epoch": 0.6283955088735965, "grad_norm": 1.3925697238128334, "learning_rate": 6.409695731609882e-06, "loss": 0.784, "step": 17350 }, { "epoch": 0.6284317276349148, "grad_norm": 0.9931905261408415, "learning_rate": 6.408600901884456e-06, "loss": 0.7122, "step": 17351 }, { "epoch": 0.6284679463962333, "grad_norm": 1.790375390222474, "learning_rate": 6.40750612157984e-06, "loss": 0.6864, "step": 17352 }, { "epoch": 0.6285041651575516, "grad_norm": 1.3041087636567417, "learning_rate": 6.406411390711094e-06, "loss": 0.7134, "step": 17353 }, { "epoch": 0.62854038391887, "grad_norm": 1.313170230695644, "learning_rate": 6.405316709293287e-06, "loss": 0.6699, "step": 17354 }, { "epoch": 0.6285766026801883, "grad_norm": 1.3816162352583772, "learning_rate": 6.404222077341478e-06, "loss": 0.6764, "step": 17355 }, { "epoch": 0.6286128214415067, "grad_norm": 1.3851678967456897, "learning_rate": 6.403127494870735e-06, "loss": 0.7434, "step": 17356 }, { "epoch": 0.6286490402028251, "grad_norm": 1.4395100195418573, "learning_rate": 6.402032961896116e-06, "loss": 0.6708, "step": 17357 }, { "epoch": 0.6286852589641434, "grad_norm": 1.250617740771245, "learning_rate": 6.400938478432688e-06, "loss": 0.6581, "step": 17358 }, { "epoch": 0.6287214777254618, "grad_norm": 1.444602739196014, "learning_rate": 6.399844044495504e-06, "loss": 0.7825, "step": 17359 }, { "epoch": 0.6287576964867801, "grad_norm": 1.5077152677918764, "learning_rate": 6.398749660099633e-06, "loss": 0.7652, "step": 17360 }, { "epoch": 0.6287939152480985, "grad_norm": 1.4030872886831964, "learning_rate": 6.39765532526013e-06, "loss": 0.7545, "step": 17361 }, { "epoch": 0.6288301340094169, "grad_norm": 1.3848651795051121, "learning_rate": 6.396561039992055e-06, "loss": 0.7176, "step": 17362 }, { "epoch": 0.6288663527707352, "grad_norm": 1.252767406666896, "learning_rate": 6.395466804310468e-06, "loss": 0.6763, "step": 17363 }, { "epoch": 0.6289025715320536, "grad_norm": 1.2882987714921144, "learning_rate": 6.394372618230422e-06, "loss": 0.6944, "step": 17364 }, { "epoch": 0.628938790293372, "grad_norm": 1.4320725430863261, "learning_rate": 6.39327848176698e-06, "loss": 0.7083, "step": 17365 }, { "epoch": 0.6289750090546903, "grad_norm": 1.3913993875975808, "learning_rate": 6.392184394935193e-06, "loss": 0.7383, "step": 17366 }, { "epoch": 0.6290112278160087, "grad_norm": 1.365500470589742, "learning_rate": 6.3910903577501215e-06, "loss": 0.7249, "step": 17367 }, { "epoch": 0.629047446577327, "grad_norm": 1.3633866406012256, "learning_rate": 6.389996370226814e-06, "loss": 0.6929, "step": 17368 }, { "epoch": 0.6290836653386455, "grad_norm": 1.0773234607768767, "learning_rate": 6.388902432380331e-06, "loss": 0.6843, "step": 17369 }, { "epoch": 0.6291198840999638, "grad_norm": 1.3799879347221515, "learning_rate": 6.387808544225723e-06, "loss": 0.713, "step": 17370 }, { "epoch": 0.6291561028612821, "grad_norm": 1.3527593068406534, "learning_rate": 6.3867147057780435e-06, "loss": 0.6589, "step": 17371 }, { "epoch": 0.6291923216226005, "grad_norm": 1.4387881089879686, "learning_rate": 6.385620917052344e-06, "loss": 0.6693, "step": 17372 }, { "epoch": 0.6292285403839188, "grad_norm": 1.4786472558016608, "learning_rate": 6.384527178063676e-06, "loss": 0.8237, "step": 17373 }, { "epoch": 0.6292647591452373, "grad_norm": 1.2592011436431951, "learning_rate": 6.383433488827093e-06, "loss": 0.6494, "step": 17374 }, { "epoch": 0.6293009779065556, "grad_norm": 1.5818938611414322, "learning_rate": 6.382339849357642e-06, "loss": 0.6449, "step": 17375 }, { "epoch": 0.6293371966678739, "grad_norm": 1.0345496780191432, "learning_rate": 6.3812462596703726e-06, "loss": 0.6735, "step": 17376 }, { "epoch": 0.6293734154291923, "grad_norm": 1.448217914581099, "learning_rate": 6.380152719780335e-06, "loss": 0.77, "step": 17377 }, { "epoch": 0.6294096341905107, "grad_norm": 1.0135534762925102, "learning_rate": 6.379059229702578e-06, "loss": 0.7283, "step": 17378 }, { "epoch": 0.6294458529518291, "grad_norm": 1.8223956538027943, "learning_rate": 6.377965789452145e-06, "loss": 0.6448, "step": 17379 }, { "epoch": 0.6294820717131474, "grad_norm": 1.0613855955236484, "learning_rate": 6.376872399044085e-06, "loss": 0.7688, "step": 17380 }, { "epoch": 0.6295182904744657, "grad_norm": 1.6219657127006204, "learning_rate": 6.375779058493448e-06, "loss": 0.7545, "step": 17381 }, { "epoch": 0.6295545092357842, "grad_norm": 1.1370400008825736, "learning_rate": 6.374685767815272e-06, "loss": 0.7373, "step": 17382 }, { "epoch": 0.6295907279971025, "grad_norm": 0.9376874004385581, "learning_rate": 6.3735925270246105e-06, "loss": 0.6823, "step": 17383 }, { "epoch": 0.6296269467584209, "grad_norm": 1.3526323286324995, "learning_rate": 6.3724993361365e-06, "loss": 0.6814, "step": 17384 }, { "epoch": 0.6296631655197392, "grad_norm": 1.5261783605848567, "learning_rate": 6.371406195165983e-06, "loss": 0.7259, "step": 17385 }, { "epoch": 0.6296993842810575, "grad_norm": 1.3883067830130802, "learning_rate": 6.3703131041281086e-06, "loss": 0.6882, "step": 17386 }, { "epoch": 0.629735603042376, "grad_norm": 1.3694083887960868, "learning_rate": 6.369220063037913e-06, "loss": 0.6636, "step": 17387 }, { "epoch": 0.6297718218036943, "grad_norm": 1.6056100801693112, "learning_rate": 6.368127071910441e-06, "loss": 0.7503, "step": 17388 }, { "epoch": 0.6298080405650127, "grad_norm": 1.3583167138599281, "learning_rate": 6.36703413076073e-06, "loss": 0.6815, "step": 17389 }, { "epoch": 0.629844259326331, "grad_norm": 1.3275742973870042, "learning_rate": 6.365941239603823e-06, "loss": 0.721, "step": 17390 }, { "epoch": 0.6298804780876494, "grad_norm": 0.917389212989905, "learning_rate": 6.364848398454756e-06, "loss": 0.7225, "step": 17391 }, { "epoch": 0.6299166968489678, "grad_norm": 1.2098983532173488, "learning_rate": 6.36375560732857e-06, "loss": 0.7171, "step": 17392 }, { "epoch": 0.6299529156102861, "grad_norm": 1.0461611992014463, "learning_rate": 6.362662866240303e-06, "loss": 0.688, "step": 17393 }, { "epoch": 0.6299891343716045, "grad_norm": 1.4225501760418489, "learning_rate": 6.36157017520499e-06, "loss": 0.6723, "step": 17394 }, { "epoch": 0.6300253531329228, "grad_norm": 1.3801881812681764, "learning_rate": 6.360477534237668e-06, "loss": 0.7765, "step": 17395 }, { "epoch": 0.6300615718942412, "grad_norm": 1.355481730352247, "learning_rate": 6.3593849433533726e-06, "loss": 0.7194, "step": 17396 }, { "epoch": 0.6300977906555596, "grad_norm": 1.4566601563185386, "learning_rate": 6.358292402567142e-06, "loss": 0.7586, "step": 17397 }, { "epoch": 0.6301340094168779, "grad_norm": 1.3690155525871897, "learning_rate": 6.357199911894005e-06, "loss": 0.7562, "step": 17398 }, { "epoch": 0.6301702281781963, "grad_norm": 1.392705254148928, "learning_rate": 6.356107471349e-06, "loss": 0.7142, "step": 17399 }, { "epoch": 0.6302064469395147, "grad_norm": 1.3044757142892345, "learning_rate": 6.355015080947158e-06, "loss": 0.7599, "step": 17400 }, { "epoch": 0.630242665700833, "grad_norm": 1.3954580296792425, "learning_rate": 6.353922740703511e-06, "loss": 0.7287, "step": 17401 }, { "epoch": 0.6302788844621514, "grad_norm": 1.3870006798536776, "learning_rate": 6.35283045063309e-06, "loss": 0.6653, "step": 17402 }, { "epoch": 0.6303151032234697, "grad_norm": 1.4526549977667156, "learning_rate": 6.351738210750925e-06, "loss": 0.7219, "step": 17403 }, { "epoch": 0.6303513219847882, "grad_norm": 1.3258677158914633, "learning_rate": 6.350646021072051e-06, "loss": 0.6667, "step": 17404 }, { "epoch": 0.6303875407461065, "grad_norm": 1.0361567877836875, "learning_rate": 6.349553881611492e-06, "loss": 0.6573, "step": 17405 }, { "epoch": 0.6304237595074248, "grad_norm": 1.453368931723371, "learning_rate": 6.348461792384282e-06, "loss": 0.7548, "step": 17406 }, { "epoch": 0.6304599782687432, "grad_norm": 1.4285834086689617, "learning_rate": 6.347369753405444e-06, "loss": 0.7133, "step": 17407 }, { "epoch": 0.6304961970300615, "grad_norm": 1.457850258012166, "learning_rate": 6.346277764690009e-06, "loss": 0.7345, "step": 17408 }, { "epoch": 0.63053241579138, "grad_norm": 1.3304837570419263, "learning_rate": 6.345185826253001e-06, "loss": 0.7427, "step": 17409 }, { "epoch": 0.6305686345526983, "grad_norm": 1.408839752478439, "learning_rate": 6.344093938109449e-06, "loss": 0.7679, "step": 17410 }, { "epoch": 0.6306048533140166, "grad_norm": 1.4348563966564516, "learning_rate": 6.343002100274375e-06, "loss": 0.7204, "step": 17411 }, { "epoch": 0.630641072075335, "grad_norm": 1.297679500426087, "learning_rate": 6.341910312762804e-06, "loss": 0.6971, "step": 17412 }, { "epoch": 0.6306772908366534, "grad_norm": 1.3243003611889788, "learning_rate": 6.340818575589767e-06, "loss": 0.7133, "step": 17413 }, { "epoch": 0.6307135095979718, "grad_norm": 1.4542180431639273, "learning_rate": 6.339726888770277e-06, "loss": 0.7163, "step": 17414 }, { "epoch": 0.6307497283592901, "grad_norm": 1.20969597560852, "learning_rate": 6.338635252319363e-06, "loss": 0.6647, "step": 17415 }, { "epoch": 0.6307859471206084, "grad_norm": 1.2324092005880485, "learning_rate": 6.337543666252044e-06, "loss": 0.6286, "step": 17416 }, { "epoch": 0.6308221658819269, "grad_norm": 1.3058456431587449, "learning_rate": 6.336452130583344e-06, "loss": 0.6961, "step": 17417 }, { "epoch": 0.6308583846432452, "grad_norm": 1.5649455649191224, "learning_rate": 6.335360645328279e-06, "loss": 0.7218, "step": 17418 }, { "epoch": 0.6308946034045636, "grad_norm": 1.3595919929160032, "learning_rate": 6.334269210501876e-06, "loss": 0.6375, "step": 17419 }, { "epoch": 0.6309308221658819, "grad_norm": 1.3561507043109575, "learning_rate": 6.333177826119145e-06, "loss": 0.7157, "step": 17420 }, { "epoch": 0.6309670409272002, "grad_norm": 1.2863040135241142, "learning_rate": 6.332086492195112e-06, "loss": 0.7459, "step": 17421 }, { "epoch": 0.6310032596885187, "grad_norm": 1.3969456037710686, "learning_rate": 6.33099520874479e-06, "loss": 0.6421, "step": 17422 }, { "epoch": 0.631039478449837, "grad_norm": 1.335024359172536, "learning_rate": 6.329903975783197e-06, "loss": 0.7242, "step": 17423 }, { "epoch": 0.6310756972111554, "grad_norm": 1.147530742764127, "learning_rate": 6.3288127933253526e-06, "loss": 0.6914, "step": 17424 }, { "epoch": 0.6311119159724737, "grad_norm": 1.8224151709421783, "learning_rate": 6.327721661386268e-06, "loss": 0.6961, "step": 17425 }, { "epoch": 0.631148134733792, "grad_norm": 1.524297679854769, "learning_rate": 6.326630579980962e-06, "loss": 0.7072, "step": 17426 }, { "epoch": 0.6311843534951105, "grad_norm": 1.2439383618583917, "learning_rate": 6.325539549124445e-06, "loss": 0.7305, "step": 17427 }, { "epoch": 0.6312205722564288, "grad_norm": 1.3712090201978866, "learning_rate": 6.3244485688317355e-06, "loss": 0.666, "step": 17428 }, { "epoch": 0.6312567910177472, "grad_norm": 1.695918592319839, "learning_rate": 6.32335763911784e-06, "loss": 0.7709, "step": 17429 }, { "epoch": 0.6312930097790656, "grad_norm": 1.4730481238550817, "learning_rate": 6.322266759997776e-06, "loss": 0.7279, "step": 17430 }, { "epoch": 0.6313292285403839, "grad_norm": 1.2639049959135253, "learning_rate": 6.321175931486552e-06, "loss": 0.6582, "step": 17431 }, { "epoch": 0.6313654473017023, "grad_norm": 1.4701215321028243, "learning_rate": 6.320085153599181e-06, "loss": 0.7226, "step": 17432 }, { "epoch": 0.6314016660630206, "grad_norm": 0.98064710667318, "learning_rate": 6.318994426350669e-06, "loss": 0.7181, "step": 17433 }, { "epoch": 0.631437884824339, "grad_norm": 1.300760461595314, "learning_rate": 6.3179037497560295e-06, "loss": 0.6285, "step": 17434 }, { "epoch": 0.6314741035856574, "grad_norm": 1.2243335997430123, "learning_rate": 6.31681312383027e-06, "loss": 0.6561, "step": 17435 }, { "epoch": 0.6315103223469757, "grad_norm": 1.3454764539276347, "learning_rate": 6.315722548588398e-06, "loss": 0.7371, "step": 17436 }, { "epoch": 0.6315465411082941, "grad_norm": 1.1915344074571743, "learning_rate": 6.314632024045422e-06, "loss": 0.7131, "step": 17437 }, { "epoch": 0.6315827598696124, "grad_norm": 1.0234292442684032, "learning_rate": 6.3135415502163466e-06, "loss": 0.7235, "step": 17438 }, { "epoch": 0.6316189786309309, "grad_norm": 1.1029093228779325, "learning_rate": 6.3124511271161794e-06, "loss": 0.7545, "step": 17439 }, { "epoch": 0.6316551973922492, "grad_norm": 1.4759798363617371, "learning_rate": 6.311360754759923e-06, "loss": 0.7892, "step": 17440 }, { "epoch": 0.6316914161535675, "grad_norm": 1.4738932525405826, "learning_rate": 6.3102704331625865e-06, "loss": 0.7055, "step": 17441 }, { "epoch": 0.6317276349148859, "grad_norm": 1.2218699326666986, "learning_rate": 6.3091801623391676e-06, "loss": 0.699, "step": 17442 }, { "epoch": 0.6317638536762042, "grad_norm": 1.5274178128444322, "learning_rate": 6.308089942304675e-06, "loss": 0.7392, "step": 17443 }, { "epoch": 0.6318000724375227, "grad_norm": 1.558396028777256, "learning_rate": 6.306999773074109e-06, "loss": 0.805, "step": 17444 }, { "epoch": 0.631836291198841, "grad_norm": 1.296220742213195, "learning_rate": 6.30590965466247e-06, "loss": 0.7525, "step": 17445 }, { "epoch": 0.6318725099601593, "grad_norm": 1.476875207727848, "learning_rate": 6.30481958708476e-06, "loss": 0.742, "step": 17446 }, { "epoch": 0.6319087287214777, "grad_norm": 1.0810042291868707, "learning_rate": 6.303729570355981e-06, "loss": 0.7093, "step": 17447 }, { "epoch": 0.6319449474827961, "grad_norm": 1.6297609821953996, "learning_rate": 6.30263960449113e-06, "loss": 0.7428, "step": 17448 }, { "epoch": 0.6319811662441145, "grad_norm": 1.0187377589150741, "learning_rate": 6.301549689505207e-06, "loss": 0.6899, "step": 17449 }, { "epoch": 0.6320173850054328, "grad_norm": 1.553755804501415, "learning_rate": 6.30045982541321e-06, "loss": 0.7046, "step": 17450 }, { "epoch": 0.6320536037667511, "grad_norm": 1.2908391932952088, "learning_rate": 6.299370012230135e-06, "loss": 0.6841, "step": 17451 }, { "epoch": 0.6320898225280696, "grad_norm": 1.2976602035797542, "learning_rate": 6.298280249970983e-06, "loss": 0.6896, "step": 17452 }, { "epoch": 0.6321260412893879, "grad_norm": 1.3098158830135394, "learning_rate": 6.297190538650745e-06, "loss": 0.7168, "step": 17453 }, { "epoch": 0.6321622600507063, "grad_norm": 1.4619550756538182, "learning_rate": 6.296100878284419e-06, "loss": 0.7376, "step": 17454 }, { "epoch": 0.6321984788120246, "grad_norm": 1.3938785649776497, "learning_rate": 6.295011268887e-06, "loss": 0.6455, "step": 17455 }, { "epoch": 0.6322346975733429, "grad_norm": 1.5512956923841494, "learning_rate": 6.293921710473481e-06, "loss": 0.7437, "step": 17456 }, { "epoch": 0.6322709163346614, "grad_norm": 1.292437198521326, "learning_rate": 6.292832203058857e-06, "loss": 0.7401, "step": 17457 }, { "epoch": 0.6323071350959797, "grad_norm": 1.3293411324862956, "learning_rate": 6.2917427466581184e-06, "loss": 0.6336, "step": 17458 }, { "epoch": 0.6323433538572981, "grad_norm": 1.3240677338126006, "learning_rate": 6.290653341286259e-06, "loss": 0.6711, "step": 17459 }, { "epoch": 0.6323795726186164, "grad_norm": 1.472858802060307, "learning_rate": 6.289563986958267e-06, "loss": 0.6979, "step": 17460 }, { "epoch": 0.6324157913799348, "grad_norm": 1.2882826975137087, "learning_rate": 6.2884746836891365e-06, "loss": 0.6908, "step": 17461 }, { "epoch": 0.6324520101412532, "grad_norm": 1.3537101942915355, "learning_rate": 6.2873854314938544e-06, "loss": 0.7503, "step": 17462 }, { "epoch": 0.6324882289025715, "grad_norm": 1.4623739134710865, "learning_rate": 6.286296230387412e-06, "loss": 0.7107, "step": 17463 }, { "epoch": 0.6325244476638899, "grad_norm": 1.272639184530596, "learning_rate": 6.285207080384795e-06, "loss": 0.7072, "step": 17464 }, { "epoch": 0.6325606664252083, "grad_norm": 1.3903536156722338, "learning_rate": 6.284117981500993e-06, "loss": 0.747, "step": 17465 }, { "epoch": 0.6325968851865266, "grad_norm": 1.3989264776151566, "learning_rate": 6.283028933750994e-06, "loss": 0.7288, "step": 17466 }, { "epoch": 0.632633103947845, "grad_norm": 1.1605701749700674, "learning_rate": 6.281939937149781e-06, "loss": 0.6794, "step": 17467 }, { "epoch": 0.6326693227091633, "grad_norm": 1.3769686745381124, "learning_rate": 6.280850991712345e-06, "loss": 0.6521, "step": 17468 }, { "epoch": 0.6327055414704817, "grad_norm": 1.0436268961281976, "learning_rate": 6.279762097453664e-06, "loss": 0.7021, "step": 17469 }, { "epoch": 0.6327417602318001, "grad_norm": 1.2700884843211286, "learning_rate": 6.2786732543887274e-06, "loss": 0.7139, "step": 17470 }, { "epoch": 0.6327779789931184, "grad_norm": 1.3118370693762926, "learning_rate": 6.277584462532516e-06, "loss": 0.6776, "step": 17471 }, { "epoch": 0.6328141977544368, "grad_norm": 1.356574305064708, "learning_rate": 6.276495721900013e-06, "loss": 0.6863, "step": 17472 }, { "epoch": 0.6328504165157551, "grad_norm": 1.3875562616971968, "learning_rate": 6.2754070325062e-06, "loss": 0.692, "step": 17473 }, { "epoch": 0.6328866352770736, "grad_norm": 1.1751668121038605, "learning_rate": 6.274318394366058e-06, "loss": 0.6594, "step": 17474 }, { "epoch": 0.6329228540383919, "grad_norm": 1.3919747350836973, "learning_rate": 6.27322980749457e-06, "loss": 0.8112, "step": 17475 }, { "epoch": 0.6329590727997102, "grad_norm": 1.4033885732147948, "learning_rate": 6.272141271906714e-06, "loss": 0.8252, "step": 17476 }, { "epoch": 0.6329952915610286, "grad_norm": 1.5426861872463642, "learning_rate": 6.271052787617471e-06, "loss": 0.7624, "step": 17477 }, { "epoch": 0.633031510322347, "grad_norm": 1.4920849040846167, "learning_rate": 6.269964354641817e-06, "loss": 0.7611, "step": 17478 }, { "epoch": 0.6330677290836654, "grad_norm": 1.0015635641701621, "learning_rate": 6.2688759729947325e-06, "loss": 0.6719, "step": 17479 }, { "epoch": 0.6331039478449837, "grad_norm": 1.1609779492554906, "learning_rate": 6.267787642691191e-06, "loss": 0.731, "step": 17480 }, { "epoch": 0.633140166606302, "grad_norm": 1.4225862971855727, "learning_rate": 6.266699363746173e-06, "loss": 0.7217, "step": 17481 }, { "epoch": 0.6331763853676204, "grad_norm": 1.5133137574640632, "learning_rate": 6.26561113617465e-06, "loss": 0.7875, "step": 17482 }, { "epoch": 0.6332126041289388, "grad_norm": 0.9066963055448938, "learning_rate": 6.264522959991601e-06, "loss": 0.7333, "step": 17483 }, { "epoch": 0.6332488228902572, "grad_norm": 0.9680490625451863, "learning_rate": 6.2634348352119966e-06, "loss": 0.6679, "step": 17484 }, { "epoch": 0.6332850416515755, "grad_norm": 1.355014420504965, "learning_rate": 6.262346761850813e-06, "loss": 0.7471, "step": 17485 }, { "epoch": 0.6333212604128938, "grad_norm": 1.3119286019986363, "learning_rate": 6.261258739923022e-06, "loss": 0.6554, "step": 17486 }, { "epoch": 0.6333574791742123, "grad_norm": 1.3377410979141824, "learning_rate": 6.2601707694435945e-06, "loss": 0.7067, "step": 17487 }, { "epoch": 0.6333936979355306, "grad_norm": 1.3134799411262252, "learning_rate": 6.259082850427507e-06, "loss": 0.7706, "step": 17488 }, { "epoch": 0.633429916696849, "grad_norm": 1.0389922263523461, "learning_rate": 6.257994982889723e-06, "loss": 0.7063, "step": 17489 }, { "epoch": 0.6334661354581673, "grad_norm": 1.3993253766300202, "learning_rate": 6.256907166845219e-06, "loss": 0.6609, "step": 17490 }, { "epoch": 0.6335023542194856, "grad_norm": 0.9799099251978727, "learning_rate": 6.255819402308958e-06, "loss": 0.7043, "step": 17491 }, { "epoch": 0.6335385729808041, "grad_norm": 1.1377761536331288, "learning_rate": 6.254731689295914e-06, "loss": 0.7348, "step": 17492 }, { "epoch": 0.6335747917421224, "grad_norm": 1.27725446313163, "learning_rate": 6.253644027821052e-06, "loss": 0.6985, "step": 17493 }, { "epoch": 0.6336110105034408, "grad_norm": 1.0953076427888717, "learning_rate": 6.252556417899342e-06, "loss": 0.7, "step": 17494 }, { "epoch": 0.6336472292647591, "grad_norm": 1.1848302428342818, "learning_rate": 6.251468859545746e-06, "loss": 0.646, "step": 17495 }, { "epoch": 0.6336834480260775, "grad_norm": 1.3240197040546302, "learning_rate": 6.250381352775232e-06, "loss": 0.7728, "step": 17496 }, { "epoch": 0.6337196667873959, "grad_norm": 1.451395220403028, "learning_rate": 6.249293897602769e-06, "loss": 0.8028, "step": 17497 }, { "epoch": 0.6337558855487142, "grad_norm": 1.3112971969232188, "learning_rate": 6.248206494043313e-06, "loss": 0.6699, "step": 17498 }, { "epoch": 0.6337921043100326, "grad_norm": 1.3923599238072208, "learning_rate": 6.247119142111835e-06, "loss": 0.6283, "step": 17499 }, { "epoch": 0.633828323071351, "grad_norm": 1.28213168037946, "learning_rate": 6.246031841823295e-06, "loss": 0.6668, "step": 17500 }, { "epoch": 0.6338645418326693, "grad_norm": 1.5777359225536367, "learning_rate": 6.244944593192655e-06, "loss": 0.6857, "step": 17501 }, { "epoch": 0.6339007605939877, "grad_norm": 1.2518392732779164, "learning_rate": 6.243857396234876e-06, "loss": 0.7119, "step": 17502 }, { "epoch": 0.633936979355306, "grad_norm": 1.087147227248377, "learning_rate": 6.242770250964921e-06, "loss": 0.6649, "step": 17503 }, { "epoch": 0.6339731981166244, "grad_norm": 1.2789392422782269, "learning_rate": 6.241683157397748e-06, "loss": 0.6641, "step": 17504 }, { "epoch": 0.6340094168779428, "grad_norm": 1.2813118187016976, "learning_rate": 6.240596115548316e-06, "loss": 0.7246, "step": 17505 }, { "epoch": 0.6340456356392611, "grad_norm": 1.7832604803281136, "learning_rate": 6.239509125431586e-06, "loss": 0.7573, "step": 17506 }, { "epoch": 0.6340818544005795, "grad_norm": 1.4967650769852845, "learning_rate": 6.238422187062515e-06, "loss": 0.7318, "step": 17507 }, { "epoch": 0.6341180731618978, "grad_norm": 1.2093906066747562, "learning_rate": 6.23733530045606e-06, "loss": 0.6469, "step": 17508 }, { "epoch": 0.6341542919232163, "grad_norm": 1.242212738374514, "learning_rate": 6.236248465627177e-06, "loss": 0.6663, "step": 17509 }, { "epoch": 0.6341905106845346, "grad_norm": 1.3996486403156725, "learning_rate": 6.235161682590824e-06, "loss": 0.6718, "step": 17510 }, { "epoch": 0.6342267294458529, "grad_norm": 1.5493769040274914, "learning_rate": 6.23407495136195e-06, "loss": 0.8129, "step": 17511 }, { "epoch": 0.6342629482071713, "grad_norm": 1.3450170298831243, "learning_rate": 6.232988271955518e-06, "loss": 0.6581, "step": 17512 }, { "epoch": 0.6342991669684896, "grad_norm": 1.4357019950462453, "learning_rate": 6.231901644386476e-06, "loss": 0.655, "step": 17513 }, { "epoch": 0.6343353857298081, "grad_norm": 1.3524451818832857, "learning_rate": 6.2308150686697775e-06, "loss": 0.7666, "step": 17514 }, { "epoch": 0.6343716044911264, "grad_norm": 1.3705400852906502, "learning_rate": 6.229728544820376e-06, "loss": 0.691, "step": 17515 }, { "epoch": 0.6344078232524447, "grad_norm": 1.3627662273367906, "learning_rate": 6.2286420728532215e-06, "loss": 0.6842, "step": 17516 }, { "epoch": 0.6344440420137631, "grad_norm": 1.2231483870576856, "learning_rate": 6.227555652783268e-06, "loss": 0.6587, "step": 17517 }, { "epoch": 0.6344802607750815, "grad_norm": 1.4401736073856697, "learning_rate": 6.226469284625462e-06, "loss": 0.7186, "step": 17518 }, { "epoch": 0.6345164795363999, "grad_norm": 1.1131952284949425, "learning_rate": 6.225382968394755e-06, "loss": 0.6576, "step": 17519 }, { "epoch": 0.6345526982977182, "grad_norm": 1.4342643237475887, "learning_rate": 6.224296704106096e-06, "loss": 0.7038, "step": 17520 }, { "epoch": 0.6345889170590365, "grad_norm": 1.07255715565714, "learning_rate": 6.223210491774432e-06, "loss": 0.7143, "step": 17521 }, { "epoch": 0.634625135820355, "grad_norm": 1.3729771953174115, "learning_rate": 6.222124331414707e-06, "loss": 0.6858, "step": 17522 }, { "epoch": 0.6346613545816733, "grad_norm": 1.2794611687030897, "learning_rate": 6.2210382230418744e-06, "loss": 0.6149, "step": 17523 }, { "epoch": 0.6346975733429917, "grad_norm": 1.2801812267485924, "learning_rate": 6.219952166670873e-06, "loss": 0.7692, "step": 17524 }, { "epoch": 0.63473379210431, "grad_norm": 1.000265482962776, "learning_rate": 6.2188661623166545e-06, "loss": 0.6878, "step": 17525 }, { "epoch": 0.6347700108656283, "grad_norm": 0.8902106832756035, "learning_rate": 6.217780209994157e-06, "loss": 0.6906, "step": 17526 }, { "epoch": 0.6348062296269468, "grad_norm": 1.1459009586040403, "learning_rate": 6.2166943097183285e-06, "loss": 0.7284, "step": 17527 }, { "epoch": 0.6348424483882651, "grad_norm": 1.4375809900646974, "learning_rate": 6.215608461504112e-06, "loss": 0.6993, "step": 17528 }, { "epoch": 0.6348786671495835, "grad_norm": 1.3369432143828919, "learning_rate": 6.214522665366446e-06, "loss": 0.6779, "step": 17529 }, { "epoch": 0.6349148859109018, "grad_norm": 1.3036343619899475, "learning_rate": 6.213436921320277e-06, "loss": 0.6663, "step": 17530 }, { "epoch": 0.6349511046722202, "grad_norm": 1.3000754590955899, "learning_rate": 6.21235122938054e-06, "loss": 0.6565, "step": 17531 }, { "epoch": 0.6349873234335386, "grad_norm": 1.0041322302457187, "learning_rate": 6.2112655895621806e-06, "loss": 0.6748, "step": 17532 }, { "epoch": 0.6350235421948569, "grad_norm": 1.470516887084021, "learning_rate": 6.210180001880135e-06, "loss": 0.7142, "step": 17533 }, { "epoch": 0.6350597609561753, "grad_norm": 1.111683032032916, "learning_rate": 6.209094466349344e-06, "loss": 0.717, "step": 17534 }, { "epoch": 0.6350959797174937, "grad_norm": 1.434239805827318, "learning_rate": 6.208008982984741e-06, "loss": 0.6989, "step": 17535 }, { "epoch": 0.635132198478812, "grad_norm": 1.1919493911183154, "learning_rate": 6.206923551801267e-06, "loss": 0.783, "step": 17536 }, { "epoch": 0.6351684172401304, "grad_norm": 0.9982418570982562, "learning_rate": 6.20583817281386e-06, "loss": 0.6488, "step": 17537 }, { "epoch": 0.6352046360014487, "grad_norm": 1.5141745698569409, "learning_rate": 6.204752846037452e-06, "loss": 0.7491, "step": 17538 }, { "epoch": 0.6352408547627671, "grad_norm": 1.2519128570909706, "learning_rate": 6.20366757148698e-06, "loss": 0.7376, "step": 17539 }, { "epoch": 0.6352770735240855, "grad_norm": 1.2099164050353235, "learning_rate": 6.202582349177377e-06, "loss": 0.7107, "step": 17540 }, { "epoch": 0.6353132922854038, "grad_norm": 1.2990616072267847, "learning_rate": 6.201497179123579e-06, "loss": 0.6669, "step": 17541 }, { "epoch": 0.6353495110467222, "grad_norm": 1.800044531734942, "learning_rate": 6.200412061340517e-06, "loss": 0.7341, "step": 17542 }, { "epoch": 0.6353857298080405, "grad_norm": 1.3489991318272287, "learning_rate": 6.199326995843125e-06, "loss": 0.713, "step": 17543 }, { "epoch": 0.635421948569359, "grad_norm": 1.2986555056483198, "learning_rate": 6.1982419826463304e-06, "loss": 0.5736, "step": 17544 }, { "epoch": 0.6354581673306773, "grad_norm": 1.1913494917372955, "learning_rate": 6.197157021765069e-06, "loss": 0.6583, "step": 17545 }, { "epoch": 0.6354943860919956, "grad_norm": 1.2152662496496596, "learning_rate": 6.196072113214266e-06, "loss": 0.6297, "step": 17546 }, { "epoch": 0.635530604853314, "grad_norm": 0.9649301373340389, "learning_rate": 6.194987257008855e-06, "loss": 0.7584, "step": 17547 }, { "epoch": 0.6355668236146323, "grad_norm": 1.1873272912173214, "learning_rate": 6.193902453163762e-06, "loss": 0.5746, "step": 17548 }, { "epoch": 0.6356030423759508, "grad_norm": 1.2288281739836562, "learning_rate": 6.1928177016939165e-06, "loss": 0.6305, "step": 17549 }, { "epoch": 0.6356392611372691, "grad_norm": 1.2515742090060877, "learning_rate": 6.191733002614244e-06, "loss": 0.7488, "step": 17550 }, { "epoch": 0.6356754798985874, "grad_norm": 1.4136025644995036, "learning_rate": 6.190648355939671e-06, "loss": 0.7057, "step": 17551 }, { "epoch": 0.6357116986599058, "grad_norm": 1.0302660546897033, "learning_rate": 6.189563761685125e-06, "loss": 0.6698, "step": 17552 }, { "epoch": 0.6357479174212242, "grad_norm": 1.266054765192779, "learning_rate": 6.188479219865529e-06, "loss": 0.6875, "step": 17553 }, { "epoch": 0.6357841361825426, "grad_norm": 1.3558259556530423, "learning_rate": 6.18739473049581e-06, "loss": 0.7134, "step": 17554 }, { "epoch": 0.6358203549438609, "grad_norm": 0.927105508472534, "learning_rate": 6.186310293590887e-06, "loss": 0.6844, "step": 17555 }, { "epoch": 0.6358565737051792, "grad_norm": 1.565285909883426, "learning_rate": 6.185225909165686e-06, "loss": 0.7563, "step": 17556 }, { "epoch": 0.6358927924664977, "grad_norm": 1.1483543788216461, "learning_rate": 6.184141577235129e-06, "loss": 0.7386, "step": 17557 }, { "epoch": 0.635929011227816, "grad_norm": 1.040517250964866, "learning_rate": 6.183057297814135e-06, "loss": 0.6991, "step": 17558 }, { "epoch": 0.6359652299891344, "grad_norm": 1.4220648104395273, "learning_rate": 6.18197307091763e-06, "loss": 0.7137, "step": 17559 }, { "epoch": 0.6360014487504527, "grad_norm": 1.048572406827008, "learning_rate": 6.1808888965605265e-06, "loss": 0.6988, "step": 17560 }, { "epoch": 0.636037667511771, "grad_norm": 1.361451339755402, "learning_rate": 6.179804774757751e-06, "loss": 0.6894, "step": 17561 }, { "epoch": 0.6360738862730895, "grad_norm": 1.4550462373657342, "learning_rate": 6.178720705524216e-06, "loss": 0.6604, "step": 17562 }, { "epoch": 0.6361101050344078, "grad_norm": 1.4993168292134755, "learning_rate": 6.177636688874844e-06, "loss": 0.8227, "step": 17563 }, { "epoch": 0.6361463237957262, "grad_norm": 1.2169500762725922, "learning_rate": 6.176552724824547e-06, "loss": 0.6711, "step": 17564 }, { "epoch": 0.6361825425570445, "grad_norm": 1.4118071611266236, "learning_rate": 6.175468813388247e-06, "loss": 0.7159, "step": 17565 }, { "epoch": 0.6362187613183629, "grad_norm": 1.3593742285649535, "learning_rate": 6.1743849545808545e-06, "loss": 0.696, "step": 17566 }, { "epoch": 0.6362549800796813, "grad_norm": 1.2974475291976402, "learning_rate": 6.173301148417286e-06, "loss": 0.6514, "step": 17567 }, { "epoch": 0.6362911988409996, "grad_norm": 1.0177367596065028, "learning_rate": 6.172217394912459e-06, "loss": 0.6867, "step": 17568 }, { "epoch": 0.636327417602318, "grad_norm": 1.0076338752713043, "learning_rate": 6.17113369408128e-06, "loss": 0.701, "step": 17569 }, { "epoch": 0.6363636363636364, "grad_norm": 1.3288559942537426, "learning_rate": 6.17005004593867e-06, "loss": 0.7451, "step": 17570 }, { "epoch": 0.6363998551249547, "grad_norm": 1.4710945986331512, "learning_rate": 6.168966450499534e-06, "loss": 0.6542, "step": 17571 }, { "epoch": 0.6364360738862731, "grad_norm": 1.2446196882371248, "learning_rate": 6.1678829077787864e-06, "loss": 0.618, "step": 17572 }, { "epoch": 0.6364722926475914, "grad_norm": 1.2521633122021343, "learning_rate": 6.166799417791337e-06, "loss": 0.6766, "step": 17573 }, { "epoch": 0.6365085114089098, "grad_norm": 1.3132308715992713, "learning_rate": 6.165715980552097e-06, "loss": 0.6312, "step": 17574 }, { "epoch": 0.6365447301702282, "grad_norm": 1.2849656748383553, "learning_rate": 6.164632596075972e-06, "loss": 0.7627, "step": 17575 }, { "epoch": 0.6365809489315465, "grad_norm": 1.1882033998022665, "learning_rate": 6.163549264377875e-06, "loss": 0.6955, "step": 17576 }, { "epoch": 0.6366171676928649, "grad_norm": 1.3277306188534173, "learning_rate": 6.16246598547271e-06, "loss": 0.6863, "step": 17577 }, { "epoch": 0.6366533864541832, "grad_norm": 1.4039895705619998, "learning_rate": 6.161382759375383e-06, "loss": 0.7663, "step": 17578 }, { "epoch": 0.6366896052155017, "grad_norm": 1.2418675186730288, "learning_rate": 6.160299586100806e-06, "loss": 0.747, "step": 17579 }, { "epoch": 0.63672582397682, "grad_norm": 1.2594963273820756, "learning_rate": 6.1592164656638775e-06, "loss": 0.7308, "step": 17580 }, { "epoch": 0.6367620427381383, "grad_norm": 1.2637663817699922, "learning_rate": 6.1581333980795065e-06, "loss": 0.6856, "step": 17581 }, { "epoch": 0.6367982614994567, "grad_norm": 1.3773505989927874, "learning_rate": 6.157050383362595e-06, "loss": 0.7311, "step": 17582 }, { "epoch": 0.636834480260775, "grad_norm": 1.3431555977505114, "learning_rate": 6.155967421528048e-06, "loss": 0.6756, "step": 17583 }, { "epoch": 0.6368706990220935, "grad_norm": 1.3323361559615985, "learning_rate": 6.154884512590766e-06, "loss": 0.642, "step": 17584 }, { "epoch": 0.6369069177834118, "grad_norm": 1.1375158853517977, "learning_rate": 6.153801656565654e-06, "loss": 0.7285, "step": 17585 }, { "epoch": 0.6369431365447301, "grad_norm": 1.234377861774121, "learning_rate": 6.152718853467607e-06, "loss": 0.7804, "step": 17586 }, { "epoch": 0.6369793553060485, "grad_norm": 1.3210269840046978, "learning_rate": 6.1516361033115335e-06, "loss": 0.6717, "step": 17587 }, { "epoch": 0.6370155740673669, "grad_norm": 1.3497361801242436, "learning_rate": 6.150553406112325e-06, "loss": 0.7197, "step": 17588 }, { "epoch": 0.6370517928286853, "grad_norm": 1.2704186629062524, "learning_rate": 6.149470761884884e-06, "loss": 0.7384, "step": 17589 }, { "epoch": 0.6370880115900036, "grad_norm": 1.0979953285944812, "learning_rate": 6.1483881706441105e-06, "loss": 0.6745, "step": 17590 }, { "epoch": 0.6371242303513219, "grad_norm": 1.373484315280456, "learning_rate": 6.1473056324049e-06, "loss": 0.6467, "step": 17591 }, { "epoch": 0.6371604491126404, "grad_norm": 1.4256840124024461, "learning_rate": 6.146223147182148e-06, "loss": 0.6815, "step": 17592 }, { "epoch": 0.6371966678739587, "grad_norm": 1.0805968532633874, "learning_rate": 6.145140714990754e-06, "loss": 0.688, "step": 17593 }, { "epoch": 0.6372328866352771, "grad_norm": 1.3229935852414814, "learning_rate": 6.144058335845612e-06, "loss": 0.7303, "step": 17594 }, { "epoch": 0.6372691053965954, "grad_norm": 1.2466578320454063, "learning_rate": 6.142976009761614e-06, "loss": 0.6814, "step": 17595 }, { "epoch": 0.6373053241579137, "grad_norm": 1.2167354372290333, "learning_rate": 6.141893736753652e-06, "loss": 0.6963, "step": 17596 }, { "epoch": 0.6373415429192322, "grad_norm": 1.2271855223785235, "learning_rate": 6.140811516836627e-06, "loss": 0.6565, "step": 17597 }, { "epoch": 0.6373777616805505, "grad_norm": 1.137896553442997, "learning_rate": 6.139729350025423e-06, "loss": 0.6862, "step": 17598 }, { "epoch": 0.6374139804418689, "grad_norm": 1.2771916969744648, "learning_rate": 6.138647236334936e-06, "loss": 0.7023, "step": 17599 }, { "epoch": 0.6374501992031872, "grad_norm": 1.3166170955998524, "learning_rate": 6.137565175780056e-06, "loss": 0.7138, "step": 17600 }, { "epoch": 0.6374864179645056, "grad_norm": 1.2736396219591852, "learning_rate": 6.1364831683756705e-06, "loss": 0.7039, "step": 17601 }, { "epoch": 0.637522636725824, "grad_norm": 1.43708832692449, "learning_rate": 6.135401214136676e-06, "loss": 0.7108, "step": 17602 }, { "epoch": 0.6375588554871423, "grad_norm": 1.0738083491654418, "learning_rate": 6.134319313077952e-06, "loss": 0.6953, "step": 17603 }, { "epoch": 0.6375950742484607, "grad_norm": 0.9735594879849355, "learning_rate": 6.133237465214394e-06, "loss": 0.7062, "step": 17604 }, { "epoch": 0.637631293009779, "grad_norm": 1.299961698952508, "learning_rate": 6.1321556705608856e-06, "loss": 0.6817, "step": 17605 }, { "epoch": 0.6376675117710975, "grad_norm": 0.9878632726225646, "learning_rate": 6.131073929132314e-06, "loss": 0.6793, "step": 17606 }, { "epoch": 0.6377037305324158, "grad_norm": 1.5082604232135797, "learning_rate": 6.129992240943563e-06, "loss": 0.7325, "step": 17607 }, { "epoch": 0.6377399492937341, "grad_norm": 0.9492390042332537, "learning_rate": 6.128910606009522e-06, "loss": 0.6414, "step": 17608 }, { "epoch": 0.6377761680550526, "grad_norm": 1.0812512957068698, "learning_rate": 6.12782902434507e-06, "loss": 0.6645, "step": 17609 }, { "epoch": 0.6378123868163709, "grad_norm": 0.9758220364008061, "learning_rate": 6.126747495965093e-06, "loss": 0.703, "step": 17610 }, { "epoch": 0.6378486055776893, "grad_norm": 1.4267560095909473, "learning_rate": 6.125666020884476e-06, "loss": 0.7259, "step": 17611 }, { "epoch": 0.6378848243390076, "grad_norm": 1.4411203959000465, "learning_rate": 6.124584599118096e-06, "loss": 0.7216, "step": 17612 }, { "epoch": 0.6379210431003259, "grad_norm": 1.0203376823276176, "learning_rate": 6.123503230680841e-06, "loss": 0.6825, "step": 17613 }, { "epoch": 0.6379572618616444, "grad_norm": 1.4843523607284357, "learning_rate": 6.122421915587584e-06, "loss": 0.7547, "step": 17614 }, { "epoch": 0.6379934806229627, "grad_norm": 0.9467603057672955, "learning_rate": 6.1213406538532105e-06, "loss": 0.7225, "step": 17615 }, { "epoch": 0.6380296993842811, "grad_norm": 1.371885777346394, "learning_rate": 6.120259445492596e-06, "loss": 0.7391, "step": 17616 }, { "epoch": 0.6380659181455994, "grad_norm": 1.2659133388414316, "learning_rate": 6.119178290520623e-06, "loss": 0.7248, "step": 17617 }, { "epoch": 0.6381021369069177, "grad_norm": 1.0811539847624498, "learning_rate": 6.1180971889521644e-06, "loss": 0.7893, "step": 17618 }, { "epoch": 0.6381383556682362, "grad_norm": 1.3275803472989363, "learning_rate": 6.117016140802101e-06, "loss": 0.7197, "step": 17619 }, { "epoch": 0.6381745744295545, "grad_norm": 0.9810074932796229, "learning_rate": 6.115935146085307e-06, "loss": 0.7087, "step": 17620 }, { "epoch": 0.6382107931908729, "grad_norm": 1.423162036998282, "learning_rate": 6.114854204816657e-06, "loss": 0.6705, "step": 17621 }, { "epoch": 0.6382470119521912, "grad_norm": 1.2029014989786682, "learning_rate": 6.1137733170110295e-06, "loss": 0.6423, "step": 17622 }, { "epoch": 0.6382832307135096, "grad_norm": 1.5029130269218394, "learning_rate": 6.112692482683293e-06, "loss": 0.7353, "step": 17623 }, { "epoch": 0.638319449474828, "grad_norm": 1.2633916412947743, "learning_rate": 6.1116117018483256e-06, "loss": 0.6718, "step": 17624 }, { "epoch": 0.6383556682361463, "grad_norm": 1.0561949674031614, "learning_rate": 6.1105309745209965e-06, "loss": 0.723, "step": 17625 }, { "epoch": 0.6383918869974647, "grad_norm": 1.523072343247931, "learning_rate": 6.109450300716181e-06, "loss": 0.7489, "step": 17626 }, { "epoch": 0.6384281057587831, "grad_norm": 1.2361391866263156, "learning_rate": 6.1083696804487445e-06, "loss": 0.6973, "step": 17627 }, { "epoch": 0.6384643245201014, "grad_norm": 1.3233734220622546, "learning_rate": 6.107289113733566e-06, "loss": 0.651, "step": 17628 }, { "epoch": 0.6385005432814198, "grad_norm": 1.0580951177903157, "learning_rate": 6.106208600585504e-06, "loss": 0.7265, "step": 17629 }, { "epoch": 0.6385367620427381, "grad_norm": 1.397314322875817, "learning_rate": 6.105128141019434e-06, "loss": 0.7231, "step": 17630 }, { "epoch": 0.6385729808040566, "grad_norm": 1.3058360891636656, "learning_rate": 6.104047735050227e-06, "loss": 0.7478, "step": 17631 }, { "epoch": 0.6386091995653749, "grad_norm": 1.3380864188054626, "learning_rate": 6.102967382692742e-06, "loss": 0.6972, "step": 17632 }, { "epoch": 0.6386454183266932, "grad_norm": 1.0735221986230241, "learning_rate": 6.101887083961852e-06, "loss": 0.7009, "step": 17633 }, { "epoch": 0.6386816370880116, "grad_norm": 2.0007879951211422, "learning_rate": 6.100806838872422e-06, "loss": 0.6946, "step": 17634 }, { "epoch": 0.6387178558493299, "grad_norm": 0.9932303952434659, "learning_rate": 6.0997266474393145e-06, "loss": 0.6912, "step": 17635 }, { "epoch": 0.6387540746106484, "grad_norm": 1.3451117425860721, "learning_rate": 6.098646509677395e-06, "loss": 0.7071, "step": 17636 }, { "epoch": 0.6387902933719667, "grad_norm": 1.4544955373812287, "learning_rate": 6.097566425601531e-06, "loss": 0.7769, "step": 17637 }, { "epoch": 0.638826512133285, "grad_norm": 1.5786971144594306, "learning_rate": 6.096486395226577e-06, "loss": 0.741, "step": 17638 }, { "epoch": 0.6388627308946034, "grad_norm": 1.2456324300386485, "learning_rate": 6.095406418567405e-06, "loss": 0.7414, "step": 17639 }, { "epoch": 0.6388989496559218, "grad_norm": 1.0733717146238866, "learning_rate": 6.094326495638867e-06, "loss": 0.6851, "step": 17640 }, { "epoch": 0.6389351684172402, "grad_norm": 1.240581135821151, "learning_rate": 6.09324662645583e-06, "loss": 0.697, "step": 17641 }, { "epoch": 0.6389713871785585, "grad_norm": 1.0334933343277999, "learning_rate": 6.092166811033154e-06, "loss": 0.7518, "step": 17642 }, { "epoch": 0.6390076059398768, "grad_norm": 1.3025568121900357, "learning_rate": 6.091087049385693e-06, "loss": 0.6605, "step": 17643 }, { "epoch": 0.6390438247011953, "grad_norm": 1.0881102088226797, "learning_rate": 6.090007341528313e-06, "loss": 0.7313, "step": 17644 }, { "epoch": 0.6390800434625136, "grad_norm": 1.1265671260508276, "learning_rate": 6.088927687475863e-06, "loss": 0.7456, "step": 17645 }, { "epoch": 0.639116262223832, "grad_norm": 1.205690519279595, "learning_rate": 6.087848087243209e-06, "loss": 0.7069, "step": 17646 }, { "epoch": 0.6391524809851503, "grad_norm": 1.35826686109458, "learning_rate": 6.0867685408452005e-06, "loss": 0.6974, "step": 17647 }, { "epoch": 0.6391886997464686, "grad_norm": 1.30110709373528, "learning_rate": 6.085689048296695e-06, "loss": 0.7148, "step": 17648 }, { "epoch": 0.6392249185077871, "grad_norm": 1.2126893555689475, "learning_rate": 6.0846096096125486e-06, "loss": 0.6408, "step": 17649 }, { "epoch": 0.6392611372691054, "grad_norm": 1.4068320672425387, "learning_rate": 6.083530224807614e-06, "loss": 0.772, "step": 17650 }, { "epoch": 0.6392973560304238, "grad_norm": 1.295900939185826, "learning_rate": 6.082450893896746e-06, "loss": 0.7287, "step": 17651 }, { "epoch": 0.6393335747917421, "grad_norm": 1.622269408646196, "learning_rate": 6.081371616894792e-06, "loss": 0.7256, "step": 17652 }, { "epoch": 0.6393697935530605, "grad_norm": 1.482500689877869, "learning_rate": 6.080292393816612e-06, "loss": 0.722, "step": 17653 }, { "epoch": 0.6394060123143789, "grad_norm": 1.3078552404037338, "learning_rate": 6.079213224677049e-06, "loss": 0.6936, "step": 17654 }, { "epoch": 0.6394422310756972, "grad_norm": 1.4350336779404023, "learning_rate": 6.078134109490962e-06, "loss": 0.679, "step": 17655 }, { "epoch": 0.6394784498370156, "grad_norm": 1.4486968830819709, "learning_rate": 6.077055048273193e-06, "loss": 0.6784, "step": 17656 }, { "epoch": 0.639514668598334, "grad_norm": 1.1727343889032766, "learning_rate": 6.075976041038594e-06, "loss": 0.621, "step": 17657 }, { "epoch": 0.6395508873596523, "grad_norm": 1.390064068776085, "learning_rate": 6.074897087802012e-06, "loss": 0.7313, "step": 17658 }, { "epoch": 0.6395871061209707, "grad_norm": 1.1070167015905084, "learning_rate": 6.073818188578295e-06, "loss": 0.7316, "step": 17659 }, { "epoch": 0.639623324882289, "grad_norm": 1.2832217051846142, "learning_rate": 6.07273934338229e-06, "loss": 0.7595, "step": 17660 }, { "epoch": 0.6396595436436074, "grad_norm": 1.4018630489621429, "learning_rate": 6.071660552228842e-06, "loss": 0.7053, "step": 17661 }, { "epoch": 0.6396957624049258, "grad_norm": 1.3701435074347992, "learning_rate": 6.070581815132797e-06, "loss": 0.7476, "step": 17662 }, { "epoch": 0.6397319811662441, "grad_norm": 1.4550529087365076, "learning_rate": 6.069503132108998e-06, "loss": 0.6523, "step": 17663 }, { "epoch": 0.6397681999275625, "grad_norm": 1.0259814607169992, "learning_rate": 6.068424503172292e-06, "loss": 0.6751, "step": 17664 }, { "epoch": 0.6398044186888808, "grad_norm": 1.3456525142548104, "learning_rate": 6.067345928337516e-06, "loss": 0.6171, "step": 17665 }, { "epoch": 0.6398406374501993, "grad_norm": 1.3936650954887322, "learning_rate": 6.06626740761952e-06, "loss": 0.6579, "step": 17666 }, { "epoch": 0.6398768562115176, "grad_norm": 1.43467913631003, "learning_rate": 6.0651889410331355e-06, "loss": 0.6901, "step": 17667 }, { "epoch": 0.6399130749728359, "grad_norm": 1.5348550583551857, "learning_rate": 6.064110528593212e-06, "loss": 0.7401, "step": 17668 }, { "epoch": 0.6399492937341543, "grad_norm": 0.969677556099499, "learning_rate": 6.063032170314585e-06, "loss": 0.6754, "step": 17669 }, { "epoch": 0.6399855124954726, "grad_norm": 1.2674158555248642, "learning_rate": 6.061953866212096e-06, "loss": 0.6851, "step": 17670 }, { "epoch": 0.6400217312567911, "grad_norm": 1.4133327139472205, "learning_rate": 6.060875616300581e-06, "loss": 0.7471, "step": 17671 }, { "epoch": 0.6400579500181094, "grad_norm": 1.296293577163649, "learning_rate": 6.0597974205948775e-06, "loss": 0.6758, "step": 17672 }, { "epoch": 0.6400941687794277, "grad_norm": 1.3609630869727691, "learning_rate": 6.058719279109826e-06, "loss": 0.6883, "step": 17673 }, { "epoch": 0.6401303875407461, "grad_norm": 1.0555803783145201, "learning_rate": 6.057641191860259e-06, "loss": 0.689, "step": 17674 }, { "epoch": 0.6401666063020645, "grad_norm": 1.1461229622612765, "learning_rate": 6.056563158861014e-06, "loss": 0.7113, "step": 17675 }, { "epoch": 0.6402028250633829, "grad_norm": 1.4109730250563817, "learning_rate": 6.0554851801269236e-06, "loss": 0.6768, "step": 17676 }, { "epoch": 0.6402390438247012, "grad_norm": 1.3567325495268865, "learning_rate": 6.054407255672825e-06, "loss": 0.6756, "step": 17677 }, { "epoch": 0.6402752625860195, "grad_norm": 1.3808702794876004, "learning_rate": 6.053329385513548e-06, "loss": 0.7169, "step": 17678 }, { "epoch": 0.640311481347338, "grad_norm": 1.3174602943155531, "learning_rate": 6.052251569663927e-06, "loss": 0.6996, "step": 17679 }, { "epoch": 0.6403477001086563, "grad_norm": 1.3076791143247397, "learning_rate": 6.051173808138793e-06, "loss": 0.7173, "step": 17680 }, { "epoch": 0.6403839188699747, "grad_norm": 1.2493302465796516, "learning_rate": 6.050096100952978e-06, "loss": 0.6451, "step": 17681 }, { "epoch": 0.640420137631293, "grad_norm": 1.0675551824131873, "learning_rate": 6.0490184481213095e-06, "loss": 0.7401, "step": 17682 }, { "epoch": 0.6404563563926113, "grad_norm": 1.3711528045347925, "learning_rate": 6.04794084965862e-06, "loss": 0.6912, "step": 17683 }, { "epoch": 0.6404925751539298, "grad_norm": 1.5244597300471088, "learning_rate": 6.046863305579738e-06, "loss": 0.7488, "step": 17684 }, { "epoch": 0.6405287939152481, "grad_norm": 1.0967125223474297, "learning_rate": 6.045785815899488e-06, "loss": 0.7341, "step": 17685 }, { "epoch": 0.6405650126765665, "grad_norm": 1.266228908521323, "learning_rate": 6.0447083806327025e-06, "loss": 0.7365, "step": 17686 }, { "epoch": 0.6406012314378848, "grad_norm": 1.34859599920215, "learning_rate": 6.043630999794202e-06, "loss": 0.6263, "step": 17687 }, { "epoch": 0.6406374501992032, "grad_norm": 1.2029851915047527, "learning_rate": 6.042553673398819e-06, "loss": 0.6867, "step": 17688 }, { "epoch": 0.6406736689605216, "grad_norm": 1.4994902340620886, "learning_rate": 6.04147640146137e-06, "loss": 0.6195, "step": 17689 }, { "epoch": 0.6407098877218399, "grad_norm": 1.2893286711983603, "learning_rate": 6.040399183996688e-06, "loss": 0.7019, "step": 17690 }, { "epoch": 0.6407461064831583, "grad_norm": 1.4294627773878246, "learning_rate": 6.039322021019591e-06, "loss": 0.7142, "step": 17691 }, { "epoch": 0.6407823252444766, "grad_norm": 1.4931842167643652, "learning_rate": 6.038244912544902e-06, "loss": 0.757, "step": 17692 }, { "epoch": 0.640818544005795, "grad_norm": 1.4061822964344461, "learning_rate": 6.0371678585874475e-06, "loss": 0.7483, "step": 17693 }, { "epoch": 0.6408547627671134, "grad_norm": 1.2789001162631828, "learning_rate": 6.036090859162041e-06, "loss": 0.7209, "step": 17694 }, { "epoch": 0.6408909815284317, "grad_norm": 1.3971899370954193, "learning_rate": 6.0350139142835105e-06, "loss": 0.6829, "step": 17695 }, { "epoch": 0.6409272002897501, "grad_norm": 1.035305219572262, "learning_rate": 6.033937023966671e-06, "loss": 0.7542, "step": 17696 }, { "epoch": 0.6409634190510685, "grad_norm": 1.034672402493463, "learning_rate": 6.032860188226344e-06, "loss": 0.6879, "step": 17697 }, { "epoch": 0.6409996378123868, "grad_norm": 1.3704637108293605, "learning_rate": 6.031783407077345e-06, "loss": 0.7076, "step": 17698 }, { "epoch": 0.6410358565737052, "grad_norm": 1.260374587383844, "learning_rate": 6.030706680534495e-06, "loss": 0.6572, "step": 17699 }, { "epoch": 0.6410720753350235, "grad_norm": 1.502944891681072, "learning_rate": 6.029630008612608e-06, "loss": 0.7945, "step": 17700 }, { "epoch": 0.641108294096342, "grad_norm": 1.0692023851440342, "learning_rate": 6.028553391326502e-06, "loss": 0.6962, "step": 17701 }, { "epoch": 0.6411445128576603, "grad_norm": 1.422924897314156, "learning_rate": 6.027476828690988e-06, "loss": 0.6413, "step": 17702 }, { "epoch": 0.6411807316189786, "grad_norm": 1.3301562210716174, "learning_rate": 6.026400320720884e-06, "loss": 0.6581, "step": 17703 }, { "epoch": 0.641216950380297, "grad_norm": 1.641670872338793, "learning_rate": 6.025323867431005e-06, "loss": 0.7708, "step": 17704 }, { "epoch": 0.6412531691416153, "grad_norm": 1.0826871437414967, "learning_rate": 6.024247468836161e-06, "loss": 0.6947, "step": 17705 }, { "epoch": 0.6412893879029338, "grad_norm": 1.3243152293502498, "learning_rate": 6.0231711249511655e-06, "loss": 0.6834, "step": 17706 }, { "epoch": 0.6413256066642521, "grad_norm": 0.9607152713375552, "learning_rate": 6.022094835790829e-06, "loss": 0.7266, "step": 17707 }, { "epoch": 0.6413618254255704, "grad_norm": 1.433394720985178, "learning_rate": 6.021018601369966e-06, "loss": 0.7295, "step": 17708 }, { "epoch": 0.6413980441868888, "grad_norm": 1.3861627283371272, "learning_rate": 6.019942421703379e-06, "loss": 0.7054, "step": 17709 }, { "epoch": 0.6414342629482072, "grad_norm": 1.2674978727495902, "learning_rate": 6.018866296805883e-06, "loss": 0.6908, "step": 17710 }, { "epoch": 0.6414704817095256, "grad_norm": 1.525717492683217, "learning_rate": 6.017790226692285e-06, "loss": 0.7115, "step": 17711 }, { "epoch": 0.6415067004708439, "grad_norm": 1.4463277627428015, "learning_rate": 6.016714211377393e-06, "loss": 0.7138, "step": 17712 }, { "epoch": 0.6415429192321622, "grad_norm": 1.3808819028427686, "learning_rate": 6.015638250876013e-06, "loss": 0.7306, "step": 17713 }, { "epoch": 0.6415791379934807, "grad_norm": 1.7120882795419388, "learning_rate": 6.01456234520295e-06, "loss": 0.7302, "step": 17714 }, { "epoch": 0.641615356754799, "grad_norm": 1.2616609642803014, "learning_rate": 6.013486494373014e-06, "loss": 0.6796, "step": 17715 }, { "epoch": 0.6416515755161174, "grad_norm": 1.4485543879778773, "learning_rate": 6.012410698401006e-06, "loss": 0.7756, "step": 17716 }, { "epoch": 0.6416877942774357, "grad_norm": 1.444407768862006, "learning_rate": 6.011334957301731e-06, "loss": 0.6636, "step": 17717 }, { "epoch": 0.641724013038754, "grad_norm": 1.3460798331463948, "learning_rate": 6.010259271089989e-06, "loss": 0.6503, "step": 17718 }, { "epoch": 0.6417602318000725, "grad_norm": 1.051992563518726, "learning_rate": 6.009183639780588e-06, "loss": 0.6959, "step": 17719 }, { "epoch": 0.6417964505613908, "grad_norm": 1.298464915035503, "learning_rate": 6.008108063388325e-06, "loss": 0.7026, "step": 17720 }, { "epoch": 0.6418326693227092, "grad_norm": 1.4080253983615265, "learning_rate": 6.007032541928004e-06, "loss": 0.6547, "step": 17721 }, { "epoch": 0.6418688880840275, "grad_norm": 1.4952421615762985, "learning_rate": 6.005957075414423e-06, "loss": 0.763, "step": 17722 }, { "epoch": 0.6419051068453459, "grad_norm": 1.179997963721575, "learning_rate": 6.004881663862381e-06, "loss": 0.7465, "step": 17723 }, { "epoch": 0.6419413256066643, "grad_norm": 1.0987190426788491, "learning_rate": 6.003806307286679e-06, "loss": 0.7166, "step": 17724 }, { "epoch": 0.6419775443679826, "grad_norm": 1.4933540230888878, "learning_rate": 6.002731005702114e-06, "loss": 0.6802, "step": 17725 }, { "epoch": 0.642013763129301, "grad_norm": 1.4605750906831434, "learning_rate": 6.001655759123481e-06, "loss": 0.7152, "step": 17726 }, { "epoch": 0.6420499818906193, "grad_norm": 1.3702640862541249, "learning_rate": 6.000580567565579e-06, "loss": 0.7011, "step": 17727 }, { "epoch": 0.6420862006519377, "grad_norm": 1.0840709341345636, "learning_rate": 5.999505431043204e-06, "loss": 0.7787, "step": 17728 }, { "epoch": 0.6421224194132561, "grad_norm": 1.4655495329340198, "learning_rate": 5.9984303495711474e-06, "loss": 0.6887, "step": 17729 }, { "epoch": 0.6421586381745744, "grad_norm": 1.3282102199870025, "learning_rate": 5.997355323164207e-06, "loss": 0.6598, "step": 17730 }, { "epoch": 0.6421948569358928, "grad_norm": 1.3117496759757519, "learning_rate": 5.996280351837173e-06, "loss": 0.7088, "step": 17731 }, { "epoch": 0.6422310756972112, "grad_norm": 1.5305184799242793, "learning_rate": 5.995205435604841e-06, "loss": 0.6533, "step": 17732 }, { "epoch": 0.6422672944585295, "grad_norm": 1.4963744749232883, "learning_rate": 5.994130574481997e-06, "loss": 0.7446, "step": 17733 }, { "epoch": 0.6423035132198479, "grad_norm": 1.3694759133851029, "learning_rate": 5.99305576848344e-06, "loss": 0.768, "step": 17734 }, { "epoch": 0.6423397319811662, "grad_norm": 1.2798505516765688, "learning_rate": 5.9919810176239554e-06, "loss": 0.713, "step": 17735 }, { "epoch": 0.6423759507424847, "grad_norm": 1.5383010774344954, "learning_rate": 5.990906321918333e-06, "loss": 0.7342, "step": 17736 }, { "epoch": 0.642412169503803, "grad_norm": 1.2515583394102883, "learning_rate": 5.989831681381364e-06, "loss": 0.7281, "step": 17737 }, { "epoch": 0.6424483882651213, "grad_norm": 1.044607894953971, "learning_rate": 5.988757096027833e-06, "loss": 0.6964, "step": 17738 }, { "epoch": 0.6424846070264397, "grad_norm": 1.1116087564741204, "learning_rate": 5.987682565872532e-06, "loss": 0.7806, "step": 17739 }, { "epoch": 0.642520825787758, "grad_norm": 1.3921643675896962, "learning_rate": 5.986608090930241e-06, "loss": 0.7213, "step": 17740 }, { "epoch": 0.6425570445490765, "grad_norm": 1.2220064499214813, "learning_rate": 5.985533671215751e-06, "loss": 0.6753, "step": 17741 }, { "epoch": 0.6425932633103948, "grad_norm": 1.3890590483436995, "learning_rate": 5.984459306743844e-06, "loss": 0.6963, "step": 17742 }, { "epoch": 0.6426294820717131, "grad_norm": 1.2608766713058992, "learning_rate": 5.9833849975293065e-06, "loss": 0.6388, "step": 17743 }, { "epoch": 0.6426657008330315, "grad_norm": 1.416880522486239, "learning_rate": 5.9823107435869184e-06, "loss": 0.7979, "step": 17744 }, { "epoch": 0.6427019195943499, "grad_norm": 1.235076856037009, "learning_rate": 5.981236544931466e-06, "loss": 0.681, "step": 17745 }, { "epoch": 0.6427381383556683, "grad_norm": 1.2913390580814244, "learning_rate": 5.980162401577732e-06, "loss": 0.6928, "step": 17746 }, { "epoch": 0.6427743571169866, "grad_norm": 1.307529352498676, "learning_rate": 5.979088313540492e-06, "loss": 0.7139, "step": 17747 }, { "epoch": 0.6428105758783049, "grad_norm": 1.1653273229890295, "learning_rate": 5.978014280834532e-06, "loss": 0.7367, "step": 17748 }, { "epoch": 0.6428467946396234, "grad_norm": 1.3195275529593826, "learning_rate": 5.9769403034746295e-06, "loss": 0.7426, "step": 17749 }, { "epoch": 0.6428830134009417, "grad_norm": 0.9036074236217494, "learning_rate": 5.975866381475564e-06, "loss": 0.6994, "step": 17750 }, { "epoch": 0.6429192321622601, "grad_norm": 1.331389644623561, "learning_rate": 5.974792514852111e-06, "loss": 0.634, "step": 17751 }, { "epoch": 0.6429554509235784, "grad_norm": 1.2920382203878145, "learning_rate": 5.973718703619051e-06, "loss": 0.7011, "step": 17752 }, { "epoch": 0.6429916696848967, "grad_norm": 1.2498050679264245, "learning_rate": 5.97264494779116e-06, "loss": 0.7209, "step": 17753 }, { "epoch": 0.6430278884462152, "grad_norm": 0.9815898172373949, "learning_rate": 5.971571247383211e-06, "loss": 0.7115, "step": 17754 }, { "epoch": 0.6430641072075335, "grad_norm": 1.2720326605351926, "learning_rate": 5.970497602409984e-06, "loss": 0.7052, "step": 17755 }, { "epoch": 0.6431003259688519, "grad_norm": 1.3158997698995767, "learning_rate": 5.969424012886249e-06, "loss": 0.7128, "step": 17756 }, { "epoch": 0.6431365447301702, "grad_norm": 1.2353390772192365, "learning_rate": 5.968350478826782e-06, "loss": 0.6719, "step": 17757 }, { "epoch": 0.6431727634914886, "grad_norm": 1.2532970546985707, "learning_rate": 5.9672770002463545e-06, "loss": 0.7538, "step": 17758 }, { "epoch": 0.643208982252807, "grad_norm": 1.4741658080459532, "learning_rate": 5.96620357715974e-06, "loss": 0.7196, "step": 17759 }, { "epoch": 0.6432452010141253, "grad_norm": 1.3344657051918736, "learning_rate": 5.965130209581706e-06, "loss": 0.764, "step": 17760 }, { "epoch": 0.6432814197754437, "grad_norm": 1.0397683729639202, "learning_rate": 5.964056897527028e-06, "loss": 0.7366, "step": 17761 }, { "epoch": 0.643317638536762, "grad_norm": 1.336051994570623, "learning_rate": 5.962983641010473e-06, "loss": 0.714, "step": 17762 }, { "epoch": 0.6433538572980804, "grad_norm": 1.277779081367767, "learning_rate": 5.961910440046811e-06, "loss": 0.7194, "step": 17763 }, { "epoch": 0.6433900760593988, "grad_norm": 0.9695394498798099, "learning_rate": 5.960837294650807e-06, "loss": 0.6519, "step": 17764 }, { "epoch": 0.6434262948207171, "grad_norm": 1.339702883091063, "learning_rate": 5.959764204837233e-06, "loss": 0.7567, "step": 17765 }, { "epoch": 0.6434625135820355, "grad_norm": 1.3713995272550927, "learning_rate": 5.958691170620854e-06, "loss": 0.7519, "step": 17766 }, { "epoch": 0.6434987323433539, "grad_norm": 1.0485066218808934, "learning_rate": 5.9576181920164336e-06, "loss": 0.7335, "step": 17767 }, { "epoch": 0.6435349511046722, "grad_norm": 1.4238498925479952, "learning_rate": 5.95654526903874e-06, "loss": 0.7417, "step": 17768 }, { "epoch": 0.6435711698659906, "grad_norm": 1.3282208281894934, "learning_rate": 5.955472401702535e-06, "loss": 0.6926, "step": 17769 }, { "epoch": 0.6436073886273089, "grad_norm": 1.4549959281132359, "learning_rate": 5.954399590022587e-06, "loss": 0.7956, "step": 17770 }, { "epoch": 0.6436436073886274, "grad_norm": 1.3952207583739082, "learning_rate": 5.953326834013653e-06, "loss": 0.6739, "step": 17771 }, { "epoch": 0.6436798261499457, "grad_norm": 1.2879768741873918, "learning_rate": 5.952254133690498e-06, "loss": 0.7206, "step": 17772 }, { "epoch": 0.643716044911264, "grad_norm": 1.2888366421882127, "learning_rate": 5.951181489067882e-06, "loss": 0.6746, "step": 17773 }, { "epoch": 0.6437522636725824, "grad_norm": 1.3222125759780767, "learning_rate": 5.950108900160568e-06, "loss": 0.6793, "step": 17774 }, { "epoch": 0.6437884824339007, "grad_norm": 1.3332829844634395, "learning_rate": 5.949036366983311e-06, "loss": 0.7037, "step": 17775 }, { "epoch": 0.6438247011952192, "grad_norm": 1.3234239066370526, "learning_rate": 5.947963889550875e-06, "loss": 0.6639, "step": 17776 }, { "epoch": 0.6438609199565375, "grad_norm": 1.2785394195209574, "learning_rate": 5.946891467878018e-06, "loss": 0.6857, "step": 17777 }, { "epoch": 0.6438971387178558, "grad_norm": 1.2824515387212274, "learning_rate": 5.945819101979493e-06, "loss": 0.6881, "step": 17778 }, { "epoch": 0.6439333574791742, "grad_norm": 1.4455877834496609, "learning_rate": 5.944746791870062e-06, "loss": 0.7902, "step": 17779 }, { "epoch": 0.6439695762404926, "grad_norm": 1.3375380566621806, "learning_rate": 5.943674537564475e-06, "loss": 0.6911, "step": 17780 }, { "epoch": 0.644005795001811, "grad_norm": 1.4204332225889744, "learning_rate": 5.942602339077495e-06, "loss": 0.6575, "step": 17781 }, { "epoch": 0.6440420137631293, "grad_norm": 1.3396774194310117, "learning_rate": 5.9415301964238685e-06, "loss": 0.6912, "step": 17782 }, { "epoch": 0.6440782325244476, "grad_norm": 1.6650737814765393, "learning_rate": 5.940458109618354e-06, "loss": 0.7751, "step": 17783 }, { "epoch": 0.644114451285766, "grad_norm": 1.261619820864694, "learning_rate": 5.939386078675702e-06, "loss": 0.7145, "step": 17784 }, { "epoch": 0.6441506700470844, "grad_norm": 1.2689812429812535, "learning_rate": 5.938314103610665e-06, "loss": 0.7045, "step": 17785 }, { "epoch": 0.6441868888084028, "grad_norm": 1.3581704456412274, "learning_rate": 5.937242184437997e-06, "loss": 0.7094, "step": 17786 }, { "epoch": 0.6442231075697211, "grad_norm": 1.3060632343734218, "learning_rate": 5.936170321172444e-06, "loss": 0.7132, "step": 17787 }, { "epoch": 0.6442593263310394, "grad_norm": 1.3668845086187789, "learning_rate": 5.935098513828761e-06, "loss": 0.7381, "step": 17788 }, { "epoch": 0.6442955450923579, "grad_norm": 1.4347965615185656, "learning_rate": 5.934026762421692e-06, "loss": 0.6821, "step": 17789 }, { "epoch": 0.6443317638536762, "grad_norm": 0.9984268794894997, "learning_rate": 5.932955066965989e-06, "loss": 0.6786, "step": 17790 }, { "epoch": 0.6443679826149946, "grad_norm": 1.4269875648889183, "learning_rate": 5.931883427476396e-06, "loss": 0.724, "step": 17791 }, { "epoch": 0.6444042013763129, "grad_norm": 1.2507545257765642, "learning_rate": 5.930811843967663e-06, "loss": 0.657, "step": 17792 }, { "epoch": 0.6444404201376313, "grad_norm": 1.1759480455053237, "learning_rate": 5.929740316454533e-06, "loss": 0.699, "step": 17793 }, { "epoch": 0.6444766388989497, "grad_norm": 1.2656424967857376, "learning_rate": 5.928668844951755e-06, "loss": 0.7416, "step": 17794 }, { "epoch": 0.644512857660268, "grad_norm": 1.1306115126502212, "learning_rate": 5.9275974294740705e-06, "loss": 0.7356, "step": 17795 }, { "epoch": 0.6445490764215864, "grad_norm": 1.3206501252564087, "learning_rate": 5.926526070036221e-06, "loss": 0.7113, "step": 17796 }, { "epoch": 0.6445852951829047, "grad_norm": 1.301777086165333, "learning_rate": 5.925454766652955e-06, "loss": 0.6788, "step": 17797 }, { "epoch": 0.6446215139442231, "grad_norm": 1.3834674356443168, "learning_rate": 5.92438351933901e-06, "loss": 0.6536, "step": 17798 }, { "epoch": 0.6446577327055415, "grad_norm": 1.2516438493472217, "learning_rate": 5.923312328109131e-06, "loss": 0.6384, "step": 17799 }, { "epoch": 0.6446939514668598, "grad_norm": 1.3660010434954994, "learning_rate": 5.922241192978053e-06, "loss": 0.7243, "step": 17800 }, { "epoch": 0.6447301702281782, "grad_norm": 1.3752469794665467, "learning_rate": 5.921170113960524e-06, "loss": 0.7354, "step": 17801 }, { "epoch": 0.6447663889894966, "grad_norm": 1.181121962435696, "learning_rate": 5.920099091071274e-06, "loss": 0.6638, "step": 17802 }, { "epoch": 0.6448026077508149, "grad_norm": 1.2970632242722218, "learning_rate": 5.919028124325048e-06, "loss": 0.6429, "step": 17803 }, { "epoch": 0.6448388265121333, "grad_norm": 1.3481342380217776, "learning_rate": 5.917957213736578e-06, "loss": 0.7508, "step": 17804 }, { "epoch": 0.6448750452734516, "grad_norm": 1.1246357148349728, "learning_rate": 5.91688635932061e-06, "loss": 0.6794, "step": 17805 }, { "epoch": 0.6449112640347701, "grad_norm": 1.2988095762606215, "learning_rate": 5.9158155610918685e-06, "loss": 0.7025, "step": 17806 }, { "epoch": 0.6449474827960884, "grad_norm": 1.0355756606125217, "learning_rate": 5.914744819065093e-06, "loss": 0.6073, "step": 17807 }, { "epoch": 0.6449837015574067, "grad_norm": 1.4152685193926628, "learning_rate": 5.913674133255019e-06, "loss": 0.6397, "step": 17808 }, { "epoch": 0.6450199203187251, "grad_norm": 1.525625726065957, "learning_rate": 5.912603503676381e-06, "loss": 0.7407, "step": 17809 }, { "epoch": 0.6450561390800434, "grad_norm": 1.0289341669908916, "learning_rate": 5.911532930343909e-06, "loss": 0.7143, "step": 17810 }, { "epoch": 0.6450923578413619, "grad_norm": 1.1048358350272047, "learning_rate": 5.910462413272337e-06, "loss": 0.7839, "step": 17811 }, { "epoch": 0.6451285766026802, "grad_norm": 1.1267939868719363, "learning_rate": 5.9093919524763935e-06, "loss": 0.6285, "step": 17812 }, { "epoch": 0.6451647953639985, "grad_norm": 1.2735920341315403, "learning_rate": 5.9083215479708134e-06, "loss": 0.684, "step": 17813 }, { "epoch": 0.6452010141253169, "grad_norm": 0.930213876379691, "learning_rate": 5.907251199770323e-06, "loss": 0.7153, "step": 17814 }, { "epoch": 0.6452372328866353, "grad_norm": 1.3863805976454062, "learning_rate": 5.906180907889654e-06, "loss": 0.6598, "step": 17815 }, { "epoch": 0.6452734516479537, "grad_norm": 1.3508542484594093, "learning_rate": 5.905110672343529e-06, "loss": 0.6628, "step": 17816 }, { "epoch": 0.645309670409272, "grad_norm": 1.1027684957094854, "learning_rate": 5.904040493146683e-06, "loss": 0.669, "step": 17817 }, { "epoch": 0.6453458891705903, "grad_norm": 1.2381623952666452, "learning_rate": 5.902970370313837e-06, "loss": 0.7072, "step": 17818 }, { "epoch": 0.6453821079319088, "grad_norm": 1.336701676501021, "learning_rate": 5.901900303859717e-06, "loss": 0.7286, "step": 17819 }, { "epoch": 0.6454183266932271, "grad_norm": 1.4400183119340106, "learning_rate": 5.900830293799054e-06, "loss": 0.774, "step": 17820 }, { "epoch": 0.6454545454545455, "grad_norm": 1.3399330017680342, "learning_rate": 5.899760340146564e-06, "loss": 0.6724, "step": 17821 }, { "epoch": 0.6454907642158638, "grad_norm": 0.9826360081006388, "learning_rate": 5.898690442916978e-06, "loss": 0.7803, "step": 17822 }, { "epoch": 0.6455269829771821, "grad_norm": 1.191322752105987, "learning_rate": 5.897620602125011e-06, "loss": 0.6787, "step": 17823 }, { "epoch": 0.6455632017385006, "grad_norm": 1.2170618742262076, "learning_rate": 5.896550817785394e-06, "loss": 0.6406, "step": 17824 }, { "epoch": 0.6455994204998189, "grad_norm": 1.1081728224583076, "learning_rate": 5.89548108991284e-06, "loss": 0.7134, "step": 17825 }, { "epoch": 0.6456356392611373, "grad_norm": 1.4910916528664049, "learning_rate": 5.894411418522075e-06, "loss": 0.7606, "step": 17826 }, { "epoch": 0.6456718580224556, "grad_norm": 1.2026936706922784, "learning_rate": 5.893341803627814e-06, "loss": 0.6614, "step": 17827 }, { "epoch": 0.645708076783774, "grad_norm": 0.9382330687007218, "learning_rate": 5.892272245244779e-06, "loss": 0.6739, "step": 17828 }, { "epoch": 0.6457442955450924, "grad_norm": 1.2716213304912376, "learning_rate": 5.891202743387689e-06, "loss": 0.7046, "step": 17829 }, { "epoch": 0.6457805143064107, "grad_norm": 1.4874593387584047, "learning_rate": 5.890133298071257e-06, "loss": 0.7456, "step": 17830 }, { "epoch": 0.6458167330677291, "grad_norm": 1.3795052137952115, "learning_rate": 5.889063909310204e-06, "loss": 0.7728, "step": 17831 }, { "epoch": 0.6458529518290475, "grad_norm": 1.0563481986328422, "learning_rate": 5.887994577119242e-06, "loss": 0.6825, "step": 17832 }, { "epoch": 0.6458891705903658, "grad_norm": 0.9957125944868245, "learning_rate": 5.88692530151309e-06, "loss": 0.6826, "step": 17833 }, { "epoch": 0.6459253893516842, "grad_norm": 0.9671927757417319, "learning_rate": 5.885856082506458e-06, "loss": 0.6953, "step": 17834 }, { "epoch": 0.6459616081130025, "grad_norm": 1.356079449071109, "learning_rate": 5.884786920114063e-06, "loss": 0.7382, "step": 17835 }, { "epoch": 0.645997826874321, "grad_norm": 1.3759211156109137, "learning_rate": 5.883717814350614e-06, "loss": 0.6789, "step": 17836 }, { "epoch": 0.6460340456356393, "grad_norm": 1.747958192760189, "learning_rate": 5.8826487652308275e-06, "loss": 0.7454, "step": 17837 }, { "epoch": 0.6460702643969576, "grad_norm": 1.3288518648663905, "learning_rate": 5.881579772769408e-06, "loss": 0.7448, "step": 17838 }, { "epoch": 0.646106483158276, "grad_norm": 1.2001812485343442, "learning_rate": 5.880510836981071e-06, "loss": 0.6998, "step": 17839 }, { "epoch": 0.6461427019195943, "grad_norm": 1.2575860278681135, "learning_rate": 5.879441957880526e-06, "loss": 0.6537, "step": 17840 }, { "epoch": 0.6461789206809128, "grad_norm": 0.9272588850256983, "learning_rate": 5.878373135482479e-06, "loss": 0.6792, "step": 17841 }, { "epoch": 0.6462151394422311, "grad_norm": 1.045167011903423, "learning_rate": 5.877304369801641e-06, "loss": 0.7889, "step": 17842 }, { "epoch": 0.6462513582035494, "grad_norm": 1.0423733414430907, "learning_rate": 5.876235660852715e-06, "loss": 0.7454, "step": 17843 }, { "epoch": 0.6462875769648678, "grad_norm": 1.3241048759684355, "learning_rate": 5.875167008650412e-06, "loss": 0.7282, "step": 17844 }, { "epoch": 0.6463237957261861, "grad_norm": 1.38775241634988, "learning_rate": 5.8740984132094326e-06, "loss": 0.6992, "step": 17845 }, { "epoch": 0.6463600144875046, "grad_norm": 1.45447704137746, "learning_rate": 5.873029874544487e-06, "loss": 0.7468, "step": 17846 }, { "epoch": 0.6463962332488229, "grad_norm": 1.3618124157890539, "learning_rate": 5.871961392670275e-06, "loss": 0.6467, "step": 17847 }, { "epoch": 0.6464324520101412, "grad_norm": 1.243154196838644, "learning_rate": 5.870892967601504e-06, "loss": 0.6696, "step": 17848 }, { "epoch": 0.6464686707714596, "grad_norm": 1.3805078506869097, "learning_rate": 5.8698245993528715e-06, "loss": 0.6584, "step": 17849 }, { "epoch": 0.646504889532778, "grad_norm": 1.697549596456633, "learning_rate": 5.86875628793908e-06, "loss": 0.728, "step": 17850 }, { "epoch": 0.6465411082940964, "grad_norm": 1.2477727269531689, "learning_rate": 5.867688033374836e-06, "loss": 0.641, "step": 17851 }, { "epoch": 0.6465773270554147, "grad_norm": 1.5089953679635564, "learning_rate": 5.866619835674832e-06, "loss": 0.7181, "step": 17852 }, { "epoch": 0.646613545816733, "grad_norm": 1.3576326323261663, "learning_rate": 5.865551694853774e-06, "loss": 0.6813, "step": 17853 }, { "epoch": 0.6466497645780515, "grad_norm": 1.3138238001170013, "learning_rate": 5.864483610926355e-06, "loss": 0.7343, "step": 17854 }, { "epoch": 0.6466859833393698, "grad_norm": 1.3569451894741682, "learning_rate": 5.863415583907277e-06, "loss": 0.755, "step": 17855 }, { "epoch": 0.6467222021006882, "grad_norm": 1.2828789472794746, "learning_rate": 5.862347613811234e-06, "loss": 0.6736, "step": 17856 }, { "epoch": 0.6467584208620065, "grad_norm": 1.382424358673969, "learning_rate": 5.861279700652927e-06, "loss": 0.6805, "step": 17857 }, { "epoch": 0.6467946396233248, "grad_norm": 1.2143566851922336, "learning_rate": 5.860211844447043e-06, "loss": 0.673, "step": 17858 }, { "epoch": 0.6468308583846433, "grad_norm": 1.2961015513370997, "learning_rate": 5.859144045208283e-06, "loss": 0.72, "step": 17859 }, { "epoch": 0.6468670771459616, "grad_norm": 1.0679270333288828, "learning_rate": 5.858076302951341e-06, "loss": 0.66, "step": 17860 }, { "epoch": 0.64690329590728, "grad_norm": 1.4004721444901818, "learning_rate": 5.857008617690907e-06, "loss": 0.6886, "step": 17861 }, { "epoch": 0.6469395146685983, "grad_norm": 1.4099789293977263, "learning_rate": 5.8559409894416765e-06, "loss": 0.6887, "step": 17862 }, { "epoch": 0.6469757334299167, "grad_norm": 1.1819929787704917, "learning_rate": 5.854873418218337e-06, "loss": 0.6817, "step": 17863 }, { "epoch": 0.6470119521912351, "grad_norm": 1.6350281047775117, "learning_rate": 5.853805904035583e-06, "loss": 0.7113, "step": 17864 }, { "epoch": 0.6470481709525534, "grad_norm": 1.243257393221559, "learning_rate": 5.852738446908103e-06, "loss": 0.6174, "step": 17865 }, { "epoch": 0.6470843897138718, "grad_norm": 1.4039604065118372, "learning_rate": 5.851671046850587e-06, "loss": 0.6996, "step": 17866 }, { "epoch": 0.6471206084751902, "grad_norm": 1.2865911889265356, "learning_rate": 5.85060370387772e-06, "loss": 0.6875, "step": 17867 }, { "epoch": 0.6471568272365085, "grad_norm": 1.3452842749870204, "learning_rate": 5.849536418004194e-06, "loss": 0.7449, "step": 17868 }, { "epoch": 0.6471930459978269, "grad_norm": 1.258853188620732, "learning_rate": 5.8484691892446925e-06, "loss": 0.6955, "step": 17869 }, { "epoch": 0.6472292647591452, "grad_norm": 1.4287006471671395, "learning_rate": 5.847402017613902e-06, "loss": 0.7697, "step": 17870 }, { "epoch": 0.6472654835204636, "grad_norm": 1.0082222496640059, "learning_rate": 5.846334903126511e-06, "loss": 0.7057, "step": 17871 }, { "epoch": 0.647301702281782, "grad_norm": 1.3913433451090205, "learning_rate": 5.845267845797199e-06, "loss": 0.6955, "step": 17872 }, { "epoch": 0.6473379210431003, "grad_norm": 1.478125822543869, "learning_rate": 5.844200845640654e-06, "loss": 0.7287, "step": 17873 }, { "epoch": 0.6473741398044187, "grad_norm": 1.4485954103428447, "learning_rate": 5.843133902671556e-06, "loss": 0.7103, "step": 17874 }, { "epoch": 0.647410358565737, "grad_norm": 1.05301067389203, "learning_rate": 5.842067016904589e-06, "loss": 0.8012, "step": 17875 }, { "epoch": 0.6474465773270555, "grad_norm": 1.3303390560676223, "learning_rate": 5.841000188354432e-06, "loss": 0.6858, "step": 17876 }, { "epoch": 0.6474827960883738, "grad_norm": 1.3571455914471624, "learning_rate": 5.839933417035768e-06, "loss": 0.727, "step": 17877 }, { "epoch": 0.6475190148496921, "grad_norm": 1.3251899242642973, "learning_rate": 5.838866702963275e-06, "loss": 0.6826, "step": 17878 }, { "epoch": 0.6475552336110105, "grad_norm": 1.130426741774063, "learning_rate": 5.837800046151631e-06, "loss": 0.7319, "step": 17879 }, { "epoch": 0.6475914523723288, "grad_norm": 1.2439966934464792, "learning_rate": 5.836733446615519e-06, "loss": 0.6638, "step": 17880 }, { "epoch": 0.6476276711336473, "grad_norm": 1.4320632250993806, "learning_rate": 5.835666904369611e-06, "loss": 0.7249, "step": 17881 }, { "epoch": 0.6476638898949656, "grad_norm": 1.3120290537158537, "learning_rate": 5.834600419428583e-06, "loss": 0.7521, "step": 17882 }, { "epoch": 0.6477001086562839, "grad_norm": 1.317080932921364, "learning_rate": 5.833533991807117e-06, "loss": 0.6189, "step": 17883 }, { "epoch": 0.6477363274176023, "grad_norm": 1.2690407964695902, "learning_rate": 5.832467621519884e-06, "loss": 0.702, "step": 17884 }, { "epoch": 0.6477725461789207, "grad_norm": 1.2554785928358163, "learning_rate": 5.831401308581559e-06, "loss": 0.6824, "step": 17885 }, { "epoch": 0.6478087649402391, "grad_norm": 1.310738123911739, "learning_rate": 5.83033505300681e-06, "loss": 0.6915, "step": 17886 }, { "epoch": 0.6478449837015574, "grad_norm": 1.2986473889909045, "learning_rate": 5.829268854810319e-06, "loss": 0.6672, "step": 17887 }, { "epoch": 0.6478812024628757, "grad_norm": 1.3602850131634567, "learning_rate": 5.828202714006753e-06, "loss": 0.5718, "step": 17888 }, { "epoch": 0.6479174212241942, "grad_norm": 1.2797263996851567, "learning_rate": 5.827136630610783e-06, "loss": 0.6608, "step": 17889 }, { "epoch": 0.6479536399855125, "grad_norm": 1.490452546714353, "learning_rate": 5.826070604637078e-06, "loss": 0.7571, "step": 17890 }, { "epoch": 0.6479898587468309, "grad_norm": 1.3821364322043181, "learning_rate": 5.8250046361003095e-06, "loss": 0.7373, "step": 17891 }, { "epoch": 0.6480260775081492, "grad_norm": 1.1876803405500564, "learning_rate": 5.823938725015148e-06, "loss": 0.631, "step": 17892 }, { "epoch": 0.6480622962694675, "grad_norm": 1.0979040103188484, "learning_rate": 5.822872871396255e-06, "loss": 0.729, "step": 17893 }, { "epoch": 0.648098515030786, "grad_norm": 1.555770682538411, "learning_rate": 5.821807075258303e-06, "loss": 0.7452, "step": 17894 }, { "epoch": 0.6481347337921043, "grad_norm": 1.4157231351066537, "learning_rate": 5.820741336615961e-06, "loss": 0.6719, "step": 17895 }, { "epoch": 0.6481709525534227, "grad_norm": 1.3940607286915305, "learning_rate": 5.819675655483888e-06, "loss": 0.6975, "step": 17896 }, { "epoch": 0.648207171314741, "grad_norm": 1.5827580513207846, "learning_rate": 5.818610031876746e-06, "loss": 0.7309, "step": 17897 }, { "epoch": 0.6482433900760594, "grad_norm": 1.31215952745867, "learning_rate": 5.817544465809209e-06, "loss": 0.6818, "step": 17898 }, { "epoch": 0.6482796088373778, "grad_norm": 1.3268357807567788, "learning_rate": 5.816478957295937e-06, "loss": 0.7626, "step": 17899 }, { "epoch": 0.6483158275986961, "grad_norm": 1.262944065177198, "learning_rate": 5.815413506351584e-06, "loss": 0.7063, "step": 17900 }, { "epoch": 0.6483520463600145, "grad_norm": 1.8491994852755567, "learning_rate": 5.814348112990824e-06, "loss": 0.7511, "step": 17901 }, { "epoch": 0.6483882651213329, "grad_norm": 1.500022183939685, "learning_rate": 5.8132827772283105e-06, "loss": 0.6882, "step": 17902 }, { "epoch": 0.6484244838826512, "grad_norm": 1.3299329919377876, "learning_rate": 5.812217499078705e-06, "loss": 0.759, "step": 17903 }, { "epoch": 0.6484607026439696, "grad_norm": 1.3179793025449642, "learning_rate": 5.8111522785566635e-06, "loss": 0.7221, "step": 17904 }, { "epoch": 0.6484969214052879, "grad_norm": 1.3432191362221566, "learning_rate": 5.81008711567685e-06, "loss": 0.6725, "step": 17905 }, { "epoch": 0.6485331401666063, "grad_norm": 0.8880065773486346, "learning_rate": 5.80902201045392e-06, "loss": 0.6949, "step": 17906 }, { "epoch": 0.6485693589279247, "grad_norm": 1.64178375127259, "learning_rate": 5.80795696290253e-06, "loss": 0.7242, "step": 17907 }, { "epoch": 0.648605577689243, "grad_norm": 1.359888186817399, "learning_rate": 5.806891973037331e-06, "loss": 0.7205, "step": 17908 }, { "epoch": 0.6486417964505614, "grad_norm": 1.3883445933525722, "learning_rate": 5.805827040872986e-06, "loss": 0.6202, "step": 17909 }, { "epoch": 0.6486780152118797, "grad_norm": 1.3840809327847026, "learning_rate": 5.804762166424147e-06, "loss": 0.7382, "step": 17910 }, { "epoch": 0.6487142339731982, "grad_norm": 1.300132527410597, "learning_rate": 5.803697349705464e-06, "loss": 0.6861, "step": 17911 }, { "epoch": 0.6487504527345165, "grad_norm": 1.0290669638497252, "learning_rate": 5.802632590731596e-06, "loss": 0.6781, "step": 17912 }, { "epoch": 0.6487866714958348, "grad_norm": 1.493693638929383, "learning_rate": 5.801567889517191e-06, "loss": 0.726, "step": 17913 }, { "epoch": 0.6488228902571532, "grad_norm": 1.3124600472500172, "learning_rate": 5.800503246076902e-06, "loss": 0.6203, "step": 17914 }, { "epoch": 0.6488591090184715, "grad_norm": 1.4009430592401309, "learning_rate": 5.799438660425374e-06, "loss": 0.8041, "step": 17915 }, { "epoch": 0.64889532777979, "grad_norm": 1.0236749893334487, "learning_rate": 5.798374132577264e-06, "loss": 0.6659, "step": 17916 }, { "epoch": 0.6489315465411083, "grad_norm": 1.0902019042498066, "learning_rate": 5.797309662547218e-06, "loss": 0.6939, "step": 17917 }, { "epoch": 0.6489677653024266, "grad_norm": 1.2466738151951056, "learning_rate": 5.796245250349885e-06, "loss": 0.6447, "step": 17918 }, { "epoch": 0.649003984063745, "grad_norm": 1.2222690769060254, "learning_rate": 5.795180895999905e-06, "loss": 0.6275, "step": 17919 }, { "epoch": 0.6490402028250634, "grad_norm": 1.0884626884909296, "learning_rate": 5.7941165995119364e-06, "loss": 0.7169, "step": 17920 }, { "epoch": 0.6490764215863818, "grad_norm": 1.7347787599725595, "learning_rate": 5.793052360900619e-06, "loss": 0.7244, "step": 17921 }, { "epoch": 0.6491126403477001, "grad_norm": 1.045416916692185, "learning_rate": 5.791988180180593e-06, "loss": 0.6668, "step": 17922 }, { "epoch": 0.6491488591090184, "grad_norm": 0.8871483449770313, "learning_rate": 5.790924057366511e-06, "loss": 0.6895, "step": 17923 }, { "epoch": 0.6491850778703369, "grad_norm": 1.4951677918045965, "learning_rate": 5.789859992473013e-06, "loss": 0.7484, "step": 17924 }, { "epoch": 0.6492212966316552, "grad_norm": 1.4213173134153065, "learning_rate": 5.788795985514742e-06, "loss": 0.7411, "step": 17925 }, { "epoch": 0.6492575153929736, "grad_norm": 1.2752867166360216, "learning_rate": 5.787732036506332e-06, "loss": 0.7251, "step": 17926 }, { "epoch": 0.6492937341542919, "grad_norm": 1.2553172832575465, "learning_rate": 5.786668145462435e-06, "loss": 0.6695, "step": 17927 }, { "epoch": 0.6493299529156102, "grad_norm": 1.4293296097632073, "learning_rate": 5.785604312397688e-06, "loss": 0.7221, "step": 17928 }, { "epoch": 0.6493661716769287, "grad_norm": 1.3991958344189541, "learning_rate": 5.784540537326727e-06, "loss": 0.6779, "step": 17929 }, { "epoch": 0.649402390438247, "grad_norm": 1.3565340349527826, "learning_rate": 5.783476820264189e-06, "loss": 0.6643, "step": 17930 }, { "epoch": 0.6494386091995654, "grad_norm": 1.3789182492239285, "learning_rate": 5.782413161224718e-06, "loss": 0.6695, "step": 17931 }, { "epoch": 0.6494748279608837, "grad_norm": 1.0367741478245303, "learning_rate": 5.7813495602229495e-06, "loss": 0.6664, "step": 17932 }, { "epoch": 0.6495110467222021, "grad_norm": 1.4311706225536893, "learning_rate": 5.780286017273513e-06, "loss": 0.6997, "step": 17933 }, { "epoch": 0.6495472654835205, "grad_norm": 1.4349474953151236, "learning_rate": 5.779222532391054e-06, "loss": 0.7058, "step": 17934 }, { "epoch": 0.6495834842448388, "grad_norm": 1.3604269958416013, "learning_rate": 5.7781591055902e-06, "loss": 0.7638, "step": 17935 }, { "epoch": 0.6496197030061572, "grad_norm": 1.382395510383045, "learning_rate": 5.777095736885585e-06, "loss": 0.679, "step": 17936 }, { "epoch": 0.6496559217674756, "grad_norm": 0.8493524135058919, "learning_rate": 5.776032426291842e-06, "loss": 0.7191, "step": 17937 }, { "epoch": 0.6496921405287939, "grad_norm": 1.3493443027264975, "learning_rate": 5.774969173823607e-06, "loss": 0.7152, "step": 17938 }, { "epoch": 0.6497283592901123, "grad_norm": 1.7840648587028025, "learning_rate": 5.773905979495509e-06, "loss": 0.7255, "step": 17939 }, { "epoch": 0.6497645780514306, "grad_norm": 1.3949504203621044, "learning_rate": 5.772842843322177e-06, "loss": 0.7294, "step": 17940 }, { "epoch": 0.649800796812749, "grad_norm": 1.2285827685082102, "learning_rate": 5.771779765318239e-06, "loss": 0.7426, "step": 17941 }, { "epoch": 0.6498370155740674, "grad_norm": 1.4572917773066305, "learning_rate": 5.77071674549833e-06, "loss": 0.7163, "step": 17942 }, { "epoch": 0.6498732343353857, "grad_norm": 1.389874737449336, "learning_rate": 5.769653783877076e-06, "loss": 0.7227, "step": 17943 }, { "epoch": 0.6499094530967041, "grad_norm": 1.2916740867827181, "learning_rate": 5.768590880469098e-06, "loss": 0.7084, "step": 17944 }, { "epoch": 0.6499456718580224, "grad_norm": 1.2987289677680174, "learning_rate": 5.767528035289032e-06, "loss": 0.715, "step": 17945 }, { "epoch": 0.6499818906193409, "grad_norm": 1.26674655828494, "learning_rate": 5.766465248351499e-06, "loss": 0.6633, "step": 17946 }, { "epoch": 0.6500181093806592, "grad_norm": 1.293933694985257, "learning_rate": 5.765402519671124e-06, "loss": 0.7304, "step": 17947 }, { "epoch": 0.6500543281419775, "grad_norm": 0.96935099317527, "learning_rate": 5.764339849262528e-06, "loss": 0.6828, "step": 17948 }, { "epoch": 0.6500905469032959, "grad_norm": 1.322752513582011, "learning_rate": 5.76327723714034e-06, "loss": 0.6459, "step": 17949 }, { "epoch": 0.6501267656646142, "grad_norm": 1.0470779700022355, "learning_rate": 5.76221468331918e-06, "loss": 0.699, "step": 17950 }, { "epoch": 0.6501629844259327, "grad_norm": 1.0461059162165658, "learning_rate": 5.761152187813671e-06, "loss": 0.6182, "step": 17951 }, { "epoch": 0.650199203187251, "grad_norm": 1.4137161743620652, "learning_rate": 5.760089750638427e-06, "loss": 0.7324, "step": 17952 }, { "epoch": 0.6502354219485693, "grad_norm": 1.3703889896680352, "learning_rate": 5.759027371808079e-06, "loss": 0.7444, "step": 17953 }, { "epoch": 0.6502716407098877, "grad_norm": 1.50000591450757, "learning_rate": 5.7579650513372395e-06, "loss": 0.6678, "step": 17954 }, { "epoch": 0.6503078594712061, "grad_norm": 1.0275716780337256, "learning_rate": 5.756902789240525e-06, "loss": 0.6802, "step": 17955 }, { "epoch": 0.6503440782325245, "grad_norm": 1.2969728944208216, "learning_rate": 5.7558405855325594e-06, "loss": 0.6799, "step": 17956 }, { "epoch": 0.6503802969938428, "grad_norm": 1.2973611032670833, "learning_rate": 5.754778440227959e-06, "loss": 0.7673, "step": 17957 }, { "epoch": 0.6504165157551611, "grad_norm": 1.3569649565408934, "learning_rate": 5.753716353341336e-06, "loss": 0.7616, "step": 17958 }, { "epoch": 0.6504527345164796, "grad_norm": 1.2317539073461314, "learning_rate": 5.752654324887302e-06, "loss": 0.6444, "step": 17959 }, { "epoch": 0.6504889532777979, "grad_norm": 1.4734763257644459, "learning_rate": 5.7515923548804816e-06, "loss": 0.7076, "step": 17960 }, { "epoch": 0.6505251720391163, "grad_norm": 1.3744020401338173, "learning_rate": 5.7505304433354825e-06, "loss": 0.6866, "step": 17961 }, { "epoch": 0.6505613908004346, "grad_norm": 1.3563873520312348, "learning_rate": 5.749468590266916e-06, "loss": 0.7043, "step": 17962 }, { "epoch": 0.6505976095617529, "grad_norm": 1.283986653208271, "learning_rate": 5.748406795689399e-06, "loss": 0.6815, "step": 17963 }, { "epoch": 0.6506338283230714, "grad_norm": 1.3331744470649856, "learning_rate": 5.747345059617541e-06, "loss": 0.7185, "step": 17964 }, { "epoch": 0.6506700470843897, "grad_norm": 1.3050760051923591, "learning_rate": 5.746283382065951e-06, "loss": 0.6811, "step": 17965 }, { "epoch": 0.6507062658457081, "grad_norm": 1.370341207878852, "learning_rate": 5.745221763049237e-06, "loss": 0.71, "step": 17966 }, { "epoch": 0.6507424846070264, "grad_norm": 1.3469540639459965, "learning_rate": 5.7441602025820135e-06, "loss": 0.6829, "step": 17967 }, { "epoch": 0.6507787033683448, "grad_norm": 1.4792452892599484, "learning_rate": 5.743098700678884e-06, "loss": 0.7961, "step": 17968 }, { "epoch": 0.6508149221296632, "grad_norm": 1.3596648456702283, "learning_rate": 5.742037257354458e-06, "loss": 0.6662, "step": 17969 }, { "epoch": 0.6508511408909815, "grad_norm": 1.3549982293142033, "learning_rate": 5.740975872623335e-06, "loss": 0.7467, "step": 17970 }, { "epoch": 0.6508873596522999, "grad_norm": 1.3825575545042186, "learning_rate": 5.739914546500133e-06, "loss": 0.7138, "step": 17971 }, { "epoch": 0.6509235784136183, "grad_norm": 1.3662177202275865, "learning_rate": 5.7388532789994476e-06, "loss": 0.7628, "step": 17972 }, { "epoch": 0.6509597971749366, "grad_norm": 1.0778416438815508, "learning_rate": 5.737792070135883e-06, "loss": 0.6922, "step": 17973 }, { "epoch": 0.650996015936255, "grad_norm": 1.0146490353820259, "learning_rate": 5.736730919924047e-06, "loss": 0.7381, "step": 17974 }, { "epoch": 0.6510322346975733, "grad_norm": 1.3773118140358327, "learning_rate": 5.735669828378541e-06, "loss": 0.7425, "step": 17975 }, { "epoch": 0.6510684534588917, "grad_norm": 1.4559178623869655, "learning_rate": 5.734608795513965e-06, "loss": 0.7253, "step": 17976 }, { "epoch": 0.6511046722202101, "grad_norm": 0.9546567726941079, "learning_rate": 5.733547821344916e-06, "loss": 0.6564, "step": 17977 }, { "epoch": 0.6511408909815284, "grad_norm": 1.0728348514584265, "learning_rate": 5.732486905886001e-06, "loss": 0.6507, "step": 17978 }, { "epoch": 0.6511771097428468, "grad_norm": 1.4983681442451255, "learning_rate": 5.731426049151816e-06, "loss": 0.7164, "step": 17979 }, { "epoch": 0.6512133285041651, "grad_norm": 1.3070753782681548, "learning_rate": 5.73036525115696e-06, "loss": 0.7006, "step": 17980 }, { "epoch": 0.6512495472654836, "grad_norm": 1.2374492107731503, "learning_rate": 5.729304511916025e-06, "loss": 0.6509, "step": 17981 }, { "epoch": 0.6512857660268019, "grad_norm": 1.5054725949308099, "learning_rate": 5.7282438314436165e-06, "loss": 0.776, "step": 17982 }, { "epoch": 0.6513219847881202, "grad_norm": 1.0817069139286601, "learning_rate": 5.7271832097543256e-06, "loss": 0.6766, "step": 17983 }, { "epoch": 0.6513582035494386, "grad_norm": 0.9963714063620858, "learning_rate": 5.7261226468627466e-06, "loss": 0.6766, "step": 17984 }, { "epoch": 0.651394422310757, "grad_norm": 1.3429341001628696, "learning_rate": 5.725062142783477e-06, "loss": 0.713, "step": 17985 }, { "epoch": 0.6514306410720754, "grad_norm": 1.4272541720631198, "learning_rate": 5.7240016975311094e-06, "loss": 0.7192, "step": 17986 }, { "epoch": 0.6514668598333937, "grad_norm": 1.0744640213384662, "learning_rate": 5.722941311120236e-06, "loss": 0.6876, "step": 17987 }, { "epoch": 0.651503078594712, "grad_norm": 1.440147440784998, "learning_rate": 5.721880983565444e-06, "loss": 0.7046, "step": 17988 }, { "epoch": 0.6515392973560304, "grad_norm": 1.3435342318828036, "learning_rate": 5.7208207148813334e-06, "loss": 0.7021, "step": 17989 }, { "epoch": 0.6515755161173488, "grad_norm": 1.2900287250910296, "learning_rate": 5.719760505082489e-06, "loss": 0.7215, "step": 17990 }, { "epoch": 0.6516117348786672, "grad_norm": 1.5435939632739597, "learning_rate": 5.7187003541835014e-06, "loss": 0.7949, "step": 17991 }, { "epoch": 0.6516479536399855, "grad_norm": 1.4245437784523738, "learning_rate": 5.7176402621989535e-06, "loss": 0.7332, "step": 17992 }, { "epoch": 0.6516841724013038, "grad_norm": 1.237057850911749, "learning_rate": 5.716580229143445e-06, "loss": 0.7069, "step": 17993 }, { "epoch": 0.6517203911626223, "grad_norm": 1.2897867848393976, "learning_rate": 5.7155202550315546e-06, "loss": 0.6921, "step": 17994 }, { "epoch": 0.6517566099239406, "grad_norm": 1.2824442388889286, "learning_rate": 5.714460339877865e-06, "loss": 0.6496, "step": 17995 }, { "epoch": 0.651792828685259, "grad_norm": 1.3178663511639657, "learning_rate": 5.713400483696974e-06, "loss": 0.6835, "step": 17996 }, { "epoch": 0.6518290474465773, "grad_norm": 1.2925981583135122, "learning_rate": 5.712340686503456e-06, "loss": 0.8319, "step": 17997 }, { "epoch": 0.6518652662078956, "grad_norm": 1.2693197543698447, "learning_rate": 5.711280948311899e-06, "loss": 0.6715, "step": 17998 }, { "epoch": 0.6519014849692141, "grad_norm": 1.340275093773132, "learning_rate": 5.71022126913688e-06, "loss": 0.7412, "step": 17999 }, { "epoch": 0.6519377037305324, "grad_norm": 1.3635815314310125, "learning_rate": 5.709161648992989e-06, "loss": 0.7155, "step": 18000 }, { "epoch": 0.6519739224918508, "grad_norm": 0.9972177981017806, "learning_rate": 5.708102087894805e-06, "loss": 0.6752, "step": 18001 }, { "epoch": 0.6520101412531691, "grad_norm": 3.0933827076217537, "learning_rate": 5.707042585856908e-06, "loss": 0.6941, "step": 18002 }, { "epoch": 0.6520463600144875, "grad_norm": 1.5560228565015657, "learning_rate": 5.7059831428938715e-06, "loss": 0.6558, "step": 18003 }, { "epoch": 0.6520825787758059, "grad_norm": 1.4104717824919584, "learning_rate": 5.704923759020285e-06, "loss": 0.816, "step": 18004 }, { "epoch": 0.6521187975371242, "grad_norm": 1.417899416444837, "learning_rate": 5.703864434250721e-06, "loss": 0.7586, "step": 18005 }, { "epoch": 0.6521550162984426, "grad_norm": 1.3021342126124553, "learning_rate": 5.702805168599754e-06, "loss": 0.669, "step": 18006 }, { "epoch": 0.652191235059761, "grad_norm": 1.2950112972194803, "learning_rate": 5.701745962081968e-06, "loss": 0.6829, "step": 18007 }, { "epoch": 0.6522274538210793, "grad_norm": 1.2696771736827412, "learning_rate": 5.700686814711933e-06, "loss": 0.7791, "step": 18008 }, { "epoch": 0.6522636725823977, "grad_norm": 1.3005784108553808, "learning_rate": 5.6996277265042265e-06, "loss": 0.6449, "step": 18009 }, { "epoch": 0.652299891343716, "grad_norm": 1.3051814412038165, "learning_rate": 5.698568697473416e-06, "loss": 0.702, "step": 18010 }, { "epoch": 0.6523361101050345, "grad_norm": 1.3744340952608198, "learning_rate": 5.697509727634085e-06, "loss": 0.7326, "step": 18011 }, { "epoch": 0.6523723288663528, "grad_norm": 0.9876129036475785, "learning_rate": 5.696450817000801e-06, "loss": 0.6454, "step": 18012 }, { "epoch": 0.6524085476276711, "grad_norm": 1.3255153148763583, "learning_rate": 5.695391965588133e-06, "loss": 0.6469, "step": 18013 }, { "epoch": 0.6524447663889895, "grad_norm": 1.5339192060650257, "learning_rate": 5.694333173410651e-06, "loss": 0.7614, "step": 18014 }, { "epoch": 0.6524809851503078, "grad_norm": 1.3862770129186963, "learning_rate": 5.693274440482932e-06, "loss": 0.7784, "step": 18015 }, { "epoch": 0.6525172039116263, "grad_norm": 1.3193620626939282, "learning_rate": 5.692215766819542e-06, "loss": 0.7926, "step": 18016 }, { "epoch": 0.6525534226729446, "grad_norm": 1.5799316484531638, "learning_rate": 5.691157152435047e-06, "loss": 0.6744, "step": 18017 }, { "epoch": 0.6525896414342629, "grad_norm": 1.0499256949213682, "learning_rate": 5.690098597344016e-06, "loss": 0.7084, "step": 18018 }, { "epoch": 0.6526258601955813, "grad_norm": 1.2515092703985327, "learning_rate": 5.689040101561014e-06, "loss": 0.7807, "step": 18019 }, { "epoch": 0.6526620789568996, "grad_norm": 1.1038265642601532, "learning_rate": 5.687981665100606e-06, "loss": 0.6834, "step": 18020 }, { "epoch": 0.6526982977182181, "grad_norm": 1.3790343842394612, "learning_rate": 5.686923287977362e-06, "loss": 0.7487, "step": 18021 }, { "epoch": 0.6527345164795364, "grad_norm": 0.9789407452283684, "learning_rate": 5.685864970205843e-06, "loss": 0.6709, "step": 18022 }, { "epoch": 0.6527707352408547, "grad_norm": 1.231488243734145, "learning_rate": 5.684806711800614e-06, "loss": 0.6944, "step": 18023 }, { "epoch": 0.6528069540021731, "grad_norm": 1.8066715547130707, "learning_rate": 5.683748512776231e-06, "loss": 0.6485, "step": 18024 }, { "epoch": 0.6528431727634915, "grad_norm": 1.3237300895316082, "learning_rate": 5.682690373147266e-06, "loss": 0.7804, "step": 18025 }, { "epoch": 0.6528793915248099, "grad_norm": 1.032692946856825, "learning_rate": 5.681632292928275e-06, "loss": 0.6873, "step": 18026 }, { "epoch": 0.6529156102861282, "grad_norm": 1.40723961606348, "learning_rate": 5.680574272133814e-06, "loss": 0.7143, "step": 18027 }, { "epoch": 0.6529518290474465, "grad_norm": 1.4281693668694235, "learning_rate": 5.67951631077845e-06, "loss": 0.7326, "step": 18028 }, { "epoch": 0.652988047808765, "grad_norm": 1.2904249613256615, "learning_rate": 5.678458408876739e-06, "loss": 0.6604, "step": 18029 }, { "epoch": 0.6530242665700833, "grad_norm": 0.9271007298223537, "learning_rate": 5.677400566443236e-06, "loss": 0.6737, "step": 18030 }, { "epoch": 0.6530604853314017, "grad_norm": 1.3100645070150472, "learning_rate": 5.676342783492496e-06, "loss": 0.6496, "step": 18031 }, { "epoch": 0.65309670409272, "grad_norm": 1.1905158409161865, "learning_rate": 5.675285060039083e-06, "loss": 0.6915, "step": 18032 }, { "epoch": 0.6531329228540383, "grad_norm": 1.388893442411376, "learning_rate": 5.6742273960975465e-06, "loss": 0.7439, "step": 18033 }, { "epoch": 0.6531691416153568, "grad_norm": 1.1633797612944008, "learning_rate": 5.673169791682442e-06, "loss": 0.7189, "step": 18034 }, { "epoch": 0.6532053603766751, "grad_norm": 1.341555578560283, "learning_rate": 5.672112246808321e-06, "loss": 0.6512, "step": 18035 }, { "epoch": 0.6532415791379935, "grad_norm": 1.3676587879687312, "learning_rate": 5.67105476148974e-06, "loss": 0.7504, "step": 18036 }, { "epoch": 0.6532777978993118, "grad_norm": 1.2597362928893887, "learning_rate": 5.6699973357412505e-06, "loss": 0.6538, "step": 18037 }, { "epoch": 0.6533140166606302, "grad_norm": 1.0178196647489117, "learning_rate": 5.668939969577398e-06, "loss": 0.7598, "step": 18038 }, { "epoch": 0.6533502354219486, "grad_norm": 1.0631778538252807, "learning_rate": 5.66788266301274e-06, "loss": 0.675, "step": 18039 }, { "epoch": 0.6533864541832669, "grad_norm": 1.014241829691016, "learning_rate": 5.666825416061825e-06, "loss": 0.6517, "step": 18040 }, { "epoch": 0.6534226729445853, "grad_norm": 1.2760151535964739, "learning_rate": 5.665768228739199e-06, "loss": 0.6744, "step": 18041 }, { "epoch": 0.6534588917059037, "grad_norm": 1.0474139604500552, "learning_rate": 5.6647111010594055e-06, "loss": 0.6791, "step": 18042 }, { "epoch": 0.653495110467222, "grad_norm": 1.4693909106555336, "learning_rate": 5.663654033037003e-06, "loss": 0.7281, "step": 18043 }, { "epoch": 0.6535313292285404, "grad_norm": 1.2932802967366668, "learning_rate": 5.6625970246865295e-06, "loss": 0.6638, "step": 18044 }, { "epoch": 0.6535675479898587, "grad_norm": 1.343691462805352, "learning_rate": 5.661540076022533e-06, "loss": 0.6808, "step": 18045 }, { "epoch": 0.6536037667511772, "grad_norm": 1.2699028560659864, "learning_rate": 5.660483187059551e-06, "loss": 0.7626, "step": 18046 }, { "epoch": 0.6536399855124955, "grad_norm": 1.334550649385597, "learning_rate": 5.65942635781214e-06, "loss": 0.6758, "step": 18047 }, { "epoch": 0.6536762042738138, "grad_norm": 1.2126846706828436, "learning_rate": 5.658369588294835e-06, "loss": 0.6029, "step": 18048 }, { "epoch": 0.6537124230351322, "grad_norm": 1.2981609251329762, "learning_rate": 5.657312878522175e-06, "loss": 0.6234, "step": 18049 }, { "epoch": 0.6537486417964505, "grad_norm": 1.3330662864743243, "learning_rate": 5.656256228508711e-06, "loss": 0.745, "step": 18050 }, { "epoch": 0.653784860557769, "grad_norm": 1.6046664769616539, "learning_rate": 5.655199638268978e-06, "loss": 0.7371, "step": 18051 }, { "epoch": 0.6538210793190873, "grad_norm": 1.3697295208045979, "learning_rate": 5.654143107817513e-06, "loss": 0.7185, "step": 18052 }, { "epoch": 0.6538572980804056, "grad_norm": 1.3844702046381994, "learning_rate": 5.653086637168857e-06, "loss": 0.675, "step": 18053 }, { "epoch": 0.653893516841724, "grad_norm": 1.1855007996160507, "learning_rate": 5.652030226337551e-06, "loss": 0.6955, "step": 18054 }, { "epoch": 0.6539297356030424, "grad_norm": 0.9023572505445681, "learning_rate": 5.650973875338128e-06, "loss": 0.5976, "step": 18055 }, { "epoch": 0.6539659543643608, "grad_norm": 1.1852901300433893, "learning_rate": 5.649917584185128e-06, "loss": 0.8246, "step": 18056 }, { "epoch": 0.6540021731256791, "grad_norm": 1.0124836607575611, "learning_rate": 5.648861352893079e-06, "loss": 0.6868, "step": 18057 }, { "epoch": 0.6540383918869974, "grad_norm": 1.1670167803168172, "learning_rate": 5.647805181476526e-06, "loss": 0.7405, "step": 18058 }, { "epoch": 0.6540746106483158, "grad_norm": 1.254801847762658, "learning_rate": 5.646749069949997e-06, "loss": 0.6589, "step": 18059 }, { "epoch": 0.6541108294096342, "grad_norm": 0.9743293662651215, "learning_rate": 5.645693018328023e-06, "loss": 0.6907, "step": 18060 }, { "epoch": 0.6541470481709526, "grad_norm": 1.4588377156266943, "learning_rate": 5.644637026625141e-06, "loss": 0.7528, "step": 18061 }, { "epoch": 0.6541832669322709, "grad_norm": 1.3911282326673768, "learning_rate": 5.643581094855883e-06, "loss": 0.6646, "step": 18062 }, { "epoch": 0.6542194856935892, "grad_norm": 1.0755060329026973, "learning_rate": 5.642525223034776e-06, "loss": 0.7231, "step": 18063 }, { "epoch": 0.6542557044549077, "grad_norm": 1.288245773954722, "learning_rate": 5.641469411176346e-06, "loss": 0.6835, "step": 18064 }, { "epoch": 0.654291923216226, "grad_norm": 1.311679873364661, "learning_rate": 5.640413659295132e-06, "loss": 0.6814, "step": 18065 }, { "epoch": 0.6543281419775444, "grad_norm": 1.2997956895409326, "learning_rate": 5.639357967405655e-06, "loss": 0.7185, "step": 18066 }, { "epoch": 0.6543643607388627, "grad_norm": 1.40575109914556, "learning_rate": 5.638302335522441e-06, "loss": 0.7679, "step": 18067 }, { "epoch": 0.654400579500181, "grad_norm": 1.2004613794818972, "learning_rate": 5.637246763660024e-06, "loss": 0.6566, "step": 18068 }, { "epoch": 0.6544367982614995, "grad_norm": 1.4373129665977629, "learning_rate": 5.636191251832923e-06, "loss": 0.7068, "step": 18069 }, { "epoch": 0.6544730170228178, "grad_norm": 1.3157125563946317, "learning_rate": 5.635135800055667e-06, "loss": 0.7393, "step": 18070 }, { "epoch": 0.6545092357841362, "grad_norm": 1.3218490917911156, "learning_rate": 5.634080408342772e-06, "loss": 0.665, "step": 18071 }, { "epoch": 0.6545454545454545, "grad_norm": 0.9572865727011585, "learning_rate": 5.633025076708773e-06, "loss": 0.713, "step": 18072 }, { "epoch": 0.6545816733067729, "grad_norm": 1.2589318096112962, "learning_rate": 5.631969805168184e-06, "loss": 0.671, "step": 18073 }, { "epoch": 0.6546178920680913, "grad_norm": 1.366779041582776, "learning_rate": 5.630914593735529e-06, "loss": 0.7156, "step": 18074 }, { "epoch": 0.6546541108294096, "grad_norm": 0.9857396448147743, "learning_rate": 5.629859442425325e-06, "loss": 0.7036, "step": 18075 }, { "epoch": 0.654690329590728, "grad_norm": 1.2577652148575653, "learning_rate": 5.628804351252099e-06, "loss": 0.6442, "step": 18076 }, { "epoch": 0.6547265483520464, "grad_norm": 1.4040144528918004, "learning_rate": 5.627749320230365e-06, "loss": 0.7552, "step": 18077 }, { "epoch": 0.6547627671133647, "grad_norm": 1.3880618884353084, "learning_rate": 5.626694349374638e-06, "loss": 0.7603, "step": 18078 }, { "epoch": 0.6547989858746831, "grad_norm": 1.2535654032551349, "learning_rate": 5.625639438699444e-06, "loss": 0.6727, "step": 18079 }, { "epoch": 0.6548352046360014, "grad_norm": 1.1891196633808345, "learning_rate": 5.624584588219295e-06, "loss": 0.6353, "step": 18080 }, { "epoch": 0.6548714233973199, "grad_norm": 1.190328084382219, "learning_rate": 5.623529797948708e-06, "loss": 0.6055, "step": 18081 }, { "epoch": 0.6549076421586382, "grad_norm": 1.1633824722673998, "learning_rate": 5.622475067902191e-06, "loss": 0.7058, "step": 18082 }, { "epoch": 0.6549438609199565, "grad_norm": 1.3273238977650696, "learning_rate": 5.621420398094267e-06, "loss": 0.6667, "step": 18083 }, { "epoch": 0.6549800796812749, "grad_norm": 0.9829854744480757, "learning_rate": 5.620365788539446e-06, "loss": 0.7019, "step": 18084 }, { "epoch": 0.6550162984425932, "grad_norm": 1.5753404156709236, "learning_rate": 5.619311239252241e-06, "loss": 0.7246, "step": 18085 }, { "epoch": 0.6550525172039117, "grad_norm": 1.2424811010247425, "learning_rate": 5.6182567502471575e-06, "loss": 0.6974, "step": 18086 }, { "epoch": 0.65508873596523, "grad_norm": 1.3753057513800602, "learning_rate": 5.617202321538716e-06, "loss": 0.7045, "step": 18087 }, { "epoch": 0.6551249547265483, "grad_norm": 0.9751171927078074, "learning_rate": 5.616147953141422e-06, "loss": 0.7113, "step": 18088 }, { "epoch": 0.6551611734878667, "grad_norm": 1.0224658905656692, "learning_rate": 5.615093645069779e-06, "loss": 0.6854, "step": 18089 }, { "epoch": 0.655197392249185, "grad_norm": 1.3032526660949912, "learning_rate": 5.6140393973383046e-06, "loss": 0.7186, "step": 18090 }, { "epoch": 0.6552336110105035, "grad_norm": 1.258707599319676, "learning_rate": 5.612985209961502e-06, "loss": 0.7701, "step": 18091 }, { "epoch": 0.6552698297718218, "grad_norm": 1.1400652343816013, "learning_rate": 5.611931082953879e-06, "loss": 0.7323, "step": 18092 }, { "epoch": 0.6553060485331401, "grad_norm": 1.1145266351546799, "learning_rate": 5.610877016329934e-06, "loss": 0.6819, "step": 18093 }, { "epoch": 0.6553422672944585, "grad_norm": 1.354227632898599, "learning_rate": 5.609823010104183e-06, "loss": 0.7722, "step": 18094 }, { "epoch": 0.6553784860557769, "grad_norm": 1.4015008716419357, "learning_rate": 5.608769064291124e-06, "loss": 0.749, "step": 18095 }, { "epoch": 0.6554147048170953, "grad_norm": 1.4230875488508354, "learning_rate": 5.607715178905262e-06, "loss": 0.6507, "step": 18096 }, { "epoch": 0.6554509235784136, "grad_norm": 1.3598260939494078, "learning_rate": 5.606661353961095e-06, "loss": 0.7234, "step": 18097 }, { "epoch": 0.6554871423397319, "grad_norm": 1.2779940399333538, "learning_rate": 5.605607589473131e-06, "loss": 0.651, "step": 18098 }, { "epoch": 0.6555233611010504, "grad_norm": 1.255313242852008, "learning_rate": 5.604553885455868e-06, "loss": 0.7325, "step": 18099 }, { "epoch": 0.6555595798623687, "grad_norm": 1.155361455192872, "learning_rate": 5.603500241923803e-06, "loss": 0.7185, "step": 18100 }, { "epoch": 0.6555957986236871, "grad_norm": 1.3411697486640153, "learning_rate": 5.602446658891441e-06, "loss": 0.7099, "step": 18101 }, { "epoch": 0.6556320173850054, "grad_norm": 1.2787108593357026, "learning_rate": 5.601393136373276e-06, "loss": 0.7, "step": 18102 }, { "epoch": 0.6556682361463237, "grad_norm": 0.9673529971052602, "learning_rate": 5.600339674383808e-06, "loss": 0.606, "step": 18103 }, { "epoch": 0.6557044549076422, "grad_norm": 1.3714355615975828, "learning_rate": 5.599286272937527e-06, "loss": 0.7724, "step": 18104 }, { "epoch": 0.6557406736689605, "grad_norm": 1.107387901671013, "learning_rate": 5.598232932048937e-06, "loss": 0.7134, "step": 18105 }, { "epoch": 0.6557768924302789, "grad_norm": 1.0930795592337446, "learning_rate": 5.5971796517325306e-06, "loss": 0.7619, "step": 18106 }, { "epoch": 0.6558131111915972, "grad_norm": 1.3387879912355674, "learning_rate": 5.5961264320028e-06, "loss": 0.6726, "step": 18107 }, { "epoch": 0.6558493299529156, "grad_norm": 1.2396350414964123, "learning_rate": 5.595073272874236e-06, "loss": 0.7379, "step": 18108 }, { "epoch": 0.655885548714234, "grad_norm": 1.4862217984158737, "learning_rate": 5.594020174361339e-06, "loss": 0.7561, "step": 18109 }, { "epoch": 0.6559217674755523, "grad_norm": 1.044452800536232, "learning_rate": 5.5929671364785955e-06, "loss": 0.6921, "step": 18110 }, { "epoch": 0.6559579862368707, "grad_norm": 1.3310642004197888, "learning_rate": 5.591914159240491e-06, "loss": 0.69, "step": 18111 }, { "epoch": 0.6559942049981891, "grad_norm": 1.3256485577865944, "learning_rate": 5.590861242661526e-06, "loss": 0.6811, "step": 18112 }, { "epoch": 0.6560304237595074, "grad_norm": 1.1393836019339592, "learning_rate": 5.589808386756184e-06, "loss": 0.6279, "step": 18113 }, { "epoch": 0.6560666425208258, "grad_norm": 1.286768637346124, "learning_rate": 5.5887555915389545e-06, "loss": 0.7162, "step": 18114 }, { "epoch": 0.6561028612821441, "grad_norm": 1.4378688574845215, "learning_rate": 5.58770285702432e-06, "loss": 0.7202, "step": 18115 }, { "epoch": 0.6561390800434626, "grad_norm": 1.3887515253115295, "learning_rate": 5.586650183226776e-06, "loss": 0.688, "step": 18116 }, { "epoch": 0.6561752988047809, "grad_norm": 1.5325498720011048, "learning_rate": 5.585597570160802e-06, "loss": 0.7558, "step": 18117 }, { "epoch": 0.6562115175660992, "grad_norm": 1.2876432553239585, "learning_rate": 5.584545017840886e-06, "loss": 0.7744, "step": 18118 }, { "epoch": 0.6562477363274176, "grad_norm": 1.1267221415940745, "learning_rate": 5.583492526281504e-06, "loss": 0.741, "step": 18119 }, { "epoch": 0.6562839550887359, "grad_norm": 1.245584208846489, "learning_rate": 5.58244009549715e-06, "loss": 0.601, "step": 18120 }, { "epoch": 0.6563201738500544, "grad_norm": 1.3208744825294638, "learning_rate": 5.581387725502303e-06, "loss": 0.6865, "step": 18121 }, { "epoch": 0.6563563926113727, "grad_norm": 1.1805009812114817, "learning_rate": 5.580335416311439e-06, "loss": 0.6444, "step": 18122 }, { "epoch": 0.656392611372691, "grad_norm": 1.1313991161916075, "learning_rate": 5.579283167939047e-06, "loss": 0.6822, "step": 18123 }, { "epoch": 0.6564288301340094, "grad_norm": 1.5611782927124216, "learning_rate": 5.5782309803996026e-06, "loss": 0.7895, "step": 18124 }, { "epoch": 0.6564650488953278, "grad_norm": 1.1865363626600276, "learning_rate": 5.577178853707586e-06, "loss": 0.6385, "step": 18125 }, { "epoch": 0.6565012676566462, "grad_norm": 1.472962527210853, "learning_rate": 5.576126787877469e-06, "loss": 0.7081, "step": 18126 }, { "epoch": 0.6565374864179645, "grad_norm": 0.959002760792961, "learning_rate": 5.575074782923741e-06, "loss": 0.6737, "step": 18127 }, { "epoch": 0.6565737051792828, "grad_norm": 1.467768892470256, "learning_rate": 5.5740228388608705e-06, "loss": 0.6676, "step": 18128 }, { "epoch": 0.6566099239406012, "grad_norm": 1.3699214976364327, "learning_rate": 5.572970955703332e-06, "loss": 0.69, "step": 18129 }, { "epoch": 0.6566461427019196, "grad_norm": 1.3266707607009731, "learning_rate": 5.571919133465605e-06, "loss": 0.78, "step": 18130 }, { "epoch": 0.656682361463238, "grad_norm": 1.3637789907922955, "learning_rate": 5.570867372162163e-06, "loss": 0.7444, "step": 18131 }, { "epoch": 0.6567185802245563, "grad_norm": 1.2635595663830717, "learning_rate": 5.569815671807477e-06, "loss": 0.7297, "step": 18132 }, { "epoch": 0.6567547989858746, "grad_norm": 0.9801240223929277, "learning_rate": 5.568764032416018e-06, "loss": 0.6996, "step": 18133 }, { "epoch": 0.6567910177471931, "grad_norm": 1.3243074528065395, "learning_rate": 5.567712454002261e-06, "loss": 0.7236, "step": 18134 }, { "epoch": 0.6568272365085114, "grad_norm": 1.430683489983315, "learning_rate": 5.566660936580677e-06, "loss": 0.7582, "step": 18135 }, { "epoch": 0.6568634552698298, "grad_norm": 1.4615232369518232, "learning_rate": 5.565609480165732e-06, "loss": 0.7224, "step": 18136 }, { "epoch": 0.6568996740311481, "grad_norm": 1.3426671613771266, "learning_rate": 5.564558084771895e-06, "loss": 0.7132, "step": 18137 }, { "epoch": 0.6569358927924664, "grad_norm": 1.3733524847896699, "learning_rate": 5.563506750413639e-06, "loss": 0.7181, "step": 18138 }, { "epoch": 0.6569721115537849, "grad_norm": 0.9845562255207181, "learning_rate": 5.562455477105428e-06, "loss": 0.6911, "step": 18139 }, { "epoch": 0.6570083303151032, "grad_norm": 1.0548837729675506, "learning_rate": 5.5614042648617265e-06, "loss": 0.7121, "step": 18140 }, { "epoch": 0.6570445490764216, "grad_norm": 1.2735581933204874, "learning_rate": 5.5603531136970054e-06, "loss": 0.6853, "step": 18141 }, { "epoch": 0.6570807678377399, "grad_norm": 1.3197375631445527, "learning_rate": 5.559302023625726e-06, "loss": 0.6607, "step": 18142 }, { "epoch": 0.6571169865990583, "grad_norm": 1.041778049021321, "learning_rate": 5.558250994662353e-06, "loss": 0.6917, "step": 18143 }, { "epoch": 0.6571532053603767, "grad_norm": 1.3125828783514455, "learning_rate": 5.557200026821346e-06, "loss": 0.6534, "step": 18144 }, { "epoch": 0.657189424121695, "grad_norm": 1.3565648220576667, "learning_rate": 5.556149120117174e-06, "loss": 0.6541, "step": 18145 }, { "epoch": 0.6572256428830134, "grad_norm": 1.2440277734114877, "learning_rate": 5.555098274564295e-06, "loss": 0.7503, "step": 18146 }, { "epoch": 0.6572618616443318, "grad_norm": 1.4231839279441825, "learning_rate": 5.554047490177168e-06, "loss": 0.7164, "step": 18147 }, { "epoch": 0.6572980804056501, "grad_norm": 1.3370218971906518, "learning_rate": 5.552996766970251e-06, "loss": 0.7988, "step": 18148 }, { "epoch": 0.6573342991669685, "grad_norm": 1.3719720357268572, "learning_rate": 5.55194610495801e-06, "loss": 0.6748, "step": 18149 }, { "epoch": 0.6573705179282868, "grad_norm": 1.296984884765827, "learning_rate": 5.550895504154898e-06, "loss": 0.6757, "step": 18150 }, { "epoch": 0.6574067366896053, "grad_norm": 1.2931900245642691, "learning_rate": 5.549844964575371e-06, "loss": 0.7041, "step": 18151 }, { "epoch": 0.6574429554509236, "grad_norm": 1.3022450488915558, "learning_rate": 5.54879448623389e-06, "loss": 0.7183, "step": 18152 }, { "epoch": 0.6574791742122419, "grad_norm": 0.9462955933073374, "learning_rate": 5.5477440691449095e-06, "loss": 0.6635, "step": 18153 }, { "epoch": 0.6575153929735603, "grad_norm": 1.387995073450419, "learning_rate": 5.5466937133228815e-06, "loss": 0.7982, "step": 18154 }, { "epoch": 0.6575516117348786, "grad_norm": 1.2306862578322557, "learning_rate": 5.5456434187822574e-06, "loss": 0.6366, "step": 18155 }, { "epoch": 0.6575878304961971, "grad_norm": 1.2781020590884853, "learning_rate": 5.544593185537498e-06, "loss": 0.6877, "step": 18156 }, { "epoch": 0.6576240492575154, "grad_norm": 1.3207286497716764, "learning_rate": 5.543543013603051e-06, "loss": 0.7185, "step": 18157 }, { "epoch": 0.6576602680188337, "grad_norm": 1.3366722720051492, "learning_rate": 5.542492902993369e-06, "loss": 0.7282, "step": 18158 }, { "epoch": 0.6576964867801521, "grad_norm": 1.2584506535120936, "learning_rate": 5.5414428537228955e-06, "loss": 0.6485, "step": 18159 }, { "epoch": 0.6577327055414705, "grad_norm": 2.2111225474075273, "learning_rate": 5.54039286580609e-06, "loss": 0.728, "step": 18160 }, { "epoch": 0.6577689243027889, "grad_norm": 1.3418420148250283, "learning_rate": 5.539342939257399e-06, "loss": 0.6457, "step": 18161 }, { "epoch": 0.6578051430641072, "grad_norm": 1.2253459505480868, "learning_rate": 5.538293074091264e-06, "loss": 0.5606, "step": 18162 }, { "epoch": 0.6578413618254255, "grad_norm": 1.3127757702351925, "learning_rate": 5.537243270322139e-06, "loss": 0.6633, "step": 18163 }, { "epoch": 0.657877580586744, "grad_norm": 1.3056792350706334, "learning_rate": 5.536193527964471e-06, "loss": 0.6538, "step": 18164 }, { "epoch": 0.6579137993480623, "grad_norm": 1.344723328175807, "learning_rate": 5.5351438470327e-06, "loss": 0.6691, "step": 18165 }, { "epoch": 0.6579500181093807, "grad_norm": 1.116257444215522, "learning_rate": 5.534094227541269e-06, "loss": 0.74, "step": 18166 }, { "epoch": 0.657986236870699, "grad_norm": 1.3779509520427544, "learning_rate": 5.53304466950463e-06, "loss": 0.7062, "step": 18167 }, { "epoch": 0.6580224556320173, "grad_norm": 1.4066944188301087, "learning_rate": 5.531995172937221e-06, "loss": 0.7089, "step": 18168 }, { "epoch": 0.6580586743933358, "grad_norm": 1.4071991340034293, "learning_rate": 5.530945737853485e-06, "loss": 0.6546, "step": 18169 }, { "epoch": 0.6580948931546541, "grad_norm": 1.2627055993070473, "learning_rate": 5.529896364267857e-06, "loss": 0.6996, "step": 18170 }, { "epoch": 0.6581311119159725, "grad_norm": 1.0416912552730413, "learning_rate": 5.528847052194789e-06, "loss": 0.8219, "step": 18171 }, { "epoch": 0.6581673306772908, "grad_norm": 1.3166331009425452, "learning_rate": 5.5277978016487135e-06, "loss": 0.6723, "step": 18172 }, { "epoch": 0.6582035494386091, "grad_norm": 1.361011815916205, "learning_rate": 5.526748612644065e-06, "loss": 0.7053, "step": 18173 }, { "epoch": 0.6582397681999276, "grad_norm": 1.4176326319184926, "learning_rate": 5.52569948519529e-06, "loss": 0.6815, "step": 18174 }, { "epoch": 0.6582759869612459, "grad_norm": 1.4417807730842978, "learning_rate": 5.524650419316824e-06, "loss": 0.7983, "step": 18175 }, { "epoch": 0.6583122057225643, "grad_norm": 1.4252139478303922, "learning_rate": 5.523601415023099e-06, "loss": 0.7375, "step": 18176 }, { "epoch": 0.6583484244838826, "grad_norm": 1.3802022753138716, "learning_rate": 5.522552472328546e-06, "loss": 0.7373, "step": 18177 }, { "epoch": 0.658384643245201, "grad_norm": 1.4205798221202348, "learning_rate": 5.521503591247612e-06, "loss": 0.6162, "step": 18178 }, { "epoch": 0.6584208620065194, "grad_norm": 1.119489524862617, "learning_rate": 5.520454771794722e-06, "loss": 0.7232, "step": 18179 }, { "epoch": 0.6584570807678377, "grad_norm": 1.0129812032711023, "learning_rate": 5.519406013984312e-06, "loss": 0.6901, "step": 18180 }, { "epoch": 0.6584932995291561, "grad_norm": 1.4146143273856622, "learning_rate": 5.518357317830807e-06, "loss": 0.6856, "step": 18181 }, { "epoch": 0.6585295182904745, "grad_norm": 1.1662179263817023, "learning_rate": 5.517308683348648e-06, "loss": 0.7723, "step": 18182 }, { "epoch": 0.6585657370517928, "grad_norm": 1.4159798216404065, "learning_rate": 5.516260110552258e-06, "loss": 0.7956, "step": 18183 }, { "epoch": 0.6586019558131112, "grad_norm": 1.3101360960680124, "learning_rate": 5.5152115994560675e-06, "loss": 0.7982, "step": 18184 }, { "epoch": 0.6586381745744295, "grad_norm": 1.2757539034608898, "learning_rate": 5.514163150074509e-06, "loss": 0.6916, "step": 18185 }, { "epoch": 0.658674393335748, "grad_norm": 1.118007819533019, "learning_rate": 5.513114762422007e-06, "loss": 0.6857, "step": 18186 }, { "epoch": 0.6587106120970663, "grad_norm": 1.2525835934361802, "learning_rate": 5.512066436512989e-06, "loss": 0.6554, "step": 18187 }, { "epoch": 0.6587468308583846, "grad_norm": 1.3743966523651086, "learning_rate": 5.511018172361876e-06, "loss": 0.7576, "step": 18188 }, { "epoch": 0.658783049619703, "grad_norm": 1.364127415027847, "learning_rate": 5.509969969983102e-06, "loss": 0.7137, "step": 18189 }, { "epoch": 0.6588192683810213, "grad_norm": 1.4353664674087554, "learning_rate": 5.5089218293910855e-06, "loss": 0.7047, "step": 18190 }, { "epoch": 0.6588554871423398, "grad_norm": 1.2704687383743452, "learning_rate": 5.507873750600247e-06, "loss": 0.6932, "step": 18191 }, { "epoch": 0.6588917059036581, "grad_norm": 1.3100828580141262, "learning_rate": 5.506825733625017e-06, "loss": 0.7023, "step": 18192 }, { "epoch": 0.6589279246649764, "grad_norm": 1.177380271264015, "learning_rate": 5.505777778479812e-06, "loss": 0.7969, "step": 18193 }, { "epoch": 0.6589641434262948, "grad_norm": 1.064919078077734, "learning_rate": 5.504729885179054e-06, "loss": 0.6737, "step": 18194 }, { "epoch": 0.6590003621876132, "grad_norm": 1.3844316399915968, "learning_rate": 5.5036820537371604e-06, "loss": 0.6427, "step": 18195 }, { "epoch": 0.6590365809489316, "grad_norm": 1.4455993689838906, "learning_rate": 5.502634284168555e-06, "loss": 0.7654, "step": 18196 }, { "epoch": 0.6590727997102499, "grad_norm": 1.2144866637810137, "learning_rate": 5.501586576487654e-06, "loss": 0.6879, "step": 18197 }, { "epoch": 0.6591090184715682, "grad_norm": 1.0399476960395246, "learning_rate": 5.500538930708874e-06, "loss": 0.6811, "step": 18198 }, { "epoch": 0.6591452372328866, "grad_norm": 1.299061031540313, "learning_rate": 5.499491346846628e-06, "loss": 0.6653, "step": 18199 }, { "epoch": 0.659181455994205, "grad_norm": 1.4737856275678243, "learning_rate": 5.4984438249153405e-06, "loss": 0.8059, "step": 18200 }, { "epoch": 0.6592176747555234, "grad_norm": 1.4167900756302567, "learning_rate": 5.497396364929421e-06, "loss": 0.7934, "step": 18201 }, { "epoch": 0.6592538935168417, "grad_norm": 1.404100644457171, "learning_rate": 5.4963489669032795e-06, "loss": 0.7014, "step": 18202 }, { "epoch": 0.65929011227816, "grad_norm": 1.046166899096811, "learning_rate": 5.4953016308513375e-06, "loss": 0.6384, "step": 18203 }, { "epoch": 0.6593263310394785, "grad_norm": 1.3325658470762765, "learning_rate": 5.4942543567880025e-06, "loss": 0.7462, "step": 18204 }, { "epoch": 0.6593625498007968, "grad_norm": 1.4587600240602019, "learning_rate": 5.493207144727688e-06, "loss": 0.7762, "step": 18205 }, { "epoch": 0.6593987685621152, "grad_norm": 1.3241239868285506, "learning_rate": 5.492159994684798e-06, "loss": 0.6874, "step": 18206 }, { "epoch": 0.6594349873234335, "grad_norm": 1.3697551403929158, "learning_rate": 5.491112906673753e-06, "loss": 0.6961, "step": 18207 }, { "epoch": 0.6594712060847518, "grad_norm": 1.398522157344284, "learning_rate": 5.490065880708955e-06, "loss": 0.7104, "step": 18208 }, { "epoch": 0.6595074248460703, "grad_norm": 1.2633648179462884, "learning_rate": 5.489018916804813e-06, "loss": 0.6768, "step": 18209 }, { "epoch": 0.6595436436073886, "grad_norm": 1.3301746694673457, "learning_rate": 5.487972014975731e-06, "loss": 0.679, "step": 18210 }, { "epoch": 0.659579862368707, "grad_norm": 1.674067904057912, "learning_rate": 5.486925175236123e-06, "loss": 0.7057, "step": 18211 }, { "epoch": 0.6596160811300253, "grad_norm": 1.3701040079702387, "learning_rate": 5.48587839760039e-06, "loss": 0.6461, "step": 18212 }, { "epoch": 0.6596522998913437, "grad_norm": 1.388649972509535, "learning_rate": 5.484831682082932e-06, "loss": 0.6579, "step": 18213 }, { "epoch": 0.6596885186526621, "grad_norm": 1.0298308852521108, "learning_rate": 5.483785028698162e-06, "loss": 0.7165, "step": 18214 }, { "epoch": 0.6597247374139804, "grad_norm": 1.00769419761822, "learning_rate": 5.482738437460477e-06, "loss": 0.7193, "step": 18215 }, { "epoch": 0.6597609561752988, "grad_norm": 1.457654013891158, "learning_rate": 5.481691908384282e-06, "loss": 0.7609, "step": 18216 }, { "epoch": 0.6597971749366172, "grad_norm": 1.320756202128777, "learning_rate": 5.480645441483973e-06, "loss": 0.632, "step": 18217 }, { "epoch": 0.6598333936979355, "grad_norm": 0.9177452241051708, "learning_rate": 5.479599036773957e-06, "loss": 0.7759, "step": 18218 }, { "epoch": 0.6598696124592539, "grad_norm": 1.3839553700766174, "learning_rate": 5.478552694268628e-06, "loss": 0.6984, "step": 18219 }, { "epoch": 0.6599058312205722, "grad_norm": 1.3506443289645096, "learning_rate": 5.47750641398239e-06, "loss": 0.696, "step": 18220 }, { "epoch": 0.6599420499818907, "grad_norm": 1.4023805755572318, "learning_rate": 5.476460195929632e-06, "loss": 0.6806, "step": 18221 }, { "epoch": 0.659978268743209, "grad_norm": 1.3637194516731663, "learning_rate": 5.475414040124761e-06, "loss": 0.6887, "step": 18222 }, { "epoch": 0.6600144875045273, "grad_norm": 1.4196271459302598, "learning_rate": 5.474367946582168e-06, "loss": 0.7062, "step": 18223 }, { "epoch": 0.6600507062658457, "grad_norm": 1.095511258061738, "learning_rate": 5.473321915316245e-06, "loss": 0.6875, "step": 18224 }, { "epoch": 0.660086925027164, "grad_norm": 1.1640184192303833, "learning_rate": 5.472275946341394e-06, "loss": 0.717, "step": 18225 }, { "epoch": 0.6601231437884825, "grad_norm": 1.3044038824083375, "learning_rate": 5.471230039672004e-06, "loss": 0.7211, "step": 18226 }, { "epoch": 0.6601593625498008, "grad_norm": 1.5421437550778532, "learning_rate": 5.470184195322468e-06, "loss": 0.7293, "step": 18227 }, { "epoch": 0.6601955813111191, "grad_norm": 1.3902279730411193, "learning_rate": 5.4691384133071775e-06, "loss": 0.7048, "step": 18228 }, { "epoch": 0.6602318000724375, "grad_norm": 1.3108442327333125, "learning_rate": 5.4680926936405245e-06, "loss": 0.6503, "step": 18229 }, { "epoch": 0.6602680188337559, "grad_norm": 1.009694987105372, "learning_rate": 5.4670470363368975e-06, "loss": 0.7274, "step": 18230 }, { "epoch": 0.6603042375950743, "grad_norm": 1.4112463209177755, "learning_rate": 5.4660014414106825e-06, "loss": 0.6414, "step": 18231 }, { "epoch": 0.6603404563563926, "grad_norm": 1.231155771374933, "learning_rate": 5.464955908876275e-06, "loss": 0.6658, "step": 18232 }, { "epoch": 0.6603766751177109, "grad_norm": 1.2559004483247354, "learning_rate": 5.463910438748059e-06, "loss": 0.6391, "step": 18233 }, { "epoch": 0.6604128938790294, "grad_norm": 1.4184064353809203, "learning_rate": 5.4628650310404184e-06, "loss": 0.7893, "step": 18234 }, { "epoch": 0.6604491126403477, "grad_norm": 1.3025351812442412, "learning_rate": 5.461819685767744e-06, "loss": 0.7182, "step": 18235 }, { "epoch": 0.6604853314016661, "grad_norm": 1.3310388527737838, "learning_rate": 5.46077440294442e-06, "loss": 0.731, "step": 18236 }, { "epoch": 0.6605215501629844, "grad_norm": 1.3601908441643662, "learning_rate": 5.459729182584828e-06, "loss": 0.687, "step": 18237 }, { "epoch": 0.6605577689243027, "grad_norm": 0.9562058479065403, "learning_rate": 5.458684024703348e-06, "loss": 0.6568, "step": 18238 }, { "epoch": 0.6605939876856212, "grad_norm": 1.5224498726355227, "learning_rate": 5.45763892931437e-06, "loss": 0.6571, "step": 18239 }, { "epoch": 0.6606302064469395, "grad_norm": 1.2668324247624823, "learning_rate": 5.456593896432272e-06, "loss": 0.6138, "step": 18240 }, { "epoch": 0.6606664252082579, "grad_norm": 1.1921934684137496, "learning_rate": 5.455548926071432e-06, "loss": 0.7117, "step": 18241 }, { "epoch": 0.6607026439695762, "grad_norm": 1.3741076673380808, "learning_rate": 5.454504018246231e-06, "loss": 0.6941, "step": 18242 }, { "epoch": 0.6607388627308945, "grad_norm": 1.0137230010016027, "learning_rate": 5.4534591729710505e-06, "loss": 0.7078, "step": 18243 }, { "epoch": 0.660775081492213, "grad_norm": 0.9787616086541892, "learning_rate": 5.452414390260267e-06, "loss": 0.6545, "step": 18244 }, { "epoch": 0.6608113002535313, "grad_norm": 1.4350484002041075, "learning_rate": 5.451369670128255e-06, "loss": 0.7165, "step": 18245 }, { "epoch": 0.6608475190148497, "grad_norm": 1.1793314953374119, "learning_rate": 5.450325012589396e-06, "loss": 0.623, "step": 18246 }, { "epoch": 0.660883737776168, "grad_norm": 1.2757407075977387, "learning_rate": 5.4492804176580625e-06, "loss": 0.6229, "step": 18247 }, { "epoch": 0.6609199565374864, "grad_norm": 1.294874334253773, "learning_rate": 5.44823588534863e-06, "loss": 0.6977, "step": 18248 }, { "epoch": 0.6609561752988048, "grad_norm": 1.3116305915335835, "learning_rate": 5.447191415675466e-06, "loss": 0.7025, "step": 18249 }, { "epoch": 0.6609923940601231, "grad_norm": 1.36170811870581, "learning_rate": 5.446147008652952e-06, "loss": 0.6621, "step": 18250 }, { "epoch": 0.6610286128214415, "grad_norm": 1.3965195151222713, "learning_rate": 5.445102664295457e-06, "loss": 0.6249, "step": 18251 }, { "epoch": 0.6610648315827599, "grad_norm": 1.268050589185564, "learning_rate": 5.4440583826173525e-06, "loss": 0.6696, "step": 18252 }, { "epoch": 0.6611010503440782, "grad_norm": 1.3435977715107001, "learning_rate": 5.443014163633004e-06, "loss": 0.7144, "step": 18253 }, { "epoch": 0.6611372691053966, "grad_norm": 1.3132496010202221, "learning_rate": 5.441970007356787e-06, "loss": 0.7399, "step": 18254 }, { "epoch": 0.6611734878667149, "grad_norm": 1.1233993229639034, "learning_rate": 5.44092591380307e-06, "loss": 0.7491, "step": 18255 }, { "epoch": 0.6612097066280334, "grad_norm": 1.3355928484567126, "learning_rate": 5.4398818829862135e-06, "loss": 0.6877, "step": 18256 }, { "epoch": 0.6612459253893517, "grad_norm": 1.2609867358137803, "learning_rate": 5.4388379149205914e-06, "loss": 0.6988, "step": 18257 }, { "epoch": 0.66128214415067, "grad_norm": 1.4003593209889795, "learning_rate": 5.43779400962057e-06, "loss": 0.7287, "step": 18258 }, { "epoch": 0.6613183629119884, "grad_norm": 1.4161680172440325, "learning_rate": 5.436750167100511e-06, "loss": 0.7599, "step": 18259 }, { "epoch": 0.6613545816733067, "grad_norm": 1.3930001575696833, "learning_rate": 5.435706387374774e-06, "loss": 0.6723, "step": 18260 }, { "epoch": 0.6613908004346252, "grad_norm": 1.1345295007930196, "learning_rate": 5.434662670457733e-06, "loss": 0.6897, "step": 18261 }, { "epoch": 0.6614270191959435, "grad_norm": 1.2869421455212755, "learning_rate": 5.433619016363745e-06, "loss": 0.6759, "step": 18262 }, { "epoch": 0.6614632379572618, "grad_norm": 1.430610285975365, "learning_rate": 5.432575425107171e-06, "loss": 0.6981, "step": 18263 }, { "epoch": 0.6614994567185802, "grad_norm": 1.3010611870111688, "learning_rate": 5.431531896702369e-06, "loss": 0.7285, "step": 18264 }, { "epoch": 0.6615356754798986, "grad_norm": 1.3641292863731467, "learning_rate": 5.430488431163705e-06, "loss": 0.7599, "step": 18265 }, { "epoch": 0.661571894241217, "grad_norm": 1.3354623664892185, "learning_rate": 5.4294450285055376e-06, "loss": 0.7416, "step": 18266 }, { "epoch": 0.6616081130025353, "grad_norm": 1.2870639346557289, "learning_rate": 5.4284016887422175e-06, "loss": 0.6131, "step": 18267 }, { "epoch": 0.6616443317638536, "grad_norm": 1.3366697922797988, "learning_rate": 5.427358411888111e-06, "loss": 0.6355, "step": 18268 }, { "epoch": 0.661680550525172, "grad_norm": 1.2158600544083498, "learning_rate": 5.42631519795757e-06, "loss": 0.7244, "step": 18269 }, { "epoch": 0.6617167692864904, "grad_norm": 1.3806591873525376, "learning_rate": 5.425272046964953e-06, "loss": 0.6742, "step": 18270 }, { "epoch": 0.6617529880478088, "grad_norm": 1.4378107735313193, "learning_rate": 5.424228958924607e-06, "loss": 0.6698, "step": 18271 }, { "epoch": 0.6617892068091271, "grad_norm": 1.2210542917588068, "learning_rate": 5.423185933850894e-06, "loss": 0.7154, "step": 18272 }, { "epoch": 0.6618254255704454, "grad_norm": 1.371599079812422, "learning_rate": 5.422142971758164e-06, "loss": 0.6918, "step": 18273 }, { "epoch": 0.6618616443317639, "grad_norm": 1.3021916734604755, "learning_rate": 5.421100072660771e-06, "loss": 0.7222, "step": 18274 }, { "epoch": 0.6618978630930822, "grad_norm": 1.0027834572148346, "learning_rate": 5.42005723657306e-06, "loss": 0.6802, "step": 18275 }, { "epoch": 0.6619340818544006, "grad_norm": 1.318911657153948, "learning_rate": 5.419014463509389e-06, "loss": 0.6068, "step": 18276 }, { "epoch": 0.6619703006157189, "grad_norm": 1.1064961882874005, "learning_rate": 5.417971753484104e-06, "loss": 0.7072, "step": 18277 }, { "epoch": 0.6620065193770373, "grad_norm": 1.3043737419695378, "learning_rate": 5.41692910651155e-06, "loss": 0.6914, "step": 18278 }, { "epoch": 0.6620427381383557, "grad_norm": 1.2526647669738287, "learning_rate": 5.415886522606084e-06, "loss": 0.7375, "step": 18279 }, { "epoch": 0.662078956899674, "grad_norm": 0.8923987309022755, "learning_rate": 5.414844001782046e-06, "loss": 0.653, "step": 18280 }, { "epoch": 0.6621151756609924, "grad_norm": 1.088267952364265, "learning_rate": 5.413801544053783e-06, "loss": 0.7378, "step": 18281 }, { "epoch": 0.6621513944223107, "grad_norm": 1.401775060172591, "learning_rate": 5.412759149435638e-06, "loss": 0.7205, "step": 18282 }, { "epoch": 0.6621876131836291, "grad_norm": 1.4273863627895755, "learning_rate": 5.411716817941961e-06, "loss": 0.7659, "step": 18283 }, { "epoch": 0.6622238319449475, "grad_norm": 1.27283789151916, "learning_rate": 5.410674549587092e-06, "loss": 0.6927, "step": 18284 }, { "epoch": 0.6622600507062658, "grad_norm": 1.2500615138325888, "learning_rate": 5.409632344385369e-06, "loss": 0.6878, "step": 18285 }, { "epoch": 0.6622962694675842, "grad_norm": 1.4945580385611976, "learning_rate": 5.4085902023511434e-06, "loss": 0.7266, "step": 18286 }, { "epoch": 0.6623324882289026, "grad_norm": 1.3189868984137267, "learning_rate": 5.407548123498751e-06, "loss": 0.6874, "step": 18287 }, { "epoch": 0.6623687069902209, "grad_norm": 1.3085991062225495, "learning_rate": 5.406506107842532e-06, "loss": 0.7003, "step": 18288 }, { "epoch": 0.6624049257515393, "grad_norm": 1.4079632931837334, "learning_rate": 5.40546415539682e-06, "loss": 0.7371, "step": 18289 }, { "epoch": 0.6624411445128576, "grad_norm": 0.9808982442963352, "learning_rate": 5.404422266175964e-06, "loss": 0.6601, "step": 18290 }, { "epoch": 0.6624773632741761, "grad_norm": 1.3314380018605165, "learning_rate": 5.403380440194295e-06, "loss": 0.6566, "step": 18291 }, { "epoch": 0.6625135820354944, "grad_norm": 0.9551324449223045, "learning_rate": 5.402338677466149e-06, "loss": 0.7032, "step": 18292 }, { "epoch": 0.6625498007968128, "grad_norm": 1.3997594988685103, "learning_rate": 5.401296978005859e-06, "loss": 0.6998, "step": 18293 }, { "epoch": 0.6625860195581311, "grad_norm": 1.1095761048959456, "learning_rate": 5.400255341827768e-06, "loss": 0.7483, "step": 18294 }, { "epoch": 0.6626222383194494, "grad_norm": 1.498089044964627, "learning_rate": 5.399213768946205e-06, "loss": 0.8174, "step": 18295 }, { "epoch": 0.6626584570807679, "grad_norm": 1.088117372965663, "learning_rate": 5.3981722593754996e-06, "loss": 0.6744, "step": 18296 }, { "epoch": 0.6626946758420862, "grad_norm": 1.3219833523507174, "learning_rate": 5.397130813129992e-06, "loss": 0.6761, "step": 18297 }, { "epoch": 0.6627308946034046, "grad_norm": 1.2948050870632088, "learning_rate": 5.396089430224006e-06, "loss": 0.6484, "step": 18298 }, { "epoch": 0.6627671133647229, "grad_norm": 1.3168644829173033, "learning_rate": 5.395048110671877e-06, "loss": 0.6812, "step": 18299 }, { "epoch": 0.6628033321260413, "grad_norm": 1.0186331190495426, "learning_rate": 5.3940068544879274e-06, "loss": 0.7702, "step": 18300 }, { "epoch": 0.6628395508873597, "grad_norm": 0.9460637258998915, "learning_rate": 5.392965661686495e-06, "loss": 0.6714, "step": 18301 }, { "epoch": 0.662875769648678, "grad_norm": 1.1477038284174044, "learning_rate": 5.391924532281905e-06, "loss": 0.685, "step": 18302 }, { "epoch": 0.6629119884099964, "grad_norm": 1.0233935877530718, "learning_rate": 5.390883466288479e-06, "loss": 0.6654, "step": 18303 }, { "epoch": 0.6629482071713148, "grad_norm": 1.2489395918751058, "learning_rate": 5.389842463720545e-06, "loss": 0.6611, "step": 18304 }, { "epoch": 0.6629844259326331, "grad_norm": 1.2303208032762014, "learning_rate": 5.388801524592433e-06, "loss": 0.6728, "step": 18305 }, { "epoch": 0.6630206446939515, "grad_norm": 1.4544645571142854, "learning_rate": 5.387760648918463e-06, "loss": 0.7183, "step": 18306 }, { "epoch": 0.6630568634552698, "grad_norm": 1.3937907042461213, "learning_rate": 5.386719836712956e-06, "loss": 0.7271, "step": 18307 }, { "epoch": 0.6630930822165882, "grad_norm": 1.3408904720601778, "learning_rate": 5.385679087990241e-06, "loss": 0.6963, "step": 18308 }, { "epoch": 0.6631293009779066, "grad_norm": 1.1841107615691902, "learning_rate": 5.384638402764637e-06, "loss": 0.7156, "step": 18309 }, { "epoch": 0.6631655197392249, "grad_norm": 1.595036946568711, "learning_rate": 5.383597781050463e-06, "loss": 0.7288, "step": 18310 }, { "epoch": 0.6632017385005433, "grad_norm": 1.406886405124344, "learning_rate": 5.382557222862037e-06, "loss": 0.7307, "step": 18311 }, { "epoch": 0.6632379572618616, "grad_norm": 1.3254769875563763, "learning_rate": 5.381516728213684e-06, "loss": 0.7098, "step": 18312 }, { "epoch": 0.6632741760231801, "grad_norm": 1.3484979937798471, "learning_rate": 5.380476297119719e-06, "loss": 0.7237, "step": 18313 }, { "epoch": 0.6633103947844984, "grad_norm": 1.0240343216663774, "learning_rate": 5.379435929594459e-06, "loss": 0.7427, "step": 18314 }, { "epoch": 0.6633466135458167, "grad_norm": 1.338892188697848, "learning_rate": 5.378395625652218e-06, "loss": 0.6741, "step": 18315 }, { "epoch": 0.6633828323071351, "grad_norm": 1.3594266399352959, "learning_rate": 5.377355385307317e-06, "loss": 0.7116, "step": 18316 }, { "epoch": 0.6634190510684534, "grad_norm": 1.3932033635955712, "learning_rate": 5.376315208574068e-06, "loss": 0.7512, "step": 18317 }, { "epoch": 0.6634552698297719, "grad_norm": 1.2339469606633129, "learning_rate": 5.375275095466779e-06, "loss": 0.6174, "step": 18318 }, { "epoch": 0.6634914885910902, "grad_norm": 1.4423458841047923, "learning_rate": 5.374235045999774e-06, "loss": 0.6485, "step": 18319 }, { "epoch": 0.6635277073524085, "grad_norm": 1.0559944177170788, "learning_rate": 5.373195060187359e-06, "loss": 0.744, "step": 18320 }, { "epoch": 0.6635639261137269, "grad_norm": 1.064727190426405, "learning_rate": 5.3721551380438455e-06, "loss": 0.7524, "step": 18321 }, { "epoch": 0.6636001448750453, "grad_norm": 0.9595014880243108, "learning_rate": 5.371115279583539e-06, "loss": 0.6189, "step": 18322 }, { "epoch": 0.6636363636363637, "grad_norm": 1.6900982997471135, "learning_rate": 5.370075484820758e-06, "loss": 0.713, "step": 18323 }, { "epoch": 0.663672582397682, "grad_norm": 1.321846448982272, "learning_rate": 5.369035753769807e-06, "loss": 0.7002, "step": 18324 }, { "epoch": 0.6637088011590003, "grad_norm": 1.3680714243009118, "learning_rate": 5.3679960864449935e-06, "loss": 0.6911, "step": 18325 }, { "epoch": 0.6637450199203188, "grad_norm": 1.3429937854650225, "learning_rate": 5.366956482860619e-06, "loss": 0.6325, "step": 18326 }, { "epoch": 0.6637812386816371, "grad_norm": 1.2774614365926693, "learning_rate": 5.365916943030998e-06, "loss": 0.6317, "step": 18327 }, { "epoch": 0.6638174574429555, "grad_norm": 1.2853630495217476, "learning_rate": 5.3648774669704315e-06, "loss": 0.6947, "step": 18328 }, { "epoch": 0.6638536762042738, "grad_norm": 1.3454408473088686, "learning_rate": 5.3638380546932215e-06, "loss": 0.6688, "step": 18329 }, { "epoch": 0.6638898949655921, "grad_norm": 1.1207146580488436, "learning_rate": 5.362798706213677e-06, "loss": 0.6975, "step": 18330 }, { "epoch": 0.6639261137269106, "grad_norm": 1.1704264994957476, "learning_rate": 5.361759421546096e-06, "loss": 0.6413, "step": 18331 }, { "epoch": 0.6639623324882289, "grad_norm": 1.3936647473282429, "learning_rate": 5.360720200704781e-06, "loss": 0.6882, "step": 18332 }, { "epoch": 0.6639985512495473, "grad_norm": 1.3989051471140448, "learning_rate": 5.359681043704028e-06, "loss": 0.7326, "step": 18333 }, { "epoch": 0.6640347700108656, "grad_norm": 1.4328809832032288, "learning_rate": 5.358641950558145e-06, "loss": 0.6524, "step": 18334 }, { "epoch": 0.664070988772184, "grad_norm": 1.5140442851538733, "learning_rate": 5.3576029212814285e-06, "loss": 0.6831, "step": 18335 }, { "epoch": 0.6641072075335024, "grad_norm": 1.3946215322532085, "learning_rate": 5.3565639558881745e-06, "loss": 0.7055, "step": 18336 }, { "epoch": 0.6641434262948207, "grad_norm": 1.2545959285737882, "learning_rate": 5.355525054392677e-06, "loss": 0.7232, "step": 18337 }, { "epoch": 0.6641796450561391, "grad_norm": 1.3747779899960684, "learning_rate": 5.354486216809239e-06, "loss": 0.7502, "step": 18338 }, { "epoch": 0.6642158638174575, "grad_norm": 1.4022153932435975, "learning_rate": 5.353447443152153e-06, "loss": 0.7819, "step": 18339 }, { "epoch": 0.6642520825787758, "grad_norm": 1.2859741647494682, "learning_rate": 5.352408733435709e-06, "loss": 0.7042, "step": 18340 }, { "epoch": 0.6642883013400942, "grad_norm": 1.429868419591974, "learning_rate": 5.351370087674208e-06, "loss": 0.6975, "step": 18341 }, { "epoch": 0.6643245201014125, "grad_norm": 1.451860662291825, "learning_rate": 5.3503315058819405e-06, "loss": 0.7377, "step": 18342 }, { "epoch": 0.664360738862731, "grad_norm": 1.3444450592119603, "learning_rate": 5.349292988073196e-06, "loss": 0.7272, "step": 18343 }, { "epoch": 0.6643969576240493, "grad_norm": 1.3206749790941779, "learning_rate": 5.348254534262262e-06, "loss": 0.714, "step": 18344 }, { "epoch": 0.6644331763853676, "grad_norm": 1.4537984008083247, "learning_rate": 5.3472161444634375e-06, "loss": 0.7468, "step": 18345 }, { "epoch": 0.664469395146686, "grad_norm": 1.0550724925711246, "learning_rate": 5.346177818691007e-06, "loss": 0.6931, "step": 18346 }, { "epoch": 0.6645056139080043, "grad_norm": 1.3185947978016341, "learning_rate": 5.345139556959255e-06, "loss": 0.6672, "step": 18347 }, { "epoch": 0.6645418326693228, "grad_norm": 1.295677701667878, "learning_rate": 5.344101359282477e-06, "loss": 0.7273, "step": 18348 }, { "epoch": 0.6645780514306411, "grad_norm": 1.366990528904159, "learning_rate": 5.343063225674956e-06, "loss": 0.7064, "step": 18349 }, { "epoch": 0.6646142701919594, "grad_norm": 0.9270317790166392, "learning_rate": 5.342025156150976e-06, "loss": 0.6796, "step": 18350 }, { "epoch": 0.6646504889532778, "grad_norm": 1.3890852614663016, "learning_rate": 5.340987150724819e-06, "loss": 0.749, "step": 18351 }, { "epoch": 0.6646867077145961, "grad_norm": 1.4911950028722907, "learning_rate": 5.339949209410775e-06, "loss": 0.7304, "step": 18352 }, { "epoch": 0.6647229264759146, "grad_norm": 1.3956280937440273, "learning_rate": 5.338911332223127e-06, "loss": 0.7478, "step": 18353 }, { "epoch": 0.6647591452372329, "grad_norm": 1.3665005797470606, "learning_rate": 5.3378735191761536e-06, "loss": 0.7273, "step": 18354 }, { "epoch": 0.6647953639985512, "grad_norm": 1.3363515354504458, "learning_rate": 5.336835770284133e-06, "loss": 0.7237, "step": 18355 }, { "epoch": 0.6648315827598696, "grad_norm": 1.2996451955685315, "learning_rate": 5.3357980855613545e-06, "loss": 0.7097, "step": 18356 }, { "epoch": 0.664867801521188, "grad_norm": 1.382821912113852, "learning_rate": 5.334760465022092e-06, "loss": 0.7131, "step": 18357 }, { "epoch": 0.6649040202825064, "grad_norm": 1.443789447724408, "learning_rate": 5.3337229086806206e-06, "loss": 0.6825, "step": 18358 }, { "epoch": 0.6649402390438247, "grad_norm": 1.0499642412231511, "learning_rate": 5.332685416551226e-06, "loss": 0.7273, "step": 18359 }, { "epoch": 0.664976457805143, "grad_norm": 1.412091294643788, "learning_rate": 5.331647988648181e-06, "loss": 0.6531, "step": 18360 }, { "epoch": 0.6650126765664615, "grad_norm": 1.0111978340616048, "learning_rate": 5.330610624985762e-06, "loss": 0.7082, "step": 18361 }, { "epoch": 0.6650488953277798, "grad_norm": 1.3464176948034825, "learning_rate": 5.32957332557824e-06, "loss": 0.7043, "step": 18362 }, { "epoch": 0.6650851140890982, "grad_norm": 1.5052749066056739, "learning_rate": 5.3285360904398965e-06, "loss": 0.7073, "step": 18363 }, { "epoch": 0.6651213328504165, "grad_norm": 1.3001841722698506, "learning_rate": 5.327498919585001e-06, "loss": 0.6904, "step": 18364 }, { "epoch": 0.6651575516117348, "grad_norm": 1.250573159479779, "learning_rate": 5.326461813027826e-06, "loss": 0.6366, "step": 18365 }, { "epoch": 0.6651937703730533, "grad_norm": 1.293802169943487, "learning_rate": 5.325424770782639e-06, "loss": 0.6431, "step": 18366 }, { "epoch": 0.6652299891343716, "grad_norm": 2.306241000036444, "learning_rate": 5.324387792863719e-06, "loss": 0.6418, "step": 18367 }, { "epoch": 0.66526620789569, "grad_norm": 1.442806243730068, "learning_rate": 5.3233508792853315e-06, "loss": 0.7289, "step": 18368 }, { "epoch": 0.6653024266570083, "grad_norm": 1.4343532699297792, "learning_rate": 5.32231403006174e-06, "loss": 0.7081, "step": 18369 }, { "epoch": 0.6653386454183267, "grad_norm": 1.4226822114122704, "learning_rate": 5.321277245207223e-06, "loss": 0.7408, "step": 18370 }, { "epoch": 0.6653748641796451, "grad_norm": 1.3719263564644277, "learning_rate": 5.3202405247360415e-06, "loss": 0.6263, "step": 18371 }, { "epoch": 0.6654110829409634, "grad_norm": 1.3188770697659882, "learning_rate": 5.319203868662462e-06, "loss": 0.6982, "step": 18372 }, { "epoch": 0.6654473017022818, "grad_norm": 1.0834092083442501, "learning_rate": 5.318167277000748e-06, "loss": 0.6879, "step": 18373 }, { "epoch": 0.6654835204636002, "grad_norm": 1.0731575376477638, "learning_rate": 5.317130749765167e-06, "loss": 0.6931, "step": 18374 }, { "epoch": 0.6655197392249185, "grad_norm": 1.5322333919842184, "learning_rate": 5.316094286969984e-06, "loss": 0.6464, "step": 18375 }, { "epoch": 0.6655559579862369, "grad_norm": 1.249382984346732, "learning_rate": 5.315057888629458e-06, "loss": 0.6882, "step": 18376 }, { "epoch": 0.6655921767475552, "grad_norm": 1.0347037248994986, "learning_rate": 5.314021554757849e-06, "loss": 0.7078, "step": 18377 }, { "epoch": 0.6656283955088736, "grad_norm": 1.3592874502895278, "learning_rate": 5.312985285369424e-06, "loss": 0.674, "step": 18378 }, { "epoch": 0.665664614270192, "grad_norm": 1.336503098062861, "learning_rate": 5.31194908047844e-06, "loss": 0.6269, "step": 18379 }, { "epoch": 0.6657008330315103, "grad_norm": 1.2701229499206332, "learning_rate": 5.310912940099153e-06, "loss": 0.7448, "step": 18380 }, { "epoch": 0.6657370517928287, "grad_norm": 1.5030925357627878, "learning_rate": 5.3098768642458264e-06, "loss": 0.6973, "step": 18381 }, { "epoch": 0.665773270554147, "grad_norm": 1.351921487276571, "learning_rate": 5.3088408529327165e-06, "loss": 0.6702, "step": 18382 }, { "epoch": 0.6658094893154655, "grad_norm": 1.4771047081048652, "learning_rate": 5.307804906174079e-06, "loss": 0.7145, "step": 18383 }, { "epoch": 0.6658457080767838, "grad_norm": 1.3643408376651762, "learning_rate": 5.306769023984163e-06, "loss": 0.7344, "step": 18384 }, { "epoch": 0.6658819268381021, "grad_norm": 1.329537276798934, "learning_rate": 5.305733206377235e-06, "loss": 0.7525, "step": 18385 }, { "epoch": 0.6659181455994205, "grad_norm": 1.3447957523195486, "learning_rate": 5.304697453367541e-06, "loss": 0.7561, "step": 18386 }, { "epoch": 0.6659543643607388, "grad_norm": 0.9909065916529624, "learning_rate": 5.303661764969336e-06, "loss": 0.6961, "step": 18387 }, { "epoch": 0.6659905831220573, "grad_norm": 1.0080698122271283, "learning_rate": 5.30262614119687e-06, "loss": 0.6505, "step": 18388 }, { "epoch": 0.6660268018833756, "grad_norm": 1.4014164826232955, "learning_rate": 5.3015905820643975e-06, "loss": 0.7606, "step": 18389 }, { "epoch": 0.6660630206446939, "grad_norm": 1.2951068789170086, "learning_rate": 5.300555087586168e-06, "loss": 0.7272, "step": 18390 }, { "epoch": 0.6660992394060123, "grad_norm": 1.4060357910828, "learning_rate": 5.299519657776424e-06, "loss": 0.6288, "step": 18391 }, { "epoch": 0.6661354581673307, "grad_norm": 1.3219181009800565, "learning_rate": 5.2984842926494265e-06, "loss": 0.6444, "step": 18392 }, { "epoch": 0.6661716769286491, "grad_norm": 1.2550971222979346, "learning_rate": 5.297448992219414e-06, "loss": 0.6658, "step": 18393 }, { "epoch": 0.6662078956899674, "grad_norm": 1.2861678148400324, "learning_rate": 5.296413756500636e-06, "loss": 0.6475, "step": 18394 }, { "epoch": 0.6662441144512857, "grad_norm": 1.5403900162355173, "learning_rate": 5.295378585507334e-06, "loss": 0.7255, "step": 18395 }, { "epoch": 0.6662803332126042, "grad_norm": 1.295933462099399, "learning_rate": 5.294343479253761e-06, "loss": 0.653, "step": 18396 }, { "epoch": 0.6663165519739225, "grad_norm": 1.29884561584283, "learning_rate": 5.293308437754155e-06, "loss": 0.6521, "step": 18397 }, { "epoch": 0.6663527707352409, "grad_norm": 1.3050044468551816, "learning_rate": 5.292273461022761e-06, "loss": 0.6562, "step": 18398 }, { "epoch": 0.6663889894965592, "grad_norm": 1.8104224239403415, "learning_rate": 5.291238549073818e-06, "loss": 0.612, "step": 18399 }, { "epoch": 0.6664252082578775, "grad_norm": 1.0771633910019198, "learning_rate": 5.290203701921572e-06, "loss": 0.694, "step": 18400 }, { "epoch": 0.666461427019196, "grad_norm": 1.3344797231219783, "learning_rate": 5.289168919580263e-06, "loss": 0.7346, "step": 18401 }, { "epoch": 0.6664976457805143, "grad_norm": 1.0833990361187273, "learning_rate": 5.288134202064126e-06, "loss": 0.7371, "step": 18402 }, { "epoch": 0.6665338645418327, "grad_norm": 1.3607451177012209, "learning_rate": 5.287099549387405e-06, "loss": 0.7736, "step": 18403 }, { "epoch": 0.666570083303151, "grad_norm": 1.7368851456513341, "learning_rate": 5.286064961564334e-06, "loss": 0.6617, "step": 18404 }, { "epoch": 0.6666063020644694, "grad_norm": 1.2275052970712577, "learning_rate": 5.285030438609153e-06, "loss": 0.7187, "step": 18405 }, { "epoch": 0.6666425208257878, "grad_norm": 0.9619703785875126, "learning_rate": 5.283995980536092e-06, "loss": 0.7185, "step": 18406 }, { "epoch": 0.6666787395871061, "grad_norm": 1.2407898094682803, "learning_rate": 5.282961587359394e-06, "loss": 0.6049, "step": 18407 }, { "epoch": 0.6667149583484245, "grad_norm": 1.2256619304486833, "learning_rate": 5.281927259093289e-06, "loss": 0.7195, "step": 18408 }, { "epoch": 0.6667511771097429, "grad_norm": 1.053043605699572, "learning_rate": 5.2808929957520075e-06, "loss": 0.6902, "step": 18409 }, { "epoch": 0.6667873958710612, "grad_norm": 1.4234892987521341, "learning_rate": 5.279858797349789e-06, "loss": 0.7665, "step": 18410 }, { "epoch": 0.6668236146323796, "grad_norm": 1.5817166733871797, "learning_rate": 5.278824663900859e-06, "loss": 0.7828, "step": 18411 }, { "epoch": 0.6668598333936979, "grad_norm": 1.436166794238825, "learning_rate": 5.277790595419451e-06, "loss": 0.6753, "step": 18412 }, { "epoch": 0.6668960521550164, "grad_norm": 1.6120163470824782, "learning_rate": 5.2767565919197895e-06, "loss": 0.7455, "step": 18413 }, { "epoch": 0.6669322709163347, "grad_norm": 1.5304694132392074, "learning_rate": 5.275722653416112e-06, "loss": 0.6881, "step": 18414 }, { "epoch": 0.666968489677653, "grad_norm": 0.9902592986859758, "learning_rate": 5.274688779922642e-06, "loss": 0.6902, "step": 18415 }, { "epoch": 0.6670047084389714, "grad_norm": 1.4028690922771294, "learning_rate": 5.273654971453606e-06, "loss": 0.7477, "step": 18416 }, { "epoch": 0.6670409272002897, "grad_norm": 1.4105048963838875, "learning_rate": 5.272621228023227e-06, "loss": 0.647, "step": 18417 }, { "epoch": 0.6670771459616082, "grad_norm": 1.3385262705669532, "learning_rate": 5.271587549645738e-06, "loss": 0.7271, "step": 18418 }, { "epoch": 0.6671133647229265, "grad_norm": 1.2390662647228605, "learning_rate": 5.270553936335359e-06, "loss": 0.6549, "step": 18419 }, { "epoch": 0.6671495834842448, "grad_norm": 1.2308758688471997, "learning_rate": 5.2695203881063085e-06, "loss": 0.7385, "step": 18420 }, { "epoch": 0.6671858022455632, "grad_norm": 1.297875241249453, "learning_rate": 5.268486904972819e-06, "loss": 0.7025, "step": 18421 }, { "epoch": 0.6672220210068815, "grad_norm": 1.211694843590121, "learning_rate": 5.2674534869491055e-06, "loss": 0.5561, "step": 18422 }, { "epoch": 0.6672582397682, "grad_norm": 1.3949687736231102, "learning_rate": 5.266420134049392e-06, "loss": 0.6478, "step": 18423 }, { "epoch": 0.6672944585295183, "grad_norm": 1.282704336227171, "learning_rate": 5.2653868462878925e-06, "loss": 0.6465, "step": 18424 }, { "epoch": 0.6673306772908366, "grad_norm": 1.236219655230782, "learning_rate": 5.2643536236788325e-06, "loss": 0.6571, "step": 18425 }, { "epoch": 0.667366896052155, "grad_norm": 1.2805257326053736, "learning_rate": 5.26332046623643e-06, "loss": 0.7187, "step": 18426 }, { "epoch": 0.6674031148134734, "grad_norm": 1.383780887932279, "learning_rate": 5.2622873739749e-06, "loss": 0.7173, "step": 18427 }, { "epoch": 0.6674393335747918, "grad_norm": 1.3608945309286093, "learning_rate": 5.261254346908453e-06, "loss": 0.6972, "step": 18428 }, { "epoch": 0.6674755523361101, "grad_norm": 1.258337887034429, "learning_rate": 5.260221385051315e-06, "loss": 0.7132, "step": 18429 }, { "epoch": 0.6675117710974284, "grad_norm": 1.347561133524889, "learning_rate": 5.259188488417695e-06, "loss": 0.6608, "step": 18430 }, { "epoch": 0.6675479898587469, "grad_norm": 1.5455126784127649, "learning_rate": 5.258155657021804e-06, "loss": 0.7015, "step": 18431 }, { "epoch": 0.6675842086200652, "grad_norm": 1.2958561776118815, "learning_rate": 5.257122890877861e-06, "loss": 0.7166, "step": 18432 }, { "epoch": 0.6676204273813836, "grad_norm": 1.4174705095036537, "learning_rate": 5.2560901900000734e-06, "loss": 0.687, "step": 18433 }, { "epoch": 0.6676566461427019, "grad_norm": 1.51551236392836, "learning_rate": 5.255057554402655e-06, "loss": 0.767, "step": 18434 }, { "epoch": 0.6676928649040202, "grad_norm": 1.1881571000850495, "learning_rate": 5.254024984099809e-06, "loss": 0.6248, "step": 18435 }, { "epoch": 0.6677290836653387, "grad_norm": 1.300720295752347, "learning_rate": 5.2529924791057545e-06, "loss": 0.6605, "step": 18436 }, { "epoch": 0.667765302426657, "grad_norm": 1.537478777014013, "learning_rate": 5.251960039434697e-06, "loss": 0.774, "step": 18437 }, { "epoch": 0.6678015211879754, "grad_norm": 1.4127731873481486, "learning_rate": 5.250927665100833e-06, "loss": 0.7341, "step": 18438 }, { "epoch": 0.6678377399492937, "grad_norm": 1.2609990875396258, "learning_rate": 5.249895356118384e-06, "loss": 0.697, "step": 18439 }, { "epoch": 0.6678739587106121, "grad_norm": 0.9696238908552506, "learning_rate": 5.248863112501547e-06, "loss": 0.7152, "step": 18440 }, { "epoch": 0.6679101774719305, "grad_norm": 1.2640797894135956, "learning_rate": 5.247830934264528e-06, "loss": 0.6891, "step": 18441 }, { "epoch": 0.6679463962332488, "grad_norm": 1.2824213645532008, "learning_rate": 5.246798821421527e-06, "loss": 0.7075, "step": 18442 }, { "epoch": 0.6679826149945672, "grad_norm": 1.3486490918502145, "learning_rate": 5.245766773986755e-06, "loss": 0.6642, "step": 18443 }, { "epoch": 0.6680188337558856, "grad_norm": 1.4193037805028619, "learning_rate": 5.244734791974409e-06, "loss": 0.6537, "step": 18444 }, { "epoch": 0.6680550525172039, "grad_norm": 1.003409408790386, "learning_rate": 5.243702875398687e-06, "loss": 0.6807, "step": 18445 }, { "epoch": 0.6680912712785223, "grad_norm": 1.3349458921040922, "learning_rate": 5.242671024273798e-06, "loss": 0.7302, "step": 18446 }, { "epoch": 0.6681274900398406, "grad_norm": 1.6583049774341476, "learning_rate": 5.241639238613934e-06, "loss": 0.6783, "step": 18447 }, { "epoch": 0.668163708801159, "grad_norm": 1.3130963594181075, "learning_rate": 5.240607518433296e-06, "loss": 0.6012, "step": 18448 }, { "epoch": 0.6681999275624774, "grad_norm": 1.390132371100436, "learning_rate": 5.239575863746076e-06, "loss": 0.7494, "step": 18449 }, { "epoch": 0.6682361463237957, "grad_norm": 1.3652702737489815, "learning_rate": 5.23854427456648e-06, "loss": 0.7096, "step": 18450 }, { "epoch": 0.6682723650851141, "grad_norm": 0.9761612467160506, "learning_rate": 5.237512750908698e-06, "loss": 0.7022, "step": 18451 }, { "epoch": 0.6683085838464324, "grad_norm": 1.6024490135877574, "learning_rate": 5.236481292786922e-06, "loss": 0.7188, "step": 18452 }, { "epoch": 0.6683448026077509, "grad_norm": 1.4328268790901537, "learning_rate": 5.235449900215351e-06, "loss": 0.7646, "step": 18453 }, { "epoch": 0.6683810213690692, "grad_norm": 1.1212876824829123, "learning_rate": 5.2344185732081776e-06, "loss": 0.7074, "step": 18454 }, { "epoch": 0.6684172401303875, "grad_norm": 1.1388990158701289, "learning_rate": 5.2333873117795905e-06, "loss": 0.7109, "step": 18455 }, { "epoch": 0.6684534588917059, "grad_norm": 1.469220779051004, "learning_rate": 5.23235611594378e-06, "loss": 0.6723, "step": 18456 }, { "epoch": 0.6684896776530243, "grad_norm": 1.3976374460707097, "learning_rate": 5.231324985714942e-06, "loss": 0.6612, "step": 18457 }, { "epoch": 0.6685258964143427, "grad_norm": 1.2988788573316834, "learning_rate": 5.2302939211072615e-06, "loss": 0.7377, "step": 18458 }, { "epoch": 0.668562115175661, "grad_norm": 1.409005676464915, "learning_rate": 5.229262922134929e-06, "loss": 0.7721, "step": 18459 }, { "epoch": 0.6685983339369793, "grad_norm": 1.318074815204243, "learning_rate": 5.228231988812126e-06, "loss": 0.7063, "step": 18460 }, { "epoch": 0.6686345526982977, "grad_norm": 1.4892072450151181, "learning_rate": 5.227201121153047e-06, "loss": 0.7423, "step": 18461 }, { "epoch": 0.6686707714596161, "grad_norm": 0.9834388492699107, "learning_rate": 5.226170319171875e-06, "loss": 0.6759, "step": 18462 }, { "epoch": 0.6687069902209345, "grad_norm": 1.2261512518796178, "learning_rate": 5.22513958288279e-06, "loss": 0.6361, "step": 18463 }, { "epoch": 0.6687432089822528, "grad_norm": 1.2415974358152082, "learning_rate": 5.224108912299984e-06, "loss": 0.6402, "step": 18464 }, { "epoch": 0.6687794277435711, "grad_norm": 1.2792561277059094, "learning_rate": 5.223078307437637e-06, "loss": 0.71, "step": 18465 }, { "epoch": 0.6688156465048896, "grad_norm": 1.3035900817173869, "learning_rate": 5.222047768309928e-06, "loss": 0.643, "step": 18466 }, { "epoch": 0.6688518652662079, "grad_norm": 1.3354245234129132, "learning_rate": 5.2210172949310365e-06, "loss": 0.6943, "step": 18467 }, { "epoch": 0.6688880840275263, "grad_norm": 1.2801821448441268, "learning_rate": 5.219986887315151e-06, "loss": 0.6227, "step": 18468 }, { "epoch": 0.6689243027888446, "grad_norm": 1.36629854389023, "learning_rate": 5.218956545476445e-06, "loss": 0.6624, "step": 18469 }, { "epoch": 0.668960521550163, "grad_norm": 1.320181784422949, "learning_rate": 5.2179262694290975e-06, "loss": 0.6881, "step": 18470 }, { "epoch": 0.6689967403114814, "grad_norm": 1.1034948344930635, "learning_rate": 5.216896059187284e-06, "loss": 0.7324, "step": 18471 }, { "epoch": 0.6690329590727997, "grad_norm": 1.2401506170859615, "learning_rate": 5.215865914765186e-06, "loss": 0.7198, "step": 18472 }, { "epoch": 0.6690691778341181, "grad_norm": 1.4245194828812735, "learning_rate": 5.214835836176978e-06, "loss": 0.7348, "step": 18473 }, { "epoch": 0.6691053965954364, "grad_norm": 1.763200428622725, "learning_rate": 5.213805823436828e-06, "loss": 0.7273, "step": 18474 }, { "epoch": 0.6691416153567548, "grad_norm": 1.3892642540301707, "learning_rate": 5.212775876558921e-06, "loss": 0.669, "step": 18475 }, { "epoch": 0.6691778341180732, "grad_norm": 1.4779632001202583, "learning_rate": 5.2117459955574244e-06, "loss": 0.712, "step": 18476 }, { "epoch": 0.6692140528793915, "grad_norm": 1.2733492396490962, "learning_rate": 5.210716180446509e-06, "loss": 0.6309, "step": 18477 }, { "epoch": 0.6692502716407099, "grad_norm": 1.5270508127931572, "learning_rate": 5.209686431240344e-06, "loss": 0.6847, "step": 18478 }, { "epoch": 0.6692864904020283, "grad_norm": 1.398736905700578, "learning_rate": 5.2086567479531066e-06, "loss": 0.6859, "step": 18479 }, { "epoch": 0.6693227091633466, "grad_norm": 1.5100217604411301, "learning_rate": 5.207627130598962e-06, "loss": 0.7409, "step": 18480 }, { "epoch": 0.669358927924665, "grad_norm": 1.493899391995646, "learning_rate": 5.206597579192078e-06, "loss": 0.7041, "step": 18481 }, { "epoch": 0.6693951466859833, "grad_norm": 1.5082371020003489, "learning_rate": 5.205568093746622e-06, "loss": 0.8349, "step": 18482 }, { "epoch": 0.6694313654473018, "grad_norm": 0.9419841600115989, "learning_rate": 5.204538674276765e-06, "loss": 0.6673, "step": 18483 }, { "epoch": 0.6694675842086201, "grad_norm": 1.3676673952912621, "learning_rate": 5.203509320796668e-06, "loss": 0.7342, "step": 18484 }, { "epoch": 0.6695038029699384, "grad_norm": 1.3591818939544094, "learning_rate": 5.202480033320494e-06, "loss": 0.7414, "step": 18485 }, { "epoch": 0.6695400217312568, "grad_norm": 1.4361974364430052, "learning_rate": 5.201450811862414e-06, "loss": 0.7243, "step": 18486 }, { "epoch": 0.6695762404925751, "grad_norm": 1.446234631850325, "learning_rate": 5.200421656436589e-06, "loss": 0.6747, "step": 18487 }, { "epoch": 0.6696124592538936, "grad_norm": 1.5002349778742134, "learning_rate": 5.199392567057178e-06, "loss": 0.7012, "step": 18488 }, { "epoch": 0.6696486780152119, "grad_norm": 1.0108124618990229, "learning_rate": 5.198363543738339e-06, "loss": 0.6807, "step": 18489 }, { "epoch": 0.6696848967765302, "grad_norm": 1.2704417250361195, "learning_rate": 5.19733458649424e-06, "loss": 0.707, "step": 18490 }, { "epoch": 0.6697211155378486, "grad_norm": 1.4033801517988322, "learning_rate": 5.196305695339039e-06, "loss": 0.7358, "step": 18491 }, { "epoch": 0.669757334299167, "grad_norm": 1.3809784469632638, "learning_rate": 5.195276870286892e-06, "loss": 0.6834, "step": 18492 }, { "epoch": 0.6697935530604854, "grad_norm": 1.3764576947262364, "learning_rate": 5.194248111351951e-06, "loss": 0.6849, "step": 18493 }, { "epoch": 0.6698297718218037, "grad_norm": 1.2833525971228859, "learning_rate": 5.193219418548386e-06, "loss": 0.629, "step": 18494 }, { "epoch": 0.669865990583122, "grad_norm": 1.0027434889847684, "learning_rate": 5.192190791890344e-06, "loss": 0.7177, "step": 18495 }, { "epoch": 0.6699022093444404, "grad_norm": 1.4270279137148336, "learning_rate": 5.191162231391977e-06, "loss": 0.7403, "step": 18496 }, { "epoch": 0.6699384281057588, "grad_norm": 1.5195078198310785, "learning_rate": 5.190133737067446e-06, "loss": 0.7254, "step": 18497 }, { "epoch": 0.6699746468670772, "grad_norm": 1.3314093070514987, "learning_rate": 5.189105308930902e-06, "loss": 0.6834, "step": 18498 }, { "epoch": 0.6700108656283955, "grad_norm": 1.0760465967560031, "learning_rate": 5.1880769469964965e-06, "loss": 0.6877, "step": 18499 }, { "epoch": 0.6700470843897138, "grad_norm": 1.2870887434873133, "learning_rate": 5.1870486512783766e-06, "loss": 0.7192, "step": 18500 }, { "epoch": 0.6700833031510323, "grad_norm": 1.0358689549260112, "learning_rate": 5.1860204217907004e-06, "loss": 0.707, "step": 18501 }, { "epoch": 0.6701195219123506, "grad_norm": 0.9886095310786449, "learning_rate": 5.184992258547612e-06, "loss": 0.6966, "step": 18502 }, { "epoch": 0.670155740673669, "grad_norm": 1.3926519184436799, "learning_rate": 5.183964161563262e-06, "loss": 0.6615, "step": 18503 }, { "epoch": 0.6701919594349873, "grad_norm": 1.3608856193472934, "learning_rate": 5.182936130851792e-06, "loss": 0.666, "step": 18504 }, { "epoch": 0.6702281781963056, "grad_norm": 1.0783433265254185, "learning_rate": 5.181908166427358e-06, "loss": 0.7084, "step": 18505 }, { "epoch": 0.6702643969576241, "grad_norm": 1.4222149200275147, "learning_rate": 5.180880268304102e-06, "loss": 0.7391, "step": 18506 }, { "epoch": 0.6703006157189424, "grad_norm": 1.3196021798935749, "learning_rate": 5.179852436496162e-06, "loss": 0.7803, "step": 18507 }, { "epoch": 0.6703368344802608, "grad_norm": 0.9876772137544787, "learning_rate": 5.178824671017693e-06, "loss": 0.6996, "step": 18508 }, { "epoch": 0.6703730532415791, "grad_norm": 3.9427013556730675, "learning_rate": 5.1777969718828336e-06, "loss": 0.7631, "step": 18509 }, { "epoch": 0.6704092720028975, "grad_norm": 1.3688324977451551, "learning_rate": 5.176769339105724e-06, "loss": 0.7238, "step": 18510 }, { "epoch": 0.6704454907642159, "grad_norm": 1.3656630484291161, "learning_rate": 5.175741772700502e-06, "loss": 0.7195, "step": 18511 }, { "epoch": 0.6704817095255342, "grad_norm": 1.4031584400108852, "learning_rate": 5.174714272681317e-06, "loss": 0.7512, "step": 18512 }, { "epoch": 0.6705179282868526, "grad_norm": 1.3735503688475643, "learning_rate": 5.173686839062304e-06, "loss": 0.7325, "step": 18513 }, { "epoch": 0.670554147048171, "grad_norm": 1.244892458456699, "learning_rate": 5.172659471857595e-06, "loss": 0.7745, "step": 18514 }, { "epoch": 0.6705903658094893, "grad_norm": 1.2703959901817332, "learning_rate": 5.1716321710813385e-06, "loss": 0.6932, "step": 18515 }, { "epoch": 0.6706265845708077, "grad_norm": 1.5514864867415143, "learning_rate": 5.170604936747665e-06, "loss": 0.6959, "step": 18516 }, { "epoch": 0.670662803332126, "grad_norm": 1.3088052984305683, "learning_rate": 5.169577768870713e-06, "loss": 0.6585, "step": 18517 }, { "epoch": 0.6706990220934445, "grad_norm": 1.3419984654965524, "learning_rate": 5.16855066746461e-06, "loss": 0.7237, "step": 18518 }, { "epoch": 0.6707352408547628, "grad_norm": 1.16683146849421, "learning_rate": 5.167523632543499e-06, "loss": 0.7069, "step": 18519 }, { "epoch": 0.6707714596160811, "grad_norm": 1.1716698428717456, "learning_rate": 5.16649666412151e-06, "loss": 0.6738, "step": 18520 }, { "epoch": 0.6708076783773995, "grad_norm": 1.2053483825351374, "learning_rate": 5.165469762212772e-06, "loss": 0.7184, "step": 18521 }, { "epoch": 0.6708438971387178, "grad_norm": 1.1369621700958727, "learning_rate": 5.1644429268314155e-06, "loss": 0.821, "step": 18522 }, { "epoch": 0.6708801159000363, "grad_norm": 1.3134795539850252, "learning_rate": 5.163416157991576e-06, "loss": 0.7056, "step": 18523 }, { "epoch": 0.6709163346613546, "grad_norm": 1.4346326206820774, "learning_rate": 5.162389455707382e-06, "loss": 0.7529, "step": 18524 }, { "epoch": 0.6709525534226729, "grad_norm": 0.9928739967751181, "learning_rate": 5.161362819992954e-06, "loss": 0.6885, "step": 18525 }, { "epoch": 0.6709887721839913, "grad_norm": 1.1882513799932224, "learning_rate": 5.16033625086243e-06, "loss": 0.655, "step": 18526 }, { "epoch": 0.6710249909453097, "grad_norm": 1.2485949300959243, "learning_rate": 5.159309748329933e-06, "loss": 0.683, "step": 18527 }, { "epoch": 0.6710612097066281, "grad_norm": 0.9970565427103476, "learning_rate": 5.158283312409584e-06, "loss": 0.6727, "step": 18528 }, { "epoch": 0.6710974284679464, "grad_norm": 1.3100523122095884, "learning_rate": 5.15725694311551e-06, "loss": 0.7089, "step": 18529 }, { "epoch": 0.6711336472292647, "grad_norm": 1.357927519115618, "learning_rate": 5.156230640461839e-06, "loss": 0.7428, "step": 18530 }, { "epoch": 0.6711698659905831, "grad_norm": 1.4158138526014508, "learning_rate": 5.155204404462689e-06, "loss": 0.6692, "step": 18531 }, { "epoch": 0.6712060847519015, "grad_norm": 1.4060291236029798, "learning_rate": 5.154178235132184e-06, "loss": 0.7189, "step": 18532 }, { "epoch": 0.6712423035132199, "grad_norm": 1.4478349865600442, "learning_rate": 5.1531521324844415e-06, "loss": 0.7504, "step": 18533 }, { "epoch": 0.6712785222745382, "grad_norm": 1.4676836815025631, "learning_rate": 5.1521260965335875e-06, "loss": 0.7312, "step": 18534 }, { "epoch": 0.6713147410358565, "grad_norm": 1.3888551762747747, "learning_rate": 5.151100127293738e-06, "loss": 0.7592, "step": 18535 }, { "epoch": 0.671350959797175, "grad_norm": 1.3647045054095774, "learning_rate": 5.150074224779008e-06, "loss": 0.6813, "step": 18536 }, { "epoch": 0.6713871785584933, "grad_norm": 1.038335875217607, "learning_rate": 5.149048389003521e-06, "loss": 0.7062, "step": 18537 }, { "epoch": 0.6714233973198117, "grad_norm": 1.3380295632608301, "learning_rate": 5.148022619981392e-06, "loss": 0.6267, "step": 18538 }, { "epoch": 0.67145961608113, "grad_norm": 1.0061133310056278, "learning_rate": 5.146996917726735e-06, "loss": 0.6834, "step": 18539 }, { "epoch": 0.6714958348424483, "grad_norm": 1.3064751541669086, "learning_rate": 5.145971282253661e-06, "loss": 0.6102, "step": 18540 }, { "epoch": 0.6715320536037668, "grad_norm": 1.3288972432696409, "learning_rate": 5.1449457135762905e-06, "loss": 0.6697, "step": 18541 }, { "epoch": 0.6715682723650851, "grad_norm": 1.4134691004960045, "learning_rate": 5.1439202117087325e-06, "loss": 0.7113, "step": 18542 }, { "epoch": 0.6716044911264035, "grad_norm": 1.3060950878107709, "learning_rate": 5.1428947766651e-06, "loss": 0.6609, "step": 18543 }, { "epoch": 0.6716407098877218, "grad_norm": 1.4783774113523744, "learning_rate": 5.141869408459499e-06, "loss": 0.6741, "step": 18544 }, { "epoch": 0.6716769286490402, "grad_norm": 1.5507029374285048, "learning_rate": 5.140844107106047e-06, "loss": 0.6763, "step": 18545 }, { "epoch": 0.6717131474103586, "grad_norm": 1.3160354915646832, "learning_rate": 5.139818872618852e-06, "loss": 0.6673, "step": 18546 }, { "epoch": 0.6717493661716769, "grad_norm": 1.437679169466659, "learning_rate": 5.138793705012013e-06, "loss": 0.6963, "step": 18547 }, { "epoch": 0.6717855849329953, "grad_norm": 1.2335810065592707, "learning_rate": 5.137768604299649e-06, "loss": 0.6685, "step": 18548 }, { "epoch": 0.6718218036943137, "grad_norm": 1.318507078868295, "learning_rate": 5.136743570495861e-06, "loss": 0.6584, "step": 18549 }, { "epoch": 0.671858022455632, "grad_norm": 2.245976430068027, "learning_rate": 5.135718603614757e-06, "loss": 0.6859, "step": 18550 }, { "epoch": 0.6718942412169504, "grad_norm": 1.0208524538402766, "learning_rate": 5.134693703670432e-06, "loss": 0.7453, "step": 18551 }, { "epoch": 0.6719304599782687, "grad_norm": 1.4306291673590417, "learning_rate": 5.1336688706770025e-06, "loss": 0.6848, "step": 18552 }, { "epoch": 0.6719666787395872, "grad_norm": 3.0713711823126917, "learning_rate": 5.132644104648565e-06, "loss": 0.6977, "step": 18553 }, { "epoch": 0.6720028975009055, "grad_norm": 1.3486808834582733, "learning_rate": 5.131619405599222e-06, "loss": 0.6363, "step": 18554 }, { "epoch": 0.6720391162622238, "grad_norm": 1.2451439618990536, "learning_rate": 5.1305947735430684e-06, "loss": 0.6586, "step": 18555 }, { "epoch": 0.6720753350235422, "grad_norm": 1.4483297814125182, "learning_rate": 5.129570208494214e-06, "loss": 0.735, "step": 18556 }, { "epoch": 0.6721115537848605, "grad_norm": 1.0604880855728631, "learning_rate": 5.1285457104667525e-06, "loss": 0.7023, "step": 18557 }, { "epoch": 0.672147772546179, "grad_norm": 1.2811595852863327, "learning_rate": 5.12752127947478e-06, "loss": 0.6784, "step": 18558 }, { "epoch": 0.6721839913074973, "grad_norm": 1.4016294360257755, "learning_rate": 5.126496915532398e-06, "loss": 0.7305, "step": 18559 }, { "epoch": 0.6722202100688156, "grad_norm": 0.9795278446908756, "learning_rate": 5.125472618653702e-06, "loss": 0.7172, "step": 18560 }, { "epoch": 0.672256428830134, "grad_norm": 1.324976716112515, "learning_rate": 5.124448388852785e-06, "loss": 0.6268, "step": 18561 }, { "epoch": 0.6722926475914524, "grad_norm": 1.3308091094520655, "learning_rate": 5.12342422614374e-06, "loss": 0.688, "step": 18562 }, { "epoch": 0.6723288663527708, "grad_norm": 1.3415860533807074, "learning_rate": 5.122400130540666e-06, "loss": 0.7169, "step": 18563 }, { "epoch": 0.6723650851140891, "grad_norm": 1.061838533487042, "learning_rate": 5.121376102057651e-06, "loss": 0.6694, "step": 18564 }, { "epoch": 0.6724013038754074, "grad_norm": 1.4204011584403775, "learning_rate": 5.120352140708788e-06, "loss": 0.7001, "step": 18565 }, { "epoch": 0.6724375226367258, "grad_norm": 1.248383069686981, "learning_rate": 5.119328246508163e-06, "loss": 0.6834, "step": 18566 }, { "epoch": 0.6724737413980442, "grad_norm": 1.306539768266682, "learning_rate": 5.1183044194698735e-06, "loss": 0.7388, "step": 18567 }, { "epoch": 0.6725099601593626, "grad_norm": 1.3632198855517323, "learning_rate": 5.117280659608006e-06, "loss": 0.7315, "step": 18568 }, { "epoch": 0.6725461789206809, "grad_norm": 1.2778632204289382, "learning_rate": 5.116256966936641e-06, "loss": 0.7124, "step": 18569 }, { "epoch": 0.6725823976819992, "grad_norm": 1.2746399902083014, "learning_rate": 5.1152333414698774e-06, "loss": 0.7074, "step": 18570 }, { "epoch": 0.6726186164433177, "grad_norm": 0.970323397323557, "learning_rate": 5.114209783221794e-06, "loss": 0.6779, "step": 18571 }, { "epoch": 0.672654835204636, "grad_norm": 0.9747855850404822, "learning_rate": 5.113186292206478e-06, "loss": 0.73, "step": 18572 }, { "epoch": 0.6726910539659544, "grad_norm": 1.3058538848140784, "learning_rate": 5.112162868438008e-06, "loss": 0.699, "step": 18573 }, { "epoch": 0.6727272727272727, "grad_norm": 1.1546225173672033, "learning_rate": 5.111139511930475e-06, "loss": 0.7003, "step": 18574 }, { "epoch": 0.672763491488591, "grad_norm": 1.4097153464410246, "learning_rate": 5.110116222697959e-06, "loss": 0.6931, "step": 18575 }, { "epoch": 0.6727997102499095, "grad_norm": 1.378797811885751, "learning_rate": 5.109093000754536e-06, "loss": 0.6903, "step": 18576 }, { "epoch": 0.6728359290112278, "grad_norm": 1.3276533216197308, "learning_rate": 5.108069846114294e-06, "loss": 0.757, "step": 18577 }, { "epoch": 0.6728721477725462, "grad_norm": 1.4333983211462709, "learning_rate": 5.107046758791311e-06, "loss": 0.6884, "step": 18578 }, { "epoch": 0.6729083665338645, "grad_norm": 1.0269553244906557, "learning_rate": 5.106023738799664e-06, "loss": 0.6228, "step": 18579 }, { "epoch": 0.6729445852951829, "grad_norm": 1.3944253340336983, "learning_rate": 5.105000786153424e-06, "loss": 0.6277, "step": 18580 }, { "epoch": 0.6729808040565013, "grad_norm": 1.3872278568432685, "learning_rate": 5.103977900866681e-06, "loss": 0.7276, "step": 18581 }, { "epoch": 0.6730170228178196, "grad_norm": 1.0682839788384706, "learning_rate": 5.102955082953502e-06, "loss": 0.7114, "step": 18582 }, { "epoch": 0.673053241579138, "grad_norm": 1.366207317759236, "learning_rate": 5.101932332427965e-06, "loss": 0.7249, "step": 18583 }, { "epoch": 0.6730894603404564, "grad_norm": 1.0456006356199141, "learning_rate": 5.10090964930414e-06, "loss": 0.6496, "step": 18584 }, { "epoch": 0.6731256791017747, "grad_norm": 1.4022651560019248, "learning_rate": 5.099887033596106e-06, "loss": 0.6426, "step": 18585 }, { "epoch": 0.6731618978630931, "grad_norm": 1.3872203913538426, "learning_rate": 5.0988644853179315e-06, "loss": 0.6667, "step": 18586 }, { "epoch": 0.6731981166244114, "grad_norm": 1.6587606757531217, "learning_rate": 5.0978420044836855e-06, "loss": 0.7038, "step": 18587 }, { "epoch": 0.6732343353857299, "grad_norm": 1.3702598114851017, "learning_rate": 5.096819591107445e-06, "loss": 0.7682, "step": 18588 }, { "epoch": 0.6732705541470482, "grad_norm": 1.3503513725519205, "learning_rate": 5.095797245203273e-06, "loss": 0.7014, "step": 18589 }, { "epoch": 0.6733067729083665, "grad_norm": 1.0644050072153524, "learning_rate": 5.094774966785242e-06, "loss": 0.7277, "step": 18590 }, { "epoch": 0.6733429916696849, "grad_norm": 1.1626555671411611, "learning_rate": 5.093752755867413e-06, "loss": 0.7587, "step": 18591 }, { "epoch": 0.6733792104310032, "grad_norm": 1.1887289368862757, "learning_rate": 5.092730612463861e-06, "loss": 0.6265, "step": 18592 }, { "epoch": 0.6734154291923217, "grad_norm": 1.27552302916104, "learning_rate": 5.091708536588648e-06, "loss": 0.6701, "step": 18593 }, { "epoch": 0.67345164795364, "grad_norm": 1.340757919354597, "learning_rate": 5.090686528255838e-06, "loss": 0.7036, "step": 18594 }, { "epoch": 0.6734878667149583, "grad_norm": 1.3623500720705526, "learning_rate": 5.089664587479491e-06, "loss": 0.7176, "step": 18595 }, { "epoch": 0.6735240854762767, "grad_norm": 1.3062566148725054, "learning_rate": 5.0886427142736785e-06, "loss": 0.7221, "step": 18596 }, { "epoch": 0.673560304237595, "grad_norm": 1.3562005236166599, "learning_rate": 5.087620908652458e-06, "loss": 0.675, "step": 18597 }, { "epoch": 0.6735965229989135, "grad_norm": 0.9904208548491822, "learning_rate": 5.086599170629885e-06, "loss": 0.7068, "step": 18598 }, { "epoch": 0.6736327417602318, "grad_norm": 1.461219022267904, "learning_rate": 5.085577500220029e-06, "loss": 0.7527, "step": 18599 }, { "epoch": 0.6736689605215501, "grad_norm": 1.375285574127863, "learning_rate": 5.0845558974369445e-06, "loss": 0.7181, "step": 18600 }, { "epoch": 0.6737051792828685, "grad_norm": 1.427581878116066, "learning_rate": 5.083534362294691e-06, "loss": 0.7844, "step": 18601 }, { "epoch": 0.6737413980441869, "grad_norm": 1.2040658788022724, "learning_rate": 5.082512894807321e-06, "loss": 0.7203, "step": 18602 }, { "epoch": 0.6737776168055053, "grad_norm": 1.2190511996886657, "learning_rate": 5.081491494988897e-06, "loss": 0.6654, "step": 18603 }, { "epoch": 0.6738138355668236, "grad_norm": 1.4120196080463967, "learning_rate": 5.080470162853473e-06, "loss": 0.643, "step": 18604 }, { "epoch": 0.6738500543281419, "grad_norm": 1.3511291798785385, "learning_rate": 5.0794488984151005e-06, "loss": 0.6776, "step": 18605 }, { "epoch": 0.6738862730894604, "grad_norm": 1.383296331660382, "learning_rate": 5.078427701687832e-06, "loss": 0.7889, "step": 18606 }, { "epoch": 0.6739224918507787, "grad_norm": 1.4709240651180675, "learning_rate": 5.077406572685727e-06, "loss": 0.7639, "step": 18607 }, { "epoch": 0.6739587106120971, "grad_norm": 1.237806889683318, "learning_rate": 5.076385511422832e-06, "loss": 0.6676, "step": 18608 }, { "epoch": 0.6739949293734154, "grad_norm": 1.8285061019228972, "learning_rate": 5.075364517913196e-06, "loss": 0.6904, "step": 18609 }, { "epoch": 0.6740311481347337, "grad_norm": 1.5235284240194424, "learning_rate": 5.074343592170874e-06, "loss": 0.6719, "step": 18610 }, { "epoch": 0.6740673668960522, "grad_norm": 1.305456323296274, "learning_rate": 5.073322734209913e-06, "loss": 0.7335, "step": 18611 }, { "epoch": 0.6741035856573705, "grad_norm": 1.5566984043861298, "learning_rate": 5.07230194404436e-06, "loss": 0.7479, "step": 18612 }, { "epoch": 0.6741398044186889, "grad_norm": 1.3436225134788125, "learning_rate": 5.071281221688258e-06, "loss": 0.6917, "step": 18613 }, { "epoch": 0.6741760231800072, "grad_norm": 1.408167995353315, "learning_rate": 5.070260567155661e-06, "loss": 0.7075, "step": 18614 }, { "epoch": 0.6742122419413256, "grad_norm": 1.4391918203998315, "learning_rate": 5.069239980460612e-06, "loss": 0.7113, "step": 18615 }, { "epoch": 0.674248460702644, "grad_norm": 1.3543461980164493, "learning_rate": 5.068219461617151e-06, "loss": 0.706, "step": 18616 }, { "epoch": 0.6742846794639623, "grad_norm": 1.1527497067477346, "learning_rate": 5.067199010639321e-06, "loss": 0.7395, "step": 18617 }, { "epoch": 0.6743208982252807, "grad_norm": 1.3824178663283513, "learning_rate": 5.06617862754117e-06, "loss": 0.6929, "step": 18618 }, { "epoch": 0.6743571169865991, "grad_norm": 1.4455806547855874, "learning_rate": 5.065158312336738e-06, "loss": 0.7199, "step": 18619 }, { "epoch": 0.6743933357479174, "grad_norm": 1.4344985299948856, "learning_rate": 5.064138065040058e-06, "loss": 0.6997, "step": 18620 }, { "epoch": 0.6744295545092358, "grad_norm": 1.2861813903185044, "learning_rate": 5.063117885665179e-06, "loss": 0.6459, "step": 18621 }, { "epoch": 0.6744657732705541, "grad_norm": 1.5733595436695043, "learning_rate": 5.062097774226137e-06, "loss": 0.6848, "step": 18622 }, { "epoch": 0.6745019920318726, "grad_norm": 1.3622683742427852, "learning_rate": 5.061077730736968e-06, "loss": 0.6676, "step": 18623 }, { "epoch": 0.6745382107931909, "grad_norm": 1.4530882202641608, "learning_rate": 5.060057755211705e-06, "loss": 0.7478, "step": 18624 }, { "epoch": 0.6745744295545092, "grad_norm": 1.4022459883337501, "learning_rate": 5.059037847664392e-06, "loss": 0.6767, "step": 18625 }, { "epoch": 0.6746106483158276, "grad_norm": 0.9450730839744003, "learning_rate": 5.05801800810906e-06, "loss": 0.6662, "step": 18626 }, { "epoch": 0.6746468670771459, "grad_norm": 1.0644992182424275, "learning_rate": 5.056998236559738e-06, "loss": 0.7035, "step": 18627 }, { "epoch": 0.6746830858384644, "grad_norm": 1.214204527274964, "learning_rate": 5.055978533030469e-06, "loss": 0.696, "step": 18628 }, { "epoch": 0.6747193045997827, "grad_norm": 1.3498798415400282, "learning_rate": 5.054958897535278e-06, "loss": 0.6394, "step": 18629 }, { "epoch": 0.674755523361101, "grad_norm": 1.034755895029264, "learning_rate": 5.053939330088199e-06, "loss": 0.662, "step": 18630 }, { "epoch": 0.6747917421224194, "grad_norm": 1.084525072129113, "learning_rate": 5.0529198307032555e-06, "loss": 0.6962, "step": 18631 }, { "epoch": 0.6748279608837378, "grad_norm": 1.3873194884249542, "learning_rate": 5.051900399394487e-06, "loss": 0.8169, "step": 18632 }, { "epoch": 0.6748641796450562, "grad_norm": 0.9864684965788952, "learning_rate": 5.0508810361759165e-06, "loss": 0.7071, "step": 18633 }, { "epoch": 0.6749003984063745, "grad_norm": 0.9430205205387432, "learning_rate": 5.049861741061571e-06, "loss": 0.6022, "step": 18634 }, { "epoch": 0.6749366171676928, "grad_norm": 1.0051726716419762, "learning_rate": 5.048842514065475e-06, "loss": 0.7147, "step": 18635 }, { "epoch": 0.6749728359290113, "grad_norm": 1.2683360342304761, "learning_rate": 5.0478233552016585e-06, "loss": 0.6889, "step": 18636 }, { "epoch": 0.6750090546903296, "grad_norm": 1.3343494960734048, "learning_rate": 5.046804264484145e-06, "loss": 0.7287, "step": 18637 }, { "epoch": 0.675045273451648, "grad_norm": 1.3939220275794157, "learning_rate": 5.045785241926954e-06, "loss": 0.6843, "step": 18638 }, { "epoch": 0.6750814922129663, "grad_norm": 1.4017448926449578, "learning_rate": 5.044766287544114e-06, "loss": 0.7301, "step": 18639 }, { "epoch": 0.6751177109742846, "grad_norm": 1.3836337630324997, "learning_rate": 5.043747401349643e-06, "loss": 0.7049, "step": 18640 }, { "epoch": 0.6751539297356031, "grad_norm": 1.357463792798614, "learning_rate": 5.042728583357566e-06, "loss": 0.7481, "step": 18641 }, { "epoch": 0.6751901484969214, "grad_norm": 1.059415862312422, "learning_rate": 5.041709833581893e-06, "loss": 0.7183, "step": 18642 }, { "epoch": 0.6752263672582398, "grad_norm": 1.1589995841252045, "learning_rate": 5.0406911520366545e-06, "loss": 0.6606, "step": 18643 }, { "epoch": 0.6752625860195581, "grad_norm": 1.842730956808452, "learning_rate": 5.039672538735861e-06, "loss": 0.6861, "step": 18644 }, { "epoch": 0.6752988047808764, "grad_norm": 1.1166231217975364, "learning_rate": 5.0386539936935345e-06, "loss": 0.7352, "step": 18645 }, { "epoch": 0.6753350235421949, "grad_norm": 1.2613942641387303, "learning_rate": 5.037635516923684e-06, "loss": 0.6434, "step": 18646 }, { "epoch": 0.6753712423035132, "grad_norm": 1.5818570476743168, "learning_rate": 5.036617108440332e-06, "loss": 0.6885, "step": 18647 }, { "epoch": 0.6754074610648316, "grad_norm": 1.2744334800380968, "learning_rate": 5.035598768257495e-06, "loss": 0.6293, "step": 18648 }, { "epoch": 0.67544367982615, "grad_norm": 1.1596589274806886, "learning_rate": 5.034580496389172e-06, "loss": 0.657, "step": 18649 }, { "epoch": 0.6754798985874683, "grad_norm": 1.129636103835525, "learning_rate": 5.03356229284939e-06, "loss": 0.6463, "step": 18650 }, { "epoch": 0.6755161173487867, "grad_norm": 1.498102841084076, "learning_rate": 5.032544157652152e-06, "loss": 0.6559, "step": 18651 }, { "epoch": 0.675552336110105, "grad_norm": 1.2072315439529722, "learning_rate": 5.031526090811469e-06, "loss": 0.6789, "step": 18652 }, { "epoch": 0.6755885548714234, "grad_norm": 1.2898185680156276, "learning_rate": 5.030508092341357e-06, "loss": 0.7315, "step": 18653 }, { "epoch": 0.6756247736327418, "grad_norm": 1.2736963845603004, "learning_rate": 5.0294901622558204e-06, "loss": 0.7148, "step": 18654 }, { "epoch": 0.6756609923940601, "grad_norm": 1.437536335050544, "learning_rate": 5.028472300568865e-06, "loss": 0.6801, "step": 18655 }, { "epoch": 0.6756972111553785, "grad_norm": 1.2443615096562708, "learning_rate": 5.027454507294497e-06, "loss": 0.6839, "step": 18656 }, { "epoch": 0.6757334299166968, "grad_norm": 1.3771181237673304, "learning_rate": 5.026436782446727e-06, "loss": 0.6906, "step": 18657 }, { "epoch": 0.6757696486780153, "grad_norm": 0.9436369790069196, "learning_rate": 5.025419126039557e-06, "loss": 0.6827, "step": 18658 }, { "epoch": 0.6758058674393336, "grad_norm": 1.2732907236968358, "learning_rate": 5.024401538086991e-06, "loss": 0.701, "step": 18659 }, { "epoch": 0.6758420862006519, "grad_norm": 1.1981514388822796, "learning_rate": 5.023384018603029e-06, "loss": 0.6531, "step": 18660 }, { "epoch": 0.6758783049619703, "grad_norm": 1.0136688949930712, "learning_rate": 5.022366567601679e-06, "loss": 0.6995, "step": 18661 }, { "epoch": 0.6759145237232886, "grad_norm": 1.1768960915290017, "learning_rate": 5.0213491850969375e-06, "loss": 0.7304, "step": 18662 }, { "epoch": 0.6759507424846071, "grad_norm": 1.0252619044368334, "learning_rate": 5.020331871102804e-06, "loss": 0.6408, "step": 18663 }, { "epoch": 0.6759869612459254, "grad_norm": 1.3701640316166581, "learning_rate": 5.019314625633283e-06, "loss": 0.6858, "step": 18664 }, { "epoch": 0.6760231800072437, "grad_norm": 1.261454440856911, "learning_rate": 5.018297448702367e-06, "loss": 0.6667, "step": 18665 }, { "epoch": 0.6760593987685621, "grad_norm": 1.3142809436251797, "learning_rate": 5.0172803403240575e-06, "loss": 0.7108, "step": 18666 }, { "epoch": 0.6760956175298805, "grad_norm": 1.4031017009748399, "learning_rate": 5.016263300512345e-06, "loss": 0.733, "step": 18667 }, { "epoch": 0.6761318362911989, "grad_norm": 1.4272732622795352, "learning_rate": 5.0152463292812316e-06, "loss": 0.6668, "step": 18668 }, { "epoch": 0.6761680550525172, "grad_norm": 1.423843064511052, "learning_rate": 5.014229426644708e-06, "loss": 0.7403, "step": 18669 }, { "epoch": 0.6762042738138355, "grad_norm": 1.4315430567594534, "learning_rate": 5.013212592616765e-06, "loss": 0.6158, "step": 18670 }, { "epoch": 0.676240492575154, "grad_norm": 1.271790216435405, "learning_rate": 5.012195827211404e-06, "loss": 0.6915, "step": 18671 }, { "epoch": 0.6762767113364723, "grad_norm": 1.034327645249671, "learning_rate": 5.011179130442609e-06, "loss": 0.6824, "step": 18672 }, { "epoch": 0.6763129300977907, "grad_norm": 1.3405238280101202, "learning_rate": 5.0101625023243734e-06, "loss": 0.7098, "step": 18673 }, { "epoch": 0.676349148859109, "grad_norm": 1.327652556874518, "learning_rate": 5.009145942870684e-06, "loss": 0.7129, "step": 18674 }, { "epoch": 0.6763853676204273, "grad_norm": 0.9521727660323297, "learning_rate": 5.008129452095533e-06, "loss": 0.7286, "step": 18675 }, { "epoch": 0.6764215863817458, "grad_norm": 1.2741785305885045, "learning_rate": 5.007113030012908e-06, "loss": 0.7483, "step": 18676 }, { "epoch": 0.6764578051430641, "grad_norm": 1.398979329786951, "learning_rate": 5.006096676636795e-06, "loss": 0.669, "step": 18677 }, { "epoch": 0.6764940239043825, "grad_norm": 1.3406147108094417, "learning_rate": 5.005080391981176e-06, "loss": 0.7328, "step": 18678 }, { "epoch": 0.6765302426657008, "grad_norm": 1.364149487658722, "learning_rate": 5.0040641760600425e-06, "loss": 0.7168, "step": 18679 }, { "epoch": 0.6765664614270192, "grad_norm": 1.2961643042569482, "learning_rate": 5.0030480288873775e-06, "loss": 0.6419, "step": 18680 }, { "epoch": 0.6766026801883376, "grad_norm": 1.2457841149154498, "learning_rate": 5.002031950477157e-06, "loss": 0.6888, "step": 18681 }, { "epoch": 0.6766388989496559, "grad_norm": 1.2707318079924594, "learning_rate": 5.001015940843374e-06, "loss": 0.6429, "step": 18682 }, { "epoch": 0.6766751177109743, "grad_norm": 1.3511318681147793, "learning_rate": 5.000000000000003e-06, "loss": 0.6943, "step": 18683 }, { "epoch": 0.6767113364722926, "grad_norm": 1.3453816485273031, "learning_rate": 4.998984127961025e-06, "loss": 0.7086, "step": 18684 }, { "epoch": 0.676747555233611, "grad_norm": 1.3834350990662767, "learning_rate": 4.997968324740416e-06, "loss": 0.6816, "step": 18685 }, { "epoch": 0.6767837739949294, "grad_norm": 1.298039299336958, "learning_rate": 4.996952590352162e-06, "loss": 0.7164, "step": 18686 }, { "epoch": 0.6768199927562477, "grad_norm": 1.252844982301802, "learning_rate": 4.995936924810237e-06, "loss": 0.63, "step": 18687 }, { "epoch": 0.6768562115175661, "grad_norm": 1.0242019794934814, "learning_rate": 4.9949213281286165e-06, "loss": 0.6769, "step": 18688 }, { "epoch": 0.6768924302788845, "grad_norm": 1.3866755478805028, "learning_rate": 4.993905800321272e-06, "loss": 0.6653, "step": 18689 }, { "epoch": 0.6769286490402028, "grad_norm": 0.9918025071639907, "learning_rate": 4.992890341402187e-06, "loss": 0.7727, "step": 18690 }, { "epoch": 0.6769648678015212, "grad_norm": 1.3207611656241398, "learning_rate": 4.99187495138533e-06, "loss": 0.7436, "step": 18691 }, { "epoch": 0.6770010865628395, "grad_norm": 1.4758231239375872, "learning_rate": 4.990859630284671e-06, "loss": 0.7396, "step": 18692 }, { "epoch": 0.677037305324158, "grad_norm": 1.298546256406438, "learning_rate": 4.989844378114188e-06, "loss": 0.7283, "step": 18693 }, { "epoch": 0.6770735240854763, "grad_norm": 1.2026166678029881, "learning_rate": 4.98882919488785e-06, "loss": 0.6832, "step": 18694 }, { "epoch": 0.6771097428467946, "grad_norm": 1.1859367246524755, "learning_rate": 4.987814080619624e-06, "loss": 0.605, "step": 18695 }, { "epoch": 0.677145961608113, "grad_norm": 1.1505135900458543, "learning_rate": 4.986799035323477e-06, "loss": 0.6764, "step": 18696 }, { "epoch": 0.6771821803694313, "grad_norm": 1.2395465407768198, "learning_rate": 4.9857840590133845e-06, "loss": 0.6772, "step": 18697 }, { "epoch": 0.6772183991307498, "grad_norm": 1.2392154638270643, "learning_rate": 4.984769151703308e-06, "loss": 0.7252, "step": 18698 }, { "epoch": 0.6772546178920681, "grad_norm": 1.3688721767748597, "learning_rate": 4.983754313407216e-06, "loss": 0.6885, "step": 18699 }, { "epoch": 0.6772908366533864, "grad_norm": 1.2212377622927937, "learning_rate": 4.982739544139068e-06, "loss": 0.6798, "step": 18700 }, { "epoch": 0.6773270554147048, "grad_norm": 1.3604767146873564, "learning_rate": 4.981724843912836e-06, "loss": 0.65, "step": 18701 }, { "epoch": 0.6773632741760232, "grad_norm": 1.274887314981441, "learning_rate": 4.98071021274248e-06, "loss": 0.7028, "step": 18702 }, { "epoch": 0.6773994929373416, "grad_norm": 1.4027874146485604, "learning_rate": 4.979695650641958e-06, "loss": 0.7096, "step": 18703 }, { "epoch": 0.6774357116986599, "grad_norm": 1.4268541727810118, "learning_rate": 4.9786811576252395e-06, "loss": 0.7867, "step": 18704 }, { "epoch": 0.6774719304599782, "grad_norm": 1.2912089625804952, "learning_rate": 4.977666733706279e-06, "loss": 0.7151, "step": 18705 }, { "epoch": 0.6775081492212967, "grad_norm": 1.34455502315608, "learning_rate": 4.976652378899038e-06, "loss": 0.7233, "step": 18706 }, { "epoch": 0.677544367982615, "grad_norm": 1.3361058491603013, "learning_rate": 4.975638093217471e-06, "loss": 0.6482, "step": 18707 }, { "epoch": 0.6775805867439334, "grad_norm": 1.3065951137000962, "learning_rate": 4.974623876675541e-06, "loss": 0.6543, "step": 18708 }, { "epoch": 0.6776168055052517, "grad_norm": 1.3613945070498215, "learning_rate": 4.973609729287204e-06, "loss": 0.7171, "step": 18709 }, { "epoch": 0.67765302426657, "grad_norm": 1.1037814853913177, "learning_rate": 4.972595651066412e-06, "loss": 0.7171, "step": 18710 }, { "epoch": 0.6776892430278885, "grad_norm": 1.1966820271906358, "learning_rate": 4.971581642027118e-06, "loss": 0.7213, "step": 18711 }, { "epoch": 0.6777254617892068, "grad_norm": 1.7531396851630037, "learning_rate": 4.970567702183283e-06, "loss": 0.6646, "step": 18712 }, { "epoch": 0.6777616805505252, "grad_norm": 1.394665387057601, "learning_rate": 4.969553831548856e-06, "loss": 0.7587, "step": 18713 }, { "epoch": 0.6777978993118435, "grad_norm": 1.4431155447910358, "learning_rate": 4.968540030137783e-06, "loss": 0.7371, "step": 18714 }, { "epoch": 0.6778341180731619, "grad_norm": 1.428714144658148, "learning_rate": 4.967526297964025e-06, "loss": 0.6708, "step": 18715 }, { "epoch": 0.6778703368344803, "grad_norm": 1.419301994372452, "learning_rate": 4.9665126350415275e-06, "loss": 0.6554, "step": 18716 }, { "epoch": 0.6779065555957986, "grad_norm": 1.0077445607748554, "learning_rate": 4.965499041384238e-06, "loss": 0.7183, "step": 18717 }, { "epoch": 0.677942774357117, "grad_norm": 1.3815121085911302, "learning_rate": 4.964485517006102e-06, "loss": 0.6972, "step": 18718 }, { "epoch": 0.6779789931184353, "grad_norm": 1.1669295518617635, "learning_rate": 4.963472061921075e-06, "loss": 0.7068, "step": 18719 }, { "epoch": 0.6780152118797537, "grad_norm": 1.448769792255146, "learning_rate": 4.962458676143097e-06, "loss": 0.7168, "step": 18720 }, { "epoch": 0.6780514306410721, "grad_norm": 1.3717236865000766, "learning_rate": 4.961445359686113e-06, "loss": 0.6677, "step": 18721 }, { "epoch": 0.6780876494023904, "grad_norm": 1.4411485313668986, "learning_rate": 4.960432112564065e-06, "loss": 0.7901, "step": 18722 }, { "epoch": 0.6781238681637088, "grad_norm": 1.4283872571467942, "learning_rate": 4.959418934790904e-06, "loss": 0.7703, "step": 18723 }, { "epoch": 0.6781600869250272, "grad_norm": 1.418161959660364, "learning_rate": 4.958405826380566e-06, "loss": 0.7213, "step": 18724 }, { "epoch": 0.6781963056863455, "grad_norm": 1.3926852400407488, "learning_rate": 4.957392787346991e-06, "loss": 0.7083, "step": 18725 }, { "epoch": 0.6782325244476639, "grad_norm": 1.3963319868692698, "learning_rate": 4.956379817704125e-06, "loss": 0.7651, "step": 18726 }, { "epoch": 0.6782687432089822, "grad_norm": 1.3895114139495264, "learning_rate": 4.955366917465906e-06, "loss": 0.6883, "step": 18727 }, { "epoch": 0.6783049619703007, "grad_norm": 1.3677577872921836, "learning_rate": 4.954354086646269e-06, "loss": 0.7217, "step": 18728 }, { "epoch": 0.678341180731619, "grad_norm": 1.309490275372923, "learning_rate": 4.95334132525915e-06, "loss": 0.6793, "step": 18729 }, { "epoch": 0.6783773994929373, "grad_norm": 1.3227980258887846, "learning_rate": 4.952328633318493e-06, "loss": 0.6819, "step": 18730 }, { "epoch": 0.6784136182542557, "grad_norm": 1.560130689265454, "learning_rate": 4.9513160108382295e-06, "loss": 0.6911, "step": 18731 }, { "epoch": 0.678449837015574, "grad_norm": 1.4529579422915222, "learning_rate": 4.9503034578322895e-06, "loss": 0.7694, "step": 18732 }, { "epoch": 0.6784860557768925, "grad_norm": 1.056320713258045, "learning_rate": 4.949290974314614e-06, "loss": 0.6709, "step": 18733 }, { "epoch": 0.6785222745382108, "grad_norm": 1.3943977919865522, "learning_rate": 4.948278560299135e-06, "loss": 0.6606, "step": 18734 }, { "epoch": 0.6785584932995291, "grad_norm": 1.3521583193371474, "learning_rate": 4.9472662157997795e-06, "loss": 0.6592, "step": 18735 }, { "epoch": 0.6785947120608475, "grad_norm": 0.974200170666698, "learning_rate": 4.9462539408304774e-06, "loss": 0.681, "step": 18736 }, { "epoch": 0.6786309308221659, "grad_norm": 1.3736795674284221, "learning_rate": 4.945241735405166e-06, "loss": 0.7203, "step": 18737 }, { "epoch": 0.6786671495834843, "grad_norm": 1.067109316913245, "learning_rate": 4.94422959953777e-06, "loss": 0.6426, "step": 18738 }, { "epoch": 0.6787033683448026, "grad_norm": 1.291173273176848, "learning_rate": 4.943217533242216e-06, "loss": 0.708, "step": 18739 }, { "epoch": 0.6787395871061209, "grad_norm": 1.4189414767943993, "learning_rate": 4.942205536532429e-06, "loss": 0.7189, "step": 18740 }, { "epoch": 0.6787758058674394, "grad_norm": 1.3736118535265713, "learning_rate": 4.941193609422342e-06, "loss": 0.6761, "step": 18741 }, { "epoch": 0.6788120246287577, "grad_norm": 0.9299818928836949, "learning_rate": 4.940181751925876e-06, "loss": 0.6857, "step": 18742 }, { "epoch": 0.6788482433900761, "grad_norm": 1.285351373185273, "learning_rate": 4.93916996405695e-06, "loss": 0.6218, "step": 18743 }, { "epoch": 0.6788844621513944, "grad_norm": 1.075312405268451, "learning_rate": 4.938158245829497e-06, "loss": 0.6551, "step": 18744 }, { "epoch": 0.6789206809127127, "grad_norm": 1.5975551281757403, "learning_rate": 4.937146597257434e-06, "loss": 0.8177, "step": 18745 }, { "epoch": 0.6789568996740312, "grad_norm": 1.550197553927404, "learning_rate": 4.936135018354683e-06, "loss": 0.6934, "step": 18746 }, { "epoch": 0.6789931184353495, "grad_norm": 1.2918217042648312, "learning_rate": 4.935123509135159e-06, "loss": 0.6912, "step": 18747 }, { "epoch": 0.6790293371966679, "grad_norm": 1.3121393466057756, "learning_rate": 4.93411206961279e-06, "loss": 0.6707, "step": 18748 }, { "epoch": 0.6790655559579862, "grad_norm": 1.2483693263742708, "learning_rate": 4.933100699801491e-06, "loss": 0.6479, "step": 18749 }, { "epoch": 0.6791017747193046, "grad_norm": 1.153526351137992, "learning_rate": 4.932089399715178e-06, "loss": 0.6847, "step": 18750 }, { "epoch": 0.679137993480623, "grad_norm": 1.334098129672378, "learning_rate": 4.931078169367763e-06, "loss": 0.6446, "step": 18751 }, { "epoch": 0.6791742122419413, "grad_norm": 0.9269059618523132, "learning_rate": 4.930067008773172e-06, "loss": 0.665, "step": 18752 }, { "epoch": 0.6792104310032597, "grad_norm": 1.3164264524590592, "learning_rate": 4.9290559179453125e-06, "loss": 0.6557, "step": 18753 }, { "epoch": 0.679246649764578, "grad_norm": 1.3027023586016406, "learning_rate": 4.928044896898096e-06, "loss": 0.7113, "step": 18754 }, { "epoch": 0.6792828685258964, "grad_norm": 1.3725117193989358, "learning_rate": 4.927033945645441e-06, "loss": 0.6056, "step": 18755 }, { "epoch": 0.6793190872872148, "grad_norm": 1.255122002393556, "learning_rate": 4.926023064201257e-06, "loss": 0.6642, "step": 18756 }, { "epoch": 0.6793553060485331, "grad_norm": 0.9753101103222493, "learning_rate": 4.9250122525794555e-06, "loss": 0.6722, "step": 18757 }, { "epoch": 0.6793915248098515, "grad_norm": 1.325526123108142, "learning_rate": 4.92400151079394e-06, "loss": 0.7192, "step": 18758 }, { "epoch": 0.6794277435711699, "grad_norm": 1.2487612748174042, "learning_rate": 4.9229908388586264e-06, "loss": 0.6672, "step": 18759 }, { "epoch": 0.6794639623324882, "grad_norm": 1.176012142060047, "learning_rate": 4.921980236787422e-06, "loss": 0.614, "step": 18760 }, { "epoch": 0.6795001810938066, "grad_norm": 1.0269690126368687, "learning_rate": 4.920969704594229e-06, "loss": 0.6881, "step": 18761 }, { "epoch": 0.6795363998551249, "grad_norm": 1.063084401700843, "learning_rate": 4.9199592422929545e-06, "loss": 0.6967, "step": 18762 }, { "epoch": 0.6795726186164434, "grad_norm": 1.4112874886770241, "learning_rate": 4.918948849897507e-06, "loss": 0.704, "step": 18763 }, { "epoch": 0.6796088373777617, "grad_norm": 0.9900261079894908, "learning_rate": 4.917938527421788e-06, "loss": 0.7067, "step": 18764 }, { "epoch": 0.67964505613908, "grad_norm": 1.3612230190546972, "learning_rate": 4.916928274879696e-06, "loss": 0.7225, "step": 18765 }, { "epoch": 0.6796812749003984, "grad_norm": 1.3456470376771268, "learning_rate": 4.915918092285142e-06, "loss": 0.7043, "step": 18766 }, { "epoch": 0.6797174936617167, "grad_norm": 1.2511561634817254, "learning_rate": 4.914907979652023e-06, "loss": 0.6818, "step": 18767 }, { "epoch": 0.6797537124230352, "grad_norm": 1.2639633981843368, "learning_rate": 4.913897936994236e-06, "loss": 0.6908, "step": 18768 }, { "epoch": 0.6797899311843535, "grad_norm": 1.4775369269739465, "learning_rate": 4.912887964325681e-06, "loss": 0.7186, "step": 18769 }, { "epoch": 0.6798261499456718, "grad_norm": 1.2934151650259804, "learning_rate": 4.91187806166026e-06, "loss": 0.6748, "step": 18770 }, { "epoch": 0.6798623687069902, "grad_norm": 1.3957075385860138, "learning_rate": 4.910868229011868e-06, "loss": 0.7098, "step": 18771 }, { "epoch": 0.6798985874683086, "grad_norm": 1.077727057078084, "learning_rate": 4.9098584663944015e-06, "loss": 0.6961, "step": 18772 }, { "epoch": 0.679934806229627, "grad_norm": 1.3136831766300865, "learning_rate": 4.908848773821751e-06, "loss": 0.7357, "step": 18773 }, { "epoch": 0.6799710249909453, "grad_norm": 1.0356557668472572, "learning_rate": 4.907839151307819e-06, "loss": 0.7701, "step": 18774 }, { "epoch": 0.6800072437522636, "grad_norm": 1.3494956096943995, "learning_rate": 4.906829598866493e-06, "loss": 0.708, "step": 18775 }, { "epoch": 0.680043462513582, "grad_norm": 1.1816422140292053, "learning_rate": 4.905820116511664e-06, "loss": 0.6705, "step": 18776 }, { "epoch": 0.6800796812749004, "grad_norm": 1.2672541282155354, "learning_rate": 4.90481070425723e-06, "loss": 0.681, "step": 18777 }, { "epoch": 0.6801159000362188, "grad_norm": 0.9899594366243197, "learning_rate": 4.903801362117079e-06, "loss": 0.6857, "step": 18778 }, { "epoch": 0.6801521187975371, "grad_norm": 1.6562961625917543, "learning_rate": 4.902792090105097e-06, "loss": 0.7331, "step": 18779 }, { "epoch": 0.6801883375588554, "grad_norm": 1.292279726154107, "learning_rate": 4.901782888235173e-06, "loss": 0.716, "step": 18780 }, { "epoch": 0.6802245563201739, "grad_norm": 1.3235253358379648, "learning_rate": 4.900773756521199e-06, "loss": 0.6811, "step": 18781 }, { "epoch": 0.6802607750814922, "grad_norm": 1.289818601230251, "learning_rate": 4.8997646949770586e-06, "loss": 0.6602, "step": 18782 }, { "epoch": 0.6802969938428106, "grad_norm": 0.9289221377742023, "learning_rate": 4.898755703616638e-06, "loss": 0.6596, "step": 18783 }, { "epoch": 0.6803332126041289, "grad_norm": 0.9917647640260241, "learning_rate": 4.897746782453818e-06, "loss": 0.646, "step": 18784 }, { "epoch": 0.6803694313654473, "grad_norm": 1.2894608662351823, "learning_rate": 4.896737931502487e-06, "loss": 0.7229, "step": 18785 }, { "epoch": 0.6804056501267657, "grad_norm": 1.4719855162259288, "learning_rate": 4.895729150776528e-06, "loss": 0.7043, "step": 18786 }, { "epoch": 0.680441868888084, "grad_norm": 1.3298318161516913, "learning_rate": 4.894720440289816e-06, "loss": 0.6956, "step": 18787 }, { "epoch": 0.6804780876494024, "grad_norm": 1.4145227856774154, "learning_rate": 4.893711800056241e-06, "loss": 0.7101, "step": 18788 }, { "epoch": 0.6805143064107207, "grad_norm": 1.3460812355816953, "learning_rate": 4.8927032300896794e-06, "loss": 0.6843, "step": 18789 }, { "epoch": 0.6805505251720391, "grad_norm": 1.2855150133633881, "learning_rate": 4.891694730404007e-06, "loss": 0.6753, "step": 18790 }, { "epoch": 0.6805867439333575, "grad_norm": 1.3437872925839824, "learning_rate": 4.890686301013101e-06, "loss": 0.7434, "step": 18791 }, { "epoch": 0.6806229626946758, "grad_norm": 1.2805923032822033, "learning_rate": 4.889677941930845e-06, "loss": 0.7101, "step": 18792 }, { "epoch": 0.6806591814559942, "grad_norm": 1.3754276022562533, "learning_rate": 4.888669653171111e-06, "loss": 0.7646, "step": 18793 }, { "epoch": 0.6806954002173126, "grad_norm": 1.4004929709976097, "learning_rate": 4.8876614347477695e-06, "loss": 0.7055, "step": 18794 }, { "epoch": 0.6807316189786309, "grad_norm": 1.339704255357887, "learning_rate": 4.886653286674702e-06, "loss": 0.7164, "step": 18795 }, { "epoch": 0.6807678377399493, "grad_norm": 1.3570845342572624, "learning_rate": 4.885645208965779e-06, "loss": 0.6992, "step": 18796 }, { "epoch": 0.6808040565012676, "grad_norm": 1.061387808912218, "learning_rate": 4.884637201634872e-06, "loss": 0.7521, "step": 18797 }, { "epoch": 0.6808402752625861, "grad_norm": 1.3152758261252842, "learning_rate": 4.883629264695847e-06, "loss": 0.6003, "step": 18798 }, { "epoch": 0.6808764940239044, "grad_norm": 1.3738929453404831, "learning_rate": 4.882621398162583e-06, "loss": 0.626, "step": 18799 }, { "epoch": 0.6809127127852227, "grad_norm": 1.3839796496339427, "learning_rate": 4.881613602048946e-06, "loss": 0.5971, "step": 18800 }, { "epoch": 0.6809489315465411, "grad_norm": 1.2926983175945075, "learning_rate": 4.8806058763688024e-06, "loss": 0.6658, "step": 18801 }, { "epoch": 0.6809851503078594, "grad_norm": 1.5466804502465792, "learning_rate": 4.8795982211360174e-06, "loss": 0.6867, "step": 18802 }, { "epoch": 0.6810213690691779, "grad_norm": 1.304256132334364, "learning_rate": 4.878590636364463e-06, "loss": 0.6157, "step": 18803 }, { "epoch": 0.6810575878304962, "grad_norm": 1.2139663264514338, "learning_rate": 4.877583122068002e-06, "loss": 0.6091, "step": 18804 }, { "epoch": 0.6810938065918145, "grad_norm": 1.0424477802127703, "learning_rate": 4.876575678260493e-06, "loss": 0.6869, "step": 18805 }, { "epoch": 0.6811300253531329, "grad_norm": 1.0365252201726474, "learning_rate": 4.87556830495581e-06, "loss": 0.7659, "step": 18806 }, { "epoch": 0.6811662441144513, "grad_norm": 1.4737644756089934, "learning_rate": 4.874561002167809e-06, "loss": 0.7162, "step": 18807 }, { "epoch": 0.6812024628757697, "grad_norm": 0.9359447766938199, "learning_rate": 4.873553769910353e-06, "loss": 0.6978, "step": 18808 }, { "epoch": 0.681238681637088, "grad_norm": 1.2571058302164315, "learning_rate": 4.872546608197297e-06, "loss": 0.6647, "step": 18809 }, { "epoch": 0.6812749003984063, "grad_norm": 1.3788018335678816, "learning_rate": 4.871539517042509e-06, "loss": 0.7331, "step": 18810 }, { "epoch": 0.6813111191597248, "grad_norm": 1.3680588833598453, "learning_rate": 4.870532496459844e-06, "loss": 0.6812, "step": 18811 }, { "epoch": 0.6813473379210431, "grad_norm": 1.5975985114383295, "learning_rate": 4.869525546463158e-06, "loss": 0.6888, "step": 18812 }, { "epoch": 0.6813835566823615, "grad_norm": 1.2889068776895765, "learning_rate": 4.868518667066307e-06, "loss": 0.6833, "step": 18813 }, { "epoch": 0.6814197754436798, "grad_norm": 1.3953568034491322, "learning_rate": 4.8675118582831496e-06, "loss": 0.7108, "step": 18814 }, { "epoch": 0.6814559942049981, "grad_norm": 1.0056639030017087, "learning_rate": 4.866505120127539e-06, "loss": 0.7178, "step": 18815 }, { "epoch": 0.6814922129663166, "grad_norm": 1.2477506329267418, "learning_rate": 4.865498452613326e-06, "loss": 0.6546, "step": 18816 }, { "epoch": 0.6815284317276349, "grad_norm": 1.2774574545610515, "learning_rate": 4.864491855754369e-06, "loss": 0.6992, "step": 18817 }, { "epoch": 0.6815646504889533, "grad_norm": 1.3116019077075056, "learning_rate": 4.863485329564516e-06, "loss": 0.6956, "step": 18818 }, { "epoch": 0.6816008692502716, "grad_norm": 1.2144155251638573, "learning_rate": 4.8624788740576176e-06, "loss": 0.7545, "step": 18819 }, { "epoch": 0.68163708801159, "grad_norm": 1.420355807904139, "learning_rate": 4.861472489247522e-06, "loss": 0.6713, "step": 18820 }, { "epoch": 0.6816733067729084, "grad_norm": 1.0677741135976593, "learning_rate": 4.860466175148083e-06, "loss": 0.6363, "step": 18821 }, { "epoch": 0.6817095255342267, "grad_norm": 1.2352204174044294, "learning_rate": 4.8594599317731465e-06, "loss": 0.6768, "step": 18822 }, { "epoch": 0.6817457442955451, "grad_norm": 1.3676245065901664, "learning_rate": 4.858453759136557e-06, "loss": 0.7047, "step": 18823 }, { "epoch": 0.6817819630568634, "grad_norm": 1.2934737335831719, "learning_rate": 4.857447657252158e-06, "loss": 0.6573, "step": 18824 }, { "epoch": 0.6818181818181818, "grad_norm": 1.3135343999785194, "learning_rate": 4.856441626133802e-06, "loss": 0.6132, "step": 18825 }, { "epoch": 0.6818544005795002, "grad_norm": 1.4180323176676815, "learning_rate": 4.855435665795329e-06, "loss": 0.674, "step": 18826 }, { "epoch": 0.6818906193408185, "grad_norm": 1.4318849191234664, "learning_rate": 4.854429776250576e-06, "loss": 0.7359, "step": 18827 }, { "epoch": 0.681926838102137, "grad_norm": 1.2582024457030194, "learning_rate": 4.853423957513396e-06, "loss": 0.6577, "step": 18828 }, { "epoch": 0.6819630568634553, "grad_norm": 1.395585860785365, "learning_rate": 4.852418209597625e-06, "loss": 0.738, "step": 18829 }, { "epoch": 0.6819992756247736, "grad_norm": 1.3855029861429373, "learning_rate": 4.851412532517102e-06, "loss": 0.6484, "step": 18830 }, { "epoch": 0.682035494386092, "grad_norm": 1.3804929130999901, "learning_rate": 4.850406926285661e-06, "loss": 0.662, "step": 18831 }, { "epoch": 0.6820717131474103, "grad_norm": 1.3953434758630006, "learning_rate": 4.849401390917151e-06, "loss": 0.6998, "step": 18832 }, { "epoch": 0.6821079319087288, "grad_norm": 1.4617063913786246, "learning_rate": 4.848395926425404e-06, "loss": 0.7574, "step": 18833 }, { "epoch": 0.6821441506700471, "grad_norm": 1.377814894395315, "learning_rate": 4.847390532824254e-06, "loss": 0.6464, "step": 18834 }, { "epoch": 0.6821803694313654, "grad_norm": 1.5643356722057717, "learning_rate": 4.8463852101275354e-06, "loss": 0.7637, "step": 18835 }, { "epoch": 0.6822165881926838, "grad_norm": 1.3158985327592845, "learning_rate": 4.845379958349088e-06, "loss": 0.6608, "step": 18836 }, { "epoch": 0.6822528069540021, "grad_norm": 1.3389353689561299, "learning_rate": 4.844374777502742e-06, "loss": 0.7251, "step": 18837 }, { "epoch": 0.6822890257153206, "grad_norm": 1.3302093821303993, "learning_rate": 4.8433696676023245e-06, "loss": 0.6691, "step": 18838 }, { "epoch": 0.6823252444766389, "grad_norm": 1.3503064054555975, "learning_rate": 4.842364628661676e-06, "loss": 0.7215, "step": 18839 }, { "epoch": 0.6823614632379572, "grad_norm": 1.32560777043396, "learning_rate": 4.841359660694623e-06, "loss": 0.6342, "step": 18840 }, { "epoch": 0.6823976819992756, "grad_norm": 1.3950186540162264, "learning_rate": 4.840354763714991e-06, "loss": 0.691, "step": 18841 }, { "epoch": 0.682433900760594, "grad_norm": 1.2602459442821254, "learning_rate": 4.839349937736609e-06, "loss": 0.5989, "step": 18842 }, { "epoch": 0.6824701195219124, "grad_norm": 1.0418536560074452, "learning_rate": 4.838345182773311e-06, "loss": 0.6444, "step": 18843 }, { "epoch": 0.6825063382832307, "grad_norm": 1.4665431629908137, "learning_rate": 4.837340498838917e-06, "loss": 0.6862, "step": 18844 }, { "epoch": 0.682542557044549, "grad_norm": 1.3965182233589686, "learning_rate": 4.836335885947255e-06, "loss": 0.6835, "step": 18845 }, { "epoch": 0.6825787758058675, "grad_norm": 1.4001376962427519, "learning_rate": 4.835331344112145e-06, "loss": 0.6309, "step": 18846 }, { "epoch": 0.6826149945671858, "grad_norm": 0.9406016389169249, "learning_rate": 4.8343268733474165e-06, "loss": 0.7204, "step": 18847 }, { "epoch": 0.6826512133285042, "grad_norm": 1.311420719274162, "learning_rate": 4.833322473666891e-06, "loss": 0.6439, "step": 18848 }, { "epoch": 0.6826874320898225, "grad_norm": 1.061316518918427, "learning_rate": 4.832318145084383e-06, "loss": 0.6787, "step": 18849 }, { "epoch": 0.6827236508511408, "grad_norm": 1.4060135642607463, "learning_rate": 4.831313887613721e-06, "loss": 0.703, "step": 18850 }, { "epoch": 0.6827598696124593, "grad_norm": 1.2554817403181815, "learning_rate": 4.8303097012687226e-06, "loss": 0.7551, "step": 18851 }, { "epoch": 0.6827960883737776, "grad_norm": 1.3279824222440328, "learning_rate": 4.829305586063206e-06, "loss": 0.6959, "step": 18852 }, { "epoch": 0.682832307135096, "grad_norm": 1.431550232881654, "learning_rate": 4.8283015420109825e-06, "loss": 0.7221, "step": 18853 }, { "epoch": 0.6828685258964143, "grad_norm": 1.1130150600139546, "learning_rate": 4.827297569125879e-06, "loss": 0.7028, "step": 18854 }, { "epoch": 0.6829047446577327, "grad_norm": 1.0316057967523715, "learning_rate": 4.826293667421705e-06, "loss": 0.6535, "step": 18855 }, { "epoch": 0.6829409634190511, "grad_norm": 1.3973027472327797, "learning_rate": 4.825289836912273e-06, "loss": 0.6469, "step": 18856 }, { "epoch": 0.6829771821803694, "grad_norm": 1.4617115624585966, "learning_rate": 4.824286077611404e-06, "loss": 0.6628, "step": 18857 }, { "epoch": 0.6830134009416878, "grad_norm": 0.9704974449981594, "learning_rate": 4.823282389532905e-06, "loss": 0.7187, "step": 18858 }, { "epoch": 0.6830496197030062, "grad_norm": 1.6445528360839354, "learning_rate": 4.82227877269059e-06, "loss": 0.7125, "step": 18859 }, { "epoch": 0.6830858384643245, "grad_norm": 1.229695704563343, "learning_rate": 4.8212752270982685e-06, "loss": 0.6973, "step": 18860 }, { "epoch": 0.6831220572256429, "grad_norm": 1.3519575748849826, "learning_rate": 4.820271752769751e-06, "loss": 0.7252, "step": 18861 }, { "epoch": 0.6831582759869612, "grad_norm": 1.4754416371849572, "learning_rate": 4.819268349718844e-06, "loss": 0.7468, "step": 18862 }, { "epoch": 0.6831944947482796, "grad_norm": 1.3512154296158942, "learning_rate": 4.8182650179593535e-06, "loss": 0.7018, "step": 18863 }, { "epoch": 0.683230713509598, "grad_norm": 1.391112697463283, "learning_rate": 4.817261757505094e-06, "loss": 0.6774, "step": 18864 }, { "epoch": 0.6832669322709163, "grad_norm": 2.094974897131719, "learning_rate": 4.816258568369866e-06, "loss": 0.7674, "step": 18865 }, { "epoch": 0.6833031510322347, "grad_norm": 1.2936376940143306, "learning_rate": 4.815255450567474e-06, "loss": 0.711, "step": 18866 }, { "epoch": 0.683339369793553, "grad_norm": 1.437902284780628, "learning_rate": 4.8142524041117215e-06, "loss": 0.7358, "step": 18867 }, { "epoch": 0.6833755885548715, "grad_norm": 1.308040068338193, "learning_rate": 4.813249429016413e-06, "loss": 0.7463, "step": 18868 }, { "epoch": 0.6834118073161898, "grad_norm": 0.9386573871843055, "learning_rate": 4.812246525295352e-06, "loss": 0.6546, "step": 18869 }, { "epoch": 0.6834480260775081, "grad_norm": 1.3517886038829596, "learning_rate": 4.811243692962333e-06, "loss": 0.7663, "step": 18870 }, { "epoch": 0.6834842448388265, "grad_norm": 0.9810819641831895, "learning_rate": 4.810240932031164e-06, "loss": 0.7112, "step": 18871 }, { "epoch": 0.6835204636001448, "grad_norm": 1.3704446155296217, "learning_rate": 4.809238242515639e-06, "loss": 0.6158, "step": 18872 }, { "epoch": 0.6835566823614633, "grad_norm": 1.477326339364908, "learning_rate": 4.808235624429558e-06, "loss": 0.7108, "step": 18873 }, { "epoch": 0.6835929011227816, "grad_norm": 1.2666323230864036, "learning_rate": 4.807233077786712e-06, "loss": 0.6759, "step": 18874 }, { "epoch": 0.6836291198840999, "grad_norm": 1.1132262662592063, "learning_rate": 4.806230602600907e-06, "loss": 0.7276, "step": 18875 }, { "epoch": 0.6836653386454183, "grad_norm": 0.9937937285033271, "learning_rate": 4.805228198885931e-06, "loss": 0.6312, "step": 18876 }, { "epoch": 0.6837015574067367, "grad_norm": 1.2753572853953348, "learning_rate": 4.804225866655578e-06, "loss": 0.6453, "step": 18877 }, { "epoch": 0.6837377761680551, "grad_norm": 1.3316486171248707, "learning_rate": 4.803223605923641e-06, "loss": 0.6772, "step": 18878 }, { "epoch": 0.6837739949293734, "grad_norm": 1.431007670779082, "learning_rate": 4.8022214167039175e-06, "loss": 0.6323, "step": 18879 }, { "epoch": 0.6838102136906917, "grad_norm": 0.9920594518084059, "learning_rate": 4.801219299010194e-06, "loss": 0.6686, "step": 18880 }, { "epoch": 0.6838464324520102, "grad_norm": 1.4687440845441395, "learning_rate": 4.800217252856256e-06, "loss": 0.6716, "step": 18881 }, { "epoch": 0.6838826512133285, "grad_norm": 1.4632407973818624, "learning_rate": 4.799215278255902e-06, "loss": 0.8072, "step": 18882 }, { "epoch": 0.6839188699746469, "grad_norm": 1.2314930225178324, "learning_rate": 4.798213375222916e-06, "loss": 0.6068, "step": 18883 }, { "epoch": 0.6839550887359652, "grad_norm": 1.3828405751864314, "learning_rate": 4.797211543771084e-06, "loss": 0.6758, "step": 18884 }, { "epoch": 0.6839913074972835, "grad_norm": 1.4369358609158527, "learning_rate": 4.7962097839141895e-06, "loss": 0.7133, "step": 18885 }, { "epoch": 0.684027526258602, "grad_norm": 1.380438029471485, "learning_rate": 4.795208095666025e-06, "loss": 0.6546, "step": 18886 }, { "epoch": 0.6840637450199203, "grad_norm": 1.320650216932416, "learning_rate": 4.79420647904037e-06, "loss": 0.6419, "step": 18887 }, { "epoch": 0.6840999637812387, "grad_norm": 1.100020892736631, "learning_rate": 4.7932049340510045e-06, "loss": 0.6732, "step": 18888 }, { "epoch": 0.684136182542557, "grad_norm": 1.1835761691056934, "learning_rate": 4.792203460711718e-06, "loss": 0.6995, "step": 18889 }, { "epoch": 0.6841724013038754, "grad_norm": 1.029401747140219, "learning_rate": 4.791202059036288e-06, "loss": 0.7133, "step": 18890 }, { "epoch": 0.6842086200651938, "grad_norm": 1.2488593067480018, "learning_rate": 4.790200729038493e-06, "loss": 0.5898, "step": 18891 }, { "epoch": 0.6842448388265121, "grad_norm": 1.4812601460338728, "learning_rate": 4.789199470732111e-06, "loss": 0.6922, "step": 18892 }, { "epoch": 0.6842810575878305, "grad_norm": 1.1845139800048292, "learning_rate": 4.788198284130927e-06, "loss": 0.6813, "step": 18893 }, { "epoch": 0.6843172763491489, "grad_norm": 1.3655252337840138, "learning_rate": 4.787197169248713e-06, "loss": 0.6848, "step": 18894 }, { "epoch": 0.6843534951104672, "grad_norm": 1.329902885953326, "learning_rate": 4.7861961260992474e-06, "loss": 0.7167, "step": 18895 }, { "epoch": 0.6843897138717856, "grad_norm": 1.3707311706467684, "learning_rate": 4.7851951546963e-06, "loss": 0.6943, "step": 18896 }, { "epoch": 0.6844259326331039, "grad_norm": 1.3562869692907926, "learning_rate": 4.7841942550536535e-06, "loss": 0.6585, "step": 18897 }, { "epoch": 0.6844621513944223, "grad_norm": 1.4994834045102052, "learning_rate": 4.783193427185078e-06, "loss": 0.702, "step": 18898 }, { "epoch": 0.6844983701557407, "grad_norm": 1.2928976344808856, "learning_rate": 4.78219267110434e-06, "loss": 0.5831, "step": 18899 }, { "epoch": 0.684534588917059, "grad_norm": 1.681919267543903, "learning_rate": 4.781191986825219e-06, "loss": 0.8038, "step": 18900 }, { "epoch": 0.6845708076783774, "grad_norm": 1.4641890016211754, "learning_rate": 4.780191374361482e-06, "loss": 0.7445, "step": 18901 }, { "epoch": 0.6846070264396957, "grad_norm": 1.3119830152170702, "learning_rate": 4.779190833726899e-06, "loss": 0.7232, "step": 18902 }, { "epoch": 0.6846432452010142, "grad_norm": 1.3692696023559212, "learning_rate": 4.778190364935232e-06, "loss": 0.7483, "step": 18903 }, { "epoch": 0.6846794639623325, "grad_norm": 1.3201555858569303, "learning_rate": 4.77718996800026e-06, "loss": 0.6295, "step": 18904 }, { "epoch": 0.6847156827236508, "grad_norm": 1.8232376842310356, "learning_rate": 4.776189642935742e-06, "loss": 0.6686, "step": 18905 }, { "epoch": 0.6847519014849692, "grad_norm": 1.433251255554938, "learning_rate": 4.775189389755445e-06, "loss": 0.7274, "step": 18906 }, { "epoch": 0.6847881202462875, "grad_norm": 1.3005756260041637, "learning_rate": 4.774189208473129e-06, "loss": 0.6968, "step": 18907 }, { "epoch": 0.684824339007606, "grad_norm": 1.2623539254396623, "learning_rate": 4.7731890991025645e-06, "loss": 0.6807, "step": 18908 }, { "epoch": 0.6848605577689243, "grad_norm": 1.4136492997491172, "learning_rate": 4.772189061657511e-06, "loss": 0.7305, "step": 18909 }, { "epoch": 0.6848967765302426, "grad_norm": 0.9961353478558003, "learning_rate": 4.771189096151726e-06, "loss": 0.6571, "step": 18910 }, { "epoch": 0.684932995291561, "grad_norm": 1.371714554682607, "learning_rate": 4.770189202598977e-06, "loss": 0.6609, "step": 18911 }, { "epoch": 0.6849692140528794, "grad_norm": 1.3547064474897121, "learning_rate": 4.769189381013019e-06, "loss": 0.6803, "step": 18912 }, { "epoch": 0.6850054328141978, "grad_norm": 1.4554368300221587, "learning_rate": 4.768189631407614e-06, "loss": 0.6488, "step": 18913 }, { "epoch": 0.6850416515755161, "grad_norm": 1.4093279202140503, "learning_rate": 4.76718995379651e-06, "loss": 0.7031, "step": 18914 }, { "epoch": 0.6850778703368344, "grad_norm": 1.274198771155813, "learning_rate": 4.766190348193475e-06, "loss": 0.6967, "step": 18915 }, { "epoch": 0.6851140890981529, "grad_norm": 1.3143225182693339, "learning_rate": 4.7651908146122605e-06, "loss": 0.662, "step": 18916 }, { "epoch": 0.6851503078594712, "grad_norm": 1.4147859140262407, "learning_rate": 4.7641913530666195e-06, "loss": 0.7121, "step": 18917 }, { "epoch": 0.6851865266207896, "grad_norm": 1.2938133208413347, "learning_rate": 4.763191963570302e-06, "loss": 0.6915, "step": 18918 }, { "epoch": 0.6852227453821079, "grad_norm": 1.6068689466030062, "learning_rate": 4.762192646137069e-06, "loss": 0.7069, "step": 18919 }, { "epoch": 0.6852589641434262, "grad_norm": 1.2834052916584544, "learning_rate": 4.761193400780667e-06, "loss": 0.7025, "step": 18920 }, { "epoch": 0.6852951829047447, "grad_norm": 1.2764716868100046, "learning_rate": 4.760194227514844e-06, "loss": 0.6523, "step": 18921 }, { "epoch": 0.685331401666063, "grad_norm": 1.316861431934994, "learning_rate": 4.7591951263533555e-06, "loss": 0.7219, "step": 18922 }, { "epoch": 0.6853676204273814, "grad_norm": 1.626559543057263, "learning_rate": 4.758196097309947e-06, "loss": 0.6613, "step": 18923 }, { "epoch": 0.6854038391886997, "grad_norm": 1.0125118057624545, "learning_rate": 4.7571971403983665e-06, "loss": 0.6723, "step": 18924 }, { "epoch": 0.6854400579500181, "grad_norm": 1.3279731803961075, "learning_rate": 4.756198255632356e-06, "loss": 0.6586, "step": 18925 }, { "epoch": 0.6854762767113365, "grad_norm": 1.286925440193906, "learning_rate": 4.755199443025669e-06, "loss": 0.7108, "step": 18926 }, { "epoch": 0.6855124954726548, "grad_norm": 1.3751824986657668, "learning_rate": 4.754200702592046e-06, "loss": 0.682, "step": 18927 }, { "epoch": 0.6855487142339732, "grad_norm": 1.4009671627204252, "learning_rate": 4.753202034345231e-06, "loss": 0.7476, "step": 18928 }, { "epoch": 0.6855849329952916, "grad_norm": 1.2173163321642297, "learning_rate": 4.752203438298962e-06, "loss": 0.6316, "step": 18929 }, { "epoch": 0.6856211517566099, "grad_norm": 1.3510956392013613, "learning_rate": 4.751204914466989e-06, "loss": 0.699, "step": 18930 }, { "epoch": 0.6856573705179283, "grad_norm": 1.3629187379300824, "learning_rate": 4.7502064628630484e-06, "loss": 0.7141, "step": 18931 }, { "epoch": 0.6856935892792466, "grad_norm": 1.435919051679668, "learning_rate": 4.749208083500876e-06, "loss": 0.7196, "step": 18932 }, { "epoch": 0.685729808040565, "grad_norm": 1.3242317384741156, "learning_rate": 4.748209776394217e-06, "loss": 0.6593, "step": 18933 }, { "epoch": 0.6857660268018834, "grad_norm": 1.3988453527368134, "learning_rate": 4.747211541556807e-06, "loss": 0.7151, "step": 18934 }, { "epoch": 0.6858022455632017, "grad_norm": 1.4095538086180697, "learning_rate": 4.746213379002381e-06, "loss": 0.7521, "step": 18935 }, { "epoch": 0.6858384643245201, "grad_norm": 1.3097272888361797, "learning_rate": 4.745215288744673e-06, "loss": 0.6765, "step": 18936 }, { "epoch": 0.6858746830858384, "grad_norm": 1.3160979726395974, "learning_rate": 4.744217270797422e-06, "loss": 0.6889, "step": 18937 }, { "epoch": 0.6859109018471569, "grad_norm": 1.4658419806335767, "learning_rate": 4.7432193251743595e-06, "loss": 0.6876, "step": 18938 }, { "epoch": 0.6859471206084752, "grad_norm": 1.5721236312910603, "learning_rate": 4.7422214518892195e-06, "loss": 0.6623, "step": 18939 }, { "epoch": 0.6859833393697935, "grad_norm": 1.5013672463556116, "learning_rate": 4.741223650955727e-06, "loss": 0.6293, "step": 18940 }, { "epoch": 0.6860195581311119, "grad_norm": 1.3112097860068672, "learning_rate": 4.740225922387623e-06, "loss": 0.7256, "step": 18941 }, { "epoch": 0.6860557768924302, "grad_norm": 0.9674201893464459, "learning_rate": 4.739228266198631e-06, "loss": 0.6673, "step": 18942 }, { "epoch": 0.6860919956537487, "grad_norm": 1.2686903317930232, "learning_rate": 4.738230682402477e-06, "loss": 0.6317, "step": 18943 }, { "epoch": 0.686128214415067, "grad_norm": 1.3276548594827455, "learning_rate": 4.737233171012896e-06, "loss": 0.8038, "step": 18944 }, { "epoch": 0.6861644331763853, "grad_norm": 1.2759493470044874, "learning_rate": 4.736235732043611e-06, "loss": 0.625, "step": 18945 }, { "epoch": 0.6862006519377037, "grad_norm": 1.4196199544758341, "learning_rate": 4.735238365508348e-06, "loss": 0.6182, "step": 18946 }, { "epoch": 0.6862368706990221, "grad_norm": 1.1718308753408753, "learning_rate": 4.734241071420826e-06, "loss": 0.6698, "step": 18947 }, { "epoch": 0.6862730894603405, "grad_norm": 1.267731872882773, "learning_rate": 4.733243849794779e-06, "loss": 0.6983, "step": 18948 }, { "epoch": 0.6863093082216588, "grad_norm": 1.4462273473148395, "learning_rate": 4.7322467006439235e-06, "loss": 0.7083, "step": 18949 }, { "epoch": 0.6863455269829771, "grad_norm": 1.2798932665498803, "learning_rate": 4.731249623981979e-06, "loss": 0.6581, "step": 18950 }, { "epoch": 0.6863817457442956, "grad_norm": 1.3111491475195107, "learning_rate": 4.73025261982267e-06, "loss": 0.7419, "step": 18951 }, { "epoch": 0.6864179645056139, "grad_norm": 1.063012963255064, "learning_rate": 4.729255688179719e-06, "loss": 0.6792, "step": 18952 }, { "epoch": 0.6864541832669323, "grad_norm": 1.3668392698402094, "learning_rate": 4.72825882906684e-06, "loss": 0.6573, "step": 18953 }, { "epoch": 0.6864904020282506, "grad_norm": 1.313109366414877, "learning_rate": 4.727262042497747e-06, "loss": 0.6595, "step": 18954 }, { "epoch": 0.6865266207895689, "grad_norm": 1.2713012460888737, "learning_rate": 4.726265328486166e-06, "loss": 0.691, "step": 18955 }, { "epoch": 0.6865628395508874, "grad_norm": 1.1066343841372972, "learning_rate": 4.7252686870458075e-06, "loss": 0.7302, "step": 18956 }, { "epoch": 0.6865990583122057, "grad_norm": 1.354051549921367, "learning_rate": 4.724272118190387e-06, "loss": 0.7098, "step": 18957 }, { "epoch": 0.6866352770735241, "grad_norm": 1.3285116621562356, "learning_rate": 4.723275621933614e-06, "loss": 0.6631, "step": 18958 }, { "epoch": 0.6866714958348424, "grad_norm": 1.301276165084677, "learning_rate": 4.722279198289209e-06, "loss": 0.6567, "step": 18959 }, { "epoch": 0.6867077145961608, "grad_norm": 1.2311887731009135, "learning_rate": 4.72128284727088e-06, "loss": 0.6344, "step": 18960 }, { "epoch": 0.6867439333574792, "grad_norm": 1.3494781845799086, "learning_rate": 4.720286568892331e-06, "loss": 0.698, "step": 18961 }, { "epoch": 0.6867801521187975, "grad_norm": 1.2658915520489398, "learning_rate": 4.7192903631672845e-06, "loss": 0.6975, "step": 18962 }, { "epoch": 0.6868163708801159, "grad_norm": 1.314930741911039, "learning_rate": 4.7182942301094416e-06, "loss": 0.7255, "step": 18963 }, { "epoch": 0.6868525896414343, "grad_norm": 1.0230434492110472, "learning_rate": 4.717298169732512e-06, "loss": 0.7257, "step": 18964 }, { "epoch": 0.6868888084027526, "grad_norm": 1.4310548359332107, "learning_rate": 4.716302182050195e-06, "loss": 0.6941, "step": 18965 }, { "epoch": 0.686925027164071, "grad_norm": 1.483250195577886, "learning_rate": 4.71530626707621e-06, "loss": 0.7895, "step": 18966 }, { "epoch": 0.6869612459253893, "grad_norm": 1.2727188769656643, "learning_rate": 4.714310424824251e-06, "loss": 0.5971, "step": 18967 }, { "epoch": 0.6869974646867077, "grad_norm": 1.0363041285459778, "learning_rate": 4.713314655308027e-06, "loss": 0.6226, "step": 18968 }, { "epoch": 0.6870336834480261, "grad_norm": 0.9702446755510695, "learning_rate": 4.712318958541235e-06, "loss": 0.7657, "step": 18969 }, { "epoch": 0.6870699022093444, "grad_norm": 1.315189816893038, "learning_rate": 4.711323334537582e-06, "loss": 0.6712, "step": 18970 }, { "epoch": 0.6871061209706628, "grad_norm": 0.9800735446894931, "learning_rate": 4.710327783310769e-06, "loss": 0.705, "step": 18971 }, { "epoch": 0.6871423397319811, "grad_norm": 1.3672569369666787, "learning_rate": 4.709332304874489e-06, "loss": 0.7302, "step": 18972 }, { "epoch": 0.6871785584932996, "grad_norm": 1.2616368132816007, "learning_rate": 4.708336899242449e-06, "loss": 0.6724, "step": 18973 }, { "epoch": 0.6872147772546179, "grad_norm": 1.4082687867896713, "learning_rate": 4.707341566428343e-06, "loss": 0.6819, "step": 18974 }, { "epoch": 0.6872509960159362, "grad_norm": 1.0436878446375024, "learning_rate": 4.706346306445868e-06, "loss": 0.6966, "step": 18975 }, { "epoch": 0.6872872147772546, "grad_norm": 1.379874120065317, "learning_rate": 4.705351119308716e-06, "loss": 0.7179, "step": 18976 }, { "epoch": 0.687323433538573, "grad_norm": 1.2975907487058256, "learning_rate": 4.704356005030587e-06, "loss": 0.5995, "step": 18977 }, { "epoch": 0.6873596522998914, "grad_norm": 1.4006996731532337, "learning_rate": 4.7033609636251735e-06, "loss": 0.6996, "step": 18978 }, { "epoch": 0.6873958710612097, "grad_norm": 1.0869115117609507, "learning_rate": 4.702365995106166e-06, "loss": 0.6994, "step": 18979 }, { "epoch": 0.687432089822528, "grad_norm": 1.3758095842332374, "learning_rate": 4.701371099487256e-06, "loss": 0.6725, "step": 18980 }, { "epoch": 0.6874683085838464, "grad_norm": 1.3101967985012217, "learning_rate": 4.700376276782137e-06, "loss": 0.7549, "step": 18981 }, { "epoch": 0.6875045273451648, "grad_norm": 1.1567127710865106, "learning_rate": 4.6993815270044985e-06, "loss": 0.7599, "step": 18982 }, { "epoch": 0.6875407461064832, "grad_norm": 1.3682046403660308, "learning_rate": 4.698386850168023e-06, "loss": 0.7441, "step": 18983 }, { "epoch": 0.6875769648678015, "grad_norm": 1.3364214069544347, "learning_rate": 4.6973922462864066e-06, "loss": 0.7017, "step": 18984 }, { "epoch": 0.6876131836291199, "grad_norm": 1.3924277212344152, "learning_rate": 4.696397715373333e-06, "loss": 0.6771, "step": 18985 }, { "epoch": 0.6876494023904383, "grad_norm": 1.3890277008304035, "learning_rate": 4.695403257442486e-06, "loss": 0.6788, "step": 18986 }, { "epoch": 0.6876856211517566, "grad_norm": 1.3802026727548318, "learning_rate": 4.6944088725075475e-06, "loss": 0.6903, "step": 18987 }, { "epoch": 0.687721839913075, "grad_norm": 1.267108178419525, "learning_rate": 4.69341456058221e-06, "loss": 0.7044, "step": 18988 }, { "epoch": 0.6877580586743933, "grad_norm": 0.8976795035262952, "learning_rate": 4.69242032168015e-06, "loss": 0.6717, "step": 18989 }, { "epoch": 0.6877942774357118, "grad_norm": 1.4061699420604754, "learning_rate": 4.69142615581505e-06, "loss": 0.7169, "step": 18990 }, { "epoch": 0.6878304961970301, "grad_norm": 1.4668040700804579, "learning_rate": 4.690432063000586e-06, "loss": 0.7016, "step": 18991 }, { "epoch": 0.6878667149583484, "grad_norm": 1.0325972021115941, "learning_rate": 4.689438043250447e-06, "loss": 0.7192, "step": 18992 }, { "epoch": 0.6879029337196668, "grad_norm": 1.346044717129235, "learning_rate": 4.688444096578307e-06, "loss": 0.7862, "step": 18993 }, { "epoch": 0.6879391524809851, "grad_norm": 1.342101086220587, "learning_rate": 4.687450222997839e-06, "loss": 0.6666, "step": 18994 }, { "epoch": 0.6879753712423036, "grad_norm": 1.0447390714788582, "learning_rate": 4.686456422522728e-06, "loss": 0.7013, "step": 18995 }, { "epoch": 0.6880115900036219, "grad_norm": 1.3267420103608145, "learning_rate": 4.685462695166645e-06, "loss": 0.6857, "step": 18996 }, { "epoch": 0.6880478087649402, "grad_norm": 1.3208888015750955, "learning_rate": 4.684469040943266e-06, "loss": 0.637, "step": 18997 }, { "epoch": 0.6880840275262586, "grad_norm": 1.928055412340509, "learning_rate": 4.683475459866259e-06, "loss": 0.6913, "step": 18998 }, { "epoch": 0.688120246287577, "grad_norm": 1.4370459106297238, "learning_rate": 4.682481951949306e-06, "loss": 0.6787, "step": 18999 }, { "epoch": 0.6881564650488954, "grad_norm": 1.5455268569472176, "learning_rate": 4.681488517206072e-06, "loss": 0.7089, "step": 19000 }, { "epoch": 0.6881926838102137, "grad_norm": 1.366879838658888, "learning_rate": 4.680495155650231e-06, "loss": 0.6875, "step": 19001 }, { "epoch": 0.688228902571532, "grad_norm": 1.3435833946895441, "learning_rate": 4.679501867295446e-06, "loss": 0.6843, "step": 19002 }, { "epoch": 0.6882651213328504, "grad_norm": 1.3613919622156843, "learning_rate": 4.678508652155395e-06, "loss": 0.6804, "step": 19003 }, { "epoch": 0.6883013400941688, "grad_norm": 1.2633816836897818, "learning_rate": 4.67751551024374e-06, "loss": 0.665, "step": 19004 }, { "epoch": 0.6883375588554872, "grad_norm": 1.3754388504435975, "learning_rate": 4.676522441574145e-06, "loss": 0.6996, "step": 19005 }, { "epoch": 0.6883737776168055, "grad_norm": 1.3934261774690444, "learning_rate": 4.675529446160282e-06, "loss": 0.7355, "step": 19006 }, { "epoch": 0.6884099963781238, "grad_norm": 1.5028885072912432, "learning_rate": 4.674536524015812e-06, "loss": 0.7371, "step": 19007 }, { "epoch": 0.6884462151394423, "grad_norm": 1.3616740712120292, "learning_rate": 4.673543675154399e-06, "loss": 0.7292, "step": 19008 }, { "epoch": 0.6884824339007606, "grad_norm": 1.102586323957993, "learning_rate": 4.6725508995897025e-06, "loss": 0.6602, "step": 19009 }, { "epoch": 0.688518652662079, "grad_norm": 0.957497089488711, "learning_rate": 4.6715581973353905e-06, "loss": 0.6987, "step": 19010 }, { "epoch": 0.6885548714233973, "grad_norm": 1.5143040266086076, "learning_rate": 4.670565568405119e-06, "loss": 0.6569, "step": 19011 }, { "epoch": 0.6885910901847156, "grad_norm": 1.3981784482053916, "learning_rate": 4.669573012812545e-06, "loss": 0.5927, "step": 19012 }, { "epoch": 0.6886273089460341, "grad_norm": 1.3882589980560196, "learning_rate": 4.668580530571333e-06, "loss": 0.7158, "step": 19013 }, { "epoch": 0.6886635277073524, "grad_norm": 1.1704701360333831, "learning_rate": 4.667588121695138e-06, "loss": 0.6519, "step": 19014 }, { "epoch": 0.6886997464686708, "grad_norm": 1.314056187052083, "learning_rate": 4.6665957861976154e-06, "loss": 0.6409, "step": 19015 }, { "epoch": 0.6887359652299891, "grad_norm": 1.3275186697361867, "learning_rate": 4.665603524092418e-06, "loss": 0.6445, "step": 19016 }, { "epoch": 0.6887721839913075, "grad_norm": 1.0868424270844745, "learning_rate": 4.664611335393207e-06, "loss": 0.7006, "step": 19017 }, { "epoch": 0.6888084027526259, "grad_norm": 1.2947695543278703, "learning_rate": 4.663619220113631e-06, "loss": 0.6049, "step": 19018 }, { "epoch": 0.6888446215139442, "grad_norm": 1.652895085809765, "learning_rate": 4.662627178267345e-06, "loss": 0.7162, "step": 19019 }, { "epoch": 0.6888808402752626, "grad_norm": 1.403077483487715, "learning_rate": 4.661635209867994e-06, "loss": 0.7348, "step": 19020 }, { "epoch": 0.688917059036581, "grad_norm": 1.0417081924943135, "learning_rate": 4.660643314929238e-06, "loss": 0.6833, "step": 19021 }, { "epoch": 0.6889532777978993, "grad_norm": 1.3201376796418156, "learning_rate": 4.659651493464721e-06, "loss": 0.686, "step": 19022 }, { "epoch": 0.6889894965592177, "grad_norm": 1.493277875356203, "learning_rate": 4.6586597454880874e-06, "loss": 0.71, "step": 19023 }, { "epoch": 0.689025715320536, "grad_norm": 1.2831173255547366, "learning_rate": 4.657668071012994e-06, "loss": 0.6783, "step": 19024 }, { "epoch": 0.6890619340818545, "grad_norm": 1.3478578472004752, "learning_rate": 4.656676470053081e-06, "loss": 0.6975, "step": 19025 }, { "epoch": 0.6890981528431728, "grad_norm": 1.286967415137178, "learning_rate": 4.655684942621995e-06, "loss": 0.7168, "step": 19026 }, { "epoch": 0.6891343716044911, "grad_norm": 1.2265717831035592, "learning_rate": 4.654693488733377e-06, "loss": 0.5737, "step": 19027 }, { "epoch": 0.6891705903658095, "grad_norm": 1.2940546176735284, "learning_rate": 4.6537021084008774e-06, "loss": 0.6936, "step": 19028 }, { "epoch": 0.6892068091271278, "grad_norm": 1.3236918814905094, "learning_rate": 4.652710801638134e-06, "loss": 0.7605, "step": 19029 }, { "epoch": 0.6892430278884463, "grad_norm": 1.4519275049561642, "learning_rate": 4.651719568458789e-06, "loss": 0.7062, "step": 19030 }, { "epoch": 0.6892792466497646, "grad_norm": 1.415127504637417, "learning_rate": 4.6507284088764774e-06, "loss": 0.6693, "step": 19031 }, { "epoch": 0.6893154654110829, "grad_norm": 1.441653054773516, "learning_rate": 4.649737322904847e-06, "loss": 0.6865, "step": 19032 }, { "epoch": 0.6893516841724013, "grad_norm": 1.355819932863438, "learning_rate": 4.648746310557533e-06, "loss": 0.688, "step": 19033 }, { "epoch": 0.6893879029337197, "grad_norm": 1.2697439485048212, "learning_rate": 4.647755371848167e-06, "loss": 0.6829, "step": 19034 }, { "epoch": 0.6894241216950381, "grad_norm": 1.238405087593363, "learning_rate": 4.646764506790395e-06, "loss": 0.624, "step": 19035 }, { "epoch": 0.6894603404563564, "grad_norm": 1.4573084641670786, "learning_rate": 4.645773715397846e-06, "loss": 0.7444, "step": 19036 }, { "epoch": 0.6894965592176747, "grad_norm": 1.0840341379367724, "learning_rate": 4.644782997684156e-06, "loss": 0.6974, "step": 19037 }, { "epoch": 0.6895327779789931, "grad_norm": 1.4490392013331976, "learning_rate": 4.643792353662954e-06, "loss": 0.682, "step": 19038 }, { "epoch": 0.6895689967403115, "grad_norm": 1.0295467988282156, "learning_rate": 4.642801783347879e-06, "loss": 0.6957, "step": 19039 }, { "epoch": 0.6896052155016299, "grad_norm": 1.2358416064071593, "learning_rate": 4.641811286752559e-06, "loss": 0.6462, "step": 19040 }, { "epoch": 0.6896414342629482, "grad_norm": 1.0464185832750685, "learning_rate": 4.640820863890623e-06, "loss": 0.6454, "step": 19041 }, { "epoch": 0.6896776530242665, "grad_norm": 0.9125862290527758, "learning_rate": 4.639830514775697e-06, "loss": 0.6844, "step": 19042 }, { "epoch": 0.689713871785585, "grad_norm": 1.3369700849901038, "learning_rate": 4.638840239421418e-06, "loss": 0.7472, "step": 19043 }, { "epoch": 0.6897500905469033, "grad_norm": 1.3781157434892217, "learning_rate": 4.637850037841407e-06, "loss": 0.6606, "step": 19044 }, { "epoch": 0.6897863093082217, "grad_norm": 1.544269621807307, "learning_rate": 4.636859910049288e-06, "loss": 0.6497, "step": 19045 }, { "epoch": 0.68982252806954, "grad_norm": 1.394738630027046, "learning_rate": 4.6358698560586925e-06, "loss": 0.7039, "step": 19046 }, { "epoch": 0.6898587468308583, "grad_norm": 1.0781075193384821, "learning_rate": 4.634879875883241e-06, "loss": 0.6486, "step": 19047 }, { "epoch": 0.6898949655921768, "grad_norm": 1.3148353865386382, "learning_rate": 4.6338899695365566e-06, "loss": 0.6702, "step": 19048 }, { "epoch": 0.6899311843534951, "grad_norm": 1.2740828174962138, "learning_rate": 4.632900137032258e-06, "loss": 0.6512, "step": 19049 }, { "epoch": 0.6899674031148135, "grad_norm": 1.029630343126321, "learning_rate": 4.6319103783839726e-06, "loss": 0.6932, "step": 19050 }, { "epoch": 0.6900036218761318, "grad_norm": 1.347344854562991, "learning_rate": 4.630920693605317e-06, "loss": 0.7659, "step": 19051 }, { "epoch": 0.6900398406374502, "grad_norm": 1.2786401985171194, "learning_rate": 4.629931082709912e-06, "loss": 0.636, "step": 19052 }, { "epoch": 0.6900760593987686, "grad_norm": 1.6794445727133633, "learning_rate": 4.628941545711369e-06, "loss": 0.7162, "step": 19053 }, { "epoch": 0.6901122781600869, "grad_norm": 1.3280627909283964, "learning_rate": 4.627952082623313e-06, "loss": 0.551, "step": 19054 }, { "epoch": 0.6901484969214053, "grad_norm": 1.3156888978271701, "learning_rate": 4.626962693459357e-06, "loss": 0.7063, "step": 19055 }, { "epoch": 0.6901847156827237, "grad_norm": 1.317295162036515, "learning_rate": 4.625973378233112e-06, "loss": 0.6941, "step": 19056 }, { "epoch": 0.690220934444042, "grad_norm": 1.29386677989522, "learning_rate": 4.624984136958199e-06, "loss": 0.7113, "step": 19057 }, { "epoch": 0.6902571532053604, "grad_norm": 1.373075369624692, "learning_rate": 4.623994969648227e-06, "loss": 0.6611, "step": 19058 }, { "epoch": 0.6902933719666787, "grad_norm": 1.429955887877932, "learning_rate": 4.623005876316808e-06, "loss": 0.7163, "step": 19059 }, { "epoch": 0.6903295907279972, "grad_norm": 1.2873511890409886, "learning_rate": 4.622016856977549e-06, "loss": 0.5917, "step": 19060 }, { "epoch": 0.6903658094893155, "grad_norm": 1.4315256876689308, "learning_rate": 4.621027911644068e-06, "loss": 0.6988, "step": 19061 }, { "epoch": 0.6904020282506338, "grad_norm": 1.311903559035178, "learning_rate": 4.620039040329969e-06, "loss": 0.7597, "step": 19062 }, { "epoch": 0.6904382470119522, "grad_norm": 1.1736211038793571, "learning_rate": 4.619050243048859e-06, "loss": 0.6282, "step": 19063 }, { "epoch": 0.6904744657732705, "grad_norm": 1.3372308426831034, "learning_rate": 4.618061519814343e-06, "loss": 0.7455, "step": 19064 }, { "epoch": 0.690510684534589, "grad_norm": 1.2615707475998807, "learning_rate": 4.617072870640034e-06, "loss": 0.7354, "step": 19065 }, { "epoch": 0.6905469032959073, "grad_norm": 1.549355109374863, "learning_rate": 4.61608429553953e-06, "loss": 0.6825, "step": 19066 }, { "epoch": 0.6905831220572256, "grad_norm": 1.4790893239102187, "learning_rate": 4.615095794526435e-06, "loss": 0.7764, "step": 19067 }, { "epoch": 0.690619340818544, "grad_norm": 1.0191230155294173, "learning_rate": 4.614107367614357e-06, "loss": 0.7136, "step": 19068 }, { "epoch": 0.6906555595798624, "grad_norm": 0.9236380088987279, "learning_rate": 4.613119014816893e-06, "loss": 0.6991, "step": 19069 }, { "epoch": 0.6906917783411808, "grad_norm": 1.3097706029258673, "learning_rate": 4.6121307361476454e-06, "loss": 0.6172, "step": 19070 }, { "epoch": 0.6907279971024991, "grad_norm": 1.3717244611741366, "learning_rate": 4.611142531620213e-06, "loss": 0.6957, "step": 19071 }, { "epoch": 0.6907642158638174, "grad_norm": 0.9951714444975135, "learning_rate": 4.610154401248194e-06, "loss": 0.68, "step": 19072 }, { "epoch": 0.6908004346251359, "grad_norm": 1.4663568368724929, "learning_rate": 4.6091663450451876e-06, "loss": 0.8036, "step": 19073 }, { "epoch": 0.6908366533864542, "grad_norm": 0.9795741563830459, "learning_rate": 4.608178363024783e-06, "loss": 0.6914, "step": 19074 }, { "epoch": 0.6908728721477726, "grad_norm": 1.3504355195825377, "learning_rate": 4.607190455200588e-06, "loss": 0.622, "step": 19075 }, { "epoch": 0.6909090909090909, "grad_norm": 1.439798593796196, "learning_rate": 4.60620262158619e-06, "loss": 0.7763, "step": 19076 }, { "epoch": 0.6909453096704092, "grad_norm": 1.3118807092299773, "learning_rate": 4.605214862195179e-06, "loss": 0.7337, "step": 19077 }, { "epoch": 0.6909815284317277, "grad_norm": 1.4619358165602927, "learning_rate": 4.604227177041156e-06, "loss": 0.7624, "step": 19078 }, { "epoch": 0.691017747193046, "grad_norm": 1.4345371698247606, "learning_rate": 4.603239566137708e-06, "loss": 0.6251, "step": 19079 }, { "epoch": 0.6910539659543644, "grad_norm": 1.4147974996732542, "learning_rate": 4.602252029498427e-06, "loss": 0.6504, "step": 19080 }, { "epoch": 0.6910901847156827, "grad_norm": 1.4091538633754785, "learning_rate": 4.601264567136895e-06, "loss": 0.7251, "step": 19081 }, { "epoch": 0.691126403477001, "grad_norm": 1.5107313290127489, "learning_rate": 4.600277179066711e-06, "loss": 0.7453, "step": 19082 }, { "epoch": 0.6911626222383195, "grad_norm": 1.325528277998917, "learning_rate": 4.5992898653014575e-06, "loss": 0.6783, "step": 19083 }, { "epoch": 0.6911988409996378, "grad_norm": 1.2624775734942013, "learning_rate": 4.598302625854722e-06, "loss": 0.7415, "step": 19084 }, { "epoch": 0.6912350597609562, "grad_norm": 1.2863049453502662, "learning_rate": 4.597315460740084e-06, "loss": 0.6448, "step": 19085 }, { "epoch": 0.6912712785222745, "grad_norm": 1.2733072462194703, "learning_rate": 4.596328369971136e-06, "loss": 0.7341, "step": 19086 }, { "epoch": 0.6913074972835929, "grad_norm": 0.9923909954660598, "learning_rate": 4.595341353561458e-06, "loss": 0.6697, "step": 19087 }, { "epoch": 0.6913437160449113, "grad_norm": 1.3826985231399926, "learning_rate": 4.594354411524627e-06, "loss": 0.7101, "step": 19088 }, { "epoch": 0.6913799348062296, "grad_norm": 1.3832227804085038, "learning_rate": 4.593367543874234e-06, "loss": 0.715, "step": 19089 }, { "epoch": 0.691416153567548, "grad_norm": 1.3442041480749984, "learning_rate": 4.592380750623855e-06, "loss": 0.6443, "step": 19090 }, { "epoch": 0.6914523723288664, "grad_norm": 1.433407551837433, "learning_rate": 4.591394031787068e-06, "loss": 0.6956, "step": 19091 }, { "epoch": 0.6914885910901847, "grad_norm": 1.29889674399821, "learning_rate": 4.590407387377447e-06, "loss": 0.6809, "step": 19092 }, { "epoch": 0.6915248098515031, "grad_norm": 1.4065869113208134, "learning_rate": 4.589420817408578e-06, "loss": 0.7291, "step": 19093 }, { "epoch": 0.6915610286128214, "grad_norm": 1.3686484460302264, "learning_rate": 4.588434321894033e-06, "loss": 0.6775, "step": 19094 }, { "epoch": 0.6915972473741399, "grad_norm": 1.425795029874081, "learning_rate": 4.587447900847386e-06, "loss": 0.7897, "step": 19095 }, { "epoch": 0.6916334661354582, "grad_norm": 1.2977500973456915, "learning_rate": 4.586461554282209e-06, "loss": 0.6637, "step": 19096 }, { "epoch": 0.6916696848967765, "grad_norm": 1.2246223445978215, "learning_rate": 4.58547528221208e-06, "loss": 0.6884, "step": 19097 }, { "epoch": 0.6917059036580949, "grad_norm": 1.5200508578877565, "learning_rate": 4.584489084650569e-06, "loss": 0.7592, "step": 19098 }, { "epoch": 0.6917421224194132, "grad_norm": 1.1923393109577265, "learning_rate": 4.583502961611245e-06, "loss": 0.6267, "step": 19099 }, { "epoch": 0.6917783411807317, "grad_norm": 1.333160857612817, "learning_rate": 4.582516913107682e-06, "loss": 0.6511, "step": 19100 }, { "epoch": 0.69181455994205, "grad_norm": 1.0082871397183242, "learning_rate": 4.581530939153446e-06, "loss": 0.6737, "step": 19101 }, { "epoch": 0.6918507787033683, "grad_norm": 1.3754930926080688, "learning_rate": 4.580545039762105e-06, "loss": 0.6572, "step": 19102 }, { "epoch": 0.6918869974646867, "grad_norm": 1.4291522204377523, "learning_rate": 4.579559214947223e-06, "loss": 0.686, "step": 19103 }, { "epoch": 0.6919232162260051, "grad_norm": 1.376523374224456, "learning_rate": 4.578573464722373e-06, "loss": 0.7158, "step": 19104 }, { "epoch": 0.6919594349873235, "grad_norm": 0.9775728179151116, "learning_rate": 4.577587789101117e-06, "loss": 0.7631, "step": 19105 }, { "epoch": 0.6919956537486418, "grad_norm": 1.2700913807492284, "learning_rate": 4.576602188097016e-06, "loss": 0.5995, "step": 19106 }, { "epoch": 0.6920318725099601, "grad_norm": 1.2667554814765623, "learning_rate": 4.5756166617236305e-06, "loss": 0.6987, "step": 19107 }, { "epoch": 0.6920680912712786, "grad_norm": 1.7439953603621956, "learning_rate": 4.57463120999453e-06, "loss": 0.7127, "step": 19108 }, { "epoch": 0.6921043100325969, "grad_norm": 0.9585583198294639, "learning_rate": 4.573645832923271e-06, "loss": 0.6493, "step": 19109 }, { "epoch": 0.6921405287939153, "grad_norm": 1.499821132560315, "learning_rate": 4.5726605305234105e-06, "loss": 0.6659, "step": 19110 }, { "epoch": 0.6921767475552336, "grad_norm": 1.3904384061912682, "learning_rate": 4.5716753028085135e-06, "loss": 0.6822, "step": 19111 }, { "epoch": 0.6922129663165519, "grad_norm": 1.4503875142666933, "learning_rate": 4.570690149792134e-06, "loss": 0.7283, "step": 19112 }, { "epoch": 0.6922491850778704, "grad_norm": 1.367934032709863, "learning_rate": 4.569705071487827e-06, "loss": 0.6878, "step": 19113 }, { "epoch": 0.6922854038391887, "grad_norm": 1.4448881577898587, "learning_rate": 4.5687200679091484e-06, "loss": 0.7041, "step": 19114 }, { "epoch": 0.6923216226005071, "grad_norm": 1.4461271683506616, "learning_rate": 4.567735139069657e-06, "loss": 0.6518, "step": 19115 }, { "epoch": 0.6923578413618254, "grad_norm": 1.1627075254746306, "learning_rate": 4.566750284982903e-06, "loss": 0.7055, "step": 19116 }, { "epoch": 0.6923940601231438, "grad_norm": 1.0702296526458903, "learning_rate": 4.565765505662435e-06, "loss": 0.6984, "step": 19117 }, { "epoch": 0.6924302788844622, "grad_norm": 1.5282475662894892, "learning_rate": 4.564780801121813e-06, "loss": 0.7298, "step": 19118 }, { "epoch": 0.6924664976457805, "grad_norm": 1.3907584383518423, "learning_rate": 4.5637961713745826e-06, "loss": 0.6704, "step": 19119 }, { "epoch": 0.6925027164070989, "grad_norm": 1.4409151041693347, "learning_rate": 4.562811616434294e-06, "loss": 0.7342, "step": 19120 }, { "epoch": 0.6925389351684172, "grad_norm": 1.5245069540241476, "learning_rate": 4.561827136314491e-06, "loss": 0.7358, "step": 19121 }, { "epoch": 0.6925751539297356, "grad_norm": 1.2084627096786185, "learning_rate": 4.5608427310287285e-06, "loss": 0.6553, "step": 19122 }, { "epoch": 0.692611372691054, "grad_norm": 0.9708452794642716, "learning_rate": 4.55985840059055e-06, "loss": 0.7592, "step": 19123 }, { "epoch": 0.6926475914523723, "grad_norm": 1.3707942431742988, "learning_rate": 4.5588741450134985e-06, "loss": 0.7461, "step": 19124 }, { "epoch": 0.6926838102136907, "grad_norm": 1.013535161810221, "learning_rate": 4.557889964311117e-06, "loss": 0.7128, "step": 19125 }, { "epoch": 0.6927200289750091, "grad_norm": 0.9898107180130732, "learning_rate": 4.556905858496956e-06, "loss": 0.6733, "step": 19126 }, { "epoch": 0.6927562477363274, "grad_norm": 1.2612686372978235, "learning_rate": 4.555921827584553e-06, "loss": 0.746, "step": 19127 }, { "epoch": 0.6927924664976458, "grad_norm": 1.3166305534131728, "learning_rate": 4.5549378715874445e-06, "loss": 0.7099, "step": 19128 }, { "epoch": 0.6928286852589641, "grad_norm": 1.2674854418075585, "learning_rate": 4.55395399051918e-06, "loss": 0.6046, "step": 19129 }, { "epoch": 0.6928649040202826, "grad_norm": 1.2840595239907444, "learning_rate": 4.552970184393292e-06, "loss": 0.6754, "step": 19130 }, { "epoch": 0.6929011227816009, "grad_norm": 1.0651585604789984, "learning_rate": 4.5519864532233224e-06, "loss": 0.7139, "step": 19131 }, { "epoch": 0.6929373415429192, "grad_norm": 0.9822184762012178, "learning_rate": 4.551002797022802e-06, "loss": 0.6973, "step": 19132 }, { "epoch": 0.6929735603042376, "grad_norm": 1.2874689301832067, "learning_rate": 4.550019215805275e-06, "loss": 0.7109, "step": 19133 }, { "epoch": 0.6930097790655559, "grad_norm": 1.2226539247025294, "learning_rate": 4.549035709584272e-06, "loss": 0.6958, "step": 19134 }, { "epoch": 0.6930459978268744, "grad_norm": 1.3131974239192705, "learning_rate": 4.548052278373327e-06, "loss": 0.74, "step": 19135 }, { "epoch": 0.6930822165881927, "grad_norm": 1.362795164488524, "learning_rate": 4.54706892218597e-06, "loss": 0.6901, "step": 19136 }, { "epoch": 0.693118435349511, "grad_norm": 0.918135093362758, "learning_rate": 4.546085641035739e-06, "loss": 0.649, "step": 19137 }, { "epoch": 0.6931546541108294, "grad_norm": 1.2011620409687014, "learning_rate": 4.545102434936163e-06, "loss": 0.7127, "step": 19138 }, { "epoch": 0.6931908728721478, "grad_norm": 1.2937155731948808, "learning_rate": 4.544119303900766e-06, "loss": 0.6446, "step": 19139 }, { "epoch": 0.6932270916334662, "grad_norm": 1.2797573413413263, "learning_rate": 4.543136247943085e-06, "loss": 0.6779, "step": 19140 }, { "epoch": 0.6932633103947845, "grad_norm": 1.2883384992386797, "learning_rate": 4.542153267076644e-06, "loss": 0.7279, "step": 19141 }, { "epoch": 0.6932995291561028, "grad_norm": 1.2715731873968994, "learning_rate": 4.54117036131497e-06, "loss": 0.6468, "step": 19142 }, { "epoch": 0.6933357479174213, "grad_norm": 1.442944878823235, "learning_rate": 4.540187530671585e-06, "loss": 0.6599, "step": 19143 }, { "epoch": 0.6933719666787396, "grad_norm": 1.1824361598018678, "learning_rate": 4.5392047751600195e-06, "loss": 0.675, "step": 19144 }, { "epoch": 0.693408185440058, "grad_norm": 0.963051640739352, "learning_rate": 4.538222094793796e-06, "loss": 0.696, "step": 19145 }, { "epoch": 0.6934444042013763, "grad_norm": 1.548303376931975, "learning_rate": 4.537239489586435e-06, "loss": 0.7706, "step": 19146 }, { "epoch": 0.6934806229626946, "grad_norm": 1.4069643643339265, "learning_rate": 4.536256959551455e-06, "loss": 0.7215, "step": 19147 }, { "epoch": 0.6935168417240131, "grad_norm": 1.3183822559207836, "learning_rate": 4.535274504702384e-06, "loss": 0.7186, "step": 19148 }, { "epoch": 0.6935530604853314, "grad_norm": 1.3468222304072384, "learning_rate": 4.534292125052737e-06, "loss": 0.6699, "step": 19149 }, { "epoch": 0.6935892792466498, "grad_norm": 1.3232284770410128, "learning_rate": 4.53330982061603e-06, "loss": 0.6909, "step": 19150 }, { "epoch": 0.6936254980079681, "grad_norm": 1.2664782553057778, "learning_rate": 4.532327591405786e-06, "loss": 0.64, "step": 19151 }, { "epoch": 0.6936617167692865, "grad_norm": 1.211194634632181, "learning_rate": 4.5313454374355195e-06, "loss": 0.7019, "step": 19152 }, { "epoch": 0.6936979355306049, "grad_norm": 1.33357277520149, "learning_rate": 4.530363358718745e-06, "loss": 0.7147, "step": 19153 }, { "epoch": 0.6937341542919232, "grad_norm": 1.3198807408250155, "learning_rate": 4.529381355268972e-06, "loss": 0.6726, "step": 19154 }, { "epoch": 0.6937703730532416, "grad_norm": 1.5798749524898485, "learning_rate": 4.528399427099723e-06, "loss": 0.6958, "step": 19155 }, { "epoch": 0.69380659181456, "grad_norm": 1.1206130373456087, "learning_rate": 4.527417574224505e-06, "loss": 0.7017, "step": 19156 }, { "epoch": 0.6938428105758783, "grad_norm": 1.423851364898446, "learning_rate": 4.52643579665683e-06, "loss": 0.6748, "step": 19157 }, { "epoch": 0.6938790293371967, "grad_norm": 1.5063170629328324, "learning_rate": 4.525454094410205e-06, "loss": 0.6955, "step": 19158 }, { "epoch": 0.693915248098515, "grad_norm": 1.3545371620890996, "learning_rate": 4.524472467498146e-06, "loss": 0.6635, "step": 19159 }, { "epoch": 0.6939514668598334, "grad_norm": 1.1342519809844807, "learning_rate": 4.5234909159341555e-06, "loss": 0.7592, "step": 19160 }, { "epoch": 0.6939876856211518, "grad_norm": 1.2739665159070943, "learning_rate": 4.522509439731739e-06, "loss": 0.7086, "step": 19161 }, { "epoch": 0.6940239043824701, "grad_norm": 1.31858473087937, "learning_rate": 4.52152803890441e-06, "loss": 0.6775, "step": 19162 }, { "epoch": 0.6940601231437885, "grad_norm": 1.0594047991735953, "learning_rate": 4.520546713465669e-06, "loss": 0.6818, "step": 19163 }, { "epoch": 0.6940963419051068, "grad_norm": 1.2986864794311723, "learning_rate": 4.5195654634290185e-06, "loss": 0.5929, "step": 19164 }, { "epoch": 0.6941325606664253, "grad_norm": 1.5144973079365291, "learning_rate": 4.51858428880796e-06, "loss": 0.7816, "step": 19165 }, { "epoch": 0.6941687794277436, "grad_norm": 1.3031783362606537, "learning_rate": 4.517603189616002e-06, "loss": 0.6871, "step": 19166 }, { "epoch": 0.6942049981890619, "grad_norm": 1.327475823196391, "learning_rate": 4.516622165866641e-06, "loss": 0.7088, "step": 19167 }, { "epoch": 0.6942412169503803, "grad_norm": 0.9819182932281451, "learning_rate": 4.515641217573377e-06, "loss": 0.6648, "step": 19168 }, { "epoch": 0.6942774357116986, "grad_norm": 1.1430071368183206, "learning_rate": 4.514660344749706e-06, "loss": 0.6577, "step": 19169 }, { "epoch": 0.6943136544730171, "grad_norm": 1.378902856752365, "learning_rate": 4.5136795474091314e-06, "loss": 0.7241, "step": 19170 }, { "epoch": 0.6943498732343354, "grad_norm": 0.9945860148820657, "learning_rate": 4.5126988255651475e-06, "loss": 0.7442, "step": 19171 }, { "epoch": 0.6943860919956537, "grad_norm": 1.4100930683215582, "learning_rate": 4.5117181792312446e-06, "loss": 0.7037, "step": 19172 }, { "epoch": 0.6944223107569721, "grad_norm": 1.2968242280569846, "learning_rate": 4.510737608420926e-06, "loss": 0.7092, "step": 19173 }, { "epoch": 0.6944585295182905, "grad_norm": 1.0233124534184137, "learning_rate": 4.509757113147681e-06, "loss": 0.7253, "step": 19174 }, { "epoch": 0.6944947482796089, "grad_norm": 0.9786027096676597, "learning_rate": 4.508776693425003e-06, "loss": 0.6708, "step": 19175 }, { "epoch": 0.6945309670409272, "grad_norm": 1.4425224315398724, "learning_rate": 4.507796349266379e-06, "loss": 0.6716, "step": 19176 }, { "epoch": 0.6945671858022455, "grad_norm": 1.339658369162941, "learning_rate": 4.506816080685307e-06, "loss": 0.661, "step": 19177 }, { "epoch": 0.694603404563564, "grad_norm": 1.2967526998648071, "learning_rate": 4.505835887695271e-06, "loss": 0.6761, "step": 19178 }, { "epoch": 0.6946396233248823, "grad_norm": 1.2537891208453635, "learning_rate": 4.504855770309758e-06, "loss": 0.6814, "step": 19179 }, { "epoch": 0.6946758420862007, "grad_norm": 1.0197489486514713, "learning_rate": 4.503875728542263e-06, "loss": 0.739, "step": 19180 }, { "epoch": 0.694712060847519, "grad_norm": 1.3886986822446115, "learning_rate": 4.502895762406265e-06, "loss": 0.6816, "step": 19181 }, { "epoch": 0.6947482796088373, "grad_norm": 1.4398058144782244, "learning_rate": 4.5019158719152535e-06, "loss": 0.6762, "step": 19182 }, { "epoch": 0.6947844983701558, "grad_norm": 1.304025491707089, "learning_rate": 4.500936057082705e-06, "loss": 0.6331, "step": 19183 }, { "epoch": 0.6948207171314741, "grad_norm": 1.0987216820800778, "learning_rate": 4.499956317922114e-06, "loss": 0.7007, "step": 19184 }, { "epoch": 0.6948569358927925, "grad_norm": 1.3472084496981174, "learning_rate": 4.498976654446956e-06, "loss": 0.6687, "step": 19185 }, { "epoch": 0.6948931546541108, "grad_norm": 1.2966842475175435, "learning_rate": 4.497997066670712e-06, "loss": 0.7446, "step": 19186 }, { "epoch": 0.6949293734154292, "grad_norm": 1.375793735928391, "learning_rate": 4.4970175546068595e-06, "loss": 0.7129, "step": 19187 }, { "epoch": 0.6949655921767476, "grad_norm": 1.4048405279358709, "learning_rate": 4.496038118268885e-06, "loss": 0.6774, "step": 19188 }, { "epoch": 0.6950018109380659, "grad_norm": 1.5060157397832608, "learning_rate": 4.495058757670262e-06, "loss": 0.693, "step": 19189 }, { "epoch": 0.6950380296993843, "grad_norm": 1.446435132786546, "learning_rate": 4.494079472824464e-06, "loss": 0.7519, "step": 19190 }, { "epoch": 0.6950742484607026, "grad_norm": 1.2313775929549637, "learning_rate": 4.493100263744974e-06, "loss": 0.6688, "step": 19191 }, { "epoch": 0.695110467222021, "grad_norm": 1.339625199652108, "learning_rate": 4.492121130445264e-06, "loss": 0.6555, "step": 19192 }, { "epoch": 0.6951466859833394, "grad_norm": 1.7371615532616702, "learning_rate": 4.491142072938806e-06, "loss": 0.7015, "step": 19193 }, { "epoch": 0.6951829047446577, "grad_norm": 1.3895034272057099, "learning_rate": 4.490163091239069e-06, "loss": 0.6732, "step": 19194 }, { "epoch": 0.6952191235059761, "grad_norm": 1.2916575890236195, "learning_rate": 4.489184185359534e-06, "loss": 0.7232, "step": 19195 }, { "epoch": 0.6952553422672945, "grad_norm": 1.3397241197012055, "learning_rate": 4.488205355313667e-06, "loss": 0.6966, "step": 19196 }, { "epoch": 0.6952915610286128, "grad_norm": 1.3768727779619414, "learning_rate": 4.487226601114937e-06, "loss": 0.7008, "step": 19197 }, { "epoch": 0.6953277797899312, "grad_norm": 1.351829842080185, "learning_rate": 4.48624792277681e-06, "loss": 0.6123, "step": 19198 }, { "epoch": 0.6953639985512495, "grad_norm": 1.0681140155240325, "learning_rate": 4.485269320312758e-06, "loss": 0.7412, "step": 19199 }, { "epoch": 0.695400217312568, "grad_norm": 1.3411356415561955, "learning_rate": 4.484290793736248e-06, "loss": 0.6977, "step": 19200 }, { "epoch": 0.6954364360738863, "grad_norm": 1.0417595483427946, "learning_rate": 4.4833123430607385e-06, "loss": 0.7765, "step": 19201 }, { "epoch": 0.6954726548352046, "grad_norm": 1.4149518917447421, "learning_rate": 4.482333968299702e-06, "loss": 0.6648, "step": 19202 }, { "epoch": 0.695508873596523, "grad_norm": 1.3560387178458944, "learning_rate": 4.481355669466599e-06, "loss": 0.6973, "step": 19203 }, { "epoch": 0.6955450923578413, "grad_norm": 1.3961032227235897, "learning_rate": 4.480377446574891e-06, "loss": 0.7443, "step": 19204 }, { "epoch": 0.6955813111191598, "grad_norm": 1.1409858157319186, "learning_rate": 4.479399299638035e-06, "loss": 0.7545, "step": 19205 }, { "epoch": 0.6956175298804781, "grad_norm": 1.3768257968280333, "learning_rate": 4.4784212286695005e-06, "loss": 0.6533, "step": 19206 }, { "epoch": 0.6956537486417964, "grad_norm": 1.392738042854302, "learning_rate": 4.47744323368274e-06, "loss": 0.6939, "step": 19207 }, { "epoch": 0.6956899674031148, "grad_norm": 1.370846565766323, "learning_rate": 4.476465314691214e-06, "loss": 0.6473, "step": 19208 }, { "epoch": 0.6957261861644332, "grad_norm": 1.5067389772855995, "learning_rate": 4.475487471708375e-06, "loss": 0.6423, "step": 19209 }, { "epoch": 0.6957624049257516, "grad_norm": 2.4133294042336466, "learning_rate": 4.474509704747686e-06, "loss": 0.6813, "step": 19210 }, { "epoch": 0.6957986236870699, "grad_norm": 1.3780477839525898, "learning_rate": 4.473532013822599e-06, "loss": 0.7461, "step": 19211 }, { "epoch": 0.6958348424483882, "grad_norm": 1.3100798182522329, "learning_rate": 4.472554398946564e-06, "loss": 0.7182, "step": 19212 }, { "epoch": 0.6958710612097067, "grad_norm": 1.0078500715379957, "learning_rate": 4.47157686013304e-06, "loss": 0.723, "step": 19213 }, { "epoch": 0.695907279971025, "grad_norm": 1.4712813274251617, "learning_rate": 4.4705993973954766e-06, "loss": 0.6982, "step": 19214 }, { "epoch": 0.6959434987323434, "grad_norm": 1.6408416287136771, "learning_rate": 4.469622010747324e-06, "loss": 0.7094, "step": 19215 }, { "epoch": 0.6959797174936617, "grad_norm": 1.3312194964278883, "learning_rate": 4.468644700202028e-06, "loss": 0.6414, "step": 19216 }, { "epoch": 0.69601593625498, "grad_norm": 1.0803699541611753, "learning_rate": 4.467667465773046e-06, "loss": 0.7192, "step": 19217 }, { "epoch": 0.6960521550162985, "grad_norm": 1.449500981839615, "learning_rate": 4.46669030747382e-06, "loss": 0.7696, "step": 19218 }, { "epoch": 0.6960883737776168, "grad_norm": 1.3977193400225907, "learning_rate": 4.465713225317797e-06, "loss": 0.7072, "step": 19219 }, { "epoch": 0.6961245925389352, "grad_norm": 1.0454341158176936, "learning_rate": 4.46473621931842e-06, "loss": 0.6977, "step": 19220 }, { "epoch": 0.6961608113002535, "grad_norm": 1.173240638377534, "learning_rate": 4.463759289489139e-06, "loss": 0.6729, "step": 19221 }, { "epoch": 0.6961970300615719, "grad_norm": 1.325722126550036, "learning_rate": 4.4627824358433956e-06, "loss": 0.7166, "step": 19222 }, { "epoch": 0.6962332488228903, "grad_norm": 1.1035138186613, "learning_rate": 4.461805658394628e-06, "loss": 0.7387, "step": 19223 }, { "epoch": 0.6962694675842086, "grad_norm": 1.4016883315977615, "learning_rate": 4.460828957156283e-06, "loss": 0.7183, "step": 19224 }, { "epoch": 0.696305686345527, "grad_norm": 1.416223347653828, "learning_rate": 4.4598523321418e-06, "loss": 0.6913, "step": 19225 }, { "epoch": 0.6963419051068453, "grad_norm": 1.0946259829905594, "learning_rate": 4.458875783364617e-06, "loss": 0.6642, "step": 19226 }, { "epoch": 0.6963781238681637, "grad_norm": 1.7394311594695198, "learning_rate": 4.4578993108381674e-06, "loss": 0.7199, "step": 19227 }, { "epoch": 0.6964143426294821, "grad_norm": 1.26259847580922, "learning_rate": 4.456922914575897e-06, "loss": 0.6314, "step": 19228 }, { "epoch": 0.6964505613908004, "grad_norm": 1.3649461691683833, "learning_rate": 4.4559465945912385e-06, "loss": 0.7266, "step": 19229 }, { "epoch": 0.6964867801521188, "grad_norm": 1.3703388991797973, "learning_rate": 4.454970350897625e-06, "loss": 0.6347, "step": 19230 }, { "epoch": 0.6965229989134372, "grad_norm": 1.3490471407170839, "learning_rate": 4.453994183508489e-06, "loss": 0.6856, "step": 19231 }, { "epoch": 0.6965592176747555, "grad_norm": 1.3926649160210731, "learning_rate": 4.45301809243727e-06, "loss": 0.682, "step": 19232 }, { "epoch": 0.6965954364360739, "grad_norm": 1.4760196535635688, "learning_rate": 4.452042077697396e-06, "loss": 0.674, "step": 19233 }, { "epoch": 0.6966316551973922, "grad_norm": 2.1631510494412582, "learning_rate": 4.451066139302294e-06, "loss": 0.7532, "step": 19234 }, { "epoch": 0.6966678739587107, "grad_norm": 1.4017722684112972, "learning_rate": 4.450090277265401e-06, "loss": 0.7015, "step": 19235 }, { "epoch": 0.696704092720029, "grad_norm": 1.37394019419012, "learning_rate": 4.449114491600143e-06, "loss": 0.7185, "step": 19236 }, { "epoch": 0.6967403114813473, "grad_norm": 1.3761552345413244, "learning_rate": 4.4481387823199465e-06, "loss": 0.6528, "step": 19237 }, { "epoch": 0.6967765302426657, "grad_norm": 1.3793149432287461, "learning_rate": 4.447163149438235e-06, "loss": 0.6791, "step": 19238 }, { "epoch": 0.696812749003984, "grad_norm": 1.4178358034068992, "learning_rate": 4.446187592968441e-06, "loss": 0.7053, "step": 19239 }, { "epoch": 0.6968489677653025, "grad_norm": 1.2970848761063891, "learning_rate": 4.445212112923988e-06, "loss": 0.6276, "step": 19240 }, { "epoch": 0.6968851865266208, "grad_norm": 1.400308042615774, "learning_rate": 4.444236709318291e-06, "loss": 0.7288, "step": 19241 }, { "epoch": 0.6969214052879391, "grad_norm": 1.3399604991172271, "learning_rate": 4.443261382164783e-06, "loss": 0.7209, "step": 19242 }, { "epoch": 0.6969576240492575, "grad_norm": 1.3471082831085583, "learning_rate": 4.442286131476882e-06, "loss": 0.7221, "step": 19243 }, { "epoch": 0.6969938428105759, "grad_norm": 1.2756283226032992, "learning_rate": 4.4413109572680054e-06, "loss": 0.6926, "step": 19244 }, { "epoch": 0.6970300615718943, "grad_norm": 1.3404546143110736, "learning_rate": 4.440335859551572e-06, "loss": 0.6798, "step": 19245 }, { "epoch": 0.6970662803332126, "grad_norm": 1.2429726573911446, "learning_rate": 4.439360838341006e-06, "loss": 0.6354, "step": 19246 }, { "epoch": 0.6971024990945309, "grad_norm": 1.3990557558618415, "learning_rate": 4.438385893649719e-06, "loss": 0.6968, "step": 19247 }, { "epoch": 0.6971387178558494, "grad_norm": 1.0907912910286168, "learning_rate": 4.437411025491131e-06, "loss": 0.6896, "step": 19248 }, { "epoch": 0.6971749366171677, "grad_norm": 1.4603792197371321, "learning_rate": 4.43643623387865e-06, "loss": 0.7246, "step": 19249 }, { "epoch": 0.6972111553784861, "grad_norm": 1.311329845078482, "learning_rate": 4.4354615188257e-06, "loss": 0.6668, "step": 19250 }, { "epoch": 0.6972473741398044, "grad_norm": 1.5242460471064543, "learning_rate": 4.434486880345689e-06, "loss": 0.7044, "step": 19251 }, { "epoch": 0.6972835929011227, "grad_norm": 1.5063263160517035, "learning_rate": 4.433512318452024e-06, "loss": 0.707, "step": 19252 }, { "epoch": 0.6973198116624412, "grad_norm": 1.3221904593815297, "learning_rate": 4.432537833158125e-06, "loss": 0.7185, "step": 19253 }, { "epoch": 0.6973560304237595, "grad_norm": 1.2607465118995922, "learning_rate": 4.4315634244773974e-06, "loss": 0.6127, "step": 19254 }, { "epoch": 0.6973922491850779, "grad_norm": 1.4338044029987618, "learning_rate": 4.43058909242325e-06, "loss": 0.7177, "step": 19255 }, { "epoch": 0.6974284679463962, "grad_norm": 1.3564736649617124, "learning_rate": 4.429614837009086e-06, "loss": 0.7738, "step": 19256 }, { "epoch": 0.6974646867077146, "grad_norm": 1.2768916538866095, "learning_rate": 4.428640658248321e-06, "loss": 0.6434, "step": 19257 }, { "epoch": 0.697500905469033, "grad_norm": 1.4389867272863388, "learning_rate": 4.427666556154355e-06, "loss": 0.6817, "step": 19258 }, { "epoch": 0.6975371242303513, "grad_norm": 1.3371070840204677, "learning_rate": 4.426692530740594e-06, "loss": 0.6618, "step": 19259 }, { "epoch": 0.6975733429916697, "grad_norm": 1.3863903812027791, "learning_rate": 4.425718582020438e-06, "loss": 0.6971, "step": 19260 }, { "epoch": 0.697609561752988, "grad_norm": 1.3583851401571543, "learning_rate": 4.424744710007295e-06, "loss": 0.6523, "step": 19261 }, { "epoch": 0.6976457805143064, "grad_norm": 1.249227533971597, "learning_rate": 4.423770914714564e-06, "loss": 0.6539, "step": 19262 }, { "epoch": 0.6976819992756248, "grad_norm": 1.3824723156632732, "learning_rate": 4.42279719615564e-06, "loss": 0.6495, "step": 19263 }, { "epoch": 0.6977182180369431, "grad_norm": 1.3594448137443047, "learning_rate": 4.421823554343933e-06, "loss": 0.7177, "step": 19264 }, { "epoch": 0.6977544367982615, "grad_norm": 1.1550146309399847, "learning_rate": 4.420849989292834e-06, "loss": 0.7012, "step": 19265 }, { "epoch": 0.6977906555595799, "grad_norm": 1.3426771621869023, "learning_rate": 4.419876501015742e-06, "loss": 0.7091, "step": 19266 }, { "epoch": 0.6978268743208982, "grad_norm": 1.2770341452751632, "learning_rate": 4.418903089526048e-06, "loss": 0.6622, "step": 19267 }, { "epoch": 0.6978630930822166, "grad_norm": 1.2213471983258046, "learning_rate": 4.4179297548371555e-06, "loss": 0.7158, "step": 19268 }, { "epoch": 0.6978993118435349, "grad_norm": 1.3687413030559292, "learning_rate": 4.416956496962455e-06, "loss": 0.6888, "step": 19269 }, { "epoch": 0.6979355306048534, "grad_norm": 0.9973501272019504, "learning_rate": 4.415983315915338e-06, "loss": 0.7416, "step": 19270 }, { "epoch": 0.6979717493661717, "grad_norm": 1.2536977296725704, "learning_rate": 4.415010211709193e-06, "loss": 0.6872, "step": 19271 }, { "epoch": 0.69800796812749, "grad_norm": 0.8585740261490513, "learning_rate": 4.414037184357419e-06, "loss": 0.6552, "step": 19272 }, { "epoch": 0.6980441868888084, "grad_norm": 1.0668366427396716, "learning_rate": 4.413064233873403e-06, "loss": 0.7379, "step": 19273 }, { "epoch": 0.6980804056501267, "grad_norm": 1.1278922462338605, "learning_rate": 4.4120913602705276e-06, "loss": 0.6764, "step": 19274 }, { "epoch": 0.6981166244114452, "grad_norm": 1.2350764914487413, "learning_rate": 4.411118563562188e-06, "loss": 0.6765, "step": 19275 }, { "epoch": 0.6981528431727635, "grad_norm": 0.9633263969334133, "learning_rate": 4.410145843761768e-06, "loss": 0.6395, "step": 19276 }, { "epoch": 0.6981890619340818, "grad_norm": 1.213924095894428, "learning_rate": 4.4091732008826535e-06, "loss": 0.6515, "step": 19277 }, { "epoch": 0.6982252806954002, "grad_norm": 1.487594115461464, "learning_rate": 4.4082006349382245e-06, "loss": 0.6913, "step": 19278 }, { "epoch": 0.6982614994567186, "grad_norm": 1.2864239948885516, "learning_rate": 4.407228145941872e-06, "loss": 0.7694, "step": 19279 }, { "epoch": 0.698297718218037, "grad_norm": 1.278129268297516, "learning_rate": 4.406255733906975e-06, "loss": 0.6797, "step": 19280 }, { "epoch": 0.6983339369793553, "grad_norm": 1.298486703440778, "learning_rate": 4.405283398846913e-06, "loss": 0.7366, "step": 19281 }, { "epoch": 0.6983701557406736, "grad_norm": 1.2732359267348734, "learning_rate": 4.404311140775067e-06, "loss": 0.662, "step": 19282 }, { "epoch": 0.698406374501992, "grad_norm": 1.3452507422723319, "learning_rate": 4.403338959704817e-06, "loss": 0.6668, "step": 19283 }, { "epoch": 0.6984425932633104, "grad_norm": 1.2478042203968613, "learning_rate": 4.402366855649537e-06, "loss": 0.6837, "step": 19284 }, { "epoch": 0.6984788120246288, "grad_norm": 1.201231489795593, "learning_rate": 4.401394828622612e-06, "loss": 0.6149, "step": 19285 }, { "epoch": 0.6985150307859471, "grad_norm": 1.2402144032469755, "learning_rate": 4.400422878637412e-06, "loss": 0.7049, "step": 19286 }, { "epoch": 0.6985512495472654, "grad_norm": 1.3378324704861682, "learning_rate": 4.399451005707315e-06, "loss": 0.6343, "step": 19287 }, { "epoch": 0.6985874683085839, "grad_norm": 1.450579197483884, "learning_rate": 4.3984792098456885e-06, "loss": 0.7852, "step": 19288 }, { "epoch": 0.6986236870699022, "grad_norm": 1.2364012476613677, "learning_rate": 4.397507491065914e-06, "loss": 0.6411, "step": 19289 }, { "epoch": 0.6986599058312206, "grad_norm": 1.298168714822205, "learning_rate": 4.396535849381359e-06, "loss": 0.6419, "step": 19290 }, { "epoch": 0.6986961245925389, "grad_norm": 1.803009638327222, "learning_rate": 4.395564284805395e-06, "loss": 0.7162, "step": 19291 }, { "epoch": 0.6987323433538573, "grad_norm": 1.3937974209159447, "learning_rate": 4.394592797351387e-06, "loss": 0.6716, "step": 19292 }, { "epoch": 0.6987685621151757, "grad_norm": 1.3365736897983267, "learning_rate": 4.393621387032712e-06, "loss": 0.6898, "step": 19293 }, { "epoch": 0.698804780876494, "grad_norm": 1.4858653240764454, "learning_rate": 4.392650053862731e-06, "loss": 0.6644, "step": 19294 }, { "epoch": 0.6988409996378124, "grad_norm": 1.4901890629070327, "learning_rate": 4.3916787978548094e-06, "loss": 0.6669, "step": 19295 }, { "epoch": 0.6988772183991308, "grad_norm": 1.3271252964856974, "learning_rate": 4.390707619022319e-06, "loss": 0.6442, "step": 19296 }, { "epoch": 0.6989134371604491, "grad_norm": 1.227516911232059, "learning_rate": 4.389736517378621e-06, "loss": 0.6913, "step": 19297 }, { "epoch": 0.6989496559217675, "grad_norm": 1.2532511147931773, "learning_rate": 4.3887654929370785e-06, "loss": 0.6801, "step": 19298 }, { "epoch": 0.6989858746830858, "grad_norm": 1.224818712195871, "learning_rate": 4.387794545711048e-06, "loss": 0.6944, "step": 19299 }, { "epoch": 0.6990220934444042, "grad_norm": 1.3411110141240268, "learning_rate": 4.3868236757139e-06, "loss": 0.6274, "step": 19300 }, { "epoch": 0.6990583122057226, "grad_norm": 0.9588820810542867, "learning_rate": 4.385852882958992e-06, "loss": 0.761, "step": 19301 }, { "epoch": 0.6990945309670409, "grad_norm": 1.0116839558231743, "learning_rate": 4.38488216745968e-06, "loss": 0.6973, "step": 19302 }, { "epoch": 0.6991307497283593, "grad_norm": 1.2867204160916925, "learning_rate": 4.38391152922932e-06, "loss": 0.6833, "step": 19303 }, { "epoch": 0.6991669684896776, "grad_norm": 1.24383684039664, "learning_rate": 4.3829409682812755e-06, "loss": 0.6531, "step": 19304 }, { "epoch": 0.6992031872509961, "grad_norm": 1.3036965744427491, "learning_rate": 4.381970484628899e-06, "loss": 0.6366, "step": 19305 }, { "epoch": 0.6992394060123144, "grad_norm": 0.9409853651776067, "learning_rate": 4.381000078285541e-06, "loss": 0.6642, "step": 19306 }, { "epoch": 0.6992756247736327, "grad_norm": 1.3581703420977327, "learning_rate": 4.380029749264564e-06, "loss": 0.7062, "step": 19307 }, { "epoch": 0.6993118435349511, "grad_norm": 1.3808116176175749, "learning_rate": 4.379059497579315e-06, "loss": 0.7042, "step": 19308 }, { "epoch": 0.6993480622962694, "grad_norm": 1.3381484159216366, "learning_rate": 4.378089323243147e-06, "loss": 0.6366, "step": 19309 }, { "epoch": 0.6993842810575879, "grad_norm": 1.4447909490242732, "learning_rate": 4.377119226269404e-06, "loss": 0.7548, "step": 19310 }, { "epoch": 0.6994204998189062, "grad_norm": 1.0088620303713758, "learning_rate": 4.376149206671446e-06, "loss": 0.7033, "step": 19311 }, { "epoch": 0.6994567185802245, "grad_norm": 1.349277586667391, "learning_rate": 4.375179264462618e-06, "loss": 0.7234, "step": 19312 }, { "epoch": 0.6994929373415429, "grad_norm": 1.3190676799732108, "learning_rate": 4.374209399656264e-06, "loss": 0.6984, "step": 19313 }, { "epoch": 0.6995291561028613, "grad_norm": 1.5730687524729698, "learning_rate": 4.373239612265729e-06, "loss": 0.6757, "step": 19314 }, { "epoch": 0.6995653748641797, "grad_norm": 1.3512920814722817, "learning_rate": 4.3722699023043634e-06, "loss": 0.7107, "step": 19315 }, { "epoch": 0.699601593625498, "grad_norm": 1.215378684148106, "learning_rate": 4.37130026978551e-06, "loss": 0.6953, "step": 19316 }, { "epoch": 0.6996378123868163, "grad_norm": 1.1206615084409073, "learning_rate": 4.370330714722507e-06, "loss": 0.7124, "step": 19317 }, { "epoch": 0.6996740311481348, "grad_norm": 1.0380480585337708, "learning_rate": 4.369361237128704e-06, "loss": 0.66, "step": 19318 }, { "epoch": 0.6997102499094531, "grad_norm": 1.5359486944269927, "learning_rate": 4.368391837017437e-06, "loss": 0.6614, "step": 19319 }, { "epoch": 0.6997464686707715, "grad_norm": 1.4366663298344642, "learning_rate": 4.367422514402048e-06, "loss": 0.6862, "step": 19320 }, { "epoch": 0.6997826874320898, "grad_norm": 0.9737801209615421, "learning_rate": 4.36645326929587e-06, "loss": 0.6279, "step": 19321 }, { "epoch": 0.6998189061934081, "grad_norm": 0.9470164609455229, "learning_rate": 4.365484101712249e-06, "loss": 0.6489, "step": 19322 }, { "epoch": 0.6998551249547266, "grad_norm": 1.24023413845871, "learning_rate": 4.364515011664517e-06, "loss": 0.6876, "step": 19323 }, { "epoch": 0.6998913437160449, "grad_norm": 1.3717984510541203, "learning_rate": 4.36354599916601e-06, "loss": 0.7116, "step": 19324 }, { "epoch": 0.6999275624773633, "grad_norm": 1.6745087170230546, "learning_rate": 4.362577064230059e-06, "loss": 0.6785, "step": 19325 }, { "epoch": 0.6999637812386816, "grad_norm": 1.3097736263400497, "learning_rate": 4.361608206870006e-06, "loss": 0.6693, "step": 19326 }, { "epoch": 0.7, "grad_norm": 1.2554150154148753, "learning_rate": 4.360639427099177e-06, "loss": 0.6533, "step": 19327 }, { "epoch": 0.7000362187613184, "grad_norm": 1.255435569528871, "learning_rate": 4.359670724930902e-06, "loss": 0.6364, "step": 19328 }, { "epoch": 0.7000724375226367, "grad_norm": 1.3569862236092407, "learning_rate": 4.358702100378517e-06, "loss": 0.613, "step": 19329 }, { "epoch": 0.7001086562839551, "grad_norm": 1.3265355725245713, "learning_rate": 4.357733553455349e-06, "loss": 0.6638, "step": 19330 }, { "epoch": 0.7001448750452735, "grad_norm": 1.2280272603271942, "learning_rate": 4.356765084174723e-06, "loss": 0.6347, "step": 19331 }, { "epoch": 0.7001810938065918, "grad_norm": 1.2529015413874411, "learning_rate": 4.355796692549966e-06, "loss": 0.6706, "step": 19332 }, { "epoch": 0.7002173125679102, "grad_norm": 1.4078147262614913, "learning_rate": 4.354828378594409e-06, "loss": 0.7094, "step": 19333 }, { "epoch": 0.7002535313292285, "grad_norm": 1.4618231847696241, "learning_rate": 4.353860142321374e-06, "loss": 0.7564, "step": 19334 }, { "epoch": 0.700289750090547, "grad_norm": 1.3385310777166186, "learning_rate": 4.352891983744182e-06, "loss": 0.6903, "step": 19335 }, { "epoch": 0.7003259688518653, "grad_norm": 1.3119277181875857, "learning_rate": 4.351923902876161e-06, "loss": 0.6778, "step": 19336 }, { "epoch": 0.7003621876131836, "grad_norm": 1.3720072803069558, "learning_rate": 4.350955899730629e-06, "loss": 0.6709, "step": 19337 }, { "epoch": 0.700398406374502, "grad_norm": 1.236022366022586, "learning_rate": 4.349987974320908e-06, "loss": 0.6465, "step": 19338 }, { "epoch": 0.7004346251358203, "grad_norm": 1.4254901824037414, "learning_rate": 4.349020126660314e-06, "loss": 0.6909, "step": 19339 }, { "epoch": 0.7004708438971388, "grad_norm": 1.278437606145509, "learning_rate": 4.3480523567621716e-06, "loss": 0.6314, "step": 19340 }, { "epoch": 0.7005070626584571, "grad_norm": 1.0332752672560972, "learning_rate": 4.347084664639795e-06, "loss": 0.6906, "step": 19341 }, { "epoch": 0.7005432814197754, "grad_norm": 1.001366249181549, "learning_rate": 4.346117050306501e-06, "loss": 0.7116, "step": 19342 }, { "epoch": 0.7005795001810938, "grad_norm": 1.4182020486741318, "learning_rate": 4.3451495137756e-06, "loss": 0.6544, "step": 19343 }, { "epoch": 0.7006157189424121, "grad_norm": 1.3444548327481523, "learning_rate": 4.344182055060414e-06, "loss": 0.6796, "step": 19344 }, { "epoch": 0.7006519377037306, "grad_norm": 0.9970231332274425, "learning_rate": 4.343214674174253e-06, "loss": 0.6482, "step": 19345 }, { "epoch": 0.7006881564650489, "grad_norm": 1.440404329884925, "learning_rate": 4.342247371130426e-06, "loss": 0.6837, "step": 19346 }, { "epoch": 0.7007243752263672, "grad_norm": 1.2516287042607177, "learning_rate": 4.3412801459422485e-06, "loss": 0.6077, "step": 19347 }, { "epoch": 0.7007605939876856, "grad_norm": 1.4433300417963684, "learning_rate": 4.340312998623029e-06, "loss": 0.7108, "step": 19348 }, { "epoch": 0.700796812749004, "grad_norm": 1.5335934087565388, "learning_rate": 4.3393459291860755e-06, "loss": 0.7187, "step": 19349 }, { "epoch": 0.7008330315103224, "grad_norm": 1.268483209835443, "learning_rate": 4.338378937644693e-06, "loss": 0.6423, "step": 19350 }, { "epoch": 0.7008692502716407, "grad_norm": 1.378453334985705, "learning_rate": 4.337412024012193e-06, "loss": 0.7664, "step": 19351 }, { "epoch": 0.700905469032959, "grad_norm": 1.234044057055487, "learning_rate": 4.336445188301881e-06, "loss": 0.5968, "step": 19352 }, { "epoch": 0.7009416877942775, "grad_norm": 1.0769903202514317, "learning_rate": 4.335478430527058e-06, "loss": 0.7403, "step": 19353 }, { "epoch": 0.7009779065555958, "grad_norm": 1.3267782711202083, "learning_rate": 4.334511750701027e-06, "loss": 0.7492, "step": 19354 }, { "epoch": 0.7010141253169142, "grad_norm": 1.3394262188767496, "learning_rate": 4.333545148837095e-06, "loss": 0.7576, "step": 19355 }, { "epoch": 0.7010503440782325, "grad_norm": 1.0890640632482718, "learning_rate": 4.332578624948561e-06, "loss": 0.7292, "step": 19356 }, { "epoch": 0.7010865628395508, "grad_norm": 1.2335685882937544, "learning_rate": 4.331612179048722e-06, "loss": 0.6839, "step": 19357 }, { "epoch": 0.7011227816008693, "grad_norm": 1.3848986494522997, "learning_rate": 4.330645811150883e-06, "loss": 0.679, "step": 19358 }, { "epoch": 0.7011590003621876, "grad_norm": 1.4202531043885211, "learning_rate": 4.329679521268339e-06, "loss": 0.7474, "step": 19359 }, { "epoch": 0.701195219123506, "grad_norm": 1.3535520168878192, "learning_rate": 4.328713309414388e-06, "loss": 0.68, "step": 19360 }, { "epoch": 0.7012314378848243, "grad_norm": 1.4055693250615353, "learning_rate": 4.327747175602321e-06, "loss": 0.7136, "step": 19361 }, { "epoch": 0.7012676566461427, "grad_norm": 1.173452838595546, "learning_rate": 4.32678111984544e-06, "loss": 0.6878, "step": 19362 }, { "epoch": 0.7013038754074611, "grad_norm": 1.2922208962743356, "learning_rate": 4.325815142157036e-06, "loss": 0.6485, "step": 19363 }, { "epoch": 0.7013400941687794, "grad_norm": 1.1743985431877446, "learning_rate": 4.324849242550402e-06, "loss": 0.7277, "step": 19364 }, { "epoch": 0.7013763129300978, "grad_norm": 1.4672230197996463, "learning_rate": 4.323883421038825e-06, "loss": 0.7803, "step": 19365 }, { "epoch": 0.7014125316914162, "grad_norm": 1.074091519656502, "learning_rate": 4.322917677635604e-06, "loss": 0.6324, "step": 19366 }, { "epoch": 0.7014487504527345, "grad_norm": 1.2804874082053121, "learning_rate": 4.321952012354023e-06, "loss": 0.7294, "step": 19367 }, { "epoch": 0.7014849692140529, "grad_norm": 1.0827095322475653, "learning_rate": 4.320986425207369e-06, "loss": 0.7124, "step": 19368 }, { "epoch": 0.7015211879753712, "grad_norm": 1.3308559991509898, "learning_rate": 4.320020916208934e-06, "loss": 0.7016, "step": 19369 }, { "epoch": 0.7015574067366896, "grad_norm": 1.009229408567876, "learning_rate": 4.3190554853720025e-06, "loss": 0.7489, "step": 19370 }, { "epoch": 0.701593625498008, "grad_norm": 1.4056686908539189, "learning_rate": 4.31809013270986e-06, "loss": 0.7566, "step": 19371 }, { "epoch": 0.7016298442593263, "grad_norm": 1.377724430144596, "learning_rate": 4.317124858235785e-06, "loss": 0.6789, "step": 19372 }, { "epoch": 0.7016660630206447, "grad_norm": 1.2768598778905653, "learning_rate": 4.31615966196307e-06, "loss": 0.6666, "step": 19373 }, { "epoch": 0.701702281781963, "grad_norm": 1.3863266113821302, "learning_rate": 4.315194543904993e-06, "loss": 0.7446, "step": 19374 }, { "epoch": 0.7017385005432815, "grad_norm": 1.4438385897039878, "learning_rate": 4.3142295040748316e-06, "loss": 0.6706, "step": 19375 }, { "epoch": 0.7017747193045998, "grad_norm": 1.0170891123552481, "learning_rate": 4.313264542485867e-06, "loss": 0.6981, "step": 19376 }, { "epoch": 0.7018109380659181, "grad_norm": 1.257426001123376, "learning_rate": 4.312299659151381e-06, "loss": 0.6473, "step": 19377 }, { "epoch": 0.7018471568272365, "grad_norm": 1.2973854592824956, "learning_rate": 4.311334854084651e-06, "loss": 0.5996, "step": 19378 }, { "epoch": 0.7018833755885548, "grad_norm": 1.2250596330754546, "learning_rate": 4.310370127298948e-06, "loss": 0.7149, "step": 19379 }, { "epoch": 0.7019195943498733, "grad_norm": 1.1264631081864556, "learning_rate": 4.309405478807553e-06, "loss": 0.7001, "step": 19380 }, { "epoch": 0.7019558131111916, "grad_norm": 1.0242243129725457, "learning_rate": 4.308440908623741e-06, "loss": 0.6668, "step": 19381 }, { "epoch": 0.7019920318725099, "grad_norm": 1.3518991643634817, "learning_rate": 4.307476416760783e-06, "loss": 0.7405, "step": 19382 }, { "epoch": 0.7020282506338283, "grad_norm": 1.272077557563503, "learning_rate": 4.306512003231946e-06, "loss": 0.7148, "step": 19383 }, { "epoch": 0.7020644693951467, "grad_norm": 1.5546157391438642, "learning_rate": 4.305547668050511e-06, "loss": 0.7294, "step": 19384 }, { "epoch": 0.7021006881564651, "grad_norm": 1.346104879576801, "learning_rate": 4.3045834112297445e-06, "loss": 0.7078, "step": 19385 }, { "epoch": 0.7021369069177834, "grad_norm": 1.049416464281555, "learning_rate": 4.303619232782915e-06, "loss": 0.7114, "step": 19386 }, { "epoch": 0.7021731256791017, "grad_norm": 1.2536913889703822, "learning_rate": 4.3026551327232854e-06, "loss": 0.6633, "step": 19387 }, { "epoch": 0.7022093444404202, "grad_norm": 1.4356147853851629, "learning_rate": 4.301691111064132e-06, "loss": 0.7713, "step": 19388 }, { "epoch": 0.7022455632017385, "grad_norm": 1.17402050945759, "learning_rate": 4.300727167818714e-06, "loss": 0.6091, "step": 19389 }, { "epoch": 0.7022817819630569, "grad_norm": 1.3604392491030721, "learning_rate": 4.299763303000296e-06, "loss": 0.749, "step": 19390 }, { "epoch": 0.7023180007243752, "grad_norm": 1.291692797110672, "learning_rate": 4.298799516622147e-06, "loss": 0.7608, "step": 19391 }, { "epoch": 0.7023542194856935, "grad_norm": 1.3924382823051702, "learning_rate": 4.297835808697526e-06, "loss": 0.7067, "step": 19392 }, { "epoch": 0.702390438247012, "grad_norm": 1.2829778980200301, "learning_rate": 4.296872179239695e-06, "loss": 0.6912, "step": 19393 }, { "epoch": 0.7024266570083303, "grad_norm": 1.2269257634721185, "learning_rate": 4.29590862826191e-06, "loss": 0.7056, "step": 19394 }, { "epoch": 0.7024628757696487, "grad_norm": 1.4351821994148692, "learning_rate": 4.294945155777438e-06, "loss": 0.719, "step": 19395 }, { "epoch": 0.702499094530967, "grad_norm": 1.4177300249603613, "learning_rate": 4.293981761799533e-06, "loss": 0.6717, "step": 19396 }, { "epoch": 0.7025353132922854, "grad_norm": 1.026640081938495, "learning_rate": 4.29301844634145e-06, "loss": 0.7027, "step": 19397 }, { "epoch": 0.7025715320536038, "grad_norm": 1.0587791527069903, "learning_rate": 4.29205520941645e-06, "loss": 0.7126, "step": 19398 }, { "epoch": 0.7026077508149221, "grad_norm": 1.3168345901997465, "learning_rate": 4.291092051037787e-06, "loss": 0.7216, "step": 19399 }, { "epoch": 0.7026439695762405, "grad_norm": 1.5566471008778726, "learning_rate": 4.290128971218713e-06, "loss": 0.746, "step": 19400 }, { "epoch": 0.7026801883375589, "grad_norm": 1.4097441681423135, "learning_rate": 4.289165969972479e-06, "loss": 0.7538, "step": 19401 }, { "epoch": 0.7027164070988772, "grad_norm": 1.4496515808858772, "learning_rate": 4.288203047312343e-06, "loss": 0.7543, "step": 19402 }, { "epoch": 0.7027526258601956, "grad_norm": 1.4891591080021425, "learning_rate": 4.287240203251551e-06, "loss": 0.7409, "step": 19403 }, { "epoch": 0.7027888446215139, "grad_norm": 1.3865034231720743, "learning_rate": 4.2862774378033525e-06, "loss": 0.6743, "step": 19404 }, { "epoch": 0.7028250633828323, "grad_norm": 1.252321704550289, "learning_rate": 4.285314750980994e-06, "loss": 0.6099, "step": 19405 }, { "epoch": 0.7028612821441507, "grad_norm": 1.3042203950751616, "learning_rate": 4.284352142797731e-06, "loss": 0.6833, "step": 19406 }, { "epoch": 0.702897500905469, "grad_norm": 1.2147015284424252, "learning_rate": 4.2833896132668025e-06, "loss": 0.5901, "step": 19407 }, { "epoch": 0.7029337196667874, "grad_norm": 1.0483899555391207, "learning_rate": 4.282427162401454e-06, "loss": 0.6963, "step": 19408 }, { "epoch": 0.7029699384281057, "grad_norm": 1.5987195181941987, "learning_rate": 4.281464790214935e-06, "loss": 0.7201, "step": 19409 }, { "epoch": 0.7030061571894242, "grad_norm": 1.4006015546909747, "learning_rate": 4.280502496720484e-06, "loss": 0.7011, "step": 19410 }, { "epoch": 0.7030423759507425, "grad_norm": 1.3563038090994375, "learning_rate": 4.279540281931346e-06, "loss": 0.7624, "step": 19411 }, { "epoch": 0.7030785947120608, "grad_norm": 1.3943776652500015, "learning_rate": 4.2785781458607555e-06, "loss": 0.6348, "step": 19412 }, { "epoch": 0.7031148134733792, "grad_norm": 1.4519912337052983, "learning_rate": 4.277616088521961e-06, "loss": 0.7534, "step": 19413 }, { "epoch": 0.7031510322346975, "grad_norm": 1.2524295906714191, "learning_rate": 4.276654109928197e-06, "loss": 0.7103, "step": 19414 }, { "epoch": 0.703187250996016, "grad_norm": 1.0773402072630918, "learning_rate": 4.275692210092701e-06, "loss": 0.7406, "step": 19415 }, { "epoch": 0.7032234697573343, "grad_norm": 1.4168385607085217, "learning_rate": 4.2747303890287076e-06, "loss": 0.6845, "step": 19416 }, { "epoch": 0.7032596885186526, "grad_norm": 1.4158508612701932, "learning_rate": 4.273768646749457e-06, "loss": 0.7432, "step": 19417 }, { "epoch": 0.703295907279971, "grad_norm": 0.9983984451687313, "learning_rate": 4.272806983268182e-06, "loss": 0.6877, "step": 19418 }, { "epoch": 0.7033321260412894, "grad_norm": 1.4354820003252995, "learning_rate": 4.271845398598111e-06, "loss": 0.6105, "step": 19419 }, { "epoch": 0.7033683448026078, "grad_norm": 1.262477327104264, "learning_rate": 4.270883892752485e-06, "loss": 0.7099, "step": 19420 }, { "epoch": 0.7034045635639261, "grad_norm": 1.0002045428622621, "learning_rate": 4.26992246574453e-06, "loss": 0.7434, "step": 19421 }, { "epoch": 0.7034407823252444, "grad_norm": 1.555998827828328, "learning_rate": 4.268961117587478e-06, "loss": 0.7645, "step": 19422 }, { "epoch": 0.7034770010865629, "grad_norm": 1.497111881671844, "learning_rate": 4.267999848294551e-06, "loss": 0.8023, "step": 19423 }, { "epoch": 0.7035132198478812, "grad_norm": 1.332825313553478, "learning_rate": 4.267038657878988e-06, "loss": 0.6584, "step": 19424 }, { "epoch": 0.7035494386091996, "grad_norm": 1.386062120062327, "learning_rate": 4.266077546354011e-06, "loss": 0.7271, "step": 19425 }, { "epoch": 0.7035856573705179, "grad_norm": 1.4010333497628977, "learning_rate": 4.2651165137328446e-06, "loss": 0.6462, "step": 19426 }, { "epoch": 0.7036218761318362, "grad_norm": 1.4308717453234678, "learning_rate": 4.264155560028709e-06, "loss": 0.6298, "step": 19427 }, { "epoch": 0.7036580948931547, "grad_norm": 1.2311463586821085, "learning_rate": 4.263194685254838e-06, "loss": 0.6527, "step": 19428 }, { "epoch": 0.703694313654473, "grad_norm": 1.2884769865575685, "learning_rate": 4.262233889424448e-06, "loss": 0.7047, "step": 19429 }, { "epoch": 0.7037305324157914, "grad_norm": 1.3359374400467887, "learning_rate": 4.2612731725507585e-06, "loss": 0.6566, "step": 19430 }, { "epoch": 0.7037667511771097, "grad_norm": 1.5396976975532406, "learning_rate": 4.260312534646996e-06, "loss": 0.6448, "step": 19431 }, { "epoch": 0.7038029699384281, "grad_norm": 1.3051193601869864, "learning_rate": 4.259351975726377e-06, "loss": 0.6617, "step": 19432 }, { "epoch": 0.7038391886997465, "grad_norm": 1.4278037821471634, "learning_rate": 4.258391495802118e-06, "loss": 0.6742, "step": 19433 }, { "epoch": 0.7038754074610648, "grad_norm": 1.3161391396164237, "learning_rate": 4.2574310948874335e-06, "loss": 0.6886, "step": 19434 }, { "epoch": 0.7039116262223832, "grad_norm": 1.349072869343612, "learning_rate": 4.256470772995547e-06, "loss": 0.6376, "step": 19435 }, { "epoch": 0.7039478449837016, "grad_norm": 1.396943821721128, "learning_rate": 4.255510530139669e-06, "loss": 0.6954, "step": 19436 }, { "epoch": 0.7039840637450199, "grad_norm": 1.4453696291174727, "learning_rate": 4.254550366333013e-06, "loss": 0.7672, "step": 19437 }, { "epoch": 0.7040202825063383, "grad_norm": 1.3630925843163817, "learning_rate": 4.25359028158879e-06, "loss": 0.6764, "step": 19438 }, { "epoch": 0.7040565012676566, "grad_norm": 1.3261251581809306, "learning_rate": 4.252630275920218e-06, "loss": 0.6252, "step": 19439 }, { "epoch": 0.704092720028975, "grad_norm": 1.116893196699278, "learning_rate": 4.251670349340502e-06, "loss": 0.7481, "step": 19440 }, { "epoch": 0.7041289387902934, "grad_norm": 1.3734168738442711, "learning_rate": 4.25071050186285e-06, "loss": 0.6997, "step": 19441 }, { "epoch": 0.7041651575516117, "grad_norm": 1.376223477693569, "learning_rate": 4.249750733500477e-06, "loss": 0.6955, "step": 19442 }, { "epoch": 0.7042013763129301, "grad_norm": 1.4006083690276396, "learning_rate": 4.248791044266587e-06, "loss": 0.6203, "step": 19443 }, { "epoch": 0.7042375950742484, "grad_norm": 1.3020797933689316, "learning_rate": 4.247831434174384e-06, "loss": 0.6748, "step": 19444 }, { "epoch": 0.7042738138355669, "grad_norm": 1.4660039039578294, "learning_rate": 4.246871903237073e-06, "loss": 0.7413, "step": 19445 }, { "epoch": 0.7043100325968852, "grad_norm": 1.4846311247858992, "learning_rate": 4.245912451467863e-06, "loss": 0.7213, "step": 19446 }, { "epoch": 0.7043462513582035, "grad_norm": 1.3378881127271245, "learning_rate": 4.244953078879953e-06, "loss": 0.684, "step": 19447 }, { "epoch": 0.7043824701195219, "grad_norm": 1.2959203711049905, "learning_rate": 4.243993785486545e-06, "loss": 0.6213, "step": 19448 }, { "epoch": 0.7044186888808402, "grad_norm": 1.404132747874257, "learning_rate": 4.243034571300838e-06, "loss": 0.6367, "step": 19449 }, { "epoch": 0.7044549076421587, "grad_norm": 1.0710820217611254, "learning_rate": 4.2420754363360364e-06, "loss": 0.631, "step": 19450 }, { "epoch": 0.704491126403477, "grad_norm": 1.232815989967883, "learning_rate": 4.241116380605337e-06, "loss": 0.705, "step": 19451 }, { "epoch": 0.7045273451647953, "grad_norm": 1.0335985834320636, "learning_rate": 4.240157404121932e-06, "loss": 0.6618, "step": 19452 }, { "epoch": 0.7045635639261137, "grad_norm": 1.2375577225311876, "learning_rate": 4.239198506899026e-06, "loss": 0.6794, "step": 19453 }, { "epoch": 0.7045997826874321, "grad_norm": 1.2845085759963835, "learning_rate": 4.23823968894981e-06, "loss": 0.6732, "step": 19454 }, { "epoch": 0.7046360014487505, "grad_norm": 1.3890928402967422, "learning_rate": 4.237280950287479e-06, "loss": 0.6344, "step": 19455 }, { "epoch": 0.7046722202100688, "grad_norm": 1.325317571843262, "learning_rate": 4.236322290925222e-06, "loss": 0.6884, "step": 19456 }, { "epoch": 0.7047084389713871, "grad_norm": 1.3600468372990822, "learning_rate": 4.2353637108762395e-06, "loss": 0.6874, "step": 19457 }, { "epoch": 0.7047446577327056, "grad_norm": 1.287001477930266, "learning_rate": 4.234405210153716e-06, "loss": 0.6699, "step": 19458 }, { "epoch": 0.7047808764940239, "grad_norm": 1.4485301202904979, "learning_rate": 4.23344678877084e-06, "loss": 0.7273, "step": 19459 }, { "epoch": 0.7048170952553423, "grad_norm": 0.9723536847722266, "learning_rate": 4.232488446740805e-06, "loss": 0.7009, "step": 19460 }, { "epoch": 0.7048533140166606, "grad_norm": 1.4091965111926428, "learning_rate": 4.2315301840768e-06, "loss": 0.6527, "step": 19461 }, { "epoch": 0.7048895327779789, "grad_norm": 1.4176899968318986, "learning_rate": 4.230572000792006e-06, "loss": 0.7168, "step": 19462 }, { "epoch": 0.7049257515392974, "grad_norm": 1.3181388769486488, "learning_rate": 4.229613896899608e-06, "loss": 0.6334, "step": 19463 }, { "epoch": 0.7049619703006157, "grad_norm": 1.3988022554994057, "learning_rate": 4.2286558724127954e-06, "loss": 0.6738, "step": 19464 }, { "epoch": 0.7049981890619341, "grad_norm": 1.1326227270140696, "learning_rate": 4.227697927344751e-06, "loss": 0.6956, "step": 19465 }, { "epoch": 0.7050344078232524, "grad_norm": 0.9661285301655611, "learning_rate": 4.226740061708654e-06, "loss": 0.6549, "step": 19466 }, { "epoch": 0.7050706265845708, "grad_norm": 1.3892293329563035, "learning_rate": 4.225782275517683e-06, "loss": 0.7652, "step": 19467 }, { "epoch": 0.7051068453458892, "grad_norm": 1.214187029824431, "learning_rate": 4.224824568785025e-06, "loss": 0.7027, "step": 19468 }, { "epoch": 0.7051430641072075, "grad_norm": 1.2721412854431824, "learning_rate": 4.223866941523857e-06, "loss": 0.6672, "step": 19469 }, { "epoch": 0.7051792828685259, "grad_norm": 1.0025000545177176, "learning_rate": 4.222909393747349e-06, "loss": 0.7283, "step": 19470 }, { "epoch": 0.7052155016298443, "grad_norm": 1.0764734787548358, "learning_rate": 4.22195192546869e-06, "loss": 0.6667, "step": 19471 }, { "epoch": 0.7052517203911626, "grad_norm": 0.9525766207385542, "learning_rate": 4.220994536701049e-06, "loss": 0.6771, "step": 19472 }, { "epoch": 0.705287939152481, "grad_norm": 1.4267267632418679, "learning_rate": 4.220037227457601e-06, "loss": 0.7503, "step": 19473 }, { "epoch": 0.7053241579137993, "grad_norm": 1.5990922834989711, "learning_rate": 4.219079997751515e-06, "loss": 0.7625, "step": 19474 }, { "epoch": 0.7053603766751178, "grad_norm": 1.0721811109686457, "learning_rate": 4.218122847595971e-06, "loss": 0.7045, "step": 19475 }, { "epoch": 0.7053965954364361, "grad_norm": 1.3143020684082327, "learning_rate": 4.2171657770041385e-06, "loss": 0.708, "step": 19476 }, { "epoch": 0.7054328141977544, "grad_norm": 1.2634284597201269, "learning_rate": 4.216208785989186e-06, "loss": 0.6504, "step": 19477 }, { "epoch": 0.7054690329590728, "grad_norm": 1.0026460199875817, "learning_rate": 4.215251874564278e-06, "loss": 0.6068, "step": 19478 }, { "epoch": 0.7055052517203911, "grad_norm": 1.3393252540042349, "learning_rate": 4.214295042742591e-06, "loss": 0.705, "step": 19479 }, { "epoch": 0.7055414704817096, "grad_norm": 1.3252526659491188, "learning_rate": 4.213338290537289e-06, "loss": 0.6885, "step": 19480 }, { "epoch": 0.7055776892430279, "grad_norm": 1.3397023582368655, "learning_rate": 4.212381617961532e-06, "loss": 0.6895, "step": 19481 }, { "epoch": 0.7056139080043462, "grad_norm": 1.734560467924366, "learning_rate": 4.211425025028494e-06, "loss": 0.7437, "step": 19482 }, { "epoch": 0.7056501267656646, "grad_norm": 1.4546088236338501, "learning_rate": 4.210468511751333e-06, "loss": 0.6989, "step": 19483 }, { "epoch": 0.705686345526983, "grad_norm": 1.7469665865441106, "learning_rate": 4.209512078143213e-06, "loss": 0.7027, "step": 19484 }, { "epoch": 0.7057225642883014, "grad_norm": 1.3898128511447518, "learning_rate": 4.208555724217292e-06, "loss": 0.6986, "step": 19485 }, { "epoch": 0.7057587830496197, "grad_norm": 1.2337181201520306, "learning_rate": 4.207599449986734e-06, "loss": 0.7444, "step": 19486 }, { "epoch": 0.705795001810938, "grad_norm": 1.3139629137187323, "learning_rate": 4.2066432554647e-06, "loss": 0.7415, "step": 19487 }, { "epoch": 0.7058312205722564, "grad_norm": 1.4481677571910854, "learning_rate": 4.205687140664343e-06, "loss": 0.6659, "step": 19488 }, { "epoch": 0.7058674393335748, "grad_norm": 1.3682449631389604, "learning_rate": 4.20473110559882e-06, "loss": 0.7353, "step": 19489 }, { "epoch": 0.7059036580948932, "grad_norm": 1.292902071031578, "learning_rate": 4.203775150281293e-06, "loss": 0.6433, "step": 19490 }, { "epoch": 0.7059398768562115, "grad_norm": 1.4042106450735063, "learning_rate": 4.202819274724912e-06, "loss": 0.7151, "step": 19491 }, { "epoch": 0.7059760956175298, "grad_norm": 1.2428959098808823, "learning_rate": 4.201863478942833e-06, "loss": 0.624, "step": 19492 }, { "epoch": 0.7060123143788483, "grad_norm": 1.0152706130301177, "learning_rate": 4.200907762948205e-06, "loss": 0.6754, "step": 19493 }, { "epoch": 0.7060485331401666, "grad_norm": 1.401085966915406, "learning_rate": 4.199952126754184e-06, "loss": 0.6893, "step": 19494 }, { "epoch": 0.706084751901485, "grad_norm": 1.1498212974581246, "learning_rate": 4.198996570373913e-06, "loss": 0.7083, "step": 19495 }, { "epoch": 0.7061209706628033, "grad_norm": 1.4449905475097742, "learning_rate": 4.198041093820551e-06, "loss": 0.7644, "step": 19496 }, { "epoch": 0.7061571894241216, "grad_norm": 1.3956802003415352, "learning_rate": 4.197085697107241e-06, "loss": 0.7065, "step": 19497 }, { "epoch": 0.7061934081854401, "grad_norm": 1.4296549313905678, "learning_rate": 4.196130380247131e-06, "loss": 0.7068, "step": 19498 }, { "epoch": 0.7062296269467584, "grad_norm": 1.4676444567406859, "learning_rate": 4.195175143253363e-06, "loss": 0.7513, "step": 19499 }, { "epoch": 0.7062658457080768, "grad_norm": 0.9336186126928263, "learning_rate": 4.194219986139089e-06, "loss": 0.66, "step": 19500 }, { "epoch": 0.7063020644693951, "grad_norm": 0.9686683977443511, "learning_rate": 4.19326490891745e-06, "loss": 0.6983, "step": 19501 }, { "epoch": 0.7063382832307135, "grad_norm": 1.248695044038536, "learning_rate": 4.1923099116015845e-06, "loss": 0.6546, "step": 19502 }, { "epoch": 0.7063745019920319, "grad_norm": 1.3954291578461149, "learning_rate": 4.191354994204641e-06, "loss": 0.71, "step": 19503 }, { "epoch": 0.7064107207533502, "grad_norm": 1.2317525199116153, "learning_rate": 4.190400156739755e-06, "loss": 0.6127, "step": 19504 }, { "epoch": 0.7064469395146686, "grad_norm": 1.2952337292255947, "learning_rate": 4.18944539922007e-06, "loss": 0.6507, "step": 19505 }, { "epoch": 0.706483158275987, "grad_norm": 1.4556222871011504, "learning_rate": 4.188490721658717e-06, "loss": 0.6936, "step": 19506 }, { "epoch": 0.7065193770373053, "grad_norm": 1.0331705472241393, "learning_rate": 4.187536124068843e-06, "loss": 0.6414, "step": 19507 }, { "epoch": 0.7065555957986237, "grad_norm": 1.2108376654109934, "learning_rate": 4.186581606463578e-06, "loss": 0.6363, "step": 19508 }, { "epoch": 0.706591814559942, "grad_norm": 1.0612527133915353, "learning_rate": 4.185627168856058e-06, "loss": 0.7518, "step": 19509 }, { "epoch": 0.7066280333212605, "grad_norm": 1.3280811625146036, "learning_rate": 4.1846728112594145e-06, "loss": 0.7045, "step": 19510 }, { "epoch": 0.7066642520825788, "grad_norm": 1.4661600505778705, "learning_rate": 4.183718533686786e-06, "loss": 0.6363, "step": 19511 }, { "epoch": 0.7067004708438971, "grad_norm": 1.280293588894356, "learning_rate": 4.1827643361513015e-06, "loss": 0.7152, "step": 19512 }, { "epoch": 0.7067366896052155, "grad_norm": 1.2263467994435973, "learning_rate": 4.181810218666087e-06, "loss": 0.6307, "step": 19513 }, { "epoch": 0.7067729083665338, "grad_norm": 1.381346663133865, "learning_rate": 4.18085618124428e-06, "loss": 0.6626, "step": 19514 }, { "epoch": 0.7068091271278523, "grad_norm": 1.5464823470134885, "learning_rate": 4.179902223899005e-06, "loss": 0.6722, "step": 19515 }, { "epoch": 0.7068453458891706, "grad_norm": 1.2895687886330982, "learning_rate": 4.17894834664339e-06, "loss": 0.63, "step": 19516 }, { "epoch": 0.7068815646504889, "grad_norm": 1.0231281568418114, "learning_rate": 4.177994549490556e-06, "loss": 0.6767, "step": 19517 }, { "epoch": 0.7069177834118073, "grad_norm": 1.418429137725494, "learning_rate": 4.177040832453637e-06, "loss": 0.7115, "step": 19518 }, { "epoch": 0.7069540021731257, "grad_norm": 1.2745480968888216, "learning_rate": 4.176087195545753e-06, "loss": 0.7066, "step": 19519 }, { "epoch": 0.7069902209344441, "grad_norm": 1.5591240962919002, "learning_rate": 4.175133638780025e-06, "loss": 0.7429, "step": 19520 }, { "epoch": 0.7070264396957624, "grad_norm": 1.402961608128458, "learning_rate": 4.174180162169575e-06, "loss": 0.7314, "step": 19521 }, { "epoch": 0.7070626584570807, "grad_norm": 1.2803128023300856, "learning_rate": 4.173226765727527e-06, "loss": 0.7187, "step": 19522 }, { "epoch": 0.7070988772183991, "grad_norm": 1.4178862035384885, "learning_rate": 4.172273449466999e-06, "loss": 0.6484, "step": 19523 }, { "epoch": 0.7071350959797175, "grad_norm": 1.3498310377706, "learning_rate": 4.171320213401105e-06, "loss": 0.7151, "step": 19524 }, { "epoch": 0.7071713147410359, "grad_norm": 1.367988415864838, "learning_rate": 4.17036705754297e-06, "loss": 0.7, "step": 19525 }, { "epoch": 0.7072075335023542, "grad_norm": 1.3931622544082638, "learning_rate": 4.169413981905707e-06, "loss": 0.7148, "step": 19526 }, { "epoch": 0.7072437522636725, "grad_norm": 1.0744777301423631, "learning_rate": 4.168460986502431e-06, "loss": 0.7192, "step": 19527 }, { "epoch": 0.707279971024991, "grad_norm": 1.2598947473636704, "learning_rate": 4.167508071346252e-06, "loss": 0.6834, "step": 19528 }, { "epoch": 0.7073161897863093, "grad_norm": 1.1107483124488635, "learning_rate": 4.16655523645029e-06, "loss": 0.7212, "step": 19529 }, { "epoch": 0.7073524085476277, "grad_norm": 1.3606190661483641, "learning_rate": 4.165602481827654e-06, "loss": 0.7376, "step": 19530 }, { "epoch": 0.707388627308946, "grad_norm": 1.3181448634788948, "learning_rate": 4.164649807491454e-06, "loss": 0.6492, "step": 19531 }, { "epoch": 0.7074248460702643, "grad_norm": 1.078784558880186, "learning_rate": 4.163697213454798e-06, "loss": 0.6718, "step": 19532 }, { "epoch": 0.7074610648315828, "grad_norm": 1.0503514915097996, "learning_rate": 4.1627446997307984e-06, "loss": 0.6809, "step": 19533 }, { "epoch": 0.7074972835929011, "grad_norm": 1.3110261060803712, "learning_rate": 4.1617922663325615e-06, "loss": 0.6446, "step": 19534 }, { "epoch": 0.7075335023542195, "grad_norm": 1.306501825940688, "learning_rate": 4.160839913273189e-06, "loss": 0.6856, "step": 19535 }, { "epoch": 0.7075697211155378, "grad_norm": 1.3188140240603454, "learning_rate": 4.159887640565794e-06, "loss": 0.7155, "step": 19536 }, { "epoch": 0.7076059398768562, "grad_norm": 1.0401946342663848, "learning_rate": 4.158935448223477e-06, "loss": 0.6497, "step": 19537 }, { "epoch": 0.7076421586381746, "grad_norm": 0.966316531224875, "learning_rate": 4.15798333625934e-06, "loss": 0.7033, "step": 19538 }, { "epoch": 0.7076783773994929, "grad_norm": 1.3772266961631685, "learning_rate": 4.1570313046864815e-06, "loss": 0.7145, "step": 19539 }, { "epoch": 0.7077145961608113, "grad_norm": 0.9961033253232882, "learning_rate": 4.156079353518011e-06, "loss": 0.6757, "step": 19540 }, { "epoch": 0.7077508149221297, "grad_norm": 1.3448894072465725, "learning_rate": 4.155127482767024e-06, "loss": 0.7156, "step": 19541 }, { "epoch": 0.707787033683448, "grad_norm": 1.0900854250590604, "learning_rate": 4.154175692446617e-06, "loss": 0.689, "step": 19542 }, { "epoch": 0.7078232524447664, "grad_norm": 0.9322565346927146, "learning_rate": 4.153223982569886e-06, "loss": 0.7406, "step": 19543 }, { "epoch": 0.7078594712060847, "grad_norm": 1.4489166556693998, "learning_rate": 4.152272353149934e-06, "loss": 0.6298, "step": 19544 }, { "epoch": 0.7078956899674032, "grad_norm": 1.083674705069571, "learning_rate": 4.151320804199853e-06, "loss": 0.6605, "step": 19545 }, { "epoch": 0.7079319087287215, "grad_norm": 1.0269391131923387, "learning_rate": 4.150369335732734e-06, "loss": 0.7088, "step": 19546 }, { "epoch": 0.7079681274900398, "grad_norm": 1.4814979145314664, "learning_rate": 4.1494179477616745e-06, "loss": 0.7554, "step": 19547 }, { "epoch": 0.7080043462513582, "grad_norm": 1.421958910383967, "learning_rate": 4.148466640299766e-06, "loss": 0.6549, "step": 19548 }, { "epoch": 0.7080405650126765, "grad_norm": 1.4464863823504095, "learning_rate": 4.147515413360098e-06, "loss": 0.7235, "step": 19549 }, { "epoch": 0.708076783773995, "grad_norm": 1.032624053233277, "learning_rate": 4.146564266955757e-06, "loss": 0.7494, "step": 19550 }, { "epoch": 0.7081130025353133, "grad_norm": 1.2911842650124123, "learning_rate": 4.145613201099839e-06, "loss": 0.624, "step": 19551 }, { "epoch": 0.7081492212966316, "grad_norm": 1.334100923431666, "learning_rate": 4.144662215805426e-06, "loss": 0.6897, "step": 19552 }, { "epoch": 0.70818544005795, "grad_norm": 1.2619672056756441, "learning_rate": 4.143711311085606e-06, "loss": 0.6655, "step": 19553 }, { "epoch": 0.7082216588192684, "grad_norm": 1.2427687462168555, "learning_rate": 4.142760486953462e-06, "loss": 0.667, "step": 19554 }, { "epoch": 0.7082578775805868, "grad_norm": 1.3869886283453898, "learning_rate": 4.141809743422083e-06, "loss": 0.6822, "step": 19555 }, { "epoch": 0.7082940963419051, "grad_norm": 1.4180305483253681, "learning_rate": 4.140859080504549e-06, "loss": 0.7604, "step": 19556 }, { "epoch": 0.7083303151032234, "grad_norm": 1.0451422378211783, "learning_rate": 4.139908498213939e-06, "loss": 0.6711, "step": 19557 }, { "epoch": 0.7083665338645418, "grad_norm": 1.3610662334293888, "learning_rate": 4.138957996563342e-06, "loss": 0.7114, "step": 19558 }, { "epoch": 0.7084027526258602, "grad_norm": 1.3752029429027508, "learning_rate": 4.1380075755658325e-06, "loss": 0.7192, "step": 19559 }, { "epoch": 0.7084389713871786, "grad_norm": 1.3134638853064036, "learning_rate": 4.137057235234488e-06, "loss": 0.6824, "step": 19560 }, { "epoch": 0.7084751901484969, "grad_norm": 1.268151706004961, "learning_rate": 4.136106975582386e-06, "loss": 0.7063, "step": 19561 }, { "epoch": 0.7085114089098152, "grad_norm": 1.3689468902456885, "learning_rate": 4.135156796622606e-06, "loss": 0.7292, "step": 19562 }, { "epoch": 0.7085476276711337, "grad_norm": 1.4188071733037566, "learning_rate": 4.134206698368223e-06, "loss": 0.6523, "step": 19563 }, { "epoch": 0.708583846432452, "grad_norm": 0.9522003456173694, "learning_rate": 4.133256680832306e-06, "loss": 0.6549, "step": 19564 }, { "epoch": 0.7086200651937704, "grad_norm": 1.3831335007673062, "learning_rate": 4.132306744027936e-06, "loss": 0.6734, "step": 19565 }, { "epoch": 0.7086562839550887, "grad_norm": 1.2459425535522917, "learning_rate": 4.131356887968181e-06, "loss": 0.6892, "step": 19566 }, { "epoch": 0.708692502716407, "grad_norm": 2.5418573461725726, "learning_rate": 4.130407112666112e-06, "loss": 0.724, "step": 19567 }, { "epoch": 0.7087287214777255, "grad_norm": 1.3393543248966129, "learning_rate": 4.129457418134795e-06, "loss": 0.6892, "step": 19568 }, { "epoch": 0.7087649402390438, "grad_norm": 1.32788101858372, "learning_rate": 4.128507804387305e-06, "loss": 0.6657, "step": 19569 }, { "epoch": 0.7088011590003622, "grad_norm": 1.3330217982526138, "learning_rate": 4.127558271436708e-06, "loss": 0.692, "step": 19570 }, { "epoch": 0.7088373777616805, "grad_norm": 1.1567275064714524, "learning_rate": 4.126608819296069e-06, "loss": 0.6645, "step": 19571 }, { "epoch": 0.7088735965229989, "grad_norm": 1.2778141317837586, "learning_rate": 4.1256594479784485e-06, "loss": 0.705, "step": 19572 }, { "epoch": 0.7089098152843173, "grad_norm": 1.2073999924961494, "learning_rate": 4.12471015749692e-06, "loss": 0.6995, "step": 19573 }, { "epoch": 0.7089460340456356, "grad_norm": 0.9847201767383604, "learning_rate": 4.123760947864542e-06, "loss": 0.7246, "step": 19574 }, { "epoch": 0.708982252806954, "grad_norm": 1.2680599620006832, "learning_rate": 4.122811819094373e-06, "loss": 0.7011, "step": 19575 }, { "epoch": 0.7090184715682724, "grad_norm": 1.302884244090599, "learning_rate": 4.121862771199482e-06, "loss": 0.7048, "step": 19576 }, { "epoch": 0.7090546903295907, "grad_norm": 1.3896267829596458, "learning_rate": 4.120913804192924e-06, "loss": 0.6976, "step": 19577 }, { "epoch": 0.7090909090909091, "grad_norm": 1.4832968812072864, "learning_rate": 4.119964918087758e-06, "loss": 0.7789, "step": 19578 }, { "epoch": 0.7091271278522274, "grad_norm": 1.3573152088547178, "learning_rate": 4.119016112897037e-06, "loss": 0.6803, "step": 19579 }, { "epoch": 0.7091633466135459, "grad_norm": 1.0516136904378797, "learning_rate": 4.118067388633827e-06, "loss": 0.6678, "step": 19580 }, { "epoch": 0.7091995653748642, "grad_norm": 1.3256748311576894, "learning_rate": 4.117118745311176e-06, "loss": 0.6867, "step": 19581 }, { "epoch": 0.7092357841361825, "grad_norm": 1.020395364338348, "learning_rate": 4.1161701829421405e-06, "loss": 0.736, "step": 19582 }, { "epoch": 0.7092720028975009, "grad_norm": 1.4045566403471248, "learning_rate": 4.115221701539771e-06, "loss": 0.6528, "step": 19583 }, { "epoch": 0.7093082216588192, "grad_norm": 1.2802672884318416, "learning_rate": 4.1142733011171245e-06, "loss": 0.6905, "step": 19584 }, { "epoch": 0.7093444404201377, "grad_norm": 0.9604424416143438, "learning_rate": 4.113324981687249e-06, "loss": 0.6966, "step": 19585 }, { "epoch": 0.709380659181456, "grad_norm": 1.2096265110540774, "learning_rate": 4.11237674326319e-06, "loss": 0.6645, "step": 19586 }, { "epoch": 0.7094168779427743, "grad_norm": 1.3641203822694221, "learning_rate": 4.111428585858005e-06, "loss": 0.7043, "step": 19587 }, { "epoch": 0.7094530967040927, "grad_norm": 1.3796604771089969, "learning_rate": 4.110480509484737e-06, "loss": 0.7042, "step": 19588 }, { "epoch": 0.709489315465411, "grad_norm": 1.0318298107143942, "learning_rate": 4.109532514156431e-06, "loss": 0.6853, "step": 19589 }, { "epoch": 0.7095255342267295, "grad_norm": 1.1711687502979333, "learning_rate": 4.108584599886131e-06, "loss": 0.6349, "step": 19590 }, { "epoch": 0.7095617529880478, "grad_norm": 1.3348898935005997, "learning_rate": 4.107636766686887e-06, "loss": 0.6998, "step": 19591 }, { "epoch": 0.7095979717493661, "grad_norm": 1.4525865632702883, "learning_rate": 4.106689014571737e-06, "loss": 0.7594, "step": 19592 }, { "epoch": 0.7096341905106845, "grad_norm": 1.2964103976743457, "learning_rate": 4.105741343553727e-06, "loss": 0.7612, "step": 19593 }, { "epoch": 0.7096704092720029, "grad_norm": 1.0923350937823066, "learning_rate": 4.1047937536458895e-06, "loss": 0.6813, "step": 19594 }, { "epoch": 0.7097066280333213, "grad_norm": 1.16077269086084, "learning_rate": 4.103846244861273e-06, "loss": 0.683, "step": 19595 }, { "epoch": 0.7097428467946396, "grad_norm": 1.3789477490672415, "learning_rate": 4.102898817212916e-06, "loss": 0.6622, "step": 19596 }, { "epoch": 0.7097790655559579, "grad_norm": 1.0497166926077504, "learning_rate": 4.101951470713847e-06, "loss": 0.6739, "step": 19597 }, { "epoch": 0.7098152843172764, "grad_norm": 2.892595364495419, "learning_rate": 4.101004205377113e-06, "loss": 0.7682, "step": 19598 }, { "epoch": 0.7098515030785947, "grad_norm": 1.2785524043938978, "learning_rate": 4.100057021215743e-06, "loss": 0.6768, "step": 19599 }, { "epoch": 0.7098877218399131, "grad_norm": 1.4064423919571127, "learning_rate": 4.099109918242774e-06, "loss": 0.6602, "step": 19600 }, { "epoch": 0.7099239406012314, "grad_norm": 1.0487010075613437, "learning_rate": 4.098162896471234e-06, "loss": 0.6799, "step": 19601 }, { "epoch": 0.7099601593625497, "grad_norm": 1.294082812075235, "learning_rate": 4.097215955914161e-06, "loss": 0.6627, "step": 19602 }, { "epoch": 0.7099963781238682, "grad_norm": 0.9986933379748812, "learning_rate": 4.096269096584583e-06, "loss": 0.7192, "step": 19603 }, { "epoch": 0.7100325968851865, "grad_norm": 1.0556381009669558, "learning_rate": 4.095322318495531e-06, "loss": 0.668, "step": 19604 }, { "epoch": 0.7100688156465049, "grad_norm": 0.9250483141626101, "learning_rate": 4.094375621660027e-06, "loss": 0.6661, "step": 19605 }, { "epoch": 0.7101050344078232, "grad_norm": 1.3685621135908157, "learning_rate": 4.093429006091108e-06, "loss": 0.6695, "step": 19606 }, { "epoch": 0.7101412531691416, "grad_norm": 1.3039699240051692, "learning_rate": 4.092482471801795e-06, "loss": 0.6821, "step": 19607 }, { "epoch": 0.71017747193046, "grad_norm": 1.5458013006406226, "learning_rate": 4.091536018805111e-06, "loss": 0.6753, "step": 19608 }, { "epoch": 0.7102136906917783, "grad_norm": 1.2578769964955057, "learning_rate": 4.090589647114087e-06, "loss": 0.7192, "step": 19609 }, { "epoch": 0.7102499094530967, "grad_norm": 1.3398621546757428, "learning_rate": 4.0896433567417424e-06, "loss": 0.6533, "step": 19610 }, { "epoch": 0.7102861282144151, "grad_norm": 1.0015283632570189, "learning_rate": 4.088697147701097e-06, "loss": 0.6486, "step": 19611 }, { "epoch": 0.7103223469757334, "grad_norm": 1.3771207518033786, "learning_rate": 4.087751020005171e-06, "loss": 0.7134, "step": 19612 }, { "epoch": 0.7103585657370518, "grad_norm": 1.2414378698576076, "learning_rate": 4.086804973666988e-06, "loss": 0.6679, "step": 19613 }, { "epoch": 0.7103947844983701, "grad_norm": 1.351281016177546, "learning_rate": 4.085859008699565e-06, "loss": 0.7121, "step": 19614 }, { "epoch": 0.7104310032596886, "grad_norm": 1.2537543068833934, "learning_rate": 4.084913125115919e-06, "loss": 0.6391, "step": 19615 }, { "epoch": 0.7104672220210069, "grad_norm": 1.1452329942202653, "learning_rate": 4.083967322929062e-06, "loss": 0.6855, "step": 19616 }, { "epoch": 0.7105034407823252, "grad_norm": 1.2404768443945664, "learning_rate": 4.083021602152016e-06, "loss": 0.5997, "step": 19617 }, { "epoch": 0.7105396595436436, "grad_norm": 1.026492491556083, "learning_rate": 4.0820759627977925e-06, "loss": 0.6558, "step": 19618 }, { "epoch": 0.7105758783049619, "grad_norm": 1.4482006843880941, "learning_rate": 4.0811304048794e-06, "loss": 0.7504, "step": 19619 }, { "epoch": 0.7106120970662804, "grad_norm": 1.4663880485782987, "learning_rate": 4.080184928409857e-06, "loss": 0.6462, "step": 19620 }, { "epoch": 0.7106483158275987, "grad_norm": 1.6246208243712996, "learning_rate": 4.079239533402172e-06, "loss": 0.7679, "step": 19621 }, { "epoch": 0.710684534588917, "grad_norm": 1.3896419056868232, "learning_rate": 4.078294219869353e-06, "loss": 0.7138, "step": 19622 }, { "epoch": 0.7107207533502354, "grad_norm": 0.9827634344917072, "learning_rate": 4.077348987824404e-06, "loss": 0.7544, "step": 19623 }, { "epoch": 0.7107569721115538, "grad_norm": 1.2511416630784873, "learning_rate": 4.076403837280342e-06, "loss": 0.6696, "step": 19624 }, { "epoch": 0.7107931908728722, "grad_norm": 1.4003123151093326, "learning_rate": 4.0754587682501675e-06, "loss": 0.6999, "step": 19625 }, { "epoch": 0.7108294096341905, "grad_norm": 1.4143502182439323, "learning_rate": 4.074513780746884e-06, "loss": 0.6864, "step": 19626 }, { "epoch": 0.7108656283955088, "grad_norm": 1.2902600146459267, "learning_rate": 4.073568874783499e-06, "loss": 0.7256, "step": 19627 }, { "epoch": 0.7109018471568272, "grad_norm": 1.3539841172773466, "learning_rate": 4.072624050373015e-06, "loss": 0.6268, "step": 19628 }, { "epoch": 0.7109380659181456, "grad_norm": 1.4348943083478667, "learning_rate": 4.0716793075284324e-06, "loss": 0.7298, "step": 19629 }, { "epoch": 0.710974284679464, "grad_norm": 1.4145383251617494, "learning_rate": 4.0707346462627474e-06, "loss": 0.6611, "step": 19630 }, { "epoch": 0.7110105034407823, "grad_norm": 1.2465889157659096, "learning_rate": 4.069790066588966e-06, "loss": 0.7265, "step": 19631 }, { "epoch": 0.7110467222021006, "grad_norm": 1.455180979300146, "learning_rate": 4.068845568520086e-06, "loss": 0.6718, "step": 19632 }, { "epoch": 0.7110829409634191, "grad_norm": 1.3304665645272296, "learning_rate": 4.067901152069102e-06, "loss": 0.7159, "step": 19633 }, { "epoch": 0.7111191597247374, "grad_norm": 1.341620446145964, "learning_rate": 4.066956817249007e-06, "loss": 0.7481, "step": 19634 }, { "epoch": 0.7111553784860558, "grad_norm": 1.3745909774722047, "learning_rate": 4.0660125640728024e-06, "loss": 0.7826, "step": 19635 }, { "epoch": 0.7111915972473741, "grad_norm": 1.4757364752960906, "learning_rate": 4.065068392553479e-06, "loss": 0.7016, "step": 19636 }, { "epoch": 0.7112278160086924, "grad_norm": 1.333647349581648, "learning_rate": 4.064124302704027e-06, "loss": 0.6891, "step": 19637 }, { "epoch": 0.7112640347700109, "grad_norm": 1.3449273156968153, "learning_rate": 4.0631802945374435e-06, "loss": 0.6174, "step": 19638 }, { "epoch": 0.7113002535313292, "grad_norm": 0.993098604538115, "learning_rate": 4.0622363680667145e-06, "loss": 0.6284, "step": 19639 }, { "epoch": 0.7113364722926476, "grad_norm": 1.3927169265163362, "learning_rate": 4.0612925233048316e-06, "loss": 0.7055, "step": 19640 }, { "epoch": 0.7113726910539659, "grad_norm": 1.2741929278952768, "learning_rate": 4.060348760264777e-06, "loss": 0.6917, "step": 19641 }, { "epoch": 0.7114089098152843, "grad_norm": 1.3249800443271158, "learning_rate": 4.059405078959547e-06, "loss": 0.6556, "step": 19642 }, { "epoch": 0.7114451285766027, "grad_norm": 1.323898133143045, "learning_rate": 4.058461479402124e-06, "loss": 0.7316, "step": 19643 }, { "epoch": 0.711481347337921, "grad_norm": 1.17497653771506, "learning_rate": 4.0575179616054904e-06, "loss": 0.6773, "step": 19644 }, { "epoch": 0.7115175660992394, "grad_norm": 1.0357565554793087, "learning_rate": 4.0565745255826274e-06, "loss": 0.7123, "step": 19645 }, { "epoch": 0.7115537848605578, "grad_norm": 1.315582369897492, "learning_rate": 4.055631171346526e-06, "loss": 0.7122, "step": 19646 }, { "epoch": 0.7115900036218761, "grad_norm": 1.0913530567891394, "learning_rate": 4.054687898910161e-06, "loss": 0.6545, "step": 19647 }, { "epoch": 0.7116262223831945, "grad_norm": 1.2598304598780419, "learning_rate": 4.053744708286512e-06, "loss": 0.6726, "step": 19648 }, { "epoch": 0.7116624411445128, "grad_norm": 1.2349754401722957, "learning_rate": 4.052801599488565e-06, "loss": 0.6812, "step": 19649 }, { "epoch": 0.7116986599058313, "grad_norm": 1.3314131871796953, "learning_rate": 4.0518585725292924e-06, "loss": 0.706, "step": 19650 }, { "epoch": 0.7117348786671496, "grad_norm": 1.2308280872371333, "learning_rate": 4.050915627421673e-06, "loss": 0.6352, "step": 19651 }, { "epoch": 0.7117710974284679, "grad_norm": 1.421764226675166, "learning_rate": 4.049972764178679e-06, "loss": 0.7642, "step": 19652 }, { "epoch": 0.7118073161897863, "grad_norm": 1.0644452542595835, "learning_rate": 4.049029982813291e-06, "loss": 0.7551, "step": 19653 }, { "epoch": 0.7118435349511046, "grad_norm": 1.4791711553373421, "learning_rate": 4.0480872833384786e-06, "loss": 0.7144, "step": 19654 }, { "epoch": 0.7118797537124231, "grad_norm": 1.4532142529713443, "learning_rate": 4.047144665767216e-06, "loss": 0.6767, "step": 19655 }, { "epoch": 0.7119159724737414, "grad_norm": 1.2699149908960488, "learning_rate": 4.046202130112468e-06, "loss": 0.6671, "step": 19656 }, { "epoch": 0.7119521912350597, "grad_norm": 1.296736207107467, "learning_rate": 4.045259676387217e-06, "loss": 0.6539, "step": 19657 }, { "epoch": 0.7119884099963781, "grad_norm": 1.2742869204096756, "learning_rate": 4.044317304604423e-06, "loss": 0.6965, "step": 19658 }, { "epoch": 0.7120246287576965, "grad_norm": 1.090135381823886, "learning_rate": 4.043375014777052e-06, "loss": 0.6784, "step": 19659 }, { "epoch": 0.7120608475190149, "grad_norm": 1.2848751097941633, "learning_rate": 4.042432806918078e-06, "loss": 0.6694, "step": 19660 }, { "epoch": 0.7120970662803332, "grad_norm": 1.3151018269363843, "learning_rate": 4.041490681040465e-06, "loss": 0.6875, "step": 19661 }, { "epoch": 0.7121332850416515, "grad_norm": 1.4627101287723019, "learning_rate": 4.040548637157175e-06, "loss": 0.6984, "step": 19662 }, { "epoch": 0.71216950380297, "grad_norm": 1.3685827070819165, "learning_rate": 4.039606675281169e-06, "loss": 0.7075, "step": 19663 }, { "epoch": 0.7122057225642883, "grad_norm": 1.1011610145467157, "learning_rate": 4.038664795425417e-06, "loss": 0.7137, "step": 19664 }, { "epoch": 0.7122419413256067, "grad_norm": 0.958639549238831, "learning_rate": 4.037722997602873e-06, "loss": 0.6778, "step": 19665 }, { "epoch": 0.712278160086925, "grad_norm": 1.3877457878656274, "learning_rate": 4.036781281826502e-06, "loss": 0.7432, "step": 19666 }, { "epoch": 0.7123143788482433, "grad_norm": 1.3084027649163996, "learning_rate": 4.035839648109254e-06, "loss": 0.6756, "step": 19667 }, { "epoch": 0.7123505976095618, "grad_norm": 1.413083609003707, "learning_rate": 4.0348980964640995e-06, "loss": 0.7019, "step": 19668 }, { "epoch": 0.7123868163708801, "grad_norm": 1.2640341360342662, "learning_rate": 4.0339566269039864e-06, "loss": 0.6345, "step": 19669 }, { "epoch": 0.7124230351321985, "grad_norm": 1.2996096962738326, "learning_rate": 4.03301523944187e-06, "loss": 0.6954, "step": 19670 }, { "epoch": 0.7124592538935168, "grad_norm": 1.2754405999141079, "learning_rate": 4.03207393409071e-06, "loss": 0.6568, "step": 19671 }, { "epoch": 0.7124954726548351, "grad_norm": 1.2441624522727315, "learning_rate": 4.031132710863457e-06, "loss": 0.6372, "step": 19672 }, { "epoch": 0.7125316914161536, "grad_norm": 1.4130062068915008, "learning_rate": 4.030191569773062e-06, "loss": 0.6985, "step": 19673 }, { "epoch": 0.7125679101774719, "grad_norm": 1.2818970580077054, "learning_rate": 4.029250510832473e-06, "loss": 0.7418, "step": 19674 }, { "epoch": 0.7126041289387903, "grad_norm": 1.3829743790972586, "learning_rate": 4.028309534054647e-06, "loss": 0.7102, "step": 19675 }, { "epoch": 0.7126403477001086, "grad_norm": 1.391430519032646, "learning_rate": 4.027368639452529e-06, "loss": 0.7589, "step": 19676 }, { "epoch": 0.7126765664614271, "grad_norm": 1.2890131030588468, "learning_rate": 4.026427827039066e-06, "loss": 0.699, "step": 19677 }, { "epoch": 0.7127127852227454, "grad_norm": 1.0794610886915204, "learning_rate": 4.025487096827203e-06, "loss": 0.6749, "step": 19678 }, { "epoch": 0.7127490039840637, "grad_norm": 1.335306595852416, "learning_rate": 4.024546448829889e-06, "loss": 0.7321, "step": 19679 }, { "epoch": 0.7127852227453821, "grad_norm": 1.0401190933477158, "learning_rate": 4.0236058830600675e-06, "loss": 0.6831, "step": 19680 }, { "epoch": 0.7128214415067005, "grad_norm": 1.3364854303831792, "learning_rate": 4.022665399530676e-06, "loss": 0.7824, "step": 19681 }, { "epoch": 0.7128576602680189, "grad_norm": 1.448188173028894, "learning_rate": 4.021724998254665e-06, "loss": 0.7076, "step": 19682 }, { "epoch": 0.7128938790293372, "grad_norm": 1.0151396366497634, "learning_rate": 4.020784679244971e-06, "loss": 0.6974, "step": 19683 }, { "epoch": 0.7129300977906555, "grad_norm": 1.3478861087539231, "learning_rate": 4.019844442514534e-06, "loss": 0.6892, "step": 19684 }, { "epoch": 0.712966316551974, "grad_norm": 1.368800797238388, "learning_rate": 4.018904288076287e-06, "loss": 0.7133, "step": 19685 }, { "epoch": 0.7130025353132923, "grad_norm": 1.4591034555927145, "learning_rate": 4.017964215943178e-06, "loss": 0.6524, "step": 19686 }, { "epoch": 0.7130387540746107, "grad_norm": 1.4091946778851645, "learning_rate": 4.017024226128136e-06, "loss": 0.7571, "step": 19687 }, { "epoch": 0.713074972835929, "grad_norm": 1.119469918066694, "learning_rate": 4.0160843186440945e-06, "loss": 0.6838, "step": 19688 }, { "epoch": 0.7131111915972473, "grad_norm": 1.5852699924504219, "learning_rate": 4.015144493503994e-06, "loss": 0.7381, "step": 19689 }, { "epoch": 0.7131474103585658, "grad_norm": 1.5610865780472154, "learning_rate": 4.014204750720765e-06, "loss": 0.7794, "step": 19690 }, { "epoch": 0.7131836291198841, "grad_norm": 1.3480718801031377, "learning_rate": 4.013265090307338e-06, "loss": 0.7158, "step": 19691 }, { "epoch": 0.7132198478812025, "grad_norm": 1.5936092081581033, "learning_rate": 4.01232551227664e-06, "loss": 0.7594, "step": 19692 }, { "epoch": 0.7132560666425208, "grad_norm": 1.385920335046212, "learning_rate": 4.0113860166416075e-06, "loss": 0.6783, "step": 19693 }, { "epoch": 0.7132922854038392, "grad_norm": 1.4052713684427327, "learning_rate": 4.010446603415166e-06, "loss": 0.6854, "step": 19694 }, { "epoch": 0.7133285041651576, "grad_norm": 0.9250837167446515, "learning_rate": 4.009507272610241e-06, "loss": 0.703, "step": 19695 }, { "epoch": 0.7133647229264759, "grad_norm": 1.446491249216482, "learning_rate": 4.008568024239757e-06, "loss": 0.6896, "step": 19696 }, { "epoch": 0.7134009416877943, "grad_norm": 1.3528156273189533, "learning_rate": 4.007628858316645e-06, "loss": 0.7009, "step": 19697 }, { "epoch": 0.7134371604491127, "grad_norm": 0.993954951781065, "learning_rate": 4.006689774853826e-06, "loss": 0.639, "step": 19698 }, { "epoch": 0.713473379210431, "grad_norm": 1.3385002473689163, "learning_rate": 4.005750773864217e-06, "loss": 0.6843, "step": 19699 }, { "epoch": 0.7135095979717494, "grad_norm": 1.3482197405855674, "learning_rate": 4.0048118553607485e-06, "loss": 0.6775, "step": 19700 }, { "epoch": 0.7135458167330677, "grad_norm": 1.3700764603111735, "learning_rate": 4.003873019356337e-06, "loss": 0.7544, "step": 19701 }, { "epoch": 0.7135820354943861, "grad_norm": 1.2964511122578566, "learning_rate": 4.0029342658639e-06, "loss": 0.7671, "step": 19702 }, { "epoch": 0.7136182542557045, "grad_norm": 1.6459252074307518, "learning_rate": 4.001995594896358e-06, "loss": 0.673, "step": 19703 }, { "epoch": 0.7136544730170228, "grad_norm": 1.360784416589262, "learning_rate": 4.001057006466628e-06, "loss": 0.7326, "step": 19704 }, { "epoch": 0.7136906917783412, "grad_norm": 1.393366317742421, "learning_rate": 4.000118500587623e-06, "loss": 0.6414, "step": 19705 }, { "epoch": 0.7137269105396595, "grad_norm": 1.03238546206079, "learning_rate": 3.999180077272257e-06, "loss": 0.687, "step": 19706 }, { "epoch": 0.713763129300978, "grad_norm": 0.941554930114451, "learning_rate": 3.998241736533449e-06, "loss": 0.6547, "step": 19707 }, { "epoch": 0.7137993480622963, "grad_norm": 1.428172845868483, "learning_rate": 3.997303478384108e-06, "loss": 0.6792, "step": 19708 }, { "epoch": 0.7138355668236146, "grad_norm": 0.9188141125508875, "learning_rate": 3.9963653028371465e-06, "loss": 0.6372, "step": 19709 }, { "epoch": 0.713871785584933, "grad_norm": 1.3118742660340488, "learning_rate": 3.9954272099054696e-06, "loss": 0.6644, "step": 19710 }, { "epoch": 0.7139080043462513, "grad_norm": 1.3364394545770286, "learning_rate": 3.994489199601993e-06, "loss": 0.7232, "step": 19711 }, { "epoch": 0.7139442231075698, "grad_norm": 1.2539511855264687, "learning_rate": 3.993551271939623e-06, "loss": 0.6452, "step": 19712 }, { "epoch": 0.7139804418688881, "grad_norm": 1.3012028789806098, "learning_rate": 3.992613426931261e-06, "loss": 0.6913, "step": 19713 }, { "epoch": 0.7140166606302064, "grad_norm": 1.306790735263738, "learning_rate": 3.991675664589821e-06, "loss": 0.7175, "step": 19714 }, { "epoch": 0.7140528793915248, "grad_norm": 1.363373076999767, "learning_rate": 3.990737984928204e-06, "loss": 0.709, "step": 19715 }, { "epoch": 0.7140890981528432, "grad_norm": 1.373039017677887, "learning_rate": 3.989800387959311e-06, "loss": 0.6492, "step": 19716 }, { "epoch": 0.7141253169141616, "grad_norm": 2.195482196561709, "learning_rate": 3.988862873696042e-06, "loss": 0.6816, "step": 19717 }, { "epoch": 0.7141615356754799, "grad_norm": 1.0239135584445067, "learning_rate": 3.987925442151307e-06, "loss": 0.6805, "step": 19718 }, { "epoch": 0.7141977544367982, "grad_norm": 1.278558477122261, "learning_rate": 3.986988093337999e-06, "loss": 0.6646, "step": 19719 }, { "epoch": 0.7142339731981167, "grad_norm": 1.390201045893811, "learning_rate": 3.986050827269015e-06, "loss": 0.6735, "step": 19720 }, { "epoch": 0.714270191959435, "grad_norm": 1.366222076834148, "learning_rate": 3.9851136439572605e-06, "loss": 0.6386, "step": 19721 }, { "epoch": 0.7143064107207534, "grad_norm": 1.38308074303927, "learning_rate": 3.984176543415626e-06, "loss": 0.6698, "step": 19722 }, { "epoch": 0.7143426294820717, "grad_norm": 1.3035354803366164, "learning_rate": 3.983239525657009e-06, "loss": 0.6949, "step": 19723 }, { "epoch": 0.71437884824339, "grad_norm": 1.2205396634330588, "learning_rate": 3.982302590694299e-06, "loss": 0.6989, "step": 19724 }, { "epoch": 0.7144150670047085, "grad_norm": 1.2475215279285456, "learning_rate": 3.9813657385403966e-06, "loss": 0.7331, "step": 19725 }, { "epoch": 0.7144512857660268, "grad_norm": 1.417684225036361, "learning_rate": 3.980428969208191e-06, "loss": 0.7062, "step": 19726 }, { "epoch": 0.7144875045273452, "grad_norm": 1.5622249599619593, "learning_rate": 3.979492282710572e-06, "loss": 0.7193, "step": 19727 }, { "epoch": 0.7145237232886635, "grad_norm": 1.424596407192988, "learning_rate": 3.978555679060424e-06, "loss": 0.6928, "step": 19728 }, { "epoch": 0.7145599420499819, "grad_norm": 1.3651489584456162, "learning_rate": 3.977619158270646e-06, "loss": 0.6458, "step": 19729 }, { "epoch": 0.7145961608113003, "grad_norm": 1.269324502249241, "learning_rate": 3.9766827203541195e-06, "loss": 0.6906, "step": 19730 }, { "epoch": 0.7146323795726186, "grad_norm": 1.240974262735925, "learning_rate": 3.975746365323727e-06, "loss": 0.6947, "step": 19731 }, { "epoch": 0.714668598333937, "grad_norm": 1.1672492040535556, "learning_rate": 3.974810093192363e-06, "loss": 0.6834, "step": 19732 }, { "epoch": 0.7147048170952554, "grad_norm": 1.2799142076031784, "learning_rate": 3.973873903972905e-06, "loss": 0.7282, "step": 19733 }, { "epoch": 0.7147410358565737, "grad_norm": 1.3335908124899176, "learning_rate": 3.972937797678238e-06, "loss": 0.6922, "step": 19734 }, { "epoch": 0.7147772546178921, "grad_norm": 1.4589213787117223, "learning_rate": 3.972001774321238e-06, "loss": 0.6707, "step": 19735 }, { "epoch": 0.7148134733792104, "grad_norm": 1.3159415771971086, "learning_rate": 3.971065833914796e-06, "loss": 0.7197, "step": 19736 }, { "epoch": 0.7148496921405288, "grad_norm": 1.4458395551115701, "learning_rate": 3.970129976471784e-06, "loss": 0.7721, "step": 19737 }, { "epoch": 0.7148859109018472, "grad_norm": 1.120887727396701, "learning_rate": 3.969194202005081e-06, "loss": 0.712, "step": 19738 }, { "epoch": 0.7149221296631655, "grad_norm": 1.3249010463530264, "learning_rate": 3.968258510527563e-06, "loss": 0.6917, "step": 19739 }, { "epoch": 0.7149583484244839, "grad_norm": 1.4095168328338135, "learning_rate": 3.967322902052111e-06, "loss": 0.7223, "step": 19740 }, { "epoch": 0.7149945671858022, "grad_norm": 1.504998406260065, "learning_rate": 3.966387376591596e-06, "loss": 0.7091, "step": 19741 }, { "epoch": 0.7150307859471207, "grad_norm": 1.337078710606164, "learning_rate": 3.965451934158889e-06, "loss": 0.6979, "step": 19742 }, { "epoch": 0.715067004708439, "grad_norm": 1.315621793889151, "learning_rate": 3.964516574766869e-06, "loss": 0.7195, "step": 19743 }, { "epoch": 0.7151032234697573, "grad_norm": 1.3553343480890914, "learning_rate": 3.963581298428405e-06, "loss": 0.6973, "step": 19744 }, { "epoch": 0.7151394422310757, "grad_norm": 1.3116646986702831, "learning_rate": 3.9626461051563655e-06, "loss": 0.6253, "step": 19745 }, { "epoch": 0.715175660992394, "grad_norm": 1.3370567117495407, "learning_rate": 3.961710994963618e-06, "loss": 0.6756, "step": 19746 }, { "epoch": 0.7152118797537125, "grad_norm": 1.5012751116705052, "learning_rate": 3.960775967863037e-06, "loss": 0.7839, "step": 19747 }, { "epoch": 0.7152480985150308, "grad_norm": 1.4947725648828063, "learning_rate": 3.959841023867484e-06, "loss": 0.6766, "step": 19748 }, { "epoch": 0.7152843172763491, "grad_norm": 0.9679628999247452, "learning_rate": 3.958906162989825e-06, "loss": 0.6696, "step": 19749 }, { "epoch": 0.7153205360376675, "grad_norm": 1.2703637870532818, "learning_rate": 3.957971385242924e-06, "loss": 0.7075, "step": 19750 }, { "epoch": 0.7153567547989859, "grad_norm": 1.400972685628388, "learning_rate": 3.957036690639647e-06, "loss": 0.6822, "step": 19751 }, { "epoch": 0.7153929735603043, "grad_norm": 1.0197853874322045, "learning_rate": 3.956102079192856e-06, "loss": 0.6862, "step": 19752 }, { "epoch": 0.7154291923216226, "grad_norm": 1.4159395980156413, "learning_rate": 3.955167550915409e-06, "loss": 0.7571, "step": 19753 }, { "epoch": 0.7154654110829409, "grad_norm": 0.964479169094638, "learning_rate": 3.954233105820169e-06, "loss": 0.711, "step": 19754 }, { "epoch": 0.7155016298442594, "grad_norm": 1.3225006094748168, "learning_rate": 3.953298743919996e-06, "loss": 0.7175, "step": 19755 }, { "epoch": 0.7155378486055777, "grad_norm": 1.2920403583280295, "learning_rate": 3.952364465227742e-06, "loss": 0.643, "step": 19756 }, { "epoch": 0.7155740673668961, "grad_norm": 1.4438557913237875, "learning_rate": 3.951430269756267e-06, "loss": 0.6779, "step": 19757 }, { "epoch": 0.7156102861282144, "grad_norm": 1.2248743750058397, "learning_rate": 3.950496157518427e-06, "loss": 0.7048, "step": 19758 }, { "epoch": 0.7156465048895327, "grad_norm": 1.046472985171215, "learning_rate": 3.949562128527078e-06, "loss": 0.7072, "step": 19759 }, { "epoch": 0.7156827236508512, "grad_norm": 1.4072586045333357, "learning_rate": 3.948628182795069e-06, "loss": 0.73, "step": 19760 }, { "epoch": 0.7157189424121695, "grad_norm": 1.2703589800100255, "learning_rate": 3.947694320335251e-06, "loss": 0.6863, "step": 19761 }, { "epoch": 0.7157551611734879, "grad_norm": 2.563063469981396, "learning_rate": 3.946760541160479e-06, "loss": 0.6727, "step": 19762 }, { "epoch": 0.7157913799348062, "grad_norm": 1.225226465829799, "learning_rate": 3.945826845283601e-06, "loss": 0.7212, "step": 19763 }, { "epoch": 0.7158275986961246, "grad_norm": 1.0080770606316847, "learning_rate": 3.9448932327174625e-06, "loss": 0.7213, "step": 19764 }, { "epoch": 0.715863817457443, "grad_norm": 1.2994160841503668, "learning_rate": 3.943959703474917e-06, "loss": 0.6755, "step": 19765 }, { "epoch": 0.7159000362187613, "grad_norm": 1.2785290849217832, "learning_rate": 3.943026257568807e-06, "loss": 0.5783, "step": 19766 }, { "epoch": 0.7159362549800797, "grad_norm": 1.3701609361859668, "learning_rate": 3.942092895011979e-06, "loss": 0.7382, "step": 19767 }, { "epoch": 0.715972473741398, "grad_norm": 0.992981204052153, "learning_rate": 3.94115961581727e-06, "loss": 0.6582, "step": 19768 }, { "epoch": 0.7160086925027164, "grad_norm": 1.1709766764466891, "learning_rate": 3.940226419997534e-06, "loss": 0.6553, "step": 19769 }, { "epoch": 0.7160449112640348, "grad_norm": 1.223913577627753, "learning_rate": 3.9392933075656055e-06, "loss": 0.6817, "step": 19770 }, { "epoch": 0.7160811300253531, "grad_norm": 1.3865752763806491, "learning_rate": 3.938360278534327e-06, "loss": 0.7226, "step": 19771 }, { "epoch": 0.7161173487866715, "grad_norm": 1.331216513325614, "learning_rate": 3.937427332916535e-06, "loss": 0.7143, "step": 19772 }, { "epoch": 0.7161535675479899, "grad_norm": 1.5732754988550552, "learning_rate": 3.936494470725072e-06, "loss": 0.7364, "step": 19773 }, { "epoch": 0.7161897863093082, "grad_norm": 0.9982229725268965, "learning_rate": 3.935561691972773e-06, "loss": 0.7122, "step": 19774 }, { "epoch": 0.7162260050706266, "grad_norm": 1.3527634995641198, "learning_rate": 3.9346289966724715e-06, "loss": 0.6316, "step": 19775 }, { "epoch": 0.7162622238319449, "grad_norm": 1.0113550843970547, "learning_rate": 3.933696384837007e-06, "loss": 0.7391, "step": 19776 }, { "epoch": 0.7162984425932634, "grad_norm": 0.9703907269639104, "learning_rate": 3.9327638564792115e-06, "loss": 0.6688, "step": 19777 }, { "epoch": 0.7163346613545817, "grad_norm": 1.2681951365680177, "learning_rate": 3.931831411611915e-06, "loss": 0.6839, "step": 19778 }, { "epoch": 0.7163708801159, "grad_norm": 1.3039327418491427, "learning_rate": 3.930899050247948e-06, "loss": 0.7172, "step": 19779 }, { "epoch": 0.7164070988772184, "grad_norm": 1.5014587588303245, "learning_rate": 3.929966772400146e-06, "loss": 0.7276, "step": 19780 }, { "epoch": 0.7164433176385367, "grad_norm": 1.2548511263635653, "learning_rate": 3.9290345780813345e-06, "loss": 0.6385, "step": 19781 }, { "epoch": 0.7164795363998552, "grad_norm": 1.4365401562016722, "learning_rate": 3.9281024673043375e-06, "loss": 0.644, "step": 19782 }, { "epoch": 0.7165157551611735, "grad_norm": 1.1859621596379792, "learning_rate": 3.927170440081991e-06, "loss": 0.624, "step": 19783 }, { "epoch": 0.7165519739224918, "grad_norm": 1.8064269715350498, "learning_rate": 3.9262384964271135e-06, "loss": 0.7243, "step": 19784 }, { "epoch": 0.7165881926838102, "grad_norm": 1.0347807869065002, "learning_rate": 3.925306636352533e-06, "loss": 0.6925, "step": 19785 }, { "epoch": 0.7166244114451286, "grad_norm": 1.302881865895416, "learning_rate": 3.924374859871066e-06, "loss": 0.7407, "step": 19786 }, { "epoch": 0.716660630206447, "grad_norm": 1.2403954408843703, "learning_rate": 3.923443166995543e-06, "loss": 0.6714, "step": 19787 }, { "epoch": 0.7166968489677653, "grad_norm": 1.2314056893795031, "learning_rate": 3.922511557738782e-06, "loss": 0.6463, "step": 19788 }, { "epoch": 0.7167330677290836, "grad_norm": 1.318748250051666, "learning_rate": 3.921580032113602e-06, "loss": 0.6727, "step": 19789 }, { "epoch": 0.7167692864904021, "grad_norm": 1.312088339115424, "learning_rate": 3.920648590132819e-06, "loss": 0.661, "step": 19790 }, { "epoch": 0.7168055052517204, "grad_norm": 1.1433097721034506, "learning_rate": 3.919717231809255e-06, "loss": 0.7353, "step": 19791 }, { "epoch": 0.7168417240130388, "grad_norm": 1.2552232384456266, "learning_rate": 3.918785957155726e-06, "loss": 0.7376, "step": 19792 }, { "epoch": 0.7168779427743571, "grad_norm": 1.4143140303058008, "learning_rate": 3.917854766185041e-06, "loss": 0.6156, "step": 19793 }, { "epoch": 0.7169141615356754, "grad_norm": 1.1514789342139446, "learning_rate": 3.916923658910022e-06, "loss": 0.7327, "step": 19794 }, { "epoch": 0.7169503802969939, "grad_norm": 1.4708005295372009, "learning_rate": 3.9159926353434786e-06, "loss": 0.8048, "step": 19795 }, { "epoch": 0.7169865990583122, "grad_norm": 1.1150172141360652, "learning_rate": 3.915061695498224e-06, "loss": 0.6932, "step": 19796 }, { "epoch": 0.7170228178196306, "grad_norm": 1.3913754523776216, "learning_rate": 3.914130839387061e-06, "loss": 0.688, "step": 19797 }, { "epoch": 0.7170590365809489, "grad_norm": 1.4014451500343126, "learning_rate": 3.913200067022811e-06, "loss": 0.7756, "step": 19798 }, { "epoch": 0.7170952553422673, "grad_norm": 1.3194095399036323, "learning_rate": 3.912269378418275e-06, "loss": 0.6606, "step": 19799 }, { "epoch": 0.7171314741035857, "grad_norm": 1.3201152799524531, "learning_rate": 3.911338773586263e-06, "loss": 0.7148, "step": 19800 }, { "epoch": 0.717167692864904, "grad_norm": 1.4691367350382023, "learning_rate": 3.910408252539575e-06, "loss": 0.7392, "step": 19801 }, { "epoch": 0.7172039116262224, "grad_norm": 1.4093688327692435, "learning_rate": 3.909477815291023e-06, "loss": 0.6762, "step": 19802 }, { "epoch": 0.7172401303875408, "grad_norm": 1.0358604424046793, "learning_rate": 3.9085474618534095e-06, "loss": 0.6534, "step": 19803 }, { "epoch": 0.7172763491488591, "grad_norm": 1.4290514029906818, "learning_rate": 3.90761719223953e-06, "loss": 0.6987, "step": 19804 }, { "epoch": 0.7173125679101775, "grad_norm": 1.4013326600769076, "learning_rate": 3.906687006462196e-06, "loss": 0.7098, "step": 19805 }, { "epoch": 0.7173487866714958, "grad_norm": 1.3392154016798936, "learning_rate": 3.905756904534203e-06, "loss": 0.6683, "step": 19806 }, { "epoch": 0.7173850054328142, "grad_norm": 1.1992258528334425, "learning_rate": 3.904826886468349e-06, "loss": 0.7224, "step": 19807 }, { "epoch": 0.7174212241941326, "grad_norm": 1.3506565289312513, "learning_rate": 3.9038969522774295e-06, "loss": 0.637, "step": 19808 }, { "epoch": 0.7174574429554509, "grad_norm": 1.6404879711464617, "learning_rate": 3.902967101974248e-06, "loss": 0.6656, "step": 19809 }, { "epoch": 0.7174936617167693, "grad_norm": 1.470816826952888, "learning_rate": 3.902037335571597e-06, "loss": 0.7202, "step": 19810 }, { "epoch": 0.7175298804780876, "grad_norm": 1.4500595909084293, "learning_rate": 3.90110765308227e-06, "loss": 0.668, "step": 19811 }, { "epoch": 0.7175660992394061, "grad_norm": 0.9875058362524185, "learning_rate": 3.900178054519057e-06, "loss": 0.7128, "step": 19812 }, { "epoch": 0.7176023180007244, "grad_norm": 1.5334360322387461, "learning_rate": 3.899248539894756e-06, "loss": 0.6187, "step": 19813 }, { "epoch": 0.7176385367620427, "grad_norm": 1.3221110573361012, "learning_rate": 3.898319109222158e-06, "loss": 0.6857, "step": 19814 }, { "epoch": 0.7176747555233611, "grad_norm": 0.971252768327165, "learning_rate": 3.897389762514046e-06, "loss": 0.7045, "step": 19815 }, { "epoch": 0.7177109742846794, "grad_norm": 1.405445402491118, "learning_rate": 3.8964604997832165e-06, "loss": 0.6898, "step": 19816 }, { "epoch": 0.7177471930459979, "grad_norm": 1.3909073593890982, "learning_rate": 3.895531321042454e-06, "loss": 0.6753, "step": 19817 }, { "epoch": 0.7177834118073162, "grad_norm": 1.3567058005522772, "learning_rate": 3.894602226304543e-06, "loss": 0.6554, "step": 19818 }, { "epoch": 0.7178196305686345, "grad_norm": 1.255812557320747, "learning_rate": 3.893673215582269e-06, "loss": 0.6258, "step": 19819 }, { "epoch": 0.7178558493299529, "grad_norm": 1.419091324142799, "learning_rate": 3.892744288888418e-06, "loss": 0.7259, "step": 19820 }, { "epoch": 0.7178920680912713, "grad_norm": 1.1568772022017948, "learning_rate": 3.891815446235773e-06, "loss": 0.6236, "step": 19821 }, { "epoch": 0.7179282868525897, "grad_norm": 1.3045426333971712, "learning_rate": 3.8908866876371144e-06, "loss": 0.6905, "step": 19822 }, { "epoch": 0.717964505613908, "grad_norm": 1.463470097202594, "learning_rate": 3.889958013105219e-06, "loss": 0.6605, "step": 19823 }, { "epoch": 0.7180007243752263, "grad_norm": 1.3209958699216955, "learning_rate": 3.889029422652875e-06, "loss": 0.6941, "step": 19824 }, { "epoch": 0.7180369431365448, "grad_norm": 1.5152900401858422, "learning_rate": 3.888100916292855e-06, "loss": 0.708, "step": 19825 }, { "epoch": 0.7180731618978631, "grad_norm": 1.2705444892145954, "learning_rate": 3.887172494037933e-06, "loss": 0.6131, "step": 19826 }, { "epoch": 0.7181093806591815, "grad_norm": 1.4159249349031129, "learning_rate": 3.886244155900893e-06, "loss": 0.637, "step": 19827 }, { "epoch": 0.7181455994204998, "grad_norm": 1.3410258903077183, "learning_rate": 3.885315901894504e-06, "loss": 0.6677, "step": 19828 }, { "epoch": 0.7181818181818181, "grad_norm": 1.3003292508422677, "learning_rate": 3.884387732031542e-06, "loss": 0.6981, "step": 19829 }, { "epoch": 0.7182180369431366, "grad_norm": 1.0725344938186259, "learning_rate": 3.883459646324775e-06, "loss": 0.7259, "step": 19830 }, { "epoch": 0.7182542557044549, "grad_norm": 1.1522448008291035, "learning_rate": 3.882531644786981e-06, "loss": 0.7025, "step": 19831 }, { "epoch": 0.7182904744657733, "grad_norm": 1.4464155073612435, "learning_rate": 3.881603727430927e-06, "loss": 0.6444, "step": 19832 }, { "epoch": 0.7183266932270916, "grad_norm": 0.9992266973273759, "learning_rate": 3.880675894269381e-06, "loss": 0.6698, "step": 19833 }, { "epoch": 0.71836291198841, "grad_norm": 1.459677832721915, "learning_rate": 3.879748145315109e-06, "loss": 0.7781, "step": 19834 }, { "epoch": 0.7183991307497284, "grad_norm": 1.2920024392780174, "learning_rate": 3.878820480580884e-06, "loss": 0.7143, "step": 19835 }, { "epoch": 0.7184353495110467, "grad_norm": 0.9951265130972465, "learning_rate": 3.877892900079466e-06, "loss": 0.6928, "step": 19836 }, { "epoch": 0.7184715682723651, "grad_norm": 1.31784069878493, "learning_rate": 3.8769654038236195e-06, "loss": 0.6611, "step": 19837 }, { "epoch": 0.7185077870336835, "grad_norm": 1.5067185524925824, "learning_rate": 3.876037991826111e-06, "loss": 0.7767, "step": 19838 }, { "epoch": 0.7185440057950018, "grad_norm": 1.3299348574765884, "learning_rate": 3.875110664099703e-06, "loss": 0.7069, "step": 19839 }, { "epoch": 0.7185802245563202, "grad_norm": 1.3920997571107183, "learning_rate": 3.874183420657152e-06, "loss": 0.7082, "step": 19840 }, { "epoch": 0.7186164433176385, "grad_norm": 1.2930832796643081, "learning_rate": 3.873256261511217e-06, "loss": 0.6966, "step": 19841 }, { "epoch": 0.718652662078957, "grad_norm": 1.3306527317986332, "learning_rate": 3.8723291866746625e-06, "loss": 0.6499, "step": 19842 }, { "epoch": 0.7186888808402753, "grad_norm": 1.2733218677728455, "learning_rate": 3.871402196160243e-06, "loss": 0.7187, "step": 19843 }, { "epoch": 0.7187250996015936, "grad_norm": 0.9041948571189036, "learning_rate": 3.870475289980711e-06, "loss": 0.7122, "step": 19844 }, { "epoch": 0.718761318362912, "grad_norm": 1.4297410096384706, "learning_rate": 3.869548468148828e-06, "loss": 0.7525, "step": 19845 }, { "epoch": 0.7187975371242303, "grad_norm": 1.2725422449695982, "learning_rate": 3.8686217306773455e-06, "loss": 0.686, "step": 19846 }, { "epoch": 0.7188337558855488, "grad_norm": 1.0338677387453716, "learning_rate": 3.867695077579015e-06, "loss": 0.659, "step": 19847 }, { "epoch": 0.7188699746468671, "grad_norm": 1.170343354669761, "learning_rate": 3.866768508866584e-06, "loss": 0.7009, "step": 19848 }, { "epoch": 0.7189061934081854, "grad_norm": 1.0101726973759952, "learning_rate": 3.865842024552813e-06, "loss": 0.6865, "step": 19849 }, { "epoch": 0.7189424121695038, "grad_norm": 1.3157947861266572, "learning_rate": 3.864915624650444e-06, "loss": 0.7456, "step": 19850 }, { "epoch": 0.7189786309308221, "grad_norm": 1.0434328583718606, "learning_rate": 3.863989309172227e-06, "loss": 0.6291, "step": 19851 }, { "epoch": 0.7190148496921406, "grad_norm": 1.2538935535893314, "learning_rate": 3.863063078130905e-06, "loss": 0.6126, "step": 19852 }, { "epoch": 0.7190510684534589, "grad_norm": 1.4457882072522923, "learning_rate": 3.862136931539232e-06, "loss": 0.7303, "step": 19853 }, { "epoch": 0.7190872872147772, "grad_norm": 1.3285071874932164, "learning_rate": 3.861210869409947e-06, "loss": 0.7284, "step": 19854 }, { "epoch": 0.7191235059760956, "grad_norm": 1.4562074618204812, "learning_rate": 3.860284891755792e-06, "loss": 0.6587, "step": 19855 }, { "epoch": 0.719159724737414, "grad_norm": 1.0039457995699013, "learning_rate": 3.859358998589514e-06, "loss": 0.7089, "step": 19856 }, { "epoch": 0.7191959434987324, "grad_norm": 1.3029910082161913, "learning_rate": 3.858433189923852e-06, "loss": 0.7245, "step": 19857 }, { "epoch": 0.7192321622600507, "grad_norm": 0.959633029391189, "learning_rate": 3.857507465771547e-06, "loss": 0.6797, "step": 19858 }, { "epoch": 0.719268381021369, "grad_norm": 0.9062768256110985, "learning_rate": 3.856581826145333e-06, "loss": 0.6645, "step": 19859 }, { "epoch": 0.7193045997826875, "grad_norm": 1.6388191400221548, "learning_rate": 3.855656271057953e-06, "loss": 0.6298, "step": 19860 }, { "epoch": 0.7193408185440058, "grad_norm": 1.2389120648444045, "learning_rate": 3.854730800522143e-06, "loss": 0.6805, "step": 19861 }, { "epoch": 0.7193770373053242, "grad_norm": 1.1211176119567703, "learning_rate": 3.853805414550636e-06, "loss": 0.7076, "step": 19862 }, { "epoch": 0.7194132560666425, "grad_norm": 1.1864488189687397, "learning_rate": 3.852880113156163e-06, "loss": 0.7223, "step": 19863 }, { "epoch": 0.7194494748279608, "grad_norm": 1.0291594546131073, "learning_rate": 3.851954896351467e-06, "loss": 0.6436, "step": 19864 }, { "epoch": 0.7194856935892793, "grad_norm": 1.275747037263558, "learning_rate": 3.851029764149271e-06, "loss": 0.6535, "step": 19865 }, { "epoch": 0.7195219123505976, "grad_norm": 1.3423574115265346, "learning_rate": 3.850104716562307e-06, "loss": 0.7002, "step": 19866 }, { "epoch": 0.719558131111916, "grad_norm": 1.5937883463622085, "learning_rate": 3.849179753603309e-06, "loss": 0.7398, "step": 19867 }, { "epoch": 0.7195943498732343, "grad_norm": 1.2410542999283303, "learning_rate": 3.848254875285e-06, "loss": 0.6608, "step": 19868 }, { "epoch": 0.7196305686345527, "grad_norm": 1.3587151041063783, "learning_rate": 3.847330081620111e-06, "loss": 0.6756, "step": 19869 }, { "epoch": 0.7196667873958711, "grad_norm": 1.3513189575965887, "learning_rate": 3.846405372621363e-06, "loss": 0.708, "step": 19870 }, { "epoch": 0.7197030061571894, "grad_norm": 1.362401578732223, "learning_rate": 3.845480748301487e-06, "loss": 0.67, "step": 19871 }, { "epoch": 0.7197392249185078, "grad_norm": 1.2839873398298913, "learning_rate": 3.844556208673204e-06, "loss": 0.6892, "step": 19872 }, { "epoch": 0.7197754436798262, "grad_norm": 1.3063489363131455, "learning_rate": 3.8436317537492364e-06, "loss": 0.6785, "step": 19873 }, { "epoch": 0.7198116624411445, "grad_norm": 1.2866178762261768, "learning_rate": 3.842707383542301e-06, "loss": 0.7316, "step": 19874 }, { "epoch": 0.7198478812024629, "grad_norm": 1.3970783397668465, "learning_rate": 3.841783098065127e-06, "loss": 0.7557, "step": 19875 }, { "epoch": 0.7198840999637812, "grad_norm": 1.2769998115075551, "learning_rate": 3.840858897330429e-06, "loss": 0.7069, "step": 19876 }, { "epoch": 0.7199203187250997, "grad_norm": 1.1862763831991343, "learning_rate": 3.839934781350919e-06, "loss": 0.6912, "step": 19877 }, { "epoch": 0.719956537486418, "grad_norm": 1.1409873272613518, "learning_rate": 3.839010750139324e-06, "loss": 0.6119, "step": 19878 }, { "epoch": 0.7199927562477363, "grad_norm": 1.4079200007646233, "learning_rate": 3.838086803708354e-06, "loss": 0.6668, "step": 19879 }, { "epoch": 0.7200289750090547, "grad_norm": 1.273922091559509, "learning_rate": 3.837162942070725e-06, "loss": 0.7311, "step": 19880 }, { "epoch": 0.720065193770373, "grad_norm": 1.2276326935110176, "learning_rate": 3.836239165239144e-06, "loss": 0.6865, "step": 19881 }, { "epoch": 0.7201014125316915, "grad_norm": 1.3955694722888048, "learning_rate": 3.835315473226332e-06, "loss": 0.7174, "step": 19882 }, { "epoch": 0.7201376312930098, "grad_norm": 1.414827390204412, "learning_rate": 3.834391866044996e-06, "loss": 0.7013, "step": 19883 }, { "epoch": 0.7201738500543281, "grad_norm": 1.2754151727495922, "learning_rate": 3.833468343707846e-06, "loss": 0.6125, "step": 19884 }, { "epoch": 0.7202100688156465, "grad_norm": 1.3536313655628567, "learning_rate": 3.832544906227585e-06, "loss": 0.7412, "step": 19885 }, { "epoch": 0.7202462875769648, "grad_norm": 1.0367219138800117, "learning_rate": 3.831621553616931e-06, "loss": 0.6372, "step": 19886 }, { "epoch": 0.7202825063382833, "grad_norm": 1.640627991582777, "learning_rate": 3.830698285888583e-06, "loss": 0.753, "step": 19887 }, { "epoch": 0.7203187250996016, "grad_norm": 1.4520580325175525, "learning_rate": 3.8297751030552445e-06, "loss": 0.6909, "step": 19888 }, { "epoch": 0.7203549438609199, "grad_norm": 1.4906557694744211, "learning_rate": 3.828852005129625e-06, "loss": 0.696, "step": 19889 }, { "epoch": 0.7203911626222383, "grad_norm": 1.4217980908983376, "learning_rate": 3.827928992124425e-06, "loss": 0.7147, "step": 19890 }, { "epoch": 0.7204273813835567, "grad_norm": 1.3562804877902064, "learning_rate": 3.827006064052346e-06, "loss": 0.6341, "step": 19891 }, { "epoch": 0.7204636001448751, "grad_norm": 1.3005937342331986, "learning_rate": 3.8260832209260825e-06, "loss": 0.6956, "step": 19892 }, { "epoch": 0.7204998189061934, "grad_norm": 1.2788725724443786, "learning_rate": 3.825160462758343e-06, "loss": 0.6404, "step": 19893 }, { "epoch": 0.7205360376675117, "grad_norm": 1.395666524512348, "learning_rate": 3.824237789561821e-06, "loss": 0.6431, "step": 19894 }, { "epoch": 0.7205722564288302, "grad_norm": 1.350060482789546, "learning_rate": 3.823315201349214e-06, "loss": 0.7379, "step": 19895 }, { "epoch": 0.7206084751901485, "grad_norm": 1.0510745806818786, "learning_rate": 3.822392698133214e-06, "loss": 0.7078, "step": 19896 }, { "epoch": 0.7206446939514669, "grad_norm": 1.08110505133633, "learning_rate": 3.821470279926521e-06, "loss": 0.6764, "step": 19897 }, { "epoch": 0.7206809127127852, "grad_norm": 1.156258751982674, "learning_rate": 3.820547946741826e-06, "loss": 0.6988, "step": 19898 }, { "epoch": 0.7207171314741035, "grad_norm": 1.5791686910162945, "learning_rate": 3.819625698591818e-06, "loss": 0.6605, "step": 19899 }, { "epoch": 0.720753350235422, "grad_norm": 1.3420305555274987, "learning_rate": 3.818703535489195e-06, "loss": 0.7174, "step": 19900 }, { "epoch": 0.7207895689967403, "grad_norm": 1.1903115762349954, "learning_rate": 3.817781457446642e-06, "loss": 0.6939, "step": 19901 }, { "epoch": 0.7208257877580587, "grad_norm": 1.388482625489004, "learning_rate": 3.8168594644768475e-06, "loss": 0.7065, "step": 19902 }, { "epoch": 0.720862006519377, "grad_norm": 1.4745887754908837, "learning_rate": 3.8159375565924975e-06, "loss": 0.6674, "step": 19903 }, { "epoch": 0.7208982252806954, "grad_norm": 2.3897304902762384, "learning_rate": 3.815015733806283e-06, "loss": 0.7068, "step": 19904 }, { "epoch": 0.7209344440420138, "grad_norm": 1.259640011697868, "learning_rate": 3.8140939961308876e-06, "loss": 0.6672, "step": 19905 }, { "epoch": 0.7209706628033321, "grad_norm": 1.2703299401553543, "learning_rate": 3.81317234357899e-06, "loss": 0.6102, "step": 19906 }, { "epoch": 0.7210068815646505, "grad_norm": 0.9860654802279165, "learning_rate": 3.812250776163282e-06, "loss": 0.6996, "step": 19907 }, { "epoch": 0.7210431003259689, "grad_norm": 1.4780259751463831, "learning_rate": 3.8113292938964384e-06, "loss": 0.6606, "step": 19908 }, { "epoch": 0.7210793190872872, "grad_norm": 2.5338133817326196, "learning_rate": 3.810407896791143e-06, "loss": 0.696, "step": 19909 }, { "epoch": 0.7211155378486056, "grad_norm": 1.042415252137532, "learning_rate": 3.8094865848600694e-06, "loss": 0.707, "step": 19910 }, { "epoch": 0.7211517566099239, "grad_norm": 1.2668743817664765, "learning_rate": 3.8085653581159043e-06, "loss": 0.6834, "step": 19911 }, { "epoch": 0.7211879753712424, "grad_norm": 1.4538002957411977, "learning_rate": 3.807644216571319e-06, "loss": 0.8125, "step": 19912 }, { "epoch": 0.7212241941325607, "grad_norm": 1.4466166798129771, "learning_rate": 3.806723160238991e-06, "loss": 0.7001, "step": 19913 }, { "epoch": 0.721260412893879, "grad_norm": 1.4533094710755794, "learning_rate": 3.805802189131593e-06, "loss": 0.7525, "step": 19914 }, { "epoch": 0.7212966316551974, "grad_norm": 1.3043161544734443, "learning_rate": 3.8048813032618016e-06, "loss": 0.6184, "step": 19915 }, { "epoch": 0.7213328504165157, "grad_norm": 1.558826751788677, "learning_rate": 3.8039605026422856e-06, "loss": 0.7666, "step": 19916 }, { "epoch": 0.7213690691778342, "grad_norm": 1.5427353549636773, "learning_rate": 3.803039787285714e-06, "loss": 0.6387, "step": 19917 }, { "epoch": 0.7214052879391525, "grad_norm": 1.4156302466990018, "learning_rate": 3.8021191572047646e-06, "loss": 0.7215, "step": 19918 }, { "epoch": 0.7214415067004708, "grad_norm": 1.729303049163102, "learning_rate": 3.8011986124121003e-06, "loss": 0.7243, "step": 19919 }, { "epoch": 0.7214777254617892, "grad_norm": 1.2356391413691776, "learning_rate": 3.800278152920387e-06, "loss": 0.6464, "step": 19920 }, { "epoch": 0.7215139442231076, "grad_norm": 1.434346635388432, "learning_rate": 3.7993577787422964e-06, "loss": 0.7668, "step": 19921 }, { "epoch": 0.721550162984426, "grad_norm": 1.3220594385709799, "learning_rate": 3.798437489890492e-06, "loss": 0.7167, "step": 19922 }, { "epoch": 0.7215863817457443, "grad_norm": 1.2707586715183514, "learning_rate": 3.797517286377637e-06, "loss": 0.6657, "step": 19923 }, { "epoch": 0.7216226005070626, "grad_norm": 1.4072352951117701, "learning_rate": 3.7965971682163903e-06, "loss": 0.729, "step": 19924 }, { "epoch": 0.721658819268381, "grad_norm": 1.1247590076136678, "learning_rate": 3.795677135419421e-06, "loss": 0.6678, "step": 19925 }, { "epoch": 0.7216950380296994, "grad_norm": 1.4142563737035627, "learning_rate": 3.794757187999386e-06, "loss": 0.7154, "step": 19926 }, { "epoch": 0.7217312567910178, "grad_norm": 1.2412335328477284, "learning_rate": 3.7938373259689445e-06, "loss": 0.6302, "step": 19927 }, { "epoch": 0.7217674755523361, "grad_norm": 1.063941713942169, "learning_rate": 3.792917549340751e-06, "loss": 0.5831, "step": 19928 }, { "epoch": 0.7218036943136544, "grad_norm": 1.0778253789968457, "learning_rate": 3.7919978581274686e-06, "loss": 0.7109, "step": 19929 }, { "epoch": 0.7218399130749729, "grad_norm": 1.402007037306247, "learning_rate": 3.7910782523417523e-06, "loss": 0.5951, "step": 19930 }, { "epoch": 0.7218761318362912, "grad_norm": 1.5788076123623709, "learning_rate": 3.7901587319962497e-06, "loss": 0.6802, "step": 19931 }, { "epoch": 0.7219123505976096, "grad_norm": 1.2331148024008125, "learning_rate": 3.789239297103624e-06, "loss": 0.655, "step": 19932 }, { "epoch": 0.7219485693589279, "grad_norm": 0.9879706206972166, "learning_rate": 3.788319947676522e-06, "loss": 0.6415, "step": 19933 }, { "epoch": 0.7219847881202462, "grad_norm": 1.0305448951604845, "learning_rate": 3.7874006837275966e-06, "loss": 0.6993, "step": 19934 }, { "epoch": 0.7220210068815647, "grad_norm": 1.4218481468714643, "learning_rate": 3.7864815052694927e-06, "loss": 0.6575, "step": 19935 }, { "epoch": 0.722057225642883, "grad_norm": 1.4536744855425845, "learning_rate": 3.785562412314866e-06, "loss": 0.7469, "step": 19936 }, { "epoch": 0.7220934444042014, "grad_norm": 1.4312950786733678, "learning_rate": 3.7846434048763615e-06, "loss": 0.7252, "step": 19937 }, { "epoch": 0.7221296631655197, "grad_norm": 1.3756000235076158, "learning_rate": 3.7837244829666255e-06, "loss": 0.7335, "step": 19938 }, { "epoch": 0.7221658819268381, "grad_norm": 1.3713529421955046, "learning_rate": 3.7828056465982974e-06, "loss": 0.7121, "step": 19939 }, { "epoch": 0.7222021006881565, "grad_norm": 1.3807598959014147, "learning_rate": 3.7818868957840316e-06, "loss": 0.7291, "step": 19940 }, { "epoch": 0.7222383194494748, "grad_norm": 1.3130902910025612, "learning_rate": 3.780968230536466e-06, "loss": 0.6692, "step": 19941 }, { "epoch": 0.7222745382107932, "grad_norm": 1.3479670522104734, "learning_rate": 3.780049650868238e-06, "loss": 0.7282, "step": 19942 }, { "epoch": 0.7223107569721116, "grad_norm": 1.08562841485475, "learning_rate": 3.779131156791995e-06, "loss": 0.6505, "step": 19943 }, { "epoch": 0.7223469757334299, "grad_norm": 1.1007112375770236, "learning_rate": 3.7782127483203747e-06, "loss": 0.7001, "step": 19944 }, { "epoch": 0.7223831944947483, "grad_norm": 1.2923140092567782, "learning_rate": 3.7772944254660136e-06, "loss": 0.6759, "step": 19945 }, { "epoch": 0.7224194132560666, "grad_norm": 1.2865097989439054, "learning_rate": 3.776376188241545e-06, "loss": 0.6435, "step": 19946 }, { "epoch": 0.722455632017385, "grad_norm": 1.2738389960329308, "learning_rate": 3.7754580366596116e-06, "loss": 0.6491, "step": 19947 }, { "epoch": 0.7224918507787034, "grad_norm": 1.2360646040842067, "learning_rate": 3.7745399707328457e-06, "loss": 0.6586, "step": 19948 }, { "epoch": 0.7225280695400217, "grad_norm": 1.0510892657565067, "learning_rate": 3.773621990473877e-06, "loss": 0.6721, "step": 19949 }, { "epoch": 0.7225642883013401, "grad_norm": 1.330530286526167, "learning_rate": 3.772704095895343e-06, "loss": 0.6983, "step": 19950 }, { "epoch": 0.7226005070626584, "grad_norm": 1.3823510776768386, "learning_rate": 3.771786287009873e-06, "loss": 0.6584, "step": 19951 }, { "epoch": 0.7226367258239769, "grad_norm": 1.2308680991369356, "learning_rate": 3.7708685638300967e-06, "loss": 0.6573, "step": 19952 }, { "epoch": 0.7226729445852952, "grad_norm": 1.039711135809048, "learning_rate": 3.7699509263686385e-06, "loss": 0.6635, "step": 19953 }, { "epoch": 0.7227091633466135, "grad_norm": 1.4303199562551834, "learning_rate": 3.7690333746381334e-06, "loss": 0.7232, "step": 19954 }, { "epoch": 0.7227453821079319, "grad_norm": 1.3955405066619873, "learning_rate": 3.7681159086512032e-06, "loss": 0.8101, "step": 19955 }, { "epoch": 0.7227816008692503, "grad_norm": 0.9593032698012235, "learning_rate": 3.7671985284204746e-06, "loss": 0.7067, "step": 19956 }, { "epoch": 0.7228178196305687, "grad_norm": 1.333176856448592, "learning_rate": 3.7662812339585675e-06, "loss": 0.6882, "step": 19957 }, { "epoch": 0.722854038391887, "grad_norm": 1.3290836065342135, "learning_rate": 3.765364025278112e-06, "loss": 0.7042, "step": 19958 }, { "epoch": 0.7228902571532053, "grad_norm": 1.403280750717798, "learning_rate": 3.764446902391725e-06, "loss": 0.7146, "step": 19959 }, { "epoch": 0.7229264759145237, "grad_norm": 1.4076767280173914, "learning_rate": 3.7635298653120244e-06, "loss": 0.6546, "step": 19960 }, { "epoch": 0.7229626946758421, "grad_norm": 1.3742119890884894, "learning_rate": 3.7626129140516354e-06, "loss": 0.6408, "step": 19961 }, { "epoch": 0.7229989134371605, "grad_norm": 1.0704481571217024, "learning_rate": 3.761696048623175e-06, "loss": 0.6911, "step": 19962 }, { "epoch": 0.7230351321984788, "grad_norm": 1.3999158561222063, "learning_rate": 3.7607792690392576e-06, "loss": 0.6692, "step": 19963 }, { "epoch": 0.7230713509597971, "grad_norm": 1.427627823033008, "learning_rate": 3.759862575312495e-06, "loss": 0.6568, "step": 19964 }, { "epoch": 0.7231075697211156, "grad_norm": 1.4975149908891647, "learning_rate": 3.7589459674555116e-06, "loss": 0.7329, "step": 19965 }, { "epoch": 0.7231437884824339, "grad_norm": 2.484599043382612, "learning_rate": 3.758029445480915e-06, "loss": 0.6353, "step": 19966 }, { "epoch": 0.7231800072437523, "grad_norm": 1.3629401553582023, "learning_rate": 3.7571130094013175e-06, "loss": 0.6844, "step": 19967 }, { "epoch": 0.7232162260050706, "grad_norm": 1.292099732689896, "learning_rate": 3.7561966592293274e-06, "loss": 0.7021, "step": 19968 }, { "epoch": 0.723252444766389, "grad_norm": 1.2673485074784667, "learning_rate": 3.7552803949775617e-06, "loss": 0.6995, "step": 19969 }, { "epoch": 0.7232886635277074, "grad_norm": 1.101734087008231, "learning_rate": 3.754364216658624e-06, "loss": 0.6794, "step": 19970 }, { "epoch": 0.7233248822890257, "grad_norm": 1.2051494873253539, "learning_rate": 3.7534481242851193e-06, "loss": 0.748, "step": 19971 }, { "epoch": 0.7233611010503441, "grad_norm": 1.3464493668677862, "learning_rate": 3.7525321178696606e-06, "loss": 0.7006, "step": 19972 }, { "epoch": 0.7233973198116624, "grad_norm": 1.4144375676445462, "learning_rate": 3.751616197424849e-06, "loss": 0.7171, "step": 19973 }, { "epoch": 0.7234335385729808, "grad_norm": 1.354484244545418, "learning_rate": 3.750700362963289e-06, "loss": 0.6707, "step": 19974 }, { "epoch": 0.7234697573342992, "grad_norm": 1.4128055768918133, "learning_rate": 3.74978461449758e-06, "loss": 0.7396, "step": 19975 }, { "epoch": 0.7235059760956175, "grad_norm": 1.1312518137868217, "learning_rate": 3.74886895204033e-06, "loss": 0.6354, "step": 19976 }, { "epoch": 0.7235421948569359, "grad_norm": 1.2615642910709097, "learning_rate": 3.7479533756041364e-06, "loss": 0.6919, "step": 19977 }, { "epoch": 0.7235784136182543, "grad_norm": 1.4543635489083084, "learning_rate": 3.7470378852015976e-06, "loss": 0.6987, "step": 19978 }, { "epoch": 0.7236146323795726, "grad_norm": 1.1449703381876308, "learning_rate": 3.7461224808453078e-06, "loss": 0.7034, "step": 19979 }, { "epoch": 0.723650851140891, "grad_norm": 1.511226289216457, "learning_rate": 3.7452071625478716e-06, "loss": 0.7313, "step": 19980 }, { "epoch": 0.7236870699022093, "grad_norm": 1.2529484531130102, "learning_rate": 3.7442919303218806e-06, "loss": 0.645, "step": 19981 }, { "epoch": 0.7237232886635278, "grad_norm": 1.424174343866984, "learning_rate": 3.7433767841799264e-06, "loss": 0.7399, "step": 19982 }, { "epoch": 0.7237595074248461, "grad_norm": 1.321532651322195, "learning_rate": 3.7424617241346086e-06, "loss": 0.6998, "step": 19983 }, { "epoch": 0.7237957261861644, "grad_norm": 1.0183706636799075, "learning_rate": 3.7415467501985158e-06, "loss": 0.6849, "step": 19984 }, { "epoch": 0.7238319449474828, "grad_norm": 1.4656055230806744, "learning_rate": 3.7406318623842387e-06, "loss": 0.6912, "step": 19985 }, { "epoch": 0.7238681637088011, "grad_norm": 1.0636777652447105, "learning_rate": 3.739717060704363e-06, "loss": 0.6527, "step": 19986 }, { "epoch": 0.7239043824701196, "grad_norm": 1.433972877657944, "learning_rate": 3.738802345171486e-06, "loss": 0.7788, "step": 19987 }, { "epoch": 0.7239406012314379, "grad_norm": 1.3360093429276756, "learning_rate": 3.7378877157981895e-06, "loss": 0.7384, "step": 19988 }, { "epoch": 0.7239768199927562, "grad_norm": 1.2465655575660013, "learning_rate": 3.7369731725970605e-06, "loss": 0.6847, "step": 19989 }, { "epoch": 0.7240130387540746, "grad_norm": 0.9843851743656722, "learning_rate": 3.7360587155806804e-06, "loss": 0.6613, "step": 19990 }, { "epoch": 0.724049257515393, "grad_norm": 1.3317462142134049, "learning_rate": 3.735144344761641e-06, "loss": 0.6791, "step": 19991 }, { "epoch": 0.7240854762767114, "grad_norm": 1.2648353202920626, "learning_rate": 3.7342300601525184e-06, "loss": 0.6948, "step": 19992 }, { "epoch": 0.7241216950380297, "grad_norm": 1.3091886548818326, "learning_rate": 3.7333158617658937e-06, "loss": 0.7068, "step": 19993 }, { "epoch": 0.724157913799348, "grad_norm": 1.4733096375715724, "learning_rate": 3.7324017496143517e-06, "loss": 0.6744, "step": 19994 }, { "epoch": 0.7241941325606664, "grad_norm": 1.1087632671155665, "learning_rate": 3.7314877237104686e-06, "loss": 0.6913, "step": 19995 }, { "epoch": 0.7242303513219848, "grad_norm": 1.3479354354981703, "learning_rate": 3.7305737840668234e-06, "loss": 0.661, "step": 19996 }, { "epoch": 0.7242665700833032, "grad_norm": 1.3525143245817204, "learning_rate": 3.729659930695988e-06, "loss": 0.709, "step": 19997 }, { "epoch": 0.7243027888446215, "grad_norm": 1.2900101869977016, "learning_rate": 3.728746163610545e-06, "loss": 0.6566, "step": 19998 }, { "epoch": 0.7243390076059398, "grad_norm": 0.9976117424056584, "learning_rate": 3.727832482823064e-06, "loss": 0.6528, "step": 19999 }, { "epoch": 0.7243752263672583, "grad_norm": 1.3084328865701536, "learning_rate": 3.7269188883461203e-06, "loss": 0.682, "step": 20000 }, { "epoch": 0.7244114451285766, "grad_norm": 1.1773341050370747, "learning_rate": 3.726005380192281e-06, "loss": 0.7136, "step": 20001 }, { "epoch": 0.724447663889895, "grad_norm": 1.2218980891273072, "learning_rate": 3.7250919583741243e-06, "loss": 0.6578, "step": 20002 }, { "epoch": 0.7244838826512133, "grad_norm": 1.4076756072621515, "learning_rate": 3.7241786229042153e-06, "loss": 0.642, "step": 20003 }, { "epoch": 0.7245201014125316, "grad_norm": 1.3842923239245561, "learning_rate": 3.723265373795121e-06, "loss": 0.7151, "step": 20004 }, { "epoch": 0.7245563201738501, "grad_norm": 1.2566254750050896, "learning_rate": 3.7223522110594123e-06, "loss": 0.6293, "step": 20005 }, { "epoch": 0.7245925389351684, "grad_norm": 1.2943513017395576, "learning_rate": 3.7214391347096535e-06, "loss": 0.6869, "step": 20006 }, { "epoch": 0.7246287576964868, "grad_norm": 1.0403828682837335, "learning_rate": 3.7205261447584085e-06, "loss": 0.7014, "step": 20007 }, { "epoch": 0.7246649764578051, "grad_norm": 1.3588369362444055, "learning_rate": 3.7196132412182385e-06, "loss": 0.6578, "step": 20008 }, { "epoch": 0.7247011952191235, "grad_norm": 1.2344821388490608, "learning_rate": 3.718700424101712e-06, "loss": 0.6226, "step": 20009 }, { "epoch": 0.7247374139804419, "grad_norm": 1.4351487522417388, "learning_rate": 3.7177876934213874e-06, "loss": 0.657, "step": 20010 }, { "epoch": 0.7247736327417602, "grad_norm": 1.7692827045969601, "learning_rate": 3.716875049189821e-06, "loss": 0.6984, "step": 20011 }, { "epoch": 0.7248098515030786, "grad_norm": 1.3555011195734339, "learning_rate": 3.715962491419577e-06, "loss": 0.7438, "step": 20012 }, { "epoch": 0.724846070264397, "grad_norm": 1.4559985686654742, "learning_rate": 3.715050020123212e-06, "loss": 0.7205, "step": 20013 }, { "epoch": 0.7248822890257153, "grad_norm": 1.2877060580922668, "learning_rate": 3.7141376353132798e-06, "loss": 0.6419, "step": 20014 }, { "epoch": 0.7249185077870337, "grad_norm": 1.2839539579706054, "learning_rate": 3.713225337002334e-06, "loss": 0.599, "step": 20015 }, { "epoch": 0.724954726548352, "grad_norm": 1.3669560195600932, "learning_rate": 3.712313125202935e-06, "loss": 0.6964, "step": 20016 }, { "epoch": 0.7249909453096705, "grad_norm": 1.3753667499449427, "learning_rate": 3.7114009999276333e-06, "loss": 0.7303, "step": 20017 }, { "epoch": 0.7250271640709888, "grad_norm": 1.0870154629969198, "learning_rate": 3.710488961188978e-06, "loss": 0.6624, "step": 20018 }, { "epoch": 0.7250633828323071, "grad_norm": 0.9952989224301475, "learning_rate": 3.709577008999519e-06, "loss": 0.6422, "step": 20019 }, { "epoch": 0.7250996015936255, "grad_norm": 0.9318587695796464, "learning_rate": 3.7086651433718102e-06, "loss": 0.6379, "step": 20020 }, { "epoch": 0.7251358203549438, "grad_norm": 1.2194375168300469, "learning_rate": 3.7077533643183973e-06, "loss": 0.6945, "step": 20021 }, { "epoch": 0.7251720391162623, "grad_norm": 1.324112077653198, "learning_rate": 3.706841671851824e-06, "loss": 0.6268, "step": 20022 }, { "epoch": 0.7252082578775806, "grad_norm": 1.4636605875961042, "learning_rate": 3.705930065984642e-06, "loss": 0.7033, "step": 20023 }, { "epoch": 0.7252444766388989, "grad_norm": 1.0428686112869872, "learning_rate": 3.705018546729392e-06, "loss": 0.7128, "step": 20024 }, { "epoch": 0.7252806954002173, "grad_norm": 1.0972650819843641, "learning_rate": 3.7041071140986184e-06, "loss": 0.704, "step": 20025 }, { "epoch": 0.7253169141615357, "grad_norm": 1.250590011557872, "learning_rate": 3.7031957681048604e-06, "loss": 0.6848, "step": 20026 }, { "epoch": 0.7253531329228541, "grad_norm": 1.1642824261643636, "learning_rate": 3.702284508760664e-06, "loss": 0.6994, "step": 20027 }, { "epoch": 0.7253893516841724, "grad_norm": 1.4393661102640725, "learning_rate": 3.701373336078566e-06, "loss": 0.6918, "step": 20028 }, { "epoch": 0.7254255704454907, "grad_norm": 1.3368246872022818, "learning_rate": 3.700462250071107e-06, "loss": 0.6708, "step": 20029 }, { "epoch": 0.7254617892068091, "grad_norm": 1.3378680812022286, "learning_rate": 3.699551250750818e-06, "loss": 0.7513, "step": 20030 }, { "epoch": 0.7254980079681275, "grad_norm": 1.315872821493421, "learning_rate": 3.6986403381302436e-06, "loss": 0.6569, "step": 20031 }, { "epoch": 0.7255342267294459, "grad_norm": 1.2253328441985225, "learning_rate": 3.697729512221916e-06, "loss": 0.6623, "step": 20032 }, { "epoch": 0.7255704454907642, "grad_norm": 1.4115235556156667, "learning_rate": 3.696818773038363e-06, "loss": 0.7115, "step": 20033 }, { "epoch": 0.7256066642520825, "grad_norm": 1.5082137705465004, "learning_rate": 3.6959081205921267e-06, "loss": 0.6562, "step": 20034 }, { "epoch": 0.725642883013401, "grad_norm": 1.005171359501001, "learning_rate": 3.6949975548957327e-06, "loss": 0.7148, "step": 20035 }, { "epoch": 0.7256791017747193, "grad_norm": 1.4366800610538106, "learning_rate": 3.6940870759617133e-06, "loss": 0.7337, "step": 20036 }, { "epoch": 0.7257153205360377, "grad_norm": 1.3959382551566197, "learning_rate": 3.693176683802593e-06, "loss": 0.6665, "step": 20037 }, { "epoch": 0.725751539297356, "grad_norm": 1.2401532446990788, "learning_rate": 3.6922663784309064e-06, "loss": 0.6747, "step": 20038 }, { "epoch": 0.7257877580586743, "grad_norm": 1.004471744983774, "learning_rate": 3.6913561598591775e-06, "loss": 0.6724, "step": 20039 }, { "epoch": 0.7258239768199928, "grad_norm": 1.0842738961835205, "learning_rate": 3.69044602809993e-06, "loss": 0.6583, "step": 20040 }, { "epoch": 0.7258601955813111, "grad_norm": 1.2852956136836586, "learning_rate": 3.689535983165686e-06, "loss": 0.6191, "step": 20041 }, { "epoch": 0.7258964143426295, "grad_norm": 1.4256123076902003, "learning_rate": 3.6886260250689764e-06, "loss": 0.6654, "step": 20042 }, { "epoch": 0.7259326331039478, "grad_norm": 1.3671458367402791, "learning_rate": 3.6877161538223173e-06, "loss": 0.7057, "step": 20043 }, { "epoch": 0.7259688518652662, "grad_norm": 2.5063388154709805, "learning_rate": 3.686806369438227e-06, "loss": 0.779, "step": 20044 }, { "epoch": 0.7260050706265846, "grad_norm": 1.278528333543086, "learning_rate": 3.685896671929232e-06, "loss": 0.6419, "step": 20045 }, { "epoch": 0.7260412893879029, "grad_norm": 1.1007566007138854, "learning_rate": 3.6849870613078476e-06, "loss": 0.7074, "step": 20046 }, { "epoch": 0.7260775081492213, "grad_norm": 1.2464719933708086, "learning_rate": 3.684077537586589e-06, "loss": 0.5615, "step": 20047 }, { "epoch": 0.7261137269105397, "grad_norm": 1.431957783508049, "learning_rate": 3.6831681007779698e-06, "loss": 0.774, "step": 20048 }, { "epoch": 0.726149945671858, "grad_norm": 1.2441895653457504, "learning_rate": 3.6822587508945116e-06, "loss": 0.6625, "step": 20049 }, { "epoch": 0.7261861644331764, "grad_norm": 1.3485347192821446, "learning_rate": 3.681349487948724e-06, "loss": 0.6467, "step": 20050 }, { "epoch": 0.7262223831944947, "grad_norm": 1.283034409357523, "learning_rate": 3.6804403119531206e-06, "loss": 0.679, "step": 20051 }, { "epoch": 0.7262586019558132, "grad_norm": 1.210491741336798, "learning_rate": 3.679531222920206e-06, "loss": 0.6639, "step": 20052 }, { "epoch": 0.7262948207171315, "grad_norm": 1.3163174336069583, "learning_rate": 3.678622220862501e-06, "loss": 0.6946, "step": 20053 }, { "epoch": 0.7263310394784498, "grad_norm": 0.9702183882250688, "learning_rate": 3.6777133057925076e-06, "loss": 0.6732, "step": 20054 }, { "epoch": 0.7263672582397682, "grad_norm": 1.0225598184183688, "learning_rate": 3.6768044777227306e-06, "loss": 0.6549, "step": 20055 }, { "epoch": 0.7264034770010865, "grad_norm": 1.3806115282786415, "learning_rate": 3.6758957366656832e-06, "loss": 0.689, "step": 20056 }, { "epoch": 0.726439695762405, "grad_norm": 1.0772744595290369, "learning_rate": 3.6749870826338685e-06, "loss": 0.672, "step": 20057 }, { "epoch": 0.7264759145237233, "grad_norm": 1.4419949151377203, "learning_rate": 3.6740785156397883e-06, "loss": 0.6889, "step": 20058 }, { "epoch": 0.7265121332850416, "grad_norm": 1.0699394281888734, "learning_rate": 3.673170035695943e-06, "loss": 0.7395, "step": 20059 }, { "epoch": 0.72654835204636, "grad_norm": 1.3666613274732633, "learning_rate": 3.672261642814842e-06, "loss": 0.7488, "step": 20060 }, { "epoch": 0.7265845708076784, "grad_norm": 0.9627406461594915, "learning_rate": 3.671353337008979e-06, "loss": 0.6888, "step": 20061 }, { "epoch": 0.7266207895689968, "grad_norm": 1.301697077840335, "learning_rate": 3.6704451182908517e-06, "loss": 0.6158, "step": 20062 }, { "epoch": 0.7266570083303151, "grad_norm": 1.3378368074494, "learning_rate": 3.6695369866729646e-06, "loss": 0.7306, "step": 20063 }, { "epoch": 0.7266932270916334, "grad_norm": 1.3897956775502, "learning_rate": 3.668628942167812e-06, "loss": 0.6318, "step": 20064 }, { "epoch": 0.7267294458529518, "grad_norm": 1.405005818559436, "learning_rate": 3.667720984787888e-06, "loss": 0.6982, "step": 20065 }, { "epoch": 0.7267656646142702, "grad_norm": 1.0504463088076474, "learning_rate": 3.6668131145456833e-06, "loss": 0.6929, "step": 20066 }, { "epoch": 0.7268018833755886, "grad_norm": 1.3094009287026327, "learning_rate": 3.6659053314536987e-06, "loss": 0.6855, "step": 20067 }, { "epoch": 0.7268381021369069, "grad_norm": 1.3592209605383074, "learning_rate": 3.6649976355244223e-06, "loss": 0.7333, "step": 20068 }, { "epoch": 0.7268743208982252, "grad_norm": 1.458164393199031, "learning_rate": 3.664090026770345e-06, "loss": 0.7059, "step": 20069 }, { "epoch": 0.7269105396595437, "grad_norm": 1.2962560383077593, "learning_rate": 3.6631825052039527e-06, "loss": 0.7647, "step": 20070 }, { "epoch": 0.726946758420862, "grad_norm": 1.1137212115044992, "learning_rate": 3.6622750708377406e-06, "loss": 0.7206, "step": 20071 }, { "epoch": 0.7269829771821804, "grad_norm": 1.3328503554134594, "learning_rate": 3.6613677236841927e-06, "loss": 0.6782, "step": 20072 }, { "epoch": 0.7270191959434987, "grad_norm": 1.2530750501040742, "learning_rate": 3.6604604637557908e-06, "loss": 0.6817, "step": 20073 }, { "epoch": 0.727055414704817, "grad_norm": 1.6319560013073267, "learning_rate": 3.6595532910650265e-06, "loss": 0.6896, "step": 20074 }, { "epoch": 0.7270916334661355, "grad_norm": 1.430323804421454, "learning_rate": 3.65864620562438e-06, "loss": 0.6886, "step": 20075 }, { "epoch": 0.7271278522274538, "grad_norm": 1.30202520018188, "learning_rate": 3.6577392074463336e-06, "loss": 0.6583, "step": 20076 }, { "epoch": 0.7271640709887722, "grad_norm": 1.3286459728454234, "learning_rate": 3.6568322965433654e-06, "loss": 0.6692, "step": 20077 }, { "epoch": 0.7272002897500905, "grad_norm": 1.2873059290557416, "learning_rate": 3.6559254729279624e-06, "loss": 0.7234, "step": 20078 }, { "epoch": 0.7272365085114089, "grad_norm": 1.1704706459662781, "learning_rate": 3.6550187366125987e-06, "loss": 0.6862, "step": 20079 }, { "epoch": 0.7272727272727273, "grad_norm": 1.356875188777863, "learning_rate": 3.6541120876097525e-06, "loss": 0.6316, "step": 20080 }, { "epoch": 0.7273089460340456, "grad_norm": 1.157758652611558, "learning_rate": 3.6532055259318964e-06, "loss": 0.7164, "step": 20081 }, { "epoch": 0.727345164795364, "grad_norm": 1.4953293877129314, "learning_rate": 3.6522990515915125e-06, "loss": 0.7533, "step": 20082 }, { "epoch": 0.7273813835566824, "grad_norm": 0.9691746671036808, "learning_rate": 3.651392664601071e-06, "loss": 0.7023, "step": 20083 }, { "epoch": 0.7274176023180007, "grad_norm": 1.8103669374467226, "learning_rate": 3.6504863649730415e-06, "loss": 0.7746, "step": 20084 }, { "epoch": 0.7274538210793191, "grad_norm": 1.2553044757356264, "learning_rate": 3.649580152719903e-06, "loss": 0.6468, "step": 20085 }, { "epoch": 0.7274900398406374, "grad_norm": 1.3789623467919154, "learning_rate": 3.648674027854121e-06, "loss": 0.6763, "step": 20086 }, { "epoch": 0.7275262586019559, "grad_norm": 1.6480147612924625, "learning_rate": 3.6477679903881645e-06, "loss": 0.6739, "step": 20087 }, { "epoch": 0.7275624773632742, "grad_norm": 1.301861921241067, "learning_rate": 3.646862040334499e-06, "loss": 0.749, "step": 20088 }, { "epoch": 0.7275986961245925, "grad_norm": 1.4217029221128783, "learning_rate": 3.6459561777055984e-06, "loss": 0.6319, "step": 20089 }, { "epoch": 0.7276349148859109, "grad_norm": 1.2520727072214342, "learning_rate": 3.645050402513923e-06, "loss": 0.642, "step": 20090 }, { "epoch": 0.7276711336472292, "grad_norm": 0.956885717448383, "learning_rate": 3.6441447147719377e-06, "loss": 0.6976, "step": 20091 }, { "epoch": 0.7277073524085477, "grad_norm": 0.9478185620638869, "learning_rate": 3.643239114492103e-06, "loss": 0.7008, "step": 20092 }, { "epoch": 0.727743571169866, "grad_norm": 1.1862344838361245, "learning_rate": 3.6423336016868883e-06, "loss": 0.6867, "step": 20093 }, { "epoch": 0.7277797899311843, "grad_norm": 1.4792205051213456, "learning_rate": 3.6414281763687474e-06, "loss": 0.743, "step": 20094 }, { "epoch": 0.7278160086925027, "grad_norm": 1.1310748258332564, "learning_rate": 3.64052283855014e-06, "loss": 0.7201, "step": 20095 }, { "epoch": 0.727852227453821, "grad_norm": 1.307630182761473, "learning_rate": 3.6396175882435293e-06, "loss": 0.7349, "step": 20096 }, { "epoch": 0.7278884462151395, "grad_norm": 1.0052296791230466, "learning_rate": 3.638712425461369e-06, "loss": 0.7162, "step": 20097 }, { "epoch": 0.7279246649764578, "grad_norm": 1.2834763997997902, "learning_rate": 3.637807350216115e-06, "loss": 0.6124, "step": 20098 }, { "epoch": 0.7279608837377761, "grad_norm": 1.265936107574701, "learning_rate": 3.6369023625202194e-06, "loss": 0.7046, "step": 20099 }, { "epoch": 0.7279971024990946, "grad_norm": 1.1989522292301602, "learning_rate": 3.6359974623861415e-06, "loss": 0.716, "step": 20100 }, { "epoch": 0.7280333212604129, "grad_norm": 1.3794294437277588, "learning_rate": 3.6350926498263307e-06, "loss": 0.632, "step": 20101 }, { "epoch": 0.7280695400217313, "grad_norm": 1.2355084353572057, "learning_rate": 3.6341879248532376e-06, "loss": 0.6692, "step": 20102 }, { "epoch": 0.7281057587830496, "grad_norm": 1.1152196049286305, "learning_rate": 3.633283287479309e-06, "loss": 0.7496, "step": 20103 }, { "epoch": 0.7281419775443679, "grad_norm": 1.462283541991551, "learning_rate": 3.6323787377169994e-06, "loss": 0.6991, "step": 20104 }, { "epoch": 0.7281781963056864, "grad_norm": 1.4267943571451038, "learning_rate": 3.6314742755787537e-06, "loss": 0.7028, "step": 20105 }, { "epoch": 0.7282144150670047, "grad_norm": 1.1184422305230373, "learning_rate": 3.6305699010770148e-06, "loss": 0.7074, "step": 20106 }, { "epoch": 0.7282506338283231, "grad_norm": 1.4481032562753462, "learning_rate": 3.6296656142242335e-06, "loss": 0.7443, "step": 20107 }, { "epoch": 0.7282868525896414, "grad_norm": 1.086339618395045, "learning_rate": 3.628761415032851e-06, "loss": 0.7244, "step": 20108 }, { "epoch": 0.7283230713509597, "grad_norm": 0.8945645157357945, "learning_rate": 3.6278573035153097e-06, "loss": 0.7071, "step": 20109 }, { "epoch": 0.7283592901122782, "grad_norm": 1.060743869045458, "learning_rate": 3.626953279684048e-06, "loss": 0.6805, "step": 20110 }, { "epoch": 0.7283955088735965, "grad_norm": 1.3250749761077205, "learning_rate": 3.626049343551512e-06, "loss": 0.7296, "step": 20111 }, { "epoch": 0.7284317276349149, "grad_norm": 1.3399063393234507, "learning_rate": 3.625145495130138e-06, "loss": 0.6758, "step": 20112 }, { "epoch": 0.7284679463962332, "grad_norm": 1.0312625621668112, "learning_rate": 3.6242417344323633e-06, "loss": 0.6832, "step": 20113 }, { "epoch": 0.7285041651575516, "grad_norm": 1.4744340842945896, "learning_rate": 3.6233380614706217e-06, "loss": 0.6978, "step": 20114 }, { "epoch": 0.72854038391887, "grad_norm": 1.0410172065907144, "learning_rate": 3.622434476257354e-06, "loss": 0.6381, "step": 20115 }, { "epoch": 0.7285766026801883, "grad_norm": 1.2531139895319308, "learning_rate": 3.6215309788049912e-06, "loss": 0.6395, "step": 20116 }, { "epoch": 0.7286128214415067, "grad_norm": 1.0551654264298576, "learning_rate": 3.6206275691259653e-06, "loss": 0.7015, "step": 20117 }, { "epoch": 0.7286490402028251, "grad_norm": 1.141948289057657, "learning_rate": 3.619724247232711e-06, "loss": 0.6788, "step": 20118 }, { "epoch": 0.7286852589641434, "grad_norm": 1.3747051035217221, "learning_rate": 3.618821013137658e-06, "loss": 0.676, "step": 20119 }, { "epoch": 0.7287214777254618, "grad_norm": 1.4491446057089696, "learning_rate": 3.6179178668532344e-06, "loss": 0.7152, "step": 20120 }, { "epoch": 0.7287576964867801, "grad_norm": 1.2834525156303773, "learning_rate": 3.6170148083918656e-06, "loss": 0.6789, "step": 20121 }, { "epoch": 0.7287939152480986, "grad_norm": 1.3410347852440965, "learning_rate": 3.6161118377659855e-06, "loss": 0.6615, "step": 20122 }, { "epoch": 0.7288301340094169, "grad_norm": 1.4994240154131309, "learning_rate": 3.6152089549880154e-06, "loss": 0.6909, "step": 20123 }, { "epoch": 0.7288663527707352, "grad_norm": 1.458800336166255, "learning_rate": 3.61430616007038e-06, "loss": 0.7904, "step": 20124 }, { "epoch": 0.7289025715320536, "grad_norm": 1.2541930466360223, "learning_rate": 3.6134034530255034e-06, "loss": 0.5839, "step": 20125 }, { "epoch": 0.7289387902933719, "grad_norm": 1.2273614196488845, "learning_rate": 3.6125008338658075e-06, "loss": 0.6733, "step": 20126 }, { "epoch": 0.7289750090546904, "grad_norm": 1.8355817835724002, "learning_rate": 3.611598302603708e-06, "loss": 0.6987, "step": 20127 }, { "epoch": 0.7290112278160087, "grad_norm": 1.0693441191723898, "learning_rate": 3.6106958592516326e-06, "loss": 0.6674, "step": 20128 }, { "epoch": 0.729047446577327, "grad_norm": 1.3920609515673101, "learning_rate": 3.6097935038219976e-06, "loss": 0.7637, "step": 20129 }, { "epoch": 0.7290836653386454, "grad_norm": 1.3412017043676243, "learning_rate": 3.6088912363272176e-06, "loss": 0.6674, "step": 20130 }, { "epoch": 0.7291198840999638, "grad_norm": 0.9925948715518997, "learning_rate": 3.607989056779707e-06, "loss": 0.6873, "step": 20131 }, { "epoch": 0.7291561028612822, "grad_norm": 0.9827739663389709, "learning_rate": 3.607086965191887e-06, "loss": 0.7056, "step": 20132 }, { "epoch": 0.7291923216226005, "grad_norm": 1.3893226448054048, "learning_rate": 3.6061849615761678e-06, "loss": 0.7141, "step": 20133 }, { "epoch": 0.7292285403839188, "grad_norm": 1.4103849488413926, "learning_rate": 3.6052830459449616e-06, "loss": 0.6473, "step": 20134 }, { "epoch": 0.7292647591452373, "grad_norm": 1.339916577598105, "learning_rate": 3.6043812183106764e-06, "loss": 0.6734, "step": 20135 }, { "epoch": 0.7293009779065556, "grad_norm": 1.3858717507675828, "learning_rate": 3.6034794786857284e-06, "loss": 0.7102, "step": 20136 }, { "epoch": 0.729337196667874, "grad_norm": 1.3533053406800986, "learning_rate": 3.6025778270825228e-06, "loss": 0.6986, "step": 20137 }, { "epoch": 0.7293734154291923, "grad_norm": 1.2850324558794763, "learning_rate": 3.601676263513465e-06, "loss": 0.6964, "step": 20138 }, { "epoch": 0.7294096341905106, "grad_norm": 1.402842298231784, "learning_rate": 3.600774787990967e-06, "loss": 0.6772, "step": 20139 }, { "epoch": 0.7294458529518291, "grad_norm": 1.3368913970430352, "learning_rate": 3.5998734005274306e-06, "loss": 0.6411, "step": 20140 }, { "epoch": 0.7294820717131474, "grad_norm": 1.2956343031642166, "learning_rate": 3.59897210113526e-06, "loss": 0.6832, "step": 20141 }, { "epoch": 0.7295182904744658, "grad_norm": 1.4307579165975703, "learning_rate": 3.5980708898268533e-06, "loss": 0.659, "step": 20142 }, { "epoch": 0.7295545092357841, "grad_norm": 1.336342878352229, "learning_rate": 3.597169766614621e-06, "loss": 0.7317, "step": 20143 }, { "epoch": 0.7295907279971025, "grad_norm": 1.5034250399852847, "learning_rate": 3.596268731510959e-06, "loss": 0.7217, "step": 20144 }, { "epoch": 0.7296269467584209, "grad_norm": 1.0367258268752961, "learning_rate": 3.595367784528265e-06, "loss": 0.6841, "step": 20145 }, { "epoch": 0.7296631655197392, "grad_norm": 1.033452594686115, "learning_rate": 3.5944669256789357e-06, "loss": 0.6897, "step": 20146 }, { "epoch": 0.7296993842810576, "grad_norm": 1.2913058924176133, "learning_rate": 3.593566154975372e-06, "loss": 0.683, "step": 20147 }, { "epoch": 0.729735603042376, "grad_norm": 1.3892059606901546, "learning_rate": 3.592665472429968e-06, "loss": 0.7087, "step": 20148 }, { "epoch": 0.7297718218036943, "grad_norm": 1.310373617450891, "learning_rate": 3.591764878055113e-06, "loss": 0.6423, "step": 20149 }, { "epoch": 0.7298080405650127, "grad_norm": 1.297745873286216, "learning_rate": 3.5908643718632086e-06, "loss": 0.6948, "step": 20150 }, { "epoch": 0.729844259326331, "grad_norm": 1.322459168687189, "learning_rate": 3.589963953866641e-06, "loss": 0.6489, "step": 20151 }, { "epoch": 0.7298804780876494, "grad_norm": 1.3701082862842195, "learning_rate": 3.589063624077802e-06, "loss": 0.6663, "step": 20152 }, { "epoch": 0.7299166968489678, "grad_norm": 1.343717479035049, "learning_rate": 3.588163382509078e-06, "loss": 0.6309, "step": 20153 }, { "epoch": 0.7299529156102861, "grad_norm": 1.3200084234550502, "learning_rate": 3.5872632291728615e-06, "loss": 0.6552, "step": 20154 }, { "epoch": 0.7299891343716045, "grad_norm": 1.4423211716670057, "learning_rate": 3.5863631640815387e-06, "loss": 0.6823, "step": 20155 }, { "epoch": 0.7300253531329228, "grad_norm": 1.4699918178073967, "learning_rate": 3.5854631872474943e-06, "loss": 0.6364, "step": 20156 }, { "epoch": 0.7300615718942413, "grad_norm": 1.5775585143370792, "learning_rate": 3.5845632986831092e-06, "loss": 0.7166, "step": 20157 }, { "epoch": 0.7300977906555596, "grad_norm": 1.5454854455720424, "learning_rate": 3.5836634984007745e-06, "loss": 0.6939, "step": 20158 }, { "epoch": 0.7301340094168779, "grad_norm": 1.3510304917130946, "learning_rate": 3.582763786412867e-06, "loss": 0.7315, "step": 20159 }, { "epoch": 0.7301702281781963, "grad_norm": 1.3987472313716307, "learning_rate": 3.581864162731765e-06, "loss": 0.7665, "step": 20160 }, { "epoch": 0.7302064469395146, "grad_norm": 1.2500692907142443, "learning_rate": 3.580964627369855e-06, "loss": 0.6433, "step": 20161 }, { "epoch": 0.7302426657008331, "grad_norm": 1.219131681237726, "learning_rate": 3.5800651803395126e-06, "loss": 0.6052, "step": 20162 }, { "epoch": 0.7302788844621514, "grad_norm": 1.063426972851839, "learning_rate": 3.5791658216531145e-06, "loss": 0.6485, "step": 20163 }, { "epoch": 0.7303151032234697, "grad_norm": 0.978944875389391, "learning_rate": 3.578266551323032e-06, "loss": 0.6285, "step": 20164 }, { "epoch": 0.7303513219847881, "grad_norm": 1.5381202899790218, "learning_rate": 3.5773673693616483e-06, "loss": 0.7076, "step": 20165 }, { "epoch": 0.7303875407461065, "grad_norm": 0.9373383854601078, "learning_rate": 3.576468275781333e-06, "loss": 0.6613, "step": 20166 }, { "epoch": 0.7304237595074249, "grad_norm": 1.4682324593371652, "learning_rate": 3.5755692705944556e-06, "loss": 0.7063, "step": 20167 }, { "epoch": 0.7304599782687432, "grad_norm": 1.5775438928996561, "learning_rate": 3.5746703538133932e-06, "loss": 0.7174, "step": 20168 }, { "epoch": 0.7304961970300615, "grad_norm": 1.2824263976247134, "learning_rate": 3.573771525450512e-06, "loss": 0.7065, "step": 20169 }, { "epoch": 0.73053241579138, "grad_norm": 1.358352545861853, "learning_rate": 3.572872785518182e-06, "loss": 0.6466, "step": 20170 }, { "epoch": 0.7305686345526983, "grad_norm": 1.352598384561569, "learning_rate": 3.5719741340287663e-06, "loss": 0.6512, "step": 20171 }, { "epoch": 0.7306048533140167, "grad_norm": 1.35261584108757, "learning_rate": 3.5710755709946366e-06, "loss": 0.6582, "step": 20172 }, { "epoch": 0.730641072075335, "grad_norm": 1.4124122818414748, "learning_rate": 3.570177096428158e-06, "loss": 0.729, "step": 20173 }, { "epoch": 0.7306772908366533, "grad_norm": 1.3845426113539447, "learning_rate": 3.5692787103416913e-06, "loss": 0.7166, "step": 20174 }, { "epoch": 0.7307135095979718, "grad_norm": 1.5396250674506413, "learning_rate": 3.5683804127475965e-06, "loss": 0.6587, "step": 20175 }, { "epoch": 0.7307497283592901, "grad_norm": 1.3180836742485587, "learning_rate": 3.5674822036582425e-06, "loss": 0.6889, "step": 20176 }, { "epoch": 0.7307859471206085, "grad_norm": 1.4375418671181297, "learning_rate": 3.5665840830859843e-06, "loss": 0.7342, "step": 20177 }, { "epoch": 0.7308221658819268, "grad_norm": 1.4924016893658512, "learning_rate": 3.56568605104318e-06, "loss": 0.7111, "step": 20178 }, { "epoch": 0.7308583846432452, "grad_norm": 1.0952464193802278, "learning_rate": 3.564788107542192e-06, "loss": 0.6898, "step": 20179 }, { "epoch": 0.7308946034045636, "grad_norm": 1.3135765207820804, "learning_rate": 3.5638902525953745e-06, "loss": 0.712, "step": 20180 }, { "epoch": 0.7309308221658819, "grad_norm": 1.5457412700227875, "learning_rate": 3.562992486215082e-06, "loss": 0.6905, "step": 20181 }, { "epoch": 0.7309670409272003, "grad_norm": 1.4460203576694497, "learning_rate": 3.5620948084136663e-06, "loss": 0.7005, "step": 20182 }, { "epoch": 0.7310032596885186, "grad_norm": 1.35823146430098, "learning_rate": 3.561197219203486e-06, "loss": 0.6901, "step": 20183 }, { "epoch": 0.731039478449837, "grad_norm": 1.3715427993318652, "learning_rate": 3.560299718596889e-06, "loss": 0.7381, "step": 20184 }, { "epoch": 0.7310756972111554, "grad_norm": 1.4560252429075269, "learning_rate": 3.559402306606227e-06, "loss": 0.6332, "step": 20185 }, { "epoch": 0.7311119159724737, "grad_norm": 1.3780361191207464, "learning_rate": 3.5585049832438446e-06, "loss": 0.7524, "step": 20186 }, { "epoch": 0.7311481347337921, "grad_norm": 1.2760472484064314, "learning_rate": 3.5576077485220973e-06, "loss": 0.738, "step": 20187 }, { "epoch": 0.7311843534951105, "grad_norm": 1.2314159758882794, "learning_rate": 3.5567106024533273e-06, "loss": 0.5795, "step": 20188 }, { "epoch": 0.7312205722564288, "grad_norm": 1.3628497000858704, "learning_rate": 3.5558135450498786e-06, "loss": 0.6582, "step": 20189 }, { "epoch": 0.7312567910177472, "grad_norm": 1.33905337998535, "learning_rate": 3.554916576324101e-06, "loss": 0.6672, "step": 20190 }, { "epoch": 0.7312930097790655, "grad_norm": 1.2595590821326133, "learning_rate": 3.5540196962883343e-06, "loss": 0.637, "step": 20191 }, { "epoch": 0.731329228540384, "grad_norm": 1.3233947004425757, "learning_rate": 3.553122904954921e-06, "loss": 0.6303, "step": 20192 }, { "epoch": 0.7313654473017023, "grad_norm": 1.4821429675037767, "learning_rate": 3.5522262023361963e-06, "loss": 0.6447, "step": 20193 }, { "epoch": 0.7314016660630206, "grad_norm": 1.123307570385422, "learning_rate": 3.5513295884445097e-06, "loss": 0.6889, "step": 20194 }, { "epoch": 0.731437884824339, "grad_norm": 1.115261282049262, "learning_rate": 3.5504330632921934e-06, "loss": 0.6957, "step": 20195 }, { "epoch": 0.7314741035856573, "grad_norm": 1.0626628294670035, "learning_rate": 3.5495366268915853e-06, "loss": 0.6783, "step": 20196 }, { "epoch": 0.7315103223469758, "grad_norm": 1.8229677773304092, "learning_rate": 3.548640279255019e-06, "loss": 0.6911, "step": 20197 }, { "epoch": 0.7315465411082941, "grad_norm": 1.1879889891560016, "learning_rate": 3.5477440203948333e-06, "loss": 0.7026, "step": 20198 }, { "epoch": 0.7315827598696124, "grad_norm": 1.1616562304448148, "learning_rate": 3.5468478503233593e-06, "loss": 0.6394, "step": 20199 }, { "epoch": 0.7316189786309308, "grad_norm": 1.393932904710316, "learning_rate": 3.5459517690529254e-06, "loss": 0.6631, "step": 20200 }, { "epoch": 0.7316551973922492, "grad_norm": 1.236600424021351, "learning_rate": 3.5450557765958706e-06, "loss": 0.6752, "step": 20201 }, { "epoch": 0.7316914161535676, "grad_norm": 1.4864318676107102, "learning_rate": 3.544159872964519e-06, "loss": 0.7242, "step": 20202 }, { "epoch": 0.7317276349148859, "grad_norm": 1.3007425898170955, "learning_rate": 3.543264058171201e-06, "loss": 0.6664, "step": 20203 }, { "epoch": 0.7317638536762042, "grad_norm": 1.2408930651344627, "learning_rate": 3.5423683322282386e-06, "loss": 0.637, "step": 20204 }, { "epoch": 0.7318000724375227, "grad_norm": 1.0594531972936057, "learning_rate": 3.5414726951479674e-06, "loss": 0.6582, "step": 20205 }, { "epoch": 0.731836291198841, "grad_norm": 1.3172206785960288, "learning_rate": 3.5405771469427052e-06, "loss": 0.7058, "step": 20206 }, { "epoch": 0.7318725099601594, "grad_norm": 0.95754664763596, "learning_rate": 3.5396816876247774e-06, "loss": 0.6811, "step": 20207 }, { "epoch": 0.7319087287214777, "grad_norm": 1.4275878682354917, "learning_rate": 3.538786317206503e-06, "loss": 0.6568, "step": 20208 }, { "epoch": 0.731944947482796, "grad_norm": 1.4482866384587274, "learning_rate": 3.5378910357002093e-06, "loss": 0.7037, "step": 20209 }, { "epoch": 0.7319811662441145, "grad_norm": 1.4274837238651965, "learning_rate": 3.536995843118214e-06, "loss": 0.8147, "step": 20210 }, { "epoch": 0.7320173850054328, "grad_norm": 1.181412448113009, "learning_rate": 3.536100739472831e-06, "loss": 0.6639, "step": 20211 }, { "epoch": 0.7320536037667512, "grad_norm": 1.4649866574832455, "learning_rate": 3.5352057247763837e-06, "loss": 0.6678, "step": 20212 }, { "epoch": 0.7320898225280695, "grad_norm": 1.1120506161603323, "learning_rate": 3.5343107990411875e-06, "loss": 0.6489, "step": 20213 }, { "epoch": 0.7321260412893879, "grad_norm": 1.3483064016163933, "learning_rate": 3.5334159622795548e-06, "loss": 0.6774, "step": 20214 }, { "epoch": 0.7321622600507063, "grad_norm": 1.65424624862974, "learning_rate": 3.532521214503798e-06, "loss": 0.6925, "step": 20215 }, { "epoch": 0.7321984788120246, "grad_norm": 1.2662951915799865, "learning_rate": 3.531626555726235e-06, "loss": 0.6316, "step": 20216 }, { "epoch": 0.732234697573343, "grad_norm": 1.3364746703633266, "learning_rate": 3.530731985959175e-06, "loss": 0.71, "step": 20217 }, { "epoch": 0.7322709163346613, "grad_norm": 1.1266034762948325, "learning_rate": 3.529837505214927e-06, "loss": 0.6572, "step": 20218 }, { "epoch": 0.7323071350959797, "grad_norm": 0.9480349406186512, "learning_rate": 3.528943113505796e-06, "loss": 0.7157, "step": 20219 }, { "epoch": 0.7323433538572981, "grad_norm": 1.3326063417901592, "learning_rate": 3.528048810844098e-06, "loss": 0.7502, "step": 20220 }, { "epoch": 0.7323795726186164, "grad_norm": 1.0144242497029985, "learning_rate": 3.527154597242135e-06, "loss": 0.6582, "step": 20221 }, { "epoch": 0.7324157913799348, "grad_norm": 1.0021315740714942, "learning_rate": 3.5262604727122084e-06, "loss": 0.7, "step": 20222 }, { "epoch": 0.7324520101412532, "grad_norm": 1.0417109909847033, "learning_rate": 3.5253664372666297e-06, "loss": 0.7468, "step": 20223 }, { "epoch": 0.7324882289025715, "grad_norm": 1.4702890276682177, "learning_rate": 3.5244724909176973e-06, "loss": 0.6501, "step": 20224 }, { "epoch": 0.7325244476638899, "grad_norm": 1.305379728886717, "learning_rate": 3.5235786336777144e-06, "loss": 0.6547, "step": 20225 }, { "epoch": 0.7325606664252082, "grad_norm": 1.0535834444284577, "learning_rate": 3.5226848655589763e-06, "loss": 0.755, "step": 20226 }, { "epoch": 0.7325968851865267, "grad_norm": 1.369614496682265, "learning_rate": 3.5217911865737885e-06, "loss": 0.6794, "step": 20227 }, { "epoch": 0.732633103947845, "grad_norm": 1.4436996043479917, "learning_rate": 3.5208975967344472e-06, "loss": 0.6816, "step": 20228 }, { "epoch": 0.7326693227091633, "grad_norm": 0.9817949211469861, "learning_rate": 3.5200040960532443e-06, "loss": 0.6974, "step": 20229 }, { "epoch": 0.7327055414704817, "grad_norm": 1.3512480754545253, "learning_rate": 3.519110684542483e-06, "loss": 0.7016, "step": 20230 }, { "epoch": 0.7327417602318, "grad_norm": 1.4066317172670195, "learning_rate": 3.5182173622144524e-06, "loss": 0.718, "step": 20231 }, { "epoch": 0.7327779789931185, "grad_norm": 1.0725885559194512, "learning_rate": 3.5173241290814464e-06, "loss": 0.7555, "step": 20232 }, { "epoch": 0.7328141977544368, "grad_norm": 1.397884637218288, "learning_rate": 3.5164309851557532e-06, "loss": 0.7808, "step": 20233 }, { "epoch": 0.7328504165157551, "grad_norm": 1.3018049129999585, "learning_rate": 3.5155379304496705e-06, "loss": 0.6811, "step": 20234 }, { "epoch": 0.7328866352770735, "grad_norm": 1.3682133701814123, "learning_rate": 3.514644964975483e-06, "loss": 0.7025, "step": 20235 }, { "epoch": 0.7329228540383919, "grad_norm": 1.3439034343881142, "learning_rate": 3.513752088745479e-06, "loss": 0.7068, "step": 20236 }, { "epoch": 0.7329590727997103, "grad_norm": 1.5079185320083803, "learning_rate": 3.512859301771944e-06, "loss": 0.7456, "step": 20237 }, { "epoch": 0.7329952915610286, "grad_norm": 1.0485015761263698, "learning_rate": 3.5119666040671663e-06, "loss": 0.7217, "step": 20238 }, { "epoch": 0.7330315103223469, "grad_norm": 1.2893207768144022, "learning_rate": 3.5110739956434303e-06, "loss": 0.7427, "step": 20239 }, { "epoch": 0.7330677290836654, "grad_norm": 1.0477411907715832, "learning_rate": 3.510181476513014e-06, "loss": 0.6928, "step": 20240 }, { "epoch": 0.7331039478449837, "grad_norm": 0.9985035871491282, "learning_rate": 3.509289046688207e-06, "loss": 0.6868, "step": 20241 }, { "epoch": 0.7331401666063021, "grad_norm": 1.3706664101398092, "learning_rate": 3.5083967061812852e-06, "loss": 0.6406, "step": 20242 }, { "epoch": 0.7331763853676204, "grad_norm": 1.262313519529137, "learning_rate": 3.507504455004529e-06, "loss": 0.6286, "step": 20243 }, { "epoch": 0.7332126041289387, "grad_norm": 1.0421324662475229, "learning_rate": 3.506612293170213e-06, "loss": 0.6887, "step": 20244 }, { "epoch": 0.7332488228902572, "grad_norm": 1.4462730334705387, "learning_rate": 3.5057202206906205e-06, "loss": 0.6626, "step": 20245 }, { "epoch": 0.7332850416515755, "grad_norm": 1.3505382612175143, "learning_rate": 3.504828237578024e-06, "loss": 0.6068, "step": 20246 }, { "epoch": 0.7333212604128939, "grad_norm": 0.951523313780824, "learning_rate": 3.5039363438446984e-06, "loss": 0.6906, "step": 20247 }, { "epoch": 0.7333574791742122, "grad_norm": 1.5464654992154376, "learning_rate": 3.503044539502913e-06, "loss": 0.7201, "step": 20248 }, { "epoch": 0.7333936979355306, "grad_norm": 1.2970829949134466, "learning_rate": 3.5021528245649473e-06, "loss": 0.6656, "step": 20249 }, { "epoch": 0.733429916696849, "grad_norm": 1.33644938979448, "learning_rate": 3.501261199043069e-06, "loss": 0.7068, "step": 20250 }, { "epoch": 0.7334661354581673, "grad_norm": 1.3058809715080422, "learning_rate": 3.500369662949542e-06, "loss": 0.6953, "step": 20251 }, { "epoch": 0.7335023542194857, "grad_norm": 1.229977078828763, "learning_rate": 3.4994782162966435e-06, "loss": 0.6708, "step": 20252 }, { "epoch": 0.733538572980804, "grad_norm": 1.2612048525688115, "learning_rate": 3.4985868590966364e-06, "loss": 0.6316, "step": 20253 }, { "epoch": 0.7335747917421224, "grad_norm": 1.3618463709213033, "learning_rate": 3.497695591361787e-06, "loss": 0.7259, "step": 20254 }, { "epoch": 0.7336110105034408, "grad_norm": 1.3070281955024339, "learning_rate": 3.4968044131043565e-06, "loss": 0.6635, "step": 20255 }, { "epoch": 0.7336472292647591, "grad_norm": 1.2945179345200646, "learning_rate": 3.495913324336615e-06, "loss": 0.6291, "step": 20256 }, { "epoch": 0.7336834480260775, "grad_norm": 1.4879613968173682, "learning_rate": 3.4950223250708204e-06, "loss": 0.6668, "step": 20257 }, { "epoch": 0.7337196667873959, "grad_norm": 1.4216689612896776, "learning_rate": 3.494131415319235e-06, "loss": 0.7276, "step": 20258 }, { "epoch": 0.7337558855487142, "grad_norm": 1.3514975783019962, "learning_rate": 3.493240595094114e-06, "loss": 0.6802, "step": 20259 }, { "epoch": 0.7337921043100326, "grad_norm": 1.1574212055448976, "learning_rate": 3.4923498644077236e-06, "loss": 0.6537, "step": 20260 }, { "epoch": 0.7338283230713509, "grad_norm": 1.2542110485626015, "learning_rate": 3.4914592232723164e-06, "loss": 0.6922, "step": 20261 }, { "epoch": 0.7338645418326694, "grad_norm": 1.2541392940173652, "learning_rate": 3.4905686717001473e-06, "loss": 0.6671, "step": 20262 }, { "epoch": 0.7339007605939877, "grad_norm": 1.0733000104746018, "learning_rate": 3.4896782097034755e-06, "loss": 0.7005, "step": 20263 }, { "epoch": 0.733936979355306, "grad_norm": 1.0622008074218356, "learning_rate": 3.4887878372945517e-06, "loss": 0.683, "step": 20264 }, { "epoch": 0.7339731981166244, "grad_norm": 1.2407102519007889, "learning_rate": 3.4878975544856285e-06, "loss": 0.6641, "step": 20265 }, { "epoch": 0.7340094168779427, "grad_norm": 1.2470994645716489, "learning_rate": 3.4870073612889544e-06, "loss": 0.6019, "step": 20266 }, { "epoch": 0.7340456356392612, "grad_norm": 0.9752355329087556, "learning_rate": 3.4861172577167845e-06, "loss": 0.6812, "step": 20267 }, { "epoch": 0.7340818544005795, "grad_norm": 1.3218728239918558, "learning_rate": 3.485227243781365e-06, "loss": 0.7089, "step": 20268 }, { "epoch": 0.7341180731618978, "grad_norm": 1.466925158268876, "learning_rate": 3.4843373194949427e-06, "loss": 0.6861, "step": 20269 }, { "epoch": 0.7341542919232162, "grad_norm": 1.2893477810560645, "learning_rate": 3.483447484869761e-06, "loss": 0.6149, "step": 20270 }, { "epoch": 0.7341905106845346, "grad_norm": 1.3662987981416084, "learning_rate": 3.482557739918071e-06, "loss": 0.6517, "step": 20271 }, { "epoch": 0.734226729445853, "grad_norm": 1.1348086002123088, "learning_rate": 3.481668084652113e-06, "loss": 0.6267, "step": 20272 }, { "epoch": 0.7342629482071713, "grad_norm": 1.4179146325474383, "learning_rate": 3.480778519084127e-06, "loss": 0.6691, "step": 20273 }, { "epoch": 0.7342991669684896, "grad_norm": 1.3355149784669667, "learning_rate": 3.47988904322636e-06, "loss": 0.7734, "step": 20274 }, { "epoch": 0.734335385729808, "grad_norm": 1.3275216330859716, "learning_rate": 3.478999657091049e-06, "loss": 0.6257, "step": 20275 }, { "epoch": 0.7343716044911264, "grad_norm": 1.3412260891841932, "learning_rate": 3.478110360690432e-06, "loss": 0.6773, "step": 20276 }, { "epoch": 0.7344078232524448, "grad_norm": 1.5947469043076725, "learning_rate": 3.477221154036744e-06, "loss": 0.7194, "step": 20277 }, { "epoch": 0.7344440420137631, "grad_norm": 1.3365455290861035, "learning_rate": 3.476332037142227e-06, "loss": 0.7261, "step": 20278 }, { "epoch": 0.7344802607750814, "grad_norm": 1.3781648208148118, "learning_rate": 3.475443010019115e-06, "loss": 0.7438, "step": 20279 }, { "epoch": 0.7345164795363999, "grad_norm": 1.088742792557092, "learning_rate": 3.474554072679639e-06, "loss": 0.6665, "step": 20280 }, { "epoch": 0.7345526982977182, "grad_norm": 1.413158225086916, "learning_rate": 3.473665225136029e-06, "loss": 0.6722, "step": 20281 }, { "epoch": 0.7345889170590366, "grad_norm": 1.3294761726801287, "learning_rate": 3.472776467400525e-06, "loss": 0.632, "step": 20282 }, { "epoch": 0.7346251358203549, "grad_norm": 1.2878902863689206, "learning_rate": 3.4718877994853508e-06, "loss": 0.6439, "step": 20283 }, { "epoch": 0.7346613545816733, "grad_norm": 1.4662373970421825, "learning_rate": 3.470999221402733e-06, "loss": 0.6749, "step": 20284 }, { "epoch": 0.7346975733429917, "grad_norm": 1.3970364259727002, "learning_rate": 3.4701107331649066e-06, "loss": 0.6254, "step": 20285 }, { "epoch": 0.73473379210431, "grad_norm": 1.4348602229203689, "learning_rate": 3.4692223347840947e-06, "loss": 0.6343, "step": 20286 }, { "epoch": 0.7347700108656284, "grad_norm": 1.3753568358307906, "learning_rate": 3.4683340262725217e-06, "loss": 0.7058, "step": 20287 }, { "epoch": 0.7348062296269467, "grad_norm": 1.3644602849921539, "learning_rate": 3.467445807642408e-06, "loss": 0.6816, "step": 20288 }, { "epoch": 0.7348424483882651, "grad_norm": 1.3025834977027664, "learning_rate": 3.466557678905983e-06, "loss": 0.6248, "step": 20289 }, { "epoch": 0.7348786671495835, "grad_norm": 1.3717333790600528, "learning_rate": 3.4656696400754652e-06, "loss": 0.7174, "step": 20290 }, { "epoch": 0.7349148859109018, "grad_norm": 1.367677900373294, "learning_rate": 3.464781691163073e-06, "loss": 0.6501, "step": 20291 }, { "epoch": 0.7349511046722202, "grad_norm": 0.9024126938778745, "learning_rate": 3.463893832181029e-06, "loss": 0.636, "step": 20292 }, { "epoch": 0.7349873234335386, "grad_norm": 1.3126569083631836, "learning_rate": 3.4630060631415487e-06, "loss": 0.6897, "step": 20293 }, { "epoch": 0.7350235421948569, "grad_norm": 1.335426017238887, "learning_rate": 3.4621183840568495e-06, "loss": 0.617, "step": 20294 }, { "epoch": 0.7350597609561753, "grad_norm": 1.4576391148671777, "learning_rate": 3.461230794939142e-06, "loss": 0.6782, "step": 20295 }, { "epoch": 0.7350959797174936, "grad_norm": 1.3260322217768488, "learning_rate": 3.4603432958006488e-06, "loss": 0.6892, "step": 20296 }, { "epoch": 0.7351321984788121, "grad_norm": 1.3824693165054942, "learning_rate": 3.4594558866535778e-06, "loss": 0.6913, "step": 20297 }, { "epoch": 0.7351684172401304, "grad_norm": 1.3047784727896263, "learning_rate": 3.45856856751014e-06, "loss": 0.696, "step": 20298 }, { "epoch": 0.7352046360014487, "grad_norm": 1.089243957179142, "learning_rate": 3.4576813383825436e-06, "loss": 0.6713, "step": 20299 }, { "epoch": 0.7352408547627671, "grad_norm": 1.2310764204912397, "learning_rate": 3.4567941992830034e-06, "loss": 0.6528, "step": 20300 }, { "epoch": 0.7352770735240854, "grad_norm": 1.4357218116774688, "learning_rate": 3.4559071502237252e-06, "loss": 0.636, "step": 20301 }, { "epoch": 0.7353132922854039, "grad_norm": 2.335069779077354, "learning_rate": 3.455020191216911e-06, "loss": 0.7583, "step": 20302 }, { "epoch": 0.7353495110467222, "grad_norm": 1.2929947899446452, "learning_rate": 3.454133322274773e-06, "loss": 0.5961, "step": 20303 }, { "epoch": 0.7353857298080405, "grad_norm": 1.284631049315476, "learning_rate": 3.453246543409513e-06, "loss": 0.629, "step": 20304 }, { "epoch": 0.7354219485693589, "grad_norm": 1.2920174531347626, "learning_rate": 3.4523598546333325e-06, "loss": 0.66, "step": 20305 }, { "epoch": 0.7354581673306773, "grad_norm": 0.9766972893438716, "learning_rate": 3.451473255958431e-06, "loss": 0.7366, "step": 20306 }, { "epoch": 0.7354943860919957, "grad_norm": 1.324800800867241, "learning_rate": 3.450586747397013e-06, "loss": 0.6719, "step": 20307 }, { "epoch": 0.735530604853314, "grad_norm": 1.3819999927686455, "learning_rate": 3.4497003289612783e-06, "loss": 0.6952, "step": 20308 }, { "epoch": 0.7355668236146323, "grad_norm": 1.3860007925859577, "learning_rate": 3.448814000663422e-06, "loss": 0.7241, "step": 20309 }, { "epoch": 0.7356030423759508, "grad_norm": 1.018671843549464, "learning_rate": 3.4479277625156383e-06, "loss": 0.6898, "step": 20310 }, { "epoch": 0.7356392611372691, "grad_norm": 1.505095831308731, "learning_rate": 3.4470416145301276e-06, "loss": 0.7418, "step": 20311 }, { "epoch": 0.7356754798985875, "grad_norm": 0.986387245134348, "learning_rate": 3.4461555567190842e-06, "loss": 0.6834, "step": 20312 }, { "epoch": 0.7357116986599058, "grad_norm": 0.9007128899820689, "learning_rate": 3.445269589094694e-06, "loss": 0.6845, "step": 20313 }, { "epoch": 0.7357479174212241, "grad_norm": 1.342869397982698, "learning_rate": 3.444383711669158e-06, "loss": 0.6439, "step": 20314 }, { "epoch": 0.7357841361825426, "grad_norm": 1.3894619256942824, "learning_rate": 3.443497924454663e-06, "loss": 0.6989, "step": 20315 }, { "epoch": 0.7358203549438609, "grad_norm": 1.5041898881047897, "learning_rate": 3.4426122274633968e-06, "loss": 0.63, "step": 20316 }, { "epoch": 0.7358565737051793, "grad_norm": 1.4229054228067064, "learning_rate": 3.441726620707545e-06, "loss": 0.6637, "step": 20317 }, { "epoch": 0.7358927924664976, "grad_norm": 1.4515312912627722, "learning_rate": 3.4408411041993007e-06, "loss": 0.6855, "step": 20318 }, { "epoch": 0.735929011227816, "grad_norm": 1.4504387558076752, "learning_rate": 3.439955677950846e-06, "loss": 0.7178, "step": 20319 }, { "epoch": 0.7359652299891344, "grad_norm": 1.5702688438837809, "learning_rate": 3.4390703419743654e-06, "loss": 0.6975, "step": 20320 }, { "epoch": 0.7360014487504527, "grad_norm": 1.5062494607633599, "learning_rate": 3.4381850962820396e-06, "loss": 0.7431, "step": 20321 }, { "epoch": 0.7360376675117711, "grad_norm": 1.3260360249306686, "learning_rate": 3.437299940886054e-06, "loss": 0.7142, "step": 20322 }, { "epoch": 0.7360738862730895, "grad_norm": 1.4050050390355102, "learning_rate": 3.4364148757985895e-06, "loss": 0.6611, "step": 20323 }, { "epoch": 0.7361101050344078, "grad_norm": 1.333342362785561, "learning_rate": 3.4355299010318187e-06, "loss": 0.7332, "step": 20324 }, { "epoch": 0.7361463237957262, "grad_norm": 1.0095058473969678, "learning_rate": 3.4346450165979294e-06, "loss": 0.6718, "step": 20325 }, { "epoch": 0.7361825425570445, "grad_norm": 1.3872624573851604, "learning_rate": 3.4337602225090926e-06, "loss": 0.6692, "step": 20326 }, { "epoch": 0.736218761318363, "grad_norm": 1.5184699301662774, "learning_rate": 3.4328755187774853e-06, "loss": 0.688, "step": 20327 }, { "epoch": 0.7362549800796813, "grad_norm": 1.2452456142538906, "learning_rate": 3.431990905415277e-06, "loss": 0.6573, "step": 20328 }, { "epoch": 0.7362911988409996, "grad_norm": 1.346402239971159, "learning_rate": 3.431106382434649e-06, "loss": 0.7022, "step": 20329 }, { "epoch": 0.736327417602318, "grad_norm": 1.4401376331181843, "learning_rate": 3.4302219498477683e-06, "loss": 0.7642, "step": 20330 }, { "epoch": 0.7363636363636363, "grad_norm": 1.3483911556046275, "learning_rate": 3.4293376076668073e-06, "loss": 0.6822, "step": 20331 }, { "epoch": 0.7363998551249548, "grad_norm": 1.0752059763731232, "learning_rate": 3.42845335590393e-06, "loss": 0.7425, "step": 20332 }, { "epoch": 0.7364360738862731, "grad_norm": 1.0374845474525796, "learning_rate": 3.427569194571313e-06, "loss": 0.7283, "step": 20333 }, { "epoch": 0.7364722926475914, "grad_norm": 1.4114022175972942, "learning_rate": 3.426685123681119e-06, "loss": 0.7289, "step": 20334 }, { "epoch": 0.7365085114089098, "grad_norm": 1.5545774114152264, "learning_rate": 3.425801143245513e-06, "loss": 0.6547, "step": 20335 }, { "epoch": 0.7365447301702281, "grad_norm": 1.2650202771556431, "learning_rate": 3.4249172532766607e-06, "loss": 0.5807, "step": 20336 }, { "epoch": 0.7365809489315466, "grad_norm": 1.1428576311455148, "learning_rate": 3.4240334537867236e-06, "loss": 0.7458, "step": 20337 }, { "epoch": 0.7366171676928649, "grad_norm": 1.413074866398464, "learning_rate": 3.4231497447878604e-06, "loss": 0.6817, "step": 20338 }, { "epoch": 0.7366533864541832, "grad_norm": 1.3556514034276497, "learning_rate": 3.4222661262922395e-06, "loss": 0.7019, "step": 20339 }, { "epoch": 0.7366896052155016, "grad_norm": 1.2820098720002506, "learning_rate": 3.4213825983120175e-06, "loss": 0.621, "step": 20340 }, { "epoch": 0.73672582397682, "grad_norm": 1.1154176372053086, "learning_rate": 3.4204991608593497e-06, "loss": 0.6914, "step": 20341 }, { "epoch": 0.7367620427381384, "grad_norm": 1.5248032280245656, "learning_rate": 3.419615813946392e-06, "loss": 0.6666, "step": 20342 }, { "epoch": 0.7367982614994567, "grad_norm": 0.9404257157673653, "learning_rate": 3.4187325575853068e-06, "loss": 0.6001, "step": 20343 }, { "epoch": 0.736834480260775, "grad_norm": 1.3425384177095656, "learning_rate": 3.417849391788244e-06, "loss": 0.7269, "step": 20344 }, { "epoch": 0.7368706990220935, "grad_norm": 1.2995139906816007, "learning_rate": 3.416966316567354e-06, "loss": 0.7063, "step": 20345 }, { "epoch": 0.7369069177834118, "grad_norm": 1.3263914210263172, "learning_rate": 3.416083331934794e-06, "loss": 0.7192, "step": 20346 }, { "epoch": 0.7369431365447302, "grad_norm": 1.3145190331541905, "learning_rate": 3.4152004379027136e-06, "loss": 0.598, "step": 20347 }, { "epoch": 0.7369793553060485, "grad_norm": 1.4515345493742264, "learning_rate": 3.4143176344832605e-06, "loss": 0.6564, "step": 20348 }, { "epoch": 0.7370155740673668, "grad_norm": 1.4158297725913467, "learning_rate": 3.4134349216885808e-06, "loss": 0.6022, "step": 20349 }, { "epoch": 0.7370517928286853, "grad_norm": 1.3607117442381351, "learning_rate": 3.412552299530827e-06, "loss": 0.6877, "step": 20350 }, { "epoch": 0.7370880115900036, "grad_norm": 1.349691286375401, "learning_rate": 3.411669768022141e-06, "loss": 0.6678, "step": 20351 }, { "epoch": 0.737124230351322, "grad_norm": 1.2699905400998623, "learning_rate": 3.4107873271746695e-06, "loss": 0.6169, "step": 20352 }, { "epoch": 0.7371604491126403, "grad_norm": 1.3646609359950688, "learning_rate": 3.40990497700055e-06, "loss": 0.7451, "step": 20353 }, { "epoch": 0.7371966678739587, "grad_norm": 1.030354756466964, "learning_rate": 3.4090227175119315e-06, "loss": 0.7007, "step": 20354 }, { "epoch": 0.7372328866352771, "grad_norm": 1.3786457426255843, "learning_rate": 3.408140548720953e-06, "loss": 0.7192, "step": 20355 }, { "epoch": 0.7372691053965954, "grad_norm": 1.2657956270301518, "learning_rate": 3.407258470639748e-06, "loss": 0.6527, "step": 20356 }, { "epoch": 0.7373053241579138, "grad_norm": 1.273683480335828, "learning_rate": 3.4063764832804637e-06, "loss": 0.6585, "step": 20357 }, { "epoch": 0.7373415429192322, "grad_norm": 1.3386291894766567, "learning_rate": 3.4054945866552324e-06, "loss": 0.6658, "step": 20358 }, { "epoch": 0.7373777616805505, "grad_norm": 1.2652057296145998, "learning_rate": 3.40461278077619e-06, "loss": 0.6551, "step": 20359 }, { "epoch": 0.7374139804418689, "grad_norm": 1.3822761926153797, "learning_rate": 3.4037310656554668e-06, "loss": 0.734, "step": 20360 }, { "epoch": 0.7374501992031872, "grad_norm": 1.495402168183337, "learning_rate": 3.402849441305205e-06, "loss": 0.7599, "step": 20361 }, { "epoch": 0.7374864179645056, "grad_norm": 1.4434998305845543, "learning_rate": 3.40196790773753e-06, "loss": 0.7737, "step": 20362 }, { "epoch": 0.737522636725824, "grad_norm": 1.3758910608089818, "learning_rate": 3.401086464964575e-06, "loss": 0.7023, "step": 20363 }, { "epoch": 0.7375588554871424, "grad_norm": 1.632162164237305, "learning_rate": 3.400205112998466e-06, "loss": 0.6902, "step": 20364 }, { "epoch": 0.7375950742484607, "grad_norm": 1.0270180732468237, "learning_rate": 3.399323851851336e-06, "loss": 0.6354, "step": 20365 }, { "epoch": 0.737631293009779, "grad_norm": 1.1611533159157363, "learning_rate": 3.39844268153531e-06, "loss": 0.6412, "step": 20366 }, { "epoch": 0.7376675117710975, "grad_norm": 1.3675642257459177, "learning_rate": 3.397561602062509e-06, "loss": 0.7327, "step": 20367 }, { "epoch": 0.7377037305324158, "grad_norm": 1.228129371666832, "learning_rate": 3.3966806134450657e-06, "loss": 0.6865, "step": 20368 }, { "epoch": 0.7377399492937342, "grad_norm": 1.2787332297742209, "learning_rate": 3.3957997156951007e-06, "loss": 0.6738, "step": 20369 }, { "epoch": 0.7377761680550525, "grad_norm": 1.3452154592246086, "learning_rate": 3.3949189088247326e-06, "loss": 0.7133, "step": 20370 }, { "epoch": 0.7378123868163708, "grad_norm": 1.3736378000462564, "learning_rate": 3.394038192846082e-06, "loss": 0.6326, "step": 20371 }, { "epoch": 0.7378486055776893, "grad_norm": 1.4874903512289155, "learning_rate": 3.3931575677712738e-06, "loss": 0.7309, "step": 20372 }, { "epoch": 0.7378848243390076, "grad_norm": 0.9539977647296873, "learning_rate": 3.3922770336124223e-06, "loss": 0.6698, "step": 20373 }, { "epoch": 0.737921043100326, "grad_norm": 1.3977900290180778, "learning_rate": 3.3913965903816446e-06, "loss": 0.636, "step": 20374 }, { "epoch": 0.7379572618616443, "grad_norm": 1.1423469194396112, "learning_rate": 3.390516238091053e-06, "loss": 0.6786, "step": 20375 }, { "epoch": 0.7379934806229627, "grad_norm": 1.3446020568742487, "learning_rate": 3.38963597675277e-06, "loss": 0.6829, "step": 20376 }, { "epoch": 0.7380296993842811, "grad_norm": 1.456791812051017, "learning_rate": 3.3887558063789037e-06, "loss": 0.6679, "step": 20377 }, { "epoch": 0.7380659181455994, "grad_norm": 1.0547648917221786, "learning_rate": 3.387875726981563e-06, "loss": 0.7234, "step": 20378 }, { "epoch": 0.7381021369069178, "grad_norm": 1.4746985553780554, "learning_rate": 3.3869957385728657e-06, "loss": 0.7374, "step": 20379 }, { "epoch": 0.7381383556682362, "grad_norm": 1.7375378680047664, "learning_rate": 3.386115841164919e-06, "loss": 0.7118, "step": 20380 }, { "epoch": 0.7381745744295545, "grad_norm": 1.4379474472180687, "learning_rate": 3.3852360347698274e-06, "loss": 0.6767, "step": 20381 }, { "epoch": 0.7382107931908729, "grad_norm": 1.349810588041625, "learning_rate": 3.3843563193996985e-06, "loss": 0.6919, "step": 20382 }, { "epoch": 0.7382470119521912, "grad_norm": 1.3260908993984437, "learning_rate": 3.383476695066643e-06, "loss": 0.7044, "step": 20383 }, { "epoch": 0.7382832307135097, "grad_norm": 1.1032580542417008, "learning_rate": 3.382597161782761e-06, "loss": 0.6904, "step": 20384 }, { "epoch": 0.738319449474828, "grad_norm": 1.2231147558810183, "learning_rate": 3.381717719560155e-06, "loss": 0.7421, "step": 20385 }, { "epoch": 0.7383556682361463, "grad_norm": 1.3422446954262306, "learning_rate": 3.38083836841093e-06, "loss": 0.7064, "step": 20386 }, { "epoch": 0.7383918869974647, "grad_norm": 1.4383299138406116, "learning_rate": 3.3799591083471862e-06, "loss": 0.7669, "step": 20387 }, { "epoch": 0.738428105758783, "grad_norm": 1.504434097844427, "learning_rate": 3.3790799393810227e-06, "loss": 0.7483, "step": 20388 }, { "epoch": 0.7384643245201015, "grad_norm": 0.9539046105608766, "learning_rate": 3.378200861524532e-06, "loss": 0.6784, "step": 20389 }, { "epoch": 0.7385005432814198, "grad_norm": 1.5066815295315343, "learning_rate": 3.3773218747898195e-06, "loss": 0.6994, "step": 20390 }, { "epoch": 0.7385367620427381, "grad_norm": 1.4054611619885118, "learning_rate": 3.3764429791889774e-06, "loss": 0.7503, "step": 20391 }, { "epoch": 0.7385729808040565, "grad_norm": 0.960276726993995, "learning_rate": 3.3755641747341008e-06, "loss": 0.6727, "step": 20392 }, { "epoch": 0.7386091995653749, "grad_norm": 1.4610921980556193, "learning_rate": 3.374685461437277e-06, "loss": 0.728, "step": 20393 }, { "epoch": 0.7386454183266933, "grad_norm": 1.3470777444743698, "learning_rate": 3.373806839310607e-06, "loss": 0.6842, "step": 20394 }, { "epoch": 0.7386816370880116, "grad_norm": 0.9671398071810683, "learning_rate": 3.3729283083661767e-06, "loss": 0.6395, "step": 20395 }, { "epoch": 0.7387178558493299, "grad_norm": 1.3094678549970475, "learning_rate": 3.372049868616073e-06, "loss": 0.7524, "step": 20396 }, { "epoch": 0.7387540746106483, "grad_norm": 1.4881117604082081, "learning_rate": 3.3711715200723904e-06, "loss": 0.7047, "step": 20397 }, { "epoch": 0.7387902933719667, "grad_norm": 1.4214846696195536, "learning_rate": 3.370293262747212e-06, "loss": 0.6891, "step": 20398 }, { "epoch": 0.7388265121332851, "grad_norm": 1.0217373333713407, "learning_rate": 3.369415096652624e-06, "loss": 0.8086, "step": 20399 }, { "epoch": 0.7388627308946034, "grad_norm": 1.3447181366190393, "learning_rate": 3.368537021800706e-06, "loss": 0.7219, "step": 20400 }, { "epoch": 0.7388989496559217, "grad_norm": 1.4722077599963217, "learning_rate": 3.3676590382035503e-06, "loss": 0.6896, "step": 20401 }, { "epoch": 0.7389351684172402, "grad_norm": 1.058302873214211, "learning_rate": 3.3667811458732323e-06, "loss": 0.6851, "step": 20402 }, { "epoch": 0.7389713871785585, "grad_norm": 1.0178579366284324, "learning_rate": 3.365903344821836e-06, "loss": 0.676, "step": 20403 }, { "epoch": 0.7390076059398769, "grad_norm": 1.3999646275758362, "learning_rate": 3.365025635061434e-06, "loss": 0.7142, "step": 20404 }, { "epoch": 0.7390438247011952, "grad_norm": 1.3494692997483992, "learning_rate": 3.364148016604114e-06, "loss": 0.6828, "step": 20405 }, { "epoch": 0.7390800434625135, "grad_norm": 1.1401981467907998, "learning_rate": 3.3632704894619473e-06, "loss": 0.6625, "step": 20406 }, { "epoch": 0.739116262223832, "grad_norm": 1.3971670802801412, "learning_rate": 3.362393053647006e-06, "loss": 0.6791, "step": 20407 }, { "epoch": 0.7391524809851503, "grad_norm": 1.5250542595856729, "learning_rate": 3.3615157091713734e-06, "loss": 0.6913, "step": 20408 }, { "epoch": 0.7391886997464687, "grad_norm": 1.446277706583325, "learning_rate": 3.3606384560471173e-06, "loss": 0.6666, "step": 20409 }, { "epoch": 0.739224918507787, "grad_norm": 1.9382237805261455, "learning_rate": 3.35976129428631e-06, "loss": 0.7915, "step": 20410 }, { "epoch": 0.7392611372691054, "grad_norm": 1.3298340060852283, "learning_rate": 3.358884223901018e-06, "loss": 0.7269, "step": 20411 }, { "epoch": 0.7392973560304238, "grad_norm": 1.2618351667096144, "learning_rate": 3.3580072449033186e-06, "loss": 0.6935, "step": 20412 }, { "epoch": 0.7393335747917421, "grad_norm": 1.2607007688401155, "learning_rate": 3.357130357305275e-06, "loss": 0.6689, "step": 20413 }, { "epoch": 0.7393697935530605, "grad_norm": 1.6705551714535711, "learning_rate": 3.356253561118955e-06, "loss": 0.7225, "step": 20414 }, { "epoch": 0.7394060123143789, "grad_norm": 1.3132658083845834, "learning_rate": 3.355376856356419e-06, "loss": 0.6692, "step": 20415 }, { "epoch": 0.7394422310756972, "grad_norm": 1.1875836372724364, "learning_rate": 3.3545002430297414e-06, "loss": 0.6317, "step": 20416 }, { "epoch": 0.7394784498370156, "grad_norm": 1.4934315457467817, "learning_rate": 3.353623721150978e-06, "loss": 0.7564, "step": 20417 }, { "epoch": 0.7395146685983339, "grad_norm": 1.6041267893735218, "learning_rate": 3.35274729073219e-06, "loss": 0.7201, "step": 20418 }, { "epoch": 0.7395508873596524, "grad_norm": 1.4752470052625206, "learning_rate": 3.3518709517854433e-06, "loss": 0.7281, "step": 20419 }, { "epoch": 0.7395871061209707, "grad_norm": 1.2650453406783735, "learning_rate": 3.350994704322794e-06, "loss": 0.6782, "step": 20420 }, { "epoch": 0.739623324882289, "grad_norm": 1.0398589262528393, "learning_rate": 3.350118548356299e-06, "loss": 0.7047, "step": 20421 }, { "epoch": 0.7396595436436074, "grad_norm": 1.3662852539071366, "learning_rate": 3.349242483898013e-06, "loss": 0.7211, "step": 20422 }, { "epoch": 0.7396957624049257, "grad_norm": 1.267637256571477, "learning_rate": 3.348366510959998e-06, "loss": 0.6131, "step": 20423 }, { "epoch": 0.7397319811662442, "grad_norm": 1.415941538352444, "learning_rate": 3.347490629554304e-06, "loss": 0.7073, "step": 20424 }, { "epoch": 0.7397681999275625, "grad_norm": 0.8486125739439427, "learning_rate": 3.3466148396929843e-06, "loss": 0.6311, "step": 20425 }, { "epoch": 0.7398044186888808, "grad_norm": 1.5123640932842555, "learning_rate": 3.345739141388088e-06, "loss": 0.6812, "step": 20426 }, { "epoch": 0.7398406374501992, "grad_norm": 1.3173127633828174, "learning_rate": 3.344863534651671e-06, "loss": 0.7239, "step": 20427 }, { "epoch": 0.7398768562115176, "grad_norm": 1.3670312560584317, "learning_rate": 3.3439880194957785e-06, "loss": 0.6847, "step": 20428 }, { "epoch": 0.739913074972836, "grad_norm": 1.0243434719178024, "learning_rate": 3.3431125959324574e-06, "loss": 0.7281, "step": 20429 }, { "epoch": 0.7399492937341543, "grad_norm": 1.2528459011020496, "learning_rate": 3.342237263973759e-06, "loss": 0.5995, "step": 20430 }, { "epoch": 0.7399855124954726, "grad_norm": 1.313785520187056, "learning_rate": 3.3413620236317255e-06, "loss": 0.7129, "step": 20431 }, { "epoch": 0.740021731256791, "grad_norm": 1.3697385842646883, "learning_rate": 3.3404868749184015e-06, "loss": 0.6685, "step": 20432 }, { "epoch": 0.7400579500181094, "grad_norm": 1.4059767258876936, "learning_rate": 3.339611817845827e-06, "loss": 0.6914, "step": 20433 }, { "epoch": 0.7400941687794278, "grad_norm": 1.2314321477913877, "learning_rate": 3.338736852426049e-06, "loss": 0.6188, "step": 20434 }, { "epoch": 0.7401303875407461, "grad_norm": 1.5695592012854915, "learning_rate": 3.3378619786711052e-06, "loss": 0.6892, "step": 20435 }, { "epoch": 0.7401666063020644, "grad_norm": 1.391590406138891, "learning_rate": 3.3369871965930346e-06, "loss": 0.7533, "step": 20436 }, { "epoch": 0.7402028250633829, "grad_norm": 1.2920294220706503, "learning_rate": 3.33611250620387e-06, "loss": 0.6195, "step": 20437 }, { "epoch": 0.7402390438247012, "grad_norm": 1.3676904856644834, "learning_rate": 3.3352379075156582e-06, "loss": 0.6525, "step": 20438 }, { "epoch": 0.7402752625860196, "grad_norm": 1.017364684409082, "learning_rate": 3.3343634005404278e-06, "loss": 0.675, "step": 20439 }, { "epoch": 0.7403114813473379, "grad_norm": 1.2481984224079856, "learning_rate": 3.333488985290211e-06, "loss": 0.6314, "step": 20440 }, { "epoch": 0.7403477001086562, "grad_norm": 1.404479199326752, "learning_rate": 3.3326146617770462e-06, "loss": 0.6595, "step": 20441 }, { "epoch": 0.7403839188699747, "grad_norm": 1.3863269433731025, "learning_rate": 3.331740430012963e-06, "loss": 0.7419, "step": 20442 }, { "epoch": 0.740420137631293, "grad_norm": 1.3637298382897678, "learning_rate": 3.3308662900099895e-06, "loss": 0.7412, "step": 20443 }, { "epoch": 0.7404563563926114, "grad_norm": 1.2419453188395355, "learning_rate": 3.3299922417801533e-06, "loss": 0.6806, "step": 20444 }, { "epoch": 0.7404925751539297, "grad_norm": 1.5048421486640167, "learning_rate": 3.3291182853354888e-06, "loss": 0.6957, "step": 20445 }, { "epoch": 0.7405287939152481, "grad_norm": 1.3433047495376433, "learning_rate": 3.3282444206880173e-06, "loss": 0.5903, "step": 20446 }, { "epoch": 0.7405650126765665, "grad_norm": 1.3729417612801371, "learning_rate": 3.3273706478497613e-06, "loss": 0.7076, "step": 20447 }, { "epoch": 0.7406012314378848, "grad_norm": 1.3211123396943336, "learning_rate": 3.3264969668327516e-06, "loss": 0.716, "step": 20448 }, { "epoch": 0.7406374501992032, "grad_norm": 1.047449641429771, "learning_rate": 3.325623377649009e-06, "loss": 0.7394, "step": 20449 }, { "epoch": 0.7406736689605216, "grad_norm": 1.3190044209631782, "learning_rate": 3.3247498803105517e-06, "loss": 0.6309, "step": 20450 }, { "epoch": 0.7407098877218399, "grad_norm": 1.0630345100124694, "learning_rate": 3.323876474829398e-06, "loss": 0.6537, "step": 20451 }, { "epoch": 0.7407461064831583, "grad_norm": 1.390138092817479, "learning_rate": 3.323003161217575e-06, "loss": 0.7009, "step": 20452 }, { "epoch": 0.7407823252444766, "grad_norm": 1.3242121417800623, "learning_rate": 3.322129939487094e-06, "loss": 0.6816, "step": 20453 }, { "epoch": 0.740818544005795, "grad_norm": 0.9438506069990746, "learning_rate": 3.321256809649973e-06, "loss": 0.6352, "step": 20454 }, { "epoch": 0.7408547627671134, "grad_norm": 1.059162540572213, "learning_rate": 3.3203837717182243e-06, "loss": 0.6655, "step": 20455 }, { "epoch": 0.7408909815284317, "grad_norm": 1.26037431102578, "learning_rate": 3.3195108257038665e-06, "loss": 0.7155, "step": 20456 }, { "epoch": 0.7409272002897501, "grad_norm": 1.279742744147643, "learning_rate": 3.318637971618911e-06, "loss": 0.6825, "step": 20457 }, { "epoch": 0.7409634190510684, "grad_norm": 1.286067345472934, "learning_rate": 3.317765209475363e-06, "loss": 0.6575, "step": 20458 }, { "epoch": 0.7409996378123869, "grad_norm": 0.9897693167351117, "learning_rate": 3.3168925392852414e-06, "loss": 0.709, "step": 20459 }, { "epoch": 0.7410358565737052, "grad_norm": 1.3877554101299485, "learning_rate": 3.3160199610605506e-06, "loss": 0.6859, "step": 20460 }, { "epoch": 0.7410720753350235, "grad_norm": 1.3446052097590502, "learning_rate": 3.315147474813297e-06, "loss": 0.6906, "step": 20461 }, { "epoch": 0.7411082940963419, "grad_norm": 1.5306478807870028, "learning_rate": 3.3142750805554867e-06, "loss": 0.686, "step": 20462 }, { "epoch": 0.7411445128576603, "grad_norm": 1.2608104180592672, "learning_rate": 3.313402778299127e-06, "loss": 0.5859, "step": 20463 }, { "epoch": 0.7411807316189787, "grad_norm": 1.4257629661171558, "learning_rate": 3.3125305680562226e-06, "loss": 0.7377, "step": 20464 }, { "epoch": 0.741216950380297, "grad_norm": 1.467515317412082, "learning_rate": 3.3116584498387726e-06, "loss": 0.6052, "step": 20465 }, { "epoch": 0.7412531691416153, "grad_norm": 1.2482993880852298, "learning_rate": 3.310786423658776e-06, "loss": 0.6997, "step": 20466 }, { "epoch": 0.7412893879029337, "grad_norm": 1.4807331533531527, "learning_rate": 3.3099144895282396e-06, "loss": 0.6755, "step": 20467 }, { "epoch": 0.7413256066642521, "grad_norm": 1.1992156582234876, "learning_rate": 3.3090426474591573e-06, "loss": 0.6717, "step": 20468 }, { "epoch": 0.7413618254255705, "grad_norm": 1.0099600522836105, "learning_rate": 3.308170897463525e-06, "loss": 0.6466, "step": 20469 }, { "epoch": 0.7413980441868888, "grad_norm": 1.296531067191592, "learning_rate": 3.307299239553344e-06, "loss": 0.6027, "step": 20470 }, { "epoch": 0.7414342629482071, "grad_norm": 1.3770577830588475, "learning_rate": 3.3064276737406064e-06, "loss": 0.6893, "step": 20471 }, { "epoch": 0.7414704817095256, "grad_norm": 1.3310728246806338, "learning_rate": 3.3055562000373055e-06, "loss": 0.6902, "step": 20472 }, { "epoch": 0.7415067004708439, "grad_norm": 1.5877904285570366, "learning_rate": 3.3046848184554294e-06, "loss": 0.7626, "step": 20473 }, { "epoch": 0.7415429192321623, "grad_norm": 1.2811835921627712, "learning_rate": 3.3038135290069773e-06, "loss": 0.6431, "step": 20474 }, { "epoch": 0.7415791379934806, "grad_norm": 1.312690874962613, "learning_rate": 3.3029423317039345e-06, "loss": 0.6557, "step": 20475 }, { "epoch": 0.741615356754799, "grad_norm": 1.4066589535687277, "learning_rate": 3.30207122655829e-06, "loss": 0.6933, "step": 20476 }, { "epoch": 0.7416515755161174, "grad_norm": 1.612625653893563, "learning_rate": 3.3012002135820278e-06, "loss": 0.7013, "step": 20477 }, { "epoch": 0.7416877942774357, "grad_norm": 1.057055789773406, "learning_rate": 3.30032929278714e-06, "loss": 0.6771, "step": 20478 }, { "epoch": 0.7417240130387541, "grad_norm": 1.2382513593200863, "learning_rate": 3.2994584641856074e-06, "loss": 0.7196, "step": 20479 }, { "epoch": 0.7417602318000724, "grad_norm": 1.4392930398585668, "learning_rate": 3.2985877277894107e-06, "loss": 0.6906, "step": 20480 }, { "epoch": 0.7417964505613908, "grad_norm": 1.319773667045167, "learning_rate": 3.297717083610539e-06, "loss": 0.6688, "step": 20481 }, { "epoch": 0.7418326693227092, "grad_norm": 1.0524866810417675, "learning_rate": 3.296846531660969e-06, "loss": 0.7313, "step": 20482 }, { "epoch": 0.7418688880840275, "grad_norm": 1.343982023622806, "learning_rate": 3.29597607195268e-06, "loss": 0.7512, "step": 20483 }, { "epoch": 0.7419051068453459, "grad_norm": 1.536011212696063, "learning_rate": 3.2951057044976487e-06, "loss": 0.7344, "step": 20484 }, { "epoch": 0.7419413256066643, "grad_norm": 1.3922469292317585, "learning_rate": 3.294235429307857e-06, "loss": 0.717, "step": 20485 }, { "epoch": 0.7419775443679826, "grad_norm": 1.2237821207796638, "learning_rate": 3.293365246395278e-06, "loss": 0.7261, "step": 20486 }, { "epoch": 0.742013763129301, "grad_norm": 1.3627630067823544, "learning_rate": 3.2924951557718877e-06, "loss": 0.676, "step": 20487 }, { "epoch": 0.7420499818906193, "grad_norm": 1.164212784410766, "learning_rate": 3.2916251574496526e-06, "loss": 0.6828, "step": 20488 }, { "epoch": 0.7420862006519378, "grad_norm": 1.7737827955439285, "learning_rate": 3.290755251440554e-06, "loss": 0.7394, "step": 20489 }, { "epoch": 0.7421224194132561, "grad_norm": 1.2949118038436018, "learning_rate": 3.289885437756558e-06, "loss": 0.6414, "step": 20490 }, { "epoch": 0.7421586381745744, "grad_norm": 1.2690795968007116, "learning_rate": 3.2890157164096315e-06, "loss": 0.6828, "step": 20491 }, { "epoch": 0.7421948569358928, "grad_norm": 1.2994199200510634, "learning_rate": 3.288146087411749e-06, "loss": 0.6615, "step": 20492 }, { "epoch": 0.7422310756972111, "grad_norm": 1.1761218606650121, "learning_rate": 3.2872765507748726e-06, "loss": 0.6529, "step": 20493 }, { "epoch": 0.7422672944585296, "grad_norm": 1.414359781987881, "learning_rate": 3.2864071065109702e-06, "loss": 0.5974, "step": 20494 }, { "epoch": 0.7423035132198479, "grad_norm": 1.2312546385033851, "learning_rate": 3.2855377546320023e-06, "loss": 0.6788, "step": 20495 }, { "epoch": 0.7423397319811662, "grad_norm": 1.303141371649305, "learning_rate": 3.2846684951499365e-06, "loss": 0.7219, "step": 20496 }, { "epoch": 0.7423759507424846, "grad_norm": 1.3724038176934656, "learning_rate": 3.2837993280767343e-06, "loss": 0.6791, "step": 20497 }, { "epoch": 0.742412169503803, "grad_norm": 1.3018375752903137, "learning_rate": 3.2829302534243537e-06, "loss": 0.6961, "step": 20498 }, { "epoch": 0.7424483882651214, "grad_norm": 1.268417733136047, "learning_rate": 3.2820612712047517e-06, "loss": 0.6959, "step": 20499 }, { "epoch": 0.7424846070264397, "grad_norm": 1.283904138163844, "learning_rate": 3.281192381429894e-06, "loss": 0.61, "step": 20500 }, { "epoch": 0.742520825787758, "grad_norm": 1.3841892261480586, "learning_rate": 3.280323584111732e-06, "loss": 0.7361, "step": 20501 }, { "epoch": 0.7425570445490765, "grad_norm": 1.3415702860328813, "learning_rate": 3.279454879262218e-06, "loss": 0.6589, "step": 20502 }, { "epoch": 0.7425932633103948, "grad_norm": 1.377243770198976, "learning_rate": 3.278586266893314e-06, "loss": 0.7103, "step": 20503 }, { "epoch": 0.7426294820717132, "grad_norm": 1.3040993233416263, "learning_rate": 3.2777177470169687e-06, "loss": 0.6885, "step": 20504 }, { "epoch": 0.7426657008330315, "grad_norm": 1.39652491690734, "learning_rate": 3.2768493196451346e-06, "loss": 0.7644, "step": 20505 }, { "epoch": 0.7427019195943498, "grad_norm": 1.3814783395200574, "learning_rate": 3.275980984789757e-06, "loss": 0.6422, "step": 20506 }, { "epoch": 0.7427381383556683, "grad_norm": 1.321991622319842, "learning_rate": 3.275112742462794e-06, "loss": 0.6448, "step": 20507 }, { "epoch": 0.7427743571169866, "grad_norm": 0.9689003373860386, "learning_rate": 3.2742445926761867e-06, "loss": 0.5779, "step": 20508 }, { "epoch": 0.742810575878305, "grad_norm": 1.2820935060447542, "learning_rate": 3.2733765354418822e-06, "loss": 0.6471, "step": 20509 }, { "epoch": 0.7428467946396233, "grad_norm": 1.2757896292832354, "learning_rate": 3.2725085707718295e-06, "loss": 0.6581, "step": 20510 }, { "epoch": 0.7428830134009416, "grad_norm": 0.9607233919517064, "learning_rate": 3.2716406986779704e-06, "loss": 0.6663, "step": 20511 }, { "epoch": 0.7429192321622601, "grad_norm": 1.2790949053710516, "learning_rate": 3.270772919172247e-06, "loss": 0.6535, "step": 20512 }, { "epoch": 0.7429554509235784, "grad_norm": 1.2680322336227445, "learning_rate": 3.269905232266597e-06, "loss": 0.6843, "step": 20513 }, { "epoch": 0.7429916696848968, "grad_norm": 1.3018116549941918, "learning_rate": 3.2690376379729694e-06, "loss": 0.691, "step": 20514 }, { "epoch": 0.7430278884462151, "grad_norm": 1.1831700547081163, "learning_rate": 3.268170136303297e-06, "loss": 0.619, "step": 20515 }, { "epoch": 0.7430641072075335, "grad_norm": 1.2910184499504818, "learning_rate": 3.267302727269519e-06, "loss": 0.7063, "step": 20516 }, { "epoch": 0.7431003259688519, "grad_norm": 1.263618974448049, "learning_rate": 3.266435410883568e-06, "loss": 0.6645, "step": 20517 }, { "epoch": 0.7431365447301702, "grad_norm": 1.4781428687501794, "learning_rate": 3.2655681871573864e-06, "loss": 0.6599, "step": 20518 }, { "epoch": 0.7431727634914886, "grad_norm": 1.3982706116878971, "learning_rate": 3.2647010561029034e-06, "loss": 0.6959, "step": 20519 }, { "epoch": 0.743208982252807, "grad_norm": 1.3898884210167568, "learning_rate": 3.2638340177320495e-06, "loss": 0.7658, "step": 20520 }, { "epoch": 0.7432452010141253, "grad_norm": 1.3612446253110413, "learning_rate": 3.262967072056761e-06, "loss": 0.7255, "step": 20521 }, { "epoch": 0.7432814197754437, "grad_norm": 1.3950088615396659, "learning_rate": 3.262100219088965e-06, "loss": 0.676, "step": 20522 }, { "epoch": 0.743317638536762, "grad_norm": 1.2418298498841247, "learning_rate": 3.2612334588405902e-06, "loss": 0.6579, "step": 20523 }, { "epoch": 0.7433538572980805, "grad_norm": 1.122214160775509, "learning_rate": 3.2603667913235616e-06, "loss": 0.5944, "step": 20524 }, { "epoch": 0.7433900760593988, "grad_norm": 1.0465109972617355, "learning_rate": 3.2595002165498114e-06, "loss": 0.6519, "step": 20525 }, { "epoch": 0.7434262948207171, "grad_norm": 1.408708354114152, "learning_rate": 3.25863373453126e-06, "loss": 0.7179, "step": 20526 }, { "epoch": 0.7434625135820355, "grad_norm": 1.4215228278653118, "learning_rate": 3.257767345279833e-06, "loss": 0.6582, "step": 20527 }, { "epoch": 0.7434987323433538, "grad_norm": 1.2970452642126002, "learning_rate": 3.2569010488074482e-06, "loss": 0.6749, "step": 20528 }, { "epoch": 0.7435349511046723, "grad_norm": 1.2793650921986304, "learning_rate": 3.2560348451260325e-06, "loss": 0.6988, "step": 20529 }, { "epoch": 0.7435711698659906, "grad_norm": 1.4330124949981482, "learning_rate": 3.2551687342475035e-06, "loss": 0.6406, "step": 20530 }, { "epoch": 0.7436073886273089, "grad_norm": 1.3558822071336771, "learning_rate": 3.254302716183776e-06, "loss": 0.7144, "step": 20531 }, { "epoch": 0.7436436073886273, "grad_norm": 1.100719956651187, "learning_rate": 3.2534367909467735e-06, "loss": 0.6067, "step": 20532 }, { "epoch": 0.7436798261499457, "grad_norm": 1.2811936160363846, "learning_rate": 3.252570958548409e-06, "loss": 0.6764, "step": 20533 }, { "epoch": 0.7437160449112641, "grad_norm": 1.2400067984174477, "learning_rate": 3.251705219000596e-06, "loss": 0.663, "step": 20534 }, { "epoch": 0.7437522636725824, "grad_norm": 1.4074052264425003, "learning_rate": 3.2508395723152464e-06, "loss": 0.6649, "step": 20535 }, { "epoch": 0.7437884824339007, "grad_norm": 1.444500209037456, "learning_rate": 3.2499740185042783e-06, "loss": 0.7365, "step": 20536 }, { "epoch": 0.7438247011952192, "grad_norm": 1.133099681527468, "learning_rate": 3.249108557579598e-06, "loss": 0.6514, "step": 20537 }, { "epoch": 0.7438609199565375, "grad_norm": 1.3209315801081274, "learning_rate": 3.2482431895531163e-06, "loss": 0.7126, "step": 20538 }, { "epoch": 0.7438971387178559, "grad_norm": 1.6587281468690365, "learning_rate": 3.2473779144367367e-06, "loss": 0.7478, "step": 20539 }, { "epoch": 0.7439333574791742, "grad_norm": 1.356107682131138, "learning_rate": 3.246512732242375e-06, "loss": 0.6466, "step": 20540 }, { "epoch": 0.7439695762404925, "grad_norm": 1.44567013517112, "learning_rate": 3.2456476429819315e-06, "loss": 0.6529, "step": 20541 }, { "epoch": 0.744005795001811, "grad_norm": 1.4574729950343133, "learning_rate": 3.2447826466673084e-06, "loss": 0.6432, "step": 20542 }, { "epoch": 0.7440420137631293, "grad_norm": 1.379677516300632, "learning_rate": 3.243917743310414e-06, "loss": 0.6546, "step": 20543 }, { "epoch": 0.7440782325244477, "grad_norm": 1.231014863561908, "learning_rate": 3.2430529329231497e-06, "loss": 0.7416, "step": 20544 }, { "epoch": 0.744114451285766, "grad_norm": 1.2694885525492487, "learning_rate": 3.2421882155174124e-06, "loss": 0.6122, "step": 20545 }, { "epoch": 0.7441506700470844, "grad_norm": 1.3759887984458028, "learning_rate": 3.2413235911051044e-06, "loss": 0.6816, "step": 20546 }, { "epoch": 0.7441868888084028, "grad_norm": 8.986394889001502, "learning_rate": 3.2404590596981224e-06, "loss": 0.7703, "step": 20547 }, { "epoch": 0.7442231075697211, "grad_norm": 1.298251894711846, "learning_rate": 3.2395946213083617e-06, "loss": 0.697, "step": 20548 }, { "epoch": 0.7442593263310395, "grad_norm": 0.9255225026785048, "learning_rate": 3.238730275947718e-06, "loss": 0.6673, "step": 20549 }, { "epoch": 0.7442955450923578, "grad_norm": 1.308472501062491, "learning_rate": 3.237866023628089e-06, "loss": 0.6414, "step": 20550 }, { "epoch": 0.7443317638536762, "grad_norm": 1.3432997979299837, "learning_rate": 3.2370018643613642e-06, "loss": 0.7336, "step": 20551 }, { "epoch": 0.7443679826149946, "grad_norm": 1.374596505065556, "learning_rate": 3.2361377981594344e-06, "loss": 0.6852, "step": 20552 }, { "epoch": 0.7444042013763129, "grad_norm": 1.486745443670768, "learning_rate": 3.235273825034195e-06, "loss": 0.7364, "step": 20553 }, { "epoch": 0.7444404201376313, "grad_norm": 1.3696725223008686, "learning_rate": 3.2344099449975308e-06, "loss": 0.7308, "step": 20554 }, { "epoch": 0.7444766388989497, "grad_norm": 0.984343364385568, "learning_rate": 3.2335461580613315e-06, "loss": 0.6481, "step": 20555 }, { "epoch": 0.744512857660268, "grad_norm": 1.2500280116340097, "learning_rate": 3.2326824642374787e-06, "loss": 0.6947, "step": 20556 }, { "epoch": 0.7445490764215864, "grad_norm": 1.2358885182381374, "learning_rate": 3.2318188635378655e-06, "loss": 0.6782, "step": 20557 }, { "epoch": 0.7445852951829047, "grad_norm": 1.210094968655275, "learning_rate": 3.2309553559743712e-06, "loss": 0.6697, "step": 20558 }, { "epoch": 0.7446215139442232, "grad_norm": 1.4675486101203798, "learning_rate": 3.230091941558879e-06, "loss": 0.6616, "step": 20559 }, { "epoch": 0.7446577327055415, "grad_norm": 1.3870626358992495, "learning_rate": 3.2292286203032674e-06, "loss": 0.7518, "step": 20560 }, { "epoch": 0.7446939514668598, "grad_norm": 1.3010909825800994, "learning_rate": 3.2283653922194223e-06, "loss": 0.6651, "step": 20561 }, { "epoch": 0.7447301702281782, "grad_norm": 1.0808434819045474, "learning_rate": 3.2275022573192192e-06, "loss": 0.7492, "step": 20562 }, { "epoch": 0.7447663889894965, "grad_norm": 1.1060425080600158, "learning_rate": 3.226639215614533e-06, "loss": 0.6344, "step": 20563 }, { "epoch": 0.744802607750815, "grad_norm": 1.2414336608719712, "learning_rate": 3.2257762671172455e-06, "loss": 0.6359, "step": 20564 }, { "epoch": 0.7448388265121333, "grad_norm": 1.5224751795407843, "learning_rate": 3.224913411839229e-06, "loss": 0.8281, "step": 20565 }, { "epoch": 0.7448750452734516, "grad_norm": 1.3149836335482716, "learning_rate": 3.224050649792357e-06, "loss": 0.7107, "step": 20566 }, { "epoch": 0.74491126403477, "grad_norm": 1.3867673260951912, "learning_rate": 3.2231879809885e-06, "loss": 0.6678, "step": 20567 }, { "epoch": 0.7449474827960884, "grad_norm": 1.0049212838103485, "learning_rate": 3.2223254054395336e-06, "loss": 0.7531, "step": 20568 }, { "epoch": 0.7449837015574068, "grad_norm": 1.3795089890246626, "learning_rate": 3.2214629231573248e-06, "loss": 0.7421, "step": 20569 }, { "epoch": 0.7450199203187251, "grad_norm": 5.256089496429953, "learning_rate": 3.2206005341537417e-06, "loss": 0.6638, "step": 20570 }, { "epoch": 0.7450561390800434, "grad_norm": 1.4021127784212852, "learning_rate": 3.2197382384406493e-06, "loss": 0.7279, "step": 20571 }, { "epoch": 0.7450923578413619, "grad_norm": 1.3588899659677596, "learning_rate": 3.21887603602992e-06, "loss": 0.6773, "step": 20572 }, { "epoch": 0.7451285766026802, "grad_norm": 0.9021968485556199, "learning_rate": 3.2180139269334145e-06, "loss": 0.637, "step": 20573 }, { "epoch": 0.7451647953639986, "grad_norm": 1.805225185907932, "learning_rate": 3.2171519111629933e-06, "loss": 0.6432, "step": 20574 }, { "epoch": 0.7452010141253169, "grad_norm": 1.0124487588254396, "learning_rate": 3.2162899887305254e-06, "loss": 0.7146, "step": 20575 }, { "epoch": 0.7452372328866352, "grad_norm": 1.2801058919291335, "learning_rate": 3.2154281596478675e-06, "loss": 0.7118, "step": 20576 }, { "epoch": 0.7452734516479537, "grad_norm": 1.1535802152354064, "learning_rate": 3.21456642392688e-06, "loss": 0.6681, "step": 20577 }, { "epoch": 0.745309670409272, "grad_norm": 1.3470300159287185, "learning_rate": 3.2137047815794177e-06, "loss": 0.7044, "step": 20578 }, { "epoch": 0.7453458891705904, "grad_norm": 1.3759520144739947, "learning_rate": 3.212843232617343e-06, "loss": 0.6854, "step": 20579 }, { "epoch": 0.7453821079319087, "grad_norm": 1.1128813789861614, "learning_rate": 3.2119817770525086e-06, "loss": 0.7021, "step": 20580 }, { "epoch": 0.745418326693227, "grad_norm": 0.9953852811262925, "learning_rate": 3.2111204148967713e-06, "loss": 0.6935, "step": 20581 }, { "epoch": 0.7454545454545455, "grad_norm": 1.313021783908463, "learning_rate": 3.210259146161978e-06, "loss": 0.6569, "step": 20582 }, { "epoch": 0.7454907642158638, "grad_norm": 0.8876527467469298, "learning_rate": 3.2093979708599877e-06, "loss": 0.6899, "step": 20583 }, { "epoch": 0.7455269829771822, "grad_norm": 1.057880335631976, "learning_rate": 3.208536889002648e-06, "loss": 0.6754, "step": 20584 }, { "epoch": 0.7455632017385005, "grad_norm": 1.536527872288284, "learning_rate": 3.2076759006018055e-06, "loss": 0.7325, "step": 20585 }, { "epoch": 0.7455994204998189, "grad_norm": 1.4153737033812197, "learning_rate": 3.206815005669315e-06, "loss": 0.703, "step": 20586 }, { "epoch": 0.7456356392611373, "grad_norm": 1.3944784449616094, "learning_rate": 3.2059542042170177e-06, "loss": 0.6917, "step": 20587 }, { "epoch": 0.7456718580224556, "grad_norm": 1.2883986054918755, "learning_rate": 3.2050934962567604e-06, "loss": 0.7068, "step": 20588 }, { "epoch": 0.745708076783774, "grad_norm": 1.2758754747697165, "learning_rate": 3.204232881800384e-06, "loss": 0.688, "step": 20589 }, { "epoch": 0.7457442955450924, "grad_norm": 1.0094641472010484, "learning_rate": 3.203372360859739e-06, "loss": 0.6368, "step": 20590 }, { "epoch": 0.7457805143064107, "grad_norm": 0.9418712721427754, "learning_rate": 3.20251193344666e-06, "loss": 0.6699, "step": 20591 }, { "epoch": 0.7458167330677291, "grad_norm": 1.5858680599494819, "learning_rate": 3.201651599572991e-06, "loss": 0.735, "step": 20592 }, { "epoch": 0.7458529518290474, "grad_norm": 1.4662649254230296, "learning_rate": 3.2007913592505668e-06, "loss": 0.7202, "step": 20593 }, { "epoch": 0.7458891705903659, "grad_norm": 1.0040937582101146, "learning_rate": 3.1999312124912306e-06, "loss": 0.7107, "step": 20594 }, { "epoch": 0.7459253893516842, "grad_norm": 1.584308246611429, "learning_rate": 3.1990711593068158e-06, "loss": 0.6835, "step": 20595 }, { "epoch": 0.7459616081130025, "grad_norm": 0.9623308453983326, "learning_rate": 3.1982111997091536e-06, "loss": 0.7057, "step": 20596 }, { "epoch": 0.7459978268743209, "grad_norm": 1.0958015586321888, "learning_rate": 3.197351333710086e-06, "loss": 0.657, "step": 20597 }, { "epoch": 0.7460340456356392, "grad_norm": 2.0498707830160177, "learning_rate": 3.196491561321442e-06, "loss": 0.6882, "step": 20598 }, { "epoch": 0.7460702643969577, "grad_norm": 1.3723637419832657, "learning_rate": 3.1956318825550516e-06, "loss": 0.638, "step": 20599 }, { "epoch": 0.746106483158276, "grad_norm": 1.3590563943348155, "learning_rate": 3.194772297422741e-06, "loss": 0.6444, "step": 20600 }, { "epoch": 0.7461427019195943, "grad_norm": 1.451704538477165, "learning_rate": 3.1939128059363477e-06, "loss": 0.686, "step": 20601 }, { "epoch": 0.7461789206809127, "grad_norm": 1.3264020192000974, "learning_rate": 3.1930534081076947e-06, "loss": 0.7078, "step": 20602 }, { "epoch": 0.7462151394422311, "grad_norm": 0.9774227581066107, "learning_rate": 3.192194103948607e-06, "loss": 0.6626, "step": 20603 }, { "epoch": 0.7462513582035495, "grad_norm": 1.1999507698376624, "learning_rate": 3.1913348934709076e-06, "loss": 0.663, "step": 20604 }, { "epoch": 0.7462875769648678, "grad_norm": 1.3372863623142257, "learning_rate": 3.1904757766864258e-06, "loss": 0.6, "step": 20605 }, { "epoch": 0.7463237957261861, "grad_norm": 1.2662485096805949, "learning_rate": 3.1896167536069813e-06, "loss": 0.7017, "step": 20606 }, { "epoch": 0.7463600144875046, "grad_norm": 1.1563677973443305, "learning_rate": 3.1887578242443895e-06, "loss": 0.6906, "step": 20607 }, { "epoch": 0.7463962332488229, "grad_norm": 1.3181211100837897, "learning_rate": 3.1878989886104792e-06, "loss": 0.6701, "step": 20608 }, { "epoch": 0.7464324520101413, "grad_norm": 1.3727379203039813, "learning_rate": 3.187040246717065e-06, "loss": 0.7341, "step": 20609 }, { "epoch": 0.7464686707714596, "grad_norm": 0.9244810767277162, "learning_rate": 3.1861815985759625e-06, "loss": 0.7108, "step": 20610 }, { "epoch": 0.7465048895327779, "grad_norm": 1.4212878090277739, "learning_rate": 3.185323044198986e-06, "loss": 0.7147, "step": 20611 }, { "epoch": 0.7465411082940964, "grad_norm": 1.510051412169986, "learning_rate": 3.184464583597955e-06, "loss": 0.775, "step": 20612 }, { "epoch": 0.7465773270554147, "grad_norm": 1.1209855819773906, "learning_rate": 3.1836062167846803e-06, "loss": 0.6537, "step": 20613 }, { "epoch": 0.7466135458167331, "grad_norm": 1.3744000696128855, "learning_rate": 3.18274794377097e-06, "loss": 0.7991, "step": 20614 }, { "epoch": 0.7466497645780514, "grad_norm": 1.367698603871708, "learning_rate": 3.181889764568641e-06, "loss": 0.7015, "step": 20615 }, { "epoch": 0.7466859833393698, "grad_norm": 1.364475478637081, "learning_rate": 3.1810316791895003e-06, "loss": 0.6119, "step": 20616 }, { "epoch": 0.7467222021006882, "grad_norm": 1.3872921859455047, "learning_rate": 3.180173687645355e-06, "loss": 0.7154, "step": 20617 }, { "epoch": 0.7467584208620065, "grad_norm": 0.9796392665772439, "learning_rate": 3.179315789948009e-06, "loss": 0.7448, "step": 20618 }, { "epoch": 0.7467946396233249, "grad_norm": 1.293363698844399, "learning_rate": 3.1784579861092744e-06, "loss": 0.7518, "step": 20619 }, { "epoch": 0.7468308583846432, "grad_norm": 0.9903561698953629, "learning_rate": 3.177600276140952e-06, "loss": 0.659, "step": 20620 }, { "epoch": 0.7468670771459616, "grad_norm": 1.0784725241021735, "learning_rate": 3.1767426600548434e-06, "loss": 0.7037, "step": 20621 }, { "epoch": 0.74690329590728, "grad_norm": 0.9847650452955992, "learning_rate": 3.1758851378627485e-06, "loss": 0.7164, "step": 20622 }, { "epoch": 0.7469395146685983, "grad_norm": 1.5998287958697504, "learning_rate": 3.1750277095764747e-06, "loss": 0.6983, "step": 20623 }, { "epoch": 0.7469757334299167, "grad_norm": 1.4114442369477391, "learning_rate": 3.1741703752078155e-06, "loss": 0.5918, "step": 20624 }, { "epoch": 0.7470119521912351, "grad_norm": 1.7353865325315165, "learning_rate": 3.1733131347685665e-06, "loss": 0.6539, "step": 20625 }, { "epoch": 0.7470481709525534, "grad_norm": 1.2955554926641482, "learning_rate": 3.1724559882705306e-06, "loss": 0.6687, "step": 20626 }, { "epoch": 0.7470843897138718, "grad_norm": 1.3214235173937903, "learning_rate": 3.1715989357254993e-06, "loss": 0.6647, "step": 20627 }, { "epoch": 0.7471206084751901, "grad_norm": 1.324851721016306, "learning_rate": 3.1707419771452663e-06, "loss": 0.6679, "step": 20628 }, { "epoch": 0.7471568272365086, "grad_norm": 1.2391133919880695, "learning_rate": 3.169885112541622e-06, "loss": 0.6508, "step": 20629 }, { "epoch": 0.7471930459978269, "grad_norm": 1.3214933321364821, "learning_rate": 3.1690283419263634e-06, "loss": 0.7292, "step": 20630 }, { "epoch": 0.7472292647591452, "grad_norm": 1.1902178952638351, "learning_rate": 3.1681716653112757e-06, "loss": 0.712, "step": 20631 }, { "epoch": 0.7472654835204636, "grad_norm": 1.1362571287213894, "learning_rate": 3.16731508270815e-06, "loss": 0.7414, "step": 20632 }, { "epoch": 0.7473017022817819, "grad_norm": 1.2679511514724562, "learning_rate": 3.1664585941287685e-06, "loss": 0.6685, "step": 20633 }, { "epoch": 0.7473379210431004, "grad_norm": 1.2304432154836762, "learning_rate": 3.1656021995849252e-06, "loss": 0.6456, "step": 20634 }, { "epoch": 0.7473741398044187, "grad_norm": 1.4189372984626385, "learning_rate": 3.1647458990884007e-06, "loss": 0.7052, "step": 20635 }, { "epoch": 0.747410358565737, "grad_norm": 1.5666749478805677, "learning_rate": 3.1638896926509755e-06, "loss": 0.6939, "step": 20636 }, { "epoch": 0.7474465773270554, "grad_norm": 1.3136337479590563, "learning_rate": 3.163033580284437e-06, "loss": 0.6754, "step": 20637 }, { "epoch": 0.7474827960883738, "grad_norm": 1.3342277666458024, "learning_rate": 3.1621775620005647e-06, "loss": 0.6959, "step": 20638 }, { "epoch": 0.7475190148496922, "grad_norm": 1.45802555469467, "learning_rate": 3.161321637811138e-06, "loss": 0.7204, "step": 20639 }, { "epoch": 0.7475552336110105, "grad_norm": 1.0697848867962727, "learning_rate": 3.1604658077279306e-06, "loss": 0.766, "step": 20640 }, { "epoch": 0.7475914523723288, "grad_norm": 1.4450931215041742, "learning_rate": 3.1596100717627265e-06, "loss": 0.7013, "step": 20641 }, { "epoch": 0.7476276711336473, "grad_norm": 1.4072222693044918, "learning_rate": 3.1587544299272985e-06, "loss": 0.7058, "step": 20642 }, { "epoch": 0.7476638898949656, "grad_norm": 1.2614538398865298, "learning_rate": 3.157898882233421e-06, "loss": 0.7217, "step": 20643 }, { "epoch": 0.747700108656284, "grad_norm": 1.3117657385777182, "learning_rate": 3.157043428692864e-06, "loss": 0.6622, "step": 20644 }, { "epoch": 0.7477363274176023, "grad_norm": 1.316674391239354, "learning_rate": 3.156188069317404e-06, "loss": 0.7781, "step": 20645 }, { "epoch": 0.7477725461789206, "grad_norm": 1.2917151211695344, "learning_rate": 3.1553328041188113e-06, "loss": 0.7692, "step": 20646 }, { "epoch": 0.7478087649402391, "grad_norm": 0.9816910644798839, "learning_rate": 3.154477633108849e-06, "loss": 0.6412, "step": 20647 }, { "epoch": 0.7478449837015574, "grad_norm": 1.404247563601228, "learning_rate": 3.153622556299294e-06, "loss": 0.7704, "step": 20648 }, { "epoch": 0.7478812024628758, "grad_norm": 1.3268677286199493, "learning_rate": 3.152767573701907e-06, "loss": 0.6383, "step": 20649 }, { "epoch": 0.7479174212241941, "grad_norm": 1.4587928922079167, "learning_rate": 3.1519126853284552e-06, "loss": 0.7007, "step": 20650 }, { "epoch": 0.7479536399855125, "grad_norm": 0.9246935666886974, "learning_rate": 3.1510578911906998e-06, "loss": 0.6378, "step": 20651 }, { "epoch": 0.7479898587468309, "grad_norm": 1.4162094812223927, "learning_rate": 3.150203191300407e-06, "loss": 0.6477, "step": 20652 }, { "epoch": 0.7480260775081492, "grad_norm": 1.3909326531812785, "learning_rate": 3.1493485856693386e-06, "loss": 0.6163, "step": 20653 }, { "epoch": 0.7480622962694676, "grad_norm": 1.4526034471564313, "learning_rate": 3.148494074309253e-06, "loss": 0.7114, "step": 20654 }, { "epoch": 0.748098515030786, "grad_norm": 1.2610621818187682, "learning_rate": 3.1476396572319057e-06, "loss": 0.6598, "step": 20655 }, { "epoch": 0.7481347337921043, "grad_norm": 1.3363018034132708, "learning_rate": 3.1467853344490618e-06, "loss": 0.638, "step": 20656 }, { "epoch": 0.7481709525534227, "grad_norm": 0.8933904191319652, "learning_rate": 3.1459311059724716e-06, "loss": 0.7343, "step": 20657 }, { "epoch": 0.748207171314741, "grad_norm": 1.362122321826262, "learning_rate": 3.145076971813891e-06, "loss": 0.6446, "step": 20658 }, { "epoch": 0.7482433900760594, "grad_norm": 1.3547056850831938, "learning_rate": 3.1442229319850758e-06, "loss": 0.6852, "step": 20659 }, { "epoch": 0.7482796088373778, "grad_norm": 1.0926086395372436, "learning_rate": 3.1433689864977777e-06, "loss": 0.7866, "step": 20660 }, { "epoch": 0.7483158275986961, "grad_norm": 1.313951158399132, "learning_rate": 3.1425151353637472e-06, "loss": 0.6921, "step": 20661 }, { "epoch": 0.7483520463600145, "grad_norm": 1.594036623481695, "learning_rate": 3.1416613785947313e-06, "loss": 0.6438, "step": 20662 }, { "epoch": 0.7483882651213328, "grad_norm": 1.280559603835143, "learning_rate": 3.140807716202483e-06, "loss": 0.6774, "step": 20663 }, { "epoch": 0.7484244838826513, "grad_norm": 1.4314295533778336, "learning_rate": 3.1399541481987484e-06, "loss": 0.6964, "step": 20664 }, { "epoch": 0.7484607026439696, "grad_norm": 1.5395498593343047, "learning_rate": 3.139100674595272e-06, "loss": 0.7071, "step": 20665 }, { "epoch": 0.7484969214052879, "grad_norm": 1.410193246002133, "learning_rate": 3.1382472954037967e-06, "loss": 0.6393, "step": 20666 }, { "epoch": 0.7485331401666063, "grad_norm": 1.318954788257573, "learning_rate": 3.13739401063607e-06, "loss": 0.7099, "step": 20667 }, { "epoch": 0.7485693589279246, "grad_norm": 1.4558325085669916, "learning_rate": 3.1365408203038326e-06, "loss": 0.6962, "step": 20668 }, { "epoch": 0.7486055776892431, "grad_norm": 1.0735160932173076, "learning_rate": 3.135687724418821e-06, "loss": 0.6646, "step": 20669 }, { "epoch": 0.7486417964505614, "grad_norm": 1.39705977556566, "learning_rate": 3.13483472299278e-06, "loss": 0.6297, "step": 20670 }, { "epoch": 0.7486780152118797, "grad_norm": 1.0652741712921772, "learning_rate": 3.133981816037447e-06, "loss": 0.6714, "step": 20671 }, { "epoch": 0.7487142339731981, "grad_norm": 1.0222418546623362, "learning_rate": 3.1331290035645555e-06, "loss": 0.6198, "step": 20672 }, { "epoch": 0.7487504527345165, "grad_norm": 1.4422320259325307, "learning_rate": 3.13227628558584e-06, "loss": 0.6936, "step": 20673 }, { "epoch": 0.7487866714958349, "grad_norm": 0.9692934276701468, "learning_rate": 3.131423662113041e-06, "loss": 0.6399, "step": 20674 }, { "epoch": 0.7488228902571532, "grad_norm": 1.304953263962731, "learning_rate": 3.130571133157888e-06, "loss": 0.6295, "step": 20675 }, { "epoch": 0.7488591090184715, "grad_norm": 0.893614763672734, "learning_rate": 3.129718698732108e-06, "loss": 0.6286, "step": 20676 }, { "epoch": 0.74889532777979, "grad_norm": 1.3842612519581974, "learning_rate": 3.12886635884744e-06, "loss": 0.6748, "step": 20677 }, { "epoch": 0.7489315465411083, "grad_norm": 1.3478075032425703, "learning_rate": 3.1280141135156074e-06, "loss": 0.65, "step": 20678 }, { "epoch": 0.7489677653024267, "grad_norm": 1.556122431543887, "learning_rate": 3.1271619627483385e-06, "loss": 0.6612, "step": 20679 }, { "epoch": 0.749003984063745, "grad_norm": 1.3035278177419873, "learning_rate": 3.1263099065573576e-06, "loss": 0.6167, "step": 20680 }, { "epoch": 0.7490402028250633, "grad_norm": 1.0049437140317863, "learning_rate": 3.1254579449543955e-06, "loss": 0.6832, "step": 20681 }, { "epoch": 0.7490764215863818, "grad_norm": 1.465243061953172, "learning_rate": 3.124606077951171e-06, "loss": 0.6776, "step": 20682 }, { "epoch": 0.7491126403477001, "grad_norm": 1.3459839050409765, "learning_rate": 3.12375430555941e-06, "loss": 0.6741, "step": 20683 }, { "epoch": 0.7491488591090185, "grad_norm": 1.3800080654912368, "learning_rate": 3.122902627790828e-06, "loss": 0.7603, "step": 20684 }, { "epoch": 0.7491850778703368, "grad_norm": 1.4392857855445336, "learning_rate": 3.122051044657152e-06, "loss": 0.6772, "step": 20685 }, { "epoch": 0.7492212966316552, "grad_norm": 1.3077276965189029, "learning_rate": 3.121199556170096e-06, "loss": 0.6864, "step": 20686 }, { "epoch": 0.7492575153929736, "grad_norm": 1.0466773587520575, "learning_rate": 3.1203481623413755e-06, "loss": 0.6263, "step": 20687 }, { "epoch": 0.7492937341542919, "grad_norm": 1.4238503320948017, "learning_rate": 3.1194968631827127e-06, "loss": 0.7493, "step": 20688 }, { "epoch": 0.7493299529156103, "grad_norm": 1.371479448859946, "learning_rate": 3.118645658705819e-06, "loss": 0.6455, "step": 20689 }, { "epoch": 0.7493661716769286, "grad_norm": 1.1889384053324146, "learning_rate": 3.117794548922408e-06, "loss": 0.698, "step": 20690 }, { "epoch": 0.749402390438247, "grad_norm": 1.3930277666949948, "learning_rate": 3.1169435338441878e-06, "loss": 0.6968, "step": 20691 }, { "epoch": 0.7494386091995654, "grad_norm": 1.3254521837113222, "learning_rate": 3.116092613482875e-06, "loss": 0.6484, "step": 20692 }, { "epoch": 0.7494748279608837, "grad_norm": 1.381027489320039, "learning_rate": 3.1152417878501775e-06, "loss": 0.6475, "step": 20693 }, { "epoch": 0.7495110467222021, "grad_norm": 1.028165138127923, "learning_rate": 3.114391056957802e-06, "loss": 0.6976, "step": 20694 }, { "epoch": 0.7495472654835205, "grad_norm": 1.2099897648299884, "learning_rate": 3.1135404208174524e-06, "loss": 0.6139, "step": 20695 }, { "epoch": 0.7495834842448388, "grad_norm": 1.0521582850042672, "learning_rate": 3.112689879440841e-06, "loss": 0.6207, "step": 20696 }, { "epoch": 0.7496197030061572, "grad_norm": 1.356472157239474, "learning_rate": 3.1118394328396683e-06, "loss": 0.7233, "step": 20697 }, { "epoch": 0.7496559217674755, "grad_norm": 1.494872528529634, "learning_rate": 3.110989081025635e-06, "loss": 0.7006, "step": 20698 }, { "epoch": 0.749692140528794, "grad_norm": 1.7104005226784906, "learning_rate": 3.110138824010448e-06, "loss": 0.6406, "step": 20699 }, { "epoch": 0.7497283592901123, "grad_norm": 1.3618696461553057, "learning_rate": 3.109288661805804e-06, "loss": 0.7033, "step": 20700 }, { "epoch": 0.7497645780514306, "grad_norm": 1.3094607761836161, "learning_rate": 3.108438594423404e-06, "loss": 0.7204, "step": 20701 }, { "epoch": 0.749800796812749, "grad_norm": 1.2554477362604044, "learning_rate": 3.107588621874941e-06, "loss": 0.6754, "step": 20702 }, { "epoch": 0.7498370155740673, "grad_norm": 1.792510511187602, "learning_rate": 3.1067387441721164e-06, "loss": 0.7089, "step": 20703 }, { "epoch": 0.7498732343353858, "grad_norm": 1.457717483343317, "learning_rate": 3.105888961326625e-06, "loss": 0.6442, "step": 20704 }, { "epoch": 0.7499094530967041, "grad_norm": 1.2741296567765807, "learning_rate": 3.1050392733501587e-06, "loss": 0.6747, "step": 20705 }, { "epoch": 0.7499456718580224, "grad_norm": 1.4320100123011148, "learning_rate": 3.104189680254408e-06, "loss": 0.7019, "step": 20706 }, { "epoch": 0.7499818906193408, "grad_norm": 1.3509494157230777, "learning_rate": 3.1033401820510687e-06, "loss": 0.6673, "step": 20707 }, { "epoch": 0.7500181093806592, "grad_norm": 1.2608834726016431, "learning_rate": 3.1024907787518287e-06, "loss": 0.6834, "step": 20708 }, { "epoch": 0.7500543281419776, "grad_norm": 1.2841377472646538, "learning_rate": 3.1016414703683737e-06, "loss": 0.6612, "step": 20709 }, { "epoch": 0.7500905469032959, "grad_norm": 1.5866796842768676, "learning_rate": 3.1007922569123962e-06, "loss": 0.6975, "step": 20710 }, { "epoch": 0.7501267656646142, "grad_norm": 1.2914926755593525, "learning_rate": 3.0999431383955802e-06, "loss": 0.7199, "step": 20711 }, { "epoch": 0.7501629844259327, "grad_norm": 1.4018402085491786, "learning_rate": 3.0990941148296084e-06, "loss": 0.7169, "step": 20712 }, { "epoch": 0.750199203187251, "grad_norm": 1.3110613317735829, "learning_rate": 3.098245186226162e-06, "loss": 0.6773, "step": 20713 }, { "epoch": 0.7502354219485694, "grad_norm": 1.0505190029817968, "learning_rate": 3.097396352596931e-06, "loss": 0.6794, "step": 20714 }, { "epoch": 0.7502716407098877, "grad_norm": 1.1011953980252283, "learning_rate": 3.09654761395359e-06, "loss": 0.6945, "step": 20715 }, { "epoch": 0.750307859471206, "grad_norm": 1.2443879396706095, "learning_rate": 3.09569897030782e-06, "loss": 0.6349, "step": 20716 }, { "epoch": 0.7503440782325245, "grad_norm": 1.3088871712311358, "learning_rate": 3.094850421671295e-06, "loss": 0.6892, "step": 20717 }, { "epoch": 0.7503802969938428, "grad_norm": 1.5085544267939315, "learning_rate": 3.0940019680557e-06, "loss": 0.6644, "step": 20718 }, { "epoch": 0.7504165157551612, "grad_norm": 1.2842241613135046, "learning_rate": 3.0931536094727054e-06, "loss": 0.6841, "step": 20719 }, { "epoch": 0.7504527345164795, "grad_norm": 1.3632827648793853, "learning_rate": 3.0923053459339824e-06, "loss": 0.7141, "step": 20720 }, { "epoch": 0.7504889532777979, "grad_norm": 1.3651505268527528, "learning_rate": 3.0914571774512114e-06, "loss": 0.5884, "step": 20721 }, { "epoch": 0.7505251720391163, "grad_norm": 1.0493309002022166, "learning_rate": 3.0906091040360596e-06, "loss": 0.7039, "step": 20722 }, { "epoch": 0.7505613908004346, "grad_norm": 1.0783103673592405, "learning_rate": 3.089761125700198e-06, "loss": 0.6871, "step": 20723 }, { "epoch": 0.750597609561753, "grad_norm": 1.51181764119028, "learning_rate": 3.0889132424552914e-06, "loss": 0.7624, "step": 20724 }, { "epoch": 0.7506338283230714, "grad_norm": 1.191520098226103, "learning_rate": 3.088065454313015e-06, "loss": 0.7285, "step": 20725 }, { "epoch": 0.7506700470843897, "grad_norm": 1.1109349433832454, "learning_rate": 3.087217761285032e-06, "loss": 0.7059, "step": 20726 }, { "epoch": 0.7507062658457081, "grad_norm": 1.0393910175226875, "learning_rate": 3.0863701633830055e-06, "loss": 0.7203, "step": 20727 }, { "epoch": 0.7507424846070264, "grad_norm": 1.4366801686626967, "learning_rate": 3.0855226606185975e-06, "loss": 0.6968, "step": 20728 }, { "epoch": 0.7507787033683448, "grad_norm": 1.3623014456871045, "learning_rate": 3.0846752530034764e-06, "loss": 0.6774, "step": 20729 }, { "epoch": 0.7508149221296632, "grad_norm": 1.3515199071589479, "learning_rate": 3.0838279405492997e-06, "loss": 0.7482, "step": 20730 }, { "epoch": 0.7508511408909815, "grad_norm": 1.4230150987562722, "learning_rate": 3.0829807232677255e-06, "loss": 0.6602, "step": 20731 }, { "epoch": 0.7508873596522999, "grad_norm": 1.435484229556931, "learning_rate": 3.0821336011704174e-06, "loss": 0.6469, "step": 20732 }, { "epoch": 0.7509235784136182, "grad_norm": 1.3923937379876872, "learning_rate": 3.081286574269029e-06, "loss": 0.7257, "step": 20733 }, { "epoch": 0.7509597971749367, "grad_norm": 1.2286305563312334, "learning_rate": 3.080439642575217e-06, "loss": 0.6938, "step": 20734 }, { "epoch": 0.750996015936255, "grad_norm": 1.3305615752246192, "learning_rate": 3.079592806100632e-06, "loss": 0.7015, "step": 20735 }, { "epoch": 0.7510322346975733, "grad_norm": 1.5519757678907498, "learning_rate": 3.0787460648569344e-06, "loss": 0.7533, "step": 20736 }, { "epoch": 0.7510684534588917, "grad_norm": 1.3301494133956304, "learning_rate": 3.0778994188557722e-06, "loss": 0.7069, "step": 20737 }, { "epoch": 0.75110467222021, "grad_norm": 1.3075227190497807, "learning_rate": 3.077052868108793e-06, "loss": 0.645, "step": 20738 }, { "epoch": 0.7511408909815285, "grad_norm": 1.496410565409393, "learning_rate": 3.0762064126276525e-06, "loss": 0.6244, "step": 20739 }, { "epoch": 0.7511771097428468, "grad_norm": 1.390508867890246, "learning_rate": 3.0753600524239956e-06, "loss": 0.7049, "step": 20740 }, { "epoch": 0.7512133285041651, "grad_norm": 1.47220445654142, "learning_rate": 3.07451378750947e-06, "loss": 0.7396, "step": 20741 }, { "epoch": 0.7512495472654835, "grad_norm": 0.9846622168300033, "learning_rate": 3.073667617895716e-06, "loss": 0.6813, "step": 20742 }, { "epoch": 0.7512857660268019, "grad_norm": 1.2662891883809144, "learning_rate": 3.072821543594384e-06, "loss": 0.6257, "step": 20743 }, { "epoch": 0.7513219847881203, "grad_norm": 1.3695825205137209, "learning_rate": 3.0719755646171146e-06, "loss": 0.7136, "step": 20744 }, { "epoch": 0.7513582035494386, "grad_norm": 1.4516099221179342, "learning_rate": 3.0711296809755497e-06, "loss": 0.7377, "step": 20745 }, { "epoch": 0.7513944223107569, "grad_norm": 1.1586080998238155, "learning_rate": 3.070283892681325e-06, "loss": 0.7165, "step": 20746 }, { "epoch": 0.7514306410720754, "grad_norm": 1.439871425745598, "learning_rate": 3.0694381997460865e-06, "loss": 0.6669, "step": 20747 }, { "epoch": 0.7514668598333937, "grad_norm": 1.0027394160681553, "learning_rate": 3.0685926021814673e-06, "loss": 0.6834, "step": 20748 }, { "epoch": 0.7515030785947121, "grad_norm": 1.3808222594392183, "learning_rate": 3.067747099999101e-06, "loss": 0.6836, "step": 20749 }, { "epoch": 0.7515392973560304, "grad_norm": 1.4037409273507868, "learning_rate": 3.06690169321063e-06, "loss": 0.7929, "step": 20750 }, { "epoch": 0.7515755161173487, "grad_norm": 1.3319612761545727, "learning_rate": 3.0660563818276836e-06, "loss": 0.7261, "step": 20751 }, { "epoch": 0.7516117348786672, "grad_norm": 1.3819919039859858, "learning_rate": 3.0652111658618933e-06, "loss": 0.7122, "step": 20752 }, { "epoch": 0.7516479536399855, "grad_norm": 1.2969102578224965, "learning_rate": 3.064366045324888e-06, "loss": 0.6668, "step": 20753 }, { "epoch": 0.7516841724013039, "grad_norm": 1.1144392945954538, "learning_rate": 3.0635210202283037e-06, "loss": 0.7067, "step": 20754 }, { "epoch": 0.7517203911626222, "grad_norm": 1.0833437334043936, "learning_rate": 3.0626760905837647e-06, "loss": 0.7478, "step": 20755 }, { "epoch": 0.7517566099239406, "grad_norm": 1.6972628682314976, "learning_rate": 3.0618312564028984e-06, "loss": 0.7565, "step": 20756 }, { "epoch": 0.751792828685259, "grad_norm": 0.9368387101270692, "learning_rate": 3.0609865176973297e-06, "loss": 0.6538, "step": 20757 }, { "epoch": 0.7518290474465773, "grad_norm": 4.060186935869086, "learning_rate": 3.0601418744786848e-06, "loss": 0.6776, "step": 20758 }, { "epoch": 0.7518652662078957, "grad_norm": 1.984559238321473, "learning_rate": 3.0592973267585844e-06, "loss": 0.6559, "step": 20759 }, { "epoch": 0.751901484969214, "grad_norm": 1.0202036156679617, "learning_rate": 3.05845287454865e-06, "loss": 0.6481, "step": 20760 }, { "epoch": 0.7519377037305324, "grad_norm": 1.66245934559476, "learning_rate": 3.0576085178605055e-06, "loss": 0.6862, "step": 20761 }, { "epoch": 0.7519739224918508, "grad_norm": 1.218982236780376, "learning_rate": 3.056764256705768e-06, "loss": 0.7138, "step": 20762 }, { "epoch": 0.7520101412531691, "grad_norm": 1.3425989363597057, "learning_rate": 3.055920091096052e-06, "loss": 0.624, "step": 20763 }, { "epoch": 0.7520463600144875, "grad_norm": 1.6777549282764561, "learning_rate": 3.0550760210429796e-06, "loss": 0.722, "step": 20764 }, { "epoch": 0.7520825787758059, "grad_norm": 1.3676201653467859, "learning_rate": 3.054232046558164e-06, "loss": 0.6304, "step": 20765 }, { "epoch": 0.7521187975371242, "grad_norm": 1.4139803986398884, "learning_rate": 3.0533881676532193e-06, "loss": 0.7425, "step": 20766 }, { "epoch": 0.7521550162984426, "grad_norm": 1.5026491100315784, "learning_rate": 3.052544384339754e-06, "loss": 0.7168, "step": 20767 }, { "epoch": 0.7521912350597609, "grad_norm": 1.389272564595559, "learning_rate": 3.0517006966293848e-06, "loss": 0.7017, "step": 20768 }, { "epoch": 0.7522274538210794, "grad_norm": 1.2732013398575954, "learning_rate": 3.05085710453372e-06, "loss": 0.6513, "step": 20769 }, { "epoch": 0.7522636725823977, "grad_norm": 1.270737220466219, "learning_rate": 3.0500136080643637e-06, "loss": 0.6916, "step": 20770 }, { "epoch": 0.752299891343716, "grad_norm": 1.2571182215555186, "learning_rate": 3.049170207232931e-06, "loss": 0.683, "step": 20771 }, { "epoch": 0.7523361101050344, "grad_norm": 1.343291584266667, "learning_rate": 3.0483269020510232e-06, "loss": 0.729, "step": 20772 }, { "epoch": 0.7523723288663527, "grad_norm": 1.4493715239146348, "learning_rate": 3.0474836925302453e-06, "loss": 0.7008, "step": 20773 }, { "epoch": 0.7524085476276712, "grad_norm": 1.3711029593112238, "learning_rate": 3.0466405786821984e-06, "loss": 0.6833, "step": 20774 }, { "epoch": 0.7524447663889895, "grad_norm": 1.1236541924391628, "learning_rate": 3.04579756051849e-06, "loss": 0.639, "step": 20775 }, { "epoch": 0.7524809851503078, "grad_norm": 1.2485863149617538, "learning_rate": 3.0449546380507166e-06, "loss": 0.7492, "step": 20776 }, { "epoch": 0.7525172039116262, "grad_norm": 1.206508400600162, "learning_rate": 3.044111811290479e-06, "loss": 0.6794, "step": 20777 }, { "epoch": 0.7525534226729446, "grad_norm": 1.383651341575685, "learning_rate": 3.0432690802493727e-06, "loss": 0.6208, "step": 20778 }, { "epoch": 0.752589641434263, "grad_norm": 1.3386446477805458, "learning_rate": 3.0424264449389984e-06, "loss": 0.7115, "step": 20779 }, { "epoch": 0.7526258601955813, "grad_norm": 1.1412938458284707, "learning_rate": 3.0415839053709506e-06, "loss": 0.6998, "step": 20780 }, { "epoch": 0.7526620789568996, "grad_norm": 1.3479326641943017, "learning_rate": 3.0407414615568186e-06, "loss": 0.7038, "step": 20781 }, { "epoch": 0.7526982977182181, "grad_norm": 1.306909779338304, "learning_rate": 3.0398991135082024e-06, "loss": 0.6172, "step": 20782 }, { "epoch": 0.7527345164795364, "grad_norm": 1.2016803711853672, "learning_rate": 3.039056861236691e-06, "loss": 0.6928, "step": 20783 }, { "epoch": 0.7527707352408548, "grad_norm": 1.3966595146419807, "learning_rate": 3.0382147047538724e-06, "loss": 0.6532, "step": 20784 }, { "epoch": 0.7528069540021731, "grad_norm": 1.3757049483337485, "learning_rate": 3.0373726440713335e-06, "loss": 0.7126, "step": 20785 }, { "epoch": 0.7528431727634914, "grad_norm": 1.4263589361496154, "learning_rate": 3.036530679200669e-06, "loss": 0.6221, "step": 20786 }, { "epoch": 0.7528793915248099, "grad_norm": 1.4380231733250144, "learning_rate": 3.035688810153461e-06, "loss": 0.6834, "step": 20787 }, { "epoch": 0.7529156102861282, "grad_norm": 1.3775503784151926, "learning_rate": 3.034847036941293e-06, "loss": 0.6245, "step": 20788 }, { "epoch": 0.7529518290474466, "grad_norm": 1.2926561216775985, "learning_rate": 3.034005359575748e-06, "loss": 0.6936, "step": 20789 }, { "epoch": 0.7529880478087649, "grad_norm": 1.3431930815654545, "learning_rate": 3.0331637780684133e-06, "loss": 0.691, "step": 20790 }, { "epoch": 0.7530242665700833, "grad_norm": 1.2713846491765242, "learning_rate": 3.032322292430866e-06, "loss": 0.6671, "step": 20791 }, { "epoch": 0.7530604853314017, "grad_norm": 1.0881875054305734, "learning_rate": 3.0314809026746826e-06, "loss": 0.6659, "step": 20792 }, { "epoch": 0.75309670409272, "grad_norm": 0.9677299029362705, "learning_rate": 3.0306396088114486e-06, "loss": 0.6984, "step": 20793 }, { "epoch": 0.7531329228540384, "grad_norm": 1.5417641983938, "learning_rate": 3.029798410852738e-06, "loss": 0.7048, "step": 20794 }, { "epoch": 0.7531691416153568, "grad_norm": 1.4546007672655548, "learning_rate": 3.0289573088101255e-06, "loss": 0.7058, "step": 20795 }, { "epoch": 0.7532053603766751, "grad_norm": 1.2252603217067302, "learning_rate": 3.028116302695183e-06, "loss": 0.6081, "step": 20796 }, { "epoch": 0.7532415791379935, "grad_norm": 1.044780003779769, "learning_rate": 3.0272753925194886e-06, "loss": 0.6686, "step": 20797 }, { "epoch": 0.7532777978993118, "grad_norm": 1.227399953214574, "learning_rate": 3.0264345782946127e-06, "loss": 0.6681, "step": 20798 }, { "epoch": 0.7533140166606302, "grad_norm": 1.422105956446229, "learning_rate": 3.0255938600321234e-06, "loss": 0.6462, "step": 20799 }, { "epoch": 0.7533502354219486, "grad_norm": 1.381920252542779, "learning_rate": 3.0247532377435885e-06, "loss": 0.6823, "step": 20800 }, { "epoch": 0.7533864541832669, "grad_norm": 1.4345999598442147, "learning_rate": 3.0239127114405795e-06, "loss": 0.6957, "step": 20801 }, { "epoch": 0.7534226729445853, "grad_norm": 1.3894013321194199, "learning_rate": 3.0230722811346635e-06, "loss": 0.7228, "step": 20802 }, { "epoch": 0.7534588917059036, "grad_norm": 1.0509382863397423, "learning_rate": 3.0222319468373985e-06, "loss": 0.6468, "step": 20803 }, { "epoch": 0.7534951104672221, "grad_norm": 1.5428798382076256, "learning_rate": 3.021391708560357e-06, "loss": 0.6671, "step": 20804 }, { "epoch": 0.7535313292285404, "grad_norm": 1.4069606018757388, "learning_rate": 3.0205515663150973e-06, "loss": 0.7145, "step": 20805 }, { "epoch": 0.7535675479898587, "grad_norm": 1.630166117717721, "learning_rate": 3.019711520113182e-06, "loss": 0.7485, "step": 20806 }, { "epoch": 0.7536037667511771, "grad_norm": 1.2104625336186703, "learning_rate": 3.0188715699661653e-06, "loss": 0.6785, "step": 20807 }, { "epoch": 0.7536399855124954, "grad_norm": 1.3764496938109199, "learning_rate": 3.0180317158856133e-06, "loss": 0.7078, "step": 20808 }, { "epoch": 0.7536762042738139, "grad_norm": 1.1632674703733004, "learning_rate": 3.0171919578830797e-06, "loss": 0.6677, "step": 20809 }, { "epoch": 0.7537124230351322, "grad_norm": 1.3751659150971027, "learning_rate": 3.0163522959701207e-06, "loss": 0.7047, "step": 20810 }, { "epoch": 0.7537486417964505, "grad_norm": 1.464777897308234, "learning_rate": 3.0155127301582863e-06, "loss": 0.6631, "step": 20811 }, { "epoch": 0.7537848605577689, "grad_norm": 1.270464401617691, "learning_rate": 3.0146732604591377e-06, "loss": 0.7214, "step": 20812 }, { "epoch": 0.7538210793190873, "grad_norm": 1.2650181719568838, "learning_rate": 3.0138338868842234e-06, "loss": 0.5864, "step": 20813 }, { "epoch": 0.7538572980804057, "grad_norm": 1.574751132081408, "learning_rate": 3.01299460944509e-06, "loss": 0.6729, "step": 20814 }, { "epoch": 0.753893516841724, "grad_norm": 1.447569268274829, "learning_rate": 3.012155428153293e-06, "loss": 0.6835, "step": 20815 }, { "epoch": 0.7539297356030423, "grad_norm": 1.0274071825115414, "learning_rate": 3.0113163430203775e-06, "loss": 0.6885, "step": 20816 }, { "epoch": 0.7539659543643608, "grad_norm": 1.415670723181958, "learning_rate": 3.0104773540578892e-06, "loss": 0.7256, "step": 20817 }, { "epoch": 0.7540021731256791, "grad_norm": 1.0212727324818378, "learning_rate": 3.009638461277372e-06, "loss": 0.6788, "step": 20818 }, { "epoch": 0.7540383918869975, "grad_norm": 1.3189220937082975, "learning_rate": 3.0087996646903737e-06, "loss": 0.6591, "step": 20819 }, { "epoch": 0.7540746106483158, "grad_norm": 1.3213306894110015, "learning_rate": 3.007960964308436e-06, "loss": 0.7065, "step": 20820 }, { "epoch": 0.7541108294096341, "grad_norm": 0.994204138043646, "learning_rate": 3.0071223601430988e-06, "loss": 0.6796, "step": 20821 }, { "epoch": 0.7541470481709526, "grad_norm": 1.2980549533109325, "learning_rate": 3.0062838522059e-06, "loss": 0.6597, "step": 20822 }, { "epoch": 0.7541832669322709, "grad_norm": 1.2241869377319083, "learning_rate": 3.0054454405083834e-06, "loss": 0.6284, "step": 20823 }, { "epoch": 0.7542194856935893, "grad_norm": 1.093120438680121, "learning_rate": 3.0046071250620836e-06, "loss": 0.6272, "step": 20824 }, { "epoch": 0.7542557044549076, "grad_norm": 1.0350935429178796, "learning_rate": 3.0037689058785325e-06, "loss": 0.6123, "step": 20825 }, { "epoch": 0.754291923216226, "grad_norm": 1.0828481133788177, "learning_rate": 3.0029307829692722e-06, "loss": 0.7321, "step": 20826 }, { "epoch": 0.7543281419775444, "grad_norm": 1.2310435806770774, "learning_rate": 3.0020927563458325e-06, "loss": 0.6211, "step": 20827 }, { "epoch": 0.7543643607388627, "grad_norm": 1.0395242777586393, "learning_rate": 3.0012548260197462e-06, "loss": 0.7023, "step": 20828 }, { "epoch": 0.7544005795001811, "grad_norm": 1.2933808567460248, "learning_rate": 3.0004169920025385e-06, "loss": 0.6521, "step": 20829 }, { "epoch": 0.7544367982614995, "grad_norm": 1.1458249164887628, "learning_rate": 2.999579254305748e-06, "loss": 0.702, "step": 20830 }, { "epoch": 0.7544730170228178, "grad_norm": 0.9059365313327162, "learning_rate": 2.9987416129408964e-06, "loss": 0.6586, "step": 20831 }, { "epoch": 0.7545092357841362, "grad_norm": 1.242155309325999, "learning_rate": 2.9979040679195103e-06, "loss": 0.641, "step": 20832 }, { "epoch": 0.7545454545454545, "grad_norm": 1.5047451849998408, "learning_rate": 2.9970666192531195e-06, "loss": 0.7259, "step": 20833 }, { "epoch": 0.754581673306773, "grad_norm": 1.2832640454885316, "learning_rate": 2.9962292669532457e-06, "loss": 0.6834, "step": 20834 }, { "epoch": 0.7546178920680913, "grad_norm": 2.3341605211769694, "learning_rate": 2.9953920110314106e-06, "loss": 0.6698, "step": 20835 }, { "epoch": 0.7546541108294096, "grad_norm": 1.3991571358246442, "learning_rate": 2.994554851499133e-06, "loss": 0.6837, "step": 20836 }, { "epoch": 0.754690329590728, "grad_norm": 1.2779711639201001, "learning_rate": 2.99371778836794e-06, "loss": 0.6841, "step": 20837 }, { "epoch": 0.7547265483520463, "grad_norm": 1.1569333674827567, "learning_rate": 2.9928808216493456e-06, "loss": 0.5462, "step": 20838 }, { "epoch": 0.7547627671133648, "grad_norm": 1.3908774663760683, "learning_rate": 2.9920439513548693e-06, "loss": 0.6434, "step": 20839 }, { "epoch": 0.7547989858746831, "grad_norm": 1.3464635131919669, "learning_rate": 2.9912071774960216e-06, "loss": 0.6189, "step": 20840 }, { "epoch": 0.7548352046360014, "grad_norm": 1.4478001797825204, "learning_rate": 2.9903705000843254e-06, "loss": 0.749, "step": 20841 }, { "epoch": 0.7548714233973198, "grad_norm": 1.39391678073648, "learning_rate": 2.98953391913129e-06, "loss": 0.6828, "step": 20842 }, { "epoch": 0.7549076421586381, "grad_norm": 1.3496973558336764, "learning_rate": 2.9886974346484254e-06, "loss": 0.7856, "step": 20843 }, { "epoch": 0.7549438609199566, "grad_norm": 1.2310306870052212, "learning_rate": 2.987861046647247e-06, "loss": 0.6429, "step": 20844 }, { "epoch": 0.7549800796812749, "grad_norm": 1.1272419804860134, "learning_rate": 2.987024755139263e-06, "loss": 0.6515, "step": 20845 }, { "epoch": 0.7550162984425932, "grad_norm": 1.211065673966547, "learning_rate": 2.9861885601359796e-06, "loss": 0.6411, "step": 20846 }, { "epoch": 0.7550525172039116, "grad_norm": 1.0584797858331574, "learning_rate": 2.9853524616489014e-06, "loss": 0.6881, "step": 20847 }, { "epoch": 0.75508873596523, "grad_norm": 1.4698669556640835, "learning_rate": 2.9845164596895404e-06, "loss": 0.7281, "step": 20848 }, { "epoch": 0.7551249547265484, "grad_norm": 1.3425888288347045, "learning_rate": 2.983680554269397e-06, "loss": 0.7068, "step": 20849 }, { "epoch": 0.7551611734878667, "grad_norm": 1.3399408891903584, "learning_rate": 2.982844745399974e-06, "loss": 0.6773, "step": 20850 }, { "epoch": 0.755197392249185, "grad_norm": 1.0276612901394102, "learning_rate": 2.982009033092771e-06, "loss": 0.749, "step": 20851 }, { "epoch": 0.7552336110105035, "grad_norm": 1.3447982073656084, "learning_rate": 2.981173417359293e-06, "loss": 0.6218, "step": 20852 }, { "epoch": 0.7552698297718218, "grad_norm": 1.2955091052578729, "learning_rate": 2.9803378982110354e-06, "loss": 0.6027, "step": 20853 }, { "epoch": 0.7553060485331402, "grad_norm": 1.3862045718330762, "learning_rate": 2.9795024756594937e-06, "loss": 0.6914, "step": 20854 }, { "epoch": 0.7553422672944585, "grad_norm": 0.9685447949246405, "learning_rate": 2.9786671497161703e-06, "loss": 0.6971, "step": 20855 }, { "epoch": 0.7553784860557768, "grad_norm": 1.3534093212311367, "learning_rate": 2.9778319203925556e-06, "loss": 0.6858, "step": 20856 }, { "epoch": 0.7554147048170953, "grad_norm": 1.3621729216207925, "learning_rate": 2.976996787700144e-06, "loss": 0.5981, "step": 20857 }, { "epoch": 0.7554509235784136, "grad_norm": 1.252222369933988, "learning_rate": 2.976161751650425e-06, "loss": 0.6933, "step": 20858 }, { "epoch": 0.755487142339732, "grad_norm": 1.4076183337612787, "learning_rate": 2.975326812254895e-06, "loss": 0.6914, "step": 20859 }, { "epoch": 0.7555233611010503, "grad_norm": 1.3018005454224522, "learning_rate": 2.9744919695250398e-06, "loss": 0.7193, "step": 20860 }, { "epoch": 0.7555595798623687, "grad_norm": 0.94626544966236, "learning_rate": 2.9736572234723483e-06, "loss": 0.7193, "step": 20861 }, { "epoch": 0.7555957986236871, "grad_norm": 0.9267252451180923, "learning_rate": 2.972822574108305e-06, "loss": 0.6428, "step": 20862 }, { "epoch": 0.7556320173850054, "grad_norm": 1.2876286935085333, "learning_rate": 2.9719880214444e-06, "loss": 0.7374, "step": 20863 }, { "epoch": 0.7556682361463238, "grad_norm": 1.3685535187006486, "learning_rate": 2.971153565492115e-06, "loss": 0.7014, "step": 20864 }, { "epoch": 0.7557044549076422, "grad_norm": 1.3833970394232478, "learning_rate": 2.9703192062629315e-06, "loss": 0.7227, "step": 20865 }, { "epoch": 0.7557406736689605, "grad_norm": 1.546101660994944, "learning_rate": 2.9694849437683337e-06, "loss": 0.6761, "step": 20866 }, { "epoch": 0.7557768924302789, "grad_norm": 1.3966041833662362, "learning_rate": 2.9686507780198015e-06, "loss": 0.6319, "step": 20867 }, { "epoch": 0.7558131111915972, "grad_norm": 1.5530038599765699, "learning_rate": 2.9678167090288137e-06, "loss": 0.6958, "step": 20868 }, { "epoch": 0.7558493299529156, "grad_norm": 0.9087460017599027, "learning_rate": 2.9669827368068427e-06, "loss": 0.669, "step": 20869 }, { "epoch": 0.755885548714234, "grad_norm": 1.4157940539938925, "learning_rate": 2.9661488613653723e-06, "loss": 0.6507, "step": 20870 }, { "epoch": 0.7559217674755523, "grad_norm": 1.3495431282025983, "learning_rate": 2.965315082715874e-06, "loss": 0.7233, "step": 20871 }, { "epoch": 0.7559579862368707, "grad_norm": 1.3634717515795807, "learning_rate": 2.964481400869822e-06, "loss": 0.7111, "step": 20872 }, { "epoch": 0.755994204998189, "grad_norm": 1.202372011115294, "learning_rate": 2.963647815838684e-06, "loss": 0.6166, "step": 20873 }, { "epoch": 0.7560304237595075, "grad_norm": 1.3512446603087258, "learning_rate": 2.9628143276339383e-06, "loss": 0.6936, "step": 20874 }, { "epoch": 0.7560666425208258, "grad_norm": 1.302803478095352, "learning_rate": 2.9619809362670513e-06, "loss": 0.6901, "step": 20875 }, { "epoch": 0.7561028612821441, "grad_norm": 1.340558731557556, "learning_rate": 2.9611476417494868e-06, "loss": 0.6547, "step": 20876 }, { "epoch": 0.7561390800434625, "grad_norm": 1.3067795917642682, "learning_rate": 2.960314444092719e-06, "loss": 0.7323, "step": 20877 }, { "epoch": 0.7561752988047808, "grad_norm": 1.788534620650144, "learning_rate": 2.9594813433082103e-06, "loss": 0.7696, "step": 20878 }, { "epoch": 0.7562115175660993, "grad_norm": 1.2490287337738917, "learning_rate": 2.9586483394074237e-06, "loss": 0.7164, "step": 20879 }, { "epoch": 0.7562477363274176, "grad_norm": 1.3219979640738788, "learning_rate": 2.9578154324018214e-06, "loss": 0.6954, "step": 20880 }, { "epoch": 0.7562839550887359, "grad_norm": 1.2922103323556782, "learning_rate": 2.9569826223028686e-06, "loss": 0.6813, "step": 20881 }, { "epoch": 0.7563201738500543, "grad_norm": 0.9777533173309765, "learning_rate": 2.9561499091220236e-06, "loss": 0.695, "step": 20882 }, { "epoch": 0.7563563926113727, "grad_norm": 1.36622402847486, "learning_rate": 2.9553172928707464e-06, "loss": 0.6766, "step": 20883 }, { "epoch": 0.7563926113726911, "grad_norm": 1.3445779611651414, "learning_rate": 2.954484773560489e-06, "loss": 0.6757, "step": 20884 }, { "epoch": 0.7564288301340094, "grad_norm": 1.491272046349086, "learning_rate": 2.9536523512027147e-06, "loss": 0.7258, "step": 20885 }, { "epoch": 0.7564650488953277, "grad_norm": 1.2588732996796461, "learning_rate": 2.952820025808876e-06, "loss": 0.6557, "step": 20886 }, { "epoch": 0.7565012676566462, "grad_norm": 1.3220137450607194, "learning_rate": 2.951987797390422e-06, "loss": 0.6746, "step": 20887 }, { "epoch": 0.7565374864179645, "grad_norm": 1.416637231600459, "learning_rate": 2.9511556659588127e-06, "loss": 0.7134, "step": 20888 }, { "epoch": 0.7565737051792829, "grad_norm": 1.404185549338333, "learning_rate": 2.9503236315254956e-06, "loss": 0.6549, "step": 20889 }, { "epoch": 0.7566099239406012, "grad_norm": 1.4558176811026822, "learning_rate": 2.9494916941019193e-06, "loss": 0.6908, "step": 20890 }, { "epoch": 0.7566461427019195, "grad_norm": 1.3488997934746543, "learning_rate": 2.9486598536995283e-06, "loss": 0.6227, "step": 20891 }, { "epoch": 0.756682361463238, "grad_norm": 1.3043002157691197, "learning_rate": 2.947828110329778e-06, "loss": 0.6821, "step": 20892 }, { "epoch": 0.7567185802245563, "grad_norm": 0.9681909791589562, "learning_rate": 2.9469964640041083e-06, "loss": 0.6997, "step": 20893 }, { "epoch": 0.7567547989858747, "grad_norm": 1.4982973608013455, "learning_rate": 2.9461649147339624e-06, "loss": 0.7396, "step": 20894 }, { "epoch": 0.756791017747193, "grad_norm": 1.52876304088311, "learning_rate": 2.945333462530788e-06, "loss": 0.7209, "step": 20895 }, { "epoch": 0.7568272365085114, "grad_norm": 1.3167795117607553, "learning_rate": 2.9445021074060245e-06, "loss": 0.6261, "step": 20896 }, { "epoch": 0.7568634552698298, "grad_norm": 0.9997038120012712, "learning_rate": 2.943670849371111e-06, "loss": 0.6817, "step": 20897 }, { "epoch": 0.7568996740311481, "grad_norm": 1.453582270646737, "learning_rate": 2.942839688437483e-06, "loss": 0.665, "step": 20898 }, { "epoch": 0.7569358927924665, "grad_norm": 1.4813814250017523, "learning_rate": 2.942008624616587e-06, "loss": 0.6741, "step": 20899 }, { "epoch": 0.7569721115537849, "grad_norm": 1.0360773118423263, "learning_rate": 2.9411776579198524e-06, "loss": 0.6278, "step": 20900 }, { "epoch": 0.7570083303151032, "grad_norm": 1.3723139529591226, "learning_rate": 2.9403467883587154e-06, "loss": 0.7256, "step": 20901 }, { "epoch": 0.7570445490764216, "grad_norm": 1.2995073793848646, "learning_rate": 2.939516015944608e-06, "loss": 0.6418, "step": 20902 }, { "epoch": 0.7570807678377399, "grad_norm": 1.5297125401264178, "learning_rate": 2.938685340688966e-06, "loss": 0.7397, "step": 20903 }, { "epoch": 0.7571169865990584, "grad_norm": 1.0432093056916913, "learning_rate": 2.9378547626032195e-06, "loss": 0.7425, "step": 20904 }, { "epoch": 0.7571532053603767, "grad_norm": 1.2908757807813396, "learning_rate": 2.9370242816987927e-06, "loss": 0.6948, "step": 20905 }, { "epoch": 0.757189424121695, "grad_norm": 1.1497700094705914, "learning_rate": 2.9361938979871218e-06, "loss": 0.6682, "step": 20906 }, { "epoch": 0.7572256428830134, "grad_norm": 1.3334604263424477, "learning_rate": 2.935363611479629e-06, "loss": 0.7262, "step": 20907 }, { "epoch": 0.7572618616443317, "grad_norm": 1.3994349060848912, "learning_rate": 2.934533422187741e-06, "loss": 0.6555, "step": 20908 }, { "epoch": 0.7572980804056502, "grad_norm": 1.2624145647096148, "learning_rate": 2.933703330122877e-06, "loss": 0.703, "step": 20909 }, { "epoch": 0.7573342991669685, "grad_norm": 1.563182460040722, "learning_rate": 2.932873335296469e-06, "loss": 0.7056, "step": 20910 }, { "epoch": 0.7573705179282868, "grad_norm": 1.304327201919113, "learning_rate": 2.9320434377199327e-06, "loss": 0.6825, "step": 20911 }, { "epoch": 0.7574067366896052, "grad_norm": 1.3453698146885926, "learning_rate": 2.9312136374046884e-06, "loss": 0.6718, "step": 20912 }, { "epoch": 0.7574429554509235, "grad_norm": 1.453502454111101, "learning_rate": 2.930383934362153e-06, "loss": 0.694, "step": 20913 }, { "epoch": 0.757479174212242, "grad_norm": 1.3876600146921774, "learning_rate": 2.929554328603751e-06, "loss": 0.721, "step": 20914 }, { "epoch": 0.7575153929735603, "grad_norm": 0.9940371012722976, "learning_rate": 2.9287248201408923e-06, "loss": 0.7029, "step": 20915 }, { "epoch": 0.7575516117348786, "grad_norm": 1.4165059891441834, "learning_rate": 2.927895408984992e-06, "loss": 0.6808, "step": 20916 }, { "epoch": 0.757587830496197, "grad_norm": 1.4566413911360683, "learning_rate": 2.927066095147467e-06, "loss": 0.6482, "step": 20917 }, { "epoch": 0.7576240492575154, "grad_norm": 1.2067986128357384, "learning_rate": 2.9262368786397278e-06, "loss": 0.6453, "step": 20918 }, { "epoch": 0.7576602680188338, "grad_norm": 1.2609991237290297, "learning_rate": 2.9254077594731846e-06, "loss": 0.7034, "step": 20919 }, { "epoch": 0.7576964867801521, "grad_norm": 1.455045718273816, "learning_rate": 2.9245787376592425e-06, "loss": 0.7005, "step": 20920 }, { "epoch": 0.7577327055414704, "grad_norm": 1.0071930112806873, "learning_rate": 2.923749813209319e-06, "loss": 0.7377, "step": 20921 }, { "epoch": 0.7577689243027889, "grad_norm": 1.2133680499529502, "learning_rate": 2.922920986134815e-06, "loss": 0.6861, "step": 20922 }, { "epoch": 0.7578051430641072, "grad_norm": 1.2336622501196166, "learning_rate": 2.9220922564471366e-06, "loss": 0.6811, "step": 20923 }, { "epoch": 0.7578413618254256, "grad_norm": 0.9127755238603699, "learning_rate": 2.9212636241576854e-06, "loss": 0.6598, "step": 20924 }, { "epoch": 0.7578775805867439, "grad_norm": 1.2751846519570316, "learning_rate": 2.920435089277869e-06, "loss": 0.6906, "step": 20925 }, { "epoch": 0.7579137993480622, "grad_norm": 1.2955287529656099, "learning_rate": 2.919606651819087e-06, "loss": 0.652, "step": 20926 }, { "epoch": 0.7579500181093807, "grad_norm": 0.8653202407493978, "learning_rate": 2.918778311792736e-06, "loss": 0.6236, "step": 20927 }, { "epoch": 0.757986236870699, "grad_norm": 1.3615621908744353, "learning_rate": 2.917950069210219e-06, "loss": 0.675, "step": 20928 }, { "epoch": 0.7580224556320174, "grad_norm": 1.963633162460443, "learning_rate": 2.917121924082934e-06, "loss": 0.6486, "step": 20929 }, { "epoch": 0.7580586743933357, "grad_norm": 1.3349352654846893, "learning_rate": 2.9162938764222736e-06, "loss": 0.6433, "step": 20930 }, { "epoch": 0.7580948931546541, "grad_norm": 1.2654453675514001, "learning_rate": 2.9154659262396313e-06, "loss": 0.6665, "step": 20931 }, { "epoch": 0.7581311119159725, "grad_norm": 1.0453862680724153, "learning_rate": 2.9146380735464052e-06, "loss": 0.7125, "step": 20932 }, { "epoch": 0.7581673306772908, "grad_norm": 1.3104587189861394, "learning_rate": 2.9138103183539856e-06, "loss": 0.642, "step": 20933 }, { "epoch": 0.7582035494386092, "grad_norm": 1.3671211172064266, "learning_rate": 2.9129826606737623e-06, "loss": 0.6748, "step": 20934 }, { "epoch": 0.7582397681999276, "grad_norm": 1.4263365421923326, "learning_rate": 2.912155100517122e-06, "loss": 0.7569, "step": 20935 }, { "epoch": 0.7582759869612459, "grad_norm": 1.360017549266995, "learning_rate": 2.911327637895458e-06, "loss": 0.7169, "step": 20936 }, { "epoch": 0.7583122057225643, "grad_norm": 1.3221399749039624, "learning_rate": 2.910500272820155e-06, "loss": 0.6395, "step": 20937 }, { "epoch": 0.7583484244838826, "grad_norm": 1.238494832815979, "learning_rate": 2.9096730053025933e-06, "loss": 0.63, "step": 20938 }, { "epoch": 0.758384643245201, "grad_norm": 1.3096186581306342, "learning_rate": 2.9088458353541647e-06, "loss": 0.6769, "step": 20939 }, { "epoch": 0.7584208620065194, "grad_norm": 1.4901549129374088, "learning_rate": 2.9080187629862488e-06, "loss": 0.7099, "step": 20940 }, { "epoch": 0.7584570807678377, "grad_norm": 1.087822195913103, "learning_rate": 2.907191788210226e-06, "loss": 0.7006, "step": 20941 }, { "epoch": 0.7584932995291561, "grad_norm": 0.872607021863738, "learning_rate": 2.9063649110374737e-06, "loss": 0.6801, "step": 20942 }, { "epoch": 0.7585295182904744, "grad_norm": 1.0052995299648997, "learning_rate": 2.905538131479376e-06, "loss": 0.6923, "step": 20943 }, { "epoch": 0.7585657370517929, "grad_norm": 1.7801183057784744, "learning_rate": 2.9047114495473085e-06, "loss": 0.7282, "step": 20944 }, { "epoch": 0.7586019558131112, "grad_norm": 1.281200168094135, "learning_rate": 2.903884865252644e-06, "loss": 0.6231, "step": 20945 }, { "epoch": 0.7586381745744295, "grad_norm": 0.9753644783643612, "learning_rate": 2.9030583786067566e-06, "loss": 0.6513, "step": 20946 }, { "epoch": 0.7586743933357479, "grad_norm": 1.4016915394579852, "learning_rate": 2.902231989621026e-06, "loss": 0.7377, "step": 20947 }, { "epoch": 0.7587106120970663, "grad_norm": 1.3335758584860997, "learning_rate": 2.901405698306818e-06, "loss": 0.661, "step": 20948 }, { "epoch": 0.7587468308583847, "grad_norm": 1.2998350033853843, "learning_rate": 2.9005795046755037e-06, "loss": 0.7212, "step": 20949 }, { "epoch": 0.758783049619703, "grad_norm": 1.3074875882049233, "learning_rate": 2.899753408738455e-06, "loss": 0.6621, "step": 20950 }, { "epoch": 0.7588192683810213, "grad_norm": 0.9125670284522103, "learning_rate": 2.89892741050704e-06, "loss": 0.6593, "step": 20951 }, { "epoch": 0.7588554871423397, "grad_norm": 1.0868603448026457, "learning_rate": 2.898101509992621e-06, "loss": 0.6638, "step": 20952 }, { "epoch": 0.7588917059036581, "grad_norm": 1.314582480375791, "learning_rate": 2.8972757072065637e-06, "loss": 0.6549, "step": 20953 }, { "epoch": 0.7589279246649765, "grad_norm": 1.1103019553344964, "learning_rate": 2.8964500021602373e-06, "loss": 0.6895, "step": 20954 }, { "epoch": 0.7589641434262948, "grad_norm": 1.294268182023273, "learning_rate": 2.895624394865e-06, "loss": 0.6734, "step": 20955 }, { "epoch": 0.7590003621876131, "grad_norm": 1.3152557391138637, "learning_rate": 2.894798885332211e-06, "loss": 0.6361, "step": 20956 }, { "epoch": 0.7590365809489316, "grad_norm": 1.3542064029720084, "learning_rate": 2.893973473573235e-06, "loss": 0.7029, "step": 20957 }, { "epoch": 0.7590727997102499, "grad_norm": 1.3373854476958904, "learning_rate": 2.893148159599427e-06, "loss": 0.655, "step": 20958 }, { "epoch": 0.7591090184715683, "grad_norm": 1.2558980134178586, "learning_rate": 2.8923229434221465e-06, "loss": 0.6181, "step": 20959 }, { "epoch": 0.7591452372328866, "grad_norm": 1.3775079880788292, "learning_rate": 2.8914978250527437e-06, "loss": 0.761, "step": 20960 }, { "epoch": 0.759181455994205, "grad_norm": 1.1954592185418897, "learning_rate": 2.89067280450258e-06, "loss": 0.6583, "step": 20961 }, { "epoch": 0.7592176747555234, "grad_norm": 1.4345832535878678, "learning_rate": 2.8898478817830055e-06, "loss": 0.6816, "step": 20962 }, { "epoch": 0.7592538935168417, "grad_norm": 1.4169778869138636, "learning_rate": 2.8890230569053723e-06, "loss": 0.7212, "step": 20963 }, { "epoch": 0.7592901122781601, "grad_norm": 0.9728427122302175, "learning_rate": 2.888198329881027e-06, "loss": 0.6587, "step": 20964 }, { "epoch": 0.7593263310394784, "grad_norm": 1.2611213597012028, "learning_rate": 2.887373700721323e-06, "loss": 0.6456, "step": 20965 }, { "epoch": 0.7593625498007968, "grad_norm": 1.4098357486275843, "learning_rate": 2.886549169437608e-06, "loss": 0.6756, "step": 20966 }, { "epoch": 0.7593987685621152, "grad_norm": 1.3052911525473434, "learning_rate": 2.8857247360412244e-06, "loss": 0.6881, "step": 20967 }, { "epoch": 0.7594349873234335, "grad_norm": 1.0333891947913691, "learning_rate": 2.8849004005435244e-06, "loss": 0.6841, "step": 20968 }, { "epoch": 0.7594712060847519, "grad_norm": 1.4928928974790494, "learning_rate": 2.8840761629558447e-06, "loss": 0.7853, "step": 20969 }, { "epoch": 0.7595074248460703, "grad_norm": 1.2712161269007425, "learning_rate": 2.8832520232895257e-06, "loss": 0.6888, "step": 20970 }, { "epoch": 0.7595436436073886, "grad_norm": 1.0156881948328564, "learning_rate": 2.8824279815559152e-06, "loss": 0.6957, "step": 20971 }, { "epoch": 0.759579862368707, "grad_norm": 1.2388816567733243, "learning_rate": 2.8816040377663503e-06, "loss": 0.6243, "step": 20972 }, { "epoch": 0.7596160811300253, "grad_norm": 1.3914242085995887, "learning_rate": 2.880780191932169e-06, "loss": 0.6283, "step": 20973 }, { "epoch": 0.7596522998913438, "grad_norm": 1.2282508184142291, "learning_rate": 2.879956444064703e-06, "loss": 0.6393, "step": 20974 }, { "epoch": 0.7596885186526621, "grad_norm": 1.2956943660310074, "learning_rate": 2.8791327941752967e-06, "loss": 0.6449, "step": 20975 }, { "epoch": 0.7597247374139804, "grad_norm": 1.4002207755456566, "learning_rate": 2.87830924227528e-06, "loss": 0.6879, "step": 20976 }, { "epoch": 0.7597609561752988, "grad_norm": 1.6171696346328144, "learning_rate": 2.8774857883759866e-06, "loss": 0.7497, "step": 20977 }, { "epoch": 0.7597971749366171, "grad_norm": 1.2030169311513772, "learning_rate": 2.8766624324887427e-06, "loss": 0.6452, "step": 20978 }, { "epoch": 0.7598333936979356, "grad_norm": 1.2811267389343748, "learning_rate": 2.875839174624887e-06, "loss": 0.6759, "step": 20979 }, { "epoch": 0.7598696124592539, "grad_norm": 1.3113639568413944, "learning_rate": 2.875016014795744e-06, "loss": 0.6827, "step": 20980 }, { "epoch": 0.7599058312205722, "grad_norm": 1.2698122241797272, "learning_rate": 2.874192953012638e-06, "loss": 0.6348, "step": 20981 }, { "epoch": 0.7599420499818906, "grad_norm": 1.4469247099862421, "learning_rate": 2.8733699892869004e-06, "loss": 0.6891, "step": 20982 }, { "epoch": 0.759978268743209, "grad_norm": 1.3093872874043033, "learning_rate": 2.872547123629854e-06, "loss": 0.6618, "step": 20983 }, { "epoch": 0.7600144875045274, "grad_norm": 1.4247321933347132, "learning_rate": 2.871724356052823e-06, "loss": 0.6936, "step": 20984 }, { "epoch": 0.7600507062658457, "grad_norm": 1.3669623284881172, "learning_rate": 2.8709016865671244e-06, "loss": 0.6983, "step": 20985 }, { "epoch": 0.760086925027164, "grad_norm": 1.0439289204806619, "learning_rate": 2.8700791151840857e-06, "loss": 0.6736, "step": 20986 }, { "epoch": 0.7601231437884824, "grad_norm": 0.9763001988527924, "learning_rate": 2.869256641915024e-06, "loss": 0.7487, "step": 20987 }, { "epoch": 0.7601593625498008, "grad_norm": 1.2946671556144531, "learning_rate": 2.868434266771256e-06, "loss": 0.619, "step": 20988 }, { "epoch": 0.7601955813111192, "grad_norm": 1.1905900799816171, "learning_rate": 2.867611989764096e-06, "loss": 0.6854, "step": 20989 }, { "epoch": 0.7602318000724375, "grad_norm": 1.4451154522989151, "learning_rate": 2.8667898109048653e-06, "loss": 0.7063, "step": 20990 }, { "epoch": 0.7602680188337558, "grad_norm": 1.4165202366368528, "learning_rate": 2.8659677302048738e-06, "loss": 0.6985, "step": 20991 }, { "epoch": 0.7603042375950743, "grad_norm": 1.0135468623113897, "learning_rate": 2.865145747675433e-06, "loss": 0.6857, "step": 20992 }, { "epoch": 0.7603404563563926, "grad_norm": 1.2446436658532594, "learning_rate": 2.864323863327858e-06, "loss": 0.6193, "step": 20993 }, { "epoch": 0.760376675117711, "grad_norm": 1.104206159262345, "learning_rate": 2.8635020771734567e-06, "loss": 0.681, "step": 20994 }, { "epoch": 0.7604128938790293, "grad_norm": 1.4048476710777646, "learning_rate": 2.8626803892235375e-06, "loss": 0.6424, "step": 20995 }, { "epoch": 0.7604491126403476, "grad_norm": 1.3949639766955155, "learning_rate": 2.8618587994894053e-06, "loss": 0.7395, "step": 20996 }, { "epoch": 0.7604853314016661, "grad_norm": 1.362744939074569, "learning_rate": 2.8610373079823707e-06, "loss": 0.6769, "step": 20997 }, { "epoch": 0.7605215501629844, "grad_norm": 1.3747108030717987, "learning_rate": 2.8602159147137355e-06, "loss": 0.6249, "step": 20998 }, { "epoch": 0.7605577689243028, "grad_norm": 3.7646893114377087, "learning_rate": 2.8593946196947997e-06, "loss": 0.7001, "step": 20999 }, { "epoch": 0.7605939876856211, "grad_norm": 0.98108154477884, "learning_rate": 2.858573422936871e-06, "loss": 0.6747, "step": 21000 }, { "epoch": 0.7606302064469395, "grad_norm": 1.3994174299712037, "learning_rate": 2.8577523244512473e-06, "loss": 0.7153, "step": 21001 }, { "epoch": 0.7606664252082579, "grad_norm": 1.441534628488062, "learning_rate": 2.856931324249228e-06, "loss": 0.6574, "step": 21002 }, { "epoch": 0.7607026439695762, "grad_norm": 1.171180770299789, "learning_rate": 2.8561104223421056e-06, "loss": 0.6471, "step": 21003 }, { "epoch": 0.7607388627308946, "grad_norm": 1.3810491836108014, "learning_rate": 2.855289618741185e-06, "loss": 0.7003, "step": 21004 }, { "epoch": 0.760775081492213, "grad_norm": 1.5158545944491584, "learning_rate": 2.8544689134577575e-06, "loss": 0.71, "step": 21005 }, { "epoch": 0.7608113002535313, "grad_norm": 0.9257421396616564, "learning_rate": 2.853648306503116e-06, "loss": 0.6742, "step": 21006 }, { "epoch": 0.7608475190148497, "grad_norm": 1.4647722131427456, "learning_rate": 2.85282779788855e-06, "loss": 0.7007, "step": 21007 }, { "epoch": 0.760883737776168, "grad_norm": 1.304050503488653, "learning_rate": 2.852007387625356e-06, "loss": 0.6924, "step": 21008 }, { "epoch": 0.7609199565374865, "grad_norm": 1.42961219813377, "learning_rate": 2.8511870757248215e-06, "loss": 0.6431, "step": 21009 }, { "epoch": 0.7609561752988048, "grad_norm": 0.9923352157135771, "learning_rate": 2.8503668621982317e-06, "loss": 0.6956, "step": 21010 }, { "epoch": 0.7609923940601231, "grad_norm": 1.2329945888502067, "learning_rate": 2.8495467470568794e-06, "loss": 0.6771, "step": 21011 }, { "epoch": 0.7610286128214415, "grad_norm": 1.2889305885330151, "learning_rate": 2.848726730312046e-06, "loss": 0.6468, "step": 21012 }, { "epoch": 0.7610648315827598, "grad_norm": 1.008591305832631, "learning_rate": 2.8479068119750176e-06, "loss": 0.6877, "step": 21013 }, { "epoch": 0.7611010503440783, "grad_norm": 1.2319849943247774, "learning_rate": 2.847086992057071e-06, "loss": 0.683, "step": 21014 }, { "epoch": 0.7611372691053966, "grad_norm": 1.3801123069851153, "learning_rate": 2.8462672705694973e-06, "loss": 0.6538, "step": 21015 }, { "epoch": 0.7611734878667149, "grad_norm": 1.3213186394650915, "learning_rate": 2.8454476475235706e-06, "loss": 0.6429, "step": 21016 }, { "epoch": 0.7612097066280333, "grad_norm": 1.1323779763912263, "learning_rate": 2.8446281229305716e-06, "loss": 0.6601, "step": 21017 }, { "epoch": 0.7612459253893517, "grad_norm": 1.2943682594747772, "learning_rate": 2.843808696801773e-06, "loss": 0.6884, "step": 21018 }, { "epoch": 0.7612821441506701, "grad_norm": 1.3115120409358874, "learning_rate": 2.842989369148459e-06, "loss": 0.6305, "step": 21019 }, { "epoch": 0.7613183629119884, "grad_norm": 1.4002426174323277, "learning_rate": 2.842170139981899e-06, "loss": 0.6838, "step": 21020 }, { "epoch": 0.7613545816733067, "grad_norm": 0.9861615645823404, "learning_rate": 2.8413510093133646e-06, "loss": 0.68, "step": 21021 }, { "epoch": 0.7613908004346251, "grad_norm": 1.2329006731027425, "learning_rate": 2.840531977154133e-06, "loss": 0.6422, "step": 21022 }, { "epoch": 0.7614270191959435, "grad_norm": 1.131343879928016, "learning_rate": 2.839713043515473e-06, "loss": 0.6633, "step": 21023 }, { "epoch": 0.7614632379572619, "grad_norm": 1.0471909699159108, "learning_rate": 2.8388942084086525e-06, "loss": 0.7545, "step": 21024 }, { "epoch": 0.7614994567185802, "grad_norm": 1.1724786514227807, "learning_rate": 2.8380754718449366e-06, "loss": 0.652, "step": 21025 }, { "epoch": 0.7615356754798985, "grad_norm": 1.0128448742251184, "learning_rate": 2.8372568338355988e-06, "loss": 0.6649, "step": 21026 }, { "epoch": 0.761571894241217, "grad_norm": 1.239313097322599, "learning_rate": 2.8364382943919012e-06, "loss": 0.6744, "step": 21027 }, { "epoch": 0.7616081130025353, "grad_norm": 1.4260462958748608, "learning_rate": 2.8356198535251066e-06, "loss": 0.6188, "step": 21028 }, { "epoch": 0.7616443317638537, "grad_norm": 1.4012718979075593, "learning_rate": 2.8348015112464743e-06, "loss": 0.7035, "step": 21029 }, { "epoch": 0.761680550525172, "grad_norm": 1.8983639125925948, "learning_rate": 2.833983267567273e-06, "loss": 0.6579, "step": 21030 }, { "epoch": 0.7617167692864903, "grad_norm": 1.0404411963961677, "learning_rate": 2.833165122498759e-06, "loss": 0.6891, "step": 21031 }, { "epoch": 0.7617529880478088, "grad_norm": 1.3023296754187492, "learning_rate": 2.8323470760521866e-06, "loss": 0.6542, "step": 21032 }, { "epoch": 0.7617892068091271, "grad_norm": 1.4175619587484947, "learning_rate": 2.831529128238819e-06, "loss": 0.6563, "step": 21033 }, { "epoch": 0.7618254255704455, "grad_norm": 1.1250313194248247, "learning_rate": 2.8307112790699097e-06, "loss": 0.6438, "step": 21034 }, { "epoch": 0.7618616443317638, "grad_norm": 1.4848287127408877, "learning_rate": 2.829893528556713e-06, "loss": 0.6952, "step": 21035 }, { "epoch": 0.7618978630930822, "grad_norm": 1.2932209153634835, "learning_rate": 2.829075876710479e-06, "loss": 0.6653, "step": 21036 }, { "epoch": 0.7619340818544006, "grad_norm": 1.4908167462310655, "learning_rate": 2.8282583235424643e-06, "loss": 0.7005, "step": 21037 }, { "epoch": 0.7619703006157189, "grad_norm": 1.2594400620988404, "learning_rate": 2.8274408690639164e-06, "loss": 0.7204, "step": 21038 }, { "epoch": 0.7620065193770373, "grad_norm": 1.0470822224776184, "learning_rate": 2.826623513286085e-06, "loss": 0.6366, "step": 21039 }, { "epoch": 0.7620427381383557, "grad_norm": 1.4822768145417187, "learning_rate": 2.8258062562202137e-06, "loss": 0.7599, "step": 21040 }, { "epoch": 0.762078956899674, "grad_norm": 1.2895256756465951, "learning_rate": 2.824989097877555e-06, "loss": 0.6461, "step": 21041 }, { "epoch": 0.7621151756609924, "grad_norm": 1.414375855306113, "learning_rate": 2.8241720382693514e-06, "loss": 0.7498, "step": 21042 }, { "epoch": 0.7621513944223107, "grad_norm": 1.3968635857839007, "learning_rate": 2.823355077406843e-06, "loss": 0.6895, "step": 21043 }, { "epoch": 0.7621876131836292, "grad_norm": 1.3421282078928396, "learning_rate": 2.8225382153012772e-06, "loss": 0.6278, "step": 21044 }, { "epoch": 0.7622238319449475, "grad_norm": 1.3414169707950176, "learning_rate": 2.8217214519638926e-06, "loss": 0.5987, "step": 21045 }, { "epoch": 0.7622600507062658, "grad_norm": 1.2957769960823275, "learning_rate": 2.820904787405927e-06, "loss": 0.6575, "step": 21046 }, { "epoch": 0.7622962694675842, "grad_norm": 1.3501814148843068, "learning_rate": 2.8200882216386182e-06, "loss": 0.6694, "step": 21047 }, { "epoch": 0.7623324882289025, "grad_norm": 1.319148150533549, "learning_rate": 2.819271754673206e-06, "loss": 0.6747, "step": 21048 }, { "epoch": 0.762368706990221, "grad_norm": 1.3787200602110592, "learning_rate": 2.818455386520925e-06, "loss": 0.64, "step": 21049 }, { "epoch": 0.7624049257515393, "grad_norm": 1.3819497590652372, "learning_rate": 2.8176391171930073e-06, "loss": 0.6925, "step": 21050 }, { "epoch": 0.7624411445128576, "grad_norm": 1.396306374683061, "learning_rate": 2.8168229467006846e-06, "loss": 0.7286, "step": 21051 }, { "epoch": 0.762477363274176, "grad_norm": 1.3987114527423832, "learning_rate": 2.8160068750551917e-06, "loss": 0.6693, "step": 21052 }, { "epoch": 0.7625135820354944, "grad_norm": 1.4525571439088774, "learning_rate": 2.815190902267757e-06, "loss": 0.7362, "step": 21053 }, { "epoch": 0.7625498007968128, "grad_norm": 1.1419814891389095, "learning_rate": 2.8143750283496054e-06, "loss": 0.6386, "step": 21054 }, { "epoch": 0.7625860195581311, "grad_norm": 1.5799962636727656, "learning_rate": 2.8135592533119706e-06, "loss": 0.7046, "step": 21055 }, { "epoch": 0.7626222383194495, "grad_norm": 1.5076523680967584, "learning_rate": 2.812743577166075e-06, "loss": 0.6297, "step": 21056 }, { "epoch": 0.7626584570807678, "grad_norm": 1.363980080003322, "learning_rate": 2.811927999923143e-06, "loss": 0.6948, "step": 21057 }, { "epoch": 0.7626946758420862, "grad_norm": 1.2042432950672421, "learning_rate": 2.8111125215943944e-06, "loss": 0.6162, "step": 21058 }, { "epoch": 0.7627308946034046, "grad_norm": 1.4062386293746283, "learning_rate": 2.8102971421910576e-06, "loss": 0.6829, "step": 21059 }, { "epoch": 0.7627671133647229, "grad_norm": 1.3316855464429433, "learning_rate": 2.8094818617243492e-06, "loss": 0.6323, "step": 21060 }, { "epoch": 0.7628033321260413, "grad_norm": 0.8670203524003465, "learning_rate": 2.8086666802054853e-06, "loss": 0.71, "step": 21061 }, { "epoch": 0.7628395508873597, "grad_norm": 1.048832086037503, "learning_rate": 2.8078515976456897e-06, "loss": 0.6269, "step": 21062 }, { "epoch": 0.762875769648678, "grad_norm": 1.1627633170219493, "learning_rate": 2.8070366140561765e-06, "loss": 0.7301, "step": 21063 }, { "epoch": 0.7629119884099964, "grad_norm": 1.4427278734827957, "learning_rate": 2.80622172944816e-06, "loss": 0.6914, "step": 21064 }, { "epoch": 0.7629482071713147, "grad_norm": 0.9421266562186686, "learning_rate": 2.805406943832849e-06, "loss": 0.6365, "step": 21065 }, { "epoch": 0.7629844259326332, "grad_norm": 1.3469860290027247, "learning_rate": 2.8045922572214644e-06, "loss": 0.6499, "step": 21066 }, { "epoch": 0.7630206446939515, "grad_norm": 1.3292952647943466, "learning_rate": 2.8037776696252127e-06, "loss": 0.6382, "step": 21067 }, { "epoch": 0.7630568634552698, "grad_norm": 1.3676874649872566, "learning_rate": 2.8029631810553026e-06, "loss": 0.6702, "step": 21068 }, { "epoch": 0.7630930822165882, "grad_norm": 1.1953240792174877, "learning_rate": 2.8021487915229407e-06, "loss": 0.6791, "step": 21069 }, { "epoch": 0.7631293009779065, "grad_norm": 1.348559795394606, "learning_rate": 2.801334501039339e-06, "loss": 0.7529, "step": 21070 }, { "epoch": 0.763165519739225, "grad_norm": 1.3252549997829803, "learning_rate": 2.8005203096156998e-06, "loss": 0.7216, "step": 21071 }, { "epoch": 0.7632017385005433, "grad_norm": 1.3025621753620562, "learning_rate": 2.799706217263224e-06, "loss": 0.6427, "step": 21072 }, { "epoch": 0.7632379572618616, "grad_norm": 1.520646085926119, "learning_rate": 2.7988922239931206e-06, "loss": 0.6774, "step": 21073 }, { "epoch": 0.76327417602318, "grad_norm": 1.3559944929278613, "learning_rate": 2.798078329816588e-06, "loss": 0.6529, "step": 21074 }, { "epoch": 0.7633103947844984, "grad_norm": 1.3635317031318075, "learning_rate": 2.7972645347448246e-06, "loss": 0.6474, "step": 21075 }, { "epoch": 0.7633466135458168, "grad_norm": 1.4485177082190603, "learning_rate": 2.7964508387890277e-06, "loss": 0.6868, "step": 21076 }, { "epoch": 0.7633828323071351, "grad_norm": 1.1132065932804514, "learning_rate": 2.7956372419604005e-06, "loss": 0.6732, "step": 21077 }, { "epoch": 0.7634190510684534, "grad_norm": 1.397340426790299, "learning_rate": 2.794823744270134e-06, "loss": 0.726, "step": 21078 }, { "epoch": 0.7634552698297719, "grad_norm": 1.3906006214866016, "learning_rate": 2.7940103457294254e-06, "loss": 0.6693, "step": 21079 }, { "epoch": 0.7634914885910902, "grad_norm": 1.3914454871208326, "learning_rate": 2.793197046349462e-06, "loss": 0.6749, "step": 21080 }, { "epoch": 0.7635277073524086, "grad_norm": 1.390322831896382, "learning_rate": 2.792383846141442e-06, "loss": 0.6931, "step": 21081 }, { "epoch": 0.7635639261137269, "grad_norm": 0.9564729168834294, "learning_rate": 2.7915707451165553e-06, "loss": 0.7253, "step": 21082 }, { "epoch": 0.7636001448750452, "grad_norm": 1.2622821269025328, "learning_rate": 2.7907577432859843e-06, "loss": 0.6703, "step": 21083 }, { "epoch": 0.7636363636363637, "grad_norm": 1.3118895696375727, "learning_rate": 2.789944840660925e-06, "loss": 0.624, "step": 21084 }, { "epoch": 0.763672582397682, "grad_norm": 1.350752520269482, "learning_rate": 2.789132037252561e-06, "loss": 0.6759, "step": 21085 }, { "epoch": 0.7637088011590004, "grad_norm": 1.3150183580435224, "learning_rate": 2.7883193330720747e-06, "loss": 0.6833, "step": 21086 }, { "epoch": 0.7637450199203187, "grad_norm": 1.4171060507184525, "learning_rate": 2.7875067281306487e-06, "loss": 0.7258, "step": 21087 }, { "epoch": 0.763781238681637, "grad_norm": 1.2470398636299265, "learning_rate": 2.786694222439471e-06, "loss": 0.6402, "step": 21088 }, { "epoch": 0.7638174574429555, "grad_norm": 1.331034237920749, "learning_rate": 2.785881816009718e-06, "loss": 0.6771, "step": 21089 }, { "epoch": 0.7638536762042738, "grad_norm": 1.3777831112488848, "learning_rate": 2.78506950885257e-06, "loss": 0.7078, "step": 21090 }, { "epoch": 0.7638898949655922, "grad_norm": 1.376348398539685, "learning_rate": 2.784257300979204e-06, "loss": 0.6422, "step": 21091 }, { "epoch": 0.7639261137269105, "grad_norm": 1.3610760641509068, "learning_rate": 2.783445192400799e-06, "loss": 0.6354, "step": 21092 }, { "epoch": 0.7639623324882289, "grad_norm": 1.4412705233490644, "learning_rate": 2.78263318312853e-06, "loss": 0.7168, "step": 21093 }, { "epoch": 0.7639985512495473, "grad_norm": 1.3660764924594087, "learning_rate": 2.7818212731735674e-06, "loss": 0.7164, "step": 21094 }, { "epoch": 0.7640347700108656, "grad_norm": 1.2199930796270613, "learning_rate": 2.7810094625470884e-06, "loss": 0.6934, "step": 21095 }, { "epoch": 0.764070988772184, "grad_norm": 1.3767366252626685, "learning_rate": 2.780197751260263e-06, "loss": 0.6682, "step": 21096 }, { "epoch": 0.7641072075335024, "grad_norm": 1.5131127290728419, "learning_rate": 2.7793861393242605e-06, "loss": 0.6772, "step": 21097 }, { "epoch": 0.7641434262948207, "grad_norm": 1.127237732075999, "learning_rate": 2.7785746267502468e-06, "loss": 0.6249, "step": 21098 }, { "epoch": 0.7641796450561391, "grad_norm": 1.0595174985019808, "learning_rate": 2.777763213549394e-06, "loss": 0.6441, "step": 21099 }, { "epoch": 0.7642158638174574, "grad_norm": 1.0901824836975815, "learning_rate": 2.7769518997328647e-06, "loss": 0.6851, "step": 21100 }, { "epoch": 0.7642520825787759, "grad_norm": 1.367693104578068, "learning_rate": 2.7761406853118242e-06, "loss": 0.695, "step": 21101 }, { "epoch": 0.7642883013400942, "grad_norm": 0.9746251208611552, "learning_rate": 2.7753295702974338e-06, "loss": 0.7224, "step": 21102 }, { "epoch": 0.7643245201014125, "grad_norm": 1.0972977259858208, "learning_rate": 2.774518554700858e-06, "loss": 0.661, "step": 21103 }, { "epoch": 0.7643607388627309, "grad_norm": 1.417559807764967, "learning_rate": 2.773707638533256e-06, "loss": 0.7311, "step": 21104 }, { "epoch": 0.7643969576240492, "grad_norm": 1.626814972025235, "learning_rate": 2.772896821805784e-06, "loss": 0.6654, "step": 21105 }, { "epoch": 0.7644331763853677, "grad_norm": 1.042882696848365, "learning_rate": 2.7720861045296053e-06, "loss": 0.7012, "step": 21106 }, { "epoch": 0.764469395146686, "grad_norm": 1.2519423451606742, "learning_rate": 2.771275486715872e-06, "loss": 0.6546, "step": 21107 }, { "epoch": 0.7645056139080043, "grad_norm": 1.2739004846870767, "learning_rate": 2.77046496837574e-06, "loss": 0.6146, "step": 21108 }, { "epoch": 0.7645418326693227, "grad_norm": 1.5008647004626194, "learning_rate": 2.7696545495203596e-06, "loss": 0.6679, "step": 21109 }, { "epoch": 0.7645780514306411, "grad_norm": 1.347136181535497, "learning_rate": 2.7688442301608886e-06, "loss": 0.7094, "step": 21110 }, { "epoch": 0.7646142701919595, "grad_norm": 1.5490346628950424, "learning_rate": 2.7680340103084758e-06, "loss": 0.6453, "step": 21111 }, { "epoch": 0.7646504889532778, "grad_norm": 1.3492835091547108, "learning_rate": 2.7672238899742676e-06, "loss": 0.6323, "step": 21112 }, { "epoch": 0.7646867077145961, "grad_norm": 1.0060170090055687, "learning_rate": 2.7664138691694122e-06, "loss": 0.7388, "step": 21113 }, { "epoch": 0.7647229264759146, "grad_norm": 0.9599946636470478, "learning_rate": 2.7656039479050613e-06, "loss": 0.6415, "step": 21114 }, { "epoch": 0.7647591452372329, "grad_norm": 1.180581041304986, "learning_rate": 2.764794126192356e-06, "loss": 0.7231, "step": 21115 }, { "epoch": 0.7647953639985513, "grad_norm": 1.4313341127483492, "learning_rate": 2.763984404042438e-06, "loss": 0.7258, "step": 21116 }, { "epoch": 0.7648315827598696, "grad_norm": 1.528636243420229, "learning_rate": 2.7631747814664568e-06, "loss": 0.6637, "step": 21117 }, { "epoch": 0.7648678015211879, "grad_norm": 0.9331600750188906, "learning_rate": 2.762365258475548e-06, "loss": 0.6197, "step": 21118 }, { "epoch": 0.7649040202825064, "grad_norm": 1.3860887479879982, "learning_rate": 2.7615558350808535e-06, "loss": 0.7198, "step": 21119 }, { "epoch": 0.7649402390438247, "grad_norm": 1.3532927344737478, "learning_rate": 2.7607465112935074e-06, "loss": 0.7178, "step": 21120 }, { "epoch": 0.7649764578051431, "grad_norm": 1.3676781945443917, "learning_rate": 2.7599372871246543e-06, "loss": 0.6634, "step": 21121 }, { "epoch": 0.7650126765664614, "grad_norm": 1.232450364368466, "learning_rate": 2.759128162585425e-06, "loss": 0.7062, "step": 21122 }, { "epoch": 0.7650488953277798, "grad_norm": 1.0559702450104895, "learning_rate": 2.758319137686951e-06, "loss": 0.6803, "step": 21123 }, { "epoch": 0.7650851140890982, "grad_norm": 0.9537519551526983, "learning_rate": 2.757510212440373e-06, "loss": 0.6964, "step": 21124 }, { "epoch": 0.7651213328504165, "grad_norm": 1.3297959664199472, "learning_rate": 2.7567013868568182e-06, "loss": 0.5719, "step": 21125 }, { "epoch": 0.7651575516117349, "grad_norm": 1.3672957227652947, "learning_rate": 2.755892660947416e-06, "loss": 0.6962, "step": 21126 }, { "epoch": 0.7651937703730533, "grad_norm": 1.2053719568855998, "learning_rate": 2.7550840347232933e-06, "loss": 0.6987, "step": 21127 }, { "epoch": 0.7652299891343716, "grad_norm": 1.3886577307709118, "learning_rate": 2.7542755081955843e-06, "loss": 0.628, "step": 21128 }, { "epoch": 0.76526620789569, "grad_norm": 0.9544240765839468, "learning_rate": 2.7534670813754096e-06, "loss": 0.6434, "step": 21129 }, { "epoch": 0.7653024266570083, "grad_norm": 1.445663423722313, "learning_rate": 2.7526587542738957e-06, "loss": 0.6947, "step": 21130 }, { "epoch": 0.7653386454183267, "grad_norm": 1.329644809697667, "learning_rate": 2.7518505269021623e-06, "loss": 0.7617, "step": 21131 }, { "epoch": 0.7653748641796451, "grad_norm": 1.4013730736851435, "learning_rate": 2.7510423992713374e-06, "loss": 0.7225, "step": 21132 }, { "epoch": 0.7654110829409634, "grad_norm": 1.378215739826738, "learning_rate": 2.750234371392538e-06, "loss": 0.6943, "step": 21133 }, { "epoch": 0.7654473017022818, "grad_norm": 1.2503532245873659, "learning_rate": 2.7494264432768826e-06, "loss": 0.7442, "step": 21134 }, { "epoch": 0.7654835204636001, "grad_norm": 1.277340043205459, "learning_rate": 2.748618614935492e-06, "loss": 0.7075, "step": 21135 }, { "epoch": 0.7655197392249186, "grad_norm": 1.376904329152271, "learning_rate": 2.7478108863794806e-06, "loss": 0.5791, "step": 21136 }, { "epoch": 0.7655559579862369, "grad_norm": 1.3468620530898003, "learning_rate": 2.7470032576199646e-06, "loss": 0.6514, "step": 21137 }, { "epoch": 0.7655921767475552, "grad_norm": 1.0653229693811015, "learning_rate": 2.746195728668053e-06, "loss": 0.6965, "step": 21138 }, { "epoch": 0.7656283955088736, "grad_norm": 1.3693506369551967, "learning_rate": 2.7453882995348657e-06, "loss": 0.7051, "step": 21139 }, { "epoch": 0.765664614270192, "grad_norm": 1.1151528727434326, "learning_rate": 2.7445809702315095e-06, "loss": 0.6904, "step": 21140 }, { "epoch": 0.7657008330315104, "grad_norm": 1.5372290037589085, "learning_rate": 2.7437737407690944e-06, "loss": 0.7677, "step": 21141 }, { "epoch": 0.7657370517928287, "grad_norm": 1.327300656664864, "learning_rate": 2.742966611158725e-06, "loss": 0.6259, "step": 21142 }, { "epoch": 0.765773270554147, "grad_norm": 1.3095700891936906, "learning_rate": 2.7421595814115153e-06, "loss": 0.6349, "step": 21143 }, { "epoch": 0.7658094893154654, "grad_norm": 1.3630182837857325, "learning_rate": 2.741352651538567e-06, "loss": 0.711, "step": 21144 }, { "epoch": 0.7658457080767838, "grad_norm": 1.0056227889857299, "learning_rate": 2.7405458215509805e-06, "loss": 0.6866, "step": 21145 }, { "epoch": 0.7658819268381022, "grad_norm": 1.2833819663895432, "learning_rate": 2.7397390914598663e-06, "loss": 0.6569, "step": 21146 }, { "epoch": 0.7659181455994205, "grad_norm": 1.4516298808646724, "learning_rate": 2.7389324612763213e-06, "loss": 0.7353, "step": 21147 }, { "epoch": 0.7659543643607388, "grad_norm": 1.2978776774136334, "learning_rate": 2.738125931011445e-06, "loss": 0.737, "step": 21148 }, { "epoch": 0.7659905831220573, "grad_norm": 1.297874967135848, "learning_rate": 2.7373195006763343e-06, "loss": 0.7215, "step": 21149 }, { "epoch": 0.7660268018833756, "grad_norm": 1.313158591684838, "learning_rate": 2.7365131702820913e-06, "loss": 0.7671, "step": 21150 }, { "epoch": 0.766063020644694, "grad_norm": 1.2459280403113928, "learning_rate": 2.7357069398398105e-06, "loss": 0.7109, "step": 21151 }, { "epoch": 0.7660992394060123, "grad_norm": 1.3890842376773849, "learning_rate": 2.7349008093605846e-06, "loss": 0.6741, "step": 21152 }, { "epoch": 0.7661354581673306, "grad_norm": 1.3483927131135032, "learning_rate": 2.734094778855504e-06, "loss": 0.635, "step": 21153 }, { "epoch": 0.7661716769286491, "grad_norm": 1.4478332491356785, "learning_rate": 2.733288848335667e-06, "loss": 0.7077, "step": 21154 }, { "epoch": 0.7662078956899674, "grad_norm": 1.4276875567815914, "learning_rate": 2.7324830178121595e-06, "loss": 0.7087, "step": 21155 }, { "epoch": 0.7662441144512858, "grad_norm": 1.455416811499064, "learning_rate": 2.7316772872960697e-06, "loss": 0.6951, "step": 21156 }, { "epoch": 0.7662803332126041, "grad_norm": 1.376265927895184, "learning_rate": 2.7308716567984896e-06, "loss": 0.6361, "step": 21157 }, { "epoch": 0.7663165519739225, "grad_norm": 1.315260627420567, "learning_rate": 2.730066126330503e-06, "loss": 0.6921, "step": 21158 }, { "epoch": 0.7663527707352409, "grad_norm": 1.2663029235562273, "learning_rate": 2.729260695903193e-06, "loss": 0.6699, "step": 21159 }, { "epoch": 0.7663889894965592, "grad_norm": 0.9477915074059514, "learning_rate": 2.7284553655276426e-06, "loss": 0.701, "step": 21160 }, { "epoch": 0.7664252082578776, "grad_norm": 1.383874200720294, "learning_rate": 2.7276501352149386e-06, "loss": 0.7739, "step": 21161 }, { "epoch": 0.766461427019196, "grad_norm": 1.7303310214617533, "learning_rate": 2.7268450049761585e-06, "loss": 0.7682, "step": 21162 }, { "epoch": 0.7664976457805143, "grad_norm": 1.375197183764582, "learning_rate": 2.7260399748223832e-06, "loss": 0.7108, "step": 21163 }, { "epoch": 0.7665338645418327, "grad_norm": 1.4312642085315457, "learning_rate": 2.7252350447646847e-06, "loss": 0.6564, "step": 21164 }, { "epoch": 0.766570083303151, "grad_norm": 1.3909141335049149, "learning_rate": 2.7244302148141476e-06, "loss": 0.7161, "step": 21165 }, { "epoch": 0.7666063020644694, "grad_norm": 1.2691669704418524, "learning_rate": 2.7236254849818443e-06, "loss": 0.5973, "step": 21166 }, { "epoch": 0.7666425208257878, "grad_norm": 1.3368722736356256, "learning_rate": 2.7228208552788436e-06, "loss": 0.6852, "step": 21167 }, { "epoch": 0.7666787395871061, "grad_norm": 1.4292105428952775, "learning_rate": 2.7220163257162267e-06, "loss": 0.6627, "step": 21168 }, { "epoch": 0.7667149583484245, "grad_norm": 2.1075899412437114, "learning_rate": 2.721211896305059e-06, "loss": 0.6416, "step": 21169 }, { "epoch": 0.7667511771097428, "grad_norm": 1.3789570052069724, "learning_rate": 2.7204075670564122e-06, "loss": 0.7594, "step": 21170 }, { "epoch": 0.7667873958710613, "grad_norm": 1.4891723082964248, "learning_rate": 2.7196033379813503e-06, "loss": 0.7003, "step": 21171 }, { "epoch": 0.7668236146323796, "grad_norm": 1.249709773830103, "learning_rate": 2.7187992090909465e-06, "loss": 0.6375, "step": 21172 }, { "epoch": 0.7668598333936979, "grad_norm": 1.5728606904618938, "learning_rate": 2.7179951803962636e-06, "loss": 0.7063, "step": 21173 }, { "epoch": 0.7668960521550163, "grad_norm": 1.0437780798355436, "learning_rate": 2.717191251908362e-06, "loss": 0.7247, "step": 21174 }, { "epoch": 0.7669322709163346, "grad_norm": 1.2484805871334024, "learning_rate": 2.7163874236383115e-06, "loss": 0.6651, "step": 21175 }, { "epoch": 0.7669684896776531, "grad_norm": 1.3246881460438868, "learning_rate": 2.7155836955971694e-06, "loss": 0.7198, "step": 21176 }, { "epoch": 0.7670047084389714, "grad_norm": 1.0336891786049476, "learning_rate": 2.7147800677959967e-06, "loss": 0.6244, "step": 21177 }, { "epoch": 0.7670409272002897, "grad_norm": 1.8055675639666116, "learning_rate": 2.7139765402458494e-06, "loss": 0.6365, "step": 21178 }, { "epoch": 0.7670771459616081, "grad_norm": 1.3977308785668419, "learning_rate": 2.7131731129577922e-06, "loss": 0.7189, "step": 21179 }, { "epoch": 0.7671133647229265, "grad_norm": 1.065430374619501, "learning_rate": 2.7123697859428722e-06, "loss": 0.7112, "step": 21180 }, { "epoch": 0.7671495834842449, "grad_norm": 1.3097405163036056, "learning_rate": 2.7115665592121453e-06, "loss": 0.6248, "step": 21181 }, { "epoch": 0.7671858022455632, "grad_norm": 1.246651982587215, "learning_rate": 2.7107634327766698e-06, "loss": 0.6812, "step": 21182 }, { "epoch": 0.7672220210068815, "grad_norm": 1.3721180347546635, "learning_rate": 2.7099604066474936e-06, "loss": 0.6492, "step": 21183 }, { "epoch": 0.7672582397682, "grad_norm": 1.2792514974976457, "learning_rate": 2.709157480835668e-06, "loss": 0.7235, "step": 21184 }, { "epoch": 0.7672944585295183, "grad_norm": 1.1944629926387451, "learning_rate": 2.7083546553522388e-06, "loss": 0.6104, "step": 21185 }, { "epoch": 0.7673306772908367, "grad_norm": 0.9949521616037285, "learning_rate": 2.7075519302082596e-06, "loss": 0.6705, "step": 21186 }, { "epoch": 0.767366896052155, "grad_norm": 1.3162441465142594, "learning_rate": 2.706749305414774e-06, "loss": 0.7046, "step": 21187 }, { "epoch": 0.7674031148134733, "grad_norm": 1.2613190262821625, "learning_rate": 2.7059467809828234e-06, "loss": 0.6478, "step": 21188 }, { "epoch": 0.7674393335747918, "grad_norm": 1.3310362624966339, "learning_rate": 2.705144356923457e-06, "loss": 0.6079, "step": 21189 }, { "epoch": 0.7674755523361101, "grad_norm": 1.350487307907395, "learning_rate": 2.704342033247713e-06, "loss": 0.6657, "step": 21190 }, { "epoch": 0.7675117710974285, "grad_norm": 1.3298817236104947, "learning_rate": 2.7035398099666343e-06, "loss": 0.6947, "step": 21191 }, { "epoch": 0.7675479898587468, "grad_norm": 1.2250577441966821, "learning_rate": 2.702737687091256e-06, "loss": 0.6461, "step": 21192 }, { "epoch": 0.7675842086200652, "grad_norm": 1.056919829647644, "learning_rate": 2.7019356646326223e-06, "loss": 0.7365, "step": 21193 }, { "epoch": 0.7676204273813836, "grad_norm": 1.469708079478491, "learning_rate": 2.7011337426017657e-06, "loss": 0.7018, "step": 21194 }, { "epoch": 0.7676566461427019, "grad_norm": 1.3094257649035634, "learning_rate": 2.700331921009722e-06, "loss": 0.668, "step": 21195 }, { "epoch": 0.7676928649040203, "grad_norm": 1.307059963433079, "learning_rate": 2.699530199867523e-06, "loss": 0.6595, "step": 21196 }, { "epoch": 0.7677290836653387, "grad_norm": 1.3457274185612178, "learning_rate": 2.6987285791862063e-06, "loss": 0.6846, "step": 21197 }, { "epoch": 0.767765302426657, "grad_norm": 1.3215338867118078, "learning_rate": 2.6979270589767994e-06, "loss": 0.7067, "step": 21198 }, { "epoch": 0.7678015211879754, "grad_norm": 1.3513587911151745, "learning_rate": 2.6971256392503286e-06, "loss": 0.6988, "step": 21199 }, { "epoch": 0.7678377399492937, "grad_norm": 1.3139373722557206, "learning_rate": 2.6963243200178292e-06, "loss": 0.6096, "step": 21200 }, { "epoch": 0.7678739587106121, "grad_norm": 0.9085064568275034, "learning_rate": 2.695523101290325e-06, "loss": 0.6959, "step": 21201 }, { "epoch": 0.7679101774719305, "grad_norm": 1.399001649060881, "learning_rate": 2.694721983078842e-06, "loss": 0.6549, "step": 21202 }, { "epoch": 0.7679463962332488, "grad_norm": 1.4479067511070363, "learning_rate": 2.6939209653944e-06, "loss": 0.6793, "step": 21203 }, { "epoch": 0.7679826149945672, "grad_norm": 1.430470706775854, "learning_rate": 2.693120048248027e-06, "loss": 0.7216, "step": 21204 }, { "epoch": 0.7680188337558855, "grad_norm": 1.480549211253145, "learning_rate": 2.692319231650744e-06, "loss": 0.6731, "step": 21205 }, { "epoch": 0.768055052517204, "grad_norm": 1.4890540411950073, "learning_rate": 2.6915185156135694e-06, "loss": 0.8099, "step": 21206 }, { "epoch": 0.7680912712785223, "grad_norm": 1.4345526749349198, "learning_rate": 2.690717900147518e-06, "loss": 0.7337, "step": 21207 }, { "epoch": 0.7681274900398406, "grad_norm": 1.442509936139605, "learning_rate": 2.689917385263614e-06, "loss": 0.6674, "step": 21208 }, { "epoch": 0.768163708801159, "grad_norm": 1.461013984463032, "learning_rate": 2.6891169709728704e-06, "loss": 0.7116, "step": 21209 }, { "epoch": 0.7681999275624773, "grad_norm": 1.4713934080129198, "learning_rate": 2.6883166572862975e-06, "loss": 0.7328, "step": 21210 }, { "epoch": 0.7682361463237958, "grad_norm": 1.4828897524407507, "learning_rate": 2.6875164442149147e-06, "loss": 0.7111, "step": 21211 }, { "epoch": 0.7682723650851141, "grad_norm": 1.3769782476635892, "learning_rate": 2.686716331769732e-06, "loss": 0.6312, "step": 21212 }, { "epoch": 0.7683085838464324, "grad_norm": 1.2564153819801327, "learning_rate": 2.685916319961759e-06, "loss": 0.6767, "step": 21213 }, { "epoch": 0.7683448026077508, "grad_norm": 1.3995069924143584, "learning_rate": 2.6851164088019997e-06, "loss": 0.6432, "step": 21214 }, { "epoch": 0.7683810213690692, "grad_norm": 1.2530882852361478, "learning_rate": 2.6843165983014698e-06, "loss": 0.7585, "step": 21215 }, { "epoch": 0.7684172401303876, "grad_norm": 1.3360461084934532, "learning_rate": 2.6835168884711715e-06, "loss": 0.6431, "step": 21216 }, { "epoch": 0.7684534588917059, "grad_norm": 1.3243086356767768, "learning_rate": 2.682717279322107e-06, "loss": 0.6646, "step": 21217 }, { "epoch": 0.7684896776530242, "grad_norm": 1.1831327057665808, "learning_rate": 2.6819177708652844e-06, "loss": 0.6944, "step": 21218 }, { "epoch": 0.7685258964143427, "grad_norm": 1.5372414344372773, "learning_rate": 2.6811183631117044e-06, "loss": 0.6985, "step": 21219 }, { "epoch": 0.768562115175661, "grad_norm": 1.1358707715790115, "learning_rate": 2.6803190560723668e-06, "loss": 0.6784, "step": 21220 }, { "epoch": 0.7685983339369794, "grad_norm": 1.430673704668173, "learning_rate": 2.6795198497582663e-06, "loss": 0.6507, "step": 21221 }, { "epoch": 0.7686345526982977, "grad_norm": 1.0013520251284662, "learning_rate": 2.6787207441804097e-06, "loss": 0.653, "step": 21222 }, { "epoch": 0.768670771459616, "grad_norm": 1.3199199051911445, "learning_rate": 2.677921739349787e-06, "loss": 0.6892, "step": 21223 }, { "epoch": 0.7687069902209345, "grad_norm": 1.1799840733441431, "learning_rate": 2.6771228352773958e-06, "loss": 0.7041, "step": 21224 }, { "epoch": 0.7687432089822528, "grad_norm": 0.8734549258131933, "learning_rate": 2.6763240319742257e-06, "loss": 0.6227, "step": 21225 }, { "epoch": 0.7687794277435712, "grad_norm": 1.281147469366562, "learning_rate": 2.675525329451275e-06, "loss": 0.681, "step": 21226 }, { "epoch": 0.7688156465048895, "grad_norm": 1.3226702478144667, "learning_rate": 2.674726727719532e-06, "loss": 0.6437, "step": 21227 }, { "epoch": 0.7688518652662079, "grad_norm": 1.4874207344293242, "learning_rate": 2.6739282267899824e-06, "loss": 0.7478, "step": 21228 }, { "epoch": 0.7688880840275263, "grad_norm": 1.1741442472872978, "learning_rate": 2.6731298266736214e-06, "loss": 0.5785, "step": 21229 }, { "epoch": 0.7689243027888446, "grad_norm": 1.045643649201798, "learning_rate": 2.672331527381431e-06, "loss": 0.6938, "step": 21230 }, { "epoch": 0.768960521550163, "grad_norm": 1.3795318373137597, "learning_rate": 2.6715333289243983e-06, "loss": 0.7274, "step": 21231 }, { "epoch": 0.7689967403114814, "grad_norm": 1.3754787742657242, "learning_rate": 2.6707352313135027e-06, "loss": 0.7125, "step": 21232 }, { "epoch": 0.7690329590727997, "grad_norm": 1.2856447963808497, "learning_rate": 2.669937234559733e-06, "loss": 0.6733, "step": 21233 }, { "epoch": 0.7690691778341181, "grad_norm": 1.411995000678923, "learning_rate": 2.669139338674068e-06, "loss": 0.6432, "step": 21234 }, { "epoch": 0.7691053965954364, "grad_norm": 1.3923609493680598, "learning_rate": 2.6683415436674873e-06, "loss": 0.6539, "step": 21235 }, { "epoch": 0.7691416153567548, "grad_norm": 1.3345409400964654, "learning_rate": 2.6675438495509666e-06, "loss": 0.7017, "step": 21236 }, { "epoch": 0.7691778341180732, "grad_norm": 0.9990144967632762, "learning_rate": 2.6667462563354885e-06, "loss": 0.6712, "step": 21237 }, { "epoch": 0.7692140528793915, "grad_norm": 1.3175861803990974, "learning_rate": 2.6659487640320247e-06, "loss": 0.6057, "step": 21238 }, { "epoch": 0.7692502716407099, "grad_norm": 5.593925452344448, "learning_rate": 2.6651513726515468e-06, "loss": 0.7163, "step": 21239 }, { "epoch": 0.7692864904020282, "grad_norm": 1.3965780853627592, "learning_rate": 2.6643540822050353e-06, "loss": 0.6385, "step": 21240 }, { "epoch": 0.7693227091633467, "grad_norm": 1.3219082545207566, "learning_rate": 2.663556892703456e-06, "loss": 0.7379, "step": 21241 }, { "epoch": 0.769358927924665, "grad_norm": 1.0261316598022026, "learning_rate": 2.6627598041577807e-06, "loss": 0.7656, "step": 21242 }, { "epoch": 0.7693951466859833, "grad_norm": 1.413315449222326, "learning_rate": 2.6619628165789744e-06, "loss": 0.6331, "step": 21243 }, { "epoch": 0.7694313654473017, "grad_norm": 1.3621253411703147, "learning_rate": 2.6611659299780113e-06, "loss": 0.6706, "step": 21244 }, { "epoch": 0.76946758420862, "grad_norm": 1.398685109737953, "learning_rate": 2.660369144365853e-06, "loss": 0.7399, "step": 21245 }, { "epoch": 0.7695038029699385, "grad_norm": 1.511856065472281, "learning_rate": 2.6595724597534642e-06, "loss": 0.6578, "step": 21246 }, { "epoch": 0.7695400217312568, "grad_norm": 1.2568209305632771, "learning_rate": 2.6587758761518057e-06, "loss": 0.7012, "step": 21247 }, { "epoch": 0.7695762404925751, "grad_norm": 1.3495032721462927, "learning_rate": 2.6579793935718445e-06, "loss": 0.6812, "step": 21248 }, { "epoch": 0.7696124592538935, "grad_norm": 1.0403332306561242, "learning_rate": 2.657183012024538e-06, "loss": 0.6944, "step": 21249 }, { "epoch": 0.7696486780152119, "grad_norm": 8.407164185831293, "learning_rate": 2.6563867315208413e-06, "loss": 0.7318, "step": 21250 }, { "epoch": 0.7696848967765303, "grad_norm": 1.3736425561883414, "learning_rate": 2.6555905520717205e-06, "loss": 0.6957, "step": 21251 }, { "epoch": 0.7697211155378486, "grad_norm": 1.5013412201393628, "learning_rate": 2.654794473688126e-06, "loss": 0.6704, "step": 21252 }, { "epoch": 0.7697573342991669, "grad_norm": 1.3353139219492975, "learning_rate": 2.653998496381014e-06, "loss": 0.6578, "step": 21253 }, { "epoch": 0.7697935530604854, "grad_norm": 1.1956760322731628, "learning_rate": 2.653202620161335e-06, "loss": 0.6334, "step": 21254 }, { "epoch": 0.7698297718218037, "grad_norm": 1.3072312806216864, "learning_rate": 2.652406845040045e-06, "loss": 0.6414, "step": 21255 }, { "epoch": 0.7698659905831221, "grad_norm": 1.0036698560640782, "learning_rate": 2.6516111710280932e-06, "loss": 0.6644, "step": 21256 }, { "epoch": 0.7699022093444404, "grad_norm": 1.329141571373295, "learning_rate": 2.6508155981364305e-06, "loss": 0.6531, "step": 21257 }, { "epoch": 0.7699384281057587, "grad_norm": 1.346782045575384, "learning_rate": 2.650020126375997e-06, "loss": 0.6221, "step": 21258 }, { "epoch": 0.7699746468670772, "grad_norm": 1.593141287607775, "learning_rate": 2.64922475575775e-06, "loss": 0.6805, "step": 21259 }, { "epoch": 0.7700108656283955, "grad_norm": 1.3477606670728883, "learning_rate": 2.6484294862926286e-06, "loss": 0.6868, "step": 21260 }, { "epoch": 0.7700470843897139, "grad_norm": 1.4963957118288358, "learning_rate": 2.647634317991573e-06, "loss": 0.7434, "step": 21261 }, { "epoch": 0.7700833031510322, "grad_norm": 1.253859575537821, "learning_rate": 2.646839250865534e-06, "loss": 0.6703, "step": 21262 }, { "epoch": 0.7701195219123506, "grad_norm": 1.5721821484411154, "learning_rate": 2.6460442849254473e-06, "loss": 0.7127, "step": 21263 }, { "epoch": 0.770155740673669, "grad_norm": 1.268850969568467, "learning_rate": 2.6452494201822533e-06, "loss": 0.6863, "step": 21264 }, { "epoch": 0.7701919594349873, "grad_norm": 1.403239912200027, "learning_rate": 2.6444546566468864e-06, "loss": 0.7414, "step": 21265 }, { "epoch": 0.7702281781963057, "grad_norm": 1.402825713682837, "learning_rate": 2.64365999433029e-06, "loss": 0.728, "step": 21266 }, { "epoch": 0.770264396957624, "grad_norm": 1.0019568233277663, "learning_rate": 2.6428654332433958e-06, "loss": 0.6821, "step": 21267 }, { "epoch": 0.7703006157189424, "grad_norm": 1.3336692035606506, "learning_rate": 2.6420709733971375e-06, "loss": 0.6995, "step": 21268 }, { "epoch": 0.7703368344802608, "grad_norm": 1.0016131457485855, "learning_rate": 2.641276614802445e-06, "loss": 0.7593, "step": 21269 }, { "epoch": 0.7703730532415791, "grad_norm": 1.4337647073659876, "learning_rate": 2.6404823574702556e-06, "loss": 0.6874, "step": 21270 }, { "epoch": 0.7704092720028975, "grad_norm": 1.3367196576345584, "learning_rate": 2.6396882014114956e-06, "loss": 0.6868, "step": 21271 }, { "epoch": 0.7704454907642159, "grad_norm": 1.2552950304476238, "learning_rate": 2.63889414663709e-06, "loss": 0.7033, "step": 21272 }, { "epoch": 0.7704817095255342, "grad_norm": 0.9164417984873935, "learning_rate": 2.638100193157972e-06, "loss": 0.6868, "step": 21273 }, { "epoch": 0.7705179282868526, "grad_norm": 1.178359311425636, "learning_rate": 2.637306340985063e-06, "loss": 0.6496, "step": 21274 }, { "epoch": 0.7705541470481709, "grad_norm": 1.4390879016315947, "learning_rate": 2.6365125901292896e-06, "loss": 0.6895, "step": 21275 }, { "epoch": 0.7705903658094894, "grad_norm": 1.5280785085962227, "learning_rate": 2.6357189406015693e-06, "loss": 0.6409, "step": 21276 }, { "epoch": 0.7706265845708077, "grad_norm": 1.3269471702913718, "learning_rate": 2.6349253924128313e-06, "loss": 0.6223, "step": 21277 }, { "epoch": 0.770662803332126, "grad_norm": 1.4135524991858928, "learning_rate": 2.63413194557399e-06, "loss": 0.701, "step": 21278 }, { "epoch": 0.7706990220934444, "grad_norm": 1.3388226798719751, "learning_rate": 2.633338600095964e-06, "loss": 0.6568, "step": 21279 }, { "epoch": 0.7707352408547627, "grad_norm": 1.4373654193698213, "learning_rate": 2.632545355989673e-06, "loss": 0.7273, "step": 21280 }, { "epoch": 0.7707714596160812, "grad_norm": 1.3690668227674536, "learning_rate": 2.6317522132660324e-06, "loss": 0.7296, "step": 21281 }, { "epoch": 0.7708076783773995, "grad_norm": 1.3397764642241283, "learning_rate": 2.6309591719359563e-06, "loss": 0.7866, "step": 21282 }, { "epoch": 0.7708438971387178, "grad_norm": 1.6318584163220158, "learning_rate": 2.6301662320103526e-06, "loss": 0.7161, "step": 21283 }, { "epoch": 0.7708801159000362, "grad_norm": 1.3468628835520327, "learning_rate": 2.629373393500141e-06, "loss": 0.6561, "step": 21284 }, { "epoch": 0.7709163346613546, "grad_norm": 1.2940840146561554, "learning_rate": 2.6285806564162276e-06, "loss": 0.6576, "step": 21285 }, { "epoch": 0.770952553422673, "grad_norm": 1.3621626352608032, "learning_rate": 2.6277880207695216e-06, "loss": 0.7235, "step": 21286 }, { "epoch": 0.7709887721839913, "grad_norm": 1.5323420104861682, "learning_rate": 2.626995486570928e-06, "loss": 0.6328, "step": 21287 }, { "epoch": 0.7710249909453096, "grad_norm": 1.340683754387074, "learning_rate": 2.6262030538313564e-06, "loss": 0.6752, "step": 21288 }, { "epoch": 0.7710612097066281, "grad_norm": 1.3923539774457763, "learning_rate": 2.625410722561712e-06, "loss": 0.6783, "step": 21289 }, { "epoch": 0.7710974284679464, "grad_norm": 1.3350233839173462, "learning_rate": 2.6246184927728913e-06, "loss": 0.6814, "step": 21290 }, { "epoch": 0.7711336472292648, "grad_norm": 1.266860562640055, "learning_rate": 2.6238263644758043e-06, "loss": 0.6723, "step": 21291 }, { "epoch": 0.7711698659905831, "grad_norm": 0.9428474861258523, "learning_rate": 2.623034337681347e-06, "loss": 0.6853, "step": 21292 }, { "epoch": 0.7712060847519014, "grad_norm": 1.2620956165126973, "learning_rate": 2.622242412400421e-06, "loss": 0.6256, "step": 21293 }, { "epoch": 0.7712423035132199, "grad_norm": 1.4833567665580647, "learning_rate": 2.621450588643918e-06, "loss": 0.694, "step": 21294 }, { "epoch": 0.7712785222745382, "grad_norm": 1.31661949779301, "learning_rate": 2.6206588664227404e-06, "loss": 0.6912, "step": 21295 }, { "epoch": 0.7713147410358566, "grad_norm": 1.134782933708032, "learning_rate": 2.6198672457477825e-06, "loss": 0.6999, "step": 21296 }, { "epoch": 0.7713509597971749, "grad_norm": 1.3631765317183973, "learning_rate": 2.619075726629935e-06, "loss": 0.7104, "step": 21297 }, { "epoch": 0.7713871785584933, "grad_norm": 1.225273342466398, "learning_rate": 2.618284309080088e-06, "loss": 0.6875, "step": 21298 }, { "epoch": 0.7714233973198117, "grad_norm": 1.3298957513422638, "learning_rate": 2.6174929931091387e-06, "loss": 0.629, "step": 21299 }, { "epoch": 0.77145961608113, "grad_norm": 1.9422883691166088, "learning_rate": 2.6167017787279725e-06, "loss": 0.6134, "step": 21300 }, { "epoch": 0.7714958348424484, "grad_norm": 1.3026745773900799, "learning_rate": 2.615910665947473e-06, "loss": 0.5979, "step": 21301 }, { "epoch": 0.7715320536037668, "grad_norm": 1.3286784961120732, "learning_rate": 2.615119654778535e-06, "loss": 0.7483, "step": 21302 }, { "epoch": 0.7715682723650851, "grad_norm": 1.2852749378392798, "learning_rate": 2.6143287452320397e-06, "loss": 0.6672, "step": 21303 }, { "epoch": 0.7716044911264035, "grad_norm": 1.3599306086476022, "learning_rate": 2.6135379373188706e-06, "loss": 0.6863, "step": 21304 }, { "epoch": 0.7716407098877218, "grad_norm": 1.4863690766646869, "learning_rate": 2.612747231049907e-06, "loss": 0.6608, "step": 21305 }, { "epoch": 0.7716769286490403, "grad_norm": 1.257878388432457, "learning_rate": 2.611956626436034e-06, "loss": 0.6115, "step": 21306 }, { "epoch": 0.7717131474103586, "grad_norm": 1.4374700030016594, "learning_rate": 2.6111661234881312e-06, "loss": 0.6733, "step": 21307 }, { "epoch": 0.7717493661716769, "grad_norm": 1.2974901761819373, "learning_rate": 2.6103757222170745e-06, "loss": 0.688, "step": 21308 }, { "epoch": 0.7717855849329953, "grad_norm": 1.2415445653720396, "learning_rate": 2.6095854226337367e-06, "loss": 0.6957, "step": 21309 }, { "epoch": 0.7718218036943136, "grad_norm": 1.547791013526156, "learning_rate": 2.6087952247490024e-06, "loss": 0.6374, "step": 21310 }, { "epoch": 0.7718580224556321, "grad_norm": 1.3447052504458121, "learning_rate": 2.6080051285737384e-06, "loss": 0.6472, "step": 21311 }, { "epoch": 0.7718942412169504, "grad_norm": 1.3619402968936527, "learning_rate": 2.607215134118817e-06, "loss": 0.6051, "step": 21312 }, { "epoch": 0.7719304599782687, "grad_norm": 1.1537979250494164, "learning_rate": 2.606425241395114e-06, "loss": 0.6697, "step": 21313 }, { "epoch": 0.7719666787395871, "grad_norm": 1.254512558578208, "learning_rate": 2.6056354504134962e-06, "loss": 0.6641, "step": 21314 }, { "epoch": 0.7720028975009054, "grad_norm": 1.4575532107399682, "learning_rate": 2.6048457611848322e-06, "loss": 0.658, "step": 21315 }, { "epoch": 0.7720391162622239, "grad_norm": 0.8668625711915199, "learning_rate": 2.6040561737199855e-06, "loss": 0.6703, "step": 21316 }, { "epoch": 0.7720753350235422, "grad_norm": 1.1117973787244309, "learning_rate": 2.603266688029826e-06, "loss": 0.7252, "step": 21317 }, { "epoch": 0.7721115537848605, "grad_norm": 1.3330818358328689, "learning_rate": 2.602477304125218e-06, "loss": 0.6655, "step": 21318 }, { "epoch": 0.772147772546179, "grad_norm": 1.4763416574215953, "learning_rate": 2.6016880220170215e-06, "loss": 0.7229, "step": 21319 }, { "epoch": 0.7721839913074973, "grad_norm": 1.3046229654900985, "learning_rate": 2.6008988417160964e-06, "loss": 0.5655, "step": 21320 }, { "epoch": 0.7722202100688157, "grad_norm": 1.3169543910751689, "learning_rate": 2.600109763233307e-06, "loss": 0.6711, "step": 21321 }, { "epoch": 0.772256428830134, "grad_norm": 1.2271119727509576, "learning_rate": 2.599320786579509e-06, "loss": 0.7306, "step": 21322 }, { "epoch": 0.7722926475914523, "grad_norm": 1.4303214531996042, "learning_rate": 2.5985319117655573e-06, "loss": 0.6762, "step": 21323 }, { "epoch": 0.7723288663527708, "grad_norm": 1.4852153607785155, "learning_rate": 2.597743138802313e-06, "loss": 0.7176, "step": 21324 }, { "epoch": 0.7723650851140891, "grad_norm": 1.2456534416444316, "learning_rate": 2.5969544677006275e-06, "loss": 0.6974, "step": 21325 }, { "epoch": 0.7724013038754075, "grad_norm": 1.501918039544579, "learning_rate": 2.5961658984713522e-06, "loss": 0.6949, "step": 21326 }, { "epoch": 0.7724375226367258, "grad_norm": 1.035246685133065, "learning_rate": 2.5953774311253377e-06, "loss": 0.6491, "step": 21327 }, { "epoch": 0.7724737413980441, "grad_norm": 1.2631668072010154, "learning_rate": 2.5945890656734396e-06, "loss": 0.5985, "step": 21328 }, { "epoch": 0.7725099601593626, "grad_norm": 1.5212397727665625, "learning_rate": 2.5938008021265016e-06, "loss": 0.7275, "step": 21329 }, { "epoch": 0.7725461789206809, "grad_norm": 1.1870078177409562, "learning_rate": 2.5930126404953738e-06, "loss": 0.6655, "step": 21330 }, { "epoch": 0.7725823976819993, "grad_norm": 0.9291151086863432, "learning_rate": 2.5922245807908964e-06, "loss": 0.5911, "step": 21331 }, { "epoch": 0.7726186164433176, "grad_norm": 1.0428449611075867, "learning_rate": 2.591436623023921e-06, "loss": 0.6768, "step": 21332 }, { "epoch": 0.772654835204636, "grad_norm": 1.3057509509658216, "learning_rate": 2.590648767205287e-06, "loss": 0.6837, "step": 21333 }, { "epoch": 0.7726910539659544, "grad_norm": 1.0451125538618475, "learning_rate": 2.5898610133458337e-06, "loss": 0.6632, "step": 21334 }, { "epoch": 0.7727272727272727, "grad_norm": 1.3129516204902358, "learning_rate": 2.589073361456407e-06, "loss": 0.5913, "step": 21335 }, { "epoch": 0.7727634914885911, "grad_norm": 1.3056311435586878, "learning_rate": 2.5882858115478424e-06, "loss": 0.6823, "step": 21336 }, { "epoch": 0.7727997102499095, "grad_norm": 1.0140478590717834, "learning_rate": 2.587498363630978e-06, "loss": 0.7498, "step": 21337 }, { "epoch": 0.7728359290112278, "grad_norm": 1.3255852692673902, "learning_rate": 2.5867110177166456e-06, "loss": 0.6604, "step": 21338 }, { "epoch": 0.7728721477725462, "grad_norm": 1.4399234671007237, "learning_rate": 2.5859237738156875e-06, "loss": 0.6808, "step": 21339 }, { "epoch": 0.7729083665338645, "grad_norm": 1.1261047675319724, "learning_rate": 2.5851366319389314e-06, "loss": 0.7651, "step": 21340 }, { "epoch": 0.772944585295183, "grad_norm": 1.428460083487537, "learning_rate": 2.5843495920972083e-06, "loss": 0.6641, "step": 21341 }, { "epoch": 0.7729808040565013, "grad_norm": 1.3490403489792147, "learning_rate": 2.583562654301355e-06, "loss": 0.6594, "step": 21342 }, { "epoch": 0.7730170228178196, "grad_norm": 1.3289564755618086, "learning_rate": 2.582775818562194e-06, "loss": 0.6179, "step": 21343 }, { "epoch": 0.773053241579138, "grad_norm": 1.3990076264144737, "learning_rate": 2.581989084890557e-06, "loss": 0.6819, "step": 21344 }, { "epoch": 0.7730894603404563, "grad_norm": 1.515970522537197, "learning_rate": 2.581202453297265e-06, "loss": 0.7536, "step": 21345 }, { "epoch": 0.7731256791017748, "grad_norm": 1.466534696531238, "learning_rate": 2.580415923793147e-06, "loss": 0.6814, "step": 21346 }, { "epoch": 0.7731618978630931, "grad_norm": 1.033442509915327, "learning_rate": 2.5796294963890267e-06, "loss": 0.6708, "step": 21347 }, { "epoch": 0.7731981166244114, "grad_norm": 0.9885475358062933, "learning_rate": 2.578843171095725e-06, "loss": 0.6835, "step": 21348 }, { "epoch": 0.7732343353857298, "grad_norm": 1.2841669176286274, "learning_rate": 2.5780569479240593e-06, "loss": 0.6834, "step": 21349 }, { "epoch": 0.7732705541470482, "grad_norm": 1.0381349075323207, "learning_rate": 2.5772708268848535e-06, "loss": 0.6276, "step": 21350 }, { "epoch": 0.7733067729083666, "grad_norm": 1.3313485854543026, "learning_rate": 2.576484807988924e-06, "loss": 0.6071, "step": 21351 }, { "epoch": 0.7733429916696849, "grad_norm": 1.3426678744349716, "learning_rate": 2.5756988912470825e-06, "loss": 0.5986, "step": 21352 }, { "epoch": 0.7733792104310032, "grad_norm": 1.885184890681271, "learning_rate": 2.5749130766701514e-06, "loss": 0.6706, "step": 21353 }, { "epoch": 0.7734154291923216, "grad_norm": 1.401134154585981, "learning_rate": 2.5741273642689412e-06, "loss": 0.7002, "step": 21354 }, { "epoch": 0.77345164795364, "grad_norm": 1.394342143863449, "learning_rate": 2.5733417540542627e-06, "loss": 0.6925, "step": 21355 }, { "epoch": 0.7734878667149584, "grad_norm": 1.468180903807195, "learning_rate": 2.5725562460369247e-06, "loss": 0.6564, "step": 21356 }, { "epoch": 0.7735240854762767, "grad_norm": 1.4439547166552933, "learning_rate": 2.5717708402277407e-06, "loss": 0.6973, "step": 21357 }, { "epoch": 0.773560304237595, "grad_norm": 1.3029856060679292, "learning_rate": 2.570985536637519e-06, "loss": 0.5794, "step": 21358 }, { "epoch": 0.7735965229989135, "grad_norm": 1.0248226500082567, "learning_rate": 2.5702003352770632e-06, "loss": 0.7335, "step": 21359 }, { "epoch": 0.7736327417602318, "grad_norm": 1.5213963485590287, "learning_rate": 2.5694152361571756e-06, "loss": 0.6827, "step": 21360 }, { "epoch": 0.7736689605215502, "grad_norm": 1.110476836480548, "learning_rate": 2.5686302392886675e-06, "loss": 0.701, "step": 21361 }, { "epoch": 0.7737051792828685, "grad_norm": 1.0114503502957863, "learning_rate": 2.567845344682336e-06, "loss": 0.7039, "step": 21362 }, { "epoch": 0.7737413980441868, "grad_norm": 1.3784983351734679, "learning_rate": 2.56706055234898e-06, "loss": 0.6627, "step": 21363 }, { "epoch": 0.7737776168055053, "grad_norm": 1.4309277053959322, "learning_rate": 2.5662758622994056e-06, "loss": 0.7433, "step": 21364 }, { "epoch": 0.7738138355668236, "grad_norm": 1.3727303013831502, "learning_rate": 2.5654912745444073e-06, "loss": 0.6491, "step": 21365 }, { "epoch": 0.773850054328142, "grad_norm": 1.3719411151102474, "learning_rate": 2.564706789094782e-06, "loss": 0.702, "step": 21366 }, { "epoch": 0.7738862730894603, "grad_norm": 1.323077255612021, "learning_rate": 2.56392240596132e-06, "loss": 0.6681, "step": 21367 }, { "epoch": 0.7739224918507787, "grad_norm": 2.7957466291131774, "learning_rate": 2.563138125154824e-06, "loss": 0.6558, "step": 21368 }, { "epoch": 0.7739587106120971, "grad_norm": 1.3649251816950625, "learning_rate": 2.5623539466860813e-06, "loss": 0.7021, "step": 21369 }, { "epoch": 0.7739949293734154, "grad_norm": 1.4383094539031982, "learning_rate": 2.561569870565883e-06, "loss": 0.7206, "step": 21370 }, { "epoch": 0.7740311481347338, "grad_norm": 1.139519825994743, "learning_rate": 2.560785896805018e-06, "loss": 0.5795, "step": 21371 }, { "epoch": 0.7740673668960522, "grad_norm": 1.2872241349643503, "learning_rate": 2.5600020254142776e-06, "loss": 0.5841, "step": 21372 }, { "epoch": 0.7741035856573705, "grad_norm": 1.3364858526709036, "learning_rate": 2.5592182564044465e-06, "loss": 0.6101, "step": 21373 }, { "epoch": 0.7741398044186889, "grad_norm": 1.4224460746641283, "learning_rate": 2.5584345897863074e-06, "loss": 0.619, "step": 21374 }, { "epoch": 0.7741760231800072, "grad_norm": 1.3466427018561815, "learning_rate": 2.55765102557065e-06, "loss": 0.6587, "step": 21375 }, { "epoch": 0.7742122419413257, "grad_norm": 1.349064236580829, "learning_rate": 2.556867563768254e-06, "loss": 0.6772, "step": 21376 }, { "epoch": 0.774248460702644, "grad_norm": 1.3485984279936836, "learning_rate": 2.5560842043899005e-06, "loss": 0.7553, "step": 21377 }, { "epoch": 0.7742846794639623, "grad_norm": 1.2687975318466063, "learning_rate": 2.5553009474463665e-06, "loss": 0.6492, "step": 21378 }, { "epoch": 0.7743208982252807, "grad_norm": 1.4649937686395023, "learning_rate": 2.5545177929484355e-06, "loss": 0.6556, "step": 21379 }, { "epoch": 0.774357116986599, "grad_norm": 1.1630778539110038, "learning_rate": 2.5537347409068826e-06, "loss": 0.7412, "step": 21380 }, { "epoch": 0.7743933357479175, "grad_norm": 1.4017014719512968, "learning_rate": 2.552951791332482e-06, "loss": 0.6472, "step": 21381 }, { "epoch": 0.7744295545092358, "grad_norm": 1.017052129563838, "learning_rate": 2.5521689442360044e-06, "loss": 0.7242, "step": 21382 }, { "epoch": 0.7744657732705541, "grad_norm": 1.0300705679204272, "learning_rate": 2.5513861996282306e-06, "loss": 0.659, "step": 21383 }, { "epoch": 0.7745019920318725, "grad_norm": 1.3858668181900489, "learning_rate": 2.5506035575199273e-06, "loss": 0.7538, "step": 21384 }, { "epoch": 0.7745382107931909, "grad_norm": 1.3589446678786206, "learning_rate": 2.549821017921862e-06, "loss": 0.7241, "step": 21385 }, { "epoch": 0.7745744295545093, "grad_norm": 1.368433959065186, "learning_rate": 2.549038580844808e-06, "loss": 0.6806, "step": 21386 }, { "epoch": 0.7746106483158276, "grad_norm": 1.2605611196870976, "learning_rate": 2.5482562462995296e-06, "loss": 0.6493, "step": 21387 }, { "epoch": 0.7746468670771459, "grad_norm": 1.228741736491056, "learning_rate": 2.5474740142967946e-06, "loss": 0.7047, "step": 21388 }, { "epoch": 0.7746830858384643, "grad_norm": 1.3199074387642882, "learning_rate": 2.546691884847361e-06, "loss": 0.7267, "step": 21389 }, { "epoch": 0.7747193045997827, "grad_norm": 1.3620359039402177, "learning_rate": 2.5459098579620034e-06, "loss": 0.7356, "step": 21390 }, { "epoch": 0.7747555233611011, "grad_norm": 0.9683117888437224, "learning_rate": 2.545127933651472e-06, "loss": 0.6766, "step": 21391 }, { "epoch": 0.7747917421224194, "grad_norm": 1.3808540657081674, "learning_rate": 2.544346111926528e-06, "loss": 0.6641, "step": 21392 }, { "epoch": 0.7748279608837377, "grad_norm": 1.332461694138071, "learning_rate": 2.543564392797936e-06, "loss": 0.7113, "step": 21393 }, { "epoch": 0.7748641796450562, "grad_norm": 1.3201211813279177, "learning_rate": 2.542782776276449e-06, "loss": 0.6516, "step": 21394 }, { "epoch": 0.7749003984063745, "grad_norm": 1.4265574393543006, "learning_rate": 2.542001262372821e-06, "loss": 0.7689, "step": 21395 }, { "epoch": 0.7749366171676929, "grad_norm": 1.38685595008287, "learning_rate": 2.5412198510978116e-06, "loss": 0.7305, "step": 21396 }, { "epoch": 0.7749728359290112, "grad_norm": 1.3058813566228629, "learning_rate": 2.54043854246217e-06, "loss": 0.7119, "step": 21397 }, { "epoch": 0.7750090546903295, "grad_norm": 1.2197529612021674, "learning_rate": 2.5396573364766507e-06, "loss": 0.6708, "step": 21398 }, { "epoch": 0.775045273451648, "grad_norm": 0.976235334191034, "learning_rate": 2.5388762331519967e-06, "loss": 0.6769, "step": 21399 }, { "epoch": 0.7750814922129663, "grad_norm": 0.9374562040917453, "learning_rate": 2.538095232498966e-06, "loss": 0.6812, "step": 21400 }, { "epoch": 0.7751177109742847, "grad_norm": 1.2681631861783116, "learning_rate": 2.537314334528301e-06, "loss": 0.7168, "step": 21401 }, { "epoch": 0.775153929735603, "grad_norm": 1.382774091176386, "learning_rate": 2.536533539250747e-06, "loss": 0.6452, "step": 21402 }, { "epoch": 0.7751901484969214, "grad_norm": 1.3261032992682467, "learning_rate": 2.535752846677049e-06, "loss": 0.6755, "step": 21403 }, { "epoch": 0.7752263672582398, "grad_norm": 1.2397243434745395, "learning_rate": 2.5349722568179514e-06, "loss": 0.6244, "step": 21404 }, { "epoch": 0.7752625860195581, "grad_norm": 1.3026381733044943, "learning_rate": 2.5341917696841954e-06, "loss": 0.6966, "step": 21405 }, { "epoch": 0.7752988047808765, "grad_norm": 1.3716404957735118, "learning_rate": 2.5334113852865185e-06, "loss": 0.7386, "step": 21406 }, { "epoch": 0.7753350235421949, "grad_norm": 3.676677164963847, "learning_rate": 2.532631103635663e-06, "loss": 0.684, "step": 21407 }, { "epoch": 0.7753712423035132, "grad_norm": 1.314392709122993, "learning_rate": 2.531850924742366e-06, "loss": 0.6157, "step": 21408 }, { "epoch": 0.7754074610648316, "grad_norm": 1.0365362795247546, "learning_rate": 2.531070848617362e-06, "loss": 0.6606, "step": 21409 }, { "epoch": 0.7754436798261499, "grad_norm": 1.027254591293902, "learning_rate": 2.5302908752713828e-06, "loss": 0.6483, "step": 21410 }, { "epoch": 0.7754798985874684, "grad_norm": 1.4925397312404716, "learning_rate": 2.5295110047151683e-06, "loss": 0.6775, "step": 21411 }, { "epoch": 0.7755161173487867, "grad_norm": 1.5604204883409805, "learning_rate": 2.5287312369594463e-06, "loss": 0.7244, "step": 21412 }, { "epoch": 0.775552336110105, "grad_norm": 1.2888741973412323, "learning_rate": 2.527951572014947e-06, "loss": 0.6992, "step": 21413 }, { "epoch": 0.7755885548714234, "grad_norm": 0.976372557049367, "learning_rate": 2.5271720098923968e-06, "loss": 0.6583, "step": 21414 }, { "epoch": 0.7756247736327417, "grad_norm": 1.340452488279123, "learning_rate": 2.5263925506025288e-06, "loss": 0.7064, "step": 21415 }, { "epoch": 0.7756609923940602, "grad_norm": 1.3075977940791434, "learning_rate": 2.525613194156067e-06, "loss": 0.6904, "step": 21416 }, { "epoch": 0.7756972111553785, "grad_norm": 1.5213799483635089, "learning_rate": 2.524833940563731e-06, "loss": 0.647, "step": 21417 }, { "epoch": 0.7757334299166968, "grad_norm": 1.063058761310075, "learning_rate": 2.5240547898362522e-06, "loss": 0.6951, "step": 21418 }, { "epoch": 0.7757696486780152, "grad_norm": 1.3066073977945591, "learning_rate": 2.523275741984348e-06, "loss": 0.6538, "step": 21419 }, { "epoch": 0.7758058674393336, "grad_norm": 1.3857469014254953, "learning_rate": 2.5224967970187386e-06, "loss": 0.6314, "step": 21420 }, { "epoch": 0.775842086200652, "grad_norm": 1.3193529196407916, "learning_rate": 2.5217179549501415e-06, "loss": 0.6686, "step": 21421 }, { "epoch": 0.7758783049619703, "grad_norm": 1.563841133334186, "learning_rate": 2.5209392157892788e-06, "loss": 0.5949, "step": 21422 }, { "epoch": 0.7759145237232886, "grad_norm": 1.3294750819724699, "learning_rate": 2.5201605795468642e-06, "loss": 0.679, "step": 21423 }, { "epoch": 0.775950742484607, "grad_norm": 1.3271617186992646, "learning_rate": 2.519382046233613e-06, "loss": 0.6839, "step": 21424 }, { "epoch": 0.7759869612459254, "grad_norm": 1.3541594518108613, "learning_rate": 2.5186036158602344e-06, "loss": 0.6801, "step": 21425 }, { "epoch": 0.7760231800072438, "grad_norm": 1.293393035204106, "learning_rate": 2.5178252884374465e-06, "loss": 0.6395, "step": 21426 }, { "epoch": 0.7760593987685621, "grad_norm": 1.4486730748691927, "learning_rate": 2.517047063975957e-06, "loss": 0.7104, "step": 21427 }, { "epoch": 0.7760956175298804, "grad_norm": 1.3639046997763347, "learning_rate": 2.5162689424864728e-06, "loss": 0.6444, "step": 21428 }, { "epoch": 0.7761318362911989, "grad_norm": 1.2693452896956965, "learning_rate": 2.5154909239797064e-06, "loss": 0.584, "step": 21429 }, { "epoch": 0.7761680550525172, "grad_norm": 1.4340996635633707, "learning_rate": 2.5147130084663607e-06, "loss": 0.7785, "step": 21430 }, { "epoch": 0.7762042738138356, "grad_norm": 1.4580857964450034, "learning_rate": 2.513935195957142e-06, "loss": 0.7173, "step": 21431 }, { "epoch": 0.7762404925751539, "grad_norm": 1.4325690420650947, "learning_rate": 2.5131574864627494e-06, "loss": 0.734, "step": 21432 }, { "epoch": 0.7762767113364722, "grad_norm": 1.0405532187807391, "learning_rate": 2.5123798799938913e-06, "loss": 0.65, "step": 21433 }, { "epoch": 0.7763129300977907, "grad_norm": 1.275159559305181, "learning_rate": 2.511602376561265e-06, "loss": 0.6358, "step": 21434 }, { "epoch": 0.776349148859109, "grad_norm": 0.9432376819855202, "learning_rate": 2.5108249761755676e-06, "loss": 0.7962, "step": 21435 }, { "epoch": 0.7763853676204274, "grad_norm": 1.3951832892491032, "learning_rate": 2.5100476788475027e-06, "loss": 0.6758, "step": 21436 }, { "epoch": 0.7764215863817457, "grad_norm": 1.2774874871718016, "learning_rate": 2.509270484587761e-06, "loss": 0.7164, "step": 21437 }, { "epoch": 0.7764578051430641, "grad_norm": 1.259243902950889, "learning_rate": 2.508493393407041e-06, "loss": 0.6779, "step": 21438 }, { "epoch": 0.7764940239043825, "grad_norm": 1.43875347820594, "learning_rate": 2.507716405316031e-06, "loss": 0.6973, "step": 21439 }, { "epoch": 0.7765302426657008, "grad_norm": 1.2528036943210892, "learning_rate": 2.5069395203254287e-06, "loss": 0.6471, "step": 21440 }, { "epoch": 0.7765664614270192, "grad_norm": 1.3414404560897752, "learning_rate": 2.5061627384459243e-06, "loss": 0.678, "step": 21441 }, { "epoch": 0.7766026801883376, "grad_norm": 1.4047595152041312, "learning_rate": 2.5053860596882042e-06, "loss": 0.6134, "step": 21442 }, { "epoch": 0.7766388989496559, "grad_norm": 1.2954951747565864, "learning_rate": 2.5046094840629543e-06, "loss": 0.7316, "step": 21443 }, { "epoch": 0.7766751177109743, "grad_norm": 1.5440406504468818, "learning_rate": 2.5038330115808673e-06, "loss": 0.7457, "step": 21444 }, { "epoch": 0.7767113364722926, "grad_norm": 1.2895097988734563, "learning_rate": 2.5030566422526236e-06, "loss": 0.7004, "step": 21445 }, { "epoch": 0.776747555233611, "grad_norm": 1.9625151697410779, "learning_rate": 2.502280376088906e-06, "loss": 0.6769, "step": 21446 }, { "epoch": 0.7767837739949294, "grad_norm": 1.3736116867361008, "learning_rate": 2.501504213100402e-06, "loss": 0.6875, "step": 21447 }, { "epoch": 0.7768199927562477, "grad_norm": 1.129638259885353, "learning_rate": 2.500728153297788e-06, "loss": 0.6514, "step": 21448 }, { "epoch": 0.7768562115175661, "grad_norm": 1.2228325063210417, "learning_rate": 2.499952196691744e-06, "loss": 0.702, "step": 21449 }, { "epoch": 0.7768924302788844, "grad_norm": 1.5768549052510499, "learning_rate": 2.499176343292945e-06, "loss": 0.7366, "step": 21450 }, { "epoch": 0.7769286490402029, "grad_norm": 1.3173169528908364, "learning_rate": 2.4984005931120735e-06, "loss": 0.6799, "step": 21451 }, { "epoch": 0.7769648678015212, "grad_norm": 1.0630412543041967, "learning_rate": 2.4976249461598013e-06, "loss": 0.7499, "step": 21452 }, { "epoch": 0.7770010865628395, "grad_norm": 0.995320110593948, "learning_rate": 2.4968494024468016e-06, "loss": 0.6552, "step": 21453 }, { "epoch": 0.7770373053241579, "grad_norm": 1.3710269476981503, "learning_rate": 2.4960739619837436e-06, "loss": 0.6976, "step": 21454 }, { "epoch": 0.7770735240854763, "grad_norm": 1.0970547531303714, "learning_rate": 2.495298624781306e-06, "loss": 0.6561, "step": 21455 }, { "epoch": 0.7771097428467947, "grad_norm": 1.3952238636228587, "learning_rate": 2.494523390850152e-06, "loss": 0.6676, "step": 21456 }, { "epoch": 0.777145961608113, "grad_norm": 1.3302577418408263, "learning_rate": 2.493748260200949e-06, "loss": 0.587, "step": 21457 }, { "epoch": 0.7771821803694313, "grad_norm": 1.4437936902402289, "learning_rate": 2.4929732328443677e-06, "loss": 0.672, "step": 21458 }, { "epoch": 0.7772183991307497, "grad_norm": 1.5116426870150723, "learning_rate": 2.492198308791072e-06, "loss": 0.6943, "step": 21459 }, { "epoch": 0.7772546178920681, "grad_norm": 1.0866810755863248, "learning_rate": 2.491423488051723e-06, "loss": 0.6621, "step": 21460 }, { "epoch": 0.7772908366533865, "grad_norm": 1.5169477663280972, "learning_rate": 2.490648770636983e-06, "loss": 0.6617, "step": 21461 }, { "epoch": 0.7773270554147048, "grad_norm": 1.4513195445321525, "learning_rate": 2.4898741565575157e-06, "loss": 0.6571, "step": 21462 }, { "epoch": 0.7773632741760231, "grad_norm": 1.0852601511779532, "learning_rate": 2.4890996458239804e-06, "loss": 0.6984, "step": 21463 }, { "epoch": 0.7773994929373416, "grad_norm": 1.431885762063905, "learning_rate": 2.4883252384470325e-06, "loss": 0.6808, "step": 21464 }, { "epoch": 0.7774357116986599, "grad_norm": 1.3811649304134503, "learning_rate": 2.487550934437327e-06, "loss": 0.6136, "step": 21465 }, { "epoch": 0.7774719304599783, "grad_norm": 1.3090010642510639, "learning_rate": 2.486776733805526e-06, "loss": 0.6717, "step": 21466 }, { "epoch": 0.7775081492212966, "grad_norm": 1.2796200278593513, "learning_rate": 2.486002636562277e-06, "loss": 0.687, "step": 21467 }, { "epoch": 0.777544367982615, "grad_norm": 0.9293189189664701, "learning_rate": 2.4852286427182313e-06, "loss": 0.5994, "step": 21468 }, { "epoch": 0.7775805867439334, "grad_norm": 1.2963308153473418, "learning_rate": 2.484454752284047e-06, "loss": 0.6976, "step": 21469 }, { "epoch": 0.7776168055052517, "grad_norm": 1.4089754172014737, "learning_rate": 2.4836809652703673e-06, "loss": 0.7349, "step": 21470 }, { "epoch": 0.7776530242665701, "grad_norm": 1.5819512335239285, "learning_rate": 2.482907281687843e-06, "loss": 0.6375, "step": 21471 }, { "epoch": 0.7776892430278884, "grad_norm": 0.9725034300277481, "learning_rate": 2.482133701547117e-06, "loss": 0.647, "step": 21472 }, { "epoch": 0.7777254617892068, "grad_norm": 1.2743809648213305, "learning_rate": 2.4813602248588387e-06, "loss": 0.6767, "step": 21473 }, { "epoch": 0.7777616805505252, "grad_norm": 1.083856729776021, "learning_rate": 2.480586851633651e-06, "loss": 0.6701, "step": 21474 }, { "epoch": 0.7777978993118435, "grad_norm": 2.1649598351050012, "learning_rate": 2.4798135818821945e-06, "loss": 0.7342, "step": 21475 }, { "epoch": 0.7778341180731619, "grad_norm": 0.9138952794111959, "learning_rate": 2.4790404156151095e-06, "loss": 0.6646, "step": 21476 }, { "epoch": 0.7778703368344803, "grad_norm": 1.4379386336300202, "learning_rate": 2.478267352843039e-06, "loss": 0.7186, "step": 21477 }, { "epoch": 0.7779065555957986, "grad_norm": 1.3333767153405343, "learning_rate": 2.4774943935766184e-06, "loss": 0.6587, "step": 21478 }, { "epoch": 0.777942774357117, "grad_norm": 1.4327113816824846, "learning_rate": 2.476721537826482e-06, "loss": 0.6814, "step": 21479 }, { "epoch": 0.7779789931184353, "grad_norm": 0.9879378037827083, "learning_rate": 2.4759487856032714e-06, "loss": 0.6989, "step": 21480 }, { "epoch": 0.7780152118797538, "grad_norm": 1.3885894233165754, "learning_rate": 2.4751761369176165e-06, "loss": 0.6533, "step": 21481 }, { "epoch": 0.7780514306410721, "grad_norm": 1.4078020226949233, "learning_rate": 2.474403591780149e-06, "loss": 0.6452, "step": 21482 }, { "epoch": 0.7780876494023904, "grad_norm": 1.1044173066661929, "learning_rate": 2.4736311502014986e-06, "loss": 0.6937, "step": 21483 }, { "epoch": 0.7781238681637088, "grad_norm": 0.9553330483536066, "learning_rate": 2.472858812192299e-06, "loss": 0.6772, "step": 21484 }, { "epoch": 0.7781600869250271, "grad_norm": 2.7426740752641514, "learning_rate": 2.4720865777631775e-06, "loss": 0.6601, "step": 21485 }, { "epoch": 0.7781963056863456, "grad_norm": 1.2464295595477837, "learning_rate": 2.4713144469247584e-06, "loss": 0.6065, "step": 21486 }, { "epoch": 0.7782325244476639, "grad_norm": 1.3304853152141523, "learning_rate": 2.4705424196876647e-06, "loss": 0.7025, "step": 21487 }, { "epoch": 0.7782687432089822, "grad_norm": 1.0330090084644203, "learning_rate": 2.4697704960625256e-06, "loss": 0.6883, "step": 21488 }, { "epoch": 0.7783049619703006, "grad_norm": 1.596822637755813, "learning_rate": 2.4689986760599626e-06, "loss": 0.6216, "step": 21489 }, { "epoch": 0.778341180731619, "grad_norm": 1.3761956938588837, "learning_rate": 2.4682269596905915e-06, "loss": 0.68, "step": 21490 }, { "epoch": 0.7783773994929374, "grad_norm": 1.420241684559342, "learning_rate": 2.4674553469650387e-06, "loss": 0.6899, "step": 21491 }, { "epoch": 0.7784136182542557, "grad_norm": 1.241095229877616, "learning_rate": 2.4666838378939185e-06, "loss": 0.6601, "step": 21492 }, { "epoch": 0.778449837015574, "grad_norm": 1.113269908119451, "learning_rate": 2.465912432487848e-06, "loss": 0.6933, "step": 21493 }, { "epoch": 0.7784860557768924, "grad_norm": 1.3643250711637251, "learning_rate": 2.465141130757439e-06, "loss": 0.7093, "step": 21494 }, { "epoch": 0.7785222745382108, "grad_norm": 1.2015465068822404, "learning_rate": 2.464369932713313e-06, "loss": 0.6998, "step": 21495 }, { "epoch": 0.7785584932995292, "grad_norm": 1.284698198891465, "learning_rate": 2.463598838366077e-06, "loss": 0.6484, "step": 21496 }, { "epoch": 0.7785947120608475, "grad_norm": 1.4263678441417678, "learning_rate": 2.462827847726341e-06, "loss": 0.6466, "step": 21497 }, { "epoch": 0.7786309308221658, "grad_norm": 1.3325606878291218, "learning_rate": 2.4620569608047197e-06, "loss": 0.6691, "step": 21498 }, { "epoch": 0.7786671495834843, "grad_norm": 1.4279176774404951, "learning_rate": 2.4612861776118168e-06, "loss": 0.6382, "step": 21499 }, { "epoch": 0.7787033683448026, "grad_norm": 1.4634783287500963, "learning_rate": 2.4605154981582415e-06, "loss": 0.7577, "step": 21500 }, { "epoch": 0.778739587106121, "grad_norm": 1.320736354307178, "learning_rate": 2.4597449224545933e-06, "loss": 0.7138, "step": 21501 }, { "epoch": 0.7787758058674393, "grad_norm": 1.3609640901729325, "learning_rate": 2.458974450511484e-06, "loss": 0.7061, "step": 21502 }, { "epoch": 0.7788120246287576, "grad_norm": 1.3747287395974819, "learning_rate": 2.4582040823395126e-06, "loss": 0.665, "step": 21503 }, { "epoch": 0.7788482433900761, "grad_norm": 1.2821423436824197, "learning_rate": 2.4574338179492797e-06, "loss": 0.6807, "step": 21504 }, { "epoch": 0.7788844621513944, "grad_norm": 1.4544544393846275, "learning_rate": 2.4566636573513826e-06, "loss": 0.7227, "step": 21505 }, { "epoch": 0.7789206809127128, "grad_norm": 1.2967085957770235, "learning_rate": 2.4558936005564237e-06, "loss": 0.6447, "step": 21506 }, { "epoch": 0.7789568996740311, "grad_norm": 1.2880229826112404, "learning_rate": 2.4551236475749985e-06, "loss": 0.68, "step": 21507 }, { "epoch": 0.7789931184353495, "grad_norm": 1.3703302515482974, "learning_rate": 2.454353798417698e-06, "loss": 0.692, "step": 21508 }, { "epoch": 0.7790293371966679, "grad_norm": 1.5544027925628772, "learning_rate": 2.453584053095123e-06, "loss": 0.6877, "step": 21509 }, { "epoch": 0.7790655559579862, "grad_norm": 0.927299033153303, "learning_rate": 2.452814411617861e-06, "loss": 0.6575, "step": 21510 }, { "epoch": 0.7791017747193046, "grad_norm": 1.3088078273913208, "learning_rate": 2.452044873996505e-06, "loss": 0.6785, "step": 21511 }, { "epoch": 0.779137993480623, "grad_norm": 1.6769780651417643, "learning_rate": 2.4512754402416407e-06, "loss": 0.7167, "step": 21512 }, { "epoch": 0.7791742122419413, "grad_norm": 1.41609146638539, "learning_rate": 2.450506110363863e-06, "loss": 0.6873, "step": 21513 }, { "epoch": 0.7792104310032597, "grad_norm": 0.985431581883162, "learning_rate": 2.4497368843737533e-06, "loss": 0.7288, "step": 21514 }, { "epoch": 0.779246649764578, "grad_norm": 1.4583487587919044, "learning_rate": 2.4489677622818997e-06, "loss": 0.6851, "step": 21515 }, { "epoch": 0.7792828685258965, "grad_norm": 1.4863600282591138, "learning_rate": 2.4481987440988807e-06, "loss": 0.6447, "step": 21516 }, { "epoch": 0.7793190872872148, "grad_norm": 1.4021948971817848, "learning_rate": 2.447429829835286e-06, "loss": 0.6915, "step": 21517 }, { "epoch": 0.7793553060485331, "grad_norm": 1.4947769829453075, "learning_rate": 2.4466610195016925e-06, "loss": 0.6705, "step": 21518 }, { "epoch": 0.7793915248098515, "grad_norm": 1.3132101802782783, "learning_rate": 2.4458923131086787e-06, "loss": 0.6543, "step": 21519 }, { "epoch": 0.7794277435711698, "grad_norm": 0.9731251582616122, "learning_rate": 2.4451237106668258e-06, "loss": 0.7084, "step": 21520 }, { "epoch": 0.7794639623324883, "grad_norm": 1.431046042371304, "learning_rate": 2.4443552121867098e-06, "loss": 0.7003, "step": 21521 }, { "epoch": 0.7795001810938066, "grad_norm": 1.3461545061236575, "learning_rate": 2.4435868176789047e-06, "loss": 0.6299, "step": 21522 }, { "epoch": 0.7795363998551249, "grad_norm": 1.4306974335511804, "learning_rate": 2.442818527153982e-06, "loss": 0.7396, "step": 21523 }, { "epoch": 0.7795726186164433, "grad_norm": 1.3567445780531704, "learning_rate": 2.4420503406225194e-06, "loss": 0.6354, "step": 21524 }, { "epoch": 0.7796088373777617, "grad_norm": 1.4231724072967336, "learning_rate": 2.4412822580950846e-06, "loss": 0.7507, "step": 21525 }, { "epoch": 0.7796450561390801, "grad_norm": 1.4000505090227855, "learning_rate": 2.440514279582248e-06, "loss": 0.6528, "step": 21526 }, { "epoch": 0.7796812749003984, "grad_norm": 1.28983894863656, "learning_rate": 2.4397464050945753e-06, "loss": 0.5842, "step": 21527 }, { "epoch": 0.7797174936617167, "grad_norm": 1.4097640107553342, "learning_rate": 2.4389786346426368e-06, "loss": 0.6578, "step": 21528 }, { "epoch": 0.7797537124230352, "grad_norm": 1.3486552426505272, "learning_rate": 2.4382109682369958e-06, "loss": 0.6806, "step": 21529 }, { "epoch": 0.7797899311843535, "grad_norm": 0.9822361661681781, "learning_rate": 2.437443405888215e-06, "loss": 0.618, "step": 21530 }, { "epoch": 0.7798261499456719, "grad_norm": 1.4285433743321414, "learning_rate": 2.4366759476068593e-06, "loss": 0.6822, "step": 21531 }, { "epoch": 0.7798623687069902, "grad_norm": 1.2797330877022968, "learning_rate": 2.4359085934034888e-06, "loss": 0.7026, "step": 21532 }, { "epoch": 0.7798985874683085, "grad_norm": 1.0486050349976912, "learning_rate": 2.4351413432886616e-06, "loss": 0.6952, "step": 21533 }, { "epoch": 0.779934806229627, "grad_norm": 1.4614735247265869, "learning_rate": 2.434374197272935e-06, "loss": 0.679, "step": 21534 }, { "epoch": 0.7799710249909453, "grad_norm": 1.4174210814868182, "learning_rate": 2.4336071553668696e-06, "loss": 0.6605, "step": 21535 }, { "epoch": 0.7800072437522637, "grad_norm": 1.4442664910592926, "learning_rate": 2.4328402175810175e-06, "loss": 0.6704, "step": 21536 }, { "epoch": 0.780043462513582, "grad_norm": 1.3597967497954557, "learning_rate": 2.4320733839259337e-06, "loss": 0.6738, "step": 21537 }, { "epoch": 0.7800796812749003, "grad_norm": 1.295456705360612, "learning_rate": 2.4313066544121676e-06, "loss": 0.6985, "step": 21538 }, { "epoch": 0.7801159000362188, "grad_norm": 1.0258755175458742, "learning_rate": 2.4305400290502745e-06, "loss": 0.6747, "step": 21539 }, { "epoch": 0.7801521187975371, "grad_norm": 1.096829213842638, "learning_rate": 2.429773507850801e-06, "loss": 0.6703, "step": 21540 }, { "epoch": 0.7801883375588555, "grad_norm": 1.2758234561647726, "learning_rate": 2.429007090824295e-06, "loss": 0.6712, "step": 21541 }, { "epoch": 0.7802245563201738, "grad_norm": 1.3148044078525933, "learning_rate": 2.4282407779813054e-06, "loss": 0.7069, "step": 21542 }, { "epoch": 0.7802607750814922, "grad_norm": 1.2646689667379432, "learning_rate": 2.4274745693323753e-06, "loss": 0.6247, "step": 21543 }, { "epoch": 0.7802969938428106, "grad_norm": 1.1693785625104327, "learning_rate": 2.4267084648880503e-06, "loss": 0.6784, "step": 21544 }, { "epoch": 0.7803332126041289, "grad_norm": 1.0677960733684575, "learning_rate": 2.425942464658867e-06, "loss": 0.7093, "step": 21545 }, { "epoch": 0.7803694313654473, "grad_norm": 1.604100955715457, "learning_rate": 2.425176568655374e-06, "loss": 0.7159, "step": 21546 }, { "epoch": 0.7804056501267657, "grad_norm": 1.476626795928058, "learning_rate": 2.4244107768881074e-06, "loss": 0.6464, "step": 21547 }, { "epoch": 0.780441868888084, "grad_norm": 1.0085607925436961, "learning_rate": 2.4236450893676055e-06, "loss": 0.6625, "step": 21548 }, { "epoch": 0.7804780876494024, "grad_norm": 1.4019261298845012, "learning_rate": 2.4228795061044007e-06, "loss": 0.6552, "step": 21549 }, { "epoch": 0.7805143064107207, "grad_norm": 1.3596823059961267, "learning_rate": 2.422114027109035e-06, "loss": 0.6367, "step": 21550 }, { "epoch": 0.7805505251720392, "grad_norm": 1.4278360374983603, "learning_rate": 2.4213486523920393e-06, "loss": 0.6674, "step": 21551 }, { "epoch": 0.7805867439333575, "grad_norm": 1.4474700456465759, "learning_rate": 2.420583381963941e-06, "loss": 0.6797, "step": 21552 }, { "epoch": 0.7806229626946758, "grad_norm": 1.4583571412209413, "learning_rate": 2.419818215835279e-06, "loss": 0.6286, "step": 21553 }, { "epoch": 0.7806591814559942, "grad_norm": 1.331964422602379, "learning_rate": 2.4190531540165798e-06, "loss": 0.6852, "step": 21554 }, { "epoch": 0.7806954002173125, "grad_norm": 1.2659559598770371, "learning_rate": 2.418288196518369e-06, "loss": 0.6995, "step": 21555 }, { "epoch": 0.780731618978631, "grad_norm": 0.9652296830513477, "learning_rate": 2.417523343351171e-06, "loss": 0.7075, "step": 21556 }, { "epoch": 0.7807678377399493, "grad_norm": 1.4520725590853687, "learning_rate": 2.416758594525519e-06, "loss": 0.7623, "step": 21557 }, { "epoch": 0.7808040565012676, "grad_norm": 1.3839924110353716, "learning_rate": 2.4159939500519313e-06, "loss": 0.72, "step": 21558 }, { "epoch": 0.780840275262586, "grad_norm": 1.351755546691389, "learning_rate": 2.415229409940928e-06, "loss": 0.7196, "step": 21559 }, { "epoch": 0.7808764940239044, "grad_norm": 1.372774384701281, "learning_rate": 2.4144649742030348e-06, "loss": 0.7024, "step": 21560 }, { "epoch": 0.7809127127852228, "grad_norm": 1.2058006768171199, "learning_rate": 2.4137006428487696e-06, "loss": 0.6692, "step": 21561 }, { "epoch": 0.7809489315465411, "grad_norm": 1.172703068152506, "learning_rate": 2.4129364158886492e-06, "loss": 0.6405, "step": 21562 }, { "epoch": 0.7809851503078594, "grad_norm": 1.2494141532011778, "learning_rate": 2.412172293333187e-06, "loss": 0.6036, "step": 21563 }, { "epoch": 0.7810213690691779, "grad_norm": 1.5626493466051887, "learning_rate": 2.4114082751929045e-06, "loss": 0.6667, "step": 21564 }, { "epoch": 0.7810575878304962, "grad_norm": 1.3262687375253288, "learning_rate": 2.4106443614783116e-06, "loss": 0.722, "step": 21565 }, { "epoch": 0.7810938065918146, "grad_norm": 1.0944227167082823, "learning_rate": 2.409880552199921e-06, "loss": 0.6474, "step": 21566 }, { "epoch": 0.7811300253531329, "grad_norm": 1.0572990740287462, "learning_rate": 2.4091168473682414e-06, "loss": 0.695, "step": 21567 }, { "epoch": 0.7811662441144512, "grad_norm": 1.2498750847747673, "learning_rate": 2.4083532469937866e-06, "loss": 0.7013, "step": 21568 }, { "epoch": 0.7812024628757697, "grad_norm": 1.288164373559102, "learning_rate": 2.4075897510870604e-06, "loss": 0.6844, "step": 21569 }, { "epoch": 0.781238681637088, "grad_norm": 1.0020137047253952, "learning_rate": 2.4068263596585685e-06, "loss": 0.7067, "step": 21570 }, { "epoch": 0.7812749003984064, "grad_norm": 1.3588052499730003, "learning_rate": 2.4060630727188205e-06, "loss": 0.656, "step": 21571 }, { "epoch": 0.7813111191597247, "grad_norm": 1.4895059772781611, "learning_rate": 2.4052998902783176e-06, "loss": 0.6945, "step": 21572 }, { "epoch": 0.781347337921043, "grad_norm": 1.123456665718335, "learning_rate": 2.404536812347561e-06, "loss": 0.6667, "step": 21573 }, { "epoch": 0.7813835566823615, "grad_norm": 1.328838187756055, "learning_rate": 2.40377383893705e-06, "loss": 0.6278, "step": 21574 }, { "epoch": 0.7814197754436798, "grad_norm": 1.3141421468123937, "learning_rate": 2.4030109700572877e-06, "loss": 0.6882, "step": 21575 }, { "epoch": 0.7814559942049982, "grad_norm": 1.3483814665467002, "learning_rate": 2.402248205718769e-06, "loss": 0.7013, "step": 21576 }, { "epoch": 0.7814922129663165, "grad_norm": 1.3323886027022, "learning_rate": 2.401485545931992e-06, "loss": 0.7515, "step": 21577 }, { "epoch": 0.7815284317276349, "grad_norm": 1.303483584578647, "learning_rate": 2.400722990707447e-06, "loss": 0.6482, "step": 21578 }, { "epoch": 0.7815646504889533, "grad_norm": 1.3732623966201967, "learning_rate": 2.3999605400556347e-06, "loss": 0.702, "step": 21579 }, { "epoch": 0.7816008692502716, "grad_norm": 1.4155573017678076, "learning_rate": 2.3991981939870424e-06, "loss": 0.6819, "step": 21580 }, { "epoch": 0.78163708801159, "grad_norm": 1.3617071769242262, "learning_rate": 2.398435952512159e-06, "loss": 0.6867, "step": 21581 }, { "epoch": 0.7816733067729084, "grad_norm": 1.4280237290830269, "learning_rate": 2.3976738156414804e-06, "loss": 0.6774, "step": 21582 }, { "epoch": 0.7817095255342267, "grad_norm": 0.9948560611175857, "learning_rate": 2.3969117833854884e-06, "loss": 0.6354, "step": 21583 }, { "epoch": 0.7817457442955451, "grad_norm": 1.3711675395835092, "learning_rate": 2.3961498557546716e-06, "loss": 0.664, "step": 21584 }, { "epoch": 0.7817819630568634, "grad_norm": 0.9788723562333166, "learning_rate": 2.3953880327595113e-06, "loss": 0.6454, "step": 21585 }, { "epoch": 0.7818181818181819, "grad_norm": 1.4575187440291792, "learning_rate": 2.3946263144104966e-06, "loss": 0.7351, "step": 21586 }, { "epoch": 0.7818544005795002, "grad_norm": 1.099825549537508, "learning_rate": 2.393864700718106e-06, "loss": 0.6849, "step": 21587 }, { "epoch": 0.7818906193408185, "grad_norm": 1.4291751319985102, "learning_rate": 2.3931031916928203e-06, "loss": 0.6861, "step": 21588 }, { "epoch": 0.7819268381021369, "grad_norm": 1.096995404932018, "learning_rate": 2.392341787345116e-06, "loss": 0.6512, "step": 21589 }, { "epoch": 0.7819630568634552, "grad_norm": 1.3394122867861156, "learning_rate": 2.3915804876854764e-06, "loss": 0.6744, "step": 21590 }, { "epoch": 0.7819992756247737, "grad_norm": 1.0169421494849895, "learning_rate": 2.3908192927243747e-06, "loss": 0.6762, "step": 21591 }, { "epoch": 0.782035494386092, "grad_norm": 1.4556036083505348, "learning_rate": 2.390058202472283e-06, "loss": 0.6908, "step": 21592 }, { "epoch": 0.7820717131474103, "grad_norm": 1.4091217996678878, "learning_rate": 2.38929721693968e-06, "loss": 0.6743, "step": 21593 }, { "epoch": 0.7821079319087287, "grad_norm": 1.356007006709518, "learning_rate": 2.388536336137034e-06, "loss": 0.7001, "step": 21594 }, { "epoch": 0.7821441506700471, "grad_norm": 1.1100066047772463, "learning_rate": 2.3877755600748164e-06, "loss": 0.7547, "step": 21595 }, { "epoch": 0.7821803694313655, "grad_norm": 1.4286164013242282, "learning_rate": 2.3870148887634926e-06, "loss": 0.6397, "step": 21596 }, { "epoch": 0.7822165881926838, "grad_norm": 1.0753221086269833, "learning_rate": 2.3862543222135358e-06, "loss": 0.701, "step": 21597 }, { "epoch": 0.7822528069540021, "grad_norm": 1.358389619240567, "learning_rate": 2.3854938604354105e-06, "loss": 0.6726, "step": 21598 }, { "epoch": 0.7822890257153206, "grad_norm": 1.25763270843521, "learning_rate": 2.3847335034395803e-06, "loss": 0.6174, "step": 21599 }, { "epoch": 0.7823252444766389, "grad_norm": 1.3144072613926614, "learning_rate": 2.3839732512365056e-06, "loss": 0.637, "step": 21600 }, { "epoch": 0.7823614632379573, "grad_norm": 1.5383564447765699, "learning_rate": 2.3832131038366556e-06, "loss": 0.6609, "step": 21601 }, { "epoch": 0.7823976819992756, "grad_norm": 1.05038648715606, "learning_rate": 2.3824530612504816e-06, "loss": 0.6868, "step": 21602 }, { "epoch": 0.7824339007605939, "grad_norm": 1.54166902111546, "learning_rate": 2.381693123488449e-06, "loss": 0.6862, "step": 21603 }, { "epoch": 0.7824701195219124, "grad_norm": 1.3696036300774785, "learning_rate": 2.380933290561014e-06, "loss": 0.6757, "step": 21604 }, { "epoch": 0.7825063382832307, "grad_norm": 1.3350333021177903, "learning_rate": 2.3801735624786305e-06, "loss": 0.6855, "step": 21605 }, { "epoch": 0.7825425570445491, "grad_norm": 1.375827488834349, "learning_rate": 2.379413939251751e-06, "loss": 0.644, "step": 21606 }, { "epoch": 0.7825787758058674, "grad_norm": 1.3462191867002806, "learning_rate": 2.3786544208908356e-06, "loss": 0.5944, "step": 21607 }, { "epoch": 0.7826149945671858, "grad_norm": 1.3739399833793844, "learning_rate": 2.3778950074063314e-06, "loss": 0.7053, "step": 21608 }, { "epoch": 0.7826512133285042, "grad_norm": 1.3065157824977578, "learning_rate": 2.37713569880869e-06, "loss": 0.6883, "step": 21609 }, { "epoch": 0.7826874320898225, "grad_norm": 1.1160876476352524, "learning_rate": 2.3763764951083556e-06, "loss": 0.7221, "step": 21610 }, { "epoch": 0.7827236508511409, "grad_norm": 1.0143867501486805, "learning_rate": 2.3756173963157825e-06, "loss": 0.7399, "step": 21611 }, { "epoch": 0.7827598696124592, "grad_norm": 1.211140734711883, "learning_rate": 2.3748584024414147e-06, "loss": 0.6365, "step": 21612 }, { "epoch": 0.7827960883737776, "grad_norm": 0.9306877610800836, "learning_rate": 2.3740995134956914e-06, "loss": 0.6815, "step": 21613 }, { "epoch": 0.782832307135096, "grad_norm": 1.40048155562545, "learning_rate": 2.373340729489062e-06, "loss": 0.6958, "step": 21614 }, { "epoch": 0.7828685258964143, "grad_norm": 1.4346452067634279, "learning_rate": 2.372582050431965e-06, "loss": 0.675, "step": 21615 }, { "epoch": 0.7829047446577327, "grad_norm": 1.3354088813784915, "learning_rate": 2.3718234763348426e-06, "loss": 0.7081, "step": 21616 }, { "epoch": 0.7829409634190511, "grad_norm": 1.3851213200140469, "learning_rate": 2.3710650072081277e-06, "loss": 0.6392, "step": 21617 }, { "epoch": 0.7829771821803694, "grad_norm": 1.3707934360425575, "learning_rate": 2.3703066430622654e-06, "loss": 0.6347, "step": 21618 }, { "epoch": 0.7830134009416878, "grad_norm": 1.3200792143205915, "learning_rate": 2.369548383907687e-06, "loss": 0.6642, "step": 21619 }, { "epoch": 0.7830496197030061, "grad_norm": 1.3214943261861418, "learning_rate": 2.368790229754828e-06, "loss": 0.696, "step": 21620 }, { "epoch": 0.7830858384643246, "grad_norm": 1.3686036037622105, "learning_rate": 2.3680321806141182e-06, "loss": 0.6516, "step": 21621 }, { "epoch": 0.7831220572256429, "grad_norm": 1.1321440497758288, "learning_rate": 2.3672742364959944e-06, "loss": 0.6562, "step": 21622 }, { "epoch": 0.7831582759869612, "grad_norm": 1.3705437581432904, "learning_rate": 2.366516397410884e-06, "loss": 0.6647, "step": 21623 }, { "epoch": 0.7831944947482796, "grad_norm": 1.382257459935806, "learning_rate": 2.365758663369212e-06, "loss": 0.6955, "step": 21624 }, { "epoch": 0.7832307135095979, "grad_norm": 1.0445987771202183, "learning_rate": 2.3650010343814113e-06, "loss": 0.6498, "step": 21625 }, { "epoch": 0.7832669322709164, "grad_norm": 1.0962864662532488, "learning_rate": 2.3642435104579054e-06, "loss": 0.7019, "step": 21626 }, { "epoch": 0.7833031510322347, "grad_norm": 1.2146148952883278, "learning_rate": 2.3634860916091183e-06, "loss": 0.6571, "step": 21627 }, { "epoch": 0.783339369793553, "grad_norm": 1.0396569450277302, "learning_rate": 2.36272877784547e-06, "loss": 0.7478, "step": 21628 }, { "epoch": 0.7833755885548714, "grad_norm": 1.5229043492259606, "learning_rate": 2.361971569177387e-06, "loss": 0.6311, "step": 21629 }, { "epoch": 0.7834118073161898, "grad_norm": 1.3954313095724487, "learning_rate": 2.361214465615287e-06, "loss": 0.6792, "step": 21630 }, { "epoch": 0.7834480260775082, "grad_norm": 1.2801674174065312, "learning_rate": 2.3604574671695867e-06, "loss": 0.6449, "step": 21631 }, { "epoch": 0.7834842448388265, "grad_norm": 1.4598645392140783, "learning_rate": 2.359700573850703e-06, "loss": 0.7006, "step": 21632 }, { "epoch": 0.7835204636001448, "grad_norm": 2.3827781534064636, "learning_rate": 2.3589437856690543e-06, "loss": 0.7572, "step": 21633 }, { "epoch": 0.7835566823614633, "grad_norm": 0.9649729084813797, "learning_rate": 2.3581871026350546e-06, "loss": 0.7276, "step": 21634 }, { "epoch": 0.7835929011227816, "grad_norm": 1.383028099122451, "learning_rate": 2.357430524759111e-06, "loss": 0.6639, "step": 21635 }, { "epoch": 0.7836291198841, "grad_norm": 1.3085740643059662, "learning_rate": 2.3566740520516416e-06, "loss": 0.5813, "step": 21636 }, { "epoch": 0.7836653386454183, "grad_norm": 1.1418234328691284, "learning_rate": 2.3559176845230534e-06, "loss": 0.7201, "step": 21637 }, { "epoch": 0.7837015574067366, "grad_norm": 1.6041115874169154, "learning_rate": 2.355161422183756e-06, "loss": 0.7412, "step": 21638 }, { "epoch": 0.7837377761680551, "grad_norm": 1.4580227594957245, "learning_rate": 2.3544052650441505e-06, "loss": 0.7031, "step": 21639 }, { "epoch": 0.7837739949293734, "grad_norm": 1.3399113771482098, "learning_rate": 2.353649213114649e-06, "loss": 0.643, "step": 21640 }, { "epoch": 0.7838102136906918, "grad_norm": 1.371230803945827, "learning_rate": 2.3528932664056537e-06, "loss": 0.6497, "step": 21641 }, { "epoch": 0.7838464324520101, "grad_norm": 1.4010969865139824, "learning_rate": 2.3521374249275664e-06, "loss": 0.6723, "step": 21642 }, { "epoch": 0.7838826512133285, "grad_norm": 1.5423515884255112, "learning_rate": 2.3513816886907847e-06, "loss": 0.6759, "step": 21643 }, { "epoch": 0.7839188699746469, "grad_norm": 0.9231378939888475, "learning_rate": 2.3506260577057148e-06, "loss": 0.6483, "step": 21644 }, { "epoch": 0.7839550887359652, "grad_norm": 1.4360392400844995, "learning_rate": 2.349870531982753e-06, "loss": 0.7375, "step": 21645 }, { "epoch": 0.7839913074972836, "grad_norm": 1.0484954556251858, "learning_rate": 2.34911511153229e-06, "loss": 0.6824, "step": 21646 }, { "epoch": 0.784027526258602, "grad_norm": 1.0555517068658287, "learning_rate": 2.34835979636473e-06, "loss": 0.6978, "step": 21647 }, { "epoch": 0.7840637450199203, "grad_norm": 1.3006864304106698, "learning_rate": 2.3476045864904605e-06, "loss": 0.6676, "step": 21648 }, { "epoch": 0.7840999637812387, "grad_norm": 1.423138237443683, "learning_rate": 2.3468494819198773e-06, "loss": 0.5896, "step": 21649 }, { "epoch": 0.784136182542557, "grad_norm": 1.4724317804564933, "learning_rate": 2.346094482663367e-06, "loss": 0.7015, "step": 21650 }, { "epoch": 0.7841724013038754, "grad_norm": 1.5537318882227993, "learning_rate": 2.3453395887313247e-06, "loss": 0.6706, "step": 21651 }, { "epoch": 0.7842086200651938, "grad_norm": 1.3257106953866225, "learning_rate": 2.344584800134135e-06, "loss": 0.6068, "step": 21652 }, { "epoch": 0.7842448388265121, "grad_norm": 1.4864725939985084, "learning_rate": 2.3438301168821864e-06, "loss": 0.6567, "step": 21653 }, { "epoch": 0.7842810575878305, "grad_norm": 1.4023486186450398, "learning_rate": 2.3430755389858596e-06, "loss": 0.7116, "step": 21654 }, { "epoch": 0.7843172763491488, "grad_norm": 1.2295523773440413, "learning_rate": 2.3423210664555428e-06, "loss": 0.626, "step": 21655 }, { "epoch": 0.7843534951104673, "grad_norm": 1.490224307014983, "learning_rate": 2.3415666993016184e-06, "loss": 0.7359, "step": 21656 }, { "epoch": 0.7843897138717856, "grad_norm": 1.3499378544358236, "learning_rate": 2.340812437534462e-06, "loss": 0.7397, "step": 21657 }, { "epoch": 0.7844259326331039, "grad_norm": 1.4177652671453471, "learning_rate": 2.340058281164459e-06, "loss": 0.7491, "step": 21658 }, { "epoch": 0.7844621513944223, "grad_norm": 1.315033130017645, "learning_rate": 2.339304230201984e-06, "loss": 0.6891, "step": 21659 }, { "epoch": 0.7844983701557406, "grad_norm": 1.3765048656890317, "learning_rate": 2.338550284657416e-06, "loss": 0.6875, "step": 21660 }, { "epoch": 0.7845345889170591, "grad_norm": 1.4045383560099785, "learning_rate": 2.3377964445411237e-06, "loss": 0.6709, "step": 21661 }, { "epoch": 0.7845708076783774, "grad_norm": 1.4410846204053862, "learning_rate": 2.3370427098634883e-06, "loss": 0.7533, "step": 21662 }, { "epoch": 0.7846070264396957, "grad_norm": 1.0587625373652418, "learning_rate": 2.336289080634878e-06, "loss": 0.7274, "step": 21663 }, { "epoch": 0.7846432452010141, "grad_norm": 0.8690430565313815, "learning_rate": 2.335535556865661e-06, "loss": 0.6871, "step": 21664 }, { "epoch": 0.7846794639623325, "grad_norm": 1.2603738074480284, "learning_rate": 2.334782138566213e-06, "loss": 0.6004, "step": 21665 }, { "epoch": 0.7847156827236509, "grad_norm": 1.1714588133556676, "learning_rate": 2.334028825746897e-06, "loss": 0.6574, "step": 21666 }, { "epoch": 0.7847519014849692, "grad_norm": 1.2963052943174567, "learning_rate": 2.3332756184180815e-06, "loss": 0.678, "step": 21667 }, { "epoch": 0.7847881202462875, "grad_norm": 1.3359787415948199, "learning_rate": 2.3325225165901266e-06, "loss": 0.6795, "step": 21668 }, { "epoch": 0.784824339007606, "grad_norm": 1.0204024166912913, "learning_rate": 2.3317695202734013e-06, "loss": 0.623, "step": 21669 }, { "epoch": 0.7848605577689243, "grad_norm": 3.0331073119947254, "learning_rate": 2.3310166294782664e-06, "loss": 0.7202, "step": 21670 }, { "epoch": 0.7848967765302427, "grad_norm": 1.3511661212905612, "learning_rate": 2.3302638442150804e-06, "loss": 0.695, "step": 21671 }, { "epoch": 0.784932995291561, "grad_norm": 1.0217428464465592, "learning_rate": 2.3295111644942003e-06, "loss": 0.7113, "step": 21672 }, { "epoch": 0.7849692140528793, "grad_norm": 1.0994642520930333, "learning_rate": 2.3287585903259903e-06, "loss": 0.6262, "step": 21673 }, { "epoch": 0.7850054328141978, "grad_norm": 1.5223429758351246, "learning_rate": 2.328006121720802e-06, "loss": 0.7086, "step": 21674 }, { "epoch": 0.7850416515755161, "grad_norm": 1.3179955905991956, "learning_rate": 2.3272537586889875e-06, "loss": 0.6247, "step": 21675 }, { "epoch": 0.7850778703368345, "grad_norm": 0.9499323840478245, "learning_rate": 2.3265015012409063e-06, "loss": 0.6244, "step": 21676 }, { "epoch": 0.7851140890981528, "grad_norm": 1.3297468520701157, "learning_rate": 2.3257493493869066e-06, "loss": 0.7157, "step": 21677 }, { "epoch": 0.7851503078594712, "grad_norm": 1.0799352573911667, "learning_rate": 2.3249973031373394e-06, "loss": 0.7236, "step": 21678 }, { "epoch": 0.7851865266207896, "grad_norm": 1.1462198154064824, "learning_rate": 2.3242453625025507e-06, "loss": 0.6604, "step": 21679 }, { "epoch": 0.7852227453821079, "grad_norm": 0.9244313310611213, "learning_rate": 2.3234935274928925e-06, "loss": 0.637, "step": 21680 }, { "epoch": 0.7852589641434263, "grad_norm": 1.649923773200504, "learning_rate": 2.3227417981187083e-06, "loss": 0.698, "step": 21681 }, { "epoch": 0.7852951829047446, "grad_norm": 1.3944685525183413, "learning_rate": 2.321990174390344e-06, "loss": 0.6721, "step": 21682 }, { "epoch": 0.785331401666063, "grad_norm": 1.1748764938898388, "learning_rate": 2.321238656318138e-06, "loss": 0.6168, "step": 21683 }, { "epoch": 0.7853676204273814, "grad_norm": 1.3566982645871364, "learning_rate": 2.320487243912437e-06, "loss": 0.6909, "step": 21684 }, { "epoch": 0.7854038391886997, "grad_norm": 1.4537366706418697, "learning_rate": 2.3197359371835802e-06, "loss": 0.6883, "step": 21685 }, { "epoch": 0.7854400579500181, "grad_norm": 0.9426305122612109, "learning_rate": 2.318984736141903e-06, "loss": 0.7087, "step": 21686 }, { "epoch": 0.7854762767113365, "grad_norm": 1.3750564891417139, "learning_rate": 2.318233640797747e-06, "loss": 0.7175, "step": 21687 }, { "epoch": 0.7855124954726548, "grad_norm": 1.3765275852265695, "learning_rate": 2.3174826511614467e-06, "loss": 0.7218, "step": 21688 }, { "epoch": 0.7855487142339732, "grad_norm": 1.083324036210618, "learning_rate": 2.3167317672433355e-06, "loss": 0.7129, "step": 21689 }, { "epoch": 0.7855849329952915, "grad_norm": 1.220570010747503, "learning_rate": 2.315980989053742e-06, "loss": 0.5782, "step": 21690 }, { "epoch": 0.78562115175661, "grad_norm": 1.4578603745334193, "learning_rate": 2.3152303166030067e-06, "loss": 0.6605, "step": 21691 }, { "epoch": 0.7856573705179283, "grad_norm": 1.181753769494011, "learning_rate": 2.3144797499014537e-06, "loss": 0.6368, "step": 21692 }, { "epoch": 0.7856935892792466, "grad_norm": 1.3566971168212372, "learning_rate": 2.313729288959412e-06, "loss": 0.6381, "step": 21693 }, { "epoch": 0.785729808040565, "grad_norm": 1.2680824478539003, "learning_rate": 2.3129789337872064e-06, "loss": 0.7126, "step": 21694 }, { "epoch": 0.7857660268018833, "grad_norm": 1.0377824636134776, "learning_rate": 2.312228684395168e-06, "loss": 0.711, "step": 21695 }, { "epoch": 0.7858022455632018, "grad_norm": 1.4827459744985043, "learning_rate": 2.311478540793619e-06, "loss": 0.6836, "step": 21696 }, { "epoch": 0.7858384643245201, "grad_norm": 1.450531063323413, "learning_rate": 2.310728502992877e-06, "loss": 0.6052, "step": 21697 }, { "epoch": 0.7858746830858384, "grad_norm": 1.2799877528889003, "learning_rate": 2.3099785710032706e-06, "loss": 0.6551, "step": 21698 }, { "epoch": 0.7859109018471568, "grad_norm": 1.3218637775672981, "learning_rate": 2.309228744835116e-06, "loss": 0.6592, "step": 21699 }, { "epoch": 0.7859471206084752, "grad_norm": 1.143093715799386, "learning_rate": 2.308479024498731e-06, "loss": 0.6919, "step": 21700 }, { "epoch": 0.7859833393697936, "grad_norm": 1.4281026584392718, "learning_rate": 2.3077294100044314e-06, "loss": 0.6885, "step": 21701 }, { "epoch": 0.7860195581311119, "grad_norm": 1.2091254296851186, "learning_rate": 2.306979901362535e-06, "loss": 0.6305, "step": 21702 }, { "epoch": 0.7860557768924302, "grad_norm": 1.3546906396693064, "learning_rate": 2.306230498583356e-06, "loss": 0.7059, "step": 21703 }, { "epoch": 0.7860919956537487, "grad_norm": 1.4200464689851613, "learning_rate": 2.3054812016772067e-06, "loss": 0.6865, "step": 21704 }, { "epoch": 0.786128214415067, "grad_norm": 1.4707823829520885, "learning_rate": 2.3047320106543922e-06, "loss": 0.7069, "step": 21705 }, { "epoch": 0.7861644331763854, "grad_norm": 1.2990996270613104, "learning_rate": 2.3039829255252312e-06, "loss": 0.6936, "step": 21706 }, { "epoch": 0.7862006519377037, "grad_norm": 1.3570026354059606, "learning_rate": 2.303233946300026e-06, "loss": 0.7328, "step": 21707 }, { "epoch": 0.786236870699022, "grad_norm": 1.108780373767641, "learning_rate": 2.3024850729890834e-06, "loss": 0.6451, "step": 21708 }, { "epoch": 0.7862730894603405, "grad_norm": 1.271106934962619, "learning_rate": 2.301736305602712e-06, "loss": 0.6594, "step": 21709 }, { "epoch": 0.7863093082216588, "grad_norm": 1.2178245960432168, "learning_rate": 2.3009876441512137e-06, "loss": 0.6102, "step": 21710 }, { "epoch": 0.7863455269829772, "grad_norm": 1.4880269251311407, "learning_rate": 2.30023908864489e-06, "loss": 0.7102, "step": 21711 }, { "epoch": 0.7863817457442955, "grad_norm": 1.2049783618227508, "learning_rate": 2.2994906390940395e-06, "loss": 0.709, "step": 21712 }, { "epoch": 0.7864179645056139, "grad_norm": 1.34470758306979, "learning_rate": 2.2987422955089668e-06, "loss": 0.7349, "step": 21713 }, { "epoch": 0.7864541832669323, "grad_norm": 1.2966328037127435, "learning_rate": 2.297994057899967e-06, "loss": 0.6666, "step": 21714 }, { "epoch": 0.7864904020282506, "grad_norm": 1.083288835292476, "learning_rate": 2.297245926277336e-06, "loss": 0.6229, "step": 21715 }, { "epoch": 0.786526620789569, "grad_norm": 1.0334428601944943, "learning_rate": 2.2964979006513666e-06, "loss": 0.6755, "step": 21716 }, { "epoch": 0.7865628395508873, "grad_norm": 1.3369293727513594, "learning_rate": 2.295749981032358e-06, "loss": 0.6194, "step": 21717 }, { "epoch": 0.7865990583122057, "grad_norm": 1.2842439584640073, "learning_rate": 2.2950021674306e-06, "loss": 0.6611, "step": 21718 }, { "epoch": 0.7866352770735241, "grad_norm": 1.1046512885583661, "learning_rate": 2.2942544598563785e-06, "loss": 0.6693, "step": 21719 }, { "epoch": 0.7866714958348424, "grad_norm": 1.114506210285592, "learning_rate": 2.2935068583199902e-06, "loss": 0.6946, "step": 21720 }, { "epoch": 0.7867077145961608, "grad_norm": 1.710163343030598, "learning_rate": 2.292759362831718e-06, "loss": 0.6463, "step": 21721 }, { "epoch": 0.7867439333574792, "grad_norm": 1.4881490964112194, "learning_rate": 2.2920119734018497e-06, "loss": 0.6634, "step": 21722 }, { "epoch": 0.7867801521187975, "grad_norm": 1.3744084581984823, "learning_rate": 2.2912646900406665e-06, "loss": 0.7267, "step": 21723 }, { "epoch": 0.7868163708801159, "grad_norm": 1.0586156359504875, "learning_rate": 2.2905175127584578e-06, "loss": 0.698, "step": 21724 }, { "epoch": 0.7868525896414342, "grad_norm": 1.1279715670617696, "learning_rate": 2.289770441565502e-06, "loss": 0.687, "step": 21725 }, { "epoch": 0.7868888084027527, "grad_norm": 1.1335859014521534, "learning_rate": 2.2890234764720778e-06, "loss": 0.669, "step": 21726 }, { "epoch": 0.786925027164071, "grad_norm": 1.43540336903663, "learning_rate": 2.2882766174884673e-06, "loss": 0.656, "step": 21727 }, { "epoch": 0.7869612459253893, "grad_norm": 1.289368956435431, "learning_rate": 2.2875298646249477e-06, "loss": 0.6893, "step": 21728 }, { "epoch": 0.7869974646867077, "grad_norm": 1.4074848693769968, "learning_rate": 2.286783217891794e-06, "loss": 0.662, "step": 21729 }, { "epoch": 0.787033683448026, "grad_norm": 0.9725220834081063, "learning_rate": 2.286036677299277e-06, "loss": 0.6484, "step": 21730 }, { "epoch": 0.7870699022093445, "grad_norm": 1.5378794044005846, "learning_rate": 2.2852902428576763e-06, "loss": 0.6656, "step": 21731 }, { "epoch": 0.7871061209706628, "grad_norm": 1.3016698804720073, "learning_rate": 2.2845439145772607e-06, "loss": 0.6279, "step": 21732 }, { "epoch": 0.7871423397319811, "grad_norm": 1.4749905876641287, "learning_rate": 2.2837976924683014e-06, "loss": 0.7345, "step": 21733 }, { "epoch": 0.7871785584932995, "grad_norm": 1.3524264319722497, "learning_rate": 2.283051576541062e-06, "loss": 0.6858, "step": 21734 }, { "epoch": 0.7872147772546179, "grad_norm": 1.2551180244310753, "learning_rate": 2.2823055668058157e-06, "loss": 0.6259, "step": 21735 }, { "epoch": 0.7872509960159363, "grad_norm": 0.9240640874536056, "learning_rate": 2.281559663272828e-06, "loss": 0.6889, "step": 21736 }, { "epoch": 0.7872872147772546, "grad_norm": 1.245629611317817, "learning_rate": 2.2808138659523573e-06, "loss": 0.6218, "step": 21737 }, { "epoch": 0.7873234335385729, "grad_norm": 1.379795930237048, "learning_rate": 2.280068174854674e-06, "loss": 0.6975, "step": 21738 }, { "epoch": 0.7873596522998914, "grad_norm": 1.3908365951247086, "learning_rate": 2.279322589990035e-06, "loss": 0.6287, "step": 21739 }, { "epoch": 0.7873958710612097, "grad_norm": 1.3178403934890024, "learning_rate": 2.2785771113687018e-06, "loss": 0.6531, "step": 21740 }, { "epoch": 0.7874320898225281, "grad_norm": 1.4338981489746836, "learning_rate": 2.2778317390009284e-06, "loss": 0.7219, "step": 21741 }, { "epoch": 0.7874683085838464, "grad_norm": 1.2609024813934726, "learning_rate": 2.2770864728969798e-06, "loss": 0.7155, "step": 21742 }, { "epoch": 0.7875045273451649, "grad_norm": 1.3787680163540044, "learning_rate": 2.2763413130671066e-06, "loss": 0.7, "step": 21743 }, { "epoch": 0.7875407461064832, "grad_norm": 1.4155706962849819, "learning_rate": 2.2755962595215643e-06, "loss": 0.6975, "step": 21744 }, { "epoch": 0.7875769648678015, "grad_norm": 0.9653317665715739, "learning_rate": 2.2748513122706008e-06, "loss": 0.6355, "step": 21745 }, { "epoch": 0.7876131836291199, "grad_norm": 0.9160299398522765, "learning_rate": 2.2741064713244744e-06, "loss": 0.6961, "step": 21746 }, { "epoch": 0.7876494023904382, "grad_norm": 1.064242958206753, "learning_rate": 2.2733617366934314e-06, "loss": 0.6944, "step": 21747 }, { "epoch": 0.7876856211517567, "grad_norm": 1.3422613936257037, "learning_rate": 2.272617108387718e-06, "loss": 0.647, "step": 21748 }, { "epoch": 0.787721839913075, "grad_norm": 1.3930501392055201, "learning_rate": 2.2718725864175852e-06, "loss": 0.6682, "step": 21749 }, { "epoch": 0.7877580586743933, "grad_norm": 1.278611284832906, "learning_rate": 2.2711281707932752e-06, "loss": 0.6502, "step": 21750 }, { "epoch": 0.7877942774357117, "grad_norm": 1.4187382082879747, "learning_rate": 2.270383861525034e-06, "loss": 0.7291, "step": 21751 }, { "epoch": 0.78783049619703, "grad_norm": 1.3544670149396854, "learning_rate": 2.2696396586230994e-06, "loss": 0.6938, "step": 21752 }, { "epoch": 0.7878667149583485, "grad_norm": 1.2642784059663308, "learning_rate": 2.2688955620977194e-06, "loss": 0.652, "step": 21753 }, { "epoch": 0.7879029337196668, "grad_norm": 1.2470344503387274, "learning_rate": 2.2681515719591284e-06, "loss": 0.6663, "step": 21754 }, { "epoch": 0.7879391524809851, "grad_norm": 0.9919084177627728, "learning_rate": 2.267407688217567e-06, "loss": 0.7414, "step": 21755 }, { "epoch": 0.7879753712423035, "grad_norm": 1.4493755309752478, "learning_rate": 2.2666639108832656e-06, "loss": 0.7209, "step": 21756 }, { "epoch": 0.7880115900036219, "grad_norm": 1.2332932857050347, "learning_rate": 2.265920239966468e-06, "loss": 0.6601, "step": 21757 }, { "epoch": 0.7880478087649403, "grad_norm": 1.2282246645376962, "learning_rate": 2.265176675477403e-06, "loss": 0.6418, "step": 21758 }, { "epoch": 0.7880840275262586, "grad_norm": 1.263192218445839, "learning_rate": 2.2644332174263006e-06, "loss": 0.5522, "step": 21759 }, { "epoch": 0.7881202462875769, "grad_norm": 1.3069597654238374, "learning_rate": 2.2636898658233973e-06, "loss": 0.6507, "step": 21760 }, { "epoch": 0.7881564650488954, "grad_norm": 0.9986404967879715, "learning_rate": 2.26294662067892e-06, "loss": 0.6499, "step": 21761 }, { "epoch": 0.7881926838102137, "grad_norm": 1.1354500371957592, "learning_rate": 2.2622034820030936e-06, "loss": 0.7199, "step": 21762 }, { "epoch": 0.7882289025715321, "grad_norm": 1.3226253521685074, "learning_rate": 2.2614604498061453e-06, "loss": 0.6516, "step": 21763 }, { "epoch": 0.7882651213328504, "grad_norm": 1.3061112955363594, "learning_rate": 2.2607175240983027e-06, "loss": 0.617, "step": 21764 }, { "epoch": 0.7883013400941687, "grad_norm": 1.3812931693244117, "learning_rate": 2.259974704889788e-06, "loss": 0.7276, "step": 21765 }, { "epoch": 0.7883375588554872, "grad_norm": 1.4483431351990612, "learning_rate": 2.259231992190821e-06, "loss": 0.7086, "step": 21766 }, { "epoch": 0.7883737776168055, "grad_norm": 1.4177598089962815, "learning_rate": 2.258489386011622e-06, "loss": 0.6795, "step": 21767 }, { "epoch": 0.7884099963781239, "grad_norm": 1.352846704881357, "learning_rate": 2.257746886362413e-06, "loss": 0.7151, "step": 21768 }, { "epoch": 0.7884462151394422, "grad_norm": 1.2290555604393976, "learning_rate": 2.257004493253411e-06, "loss": 0.6264, "step": 21769 }, { "epoch": 0.7884824339007606, "grad_norm": 1.448260689821076, "learning_rate": 2.256262206694827e-06, "loss": 0.6519, "step": 21770 }, { "epoch": 0.788518652662079, "grad_norm": 1.3315976988088916, "learning_rate": 2.2555200266968822e-06, "loss": 0.7107, "step": 21771 }, { "epoch": 0.7885548714233973, "grad_norm": 1.3452629371533893, "learning_rate": 2.2547779532697876e-06, "loss": 0.638, "step": 21772 }, { "epoch": 0.7885910901847157, "grad_norm": 1.4325882350333763, "learning_rate": 2.254035986423754e-06, "loss": 0.6258, "step": 21773 }, { "epoch": 0.7886273089460341, "grad_norm": 1.0536260520950609, "learning_rate": 2.2532941261689866e-06, "loss": 0.6573, "step": 21774 }, { "epoch": 0.7886635277073524, "grad_norm": 1.1437956177901545, "learning_rate": 2.252552372515704e-06, "loss": 0.7266, "step": 21775 }, { "epoch": 0.7886997464686708, "grad_norm": 1.3586757666140723, "learning_rate": 2.2518107254741072e-06, "loss": 0.668, "step": 21776 }, { "epoch": 0.7887359652299891, "grad_norm": 1.3627080204073025, "learning_rate": 2.251069185054403e-06, "loss": 0.7018, "step": 21777 }, { "epoch": 0.7887721839913076, "grad_norm": 1.3475716826953739, "learning_rate": 2.250327751266792e-06, "loss": 0.6539, "step": 21778 }, { "epoch": 0.7888084027526259, "grad_norm": 0.9511011291827772, "learning_rate": 2.2495864241214848e-06, "loss": 0.7162, "step": 21779 }, { "epoch": 0.7888446215139442, "grad_norm": 1.45883449361634, "learning_rate": 2.2488452036286777e-06, "loss": 0.6152, "step": 21780 }, { "epoch": 0.7888808402752626, "grad_norm": 1.272555427526839, "learning_rate": 2.2481040897985696e-06, "loss": 0.6208, "step": 21781 }, { "epoch": 0.7889170590365809, "grad_norm": 0.9823737895831371, "learning_rate": 2.247363082641362e-06, "loss": 0.7003, "step": 21782 }, { "epoch": 0.7889532777978994, "grad_norm": 1.0406477194294037, "learning_rate": 2.2466221821672517e-06, "loss": 0.679, "step": 21783 }, { "epoch": 0.7889894965592177, "grad_norm": 1.3081466444767815, "learning_rate": 2.2458813883864315e-06, "loss": 0.6255, "step": 21784 }, { "epoch": 0.789025715320536, "grad_norm": 1.391826253452985, "learning_rate": 2.2451407013090955e-06, "loss": 0.6997, "step": 21785 }, { "epoch": 0.7890619340818544, "grad_norm": 0.9691077981096708, "learning_rate": 2.244400120945439e-06, "loss": 0.6183, "step": 21786 }, { "epoch": 0.7890981528431728, "grad_norm": 1.3205514471537665, "learning_rate": 2.2436596473056527e-06, "loss": 0.6937, "step": 21787 }, { "epoch": 0.7891343716044912, "grad_norm": 1.398180960721301, "learning_rate": 2.2429192803999213e-06, "loss": 0.6551, "step": 21788 }, { "epoch": 0.7891705903658095, "grad_norm": 1.3786995654509213, "learning_rate": 2.24217902023844e-06, "loss": 0.7023, "step": 21789 }, { "epoch": 0.7892068091271278, "grad_norm": 1.2517392411731498, "learning_rate": 2.2414388668313912e-06, "loss": 0.6507, "step": 21790 }, { "epoch": 0.7892430278884462, "grad_norm": 1.352321497085634, "learning_rate": 2.240698820188961e-06, "loss": 0.6852, "step": 21791 }, { "epoch": 0.7892792466497646, "grad_norm": 1.525389004607567, "learning_rate": 2.23995888032133e-06, "loss": 0.7051, "step": 21792 }, { "epoch": 0.789315465411083, "grad_norm": 1.316152759052477, "learning_rate": 2.2392190472386865e-06, "loss": 0.7195, "step": 21793 }, { "epoch": 0.7893516841724013, "grad_norm": 1.301689911693354, "learning_rate": 2.238479320951209e-06, "loss": 0.6592, "step": 21794 }, { "epoch": 0.7893879029337196, "grad_norm": 1.3520566596457557, "learning_rate": 2.237739701469075e-06, "loss": 0.661, "step": 21795 }, { "epoch": 0.7894241216950381, "grad_norm": 1.3585843432782465, "learning_rate": 2.2370001888024606e-06, "loss": 0.6507, "step": 21796 }, { "epoch": 0.7894603404563564, "grad_norm": 1.7091038950951551, "learning_rate": 2.2362607829615466e-06, "loss": 0.6709, "step": 21797 }, { "epoch": 0.7894965592176748, "grad_norm": 1.3723901776288328, "learning_rate": 2.2355214839565074e-06, "loss": 0.6844, "step": 21798 }, { "epoch": 0.7895327779789931, "grad_norm": 1.4410361445437732, "learning_rate": 2.2347822917975116e-06, "loss": 0.6807, "step": 21799 }, { "epoch": 0.7895689967403114, "grad_norm": 1.3973381633661797, "learning_rate": 2.2340432064947383e-06, "loss": 0.6505, "step": 21800 }, { "epoch": 0.7896052155016299, "grad_norm": 1.304788539166316, "learning_rate": 2.233304228058353e-06, "loss": 0.6259, "step": 21801 }, { "epoch": 0.7896414342629482, "grad_norm": 1.4467107156892391, "learning_rate": 2.232565356498526e-06, "loss": 0.7331, "step": 21802 }, { "epoch": 0.7896776530242666, "grad_norm": 1.0123606435645058, "learning_rate": 2.2318265918254234e-06, "loss": 0.7566, "step": 21803 }, { "epoch": 0.7897138717855849, "grad_norm": 1.605053317717787, "learning_rate": 2.231087934049214e-06, "loss": 0.7021, "step": 21804 }, { "epoch": 0.7897500905469033, "grad_norm": 1.1189131090442772, "learning_rate": 2.2303493831800616e-06, "loss": 0.6887, "step": 21805 }, { "epoch": 0.7897863093082217, "grad_norm": 1.4402793223892003, "learning_rate": 2.2296109392281286e-06, "loss": 0.7044, "step": 21806 }, { "epoch": 0.78982252806954, "grad_norm": 1.400883764695325, "learning_rate": 2.2288726022035732e-06, "loss": 0.7069, "step": 21807 }, { "epoch": 0.7898587468308584, "grad_norm": 0.9362639553374852, "learning_rate": 2.228134372116563e-06, "loss": 0.6733, "step": 21808 }, { "epoch": 0.7898949655921768, "grad_norm": 1.192412236231024, "learning_rate": 2.227396248977253e-06, "loss": 0.6818, "step": 21809 }, { "epoch": 0.7899311843534951, "grad_norm": 1.366212116810833, "learning_rate": 2.2266582327957962e-06, "loss": 0.7025, "step": 21810 }, { "epoch": 0.7899674031148135, "grad_norm": 1.2832178517081536, "learning_rate": 2.225920323582359e-06, "loss": 0.6595, "step": 21811 }, { "epoch": 0.7900036218761318, "grad_norm": 1.242754260359899, "learning_rate": 2.2251825213470856e-06, "loss": 0.6967, "step": 21812 }, { "epoch": 0.7900398406374503, "grad_norm": 1.3120685887336057, "learning_rate": 2.2244448261001295e-06, "loss": 0.7086, "step": 21813 }, { "epoch": 0.7900760593987686, "grad_norm": 1.3116954241911467, "learning_rate": 2.223707237851648e-06, "loss": 0.7032, "step": 21814 }, { "epoch": 0.7901122781600869, "grad_norm": 1.3624109869335475, "learning_rate": 2.222969756611787e-06, "loss": 0.6545, "step": 21815 }, { "epoch": 0.7901484969214053, "grad_norm": 1.0847039053453738, "learning_rate": 2.2222323823906965e-06, "loss": 0.662, "step": 21816 }, { "epoch": 0.7901847156827236, "grad_norm": 1.3350516806780592, "learning_rate": 2.22149511519852e-06, "loss": 0.6484, "step": 21817 }, { "epoch": 0.7902209344440421, "grad_norm": 1.3566180833898378, "learning_rate": 2.2207579550454074e-06, "loss": 0.6744, "step": 21818 }, { "epoch": 0.7902571532053604, "grad_norm": 1.3165043207933638, "learning_rate": 2.220020901941501e-06, "loss": 0.6723, "step": 21819 }, { "epoch": 0.7902933719666787, "grad_norm": 1.3299555427931928, "learning_rate": 2.2192839558969406e-06, "loss": 0.6779, "step": 21820 }, { "epoch": 0.7903295907279971, "grad_norm": 1.2193983085270395, "learning_rate": 2.218547116921872e-06, "loss": 0.6175, "step": 21821 }, { "epoch": 0.7903658094893155, "grad_norm": 1.2901617969284205, "learning_rate": 2.2178103850264334e-06, "loss": 0.6407, "step": 21822 }, { "epoch": 0.7904020282506339, "grad_norm": 1.1028492443079403, "learning_rate": 2.2170737602207605e-06, "loss": 0.6656, "step": 21823 }, { "epoch": 0.7904382470119522, "grad_norm": 1.534079176283131, "learning_rate": 2.2163372425149897e-06, "loss": 0.6563, "step": 21824 }, { "epoch": 0.7904744657732705, "grad_norm": 1.5245868430988183, "learning_rate": 2.21560083191926e-06, "loss": 0.6369, "step": 21825 }, { "epoch": 0.790510684534589, "grad_norm": 1.0790037886526027, "learning_rate": 2.214864528443703e-06, "loss": 0.6828, "step": 21826 }, { "epoch": 0.7905469032959073, "grad_norm": 1.092395440964228, "learning_rate": 2.2141283320984506e-06, "loss": 0.6978, "step": 21827 }, { "epoch": 0.7905831220572257, "grad_norm": 0.9655041433124832, "learning_rate": 2.2133922428936306e-06, "loss": 0.6218, "step": 21828 }, { "epoch": 0.790619340818544, "grad_norm": 1.4669795157533196, "learning_rate": 2.2126562608393785e-06, "loss": 0.6983, "step": 21829 }, { "epoch": 0.7906555595798623, "grad_norm": 1.588661893349424, "learning_rate": 2.2119203859458195e-06, "loss": 0.7885, "step": 21830 }, { "epoch": 0.7906917783411808, "grad_norm": 1.0998046540257091, "learning_rate": 2.2111846182230754e-06, "loss": 0.6766, "step": 21831 }, { "epoch": 0.7907279971024991, "grad_norm": 1.3256649563991465, "learning_rate": 2.210448957681277e-06, "loss": 0.6852, "step": 21832 }, { "epoch": 0.7907642158638175, "grad_norm": 0.9804754072344222, "learning_rate": 2.2097134043305467e-06, "loss": 0.6678, "step": 21833 }, { "epoch": 0.7908004346251358, "grad_norm": 0.9940115285904068, "learning_rate": 2.2089779581810044e-06, "loss": 0.6699, "step": 21834 }, { "epoch": 0.7908366533864541, "grad_norm": 1.430258608184705, "learning_rate": 2.2082426192427685e-06, "loss": 0.7238, "step": 21835 }, { "epoch": 0.7908728721477726, "grad_norm": 1.4038285713432859, "learning_rate": 2.2075073875259645e-06, "loss": 0.6624, "step": 21836 }, { "epoch": 0.7909090909090909, "grad_norm": 1.3064397192715667, "learning_rate": 2.2067722630407042e-06, "loss": 0.6299, "step": 21837 }, { "epoch": 0.7909453096704093, "grad_norm": 1.3305292554796004, "learning_rate": 2.2060372457971067e-06, "loss": 0.6834, "step": 21838 }, { "epoch": 0.7909815284317276, "grad_norm": 1.2502701229054263, "learning_rate": 2.205302335805282e-06, "loss": 0.6555, "step": 21839 }, { "epoch": 0.791017747193046, "grad_norm": 1.2424438574908612, "learning_rate": 2.204567533075348e-06, "loss": 0.6592, "step": 21840 }, { "epoch": 0.7910539659543644, "grad_norm": 1.3680988013027593, "learning_rate": 2.2038328376174146e-06, "loss": 0.6048, "step": 21841 }, { "epoch": 0.7910901847156827, "grad_norm": 1.3755602535096172, "learning_rate": 2.2030982494415896e-06, "loss": 0.6987, "step": 21842 }, { "epoch": 0.7911264034770011, "grad_norm": 1.2888605421981456, "learning_rate": 2.2023637685579856e-06, "loss": 0.6657, "step": 21843 }, { "epoch": 0.7911626222383195, "grad_norm": 1.2980592251527916, "learning_rate": 2.2016293949767085e-06, "loss": 0.6499, "step": 21844 }, { "epoch": 0.7911988409996378, "grad_norm": 1.3685605749529242, "learning_rate": 2.200895128707863e-06, "loss": 0.6836, "step": 21845 }, { "epoch": 0.7912350597609562, "grad_norm": 1.2553404936831591, "learning_rate": 2.200160969761551e-06, "loss": 0.7349, "step": 21846 }, { "epoch": 0.7912712785222745, "grad_norm": 1.3966533785371766, "learning_rate": 2.19942691814788e-06, "loss": 0.6771, "step": 21847 }, { "epoch": 0.791307497283593, "grad_norm": 1.3583095611962086, "learning_rate": 2.198692973876949e-06, "loss": 0.6806, "step": 21848 }, { "epoch": 0.7913437160449113, "grad_norm": 1.410528984863565, "learning_rate": 2.197959136958857e-06, "loss": 0.6739, "step": 21849 }, { "epoch": 0.7913799348062296, "grad_norm": 1.0681485707121663, "learning_rate": 2.1972254074037004e-06, "loss": 0.6138, "step": 21850 }, { "epoch": 0.791416153567548, "grad_norm": 1.2825992548121568, "learning_rate": 2.1964917852215807e-06, "loss": 0.6136, "step": 21851 }, { "epoch": 0.7914523723288663, "grad_norm": 1.3417380929478628, "learning_rate": 2.195758270422591e-06, "loss": 0.6373, "step": 21852 }, { "epoch": 0.7914885910901848, "grad_norm": 1.5333936730809934, "learning_rate": 2.1950248630168223e-06, "loss": 0.6481, "step": 21853 }, { "epoch": 0.7915248098515031, "grad_norm": 1.0200269478783472, "learning_rate": 2.1942915630143723e-06, "loss": 0.6478, "step": 21854 }, { "epoch": 0.7915610286128214, "grad_norm": 1.0875080988194634, "learning_rate": 2.1935583704253283e-06, "loss": 0.6862, "step": 21855 }, { "epoch": 0.7915972473741398, "grad_norm": 1.3166142024933236, "learning_rate": 2.192825285259782e-06, "loss": 0.6305, "step": 21856 }, { "epoch": 0.7916334661354582, "grad_norm": 1.394495192897758, "learning_rate": 2.1920923075278153e-06, "loss": 0.7217, "step": 21857 }, { "epoch": 0.7916696848967766, "grad_norm": 1.3278713327673832, "learning_rate": 2.191359437239523e-06, "loss": 0.7024, "step": 21858 }, { "epoch": 0.7917059036580949, "grad_norm": 1.2977049337874231, "learning_rate": 2.1906266744049853e-06, "loss": 0.5353, "step": 21859 }, { "epoch": 0.7917421224194132, "grad_norm": 0.989652139549679, "learning_rate": 2.1898940190342875e-06, "loss": 0.6698, "step": 21860 }, { "epoch": 0.7917783411807316, "grad_norm": 1.4579683044252214, "learning_rate": 2.1891614711375076e-06, "loss": 0.7273, "step": 21861 }, { "epoch": 0.79181455994205, "grad_norm": 1.34242886431049, "learning_rate": 2.1884290307247313e-06, "loss": 0.6677, "step": 21862 }, { "epoch": 0.7918507787033684, "grad_norm": 1.4289468125978106, "learning_rate": 2.1876966978060353e-06, "loss": 0.7602, "step": 21863 }, { "epoch": 0.7918869974646867, "grad_norm": 1.559741066391355, "learning_rate": 2.186964472391495e-06, "loss": 0.6545, "step": 21864 }, { "epoch": 0.791923216226005, "grad_norm": 1.3562381028607216, "learning_rate": 2.186232354491191e-06, "loss": 0.6508, "step": 21865 }, { "epoch": 0.7919594349873235, "grad_norm": 1.5343311023842352, "learning_rate": 2.1855003441151957e-06, "loss": 0.7342, "step": 21866 }, { "epoch": 0.7919956537486418, "grad_norm": 1.1122340363783543, "learning_rate": 2.1847684412735828e-06, "loss": 0.7085, "step": 21867 }, { "epoch": 0.7920318725099602, "grad_norm": 1.2956622159236795, "learning_rate": 2.1840366459764194e-06, "loss": 0.6799, "step": 21868 }, { "epoch": 0.7920680912712785, "grad_norm": 1.559123829879557, "learning_rate": 2.183304958233782e-06, "loss": 0.6856, "step": 21869 }, { "epoch": 0.7921043100325968, "grad_norm": 1.357984318606114, "learning_rate": 2.182573378055738e-06, "loss": 0.6115, "step": 21870 }, { "epoch": 0.7921405287939153, "grad_norm": 1.2498198297414633, "learning_rate": 2.181841905452352e-06, "loss": 0.63, "step": 21871 }, { "epoch": 0.7921767475552336, "grad_norm": 1.419747294643123, "learning_rate": 2.1811105404336896e-06, "loss": 0.6113, "step": 21872 }, { "epoch": 0.792212966316552, "grad_norm": 1.2077316694798745, "learning_rate": 2.180379283009818e-06, "loss": 0.6817, "step": 21873 }, { "epoch": 0.7922491850778703, "grad_norm": 1.3729047566060326, "learning_rate": 2.179648133190799e-06, "loss": 0.6885, "step": 21874 }, { "epoch": 0.7922854038391887, "grad_norm": 1.4497457456060916, "learning_rate": 2.178917090986691e-06, "loss": 0.6781, "step": 21875 }, { "epoch": 0.7923216226005071, "grad_norm": 1.351886671803352, "learning_rate": 2.178186156407558e-06, "loss": 0.6871, "step": 21876 }, { "epoch": 0.7923578413618254, "grad_norm": 1.3704092539662878, "learning_rate": 2.177455329463456e-06, "loss": 0.6927, "step": 21877 }, { "epoch": 0.7923940601231438, "grad_norm": 1.0340716956932778, "learning_rate": 2.1767246101644425e-06, "loss": 0.6844, "step": 21878 }, { "epoch": 0.7924302788844622, "grad_norm": 1.4443856321687971, "learning_rate": 2.1759939985205704e-06, "loss": 0.7096, "step": 21879 }, { "epoch": 0.7924664976457805, "grad_norm": 1.248692393332137, "learning_rate": 2.175263494541897e-06, "loss": 0.6487, "step": 21880 }, { "epoch": 0.7925027164070989, "grad_norm": 1.3686350563800243, "learning_rate": 2.174533098238475e-06, "loss": 0.6484, "step": 21881 }, { "epoch": 0.7925389351684172, "grad_norm": 1.022979604206058, "learning_rate": 2.17380280962035e-06, "loss": 0.7147, "step": 21882 }, { "epoch": 0.7925751539297357, "grad_norm": 1.1715503385949981, "learning_rate": 2.173072628697578e-06, "loss": 0.7325, "step": 21883 }, { "epoch": 0.792611372691054, "grad_norm": 1.261217244249129, "learning_rate": 2.1723425554802036e-06, "loss": 0.5948, "step": 21884 }, { "epoch": 0.7926475914523723, "grad_norm": 1.4155349204895897, "learning_rate": 2.1716125899782734e-06, "loss": 0.7236, "step": 21885 }, { "epoch": 0.7926838102136907, "grad_norm": 1.0203805819909284, "learning_rate": 2.1708827322018313e-06, "loss": 0.6953, "step": 21886 }, { "epoch": 0.792720028975009, "grad_norm": 1.2981817042294992, "learning_rate": 2.1701529821609235e-06, "loss": 0.593, "step": 21887 }, { "epoch": 0.7927562477363275, "grad_norm": 1.1062910325339765, "learning_rate": 2.1694233398655906e-06, "loss": 0.6763, "step": 21888 }, { "epoch": 0.7927924664976458, "grad_norm": 0.9783097096175336, "learning_rate": 2.1686938053258734e-06, "loss": 0.6088, "step": 21889 }, { "epoch": 0.7928286852589641, "grad_norm": 1.2990386871087467, "learning_rate": 2.1679643785518077e-06, "loss": 0.6956, "step": 21890 }, { "epoch": 0.7928649040202825, "grad_norm": 1.3068008552691799, "learning_rate": 2.1672350595534373e-06, "loss": 0.6287, "step": 21891 }, { "epoch": 0.7929011227816009, "grad_norm": 1.0741468767483897, "learning_rate": 2.166505848340794e-06, "loss": 0.7274, "step": 21892 }, { "epoch": 0.7929373415429193, "grad_norm": 1.3553920732407208, "learning_rate": 2.1657767449239118e-06, "loss": 0.6417, "step": 21893 }, { "epoch": 0.7929735603042376, "grad_norm": 1.2903622270130184, "learning_rate": 2.1650477493128274e-06, "loss": 0.7126, "step": 21894 }, { "epoch": 0.7930097790655559, "grad_norm": 1.379433963178268, "learning_rate": 2.1643188615175705e-06, "loss": 0.6435, "step": 21895 }, { "epoch": 0.7930459978268743, "grad_norm": 1.384011183486329, "learning_rate": 2.1635900815481716e-06, "loss": 0.6678, "step": 21896 }, { "epoch": 0.7930822165881927, "grad_norm": 1.390589392800891, "learning_rate": 2.162861409414655e-06, "loss": 0.6157, "step": 21897 }, { "epoch": 0.7931184353495111, "grad_norm": 1.124587799758364, "learning_rate": 2.162132845127055e-06, "loss": 0.6519, "step": 21898 }, { "epoch": 0.7931546541108294, "grad_norm": 1.4265986098516958, "learning_rate": 2.1614043886953937e-06, "loss": 0.6801, "step": 21899 }, { "epoch": 0.7931908728721477, "grad_norm": 1.4230088347974512, "learning_rate": 2.1606760401296954e-06, "loss": 0.6986, "step": 21900 }, { "epoch": 0.7932270916334662, "grad_norm": 1.07296087951431, "learning_rate": 2.159947799439981e-06, "loss": 0.7344, "step": 21901 }, { "epoch": 0.7932633103947845, "grad_norm": 1.356636576285391, "learning_rate": 2.1592196666362763e-06, "loss": 0.6879, "step": 21902 }, { "epoch": 0.7932995291561029, "grad_norm": 1.5880853010445395, "learning_rate": 2.1584916417285985e-06, "loss": 0.6767, "step": 21903 }, { "epoch": 0.7933357479174212, "grad_norm": 1.3748451153804442, "learning_rate": 2.1577637247269635e-06, "loss": 0.6378, "step": 21904 }, { "epoch": 0.7933719666787395, "grad_norm": 1.0028548731613849, "learning_rate": 2.157035915641392e-06, "loss": 0.6726, "step": 21905 }, { "epoch": 0.793408185440058, "grad_norm": 1.3483963047013925, "learning_rate": 2.156308214481899e-06, "loss": 0.7542, "step": 21906 }, { "epoch": 0.7934444042013763, "grad_norm": 0.9645010472824349, "learning_rate": 2.1555806212584963e-06, "loss": 0.6761, "step": 21907 }, { "epoch": 0.7934806229626947, "grad_norm": 1.3254435130415674, "learning_rate": 2.154853135981195e-06, "loss": 0.6731, "step": 21908 }, { "epoch": 0.793516841724013, "grad_norm": 1.3794598818802168, "learning_rate": 2.1541257586600097e-06, "loss": 0.6808, "step": 21909 }, { "epoch": 0.7935530604853314, "grad_norm": 1.4767521416018303, "learning_rate": 2.1533984893049487e-06, "loss": 0.6969, "step": 21910 }, { "epoch": 0.7935892792466498, "grad_norm": 1.514547701301274, "learning_rate": 2.152671327926019e-06, "loss": 0.7136, "step": 21911 }, { "epoch": 0.7936254980079681, "grad_norm": 1.3461949705035898, "learning_rate": 2.151944274533224e-06, "loss": 0.7137, "step": 21912 }, { "epoch": 0.7936617167692865, "grad_norm": 1.4888744605030748, "learning_rate": 2.1512173291365747e-06, "loss": 0.7374, "step": 21913 }, { "epoch": 0.7936979355306049, "grad_norm": 1.4874799578783053, "learning_rate": 2.150490491746071e-06, "loss": 0.7235, "step": 21914 }, { "epoch": 0.7937341542919232, "grad_norm": 1.4683972141731332, "learning_rate": 2.1497637623717126e-06, "loss": 0.6574, "step": 21915 }, { "epoch": 0.7937703730532416, "grad_norm": 1.3542863225952142, "learning_rate": 2.1490371410235056e-06, "loss": 0.6385, "step": 21916 }, { "epoch": 0.7938065918145599, "grad_norm": 1.4701135784838213, "learning_rate": 2.148310627711446e-06, "loss": 0.6944, "step": 21917 }, { "epoch": 0.7938428105758784, "grad_norm": 1.1626313603981808, "learning_rate": 2.1475842224455323e-06, "loss": 0.564, "step": 21918 }, { "epoch": 0.7938790293371967, "grad_norm": 1.3508207533538823, "learning_rate": 2.146857925235756e-06, "loss": 0.6056, "step": 21919 }, { "epoch": 0.793915248098515, "grad_norm": 1.0027096673234472, "learning_rate": 2.1461317360921174e-06, "loss": 0.6245, "step": 21920 }, { "epoch": 0.7939514668598334, "grad_norm": 1.3437645254348092, "learning_rate": 2.1454056550246073e-06, "loss": 0.6121, "step": 21921 }, { "epoch": 0.7939876856211517, "grad_norm": 1.0114582036209674, "learning_rate": 2.144679682043217e-06, "loss": 0.7517, "step": 21922 }, { "epoch": 0.7940239043824702, "grad_norm": 1.0797625085341982, "learning_rate": 2.1439538171579333e-06, "loss": 0.6747, "step": 21923 }, { "epoch": 0.7940601231437885, "grad_norm": 1.363899265077605, "learning_rate": 2.1432280603787514e-06, "loss": 0.7197, "step": 21924 }, { "epoch": 0.7940963419051068, "grad_norm": 1.0805508404643576, "learning_rate": 2.1425024117156545e-06, "loss": 0.6844, "step": 21925 }, { "epoch": 0.7941325606664252, "grad_norm": 1.7050825088374708, "learning_rate": 2.141776871178627e-06, "loss": 0.6065, "step": 21926 }, { "epoch": 0.7941687794277436, "grad_norm": 1.4681966067542795, "learning_rate": 2.141051438777656e-06, "loss": 0.6224, "step": 21927 }, { "epoch": 0.794204998189062, "grad_norm": 1.3034292323902732, "learning_rate": 2.140326114522723e-06, "loss": 0.7312, "step": 21928 }, { "epoch": 0.7942412169503803, "grad_norm": 1.010350016573485, "learning_rate": 2.1396008984238094e-06, "loss": 0.6797, "step": 21929 }, { "epoch": 0.7942774357116986, "grad_norm": 1.3129854886730001, "learning_rate": 2.1388757904908908e-06, "loss": 0.713, "step": 21930 }, { "epoch": 0.794313654473017, "grad_norm": 1.4186775902721689, "learning_rate": 2.1381507907339516e-06, "loss": 0.7217, "step": 21931 }, { "epoch": 0.7943498732343354, "grad_norm": 1.4896069733343291, "learning_rate": 2.137425899162966e-06, "loss": 0.6742, "step": 21932 }, { "epoch": 0.7943860919956538, "grad_norm": 1.2571225050810133, "learning_rate": 2.136701115787908e-06, "loss": 0.6634, "step": 21933 }, { "epoch": 0.7944223107569721, "grad_norm": 1.3233782694757195, "learning_rate": 2.1359764406187487e-06, "loss": 0.6911, "step": 21934 }, { "epoch": 0.7944585295182904, "grad_norm": 1.2298735155260343, "learning_rate": 2.135251873665468e-06, "loss": 0.6593, "step": 21935 }, { "epoch": 0.7944947482796089, "grad_norm": 1.707372960734209, "learning_rate": 2.134527414938031e-06, "loss": 0.6546, "step": 21936 }, { "epoch": 0.7945309670409272, "grad_norm": 1.3447337998782047, "learning_rate": 2.1338030644464057e-06, "loss": 0.7526, "step": 21937 }, { "epoch": 0.7945671858022456, "grad_norm": 1.4492239759201875, "learning_rate": 2.133078822200564e-06, "loss": 0.7249, "step": 21938 }, { "epoch": 0.7946034045635639, "grad_norm": 1.0749025603519684, "learning_rate": 2.132354688210471e-06, "loss": 0.7152, "step": 21939 }, { "epoch": 0.7946396233248822, "grad_norm": 1.417777418989431, "learning_rate": 2.1316306624860907e-06, "loss": 0.6622, "step": 21940 }, { "epoch": 0.7946758420862007, "grad_norm": 1.4495563075395892, "learning_rate": 2.130906745037382e-06, "loss": 0.685, "step": 21941 }, { "epoch": 0.794712060847519, "grad_norm": 1.3882193089471957, "learning_rate": 2.1301829358743154e-06, "loss": 0.7267, "step": 21942 }, { "epoch": 0.7947482796088374, "grad_norm": 1.3138243353093157, "learning_rate": 2.1294592350068454e-06, "loss": 0.6508, "step": 21943 }, { "epoch": 0.7947844983701557, "grad_norm": 1.389669093811538, "learning_rate": 2.1287356424449303e-06, "loss": 0.7038, "step": 21944 }, { "epoch": 0.7948207171314741, "grad_norm": 1.384749225470923, "learning_rate": 2.1280121581985314e-06, "loss": 0.6717, "step": 21945 }, { "epoch": 0.7948569358927925, "grad_norm": 1.3223067859007462, "learning_rate": 2.1272887822776025e-06, "loss": 0.6511, "step": 21946 }, { "epoch": 0.7948931546541108, "grad_norm": 1.3402896281172485, "learning_rate": 2.1265655146920972e-06, "loss": 0.7024, "step": 21947 }, { "epoch": 0.7949293734154292, "grad_norm": 1.106522654846108, "learning_rate": 2.1258423554519646e-06, "loss": 0.6817, "step": 21948 }, { "epoch": 0.7949655921767476, "grad_norm": 1.4282682212668945, "learning_rate": 2.1251193045671647e-06, "loss": 0.6877, "step": 21949 }, { "epoch": 0.7950018109380659, "grad_norm": 1.3094173216865264, "learning_rate": 2.1243963620476415e-06, "loss": 0.6792, "step": 21950 }, { "epoch": 0.7950380296993843, "grad_norm": 1.0738841805930122, "learning_rate": 2.123673527903345e-06, "loss": 0.6644, "step": 21951 }, { "epoch": 0.7950742484607026, "grad_norm": 1.326300870383134, "learning_rate": 2.122950802144219e-06, "loss": 0.6299, "step": 21952 }, { "epoch": 0.7951104672220211, "grad_norm": 1.4519382239996719, "learning_rate": 2.122228184780213e-06, "loss": 0.6851, "step": 21953 }, { "epoch": 0.7951466859833394, "grad_norm": 1.3380860687119935, "learning_rate": 2.1215056758212703e-06, "loss": 0.6576, "step": 21954 }, { "epoch": 0.7951829047446577, "grad_norm": 1.2806275690512816, "learning_rate": 2.1207832752773293e-06, "loss": 0.744, "step": 21955 }, { "epoch": 0.7952191235059761, "grad_norm": 1.0590033818558104, "learning_rate": 2.120060983158336e-06, "loss": 0.664, "step": 21956 }, { "epoch": 0.7952553422672944, "grad_norm": 1.2662042459337284, "learning_rate": 2.1193387994742275e-06, "loss": 0.6942, "step": 21957 }, { "epoch": 0.7952915610286129, "grad_norm": 1.3680657209168225, "learning_rate": 2.118616724234942e-06, "loss": 0.7225, "step": 21958 }, { "epoch": 0.7953277797899312, "grad_norm": 1.3314649477217744, "learning_rate": 2.1178947574504128e-06, "loss": 0.6781, "step": 21959 }, { "epoch": 0.7953639985512495, "grad_norm": 1.3992604146704082, "learning_rate": 2.1171728991305797e-06, "loss": 0.6574, "step": 21960 }, { "epoch": 0.7954002173125679, "grad_norm": 1.359611008425076, "learning_rate": 2.1164511492853744e-06, "loss": 0.6817, "step": 21961 }, { "epoch": 0.7954364360738863, "grad_norm": 1.1092448885477073, "learning_rate": 2.1157295079247287e-06, "loss": 0.6538, "step": 21962 }, { "epoch": 0.7954726548352047, "grad_norm": 1.1205298403700747, "learning_rate": 2.1150079750585697e-06, "loss": 0.6498, "step": 21963 }, { "epoch": 0.795508873596523, "grad_norm": 1.3630098391071155, "learning_rate": 2.1142865506968315e-06, "loss": 0.6974, "step": 21964 }, { "epoch": 0.7955450923578413, "grad_norm": 1.1727910168120896, "learning_rate": 2.1135652348494405e-06, "loss": 0.7532, "step": 21965 }, { "epoch": 0.7955813111191598, "grad_norm": 1.067690535498533, "learning_rate": 2.1128440275263175e-06, "loss": 0.632, "step": 21966 }, { "epoch": 0.7956175298804781, "grad_norm": 1.4421633396522042, "learning_rate": 2.1121229287373945e-06, "loss": 0.7143, "step": 21967 }, { "epoch": 0.7956537486417965, "grad_norm": 1.4322336284120334, "learning_rate": 2.111401938492591e-06, "loss": 0.718, "step": 21968 }, { "epoch": 0.7956899674031148, "grad_norm": 1.3574721004607462, "learning_rate": 2.1106810568018275e-06, "loss": 0.6725, "step": 21969 }, { "epoch": 0.7957261861644331, "grad_norm": 1.4456277632326138, "learning_rate": 2.109960283675023e-06, "loss": 0.6606, "step": 21970 }, { "epoch": 0.7957624049257516, "grad_norm": 1.3093658646103272, "learning_rate": 2.1092396191220997e-06, "loss": 0.6539, "step": 21971 }, { "epoch": 0.7957986236870699, "grad_norm": 1.2826073372423286, "learning_rate": 2.1085190631529738e-06, "loss": 0.6718, "step": 21972 }, { "epoch": 0.7958348424483883, "grad_norm": 0.9815218766912984, "learning_rate": 2.1077986157775586e-06, "loss": 0.6892, "step": 21973 }, { "epoch": 0.7958710612097066, "grad_norm": 1.30197413121699, "learning_rate": 2.1070782770057674e-06, "loss": 0.6307, "step": 21974 }, { "epoch": 0.795907279971025, "grad_norm": 1.4836237229678721, "learning_rate": 2.1063580468475165e-06, "loss": 0.7281, "step": 21975 }, { "epoch": 0.7959434987323434, "grad_norm": 1.3566646906604405, "learning_rate": 2.1056379253127144e-06, "loss": 0.6398, "step": 21976 }, { "epoch": 0.7959797174936617, "grad_norm": 1.3847289430255139, "learning_rate": 2.1049179124112694e-06, "loss": 0.6593, "step": 21977 }, { "epoch": 0.7960159362549801, "grad_norm": 1.334069488066553, "learning_rate": 2.1041980081530932e-06, "loss": 0.6016, "step": 21978 }, { "epoch": 0.7960521550162984, "grad_norm": 0.945588925507565, "learning_rate": 2.10347821254809e-06, "loss": 0.6713, "step": 21979 }, { "epoch": 0.7960883737776168, "grad_norm": 1.385712953948699, "learning_rate": 2.1027585256061654e-06, "loss": 0.6849, "step": 21980 }, { "epoch": 0.7961245925389352, "grad_norm": 1.4654325929308123, "learning_rate": 2.1020389473372206e-06, "loss": 0.7532, "step": 21981 }, { "epoch": 0.7961608113002535, "grad_norm": 1.2970668983960236, "learning_rate": 2.1013194777511615e-06, "loss": 0.6801, "step": 21982 }, { "epoch": 0.7961970300615719, "grad_norm": 1.485916406499805, "learning_rate": 2.100600116857887e-06, "loss": 0.7124, "step": 21983 }, { "epoch": 0.7962332488228903, "grad_norm": 1.0059240699999146, "learning_rate": 2.0998808646672964e-06, "loss": 0.6238, "step": 21984 }, { "epoch": 0.7962694675842086, "grad_norm": 1.3118704708092968, "learning_rate": 2.0991617211892836e-06, "loss": 0.6532, "step": 21985 }, { "epoch": 0.796305686345527, "grad_norm": 1.4047105183959012, "learning_rate": 2.0984426864337505e-06, "loss": 0.7314, "step": 21986 }, { "epoch": 0.7963419051068453, "grad_norm": 1.4278693748502491, "learning_rate": 2.0977237604105883e-06, "loss": 0.7001, "step": 21987 }, { "epoch": 0.7963781238681638, "grad_norm": 1.1228428844034277, "learning_rate": 2.097004943129689e-06, "loss": 0.6901, "step": 21988 }, { "epoch": 0.7964143426294821, "grad_norm": 1.4568795260674514, "learning_rate": 2.0962862346009483e-06, "loss": 0.6601, "step": 21989 }, { "epoch": 0.7964505613908004, "grad_norm": 1.2438541266629528, "learning_rate": 2.0955676348342533e-06, "loss": 0.6611, "step": 21990 }, { "epoch": 0.7964867801521188, "grad_norm": 1.0084152479408048, "learning_rate": 2.0948491438394934e-06, "loss": 0.6141, "step": 21991 }, { "epoch": 0.7965229989134371, "grad_norm": 1.5312779555396334, "learning_rate": 2.094130761626553e-06, "loss": 0.7683, "step": 21992 }, { "epoch": 0.7965592176747556, "grad_norm": 1.3445557677678461, "learning_rate": 2.0934124882053218e-06, "loss": 0.6684, "step": 21993 }, { "epoch": 0.7965954364360739, "grad_norm": 1.3538446735141363, "learning_rate": 2.092694323585683e-06, "loss": 0.6399, "step": 21994 }, { "epoch": 0.7966316551973922, "grad_norm": 1.0710045506586172, "learning_rate": 2.091976267777518e-06, "loss": 0.6345, "step": 21995 }, { "epoch": 0.7966678739587106, "grad_norm": 1.4107494123797502, "learning_rate": 2.0912583207907055e-06, "loss": 0.695, "step": 21996 }, { "epoch": 0.796704092720029, "grad_norm": 1.3800338148199611, "learning_rate": 2.0905404826351305e-06, "loss": 0.7404, "step": 21997 }, { "epoch": 0.7967403114813474, "grad_norm": 1.4553251146462385, "learning_rate": 2.089822753320668e-06, "loss": 0.6675, "step": 21998 }, { "epoch": 0.7967765302426657, "grad_norm": 1.4669298201909655, "learning_rate": 2.089105132857193e-06, "loss": 0.79, "step": 21999 }, { "epoch": 0.796812749003984, "grad_norm": 1.3713361221581877, "learning_rate": 2.088387621254584e-06, "loss": 0.6391, "step": 22000 }, { "epoch": 0.7968489677653025, "grad_norm": 1.4773125030424619, "learning_rate": 2.087670218522714e-06, "loss": 0.6792, "step": 22001 }, { "epoch": 0.7968851865266208, "grad_norm": 1.3638649349530008, "learning_rate": 2.0869529246714536e-06, "loss": 0.6314, "step": 22002 }, { "epoch": 0.7969214052879392, "grad_norm": 1.5668685283268562, "learning_rate": 2.0862357397106727e-06, "loss": 0.7145, "step": 22003 }, { "epoch": 0.7969576240492575, "grad_norm": 1.4121006573609787, "learning_rate": 2.0855186636502424e-06, "loss": 0.6852, "step": 22004 }, { "epoch": 0.7969938428105758, "grad_norm": 1.4870726284750682, "learning_rate": 2.084801696500031e-06, "loss": 0.7241, "step": 22005 }, { "epoch": 0.7970300615718943, "grad_norm": 1.0928223881163865, "learning_rate": 2.0840848382698998e-06, "loss": 0.683, "step": 22006 }, { "epoch": 0.7970662803332126, "grad_norm": 1.278202772747248, "learning_rate": 2.0833680889697194e-06, "loss": 0.6931, "step": 22007 }, { "epoch": 0.797102499094531, "grad_norm": 1.3386625471588802, "learning_rate": 2.0826514486093496e-06, "loss": 0.6483, "step": 22008 }, { "epoch": 0.7971387178558493, "grad_norm": 1.2835314496117995, "learning_rate": 2.0819349171986536e-06, "loss": 0.6909, "step": 22009 }, { "epoch": 0.7971749366171677, "grad_norm": 1.5047527667970186, "learning_rate": 2.081218494747488e-06, "loss": 0.7104, "step": 22010 }, { "epoch": 0.7972111553784861, "grad_norm": 1.5016310821554673, "learning_rate": 2.0805021812657168e-06, "loss": 0.6876, "step": 22011 }, { "epoch": 0.7972473741398044, "grad_norm": 1.3610226235008047, "learning_rate": 2.0797859767631936e-06, "loss": 0.7419, "step": 22012 }, { "epoch": 0.7972835929011228, "grad_norm": 1.3550423217952416, "learning_rate": 2.0790698812497744e-06, "loss": 0.6933, "step": 22013 }, { "epoch": 0.7973198116624411, "grad_norm": 0.9720413207780261, "learning_rate": 2.078353894735312e-06, "loss": 0.6769, "step": 22014 }, { "epoch": 0.7973560304237595, "grad_norm": 1.3603993610766896, "learning_rate": 2.0776380172296628e-06, "loss": 0.6263, "step": 22015 }, { "epoch": 0.7973922491850779, "grad_norm": 1.524737486293555, "learning_rate": 2.0769222487426753e-06, "loss": 0.6675, "step": 22016 }, { "epoch": 0.7974284679463962, "grad_norm": 1.3636364665333933, "learning_rate": 2.0762065892841967e-06, "loss": 0.6359, "step": 22017 }, { "epoch": 0.7974646867077146, "grad_norm": 1.3950390972919016, "learning_rate": 2.0754910388640814e-06, "loss": 0.7424, "step": 22018 }, { "epoch": 0.797500905469033, "grad_norm": 1.3620075530734286, "learning_rate": 2.0747755974921714e-06, "loss": 0.7057, "step": 22019 }, { "epoch": 0.7975371242303513, "grad_norm": 0.9646697514896688, "learning_rate": 2.074060265178314e-06, "loss": 0.6393, "step": 22020 }, { "epoch": 0.7975733429916697, "grad_norm": 1.3950138205417046, "learning_rate": 2.0733450419323477e-06, "loss": 0.6385, "step": 22021 }, { "epoch": 0.797609561752988, "grad_norm": 1.2295157534073213, "learning_rate": 2.072629927764125e-06, "loss": 0.6651, "step": 22022 }, { "epoch": 0.7976457805143065, "grad_norm": 1.1963187887429951, "learning_rate": 2.0719149226834765e-06, "loss": 0.6005, "step": 22023 }, { "epoch": 0.7976819992756248, "grad_norm": 1.0893551082901165, "learning_rate": 2.071200026700243e-06, "loss": 0.6542, "step": 22024 }, { "epoch": 0.7977182180369431, "grad_norm": 0.9636479191630287, "learning_rate": 2.070485239824267e-06, "loss": 0.6458, "step": 22025 }, { "epoch": 0.7977544367982615, "grad_norm": 1.3127106337184584, "learning_rate": 2.0697705620653806e-06, "loss": 0.6479, "step": 22026 }, { "epoch": 0.7977906555595798, "grad_norm": 1.9364350607015441, "learning_rate": 2.069055993433421e-06, "loss": 0.6895, "step": 22027 }, { "epoch": 0.7978268743208983, "grad_norm": 1.241120296863355, "learning_rate": 2.068341533938215e-06, "loss": 0.6547, "step": 22028 }, { "epoch": 0.7978630930822166, "grad_norm": 1.2449274344869243, "learning_rate": 2.067627183589602e-06, "loss": 0.6291, "step": 22029 }, { "epoch": 0.7978993118435349, "grad_norm": 1.0630421906825047, "learning_rate": 2.06691294239741e-06, "loss": 0.6811, "step": 22030 }, { "epoch": 0.7979355306048533, "grad_norm": 1.2673068116224953, "learning_rate": 2.0661988103714637e-06, "loss": 0.6532, "step": 22031 }, { "epoch": 0.7979717493661717, "grad_norm": 1.129320307499324, "learning_rate": 2.0654847875215944e-06, "loss": 0.6555, "step": 22032 }, { "epoch": 0.7980079681274901, "grad_norm": 1.2864771348014348, "learning_rate": 2.064770873857628e-06, "loss": 0.6909, "step": 22033 }, { "epoch": 0.7980441868888084, "grad_norm": 1.3634561328152466, "learning_rate": 2.0640570693893856e-06, "loss": 0.7389, "step": 22034 }, { "epoch": 0.7980804056501267, "grad_norm": 1.2816408886919388, "learning_rate": 2.0633433741266884e-06, "loss": 0.6302, "step": 22035 }, { "epoch": 0.7981166244114452, "grad_norm": 1.3621852029337793, "learning_rate": 2.062629788079363e-06, "loss": 0.6395, "step": 22036 }, { "epoch": 0.7981528431727635, "grad_norm": 1.3384861267316746, "learning_rate": 2.0619163112572262e-06, "loss": 0.716, "step": 22037 }, { "epoch": 0.7981890619340819, "grad_norm": 1.4952394621100944, "learning_rate": 2.0612029436700963e-06, "loss": 0.7087, "step": 22038 }, { "epoch": 0.7982252806954002, "grad_norm": 0.9458209831008311, "learning_rate": 2.0604896853277867e-06, "loss": 0.6457, "step": 22039 }, { "epoch": 0.7982614994567185, "grad_norm": 1.2839138452059897, "learning_rate": 2.059776536240118e-06, "loss": 0.6146, "step": 22040 }, { "epoch": 0.798297718218037, "grad_norm": 1.1774648164951582, "learning_rate": 2.0590634964169008e-06, "loss": 0.5966, "step": 22041 }, { "epoch": 0.7983339369793553, "grad_norm": 1.4309736053667, "learning_rate": 2.0583505658679447e-06, "loss": 0.6171, "step": 22042 }, { "epoch": 0.7983701557406737, "grad_norm": 0.956494483478377, "learning_rate": 2.057637744603066e-06, "loss": 0.6782, "step": 22043 }, { "epoch": 0.798406374501992, "grad_norm": 1.3989097354627769, "learning_rate": 2.05692503263207e-06, "loss": 0.6729, "step": 22044 }, { "epoch": 0.7984425932633104, "grad_norm": 1.0641061382979338, "learning_rate": 2.056212429964766e-06, "loss": 0.7163, "step": 22045 }, { "epoch": 0.7984788120246288, "grad_norm": 1.447332934051198, "learning_rate": 2.055499936610955e-06, "loss": 0.6364, "step": 22046 }, { "epoch": 0.7985150307859471, "grad_norm": 1.4070397464129236, "learning_rate": 2.0547875525804496e-06, "loss": 0.7168, "step": 22047 }, { "epoch": 0.7985512495472655, "grad_norm": 1.3869718334122554, "learning_rate": 2.0540752778830475e-06, "loss": 0.6487, "step": 22048 }, { "epoch": 0.7985874683085838, "grad_norm": 0.9871214533061422, "learning_rate": 2.0533631125285504e-06, "loss": 0.6885, "step": 22049 }, { "epoch": 0.7986236870699022, "grad_norm": 1.5410856260058285, "learning_rate": 2.05265105652676e-06, "loss": 0.7181, "step": 22050 }, { "epoch": 0.7986599058312206, "grad_norm": 1.46098213459108, "learning_rate": 2.051939109887475e-06, "loss": 0.7181, "step": 22051 }, { "epoch": 0.7986961245925389, "grad_norm": 1.4167898994612746, "learning_rate": 2.051227272620493e-06, "loss": 0.7057, "step": 22052 }, { "epoch": 0.7987323433538573, "grad_norm": 1.36144472960958, "learning_rate": 2.050515544735604e-06, "loss": 0.6372, "step": 22053 }, { "epoch": 0.7987685621151757, "grad_norm": 1.3420941031970681, "learning_rate": 2.0498039262426096e-06, "loss": 0.6461, "step": 22054 }, { "epoch": 0.798804780876494, "grad_norm": 1.337973564250214, "learning_rate": 2.049092417151298e-06, "loss": 0.5988, "step": 22055 }, { "epoch": 0.7988409996378124, "grad_norm": 1.144182515496719, "learning_rate": 2.0483810174714613e-06, "loss": 0.7344, "step": 22056 }, { "epoch": 0.7988772183991307, "grad_norm": 1.7070415865325483, "learning_rate": 2.0476697272128864e-06, "loss": 0.7494, "step": 22057 }, { "epoch": 0.7989134371604492, "grad_norm": 1.0237243529085407, "learning_rate": 2.046958546385366e-06, "loss": 0.6796, "step": 22058 }, { "epoch": 0.7989496559217675, "grad_norm": 1.3660342625106605, "learning_rate": 2.0462474749986837e-06, "loss": 0.6096, "step": 22059 }, { "epoch": 0.7989858746830858, "grad_norm": 1.3099351718871797, "learning_rate": 2.0455365130626228e-06, "loss": 0.674, "step": 22060 }, { "epoch": 0.7990220934444042, "grad_norm": 1.3240325256656995, "learning_rate": 2.044825660586972e-06, "loss": 0.6369, "step": 22061 }, { "epoch": 0.7990583122057225, "grad_norm": 1.371409028775096, "learning_rate": 2.044114917581509e-06, "loss": 0.6721, "step": 22062 }, { "epoch": 0.799094530967041, "grad_norm": 1.408233257346969, "learning_rate": 2.0434042840560165e-06, "loss": 0.71, "step": 22063 }, { "epoch": 0.7991307497283593, "grad_norm": 0.9659605429485036, "learning_rate": 2.042693760020269e-06, "loss": 0.6294, "step": 22064 }, { "epoch": 0.7991669684896776, "grad_norm": 0.9661126472414574, "learning_rate": 2.04198334548405e-06, "loss": 0.6556, "step": 22065 }, { "epoch": 0.799203187250996, "grad_norm": 1.3595106150406335, "learning_rate": 2.0412730404571325e-06, "loss": 0.6927, "step": 22066 }, { "epoch": 0.7992394060123144, "grad_norm": 1.3582164778284804, "learning_rate": 2.040562844949292e-06, "loss": 0.6956, "step": 22067 }, { "epoch": 0.7992756247736328, "grad_norm": 1.3529721458406538, "learning_rate": 2.0398527589702967e-06, "loss": 0.6907, "step": 22068 }, { "epoch": 0.7993118435349511, "grad_norm": 1.013384597614283, "learning_rate": 2.0391427825299246e-06, "loss": 0.6543, "step": 22069 }, { "epoch": 0.7993480622962694, "grad_norm": 1.2786786150630851, "learning_rate": 2.0384329156379436e-06, "loss": 0.6694, "step": 22070 }, { "epoch": 0.7993842810575879, "grad_norm": 1.5187798090941045, "learning_rate": 2.037723158304119e-06, "loss": 0.6204, "step": 22071 }, { "epoch": 0.7994204998189062, "grad_norm": 1.5014912362609085, "learning_rate": 2.0370135105382215e-06, "loss": 0.6619, "step": 22072 }, { "epoch": 0.7994567185802246, "grad_norm": 1.3061418493687094, "learning_rate": 2.0363039723500155e-06, "loss": 0.6397, "step": 22073 }, { "epoch": 0.7994929373415429, "grad_norm": 1.3905085651537017, "learning_rate": 2.035594543749264e-06, "loss": 0.6789, "step": 22074 }, { "epoch": 0.7995291561028612, "grad_norm": 1.7794569584157516, "learning_rate": 2.0348852247457273e-06, "loss": 0.631, "step": 22075 }, { "epoch": 0.7995653748641797, "grad_norm": 1.3034861581596586, "learning_rate": 2.034176015349172e-06, "loss": 0.6794, "step": 22076 }, { "epoch": 0.799601593625498, "grad_norm": 1.3425246566578501, "learning_rate": 2.033466915569353e-06, "loss": 0.7158, "step": 22077 }, { "epoch": 0.7996378123868164, "grad_norm": 1.3405304982427604, "learning_rate": 2.03275792541603e-06, "loss": 0.6737, "step": 22078 }, { "epoch": 0.7996740311481347, "grad_norm": 1.7705756341197334, "learning_rate": 2.032049044898956e-06, "loss": 0.6961, "step": 22079 }, { "epoch": 0.799710249909453, "grad_norm": 1.0779671142536928, "learning_rate": 2.031340274027891e-06, "loss": 0.6852, "step": 22080 }, { "epoch": 0.7997464686707715, "grad_norm": 2.1064389175226004, "learning_rate": 2.030631612812586e-06, "loss": 0.6948, "step": 22081 }, { "epoch": 0.7997826874320898, "grad_norm": 1.3282722980496915, "learning_rate": 2.0299230612627885e-06, "loss": 0.7216, "step": 22082 }, { "epoch": 0.7998189061934082, "grad_norm": 1.0652225484351505, "learning_rate": 2.029214619388257e-06, "loss": 0.7096, "step": 22083 }, { "epoch": 0.7998551249547265, "grad_norm": 1.3426140264040107, "learning_rate": 2.028506287198735e-06, "loss": 0.6874, "step": 22084 }, { "epoch": 0.7998913437160449, "grad_norm": 1.4027670849571847, "learning_rate": 2.027798064703971e-06, "loss": 0.7455, "step": 22085 }, { "epoch": 0.7999275624773633, "grad_norm": 1.3463013306278335, "learning_rate": 2.027089951913709e-06, "loss": 0.6388, "step": 22086 }, { "epoch": 0.7999637812386816, "grad_norm": 1.453485423683902, "learning_rate": 2.0263819488376956e-06, "loss": 0.7148, "step": 22087 }, { "epoch": 0.8, "grad_norm": 1.5187948815852959, "learning_rate": 2.0256740554856747e-06, "loss": 0.6666, "step": 22088 }, { "epoch": 0.8000362187613184, "grad_norm": 1.3739078041986728, "learning_rate": 2.0249662718673844e-06, "loss": 0.6152, "step": 22089 }, { "epoch": 0.8000724375226367, "grad_norm": 1.4705936244206699, "learning_rate": 2.0242585979925633e-06, "loss": 0.7189, "step": 22090 }, { "epoch": 0.8001086562839551, "grad_norm": 1.6154325085793924, "learning_rate": 2.023551033870955e-06, "loss": 0.7326, "step": 22091 }, { "epoch": 0.8001448750452734, "grad_norm": 1.5921265370813222, "learning_rate": 2.0228435795122937e-06, "loss": 0.6901, "step": 22092 }, { "epoch": 0.8001810938065919, "grad_norm": 1.3990477623609556, "learning_rate": 2.0221362349263106e-06, "loss": 0.6867, "step": 22093 }, { "epoch": 0.8002173125679102, "grad_norm": 1.4126542518794836, "learning_rate": 2.0214290001227457e-06, "loss": 0.6713, "step": 22094 }, { "epoch": 0.8002535313292285, "grad_norm": 1.3444411959128053, "learning_rate": 2.0207218751113287e-06, "loss": 0.6667, "step": 22095 }, { "epoch": 0.8002897500905469, "grad_norm": 1.375995408437416, "learning_rate": 2.020014859901789e-06, "loss": 0.7316, "step": 22096 }, { "epoch": 0.8003259688518652, "grad_norm": 1.5183464668025544, "learning_rate": 2.019307954503855e-06, "loss": 0.6737, "step": 22097 }, { "epoch": 0.8003621876131837, "grad_norm": 1.385281223354517, "learning_rate": 2.018601158927258e-06, "loss": 0.6461, "step": 22098 }, { "epoch": 0.800398406374502, "grad_norm": 1.3292713098928957, "learning_rate": 2.017894473181723e-06, "loss": 0.6823, "step": 22099 }, { "epoch": 0.8004346251358203, "grad_norm": 1.0427576096182054, "learning_rate": 2.017187897276972e-06, "loss": 0.7196, "step": 22100 }, { "epoch": 0.8004708438971387, "grad_norm": 1.3149157662199051, "learning_rate": 2.0164814312227288e-06, "loss": 0.6338, "step": 22101 }, { "epoch": 0.8005070626584571, "grad_norm": 1.4307856016643818, "learning_rate": 2.015775075028719e-06, "loss": 0.6476, "step": 22102 }, { "epoch": 0.8005432814197755, "grad_norm": 1.5256711040142603, "learning_rate": 2.015068828704658e-06, "loss": 0.6823, "step": 22103 }, { "epoch": 0.8005795001810938, "grad_norm": 0.9976995616116335, "learning_rate": 2.0143626922602643e-06, "loss": 0.7382, "step": 22104 }, { "epoch": 0.8006157189424121, "grad_norm": 1.3615707443964795, "learning_rate": 2.0136566657052592e-06, "loss": 0.6543, "step": 22105 }, { "epoch": 0.8006519377037306, "grad_norm": 1.4764971851102195, "learning_rate": 2.012950749049356e-06, "loss": 0.7033, "step": 22106 }, { "epoch": 0.8006881564650489, "grad_norm": 1.1242351552417502, "learning_rate": 2.0122449423022694e-06, "loss": 0.6323, "step": 22107 }, { "epoch": 0.8007243752263673, "grad_norm": 1.1842170024417233, "learning_rate": 2.011539245473707e-06, "loss": 0.6747, "step": 22108 }, { "epoch": 0.8007605939876856, "grad_norm": 1.3537105279607322, "learning_rate": 2.0108336585733877e-06, "loss": 0.6717, "step": 22109 }, { "epoch": 0.8007968127490039, "grad_norm": 1.2871796938973668, "learning_rate": 2.010128181611016e-06, "loss": 0.6826, "step": 22110 }, { "epoch": 0.8008330315103224, "grad_norm": 1.2396183135869852, "learning_rate": 2.009422814596299e-06, "loss": 0.6505, "step": 22111 }, { "epoch": 0.8008692502716407, "grad_norm": 1.4183380070450158, "learning_rate": 2.0087175575389484e-06, "loss": 0.6759, "step": 22112 }, { "epoch": 0.8009054690329591, "grad_norm": 1.359368348927505, "learning_rate": 2.0080124104486653e-06, "loss": 0.6411, "step": 22113 }, { "epoch": 0.8009416877942774, "grad_norm": 1.824586272741668, "learning_rate": 2.0073073733351534e-06, "loss": 0.7219, "step": 22114 }, { "epoch": 0.8009779065555958, "grad_norm": 1.3529729264328474, "learning_rate": 2.0066024462081124e-06, "loss": 0.5303, "step": 22115 }, { "epoch": 0.8010141253169142, "grad_norm": 1.493004053697525, "learning_rate": 2.0058976290772493e-06, "loss": 0.6739, "step": 22116 }, { "epoch": 0.8010503440782325, "grad_norm": 1.0489717890213612, "learning_rate": 2.005192921952258e-06, "loss": 0.6967, "step": 22117 }, { "epoch": 0.8010865628395509, "grad_norm": 1.449085574010724, "learning_rate": 2.004488324842837e-06, "loss": 0.6769, "step": 22118 }, { "epoch": 0.8011227816008692, "grad_norm": 1.5718163766603452, "learning_rate": 2.0037838377586793e-06, "loss": 0.6292, "step": 22119 }, { "epoch": 0.8011590003621876, "grad_norm": 1.41820170522394, "learning_rate": 2.003079460709485e-06, "loss": 0.6783, "step": 22120 }, { "epoch": 0.801195219123506, "grad_norm": 1.3658145113985054, "learning_rate": 2.0023751937049442e-06, "loss": 0.6885, "step": 22121 }, { "epoch": 0.8012314378848243, "grad_norm": 1.4263719308432927, "learning_rate": 2.0016710367547455e-06, "loss": 0.7004, "step": 22122 }, { "epoch": 0.8012676566461427, "grad_norm": 1.4348287505417137, "learning_rate": 2.0009669898685836e-06, "loss": 0.668, "step": 22123 }, { "epoch": 0.8013038754074611, "grad_norm": 1.4512657427816944, "learning_rate": 2.000263053056145e-06, "loss": 0.6965, "step": 22124 }, { "epoch": 0.8013400941687794, "grad_norm": 1.415301675036013, "learning_rate": 1.9995592263271156e-06, "loss": 0.7592, "step": 22125 }, { "epoch": 0.8013763129300978, "grad_norm": 1.3408317287595302, "learning_rate": 1.9988555096911775e-06, "loss": 0.7065, "step": 22126 }, { "epoch": 0.8014125316914161, "grad_norm": 1.142120572219452, "learning_rate": 1.998151903158022e-06, "loss": 0.6609, "step": 22127 }, { "epoch": 0.8014487504527346, "grad_norm": 1.3707074398191368, "learning_rate": 1.9974484067373267e-06, "loss": 0.6411, "step": 22128 }, { "epoch": 0.8014849692140529, "grad_norm": 1.5262067958252952, "learning_rate": 1.9967450204387738e-06, "loss": 0.6753, "step": 22129 }, { "epoch": 0.8015211879753712, "grad_norm": 1.4653508453025708, "learning_rate": 1.996041744272038e-06, "loss": 0.6269, "step": 22130 }, { "epoch": 0.8015574067366896, "grad_norm": 1.5373760401659986, "learning_rate": 1.995338578246804e-06, "loss": 0.6903, "step": 22131 }, { "epoch": 0.801593625498008, "grad_norm": 1.2193435903037435, "learning_rate": 1.9946355223727443e-06, "loss": 0.6867, "step": 22132 }, { "epoch": 0.8016298442593264, "grad_norm": 1.3271561244155325, "learning_rate": 1.9939325766595317e-06, "loss": 0.6649, "step": 22133 }, { "epoch": 0.8016660630206447, "grad_norm": 1.4191749469639792, "learning_rate": 1.993229741116843e-06, "loss": 0.7198, "step": 22134 }, { "epoch": 0.801702281781963, "grad_norm": 1.1333375492976976, "learning_rate": 1.992527015754351e-06, "loss": 0.6453, "step": 22135 }, { "epoch": 0.8017385005432814, "grad_norm": 1.393408681469267, "learning_rate": 1.991824400581721e-06, "loss": 0.6216, "step": 22136 }, { "epoch": 0.8017747193045998, "grad_norm": 1.290670821361325, "learning_rate": 1.991121895608622e-06, "loss": 0.6593, "step": 22137 }, { "epoch": 0.8018109380659182, "grad_norm": 1.3276705497311911, "learning_rate": 1.990419500844725e-06, "loss": 0.6497, "step": 22138 }, { "epoch": 0.8018471568272365, "grad_norm": 1.195913466833068, "learning_rate": 1.9897172162996938e-06, "loss": 0.6366, "step": 22139 }, { "epoch": 0.8018833755885548, "grad_norm": 1.390182539918641, "learning_rate": 1.989015041983192e-06, "loss": 0.5997, "step": 22140 }, { "epoch": 0.8019195943498733, "grad_norm": 1.315861321045659, "learning_rate": 1.9883129779048803e-06, "loss": 0.6204, "step": 22141 }, { "epoch": 0.8019558131111916, "grad_norm": 1.3028223638545446, "learning_rate": 1.987611024074424e-06, "loss": 0.6166, "step": 22142 }, { "epoch": 0.80199203187251, "grad_norm": 1.0980398141741832, "learning_rate": 1.9869091805014806e-06, "loss": 0.7177, "step": 22143 }, { "epoch": 0.8020282506338283, "grad_norm": 1.1309933015888471, "learning_rate": 1.986207447195705e-06, "loss": 0.6654, "step": 22144 }, { "epoch": 0.8020644693951466, "grad_norm": 1.3560936844855727, "learning_rate": 1.985505824166759e-06, "loss": 0.6638, "step": 22145 }, { "epoch": 0.8021006881564651, "grad_norm": 1.4916418048904125, "learning_rate": 1.984804311424294e-06, "loss": 0.7355, "step": 22146 }, { "epoch": 0.8021369069177834, "grad_norm": 1.4282284806613674, "learning_rate": 1.984102908977966e-06, "loss": 0.651, "step": 22147 }, { "epoch": 0.8021731256791018, "grad_norm": 1.2663660212780612, "learning_rate": 1.983401616837423e-06, "loss": 0.6564, "step": 22148 }, { "epoch": 0.8022093444404201, "grad_norm": 1.0298411615009562, "learning_rate": 1.9827004350123193e-06, "loss": 0.6455, "step": 22149 }, { "epoch": 0.8022455632017385, "grad_norm": 1.3113234290049411, "learning_rate": 1.9819993635123014e-06, "loss": 0.6621, "step": 22150 }, { "epoch": 0.8022817819630569, "grad_norm": 1.3867592503449495, "learning_rate": 1.9812984023470196e-06, "loss": 0.7237, "step": 22151 }, { "epoch": 0.8023180007243752, "grad_norm": 1.388515604457341, "learning_rate": 1.980597551526113e-06, "loss": 0.716, "step": 22152 }, { "epoch": 0.8023542194856936, "grad_norm": 1.3837961511019834, "learning_rate": 1.979896811059234e-06, "loss": 0.7616, "step": 22153 }, { "epoch": 0.802390438247012, "grad_norm": 1.4028286241754888, "learning_rate": 1.979196180956022e-06, "loss": 0.6441, "step": 22154 }, { "epoch": 0.8024266570083303, "grad_norm": 1.368644037265728, "learning_rate": 1.9784956612261163e-06, "loss": 0.674, "step": 22155 }, { "epoch": 0.8024628757696487, "grad_norm": 1.4688658985612837, "learning_rate": 1.977795251879161e-06, "loss": 0.7216, "step": 22156 }, { "epoch": 0.802499094530967, "grad_norm": 1.3140735467850133, "learning_rate": 1.9770949529247906e-06, "loss": 0.6881, "step": 22157 }, { "epoch": 0.8025353132922854, "grad_norm": 1.3874150750926741, "learning_rate": 1.9763947643726444e-06, "loss": 0.6264, "step": 22158 }, { "epoch": 0.8025715320536038, "grad_norm": 0.9240582036091077, "learning_rate": 1.9756946862323534e-06, "loss": 0.7404, "step": 22159 }, { "epoch": 0.8026077508149221, "grad_norm": 1.44098619217465, "learning_rate": 1.9749947185135576e-06, "loss": 0.7456, "step": 22160 }, { "epoch": 0.8026439695762405, "grad_norm": 1.4443986806853364, "learning_rate": 1.974294861225885e-06, "loss": 0.6884, "step": 22161 }, { "epoch": 0.8026801883375588, "grad_norm": 0.9855737649461793, "learning_rate": 1.973595114378968e-06, "loss": 0.7085, "step": 22162 }, { "epoch": 0.8027164070988773, "grad_norm": 1.4062548571078872, "learning_rate": 1.972895477982432e-06, "loss": 0.7153, "step": 22163 }, { "epoch": 0.8027526258601956, "grad_norm": 1.3254118098076095, "learning_rate": 1.972195952045909e-06, "loss": 0.7507, "step": 22164 }, { "epoch": 0.8027888446215139, "grad_norm": 1.344943615662608, "learning_rate": 1.971496536579025e-06, "loss": 0.7143, "step": 22165 }, { "epoch": 0.8028250633828323, "grad_norm": 1.7437955722186151, "learning_rate": 1.9707972315913993e-06, "loss": 0.6588, "step": 22166 }, { "epoch": 0.8028612821441506, "grad_norm": 1.4405033635839535, "learning_rate": 1.9700980370926616e-06, "loss": 0.666, "step": 22167 }, { "epoch": 0.8028975009054691, "grad_norm": 1.2798407999900507, "learning_rate": 1.969398953092431e-06, "loss": 0.7199, "step": 22168 }, { "epoch": 0.8029337196667874, "grad_norm": 1.5280560591032746, "learning_rate": 1.968699979600327e-06, "loss": 0.6576, "step": 22169 }, { "epoch": 0.8029699384281057, "grad_norm": 1.2402254400907653, "learning_rate": 1.968001116625966e-06, "loss": 0.624, "step": 22170 }, { "epoch": 0.8030061571894241, "grad_norm": 1.0385228471589423, "learning_rate": 1.967302364178969e-06, "loss": 0.7437, "step": 22171 }, { "epoch": 0.8030423759507425, "grad_norm": 1.5500304376782275, "learning_rate": 1.9666037222689504e-06, "loss": 0.7055, "step": 22172 }, { "epoch": 0.8030785947120609, "grad_norm": 1.1611303305078036, "learning_rate": 1.9659051909055194e-06, "loss": 0.7038, "step": 22173 }, { "epoch": 0.8031148134733792, "grad_norm": 1.3669710989397916, "learning_rate": 1.965206770098297e-06, "loss": 0.6953, "step": 22174 }, { "epoch": 0.8031510322346975, "grad_norm": 1.2950700931100956, "learning_rate": 1.9645084598568876e-06, "loss": 0.6804, "step": 22175 }, { "epoch": 0.803187250996016, "grad_norm": 1.5493243722503474, "learning_rate": 1.9638102601909027e-06, "loss": 0.6736, "step": 22176 }, { "epoch": 0.8032234697573343, "grad_norm": 1.3077315229950766, "learning_rate": 1.9631121711099475e-06, "loss": 0.626, "step": 22177 }, { "epoch": 0.8032596885186527, "grad_norm": 1.3587917962265426, "learning_rate": 1.962414192623634e-06, "loss": 0.715, "step": 22178 }, { "epoch": 0.803295907279971, "grad_norm": 1.484610069634871, "learning_rate": 1.9617163247415627e-06, "loss": 0.6603, "step": 22179 }, { "epoch": 0.8033321260412893, "grad_norm": 1.4363431292246345, "learning_rate": 1.9610185674733386e-06, "loss": 0.6872, "step": 22180 }, { "epoch": 0.8033683448026078, "grad_norm": 1.7620369228452017, "learning_rate": 1.960320920828559e-06, "loss": 0.5977, "step": 22181 }, { "epoch": 0.8034045635639261, "grad_norm": 1.2821130150274986, "learning_rate": 1.959623384816831e-06, "loss": 0.693, "step": 22182 }, { "epoch": 0.8034407823252445, "grad_norm": 1.0315576968560904, "learning_rate": 1.9589259594477503e-06, "loss": 0.6565, "step": 22183 }, { "epoch": 0.8034770010865628, "grad_norm": 1.2989045522188407, "learning_rate": 1.958228644730912e-06, "loss": 0.695, "step": 22184 }, { "epoch": 0.8035132198478812, "grad_norm": 1.3544155599733099, "learning_rate": 1.957531440675915e-06, "loss": 0.626, "step": 22185 }, { "epoch": 0.8035494386091996, "grad_norm": 1.291928123578342, "learning_rate": 1.9568343472923524e-06, "loss": 0.6907, "step": 22186 }, { "epoch": 0.8035856573705179, "grad_norm": 1.3207073404255578, "learning_rate": 1.9561373645898164e-06, "loss": 0.644, "step": 22187 }, { "epoch": 0.8036218761318363, "grad_norm": 1.2497265941414903, "learning_rate": 1.9554404925778957e-06, "loss": 0.7057, "step": 22188 }, { "epoch": 0.8036580948931547, "grad_norm": 1.3870845231687752, "learning_rate": 1.9547437312661856e-06, "loss": 0.6257, "step": 22189 }, { "epoch": 0.803694313654473, "grad_norm": 1.2867961945357385, "learning_rate": 1.9540470806642707e-06, "loss": 0.6166, "step": 22190 }, { "epoch": 0.8037305324157914, "grad_norm": 1.2920130441817463, "learning_rate": 1.9533505407817377e-06, "loss": 0.6198, "step": 22191 }, { "epoch": 0.8037667511771097, "grad_norm": 1.3099208201994932, "learning_rate": 1.952654111628168e-06, "loss": 0.6859, "step": 22192 }, { "epoch": 0.8038029699384281, "grad_norm": 1.2951858209790914, "learning_rate": 1.9519577932131527e-06, "loss": 0.7019, "step": 22193 }, { "epoch": 0.8038391886997465, "grad_norm": 1.3843046774343712, "learning_rate": 1.951261585546269e-06, "loss": 0.6038, "step": 22194 }, { "epoch": 0.8038754074610648, "grad_norm": 1.4644145158677813, "learning_rate": 1.950565488637096e-06, "loss": 0.655, "step": 22195 }, { "epoch": 0.8039116262223832, "grad_norm": 1.4971083258613374, "learning_rate": 1.9498695024952173e-06, "loss": 0.662, "step": 22196 }, { "epoch": 0.8039478449837015, "grad_norm": 1.264132140237044, "learning_rate": 1.9491736271302077e-06, "loss": 0.6541, "step": 22197 }, { "epoch": 0.80398406374502, "grad_norm": 1.1137883863312255, "learning_rate": 1.9484778625516434e-06, "loss": 0.7122, "step": 22198 }, { "epoch": 0.8040202825063383, "grad_norm": 1.0801126440257185, "learning_rate": 1.947782208769096e-06, "loss": 0.6521, "step": 22199 }, { "epoch": 0.8040565012676566, "grad_norm": 1.2999469147615959, "learning_rate": 1.9470866657921416e-06, "loss": 0.6953, "step": 22200 }, { "epoch": 0.804092720028975, "grad_norm": 1.2356799657913315, "learning_rate": 1.9463912336303526e-06, "loss": 0.6188, "step": 22201 }, { "epoch": 0.8041289387902933, "grad_norm": 1.3655800456825073, "learning_rate": 1.945695912293295e-06, "loss": 0.6646, "step": 22202 }, { "epoch": 0.8041651575516118, "grad_norm": 1.2623132130233734, "learning_rate": 1.945000701790537e-06, "loss": 0.6888, "step": 22203 }, { "epoch": 0.8042013763129301, "grad_norm": 0.9666488902668091, "learning_rate": 1.9443056021316496e-06, "loss": 0.7031, "step": 22204 }, { "epoch": 0.8042375950742484, "grad_norm": 2.178151586151759, "learning_rate": 1.9436106133261955e-06, "loss": 0.6671, "step": 22205 }, { "epoch": 0.8042738138355668, "grad_norm": 1.4248871246711032, "learning_rate": 1.9429157353837356e-06, "loss": 0.6706, "step": 22206 }, { "epoch": 0.8043100325968852, "grad_norm": 1.3415674571512848, "learning_rate": 1.942220968313838e-06, "loss": 0.6554, "step": 22207 }, { "epoch": 0.8043462513582036, "grad_norm": 1.4862429084360003, "learning_rate": 1.94152631212606e-06, "loss": 0.6828, "step": 22208 }, { "epoch": 0.8043824701195219, "grad_norm": 1.3805463266082862, "learning_rate": 1.9408317668299605e-06, "loss": 0.7125, "step": 22209 }, { "epoch": 0.8044186888808402, "grad_norm": 1.3110629066357635, "learning_rate": 1.9401373324350947e-06, "loss": 0.5913, "step": 22210 }, { "epoch": 0.8044549076421587, "grad_norm": 0.9329812775297912, "learning_rate": 1.939443008951023e-06, "loss": 0.6703, "step": 22211 }, { "epoch": 0.804491126403477, "grad_norm": 1.3918305154315378, "learning_rate": 1.9387487963872988e-06, "loss": 0.6695, "step": 22212 }, { "epoch": 0.8045273451647954, "grad_norm": 1.0111024754325533, "learning_rate": 1.938054694753475e-06, "loss": 0.6251, "step": 22213 }, { "epoch": 0.8045635639261137, "grad_norm": 1.0638673533840797, "learning_rate": 1.9373607040590992e-06, "loss": 0.6475, "step": 22214 }, { "epoch": 0.804599782687432, "grad_norm": 2.3710195931994043, "learning_rate": 1.9366668243137266e-06, "loss": 0.6614, "step": 22215 }, { "epoch": 0.8046360014487505, "grad_norm": 1.504590136605422, "learning_rate": 1.935973055526904e-06, "loss": 0.6728, "step": 22216 }, { "epoch": 0.8046722202100688, "grad_norm": 1.3023658514292804, "learning_rate": 1.935279397708175e-06, "loss": 0.6457, "step": 22217 }, { "epoch": 0.8047084389713872, "grad_norm": 0.9336013016784376, "learning_rate": 1.9345858508670913e-06, "loss": 0.6864, "step": 22218 }, { "epoch": 0.8047446577327055, "grad_norm": 1.519868899000359, "learning_rate": 1.9338924150131933e-06, "loss": 0.6547, "step": 22219 }, { "epoch": 0.8047808764940239, "grad_norm": 1.5616327078151646, "learning_rate": 1.9331990901560216e-06, "loss": 0.7047, "step": 22220 }, { "epoch": 0.8048170952553423, "grad_norm": 1.427089029988755, "learning_rate": 1.9325058763051176e-06, "loss": 0.6696, "step": 22221 }, { "epoch": 0.8048533140166606, "grad_norm": 0.9340026691390924, "learning_rate": 1.931812773470023e-06, "loss": 0.6403, "step": 22222 }, { "epoch": 0.804889532777979, "grad_norm": 1.443390789128892, "learning_rate": 1.9311197816602746e-06, "loss": 0.6658, "step": 22223 }, { "epoch": 0.8049257515392974, "grad_norm": 1.3838748494748925, "learning_rate": 1.9304269008854073e-06, "loss": 0.6873, "step": 22224 }, { "epoch": 0.8049619703006157, "grad_norm": 1.4349827611719315, "learning_rate": 1.9297341311549543e-06, "loss": 0.6905, "step": 22225 }, { "epoch": 0.8049981890619341, "grad_norm": 1.322193169827342, "learning_rate": 1.9290414724784535e-06, "loss": 0.7178, "step": 22226 }, { "epoch": 0.8050344078232524, "grad_norm": 1.5166438969428615, "learning_rate": 1.928348924865433e-06, "loss": 0.6112, "step": 22227 }, { "epoch": 0.8050706265845708, "grad_norm": 0.9862494862834672, "learning_rate": 1.927656488325421e-06, "loss": 0.6827, "step": 22228 }, { "epoch": 0.8051068453458892, "grad_norm": 1.7843946485055588, "learning_rate": 1.9269641628679527e-06, "loss": 0.7306, "step": 22229 }, { "epoch": 0.8051430641072075, "grad_norm": 1.356051067763383, "learning_rate": 1.9262719485025506e-06, "loss": 0.6912, "step": 22230 }, { "epoch": 0.8051792828685259, "grad_norm": 1.303085260591463, "learning_rate": 1.92557984523874e-06, "loss": 0.6182, "step": 22231 }, { "epoch": 0.8052155016298442, "grad_norm": 1.295866748103265, "learning_rate": 1.9248878530860436e-06, "loss": 0.668, "step": 22232 }, { "epoch": 0.8052517203911627, "grad_norm": 1.4079026908329964, "learning_rate": 1.9241959720539902e-06, "loss": 0.6787, "step": 22233 }, { "epoch": 0.805287939152481, "grad_norm": 1.2787975283126156, "learning_rate": 1.9235042021520946e-06, "loss": 0.6639, "step": 22234 }, { "epoch": 0.8053241579137993, "grad_norm": 1.3083964718674208, "learning_rate": 1.9228125433898748e-06, "loss": 0.7134, "step": 22235 }, { "epoch": 0.8053603766751177, "grad_norm": 1.1477327943939264, "learning_rate": 1.922120995776854e-06, "loss": 0.6612, "step": 22236 }, { "epoch": 0.805396595436436, "grad_norm": 1.3325869947488174, "learning_rate": 1.9214295593225464e-06, "loss": 0.6606, "step": 22237 }, { "epoch": 0.8054328141977545, "grad_norm": 0.8920157915458256, "learning_rate": 1.920738234036463e-06, "loss": 0.6663, "step": 22238 }, { "epoch": 0.8054690329590728, "grad_norm": 1.2763563168799013, "learning_rate": 1.920047019928125e-06, "loss": 0.6665, "step": 22239 }, { "epoch": 0.8055052517203911, "grad_norm": 1.3780170780643226, "learning_rate": 1.9193559170070387e-06, "loss": 0.6099, "step": 22240 }, { "epoch": 0.8055414704817095, "grad_norm": 1.3563341626628982, "learning_rate": 1.918664925282715e-06, "loss": 0.6822, "step": 22241 }, { "epoch": 0.8055776892430279, "grad_norm": 1.287606408180232, "learning_rate": 1.91797404476466e-06, "loss": 0.6541, "step": 22242 }, { "epoch": 0.8056139080043463, "grad_norm": 1.0286573333125066, "learning_rate": 1.917283275462387e-06, "loss": 0.6967, "step": 22243 }, { "epoch": 0.8056501267656646, "grad_norm": 1.4027316186634031, "learning_rate": 1.9165926173853977e-06, "loss": 0.6398, "step": 22244 }, { "epoch": 0.8056863455269829, "grad_norm": 1.4146465826068468, "learning_rate": 1.9159020705431974e-06, "loss": 0.6968, "step": 22245 }, { "epoch": 0.8057225642883014, "grad_norm": 1.408375668635912, "learning_rate": 1.9152116349452845e-06, "loss": 0.6777, "step": 22246 }, { "epoch": 0.8057587830496197, "grad_norm": 1.3869662842980142, "learning_rate": 1.914521310601166e-06, "loss": 0.6403, "step": 22247 }, { "epoch": 0.8057950018109381, "grad_norm": 1.4011679730388928, "learning_rate": 1.9138310975203402e-06, "loss": 0.6889, "step": 22248 }, { "epoch": 0.8058312205722564, "grad_norm": 1.3569894431179867, "learning_rate": 1.9131409957122993e-06, "loss": 0.6288, "step": 22249 }, { "epoch": 0.8058674393335747, "grad_norm": 1.323678934859392, "learning_rate": 1.9124510051865475e-06, "loss": 0.6517, "step": 22250 }, { "epoch": 0.8059036580948932, "grad_norm": 1.3436598248498408, "learning_rate": 1.9117611259525758e-06, "loss": 0.61, "step": 22251 }, { "epoch": 0.8059398768562115, "grad_norm": 1.371757555846249, "learning_rate": 1.911071358019878e-06, "loss": 0.6342, "step": 22252 }, { "epoch": 0.8059760956175299, "grad_norm": 1.3415516041966988, "learning_rate": 1.910381701397944e-06, "loss": 0.6767, "step": 22253 }, { "epoch": 0.8060123143788482, "grad_norm": 1.3517890585936994, "learning_rate": 1.909692156096268e-06, "loss": 0.6422, "step": 22254 }, { "epoch": 0.8060485331401666, "grad_norm": 1.2680304848654809, "learning_rate": 1.9090027221243366e-06, "loss": 0.6438, "step": 22255 }, { "epoch": 0.806084751901485, "grad_norm": 1.1024453467113566, "learning_rate": 1.908313399491637e-06, "loss": 0.6824, "step": 22256 }, { "epoch": 0.8061209706628033, "grad_norm": 1.335016439912373, "learning_rate": 1.907624188207653e-06, "loss": 0.7245, "step": 22257 }, { "epoch": 0.8061571894241217, "grad_norm": 1.2806925156832856, "learning_rate": 1.9069350882818726e-06, "loss": 0.645, "step": 22258 }, { "epoch": 0.80619340818544, "grad_norm": 1.3317295701895278, "learning_rate": 1.9062460997237775e-06, "loss": 0.6468, "step": 22259 }, { "epoch": 0.8062296269467584, "grad_norm": 1.6009643626937868, "learning_rate": 1.9055572225428442e-06, "loss": 0.7267, "step": 22260 }, { "epoch": 0.8062658457080768, "grad_norm": 1.58533236275909, "learning_rate": 1.9048684567485587e-06, "loss": 0.658, "step": 22261 }, { "epoch": 0.8063020644693951, "grad_norm": 1.0692245778142806, "learning_rate": 1.9041798023503965e-06, "loss": 0.6266, "step": 22262 }, { "epoch": 0.8063382832307135, "grad_norm": 1.3612552750577647, "learning_rate": 1.9034912593578348e-06, "loss": 0.6559, "step": 22263 }, { "epoch": 0.8063745019920319, "grad_norm": 1.3583684557328566, "learning_rate": 1.9028028277803435e-06, "loss": 0.7091, "step": 22264 }, { "epoch": 0.8064107207533502, "grad_norm": 1.4052303365919878, "learning_rate": 1.9021145076274028e-06, "loss": 0.6388, "step": 22265 }, { "epoch": 0.8064469395146686, "grad_norm": 1.65337531713118, "learning_rate": 1.901426298908483e-06, "loss": 0.6625, "step": 22266 }, { "epoch": 0.8064831582759869, "grad_norm": 1.3981270152398475, "learning_rate": 1.90073820163305e-06, "loss": 0.7099, "step": 22267 }, { "epoch": 0.8065193770373054, "grad_norm": 1.378192515622427, "learning_rate": 1.9000502158105783e-06, "loss": 0.6589, "step": 22268 }, { "epoch": 0.8065555957986237, "grad_norm": 1.3588125816569956, "learning_rate": 1.8993623414505335e-06, "loss": 0.6277, "step": 22269 }, { "epoch": 0.806591814559942, "grad_norm": 1.015146291772742, "learning_rate": 1.8986745785623806e-06, "loss": 0.678, "step": 22270 }, { "epoch": 0.8066280333212604, "grad_norm": 0.951321408729422, "learning_rate": 1.8979869271555817e-06, "loss": 0.6731, "step": 22271 }, { "epoch": 0.8066642520825787, "grad_norm": 1.4298871088608223, "learning_rate": 1.8972993872396038e-06, "loss": 0.7549, "step": 22272 }, { "epoch": 0.8067004708438972, "grad_norm": 1.2579030051012199, "learning_rate": 1.8966119588239064e-06, "loss": 0.6796, "step": 22273 }, { "epoch": 0.8067366896052155, "grad_norm": 1.3896529317583444, "learning_rate": 1.8959246419179479e-06, "loss": 0.6785, "step": 22274 }, { "epoch": 0.8067729083665338, "grad_norm": 1.2768458167271108, "learning_rate": 1.8952374365311855e-06, "loss": 0.6764, "step": 22275 }, { "epoch": 0.8068091271278522, "grad_norm": 1.3662988444930608, "learning_rate": 1.8945503426730794e-06, "loss": 0.7149, "step": 22276 }, { "epoch": 0.8068453458891706, "grad_norm": 2.4259122745554245, "learning_rate": 1.8938633603530832e-06, "loss": 0.692, "step": 22277 }, { "epoch": 0.806881564650489, "grad_norm": 1.427691033693087, "learning_rate": 1.8931764895806458e-06, "loss": 0.6741, "step": 22278 }, { "epoch": 0.8069177834118073, "grad_norm": 1.2425250196156554, "learning_rate": 1.8924897303652278e-06, "loss": 0.6141, "step": 22279 }, { "epoch": 0.8069540021731256, "grad_norm": 4.692725828972936, "learning_rate": 1.8918030827162737e-06, "loss": 0.6815, "step": 22280 }, { "epoch": 0.8069902209344441, "grad_norm": 1.4354727771345863, "learning_rate": 1.8911165466432346e-06, "loss": 0.6605, "step": 22281 }, { "epoch": 0.8070264396957624, "grad_norm": 1.0699590083528814, "learning_rate": 1.8904301221555533e-06, "loss": 0.6357, "step": 22282 }, { "epoch": 0.8070626584570808, "grad_norm": 1.2260756005515239, "learning_rate": 1.889743809262682e-06, "loss": 0.7295, "step": 22283 }, { "epoch": 0.8070988772183991, "grad_norm": 1.3902081155346868, "learning_rate": 1.889057607974063e-06, "loss": 0.654, "step": 22284 }, { "epoch": 0.8071350959797174, "grad_norm": 1.3008571164457094, "learning_rate": 1.888371518299137e-06, "loss": 0.6672, "step": 22285 }, { "epoch": 0.8071713147410359, "grad_norm": 1.2660218671817252, "learning_rate": 1.8876855402473448e-06, "loss": 0.5907, "step": 22286 }, { "epoch": 0.8072075335023542, "grad_norm": 1.3831967881280274, "learning_rate": 1.8869996738281304e-06, "loss": 0.7102, "step": 22287 }, { "epoch": 0.8072437522636726, "grad_norm": 1.338876267895558, "learning_rate": 1.8863139190509283e-06, "loss": 0.6142, "step": 22288 }, { "epoch": 0.8072799710249909, "grad_norm": 1.50730760303826, "learning_rate": 1.8856282759251742e-06, "loss": 0.6645, "step": 22289 }, { "epoch": 0.8073161897863093, "grad_norm": 1.184695156276941, "learning_rate": 1.8849427444603063e-06, "loss": 0.5865, "step": 22290 }, { "epoch": 0.8073524085476277, "grad_norm": 1.5266903189265593, "learning_rate": 1.884257324665758e-06, "loss": 0.7462, "step": 22291 }, { "epoch": 0.807388627308946, "grad_norm": 1.045555565017393, "learning_rate": 1.8835720165509598e-06, "loss": 0.6371, "step": 22292 }, { "epoch": 0.8074248460702644, "grad_norm": 1.503846895607189, "learning_rate": 1.8828868201253391e-06, "loss": 0.6907, "step": 22293 }, { "epoch": 0.8074610648315828, "grad_norm": 1.2696426804863903, "learning_rate": 1.8822017353983312e-06, "loss": 0.622, "step": 22294 }, { "epoch": 0.8074972835929011, "grad_norm": 2.037443153781899, "learning_rate": 1.8815167623793606e-06, "loss": 0.5921, "step": 22295 }, { "epoch": 0.8075335023542195, "grad_norm": 1.3572785638038665, "learning_rate": 1.880831901077853e-06, "loss": 0.6834, "step": 22296 }, { "epoch": 0.8075697211155378, "grad_norm": 1.1589028222450402, "learning_rate": 1.88014715150323e-06, "loss": 0.6093, "step": 22297 }, { "epoch": 0.8076059398768562, "grad_norm": 1.4255911295344574, "learning_rate": 1.8794625136649181e-06, "loss": 0.6914, "step": 22298 }, { "epoch": 0.8076421586381746, "grad_norm": 1.3214812635880706, "learning_rate": 1.8787779875723389e-06, "loss": 0.682, "step": 22299 }, { "epoch": 0.8076783773994929, "grad_norm": 1.26810449316079, "learning_rate": 1.878093573234907e-06, "loss": 0.6594, "step": 22300 }, { "epoch": 0.8077145961608113, "grad_norm": 0.9398864528457505, "learning_rate": 1.8774092706620473e-06, "loss": 0.6528, "step": 22301 }, { "epoch": 0.8077508149221296, "grad_norm": 1.3295401663015547, "learning_rate": 1.8767250798631731e-06, "loss": 0.6525, "step": 22302 }, { "epoch": 0.8077870336834481, "grad_norm": 0.9647458745105887, "learning_rate": 1.8760410008476993e-06, "loss": 0.6617, "step": 22303 }, { "epoch": 0.8078232524447664, "grad_norm": 1.485163104248685, "learning_rate": 1.875357033625037e-06, "loss": 0.7188, "step": 22304 }, { "epoch": 0.8078594712060847, "grad_norm": 1.3649470848341467, "learning_rate": 1.8746731782046035e-06, "loss": 0.7062, "step": 22305 }, { "epoch": 0.8078956899674031, "grad_norm": 1.385888063093105, "learning_rate": 1.8739894345958065e-06, "loss": 0.6522, "step": 22306 }, { "epoch": 0.8079319087287214, "grad_norm": 1.4629273142625665, "learning_rate": 1.8733058028080552e-06, "loss": 0.6469, "step": 22307 }, { "epoch": 0.8079681274900399, "grad_norm": 1.368826102863867, "learning_rate": 1.8726222828507546e-06, "loss": 0.6712, "step": 22308 }, { "epoch": 0.8080043462513582, "grad_norm": 1.4725971032971394, "learning_rate": 1.8719388747333145e-06, "loss": 0.6923, "step": 22309 }, { "epoch": 0.8080405650126765, "grad_norm": 1.1389499397070504, "learning_rate": 1.8712555784651375e-06, "loss": 0.6464, "step": 22310 }, { "epoch": 0.8080767837739949, "grad_norm": 1.3799468428542847, "learning_rate": 1.8705723940556242e-06, "loss": 0.7215, "step": 22311 }, { "epoch": 0.8081130025353133, "grad_norm": 1.3610996522557606, "learning_rate": 1.86988932151418e-06, "loss": 0.6942, "step": 22312 }, { "epoch": 0.8081492212966317, "grad_norm": 1.4634349698976115, "learning_rate": 1.8692063608502032e-06, "loss": 0.735, "step": 22313 }, { "epoch": 0.80818544005795, "grad_norm": 1.251764057064685, "learning_rate": 1.8685235120730904e-06, "loss": 0.6434, "step": 22314 }, { "epoch": 0.8082216588192683, "grad_norm": 1.0757881228264472, "learning_rate": 1.8678407751922368e-06, "loss": 0.6632, "step": 22315 }, { "epoch": 0.8082578775805868, "grad_norm": 1.4290619312759059, "learning_rate": 1.8671581502170421e-06, "loss": 0.7322, "step": 22316 }, { "epoch": 0.8082940963419051, "grad_norm": 0.9807735456527306, "learning_rate": 1.8664756371568981e-06, "loss": 0.7766, "step": 22317 }, { "epoch": 0.8083303151032235, "grad_norm": 1.434194484428119, "learning_rate": 1.865793236021195e-06, "loss": 0.7337, "step": 22318 }, { "epoch": 0.8083665338645418, "grad_norm": 0.986404853641905, "learning_rate": 1.865110946819323e-06, "loss": 0.6755, "step": 22319 }, { "epoch": 0.8084027526258601, "grad_norm": 1.1881668407987929, "learning_rate": 1.8644287695606744e-06, "loss": 0.6778, "step": 22320 }, { "epoch": 0.8084389713871786, "grad_norm": 1.4446948046028225, "learning_rate": 1.8637467042546343e-06, "loss": 0.7245, "step": 22321 }, { "epoch": 0.8084751901484969, "grad_norm": 1.3967868317337357, "learning_rate": 1.8630647509105859e-06, "loss": 0.6217, "step": 22322 }, { "epoch": 0.8085114089098153, "grad_norm": 1.0371119346571351, "learning_rate": 1.8623829095379188e-06, "loss": 0.684, "step": 22323 }, { "epoch": 0.8085476276711336, "grad_norm": 1.3519431450549029, "learning_rate": 1.8617011801460128e-06, "loss": 0.6748, "step": 22324 }, { "epoch": 0.808583846432452, "grad_norm": 1.4081347392033672, "learning_rate": 1.8610195627442507e-06, "loss": 0.7295, "step": 22325 }, { "epoch": 0.8086200651937704, "grad_norm": 0.9483769503483185, "learning_rate": 1.8603380573420072e-06, "loss": 0.5874, "step": 22326 }, { "epoch": 0.8086562839550887, "grad_norm": 1.4435200505003778, "learning_rate": 1.859656663948668e-06, "loss": 0.6727, "step": 22327 }, { "epoch": 0.8086925027164071, "grad_norm": 1.2826752549456253, "learning_rate": 1.8589753825736045e-06, "loss": 0.5783, "step": 22328 }, { "epoch": 0.8087287214777255, "grad_norm": 1.190645435394421, "learning_rate": 1.8582942132261916e-06, "loss": 0.649, "step": 22329 }, { "epoch": 0.8087649402390438, "grad_norm": 1.3729277792459404, "learning_rate": 1.857613155915806e-06, "loss": 0.6525, "step": 22330 }, { "epoch": 0.8088011590003622, "grad_norm": 1.0088409619815197, "learning_rate": 1.8569322106518184e-06, "loss": 0.6708, "step": 22331 }, { "epoch": 0.8088373777616805, "grad_norm": 1.3224224225814587, "learning_rate": 1.8562513774435986e-06, "loss": 0.693, "step": 22332 }, { "epoch": 0.808873596522999, "grad_norm": 1.4244868033051574, "learning_rate": 1.8555706563005126e-06, "loss": 0.7182, "step": 22333 }, { "epoch": 0.8089098152843173, "grad_norm": 1.3339837375879509, "learning_rate": 1.8548900472319342e-06, "loss": 0.6779, "step": 22334 }, { "epoch": 0.8089460340456356, "grad_norm": 1.3693988110384272, "learning_rate": 1.8542095502472247e-06, "loss": 0.6583, "step": 22335 }, { "epoch": 0.808982252806954, "grad_norm": 0.9319507522671323, "learning_rate": 1.8535291653557498e-06, "loss": 0.6851, "step": 22336 }, { "epoch": 0.8090184715682723, "grad_norm": 1.3616178572314692, "learning_rate": 1.8528488925668685e-06, "loss": 0.6455, "step": 22337 }, { "epoch": 0.8090546903295908, "grad_norm": 1.4724686064180796, "learning_rate": 1.8521687318899483e-06, "loss": 0.6917, "step": 22338 }, { "epoch": 0.8090909090909091, "grad_norm": 1.4457833025905478, "learning_rate": 1.851488683334346e-06, "loss": 0.7187, "step": 22339 }, { "epoch": 0.8091271278522274, "grad_norm": 1.334506634957637, "learning_rate": 1.850808746909416e-06, "loss": 0.7339, "step": 22340 }, { "epoch": 0.8091633466135458, "grad_norm": 1.42082293726058, "learning_rate": 1.8501289226245212e-06, "loss": 0.657, "step": 22341 }, { "epoch": 0.8091995653748641, "grad_norm": 1.4023178378449996, "learning_rate": 1.8494492104890137e-06, "loss": 0.6822, "step": 22342 }, { "epoch": 0.8092357841361826, "grad_norm": 1.3622038656957263, "learning_rate": 1.8487696105122466e-06, "loss": 0.7283, "step": 22343 }, { "epoch": 0.8092720028975009, "grad_norm": 1.3709958277890308, "learning_rate": 1.8480901227035696e-06, "loss": 0.721, "step": 22344 }, { "epoch": 0.8093082216588192, "grad_norm": 1.5573269217837213, "learning_rate": 1.8474107470723368e-06, "loss": 0.639, "step": 22345 }, { "epoch": 0.8093444404201376, "grad_norm": 1.4111374693269019, "learning_rate": 1.846731483627897e-06, "loss": 0.6668, "step": 22346 }, { "epoch": 0.809380659181456, "grad_norm": 1.390962841620061, "learning_rate": 1.8460523323795954e-06, "loss": 0.6554, "step": 22347 }, { "epoch": 0.8094168779427744, "grad_norm": 1.5610424145388708, "learning_rate": 1.8453732933367762e-06, "loss": 0.6338, "step": 22348 }, { "epoch": 0.8094530967040927, "grad_norm": 1.2734456398475598, "learning_rate": 1.844694366508788e-06, "loss": 0.6956, "step": 22349 }, { "epoch": 0.809489315465411, "grad_norm": 1.220947956709732, "learning_rate": 1.8440155519049707e-06, "loss": 0.6537, "step": 22350 }, { "epoch": 0.8095255342267295, "grad_norm": 1.5581613791825017, "learning_rate": 1.843336849534665e-06, "loss": 0.6614, "step": 22351 }, { "epoch": 0.8095617529880478, "grad_norm": 1.0369117853294962, "learning_rate": 1.8426582594072118e-06, "loss": 0.7243, "step": 22352 }, { "epoch": 0.8095979717493662, "grad_norm": 1.3740172572427343, "learning_rate": 1.8419797815319495e-06, "loss": 0.6577, "step": 22353 }, { "epoch": 0.8096341905106845, "grad_norm": 1.3887033271178637, "learning_rate": 1.841301415918214e-06, "loss": 0.6917, "step": 22354 }, { "epoch": 0.8096704092720028, "grad_norm": 0.9827436800884164, "learning_rate": 1.840623162575338e-06, "loss": 0.652, "step": 22355 }, { "epoch": 0.8097066280333213, "grad_norm": 1.4130417560506396, "learning_rate": 1.839945021512658e-06, "loss": 0.5907, "step": 22356 }, { "epoch": 0.8097428467946396, "grad_norm": 1.4211797429218709, "learning_rate": 1.8392669927395047e-06, "loss": 0.7223, "step": 22357 }, { "epoch": 0.809779065555958, "grad_norm": 1.4469485855855198, "learning_rate": 1.8385890762652082e-06, "loss": 0.6616, "step": 22358 }, { "epoch": 0.8098152843172763, "grad_norm": 1.4699311072885561, "learning_rate": 1.8379112720990955e-06, "loss": 0.6857, "step": 22359 }, { "epoch": 0.8098515030785947, "grad_norm": 1.4182956958355304, "learning_rate": 1.837233580250497e-06, "loss": 0.682, "step": 22360 }, { "epoch": 0.8098877218399131, "grad_norm": 1.1574506510737836, "learning_rate": 1.8365560007287376e-06, "loss": 0.744, "step": 22361 }, { "epoch": 0.8099239406012314, "grad_norm": 1.4940410817714918, "learning_rate": 1.8358785335431383e-06, "loss": 0.7414, "step": 22362 }, { "epoch": 0.8099601593625498, "grad_norm": 1.4368386703961775, "learning_rate": 1.8352011787030266e-06, "loss": 0.7283, "step": 22363 }, { "epoch": 0.8099963781238682, "grad_norm": 1.3594355542299348, "learning_rate": 1.8345239362177203e-06, "loss": 0.6742, "step": 22364 }, { "epoch": 0.8100325968851865, "grad_norm": 1.3621863365037101, "learning_rate": 1.8338468060965398e-06, "loss": 0.6722, "step": 22365 }, { "epoch": 0.8100688156465049, "grad_norm": 1.2715018120315689, "learning_rate": 1.8331697883487999e-06, "loss": 0.6288, "step": 22366 }, { "epoch": 0.8101050344078232, "grad_norm": 1.2400433303091438, "learning_rate": 1.8324928829838218e-06, "loss": 0.6761, "step": 22367 }, { "epoch": 0.8101412531691417, "grad_norm": 1.016290315968801, "learning_rate": 1.8318160900109194e-06, "loss": 0.6854, "step": 22368 }, { "epoch": 0.81017747193046, "grad_norm": 1.3376813237554246, "learning_rate": 1.8311394094394041e-06, "loss": 0.7805, "step": 22369 }, { "epoch": 0.8102136906917783, "grad_norm": 1.3201646303185908, "learning_rate": 1.8304628412785852e-06, "loss": 0.729, "step": 22370 }, { "epoch": 0.8102499094530967, "grad_norm": 1.2407647341168953, "learning_rate": 1.8297863855377796e-06, "loss": 0.6804, "step": 22371 }, { "epoch": 0.810286128214415, "grad_norm": 1.912609065248003, "learning_rate": 1.829110042226292e-06, "loss": 0.6537, "step": 22372 }, { "epoch": 0.8103223469757335, "grad_norm": 1.0725384162303713, "learning_rate": 1.8284338113534273e-06, "loss": 0.6662, "step": 22373 }, { "epoch": 0.8103585657370518, "grad_norm": 1.3704632820222116, "learning_rate": 1.8277576929284956e-06, "loss": 0.6352, "step": 22374 }, { "epoch": 0.8103947844983701, "grad_norm": 1.3262598081602683, "learning_rate": 1.827081686960799e-06, "loss": 0.6483, "step": 22375 }, { "epoch": 0.8104310032596885, "grad_norm": 1.3812595546993922, "learning_rate": 1.8264057934596403e-06, "loss": 0.636, "step": 22376 }, { "epoch": 0.8104672220210069, "grad_norm": 1.3615823861290304, "learning_rate": 1.825730012434316e-06, "loss": 0.6613, "step": 22377 }, { "epoch": 0.8105034407823253, "grad_norm": 0.8982534752009211, "learning_rate": 1.8250543438941326e-06, "loss": 0.6735, "step": 22378 }, { "epoch": 0.8105396595436436, "grad_norm": 1.2700966991606384, "learning_rate": 1.8243787878483853e-06, "loss": 0.6529, "step": 22379 }, { "epoch": 0.8105758783049619, "grad_norm": 1.2822445463617542, "learning_rate": 1.823703344306369e-06, "loss": 0.6613, "step": 22380 }, { "epoch": 0.8106120970662803, "grad_norm": 1.4644147460886248, "learning_rate": 1.8230280132773758e-06, "loss": 0.6129, "step": 22381 }, { "epoch": 0.8106483158275987, "grad_norm": 1.4920076759943668, "learning_rate": 1.8223527947707053e-06, "loss": 0.6942, "step": 22382 }, { "epoch": 0.8106845345889171, "grad_norm": 1.297575482094702, "learning_rate": 1.8216776887956456e-06, "loss": 0.6119, "step": 22383 }, { "epoch": 0.8107207533502354, "grad_norm": 1.3388629651716057, "learning_rate": 1.8210026953614846e-06, "loss": 0.7174, "step": 22384 }, { "epoch": 0.8107569721115537, "grad_norm": 1.2257488681960418, "learning_rate": 1.8203278144775162e-06, "loss": 0.6241, "step": 22385 }, { "epoch": 0.8107931908728722, "grad_norm": 1.3875503766359014, "learning_rate": 1.8196530461530238e-06, "loss": 0.7485, "step": 22386 }, { "epoch": 0.8108294096341905, "grad_norm": 1.401297915138842, "learning_rate": 1.8189783903972934e-06, "loss": 0.7091, "step": 22387 }, { "epoch": 0.8108656283955089, "grad_norm": 1.4108841026801167, "learning_rate": 1.8183038472196057e-06, "loss": 0.7143, "step": 22388 }, { "epoch": 0.8109018471568272, "grad_norm": 1.437880277868472, "learning_rate": 1.8176294166292497e-06, "loss": 0.7474, "step": 22389 }, { "epoch": 0.8109380659181455, "grad_norm": 1.2340633445557183, "learning_rate": 1.8169550986355022e-06, "loss": 0.6281, "step": 22390 }, { "epoch": 0.810974284679464, "grad_norm": 1.2326344204376347, "learning_rate": 1.8162808932476406e-06, "loss": 0.6884, "step": 22391 }, { "epoch": 0.8110105034407823, "grad_norm": 1.3570591496831765, "learning_rate": 1.8156068004749472e-06, "loss": 0.6305, "step": 22392 }, { "epoch": 0.8110467222021007, "grad_norm": 1.4308560812139235, "learning_rate": 1.8149328203266947e-06, "loss": 0.7045, "step": 22393 }, { "epoch": 0.811082940963419, "grad_norm": 1.3591995256173672, "learning_rate": 1.81425895281216e-06, "loss": 0.6379, "step": 22394 }, { "epoch": 0.8111191597247374, "grad_norm": 1.3424032387635636, "learning_rate": 1.8135851979406115e-06, "loss": 0.7047, "step": 22395 }, { "epoch": 0.8111553784860558, "grad_norm": 1.364433230875639, "learning_rate": 1.8129115557213262e-06, "loss": 0.6403, "step": 22396 }, { "epoch": 0.8111915972473741, "grad_norm": 1.3748808214437818, "learning_rate": 1.8122380261635719e-06, "loss": 0.64, "step": 22397 }, { "epoch": 0.8112278160086925, "grad_norm": 0.9828216546468365, "learning_rate": 1.811564609276617e-06, "loss": 0.7109, "step": 22398 }, { "epoch": 0.8112640347700109, "grad_norm": 0.9555593761654287, "learning_rate": 1.8108913050697252e-06, "loss": 0.6333, "step": 22399 }, { "epoch": 0.8113002535313292, "grad_norm": 1.288043586867668, "learning_rate": 1.8102181135521668e-06, "loss": 0.7336, "step": 22400 }, { "epoch": 0.8113364722926476, "grad_norm": 1.0901679655697347, "learning_rate": 1.8095450347332045e-06, "loss": 0.6778, "step": 22401 }, { "epoch": 0.8113726910539659, "grad_norm": 1.2872306566882623, "learning_rate": 1.8088720686220962e-06, "loss": 0.6307, "step": 22402 }, { "epoch": 0.8114089098152844, "grad_norm": 1.2647399501019512, "learning_rate": 1.808199215228108e-06, "loss": 0.5665, "step": 22403 }, { "epoch": 0.8114451285766027, "grad_norm": 1.3388025037247535, "learning_rate": 1.8075264745604958e-06, "loss": 0.7765, "step": 22404 }, { "epoch": 0.811481347337921, "grad_norm": 1.4193661807927218, "learning_rate": 1.806853846628519e-06, "loss": 0.6445, "step": 22405 }, { "epoch": 0.8115175660992394, "grad_norm": 1.4278242295583454, "learning_rate": 1.80618133144143e-06, "loss": 0.6554, "step": 22406 }, { "epoch": 0.8115537848605577, "grad_norm": 1.354818017517475, "learning_rate": 1.805508929008487e-06, "loss": 0.6442, "step": 22407 }, { "epoch": 0.8115900036218762, "grad_norm": 1.2127298024220348, "learning_rate": 1.8048366393389427e-06, "loss": 0.6557, "step": 22408 }, { "epoch": 0.8116262223831945, "grad_norm": 1.2948698792871418, "learning_rate": 1.8041644624420462e-06, "loss": 0.666, "step": 22409 }, { "epoch": 0.8116624411445128, "grad_norm": 1.0560659900366993, "learning_rate": 1.803492398327046e-06, "loss": 0.6957, "step": 22410 }, { "epoch": 0.8116986599058312, "grad_norm": 1.3679523290566853, "learning_rate": 1.802820447003195e-06, "loss": 0.6372, "step": 22411 }, { "epoch": 0.8117348786671496, "grad_norm": 1.386678237054606, "learning_rate": 1.8021486084797368e-06, "loss": 0.6877, "step": 22412 }, { "epoch": 0.811771097428468, "grad_norm": 0.8849597210382164, "learning_rate": 1.8014768827659158e-06, "loss": 0.6486, "step": 22413 }, { "epoch": 0.8118073161897863, "grad_norm": 0.8607275676481131, "learning_rate": 1.8008052698709776e-06, "loss": 0.6811, "step": 22414 }, { "epoch": 0.8118435349511046, "grad_norm": 1.2910559860014745, "learning_rate": 1.8001337698041643e-06, "loss": 0.6779, "step": 22415 }, { "epoch": 0.811879753712423, "grad_norm": 1.404497888261994, "learning_rate": 1.7994623825747159e-06, "loss": 0.6842, "step": 22416 }, { "epoch": 0.8119159724737414, "grad_norm": 1.2252862047048412, "learning_rate": 1.7987911081918673e-06, "loss": 0.6805, "step": 22417 }, { "epoch": 0.8119521912350598, "grad_norm": 1.730541124401911, "learning_rate": 1.7981199466648613e-06, "loss": 0.6592, "step": 22418 }, { "epoch": 0.8119884099963781, "grad_norm": 1.0990092712838928, "learning_rate": 1.7974488980029325e-06, "loss": 0.6953, "step": 22419 }, { "epoch": 0.8120246287576964, "grad_norm": 1.1847487182233118, "learning_rate": 1.7967779622153147e-06, "loss": 0.6308, "step": 22420 }, { "epoch": 0.8120608475190149, "grad_norm": 1.3862208063394188, "learning_rate": 1.7961071393112372e-06, "loss": 0.7423, "step": 22421 }, { "epoch": 0.8120970662803332, "grad_norm": 1.1884346138032398, "learning_rate": 1.795436429299936e-06, "loss": 0.6317, "step": 22422 }, { "epoch": 0.8121332850416516, "grad_norm": 1.2619319033014735, "learning_rate": 1.7947658321906393e-06, "loss": 0.6239, "step": 22423 }, { "epoch": 0.8121695038029699, "grad_norm": 1.3893865109278711, "learning_rate": 1.7940953479925715e-06, "loss": 0.698, "step": 22424 }, { "epoch": 0.8122057225642882, "grad_norm": 1.390836994214132, "learning_rate": 1.7934249767149647e-06, "loss": 0.7, "step": 22425 }, { "epoch": 0.8122419413256067, "grad_norm": 1.5258952464945397, "learning_rate": 1.7927547183670413e-06, "loss": 0.6771, "step": 22426 }, { "epoch": 0.812278160086925, "grad_norm": 1.2828778046132954, "learning_rate": 1.7920845729580249e-06, "loss": 0.637, "step": 22427 }, { "epoch": 0.8123143788482434, "grad_norm": 1.425652357049676, "learning_rate": 1.791414540497134e-06, "loss": 0.7005, "step": 22428 }, { "epoch": 0.8123505976095617, "grad_norm": 1.0548578263872606, "learning_rate": 1.7907446209935941e-06, "loss": 0.6554, "step": 22429 }, { "epoch": 0.8123868163708801, "grad_norm": 1.3730539590319109, "learning_rate": 1.7900748144566216e-06, "loss": 0.6969, "step": 22430 }, { "epoch": 0.8124230351321985, "grad_norm": 1.447783050758388, "learning_rate": 1.7894051208954334e-06, "loss": 0.6971, "step": 22431 }, { "epoch": 0.8124592538935168, "grad_norm": 1.4119839935778384, "learning_rate": 1.788735540319243e-06, "loss": 0.7386, "step": 22432 }, { "epoch": 0.8124954726548352, "grad_norm": 1.407265661941974, "learning_rate": 1.788066072737269e-06, "loss": 0.6917, "step": 22433 }, { "epoch": 0.8125316914161536, "grad_norm": 1.4274329912384291, "learning_rate": 1.7873967181587215e-06, "loss": 0.6578, "step": 22434 }, { "epoch": 0.812567910177472, "grad_norm": 1.3255205184084276, "learning_rate": 1.7867274765928088e-06, "loss": 0.6411, "step": 22435 }, { "epoch": 0.8126041289387903, "grad_norm": 1.3136856777680883, "learning_rate": 1.7860583480487449e-06, "loss": 0.6234, "step": 22436 }, { "epoch": 0.8126403477001086, "grad_norm": 1.2980887574445357, "learning_rate": 1.7853893325357364e-06, "loss": 0.6641, "step": 22437 }, { "epoch": 0.812676566461427, "grad_norm": 1.2337400653428492, "learning_rate": 1.7847204300629873e-06, "loss": 0.6759, "step": 22438 }, { "epoch": 0.8127127852227454, "grad_norm": 0.968459192516696, "learning_rate": 1.7840516406397025e-06, "loss": 0.7137, "step": 22439 }, { "epoch": 0.8127490039840638, "grad_norm": 1.4191520616431221, "learning_rate": 1.7833829642750888e-06, "loss": 0.7171, "step": 22440 }, { "epoch": 0.8127852227453821, "grad_norm": 1.1178552077115955, "learning_rate": 1.782714400978345e-06, "loss": 0.6985, "step": 22441 }, { "epoch": 0.8128214415067004, "grad_norm": 1.2551763977806685, "learning_rate": 1.7820459507586708e-06, "loss": 0.6074, "step": 22442 }, { "epoch": 0.8128576602680189, "grad_norm": 1.0553538301534393, "learning_rate": 1.7813776136252626e-06, "loss": 0.6427, "step": 22443 }, { "epoch": 0.8128938790293372, "grad_norm": 1.0384363657475193, "learning_rate": 1.780709389587326e-06, "loss": 0.6386, "step": 22444 }, { "epoch": 0.8129300977906556, "grad_norm": 1.1196755896724884, "learning_rate": 1.7800412786540445e-06, "loss": 0.6762, "step": 22445 }, { "epoch": 0.8129663165519739, "grad_norm": 1.3954924174821706, "learning_rate": 1.7793732808346209e-06, "loss": 0.6861, "step": 22446 }, { "epoch": 0.8130025353132923, "grad_norm": 1.5542328771265723, "learning_rate": 1.778705396138244e-06, "loss": 0.7312, "step": 22447 }, { "epoch": 0.8130387540746107, "grad_norm": 1.1902657927335099, "learning_rate": 1.7780376245741038e-06, "loss": 0.7129, "step": 22448 }, { "epoch": 0.813074972835929, "grad_norm": 0.8873918326228047, "learning_rate": 1.7773699661513887e-06, "loss": 0.6714, "step": 22449 }, { "epoch": 0.8131111915972474, "grad_norm": 1.2920080408866665, "learning_rate": 1.7767024208792905e-06, "loss": 0.682, "step": 22450 }, { "epoch": 0.8131474103585657, "grad_norm": 1.0438854659927488, "learning_rate": 1.7760349887669925e-06, "loss": 0.6843, "step": 22451 }, { "epoch": 0.8131836291198841, "grad_norm": 1.3932026102232689, "learning_rate": 1.7753676698236789e-06, "loss": 0.6844, "step": 22452 }, { "epoch": 0.8132198478812025, "grad_norm": 1.129162341595889, "learning_rate": 1.7747004640585307e-06, "loss": 0.6662, "step": 22453 }, { "epoch": 0.8132560666425208, "grad_norm": 0.8912472249964237, "learning_rate": 1.7740333714807345e-06, "loss": 0.696, "step": 22454 }, { "epoch": 0.8132922854038392, "grad_norm": 1.3395785105333182, "learning_rate": 1.773366392099466e-06, "loss": 0.6663, "step": 22455 }, { "epoch": 0.8133285041651576, "grad_norm": 1.4717008757505665, "learning_rate": 1.7726995259239034e-06, "loss": 0.6663, "step": 22456 }, { "epoch": 0.8133647229264759, "grad_norm": 1.337395040931554, "learning_rate": 1.7720327729632258e-06, "loss": 0.5871, "step": 22457 }, { "epoch": 0.8134009416877943, "grad_norm": 1.2580430799804976, "learning_rate": 1.7713661332266075e-06, "loss": 0.61, "step": 22458 }, { "epoch": 0.8134371604491126, "grad_norm": 1.585259978734464, "learning_rate": 1.7706996067232208e-06, "loss": 0.6976, "step": 22459 }, { "epoch": 0.8134733792104311, "grad_norm": 1.3636709089502494, "learning_rate": 1.770033193462236e-06, "loss": 0.6346, "step": 22460 }, { "epoch": 0.8135095979717494, "grad_norm": 1.2630396127943293, "learning_rate": 1.7693668934528296e-06, "loss": 0.6206, "step": 22461 }, { "epoch": 0.8135458167330677, "grad_norm": 1.320262875555471, "learning_rate": 1.7687007067041661e-06, "loss": 0.6593, "step": 22462 }, { "epoch": 0.8135820354943861, "grad_norm": 1.4828269138781236, "learning_rate": 1.7680346332254128e-06, "loss": 0.688, "step": 22463 }, { "epoch": 0.8136182542557044, "grad_norm": 1.4066456880454503, "learning_rate": 1.7673686730257345e-06, "loss": 0.735, "step": 22464 }, { "epoch": 0.8136544730170229, "grad_norm": 1.426045817167007, "learning_rate": 1.7667028261142993e-06, "loss": 0.7321, "step": 22465 }, { "epoch": 0.8136906917783412, "grad_norm": 1.2985539103850126, "learning_rate": 1.7660370925002678e-06, "loss": 0.6877, "step": 22466 }, { "epoch": 0.8137269105396595, "grad_norm": 1.4332364219089395, "learning_rate": 1.7653714721927984e-06, "loss": 0.7363, "step": 22467 }, { "epoch": 0.8137631293009779, "grad_norm": 1.0002921841610035, "learning_rate": 1.7647059652010557e-06, "loss": 0.6681, "step": 22468 }, { "epoch": 0.8137993480622963, "grad_norm": 1.3639177807936582, "learning_rate": 1.7640405715341958e-06, "loss": 0.7093, "step": 22469 }, { "epoch": 0.8138355668236147, "grad_norm": 1.4597761630503994, "learning_rate": 1.7633752912013736e-06, "loss": 0.6635, "step": 22470 }, { "epoch": 0.813871785584933, "grad_norm": 1.390883565020714, "learning_rate": 1.7627101242117417e-06, "loss": 0.648, "step": 22471 }, { "epoch": 0.8139080043462513, "grad_norm": 1.2609949486466439, "learning_rate": 1.7620450705744597e-06, "loss": 0.6815, "step": 22472 }, { "epoch": 0.8139442231075698, "grad_norm": 1.2937756049230054, "learning_rate": 1.7613801302986756e-06, "loss": 0.6094, "step": 22473 }, { "epoch": 0.8139804418688881, "grad_norm": 0.9822922414533769, "learning_rate": 1.7607153033935398e-06, "loss": 0.6592, "step": 22474 }, { "epoch": 0.8140166606302065, "grad_norm": 1.4201755546313732, "learning_rate": 1.7600505898681996e-06, "loss": 0.7336, "step": 22475 }, { "epoch": 0.8140528793915248, "grad_norm": 1.3272415385316516, "learning_rate": 1.7593859897318055e-06, "loss": 0.6645, "step": 22476 }, { "epoch": 0.8140890981528431, "grad_norm": 1.300145178434719, "learning_rate": 1.7587215029935e-06, "loss": 0.6932, "step": 22477 }, { "epoch": 0.8141253169141616, "grad_norm": 1.329819602321271, "learning_rate": 1.7580571296624272e-06, "loss": 0.6487, "step": 22478 }, { "epoch": 0.8141615356754799, "grad_norm": 1.3944312855877539, "learning_rate": 1.7573928697477316e-06, "loss": 0.6318, "step": 22479 }, { "epoch": 0.8141977544367983, "grad_norm": 0.9798929237621882, "learning_rate": 1.7567287232585527e-06, "loss": 0.6925, "step": 22480 }, { "epoch": 0.8142339731981166, "grad_norm": 1.4648894875061729, "learning_rate": 1.7560646902040312e-06, "loss": 0.7283, "step": 22481 }, { "epoch": 0.814270191959435, "grad_norm": 1.5125961221936777, "learning_rate": 1.7554007705932996e-06, "loss": 0.5939, "step": 22482 }, { "epoch": 0.8143064107207534, "grad_norm": 1.3402882838101662, "learning_rate": 1.7547369644355007e-06, "loss": 0.6814, "step": 22483 }, { "epoch": 0.8143426294820717, "grad_norm": 1.3644264536112043, "learning_rate": 1.7540732717397657e-06, "loss": 0.6909, "step": 22484 }, { "epoch": 0.8143788482433901, "grad_norm": 1.2023353841898317, "learning_rate": 1.7534096925152277e-06, "loss": 0.6112, "step": 22485 }, { "epoch": 0.8144150670047084, "grad_norm": 1.3696203445738158, "learning_rate": 1.7527462267710161e-06, "loss": 0.7278, "step": 22486 }, { "epoch": 0.8144512857660268, "grad_norm": 1.3718873998281442, "learning_rate": 1.7520828745162666e-06, "loss": 0.611, "step": 22487 }, { "epoch": 0.8144875045273452, "grad_norm": 1.5161151621832982, "learning_rate": 1.7514196357601033e-06, "loss": 0.6816, "step": 22488 }, { "epoch": 0.8145237232886635, "grad_norm": 1.297392743960219, "learning_rate": 1.750756510511651e-06, "loss": 0.6555, "step": 22489 }, { "epoch": 0.8145599420499819, "grad_norm": 1.3752477568963681, "learning_rate": 1.75009349878004e-06, "loss": 0.591, "step": 22490 }, { "epoch": 0.8145961608113003, "grad_norm": 0.8550172659459582, "learning_rate": 1.749430600574391e-06, "loss": 0.6749, "step": 22491 }, { "epoch": 0.8146323795726186, "grad_norm": 1.444210929180836, "learning_rate": 1.7487678159038268e-06, "loss": 0.5874, "step": 22492 }, { "epoch": 0.814668598333937, "grad_norm": 1.4305231037938768, "learning_rate": 1.7481051447774656e-06, "loss": 0.6676, "step": 22493 }, { "epoch": 0.8147048170952553, "grad_norm": 1.4198266722864363, "learning_rate": 1.747442587204431e-06, "loss": 0.678, "step": 22494 }, { "epoch": 0.8147410358565738, "grad_norm": 1.3261139608442931, "learning_rate": 1.746780143193837e-06, "loss": 0.6859, "step": 22495 }, { "epoch": 0.8147772546178921, "grad_norm": 1.2735651587055883, "learning_rate": 1.7461178127547984e-06, "loss": 0.6428, "step": 22496 }, { "epoch": 0.8148134733792104, "grad_norm": 0.964523027536504, "learning_rate": 1.7454555958964336e-06, "loss": 0.6529, "step": 22497 }, { "epoch": 0.8148496921405288, "grad_norm": 1.3023613083873773, "learning_rate": 1.744793492627852e-06, "loss": 0.6544, "step": 22498 }, { "epoch": 0.8148859109018471, "grad_norm": 1.4193051880140373, "learning_rate": 1.7441315029581663e-06, "loss": 0.7253, "step": 22499 }, { "epoch": 0.8149221296631656, "grad_norm": 1.3618969276287987, "learning_rate": 1.7434696268964834e-06, "loss": 0.7263, "step": 22500 }, { "epoch": 0.8149583484244839, "grad_norm": 1.2425033925664235, "learning_rate": 1.7428078644519154e-06, "loss": 0.6573, "step": 22501 }, { "epoch": 0.8149945671858022, "grad_norm": 1.4188409525384515, "learning_rate": 1.7421462156335657e-06, "loss": 0.7132, "step": 22502 }, { "epoch": 0.8150307859471206, "grad_norm": 1.4921525881213298, "learning_rate": 1.7414846804505403e-06, "loss": 0.7303, "step": 22503 }, { "epoch": 0.815067004708439, "grad_norm": 1.5253129138273849, "learning_rate": 1.74082325891194e-06, "loss": 0.6764, "step": 22504 }, { "epoch": 0.8151032234697574, "grad_norm": 1.3224005327470514, "learning_rate": 1.740161951026872e-06, "loss": 0.6745, "step": 22505 }, { "epoch": 0.8151394422310757, "grad_norm": 1.1272230025574073, "learning_rate": 1.7395007568044331e-06, "loss": 0.672, "step": 22506 }, { "epoch": 0.815175660992394, "grad_norm": 0.8871159435912577, "learning_rate": 1.7388396762537185e-06, "loss": 0.6433, "step": 22507 }, { "epoch": 0.8152118797537125, "grad_norm": 1.3015575619783644, "learning_rate": 1.738178709383832e-06, "loss": 0.6643, "step": 22508 }, { "epoch": 0.8152480985150308, "grad_norm": 1.6427540611182931, "learning_rate": 1.7375178562038663e-06, "loss": 0.6551, "step": 22509 }, { "epoch": 0.8152843172763492, "grad_norm": 1.0038029024604094, "learning_rate": 1.7368571167229154e-06, "loss": 0.678, "step": 22510 }, { "epoch": 0.8153205360376675, "grad_norm": 1.3772113084542057, "learning_rate": 1.7361964909500673e-06, "loss": 0.7, "step": 22511 }, { "epoch": 0.8153567547989858, "grad_norm": 1.3838521994143789, "learning_rate": 1.7355359788944204e-06, "loss": 0.6981, "step": 22512 }, { "epoch": 0.8153929735603043, "grad_norm": 1.247841013824281, "learning_rate": 1.7348755805650597e-06, "loss": 0.6964, "step": 22513 }, { "epoch": 0.8154291923216226, "grad_norm": 1.3554588456300458, "learning_rate": 1.7342152959710735e-06, "loss": 0.6014, "step": 22514 }, { "epoch": 0.815465411082941, "grad_norm": 1.3204805870956502, "learning_rate": 1.7335551251215443e-06, "loss": 0.6217, "step": 22515 }, { "epoch": 0.8155016298442593, "grad_norm": 1.2747635658087617, "learning_rate": 1.7328950680255642e-06, "loss": 0.6719, "step": 22516 }, { "epoch": 0.8155378486055777, "grad_norm": 1.3232510612162103, "learning_rate": 1.7322351246922108e-06, "loss": 0.6618, "step": 22517 }, { "epoch": 0.8155740673668961, "grad_norm": 1.4846099253517306, "learning_rate": 1.7315752951305654e-06, "loss": 0.6872, "step": 22518 }, { "epoch": 0.8156102861282144, "grad_norm": 1.4711083845483315, "learning_rate": 1.7309155793497112e-06, "loss": 0.7491, "step": 22519 }, { "epoch": 0.8156465048895328, "grad_norm": 1.0661210877056797, "learning_rate": 1.7302559773587236e-06, "loss": 0.6799, "step": 22520 }, { "epoch": 0.8156827236508511, "grad_norm": 1.3840835314509523, "learning_rate": 1.7295964891666817e-06, "loss": 0.7752, "step": 22521 }, { "epoch": 0.8157189424121695, "grad_norm": 1.3179674908327286, "learning_rate": 1.7289371147826561e-06, "loss": 0.6425, "step": 22522 }, { "epoch": 0.8157551611734879, "grad_norm": 1.4454610152234006, "learning_rate": 1.7282778542157253e-06, "loss": 0.7205, "step": 22523 }, { "epoch": 0.8157913799348062, "grad_norm": 1.0814056672073118, "learning_rate": 1.7276187074749595e-06, "loss": 0.6327, "step": 22524 }, { "epoch": 0.8158275986961246, "grad_norm": 1.4819310397231389, "learning_rate": 1.7269596745694295e-06, "loss": 0.6778, "step": 22525 }, { "epoch": 0.815863817457443, "grad_norm": 1.4486914360411842, "learning_rate": 1.7263007555082011e-06, "loss": 0.7372, "step": 22526 }, { "epoch": 0.8159000362187613, "grad_norm": 1.373762891612815, "learning_rate": 1.7256419503003464e-06, "loss": 0.6939, "step": 22527 }, { "epoch": 0.8159362549800797, "grad_norm": 1.4656853693147958, "learning_rate": 1.7249832589549297e-06, "loss": 0.6738, "step": 22528 }, { "epoch": 0.815972473741398, "grad_norm": 1.2888868577100419, "learning_rate": 1.724324681481011e-06, "loss": 0.6803, "step": 22529 }, { "epoch": 0.8160086925027165, "grad_norm": 1.3268901666276653, "learning_rate": 1.7236662178876595e-06, "loss": 0.6915, "step": 22530 }, { "epoch": 0.8160449112640348, "grad_norm": 1.4357141890304228, "learning_rate": 1.7230078681839334e-06, "loss": 0.6684, "step": 22531 }, { "epoch": 0.8160811300253531, "grad_norm": 1.471811430412781, "learning_rate": 1.7223496323788912e-06, "loss": 0.7151, "step": 22532 }, { "epoch": 0.8161173487866715, "grad_norm": 1.343841824055611, "learning_rate": 1.721691510481589e-06, "loss": 0.5858, "step": 22533 }, { "epoch": 0.8161535675479898, "grad_norm": 1.4753375851099022, "learning_rate": 1.7210335025010883e-06, "loss": 0.6956, "step": 22534 }, { "epoch": 0.8161897863093083, "grad_norm": 0.9761207348775319, "learning_rate": 1.7203756084464407e-06, "loss": 0.7111, "step": 22535 }, { "epoch": 0.8162260050706266, "grad_norm": 1.5182664407524296, "learning_rate": 1.7197178283267003e-06, "loss": 0.7075, "step": 22536 }, { "epoch": 0.8162622238319449, "grad_norm": 0.9859245126007897, "learning_rate": 1.7190601621509151e-06, "loss": 0.6779, "step": 22537 }, { "epoch": 0.8162984425932633, "grad_norm": 1.3299118146676634, "learning_rate": 1.7184026099281414e-06, "loss": 0.6826, "step": 22538 }, { "epoch": 0.8163346613545817, "grad_norm": 1.3663819787458207, "learning_rate": 1.7177451716674242e-06, "loss": 0.6743, "step": 22539 }, { "epoch": 0.8163708801159001, "grad_norm": 1.083480520252531, "learning_rate": 1.7170878473778086e-06, "loss": 0.6626, "step": 22540 }, { "epoch": 0.8164070988772184, "grad_norm": 1.4866916902895846, "learning_rate": 1.7164306370683448e-06, "loss": 0.6538, "step": 22541 }, { "epoch": 0.8164433176385367, "grad_norm": 1.489443891813291, "learning_rate": 1.7157735407480746e-06, "loss": 0.714, "step": 22542 }, { "epoch": 0.8164795363998552, "grad_norm": 1.408314170109803, "learning_rate": 1.7151165584260388e-06, "loss": 0.6829, "step": 22543 }, { "epoch": 0.8165157551611735, "grad_norm": 1.257163480835731, "learning_rate": 1.7144596901112764e-06, "loss": 0.6376, "step": 22544 }, { "epoch": 0.8165519739224919, "grad_norm": 1.401406111492329, "learning_rate": 1.7138029358128316e-06, "loss": 0.7486, "step": 22545 }, { "epoch": 0.8165881926838102, "grad_norm": 1.3964370852913135, "learning_rate": 1.7131462955397405e-06, "loss": 0.71, "step": 22546 }, { "epoch": 0.8166244114451285, "grad_norm": 0.9688896403138004, "learning_rate": 1.7124897693010334e-06, "loss": 0.6788, "step": 22547 }, { "epoch": 0.816660630206447, "grad_norm": 1.4470457966710455, "learning_rate": 1.7118333571057532e-06, "loss": 0.6904, "step": 22548 }, { "epoch": 0.8166968489677653, "grad_norm": 1.460354409413396, "learning_rate": 1.7111770589629272e-06, "loss": 0.6941, "step": 22549 }, { "epoch": 0.8167330677290837, "grad_norm": 1.5233542517389016, "learning_rate": 1.7105208748815894e-06, "loss": 0.6868, "step": 22550 }, { "epoch": 0.816769286490402, "grad_norm": 1.4015000695684643, "learning_rate": 1.7098648048707644e-06, "loss": 0.623, "step": 22551 }, { "epoch": 0.8168055052517204, "grad_norm": 1.3407702884663877, "learning_rate": 1.7092088489394876e-06, "loss": 0.6062, "step": 22552 }, { "epoch": 0.8168417240130388, "grad_norm": 1.3688582906060245, "learning_rate": 1.7085530070967815e-06, "loss": 0.6665, "step": 22553 }, { "epoch": 0.8168779427743571, "grad_norm": 1.9231198144563246, "learning_rate": 1.7078972793516713e-06, "loss": 0.6505, "step": 22554 }, { "epoch": 0.8169141615356755, "grad_norm": 1.1845867276921596, "learning_rate": 1.7072416657131784e-06, "loss": 0.6306, "step": 22555 }, { "epoch": 0.8169503802969938, "grad_norm": 1.1723370499478547, "learning_rate": 1.7065861661903293e-06, "loss": 0.6851, "step": 22556 }, { "epoch": 0.8169865990583122, "grad_norm": 0.9735526733576007, "learning_rate": 1.7059307807921421e-06, "loss": 0.7132, "step": 22557 }, { "epoch": 0.8170228178196306, "grad_norm": 1.443513329055, "learning_rate": 1.7052755095276319e-06, "loss": 0.673, "step": 22558 }, { "epoch": 0.8170590365809489, "grad_norm": 1.2422354946821168, "learning_rate": 1.7046203524058225e-06, "loss": 0.6822, "step": 22559 }, { "epoch": 0.8170952553422673, "grad_norm": 1.1920039512961136, "learning_rate": 1.7039653094357256e-06, "loss": 0.7043, "step": 22560 }, { "epoch": 0.8171314741035857, "grad_norm": 1.4103995889902767, "learning_rate": 1.703310380626355e-06, "loss": 0.7627, "step": 22561 }, { "epoch": 0.817167692864904, "grad_norm": 0.9100507883402257, "learning_rate": 1.7026555659867217e-06, "loss": 0.6969, "step": 22562 }, { "epoch": 0.8172039116262224, "grad_norm": 1.4759644111980146, "learning_rate": 1.7020008655258413e-06, "loss": 0.6757, "step": 22563 }, { "epoch": 0.8172401303875407, "grad_norm": 1.3240907545318505, "learning_rate": 1.7013462792527203e-06, "loss": 0.6436, "step": 22564 }, { "epoch": 0.8172763491488592, "grad_norm": 1.3258189573079013, "learning_rate": 1.7006918071763657e-06, "loss": 0.6748, "step": 22565 }, { "epoch": 0.8173125679101775, "grad_norm": 1.4182999561684122, "learning_rate": 1.7000374493057825e-06, "loss": 0.6702, "step": 22566 }, { "epoch": 0.8173487866714958, "grad_norm": 1.3447463422662762, "learning_rate": 1.699383205649978e-06, "loss": 0.692, "step": 22567 }, { "epoch": 0.8173850054328142, "grad_norm": 1.2146651337551515, "learning_rate": 1.6987290762179553e-06, "loss": 0.6609, "step": 22568 }, { "epoch": 0.8174212241941325, "grad_norm": 1.0007136328720891, "learning_rate": 1.6980750610187124e-06, "loss": 0.6803, "step": 22569 }, { "epoch": 0.817457442955451, "grad_norm": 1.3627662744918014, "learning_rate": 1.6974211600612533e-06, "loss": 0.6692, "step": 22570 }, { "epoch": 0.8174936617167693, "grad_norm": 1.5121868595333599, "learning_rate": 1.6967673733545743e-06, "loss": 0.6706, "step": 22571 }, { "epoch": 0.8175298804780876, "grad_norm": 1.3185249959263798, "learning_rate": 1.6961137009076723e-06, "loss": 0.6808, "step": 22572 }, { "epoch": 0.817566099239406, "grad_norm": 1.3223246042854953, "learning_rate": 1.6954601427295403e-06, "loss": 0.6794, "step": 22573 }, { "epoch": 0.8176023180007244, "grad_norm": 1.4102960243853955, "learning_rate": 1.6948066988291756e-06, "loss": 0.6763, "step": 22574 }, { "epoch": 0.8176385367620428, "grad_norm": 1.1766230676203633, "learning_rate": 1.6941533692155675e-06, "loss": 0.6201, "step": 22575 }, { "epoch": 0.8176747555233611, "grad_norm": 1.0422221128298987, "learning_rate": 1.6935001538977092e-06, "loss": 0.715, "step": 22576 }, { "epoch": 0.8177109742846794, "grad_norm": 1.3808904007659497, "learning_rate": 1.692847052884583e-06, "loss": 0.6771, "step": 22577 }, { "epoch": 0.8177471930459979, "grad_norm": 1.4640259430798555, "learning_rate": 1.6921940661851844e-06, "loss": 0.6547, "step": 22578 }, { "epoch": 0.8177834118073162, "grad_norm": 1.2441388778434714, "learning_rate": 1.6915411938084947e-06, "loss": 0.6411, "step": 22579 }, { "epoch": 0.8178196305686346, "grad_norm": 1.9573031177629034, "learning_rate": 1.690888435763497e-06, "loss": 0.666, "step": 22580 }, { "epoch": 0.8178558493299529, "grad_norm": 1.322998067289508, "learning_rate": 1.6902357920591773e-06, "loss": 0.6898, "step": 22581 }, { "epoch": 0.8178920680912712, "grad_norm": 1.3098021365366097, "learning_rate": 1.6895832627045139e-06, "loss": 0.6444, "step": 22582 }, { "epoch": 0.8179282868525897, "grad_norm": 1.539829977739389, "learning_rate": 1.6889308477084887e-06, "loss": 0.6829, "step": 22583 }, { "epoch": 0.817964505613908, "grad_norm": 1.5557584141281975, "learning_rate": 1.6882785470800734e-06, "loss": 0.6557, "step": 22584 }, { "epoch": 0.8180007243752264, "grad_norm": 1.4876339968014705, "learning_rate": 1.6876263608282528e-06, "loss": 0.7818, "step": 22585 }, { "epoch": 0.8180369431365447, "grad_norm": 1.4875080598284245, "learning_rate": 1.6869742889619967e-06, "loss": 0.6425, "step": 22586 }, { "epoch": 0.818073161897863, "grad_norm": 1.3368708845289659, "learning_rate": 1.6863223314902788e-06, "loss": 0.6633, "step": 22587 }, { "epoch": 0.8181093806591815, "grad_norm": 1.4266893408270427, "learning_rate": 1.6856704884220675e-06, "loss": 0.6948, "step": 22588 }, { "epoch": 0.8181455994204998, "grad_norm": 1.5190092970213243, "learning_rate": 1.6850187597663392e-06, "loss": 0.6938, "step": 22589 }, { "epoch": 0.8181818181818182, "grad_norm": 1.3657723323337558, "learning_rate": 1.6843671455320587e-06, "loss": 0.5806, "step": 22590 }, { "epoch": 0.8182180369431366, "grad_norm": 1.3389080981814871, "learning_rate": 1.6837156457281912e-06, "loss": 0.6095, "step": 22591 }, { "epoch": 0.8182542557044549, "grad_norm": 1.4131137921607166, "learning_rate": 1.683064260363705e-06, "loss": 0.6809, "step": 22592 }, { "epoch": 0.8182904744657733, "grad_norm": 1.310035263222382, "learning_rate": 1.6824129894475637e-06, "loss": 0.6815, "step": 22593 }, { "epoch": 0.8183266932270916, "grad_norm": 1.4243088834169622, "learning_rate": 1.6817618329887276e-06, "loss": 0.7442, "step": 22594 }, { "epoch": 0.81836291198841, "grad_norm": 1.3460572448807357, "learning_rate": 1.6811107909961555e-06, "loss": 0.6385, "step": 22595 }, { "epoch": 0.8183991307497284, "grad_norm": 1.325987917567288, "learning_rate": 1.6804598634788106e-06, "loss": 0.6198, "step": 22596 }, { "epoch": 0.8184353495110467, "grad_norm": 1.328873857981378, "learning_rate": 1.6798090504456478e-06, "loss": 0.6286, "step": 22597 }, { "epoch": 0.8184715682723651, "grad_norm": 1.4757982772796012, "learning_rate": 1.6791583519056243e-06, "loss": 0.7175, "step": 22598 }, { "epoch": 0.8185077870336834, "grad_norm": 1.394612822831574, "learning_rate": 1.6785077678676898e-06, "loss": 0.6914, "step": 22599 }, { "epoch": 0.8185440057950019, "grad_norm": 1.5958904870740582, "learning_rate": 1.6778572983408026e-06, "loss": 0.6539, "step": 22600 }, { "epoch": 0.8185802245563202, "grad_norm": 1.3184432589370585, "learning_rate": 1.6772069433339121e-06, "loss": 0.5867, "step": 22601 }, { "epoch": 0.8186164433176385, "grad_norm": 1.4256866737247271, "learning_rate": 1.676556702855965e-06, "loss": 0.7533, "step": 22602 }, { "epoch": 0.8186526620789569, "grad_norm": 1.4933426365374531, "learning_rate": 1.6759065769159133e-06, "loss": 0.7083, "step": 22603 }, { "epoch": 0.8186888808402752, "grad_norm": 1.440322332745917, "learning_rate": 1.6752565655227015e-06, "loss": 0.6533, "step": 22604 }, { "epoch": 0.8187250996015937, "grad_norm": 1.3351348904541205, "learning_rate": 1.6746066686852734e-06, "loss": 0.6642, "step": 22605 }, { "epoch": 0.818761318362912, "grad_norm": 1.3963836740318787, "learning_rate": 1.673956886412571e-06, "loss": 0.6549, "step": 22606 }, { "epoch": 0.8187975371242303, "grad_norm": 1.2481969614778679, "learning_rate": 1.6733072187135402e-06, "loss": 0.7192, "step": 22607 }, { "epoch": 0.8188337558855487, "grad_norm": 1.4728320709929845, "learning_rate": 1.6726576655971184e-06, "loss": 0.6959, "step": 22608 }, { "epoch": 0.8188699746468671, "grad_norm": 1.296185540000846, "learning_rate": 1.6720082270722415e-06, "loss": 0.6369, "step": 22609 }, { "epoch": 0.8189061934081855, "grad_norm": 1.307116013592652, "learning_rate": 1.6713589031478516e-06, "loss": 0.6286, "step": 22610 }, { "epoch": 0.8189424121695038, "grad_norm": 1.1998817785654892, "learning_rate": 1.6707096938328814e-06, "loss": 0.6906, "step": 22611 }, { "epoch": 0.8189786309308221, "grad_norm": 1.130387408337565, "learning_rate": 1.670060599136265e-06, "loss": 0.7041, "step": 22612 }, { "epoch": 0.8190148496921406, "grad_norm": 0.9684188875603265, "learning_rate": 1.6694116190669319e-06, "loss": 0.6805, "step": 22613 }, { "epoch": 0.8190510684534589, "grad_norm": 1.6809760728928913, "learning_rate": 1.6687627536338158e-06, "loss": 0.6349, "step": 22614 }, { "epoch": 0.8190872872147773, "grad_norm": 1.4543166248129098, "learning_rate": 1.6681140028458464e-06, "loss": 0.6401, "step": 22615 }, { "epoch": 0.8191235059760956, "grad_norm": 1.383071077815204, "learning_rate": 1.6674653667119478e-06, "loss": 0.7069, "step": 22616 }, { "epoch": 0.8191597247374139, "grad_norm": 1.2966762492402475, "learning_rate": 1.666816845241046e-06, "loss": 0.6503, "step": 22617 }, { "epoch": 0.8191959434987324, "grad_norm": 1.8244723076400595, "learning_rate": 1.6661684384420685e-06, "loss": 0.7302, "step": 22618 }, { "epoch": 0.8192321622600507, "grad_norm": 0.9378435059090094, "learning_rate": 1.6655201463239367e-06, "loss": 0.6985, "step": 22619 }, { "epoch": 0.8192683810213691, "grad_norm": 0.9788575852357766, "learning_rate": 1.6648719688955672e-06, "loss": 0.6755, "step": 22620 }, { "epoch": 0.8193045997826874, "grad_norm": 1.2433436998870706, "learning_rate": 1.6642239061658872e-06, "loss": 0.6278, "step": 22621 }, { "epoch": 0.8193408185440058, "grad_norm": 0.9595952107998837, "learning_rate": 1.6635759581438093e-06, "loss": 0.6629, "step": 22622 }, { "epoch": 0.8193770373053242, "grad_norm": 1.2005253541033207, "learning_rate": 1.6629281248382523e-06, "loss": 0.7111, "step": 22623 }, { "epoch": 0.8194132560666425, "grad_norm": 1.3110918853250062, "learning_rate": 1.6622804062581276e-06, "loss": 0.6127, "step": 22624 }, { "epoch": 0.8194494748279609, "grad_norm": 1.4033859877326536, "learning_rate": 1.6616328024123529e-06, "loss": 0.719, "step": 22625 }, { "epoch": 0.8194856935892793, "grad_norm": 1.645416261746661, "learning_rate": 1.6609853133098375e-06, "loss": 0.7912, "step": 22626 }, { "epoch": 0.8195219123505976, "grad_norm": 1.2887175980013343, "learning_rate": 1.660337938959491e-06, "loss": 0.6228, "step": 22627 }, { "epoch": 0.819558131111916, "grad_norm": 1.3176713306586778, "learning_rate": 1.6596906793702205e-06, "loss": 0.6754, "step": 22628 }, { "epoch": 0.8195943498732343, "grad_norm": 1.0491540293859776, "learning_rate": 1.659043534550936e-06, "loss": 0.6029, "step": 22629 }, { "epoch": 0.8196305686345527, "grad_norm": 1.4322379337693272, "learning_rate": 1.6583965045105433e-06, "loss": 0.7056, "step": 22630 }, { "epoch": 0.8196667873958711, "grad_norm": 1.358601686559773, "learning_rate": 1.6577495892579398e-06, "loss": 0.596, "step": 22631 }, { "epoch": 0.8197030061571894, "grad_norm": 1.0148434365547714, "learning_rate": 1.6571027888020364e-06, "loss": 0.656, "step": 22632 }, { "epoch": 0.8197392249185078, "grad_norm": 1.33633320299247, "learning_rate": 1.6564561031517278e-06, "loss": 0.6564, "step": 22633 }, { "epoch": 0.8197754436798261, "grad_norm": 1.292803194072694, "learning_rate": 1.6558095323159161e-06, "loss": 0.661, "step": 22634 }, { "epoch": 0.8198116624411446, "grad_norm": 1.2037743384737511, "learning_rate": 1.6551630763034931e-06, "loss": 0.6535, "step": 22635 }, { "epoch": 0.8198478812024629, "grad_norm": 1.4379003358641014, "learning_rate": 1.6545167351233615e-06, "loss": 0.7494, "step": 22636 }, { "epoch": 0.8198840999637812, "grad_norm": 1.464427036776488, "learning_rate": 1.6538705087844132e-06, "loss": 0.7209, "step": 22637 }, { "epoch": 0.8199203187250996, "grad_norm": 1.354949691697859, "learning_rate": 1.6532243972955397e-06, "loss": 0.6375, "step": 22638 }, { "epoch": 0.819956537486418, "grad_norm": 1.074265727890767, "learning_rate": 1.652578400665631e-06, "loss": 0.6774, "step": 22639 }, { "epoch": 0.8199927562477364, "grad_norm": 1.2830989895549678, "learning_rate": 1.6519325189035796e-06, "loss": 0.6608, "step": 22640 }, { "epoch": 0.8200289750090547, "grad_norm": 1.1073906292157227, "learning_rate": 1.651286752018273e-06, "loss": 0.6446, "step": 22641 }, { "epoch": 0.820065193770373, "grad_norm": 1.4230222815609164, "learning_rate": 1.650641100018593e-06, "loss": 0.647, "step": 22642 }, { "epoch": 0.8201014125316914, "grad_norm": 1.2652158112502292, "learning_rate": 1.64999556291343e-06, "loss": 0.6246, "step": 22643 }, { "epoch": 0.8201376312930098, "grad_norm": 1.0499723932580165, "learning_rate": 1.6493501407116663e-06, "loss": 0.6112, "step": 22644 }, { "epoch": 0.8201738500543282, "grad_norm": 1.1443460151534137, "learning_rate": 1.6487048334221812e-06, "loss": 0.6933, "step": 22645 }, { "epoch": 0.8202100688156465, "grad_norm": 0.9626287856084597, "learning_rate": 1.6480596410538541e-06, "loss": 0.7388, "step": 22646 }, { "epoch": 0.8202462875769648, "grad_norm": 0.9536226153283137, "learning_rate": 1.6474145636155669e-06, "loss": 0.6487, "step": 22647 }, { "epoch": 0.8202825063382833, "grad_norm": 1.295733370811037, "learning_rate": 1.646769601116195e-06, "loss": 0.6714, "step": 22648 }, { "epoch": 0.8203187250996016, "grad_norm": 1.1729215115517486, "learning_rate": 1.646124753564613e-06, "loss": 0.6317, "step": 22649 }, { "epoch": 0.82035494386092, "grad_norm": 1.3504021423682913, "learning_rate": 1.645480020969692e-06, "loss": 0.6669, "step": 22650 }, { "epoch": 0.8203911626222383, "grad_norm": 1.3515590181195607, "learning_rate": 1.6448354033403091e-06, "loss": 0.6668, "step": 22651 }, { "epoch": 0.8204273813835566, "grad_norm": 1.3244298876329001, "learning_rate": 1.644190900685333e-06, "loss": 0.6876, "step": 22652 }, { "epoch": 0.8204636001448751, "grad_norm": 1.740511851768037, "learning_rate": 1.6435465130136297e-06, "loss": 0.6704, "step": 22653 }, { "epoch": 0.8204998189061934, "grad_norm": 1.4329250657553776, "learning_rate": 1.6429022403340711e-06, "loss": 0.6884, "step": 22654 }, { "epoch": 0.8205360376675118, "grad_norm": 1.4235636448474338, "learning_rate": 1.6422580826555234e-06, "loss": 0.6444, "step": 22655 }, { "epoch": 0.8205722564288301, "grad_norm": 1.423182914942398, "learning_rate": 1.641614039986844e-06, "loss": 0.7672, "step": 22656 }, { "epoch": 0.8206084751901485, "grad_norm": 1.3612318415367692, "learning_rate": 1.6409701123369015e-06, "loss": 0.6672, "step": 22657 }, { "epoch": 0.8206446939514669, "grad_norm": 1.2945099437217242, "learning_rate": 1.6403262997145553e-06, "loss": 0.729, "step": 22658 }, { "epoch": 0.8206809127127852, "grad_norm": 1.0219820083838644, "learning_rate": 1.6396826021286649e-06, "loss": 0.7146, "step": 22659 }, { "epoch": 0.8207171314741036, "grad_norm": 1.1450620926997062, "learning_rate": 1.6390390195880845e-06, "loss": 0.6328, "step": 22660 }, { "epoch": 0.820753350235422, "grad_norm": 1.3827665551683352, "learning_rate": 1.638395552101677e-06, "loss": 0.6466, "step": 22661 }, { "epoch": 0.8207895689967403, "grad_norm": 1.4688970411028428, "learning_rate": 1.6377521996782942e-06, "loss": 0.6431, "step": 22662 }, { "epoch": 0.8208257877580587, "grad_norm": 1.4016684710055942, "learning_rate": 1.6371089623267856e-06, "loss": 0.7129, "step": 22663 }, { "epoch": 0.820862006519377, "grad_norm": 1.3875897848606855, "learning_rate": 1.6364658400560085e-06, "loss": 0.6757, "step": 22664 }, { "epoch": 0.8208982252806954, "grad_norm": 1.201641154691221, "learning_rate": 1.6358228328748115e-06, "loss": 0.7121, "step": 22665 }, { "epoch": 0.8209344440420138, "grad_norm": 1.2836497620378617, "learning_rate": 1.6351799407920399e-06, "loss": 0.6566, "step": 22666 }, { "epoch": 0.8209706628033321, "grad_norm": 1.4143002070539568, "learning_rate": 1.6345371638165408e-06, "loss": 0.6712, "step": 22667 }, { "epoch": 0.8210068815646505, "grad_norm": 1.0302920803647762, "learning_rate": 1.633894501957164e-06, "loss": 0.6611, "step": 22668 }, { "epoch": 0.8210431003259688, "grad_norm": 1.0532723539111062, "learning_rate": 1.633251955222749e-06, "loss": 0.6509, "step": 22669 }, { "epoch": 0.8210793190872873, "grad_norm": 1.4644084983145467, "learning_rate": 1.6326095236221385e-06, "loss": 0.7058, "step": 22670 }, { "epoch": 0.8211155378486056, "grad_norm": 1.5306855287710133, "learning_rate": 1.6319672071641712e-06, "loss": 0.7195, "step": 22671 }, { "epoch": 0.8211517566099239, "grad_norm": 1.3187158568875355, "learning_rate": 1.6313250058576913e-06, "loss": 0.6358, "step": 22672 }, { "epoch": 0.8211879753712423, "grad_norm": 1.3734190858250979, "learning_rate": 1.6306829197115315e-06, "loss": 0.6479, "step": 22673 }, { "epoch": 0.8212241941325606, "grad_norm": 1.3913925549338113, "learning_rate": 1.630040948734526e-06, "loss": 0.6777, "step": 22674 }, { "epoch": 0.8212604128938791, "grad_norm": 1.2817713684171077, "learning_rate": 1.6293990929355152e-06, "loss": 0.6475, "step": 22675 }, { "epoch": 0.8212966316551974, "grad_norm": 1.4391751680939175, "learning_rate": 1.628757352323327e-06, "loss": 0.6932, "step": 22676 }, { "epoch": 0.8213328504165157, "grad_norm": 1.4601794745913297, "learning_rate": 1.6281157269067937e-06, "loss": 0.6736, "step": 22677 }, { "epoch": 0.8213690691778341, "grad_norm": 1.1989696947671675, "learning_rate": 1.627474216694741e-06, "loss": 0.6375, "step": 22678 }, { "epoch": 0.8214052879391525, "grad_norm": 1.332077408016017, "learning_rate": 1.6268328216960027e-06, "loss": 0.6357, "step": 22679 }, { "epoch": 0.8214415067004709, "grad_norm": 1.291015419765231, "learning_rate": 1.6261915419194019e-06, "loss": 0.6428, "step": 22680 }, { "epoch": 0.8214777254617892, "grad_norm": 1.0120385089962598, "learning_rate": 1.6255503773737635e-06, "loss": 0.6586, "step": 22681 }, { "epoch": 0.8215139442231075, "grad_norm": 1.3576198639319892, "learning_rate": 1.6249093280679084e-06, "loss": 0.6823, "step": 22682 }, { "epoch": 0.821550162984426, "grad_norm": 1.0104863860424538, "learning_rate": 1.6242683940106618e-06, "loss": 0.6886, "step": 22683 }, { "epoch": 0.8215863817457443, "grad_norm": 1.2358565681679332, "learning_rate": 1.6236275752108421e-06, "loss": 0.6215, "step": 22684 }, { "epoch": 0.8216226005070627, "grad_norm": 1.0622539857792184, "learning_rate": 1.6229868716772646e-06, "loss": 0.7077, "step": 22685 }, { "epoch": 0.821658819268381, "grad_norm": 1.378863220516051, "learning_rate": 1.6223462834187508e-06, "loss": 0.6249, "step": 22686 }, { "epoch": 0.8216950380296993, "grad_norm": 1.1705665698708598, "learning_rate": 1.6217058104441131e-06, "loss": 0.6147, "step": 22687 }, { "epoch": 0.8217312567910178, "grad_norm": 1.4012293651475385, "learning_rate": 1.621065452762165e-06, "loss": 0.6737, "step": 22688 }, { "epoch": 0.8217674755523361, "grad_norm": 1.3850779685541548, "learning_rate": 1.6204252103817175e-06, "loss": 0.6737, "step": 22689 }, { "epoch": 0.8218036943136545, "grad_norm": 1.4762901705937905, "learning_rate": 1.6197850833115835e-06, "loss": 0.6986, "step": 22690 }, { "epoch": 0.8218399130749728, "grad_norm": 1.331787966020765, "learning_rate": 1.6191450715605706e-06, "loss": 0.6642, "step": 22691 }, { "epoch": 0.8218761318362912, "grad_norm": 1.4840963999404135, "learning_rate": 1.6185051751374847e-06, "loss": 0.6459, "step": 22692 }, { "epoch": 0.8219123505976096, "grad_norm": 1.5105409210715135, "learning_rate": 1.6178653940511301e-06, "loss": 0.7278, "step": 22693 }, { "epoch": 0.8219485693589279, "grad_norm": 1.3754249059819301, "learning_rate": 1.6172257283103166e-06, "loss": 0.6873, "step": 22694 }, { "epoch": 0.8219847881202463, "grad_norm": 1.0864449708829742, "learning_rate": 1.616586177923841e-06, "loss": 0.6856, "step": 22695 }, { "epoch": 0.8220210068815647, "grad_norm": 1.344494717238511, "learning_rate": 1.6159467429005049e-06, "loss": 0.6459, "step": 22696 }, { "epoch": 0.822057225642883, "grad_norm": 0.9606381016293405, "learning_rate": 1.6153074232491106e-06, "loss": 0.656, "step": 22697 }, { "epoch": 0.8220934444042014, "grad_norm": 1.1145802161197842, "learning_rate": 1.6146682189784523e-06, "loss": 0.6488, "step": 22698 }, { "epoch": 0.8221296631655197, "grad_norm": 1.4439058766190676, "learning_rate": 1.6140291300973287e-06, "loss": 0.654, "step": 22699 }, { "epoch": 0.8221658819268381, "grad_norm": 1.327104510211208, "learning_rate": 1.6133901566145304e-06, "loss": 0.7351, "step": 22700 }, { "epoch": 0.8222021006881565, "grad_norm": 1.3107092212458604, "learning_rate": 1.612751298538855e-06, "loss": 0.6753, "step": 22701 }, { "epoch": 0.8222383194494748, "grad_norm": 1.4415151611781882, "learning_rate": 1.6121125558790907e-06, "loss": 0.6351, "step": 22702 }, { "epoch": 0.8222745382107932, "grad_norm": 1.4788541677093248, "learning_rate": 1.6114739286440284e-06, "loss": 0.7397, "step": 22703 }, { "epoch": 0.8223107569721115, "grad_norm": 1.0232316373279182, "learning_rate": 1.6108354168424533e-06, "loss": 0.6617, "step": 22704 }, { "epoch": 0.82234697573343, "grad_norm": 1.3913541339621893, "learning_rate": 1.6101970204831563e-06, "loss": 0.6431, "step": 22705 }, { "epoch": 0.8223831944947483, "grad_norm": 1.527319389646563, "learning_rate": 1.6095587395749202e-06, "loss": 0.6751, "step": 22706 }, { "epoch": 0.8224194132560666, "grad_norm": 1.3196561609684732, "learning_rate": 1.6089205741265269e-06, "loss": 0.6108, "step": 22707 }, { "epoch": 0.822455632017385, "grad_norm": 1.3099296568933283, "learning_rate": 1.6082825241467614e-06, "loss": 0.6134, "step": 22708 }, { "epoch": 0.8224918507787033, "grad_norm": 1.4381814441572887, "learning_rate": 1.6076445896444014e-06, "loss": 0.6806, "step": 22709 }, { "epoch": 0.8225280695400218, "grad_norm": 1.4256616653301823, "learning_rate": 1.6070067706282266e-06, "loss": 0.6793, "step": 22710 }, { "epoch": 0.8225642883013401, "grad_norm": 1.4290921105455054, "learning_rate": 1.6063690671070109e-06, "loss": 0.6668, "step": 22711 }, { "epoch": 0.8226005070626584, "grad_norm": 1.5305496640842253, "learning_rate": 1.605731479089534e-06, "loss": 0.6834, "step": 22712 }, { "epoch": 0.8226367258239768, "grad_norm": 1.512977800288025, "learning_rate": 1.6050940065845677e-06, "loss": 0.6962, "step": 22713 }, { "epoch": 0.8226729445852952, "grad_norm": 1.2015744081429214, "learning_rate": 1.6044566496008829e-06, "loss": 0.6427, "step": 22714 }, { "epoch": 0.8227091633466136, "grad_norm": 1.0774430119556868, "learning_rate": 1.6038194081472536e-06, "loss": 0.6961, "step": 22715 }, { "epoch": 0.8227453821079319, "grad_norm": 1.262931583674256, "learning_rate": 1.603182282232446e-06, "loss": 0.6536, "step": 22716 }, { "epoch": 0.8227816008692502, "grad_norm": 1.4189569869220815, "learning_rate": 1.602545271865229e-06, "loss": 0.7012, "step": 22717 }, { "epoch": 0.8228178196305687, "grad_norm": 1.3130971594595797, "learning_rate": 1.6019083770543654e-06, "loss": 0.6943, "step": 22718 }, { "epoch": 0.822854038391887, "grad_norm": 1.3973289743331367, "learning_rate": 1.6012715978086247e-06, "loss": 0.6517, "step": 22719 }, { "epoch": 0.8228902571532054, "grad_norm": 1.288557800558424, "learning_rate": 1.6006349341367656e-06, "loss": 0.6694, "step": 22720 }, { "epoch": 0.8229264759145237, "grad_norm": 1.055315495244802, "learning_rate": 1.5999983860475499e-06, "loss": 0.7339, "step": 22721 }, { "epoch": 0.822962694675842, "grad_norm": 1.3432666580181492, "learning_rate": 1.599361953549735e-06, "loss": 0.6873, "step": 22722 }, { "epoch": 0.8229989134371605, "grad_norm": 1.3878503702071552, "learning_rate": 1.5987256366520842e-06, "loss": 0.7186, "step": 22723 }, { "epoch": 0.8230351321984788, "grad_norm": 1.3148475065251326, "learning_rate": 1.5980894353633503e-06, "loss": 0.6842, "step": 22724 }, { "epoch": 0.8230713509597972, "grad_norm": 1.5923998316764565, "learning_rate": 1.5974533496922862e-06, "loss": 0.7093, "step": 22725 }, { "epoch": 0.8231075697211155, "grad_norm": 1.1528264368823078, "learning_rate": 1.5968173796476494e-06, "loss": 0.63, "step": 22726 }, { "epoch": 0.8231437884824339, "grad_norm": 1.449961998663524, "learning_rate": 1.5961815252381885e-06, "loss": 0.6823, "step": 22727 }, { "epoch": 0.8231800072437523, "grad_norm": 1.6648917728731691, "learning_rate": 1.5955457864726553e-06, "loss": 0.7305, "step": 22728 }, { "epoch": 0.8232162260050706, "grad_norm": 1.4360366712326782, "learning_rate": 1.594910163359793e-06, "loss": 0.6559, "step": 22729 }, { "epoch": 0.823252444766389, "grad_norm": 1.3648307241769182, "learning_rate": 1.5942746559083556e-06, "loss": 0.7087, "step": 22730 }, { "epoch": 0.8232886635277074, "grad_norm": 1.0858859423663936, "learning_rate": 1.5936392641270848e-06, "loss": 0.7129, "step": 22731 }, { "epoch": 0.8233248822890257, "grad_norm": 1.341644120856835, "learning_rate": 1.593003988024724e-06, "loss": 0.6272, "step": 22732 }, { "epoch": 0.8233611010503441, "grad_norm": 1.4577041649607758, "learning_rate": 1.5923688276100125e-06, "loss": 0.7367, "step": 22733 }, { "epoch": 0.8233973198116624, "grad_norm": 1.2725271147750732, "learning_rate": 1.5917337828916969e-06, "loss": 0.6725, "step": 22734 }, { "epoch": 0.8234335385729808, "grad_norm": 1.3596691081995684, "learning_rate": 1.5910988538785122e-06, "loss": 0.6006, "step": 22735 }, { "epoch": 0.8234697573342992, "grad_norm": 1.4673667491499223, "learning_rate": 1.5904640405791927e-06, "loss": 0.71, "step": 22736 }, { "epoch": 0.8235059760956175, "grad_norm": 1.2335243961172906, "learning_rate": 1.5898293430024804e-06, "loss": 0.6247, "step": 22737 }, { "epoch": 0.8235421948569359, "grad_norm": 1.3209508447489677, "learning_rate": 1.5891947611571067e-06, "loss": 0.6985, "step": 22738 }, { "epoch": 0.8235784136182542, "grad_norm": 1.2931642384565725, "learning_rate": 1.588560295051803e-06, "loss": 0.6489, "step": 22739 }, { "epoch": 0.8236146323795727, "grad_norm": 1.3004063353788404, "learning_rate": 1.5879259446952978e-06, "loss": 0.6646, "step": 22740 }, { "epoch": 0.823650851140891, "grad_norm": 1.608955059423474, "learning_rate": 1.5872917100963259e-06, "loss": 0.6968, "step": 22741 }, { "epoch": 0.8236870699022093, "grad_norm": 1.3642210795018141, "learning_rate": 1.5866575912636117e-06, "loss": 0.7448, "step": 22742 }, { "epoch": 0.8237232886635277, "grad_norm": 2.074082612564764, "learning_rate": 1.5860235882058827e-06, "loss": 0.6373, "step": 22743 }, { "epoch": 0.823759507424846, "grad_norm": 1.000879261881873, "learning_rate": 1.5853897009318586e-06, "loss": 0.6408, "step": 22744 }, { "epoch": 0.8237957261861645, "grad_norm": 1.4771049086055859, "learning_rate": 1.584755929450269e-06, "loss": 0.6955, "step": 22745 }, { "epoch": 0.8238319449474828, "grad_norm": 1.3703204933333544, "learning_rate": 1.5841222737698325e-06, "loss": 0.6249, "step": 22746 }, { "epoch": 0.8238681637088011, "grad_norm": 1.3721842552197516, "learning_rate": 1.5834887338992655e-06, "loss": 0.6977, "step": 22747 }, { "epoch": 0.8239043824701195, "grad_norm": 1.4619540997095608, "learning_rate": 1.582855309847291e-06, "loss": 0.6884, "step": 22748 }, { "epoch": 0.8239406012314379, "grad_norm": 1.333800870133975, "learning_rate": 1.5822220016226242e-06, "loss": 0.6862, "step": 22749 }, { "epoch": 0.8239768199927563, "grad_norm": 0.9606786904917679, "learning_rate": 1.5815888092339782e-06, "loss": 0.6765, "step": 22750 }, { "epoch": 0.8240130387540746, "grad_norm": 1.0647966820230044, "learning_rate": 1.580955732690065e-06, "loss": 0.6873, "step": 22751 }, { "epoch": 0.8240492575153929, "grad_norm": 1.5470213056261477, "learning_rate": 1.5803227719996016e-06, "loss": 0.6941, "step": 22752 }, { "epoch": 0.8240854762767114, "grad_norm": 1.3811487895469676, "learning_rate": 1.579689927171294e-06, "loss": 0.6617, "step": 22753 }, { "epoch": 0.8241216950380297, "grad_norm": 1.2747115010982681, "learning_rate": 1.5790571982138525e-06, "loss": 0.6352, "step": 22754 }, { "epoch": 0.8241579137993481, "grad_norm": 1.3893688118947407, "learning_rate": 1.5784245851359813e-06, "loss": 0.6648, "step": 22755 }, { "epoch": 0.8241941325606664, "grad_norm": 1.0046023865771785, "learning_rate": 1.5777920879463892e-06, "loss": 0.6714, "step": 22756 }, { "epoch": 0.8242303513219847, "grad_norm": 1.55734653692531, "learning_rate": 1.577159706653779e-06, "loss": 0.7025, "step": 22757 }, { "epoch": 0.8242665700833032, "grad_norm": 1.2578919845398622, "learning_rate": 1.5765274412668497e-06, "loss": 0.6718, "step": 22758 }, { "epoch": 0.8243027888446215, "grad_norm": 0.9353831630519631, "learning_rate": 1.575895291794306e-06, "loss": 0.6506, "step": 22759 }, { "epoch": 0.8243390076059399, "grad_norm": 1.3856340114746122, "learning_rate": 1.5752632582448458e-06, "loss": 0.6371, "step": 22760 }, { "epoch": 0.8243752263672582, "grad_norm": 1.3680421787185928, "learning_rate": 1.5746313406271651e-06, "loss": 0.6757, "step": 22761 }, { "epoch": 0.8244114451285766, "grad_norm": 1.4075334890609537, "learning_rate": 1.5739995389499586e-06, "loss": 0.6785, "step": 22762 }, { "epoch": 0.824447663889895, "grad_norm": 1.476534517719396, "learning_rate": 1.5733678532219243e-06, "loss": 0.6479, "step": 22763 }, { "epoch": 0.8244838826512133, "grad_norm": 1.369568884991253, "learning_rate": 1.5727362834517534e-06, "loss": 0.6913, "step": 22764 }, { "epoch": 0.8245201014125317, "grad_norm": 1.298765264151553, "learning_rate": 1.5721048296481356e-06, "loss": 0.6726, "step": 22765 }, { "epoch": 0.82455632017385, "grad_norm": 1.427843100169166, "learning_rate": 1.5714734918197572e-06, "loss": 0.6757, "step": 22766 }, { "epoch": 0.8245925389351684, "grad_norm": 1.3987006460647453, "learning_rate": 1.5708422699753135e-06, "loss": 0.6897, "step": 22767 }, { "epoch": 0.8246287576964868, "grad_norm": 1.4846182786951259, "learning_rate": 1.5702111641234863e-06, "loss": 0.6808, "step": 22768 }, { "epoch": 0.8246649764578051, "grad_norm": 1.389229104018807, "learning_rate": 1.5695801742729577e-06, "loss": 0.6596, "step": 22769 }, { "epoch": 0.8247011952191236, "grad_norm": 1.1374152952236178, "learning_rate": 1.5689493004324163e-06, "loss": 0.6081, "step": 22770 }, { "epoch": 0.8247374139804419, "grad_norm": 1.3899339575599086, "learning_rate": 1.5683185426105397e-06, "loss": 0.6848, "step": 22771 }, { "epoch": 0.8247736327417602, "grad_norm": 1.4459004829966302, "learning_rate": 1.5676879008160095e-06, "loss": 0.6842, "step": 22772 }, { "epoch": 0.8248098515030786, "grad_norm": 0.9594810569191584, "learning_rate": 1.5670573750575013e-06, "loss": 0.6559, "step": 22773 }, { "epoch": 0.8248460702643969, "grad_norm": 1.3007089778079488, "learning_rate": 1.5664269653436947e-06, "loss": 0.6201, "step": 22774 }, { "epoch": 0.8248822890257154, "grad_norm": 1.344933549994839, "learning_rate": 1.565796671683264e-06, "loss": 0.7072, "step": 22775 }, { "epoch": 0.8249185077870337, "grad_norm": 1.2988551932934629, "learning_rate": 1.5651664940848798e-06, "loss": 0.6745, "step": 22776 }, { "epoch": 0.824954726548352, "grad_norm": 1.29190884967966, "learning_rate": 1.5645364325572188e-06, "loss": 0.6993, "step": 22777 }, { "epoch": 0.8249909453096704, "grad_norm": 1.37172104880291, "learning_rate": 1.5639064871089472e-06, "loss": 0.701, "step": 22778 }, { "epoch": 0.8250271640709887, "grad_norm": 1.3625482539722715, "learning_rate": 1.5632766577487357e-06, "loss": 0.6594, "step": 22779 }, { "epoch": 0.8250633828323072, "grad_norm": 1.3799715209329628, "learning_rate": 1.5626469444852488e-06, "loss": 0.5877, "step": 22780 }, { "epoch": 0.8250996015936255, "grad_norm": 1.4353300281154537, "learning_rate": 1.562017347327155e-06, "loss": 0.7432, "step": 22781 }, { "epoch": 0.8251358203549438, "grad_norm": 1.3299045547520978, "learning_rate": 1.5613878662831172e-06, "loss": 0.6188, "step": 22782 }, { "epoch": 0.8251720391162622, "grad_norm": 1.589729483057364, "learning_rate": 1.5607585013617977e-06, "loss": 0.6676, "step": 22783 }, { "epoch": 0.8252082578775806, "grad_norm": 1.0661424337370582, "learning_rate": 1.5601292525718537e-06, "loss": 0.6617, "step": 22784 }, { "epoch": 0.825244476638899, "grad_norm": 1.5157398408955614, "learning_rate": 1.5595001199219496e-06, "loss": 0.7052, "step": 22785 }, { "epoch": 0.8252806954002173, "grad_norm": 0.9907378969435002, "learning_rate": 1.5588711034207404e-06, "loss": 0.6619, "step": 22786 }, { "epoch": 0.8253169141615356, "grad_norm": 1.2977673099229523, "learning_rate": 1.5582422030768785e-06, "loss": 0.6687, "step": 22787 }, { "epoch": 0.8253531329228541, "grad_norm": 1.468494025204128, "learning_rate": 1.5576134188990254e-06, "loss": 0.6343, "step": 22788 }, { "epoch": 0.8253893516841724, "grad_norm": 1.304316403670348, "learning_rate": 1.556984750895829e-06, "loss": 0.6173, "step": 22789 }, { "epoch": 0.8254255704454908, "grad_norm": 1.1181921933000487, "learning_rate": 1.5563561990759412e-06, "loss": 0.7059, "step": 22790 }, { "epoch": 0.8254617892068091, "grad_norm": 1.4847867639276782, "learning_rate": 1.5557277634480084e-06, "loss": 0.7723, "step": 22791 }, { "epoch": 0.8254980079681274, "grad_norm": 1.7338070927703346, "learning_rate": 1.5550994440206846e-06, "loss": 0.7457, "step": 22792 }, { "epoch": 0.8255342267294459, "grad_norm": 1.2198340175443683, "learning_rate": 1.554471240802612e-06, "loss": 0.6107, "step": 22793 }, { "epoch": 0.8255704454907642, "grad_norm": 1.3727960832310215, "learning_rate": 1.5538431538024369e-06, "loss": 0.6766, "step": 22794 }, { "epoch": 0.8256066642520826, "grad_norm": 1.128223061183856, "learning_rate": 1.5532151830287978e-06, "loss": 0.7003, "step": 22795 }, { "epoch": 0.8256428830134009, "grad_norm": 1.418359688785343, "learning_rate": 1.5525873284903437e-06, "loss": 0.7057, "step": 22796 }, { "epoch": 0.8256791017747193, "grad_norm": 1.2723190231408061, "learning_rate": 1.5519595901957096e-06, "loss": 0.6517, "step": 22797 }, { "epoch": 0.8257153205360377, "grad_norm": 1.4430678141532143, "learning_rate": 1.5513319681535321e-06, "loss": 0.6836, "step": 22798 }, { "epoch": 0.825751539297356, "grad_norm": 1.5891554148412286, "learning_rate": 1.5507044623724532e-06, "loss": 0.7317, "step": 22799 }, { "epoch": 0.8257877580586744, "grad_norm": 1.4550981712240552, "learning_rate": 1.5500770728611047e-06, "loss": 0.6985, "step": 22800 }, { "epoch": 0.8258239768199928, "grad_norm": 1.3458298725903257, "learning_rate": 1.549449799628121e-06, "loss": 0.6976, "step": 22801 }, { "epoch": 0.8258601955813111, "grad_norm": 1.3760424720676134, "learning_rate": 1.5488226426821307e-06, "loss": 0.7217, "step": 22802 }, { "epoch": 0.8258964143426295, "grad_norm": 1.4012647425193345, "learning_rate": 1.548195602031769e-06, "loss": 0.6865, "step": 22803 }, { "epoch": 0.8259326331039478, "grad_norm": 0.9766754033567108, "learning_rate": 1.5475686776856624e-06, "loss": 0.6696, "step": 22804 }, { "epoch": 0.8259688518652663, "grad_norm": 1.402255655884792, "learning_rate": 1.5469418696524374e-06, "loss": 0.688, "step": 22805 }, { "epoch": 0.8260050706265846, "grad_norm": 1.3850128904052215, "learning_rate": 1.5463151779407182e-06, "loss": 0.652, "step": 22806 }, { "epoch": 0.8260412893879029, "grad_norm": 1.4868821478780516, "learning_rate": 1.5456886025591311e-06, "loss": 0.7148, "step": 22807 }, { "epoch": 0.8260775081492213, "grad_norm": 1.3596229614817874, "learning_rate": 1.5450621435162994e-06, "loss": 0.6781, "step": 22808 }, { "epoch": 0.8261137269105396, "grad_norm": 1.2257274074897884, "learning_rate": 1.5444358008208383e-06, "loss": 0.6132, "step": 22809 }, { "epoch": 0.8261499456718581, "grad_norm": 1.198623873133562, "learning_rate": 1.543809574481372e-06, "loss": 0.5979, "step": 22810 }, { "epoch": 0.8261861644331764, "grad_norm": 1.417820120081142, "learning_rate": 1.5431834645065169e-06, "loss": 0.6907, "step": 22811 }, { "epoch": 0.8262223831944947, "grad_norm": 1.4004321923464549, "learning_rate": 1.5425574709048884e-06, "loss": 0.6901, "step": 22812 }, { "epoch": 0.8262586019558131, "grad_norm": 1.1977328847975763, "learning_rate": 1.5419315936850976e-06, "loss": 0.7448, "step": 22813 }, { "epoch": 0.8262948207171315, "grad_norm": 1.2055227655899194, "learning_rate": 1.5413058328557628e-06, "loss": 0.6383, "step": 22814 }, { "epoch": 0.8263310394784499, "grad_norm": 1.6414742973598906, "learning_rate": 1.5406801884254918e-06, "loss": 0.6455, "step": 22815 }, { "epoch": 0.8263672582397682, "grad_norm": 1.3211431295831726, "learning_rate": 1.5400546604028954e-06, "loss": 0.6112, "step": 22816 }, { "epoch": 0.8264034770010865, "grad_norm": 1.3529092647001764, "learning_rate": 1.5394292487965767e-06, "loss": 0.6829, "step": 22817 }, { "epoch": 0.826439695762405, "grad_norm": 1.2480263820273103, "learning_rate": 1.5388039536151488e-06, "loss": 0.6519, "step": 22818 }, { "epoch": 0.8264759145237233, "grad_norm": 1.3551963276799233, "learning_rate": 1.5381787748672127e-06, "loss": 0.633, "step": 22819 }, { "epoch": 0.8265121332850417, "grad_norm": 1.1510824444609513, "learning_rate": 1.5375537125613704e-06, "loss": 0.6722, "step": 22820 }, { "epoch": 0.82654835204636, "grad_norm": 1.3892424507101606, "learning_rate": 1.536928766706226e-06, "loss": 0.6643, "step": 22821 }, { "epoch": 0.8265845708076783, "grad_norm": 1.3357710811347514, "learning_rate": 1.5363039373103783e-06, "loss": 0.6925, "step": 22822 }, { "epoch": 0.8266207895689968, "grad_norm": 1.4272826820092233, "learning_rate": 1.5356792243824258e-06, "loss": 0.7142, "step": 22823 }, { "epoch": 0.8266570083303151, "grad_norm": 1.3883580746590138, "learning_rate": 1.535054627930962e-06, "loss": 0.7337, "step": 22824 }, { "epoch": 0.8266932270916335, "grad_norm": 1.4175981656053624, "learning_rate": 1.5344301479645862e-06, "loss": 0.5651, "step": 22825 }, { "epoch": 0.8267294458529518, "grad_norm": 1.1801326595966124, "learning_rate": 1.5338057844918907e-06, "loss": 0.6793, "step": 22826 }, { "epoch": 0.8267656646142701, "grad_norm": 1.3851444092194432, "learning_rate": 1.5331815375214676e-06, "loss": 0.6218, "step": 22827 }, { "epoch": 0.8268018833755886, "grad_norm": 1.4906694869434591, "learning_rate": 1.532557407061902e-06, "loss": 0.6721, "step": 22828 }, { "epoch": 0.8268381021369069, "grad_norm": 1.3716232976684308, "learning_rate": 1.5319333931217906e-06, "loss": 0.6404, "step": 22829 }, { "epoch": 0.8268743208982253, "grad_norm": 1.3337349187167895, "learning_rate": 1.5313094957097152e-06, "loss": 0.7132, "step": 22830 }, { "epoch": 0.8269105396595436, "grad_norm": 1.5738773259671444, "learning_rate": 1.5306857148342602e-06, "loss": 0.6488, "step": 22831 }, { "epoch": 0.826946758420862, "grad_norm": 1.2246093980053012, "learning_rate": 1.5300620505040142e-06, "loss": 0.6732, "step": 22832 }, { "epoch": 0.8269829771821804, "grad_norm": 1.348692132362041, "learning_rate": 1.529438502727557e-06, "loss": 0.6615, "step": 22833 }, { "epoch": 0.8270191959434987, "grad_norm": 1.3783474349090863, "learning_rate": 1.5288150715134686e-06, "loss": 0.6907, "step": 22834 }, { "epoch": 0.8270554147048171, "grad_norm": 1.0883808879355064, "learning_rate": 1.5281917568703253e-06, "loss": 0.636, "step": 22835 }, { "epoch": 0.8270916334661355, "grad_norm": 1.4263786893169579, "learning_rate": 1.5275685588067102e-06, "loss": 0.658, "step": 22836 }, { "epoch": 0.8271278522274538, "grad_norm": 1.4834836052660292, "learning_rate": 1.5269454773311965e-06, "loss": 0.6834, "step": 22837 }, { "epoch": 0.8271640709887722, "grad_norm": 1.36211584894815, "learning_rate": 1.5263225124523563e-06, "loss": 0.6278, "step": 22838 }, { "epoch": 0.8272002897500905, "grad_norm": 1.39557543431239, "learning_rate": 1.5256996641787647e-06, "loss": 0.704, "step": 22839 }, { "epoch": 0.827236508511409, "grad_norm": 1.3930942166263747, "learning_rate": 1.525076932518993e-06, "loss": 0.7191, "step": 22840 }, { "epoch": 0.8272727272727273, "grad_norm": 1.4233744792093908, "learning_rate": 1.5244543174816096e-06, "loss": 0.6504, "step": 22841 }, { "epoch": 0.8273089460340456, "grad_norm": 1.360222110627272, "learning_rate": 1.5238318190751789e-06, "loss": 0.6811, "step": 22842 }, { "epoch": 0.827345164795364, "grad_norm": 1.5250224769413263, "learning_rate": 1.5232094373082729e-06, "loss": 0.6753, "step": 22843 }, { "epoch": 0.8273813835566823, "grad_norm": 1.249055957445404, "learning_rate": 1.5225871721894536e-06, "loss": 0.6428, "step": 22844 }, { "epoch": 0.8274176023180008, "grad_norm": 1.4841423448387094, "learning_rate": 1.521965023727283e-06, "loss": 0.7391, "step": 22845 }, { "epoch": 0.8274538210793191, "grad_norm": 1.3415348621042225, "learning_rate": 1.5213429919303225e-06, "loss": 0.7548, "step": 22846 }, { "epoch": 0.8274900398406374, "grad_norm": 1.2507550939127958, "learning_rate": 1.5207210768071334e-06, "loss": 0.6136, "step": 22847 }, { "epoch": 0.8275262586019558, "grad_norm": 1.3610850025395573, "learning_rate": 1.5200992783662727e-06, "loss": 0.6422, "step": 22848 }, { "epoch": 0.8275624773632742, "grad_norm": 1.3435408975675696, "learning_rate": 1.5194775966162955e-06, "loss": 0.6756, "step": 22849 }, { "epoch": 0.8275986961245926, "grad_norm": 1.0887017102938783, "learning_rate": 1.5188560315657596e-06, "loss": 0.7124, "step": 22850 }, { "epoch": 0.8276349148859109, "grad_norm": 1.4463351459360583, "learning_rate": 1.518234583223218e-06, "loss": 0.7092, "step": 22851 }, { "epoch": 0.8276711336472292, "grad_norm": 1.2770630684133362, "learning_rate": 1.5176132515972209e-06, "loss": 0.6666, "step": 22852 }, { "epoch": 0.8277073524085476, "grad_norm": 1.2586309786511176, "learning_rate": 1.5169920366963152e-06, "loss": 0.5618, "step": 22853 }, { "epoch": 0.827743571169866, "grad_norm": 1.3247692805005433, "learning_rate": 1.516370938529057e-06, "loss": 0.7557, "step": 22854 }, { "epoch": 0.8277797899311844, "grad_norm": 1.4289344311631458, "learning_rate": 1.515749957103988e-06, "loss": 0.7301, "step": 22855 }, { "epoch": 0.8278160086925027, "grad_norm": 1.4635231321208788, "learning_rate": 1.5151290924296547e-06, "loss": 0.6286, "step": 22856 }, { "epoch": 0.827852227453821, "grad_norm": 1.4433281053709908, "learning_rate": 1.514508344514598e-06, "loss": 0.6759, "step": 22857 }, { "epoch": 0.8278884462151395, "grad_norm": 1.4896360567377902, "learning_rate": 1.5138877133673656e-06, "loss": 0.5735, "step": 22858 }, { "epoch": 0.8279246649764578, "grad_norm": 1.3935745835304116, "learning_rate": 1.5132671989964942e-06, "loss": 0.695, "step": 22859 }, { "epoch": 0.8279608837377762, "grad_norm": 1.3057649226809458, "learning_rate": 1.512646801410521e-06, "loss": 0.6608, "step": 22860 }, { "epoch": 0.8279971024990945, "grad_norm": 1.2366577887074508, "learning_rate": 1.5120265206179873e-06, "loss": 0.6635, "step": 22861 }, { "epoch": 0.8280333212604128, "grad_norm": 1.3925960841900114, "learning_rate": 1.5114063566274274e-06, "loss": 0.6917, "step": 22862 }, { "epoch": 0.8280695400217313, "grad_norm": 1.4260835496941626, "learning_rate": 1.510786309447375e-06, "loss": 0.6571, "step": 22863 }, { "epoch": 0.8281057587830496, "grad_norm": 1.3731316822444597, "learning_rate": 1.5101663790863597e-06, "loss": 0.66, "step": 22864 }, { "epoch": 0.828141977544368, "grad_norm": 1.3801696004171107, "learning_rate": 1.5095465655529174e-06, "loss": 0.6408, "step": 22865 }, { "epoch": 0.8281781963056863, "grad_norm": 1.3442758355920708, "learning_rate": 1.5089268688555769e-06, "loss": 0.6698, "step": 22866 }, { "epoch": 0.8282144150670047, "grad_norm": 1.29681178773367, "learning_rate": 1.5083072890028582e-06, "loss": 0.6379, "step": 22867 }, { "epoch": 0.8282506338283231, "grad_norm": 1.358157617159304, "learning_rate": 1.5076878260032957e-06, "loss": 0.7052, "step": 22868 }, { "epoch": 0.8282868525896414, "grad_norm": 1.5059058012775743, "learning_rate": 1.507068479865411e-06, "loss": 0.7649, "step": 22869 }, { "epoch": 0.8283230713509598, "grad_norm": 1.3971823081999604, "learning_rate": 1.5064492505977234e-06, "loss": 0.6949, "step": 22870 }, { "epoch": 0.8283592901122782, "grad_norm": 1.5351006956458777, "learning_rate": 1.5058301382087603e-06, "loss": 0.6786, "step": 22871 }, { "epoch": 0.8283955088735965, "grad_norm": 1.359104587937606, "learning_rate": 1.505211142707038e-06, "loss": 0.6284, "step": 22872 }, { "epoch": 0.8284317276349149, "grad_norm": 1.4287707693696305, "learning_rate": 1.5045922641010746e-06, "loss": 0.6868, "step": 22873 }, { "epoch": 0.8284679463962332, "grad_norm": 1.3280625489923075, "learning_rate": 1.503973502399384e-06, "loss": 0.7082, "step": 22874 }, { "epoch": 0.8285041651575517, "grad_norm": 1.3184887373999585, "learning_rate": 1.5033548576104851e-06, "loss": 0.6783, "step": 22875 }, { "epoch": 0.82854038391887, "grad_norm": 1.1501646209220868, "learning_rate": 1.5027363297428898e-06, "loss": 0.6343, "step": 22876 }, { "epoch": 0.8285766026801883, "grad_norm": 0.9627702452300272, "learning_rate": 1.5021179188051082e-06, "loss": 0.6774, "step": 22877 }, { "epoch": 0.8286128214415067, "grad_norm": 1.1858396702775482, "learning_rate": 1.501499624805649e-06, "loss": 0.621, "step": 22878 }, { "epoch": 0.828649040202825, "grad_norm": 1.4306441165718438, "learning_rate": 1.5008814477530242e-06, "loss": 0.673, "step": 22879 }, { "epoch": 0.8286852589641435, "grad_norm": 0.9454177819245813, "learning_rate": 1.5002633876557394e-06, "loss": 0.6592, "step": 22880 }, { "epoch": 0.8287214777254618, "grad_norm": 1.4149779970670686, "learning_rate": 1.4996454445222963e-06, "loss": 0.6674, "step": 22881 }, { "epoch": 0.8287576964867801, "grad_norm": 1.1506141710431461, "learning_rate": 1.499027618361204e-06, "loss": 0.6821, "step": 22882 }, { "epoch": 0.8287939152480985, "grad_norm": 1.5630009211821916, "learning_rate": 1.49840990918096e-06, "loss": 0.749, "step": 22883 }, { "epoch": 0.8288301340094169, "grad_norm": 0.9602890913564878, "learning_rate": 1.4977923169900665e-06, "loss": 0.7127, "step": 22884 }, { "epoch": 0.8288663527707353, "grad_norm": 1.4625972732520012, "learning_rate": 1.497174841797019e-06, "loss": 0.6563, "step": 22885 }, { "epoch": 0.8289025715320536, "grad_norm": 1.3593869928126934, "learning_rate": 1.4965574836103192e-06, "loss": 0.7529, "step": 22886 }, { "epoch": 0.8289387902933719, "grad_norm": 1.1698644556604738, "learning_rate": 1.4959402424384607e-06, "loss": 0.6613, "step": 22887 }, { "epoch": 0.8289750090546903, "grad_norm": 1.1025143460931983, "learning_rate": 1.4953231182899363e-06, "loss": 0.6412, "step": 22888 }, { "epoch": 0.8290112278160087, "grad_norm": 1.350249917548355, "learning_rate": 1.4947061111732363e-06, "loss": 0.6724, "step": 22889 }, { "epoch": 0.8290474465773271, "grad_norm": 1.3986473758473605, "learning_rate": 1.494089221096856e-06, "loss": 0.6332, "step": 22890 }, { "epoch": 0.8290836653386454, "grad_norm": 1.3969804376240786, "learning_rate": 1.493472448069282e-06, "loss": 0.7558, "step": 22891 }, { "epoch": 0.8291198840999637, "grad_norm": 1.4228334360270614, "learning_rate": 1.4928557920989994e-06, "loss": 0.6721, "step": 22892 }, { "epoch": 0.8291561028612822, "grad_norm": 0.9356646649038172, "learning_rate": 1.4922392531944985e-06, "loss": 0.6916, "step": 22893 }, { "epoch": 0.8291923216226005, "grad_norm": 1.474157334192111, "learning_rate": 1.4916228313642612e-06, "loss": 0.6969, "step": 22894 }, { "epoch": 0.8292285403839189, "grad_norm": 1.5880502031758426, "learning_rate": 1.4910065266167695e-06, "loss": 0.6762, "step": 22895 }, { "epoch": 0.8292647591452372, "grad_norm": 1.4037075238539967, "learning_rate": 1.4903903389605024e-06, "loss": 0.5988, "step": 22896 }, { "epoch": 0.8293009779065555, "grad_norm": 1.139047655601558, "learning_rate": 1.489774268403943e-06, "loss": 0.7022, "step": 22897 }, { "epoch": 0.829337196667874, "grad_norm": 1.3078769229303104, "learning_rate": 1.4891583149555688e-06, "loss": 0.6538, "step": 22898 }, { "epoch": 0.8293734154291923, "grad_norm": 1.290173447054168, "learning_rate": 1.4885424786238533e-06, "loss": 0.6334, "step": 22899 }, { "epoch": 0.8294096341905107, "grad_norm": 1.1777115372292735, "learning_rate": 1.4879267594172708e-06, "loss": 0.6585, "step": 22900 }, { "epoch": 0.829445852951829, "grad_norm": 1.409025589625953, "learning_rate": 1.4873111573442967e-06, "loss": 0.6689, "step": 22901 }, { "epoch": 0.8294820717131474, "grad_norm": 1.3908537371175709, "learning_rate": 1.486695672413402e-06, "loss": 0.7263, "step": 22902 }, { "epoch": 0.8295182904744658, "grad_norm": 1.261523989062935, "learning_rate": 1.4860803046330519e-06, "loss": 0.5354, "step": 22903 }, { "epoch": 0.8295545092357841, "grad_norm": 1.3635324080106501, "learning_rate": 1.4854650540117211e-06, "loss": 0.6364, "step": 22904 }, { "epoch": 0.8295907279971025, "grad_norm": 1.5603167969415115, "learning_rate": 1.4848499205578726e-06, "loss": 0.6555, "step": 22905 }, { "epoch": 0.8296269467584209, "grad_norm": 1.1003114517544, "learning_rate": 1.4842349042799708e-06, "loss": 0.6688, "step": 22906 }, { "epoch": 0.8296631655197392, "grad_norm": 1.38763253031701, "learning_rate": 1.483620005186477e-06, "loss": 0.674, "step": 22907 }, { "epoch": 0.8296993842810576, "grad_norm": 1.4254392141788141, "learning_rate": 1.4830052232858582e-06, "loss": 0.654, "step": 22908 }, { "epoch": 0.8297356030423759, "grad_norm": 1.1447341498237555, "learning_rate": 1.4823905585865706e-06, "loss": 0.6522, "step": 22909 }, { "epoch": 0.8297718218036944, "grad_norm": 1.3208989824324662, "learning_rate": 1.4817760110970726e-06, "loss": 0.6527, "step": 22910 }, { "epoch": 0.8298080405650127, "grad_norm": 1.6552385838532921, "learning_rate": 1.4811615808258195e-06, "loss": 0.7199, "step": 22911 }, { "epoch": 0.829844259326331, "grad_norm": 1.1168049740139063, "learning_rate": 1.4805472677812715e-06, "loss": 0.7138, "step": 22912 }, { "epoch": 0.8298804780876494, "grad_norm": 1.3717987626720647, "learning_rate": 1.4799330719718774e-06, "loss": 0.6385, "step": 22913 }, { "epoch": 0.8299166968489677, "grad_norm": 1.454211428804086, "learning_rate": 1.4793189934060891e-06, "loss": 0.6891, "step": 22914 }, { "epoch": 0.8299529156102862, "grad_norm": 1.5276008624915662, "learning_rate": 1.4787050320923613e-06, "loss": 0.7104, "step": 22915 }, { "epoch": 0.8299891343716045, "grad_norm": 1.4142773455801443, "learning_rate": 1.4780911880391392e-06, "loss": 0.6141, "step": 22916 }, { "epoch": 0.8300253531329228, "grad_norm": 1.3160271218913833, "learning_rate": 1.4774774612548704e-06, "loss": 0.6557, "step": 22917 }, { "epoch": 0.8300615718942412, "grad_norm": 1.2667812747124148, "learning_rate": 1.4768638517479983e-06, "loss": 0.6737, "step": 22918 }, { "epoch": 0.8300977906555596, "grad_norm": 1.3520819817797347, "learning_rate": 1.4762503595269705e-06, "loss": 0.7101, "step": 22919 }, { "epoch": 0.830134009416878, "grad_norm": 1.2461380106141715, "learning_rate": 1.4756369846002273e-06, "loss": 0.6619, "step": 22920 }, { "epoch": 0.8301702281781963, "grad_norm": 1.3858545431191234, "learning_rate": 1.475023726976209e-06, "loss": 0.6787, "step": 22921 }, { "epoch": 0.8302064469395146, "grad_norm": 1.4624252954379144, "learning_rate": 1.474410586663354e-06, "loss": 0.7185, "step": 22922 }, { "epoch": 0.830242665700833, "grad_norm": 1.2555218216254618, "learning_rate": 1.4737975636701019e-06, "loss": 0.6337, "step": 22923 }, { "epoch": 0.8302788844621514, "grad_norm": 1.372605581846514, "learning_rate": 1.473184658004887e-06, "loss": 0.7116, "step": 22924 }, { "epoch": 0.8303151032234698, "grad_norm": 1.2168713442760253, "learning_rate": 1.4725718696761415e-06, "loss": 0.6346, "step": 22925 }, { "epoch": 0.8303513219847881, "grad_norm": 1.4828000572615105, "learning_rate": 1.4719591986923021e-06, "loss": 0.7176, "step": 22926 }, { "epoch": 0.8303875407461064, "grad_norm": 1.337775348653388, "learning_rate": 1.4713466450617975e-06, "loss": 0.7058, "step": 22927 }, { "epoch": 0.8304237595074249, "grad_norm": 1.3345742492305084, "learning_rate": 1.4707342087930576e-06, "loss": 0.6964, "step": 22928 }, { "epoch": 0.8304599782687432, "grad_norm": 0.9058846290691405, "learning_rate": 1.4701218898945058e-06, "loss": 0.6756, "step": 22929 }, { "epoch": 0.8304961970300616, "grad_norm": 1.3323750024388974, "learning_rate": 1.4695096883745752e-06, "loss": 0.7002, "step": 22930 }, { "epoch": 0.8305324157913799, "grad_norm": 1.3532068105318962, "learning_rate": 1.468897604241687e-06, "loss": 0.6609, "step": 22931 }, { "epoch": 0.8305686345526982, "grad_norm": 0.987583969042677, "learning_rate": 1.4682856375042608e-06, "loss": 0.6936, "step": 22932 }, { "epoch": 0.8306048533140167, "grad_norm": 1.426521557171292, "learning_rate": 1.4676737881707237e-06, "loss": 0.7563, "step": 22933 }, { "epoch": 0.830641072075335, "grad_norm": 1.2329352072257511, "learning_rate": 1.467062056249492e-06, "loss": 0.6678, "step": 22934 }, { "epoch": 0.8306772908366534, "grad_norm": 1.0652958193775155, "learning_rate": 1.4664504417489844e-06, "loss": 0.6443, "step": 22935 }, { "epoch": 0.8307135095979717, "grad_norm": 1.3316492465627783, "learning_rate": 1.4658389446776145e-06, "loss": 0.6086, "step": 22936 }, { "epoch": 0.8307497283592901, "grad_norm": 1.5035008676785382, "learning_rate": 1.4652275650438019e-06, "loss": 0.6757, "step": 22937 }, { "epoch": 0.8307859471206085, "grad_norm": 1.3732403812822473, "learning_rate": 1.4646163028559568e-06, "loss": 0.6742, "step": 22938 }, { "epoch": 0.8308221658819268, "grad_norm": 1.074713616067176, "learning_rate": 1.4640051581224923e-06, "loss": 0.6291, "step": 22939 }, { "epoch": 0.8308583846432452, "grad_norm": 1.4670740950066112, "learning_rate": 1.463394130851814e-06, "loss": 0.718, "step": 22940 }, { "epoch": 0.8308946034045636, "grad_norm": 1.3604398980799306, "learning_rate": 1.4627832210523363e-06, "loss": 0.7104, "step": 22941 }, { "epoch": 0.8309308221658819, "grad_norm": 1.3609215705346638, "learning_rate": 1.4621724287324623e-06, "loss": 0.6339, "step": 22942 }, { "epoch": 0.8309670409272003, "grad_norm": 0.9673443938407151, "learning_rate": 1.4615617539005955e-06, "loss": 0.6618, "step": 22943 }, { "epoch": 0.8310032596885186, "grad_norm": 1.4093932718902622, "learning_rate": 1.4609511965651447e-06, "loss": 0.7196, "step": 22944 }, { "epoch": 0.831039478449837, "grad_norm": 0.9978238899057069, "learning_rate": 1.4603407567345074e-06, "loss": 0.7171, "step": 22945 }, { "epoch": 0.8310756972111554, "grad_norm": 1.4380637286718672, "learning_rate": 1.4597304344170848e-06, "loss": 0.6181, "step": 22946 }, { "epoch": 0.8311119159724737, "grad_norm": 1.3895536217475162, "learning_rate": 1.4591202296212736e-06, "loss": 0.7322, "step": 22947 }, { "epoch": 0.8311481347337921, "grad_norm": 1.570042824244651, "learning_rate": 1.4585101423554758e-06, "loss": 0.6377, "step": 22948 }, { "epoch": 0.8311843534951104, "grad_norm": 1.5115490998089576, "learning_rate": 1.4579001726280828e-06, "loss": 0.7287, "step": 22949 }, { "epoch": 0.8312205722564289, "grad_norm": 1.2511442200149794, "learning_rate": 1.4572903204474897e-06, "loss": 0.7002, "step": 22950 }, { "epoch": 0.8312567910177472, "grad_norm": 1.1642649441234096, "learning_rate": 1.4566805858220857e-06, "loss": 0.7638, "step": 22951 }, { "epoch": 0.8312930097790655, "grad_norm": 0.9927139534356538, "learning_rate": 1.4560709687602649e-06, "loss": 0.6696, "step": 22952 }, { "epoch": 0.8313292285403839, "grad_norm": 1.3265900465236091, "learning_rate": 1.455461469270416e-06, "loss": 0.6136, "step": 22953 }, { "epoch": 0.8313654473017023, "grad_norm": 1.34938157252716, "learning_rate": 1.4548520873609228e-06, "loss": 0.6756, "step": 22954 }, { "epoch": 0.8314016660630207, "grad_norm": 1.1333069630642862, "learning_rate": 1.454242823040175e-06, "loss": 0.6869, "step": 22955 }, { "epoch": 0.831437884824339, "grad_norm": 1.3532420978799928, "learning_rate": 1.453633676316556e-06, "loss": 0.7006, "step": 22956 }, { "epoch": 0.8314741035856573, "grad_norm": 1.3394445778370176, "learning_rate": 1.4530246471984455e-06, "loss": 0.618, "step": 22957 }, { "epoch": 0.8315103223469757, "grad_norm": 1.2629785470831485, "learning_rate": 1.4524157356942249e-06, "loss": 0.6413, "step": 22958 }, { "epoch": 0.8315465411082941, "grad_norm": 1.4065603364140895, "learning_rate": 1.451806941812276e-06, "loss": 0.6834, "step": 22959 }, { "epoch": 0.8315827598696125, "grad_norm": 1.3604906638146066, "learning_rate": 1.4511982655609746e-06, "loss": 0.7006, "step": 22960 }, { "epoch": 0.8316189786309308, "grad_norm": 1.3166876967464327, "learning_rate": 1.4505897069486973e-06, "loss": 0.601, "step": 22961 }, { "epoch": 0.8316551973922491, "grad_norm": 1.2984583084387753, "learning_rate": 1.4499812659838152e-06, "loss": 0.5996, "step": 22962 }, { "epoch": 0.8316914161535676, "grad_norm": 1.115109247027017, "learning_rate": 1.4493729426747062e-06, "loss": 0.6696, "step": 22963 }, { "epoch": 0.8317276349148859, "grad_norm": 1.146429781582642, "learning_rate": 1.4487647370297386e-06, "loss": 0.6957, "step": 22964 }, { "epoch": 0.8317638536762043, "grad_norm": 1.3730212354368831, "learning_rate": 1.4481566490572797e-06, "loss": 0.6893, "step": 22965 }, { "epoch": 0.8318000724375226, "grad_norm": 1.2131678100188066, "learning_rate": 1.4475486787657011e-06, "loss": 0.6396, "step": 22966 }, { "epoch": 0.831836291198841, "grad_norm": 1.3101249390820384, "learning_rate": 1.4469408261633689e-06, "loss": 0.6953, "step": 22967 }, { "epoch": 0.8318725099601594, "grad_norm": 0.9951591626739867, "learning_rate": 1.4463330912586448e-06, "loss": 0.6026, "step": 22968 }, { "epoch": 0.8319087287214777, "grad_norm": 1.397536219762358, "learning_rate": 1.4457254740598925e-06, "loss": 0.658, "step": 22969 }, { "epoch": 0.8319449474827961, "grad_norm": 1.3940263737160092, "learning_rate": 1.4451179745754752e-06, "loss": 0.6731, "step": 22970 }, { "epoch": 0.8319811662441144, "grad_norm": 1.390016897656497, "learning_rate": 1.4445105928137526e-06, "loss": 0.6612, "step": 22971 }, { "epoch": 0.8320173850054328, "grad_norm": 1.332759308270664, "learning_rate": 1.4439033287830805e-06, "loss": 0.6684, "step": 22972 }, { "epoch": 0.8320536037667512, "grad_norm": 1.673031585146099, "learning_rate": 1.4432961824918156e-06, "loss": 0.724, "step": 22973 }, { "epoch": 0.8320898225280695, "grad_norm": 1.034884166019693, "learning_rate": 1.4426891539483158e-06, "loss": 0.6986, "step": 22974 }, { "epoch": 0.8321260412893879, "grad_norm": 1.340821141147566, "learning_rate": 1.4420822431609316e-06, "loss": 0.6299, "step": 22975 }, { "epoch": 0.8321622600507063, "grad_norm": 1.3146402564479254, "learning_rate": 1.4414754501380134e-06, "loss": 0.647, "step": 22976 }, { "epoch": 0.8321984788120246, "grad_norm": 1.3860429097422837, "learning_rate": 1.4408687748879157e-06, "loss": 0.6227, "step": 22977 }, { "epoch": 0.832234697573343, "grad_norm": 1.5344595268424388, "learning_rate": 1.4402622174189851e-06, "loss": 0.6877, "step": 22978 }, { "epoch": 0.8322709163346613, "grad_norm": 1.5649444139662954, "learning_rate": 1.4396557777395658e-06, "loss": 0.6909, "step": 22979 }, { "epoch": 0.8323071350959798, "grad_norm": 1.1281776297882273, "learning_rate": 1.4390494558580038e-06, "loss": 0.6961, "step": 22980 }, { "epoch": 0.8323433538572981, "grad_norm": 1.3929545394046374, "learning_rate": 1.4384432517826463e-06, "loss": 0.6391, "step": 22981 }, { "epoch": 0.8323795726186164, "grad_norm": 1.1198769687420083, "learning_rate": 1.4378371655218314e-06, "loss": 0.735, "step": 22982 }, { "epoch": 0.8324157913799348, "grad_norm": 1.366488389697444, "learning_rate": 1.4372311970839015e-06, "loss": 0.6551, "step": 22983 }, { "epoch": 0.8324520101412531, "grad_norm": 1.020021702009478, "learning_rate": 1.4366253464771906e-06, "loss": 0.72, "step": 22984 }, { "epoch": 0.8324882289025716, "grad_norm": 1.24581720122015, "learning_rate": 1.4360196137100423e-06, "loss": 0.6395, "step": 22985 }, { "epoch": 0.8325244476638899, "grad_norm": 1.4320605060410927, "learning_rate": 1.43541399879079e-06, "loss": 0.671, "step": 22986 }, { "epoch": 0.8325606664252082, "grad_norm": 1.5778143737564247, "learning_rate": 1.4348085017277635e-06, "loss": 0.5994, "step": 22987 }, { "epoch": 0.8325968851865266, "grad_norm": 1.3692139444024174, "learning_rate": 1.4342031225292995e-06, "loss": 0.6456, "step": 22988 }, { "epoch": 0.832633103947845, "grad_norm": 1.348660736015887, "learning_rate": 1.4335978612037283e-06, "loss": 0.6331, "step": 22989 }, { "epoch": 0.8326693227091634, "grad_norm": 1.2731383965571144, "learning_rate": 1.4329927177593771e-06, "loss": 0.6485, "step": 22990 }, { "epoch": 0.8327055414704817, "grad_norm": 1.1811427981963531, "learning_rate": 1.432387692204572e-06, "loss": 0.7, "step": 22991 }, { "epoch": 0.8327417602318, "grad_norm": 1.2820276772904349, "learning_rate": 1.4317827845476418e-06, "loss": 0.6739, "step": 22992 }, { "epoch": 0.8327779789931185, "grad_norm": 1.4492582948127175, "learning_rate": 1.4311779947969095e-06, "loss": 0.5988, "step": 22993 }, { "epoch": 0.8328141977544368, "grad_norm": 1.457400843965205, "learning_rate": 1.4305733229606956e-06, "loss": 0.6917, "step": 22994 }, { "epoch": 0.8328504165157552, "grad_norm": 1.362968292767589, "learning_rate": 1.4299687690473241e-06, "loss": 0.7111, "step": 22995 }, { "epoch": 0.8328866352770735, "grad_norm": 1.0021249482795105, "learning_rate": 1.429364333065113e-06, "loss": 0.7256, "step": 22996 }, { "epoch": 0.8329228540383918, "grad_norm": 1.3338157699453805, "learning_rate": 1.4287600150223791e-06, "loss": 0.6953, "step": 22997 }, { "epoch": 0.8329590727997103, "grad_norm": 1.4171729987926125, "learning_rate": 1.4281558149274377e-06, "loss": 0.6617, "step": 22998 }, { "epoch": 0.8329952915610286, "grad_norm": 1.387234219932846, "learning_rate": 1.4275517327886045e-06, "loss": 0.7357, "step": 22999 }, { "epoch": 0.833031510322347, "grad_norm": 1.6708971978901423, "learning_rate": 1.426947768614193e-06, "loss": 0.6732, "step": 23000 }, { "epoch": 0.8330677290836653, "grad_norm": 1.4232372266573803, "learning_rate": 1.426343922412513e-06, "loss": 0.6578, "step": 23001 }, { "epoch": 0.8331039478449836, "grad_norm": 1.402913494513209, "learning_rate": 1.4257401941918725e-06, "loss": 0.665, "step": 23002 }, { "epoch": 0.8331401666063021, "grad_norm": 1.049389537519023, "learning_rate": 1.4251365839605823e-06, "loss": 0.7402, "step": 23003 }, { "epoch": 0.8331763853676204, "grad_norm": 1.0962923759568046, "learning_rate": 1.4245330917269473e-06, "loss": 0.7016, "step": 23004 }, { "epoch": 0.8332126041289388, "grad_norm": 1.3791264589150793, "learning_rate": 1.4239297174992706e-06, "loss": 0.6566, "step": 23005 }, { "epoch": 0.8332488228902571, "grad_norm": 1.53447215657386, "learning_rate": 1.4233264612858577e-06, "loss": 0.677, "step": 23006 }, { "epoch": 0.8332850416515755, "grad_norm": 1.3928604307972088, "learning_rate": 1.4227233230950089e-06, "loss": 0.6864, "step": 23007 }, { "epoch": 0.8333212604128939, "grad_norm": 1.396160367578131, "learning_rate": 1.422120302935025e-06, "loss": 0.6745, "step": 23008 }, { "epoch": 0.8333574791742122, "grad_norm": 1.3511811197918016, "learning_rate": 1.4215174008141997e-06, "loss": 0.7058, "step": 23009 }, { "epoch": 0.8333936979355306, "grad_norm": 1.4033496604085836, "learning_rate": 1.420914616740835e-06, "loss": 0.6101, "step": 23010 }, { "epoch": 0.833429916696849, "grad_norm": 1.3413615027103114, "learning_rate": 1.4203119507232233e-06, "loss": 0.6616, "step": 23011 }, { "epoch": 0.8334661354581673, "grad_norm": 1.2547916646428336, "learning_rate": 1.419709402769658e-06, "loss": 0.5564, "step": 23012 }, { "epoch": 0.8335023542194857, "grad_norm": 1.3650085469823705, "learning_rate": 1.4191069728884287e-06, "loss": 0.7368, "step": 23013 }, { "epoch": 0.833538572980804, "grad_norm": 1.4036119691559565, "learning_rate": 1.4185046610878294e-06, "loss": 0.6431, "step": 23014 }, { "epoch": 0.8335747917421225, "grad_norm": 1.3709583881322494, "learning_rate": 1.4179024673761465e-06, "loss": 0.6683, "step": 23015 }, { "epoch": 0.8336110105034408, "grad_norm": 1.3954024044827753, "learning_rate": 1.4173003917616635e-06, "loss": 0.7121, "step": 23016 }, { "epoch": 0.8336472292647591, "grad_norm": 1.2552965862430199, "learning_rate": 1.4166984342526724e-06, "loss": 0.6694, "step": 23017 }, { "epoch": 0.8336834480260775, "grad_norm": 1.388829570379394, "learning_rate": 1.4160965948574513e-06, "loss": 0.6319, "step": 23018 }, { "epoch": 0.8337196667873958, "grad_norm": 1.0733329845290338, "learning_rate": 1.4154948735842844e-06, "loss": 0.7118, "step": 23019 }, { "epoch": 0.8337558855487143, "grad_norm": 1.272374118703441, "learning_rate": 1.4148932704414475e-06, "loss": 0.6872, "step": 23020 }, { "epoch": 0.8337921043100326, "grad_norm": 1.426520827270717, "learning_rate": 1.4142917854372262e-06, "loss": 0.6576, "step": 23021 }, { "epoch": 0.8338283230713509, "grad_norm": 1.625324854107952, "learning_rate": 1.4136904185798939e-06, "loss": 0.7269, "step": 23022 }, { "epoch": 0.8338645418326693, "grad_norm": 1.0367287905480567, "learning_rate": 1.413089169877726e-06, "loss": 0.7656, "step": 23023 }, { "epoch": 0.8339007605939877, "grad_norm": 1.3756413752054797, "learning_rate": 1.412488039338994e-06, "loss": 0.6836, "step": 23024 }, { "epoch": 0.8339369793553061, "grad_norm": 1.394475487654705, "learning_rate": 1.4118870269719743e-06, "loss": 0.6311, "step": 23025 }, { "epoch": 0.8339731981166244, "grad_norm": 0.9189726856362719, "learning_rate": 1.4112861327849347e-06, "loss": 0.6483, "step": 23026 }, { "epoch": 0.8340094168779427, "grad_norm": 1.3837689604877368, "learning_rate": 1.4106853567861435e-06, "loss": 0.5996, "step": 23027 }, { "epoch": 0.8340456356392612, "grad_norm": 1.3355969760475763, "learning_rate": 1.41008469898387e-06, "loss": 0.572, "step": 23028 }, { "epoch": 0.8340818544005795, "grad_norm": 1.4842819918806884, "learning_rate": 1.4094841593863794e-06, "loss": 0.6907, "step": 23029 }, { "epoch": 0.8341180731618979, "grad_norm": 1.3289374166755128, "learning_rate": 1.408883738001936e-06, "loss": 0.6538, "step": 23030 }, { "epoch": 0.8341542919232162, "grad_norm": 1.2517891112860986, "learning_rate": 1.4082834348387974e-06, "loss": 0.6038, "step": 23031 }, { "epoch": 0.8341905106845345, "grad_norm": 1.253493504018484, "learning_rate": 1.4076832499052317e-06, "loss": 0.6422, "step": 23032 }, { "epoch": 0.834226729445853, "grad_norm": 1.5251243046069405, "learning_rate": 1.4070831832094934e-06, "loss": 0.7063, "step": 23033 }, { "epoch": 0.8342629482071713, "grad_norm": 1.3649968362210887, "learning_rate": 1.4064832347598412e-06, "loss": 0.6558, "step": 23034 }, { "epoch": 0.8342991669684897, "grad_norm": 1.2754446554203822, "learning_rate": 1.4058834045645287e-06, "loss": 0.6502, "step": 23035 }, { "epoch": 0.834335385729808, "grad_norm": 1.326934826033816, "learning_rate": 1.4052836926318136e-06, "loss": 0.7444, "step": 23036 }, { "epoch": 0.8343716044911264, "grad_norm": 1.3766621539246253, "learning_rate": 1.4046840989699474e-06, "loss": 0.7207, "step": 23037 }, { "epoch": 0.8344078232524448, "grad_norm": 0.9663275934999983, "learning_rate": 1.4040846235871775e-06, "loss": 0.6379, "step": 23038 }, { "epoch": 0.8344440420137631, "grad_norm": 1.343297894533937, "learning_rate": 1.4034852664917586e-06, "loss": 0.6733, "step": 23039 }, { "epoch": 0.8344802607750815, "grad_norm": 1.0948098254272085, "learning_rate": 1.4028860276919365e-06, "loss": 0.7203, "step": 23040 }, { "epoch": 0.8345164795363998, "grad_norm": 1.3957388108071629, "learning_rate": 1.4022869071959566e-06, "loss": 0.7282, "step": 23041 }, { "epoch": 0.8345526982977182, "grad_norm": 0.9890487743710336, "learning_rate": 1.4016879050120612e-06, "loss": 0.7085, "step": 23042 }, { "epoch": 0.8345889170590366, "grad_norm": 1.011341140201679, "learning_rate": 1.4010890211484983e-06, "loss": 0.6208, "step": 23043 }, { "epoch": 0.8346251358203549, "grad_norm": 1.412356504076118, "learning_rate": 1.4004902556135048e-06, "loss": 0.672, "step": 23044 }, { "epoch": 0.8346613545816733, "grad_norm": 1.6540821075646635, "learning_rate": 1.3998916084153225e-06, "loss": 0.619, "step": 23045 }, { "epoch": 0.8346975733429917, "grad_norm": 1.5314845816558542, "learning_rate": 1.3992930795621862e-06, "loss": 0.6372, "step": 23046 }, { "epoch": 0.83473379210431, "grad_norm": 1.4408833296932086, "learning_rate": 1.398694669062336e-06, "loss": 0.6699, "step": 23047 }, { "epoch": 0.8347700108656284, "grad_norm": 1.4002324950083145, "learning_rate": 1.3980963769240051e-06, "loss": 0.6658, "step": 23048 }, { "epoch": 0.8348062296269467, "grad_norm": 1.3119705138357156, "learning_rate": 1.397498203155424e-06, "loss": 0.6953, "step": 23049 }, { "epoch": 0.8348424483882652, "grad_norm": 1.3643751674194164, "learning_rate": 1.3969001477648292e-06, "loss": 0.7283, "step": 23050 }, { "epoch": 0.8348786671495835, "grad_norm": 1.3387542151069534, "learning_rate": 1.3963022107604474e-06, "loss": 0.6689, "step": 23051 }, { "epoch": 0.8349148859109018, "grad_norm": 1.3229465731546957, "learning_rate": 1.3957043921505075e-06, "loss": 0.623, "step": 23052 }, { "epoch": 0.8349511046722202, "grad_norm": 1.282771898125442, "learning_rate": 1.3951066919432332e-06, "loss": 0.6597, "step": 23053 }, { "epoch": 0.8349873234335385, "grad_norm": 1.2935041928072686, "learning_rate": 1.3945091101468532e-06, "loss": 0.6717, "step": 23054 }, { "epoch": 0.835023542194857, "grad_norm": 1.7062644895001242, "learning_rate": 1.3939116467695902e-06, "loss": 0.6632, "step": 23055 }, { "epoch": 0.8350597609561753, "grad_norm": 0.9914630107801826, "learning_rate": 1.3933143018196615e-06, "loss": 0.6722, "step": 23056 }, { "epoch": 0.8350959797174936, "grad_norm": 1.2243984312993856, "learning_rate": 1.3927170753052944e-06, "loss": 0.6642, "step": 23057 }, { "epoch": 0.835132198478812, "grad_norm": 1.8283388805393999, "learning_rate": 1.392119967234702e-06, "loss": 0.6087, "step": 23058 }, { "epoch": 0.8351684172401304, "grad_norm": 1.513584842025599, "learning_rate": 1.3915229776161031e-06, "loss": 0.6823, "step": 23059 }, { "epoch": 0.8352046360014488, "grad_norm": 1.1091615987817633, "learning_rate": 1.3909261064577094e-06, "loss": 0.6957, "step": 23060 }, { "epoch": 0.8352408547627671, "grad_norm": 1.0599589306480413, "learning_rate": 1.3903293537677398e-06, "loss": 0.6707, "step": 23061 }, { "epoch": 0.8352770735240854, "grad_norm": 0.8655197525154201, "learning_rate": 1.389732719554403e-06, "loss": 0.635, "step": 23062 }, { "epoch": 0.8353132922854039, "grad_norm": 1.2996940580043452, "learning_rate": 1.3891362038259093e-06, "loss": 0.6422, "step": 23063 }, { "epoch": 0.8353495110467222, "grad_norm": 1.2235008489777204, "learning_rate": 1.3885398065904666e-06, "loss": 0.6724, "step": 23064 }, { "epoch": 0.8353857298080406, "grad_norm": 1.5514531643741771, "learning_rate": 1.3879435278562835e-06, "loss": 0.7584, "step": 23065 }, { "epoch": 0.8354219485693589, "grad_norm": 1.332934040082261, "learning_rate": 1.3873473676315662e-06, "loss": 0.6302, "step": 23066 }, { "epoch": 0.8354581673306772, "grad_norm": 1.3448716710736661, "learning_rate": 1.3867513259245135e-06, "loss": 0.6295, "step": 23067 }, { "epoch": 0.8354943860919957, "grad_norm": 1.2937625568812516, "learning_rate": 1.3861554027433332e-06, "loss": 0.6749, "step": 23068 }, { "epoch": 0.835530604853314, "grad_norm": 1.0052644329187028, "learning_rate": 1.3855595980962243e-06, "loss": 0.6913, "step": 23069 }, { "epoch": 0.8355668236146324, "grad_norm": 1.3063086057935906, "learning_rate": 1.3849639119913838e-06, "loss": 0.6509, "step": 23070 }, { "epoch": 0.8356030423759507, "grad_norm": 1.1680166000123946, "learning_rate": 1.3843683444370071e-06, "loss": 0.6562, "step": 23071 }, { "epoch": 0.835639261137269, "grad_norm": 1.3815882481032122, "learning_rate": 1.3837728954412955e-06, "loss": 0.6585, "step": 23072 }, { "epoch": 0.8356754798985875, "grad_norm": 1.3131340119053998, "learning_rate": 1.383177565012439e-06, "loss": 0.5832, "step": 23073 }, { "epoch": 0.8357116986599058, "grad_norm": 1.3284842601293891, "learning_rate": 1.3825823531586312e-06, "loss": 0.6807, "step": 23074 }, { "epoch": 0.8357479174212242, "grad_norm": 1.0731808703038763, "learning_rate": 1.3819872598880602e-06, "loss": 0.6864, "step": 23075 }, { "epoch": 0.8357841361825425, "grad_norm": 0.892361852299425, "learning_rate": 1.3813922852089178e-06, "loss": 0.6715, "step": 23076 }, { "epoch": 0.8358203549438609, "grad_norm": 1.2648983712092223, "learning_rate": 1.3807974291293947e-06, "loss": 0.6318, "step": 23077 }, { "epoch": 0.8358565737051793, "grad_norm": 1.4259234710202213, "learning_rate": 1.380202691657666e-06, "loss": 0.719, "step": 23078 }, { "epoch": 0.8358927924664976, "grad_norm": 1.4466562886725818, "learning_rate": 1.3796080728019256e-06, "loss": 0.7261, "step": 23079 }, { "epoch": 0.835929011227816, "grad_norm": 1.3634012942760605, "learning_rate": 1.3790135725703523e-06, "loss": 0.6454, "step": 23080 }, { "epoch": 0.8359652299891344, "grad_norm": 1.4106693747588406, "learning_rate": 1.3784191909711253e-06, "loss": 0.6823, "step": 23081 }, { "epoch": 0.8360014487504527, "grad_norm": 1.1052178716464705, "learning_rate": 1.3778249280124278e-06, "loss": 0.7071, "step": 23082 }, { "epoch": 0.8360376675117711, "grad_norm": 1.4050339222292494, "learning_rate": 1.3772307837024356e-06, "loss": 0.672, "step": 23083 }, { "epoch": 0.8360738862730894, "grad_norm": 1.3457775586752037, "learning_rate": 1.3766367580493244e-06, "loss": 0.6598, "step": 23084 }, { "epoch": 0.8361101050344079, "grad_norm": 1.3687238126801484, "learning_rate": 1.3760428510612667e-06, "loss": 0.6867, "step": 23085 }, { "epoch": 0.8361463237957262, "grad_norm": 1.2904659398077536, "learning_rate": 1.375449062746439e-06, "loss": 0.6431, "step": 23086 }, { "epoch": 0.8361825425570445, "grad_norm": 1.0126880427764506, "learning_rate": 1.3748553931130103e-06, "loss": 0.678, "step": 23087 }, { "epoch": 0.8362187613183629, "grad_norm": 1.447029510856805, "learning_rate": 1.3742618421691501e-06, "loss": 0.612, "step": 23088 }, { "epoch": 0.8362549800796812, "grad_norm": 1.0147803290731734, "learning_rate": 1.3736684099230247e-06, "loss": 0.6562, "step": 23089 }, { "epoch": 0.8362911988409997, "grad_norm": 1.4567427081026574, "learning_rate": 1.3730750963828033e-06, "loss": 0.6656, "step": 23090 }, { "epoch": 0.836327417602318, "grad_norm": 1.5044122618732627, "learning_rate": 1.3724819015566481e-06, "loss": 0.695, "step": 23091 }, { "epoch": 0.8363636363636363, "grad_norm": 1.247319578950586, "learning_rate": 1.371888825452722e-06, "loss": 0.6292, "step": 23092 }, { "epoch": 0.8363998551249547, "grad_norm": 1.0907191105379461, "learning_rate": 1.3712958680791878e-06, "loss": 0.6769, "step": 23093 }, { "epoch": 0.8364360738862731, "grad_norm": 1.4205540957516742, "learning_rate": 1.3707030294442048e-06, "loss": 0.689, "step": 23094 }, { "epoch": 0.8364722926475915, "grad_norm": 1.2146976976228658, "learning_rate": 1.3701103095559308e-06, "loss": 0.6651, "step": 23095 }, { "epoch": 0.8365085114089098, "grad_norm": 1.4482781694888176, "learning_rate": 1.3695177084225198e-06, "loss": 0.7005, "step": 23096 }, { "epoch": 0.8365447301702281, "grad_norm": 1.3486403976768528, "learning_rate": 1.3689252260521302e-06, "loss": 0.7113, "step": 23097 }, { "epoch": 0.8365809489315466, "grad_norm": 1.1098318369288274, "learning_rate": 1.3683328624529135e-06, "loss": 0.6372, "step": 23098 }, { "epoch": 0.8366171676928649, "grad_norm": 1.4245760703528147, "learning_rate": 1.3677406176330187e-06, "loss": 0.6685, "step": 23099 }, { "epoch": 0.8366533864541833, "grad_norm": 2.1435614079606844, "learning_rate": 1.3671484916006006e-06, "loss": 0.762, "step": 23100 }, { "epoch": 0.8366896052155016, "grad_norm": 0.9903827420066186, "learning_rate": 1.3665564843638045e-06, "loss": 0.6722, "step": 23101 }, { "epoch": 0.8367258239768199, "grad_norm": 1.2785776926855337, "learning_rate": 1.3659645959307777e-06, "loss": 0.6203, "step": 23102 }, { "epoch": 0.8367620427381384, "grad_norm": 1.4959436271586626, "learning_rate": 1.365372826309661e-06, "loss": 0.7595, "step": 23103 }, { "epoch": 0.8367982614994567, "grad_norm": 1.297770123764716, "learning_rate": 1.3647811755086048e-06, "loss": 0.6899, "step": 23104 }, { "epoch": 0.8368344802607751, "grad_norm": 1.3802682955030607, "learning_rate": 1.364189643535747e-06, "loss": 0.6507, "step": 23105 }, { "epoch": 0.8368706990220934, "grad_norm": 1.4485296978767612, "learning_rate": 1.363598230399228e-06, "loss": 0.6689, "step": 23106 }, { "epoch": 0.8369069177834118, "grad_norm": 1.3378571854192274, "learning_rate": 1.363006936107183e-06, "loss": 0.6162, "step": 23107 }, { "epoch": 0.8369431365447302, "grad_norm": 1.578160156906491, "learning_rate": 1.3624157606677534e-06, "loss": 0.7562, "step": 23108 }, { "epoch": 0.8369793553060485, "grad_norm": 1.408564043816125, "learning_rate": 1.3618247040890743e-06, "loss": 0.6813, "step": 23109 }, { "epoch": 0.8370155740673669, "grad_norm": 1.3697826034067173, "learning_rate": 1.3612337663792741e-06, "loss": 0.7003, "step": 23110 }, { "epoch": 0.8370517928286852, "grad_norm": 1.367100783731519, "learning_rate": 1.36064294754649e-06, "loss": 0.625, "step": 23111 }, { "epoch": 0.8370880115900036, "grad_norm": 1.0536947103833858, "learning_rate": 1.3600522475988508e-06, "loss": 0.6424, "step": 23112 }, { "epoch": 0.837124230351322, "grad_norm": 1.785236982861434, "learning_rate": 1.3594616665444837e-06, "loss": 0.661, "step": 23113 }, { "epoch": 0.8371604491126403, "grad_norm": 1.4108397301566706, "learning_rate": 1.3588712043915153e-06, "loss": 0.6711, "step": 23114 }, { "epoch": 0.8371966678739587, "grad_norm": 1.4609521365325, "learning_rate": 1.3582808611480735e-06, "loss": 0.658, "step": 23115 }, { "epoch": 0.8372328866352771, "grad_norm": 1.3692776486264786, "learning_rate": 1.3576906368222797e-06, "loss": 0.6244, "step": 23116 }, { "epoch": 0.8372691053965954, "grad_norm": 1.2511842573952123, "learning_rate": 1.3571005314222574e-06, "loss": 0.6997, "step": 23117 }, { "epoch": 0.8373053241579138, "grad_norm": 1.4318661635703573, "learning_rate": 1.3565105449561234e-06, "loss": 0.7457, "step": 23118 }, { "epoch": 0.8373415429192321, "grad_norm": 2.110662173564257, "learning_rate": 1.3559206774320012e-06, "loss": 0.6634, "step": 23119 }, { "epoch": 0.8373777616805506, "grad_norm": 1.013752489660539, "learning_rate": 1.3553309288580064e-06, "loss": 0.6633, "step": 23120 }, { "epoch": 0.8374139804418689, "grad_norm": 1.234068177043484, "learning_rate": 1.3547412992422515e-06, "loss": 0.5819, "step": 23121 }, { "epoch": 0.8374501992031872, "grad_norm": 1.3519204808107363, "learning_rate": 1.3541517885928546e-06, "loss": 0.6321, "step": 23122 }, { "epoch": 0.8374864179645056, "grad_norm": 1.4566095575450153, "learning_rate": 1.3535623969179256e-06, "loss": 0.6822, "step": 23123 }, { "epoch": 0.8375226367258239, "grad_norm": 1.2943205717540975, "learning_rate": 1.3529731242255762e-06, "loss": 0.6141, "step": 23124 }, { "epoch": 0.8375588554871424, "grad_norm": 1.1454354680124474, "learning_rate": 1.3523839705239128e-06, "loss": 0.7227, "step": 23125 }, { "epoch": 0.8375950742484607, "grad_norm": 1.4721797329004012, "learning_rate": 1.3517949358210458e-06, "loss": 0.7008, "step": 23126 }, { "epoch": 0.8376312930097791, "grad_norm": 1.0087988833127568, "learning_rate": 1.3512060201250798e-06, "loss": 0.6894, "step": 23127 }, { "epoch": 0.8376675117710974, "grad_norm": 1.4957046712517186, "learning_rate": 1.3506172234441184e-06, "loss": 0.7645, "step": 23128 }, { "epoch": 0.8377037305324158, "grad_norm": 1.2106462037425803, "learning_rate": 1.3500285457862616e-06, "loss": 0.6767, "step": 23129 }, { "epoch": 0.8377399492937342, "grad_norm": 1.3113950416563653, "learning_rate": 1.3494399871596143e-06, "loss": 0.6228, "step": 23130 }, { "epoch": 0.8377761680550525, "grad_norm": 1.3914862333994995, "learning_rate": 1.3488515475722752e-06, "loss": 0.6682, "step": 23131 }, { "epoch": 0.8378123868163709, "grad_norm": 0.9553101088201004, "learning_rate": 1.348263227032337e-06, "loss": 0.7035, "step": 23132 }, { "epoch": 0.8378486055776893, "grad_norm": 7.626141306401585, "learning_rate": 1.3476750255479009e-06, "loss": 0.6953, "step": 23133 }, { "epoch": 0.8378848243390076, "grad_norm": 1.480715581981103, "learning_rate": 1.3470869431270606e-06, "loss": 0.7018, "step": 23134 }, { "epoch": 0.837921043100326, "grad_norm": 1.2185367908193814, "learning_rate": 1.346498979777906e-06, "loss": 0.6395, "step": 23135 }, { "epoch": 0.8379572618616443, "grad_norm": 1.4677815706759945, "learning_rate": 1.3459111355085263e-06, "loss": 0.6929, "step": 23136 }, { "epoch": 0.8379934806229627, "grad_norm": 1.243159294868431, "learning_rate": 1.3453234103270162e-06, "loss": 0.6263, "step": 23137 }, { "epoch": 0.8380296993842811, "grad_norm": 1.376655064781977, "learning_rate": 1.3447358042414617e-06, "loss": 0.728, "step": 23138 }, { "epoch": 0.8380659181455994, "grad_norm": 0.9787777316589523, "learning_rate": 1.3441483172599468e-06, "loss": 0.7575, "step": 23139 }, { "epoch": 0.8381021369069178, "grad_norm": 1.4081469232293462, "learning_rate": 1.3435609493905544e-06, "loss": 0.6704, "step": 23140 }, { "epoch": 0.8381383556682361, "grad_norm": 1.288739577488947, "learning_rate": 1.3429737006413712e-06, "loss": 0.6656, "step": 23141 }, { "epoch": 0.8381745744295546, "grad_norm": 1.341859745235199, "learning_rate": 1.3423865710204776e-06, "loss": 0.6553, "step": 23142 }, { "epoch": 0.8382107931908729, "grad_norm": 1.4443271009937046, "learning_rate": 1.3417995605359503e-06, "loss": 0.654, "step": 23143 }, { "epoch": 0.8382470119521912, "grad_norm": 1.4986186304145965, "learning_rate": 1.3412126691958692e-06, "loss": 0.6331, "step": 23144 }, { "epoch": 0.8382832307135096, "grad_norm": 1.359547361594754, "learning_rate": 1.3406258970083119e-06, "loss": 0.7112, "step": 23145 }, { "epoch": 0.838319449474828, "grad_norm": 1.2250828496920816, "learning_rate": 1.3400392439813504e-06, "loss": 0.6584, "step": 23146 }, { "epoch": 0.8383556682361464, "grad_norm": 1.4869130947200582, "learning_rate": 1.339452710123056e-06, "loss": 0.6784, "step": 23147 }, { "epoch": 0.8383918869974647, "grad_norm": 1.2784309572459447, "learning_rate": 1.3388662954415032e-06, "loss": 0.6405, "step": 23148 }, { "epoch": 0.838428105758783, "grad_norm": 1.359183498711059, "learning_rate": 1.3382799999447616e-06, "loss": 0.6737, "step": 23149 }, { "epoch": 0.8384643245201014, "grad_norm": 1.0453998614387392, "learning_rate": 1.3376938236408986e-06, "loss": 0.6396, "step": 23150 }, { "epoch": 0.8385005432814198, "grad_norm": 1.3277824240528509, "learning_rate": 1.3371077665379772e-06, "loss": 0.6121, "step": 23151 }, { "epoch": 0.8385367620427382, "grad_norm": 1.4667692639003607, "learning_rate": 1.3365218286440663e-06, "loss": 0.7446, "step": 23152 }, { "epoch": 0.8385729808040565, "grad_norm": 1.2407946133192702, "learning_rate": 1.335936009967228e-06, "loss": 0.6885, "step": 23153 }, { "epoch": 0.8386091995653748, "grad_norm": 1.9083839172726011, "learning_rate": 1.335350310515522e-06, "loss": 0.6922, "step": 23154 }, { "epoch": 0.8386454183266933, "grad_norm": 1.3333505029305504, "learning_rate": 1.33476473029701e-06, "loss": 0.684, "step": 23155 }, { "epoch": 0.8386816370880116, "grad_norm": 1.4329451034216223, "learning_rate": 1.3341792693197498e-06, "loss": 0.6283, "step": 23156 }, { "epoch": 0.83871785584933, "grad_norm": 1.2763940323835525, "learning_rate": 1.3335939275917964e-06, "loss": 0.6442, "step": 23157 }, { "epoch": 0.8387540746106483, "grad_norm": 1.3588397641281946, "learning_rate": 1.3330087051212048e-06, "loss": 0.5645, "step": 23158 }, { "epoch": 0.8387902933719666, "grad_norm": 1.3160510369221592, "learning_rate": 1.3324236019160296e-06, "loss": 0.6683, "step": 23159 }, { "epoch": 0.8388265121332851, "grad_norm": 1.0795529576844993, "learning_rate": 1.3318386179843223e-06, "loss": 0.6724, "step": 23160 }, { "epoch": 0.8388627308946034, "grad_norm": 1.147597020226703, "learning_rate": 1.3312537533341307e-06, "loss": 0.6689, "step": 23161 }, { "epoch": 0.8388989496559218, "grad_norm": 1.2750110200038742, "learning_rate": 1.3306690079735063e-06, "loss": 0.6213, "step": 23162 }, { "epoch": 0.8389351684172401, "grad_norm": 1.2507440085081951, "learning_rate": 1.3300843819104937e-06, "loss": 0.6304, "step": 23163 }, { "epoch": 0.8389713871785585, "grad_norm": 1.3099344339017243, "learning_rate": 1.3294998751531374e-06, "loss": 0.6381, "step": 23164 }, { "epoch": 0.8390076059398769, "grad_norm": 1.4288181830496285, "learning_rate": 1.3289154877094801e-06, "loss": 0.6466, "step": 23165 }, { "epoch": 0.8390438247011952, "grad_norm": 1.0094974036959885, "learning_rate": 1.3283312195875663e-06, "loss": 0.6475, "step": 23166 }, { "epoch": 0.8390800434625136, "grad_norm": 1.1630909667999343, "learning_rate": 1.3277470707954343e-06, "loss": 0.6709, "step": 23167 }, { "epoch": 0.839116262223832, "grad_norm": 1.348292452148627, "learning_rate": 1.3271630413411241e-06, "loss": 0.6992, "step": 23168 }, { "epoch": 0.8391524809851503, "grad_norm": 1.3338070370380068, "learning_rate": 1.326579131232668e-06, "loss": 0.6038, "step": 23169 }, { "epoch": 0.8391886997464687, "grad_norm": 1.3694858306299387, "learning_rate": 1.3259953404781055e-06, "loss": 0.6051, "step": 23170 }, { "epoch": 0.839224918507787, "grad_norm": 1.2815679774903317, "learning_rate": 1.3254116690854702e-06, "loss": 0.6645, "step": 23171 }, { "epoch": 0.8392611372691055, "grad_norm": 1.078954644903778, "learning_rate": 1.3248281170627897e-06, "loss": 0.6476, "step": 23172 }, { "epoch": 0.8392973560304238, "grad_norm": 1.4041539757487111, "learning_rate": 1.324244684418099e-06, "loss": 0.6289, "step": 23173 }, { "epoch": 0.8393335747917421, "grad_norm": 1.3999889068930371, "learning_rate": 1.3236613711594248e-06, "loss": 0.727, "step": 23174 }, { "epoch": 0.8393697935530605, "grad_norm": 1.311379369377049, "learning_rate": 1.3230781772947942e-06, "loss": 0.6966, "step": 23175 }, { "epoch": 0.8394060123143788, "grad_norm": 1.0704622449983126, "learning_rate": 1.3224951028322296e-06, "loss": 0.6583, "step": 23176 }, { "epoch": 0.8394422310756973, "grad_norm": 1.126709160747907, "learning_rate": 1.3219121477797592e-06, "loss": 0.6871, "step": 23177 }, { "epoch": 0.8394784498370156, "grad_norm": 0.9553599422986578, "learning_rate": 1.321329312145403e-06, "loss": 0.644, "step": 23178 }, { "epoch": 0.8395146685983339, "grad_norm": 1.3611289335107455, "learning_rate": 1.3207465959371812e-06, "loss": 0.7255, "step": 23179 }, { "epoch": 0.8395508873596523, "grad_norm": 1.2953592629038577, "learning_rate": 1.320163999163111e-06, "loss": 0.5973, "step": 23180 }, { "epoch": 0.8395871061209706, "grad_norm": 1.37866036178724, "learning_rate": 1.3195815218312125e-06, "loss": 0.6619, "step": 23181 }, { "epoch": 0.8396233248822891, "grad_norm": 1.3159174407788068, "learning_rate": 1.3189991639494992e-06, "loss": 0.618, "step": 23182 }, { "epoch": 0.8396595436436074, "grad_norm": 1.08114507419212, "learning_rate": 1.3184169255259827e-06, "loss": 0.6656, "step": 23183 }, { "epoch": 0.8396957624049257, "grad_norm": 1.469037866662793, "learning_rate": 1.3178348065686807e-06, "loss": 0.6791, "step": 23184 }, { "epoch": 0.8397319811662441, "grad_norm": 1.3879875986678236, "learning_rate": 1.3172528070856006e-06, "loss": 0.704, "step": 23185 }, { "epoch": 0.8397681999275625, "grad_norm": 0.9346568617859791, "learning_rate": 1.316670927084751e-06, "loss": 0.6776, "step": 23186 }, { "epoch": 0.8398044186888809, "grad_norm": 1.3891649919905071, "learning_rate": 1.316089166574136e-06, "loss": 0.6326, "step": 23187 }, { "epoch": 0.8398406374501992, "grad_norm": 1.3293130658699719, "learning_rate": 1.3155075255617677e-06, "loss": 0.6141, "step": 23188 }, { "epoch": 0.8398768562115175, "grad_norm": 0.9448500998412114, "learning_rate": 1.3149260040556467e-06, "loss": 0.6297, "step": 23189 }, { "epoch": 0.839913074972836, "grad_norm": 1.1358540368478636, "learning_rate": 1.314344602063775e-06, "loss": 0.6762, "step": 23190 }, { "epoch": 0.8399492937341543, "grad_norm": 1.2882099395595767, "learning_rate": 1.3137633195941524e-06, "loss": 0.6444, "step": 23191 }, { "epoch": 0.8399855124954727, "grad_norm": 1.4730769993188515, "learning_rate": 1.3131821566547798e-06, "loss": 0.7052, "step": 23192 }, { "epoch": 0.840021731256791, "grad_norm": 1.1978779902295975, "learning_rate": 1.3126011132536542e-06, "loss": 0.6309, "step": 23193 }, { "epoch": 0.8400579500181093, "grad_norm": 1.4332989011386894, "learning_rate": 1.3120201893987683e-06, "loss": 0.695, "step": 23194 }, { "epoch": 0.8400941687794278, "grad_norm": 1.2681222506019076, "learning_rate": 1.31143938509812e-06, "loss": 0.6074, "step": 23195 }, { "epoch": 0.8401303875407461, "grad_norm": 1.375521026048234, "learning_rate": 1.3108587003597007e-06, "loss": 0.6636, "step": 23196 }, { "epoch": 0.8401666063020645, "grad_norm": 1.1381188435181129, "learning_rate": 1.3102781351915016e-06, "loss": 0.7062, "step": 23197 }, { "epoch": 0.8402028250633828, "grad_norm": 0.969385189079986, "learning_rate": 1.3096976896015079e-06, "loss": 0.6609, "step": 23198 }, { "epoch": 0.8402390438247012, "grad_norm": 1.0308805738650786, "learning_rate": 1.3091173635977116e-06, "loss": 0.6229, "step": 23199 }, { "epoch": 0.8402752625860196, "grad_norm": 1.362414510805352, "learning_rate": 1.308537157188098e-06, "loss": 0.6019, "step": 23200 }, { "epoch": 0.8403114813473379, "grad_norm": 1.2800908856924318, "learning_rate": 1.3079570703806487e-06, "loss": 0.6213, "step": 23201 }, { "epoch": 0.8403477001086563, "grad_norm": 1.293087524536623, "learning_rate": 1.307377103183346e-06, "loss": 0.6348, "step": 23202 }, { "epoch": 0.8403839188699747, "grad_norm": 1.3020848162563143, "learning_rate": 1.3067972556041753e-06, "loss": 0.6448, "step": 23203 }, { "epoch": 0.840420137631293, "grad_norm": 0.987001495967949, "learning_rate": 1.3062175276511124e-06, "loss": 0.6585, "step": 23204 }, { "epoch": 0.8404563563926114, "grad_norm": 1.0272705124316228, "learning_rate": 1.3056379193321333e-06, "loss": 0.6288, "step": 23205 }, { "epoch": 0.8404925751539297, "grad_norm": 1.375351214799006, "learning_rate": 1.3050584306552172e-06, "loss": 0.702, "step": 23206 }, { "epoch": 0.8405287939152482, "grad_norm": 1.3281891526828598, "learning_rate": 1.3044790616283376e-06, "loss": 0.6118, "step": 23207 }, { "epoch": 0.8405650126765665, "grad_norm": 1.2835233766288159, "learning_rate": 1.303899812259467e-06, "loss": 0.6634, "step": 23208 }, { "epoch": 0.8406012314378848, "grad_norm": 1.396856350475559, "learning_rate": 1.3033206825565747e-06, "loss": 0.7063, "step": 23209 }, { "epoch": 0.8406374501992032, "grad_norm": 1.3796493726972514, "learning_rate": 1.302741672527632e-06, "loss": 0.6337, "step": 23210 }, { "epoch": 0.8406736689605215, "grad_norm": 1.510289689733762, "learning_rate": 1.302162782180607e-06, "loss": 0.6565, "step": 23211 }, { "epoch": 0.84070988772184, "grad_norm": 1.1970129749945617, "learning_rate": 1.3015840115234645e-06, "loss": 0.6601, "step": 23212 }, { "epoch": 0.8407461064831583, "grad_norm": 1.345011670507794, "learning_rate": 1.301005360564166e-06, "loss": 0.6554, "step": 23213 }, { "epoch": 0.8407823252444766, "grad_norm": 1.1486234883294986, "learning_rate": 1.3004268293106803e-06, "loss": 0.6494, "step": 23214 }, { "epoch": 0.840818544005795, "grad_norm": 2.39925811285623, "learning_rate": 1.2998484177709658e-06, "loss": 0.7204, "step": 23215 }, { "epoch": 0.8408547627671134, "grad_norm": 1.4534028144061244, "learning_rate": 1.2992701259529793e-06, "loss": 0.6803, "step": 23216 }, { "epoch": 0.8408909815284318, "grad_norm": 1.3073889865847426, "learning_rate": 1.2986919538646835e-06, "loss": 0.7274, "step": 23217 }, { "epoch": 0.8409272002897501, "grad_norm": 1.0062152929588288, "learning_rate": 1.2981139015140309e-06, "loss": 0.683, "step": 23218 }, { "epoch": 0.8409634190510684, "grad_norm": 1.365974686014338, "learning_rate": 1.2975359689089785e-06, "loss": 0.6702, "step": 23219 }, { "epoch": 0.8409996378123868, "grad_norm": 1.40297136633863, "learning_rate": 1.296958156057475e-06, "loss": 0.6317, "step": 23220 }, { "epoch": 0.8410358565737052, "grad_norm": 1.2541048889341797, "learning_rate": 1.296380462967477e-06, "loss": 0.6386, "step": 23221 }, { "epoch": 0.8410720753350236, "grad_norm": 1.3389843667558559, "learning_rate": 1.295802889646931e-06, "loss": 0.6293, "step": 23222 }, { "epoch": 0.8411082940963419, "grad_norm": 1.4012052276274642, "learning_rate": 1.295225436103783e-06, "loss": 0.6554, "step": 23223 }, { "epoch": 0.8411445128576602, "grad_norm": 0.9375908143915627, "learning_rate": 1.2946481023459844e-06, "loss": 0.6584, "step": 23224 }, { "epoch": 0.8411807316189787, "grad_norm": 1.389853326325986, "learning_rate": 1.2940708883814768e-06, "loss": 0.6554, "step": 23225 }, { "epoch": 0.841216950380297, "grad_norm": 1.3718409132687315, "learning_rate": 1.2934937942182034e-06, "loss": 0.5895, "step": 23226 }, { "epoch": 0.8412531691416154, "grad_norm": 1.0755293691632581, "learning_rate": 1.292916819864104e-06, "loss": 0.702, "step": 23227 }, { "epoch": 0.8412893879029337, "grad_norm": 1.4659307375999742, "learning_rate": 1.2923399653271217e-06, "loss": 0.6176, "step": 23228 }, { "epoch": 0.841325606664252, "grad_norm": 1.1246285101993265, "learning_rate": 1.291763230615193e-06, "loss": 0.6874, "step": 23229 }, { "epoch": 0.8413618254255705, "grad_norm": 1.2908340317731122, "learning_rate": 1.291186615736254e-06, "loss": 0.6359, "step": 23230 }, { "epoch": 0.8413980441868888, "grad_norm": 1.4029444441350523, "learning_rate": 1.2906101206982368e-06, "loss": 0.6963, "step": 23231 }, { "epoch": 0.8414342629482072, "grad_norm": 1.9420097073283893, "learning_rate": 1.2900337455090793e-06, "loss": 0.7365, "step": 23232 }, { "epoch": 0.8414704817095255, "grad_norm": 1.2806353388513725, "learning_rate": 1.289457490176711e-06, "loss": 0.6396, "step": 23233 }, { "epoch": 0.8415067004708439, "grad_norm": 1.0501935229612995, "learning_rate": 1.2888813547090585e-06, "loss": 0.7203, "step": 23234 }, { "epoch": 0.8415429192321623, "grad_norm": 1.2663592244229744, "learning_rate": 1.2883053391140553e-06, "loss": 0.687, "step": 23235 }, { "epoch": 0.8415791379934806, "grad_norm": 0.9964056216916041, "learning_rate": 1.2877294433996256e-06, "loss": 0.7202, "step": 23236 }, { "epoch": 0.841615356754799, "grad_norm": 1.2805508350166557, "learning_rate": 1.287153667573694e-06, "loss": 0.6528, "step": 23237 }, { "epoch": 0.8416515755161174, "grad_norm": 1.8161873100984554, "learning_rate": 1.2865780116441817e-06, "loss": 0.634, "step": 23238 }, { "epoch": 0.8416877942774357, "grad_norm": 1.0337931781277785, "learning_rate": 1.2860024756190149e-06, "loss": 0.6282, "step": 23239 }, { "epoch": 0.8417240130387541, "grad_norm": 1.4290594150521176, "learning_rate": 1.2854270595061102e-06, "loss": 0.6991, "step": 23240 }, { "epoch": 0.8417602318000724, "grad_norm": 1.3962552049304182, "learning_rate": 1.2848517633133861e-06, "loss": 0.6955, "step": 23241 }, { "epoch": 0.8417964505613909, "grad_norm": 1.417130856528079, "learning_rate": 1.2842765870487571e-06, "loss": 0.6973, "step": 23242 }, { "epoch": 0.8418326693227092, "grad_norm": 1.4921960006591568, "learning_rate": 1.283701530720144e-06, "loss": 0.7261, "step": 23243 }, { "epoch": 0.8418688880840275, "grad_norm": 1.347331310572241, "learning_rate": 1.2831265943354553e-06, "loss": 0.6526, "step": 23244 }, { "epoch": 0.8419051068453459, "grad_norm": 1.2771438725143873, "learning_rate": 1.282551777902603e-06, "loss": 0.6538, "step": 23245 }, { "epoch": 0.8419413256066642, "grad_norm": 1.584626188906965, "learning_rate": 1.2819770814294996e-06, "loss": 0.7349, "step": 23246 }, { "epoch": 0.8419775443679827, "grad_norm": 1.075866962592325, "learning_rate": 1.2814025049240519e-06, "loss": 0.6349, "step": 23247 }, { "epoch": 0.842013763129301, "grad_norm": 1.0400475485437586, "learning_rate": 1.2808280483941672e-06, "loss": 0.7243, "step": 23248 }, { "epoch": 0.8420499818906193, "grad_norm": 1.0943482839256211, "learning_rate": 1.2802537118477476e-06, "loss": 0.7001, "step": 23249 }, { "epoch": 0.8420862006519377, "grad_norm": 1.5933590437506957, "learning_rate": 1.2796794952927005e-06, "loss": 0.6166, "step": 23250 }, { "epoch": 0.842122419413256, "grad_norm": 1.4085628973194442, "learning_rate": 1.279105398736925e-06, "loss": 0.6628, "step": 23251 }, { "epoch": 0.8421586381745745, "grad_norm": 1.4414809577210437, "learning_rate": 1.2785314221883238e-06, "loss": 0.7051, "step": 23252 }, { "epoch": 0.8421948569358928, "grad_norm": 1.3461253475492423, "learning_rate": 1.2779575656547906e-06, "loss": 0.6478, "step": 23253 }, { "epoch": 0.8422310756972111, "grad_norm": 1.0548468373393245, "learning_rate": 1.2773838291442276e-06, "loss": 0.6747, "step": 23254 }, { "epoch": 0.8422672944585295, "grad_norm": 1.3553448986009813, "learning_rate": 1.2768102126645266e-06, "loss": 0.6318, "step": 23255 }, { "epoch": 0.8423035132198479, "grad_norm": 1.352452037091066, "learning_rate": 1.27623671622358e-06, "loss": 0.6856, "step": 23256 }, { "epoch": 0.8423397319811663, "grad_norm": 1.1828730083632935, "learning_rate": 1.275663339829285e-06, "loss": 0.6239, "step": 23257 }, { "epoch": 0.8423759507424846, "grad_norm": 1.0305462011742186, "learning_rate": 1.2750900834895274e-06, "loss": 0.6509, "step": 23258 }, { "epoch": 0.8424121695038029, "grad_norm": 1.3299384737012552, "learning_rate": 1.2745169472121966e-06, "loss": 0.6503, "step": 23259 }, { "epoch": 0.8424483882651214, "grad_norm": 1.4185501553929067, "learning_rate": 1.273943931005177e-06, "loss": 0.6298, "step": 23260 }, { "epoch": 0.8424846070264397, "grad_norm": 1.1088798738270034, "learning_rate": 1.2733710348763584e-06, "loss": 0.6519, "step": 23261 }, { "epoch": 0.8425208257877581, "grad_norm": 0.954678938505544, "learning_rate": 1.2727982588336229e-06, "loss": 0.6509, "step": 23262 }, { "epoch": 0.8425570445490764, "grad_norm": 0.9895444308391241, "learning_rate": 1.2722256028848513e-06, "loss": 0.6252, "step": 23263 }, { "epoch": 0.8425932633103947, "grad_norm": 1.351090460411859, "learning_rate": 1.2716530670379224e-06, "loss": 0.6353, "step": 23264 }, { "epoch": 0.8426294820717132, "grad_norm": 1.4116179374836353, "learning_rate": 1.271080651300719e-06, "loss": 0.6784, "step": 23265 }, { "epoch": 0.8426657008330315, "grad_norm": 1.3418310714928399, "learning_rate": 1.270508355681116e-06, "loss": 0.6326, "step": 23266 }, { "epoch": 0.8427019195943499, "grad_norm": 1.3597799450252614, "learning_rate": 1.2699361801869858e-06, "loss": 0.6942, "step": 23267 }, { "epoch": 0.8427381383556682, "grad_norm": 1.3326901417140595, "learning_rate": 1.2693641248262067e-06, "loss": 0.6702, "step": 23268 }, { "epoch": 0.8427743571169866, "grad_norm": 1.4418974739765504, "learning_rate": 1.268792189606649e-06, "loss": 0.6662, "step": 23269 }, { "epoch": 0.842810575878305, "grad_norm": 1.2550396060583264, "learning_rate": 1.268220374536182e-06, "loss": 0.68, "step": 23270 }, { "epoch": 0.8428467946396233, "grad_norm": 1.5753321248265921, "learning_rate": 1.2676486796226738e-06, "loss": 0.6917, "step": 23271 }, { "epoch": 0.8428830134009417, "grad_norm": 1.2963077786939312, "learning_rate": 1.2670771048739939e-06, "loss": 0.6266, "step": 23272 }, { "epoch": 0.8429192321622601, "grad_norm": 1.4016632943552465, "learning_rate": 1.2665056502980078e-06, "loss": 0.6774, "step": 23273 }, { "epoch": 0.8429554509235784, "grad_norm": 1.1949832336044859, "learning_rate": 1.2659343159025762e-06, "loss": 0.727, "step": 23274 }, { "epoch": 0.8429916696848968, "grad_norm": 0.9615347491358739, "learning_rate": 1.2653631016955615e-06, "loss": 0.6784, "step": 23275 }, { "epoch": 0.8430278884462151, "grad_norm": 1.327750954127467, "learning_rate": 1.2647920076848275e-06, "loss": 0.6179, "step": 23276 }, { "epoch": 0.8430641072075336, "grad_norm": 1.293309598982799, "learning_rate": 1.2642210338782314e-06, "loss": 0.6407, "step": 23277 }, { "epoch": 0.8431003259688519, "grad_norm": 1.3591138640969105, "learning_rate": 1.2636501802836277e-06, "loss": 0.6442, "step": 23278 }, { "epoch": 0.8431365447301702, "grad_norm": 1.2888633068388746, "learning_rate": 1.2630794469088759e-06, "loss": 0.6616, "step": 23279 }, { "epoch": 0.8431727634914886, "grad_norm": 1.26538176171287, "learning_rate": 1.2625088337618275e-06, "loss": 0.6051, "step": 23280 }, { "epoch": 0.8432089822528069, "grad_norm": 1.2766918930979656, "learning_rate": 1.261938340850335e-06, "loss": 0.625, "step": 23281 }, { "epoch": 0.8432452010141254, "grad_norm": 1.0676557551510986, "learning_rate": 1.2613679681822465e-06, "loss": 0.6466, "step": 23282 }, { "epoch": 0.8432814197754437, "grad_norm": 0.9378680524034656, "learning_rate": 1.2607977157654162e-06, "loss": 0.6666, "step": 23283 }, { "epoch": 0.843317638536762, "grad_norm": 1.3325410214232687, "learning_rate": 1.260227583607687e-06, "loss": 0.6368, "step": 23284 }, { "epoch": 0.8433538572980804, "grad_norm": 1.247028448706108, "learning_rate": 1.2596575717169047e-06, "loss": 0.6404, "step": 23285 }, { "epoch": 0.8433900760593988, "grad_norm": 1.3211223217961119, "learning_rate": 1.2590876801009154e-06, "loss": 0.5894, "step": 23286 }, { "epoch": 0.8434262948207172, "grad_norm": 1.3543808086122835, "learning_rate": 1.2585179087675603e-06, "loss": 0.6955, "step": 23287 }, { "epoch": 0.8434625135820355, "grad_norm": 1.2169648868851726, "learning_rate": 1.2579482577246804e-06, "loss": 0.6663, "step": 23288 }, { "epoch": 0.8434987323433538, "grad_norm": 1.4490415414650262, "learning_rate": 1.257378726980113e-06, "loss": 0.7371, "step": 23289 }, { "epoch": 0.8435349511046722, "grad_norm": 1.2903067371023065, "learning_rate": 1.2568093165416972e-06, "loss": 0.6358, "step": 23290 }, { "epoch": 0.8435711698659906, "grad_norm": 1.4678645089855589, "learning_rate": 1.256240026417267e-06, "loss": 0.7396, "step": 23291 }, { "epoch": 0.843607388627309, "grad_norm": 1.038438041427968, "learning_rate": 1.2556708566146548e-06, "loss": 0.6321, "step": 23292 }, { "epoch": 0.8436436073886273, "grad_norm": 1.3428222897052677, "learning_rate": 1.2551018071416975e-06, "loss": 0.6475, "step": 23293 }, { "epoch": 0.8436798261499456, "grad_norm": 1.2982605181473543, "learning_rate": 1.2545328780062238e-06, "loss": 0.6465, "step": 23294 }, { "epoch": 0.8437160449112641, "grad_norm": 1.2626447674871528, "learning_rate": 1.2539640692160616e-06, "loss": 0.6965, "step": 23295 }, { "epoch": 0.8437522636725824, "grad_norm": 1.366387505127376, "learning_rate": 1.2533953807790356e-06, "loss": 0.6054, "step": 23296 }, { "epoch": 0.8437884824339008, "grad_norm": 2.5241095353694023, "learning_rate": 1.2528268127029786e-06, "loss": 0.6955, "step": 23297 }, { "epoch": 0.8438247011952191, "grad_norm": 1.0054181637011501, "learning_rate": 1.2522583649957088e-06, "loss": 0.6214, "step": 23298 }, { "epoch": 0.8438609199565374, "grad_norm": 1.3111065999602574, "learning_rate": 1.2516900376650498e-06, "loss": 0.6908, "step": 23299 }, { "epoch": 0.8438971387178559, "grad_norm": 1.4526787400779386, "learning_rate": 1.2511218307188233e-06, "loss": 0.7516, "step": 23300 }, { "epoch": 0.8439333574791742, "grad_norm": 1.3565017816833613, "learning_rate": 1.2505537441648496e-06, "loss": 0.6345, "step": 23301 }, { "epoch": 0.8439695762404926, "grad_norm": 1.2469112082652145, "learning_rate": 1.2499857780109426e-06, "loss": 0.6962, "step": 23302 }, { "epoch": 0.8440057950018109, "grad_norm": 1.2359074075241854, "learning_rate": 1.2494179322649192e-06, "loss": 0.7198, "step": 23303 }, { "epoch": 0.8440420137631293, "grad_norm": 1.383409831780927, "learning_rate": 1.2488502069345954e-06, "loss": 0.6176, "step": 23304 }, { "epoch": 0.8440782325244477, "grad_norm": 1.43303075332445, "learning_rate": 1.2482826020277816e-06, "loss": 0.6922, "step": 23305 }, { "epoch": 0.844114451285766, "grad_norm": 0.9946089837612335, "learning_rate": 1.2477151175522894e-06, "loss": 0.6097, "step": 23306 }, { "epoch": 0.8441506700470844, "grad_norm": 1.1890657017911597, "learning_rate": 1.247147753515926e-06, "loss": 0.6365, "step": 23307 }, { "epoch": 0.8441868888084028, "grad_norm": 1.3780023912516248, "learning_rate": 1.2465805099265026e-06, "loss": 0.6838, "step": 23308 }, { "epoch": 0.8442231075697211, "grad_norm": 1.4573312174488346, "learning_rate": 1.2460133867918224e-06, "loss": 0.7153, "step": 23309 }, { "epoch": 0.8442593263310395, "grad_norm": 1.216452057521596, "learning_rate": 1.2454463841196884e-06, "loss": 0.6514, "step": 23310 }, { "epoch": 0.8442955450923578, "grad_norm": 1.2958359268451027, "learning_rate": 1.244879501917906e-06, "loss": 0.6849, "step": 23311 }, { "epoch": 0.8443317638536763, "grad_norm": 1.2533522091932774, "learning_rate": 1.2443127401942746e-06, "loss": 0.6545, "step": 23312 }, { "epoch": 0.8443679826149946, "grad_norm": 1.2472645093827437, "learning_rate": 1.243746098956594e-06, "loss": 0.6912, "step": 23313 }, { "epoch": 0.8444042013763129, "grad_norm": 1.1204819165997255, "learning_rate": 1.2431795782126576e-06, "loss": 0.6489, "step": 23314 }, { "epoch": 0.8444404201376313, "grad_norm": 1.3817253756016477, "learning_rate": 1.2426131779702678e-06, "loss": 0.6784, "step": 23315 }, { "epoch": 0.8444766388989496, "grad_norm": 1.1273036957251175, "learning_rate": 1.2420468982372158e-06, "loss": 0.6957, "step": 23316 }, { "epoch": 0.8445128576602681, "grad_norm": 1.0484581426063435, "learning_rate": 1.2414807390212912e-06, "loss": 0.7443, "step": 23317 }, { "epoch": 0.8445490764215864, "grad_norm": 1.3997810452256938, "learning_rate": 1.2409147003302901e-06, "loss": 0.6678, "step": 23318 }, { "epoch": 0.8445852951829047, "grad_norm": 1.2541379714478644, "learning_rate": 1.2403487821719984e-06, "loss": 0.665, "step": 23319 }, { "epoch": 0.8446215139442231, "grad_norm": 1.2759081409782478, "learning_rate": 1.2397829845542053e-06, "loss": 0.6651, "step": 23320 }, { "epoch": 0.8446577327055415, "grad_norm": 1.5753054949826601, "learning_rate": 1.2392173074846925e-06, "loss": 0.7321, "step": 23321 }, { "epoch": 0.8446939514668599, "grad_norm": 1.310713854014408, "learning_rate": 1.2386517509712503e-06, "loss": 0.6512, "step": 23322 }, { "epoch": 0.8447301702281782, "grad_norm": 1.3649012980641277, "learning_rate": 1.2380863150216583e-06, "loss": 0.6557, "step": 23323 }, { "epoch": 0.8447663889894965, "grad_norm": 1.404281374722437, "learning_rate": 1.2375209996436977e-06, "loss": 0.6732, "step": 23324 }, { "epoch": 0.844802607750815, "grad_norm": 0.8795545690919904, "learning_rate": 1.2369558048451447e-06, "loss": 0.6835, "step": 23325 }, { "epoch": 0.8448388265121333, "grad_norm": 1.5286165919152588, "learning_rate": 1.236390730633783e-06, "loss": 0.6573, "step": 23326 }, { "epoch": 0.8448750452734517, "grad_norm": 1.4038968186228853, "learning_rate": 1.2358257770173842e-06, "loss": 0.7046, "step": 23327 }, { "epoch": 0.84491126403477, "grad_norm": 1.0250444096819045, "learning_rate": 1.235260944003721e-06, "loss": 0.6066, "step": 23328 }, { "epoch": 0.8449474827960883, "grad_norm": 1.4248691829231683, "learning_rate": 1.2346962316005718e-06, "loss": 0.7263, "step": 23329 }, { "epoch": 0.8449837015574068, "grad_norm": 1.3143588070522718, "learning_rate": 1.2341316398157044e-06, "loss": 0.6213, "step": 23330 }, { "epoch": 0.8450199203187251, "grad_norm": 1.444307147726335, "learning_rate": 1.2335671686568872e-06, "loss": 0.6745, "step": 23331 }, { "epoch": 0.8450561390800435, "grad_norm": 1.3658702417360424, "learning_rate": 1.2330028181318876e-06, "loss": 0.6691, "step": 23332 }, { "epoch": 0.8450923578413618, "grad_norm": 1.2205635944731024, "learning_rate": 1.2324385882484736e-06, "loss": 0.6194, "step": 23333 }, { "epoch": 0.8451285766026801, "grad_norm": 1.079647846811792, "learning_rate": 1.2318744790144087e-06, "loss": 0.681, "step": 23334 }, { "epoch": 0.8451647953639986, "grad_norm": 1.2673570854596286, "learning_rate": 1.2313104904374562e-06, "loss": 0.6611, "step": 23335 }, { "epoch": 0.8452010141253169, "grad_norm": 1.4147718462242962, "learning_rate": 1.2307466225253729e-06, "loss": 0.7087, "step": 23336 }, { "epoch": 0.8452372328866353, "grad_norm": 1.402345632110863, "learning_rate": 1.2301828752859246e-06, "loss": 0.6542, "step": 23337 }, { "epoch": 0.8452734516479536, "grad_norm": 1.3201758424020908, "learning_rate": 1.2296192487268644e-06, "loss": 0.6334, "step": 23338 }, { "epoch": 0.845309670409272, "grad_norm": 1.3097471498971978, "learning_rate": 1.229055742855949e-06, "loss": 0.6629, "step": 23339 }, { "epoch": 0.8453458891705904, "grad_norm": 1.4289469879197714, "learning_rate": 1.228492357680935e-06, "loss": 0.722, "step": 23340 }, { "epoch": 0.8453821079319087, "grad_norm": 1.4149251724861058, "learning_rate": 1.2279290932095734e-06, "loss": 0.7196, "step": 23341 }, { "epoch": 0.8454183266932271, "grad_norm": 1.4472528616668032, "learning_rate": 1.2273659494496148e-06, "loss": 0.673, "step": 23342 }, { "epoch": 0.8454545454545455, "grad_norm": 1.38547851716784, "learning_rate": 1.2268029264088076e-06, "loss": 0.6752, "step": 23343 }, { "epoch": 0.8454907642158638, "grad_norm": 1.3869429682405556, "learning_rate": 1.2262400240949023e-06, "loss": 0.6917, "step": 23344 }, { "epoch": 0.8455269829771822, "grad_norm": 1.3773672725575143, "learning_rate": 1.2256772425156449e-06, "loss": 0.6781, "step": 23345 }, { "epoch": 0.8455632017385005, "grad_norm": 1.4988897785900999, "learning_rate": 1.2251145816787768e-06, "loss": 0.6712, "step": 23346 }, { "epoch": 0.845599420499819, "grad_norm": 1.1021815274402316, "learning_rate": 1.224552041592042e-06, "loss": 0.6992, "step": 23347 }, { "epoch": 0.8456356392611373, "grad_norm": 1.341466953357074, "learning_rate": 1.223989622263183e-06, "loss": 0.5942, "step": 23348 }, { "epoch": 0.8456718580224556, "grad_norm": 1.3731559822663653, "learning_rate": 1.2234273236999371e-06, "loss": 0.716, "step": 23349 }, { "epoch": 0.845708076783774, "grad_norm": 1.2303845056955942, "learning_rate": 1.2228651459100427e-06, "loss": 0.6255, "step": 23350 }, { "epoch": 0.8457442955450923, "grad_norm": 1.3900422270817745, "learning_rate": 1.2223030889012377e-06, "loss": 0.6746, "step": 23351 }, { "epoch": 0.8457805143064108, "grad_norm": 1.4660068726170348, "learning_rate": 1.2217411526812539e-06, "loss": 0.6601, "step": 23352 }, { "epoch": 0.8458167330677291, "grad_norm": 1.299828684618981, "learning_rate": 1.2211793372578262e-06, "loss": 0.6551, "step": 23353 }, { "epoch": 0.8458529518290474, "grad_norm": 1.3089203035650485, "learning_rate": 1.2206176426386828e-06, "loss": 0.6303, "step": 23354 }, { "epoch": 0.8458891705903658, "grad_norm": 1.1526278158289405, "learning_rate": 1.2200560688315567e-06, "loss": 0.6883, "step": 23355 }, { "epoch": 0.8459253893516842, "grad_norm": 1.4086594159809414, "learning_rate": 1.2194946158441733e-06, "loss": 0.6121, "step": 23356 }, { "epoch": 0.8459616081130026, "grad_norm": 1.4471052147827586, "learning_rate": 1.2189332836842593e-06, "loss": 0.6992, "step": 23357 }, { "epoch": 0.8459978268743209, "grad_norm": 1.3806457259827272, "learning_rate": 1.218372072359537e-06, "loss": 0.5826, "step": 23358 }, { "epoch": 0.8460340456356392, "grad_norm": 1.3852977328919285, "learning_rate": 1.2178109818777327e-06, "loss": 0.6238, "step": 23359 }, { "epoch": 0.8460702643969576, "grad_norm": 1.2848562670240298, "learning_rate": 1.217250012246567e-06, "loss": 0.6564, "step": 23360 }, { "epoch": 0.846106483158276, "grad_norm": 1.3486005556860563, "learning_rate": 1.2166891634737564e-06, "loss": 0.6509, "step": 23361 }, { "epoch": 0.8461427019195944, "grad_norm": 1.4444399293935062, "learning_rate": 1.2161284355670223e-06, "loss": 0.737, "step": 23362 }, { "epoch": 0.8461789206809127, "grad_norm": 1.503940472275397, "learning_rate": 1.2155678285340789e-06, "loss": 0.7458, "step": 23363 }, { "epoch": 0.846215139442231, "grad_norm": 1.3417491224140186, "learning_rate": 1.2150073423826403e-06, "loss": 0.6848, "step": 23364 }, { "epoch": 0.8462513582035495, "grad_norm": 1.3612425365387892, "learning_rate": 1.2144469771204193e-06, "loss": 0.6838, "step": 23365 }, { "epoch": 0.8462875769648678, "grad_norm": 1.1266958134667653, "learning_rate": 1.2138867327551285e-06, "loss": 0.6529, "step": 23366 }, { "epoch": 0.8463237957261862, "grad_norm": 1.2880573002001223, "learning_rate": 1.2133266092944763e-06, "loss": 0.6666, "step": 23367 }, { "epoch": 0.8463600144875045, "grad_norm": 1.5485261042984029, "learning_rate": 1.212766606746172e-06, "loss": 0.6837, "step": 23368 }, { "epoch": 0.8463962332488228, "grad_norm": 1.525200885127963, "learning_rate": 1.212206725117917e-06, "loss": 0.7256, "step": 23369 }, { "epoch": 0.8464324520101413, "grad_norm": 1.3967518656978757, "learning_rate": 1.211646964417421e-06, "loss": 0.6598, "step": 23370 }, { "epoch": 0.8464686707714596, "grad_norm": 1.3409531208731786, "learning_rate": 1.2110873246523857e-06, "loss": 0.6911, "step": 23371 }, { "epoch": 0.846504889532778, "grad_norm": 1.366804161826413, "learning_rate": 1.210527805830508e-06, "loss": 0.6777, "step": 23372 }, { "epoch": 0.8465411082940963, "grad_norm": 1.3648459641930493, "learning_rate": 1.2099684079594943e-06, "loss": 0.6748, "step": 23373 }, { "epoch": 0.8465773270554147, "grad_norm": 1.5665185090219027, "learning_rate": 1.2094091310470391e-06, "loss": 0.7353, "step": 23374 }, { "epoch": 0.8466135458167331, "grad_norm": 1.3065853142827895, "learning_rate": 1.2088499751008377e-06, "loss": 0.6602, "step": 23375 }, { "epoch": 0.8466497645780514, "grad_norm": 1.2905993937637739, "learning_rate": 1.2082909401285825e-06, "loss": 0.6392, "step": 23376 }, { "epoch": 0.8466859833393698, "grad_norm": 1.3568029639368344, "learning_rate": 1.2077320261379722e-06, "loss": 0.6649, "step": 23377 }, { "epoch": 0.8467222021006882, "grad_norm": 1.3544257875874057, "learning_rate": 1.2071732331366936e-06, "loss": 0.6447, "step": 23378 }, { "epoch": 0.8467584208620065, "grad_norm": 1.4018382367357518, "learning_rate": 1.2066145611324366e-06, "loss": 0.6318, "step": 23379 }, { "epoch": 0.8467946396233249, "grad_norm": 0.9513838757743556, "learning_rate": 1.206056010132891e-06, "loss": 0.6628, "step": 23380 }, { "epoch": 0.8468308583846432, "grad_norm": 1.322712636749516, "learning_rate": 1.2054975801457424e-06, "loss": 0.6735, "step": 23381 }, { "epoch": 0.8468670771459617, "grad_norm": 1.3583359940503785, "learning_rate": 1.2049392711786733e-06, "loss": 0.6714, "step": 23382 }, { "epoch": 0.84690329590728, "grad_norm": 1.3373716983856474, "learning_rate": 1.2043810832393676e-06, "loss": 0.6435, "step": 23383 }, { "epoch": 0.8469395146685983, "grad_norm": 1.0248237647778604, "learning_rate": 1.203823016335508e-06, "loss": 0.6241, "step": 23384 }, { "epoch": 0.8469757334299167, "grad_norm": 1.3901466279502386, "learning_rate": 1.2032650704747716e-06, "loss": 0.6713, "step": 23385 }, { "epoch": 0.847011952191235, "grad_norm": 1.3930141159710434, "learning_rate": 1.2027072456648392e-06, "loss": 0.6497, "step": 23386 }, { "epoch": 0.8470481709525535, "grad_norm": 1.335449080667715, "learning_rate": 1.2021495419133822e-06, "loss": 0.7006, "step": 23387 }, { "epoch": 0.8470843897138718, "grad_norm": 1.3653500007269257, "learning_rate": 1.20159195922808e-06, "loss": 0.7531, "step": 23388 }, { "epoch": 0.8471206084751901, "grad_norm": 0.9928141993850343, "learning_rate": 1.2010344976166033e-06, "loss": 0.7133, "step": 23389 }, { "epoch": 0.8471568272365085, "grad_norm": 1.4035918743529108, "learning_rate": 1.2004771570866214e-06, "loss": 0.6987, "step": 23390 }, { "epoch": 0.8471930459978269, "grad_norm": 1.3583339405722763, "learning_rate": 1.199919937645807e-06, "loss": 0.7184, "step": 23391 }, { "epoch": 0.8472292647591453, "grad_norm": 1.521084271931912, "learning_rate": 1.1993628393018276e-06, "loss": 0.7112, "step": 23392 }, { "epoch": 0.8472654835204636, "grad_norm": 0.9150016810683247, "learning_rate": 1.1988058620623488e-06, "loss": 0.6859, "step": 23393 }, { "epoch": 0.8473017022817819, "grad_norm": 1.2615237734933402, "learning_rate": 1.1982490059350316e-06, "loss": 0.7215, "step": 23394 }, { "epoch": 0.8473379210431004, "grad_norm": 1.429705482249961, "learning_rate": 1.1976922709275452e-06, "loss": 0.6707, "step": 23395 }, { "epoch": 0.8473741398044187, "grad_norm": 1.290149046564892, "learning_rate": 1.1971356570475467e-06, "loss": 0.6368, "step": 23396 }, { "epoch": 0.8474103585657371, "grad_norm": 1.3472098882885917, "learning_rate": 1.1965791643026969e-06, "loss": 0.6533, "step": 23397 }, { "epoch": 0.8474465773270554, "grad_norm": 1.4348145206302758, "learning_rate": 1.1960227927006508e-06, "loss": 0.6664, "step": 23398 }, { "epoch": 0.8474827960883737, "grad_norm": 0.993944528657033, "learning_rate": 1.1954665422490686e-06, "loss": 0.7095, "step": 23399 }, { "epoch": 0.8475190148496922, "grad_norm": 0.9960848681086627, "learning_rate": 1.1949104129556034e-06, "loss": 0.6658, "step": 23400 }, { "epoch": 0.8475552336110105, "grad_norm": 1.4310037157652669, "learning_rate": 1.1943544048279065e-06, "loss": 0.6645, "step": 23401 }, { "epoch": 0.8475914523723289, "grad_norm": 1.387725588141418, "learning_rate": 1.1937985178736312e-06, "loss": 0.6719, "step": 23402 }, { "epoch": 0.8476276711336472, "grad_norm": 1.4424680356654676, "learning_rate": 1.1932427521004275e-06, "loss": 0.6717, "step": 23403 }, { "epoch": 0.8476638898949655, "grad_norm": 1.4216772350652258, "learning_rate": 1.1926871075159406e-06, "loss": 0.7194, "step": 23404 }, { "epoch": 0.847700108656284, "grad_norm": 1.3597594825947803, "learning_rate": 1.1921315841278158e-06, "loss": 0.6736, "step": 23405 }, { "epoch": 0.8477363274176023, "grad_norm": 1.4393256453194958, "learning_rate": 1.1915761819437021e-06, "loss": 0.6702, "step": 23406 }, { "epoch": 0.8477725461789207, "grad_norm": 1.0923694275188252, "learning_rate": 1.1910209009712403e-06, "loss": 0.6564, "step": 23407 }, { "epoch": 0.847808764940239, "grad_norm": 1.3796628613394257, "learning_rate": 1.190465741218071e-06, "loss": 0.6886, "step": 23408 }, { "epoch": 0.8478449837015574, "grad_norm": 1.478356481954591, "learning_rate": 1.189910702691831e-06, "loss": 0.6492, "step": 23409 }, { "epoch": 0.8478812024628758, "grad_norm": 1.2892079372435705, "learning_rate": 1.1893557854001636e-06, "loss": 0.6848, "step": 23410 }, { "epoch": 0.8479174212241941, "grad_norm": 1.2844488659557172, "learning_rate": 1.1888009893507024e-06, "loss": 0.6834, "step": 23411 }, { "epoch": 0.8479536399855125, "grad_norm": 1.340794826996072, "learning_rate": 1.1882463145510792e-06, "loss": 0.686, "step": 23412 }, { "epoch": 0.8479898587468309, "grad_norm": 1.3404747496118377, "learning_rate": 1.1876917610089323e-06, "loss": 0.6946, "step": 23413 }, { "epoch": 0.8480260775081492, "grad_norm": 1.0208337943632457, "learning_rate": 1.18713732873189e-06, "loss": 0.7467, "step": 23414 }, { "epoch": 0.8480622962694676, "grad_norm": 1.3161792347140078, "learning_rate": 1.1865830177275805e-06, "loss": 0.6335, "step": 23415 }, { "epoch": 0.8480985150307859, "grad_norm": 1.4130171223237964, "learning_rate": 1.1860288280036314e-06, "loss": 0.6374, "step": 23416 }, { "epoch": 0.8481347337921044, "grad_norm": 1.8569808019093512, "learning_rate": 1.185474759567673e-06, "loss": 0.6807, "step": 23417 }, { "epoch": 0.8481709525534227, "grad_norm": 1.4194558400712824, "learning_rate": 1.184920812427326e-06, "loss": 0.6419, "step": 23418 }, { "epoch": 0.848207171314741, "grad_norm": 1.3105801522681906, "learning_rate": 1.1843669865902152e-06, "loss": 0.5902, "step": 23419 }, { "epoch": 0.8482433900760594, "grad_norm": 0.9470185184899442, "learning_rate": 1.183813282063958e-06, "loss": 0.6275, "step": 23420 }, { "epoch": 0.8482796088373777, "grad_norm": 1.0897992065814113, "learning_rate": 1.1832596988561785e-06, "loss": 0.5994, "step": 23421 }, { "epoch": 0.8483158275986962, "grad_norm": 1.3784060543571022, "learning_rate": 1.1827062369744924e-06, "loss": 0.676, "step": 23422 }, { "epoch": 0.8483520463600145, "grad_norm": 1.5453765389400378, "learning_rate": 1.182152896426515e-06, "loss": 0.6529, "step": 23423 }, { "epoch": 0.8483882651213328, "grad_norm": 1.4069693368001772, "learning_rate": 1.1815996772198635e-06, "loss": 0.6997, "step": 23424 }, { "epoch": 0.8484244838826512, "grad_norm": 1.394253900874247, "learning_rate": 1.1810465793621484e-06, "loss": 0.6578, "step": 23425 }, { "epoch": 0.8484607026439696, "grad_norm": 1.0266597934241657, "learning_rate": 1.1804936028609826e-06, "loss": 0.6565, "step": 23426 }, { "epoch": 0.848496921405288, "grad_norm": 0.984132377983659, "learning_rate": 1.179940747723971e-06, "loss": 0.6295, "step": 23427 }, { "epoch": 0.8485331401666063, "grad_norm": 1.427748678738479, "learning_rate": 1.1793880139587266e-06, "loss": 0.6443, "step": 23428 }, { "epoch": 0.8485693589279246, "grad_norm": 1.5412496478520445, "learning_rate": 1.1788354015728543e-06, "loss": 0.5785, "step": 23429 }, { "epoch": 0.848605577689243, "grad_norm": 1.4081398199722113, "learning_rate": 1.1782829105739578e-06, "loss": 0.6788, "step": 23430 }, { "epoch": 0.8486417964505614, "grad_norm": 1.2079406240060873, "learning_rate": 1.1777305409696382e-06, "loss": 0.6128, "step": 23431 }, { "epoch": 0.8486780152118798, "grad_norm": 1.4527997657888991, "learning_rate": 1.1771782927675001e-06, "loss": 0.6119, "step": 23432 }, { "epoch": 0.8487142339731981, "grad_norm": 1.3543070300117332, "learning_rate": 1.176626165975141e-06, "loss": 0.6379, "step": 23433 }, { "epoch": 0.8487504527345164, "grad_norm": 1.4241148708332845, "learning_rate": 1.1760741606001558e-06, "loss": 0.6541, "step": 23434 }, { "epoch": 0.8487866714958349, "grad_norm": 1.060883389172438, "learning_rate": 1.1755222766501473e-06, "loss": 0.624, "step": 23435 }, { "epoch": 0.8488228902571532, "grad_norm": 1.343003893409608, "learning_rate": 1.174970514132705e-06, "loss": 0.6697, "step": 23436 }, { "epoch": 0.8488591090184716, "grad_norm": 1.0395246648766674, "learning_rate": 1.174418873055423e-06, "loss": 0.6035, "step": 23437 }, { "epoch": 0.8488953277797899, "grad_norm": 1.4181422805216628, "learning_rate": 1.1738673534258894e-06, "loss": 0.6152, "step": 23438 }, { "epoch": 0.8489315465411083, "grad_norm": 1.0225265440550912, "learning_rate": 1.1733159552516992e-06, "loss": 0.6627, "step": 23439 }, { "epoch": 0.8489677653024267, "grad_norm": 1.3322728698605957, "learning_rate": 1.172764678540438e-06, "loss": 0.6401, "step": 23440 }, { "epoch": 0.849003984063745, "grad_norm": 1.4922610702902481, "learning_rate": 1.1722135232996879e-06, "loss": 0.6093, "step": 23441 }, { "epoch": 0.8490402028250634, "grad_norm": 1.514802340787004, "learning_rate": 1.1716624895370387e-06, "loss": 0.6423, "step": 23442 }, { "epoch": 0.8490764215863817, "grad_norm": 1.3072284907075007, "learning_rate": 1.1711115772600712e-06, "loss": 0.6676, "step": 23443 }, { "epoch": 0.8491126403477001, "grad_norm": 1.385141546286456, "learning_rate": 1.1705607864763657e-06, "loss": 0.7217, "step": 23444 }, { "epoch": 0.8491488591090185, "grad_norm": 1.256271431156517, "learning_rate": 1.1700101171935008e-06, "loss": 0.6047, "step": 23445 }, { "epoch": 0.8491850778703368, "grad_norm": 1.430110810752144, "learning_rate": 1.1694595694190557e-06, "loss": 0.6477, "step": 23446 }, { "epoch": 0.8492212966316552, "grad_norm": 1.4436835078620187, "learning_rate": 1.1689091431606082e-06, "loss": 0.6628, "step": 23447 }, { "epoch": 0.8492575153929736, "grad_norm": 1.5562800093806137, "learning_rate": 1.1683588384257294e-06, "loss": 0.6462, "step": 23448 }, { "epoch": 0.8492937341542919, "grad_norm": 1.4739392317813622, "learning_rate": 1.1678086552219915e-06, "loss": 0.6649, "step": 23449 }, { "epoch": 0.8493299529156103, "grad_norm": 1.5005926523551167, "learning_rate": 1.1672585935569691e-06, "loss": 0.7269, "step": 23450 }, { "epoch": 0.8493661716769286, "grad_norm": 1.392708926725606, "learning_rate": 1.1667086534382287e-06, "loss": 0.6859, "step": 23451 }, { "epoch": 0.8494023904382471, "grad_norm": 1.3635276350448564, "learning_rate": 1.1661588348733377e-06, "loss": 0.631, "step": 23452 }, { "epoch": 0.8494386091995654, "grad_norm": 1.5528914510102256, "learning_rate": 1.1656091378698653e-06, "loss": 0.7001, "step": 23453 }, { "epoch": 0.8494748279608837, "grad_norm": 1.259799116266096, "learning_rate": 1.1650595624353744e-06, "loss": 0.6628, "step": 23454 }, { "epoch": 0.8495110467222021, "grad_norm": 1.5469449923314884, "learning_rate": 1.1645101085774258e-06, "loss": 0.5786, "step": 23455 }, { "epoch": 0.8495472654835204, "grad_norm": 1.2749906653233158, "learning_rate": 1.1639607763035798e-06, "loss": 0.6349, "step": 23456 }, { "epoch": 0.8495834842448389, "grad_norm": 1.3520488635676744, "learning_rate": 1.1634115656213995e-06, "loss": 0.7027, "step": 23457 }, { "epoch": 0.8496197030061572, "grad_norm": 1.2862688548753884, "learning_rate": 1.162862476538441e-06, "loss": 0.5902, "step": 23458 }, { "epoch": 0.8496559217674755, "grad_norm": 0.9275876358332281, "learning_rate": 1.1623135090622606e-06, "loss": 0.6804, "step": 23459 }, { "epoch": 0.8496921405287939, "grad_norm": 1.3791328789667368, "learning_rate": 1.1617646632004087e-06, "loss": 0.6134, "step": 23460 }, { "epoch": 0.8497283592901123, "grad_norm": 1.3308319736701986, "learning_rate": 1.1612159389604426e-06, "loss": 0.6169, "step": 23461 }, { "epoch": 0.8497645780514307, "grad_norm": 1.3964315751237781, "learning_rate": 1.1606673363499121e-06, "loss": 0.6932, "step": 23462 }, { "epoch": 0.849800796812749, "grad_norm": 1.2640774170463374, "learning_rate": 1.1601188553763643e-06, "loss": 0.6429, "step": 23463 }, { "epoch": 0.8498370155740673, "grad_norm": 1.3490438684264245, "learning_rate": 1.1595704960473498e-06, "loss": 0.7333, "step": 23464 }, { "epoch": 0.8498732343353858, "grad_norm": 1.403078043710586, "learning_rate": 1.1590222583704135e-06, "loss": 0.727, "step": 23465 }, { "epoch": 0.8499094530967041, "grad_norm": 1.2555904419357742, "learning_rate": 1.1584741423530999e-06, "loss": 0.6581, "step": 23466 }, { "epoch": 0.8499456718580225, "grad_norm": 1.2830126612329638, "learning_rate": 1.157926148002948e-06, "loss": 0.6247, "step": 23467 }, { "epoch": 0.8499818906193408, "grad_norm": 0.998226441284082, "learning_rate": 1.1573782753275042e-06, "loss": 0.6736, "step": 23468 }, { "epoch": 0.8500181093806591, "grad_norm": 1.388657893406892, "learning_rate": 1.1568305243343047e-06, "loss": 0.7203, "step": 23469 }, { "epoch": 0.8500543281419776, "grad_norm": 1.4147796709030245, "learning_rate": 1.156282895030888e-06, "loss": 0.6562, "step": 23470 }, { "epoch": 0.8500905469032959, "grad_norm": 1.2103878759133986, "learning_rate": 1.1557353874247868e-06, "loss": 0.681, "step": 23471 }, { "epoch": 0.8501267656646143, "grad_norm": 1.3653169035354868, "learning_rate": 1.1551880015235406e-06, "loss": 0.6824, "step": 23472 }, { "epoch": 0.8501629844259326, "grad_norm": 1.4125322286549657, "learning_rate": 1.154640737334679e-06, "loss": 0.689, "step": 23473 }, { "epoch": 0.850199203187251, "grad_norm": 1.3255067584434053, "learning_rate": 1.1540935948657305e-06, "loss": 0.6624, "step": 23474 }, { "epoch": 0.8502354219485694, "grad_norm": 1.2232314491512504, "learning_rate": 1.153546574124229e-06, "loss": 0.6232, "step": 23475 }, { "epoch": 0.8502716407098877, "grad_norm": 1.0375028587120239, "learning_rate": 1.1529996751177007e-06, "loss": 0.6548, "step": 23476 }, { "epoch": 0.8503078594712061, "grad_norm": 1.3136753336032572, "learning_rate": 1.1524528978536698e-06, "loss": 0.6528, "step": 23477 }, { "epoch": 0.8503440782325244, "grad_norm": 1.4169507351700454, "learning_rate": 1.15190624233966e-06, "loss": 0.6696, "step": 23478 }, { "epoch": 0.8503802969938428, "grad_norm": 1.3437949723380753, "learning_rate": 1.1513597085831952e-06, "loss": 0.6623, "step": 23479 }, { "epoch": 0.8504165157551612, "grad_norm": 1.3877894559545594, "learning_rate": 1.1508132965917974e-06, "loss": 0.5951, "step": 23480 }, { "epoch": 0.8504527345164795, "grad_norm": 1.0042732077521954, "learning_rate": 1.1502670063729837e-06, "loss": 0.7063, "step": 23481 }, { "epoch": 0.8504889532777979, "grad_norm": 1.3988059147315488, "learning_rate": 1.1497208379342706e-06, "loss": 0.6723, "step": 23482 }, { "epoch": 0.8505251720391163, "grad_norm": 1.369317224298795, "learning_rate": 1.1491747912831763e-06, "loss": 0.7006, "step": 23483 }, { "epoch": 0.8505613908004346, "grad_norm": 1.2759537509279484, "learning_rate": 1.1486288664272138e-06, "loss": 0.6873, "step": 23484 }, { "epoch": 0.850597609561753, "grad_norm": 0.9892891866695137, "learning_rate": 1.1480830633738949e-06, "loss": 0.7162, "step": 23485 }, { "epoch": 0.8506338283230713, "grad_norm": 1.2835901321271919, "learning_rate": 1.1475373821307322e-06, "loss": 0.67, "step": 23486 }, { "epoch": 0.8506700470843898, "grad_norm": 1.390762389286263, "learning_rate": 1.1469918227052335e-06, "loss": 0.6666, "step": 23487 }, { "epoch": 0.8507062658457081, "grad_norm": 1.0672798774044008, "learning_rate": 1.146446385104907e-06, "loss": 0.6729, "step": 23488 }, { "epoch": 0.8507424846070264, "grad_norm": 1.07696280021078, "learning_rate": 1.145901069337254e-06, "loss": 0.6372, "step": 23489 }, { "epoch": 0.8507787033683448, "grad_norm": 1.3655788644618223, "learning_rate": 1.145355875409786e-06, "loss": 0.6981, "step": 23490 }, { "epoch": 0.8508149221296631, "grad_norm": 1.5071897208829996, "learning_rate": 1.144810803330001e-06, "loss": 0.7397, "step": 23491 }, { "epoch": 0.8508511408909816, "grad_norm": 1.4384687621610206, "learning_rate": 1.1442658531053996e-06, "loss": 0.7084, "step": 23492 }, { "epoch": 0.8508873596522999, "grad_norm": 1.0139859361649033, "learning_rate": 1.1437210247434793e-06, "loss": 0.6602, "step": 23493 }, { "epoch": 0.8509235784136182, "grad_norm": 1.4192334392561865, "learning_rate": 1.143176318251743e-06, "loss": 0.6877, "step": 23494 }, { "epoch": 0.8509597971749366, "grad_norm": 1.2904802740305108, "learning_rate": 1.1426317336376812e-06, "loss": 0.6266, "step": 23495 }, { "epoch": 0.850996015936255, "grad_norm": 0.904794081591409, "learning_rate": 1.1420872709087882e-06, "loss": 0.6403, "step": 23496 }, { "epoch": 0.8510322346975734, "grad_norm": 1.3884322735018155, "learning_rate": 1.14154293007256e-06, "loss": 0.6886, "step": 23497 }, { "epoch": 0.8510684534588917, "grad_norm": 1.4078879043261119, "learning_rate": 1.1409987111364863e-06, "loss": 0.7063, "step": 23498 }, { "epoch": 0.85110467222021, "grad_norm": 1.87507661821861, "learning_rate": 1.1404546141080497e-06, "loss": 0.6321, "step": 23499 }, { "epoch": 0.8511408909815285, "grad_norm": 1.4576754999336194, "learning_rate": 1.1399106389947457e-06, "loss": 0.6498, "step": 23500 }, { "epoch": 0.8511771097428468, "grad_norm": 1.1619788638025852, "learning_rate": 1.1393667858040557e-06, "loss": 0.6864, "step": 23501 }, { "epoch": 0.8512133285041652, "grad_norm": 1.4013221228394923, "learning_rate": 1.1388230545434652e-06, "loss": 0.6655, "step": 23502 }, { "epoch": 0.8512495472654835, "grad_norm": 1.2944920834939826, "learning_rate": 1.1382794452204527e-06, "loss": 0.6711, "step": 23503 }, { "epoch": 0.8512857660268018, "grad_norm": 1.387146635261429, "learning_rate": 1.137735957842504e-06, "loss": 0.6581, "step": 23504 }, { "epoch": 0.8513219847881203, "grad_norm": 1.3615112934000524, "learning_rate": 1.1371925924170957e-06, "loss": 0.6424, "step": 23505 }, { "epoch": 0.8513582035494386, "grad_norm": 0.9498089846248189, "learning_rate": 1.1366493489517028e-06, "loss": 0.6836, "step": 23506 }, { "epoch": 0.851394422310757, "grad_norm": 1.2864274727312435, "learning_rate": 1.1361062274538037e-06, "loss": 0.6309, "step": 23507 }, { "epoch": 0.8514306410720753, "grad_norm": 1.0579737006685208, "learning_rate": 1.1355632279308726e-06, "loss": 0.6799, "step": 23508 }, { "epoch": 0.8514668598333937, "grad_norm": 1.3443500113481428, "learning_rate": 1.1350203503903812e-06, "loss": 0.5762, "step": 23509 }, { "epoch": 0.8515030785947121, "grad_norm": 1.4066620832235042, "learning_rate": 1.1344775948397956e-06, "loss": 0.6805, "step": 23510 }, { "epoch": 0.8515392973560304, "grad_norm": 1.392319574336694, "learning_rate": 1.1339349612865912e-06, "loss": 0.65, "step": 23511 }, { "epoch": 0.8515755161173488, "grad_norm": 1.039269009516657, "learning_rate": 1.133392449738232e-06, "loss": 0.6511, "step": 23512 }, { "epoch": 0.8516117348786671, "grad_norm": 1.3745920716779605, "learning_rate": 1.132850060202183e-06, "loss": 0.6478, "step": 23513 }, { "epoch": 0.8516479536399855, "grad_norm": 0.977680396034204, "learning_rate": 1.1323077926859072e-06, "loss": 0.675, "step": 23514 }, { "epoch": 0.8516841724013039, "grad_norm": 1.3326315002813336, "learning_rate": 1.1317656471968696e-06, "loss": 0.631, "step": 23515 }, { "epoch": 0.8517203911626222, "grad_norm": 1.5241418212959452, "learning_rate": 1.13122362374253e-06, "loss": 0.6216, "step": 23516 }, { "epoch": 0.8517566099239406, "grad_norm": 1.237539031335581, "learning_rate": 1.1306817223303424e-06, "loss": 0.6021, "step": 23517 }, { "epoch": 0.851792828685259, "grad_norm": 1.4285864146704648, "learning_rate": 1.1301399429677705e-06, "loss": 0.6076, "step": 23518 }, { "epoch": 0.8518290474465773, "grad_norm": 0.9982271776652719, "learning_rate": 1.1295982856622677e-06, "loss": 0.652, "step": 23519 }, { "epoch": 0.8518652662078957, "grad_norm": 1.4268170970425111, "learning_rate": 1.1290567504212857e-06, "loss": 0.6779, "step": 23520 }, { "epoch": 0.851901484969214, "grad_norm": 1.2573663671151, "learning_rate": 1.1285153372522751e-06, "loss": 0.6685, "step": 23521 }, { "epoch": 0.8519377037305325, "grad_norm": 1.3584066537461859, "learning_rate": 1.1279740461626908e-06, "loss": 0.7271, "step": 23522 }, { "epoch": 0.8519739224918508, "grad_norm": 1.3352320625283365, "learning_rate": 1.1274328771599797e-06, "loss": 0.638, "step": 23523 }, { "epoch": 0.8520101412531691, "grad_norm": 1.471220455522253, "learning_rate": 1.1268918302515884e-06, "loss": 0.6408, "step": 23524 }, { "epoch": 0.8520463600144875, "grad_norm": 1.2558315875262966, "learning_rate": 1.1263509054449596e-06, "loss": 0.6728, "step": 23525 }, { "epoch": 0.8520825787758058, "grad_norm": 1.412412014790782, "learning_rate": 1.125810102747542e-06, "loss": 0.6392, "step": 23526 }, { "epoch": 0.8521187975371243, "grad_norm": 1.3238647973920725, "learning_rate": 1.1252694221667747e-06, "loss": 0.6528, "step": 23527 }, { "epoch": 0.8521550162984426, "grad_norm": 1.387348440415627, "learning_rate": 1.124728863710096e-06, "loss": 0.6733, "step": 23528 }, { "epoch": 0.8521912350597609, "grad_norm": 0.9968663992789157, "learning_rate": 1.124188427384949e-06, "loss": 0.6686, "step": 23529 }, { "epoch": 0.8522274538210793, "grad_norm": 1.3858633116140355, "learning_rate": 1.1236481131987676e-06, "loss": 0.7472, "step": 23530 }, { "epoch": 0.8522636725823977, "grad_norm": 1.386838829187509, "learning_rate": 1.123107921158988e-06, "loss": 0.6976, "step": 23531 }, { "epoch": 0.8522998913437161, "grad_norm": 1.2271151990637539, "learning_rate": 1.1225678512730409e-06, "loss": 0.665, "step": 23532 }, { "epoch": 0.8523361101050344, "grad_norm": 1.4502917954430699, "learning_rate": 1.122027903548364e-06, "loss": 0.6915, "step": 23533 }, { "epoch": 0.8523723288663527, "grad_norm": 1.5195492201578527, "learning_rate": 1.121488077992382e-06, "loss": 0.7119, "step": 23534 }, { "epoch": 0.8524085476276712, "grad_norm": 1.0836554447601048, "learning_rate": 1.120948374612527e-06, "loss": 0.6486, "step": 23535 }, { "epoch": 0.8524447663889895, "grad_norm": 1.440077311200446, "learning_rate": 1.1204087934162211e-06, "loss": 0.6807, "step": 23536 }, { "epoch": 0.8524809851503079, "grad_norm": 1.4104333582837292, "learning_rate": 1.1198693344108957e-06, "loss": 0.6458, "step": 23537 }, { "epoch": 0.8525172039116262, "grad_norm": 1.0778303351068799, "learning_rate": 1.1193299976039695e-06, "loss": 0.6638, "step": 23538 }, { "epoch": 0.8525534226729445, "grad_norm": 1.2756821794231197, "learning_rate": 1.1187907830028644e-06, "loss": 0.5744, "step": 23539 }, { "epoch": 0.852589641434263, "grad_norm": 0.9574846646031552, "learning_rate": 1.1182516906150042e-06, "loss": 0.693, "step": 23540 }, { "epoch": 0.8526258601955813, "grad_norm": 0.9528474169019994, "learning_rate": 1.1177127204478044e-06, "loss": 0.6497, "step": 23541 }, { "epoch": 0.8526620789568997, "grad_norm": 1.080104930104874, "learning_rate": 1.1171738725086833e-06, "loss": 0.6701, "step": 23542 }, { "epoch": 0.852698297718218, "grad_norm": 1.0980579017445458, "learning_rate": 1.1166351468050517e-06, "loss": 0.6941, "step": 23543 }, { "epoch": 0.8527345164795364, "grad_norm": 1.3228549116062116, "learning_rate": 1.1160965433443283e-06, "loss": 0.6461, "step": 23544 }, { "epoch": 0.8527707352408548, "grad_norm": 1.315157643417485, "learning_rate": 1.1155580621339234e-06, "loss": 0.656, "step": 23545 }, { "epoch": 0.8528069540021731, "grad_norm": 1.3150010249479274, "learning_rate": 1.1150197031812427e-06, "loss": 0.6558, "step": 23546 }, { "epoch": 0.8528431727634915, "grad_norm": 1.3473449762191245, "learning_rate": 1.114481466493701e-06, "loss": 0.6416, "step": 23547 }, { "epoch": 0.8528793915248098, "grad_norm": 1.4462852014339407, "learning_rate": 1.1139433520787014e-06, "loss": 0.6696, "step": 23548 }, { "epoch": 0.8529156102861282, "grad_norm": 1.2625844925990781, "learning_rate": 1.113405359943649e-06, "loss": 0.7105, "step": 23549 }, { "epoch": 0.8529518290474466, "grad_norm": 1.3147300170672613, "learning_rate": 1.1128674900959457e-06, "loss": 0.6517, "step": 23550 }, { "epoch": 0.8529880478087649, "grad_norm": 0.9828850627591619, "learning_rate": 1.1123297425429968e-06, "loss": 0.6447, "step": 23551 }, { "epoch": 0.8530242665700833, "grad_norm": 1.3362657432358052, "learning_rate": 1.1117921172921997e-06, "loss": 0.5869, "step": 23552 }, { "epoch": 0.8530604853314017, "grad_norm": 1.136537902393027, "learning_rate": 1.1112546143509528e-06, "loss": 0.6424, "step": 23553 }, { "epoch": 0.85309670409272, "grad_norm": 1.1902734419327812, "learning_rate": 1.1107172337266515e-06, "loss": 0.7045, "step": 23554 }, { "epoch": 0.8531329228540384, "grad_norm": 1.4023305419469885, "learning_rate": 1.110179975426693e-06, "loss": 0.6044, "step": 23555 }, { "epoch": 0.8531691416153567, "grad_norm": 1.5280643919176673, "learning_rate": 1.1096428394584701e-06, "loss": 0.7003, "step": 23556 }, { "epoch": 0.8532053603766752, "grad_norm": 1.3629285874058514, "learning_rate": 1.1091058258293708e-06, "loss": 0.7109, "step": 23557 }, { "epoch": 0.8532415791379935, "grad_norm": 1.5370888314819222, "learning_rate": 1.1085689345467899e-06, "loss": 0.7209, "step": 23558 }, { "epoch": 0.8532777978993118, "grad_norm": 0.9894178953581394, "learning_rate": 1.1080321656181125e-06, "loss": 0.7132, "step": 23559 }, { "epoch": 0.8533140166606302, "grad_norm": 1.0124141335295338, "learning_rate": 1.1074955190507264e-06, "loss": 0.6809, "step": 23560 }, { "epoch": 0.8533502354219485, "grad_norm": 1.4230599142833753, "learning_rate": 1.1069589948520131e-06, "loss": 0.6631, "step": 23561 }, { "epoch": 0.853386454183267, "grad_norm": 2.6979839161306143, "learning_rate": 1.1064225930293616e-06, "loss": 0.6491, "step": 23562 }, { "epoch": 0.8534226729445853, "grad_norm": 1.305067167345643, "learning_rate": 1.1058863135901476e-06, "loss": 0.6566, "step": 23563 }, { "epoch": 0.8534588917059036, "grad_norm": 1.3931502802669935, "learning_rate": 1.1053501565417545e-06, "loss": 0.6844, "step": 23564 }, { "epoch": 0.853495110467222, "grad_norm": 1.304476017001444, "learning_rate": 1.1048141218915565e-06, "loss": 0.6193, "step": 23565 }, { "epoch": 0.8535313292285404, "grad_norm": 1.7185895584822501, "learning_rate": 1.104278209646934e-06, "loss": 0.6067, "step": 23566 }, { "epoch": 0.8535675479898588, "grad_norm": 1.4446232602136695, "learning_rate": 1.103742419815259e-06, "loss": 0.7209, "step": 23567 }, { "epoch": 0.8536037667511771, "grad_norm": 1.3434686698944607, "learning_rate": 1.1032067524039047e-06, "loss": 0.6259, "step": 23568 }, { "epoch": 0.8536399855124954, "grad_norm": 1.3743435751753685, "learning_rate": 1.1026712074202439e-06, "loss": 0.6401, "step": 23569 }, { "epoch": 0.8536762042738139, "grad_norm": 1.2504267839431, "learning_rate": 1.1021357848716462e-06, "loss": 0.6349, "step": 23570 }, { "epoch": 0.8537124230351322, "grad_norm": 1.5597663163092352, "learning_rate": 1.101600484765477e-06, "loss": 0.6398, "step": 23571 }, { "epoch": 0.8537486417964506, "grad_norm": 1.5444131042394906, "learning_rate": 1.1010653071091027e-06, "loss": 0.6838, "step": 23572 }, { "epoch": 0.8537848605577689, "grad_norm": 1.4999441096701724, "learning_rate": 1.1005302519098915e-06, "loss": 0.6022, "step": 23573 }, { "epoch": 0.8538210793190872, "grad_norm": 0.9628287229821222, "learning_rate": 1.0999953191752034e-06, "loss": 0.6579, "step": 23574 }, { "epoch": 0.8538572980804057, "grad_norm": 1.4371986837858126, "learning_rate": 1.0994605089123999e-06, "loss": 0.6895, "step": 23575 }, { "epoch": 0.853893516841724, "grad_norm": 1.034653045484568, "learning_rate": 1.0989258211288379e-06, "loss": 0.6543, "step": 23576 }, { "epoch": 0.8539297356030424, "grad_norm": 1.395327673217616, "learning_rate": 1.098391255831881e-06, "loss": 0.7631, "step": 23577 }, { "epoch": 0.8539659543643607, "grad_norm": 1.4617806260975132, "learning_rate": 1.0978568130288813e-06, "loss": 0.7127, "step": 23578 }, { "epoch": 0.854002173125679, "grad_norm": 1.0749923540917026, "learning_rate": 1.0973224927271919e-06, "loss": 0.6454, "step": 23579 }, { "epoch": 0.8540383918869975, "grad_norm": 0.9819501632864968, "learning_rate": 1.0967882949341701e-06, "loss": 0.627, "step": 23580 }, { "epoch": 0.8540746106483158, "grad_norm": 1.7659091762970835, "learning_rate": 1.0962542196571636e-06, "loss": 0.6717, "step": 23581 }, { "epoch": 0.8541108294096342, "grad_norm": 1.351237238626096, "learning_rate": 1.0957202669035228e-06, "loss": 0.7331, "step": 23582 }, { "epoch": 0.8541470481709525, "grad_norm": 1.3442207471076013, "learning_rate": 1.0951864366805931e-06, "loss": 0.6594, "step": 23583 }, { "epoch": 0.8541832669322709, "grad_norm": 1.2675160637491023, "learning_rate": 1.0946527289957255e-06, "loss": 0.6638, "step": 23584 }, { "epoch": 0.8542194856935893, "grad_norm": 1.302116602154736, "learning_rate": 1.0941191438562604e-06, "loss": 0.6418, "step": 23585 }, { "epoch": 0.8542557044549076, "grad_norm": 1.377995517842503, "learning_rate": 1.0935856812695412e-06, "loss": 0.615, "step": 23586 }, { "epoch": 0.854291923216226, "grad_norm": 1.4268444238649975, "learning_rate": 1.0930523412429072e-06, "loss": 0.6808, "step": 23587 }, { "epoch": 0.8543281419775444, "grad_norm": 1.3456555449458085, "learning_rate": 1.0925191237837007e-06, "loss": 0.6443, "step": 23588 }, { "epoch": 0.8543643607388627, "grad_norm": 1.4212620063566495, "learning_rate": 1.0919860288992578e-06, "loss": 0.6567, "step": 23589 }, { "epoch": 0.8544005795001811, "grad_norm": 1.0402325900665403, "learning_rate": 1.0914530565969139e-06, "loss": 0.7661, "step": 23590 }, { "epoch": 0.8544367982614994, "grad_norm": 1.416229222728997, "learning_rate": 1.090920206884004e-06, "loss": 0.6357, "step": 23591 }, { "epoch": 0.8544730170228179, "grad_norm": 1.662922840554592, "learning_rate": 1.0903874797678616e-06, "loss": 0.6494, "step": 23592 }, { "epoch": 0.8545092357841362, "grad_norm": 1.4753803031626986, "learning_rate": 1.0898548752558148e-06, "loss": 0.6782, "step": 23593 }, { "epoch": 0.8545454545454545, "grad_norm": 1.3852374318232046, "learning_rate": 1.0893223933551932e-06, "loss": 0.6751, "step": 23594 }, { "epoch": 0.8545816733067729, "grad_norm": 0.9783223452863938, "learning_rate": 1.0887900340733271e-06, "loss": 0.706, "step": 23595 }, { "epoch": 0.8546178920680912, "grad_norm": 1.2951591642523925, "learning_rate": 1.088257797417539e-06, "loss": 0.6756, "step": 23596 }, { "epoch": 0.8546541108294097, "grad_norm": 1.478752535505854, "learning_rate": 1.0877256833951556e-06, "loss": 0.7132, "step": 23597 }, { "epoch": 0.854690329590728, "grad_norm": 1.5603090159887087, "learning_rate": 1.087193692013494e-06, "loss": 0.6655, "step": 23598 }, { "epoch": 0.8547265483520463, "grad_norm": 1.3586114665648172, "learning_rate": 1.086661823279882e-06, "loss": 0.7088, "step": 23599 }, { "epoch": 0.8547627671133647, "grad_norm": 1.545700929354327, "learning_rate": 1.0861300772016347e-06, "loss": 0.628, "step": 23600 }, { "epoch": 0.8547989858746831, "grad_norm": 1.3654266216607611, "learning_rate": 1.0855984537860686e-06, "loss": 0.6356, "step": 23601 }, { "epoch": 0.8548352046360015, "grad_norm": 1.430525127864532, "learning_rate": 1.0850669530405022e-06, "loss": 0.6808, "step": 23602 }, { "epoch": 0.8548714233973198, "grad_norm": 1.3841886051268044, "learning_rate": 1.0845355749722486e-06, "loss": 0.7199, "step": 23603 }, { "epoch": 0.8549076421586381, "grad_norm": 1.2469906240049018, "learning_rate": 1.0840043195886185e-06, "loss": 0.6166, "step": 23604 }, { "epoch": 0.8549438609199566, "grad_norm": 1.655429843951278, "learning_rate": 1.0834731868969217e-06, "loss": 0.7422, "step": 23605 }, { "epoch": 0.8549800796812749, "grad_norm": 1.50498632932748, "learning_rate": 1.0829421769044702e-06, "loss": 0.6698, "step": 23606 }, { "epoch": 0.8550162984425933, "grad_norm": 1.334242300815689, "learning_rate": 1.082411289618569e-06, "loss": 0.7162, "step": 23607 }, { "epoch": 0.8550525172039116, "grad_norm": 1.2190046159082923, "learning_rate": 1.0818805250465237e-06, "loss": 0.59, "step": 23608 }, { "epoch": 0.8550887359652299, "grad_norm": 1.012126543147607, "learning_rate": 1.0813498831956393e-06, "loss": 0.6897, "step": 23609 }, { "epoch": 0.8551249547265484, "grad_norm": 1.4036501096744072, "learning_rate": 1.0808193640732178e-06, "loss": 0.6626, "step": 23610 }, { "epoch": 0.8551611734878667, "grad_norm": 1.2958016614560932, "learning_rate": 1.080288967686558e-06, "loss": 0.6003, "step": 23611 }, { "epoch": 0.8551973922491851, "grad_norm": 1.0911166473532385, "learning_rate": 1.0797586940429582e-06, "loss": 0.6381, "step": 23612 }, { "epoch": 0.8552336110105034, "grad_norm": 1.7790040891217653, "learning_rate": 1.0792285431497184e-06, "loss": 0.7408, "step": 23613 }, { "epoch": 0.8552698297718218, "grad_norm": 1.398855751508341, "learning_rate": 1.0786985150141327e-06, "loss": 0.6811, "step": 23614 }, { "epoch": 0.8553060485331402, "grad_norm": 1.2917106617961225, "learning_rate": 1.0781686096434952e-06, "loss": 0.6515, "step": 23615 }, { "epoch": 0.8553422672944585, "grad_norm": 1.347395994791946, "learning_rate": 1.0776388270450933e-06, "loss": 0.6371, "step": 23616 }, { "epoch": 0.8553784860557769, "grad_norm": 1.2404147527591787, "learning_rate": 1.0771091672262246e-06, "loss": 0.6416, "step": 23617 }, { "epoch": 0.8554147048170953, "grad_norm": 1.4873589877099351, "learning_rate": 1.0765796301941734e-06, "loss": 0.7704, "step": 23618 }, { "epoch": 0.8554509235784136, "grad_norm": 1.3970108000352177, "learning_rate": 1.0760502159562247e-06, "loss": 0.6479, "step": 23619 }, { "epoch": 0.855487142339732, "grad_norm": 1.3738180690274784, "learning_rate": 1.0755209245196695e-06, "loss": 0.6911, "step": 23620 }, { "epoch": 0.8555233611010503, "grad_norm": 1.4380712708649186, "learning_rate": 1.0749917558917867e-06, "loss": 0.687, "step": 23621 }, { "epoch": 0.8555595798623687, "grad_norm": 1.3744973949275512, "learning_rate": 1.0744627100798611e-06, "loss": 0.6303, "step": 23622 }, { "epoch": 0.8555957986236871, "grad_norm": 1.4298823880425686, "learning_rate": 1.0739337870911681e-06, "loss": 0.6223, "step": 23623 }, { "epoch": 0.8556320173850054, "grad_norm": 1.3045562846620953, "learning_rate": 1.073404986932992e-06, "loss": 0.6138, "step": 23624 }, { "epoch": 0.8556682361463238, "grad_norm": 1.1872161436566677, "learning_rate": 1.072876309612606e-06, "loss": 0.6995, "step": 23625 }, { "epoch": 0.8557044549076421, "grad_norm": 1.3420097759650425, "learning_rate": 1.0723477551372873e-06, "loss": 0.6705, "step": 23626 }, { "epoch": 0.8557406736689606, "grad_norm": 1.1125706592466296, "learning_rate": 1.0718193235143048e-06, "loss": 0.6442, "step": 23627 }, { "epoch": 0.8557768924302789, "grad_norm": 1.3818400622195093, "learning_rate": 1.0712910147509358e-06, "loss": 0.6695, "step": 23628 }, { "epoch": 0.8558131111915972, "grad_norm": 1.4066781292984283, "learning_rate": 1.0707628288544481e-06, "loss": 0.6845, "step": 23629 }, { "epoch": 0.8558493299529156, "grad_norm": 0.9965313397294365, "learning_rate": 1.0702347658321077e-06, "loss": 0.6308, "step": 23630 }, { "epoch": 0.855885548714234, "grad_norm": 1.3176836394237748, "learning_rate": 1.0697068256911859e-06, "loss": 0.688, "step": 23631 }, { "epoch": 0.8559217674755524, "grad_norm": 0.9579652775961649, "learning_rate": 1.0691790084389443e-06, "loss": 0.6505, "step": 23632 }, { "epoch": 0.8559579862368707, "grad_norm": 0.889663519310122, "learning_rate": 1.0686513140826472e-06, "loss": 0.6646, "step": 23633 }, { "epoch": 0.855994204998189, "grad_norm": 1.4063285448073608, "learning_rate": 1.0681237426295531e-06, "loss": 0.7234, "step": 23634 }, { "epoch": 0.8560304237595074, "grad_norm": 1.165323619193594, "learning_rate": 1.0675962940869266e-06, "loss": 0.6278, "step": 23635 }, { "epoch": 0.8560666425208258, "grad_norm": 1.4073186793941959, "learning_rate": 1.067068968462025e-06, "loss": 0.6607, "step": 23636 }, { "epoch": 0.8561028612821442, "grad_norm": 1.4401100664006448, "learning_rate": 1.0665417657621025e-06, "loss": 0.6437, "step": 23637 }, { "epoch": 0.8561390800434625, "grad_norm": 1.4340290453340512, "learning_rate": 1.0660146859944131e-06, "loss": 0.6424, "step": 23638 }, { "epoch": 0.8561752988047808, "grad_norm": 1.1789881471426544, "learning_rate": 1.0654877291662136e-06, "loss": 0.6951, "step": 23639 }, { "epoch": 0.8562115175660993, "grad_norm": 1.3366366628750557, "learning_rate": 1.0649608952847535e-06, "loss": 0.7707, "step": 23640 }, { "epoch": 0.8562477363274176, "grad_norm": 1.3734869692554557, "learning_rate": 1.0644341843572815e-06, "loss": 0.6222, "step": 23641 }, { "epoch": 0.856283955088736, "grad_norm": 1.2580354541820888, "learning_rate": 1.0639075963910473e-06, "loss": 0.6478, "step": 23642 }, { "epoch": 0.8563201738500543, "grad_norm": 1.4138127173533677, "learning_rate": 1.0633811313932974e-06, "loss": 0.6727, "step": 23643 }, { "epoch": 0.8563563926113726, "grad_norm": 1.4693391409700485, "learning_rate": 1.062854789371275e-06, "loss": 0.6902, "step": 23644 }, { "epoch": 0.8563926113726911, "grad_norm": 0.8832238021048385, "learning_rate": 1.0623285703322216e-06, "loss": 0.6767, "step": 23645 }, { "epoch": 0.8564288301340094, "grad_norm": 2.7091023069480826, "learning_rate": 1.061802474283382e-06, "loss": 0.7351, "step": 23646 }, { "epoch": 0.8564650488953278, "grad_norm": 1.3808082546823617, "learning_rate": 1.0612765012319948e-06, "loss": 0.7296, "step": 23647 }, { "epoch": 0.8565012676566461, "grad_norm": 1.5435633419102688, "learning_rate": 1.0607506511852972e-06, "loss": 0.671, "step": 23648 }, { "epoch": 0.8565374864179645, "grad_norm": 1.0966430114800347, "learning_rate": 1.0602249241505224e-06, "loss": 0.6762, "step": 23649 }, { "epoch": 0.8565737051792829, "grad_norm": 1.3545914038120517, "learning_rate": 1.0596993201349115e-06, "loss": 0.6414, "step": 23650 }, { "epoch": 0.8566099239406012, "grad_norm": 1.4952957985980675, "learning_rate": 1.0591738391456918e-06, "loss": 0.7202, "step": 23651 }, { "epoch": 0.8566461427019196, "grad_norm": 1.0142380418081884, "learning_rate": 1.0586484811900955e-06, "loss": 0.6521, "step": 23652 }, { "epoch": 0.856682361463238, "grad_norm": 1.300201106936021, "learning_rate": 1.0581232462753544e-06, "loss": 0.6816, "step": 23653 }, { "epoch": 0.8567185802245563, "grad_norm": 1.2919562349918117, "learning_rate": 1.0575981344086939e-06, "loss": 0.5982, "step": 23654 }, { "epoch": 0.8567547989858747, "grad_norm": 1.3889937443607367, "learning_rate": 1.0570731455973415e-06, "loss": 0.7375, "step": 23655 }, { "epoch": 0.856791017747193, "grad_norm": 0.9889987285495506, "learning_rate": 1.0565482798485172e-06, "loss": 0.6869, "step": 23656 }, { "epoch": 0.8568272365085114, "grad_norm": 1.4151543688079808, "learning_rate": 1.0560235371694505e-06, "loss": 0.6637, "step": 23657 }, { "epoch": 0.8568634552698298, "grad_norm": 1.4748525532676868, "learning_rate": 1.055498917567358e-06, "loss": 0.6939, "step": 23658 }, { "epoch": 0.8568996740311481, "grad_norm": 1.320896181902496, "learning_rate": 1.0549744210494573e-06, "loss": 0.6197, "step": 23659 }, { "epoch": 0.8569358927924665, "grad_norm": 1.006578276345877, "learning_rate": 1.0544500476229713e-06, "loss": 0.732, "step": 23660 }, { "epoch": 0.8569721115537848, "grad_norm": 0.9806526553807862, "learning_rate": 1.0539257972951123e-06, "loss": 0.6647, "step": 23661 }, { "epoch": 0.8570083303151033, "grad_norm": 1.350567142037897, "learning_rate": 1.0534016700730954e-06, "loss": 0.7027, "step": 23662 }, { "epoch": 0.8570445490764216, "grad_norm": 1.3044613860689431, "learning_rate": 1.0528776659641305e-06, "loss": 0.645, "step": 23663 }, { "epoch": 0.8570807678377399, "grad_norm": 1.3550330370079242, "learning_rate": 1.0523537849754316e-06, "loss": 0.7047, "step": 23664 }, { "epoch": 0.8571169865990583, "grad_norm": 1.3988308053788596, "learning_rate": 1.051830027114208e-06, "loss": 0.6192, "step": 23665 }, { "epoch": 0.8571532053603766, "grad_norm": 1.4784680652401943, "learning_rate": 1.0513063923876655e-06, "loss": 0.7653, "step": 23666 }, { "epoch": 0.8571894241216951, "grad_norm": 1.3088231924662312, "learning_rate": 1.0507828808030075e-06, "loss": 0.6101, "step": 23667 }, { "epoch": 0.8572256428830134, "grad_norm": 1.4625971732807708, "learning_rate": 1.0502594923674437e-06, "loss": 0.7031, "step": 23668 }, { "epoch": 0.8572618616443317, "grad_norm": 1.4578388471358839, "learning_rate": 1.0497362270881718e-06, "loss": 0.663, "step": 23669 }, { "epoch": 0.8572980804056501, "grad_norm": 1.3775474442616824, "learning_rate": 1.0492130849723924e-06, "loss": 0.599, "step": 23670 }, { "epoch": 0.8573342991669685, "grad_norm": 1.365679426876328, "learning_rate": 1.0486900660273081e-06, "loss": 0.7236, "step": 23671 }, { "epoch": 0.8573705179282869, "grad_norm": 1.432117063471422, "learning_rate": 1.0481671702601138e-06, "loss": 0.6553, "step": 23672 }, { "epoch": 0.8574067366896052, "grad_norm": 1.2114711252895585, "learning_rate": 1.047644397678005e-06, "loss": 0.6187, "step": 23673 }, { "epoch": 0.8574429554509235, "grad_norm": 1.4311478158234285, "learning_rate": 1.0471217482881723e-06, "loss": 0.677, "step": 23674 }, { "epoch": 0.857479174212242, "grad_norm": 2.4853908789584542, "learning_rate": 1.046599222097815e-06, "loss": 0.6494, "step": 23675 }, { "epoch": 0.8575153929735603, "grad_norm": 0.9749022135380533, "learning_rate": 1.046076819114118e-06, "loss": 0.6924, "step": 23676 }, { "epoch": 0.8575516117348787, "grad_norm": 1.0659522114643063, "learning_rate": 1.0455545393442722e-06, "loss": 0.6806, "step": 23677 }, { "epoch": 0.857587830496197, "grad_norm": 1.3561489297743226, "learning_rate": 1.0450323827954623e-06, "loss": 0.6421, "step": 23678 }, { "epoch": 0.8576240492575153, "grad_norm": 1.325036108198612, "learning_rate": 1.0445103494748766e-06, "loss": 0.6873, "step": 23679 }, { "epoch": 0.8576602680188338, "grad_norm": 1.3237859129683207, "learning_rate": 1.0439884393896982e-06, "loss": 0.6573, "step": 23680 }, { "epoch": 0.8576964867801521, "grad_norm": 1.4242689283668568, "learning_rate": 1.043466652547106e-06, "loss": 0.627, "step": 23681 }, { "epoch": 0.8577327055414705, "grad_norm": 1.4489974690974485, "learning_rate": 1.0429449889542842e-06, "loss": 0.6316, "step": 23682 }, { "epoch": 0.8577689243027888, "grad_norm": 1.297807406412725, "learning_rate": 1.0424234486184104e-06, "loss": 0.6581, "step": 23683 }, { "epoch": 0.8578051430641072, "grad_norm": 1.6162075803324243, "learning_rate": 1.04190203154666e-06, "loss": 0.6384, "step": 23684 }, { "epoch": 0.8578413618254256, "grad_norm": 1.2465209635932686, "learning_rate": 1.041380737746207e-06, "loss": 0.6339, "step": 23685 }, { "epoch": 0.8578775805867439, "grad_norm": 1.2795152754397263, "learning_rate": 1.0408595672242293e-06, "loss": 0.6627, "step": 23686 }, { "epoch": 0.8579137993480623, "grad_norm": 1.4174845777892904, "learning_rate": 1.0403385199878957e-06, "loss": 0.7296, "step": 23687 }, { "epoch": 0.8579500181093807, "grad_norm": 0.9721737401449446, "learning_rate": 1.039817596044378e-06, "loss": 0.7321, "step": 23688 }, { "epoch": 0.857986236870699, "grad_norm": 1.4155581399568424, "learning_rate": 1.0392967954008394e-06, "loss": 0.5997, "step": 23689 }, { "epoch": 0.8580224556320174, "grad_norm": 1.3044088460107213, "learning_rate": 1.038776118064454e-06, "loss": 0.631, "step": 23690 }, { "epoch": 0.8580586743933357, "grad_norm": 1.053572651063517, "learning_rate": 1.038255564042382e-06, "loss": 0.6594, "step": 23691 }, { "epoch": 0.8580948931546541, "grad_norm": 1.4283150882643663, "learning_rate": 1.0377351333417862e-06, "loss": 0.6399, "step": 23692 }, { "epoch": 0.8581311119159725, "grad_norm": 1.313645202418119, "learning_rate": 1.0372148259698323e-06, "loss": 0.6462, "step": 23693 }, { "epoch": 0.8581673306772908, "grad_norm": 1.313384811571891, "learning_rate": 1.0366946419336775e-06, "loss": 0.6107, "step": 23694 }, { "epoch": 0.8582035494386092, "grad_norm": 1.0660738933625564, "learning_rate": 1.036174581240481e-06, "loss": 0.6309, "step": 23695 }, { "epoch": 0.8582397681999275, "grad_norm": 1.5918146541983336, "learning_rate": 1.0356546438973958e-06, "loss": 0.6528, "step": 23696 }, { "epoch": 0.858275986961246, "grad_norm": 1.0692430086777818, "learning_rate": 1.0351348299115816e-06, "loss": 0.7767, "step": 23697 }, { "epoch": 0.8583122057225643, "grad_norm": 1.2914129538861117, "learning_rate": 1.0346151392901892e-06, "loss": 0.6517, "step": 23698 }, { "epoch": 0.8583484244838826, "grad_norm": 0.9656507626621187, "learning_rate": 1.034095572040371e-06, "loss": 0.6937, "step": 23699 }, { "epoch": 0.858384643245201, "grad_norm": 1.3510421856666202, "learning_rate": 1.0335761281692736e-06, "loss": 0.6823, "step": 23700 }, { "epoch": 0.8584208620065193, "grad_norm": 1.2929879429133162, "learning_rate": 1.0330568076840486e-06, "loss": 0.6952, "step": 23701 }, { "epoch": 0.8584570807678378, "grad_norm": 1.3191058207374855, "learning_rate": 1.0325376105918417e-06, "loss": 0.6244, "step": 23702 }, { "epoch": 0.8584932995291561, "grad_norm": 1.4409526354123685, "learning_rate": 1.0320185368997937e-06, "loss": 0.7028, "step": 23703 }, { "epoch": 0.8585295182904744, "grad_norm": 1.0437392178232137, "learning_rate": 1.0314995866150535e-06, "loss": 0.6968, "step": 23704 }, { "epoch": 0.8585657370517928, "grad_norm": 1.301225784597023, "learning_rate": 1.0309807597447585e-06, "loss": 0.6137, "step": 23705 }, { "epoch": 0.8586019558131112, "grad_norm": 1.362134259682823, "learning_rate": 1.0304620562960498e-06, "loss": 0.6193, "step": 23706 }, { "epoch": 0.8586381745744296, "grad_norm": 1.549847674960468, "learning_rate": 1.0299434762760629e-06, "loss": 0.7175, "step": 23707 }, { "epoch": 0.8586743933357479, "grad_norm": 1.110187699022562, "learning_rate": 1.0294250196919375e-06, "loss": 0.6906, "step": 23708 }, { "epoch": 0.8587106120970662, "grad_norm": 1.2996123242853568, "learning_rate": 1.0289066865508079e-06, "loss": 0.7, "step": 23709 }, { "epoch": 0.8587468308583847, "grad_norm": 1.1526516327284584, "learning_rate": 1.0283884768598008e-06, "loss": 0.6745, "step": 23710 }, { "epoch": 0.858783049619703, "grad_norm": 1.2861516729549112, "learning_rate": 1.0278703906260534e-06, "loss": 0.671, "step": 23711 }, { "epoch": 0.8588192683810214, "grad_norm": 1.3948111173146263, "learning_rate": 1.0273524278566927e-06, "loss": 0.6491, "step": 23712 }, { "epoch": 0.8588554871423397, "grad_norm": 1.2581060463879508, "learning_rate": 1.026834588558845e-06, "loss": 0.6155, "step": 23713 }, { "epoch": 0.858891705903658, "grad_norm": 0.8793436896328484, "learning_rate": 1.02631687273964e-06, "loss": 0.7143, "step": 23714 }, { "epoch": 0.8589279246649765, "grad_norm": 1.311627581985555, "learning_rate": 1.0257992804062e-06, "loss": 0.6538, "step": 23715 }, { "epoch": 0.8589641434262948, "grad_norm": 0.9723312643378741, "learning_rate": 1.025281811565646e-06, "loss": 0.6495, "step": 23716 }, { "epoch": 0.8590003621876132, "grad_norm": 1.0025805644073884, "learning_rate": 1.0247644662251e-06, "loss": 0.6689, "step": 23717 }, { "epoch": 0.8590365809489315, "grad_norm": 1.3234818174717735, "learning_rate": 1.0242472443916817e-06, "loss": 0.6424, "step": 23718 }, { "epoch": 0.8590727997102499, "grad_norm": 1.3666750206258294, "learning_rate": 1.0237301460725091e-06, "loss": 0.6447, "step": 23719 }, { "epoch": 0.8591090184715683, "grad_norm": 1.5600010940390503, "learning_rate": 1.0232131712746973e-06, "loss": 0.7091, "step": 23720 }, { "epoch": 0.8591452372328866, "grad_norm": 0.9813371082400597, "learning_rate": 1.0226963200053564e-06, "loss": 0.6742, "step": 23721 }, { "epoch": 0.859181455994205, "grad_norm": 1.3296724161431104, "learning_rate": 1.022179592271606e-06, "loss": 0.6937, "step": 23722 }, { "epoch": 0.8592176747555234, "grad_norm": 1.511566711389946, "learning_rate": 1.0216629880805517e-06, "loss": 0.5632, "step": 23723 }, { "epoch": 0.8592538935168417, "grad_norm": 1.0553377304462814, "learning_rate": 1.0211465074393023e-06, "loss": 0.6635, "step": 23724 }, { "epoch": 0.8592901122781601, "grad_norm": 1.0408196925801918, "learning_rate": 1.0206301503549688e-06, "loss": 0.7093, "step": 23725 }, { "epoch": 0.8593263310394784, "grad_norm": 1.3473248751589104, "learning_rate": 1.020113916834654e-06, "loss": 0.6879, "step": 23726 }, { "epoch": 0.8593625498007968, "grad_norm": 1.4183641045443454, "learning_rate": 1.0195978068854618e-06, "loss": 0.6755, "step": 23727 }, { "epoch": 0.8593987685621152, "grad_norm": 1.359545190066933, "learning_rate": 1.0190818205144938e-06, "loss": 0.6765, "step": 23728 }, { "epoch": 0.8594349873234335, "grad_norm": 1.325724649935169, "learning_rate": 1.018565957728852e-06, "loss": 0.621, "step": 23729 }, { "epoch": 0.8594712060847519, "grad_norm": 1.4271312416839712, "learning_rate": 1.0180502185356345e-06, "loss": 0.654, "step": 23730 }, { "epoch": 0.8595074248460702, "grad_norm": 1.2686028922879558, "learning_rate": 1.0175346029419387e-06, "loss": 0.6595, "step": 23731 }, { "epoch": 0.8595436436073887, "grad_norm": 1.3303787428078693, "learning_rate": 1.0170191109548577e-06, "loss": 0.6971, "step": 23732 }, { "epoch": 0.859579862368707, "grad_norm": 1.0163951432551648, "learning_rate": 1.0165037425814883e-06, "loss": 0.711, "step": 23733 }, { "epoch": 0.8596160811300253, "grad_norm": 1.5775529395868997, "learning_rate": 1.0159884978289215e-06, "loss": 0.6989, "step": 23734 }, { "epoch": 0.8596522998913437, "grad_norm": 1.5482152264941098, "learning_rate": 1.0154733767042446e-06, "loss": 0.5657, "step": 23735 }, { "epoch": 0.859688518652662, "grad_norm": 1.3507732556593557, "learning_rate": 1.0149583792145502e-06, "loss": 0.6557, "step": 23736 }, { "epoch": 0.8597247374139805, "grad_norm": 1.4130785651235442, "learning_rate": 1.0144435053669244e-06, "loss": 0.614, "step": 23737 }, { "epoch": 0.8597609561752988, "grad_norm": 1.4435403320688773, "learning_rate": 1.0139287551684507e-06, "loss": 0.6779, "step": 23738 }, { "epoch": 0.8597971749366171, "grad_norm": 1.340083671447817, "learning_rate": 1.0134141286262112e-06, "loss": 0.6814, "step": 23739 }, { "epoch": 0.8598333936979355, "grad_norm": 0.9660913080108127, "learning_rate": 1.0128996257472924e-06, "loss": 0.673, "step": 23740 }, { "epoch": 0.8598696124592539, "grad_norm": 1.0240832124999049, "learning_rate": 1.0123852465387706e-06, "loss": 0.6533, "step": 23741 }, { "epoch": 0.8599058312205723, "grad_norm": 1.3244449568558312, "learning_rate": 1.011870991007725e-06, "loss": 0.6718, "step": 23742 }, { "epoch": 0.8599420499818906, "grad_norm": 1.4557834976500754, "learning_rate": 1.011356859161231e-06, "loss": 0.5998, "step": 23743 }, { "epoch": 0.8599782687432089, "grad_norm": 1.3901508841079273, "learning_rate": 1.0108428510063661e-06, "loss": 0.6899, "step": 23744 }, { "epoch": 0.8600144875045274, "grad_norm": 1.278334526330144, "learning_rate": 1.0103289665502025e-06, "loss": 0.6391, "step": 23745 }, { "epoch": 0.8600507062658457, "grad_norm": 1.3782719741892315, "learning_rate": 1.009815205799809e-06, "loss": 0.6611, "step": 23746 }, { "epoch": 0.8600869250271641, "grad_norm": 1.2712169042138433, "learning_rate": 1.0093015687622598e-06, "loss": 0.6649, "step": 23747 }, { "epoch": 0.8601231437884824, "grad_norm": 1.3555473392800828, "learning_rate": 1.0087880554446205e-06, "loss": 0.674, "step": 23748 }, { "epoch": 0.8601593625498007, "grad_norm": 1.0334944244536375, "learning_rate": 1.0082746658539588e-06, "loss": 0.6969, "step": 23749 }, { "epoch": 0.8601955813111192, "grad_norm": 1.288123217745274, "learning_rate": 1.0077613999973357e-06, "loss": 0.5921, "step": 23750 }, { "epoch": 0.8602318000724375, "grad_norm": 1.4444991897080708, "learning_rate": 1.0072482578818198e-06, "loss": 0.6708, "step": 23751 }, { "epoch": 0.8602680188337559, "grad_norm": 1.273218108948825, "learning_rate": 1.0067352395144692e-06, "loss": 0.6234, "step": 23752 }, { "epoch": 0.8603042375950742, "grad_norm": 1.2890801905371807, "learning_rate": 1.0062223449023446e-06, "loss": 0.697, "step": 23753 }, { "epoch": 0.8603404563563926, "grad_norm": 1.4782464111116984, "learning_rate": 1.0057095740525003e-06, "loss": 0.6686, "step": 23754 }, { "epoch": 0.860376675117711, "grad_norm": 1.3611185379089934, "learning_rate": 1.0051969269719974e-06, "loss": 0.7034, "step": 23755 }, { "epoch": 0.8604128938790293, "grad_norm": 0.9925555426332394, "learning_rate": 1.0046844036678882e-06, "loss": 0.7001, "step": 23756 }, { "epoch": 0.8604491126403477, "grad_norm": 1.0716546578779755, "learning_rate": 1.0041720041472236e-06, "loss": 0.6605, "step": 23757 }, { "epoch": 0.860485331401666, "grad_norm": 1.278505217232724, "learning_rate": 1.003659728417059e-06, "loss": 0.6647, "step": 23758 }, { "epoch": 0.8605215501629844, "grad_norm": 0.876865781924573, "learning_rate": 1.0031475764844412e-06, "loss": 0.6172, "step": 23759 }, { "epoch": 0.8605577689243028, "grad_norm": 1.4358523420116684, "learning_rate": 1.0026355483564176e-06, "loss": 0.7473, "step": 23760 }, { "epoch": 0.8605939876856211, "grad_norm": 1.496391657658257, "learning_rate": 1.0021236440400339e-06, "loss": 0.664, "step": 23761 }, { "epoch": 0.8606302064469395, "grad_norm": 1.3537856709209157, "learning_rate": 1.0016118635423366e-06, "loss": 0.7096, "step": 23762 }, { "epoch": 0.8606664252082579, "grad_norm": 1.129671013946012, "learning_rate": 1.0011002068703657e-06, "loss": 0.6547, "step": 23763 }, { "epoch": 0.8607026439695762, "grad_norm": 1.3849160693755955, "learning_rate": 1.0005886740311621e-06, "loss": 0.7058, "step": 23764 }, { "epoch": 0.8607388627308946, "grad_norm": 1.4293956575977589, "learning_rate": 1.0000772650317682e-06, "loss": 0.7721, "step": 23765 }, { "epoch": 0.8607750814922129, "grad_norm": 1.4623932451273993, "learning_rate": 9.99565979879218e-07, "loss": 0.6602, "step": 23766 }, { "epoch": 0.8608113002535314, "grad_norm": 1.4122299555716982, "learning_rate": 9.990548185805493e-07, "loss": 0.6774, "step": 23767 }, { "epoch": 0.8608475190148497, "grad_norm": 0.9393786928888274, "learning_rate": 9.985437811427934e-07, "loss": 0.652, "step": 23768 }, { "epoch": 0.860883737776168, "grad_norm": 1.3713387524461338, "learning_rate": 9.980328675729855e-07, "loss": 0.6348, "step": 23769 }, { "epoch": 0.8609199565374864, "grad_norm": 1.3683685630481155, "learning_rate": 9.975220778781558e-07, "loss": 0.6728, "step": 23770 }, { "epoch": 0.8609561752988047, "grad_norm": 0.9746782553485636, "learning_rate": 9.97011412065333e-07, "loss": 0.6196, "step": 23771 }, { "epoch": 0.8609923940601232, "grad_norm": 1.2887319863583935, "learning_rate": 9.965008701415412e-07, "loss": 0.604, "step": 23772 }, { "epoch": 0.8610286128214415, "grad_norm": 1.2133030782335403, "learning_rate": 9.95990452113811e-07, "loss": 0.7098, "step": 23773 }, { "epoch": 0.8610648315827598, "grad_norm": 0.9405656799620298, "learning_rate": 9.954801579891627e-07, "loss": 0.6749, "step": 23774 }, { "epoch": 0.8611010503440782, "grad_norm": 1.2891714165652868, "learning_rate": 9.94969987774619e-07, "loss": 0.5856, "step": 23775 }, { "epoch": 0.8611372691053966, "grad_norm": 1.458089125499939, "learning_rate": 9.944599414772017e-07, "loss": 0.6802, "step": 23776 }, { "epoch": 0.861173487866715, "grad_norm": 1.4506436871792256, "learning_rate": 9.939500191039286e-07, "loss": 0.665, "step": 23777 }, { "epoch": 0.8612097066280333, "grad_norm": 1.3957052511609525, "learning_rate": 9.934402206618176e-07, "loss": 0.6815, "step": 23778 }, { "epoch": 0.8612459253893516, "grad_norm": 1.5011217584416159, "learning_rate": 9.929305461578799e-07, "loss": 0.6339, "step": 23779 }, { "epoch": 0.8612821441506701, "grad_norm": 1.3023986133502392, "learning_rate": 9.924209955991348e-07, "loss": 0.6942, "step": 23780 }, { "epoch": 0.8613183629119884, "grad_norm": 0.9447423182951652, "learning_rate": 9.919115689925918e-07, "loss": 0.6979, "step": 23781 }, { "epoch": 0.8613545816733068, "grad_norm": 1.4841927090191707, "learning_rate": 9.914022663452605e-07, "loss": 0.6621, "step": 23782 }, { "epoch": 0.8613908004346251, "grad_norm": 1.3549523022447687, "learning_rate": 9.908930876641477e-07, "loss": 0.7195, "step": 23783 }, { "epoch": 0.8614270191959434, "grad_norm": 1.3931298587439664, "learning_rate": 9.903840329562642e-07, "loss": 0.7384, "step": 23784 }, { "epoch": 0.8614632379572619, "grad_norm": 1.3979899266931717, "learning_rate": 9.898751022286124e-07, "loss": 0.6869, "step": 23785 }, { "epoch": 0.8614994567185802, "grad_norm": 1.5579453269170256, "learning_rate": 9.893662954881945e-07, "loss": 0.7089, "step": 23786 }, { "epoch": 0.8615356754798986, "grad_norm": 1.164294626196371, "learning_rate": 9.888576127420158e-07, "loss": 0.6657, "step": 23787 }, { "epoch": 0.8615718942412169, "grad_norm": 1.49956001081622, "learning_rate": 9.883490539970741e-07, "loss": 0.7235, "step": 23788 }, { "epoch": 0.8616081130025353, "grad_norm": 1.3769934749778514, "learning_rate": 9.878406192603684e-07, "loss": 0.6099, "step": 23789 }, { "epoch": 0.8616443317638537, "grad_norm": 1.360241254947198, "learning_rate": 9.87332308538892e-07, "loss": 0.656, "step": 23790 }, { "epoch": 0.861680550525172, "grad_norm": 1.417721601364442, "learning_rate": 9.868241218396447e-07, "loss": 0.6657, "step": 23791 }, { "epoch": 0.8617167692864904, "grad_norm": 1.4783671448151676, "learning_rate": 9.863160591696163e-07, "loss": 0.7504, "step": 23792 }, { "epoch": 0.8617529880478088, "grad_norm": 0.9535916957924612, "learning_rate": 9.858081205357993e-07, "loss": 0.6836, "step": 23793 }, { "epoch": 0.8617892068091271, "grad_norm": 1.4053073902813404, "learning_rate": 9.853003059451827e-07, "loss": 0.6483, "step": 23794 }, { "epoch": 0.8618254255704455, "grad_norm": 1.1987943506020107, "learning_rate": 9.84792615404755e-07, "loss": 0.6061, "step": 23795 }, { "epoch": 0.8618616443317638, "grad_norm": 1.4297830192779752, "learning_rate": 9.842850489215028e-07, "loss": 0.6853, "step": 23796 }, { "epoch": 0.8618978630930823, "grad_norm": 1.4211338778886282, "learning_rate": 9.837776065024086e-07, "loss": 0.6596, "step": 23797 }, { "epoch": 0.8619340818544006, "grad_norm": 1.3636673596410376, "learning_rate": 9.832702881544588e-07, "loss": 0.6824, "step": 23798 }, { "epoch": 0.8619703006157189, "grad_norm": 1.4872178874545443, "learning_rate": 9.827630938846323e-07, "loss": 0.6606, "step": 23799 }, { "epoch": 0.8620065193770373, "grad_norm": 1.2405283151565878, "learning_rate": 9.82256023699909e-07, "loss": 0.6667, "step": 23800 }, { "epoch": 0.8620427381383556, "grad_norm": 1.4908352948943597, "learning_rate": 9.817490776072647e-07, "loss": 0.6294, "step": 23801 }, { "epoch": 0.8620789568996741, "grad_norm": 1.5619968879856603, "learning_rate": 9.81242255613679e-07, "loss": 0.6324, "step": 23802 }, { "epoch": 0.8621151756609924, "grad_norm": 0.9902290422803388, "learning_rate": 9.807355577261246e-07, "loss": 0.6526, "step": 23803 }, { "epoch": 0.8621513944223107, "grad_norm": 1.436711190801734, "learning_rate": 9.802289839515743e-07, "loss": 0.6554, "step": 23804 }, { "epoch": 0.8621876131836291, "grad_norm": 1.3789998852112177, "learning_rate": 9.797225342969963e-07, "loss": 0.6657, "step": 23805 }, { "epoch": 0.8622238319449474, "grad_norm": 0.9873196432002493, "learning_rate": 9.792162087693634e-07, "loss": 0.6875, "step": 23806 }, { "epoch": 0.8622600507062659, "grad_norm": 1.239442670996263, "learning_rate": 9.787100073756428e-07, "loss": 0.6583, "step": 23807 }, { "epoch": 0.8622962694675842, "grad_norm": 1.4145090625682641, "learning_rate": 9.782039301227974e-07, "loss": 0.6753, "step": 23808 }, { "epoch": 0.8623324882289025, "grad_norm": 1.3495563270312132, "learning_rate": 9.776979770177942e-07, "loss": 0.6931, "step": 23809 }, { "epoch": 0.862368706990221, "grad_norm": 1.547243983399806, "learning_rate": 9.77192148067595e-07, "loss": 0.6062, "step": 23810 }, { "epoch": 0.8624049257515393, "grad_norm": 1.1620676530533138, "learning_rate": 9.7668644327916e-07, "loss": 0.5997, "step": 23811 }, { "epoch": 0.8624411445128577, "grad_norm": 1.0985966285998776, "learning_rate": 9.76180862659446e-07, "loss": 0.6846, "step": 23812 }, { "epoch": 0.862477363274176, "grad_norm": 1.3035989877590322, "learning_rate": 9.75675406215415e-07, "loss": 0.6694, "step": 23813 }, { "epoch": 0.8625135820354944, "grad_norm": 1.2992892210509415, "learning_rate": 9.751700739540193e-07, "loss": 0.6744, "step": 23814 }, { "epoch": 0.8625498007968128, "grad_norm": 1.4347328648356985, "learning_rate": 9.746648658822133e-07, "loss": 0.7548, "step": 23815 }, { "epoch": 0.8625860195581311, "grad_norm": 1.435266375223101, "learning_rate": 9.74159782006947e-07, "loss": 0.6735, "step": 23816 }, { "epoch": 0.8626222383194495, "grad_norm": 1.2909277844230793, "learning_rate": 9.736548223351749e-07, "loss": 0.6486, "step": 23817 }, { "epoch": 0.8626584570807678, "grad_norm": 1.365385981015622, "learning_rate": 9.731499868738448e-07, "loss": 0.6251, "step": 23818 }, { "epoch": 0.8626946758420863, "grad_norm": 1.3876078253918467, "learning_rate": 9.726452756298988e-07, "loss": 0.6906, "step": 23819 }, { "epoch": 0.8627308946034046, "grad_norm": 0.9861716810290052, "learning_rate": 9.721406886102891e-07, "loss": 0.7073, "step": 23820 }, { "epoch": 0.8627671133647229, "grad_norm": 1.3494038456402906, "learning_rate": 9.716362258219559e-07, "loss": 0.6755, "step": 23821 }, { "epoch": 0.8628033321260413, "grad_norm": 1.0503954343057933, "learning_rate": 9.71131887271841e-07, "loss": 0.7083, "step": 23822 }, { "epoch": 0.8628395508873596, "grad_norm": 1.3892135857898105, "learning_rate": 9.70627672966883e-07, "loss": 0.7355, "step": 23823 }, { "epoch": 0.8628757696486781, "grad_norm": 1.342929824738773, "learning_rate": 9.701235829140244e-07, "loss": 0.6891, "step": 23824 }, { "epoch": 0.8629119884099964, "grad_norm": 1.0411219835801035, "learning_rate": 9.696196171201987e-07, "loss": 0.6166, "step": 23825 }, { "epoch": 0.8629482071713147, "grad_norm": 1.3725945927022187, "learning_rate": 9.691157755923408e-07, "loss": 0.6835, "step": 23826 }, { "epoch": 0.8629844259326331, "grad_norm": 1.4258157621413319, "learning_rate": 9.68612058337387e-07, "loss": 0.688, "step": 23827 }, { "epoch": 0.8630206446939515, "grad_norm": 1.4581494961964763, "learning_rate": 9.681084653622662e-07, "loss": 0.6489, "step": 23828 }, { "epoch": 0.8630568634552699, "grad_norm": 1.3207002982312708, "learning_rate": 9.676049966739098e-07, "loss": 0.681, "step": 23829 }, { "epoch": 0.8630930822165882, "grad_norm": 0.9585741854974595, "learning_rate": 9.67101652279242e-07, "loss": 0.6691, "step": 23830 }, { "epoch": 0.8631293009779065, "grad_norm": 1.068750443814873, "learning_rate": 9.665984321851961e-07, "loss": 0.7032, "step": 23831 }, { "epoch": 0.863165519739225, "grad_norm": 1.279266809881892, "learning_rate": 9.660953363986925e-07, "loss": 0.6278, "step": 23832 }, { "epoch": 0.8632017385005433, "grad_norm": 1.0961827623476794, "learning_rate": 9.65592364926654e-07, "loss": 0.6772, "step": 23833 }, { "epoch": 0.8632379572618617, "grad_norm": 1.4220834739043524, "learning_rate": 9.650895177760023e-07, "loss": 0.6712, "step": 23834 }, { "epoch": 0.86327417602318, "grad_norm": 1.3064705562676242, "learning_rate": 9.645867949536591e-07, "loss": 0.5692, "step": 23835 }, { "epoch": 0.8633103947844983, "grad_norm": 1.4327195559373387, "learning_rate": 9.640841964665403e-07, "loss": 0.6567, "step": 23836 }, { "epoch": 0.8633466135458168, "grad_norm": 1.019597949697635, "learning_rate": 9.635817223215616e-07, "loss": 0.6274, "step": 23837 }, { "epoch": 0.8633828323071351, "grad_norm": 1.348647520826214, "learning_rate": 9.630793725256403e-07, "loss": 0.6697, "step": 23838 }, { "epoch": 0.8634190510684535, "grad_norm": 0.932192034579467, "learning_rate": 9.62577147085687e-07, "loss": 0.6616, "step": 23839 }, { "epoch": 0.8634552698297718, "grad_norm": 1.3968821619264113, "learning_rate": 9.620750460086126e-07, "loss": 0.7429, "step": 23840 }, { "epoch": 0.8634914885910902, "grad_norm": 1.430709374657394, "learning_rate": 9.61573069301326e-07, "loss": 0.6588, "step": 23841 }, { "epoch": 0.8635277073524086, "grad_norm": 1.428826499510245, "learning_rate": 9.61071216970737e-07, "loss": 0.7438, "step": 23842 }, { "epoch": 0.8635639261137269, "grad_norm": 1.030902079233065, "learning_rate": 9.605694890237516e-07, "loss": 0.6839, "step": 23843 }, { "epoch": 0.8636001448750453, "grad_norm": 1.305568650211138, "learning_rate": 9.600678854672719e-07, "loss": 0.652, "step": 23844 }, { "epoch": 0.8636363636363636, "grad_norm": 1.339552687151157, "learning_rate": 9.595664063081989e-07, "loss": 0.597, "step": 23845 }, { "epoch": 0.863672582397682, "grad_norm": 1.246236386401739, "learning_rate": 9.590650515534382e-07, "loss": 0.6023, "step": 23846 }, { "epoch": 0.8637088011590004, "grad_norm": 1.3461507106879746, "learning_rate": 9.58563821209887e-07, "loss": 0.6529, "step": 23847 }, { "epoch": 0.8637450199203187, "grad_norm": 1.2196813518492362, "learning_rate": 9.580627152844391e-07, "loss": 0.6165, "step": 23848 }, { "epoch": 0.8637812386816371, "grad_norm": 1.3704409186260167, "learning_rate": 9.57561733783996e-07, "loss": 0.6706, "step": 23849 }, { "epoch": 0.8638174574429555, "grad_norm": 1.4204525356061761, "learning_rate": 9.570608767154477e-07, "loss": 0.7236, "step": 23850 }, { "epoch": 0.8638536762042738, "grad_norm": 2.063790407157785, "learning_rate": 9.565601440856887e-07, "loss": 0.6745, "step": 23851 }, { "epoch": 0.8638898949655922, "grad_norm": 1.5645591518939401, "learning_rate": 9.560595359016057e-07, "loss": 0.7457, "step": 23852 }, { "epoch": 0.8639261137269105, "grad_norm": 1.3859826753644466, "learning_rate": 9.55559052170093e-07, "loss": 0.6708, "step": 23853 }, { "epoch": 0.863962332488229, "grad_norm": 1.458808790056409, "learning_rate": 9.550586928980343e-07, "loss": 0.6353, "step": 23854 }, { "epoch": 0.8639985512495473, "grad_norm": 0.9609681735507608, "learning_rate": 9.545584580923152e-07, "loss": 0.6419, "step": 23855 }, { "epoch": 0.8640347700108656, "grad_norm": 1.1149561646382584, "learning_rate": 9.540583477598187e-07, "loss": 0.6476, "step": 23856 }, { "epoch": 0.864070988772184, "grad_norm": 1.380755751092583, "learning_rate": 9.535583619074296e-07, "loss": 0.6102, "step": 23857 }, { "epoch": 0.8641072075335023, "grad_norm": 1.32824446126876, "learning_rate": 9.530585005420256e-07, "loss": 0.631, "step": 23858 }, { "epoch": 0.8641434262948208, "grad_norm": 1.4758517507471571, "learning_rate": 9.525587636704847e-07, "loss": 0.6834, "step": 23859 }, { "epoch": 0.8641796450561391, "grad_norm": 1.020235056309252, "learning_rate": 9.520591512996858e-07, "loss": 0.627, "step": 23860 }, { "epoch": 0.8642158638174574, "grad_norm": 1.504446548161887, "learning_rate": 9.515596634365043e-07, "loss": 0.6939, "step": 23861 }, { "epoch": 0.8642520825787758, "grad_norm": 0.9407418355099922, "learning_rate": 9.510603000878127e-07, "loss": 0.6577, "step": 23862 }, { "epoch": 0.8642883013400942, "grad_norm": 1.2355800764101568, "learning_rate": 9.505610612604798e-07, "loss": 0.627, "step": 23863 }, { "epoch": 0.8643245201014126, "grad_norm": 1.4733408947825404, "learning_rate": 9.500619469613803e-07, "loss": 0.7454, "step": 23864 }, { "epoch": 0.8643607388627309, "grad_norm": 0.9730610851173285, "learning_rate": 9.495629571973808e-07, "loss": 0.6295, "step": 23865 }, { "epoch": 0.8643969576240492, "grad_norm": 1.3935325102187572, "learning_rate": 9.490640919753469e-07, "loss": 0.649, "step": 23866 }, { "epoch": 0.8644331763853677, "grad_norm": 1.5808823982317366, "learning_rate": 9.48565351302142e-07, "loss": 0.6846, "step": 23867 }, { "epoch": 0.864469395146686, "grad_norm": 1.3385466008166522, "learning_rate": 9.48066735184634e-07, "loss": 0.6028, "step": 23868 }, { "epoch": 0.8645056139080044, "grad_norm": 1.2673721440007752, "learning_rate": 9.475682436296807e-07, "loss": 0.663, "step": 23869 }, { "epoch": 0.8645418326693227, "grad_norm": 1.000021151217928, "learning_rate": 9.4706987664414e-07, "loss": 0.6435, "step": 23870 }, { "epoch": 0.864578051430641, "grad_norm": 1.4611495216045505, "learning_rate": 9.465716342348752e-07, "loss": 0.7405, "step": 23871 }, { "epoch": 0.8646142701919595, "grad_norm": 1.1419167554622829, "learning_rate": 9.460735164087398e-07, "loss": 0.7001, "step": 23872 }, { "epoch": 0.8646504889532778, "grad_norm": 1.415972652440954, "learning_rate": 9.455755231725883e-07, "loss": 0.6674, "step": 23873 }, { "epoch": 0.8646867077145962, "grad_norm": 1.3433306838219279, "learning_rate": 9.450776545332707e-07, "loss": 0.6127, "step": 23874 }, { "epoch": 0.8647229264759145, "grad_norm": 1.3457398139702588, "learning_rate": 9.445799104976438e-07, "loss": 0.6922, "step": 23875 }, { "epoch": 0.8647591452372329, "grad_norm": 1.3421741285646127, "learning_rate": 9.440822910725534e-07, "loss": 0.6718, "step": 23876 }, { "epoch": 0.8647953639985513, "grad_norm": 1.3267827403830645, "learning_rate": 9.435847962648493e-07, "loss": 0.6735, "step": 23877 }, { "epoch": 0.8648315827598696, "grad_norm": 1.3489140730708546, "learning_rate": 9.430874260813727e-07, "loss": 0.6796, "step": 23878 }, { "epoch": 0.864867801521188, "grad_norm": 1.004361747291036, "learning_rate": 9.425901805289739e-07, "loss": 0.7089, "step": 23879 }, { "epoch": 0.8649040202825063, "grad_norm": 1.495576292818835, "learning_rate": 9.420930596144917e-07, "loss": 0.7158, "step": 23880 }, { "epoch": 0.8649402390438247, "grad_norm": 1.000986084427553, "learning_rate": 9.415960633447674e-07, "loss": 0.6837, "step": 23881 }, { "epoch": 0.8649764578051431, "grad_norm": 1.8318534471980803, "learning_rate": 9.410991917266421e-07, "loss": 0.6869, "step": 23882 }, { "epoch": 0.8650126765664614, "grad_norm": 1.2895466718857391, "learning_rate": 9.406024447669515e-07, "loss": 0.6276, "step": 23883 }, { "epoch": 0.8650488953277798, "grad_norm": 0.9665985021308608, "learning_rate": 9.401058224725313e-07, "loss": 0.6523, "step": 23884 }, { "epoch": 0.8650851140890982, "grad_norm": 1.5258756811489356, "learning_rate": 9.396093248502147e-07, "loss": 0.6261, "step": 23885 }, { "epoch": 0.8651213328504165, "grad_norm": 1.1048552149350728, "learning_rate": 9.391129519068364e-07, "loss": 0.6441, "step": 23886 }, { "epoch": 0.8651575516117349, "grad_norm": 1.362163911060818, "learning_rate": 9.386167036492255e-07, "loss": 0.6786, "step": 23887 }, { "epoch": 0.8651937703730532, "grad_norm": 1.414957647699695, "learning_rate": 9.381205800842086e-07, "loss": 0.7426, "step": 23888 }, { "epoch": 0.8652299891343717, "grad_norm": 1.4501821799919887, "learning_rate": 9.376245812186169e-07, "loss": 0.7263, "step": 23889 }, { "epoch": 0.86526620789569, "grad_norm": 1.3871533411331476, "learning_rate": 9.371287070592738e-07, "loss": 0.6495, "step": 23890 }, { "epoch": 0.8653024266570083, "grad_norm": 1.2767573813113378, "learning_rate": 9.36632957613004e-07, "loss": 0.6041, "step": 23891 }, { "epoch": 0.8653386454183267, "grad_norm": 1.011852712264507, "learning_rate": 9.361373328866252e-07, "loss": 0.7472, "step": 23892 }, { "epoch": 0.865374864179645, "grad_norm": 1.0509963249491818, "learning_rate": 9.356418328869643e-07, "loss": 0.7245, "step": 23893 }, { "epoch": 0.8654110829409635, "grad_norm": 1.3962466757889744, "learning_rate": 9.351464576208346e-07, "loss": 0.7322, "step": 23894 }, { "epoch": 0.8654473017022818, "grad_norm": 1.3922732957205146, "learning_rate": 9.346512070950564e-07, "loss": 0.6071, "step": 23895 }, { "epoch": 0.8654835204636001, "grad_norm": 1.366628538718414, "learning_rate": 9.341560813164407e-07, "loss": 0.7017, "step": 23896 }, { "epoch": 0.8655197392249185, "grad_norm": 1.454571005118837, "learning_rate": 9.336610802918044e-07, "loss": 0.6428, "step": 23897 }, { "epoch": 0.8655559579862369, "grad_norm": 1.477297215987142, "learning_rate": 9.331662040279588e-07, "loss": 0.7553, "step": 23898 }, { "epoch": 0.8655921767475553, "grad_norm": 1.3103660012400793, "learning_rate": 9.326714525317104e-07, "loss": 0.6715, "step": 23899 }, { "epoch": 0.8656283955088736, "grad_norm": 1.0416596250867265, "learning_rate": 9.321768258098729e-07, "loss": 0.7137, "step": 23900 }, { "epoch": 0.8656646142701919, "grad_norm": 1.958930474069384, "learning_rate": 9.316823238692496e-07, "loss": 0.7412, "step": 23901 }, { "epoch": 0.8657008330315104, "grad_norm": 1.3535905549094398, "learning_rate": 9.311879467166463e-07, "loss": 0.6549, "step": 23902 }, { "epoch": 0.8657370517928287, "grad_norm": 1.3182230907217516, "learning_rate": 9.306936943588629e-07, "loss": 0.673, "step": 23903 }, { "epoch": 0.8657732705541471, "grad_norm": 1.4127439857588038, "learning_rate": 9.301995668027041e-07, "loss": 0.6469, "step": 23904 }, { "epoch": 0.8658094893154654, "grad_norm": 1.1352928643053732, "learning_rate": 9.297055640549701e-07, "loss": 0.6908, "step": 23905 }, { "epoch": 0.8658457080767837, "grad_norm": 1.4077017429892036, "learning_rate": 9.292116861224565e-07, "loss": 0.6858, "step": 23906 }, { "epoch": 0.8658819268381022, "grad_norm": 1.3633985059094562, "learning_rate": 9.2871793301196e-07, "loss": 0.6852, "step": 23907 }, { "epoch": 0.8659181455994205, "grad_norm": 1.3089155833348836, "learning_rate": 9.282243047302764e-07, "loss": 0.6502, "step": 23908 }, { "epoch": 0.8659543643607389, "grad_norm": 1.3885991041213959, "learning_rate": 9.277308012841968e-07, "loss": 0.6965, "step": 23909 }, { "epoch": 0.8659905831220572, "grad_norm": 1.3127625612907194, "learning_rate": 9.272374226805115e-07, "loss": 0.6168, "step": 23910 }, { "epoch": 0.8660268018833756, "grad_norm": 1.2594799479293965, "learning_rate": 9.267441689260137e-07, "loss": 0.6857, "step": 23911 }, { "epoch": 0.866063020644694, "grad_norm": 1.1789704276590738, "learning_rate": 9.262510400274882e-07, "loss": 0.5842, "step": 23912 }, { "epoch": 0.8660992394060123, "grad_norm": 1.4755895413872913, "learning_rate": 9.257580359917218e-07, "loss": 0.7998, "step": 23913 }, { "epoch": 0.8661354581673307, "grad_norm": 2.7077702659861647, "learning_rate": 9.252651568254955e-07, "loss": 0.6969, "step": 23914 }, { "epoch": 0.866171676928649, "grad_norm": 1.3687777807081536, "learning_rate": 9.247724025355964e-07, "loss": 0.6659, "step": 23915 }, { "epoch": 0.8662078956899674, "grad_norm": 1.421916527456078, "learning_rate": 9.242797731288045e-07, "loss": 0.7278, "step": 23916 }, { "epoch": 0.8662441144512858, "grad_norm": 1.3983582595400041, "learning_rate": 9.237872686118965e-07, "loss": 0.7491, "step": 23917 }, { "epoch": 0.8662803332126041, "grad_norm": 1.4558450636225542, "learning_rate": 9.232948889916483e-07, "loss": 0.739, "step": 23918 }, { "epoch": 0.8663165519739225, "grad_norm": 1.1201562051537604, "learning_rate": 9.228026342748408e-07, "loss": 0.6793, "step": 23919 }, { "epoch": 0.8663527707352409, "grad_norm": 1.4328731511028476, "learning_rate": 9.223105044682457e-07, "loss": 0.661, "step": 23920 }, { "epoch": 0.8663889894965592, "grad_norm": 1.4822181327943038, "learning_rate": 9.218184995786317e-07, "loss": 0.6383, "step": 23921 }, { "epoch": 0.8664252082578776, "grad_norm": 1.0534968892625347, "learning_rate": 9.213266196127735e-07, "loss": 0.6398, "step": 23922 }, { "epoch": 0.8664614270191959, "grad_norm": 1.257927737849086, "learning_rate": 9.20834864577439e-07, "loss": 0.667, "step": 23923 }, { "epoch": 0.8664976457805144, "grad_norm": 0.9241396030032, "learning_rate": 9.203432344793917e-07, "loss": 0.634, "step": 23924 }, { "epoch": 0.8665338645418327, "grad_norm": 1.1526792766664549, "learning_rate": 9.198517293254006e-07, "loss": 0.6367, "step": 23925 }, { "epoch": 0.866570083303151, "grad_norm": 1.3076801163827723, "learning_rate": 9.193603491222292e-07, "loss": 0.6495, "step": 23926 }, { "epoch": 0.8666063020644694, "grad_norm": 1.5802345198479733, "learning_rate": 9.188690938766387e-07, "loss": 0.5984, "step": 23927 }, { "epoch": 0.8666425208257877, "grad_norm": 1.5249832857400278, "learning_rate": 9.183779635953872e-07, "loss": 0.6639, "step": 23928 }, { "epoch": 0.8666787395871062, "grad_norm": 1.247593473853864, "learning_rate": 9.178869582852356e-07, "loss": 0.6781, "step": 23929 }, { "epoch": 0.8667149583484245, "grad_norm": 1.1611026285962953, "learning_rate": 9.173960779529401e-07, "loss": 0.5861, "step": 23930 }, { "epoch": 0.8667511771097428, "grad_norm": 1.2361729896118545, "learning_rate": 9.169053226052537e-07, "loss": 0.639, "step": 23931 }, { "epoch": 0.8667873958710612, "grad_norm": 1.3804662372790764, "learning_rate": 9.164146922489326e-07, "loss": 0.7163, "step": 23932 }, { "epoch": 0.8668236146323796, "grad_norm": 1.463299370769067, "learning_rate": 9.159241868907265e-07, "loss": 0.6688, "step": 23933 }, { "epoch": 0.866859833393698, "grad_norm": 1.3960970788370588, "learning_rate": 9.154338065373857e-07, "loss": 0.679, "step": 23934 }, { "epoch": 0.8668960521550163, "grad_norm": 1.3688457024443244, "learning_rate": 9.149435511956572e-07, "loss": 0.6638, "step": 23935 }, { "epoch": 0.8669322709163346, "grad_norm": 1.5377682233580503, "learning_rate": 9.1445342087229e-07, "loss": 0.6923, "step": 23936 }, { "epoch": 0.866968489677653, "grad_norm": 1.1445542682482883, "learning_rate": 9.139634155740262e-07, "loss": 0.6959, "step": 23937 }, { "epoch": 0.8670047084389714, "grad_norm": 1.3108575483955431, "learning_rate": 9.134735353076096e-07, "loss": 0.6909, "step": 23938 }, { "epoch": 0.8670409272002898, "grad_norm": 1.1106242167614846, "learning_rate": 9.129837800797792e-07, "loss": 0.6423, "step": 23939 }, { "epoch": 0.8670771459616081, "grad_norm": 1.355418995870287, "learning_rate": 9.124941498972795e-07, "loss": 0.6035, "step": 23940 }, { "epoch": 0.8671133647229264, "grad_norm": 1.743935708613252, "learning_rate": 9.12004644766844e-07, "loss": 0.5963, "step": 23941 }, { "epoch": 0.8671495834842449, "grad_norm": 1.3292643491513705, "learning_rate": 9.115152646952086e-07, "loss": 0.7033, "step": 23942 }, { "epoch": 0.8671858022455632, "grad_norm": 1.3741734571241662, "learning_rate": 9.11026009689111e-07, "loss": 0.6702, "step": 23943 }, { "epoch": 0.8672220210068816, "grad_norm": 1.3154432548100818, "learning_rate": 9.105368797552827e-07, "loss": 0.6652, "step": 23944 }, { "epoch": 0.8672582397681999, "grad_norm": 1.1019558104532556, "learning_rate": 9.100478749004527e-07, "loss": 0.6377, "step": 23945 }, { "epoch": 0.8672944585295183, "grad_norm": 1.3459812757158953, "learning_rate": 9.095589951313488e-07, "loss": 0.6899, "step": 23946 }, { "epoch": 0.8673306772908367, "grad_norm": 1.4897442457899688, "learning_rate": 9.090702404547025e-07, "loss": 0.6942, "step": 23947 }, { "epoch": 0.867366896052155, "grad_norm": 1.3707046941724847, "learning_rate": 9.085816108772383e-07, "loss": 0.6746, "step": 23948 }, { "epoch": 0.8674031148134734, "grad_norm": 1.3818570273579367, "learning_rate": 9.080931064056797e-07, "loss": 0.6766, "step": 23949 }, { "epoch": 0.8674393335747917, "grad_norm": 1.3183098119673409, "learning_rate": 9.07604727046747e-07, "loss": 0.6787, "step": 23950 }, { "epoch": 0.8674755523361101, "grad_norm": 1.3418485750902154, "learning_rate": 9.071164728071635e-07, "loss": 0.6721, "step": 23951 }, { "epoch": 0.8675117710974285, "grad_norm": 1.3300399557734293, "learning_rate": 9.066283436936474e-07, "loss": 0.6832, "step": 23952 }, { "epoch": 0.8675479898587468, "grad_norm": 1.2962302363148022, "learning_rate": 9.061403397129143e-07, "loss": 0.6494, "step": 23953 }, { "epoch": 0.8675842086200652, "grad_norm": 1.2343208605363305, "learning_rate": 9.056524608716821e-07, "loss": 0.646, "step": 23954 }, { "epoch": 0.8676204273813836, "grad_norm": 1.3365818235027516, "learning_rate": 9.051647071766623e-07, "loss": 0.6112, "step": 23955 }, { "epoch": 0.8676566461427019, "grad_norm": 1.3506377645328038, "learning_rate": 9.046770786345682e-07, "loss": 0.6138, "step": 23956 }, { "epoch": 0.8676928649040203, "grad_norm": 1.356676515905737, "learning_rate": 9.04189575252108e-07, "loss": 0.6209, "step": 23957 }, { "epoch": 0.8677290836653386, "grad_norm": 1.3433079727179802, "learning_rate": 9.037021970359927e-07, "loss": 0.6028, "step": 23958 }, { "epoch": 0.8677653024266571, "grad_norm": 1.380054117005273, "learning_rate": 9.032149439929283e-07, "loss": 0.6713, "step": 23959 }, { "epoch": 0.8678015211879754, "grad_norm": 1.406644646581374, "learning_rate": 9.027278161296194e-07, "loss": 0.6915, "step": 23960 }, { "epoch": 0.8678377399492937, "grad_norm": 1.4204699115897386, "learning_rate": 9.022408134527671e-07, "loss": 0.6284, "step": 23961 }, { "epoch": 0.8678739587106121, "grad_norm": 1.3416403196031612, "learning_rate": 9.017539359690764e-07, "loss": 0.6568, "step": 23962 }, { "epoch": 0.8679101774719304, "grad_norm": 1.4270481229415501, "learning_rate": 9.012671836852471e-07, "loss": 0.6011, "step": 23963 }, { "epoch": 0.8679463962332489, "grad_norm": 1.4736989810970318, "learning_rate": 9.007805566079741e-07, "loss": 0.6709, "step": 23964 }, { "epoch": 0.8679826149945672, "grad_norm": 1.9021633775490379, "learning_rate": 9.002940547439576e-07, "loss": 0.6875, "step": 23965 }, { "epoch": 0.8680188337558855, "grad_norm": 1.419341935701138, "learning_rate": 8.99807678099891e-07, "loss": 0.6418, "step": 23966 }, { "epoch": 0.8680550525172039, "grad_norm": 1.2811168479349277, "learning_rate": 8.993214266824657e-07, "loss": 0.6977, "step": 23967 }, { "epoch": 0.8680912712785223, "grad_norm": 1.332619951573211, "learning_rate": 8.98835300498373e-07, "loss": 0.7468, "step": 23968 }, { "epoch": 0.8681274900398407, "grad_norm": 1.4058032736908603, "learning_rate": 8.983492995543063e-07, "loss": 0.6351, "step": 23969 }, { "epoch": 0.868163708801159, "grad_norm": 1.3015833435129829, "learning_rate": 8.978634238569495e-07, "loss": 0.5509, "step": 23970 }, { "epoch": 0.8681999275624773, "grad_norm": 0.980851405717427, "learning_rate": 8.973776734129902e-07, "loss": 0.67, "step": 23971 }, { "epoch": 0.8682361463237958, "grad_norm": 1.326957292384582, "learning_rate": 8.968920482291099e-07, "loss": 0.641, "step": 23972 }, { "epoch": 0.8682723650851141, "grad_norm": 1.1643448775078575, "learning_rate": 8.964065483119966e-07, "loss": 0.6666, "step": 23973 }, { "epoch": 0.8683085838464325, "grad_norm": 1.393187130237431, "learning_rate": 8.959211736683282e-07, "loss": 0.7179, "step": 23974 }, { "epoch": 0.8683448026077508, "grad_norm": 1.5458486359816492, "learning_rate": 8.954359243047816e-07, "loss": 0.6578, "step": 23975 }, { "epoch": 0.8683810213690691, "grad_norm": 1.4376136149206127, "learning_rate": 8.949508002280383e-07, "loss": 0.6721, "step": 23976 }, { "epoch": 0.8684172401303876, "grad_norm": 1.451908557373077, "learning_rate": 8.944658014447727e-07, "loss": 0.6879, "step": 23977 }, { "epoch": 0.8684534588917059, "grad_norm": 1.0669271783558432, "learning_rate": 8.939809279616596e-07, "loss": 0.6432, "step": 23978 }, { "epoch": 0.8684896776530243, "grad_norm": 1.3887471468359243, "learning_rate": 8.934961797853669e-07, "loss": 0.6936, "step": 23979 }, { "epoch": 0.8685258964143426, "grad_norm": 1.5136503254986962, "learning_rate": 8.930115569225706e-07, "loss": 0.5987, "step": 23980 }, { "epoch": 0.868562115175661, "grad_norm": 0.9530988047116632, "learning_rate": 8.925270593799373e-07, "loss": 0.6533, "step": 23981 }, { "epoch": 0.8685983339369794, "grad_norm": 1.545581382679001, "learning_rate": 8.92042687164133e-07, "loss": 0.729, "step": 23982 }, { "epoch": 0.8686345526982977, "grad_norm": 1.3662049168637445, "learning_rate": 8.915584402818256e-07, "loss": 0.6536, "step": 23983 }, { "epoch": 0.8686707714596161, "grad_norm": 1.0623664824657921, "learning_rate": 8.910743187396775e-07, "loss": 0.6342, "step": 23984 }, { "epoch": 0.8687069902209344, "grad_norm": 1.4197791682421137, "learning_rate": 8.905903225443502e-07, "loss": 0.6672, "step": 23985 }, { "epoch": 0.8687432089822528, "grad_norm": 1.5285228430521258, "learning_rate": 8.901064517025026e-07, "loss": 0.7604, "step": 23986 }, { "epoch": 0.8687794277435712, "grad_norm": 1.4158039434652083, "learning_rate": 8.896227062207974e-07, "loss": 0.693, "step": 23987 }, { "epoch": 0.8688156465048895, "grad_norm": 1.492871532139427, "learning_rate": 8.891390861058879e-07, "loss": 0.6762, "step": 23988 }, { "epoch": 0.868851865266208, "grad_norm": 1.3876493796706932, "learning_rate": 8.886555913644313e-07, "loss": 0.624, "step": 23989 }, { "epoch": 0.8688880840275263, "grad_norm": 1.315867184170364, "learning_rate": 8.881722220030764e-07, "loss": 0.649, "step": 23990 }, { "epoch": 0.8689243027888446, "grad_norm": 1.0132445516977486, "learning_rate": 8.876889780284814e-07, "loss": 0.6507, "step": 23991 }, { "epoch": 0.868960521550163, "grad_norm": 1.1100062738345282, "learning_rate": 8.872058594472921e-07, "loss": 0.6812, "step": 23992 }, { "epoch": 0.8689967403114813, "grad_norm": 1.4794903418400398, "learning_rate": 8.867228662661553e-07, "loss": 0.6633, "step": 23993 }, { "epoch": 0.8690329590727998, "grad_norm": 0.9357469996153132, "learning_rate": 8.862399984917214e-07, "loss": 0.6378, "step": 23994 }, { "epoch": 0.8690691778341181, "grad_norm": 1.0464498321577091, "learning_rate": 8.857572561306327e-07, "loss": 0.6718, "step": 23995 }, { "epoch": 0.8691053965954364, "grad_norm": 1.0347412689527047, "learning_rate": 8.852746391895339e-07, "loss": 0.6921, "step": 23996 }, { "epoch": 0.8691416153567548, "grad_norm": 1.3765725245581644, "learning_rate": 8.84792147675062e-07, "loss": 0.6794, "step": 23997 }, { "epoch": 0.8691778341180731, "grad_norm": 1.3869481481283301, "learning_rate": 8.843097815938617e-07, "loss": 0.6721, "step": 23998 }, { "epoch": 0.8692140528793916, "grad_norm": 1.3540305228906766, "learning_rate": 8.838275409525688e-07, "loss": 0.6324, "step": 23999 }, { "epoch": 0.8692502716407099, "grad_norm": 0.9380110670129055, "learning_rate": 8.833454257578189e-07, "loss": 0.6288, "step": 24000 }, { "epoch": 0.8692864904020282, "grad_norm": 1.3790193746377417, "learning_rate": 8.828634360162447e-07, "loss": 0.6449, "step": 24001 }, { "epoch": 0.8693227091633466, "grad_norm": 0.9072281749678162, "learning_rate": 8.82381571734483e-07, "loss": 0.6639, "step": 24002 }, { "epoch": 0.869358927924665, "grad_norm": 1.4611828694860252, "learning_rate": 8.818998329191619e-07, "loss": 0.6379, "step": 24003 }, { "epoch": 0.8693951466859834, "grad_norm": 1.397862784471355, "learning_rate": 8.814182195769083e-07, "loss": 0.6575, "step": 24004 }, { "epoch": 0.8694313654473017, "grad_norm": 1.1056091954748186, "learning_rate": 8.809367317143558e-07, "loss": 0.6544, "step": 24005 }, { "epoch": 0.86946758420862, "grad_norm": 1.1291648137743697, "learning_rate": 8.804553693381257e-07, "loss": 0.7085, "step": 24006 }, { "epoch": 0.8695038029699385, "grad_norm": 1.4933347679294096, "learning_rate": 8.799741324548427e-07, "loss": 0.7015, "step": 24007 }, { "epoch": 0.8695400217312568, "grad_norm": 1.3562922065184024, "learning_rate": 8.794930210711273e-07, "loss": 0.6765, "step": 24008 }, { "epoch": 0.8695762404925752, "grad_norm": 1.5418869590247175, "learning_rate": 8.790120351936038e-07, "loss": 0.7141, "step": 24009 }, { "epoch": 0.8696124592538935, "grad_norm": 1.379748710465792, "learning_rate": 8.785311748288894e-07, "loss": 0.6729, "step": 24010 }, { "epoch": 0.8696486780152118, "grad_norm": 0.9070439886719579, "learning_rate": 8.780504399835999e-07, "loss": 0.6569, "step": 24011 }, { "epoch": 0.8696848967765303, "grad_norm": 1.0423453686229593, "learning_rate": 8.775698306643509e-07, "loss": 0.6637, "step": 24012 }, { "epoch": 0.8697211155378486, "grad_norm": 0.9460050181617855, "learning_rate": 8.770893468777586e-07, "loss": 0.6558, "step": 24013 }, { "epoch": 0.869757334299167, "grad_norm": 1.3930230165684485, "learning_rate": 8.766089886304319e-07, "loss": 0.7078, "step": 24014 }, { "epoch": 0.8697935530604853, "grad_norm": 1.452442637068361, "learning_rate": 8.76128755928981e-07, "loss": 0.7076, "step": 24015 }, { "epoch": 0.8698297718218037, "grad_norm": 1.2138112907705643, "learning_rate": 8.756486487800175e-07, "loss": 0.6059, "step": 24016 }, { "epoch": 0.8698659905831221, "grad_norm": 1.333005758430659, "learning_rate": 8.75168667190145e-07, "loss": 0.7672, "step": 24017 }, { "epoch": 0.8699022093444404, "grad_norm": 1.6090736153256278, "learning_rate": 8.746888111659701e-07, "loss": 0.7021, "step": 24018 }, { "epoch": 0.8699384281057588, "grad_norm": 1.2718132980889603, "learning_rate": 8.742090807140935e-07, "loss": 0.6507, "step": 24019 }, { "epoch": 0.8699746468670772, "grad_norm": 1.4664575334095966, "learning_rate": 8.737294758411197e-07, "loss": 0.6134, "step": 24020 }, { "epoch": 0.8700108656283955, "grad_norm": 1.2203167273280913, "learning_rate": 8.732499965536478e-07, "loss": 0.6192, "step": 24021 }, { "epoch": 0.8700470843897139, "grad_norm": 1.3118241448970354, "learning_rate": 8.727706428582761e-07, "loss": 0.6618, "step": 24022 }, { "epoch": 0.8700833031510322, "grad_norm": 1.0487191720399032, "learning_rate": 8.722914147615968e-07, "loss": 0.6579, "step": 24023 }, { "epoch": 0.8701195219123506, "grad_norm": 1.4204053868869035, "learning_rate": 8.718123122702116e-07, "loss": 0.6104, "step": 24024 }, { "epoch": 0.870155740673669, "grad_norm": 1.3075951246034156, "learning_rate": 8.713333353907083e-07, "loss": 0.7194, "step": 24025 }, { "epoch": 0.8701919594349873, "grad_norm": 1.2939294691735317, "learning_rate": 8.708544841296785e-07, "loss": 0.6272, "step": 24026 }, { "epoch": 0.8702281781963057, "grad_norm": 0.9709648145006274, "learning_rate": 8.703757584937144e-07, "loss": 0.6852, "step": 24027 }, { "epoch": 0.870264396957624, "grad_norm": 1.1533679373438408, "learning_rate": 8.698971584894011e-07, "loss": 0.6164, "step": 24028 }, { "epoch": 0.8703006157189425, "grad_norm": 1.418582171436885, "learning_rate": 8.694186841233255e-07, "loss": 0.71, "step": 24029 }, { "epoch": 0.8703368344802608, "grad_norm": 1.279058949717001, "learning_rate": 8.689403354020709e-07, "loss": 0.6633, "step": 24030 }, { "epoch": 0.8703730532415791, "grad_norm": 1.4482170380603485, "learning_rate": 8.684621123322213e-07, "loss": 0.7203, "step": 24031 }, { "epoch": 0.8704092720028975, "grad_norm": 1.0322792275837749, "learning_rate": 8.679840149203567e-07, "loss": 0.7492, "step": 24032 }, { "epoch": 0.8704454907642158, "grad_norm": 1.2265544918335698, "learning_rate": 8.675060431730564e-07, "loss": 0.6047, "step": 24033 }, { "epoch": 0.8704817095255343, "grad_norm": 1.2520225054876966, "learning_rate": 8.670281970968952e-07, "loss": 0.5698, "step": 24034 }, { "epoch": 0.8705179282868526, "grad_norm": 1.451525295368996, "learning_rate": 8.665504766984533e-07, "loss": 0.6471, "step": 24035 }, { "epoch": 0.8705541470481709, "grad_norm": 1.0238604948293828, "learning_rate": 8.660728819843023e-07, "loss": 0.7047, "step": 24036 }, { "epoch": 0.8705903658094893, "grad_norm": 1.494113221260039, "learning_rate": 8.655954129610122e-07, "loss": 0.7695, "step": 24037 }, { "epoch": 0.8706265845708077, "grad_norm": 1.0607899752525083, "learning_rate": 8.651180696351569e-07, "loss": 0.7053, "step": 24038 }, { "epoch": 0.8706628033321261, "grad_norm": 1.4774197229572734, "learning_rate": 8.646408520133043e-07, "loss": 0.6285, "step": 24039 }, { "epoch": 0.8706990220934444, "grad_norm": 1.1691301093391724, "learning_rate": 8.641637601020203e-07, "loss": 0.7628, "step": 24040 }, { "epoch": 0.8707352408547627, "grad_norm": 1.4689835612528062, "learning_rate": 8.636867939078686e-07, "loss": 0.6436, "step": 24041 }, { "epoch": 0.8707714596160812, "grad_norm": 0.9194515763872699, "learning_rate": 8.632099534374161e-07, "loss": 0.6628, "step": 24042 }, { "epoch": 0.8708076783773995, "grad_norm": 1.0259210898679063, "learning_rate": 8.627332386972231e-07, "loss": 0.5949, "step": 24043 }, { "epoch": 0.8708438971387179, "grad_norm": 1.2855782830439129, "learning_rate": 8.622566496938467e-07, "loss": 0.6913, "step": 24044 }, { "epoch": 0.8708801159000362, "grad_norm": 1.4798905626285124, "learning_rate": 8.617801864338515e-07, "loss": 0.6854, "step": 24045 }, { "epoch": 0.8709163346613545, "grad_norm": 1.3671905374493236, "learning_rate": 8.613038489237891e-07, "loss": 0.7043, "step": 24046 }, { "epoch": 0.870952553422673, "grad_norm": 1.2975407006379458, "learning_rate": 8.608276371702162e-07, "loss": 0.6576, "step": 24047 }, { "epoch": 0.8709887721839913, "grad_norm": 1.5164608039555016, "learning_rate": 8.603515511796834e-07, "loss": 0.6781, "step": 24048 }, { "epoch": 0.8710249909453097, "grad_norm": 0.9194054982931574, "learning_rate": 8.598755909587464e-07, "loss": 0.6204, "step": 24049 }, { "epoch": 0.871061209706628, "grad_norm": 1.2822998916390338, "learning_rate": 8.593997565139522e-07, "loss": 0.6912, "step": 24050 }, { "epoch": 0.8710974284679464, "grad_norm": 1.4308420168245275, "learning_rate": 8.589240478518479e-07, "loss": 0.6292, "step": 24051 }, { "epoch": 0.8711336472292648, "grad_norm": 1.3649061801761995, "learning_rate": 8.584484649789804e-07, "loss": 0.6732, "step": 24052 }, { "epoch": 0.8711698659905831, "grad_norm": 1.044325803038167, "learning_rate": 8.579730079018955e-07, "loss": 0.6679, "step": 24053 }, { "epoch": 0.8712060847519015, "grad_norm": 1.3970255196738297, "learning_rate": 8.574976766271359e-07, "loss": 0.6216, "step": 24054 }, { "epoch": 0.8712423035132199, "grad_norm": 1.3311550784018216, "learning_rate": 8.570224711612385e-07, "loss": 0.7576, "step": 24055 }, { "epoch": 0.8712785222745382, "grad_norm": 1.2794532163470538, "learning_rate": 8.565473915107481e-07, "loss": 0.6254, "step": 24056 }, { "epoch": 0.8713147410358566, "grad_norm": 1.3508193551497505, "learning_rate": 8.560724376821994e-07, "loss": 0.5943, "step": 24057 }, { "epoch": 0.8713509597971749, "grad_norm": 1.3466645676315243, "learning_rate": 8.555976096821294e-07, "loss": 0.6349, "step": 24058 }, { "epoch": 0.8713871785584933, "grad_norm": 1.1742985716117773, "learning_rate": 8.551229075170686e-07, "loss": 0.6533, "step": 24059 }, { "epoch": 0.8714233973198117, "grad_norm": 1.4212747947162299, "learning_rate": 8.546483311935538e-07, "loss": 0.6892, "step": 24060 }, { "epoch": 0.87145961608113, "grad_norm": 0.9510180341084785, "learning_rate": 8.541738807181143e-07, "loss": 0.6759, "step": 24061 }, { "epoch": 0.8714958348424484, "grad_norm": 1.2914674836703446, "learning_rate": 8.536995560972783e-07, "loss": 0.6509, "step": 24062 }, { "epoch": 0.8715320536037667, "grad_norm": 1.0113303840215717, "learning_rate": 8.532253573375704e-07, "loss": 0.6859, "step": 24063 }, { "epoch": 0.8715682723650852, "grad_norm": 1.2732929048181518, "learning_rate": 8.52751284445521e-07, "loss": 0.581, "step": 24064 }, { "epoch": 0.8716044911264035, "grad_norm": 1.2001419569122012, "learning_rate": 8.522773374276516e-07, "loss": 0.6721, "step": 24065 }, { "epoch": 0.8716407098877218, "grad_norm": 1.4846736456951877, "learning_rate": 8.518035162904825e-07, "loss": 0.6803, "step": 24066 }, { "epoch": 0.8716769286490402, "grad_norm": 1.414776436106499, "learning_rate": 8.513298210405363e-07, "loss": 0.647, "step": 24067 }, { "epoch": 0.8717131474103585, "grad_norm": 0.9076033088480496, "learning_rate": 8.50856251684331e-07, "loss": 0.6077, "step": 24068 }, { "epoch": 0.871749366171677, "grad_norm": 1.0283618337969267, "learning_rate": 8.503828082283827e-07, "loss": 0.6474, "step": 24069 }, { "epoch": 0.8717855849329953, "grad_norm": 1.2802473104853167, "learning_rate": 8.499094906792049e-07, "loss": 0.5936, "step": 24070 }, { "epoch": 0.8718218036943136, "grad_norm": 1.5537296716026965, "learning_rate": 8.494362990433136e-07, "loss": 0.6949, "step": 24071 }, { "epoch": 0.871858022455632, "grad_norm": 1.3370467636311276, "learning_rate": 8.489632333272202e-07, "loss": 0.6712, "step": 24072 }, { "epoch": 0.8718942412169504, "grad_norm": 1.4957418812343024, "learning_rate": 8.484902935374328e-07, "loss": 0.6532, "step": 24073 }, { "epoch": 0.8719304599782688, "grad_norm": 1.377596264315032, "learning_rate": 8.480174796804586e-07, "loss": 0.6731, "step": 24074 }, { "epoch": 0.8719666787395871, "grad_norm": 1.4738071451025954, "learning_rate": 8.475447917628065e-07, "loss": 0.7043, "step": 24075 }, { "epoch": 0.8720028975009054, "grad_norm": 1.417875557456369, "learning_rate": 8.470722297909817e-07, "loss": 0.606, "step": 24076 }, { "epoch": 0.8720391162622239, "grad_norm": 1.17325994034569, "learning_rate": 8.46599793771482e-07, "loss": 0.6065, "step": 24077 }, { "epoch": 0.8720753350235422, "grad_norm": 1.3580735721080608, "learning_rate": 8.461274837108158e-07, "loss": 0.6793, "step": 24078 }, { "epoch": 0.8721115537848606, "grad_norm": 1.465018403767473, "learning_rate": 8.456552996154765e-07, "loss": 0.6478, "step": 24079 }, { "epoch": 0.8721477725461789, "grad_norm": 1.0006879081828044, "learning_rate": 8.451832414919658e-07, "loss": 0.6293, "step": 24080 }, { "epoch": 0.8721839913074972, "grad_norm": 1.437916016503453, "learning_rate": 8.447113093467752e-07, "loss": 0.6124, "step": 24081 }, { "epoch": 0.8722202100688157, "grad_norm": 1.016147056844971, "learning_rate": 8.442395031864037e-07, "loss": 0.6979, "step": 24082 }, { "epoch": 0.872256428830134, "grad_norm": 1.4406089347907614, "learning_rate": 8.437678230173407e-07, "loss": 0.6794, "step": 24083 }, { "epoch": 0.8722926475914524, "grad_norm": 1.0165699602101648, "learning_rate": 8.432962688460788e-07, "loss": 0.6692, "step": 24084 }, { "epoch": 0.8723288663527707, "grad_norm": 1.4222146011206194, "learning_rate": 8.42824840679104e-07, "loss": 0.6923, "step": 24085 }, { "epoch": 0.8723650851140891, "grad_norm": 1.3967613625760693, "learning_rate": 8.423535385229076e-07, "loss": 0.6595, "step": 24086 }, { "epoch": 0.8724013038754075, "grad_norm": 1.4376518490656278, "learning_rate": 8.418823623839722e-07, "loss": 0.6595, "step": 24087 }, { "epoch": 0.8724375226367258, "grad_norm": 1.2711407970720339, "learning_rate": 8.414113122687817e-07, "loss": 0.629, "step": 24088 }, { "epoch": 0.8724737413980442, "grad_norm": 1.3181642161844163, "learning_rate": 8.409403881838207e-07, "loss": 0.6401, "step": 24089 }, { "epoch": 0.8725099601593626, "grad_norm": 1.3868419820805673, "learning_rate": 8.404695901355675e-07, "loss": 0.6649, "step": 24090 }, { "epoch": 0.8725461789206809, "grad_norm": 1.425743220749662, "learning_rate": 8.399989181305013e-07, "loss": 0.6973, "step": 24091 }, { "epoch": 0.8725823976819993, "grad_norm": 1.4440348990445715, "learning_rate": 8.395283721750958e-07, "loss": 0.6382, "step": 24092 }, { "epoch": 0.8726186164433176, "grad_norm": 1.5291193150157227, "learning_rate": 8.390579522758324e-07, "loss": 0.6388, "step": 24093 }, { "epoch": 0.872654835204636, "grad_norm": 0.9851984502257274, "learning_rate": 8.385876584391794e-07, "loss": 0.6496, "step": 24094 }, { "epoch": 0.8726910539659544, "grad_norm": 1.0356901894247925, "learning_rate": 8.381174906716117e-07, "loss": 0.7208, "step": 24095 }, { "epoch": 0.8727272727272727, "grad_norm": 1.0607478038603686, "learning_rate": 8.37647448979595e-07, "loss": 0.6614, "step": 24096 }, { "epoch": 0.8727634914885911, "grad_norm": 1.282592424171366, "learning_rate": 8.371775333696019e-07, "loss": 0.621, "step": 24097 }, { "epoch": 0.8727997102499094, "grad_norm": 1.281897906581622, "learning_rate": 8.367077438480975e-07, "loss": 0.7264, "step": 24098 }, { "epoch": 0.8728359290112279, "grad_norm": 1.3232782830527028, "learning_rate": 8.362380804215442e-07, "loss": 0.6721, "step": 24099 }, { "epoch": 0.8728721477725462, "grad_norm": 1.1672214024580252, "learning_rate": 8.357685430964091e-07, "loss": 0.6725, "step": 24100 }, { "epoch": 0.8729083665338645, "grad_norm": 1.2496348749297466, "learning_rate": 8.352991318791503e-07, "loss": 0.6371, "step": 24101 }, { "epoch": 0.8729445852951829, "grad_norm": 1.5295620173224895, "learning_rate": 8.348298467762284e-07, "loss": 0.6606, "step": 24102 }, { "epoch": 0.8729808040565012, "grad_norm": 1.3253700426483965, "learning_rate": 8.343606877941002e-07, "loss": 0.6622, "step": 24103 }, { "epoch": 0.8730170228178197, "grad_norm": 1.334534792634661, "learning_rate": 8.33891654939224e-07, "loss": 0.6829, "step": 24104 }, { "epoch": 0.873053241579138, "grad_norm": 1.4861444154211985, "learning_rate": 8.334227482180524e-07, "loss": 0.7443, "step": 24105 }, { "epoch": 0.8730894603404563, "grad_norm": 1.4386986601814407, "learning_rate": 8.329539676370357e-07, "loss": 0.6857, "step": 24106 }, { "epoch": 0.8731256791017747, "grad_norm": 1.3677320513033062, "learning_rate": 8.3248531320263e-07, "loss": 0.697, "step": 24107 }, { "epoch": 0.8731618978630931, "grad_norm": 1.405363126415014, "learning_rate": 8.320167849212823e-07, "loss": 0.5992, "step": 24108 }, { "epoch": 0.8731981166244115, "grad_norm": 1.2255764485937934, "learning_rate": 8.315483827994387e-07, "loss": 0.5807, "step": 24109 }, { "epoch": 0.8732343353857298, "grad_norm": 1.2717125193266703, "learning_rate": 8.310801068435426e-07, "loss": 0.6737, "step": 24110 }, { "epoch": 0.8732705541470481, "grad_norm": 1.0492708495336849, "learning_rate": 8.306119570600435e-07, "loss": 0.7037, "step": 24111 }, { "epoch": 0.8733067729083666, "grad_norm": 1.459902808138934, "learning_rate": 8.301439334553807e-07, "loss": 0.7142, "step": 24112 }, { "epoch": 0.8733429916696849, "grad_norm": 0.9373640092798975, "learning_rate": 8.296760360359957e-07, "loss": 0.6824, "step": 24113 }, { "epoch": 0.8733792104310033, "grad_norm": 1.030424099059912, "learning_rate": 8.292082648083233e-07, "loss": 0.6817, "step": 24114 }, { "epoch": 0.8734154291923216, "grad_norm": 1.1157241389308568, "learning_rate": 8.28740619778805e-07, "loss": 0.7178, "step": 24115 }, { "epoch": 0.8734516479536399, "grad_norm": 1.1300068360042186, "learning_rate": 8.282731009538747e-07, "loss": 0.6682, "step": 24116 }, { "epoch": 0.8734878667149584, "grad_norm": 1.4688734537855348, "learning_rate": 8.278057083399638e-07, "loss": 0.6884, "step": 24117 }, { "epoch": 0.8735240854762767, "grad_norm": 1.254403844481776, "learning_rate": 8.273384419435071e-07, "loss": 0.6811, "step": 24118 }, { "epoch": 0.8735603042375951, "grad_norm": 1.3615987897221387, "learning_rate": 8.26871301770934e-07, "loss": 0.651, "step": 24119 }, { "epoch": 0.8735965229989134, "grad_norm": 1.887225276567664, "learning_rate": 8.264042878286715e-07, "loss": 0.7118, "step": 24120 }, { "epoch": 0.8736327417602318, "grad_norm": 1.348222013939025, "learning_rate": 8.259374001231447e-07, "loss": 0.6283, "step": 24121 }, { "epoch": 0.8736689605215502, "grad_norm": 1.439812113382773, "learning_rate": 8.254706386607825e-07, "loss": 0.6874, "step": 24122 }, { "epoch": 0.8737051792828685, "grad_norm": 1.3196832410096428, "learning_rate": 8.250040034480055e-07, "loss": 0.6552, "step": 24123 }, { "epoch": 0.8737413980441869, "grad_norm": 1.0292026154388507, "learning_rate": 8.245374944912354e-07, "loss": 0.755, "step": 24124 }, { "epoch": 0.8737776168055053, "grad_norm": 1.1157938514006098, "learning_rate": 8.240711117968903e-07, "loss": 0.6612, "step": 24125 }, { "epoch": 0.8738138355668236, "grad_norm": 1.4744895570593948, "learning_rate": 8.236048553713905e-07, "loss": 0.7081, "step": 24126 }, { "epoch": 0.873850054328142, "grad_norm": 1.185472504678705, "learning_rate": 8.231387252211509e-07, "loss": 0.6637, "step": 24127 }, { "epoch": 0.8738862730894603, "grad_norm": 1.3694755604351359, "learning_rate": 8.226727213525843e-07, "loss": 0.6587, "step": 24128 }, { "epoch": 0.8739224918507787, "grad_norm": 1.458671546499998, "learning_rate": 8.222068437721076e-07, "loss": 0.7156, "step": 24129 }, { "epoch": 0.8739587106120971, "grad_norm": 0.882399597758542, "learning_rate": 8.21741092486128e-07, "loss": 0.6486, "step": 24130 }, { "epoch": 0.8739949293734154, "grad_norm": 1.4616178532290045, "learning_rate": 8.212754675010559e-07, "loss": 0.7878, "step": 24131 }, { "epoch": 0.8740311481347338, "grad_norm": 1.3750620687087094, "learning_rate": 8.208099688232984e-07, "loss": 0.6669, "step": 24132 }, { "epoch": 0.8740673668960521, "grad_norm": 1.43955712867902, "learning_rate": 8.203445964592616e-07, "loss": 0.6496, "step": 24133 }, { "epoch": 0.8741035856573706, "grad_norm": 1.2845097976914515, "learning_rate": 8.198793504153491e-07, "loss": 0.6418, "step": 24134 }, { "epoch": 0.8741398044186889, "grad_norm": 1.4002720657609231, "learning_rate": 8.194142306979613e-07, "loss": 0.7182, "step": 24135 }, { "epoch": 0.8741760231800072, "grad_norm": 1.1863776073692651, "learning_rate": 8.189492373135021e-07, "loss": 0.6588, "step": 24136 }, { "epoch": 0.8742122419413256, "grad_norm": 1.468529319417242, "learning_rate": 8.184843702683687e-07, "loss": 0.6633, "step": 24137 }, { "epoch": 0.874248460702644, "grad_norm": 1.3677319750397345, "learning_rate": 8.18019629568958e-07, "loss": 0.6098, "step": 24138 }, { "epoch": 0.8742846794639624, "grad_norm": 0.9067968066929821, "learning_rate": 8.175550152216627e-07, "loss": 0.6327, "step": 24139 }, { "epoch": 0.8743208982252807, "grad_norm": 0.9437611416471763, "learning_rate": 8.17090527232881e-07, "loss": 0.6866, "step": 24140 }, { "epoch": 0.874357116986599, "grad_norm": 1.4686932552820162, "learning_rate": 8.166261656090035e-07, "loss": 0.6576, "step": 24141 }, { "epoch": 0.8743933357479174, "grad_norm": 1.1585663905561985, "learning_rate": 8.161619303564161e-07, "loss": 0.6973, "step": 24142 }, { "epoch": 0.8744295545092358, "grad_norm": 1.3272059563585596, "learning_rate": 8.156978214815125e-07, "loss": 0.641, "step": 24143 }, { "epoch": 0.8744657732705542, "grad_norm": 1.3038006933797377, "learning_rate": 8.152338389906767e-07, "loss": 0.6638, "step": 24144 }, { "epoch": 0.8745019920318725, "grad_norm": 1.433733909885813, "learning_rate": 8.147699828902933e-07, "loss": 0.6718, "step": 24145 }, { "epoch": 0.8745382107931908, "grad_norm": 1.2842712942599797, "learning_rate": 8.143062531867451e-07, "loss": 0.6013, "step": 24146 }, { "epoch": 0.8745744295545093, "grad_norm": 1.4254691749165689, "learning_rate": 8.13842649886415e-07, "loss": 0.68, "step": 24147 }, { "epoch": 0.8746106483158276, "grad_norm": 1.3856774584523495, "learning_rate": 8.133791729956819e-07, "loss": 0.6527, "step": 24148 }, { "epoch": 0.874646867077146, "grad_norm": 1.2980266607927888, "learning_rate": 8.129158225209222e-07, "loss": 0.6582, "step": 24149 }, { "epoch": 0.8746830858384643, "grad_norm": 1.2572165548653127, "learning_rate": 8.124525984685139e-07, "loss": 0.731, "step": 24150 }, { "epoch": 0.8747193045997826, "grad_norm": 1.3820489159894904, "learning_rate": 8.11989500844832e-07, "loss": 0.646, "step": 24151 }, { "epoch": 0.8747555233611011, "grad_norm": 1.7481786987211607, "learning_rate": 8.115265296562469e-07, "loss": 0.6677, "step": 24152 }, { "epoch": 0.8747917421224194, "grad_norm": 1.477307014623878, "learning_rate": 8.110636849091291e-07, "loss": 0.7411, "step": 24153 }, { "epoch": 0.8748279608837378, "grad_norm": 1.5589475684240557, "learning_rate": 8.106009666098514e-07, "loss": 0.658, "step": 24154 }, { "epoch": 0.8748641796450561, "grad_norm": 1.350441497421607, "learning_rate": 8.101383747647784e-07, "loss": 0.7013, "step": 24155 }, { "epoch": 0.8749003984063745, "grad_norm": 1.4033018906033887, "learning_rate": 8.096759093802752e-07, "loss": 0.7076, "step": 24156 }, { "epoch": 0.8749366171676929, "grad_norm": 1.4404238703982033, "learning_rate": 8.092135704627069e-07, "loss": 0.6898, "step": 24157 }, { "epoch": 0.8749728359290112, "grad_norm": 0.9897611175350991, "learning_rate": 8.087513580184358e-07, "loss": 0.6859, "step": 24158 }, { "epoch": 0.8750090546903296, "grad_norm": 0.9960537600320343, "learning_rate": 8.082892720538238e-07, "loss": 0.6498, "step": 24159 }, { "epoch": 0.875045273451648, "grad_norm": 1.3097086526850537, "learning_rate": 8.078273125752245e-07, "loss": 0.6929, "step": 24160 }, { "epoch": 0.8750814922129663, "grad_norm": 1.302068698959763, "learning_rate": 8.073654795890007e-07, "loss": 0.6556, "step": 24161 }, { "epoch": 0.8751177109742847, "grad_norm": 0.9060335424265102, "learning_rate": 8.069037731015062e-07, "loss": 0.6544, "step": 24162 }, { "epoch": 0.875153929735603, "grad_norm": 1.2660463832566284, "learning_rate": 8.064421931190925e-07, "loss": 0.62, "step": 24163 }, { "epoch": 0.8751901484969214, "grad_norm": 1.443414296046081, "learning_rate": 8.059807396481112e-07, "loss": 0.7129, "step": 24164 }, { "epoch": 0.8752263672582398, "grad_norm": 1.0747999141289215, "learning_rate": 8.055194126949151e-07, "loss": 0.6948, "step": 24165 }, { "epoch": 0.8752625860195581, "grad_norm": 1.198852361650515, "learning_rate": 8.050582122658501e-07, "loss": 0.6523, "step": 24166 }, { "epoch": 0.8752988047808765, "grad_norm": 1.4477890241972828, "learning_rate": 8.045971383672646e-07, "loss": 0.6661, "step": 24167 }, { "epoch": 0.8753350235421948, "grad_norm": 1.4256205494341687, "learning_rate": 8.041361910055002e-07, "loss": 0.6759, "step": 24168 }, { "epoch": 0.8753712423035133, "grad_norm": 1.329240536623876, "learning_rate": 8.036753701869026e-07, "loss": 0.706, "step": 24169 }, { "epoch": 0.8754074610648316, "grad_norm": 1.263415819194834, "learning_rate": 8.032146759178138e-07, "loss": 0.5936, "step": 24170 }, { "epoch": 0.8754436798261499, "grad_norm": 1.6230630655777971, "learning_rate": 8.027541082045698e-07, "loss": 0.68, "step": 24171 }, { "epoch": 0.8754798985874683, "grad_norm": 1.3933236185185585, "learning_rate": 8.02293667053513e-07, "loss": 0.706, "step": 24172 }, { "epoch": 0.8755161173487866, "grad_norm": 1.34764941738995, "learning_rate": 8.018333524709765e-07, "loss": 0.6451, "step": 24173 }, { "epoch": 0.8755523361101051, "grad_norm": 1.4024720113376736, "learning_rate": 8.01373164463296e-07, "loss": 0.7256, "step": 24174 }, { "epoch": 0.8755885548714234, "grad_norm": 1.4230974360658488, "learning_rate": 8.009131030368012e-07, "loss": 0.6664, "step": 24175 }, { "epoch": 0.8756247736327417, "grad_norm": 1.374723265716586, "learning_rate": 8.004531681978267e-07, "loss": 0.668, "step": 24176 }, { "epoch": 0.8756609923940601, "grad_norm": 1.2992070871358408, "learning_rate": 7.999933599526999e-07, "loss": 0.6252, "step": 24177 }, { "epoch": 0.8756972111553785, "grad_norm": 1.2394834963746972, "learning_rate": 7.99533678307749e-07, "loss": 0.6475, "step": 24178 }, { "epoch": 0.8757334299166969, "grad_norm": 1.353628649953583, "learning_rate": 7.990741232692967e-07, "loss": 0.6904, "step": 24179 }, { "epoch": 0.8757696486780152, "grad_norm": 1.5093085706301006, "learning_rate": 7.986146948436713e-07, "loss": 0.7083, "step": 24180 }, { "epoch": 0.8758058674393335, "grad_norm": 1.5248074851701199, "learning_rate": 7.981553930371922e-07, "loss": 0.6693, "step": 24181 }, { "epoch": 0.875842086200652, "grad_norm": 1.308481836515321, "learning_rate": 7.976962178561787e-07, "loss": 0.6582, "step": 24182 }, { "epoch": 0.8758783049619703, "grad_norm": 1.5552327359727873, "learning_rate": 7.972371693069525e-07, "loss": 0.6326, "step": 24183 }, { "epoch": 0.8759145237232887, "grad_norm": 1.530955569755731, "learning_rate": 7.967782473958296e-07, "loss": 0.7072, "step": 24184 }, { "epoch": 0.875950742484607, "grad_norm": 0.9701897733585293, "learning_rate": 7.963194521291251e-07, "loss": 0.6863, "step": 24185 }, { "epoch": 0.8759869612459253, "grad_norm": 1.16547604859191, "learning_rate": 7.958607835131493e-07, "loss": 0.6652, "step": 24186 }, { "epoch": 0.8760231800072438, "grad_norm": 0.9131683908108666, "learning_rate": 7.954022415542195e-07, "loss": 0.6869, "step": 24187 }, { "epoch": 0.8760593987685621, "grad_norm": 1.4095518634524897, "learning_rate": 7.949438262586418e-07, "loss": 0.7064, "step": 24188 }, { "epoch": 0.8760956175298805, "grad_norm": 1.3170112067372066, "learning_rate": 7.944855376327254e-07, "loss": 0.619, "step": 24189 }, { "epoch": 0.8761318362911988, "grad_norm": 1.1715527656706402, "learning_rate": 7.940273756827754e-07, "loss": 0.6823, "step": 24190 }, { "epoch": 0.8761680550525172, "grad_norm": 1.3829774697547899, "learning_rate": 7.93569340415099e-07, "loss": 0.7188, "step": 24191 }, { "epoch": 0.8762042738138356, "grad_norm": 1.2705386185191605, "learning_rate": 7.931114318359978e-07, "loss": 0.5823, "step": 24192 }, { "epoch": 0.8762404925751539, "grad_norm": 1.630807287560774, "learning_rate": 7.926536499517723e-07, "loss": 0.6455, "step": 24193 }, { "epoch": 0.8762767113364723, "grad_norm": 1.3511117296488304, "learning_rate": 7.921959947687241e-07, "loss": 0.6186, "step": 24194 }, { "epoch": 0.8763129300977907, "grad_norm": 1.3938571438639857, "learning_rate": 7.917384662931493e-07, "loss": 0.6979, "step": 24195 }, { "epoch": 0.876349148859109, "grad_norm": 1.3902939833777528, "learning_rate": 7.912810645313451e-07, "loss": 0.6861, "step": 24196 }, { "epoch": 0.8763853676204274, "grad_norm": 1.3259416473057692, "learning_rate": 7.90823789489602e-07, "loss": 0.7333, "step": 24197 }, { "epoch": 0.8764215863817457, "grad_norm": 1.3106868124837674, "learning_rate": 7.903666411742184e-07, "loss": 0.7029, "step": 24198 }, { "epoch": 0.8764578051430642, "grad_norm": 1.3101909597595287, "learning_rate": 7.899096195914813e-07, "loss": 0.6319, "step": 24199 }, { "epoch": 0.8764940239043825, "grad_norm": 1.4025508692298494, "learning_rate": 7.894527247476813e-07, "loss": 0.6986, "step": 24200 }, { "epoch": 0.8765302426657008, "grad_norm": 1.4619405211485987, "learning_rate": 7.889959566491023e-07, "loss": 0.677, "step": 24201 }, { "epoch": 0.8765664614270192, "grad_norm": 1.5649284772889336, "learning_rate": 7.885393153020348e-07, "loss": 0.7075, "step": 24202 }, { "epoch": 0.8766026801883375, "grad_norm": 1.3642462031553826, "learning_rate": 7.880828007127605e-07, "loss": 0.6496, "step": 24203 }, { "epoch": 0.876638898949656, "grad_norm": 1.342013951612131, "learning_rate": 7.876264128875588e-07, "loss": 0.6474, "step": 24204 }, { "epoch": 0.8766751177109743, "grad_norm": 1.4271445230054702, "learning_rate": 7.871701518327146e-07, "loss": 0.6596, "step": 24205 }, { "epoch": 0.8767113364722926, "grad_norm": 1.4276162116888955, "learning_rate": 7.86714017554503e-07, "loss": 0.6854, "step": 24206 }, { "epoch": 0.876747555233611, "grad_norm": 1.254939646491449, "learning_rate": 7.862580100592032e-07, "loss": 0.6291, "step": 24207 }, { "epoch": 0.8767837739949293, "grad_norm": 1.43575219612852, "learning_rate": 7.858021293530871e-07, "loss": 0.7241, "step": 24208 }, { "epoch": 0.8768199927562478, "grad_norm": 1.3068949084188524, "learning_rate": 7.853463754424318e-07, "loss": 0.6834, "step": 24209 }, { "epoch": 0.8768562115175661, "grad_norm": 1.0425374845194235, "learning_rate": 7.848907483335066e-07, "loss": 0.685, "step": 24210 }, { "epoch": 0.8768924302788844, "grad_norm": 1.3739376081976722, "learning_rate": 7.844352480325813e-07, "loss": 0.7211, "step": 24211 }, { "epoch": 0.8769286490402028, "grad_norm": 1.403543727932083, "learning_rate": 7.839798745459249e-07, "loss": 0.6651, "step": 24212 }, { "epoch": 0.8769648678015212, "grad_norm": 1.4889757871147473, "learning_rate": 7.835246278798037e-07, "loss": 0.7499, "step": 24213 }, { "epoch": 0.8770010865628396, "grad_norm": 1.494656471264947, "learning_rate": 7.830695080404815e-07, "loss": 0.6783, "step": 24214 }, { "epoch": 0.8770373053241579, "grad_norm": 1.3923268894079668, "learning_rate": 7.826145150342213e-07, "loss": 0.6388, "step": 24215 }, { "epoch": 0.8770735240854762, "grad_norm": 1.3186830513593708, "learning_rate": 7.821596488672845e-07, "loss": 0.7217, "step": 24216 }, { "epoch": 0.8771097428467947, "grad_norm": 1.493747761133144, "learning_rate": 7.817049095459317e-07, "loss": 0.7446, "step": 24217 }, { "epoch": 0.877145961608113, "grad_norm": 1.3332182581858865, "learning_rate": 7.81250297076419e-07, "loss": 0.6568, "step": 24218 }, { "epoch": 0.8771821803694314, "grad_norm": 1.2822962195076468, "learning_rate": 7.807958114650005e-07, "loss": 0.6432, "step": 24219 }, { "epoch": 0.8772183991307497, "grad_norm": 1.3681319060756896, "learning_rate": 7.803414527179343e-07, "loss": 0.6377, "step": 24220 }, { "epoch": 0.877254617892068, "grad_norm": 1.0156897391836341, "learning_rate": 7.798872208414698e-07, "loss": 0.6403, "step": 24221 }, { "epoch": 0.8772908366533865, "grad_norm": 1.8061370054874044, "learning_rate": 7.794331158418577e-07, "loss": 0.6319, "step": 24222 }, { "epoch": 0.8773270554147048, "grad_norm": 1.4015008869258496, "learning_rate": 7.789791377253497e-07, "loss": 0.6946, "step": 24223 }, { "epoch": 0.8773632741760232, "grad_norm": 1.7110376237560052, "learning_rate": 7.785252864981907e-07, "loss": 0.6984, "step": 24224 }, { "epoch": 0.8773994929373415, "grad_norm": 1.2705146498353195, "learning_rate": 7.780715621666268e-07, "loss": 0.6411, "step": 24225 }, { "epoch": 0.8774357116986599, "grad_norm": 0.9617061370902596, "learning_rate": 7.776179647368997e-07, "loss": 0.6528, "step": 24226 }, { "epoch": 0.8774719304599783, "grad_norm": 1.303473757482143, "learning_rate": 7.771644942152546e-07, "loss": 0.6158, "step": 24227 }, { "epoch": 0.8775081492212966, "grad_norm": 1.4328278729239123, "learning_rate": 7.767111506079294e-07, "loss": 0.7103, "step": 24228 }, { "epoch": 0.877544367982615, "grad_norm": 1.3433466884094956, "learning_rate": 7.76257933921164e-07, "loss": 0.6328, "step": 24229 }, { "epoch": 0.8775805867439334, "grad_norm": 1.071614466230358, "learning_rate": 7.758048441611921e-07, "loss": 0.6461, "step": 24230 }, { "epoch": 0.8776168055052517, "grad_norm": 1.0745532019109294, "learning_rate": 7.75351881334252e-07, "loss": 0.7037, "step": 24231 }, { "epoch": 0.8776530242665701, "grad_norm": 1.4351004431853613, "learning_rate": 7.748990454465755e-07, "loss": 0.7336, "step": 24232 }, { "epoch": 0.8776892430278884, "grad_norm": 1.021803682164724, "learning_rate": 7.744463365043919e-07, "loss": 0.6904, "step": 24233 }, { "epoch": 0.8777254617892069, "grad_norm": 1.2365494050965902, "learning_rate": 7.739937545139354e-07, "loss": 0.5896, "step": 24234 }, { "epoch": 0.8777616805505252, "grad_norm": 1.276161383021451, "learning_rate": 7.735412994814318e-07, "loss": 0.7303, "step": 24235 }, { "epoch": 0.8777978993118435, "grad_norm": 0.9778930588360456, "learning_rate": 7.730889714131062e-07, "loss": 0.7065, "step": 24236 }, { "epoch": 0.8778341180731619, "grad_norm": 1.2731578701263004, "learning_rate": 7.726367703151816e-07, "loss": 0.6266, "step": 24237 }, { "epoch": 0.8778703368344802, "grad_norm": 1.412571000587089, "learning_rate": 7.721846961938861e-07, "loss": 0.6906, "step": 24238 }, { "epoch": 0.8779065555957987, "grad_norm": 1.4524916671010804, "learning_rate": 7.717327490554372e-07, "loss": 0.6445, "step": 24239 }, { "epoch": 0.877942774357117, "grad_norm": 1.309764413864345, "learning_rate": 7.71280928906053e-07, "loss": 0.7105, "step": 24240 }, { "epoch": 0.8779789931184353, "grad_norm": 1.5003828762085754, "learning_rate": 7.70829235751952e-07, "loss": 0.5617, "step": 24241 }, { "epoch": 0.8780152118797537, "grad_norm": 1.0054967646773432, "learning_rate": 7.703776695993515e-07, "loss": 0.6888, "step": 24242 }, { "epoch": 0.878051430641072, "grad_norm": 1.4379576843322577, "learning_rate": 7.699262304544641e-07, "loss": 0.6061, "step": 24243 }, { "epoch": 0.8780876494023905, "grad_norm": 1.3460303339242248, "learning_rate": 7.694749183234995e-07, "loss": 0.6515, "step": 24244 }, { "epoch": 0.8781238681637088, "grad_norm": 1.3031670296589688, "learning_rate": 7.690237332126737e-07, "loss": 0.6719, "step": 24245 }, { "epoch": 0.8781600869250271, "grad_norm": 1.0897245317329696, "learning_rate": 7.685726751281919e-07, "loss": 0.6913, "step": 24246 }, { "epoch": 0.8781963056863455, "grad_norm": 0.9688442057279055, "learning_rate": 7.681217440762611e-07, "loss": 0.6718, "step": 24247 }, { "epoch": 0.8782325244476639, "grad_norm": 1.4286299198200516, "learning_rate": 7.676709400630855e-07, "loss": 0.7657, "step": 24248 }, { "epoch": 0.8782687432089823, "grad_norm": 1.4036103806916558, "learning_rate": 7.672202630948722e-07, "loss": 0.6783, "step": 24249 }, { "epoch": 0.8783049619703006, "grad_norm": 1.1911706138365104, "learning_rate": 7.667697131778196e-07, "loss": 0.6577, "step": 24250 }, { "epoch": 0.8783411807316189, "grad_norm": 1.3728302164734256, "learning_rate": 7.663192903181293e-07, "loss": 0.5405, "step": 24251 }, { "epoch": 0.8783773994929374, "grad_norm": 1.0444919302156577, "learning_rate": 7.658689945219977e-07, "loss": 0.6719, "step": 24252 }, { "epoch": 0.8784136182542557, "grad_norm": 1.4288342488514472, "learning_rate": 7.654188257956241e-07, "loss": 0.6838, "step": 24253 }, { "epoch": 0.8784498370155741, "grad_norm": 0.9427771341525706, "learning_rate": 7.649687841452014e-07, "loss": 0.681, "step": 24254 }, { "epoch": 0.8784860557768924, "grad_norm": 1.4479977426510924, "learning_rate": 7.645188695769212e-07, "loss": 0.7006, "step": 24255 }, { "epoch": 0.8785222745382107, "grad_norm": 1.4716370790941784, "learning_rate": 7.640690820969776e-07, "loss": 0.5833, "step": 24256 }, { "epoch": 0.8785584932995292, "grad_norm": 1.199576340863028, "learning_rate": 7.6361942171156e-07, "loss": 0.59, "step": 24257 }, { "epoch": 0.8785947120608475, "grad_norm": 1.339107429269211, "learning_rate": 7.631698884268546e-07, "loss": 0.6642, "step": 24258 }, { "epoch": 0.8786309308221659, "grad_norm": 1.2913989408416586, "learning_rate": 7.627204822490453e-07, "loss": 0.5917, "step": 24259 }, { "epoch": 0.8786671495834842, "grad_norm": 1.3330421387441236, "learning_rate": 7.622712031843204e-07, "loss": 0.706, "step": 24260 }, { "epoch": 0.8787033683448026, "grad_norm": 1.2828906091405399, "learning_rate": 7.618220512388619e-07, "loss": 0.6508, "step": 24261 }, { "epoch": 0.878739587106121, "grad_norm": 1.1320894920462827, "learning_rate": 7.61373026418849e-07, "loss": 0.7236, "step": 24262 }, { "epoch": 0.8787758058674393, "grad_norm": 1.111326802593175, "learning_rate": 7.609241287304581e-07, "loss": 0.6614, "step": 24263 }, { "epoch": 0.8788120246287577, "grad_norm": 1.2667255973230886, "learning_rate": 7.604753581798718e-07, "loss": 0.6231, "step": 24264 }, { "epoch": 0.8788482433900761, "grad_norm": 1.0160617711055038, "learning_rate": 7.600267147732643e-07, "loss": 0.681, "step": 24265 }, { "epoch": 0.8788844621513944, "grad_norm": 1.5330319351551522, "learning_rate": 7.595781985168049e-07, "loss": 0.6975, "step": 24266 }, { "epoch": 0.8789206809127128, "grad_norm": 1.4449335574502575, "learning_rate": 7.591298094166711e-07, "loss": 0.6589, "step": 24267 }, { "epoch": 0.8789568996740311, "grad_norm": 1.5812248894710563, "learning_rate": 7.58681547479031e-07, "loss": 0.6792, "step": 24268 }, { "epoch": 0.8789931184353496, "grad_norm": 1.1642285110998525, "learning_rate": 7.582334127100521e-07, "loss": 0.6735, "step": 24269 }, { "epoch": 0.8790293371966679, "grad_norm": 1.4150962607418296, "learning_rate": 7.577854051159017e-07, "loss": 0.7184, "step": 24270 }, { "epoch": 0.8790655559579862, "grad_norm": 1.457559406751026, "learning_rate": 7.573375247027459e-07, "loss": 0.6638, "step": 24271 }, { "epoch": 0.8791017747193046, "grad_norm": 1.4157774810479828, "learning_rate": 7.568897714767464e-07, "loss": 0.72, "step": 24272 }, { "epoch": 0.8791379934806229, "grad_norm": 1.442518666919718, "learning_rate": 7.564421454440651e-07, "loss": 0.754, "step": 24273 }, { "epoch": 0.8791742122419414, "grad_norm": 1.4242846564205658, "learning_rate": 7.559946466108625e-07, "loss": 0.6635, "step": 24274 }, { "epoch": 0.8792104310032597, "grad_norm": 1.0919756024386806, "learning_rate": 7.55547274983297e-07, "loss": 0.6342, "step": 24275 }, { "epoch": 0.879246649764578, "grad_norm": 1.5323377049862588, "learning_rate": 7.551000305675238e-07, "loss": 0.648, "step": 24276 }, { "epoch": 0.8792828685258964, "grad_norm": 1.4626742108677417, "learning_rate": 7.546529133696944e-07, "loss": 0.6409, "step": 24277 }, { "epoch": 0.8793190872872148, "grad_norm": 1.291318779568894, "learning_rate": 7.542059233959676e-07, "loss": 0.6612, "step": 24278 }, { "epoch": 0.8793553060485332, "grad_norm": 1.39235869373521, "learning_rate": 7.537590606524914e-07, "loss": 0.7073, "step": 24279 }, { "epoch": 0.8793915248098515, "grad_norm": 0.9703038530695827, "learning_rate": 7.533123251454145e-07, "loss": 0.6489, "step": 24280 }, { "epoch": 0.8794277435711698, "grad_norm": 1.7311953424956228, "learning_rate": 7.528657168808829e-07, "loss": 0.6768, "step": 24281 }, { "epoch": 0.8794639623324882, "grad_norm": 1.1012010036139657, "learning_rate": 7.524192358650461e-07, "loss": 0.5728, "step": 24282 }, { "epoch": 0.8795001810938066, "grad_norm": 1.3438131336972503, "learning_rate": 7.519728821040473e-07, "loss": 0.6703, "step": 24283 }, { "epoch": 0.879536399855125, "grad_norm": 1.4054504359940887, "learning_rate": 7.515266556040246e-07, "loss": 0.6716, "step": 24284 }, { "epoch": 0.8795726186164433, "grad_norm": 1.3795751915839023, "learning_rate": 7.510805563711243e-07, "loss": 0.6658, "step": 24285 }, { "epoch": 0.8796088373777616, "grad_norm": 1.4462425512939696, "learning_rate": 7.506345844114816e-07, "loss": 0.6499, "step": 24286 }, { "epoch": 0.8796450561390801, "grad_norm": 1.0842672543819443, "learning_rate": 7.501887397312346e-07, "loss": 0.6551, "step": 24287 }, { "epoch": 0.8796812749003984, "grad_norm": 1.050609268382673, "learning_rate": 7.497430223365165e-07, "loss": 0.5993, "step": 24288 }, { "epoch": 0.8797174936617168, "grad_norm": 1.3128699096457335, "learning_rate": 7.492974322334634e-07, "loss": 0.608, "step": 24289 }, { "epoch": 0.8797537124230351, "grad_norm": 1.0650189015479967, "learning_rate": 7.48851969428207e-07, "loss": 0.6797, "step": 24290 }, { "epoch": 0.8797899311843534, "grad_norm": 1.2418647236541647, "learning_rate": 7.484066339268759e-07, "loss": 0.7159, "step": 24291 }, { "epoch": 0.8798261499456719, "grad_norm": 1.3932252747193163, "learning_rate": 7.479614257355972e-07, "loss": 0.6426, "step": 24292 }, { "epoch": 0.8798623687069902, "grad_norm": 1.4423687197462316, "learning_rate": 7.475163448604994e-07, "loss": 0.595, "step": 24293 }, { "epoch": 0.8798985874683086, "grad_norm": 1.4699484952228892, "learning_rate": 7.470713913077077e-07, "loss": 0.6659, "step": 24294 }, { "epoch": 0.8799348062296269, "grad_norm": 1.4974769530230363, "learning_rate": 7.466265650833416e-07, "loss": 0.7215, "step": 24295 }, { "epoch": 0.8799710249909453, "grad_norm": 1.580751970892309, "learning_rate": 7.461818661935271e-07, "loss": 0.6037, "step": 24296 }, { "epoch": 0.8800072437522637, "grad_norm": 1.3265134233937943, "learning_rate": 7.457372946443808e-07, "loss": 0.6847, "step": 24297 }, { "epoch": 0.880043462513582, "grad_norm": 1.5306074323274919, "learning_rate": 7.45292850442021e-07, "loss": 0.6685, "step": 24298 }, { "epoch": 0.8800796812749004, "grad_norm": 1.2820197299576845, "learning_rate": 7.448485335925626e-07, "loss": 0.6169, "step": 24299 }, { "epoch": 0.8801159000362188, "grad_norm": 1.3434573911186811, "learning_rate": 7.444043441021209e-07, "loss": 0.6647, "step": 24300 }, { "epoch": 0.8801521187975371, "grad_norm": 0.9140165169802688, "learning_rate": 7.439602819768088e-07, "loss": 0.6622, "step": 24301 }, { "epoch": 0.8801883375588555, "grad_norm": 1.411388697407902, "learning_rate": 7.435163472227369e-07, "loss": 0.637, "step": 24302 }, { "epoch": 0.8802245563201738, "grad_norm": 1.2830549260497581, "learning_rate": 7.430725398460103e-07, "loss": 0.6067, "step": 24303 }, { "epoch": 0.8802607750814923, "grad_norm": 1.3495045053946402, "learning_rate": 7.42628859852742e-07, "loss": 0.668, "step": 24304 }, { "epoch": 0.8802969938428106, "grad_norm": 1.6559213861119488, "learning_rate": 7.421853072490348e-07, "loss": 0.6945, "step": 24305 }, { "epoch": 0.8803332126041289, "grad_norm": 1.2540496769396126, "learning_rate": 7.417418820409917e-07, "loss": 0.6594, "step": 24306 }, { "epoch": 0.8803694313654473, "grad_norm": 0.9929994695066667, "learning_rate": 7.412985842347153e-07, "loss": 0.677, "step": 24307 }, { "epoch": 0.8804056501267656, "grad_norm": 1.3776031177447052, "learning_rate": 7.408554138363067e-07, "loss": 0.6932, "step": 24308 }, { "epoch": 0.8804418688880841, "grad_norm": 1.2883052769323693, "learning_rate": 7.404123708518629e-07, "loss": 0.6683, "step": 24309 }, { "epoch": 0.8804780876494024, "grad_norm": 1.0395474531273627, "learning_rate": 7.399694552874803e-07, "loss": 0.6684, "step": 24310 }, { "epoch": 0.8805143064107207, "grad_norm": 1.4033111148551396, "learning_rate": 7.395266671492552e-07, "loss": 0.684, "step": 24311 }, { "epoch": 0.8805505251720391, "grad_norm": 1.3616963761407128, "learning_rate": 7.390840064432802e-07, "loss": 0.7632, "step": 24312 }, { "epoch": 0.8805867439333575, "grad_norm": 1.3797003563987185, "learning_rate": 7.386414731756475e-07, "loss": 0.6753, "step": 24313 }, { "epoch": 0.8806229626946759, "grad_norm": 1.1503182340702958, "learning_rate": 7.381990673524431e-07, "loss": 0.7472, "step": 24314 }, { "epoch": 0.8806591814559942, "grad_norm": 1.3784657036640264, "learning_rate": 7.377567889797598e-07, "loss": 0.6584, "step": 24315 }, { "epoch": 0.8806954002173125, "grad_norm": 1.3507892044698042, "learning_rate": 7.37314638063682e-07, "loss": 0.6642, "step": 24316 }, { "epoch": 0.880731618978631, "grad_norm": 1.4390889383842669, "learning_rate": 7.368726146102923e-07, "loss": 0.6395, "step": 24317 }, { "epoch": 0.8807678377399493, "grad_norm": 1.373493943204664, "learning_rate": 7.364307186256758e-07, "loss": 0.6714, "step": 24318 }, { "epoch": 0.8808040565012677, "grad_norm": 1.5374284047227942, "learning_rate": 7.359889501159123e-07, "loss": 0.6829, "step": 24319 }, { "epoch": 0.880840275262586, "grad_norm": 1.111277004356321, "learning_rate": 7.355473090870813e-07, "loss": 0.6951, "step": 24320 }, { "epoch": 0.8808764940239043, "grad_norm": 1.7508559262556331, "learning_rate": 7.351057955452567e-07, "loss": 0.6875, "step": 24321 }, { "epoch": 0.8809127127852228, "grad_norm": 1.5285705995215255, "learning_rate": 7.346644094965205e-07, "loss": 0.6728, "step": 24322 }, { "epoch": 0.8809489315465411, "grad_norm": 1.3106998668404373, "learning_rate": 7.342231509469422e-07, "loss": 0.6712, "step": 24323 }, { "epoch": 0.8809851503078595, "grad_norm": 1.303003662221242, "learning_rate": 7.337820199025958e-07, "loss": 0.7005, "step": 24324 }, { "epoch": 0.8810213690691778, "grad_norm": 1.521748915213801, "learning_rate": 7.333410163695487e-07, "loss": 0.6442, "step": 24325 }, { "epoch": 0.8810575878304961, "grad_norm": 1.233924559453372, "learning_rate": 7.329001403538738e-07, "loss": 0.6712, "step": 24326 }, { "epoch": 0.8810938065918146, "grad_norm": 1.3307683263032237, "learning_rate": 7.324593918616352e-07, "loss": 0.5939, "step": 24327 }, { "epoch": 0.8811300253531329, "grad_norm": 1.3014403025380745, "learning_rate": 7.320187708988968e-07, "loss": 0.6355, "step": 24328 }, { "epoch": 0.8811662441144513, "grad_norm": 1.7852277657939617, "learning_rate": 7.315782774717262e-07, "loss": 0.7009, "step": 24329 }, { "epoch": 0.8812024628757696, "grad_norm": 1.3493885102463676, "learning_rate": 7.311379115861816e-07, "loss": 0.6351, "step": 24330 }, { "epoch": 0.881238681637088, "grad_norm": 1.2047494574581907, "learning_rate": 7.30697673248324e-07, "loss": 0.676, "step": 24331 }, { "epoch": 0.8812749003984064, "grad_norm": 1.3851638995979738, "learning_rate": 7.302575624642094e-07, "loss": 0.6724, "step": 24332 }, { "epoch": 0.8813111191597247, "grad_norm": 1.266296974358873, "learning_rate": 7.298175792398976e-07, "loss": 0.7175, "step": 24333 }, { "epoch": 0.8813473379210431, "grad_norm": 1.4907964418894823, "learning_rate": 7.293777235814415e-07, "loss": 0.6968, "step": 24334 }, { "epoch": 0.8813835566823615, "grad_norm": 1.360565624660959, "learning_rate": 7.289379954948916e-07, "loss": 0.682, "step": 24335 }, { "epoch": 0.8814197754436798, "grad_norm": 1.401793132739357, "learning_rate": 7.284983949863034e-07, "loss": 0.6702, "step": 24336 }, { "epoch": 0.8814559942049982, "grad_norm": 1.322678819780384, "learning_rate": 7.280589220617229e-07, "loss": 0.5487, "step": 24337 }, { "epoch": 0.8814922129663165, "grad_norm": 1.3192483010258524, "learning_rate": 7.276195767271998e-07, "loss": 0.6396, "step": 24338 }, { "epoch": 0.881528431727635, "grad_norm": 1.4573008540979355, "learning_rate": 7.271803589887771e-07, "loss": 0.7569, "step": 24339 }, { "epoch": 0.8815646504889533, "grad_norm": 0.9816627893714066, "learning_rate": 7.26741268852501e-07, "loss": 0.6899, "step": 24340 }, { "epoch": 0.8816008692502716, "grad_norm": 1.2571824347541916, "learning_rate": 7.263023063244146e-07, "loss": 0.7013, "step": 24341 }, { "epoch": 0.88163708801159, "grad_norm": 1.0630825315638077, "learning_rate": 7.258634714105561e-07, "loss": 0.6717, "step": 24342 }, { "epoch": 0.8816733067729083, "grad_norm": 1.5010250836970152, "learning_rate": 7.254247641169654e-07, "loss": 0.7423, "step": 24343 }, { "epoch": 0.8817095255342268, "grad_norm": 1.5704397474990115, "learning_rate": 7.249861844496797e-07, "loss": 0.6515, "step": 24344 }, { "epoch": 0.8817457442955451, "grad_norm": 1.3638285287127803, "learning_rate": 7.245477324147343e-07, "loss": 0.7258, "step": 24345 }, { "epoch": 0.8817819630568634, "grad_norm": 1.2557728190279656, "learning_rate": 7.241094080181599e-07, "loss": 0.6414, "step": 24346 }, { "epoch": 0.8818181818181818, "grad_norm": 1.0995891087046694, "learning_rate": 7.236712112659916e-07, "loss": 0.7031, "step": 24347 }, { "epoch": 0.8818544005795002, "grad_norm": 1.4086093832429551, "learning_rate": 7.232331421642591e-07, "loss": 0.7005, "step": 24348 }, { "epoch": 0.8818906193408186, "grad_norm": 0.9529453195187715, "learning_rate": 7.227952007189876e-07, "loss": 0.6909, "step": 24349 }, { "epoch": 0.8819268381021369, "grad_norm": 1.3619301373359018, "learning_rate": 7.223573869362077e-07, "loss": 0.675, "step": 24350 }, { "epoch": 0.8819630568634552, "grad_norm": 1.2876799097504001, "learning_rate": 7.219197008219425e-07, "loss": 0.6355, "step": 24351 }, { "epoch": 0.8819992756247736, "grad_norm": 1.3830380760892995, "learning_rate": 7.21482142382215e-07, "loss": 0.5848, "step": 24352 }, { "epoch": 0.882035494386092, "grad_norm": 1.3168636300008467, "learning_rate": 7.210447116230435e-07, "loss": 0.6346, "step": 24353 }, { "epoch": 0.8820717131474104, "grad_norm": 1.4878586196561787, "learning_rate": 7.206074085504522e-07, "loss": 0.6603, "step": 24354 }, { "epoch": 0.8821079319087287, "grad_norm": 1.3094147867904211, "learning_rate": 7.201702331704563e-07, "loss": 0.6766, "step": 24355 }, { "epoch": 0.882144150670047, "grad_norm": 1.0346186611727666, "learning_rate": 7.197331854890721e-07, "loss": 0.6822, "step": 24356 }, { "epoch": 0.8821803694313655, "grad_norm": 1.3164285260305157, "learning_rate": 7.192962655123126e-07, "loss": 0.5942, "step": 24357 }, { "epoch": 0.8822165881926838, "grad_norm": 1.5571879402326911, "learning_rate": 7.188594732461918e-07, "loss": 0.6709, "step": 24358 }, { "epoch": 0.8822528069540022, "grad_norm": 1.4099381217373474, "learning_rate": 7.184228086967204e-07, "loss": 0.6483, "step": 24359 }, { "epoch": 0.8822890257153205, "grad_norm": 1.4794427303193391, "learning_rate": 7.179862718699049e-07, "loss": 0.7195, "step": 24360 }, { "epoch": 0.8823252444766388, "grad_norm": 1.4874860006657113, "learning_rate": 7.175498627717559e-07, "loss": 0.7035, "step": 24361 }, { "epoch": 0.8823614632379573, "grad_norm": 1.7201726812457507, "learning_rate": 7.171135814082764e-07, "loss": 0.724, "step": 24362 }, { "epoch": 0.8823976819992756, "grad_norm": 1.428187657767004, "learning_rate": 7.166774277854716e-07, "loss": 0.6781, "step": 24363 }, { "epoch": 0.882433900760594, "grad_norm": 1.3670437055991116, "learning_rate": 7.1624140190934e-07, "loss": 0.6642, "step": 24364 }, { "epoch": 0.8824701195219123, "grad_norm": 1.4592950916468292, "learning_rate": 7.158055037858858e-07, "loss": 0.6611, "step": 24365 }, { "epoch": 0.8825063382832307, "grad_norm": 1.260904654030162, "learning_rate": 7.153697334211063e-07, "loss": 0.6076, "step": 24366 }, { "epoch": 0.8825425570445491, "grad_norm": 1.4462886014392846, "learning_rate": 7.149340908209956e-07, "loss": 0.6418, "step": 24367 }, { "epoch": 0.8825787758058674, "grad_norm": 1.330474424770078, "learning_rate": 7.144985759915524e-07, "loss": 0.6533, "step": 24368 }, { "epoch": 0.8826149945671858, "grad_norm": 0.9812366834988043, "learning_rate": 7.140631889387672e-07, "loss": 0.7234, "step": 24369 }, { "epoch": 0.8826512133285042, "grad_norm": 0.9911668678408884, "learning_rate": 7.136279296686322e-07, "loss": 0.6486, "step": 24370 }, { "epoch": 0.8826874320898225, "grad_norm": 1.4164595177844632, "learning_rate": 7.131927981871345e-07, "loss": 0.6134, "step": 24371 }, { "epoch": 0.8827236508511409, "grad_norm": 1.4384793535970735, "learning_rate": 7.127577945002662e-07, "loss": 0.6869, "step": 24372 }, { "epoch": 0.8827598696124592, "grad_norm": 1.4549547335131177, "learning_rate": 7.123229186140101e-07, "loss": 0.6862, "step": 24373 }, { "epoch": 0.8827960883737777, "grad_norm": 1.3420123867002105, "learning_rate": 7.118881705343527e-07, "loss": 0.6061, "step": 24374 }, { "epoch": 0.882832307135096, "grad_norm": 1.3209490158751376, "learning_rate": 7.114535502672726e-07, "loss": 0.6395, "step": 24375 }, { "epoch": 0.8828685258964143, "grad_norm": 1.4159733654161764, "learning_rate": 7.11019057818756e-07, "loss": 0.6315, "step": 24376 }, { "epoch": 0.8829047446577327, "grad_norm": 1.4987077076081436, "learning_rate": 7.105846931947791e-07, "loss": 0.6978, "step": 24377 }, { "epoch": 0.882940963419051, "grad_norm": 1.3926502445855002, "learning_rate": 7.101504564013162e-07, "loss": 0.709, "step": 24378 }, { "epoch": 0.8829771821803695, "grad_norm": 1.0350154766368453, "learning_rate": 7.097163474443491e-07, "loss": 0.671, "step": 24379 }, { "epoch": 0.8830134009416878, "grad_norm": 1.042464407310221, "learning_rate": 7.092823663298476e-07, "loss": 0.6353, "step": 24380 }, { "epoch": 0.8830496197030061, "grad_norm": 1.300920811666468, "learning_rate": 7.088485130637835e-07, "loss": 0.6505, "step": 24381 }, { "epoch": 0.8830858384643245, "grad_norm": 1.3429051629392004, "learning_rate": 7.084147876521274e-07, "loss": 0.6981, "step": 24382 }, { "epoch": 0.8831220572256429, "grad_norm": 1.5161722463545515, "learning_rate": 7.079811901008493e-07, "loss": 0.7254, "step": 24383 }, { "epoch": 0.8831582759869613, "grad_norm": 1.3109870749097812, "learning_rate": 7.075477204159142e-07, "loss": 0.5918, "step": 24384 }, { "epoch": 0.8831944947482796, "grad_norm": 1.283930505574744, "learning_rate": 7.071143786032886e-07, "loss": 0.6584, "step": 24385 }, { "epoch": 0.8832307135095979, "grad_norm": 1.4247655993898412, "learning_rate": 7.06681164668932e-07, "loss": 0.603, "step": 24386 }, { "epoch": 0.8832669322709163, "grad_norm": 1.3471087311029293, "learning_rate": 7.062480786188097e-07, "loss": 0.671, "step": 24387 }, { "epoch": 0.8833031510322347, "grad_norm": 1.3709550954532161, "learning_rate": 7.058151204588815e-07, "loss": 0.7025, "step": 24388 }, { "epoch": 0.8833393697935531, "grad_norm": 0.9262303878296055, "learning_rate": 7.053822901951013e-07, "loss": 0.7284, "step": 24389 }, { "epoch": 0.8833755885548714, "grad_norm": 1.2419196074175296, "learning_rate": 7.04949587833429e-07, "loss": 0.713, "step": 24390 }, { "epoch": 0.8834118073161897, "grad_norm": 1.4128325742028656, "learning_rate": 7.045170133798184e-07, "loss": 0.708, "step": 24391 }, { "epoch": 0.8834480260775082, "grad_norm": 1.3444058216524162, "learning_rate": 7.040845668402218e-07, "loss": 0.6465, "step": 24392 }, { "epoch": 0.8834842448388265, "grad_norm": 1.3903944910735928, "learning_rate": 7.036522482205866e-07, "loss": 0.6716, "step": 24393 }, { "epoch": 0.8835204636001449, "grad_norm": 1.390517096672931, "learning_rate": 7.032200575268677e-07, "loss": 0.6566, "step": 24394 }, { "epoch": 0.8835566823614632, "grad_norm": 1.4301572546470378, "learning_rate": 7.027879947650096e-07, "loss": 0.6655, "step": 24395 }, { "epoch": 0.8835929011227815, "grad_norm": 1.3965463484504241, "learning_rate": 7.023560599409574e-07, "loss": 0.6423, "step": 24396 }, { "epoch": 0.8836291198841, "grad_norm": 1.0122172868773107, "learning_rate": 7.01924253060654e-07, "loss": 0.6759, "step": 24397 }, { "epoch": 0.8836653386454183, "grad_norm": 0.9967287427597542, "learning_rate": 7.014925741300449e-07, "loss": 0.6706, "step": 24398 }, { "epoch": 0.8837015574067367, "grad_norm": 0.9910091209665725, "learning_rate": 7.010610231550686e-07, "loss": 0.7093, "step": 24399 }, { "epoch": 0.883737776168055, "grad_norm": 1.0780394705288938, "learning_rate": 7.006296001416613e-07, "loss": 0.7018, "step": 24400 }, { "epoch": 0.8837739949293734, "grad_norm": 1.304853235356906, "learning_rate": 7.001983050957628e-07, "loss": 0.6616, "step": 24401 }, { "epoch": 0.8838102136906918, "grad_norm": 1.0092242526954938, "learning_rate": 6.997671380233084e-07, "loss": 0.6719, "step": 24402 }, { "epoch": 0.8838464324520101, "grad_norm": 1.069062892682506, "learning_rate": 6.993360989302289e-07, "loss": 0.6707, "step": 24403 }, { "epoch": 0.8838826512133285, "grad_norm": 1.254254793288641, "learning_rate": 6.989051878224562e-07, "loss": 0.6655, "step": 24404 }, { "epoch": 0.8839188699746469, "grad_norm": 1.3250483626073053, "learning_rate": 6.984744047059222e-07, "loss": 0.7126, "step": 24405 }, { "epoch": 0.8839550887359652, "grad_norm": 1.0576681304125228, "learning_rate": 6.980437495865533e-07, "loss": 0.6714, "step": 24406 }, { "epoch": 0.8839913074972836, "grad_norm": 1.366583003193017, "learning_rate": 6.976132224702759e-07, "loss": 0.7189, "step": 24407 }, { "epoch": 0.8840275262586019, "grad_norm": 1.6069752152557697, "learning_rate": 6.971828233630129e-07, "loss": 0.6796, "step": 24408 }, { "epoch": 0.8840637450199204, "grad_norm": 1.1763286691999795, "learning_rate": 6.967525522706897e-07, "loss": 0.677, "step": 24409 }, { "epoch": 0.8840999637812387, "grad_norm": 1.4057101723540892, "learning_rate": 6.96322409199226e-07, "loss": 0.6181, "step": 24410 }, { "epoch": 0.884136182542557, "grad_norm": 1.244718870141888, "learning_rate": 6.958923941545393e-07, "loss": 0.6655, "step": 24411 }, { "epoch": 0.8841724013038754, "grad_norm": 0.9299763758628483, "learning_rate": 6.954625071425502e-07, "loss": 0.6386, "step": 24412 }, { "epoch": 0.8842086200651937, "grad_norm": 1.478928143086835, "learning_rate": 6.950327481691732e-07, "loss": 0.692, "step": 24413 }, { "epoch": 0.8842448388265122, "grad_norm": 1.3059920068175122, "learning_rate": 6.946031172403211e-07, "loss": 0.6497, "step": 24414 }, { "epoch": 0.8842810575878305, "grad_norm": 1.0431814790483118, "learning_rate": 6.941736143619048e-07, "loss": 0.6258, "step": 24415 }, { "epoch": 0.8843172763491488, "grad_norm": 1.2461297955277228, "learning_rate": 6.937442395398375e-07, "loss": 0.6613, "step": 24416 }, { "epoch": 0.8843534951104672, "grad_norm": 1.1403876529826897, "learning_rate": 6.933149927800275e-07, "loss": 0.6527, "step": 24417 }, { "epoch": 0.8843897138717856, "grad_norm": 1.3444279849667713, "learning_rate": 6.928858740883803e-07, "loss": 0.6761, "step": 24418 }, { "epoch": 0.884425932633104, "grad_norm": 1.4593072684578718, "learning_rate": 6.924568834708001e-07, "loss": 0.6926, "step": 24419 }, { "epoch": 0.8844621513944223, "grad_norm": 1.4407356111911997, "learning_rate": 6.920280209331931e-07, "loss": 0.721, "step": 24420 }, { "epoch": 0.8844983701557406, "grad_norm": 1.317005328905283, "learning_rate": 6.915992864814602e-07, "loss": 0.6539, "step": 24421 }, { "epoch": 0.884534588917059, "grad_norm": 1.3783147383387027, "learning_rate": 6.911706801214968e-07, "loss": 0.6324, "step": 24422 }, { "epoch": 0.8845708076783774, "grad_norm": 1.403738860678868, "learning_rate": 6.907422018592069e-07, "loss": 0.6902, "step": 24423 }, { "epoch": 0.8846070264396958, "grad_norm": 1.3690775232289687, "learning_rate": 6.903138517004837e-07, "loss": 0.6913, "step": 24424 }, { "epoch": 0.8846432452010141, "grad_norm": 1.4767125074946674, "learning_rate": 6.898856296512225e-07, "loss": 0.7444, "step": 24425 }, { "epoch": 0.8846794639623324, "grad_norm": 1.341008860034586, "learning_rate": 6.894575357173139e-07, "loss": 0.6316, "step": 24426 }, { "epoch": 0.8847156827236509, "grad_norm": 1.3778671707365029, "learning_rate": 6.890295699046524e-07, "loss": 0.6792, "step": 24427 }, { "epoch": 0.8847519014849692, "grad_norm": 1.3660841379227808, "learning_rate": 6.886017322191252e-07, "loss": 0.6522, "step": 24428 }, { "epoch": 0.8847881202462876, "grad_norm": 1.3048526179738966, "learning_rate": 6.881740226666178e-07, "loss": 0.6916, "step": 24429 }, { "epoch": 0.8848243390076059, "grad_norm": 1.4221071548939157, "learning_rate": 6.8774644125302e-07, "loss": 0.7117, "step": 24430 }, { "epoch": 0.8848605577689242, "grad_norm": 1.321046463333271, "learning_rate": 6.873189879842134e-07, "loss": 0.6374, "step": 24431 }, { "epoch": 0.8848967765302427, "grad_norm": 0.9339572696799084, "learning_rate": 6.868916628660815e-07, "loss": 0.6117, "step": 24432 }, { "epoch": 0.884932995291561, "grad_norm": 1.3387360525985241, "learning_rate": 6.864644659045006e-07, "loss": 0.7109, "step": 24433 }, { "epoch": 0.8849692140528794, "grad_norm": 0.9399589522074103, "learning_rate": 6.860373971053535e-07, "loss": 0.623, "step": 24434 }, { "epoch": 0.8850054328141977, "grad_norm": 1.4058155554333438, "learning_rate": 6.856104564745169e-07, "loss": 0.6953, "step": 24435 }, { "epoch": 0.8850416515755161, "grad_norm": 1.0821294560842851, "learning_rate": 6.85183644017865e-07, "loss": 0.6389, "step": 24436 }, { "epoch": 0.8850778703368345, "grad_norm": 0.9482356344653655, "learning_rate": 6.847569597412684e-07, "loss": 0.6661, "step": 24437 }, { "epoch": 0.8851140890981528, "grad_norm": 1.3582945865558684, "learning_rate": 6.843304036506026e-07, "loss": 0.6637, "step": 24438 }, { "epoch": 0.8851503078594712, "grad_norm": 2.2647832951235864, "learning_rate": 6.839039757517363e-07, "loss": 0.6874, "step": 24439 }, { "epoch": 0.8851865266207896, "grad_norm": 1.5361304164899316, "learning_rate": 6.834776760505357e-07, "loss": 0.6616, "step": 24440 }, { "epoch": 0.8852227453821079, "grad_norm": 1.3394664811575083, "learning_rate": 6.830515045528696e-07, "loss": 0.6788, "step": 24441 }, { "epoch": 0.8852589641434263, "grad_norm": 1.2746948941248482, "learning_rate": 6.826254612646011e-07, "loss": 0.6878, "step": 24442 }, { "epoch": 0.8852951829047446, "grad_norm": 1.4784782176910667, "learning_rate": 6.821995461915942e-07, "loss": 0.7401, "step": 24443 }, { "epoch": 0.8853314016660631, "grad_norm": 1.032657272118413, "learning_rate": 6.817737593397067e-07, "loss": 0.6872, "step": 24444 }, { "epoch": 0.8853676204273814, "grad_norm": 1.3016822010296827, "learning_rate": 6.813481007148015e-07, "loss": 0.7052, "step": 24445 }, { "epoch": 0.8854038391886997, "grad_norm": 1.4583811930199773, "learning_rate": 6.809225703227352e-07, "loss": 0.7578, "step": 24446 }, { "epoch": 0.8854400579500181, "grad_norm": 1.3694006595379504, "learning_rate": 6.804971681693618e-07, "loss": 0.7237, "step": 24447 }, { "epoch": 0.8854762767113364, "grad_norm": 1.2257606974471837, "learning_rate": 6.800718942605355e-07, "loss": 0.6541, "step": 24448 }, { "epoch": 0.8855124954726549, "grad_norm": 1.481755267605092, "learning_rate": 6.796467486021108e-07, "loss": 0.6964, "step": 24449 }, { "epoch": 0.8855487142339732, "grad_norm": 1.2726116596170833, "learning_rate": 6.79221731199936e-07, "loss": 0.7028, "step": 24450 }, { "epoch": 0.8855849329952915, "grad_norm": 1.3869268999900994, "learning_rate": 6.787968420598579e-07, "loss": 0.706, "step": 24451 }, { "epoch": 0.8856211517566099, "grad_norm": 1.3224999939922124, "learning_rate": 6.783720811877281e-07, "loss": 0.6928, "step": 24452 }, { "epoch": 0.8856573705179283, "grad_norm": 1.3832467341611567, "learning_rate": 6.779474485893889e-07, "loss": 0.6278, "step": 24453 }, { "epoch": 0.8856935892792467, "grad_norm": 1.3738441363853153, "learning_rate": 6.775229442706843e-07, "loss": 0.7019, "step": 24454 }, { "epoch": 0.885729808040565, "grad_norm": 1.0644353261386972, "learning_rate": 6.770985682374531e-07, "loss": 0.6695, "step": 24455 }, { "epoch": 0.8857660268018833, "grad_norm": 1.371565781684134, "learning_rate": 6.766743204955395e-07, "loss": 0.706, "step": 24456 }, { "epoch": 0.8858022455632018, "grad_norm": 1.5183587159414704, "learning_rate": 6.762502010507788e-07, "loss": 0.6626, "step": 24457 }, { "epoch": 0.8858384643245201, "grad_norm": 1.332123989501363, "learning_rate": 6.758262099090085e-07, "loss": 0.608, "step": 24458 }, { "epoch": 0.8858746830858385, "grad_norm": 1.3237051783741576, "learning_rate": 6.754023470760607e-07, "loss": 0.67, "step": 24459 }, { "epoch": 0.8859109018471568, "grad_norm": 1.0555198430979684, "learning_rate": 6.749786125577717e-07, "loss": 0.6257, "step": 24460 }, { "epoch": 0.8859471206084751, "grad_norm": 1.4762253805347672, "learning_rate": 6.745550063599704e-07, "loss": 0.6773, "step": 24461 }, { "epoch": 0.8859833393697936, "grad_norm": 1.034432344759467, "learning_rate": 6.741315284884842e-07, "loss": 0.6941, "step": 24462 }, { "epoch": 0.8860195581311119, "grad_norm": 1.4187535210809412, "learning_rate": 6.737081789491451e-07, "loss": 0.6756, "step": 24463 }, { "epoch": 0.8860557768924303, "grad_norm": 1.4534742950739172, "learning_rate": 6.732849577477763e-07, "loss": 0.6505, "step": 24464 }, { "epoch": 0.8860919956537486, "grad_norm": 1.1792835205461725, "learning_rate": 6.728618648902008e-07, "loss": 0.5975, "step": 24465 }, { "epoch": 0.886128214415067, "grad_norm": 1.0367631058317328, "learning_rate": 6.724389003822406e-07, "loss": 0.6633, "step": 24466 }, { "epoch": 0.8861644331763854, "grad_norm": 1.2641087742983428, "learning_rate": 6.72016064229718e-07, "loss": 0.6473, "step": 24467 }, { "epoch": 0.8862006519377037, "grad_norm": 1.3822116253487613, "learning_rate": 6.715933564384514e-07, "loss": 0.6353, "step": 24468 }, { "epoch": 0.8862368706990221, "grad_norm": 1.372809979092163, "learning_rate": 6.711707770142573e-07, "loss": 0.7181, "step": 24469 }, { "epoch": 0.8862730894603404, "grad_norm": 1.4055632218898835, "learning_rate": 6.707483259629477e-07, "loss": 0.7203, "step": 24470 }, { "epoch": 0.8863093082216588, "grad_norm": 1.0154058187819475, "learning_rate": 6.703260032903403e-07, "loss": 0.6006, "step": 24471 }, { "epoch": 0.8863455269829772, "grad_norm": 1.404658964586498, "learning_rate": 6.699038090022447e-07, "loss": 0.6991, "step": 24472 }, { "epoch": 0.8863817457442955, "grad_norm": 1.5125162500107536, "learning_rate": 6.694817431044698e-07, "loss": 0.7319, "step": 24473 }, { "epoch": 0.8864179645056139, "grad_norm": 1.3360395427403484, "learning_rate": 6.690598056028263e-07, "loss": 0.6972, "step": 24474 }, { "epoch": 0.8864541832669323, "grad_norm": 1.4270877234769617, "learning_rate": 6.686379965031186e-07, "loss": 0.7014, "step": 24475 }, { "epoch": 0.8864904020282506, "grad_norm": 1.4740707815959457, "learning_rate": 6.68216315811151e-07, "loss": 0.6817, "step": 24476 }, { "epoch": 0.886526620789569, "grad_norm": 1.304310327964927, "learning_rate": 6.677947635327253e-07, "loss": 0.6422, "step": 24477 }, { "epoch": 0.8865628395508873, "grad_norm": 1.4315767007220017, "learning_rate": 6.673733396736448e-07, "loss": 0.7352, "step": 24478 }, { "epoch": 0.8865990583122058, "grad_norm": 1.3484644560313135, "learning_rate": 6.669520442397081e-07, "loss": 0.6111, "step": 24479 }, { "epoch": 0.8866352770735241, "grad_norm": 1.3254162935154299, "learning_rate": 6.665308772367118e-07, "loss": 0.6438, "step": 24480 }, { "epoch": 0.8866714958348424, "grad_norm": 1.5231847840777166, "learning_rate": 6.66109838670449e-07, "loss": 0.7088, "step": 24481 }, { "epoch": 0.8867077145961608, "grad_norm": 1.5462151839693303, "learning_rate": 6.656889285467194e-07, "loss": 0.6304, "step": 24482 }, { "epoch": 0.8867439333574791, "grad_norm": 1.397114808157822, "learning_rate": 6.652681468713118e-07, "loss": 0.6834, "step": 24483 }, { "epoch": 0.8867801521187976, "grad_norm": 1.295938762132266, "learning_rate": 6.648474936500149e-07, "loss": 0.6397, "step": 24484 }, { "epoch": 0.8868163708801159, "grad_norm": 1.4131453500950164, "learning_rate": 6.644269688886196e-07, "loss": 0.6656, "step": 24485 }, { "epoch": 0.8868525896414342, "grad_norm": 1.5815520137611838, "learning_rate": 6.640065725929135e-07, "loss": 0.6455, "step": 24486 }, { "epoch": 0.8868888084027526, "grad_norm": 1.3415028161019953, "learning_rate": 6.635863047686797e-07, "loss": 0.7001, "step": 24487 }, { "epoch": 0.886925027164071, "grad_norm": 1.4777771820293577, "learning_rate": 6.631661654217014e-07, "loss": 0.7326, "step": 24488 }, { "epoch": 0.8869612459253894, "grad_norm": 1.5948841147959492, "learning_rate": 6.627461545577607e-07, "loss": 0.7511, "step": 24489 }, { "epoch": 0.8869974646867077, "grad_norm": 1.5396891325184565, "learning_rate": 6.623262721826384e-07, "loss": 0.712, "step": 24490 }, { "epoch": 0.887033683448026, "grad_norm": 1.410858893950729, "learning_rate": 6.619065183021101e-07, "loss": 0.6677, "step": 24491 }, { "epoch": 0.8870699022093445, "grad_norm": 1.5672519046864972, "learning_rate": 6.614868929219542e-07, "loss": 0.6202, "step": 24492 }, { "epoch": 0.8871061209706628, "grad_norm": 1.4547022065106607, "learning_rate": 6.610673960479441e-07, "loss": 0.6534, "step": 24493 }, { "epoch": 0.8871423397319812, "grad_norm": 1.245048394546714, "learning_rate": 6.606480276858529e-07, "loss": 0.7305, "step": 24494 }, { "epoch": 0.8871785584932995, "grad_norm": 1.415696734332987, "learning_rate": 6.602287878414493e-07, "loss": 0.6736, "step": 24495 }, { "epoch": 0.8872147772546178, "grad_norm": 1.5572825942418447, "learning_rate": 6.598096765205065e-07, "loss": 0.6694, "step": 24496 }, { "epoch": 0.8872509960159363, "grad_norm": 1.4421961346530012, "learning_rate": 6.593906937287886e-07, "loss": 0.72, "step": 24497 }, { "epoch": 0.8872872147772546, "grad_norm": 1.3181685514454184, "learning_rate": 6.589718394720623e-07, "loss": 0.6098, "step": 24498 }, { "epoch": 0.887323433538573, "grad_norm": 1.0066639658990406, "learning_rate": 6.585531137560897e-07, "loss": 0.6668, "step": 24499 }, { "epoch": 0.8873596522998913, "grad_norm": 1.4355957554909669, "learning_rate": 6.581345165866371e-07, "loss": 0.7053, "step": 24500 }, { "epoch": 0.8873958710612097, "grad_norm": 1.8399028890058675, "learning_rate": 6.5771604796946e-07, "loss": 0.6852, "step": 24501 }, { "epoch": 0.8874320898225281, "grad_norm": 1.4081754298138767, "learning_rate": 6.572977079103193e-07, "loss": 0.6886, "step": 24502 }, { "epoch": 0.8874683085838464, "grad_norm": 0.9967636469044495, "learning_rate": 6.568794964149716e-07, "loss": 0.6363, "step": 24503 }, { "epoch": 0.8875045273451648, "grad_norm": 1.509375671125938, "learning_rate": 6.564614134891722e-07, "loss": 0.6579, "step": 24504 }, { "epoch": 0.8875407461064831, "grad_norm": 1.2903008870308763, "learning_rate": 6.560434591386733e-07, "loss": 0.6805, "step": 24505 }, { "epoch": 0.8875769648678016, "grad_norm": 1.262132042321102, "learning_rate": 6.556256333692257e-07, "loss": 0.6315, "step": 24506 }, { "epoch": 0.8876131836291199, "grad_norm": 1.4087006884290565, "learning_rate": 6.552079361865815e-07, "loss": 0.6076, "step": 24507 }, { "epoch": 0.8876494023904382, "grad_norm": 1.4240642353147324, "learning_rate": 6.547903675964873e-07, "loss": 0.7111, "step": 24508 }, { "epoch": 0.8876856211517566, "grad_norm": 1.454023229719247, "learning_rate": 6.543729276046884e-07, "loss": 0.6761, "step": 24509 }, { "epoch": 0.887721839913075, "grad_norm": 1.4558818504762363, "learning_rate": 6.53955616216928e-07, "loss": 0.684, "step": 24510 }, { "epoch": 0.8877580586743934, "grad_norm": 1.4311191895990705, "learning_rate": 6.535384334389527e-07, "loss": 0.6439, "step": 24511 }, { "epoch": 0.8877942774357117, "grad_norm": 1.339047795390801, "learning_rate": 6.531213792765012e-07, "loss": 0.697, "step": 24512 }, { "epoch": 0.88783049619703, "grad_norm": 1.416161523077098, "learning_rate": 6.527044537353111e-07, "loss": 0.7286, "step": 24513 }, { "epoch": 0.8878667149583485, "grad_norm": 1.3560163988380534, "learning_rate": 6.522876568211223e-07, "loss": 0.6527, "step": 24514 }, { "epoch": 0.8879029337196668, "grad_norm": 1.5001468363060178, "learning_rate": 6.518709885396679e-07, "loss": 0.6858, "step": 24515 }, { "epoch": 0.8879391524809852, "grad_norm": 1.0852446453403382, "learning_rate": 6.514544488966845e-07, "loss": 0.6208, "step": 24516 }, { "epoch": 0.8879753712423035, "grad_norm": 1.4176963887428544, "learning_rate": 6.510380378978987e-07, "loss": 0.6932, "step": 24517 }, { "epoch": 0.8880115900036218, "grad_norm": 1.4115598712155175, "learning_rate": 6.506217555490469e-07, "loss": 0.6287, "step": 24518 }, { "epoch": 0.8880478087649403, "grad_norm": 1.4761372160647137, "learning_rate": 6.502056018558556e-07, "loss": 0.7061, "step": 24519 }, { "epoch": 0.8880840275262586, "grad_norm": 1.5982033496992099, "learning_rate": 6.497895768240492e-07, "loss": 0.6827, "step": 24520 }, { "epoch": 0.888120246287577, "grad_norm": 1.398647395157699, "learning_rate": 6.493736804593531e-07, "loss": 0.582, "step": 24521 }, { "epoch": 0.8881564650488953, "grad_norm": 1.6117722656344553, "learning_rate": 6.489579127674928e-07, "loss": 0.6032, "step": 24522 }, { "epoch": 0.8881926838102137, "grad_norm": 1.354697962685323, "learning_rate": 6.48542273754188e-07, "loss": 0.6459, "step": 24523 }, { "epoch": 0.8882289025715321, "grad_norm": 1.4074797187509809, "learning_rate": 6.481267634251576e-07, "loss": 0.7177, "step": 24524 }, { "epoch": 0.8882651213328504, "grad_norm": 1.4731187283965477, "learning_rate": 6.477113817861214e-07, "loss": 0.6738, "step": 24525 }, { "epoch": 0.8883013400941688, "grad_norm": 1.4220570569546378, "learning_rate": 6.472961288427937e-07, "loss": 0.6792, "step": 24526 }, { "epoch": 0.8883375588554872, "grad_norm": 1.2751532300098372, "learning_rate": 6.4688100460089e-07, "loss": 0.6771, "step": 24527 }, { "epoch": 0.8883737776168055, "grad_norm": 1.4384917059656122, "learning_rate": 6.464660090661202e-07, "loss": 0.6265, "step": 24528 }, { "epoch": 0.8884099963781239, "grad_norm": 1.272474651160305, "learning_rate": 6.460511422441984e-07, "loss": 0.7133, "step": 24529 }, { "epoch": 0.8884462151394422, "grad_norm": 1.4461394953228766, "learning_rate": 6.456364041408314e-07, "loss": 0.6473, "step": 24530 }, { "epoch": 0.8884824339007606, "grad_norm": 1.0937542405429446, "learning_rate": 6.452217947617279e-07, "loss": 0.6746, "step": 24531 }, { "epoch": 0.888518652662079, "grad_norm": 1.392506005692201, "learning_rate": 6.448073141125899e-07, "loss": 0.6325, "step": 24532 }, { "epoch": 0.8885548714233973, "grad_norm": 1.30174750662385, "learning_rate": 6.443929621991252e-07, "loss": 0.6121, "step": 24533 }, { "epoch": 0.8885910901847157, "grad_norm": 1.0829173844476616, "learning_rate": 6.439787390270347e-07, "loss": 0.6149, "step": 24534 }, { "epoch": 0.888627308946034, "grad_norm": 0.9507720387346331, "learning_rate": 6.435646446020149e-07, "loss": 0.6775, "step": 24535 }, { "epoch": 0.8886635277073525, "grad_norm": 1.3414156197202345, "learning_rate": 6.431506789297693e-07, "loss": 0.7213, "step": 24536 }, { "epoch": 0.8886997464686708, "grad_norm": 1.348266862668055, "learning_rate": 6.427368420159918e-07, "loss": 0.6562, "step": 24537 }, { "epoch": 0.8887359652299891, "grad_norm": 1.4216845359934438, "learning_rate": 6.423231338663783e-07, "loss": 0.6111, "step": 24538 }, { "epoch": 0.8887721839913075, "grad_norm": 1.2698947195448689, "learning_rate": 6.419095544866183e-07, "loss": 0.6021, "step": 24539 }, { "epoch": 0.8888084027526258, "grad_norm": 1.372308958076859, "learning_rate": 6.414961038824064e-07, "loss": 0.7316, "step": 24540 }, { "epoch": 0.8888446215139443, "grad_norm": 1.288922269736411, "learning_rate": 6.410827820594323e-07, "loss": 0.6726, "step": 24541 }, { "epoch": 0.8888808402752626, "grad_norm": 1.4803991527856304, "learning_rate": 6.406695890233827e-07, "loss": 0.7549, "step": 24542 }, { "epoch": 0.8889170590365809, "grad_norm": 1.3194412186615256, "learning_rate": 6.402565247799408e-07, "loss": 0.689, "step": 24543 }, { "epoch": 0.8889532777978993, "grad_norm": 1.3553236724321773, "learning_rate": 6.398435893347954e-07, "loss": 0.6601, "step": 24544 }, { "epoch": 0.8889894965592177, "grad_norm": 1.2415425835211464, "learning_rate": 6.394307826936275e-07, "loss": 0.6398, "step": 24545 }, { "epoch": 0.8890257153205361, "grad_norm": 1.3945695700843417, "learning_rate": 6.390181048621147e-07, "loss": 0.648, "step": 24546 }, { "epoch": 0.8890619340818544, "grad_norm": 0.9956025189149756, "learning_rate": 6.386055558459392e-07, "loss": 0.6169, "step": 24547 }, { "epoch": 0.8890981528431727, "grad_norm": 1.3250196251900306, "learning_rate": 6.381931356507776e-07, "loss": 0.6723, "step": 24548 }, { "epoch": 0.8891343716044912, "grad_norm": 1.323542154806319, "learning_rate": 6.377808442823041e-07, "loss": 0.6403, "step": 24549 }, { "epoch": 0.8891705903658095, "grad_norm": 1.533825463742625, "learning_rate": 6.373686817461911e-07, "loss": 0.6556, "step": 24550 }, { "epoch": 0.8892068091271279, "grad_norm": 1.3265810378838363, "learning_rate": 6.369566480481138e-07, "loss": 0.6241, "step": 24551 }, { "epoch": 0.8892430278884462, "grad_norm": 1.1870819371723549, "learning_rate": 6.365447431937388e-07, "loss": 0.665, "step": 24552 }, { "epoch": 0.8892792466497645, "grad_norm": 1.0020345432633946, "learning_rate": 6.361329671887373e-07, "loss": 0.6966, "step": 24553 }, { "epoch": 0.889315465411083, "grad_norm": 1.424210056313971, "learning_rate": 6.357213200387724e-07, "loss": 0.7066, "step": 24554 }, { "epoch": 0.8893516841724013, "grad_norm": 1.1220847617867398, "learning_rate": 6.353098017495119e-07, "loss": 0.6792, "step": 24555 }, { "epoch": 0.8893879029337197, "grad_norm": 1.3656003765673113, "learning_rate": 6.348984123266144e-07, "loss": 0.6269, "step": 24556 }, { "epoch": 0.889424121695038, "grad_norm": 1.1062032057451516, "learning_rate": 6.344871517757456e-07, "loss": 0.7223, "step": 24557 }, { "epoch": 0.8894603404563564, "grad_norm": 1.2628176188111486, "learning_rate": 6.34076020102562e-07, "loss": 0.5882, "step": 24558 }, { "epoch": 0.8894965592176748, "grad_norm": 1.0758119608862429, "learning_rate": 6.336650173127224e-07, "loss": 0.6646, "step": 24559 }, { "epoch": 0.8895327779789931, "grad_norm": 1.3181685987827723, "learning_rate": 6.33254143411881e-07, "loss": 0.5963, "step": 24560 }, { "epoch": 0.8895689967403115, "grad_norm": 1.3718666794379042, "learning_rate": 6.328433984056948e-07, "loss": 0.6794, "step": 24561 }, { "epoch": 0.8896052155016299, "grad_norm": 1.3709556885028604, "learning_rate": 6.324327822998133e-07, "loss": 0.712, "step": 24562 }, { "epoch": 0.8896414342629482, "grad_norm": 1.0970692051380977, "learning_rate": 6.320222950998889e-07, "loss": 0.6957, "step": 24563 }, { "epoch": 0.8896776530242666, "grad_norm": 1.412145366183657, "learning_rate": 6.316119368115659e-07, "loss": 0.6846, "step": 24564 }, { "epoch": 0.8897138717855849, "grad_norm": 1.310102749317477, "learning_rate": 6.312017074404975e-07, "loss": 0.6496, "step": 24565 }, { "epoch": 0.8897500905469033, "grad_norm": 1.4128329631238556, "learning_rate": 6.307916069923248e-07, "loss": 0.627, "step": 24566 }, { "epoch": 0.8897863093082217, "grad_norm": 1.385903540607006, "learning_rate": 6.303816354726911e-07, "loss": 0.6718, "step": 24567 }, { "epoch": 0.88982252806954, "grad_norm": 1.1721016927344965, "learning_rate": 6.299717928872406e-07, "loss": 0.6873, "step": 24568 }, { "epoch": 0.8898587468308584, "grad_norm": 0.9999664739207906, "learning_rate": 6.295620792416112e-07, "loss": 0.6953, "step": 24569 }, { "epoch": 0.8898949655921767, "grad_norm": 1.3112325196560166, "learning_rate": 6.291524945414407e-07, "loss": 0.7236, "step": 24570 }, { "epoch": 0.8899311843534952, "grad_norm": 1.4507806990051866, "learning_rate": 6.287430387923654e-07, "loss": 0.6052, "step": 24571 }, { "epoch": 0.8899674031148135, "grad_norm": 1.1835631785205742, "learning_rate": 6.28333712000021e-07, "loss": 0.5733, "step": 24572 }, { "epoch": 0.8900036218761318, "grad_norm": 1.315813579205191, "learning_rate": 6.279245141700397e-07, "loss": 0.6353, "step": 24573 }, { "epoch": 0.8900398406374502, "grad_norm": 1.0882827373809056, "learning_rate": 6.275154453080512e-07, "loss": 0.6749, "step": 24574 }, { "epoch": 0.8900760593987685, "grad_norm": 1.3101238372401036, "learning_rate": 6.271065054196846e-07, "loss": 0.6289, "step": 24575 }, { "epoch": 0.890112278160087, "grad_norm": 1.4619336238411123, "learning_rate": 6.266976945105707e-07, "loss": 0.7111, "step": 24576 }, { "epoch": 0.8901484969214053, "grad_norm": 1.4147849029107273, "learning_rate": 6.262890125863319e-07, "loss": 0.7459, "step": 24577 }, { "epoch": 0.8901847156827236, "grad_norm": 1.5426015115058158, "learning_rate": 6.258804596525902e-07, "loss": 0.7013, "step": 24578 }, { "epoch": 0.890220934444042, "grad_norm": 1.2863912576172214, "learning_rate": 6.254720357149735e-07, "loss": 0.7692, "step": 24579 }, { "epoch": 0.8902571532053604, "grad_norm": 1.5111370347432296, "learning_rate": 6.250637407790971e-07, "loss": 0.6644, "step": 24580 }, { "epoch": 0.8902933719666788, "grad_norm": 1.0750455005677158, "learning_rate": 6.246555748505822e-07, "loss": 0.6799, "step": 24581 }, { "epoch": 0.8903295907279971, "grad_norm": 1.289055127787461, "learning_rate": 6.242475379350432e-07, "loss": 0.5532, "step": 24582 }, { "epoch": 0.8903658094893154, "grad_norm": 1.1041778912841549, "learning_rate": 6.238396300380978e-07, "loss": 0.6324, "step": 24583 }, { "epoch": 0.8904020282506339, "grad_norm": 1.748703659532954, "learning_rate": 6.23431851165357e-07, "loss": 0.6599, "step": 24584 }, { "epoch": 0.8904382470119522, "grad_norm": 1.0770285241128006, "learning_rate": 6.230242013224331e-07, "loss": 0.6385, "step": 24585 }, { "epoch": 0.8904744657732706, "grad_norm": 1.3817318705270472, "learning_rate": 6.226166805149348e-07, "loss": 0.6923, "step": 24586 }, { "epoch": 0.8905106845345889, "grad_norm": 1.2826030676691753, "learning_rate": 6.22209288748471e-07, "loss": 0.6938, "step": 24587 }, { "epoch": 0.8905469032959072, "grad_norm": 1.1466392783779522, "learning_rate": 6.218020260286484e-07, "loss": 0.6233, "step": 24588 }, { "epoch": 0.8905831220572257, "grad_norm": 1.104267789106309, "learning_rate": 6.21394892361068e-07, "loss": 0.6625, "step": 24589 }, { "epoch": 0.890619340818544, "grad_norm": 1.3332502039932992, "learning_rate": 6.209878877513364e-07, "loss": 0.6599, "step": 24590 }, { "epoch": 0.8906555595798624, "grad_norm": 1.3088227769078324, "learning_rate": 6.205810122050526e-07, "loss": 0.5651, "step": 24591 }, { "epoch": 0.8906917783411807, "grad_norm": 1.4120270453125898, "learning_rate": 6.201742657278153e-07, "loss": 0.6344, "step": 24592 }, { "epoch": 0.8907279971024991, "grad_norm": 1.0042711337678802, "learning_rate": 6.197676483252213e-07, "loss": 0.7072, "step": 24593 }, { "epoch": 0.8907642158638175, "grad_norm": 1.4096262363748253, "learning_rate": 6.193611600028681e-07, "loss": 0.6907, "step": 24594 }, { "epoch": 0.8908004346251358, "grad_norm": 1.1923296355258755, "learning_rate": 6.189548007663471e-07, "loss": 0.68, "step": 24595 }, { "epoch": 0.8908366533864542, "grad_norm": 1.3060372467667778, "learning_rate": 6.185485706212502e-07, "loss": 0.6467, "step": 24596 }, { "epoch": 0.8908728721477726, "grad_norm": 1.3106144462709908, "learning_rate": 6.181424695731686e-07, "loss": 0.6441, "step": 24597 }, { "epoch": 0.8909090909090909, "grad_norm": 1.3759868484719064, "learning_rate": 6.177364976276912e-07, "loss": 0.6578, "step": 24598 }, { "epoch": 0.8909453096704093, "grad_norm": 1.3138612928239841, "learning_rate": 6.173306547904034e-07, "loss": 0.6132, "step": 24599 }, { "epoch": 0.8909815284317276, "grad_norm": 1.3394488129333615, "learning_rate": 6.169249410668887e-07, "loss": 0.6323, "step": 24600 }, { "epoch": 0.891017747193046, "grad_norm": 0.9468554065702516, "learning_rate": 6.165193564627326e-07, "loss": 0.6345, "step": 24601 }, { "epoch": 0.8910539659543644, "grad_norm": 1.378747297370368, "learning_rate": 6.161139009835159e-07, "loss": 0.6213, "step": 24602 }, { "epoch": 0.8910901847156827, "grad_norm": 1.2739284576762204, "learning_rate": 6.157085746348169e-07, "loss": 0.6149, "step": 24603 }, { "epoch": 0.8911264034770011, "grad_norm": 1.464873155934501, "learning_rate": 6.153033774222117e-07, "loss": 0.6694, "step": 24604 }, { "epoch": 0.8911626222383194, "grad_norm": 1.295956087878294, "learning_rate": 6.148983093512806e-07, "loss": 0.6537, "step": 24605 }, { "epoch": 0.8911988409996379, "grad_norm": 1.4770152089272799, "learning_rate": 6.144933704275946e-07, "loss": 0.7089, "step": 24606 }, { "epoch": 0.8912350597609562, "grad_norm": 1.1527272918660327, "learning_rate": 6.140885606567248e-07, "loss": 0.6261, "step": 24607 }, { "epoch": 0.8912712785222745, "grad_norm": 1.468883115869086, "learning_rate": 6.136838800442457e-07, "loss": 0.6816, "step": 24608 }, { "epoch": 0.8913074972835929, "grad_norm": 0.9692671250248129, "learning_rate": 6.132793285957239e-07, "loss": 0.6979, "step": 24609 }, { "epoch": 0.8913437160449112, "grad_norm": 1.1127391916640617, "learning_rate": 6.128749063167272e-07, "loss": 0.6629, "step": 24610 }, { "epoch": 0.8913799348062297, "grad_norm": 1.4254829826342745, "learning_rate": 6.124706132128177e-07, "loss": 0.6053, "step": 24611 }, { "epoch": 0.891416153567548, "grad_norm": 1.5929943220406813, "learning_rate": 6.120664492895622e-07, "loss": 0.6866, "step": 24612 }, { "epoch": 0.8914523723288663, "grad_norm": 1.0106487716476555, "learning_rate": 6.116624145525218e-07, "loss": 0.6489, "step": 24613 }, { "epoch": 0.8914885910901847, "grad_norm": 1.263034134889804, "learning_rate": 6.112585090072565e-07, "loss": 0.6756, "step": 24614 }, { "epoch": 0.8915248098515031, "grad_norm": 1.6345225965910612, "learning_rate": 6.108547326593206e-07, "loss": 0.6594, "step": 24615 }, { "epoch": 0.8915610286128215, "grad_norm": 1.496558678648621, "learning_rate": 6.104510855142764e-07, "loss": 0.6538, "step": 24616 }, { "epoch": 0.8915972473741398, "grad_norm": 1.165449498383223, "learning_rate": 6.100475675776762e-07, "loss": 0.7009, "step": 24617 }, { "epoch": 0.8916334661354581, "grad_norm": 1.3409917545388688, "learning_rate": 6.0964417885507e-07, "loss": 0.6563, "step": 24618 }, { "epoch": 0.8916696848967766, "grad_norm": 1.0343898401738154, "learning_rate": 6.092409193520121e-07, "loss": 0.6828, "step": 24619 }, { "epoch": 0.8917059036580949, "grad_norm": 1.040588836696671, "learning_rate": 6.088377890740504e-07, "loss": 0.696, "step": 24620 }, { "epoch": 0.8917421224194133, "grad_norm": 1.3619823511996945, "learning_rate": 6.084347880267338e-07, "loss": 0.5843, "step": 24621 }, { "epoch": 0.8917783411807316, "grad_norm": 1.4166731182704138, "learning_rate": 6.080319162156045e-07, "loss": 0.6539, "step": 24622 }, { "epoch": 0.89181455994205, "grad_norm": 1.4445612914085655, "learning_rate": 6.076291736462092e-07, "loss": 0.6755, "step": 24623 }, { "epoch": 0.8918507787033684, "grad_norm": 1.1018105637738105, "learning_rate": 6.072265603240901e-07, "loss": 0.5779, "step": 24624 }, { "epoch": 0.8918869974646867, "grad_norm": 1.1651092847975162, "learning_rate": 6.068240762547872e-07, "loss": 0.6704, "step": 24625 }, { "epoch": 0.8919232162260051, "grad_norm": 1.3906541262748209, "learning_rate": 6.064217214438361e-07, "loss": 0.6852, "step": 24626 }, { "epoch": 0.8919594349873234, "grad_norm": 1.323191354626825, "learning_rate": 6.06019495896778e-07, "loss": 0.6569, "step": 24627 }, { "epoch": 0.8919956537486418, "grad_norm": 1.4058756363117189, "learning_rate": 6.056173996191461e-07, "loss": 0.652, "step": 24628 }, { "epoch": 0.8920318725099602, "grad_norm": 1.6338653706243567, "learning_rate": 6.052154326164717e-07, "loss": 0.6582, "step": 24629 }, { "epoch": 0.8920680912712785, "grad_norm": 1.8484268234924872, "learning_rate": 6.048135948942902e-07, "loss": 0.6936, "step": 24630 }, { "epoch": 0.8921043100325969, "grad_norm": 1.3838257391032591, "learning_rate": 6.044118864581284e-07, "loss": 0.6498, "step": 24631 }, { "epoch": 0.8921405287939153, "grad_norm": 1.3725886428132998, "learning_rate": 6.04010307313514e-07, "loss": 0.5994, "step": 24632 }, { "epoch": 0.8921767475552336, "grad_norm": 0.9531424785353965, "learning_rate": 6.036088574659726e-07, "loss": 0.6732, "step": 24633 }, { "epoch": 0.892212966316552, "grad_norm": 1.504266987921972, "learning_rate": 6.032075369210322e-07, "loss": 0.5563, "step": 24634 }, { "epoch": 0.8922491850778703, "grad_norm": 1.3593965623230704, "learning_rate": 6.028063456842115e-07, "loss": 0.6227, "step": 24635 }, { "epoch": 0.8922854038391888, "grad_norm": 1.5054666313766198, "learning_rate": 6.02405283761034e-07, "loss": 0.7209, "step": 24636 }, { "epoch": 0.8923216226005071, "grad_norm": 1.005079873531891, "learning_rate": 6.02004351157014e-07, "loss": 0.6471, "step": 24637 }, { "epoch": 0.8923578413618254, "grad_norm": 1.371974712312993, "learning_rate": 6.016035478776749e-07, "loss": 0.7444, "step": 24638 }, { "epoch": 0.8923940601231438, "grad_norm": 1.4403082185380747, "learning_rate": 6.01202873928528e-07, "loss": 0.637, "step": 24639 }, { "epoch": 0.8924302788844621, "grad_norm": 1.3668808715839849, "learning_rate": 6.008023293150878e-07, "loss": 0.6826, "step": 24640 }, { "epoch": 0.8924664976457806, "grad_norm": 1.0189597871446758, "learning_rate": 6.004019140428674e-07, "loss": 0.6904, "step": 24641 }, { "epoch": 0.8925027164070989, "grad_norm": 1.4247356147050767, "learning_rate": 6.00001628117376e-07, "loss": 0.6902, "step": 24642 }, { "epoch": 0.8925389351684172, "grad_norm": 1.3605255568852428, "learning_rate": 5.996014715441211e-07, "loss": 0.6591, "step": 24643 }, { "epoch": 0.8925751539297356, "grad_norm": 1.1134532311470404, "learning_rate": 5.992014443286087e-07, "loss": 0.6923, "step": 24644 }, { "epoch": 0.892611372691054, "grad_norm": 1.3943159307147133, "learning_rate": 5.988015464763463e-07, "loss": 0.7123, "step": 24645 }, { "epoch": 0.8926475914523724, "grad_norm": 1.4204938135463034, "learning_rate": 5.984017779928353e-07, "loss": 0.6647, "step": 24646 }, { "epoch": 0.8926838102136907, "grad_norm": 1.387372906283232, "learning_rate": 5.980021388835767e-07, "loss": 0.6419, "step": 24647 }, { "epoch": 0.892720028975009, "grad_norm": 1.3508269108125504, "learning_rate": 5.976026291540682e-07, "loss": 0.6581, "step": 24648 }, { "epoch": 0.8927562477363274, "grad_norm": 1.2550798268904724, "learning_rate": 5.972032488098111e-07, "loss": 0.6379, "step": 24649 }, { "epoch": 0.8927924664976458, "grad_norm": 1.0996760948969204, "learning_rate": 5.968039978563001e-07, "loss": 0.6978, "step": 24650 }, { "epoch": 0.8928286852589642, "grad_norm": 1.285620333121123, "learning_rate": 5.96404876299026e-07, "loss": 0.6625, "step": 24651 }, { "epoch": 0.8928649040202825, "grad_norm": 1.4772297307844033, "learning_rate": 5.960058841434846e-07, "loss": 0.6684, "step": 24652 }, { "epoch": 0.8929011227816008, "grad_norm": 0.9735388949936407, "learning_rate": 5.956070213951659e-07, "loss": 0.6551, "step": 24653 }, { "epoch": 0.8929373415429193, "grad_norm": 0.8802106088729041, "learning_rate": 5.952082880595578e-07, "loss": 0.6763, "step": 24654 }, { "epoch": 0.8929735603042376, "grad_norm": 1.083732877479911, "learning_rate": 5.948096841421458e-07, "loss": 0.6705, "step": 24655 }, { "epoch": 0.893009779065556, "grad_norm": 1.0520361815140835, "learning_rate": 5.944112096484189e-07, "loss": 0.6777, "step": 24656 }, { "epoch": 0.8930459978268743, "grad_norm": 1.0185925916913716, "learning_rate": 5.940128645838572e-07, "loss": 0.7048, "step": 24657 }, { "epoch": 0.8930822165881926, "grad_norm": 1.3966388190169718, "learning_rate": 5.936146489539429e-07, "loss": 0.6843, "step": 24658 }, { "epoch": 0.8931184353495111, "grad_norm": 1.0594371579593957, "learning_rate": 5.932165627641573e-07, "loss": 0.6974, "step": 24659 }, { "epoch": 0.8931546541108294, "grad_norm": 0.9038181837032709, "learning_rate": 5.92818606019977e-07, "loss": 0.7039, "step": 24660 }, { "epoch": 0.8931908728721478, "grad_norm": 1.328899696850864, "learning_rate": 5.924207787268788e-07, "loss": 0.5503, "step": 24661 }, { "epoch": 0.8932270916334661, "grad_norm": 1.349593579935482, "learning_rate": 5.920230808903348e-07, "loss": 0.6511, "step": 24662 }, { "epoch": 0.8932633103947845, "grad_norm": 1.2421399863171032, "learning_rate": 5.916255125158221e-07, "loss": 0.6577, "step": 24663 }, { "epoch": 0.8932995291561029, "grad_norm": 1.4435391668993207, "learning_rate": 5.912280736088094e-07, "loss": 0.7373, "step": 24664 }, { "epoch": 0.8933357479174212, "grad_norm": 1.3673541675886114, "learning_rate": 5.908307641747657e-07, "loss": 0.6709, "step": 24665 }, { "epoch": 0.8933719666787396, "grad_norm": 1.5492008717977452, "learning_rate": 5.904335842191555e-07, "loss": 0.6836, "step": 24666 }, { "epoch": 0.893408185440058, "grad_norm": 1.37545928951608, "learning_rate": 5.900365337474501e-07, "loss": 0.7484, "step": 24667 }, { "epoch": 0.8934444042013763, "grad_norm": 1.2733335072233711, "learning_rate": 5.896396127651093e-07, "loss": 0.5655, "step": 24668 }, { "epoch": 0.8934806229626947, "grad_norm": 1.3332347264527735, "learning_rate": 5.892428212775958e-07, "loss": 0.6782, "step": 24669 }, { "epoch": 0.893516841724013, "grad_norm": 1.2689660664290643, "learning_rate": 5.888461592903704e-07, "loss": 0.7276, "step": 24670 }, { "epoch": 0.8935530604853315, "grad_norm": 1.354602919207363, "learning_rate": 5.884496268088912e-07, "loss": 0.737, "step": 24671 }, { "epoch": 0.8935892792466498, "grad_norm": 1.1216555445920904, "learning_rate": 5.880532238386161e-07, "loss": 0.6834, "step": 24672 }, { "epoch": 0.8936254980079681, "grad_norm": 1.4347879518661424, "learning_rate": 5.876569503849961e-07, "loss": 0.6747, "step": 24673 }, { "epoch": 0.8936617167692865, "grad_norm": 1.2518746735307666, "learning_rate": 5.87260806453489e-07, "loss": 0.6216, "step": 24674 }, { "epoch": 0.8936979355306048, "grad_norm": 1.384024252116513, "learning_rate": 5.86864792049543e-07, "loss": 0.662, "step": 24675 }, { "epoch": 0.8937341542919233, "grad_norm": 1.3797529392988717, "learning_rate": 5.8646890717861e-07, "loss": 0.6735, "step": 24676 }, { "epoch": 0.8937703730532416, "grad_norm": 1.2983106830101059, "learning_rate": 5.860731518461338e-07, "loss": 0.6822, "step": 24677 }, { "epoch": 0.8938065918145599, "grad_norm": 2.5173869937172704, "learning_rate": 5.856775260575642e-07, "loss": 0.723, "step": 24678 }, { "epoch": 0.8938428105758783, "grad_norm": 1.3573988706711475, "learning_rate": 5.852820298183449e-07, "loss": 0.629, "step": 24679 }, { "epoch": 0.8938790293371967, "grad_norm": 1.3648018670329314, "learning_rate": 5.848866631339145e-07, "loss": 0.6044, "step": 24680 }, { "epoch": 0.8939152480985151, "grad_norm": 1.4043531254046036, "learning_rate": 5.84491426009719e-07, "loss": 0.6379, "step": 24681 }, { "epoch": 0.8939514668598334, "grad_norm": 1.3997296225762625, "learning_rate": 5.840963184511938e-07, "loss": 0.6658, "step": 24682 }, { "epoch": 0.8939876856211517, "grad_norm": 1.4046851491484103, "learning_rate": 5.837013404637782e-07, "loss": 0.6362, "step": 24683 }, { "epoch": 0.8940239043824701, "grad_norm": 1.4370855895072914, "learning_rate": 5.833064920529031e-07, "loss": 0.6925, "step": 24684 }, { "epoch": 0.8940601231437885, "grad_norm": 1.5171323481149848, "learning_rate": 5.829117732240075e-07, "loss": 0.6499, "step": 24685 }, { "epoch": 0.8940963419051069, "grad_norm": 1.357187440236701, "learning_rate": 5.825171839825195e-07, "loss": 0.6831, "step": 24686 }, { "epoch": 0.8941325606664252, "grad_norm": 1.2889034469332152, "learning_rate": 5.821227243338712e-07, "loss": 0.6942, "step": 24687 }, { "epoch": 0.8941687794277435, "grad_norm": 1.414778409636069, "learning_rate": 5.817283942834873e-07, "loss": 0.6396, "step": 24688 }, { "epoch": 0.894204998189062, "grad_norm": 1.464442848022983, "learning_rate": 5.813341938367978e-07, "loss": 0.6648, "step": 24689 }, { "epoch": 0.8942412169503803, "grad_norm": 1.4832805388298327, "learning_rate": 5.809401229992251e-07, "loss": 0.7053, "step": 24690 }, { "epoch": 0.8942774357116987, "grad_norm": 1.3596922535759508, "learning_rate": 5.805461817761926e-07, "loss": 0.7115, "step": 24691 }, { "epoch": 0.894313654473017, "grad_norm": 1.4039262752134332, "learning_rate": 5.801523701731216e-07, "loss": 0.7123, "step": 24692 }, { "epoch": 0.8943498732343353, "grad_norm": 1.3499213170786086, "learning_rate": 5.797586881954309e-07, "loss": 0.7278, "step": 24693 }, { "epoch": 0.8943860919956538, "grad_norm": 1.3512195333924402, "learning_rate": 5.793651358485386e-07, "loss": 0.6991, "step": 24694 }, { "epoch": 0.8944223107569721, "grad_norm": 1.3529661681691039, "learning_rate": 5.789717131378581e-07, "loss": 0.6812, "step": 24695 }, { "epoch": 0.8944585295182905, "grad_norm": 1.3258493098591215, "learning_rate": 5.785784200688061e-07, "loss": 0.6599, "step": 24696 }, { "epoch": 0.8944947482796088, "grad_norm": 1.3496258537291193, "learning_rate": 5.781852566467938e-07, "loss": 0.5806, "step": 24697 }, { "epoch": 0.8945309670409272, "grad_norm": 1.2989604904014902, "learning_rate": 5.777922228772314e-07, "loss": 0.6398, "step": 24698 }, { "epoch": 0.8945671858022456, "grad_norm": 1.4154701023956484, "learning_rate": 5.773993187655247e-07, "loss": 0.7241, "step": 24699 }, { "epoch": 0.8946034045635639, "grad_norm": 1.4575514785102417, "learning_rate": 5.770065443170847e-07, "loss": 0.6824, "step": 24700 }, { "epoch": 0.8946396233248823, "grad_norm": 1.1103517204496762, "learning_rate": 5.76613899537315e-07, "loss": 0.6134, "step": 24701 }, { "epoch": 0.8946758420862007, "grad_norm": 1.049293622577895, "learning_rate": 5.762213844316155e-07, "loss": 0.6909, "step": 24702 }, { "epoch": 0.894712060847519, "grad_norm": 1.336031122719483, "learning_rate": 5.758289990053934e-07, "loss": 0.6538, "step": 24703 }, { "epoch": 0.8947482796088374, "grad_norm": 1.2895868578728111, "learning_rate": 5.75436743264044e-07, "loss": 0.6454, "step": 24704 }, { "epoch": 0.8947844983701557, "grad_norm": 1.3536859401045487, "learning_rate": 5.750446172129653e-07, "loss": 0.7179, "step": 24705 }, { "epoch": 0.8948207171314742, "grad_norm": 1.5347092174165549, "learning_rate": 5.746526208575542e-07, "loss": 0.6837, "step": 24706 }, { "epoch": 0.8948569358927925, "grad_norm": 1.3971761470615844, "learning_rate": 5.742607542032053e-07, "loss": 0.7035, "step": 24707 }, { "epoch": 0.8948931546541108, "grad_norm": 1.4817555780971912, "learning_rate": 5.738690172553107e-07, "loss": 0.6698, "step": 24708 }, { "epoch": 0.8949293734154292, "grad_norm": 0.9778105452461479, "learning_rate": 5.73477410019262e-07, "loss": 0.7491, "step": 24709 }, { "epoch": 0.8949655921767475, "grad_norm": 1.306692244275611, "learning_rate": 5.730859325004445e-07, "loss": 0.6206, "step": 24710 }, { "epoch": 0.895001810938066, "grad_norm": 1.2753515368980988, "learning_rate": 5.726945847042497e-07, "loss": 0.6388, "step": 24711 }, { "epoch": 0.8950380296993843, "grad_norm": 1.0468892857236065, "learning_rate": 5.723033666360611e-07, "loss": 0.6831, "step": 24712 }, { "epoch": 0.8950742484607026, "grad_norm": 1.3426705775311645, "learning_rate": 5.719122783012598e-07, "loss": 0.6656, "step": 24713 }, { "epoch": 0.895110467222021, "grad_norm": 0.9019317406560438, "learning_rate": 5.715213197052316e-07, "loss": 0.6476, "step": 24714 }, { "epoch": 0.8951466859833394, "grad_norm": 1.3564002308999248, "learning_rate": 5.711304908533555e-07, "loss": 0.5749, "step": 24715 }, { "epoch": 0.8951829047446578, "grad_norm": 1.3374791675130633, "learning_rate": 5.707397917510072e-07, "loss": 0.6447, "step": 24716 }, { "epoch": 0.8952191235059761, "grad_norm": 1.4476205979932415, "learning_rate": 5.703492224035645e-07, "loss": 0.6993, "step": 24717 }, { "epoch": 0.8952553422672944, "grad_norm": 1.3894111701501732, "learning_rate": 5.699587828164033e-07, "loss": 0.6236, "step": 24718 }, { "epoch": 0.8952915610286128, "grad_norm": 1.016981389019018, "learning_rate": 5.695684729948947e-07, "loss": 0.6928, "step": 24719 }, { "epoch": 0.8953277797899312, "grad_norm": 1.5235665743507663, "learning_rate": 5.691782929444089e-07, "loss": 0.6454, "step": 24720 }, { "epoch": 0.8953639985512496, "grad_norm": 1.3463333319343778, "learning_rate": 5.687882426703185e-07, "loss": 0.6907, "step": 24721 }, { "epoch": 0.8954002173125679, "grad_norm": 1.394738292818377, "learning_rate": 5.683983221779887e-07, "loss": 0.6632, "step": 24722 }, { "epoch": 0.8954364360738862, "grad_norm": 1.3331853797216708, "learning_rate": 5.680085314727846e-07, "loss": 0.6197, "step": 24723 }, { "epoch": 0.8954726548352047, "grad_norm": 1.571783189623444, "learning_rate": 5.676188705600693e-07, "loss": 0.7559, "step": 24724 }, { "epoch": 0.895508873596523, "grad_norm": 1.36328902829856, "learning_rate": 5.672293394452088e-07, "loss": 0.6484, "step": 24725 }, { "epoch": 0.8955450923578414, "grad_norm": 1.0053958607235778, "learning_rate": 5.66839938133561e-07, "loss": 0.6296, "step": 24726 }, { "epoch": 0.8955813111191597, "grad_norm": 1.2538892174392842, "learning_rate": 5.664506666304837e-07, "loss": 0.6181, "step": 24727 }, { "epoch": 0.895617529880478, "grad_norm": 1.273689410021621, "learning_rate": 5.660615249413326e-07, "loss": 0.6578, "step": 24728 }, { "epoch": 0.8956537486417965, "grad_norm": 0.9687906192331591, "learning_rate": 5.656725130714657e-07, "loss": 0.6682, "step": 24729 }, { "epoch": 0.8956899674031148, "grad_norm": 1.4544877057744727, "learning_rate": 5.652836310262355e-07, "loss": 0.642, "step": 24730 }, { "epoch": 0.8957261861644332, "grad_norm": 1.323622710774964, "learning_rate": 5.648948788109909e-07, "loss": 0.6787, "step": 24731 }, { "epoch": 0.8957624049257515, "grad_norm": 1.0357522639224686, "learning_rate": 5.645062564310844e-07, "loss": 0.6678, "step": 24732 }, { "epoch": 0.8957986236870699, "grad_norm": 1.156929536316307, "learning_rate": 5.641177638918627e-07, "loss": 0.6543, "step": 24733 }, { "epoch": 0.8958348424483883, "grad_norm": 1.1657900810256328, "learning_rate": 5.637294011986716e-07, "loss": 0.6373, "step": 24734 }, { "epoch": 0.8958710612097066, "grad_norm": 1.3614731024572453, "learning_rate": 5.633411683568535e-07, "loss": 0.7161, "step": 24735 }, { "epoch": 0.895907279971025, "grad_norm": 1.0781109356956244, "learning_rate": 5.629530653717552e-07, "loss": 0.7675, "step": 24736 }, { "epoch": 0.8959434987323434, "grad_norm": 1.4873785850917056, "learning_rate": 5.625650922487136e-07, "loss": 0.6701, "step": 24737 }, { "epoch": 0.8959797174936617, "grad_norm": 1.312140326803495, "learning_rate": 5.621772489930699e-07, "loss": 0.656, "step": 24738 }, { "epoch": 0.8960159362549801, "grad_norm": 1.3845656671606208, "learning_rate": 5.617895356101577e-07, "loss": 0.7144, "step": 24739 }, { "epoch": 0.8960521550162984, "grad_norm": 1.5489493727879127, "learning_rate": 5.614019521053171e-07, "loss": 0.6484, "step": 24740 }, { "epoch": 0.8960883737776169, "grad_norm": 1.3034886910099683, "learning_rate": 5.610144984838783e-07, "loss": 0.6288, "step": 24741 }, { "epoch": 0.8961245925389352, "grad_norm": 0.9792934593304976, "learning_rate": 5.606271747511727e-07, "loss": 0.6961, "step": 24742 }, { "epoch": 0.8961608113002535, "grad_norm": 1.2421892903872849, "learning_rate": 5.602399809125325e-07, "loss": 0.5497, "step": 24743 }, { "epoch": 0.8961970300615719, "grad_norm": 1.3287882254212657, "learning_rate": 5.598529169732847e-07, "loss": 0.6735, "step": 24744 }, { "epoch": 0.8962332488228902, "grad_norm": 1.4067844236327807, "learning_rate": 5.594659829387561e-07, "loss": 0.6674, "step": 24745 }, { "epoch": 0.8962694675842087, "grad_norm": 1.3726893145270413, "learning_rate": 5.590791788142691e-07, "loss": 0.6773, "step": 24746 }, { "epoch": 0.896305686345527, "grad_norm": 1.3913273162985105, "learning_rate": 5.586925046051494e-07, "loss": 0.6733, "step": 24747 }, { "epoch": 0.8963419051068453, "grad_norm": 1.3359534021457293, "learning_rate": 5.583059603167162e-07, "loss": 0.5832, "step": 24748 }, { "epoch": 0.8963781238681637, "grad_norm": 1.2210014660564603, "learning_rate": 5.579195459542896e-07, "loss": 0.62, "step": 24749 }, { "epoch": 0.896414342629482, "grad_norm": 1.1394709275300505, "learning_rate": 5.575332615231855e-07, "loss": 0.646, "step": 24750 }, { "epoch": 0.8964505613908005, "grad_norm": 1.0823752866263119, "learning_rate": 5.571471070287215e-07, "loss": 0.6714, "step": 24751 }, { "epoch": 0.8964867801521188, "grad_norm": 1.4621278916046871, "learning_rate": 5.567610824762115e-07, "loss": 0.7283, "step": 24752 }, { "epoch": 0.8965229989134371, "grad_norm": 1.4716507017170644, "learning_rate": 5.563751878709645e-07, "loss": 0.6348, "step": 24753 }, { "epoch": 0.8965592176747555, "grad_norm": 2.0017612718903433, "learning_rate": 5.559894232182939e-07, "loss": 0.6965, "step": 24754 }, { "epoch": 0.8965954364360739, "grad_norm": 1.4090135890173079, "learning_rate": 5.556037885235078e-07, "loss": 0.6672, "step": 24755 }, { "epoch": 0.8966316551973923, "grad_norm": 1.506775286795712, "learning_rate": 5.552182837919106e-07, "loss": 0.6773, "step": 24756 }, { "epoch": 0.8966678739587106, "grad_norm": 1.3773344377688663, "learning_rate": 5.548329090288085e-07, "loss": 0.71, "step": 24757 }, { "epoch": 0.8967040927200289, "grad_norm": 0.9823086823988686, "learning_rate": 5.544476642395069e-07, "loss": 0.6445, "step": 24758 }, { "epoch": 0.8967403114813474, "grad_norm": 1.41805255395289, "learning_rate": 5.540625494293039e-07, "loss": 0.6531, "step": 24759 }, { "epoch": 0.8967765302426657, "grad_norm": 1.3333400989866218, "learning_rate": 5.536775646035008e-07, "loss": 0.6359, "step": 24760 }, { "epoch": 0.8968127490039841, "grad_norm": 1.3597998243007534, "learning_rate": 5.532927097673923e-07, "loss": 0.7071, "step": 24761 }, { "epoch": 0.8968489677653024, "grad_norm": 1.543179409519055, "learning_rate": 5.529079849262797e-07, "loss": 0.7186, "step": 24762 }, { "epoch": 0.8968851865266207, "grad_norm": 0.9342781394202695, "learning_rate": 5.525233900854532e-07, "loss": 0.6544, "step": 24763 }, { "epoch": 0.8969214052879392, "grad_norm": 1.4091252946302597, "learning_rate": 5.521389252502063e-07, "loss": 0.7086, "step": 24764 }, { "epoch": 0.8969576240492575, "grad_norm": 1.1509701048744139, "learning_rate": 5.517545904258293e-07, "loss": 0.7042, "step": 24765 }, { "epoch": 0.8969938428105759, "grad_norm": 1.2735453089030084, "learning_rate": 5.513703856176112e-07, "loss": 0.6165, "step": 24766 }, { "epoch": 0.8970300615718942, "grad_norm": 1.4904337545886042, "learning_rate": 5.509863108308366e-07, "loss": 0.6618, "step": 24767 }, { "epoch": 0.8970662803332126, "grad_norm": 1.2341582750846616, "learning_rate": 5.506023660707949e-07, "loss": 0.6032, "step": 24768 }, { "epoch": 0.897102499094531, "grad_norm": 1.4247996034067394, "learning_rate": 5.502185513427683e-07, "loss": 0.6791, "step": 24769 }, { "epoch": 0.8971387178558493, "grad_norm": 1.079346003345775, "learning_rate": 5.49834866652037e-07, "loss": 0.6439, "step": 24770 }, { "epoch": 0.8971749366171677, "grad_norm": 1.5288463089566366, "learning_rate": 5.494513120038814e-07, "loss": 0.6515, "step": 24771 }, { "epoch": 0.8972111553784861, "grad_norm": 1.412460648834605, "learning_rate": 5.490678874035793e-07, "loss": 0.6997, "step": 24772 }, { "epoch": 0.8972473741398044, "grad_norm": 1.6778606060888184, "learning_rate": 5.486845928564088e-07, "loss": 0.5387, "step": 24773 }, { "epoch": 0.8972835929011228, "grad_norm": 1.3897301736305403, "learning_rate": 5.483014283676413e-07, "loss": 0.638, "step": 24774 }, { "epoch": 0.8973198116624411, "grad_norm": 1.3743219112431546, "learning_rate": 5.479183939425525e-07, "loss": 0.6817, "step": 24775 }, { "epoch": 0.8973560304237596, "grad_norm": 1.3619887118859049, "learning_rate": 5.475354895864115e-07, "loss": 0.6209, "step": 24776 }, { "epoch": 0.8973922491850779, "grad_norm": 1.3241135103892105, "learning_rate": 5.471527153044875e-07, "loss": 0.6328, "step": 24777 }, { "epoch": 0.8974284679463962, "grad_norm": 1.3444294784869733, "learning_rate": 5.467700711020474e-07, "loss": 0.7239, "step": 24778 }, { "epoch": 0.8974646867077146, "grad_norm": 1.436440597054232, "learning_rate": 5.46387556984359e-07, "loss": 0.5841, "step": 24779 }, { "epoch": 0.8975009054690329, "grad_norm": 1.3474101256480184, "learning_rate": 5.46005172956684e-07, "loss": 0.6365, "step": 24780 }, { "epoch": 0.8975371242303514, "grad_norm": 1.0565948754064836, "learning_rate": 5.456229190242856e-07, "loss": 0.6851, "step": 24781 }, { "epoch": 0.8975733429916697, "grad_norm": 0.9682520994202969, "learning_rate": 5.452407951924211e-07, "loss": 0.68, "step": 24782 }, { "epoch": 0.897609561752988, "grad_norm": 1.3096161968718245, "learning_rate": 5.448588014663525e-07, "loss": 0.6716, "step": 24783 }, { "epoch": 0.8976457805143064, "grad_norm": 0.9800633812727957, "learning_rate": 5.444769378513348e-07, "loss": 0.6432, "step": 24784 }, { "epoch": 0.8976819992756248, "grad_norm": 1.236179492025071, "learning_rate": 5.440952043526215e-07, "loss": 0.6122, "step": 24785 }, { "epoch": 0.8977182180369432, "grad_norm": 1.3602782706886396, "learning_rate": 5.437136009754674e-07, "loss": 0.6783, "step": 24786 }, { "epoch": 0.8977544367982615, "grad_norm": 1.3881083765627578, "learning_rate": 5.433321277251247e-07, "loss": 0.66, "step": 24787 }, { "epoch": 0.8977906555595798, "grad_norm": 1.4369593958154498, "learning_rate": 5.429507846068405e-07, "loss": 0.6444, "step": 24788 }, { "epoch": 0.8978268743208982, "grad_norm": 1.8655827940826493, "learning_rate": 5.425695716258617e-07, "loss": 0.6603, "step": 24789 }, { "epoch": 0.8978630930822166, "grad_norm": 1.311395519605388, "learning_rate": 5.421884887874373e-07, "loss": 0.5798, "step": 24790 }, { "epoch": 0.897899311843535, "grad_norm": 0.9004296763794193, "learning_rate": 5.418075360968089e-07, "loss": 0.6312, "step": 24791 }, { "epoch": 0.8979355306048533, "grad_norm": 1.3662869130319522, "learning_rate": 5.4142671355922e-07, "loss": 0.6701, "step": 24792 }, { "epoch": 0.8979717493661716, "grad_norm": 1.4694826373587973, "learning_rate": 5.410460211799095e-07, "loss": 0.6456, "step": 24793 }, { "epoch": 0.8980079681274901, "grad_norm": 1.4539697643726774, "learning_rate": 5.40665458964118e-07, "loss": 0.7256, "step": 24794 }, { "epoch": 0.8980441868888084, "grad_norm": 1.416296112299089, "learning_rate": 5.402850269170823e-07, "loss": 0.6445, "step": 24795 }, { "epoch": 0.8980804056501268, "grad_norm": 1.2641845419348825, "learning_rate": 5.399047250440348e-07, "loss": 0.6135, "step": 24796 }, { "epoch": 0.8981166244114451, "grad_norm": 1.4283626963762446, "learning_rate": 5.395245533502114e-07, "loss": 0.6869, "step": 24797 }, { "epoch": 0.8981528431727634, "grad_norm": 1.4074786706253932, "learning_rate": 5.391445118408434e-07, "loss": 0.7094, "step": 24798 }, { "epoch": 0.8981890619340819, "grad_norm": 1.316368955004645, "learning_rate": 5.387646005211599e-07, "loss": 0.6148, "step": 24799 }, { "epoch": 0.8982252806954002, "grad_norm": 1.300214348943484, "learning_rate": 5.38384819396387e-07, "loss": 0.6961, "step": 24800 }, { "epoch": 0.8982614994567186, "grad_norm": 1.5128082617401093, "learning_rate": 5.380051684717547e-07, "loss": 0.6518, "step": 24801 }, { "epoch": 0.898297718218037, "grad_norm": 1.439873879738902, "learning_rate": 5.376256477524855e-07, "loss": 0.6473, "step": 24802 }, { "epoch": 0.8983339369793553, "grad_norm": 1.5951891705257233, "learning_rate": 5.37246257243802e-07, "loss": 0.6809, "step": 24803 }, { "epoch": 0.8983701557406737, "grad_norm": 1.3437945973957712, "learning_rate": 5.368669969509222e-07, "loss": 0.6608, "step": 24804 }, { "epoch": 0.898406374501992, "grad_norm": 1.2267812061520034, "learning_rate": 5.364878668790696e-07, "loss": 0.6636, "step": 24805 }, { "epoch": 0.8984425932633104, "grad_norm": 1.282273120281821, "learning_rate": 5.361088670334591e-07, "loss": 0.6452, "step": 24806 }, { "epoch": 0.8984788120246288, "grad_norm": 1.3814367910775174, "learning_rate": 5.357299974193053e-07, "loss": 0.5942, "step": 24807 }, { "epoch": 0.8985150307859471, "grad_norm": 1.0669919823849372, "learning_rate": 5.35351258041824e-07, "loss": 0.6875, "step": 24808 }, { "epoch": 0.8985512495472655, "grad_norm": 1.0712226024005846, "learning_rate": 5.349726489062257e-07, "loss": 0.6789, "step": 24809 }, { "epoch": 0.8985874683085838, "grad_norm": 1.4607654716238425, "learning_rate": 5.345941700177204e-07, "loss": 0.6779, "step": 24810 }, { "epoch": 0.8986236870699023, "grad_norm": 0.9798845475900296, "learning_rate": 5.342158213815152e-07, "loss": 0.6738, "step": 24811 }, { "epoch": 0.8986599058312206, "grad_norm": 1.060677848586038, "learning_rate": 5.338376030028181e-07, "loss": 0.6972, "step": 24812 }, { "epoch": 0.8986961245925389, "grad_norm": 0.9633029214394719, "learning_rate": 5.334595148868349e-07, "loss": 0.6205, "step": 24813 }, { "epoch": 0.8987323433538573, "grad_norm": 1.335689186977981, "learning_rate": 5.330815570387638e-07, "loss": 0.6961, "step": 24814 }, { "epoch": 0.8987685621151756, "grad_norm": 1.387073513105629, "learning_rate": 5.327037294638115e-07, "loss": 0.7155, "step": 24815 }, { "epoch": 0.8988047808764941, "grad_norm": 0.95421567413717, "learning_rate": 5.323260321671742e-07, "loss": 0.6393, "step": 24816 }, { "epoch": 0.8988409996378124, "grad_norm": 1.3451376489551126, "learning_rate": 5.319484651540497e-07, "loss": 0.6576, "step": 24817 }, { "epoch": 0.8988772183991307, "grad_norm": 1.4378274656050452, "learning_rate": 5.315710284296316e-07, "loss": 0.6504, "step": 24818 }, { "epoch": 0.8989134371604491, "grad_norm": 1.4445315809491917, "learning_rate": 5.311937219991181e-07, "loss": 0.7257, "step": 24819 }, { "epoch": 0.8989496559217675, "grad_norm": 1.2947470293773777, "learning_rate": 5.308165458676984e-07, "loss": 0.6298, "step": 24820 }, { "epoch": 0.8989858746830859, "grad_norm": 1.3522604810354732, "learning_rate": 5.304395000405637e-07, "loss": 0.6275, "step": 24821 }, { "epoch": 0.8990220934444042, "grad_norm": 1.4291327265216354, "learning_rate": 5.300625845229013e-07, "loss": 0.6109, "step": 24822 }, { "epoch": 0.8990583122057225, "grad_norm": 1.1393893984516785, "learning_rate": 5.29685799319899e-07, "loss": 0.7091, "step": 24823 }, { "epoch": 0.899094530967041, "grad_norm": 1.0727555513857683, "learning_rate": 5.293091444367415e-07, "loss": 0.6251, "step": 24824 }, { "epoch": 0.8991307497283593, "grad_norm": 1.4740682000459826, "learning_rate": 5.289326198786116e-07, "loss": 0.6934, "step": 24825 }, { "epoch": 0.8991669684896777, "grad_norm": 1.0230351461474292, "learning_rate": 5.285562256506915e-07, "loss": 0.7288, "step": 24826 }, { "epoch": 0.899203187250996, "grad_norm": 1.435623927870588, "learning_rate": 5.281799617581607e-07, "loss": 0.668, "step": 24827 }, { "epoch": 0.8992394060123143, "grad_norm": 1.3812211458452561, "learning_rate": 5.27803828206196e-07, "loss": 0.6313, "step": 24828 }, { "epoch": 0.8992756247736328, "grad_norm": 1.3895403269723623, "learning_rate": 5.27427824999972e-07, "loss": 0.6688, "step": 24829 }, { "epoch": 0.8993118435349511, "grad_norm": 1.353051803846846, "learning_rate": 5.270519521446648e-07, "loss": 0.6713, "step": 24830 }, { "epoch": 0.8993480622962695, "grad_norm": 1.0710880861264762, "learning_rate": 5.266762096454481e-07, "loss": 0.633, "step": 24831 }, { "epoch": 0.8993842810575878, "grad_norm": 1.5795850055036782, "learning_rate": 5.263005975074897e-07, "loss": 0.7093, "step": 24832 }, { "epoch": 0.8994204998189061, "grad_norm": 1.3715011305289366, "learning_rate": 5.25925115735959e-07, "loss": 0.6583, "step": 24833 }, { "epoch": 0.8994567185802246, "grad_norm": 1.576377158050143, "learning_rate": 5.255497643360241e-07, "loss": 0.6168, "step": 24834 }, { "epoch": 0.8994929373415429, "grad_norm": 1.0778658190808705, "learning_rate": 5.251745433128497e-07, "loss": 0.717, "step": 24835 }, { "epoch": 0.8995291561028613, "grad_norm": 1.085523804040331, "learning_rate": 5.247994526715971e-07, "loss": 0.6606, "step": 24836 }, { "epoch": 0.8995653748641796, "grad_norm": 1.1345036984763748, "learning_rate": 5.244244924174313e-07, "loss": 0.7119, "step": 24837 }, { "epoch": 0.899601593625498, "grad_norm": 2.1203734753864145, "learning_rate": 5.240496625555103e-07, "loss": 0.6512, "step": 24838 }, { "epoch": 0.8996378123868164, "grad_norm": 1.5387596519827555, "learning_rate": 5.236749630909932e-07, "loss": 0.6653, "step": 24839 }, { "epoch": 0.8996740311481347, "grad_norm": 1.4676060698536564, "learning_rate": 5.233003940290327e-07, "loss": 0.7149, "step": 24840 }, { "epoch": 0.8997102499094531, "grad_norm": 1.4143432396463622, "learning_rate": 5.229259553747867e-07, "loss": 0.6512, "step": 24841 }, { "epoch": 0.8997464686707715, "grad_norm": 0.9712921868115245, "learning_rate": 5.22551647133408e-07, "loss": 0.6771, "step": 24842 }, { "epoch": 0.8997826874320898, "grad_norm": 1.337228081231824, "learning_rate": 5.221774693100457e-07, "loss": 0.6538, "step": 24843 }, { "epoch": 0.8998189061934082, "grad_norm": 1.0218179523469009, "learning_rate": 5.218034219098489e-07, "loss": 0.6254, "step": 24844 }, { "epoch": 0.8998551249547265, "grad_norm": 0.9942225955723971, "learning_rate": 5.214295049379658e-07, "loss": 0.6457, "step": 24845 }, { "epoch": 0.899891343716045, "grad_norm": 1.5454892763491526, "learning_rate": 5.210557183995413e-07, "loss": 0.6306, "step": 24846 }, { "epoch": 0.8999275624773633, "grad_norm": 1.5166811919078997, "learning_rate": 5.206820622997177e-07, "loss": 0.7055, "step": 24847 }, { "epoch": 0.8999637812386816, "grad_norm": 1.445040282182759, "learning_rate": 5.2030853664364e-07, "loss": 0.6545, "step": 24848 }, { "epoch": 0.9, "grad_norm": 2.6970115199405256, "learning_rate": 5.19935141436445e-07, "loss": 0.6386, "step": 24849 }, { "epoch": 0.9000362187613183, "grad_norm": 1.1143151337423332, "learning_rate": 5.195618766832744e-07, "loss": 0.6677, "step": 24850 }, { "epoch": 0.9000724375226368, "grad_norm": 1.305255744730833, "learning_rate": 5.191887423892595e-07, "loss": 0.7173, "step": 24851 }, { "epoch": 0.9001086562839551, "grad_norm": 1.3279066313009895, "learning_rate": 5.188157385595405e-07, "loss": 0.5795, "step": 24852 }, { "epoch": 0.9001448750452734, "grad_norm": 1.4871437601854243, "learning_rate": 5.184428651992479e-07, "loss": 0.7852, "step": 24853 }, { "epoch": 0.9001810938065918, "grad_norm": 1.370981747788395, "learning_rate": 5.180701223135121e-07, "loss": 0.6454, "step": 24854 }, { "epoch": 0.9002173125679102, "grad_norm": 1.259304529276019, "learning_rate": 5.176975099074622e-07, "loss": 0.5496, "step": 24855 }, { "epoch": 0.9002535313292286, "grad_norm": 0.9197863994221462, "learning_rate": 5.173250279862274e-07, "loss": 0.687, "step": 24856 }, { "epoch": 0.9002897500905469, "grad_norm": 1.4145892397755373, "learning_rate": 5.169526765549326e-07, "loss": 0.6867, "step": 24857 }, { "epoch": 0.9003259688518652, "grad_norm": 0.9626379492180821, "learning_rate": 5.165804556186993e-07, "loss": 0.6534, "step": 24858 }, { "epoch": 0.9003621876131837, "grad_norm": 1.5928467102785149, "learning_rate": 5.162083651826544e-07, "loss": 0.6185, "step": 24859 }, { "epoch": 0.900398406374502, "grad_norm": 1.4955683076820603, "learning_rate": 5.158364052519149e-07, "loss": 0.7287, "step": 24860 }, { "epoch": 0.9004346251358204, "grad_norm": 1.1077762151245003, "learning_rate": 5.154645758316001e-07, "loss": 0.6646, "step": 24861 }, { "epoch": 0.9004708438971387, "grad_norm": 1.2073326961235447, "learning_rate": 5.150928769268259e-07, "loss": 0.6301, "step": 24862 }, { "epoch": 0.900507062658457, "grad_norm": 1.0321967313821885, "learning_rate": 5.147213085427083e-07, "loss": 0.6886, "step": 24863 }, { "epoch": 0.9005432814197755, "grad_norm": 1.2994526408558162, "learning_rate": 5.143498706843609e-07, "loss": 0.6941, "step": 24864 }, { "epoch": 0.9005795001810938, "grad_norm": 1.3020847307833485, "learning_rate": 5.139785633568928e-07, "loss": 0.6302, "step": 24865 }, { "epoch": 0.9006157189424122, "grad_norm": 1.0558096595215385, "learning_rate": 5.136073865654146e-07, "loss": 0.7086, "step": 24866 }, { "epoch": 0.9006519377037305, "grad_norm": 1.2571877211022473, "learning_rate": 5.132363403150353e-07, "loss": 0.6823, "step": 24867 }, { "epoch": 0.9006881564650489, "grad_norm": 1.2127531371746905, "learning_rate": 5.128654246108589e-07, "loss": 0.6772, "step": 24868 }, { "epoch": 0.9007243752263673, "grad_norm": 1.3360315800048665, "learning_rate": 5.1249463945799e-07, "loss": 0.6662, "step": 24869 }, { "epoch": 0.9007605939876856, "grad_norm": 1.1453097719315977, "learning_rate": 5.121239848615323e-07, "loss": 0.7299, "step": 24870 }, { "epoch": 0.900796812749004, "grad_norm": 1.3296956332985619, "learning_rate": 5.117534608265862e-07, "loss": 0.682, "step": 24871 }, { "epoch": 0.9008330315103223, "grad_norm": 1.5844525069967277, "learning_rate": 5.113830673582487e-07, "loss": 0.6613, "step": 24872 }, { "epoch": 0.9008692502716407, "grad_norm": 1.538905807448188, "learning_rate": 5.110128044616158e-07, "loss": 0.6631, "step": 24873 }, { "epoch": 0.9009054690329591, "grad_norm": 1.3240114931782443, "learning_rate": 5.106426721417868e-07, "loss": 0.6481, "step": 24874 }, { "epoch": 0.9009416877942774, "grad_norm": 1.3936183663610753, "learning_rate": 5.102726704038518e-07, "loss": 0.6838, "step": 24875 }, { "epoch": 0.9009779065555958, "grad_norm": 1.343186238732486, "learning_rate": 5.099027992529026e-07, "loss": 0.6674, "step": 24876 }, { "epoch": 0.9010141253169142, "grad_norm": 1.5427013684632198, "learning_rate": 5.095330586940305e-07, "loss": 0.6455, "step": 24877 }, { "epoch": 0.9010503440782325, "grad_norm": 1.0221909796377084, "learning_rate": 5.091634487323238e-07, "loss": 0.622, "step": 24878 }, { "epoch": 0.9010865628395509, "grad_norm": 1.497786314609254, "learning_rate": 5.087939693728661e-07, "loss": 0.6645, "step": 24879 }, { "epoch": 0.9011227816008692, "grad_norm": 1.421124357110961, "learning_rate": 5.084246206207422e-07, "loss": 0.7245, "step": 24880 }, { "epoch": 0.9011590003621877, "grad_norm": 1.0023984091559073, "learning_rate": 5.08055402481037e-07, "loss": 0.6732, "step": 24881 }, { "epoch": 0.901195219123506, "grad_norm": 1.5608417401901626, "learning_rate": 5.076863149588296e-07, "loss": 0.7227, "step": 24882 }, { "epoch": 0.9012314378848243, "grad_norm": 1.4017849369302908, "learning_rate": 5.073173580591994e-07, "loss": 0.7037, "step": 24883 }, { "epoch": 0.9012676566461427, "grad_norm": 1.4864726867252342, "learning_rate": 5.069485317872213e-07, "loss": 0.6768, "step": 24884 }, { "epoch": 0.901303875407461, "grad_norm": 1.003925038217094, "learning_rate": 5.065798361479746e-07, "loss": 0.6725, "step": 24885 }, { "epoch": 0.9013400941687795, "grad_norm": 1.3741396696900785, "learning_rate": 5.062112711465317e-07, "loss": 0.6024, "step": 24886 }, { "epoch": 0.9013763129300978, "grad_norm": 1.5888607778812636, "learning_rate": 5.05842836787962e-07, "loss": 0.6948, "step": 24887 }, { "epoch": 0.9014125316914161, "grad_norm": 1.261479373031047, "learning_rate": 5.05474533077338e-07, "loss": 0.614, "step": 24888 }, { "epoch": 0.9014487504527345, "grad_norm": 1.3071200862468677, "learning_rate": 5.05106360019727e-07, "loss": 0.5843, "step": 24889 }, { "epoch": 0.9014849692140529, "grad_norm": 1.444375951288786, "learning_rate": 5.047383176201959e-07, "loss": 0.7053, "step": 24890 }, { "epoch": 0.9015211879753713, "grad_norm": 1.3048715825618977, "learning_rate": 5.043704058838073e-07, "loss": 0.6545, "step": 24891 }, { "epoch": 0.9015574067366896, "grad_norm": 1.3887440217399767, "learning_rate": 5.04002624815626e-07, "loss": 0.7187, "step": 24892 }, { "epoch": 0.9015936254980079, "grad_norm": 1.349701781635634, "learning_rate": 5.036349744207136e-07, "loss": 0.6611, "step": 24893 }, { "epoch": 0.9016298442593264, "grad_norm": 1.4542214220603855, "learning_rate": 5.032674547041283e-07, "loss": 0.6834, "step": 24894 }, { "epoch": 0.9016660630206447, "grad_norm": 0.9882904126403148, "learning_rate": 5.029000656709249e-07, "loss": 0.691, "step": 24895 }, { "epoch": 0.9017022817819631, "grad_norm": 1.369195864892222, "learning_rate": 5.025328073261638e-07, "loss": 0.6595, "step": 24896 }, { "epoch": 0.9017385005432814, "grad_norm": 1.1393915330058717, "learning_rate": 5.021656796748964e-07, "loss": 0.6527, "step": 24897 }, { "epoch": 0.9017747193045997, "grad_norm": 1.3426926397810233, "learning_rate": 5.017986827221733e-07, "loss": 0.6538, "step": 24898 }, { "epoch": 0.9018109380659182, "grad_norm": 1.4429933615580222, "learning_rate": 5.014318164730469e-07, "loss": 0.6817, "step": 24899 }, { "epoch": 0.9018471568272365, "grad_norm": 1.4103467071654598, "learning_rate": 5.010650809325657e-07, "loss": 0.6515, "step": 24900 }, { "epoch": 0.9018833755885549, "grad_norm": 1.5777776342922982, "learning_rate": 5.006984761057754e-07, "loss": 0.66, "step": 24901 }, { "epoch": 0.9019195943498732, "grad_norm": 1.5979656127487842, "learning_rate": 5.003320019977187e-07, "loss": 0.6664, "step": 24902 }, { "epoch": 0.9019558131111916, "grad_norm": 1.451056759505454, "learning_rate": 4.999656586134438e-07, "loss": 0.7038, "step": 24903 }, { "epoch": 0.90199203187251, "grad_norm": 1.5275343205327514, "learning_rate": 4.995994459579879e-07, "loss": 0.695, "step": 24904 }, { "epoch": 0.9020282506338283, "grad_norm": 1.4532992139334973, "learning_rate": 4.992333640363911e-07, "loss": 0.6478, "step": 24905 }, { "epoch": 0.9020644693951467, "grad_norm": 1.0119328172549809, "learning_rate": 4.988674128536896e-07, "loss": 0.6568, "step": 24906 }, { "epoch": 0.902100688156465, "grad_norm": 1.4400313521654386, "learning_rate": 4.985015924149239e-07, "loss": 0.6998, "step": 24907 }, { "epoch": 0.9021369069177834, "grad_norm": 1.5260281421161885, "learning_rate": 4.981359027251231e-07, "loss": 0.6984, "step": 24908 }, { "epoch": 0.9021731256791018, "grad_norm": 1.4931907139976228, "learning_rate": 4.977703437893211e-07, "loss": 0.6849, "step": 24909 }, { "epoch": 0.9022093444404201, "grad_norm": 1.385522527940551, "learning_rate": 4.974049156125504e-07, "loss": 0.6832, "step": 24910 }, { "epoch": 0.9022455632017385, "grad_norm": 0.937843621133474, "learning_rate": 4.970396181998372e-07, "loss": 0.6636, "step": 24911 }, { "epoch": 0.9022817819630569, "grad_norm": 1.2466092642750397, "learning_rate": 4.966744515562083e-07, "loss": 0.6495, "step": 24912 }, { "epoch": 0.9023180007243752, "grad_norm": 0.8795711008720738, "learning_rate": 4.963094156866887e-07, "loss": 0.689, "step": 24913 }, { "epoch": 0.9023542194856936, "grad_norm": 1.357866940158735, "learning_rate": 4.959445105963034e-07, "loss": 0.6713, "step": 24914 }, { "epoch": 0.9023904382470119, "grad_norm": 1.4582616281451892, "learning_rate": 4.955797362900727e-07, "loss": 0.6954, "step": 24915 }, { "epoch": 0.9024266570083304, "grad_norm": 1.4895288814804757, "learning_rate": 4.95215092773016e-07, "loss": 0.7538, "step": 24916 }, { "epoch": 0.9024628757696487, "grad_norm": 1.4495195652716524, "learning_rate": 4.948505800501502e-07, "loss": 0.701, "step": 24917 }, { "epoch": 0.902499094530967, "grad_norm": 1.3318988511911591, "learning_rate": 4.944861981264937e-07, "loss": 0.646, "step": 24918 }, { "epoch": 0.9025353132922854, "grad_norm": 1.3769490668054394, "learning_rate": 4.94121947007059e-07, "loss": 0.6209, "step": 24919 }, { "epoch": 0.9025715320536037, "grad_norm": 1.0891673219691842, "learning_rate": 4.937578266968568e-07, "loss": 0.6877, "step": 24920 }, { "epoch": 0.9026077508149222, "grad_norm": 1.2963892613402181, "learning_rate": 4.933938372009029e-07, "loss": 0.6628, "step": 24921 }, { "epoch": 0.9026439695762405, "grad_norm": 1.0315833820585372, "learning_rate": 4.930299785242021e-07, "loss": 0.7279, "step": 24922 }, { "epoch": 0.9026801883375588, "grad_norm": 0.9770922997540434, "learning_rate": 4.926662506717628e-07, "loss": 0.6533, "step": 24923 }, { "epoch": 0.9027164070988772, "grad_norm": 1.134615830384894, "learning_rate": 4.923026536485876e-07, "loss": 0.7186, "step": 24924 }, { "epoch": 0.9027526258601956, "grad_norm": 1.3497763860244516, "learning_rate": 4.919391874596835e-07, "loss": 0.6583, "step": 24925 }, { "epoch": 0.902788844621514, "grad_norm": 1.5374560631651704, "learning_rate": 4.915758521100511e-07, "loss": 0.6887, "step": 24926 }, { "epoch": 0.9028250633828323, "grad_norm": 1.117369887360667, "learning_rate": 4.912126476046897e-07, "loss": 0.6747, "step": 24927 }, { "epoch": 0.9028612821441506, "grad_norm": 1.1020416216391464, "learning_rate": 4.908495739485963e-07, "loss": 0.6604, "step": 24928 }, { "epoch": 0.902897500905469, "grad_norm": 1.282985137469631, "learning_rate": 4.904866311467693e-07, "loss": 0.6144, "step": 24929 }, { "epoch": 0.9029337196667874, "grad_norm": 1.5790221666748483, "learning_rate": 4.901238192042024e-07, "loss": 0.6872, "step": 24930 }, { "epoch": 0.9029699384281058, "grad_norm": 1.3394530911249538, "learning_rate": 4.897611381258871e-07, "loss": 0.6629, "step": 24931 }, { "epoch": 0.9030061571894241, "grad_norm": 1.3293971010610774, "learning_rate": 4.89398587916815e-07, "loss": 0.7, "step": 24932 }, { "epoch": 0.9030423759507424, "grad_norm": 1.0750677863487532, "learning_rate": 4.890361685819756e-07, "loss": 0.6821, "step": 24933 }, { "epoch": 0.9030785947120609, "grad_norm": 1.3995830369984328, "learning_rate": 4.886738801263557e-07, "loss": 0.6804, "step": 24934 }, { "epoch": 0.9031148134733792, "grad_norm": 1.4872193333911465, "learning_rate": 4.883117225549395e-07, "loss": 0.6605, "step": 24935 }, { "epoch": 0.9031510322346976, "grad_norm": 1.206250017001221, "learning_rate": 4.879496958727125e-07, "loss": 0.7103, "step": 24936 }, { "epoch": 0.9031872509960159, "grad_norm": 0.9840547746462107, "learning_rate": 4.875878000846556e-07, "loss": 0.6101, "step": 24937 }, { "epoch": 0.9032234697573343, "grad_norm": 1.1534759837492705, "learning_rate": 4.872260351957481e-07, "loss": 0.6768, "step": 24938 }, { "epoch": 0.9032596885186527, "grad_norm": 1.0026397411032533, "learning_rate": 4.868644012109703e-07, "loss": 0.6711, "step": 24939 }, { "epoch": 0.903295907279971, "grad_norm": 1.4559246800880297, "learning_rate": 4.86502898135297e-07, "loss": 0.7298, "step": 24940 }, { "epoch": 0.9033321260412894, "grad_norm": 1.2682257739145475, "learning_rate": 4.861415259737035e-07, "loss": 0.6741, "step": 24941 }, { "epoch": 0.9033683448026077, "grad_norm": 1.4732005001533681, "learning_rate": 4.857802847311599e-07, "loss": 0.6747, "step": 24942 }, { "epoch": 0.9034045635639261, "grad_norm": 1.5037247021817195, "learning_rate": 4.854191744126424e-07, "loss": 0.6305, "step": 24943 }, { "epoch": 0.9034407823252445, "grad_norm": 1.5564113361921597, "learning_rate": 4.850581950231159e-07, "loss": 0.676, "step": 24944 }, { "epoch": 0.9034770010865628, "grad_norm": 1.1000115493542244, "learning_rate": 4.846973465675498e-07, "loss": 0.6209, "step": 24945 }, { "epoch": 0.9035132198478812, "grad_norm": 1.0280331305605233, "learning_rate": 4.843366290509077e-07, "loss": 0.7371, "step": 24946 }, { "epoch": 0.9035494386091996, "grad_norm": 1.431816393430444, "learning_rate": 4.839760424781559e-07, "loss": 0.6591, "step": 24947 }, { "epoch": 0.9035856573705179, "grad_norm": 1.327025360789201, "learning_rate": 4.836155868542558e-07, "loss": 0.6955, "step": 24948 }, { "epoch": 0.9036218761318363, "grad_norm": 1.3311455325487682, "learning_rate": 4.832552621841657e-07, "loss": 0.6751, "step": 24949 }, { "epoch": 0.9036580948931546, "grad_norm": 1.4090698295856519, "learning_rate": 4.828950684728462e-07, "loss": 0.7567, "step": 24950 }, { "epoch": 0.9036943136544731, "grad_norm": 1.3274974686837353, "learning_rate": 4.825350057252532e-07, "loss": 0.6926, "step": 24951 }, { "epoch": 0.9037305324157914, "grad_norm": 1.2438327375217983, "learning_rate": 4.821750739463415e-07, "loss": 0.7404, "step": 24952 }, { "epoch": 0.9037667511771097, "grad_norm": 1.069574003195822, "learning_rate": 4.818152731410619e-07, "loss": 0.6493, "step": 24953 }, { "epoch": 0.9038029699384281, "grad_norm": 1.309858703576842, "learning_rate": 4.814556033143691e-07, "loss": 0.5962, "step": 24954 }, { "epoch": 0.9038391886997464, "grad_norm": 1.3330129350836282, "learning_rate": 4.810960644712115e-07, "loss": 0.6251, "step": 24955 }, { "epoch": 0.9038754074610649, "grad_norm": 1.5242406788035052, "learning_rate": 4.807366566165361e-07, "loss": 0.7644, "step": 24956 }, { "epoch": 0.9039116262223832, "grad_norm": 1.1123281332815431, "learning_rate": 4.803773797552869e-07, "loss": 0.6526, "step": 24957 }, { "epoch": 0.9039478449837015, "grad_norm": 1.4846772689242, "learning_rate": 4.800182338924109e-07, "loss": 0.7035, "step": 24958 }, { "epoch": 0.9039840637450199, "grad_norm": 1.342409105595308, "learning_rate": 4.796592190328486e-07, "loss": 0.6424, "step": 24959 }, { "epoch": 0.9040202825063383, "grad_norm": 1.335226628305217, "learning_rate": 4.793003351815406e-07, "loss": 0.6334, "step": 24960 }, { "epoch": 0.9040565012676567, "grad_norm": 1.0410631747774917, "learning_rate": 4.789415823434263e-07, "loss": 0.6346, "step": 24961 }, { "epoch": 0.904092720028975, "grad_norm": 1.4733043213523718, "learning_rate": 4.785829605234416e-07, "loss": 0.6919, "step": 24962 }, { "epoch": 0.9041289387902933, "grad_norm": 1.3813893233346206, "learning_rate": 4.782244697265215e-07, "loss": 0.704, "step": 24963 }, { "epoch": 0.9041651575516118, "grad_norm": 1.396562320027838, "learning_rate": 4.778661099575976e-07, "loss": 0.7517, "step": 24964 }, { "epoch": 0.9042013763129301, "grad_norm": 1.3990870956962527, "learning_rate": 4.77507881221605e-07, "loss": 0.6957, "step": 24965 }, { "epoch": 0.9042375950742485, "grad_norm": 1.4066260020590535, "learning_rate": 4.771497835234718e-07, "loss": 0.6881, "step": 24966 }, { "epoch": 0.9042738138355668, "grad_norm": 1.3038588822105237, "learning_rate": 4.7679181686812406e-07, "loss": 0.7302, "step": 24967 }, { "epoch": 0.9043100325968851, "grad_norm": 1.543768733915759, "learning_rate": 4.764339812604868e-07, "loss": 0.7259, "step": 24968 }, { "epoch": 0.9043462513582036, "grad_norm": 1.1754357434872922, "learning_rate": 4.760762767054883e-07, "loss": 0.7175, "step": 24969 }, { "epoch": 0.9043824701195219, "grad_norm": 1.4842977302918647, "learning_rate": 4.75718703208049e-07, "loss": 0.7355, "step": 24970 }, { "epoch": 0.9044186888808403, "grad_norm": 1.3881773772420842, "learning_rate": 4.753612607730873e-07, "loss": 0.6797, "step": 24971 }, { "epoch": 0.9044549076421586, "grad_norm": 1.015905282886421, "learning_rate": 4.7500394940552585e-07, "loss": 0.6715, "step": 24972 }, { "epoch": 0.904491126403477, "grad_norm": 1.309187925723433, "learning_rate": 4.746467691102785e-07, "loss": 0.5839, "step": 24973 }, { "epoch": 0.9045273451647954, "grad_norm": 1.4215987731048385, "learning_rate": 4.7428971989226136e-07, "loss": 0.6943, "step": 24974 }, { "epoch": 0.9045635639261137, "grad_norm": 1.4102170738248123, "learning_rate": 4.7393280175638825e-07, "loss": 0.6741, "step": 24975 }, { "epoch": 0.9045997826874321, "grad_norm": 1.332146320099162, "learning_rate": 4.7357601470756964e-07, "loss": 0.6336, "step": 24976 }, { "epoch": 0.9046360014487504, "grad_norm": 0.9817887024857755, "learning_rate": 4.7321935875071614e-07, "loss": 0.7091, "step": 24977 }, { "epoch": 0.9046722202100688, "grad_norm": 1.299133907978399, "learning_rate": 4.728628338907326e-07, "loss": 0.6215, "step": 24978 }, { "epoch": 0.9047084389713872, "grad_norm": 1.3759811345789954, "learning_rate": 4.725064401325308e-07, "loss": 0.67, "step": 24979 }, { "epoch": 0.9047446577327055, "grad_norm": 1.0363614452423937, "learning_rate": 4.721501774810111e-07, "loss": 0.6385, "step": 24980 }, { "epoch": 0.904780876494024, "grad_norm": 1.4584478096365652, "learning_rate": 4.7179404594107523e-07, "loss": 0.6592, "step": 24981 }, { "epoch": 0.9048170952553423, "grad_norm": 0.9924799124683685, "learning_rate": 4.714380455176282e-07, "loss": 0.6266, "step": 24982 }, { "epoch": 0.9048533140166606, "grad_norm": 1.241092003887494, "learning_rate": 4.7108217621556487e-07, "loss": 0.6251, "step": 24983 }, { "epoch": 0.904889532777979, "grad_norm": 1.4728332028713573, "learning_rate": 4.707264380397836e-07, "loss": 0.6857, "step": 24984 }, { "epoch": 0.9049257515392973, "grad_norm": 1.0720407071146112, "learning_rate": 4.703708309951793e-07, "loss": 0.692, "step": 24985 }, { "epoch": 0.9049619703006158, "grad_norm": 1.4263080707946976, "learning_rate": 4.700153550866471e-07, "loss": 0.6053, "step": 24986 }, { "epoch": 0.9049981890619341, "grad_norm": 1.3492783983869576, "learning_rate": 4.696600103190763e-07, "loss": 0.7261, "step": 24987 }, { "epoch": 0.9050344078232524, "grad_norm": 1.584170630401731, "learning_rate": 4.693047966973585e-07, "loss": 0.6865, "step": 24988 }, { "epoch": 0.9050706265845708, "grad_norm": 1.554773214507715, "learning_rate": 4.6894971422637993e-07, "loss": 0.6881, "step": 24989 }, { "epoch": 0.9051068453458891, "grad_norm": 1.3316005264896353, "learning_rate": 4.685947629110299e-07, "loss": 0.6205, "step": 24990 }, { "epoch": 0.9051430641072076, "grad_norm": 1.4513806729389003, "learning_rate": 4.682399427561901e-07, "loss": 0.7338, "step": 24991 }, { "epoch": 0.9051792828685259, "grad_norm": 1.4699431985265923, "learning_rate": 4.678852537667422e-07, "loss": 0.6762, "step": 24992 }, { "epoch": 0.9052155016298442, "grad_norm": 1.4316177294775179, "learning_rate": 4.675306959475712e-07, "loss": 0.6661, "step": 24993 }, { "epoch": 0.9052517203911626, "grad_norm": 1.3332053796149481, "learning_rate": 4.671762693035531e-07, "loss": 0.5431, "step": 24994 }, { "epoch": 0.905287939152481, "grad_norm": 1.1281909079704486, "learning_rate": 4.668219738395663e-07, "loss": 0.6727, "step": 24995 }, { "epoch": 0.9053241579137994, "grad_norm": 1.4526185363144497, "learning_rate": 4.6646780956048463e-07, "loss": 0.728, "step": 24996 }, { "epoch": 0.9053603766751177, "grad_norm": 1.3281853977663054, "learning_rate": 4.6611377647118317e-07, "loss": 0.6908, "step": 24997 }, { "epoch": 0.905396595436436, "grad_norm": 1.327812357175536, "learning_rate": 4.657598745765346e-07, "loss": 0.648, "step": 24998 }, { "epoch": 0.9054328141977545, "grad_norm": 1.3746859580634232, "learning_rate": 4.6540610388140616e-07, "loss": 0.7427, "step": 24999 }, { "epoch": 0.9054690329590728, "grad_norm": 1.4101565815404782, "learning_rate": 4.650524643906673e-07, "loss": 0.7166, "step": 25000 }, { "epoch": 0.9055052517203912, "grad_norm": 1.3376368365468125, "learning_rate": 4.6469895610918636e-07, "loss": 0.711, "step": 25001 }, { "epoch": 0.9055414704817095, "grad_norm": 1.3065083326987526, "learning_rate": 4.6434557904182495e-07, "loss": 0.604, "step": 25002 }, { "epoch": 0.9055776892430278, "grad_norm": 0.9973978709256306, "learning_rate": 4.6399233319344703e-07, "loss": 0.6691, "step": 25003 }, { "epoch": 0.9056139080043463, "grad_norm": 1.31323389981228, "learning_rate": 4.6363921856891424e-07, "loss": 0.6945, "step": 25004 }, { "epoch": 0.9056501267656646, "grad_norm": 1.2986438677887064, "learning_rate": 4.632862351730849e-07, "loss": 0.6751, "step": 25005 }, { "epoch": 0.905686345526983, "grad_norm": 1.2531661771551512, "learning_rate": 4.6293338301081737e-07, "loss": 0.6133, "step": 25006 }, { "epoch": 0.9057225642883013, "grad_norm": 1.224336954633605, "learning_rate": 4.625806620869655e-07, "loss": 0.6653, "step": 25007 }, { "epoch": 0.9057587830496197, "grad_norm": 1.4105482864161538, "learning_rate": 4.6222807240638436e-07, "loss": 0.6696, "step": 25008 }, { "epoch": 0.9057950018109381, "grad_norm": 1.3240866581124424, "learning_rate": 4.6187561397392555e-07, "loss": 0.6308, "step": 25009 }, { "epoch": 0.9058312205722564, "grad_norm": 1.2881890480528126, "learning_rate": 4.615232867944397e-07, "loss": 0.7092, "step": 25010 }, { "epoch": 0.9058674393335748, "grad_norm": 1.3823682548275775, "learning_rate": 4.61171090872774e-07, "loss": 0.7191, "step": 25011 }, { "epoch": 0.9059036580948931, "grad_norm": 1.0235888173855832, "learning_rate": 4.6081902621377574e-07, "loss": 0.6456, "step": 25012 }, { "epoch": 0.9059398768562115, "grad_norm": 1.3059363189205218, "learning_rate": 4.604670928222899e-07, "loss": 0.7194, "step": 25013 }, { "epoch": 0.9059760956175299, "grad_norm": 1.35790402816712, "learning_rate": 4.601152907031581e-07, "loss": 0.669, "step": 25014 }, { "epoch": 0.9060123143788482, "grad_norm": 1.322633285063974, "learning_rate": 4.597636198612232e-07, "loss": 0.6529, "step": 25015 }, { "epoch": 0.9060485331401666, "grad_norm": 1.4916211296657258, "learning_rate": 4.5941208030132465e-07, "loss": 0.7175, "step": 25016 }, { "epoch": 0.906084751901485, "grad_norm": 0.9948426343797905, "learning_rate": 4.590606720282975e-07, "loss": 0.6436, "step": 25017 }, { "epoch": 0.9061209706628033, "grad_norm": 1.3167418263023511, "learning_rate": 4.58709395046979e-07, "loss": 0.5908, "step": 25018 }, { "epoch": 0.9061571894241217, "grad_norm": 1.435006601820191, "learning_rate": 4.5835824936220297e-07, "loss": 0.6608, "step": 25019 }, { "epoch": 0.90619340818544, "grad_norm": 1.3377106789190767, "learning_rate": 4.580072349788023e-07, "loss": 0.61, "step": 25020 }, { "epoch": 0.9062296269467585, "grad_norm": 1.1470199510149917, "learning_rate": 4.576563519016064e-07, "loss": 0.6704, "step": 25021 }, { "epoch": 0.9062658457080768, "grad_norm": 1.4503297814085736, "learning_rate": 4.5730560013544256e-07, "loss": 0.6853, "step": 25022 }, { "epoch": 0.9063020644693951, "grad_norm": 1.438104099602997, "learning_rate": 4.5695497968513915e-07, "loss": 0.6884, "step": 25023 }, { "epoch": 0.9063382832307135, "grad_norm": 1.358643011348412, "learning_rate": 4.5660449055552114e-07, "loss": 0.7144, "step": 25024 }, { "epoch": 0.9063745019920318, "grad_norm": 1.3644602993258608, "learning_rate": 4.5625413275140915e-07, "loss": 0.6128, "step": 25025 }, { "epoch": 0.9064107207533503, "grad_norm": 1.4656779825642432, "learning_rate": 4.5590390627762826e-07, "loss": 0.7208, "step": 25026 }, { "epoch": 0.9064469395146686, "grad_norm": 1.315751120724401, "learning_rate": 4.5555381113899454e-07, "loss": 0.6766, "step": 25027 }, { "epoch": 0.9064831582759869, "grad_norm": 1.611821661995192, "learning_rate": 4.5520384734032755e-07, "loss": 0.7376, "step": 25028 }, { "epoch": 0.9065193770373053, "grad_norm": 1.4261421935566958, "learning_rate": 4.548540148864411e-07, "loss": 0.6763, "step": 25029 }, { "epoch": 0.9065555957986237, "grad_norm": 1.14588705721715, "learning_rate": 4.5450431378215253e-07, "loss": 0.6335, "step": 25030 }, { "epoch": 0.9065918145599421, "grad_norm": 1.3416417863774492, "learning_rate": 4.5415474403227135e-07, "loss": 0.632, "step": 25031 }, { "epoch": 0.9066280333212604, "grad_norm": 1.2931817691767808, "learning_rate": 4.5380530564160806e-07, "loss": 0.6203, "step": 25032 }, { "epoch": 0.9066642520825787, "grad_norm": 1.3395292243799815, "learning_rate": 4.5345599861497115e-07, "loss": 0.6183, "step": 25033 }, { "epoch": 0.9067004708438972, "grad_norm": 1.2691328773485133, "learning_rate": 4.531068229571689e-07, "loss": 0.6503, "step": 25034 }, { "epoch": 0.9067366896052155, "grad_norm": 1.325745067260046, "learning_rate": 4.527577786730053e-07, "loss": 0.6633, "step": 25035 }, { "epoch": 0.9067729083665339, "grad_norm": 0.9765596638280691, "learning_rate": 4.524088657672832e-07, "loss": 0.6889, "step": 25036 }, { "epoch": 0.9068091271278522, "grad_norm": 0.9565987823227964, "learning_rate": 4.520600842448053e-07, "loss": 0.6819, "step": 25037 }, { "epoch": 0.9068453458891705, "grad_norm": 1.2725865791691493, "learning_rate": 4.517114341103701e-07, "loss": 0.5508, "step": 25038 }, { "epoch": 0.906881564650489, "grad_norm": 1.2239157539090233, "learning_rate": 4.513629153687749e-07, "loss": 0.6622, "step": 25039 }, { "epoch": 0.9069177834118073, "grad_norm": 1.2334795360376525, "learning_rate": 4.510145280248157e-07, "loss": 0.6339, "step": 25040 }, { "epoch": 0.9069540021731257, "grad_norm": 1.4056100634344844, "learning_rate": 4.506662720832877e-07, "loss": 0.6256, "step": 25041 }, { "epoch": 0.906990220934444, "grad_norm": 1.5088154429379415, "learning_rate": 4.503181475489826e-07, "loss": 0.675, "step": 25042 }, { "epoch": 0.9070264396957624, "grad_norm": 0.9891210529870743, "learning_rate": 4.499701544266899e-07, "loss": 0.6696, "step": 25043 }, { "epoch": 0.9070626584570808, "grad_norm": 1.5525942589261241, "learning_rate": 4.496222927212002e-07, "loss": 0.7358, "step": 25044 }, { "epoch": 0.9070988772183991, "grad_norm": 1.4263284469289041, "learning_rate": 4.4927456243729964e-07, "loss": 0.6909, "step": 25045 }, { "epoch": 0.9071350959797175, "grad_norm": 1.6014644999875618, "learning_rate": 4.489269635797733e-07, "loss": 0.7, "step": 25046 }, { "epoch": 0.9071713147410359, "grad_norm": 1.5742150346887256, "learning_rate": 4.485794961534029e-07, "loss": 0.6773, "step": 25047 }, { "epoch": 0.9072075335023542, "grad_norm": 1.3726145896766537, "learning_rate": 4.4823216016297246e-07, "loss": 0.6942, "step": 25048 }, { "epoch": 0.9072437522636726, "grad_norm": 1.5558968170529939, "learning_rate": 4.478849556132603e-07, "loss": 0.6966, "step": 25049 }, { "epoch": 0.9072799710249909, "grad_norm": 1.3951881121236485, "learning_rate": 4.475378825090448e-07, "loss": 0.6591, "step": 25050 }, { "epoch": 0.9073161897863093, "grad_norm": 1.1067759970528157, "learning_rate": 4.471909408551001e-07, "loss": 0.6832, "step": 25051 }, { "epoch": 0.9073524085476277, "grad_norm": 0.8916959872846274, "learning_rate": 4.4684413065620325e-07, "loss": 0.6553, "step": 25052 }, { "epoch": 0.907388627308946, "grad_norm": 0.9795910527698447, "learning_rate": 4.4649745191712503e-07, "loss": 0.672, "step": 25053 }, { "epoch": 0.9074248460702644, "grad_norm": 1.271103668643116, "learning_rate": 4.461509046426349e-07, "loss": 0.6054, "step": 25054 }, { "epoch": 0.9074610648315827, "grad_norm": 1.4381525907570019, "learning_rate": 4.458044888375046e-07, "loss": 0.6946, "step": 25055 }, { "epoch": 0.9074972835929012, "grad_norm": 1.3589171420936457, "learning_rate": 4.4545820450649925e-07, "loss": 0.6468, "step": 25056 }, { "epoch": 0.9075335023542195, "grad_norm": 1.3973946345692982, "learning_rate": 4.4511205165438496e-07, "loss": 0.661, "step": 25057 }, { "epoch": 0.9075697211155378, "grad_norm": 1.1470524332649437, "learning_rate": 4.4476603028592246e-07, "loss": 0.7406, "step": 25058 }, { "epoch": 0.9076059398768562, "grad_norm": 1.3727143378090123, "learning_rate": 4.444201404058768e-07, "loss": 0.6109, "step": 25059 }, { "epoch": 0.9076421586381745, "grad_norm": 0.9533507933603473, "learning_rate": 4.440743820190074e-07, "loss": 0.6903, "step": 25060 }, { "epoch": 0.907678377399493, "grad_norm": 0.9698501771154766, "learning_rate": 4.4372875513006953e-07, "loss": 0.6956, "step": 25061 }, { "epoch": 0.9077145961608113, "grad_norm": 1.4726740433809449, "learning_rate": 4.4338325974382035e-07, "loss": 0.6541, "step": 25062 }, { "epoch": 0.9077508149221296, "grad_norm": 1.2873942873145015, "learning_rate": 4.430378958650161e-07, "loss": 0.6172, "step": 25063 }, { "epoch": 0.907787033683448, "grad_norm": 1.240366499163376, "learning_rate": 4.4269266349840745e-07, "loss": 0.6181, "step": 25064 }, { "epoch": 0.9078232524447664, "grad_norm": 1.3077644481741377, "learning_rate": 4.42347562648745e-07, "loss": 0.6246, "step": 25065 }, { "epoch": 0.9078594712060848, "grad_norm": 0.9432045892156712, "learning_rate": 4.420025933207783e-07, "loss": 0.6545, "step": 25066 }, { "epoch": 0.9078956899674031, "grad_norm": 1.3734265130908778, "learning_rate": 4.416577555192558e-07, "loss": 0.6476, "step": 25067 }, { "epoch": 0.9079319087287214, "grad_norm": 1.3407986646823142, "learning_rate": 4.413130492489204e-07, "loss": 0.6208, "step": 25068 }, { "epoch": 0.9079681274900399, "grad_norm": 1.1694325842783322, "learning_rate": 4.4096847451451485e-07, "loss": 0.6844, "step": 25069 }, { "epoch": 0.9080043462513582, "grad_norm": 1.2919885710292958, "learning_rate": 4.4062403132078434e-07, "loss": 0.6794, "step": 25070 }, { "epoch": 0.9080405650126766, "grad_norm": 1.0399505127565307, "learning_rate": 4.402797196724662e-07, "loss": 0.6564, "step": 25071 }, { "epoch": 0.9080767837739949, "grad_norm": 1.3307258916343008, "learning_rate": 4.399355395742988e-07, "loss": 0.6184, "step": 25072 }, { "epoch": 0.9081130025353132, "grad_norm": 1.4279477903178386, "learning_rate": 4.395914910310184e-07, "loss": 0.6774, "step": 25073 }, { "epoch": 0.9081492212966317, "grad_norm": 1.3347508928693312, "learning_rate": 4.3924757404736005e-07, "loss": 0.6939, "step": 25074 }, { "epoch": 0.90818544005795, "grad_norm": 1.3608147824501338, "learning_rate": 4.3890378862805563e-07, "loss": 0.7039, "step": 25075 }, { "epoch": 0.9082216588192684, "grad_norm": 1.5013163038761974, "learning_rate": 4.3856013477783564e-07, "loss": 0.6793, "step": 25076 }, { "epoch": 0.9082578775805867, "grad_norm": 1.5901069001558437, "learning_rate": 4.3821661250143086e-07, "loss": 0.7183, "step": 25077 }, { "epoch": 0.908294096341905, "grad_norm": 1.5268062538483926, "learning_rate": 4.3787322180356753e-07, "loss": 0.6799, "step": 25078 }, { "epoch": 0.9083303151032235, "grad_norm": 1.231506155049661, "learning_rate": 4.375299626889695e-07, "loss": 0.6528, "step": 25079 }, { "epoch": 0.9083665338645418, "grad_norm": 1.3726895438484183, "learning_rate": 4.3718683516236094e-07, "loss": 0.6613, "step": 25080 }, { "epoch": 0.9084027526258602, "grad_norm": 1.4233250579589463, "learning_rate": 4.3684383922846574e-07, "loss": 0.6089, "step": 25081 }, { "epoch": 0.9084389713871786, "grad_norm": 1.5159097961881773, "learning_rate": 4.3650097489200125e-07, "loss": 0.686, "step": 25082 }, { "epoch": 0.9084751901484969, "grad_norm": 1.113570419958599, "learning_rate": 4.361582421576871e-07, "loss": 0.6678, "step": 25083 }, { "epoch": 0.9085114089098153, "grad_norm": 0.9505265506127175, "learning_rate": 4.3581564103023835e-07, "loss": 0.647, "step": 25084 }, { "epoch": 0.9085476276711336, "grad_norm": 1.117708663182801, "learning_rate": 4.3547317151437117e-07, "loss": 0.6597, "step": 25085 }, { "epoch": 0.908583846432452, "grad_norm": 1.3888490499471682, "learning_rate": 4.351308336147964e-07, "loss": 0.6852, "step": 25086 }, { "epoch": 0.9086200651937704, "grad_norm": 1.2400688248627285, "learning_rate": 4.3478862733622564e-07, "loss": 0.6893, "step": 25087 }, { "epoch": 0.9086562839550887, "grad_norm": 1.3405495291698686, "learning_rate": 4.3444655268336857e-07, "loss": 0.671, "step": 25088 }, { "epoch": 0.9086925027164071, "grad_norm": 1.3169364604279168, "learning_rate": 4.341046096609325e-07, "loss": 0.6501, "step": 25089 }, { "epoch": 0.9087287214777254, "grad_norm": 1.4474778836953976, "learning_rate": 4.3376279827362253e-07, "loss": 0.63, "step": 25090 }, { "epoch": 0.9087649402390439, "grad_norm": 1.0581462276124751, "learning_rate": 4.3342111852614056e-07, "loss": 0.6694, "step": 25091 }, { "epoch": 0.9088011590003622, "grad_norm": 1.6166318305051999, "learning_rate": 4.330795704231905e-07, "loss": 0.6979, "step": 25092 }, { "epoch": 0.9088373777616805, "grad_norm": 1.1268842694082917, "learning_rate": 4.3273815396947306e-07, "loss": 0.7115, "step": 25093 }, { "epoch": 0.9088735965229989, "grad_norm": 1.3063847500825334, "learning_rate": 4.3239686916968336e-07, "loss": 0.6166, "step": 25094 }, { "epoch": 0.9089098152843172, "grad_norm": 1.2882322816812892, "learning_rate": 4.320557160285199e-07, "loss": 0.6524, "step": 25095 }, { "epoch": 0.9089460340456357, "grad_norm": 1.2941460047137094, "learning_rate": 4.3171469455067785e-07, "loss": 0.6711, "step": 25096 }, { "epoch": 0.908982252806954, "grad_norm": 1.4855165358601763, "learning_rate": 4.313738047408489e-07, "loss": 0.7234, "step": 25097 }, { "epoch": 0.9090184715682723, "grad_norm": 1.477911312916131, "learning_rate": 4.310330466037227e-07, "loss": 0.6862, "step": 25098 }, { "epoch": 0.9090546903295907, "grad_norm": 1.1666930314012753, "learning_rate": 4.3069242014399104e-07, "loss": 0.621, "step": 25099 }, { "epoch": 0.9090909090909091, "grad_norm": 1.4640832630830694, "learning_rate": 4.303519253663402e-07, "loss": 0.6279, "step": 25100 }, { "epoch": 0.9091271278522275, "grad_norm": 1.179409896401151, "learning_rate": 4.3001156227545636e-07, "loss": 0.6277, "step": 25101 }, { "epoch": 0.9091633466135458, "grad_norm": 1.1176064834551787, "learning_rate": 4.296713308760203e-07, "loss": 0.6679, "step": 25102 }, { "epoch": 0.9091995653748641, "grad_norm": 1.070953037000047, "learning_rate": 4.293312311727171e-07, "loss": 0.6946, "step": 25103 }, { "epoch": 0.9092357841361826, "grad_norm": 1.3006608165514066, "learning_rate": 4.289912631702253e-07, "loss": 0.6303, "step": 25104 }, { "epoch": 0.9092720028975009, "grad_norm": 1.4142833966597697, "learning_rate": 4.2865142687322225e-07, "loss": 0.6451, "step": 25105 }, { "epoch": 0.9093082216588193, "grad_norm": 1.573597983223065, "learning_rate": 4.283117222863875e-07, "loss": 0.6864, "step": 25106 }, { "epoch": 0.9093444404201376, "grad_norm": 1.2893755937141314, "learning_rate": 4.2797214941439403e-07, "loss": 0.5877, "step": 25107 }, { "epoch": 0.9093806591814559, "grad_norm": 0.8782651227485879, "learning_rate": 4.276327082619147e-07, "loss": 0.6404, "step": 25108 }, { "epoch": 0.9094168779427744, "grad_norm": 1.3483808835985933, "learning_rate": 4.2729339883361806e-07, "loss": 0.6713, "step": 25109 }, { "epoch": 0.9094530967040927, "grad_norm": 1.313043712355227, "learning_rate": 4.269542211341782e-07, "loss": 0.6547, "step": 25110 }, { "epoch": 0.9094893154654111, "grad_norm": 1.2484118460763474, "learning_rate": 4.2661517516825903e-07, "loss": 0.6516, "step": 25111 }, { "epoch": 0.9095255342267294, "grad_norm": 1.3776766249296633, "learning_rate": 4.2627626094052687e-07, "loss": 0.6281, "step": 25112 }, { "epoch": 0.9095617529880478, "grad_norm": 1.5275961125954136, "learning_rate": 4.259374784556447e-07, "loss": 0.6597, "step": 25113 }, { "epoch": 0.9095979717493662, "grad_norm": 1.4494336256838443, "learning_rate": 4.2559882771827545e-07, "loss": 0.6655, "step": 25114 }, { "epoch": 0.9096341905106845, "grad_norm": 1.8708936349337417, "learning_rate": 4.252603087330809e-07, "loss": 0.6763, "step": 25115 }, { "epoch": 0.9096704092720029, "grad_norm": 1.0089727671951843, "learning_rate": 4.249219215047151e-07, "loss": 0.7089, "step": 25116 }, { "epoch": 0.9097066280333213, "grad_norm": 1.433018693782552, "learning_rate": 4.2458366603783883e-07, "loss": 0.6169, "step": 25117 }, { "epoch": 0.9097428467946396, "grad_norm": 1.3286604615301818, "learning_rate": 4.24245542337105e-07, "loss": 0.6612, "step": 25118 }, { "epoch": 0.909779065555958, "grad_norm": 1.450182101000065, "learning_rate": 4.2390755040716545e-07, "loss": 0.6714, "step": 25119 }, { "epoch": 0.9098152843172763, "grad_norm": 1.08799529465626, "learning_rate": 4.23569690252672e-07, "loss": 0.6894, "step": 25120 }, { "epoch": 0.9098515030785947, "grad_norm": 1.1079553373569795, "learning_rate": 4.232319618782754e-07, "loss": 0.6728, "step": 25121 }, { "epoch": 0.9098877218399131, "grad_norm": 1.298865241760869, "learning_rate": 4.228943652886219e-07, "loss": 0.6409, "step": 25122 }, { "epoch": 0.9099239406012314, "grad_norm": 1.3019997085102732, "learning_rate": 4.225569004883578e-07, "loss": 0.6659, "step": 25123 }, { "epoch": 0.9099601593625498, "grad_norm": 4.718737301430447, "learning_rate": 4.222195674821239e-07, "loss": 0.7041, "step": 25124 }, { "epoch": 0.9099963781238681, "grad_norm": 1.2735131824856813, "learning_rate": 4.2188236627456633e-07, "loss": 0.606, "step": 25125 }, { "epoch": 0.9100325968851866, "grad_norm": 1.385389184245982, "learning_rate": 4.2154529687032375e-07, "loss": 0.6448, "step": 25126 }, { "epoch": 0.9100688156465049, "grad_norm": 1.2278788400958731, "learning_rate": 4.2120835927403234e-07, "loss": 0.6734, "step": 25127 }, { "epoch": 0.9101050344078232, "grad_norm": 1.0537820167626244, "learning_rate": 4.208715534903318e-07, "loss": 0.6251, "step": 25128 }, { "epoch": 0.9101412531691416, "grad_norm": 1.4095975589378524, "learning_rate": 4.205348795238562e-07, "loss": 0.6692, "step": 25129 }, { "epoch": 0.91017747193046, "grad_norm": 1.3171031384982554, "learning_rate": 4.201983373792373e-07, "loss": 0.5488, "step": 25130 }, { "epoch": 0.9102136906917784, "grad_norm": 1.5971587975896742, "learning_rate": 4.1986192706110597e-07, "loss": 0.6713, "step": 25131 }, { "epoch": 0.9102499094530967, "grad_norm": 1.127133395624771, "learning_rate": 4.1952564857409393e-07, "loss": 0.6535, "step": 25132 }, { "epoch": 0.910286128214415, "grad_norm": 1.0681759521665255, "learning_rate": 4.191895019228265e-07, "loss": 0.7059, "step": 25133 }, { "epoch": 0.9103223469757334, "grad_norm": 1.5420090382714036, "learning_rate": 4.1885348711192986e-07, "loss": 0.6835, "step": 25134 }, { "epoch": 0.9103585657370518, "grad_norm": 1.3919280669676037, "learning_rate": 4.1851760414602706e-07, "loss": 0.7206, "step": 25135 }, { "epoch": 0.9103947844983702, "grad_norm": 1.3930174146742582, "learning_rate": 4.1818185302974213e-07, "loss": 0.6208, "step": 25136 }, { "epoch": 0.9104310032596885, "grad_norm": 0.8549569108044185, "learning_rate": 4.1784623376769475e-07, "loss": 0.6614, "step": 25137 }, { "epoch": 0.9104672220210068, "grad_norm": 1.313996122194201, "learning_rate": 4.175107463645012e-07, "loss": 0.6241, "step": 25138 }, { "epoch": 0.9105034407823253, "grad_norm": 1.1251633529853742, "learning_rate": 4.171753908247811e-07, "loss": 0.6392, "step": 25139 }, { "epoch": 0.9105396595436436, "grad_norm": 1.3554421025111105, "learning_rate": 4.1684016715314746e-07, "loss": 0.6593, "step": 25140 }, { "epoch": 0.910575878304962, "grad_norm": 0.9164117826574154, "learning_rate": 4.1650507535421435e-07, "loss": 0.6551, "step": 25141 }, { "epoch": 0.9106120970662803, "grad_norm": 1.525490033455371, "learning_rate": 4.1617011543259036e-07, "loss": 0.6872, "step": 25142 }, { "epoch": 0.9106483158275986, "grad_norm": 1.3655771860627754, "learning_rate": 4.1583528739288723e-07, "loss": 0.6986, "step": 25143 }, { "epoch": 0.9106845345889171, "grad_norm": 1.2591167064978903, "learning_rate": 4.1550059123971256e-07, "loss": 0.6854, "step": 25144 }, { "epoch": 0.9107207533502354, "grad_norm": 1.3068993216818692, "learning_rate": 4.151660269776725e-07, "loss": 0.6104, "step": 25145 }, { "epoch": 0.9107569721115538, "grad_norm": 0.9825841723887467, "learning_rate": 4.148315946113668e-07, "loss": 0.7247, "step": 25146 }, { "epoch": 0.9107931908728721, "grad_norm": 1.3518763485104628, "learning_rate": 4.1449729414540283e-07, "loss": 0.6497, "step": 25147 }, { "epoch": 0.9108294096341905, "grad_norm": 1.3061506667081815, "learning_rate": 4.141631255843792e-07, "loss": 0.6181, "step": 25148 }, { "epoch": 0.9108656283955089, "grad_norm": 1.1960084454175144, "learning_rate": 4.138290889328922e-07, "loss": 0.6699, "step": 25149 }, { "epoch": 0.9109018471568272, "grad_norm": 1.2224280693638157, "learning_rate": 4.1349518419554035e-07, "loss": 0.6377, "step": 25150 }, { "epoch": 0.9109380659181456, "grad_norm": 1.5824299156947066, "learning_rate": 4.1316141137691887e-07, "loss": 0.6244, "step": 25151 }, { "epoch": 0.910974284679464, "grad_norm": 1.3625395575793944, "learning_rate": 4.128277704816208e-07, "loss": 0.6806, "step": 25152 }, { "epoch": 0.9110105034407823, "grad_norm": 1.1307282864424004, "learning_rate": 4.1249426151423466e-07, "loss": 0.6293, "step": 25153 }, { "epoch": 0.9110467222021007, "grad_norm": 0.9929768241115395, "learning_rate": 4.1216088447935345e-07, "loss": 0.7022, "step": 25154 }, { "epoch": 0.911082940963419, "grad_norm": 1.3271643958349808, "learning_rate": 4.1182763938156236e-07, "loss": 0.6548, "step": 25155 }, { "epoch": 0.9111191597247374, "grad_norm": 1.4654722696407052, "learning_rate": 4.114945262254466e-07, "loss": 0.6577, "step": 25156 }, { "epoch": 0.9111553784860558, "grad_norm": 1.4313276008532956, "learning_rate": 4.111615450155937e-07, "loss": 0.7271, "step": 25157 }, { "epoch": 0.9111915972473741, "grad_norm": 1.3934795697556477, "learning_rate": 4.108286957565821e-07, "loss": 0.6978, "step": 25158 }, { "epoch": 0.9112278160086925, "grad_norm": 1.3257852757695703, "learning_rate": 4.104959784529949e-07, "loss": 0.6306, "step": 25159 }, { "epoch": 0.9112640347700108, "grad_norm": 1.3965418363633557, "learning_rate": 4.101633931094062e-07, "loss": 0.6736, "step": 25160 }, { "epoch": 0.9113002535313293, "grad_norm": 1.304467009141321, "learning_rate": 4.098309397303979e-07, "loss": 0.5686, "step": 25161 }, { "epoch": 0.9113364722926476, "grad_norm": 1.2528800373337838, "learning_rate": 4.0949861832054297e-07, "loss": 0.6445, "step": 25162 }, { "epoch": 0.9113726910539659, "grad_norm": 1.459780886720617, "learning_rate": 4.0916642888441327e-07, "loss": 0.661, "step": 25163 }, { "epoch": 0.9114089098152843, "grad_norm": 1.4397913795464097, "learning_rate": 4.088343714265797e-07, "loss": 0.6916, "step": 25164 }, { "epoch": 0.9114451285766026, "grad_norm": 1.494471437321189, "learning_rate": 4.0850244595161294e-07, "loss": 0.6271, "step": 25165 }, { "epoch": 0.9114813473379211, "grad_norm": 1.5544378910534697, "learning_rate": 4.081706524640816e-07, "loss": 0.7895, "step": 25166 }, { "epoch": 0.9115175660992394, "grad_norm": 1.5000791950975974, "learning_rate": 4.0783899096854874e-07, "loss": 0.675, "step": 25167 }, { "epoch": 0.9115537848605577, "grad_norm": 1.4217750925554185, "learning_rate": 4.0750746146958064e-07, "loss": 0.6538, "step": 25168 }, { "epoch": 0.9115900036218761, "grad_norm": 1.3932370784273436, "learning_rate": 4.0717606397173926e-07, "loss": 0.6161, "step": 25169 }, { "epoch": 0.9116262223831945, "grad_norm": 1.4522876044302158, "learning_rate": 4.0684479847958423e-07, "loss": 0.6616, "step": 25170 }, { "epoch": 0.9116624411445129, "grad_norm": 1.561360484883765, "learning_rate": 4.06513664997672e-07, "loss": 0.6345, "step": 25171 }, { "epoch": 0.9116986599058312, "grad_norm": 1.123908067954448, "learning_rate": 4.0618266353056325e-07, "loss": 0.6639, "step": 25172 }, { "epoch": 0.9117348786671495, "grad_norm": 1.3622533521926885, "learning_rate": 4.0585179408280996e-07, "loss": 0.7348, "step": 25173 }, { "epoch": 0.911771097428468, "grad_norm": 1.2396470866459648, "learning_rate": 4.0552105665896735e-07, "loss": 0.7076, "step": 25174 }, { "epoch": 0.9118073161897863, "grad_norm": 1.1925980757708847, "learning_rate": 4.0519045126358293e-07, "loss": 0.6555, "step": 25175 }, { "epoch": 0.9118435349511047, "grad_norm": 0.8739362317207042, "learning_rate": 4.0485997790121077e-07, "loss": 0.6181, "step": 25176 }, { "epoch": 0.911879753712423, "grad_norm": 1.235183919210497, "learning_rate": 4.045296365763962e-07, "loss": 0.718, "step": 25177 }, { "epoch": 0.9119159724737413, "grad_norm": 1.3785378332719151, "learning_rate": 4.0419942729368445e-07, "loss": 0.629, "step": 25178 }, { "epoch": 0.9119521912350598, "grad_norm": 1.2616904600820538, "learning_rate": 4.0386935005762074e-07, "loss": 0.6869, "step": 25179 }, { "epoch": 0.9119884099963781, "grad_norm": 1.4114474859794466, "learning_rate": 4.03539404872747e-07, "loss": 0.6445, "step": 25180 }, { "epoch": 0.9120246287576965, "grad_norm": 1.032992122448137, "learning_rate": 4.03209591743603e-07, "loss": 0.6442, "step": 25181 }, { "epoch": 0.9120608475190148, "grad_norm": 1.1847521309425297, "learning_rate": 4.0287991067472607e-07, "loss": 0.6836, "step": 25182 }, { "epoch": 0.9120970662803332, "grad_norm": 1.6758288441192741, "learning_rate": 4.0255036167065607e-07, "loss": 0.7433, "step": 25183 }, { "epoch": 0.9121332850416516, "grad_norm": 0.9915520574336616, "learning_rate": 4.02220944735926e-07, "loss": 0.7034, "step": 25184 }, { "epoch": 0.9121695038029699, "grad_norm": 1.0269778829599174, "learning_rate": 4.0189165987506883e-07, "loss": 0.7376, "step": 25185 }, { "epoch": 0.9122057225642883, "grad_norm": 1.5319110090777468, "learning_rate": 4.015625070926155e-07, "loss": 0.646, "step": 25186 }, { "epoch": 0.9122419413256067, "grad_norm": 1.4033184887708947, "learning_rate": 4.0123348639309665e-07, "loss": 0.7227, "step": 25187 }, { "epoch": 0.912278160086925, "grad_norm": 1.2675237761544464, "learning_rate": 4.0090459778103997e-07, "loss": 0.642, "step": 25188 }, { "epoch": 0.9123143788482434, "grad_norm": 1.4391981064609416, "learning_rate": 4.0057584126096837e-07, "loss": 0.7355, "step": 25189 }, { "epoch": 0.9123505976095617, "grad_norm": 1.1274606601412136, "learning_rate": 4.002472168374094e-07, "loss": 0.686, "step": 25190 }, { "epoch": 0.9123868163708801, "grad_norm": 1.3809615344252044, "learning_rate": 3.999187245148839e-07, "loss": 0.6485, "step": 25191 }, { "epoch": 0.9124230351321985, "grad_norm": 1.4133250006114841, "learning_rate": 3.995903642979104e-07, "loss": 0.644, "step": 25192 }, { "epoch": 0.9124592538935168, "grad_norm": 1.269302444380047, "learning_rate": 3.9926213619101097e-07, "loss": 0.6695, "step": 25193 }, { "epoch": 0.9124954726548352, "grad_norm": 1.2128763792142745, "learning_rate": 3.989340401987007e-07, "loss": 0.6936, "step": 25194 }, { "epoch": 0.9125316914161535, "grad_norm": 1.318956127034854, "learning_rate": 3.9860607632549396e-07, "loss": 0.6943, "step": 25195 }, { "epoch": 0.912567910177472, "grad_norm": 1.4175900434348812, "learning_rate": 3.982782445759026e-07, "loss": 0.7035, "step": 25196 }, { "epoch": 0.9126041289387903, "grad_norm": 1.3856585869226454, "learning_rate": 3.9795054495444075e-07, "loss": 0.6677, "step": 25197 }, { "epoch": 0.9126403477001087, "grad_norm": 1.1442320416496798, "learning_rate": 3.9762297746561706e-07, "loss": 0.687, "step": 25198 }, { "epoch": 0.912676566461427, "grad_norm": 1.3726209784423131, "learning_rate": 3.9729554211393686e-07, "loss": 0.6265, "step": 25199 }, { "epoch": 0.9127127852227453, "grad_norm": 1.2180643140558804, "learning_rate": 3.9696823890390867e-07, "loss": 0.7194, "step": 25200 }, { "epoch": 0.9127490039840638, "grad_norm": 0.9249523276864684, "learning_rate": 3.9664106784003565e-07, "loss": 0.6758, "step": 25201 }, { "epoch": 0.9127852227453821, "grad_norm": 1.297841272248749, "learning_rate": 3.963140289268208e-07, "loss": 0.7029, "step": 25202 }, { "epoch": 0.9128214415067005, "grad_norm": 1.402769447840009, "learning_rate": 3.9598712216876054e-07, "loss": 0.6688, "step": 25203 }, { "epoch": 0.9128576602680188, "grad_norm": 1.411054103464585, "learning_rate": 3.956603475703591e-07, "loss": 0.65, "step": 25204 }, { "epoch": 0.9128938790293372, "grad_norm": 0.9770727447040954, "learning_rate": 3.953337051361095e-07, "loss": 0.6467, "step": 25205 }, { "epoch": 0.9129300977906556, "grad_norm": 1.3678063320457048, "learning_rate": 3.9500719487050696e-07, "loss": 0.6821, "step": 25206 }, { "epoch": 0.9129663165519739, "grad_norm": 1.290358274327652, "learning_rate": 3.946808167780447e-07, "loss": 0.6103, "step": 25207 }, { "epoch": 0.9130025353132923, "grad_norm": 1.3428300838659688, "learning_rate": 3.943545708632157e-07, "loss": 0.5889, "step": 25208 }, { "epoch": 0.9130387540746107, "grad_norm": 1.3413070002415497, "learning_rate": 3.940284571305075e-07, "loss": 0.6782, "step": 25209 }, { "epoch": 0.913074972835929, "grad_norm": 1.4917452653686207, "learning_rate": 3.9370247558440765e-07, "loss": 0.6717, "step": 25210 }, { "epoch": 0.9131111915972474, "grad_norm": 1.2000134895748198, "learning_rate": 3.9337662622940366e-07, "loss": 0.6767, "step": 25211 }, { "epoch": 0.9131474103585657, "grad_norm": 1.4161531631497029, "learning_rate": 3.930509090699786e-07, "loss": 0.6853, "step": 25212 }, { "epoch": 0.9131836291198842, "grad_norm": 1.382246597385246, "learning_rate": 3.927253241106144e-07, "loss": 0.6383, "step": 25213 }, { "epoch": 0.9132198478812025, "grad_norm": 1.32960262215393, "learning_rate": 3.9239987135578973e-07, "loss": 0.7006, "step": 25214 }, { "epoch": 0.9132560666425208, "grad_norm": 1.489018477922956, "learning_rate": 3.9207455080998545e-07, "loss": 0.719, "step": 25215 }, { "epoch": 0.9132922854038392, "grad_norm": 1.3004636534320917, "learning_rate": 3.9174936247767916e-07, "loss": 0.6957, "step": 25216 }, { "epoch": 0.9133285041651575, "grad_norm": 1.356555780426199, "learning_rate": 3.914243063633427e-07, "loss": 0.6749, "step": 25217 }, { "epoch": 0.913364722926476, "grad_norm": 0.9625063786241107, "learning_rate": 3.910993824714504e-07, "loss": 0.6831, "step": 25218 }, { "epoch": 0.9134009416877943, "grad_norm": 1.3484344495371927, "learning_rate": 3.907745908064742e-07, "loss": 0.67, "step": 25219 }, { "epoch": 0.9134371604491126, "grad_norm": 1.413341520193612, "learning_rate": 3.904499313728838e-07, "loss": 0.6822, "step": 25220 }, { "epoch": 0.913473379210431, "grad_norm": 1.4126095327912493, "learning_rate": 3.9012540417514457e-07, "loss": 0.6504, "step": 25221 }, { "epoch": 0.9135095979717494, "grad_norm": 1.2864542156190906, "learning_rate": 3.898010092177251e-07, "loss": 0.6756, "step": 25222 }, { "epoch": 0.9135458167330678, "grad_norm": 1.369805728833525, "learning_rate": 3.894767465050875e-07, "loss": 0.6707, "step": 25223 }, { "epoch": 0.9135820354943861, "grad_norm": 1.4741377562635936, "learning_rate": 3.8915261604169475e-07, "loss": 0.7024, "step": 25224 }, { "epoch": 0.9136182542557044, "grad_norm": 1.5236959886112205, "learning_rate": 3.8882861783200556e-07, "loss": 0.7566, "step": 25225 }, { "epoch": 0.9136544730170229, "grad_norm": 1.9482771562005088, "learning_rate": 3.8850475188048074e-07, "loss": 0.666, "step": 25226 }, { "epoch": 0.9136906917783412, "grad_norm": 1.3948177135582647, "learning_rate": 3.8818101819157684e-07, "loss": 0.6631, "step": 25227 }, { "epoch": 0.9137269105396596, "grad_norm": 1.453757111414312, "learning_rate": 3.878574167697469e-07, "loss": 0.651, "step": 25228 }, { "epoch": 0.9137631293009779, "grad_norm": 1.3894473451103257, "learning_rate": 3.87533947619444e-07, "loss": 0.6998, "step": 25229 }, { "epoch": 0.9137993480622962, "grad_norm": 1.9583896909069312, "learning_rate": 3.8721061074512125e-07, "loss": 0.68, "step": 25230 }, { "epoch": 0.9138355668236147, "grad_norm": 1.1011324779339549, "learning_rate": 3.868874061512273e-07, "loss": 0.6439, "step": 25231 }, { "epoch": 0.913871785584933, "grad_norm": 1.3384498610784095, "learning_rate": 3.865643338422076e-07, "loss": 0.6773, "step": 25232 }, { "epoch": 0.9139080043462514, "grad_norm": 1.3197455155153621, "learning_rate": 3.862413938225118e-07, "loss": 0.6702, "step": 25233 }, { "epoch": 0.9139442231075697, "grad_norm": 1.5374306630185284, "learning_rate": 3.859185860965808e-07, "loss": 0.6632, "step": 25234 }, { "epoch": 0.913980441868888, "grad_norm": 1.3165996413704257, "learning_rate": 3.8559591066885893e-07, "loss": 0.5937, "step": 25235 }, { "epoch": 0.9140166606302065, "grad_norm": 1.3320124134007147, "learning_rate": 3.8527336754378366e-07, "loss": 0.6708, "step": 25236 }, { "epoch": 0.9140528793915248, "grad_norm": 1.0148226143996035, "learning_rate": 3.8495095672579584e-07, "loss": 0.6689, "step": 25237 }, { "epoch": 0.9140890981528432, "grad_norm": 1.318604068653575, "learning_rate": 3.8462867821933203e-07, "loss": 0.6185, "step": 25238 }, { "epoch": 0.9141253169141615, "grad_norm": 1.0759702056399059, "learning_rate": 3.843065320288264e-07, "loss": 0.6147, "step": 25239 }, { "epoch": 0.9141615356754799, "grad_norm": 0.9714516870315953, "learning_rate": 3.8398451815870984e-07, "loss": 0.6721, "step": 25240 }, { "epoch": 0.9141977544367983, "grad_norm": 1.4275368052819795, "learning_rate": 3.8366263661341773e-07, "loss": 0.7383, "step": 25241 }, { "epoch": 0.9142339731981166, "grad_norm": 1.1791247876491044, "learning_rate": 3.8334088739737654e-07, "loss": 0.6886, "step": 25242 }, { "epoch": 0.914270191959435, "grad_norm": 1.0886938052473845, "learning_rate": 3.830192705150149e-07, "loss": 0.6909, "step": 25243 }, { "epoch": 0.9143064107207534, "grad_norm": 1.318444057853133, "learning_rate": 3.826977859707581e-07, "loss": 0.6185, "step": 25244 }, { "epoch": 0.9143426294820717, "grad_norm": 1.446315388436687, "learning_rate": 3.823764337690317e-07, "loss": 0.7022, "step": 25245 }, { "epoch": 0.9143788482433901, "grad_norm": 1.6541112764140629, "learning_rate": 3.8205521391425525e-07, "loss": 0.6889, "step": 25246 }, { "epoch": 0.9144150670047084, "grad_norm": 1.4138066023382383, "learning_rate": 3.8173412641084873e-07, "loss": 0.6646, "step": 25247 }, { "epoch": 0.9144512857660269, "grad_norm": 1.311083958603666, "learning_rate": 3.8141317126323406e-07, "loss": 0.7025, "step": 25248 }, { "epoch": 0.9144875045273452, "grad_norm": 1.458541151052542, "learning_rate": 3.8109234847582555e-07, "loss": 0.703, "step": 25249 }, { "epoch": 0.9145237232886635, "grad_norm": 1.344671100727669, "learning_rate": 3.807716580530374e-07, "loss": 0.6849, "step": 25250 }, { "epoch": 0.9145599420499819, "grad_norm": 1.0719458490459195, "learning_rate": 3.8045109999928274e-07, "loss": 0.7023, "step": 25251 }, { "epoch": 0.9145961608113002, "grad_norm": 1.3958987388541129, "learning_rate": 3.801306743189759e-07, "loss": 0.7232, "step": 25252 }, { "epoch": 0.9146323795726187, "grad_norm": 1.3756236563931203, "learning_rate": 3.7981038101652214e-07, "loss": 0.6605, "step": 25253 }, { "epoch": 0.914668598333937, "grad_norm": 1.3473725227558264, "learning_rate": 3.7949022009633017e-07, "loss": 0.6664, "step": 25254 }, { "epoch": 0.9147048170952553, "grad_norm": 1.4502474431901637, "learning_rate": 3.791701915628065e-07, "loss": 0.6724, "step": 25255 }, { "epoch": 0.9147410358565737, "grad_norm": 1.2618726490335637, "learning_rate": 3.7885029542035544e-07, "loss": 0.6306, "step": 25256 }, { "epoch": 0.914777254617892, "grad_norm": 1.3267619978783816, "learning_rate": 3.785305316733778e-07, "loss": 0.6261, "step": 25257 }, { "epoch": 0.9148134733792105, "grad_norm": 1.3694642265444805, "learning_rate": 3.782109003262724e-07, "loss": 0.6328, "step": 25258 }, { "epoch": 0.9148496921405288, "grad_norm": 1.4191487226247306, "learning_rate": 3.778914013834412e-07, "loss": 0.6681, "step": 25259 }, { "epoch": 0.9148859109018471, "grad_norm": 1.6799138970055454, "learning_rate": 3.775720348492795e-07, "loss": 0.7136, "step": 25260 }, { "epoch": 0.9149221296631656, "grad_norm": 3.149649431763303, "learning_rate": 3.772528007281806e-07, "loss": 0.6331, "step": 25261 }, { "epoch": 0.9149583484244839, "grad_norm": 1.3051488568875111, "learning_rate": 3.769336990245387e-07, "loss": 0.6666, "step": 25262 }, { "epoch": 0.9149945671858023, "grad_norm": 1.355725093754344, "learning_rate": 3.766147297427447e-07, "loss": 0.6852, "step": 25263 }, { "epoch": 0.9150307859471206, "grad_norm": 1.4023939172861812, "learning_rate": 3.762958928871885e-07, "loss": 0.6088, "step": 25264 }, { "epoch": 0.9150670047084389, "grad_norm": 1.5263374420899316, "learning_rate": 3.759771884622554e-07, "loss": 0.7694, "step": 25265 }, { "epoch": 0.9151032234697574, "grad_norm": 1.0173374537685036, "learning_rate": 3.7565861647233413e-07, "loss": 0.6321, "step": 25266 }, { "epoch": 0.9151394422310757, "grad_norm": 1.622875112724697, "learning_rate": 3.753401769218068e-07, "loss": 0.7379, "step": 25267 }, { "epoch": 0.9151756609923941, "grad_norm": 1.0242118867772223, "learning_rate": 3.7502186981505537e-07, "loss": 0.6703, "step": 25268 }, { "epoch": 0.9152118797537124, "grad_norm": 1.0214927940359046, "learning_rate": 3.747036951564587e-07, "loss": 0.6912, "step": 25269 }, { "epoch": 0.9152480985150308, "grad_norm": 1.2578322241699726, "learning_rate": 3.743856529503986e-07, "loss": 0.6102, "step": 25270 }, { "epoch": 0.9152843172763492, "grad_norm": 1.418256165009554, "learning_rate": 3.740677432012496e-07, "loss": 0.6528, "step": 25271 }, { "epoch": 0.9153205360376675, "grad_norm": 1.2504279424848177, "learning_rate": 3.7374996591338586e-07, "loss": 0.6254, "step": 25272 }, { "epoch": 0.9153567547989859, "grad_norm": 2.735930207622402, "learning_rate": 3.7343232109118165e-07, "loss": 0.6336, "step": 25273 }, { "epoch": 0.9153929735603042, "grad_norm": 1.3190606117268908, "learning_rate": 3.7311480873900687e-07, "loss": 0.6484, "step": 25274 }, { "epoch": 0.9154291923216226, "grad_norm": 1.3583870955157582, "learning_rate": 3.727974288612313e-07, "loss": 0.7213, "step": 25275 }, { "epoch": 0.915465411082941, "grad_norm": 1.4381619845593232, "learning_rate": 3.724801814622214e-07, "loss": 0.6921, "step": 25276 }, { "epoch": 0.9155016298442593, "grad_norm": 1.4289471945259207, "learning_rate": 3.7216306654634496e-07, "loss": 0.6642, "step": 25277 }, { "epoch": 0.9155378486055777, "grad_norm": 1.2730275720055821, "learning_rate": 3.71846084117965e-07, "loss": 0.6384, "step": 25278 }, { "epoch": 0.9155740673668961, "grad_norm": 1.4380789179321711, "learning_rate": 3.7152923418144247e-07, "loss": 0.6316, "step": 25279 }, { "epoch": 0.9156102861282144, "grad_norm": 1.4638021695163523, "learning_rate": 3.712125167411362e-07, "loss": 0.714, "step": 25280 }, { "epoch": 0.9156465048895328, "grad_norm": 1.2318220477381836, "learning_rate": 3.7089593180140823e-07, "loss": 0.5827, "step": 25281 }, { "epoch": 0.9156827236508511, "grad_norm": 1.4017387894318556, "learning_rate": 3.705794793666129e-07, "loss": 0.6798, "step": 25282 }, { "epoch": 0.9157189424121696, "grad_norm": 1.3374006780396053, "learning_rate": 3.702631594411044e-07, "loss": 0.6735, "step": 25283 }, { "epoch": 0.9157551611734879, "grad_norm": 1.4425311357740627, "learning_rate": 3.699469720292359e-07, "loss": 0.64, "step": 25284 }, { "epoch": 0.9157913799348062, "grad_norm": 1.3891518430008556, "learning_rate": 3.696309171353607e-07, "loss": 0.5963, "step": 25285 }, { "epoch": 0.9158275986961246, "grad_norm": 1.2673418988734138, "learning_rate": 3.693149947638253e-07, "loss": 0.6214, "step": 25286 }, { "epoch": 0.9158638174574429, "grad_norm": 1.353323022754635, "learning_rate": 3.689992049189761e-07, "loss": 0.7162, "step": 25287 }, { "epoch": 0.9159000362187614, "grad_norm": 1.4548360629642219, "learning_rate": 3.68683547605162e-07, "loss": 0.6778, "step": 25288 }, { "epoch": 0.9159362549800797, "grad_norm": 1.3663928201930398, "learning_rate": 3.68368022826725e-07, "loss": 0.7234, "step": 25289 }, { "epoch": 0.915972473741398, "grad_norm": 1.2970316387282086, "learning_rate": 3.6805263058800723e-07, "loss": 0.7163, "step": 25290 }, { "epoch": 0.9160086925027164, "grad_norm": 1.0601696272425438, "learning_rate": 3.6773737089334736e-07, "loss": 0.6276, "step": 25291 }, { "epoch": 0.9160449112640348, "grad_norm": 1.4096825998284954, "learning_rate": 3.674222437470865e-07, "loss": 0.6154, "step": 25292 }, { "epoch": 0.9160811300253532, "grad_norm": 1.3071309169088354, "learning_rate": 3.671072491535588e-07, "loss": 0.6645, "step": 25293 }, { "epoch": 0.9161173487866715, "grad_norm": 1.3684035608245209, "learning_rate": 3.6679238711709865e-07, "loss": 0.6835, "step": 25294 }, { "epoch": 0.9161535675479898, "grad_norm": 1.4179974203652899, "learning_rate": 3.6647765764204145e-07, "loss": 0.6015, "step": 25295 }, { "epoch": 0.9161897863093083, "grad_norm": 1.4886436446713014, "learning_rate": 3.6616306073271603e-07, "loss": 0.7181, "step": 25296 }, { "epoch": 0.9162260050706266, "grad_norm": 1.3403078226429, "learning_rate": 3.6584859639345104e-07, "loss": 0.6748, "step": 25297 }, { "epoch": 0.916262223831945, "grad_norm": 1.2983724375546912, "learning_rate": 3.6553426462857423e-07, "loss": 0.6885, "step": 25298 }, { "epoch": 0.9162984425932633, "grad_norm": 1.3664686165868736, "learning_rate": 3.6522006544241315e-07, "loss": 0.6308, "step": 25299 }, { "epoch": 0.9163346613545816, "grad_norm": 1.6021894390882405, "learning_rate": 3.649059988392889e-07, "loss": 0.6455, "step": 25300 }, { "epoch": 0.9163708801159001, "grad_norm": 0.9986692410555237, "learning_rate": 3.645920648235257e-07, "loss": 0.7001, "step": 25301 }, { "epoch": 0.9164070988772184, "grad_norm": 1.5890537883909845, "learning_rate": 3.6427826339944017e-07, "loss": 0.7223, "step": 25302 }, { "epoch": 0.9164433176385368, "grad_norm": 1.4186937774743842, "learning_rate": 3.6396459457135324e-07, "loss": 0.5992, "step": 25303 }, { "epoch": 0.9164795363998551, "grad_norm": 0.9128232000363984, "learning_rate": 3.6365105834358036e-07, "loss": 0.6084, "step": 25304 }, { "epoch": 0.9165157551611735, "grad_norm": 1.3786228316872648, "learning_rate": 3.633376547204348e-07, "loss": 0.6995, "step": 25305 }, { "epoch": 0.9165519739224919, "grad_norm": 1.693255940526516, "learning_rate": 3.6302438370623304e-07, "loss": 0.6478, "step": 25306 }, { "epoch": 0.9165881926838102, "grad_norm": 1.3108211506054435, "learning_rate": 3.627112453052828e-07, "loss": 0.5854, "step": 25307 }, { "epoch": 0.9166244114451286, "grad_norm": 2.3152343976551952, "learning_rate": 3.623982395218939e-07, "loss": 0.6834, "step": 25308 }, { "epoch": 0.916660630206447, "grad_norm": 1.3841163353237524, "learning_rate": 3.620853663603718e-07, "loss": 0.6474, "step": 25309 }, { "epoch": 0.9166968489677653, "grad_norm": 1.280853830632655, "learning_rate": 3.6177262582502537e-07, "loss": 0.6625, "step": 25310 }, { "epoch": 0.9167330677290837, "grad_norm": 1.5056532658140258, "learning_rate": 3.6146001792015663e-07, "loss": 0.697, "step": 25311 }, { "epoch": 0.916769286490402, "grad_norm": 1.3049616418571053, "learning_rate": 3.6114754265006767e-07, "loss": 0.6488, "step": 25312 }, { "epoch": 0.9168055052517204, "grad_norm": 1.4594501533190123, "learning_rate": 3.6083520001905623e-07, "loss": 0.7158, "step": 25313 }, { "epoch": 0.9168417240130388, "grad_norm": 1.3034922462804077, "learning_rate": 3.6052299003142333e-07, "loss": 0.6684, "step": 25314 }, { "epoch": 0.9168779427743571, "grad_norm": 1.1262545336602376, "learning_rate": 3.6021091269146327e-07, "loss": 0.6251, "step": 25315 }, { "epoch": 0.9169141615356755, "grad_norm": 1.298244047414665, "learning_rate": 3.5989896800347146e-07, "loss": 0.6569, "step": 25316 }, { "epoch": 0.9169503802969938, "grad_norm": 1.407156122524094, "learning_rate": 3.595871559717412e-07, "loss": 0.6589, "step": 25317 }, { "epoch": 0.9169865990583123, "grad_norm": 0.938361881149348, "learning_rate": 3.5927547660056237e-07, "loss": 0.632, "step": 25318 }, { "epoch": 0.9170228178196306, "grad_norm": 0.8901736686450866, "learning_rate": 3.589639298942238e-07, "loss": 0.6516, "step": 25319 }, { "epoch": 0.9170590365809489, "grad_norm": 1.0058012877862714, "learning_rate": 3.58652515857012e-07, "loss": 0.6974, "step": 25320 }, { "epoch": 0.9170952553422673, "grad_norm": 1.4895072065309833, "learning_rate": 3.5834123449321357e-07, "loss": 0.6619, "step": 25321 }, { "epoch": 0.9171314741035856, "grad_norm": 1.196768756078463, "learning_rate": 3.5803008580711284e-07, "loss": 0.6347, "step": 25322 }, { "epoch": 0.9171676928649041, "grad_norm": 1.494343448411132, "learning_rate": 3.577190698029886e-07, "loss": 0.6615, "step": 25323 }, { "epoch": 0.9172039116262224, "grad_norm": 1.0176930353803453, "learning_rate": 3.574081864851242e-07, "loss": 0.6518, "step": 25324 }, { "epoch": 0.9172401303875407, "grad_norm": 1.0524703686272177, "learning_rate": 3.57097435857795e-07, "loss": 0.658, "step": 25325 }, { "epoch": 0.9172763491488591, "grad_norm": 1.2374021676916682, "learning_rate": 3.567868179252787e-07, "loss": 0.6542, "step": 25326 }, { "epoch": 0.9173125679101775, "grad_norm": 1.3697327463969682, "learning_rate": 3.5647633269184747e-07, "loss": 0.6822, "step": 25327 }, { "epoch": 0.9173487866714959, "grad_norm": 1.4236611607199745, "learning_rate": 3.5616598016177675e-07, "loss": 0.6962, "step": 25328 }, { "epoch": 0.9173850054328142, "grad_norm": 0.9746810626861362, "learning_rate": 3.5585576033933536e-07, "loss": 0.7056, "step": 25329 }, { "epoch": 0.9174212241941325, "grad_norm": 1.2890354442647036, "learning_rate": 3.555456732287943e-07, "loss": 0.6283, "step": 25330 }, { "epoch": 0.917457442955451, "grad_norm": 1.3947655163132815, "learning_rate": 3.552357188344169e-07, "loss": 0.6521, "step": 25331 }, { "epoch": 0.9174936617167693, "grad_norm": 1.3764755854842108, "learning_rate": 3.549258971604719e-07, "loss": 0.6556, "step": 25332 }, { "epoch": 0.9175298804780877, "grad_norm": 1.2827259893844332, "learning_rate": 3.546162082112214e-07, "loss": 0.6498, "step": 25333 }, { "epoch": 0.917566099239406, "grad_norm": 1.472047792656965, "learning_rate": 3.543066519909266e-07, "loss": 0.6339, "step": 25334 }, { "epoch": 0.9176023180007243, "grad_norm": 1.3280614409945832, "learning_rate": 3.539972285038473e-07, "loss": 0.714, "step": 25335 }, { "epoch": 0.9176385367620428, "grad_norm": 1.287216460284089, "learning_rate": 3.5368793775424347e-07, "loss": 0.6088, "step": 25336 }, { "epoch": 0.9176747555233611, "grad_norm": 1.3909834375704084, "learning_rate": 3.5337877974636946e-07, "loss": 0.6779, "step": 25337 }, { "epoch": 0.9177109742846795, "grad_norm": 1.302518235811021, "learning_rate": 3.5306975448447747e-07, "loss": 0.689, "step": 25338 }, { "epoch": 0.9177471930459978, "grad_norm": 1.331106561334248, "learning_rate": 3.52760861972824e-07, "loss": 0.6704, "step": 25339 }, { "epoch": 0.9177834118073162, "grad_norm": 0.9823179799732304, "learning_rate": 3.5245210221565793e-07, "loss": 0.6399, "step": 25340 }, { "epoch": 0.9178196305686346, "grad_norm": 1.4060410973567332, "learning_rate": 3.5214347521722815e-07, "loss": 0.6636, "step": 25341 }, { "epoch": 0.9178558493299529, "grad_norm": 1.3787874644773548, "learning_rate": 3.5183498098178e-07, "loss": 0.6887, "step": 25342 }, { "epoch": 0.9178920680912713, "grad_norm": 1.3757405054892364, "learning_rate": 3.5152661951356136e-07, "loss": 0.7277, "step": 25343 }, { "epoch": 0.9179282868525896, "grad_norm": 1.625886216961451, "learning_rate": 3.512183908168143e-07, "loss": 0.7989, "step": 25344 }, { "epoch": 0.917964505613908, "grad_norm": 0.9502844919639188, "learning_rate": 3.5091029489577876e-07, "loss": 0.6491, "step": 25345 }, { "epoch": 0.9180007243752264, "grad_norm": 1.3007717157277956, "learning_rate": 3.5060233175469806e-07, "loss": 0.6707, "step": 25346 }, { "epoch": 0.9180369431365447, "grad_norm": 1.4861994730276895, "learning_rate": 3.502945013978076e-07, "loss": 0.6872, "step": 25347 }, { "epoch": 0.9180731618978631, "grad_norm": 1.42030063782618, "learning_rate": 3.499868038293441e-07, "loss": 0.7449, "step": 25348 }, { "epoch": 0.9181093806591815, "grad_norm": 1.3271046358280918, "learning_rate": 3.496792390535397e-07, "loss": 0.7418, "step": 25349 }, { "epoch": 0.9181455994204998, "grad_norm": 1.2934919210749094, "learning_rate": 3.493718070746299e-07, "loss": 0.6125, "step": 25350 }, { "epoch": 0.9181818181818182, "grad_norm": 0.9552826688585034, "learning_rate": 3.4906450789684353e-07, "loss": 0.7291, "step": 25351 }, { "epoch": 0.9182180369431365, "grad_norm": 1.4968760116111826, "learning_rate": 3.4875734152441054e-07, "loss": 0.6562, "step": 25352 }, { "epoch": 0.918254255704455, "grad_norm": 1.3872031789509323, "learning_rate": 3.484503079615542e-07, "loss": 0.6991, "step": 25353 }, { "epoch": 0.9182904744657733, "grad_norm": 1.4932143937005384, "learning_rate": 3.481434072125045e-07, "loss": 0.6533, "step": 25354 }, { "epoch": 0.9183266932270916, "grad_norm": 1.1923800065384824, "learning_rate": 3.478366392814825e-07, "loss": 0.6622, "step": 25355 }, { "epoch": 0.91836291198841, "grad_norm": 1.3700630322749812, "learning_rate": 3.47530004172707e-07, "loss": 0.7371, "step": 25356 }, { "epoch": 0.9183991307497283, "grad_norm": 1.3142499554133213, "learning_rate": 3.472235018904024e-07, "loss": 0.6561, "step": 25357 }, { "epoch": 0.9184353495110468, "grad_norm": 1.3404978505384226, "learning_rate": 3.469171324387843e-07, "loss": 0.612, "step": 25358 }, { "epoch": 0.9184715682723651, "grad_norm": 1.542555632998414, "learning_rate": 3.466108958220671e-07, "loss": 0.6575, "step": 25359 }, { "epoch": 0.9185077870336834, "grad_norm": 1.5423496295904613, "learning_rate": 3.463047920444662e-07, "loss": 0.6206, "step": 25360 }, { "epoch": 0.9185440057950018, "grad_norm": 1.4582825240298536, "learning_rate": 3.4599882111019503e-07, "loss": 0.6791, "step": 25361 }, { "epoch": 0.9185802245563202, "grad_norm": 1.030087630185924, "learning_rate": 3.456929830234623e-07, "loss": 0.6717, "step": 25362 }, { "epoch": 0.9186164433176386, "grad_norm": 1.241814000880683, "learning_rate": 3.4538727778847705e-07, "loss": 0.6669, "step": 25363 }, { "epoch": 0.9186526620789569, "grad_norm": 1.7037091489251361, "learning_rate": 3.450817054094446e-07, "loss": 0.6882, "step": 25364 }, { "epoch": 0.9186888808402752, "grad_norm": 1.3418409306985564, "learning_rate": 3.44776265890574e-07, "loss": 0.6537, "step": 25365 }, { "epoch": 0.9187250996015937, "grad_norm": 1.3740595736221928, "learning_rate": 3.44470959236064e-07, "loss": 0.6853, "step": 25366 }, { "epoch": 0.918761318362912, "grad_norm": 1.3347628634164412, "learning_rate": 3.441657854501179e-07, "loss": 0.6305, "step": 25367 }, { "epoch": 0.9187975371242304, "grad_norm": 1.349847610386202, "learning_rate": 3.4386074453693573e-07, "loss": 0.7544, "step": 25368 }, { "epoch": 0.9188337558855487, "grad_norm": 1.3937256380056076, "learning_rate": 3.4355583650071413e-07, "loss": 0.6773, "step": 25369 }, { "epoch": 0.918869974646867, "grad_norm": 1.0225025594245243, "learning_rate": 3.432510613456497e-07, "loss": 0.6632, "step": 25370 }, { "epoch": 0.9189061934081855, "grad_norm": 1.684393020367169, "learning_rate": 3.429464190759346e-07, "loss": 0.714, "step": 25371 }, { "epoch": 0.9189424121695038, "grad_norm": 1.3498120452477198, "learning_rate": 3.426419096957623e-07, "loss": 0.6628, "step": 25372 }, { "epoch": 0.9189786309308222, "grad_norm": 1.1210218942460295, "learning_rate": 3.4233753320932374e-07, "loss": 0.6345, "step": 25373 }, { "epoch": 0.9190148496921405, "grad_norm": 1.1942808496460973, "learning_rate": 3.420332896208067e-07, "loss": 0.663, "step": 25374 }, { "epoch": 0.9190510684534589, "grad_norm": 1.2805859332027494, "learning_rate": 3.4172917893439684e-07, "loss": 0.5998, "step": 25375 }, { "epoch": 0.9190872872147773, "grad_norm": 1.338128998675601, "learning_rate": 3.4142520115428066e-07, "loss": 0.6402, "step": 25376 }, { "epoch": 0.9191235059760956, "grad_norm": 1.1644144990139513, "learning_rate": 3.411213562846416e-07, "loss": 0.5366, "step": 25377 }, { "epoch": 0.919159724737414, "grad_norm": 0.9549680101630549, "learning_rate": 3.4081764432965737e-07, "loss": 0.6203, "step": 25378 }, { "epoch": 0.9191959434987323, "grad_norm": 1.4307976390205752, "learning_rate": 3.4051406529351126e-07, "loss": 0.7359, "step": 25379 }, { "epoch": 0.9192321622600507, "grad_norm": 1.3791653022358397, "learning_rate": 3.402106191803789e-07, "loss": 0.6588, "step": 25380 }, { "epoch": 0.9192683810213691, "grad_norm": 1.6986039847324501, "learning_rate": 3.399073059944369e-07, "loss": 0.6762, "step": 25381 }, { "epoch": 0.9193045997826874, "grad_norm": 1.069405054499815, "learning_rate": 3.3960412573985636e-07, "loss": 0.6296, "step": 25382 }, { "epoch": 0.9193408185440058, "grad_norm": 1.0936547329404187, "learning_rate": 3.393010784208128e-07, "loss": 0.6762, "step": 25383 }, { "epoch": 0.9193770373053242, "grad_norm": 0.9802608232708834, "learning_rate": 3.3899816404147636e-07, "loss": 0.6137, "step": 25384 }, { "epoch": 0.9194132560666425, "grad_norm": 1.5231048025076424, "learning_rate": 3.386953826060113e-07, "loss": 0.7325, "step": 25385 }, { "epoch": 0.9194494748279609, "grad_norm": 1.4449359364584655, "learning_rate": 3.3839273411858887e-07, "loss": 0.6685, "step": 25386 }, { "epoch": 0.9194856935892792, "grad_norm": 1.2709902082871922, "learning_rate": 3.380902185833712e-07, "loss": 0.6108, "step": 25387 }, { "epoch": 0.9195219123505977, "grad_norm": 1.0375117327511632, "learning_rate": 3.3778783600452283e-07, "loss": 0.6273, "step": 25388 }, { "epoch": 0.919558131111916, "grad_norm": 1.2902046749936364, "learning_rate": 3.3748558638620146e-07, "loss": 0.5445, "step": 25389 }, { "epoch": 0.9195943498732343, "grad_norm": 1.37594307722685, "learning_rate": 3.371834697325704e-07, "loss": 0.6734, "step": 25390 }, { "epoch": 0.9196305686345527, "grad_norm": 0.9389041956925737, "learning_rate": 3.368814860477854e-07, "loss": 0.6783, "step": 25391 }, { "epoch": 0.919666787395871, "grad_norm": 1.3915088020328, "learning_rate": 3.365796353360018e-07, "loss": 0.6378, "step": 25392 }, { "epoch": 0.9197030061571895, "grad_norm": 1.4384398005722183, "learning_rate": 3.3627791760137307e-07, "loss": 0.704, "step": 25393 }, { "epoch": 0.9197392249185078, "grad_norm": 1.3241719797477738, "learning_rate": 3.359763328480514e-07, "loss": 0.7105, "step": 25394 }, { "epoch": 0.9197754436798261, "grad_norm": 1.3172399473099934, "learning_rate": 3.35674881080188e-07, "loss": 0.7003, "step": 25395 }, { "epoch": 0.9198116624411445, "grad_norm": 1.4442694792504422, "learning_rate": 3.353735623019294e-07, "loss": 0.7157, "step": 25396 }, { "epoch": 0.9198478812024629, "grad_norm": 1.0727328899944146, "learning_rate": 3.350723765174224e-07, "loss": 0.6121, "step": 25397 }, { "epoch": 0.9198840999637813, "grad_norm": 1.4206116245939095, "learning_rate": 3.3477132373081257e-07, "loss": 0.6439, "step": 25398 }, { "epoch": 0.9199203187250996, "grad_norm": 1.413887596641774, "learning_rate": 3.3447040394623986e-07, "loss": 0.6549, "step": 25399 }, { "epoch": 0.9199565374864179, "grad_norm": 1.525599069798875, "learning_rate": 3.341696171678499e-07, "loss": 0.6869, "step": 25400 }, { "epoch": 0.9199927562477364, "grad_norm": 1.469741806001917, "learning_rate": 3.338689633997783e-07, "loss": 0.6831, "step": 25401 }, { "epoch": 0.9200289750090547, "grad_norm": 1.3829822793011546, "learning_rate": 3.3356844264616274e-07, "loss": 0.7278, "step": 25402 }, { "epoch": 0.9200651937703731, "grad_norm": 1.2416893113543956, "learning_rate": 3.332680549111378e-07, "loss": 0.6056, "step": 25403 }, { "epoch": 0.9201014125316914, "grad_norm": 1.0622929157486631, "learning_rate": 3.329678001988401e-07, "loss": 0.6158, "step": 25404 }, { "epoch": 0.9201376312930097, "grad_norm": 0.9674222973272393, "learning_rate": 3.3266767851339977e-07, "loss": 0.6707, "step": 25405 }, { "epoch": 0.9201738500543282, "grad_norm": 1.0365332382273231, "learning_rate": 3.323676898589467e-07, "loss": 0.6133, "step": 25406 }, { "epoch": 0.9202100688156465, "grad_norm": 1.3288612064305578, "learning_rate": 3.3206783423960774e-07, "loss": 0.5551, "step": 25407 }, { "epoch": 0.9202462875769649, "grad_norm": 0.9931541540721563, "learning_rate": 3.3176811165951064e-07, "loss": 0.7199, "step": 25408 }, { "epoch": 0.9202825063382832, "grad_norm": 1.330240371376037, "learning_rate": 3.314685221227809e-07, "loss": 0.6959, "step": 25409 }, { "epoch": 0.9203187250996016, "grad_norm": 1.3039026326608736, "learning_rate": 3.3116906563353757e-07, "loss": 0.6191, "step": 25410 }, { "epoch": 0.92035494386092, "grad_norm": 1.095940891773353, "learning_rate": 3.308697421959051e-07, "loss": 0.7359, "step": 25411 }, { "epoch": 0.9203911626222383, "grad_norm": 1.4416571763126844, "learning_rate": 3.305705518140012e-07, "loss": 0.6785, "step": 25412 }, { "epoch": 0.9204273813835567, "grad_norm": 1.428879512142512, "learning_rate": 3.3027149449194273e-07, "loss": 0.6649, "step": 25413 }, { "epoch": 0.920463600144875, "grad_norm": 1.3074399330845579, "learning_rate": 3.299725702338441e-07, "loss": 0.6456, "step": 25414 }, { "epoch": 0.9204998189061934, "grad_norm": 1.3787059165624307, "learning_rate": 3.2967377904382094e-07, "loss": 0.5884, "step": 25415 }, { "epoch": 0.9205360376675118, "grad_norm": 1.4071972676496642, "learning_rate": 3.2937512092598323e-07, "loss": 0.6676, "step": 25416 }, { "epoch": 0.9205722564288301, "grad_norm": 1.3848715682454615, "learning_rate": 3.2907659588444106e-07, "loss": 0.6597, "step": 25417 }, { "epoch": 0.9206084751901485, "grad_norm": 1.345364283657209, "learning_rate": 3.287782039233023e-07, "loss": 0.6165, "step": 25418 }, { "epoch": 0.9206446939514669, "grad_norm": 1.3402609008925856, "learning_rate": 3.284799450466747e-07, "loss": 0.686, "step": 25419 }, { "epoch": 0.9206809127127852, "grad_norm": 1.5164720488574952, "learning_rate": 3.2818181925866057e-07, "loss": 0.6489, "step": 25420 }, { "epoch": 0.9207171314741036, "grad_norm": 1.3711825998160814, "learning_rate": 3.2788382656336103e-07, "loss": 0.6627, "step": 25421 }, { "epoch": 0.9207533502354219, "grad_norm": 1.046188236901605, "learning_rate": 3.2758596696488066e-07, "loss": 0.6616, "step": 25422 }, { "epoch": 0.9207895689967404, "grad_norm": 1.070153089890959, "learning_rate": 3.272882404673172e-07, "loss": 0.7057, "step": 25423 }, { "epoch": 0.9208257877580587, "grad_norm": 1.288801734009199, "learning_rate": 3.269906470747652e-07, "loss": 0.6607, "step": 25424 }, { "epoch": 0.920862006519377, "grad_norm": 1.2582599839038573, "learning_rate": 3.2669318679132146e-07, "loss": 0.7126, "step": 25425 }, { "epoch": 0.9208982252806954, "grad_norm": 1.3561377914822539, "learning_rate": 3.263958596210792e-07, "loss": 0.6283, "step": 25426 }, { "epoch": 0.9209344440420137, "grad_norm": 1.519911519277329, "learning_rate": 3.2609866556813085e-07, "loss": 0.7227, "step": 25427 }, { "epoch": 0.9209706628033322, "grad_norm": 0.9248381970710737, "learning_rate": 3.2580160463656306e-07, "loss": 0.6871, "step": 25428 }, { "epoch": 0.9210068815646505, "grad_norm": 1.6716862531876904, "learning_rate": 3.2550467683046707e-07, "loss": 0.7196, "step": 25429 }, { "epoch": 0.9210431003259688, "grad_norm": 1.432639207651232, "learning_rate": 3.252078821539284e-07, "loss": 0.6982, "step": 25430 }, { "epoch": 0.9210793190872872, "grad_norm": 1.3456284664244949, "learning_rate": 3.2491122061102945e-07, "loss": 0.642, "step": 25431 }, { "epoch": 0.9211155378486056, "grad_norm": 1.3494614804409553, "learning_rate": 3.246146922058524e-07, "loss": 0.7057, "step": 25432 }, { "epoch": 0.921151756609924, "grad_norm": 0.9808620298706037, "learning_rate": 3.2431829694247963e-07, "loss": 0.6469, "step": 25433 }, { "epoch": 0.9211879753712423, "grad_norm": 1.3280769676049873, "learning_rate": 3.2402203482498895e-07, "loss": 0.6187, "step": 25434 }, { "epoch": 0.9212241941325606, "grad_norm": 1.4824403477700152, "learning_rate": 3.237259058574571e-07, "loss": 0.7478, "step": 25435 }, { "epoch": 0.921260412893879, "grad_norm": 1.3218878786832688, "learning_rate": 3.2342991004395754e-07, "loss": 0.6562, "step": 25436 }, { "epoch": 0.9212966316551974, "grad_norm": 1.1120577720037987, "learning_rate": 3.231340473885669e-07, "loss": 0.6346, "step": 25437 }, { "epoch": 0.9213328504165158, "grad_norm": 1.3720245496238512, "learning_rate": 3.2283831789535425e-07, "loss": 0.5866, "step": 25438 }, { "epoch": 0.9213690691778341, "grad_norm": 1.3700603539934393, "learning_rate": 3.225427215683885e-07, "loss": 0.6706, "step": 25439 }, { "epoch": 0.9214052879391524, "grad_norm": 1.4064692975290107, "learning_rate": 3.222472584117398e-07, "loss": 0.6917, "step": 25440 }, { "epoch": 0.9214415067004709, "grad_norm": 1.0081952979098263, "learning_rate": 3.2195192842947254e-07, "loss": 0.6472, "step": 25441 }, { "epoch": 0.9214777254617892, "grad_norm": 1.2893516729140508, "learning_rate": 3.216567316256491e-07, "loss": 0.6566, "step": 25442 }, { "epoch": 0.9215139442231076, "grad_norm": 1.4236493763749651, "learning_rate": 3.2136166800433297e-07, "loss": 0.6876, "step": 25443 }, { "epoch": 0.9215501629844259, "grad_norm": 1.3907323367907827, "learning_rate": 3.210667375695864e-07, "loss": 0.6418, "step": 25444 }, { "epoch": 0.9215863817457443, "grad_norm": 1.2784209834675133, "learning_rate": 3.207719403254661e-07, "loss": 0.5647, "step": 25445 }, { "epoch": 0.9216226005070627, "grad_norm": 1.3448902950185635, "learning_rate": 3.2047727627602775e-07, "loss": 0.6665, "step": 25446 }, { "epoch": 0.921658819268381, "grad_norm": 1.3611805246097266, "learning_rate": 3.20182745425327e-07, "loss": 0.6428, "step": 25447 }, { "epoch": 0.9216950380296994, "grad_norm": 1.2504163868117386, "learning_rate": 3.198883477774184e-07, "loss": 0.6723, "step": 25448 }, { "epoch": 0.9217312567910178, "grad_norm": 1.275970919912052, "learning_rate": 3.195940833363509e-07, "loss": 0.6575, "step": 25449 }, { "epoch": 0.9217674755523361, "grad_norm": 1.490458283533094, "learning_rate": 3.192999521061746e-07, "loss": 0.6541, "step": 25450 }, { "epoch": 0.9218036943136545, "grad_norm": 1.449959780129082, "learning_rate": 3.190059540909374e-07, "loss": 0.6623, "step": 25451 }, { "epoch": 0.9218399130749728, "grad_norm": 1.2755225655159872, "learning_rate": 3.187120892946849e-07, "loss": 0.7043, "step": 25452 }, { "epoch": 0.9218761318362912, "grad_norm": 1.738111374087319, "learning_rate": 3.184183577214617e-07, "loss": 0.6334, "step": 25453 }, { "epoch": 0.9219123505976096, "grad_norm": 1.490833263998148, "learning_rate": 3.1812475937530674e-07, "loss": 0.7037, "step": 25454 }, { "epoch": 0.9219485693589279, "grad_norm": 1.4404742276045874, "learning_rate": 3.178312942602635e-07, "loss": 0.6154, "step": 25455 }, { "epoch": 0.9219847881202463, "grad_norm": 0.882463059567441, "learning_rate": 3.175379623803698e-07, "loss": 0.6183, "step": 25456 }, { "epoch": 0.9220210068815646, "grad_norm": 1.3922139369710753, "learning_rate": 3.172447637396614e-07, "loss": 0.6917, "step": 25457 }, { "epoch": 0.9220572256428831, "grad_norm": 1.4000988793366103, "learning_rate": 3.1695169834217167e-07, "loss": 0.7277, "step": 25458 }, { "epoch": 0.9220934444042014, "grad_norm": 1.7726741979053122, "learning_rate": 3.1665876619193624e-07, "loss": 0.6682, "step": 25459 }, { "epoch": 0.9221296631655197, "grad_norm": 1.3958857954068475, "learning_rate": 3.163659672929842e-07, "loss": 0.6374, "step": 25460 }, { "epoch": 0.9221658819268381, "grad_norm": 1.296230032264916, "learning_rate": 3.1607330164934446e-07, "loss": 0.6667, "step": 25461 }, { "epoch": 0.9222021006881564, "grad_norm": 1.2570795276645825, "learning_rate": 3.157807692650461e-07, "loss": 0.6808, "step": 25462 }, { "epoch": 0.9222383194494749, "grad_norm": 1.4864386493628856, "learning_rate": 3.154883701441136e-07, "loss": 0.6318, "step": 25463 }, { "epoch": 0.9222745382107932, "grad_norm": 1.3096096159508033, "learning_rate": 3.1519610429057044e-07, "loss": 0.6305, "step": 25464 }, { "epoch": 0.9223107569721115, "grad_norm": 1.3910077899582582, "learning_rate": 3.1490397170843787e-07, "loss": 0.6127, "step": 25465 }, { "epoch": 0.9223469757334299, "grad_norm": 1.4137591002265846, "learning_rate": 3.1461197240173824e-07, "loss": 0.7024, "step": 25466 }, { "epoch": 0.9223831944947483, "grad_norm": 1.4065254723095593, "learning_rate": 3.143201063744883e-07, "loss": 0.6126, "step": 25467 }, { "epoch": 0.9224194132560667, "grad_norm": 1.154126787725032, "learning_rate": 3.140283736307037e-07, "loss": 0.6899, "step": 25468 }, { "epoch": 0.922455632017385, "grad_norm": 0.9982830005246095, "learning_rate": 3.1373677417439794e-07, "loss": 0.6522, "step": 25469 }, { "epoch": 0.9224918507787033, "grad_norm": 1.3357437569131678, "learning_rate": 3.1344530800958785e-07, "loss": 0.6954, "step": 25470 }, { "epoch": 0.9225280695400218, "grad_norm": 1.3336011879405705, "learning_rate": 3.1315397514028013e-07, "loss": 0.5955, "step": 25471 }, { "epoch": 0.9225642883013401, "grad_norm": 0.8715571809680123, "learning_rate": 3.1286277557048497e-07, "loss": 0.6276, "step": 25472 }, { "epoch": 0.9226005070626585, "grad_norm": 1.3775800752125684, "learning_rate": 3.125717093042113e-07, "loss": 0.6898, "step": 25473 }, { "epoch": 0.9226367258239768, "grad_norm": 1.1147169862001434, "learning_rate": 3.122807763454627e-07, "loss": 0.7092, "step": 25474 }, { "epoch": 0.9226729445852951, "grad_norm": 1.3467055685476177, "learning_rate": 3.119899766982426e-07, "loss": 0.6061, "step": 25475 }, { "epoch": 0.9227091633466136, "grad_norm": 1.4276986750406802, "learning_rate": 3.116993103665522e-07, "loss": 0.669, "step": 25476 }, { "epoch": 0.9227453821079319, "grad_norm": 1.3226871839771328, "learning_rate": 3.114087773543939e-07, "loss": 0.6766, "step": 25477 }, { "epoch": 0.9227816008692503, "grad_norm": 1.329834311429049, "learning_rate": 3.111183776657634e-07, "loss": 0.6975, "step": 25478 }, { "epoch": 0.9228178196305686, "grad_norm": 1.535858217057804, "learning_rate": 3.1082811130465627e-07, "loss": 0.5788, "step": 25479 }, { "epoch": 0.922854038391887, "grad_norm": 1.3344601087262942, "learning_rate": 3.105379782750695e-07, "loss": 0.648, "step": 25480 }, { "epoch": 0.9228902571532054, "grad_norm": 1.309697660587886, "learning_rate": 3.1024797858099307e-07, "loss": 0.5909, "step": 25481 }, { "epoch": 0.9229264759145237, "grad_norm": 1.4008736329429834, "learning_rate": 3.0995811222641946e-07, "loss": 0.6953, "step": 25482 }, { "epoch": 0.9229626946758421, "grad_norm": 1.2908499260782662, "learning_rate": 3.0966837921533546e-07, "loss": 0.6486, "step": 25483 }, { "epoch": 0.9229989134371605, "grad_norm": 1.1140487389183598, "learning_rate": 3.0937877955173e-07, "loss": 0.6748, "step": 25484 }, { "epoch": 0.9230351321984788, "grad_norm": 1.3714078953583055, "learning_rate": 3.090893132395878e-07, "loss": 0.7044, "step": 25485 }, { "epoch": 0.9230713509597972, "grad_norm": 1.3546193382220852, "learning_rate": 3.0879998028289227e-07, "loss": 0.7177, "step": 25486 }, { "epoch": 0.9231075697211155, "grad_norm": 1.3657586137649926, "learning_rate": 3.085107806856225e-07, "loss": 0.6316, "step": 25487 }, { "epoch": 0.923143788482434, "grad_norm": 1.3325286982337368, "learning_rate": 3.0822171445176186e-07, "loss": 0.6156, "step": 25488 }, { "epoch": 0.9231800072437523, "grad_norm": 1.3493320101407031, "learning_rate": 3.0793278158528506e-07, "loss": 0.6287, "step": 25489 }, { "epoch": 0.9232162260050706, "grad_norm": 4.057511193535596, "learning_rate": 3.0764398209017e-07, "loss": 0.6106, "step": 25490 }, { "epoch": 0.923252444766389, "grad_norm": 1.2788030100639007, "learning_rate": 3.0735531597039013e-07, "loss": 0.6637, "step": 25491 }, { "epoch": 0.9232886635277073, "grad_norm": 1.3957475775929853, "learning_rate": 3.070667832299179e-07, "loss": 0.6327, "step": 25492 }, { "epoch": 0.9233248822890258, "grad_norm": 1.0131624045224512, "learning_rate": 3.067783838727245e-07, "loss": 0.6639, "step": 25493 }, { "epoch": 0.9233611010503441, "grad_norm": 1.2737096652164597, "learning_rate": 3.064901179027757e-07, "loss": 0.6922, "step": 25494 }, { "epoch": 0.9233973198116624, "grad_norm": 1.3779686483637699, "learning_rate": 3.0620198532404164e-07, "loss": 0.6718, "step": 25495 }, { "epoch": 0.9234335385729808, "grad_norm": 0.9301198615263855, "learning_rate": 3.0591398614048695e-07, "loss": 0.6627, "step": 25496 }, { "epoch": 0.9234697573342991, "grad_norm": 1.0515168187519686, "learning_rate": 3.056261203560729e-07, "loss": 0.6696, "step": 25497 }, { "epoch": 0.9235059760956176, "grad_norm": 1.3427284982270455, "learning_rate": 3.053383879747607e-07, "loss": 0.6527, "step": 25498 }, { "epoch": 0.9235421948569359, "grad_norm": 1.173675025494076, "learning_rate": 3.0505078900051166e-07, "loss": 0.7506, "step": 25499 }, { "epoch": 0.9235784136182542, "grad_norm": 1.0904568086095399, "learning_rate": 3.0476332343728263e-07, "loss": 0.6663, "step": 25500 }, { "epoch": 0.9236146323795726, "grad_norm": 1.3311695215867214, "learning_rate": 3.044759912890283e-07, "loss": 0.5942, "step": 25501 }, { "epoch": 0.923650851140891, "grad_norm": 1.4072381772919378, "learning_rate": 3.041887925597042e-07, "loss": 0.6803, "step": 25502 }, { "epoch": 0.9236870699022094, "grad_norm": 1.383873847404643, "learning_rate": 3.0390172725326294e-07, "loss": 0.6469, "step": 25503 }, { "epoch": 0.9237232886635277, "grad_norm": 1.4900717915330521, "learning_rate": 3.0361479537365237e-07, "loss": 0.6294, "step": 25504 }, { "epoch": 0.923759507424846, "grad_norm": 1.5634579356583278, "learning_rate": 3.0332799692482153e-07, "loss": 0.73, "step": 25505 }, { "epoch": 0.9237957261861645, "grad_norm": 1.332082553054548, "learning_rate": 3.030413319107195e-07, "loss": 0.6152, "step": 25506 }, { "epoch": 0.9238319449474828, "grad_norm": 1.5774576409587926, "learning_rate": 3.027548003352887e-07, "loss": 0.6889, "step": 25507 }, { "epoch": 0.9238681637088012, "grad_norm": 1.02294330435889, "learning_rate": 3.0246840220247153e-07, "loss": 0.6217, "step": 25508 }, { "epoch": 0.9239043824701195, "grad_norm": 1.0055264220793305, "learning_rate": 3.021821375162104e-07, "loss": 0.7058, "step": 25509 }, { "epoch": 0.9239406012314378, "grad_norm": 1.3741198094508174, "learning_rate": 3.0189600628044545e-07, "loss": 0.6445, "step": 25510 }, { "epoch": 0.9239768199927563, "grad_norm": 1.7948144303583211, "learning_rate": 3.0161000849911246e-07, "loss": 0.7058, "step": 25511 }, { "epoch": 0.9240130387540746, "grad_norm": 1.4142623761423079, "learning_rate": 3.0132414417614606e-07, "loss": 0.7044, "step": 25512 }, { "epoch": 0.924049257515393, "grad_norm": 1.1313071605874885, "learning_rate": 3.010384133154831e-07, "loss": 0.6841, "step": 25513 }, { "epoch": 0.9240854762767113, "grad_norm": 0.9825786363953972, "learning_rate": 3.0075281592105376e-07, "loss": 0.579, "step": 25514 }, { "epoch": 0.9241216950380297, "grad_norm": 1.5252495650398836, "learning_rate": 3.0046735199678823e-07, "loss": 0.6184, "step": 25515 }, { "epoch": 0.9241579137993481, "grad_norm": 1.3437135422551416, "learning_rate": 3.0018202154661334e-07, "loss": 0.7128, "step": 25516 }, { "epoch": 0.9241941325606664, "grad_norm": 1.2617975164634665, "learning_rate": 2.998968245744582e-07, "loss": 0.5746, "step": 25517 }, { "epoch": 0.9242303513219848, "grad_norm": 1.3672456320256743, "learning_rate": 2.9961176108424526e-07, "loss": 0.6342, "step": 25518 }, { "epoch": 0.9242665700833032, "grad_norm": 1.201224776543274, "learning_rate": 2.9932683107989915e-07, "loss": 0.7559, "step": 25519 }, { "epoch": 0.9243027888446215, "grad_norm": 1.1066571095127047, "learning_rate": 2.9904203456533777e-07, "loss": 0.6494, "step": 25520 }, { "epoch": 0.9243390076059399, "grad_norm": 1.4326876549529513, "learning_rate": 2.987573715444836e-07, "loss": 0.6906, "step": 25521 }, { "epoch": 0.9243752263672582, "grad_norm": 1.5380108260933545, "learning_rate": 2.9847284202125236e-07, "loss": 0.731, "step": 25522 }, { "epoch": 0.9244114451285766, "grad_norm": 1.2929742545042393, "learning_rate": 2.9818844599955766e-07, "loss": 0.6427, "step": 25523 }, { "epoch": 0.924447663889895, "grad_norm": 1.4132361038972459, "learning_rate": 2.9790418348331517e-07, "loss": 0.6785, "step": 25524 }, { "epoch": 0.9244838826512133, "grad_norm": 1.4592715195281314, "learning_rate": 2.9762005447643737e-07, "loss": 0.6934, "step": 25525 }, { "epoch": 0.9245201014125317, "grad_norm": 1.4915694334818879, "learning_rate": 2.9733605898283227e-07, "loss": 0.7192, "step": 25526 }, { "epoch": 0.92455632017385, "grad_norm": 1.592522056945597, "learning_rate": 2.9705219700640776e-07, "loss": 0.6556, "step": 25527 }, { "epoch": 0.9245925389351685, "grad_norm": 1.547749535239553, "learning_rate": 2.967684685510719e-07, "loss": 0.6959, "step": 25528 }, { "epoch": 0.9246287576964868, "grad_norm": 1.3198911414924683, "learning_rate": 2.964848736207271e-07, "loss": 0.5975, "step": 25529 }, { "epoch": 0.9246649764578051, "grad_norm": 0.9958238782770872, "learning_rate": 2.96201412219278e-07, "loss": 0.6019, "step": 25530 }, { "epoch": 0.9247011952191235, "grad_norm": 1.3855947378423865, "learning_rate": 2.959180843506215e-07, "loss": 0.7422, "step": 25531 }, { "epoch": 0.9247374139804418, "grad_norm": 1.0834278969600273, "learning_rate": 2.956348900186612e-07, "loss": 0.6562, "step": 25532 }, { "epoch": 0.9247736327417603, "grad_norm": 1.8101155097436836, "learning_rate": 2.9535182922729057e-07, "loss": 0.623, "step": 25533 }, { "epoch": 0.9248098515030786, "grad_norm": 1.4525328071678056, "learning_rate": 2.950689019804065e-07, "loss": 0.6397, "step": 25534 }, { "epoch": 0.9248460702643969, "grad_norm": 1.5746850466117663, "learning_rate": 2.947861082819015e-07, "loss": 0.7316, "step": 25535 }, { "epoch": 0.9248822890257153, "grad_norm": 1.4091959443975386, "learning_rate": 2.945034481356679e-07, "loss": 0.6832, "step": 25536 }, { "epoch": 0.9249185077870337, "grad_norm": 1.3295455474143045, "learning_rate": 2.942209215455949e-07, "loss": 0.6385, "step": 25537 }, { "epoch": 0.9249547265483521, "grad_norm": 1.1026732964479833, "learning_rate": 2.9393852851556937e-07, "loss": 0.5895, "step": 25538 }, { "epoch": 0.9249909453096704, "grad_norm": 1.0175414697141474, "learning_rate": 2.936562690494793e-07, "loss": 0.7191, "step": 25539 }, { "epoch": 0.9250271640709887, "grad_norm": 1.5911301685182326, "learning_rate": 2.9337414315120717e-07, "loss": 0.6787, "step": 25540 }, { "epoch": 0.9250633828323072, "grad_norm": 1.4598623682639713, "learning_rate": 2.930921508246354e-07, "loss": 0.6741, "step": 25541 }, { "epoch": 0.9250996015936255, "grad_norm": 1.3267858029460446, "learning_rate": 2.928102920736453e-07, "loss": 0.677, "step": 25542 }, { "epoch": 0.9251358203549439, "grad_norm": 1.3146716675882935, "learning_rate": 2.9252856690211605e-07, "loss": 0.695, "step": 25543 }, { "epoch": 0.9251720391162622, "grad_norm": 1.33427427864295, "learning_rate": 2.9224697531392235e-07, "loss": 0.6242, "step": 25544 }, { "epoch": 0.9252082578775805, "grad_norm": 1.4526620865858761, "learning_rate": 2.9196551731293985e-07, "loss": 0.6893, "step": 25545 }, { "epoch": 0.925244476638899, "grad_norm": 1.3757996834641593, "learning_rate": 2.916841929030423e-07, "loss": 0.6552, "step": 25546 }, { "epoch": 0.9252806954002173, "grad_norm": 1.4412584810574003, "learning_rate": 2.9140300208810203e-07, "loss": 0.6435, "step": 25547 }, { "epoch": 0.9253169141615357, "grad_norm": 1.3913861700308656, "learning_rate": 2.91121944871986e-07, "loss": 0.7094, "step": 25548 }, { "epoch": 0.925353132922854, "grad_norm": 1.4956025737867693, "learning_rate": 2.908410212585622e-07, "loss": 0.7219, "step": 25549 }, { "epoch": 0.9253893516841724, "grad_norm": 1.6282815078224129, "learning_rate": 2.905602312516986e-07, "loss": 0.6774, "step": 25550 }, { "epoch": 0.9254255704454908, "grad_norm": 1.349357549743139, "learning_rate": 2.902795748552578e-07, "loss": 0.6456, "step": 25551 }, { "epoch": 0.9254617892068091, "grad_norm": 1.4711469434013555, "learning_rate": 2.899990520731e-07, "loss": 0.6585, "step": 25552 }, { "epoch": 0.9254980079681275, "grad_norm": 1.3758029780873502, "learning_rate": 2.897186629090887e-07, "loss": 0.6447, "step": 25553 }, { "epoch": 0.9255342267294459, "grad_norm": 0.9874293942368533, "learning_rate": 2.8943840736708086e-07, "loss": 0.7284, "step": 25554 }, { "epoch": 0.9255704454907642, "grad_norm": 1.6710785791529716, "learning_rate": 2.891582854509323e-07, "loss": 0.7205, "step": 25555 }, { "epoch": 0.9256066642520826, "grad_norm": 1.480524272179529, "learning_rate": 2.8887829716449877e-07, "loss": 0.7448, "step": 25556 }, { "epoch": 0.9256428830134009, "grad_norm": 1.3345059603277063, "learning_rate": 2.885984425116328e-07, "loss": 0.6515, "step": 25557 }, { "epoch": 0.9256791017747193, "grad_norm": 1.497751609719401, "learning_rate": 2.883187214961858e-07, "loss": 0.6511, "step": 25558 }, { "epoch": 0.9257153205360377, "grad_norm": 1.1322411225595008, "learning_rate": 2.880391341220057e-07, "loss": 0.6854, "step": 25559 }, { "epoch": 0.925751539297356, "grad_norm": 1.4507222897489478, "learning_rate": 2.877596803929406e-07, "loss": 0.6562, "step": 25560 }, { "epoch": 0.9257877580586744, "grad_norm": 1.1261884304880907, "learning_rate": 2.8748036031283734e-07, "loss": 0.6274, "step": 25561 }, { "epoch": 0.9258239768199927, "grad_norm": 1.5798407648654642, "learning_rate": 2.872011738855385e-07, "loss": 0.6378, "step": 25562 }, { "epoch": 0.9258601955813112, "grad_norm": 1.3427898496145005, "learning_rate": 2.869221211148843e-07, "loss": 0.6388, "step": 25563 }, { "epoch": 0.9258964143426295, "grad_norm": 1.390641832517698, "learning_rate": 2.866432020047183e-07, "loss": 0.649, "step": 25564 }, { "epoch": 0.9259326331039478, "grad_norm": 1.0138086518570804, "learning_rate": 2.8636441655887637e-07, "loss": 0.6685, "step": 25565 }, { "epoch": 0.9259688518652662, "grad_norm": 1.5343863944720202, "learning_rate": 2.8608576478119543e-07, "loss": 0.6972, "step": 25566 }, { "epoch": 0.9260050706265845, "grad_norm": 1.3138531985104545, "learning_rate": 2.85807246675508e-07, "loss": 0.6448, "step": 25567 }, { "epoch": 0.926041289387903, "grad_norm": 1.4436414111418288, "learning_rate": 2.855288622456498e-07, "loss": 0.7295, "step": 25568 }, { "epoch": 0.9260775081492213, "grad_norm": 1.4333057763972412, "learning_rate": 2.852506114954501e-07, "loss": 0.68, "step": 25569 }, { "epoch": 0.9261137269105396, "grad_norm": 1.3877591957995552, "learning_rate": 2.849724944287391e-07, "loss": 0.7072, "step": 25570 }, { "epoch": 0.926149945671858, "grad_norm": 1.0849879768557675, "learning_rate": 2.846945110493415e-07, "loss": 0.7339, "step": 25571 }, { "epoch": 0.9261861644331764, "grad_norm": 1.4370858675264635, "learning_rate": 2.844166613610844e-07, "loss": 0.6489, "step": 25572 }, { "epoch": 0.9262223831944948, "grad_norm": 1.4194761466047392, "learning_rate": 2.8413894536779117e-07, "loss": 0.6795, "step": 25573 }, { "epoch": 0.9262586019558131, "grad_norm": 1.3957409046717342, "learning_rate": 2.838613630732823e-07, "loss": 0.6844, "step": 25574 }, { "epoch": 0.9262948207171314, "grad_norm": 1.373850670603216, "learning_rate": 2.835839144813801e-07, "loss": 0.7151, "step": 25575 }, { "epoch": 0.9263310394784499, "grad_norm": 1.363194091901835, "learning_rate": 2.8330659959589944e-07, "loss": 0.6661, "step": 25576 }, { "epoch": 0.9263672582397682, "grad_norm": 1.3805772622685362, "learning_rate": 2.8302941842065945e-07, "loss": 0.6715, "step": 25577 }, { "epoch": 0.9264034770010866, "grad_norm": 1.3862346314532001, "learning_rate": 2.827523709594704e-07, "loss": 0.6787, "step": 25578 }, { "epoch": 0.9264396957624049, "grad_norm": 1.5502888590840815, "learning_rate": 2.824754572161481e-07, "loss": 0.6516, "step": 25579 }, { "epoch": 0.9264759145237232, "grad_norm": 1.5121376697915394, "learning_rate": 2.821986771945029e-07, "loss": 0.6662, "step": 25580 }, { "epoch": 0.9265121332850417, "grad_norm": 1.5394643099305205, "learning_rate": 2.819220308983428e-07, "loss": 0.6241, "step": 25581 }, { "epoch": 0.92654835204636, "grad_norm": 1.2927435009938488, "learning_rate": 2.8164551833147256e-07, "loss": 0.6725, "step": 25582 }, { "epoch": 0.9265845708076784, "grad_norm": 1.368027894720204, "learning_rate": 2.813691394977014e-07, "loss": 0.6727, "step": 25583 }, { "epoch": 0.9266207895689967, "grad_norm": 1.3163522163077916, "learning_rate": 2.8109289440082954e-07, "loss": 0.6766, "step": 25584 }, { "epoch": 0.9266570083303151, "grad_norm": 1.3067938225863815, "learning_rate": 2.808167830446584e-07, "loss": 0.6909, "step": 25585 }, { "epoch": 0.9266932270916335, "grad_norm": 1.341299349291606, "learning_rate": 2.8054080543298943e-07, "loss": 0.6724, "step": 25586 }, { "epoch": 0.9267294458529518, "grad_norm": 1.4994506773826584, "learning_rate": 2.802649615696196e-07, "loss": 0.6633, "step": 25587 }, { "epoch": 0.9267656646142702, "grad_norm": 1.4134414136099924, "learning_rate": 2.7998925145834355e-07, "loss": 0.6661, "step": 25588 }, { "epoch": 0.9268018833755886, "grad_norm": 1.285120832395804, "learning_rate": 2.797136751029561e-07, "loss": 0.643, "step": 25589 }, { "epoch": 0.9268381021369069, "grad_norm": 1.3833507053510532, "learning_rate": 2.7943823250724977e-07, "loss": 0.6213, "step": 25590 }, { "epoch": 0.9268743208982253, "grad_norm": 1.0526760483040776, "learning_rate": 2.7916292367501483e-07, "loss": 0.6851, "step": 25591 }, { "epoch": 0.9269105396595436, "grad_norm": 1.4375594207154698, "learning_rate": 2.7888774861003944e-07, "loss": 0.6227, "step": 25592 }, { "epoch": 0.926946758420862, "grad_norm": 1.3960336484657683, "learning_rate": 2.7861270731610936e-07, "loss": 0.7121, "step": 25593 }, { "epoch": 0.9269829771821804, "grad_norm": 1.0617042169767483, "learning_rate": 2.783377997970116e-07, "loss": 0.6385, "step": 25594 }, { "epoch": 0.9270191959434987, "grad_norm": 1.394579275490346, "learning_rate": 2.780630260565287e-07, "loss": 0.6662, "step": 25595 }, { "epoch": 0.9270554147048171, "grad_norm": 1.5238146804063921, "learning_rate": 2.777883860984387e-07, "loss": 0.7123, "step": 25596 }, { "epoch": 0.9270916334661354, "grad_norm": 1.5114584977232908, "learning_rate": 2.7751387992652423e-07, "loss": 0.7761, "step": 25597 }, { "epoch": 0.9271278522274539, "grad_norm": 1.2766397060038683, "learning_rate": 2.772395075445622e-07, "loss": 0.6937, "step": 25598 }, { "epoch": 0.9271640709887722, "grad_norm": 1.0247858538086605, "learning_rate": 2.7696526895632737e-07, "loss": 0.7284, "step": 25599 }, { "epoch": 0.9272002897500905, "grad_norm": 1.2910025086652763, "learning_rate": 2.7669116416559336e-07, "loss": 0.7139, "step": 25600 }, { "epoch": 0.9272365085114089, "grad_norm": 1.1609513877550446, "learning_rate": 2.764171931761339e-07, "loss": 0.6929, "step": 25601 }, { "epoch": 0.9272727272727272, "grad_norm": 1.390068766035503, "learning_rate": 2.7614335599171706e-07, "loss": 0.6599, "step": 25602 }, { "epoch": 0.9273089460340457, "grad_norm": 1.284684954262366, "learning_rate": 2.7586965261611086e-07, "loss": 0.6786, "step": 25603 }, { "epoch": 0.927345164795364, "grad_norm": 1.3663007987506692, "learning_rate": 2.755960830530835e-07, "loss": 0.681, "step": 25604 }, { "epoch": 0.9273813835566823, "grad_norm": 0.9777108582094286, "learning_rate": 2.753226473063986e-07, "loss": 0.6434, "step": 25605 }, { "epoch": 0.9274176023180007, "grad_norm": 1.379399238510877, "learning_rate": 2.7504934537981865e-07, "loss": 0.6098, "step": 25606 }, { "epoch": 0.9274538210793191, "grad_norm": 1.4591498959279938, "learning_rate": 2.747761772771063e-07, "loss": 0.6538, "step": 25607 }, { "epoch": 0.9274900398406375, "grad_norm": 1.3318770073218147, "learning_rate": 2.745031430020173e-07, "loss": 0.6625, "step": 25608 }, { "epoch": 0.9275262586019558, "grad_norm": 1.494060321399837, "learning_rate": 2.74230242558311e-07, "loss": 0.7041, "step": 25609 }, { "epoch": 0.9275624773632741, "grad_norm": 1.7789706727794705, "learning_rate": 2.739574759497421e-07, "loss": 0.6544, "step": 25610 }, { "epoch": 0.9275986961245926, "grad_norm": 1.3941506746709618, "learning_rate": 2.736848431800654e-07, "loss": 0.6799, "step": 25611 }, { "epoch": 0.9276349148859109, "grad_norm": 1.3545838320201804, "learning_rate": 2.734123442530312e-07, "loss": 0.5969, "step": 25612 }, { "epoch": 0.9276711336472293, "grad_norm": 1.0369463356798374, "learning_rate": 2.731399791723888e-07, "loss": 0.6586, "step": 25613 }, { "epoch": 0.9277073524085476, "grad_norm": 1.0887201767787285, "learning_rate": 2.728677479418873e-07, "loss": 0.6261, "step": 25614 }, { "epoch": 0.9277435711698659, "grad_norm": 1.3739289588140413, "learning_rate": 2.725956505652727e-07, "loss": 0.6773, "step": 25615 }, { "epoch": 0.9277797899311844, "grad_norm": 0.9209653420418964, "learning_rate": 2.7232368704628974e-07, "loss": 0.618, "step": 25616 }, { "epoch": 0.9278160086925027, "grad_norm": 1.0594731698354483, "learning_rate": 2.720518573886799e-07, "loss": 0.6805, "step": 25617 }, { "epoch": 0.9278522274538211, "grad_norm": 2.1021869548858145, "learning_rate": 2.717801615961846e-07, "loss": 0.6441, "step": 25618 }, { "epoch": 0.9278884462151394, "grad_norm": 1.324462453956356, "learning_rate": 2.7150859967254196e-07, "loss": 0.6561, "step": 25619 }, { "epoch": 0.9279246649764578, "grad_norm": 1.3595880836067382, "learning_rate": 2.71237171621489e-07, "loss": 0.6821, "step": 25620 }, { "epoch": 0.9279608837377762, "grad_norm": 0.9440029763197093, "learning_rate": 2.709658774467605e-07, "loss": 0.6221, "step": 25621 }, { "epoch": 0.9279971024990945, "grad_norm": 1.321722660124303, "learning_rate": 2.706947171520913e-07, "loss": 0.5882, "step": 25622 }, { "epoch": 0.9280333212604129, "grad_norm": 1.3119324460373631, "learning_rate": 2.7042369074121164e-07, "loss": 0.6562, "step": 25623 }, { "epoch": 0.9280695400217313, "grad_norm": 1.5100185742502934, "learning_rate": 2.701527982178509e-07, "loss": 0.7295, "step": 25624 }, { "epoch": 0.9281057587830496, "grad_norm": 1.3662175225405253, "learning_rate": 2.69882039585736e-07, "loss": 0.706, "step": 25625 }, { "epoch": 0.928141977544368, "grad_norm": 0.9596454240348536, "learning_rate": 2.696114148485951e-07, "loss": 0.5922, "step": 25626 }, { "epoch": 0.9281781963056863, "grad_norm": 1.1329529327357055, "learning_rate": 2.6934092401015187e-07, "loss": 0.6508, "step": 25627 }, { "epoch": 0.9282144150670047, "grad_norm": 1.2475031368224778, "learning_rate": 2.690705670741256e-07, "loss": 0.5624, "step": 25628 }, { "epoch": 0.9282506338283231, "grad_norm": 1.46885892725368, "learning_rate": 2.6880034404423995e-07, "loss": 0.6714, "step": 25629 }, { "epoch": 0.9282868525896414, "grad_norm": 1.3776565351309606, "learning_rate": 2.685302549242119e-07, "loss": 0.6266, "step": 25630 }, { "epoch": 0.9283230713509598, "grad_norm": 1.4157124503915683, "learning_rate": 2.682602997177586e-07, "loss": 0.6847, "step": 25631 }, { "epoch": 0.9283592901122781, "grad_norm": 1.21108220143374, "learning_rate": 2.6799047842859247e-07, "loss": 0.6473, "step": 25632 }, { "epoch": 0.9283955088735966, "grad_norm": 1.5037962117748425, "learning_rate": 2.677207910604307e-07, "loss": 0.66, "step": 25633 }, { "epoch": 0.9284317276349149, "grad_norm": 1.4418125523283554, "learning_rate": 2.674512376169824e-07, "loss": 0.6234, "step": 25634 }, { "epoch": 0.9284679463962332, "grad_norm": 1.3875193138647008, "learning_rate": 2.67181818101957e-07, "loss": 0.6814, "step": 25635 }, { "epoch": 0.9285041651575516, "grad_norm": 1.0081438149065125, "learning_rate": 2.6691253251906023e-07, "loss": 0.6672, "step": 25636 }, { "epoch": 0.92854038391887, "grad_norm": 1.2988056051840318, "learning_rate": 2.6664338087200146e-07, "loss": 0.619, "step": 25637 }, { "epoch": 0.9285766026801884, "grad_norm": 1.3108650594436069, "learning_rate": 2.6637436316448106e-07, "loss": 0.6709, "step": 25638 }, { "epoch": 0.9286128214415067, "grad_norm": 0.9929065200165248, "learning_rate": 2.661054794002016e-07, "loss": 0.6257, "step": 25639 }, { "epoch": 0.928649040202825, "grad_norm": 1.1116261471274325, "learning_rate": 2.6583672958286454e-07, "loss": 0.6394, "step": 25640 }, { "epoch": 0.9286852589641434, "grad_norm": 1.3336196040444708, "learning_rate": 2.65568113716167e-07, "loss": 0.6389, "step": 25641 }, { "epoch": 0.9287214777254618, "grad_norm": 1.5405717024182615, "learning_rate": 2.652996318038059e-07, "loss": 0.7194, "step": 25642 }, { "epoch": 0.9287576964867802, "grad_norm": 1.2664374037034112, "learning_rate": 2.650312838494751e-07, "loss": 0.6151, "step": 25643 }, { "epoch": 0.9287939152480985, "grad_norm": 1.2979124914582312, "learning_rate": 2.647630698568682e-07, "loss": 0.5594, "step": 25644 }, { "epoch": 0.9288301340094168, "grad_norm": 1.5259674150355578, "learning_rate": 2.644949898296756e-07, "loss": 0.6998, "step": 25645 }, { "epoch": 0.9288663527707353, "grad_norm": 1.3830052001760371, "learning_rate": 2.642270437715844e-07, "loss": 0.6753, "step": 25646 }, { "epoch": 0.9289025715320536, "grad_norm": 1.4079454175607344, "learning_rate": 2.63959231686286e-07, "loss": 0.6948, "step": 25647 }, { "epoch": 0.928938790293372, "grad_norm": 1.4361653481321504, "learning_rate": 2.6369155357746315e-07, "loss": 0.6832, "step": 25648 }, { "epoch": 0.9289750090546903, "grad_norm": 1.419563185186174, "learning_rate": 2.634240094487994e-07, "loss": 0.6459, "step": 25649 }, { "epoch": 0.9290112278160086, "grad_norm": 1.3514473258266275, "learning_rate": 2.631565993039753e-07, "loss": 0.647, "step": 25650 }, { "epoch": 0.9290474465773271, "grad_norm": 1.3371783163254998, "learning_rate": 2.6288932314667336e-07, "loss": 0.6648, "step": 25651 }, { "epoch": 0.9290836653386454, "grad_norm": 1.3922238802197007, "learning_rate": 2.6262218098056957e-07, "loss": 0.7203, "step": 25652 }, { "epoch": 0.9291198840999638, "grad_norm": 1.3843164219798336, "learning_rate": 2.6235517280934097e-07, "loss": 0.6849, "step": 25653 }, { "epoch": 0.9291561028612821, "grad_norm": 1.4984230056714036, "learning_rate": 2.620882986366602e-07, "loss": 0.6556, "step": 25654 }, { "epoch": 0.9291923216226005, "grad_norm": 1.5442803916238652, "learning_rate": 2.6182155846620205e-07, "loss": 0.6851, "step": 25655 }, { "epoch": 0.9292285403839189, "grad_norm": 1.1504907694733342, "learning_rate": 2.61554952301637e-07, "loss": 0.7271, "step": 25656 }, { "epoch": 0.9292647591452372, "grad_norm": 1.3854703334103646, "learning_rate": 2.612884801466298e-07, "loss": 0.6472, "step": 25657 }, { "epoch": 0.9293009779065556, "grad_norm": 1.2426433873778964, "learning_rate": 2.610221420048531e-07, "loss": 0.6699, "step": 25658 }, { "epoch": 0.929337196667874, "grad_norm": 1.3588880257252403, "learning_rate": 2.6075593787996844e-07, "loss": 0.6583, "step": 25659 }, { "epoch": 0.9293734154291923, "grad_norm": 1.4311823729504827, "learning_rate": 2.604898677756396e-07, "loss": 0.6805, "step": 25660 }, { "epoch": 0.9294096341905107, "grad_norm": 1.3613968699103767, "learning_rate": 2.6022393169552687e-07, "loss": 0.5838, "step": 25661 }, { "epoch": 0.929445852951829, "grad_norm": 1.543883920229531, "learning_rate": 2.599581296432918e-07, "loss": 0.6227, "step": 25662 }, { "epoch": 0.9294820717131475, "grad_norm": 1.462453083438152, "learning_rate": 2.5969246162259156e-07, "loss": 0.6737, "step": 25663 }, { "epoch": 0.9295182904744658, "grad_norm": 1.022161834932461, "learning_rate": 2.5942692763708197e-07, "loss": 0.6369, "step": 25664 }, { "epoch": 0.9295545092357841, "grad_norm": 1.3921403598082767, "learning_rate": 2.591615276904158e-07, "loss": 0.685, "step": 25665 }, { "epoch": 0.9295907279971025, "grad_norm": 1.4806038975993188, "learning_rate": 2.588962617862467e-07, "loss": 0.641, "step": 25666 }, { "epoch": 0.9296269467584208, "grad_norm": 1.4677757560117763, "learning_rate": 2.58631129928224e-07, "loss": 0.669, "step": 25667 }, { "epoch": 0.9296631655197393, "grad_norm": 1.3909534591499741, "learning_rate": 2.5836613211999593e-07, "loss": 0.6238, "step": 25668 }, { "epoch": 0.9296993842810576, "grad_norm": 1.0764025454047146, "learning_rate": 2.5810126836520954e-07, "loss": 0.7368, "step": 25669 }, { "epoch": 0.9297356030423759, "grad_norm": 1.316250030575541, "learning_rate": 2.578365386675108e-07, "loss": 0.5937, "step": 25670 }, { "epoch": 0.9297718218036943, "grad_norm": 0.9733273169833409, "learning_rate": 2.575719430305412e-07, "loss": 0.6396, "step": 25671 }, { "epoch": 0.9298080405650127, "grad_norm": 1.3577757449602252, "learning_rate": 2.5730748145794126e-07, "loss": 0.6623, "step": 25672 }, { "epoch": 0.9298442593263311, "grad_norm": 1.344643620057216, "learning_rate": 2.5704315395335245e-07, "loss": 0.629, "step": 25673 }, { "epoch": 0.9298804780876494, "grad_norm": 1.6269227113205122, "learning_rate": 2.567789605204096e-07, "loss": 0.714, "step": 25674 }, { "epoch": 0.9299166968489677, "grad_norm": 1.398964522858867, "learning_rate": 2.565149011627499e-07, "loss": 0.6507, "step": 25675 }, { "epoch": 0.9299529156102861, "grad_norm": 1.5161672954182541, "learning_rate": 2.5625097588400596e-07, "loss": 0.635, "step": 25676 }, { "epoch": 0.9299891343716045, "grad_norm": 1.4059230424350029, "learning_rate": 2.5598718468781146e-07, "loss": 0.6815, "step": 25677 }, { "epoch": 0.9300253531329229, "grad_norm": 1.575098048906681, "learning_rate": 2.5572352757779474e-07, "loss": 0.7462, "step": 25678 }, { "epoch": 0.9300615718942412, "grad_norm": 1.0191311903726483, "learning_rate": 2.554600045575828e-07, "loss": 0.6754, "step": 25679 }, { "epoch": 0.9300977906555595, "grad_norm": 1.3396409405338503, "learning_rate": 2.55196615630805e-07, "loss": 0.7243, "step": 25680 }, { "epoch": 0.930134009416878, "grad_norm": 1.130505979432579, "learning_rate": 2.54933360801084e-07, "loss": 0.7083, "step": 25681 }, { "epoch": 0.9301702281781963, "grad_norm": 1.5825049727163631, "learning_rate": 2.5467024007204355e-07, "loss": 0.6016, "step": 25682 }, { "epoch": 0.9302064469395147, "grad_norm": 1.4706370970860336, "learning_rate": 2.544072534473019e-07, "loss": 0.6363, "step": 25683 }, { "epoch": 0.930242665700833, "grad_norm": 1.308064546617794, "learning_rate": 2.541444009304805e-07, "loss": 0.6078, "step": 25684 }, { "epoch": 0.9302788844621513, "grad_norm": 1.4080875351934035, "learning_rate": 2.5388168252519553e-07, "loss": 0.6864, "step": 25685 }, { "epoch": 0.9303151032234698, "grad_norm": 1.945317238683176, "learning_rate": 2.536190982350628e-07, "loss": 0.6761, "step": 25686 }, { "epoch": 0.9303513219847881, "grad_norm": 1.4706317123107102, "learning_rate": 2.5335664806369397e-07, "loss": 0.685, "step": 25687 }, { "epoch": 0.9303875407461065, "grad_norm": 1.4374051218234707, "learning_rate": 2.530943320147028e-07, "loss": 0.6381, "step": 25688 }, { "epoch": 0.9304237595074248, "grad_norm": 1.4217999931086198, "learning_rate": 2.528321500916986e-07, "loss": 0.7156, "step": 25689 }, { "epoch": 0.9304599782687432, "grad_norm": 1.2992752252164672, "learning_rate": 2.5257010229828735e-07, "loss": 0.7107, "step": 25690 }, { "epoch": 0.9304961970300616, "grad_norm": 1.4100104782335718, "learning_rate": 2.523081886380763e-07, "loss": 0.6767, "step": 25691 }, { "epoch": 0.9305324157913799, "grad_norm": 1.3826404042833462, "learning_rate": 2.5204640911467125e-07, "loss": 0.6722, "step": 25692 }, { "epoch": 0.9305686345526983, "grad_norm": 1.3586502986664606, "learning_rate": 2.5178476373167173e-07, "loss": 0.6869, "step": 25693 }, { "epoch": 0.9306048533140167, "grad_norm": 1.2557871710279458, "learning_rate": 2.5152325249267916e-07, "loss": 0.613, "step": 25694 }, { "epoch": 0.930641072075335, "grad_norm": 1.4151952853425647, "learning_rate": 2.51261875401293e-07, "loss": 0.6786, "step": 25695 }, { "epoch": 0.9306772908366534, "grad_norm": 1.306145381730679, "learning_rate": 2.5100063246110915e-07, "loss": 0.6772, "step": 25696 }, { "epoch": 0.9307135095979717, "grad_norm": 1.0829431160408758, "learning_rate": 2.5073952367572265e-07, "loss": 0.6244, "step": 25697 }, { "epoch": 0.9307497283592902, "grad_norm": 1.6336153260661974, "learning_rate": 2.5047854904872603e-07, "loss": 0.7352, "step": 25698 }, { "epoch": 0.9307859471206085, "grad_norm": 1.452080776354533, "learning_rate": 2.502177085837132e-07, "loss": 0.7562, "step": 25699 }, { "epoch": 0.9308221658819268, "grad_norm": 1.3481435333629592, "learning_rate": 2.499570022842701e-07, "loss": 0.6458, "step": 25700 }, { "epoch": 0.9308583846432452, "grad_norm": 1.425755904246755, "learning_rate": 2.4969643015398613e-07, "loss": 0.6646, "step": 25701 }, { "epoch": 0.9308946034045635, "grad_norm": 1.4292101871029195, "learning_rate": 2.4943599219644733e-07, "loss": 0.6615, "step": 25702 }, { "epoch": 0.930930822165882, "grad_norm": 1.3655604026408537, "learning_rate": 2.491756884152363e-07, "loss": 0.6505, "step": 25703 }, { "epoch": 0.9309670409272003, "grad_norm": 1.4417862352522781, "learning_rate": 2.4891551881393585e-07, "loss": 0.6401, "step": 25704 }, { "epoch": 0.9310032596885186, "grad_norm": 1.519481008429651, "learning_rate": 2.486554833961252e-07, "loss": 0.6633, "step": 25705 }, { "epoch": 0.931039478449837, "grad_norm": 1.3990967859068482, "learning_rate": 2.4839558216538386e-07, "loss": 0.7342, "step": 25706 }, { "epoch": 0.9310756972111554, "grad_norm": 1.6662962186410053, "learning_rate": 2.481358151252877e-07, "loss": 0.6911, "step": 25707 }, { "epoch": 0.9311119159724738, "grad_norm": 1.3112733961905483, "learning_rate": 2.478761822794118e-07, "loss": 0.7023, "step": 25708 }, { "epoch": 0.9311481347337921, "grad_norm": 1.3079772621541308, "learning_rate": 2.4761668363132873e-07, "loss": 0.6519, "step": 25709 }, { "epoch": 0.9311843534951104, "grad_norm": 1.0635009657244459, "learning_rate": 2.4735731918460906e-07, "loss": 0.6324, "step": 25710 }, { "epoch": 0.9312205722564288, "grad_norm": 1.008278493253885, "learning_rate": 2.4709808894282203e-07, "loss": 0.7479, "step": 25711 }, { "epoch": 0.9312567910177472, "grad_norm": 1.7772793118879124, "learning_rate": 2.468389929095338e-07, "loss": 0.6609, "step": 25712 }, { "epoch": 0.9312930097790656, "grad_norm": 1.3698750653793375, "learning_rate": 2.4658003108831264e-07, "loss": 0.6791, "step": 25713 }, { "epoch": 0.9313292285403839, "grad_norm": 1.3456670111944153, "learning_rate": 2.4632120348272e-07, "loss": 0.5727, "step": 25714 }, { "epoch": 0.9313654473017022, "grad_norm": 1.3815962553247272, "learning_rate": 2.4606251009631767e-07, "loss": 0.6558, "step": 25715 }, { "epoch": 0.9314016660630207, "grad_norm": 1.3569785903712404, "learning_rate": 2.4580395093266483e-07, "loss": 0.6928, "step": 25716 }, { "epoch": 0.931437884824339, "grad_norm": 1.5537464610126237, "learning_rate": 2.455455259953221e-07, "loss": 0.7033, "step": 25717 }, { "epoch": 0.9314741035856574, "grad_norm": 1.4044110725245167, "learning_rate": 2.4528723528784326e-07, "loss": 0.6173, "step": 25718 }, { "epoch": 0.9315103223469757, "grad_norm": 1.3435940575768448, "learning_rate": 2.4502907881378326e-07, "loss": 0.6133, "step": 25719 }, { "epoch": 0.931546541108294, "grad_norm": 1.3091715953143723, "learning_rate": 2.4477105657669477e-07, "loss": 0.6423, "step": 25720 }, { "epoch": 0.9315827598696125, "grad_norm": 1.3379496568339655, "learning_rate": 2.4451316858012833e-07, "loss": 0.6987, "step": 25721 }, { "epoch": 0.9316189786309308, "grad_norm": 1.5212696888667456, "learning_rate": 2.4425541482763217e-07, "loss": 0.7338, "step": 25722 }, { "epoch": 0.9316551973922492, "grad_norm": 1.5456777693466979, "learning_rate": 2.439977953227535e-07, "loss": 0.6992, "step": 25723 }, { "epoch": 0.9316914161535675, "grad_norm": 1.4871052271998293, "learning_rate": 2.437403100690372e-07, "loss": 0.6629, "step": 25724 }, { "epoch": 0.9317276349148859, "grad_norm": 1.5615672331412485, "learning_rate": 2.434829590700283e-07, "loss": 0.6456, "step": 25725 }, { "epoch": 0.9317638536762043, "grad_norm": 1.0661821912695963, "learning_rate": 2.432257423292661e-07, "loss": 0.6546, "step": 25726 }, { "epoch": 0.9318000724375226, "grad_norm": 1.2341348831005514, "learning_rate": 2.4296865985029007e-07, "loss": 0.6422, "step": 25727 }, { "epoch": 0.931836291198841, "grad_norm": 0.9505927651575345, "learning_rate": 2.4271171163663954e-07, "loss": 0.7117, "step": 25728 }, { "epoch": 0.9318725099601594, "grad_norm": 1.083905723495642, "learning_rate": 2.424548976918484e-07, "loss": 0.6341, "step": 25729 }, { "epoch": 0.9319087287214777, "grad_norm": 1.498403152463971, "learning_rate": 2.4219821801945156e-07, "loss": 0.6317, "step": 25730 }, { "epoch": 0.9319449474827961, "grad_norm": 1.4134919717477719, "learning_rate": 2.4194167262298176e-07, "loss": 0.6487, "step": 25731 }, { "epoch": 0.9319811662441144, "grad_norm": 1.2638725397075947, "learning_rate": 2.4168526150596947e-07, "loss": 0.6447, "step": 25732 }, { "epoch": 0.9320173850054329, "grad_norm": 1.0584246055580222, "learning_rate": 2.41428984671942e-07, "loss": 0.7021, "step": 25733 }, { "epoch": 0.9320536037667512, "grad_norm": 1.3439973493324664, "learning_rate": 2.4117284212442413e-07, "loss": 0.6687, "step": 25734 }, { "epoch": 0.9320898225280695, "grad_norm": 2.008622690672606, "learning_rate": 2.4091683386694544e-07, "loss": 0.6161, "step": 25735 }, { "epoch": 0.9321260412893879, "grad_norm": 1.4282286691131525, "learning_rate": 2.406609599030252e-07, "loss": 0.658, "step": 25736 }, { "epoch": 0.9321622600507062, "grad_norm": 1.3541067822371025, "learning_rate": 2.404052202361862e-07, "loss": 0.641, "step": 25737 }, { "epoch": 0.9321984788120247, "grad_norm": 1.0200143646070121, "learning_rate": 2.401496148699456e-07, "loss": 0.6513, "step": 25738 }, { "epoch": 0.932234697573343, "grad_norm": 1.0465943143464986, "learning_rate": 2.3989414380782285e-07, "loss": 0.6828, "step": 25739 }, { "epoch": 0.9322709163346613, "grad_norm": 1.1927983460512266, "learning_rate": 2.3963880705333285e-07, "loss": 0.6486, "step": 25740 }, { "epoch": 0.9323071350959797, "grad_norm": 1.03748398040318, "learning_rate": 2.3938360460998846e-07, "loss": 0.7193, "step": 25741 }, { "epoch": 0.932343353857298, "grad_norm": 0.9812815801254279, "learning_rate": 2.391285364813023e-07, "loss": 0.7029, "step": 25742 }, { "epoch": 0.9323795726186165, "grad_norm": 1.3308112193980501, "learning_rate": 2.38873602670785e-07, "loss": 0.6911, "step": 25743 }, { "epoch": 0.9324157913799348, "grad_norm": 0.9907408948390335, "learning_rate": 2.386188031819425e-07, "loss": 0.5852, "step": 25744 }, { "epoch": 0.9324520101412531, "grad_norm": 0.9373406202097231, "learning_rate": 2.3836413801828217e-07, "loss": 0.6222, "step": 25745 }, { "epoch": 0.9324882289025715, "grad_norm": 1.0571363427636233, "learning_rate": 2.3810960718331e-07, "loss": 0.6908, "step": 25746 }, { "epoch": 0.9325244476638899, "grad_norm": 1.3553384393941246, "learning_rate": 2.3785521068052652e-07, "loss": 0.6331, "step": 25747 }, { "epoch": 0.9325606664252083, "grad_norm": 1.4242233635685908, "learning_rate": 2.376009485134334e-07, "loss": 0.6892, "step": 25748 }, { "epoch": 0.9325968851865266, "grad_norm": 1.1977084662864506, "learning_rate": 2.3734682068552893e-07, "loss": 0.6266, "step": 25749 }, { "epoch": 0.9326331039478449, "grad_norm": 1.3482620743433322, "learning_rate": 2.3709282720031035e-07, "loss": 0.7112, "step": 25750 }, { "epoch": 0.9326693227091634, "grad_norm": 1.4166082431220712, "learning_rate": 2.368389680612737e-07, "loss": 0.6206, "step": 25751 }, { "epoch": 0.9327055414704817, "grad_norm": 1.3368278800265192, "learning_rate": 2.3658524327190957e-07, "loss": 0.6258, "step": 25752 }, { "epoch": 0.9327417602318001, "grad_norm": 1.3066217131735742, "learning_rate": 2.3633165283571292e-07, "loss": 0.5608, "step": 25753 }, { "epoch": 0.9327779789931184, "grad_norm": 1.0908154795504599, "learning_rate": 2.3607819675617204e-07, "loss": 0.7003, "step": 25754 }, { "epoch": 0.9328141977544367, "grad_norm": 1.496554651056586, "learning_rate": 2.358248750367742e-07, "loss": 0.6914, "step": 25755 }, { "epoch": 0.9328504165157552, "grad_norm": 1.4541161775334879, "learning_rate": 2.3557168768100548e-07, "loss": 0.617, "step": 25756 }, { "epoch": 0.9328866352770735, "grad_norm": 1.0140954734766492, "learning_rate": 2.3531863469234974e-07, "loss": 0.6214, "step": 25757 }, { "epoch": 0.9329228540383919, "grad_norm": 1.3340539760500465, "learning_rate": 2.3506571607429085e-07, "loss": 0.6114, "step": 25758 }, { "epoch": 0.9329590727997102, "grad_norm": 1.414254176863881, "learning_rate": 2.3481293183030718e-07, "loss": 0.7073, "step": 25759 }, { "epoch": 0.9329952915610286, "grad_norm": 1.324391285591169, "learning_rate": 2.3456028196387703e-07, "loss": 0.6921, "step": 25760 }, { "epoch": 0.933031510322347, "grad_norm": 1.6104870525799406, "learning_rate": 2.3430776647847876e-07, "loss": 0.7149, "step": 25761 }, { "epoch": 0.9330677290836653, "grad_norm": 1.438517966778448, "learning_rate": 2.340553853775873e-07, "loss": 0.6613, "step": 25762 }, { "epoch": 0.9331039478449837, "grad_norm": 1.1450641244671973, "learning_rate": 2.338031386646733e-07, "loss": 0.7308, "step": 25763 }, { "epoch": 0.9331401666063021, "grad_norm": 1.432562275089619, "learning_rate": 2.335510263432106e-07, "loss": 0.645, "step": 25764 }, { "epoch": 0.9331763853676204, "grad_norm": 1.2558241879293255, "learning_rate": 2.3329904841666752e-07, "loss": 0.564, "step": 25765 }, { "epoch": 0.9332126041289388, "grad_norm": 1.2396712120912723, "learning_rate": 2.3304720488851128e-07, "loss": 0.6406, "step": 25766 }, { "epoch": 0.9332488228902571, "grad_norm": 1.3501407940614176, "learning_rate": 2.3279549576220584e-07, "loss": 0.6376, "step": 25767 }, { "epoch": 0.9332850416515756, "grad_norm": 1.3034271738646628, "learning_rate": 2.3254392104121837e-07, "loss": 0.6321, "step": 25768 }, { "epoch": 0.9333212604128939, "grad_norm": 1.2783974568874392, "learning_rate": 2.3229248072900835e-07, "loss": 0.678, "step": 25769 }, { "epoch": 0.9333574791742122, "grad_norm": 1.2768222855583164, "learning_rate": 2.3204117482903633e-07, "loss": 0.6762, "step": 25770 }, { "epoch": 0.9333936979355306, "grad_norm": 1.3965210482977126, "learning_rate": 2.3179000334476064e-07, "loss": 0.6663, "step": 25771 }, { "epoch": 0.9334299166968489, "grad_norm": 1.444188298719787, "learning_rate": 2.3153896627963856e-07, "loss": 0.7042, "step": 25772 }, { "epoch": 0.9334661354581674, "grad_norm": 1.5491100989036146, "learning_rate": 2.3128806363712287e-07, "loss": 0.7087, "step": 25773 }, { "epoch": 0.9335023542194857, "grad_norm": 1.226704396404525, "learning_rate": 2.3103729542066633e-07, "loss": 0.595, "step": 25774 }, { "epoch": 0.933538572980804, "grad_norm": 1.4315461847087143, "learning_rate": 2.3078666163372176e-07, "loss": 0.6759, "step": 25775 }, { "epoch": 0.9335747917421224, "grad_norm": 1.2473515527254908, "learning_rate": 2.305361622797353e-07, "loss": 0.6479, "step": 25776 }, { "epoch": 0.9336110105034408, "grad_norm": 0.9298423605054186, "learning_rate": 2.3028579736215638e-07, "loss": 0.6171, "step": 25777 }, { "epoch": 0.9336472292647592, "grad_norm": 1.4664120672160232, "learning_rate": 2.3003556688442897e-07, "loss": 0.6811, "step": 25778 }, { "epoch": 0.9336834480260775, "grad_norm": 1.2728667823311208, "learning_rate": 2.2978547084999693e-07, "loss": 0.6339, "step": 25779 }, { "epoch": 0.9337196667873958, "grad_norm": 1.4692346009309463, "learning_rate": 2.2953550926230195e-07, "loss": 0.6432, "step": 25780 }, { "epoch": 0.9337558855487142, "grad_norm": 1.6447110700721357, "learning_rate": 2.2928568212478242e-07, "loss": 0.69, "step": 25781 }, { "epoch": 0.9337921043100326, "grad_norm": 1.5262016312921842, "learning_rate": 2.290359894408778e-07, "loss": 0.6368, "step": 25782 }, { "epoch": 0.933828323071351, "grad_norm": 1.5738288512504002, "learning_rate": 2.287864312140242e-07, "loss": 0.7306, "step": 25783 }, { "epoch": 0.9338645418326693, "grad_norm": 0.994871632920179, "learning_rate": 2.285370074476545e-07, "loss": 0.6784, "step": 25784 }, { "epoch": 0.9339007605939876, "grad_norm": 0.9759931287934978, "learning_rate": 2.2828771814520145e-07, "loss": 0.6186, "step": 25785 }, { "epoch": 0.9339369793553061, "grad_norm": 1.4986718980208218, "learning_rate": 2.2803856331009566e-07, "loss": 0.6448, "step": 25786 }, { "epoch": 0.9339731981166244, "grad_norm": 1.3757491028662077, "learning_rate": 2.277895429457655e-07, "loss": 0.6502, "step": 25787 }, { "epoch": 0.9340094168779428, "grad_norm": 1.1232771614043644, "learning_rate": 2.2754065705563822e-07, "loss": 0.6893, "step": 25788 }, { "epoch": 0.9340456356392611, "grad_norm": 1.4786297188481272, "learning_rate": 2.2729190564313774e-07, "loss": 0.7247, "step": 25789 }, { "epoch": 0.9340818544005794, "grad_norm": 1.5853895358831607, "learning_rate": 2.27043288711688e-07, "loss": 0.6879, "step": 25790 }, { "epoch": 0.9341180731618979, "grad_norm": 1.4558286689731983, "learning_rate": 2.2679480626470963e-07, "loss": 0.5703, "step": 25791 }, { "epoch": 0.9341542919232162, "grad_norm": 1.3082746913877183, "learning_rate": 2.2654645830562205e-07, "loss": 0.6097, "step": 25792 }, { "epoch": 0.9341905106845346, "grad_norm": 1.5017801415843188, "learning_rate": 2.262982448378437e-07, "loss": 0.6271, "step": 25793 }, { "epoch": 0.9342267294458529, "grad_norm": 1.3940247516143263, "learning_rate": 2.2605016586478957e-07, "loss": 0.6808, "step": 25794 }, { "epoch": 0.9342629482071713, "grad_norm": 1.3567706508203141, "learning_rate": 2.2580222138987252e-07, "loss": 0.6196, "step": 25795 }, { "epoch": 0.9342991669684897, "grad_norm": 1.3488991520638678, "learning_rate": 2.2555441141650536e-07, "loss": 0.6802, "step": 25796 }, { "epoch": 0.934335385729808, "grad_norm": 1.0356131790976129, "learning_rate": 2.253067359480987e-07, "loss": 0.6909, "step": 25797 }, { "epoch": 0.9343716044911264, "grad_norm": 1.4053260337418823, "learning_rate": 2.2505919498805985e-07, "loss": 0.6111, "step": 25798 }, { "epoch": 0.9344078232524448, "grad_norm": 0.957768831499269, "learning_rate": 2.248117885397949e-07, "loss": 0.6215, "step": 25799 }, { "epoch": 0.9344440420137631, "grad_norm": 1.41604233157525, "learning_rate": 2.2456451660670897e-07, "loss": 0.6691, "step": 25800 }, { "epoch": 0.9344802607750815, "grad_norm": 1.2642510179979143, "learning_rate": 2.2431737919220597e-07, "loss": 0.6816, "step": 25801 }, { "epoch": 0.9345164795363998, "grad_norm": 1.2281350249750047, "learning_rate": 2.2407037629968431e-07, "loss": 0.6152, "step": 25802 }, { "epoch": 0.9345526982977183, "grad_norm": 1.4321982994136466, "learning_rate": 2.238235079325446e-07, "loss": 0.6209, "step": 25803 }, { "epoch": 0.9345889170590366, "grad_norm": 1.0936500888984817, "learning_rate": 2.23576774094183e-07, "loss": 0.7158, "step": 25804 }, { "epoch": 0.9346251358203549, "grad_norm": 0.9550778292317118, "learning_rate": 2.233301747879968e-07, "loss": 0.6654, "step": 25805 }, { "epoch": 0.9346613545816733, "grad_norm": 1.0254188841944782, "learning_rate": 2.230837100173766e-07, "loss": 0.7171, "step": 25806 }, { "epoch": 0.9346975733429916, "grad_norm": 1.411399128476879, "learning_rate": 2.228373797857153e-07, "loss": 0.6897, "step": 25807 }, { "epoch": 0.9347337921043101, "grad_norm": 0.9738429540062854, "learning_rate": 2.2259118409640235e-07, "loss": 0.7012, "step": 25808 }, { "epoch": 0.9347700108656284, "grad_norm": 1.206380585025194, "learning_rate": 2.223451229528273e-07, "loss": 0.7431, "step": 25809 }, { "epoch": 0.9348062296269467, "grad_norm": 1.3786408416150637, "learning_rate": 2.2209919635837406e-07, "loss": 0.721, "step": 25810 }, { "epoch": 0.9348424483882651, "grad_norm": 1.3888183696215588, "learning_rate": 2.2185340431642666e-07, "loss": 0.6725, "step": 25811 }, { "epoch": 0.9348786671495835, "grad_norm": 1.4243298877474284, "learning_rate": 2.2160774683036902e-07, "loss": 0.6898, "step": 25812 }, { "epoch": 0.9349148859109019, "grad_norm": 1.0418519230792036, "learning_rate": 2.2136222390358175e-07, "loss": 0.7183, "step": 25813 }, { "epoch": 0.9349511046722202, "grad_norm": 1.3512990631407444, "learning_rate": 2.2111683553943996e-07, "loss": 0.6439, "step": 25814 }, { "epoch": 0.9349873234335385, "grad_norm": 1.0148276886591785, "learning_rate": 2.208715817413254e-07, "loss": 0.7012, "step": 25815 }, { "epoch": 0.935023542194857, "grad_norm": 1.0045779414668699, "learning_rate": 2.2062646251260977e-07, "loss": 0.6825, "step": 25816 }, { "epoch": 0.9350597609561753, "grad_norm": 1.1416006944082417, "learning_rate": 2.2038147785666708e-07, "loss": 0.6865, "step": 25817 }, { "epoch": 0.9350959797174937, "grad_norm": 1.3148710215727124, "learning_rate": 2.2013662777686906e-07, "loss": 0.6218, "step": 25818 }, { "epoch": 0.935132198478812, "grad_norm": 1.4178126146721324, "learning_rate": 2.19891912276583e-07, "loss": 0.7, "step": 25819 }, { "epoch": 0.9351684172401303, "grad_norm": 1.071380691024139, "learning_rate": 2.196473313591796e-07, "loss": 0.6337, "step": 25820 }, { "epoch": 0.9352046360014488, "grad_norm": 1.337277264854985, "learning_rate": 2.1940288502802054e-07, "loss": 0.6808, "step": 25821 }, { "epoch": 0.9352408547627671, "grad_norm": 1.366237543493248, "learning_rate": 2.1915857328647205e-07, "loss": 0.7389, "step": 25822 }, { "epoch": 0.9352770735240855, "grad_norm": 1.0409906172521992, "learning_rate": 2.18914396137897e-07, "loss": 0.6187, "step": 25823 }, { "epoch": 0.9353132922854038, "grad_norm": 1.351831413539471, "learning_rate": 2.1867035358565268e-07, "loss": 0.6457, "step": 25824 }, { "epoch": 0.9353495110467221, "grad_norm": 1.525361142624567, "learning_rate": 2.1842644563310088e-07, "loss": 0.6279, "step": 25825 }, { "epoch": 0.9353857298080406, "grad_norm": 0.9038785236834668, "learning_rate": 2.1818267228359557e-07, "loss": 0.6298, "step": 25826 }, { "epoch": 0.9354219485693589, "grad_norm": 1.5323289073609554, "learning_rate": 2.1793903354049074e-07, "loss": 0.6658, "step": 25827 }, { "epoch": 0.9354581673306773, "grad_norm": 1.3114295401800888, "learning_rate": 2.1769552940714034e-07, "loss": 0.6607, "step": 25828 }, { "epoch": 0.9354943860919956, "grad_norm": 1.4029469155877927, "learning_rate": 2.1745215988689506e-07, "loss": 0.6685, "step": 25829 }, { "epoch": 0.935530604853314, "grad_norm": 1.3804434941846861, "learning_rate": 2.172089249831044e-07, "loss": 0.6167, "step": 25830 }, { "epoch": 0.9355668236146324, "grad_norm": 1.5039039004914398, "learning_rate": 2.1696582469911464e-07, "loss": 0.6959, "step": 25831 }, { "epoch": 0.9356030423759507, "grad_norm": 1.433410244300661, "learning_rate": 2.1672285903826974e-07, "loss": 0.6478, "step": 25832 }, { "epoch": 0.9356392611372691, "grad_norm": 1.3580077059814464, "learning_rate": 2.1648002800391587e-07, "loss": 0.6976, "step": 25833 }, { "epoch": 0.9356754798985875, "grad_norm": 1.2748117384113609, "learning_rate": 2.1623733159939375e-07, "loss": 0.6958, "step": 25834 }, { "epoch": 0.9357116986599058, "grad_norm": 1.6005993447431754, "learning_rate": 2.1599476982804179e-07, "loss": 0.7418, "step": 25835 }, { "epoch": 0.9357479174212242, "grad_norm": 1.2466459687945084, "learning_rate": 2.1575234269319957e-07, "loss": 0.6217, "step": 25836 }, { "epoch": 0.9357841361825425, "grad_norm": 1.3630583206485951, "learning_rate": 2.1551005019820213e-07, "loss": 0.6405, "step": 25837 }, { "epoch": 0.935820354943861, "grad_norm": 1.356232412890041, "learning_rate": 2.1526789234638467e-07, "loss": 0.6316, "step": 25838 }, { "epoch": 0.9358565737051793, "grad_norm": 1.57765252646458, "learning_rate": 2.150258691410767e-07, "loss": 0.5962, "step": 25839 }, { "epoch": 0.9358927924664976, "grad_norm": 1.3840652545549272, "learning_rate": 2.1478398058561222e-07, "loss": 0.6424, "step": 25840 }, { "epoch": 0.935929011227816, "grad_norm": 1.3691233499174502, "learning_rate": 2.1454222668331748e-07, "loss": 0.693, "step": 25841 }, { "epoch": 0.9359652299891343, "grad_norm": 1.1418725025473193, "learning_rate": 2.143006074375198e-07, "loss": 0.7454, "step": 25842 }, { "epoch": 0.9360014487504528, "grad_norm": 1.3777492926892907, "learning_rate": 2.140591228515443e-07, "loss": 0.6844, "step": 25843 }, { "epoch": 0.9360376675117711, "grad_norm": 1.3549873791767584, "learning_rate": 2.1381777292871385e-07, "loss": 0.6659, "step": 25844 }, { "epoch": 0.9360738862730894, "grad_norm": 1.4045680013816313, "learning_rate": 2.135765576723503e-07, "loss": 0.7185, "step": 25845 }, { "epoch": 0.9361101050344078, "grad_norm": 1.6203586531791316, "learning_rate": 2.1333547708577207e-07, "loss": 0.7254, "step": 25846 }, { "epoch": 0.9361463237957262, "grad_norm": 1.3884434196883448, "learning_rate": 2.1309453117229652e-07, "loss": 0.6598, "step": 25847 }, { "epoch": 0.9361825425570446, "grad_norm": 1.3469770917789694, "learning_rate": 2.1285371993524097e-07, "loss": 0.6555, "step": 25848 }, { "epoch": 0.9362187613183629, "grad_norm": 1.4012241204100437, "learning_rate": 2.1261304337791722e-07, "loss": 0.6976, "step": 25849 }, { "epoch": 0.9362549800796812, "grad_norm": 1.236813212302866, "learning_rate": 2.123725015036371e-07, "loss": 0.6603, "step": 25850 }, { "epoch": 0.9362911988409996, "grad_norm": 1.25164473201126, "learning_rate": 2.1213209431571346e-07, "loss": 0.6005, "step": 25851 }, { "epoch": 0.936327417602318, "grad_norm": 1.3878017776635958, "learning_rate": 2.118918218174515e-07, "loss": 0.6023, "step": 25852 }, { "epoch": 0.9363636363636364, "grad_norm": 1.281154442195248, "learning_rate": 2.116516840121585e-07, "loss": 0.6458, "step": 25853 }, { "epoch": 0.9363998551249547, "grad_norm": 1.4266893604106174, "learning_rate": 2.1141168090313857e-07, "loss": 0.6322, "step": 25854 }, { "epoch": 0.936436073886273, "grad_norm": 1.002112561275184, "learning_rate": 2.111718124936968e-07, "loss": 0.7332, "step": 25855 }, { "epoch": 0.9364722926475915, "grad_norm": 1.4689162869116104, "learning_rate": 2.109320787871305e-07, "loss": 0.711, "step": 25856 }, { "epoch": 0.9365085114089098, "grad_norm": 1.3765202254365017, "learning_rate": 2.1069247978674046e-07, "loss": 0.6636, "step": 25857 }, { "epoch": 0.9365447301702282, "grad_norm": 1.4346622842659593, "learning_rate": 2.1045301549582398e-07, "loss": 0.672, "step": 25858 }, { "epoch": 0.9365809489315465, "grad_norm": 1.3775209000294852, "learning_rate": 2.102136859176762e-07, "loss": 0.6456, "step": 25859 }, { "epoch": 0.9366171676928648, "grad_norm": 0.9416662612440055, "learning_rate": 2.0997449105558898e-07, "loss": 0.6076, "step": 25860 }, { "epoch": 0.9366533864541833, "grad_norm": 1.583900240111241, "learning_rate": 2.097354309128552e-07, "loss": 0.6763, "step": 25861 }, { "epoch": 0.9366896052155016, "grad_norm": 1.420394591832705, "learning_rate": 2.0949650549276556e-07, "loss": 0.6665, "step": 25862 }, { "epoch": 0.93672582397682, "grad_norm": 1.4298819751004959, "learning_rate": 2.0925771479860635e-07, "loss": 0.7093, "step": 25863 }, { "epoch": 0.9367620427381383, "grad_norm": 1.4350529398357355, "learning_rate": 2.090190588336627e-07, "loss": 0.6202, "step": 25864 }, { "epoch": 0.9367982614994567, "grad_norm": 1.4108140103529745, "learning_rate": 2.0878053760122085e-07, "loss": 0.5933, "step": 25865 }, { "epoch": 0.9368344802607751, "grad_norm": 1.3781279644966677, "learning_rate": 2.0854215110456265e-07, "loss": 0.6823, "step": 25866 }, { "epoch": 0.9368706990220934, "grad_norm": 1.3779844730064992, "learning_rate": 2.083038993469666e-07, "loss": 0.6324, "step": 25867 }, { "epoch": 0.9369069177834118, "grad_norm": 1.4332839360799876, "learning_rate": 2.0806578233171338e-07, "loss": 0.6204, "step": 25868 }, { "epoch": 0.9369431365447302, "grad_norm": 1.3227563254614836, "learning_rate": 2.0782780006207925e-07, "loss": 0.6093, "step": 25869 }, { "epoch": 0.9369793553060485, "grad_norm": 1.3032732203748105, "learning_rate": 2.0758995254133829e-07, "loss": 0.6716, "step": 25870 }, { "epoch": 0.9370155740673669, "grad_norm": 1.4670227958727324, "learning_rate": 2.0735223977276453e-07, "loss": 0.689, "step": 25871 }, { "epoch": 0.9370517928286852, "grad_norm": 1.3741681416546803, "learning_rate": 2.071146617596276e-07, "loss": 0.7092, "step": 25872 }, { "epoch": 0.9370880115900037, "grad_norm": 1.1725087303003274, "learning_rate": 2.0687721850519814e-07, "loss": 0.6814, "step": 25873 }, { "epoch": 0.937124230351322, "grad_norm": 1.493083052249543, "learning_rate": 2.0663991001274363e-07, "loss": 0.6885, "step": 25874 }, { "epoch": 0.9371604491126403, "grad_norm": 0.8840793396856809, "learning_rate": 2.064027362855281e-07, "loss": 0.6812, "step": 25875 }, { "epoch": 0.9371966678739587, "grad_norm": 1.3935767187396653, "learning_rate": 2.0616569732681669e-07, "loss": 0.6297, "step": 25876 }, { "epoch": 0.937232886635277, "grad_norm": 1.3334814823038759, "learning_rate": 2.0592879313987125e-07, "loss": 0.7114, "step": 25877 }, { "epoch": 0.9372691053965955, "grad_norm": 1.193121389919956, "learning_rate": 2.0569202372795138e-07, "loss": 0.6824, "step": 25878 }, { "epoch": 0.9373053241579138, "grad_norm": 1.3142830127868688, "learning_rate": 2.0545538909431452e-07, "loss": 0.6072, "step": 25879 }, { "epoch": 0.9373415429192321, "grad_norm": 1.3462443050355355, "learning_rate": 2.052188892422191e-07, "loss": 0.6137, "step": 25880 }, { "epoch": 0.9373777616805505, "grad_norm": 1.01950856686896, "learning_rate": 2.0498252417491815e-07, "loss": 0.7551, "step": 25881 }, { "epoch": 0.9374139804418689, "grad_norm": 1.5320407664826785, "learning_rate": 2.0474629389566347e-07, "loss": 0.6275, "step": 25882 }, { "epoch": 0.9374501992031873, "grad_norm": 1.424631056556528, "learning_rate": 2.0451019840770692e-07, "loss": 0.6779, "step": 25883 }, { "epoch": 0.9374864179645056, "grad_norm": 1.290952026630126, "learning_rate": 2.04274237714297e-07, "loss": 0.6884, "step": 25884 }, { "epoch": 0.937522636725824, "grad_norm": 1.0168506053769757, "learning_rate": 2.040384118186811e-07, "loss": 0.5966, "step": 25885 }, { "epoch": 0.9375588554871424, "grad_norm": 1.0964263969465153, "learning_rate": 2.0380272072410335e-07, "loss": 0.6709, "step": 25886 }, { "epoch": 0.9375950742484607, "grad_norm": 1.394327622327525, "learning_rate": 2.0356716443380885e-07, "loss": 0.6922, "step": 25887 }, { "epoch": 0.9376312930097791, "grad_norm": 1.45093412813501, "learning_rate": 2.033317429510384e-07, "loss": 0.6735, "step": 25888 }, { "epoch": 0.9376675117710974, "grad_norm": 1.335278066654951, "learning_rate": 2.0309645627903051e-07, "loss": 0.6691, "step": 25889 }, { "epoch": 0.9377037305324158, "grad_norm": 1.3430187706507757, "learning_rate": 2.028613044210237e-07, "loss": 0.6298, "step": 25890 }, { "epoch": 0.9377399492937342, "grad_norm": 1.4925199645091323, "learning_rate": 2.0262628738025314e-07, "loss": 0.6539, "step": 25891 }, { "epoch": 0.9377761680550525, "grad_norm": 1.4979706363103071, "learning_rate": 2.0239140515995515e-07, "loss": 0.691, "step": 25892 }, { "epoch": 0.9378123868163709, "grad_norm": 1.3940283291878104, "learning_rate": 2.0215665776335935e-07, "loss": 0.6593, "step": 25893 }, { "epoch": 0.9378486055776892, "grad_norm": 1.3210520782023072, "learning_rate": 2.0192204519369762e-07, "loss": 0.6157, "step": 25894 }, { "epoch": 0.9378848243390077, "grad_norm": 1.215893159162392, "learning_rate": 2.0168756745419737e-07, "loss": 0.6546, "step": 25895 }, { "epoch": 0.937921043100326, "grad_norm": 1.4141123747065192, "learning_rate": 2.0145322454808603e-07, "loss": 0.6573, "step": 25896 }, { "epoch": 0.9379572618616443, "grad_norm": 1.059466997781197, "learning_rate": 2.0121901647858765e-07, "loss": 0.701, "step": 25897 }, { "epoch": 0.9379934806229627, "grad_norm": 1.4340755506660663, "learning_rate": 2.0098494324892636e-07, "loss": 0.6867, "step": 25898 }, { "epoch": 0.938029699384281, "grad_norm": 1.3521020596452287, "learning_rate": 2.007510048623218e-07, "loss": 0.6387, "step": 25899 }, { "epoch": 0.9380659181455995, "grad_norm": 1.2081685618437326, "learning_rate": 2.0051720132199471e-07, "loss": 0.5759, "step": 25900 }, { "epoch": 0.9381021369069178, "grad_norm": 1.7060031762996344, "learning_rate": 2.002835326311603e-07, "loss": 0.6208, "step": 25901 }, { "epoch": 0.9381383556682361, "grad_norm": 1.3176492276022111, "learning_rate": 2.00049998793036e-07, "loss": 0.611, "step": 25902 }, { "epoch": 0.9381745744295545, "grad_norm": 1.3279410894066854, "learning_rate": 1.9981659981083478e-07, "loss": 0.6979, "step": 25903 }, { "epoch": 0.9382107931908729, "grad_norm": 1.336869537916167, "learning_rate": 1.9958333568776745e-07, "loss": 0.605, "step": 25904 }, { "epoch": 0.9382470119521913, "grad_norm": 1.1393457479261284, "learning_rate": 1.993502064270447e-07, "loss": 0.6725, "step": 25905 }, { "epoch": 0.9382832307135096, "grad_norm": 1.3269556837475542, "learning_rate": 1.9911721203187628e-07, "loss": 0.6337, "step": 25906 }, { "epoch": 0.9383194494748279, "grad_norm": 1.128360532474355, "learning_rate": 1.9888435250546623e-07, "loss": 0.6663, "step": 25907 }, { "epoch": 0.9383556682361464, "grad_norm": 1.2190717254731536, "learning_rate": 1.9865162785101866e-07, "loss": 0.6464, "step": 25908 }, { "epoch": 0.9383918869974647, "grad_norm": 1.3776709396774989, "learning_rate": 1.9841903807173769e-07, "loss": 0.6602, "step": 25909 }, { "epoch": 0.9384281057587831, "grad_norm": 1.1119551516040675, "learning_rate": 1.9818658317082296e-07, "loss": 0.6518, "step": 25910 }, { "epoch": 0.9384643245201014, "grad_norm": 1.0448877382774227, "learning_rate": 1.9795426315147414e-07, "loss": 0.6509, "step": 25911 }, { "epoch": 0.9385005432814197, "grad_norm": 0.9920064135398121, "learning_rate": 1.9772207801688647e-07, "loss": 0.6933, "step": 25912 }, { "epoch": 0.9385367620427382, "grad_norm": 1.3137924977215305, "learning_rate": 1.974900277702563e-07, "loss": 0.5667, "step": 25913 }, { "epoch": 0.9385729808040565, "grad_norm": 1.4226242648388343, "learning_rate": 1.9725811241477654e-07, "loss": 0.7798, "step": 25914 }, { "epoch": 0.9386091995653749, "grad_norm": 1.407573373958778, "learning_rate": 1.9702633195363918e-07, "loss": 0.6754, "step": 25915 }, { "epoch": 0.9386454183266932, "grad_norm": 1.3194786142163657, "learning_rate": 1.9679468639003274e-07, "loss": 0.6618, "step": 25916 }, { "epoch": 0.9386816370880116, "grad_norm": 0.9997946945328371, "learning_rate": 1.9656317572714468e-07, "loss": 0.7131, "step": 25917 }, { "epoch": 0.93871785584933, "grad_norm": 1.3569924148598471, "learning_rate": 1.9633179996816242e-07, "loss": 0.5928, "step": 25918 }, { "epoch": 0.9387540746106483, "grad_norm": 1.4796422001038443, "learning_rate": 1.961005591162679e-07, "loss": 0.6246, "step": 25919 }, { "epoch": 0.9387902933719667, "grad_norm": 1.481450080235252, "learning_rate": 1.9586945317464522e-07, "loss": 0.6688, "step": 25920 }, { "epoch": 0.938826512133285, "grad_norm": 1.309805307063072, "learning_rate": 1.9563848214647296e-07, "loss": 0.6451, "step": 25921 }, { "epoch": 0.9388627308946034, "grad_norm": 0.9672306247371446, "learning_rate": 1.9540764603492966e-07, "loss": 0.6104, "step": 25922 }, { "epoch": 0.9388989496559218, "grad_norm": 1.4619251759169762, "learning_rate": 1.9517694484319283e-07, "loss": 0.652, "step": 25923 }, { "epoch": 0.9389351684172401, "grad_norm": 1.351125405310112, "learning_rate": 1.9494637857443543e-07, "loss": 0.6955, "step": 25924 }, { "epoch": 0.9389713871785585, "grad_norm": 1.0327713431804362, "learning_rate": 1.9471594723183273e-07, "loss": 0.6737, "step": 25925 }, { "epoch": 0.9390076059398769, "grad_norm": 1.7892777675735552, "learning_rate": 1.944856508185533e-07, "loss": 0.672, "step": 25926 }, { "epoch": 0.9390438247011952, "grad_norm": 1.3681341061794425, "learning_rate": 1.9425548933776796e-07, "loss": 0.7179, "step": 25927 }, { "epoch": 0.9390800434625136, "grad_norm": 1.0611338023710875, "learning_rate": 1.94025462792643e-07, "loss": 0.7199, "step": 25928 }, { "epoch": 0.9391162622238319, "grad_norm": 1.421236753005694, "learning_rate": 1.9379557118634373e-07, "loss": 0.6491, "step": 25929 }, { "epoch": 0.9391524809851504, "grad_norm": 1.6480199728011244, "learning_rate": 1.9356581452203317e-07, "loss": 0.7292, "step": 25930 }, { "epoch": 0.9391886997464687, "grad_norm": 1.413951911606301, "learning_rate": 1.933361928028754e-07, "loss": 0.6045, "step": 25931 }, { "epoch": 0.939224918507787, "grad_norm": 1.3361926281346748, "learning_rate": 1.9310670603202796e-07, "loss": 0.6467, "step": 25932 }, { "epoch": 0.9392611372691054, "grad_norm": 1.5139762609287903, "learning_rate": 1.928773542126483e-07, "loss": 0.6645, "step": 25933 }, { "epoch": 0.9392973560304237, "grad_norm": 1.323617434345472, "learning_rate": 1.926481373478939e-07, "loss": 0.5861, "step": 25934 }, { "epoch": 0.9393335747917422, "grad_norm": 1.293597310611134, "learning_rate": 1.9241905544091888e-07, "loss": 0.6004, "step": 25935 }, { "epoch": 0.9393697935530605, "grad_norm": 1.4251398917180413, "learning_rate": 1.921901084948763e-07, "loss": 0.6093, "step": 25936 }, { "epoch": 0.9394060123143788, "grad_norm": 0.9542646277274991, "learning_rate": 1.9196129651291362e-07, "loss": 0.6629, "step": 25937 }, { "epoch": 0.9394422310756972, "grad_norm": 1.3302313027425112, "learning_rate": 1.9173261949818278e-07, "loss": 0.5941, "step": 25938 }, { "epoch": 0.9394784498370156, "grad_norm": 1.43386281552208, "learning_rate": 1.9150407745383015e-07, "loss": 0.6605, "step": 25939 }, { "epoch": 0.939514668598334, "grad_norm": 1.4452340865128461, "learning_rate": 1.9127567038299878e-07, "loss": 0.6713, "step": 25940 }, { "epoch": 0.9395508873596523, "grad_norm": 1.0858660102640292, "learning_rate": 1.9104739828883277e-07, "loss": 0.6823, "step": 25941 }, { "epoch": 0.9395871061209706, "grad_norm": 1.3256650638865983, "learning_rate": 1.9081926117447413e-07, "loss": 0.6734, "step": 25942 }, { "epoch": 0.9396233248822891, "grad_norm": 1.672131503116971, "learning_rate": 1.905912590430614e-07, "loss": 0.6953, "step": 25943 }, { "epoch": 0.9396595436436074, "grad_norm": 1.273488010826475, "learning_rate": 1.9036339189773323e-07, "loss": 0.6622, "step": 25944 }, { "epoch": 0.9396957624049258, "grad_norm": 1.0111044232899284, "learning_rate": 1.9013565974162262e-07, "loss": 0.6829, "step": 25945 }, { "epoch": 0.9397319811662441, "grad_norm": 1.129377790183489, "learning_rate": 1.89908062577866e-07, "loss": 0.6606, "step": 25946 }, { "epoch": 0.9397681999275624, "grad_norm": 1.4407578985840843, "learning_rate": 1.8968060040959413e-07, "loss": 0.7224, "step": 25947 }, { "epoch": 0.9398044186888809, "grad_norm": 1.3706127610945218, "learning_rate": 1.8945327323993678e-07, "loss": 0.6584, "step": 25948 }, { "epoch": 0.9398406374501992, "grad_norm": 1.442823332433375, "learning_rate": 1.8922608107202256e-07, "loss": 0.7065, "step": 25949 }, { "epoch": 0.9398768562115176, "grad_norm": 1.3399615333323829, "learning_rate": 1.8899902390897894e-07, "loss": 0.6771, "step": 25950 }, { "epoch": 0.9399130749728359, "grad_norm": 1.5608889845734613, "learning_rate": 1.88772101753929e-07, "loss": 0.7123, "step": 25951 }, { "epoch": 0.9399492937341543, "grad_norm": 1.107884903084502, "learning_rate": 1.885453146099958e-07, "loss": 0.6914, "step": 25952 }, { "epoch": 0.9399855124954727, "grad_norm": 1.472852735540441, "learning_rate": 1.8831866248030018e-07, "loss": 0.6088, "step": 25953 }, { "epoch": 0.940021731256791, "grad_norm": 1.329503512232701, "learning_rate": 1.8809214536796184e-07, "loss": 0.648, "step": 25954 }, { "epoch": 0.9400579500181094, "grad_norm": 1.4367242954057935, "learning_rate": 1.8786576327609606e-07, "loss": 0.7155, "step": 25955 }, { "epoch": 0.9400941687794278, "grad_norm": 1.0092375913982272, "learning_rate": 1.876395162078193e-07, "loss": 0.6657, "step": 25956 }, { "epoch": 0.9401303875407461, "grad_norm": 1.2907676675075193, "learning_rate": 1.8741340416624453e-07, "loss": 0.637, "step": 25957 }, { "epoch": 0.9401666063020645, "grad_norm": 1.312074195529898, "learning_rate": 1.8718742715448378e-07, "loss": 0.6259, "step": 25958 }, { "epoch": 0.9402028250633828, "grad_norm": 1.3777326014404303, "learning_rate": 1.8696158517564566e-07, "loss": 0.7001, "step": 25959 }, { "epoch": 0.9402390438247012, "grad_norm": 1.4858084508291673, "learning_rate": 1.867358782328399e-07, "loss": 0.7426, "step": 25960 }, { "epoch": 0.9402752625860196, "grad_norm": 1.039464962473287, "learning_rate": 1.8651030632917065e-07, "loss": 0.6548, "step": 25961 }, { "epoch": 0.9403114813473379, "grad_norm": 1.4791146567820483, "learning_rate": 1.8628486946774216e-07, "loss": 0.6053, "step": 25962 }, { "epoch": 0.9403477001086563, "grad_norm": 1.045946148677714, "learning_rate": 1.8605956765165633e-07, "loss": 0.7093, "step": 25963 }, { "epoch": 0.9403839188699746, "grad_norm": 1.5849557425225205, "learning_rate": 1.8583440088401516e-07, "loss": 0.7208, "step": 25964 }, { "epoch": 0.9404201376312931, "grad_norm": 1.3095811584558317, "learning_rate": 1.8560936916791618e-07, "loss": 0.6441, "step": 25965 }, { "epoch": 0.9404563563926114, "grad_norm": 1.0530690479409042, "learning_rate": 1.8538447250645575e-07, "loss": 0.6236, "step": 25966 }, { "epoch": 0.9404925751539297, "grad_norm": 1.4108373814501465, "learning_rate": 1.851597109027281e-07, "loss": 0.6309, "step": 25967 }, { "epoch": 0.9405287939152481, "grad_norm": 1.37313961093011, "learning_rate": 1.8493508435982743e-07, "loss": 0.7074, "step": 25968 }, { "epoch": 0.9405650126765664, "grad_norm": 1.206197442166073, "learning_rate": 1.847105928808446e-07, "loss": 0.6449, "step": 25969 }, { "epoch": 0.9406012314378849, "grad_norm": 1.3751221299118523, "learning_rate": 1.844862364688671e-07, "loss": 0.6408, "step": 25970 }, { "epoch": 0.9406374501992032, "grad_norm": 1.3785097319341073, "learning_rate": 1.8426201512698582e-07, "loss": 0.6803, "step": 25971 }, { "epoch": 0.9406736689605215, "grad_norm": 1.3756690328053127, "learning_rate": 1.8403792885828276e-07, "loss": 0.6156, "step": 25972 }, { "epoch": 0.9407098877218399, "grad_norm": 1.3496267494264693, "learning_rate": 1.8381397766584318e-07, "loss": 0.6138, "step": 25973 }, { "epoch": 0.9407461064831583, "grad_norm": 1.4231344660468024, "learning_rate": 1.83590161552748e-07, "loss": 0.6533, "step": 25974 }, { "epoch": 0.9407823252444767, "grad_norm": 1.3106761781887832, "learning_rate": 1.8336648052207807e-07, "loss": 0.6509, "step": 25975 }, { "epoch": 0.940818544005795, "grad_norm": 1.4207752161245029, "learning_rate": 1.831429345769109e-07, "loss": 0.755, "step": 25976 }, { "epoch": 0.9408547627671133, "grad_norm": 1.264083613188718, "learning_rate": 1.8291952372032406e-07, "loss": 0.6385, "step": 25977 }, { "epoch": 0.9408909815284318, "grad_norm": 1.4092707231160038, "learning_rate": 1.8269624795538843e-07, "loss": 0.6277, "step": 25978 }, { "epoch": 0.9409272002897501, "grad_norm": 1.0046363908468334, "learning_rate": 1.8247310728518042e-07, "loss": 0.6576, "step": 25979 }, { "epoch": 0.9409634190510685, "grad_norm": 1.3221003994059135, "learning_rate": 1.822501017127676e-07, "loss": 0.6994, "step": 25980 }, { "epoch": 0.9409996378123868, "grad_norm": 0.9362405326308727, "learning_rate": 1.8202723124122078e-07, "loss": 0.6714, "step": 25981 }, { "epoch": 0.9410358565737051, "grad_norm": 1.3605617701722441, "learning_rate": 1.8180449587360538e-07, "loss": 0.6673, "step": 25982 }, { "epoch": 0.9410720753350236, "grad_norm": 1.4635848898602652, "learning_rate": 1.8158189561298777e-07, "loss": 0.701, "step": 25983 }, { "epoch": 0.9411082940963419, "grad_norm": 1.4085632781627357, "learning_rate": 1.813594304624311e-07, "loss": 0.6537, "step": 25984 }, { "epoch": 0.9411445128576603, "grad_norm": 0.9336040888547028, "learning_rate": 1.811371004249951e-07, "loss": 0.6472, "step": 25985 }, { "epoch": 0.9411807316189786, "grad_norm": 1.3889035086723274, "learning_rate": 1.8091490550374069e-07, "loss": 0.6875, "step": 25986 }, { "epoch": 0.941216950380297, "grad_norm": 1.174338483801161, "learning_rate": 1.806928457017243e-07, "loss": 0.6811, "step": 25987 }, { "epoch": 0.9412531691416154, "grad_norm": 1.3309357268811337, "learning_rate": 1.8047092102200238e-07, "loss": 0.6402, "step": 25988 }, { "epoch": 0.9412893879029337, "grad_norm": 1.0001365630481338, "learning_rate": 1.8024913146762912e-07, "loss": 0.6471, "step": 25989 }, { "epoch": 0.9413256066642521, "grad_norm": 0.9503810216902195, "learning_rate": 1.8002747704165656e-07, "loss": 0.6911, "step": 25990 }, { "epoch": 0.9413618254255705, "grad_norm": 1.445323831281834, "learning_rate": 1.798059577471345e-07, "loss": 0.7175, "step": 25991 }, { "epoch": 0.9413980441868888, "grad_norm": 1.2123927232389036, "learning_rate": 1.7958457358711045e-07, "loss": 0.65, "step": 25992 }, { "epoch": 0.9414342629482072, "grad_norm": 1.4439636470271002, "learning_rate": 1.79363324564632e-07, "loss": 0.6192, "step": 25993 }, { "epoch": 0.9414704817095255, "grad_norm": 1.3626724075026075, "learning_rate": 1.791422106827445e-07, "loss": 0.6577, "step": 25994 }, { "epoch": 0.941506700470844, "grad_norm": 1.0568403168361775, "learning_rate": 1.7892123194448886e-07, "loss": 0.6264, "step": 25995 }, { "epoch": 0.9415429192321623, "grad_norm": 1.6811246168878735, "learning_rate": 1.7870038835290594e-07, "loss": 0.6636, "step": 25996 }, { "epoch": 0.9415791379934806, "grad_norm": 0.9928640080368718, "learning_rate": 1.7847967991103555e-07, "loss": 0.6928, "step": 25997 }, { "epoch": 0.941615356754799, "grad_norm": 1.4460216638207208, "learning_rate": 1.7825910662191525e-07, "loss": 0.6107, "step": 25998 }, { "epoch": 0.9416515755161173, "grad_norm": 1.3810163540888667, "learning_rate": 1.780386684885793e-07, "loss": 0.6969, "step": 25999 }, { "epoch": 0.9416877942774358, "grad_norm": 1.3246741081615112, "learning_rate": 1.7781836551406195e-07, "loss": 0.5955, "step": 26000 }, { "epoch": 0.9417240130387541, "grad_norm": 1.3374629708538721, "learning_rate": 1.7759819770139408e-07, "loss": 0.6452, "step": 26001 }, { "epoch": 0.9417602318000724, "grad_norm": 1.4330936515879773, "learning_rate": 1.773781650536066e-07, "loss": 0.594, "step": 26002 }, { "epoch": 0.9417964505613908, "grad_norm": 1.405487234488114, "learning_rate": 1.771582675737249e-07, "loss": 0.6181, "step": 26003 }, { "epoch": 0.9418326693227091, "grad_norm": 1.3415433867721553, "learning_rate": 1.7693850526477763e-07, "loss": 0.6913, "step": 26004 }, { "epoch": 0.9418688880840276, "grad_norm": 1.4299726227952319, "learning_rate": 1.7671887812978794e-07, "loss": 0.6872, "step": 26005 }, { "epoch": 0.9419051068453459, "grad_norm": 1.0090130185665773, "learning_rate": 1.7649938617177785e-07, "loss": 0.6401, "step": 26006 }, { "epoch": 0.9419413256066642, "grad_norm": 1.0450077530741053, "learning_rate": 1.762800293937672e-07, "loss": 0.701, "step": 26007 }, { "epoch": 0.9419775443679826, "grad_norm": 1.0590293149196521, "learning_rate": 1.7606080779877578e-07, "loss": 0.7524, "step": 26008 }, { "epoch": 0.942013763129301, "grad_norm": 1.7032133777243852, "learning_rate": 1.7584172138982003e-07, "loss": 0.7085, "step": 26009 }, { "epoch": 0.9420499818906194, "grad_norm": 1.6946530399902344, "learning_rate": 1.7562277016991314e-07, "loss": 0.6984, "step": 26010 }, { "epoch": 0.9420862006519377, "grad_norm": 1.3953375041714065, "learning_rate": 1.7540395414207045e-07, "loss": 0.6629, "step": 26011 }, { "epoch": 0.942122419413256, "grad_norm": 1.6487692288993907, "learning_rate": 1.751852733093029e-07, "loss": 0.6423, "step": 26012 }, { "epoch": 0.9421586381745745, "grad_norm": 1.5690664856470145, "learning_rate": 1.749667276746181e-07, "loss": 0.6856, "step": 26013 }, { "epoch": 0.9421948569358928, "grad_norm": 1.4830005368250565, "learning_rate": 1.747483172410225e-07, "loss": 0.7101, "step": 26014 }, { "epoch": 0.9422310756972112, "grad_norm": 1.6944119768916732, "learning_rate": 1.745300420115259e-07, "loss": 0.6706, "step": 26015 }, { "epoch": 0.9422672944585295, "grad_norm": 1.3549207596037978, "learning_rate": 1.7431190198912817e-07, "loss": 0.6926, "step": 26016 }, { "epoch": 0.9423035132198478, "grad_norm": 1.3280197978427724, "learning_rate": 1.7409389717683244e-07, "loss": 0.6467, "step": 26017 }, { "epoch": 0.9423397319811663, "grad_norm": 1.0661728398495216, "learning_rate": 1.7387602757763743e-07, "loss": 0.6438, "step": 26018 }, { "epoch": 0.9423759507424846, "grad_norm": 1.2556942464856018, "learning_rate": 1.7365829319454296e-07, "loss": 0.6555, "step": 26019 }, { "epoch": 0.942412169503803, "grad_norm": 1.425124644278348, "learning_rate": 1.7344069403054554e-07, "loss": 0.6065, "step": 26020 }, { "epoch": 0.9424483882651213, "grad_norm": 1.378556338833475, "learning_rate": 1.7322323008863607e-07, "loss": 0.7258, "step": 26021 }, { "epoch": 0.9424846070264397, "grad_norm": 1.3909671367923784, "learning_rate": 1.7300590137181217e-07, "loss": 0.6633, "step": 26022 }, { "epoch": 0.9425208257877581, "grad_norm": 1.366139047379558, "learning_rate": 1.7278870788306035e-07, "loss": 0.6232, "step": 26023 }, { "epoch": 0.9425570445490764, "grad_norm": 1.5539745788069832, "learning_rate": 1.7257164962537153e-07, "loss": 0.652, "step": 26024 }, { "epoch": 0.9425932633103948, "grad_norm": 1.3513799506868094, "learning_rate": 1.7235472660173114e-07, "loss": 0.6576, "step": 26025 }, { "epoch": 0.9426294820717132, "grad_norm": 1.4087615684873718, "learning_rate": 1.7213793881512563e-07, "loss": 0.6244, "step": 26026 }, { "epoch": 0.9426657008330315, "grad_norm": 1.6057784740063579, "learning_rate": 1.7192128626853822e-07, "loss": 0.6937, "step": 26027 }, { "epoch": 0.9427019195943499, "grad_norm": 1.3548521211339355, "learning_rate": 1.717047689649487e-07, "loss": 0.6306, "step": 26028 }, { "epoch": 0.9427381383556682, "grad_norm": 1.3972515579797407, "learning_rate": 1.7148838690733804e-07, "loss": 0.6881, "step": 26029 }, { "epoch": 0.9427743571169866, "grad_norm": 1.2860767182823756, "learning_rate": 1.7127214009868387e-07, "loss": 0.6024, "step": 26030 }, { "epoch": 0.942810575878305, "grad_norm": 1.4340195543176808, "learning_rate": 1.7105602854195936e-07, "loss": 0.6834, "step": 26031 }, { "epoch": 0.9428467946396233, "grad_norm": 1.4622548469965153, "learning_rate": 1.7084005224014212e-07, "loss": 0.7296, "step": 26032 }, { "epoch": 0.9428830134009417, "grad_norm": 1.6396024003418774, "learning_rate": 1.7062421119620198e-07, "loss": 0.7145, "step": 26033 }, { "epoch": 0.94291923216226, "grad_norm": 1.4203875458911992, "learning_rate": 1.7040850541310883e-07, "loss": 0.6894, "step": 26034 }, { "epoch": 0.9429554509235785, "grad_norm": 1.2759873701403088, "learning_rate": 1.7019293489383138e-07, "loss": 0.643, "step": 26035 }, { "epoch": 0.9429916696848968, "grad_norm": 1.4961876628075226, "learning_rate": 1.6997749964133724e-07, "loss": 0.6425, "step": 26036 }, { "epoch": 0.9430278884462151, "grad_norm": 1.4243454300856233, "learning_rate": 1.6976219965859076e-07, "loss": 0.6679, "step": 26037 }, { "epoch": 0.9430641072075335, "grad_norm": 0.9983293330816934, "learning_rate": 1.6954703494855286e-07, "loss": 0.6124, "step": 26038 }, { "epoch": 0.9431003259688518, "grad_norm": 1.4331838557106475, "learning_rate": 1.693320055141856e-07, "loss": 0.6861, "step": 26039 }, { "epoch": 0.9431365447301703, "grad_norm": 1.2584059097373699, "learning_rate": 1.6911711135844888e-07, "loss": 0.6219, "step": 26040 }, { "epoch": 0.9431727634914886, "grad_norm": 1.3027931790613123, "learning_rate": 1.689023524842981e-07, "loss": 0.6614, "step": 26041 }, { "epoch": 0.9432089822528069, "grad_norm": 1.1922246647388597, "learning_rate": 1.6868772889468866e-07, "loss": 0.6905, "step": 26042 }, { "epoch": 0.9432452010141253, "grad_norm": 1.5769141480828712, "learning_rate": 1.6847324059257596e-07, "loss": 0.665, "step": 26043 }, { "epoch": 0.9432814197754437, "grad_norm": 1.3414937750067295, "learning_rate": 1.6825888758090992e-07, "loss": 0.7057, "step": 26044 }, { "epoch": 0.9433176385367621, "grad_norm": 1.1714017015574671, "learning_rate": 1.6804466986264033e-07, "loss": 0.6274, "step": 26045 }, { "epoch": 0.9433538572980804, "grad_norm": 0.9578404844896823, "learning_rate": 1.678305874407149e-07, "loss": 0.7072, "step": 26046 }, { "epoch": 0.9433900760593987, "grad_norm": 1.4297629998288621, "learning_rate": 1.6761664031808012e-07, "loss": 0.6009, "step": 26047 }, { "epoch": 0.9434262948207172, "grad_norm": 1.3209269197205868, "learning_rate": 1.6740282849768031e-07, "loss": 0.7491, "step": 26048 }, { "epoch": 0.9434625135820355, "grad_norm": 1.385195597435621, "learning_rate": 1.6718915198245644e-07, "loss": 0.7182, "step": 26049 }, { "epoch": 0.9434987323433539, "grad_norm": 1.2807539613512824, "learning_rate": 1.6697561077534952e-07, "loss": 0.6331, "step": 26050 }, { "epoch": 0.9435349511046722, "grad_norm": 1.354229604575895, "learning_rate": 1.667622048792994e-07, "loss": 0.6599, "step": 26051 }, { "epoch": 0.9435711698659905, "grad_norm": 1.3699207817684584, "learning_rate": 1.665489342972415e-07, "loss": 0.6406, "step": 26052 }, { "epoch": 0.943607388627309, "grad_norm": 0.9258091550846059, "learning_rate": 1.6633579903210906e-07, "loss": 0.6411, "step": 26053 }, { "epoch": 0.9436436073886273, "grad_norm": 1.4245550055833274, "learning_rate": 1.6612279908683858e-07, "loss": 0.6286, "step": 26054 }, { "epoch": 0.9436798261499457, "grad_norm": 1.2198686547508153, "learning_rate": 1.6590993446435776e-07, "loss": 0.619, "step": 26055 }, { "epoch": 0.943716044911264, "grad_norm": 1.2880235893518668, "learning_rate": 1.6569720516759868e-07, "loss": 0.6812, "step": 26056 }, { "epoch": 0.9437522636725824, "grad_norm": 1.3825448263389477, "learning_rate": 1.6548461119948568e-07, "loss": 0.6507, "step": 26057 }, { "epoch": 0.9437884824339008, "grad_norm": 1.4711864457425043, "learning_rate": 1.6527215256294638e-07, "loss": 0.7345, "step": 26058 }, { "epoch": 0.9438247011952191, "grad_norm": 1.4262509558648977, "learning_rate": 1.6505982926090403e-07, "loss": 0.6841, "step": 26059 }, { "epoch": 0.9438609199565375, "grad_norm": 1.4538467893684104, "learning_rate": 1.6484764129628073e-07, "loss": 0.6081, "step": 26060 }, { "epoch": 0.9438971387178559, "grad_norm": 1.4037244927465842, "learning_rate": 1.6463558867199415e-07, "loss": 0.6666, "step": 26061 }, { "epoch": 0.9439333574791742, "grad_norm": 1.3514979421489324, "learning_rate": 1.6442367139096528e-07, "loss": 0.6311, "step": 26062 }, { "epoch": 0.9439695762404926, "grad_norm": 1.0509550425504939, "learning_rate": 1.6421188945610843e-07, "loss": 0.6902, "step": 26063 }, { "epoch": 0.9440057950018109, "grad_norm": 0.9515004665347447, "learning_rate": 1.64000242870338e-07, "loss": 0.69, "step": 26064 }, { "epoch": 0.9440420137631294, "grad_norm": 1.1632453868615664, "learning_rate": 1.6378873163656828e-07, "loss": 0.6674, "step": 26065 }, { "epoch": 0.9440782325244477, "grad_norm": 1.3703840081577259, "learning_rate": 1.6357735575770806e-07, "loss": 0.6248, "step": 26066 }, { "epoch": 0.944114451285766, "grad_norm": 1.1975532484957927, "learning_rate": 1.6336611523666612e-07, "loss": 0.6506, "step": 26067 }, { "epoch": 0.9441506700470844, "grad_norm": 1.3486658985849869, "learning_rate": 1.6315501007634905e-07, "loss": 0.6206, "step": 26068 }, { "epoch": 0.9441868888084027, "grad_norm": 1.5342049375043023, "learning_rate": 1.6294404027966338e-07, "loss": 0.6917, "step": 26069 }, { "epoch": 0.9442231075697212, "grad_norm": 1.5545577613332904, "learning_rate": 1.6273320584951124e-07, "loss": 0.7025, "step": 26070 }, { "epoch": 0.9442593263310395, "grad_norm": 1.3972530910184655, "learning_rate": 1.6252250678879368e-07, "loss": 0.6816, "step": 26071 }, { "epoch": 0.9442955450923578, "grad_norm": 1.3680589405438563, "learning_rate": 1.6231194310041055e-07, "loss": 0.6601, "step": 26072 }, { "epoch": 0.9443317638536762, "grad_norm": 1.3627203518283966, "learning_rate": 1.6210151478725954e-07, "loss": 0.6881, "step": 26073 }, { "epoch": 0.9443679826149945, "grad_norm": 1.4013266886531335, "learning_rate": 1.6189122185223504e-07, "loss": 0.6649, "step": 26074 }, { "epoch": 0.944404201376313, "grad_norm": 1.345317860170645, "learning_rate": 1.616810642982325e-07, "loss": 0.704, "step": 26075 }, { "epoch": 0.9444404201376313, "grad_norm": 1.4510487135816212, "learning_rate": 1.614710421281429e-07, "loss": 0.7417, "step": 26076 }, { "epoch": 0.9444766388989496, "grad_norm": 4.802556187014568, "learning_rate": 1.6126115534485733e-07, "loss": 0.6357, "step": 26077 }, { "epoch": 0.944512857660268, "grad_norm": 1.3037181925482457, "learning_rate": 1.610514039512634e-07, "loss": 0.6592, "step": 26078 }, { "epoch": 0.9445490764215864, "grad_norm": 2.0244991047244354, "learning_rate": 1.6084178795024664e-07, "loss": 0.686, "step": 26079 }, { "epoch": 0.9445852951829048, "grad_norm": 1.373174017775502, "learning_rate": 1.606323073446925e-07, "loss": 0.7158, "step": 26080 }, { "epoch": 0.9446215139442231, "grad_norm": 1.372155161097322, "learning_rate": 1.6042296213748308e-07, "loss": 0.6926, "step": 26081 }, { "epoch": 0.9446577327055414, "grad_norm": 1.2408978279636558, "learning_rate": 1.602137523315006e-07, "loss": 0.6059, "step": 26082 }, { "epoch": 0.9446939514668599, "grad_norm": 1.1384940719133352, "learning_rate": 1.600046779296216e-07, "loss": 0.6726, "step": 26083 }, { "epoch": 0.9447301702281782, "grad_norm": 1.3575679576949884, "learning_rate": 1.5979573893472488e-07, "loss": 0.5719, "step": 26084 }, { "epoch": 0.9447663889894966, "grad_norm": 1.2080870358949045, "learning_rate": 1.5958693534968595e-07, "loss": 0.7111, "step": 26085 }, { "epoch": 0.9448026077508149, "grad_norm": 1.3471931663941312, "learning_rate": 1.5937826717737582e-07, "loss": 0.6604, "step": 26086 }, { "epoch": 0.9448388265121332, "grad_norm": 1.3788373297739058, "learning_rate": 1.5916973442066886e-07, "loss": 0.6766, "step": 26087 }, { "epoch": 0.9448750452734517, "grad_norm": 1.4113984837957645, "learning_rate": 1.5896133708243276e-07, "loss": 0.6889, "step": 26088 }, { "epoch": 0.94491126403477, "grad_norm": 1.3274554233713203, "learning_rate": 1.587530751655364e-07, "loss": 0.6837, "step": 26089 }, { "epoch": 0.9449474827960884, "grad_norm": 1.5900495862220123, "learning_rate": 1.5854494867284408e-07, "loss": 0.7411, "step": 26090 }, { "epoch": 0.9449837015574067, "grad_norm": 1.4333691466515028, "learning_rate": 1.5833695760722135e-07, "loss": 0.6755, "step": 26091 }, { "epoch": 0.9450199203187251, "grad_norm": 1.4668816695612281, "learning_rate": 1.581291019715292e-07, "loss": 0.6376, "step": 26092 }, { "epoch": 0.9450561390800435, "grad_norm": 1.3703707163394163, "learning_rate": 1.579213817686276e-07, "loss": 0.694, "step": 26093 }, { "epoch": 0.9450923578413618, "grad_norm": 1.3466768755755483, "learning_rate": 1.5771379700137758e-07, "loss": 0.6941, "step": 26094 }, { "epoch": 0.9451285766026802, "grad_norm": 1.4096057436785865, "learning_rate": 1.5750634767263352e-07, "loss": 0.7279, "step": 26095 }, { "epoch": 0.9451647953639986, "grad_norm": 1.4512838228052216, "learning_rate": 1.572990337852498e-07, "loss": 0.6541, "step": 26096 }, { "epoch": 0.9452010141253169, "grad_norm": 1.291743475142215, "learning_rate": 1.570918553420797e-07, "loss": 0.6887, "step": 26097 }, { "epoch": 0.9452372328866353, "grad_norm": 1.4175745025834456, "learning_rate": 1.5688481234597542e-07, "loss": 0.7094, "step": 26098 }, { "epoch": 0.9452734516479536, "grad_norm": 1.5050713088739869, "learning_rate": 1.5667790479978462e-07, "loss": 0.7289, "step": 26099 }, { "epoch": 0.945309670409272, "grad_norm": 1.442888714832816, "learning_rate": 1.5647113270635504e-07, "loss": 0.6329, "step": 26100 }, { "epoch": 0.9453458891705904, "grad_norm": 1.3836110134437072, "learning_rate": 1.562644960685311e-07, "loss": 0.6706, "step": 26101 }, { "epoch": 0.9453821079319087, "grad_norm": 0.9911260338885677, "learning_rate": 1.560579948891583e-07, "loss": 0.6323, "step": 26102 }, { "epoch": 0.9454183266932271, "grad_norm": 1.415295041337316, "learning_rate": 1.5585162917107656e-07, "loss": 0.6685, "step": 26103 }, { "epoch": 0.9454545454545454, "grad_norm": 1.3636879915363034, "learning_rate": 1.5564539891712583e-07, "loss": 0.6586, "step": 26104 }, { "epoch": 0.9454907642158639, "grad_norm": 1.2592985573705373, "learning_rate": 1.5543930413014496e-07, "loss": 0.5974, "step": 26105 }, { "epoch": 0.9455269829771822, "grad_norm": 1.3281525987674698, "learning_rate": 1.5523334481297058e-07, "loss": 0.6396, "step": 26106 }, { "epoch": 0.9455632017385005, "grad_norm": 1.6111835368626464, "learning_rate": 1.5502752096843488e-07, "loss": 0.6515, "step": 26107 }, { "epoch": 0.9455994204998189, "grad_norm": 1.4392981995812173, "learning_rate": 1.5482183259937e-07, "loss": 0.6982, "step": 26108 }, { "epoch": 0.9456356392611373, "grad_norm": 1.2706002437093296, "learning_rate": 1.5461627970860814e-07, "loss": 0.625, "step": 26109 }, { "epoch": 0.9456718580224557, "grad_norm": 1.3815590577696375, "learning_rate": 1.5441086229897706e-07, "loss": 0.594, "step": 26110 }, { "epoch": 0.945708076783774, "grad_norm": 1.378416172206139, "learning_rate": 1.5420558037330448e-07, "loss": 0.6954, "step": 26111 }, { "epoch": 0.9457442955450923, "grad_norm": 0.9630447393364705, "learning_rate": 1.5400043393441256e-07, "loss": 0.6706, "step": 26112 }, { "epoch": 0.9457805143064107, "grad_norm": 1.2401345845967937, "learning_rate": 1.5379542298512796e-07, "loss": 0.6416, "step": 26113 }, { "epoch": 0.9458167330677291, "grad_norm": 1.4498121275538092, "learning_rate": 1.5359054752826953e-07, "loss": 0.6586, "step": 26114 }, { "epoch": 0.9458529518290475, "grad_norm": 0.9625517271009149, "learning_rate": 1.533858075666561e-07, "loss": 0.6259, "step": 26115 }, { "epoch": 0.9458891705903658, "grad_norm": 1.0057300031607381, "learning_rate": 1.531812031031066e-07, "loss": 0.6958, "step": 26116 }, { "epoch": 0.9459253893516841, "grad_norm": 1.4120231914733539, "learning_rate": 1.5297673414043644e-07, "loss": 0.6468, "step": 26117 }, { "epoch": 0.9459616081130026, "grad_norm": 1.3766580231622911, "learning_rate": 1.5277240068145795e-07, "loss": 0.6694, "step": 26118 }, { "epoch": 0.9459978268743209, "grad_norm": 1.2247347063648324, "learning_rate": 1.525682027289843e-07, "loss": 0.7275, "step": 26119 }, { "epoch": 0.9460340456356393, "grad_norm": 1.4553189916754676, "learning_rate": 1.5236414028582447e-07, "loss": 0.6568, "step": 26120 }, { "epoch": 0.9460702643969576, "grad_norm": 1.2873610646955964, "learning_rate": 1.5216021335478725e-07, "loss": 0.6875, "step": 26121 }, { "epoch": 0.946106483158276, "grad_norm": 1.5955673729665498, "learning_rate": 1.519564219386782e-07, "loss": 0.7067, "step": 26122 }, { "epoch": 0.9461427019195944, "grad_norm": 1.5985945170947573, "learning_rate": 1.5175276604030177e-07, "loss": 0.7521, "step": 26123 }, { "epoch": 0.9461789206809127, "grad_norm": 1.3069102389744354, "learning_rate": 1.5154924566246232e-07, "loss": 0.6427, "step": 26124 }, { "epoch": 0.9462151394422311, "grad_norm": 1.590759271489171, "learning_rate": 1.5134586080795766e-07, "loss": 0.7344, "step": 26125 }, { "epoch": 0.9462513582035494, "grad_norm": 1.2930197359963598, "learning_rate": 1.5114261147958665e-07, "loss": 0.6328, "step": 26126 }, { "epoch": 0.9462875769648678, "grad_norm": 1.357945579936419, "learning_rate": 1.5093949768014925e-07, "loss": 0.6823, "step": 26127 }, { "epoch": 0.9463237957261862, "grad_norm": 1.3327966767385868, "learning_rate": 1.5073651941243772e-07, "loss": 0.6818, "step": 26128 }, { "epoch": 0.9463600144875045, "grad_norm": 1.322109566076353, "learning_rate": 1.5053367667924645e-07, "loss": 0.6437, "step": 26129 }, { "epoch": 0.9463962332488229, "grad_norm": 1.2896971671849229, "learning_rate": 1.5033096948336546e-07, "loss": 0.6407, "step": 26130 }, { "epoch": 0.9464324520101413, "grad_norm": 1.3065681708280559, "learning_rate": 1.5012839782758472e-07, "loss": 0.6648, "step": 26131 }, { "epoch": 0.9464686707714596, "grad_norm": 0.9357986054285603, "learning_rate": 1.4992596171469308e-07, "loss": 0.6664, "step": 26132 }, { "epoch": 0.946504889532778, "grad_norm": 1.5028964027204095, "learning_rate": 1.4972366114747504e-07, "loss": 0.7365, "step": 26133 }, { "epoch": 0.9465411082940963, "grad_norm": 2.427263287836542, "learning_rate": 1.4952149612871391e-07, "loss": 0.7061, "step": 26134 }, { "epoch": 0.9465773270554148, "grad_norm": 2.453366220665917, "learning_rate": 1.49319466661193e-07, "loss": 0.7356, "step": 26135 }, { "epoch": 0.9466135458167331, "grad_norm": 1.3085069777059324, "learning_rate": 1.4911757274769235e-07, "loss": 0.6642, "step": 26136 }, { "epoch": 0.9466497645780514, "grad_norm": 1.17498885855107, "learning_rate": 1.4891581439098858e-07, "loss": 0.655, "step": 26137 }, { "epoch": 0.9466859833393698, "grad_norm": 1.357259379067155, "learning_rate": 1.4871419159385948e-07, "loss": 0.6178, "step": 26138 }, { "epoch": 0.9467222021006881, "grad_norm": 1.896435159647191, "learning_rate": 1.485127043590795e-07, "loss": 0.6873, "step": 26139 }, { "epoch": 0.9467584208620066, "grad_norm": 1.5512824863977126, "learning_rate": 1.4831135268942088e-07, "loss": 0.693, "step": 26140 }, { "epoch": 0.9467946396233249, "grad_norm": 1.3246603387013016, "learning_rate": 1.4811013658765471e-07, "loss": 0.7003, "step": 26141 }, { "epoch": 0.9468308583846432, "grad_norm": 1.3746504610278782, "learning_rate": 1.4790905605654882e-07, "loss": 0.6844, "step": 26142 }, { "epoch": 0.9468670771459616, "grad_norm": 1.3646822933235714, "learning_rate": 1.4770811109887206e-07, "loss": 0.604, "step": 26143 }, { "epoch": 0.94690329590728, "grad_norm": 1.3227887640127305, "learning_rate": 1.4750730171738892e-07, "loss": 0.6829, "step": 26144 }, { "epoch": 0.9469395146685984, "grad_norm": 1.3702283230963335, "learning_rate": 1.473066279148616e-07, "loss": 0.597, "step": 26145 }, { "epoch": 0.9469757334299167, "grad_norm": 1.5718845640160763, "learning_rate": 1.471060896940524e-07, "loss": 0.5985, "step": 26146 }, { "epoch": 0.947011952191235, "grad_norm": 1.374483328436299, "learning_rate": 1.4690568705772123e-07, "loss": 0.6812, "step": 26147 }, { "epoch": 0.9470481709525534, "grad_norm": 1.3991561593434052, "learning_rate": 1.4670542000862486e-07, "loss": 0.6996, "step": 26148 }, { "epoch": 0.9470843897138718, "grad_norm": 1.282831245545599, "learning_rate": 1.4650528854952105e-07, "loss": 0.6687, "step": 26149 }, { "epoch": 0.9471206084751902, "grad_norm": 0.9201022268605189, "learning_rate": 1.4630529268316207e-07, "loss": 0.6654, "step": 26150 }, { "epoch": 0.9471568272365085, "grad_norm": 1.4333115450652494, "learning_rate": 1.4610543241230012e-07, "loss": 0.6446, "step": 26151 }, { "epoch": 0.9471930459978268, "grad_norm": 1.2878744987204778, "learning_rate": 1.4590570773968526e-07, "loss": 0.6566, "step": 26152 }, { "epoch": 0.9472292647591453, "grad_norm": 1.0362401097988527, "learning_rate": 1.457061186680675e-07, "loss": 0.6889, "step": 26153 }, { "epoch": 0.9472654835204636, "grad_norm": 1.3897400549030374, "learning_rate": 1.455066652001924e-07, "loss": 0.6566, "step": 26154 }, { "epoch": 0.947301702281782, "grad_norm": 1.3807587320932992, "learning_rate": 1.4530734733880337e-07, "loss": 0.6815, "step": 26155 }, { "epoch": 0.9473379210431003, "grad_norm": 1.2409092480593296, "learning_rate": 1.4510816508664482e-07, "loss": 0.6767, "step": 26156 }, { "epoch": 0.9473741398044186, "grad_norm": 1.491644464664775, "learning_rate": 1.4490911844645793e-07, "loss": 0.7445, "step": 26157 }, { "epoch": 0.9474103585657371, "grad_norm": 1.2008486732089718, "learning_rate": 1.4471020742098052e-07, "loss": 0.6821, "step": 26158 }, { "epoch": 0.9474465773270554, "grad_norm": 1.0572830479308086, "learning_rate": 1.4451143201294925e-07, "loss": 0.6632, "step": 26159 }, { "epoch": 0.9474827960883738, "grad_norm": 1.4577998813418291, "learning_rate": 1.4431279222510196e-07, "loss": 0.644, "step": 26160 }, { "epoch": 0.9475190148496921, "grad_norm": 1.44935299593453, "learning_rate": 1.441142880601698e-07, "loss": 0.715, "step": 26161 }, { "epoch": 0.9475552336110105, "grad_norm": 1.4518679339581368, "learning_rate": 1.4391591952088503e-07, "loss": 0.6629, "step": 26162 }, { "epoch": 0.9475914523723289, "grad_norm": 1.4113421274985072, "learning_rate": 1.4371768660997764e-07, "loss": 0.6175, "step": 26163 }, { "epoch": 0.9476276711336472, "grad_norm": 1.4212453975682653, "learning_rate": 1.435195893301755e-07, "loss": 0.7133, "step": 26164 }, { "epoch": 0.9476638898949656, "grad_norm": 1.4910295972002034, "learning_rate": 1.4332162768420533e-07, "loss": 0.6423, "step": 26165 }, { "epoch": 0.947700108656284, "grad_norm": 1.4402477485712475, "learning_rate": 1.4312380167478935e-07, "loss": 0.6451, "step": 26166 }, { "epoch": 0.9477363274176023, "grad_norm": 1.3019238349659716, "learning_rate": 1.4292611130465096e-07, "loss": 0.6912, "step": 26167 }, { "epoch": 0.9477725461789207, "grad_norm": 1.5087577193131063, "learning_rate": 1.427285565765113e-07, "loss": 0.732, "step": 26168 }, { "epoch": 0.947808764940239, "grad_norm": 1.3097381135819772, "learning_rate": 1.4253113749308712e-07, "loss": 0.6818, "step": 26169 }, { "epoch": 0.9478449837015575, "grad_norm": 1.1583568288727022, "learning_rate": 1.423338540570962e-07, "loss": 0.6715, "step": 26170 }, { "epoch": 0.9478812024628758, "grad_norm": 1.4213531572736253, "learning_rate": 1.4213670627125308e-07, "loss": 0.6756, "step": 26171 }, { "epoch": 0.9479174212241941, "grad_norm": 1.2714097415683447, "learning_rate": 1.4193969413827113e-07, "loss": 0.6439, "step": 26172 }, { "epoch": 0.9479536399855125, "grad_norm": 1.3097260152889638, "learning_rate": 1.417428176608615e-07, "loss": 0.6327, "step": 26173 }, { "epoch": 0.9479898587468308, "grad_norm": 1.010264990434873, "learning_rate": 1.4154607684173094e-07, "loss": 0.6682, "step": 26174 }, { "epoch": 0.9480260775081493, "grad_norm": 1.1178470872344284, "learning_rate": 1.413494716835906e-07, "loss": 0.6847, "step": 26175 }, { "epoch": 0.9480622962694676, "grad_norm": 1.3744967837564044, "learning_rate": 1.4115300218914386e-07, "loss": 0.7774, "step": 26176 }, { "epoch": 0.9480985150307859, "grad_norm": 1.5681201709644073, "learning_rate": 1.409566683610941e-07, "loss": 0.7419, "step": 26177 }, { "epoch": 0.9481347337921043, "grad_norm": 1.1018734109131607, "learning_rate": 1.4076047020214368e-07, "loss": 0.7501, "step": 26178 }, { "epoch": 0.9481709525534227, "grad_norm": 1.3622599454252613, "learning_rate": 1.405644077149926e-07, "loss": 0.7266, "step": 26179 }, { "epoch": 0.9482071713147411, "grad_norm": 1.2424760834144966, "learning_rate": 1.403684809023387e-07, "loss": 0.6559, "step": 26180 }, { "epoch": 0.9482433900760594, "grad_norm": 1.3846690463026867, "learning_rate": 1.4017268976687648e-07, "loss": 0.7266, "step": 26181 }, { "epoch": 0.9482796088373777, "grad_norm": 1.3263846340500935, "learning_rate": 1.3997703431130271e-07, "loss": 0.6497, "step": 26182 }, { "epoch": 0.9483158275986961, "grad_norm": 1.2130386167038643, "learning_rate": 1.3978151453830856e-07, "loss": 0.658, "step": 26183 }, { "epoch": 0.9483520463600145, "grad_norm": 0.9422874466667286, "learning_rate": 1.3958613045058522e-07, "loss": 0.6441, "step": 26184 }, { "epoch": 0.9483882651213329, "grad_norm": 1.3336862381729335, "learning_rate": 1.3939088205082052e-07, "loss": 0.6216, "step": 26185 }, { "epoch": 0.9484244838826512, "grad_norm": 1.4206384965064174, "learning_rate": 1.391957693417012e-07, "loss": 0.6658, "step": 26186 }, { "epoch": 0.9484607026439695, "grad_norm": 1.5845612887616098, "learning_rate": 1.390007923259129e-07, "loss": 0.6494, "step": 26187 }, { "epoch": 0.948496921405288, "grad_norm": 1.4040723602616034, "learning_rate": 1.388059510061379e-07, "loss": 0.6696, "step": 26188 }, { "epoch": 0.9485331401666063, "grad_norm": 0.9091857889208336, "learning_rate": 1.3861124538505854e-07, "loss": 0.6423, "step": 26189 }, { "epoch": 0.9485693589279247, "grad_norm": 1.392362662870682, "learning_rate": 1.3841667546535265e-07, "loss": 0.6835, "step": 26190 }, { "epoch": 0.948605577689243, "grad_norm": 1.384374705932448, "learning_rate": 1.3822224124969918e-07, "loss": 0.6661, "step": 26191 }, { "epoch": 0.9486417964505613, "grad_norm": 1.3927379097831893, "learning_rate": 1.380279427407727e-07, "loss": 0.6908, "step": 26192 }, { "epoch": 0.9486780152118798, "grad_norm": 1.3588590745927436, "learning_rate": 1.378337799412477e-07, "loss": 0.6307, "step": 26193 }, { "epoch": 0.9487142339731981, "grad_norm": 1.5236543177101807, "learning_rate": 1.3763975285379426e-07, "loss": 0.7276, "step": 26194 }, { "epoch": 0.9487504527345165, "grad_norm": 1.4058864308960335, "learning_rate": 1.3744586148108475e-07, "loss": 0.6403, "step": 26195 }, { "epoch": 0.9487866714958348, "grad_norm": 1.476619317586367, "learning_rate": 1.3725210582578584e-07, "loss": 0.6074, "step": 26196 }, { "epoch": 0.9488228902571532, "grad_norm": 1.5546764536326194, "learning_rate": 1.3705848589056437e-07, "loss": 0.6265, "step": 26197 }, { "epoch": 0.9488591090184716, "grad_norm": 1.4589542962871418, "learning_rate": 1.368650016780837e-07, "loss": 0.6366, "step": 26198 }, { "epoch": 0.9488953277797899, "grad_norm": 1.4447418897196596, "learning_rate": 1.3667165319100729e-07, "loss": 0.6957, "step": 26199 }, { "epoch": 0.9489315465411083, "grad_norm": 1.5550940128311093, "learning_rate": 1.3647844043199633e-07, "loss": 0.6788, "step": 26200 }, { "epoch": 0.9489677653024267, "grad_norm": 1.404885936485921, "learning_rate": 1.3628536340370869e-07, "loss": 0.5683, "step": 26201 }, { "epoch": 0.949003984063745, "grad_norm": 1.433961009191375, "learning_rate": 1.3609242210880114e-07, "loss": 0.6628, "step": 26202 }, { "epoch": 0.9490402028250634, "grad_norm": 1.3474458841250747, "learning_rate": 1.3589961654992934e-07, "loss": 0.7004, "step": 26203 }, { "epoch": 0.9490764215863817, "grad_norm": 1.3646431919945436, "learning_rate": 1.357069467297456e-07, "loss": 0.6591, "step": 26204 }, { "epoch": 0.9491126403477002, "grad_norm": 1.5809686223827593, "learning_rate": 1.3551441265090336e-07, "loss": 0.7293, "step": 26205 }, { "epoch": 0.9491488591090185, "grad_norm": 1.373547348174478, "learning_rate": 1.353220143160483e-07, "loss": 0.7569, "step": 26206 }, { "epoch": 0.9491850778703368, "grad_norm": 1.514508014591434, "learning_rate": 1.351297517278316e-07, "loss": 0.6433, "step": 26207 }, { "epoch": 0.9492212966316552, "grad_norm": 1.4585747418240314, "learning_rate": 1.3493762488889784e-07, "loss": 0.7403, "step": 26208 }, { "epoch": 0.9492575153929735, "grad_norm": 0.9565950410416073, "learning_rate": 1.3474563380188932e-07, "loss": 0.6153, "step": 26209 }, { "epoch": 0.949293734154292, "grad_norm": 1.3845790204648625, "learning_rate": 1.345537784694495e-07, "loss": 0.7178, "step": 26210 }, { "epoch": 0.9493299529156103, "grad_norm": 1.316936961824605, "learning_rate": 1.343620588942196e-07, "loss": 0.6312, "step": 26211 }, { "epoch": 0.9493661716769286, "grad_norm": 1.4335598103562097, "learning_rate": 1.3417047507883528e-07, "loss": 0.659, "step": 26212 }, { "epoch": 0.949402390438247, "grad_norm": 0.9622371775354986, "learning_rate": 1.3397902702593446e-07, "loss": 0.737, "step": 26213 }, { "epoch": 0.9494386091995654, "grad_norm": 1.412232800520231, "learning_rate": 1.3378771473815054e-07, "loss": 0.6016, "step": 26214 }, { "epoch": 0.9494748279608838, "grad_norm": 1.2990100259967532, "learning_rate": 1.3359653821811812e-07, "loss": 0.6496, "step": 26215 }, { "epoch": 0.9495110467222021, "grad_norm": 1.0253659448889212, "learning_rate": 1.334054974684662e-07, "loss": 0.665, "step": 26216 }, { "epoch": 0.9495472654835204, "grad_norm": 1.3480107200613531, "learning_rate": 1.3321459249182266e-07, "loss": 0.6422, "step": 26217 }, { "epoch": 0.9495834842448388, "grad_norm": 1.3401988771096494, "learning_rate": 1.3302382329081765e-07, "loss": 0.6433, "step": 26218 }, { "epoch": 0.9496197030061572, "grad_norm": 1.0573207703687688, "learning_rate": 1.3283318986807458e-07, "loss": 0.6791, "step": 26219 }, { "epoch": 0.9496559217674756, "grad_norm": 1.078664035690927, "learning_rate": 1.3264269222621696e-07, "loss": 0.6071, "step": 26220 }, { "epoch": 0.9496921405287939, "grad_norm": 1.1844494963231111, "learning_rate": 1.3245233036786597e-07, "loss": 0.7207, "step": 26221 }, { "epoch": 0.9497283592901122, "grad_norm": 1.4767427082194953, "learning_rate": 1.322621042956418e-07, "loss": 0.6968, "step": 26222 }, { "epoch": 0.9497645780514307, "grad_norm": 1.3189472653772625, "learning_rate": 1.320720140121612e-07, "loss": 0.6464, "step": 26223 }, { "epoch": 0.949800796812749, "grad_norm": 1.3993022052758572, "learning_rate": 1.3188205952003984e-07, "loss": 0.7037, "step": 26224 }, { "epoch": 0.9498370155740674, "grad_norm": 1.301696954779837, "learning_rate": 1.3169224082189236e-07, "loss": 0.6516, "step": 26225 }, { "epoch": 0.9498732343353857, "grad_norm": 1.26937158460876, "learning_rate": 1.3150255792033107e-07, "loss": 0.7026, "step": 26226 }, { "epoch": 0.949909453096704, "grad_norm": 1.3415859048229302, "learning_rate": 1.313130108179661e-07, "loss": 0.6621, "step": 26227 }, { "epoch": 0.9499456718580225, "grad_norm": 1.3706130267764867, "learning_rate": 1.3112359951740428e-07, "loss": 0.6401, "step": 26228 }, { "epoch": 0.9499818906193408, "grad_norm": 1.3812798233609995, "learning_rate": 1.309343240212546e-07, "loss": 0.6685, "step": 26229 }, { "epoch": 0.9500181093806592, "grad_norm": 1.4356659191103913, "learning_rate": 1.307451843321206e-07, "loss": 0.7007, "step": 26230 }, { "epoch": 0.9500543281419775, "grad_norm": 1.4085693432105166, "learning_rate": 1.3055618045260344e-07, "loss": 0.6955, "step": 26231 }, { "epoch": 0.9500905469032959, "grad_norm": 1.3662439360052117, "learning_rate": 1.3036731238530553e-07, "loss": 0.6771, "step": 26232 }, { "epoch": 0.9501267656646143, "grad_norm": 1.0033283141961429, "learning_rate": 1.3017858013282702e-07, "loss": 0.6979, "step": 26233 }, { "epoch": 0.9501629844259326, "grad_norm": 1.626280165728988, "learning_rate": 1.299899836977625e-07, "loss": 0.7442, "step": 26234 }, { "epoch": 0.950199203187251, "grad_norm": 1.4501446686195811, "learning_rate": 1.2980152308270877e-07, "loss": 0.7267, "step": 26235 }, { "epoch": 0.9502354219485694, "grad_norm": 1.1722438773593928, "learning_rate": 1.2961319829025708e-07, "loss": 0.6569, "step": 26236 }, { "epoch": 0.9502716407098877, "grad_norm": 0.9761618528759399, "learning_rate": 1.2942500932300317e-07, "loss": 0.6848, "step": 26237 }, { "epoch": 0.9503078594712061, "grad_norm": 0.9307821204302246, "learning_rate": 1.2923695618353272e-07, "loss": 0.7123, "step": 26238 }, { "epoch": 0.9503440782325244, "grad_norm": 1.3521884684165801, "learning_rate": 1.2904903887443477e-07, "loss": 0.629, "step": 26239 }, { "epoch": 0.9503802969938429, "grad_norm": 1.367909317887815, "learning_rate": 1.2886125739829724e-07, "loss": 0.7119, "step": 26240 }, { "epoch": 0.9504165157551612, "grad_norm": 1.4650719731881692, "learning_rate": 1.2867361175770143e-07, "loss": 0.6961, "step": 26241 }, { "epoch": 0.9504527345164795, "grad_norm": 1.350792934473091, "learning_rate": 1.284861019552297e-07, "loss": 0.701, "step": 26242 }, { "epoch": 0.9504889532777979, "grad_norm": 1.293247597638979, "learning_rate": 1.2829872799346334e-07, "loss": 0.6545, "step": 26243 }, { "epoch": 0.9505251720391162, "grad_norm": 1.2328463109170082, "learning_rate": 1.2811148987498025e-07, "loss": 0.566, "step": 26244 }, { "epoch": 0.9505613908004347, "grad_norm": 1.4467555658231495, "learning_rate": 1.2792438760235838e-07, "loss": 0.65, "step": 26245 }, { "epoch": 0.950597609561753, "grad_norm": 1.084282923838417, "learning_rate": 1.2773742117817012e-07, "loss": 0.6242, "step": 26246 }, { "epoch": 0.9506338283230713, "grad_norm": 1.2960663667261425, "learning_rate": 1.2755059060499008e-07, "loss": 0.6614, "step": 26247 }, { "epoch": 0.9506700470843897, "grad_norm": 1.307624956112178, "learning_rate": 1.2736389588538844e-07, "loss": 0.6347, "step": 26248 }, { "epoch": 0.950706265845708, "grad_norm": 1.0365510986885178, "learning_rate": 1.2717733702193424e-07, "loss": 0.7145, "step": 26249 }, { "epoch": 0.9507424846070265, "grad_norm": 1.3622450073733159, "learning_rate": 1.2699091401719544e-07, "loss": 0.6971, "step": 26250 }, { "epoch": 0.9507787033683448, "grad_norm": 1.1146106946599663, "learning_rate": 1.2680462687373664e-07, "loss": 0.6599, "step": 26251 }, { "epoch": 0.9508149221296631, "grad_norm": 1.3064298416980464, "learning_rate": 1.2661847559412133e-07, "loss": 0.6261, "step": 26252 }, { "epoch": 0.9508511408909815, "grad_norm": 1.3136867780404402, "learning_rate": 1.264324601809108e-07, "loss": 0.6134, "step": 26253 }, { "epoch": 0.9508873596522999, "grad_norm": 1.3657285760417757, "learning_rate": 1.262465806366664e-07, "loss": 0.6503, "step": 26254 }, { "epoch": 0.9509235784136183, "grad_norm": 1.024251831380272, "learning_rate": 1.2606083696394488e-07, "loss": 0.6495, "step": 26255 }, { "epoch": 0.9509597971749366, "grad_norm": 1.2963808900087865, "learning_rate": 1.2587522916530093e-07, "loss": 0.6803, "step": 26256 }, { "epoch": 0.9509960159362549, "grad_norm": 1.4889708101727612, "learning_rate": 1.2568975724329134e-07, "loss": 0.707, "step": 26257 }, { "epoch": 0.9510322346975734, "grad_norm": 0.9918300066344745, "learning_rate": 1.2550442120046636e-07, "loss": 0.613, "step": 26258 }, { "epoch": 0.9510684534588917, "grad_norm": 0.9281511252549763, "learning_rate": 1.253192210393772e-07, "loss": 0.6348, "step": 26259 }, { "epoch": 0.9511046722202101, "grad_norm": 1.2351063262518363, "learning_rate": 1.251341567625719e-07, "loss": 0.6735, "step": 26260 }, { "epoch": 0.9511408909815284, "grad_norm": 1.458171489085254, "learning_rate": 1.2494922837259727e-07, "loss": 0.6716, "step": 26261 }, { "epoch": 0.9511771097428467, "grad_norm": 1.4851973614729095, "learning_rate": 1.2476443587199904e-07, "loss": 0.6644, "step": 26262 }, { "epoch": 0.9512133285041652, "grad_norm": 1.0570467914818884, "learning_rate": 1.2457977926331966e-07, "loss": 0.722, "step": 26263 }, { "epoch": 0.9512495472654835, "grad_norm": 1.0478838918827837, "learning_rate": 1.2439525854909817e-07, "loss": 0.6849, "step": 26264 }, { "epoch": 0.9512857660268019, "grad_norm": 1.378782904406089, "learning_rate": 1.24210873731877e-07, "loss": 0.6176, "step": 26265 }, { "epoch": 0.9513219847881202, "grad_norm": 1.3931033888011686, "learning_rate": 1.2402662481419082e-07, "loss": 0.6701, "step": 26266 }, { "epoch": 0.9513582035494386, "grad_norm": 1.35520734382562, "learning_rate": 1.2384251179857642e-07, "loss": 0.6469, "step": 26267 }, { "epoch": 0.951394422310757, "grad_norm": 1.4264826179557086, "learning_rate": 1.2365853468756628e-07, "loss": 0.6077, "step": 26268 }, { "epoch": 0.9514306410720753, "grad_norm": 1.3405957822350114, "learning_rate": 1.2347469348369389e-07, "loss": 0.6967, "step": 26269 }, { "epoch": 0.9514668598333937, "grad_norm": 1.3848897181597446, "learning_rate": 1.2329098818948836e-07, "loss": 0.6341, "step": 26270 }, { "epoch": 0.9515030785947121, "grad_norm": 1.1410256228385223, "learning_rate": 1.2310741880747545e-07, "loss": 0.6237, "step": 26271 }, { "epoch": 0.9515392973560304, "grad_norm": 1.3196718019971825, "learning_rate": 1.2292398534018423e-07, "loss": 0.659, "step": 26272 }, { "epoch": 0.9515755161173488, "grad_norm": 1.3459720758053955, "learning_rate": 1.2274068779013825e-07, "loss": 0.6794, "step": 26273 }, { "epoch": 0.9516117348786671, "grad_norm": 1.358104314455225, "learning_rate": 1.2255752615985884e-07, "loss": 0.624, "step": 26274 }, { "epoch": 0.9516479536399856, "grad_norm": 1.3359931277522439, "learning_rate": 1.2237450045186616e-07, "loss": 0.6669, "step": 26275 }, { "epoch": 0.9516841724013039, "grad_norm": 1.344407747119564, "learning_rate": 1.2219161066868047e-07, "loss": 0.6777, "step": 26276 }, { "epoch": 0.9517203911626222, "grad_norm": 1.4205976803687321, "learning_rate": 1.220088568128175e-07, "loss": 0.7211, "step": 26277 }, { "epoch": 0.9517566099239406, "grad_norm": 1.4066607511182818, "learning_rate": 1.21826238886793e-07, "loss": 0.6778, "step": 26278 }, { "epoch": 0.9517928286852589, "grad_norm": 1.0518118522412303, "learning_rate": 1.2164375689311725e-07, "loss": 0.7049, "step": 26279 }, { "epoch": 0.9518290474465774, "grad_norm": 1.370150615314265, "learning_rate": 1.2146141083430486e-07, "loss": 0.6725, "step": 26280 }, { "epoch": 0.9518652662078957, "grad_norm": 1.2323343454463773, "learning_rate": 1.212792007128638e-07, "loss": 0.5907, "step": 26281 }, { "epoch": 0.951901484969214, "grad_norm": 1.0293476776860768, "learning_rate": 1.210971265312988e-07, "loss": 0.6749, "step": 26282 }, { "epoch": 0.9519377037305324, "grad_norm": 1.257966572642737, "learning_rate": 1.2091518829212e-07, "loss": 0.6211, "step": 26283 }, { "epoch": 0.9519739224918508, "grad_norm": 1.4908999410418748, "learning_rate": 1.207333859978277e-07, "loss": 0.6953, "step": 26284 }, { "epoch": 0.9520101412531692, "grad_norm": 1.300042066959476, "learning_rate": 1.2055171965092538e-07, "loss": 0.6458, "step": 26285 }, { "epoch": 0.9520463600144875, "grad_norm": 1.494136529181358, "learning_rate": 1.2037018925391108e-07, "loss": 0.6223, "step": 26286 }, { "epoch": 0.9520825787758058, "grad_norm": 1.2465610636865179, "learning_rate": 1.2018879480928502e-07, "loss": 0.6444, "step": 26287 }, { "epoch": 0.9521187975371243, "grad_norm": 1.2713328299033606, "learning_rate": 1.2000753631954298e-07, "loss": 0.6232, "step": 26288 }, { "epoch": 0.9521550162984426, "grad_norm": 1.3869280885460722, "learning_rate": 1.198264137871774e-07, "loss": 0.6693, "step": 26289 }, { "epoch": 0.952191235059761, "grad_norm": 1.3622099388835924, "learning_rate": 1.1964542721468187e-07, "loss": 0.6815, "step": 26290 }, { "epoch": 0.9522274538210793, "grad_norm": 1.3654472146964396, "learning_rate": 1.1946457660454657e-07, "loss": 0.6286, "step": 26291 }, { "epoch": 0.9522636725823976, "grad_norm": 1.202064507356714, "learning_rate": 1.192838619592618e-07, "loss": 0.6482, "step": 26292 }, { "epoch": 0.9522998913437161, "grad_norm": 1.1679213051843786, "learning_rate": 1.1910328328131105e-07, "loss": 0.6209, "step": 26293 }, { "epoch": 0.9523361101050344, "grad_norm": 1.047634649856347, "learning_rate": 1.1892284057318349e-07, "loss": 0.6448, "step": 26294 }, { "epoch": 0.9523723288663528, "grad_norm": 1.4463094504450011, "learning_rate": 1.1874253383735824e-07, "loss": 0.6625, "step": 26295 }, { "epoch": 0.9524085476276711, "grad_norm": 1.3099019116500172, "learning_rate": 1.1856236307631996e-07, "loss": 0.6401, "step": 26296 }, { "epoch": 0.9524447663889894, "grad_norm": 1.4113190194793581, "learning_rate": 1.1838232829254447e-07, "loss": 0.6611, "step": 26297 }, { "epoch": 0.9524809851503079, "grad_norm": 1.4263635804831842, "learning_rate": 1.1820242948851091e-07, "loss": 0.7428, "step": 26298 }, { "epoch": 0.9525172039116262, "grad_norm": 1.3726263562183942, "learning_rate": 1.1802266666669616e-07, "loss": 0.7068, "step": 26299 }, { "epoch": 0.9525534226729446, "grad_norm": 1.5135807649818023, "learning_rate": 1.1784303982957158e-07, "loss": 0.6675, "step": 26300 }, { "epoch": 0.952589641434263, "grad_norm": 1.5080105819449448, "learning_rate": 1.1766354897960963e-07, "loss": 0.6691, "step": 26301 }, { "epoch": 0.9526258601955813, "grad_norm": 1.003623059816152, "learning_rate": 1.1748419411928169e-07, "loss": 0.6936, "step": 26302 }, { "epoch": 0.9526620789568997, "grad_norm": 1.3542001694117283, "learning_rate": 1.1730497525105356e-07, "loss": 0.6023, "step": 26303 }, { "epoch": 0.952698297718218, "grad_norm": 1.413173096162885, "learning_rate": 1.1712589237739324e-07, "loss": 0.6394, "step": 26304 }, { "epoch": 0.9527345164795364, "grad_norm": 1.436892363542716, "learning_rate": 1.1694694550076435e-07, "loss": 0.6562, "step": 26305 }, { "epoch": 0.9527707352408548, "grad_norm": 1.3084584707110685, "learning_rate": 1.1676813462363046e-07, "loss": 0.5862, "step": 26306 }, { "epoch": 0.9528069540021731, "grad_norm": 1.2947989422714454, "learning_rate": 1.1658945974845071e-07, "loss": 0.6849, "step": 26307 }, { "epoch": 0.9528431727634915, "grad_norm": 1.4688969168332537, "learning_rate": 1.1641092087768314e-07, "loss": 0.6671, "step": 26308 }, { "epoch": 0.9528793915248098, "grad_norm": 1.124929942518688, "learning_rate": 1.16232518013788e-07, "loss": 0.6832, "step": 26309 }, { "epoch": 0.9529156102861283, "grad_norm": 1.3858534678860084, "learning_rate": 1.1605425115921665e-07, "loss": 0.6331, "step": 26310 }, { "epoch": 0.9529518290474466, "grad_norm": 1.4792471629277213, "learning_rate": 1.1587612031642271e-07, "loss": 0.6705, "step": 26311 }, { "epoch": 0.9529880478087649, "grad_norm": 1.2857383102713171, "learning_rate": 1.1569812548785974e-07, "loss": 0.6436, "step": 26312 }, { "epoch": 0.9530242665700833, "grad_norm": 1.089157902821204, "learning_rate": 1.1552026667597582e-07, "loss": 0.6522, "step": 26313 }, { "epoch": 0.9530604853314016, "grad_norm": 1.0411591894941352, "learning_rate": 1.1534254388321786e-07, "loss": 0.6386, "step": 26314 }, { "epoch": 0.9530967040927201, "grad_norm": 1.413619118143778, "learning_rate": 1.1516495711203169e-07, "loss": 0.6602, "step": 26315 }, { "epoch": 0.9531329228540384, "grad_norm": 1.3682278863967183, "learning_rate": 1.1498750636486089e-07, "loss": 0.6501, "step": 26316 }, { "epoch": 0.9531691416153567, "grad_norm": 1.3575748871936286, "learning_rate": 1.1481019164414909e-07, "loss": 0.6753, "step": 26317 }, { "epoch": 0.9532053603766751, "grad_norm": 1.3981714055432812, "learning_rate": 1.146330129523343e-07, "loss": 0.7203, "step": 26318 }, { "epoch": 0.9532415791379935, "grad_norm": 1.370320950050355, "learning_rate": 1.1445597029185462e-07, "loss": 0.6916, "step": 26319 }, { "epoch": 0.9532777978993119, "grad_norm": 1.1834678712796358, "learning_rate": 1.1427906366514696e-07, "loss": 0.6895, "step": 26320 }, { "epoch": 0.9533140166606302, "grad_norm": 0.9533398522995804, "learning_rate": 1.1410229307464715e-07, "loss": 0.7022, "step": 26321 }, { "epoch": 0.9533502354219485, "grad_norm": 1.4055290440364538, "learning_rate": 1.1392565852278437e-07, "loss": 0.6995, "step": 26322 }, { "epoch": 0.953386454183267, "grad_norm": 1.5278874183336744, "learning_rate": 1.137491600119922e-07, "loss": 0.6468, "step": 26323 }, { "epoch": 0.9534226729445853, "grad_norm": 1.4430155273542713, "learning_rate": 1.1357279754469874e-07, "loss": 0.7325, "step": 26324 }, { "epoch": 0.9534588917059037, "grad_norm": 1.3035595973894214, "learning_rate": 1.1339657112333091e-07, "loss": 0.6273, "step": 26325 }, { "epoch": 0.953495110467222, "grad_norm": 1.4190195685094045, "learning_rate": 1.1322048075031233e-07, "loss": 0.6696, "step": 26326 }, { "epoch": 0.9535313292285403, "grad_norm": 1.3823322604480077, "learning_rate": 1.1304452642806773e-07, "loss": 0.6842, "step": 26327 }, { "epoch": 0.9535675479898588, "grad_norm": 1.5584651050203355, "learning_rate": 1.128687081590174e-07, "loss": 0.7431, "step": 26328 }, { "epoch": 0.9536037667511771, "grad_norm": 1.3887711065653736, "learning_rate": 1.1269302594558163e-07, "loss": 0.6528, "step": 26329 }, { "epoch": 0.9536399855124955, "grad_norm": 0.9251539187830502, "learning_rate": 1.1251747979017735e-07, "loss": 0.7014, "step": 26330 }, { "epoch": 0.9536762042738138, "grad_norm": 1.1362901459555834, "learning_rate": 1.1234206969522155e-07, "loss": 0.6302, "step": 26331 }, { "epoch": 0.9537124230351322, "grad_norm": 1.4272893063205498, "learning_rate": 1.121667956631256e-07, "loss": 0.6879, "step": 26332 }, { "epoch": 0.9537486417964506, "grad_norm": 1.4369212253968093, "learning_rate": 1.1199165769630316e-07, "loss": 0.6088, "step": 26333 }, { "epoch": 0.9537848605577689, "grad_norm": 1.0144916028910218, "learning_rate": 1.1181665579716339e-07, "loss": 0.6603, "step": 26334 }, { "epoch": 0.9538210793190873, "grad_norm": 1.2306885651626984, "learning_rate": 1.1164178996811659e-07, "loss": 0.6377, "step": 26335 }, { "epoch": 0.9538572980804056, "grad_norm": 1.3599319331100201, "learning_rate": 1.1146706021156639e-07, "loss": 0.6744, "step": 26336 }, { "epoch": 0.953893516841724, "grad_norm": 1.5004467280326128, "learning_rate": 1.1129246652991755e-07, "loss": 0.6643, "step": 26337 }, { "epoch": 0.9539297356030424, "grad_norm": 1.2802345115743954, "learning_rate": 1.1111800892557367e-07, "loss": 0.6475, "step": 26338 }, { "epoch": 0.9539659543643607, "grad_norm": 1.3090208096899163, "learning_rate": 1.109436874009362e-07, "loss": 0.5978, "step": 26339 }, { "epoch": 0.9540021731256791, "grad_norm": 1.4088252263495995, "learning_rate": 1.1076950195840208e-07, "loss": 0.665, "step": 26340 }, { "epoch": 0.9540383918869975, "grad_norm": 0.9909443971992706, "learning_rate": 1.105954526003683e-07, "loss": 0.6637, "step": 26341 }, { "epoch": 0.9540746106483158, "grad_norm": 0.974167956154793, "learning_rate": 1.1042153932923184e-07, "loss": 0.6907, "step": 26342 }, { "epoch": 0.9541108294096342, "grad_norm": 1.341900087187797, "learning_rate": 1.102477621473852e-07, "loss": 0.6564, "step": 26343 }, { "epoch": 0.9541470481709525, "grad_norm": 1.018585364171458, "learning_rate": 1.100741210572176e-07, "loss": 0.6852, "step": 26344 }, { "epoch": 0.954183266932271, "grad_norm": 1.4617870125257693, "learning_rate": 1.0990061606112156e-07, "loss": 0.6767, "step": 26345 }, { "epoch": 0.9542194856935893, "grad_norm": 1.3572776168902148, "learning_rate": 1.0972724716148186e-07, "loss": 0.6708, "step": 26346 }, { "epoch": 0.9542557044549076, "grad_norm": 1.306065027758769, "learning_rate": 1.0955401436068657e-07, "loss": 0.5764, "step": 26347 }, { "epoch": 0.954291923216226, "grad_norm": 1.6762056635782796, "learning_rate": 1.0938091766111825e-07, "loss": 0.7128, "step": 26348 }, { "epoch": 0.9543281419775443, "grad_norm": 1.57974848355984, "learning_rate": 1.0920795706515941e-07, "loss": 0.7078, "step": 26349 }, { "epoch": 0.9543643607388628, "grad_norm": 1.373123819981625, "learning_rate": 1.0903513257519039e-07, "loss": 0.6808, "step": 26350 }, { "epoch": 0.9544005795001811, "grad_norm": 0.9674702393981655, "learning_rate": 1.0886244419358816e-07, "loss": 0.6749, "step": 26351 }, { "epoch": 0.9544367982614994, "grad_norm": 1.336381153285604, "learning_rate": 1.0868989192272972e-07, "loss": 0.618, "step": 26352 }, { "epoch": 0.9544730170228178, "grad_norm": 1.0619819068835092, "learning_rate": 1.0851747576498983e-07, "loss": 0.6597, "step": 26353 }, { "epoch": 0.9545092357841362, "grad_norm": 1.3912030348740092, "learning_rate": 1.0834519572274105e-07, "loss": 0.6871, "step": 26354 }, { "epoch": 0.9545454545454546, "grad_norm": 0.9689788199025957, "learning_rate": 1.0817305179835368e-07, "loss": 0.6802, "step": 26355 }, { "epoch": 0.9545816733067729, "grad_norm": 1.2390080982703013, "learning_rate": 1.0800104399419698e-07, "loss": 0.5752, "step": 26356 }, { "epoch": 0.9546178920680912, "grad_norm": 1.3626244979704725, "learning_rate": 1.078291723126379e-07, "loss": 0.5941, "step": 26357 }, { "epoch": 0.9546541108294097, "grad_norm": 1.2595417104636175, "learning_rate": 1.0765743675604123e-07, "loss": 0.6397, "step": 26358 }, { "epoch": 0.954690329590728, "grad_norm": 1.0178996850680726, "learning_rate": 1.0748583732677065e-07, "loss": 0.6334, "step": 26359 }, { "epoch": 0.9547265483520464, "grad_norm": 1.2946362971652896, "learning_rate": 1.0731437402718647e-07, "loss": 0.6604, "step": 26360 }, { "epoch": 0.9547627671133647, "grad_norm": 1.120620001229111, "learning_rate": 1.0714304685965015e-07, "loss": 0.7075, "step": 26361 }, { "epoch": 0.954798985874683, "grad_norm": 1.0918949954304682, "learning_rate": 1.0697185582651759e-07, "loss": 0.6445, "step": 26362 }, { "epoch": 0.9548352046360015, "grad_norm": 1.3879200594309533, "learning_rate": 1.0680080093014466e-07, "loss": 0.6535, "step": 26363 }, { "epoch": 0.9548714233973198, "grad_norm": 1.3300287689876795, "learning_rate": 1.0662988217288617e-07, "loss": 0.6551, "step": 26364 }, { "epoch": 0.9549076421586382, "grad_norm": 1.3828545526263303, "learning_rate": 1.0645909955709355e-07, "loss": 0.6569, "step": 26365 }, { "epoch": 0.9549438609199565, "grad_norm": 1.0336626758488474, "learning_rate": 1.0628845308511714e-07, "loss": 0.6303, "step": 26366 }, { "epoch": 0.9549800796812749, "grad_norm": 1.2236127707140814, "learning_rate": 1.0611794275930398e-07, "loss": 0.6732, "step": 26367 }, { "epoch": 0.9550162984425933, "grad_norm": 1.2598491071792717, "learning_rate": 1.0594756858200328e-07, "loss": 0.6256, "step": 26368 }, { "epoch": 0.9550525172039116, "grad_norm": 1.038741223590242, "learning_rate": 1.057773305555565e-07, "loss": 0.6957, "step": 26369 }, { "epoch": 0.95508873596523, "grad_norm": 1.3471250503024919, "learning_rate": 1.0560722868230733e-07, "loss": 0.6389, "step": 26370 }, { "epoch": 0.9551249547265483, "grad_norm": 1.2991215833261418, "learning_rate": 1.0543726296459722e-07, "loss": 0.6366, "step": 26371 }, { "epoch": 0.9551611734878667, "grad_norm": 1.3043016781222962, "learning_rate": 1.0526743340476431e-07, "loss": 0.7219, "step": 26372 }, { "epoch": 0.9551973922491851, "grad_norm": 1.3023190729999015, "learning_rate": 1.050977400051445e-07, "loss": 0.6288, "step": 26373 }, { "epoch": 0.9552336110105034, "grad_norm": 1.4495659075639258, "learning_rate": 1.0492818276807592e-07, "loss": 0.6124, "step": 26374 }, { "epoch": 0.9552698297718218, "grad_norm": 1.3817074476696705, "learning_rate": 1.0475876169589005e-07, "loss": 0.6783, "step": 26375 }, { "epoch": 0.9553060485331402, "grad_norm": 1.3668930520013358, "learning_rate": 1.0458947679091724e-07, "loss": 0.6387, "step": 26376 }, { "epoch": 0.9553422672944585, "grad_norm": 1.3366894620399044, "learning_rate": 1.0442032805548896e-07, "loss": 0.666, "step": 26377 }, { "epoch": 0.9553784860557769, "grad_norm": 1.2802685389072215, "learning_rate": 1.0425131549193224e-07, "loss": 0.6603, "step": 26378 }, { "epoch": 0.9554147048170952, "grad_norm": 1.4372911612528272, "learning_rate": 1.0408243910257187e-07, "loss": 0.647, "step": 26379 }, { "epoch": 0.9554509235784137, "grad_norm": 1.3823345066713144, "learning_rate": 1.0391369888973269e-07, "loss": 0.649, "step": 26380 }, { "epoch": 0.955487142339732, "grad_norm": 1.385843134853242, "learning_rate": 1.0374509485573614e-07, "loss": 0.6921, "step": 26381 }, { "epoch": 0.9555233611010503, "grad_norm": 1.1553493880988286, "learning_rate": 1.0357662700290372e-07, "loss": 0.6722, "step": 26382 }, { "epoch": 0.9555595798623687, "grad_norm": 1.2389919203930122, "learning_rate": 1.0340829533355135e-07, "loss": 0.6919, "step": 26383 }, { "epoch": 0.955595798623687, "grad_norm": 1.29678531129775, "learning_rate": 1.0324009984999717e-07, "loss": 0.6687, "step": 26384 }, { "epoch": 0.9556320173850055, "grad_norm": 1.3092078877996722, "learning_rate": 1.03072040554556e-07, "loss": 0.7139, "step": 26385 }, { "epoch": 0.9556682361463238, "grad_norm": 1.365317824409643, "learning_rate": 1.0290411744953932e-07, "loss": 0.6825, "step": 26386 }, { "epoch": 0.9557044549076421, "grad_norm": 1.2879887443701883, "learning_rate": 1.0273633053725862e-07, "loss": 0.675, "step": 26387 }, { "epoch": 0.9557406736689605, "grad_norm": 1.192433611563309, "learning_rate": 1.0256867982002094e-07, "loss": 0.6055, "step": 26388 }, { "epoch": 0.9557768924302789, "grad_norm": 1.0253338064334896, "learning_rate": 1.0240116530013666e-07, "loss": 0.6476, "step": 26389 }, { "epoch": 0.9558131111915973, "grad_norm": 1.3105549975511601, "learning_rate": 1.0223378697990837e-07, "loss": 0.5824, "step": 26390 }, { "epoch": 0.9558493299529156, "grad_norm": 1.2758254073934683, "learning_rate": 1.0206654486164091e-07, "loss": 0.6393, "step": 26391 }, { "epoch": 0.9558855487142339, "grad_norm": 1.4901041071439658, "learning_rate": 1.0189943894763355e-07, "loss": 0.6572, "step": 26392 }, { "epoch": 0.9559217674755524, "grad_norm": 1.026884058684931, "learning_rate": 1.0173246924018775e-07, "loss": 0.7138, "step": 26393 }, { "epoch": 0.9559579862368707, "grad_norm": 1.3839951269564945, "learning_rate": 1.0156563574160172e-07, "loss": 0.5534, "step": 26394 }, { "epoch": 0.9559942049981891, "grad_norm": 1.3393814322207693, "learning_rate": 1.0139893845416804e-07, "loss": 0.6738, "step": 26395 }, { "epoch": 0.9560304237595074, "grad_norm": 0.9798662113003205, "learning_rate": 1.0123237738018377e-07, "loss": 0.6006, "step": 26396 }, { "epoch": 0.9560666425208257, "grad_norm": 1.150637953843059, "learning_rate": 1.0106595252193929e-07, "loss": 0.5982, "step": 26397 }, { "epoch": 0.9561028612821442, "grad_norm": 1.2820869356704152, "learning_rate": 1.0089966388172611e-07, "loss": 0.6584, "step": 26398 }, { "epoch": 0.9561390800434625, "grad_norm": 1.4945973975902191, "learning_rate": 1.007335114618313e-07, "loss": 0.7003, "step": 26399 }, { "epoch": 0.9561752988047809, "grad_norm": 1.326783503743167, "learning_rate": 1.0056749526454079e-07, "loss": 0.6815, "step": 26400 }, { "epoch": 0.9562115175660992, "grad_norm": 0.9545501036981889, "learning_rate": 1.0040161529214166e-07, "loss": 0.6489, "step": 26401 }, { "epoch": 0.9562477363274176, "grad_norm": 1.112570904932729, "learning_rate": 1.0023587154691317e-07, "loss": 0.6913, "step": 26402 }, { "epoch": 0.956283955088736, "grad_norm": 0.9688186758460221, "learning_rate": 1.0007026403113906e-07, "loss": 0.6998, "step": 26403 }, { "epoch": 0.9563201738500543, "grad_norm": 1.417955255312849, "learning_rate": 9.990479274709641e-08, "loss": 0.6893, "step": 26404 }, { "epoch": 0.9563563926113727, "grad_norm": 1.3038261191246214, "learning_rate": 9.973945769706228e-08, "loss": 0.6359, "step": 26405 }, { "epoch": 0.956392611372691, "grad_norm": 1.1673088697975853, "learning_rate": 9.957425888331262e-08, "loss": 0.6171, "step": 26406 }, { "epoch": 0.9564288301340094, "grad_norm": 1.2417908521075973, "learning_rate": 9.940919630812117e-08, "loss": 0.6824, "step": 26407 }, { "epoch": 0.9564650488953278, "grad_norm": 1.3623967498517895, "learning_rate": 9.924426997375836e-08, "loss": 0.6818, "step": 26408 }, { "epoch": 0.9565012676566461, "grad_norm": 1.2971620715931342, "learning_rate": 9.907947988249344e-08, "loss": 0.6674, "step": 26409 }, { "epoch": 0.9565374864179645, "grad_norm": 1.2963461048783114, "learning_rate": 9.891482603659463e-08, "loss": 0.62, "step": 26410 }, { "epoch": 0.9565737051792829, "grad_norm": 1.4865623106664665, "learning_rate": 9.875030843832789e-08, "loss": 0.7062, "step": 26411 }, { "epoch": 0.9566099239406012, "grad_norm": 1.2954306408541858, "learning_rate": 9.858592708995696e-08, "loss": 0.6599, "step": 26412 }, { "epoch": 0.9566461427019196, "grad_norm": 0.9810783740251328, "learning_rate": 9.842168199374336e-08, "loss": 0.6625, "step": 26413 }, { "epoch": 0.9566823614632379, "grad_norm": 1.3514462638772464, "learning_rate": 9.825757315194751e-08, "loss": 0.7103, "step": 26414 }, { "epoch": 0.9567185802245564, "grad_norm": 1.4711138011317475, "learning_rate": 9.80936005668287e-08, "loss": 0.6504, "step": 26415 }, { "epoch": 0.9567547989858747, "grad_norm": 1.3365005065297637, "learning_rate": 9.792976424064293e-08, "loss": 0.6393, "step": 26416 }, { "epoch": 0.956791017747193, "grad_norm": 1.4628070564235964, "learning_rate": 9.776606417564283e-08, "loss": 0.7227, "step": 26417 }, { "epoch": 0.9568272365085114, "grad_norm": 1.3251729399690788, "learning_rate": 9.760250037408325e-08, "loss": 0.679, "step": 26418 }, { "epoch": 0.9568634552698297, "grad_norm": 1.237295737126908, "learning_rate": 9.743907283821463e-08, "loss": 0.6501, "step": 26419 }, { "epoch": 0.9568996740311482, "grad_norm": 1.3448517034031084, "learning_rate": 9.727578157028627e-08, "loss": 0.6625, "step": 26420 }, { "epoch": 0.9569358927924665, "grad_norm": 1.3589770245606843, "learning_rate": 9.711262657254305e-08, "loss": 0.6123, "step": 26421 }, { "epoch": 0.9569721115537848, "grad_norm": 1.4959323057660654, "learning_rate": 9.694960784723206e-08, "loss": 0.7043, "step": 26422 }, { "epoch": 0.9570083303151032, "grad_norm": 1.0988681143337617, "learning_rate": 9.678672539659595e-08, "loss": 0.731, "step": 26423 }, { "epoch": 0.9570445490764216, "grad_norm": 1.6472679210535575, "learning_rate": 9.662397922287625e-08, "loss": 0.6778, "step": 26424 }, { "epoch": 0.95708076783774, "grad_norm": 1.3788725846155054, "learning_rate": 9.64613693283123e-08, "loss": 0.6604, "step": 26425 }, { "epoch": 0.9571169865990583, "grad_norm": 1.0171551059952098, "learning_rate": 9.629889571514228e-08, "loss": 0.703, "step": 26426 }, { "epoch": 0.9571532053603766, "grad_norm": 1.4748576024792452, "learning_rate": 9.61365583856011e-08, "loss": 0.6618, "step": 26427 }, { "epoch": 0.957189424121695, "grad_norm": 1.338172696304519, "learning_rate": 9.597435734192251e-08, "loss": 0.6083, "step": 26428 }, { "epoch": 0.9572256428830134, "grad_norm": 1.1693060012678627, "learning_rate": 9.581229258634028e-08, "loss": 0.6231, "step": 26429 }, { "epoch": 0.9572618616443318, "grad_norm": 1.7440400217395458, "learning_rate": 9.565036412108375e-08, "loss": 0.7267, "step": 26430 }, { "epoch": 0.9572980804056501, "grad_norm": 1.5270203665124273, "learning_rate": 9.548857194838112e-08, "loss": 0.6501, "step": 26431 }, { "epoch": 0.9573342991669684, "grad_norm": 1.0122318907108092, "learning_rate": 9.53269160704573e-08, "loss": 0.612, "step": 26432 }, { "epoch": 0.9573705179282869, "grad_norm": 1.336878048096134, "learning_rate": 9.516539648953937e-08, "loss": 0.6079, "step": 26433 }, { "epoch": 0.9574067366896052, "grad_norm": 0.9970701802138745, "learning_rate": 9.50040132078478e-08, "loss": 0.6123, "step": 26434 }, { "epoch": 0.9574429554509236, "grad_norm": 1.4381360617640762, "learning_rate": 9.484276622760413e-08, "loss": 0.607, "step": 26435 }, { "epoch": 0.9574791742122419, "grad_norm": 1.3089199179213173, "learning_rate": 9.468165555102882e-08, "loss": 0.6599, "step": 26436 }, { "epoch": 0.9575153929735603, "grad_norm": 1.397257898083769, "learning_rate": 9.452068118033675e-08, "loss": 0.6479, "step": 26437 }, { "epoch": 0.9575516117348787, "grad_norm": 1.32371411512201, "learning_rate": 9.435984311774394e-08, "loss": 0.6057, "step": 26438 }, { "epoch": 0.957587830496197, "grad_norm": 1.3537284414040063, "learning_rate": 9.419914136546304e-08, "loss": 0.6583, "step": 26439 }, { "epoch": 0.9576240492575154, "grad_norm": 1.4072617963871066, "learning_rate": 9.403857592570675e-08, "loss": 0.6846, "step": 26440 }, { "epoch": 0.9576602680188337, "grad_norm": 1.1661657666222205, "learning_rate": 9.38781468006833e-08, "loss": 0.6717, "step": 26441 }, { "epoch": 0.9576964867801521, "grad_norm": 1.4800896844993907, "learning_rate": 9.371785399260092e-08, "loss": 0.7401, "step": 26442 }, { "epoch": 0.9577327055414705, "grad_norm": 1.3442248105760872, "learning_rate": 9.35576975036645e-08, "loss": 0.6235, "step": 26443 }, { "epoch": 0.9577689243027888, "grad_norm": 1.6026690103216847, "learning_rate": 9.339767733608007e-08, "loss": 0.6934, "step": 26444 }, { "epoch": 0.9578051430641072, "grad_norm": 1.538662179450047, "learning_rate": 9.323779349204698e-08, "loss": 0.7178, "step": 26445 }, { "epoch": 0.9578413618254256, "grad_norm": 1.3405781969768267, "learning_rate": 9.307804597376791e-08, "loss": 0.6143, "step": 26446 }, { "epoch": 0.9578775805867439, "grad_norm": 1.4580175734165761, "learning_rate": 9.291843478343888e-08, "loss": 0.7016, "step": 26447 }, { "epoch": 0.9579137993480623, "grad_norm": 1.4791234101319406, "learning_rate": 9.275895992325812e-08, "loss": 0.6959, "step": 26448 }, { "epoch": 0.9579500181093806, "grad_norm": 1.1492416425252052, "learning_rate": 9.259962139541945e-08, "loss": 0.6745, "step": 26449 }, { "epoch": 0.9579862368706991, "grad_norm": 1.5071546300372234, "learning_rate": 9.244041920211445e-08, "loss": 0.7158, "step": 26450 }, { "epoch": 0.9580224556320174, "grad_norm": 1.41464419018662, "learning_rate": 9.228135334553579e-08, "loss": 0.6806, "step": 26451 }, { "epoch": 0.9580586743933357, "grad_norm": 1.0469536969132824, "learning_rate": 9.212242382787173e-08, "loss": 0.6842, "step": 26452 }, { "epoch": 0.9580948931546541, "grad_norm": 1.3118234286513994, "learning_rate": 9.19636306513072e-08, "loss": 0.6595, "step": 26453 }, { "epoch": 0.9581311119159724, "grad_norm": 1.3320135719491035, "learning_rate": 9.180497381803044e-08, "loss": 0.6088, "step": 26454 }, { "epoch": 0.9581673306772909, "grad_norm": 1.2392106381659345, "learning_rate": 9.164645333022303e-08, "loss": 0.6024, "step": 26455 }, { "epoch": 0.9582035494386092, "grad_norm": 1.4035973544203497, "learning_rate": 9.148806919006659e-08, "loss": 0.6978, "step": 26456 }, { "epoch": 0.9582397681999275, "grad_norm": 1.4457569753393484, "learning_rate": 9.132982139974045e-08, "loss": 0.6744, "step": 26457 }, { "epoch": 0.9582759869612459, "grad_norm": 1.4051195727129744, "learning_rate": 9.117170996142289e-08, "loss": 0.6947, "step": 26458 }, { "epoch": 0.9583122057225643, "grad_norm": 1.328732417554161, "learning_rate": 9.101373487728993e-08, "loss": 0.6722, "step": 26459 }, { "epoch": 0.9583484244838827, "grad_norm": 1.2245138532627704, "learning_rate": 9.085589614951318e-08, "loss": 0.6343, "step": 26460 }, { "epoch": 0.958384643245201, "grad_norm": 1.3725950242360945, "learning_rate": 9.069819378026757e-08, "loss": 0.6722, "step": 26461 }, { "epoch": 0.9584208620065193, "grad_norm": 1.3685530239654224, "learning_rate": 9.054062777172134e-08, "loss": 0.6641, "step": 26462 }, { "epoch": 0.9584570807678378, "grad_norm": 1.0627717994766508, "learning_rate": 9.038319812604279e-08, "loss": 0.6542, "step": 26463 }, { "epoch": 0.9584932995291561, "grad_norm": 1.3571862379872364, "learning_rate": 9.022590484539906e-08, "loss": 0.6322, "step": 26464 }, { "epoch": 0.9585295182904745, "grad_norm": 1.372415048590192, "learning_rate": 9.006874793195509e-08, "loss": 0.6458, "step": 26465 }, { "epoch": 0.9585657370517928, "grad_norm": 1.483049360416968, "learning_rate": 8.991172738787246e-08, "loss": 0.6536, "step": 26466 }, { "epoch": 0.9586019558131111, "grad_norm": 1.4153991449049854, "learning_rate": 8.97548432153117e-08, "loss": 0.6224, "step": 26467 }, { "epoch": 0.9586381745744296, "grad_norm": 0.994906995872767, "learning_rate": 8.959809541643216e-08, "loss": 0.6281, "step": 26468 }, { "epoch": 0.9586743933357479, "grad_norm": 1.0305916065529985, "learning_rate": 8.944148399339103e-08, "loss": 0.6644, "step": 26469 }, { "epoch": 0.9587106120970663, "grad_norm": 1.0047914429985232, "learning_rate": 8.928500894834436e-08, "loss": 0.6368, "step": 26470 }, { "epoch": 0.9587468308583846, "grad_norm": 1.3783101172419474, "learning_rate": 8.912867028344263e-08, "loss": 0.6699, "step": 26471 }, { "epoch": 0.958783049619703, "grad_norm": 1.0562202092974624, "learning_rate": 8.89724680008397e-08, "loss": 0.6666, "step": 26472 }, { "epoch": 0.9588192683810214, "grad_norm": 1.3273957024264404, "learning_rate": 8.881640210268494e-08, "loss": 0.6139, "step": 26473 }, { "epoch": 0.9588554871423397, "grad_norm": 1.316280152368562, "learning_rate": 8.866047259112442e-08, "loss": 0.6227, "step": 26474 }, { "epoch": 0.9588917059036581, "grad_norm": 1.3251143355567012, "learning_rate": 8.850467946830421e-08, "loss": 0.6496, "step": 26475 }, { "epoch": 0.9589279246649764, "grad_norm": 1.2527338813017113, "learning_rate": 8.834902273636925e-08, "loss": 0.7092, "step": 26476 }, { "epoch": 0.9589641434262948, "grad_norm": 1.317577514512582, "learning_rate": 8.819350239746005e-08, "loss": 0.714, "step": 26477 }, { "epoch": 0.9590003621876132, "grad_norm": 1.3519142573770182, "learning_rate": 8.803811845371823e-08, "loss": 0.613, "step": 26478 }, { "epoch": 0.9590365809489315, "grad_norm": 1.2901183861277523, "learning_rate": 8.788287090728098e-08, "loss": 0.6481, "step": 26479 }, { "epoch": 0.95907279971025, "grad_norm": 1.4384396296445408, "learning_rate": 8.772775976028547e-08, "loss": 0.6778, "step": 26480 }, { "epoch": 0.9591090184715683, "grad_norm": 1.3791964073020146, "learning_rate": 8.757278501486555e-08, "loss": 0.5943, "step": 26481 }, { "epoch": 0.9591452372328866, "grad_norm": 1.2573585439366823, "learning_rate": 8.741794667315396e-08, "loss": 0.7188, "step": 26482 }, { "epoch": 0.959181455994205, "grad_norm": 1.2394938112704932, "learning_rate": 8.726324473728121e-08, "loss": 0.5918, "step": 26483 }, { "epoch": 0.9592176747555233, "grad_norm": 1.2810947226718281, "learning_rate": 8.710867920937671e-08, "loss": 0.6613, "step": 26484 }, { "epoch": 0.9592538935168418, "grad_norm": 0.9776536845627111, "learning_rate": 8.695425009156654e-08, "loss": 0.654, "step": 26485 }, { "epoch": 0.9592901122781601, "grad_norm": 1.1068160904836075, "learning_rate": 8.679995738597568e-08, "loss": 0.6581, "step": 26486 }, { "epoch": 0.9593263310394784, "grad_norm": 1.3729840836743725, "learning_rate": 8.664580109472908e-08, "loss": 0.6251, "step": 26487 }, { "epoch": 0.9593625498007968, "grad_norm": 1.2960168120385507, "learning_rate": 8.649178121994616e-08, "loss": 0.7118, "step": 26488 }, { "epoch": 0.9593987685621151, "grad_norm": 1.3708413087316698, "learning_rate": 8.633789776374745e-08, "loss": 0.6259, "step": 26489 }, { "epoch": 0.9594349873234336, "grad_norm": 1.30785802043577, "learning_rate": 8.618415072825014e-08, "loss": 0.6208, "step": 26490 }, { "epoch": 0.9594712060847519, "grad_norm": 1.3832028162745014, "learning_rate": 8.603054011557032e-08, "loss": 0.7125, "step": 26491 }, { "epoch": 0.9595074248460702, "grad_norm": 1.0980247482956644, "learning_rate": 8.587706592782075e-08, "loss": 0.6641, "step": 26492 }, { "epoch": 0.9595436436073886, "grad_norm": 1.2775313096772754, "learning_rate": 8.572372816711416e-08, "loss": 0.6691, "step": 26493 }, { "epoch": 0.959579862368707, "grad_norm": 1.1096815403304174, "learning_rate": 8.557052683556e-08, "loss": 0.6656, "step": 26494 }, { "epoch": 0.9596160811300254, "grad_norm": 1.2553414156182856, "learning_rate": 8.541746193526879e-08, "loss": 0.6124, "step": 26495 }, { "epoch": 0.9596522998913437, "grad_norm": 1.4050085018153573, "learning_rate": 8.526453346834329e-08, "loss": 0.6391, "step": 26496 }, { "epoch": 0.959688518652662, "grad_norm": 1.3719932704215625, "learning_rate": 8.51117414368896e-08, "loss": 0.6422, "step": 26497 }, { "epoch": 0.9597247374139805, "grad_norm": 1.3291932886495994, "learning_rate": 8.49590858430116e-08, "loss": 0.5938, "step": 26498 }, { "epoch": 0.9597609561752988, "grad_norm": 1.0318220540531005, "learning_rate": 8.480656668880871e-08, "loss": 0.7053, "step": 26499 }, { "epoch": 0.9597971749366172, "grad_norm": 1.3437196643587765, "learning_rate": 8.465418397637815e-08, "loss": 0.6922, "step": 26500 }, { "epoch": 0.9598333936979355, "grad_norm": 1.3579328906129, "learning_rate": 8.450193770781933e-08, "loss": 0.6744, "step": 26501 }, { "epoch": 0.9598696124592538, "grad_norm": 1.2683748546521592, "learning_rate": 8.434982788522617e-08, "loss": 0.6888, "step": 26502 }, { "epoch": 0.9599058312205723, "grad_norm": 1.33508672475668, "learning_rate": 8.41978545106925e-08, "loss": 0.7582, "step": 26503 }, { "epoch": 0.9599420499818906, "grad_norm": 1.0159259759959955, "learning_rate": 8.404601758630892e-08, "loss": 0.605, "step": 26504 }, { "epoch": 0.959978268743209, "grad_norm": 1.5043958251016483, "learning_rate": 8.389431711416484e-08, "loss": 0.6412, "step": 26505 }, { "epoch": 0.9600144875045273, "grad_norm": 1.3319253362011878, "learning_rate": 8.374275309634861e-08, "loss": 0.6689, "step": 26506 }, { "epoch": 0.9600507062658457, "grad_norm": 1.4270544533580172, "learning_rate": 8.35913255349452e-08, "loss": 0.717, "step": 26507 }, { "epoch": 0.9600869250271641, "grad_norm": 1.3923775299716912, "learning_rate": 8.34400344320374e-08, "loss": 0.6764, "step": 26508 }, { "epoch": 0.9601231437884824, "grad_norm": 1.4063611874069495, "learning_rate": 8.328887978970912e-08, "loss": 0.7149, "step": 26509 }, { "epoch": 0.9601593625498008, "grad_norm": 1.3277079740920161, "learning_rate": 8.313786161003978e-08, "loss": 0.6506, "step": 26510 }, { "epoch": 0.9601955813111192, "grad_norm": 1.5354962384985429, "learning_rate": 8.298697989510551e-08, "loss": 0.726, "step": 26511 }, { "epoch": 0.9602318000724375, "grad_norm": 1.3314015359188476, "learning_rate": 8.283623464698576e-08, "loss": 0.6765, "step": 26512 }, { "epoch": 0.9602680188337559, "grad_norm": 1.0180159080282998, "learning_rate": 8.268562586775331e-08, "loss": 0.688, "step": 26513 }, { "epoch": 0.9603042375950742, "grad_norm": 0.9945374804322301, "learning_rate": 8.253515355948094e-08, "loss": 0.673, "step": 26514 }, { "epoch": 0.9603404563563926, "grad_norm": 1.3505469997196533, "learning_rate": 8.238481772423812e-08, "loss": 0.6458, "step": 26515 }, { "epoch": 0.960376675117711, "grad_norm": 1.451290091144412, "learning_rate": 8.223461836409429e-08, "loss": 0.7292, "step": 26516 }, { "epoch": 0.9604128938790293, "grad_norm": 1.118674633947956, "learning_rate": 8.208455548111782e-08, "loss": 0.639, "step": 26517 }, { "epoch": 0.9604491126403477, "grad_norm": 1.0491451019144027, "learning_rate": 8.19346290773726e-08, "loss": 0.6748, "step": 26518 }, { "epoch": 0.960485331401666, "grad_norm": 1.6117398516280887, "learning_rate": 8.17848391549203e-08, "loss": 0.6961, "step": 26519 }, { "epoch": 0.9605215501629845, "grad_norm": 1.4440265192544912, "learning_rate": 8.163518571582486e-08, "loss": 0.7121, "step": 26520 }, { "epoch": 0.9605577689243028, "grad_norm": 1.3576371063463677, "learning_rate": 8.148566876214348e-08, "loss": 0.6645, "step": 26521 }, { "epoch": 0.9605939876856211, "grad_norm": 1.2915436412272978, "learning_rate": 8.133628829593454e-08, "loss": 0.6027, "step": 26522 }, { "epoch": 0.9606302064469395, "grad_norm": 1.4823565625766602, "learning_rate": 8.118704431925417e-08, "loss": 0.6823, "step": 26523 }, { "epoch": 0.9606664252082578, "grad_norm": 1.4101618987258249, "learning_rate": 8.103793683415517e-08, "loss": 0.6038, "step": 26524 }, { "epoch": 0.9607026439695763, "grad_norm": 0.9941472038262614, "learning_rate": 8.088896584269035e-08, "loss": 0.727, "step": 26525 }, { "epoch": 0.9607388627308946, "grad_norm": 1.5658190071183309, "learning_rate": 8.074013134690805e-08, "loss": 0.7149, "step": 26526 }, { "epoch": 0.9607750814922129, "grad_norm": 1.4587734815567568, "learning_rate": 8.059143334885778e-08, "loss": 0.7026, "step": 26527 }, { "epoch": 0.9608113002535313, "grad_norm": 0.9793518334053416, "learning_rate": 8.044287185058564e-08, "loss": 0.6429, "step": 26528 }, { "epoch": 0.9608475190148497, "grad_norm": 1.3798863153757395, "learning_rate": 8.029444685413445e-08, "loss": 0.67, "step": 26529 }, { "epoch": 0.9608837377761681, "grad_norm": 1.0165959353708054, "learning_rate": 8.014615836154926e-08, "loss": 0.655, "step": 26530 }, { "epoch": 0.9609199565374864, "grad_norm": 1.4690152953369553, "learning_rate": 7.999800637486954e-08, "loss": 0.6153, "step": 26531 }, { "epoch": 0.9609561752988047, "grad_norm": 1.432007259973345, "learning_rate": 7.984999089613254e-08, "loss": 0.678, "step": 26532 }, { "epoch": 0.9609923940601232, "grad_norm": 1.1123869550899887, "learning_rate": 7.970211192737664e-08, "loss": 0.7286, "step": 26533 }, { "epoch": 0.9610286128214415, "grad_norm": 1.4948279887836138, "learning_rate": 7.955436947063688e-08, "loss": 0.6926, "step": 26534 }, { "epoch": 0.9610648315827599, "grad_norm": 1.2969897015530565, "learning_rate": 7.940676352794496e-08, "loss": 0.583, "step": 26535 }, { "epoch": 0.9611010503440782, "grad_norm": 1.3558734569955766, "learning_rate": 7.92592941013337e-08, "loss": 0.6118, "step": 26536 }, { "epoch": 0.9611372691053965, "grad_norm": 1.3725676663484638, "learning_rate": 7.911196119283149e-08, "loss": 0.6385, "step": 26537 }, { "epoch": 0.961173487866715, "grad_norm": 1.0635912809683739, "learning_rate": 7.896476480446558e-08, "loss": 0.7167, "step": 26538 }, { "epoch": 0.9612097066280333, "grad_norm": 1.2647948382853678, "learning_rate": 7.881770493826324e-08, "loss": 0.6166, "step": 26539 }, { "epoch": 0.9612459253893517, "grad_norm": 2.978033036706439, "learning_rate": 7.867078159624508e-08, "loss": 0.6432, "step": 26540 }, { "epoch": 0.96128214415067, "grad_norm": 1.4108327071296598, "learning_rate": 7.852399478043505e-08, "loss": 0.628, "step": 26541 }, { "epoch": 0.9613183629119884, "grad_norm": 1.0206847952563438, "learning_rate": 7.837734449285372e-08, "loss": 0.6631, "step": 26542 }, { "epoch": 0.9613545816733068, "grad_norm": 1.3839511264938498, "learning_rate": 7.823083073551729e-08, "loss": 0.6886, "step": 26543 }, { "epoch": 0.9613908004346251, "grad_norm": 1.3563474107624398, "learning_rate": 7.808445351044192e-08, "loss": 0.6193, "step": 26544 }, { "epoch": 0.9614270191959435, "grad_norm": 1.4052858464778049, "learning_rate": 7.793821281964376e-08, "loss": 0.6643, "step": 26545 }, { "epoch": 0.9614632379572619, "grad_norm": 1.0298016392187699, "learning_rate": 7.779210866513342e-08, "loss": 0.596, "step": 26546 }, { "epoch": 0.9614994567185802, "grad_norm": 1.2686613917532816, "learning_rate": 7.764614104892265e-08, "loss": 0.7024, "step": 26547 }, { "epoch": 0.9615356754798986, "grad_norm": 0.8973019264667756, "learning_rate": 7.750030997301872e-08, "loss": 0.6516, "step": 26548 }, { "epoch": 0.9615718942412169, "grad_norm": 1.3530359016679199, "learning_rate": 7.735461543943001e-08, "loss": 0.6676, "step": 26549 }, { "epoch": 0.9616081130025353, "grad_norm": 1.0308078944024979, "learning_rate": 7.720905745015939e-08, "loss": 0.6844, "step": 26550 }, { "epoch": 0.9616443317638537, "grad_norm": 1.405576267047494, "learning_rate": 7.70636360072119e-08, "loss": 0.593, "step": 26551 }, { "epoch": 0.961680550525172, "grad_norm": 1.4531710749417714, "learning_rate": 7.691835111258817e-08, "loss": 0.659, "step": 26552 }, { "epoch": 0.9617167692864904, "grad_norm": 1.2752764623000068, "learning_rate": 7.67732027682866e-08, "loss": 0.6117, "step": 26553 }, { "epoch": 0.9617529880478087, "grad_norm": 1.2568216056273351, "learning_rate": 7.662819097630558e-08, "loss": 0.609, "step": 26554 }, { "epoch": 0.9617892068091272, "grad_norm": 1.3003986701017471, "learning_rate": 7.64833157386391e-08, "loss": 0.5756, "step": 26555 }, { "epoch": 0.9618254255704455, "grad_norm": 1.0643940182995462, "learning_rate": 7.633857705728331e-08, "loss": 0.6479, "step": 26556 }, { "epoch": 0.9618616443317638, "grad_norm": 1.3648641340767542, "learning_rate": 7.619397493422665e-08, "loss": 0.6848, "step": 26557 }, { "epoch": 0.9618978630930822, "grad_norm": 1.519392872665622, "learning_rate": 7.604950937146193e-08, "loss": 0.755, "step": 26558 }, { "epoch": 0.9619340818544005, "grad_norm": 1.3054114891337776, "learning_rate": 7.590518037097539e-08, "loss": 0.6082, "step": 26559 }, { "epoch": 0.961970300615719, "grad_norm": 1.2354495092621847, "learning_rate": 7.576098793475428e-08, "loss": 0.7152, "step": 26560 }, { "epoch": 0.9620065193770373, "grad_norm": 1.3144462345787253, "learning_rate": 7.561693206478149e-08, "loss": 0.6466, "step": 26561 }, { "epoch": 0.9620427381383556, "grad_norm": 1.3156982634797552, "learning_rate": 7.547301276303987e-08, "loss": 0.6384, "step": 26562 }, { "epoch": 0.962078956899674, "grad_norm": 1.4132810983821182, "learning_rate": 7.532923003151005e-08, "loss": 0.6563, "step": 26563 }, { "epoch": 0.9621151756609924, "grad_norm": 1.4089844681465071, "learning_rate": 7.518558387217157e-08, "loss": 0.7055, "step": 26564 }, { "epoch": 0.9621513944223108, "grad_norm": 1.5147591986331241, "learning_rate": 7.50420742869995e-08, "loss": 0.6434, "step": 26565 }, { "epoch": 0.9621876131836291, "grad_norm": 1.2975613231490986, "learning_rate": 7.489870127796894e-08, "loss": 0.5935, "step": 26566 }, { "epoch": 0.9622238319449474, "grad_norm": 1.3500402904001712, "learning_rate": 7.475546484705387e-08, "loss": 0.696, "step": 26567 }, { "epoch": 0.9622600507062659, "grad_norm": 1.2619790393329682, "learning_rate": 7.461236499622493e-08, "loss": 0.6077, "step": 26568 }, { "epoch": 0.9622962694675842, "grad_norm": 1.4050555454589568, "learning_rate": 7.446940172745055e-08, "loss": 0.6249, "step": 26569 }, { "epoch": 0.9623324882289026, "grad_norm": 1.4373620397252336, "learning_rate": 7.432657504269803e-08, "loss": 0.5827, "step": 26570 }, { "epoch": 0.9623687069902209, "grad_norm": 0.9377993089124327, "learning_rate": 7.418388494393358e-08, "loss": 0.6906, "step": 26571 }, { "epoch": 0.9624049257515392, "grad_norm": 1.0347978972550291, "learning_rate": 7.40413314331212e-08, "loss": 0.6482, "step": 26572 }, { "epoch": 0.9624411445128577, "grad_norm": 1.0254071572651315, "learning_rate": 7.389891451222153e-08, "loss": 0.6155, "step": 26573 }, { "epoch": 0.962477363274176, "grad_norm": 1.3858608721931396, "learning_rate": 7.375663418319411e-08, "loss": 0.678, "step": 26574 }, { "epoch": 0.9625135820354944, "grad_norm": 1.195081451624851, "learning_rate": 7.36144904479974e-08, "loss": 0.5787, "step": 26575 }, { "epoch": 0.9625498007968127, "grad_norm": 1.4879462836369195, "learning_rate": 7.347248330858869e-08, "loss": 0.6589, "step": 26576 }, { "epoch": 0.9625860195581312, "grad_norm": 1.4719743117310051, "learning_rate": 7.333061276691866e-08, "loss": 0.6421, "step": 26577 }, { "epoch": 0.9626222383194495, "grad_norm": 1.0663459123178267, "learning_rate": 7.318887882494352e-08, "loss": 0.6317, "step": 26578 }, { "epoch": 0.9626584570807678, "grad_norm": 1.5408465597291612, "learning_rate": 7.304728148461171e-08, "loss": 0.6667, "step": 26579 }, { "epoch": 0.9626946758420862, "grad_norm": 0.9746642863821599, "learning_rate": 7.290582074787056e-08, "loss": 0.7194, "step": 26580 }, { "epoch": 0.9627308946034046, "grad_norm": 1.311188032162889, "learning_rate": 7.276449661666851e-08, "loss": 0.5962, "step": 26581 }, { "epoch": 0.962767113364723, "grad_norm": 1.3407246895218343, "learning_rate": 7.262330909295068e-08, "loss": 0.6143, "step": 26582 }, { "epoch": 0.9628033321260413, "grad_norm": 1.2909627822818748, "learning_rate": 7.248225817865883e-08, "loss": 0.5669, "step": 26583 }, { "epoch": 0.9628395508873596, "grad_norm": 1.3880750732458322, "learning_rate": 7.234134387573367e-08, "loss": 0.6489, "step": 26584 }, { "epoch": 0.962875769648678, "grad_norm": 1.4642277706748872, "learning_rate": 7.220056618611471e-08, "loss": 0.6308, "step": 26585 }, { "epoch": 0.9629119884099964, "grad_norm": 1.5975755074204576, "learning_rate": 7.205992511174043e-08, "loss": 0.6484, "step": 26586 }, { "epoch": 0.9629482071713148, "grad_norm": 1.4288979862239057, "learning_rate": 7.191942065454372e-08, "loss": 0.6981, "step": 26587 }, { "epoch": 0.9629844259326331, "grad_norm": 1.0174417696386517, "learning_rate": 7.177905281645969e-08, "loss": 0.677, "step": 26588 }, { "epoch": 0.9630206446939514, "grad_norm": 1.3441602737622484, "learning_rate": 7.163882159942015e-08, "loss": 0.667, "step": 26589 }, { "epoch": 0.9630568634552699, "grad_norm": 2.39981841273344, "learning_rate": 7.149872700535465e-08, "loss": 0.7074, "step": 26590 }, { "epoch": 0.9630930822165882, "grad_norm": 0.9751961215840713, "learning_rate": 7.135876903618944e-08, "loss": 0.6402, "step": 26591 }, { "epoch": 0.9631293009779066, "grad_norm": 1.4182199670710627, "learning_rate": 7.121894769385185e-08, "loss": 0.6115, "step": 26592 }, { "epoch": 0.9631655197392249, "grad_norm": 1.2950475158555887, "learning_rate": 7.10792629802659e-08, "loss": 0.6614, "step": 26593 }, { "epoch": 0.9632017385005432, "grad_norm": 1.034879975631076, "learning_rate": 7.093971489735341e-08, "loss": 0.6757, "step": 26594 }, { "epoch": 0.9632379572618617, "grad_norm": 1.1054043011783505, "learning_rate": 7.080030344703503e-08, "loss": 0.7107, "step": 26595 }, { "epoch": 0.96327417602318, "grad_norm": 1.3349089713540951, "learning_rate": 7.066102863122925e-08, "loss": 0.6566, "step": 26596 }, { "epoch": 0.9633103947844984, "grad_norm": 1.1966077528961139, "learning_rate": 7.052189045185232e-08, "loss": 0.6812, "step": 26597 }, { "epoch": 0.9633466135458167, "grad_norm": 0.960455269435699, "learning_rate": 7.038288891081935e-08, "loss": 0.6455, "step": 26598 }, { "epoch": 0.9633828323071351, "grad_norm": 0.888523904630208, "learning_rate": 7.024402401004216e-08, "loss": 0.639, "step": 26599 }, { "epoch": 0.9634190510684535, "grad_norm": 1.3534622897474216, "learning_rate": 7.010529575143255e-08, "loss": 0.6159, "step": 26600 }, { "epoch": 0.9634552698297718, "grad_norm": 1.4506013346808055, "learning_rate": 6.996670413689899e-08, "loss": 0.6446, "step": 26601 }, { "epoch": 0.9634914885910902, "grad_norm": 1.4372233216542556, "learning_rate": 6.982824916834886e-08, "loss": 0.6804, "step": 26602 }, { "epoch": 0.9635277073524086, "grad_norm": 1.44002528639971, "learning_rate": 6.968993084768728e-08, "loss": 0.6197, "step": 26603 }, { "epoch": 0.9635639261137269, "grad_norm": 0.9611432313330133, "learning_rate": 6.955174917681829e-08, "loss": 0.6952, "step": 26604 }, { "epoch": 0.9636001448750453, "grad_norm": 1.2970211825653888, "learning_rate": 6.941370415764259e-08, "loss": 0.646, "step": 26605 }, { "epoch": 0.9636363636363636, "grad_norm": 1.3283232400707936, "learning_rate": 6.927579579205978e-08, "loss": 0.6828, "step": 26606 }, { "epoch": 0.963672582397682, "grad_norm": 1.4247327783804762, "learning_rate": 6.913802408196835e-08, "loss": 0.7136, "step": 26607 }, { "epoch": 0.9637088011590004, "grad_norm": 1.427307553033315, "learning_rate": 6.900038902926454e-08, "loss": 0.7374, "step": 26608 }, { "epoch": 0.9637450199203187, "grad_norm": 1.586089638403211, "learning_rate": 6.886289063584017e-08, "loss": 0.673, "step": 26609 }, { "epoch": 0.9637812386816371, "grad_norm": 1.292558985430391, "learning_rate": 6.87255289035893e-08, "loss": 0.6599, "step": 26610 }, { "epoch": 0.9638174574429554, "grad_norm": 1.4564645345185987, "learning_rate": 6.858830383440152e-08, "loss": 0.7144, "step": 26611 }, { "epoch": 0.9638536762042739, "grad_norm": 1.4270141424828335, "learning_rate": 6.84512154301653e-08, "loss": 0.671, "step": 26612 }, { "epoch": 0.9638898949655922, "grad_norm": 1.4964671801856515, "learning_rate": 6.831426369276584e-08, "loss": 0.6638, "step": 26613 }, { "epoch": 0.9639261137269105, "grad_norm": 1.4274321123937463, "learning_rate": 6.817744862409049e-08, "loss": 0.5878, "step": 26614 }, { "epoch": 0.9639623324882289, "grad_norm": 1.466625609983719, "learning_rate": 6.804077022601996e-08, "loss": 0.7045, "step": 26615 }, { "epoch": 0.9639985512495473, "grad_norm": 1.3618081114124156, "learning_rate": 6.790422850043498e-08, "loss": 0.6834, "step": 26616 }, { "epoch": 0.9640347700108657, "grad_norm": 1.3531881998755482, "learning_rate": 6.776782344921407e-08, "loss": 0.6743, "step": 26617 }, { "epoch": 0.964070988772184, "grad_norm": 1.3329495312452548, "learning_rate": 6.763155507423679e-08, "loss": 0.6633, "step": 26618 }, { "epoch": 0.9641072075335023, "grad_norm": 0.9497493745489191, "learning_rate": 6.749542337737614e-08, "loss": 0.7121, "step": 26619 }, { "epoch": 0.9641434262948207, "grad_norm": 1.35737202161726, "learning_rate": 6.735942836050613e-08, "loss": 0.6689, "step": 26620 }, { "epoch": 0.9641796450561391, "grad_norm": 1.382999975217009, "learning_rate": 6.72235700254964e-08, "loss": 0.6527, "step": 26621 }, { "epoch": 0.9642158638174575, "grad_norm": 1.5766472223766756, "learning_rate": 6.708784837421989e-08, "loss": 0.7068, "step": 26622 }, { "epoch": 0.9642520825787758, "grad_norm": 1.2974641256079325, "learning_rate": 6.695226340854177e-08, "loss": 0.608, "step": 26623 }, { "epoch": 0.9642883013400941, "grad_norm": 1.420448829856124, "learning_rate": 6.681681513032834e-08, "loss": 0.574, "step": 26624 }, { "epoch": 0.9643245201014126, "grad_norm": 1.3418226288179924, "learning_rate": 6.668150354144365e-08, "loss": 0.6603, "step": 26625 }, { "epoch": 0.9643607388627309, "grad_norm": 1.1058193688473001, "learning_rate": 6.654632864374954e-08, "loss": 0.6708, "step": 26626 }, { "epoch": 0.9643969576240493, "grad_norm": 1.0408944742217263, "learning_rate": 6.641129043910677e-08, "loss": 0.6769, "step": 26627 }, { "epoch": 0.9644331763853676, "grad_norm": 1.4262313014897645, "learning_rate": 6.62763889293716e-08, "loss": 0.6061, "step": 26628 }, { "epoch": 0.964469395146686, "grad_norm": 1.2961464294617802, "learning_rate": 6.614162411640368e-08, "loss": 0.6057, "step": 26629 }, { "epoch": 0.9645056139080044, "grad_norm": 1.4073602538762697, "learning_rate": 6.600699600205485e-08, "loss": 0.7051, "step": 26630 }, { "epoch": 0.9645418326693227, "grad_norm": 1.3886779109002774, "learning_rate": 6.587250458817806e-08, "loss": 0.6225, "step": 26631 }, { "epoch": 0.9645780514306411, "grad_norm": 1.1620038791063114, "learning_rate": 6.57381498766252e-08, "loss": 0.6237, "step": 26632 }, { "epoch": 0.9646142701919594, "grad_norm": 1.2401678563099117, "learning_rate": 6.560393186924363e-08, "loss": 0.6837, "step": 26633 }, { "epoch": 0.9646504889532778, "grad_norm": 1.4188483743969693, "learning_rate": 6.546985056788191e-08, "loss": 0.6157, "step": 26634 }, { "epoch": 0.9646867077145962, "grad_norm": 1.32316054412547, "learning_rate": 6.533590597438411e-08, "loss": 0.6368, "step": 26635 }, { "epoch": 0.9647229264759145, "grad_norm": 1.3903521400010825, "learning_rate": 6.520209809059319e-08, "loss": 0.6747, "step": 26636 }, { "epoch": 0.9647591452372329, "grad_norm": 0.9698462800128839, "learning_rate": 6.506842691835214e-08, "loss": 0.657, "step": 26637 }, { "epoch": 0.9647953639985513, "grad_norm": 1.0479186230231359, "learning_rate": 6.493489245949835e-08, "loss": 0.6553, "step": 26638 }, { "epoch": 0.9648315827598696, "grad_norm": 1.4769725608606217, "learning_rate": 6.480149471586928e-08, "loss": 0.6445, "step": 26639 }, { "epoch": 0.964867801521188, "grad_norm": 1.3352743902611186, "learning_rate": 6.46682336893023e-08, "loss": 0.675, "step": 26640 }, { "epoch": 0.9649040202825063, "grad_norm": 1.2373062726908632, "learning_rate": 6.453510938163043e-08, "loss": 0.6359, "step": 26641 }, { "epoch": 0.9649402390438248, "grad_norm": 1.4723335223635663, "learning_rate": 6.440212179468553e-08, "loss": 0.6547, "step": 26642 }, { "epoch": 0.9649764578051431, "grad_norm": 1.592854985968443, "learning_rate": 6.426927093029722e-08, "loss": 0.6274, "step": 26643 }, { "epoch": 0.9650126765664614, "grad_norm": 1.1302358907951588, "learning_rate": 6.413655679029407e-08, "loss": 0.7111, "step": 26644 }, { "epoch": 0.9650488953277798, "grad_norm": 1.4803089241780278, "learning_rate": 6.400397937650238e-08, "loss": 0.6608, "step": 26645 }, { "epoch": 0.9650851140890981, "grad_norm": 1.4950126222688425, "learning_rate": 6.387153869074625e-08, "loss": 0.6317, "step": 26646 }, { "epoch": 0.9651213328504166, "grad_norm": 1.4184660118462178, "learning_rate": 6.373923473484866e-08, "loss": 0.7083, "step": 26647 }, { "epoch": 0.9651575516117349, "grad_norm": 1.4433607422766708, "learning_rate": 6.36070675106304e-08, "loss": 0.6833, "step": 26648 }, { "epoch": 0.9651937703730532, "grad_norm": 1.5112142669844575, "learning_rate": 6.347503701990998e-08, "loss": 0.651, "step": 26649 }, { "epoch": 0.9652299891343716, "grad_norm": 1.46283905692178, "learning_rate": 6.334314326450264e-08, "loss": 0.6765, "step": 26650 }, { "epoch": 0.96526620789569, "grad_norm": 1.5083315590842317, "learning_rate": 6.32113862462258e-08, "loss": 0.6764, "step": 26651 }, { "epoch": 0.9653024266570084, "grad_norm": 1.4278989010845629, "learning_rate": 6.307976596689137e-08, "loss": 0.6841, "step": 26652 }, { "epoch": 0.9653386454183267, "grad_norm": 1.4781606070330748, "learning_rate": 6.294828242831008e-08, "loss": 0.69, "step": 26653 }, { "epoch": 0.965374864179645, "grad_norm": 1.3501664342486217, "learning_rate": 6.281693563229275e-08, "loss": 0.6367, "step": 26654 }, { "epoch": 0.9654110829409634, "grad_norm": 1.461634915698708, "learning_rate": 6.268572558064567e-08, "loss": 0.7466, "step": 26655 }, { "epoch": 0.9654473017022818, "grad_norm": 1.5270760147869995, "learning_rate": 6.255465227517521e-08, "loss": 0.649, "step": 26656 }, { "epoch": 0.9654835204636002, "grad_norm": 1.2651391229934257, "learning_rate": 6.242371571768324e-08, "loss": 0.6394, "step": 26657 }, { "epoch": 0.9655197392249185, "grad_norm": 1.231670997262077, "learning_rate": 6.229291590997388e-08, "loss": 0.5743, "step": 26658 }, { "epoch": 0.9655559579862368, "grad_norm": 1.4698696833531089, "learning_rate": 6.21622528538457e-08, "loss": 0.6904, "step": 26659 }, { "epoch": 0.9655921767475553, "grad_norm": 1.4131782016541883, "learning_rate": 6.203172655109723e-08, "loss": 0.6308, "step": 26660 }, { "epoch": 0.9656283955088736, "grad_norm": 1.4731544788898432, "learning_rate": 6.190133700352486e-08, "loss": 0.6675, "step": 26661 }, { "epoch": 0.965664614270192, "grad_norm": 1.1340662890844253, "learning_rate": 6.177108421292266e-08, "loss": 0.6623, "step": 26662 }, { "epoch": 0.9657008330315103, "grad_norm": 1.409732810505543, "learning_rate": 6.164096818108145e-08, "loss": 0.6555, "step": 26663 }, { "epoch": 0.9657370517928286, "grad_norm": 1.4323835656999124, "learning_rate": 6.151098890979423e-08, "loss": 0.6961, "step": 26664 }, { "epoch": 0.9657732705541471, "grad_norm": 1.3122455734914376, "learning_rate": 6.138114640084847e-08, "loss": 0.6756, "step": 26665 }, { "epoch": 0.9658094893154654, "grad_norm": 1.413654904871093, "learning_rate": 6.12514406560305e-08, "loss": 0.6051, "step": 26666 }, { "epoch": 0.9658457080767838, "grad_norm": 1.2897116063546519, "learning_rate": 6.112187167712558e-08, "loss": 0.6281, "step": 26667 }, { "epoch": 0.9658819268381021, "grad_norm": 1.3215039509161401, "learning_rate": 6.09924394659167e-08, "loss": 0.6407, "step": 26668 }, { "epoch": 0.9659181455994205, "grad_norm": 0.9423786032908821, "learning_rate": 6.086314402418581e-08, "loss": 0.6244, "step": 26669 }, { "epoch": 0.9659543643607389, "grad_norm": 1.5141484248725727, "learning_rate": 6.073398535371034e-08, "loss": 0.6857, "step": 26670 }, { "epoch": 0.9659905831220572, "grad_norm": 1.3603802682732864, "learning_rate": 6.060496345626887e-08, "loss": 0.6554, "step": 26671 }, { "epoch": 0.9660268018833756, "grad_norm": 1.1650008066815964, "learning_rate": 6.047607833363778e-08, "loss": 0.6923, "step": 26672 }, { "epoch": 0.966063020644694, "grad_norm": 1.3989479662349402, "learning_rate": 6.034732998758785e-08, "loss": 0.6389, "step": 26673 }, { "epoch": 0.9660992394060123, "grad_norm": 1.160543377358664, "learning_rate": 6.021871841989324e-08, "loss": 0.7115, "step": 26674 }, { "epoch": 0.9661354581673307, "grad_norm": 1.289348291736689, "learning_rate": 6.009024363232252e-08, "loss": 0.6337, "step": 26675 }, { "epoch": 0.966171676928649, "grad_norm": 1.4938440888791114, "learning_rate": 5.99619056266454e-08, "loss": 0.7247, "step": 26676 }, { "epoch": 0.9662078956899675, "grad_norm": 1.2378399503320259, "learning_rate": 5.983370440462488e-08, "loss": 0.6641, "step": 26677 }, { "epoch": 0.9662441144512858, "grad_norm": 0.9640091850462018, "learning_rate": 5.970563996802736e-08, "loss": 0.695, "step": 26678 }, { "epoch": 0.9662803332126041, "grad_norm": 1.6949674732064397, "learning_rate": 5.957771231861476e-08, "loss": 0.6712, "step": 26679 }, { "epoch": 0.9663165519739225, "grad_norm": 1.3685015004295789, "learning_rate": 5.9449921458146765e-08, "loss": 0.6778, "step": 26680 }, { "epoch": 0.9663527707352408, "grad_norm": 1.4210398672446731, "learning_rate": 5.932226738838309e-08, "loss": 0.7091, "step": 26681 }, { "epoch": 0.9663889894965593, "grad_norm": 1.064529868381241, "learning_rate": 5.9194750111079e-08, "loss": 0.7434, "step": 26682 }, { "epoch": 0.9664252082578776, "grad_norm": 1.5403800833516907, "learning_rate": 5.9067369627989755e-08, "loss": 0.6235, "step": 26683 }, { "epoch": 0.9664614270191959, "grad_norm": 1.2594375552811166, "learning_rate": 5.894012594086951e-08, "loss": 0.5764, "step": 26684 }, { "epoch": 0.9664976457805143, "grad_norm": 1.3225593064669674, "learning_rate": 5.8813019051466855e-08, "loss": 0.6669, "step": 26685 }, { "epoch": 0.9665338645418327, "grad_norm": 1.484018192573684, "learning_rate": 5.8686048961532627e-08, "loss": 0.7472, "step": 26686 }, { "epoch": 0.9665700833031511, "grad_norm": 1.1014510770086725, "learning_rate": 5.85592156728132e-08, "loss": 0.6768, "step": 26687 }, { "epoch": 0.9666063020644694, "grad_norm": 0.9471019550324746, "learning_rate": 5.843251918705495e-08, "loss": 0.6608, "step": 26688 }, { "epoch": 0.9666425208257877, "grad_norm": 1.097197890881657, "learning_rate": 5.830595950600093e-08, "loss": 0.6388, "step": 26689 }, { "epoch": 0.9666787395871062, "grad_norm": 1.0390548599336942, "learning_rate": 5.817953663139198e-08, "loss": 0.6828, "step": 26690 }, { "epoch": 0.9667149583484245, "grad_norm": 1.330073334913549, "learning_rate": 5.8053250564967803e-08, "loss": 0.6769, "step": 26691 }, { "epoch": 0.9667511771097429, "grad_norm": 1.0536806654735789, "learning_rate": 5.7927101308467015e-08, "loss": 0.736, "step": 26692 }, { "epoch": 0.9667873958710612, "grad_norm": 1.360492442956392, "learning_rate": 5.78010888636249e-08, "loss": 0.6477, "step": 26693 }, { "epoch": 0.9668236146323795, "grad_norm": 0.962147341073786, "learning_rate": 5.767521323217673e-08, "loss": 0.645, "step": 26694 }, { "epoch": 0.966859833393698, "grad_norm": 1.359991944148725, "learning_rate": 5.754947441585335e-08, "loss": 0.7069, "step": 26695 }, { "epoch": 0.9668960521550163, "grad_norm": 1.3630157057694001, "learning_rate": 5.7423872416384474e-08, "loss": 0.6313, "step": 26696 }, { "epoch": 0.9669322709163347, "grad_norm": 1.5553774213395728, "learning_rate": 5.7298407235499844e-08, "loss": 0.7306, "step": 26697 }, { "epoch": 0.966968489677653, "grad_norm": 1.1854908526749195, "learning_rate": 5.7173078874925845e-08, "loss": 0.6494, "step": 26698 }, { "epoch": 0.9670047084389713, "grad_norm": 1.3813096543846266, "learning_rate": 5.7047887336386664e-08, "loss": 0.6173, "step": 26699 }, { "epoch": 0.9670409272002898, "grad_norm": 0.9998003938092386, "learning_rate": 5.6922832621605364e-08, "loss": 0.6672, "step": 26700 }, { "epoch": 0.9670771459616081, "grad_norm": 1.4716285058774203, "learning_rate": 5.679791473230278e-08, "loss": 0.6712, "step": 26701 }, { "epoch": 0.9671133647229265, "grad_norm": 1.5092843776601914, "learning_rate": 5.6673133670197555e-08, "loss": 0.6461, "step": 26702 }, { "epoch": 0.9671495834842448, "grad_norm": 1.4725436406628925, "learning_rate": 5.654848943700719e-08, "loss": 0.7165, "step": 26703 }, { "epoch": 0.9671858022455632, "grad_norm": 1.4629428119543835, "learning_rate": 5.6423982034445876e-08, "loss": 0.6819, "step": 26704 }, { "epoch": 0.9672220210068816, "grad_norm": 1.5004484925175778, "learning_rate": 5.629961146422891e-08, "loss": 0.646, "step": 26705 }, { "epoch": 0.9672582397681999, "grad_norm": 1.4989792559446793, "learning_rate": 5.617537772806603e-08, "loss": 0.7163, "step": 26706 }, { "epoch": 0.9672944585295183, "grad_norm": 1.3819264345224214, "learning_rate": 5.6051280827666975e-08, "loss": 0.6949, "step": 26707 }, { "epoch": 0.9673306772908367, "grad_norm": 1.360585709443927, "learning_rate": 5.592732076474039e-08, "loss": 0.6098, "step": 26708 }, { "epoch": 0.967366896052155, "grad_norm": 1.2565935667391257, "learning_rate": 5.580349754099157e-08, "loss": 0.5749, "step": 26709 }, { "epoch": 0.9674031148134734, "grad_norm": 1.1306774183156796, "learning_rate": 5.5679811158123597e-08, "loss": 0.6483, "step": 26710 }, { "epoch": 0.9674393335747917, "grad_norm": 1.6047847554200485, "learning_rate": 5.5556261617839555e-08, "loss": 0.7177, "step": 26711 }, { "epoch": 0.9674755523361102, "grad_norm": 1.507631067987249, "learning_rate": 5.543284892184031e-08, "loss": 0.6579, "step": 26712 }, { "epoch": 0.9675117710974285, "grad_norm": 1.4141674503348785, "learning_rate": 5.5309573071822276e-08, "loss": 0.656, "step": 26713 }, { "epoch": 0.9675479898587468, "grad_norm": 1.516803428346917, "learning_rate": 5.518643406948299e-08, "loss": 0.6725, "step": 26714 }, { "epoch": 0.9675842086200652, "grad_norm": 1.3215465675353375, "learning_rate": 5.506343191651553e-08, "loss": 0.6754, "step": 26715 }, { "epoch": 0.9676204273813835, "grad_norm": 2.0339093040061824, "learning_rate": 5.494056661461522e-08, "loss": 0.7526, "step": 26716 }, { "epoch": 0.967656646142702, "grad_norm": 1.3703948091093276, "learning_rate": 5.48178381654707e-08, "loss": 0.716, "step": 26717 }, { "epoch": 0.9676928649040203, "grad_norm": 1.884708227756139, "learning_rate": 5.469524657077063e-08, "loss": 0.5979, "step": 26718 }, { "epoch": 0.9677290836653386, "grad_norm": 1.1816507726225702, "learning_rate": 5.4572791832203655e-08, "loss": 0.6661, "step": 26719 }, { "epoch": 0.967765302426657, "grad_norm": 1.4561231048290681, "learning_rate": 5.445047395145286e-08, "loss": 0.6751, "step": 26720 }, { "epoch": 0.9678015211879754, "grad_norm": 0.9801206283596167, "learning_rate": 5.4328292930203584e-08, "loss": 0.6652, "step": 26721 }, { "epoch": 0.9678377399492938, "grad_norm": 1.4345884588144289, "learning_rate": 5.4206248770134474e-08, "loss": 0.7664, "step": 26722 }, { "epoch": 0.9678739587106121, "grad_norm": 1.3021229319908951, "learning_rate": 5.4084341472927515e-08, "loss": 0.6695, "step": 26723 }, { "epoch": 0.9679101774719304, "grad_norm": 1.3020309532279177, "learning_rate": 5.396257104026026e-08, "loss": 0.6703, "step": 26724 }, { "epoch": 0.9679463962332489, "grad_norm": 1.4230005772724592, "learning_rate": 5.384093747380581e-08, "loss": 0.665, "step": 26725 }, { "epoch": 0.9679826149945672, "grad_norm": 1.4410308895765815, "learning_rate": 5.3719440775240606e-08, "loss": 0.6984, "step": 26726 }, { "epoch": 0.9680188337558856, "grad_norm": 1.1191144827528094, "learning_rate": 5.359808094623442e-08, "loss": 0.7321, "step": 26727 }, { "epoch": 0.9680550525172039, "grad_norm": 1.3235458982678117, "learning_rate": 5.347685798845925e-08, "loss": 0.6242, "step": 26728 }, { "epoch": 0.9680912712785222, "grad_norm": 1.1721383249945185, "learning_rate": 5.335577190358154e-08, "loss": 0.6078, "step": 26729 }, { "epoch": 0.9681274900398407, "grad_norm": 1.0093182528154079, "learning_rate": 5.323482269326885e-08, "loss": 0.7184, "step": 26730 }, { "epoch": 0.968163708801159, "grad_norm": 1.3347352610822114, "learning_rate": 5.3114010359184286e-08, "loss": 0.6358, "step": 26731 }, { "epoch": 0.9681999275624774, "grad_norm": 1.5631497975071336, "learning_rate": 5.299333490299208e-08, "loss": 0.6382, "step": 26732 }, { "epoch": 0.9682361463237957, "grad_norm": 1.3580988913471637, "learning_rate": 5.2872796326350896e-08, "loss": 0.6479, "step": 26733 }, { "epoch": 0.968272365085114, "grad_norm": 1.0862105918857765, "learning_rate": 5.2752394630920526e-08, "loss": 0.6964, "step": 26734 }, { "epoch": 0.9683085838464325, "grad_norm": 1.6407425111738931, "learning_rate": 5.263212981835852e-08, "loss": 0.6483, "step": 26735 }, { "epoch": 0.9683448026077508, "grad_norm": 1.3279468690400564, "learning_rate": 5.251200189031802e-08, "loss": 0.6341, "step": 26736 }, { "epoch": 0.9683810213690692, "grad_norm": 1.401451146205917, "learning_rate": 5.239201084845324e-08, "loss": 0.6222, "step": 26737 }, { "epoch": 0.9684172401303875, "grad_norm": 1.2374591339593342, "learning_rate": 5.2272156694415097e-08, "loss": 0.6364, "step": 26738 }, { "epoch": 0.9684534588917059, "grad_norm": 1.2789160518352183, "learning_rate": 5.215243942985227e-08, "loss": 0.5918, "step": 26739 }, { "epoch": 0.9684896776530243, "grad_norm": 1.1375802453119526, "learning_rate": 5.203285905641342e-08, "loss": 0.6881, "step": 26740 }, { "epoch": 0.9685258964143426, "grad_norm": 1.1530546336370462, "learning_rate": 5.191341557574392e-08, "loss": 0.6618, "step": 26741 }, { "epoch": 0.968562115175661, "grad_norm": 0.98364225955299, "learning_rate": 5.179410898948689e-08, "loss": 0.717, "step": 26742 }, { "epoch": 0.9685983339369794, "grad_norm": 1.4296378702919528, "learning_rate": 5.1674939299283245e-08, "loss": 0.6221, "step": 26743 }, { "epoch": 0.9686345526982977, "grad_norm": 1.4053529341786666, "learning_rate": 5.1555906506775e-08, "loss": 0.6005, "step": 26744 }, { "epoch": 0.9686707714596161, "grad_norm": 0.8940745595976411, "learning_rate": 5.143701061359752e-08, "loss": 0.7266, "step": 26745 }, { "epoch": 0.9687069902209344, "grad_norm": 1.0325205584089532, "learning_rate": 5.131825162138948e-08, "loss": 0.6766, "step": 26746 }, { "epoch": 0.9687432089822529, "grad_norm": 1.4254369555173019, "learning_rate": 5.119962953178403e-08, "loss": 0.6882, "step": 26747 }, { "epoch": 0.9687794277435712, "grad_norm": 1.533462879020645, "learning_rate": 5.10811443464132e-08, "loss": 0.6636, "step": 26748 }, { "epoch": 0.9688156465048895, "grad_norm": 1.2116756004801494, "learning_rate": 5.0962796066907906e-08, "loss": 0.6044, "step": 26749 }, { "epoch": 0.9688518652662079, "grad_norm": 1.2891670358059746, "learning_rate": 5.084458469489573e-08, "loss": 0.5624, "step": 26750 }, { "epoch": 0.9688880840275262, "grad_norm": 1.3178439489101106, "learning_rate": 5.0726510232004254e-08, "loss": 0.6305, "step": 26751 }, { "epoch": 0.9689243027888447, "grad_norm": 1.1281279946918843, "learning_rate": 5.060857267985886e-08, "loss": 0.6794, "step": 26752 }, { "epoch": 0.968960521550163, "grad_norm": 1.4120627088209332, "learning_rate": 5.049077204008157e-08, "loss": 0.6597, "step": 26753 }, { "epoch": 0.9689967403114813, "grad_norm": 2.369998330844269, "learning_rate": 5.0373108314294426e-08, "loss": 0.6958, "step": 26754 }, { "epoch": 0.9690329590727997, "grad_norm": 1.2719458879227803, "learning_rate": 5.0255581504113914e-08, "loss": 0.6627, "step": 26755 }, { "epoch": 0.9690691778341181, "grad_norm": 1.4630155802838747, "learning_rate": 5.013819161116096e-08, "loss": 0.6702, "step": 26756 }, { "epoch": 0.9691053965954365, "grad_norm": 1.2729567559373582, "learning_rate": 5.002093863704871e-08, "loss": 0.5797, "step": 26757 }, { "epoch": 0.9691416153567548, "grad_norm": 1.2127074276868757, "learning_rate": 4.990382258339033e-08, "loss": 0.6568, "step": 26758 }, { "epoch": 0.9691778341180731, "grad_norm": 1.0958126153868692, "learning_rate": 4.9786843451798964e-08, "loss": 0.6995, "step": 26759 }, { "epoch": 0.9692140528793916, "grad_norm": 1.3656383211503114, "learning_rate": 4.9670001243884434e-08, "loss": 0.6125, "step": 26760 }, { "epoch": 0.9692502716407099, "grad_norm": 1.36893351848179, "learning_rate": 4.955329596125325e-08, "loss": 0.6371, "step": 26761 }, { "epoch": 0.9692864904020283, "grad_norm": 1.2453047225941423, "learning_rate": 4.943672760551077e-08, "loss": 0.6342, "step": 26762 }, { "epoch": 0.9693227091633466, "grad_norm": 1.1730447999020543, "learning_rate": 4.9320296178263506e-08, "loss": 0.6898, "step": 26763 }, { "epoch": 0.9693589279246649, "grad_norm": 1.3522502666592753, "learning_rate": 4.920400168111239e-08, "loss": 0.5662, "step": 26764 }, { "epoch": 0.9693951466859834, "grad_norm": 1.3356283092190024, "learning_rate": 4.908784411565726e-08, "loss": 0.6618, "step": 26765 }, { "epoch": 0.9694313654473017, "grad_norm": 1.3472867638867072, "learning_rate": 4.8971823483496825e-08, "loss": 0.6131, "step": 26766 }, { "epoch": 0.9694675842086201, "grad_norm": 1.4299756903536953, "learning_rate": 4.8855939786228715e-08, "loss": 0.6511, "step": 26767 }, { "epoch": 0.9695038029699384, "grad_norm": 1.4199495129210806, "learning_rate": 4.874019302544608e-08, "loss": 0.6763, "step": 26768 }, { "epoch": 0.9695400217312568, "grad_norm": 1.2020324263502784, "learning_rate": 4.8624583202742104e-08, "loss": 0.6985, "step": 26769 }, { "epoch": 0.9695762404925752, "grad_norm": 1.2141946955367149, "learning_rate": 4.8509110319708844e-08, "loss": 0.6315, "step": 26770 }, { "epoch": 0.9696124592538935, "grad_norm": 1.437013967643842, "learning_rate": 4.839377437793391e-08, "loss": 0.7524, "step": 26771 }, { "epoch": 0.9696486780152119, "grad_norm": 1.3142365841250183, "learning_rate": 4.827857537900493e-08, "loss": 0.6894, "step": 26772 }, { "epoch": 0.9696848967765302, "grad_norm": 1.2597936020930842, "learning_rate": 4.81635133245073e-08, "loss": 0.5899, "step": 26773 }, { "epoch": 0.9697211155378486, "grad_norm": 1.5006105787127633, "learning_rate": 4.80485882160231e-08, "loss": 0.7073, "step": 26774 }, { "epoch": 0.969757334299167, "grad_norm": 1.354543832758331, "learning_rate": 4.793380005513659e-08, "loss": 0.6236, "step": 26775 }, { "epoch": 0.9697935530604853, "grad_norm": 1.2375318661895305, "learning_rate": 4.781914884342431e-08, "loss": 0.6496, "step": 26776 }, { "epoch": 0.9698297718218037, "grad_norm": 1.2905059189160042, "learning_rate": 4.770463458246499e-08, "loss": 0.6794, "step": 26777 }, { "epoch": 0.9698659905831221, "grad_norm": 1.3975939644037456, "learning_rate": 4.759025727383626e-08, "loss": 0.6493, "step": 26778 }, { "epoch": 0.9699022093444404, "grad_norm": 1.12234930811455, "learning_rate": 4.747601691910908e-08, "loss": 0.7113, "step": 26779 }, { "epoch": 0.9699384281057588, "grad_norm": 1.2764732094987363, "learning_rate": 4.736191351985775e-08, "loss": 0.606, "step": 26780 }, { "epoch": 0.9699746468670771, "grad_norm": 1.647222100239866, "learning_rate": 4.724794707765212e-08, "loss": 0.6525, "step": 26781 }, { "epoch": 0.9700108656283956, "grad_norm": 1.4254577766212684, "learning_rate": 4.713411759405873e-08, "loss": 0.631, "step": 26782 }, { "epoch": 0.9700470843897139, "grad_norm": 1.3813971193397154, "learning_rate": 4.70204250706463e-08, "loss": 0.6187, "step": 26783 }, { "epoch": 0.9700833031510322, "grad_norm": 1.4016564672072447, "learning_rate": 4.6906869508976935e-08, "loss": 0.7173, "step": 26784 }, { "epoch": 0.9701195219123506, "grad_norm": 1.003961608055966, "learning_rate": 4.679345091061604e-08, "loss": 0.6239, "step": 26785 }, { "epoch": 0.9701557406736689, "grad_norm": 1.5437297598191497, "learning_rate": 4.6680169277122376e-08, "loss": 0.717, "step": 26786 }, { "epoch": 0.9701919594349874, "grad_norm": 1.5296874053144571, "learning_rate": 4.65670246100558e-08, "loss": 0.7673, "step": 26787 }, { "epoch": 0.9702281781963057, "grad_norm": 1.2844892375325052, "learning_rate": 4.645401691097173e-08, "loss": 0.6422, "step": 26788 }, { "epoch": 0.970264396957624, "grad_norm": 1.6304676613395066, "learning_rate": 4.634114618142782e-08, "loss": 0.6698, "step": 26789 }, { "epoch": 0.9703006157189424, "grad_norm": 1.3250355276313275, "learning_rate": 4.622841242297504e-08, "loss": 0.6354, "step": 26790 }, { "epoch": 0.9703368344802608, "grad_norm": 1.6529344174834517, "learning_rate": 4.6115815637164386e-08, "loss": 0.6803, "step": 26791 }, { "epoch": 0.9703730532415792, "grad_norm": 1.427030684783132, "learning_rate": 4.600335582554793e-08, "loss": 0.6409, "step": 26792 }, { "epoch": 0.9704092720028975, "grad_norm": 1.4295714336455183, "learning_rate": 4.5891032989671126e-08, "loss": 0.6321, "step": 26793 }, { "epoch": 0.9704454907642158, "grad_norm": 0.9724896895595084, "learning_rate": 4.5778847131079384e-08, "loss": 0.6529, "step": 26794 }, { "epoch": 0.9704817095255343, "grad_norm": 1.2504527903580698, "learning_rate": 4.5666798251318146e-08, "loss": 0.6562, "step": 26795 }, { "epoch": 0.9705179282868526, "grad_norm": 1.6234887660002884, "learning_rate": 4.5554886351928396e-08, "loss": 0.6736, "step": 26796 }, { "epoch": 0.970554147048171, "grad_norm": 1.1035639809778386, "learning_rate": 4.544311143445002e-08, "loss": 0.6687, "step": 26797 }, { "epoch": 0.9705903658094893, "grad_norm": 1.2737756350411227, "learning_rate": 4.533147350042178e-08, "loss": 0.6339, "step": 26798 }, { "epoch": 0.9706265845708076, "grad_norm": 1.4918267983729012, "learning_rate": 4.5219972551378e-08, "loss": 0.7006, "step": 26799 }, { "epoch": 0.9706628033321261, "grad_norm": 1.3438736060313319, "learning_rate": 4.510860858885524e-08, "loss": 0.7121, "step": 26800 }, { "epoch": 0.9706990220934444, "grad_norm": 1.364911083423102, "learning_rate": 4.49973816143856e-08, "loss": 0.6828, "step": 26801 }, { "epoch": 0.9707352408547628, "grad_norm": 1.4136611602084113, "learning_rate": 4.488629162949787e-08, "loss": 0.7063, "step": 26802 }, { "epoch": 0.9707714596160811, "grad_norm": 0.9839549527574294, "learning_rate": 4.477533863572414e-08, "loss": 0.6625, "step": 26803 }, { "epoch": 0.9708076783773995, "grad_norm": 1.1267115662670206, "learning_rate": 4.466452263458765e-08, "loss": 0.6916, "step": 26804 }, { "epoch": 0.9708438971387179, "grad_norm": 1.3027544034364809, "learning_rate": 4.455384362761495e-08, "loss": 0.6174, "step": 26805 }, { "epoch": 0.9708801159000362, "grad_norm": 1.4063208224423416, "learning_rate": 4.444330161632926e-08, "loss": 0.7109, "step": 26806 }, { "epoch": 0.9709163346613546, "grad_norm": 1.334970693266577, "learning_rate": 4.433289660225159e-08, "loss": 0.6878, "step": 26807 }, { "epoch": 0.970952553422673, "grad_norm": 1.4495475735310137, "learning_rate": 4.422262858690074e-08, "loss": 0.6576, "step": 26808 }, { "epoch": 0.9709887721839913, "grad_norm": 1.518388594347949, "learning_rate": 4.4112497571795475e-08, "loss": 0.6881, "step": 26809 }, { "epoch": 0.9710249909453097, "grad_norm": 1.3955118807926232, "learning_rate": 4.400250355844904e-08, "loss": 0.681, "step": 26810 }, { "epoch": 0.971061209706628, "grad_norm": 1.4130161877166874, "learning_rate": 4.389264654837688e-08, "loss": 0.6919, "step": 26811 }, { "epoch": 0.9710974284679464, "grad_norm": 1.323970546360946, "learning_rate": 4.378292654309002e-08, "loss": 0.652, "step": 26812 }, { "epoch": 0.9711336472292648, "grad_norm": 1.2727077929747526, "learning_rate": 4.3673343544098354e-08, "loss": 0.7189, "step": 26813 }, { "epoch": 0.9711698659905831, "grad_norm": 1.375245526186277, "learning_rate": 4.356389755290957e-08, "loss": 0.6273, "step": 26814 }, { "epoch": 0.9712060847519015, "grad_norm": 1.060397408455266, "learning_rate": 4.345458857103024e-08, "loss": 0.7005, "step": 26815 }, { "epoch": 0.9712423035132198, "grad_norm": 1.1223040154046207, "learning_rate": 4.3345416599964715e-08, "loss": 0.7035, "step": 26816 }, { "epoch": 0.9712785222745383, "grad_norm": 1.3607115573865984, "learning_rate": 4.323638164121402e-08, "loss": 0.6255, "step": 26817 }, { "epoch": 0.9713147410358566, "grad_norm": 1.5155387662167925, "learning_rate": 4.312748369628028e-08, "loss": 0.6898, "step": 26818 }, { "epoch": 0.9713509597971749, "grad_norm": 1.4029621790188136, "learning_rate": 4.3018722766661194e-08, "loss": 0.708, "step": 26819 }, { "epoch": 0.9713871785584933, "grad_norm": 1.3672454051970517, "learning_rate": 4.291009885385333e-08, "loss": 0.7187, "step": 26820 }, { "epoch": 0.9714233973198116, "grad_norm": 0.956588207027551, "learning_rate": 4.2801611959351065e-08, "loss": 0.6539, "step": 26821 }, { "epoch": 0.9714596160811301, "grad_norm": 1.2946147367460654, "learning_rate": 4.269326208464875e-08, "loss": 0.6515, "step": 26822 }, { "epoch": 0.9714958348424484, "grad_norm": 1.0106870098302503, "learning_rate": 4.2585049231236295e-08, "loss": 0.704, "step": 26823 }, { "epoch": 0.9715320536037667, "grad_norm": 1.3116721312535486, "learning_rate": 4.247697340060142e-08, "loss": 0.6023, "step": 26824 }, { "epoch": 0.9715682723650851, "grad_norm": 1.4298822173945973, "learning_rate": 4.236903459423402e-08, "loss": 0.6938, "step": 26825 }, { "epoch": 0.9716044911264035, "grad_norm": 0.9457489841646389, "learning_rate": 4.226123281361849e-08, "loss": 0.7212, "step": 26826 }, { "epoch": 0.9716407098877219, "grad_norm": 1.4247121094504593, "learning_rate": 4.215356806023807e-08, "loss": 0.6159, "step": 26827 }, { "epoch": 0.9716769286490402, "grad_norm": 1.3983693306040665, "learning_rate": 4.20460403355738e-08, "loss": 0.7063, "step": 26828 }, { "epoch": 0.9717131474103585, "grad_norm": 1.2705207790325903, "learning_rate": 4.193864964110561e-08, "loss": 0.6766, "step": 26829 }, { "epoch": 0.971749366171677, "grad_norm": 1.4375777029204049, "learning_rate": 4.1831395978312317e-08, "loss": 0.7605, "step": 26830 }, { "epoch": 0.9717855849329953, "grad_norm": 1.0691969368573377, "learning_rate": 4.17242793486683e-08, "loss": 0.7291, "step": 26831 }, { "epoch": 0.9718218036943137, "grad_norm": 1.4270268668944153, "learning_rate": 4.1617299753649035e-08, "loss": 0.691, "step": 26832 }, { "epoch": 0.971858022455632, "grad_norm": 1.358771604733151, "learning_rate": 4.151045719472557e-08, "loss": 0.6288, "step": 26833 }, { "epoch": 0.9718942412169503, "grad_norm": 1.7775359772912251, "learning_rate": 4.1403751673367855e-08, "loss": 0.539, "step": 26834 }, { "epoch": 0.9719304599782688, "grad_norm": 1.3785728241935156, "learning_rate": 4.129718319104581e-08, "loss": 0.6355, "step": 26835 }, { "epoch": 0.9719666787395871, "grad_norm": 1.450549721358833, "learning_rate": 4.119075174922493e-08, "loss": 0.6848, "step": 26836 }, { "epoch": 0.9720028975009055, "grad_norm": 1.5943522367306684, "learning_rate": 4.1084457349369614e-08, "loss": 0.6839, "step": 26837 }, { "epoch": 0.9720391162622238, "grad_norm": 0.9976255411152006, "learning_rate": 4.097829999294312e-08, "loss": 0.6419, "step": 26838 }, { "epoch": 0.9720753350235422, "grad_norm": 1.3303769597966781, "learning_rate": 4.087227968140539e-08, "loss": 0.6187, "step": 26839 }, { "epoch": 0.9721115537848606, "grad_norm": 1.4422553748032134, "learning_rate": 4.076639641621638e-08, "loss": 0.6461, "step": 26840 }, { "epoch": 0.9721477725461789, "grad_norm": 1.107323581474764, "learning_rate": 4.0660650198832696e-08, "loss": 0.7162, "step": 26841 }, { "epoch": 0.9721839913074973, "grad_norm": 1.3769147482048927, "learning_rate": 4.055504103070873e-08, "loss": 0.6783, "step": 26842 }, { "epoch": 0.9722202100688156, "grad_norm": 1.0811071198476023, "learning_rate": 4.044956891329888e-08, "loss": 0.6177, "step": 26843 }, { "epoch": 0.972256428830134, "grad_norm": 1.4786086336660815, "learning_rate": 4.03442338480553e-08, "loss": 0.6238, "step": 26844 }, { "epoch": 0.9722926475914524, "grad_norm": 1.5765314601946523, "learning_rate": 4.023903583642574e-08, "loss": 0.6807, "step": 26845 }, { "epoch": 0.9723288663527707, "grad_norm": 1.629091746673767, "learning_rate": 4.013397487985793e-08, "loss": 0.6608, "step": 26846 }, { "epoch": 0.9723650851140891, "grad_norm": 1.3882684275829804, "learning_rate": 4.0029050979798476e-08, "loss": 0.7494, "step": 26847 }, { "epoch": 0.9724013038754075, "grad_norm": 1.049796577106241, "learning_rate": 3.9924264137690684e-08, "loss": 0.6465, "step": 26848 }, { "epoch": 0.9724375226367258, "grad_norm": 1.4653065192071455, "learning_rate": 3.9819614354977834e-08, "loss": 0.6878, "step": 26849 }, { "epoch": 0.9724737413980442, "grad_norm": 1.106919142881583, "learning_rate": 3.971510163309766e-08, "loss": 0.6349, "step": 26850 }, { "epoch": 0.9725099601593625, "grad_norm": 1.4211048880754542, "learning_rate": 3.961072597348903e-08, "loss": 0.6178, "step": 26851 }, { "epoch": 0.972546178920681, "grad_norm": 1.0805548454188532, "learning_rate": 3.950648737758966e-08, "loss": 0.6588, "step": 26852 }, { "epoch": 0.9725823976819993, "grad_norm": 1.0577887224164702, "learning_rate": 3.940238584683287e-08, "loss": 0.6944, "step": 26853 }, { "epoch": 0.9726186164433176, "grad_norm": 1.3866033344863042, "learning_rate": 3.929842138265194e-08, "loss": 0.6941, "step": 26854 }, { "epoch": 0.972654835204636, "grad_norm": 1.4539149957590596, "learning_rate": 3.9194593986476845e-08, "loss": 0.6419, "step": 26855 }, { "epoch": 0.9726910539659543, "grad_norm": 1.2304733614986532, "learning_rate": 3.909090365973645e-08, "loss": 0.6789, "step": 26856 }, { "epoch": 0.9727272727272728, "grad_norm": 1.4405175638064054, "learning_rate": 3.898735040385737e-08, "loss": 0.6508, "step": 26857 }, { "epoch": 0.9727634914885911, "grad_norm": 1.312615409386471, "learning_rate": 3.888393422026515e-08, "loss": 0.6799, "step": 26858 }, { "epoch": 0.9727997102499094, "grad_norm": 1.0779215875730987, "learning_rate": 3.878065511038198e-08, "loss": 0.6736, "step": 26859 }, { "epoch": 0.9728359290112278, "grad_norm": 0.9442153512687831, "learning_rate": 3.867751307563006e-08, "loss": 0.676, "step": 26860 }, { "epoch": 0.9728721477725462, "grad_norm": 1.2239134216354155, "learning_rate": 3.857450811742824e-08, "loss": 0.7047, "step": 26861 }, { "epoch": 0.9729083665338646, "grad_norm": 1.1524551364792166, "learning_rate": 3.84716402371943e-08, "loss": 0.6667, "step": 26862 }, { "epoch": 0.9729445852951829, "grad_norm": 1.4439950815480944, "learning_rate": 3.836890943634375e-08, "loss": 0.6981, "step": 26863 }, { "epoch": 0.9729808040565012, "grad_norm": 1.346166389794648, "learning_rate": 3.8266315716288804e-08, "loss": 0.6618, "step": 26864 }, { "epoch": 0.9730170228178197, "grad_norm": 1.4689067626295278, "learning_rate": 3.816385907844389e-08, "loss": 0.6888, "step": 26865 }, { "epoch": 0.973053241579138, "grad_norm": 1.3577337565377574, "learning_rate": 3.806153952421676e-08, "loss": 0.642, "step": 26866 }, { "epoch": 0.9730894603404564, "grad_norm": 1.0565920788212662, "learning_rate": 3.795935705501741e-08, "loss": 0.666, "step": 26867 }, { "epoch": 0.9731256791017747, "grad_norm": 1.370130028396, "learning_rate": 3.785731167224915e-08, "loss": 0.6358, "step": 26868 }, { "epoch": 0.973161897863093, "grad_norm": 0.9700025056197696, "learning_rate": 3.7755403377318644e-08, "loss": 0.6333, "step": 26869 }, { "epoch": 0.9731981166244115, "grad_norm": 1.084964082217187, "learning_rate": 3.765363217162698e-08, "loss": 0.6869, "step": 26870 }, { "epoch": 0.9732343353857298, "grad_norm": 1.340784700313886, "learning_rate": 3.755199805657528e-08, "loss": 0.668, "step": 26871 }, { "epoch": 0.9732705541470482, "grad_norm": 1.3221751348333826, "learning_rate": 3.74505010335624e-08, "loss": 0.5945, "step": 26872 }, { "epoch": 0.9733067729083665, "grad_norm": 1.4139729288353369, "learning_rate": 3.734914110398391e-08, "loss": 0.6717, "step": 26873 }, { "epoch": 0.9733429916696849, "grad_norm": 1.5572011290314947, "learning_rate": 3.724791826923424e-08, "loss": 0.5988, "step": 26874 }, { "epoch": 0.9733792104310033, "grad_norm": 1.498141140315736, "learning_rate": 3.7146832530708944e-08, "loss": 0.68, "step": 26875 }, { "epoch": 0.9734154291923216, "grad_norm": 1.4207273255710666, "learning_rate": 3.7045883889796907e-08, "loss": 0.595, "step": 26876 }, { "epoch": 0.97345164795364, "grad_norm": 1.4059352562266003, "learning_rate": 3.694507234788702e-08, "loss": 0.6304, "step": 26877 }, { "epoch": 0.9734878667149583, "grad_norm": 1.1729016058593817, "learning_rate": 3.684439790636707e-08, "loss": 0.725, "step": 26878 }, { "epoch": 0.9735240854762767, "grad_norm": 1.3203082881668446, "learning_rate": 3.674386056662371e-08, "loss": 0.618, "step": 26879 }, { "epoch": 0.9735603042375951, "grad_norm": 1.3861772017072107, "learning_rate": 3.664346033003918e-08, "loss": 0.7452, "step": 26880 }, { "epoch": 0.9735965229989134, "grad_norm": 1.3407615383405416, "learning_rate": 3.65431971979946e-08, "loss": 0.6118, "step": 26881 }, { "epoch": 0.9736327417602318, "grad_norm": 1.4379706312045188, "learning_rate": 3.644307117186996e-08, "loss": 0.6514, "step": 26882 }, { "epoch": 0.9736689605215502, "grad_norm": 1.451387734614856, "learning_rate": 3.634308225304417e-08, "loss": 0.6201, "step": 26883 }, { "epoch": 0.9737051792828685, "grad_norm": 1.0319600294064886, "learning_rate": 3.624323044289169e-08, "loss": 0.6586, "step": 26884 }, { "epoch": 0.9737413980441869, "grad_norm": 1.413308837172215, "learning_rate": 3.614351574278807e-08, "loss": 0.6818, "step": 26885 }, { "epoch": 0.9737776168055052, "grad_norm": 1.3625231694134798, "learning_rate": 3.604393815410445e-08, "loss": 0.6876, "step": 26886 }, { "epoch": 0.9738138355668237, "grad_norm": 1.3934124190294832, "learning_rate": 3.594449767821084e-08, "loss": 0.6656, "step": 26887 }, { "epoch": 0.973850054328142, "grad_norm": 0.9265455313998541, "learning_rate": 3.584519431647726e-08, "loss": 0.6548, "step": 26888 }, { "epoch": 0.9738862730894603, "grad_norm": 1.3633168401690587, "learning_rate": 3.574602807026817e-08, "loss": 0.7224, "step": 26889 }, { "epoch": 0.9739224918507787, "grad_norm": 1.3658519324994667, "learning_rate": 3.564699894094914e-08, "loss": 0.6423, "step": 26890 }, { "epoch": 0.973958710612097, "grad_norm": 1.1012662532444202, "learning_rate": 3.554810692988353e-08, "loss": 0.6102, "step": 26891 }, { "epoch": 0.9739949293734155, "grad_norm": 1.2444393407190077, "learning_rate": 3.544935203843025e-08, "loss": 0.6628, "step": 26892 }, { "epoch": 0.9740311481347338, "grad_norm": 1.3692504645569337, "learning_rate": 3.535073426795044e-08, "loss": 0.6507, "step": 26893 }, { "epoch": 0.9740673668960521, "grad_norm": 0.9509134220214134, "learning_rate": 3.525225361979967e-08, "loss": 0.6986, "step": 26894 }, { "epoch": 0.9741035856573705, "grad_norm": 1.554706836944402, "learning_rate": 3.515391009533464e-08, "loss": 0.7466, "step": 26895 }, { "epoch": 0.9741398044186889, "grad_norm": 1.3582819087207811, "learning_rate": 3.5055703695906494e-08, "loss": 0.6689, "step": 26896 }, { "epoch": 0.9741760231800073, "grad_norm": 1.4296753640519655, "learning_rate": 3.495763442286859e-08, "loss": 0.622, "step": 26897 }, { "epoch": 0.9742122419413256, "grad_norm": 1.4281131806760967, "learning_rate": 3.485970227756985e-08, "loss": 0.6331, "step": 26898 }, { "epoch": 0.9742484607026439, "grad_norm": 1.4563712576083923, "learning_rate": 3.4761907261356976e-08, "loss": 0.6493, "step": 26899 }, { "epoch": 0.9742846794639624, "grad_norm": 1.493367149508408, "learning_rate": 3.466424937557667e-08, "loss": 0.6411, "step": 26900 }, { "epoch": 0.9743208982252807, "grad_norm": 1.2892467287659453, "learning_rate": 3.45667286215734e-08, "loss": 0.5942, "step": 26901 }, { "epoch": 0.9743571169865991, "grad_norm": 1.3710337656979383, "learning_rate": 3.446934500068722e-08, "loss": 0.6384, "step": 26902 }, { "epoch": 0.9743933357479174, "grad_norm": 1.0554267176098935, "learning_rate": 3.4372098514260374e-08, "loss": 0.7489, "step": 26903 }, { "epoch": 0.9744295545092357, "grad_norm": 1.3931832149242145, "learning_rate": 3.4274989163629594e-08, "loss": 0.654, "step": 26904 }, { "epoch": 0.9744657732705542, "grad_norm": 1.303898297109058, "learning_rate": 3.417801695013156e-08, "loss": 0.6418, "step": 26905 }, { "epoch": 0.9745019920318725, "grad_norm": 1.4284442279801606, "learning_rate": 3.408118187510079e-08, "loss": 0.6382, "step": 26906 }, { "epoch": 0.9745382107931909, "grad_norm": 1.4508508465784156, "learning_rate": 3.3984483939869525e-08, "loss": 0.6434, "step": 26907 }, { "epoch": 0.9745744295545092, "grad_norm": 1.4177115827184974, "learning_rate": 3.388792314576894e-08, "loss": 0.6632, "step": 26908 }, { "epoch": 0.9746106483158276, "grad_norm": 1.4339300574518672, "learning_rate": 3.3791499494127964e-08, "loss": 0.6699, "step": 26909 }, { "epoch": 0.974646867077146, "grad_norm": 0.948201441724246, "learning_rate": 3.3695212986272207e-08, "loss": 0.705, "step": 26910 }, { "epoch": 0.9746830858384643, "grad_norm": 1.3195729279302355, "learning_rate": 3.3599063623527276e-08, "loss": 0.5823, "step": 26911 }, { "epoch": 0.9747193045997827, "grad_norm": 1.4926099466337552, "learning_rate": 3.350305140721766e-08, "loss": 0.6076, "step": 26912 }, { "epoch": 0.974755523361101, "grad_norm": 1.4451107554493017, "learning_rate": 3.340717633866231e-08, "loss": 0.6671, "step": 26913 }, { "epoch": 0.9747917421224194, "grad_norm": 1.3520135168338563, "learning_rate": 3.3311438419182385e-08, "loss": 0.681, "step": 26914 }, { "epoch": 0.9748279608837378, "grad_norm": 1.4678101036938684, "learning_rate": 3.321583765009462e-08, "loss": 0.6823, "step": 26915 }, { "epoch": 0.9748641796450561, "grad_norm": 1.688611743318779, "learning_rate": 3.312037403271351e-08, "loss": 0.636, "step": 26916 }, { "epoch": 0.9749003984063745, "grad_norm": 1.6674435434120742, "learning_rate": 3.302504756835467e-08, "loss": 0.657, "step": 26917 }, { "epoch": 0.9749366171676929, "grad_norm": 1.4780206494263248, "learning_rate": 3.292985825832817e-08, "loss": 0.7236, "step": 26918 }, { "epoch": 0.9749728359290112, "grad_norm": 1.4254691827567039, "learning_rate": 3.2834806103945184e-08, "loss": 0.6613, "step": 26919 }, { "epoch": 0.9750090546903296, "grad_norm": 1.3879966448942065, "learning_rate": 3.273989110651354e-08, "loss": 0.6417, "step": 26920 }, { "epoch": 0.9750452734516479, "grad_norm": 1.2818887880151995, "learning_rate": 3.2645113267338876e-08, "loss": 0.684, "step": 26921 }, { "epoch": 0.9750814922129664, "grad_norm": 1.413974995039521, "learning_rate": 3.255047258772459e-08, "loss": 0.6188, "step": 26922 }, { "epoch": 0.9751177109742847, "grad_norm": 1.038701853007044, "learning_rate": 3.245596906897519e-08, "loss": 0.7344, "step": 26923 }, { "epoch": 0.975153929735603, "grad_norm": 1.0956655178611547, "learning_rate": 3.2361602712389637e-08, "loss": 0.6515, "step": 26924 }, { "epoch": 0.9751901484969214, "grad_norm": 1.454094811187504, "learning_rate": 3.226737351926579e-08, "loss": 0.6539, "step": 26925 }, { "epoch": 0.9752263672582397, "grad_norm": 1.4794509151732513, "learning_rate": 3.2173281490901485e-08, "loss": 0.6566, "step": 26926 }, { "epoch": 0.9752625860195582, "grad_norm": 1.3916714767829301, "learning_rate": 3.2079326628592365e-08, "loss": 0.6658, "step": 26927 }, { "epoch": 0.9752988047808765, "grad_norm": 1.2530381883424409, "learning_rate": 3.1985508933629613e-08, "loss": 0.6943, "step": 26928 }, { "epoch": 0.9753350235421948, "grad_norm": 1.3005376170941352, "learning_rate": 3.189182840730554e-08, "loss": 0.6957, "step": 26929 }, { "epoch": 0.9753712423035132, "grad_norm": 1.1192924294509432, "learning_rate": 3.179828505090798e-08, "loss": 0.6701, "step": 26930 }, { "epoch": 0.9754074610648316, "grad_norm": 1.3027994541401222, "learning_rate": 3.170487886572482e-08, "loss": 0.7308, "step": 26931 }, { "epoch": 0.97544367982615, "grad_norm": 1.4237725635566763, "learning_rate": 3.161160985304168e-08, "loss": 0.6925, "step": 26932 }, { "epoch": 0.9754798985874683, "grad_norm": 1.250364552178065, "learning_rate": 3.151847801414199e-08, "loss": 0.6317, "step": 26933 }, { "epoch": 0.9755161173487866, "grad_norm": 1.4568976629958248, "learning_rate": 3.142548335030693e-08, "loss": 0.6945, "step": 26934 }, { "epoch": 0.9755523361101051, "grad_norm": 1.3359368272690444, "learning_rate": 3.13326258628166e-08, "loss": 0.6893, "step": 26935 }, { "epoch": 0.9755885548714234, "grad_norm": 1.1261410811159165, "learning_rate": 3.123990555294776e-08, "loss": 0.6671, "step": 26936 }, { "epoch": 0.9756247736327418, "grad_norm": 1.3930132120688399, "learning_rate": 3.1147322421977154e-08, "loss": 0.6421, "step": 26937 }, { "epoch": 0.9756609923940601, "grad_norm": 1.4773942065788654, "learning_rate": 3.105487647117933e-08, "loss": 0.7548, "step": 26938 }, { "epoch": 0.9756972111553784, "grad_norm": 1.4151952366440759, "learning_rate": 3.0962567701826594e-08, "loss": 0.6751, "step": 26939 }, { "epoch": 0.9757334299166969, "grad_norm": 1.4572264697784651, "learning_rate": 3.087039611518683e-08, "loss": 0.7312, "step": 26940 }, { "epoch": 0.9757696486780152, "grad_norm": 1.3731697178147604, "learning_rate": 3.0778361712531234e-08, "loss": 0.668, "step": 26941 }, { "epoch": 0.9758058674393336, "grad_norm": 1.3388916419624952, "learning_rate": 3.068646449512547e-08, "loss": 0.633, "step": 26942 }, { "epoch": 0.9758420862006519, "grad_norm": 1.4018677179803891, "learning_rate": 3.059470446423296e-08, "loss": 0.6667, "step": 26943 }, { "epoch": 0.9758783049619703, "grad_norm": 1.3584939558341282, "learning_rate": 3.0503081621117145e-08, "loss": 0.6881, "step": 26944 }, { "epoch": 0.9759145237232887, "grad_norm": 0.9660070582175604, "learning_rate": 3.041159596703924e-08, "loss": 0.6565, "step": 26945 }, { "epoch": 0.975950742484607, "grad_norm": 1.3638024549815781, "learning_rate": 3.032024750325824e-08, "loss": 0.6618, "step": 26946 }, { "epoch": 0.9759869612459254, "grad_norm": 1.487686537768947, "learning_rate": 3.0229036231029795e-08, "loss": 0.6172, "step": 26947 }, { "epoch": 0.9760231800072438, "grad_norm": 1.4888263827910888, "learning_rate": 3.013796215161069e-08, "loss": 0.7164, "step": 26948 }, { "epoch": 0.9760593987685621, "grad_norm": 1.0532730393457714, "learning_rate": 3.004702526625325e-08, "loss": 0.7286, "step": 26949 }, { "epoch": 0.9760956175298805, "grad_norm": 1.4067112616526443, "learning_rate": 2.99562255762087e-08, "loss": 0.6426, "step": 26950 }, { "epoch": 0.9761318362911988, "grad_norm": 1.061990656260198, "learning_rate": 2.986556308272715e-08, "loss": 0.6592, "step": 26951 }, { "epoch": 0.9761680550525172, "grad_norm": 1.5194548707805198, "learning_rate": 2.9775037787056482e-08, "loss": 0.6312, "step": 26952 }, { "epoch": 0.9762042738138356, "grad_norm": 1.1387582270475256, "learning_rate": 2.968464969044238e-08, "loss": 0.6262, "step": 26953 }, { "epoch": 0.9762404925751539, "grad_norm": 1.2635020233280279, "learning_rate": 2.9594398794127178e-08, "loss": 0.632, "step": 26954 }, { "epoch": 0.9762767113364723, "grad_norm": 1.3008635240030106, "learning_rate": 2.9504285099354323e-08, "loss": 0.6419, "step": 26955 }, { "epoch": 0.9763129300977906, "grad_norm": 1.2047201615438843, "learning_rate": 2.9414308607362828e-08, "loss": 0.698, "step": 26956 }, { "epoch": 0.9763491488591091, "grad_norm": 1.0869393860389511, "learning_rate": 2.9324469319391702e-08, "loss": 0.6832, "step": 26957 }, { "epoch": 0.9763853676204274, "grad_norm": 1.5073325636088684, "learning_rate": 2.923476723667662e-08, "loss": 0.7373, "step": 26958 }, { "epoch": 0.9764215863817457, "grad_norm": 1.378802336108459, "learning_rate": 2.9145202360452153e-08, "loss": 0.6571, "step": 26959 }, { "epoch": 0.9764578051430641, "grad_norm": 0.975810253481404, "learning_rate": 2.9055774691951756e-08, "loss": 0.6408, "step": 26960 }, { "epoch": 0.9764940239043824, "grad_norm": 1.3358903410881018, "learning_rate": 2.8966484232404445e-08, "loss": 0.652, "step": 26961 }, { "epoch": 0.9765302426657009, "grad_norm": 1.482823747133987, "learning_rate": 2.887733098303924e-08, "loss": 0.6825, "step": 26962 }, { "epoch": 0.9765664614270192, "grad_norm": 1.5459647571076256, "learning_rate": 2.878831494508405e-08, "loss": 0.7657, "step": 26963 }, { "epoch": 0.9766026801883375, "grad_norm": 1.73398247349397, "learning_rate": 2.8699436119763445e-08, "loss": 0.7323, "step": 26964 }, { "epoch": 0.9766388989496559, "grad_norm": 1.4749928133797545, "learning_rate": 2.8610694508299785e-08, "loss": 0.6464, "step": 26965 }, { "epoch": 0.9766751177109743, "grad_norm": 1.447976509208056, "learning_rate": 2.8522090111914312e-08, "loss": 0.7066, "step": 26966 }, { "epoch": 0.9767113364722927, "grad_norm": 1.3798114814868583, "learning_rate": 2.8433622931826054e-08, "loss": 0.648, "step": 26967 }, { "epoch": 0.976747555233611, "grad_norm": 1.3792252110900054, "learning_rate": 2.8345292969254035e-08, "loss": 0.656, "step": 26968 }, { "epoch": 0.9767837739949293, "grad_norm": 1.3719081497749248, "learning_rate": 2.8257100225411728e-08, "loss": 0.6283, "step": 26969 }, { "epoch": 0.9768199927562478, "grad_norm": 1.4170285577792638, "learning_rate": 2.8169044701513716e-08, "loss": 0.674, "step": 26970 }, { "epoch": 0.9768562115175661, "grad_norm": 0.9552551794832371, "learning_rate": 2.8081126398772364e-08, "loss": 0.6662, "step": 26971 }, { "epoch": 0.9768924302788845, "grad_norm": 1.3721254127271938, "learning_rate": 2.7993345318396702e-08, "loss": 0.7146, "step": 26972 }, { "epoch": 0.9769286490402028, "grad_norm": 1.0571395908998484, "learning_rate": 2.790570146159466e-08, "loss": 0.6803, "step": 26973 }, { "epoch": 0.9769648678015211, "grad_norm": 1.3340656889655966, "learning_rate": 2.7818194829571932e-08, "loss": 0.6584, "step": 26974 }, { "epoch": 0.9770010865628396, "grad_norm": 1.431339803474905, "learning_rate": 2.7730825423534224e-08, "loss": 0.677, "step": 26975 }, { "epoch": 0.9770373053241579, "grad_norm": 1.3343484753065138, "learning_rate": 2.7643593244681688e-08, "loss": 0.6177, "step": 26976 }, { "epoch": 0.9770735240854763, "grad_norm": 1.27524584668466, "learning_rate": 2.7556498294216694e-08, "loss": 0.6227, "step": 26977 }, { "epoch": 0.9771097428467946, "grad_norm": 1.0105742505057136, "learning_rate": 2.746954057333606e-08, "loss": 0.6584, "step": 26978 }, { "epoch": 0.977145961608113, "grad_norm": 1.3141324961579661, "learning_rate": 2.7382720083237724e-08, "loss": 0.6357, "step": 26979 }, { "epoch": 0.9771821803694314, "grad_norm": 1.520777446283898, "learning_rate": 2.729603682511517e-08, "loss": 0.6661, "step": 26980 }, { "epoch": 0.9772183991307497, "grad_norm": 1.4194601992753588, "learning_rate": 2.7209490800162998e-08, "loss": 0.6276, "step": 26981 }, { "epoch": 0.9772546178920681, "grad_norm": 1.2983325313894447, "learning_rate": 2.7123082009570255e-08, "loss": 0.6494, "step": 26982 }, { "epoch": 0.9772908366533865, "grad_norm": 1.353771201413064, "learning_rate": 2.7036810454527106e-08, "loss": 0.7122, "step": 26983 }, { "epoch": 0.9773270554147048, "grad_norm": 1.2489408349407989, "learning_rate": 2.6950676136220378e-08, "loss": 0.6748, "step": 26984 }, { "epoch": 0.9773632741760232, "grad_norm": 1.470481242835243, "learning_rate": 2.6864679055835786e-08, "loss": 0.6451, "step": 26985 }, { "epoch": 0.9773994929373415, "grad_norm": 1.4272069876732967, "learning_rate": 2.6778819214555718e-08, "loss": 0.6845, "step": 26986 }, { "epoch": 0.97743571169866, "grad_norm": 1.4039275356158696, "learning_rate": 2.6693096613562563e-08, "loss": 0.6427, "step": 26987 }, { "epoch": 0.9774719304599783, "grad_norm": 1.3349577123049081, "learning_rate": 2.6607511254035377e-08, "loss": 0.6529, "step": 26988 }, { "epoch": 0.9775081492212966, "grad_norm": 1.0875261198953825, "learning_rate": 2.652206313715322e-08, "loss": 0.6319, "step": 26989 }, { "epoch": 0.977544367982615, "grad_norm": 1.3629090489667157, "learning_rate": 2.6436752264090702e-08, "loss": 0.7198, "step": 26990 }, { "epoch": 0.9775805867439333, "grad_norm": 0.9007681593756409, "learning_rate": 2.6351578636021335e-08, "loss": 0.6929, "step": 26991 }, { "epoch": 0.9776168055052518, "grad_norm": 1.0800311011930488, "learning_rate": 2.6266542254118622e-08, "loss": 0.69, "step": 26992 }, { "epoch": 0.9776530242665701, "grad_norm": 1.3841951254996636, "learning_rate": 2.6181643119551624e-08, "loss": 0.6137, "step": 26993 }, { "epoch": 0.9776892430278884, "grad_norm": 1.3848562110089324, "learning_rate": 2.6096881233489413e-08, "loss": 0.6743, "step": 26994 }, { "epoch": 0.9777254617892068, "grad_norm": 1.363913280260486, "learning_rate": 2.6012256597096607e-08, "loss": 0.6382, "step": 26995 }, { "epoch": 0.9777616805505251, "grad_norm": 1.5207885269908845, "learning_rate": 2.5927769211540054e-08, "loss": 0.6701, "step": 26996 }, { "epoch": 0.9777978993118436, "grad_norm": 1.3786935669269056, "learning_rate": 2.5843419077981046e-08, "loss": 0.6562, "step": 26997 }, { "epoch": 0.9778341180731619, "grad_norm": 1.4752221482698562, "learning_rate": 2.5759206197580876e-08, "loss": 0.5884, "step": 26998 }, { "epoch": 0.9778703368344802, "grad_norm": 1.4990693958208157, "learning_rate": 2.567513057149862e-08, "loss": 0.7285, "step": 26999 }, { "epoch": 0.9779065555957986, "grad_norm": 1.3138772814099102, "learning_rate": 2.5591192200890012e-08, "loss": 0.6569, "step": 27000 } ], "logging_steps": 1.0, "max_steps": 27610, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.9828973577575596e+20, "train_batch_size": 4, "trial_name": null, "trial_params": null }