{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.4636336426219696, "eval_steps": 26, "global_step": 260, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.009577970667464832, "grad_norm": 10.084560608592307, "learning_rate": 1.7241379310344828e-07, "loss": 1.579, "step": 1 }, { "epoch": 0.009577970667464832, "eval_loss": 2.5250306129455566, "eval_runtime": 107.4458, "eval_samples_per_second": 13.16, "eval_steps_per_second": 3.295, "step": 1 }, { "epoch": 0.019155941334929663, "grad_norm": 8.306669565661105, "learning_rate": 3.4482758620689656e-07, "loss": 1.5724, "step": 2 }, { "epoch": 0.02873391200239449, "grad_norm": 13.938049536284893, "learning_rate": 5.172413793103449e-07, "loss": 1.5871, "step": 3 }, { "epoch": 0.038311882669859326, "grad_norm": 12.43456292626288, "learning_rate": 6.896551724137931e-07, "loss": 1.5681, "step": 4 }, { "epoch": 0.04788985333732416, "grad_norm": 13.870879646573128, "learning_rate": 8.620689655172415e-07, "loss": 1.5744, "step": 5 }, { "epoch": 0.05746782400478898, "grad_norm": 15.247654309196745, "learning_rate": 1.0344827586206898e-06, "loss": 1.5925, "step": 6 }, { "epoch": 0.06704579467225381, "grad_norm": 15.680512101057806, "learning_rate": 1.2068965517241381e-06, "loss": 1.5704, "step": 7 }, { "epoch": 0.07662376533971865, "grad_norm": 14.30414461091009, "learning_rate": 1.3793103448275862e-06, "loss": 1.5732, "step": 8 }, { "epoch": 0.08620173600718348, "grad_norm": 11.033868746409794, "learning_rate": 1.5517241379310346e-06, "loss": 1.5325, "step": 9 }, { "epoch": 0.09577970667464832, "grad_norm": 9.293155363204939, "learning_rate": 1.724137931034483e-06, "loss": 1.5525, "step": 10 }, { "epoch": 0.10535767734211314, "grad_norm": 10.55909566144827, "learning_rate": 1.896551724137931e-06, "loss": 1.5283, "step": 11 }, { "epoch": 0.11493564800957796, "grad_norm": 7.362528707126726, "learning_rate": 2.0689655172413796e-06, "loss": 1.5246, "step": 12 }, { "epoch": 0.1245136186770428, "grad_norm": 7.368215078656617, "learning_rate": 2.241379310344828e-06, "loss": 1.5313, "step": 13 }, { "epoch": 0.13409158934450763, "grad_norm": 6.065170717786516, "learning_rate": 2.4137931034482762e-06, "loss": 1.5027, "step": 14 }, { "epoch": 0.14366956001197245, "grad_norm": 5.328528823161362, "learning_rate": 2.5862068965517246e-06, "loss": 1.481, "step": 15 }, { "epoch": 0.1532475306794373, "grad_norm": 4.425999183762783, "learning_rate": 2.7586206896551725e-06, "loss": 1.4494, "step": 16 }, { "epoch": 0.16282550134690213, "grad_norm": 2.3104583142533675, "learning_rate": 2.931034482758621e-06, "loss": 1.4645, "step": 17 }, { "epoch": 0.17240347201436695, "grad_norm": 1.595394748941364, "learning_rate": 3.103448275862069e-06, "loss": 1.4619, "step": 18 }, { "epoch": 0.18198144268183178, "grad_norm": 1.2488731383034972, "learning_rate": 3.2758620689655175e-06, "loss": 1.4641, "step": 19 }, { "epoch": 0.19155941334929663, "grad_norm": 1.5772662843657, "learning_rate": 3.448275862068966e-06, "loss": 1.4029, "step": 20 }, { "epoch": 0.20113738401676146, "grad_norm": 2.556424014112241, "learning_rate": 3.620689655172414e-06, "loss": 1.4453, "step": 21 }, { "epoch": 0.21071535468422628, "grad_norm": 2.0581192872654483, "learning_rate": 3.793103448275862e-06, "loss": 1.4135, "step": 22 }, { "epoch": 0.2202933253516911, "grad_norm": 1.6613052346475512, "learning_rate": 3.96551724137931e-06, "loss": 1.4336, "step": 23 }, { "epoch": 0.22987129601915593, "grad_norm": 1.2670811596205898, "learning_rate": 4.137931034482759e-06, "loss": 1.3898, "step": 24 }, { "epoch": 0.23944926668662078, "grad_norm": 1.4594637064715403, "learning_rate": 4.310344827586207e-06, "loss": 1.392, "step": 25 }, { "epoch": 0.2490272373540856, "grad_norm": 1.6947460151500366, "learning_rate": 4.482758620689656e-06, "loss": 1.3967, "step": 26 }, { "epoch": 0.2490272373540856, "eval_loss": 2.319483518600464, "eval_runtime": 107.1009, "eval_samples_per_second": 13.202, "eval_steps_per_second": 3.305, "step": 26 }, { "epoch": 0.25860520802155046, "grad_norm": 1.4794556532045955, "learning_rate": 4.655172413793104e-06, "loss": 1.3882, "step": 27 }, { "epoch": 0.26818317868901526, "grad_norm": 1.275878657564078, "learning_rate": 4.8275862068965525e-06, "loss": 1.4152, "step": 28 }, { "epoch": 0.2777611493564801, "grad_norm": 1.0273810925450593, "learning_rate": 5e-06, "loss": 1.3897, "step": 29 }, { "epoch": 0.2873391200239449, "grad_norm": 1.3658855156304837, "learning_rate": 4.9998459603839726e-06, "loss": 1.3539, "step": 30 }, { "epoch": 0.29691709069140976, "grad_norm": 1.160650318212732, "learning_rate": 4.9993838605184505e-06, "loss": 1.3461, "step": 31 }, { "epoch": 0.3064950613588746, "grad_norm": 0.9334705830010439, "learning_rate": 4.998613757348784e-06, "loss": 1.3575, "step": 32 }, { "epoch": 0.3160730320263394, "grad_norm": 1.0269221075865582, "learning_rate": 4.99753574577609e-06, "loss": 1.3503, "step": 33 }, { "epoch": 0.32565100269380426, "grad_norm": 0.9951200682896573, "learning_rate": 4.996149958645559e-06, "loss": 1.3718, "step": 34 }, { "epoch": 0.33522897336126906, "grad_norm": 0.8568405246328175, "learning_rate": 4.994456566730085e-06, "loss": 1.3515, "step": 35 }, { "epoch": 0.3448069440287339, "grad_norm": 0.8752926728569858, "learning_rate": 4.992455778709222e-06, "loss": 1.3571, "step": 36 }, { "epoch": 0.35438491469619876, "grad_norm": 0.9195979878575848, "learning_rate": 4.990147841143462e-06, "loss": 1.3335, "step": 37 }, { "epoch": 0.36396288536366356, "grad_norm": 0.8848215909446233, "learning_rate": 4.98753303844386e-06, "loss": 1.3093, "step": 38 }, { "epoch": 0.3735408560311284, "grad_norm": 0.8261733197817335, "learning_rate": 4.984611692836979e-06, "loss": 1.3376, "step": 39 }, { "epoch": 0.38311882669859326, "grad_norm": 0.7643849735934586, "learning_rate": 4.981384164325184e-06, "loss": 1.3172, "step": 40 }, { "epoch": 0.39269679736605806, "grad_norm": 0.8302859072234411, "learning_rate": 4.977850850642275e-06, "loss": 1.352, "step": 41 }, { "epoch": 0.4022747680335229, "grad_norm": 0.8019795318623388, "learning_rate": 4.97401218720448e-06, "loss": 1.3271, "step": 42 }, { "epoch": 0.4118527387009877, "grad_norm": 0.7856123291749388, "learning_rate": 4.969868647056793e-06, "loss": 1.3302, "step": 43 }, { "epoch": 0.42143070936845256, "grad_norm": 0.7212471859830762, "learning_rate": 4.965420740814679e-06, "loss": 1.3215, "step": 44 }, { "epoch": 0.4310086800359174, "grad_norm": 0.7660292329930958, "learning_rate": 4.960669016601155e-06, "loss": 1.3435, "step": 45 }, { "epoch": 0.4405866507033822, "grad_norm": 0.7247198414191649, "learning_rate": 4.95561405997924e-06, "loss": 1.3163, "step": 46 }, { "epoch": 0.45016462137084706, "grad_norm": 0.7419070442778594, "learning_rate": 4.950256493879795e-06, "loss": 1.3209, "step": 47 }, { "epoch": 0.45974259203831186, "grad_norm": 0.7024643859790418, "learning_rate": 4.94459697852476e-06, "loss": 1.2684, "step": 48 }, { "epoch": 0.4693205627057767, "grad_norm": 0.7208397492740805, "learning_rate": 4.938636211345792e-06, "loss": 1.2818, "step": 49 }, { "epoch": 0.47889853337324156, "grad_norm": 0.7159719760236076, "learning_rate": 4.932374926898321e-06, "loss": 1.3094, "step": 50 }, { "epoch": 0.48847650404070636, "grad_norm": 0.7100286359014379, "learning_rate": 4.92581389677103e-06, "loss": 1.3177, "step": 51 }, { "epoch": 0.4980544747081712, "grad_norm": 0.664062518173294, "learning_rate": 4.918953929490768e-06, "loss": 1.2868, "step": 52 }, { "epoch": 0.4980544747081712, "eval_loss": 2.239407777786255, "eval_runtime": 107.263, "eval_samples_per_second": 13.183, "eval_steps_per_second": 3.3, "step": 52 }, { "epoch": 0.507632445375636, "grad_norm": 0.8658636506450442, "learning_rate": 4.911795870422916e-06, "loss": 1.2904, "step": 53 }, { "epoch": 0.5172104160431009, "grad_norm": 0.6715121564275828, "learning_rate": 4.904340601667208e-06, "loss": 1.326, "step": 54 }, { "epoch": 0.5267883867105657, "grad_norm": 0.8518222183690225, "learning_rate": 4.896589041949036e-06, "loss": 1.2757, "step": 55 }, { "epoch": 0.5363663573780305, "grad_norm": 0.6780934729098863, "learning_rate": 4.888542146506224e-06, "loss": 1.3027, "step": 56 }, { "epoch": 0.5459443280454953, "grad_norm": 0.8407110074770763, "learning_rate": 4.880200906971321e-06, "loss": 1.2965, "step": 57 }, { "epoch": 0.5555222987129602, "grad_norm": 0.654501814705368, "learning_rate": 4.8715663512493924e-06, "loss": 1.2764, "step": 58 }, { "epoch": 0.565100269380425, "grad_norm": 0.7722805216190872, "learning_rate": 4.8626395433913595e-06, "loss": 1.2799, "step": 59 }, { "epoch": 0.5746782400478898, "grad_norm": 0.6575468000608066, "learning_rate": 4.853421583462866e-06, "loss": 1.3009, "step": 60 }, { "epoch": 0.5842562107153547, "grad_norm": 0.6919845481307941, "learning_rate": 4.8439136074087165e-06, "loss": 1.2885, "step": 61 }, { "epoch": 0.5938341813828195, "grad_norm": 0.652693683934317, "learning_rate": 4.834116786912897e-06, "loss": 1.2564, "step": 62 }, { "epoch": 0.6034121520502843, "grad_norm": 0.6684643483116979, "learning_rate": 4.82403232925418e-06, "loss": 1.278, "step": 63 }, { "epoch": 0.6129901227177492, "grad_norm": 0.6735443956477082, "learning_rate": 4.813661477157355e-06, "loss": 1.2895, "step": 64 }, { "epoch": 0.622568093385214, "grad_norm": 0.6574494336528988, "learning_rate": 4.803005508640083e-06, "loss": 1.2481, "step": 65 }, { "epoch": 0.6321460640526788, "grad_norm": 0.7061153031772025, "learning_rate": 4.7920657368554e-06, "loss": 1.3023, "step": 66 }, { "epoch": 0.6417240347201437, "grad_norm": 0.6609850544647713, "learning_rate": 4.780843509929905e-06, "loss": 1.2619, "step": 67 }, { "epoch": 0.6513020053876085, "grad_norm": 0.6958172104041147, "learning_rate": 4.769340210797618e-06, "loss": 1.2633, "step": 68 }, { "epoch": 0.6608799760550733, "grad_norm": 0.6532872905224688, "learning_rate": 4.757557257029563e-06, "loss": 1.2581, "step": 69 }, { "epoch": 0.6704579467225381, "grad_norm": 0.693714390508834, "learning_rate": 4.745496100659083e-06, "loss": 1.2499, "step": 70 }, { "epoch": 0.680035917390003, "grad_norm": 0.6749996898449282, "learning_rate": 4.733158228002891e-06, "loss": 1.2536, "step": 71 }, { "epoch": 0.6896138880574678, "grad_norm": 0.6753612400656019, "learning_rate": 4.720545159477921e-06, "loss": 1.2605, "step": 72 }, { "epoch": 0.6991918587249326, "grad_norm": 0.6950386791904168, "learning_rate": 4.707658449413961e-06, "loss": 1.2489, "step": 73 }, { "epoch": 0.7087698293923975, "grad_norm": 0.6396387112266337, "learning_rate": 4.694499685862106e-06, "loss": 1.264, "step": 74 }, { "epoch": 0.7183478000598623, "grad_norm": 0.6809655013846588, "learning_rate": 4.681070490399064e-06, "loss": 1.2477, "step": 75 }, { "epoch": 0.7279257707273271, "grad_norm": 0.6814836664342683, "learning_rate": 4.667372517927323e-06, "loss": 1.2349, "step": 76 }, { "epoch": 0.737503741394792, "grad_norm": 0.6502075268222723, "learning_rate": 4.653407456471222e-06, "loss": 1.243, "step": 77 }, { "epoch": 0.7470817120622568, "grad_norm": 0.6579341200451629, "learning_rate": 4.639177026968924e-06, "loss": 1.2549, "step": 78 }, { "epoch": 0.7470817120622568, "eval_loss": 2.2078425884246826, "eval_runtime": 107.0636, "eval_samples_per_second": 13.207, "eval_steps_per_second": 3.306, "step": 78 }, { "epoch": 0.7566596827297216, "grad_norm": 0.6264741505964025, "learning_rate": 4.624682983060346e-06, "loss": 1.2903, "step": 79 }, { "epoch": 0.7662376533971865, "grad_norm": 0.6533395420906253, "learning_rate": 4.609927110871053e-06, "loss": 1.2371, "step": 80 }, { "epoch": 0.7758156240646513, "grad_norm": 0.6366166912748572, "learning_rate": 4.594911228792156e-06, "loss": 1.2554, "step": 81 }, { "epoch": 0.7853935947321161, "grad_norm": 0.6435835690637465, "learning_rate": 4.579637187256222e-06, "loss": 1.2855, "step": 82 }, { "epoch": 0.7949715653995809, "grad_norm": 0.6410872090826751, "learning_rate": 4.564106868509246e-06, "loss": 1.232, "step": 83 }, { "epoch": 0.8045495360670458, "grad_norm": 0.6260242741257913, "learning_rate": 4.5483221863786965e-06, "loss": 1.2458, "step": 84 }, { "epoch": 0.8141275067345106, "grad_norm": 0.6588265965096135, "learning_rate": 4.5322850860376744e-06, "loss": 1.2474, "step": 85 }, { "epoch": 0.8237054774019754, "grad_norm": 0.6372013969893753, "learning_rate": 4.515997543765202e-06, "loss": 1.2563, "step": 86 }, { "epoch": 0.8332834480694403, "grad_norm": 0.683356686747451, "learning_rate": 4.499461566702685e-06, "loss": 1.2447, "step": 87 }, { "epoch": 0.8428614187369051, "grad_norm": 0.6520958114219059, "learning_rate": 4.48267919260657e-06, "loss": 1.2243, "step": 88 }, { "epoch": 0.8524393894043699, "grad_norm": 0.6468861797594448, "learning_rate": 4.465652489597226e-06, "loss": 1.2254, "step": 89 }, { "epoch": 0.8620173600718348, "grad_norm": 0.6675355862176291, "learning_rate": 4.4483835559040885e-06, "loss": 1.2116, "step": 90 }, { "epoch": 0.8715953307392996, "grad_norm": 0.6318507194646, "learning_rate": 4.430874519607089e-06, "loss": 1.2634, "step": 91 }, { "epoch": 0.8811733014067644, "grad_norm": 0.6496099541936005, "learning_rate": 4.413127538374411e-06, "loss": 1.2129, "step": 92 }, { "epoch": 0.8907512720742293, "grad_norm": 0.6026396711785842, "learning_rate": 4.395144799196593e-06, "loss": 1.2483, "step": 93 }, { "epoch": 0.9003292427416941, "grad_norm": 0.6709684350468395, "learning_rate": 4.376928518117028e-06, "loss": 1.2193, "step": 94 }, { "epoch": 0.9099072134091589, "grad_norm": 0.6237262552476821, "learning_rate": 4.358480939958867e-06, "loss": 1.218, "step": 95 }, { "epoch": 0.9194851840766237, "grad_norm": 0.6582242790059232, "learning_rate": 4.339804338048397e-06, "loss": 1.229, "step": 96 }, { "epoch": 0.9290631547440886, "grad_norm": 0.6235719312223321, "learning_rate": 4.320901013934887e-06, "loss": 1.2098, "step": 97 }, { "epoch": 0.9386411254115534, "grad_norm": 0.6295163336318428, "learning_rate": 4.301773297106968e-06, "loss": 1.205, "step": 98 }, { "epoch": 0.9482190960790182, "grad_norm": 0.6250959313071772, "learning_rate": 4.282423544705564e-06, "loss": 1.2054, "step": 99 }, { "epoch": 0.9577970667464831, "grad_norm": 0.6086898991547662, "learning_rate": 4.262854141233419e-06, "loss": 1.2118, "step": 100 }, { "epoch": 0.9673750374139479, "grad_norm": 0.5764067645719498, "learning_rate": 4.243067498261251e-06, "loss": 1.2372, "step": 101 }, { "epoch": 0.9769530080814127, "grad_norm": 0.6406315852737573, "learning_rate": 4.223066054130568e-06, "loss": 1.2251, "step": 102 }, { "epoch": 0.9865309787488776, "grad_norm": 0.5834984455673559, "learning_rate": 4.2028522736531895e-06, "loss": 1.2258, "step": 103 }, { "epoch": 0.9961089494163424, "grad_norm": 0.5911139350878512, "learning_rate": 4.182428647807503e-06, "loss": 1.2286, "step": 104 }, { "epoch": 0.9961089494163424, "eval_loss": 2.184576988220215, "eval_runtime": 107.5576, "eval_samples_per_second": 13.146, "eval_steps_per_second": 3.291, "step": 104 }, { "epoch": 1.0056869200838072, "grad_norm": 0.6299976497698655, "learning_rate": 4.161797693431493e-06, "loss": 1.2383, "step": 105 }, { "epoch": 1.002095181083508, "grad_norm": 0.5986176560633782, "learning_rate": 4.140961952912594e-06, "loss": 1.2182, "step": 106 }, { "epoch": 1.0116731517509727, "grad_norm": 0.7138997909374802, "learning_rate": 4.11992399387438e-06, "loss": 1.1894, "step": 107 }, { "epoch": 1.0212511224184375, "grad_norm": 0.6431525283411005, "learning_rate": 4.098686408860157e-06, "loss": 1.1741, "step": 108 }, { "epoch": 1.0308290930859023, "grad_norm": 0.7490910983392529, "learning_rate": 4.077251815013477e-06, "loss": 1.1849, "step": 109 }, { "epoch": 1.0404070637533673, "grad_norm": 0.6667698353299697, "learning_rate": 4.055622853755627e-06, "loss": 1.1833, "step": 110 }, { "epoch": 1.0499850344208321, "grad_norm": 0.7240102351414811, "learning_rate": 4.033802190460114e-06, "loss": 1.1915, "step": 111 }, { "epoch": 1.059563005088297, "grad_norm": 0.6281393232743739, "learning_rate": 4.011792514124217e-06, "loss": 1.1557, "step": 112 }, { "epoch": 1.0691409757557617, "grad_norm": 0.6735415717178005, "learning_rate": 3.989596537037608e-06, "loss": 1.1878, "step": 113 }, { "epoch": 1.0787189464232265, "grad_norm": 0.5939146155666697, "learning_rate": 3.967216994448116e-06, "loss": 1.1639, "step": 114 }, { "epoch": 1.0882969170906913, "grad_norm": 0.6932505538102671, "learning_rate": 3.9446566442246615e-06, "loss": 1.1759, "step": 115 }, { "epoch": 1.0978748877581563, "grad_norm": 0.5763908483496408, "learning_rate": 3.921918266517392e-06, "loss": 1.1781, "step": 116 }, { "epoch": 1.1074528584256211, "grad_norm": 0.6818836608860367, "learning_rate": 3.899004663415083e-06, "loss": 1.1869, "step": 117 }, { "epoch": 1.117030829093086, "grad_norm": 0.5998154432302447, "learning_rate": 3.875918658599837e-06, "loss": 1.1692, "step": 118 }, { "epoch": 1.1266087997605507, "grad_norm": 0.6596200288243683, "learning_rate": 3.852663096999104e-06, "loss": 1.2059, "step": 119 }, { "epoch": 1.1361867704280155, "grad_norm": 0.5918812335768482, "learning_rate": 3.829240844435109e-06, "loss": 1.1798, "step": 120 }, { "epoch": 1.1457647410954803, "grad_norm": 0.6232580849345692, "learning_rate": 3.8056547872716865e-06, "loss": 1.1517, "step": 121 }, { "epoch": 1.1553427117629451, "grad_norm": 0.5903843042319051, "learning_rate": 3.7819078320585865e-06, "loss": 1.1906, "step": 122 }, { "epoch": 1.1649206824304101, "grad_norm": 0.5896678764206408, "learning_rate": 3.7580029051732992e-06, "loss": 1.1832, "step": 123 }, { "epoch": 1.174498653097875, "grad_norm": 0.5666656027289849, "learning_rate": 3.733942952460432e-06, "loss": 1.1911, "step": 124 }, { "epoch": 1.1840766237653397, "grad_norm": 0.565358825737842, "learning_rate": 3.7097309388686865e-06, "loss": 1.1945, "step": 125 }, { "epoch": 1.1936545944328045, "grad_norm": 0.645159266559964, "learning_rate": 3.6853698480854853e-06, "loss": 1.1988, "step": 126 }, { "epoch": 1.2032325651002693, "grad_norm": 0.5661828443152349, "learning_rate": 3.660862682169283e-06, "loss": 1.1683, "step": 127 }, { "epoch": 1.2128105357677341, "grad_norm": 0.5590652900384634, "learning_rate": 3.636212461179623e-06, "loss": 1.1401, "step": 128 }, { "epoch": 1.2223885064351991, "grad_norm": 0.5772830186331369, "learning_rate": 3.6114222228049657e-06, "loss": 1.1457, "step": 129 }, { "epoch": 1.231966477102664, "grad_norm": 0.5638028162671672, "learning_rate": 3.5864950219883514e-06, "loss": 1.1599, "step": 130 }, { "epoch": 1.231966477102664, "eval_loss": 2.181441068649292, "eval_runtime": 107.4543, "eval_samples_per_second": 13.159, "eval_steps_per_second": 3.294, "step": 130 }, { "epoch": 1.2415444477701287, "grad_norm": 0.5674264221381613, "learning_rate": 3.561433930550934e-06, "loss": 1.1439, "step": 131 }, { "epoch": 1.2511224184375935, "grad_norm": 0.5548457286136358, "learning_rate": 3.536242036813436e-06, "loss": 1.1455, "step": 132 }, { "epoch": 1.2607003891050583, "grad_norm": 0.5681860545302818, "learning_rate": 3.510922445215568e-06, "loss": 1.1619, "step": 133 }, { "epoch": 1.2702783597725231, "grad_norm": 0.5189655726956113, "learning_rate": 3.4854782759334625e-06, "loss": 1.1647, "step": 134 }, { "epoch": 1.279856330439988, "grad_norm": 0.5482759127528988, "learning_rate": 3.4599126644951758e-06, "loss": 1.1963, "step": 135 }, { "epoch": 1.289434301107453, "grad_norm": 0.5545710145438582, "learning_rate": 3.4342287613942804e-06, "loss": 1.1673, "step": 136 }, { "epoch": 1.2990122717749177, "grad_norm": 0.5616911560631516, "learning_rate": 3.4084297317016353e-06, "loss": 1.1482, "step": 137 }, { "epoch": 1.3085902424423825, "grad_norm": 0.5429625311889626, "learning_rate": 3.3825187546753426e-06, "loss": 1.1459, "step": 138 }, { "epoch": 1.3181682131098473, "grad_norm": 0.5775738090552808, "learning_rate": 3.3564990233689632e-06, "loss": 1.1744, "step": 139 }, { "epoch": 1.3277461837773121, "grad_norm": 0.5422962267277087, "learning_rate": 3.330373744238033e-06, "loss": 1.1796, "step": 140 }, { "epoch": 1.3373241544447771, "grad_norm": 0.5383626495155892, "learning_rate": 3.3041461367449256e-06, "loss": 1.1646, "step": 141 }, { "epoch": 1.346902125112242, "grad_norm": 0.5588657340470299, "learning_rate": 3.2778194329621104e-06, "loss": 1.1842, "step": 142 }, { "epoch": 1.3564800957797067, "grad_norm": 0.5198196148369068, "learning_rate": 3.2513968771738606e-06, "loss": 1.1708, "step": 143 }, { "epoch": 1.3660580664471715, "grad_norm": 0.5453371169769571, "learning_rate": 3.224881725476456e-06, "loss": 1.1636, "step": 144 }, { "epoch": 1.3756360371146363, "grad_norm": 0.5692897944097868, "learning_rate": 3.198277245376924e-06, "loss": 1.1273, "step": 145 }, { "epoch": 1.3852140077821011, "grad_norm": 0.5423704486470122, "learning_rate": 3.1715867153903844e-06, "loss": 1.1405, "step": 146 }, { "epoch": 1.394791978449566, "grad_norm": 0.5819177408649716, "learning_rate": 3.144813424636031e-06, "loss": 1.1665, "step": 147 }, { "epoch": 1.4043699491170307, "grad_norm": 0.554870749454361, "learning_rate": 3.1179606724318052e-06, "loss": 1.1872, "step": 148 }, { "epoch": 1.4139479197844955, "grad_norm": 0.5493659769746441, "learning_rate": 3.091031767887817e-06, "loss": 1.1906, "step": 149 }, { "epoch": 1.4235258904519605, "grad_norm": 0.6008378552179591, "learning_rate": 3.0640300294985613e-06, "loss": 1.1635, "step": 150 }, { "epoch": 1.4331038611194253, "grad_norm": 0.5078261653762177, "learning_rate": 3.036958784733967e-06, "loss": 1.1438, "step": 151 }, { "epoch": 1.4426818317868901, "grad_norm": 0.5559592542323409, "learning_rate": 3.0098213696293542e-06, "loss": 1.1642, "step": 152 }, { "epoch": 1.452259802454355, "grad_norm": 0.5461821050739424, "learning_rate": 2.982621128374325e-06, "loss": 1.1725, "step": 153 }, { "epoch": 1.46183777312182, "grad_norm": 0.5412862095154186, "learning_rate": 2.9553614129006543e-06, "loss": 1.1654, "step": 154 }, { "epoch": 1.4714157437892847, "grad_norm": 0.5658659296771973, "learning_rate": 2.9280455824692255e-06, "loss": 1.1655, "step": 155 }, { "epoch": 1.4809937144567495, "grad_norm": 0.5525850336445564, "learning_rate": 2.9006770032560637e-06, "loss": 1.1577, "step": 156 }, { "epoch": 1.4809937144567495, "eval_loss": 2.1755869388580322, "eval_runtime": 107.2159, "eval_samples_per_second": 13.188, "eval_steps_per_second": 3.302, "step": 156 }, { "epoch": 1.4905716851242143, "grad_norm": 0.5710362202768258, "learning_rate": 2.8732590479375167e-06, "loss": 1.1595, "step": 157 }, { "epoch": 1.5001496557916791, "grad_norm": 0.5369626897696785, "learning_rate": 2.8457950952746293e-06, "loss": 1.1622, "step": 158 }, { "epoch": 1.509727626459144, "grad_norm": 0.5194143574454793, "learning_rate": 2.8182885296967833e-06, "loss": 1.1313, "step": 159 }, { "epoch": 1.5193055971266087, "grad_norm": 0.5220817246963333, "learning_rate": 2.7907427408846156e-06, "loss": 1.1493, "step": 160 }, { "epoch": 1.5288835677940735, "grad_norm": 0.5307538609855902, "learning_rate": 2.763161123352314e-06, "loss": 1.1571, "step": 161 }, { "epoch": 1.5384615384615383, "grad_norm": 0.5133921578064818, "learning_rate": 2.735547076029296e-06, "loss": 1.1398, "step": 162 }, { "epoch": 1.5480395091290033, "grad_norm": 0.528392253063443, "learning_rate": 2.7079040018413586e-06, "loss": 1.169, "step": 163 }, { "epoch": 1.5576174797964681, "grad_norm": 0.5033775123091357, "learning_rate": 2.6802353072913307e-06, "loss": 1.1396, "step": 164 }, { "epoch": 1.567195450463933, "grad_norm": 0.5429413779707357, "learning_rate": 2.6525444020392794e-06, "loss": 1.1558, "step": 165 }, { "epoch": 1.5767734211313977, "grad_norm": 0.5391198899526514, "learning_rate": 2.6248346984823325e-06, "loss": 1.1584, "step": 166 }, { "epoch": 1.5863513917988628, "grad_norm": 0.5237711725405991, "learning_rate": 2.5971096113341692e-06, "loss": 1.1399, "step": 167 }, { "epoch": 1.5959293624663276, "grad_norm": 0.522431379990406, "learning_rate": 2.5693725572042135e-06, "loss": 1.146, "step": 168 }, { "epoch": 1.6055073331337923, "grad_norm": 0.540451111257001, "learning_rate": 2.5416269541765963e-06, "loss": 1.1347, "step": 169 }, { "epoch": 1.6150853038012571, "grad_norm": 0.542474309771266, "learning_rate": 2.5138762213889493e-06, "loss": 1.1507, "step": 170 }, { "epoch": 1.624663274468722, "grad_norm": 0.5339716680549861, "learning_rate": 2.486123778611051e-06, "loss": 1.1428, "step": 171 }, { "epoch": 1.6342412451361867, "grad_norm": 0.5194346219437855, "learning_rate": 2.458373045823404e-06, "loss": 1.1717, "step": 172 }, { "epoch": 1.6438192158036515, "grad_norm": 0.5486922902738444, "learning_rate": 2.4306274427957878e-06, "loss": 1.1405, "step": 173 }, { "epoch": 1.6533971864711163, "grad_norm": 0.5364703724723029, "learning_rate": 2.402890388665831e-06, "loss": 1.1397, "step": 174 }, { "epoch": 1.6629751571385811, "grad_norm": 0.5151838009534813, "learning_rate": 2.375165301517668e-06, "loss": 1.1625, "step": 175 }, { "epoch": 1.6725531278060461, "grad_norm": 0.5387178228054901, "learning_rate": 2.3474555979607214e-06, "loss": 1.1586, "step": 176 }, { "epoch": 1.682131098473511, "grad_norm": 0.5264984610535657, "learning_rate": 2.3197646927086697e-06, "loss": 1.1654, "step": 177 }, { "epoch": 1.6917090691409757, "grad_norm": 0.5272357155280125, "learning_rate": 2.2920959981586426e-06, "loss": 1.1934, "step": 178 }, { "epoch": 1.7012870398084405, "grad_norm": 0.5252339989768573, "learning_rate": 2.2644529239707054e-06, "loss": 1.1426, "step": 179 }, { "epoch": 1.7108650104759056, "grad_norm": 0.4974185735061034, "learning_rate": 2.2368388766476875e-06, "loss": 1.1597, "step": 180 }, { "epoch": 1.7204429811433704, "grad_norm": 0.5361098970394095, "learning_rate": 2.2092572591153843e-06, "loss": 1.1637, "step": 181 }, { "epoch": 1.7300209518108352, "grad_norm": 0.5305009042993176, "learning_rate": 2.1817114703032176e-06, "loss": 1.1637, "step": 182 }, { "epoch": 1.7300209518108352, "eval_loss": 2.1710658073425293, "eval_runtime": 107.1212, "eval_samples_per_second": 13.2, "eval_steps_per_second": 3.305, "step": 182 }, { "epoch": 1.7395989224783, "grad_norm": 0.5012187069773779, "learning_rate": 2.154204904725371e-06, "loss": 1.1447, "step": 183 }, { "epoch": 1.7491768931457647, "grad_norm": 0.5763812037469009, "learning_rate": 2.126740952062484e-06, "loss": 1.1565, "step": 184 }, { "epoch": 1.7587548638132295, "grad_norm": 0.5129804478325861, "learning_rate": 2.099322996743936e-06, "loss": 1.1798, "step": 185 }, { "epoch": 1.7683328344806943, "grad_norm": 0.5107704085635135, "learning_rate": 2.0719544175307754e-06, "loss": 1.1486, "step": 186 }, { "epoch": 1.7779108051481591, "grad_norm": 0.5225266432128085, "learning_rate": 2.044638587099347e-06, "loss": 1.1457, "step": 187 }, { "epoch": 1.787488775815624, "grad_norm": 0.48553711881118367, "learning_rate": 2.0173788716256758e-06, "loss": 1.1557, "step": 188 }, { "epoch": 1.797066746483089, "grad_norm": 0.5155245524580911, "learning_rate": 1.9901786303706466e-06, "loss": 1.1667, "step": 189 }, { "epoch": 1.8066447171505537, "grad_norm": 0.5394238331941211, "learning_rate": 1.9630412152660333e-06, "loss": 1.1639, "step": 190 }, { "epoch": 1.8162226878180185, "grad_norm": 0.5208012650775928, "learning_rate": 1.93596997050144e-06, "loss": 1.167, "step": 191 }, { "epoch": 1.8258006584854833, "grad_norm": 0.5084683728452081, "learning_rate": 1.9089682321121834e-06, "loss": 1.146, "step": 192 }, { "epoch": 1.8353786291529484, "grad_norm": 0.5107216674575125, "learning_rate": 1.8820393275681954e-06, "loss": 1.1299, "step": 193 }, { "epoch": 1.8449565998204132, "grad_norm": 0.5037220655522233, "learning_rate": 1.8551865753639692e-06, "loss": 1.1705, "step": 194 }, { "epoch": 1.854534570487878, "grad_norm": 0.5081083073272432, "learning_rate": 1.8284132846096164e-06, "loss": 1.1232, "step": 195 }, { "epoch": 1.8641125411553428, "grad_norm": 0.4960779996118519, "learning_rate": 1.801722754623077e-06, "loss": 1.1356, "step": 196 }, { "epoch": 1.8736905118228075, "grad_norm": 0.5194537399766056, "learning_rate": 1.775118274523545e-06, "loss": 1.1321, "step": 197 }, { "epoch": 1.8832684824902723, "grad_norm": 0.5149057994299137, "learning_rate": 1.74860312282614e-06, "loss": 1.1306, "step": 198 }, { "epoch": 1.8928464531577371, "grad_norm": 0.5061127962699723, "learning_rate": 1.72218056703789e-06, "loss": 1.1302, "step": 199 }, { "epoch": 1.902424423825202, "grad_norm": 0.49704736224795454, "learning_rate": 1.6958538632550753e-06, "loss": 1.1479, "step": 200 }, { "epoch": 1.9120023944926667, "grad_norm": 0.4976492539596855, "learning_rate": 1.6696262557619677e-06, "loss": 1.135, "step": 201 }, { "epoch": 1.9215803651601315, "grad_norm": 0.5438558597014863, "learning_rate": 1.6435009766310372e-06, "loss": 1.1677, "step": 202 }, { "epoch": 1.9311583358275966, "grad_norm": 0.49386649254244525, "learning_rate": 1.6174812453246582e-06, "loss": 1.1396, "step": 203 }, { "epoch": 1.9407363064950613, "grad_norm": 0.5039832884638089, "learning_rate": 1.5915702682983657e-06, "loss": 1.1857, "step": 204 }, { "epoch": 1.9503142771625261, "grad_norm": 0.4892382263387271, "learning_rate": 1.5657712386057202e-06, "loss": 1.15, "step": 205 }, { "epoch": 1.9598922478299912, "grad_norm": 0.5084631284159544, "learning_rate": 1.5400873355048248e-06, "loss": 1.1572, "step": 206 }, { "epoch": 1.969470218497456, "grad_norm": 0.5008750617477549, "learning_rate": 1.5145217240665373e-06, "loss": 1.1326, "step": 207 }, { "epoch": 1.9790481891649208, "grad_norm": 0.4980386882470781, "learning_rate": 1.489077554784432e-06, "loss": 1.143, "step": 208 }, { "epoch": 1.9790481891649208, "eval_loss": 2.1687815189361572, "eval_runtime": 107.1708, "eval_samples_per_second": 13.194, "eval_steps_per_second": 3.303, "step": 208 }, { "epoch": 1.9886261598323856, "grad_norm": 0.4895688225344272, "learning_rate": 1.4637579631865645e-06, "loss": 1.1171, "step": 209 }, { "epoch": 1.9982041304998504, "grad_norm": 0.49262081512228967, "learning_rate": 1.4385660694490667e-06, "loss": 1.1449, "step": 210 }, { "epoch": 2.007782101167315, "grad_norm": 0.5057383346810608, "learning_rate": 1.4135049780116496e-06, "loss": 1.1394, "step": 211 }, { "epoch": 2.0038910505836576, "grad_norm": 0.5387584817585892, "learning_rate": 1.388577777195035e-06, "loss": 1.1306, "step": 212 }, { "epoch": 2.0134690212511224, "grad_norm": 0.5623404364476285, "learning_rate": 1.3637875388203784e-06, "loss": 1.0952, "step": 213 }, { "epoch": 2.023046991918587, "grad_norm": 0.5743034832238124, "learning_rate": 1.3391373178307182e-06, "loss": 1.1261, "step": 214 }, { "epoch": 2.032624962586052, "grad_norm": 0.5461858778537674, "learning_rate": 1.3146301519145153e-06, "loss": 1.1328, "step": 215 }, { "epoch": 2.0422029332535168, "grad_norm": 0.5528333288756201, "learning_rate": 1.2902690611313135e-06, "loss": 1.1249, "step": 216 }, { "epoch": 2.0517809039209816, "grad_norm": 0.5258934842101934, "learning_rate": 1.2660570475395684e-06, "loss": 1.1109, "step": 217 }, { "epoch": 2.0613588745884464, "grad_norm": 0.5524292613274455, "learning_rate": 1.2419970948267014e-06, "loss": 1.1135, "step": 218 }, { "epoch": 2.0709368452559116, "grad_norm": 0.5405228294413486, "learning_rate": 1.2180921679414143e-06, "loss": 1.1287, "step": 219 }, { "epoch": 2.0805148159233764, "grad_norm": 0.5298775138689613, "learning_rate": 1.1943452127283145e-06, "loss": 1.124, "step": 220 }, { "epoch": 2.090092786590841, "grad_norm": 0.514214942457388, "learning_rate": 1.1707591555648905e-06, "loss": 1.1059, "step": 221 }, { "epoch": 2.099670757258306, "grad_norm": 0.5329396149825425, "learning_rate": 1.1473369030008974e-06, "loss": 1.1201, "step": 222 }, { "epoch": 2.109248727925771, "grad_norm": 0.5564862124718808, "learning_rate": 1.124081341400165e-06, "loss": 1.1032, "step": 223 }, { "epoch": 2.1188266985932356, "grad_norm": 0.5244468629630417, "learning_rate": 1.1009953365849168e-06, "loss": 1.1433, "step": 224 }, { "epoch": 2.1284046692607004, "grad_norm": 0.5087349968174719, "learning_rate": 1.078081733482609e-06, "loss": 1.1286, "step": 225 }, { "epoch": 2.137982639928165, "grad_norm": 0.522473732751717, "learning_rate": 1.055343355775339e-06, "loss": 1.084, "step": 226 }, { "epoch": 2.14756061059563, "grad_norm": 0.5213841410982886, "learning_rate": 1.0327830055518843e-06, "loss": 1.0778, "step": 227 }, { "epoch": 2.1571385812630948, "grad_norm": 0.5211792543694728, "learning_rate": 1.0104034629623933e-06, "loss": 1.0892, "step": 228 }, { "epoch": 2.1667165519305596, "grad_norm": 0.5366639328996056, "learning_rate": 9.88207485875784e-07, "loss": 1.1129, "step": 229 }, { "epoch": 2.1762945225980244, "grad_norm": 0.5072189940995689, "learning_rate": 9.661978095398854e-07, "loss": 1.1124, "step": 230 }, { "epoch": 2.1858724932654896, "grad_norm": 0.5273739329980739, "learning_rate": 9.443771462443743e-07, "loss": 1.0966, "step": 231 }, { "epoch": 2.1954504639329544, "grad_norm": 0.530434300883332, "learning_rate": 9.227481849865236e-07, "loss": 1.121, "step": 232 }, { "epoch": 2.205028434600419, "grad_norm": 0.49620125772094664, "learning_rate": 9.013135911398435e-07, "loss": 1.1227, "step": 233 }, { "epoch": 2.214606405267884, "grad_norm": 0.48930931831635505, "learning_rate": 8.800760061256205e-07, "loss": 1.1249, "step": 234 }, { "epoch": 2.214606405267884, "eval_loss": 2.177833318710327, "eval_runtime": 106.9928, "eval_samples_per_second": 13.216, "eval_steps_per_second": 3.309, "step": 234 }, { "epoch": 2.224184375935349, "grad_norm": 0.5117030753774101, "learning_rate": 8.590380470874066e-07, "loss": 1.0983, "step": 235 }, { "epoch": 2.2337623466028136, "grad_norm": 0.5334281898363374, "learning_rate": 8.382023065685071e-07, "loss": 1.1058, "step": 236 }, { "epoch": 2.2433403172702784, "grad_norm": 0.4997549069918058, "learning_rate": 8.175713521924977e-07, "loss": 1.1205, "step": 237 }, { "epoch": 2.252918287937743, "grad_norm": 0.4903764233470244, "learning_rate": 7.971477263468108e-07, "loss": 1.1166, "step": 238 }, { "epoch": 2.262496258605208, "grad_norm": 0.5111886828961109, "learning_rate": 7.769339458694319e-07, "loss": 1.1296, "step": 239 }, { "epoch": 2.2720742292726728, "grad_norm": 0.5046245576610761, "learning_rate": 7.569325017387502e-07, "loss": 1.1214, "step": 240 }, { "epoch": 2.2816521999401376, "grad_norm": 0.5012727372502416, "learning_rate": 7.371458587665822e-07, "loss": 1.1282, "step": 241 }, { "epoch": 2.2912301706076024, "grad_norm": 0.5089746600647955, "learning_rate": 7.175764552944368e-07, "loss": 1.1228, "step": 242 }, { "epoch": 2.300808141275067, "grad_norm": 0.49011501775043553, "learning_rate": 6.982267028930326e-07, "loss": 1.1019, "step": 243 }, { "epoch": 2.310386111942532, "grad_norm": 0.5062866494664521, "learning_rate": 6.790989860651143e-07, "loss": 1.1237, "step": 244 }, { "epoch": 2.3199640826099968, "grad_norm": 0.48914824725834716, "learning_rate": 6.601956619516037e-07, "loss": 1.1228, "step": 245 }, { "epoch": 2.329542053277462, "grad_norm": 0.500095846054479, "learning_rate": 6.41519060041134e-07, "loss": 1.0725, "step": 246 }, { "epoch": 2.339120023944927, "grad_norm": 0.48427264883155136, "learning_rate": 6.230714818829733e-07, "loss": 1.116, "step": 247 }, { "epoch": 2.3486979946123916, "grad_norm": 0.5009855645248527, "learning_rate": 6.048552008034073e-07, "loss": 1.1158, "step": 248 }, { "epoch": 2.3582759652798564, "grad_norm": 0.4895310310383359, "learning_rate": 5.868724616255899e-07, "loss": 1.1134, "step": 249 }, { "epoch": 2.367853935947321, "grad_norm": 0.49721503448947285, "learning_rate": 5.691254803929117e-07, "loss": 1.1178, "step": 250 }, { "epoch": 2.377431906614786, "grad_norm": 0.4908749278467018, "learning_rate": 5.516164440959118e-07, "loss": 1.0965, "step": 251 }, { "epoch": 2.387009877282251, "grad_norm": 0.49404150582673295, "learning_rate": 5.343475104027743e-07, "loss": 1.1299, "step": 252 }, { "epoch": 2.3965878479497156, "grad_norm": 0.4824591396857287, "learning_rate": 5.17320807393431e-07, "loss": 1.0795, "step": 253 }, { "epoch": 2.4061658186171804, "grad_norm": 0.49129116007089907, "learning_rate": 5.005384332973154e-07, "loss": 1.1193, "step": 254 }, { "epoch": 2.415743789284645, "grad_norm": 0.49733333626674653, "learning_rate": 4.840024562347987e-07, "loss": 1.11, "step": 255 }, { "epoch": 2.42532175995211, "grad_norm": 0.4755146663348369, "learning_rate": 4.67714913962326e-07, "loss": 1.1091, "step": 256 }, { "epoch": 2.434899730619575, "grad_norm": 0.49183045855158936, "learning_rate": 4.5167781362130374e-07, "loss": 1.1247, "step": 257 }, { "epoch": 2.44447770128704, "grad_norm": 0.48090622566109875, "learning_rate": 4.3589313149075495e-07, "loss": 1.0957, "step": 258 }, { "epoch": 2.454055671954505, "grad_norm": 0.49785336365870875, "learning_rate": 4.2036281274377865e-07, "loss": 1.1139, "step": 259 }, { "epoch": 2.4636336426219696, "grad_norm": 0.49485716633378346, "learning_rate": 4.050887712078444e-07, "loss": 1.1298, "step": 260 }, { "epoch": 2.4636336426219696, "eval_loss": 2.177307367324829, "eval_runtime": 107.0849, "eval_samples_per_second": 13.204, "eval_steps_per_second": 3.306, "step": 260 } ], "logging_steps": 1, "max_steps": 312, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 52, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 6.246866760855716e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }