{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.990024577128813, "eval_steps": 500, "global_step": 648, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004626283070695388, "grad_norm": 0.6155030131340027, "learning_rate": 9.98456790123457e-06, "loss": 17.5936, "step": 1 }, { "epoch": 0.009252566141390776, "grad_norm": 0.6313583254814148, "learning_rate": 9.969135802469136e-06, "loss": 21.7773, "step": 2 }, { "epoch": 0.013878849212086165, "grad_norm": 0.4570765793323517, "learning_rate": 9.953703703703704e-06, "loss": 17.1664, "step": 3 }, { "epoch": 0.018505132282781552, "grad_norm": 0.3836732506752014, "learning_rate": 9.938271604938273e-06, "loss": 18.3922, "step": 4 }, { "epoch": 0.02313141535347694, "grad_norm": 0.3599857985973358, "learning_rate": 9.92283950617284e-06, "loss": 16.9993, "step": 5 }, { "epoch": 0.02775769842417233, "grad_norm": 0.4342029094696045, "learning_rate": 9.907407407407408e-06, "loss": 17.4883, "step": 6 }, { "epoch": 0.032383981494867715, "grad_norm": 0.40308186411857605, "learning_rate": 9.891975308641975e-06, "loss": 18.2562, "step": 7 }, { "epoch": 0.037010264565563104, "grad_norm": 0.39312881231307983, "learning_rate": 9.876543209876543e-06, "loss": 17.9318, "step": 8 }, { "epoch": 0.04163654763625849, "grad_norm": 0.6749095916748047, "learning_rate": 9.861111111111112e-06, "loss": 18.5244, "step": 9 }, { "epoch": 0.04626283070695388, "grad_norm": 0.3622874617576599, "learning_rate": 9.84567901234568e-06, "loss": 16.4386, "step": 10 }, { "epoch": 0.05088911377764927, "grad_norm": 0.24345389008522034, "learning_rate": 9.830246913580249e-06, "loss": 17.5414, "step": 11 }, { "epoch": 0.05551539684834466, "grad_norm": 0.34016355872154236, "learning_rate": 9.814814814814815e-06, "loss": 16.9405, "step": 12 }, { "epoch": 0.06014167991904005, "grad_norm": 0.25097739696502686, "learning_rate": 9.799382716049384e-06, "loss": 14.4692, "step": 13 }, { "epoch": 0.06476796298973543, "grad_norm": 0.2524467408657074, "learning_rate": 9.783950617283952e-06, "loss": 16.2185, "step": 14 }, { "epoch": 0.06939424606043082, "grad_norm": 0.3705744743347168, "learning_rate": 9.768518518518519e-06, "loss": 16.3573, "step": 15 }, { "epoch": 0.07402052913112621, "grad_norm": 0.2848168909549713, "learning_rate": 9.753086419753087e-06, "loss": 16.331, "step": 16 }, { "epoch": 0.0786468122018216, "grad_norm": 0.300182044506073, "learning_rate": 9.737654320987654e-06, "loss": 15.6623, "step": 17 }, { "epoch": 0.08327309527251699, "grad_norm": 0.36958807706832886, "learning_rate": 9.722222222222223e-06, "loss": 18.5857, "step": 18 }, { "epoch": 0.08789937834321238, "grad_norm": 0.3009611666202545, "learning_rate": 9.706790123456791e-06, "loss": 17.0655, "step": 19 }, { "epoch": 0.09252566141390776, "grad_norm": 0.24450023472309113, "learning_rate": 9.691358024691358e-06, "loss": 15.3945, "step": 20 }, { "epoch": 0.09715194448460315, "grad_norm": 0.3174358010292053, "learning_rate": 9.675925925925926e-06, "loss": 15.3045, "step": 21 }, { "epoch": 0.10177822755529854, "grad_norm": 0.34466007351875305, "learning_rate": 9.660493827160495e-06, "loss": 16.7186, "step": 22 }, { "epoch": 0.10640451062599393, "grad_norm": 0.294209361076355, "learning_rate": 9.645061728395062e-06, "loss": 16.2893, "step": 23 }, { "epoch": 0.11103079369668932, "grad_norm": 0.3418446183204651, "learning_rate": 9.62962962962963e-06, "loss": 16.1827, "step": 24 }, { "epoch": 0.11565707676738471, "grad_norm": 0.2584611177444458, "learning_rate": 9.614197530864198e-06, "loss": 15.7934, "step": 25 }, { "epoch": 0.1202833598380801, "grad_norm": 0.3344869911670685, "learning_rate": 9.598765432098767e-06, "loss": 14.2338, "step": 26 }, { "epoch": 0.12490964290877549, "grad_norm": 0.28123536705970764, "learning_rate": 9.583333333333335e-06, "loss": 16.2229, "step": 27 }, { "epoch": 0.12953592597947086, "grad_norm": 0.22181656956672668, "learning_rate": 9.567901234567902e-06, "loss": 13.7744, "step": 28 }, { "epoch": 0.13416220905016626, "grad_norm": 0.45139142870903015, "learning_rate": 9.55246913580247e-06, "loss": 15.4782, "step": 29 }, { "epoch": 0.13878849212086164, "grad_norm": 0.21577508747577667, "learning_rate": 9.537037037037037e-06, "loss": 15.6752, "step": 30 }, { "epoch": 0.14341477519155704, "grad_norm": 0.29512590169906616, "learning_rate": 9.521604938271606e-06, "loss": 14.7965, "step": 31 }, { "epoch": 0.14804105826225242, "grad_norm": 0.28356751799583435, "learning_rate": 9.506172839506174e-06, "loss": 14.9097, "step": 32 }, { "epoch": 0.15266734133294782, "grad_norm": 0.25781649351119995, "learning_rate": 9.490740740740741e-06, "loss": 14.494, "step": 33 }, { "epoch": 0.1572936244036432, "grad_norm": 0.1967274248600006, "learning_rate": 9.47530864197531e-06, "loss": 15.0075, "step": 34 }, { "epoch": 0.1619199074743386, "grad_norm": 0.42894116044044495, "learning_rate": 9.459876543209878e-06, "loss": 14.8055, "step": 35 }, { "epoch": 0.16654619054503397, "grad_norm": 0.2417590320110321, "learning_rate": 9.444444444444445e-06, "loss": 14.7145, "step": 36 }, { "epoch": 0.17117247361572938, "grad_norm": 0.15463685989379883, "learning_rate": 9.429012345679013e-06, "loss": 14.4273, "step": 37 }, { "epoch": 0.17579875668642475, "grad_norm": 0.21183425188064575, "learning_rate": 9.413580246913581e-06, "loss": 16.6545, "step": 38 }, { "epoch": 0.18042503975712013, "grad_norm": 0.41926464438438416, "learning_rate": 9.398148148148148e-06, "loss": 13.4609, "step": 39 }, { "epoch": 0.18505132282781553, "grad_norm": 0.21585555374622345, "learning_rate": 9.382716049382717e-06, "loss": 14.1515, "step": 40 }, { "epoch": 0.1896776058985109, "grad_norm": 0.28051456809043884, "learning_rate": 9.367283950617285e-06, "loss": 14.0035, "step": 41 }, { "epoch": 0.1943038889692063, "grad_norm": 0.18281345069408417, "learning_rate": 9.351851851851854e-06, "loss": 14.6411, "step": 42 }, { "epoch": 0.19893017203990168, "grad_norm": 0.1747712790966034, "learning_rate": 9.33641975308642e-06, "loss": 14.4986, "step": 43 }, { "epoch": 0.20355645511059708, "grad_norm": 0.2613980174064636, "learning_rate": 9.320987654320989e-06, "loss": 14.2572, "step": 44 }, { "epoch": 0.20818273818129246, "grad_norm": 0.19127604365348816, "learning_rate": 9.305555555555557e-06, "loss": 14.3084, "step": 45 }, { "epoch": 0.21280902125198786, "grad_norm": 0.24545663595199585, "learning_rate": 9.290123456790124e-06, "loss": 13.3381, "step": 46 }, { "epoch": 0.21743530432268324, "grad_norm": 0.25891631841659546, "learning_rate": 9.274691358024692e-06, "loss": 13.1076, "step": 47 }, { "epoch": 0.22206158739337864, "grad_norm": 0.23256707191467285, "learning_rate": 9.25925925925926e-06, "loss": 13.3061, "step": 48 }, { "epoch": 0.22668787046407401, "grad_norm": 0.2427562177181244, "learning_rate": 9.243827160493828e-06, "loss": 16.0646, "step": 49 }, { "epoch": 0.23131415353476942, "grad_norm": 0.22760462760925293, "learning_rate": 9.228395061728396e-06, "loss": 12.937, "step": 50 }, { "epoch": 0.2359404366054648, "grad_norm": 0.27145224809646606, "learning_rate": 9.212962962962963e-06, "loss": 13.8299, "step": 51 }, { "epoch": 0.2405667196761602, "grad_norm": 0.3120661973953247, "learning_rate": 9.197530864197531e-06, "loss": 14.074, "step": 52 }, { "epoch": 0.24519300274685557, "grad_norm": 0.21604514122009277, "learning_rate": 9.1820987654321e-06, "loss": 13.6731, "step": 53 }, { "epoch": 0.24981928581755097, "grad_norm": 0.22970032691955566, "learning_rate": 9.166666666666666e-06, "loss": 14.0123, "step": 54 }, { "epoch": 0.2544455688882464, "grad_norm": 0.24079963564872742, "learning_rate": 9.151234567901235e-06, "loss": 13.5458, "step": 55 }, { "epoch": 0.2590718519589417, "grad_norm": 0.21451319754123688, "learning_rate": 9.135802469135803e-06, "loss": 12.5504, "step": 56 }, { "epoch": 0.2636981350296371, "grad_norm": 0.2508305609226227, "learning_rate": 9.120370370370372e-06, "loss": 14.8642, "step": 57 }, { "epoch": 0.26832441810033253, "grad_norm": 0.24022799730300903, "learning_rate": 9.10493827160494e-06, "loss": 15.1011, "step": 58 }, { "epoch": 0.2729507011710279, "grad_norm": 0.3794403076171875, "learning_rate": 9.089506172839507e-06, "loss": 14.3993, "step": 59 }, { "epoch": 0.2775769842417233, "grad_norm": 0.28563258051872253, "learning_rate": 9.074074074074075e-06, "loss": 14.294, "step": 60 }, { "epoch": 0.2822032673124187, "grad_norm": 0.28061386942863464, "learning_rate": 9.058641975308642e-06, "loss": 13.033, "step": 61 }, { "epoch": 0.2868295503831141, "grad_norm": 0.3351198136806488, "learning_rate": 9.04320987654321e-06, "loss": 14.5075, "step": 62 }, { "epoch": 0.29145583345380943, "grad_norm": 0.23749324679374695, "learning_rate": 9.027777777777779e-06, "loss": 13.7952, "step": 63 }, { "epoch": 0.29608211652450483, "grad_norm": 0.16919538378715515, "learning_rate": 9.012345679012346e-06, "loss": 14.7146, "step": 64 }, { "epoch": 0.30070839959520024, "grad_norm": 0.18566973507404327, "learning_rate": 8.996913580246914e-06, "loss": 13.9967, "step": 65 }, { "epoch": 0.30533468266589564, "grad_norm": 0.25144919753074646, "learning_rate": 8.981481481481483e-06, "loss": 13.0544, "step": 66 }, { "epoch": 0.309960965736591, "grad_norm": 0.1971070021390915, "learning_rate": 8.96604938271605e-06, "loss": 12.7932, "step": 67 }, { "epoch": 0.3145872488072864, "grad_norm": 0.2236046940088272, "learning_rate": 8.950617283950618e-06, "loss": 14.4338, "step": 68 }, { "epoch": 0.3192135318779818, "grad_norm": 0.20062977075576782, "learning_rate": 8.935185185185186e-06, "loss": 13.6965, "step": 69 }, { "epoch": 0.3238398149486772, "grad_norm": 0.3102545738220215, "learning_rate": 8.919753086419753e-06, "loss": 14.174, "step": 70 }, { "epoch": 0.32846609801937254, "grad_norm": 0.282172828912735, "learning_rate": 8.904320987654322e-06, "loss": 14.012, "step": 71 }, { "epoch": 0.33309238109006795, "grad_norm": 0.31374603509902954, "learning_rate": 8.888888888888888e-06, "loss": 12.9434, "step": 72 }, { "epoch": 0.33771866416076335, "grad_norm": 0.24876756966114044, "learning_rate": 8.873456790123458e-06, "loss": 13.0801, "step": 73 }, { "epoch": 0.34234494723145875, "grad_norm": 0.2828700840473175, "learning_rate": 8.858024691358025e-06, "loss": 12.3676, "step": 74 }, { "epoch": 0.3469712303021541, "grad_norm": 0.23395580053329468, "learning_rate": 8.842592592592594e-06, "loss": 12.8463, "step": 75 }, { "epoch": 0.3515975133728495, "grad_norm": 0.26694121956825256, "learning_rate": 8.827160493827162e-06, "loss": 11.5822, "step": 76 }, { "epoch": 0.3562237964435449, "grad_norm": 0.2129782736301422, "learning_rate": 8.811728395061729e-06, "loss": 12.6834, "step": 77 }, { "epoch": 0.36085007951424025, "grad_norm": 0.20400603115558624, "learning_rate": 8.796296296296297e-06, "loss": 14.2065, "step": 78 }, { "epoch": 0.36547636258493565, "grad_norm": 0.45560839772224426, "learning_rate": 8.780864197530866e-06, "loss": 13.1033, "step": 79 }, { "epoch": 0.37010264565563106, "grad_norm": 0.26329922676086426, "learning_rate": 8.765432098765432e-06, "loss": 11.9119, "step": 80 }, { "epoch": 0.37472892872632646, "grad_norm": 0.21817731857299805, "learning_rate": 8.750000000000001e-06, "loss": 12.815, "step": 81 }, { "epoch": 0.3793552117970218, "grad_norm": 0.2743465304374695, "learning_rate": 8.73456790123457e-06, "loss": 12.2468, "step": 82 }, { "epoch": 0.3839814948677172, "grad_norm": 0.2570422291755676, "learning_rate": 8.719135802469136e-06, "loss": 11.9801, "step": 83 }, { "epoch": 0.3886077779384126, "grad_norm": 0.21237581968307495, "learning_rate": 8.703703703703705e-06, "loss": 11.7092, "step": 84 }, { "epoch": 0.393234061009108, "grad_norm": 0.27444881200790405, "learning_rate": 8.688271604938271e-06, "loss": 13.3074, "step": 85 }, { "epoch": 0.39786034407980336, "grad_norm": 0.22574838995933533, "learning_rate": 8.67283950617284e-06, "loss": 13.1573, "step": 86 }, { "epoch": 0.40248662715049877, "grad_norm": 0.252755731344223, "learning_rate": 8.657407407407408e-06, "loss": 12.3472, "step": 87 }, { "epoch": 0.40711291022119417, "grad_norm": 0.2729659378528595, "learning_rate": 8.641975308641975e-06, "loss": 12.8413, "step": 88 }, { "epoch": 0.41173919329188957, "grad_norm": 0.21728309988975525, "learning_rate": 8.626543209876543e-06, "loss": 11.3523, "step": 89 }, { "epoch": 0.4163654763625849, "grad_norm": 0.264457106590271, "learning_rate": 8.611111111111112e-06, "loss": 13.1175, "step": 90 }, { "epoch": 0.4209917594332803, "grad_norm": 0.20264536142349243, "learning_rate": 8.59567901234568e-06, "loss": 12.8442, "step": 91 }, { "epoch": 0.4256180425039757, "grad_norm": 0.3003804385662079, "learning_rate": 8.580246913580249e-06, "loss": 14.5755, "step": 92 }, { "epoch": 0.4302443255746711, "grad_norm": 0.294029176235199, "learning_rate": 8.564814814814816e-06, "loss": 13.5777, "step": 93 }, { "epoch": 0.4348706086453665, "grad_norm": 0.21502816677093506, "learning_rate": 8.549382716049384e-06, "loss": 11.9804, "step": 94 }, { "epoch": 0.4394968917160619, "grad_norm": 0.21898359060287476, "learning_rate": 8.53395061728395e-06, "loss": 13.6382, "step": 95 }, { "epoch": 0.4441231747867573, "grad_norm": 0.20010338723659515, "learning_rate": 8.518518518518519e-06, "loss": 12.4733, "step": 96 }, { "epoch": 0.4487494578574526, "grad_norm": 0.2247803956270218, "learning_rate": 8.503086419753088e-06, "loss": 12.1589, "step": 97 }, { "epoch": 0.45337574092814803, "grad_norm": 0.29326918721199036, "learning_rate": 8.487654320987654e-06, "loss": 12.8177, "step": 98 }, { "epoch": 0.45800202399884343, "grad_norm": 0.18403670191764832, "learning_rate": 8.472222222222223e-06, "loss": 13.8527, "step": 99 }, { "epoch": 0.46262830706953884, "grad_norm": 0.24531710147857666, "learning_rate": 8.456790123456791e-06, "loss": 11.9126, "step": 100 }, { "epoch": 0.4672545901402342, "grad_norm": 0.24183672666549683, "learning_rate": 8.441358024691358e-06, "loss": 12.952, "step": 101 }, { "epoch": 0.4718808732109296, "grad_norm": 0.29563647508621216, "learning_rate": 8.425925925925926e-06, "loss": 14.462, "step": 102 }, { "epoch": 0.476507156281625, "grad_norm": 0.21156498789787292, "learning_rate": 8.410493827160495e-06, "loss": 13.4846, "step": 103 }, { "epoch": 0.4811334393523204, "grad_norm": 0.24183927476406097, "learning_rate": 8.395061728395062e-06, "loss": 11.428, "step": 104 }, { "epoch": 0.48575972242301574, "grad_norm": 0.21241174638271332, "learning_rate": 8.37962962962963e-06, "loss": 13.8068, "step": 105 }, { "epoch": 0.49038600549371114, "grad_norm": 0.2149106115102768, "learning_rate": 8.364197530864199e-06, "loss": 12.7659, "step": 106 }, { "epoch": 0.49501228856440654, "grad_norm": 0.21057994663715363, "learning_rate": 8.348765432098767e-06, "loss": 12.4651, "step": 107 }, { "epoch": 0.49963857163510195, "grad_norm": 0.18793097138404846, "learning_rate": 8.333333333333334e-06, "loss": 11.7854, "step": 108 }, { "epoch": 0.5042648547057973, "grad_norm": 0.1801634281873703, "learning_rate": 8.317901234567902e-06, "loss": 12.9306, "step": 109 }, { "epoch": 0.5088911377764928, "grad_norm": 0.21523930132389069, "learning_rate": 8.30246913580247e-06, "loss": 11.6702, "step": 110 }, { "epoch": 0.5135174208471881, "grad_norm": 0.22931678593158722, "learning_rate": 8.287037037037037e-06, "loss": 11.5856, "step": 111 }, { "epoch": 0.5181437039178834, "grad_norm": 0.23802338540554047, "learning_rate": 8.271604938271606e-06, "loss": 12.7508, "step": 112 }, { "epoch": 0.5227699869885789, "grad_norm": 0.23633763194084167, "learning_rate": 8.256172839506174e-06, "loss": 12.0906, "step": 113 }, { "epoch": 0.5273962700592743, "grad_norm": 0.22553692758083344, "learning_rate": 8.240740740740741e-06, "loss": 12.4982, "step": 114 }, { "epoch": 0.5320225531299696, "grad_norm": 0.2265913337469101, "learning_rate": 8.22530864197531e-06, "loss": 11.4945, "step": 115 }, { "epoch": 0.5366488362006651, "grad_norm": 0.1848449558019638, "learning_rate": 8.209876543209876e-06, "loss": 11.866, "step": 116 }, { "epoch": 0.5412751192713604, "grad_norm": 0.16533511877059937, "learning_rate": 8.194444444444445e-06, "loss": 12.3876, "step": 117 }, { "epoch": 0.5459014023420558, "grad_norm": 0.19875134527683258, "learning_rate": 8.179012345679013e-06, "loss": 12.0186, "step": 118 }, { "epoch": 0.5505276854127512, "grad_norm": 0.20982632040977478, "learning_rate": 8.16358024691358e-06, "loss": 13.1202, "step": 119 }, { "epoch": 0.5551539684834466, "grad_norm": 0.2689734101295471, "learning_rate": 8.148148148148148e-06, "loss": 13.9764, "step": 120 }, { "epoch": 0.559780251554142, "grad_norm": 0.14883318543434143, "learning_rate": 8.132716049382717e-06, "loss": 11.7656, "step": 121 }, { "epoch": 0.5644065346248374, "grad_norm": 0.18115845322608948, "learning_rate": 8.117283950617285e-06, "loss": 11.8902, "step": 122 }, { "epoch": 0.5690328176955327, "grad_norm": 0.23553551733493805, "learning_rate": 8.101851851851854e-06, "loss": 13.0425, "step": 123 }, { "epoch": 0.5736591007662282, "grad_norm": 0.2110109180212021, "learning_rate": 8.08641975308642e-06, "loss": 13.6931, "step": 124 }, { "epoch": 0.5782853838369235, "grad_norm": 0.2748431861400604, "learning_rate": 8.070987654320989e-06, "loss": 12.7852, "step": 125 }, { "epoch": 0.5829116669076189, "grad_norm": 0.1847638040781021, "learning_rate": 8.055555555555557e-06, "loss": 12.9106, "step": 126 }, { "epoch": 0.5875379499783143, "grad_norm": 0.48550117015838623, "learning_rate": 8.040123456790124e-06, "loss": 11.5484, "step": 127 }, { "epoch": 0.5921642330490097, "grad_norm": 0.2579441964626312, "learning_rate": 8.024691358024692e-06, "loss": 11.9785, "step": 128 }, { "epoch": 0.5967905161197051, "grad_norm": 0.19037479162216187, "learning_rate": 8.00925925925926e-06, "loss": 11.0466, "step": 129 }, { "epoch": 0.6014167991904005, "grad_norm": 0.22901126742362976, "learning_rate": 7.993827160493828e-06, "loss": 11.1139, "step": 130 }, { "epoch": 0.6060430822610958, "grad_norm": 0.16406911611557007, "learning_rate": 7.978395061728396e-06, "loss": 10.6613, "step": 131 }, { "epoch": 0.6106693653317913, "grad_norm": 0.23492039740085602, "learning_rate": 7.962962962962963e-06, "loss": 11.3782, "step": 132 }, { "epoch": 0.6152956484024866, "grad_norm": 0.21692270040512085, "learning_rate": 7.947530864197531e-06, "loss": 11.2788, "step": 133 }, { "epoch": 0.619921931473182, "grad_norm": 0.17832407355308533, "learning_rate": 7.9320987654321e-06, "loss": 11.7617, "step": 134 }, { "epoch": 0.6245482145438774, "grad_norm": 0.20935167372226715, "learning_rate": 7.916666666666667e-06, "loss": 10.9103, "step": 135 }, { "epoch": 0.6291744976145728, "grad_norm": 0.199855774641037, "learning_rate": 7.901234567901235e-06, "loss": 11.5339, "step": 136 }, { "epoch": 0.6338007806852681, "grad_norm": 0.1627349704504013, "learning_rate": 7.885802469135803e-06, "loss": 10.3593, "step": 137 }, { "epoch": 0.6384270637559636, "grad_norm": 0.18615403771400452, "learning_rate": 7.870370370370372e-06, "loss": 12.3204, "step": 138 }, { "epoch": 0.6430533468266589, "grad_norm": 0.17866994440555573, "learning_rate": 7.854938271604939e-06, "loss": 11.548, "step": 139 }, { "epoch": 0.6476796298973544, "grad_norm": 0.22644223272800446, "learning_rate": 7.839506172839507e-06, "loss": 12.0289, "step": 140 }, { "epoch": 0.6523059129680497, "grad_norm": 0.18510523438453674, "learning_rate": 7.824074074074076e-06, "loss": 12.1312, "step": 141 }, { "epoch": 0.6569321960387451, "grad_norm": 0.1948799043893814, "learning_rate": 7.808641975308642e-06, "loss": 13.4831, "step": 142 }, { "epoch": 0.6615584791094405, "grad_norm": 0.2084604948759079, "learning_rate": 7.79320987654321e-06, "loss": 11.1289, "step": 143 }, { "epoch": 0.6661847621801359, "grad_norm": 0.194437637925148, "learning_rate": 7.77777777777778e-06, "loss": 11.5961, "step": 144 }, { "epoch": 0.6708110452508312, "grad_norm": 0.2627541124820709, "learning_rate": 7.762345679012346e-06, "loss": 12.0198, "step": 145 }, { "epoch": 0.6754373283215267, "grad_norm": 0.21856093406677246, "learning_rate": 7.746913580246914e-06, "loss": 12.5285, "step": 146 }, { "epoch": 0.680063611392222, "grad_norm": 0.15228690207004547, "learning_rate": 7.731481481481483e-06, "loss": 10.8653, "step": 147 }, { "epoch": 0.6846898944629175, "grad_norm": 0.22951403260231018, "learning_rate": 7.71604938271605e-06, "loss": 10.8751, "step": 148 }, { "epoch": 0.6893161775336129, "grad_norm": 0.25641652941703796, "learning_rate": 7.700617283950618e-06, "loss": 11.0195, "step": 149 }, { "epoch": 0.6939424606043082, "grad_norm": 0.17132331430912018, "learning_rate": 7.685185185185185e-06, "loss": 13.9163, "step": 150 }, { "epoch": 0.6985687436750037, "grad_norm": 0.24479900300502777, "learning_rate": 7.669753086419753e-06, "loss": 11.4487, "step": 151 }, { "epoch": 0.703195026745699, "grad_norm": 0.1990821659564972, "learning_rate": 7.654320987654322e-06, "loss": 10.6267, "step": 152 }, { "epoch": 0.7078213098163944, "grad_norm": 0.17914509773254395, "learning_rate": 7.638888888888888e-06, "loss": 11.5696, "step": 153 }, { "epoch": 0.7124475928870898, "grad_norm": 0.19086718559265137, "learning_rate": 7.623456790123458e-06, "loss": 12.3853, "step": 154 }, { "epoch": 0.7170738759577852, "grad_norm": 0.1417158991098404, "learning_rate": 7.608024691358026e-06, "loss": 10.8805, "step": 155 }, { "epoch": 0.7217001590284805, "grad_norm": 0.32373055815696716, "learning_rate": 7.592592592592594e-06, "loss": 12.586, "step": 156 }, { "epoch": 0.726326442099176, "grad_norm": 0.1995164453983307, "learning_rate": 7.577160493827161e-06, "loss": 11.4977, "step": 157 }, { "epoch": 0.7309527251698713, "grad_norm": 0.16449247300624847, "learning_rate": 7.561728395061729e-06, "loss": 11.8027, "step": 158 }, { "epoch": 0.7355790082405668, "grad_norm": 0.21888476610183716, "learning_rate": 7.546296296296297e-06, "loss": 11.8753, "step": 159 }, { "epoch": 0.7402052913112621, "grad_norm": 0.19433487951755524, "learning_rate": 7.530864197530865e-06, "loss": 11.5252, "step": 160 }, { "epoch": 0.7448315743819575, "grad_norm": 0.19709180295467377, "learning_rate": 7.515432098765433e-06, "loss": 10.7059, "step": 161 }, { "epoch": 0.7494578574526529, "grad_norm": 0.17522279918193817, "learning_rate": 7.500000000000001e-06, "loss": 11.2329, "step": 162 }, { "epoch": 0.7540841405233483, "grad_norm": 0.2038147747516632, "learning_rate": 7.484567901234569e-06, "loss": 11.425, "step": 163 }, { "epoch": 0.7587104235940436, "grad_norm": 0.29561126232147217, "learning_rate": 7.469135802469136e-06, "loss": 13.192, "step": 164 }, { "epoch": 0.7633367066647391, "grad_norm": 0.2295159548521042, "learning_rate": 7.453703703703704e-06, "loss": 10.9714, "step": 165 }, { "epoch": 0.7679629897354344, "grad_norm": 0.1554422378540039, "learning_rate": 7.438271604938272e-06, "loss": 10.5617, "step": 166 }, { "epoch": 0.7725892728061299, "grad_norm": 0.19565965235233307, "learning_rate": 7.42283950617284e-06, "loss": 10.8574, "step": 167 }, { "epoch": 0.7772155558768252, "grad_norm": 0.2645648419857025, "learning_rate": 7.4074074074074075e-06, "loss": 11.4472, "step": 168 }, { "epoch": 0.7818418389475206, "grad_norm": 0.2832958996295929, "learning_rate": 7.391975308641975e-06, "loss": 11.628, "step": 169 }, { "epoch": 0.786468122018216, "grad_norm": 0.2097349464893341, "learning_rate": 7.3765432098765435e-06, "loss": 12.2379, "step": 170 }, { "epoch": 0.7910944050889114, "grad_norm": 0.19168758392333984, "learning_rate": 7.361111111111112e-06, "loss": 10.0489, "step": 171 }, { "epoch": 0.7957206881596067, "grad_norm": 0.22470812499523163, "learning_rate": 7.34567901234568e-06, "loss": 11.8545, "step": 172 }, { "epoch": 0.8003469712303022, "grad_norm": 0.2722543179988861, "learning_rate": 7.330246913580248e-06, "loss": 10.4316, "step": 173 }, { "epoch": 0.8049732543009975, "grad_norm": 0.1686207354068756, "learning_rate": 7.314814814814816e-06, "loss": 12.5917, "step": 174 }, { "epoch": 0.8095995373716929, "grad_norm": 0.2805918753147125, "learning_rate": 7.299382716049383e-06, "loss": 10.6457, "step": 175 }, { "epoch": 0.8142258204423883, "grad_norm": 0.28346166014671326, "learning_rate": 7.283950617283952e-06, "loss": 11.6686, "step": 176 }, { "epoch": 0.8188521035130837, "grad_norm": 0.23147298395633698, "learning_rate": 7.268518518518519e-06, "loss": 10.316, "step": 177 }, { "epoch": 0.8234783865837791, "grad_norm": 0.17895270884037018, "learning_rate": 7.253086419753087e-06, "loss": 10.5762, "step": 178 }, { "epoch": 0.8281046696544745, "grad_norm": 0.2260911464691162, "learning_rate": 7.2376543209876545e-06, "loss": 10.9989, "step": 179 }, { "epoch": 0.8327309527251698, "grad_norm": 0.2163412719964981, "learning_rate": 7.222222222222223e-06, "loss": 10.7737, "step": 180 }, { "epoch": 0.8373572357958653, "grad_norm": 0.21451716125011444, "learning_rate": 7.2067901234567905e-06, "loss": 12.4478, "step": 181 }, { "epoch": 0.8419835188665606, "grad_norm": 0.14181958138942719, "learning_rate": 7.191358024691358e-06, "loss": 11.1375, "step": 182 }, { "epoch": 0.846609801937256, "grad_norm": 0.18365442752838135, "learning_rate": 7.1759259259259266e-06, "loss": 11.3768, "step": 183 }, { "epoch": 0.8512360850079514, "grad_norm": 0.20560483634471893, "learning_rate": 7.160493827160494e-06, "loss": 10.6342, "step": 184 }, { "epoch": 0.8558623680786468, "grad_norm": 0.3396760821342468, "learning_rate": 7.145061728395062e-06, "loss": 12.9821, "step": 185 }, { "epoch": 0.8604886511493423, "grad_norm": 0.18169118463993073, "learning_rate": 7.129629629629629e-06, "loss": 11.067, "step": 186 }, { "epoch": 0.8651149342200376, "grad_norm": 0.18200387060642242, "learning_rate": 7.114197530864199e-06, "loss": 10.6416, "step": 187 }, { "epoch": 0.869741217290733, "grad_norm": 0.23793809115886688, "learning_rate": 7.098765432098766e-06, "loss": 12.6389, "step": 188 }, { "epoch": 0.8743675003614284, "grad_norm": 0.6948350667953491, "learning_rate": 7.083333333333335e-06, "loss": 11.5131, "step": 189 }, { "epoch": 0.8789937834321238, "grad_norm": 0.19556772708892822, "learning_rate": 7.067901234567902e-06, "loss": 12.9335, "step": 190 }, { "epoch": 0.8836200665028191, "grad_norm": 0.21277347207069397, "learning_rate": 7.05246913580247e-06, "loss": 11.533, "step": 191 }, { "epoch": 0.8882463495735146, "grad_norm": 0.13241365551948547, "learning_rate": 7.0370370370370375e-06, "loss": 11.7588, "step": 192 }, { "epoch": 0.8928726326442099, "grad_norm": 0.16065774857997894, "learning_rate": 7.021604938271606e-06, "loss": 10.907, "step": 193 }, { "epoch": 0.8974989157149053, "grad_norm": 0.2425934374332428, "learning_rate": 7.0061728395061735e-06, "loss": 12.0332, "step": 194 }, { "epoch": 0.9021251987856007, "grad_norm": 0.32493147253990173, "learning_rate": 6.990740740740741e-06, "loss": 11.0481, "step": 195 }, { "epoch": 0.9067514818562961, "grad_norm": 0.20006415247917175, "learning_rate": 6.975308641975309e-06, "loss": 10.3825, "step": 196 }, { "epoch": 0.9113777649269915, "grad_norm": 0.14633627235889435, "learning_rate": 6.959876543209877e-06, "loss": 11.3131, "step": 197 }, { "epoch": 0.9160040479976869, "grad_norm": 0.1695437729358673, "learning_rate": 6.944444444444445e-06, "loss": 11.6805, "step": 198 }, { "epoch": 0.9206303310683822, "grad_norm": 0.22382433712482452, "learning_rate": 6.929012345679012e-06, "loss": 12.4138, "step": 199 }, { "epoch": 0.9252566141390777, "grad_norm": 0.18652208149433136, "learning_rate": 6.913580246913581e-06, "loss": 11.1235, "step": 200 }, { "epoch": 0.929882897209773, "grad_norm": 0.21566657721996307, "learning_rate": 6.898148148148148e-06, "loss": 11.1278, "step": 201 }, { "epoch": 0.9345091802804684, "grad_norm": 0.17630067467689514, "learning_rate": 6.882716049382716e-06, "loss": 12.1383, "step": 202 }, { "epoch": 0.9391354633511638, "grad_norm": 0.22753016650676727, "learning_rate": 6.867283950617285e-06, "loss": 12.0065, "step": 203 }, { "epoch": 0.9437617464218592, "grad_norm": 0.12988047301769257, "learning_rate": 6.851851851851853e-06, "loss": 11.9477, "step": 204 }, { "epoch": 0.9483880294925546, "grad_norm": 0.2507847547531128, "learning_rate": 6.8364197530864205e-06, "loss": 11.5818, "step": 205 }, { "epoch": 0.95301431256325, "grad_norm": 0.18030938506126404, "learning_rate": 6.820987654320988e-06, "loss": 13.1558, "step": 206 }, { "epoch": 0.9576405956339453, "grad_norm": 0.21600230038166046, "learning_rate": 6.8055555555555566e-06, "loss": 10.7038, "step": 207 }, { "epoch": 0.9622668787046408, "grad_norm": 0.1989142745733261, "learning_rate": 6.790123456790124e-06, "loss": 11.9244, "step": 208 }, { "epoch": 0.9668931617753361, "grad_norm": 0.2517797648906708, "learning_rate": 6.774691358024692e-06, "loss": 11.0019, "step": 209 }, { "epoch": 0.9715194448460315, "grad_norm": 0.24413681030273438, "learning_rate": 6.75925925925926e-06, "loss": 12.4322, "step": 210 }, { "epoch": 0.9761457279167269, "grad_norm": 0.20667782425880432, "learning_rate": 6.743827160493828e-06, "loss": 10.6134, "step": 211 }, { "epoch": 0.9807720109874223, "grad_norm": 0.17336109280586243, "learning_rate": 6.728395061728395e-06, "loss": 10.4673, "step": 212 }, { "epoch": 0.9853982940581176, "grad_norm": 0.17548175156116486, "learning_rate": 6.712962962962963e-06, "loss": 10.1377, "step": 213 }, { "epoch": 0.9900245771288131, "grad_norm": 0.1717563420534134, "learning_rate": 6.6975308641975314e-06, "loss": 11.5888, "step": 214 }, { "epoch": 0.9946508601995084, "grad_norm": 0.1841057986021042, "learning_rate": 6.682098765432099e-06, "loss": 12.2213, "step": 215 }, { "epoch": 0.9992771432702039, "grad_norm": 0.3000975251197815, "learning_rate": 6.666666666666667e-06, "loss": 11.046, "step": 216 }, { "epoch": 1.0, "grad_norm": 0.3000975251197815, "learning_rate": 6.666666666666667e-06, "loss": 1.4199, "step": 217 }, { "epoch": 1.0046262830706953, "grad_norm": 0.4262640178203583, "learning_rate": 6.651234567901235e-06, "loss": 10.5265, "step": 218 }, { "epoch": 1.0092525661413907, "grad_norm": 0.21694016456604004, "learning_rate": 6.635802469135803e-06, "loss": 10.4657, "step": 219 }, { "epoch": 1.0138788492120863, "grad_norm": 0.2662133276462555, "learning_rate": 6.620370370370371e-06, "loss": 10.7531, "step": 220 }, { "epoch": 1.0185051322827816, "grad_norm": 0.4052404463291168, "learning_rate": 6.60493827160494e-06, "loss": 11.1598, "step": 221 }, { "epoch": 1.023131415353477, "grad_norm": 0.23900097608566284, "learning_rate": 6.589506172839507e-06, "loss": 11.5257, "step": 222 }, { "epoch": 1.0277576984241723, "grad_norm": 0.22569526731967926, "learning_rate": 6.574074074074075e-06, "loss": 11.5564, "step": 223 }, { "epoch": 1.0323839814948677, "grad_norm": 0.29018473625183105, "learning_rate": 6.558641975308642e-06, "loss": 11.8655, "step": 224 }, { "epoch": 1.037010264565563, "grad_norm": 0.2856806516647339, "learning_rate": 6.543209876543211e-06, "loss": 11.8046, "step": 225 }, { "epoch": 1.0416365476362586, "grad_norm": 0.16691464185714722, "learning_rate": 6.5277777777777784e-06, "loss": 12.3704, "step": 226 }, { "epoch": 1.046262830706954, "grad_norm": 0.24121911823749542, "learning_rate": 6.512345679012346e-06, "loss": 10.7236, "step": 227 }, { "epoch": 1.0508891137776493, "grad_norm": 0.205901101231575, "learning_rate": 6.4969135802469145e-06, "loss": 10.9316, "step": 228 }, { "epoch": 1.0555153968483446, "grad_norm": 0.18375830352306366, "learning_rate": 6.481481481481482e-06, "loss": 9.422, "step": 229 }, { "epoch": 1.06014167991904, "grad_norm": 0.2132750004529953, "learning_rate": 6.46604938271605e-06, "loss": 12.6065, "step": 230 }, { "epoch": 1.0647679629897355, "grad_norm": 0.23490285873413086, "learning_rate": 6.450617283950617e-06, "loss": 11.2282, "step": 231 }, { "epoch": 1.0693942460604309, "grad_norm": 0.2604386806488037, "learning_rate": 6.435185185185186e-06, "loss": 10.8157, "step": 232 }, { "epoch": 1.0740205291311262, "grad_norm": 0.21887636184692383, "learning_rate": 6.419753086419753e-06, "loss": 10.6996, "step": 233 }, { "epoch": 1.0786468122018216, "grad_norm": 0.16946089267730713, "learning_rate": 6.404320987654321e-06, "loss": 10.3805, "step": 234 }, { "epoch": 1.083273095272517, "grad_norm": 0.20204631984233856, "learning_rate": 6.3888888888888885e-06, "loss": 11.4941, "step": 235 }, { "epoch": 1.0878993783432125, "grad_norm": 0.202669158577919, "learning_rate": 6.373456790123458e-06, "loss": 11.7944, "step": 236 }, { "epoch": 1.0925256614139078, "grad_norm": 0.22706535458564758, "learning_rate": 6.358024691358025e-06, "loss": 12.6031, "step": 237 }, { "epoch": 1.0971519444846032, "grad_norm": 0.17785371840000153, "learning_rate": 6.342592592592594e-06, "loss": 10.5881, "step": 238 }, { "epoch": 1.1017782275552985, "grad_norm": 0.17540261149406433, "learning_rate": 6.3271604938271615e-06, "loss": 10.996, "step": 239 }, { "epoch": 1.1064045106259939, "grad_norm": 0.22038735449314117, "learning_rate": 6.311728395061729e-06, "loss": 10.0293, "step": 240 }, { "epoch": 1.1110307936966892, "grad_norm": 0.17491741478443146, "learning_rate": 6.296296296296297e-06, "loss": 10.0996, "step": 241 }, { "epoch": 1.1156570767673848, "grad_norm": 0.2827455699443817, "learning_rate": 6.280864197530865e-06, "loss": 11.0431, "step": 242 }, { "epoch": 1.1202833598380801, "grad_norm": 0.21907830238342285, "learning_rate": 6.265432098765433e-06, "loss": 10.4707, "step": 243 }, { "epoch": 1.1249096429087755, "grad_norm": 0.19320356845855713, "learning_rate": 6.25e-06, "loss": 10.7269, "step": 244 }, { "epoch": 1.1295359259794708, "grad_norm": 0.22684165835380554, "learning_rate": 6.234567901234569e-06, "loss": 11.3143, "step": 245 }, { "epoch": 1.1341622090501662, "grad_norm": 0.21541574597358704, "learning_rate": 6.219135802469136e-06, "loss": 10.397, "step": 246 }, { "epoch": 1.1387884921208618, "grad_norm": 0.21086207032203674, "learning_rate": 6.203703703703704e-06, "loss": 11.0086, "step": 247 }, { "epoch": 1.143414775191557, "grad_norm": 0.26652148365974426, "learning_rate": 6.1882716049382715e-06, "loss": 11.5252, "step": 248 }, { "epoch": 1.1480410582622524, "grad_norm": 0.1995311677455902, "learning_rate": 6.17283950617284e-06, "loss": 9.3802, "step": 249 }, { "epoch": 1.1526673413329478, "grad_norm": 0.19715267419815063, "learning_rate": 6.157407407407408e-06, "loss": 11.6554, "step": 250 }, { "epoch": 1.1572936244036431, "grad_norm": 0.16357477009296417, "learning_rate": 6.141975308641975e-06, "loss": 10.7531, "step": 251 }, { "epoch": 1.1619199074743385, "grad_norm": 0.2201809138059616, "learning_rate": 6.126543209876543e-06, "loss": 11.4639, "step": 252 }, { "epoch": 1.166546190545034, "grad_norm": 0.24760210514068604, "learning_rate": 6.111111111111112e-06, "loss": 10.1224, "step": 253 }, { "epoch": 1.1711724736157294, "grad_norm": 0.18471167981624603, "learning_rate": 6.09567901234568e-06, "loss": 10.8261, "step": 254 }, { "epoch": 1.1757987566864248, "grad_norm": 0.20257946848869324, "learning_rate": 6.080246913580248e-06, "loss": 9.8945, "step": 255 }, { "epoch": 1.18042503975712, "grad_norm": 0.15809153020381927, "learning_rate": 6.064814814814816e-06, "loss": 10.4489, "step": 256 }, { "epoch": 1.1850513228278154, "grad_norm": 0.2175232172012329, "learning_rate": 6.049382716049383e-06, "loss": 10.6301, "step": 257 }, { "epoch": 1.189677605898511, "grad_norm": 0.22652743756771088, "learning_rate": 6.033950617283951e-06, "loss": 9.6189, "step": 258 }, { "epoch": 1.1943038889692064, "grad_norm": 0.21281598508358002, "learning_rate": 6.018518518518519e-06, "loss": 10.8016, "step": 259 }, { "epoch": 1.1989301720399017, "grad_norm": 0.20046480000019073, "learning_rate": 6.003086419753087e-06, "loss": 10.1304, "step": 260 }, { "epoch": 1.203556455110597, "grad_norm": 0.19008351862430573, "learning_rate": 5.9876543209876546e-06, "loss": 10.6718, "step": 261 }, { "epoch": 1.2081827381812924, "grad_norm": 0.2159290611743927, "learning_rate": 5.972222222222222e-06, "loss": 10.1963, "step": 262 }, { "epoch": 1.2128090212519878, "grad_norm": 0.1703040450811386, "learning_rate": 5.956790123456791e-06, "loss": 10.9394, "step": 263 }, { "epoch": 1.2174353043226833, "grad_norm": 0.20243868231773376, "learning_rate": 5.941358024691358e-06, "loss": 9.7358, "step": 264 }, { "epoch": 1.2220615873933787, "grad_norm": 0.21622534096240997, "learning_rate": 5.925925925925926e-06, "loss": 9.2896, "step": 265 }, { "epoch": 1.226687870464074, "grad_norm": 0.19408589601516724, "learning_rate": 5.910493827160494e-06, "loss": 10.0141, "step": 266 }, { "epoch": 1.2313141535347694, "grad_norm": 0.19652226567268372, "learning_rate": 5.895061728395062e-06, "loss": 10.6724, "step": 267 }, { "epoch": 1.2359404366054647, "grad_norm": 0.22394828498363495, "learning_rate": 5.8796296296296295e-06, "loss": 11.7472, "step": 268 }, { "epoch": 1.2405667196761603, "grad_norm": 0.18796682357788086, "learning_rate": 5.864197530864199e-06, "loss": 11.3051, "step": 269 }, { "epoch": 1.2451930027468556, "grad_norm": 0.1897086501121521, "learning_rate": 5.848765432098766e-06, "loss": 9.9475, "step": 270 }, { "epoch": 1.249819285817551, "grad_norm": 0.19912078976631165, "learning_rate": 5.833333333333334e-06, "loss": 10.6907, "step": 271 }, { "epoch": 1.2544455688882463, "grad_norm": 0.25889354944229126, "learning_rate": 5.817901234567902e-06, "loss": 9.9069, "step": 272 }, { "epoch": 1.2590718519589417, "grad_norm": 0.19028806686401367, "learning_rate": 5.80246913580247e-06, "loss": 10.0748, "step": 273 }, { "epoch": 1.263698135029637, "grad_norm": 0.1661495566368103, "learning_rate": 5.787037037037038e-06, "loss": 10.1342, "step": 274 }, { "epoch": 1.2683244181003326, "grad_norm": 0.17706620693206787, "learning_rate": 5.771604938271605e-06, "loss": 12.4703, "step": 275 }, { "epoch": 1.272950701171028, "grad_norm": 0.26582854986190796, "learning_rate": 5.756172839506174e-06, "loss": 12.6464, "step": 276 }, { "epoch": 1.2775769842417233, "grad_norm": 0.178712397813797, "learning_rate": 5.740740740740741e-06, "loss": 9.6011, "step": 277 }, { "epoch": 1.2822032673124186, "grad_norm": 0.18910570442676544, "learning_rate": 5.725308641975309e-06, "loss": 10.8773, "step": 278 }, { "epoch": 1.2868295503831142, "grad_norm": 0.24754610657691956, "learning_rate": 5.7098765432098764e-06, "loss": 10.8981, "step": 279 }, { "epoch": 1.2914558334538095, "grad_norm": 0.14676739275455475, "learning_rate": 5.694444444444445e-06, "loss": 10.1851, "step": 280 }, { "epoch": 1.296082116524505, "grad_norm": 0.21573857963085175, "learning_rate": 5.6790123456790125e-06, "loss": 10.2565, "step": 281 }, { "epoch": 1.3007083995952002, "grad_norm": 0.24850860238075256, "learning_rate": 5.66358024691358e-06, "loss": 9.9282, "step": 282 }, { "epoch": 1.3053346826658956, "grad_norm": 0.180882066488266, "learning_rate": 5.6481481481481485e-06, "loss": 11.6456, "step": 283 }, { "epoch": 1.309960965736591, "grad_norm": 0.19305004179477692, "learning_rate": 5.632716049382716e-06, "loss": 10.0129, "step": 284 }, { "epoch": 1.3145872488072863, "grad_norm": 0.18499980866909027, "learning_rate": 5.617283950617285e-06, "loss": 10.6136, "step": 285 }, { "epoch": 1.3192135318779818, "grad_norm": 0.2523798942565918, "learning_rate": 5.601851851851853e-06, "loss": 9.1629, "step": 286 }, { "epoch": 1.3238398149486772, "grad_norm": 0.19359612464904785, "learning_rate": 5.586419753086421e-06, "loss": 9.9917, "step": 287 }, { "epoch": 1.3284660980193725, "grad_norm": 0.15306031703948975, "learning_rate": 5.570987654320988e-06, "loss": 9.6058, "step": 288 }, { "epoch": 1.333092381090068, "grad_norm": 0.2070600688457489, "learning_rate": 5.555555555555557e-06, "loss": 10.2776, "step": 289 }, { "epoch": 1.3377186641607635, "grad_norm": 0.1619105190038681, "learning_rate": 5.540123456790124e-06, "loss": 9.1465, "step": 290 }, { "epoch": 1.3423449472314588, "grad_norm": 0.16202016174793243, "learning_rate": 5.524691358024692e-06, "loss": 10.0362, "step": 291 }, { "epoch": 1.3469712303021542, "grad_norm": 0.19075003266334534, "learning_rate": 5.5092592592592595e-06, "loss": 10.0189, "step": 292 }, { "epoch": 1.3515975133728495, "grad_norm": 0.16484172642230988, "learning_rate": 5.493827160493828e-06, "loss": 12.3743, "step": 293 }, { "epoch": 1.3562237964435448, "grad_norm": 0.2151799499988556, "learning_rate": 5.4783950617283955e-06, "loss": 9.9846, "step": 294 }, { "epoch": 1.3608500795142402, "grad_norm": 0.20363835990428925, "learning_rate": 5.462962962962963e-06, "loss": 9.671, "step": 295 }, { "epoch": 1.3654763625849355, "grad_norm": 0.15196539461612701, "learning_rate": 5.447530864197531e-06, "loss": 9.97, "step": 296 }, { "epoch": 1.3701026456556311, "grad_norm": 0.17592975497245789, "learning_rate": 5.432098765432099e-06, "loss": 10.8506, "step": 297 }, { "epoch": 1.3747289287263265, "grad_norm": 0.20012719929218292, "learning_rate": 5.416666666666667e-06, "loss": 10.4503, "step": 298 }, { "epoch": 1.3793552117970218, "grad_norm": 0.19312864542007446, "learning_rate": 5.401234567901234e-06, "loss": 9.7727, "step": 299 }, { "epoch": 1.3839814948677172, "grad_norm": 0.14697419106960297, "learning_rate": 5.385802469135803e-06, "loss": 11.9123, "step": 300 }, { "epoch": 1.3886077779384127, "grad_norm": 0.15580600500106812, "learning_rate": 5.370370370370371e-06, "loss": 10.3442, "step": 301 }, { "epoch": 1.393234061009108, "grad_norm": 0.19586268067359924, "learning_rate": 5.354938271604939e-06, "loss": 8.9558, "step": 302 }, { "epoch": 1.3978603440798034, "grad_norm": 0.19394423067569733, "learning_rate": 5.339506172839507e-06, "loss": 10.8312, "step": 303 }, { "epoch": 1.4024866271504988, "grad_norm": 0.21832065284252167, "learning_rate": 5.324074074074075e-06, "loss": 9.9943, "step": 304 }, { "epoch": 1.4071129102211941, "grad_norm": 0.16751375794410706, "learning_rate": 5.3086419753086425e-06, "loss": 11.3979, "step": 305 }, { "epoch": 1.4117391932918895, "grad_norm": 0.1763811856508255, "learning_rate": 5.29320987654321e-06, "loss": 11.175, "step": 306 }, { "epoch": 1.4163654763625848, "grad_norm": 0.1916954666376114, "learning_rate": 5.2777777777777785e-06, "loss": 8.8938, "step": 307 }, { "epoch": 1.4209917594332804, "grad_norm": 0.12443282455205917, "learning_rate": 5.262345679012346e-06, "loss": 9.6263, "step": 308 }, { "epoch": 1.4256180425039757, "grad_norm": 0.18852603435516357, "learning_rate": 5.246913580246914e-06, "loss": 10.6811, "step": 309 }, { "epoch": 1.430244325574671, "grad_norm": 0.1854352355003357, "learning_rate": 5.231481481481482e-06, "loss": 9.9107, "step": 310 }, { "epoch": 1.4348706086453664, "grad_norm": 0.2382035106420517, "learning_rate": 5.21604938271605e-06, "loss": 10.8001, "step": 311 }, { "epoch": 1.439496891716062, "grad_norm": 0.18645620346069336, "learning_rate": 5.200617283950617e-06, "loss": 9.686, "step": 312 }, { "epoch": 1.4441231747867573, "grad_norm": 0.1910058856010437, "learning_rate": 5.185185185185185e-06, "loss": 10.3407, "step": 313 }, { "epoch": 1.4487494578574527, "grad_norm": 0.23428039252758026, "learning_rate": 5.1697530864197534e-06, "loss": 9.2624, "step": 314 }, { "epoch": 1.453375740928148, "grad_norm": 0.17926917970180511, "learning_rate": 5.154320987654321e-06, "loss": 9.5153, "step": 315 }, { "epoch": 1.4580020239988434, "grad_norm": 0.22386913001537323, "learning_rate": 5.138888888888889e-06, "loss": 10.6948, "step": 316 }, { "epoch": 1.4626283070695387, "grad_norm": 0.20423167943954468, "learning_rate": 5.123456790123458e-06, "loss": 10.3966, "step": 317 }, { "epoch": 1.467254590140234, "grad_norm": 0.15411533415317535, "learning_rate": 5.1080246913580255e-06, "loss": 10.2981, "step": 318 }, { "epoch": 1.4718808732109296, "grad_norm": 0.11654549837112427, "learning_rate": 5.092592592592593e-06, "loss": 10.1807, "step": 319 }, { "epoch": 1.476507156281625, "grad_norm": 0.20476582646369934, "learning_rate": 5.0771604938271616e-06, "loss": 9.9201, "step": 320 }, { "epoch": 1.4811334393523203, "grad_norm": 0.19140280783176422, "learning_rate": 5.061728395061729e-06, "loss": 10.3292, "step": 321 }, { "epoch": 1.4857597224230157, "grad_norm": 0.16844482719898224, "learning_rate": 5.046296296296297e-06, "loss": 10.0726, "step": 322 }, { "epoch": 1.4903860054937113, "grad_norm": 0.22298437356948853, "learning_rate": 5.030864197530864e-06, "loss": 10.9447, "step": 323 }, { "epoch": 1.4950122885644066, "grad_norm": 0.1775355190038681, "learning_rate": 5.015432098765433e-06, "loss": 12.1205, "step": 324 }, { "epoch": 1.499638571635102, "grad_norm": 0.198349267244339, "learning_rate": 5e-06, "loss": 9.0277, "step": 325 }, { "epoch": 1.5042648547057973, "grad_norm": 0.16838030517101288, "learning_rate": 4.984567901234568e-06, "loss": 10.3551, "step": 326 }, { "epoch": 1.5088911377764926, "grad_norm": 0.15806855261325836, "learning_rate": 4.9691358024691365e-06, "loss": 10.0048, "step": 327 }, { "epoch": 1.513517420847188, "grad_norm": 0.21915429830551147, "learning_rate": 4.953703703703704e-06, "loss": 10.554, "step": 328 }, { "epoch": 1.5181437039178833, "grad_norm": 0.23262116312980652, "learning_rate": 4.938271604938272e-06, "loss": 10.8198, "step": 329 }, { "epoch": 1.522769986988579, "grad_norm": 0.18927887082099915, "learning_rate": 4.92283950617284e-06, "loss": 10.7398, "step": 330 }, { "epoch": 1.5273962700592743, "grad_norm": 0.138357013463974, "learning_rate": 4.907407407407408e-06, "loss": 9.8479, "step": 331 }, { "epoch": 1.5320225531299696, "grad_norm": 0.20223885774612427, "learning_rate": 4.891975308641976e-06, "loss": 11.1796, "step": 332 }, { "epoch": 1.5366488362006652, "grad_norm": 0.17738379538059235, "learning_rate": 4.876543209876544e-06, "loss": 9.1805, "step": 333 }, { "epoch": 1.5412751192713605, "grad_norm": 0.18442702293395996, "learning_rate": 4.861111111111111e-06, "loss": 8.4714, "step": 334 }, { "epoch": 1.5459014023420559, "grad_norm": 0.19169731438159943, "learning_rate": 4.845679012345679e-06, "loss": 9.1503, "step": 335 }, { "epoch": 1.5505276854127512, "grad_norm": 0.19981014728546143, "learning_rate": 4.830246913580247e-06, "loss": 9.6427, "step": 336 }, { "epoch": 1.5551539684834466, "grad_norm": 0.1916283518075943, "learning_rate": 4.814814814814815e-06, "loss": 8.8991, "step": 337 }, { "epoch": 1.559780251554142, "grad_norm": 0.14049910008907318, "learning_rate": 4.7993827160493834e-06, "loss": 9.2888, "step": 338 }, { "epoch": 1.5644065346248373, "grad_norm": 0.14010216295719147, "learning_rate": 4.783950617283951e-06, "loss": 10.4533, "step": 339 }, { "epoch": 1.5690328176955326, "grad_norm": 0.1919959932565689, "learning_rate": 4.768518518518519e-06, "loss": 10.149, "step": 340 }, { "epoch": 1.5736591007662282, "grad_norm": 0.1598203033208847, "learning_rate": 4.753086419753087e-06, "loss": 11.1529, "step": 341 }, { "epoch": 1.5782853838369235, "grad_norm": 0.19053767621517181, "learning_rate": 4.737654320987655e-06, "loss": 10.5543, "step": 342 }, { "epoch": 1.5829116669076189, "grad_norm": 0.21635524928569794, "learning_rate": 4.722222222222222e-06, "loss": 8.7965, "step": 343 }, { "epoch": 1.5875379499783144, "grad_norm": 0.2088153064250946, "learning_rate": 4.706790123456791e-06, "loss": 9.1113, "step": 344 }, { "epoch": 1.5921642330490098, "grad_norm": 0.1993483006954193, "learning_rate": 4.691358024691358e-06, "loss": 9.7291, "step": 345 }, { "epoch": 1.5967905161197051, "grad_norm": 0.24835029244422913, "learning_rate": 4.675925925925927e-06, "loss": 9.5215, "step": 346 }, { "epoch": 1.6014167991904005, "grad_norm": 0.24942326545715332, "learning_rate": 4.660493827160494e-06, "loss": 9.3023, "step": 347 }, { "epoch": 1.6060430822610958, "grad_norm": 0.18829648196697235, "learning_rate": 4.645061728395062e-06, "loss": 9.6869, "step": 348 }, { "epoch": 1.6106693653317912, "grad_norm": 0.17725762724876404, "learning_rate": 4.62962962962963e-06, "loss": 10.0154, "step": 349 }, { "epoch": 1.6152956484024865, "grad_norm": 0.145020991563797, "learning_rate": 4.614197530864198e-06, "loss": 10.6015, "step": 350 }, { "epoch": 1.6199219314731819, "grad_norm": 0.12719136476516724, "learning_rate": 4.598765432098766e-06, "loss": 9.9244, "step": 351 }, { "epoch": 1.6245482145438774, "grad_norm": 0.1771680861711502, "learning_rate": 4.583333333333333e-06, "loss": 9.4468, "step": 352 }, { "epoch": 1.6291744976145728, "grad_norm": 0.1684349775314331, "learning_rate": 4.567901234567902e-06, "loss": 11.5601, "step": 353 }, { "epoch": 1.6338007806852681, "grad_norm": 0.1284870207309723, "learning_rate": 4.55246913580247e-06, "loss": 9.5658, "step": 354 }, { "epoch": 1.6384270637559637, "grad_norm": 0.15432706475257874, "learning_rate": 4.537037037037038e-06, "loss": 10.2547, "step": 355 }, { "epoch": 1.643053346826659, "grad_norm": 0.19519680738449097, "learning_rate": 4.521604938271605e-06, "loss": 9.2126, "step": 356 }, { "epoch": 1.6476796298973544, "grad_norm": 0.1388140469789505, "learning_rate": 4.506172839506173e-06, "loss": 9.9578, "step": 357 }, { "epoch": 1.6523059129680497, "grad_norm": 0.19714529812335968, "learning_rate": 4.490740740740741e-06, "loss": 11.7677, "step": 358 }, { "epoch": 1.656932196038745, "grad_norm": 0.22792014479637146, "learning_rate": 4.475308641975309e-06, "loss": 9.6164, "step": 359 }, { "epoch": 1.6615584791094404, "grad_norm": 0.1519528478384018, "learning_rate": 4.4598765432098765e-06, "loss": 9.4579, "step": 360 }, { "epoch": 1.6661847621801358, "grad_norm": 0.17547698318958282, "learning_rate": 4.444444444444444e-06, "loss": 8.2953, "step": 361 }, { "epoch": 1.6708110452508311, "grad_norm": 0.2348184585571289, "learning_rate": 4.429012345679013e-06, "loss": 9.8739, "step": 362 }, { "epoch": 1.6754373283215267, "grad_norm": 0.16753868758678436, "learning_rate": 4.413580246913581e-06, "loss": 9.8006, "step": 363 }, { "epoch": 1.680063611392222, "grad_norm": 0.16872891783714294, "learning_rate": 4.398148148148149e-06, "loss": 10.2161, "step": 364 }, { "epoch": 1.6846898944629176, "grad_norm": 0.18189671635627747, "learning_rate": 4.382716049382716e-06, "loss": 9.5235, "step": 365 }, { "epoch": 1.689316177533613, "grad_norm": 0.12333816289901733, "learning_rate": 4.367283950617285e-06, "loss": 10.2188, "step": 366 }, { "epoch": 1.6939424606043083, "grad_norm": 0.22701486945152283, "learning_rate": 4.351851851851852e-06, "loss": 10.7382, "step": 367 }, { "epoch": 1.6985687436750037, "grad_norm": 0.20510244369506836, "learning_rate": 4.33641975308642e-06, "loss": 8.7558, "step": 368 }, { "epoch": 1.703195026745699, "grad_norm": 0.1589343398809433, "learning_rate": 4.3209876543209875e-06, "loss": 9.924, "step": 369 }, { "epoch": 1.7078213098163944, "grad_norm": 0.1433723270893097, "learning_rate": 4.305555555555556e-06, "loss": 11.0008, "step": 370 }, { "epoch": 1.7124475928870897, "grad_norm": 0.17607755959033966, "learning_rate": 4.290123456790124e-06, "loss": 10.0096, "step": 371 }, { "epoch": 1.717073875957785, "grad_norm": 0.16202040016651154, "learning_rate": 4.274691358024692e-06, "loss": 9.9295, "step": 372 }, { "epoch": 1.7217001590284804, "grad_norm": 0.21179550886154175, "learning_rate": 4.2592592592592596e-06, "loss": 8.9856, "step": 373 }, { "epoch": 1.726326442099176, "grad_norm": 0.35698196291923523, "learning_rate": 4.243827160493827e-06, "loss": 9.8793, "step": 374 }, { "epoch": 1.7309527251698713, "grad_norm": 0.18081314861774445, "learning_rate": 4.228395061728396e-06, "loss": 10.8514, "step": 375 }, { "epoch": 1.7355790082405669, "grad_norm": 0.19385862350463867, "learning_rate": 4.212962962962963e-06, "loss": 8.6034, "step": 376 }, { "epoch": 1.7402052913112622, "grad_norm": 0.17623740434646606, "learning_rate": 4.197530864197531e-06, "loss": 10.7457, "step": 377 }, { "epoch": 1.7448315743819576, "grad_norm": 0.1551784873008728, "learning_rate": 4.182098765432099e-06, "loss": 9.2378, "step": 378 }, { "epoch": 1.749457857452653, "grad_norm": 0.15923646092414856, "learning_rate": 4.166666666666667e-06, "loss": 10.0731, "step": 379 }, { "epoch": 1.7540841405233483, "grad_norm": 0.16185562312602997, "learning_rate": 4.151234567901235e-06, "loss": 8.9162, "step": 380 }, { "epoch": 1.7587104235940436, "grad_norm": 0.23534594476222992, "learning_rate": 4.135802469135803e-06, "loss": 9.3805, "step": 381 }, { "epoch": 1.763336706664739, "grad_norm": 0.17162498831748962, "learning_rate": 4.1203703703703705e-06, "loss": 10.0609, "step": 382 }, { "epoch": 1.7679629897354343, "grad_norm": 0.15775223076343536, "learning_rate": 4.104938271604938e-06, "loss": 9.3534, "step": 383 }, { "epoch": 1.7725892728061299, "grad_norm": 0.2251134216785431, "learning_rate": 4.0895061728395066e-06, "loss": 9.746, "step": 384 }, { "epoch": 1.7772155558768252, "grad_norm": 0.19613635540008545, "learning_rate": 4.074074074074074e-06, "loss": 10.42, "step": 385 }, { "epoch": 1.7818418389475206, "grad_norm": 0.16256873309612274, "learning_rate": 4.058641975308643e-06, "loss": 9.4703, "step": 386 }, { "epoch": 1.7864681220182161, "grad_norm": 0.16490335762500763, "learning_rate": 4.04320987654321e-06, "loss": 9.5418, "step": 387 }, { "epoch": 1.7910944050889115, "grad_norm": 0.159736767411232, "learning_rate": 4.027777777777779e-06, "loss": 10.1403, "step": 388 }, { "epoch": 1.7957206881596068, "grad_norm": 0.1494799256324768, "learning_rate": 4.012345679012346e-06, "loss": 9.1583, "step": 389 }, { "epoch": 1.8003469712303022, "grad_norm": 0.19815048575401306, "learning_rate": 3.996913580246914e-06, "loss": 10.2705, "step": 390 }, { "epoch": 1.8049732543009975, "grad_norm": 0.17572833597660065, "learning_rate": 3.9814814814814814e-06, "loss": 9.4608, "step": 391 }, { "epoch": 1.8095995373716929, "grad_norm": 0.23244017362594604, "learning_rate": 3.96604938271605e-06, "loss": 10.6273, "step": 392 }, { "epoch": 1.8142258204423882, "grad_norm": 0.17120610177516937, "learning_rate": 3.9506172839506175e-06, "loss": 10.0444, "step": 393 }, { "epoch": 1.8188521035130836, "grad_norm": 0.27220579981803894, "learning_rate": 3.935185185185186e-06, "loss": 10.64, "step": 394 }, { "epoch": 1.8234783865837791, "grad_norm": 0.1648477017879486, "learning_rate": 3.9197530864197535e-06, "loss": 9.6264, "step": 395 }, { "epoch": 1.8281046696544745, "grad_norm": 0.15771055221557617, "learning_rate": 3.904320987654321e-06, "loss": 10.5564, "step": 396 }, { "epoch": 1.8327309527251698, "grad_norm": 0.21713097393512726, "learning_rate": 3.88888888888889e-06, "loss": 10.6551, "step": 397 }, { "epoch": 1.8373572357958654, "grad_norm": 0.17076119780540466, "learning_rate": 3.873456790123457e-06, "loss": 11.1632, "step": 398 }, { "epoch": 1.8419835188665608, "grad_norm": 0.160685196518898, "learning_rate": 3.858024691358025e-06, "loss": 9.4666, "step": 399 }, { "epoch": 1.846609801937256, "grad_norm": 0.17023253440856934, "learning_rate": 3.842592592592592e-06, "loss": 9.8118, "step": 400 }, { "epoch": 1.8512360850079514, "grad_norm": 0.2186374068260193, "learning_rate": 3.827160493827161e-06, "loss": 10.4903, "step": 401 }, { "epoch": 1.8558623680786468, "grad_norm": 0.17954406142234802, "learning_rate": 3.811728395061729e-06, "loss": 11.3069, "step": 402 }, { "epoch": 1.8604886511493421, "grad_norm": 0.2778474986553192, "learning_rate": 3.796296296296297e-06, "loss": 10.5854, "step": 403 }, { "epoch": 1.8651149342200375, "grad_norm": 0.27708959579467773, "learning_rate": 3.7808641975308645e-06, "loss": 10.2884, "step": 404 }, { "epoch": 1.8697412172907328, "grad_norm": 0.14399868249893188, "learning_rate": 3.7654320987654325e-06, "loss": 10.1108, "step": 405 }, { "epoch": 1.8743675003614284, "grad_norm": 0.19901257753372192, "learning_rate": 3.7500000000000005e-06, "loss": 10.2892, "step": 406 }, { "epoch": 1.8789937834321238, "grad_norm": 0.21875345706939697, "learning_rate": 3.734567901234568e-06, "loss": 9.0521, "step": 407 }, { "epoch": 1.883620066502819, "grad_norm": 0.14913184940814972, "learning_rate": 3.719135802469136e-06, "loss": 10.5905, "step": 408 }, { "epoch": 1.8882463495735147, "grad_norm": 0.22289720177650452, "learning_rate": 3.7037037037037037e-06, "loss": 9.1175, "step": 409 }, { "epoch": 1.89287263264421, "grad_norm": 0.16322648525238037, "learning_rate": 3.6882716049382718e-06, "loss": 10.1569, "step": 410 }, { "epoch": 1.8974989157149054, "grad_norm": 0.13389474153518677, "learning_rate": 3.67283950617284e-06, "loss": 9.7421, "step": 411 }, { "epoch": 1.9021251987856007, "grad_norm": 0.17904877662658691, "learning_rate": 3.657407407407408e-06, "loss": 10.306, "step": 412 }, { "epoch": 1.906751481856296, "grad_norm": 0.22490617632865906, "learning_rate": 3.641975308641976e-06, "loss": 10.0275, "step": 413 }, { "epoch": 1.9113777649269914, "grad_norm": 0.20119769871234894, "learning_rate": 3.6265432098765434e-06, "loss": 9.5942, "step": 414 }, { "epoch": 1.9160040479976868, "grad_norm": 0.156976118683815, "learning_rate": 3.6111111111111115e-06, "loss": 10.3435, "step": 415 }, { "epoch": 1.920630331068382, "grad_norm": 0.13303594291210175, "learning_rate": 3.595679012345679e-06, "loss": 8.8616, "step": 416 }, { "epoch": 1.9252566141390777, "grad_norm": 0.21186932921409607, "learning_rate": 3.580246913580247e-06, "loss": 9.8981, "step": 417 }, { "epoch": 1.929882897209773, "grad_norm": 0.20669791102409363, "learning_rate": 3.5648148148148147e-06, "loss": 10.3689, "step": 418 }, { "epoch": 1.9345091802804684, "grad_norm": 0.20095276832580566, "learning_rate": 3.549382716049383e-06, "loss": 9.136, "step": 419 }, { "epoch": 1.939135463351164, "grad_norm": 0.16348451375961304, "learning_rate": 3.533950617283951e-06, "loss": 10.4419, "step": 420 }, { "epoch": 1.9437617464218593, "grad_norm": 0.16439932584762573, "learning_rate": 3.5185185185185187e-06, "loss": 10.0692, "step": 421 }, { "epoch": 1.9483880294925546, "grad_norm": 0.18562501668930054, "learning_rate": 3.5030864197530868e-06, "loss": 10.4452, "step": 422 }, { "epoch": 1.95301431256325, "grad_norm": 0.15668661892414093, "learning_rate": 3.4876543209876544e-06, "loss": 9.1397, "step": 423 }, { "epoch": 1.9576405956339453, "grad_norm": 0.129827082157135, "learning_rate": 3.4722222222222224e-06, "loss": 8.428, "step": 424 }, { "epoch": 1.9622668787046407, "grad_norm": 0.14903448522090912, "learning_rate": 3.4567901234567904e-06, "loss": 9.0113, "step": 425 }, { "epoch": 1.966893161775336, "grad_norm": 0.2033424973487854, "learning_rate": 3.441358024691358e-06, "loss": 8.6822, "step": 426 }, { "epoch": 1.9715194448460314, "grad_norm": 0.1724868267774582, "learning_rate": 3.4259259259259265e-06, "loss": 9.3924, "step": 427 }, { "epoch": 1.976145727916727, "grad_norm": 0.1788475513458252, "learning_rate": 3.410493827160494e-06, "loss": 7.8849, "step": 428 }, { "epoch": 1.9807720109874223, "grad_norm": 0.13475210964679718, "learning_rate": 3.395061728395062e-06, "loss": 10.321, "step": 429 }, { "epoch": 1.9853982940581176, "grad_norm": 0.15269149839878082, "learning_rate": 3.37962962962963e-06, "loss": 9.5632, "step": 430 }, { "epoch": 1.9900245771288132, "grad_norm": 0.20584136247634888, "learning_rate": 3.3641975308641977e-06, "loss": 9.0406, "step": 431 }, { "epoch": 1.9946508601995085, "grad_norm": 0.1267031729221344, "learning_rate": 3.3487654320987657e-06, "loss": 9.1269, "step": 432 }, { "epoch": 1.999277143270204, "grad_norm": 0.20229946076869965, "learning_rate": 3.3333333333333333e-06, "loss": 9.1644, "step": 433 }, { "epoch": 2.0, "grad_norm": 0.20229946076869965, "learning_rate": 3.3333333333333333e-06, "loss": 1.6753, "step": 434 }, { "epoch": 2.0046262830706953, "grad_norm": 0.19674813747406006, "learning_rate": 3.3179012345679013e-06, "loss": 8.9472, "step": 435 }, { "epoch": 2.0092525661413907, "grad_norm": 0.17176903784275055, "learning_rate": 3.30246913580247e-06, "loss": 8.1618, "step": 436 }, { "epoch": 2.013878849212086, "grad_norm": 0.21751543879508972, "learning_rate": 3.2870370370370374e-06, "loss": 11.298, "step": 437 }, { "epoch": 2.0185051322827814, "grad_norm": 0.15177738666534424, "learning_rate": 3.2716049382716054e-06, "loss": 8.9182, "step": 438 }, { "epoch": 2.0231314153534767, "grad_norm": 0.13782791793346405, "learning_rate": 3.256172839506173e-06, "loss": 9.138, "step": 439 }, { "epoch": 2.0277576984241725, "grad_norm": 0.1871718168258667, "learning_rate": 3.240740740740741e-06, "loss": 9.3935, "step": 440 }, { "epoch": 2.032383981494868, "grad_norm": 0.13550569117069244, "learning_rate": 3.2253086419753086e-06, "loss": 9.4065, "step": 441 }, { "epoch": 2.0370102645655632, "grad_norm": 0.1414715200662613, "learning_rate": 3.2098765432098767e-06, "loss": 10.0772, "step": 442 }, { "epoch": 2.0416365476362586, "grad_norm": 0.16839131712913513, "learning_rate": 3.1944444444444443e-06, "loss": 8.4968, "step": 443 }, { "epoch": 2.046262830706954, "grad_norm": 0.181600421667099, "learning_rate": 3.1790123456790127e-06, "loss": 8.0076, "step": 444 }, { "epoch": 2.0508891137776493, "grad_norm": 0.1106899231672287, "learning_rate": 3.1635802469135807e-06, "loss": 8.5825, "step": 445 }, { "epoch": 2.0555153968483446, "grad_norm": 0.17763055860996246, "learning_rate": 3.1481481481481483e-06, "loss": 9.2981, "step": 446 }, { "epoch": 2.06014167991904, "grad_norm": 0.19595400989055634, "learning_rate": 3.1327160493827164e-06, "loss": 10.5243, "step": 447 }, { "epoch": 2.0647679629897353, "grad_norm": 0.18254916369915009, "learning_rate": 3.1172839506172844e-06, "loss": 10.9755, "step": 448 }, { "epoch": 2.0693942460604307, "grad_norm": 0.18021439015865326, "learning_rate": 3.101851851851852e-06, "loss": 9.3634, "step": 449 }, { "epoch": 2.074020529131126, "grad_norm": 0.20369423925876617, "learning_rate": 3.08641975308642e-06, "loss": 10.1676, "step": 450 }, { "epoch": 2.078646812201822, "grad_norm": 0.1611659973859787, "learning_rate": 3.0709876543209876e-06, "loss": 10.5072, "step": 451 }, { "epoch": 2.083273095272517, "grad_norm": 0.16623584926128387, "learning_rate": 3.055555555555556e-06, "loss": 8.6923, "step": 452 }, { "epoch": 2.0878993783432125, "grad_norm": 0.20521709322929382, "learning_rate": 3.040123456790124e-06, "loss": 10.3398, "step": 453 }, { "epoch": 2.092525661413908, "grad_norm": 0.150822713971138, "learning_rate": 3.0246913580246917e-06, "loss": 8.1734, "step": 454 }, { "epoch": 2.097151944484603, "grad_norm": 0.14610332250595093, "learning_rate": 3.0092592592592597e-06, "loss": 8.0033, "step": 455 }, { "epoch": 2.1017782275552985, "grad_norm": 0.16596634685993195, "learning_rate": 2.9938271604938273e-06, "loss": 9.923, "step": 456 }, { "epoch": 2.106404510625994, "grad_norm": 0.15604960918426514, "learning_rate": 2.9783950617283953e-06, "loss": 9.5128, "step": 457 }, { "epoch": 2.1110307936966892, "grad_norm": 0.1488286554813385, "learning_rate": 2.962962962962963e-06, "loss": 8.704, "step": 458 }, { "epoch": 2.1156570767673846, "grad_norm": 0.1483355164527893, "learning_rate": 2.947530864197531e-06, "loss": 8.4275, "step": 459 }, { "epoch": 2.12028335983808, "grad_norm": 0.2545453906059265, "learning_rate": 2.9320987654320994e-06, "loss": 9.332, "step": 460 }, { "epoch": 2.1249096429087757, "grad_norm": 0.18272359669208527, "learning_rate": 2.916666666666667e-06, "loss": 9.4908, "step": 461 }, { "epoch": 2.129535925979471, "grad_norm": 0.19491931796073914, "learning_rate": 2.901234567901235e-06, "loss": 8.5668, "step": 462 }, { "epoch": 2.1341622090501664, "grad_norm": 0.171804741024971, "learning_rate": 2.8858024691358026e-06, "loss": 8.9085, "step": 463 }, { "epoch": 2.1387884921208618, "grad_norm": 0.22477325797080994, "learning_rate": 2.8703703703703706e-06, "loss": 9.5643, "step": 464 }, { "epoch": 2.143414775191557, "grad_norm": 0.2512359023094177, "learning_rate": 2.8549382716049382e-06, "loss": 10.6236, "step": 465 }, { "epoch": 2.1480410582622524, "grad_norm": 0.15513500571250916, "learning_rate": 2.8395061728395062e-06, "loss": 9.3385, "step": 466 }, { "epoch": 2.152667341332948, "grad_norm": 0.18065772950649261, "learning_rate": 2.8240740740740743e-06, "loss": 9.3942, "step": 467 }, { "epoch": 2.157293624403643, "grad_norm": 0.19600172340869904, "learning_rate": 2.8086419753086423e-06, "loss": 9.0732, "step": 468 }, { "epoch": 2.1619199074743385, "grad_norm": 0.17671746015548706, "learning_rate": 2.7932098765432103e-06, "loss": 10.023, "step": 469 }, { "epoch": 2.166546190545034, "grad_norm": 0.16406480967998505, "learning_rate": 2.7777777777777783e-06, "loss": 9.8105, "step": 470 }, { "epoch": 2.171172473615729, "grad_norm": 0.11806418746709824, "learning_rate": 2.762345679012346e-06, "loss": 8.2566, "step": 471 }, { "epoch": 2.175798756686425, "grad_norm": 0.32428812980651855, "learning_rate": 2.746913580246914e-06, "loss": 9.1191, "step": 472 }, { "epoch": 2.1804250397571203, "grad_norm": 0.22312945127487183, "learning_rate": 2.7314814814814816e-06, "loss": 9.9265, "step": 473 }, { "epoch": 2.1850513228278157, "grad_norm": 0.32287222146987915, "learning_rate": 2.7160493827160496e-06, "loss": 10.6421, "step": 474 }, { "epoch": 2.189677605898511, "grad_norm": 0.15497097373008728, "learning_rate": 2.700617283950617e-06, "loss": 8.9862, "step": 475 }, { "epoch": 2.1943038889692064, "grad_norm": 0.13556216657161713, "learning_rate": 2.6851851851851856e-06, "loss": 8.658, "step": 476 }, { "epoch": 2.1989301720399017, "grad_norm": 0.17666535079479218, "learning_rate": 2.6697530864197536e-06, "loss": 10.2978, "step": 477 }, { "epoch": 2.203556455110597, "grad_norm": 0.16342324018478394, "learning_rate": 2.6543209876543212e-06, "loss": 8.4887, "step": 478 }, { "epoch": 2.2081827381812924, "grad_norm": 0.2034367471933365, "learning_rate": 2.6388888888888893e-06, "loss": 7.8828, "step": 479 }, { "epoch": 2.2128090212519878, "grad_norm": 0.23170992732048035, "learning_rate": 2.623456790123457e-06, "loss": 9.6533, "step": 480 }, { "epoch": 2.217435304322683, "grad_norm": 0.16840022802352905, "learning_rate": 2.608024691358025e-06, "loss": 8.1141, "step": 481 }, { "epoch": 2.2220615873933784, "grad_norm": 0.1619078814983368, "learning_rate": 2.5925925925925925e-06, "loss": 8.6294, "step": 482 }, { "epoch": 2.2266878704640742, "grad_norm": 0.22204962372779846, "learning_rate": 2.5771604938271605e-06, "loss": 10.4391, "step": 483 }, { "epoch": 2.2313141535347696, "grad_norm": 0.1876753270626068, "learning_rate": 2.561728395061729e-06, "loss": 9.9287, "step": 484 }, { "epoch": 2.235940436605465, "grad_norm": 0.18601296842098236, "learning_rate": 2.5462962962962966e-06, "loss": 9.3439, "step": 485 }, { "epoch": 2.2405667196761603, "grad_norm": 0.17974725365638733, "learning_rate": 2.5308641975308646e-06, "loss": 9.6812, "step": 486 }, { "epoch": 2.2451930027468556, "grad_norm": 0.1832571029663086, "learning_rate": 2.515432098765432e-06, "loss": 7.6883, "step": 487 }, { "epoch": 2.249819285817551, "grad_norm": 0.18652378022670746, "learning_rate": 2.5e-06, "loss": 8.9935, "step": 488 }, { "epoch": 2.2544455688882463, "grad_norm": 0.20331954956054688, "learning_rate": 2.4845679012345682e-06, "loss": 9.1709, "step": 489 }, { "epoch": 2.2590718519589417, "grad_norm": 0.15439340472221375, "learning_rate": 2.469135802469136e-06, "loss": 9.0578, "step": 490 }, { "epoch": 2.263698135029637, "grad_norm": 0.16174978017807007, "learning_rate": 2.453703703703704e-06, "loss": 9.2311, "step": 491 }, { "epoch": 2.2683244181003324, "grad_norm": 0.20344924926757812, "learning_rate": 2.438271604938272e-06, "loss": 9.8727, "step": 492 }, { "epoch": 2.2729507011710277, "grad_norm": 0.19355060160160065, "learning_rate": 2.4228395061728395e-06, "loss": 9.3433, "step": 493 }, { "epoch": 2.2775769842417235, "grad_norm": 0.17609727382659912, "learning_rate": 2.4074074074074075e-06, "loss": 8.8289, "step": 494 }, { "epoch": 2.282203267312419, "grad_norm": 0.16068409383296967, "learning_rate": 2.3919753086419755e-06, "loss": 9.1044, "step": 495 }, { "epoch": 2.286829550383114, "grad_norm": 0.15089063346385956, "learning_rate": 2.3765432098765435e-06, "loss": 8.8363, "step": 496 }, { "epoch": 2.2914558334538095, "grad_norm": 0.3163600265979767, "learning_rate": 2.361111111111111e-06, "loss": 9.3244, "step": 497 }, { "epoch": 2.296082116524505, "grad_norm": 0.18847279250621796, "learning_rate": 2.345679012345679e-06, "loss": 9.1338, "step": 498 }, { "epoch": 2.3007083995952002, "grad_norm": 0.23115108907222748, "learning_rate": 2.330246913580247e-06, "loss": 9.6425, "step": 499 }, { "epoch": 2.3053346826658956, "grad_norm": 0.15251374244689941, "learning_rate": 2.314814814814815e-06, "loss": 9.6591, "step": 500 }, { "epoch": 2.309960965736591, "grad_norm": 0.1706661432981491, "learning_rate": 2.299382716049383e-06, "loss": 10.3708, "step": 501 }, { "epoch": 2.3145872488072863, "grad_norm": 0.16492144763469696, "learning_rate": 2.283950617283951e-06, "loss": 8.8977, "step": 502 }, { "epoch": 2.3192135318779816, "grad_norm": 0.15373440086841583, "learning_rate": 2.268518518518519e-06, "loss": 10.9027, "step": 503 }, { "epoch": 2.323839814948677, "grad_norm": 0.18270155787467957, "learning_rate": 2.2530864197530865e-06, "loss": 9.8633, "step": 504 }, { "epoch": 2.3284660980193728, "grad_norm": 0.22537460923194885, "learning_rate": 2.2376543209876545e-06, "loss": 8.6262, "step": 505 }, { "epoch": 2.333092381090068, "grad_norm": 0.1788664162158966, "learning_rate": 2.222222222222222e-06, "loss": 8.3806, "step": 506 }, { "epoch": 2.3377186641607635, "grad_norm": 0.2230851948261261, "learning_rate": 2.2067901234567905e-06, "loss": 9.6656, "step": 507 }, { "epoch": 2.342344947231459, "grad_norm": 0.157254159450531, "learning_rate": 2.191358024691358e-06, "loss": 9.0586, "step": 508 }, { "epoch": 2.346971230302154, "grad_norm": 0.21109546720981598, "learning_rate": 2.175925925925926e-06, "loss": 10.4577, "step": 509 }, { "epoch": 2.3515975133728495, "grad_norm": 0.16909867525100708, "learning_rate": 2.1604938271604937e-06, "loss": 9.5222, "step": 510 }, { "epoch": 2.356223796443545, "grad_norm": 0.1652510166168213, "learning_rate": 2.145061728395062e-06, "loss": 9.5384, "step": 511 }, { "epoch": 2.36085007951424, "grad_norm": 0.14154984056949615, "learning_rate": 2.1296296296296298e-06, "loss": 9.1477, "step": 512 }, { "epoch": 2.3654763625849355, "grad_norm": 0.18620611727237701, "learning_rate": 2.114197530864198e-06, "loss": 9.8372, "step": 513 }, { "epoch": 2.370102645655631, "grad_norm": 0.17792826890945435, "learning_rate": 2.0987654320987654e-06, "loss": 9.4475, "step": 514 }, { "epoch": 2.3747289287263262, "grad_norm": 0.17278362810611725, "learning_rate": 2.0833333333333334e-06, "loss": 8.2241, "step": 515 }, { "epoch": 2.379355211797022, "grad_norm": 0.17132383584976196, "learning_rate": 2.0679012345679015e-06, "loss": 10.9298, "step": 516 }, { "epoch": 2.3839814948677174, "grad_norm": 0.15168847143650055, "learning_rate": 2.052469135802469e-06, "loss": 9.0072, "step": 517 }, { "epoch": 2.3886077779384127, "grad_norm": 0.186196431517601, "learning_rate": 2.037037037037037e-06, "loss": 9.0923, "step": 518 }, { "epoch": 2.393234061009108, "grad_norm": 0.12113353610038757, "learning_rate": 2.021604938271605e-06, "loss": 10.696, "step": 519 }, { "epoch": 2.3978603440798034, "grad_norm": 0.25200334191322327, "learning_rate": 2.006172839506173e-06, "loss": 8.5775, "step": 520 }, { "epoch": 2.4024866271504988, "grad_norm": 0.1717706322669983, "learning_rate": 1.9907407407407407e-06, "loss": 9.8223, "step": 521 }, { "epoch": 2.407112910221194, "grad_norm": 0.21262796223163605, "learning_rate": 1.9753086419753087e-06, "loss": 9.1006, "step": 522 }, { "epoch": 2.4117391932918895, "grad_norm": 0.15240874886512756, "learning_rate": 1.9598765432098768e-06, "loss": 8.8133, "step": 523 }, { "epoch": 2.416365476362585, "grad_norm": 0.20154468715190887, "learning_rate": 1.944444444444445e-06, "loss": 8.9349, "step": 524 }, { "epoch": 2.42099175943328, "grad_norm": 0.2465580701828003, "learning_rate": 1.9290123456790124e-06, "loss": 9.7463, "step": 525 }, { "epoch": 2.4256180425039755, "grad_norm": 0.262588769197464, "learning_rate": 1.9135802469135804e-06, "loss": 8.9882, "step": 526 }, { "epoch": 2.4302443255746713, "grad_norm": 0.1886514276266098, "learning_rate": 1.8981481481481484e-06, "loss": 10.1046, "step": 527 }, { "epoch": 2.4348706086453666, "grad_norm": 0.15457630157470703, "learning_rate": 1.8827160493827162e-06, "loss": 9.1876, "step": 528 }, { "epoch": 2.439496891716062, "grad_norm": 0.18536067008972168, "learning_rate": 1.867283950617284e-06, "loss": 8.9485, "step": 529 }, { "epoch": 2.4441231747867573, "grad_norm": 0.16247576475143433, "learning_rate": 1.8518518518518519e-06, "loss": 9.7982, "step": 530 }, { "epoch": 2.4487494578574527, "grad_norm": 0.13770771026611328, "learning_rate": 1.83641975308642e-06, "loss": 9.9136, "step": 531 }, { "epoch": 2.453375740928148, "grad_norm": 0.1654641479253769, "learning_rate": 1.820987654320988e-06, "loss": 9.1883, "step": 532 }, { "epoch": 2.4580020239988434, "grad_norm": 0.17362841963768005, "learning_rate": 1.8055555555555557e-06, "loss": 9.0925, "step": 533 }, { "epoch": 2.4626283070695387, "grad_norm": 0.15451891720294952, "learning_rate": 1.7901234567901235e-06, "loss": 9.477, "step": 534 }, { "epoch": 2.467254590140234, "grad_norm": 0.14848832786083221, "learning_rate": 1.7746913580246916e-06, "loss": 10.4986, "step": 535 }, { "epoch": 2.4718808732109294, "grad_norm": 0.18573352694511414, "learning_rate": 1.7592592592592594e-06, "loss": 9.2772, "step": 536 }, { "epoch": 2.4765071562816248, "grad_norm": 0.16295011341571808, "learning_rate": 1.7438271604938272e-06, "loss": 10.3985, "step": 537 }, { "epoch": 2.4811334393523206, "grad_norm": 0.1890224814414978, "learning_rate": 1.7283950617283952e-06, "loss": 8.1752, "step": 538 }, { "epoch": 2.485759722423016, "grad_norm": 0.1372375637292862, "learning_rate": 1.7129629629629632e-06, "loss": 8.903, "step": 539 }, { "epoch": 2.4903860054937113, "grad_norm": 0.17448656260967255, "learning_rate": 1.697530864197531e-06, "loss": 8.5784, "step": 540 }, { "epoch": 2.4950122885644066, "grad_norm": 0.1701819896697998, "learning_rate": 1.6820987654320989e-06, "loss": 8.7201, "step": 541 }, { "epoch": 2.499638571635102, "grad_norm": 0.14620746672153473, "learning_rate": 1.6666666666666667e-06, "loss": 9.211, "step": 542 }, { "epoch": 2.5042648547057973, "grad_norm": 0.16162265837192535, "learning_rate": 1.651234567901235e-06, "loss": 9.4963, "step": 543 }, { "epoch": 2.5088911377764926, "grad_norm": 0.15886524319648743, "learning_rate": 1.6358024691358027e-06, "loss": 8.8424, "step": 544 }, { "epoch": 2.513517420847188, "grad_norm": 0.14556364715099335, "learning_rate": 1.6203703703703705e-06, "loss": 7.9769, "step": 545 }, { "epoch": 2.5181437039178833, "grad_norm": 0.21578781306743622, "learning_rate": 1.6049382716049383e-06, "loss": 8.9909, "step": 546 }, { "epoch": 2.522769986988579, "grad_norm": 0.18850085139274597, "learning_rate": 1.5895061728395064e-06, "loss": 9.8237, "step": 547 }, { "epoch": 2.527396270059274, "grad_norm": 0.1415654569864273, "learning_rate": 1.5740740740740742e-06, "loss": 8.9615, "step": 548 }, { "epoch": 2.53202255312997, "grad_norm": 0.17242810130119324, "learning_rate": 1.5586419753086422e-06, "loss": 8.9936, "step": 549 }, { "epoch": 2.536648836200665, "grad_norm": 0.20581774413585663, "learning_rate": 1.54320987654321e-06, "loss": 9.376, "step": 550 }, { "epoch": 2.5412751192713605, "grad_norm": 0.17876863479614258, "learning_rate": 1.527777777777778e-06, "loss": 8.9055, "step": 551 }, { "epoch": 2.545901402342056, "grad_norm": 0.1470121592283249, "learning_rate": 1.5123456790123458e-06, "loss": 9.3424, "step": 552 }, { "epoch": 2.550527685412751, "grad_norm": 0.23757006227970123, "learning_rate": 1.4969135802469136e-06, "loss": 9.7866, "step": 553 }, { "epoch": 2.5551539684834466, "grad_norm": 0.16553503274917603, "learning_rate": 1.4814814814814815e-06, "loss": 8.6868, "step": 554 }, { "epoch": 2.559780251554142, "grad_norm": 0.12771116197109222, "learning_rate": 1.4660493827160497e-06, "loss": 8.9301, "step": 555 }, { "epoch": 2.5644065346248373, "grad_norm": 0.35117506980895996, "learning_rate": 1.4506172839506175e-06, "loss": 9.3037, "step": 556 }, { "epoch": 2.5690328176955326, "grad_norm": 0.1245264932513237, "learning_rate": 1.4351851851851853e-06, "loss": 9.0822, "step": 557 }, { "epoch": 2.5736591007662284, "grad_norm": 0.18772459030151367, "learning_rate": 1.4197530864197531e-06, "loss": 9.146, "step": 558 }, { "epoch": 2.5782853838369233, "grad_norm": 0.18729011714458466, "learning_rate": 1.4043209876543211e-06, "loss": 8.183, "step": 559 }, { "epoch": 2.582911666907619, "grad_norm": 0.12940698862075806, "learning_rate": 1.3888888888888892e-06, "loss": 9.4745, "step": 560 }, { "epoch": 2.5875379499783144, "grad_norm": 0.16963091492652893, "learning_rate": 1.373456790123457e-06, "loss": 8.8193, "step": 561 }, { "epoch": 2.59216423304901, "grad_norm": 0.17672613263130188, "learning_rate": 1.3580246913580248e-06, "loss": 8.2093, "step": 562 }, { "epoch": 2.596790516119705, "grad_norm": 0.24918967485427856, "learning_rate": 1.3425925925925928e-06, "loss": 9.2257, "step": 563 }, { "epoch": 2.6014167991904005, "grad_norm": 0.15989799797534943, "learning_rate": 1.3271604938271606e-06, "loss": 9.08, "step": 564 }, { "epoch": 2.606043082261096, "grad_norm": 0.14997775852680206, "learning_rate": 1.3117283950617284e-06, "loss": 9.8745, "step": 565 }, { "epoch": 2.610669365331791, "grad_norm": 0.17529721558094025, "learning_rate": 1.2962962962962962e-06, "loss": 8.8959, "step": 566 }, { "epoch": 2.6152956484024865, "grad_norm": 0.16525782644748688, "learning_rate": 1.2808641975308645e-06, "loss": 8.2331, "step": 567 }, { "epoch": 2.619921931473182, "grad_norm": 0.14114739000797272, "learning_rate": 1.2654320987654323e-06, "loss": 8.1592, "step": 568 }, { "epoch": 2.6245482145438777, "grad_norm": 0.21027730405330658, "learning_rate": 1.25e-06, "loss": 9.1182, "step": 569 }, { "epoch": 2.6291744976145726, "grad_norm": 0.13383808732032776, "learning_rate": 1.234567901234568e-06, "loss": 8.8642, "step": 570 }, { "epoch": 2.6338007806852684, "grad_norm": 0.15454170107841492, "learning_rate": 1.219135802469136e-06, "loss": 7.1809, "step": 571 }, { "epoch": 2.6384270637559637, "grad_norm": 0.1575554460287094, "learning_rate": 1.2037037037037037e-06, "loss": 8.5097, "step": 572 }, { "epoch": 2.643053346826659, "grad_norm": 0.16009144484996796, "learning_rate": 1.1882716049382718e-06, "loss": 8.8008, "step": 573 }, { "epoch": 2.6476796298973544, "grad_norm": 0.18514862656593323, "learning_rate": 1.1728395061728396e-06, "loss": 8.3351, "step": 574 }, { "epoch": 2.6523059129680497, "grad_norm": 0.16186843812465668, "learning_rate": 1.1574074074074076e-06, "loss": 8.702, "step": 575 }, { "epoch": 2.656932196038745, "grad_norm": 0.19223704934120178, "learning_rate": 1.1419753086419754e-06, "loss": 9.1931, "step": 576 }, { "epoch": 2.6615584791094404, "grad_norm": 0.1866094172000885, "learning_rate": 1.1265432098765432e-06, "loss": 9.5196, "step": 577 }, { "epoch": 2.666184762180136, "grad_norm": 0.2299501746892929, "learning_rate": 1.111111111111111e-06, "loss": 9.4269, "step": 578 }, { "epoch": 2.670811045250831, "grad_norm": 0.19736136496067047, "learning_rate": 1.095679012345679e-06, "loss": 9.3998, "step": 579 }, { "epoch": 2.675437328321527, "grad_norm": 0.16987043619155884, "learning_rate": 1.0802469135802469e-06, "loss": 8.2513, "step": 580 }, { "epoch": 2.680063611392222, "grad_norm": 0.18940360844135284, "learning_rate": 1.0648148148148149e-06, "loss": 7.8771, "step": 581 }, { "epoch": 2.6846898944629176, "grad_norm": 0.16819283366203308, "learning_rate": 1.0493827160493827e-06, "loss": 8.9367, "step": 582 }, { "epoch": 2.689316177533613, "grad_norm": 0.14170341193675995, "learning_rate": 1.0339506172839507e-06, "loss": 7.1555, "step": 583 }, { "epoch": 2.6939424606043083, "grad_norm": 0.16584675014019012, "learning_rate": 1.0185185185185185e-06, "loss": 9.7405, "step": 584 }, { "epoch": 2.6985687436750037, "grad_norm": 0.15231122076511383, "learning_rate": 1.0030864197530866e-06, "loss": 9.8814, "step": 585 }, { "epoch": 2.703195026745699, "grad_norm": 0.14071619510650635, "learning_rate": 9.876543209876544e-07, "loss": 8.0466, "step": 586 }, { "epoch": 2.7078213098163944, "grad_norm": 0.21130667626857758, "learning_rate": 9.722222222222224e-07, "loss": 10.1551, "step": 587 }, { "epoch": 2.7124475928870897, "grad_norm": 0.14736154675483704, "learning_rate": 9.567901234567902e-07, "loss": 7.5648, "step": 588 }, { "epoch": 2.717073875957785, "grad_norm": 0.18979448080062866, "learning_rate": 9.413580246913581e-07, "loss": 9.5863, "step": 589 }, { "epoch": 2.7217001590284804, "grad_norm": 0.14803054928779602, "learning_rate": 9.259259259259259e-07, "loss": 9.2012, "step": 590 }, { "epoch": 2.726326442099176, "grad_norm": 0.24644902348518372, "learning_rate": 9.10493827160494e-07, "loss": 7.9091, "step": 591 }, { "epoch": 2.730952725169871, "grad_norm": 0.18411195278167725, "learning_rate": 8.950617283950618e-07, "loss": 9.0894, "step": 592 }, { "epoch": 2.735579008240567, "grad_norm": 0.21505457162857056, "learning_rate": 8.796296296296297e-07, "loss": 7.4988, "step": 593 }, { "epoch": 2.7402052913112622, "grad_norm": 0.18174538016319275, "learning_rate": 8.641975308641976e-07, "loss": 8.2626, "step": 594 }, { "epoch": 2.7448315743819576, "grad_norm": 0.1982118785381317, "learning_rate": 8.487654320987655e-07, "loss": 8.415, "step": 595 }, { "epoch": 2.749457857452653, "grad_norm": 0.18941333889961243, "learning_rate": 8.333333333333333e-07, "loss": 8.931, "step": 596 }, { "epoch": 2.7540841405233483, "grad_norm": 0.2093527615070343, "learning_rate": 8.179012345679014e-07, "loss": 9.6871, "step": 597 }, { "epoch": 2.7587104235940436, "grad_norm": 0.25371572375297546, "learning_rate": 8.024691358024692e-07, "loss": 9.719, "step": 598 }, { "epoch": 2.763336706664739, "grad_norm": 0.13767634332180023, "learning_rate": 7.870370370370371e-07, "loss": 7.0931, "step": 599 }, { "epoch": 2.7679629897354343, "grad_norm": 0.18252308666706085, "learning_rate": 7.71604938271605e-07, "loss": 9.0813, "step": 600 }, { "epoch": 2.7725892728061297, "grad_norm": 0.13574370741844177, "learning_rate": 7.561728395061729e-07, "loss": 8.8218, "step": 601 }, { "epoch": 2.7772155558768254, "grad_norm": 0.17202846705913544, "learning_rate": 7.407407407407407e-07, "loss": 8.954, "step": 602 }, { "epoch": 2.7818418389475204, "grad_norm": 0.18582294881343842, "learning_rate": 7.253086419753087e-07, "loss": 7.7078, "step": 603 }, { "epoch": 2.786468122018216, "grad_norm": 0.27104949951171875, "learning_rate": 7.098765432098766e-07, "loss": 8.3435, "step": 604 }, { "epoch": 2.7910944050889115, "grad_norm": 0.32489752769470215, "learning_rate": 6.944444444444446e-07, "loss": 8.5035, "step": 605 }, { "epoch": 2.795720688159607, "grad_norm": 0.1454378217458725, "learning_rate": 6.790123456790124e-07, "loss": 8.4067, "step": 606 }, { "epoch": 2.800346971230302, "grad_norm": 0.20940132439136505, "learning_rate": 6.635802469135803e-07, "loss": 9.4978, "step": 607 }, { "epoch": 2.8049732543009975, "grad_norm": 0.17158959805965424, "learning_rate": 6.481481481481481e-07, "loss": 9.3416, "step": 608 }, { "epoch": 2.809599537371693, "grad_norm": 0.17591050267219543, "learning_rate": 6.327160493827161e-07, "loss": 8.7415, "step": 609 }, { "epoch": 2.8142258204423882, "grad_norm": 0.18380632996559143, "learning_rate": 6.17283950617284e-07, "loss": 8.2376, "step": 610 }, { "epoch": 2.8188521035130836, "grad_norm": 0.19443170726299286, "learning_rate": 6.018518518518519e-07, "loss": 7.562, "step": 611 }, { "epoch": 2.823478386583779, "grad_norm": 0.24004080891609192, "learning_rate": 5.864197530864198e-07, "loss": 9.346, "step": 612 }, { "epoch": 2.8281046696544747, "grad_norm": 0.19434408843517303, "learning_rate": 5.709876543209877e-07, "loss": 6.8662, "step": 613 }, { "epoch": 2.8327309527251696, "grad_norm": 0.1413394808769226, "learning_rate": 5.555555555555555e-07, "loss": 7.7949, "step": 614 }, { "epoch": 2.8373572357958654, "grad_norm": 0.21421702206134796, "learning_rate": 5.401234567901234e-07, "loss": 8.6497, "step": 615 }, { "epoch": 2.8419835188665608, "grad_norm": 0.18243171274662018, "learning_rate": 5.246913580246914e-07, "loss": 9.0258, "step": 616 }, { "epoch": 2.846609801937256, "grad_norm": 0.15553410351276398, "learning_rate": 5.092592592592593e-07, "loss": 9.0613, "step": 617 }, { "epoch": 2.8512360850079514, "grad_norm": 0.1925041526556015, "learning_rate": 4.938271604938272e-07, "loss": 8.8279, "step": 618 }, { "epoch": 2.855862368078647, "grad_norm": 0.25465813279151917, "learning_rate": 4.783950617283951e-07, "loss": 8.5684, "step": 619 }, { "epoch": 2.860488651149342, "grad_norm": 0.17943061888217926, "learning_rate": 4.6296296296296297e-07, "loss": 8.9017, "step": 620 }, { "epoch": 2.8651149342200375, "grad_norm": 0.22506891191005707, "learning_rate": 4.475308641975309e-07, "loss": 8.7791, "step": 621 }, { "epoch": 2.869741217290733, "grad_norm": 0.1924242228269577, "learning_rate": 4.320987654320988e-07, "loss": 8.4592, "step": 622 }, { "epoch": 2.874367500361428, "grad_norm": 0.1336522102355957, "learning_rate": 4.1666666666666667e-07, "loss": 9.2921, "step": 623 }, { "epoch": 2.878993783432124, "grad_norm": 0.16631512343883514, "learning_rate": 4.012345679012346e-07, "loss": 8.7771, "step": 624 }, { "epoch": 2.883620066502819, "grad_norm": 0.13000570237636566, "learning_rate": 3.858024691358025e-07, "loss": 7.6173, "step": 625 }, { "epoch": 2.8882463495735147, "grad_norm": 0.16806651651859283, "learning_rate": 3.7037037037037036e-07, "loss": 8.2835, "step": 626 }, { "epoch": 2.89287263264421, "grad_norm": 0.16624124348163605, "learning_rate": 3.549382716049383e-07, "loss": 9.0295, "step": 627 }, { "epoch": 2.8974989157149054, "grad_norm": 0.18300163745880127, "learning_rate": 3.395061728395062e-07, "loss": 9.3673, "step": 628 }, { "epoch": 2.9021251987856007, "grad_norm": 0.1591711938381195, "learning_rate": 3.2407407407407406e-07, "loss": 9.2437, "step": 629 }, { "epoch": 2.906751481856296, "grad_norm": 0.1935225874185562, "learning_rate": 3.08641975308642e-07, "loss": 8.1388, "step": 630 }, { "epoch": 2.9113777649269914, "grad_norm": 0.18095123767852783, "learning_rate": 2.932098765432099e-07, "loss": 8.8998, "step": 631 }, { "epoch": 2.9160040479976868, "grad_norm": 0.15141603350639343, "learning_rate": 2.7777777777777776e-07, "loss": 9.5553, "step": 632 }, { "epoch": 2.920630331068382, "grad_norm": 0.1664353758096695, "learning_rate": 2.623456790123457e-07, "loss": 9.2671, "step": 633 }, { "epoch": 2.9252566141390774, "grad_norm": 0.14757901430130005, "learning_rate": 2.469135802469136e-07, "loss": 9.3753, "step": 634 }, { "epoch": 2.9298828972097732, "grad_norm": 0.12832246720790863, "learning_rate": 2.3148148148148148e-07, "loss": 7.55, "step": 635 }, { "epoch": 2.934509180280468, "grad_norm": 0.14796899259090424, "learning_rate": 2.160493827160494e-07, "loss": 8.9455, "step": 636 }, { "epoch": 2.939135463351164, "grad_norm": 0.16375280916690826, "learning_rate": 2.006172839506173e-07, "loss": 9.1212, "step": 637 }, { "epoch": 2.9437617464218593, "grad_norm": 0.15986521542072296, "learning_rate": 1.8518518518518518e-07, "loss": 8.0966, "step": 638 }, { "epoch": 2.9483880294925546, "grad_norm": 0.14000752568244934, "learning_rate": 1.697530864197531e-07, "loss": 8.1283, "step": 639 }, { "epoch": 2.95301431256325, "grad_norm": 0.10660137236118317, "learning_rate": 1.54320987654321e-07, "loss": 7.7013, "step": 640 }, { "epoch": 2.9576405956339453, "grad_norm": 0.12879547476768494, "learning_rate": 1.3888888888888888e-07, "loss": 8.2472, "step": 641 }, { "epoch": 2.9622668787046407, "grad_norm": 0.24027042090892792, "learning_rate": 1.234567901234568e-07, "loss": 9.3202, "step": 642 }, { "epoch": 2.966893161775336, "grad_norm": 0.18989317119121552, "learning_rate": 1.080246913580247e-07, "loss": 8.6135, "step": 643 }, { "epoch": 2.9715194448460314, "grad_norm": 0.20127813518047333, "learning_rate": 9.259259259259259e-08, "loss": 8.3557, "step": 644 }, { "epoch": 2.9761457279167267, "grad_norm": 0.20816563069820404, "learning_rate": 7.71604938271605e-08, "loss": 9.6997, "step": 645 }, { "epoch": 2.9807720109874225, "grad_norm": 0.2171671837568283, "learning_rate": 6.17283950617284e-08, "loss": 9.6809, "step": 646 }, { "epoch": 2.9853982940581174, "grad_norm": 0.18569743633270264, "learning_rate": 4.6296296296296295e-08, "loss": 8.4721, "step": 647 }, { "epoch": 2.990024577128813, "grad_norm": 0.16961927711963654, "learning_rate": 3.08641975308642e-08, "loss": 9.1963, "step": 648 }, { "epoch": 2.990024577128813, "step": 648, "total_flos": 2.9427853089130414e+18, "train_loss": 10.740490390930647, "train_runtime": 52216.2733, "train_samples_per_second": 1.59, "train_steps_per_second": 0.012 } ], "logging_steps": 1.0, "max_steps": 648, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.9427853089130414e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }