{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9996783531682213, "eval_steps": 500, "global_step": 1554, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.000643293663557414, "grad_norm": 14.195452602693646, "learning_rate": 4.25531914893617e-08, "loss": 0.9377, "step": 1 }, { "epoch": 0.001286587327114828, "grad_norm": 12.562526279423325, "learning_rate": 8.51063829787234e-08, "loss": 0.7437, "step": 2 }, { "epoch": 0.0019298809906722419, "grad_norm": 13.969400136430751, "learning_rate": 1.2765957446808508e-07, "loss": 0.8448, "step": 3 }, { "epoch": 0.002573174654229656, "grad_norm": 13.144844296988808, "learning_rate": 1.702127659574468e-07, "loss": 0.8799, "step": 4 }, { "epoch": 0.0032164683177870698, "grad_norm": 14.744627427062406, "learning_rate": 2.127659574468085e-07, "loss": 0.7967, "step": 5 }, { "epoch": 0.0038597619813444837, "grad_norm": 11.031961572887615, "learning_rate": 2.5531914893617016e-07, "loss": 0.8347, "step": 6 }, { "epoch": 0.004503055644901898, "grad_norm": 12.675025434500915, "learning_rate": 2.978723404255319e-07, "loss": 0.7732, "step": 7 }, { "epoch": 0.005146349308459312, "grad_norm": 12.667052412777513, "learning_rate": 3.404255319148936e-07, "loss": 1.0298, "step": 8 }, { "epoch": 0.005789642972016726, "grad_norm": 12.52841911964234, "learning_rate": 3.829787234042553e-07, "loss": 0.768, "step": 9 }, { "epoch": 0.0064329366355741395, "grad_norm": 12.90764051200186, "learning_rate": 4.25531914893617e-07, "loss": 0.856, "step": 10 }, { "epoch": 0.007076230299131554, "grad_norm": 13.847565855218129, "learning_rate": 4.6808510638297873e-07, "loss": 0.7912, "step": 11 }, { "epoch": 0.0077195239626889674, "grad_norm": 10.362826178805946, "learning_rate": 5.106382978723403e-07, "loss": 0.9174, "step": 12 }, { "epoch": 0.008362817626246381, "grad_norm": 13.021945803791676, "learning_rate": 5.531914893617021e-07, "loss": 0.7403, "step": 13 }, { "epoch": 0.009006111289803796, "grad_norm": 12.428379909874558, "learning_rate": 5.957446808510638e-07, "loss": 0.8688, "step": 14 }, { "epoch": 0.00964940495336121, "grad_norm": 13.525652179538579, "learning_rate": 6.382978723404255e-07, "loss": 0.9104, "step": 15 }, { "epoch": 0.010292698616918623, "grad_norm": 10.837600034613665, "learning_rate": 6.808510638297872e-07, "loss": 0.724, "step": 16 }, { "epoch": 0.010935992280476037, "grad_norm": 10.914705230259882, "learning_rate": 7.23404255319149e-07, "loss": 0.8035, "step": 17 }, { "epoch": 0.011579285944033452, "grad_norm": 10.344303009612716, "learning_rate": 7.659574468085106e-07, "loss": 0.706, "step": 18 }, { "epoch": 0.012222579607590866, "grad_norm": 8.931469895580307, "learning_rate": 8.085106382978723e-07, "loss": 0.8626, "step": 19 }, { "epoch": 0.012865873271148279, "grad_norm": 9.166333334528717, "learning_rate": 8.51063829787234e-07, "loss": 0.7236, "step": 20 }, { "epoch": 0.013509166934705693, "grad_norm": 6.032964926857977, "learning_rate": 8.936170212765957e-07, "loss": 0.6926, "step": 21 }, { "epoch": 0.014152460598263108, "grad_norm": 6.771174414362514, "learning_rate": 9.361702127659575e-07, "loss": 0.6679, "step": 22 }, { "epoch": 0.014795754261820521, "grad_norm": 4.824288233389681, "learning_rate": 9.78723404255319e-07, "loss": 0.7216, "step": 23 }, { "epoch": 0.015439047925377935, "grad_norm": 4.297060430986501, "learning_rate": 1.0212765957446806e-06, "loss": 0.62, "step": 24 }, { "epoch": 0.01608234158893535, "grad_norm": 4.323595061786908, "learning_rate": 1.0638297872340424e-06, "loss": 0.5187, "step": 25 }, { "epoch": 0.016725635252492762, "grad_norm": 4.3498174796703415, "learning_rate": 1.1063829787234042e-06, "loss": 0.6889, "step": 26 }, { "epoch": 0.017368928916050175, "grad_norm": 3.354698985518625, "learning_rate": 1.148936170212766e-06, "loss": 0.6601, "step": 27 }, { "epoch": 0.018012222579607592, "grad_norm": 3.9773230778415702, "learning_rate": 1.1914893617021276e-06, "loss": 0.6117, "step": 28 }, { "epoch": 0.018655516243165006, "grad_norm": 4.0765680273835025, "learning_rate": 1.2340425531914892e-06, "loss": 0.5808, "step": 29 }, { "epoch": 0.01929880990672242, "grad_norm": 4.887322129358923, "learning_rate": 1.276595744680851e-06, "loss": 0.5784, "step": 30 }, { "epoch": 0.019942103570279833, "grad_norm": 4.633995858756539, "learning_rate": 1.3191489361702126e-06, "loss": 0.6372, "step": 31 }, { "epoch": 0.020585397233837247, "grad_norm": 6.154813160470098, "learning_rate": 1.3617021276595744e-06, "loss": 0.609, "step": 32 }, { "epoch": 0.02122869089739466, "grad_norm": 4.177768735695119, "learning_rate": 1.4042553191489362e-06, "loss": 0.6225, "step": 33 }, { "epoch": 0.021871984560952074, "grad_norm": 3.3941179320823154, "learning_rate": 1.446808510638298e-06, "loss": 0.5938, "step": 34 }, { "epoch": 0.022515278224509487, "grad_norm": 4.181664446487992, "learning_rate": 1.4893617021276594e-06, "loss": 0.6497, "step": 35 }, { "epoch": 0.023158571888066904, "grad_norm": 3.509432207535639, "learning_rate": 1.5319148936170212e-06, "loss": 0.6043, "step": 36 }, { "epoch": 0.023801865551624318, "grad_norm": 3.9116444081233106, "learning_rate": 1.574468085106383e-06, "loss": 0.6274, "step": 37 }, { "epoch": 0.02444515921518173, "grad_norm": 3.07802408776738, "learning_rate": 1.6170212765957446e-06, "loss": 0.579, "step": 38 }, { "epoch": 0.025088452878739145, "grad_norm": 3.231905423687573, "learning_rate": 1.6595744680851064e-06, "loss": 0.6675, "step": 39 }, { "epoch": 0.025731746542296558, "grad_norm": 3.453140679604751, "learning_rate": 1.702127659574468e-06, "loss": 0.6497, "step": 40 }, { "epoch": 0.02637504020585397, "grad_norm": 3.807628113411218, "learning_rate": 1.7446808510638297e-06, "loss": 0.4901, "step": 41 }, { "epoch": 0.027018333869411385, "grad_norm": 2.991136498110161, "learning_rate": 1.7872340425531913e-06, "loss": 0.654, "step": 42 }, { "epoch": 0.0276616275329688, "grad_norm": 3.339836631271973, "learning_rate": 1.8297872340425531e-06, "loss": 0.5199, "step": 43 }, { "epoch": 0.028304921196526216, "grad_norm": 2.970978612870501, "learning_rate": 1.872340425531915e-06, "loss": 0.5192, "step": 44 }, { "epoch": 0.02894821486008363, "grad_norm": 3.057986001034445, "learning_rate": 1.9148936170212767e-06, "loss": 0.607, "step": 45 }, { "epoch": 0.029591508523641043, "grad_norm": 4.132443571924768, "learning_rate": 1.957446808510638e-06, "loss": 0.6036, "step": 46 }, { "epoch": 0.030234802187198456, "grad_norm": 2.8809364973099583, "learning_rate": 2e-06, "loss": 0.5132, "step": 47 }, { "epoch": 0.03087809585075587, "grad_norm": 2.863583030549207, "learning_rate": 1.9999978270832495e-06, "loss": 0.4422, "step": 48 }, { "epoch": 0.03152138951431328, "grad_norm": 2.8391867721028023, "learning_rate": 1.9999913083424407e-06, "loss": 0.4572, "step": 49 }, { "epoch": 0.0321646831778707, "grad_norm": 2.5450203691999005, "learning_rate": 1.999980443805904e-06, "loss": 0.4908, "step": 50 }, { "epoch": 0.03280797684142811, "grad_norm": 2.6722975150931356, "learning_rate": 1.9999652335208536e-06, "loss": 0.5207, "step": 51 }, { "epoch": 0.033451270504985524, "grad_norm": 3.0212265295784433, "learning_rate": 1.9999456775533917e-06, "loss": 0.5919, "step": 52 }, { "epoch": 0.03409456416854294, "grad_norm": 2.677774522651171, "learning_rate": 1.999921775988505e-06, "loss": 0.3821, "step": 53 }, { "epoch": 0.03473785783210035, "grad_norm": 2.855582223082552, "learning_rate": 1.999893528930066e-06, "loss": 0.5335, "step": 54 }, { "epoch": 0.035381151495657764, "grad_norm": 3.903053917511972, "learning_rate": 1.999860936500831e-06, "loss": 0.4657, "step": 55 }, { "epoch": 0.036024445159215185, "grad_norm": 2.897968036492491, "learning_rate": 1.999823998842442e-06, "loss": 0.4578, "step": 56 }, { "epoch": 0.0366677388227726, "grad_norm": 2.896741877752234, "learning_rate": 1.9997827161154238e-06, "loss": 0.5503, "step": 57 }, { "epoch": 0.03731103248633001, "grad_norm": 2.179764834030277, "learning_rate": 1.999737088499184e-06, "loss": 0.389, "step": 58 }, { "epoch": 0.037954326149887425, "grad_norm": 2.8033535931454625, "learning_rate": 1.9996871161920126e-06, "loss": 0.4397, "step": 59 }, { "epoch": 0.03859761981344484, "grad_norm": 2.7905043164209706, "learning_rate": 1.9996327994110807e-06, "loss": 0.5166, "step": 60 }, { "epoch": 0.03924091347700225, "grad_norm": 2.6230024305919493, "learning_rate": 1.9995741383924412e-06, "loss": 0.5552, "step": 61 }, { "epoch": 0.039884207140559666, "grad_norm": 2.4303968289355824, "learning_rate": 1.9995111333910236e-06, "loss": 0.4453, "step": 62 }, { "epoch": 0.04052750080411708, "grad_norm": 2.7678880597006095, "learning_rate": 1.9994437846806383e-06, "loss": 0.4419, "step": 63 }, { "epoch": 0.04117079446767449, "grad_norm": 2.6890299022507844, "learning_rate": 1.999372092553971e-06, "loss": 0.5375, "step": 64 }, { "epoch": 0.04181408813123191, "grad_norm": 2.2862321981216094, "learning_rate": 1.9992960573225832e-06, "loss": 0.3622, "step": 65 }, { "epoch": 0.04245738179478932, "grad_norm": 3.2119944533254796, "learning_rate": 1.999215679316913e-06, "loss": 0.54, "step": 66 }, { "epoch": 0.043100675458346734, "grad_norm": 2.8235369662933767, "learning_rate": 1.9991309588862683e-06, "loss": 0.4827, "step": 67 }, { "epoch": 0.04374396912190415, "grad_norm": 2.661614118141481, "learning_rate": 1.9990418963988303e-06, "loss": 0.5665, "step": 68 }, { "epoch": 0.04438726278546156, "grad_norm": 2.6696823368366993, "learning_rate": 1.99894849224165e-06, "loss": 0.5273, "step": 69 }, { "epoch": 0.045030556449018974, "grad_norm": 2.962675797827742, "learning_rate": 1.9988507468206465e-06, "loss": 0.5475, "step": 70 }, { "epoch": 0.04567385011257639, "grad_norm": 3.050567893945894, "learning_rate": 1.998748660560605e-06, "loss": 0.5404, "step": 71 }, { "epoch": 0.04631714377613381, "grad_norm": 3.205258606476403, "learning_rate": 1.998642233905175e-06, "loss": 0.5232, "step": 72 }, { "epoch": 0.04696043743969122, "grad_norm": 2.7828660855759964, "learning_rate": 1.9985314673168694e-06, "loss": 0.4325, "step": 73 }, { "epoch": 0.047603731103248635, "grad_norm": 3.0583759313672765, "learning_rate": 1.9984163612770613e-06, "loss": 0.5161, "step": 74 }, { "epoch": 0.04824702476680605, "grad_norm": 2.4289515589953554, "learning_rate": 1.9982969162859823e-06, "loss": 0.4118, "step": 75 }, { "epoch": 0.04889031843036346, "grad_norm": 2.76143084990133, "learning_rate": 1.9981731328627206e-06, "loss": 0.4035, "step": 76 }, { "epoch": 0.049533612093920876, "grad_norm": 2.994871818214066, "learning_rate": 1.998045011545218e-06, "loss": 0.4957, "step": 77 }, { "epoch": 0.05017690575747829, "grad_norm": 2.680618245027871, "learning_rate": 1.9979125528902687e-06, "loss": 0.4422, "step": 78 }, { "epoch": 0.0508201994210357, "grad_norm": 2.4375681189208307, "learning_rate": 1.997775757473516e-06, "loss": 0.4754, "step": 79 }, { "epoch": 0.051463493084593116, "grad_norm": 2.949577631202856, "learning_rate": 1.99763462588945e-06, "loss": 0.523, "step": 80 }, { "epoch": 0.05210678674815053, "grad_norm": 3.2193695678707996, "learning_rate": 1.997489158751405e-06, "loss": 0.5065, "step": 81 }, { "epoch": 0.05275008041170794, "grad_norm": 3.052842628001906, "learning_rate": 1.9973393566915566e-06, "loss": 0.4694, "step": 82 }, { "epoch": 0.05339337407526536, "grad_norm": 2.552647143575962, "learning_rate": 1.99718522036092e-06, "loss": 0.4733, "step": 83 }, { "epoch": 0.05403666773882277, "grad_norm": 3.755295142598979, "learning_rate": 1.997026750429346e-06, "loss": 0.5512, "step": 84 }, { "epoch": 0.054679961402380184, "grad_norm": 2.7876823575046514, "learning_rate": 1.996863947585519e-06, "loss": 0.4087, "step": 85 }, { "epoch": 0.0553232550659376, "grad_norm": 4.888678609760293, "learning_rate": 1.996696812536952e-06, "loss": 0.5053, "step": 86 }, { "epoch": 0.05596654872949502, "grad_norm": 2.2995250943343444, "learning_rate": 1.996525346009987e-06, "loss": 0.3973, "step": 87 }, { "epoch": 0.05660984239305243, "grad_norm": 2.743132483926581, "learning_rate": 1.9963495487497885e-06, "loss": 0.4846, "step": 88 }, { "epoch": 0.057253136056609845, "grad_norm": 2.839079434159953, "learning_rate": 1.996169421520342e-06, "loss": 0.4374, "step": 89 }, { "epoch": 0.05789642972016726, "grad_norm": 2.510888710671949, "learning_rate": 1.9959849651044515e-06, "loss": 0.5191, "step": 90 }, { "epoch": 0.05853972338372467, "grad_norm": 2.4316270167425107, "learning_rate": 1.9957961803037326e-06, "loss": 0.3696, "step": 91 }, { "epoch": 0.059183017047282085, "grad_norm": 2.812493178728756, "learning_rate": 1.9956030679386127e-06, "loss": 0.4926, "step": 92 }, { "epoch": 0.0598263107108395, "grad_norm": 2.5735300856983896, "learning_rate": 1.995405628848327e-06, "loss": 0.4666, "step": 93 }, { "epoch": 0.06046960437439691, "grad_norm": 2.6101944716121936, "learning_rate": 1.995203863890912e-06, "loss": 0.4605, "step": 94 }, { "epoch": 0.061112898037954326, "grad_norm": 3.348100457415799, "learning_rate": 1.994997773943205e-06, "loss": 0.437, "step": 95 }, { "epoch": 0.06175619170151174, "grad_norm": 2.3307659416480333, "learning_rate": 1.9947873599008385e-06, "loss": 0.4623, "step": 96 }, { "epoch": 0.06239948536506915, "grad_norm": 2.575966690302254, "learning_rate": 1.9945726226782366e-06, "loss": 0.492, "step": 97 }, { "epoch": 0.06304277902862657, "grad_norm": 2.414378721033498, "learning_rate": 1.994353563208612e-06, "loss": 0.4809, "step": 98 }, { "epoch": 0.06368607269218399, "grad_norm": 2.5638877247325125, "learning_rate": 1.9941301824439606e-06, "loss": 0.3891, "step": 99 }, { "epoch": 0.0643293663557414, "grad_norm": 2.6558138161083313, "learning_rate": 1.9939024813550577e-06, "loss": 0.3939, "step": 100 }, { "epoch": 0.06497266001929881, "grad_norm": 2.4162394379097107, "learning_rate": 1.993670460931455e-06, "loss": 0.3907, "step": 101 }, { "epoch": 0.06561595368285622, "grad_norm": 2.931605329584879, "learning_rate": 1.9934341221814736e-06, "loss": 0.4513, "step": 102 }, { "epoch": 0.06625924734641364, "grad_norm": 2.7867875453884112, "learning_rate": 1.9931934661322033e-06, "loss": 0.4125, "step": 103 }, { "epoch": 0.06690254100997105, "grad_norm": 3.1729323606089324, "learning_rate": 1.9929484938294948e-06, "loss": 0.4873, "step": 104 }, { "epoch": 0.06754583467352847, "grad_norm": 2.097576989226628, "learning_rate": 1.9926992063379567e-06, "loss": 0.3343, "step": 105 }, { "epoch": 0.06818912833708587, "grad_norm": 2.404743225971747, "learning_rate": 1.9924456047409514e-06, "loss": 0.392, "step": 106 }, { "epoch": 0.0688324220006433, "grad_norm": 3.2381968998854482, "learning_rate": 1.992187690140589e-06, "loss": 0.5323, "step": 107 }, { "epoch": 0.0694757156642007, "grad_norm": 2.7180013455900927, "learning_rate": 1.9919254636577233e-06, "loss": 0.3723, "step": 108 }, { "epoch": 0.07011900932775812, "grad_norm": 3.314581183690992, "learning_rate": 1.991658926431947e-06, "loss": 0.4295, "step": 109 }, { "epoch": 0.07076230299131553, "grad_norm": 2.629550826518601, "learning_rate": 1.991388079621587e-06, "loss": 0.4098, "step": 110 }, { "epoch": 0.07140559665487295, "grad_norm": 2.7380735731804093, "learning_rate": 1.9911129244036978e-06, "loss": 0.5011, "step": 111 }, { "epoch": 0.07204889031843037, "grad_norm": 2.671157341576739, "learning_rate": 1.9908334619740585e-06, "loss": 0.4481, "step": 112 }, { "epoch": 0.07269218398198778, "grad_norm": 2.3917620141375666, "learning_rate": 1.990549693547166e-06, "loss": 0.5085, "step": 113 }, { "epoch": 0.0733354776455452, "grad_norm": 2.42359481081289, "learning_rate": 1.9902616203562306e-06, "loss": 0.4654, "step": 114 }, { "epoch": 0.0739787713091026, "grad_norm": 2.540027936728536, "learning_rate": 1.9899692436531707e-06, "loss": 0.4416, "step": 115 }, { "epoch": 0.07462206497266002, "grad_norm": 2.5632769234267156, "learning_rate": 1.989672564708607e-06, "loss": 0.4509, "step": 116 }, { "epoch": 0.07526535863621743, "grad_norm": 3.356318764264533, "learning_rate": 1.9893715848118563e-06, "loss": 0.505, "step": 117 }, { "epoch": 0.07590865229977485, "grad_norm": 2.48471042190396, "learning_rate": 1.9890663052709275e-06, "loss": 0.4412, "step": 118 }, { "epoch": 0.07655194596333226, "grad_norm": 3.3593554218964496, "learning_rate": 1.988756727412514e-06, "loss": 0.4981, "step": 119 }, { "epoch": 0.07719523962688968, "grad_norm": 2.9830700540090853, "learning_rate": 1.9884428525819903e-06, "loss": 0.4722, "step": 120 }, { "epoch": 0.07783853329044708, "grad_norm": 2.6408971957130185, "learning_rate": 1.9881246821434044e-06, "loss": 0.4765, "step": 121 }, { "epoch": 0.0784818269540045, "grad_norm": 2.7844796240046654, "learning_rate": 1.987802217479471e-06, "loss": 0.5256, "step": 122 }, { "epoch": 0.07912512061756191, "grad_norm": 2.6880267550869887, "learning_rate": 1.9874754599915686e-06, "loss": 0.4593, "step": 123 }, { "epoch": 0.07976841428111933, "grad_norm": 2.6339053636176493, "learning_rate": 1.9871444110997308e-06, "loss": 0.3693, "step": 124 }, { "epoch": 0.08041170794467674, "grad_norm": 2.9543453583963326, "learning_rate": 1.986809072242641e-06, "loss": 0.3749, "step": 125 }, { "epoch": 0.08105500160823416, "grad_norm": 2.6340359135086207, "learning_rate": 1.9864694448776257e-06, "loss": 0.4353, "step": 126 }, { "epoch": 0.08169829527179158, "grad_norm": 3.0707389205091213, "learning_rate": 1.986125530480649e-06, "loss": 0.549, "step": 127 }, { "epoch": 0.08234158893534899, "grad_norm": 3.0835367966564204, "learning_rate": 1.985777330546306e-06, "loss": 0.4508, "step": 128 }, { "epoch": 0.0829848825989064, "grad_norm": 2.8522076577749838, "learning_rate": 1.985424846587815e-06, "loss": 0.3722, "step": 129 }, { "epoch": 0.08362817626246381, "grad_norm": 2.773290050393778, "learning_rate": 1.9850680801370128e-06, "loss": 0.4469, "step": 130 }, { "epoch": 0.08427146992602123, "grad_norm": 2.786960431154452, "learning_rate": 1.9847070327443475e-06, "loss": 0.4115, "step": 131 }, { "epoch": 0.08491476358957864, "grad_norm": 2.8939612230251566, "learning_rate": 1.98434170597887e-06, "loss": 0.4276, "step": 132 }, { "epoch": 0.08555805725313606, "grad_norm": 2.5196224856457787, "learning_rate": 1.9839721014282305e-06, "loss": 0.4555, "step": 133 }, { "epoch": 0.08620135091669347, "grad_norm": 2.9020534954391, "learning_rate": 1.983598220698668e-06, "loss": 0.4904, "step": 134 }, { "epoch": 0.08684464458025089, "grad_norm": 2.3914445789038767, "learning_rate": 1.9832200654150074e-06, "loss": 0.4167, "step": 135 }, { "epoch": 0.0874879382438083, "grad_norm": 2.5104830325545984, "learning_rate": 1.982837637220647e-06, "loss": 0.4288, "step": 136 }, { "epoch": 0.08813123190736571, "grad_norm": 2.6044230559088035, "learning_rate": 1.9824509377775566e-06, "loss": 0.401, "step": 137 }, { "epoch": 0.08877452557092312, "grad_norm": 2.6146016382768265, "learning_rate": 1.9820599687662676e-06, "loss": 0.4714, "step": 138 }, { "epoch": 0.08941781923448054, "grad_norm": 2.1756490564929942, "learning_rate": 1.9816647318858665e-06, "loss": 0.3567, "step": 139 }, { "epoch": 0.09006111289803795, "grad_norm": 2.839962301734442, "learning_rate": 1.9812652288539867e-06, "loss": 0.4857, "step": 140 }, { "epoch": 0.09070440656159537, "grad_norm": 2.4373851809919427, "learning_rate": 1.9808614614068018e-06, "loss": 0.4856, "step": 141 }, { "epoch": 0.09134770022515278, "grad_norm": 2.637314199067112, "learning_rate": 1.980453431299018e-06, "loss": 0.4084, "step": 142 }, { "epoch": 0.0919909938887102, "grad_norm": 2.106524684142874, "learning_rate": 1.980041140303866e-06, "loss": 0.4058, "step": 143 }, { "epoch": 0.09263428755226762, "grad_norm": 2.691921837012061, "learning_rate": 1.9796245902130946e-06, "loss": 0.4661, "step": 144 }, { "epoch": 0.09327758121582502, "grad_norm": 2.16704862339589, "learning_rate": 1.9792037828369605e-06, "loss": 0.3783, "step": 145 }, { "epoch": 0.09392087487938244, "grad_norm": 3.0658617205245426, "learning_rate": 1.978778720004222e-06, "loss": 0.5525, "step": 146 }, { "epoch": 0.09456416854293985, "grad_norm": 3.2979404778523635, "learning_rate": 1.9783494035621325e-06, "loss": 0.4886, "step": 147 }, { "epoch": 0.09520746220649727, "grad_norm": 2.9889772026820443, "learning_rate": 1.977915835376429e-06, "loss": 0.4831, "step": 148 }, { "epoch": 0.09585075587005468, "grad_norm": 2.486707951880786, "learning_rate": 1.977478017331327e-06, "loss": 0.4397, "step": 149 }, { "epoch": 0.0964940495336121, "grad_norm": 2.4358436090883595, "learning_rate": 1.9770359513295106e-06, "loss": 0.3704, "step": 150 }, { "epoch": 0.0971373431971695, "grad_norm": 3.233218659808801, "learning_rate": 1.976589639292125e-06, "loss": 0.3438, "step": 151 }, { "epoch": 0.09778063686072692, "grad_norm": 2.6606822903829963, "learning_rate": 1.9761390831587687e-06, "loss": 0.4455, "step": 152 }, { "epoch": 0.09842393052428433, "grad_norm": 2.9417589938828446, "learning_rate": 1.9756842848874827e-06, "loss": 0.4694, "step": 153 }, { "epoch": 0.09906722418784175, "grad_norm": 3.0763765579228695, "learning_rate": 1.975225246454745e-06, "loss": 0.3621, "step": 154 }, { "epoch": 0.09971051785139916, "grad_norm": 2.7358116363985325, "learning_rate": 1.9747619698554602e-06, "loss": 0.43, "step": 155 }, { "epoch": 0.10035381151495658, "grad_norm": 2.687901237710609, "learning_rate": 1.9742944571029513e-06, "loss": 0.4409, "step": 156 }, { "epoch": 0.10099710517851399, "grad_norm": 2.2817230497608825, "learning_rate": 1.9738227102289505e-06, "loss": 0.3848, "step": 157 }, { "epoch": 0.1016403988420714, "grad_norm": 3.0899990142166076, "learning_rate": 1.9733467312835917e-06, "loss": 0.4811, "step": 158 }, { "epoch": 0.10228369250562883, "grad_norm": 2.49822420484545, "learning_rate": 1.9728665223354e-06, "loss": 0.335, "step": 159 }, { "epoch": 0.10292698616918623, "grad_norm": 2.7140528495794047, "learning_rate": 1.9723820854712833e-06, "loss": 0.3955, "step": 160 }, { "epoch": 0.10357027983274365, "grad_norm": 2.815211567089355, "learning_rate": 1.971893422796524e-06, "loss": 0.5539, "step": 161 }, { "epoch": 0.10421357349630106, "grad_norm": 2.5697375229850925, "learning_rate": 1.971400536434768e-06, "loss": 0.3997, "step": 162 }, { "epoch": 0.10485686715985848, "grad_norm": 2.786868655579082, "learning_rate": 1.9709034285280185e-06, "loss": 0.4559, "step": 163 }, { "epoch": 0.10550016082341589, "grad_norm": 2.823794616478473, "learning_rate": 1.9704021012366226e-06, "loss": 0.4267, "step": 164 }, { "epoch": 0.10614345448697331, "grad_norm": 3.309662602498236, "learning_rate": 1.9698965567392656e-06, "loss": 0.5521, "step": 165 }, { "epoch": 0.10678674815053071, "grad_norm": 2.255337059656187, "learning_rate": 1.9693867972329595e-06, "loss": 0.3754, "step": 166 }, { "epoch": 0.10743004181408813, "grad_norm": 2.9605428606442357, "learning_rate": 1.9688728249330347e-06, "loss": 0.4319, "step": 167 }, { "epoch": 0.10807333547764554, "grad_norm": 2.683870068441683, "learning_rate": 1.968354642073129e-06, "loss": 0.4847, "step": 168 }, { "epoch": 0.10871662914120296, "grad_norm": 2.967784333330616, "learning_rate": 1.9678322509051787e-06, "loss": 0.4118, "step": 169 }, { "epoch": 0.10935992280476037, "grad_norm": 2.539086599336895, "learning_rate": 1.9673056536994094e-06, "loss": 0.4549, "step": 170 }, { "epoch": 0.11000321646831779, "grad_norm": 3.125313367672374, "learning_rate": 1.9667748527443245e-06, "loss": 0.5192, "step": 171 }, { "epoch": 0.1106465101318752, "grad_norm": 2.5239188362010623, "learning_rate": 1.966239850346696e-06, "loss": 0.4673, "step": 172 }, { "epoch": 0.11128980379543262, "grad_norm": 2.8231510117869356, "learning_rate": 1.9657006488315563e-06, "loss": 0.3905, "step": 173 }, { "epoch": 0.11193309745899004, "grad_norm": 2.6259984229172546, "learning_rate": 1.9651572505421845e-06, "loss": 0.4076, "step": 174 }, { "epoch": 0.11257639112254744, "grad_norm": 2.5546423812090295, "learning_rate": 1.9646096578400995e-06, "loss": 0.4349, "step": 175 }, { "epoch": 0.11321968478610486, "grad_norm": 2.3097600462995236, "learning_rate": 1.9640578731050485e-06, "loss": 0.4375, "step": 176 }, { "epoch": 0.11386297844966227, "grad_norm": 2.8905019947083574, "learning_rate": 1.9635018987349953e-06, "loss": 0.5124, "step": 177 }, { "epoch": 0.11450627211321969, "grad_norm": 2.4709146140289486, "learning_rate": 1.9629417371461123e-06, "loss": 0.4768, "step": 178 }, { "epoch": 0.1151495657767771, "grad_norm": 2.5307502798764148, "learning_rate": 1.962377390772768e-06, "loss": 0.4337, "step": 179 }, { "epoch": 0.11579285944033452, "grad_norm": 3.0279948337406943, "learning_rate": 1.9618088620675184e-06, "loss": 0.4047, "step": 180 }, { "epoch": 0.11643615310389192, "grad_norm": 2.8905001172365496, "learning_rate": 1.9612361535010947e-06, "loss": 0.5211, "step": 181 }, { "epoch": 0.11707944676744934, "grad_norm": 2.52983462639843, "learning_rate": 1.9606592675623925e-06, "loss": 0.4651, "step": 182 }, { "epoch": 0.11772274043100675, "grad_norm": 3.039914144243252, "learning_rate": 1.960078206758462e-06, "loss": 0.5112, "step": 183 }, { "epoch": 0.11836603409456417, "grad_norm": 2.364638347509278, "learning_rate": 1.9594929736144973e-06, "loss": 0.4089, "step": 184 }, { "epoch": 0.11900932775812158, "grad_norm": 2.590008309458127, "learning_rate": 1.9589035706738236e-06, "loss": 0.4338, "step": 185 }, { "epoch": 0.119652621421679, "grad_norm": 2.8837780018093837, "learning_rate": 1.9583100004978883e-06, "loss": 0.4771, "step": 186 }, { "epoch": 0.1202959150852364, "grad_norm": 2.5515011308511517, "learning_rate": 1.9577122656662484e-06, "loss": 0.3656, "step": 187 }, { "epoch": 0.12093920874879382, "grad_norm": 2.928544150792617, "learning_rate": 1.9571103687765596e-06, "loss": 0.4544, "step": 188 }, { "epoch": 0.12158250241235123, "grad_norm": 2.4692351356498246, "learning_rate": 1.9565043124445663e-06, "loss": 0.4603, "step": 189 }, { "epoch": 0.12222579607590865, "grad_norm": 2.987769624658221, "learning_rate": 1.9558940993040883e-06, "loss": 0.3986, "step": 190 }, { "epoch": 0.12286908973946607, "grad_norm": 2.8398854803142046, "learning_rate": 1.9552797320070095e-06, "loss": 0.4133, "step": 191 }, { "epoch": 0.12351238340302348, "grad_norm": 3.051426202752587, "learning_rate": 1.954661213223269e-06, "loss": 0.4285, "step": 192 }, { "epoch": 0.1241556770665809, "grad_norm": 2.575083352874075, "learning_rate": 1.9540385456408454e-06, "loss": 0.3756, "step": 193 }, { "epoch": 0.1247989707301383, "grad_norm": 2.390609147340149, "learning_rate": 1.953411731965749e-06, "loss": 0.447, "step": 194 }, { "epoch": 0.1254422643936957, "grad_norm": 2.257888574769132, "learning_rate": 1.9527807749220077e-06, "loss": 0.3656, "step": 195 }, { "epoch": 0.12608555805725313, "grad_norm": 2.778103109912705, "learning_rate": 1.952145677251655e-06, "loss": 0.4531, "step": 196 }, { "epoch": 0.12672885172081055, "grad_norm": 2.669566852882443, "learning_rate": 1.9515064417147207e-06, "loss": 0.498, "step": 197 }, { "epoch": 0.12737214538436797, "grad_norm": 3.562635048991329, "learning_rate": 1.9508630710892152e-06, "loss": 0.4845, "step": 198 }, { "epoch": 0.12801543904792537, "grad_norm": 2.6135289003410085, "learning_rate": 1.9502155681711207e-06, "loss": 0.4426, "step": 199 }, { "epoch": 0.1286587327114828, "grad_norm": 2.292875901770694, "learning_rate": 1.949563935774377e-06, "loss": 0.3798, "step": 200 }, { "epoch": 0.1293020263750402, "grad_norm": 2.98397460116356, "learning_rate": 1.9489081767308697e-06, "loss": 0.4287, "step": 201 }, { "epoch": 0.12994532003859763, "grad_norm": 2.720651704775655, "learning_rate": 1.948248293890418e-06, "loss": 0.4334, "step": 202 }, { "epoch": 0.13058861370215502, "grad_norm": 2.823733039819591, "learning_rate": 1.9475842901207638e-06, "loss": 0.4594, "step": 203 }, { "epoch": 0.13123190736571244, "grad_norm": 2.8351509890928077, "learning_rate": 1.9469161683075565e-06, "loss": 0.456, "step": 204 }, { "epoch": 0.13187520102926986, "grad_norm": 2.5566518859249183, "learning_rate": 1.946243931354342e-06, "loss": 0.4125, "step": 205 }, { "epoch": 0.13251849469282728, "grad_norm": 2.643290766653592, "learning_rate": 1.945567582182551e-06, "loss": 0.4452, "step": 206 }, { "epoch": 0.1331617883563847, "grad_norm": 2.953802468265646, "learning_rate": 1.944887123731483e-06, "loss": 0.4343, "step": 207 }, { "epoch": 0.1338050820199421, "grad_norm": 2.3961645379432994, "learning_rate": 1.9442025589582985e-06, "loss": 0.472, "step": 208 }, { "epoch": 0.13444837568349952, "grad_norm": 2.3524372871813055, "learning_rate": 1.9435138908380013e-06, "loss": 0.3378, "step": 209 }, { "epoch": 0.13509166934705694, "grad_norm": 2.48480736649653, "learning_rate": 1.942821122363429e-06, "loss": 0.3939, "step": 210 }, { "epoch": 0.13573496301061436, "grad_norm": 2.662611686438726, "learning_rate": 1.942124256545237e-06, "loss": 0.3596, "step": 211 }, { "epoch": 0.13637825667417175, "grad_norm": 2.887739056470739, "learning_rate": 1.9414232964118892e-06, "loss": 0.4159, "step": 212 }, { "epoch": 0.13702155033772917, "grad_norm": 2.9255511601394852, "learning_rate": 1.940718245009641e-06, "loss": 0.4407, "step": 213 }, { "epoch": 0.1376648440012866, "grad_norm": 3.020676594603636, "learning_rate": 1.9400091054025282e-06, "loss": 0.4336, "step": 214 }, { "epoch": 0.138308137664844, "grad_norm": 2.5883652423350187, "learning_rate": 1.939295880672354e-06, "loss": 0.4383, "step": 215 }, { "epoch": 0.1389514313284014, "grad_norm": 2.8438460470794835, "learning_rate": 1.9385785739186746e-06, "loss": 0.5108, "step": 216 }, { "epoch": 0.13959472499195882, "grad_norm": 2.333414680282332, "learning_rate": 1.9378571882587845e-06, "loss": 0.4384, "step": 217 }, { "epoch": 0.14023801865551624, "grad_norm": 2.3287319223051615, "learning_rate": 1.937131726827707e-06, "loss": 0.4536, "step": 218 }, { "epoch": 0.14088131231907366, "grad_norm": 2.835148042161422, "learning_rate": 1.936402192778176e-06, "loss": 0.4392, "step": 219 }, { "epoch": 0.14152460598263106, "grad_norm": 2.8659023148759095, "learning_rate": 1.9356685892806252e-06, "loss": 0.4109, "step": 220 }, { "epoch": 0.14216789964618848, "grad_norm": 2.2168446900039434, "learning_rate": 1.934930919523173e-06, "loss": 0.3276, "step": 221 }, { "epoch": 0.1428111933097459, "grad_norm": 2.7744952118975403, "learning_rate": 1.93418918671161e-06, "loss": 0.4034, "step": 222 }, { "epoch": 0.14345448697330332, "grad_norm": 2.7239578937089917, "learning_rate": 1.9334433940693826e-06, "loss": 0.4538, "step": 223 }, { "epoch": 0.14409778063686074, "grad_norm": 2.4736319436371086, "learning_rate": 1.932693544837582e-06, "loss": 0.3928, "step": 224 }, { "epoch": 0.14474107430041813, "grad_norm": 2.6286649302494567, "learning_rate": 1.931939642274928e-06, "loss": 0.3867, "step": 225 }, { "epoch": 0.14538436796397555, "grad_norm": 2.675079798177804, "learning_rate": 1.9311816896577556e-06, "loss": 0.4345, "step": 226 }, { "epoch": 0.14602766162753297, "grad_norm": 3.148742890109798, "learning_rate": 1.9304196902800014e-06, "loss": 0.4862, "step": 227 }, { "epoch": 0.1466709552910904, "grad_norm": 2.5613965870541358, "learning_rate": 1.9296536474531864e-06, "loss": 0.435, "step": 228 }, { "epoch": 0.1473142489546478, "grad_norm": 2.3821725520483903, "learning_rate": 1.928883564506406e-06, "loss": 0.3949, "step": 229 }, { "epoch": 0.1479575426182052, "grad_norm": 3.061333925789048, "learning_rate": 1.9281094447863123e-06, "loss": 0.4683, "step": 230 }, { "epoch": 0.14860083628176263, "grad_norm": 2.234065162657386, "learning_rate": 1.927331291657101e-06, "loss": 0.3813, "step": 231 }, { "epoch": 0.14924412994532005, "grad_norm": 2.7501428628124227, "learning_rate": 1.926549108500495e-06, "loss": 0.4378, "step": 232 }, { "epoch": 0.14988742360887744, "grad_norm": 3.033674590745771, "learning_rate": 1.925762898715734e-06, "loss": 0.4262, "step": 233 }, { "epoch": 0.15053071727243486, "grad_norm": 2.2913491277313893, "learning_rate": 1.9249726657195533e-06, "loss": 0.4126, "step": 234 }, { "epoch": 0.15117401093599228, "grad_norm": 2.5831175161917415, "learning_rate": 1.924178412946174e-06, "loss": 0.4442, "step": 235 }, { "epoch": 0.1518173045995497, "grad_norm": 2.4951420794159853, "learning_rate": 1.9233801438472872e-06, "loss": 0.4832, "step": 236 }, { "epoch": 0.15246059826310712, "grad_norm": 2.926333156221301, "learning_rate": 1.9225778618920367e-06, "loss": 0.457, "step": 237 }, { "epoch": 0.15310389192666451, "grad_norm": 2.4840417651881053, "learning_rate": 1.921771570567007e-06, "loss": 0.4476, "step": 238 }, { "epoch": 0.15374718559022194, "grad_norm": 2.463042208581304, "learning_rate": 1.9209612733762054e-06, "loss": 0.4226, "step": 239 }, { "epoch": 0.15439047925377936, "grad_norm": 2.499309726643823, "learning_rate": 1.920146973841049e-06, "loss": 0.3701, "step": 240 }, { "epoch": 0.15503377291733678, "grad_norm": 3.0734195461858236, "learning_rate": 1.9193286755003474e-06, "loss": 0.4622, "step": 241 }, { "epoch": 0.15567706658089417, "grad_norm": 2.977527571600767, "learning_rate": 1.9185063819102895e-06, "loss": 0.4665, "step": 242 }, { "epoch": 0.1563203602444516, "grad_norm": 2.6764720834447, "learning_rate": 1.917680096644426e-06, "loss": 0.5011, "step": 243 }, { "epoch": 0.156963653908009, "grad_norm": 2.6244004858976795, "learning_rate": 1.9168498232936553e-06, "loss": 0.4265, "step": 244 }, { "epoch": 0.15760694757156643, "grad_norm": 3.0901610544119533, "learning_rate": 1.9160155654662073e-06, "loss": 0.4913, "step": 245 }, { "epoch": 0.15825024123512382, "grad_norm": 2.7712219095376063, "learning_rate": 1.915177326787627e-06, "loss": 0.4877, "step": 246 }, { "epoch": 0.15889353489868124, "grad_norm": 2.6779205315606047, "learning_rate": 1.914335110900761e-06, "loss": 0.4669, "step": 247 }, { "epoch": 0.15953682856223866, "grad_norm": 2.5951072454899715, "learning_rate": 1.9134889214657384e-06, "loss": 0.3366, "step": 248 }, { "epoch": 0.16018012222579608, "grad_norm": 2.873097652767785, "learning_rate": 1.9126387621599583e-06, "loss": 0.3817, "step": 249 }, { "epoch": 0.16082341588935348, "grad_norm": 2.42494244097029, "learning_rate": 1.911784636678071e-06, "loss": 0.4464, "step": 250 }, { "epoch": 0.1614667095529109, "grad_norm": 2.4114511445855538, "learning_rate": 1.9109265487319643e-06, "loss": 0.3848, "step": 251 }, { "epoch": 0.16211000321646832, "grad_norm": 2.455798483289963, "learning_rate": 1.910064502050745e-06, "loss": 0.4027, "step": 252 }, { "epoch": 0.16275329688002574, "grad_norm": 3.596926470387692, "learning_rate": 1.909198500380724e-06, "loss": 0.5526, "step": 253 }, { "epoch": 0.16339659054358316, "grad_norm": 2.8032000966565778, "learning_rate": 1.908328547485401e-06, "loss": 0.3828, "step": 254 }, { "epoch": 0.16403988420714055, "grad_norm": 2.295281891740367, "learning_rate": 1.907454647145447e-06, "loss": 0.4414, "step": 255 }, { "epoch": 0.16468317787069797, "grad_norm": 2.805571045697796, "learning_rate": 1.906576803158686e-06, "loss": 0.4273, "step": 256 }, { "epoch": 0.1653264715342554, "grad_norm": 2.803032642615905, "learning_rate": 1.905695019340083e-06, "loss": 0.4138, "step": 257 }, { "epoch": 0.1659697651978128, "grad_norm": 2.6935599816738067, "learning_rate": 1.904809299521723e-06, "loss": 0.453, "step": 258 }, { "epoch": 0.1666130588613702, "grad_norm": 2.30834660857966, "learning_rate": 1.9039196475527973e-06, "loss": 0.3524, "step": 259 }, { "epoch": 0.16725635252492763, "grad_norm": 2.6229578021840174, "learning_rate": 1.9030260672995846e-06, "loss": 0.4247, "step": 260 }, { "epoch": 0.16789964618848505, "grad_norm": 3.222566869011792, "learning_rate": 1.9021285626454366e-06, "loss": 0.4295, "step": 261 }, { "epoch": 0.16854293985204247, "grad_norm": 2.6715848364335764, "learning_rate": 1.9012271374907588e-06, "loss": 0.5142, "step": 262 }, { "epoch": 0.16918623351559986, "grad_norm": 2.507629992450772, "learning_rate": 1.9003217957529945e-06, "loss": 0.347, "step": 263 }, { "epoch": 0.16982952717915728, "grad_norm": 2.798374832934813, "learning_rate": 1.8994125413666086e-06, "loss": 0.3681, "step": 264 }, { "epoch": 0.1704728208427147, "grad_norm": 2.5937955162408914, "learning_rate": 1.898499378283069e-06, "loss": 0.3874, "step": 265 }, { "epoch": 0.17111611450627212, "grad_norm": 2.306541254075938, "learning_rate": 1.8975823104708311e-06, "loss": 0.406, "step": 266 }, { "epoch": 0.1717594081698295, "grad_norm": 2.8185354751245963, "learning_rate": 1.8966613419153178e-06, "loss": 0.4028, "step": 267 }, { "epoch": 0.17240270183338693, "grad_norm": 2.741366592854792, "learning_rate": 1.8957364766189059e-06, "loss": 0.4434, "step": 268 }, { "epoch": 0.17304599549694435, "grad_norm": 2.7627285013337217, "learning_rate": 1.894807718600906e-06, "loss": 0.3434, "step": 269 }, { "epoch": 0.17368928916050178, "grad_norm": 2.3410359860317826, "learning_rate": 1.8938750718975454e-06, "loss": 0.4538, "step": 270 }, { "epoch": 0.1743325828240592, "grad_norm": 2.6708866513936793, "learning_rate": 1.8929385405619515e-06, "loss": 0.4541, "step": 271 }, { "epoch": 0.1749758764876166, "grad_norm": 2.4656710767335146, "learning_rate": 1.8919981286641336e-06, "loss": 0.3731, "step": 272 }, { "epoch": 0.175619170151174, "grad_norm": 2.240470424851887, "learning_rate": 1.8910538402909654e-06, "loss": 0.4211, "step": 273 }, { "epoch": 0.17626246381473143, "grad_norm": 2.87025161016199, "learning_rate": 1.8901056795461667e-06, "loss": 0.416, "step": 274 }, { "epoch": 0.17690575747828885, "grad_norm": 2.6812721625712004, "learning_rate": 1.8891536505502862e-06, "loss": 0.4538, "step": 275 }, { "epoch": 0.17754905114184624, "grad_norm": 2.7772477416249606, "learning_rate": 1.8881977574406837e-06, "loss": 0.4355, "step": 276 }, { "epoch": 0.17819234480540366, "grad_norm": 2.5102243477464454, "learning_rate": 1.8872380043715115e-06, "loss": 0.3718, "step": 277 }, { "epoch": 0.17883563846896108, "grad_norm": 3.33977255157816, "learning_rate": 1.8862743955136963e-06, "loss": 0.4361, "step": 278 }, { "epoch": 0.1794789321325185, "grad_norm": 2.4312416622909985, "learning_rate": 1.885306935054922e-06, "loss": 0.3971, "step": 279 }, { "epoch": 0.1801222257960759, "grad_norm": 2.6073601772537156, "learning_rate": 1.8843356271996106e-06, "loss": 0.4093, "step": 280 }, { "epoch": 0.18076551945963332, "grad_norm": 2.5133912494450183, "learning_rate": 1.8833604761689043e-06, "loss": 0.5393, "step": 281 }, { "epoch": 0.18140881312319074, "grad_norm": 2.4789853852543287, "learning_rate": 1.8823814862006468e-06, "loss": 0.4433, "step": 282 }, { "epoch": 0.18205210678674816, "grad_norm": 2.890074584306752, "learning_rate": 1.8813986615493663e-06, "loss": 0.4718, "step": 283 }, { "epoch": 0.18269540045030555, "grad_norm": 2.811002023177097, "learning_rate": 1.8804120064862545e-06, "loss": 0.4559, "step": 284 }, { "epoch": 0.18333869411386297, "grad_norm": 2.3860852841858233, "learning_rate": 1.87942152529915e-06, "loss": 0.3262, "step": 285 }, { "epoch": 0.1839819877774204, "grad_norm": 2.600582011878044, "learning_rate": 1.8784272222925198e-06, "loss": 0.476, "step": 286 }, { "epoch": 0.1846252814409778, "grad_norm": 2.4577598805954315, "learning_rate": 1.8774291017874383e-06, "loss": 0.4518, "step": 287 }, { "epoch": 0.18526857510453523, "grad_norm": 2.8228748434483912, "learning_rate": 1.8764271681215713e-06, "loss": 0.4364, "step": 288 }, { "epoch": 0.18591186876809263, "grad_norm": 2.4702505306605413, "learning_rate": 1.875421425649156e-06, "loss": 0.4534, "step": 289 }, { "epoch": 0.18655516243165005, "grad_norm": 2.7315968416930194, "learning_rate": 1.8744118787409816e-06, "loss": 0.4838, "step": 290 }, { "epoch": 0.18719845609520747, "grad_norm": 2.317329919988444, "learning_rate": 1.8733985317843707e-06, "loss": 0.3979, "step": 291 }, { "epoch": 0.1878417497587649, "grad_norm": 2.6602718445609037, "learning_rate": 1.8723813891831605e-06, "loss": 0.4562, "step": 292 }, { "epoch": 0.18848504342232228, "grad_norm": 2.858501652568693, "learning_rate": 1.8713604553576838e-06, "loss": 0.4333, "step": 293 }, { "epoch": 0.1891283370858797, "grad_norm": 2.206587638461408, "learning_rate": 1.8703357347447486e-06, "loss": 0.3575, "step": 294 }, { "epoch": 0.18977163074943712, "grad_norm": 2.549123358725168, "learning_rate": 1.8693072317976201e-06, "loss": 0.4064, "step": 295 }, { "epoch": 0.19041492441299454, "grad_norm": 3.203554622608017, "learning_rate": 1.868274950986001e-06, "loss": 0.4219, "step": 296 }, { "epoch": 0.19105821807655193, "grad_norm": 2.7528950977814506, "learning_rate": 1.8672388967960114e-06, "loss": 0.3452, "step": 297 }, { "epoch": 0.19170151174010935, "grad_norm": 2.40952298416949, "learning_rate": 1.8661990737301711e-06, "loss": 0.3871, "step": 298 }, { "epoch": 0.19234480540366677, "grad_norm": 2.4234160426871836, "learning_rate": 1.8651554863073775e-06, "loss": 0.4231, "step": 299 }, { "epoch": 0.1929880990672242, "grad_norm": 2.4988018261235303, "learning_rate": 1.8641081390628876e-06, "loss": 0.4494, "step": 300 }, { "epoch": 0.19363139273078162, "grad_norm": 2.707870573740472, "learning_rate": 1.8630570365482986e-06, "loss": 0.4238, "step": 301 }, { "epoch": 0.194274686394339, "grad_norm": 2.6166703355133385, "learning_rate": 1.862002183331527e-06, "loss": 0.3737, "step": 302 }, { "epoch": 0.19491798005789643, "grad_norm": 2.4293260021537515, "learning_rate": 1.8609435839967889e-06, "loss": 0.366, "step": 303 }, { "epoch": 0.19556127372145385, "grad_norm": 2.340153108035801, "learning_rate": 1.859881243144581e-06, "loss": 0.3764, "step": 304 }, { "epoch": 0.19620456738501127, "grad_norm": 2.86988621351114, "learning_rate": 1.8588151653916595e-06, "loss": 0.3708, "step": 305 }, { "epoch": 0.19684786104856866, "grad_norm": 2.3903310579307617, "learning_rate": 1.857745355371021e-06, "loss": 0.3793, "step": 306 }, { "epoch": 0.19749115471212608, "grad_norm": 2.6352359524332227, "learning_rate": 1.856671817731882e-06, "loss": 0.4173, "step": 307 }, { "epoch": 0.1981344483756835, "grad_norm": 2.5751417515242645, "learning_rate": 1.855594557139658e-06, "loss": 0.3732, "step": 308 }, { "epoch": 0.19877774203924092, "grad_norm": 2.2179382730519355, "learning_rate": 1.8545135782759442e-06, "loss": 0.4069, "step": 309 }, { "epoch": 0.19942103570279832, "grad_norm": 2.53339092793023, "learning_rate": 1.853428885838495e-06, "loss": 0.4252, "step": 310 }, { "epoch": 0.20006432936635574, "grad_norm": 2.7708827259672404, "learning_rate": 1.8523404845412025e-06, "loss": 0.3331, "step": 311 }, { "epoch": 0.20070762302991316, "grad_norm": 2.5092079129384044, "learning_rate": 1.8512483791140782e-06, "loss": 0.4519, "step": 312 }, { "epoch": 0.20135091669347058, "grad_norm": 2.739751223185084, "learning_rate": 1.8501525743032302e-06, "loss": 0.4475, "step": 313 }, { "epoch": 0.20199421035702797, "grad_norm": 2.8872041143974925, "learning_rate": 1.849053074870844e-06, "loss": 0.5822, "step": 314 }, { "epoch": 0.2026375040205854, "grad_norm": 3.2094972249586724, "learning_rate": 1.8479498855951604e-06, "loss": 0.4084, "step": 315 }, { "epoch": 0.2032807976841428, "grad_norm": 2.7032185106427504, "learning_rate": 1.846843011270457e-06, "loss": 0.4407, "step": 316 }, { "epoch": 0.20392409134770023, "grad_norm": 2.679216569955498, "learning_rate": 1.8457324567070251e-06, "loss": 0.3701, "step": 317 }, { "epoch": 0.20456738501125765, "grad_norm": 2.73732229591967, "learning_rate": 1.8446182267311499e-06, "loss": 0.4536, "step": 318 }, { "epoch": 0.20521067867481504, "grad_norm": 2.353099482083382, "learning_rate": 1.8435003261850891e-06, "loss": 0.383, "step": 319 }, { "epoch": 0.20585397233837247, "grad_norm": 2.301831914995752, "learning_rate": 1.8423787599270528e-06, "loss": 0.3683, "step": 320 }, { "epoch": 0.20649726600192989, "grad_norm": 2.459790735769469, "learning_rate": 1.8412535328311812e-06, "loss": 0.4135, "step": 321 }, { "epoch": 0.2071405596654873, "grad_norm": 2.830726132521204, "learning_rate": 1.8401246497875235e-06, "loss": 0.4538, "step": 322 }, { "epoch": 0.2077838533290447, "grad_norm": 2.6828771843615065, "learning_rate": 1.8389921157020177e-06, "loss": 0.4571, "step": 323 }, { "epoch": 0.20842714699260212, "grad_norm": 2.240512001023228, "learning_rate": 1.8378559354964682e-06, "loss": 0.3427, "step": 324 }, { "epoch": 0.20907044065615954, "grad_norm": 2.791987409090214, "learning_rate": 1.8367161141085252e-06, "loss": 0.4604, "step": 325 }, { "epoch": 0.20971373431971696, "grad_norm": 2.621525622134155, "learning_rate": 1.8355726564916628e-06, "loss": 0.3881, "step": 326 }, { "epoch": 0.21035702798327435, "grad_norm": 2.9210090161032207, "learning_rate": 1.8344255676151569e-06, "loss": 0.4381, "step": 327 }, { "epoch": 0.21100032164683177, "grad_norm": 2.813482178288473, "learning_rate": 1.8332748524640656e-06, "loss": 0.522, "step": 328 }, { "epoch": 0.2116436153103892, "grad_norm": 2.787348760055315, "learning_rate": 1.8321205160392046e-06, "loss": 0.3562, "step": 329 }, { "epoch": 0.21228690897394661, "grad_norm": 2.817845189991954, "learning_rate": 1.830962563357128e-06, "loss": 0.5072, "step": 330 }, { "epoch": 0.212930202637504, "grad_norm": 2.5097247464978554, "learning_rate": 1.8298009994501056e-06, "loss": 0.4136, "step": 331 }, { "epoch": 0.21357349630106143, "grad_norm": 2.637653095919712, "learning_rate": 1.8286358293661004e-06, "loss": 0.4005, "step": 332 }, { "epoch": 0.21421678996461885, "grad_norm": 2.3161744940145486, "learning_rate": 1.8274670581687478e-06, "loss": 0.3702, "step": 333 }, { "epoch": 0.21486008362817627, "grad_norm": 2.5820631981289277, "learning_rate": 1.8262946909373326e-06, "loss": 0.3889, "step": 334 }, { "epoch": 0.2155033772917337, "grad_norm": 2.5724603918065254, "learning_rate": 1.8251187327667682e-06, "loss": 0.408, "step": 335 }, { "epoch": 0.21614667095529108, "grad_norm": 2.7770319030060455, "learning_rate": 1.8239391887675722e-06, "loss": 0.406, "step": 336 }, { "epoch": 0.2167899646188485, "grad_norm": 2.455738097667333, "learning_rate": 1.8227560640658469e-06, "loss": 0.3622, "step": 337 }, { "epoch": 0.21743325828240592, "grad_norm": 2.4736947604093067, "learning_rate": 1.8215693638032553e-06, "loss": 0.3121, "step": 338 }, { "epoch": 0.21807655194596334, "grad_norm": 2.315797850791143, "learning_rate": 1.820379093136999e-06, "loss": 0.4117, "step": 339 }, { "epoch": 0.21871984560952074, "grad_norm": 3.0042830290860905, "learning_rate": 1.819185257239796e-06, "loss": 0.5264, "step": 340 }, { "epoch": 0.21936313927307816, "grad_norm": 2.5280660605509127, "learning_rate": 1.8179878612998585e-06, "loss": 0.3736, "step": 341 }, { "epoch": 0.22000643293663558, "grad_norm": 2.615456508762799, "learning_rate": 1.81678691052087e-06, "loss": 0.3422, "step": 342 }, { "epoch": 0.220649726600193, "grad_norm": 2.449964843304598, "learning_rate": 1.8155824101219625e-06, "loss": 0.4259, "step": 343 }, { "epoch": 0.2212930202637504, "grad_norm": 2.6595051612859786, "learning_rate": 1.8143743653376943e-06, "loss": 0.448, "step": 344 }, { "epoch": 0.2219363139273078, "grad_norm": 2.6893313190483843, "learning_rate": 1.8131627814180263e-06, "loss": 0.4224, "step": 345 }, { "epoch": 0.22257960759086523, "grad_norm": 2.315030296025039, "learning_rate": 1.8119476636283015e-06, "loss": 0.3735, "step": 346 }, { "epoch": 0.22322290125442265, "grad_norm": 2.7108464822369682, "learning_rate": 1.810729017249219e-06, "loss": 0.4628, "step": 347 }, { "epoch": 0.22386619491798007, "grad_norm": 2.3657969901740765, "learning_rate": 1.8095068475768129e-06, "loss": 0.4178, "step": 348 }, { "epoch": 0.22450948858153746, "grad_norm": 3.154131513257161, "learning_rate": 1.8082811599224291e-06, "loss": 0.3427, "step": 349 }, { "epoch": 0.22515278224509488, "grad_norm": 2.311139277747802, "learning_rate": 1.8070519596127025e-06, "loss": 0.2975, "step": 350 }, { "epoch": 0.2257960759086523, "grad_norm": 2.2678121591948366, "learning_rate": 1.8058192519895325e-06, "loss": 0.3415, "step": 351 }, { "epoch": 0.22643936957220973, "grad_norm": 2.5559238697945985, "learning_rate": 1.8045830424100617e-06, "loss": 0.4016, "step": 352 }, { "epoch": 0.22708266323576712, "grad_norm": 2.6416337790585898, "learning_rate": 1.8033433362466503e-06, "loss": 0.4339, "step": 353 }, { "epoch": 0.22772595689932454, "grad_norm": 2.2826199672546212, "learning_rate": 1.8021001388868557e-06, "loss": 0.3443, "step": 354 }, { "epoch": 0.22836925056288196, "grad_norm": 2.6601342944122206, "learning_rate": 1.8008534557334063e-06, "loss": 0.4321, "step": 355 }, { "epoch": 0.22901254422643938, "grad_norm": 2.6563790844282336, "learning_rate": 1.7996032922041796e-06, "loss": 0.3808, "step": 356 }, { "epoch": 0.22965583788999677, "grad_norm": 2.3130308906025365, "learning_rate": 1.7983496537321778e-06, "loss": 0.3769, "step": 357 }, { "epoch": 0.2302991315535542, "grad_norm": 2.4979464517565746, "learning_rate": 1.7970925457655053e-06, "loss": 0.4243, "step": 358 }, { "epoch": 0.2309424252171116, "grad_norm": 2.428043585315322, "learning_rate": 1.7958319737673442e-06, "loss": 0.342, "step": 359 }, { "epoch": 0.23158571888066903, "grad_norm": 2.4800874795638586, "learning_rate": 1.7945679432159299e-06, "loss": 0.3736, "step": 360 }, { "epoch": 0.23222901254422643, "grad_norm": 2.4608016834023947, "learning_rate": 1.7933004596045294e-06, "loss": 0.3392, "step": 361 }, { "epoch": 0.23287230620778385, "grad_norm": 2.742474243143699, "learning_rate": 1.792029528441415e-06, "loss": 0.4031, "step": 362 }, { "epoch": 0.23351559987134127, "grad_norm": 2.7863471242832203, "learning_rate": 1.790755155249842e-06, "loss": 0.4608, "step": 363 }, { "epoch": 0.2341588935348987, "grad_norm": 2.208572809154207, "learning_rate": 1.7894773455680246e-06, "loss": 0.4036, "step": 364 }, { "epoch": 0.2348021871984561, "grad_norm": 2.8518974031822832, "learning_rate": 1.7881961049491099e-06, "loss": 0.4059, "step": 365 }, { "epoch": 0.2354454808620135, "grad_norm": 2.487782402196943, "learning_rate": 1.7869114389611573e-06, "loss": 0.3985, "step": 366 }, { "epoch": 0.23608877452557092, "grad_norm": 2.271993736736465, "learning_rate": 1.7856233531871109e-06, "loss": 0.3774, "step": 367 }, { "epoch": 0.23673206818912834, "grad_norm": 2.4355343088979917, "learning_rate": 1.7843318532247763e-06, "loss": 0.3059, "step": 368 }, { "epoch": 0.23737536185268576, "grad_norm": 2.2873436520534565, "learning_rate": 1.7830369446867988e-06, "loss": 0.3628, "step": 369 }, { "epoch": 0.23801865551624315, "grad_norm": 2.13448528745857, "learning_rate": 1.7817386332006342e-06, "loss": 0.3766, "step": 370 }, { "epoch": 0.23866194917980058, "grad_norm": 2.522129139915578, "learning_rate": 1.7804369244085283e-06, "loss": 0.3661, "step": 371 }, { "epoch": 0.239305242843358, "grad_norm": 2.4922506544377017, "learning_rate": 1.7791318239674912e-06, "loss": 0.3924, "step": 372 }, { "epoch": 0.23994853650691542, "grad_norm": 2.6159855784520682, "learning_rate": 1.7778233375492712e-06, "loss": 0.3811, "step": 373 }, { "epoch": 0.2405918301704728, "grad_norm": 2.2819207857429906, "learning_rate": 1.776511470840333e-06, "loss": 0.3309, "step": 374 }, { "epoch": 0.24123512383403023, "grad_norm": 2.2687827421619366, "learning_rate": 1.7751962295418312e-06, "loss": 0.3943, "step": 375 }, { "epoch": 0.24187841749758765, "grad_norm": 2.5742824564032802, "learning_rate": 1.773877619369585e-06, "loss": 0.3718, "step": 376 }, { "epoch": 0.24252171116114507, "grad_norm": 2.649800986774041, "learning_rate": 1.772555646054055e-06, "loss": 0.3795, "step": 377 }, { "epoch": 0.24316500482470246, "grad_norm": 2.637631672478734, "learning_rate": 1.7712303153403165e-06, "loss": 0.4143, "step": 378 }, { "epoch": 0.24380829848825988, "grad_norm": 2.225706974271865, "learning_rate": 1.769901632988037e-06, "loss": 0.3419, "step": 379 }, { "epoch": 0.2444515921518173, "grad_norm": 2.922438081654237, "learning_rate": 1.7685696047714481e-06, "loss": 0.4529, "step": 380 }, { "epoch": 0.24509488581537472, "grad_norm": 2.543811268621379, "learning_rate": 1.767234236479323e-06, "loss": 0.3748, "step": 381 }, { "epoch": 0.24573817947893214, "grad_norm": 2.454993467447398, "learning_rate": 1.7658955339149494e-06, "loss": 0.4464, "step": 382 }, { "epoch": 0.24638147314248954, "grad_norm": 2.791874748371137, "learning_rate": 1.7645535028961068e-06, "loss": 0.4206, "step": 383 }, { "epoch": 0.24702476680604696, "grad_norm": 2.4048083431434293, "learning_rate": 1.7632081492550377e-06, "loss": 0.3274, "step": 384 }, { "epoch": 0.24766806046960438, "grad_norm": 2.4653546397526007, "learning_rate": 1.7618594788384252e-06, "loss": 0.3754, "step": 385 }, { "epoch": 0.2483113541331618, "grad_norm": 2.7093432259978183, "learning_rate": 1.7605074975073663e-06, "loss": 0.4609, "step": 386 }, { "epoch": 0.2489546477967192, "grad_norm": 2.714601296529243, "learning_rate": 1.7591522111373475e-06, "loss": 0.4, "step": 387 }, { "epoch": 0.2495979414602766, "grad_norm": 2.647536844805239, "learning_rate": 1.7577936256182167e-06, "loss": 0.4068, "step": 388 }, { "epoch": 0.250241235123834, "grad_norm": 2.407334401545142, "learning_rate": 1.7564317468541607e-06, "loss": 0.3694, "step": 389 }, { "epoch": 0.2508845287873914, "grad_norm": 2.236220883385462, "learning_rate": 1.7550665807636782e-06, "loss": 0.3635, "step": 390 }, { "epoch": 0.25152782245094885, "grad_norm": 2.707912103225004, "learning_rate": 1.753698133279553e-06, "loss": 0.3566, "step": 391 }, { "epoch": 0.25217111611450627, "grad_norm": 2.394526529094081, "learning_rate": 1.752326410348831e-06, "loss": 0.3621, "step": 392 }, { "epoch": 0.2528144097780637, "grad_norm": 2.3874609679329644, "learning_rate": 1.7509514179327911e-06, "loss": 0.3883, "step": 393 }, { "epoch": 0.2534577034416211, "grad_norm": 2.822161576499888, "learning_rate": 1.7495731620069213e-06, "loss": 0.383, "step": 394 }, { "epoch": 0.2541009971051785, "grad_norm": 2.7003184150448205, "learning_rate": 1.7481916485608928e-06, "loss": 0.3713, "step": 395 }, { "epoch": 0.25474429076873595, "grad_norm": 2.969314874822505, "learning_rate": 1.7468068835985324e-06, "loss": 0.4086, "step": 396 }, { "epoch": 0.25538758443229337, "grad_norm": 2.7087967346846415, "learning_rate": 1.7454188731377985e-06, "loss": 0.3819, "step": 397 }, { "epoch": 0.25603087809585073, "grad_norm": 2.8849364622507254, "learning_rate": 1.7440276232107535e-06, "loss": 0.4001, "step": 398 }, { "epoch": 0.25667417175940815, "grad_norm": 2.0536689357729303, "learning_rate": 1.742633139863538e-06, "loss": 0.3065, "step": 399 }, { "epoch": 0.2573174654229656, "grad_norm": 2.6216607847409112, "learning_rate": 1.741235429156344e-06, "loss": 0.3611, "step": 400 }, { "epoch": 0.257960759086523, "grad_norm": 2.3793797046645477, "learning_rate": 1.73983449716339e-06, "loss": 0.3278, "step": 401 }, { "epoch": 0.2586040527500804, "grad_norm": 2.4327721311053914, "learning_rate": 1.7384303499728928e-06, "loss": 0.393, "step": 402 }, { "epoch": 0.25924734641363784, "grad_norm": 2.6997994571968773, "learning_rate": 1.7370229936870423e-06, "loss": 0.5216, "step": 403 }, { "epoch": 0.25989064007719526, "grad_norm": 2.5526930593779737, "learning_rate": 1.7356124344219755e-06, "loss": 0.4175, "step": 404 }, { "epoch": 0.2605339337407527, "grad_norm": 2.119162451622028, "learning_rate": 1.7341986783077465e-06, "loss": 0.3449, "step": 405 }, { "epoch": 0.26117722740431004, "grad_norm": 2.360917487297911, "learning_rate": 1.7327817314883052e-06, "loss": 0.4185, "step": 406 }, { "epoch": 0.26182052106786746, "grad_norm": 3.2484858795424927, "learning_rate": 1.7313616001214663e-06, "loss": 0.3955, "step": 407 }, { "epoch": 0.2624638147314249, "grad_norm": 2.1916738657345975, "learning_rate": 1.7299382903788841e-06, "loss": 0.3166, "step": 408 }, { "epoch": 0.2631071083949823, "grad_norm": 2.914052151035379, "learning_rate": 1.7285118084460263e-06, "loss": 0.3971, "step": 409 }, { "epoch": 0.2637504020585397, "grad_norm": 2.554164182489966, "learning_rate": 1.7270821605221446e-06, "loss": 0.3955, "step": 410 }, { "epoch": 0.26439369572209714, "grad_norm": 2.346166518166145, "learning_rate": 1.7256493528202522e-06, "loss": 0.3443, "step": 411 }, { "epoch": 0.26503698938565456, "grad_norm": 2.392616817199804, "learning_rate": 1.7242133915670923e-06, "loss": 0.3804, "step": 412 }, { "epoch": 0.265680283049212, "grad_norm": 2.5661151494661905, "learning_rate": 1.722774283003113e-06, "loss": 0.362, "step": 413 }, { "epoch": 0.2663235767127694, "grad_norm": 2.6811558959547797, "learning_rate": 1.7213320333824415e-06, "loss": 0.3233, "step": 414 }, { "epoch": 0.26696687037632677, "grad_norm": 2.757593559522185, "learning_rate": 1.7198866489728534e-06, "loss": 0.3456, "step": 415 }, { "epoch": 0.2676101640398842, "grad_norm": 2.2787578275534353, "learning_rate": 1.7184381360557496e-06, "loss": 0.3264, "step": 416 }, { "epoch": 0.2682534577034416, "grad_norm": 2.495690873180919, "learning_rate": 1.7169865009261255e-06, "loss": 0.3991, "step": 417 }, { "epoch": 0.26889675136699903, "grad_norm": 2.6701720796875903, "learning_rate": 1.715531749892546e-06, "loss": 0.3844, "step": 418 }, { "epoch": 0.26954004503055645, "grad_norm": 2.372153137927286, "learning_rate": 1.7140738892771164e-06, "loss": 0.373, "step": 419 }, { "epoch": 0.2701833386941139, "grad_norm": 2.7239772974374388, "learning_rate": 1.7126129254154566e-06, "loss": 0.4851, "step": 420 }, { "epoch": 0.2708266323576713, "grad_norm": 2.256323143538246, "learning_rate": 1.7111488646566725e-06, "loss": 0.34, "step": 421 }, { "epoch": 0.2714699260212287, "grad_norm": 2.632423999585627, "learning_rate": 1.7096817133633282e-06, "loss": 0.3671, "step": 422 }, { "epoch": 0.2721132196847861, "grad_norm": 2.543308965468568, "learning_rate": 1.7082114779114185e-06, "loss": 0.4118, "step": 423 }, { "epoch": 0.2727565133483435, "grad_norm": 2.4954563015245768, "learning_rate": 1.7067381646903425e-06, "loss": 0.4023, "step": 424 }, { "epoch": 0.2733998070119009, "grad_norm": 2.8797405078723104, "learning_rate": 1.705261780102874e-06, "loss": 0.3583, "step": 425 }, { "epoch": 0.27404310067545834, "grad_norm": 2.381149758079924, "learning_rate": 1.7037823305651342e-06, "loss": 0.3833, "step": 426 }, { "epoch": 0.27468639433901576, "grad_norm": 3.1473522291923133, "learning_rate": 1.7022998225065649e-06, "loss": 0.4821, "step": 427 }, { "epoch": 0.2753296880025732, "grad_norm": 2.3651137544474414, "learning_rate": 1.7008142623698988e-06, "loss": 0.499, "step": 428 }, { "epoch": 0.2759729816661306, "grad_norm": 2.7437006133140516, "learning_rate": 1.6993256566111337e-06, "loss": 0.4079, "step": 429 }, { "epoch": 0.276616275329688, "grad_norm": 2.632969523151178, "learning_rate": 1.697834011699502e-06, "loss": 0.369, "step": 430 }, { "epoch": 0.27725956899324544, "grad_norm": 2.338010706185912, "learning_rate": 1.6963393341174436e-06, "loss": 0.4189, "step": 431 }, { "epoch": 0.2779028626568028, "grad_norm": 2.691638939304797, "learning_rate": 1.6948416303605793e-06, "loss": 0.4538, "step": 432 }, { "epoch": 0.2785461563203602, "grad_norm": 2.7075919931191033, "learning_rate": 1.6933409069376796e-06, "loss": 0.3265, "step": 433 }, { "epoch": 0.27918944998391765, "grad_norm": 2.5747700847471484, "learning_rate": 1.6918371703706392e-06, "loss": 0.4294, "step": 434 }, { "epoch": 0.27983274364747507, "grad_norm": 2.924344389358787, "learning_rate": 1.6903304271944462e-06, "loss": 0.4807, "step": 435 }, { "epoch": 0.2804760373110325, "grad_norm": 2.6601567624044584, "learning_rate": 1.688820683957156e-06, "loss": 0.3957, "step": 436 }, { "epoch": 0.2811193309745899, "grad_norm": 3.1455751737506086, "learning_rate": 1.6873079472198612e-06, "loss": 0.4937, "step": 437 }, { "epoch": 0.28176262463814733, "grad_norm": 2.763639649604737, "learning_rate": 1.685792223556664e-06, "loss": 0.3895, "step": 438 }, { "epoch": 0.28240591830170475, "grad_norm": 2.32658655035169, "learning_rate": 1.6842735195546464e-06, "loss": 0.3915, "step": 439 }, { "epoch": 0.2830492119652621, "grad_norm": 2.7492251532734193, "learning_rate": 1.682751841813844e-06, "loss": 0.4224, "step": 440 }, { "epoch": 0.28369250562881954, "grad_norm": 2.6557575631775006, "learning_rate": 1.6812271969472144e-06, "loss": 0.4054, "step": 441 }, { "epoch": 0.28433579929237696, "grad_norm": 2.756700238554445, "learning_rate": 1.6796995915806105e-06, "loss": 0.4524, "step": 442 }, { "epoch": 0.2849790929559344, "grad_norm": 2.060659030079423, "learning_rate": 1.6781690323527509e-06, "loss": 0.3116, "step": 443 }, { "epoch": 0.2856223866194918, "grad_norm": 2.420610152936278, "learning_rate": 1.676635525915191e-06, "loss": 0.4158, "step": 444 }, { "epoch": 0.2862656802830492, "grad_norm": 2.8134903839235474, "learning_rate": 1.6750990789322948e-06, "loss": 0.3734, "step": 445 }, { "epoch": 0.28690897394660664, "grad_norm": 2.49407908045489, "learning_rate": 1.6735596980812045e-06, "loss": 0.3985, "step": 446 }, { "epoch": 0.28755226761016406, "grad_norm": 2.5713465689775505, "learning_rate": 1.6720173900518134e-06, "loss": 0.3636, "step": 447 }, { "epoch": 0.2881955612737215, "grad_norm": 2.8795050128715607, "learning_rate": 1.6704721615467357e-06, "loss": 0.3192, "step": 448 }, { "epoch": 0.28883885493727884, "grad_norm": 2.8257370573604055, "learning_rate": 1.6689240192812761e-06, "loss": 0.3911, "step": 449 }, { "epoch": 0.28948214860083626, "grad_norm": 2.653113072463446, "learning_rate": 1.6673729699834044e-06, "loss": 0.4115, "step": 450 }, { "epoch": 0.2901254422643937, "grad_norm": 2.5207176189864784, "learning_rate": 1.6658190203937216e-06, "loss": 0.473, "step": 451 }, { "epoch": 0.2907687359279511, "grad_norm": 2.2860454834981723, "learning_rate": 1.6642621772654345e-06, "loss": 0.3861, "step": 452 }, { "epoch": 0.2914120295915085, "grad_norm": 2.380185779468059, "learning_rate": 1.6627024473643244e-06, "loss": 0.3211, "step": 453 }, { "epoch": 0.29205532325506595, "grad_norm": 2.4693816382507014, "learning_rate": 1.661139837468717e-06, "loss": 0.4162, "step": 454 }, { "epoch": 0.29269861691862337, "grad_norm": 2.6372720767339253, "learning_rate": 1.6595743543694547e-06, "loss": 0.3734, "step": 455 }, { "epoch": 0.2933419105821808, "grad_norm": 2.6700804611193205, "learning_rate": 1.658006004869867e-06, "loss": 0.4215, "step": 456 }, { "epoch": 0.29398520424573815, "grad_norm": 2.5601392919253403, "learning_rate": 1.6564347957857394e-06, "loss": 0.4109, "step": 457 }, { "epoch": 0.2946284979092956, "grad_norm": 2.637547006631978, "learning_rate": 1.6548607339452852e-06, "loss": 0.4099, "step": 458 }, { "epoch": 0.295271791572853, "grad_norm": 2.8281989452448357, "learning_rate": 1.6532838261891144e-06, "loss": 0.3915, "step": 459 }, { "epoch": 0.2959150852364104, "grad_norm": 2.591497537058014, "learning_rate": 1.6517040793702064e-06, "loss": 0.4188, "step": 460 }, { "epoch": 0.29655837889996783, "grad_norm": 2.358330578508166, "learning_rate": 1.6501215003538774e-06, "loss": 0.4417, "step": 461 }, { "epoch": 0.29720167256352525, "grad_norm": 2.5388107096705284, "learning_rate": 1.6485360960177524e-06, "loss": 0.432, "step": 462 }, { "epoch": 0.2978449662270827, "grad_norm": 2.5074862447228883, "learning_rate": 1.6469478732517348e-06, "loss": 0.4121, "step": 463 }, { "epoch": 0.2984882598906401, "grad_norm": 2.54561028195917, "learning_rate": 1.645356838957976e-06, "loss": 0.3344, "step": 464 }, { "epoch": 0.2991315535541975, "grad_norm": 2.5752477549918606, "learning_rate": 1.6437630000508462e-06, "loss": 0.3913, "step": 465 }, { "epoch": 0.2997748472177549, "grad_norm": 2.454284760693525, "learning_rate": 1.6421663634569044e-06, "loss": 0.3691, "step": 466 }, { "epoch": 0.3004181408813123, "grad_norm": 2.5975857067637587, "learning_rate": 1.6405669361148667e-06, "loss": 0.3736, "step": 467 }, { "epoch": 0.3010614345448697, "grad_norm": 2.1900270614602677, "learning_rate": 1.6389647249755785e-06, "loss": 0.3741, "step": 468 }, { "epoch": 0.30170472820842714, "grad_norm": 2.346741298160916, "learning_rate": 1.6373597370019822e-06, "loss": 0.3356, "step": 469 }, { "epoch": 0.30234802187198456, "grad_norm": 2.9487834488238605, "learning_rate": 1.635751979169089e-06, "loss": 0.4189, "step": 470 }, { "epoch": 0.302991315535542, "grad_norm": 2.7433406932546194, "learning_rate": 1.6341414584639464e-06, "loss": 0.3715, "step": 471 }, { "epoch": 0.3036346091990994, "grad_norm": 2.451692112751837, "learning_rate": 1.6325281818856088e-06, "loss": 0.3563, "step": 472 }, { "epoch": 0.3042779028626568, "grad_norm": 2.0810218629703856, "learning_rate": 1.6309121564451084e-06, "loss": 0.3267, "step": 473 }, { "epoch": 0.30492119652621424, "grad_norm": 2.708338816263141, "learning_rate": 1.6292933891654218e-06, "loss": 0.4386, "step": 474 }, { "epoch": 0.3055644901897716, "grad_norm": 2.516816295510506, "learning_rate": 1.6276718870814427e-06, "loss": 0.3111, "step": 475 }, { "epoch": 0.30620778385332903, "grad_norm": 3.038449110967844, "learning_rate": 1.6260476572399493e-06, "loss": 0.3687, "step": 476 }, { "epoch": 0.30685107751688645, "grad_norm": 2.318352449754961, "learning_rate": 1.6244207066995736e-06, "loss": 0.3414, "step": 477 }, { "epoch": 0.30749437118044387, "grad_norm": 3.0831582019409955, "learning_rate": 1.622791042530772e-06, "loss": 0.42, "step": 478 }, { "epoch": 0.3081376648440013, "grad_norm": 2.7725031410958705, "learning_rate": 1.6211586718157935e-06, "loss": 0.4305, "step": 479 }, { "epoch": 0.3087809585075587, "grad_norm": 3.0923220432822776, "learning_rate": 1.6195236016486499e-06, "loss": 0.4718, "step": 480 }, { "epoch": 0.30942425217111613, "grad_norm": 2.938230071519889, "learning_rate": 1.617885839135083e-06, "loss": 0.4184, "step": 481 }, { "epoch": 0.31006754583467355, "grad_norm": 2.6668436529231156, "learning_rate": 1.6162453913925372e-06, "loss": 0.506, "step": 482 }, { "epoch": 0.3107108394982309, "grad_norm": 2.6989132090068293, "learning_rate": 1.6146022655501242e-06, "loss": 0.356, "step": 483 }, { "epoch": 0.31135413316178834, "grad_norm": 2.93085261484037, "learning_rate": 1.6129564687485958e-06, "loss": 0.4306, "step": 484 }, { "epoch": 0.31199742682534576, "grad_norm": 2.6028824089211438, "learning_rate": 1.6113080081403108e-06, "loss": 0.4142, "step": 485 }, { "epoch": 0.3126407204889032, "grad_norm": 2.679837577231397, "learning_rate": 1.6096568908892045e-06, "loss": 0.2924, "step": 486 }, { "epoch": 0.3132840141524606, "grad_norm": 2.292414639717237, "learning_rate": 1.6080031241707576e-06, "loss": 0.413, "step": 487 }, { "epoch": 0.313927307816018, "grad_norm": 2.548259133277209, "learning_rate": 1.606346715171965e-06, "loss": 0.3843, "step": 488 }, { "epoch": 0.31457060147957544, "grad_norm": 2.7606547155703574, "learning_rate": 1.6046876710913041e-06, "loss": 0.439, "step": 489 }, { "epoch": 0.31521389514313286, "grad_norm": 2.7493374194426363, "learning_rate": 1.6030259991387047e-06, "loss": 0.4568, "step": 490 }, { "epoch": 0.3158571888066903, "grad_norm": 2.9839182371372006, "learning_rate": 1.6013617065355165e-06, "loss": 0.413, "step": 491 }, { "epoch": 0.31650048247024765, "grad_norm": 2.147727589405176, "learning_rate": 1.5996948005144773e-06, "loss": 0.3532, "step": 492 }, { "epoch": 0.31714377613380507, "grad_norm": 2.409079901966685, "learning_rate": 1.5980252883196838e-06, "loss": 0.3476, "step": 493 }, { "epoch": 0.3177870697973625, "grad_norm": 2.4858633879073904, "learning_rate": 1.5963531772065578e-06, "loss": 0.3844, "step": 494 }, { "epoch": 0.3184303634609199, "grad_norm": 2.723288103563476, "learning_rate": 1.594678474441816e-06, "loss": 0.47, "step": 495 }, { "epoch": 0.31907365712447733, "grad_norm": 2.7136014014908723, "learning_rate": 1.5930011873034375e-06, "loss": 0.3672, "step": 496 }, { "epoch": 0.31971695078803475, "grad_norm": 2.9054471695077404, "learning_rate": 1.591321323080633e-06, "loss": 0.3852, "step": 497 }, { "epoch": 0.32036024445159217, "grad_norm": 2.4823986340671245, "learning_rate": 1.5896388890738127e-06, "loss": 0.4521, "step": 498 }, { "epoch": 0.3210035381151496, "grad_norm": 2.4519380736075003, "learning_rate": 1.5879538925945548e-06, "loss": 0.3914, "step": 499 }, { "epoch": 0.32164683177870695, "grad_norm": 2.1047073380135197, "learning_rate": 1.5862663409655731e-06, "loss": 0.3928, "step": 500 }, { "epoch": 0.3222901254422644, "grad_norm": 2.1819538109210783, "learning_rate": 1.5845762415206868e-06, "loss": 0.3549, "step": 501 }, { "epoch": 0.3229334191058218, "grad_norm": 2.242919198303355, "learning_rate": 1.5828836016047858e-06, "loss": 0.3751, "step": 502 }, { "epoch": 0.3235767127693792, "grad_norm": 2.5809004374821907, "learning_rate": 1.5811884285738017e-06, "loss": 0.4149, "step": 503 }, { "epoch": 0.32422000643293664, "grad_norm": 2.583323128127284, "learning_rate": 1.5794907297946747e-06, "loss": 0.4014, "step": 504 }, { "epoch": 0.32486330009649406, "grad_norm": 2.378587649972078, "learning_rate": 1.5777905126453202e-06, "loss": 0.358, "step": 505 }, { "epoch": 0.3255065937600515, "grad_norm": 2.237604458313931, "learning_rate": 1.5760877845145992e-06, "loss": 0.2846, "step": 506 }, { "epoch": 0.3261498874236089, "grad_norm": 2.7659792319421457, "learning_rate": 1.5743825528022848e-06, "loss": 0.4052, "step": 507 }, { "epoch": 0.3267931810871663, "grad_norm": 2.720761905163603, "learning_rate": 1.57267482491903e-06, "loss": 0.4128, "step": 508 }, { "epoch": 0.3274364747507237, "grad_norm": 2.2090709196243887, "learning_rate": 1.570964608286336e-06, "loss": 0.3377, "step": 509 }, { "epoch": 0.3280797684142811, "grad_norm": 2.6558540943392925, "learning_rate": 1.5692519103365192e-06, "loss": 0.4164, "step": 510 }, { "epoch": 0.3287230620778385, "grad_norm": 2.455592695121354, "learning_rate": 1.56753673851268e-06, "loss": 0.4946, "step": 511 }, { "epoch": 0.32936635574139594, "grad_norm": 2.6697923369473147, "learning_rate": 1.5658191002686695e-06, "loss": 0.393, "step": 512 }, { "epoch": 0.33000964940495336, "grad_norm": 2.3333799535614257, "learning_rate": 1.5640990030690572e-06, "loss": 0.3682, "step": 513 }, { "epoch": 0.3306529430685108, "grad_norm": 2.806414146805125, "learning_rate": 1.5623764543890998e-06, "loss": 0.4408, "step": 514 }, { "epoch": 0.3312962367320682, "grad_norm": 2.8449833851029522, "learning_rate": 1.5606514617147066e-06, "loss": 0.395, "step": 515 }, { "epoch": 0.3319395303956256, "grad_norm": 2.6769573720121445, "learning_rate": 1.5589240325424087e-06, "loss": 0.4213, "step": 516 }, { "epoch": 0.332582824059183, "grad_norm": 2.6379954671806183, "learning_rate": 1.5571941743793254e-06, "loss": 0.433, "step": 517 }, { "epoch": 0.3332261177227404, "grad_norm": 2.625688205802763, "learning_rate": 1.5554618947431325e-06, "loss": 0.4372, "step": 518 }, { "epoch": 0.33386941138629783, "grad_norm": 2.961648731283995, "learning_rate": 1.553727201162029e-06, "loss": 0.3827, "step": 519 }, { "epoch": 0.33451270504985525, "grad_norm": 3.1295281389757825, "learning_rate": 1.5519901011747043e-06, "loss": 0.3703, "step": 520 }, { "epoch": 0.3351559987134127, "grad_norm": 2.869729450540354, "learning_rate": 1.5502506023303058e-06, "loss": 0.4406, "step": 521 }, { "epoch": 0.3357992923769701, "grad_norm": 2.3380926358852325, "learning_rate": 1.5485087121884052e-06, "loss": 0.4566, "step": 522 }, { "epoch": 0.3364425860405275, "grad_norm": 2.467083426297788, "learning_rate": 1.546764438318968e-06, "loss": 0.3894, "step": 523 }, { "epoch": 0.33708587970408493, "grad_norm": 2.135835685337476, "learning_rate": 1.5450177883023173e-06, "loss": 0.316, "step": 524 }, { "epoch": 0.33772917336764235, "grad_norm": 2.7005312785435964, "learning_rate": 1.5432687697291035e-06, "loss": 0.3538, "step": 525 }, { "epoch": 0.3383724670311997, "grad_norm": 2.8981780886504125, "learning_rate": 1.54151739020027e-06, "loss": 0.3199, "step": 526 }, { "epoch": 0.33901576069475714, "grad_norm": 2.845834270773485, "learning_rate": 1.539763657327021e-06, "loss": 0.4056, "step": 527 }, { "epoch": 0.33965905435831456, "grad_norm": 2.5880183382974207, "learning_rate": 1.5380075787307873e-06, "loss": 0.3633, "step": 528 }, { "epoch": 0.340302348021872, "grad_norm": 2.5232472320740893, "learning_rate": 1.536249162043194e-06, "loss": 0.3043, "step": 529 }, { "epoch": 0.3409456416854294, "grad_norm": 2.7191380585485376, "learning_rate": 1.5344884149060277e-06, "loss": 0.3382, "step": 530 }, { "epoch": 0.3415889353489868, "grad_norm": 2.7348138861811413, "learning_rate": 1.5327253449712018e-06, "loss": 0.4832, "step": 531 }, { "epoch": 0.34223222901254424, "grad_norm": 2.152647339138582, "learning_rate": 1.5309599599007248e-06, "loss": 0.4037, "step": 532 }, { "epoch": 0.34287552267610166, "grad_norm": 2.445540992963504, "learning_rate": 1.529192267366666e-06, "loss": 0.4022, "step": 533 }, { "epoch": 0.343518816339659, "grad_norm": 2.929447199761597, "learning_rate": 1.5274222750511237e-06, "loss": 0.445, "step": 534 }, { "epoch": 0.34416211000321645, "grad_norm": 2.444714797025497, "learning_rate": 1.525649990646189e-06, "loss": 0.4156, "step": 535 }, { "epoch": 0.34480540366677387, "grad_norm": 2.6021335593826267, "learning_rate": 1.5238754218539153e-06, "loss": 0.3686, "step": 536 }, { "epoch": 0.3454486973303313, "grad_norm": 2.5628134923441777, "learning_rate": 1.5220985763862827e-06, "loss": 0.3736, "step": 537 }, { "epoch": 0.3460919909938887, "grad_norm": 2.643185309959748, "learning_rate": 1.5203194619651661e-06, "loss": 0.4642, "step": 538 }, { "epoch": 0.34673528465744613, "grad_norm": 2.537459313941842, "learning_rate": 1.5185380863223005e-06, "loss": 0.4158, "step": 539 }, { "epoch": 0.34737857832100355, "grad_norm": 2.4520569761292577, "learning_rate": 1.5167544571992477e-06, "loss": 0.3483, "step": 540 }, { "epoch": 0.34802187198456097, "grad_norm": 2.3251172829706688, "learning_rate": 1.514968582347363e-06, "loss": 0.2714, "step": 541 }, { "epoch": 0.3486651656481184, "grad_norm": 2.2327181261200346, "learning_rate": 1.513180469527761e-06, "loss": 0.4192, "step": 542 }, { "epoch": 0.34930845931167576, "grad_norm": 2.5173887866424285, "learning_rate": 1.5113901265112827e-06, "loss": 0.4172, "step": 543 }, { "epoch": 0.3499517529752332, "grad_norm": 2.3417594620643984, "learning_rate": 1.50959756107846e-06, "loss": 0.3769, "step": 544 }, { "epoch": 0.3505950466387906, "grad_norm": 3.2214671208270134, "learning_rate": 1.5078027810194845e-06, "loss": 0.4535, "step": 545 }, { "epoch": 0.351238340302348, "grad_norm": 2.974545175316378, "learning_rate": 1.5060057941341714e-06, "loss": 0.4327, "step": 546 }, { "epoch": 0.35188163396590544, "grad_norm": 2.9524009061838794, "learning_rate": 1.5042066082319265e-06, "loss": 0.4221, "step": 547 }, { "epoch": 0.35252492762946286, "grad_norm": 2.910925940872193, "learning_rate": 1.502405231131712e-06, "loss": 0.4462, "step": 548 }, { "epoch": 0.3531682212930203, "grad_norm": 2.612598748505983, "learning_rate": 1.500601670662013e-06, "loss": 0.4307, "step": 549 }, { "epoch": 0.3538115149565777, "grad_norm": 2.2795867753868633, "learning_rate": 1.4987959346608028e-06, "loss": 0.317, "step": 550 }, { "epoch": 0.35445480862013506, "grad_norm": 2.500561876455955, "learning_rate": 1.4969880309755096e-06, "loss": 0.2984, "step": 551 }, { "epoch": 0.3550981022836925, "grad_norm": 2.4887610032930203, "learning_rate": 1.4951779674629818e-06, "loss": 0.428, "step": 552 }, { "epoch": 0.3557413959472499, "grad_norm": 2.5647718026296054, "learning_rate": 1.493365751989454e-06, "loss": 0.3614, "step": 553 }, { "epoch": 0.3563846896108073, "grad_norm": 2.23845895441745, "learning_rate": 1.491551392430513e-06, "loss": 0.3297, "step": 554 }, { "epoch": 0.35702798327436475, "grad_norm": 2.258919862474323, "learning_rate": 1.489734896671063e-06, "loss": 0.3826, "step": 555 }, { "epoch": 0.35767127693792217, "grad_norm": 2.529666676028519, "learning_rate": 1.4879162726052924e-06, "loss": 0.3505, "step": 556 }, { "epoch": 0.3583145706014796, "grad_norm": 2.675007032481313, "learning_rate": 1.486095528136639e-06, "loss": 0.4137, "step": 557 }, { "epoch": 0.358957864265037, "grad_norm": 2.7062699038122418, "learning_rate": 1.4842726711777545e-06, "loss": 0.3767, "step": 558 }, { "epoch": 0.35960115792859443, "grad_norm": 2.893846327849919, "learning_rate": 1.482447709650472e-06, "loss": 0.3827, "step": 559 }, { "epoch": 0.3602444515921518, "grad_norm": 2.589177204973386, "learning_rate": 1.4806206514857701e-06, "loss": 0.3826, "step": 560 }, { "epoch": 0.3608877452557092, "grad_norm": 2.674900887088488, "learning_rate": 1.47879150462374e-06, "loss": 0.3608, "step": 561 }, { "epoch": 0.36153103891926663, "grad_norm": 2.5800313648196096, "learning_rate": 1.476960277013549e-06, "loss": 0.3991, "step": 562 }, { "epoch": 0.36217433258282405, "grad_norm": 2.797534176713798, "learning_rate": 1.4751269766134074e-06, "loss": 0.4298, "step": 563 }, { "epoch": 0.3628176262463815, "grad_norm": 2.687738479675439, "learning_rate": 1.4732916113905333e-06, "loss": 0.397, "step": 564 }, { "epoch": 0.3634609199099389, "grad_norm": 2.5340936231096864, "learning_rate": 1.4714541893211193e-06, "loss": 0.3347, "step": 565 }, { "epoch": 0.3641042135734963, "grad_norm": 2.707257757361265, "learning_rate": 1.4696147183902947e-06, "loss": 0.4223, "step": 566 }, { "epoch": 0.36474750723705374, "grad_norm": 2.7298793742424228, "learning_rate": 1.4677732065920942e-06, "loss": 0.3334, "step": 567 }, { "epoch": 0.3653908009006111, "grad_norm": 2.830004392947017, "learning_rate": 1.4659296619294219e-06, "loss": 0.4124, "step": 568 }, { "epoch": 0.3660340945641685, "grad_norm": 2.593128527508919, "learning_rate": 1.4640840924140155e-06, "loss": 0.3359, "step": 569 }, { "epoch": 0.36667738822772594, "grad_norm": 2.370018676538994, "learning_rate": 1.4622365060664131e-06, "loss": 0.3492, "step": 570 }, { "epoch": 0.36732068189128336, "grad_norm": 2.3387311582716244, "learning_rate": 1.460386910915917e-06, "loss": 0.3583, "step": 571 }, { "epoch": 0.3679639755548408, "grad_norm": 2.801706118291109, "learning_rate": 1.4585353150005599e-06, "loss": 0.3986, "step": 572 }, { "epoch": 0.3686072692183982, "grad_norm": 2.4059505041133598, "learning_rate": 1.4566817263670697e-06, "loss": 0.3627, "step": 573 }, { "epoch": 0.3692505628819556, "grad_norm": 2.3512022871506795, "learning_rate": 1.4548261530708335e-06, "loss": 0.4101, "step": 574 }, { "epoch": 0.36989385654551304, "grad_norm": 2.3260549832727038, "learning_rate": 1.4529686031758642e-06, "loss": 0.4369, "step": 575 }, { "epoch": 0.37053715020907046, "grad_norm": 2.5901020057029505, "learning_rate": 1.4511090847547643e-06, "loss": 0.3277, "step": 576 }, { "epoch": 0.37118044387262783, "grad_norm": 2.7317323860639795, "learning_rate": 1.4492476058886913e-06, "loss": 0.3778, "step": 577 }, { "epoch": 0.37182373753618525, "grad_norm": 2.48104528650747, "learning_rate": 1.4473841746673223e-06, "loss": 0.3803, "step": 578 }, { "epoch": 0.37246703119974267, "grad_norm": 2.8319383718404105, "learning_rate": 1.4455187991888194e-06, "loss": 0.4226, "step": 579 }, { "epoch": 0.3731103248633001, "grad_norm": 2.049691714573059, "learning_rate": 1.4436514875597934e-06, "loss": 0.3385, "step": 580 }, { "epoch": 0.3737536185268575, "grad_norm": 2.4036424313233478, "learning_rate": 1.44178224789527e-06, "loss": 0.3477, "step": 581 }, { "epoch": 0.37439691219041493, "grad_norm": 2.446931154353517, "learning_rate": 1.4399110883186535e-06, "loss": 0.3863, "step": 582 }, { "epoch": 0.37504020585397235, "grad_norm": 2.4753907259892434, "learning_rate": 1.4380380169616919e-06, "loss": 0.3568, "step": 583 }, { "epoch": 0.3756834995175298, "grad_norm": 2.962275625400323, "learning_rate": 1.4361630419644414e-06, "loss": 0.3876, "step": 584 }, { "epoch": 0.3763267931810872, "grad_norm": 3.126503156467689, "learning_rate": 1.4342861714752313e-06, "loss": 0.4409, "step": 585 }, { "epoch": 0.37697008684464456, "grad_norm": 2.5160251435520213, "learning_rate": 1.4324074136506282e-06, "loss": 0.3723, "step": 586 }, { "epoch": 0.377613380508202, "grad_norm": 2.565801048216741, "learning_rate": 1.4305267766554005e-06, "loss": 0.3858, "step": 587 }, { "epoch": 0.3782566741717594, "grad_norm": 2.686163798990735, "learning_rate": 1.428644268662484e-06, "loss": 0.2861, "step": 588 }, { "epoch": 0.3788999678353168, "grad_norm": 2.689110819600383, "learning_rate": 1.4267598978529445e-06, "loss": 0.3784, "step": 589 }, { "epoch": 0.37954326149887424, "grad_norm": 2.7433678401948964, "learning_rate": 1.4248736724159441e-06, "loss": 0.4585, "step": 590 }, { "epoch": 0.38018655516243166, "grad_norm": 2.1846185842414614, "learning_rate": 1.4229856005487042e-06, "loss": 0.3253, "step": 591 }, { "epoch": 0.3808298488259891, "grad_norm": 2.4067792295833947, "learning_rate": 1.4210956904564708e-06, "loss": 0.3969, "step": 592 }, { "epoch": 0.3814731424895465, "grad_norm": 2.5866674138333305, "learning_rate": 1.419203950352479e-06, "loss": 0.4193, "step": 593 }, { "epoch": 0.38211643615310387, "grad_norm": 2.733767455713775, "learning_rate": 1.4173103884579161e-06, "loss": 0.4487, "step": 594 }, { "epoch": 0.3827597298166613, "grad_norm": 2.439155903620872, "learning_rate": 1.4154150130018865e-06, "loss": 0.3714, "step": 595 }, { "epoch": 0.3834030234802187, "grad_norm": 2.4358888793151294, "learning_rate": 1.4135178322213763e-06, "loss": 0.393, "step": 596 }, { "epoch": 0.38404631714377613, "grad_norm": 2.5548071363048397, "learning_rate": 1.411618854361218e-06, "loss": 0.4463, "step": 597 }, { "epoch": 0.38468961080733355, "grad_norm": 3.0165672511841133, "learning_rate": 1.4097180876740524e-06, "loss": 0.5074, "step": 598 }, { "epoch": 0.38533290447089097, "grad_norm": 2.2861574154764868, "learning_rate": 1.4078155404202956e-06, "loss": 0.396, "step": 599 }, { "epoch": 0.3859761981344484, "grad_norm": 2.2131043134526873, "learning_rate": 1.405911220868101e-06, "loss": 0.37, "step": 600 }, { "epoch": 0.3866194917980058, "grad_norm": 2.5190910170999503, "learning_rate": 1.4040051372933243e-06, "loss": 0.3312, "step": 601 }, { "epoch": 0.38726278546156323, "grad_norm": 2.942627932810319, "learning_rate": 1.402097297979487e-06, "loss": 0.4361, "step": 602 }, { "epoch": 0.3879060791251206, "grad_norm": 2.6781356737693884, "learning_rate": 1.4001877112177416e-06, "loss": 0.4126, "step": 603 }, { "epoch": 0.388549372788678, "grad_norm": 2.24867322478735, "learning_rate": 1.3982763853068343e-06, "loss": 0.3354, "step": 604 }, { "epoch": 0.38919266645223544, "grad_norm": 2.812037090594078, "learning_rate": 1.3963633285530687e-06, "loss": 0.388, "step": 605 }, { "epoch": 0.38983596011579286, "grad_norm": 2.578692938805659, "learning_rate": 1.3944485492702715e-06, "loss": 0.4385, "step": 606 }, { "epoch": 0.3904792537793503, "grad_norm": 2.662008519613447, "learning_rate": 1.3925320557797545e-06, "loss": 0.3862, "step": 607 }, { "epoch": 0.3911225474429077, "grad_norm": 2.4772453316811665, "learning_rate": 1.3906138564102792e-06, "loss": 0.3497, "step": 608 }, { "epoch": 0.3917658411064651, "grad_norm": 1.939270940151947, "learning_rate": 1.3886939594980208e-06, "loss": 0.334, "step": 609 }, { "epoch": 0.39240913477002254, "grad_norm": 2.744084439097213, "learning_rate": 1.3867723733865313e-06, "loss": 0.3029, "step": 610 }, { "epoch": 0.3930524284335799, "grad_norm": 2.394264695791567, "learning_rate": 1.3848491064267044e-06, "loss": 0.3764, "step": 611 }, { "epoch": 0.3936957220971373, "grad_norm": 2.689898757560601, "learning_rate": 1.382924166976738e-06, "loss": 0.4112, "step": 612 }, { "epoch": 0.39433901576069474, "grad_norm": 2.350082710935213, "learning_rate": 1.3809975634020983e-06, "loss": 0.3704, "step": 613 }, { "epoch": 0.39498230942425216, "grad_norm": 2.2773722628250406, "learning_rate": 1.3790693040754836e-06, "loss": 0.2787, "step": 614 }, { "epoch": 0.3956256030878096, "grad_norm": 2.1481732320438582, "learning_rate": 1.3771393973767884e-06, "loss": 0.3274, "step": 615 }, { "epoch": 0.396268896751367, "grad_norm": 2.4916951936529204, "learning_rate": 1.3752078516930652e-06, "loss": 0.3869, "step": 616 }, { "epoch": 0.3969121904149244, "grad_norm": 2.987980245824785, "learning_rate": 1.3732746754184904e-06, "loss": 0.3342, "step": 617 }, { "epoch": 0.39755548407848185, "grad_norm": 2.359931050860899, "learning_rate": 1.3713398769543259e-06, "loss": 0.3513, "step": 618 }, { "epoch": 0.39819877774203927, "grad_norm": 2.6497059478698763, "learning_rate": 1.3694034647088838e-06, "loss": 0.3838, "step": 619 }, { "epoch": 0.39884207140559663, "grad_norm": 2.372671184769959, "learning_rate": 1.3674654470974896e-06, "loss": 0.38, "step": 620 }, { "epoch": 0.39948536506915405, "grad_norm": 2.5056822270917025, "learning_rate": 1.3655258325424446e-06, "loss": 0.3765, "step": 621 }, { "epoch": 0.4001286587327115, "grad_norm": 2.3489206096999924, "learning_rate": 1.3635846294729915e-06, "loss": 0.3957, "step": 622 }, { "epoch": 0.4007719523962689, "grad_norm": 2.809605160938011, "learning_rate": 1.3616418463252746e-06, "loss": 0.4597, "step": 623 }, { "epoch": 0.4014152460598263, "grad_norm": 3.2297832570812446, "learning_rate": 1.359697491542307e-06, "loss": 0.3519, "step": 624 }, { "epoch": 0.40205853972338373, "grad_norm": 2.527833611394939, "learning_rate": 1.35775157357393e-06, "loss": 0.4123, "step": 625 }, { "epoch": 0.40270183338694115, "grad_norm": 2.6132574108529876, "learning_rate": 1.3558041008767797e-06, "loss": 0.3607, "step": 626 }, { "epoch": 0.4033451270504986, "grad_norm": 2.846425327247037, "learning_rate": 1.3538550819142483e-06, "loss": 0.3917, "step": 627 }, { "epoch": 0.40398842071405594, "grad_norm": 2.6247715066429254, "learning_rate": 1.351904525156447e-06, "loss": 0.3852, "step": 628 }, { "epoch": 0.40463171437761336, "grad_norm": 2.679724041739953, "learning_rate": 1.3499524390801714e-06, "loss": 0.3311, "step": 629 }, { "epoch": 0.4052750080411708, "grad_norm": 2.2717376629439108, "learning_rate": 1.3479988321688618e-06, "loss": 0.3503, "step": 630 }, { "epoch": 0.4059183017047282, "grad_norm": 2.389674780016254, "learning_rate": 1.3460437129125692e-06, "loss": 0.4374, "step": 631 }, { "epoch": 0.4065615953682856, "grad_norm": 2.4371735017507365, "learning_rate": 1.3440870898079164e-06, "loss": 0.4441, "step": 632 }, { "epoch": 0.40720488903184304, "grad_norm": 2.5635115177158867, "learning_rate": 1.3421289713580616e-06, "loss": 0.4151, "step": 633 }, { "epoch": 0.40784818269540046, "grad_norm": 2.407126422938099, "learning_rate": 1.340169366072661e-06, "loss": 0.3525, "step": 634 }, { "epoch": 0.4084914763589579, "grad_norm": 2.9559867459054647, "learning_rate": 1.3382082824678334e-06, "loss": 0.4273, "step": 635 }, { "epoch": 0.4091347700225153, "grad_norm": 2.483401776128031, "learning_rate": 1.3362457290661213e-06, "loss": 0.3823, "step": 636 }, { "epoch": 0.40977806368607267, "grad_norm": 2.5317320977387374, "learning_rate": 1.3342817143964555e-06, "loss": 0.3507, "step": 637 }, { "epoch": 0.4104213573496301, "grad_norm": 2.664878529716042, "learning_rate": 1.3323162469941162e-06, "loss": 0.4458, "step": 638 }, { "epoch": 0.4110646510131875, "grad_norm": 2.2056675013227207, "learning_rate": 1.3303493354006982e-06, "loss": 0.328, "step": 639 }, { "epoch": 0.41170794467674493, "grad_norm": 2.607051579155456, "learning_rate": 1.3283809881640708e-06, "loss": 0.4446, "step": 640 }, { "epoch": 0.41235123834030235, "grad_norm": 2.607684567245694, "learning_rate": 1.3264112138383443e-06, "loss": 0.3833, "step": 641 }, { "epoch": 0.41299453200385977, "grad_norm": 2.4411314165968876, "learning_rate": 1.3244400209838295e-06, "loss": 0.3716, "step": 642 }, { "epoch": 0.4136378256674172, "grad_norm": 3.328208192977269, "learning_rate": 1.3224674181670023e-06, "loss": 0.4083, "step": 643 }, { "epoch": 0.4142811193309746, "grad_norm": 2.7051845162339903, "learning_rate": 1.3204934139604666e-06, "loss": 0.3691, "step": 644 }, { "epoch": 0.414924412994532, "grad_norm": 3.0126758552131125, "learning_rate": 1.3185180169429153e-06, "loss": 0.3744, "step": 645 }, { "epoch": 0.4155677066580894, "grad_norm": 2.7293697868659526, "learning_rate": 1.3165412356990952e-06, "loss": 0.4013, "step": 646 }, { "epoch": 0.4162110003216468, "grad_norm": 2.802370493444941, "learning_rate": 1.3145630788197688e-06, "loss": 0.4306, "step": 647 }, { "epoch": 0.41685429398520424, "grad_norm": 2.2970140119878892, "learning_rate": 1.312583554901676e-06, "loss": 0.3548, "step": 648 }, { "epoch": 0.41749758764876166, "grad_norm": 2.4389580977498917, "learning_rate": 1.3106026725474985e-06, "loss": 0.3806, "step": 649 }, { "epoch": 0.4181408813123191, "grad_norm": 2.34341294378266, "learning_rate": 1.3086204403658213e-06, "loss": 0.4171, "step": 650 }, { "epoch": 0.4187841749758765, "grad_norm": 2.5112750325348423, "learning_rate": 1.3066368669710951e-06, "loss": 0.3827, "step": 651 }, { "epoch": 0.4194274686394339, "grad_norm": 2.2752357988805048, "learning_rate": 1.3046519609836e-06, "loss": 0.311, "step": 652 }, { "epoch": 0.42007076230299134, "grad_norm": 2.216893731370644, "learning_rate": 1.3026657310294064e-06, "loss": 0.3473, "step": 653 }, { "epoch": 0.4207140559665487, "grad_norm": 2.697787096169345, "learning_rate": 1.3006781857403393e-06, "loss": 0.3241, "step": 654 }, { "epoch": 0.4213573496301061, "grad_norm": 2.701099644119047, "learning_rate": 1.2986893337539395e-06, "loss": 0.4009, "step": 655 }, { "epoch": 0.42200064329366355, "grad_norm": 3.7304879572065013, "learning_rate": 1.296699183713427e-06, "loss": 0.4678, "step": 656 }, { "epoch": 0.42264393695722097, "grad_norm": 3.1488929715259144, "learning_rate": 1.294707744267662e-06, "loss": 0.3838, "step": 657 }, { "epoch": 0.4232872306207784, "grad_norm": 2.4128161652477784, "learning_rate": 1.2927150240711086e-06, "loss": 0.3234, "step": 658 }, { "epoch": 0.4239305242843358, "grad_norm": 2.9984911467976274, "learning_rate": 1.2907210317837978e-06, "loss": 0.3446, "step": 659 }, { "epoch": 0.42457381794789323, "grad_norm": 2.7458510986344034, "learning_rate": 1.2887257760712873e-06, "loss": 0.4134, "step": 660 }, { "epoch": 0.42521711161145065, "grad_norm": 2.5441187038246804, "learning_rate": 1.2867292656046266e-06, "loss": 0.4343, "step": 661 }, { "epoch": 0.425860405275008, "grad_norm": 2.2575435402311235, "learning_rate": 1.2847315090603179e-06, "loss": 0.3994, "step": 662 }, { "epoch": 0.42650369893856543, "grad_norm": 2.963221055654897, "learning_rate": 1.2827325151202782e-06, "loss": 0.4057, "step": 663 }, { "epoch": 0.42714699260212285, "grad_norm": 2.7424181461234207, "learning_rate": 1.2807322924718023e-06, "loss": 0.4437, "step": 664 }, { "epoch": 0.4277902862656803, "grad_norm": 2.528447353379403, "learning_rate": 1.2787308498075251e-06, "loss": 0.471, "step": 665 }, { "epoch": 0.4284335799292377, "grad_norm": 2.5998916717476046, "learning_rate": 1.2767281958253829e-06, "loss": 0.3487, "step": 666 }, { "epoch": 0.4290768735927951, "grad_norm": 2.7611931171844493, "learning_rate": 1.2747243392285766e-06, "loss": 0.4644, "step": 667 }, { "epoch": 0.42972016725635254, "grad_norm": 2.951727968114276, "learning_rate": 1.272719288725533e-06, "loss": 0.3881, "step": 668 }, { "epoch": 0.43036346091990996, "grad_norm": 2.7788723435952587, "learning_rate": 1.2707130530298687e-06, "loss": 0.3639, "step": 669 }, { "epoch": 0.4310067545834674, "grad_norm": 2.6263555085150294, "learning_rate": 1.268705640860349e-06, "loss": 0.3869, "step": 670 }, { "epoch": 0.43165004824702474, "grad_norm": 2.243740333530608, "learning_rate": 1.2666970609408533e-06, "loss": 0.3351, "step": 671 }, { "epoch": 0.43229334191058216, "grad_norm": 2.7199643049036117, "learning_rate": 1.2646873220003357e-06, "loss": 0.4245, "step": 672 }, { "epoch": 0.4329366355741396, "grad_norm": 2.519302924884347, "learning_rate": 1.262676432772787e-06, "loss": 0.397, "step": 673 }, { "epoch": 0.433579929237697, "grad_norm": 2.770649376490532, "learning_rate": 1.2606644019971966e-06, "loss": 0.4881, "step": 674 }, { "epoch": 0.4342232229012544, "grad_norm": 2.7824673633590984, "learning_rate": 1.2586512384175155e-06, "loss": 0.3999, "step": 675 }, { "epoch": 0.43486651656481184, "grad_norm": 2.637769710154083, "learning_rate": 1.2566369507826173e-06, "loss": 0.3564, "step": 676 }, { "epoch": 0.43550981022836927, "grad_norm": 2.342179876524597, "learning_rate": 1.254621547846261e-06, "loss": 0.332, "step": 677 }, { "epoch": 0.4361531038919267, "grad_norm": 2.5630667539577883, "learning_rate": 1.2526050383670515e-06, "loss": 0.4479, "step": 678 }, { "epoch": 0.4367963975554841, "grad_norm": 2.8135634472272857, "learning_rate": 1.2505874311084039e-06, "loss": 0.4173, "step": 679 }, { "epoch": 0.43743969121904147, "grad_norm": 2.3253440609310876, "learning_rate": 1.2485687348385036e-06, "loss": 0.3612, "step": 680 }, { "epoch": 0.4380829848825989, "grad_norm": 2.614618928445224, "learning_rate": 1.2465489583302679e-06, "loss": 0.4103, "step": 681 }, { "epoch": 0.4387262785461563, "grad_norm": 2.7945897263533754, "learning_rate": 1.2445281103613095e-06, "loss": 0.426, "step": 682 }, { "epoch": 0.43936957220971373, "grad_norm": 2.9646364554802602, "learning_rate": 1.242506199713897e-06, "loss": 0.3954, "step": 683 }, { "epoch": 0.44001286587327115, "grad_norm": 2.394167790191542, "learning_rate": 1.2404832351749177e-06, "loss": 0.3839, "step": 684 }, { "epoch": 0.4406561595368286, "grad_norm": 2.4778767638496864, "learning_rate": 1.2384592255358384e-06, "loss": 0.3855, "step": 685 }, { "epoch": 0.441299453200386, "grad_norm": 2.570653027249716, "learning_rate": 1.2364341795926682e-06, "loss": 0.4038, "step": 686 }, { "epoch": 0.4419427468639434, "grad_norm": 1.9773640400814734, "learning_rate": 1.2344081061459194e-06, "loss": 0.3347, "step": 687 }, { "epoch": 0.4425860405275008, "grad_norm": 3.2370843955629915, "learning_rate": 1.2323810140005698e-06, "loss": 0.4241, "step": 688 }, { "epoch": 0.4432293341910582, "grad_norm": 2.381996681781819, "learning_rate": 1.230352911966025e-06, "loss": 0.3733, "step": 689 }, { "epoch": 0.4438726278546156, "grad_norm": 2.407865654405103, "learning_rate": 1.2283238088560781e-06, "loss": 0.3586, "step": 690 }, { "epoch": 0.44451592151817304, "grad_norm": 2.62089018627661, "learning_rate": 1.2262937134888737e-06, "loss": 0.3284, "step": 691 }, { "epoch": 0.44515921518173046, "grad_norm": 2.5275345277126764, "learning_rate": 1.2242626346868682e-06, "loss": 0.3902, "step": 692 }, { "epoch": 0.4458025088452879, "grad_norm": 2.37168161997476, "learning_rate": 1.2222305812767917e-06, "loss": 0.3521, "step": 693 }, { "epoch": 0.4464458025088453, "grad_norm": 2.8473080362072847, "learning_rate": 1.2201975620896102e-06, "loss": 0.3701, "step": 694 }, { "epoch": 0.4470890961724027, "grad_norm": 2.3760559421406238, "learning_rate": 1.2181635859604862e-06, "loss": 0.3735, "step": 695 }, { "epoch": 0.44773238983596014, "grad_norm": 2.5174487174002866, "learning_rate": 1.2161286617287418e-06, "loss": 0.4077, "step": 696 }, { "epoch": 0.4483756834995175, "grad_norm": 2.6100141219767456, "learning_rate": 1.2140927982378186e-06, "loss": 0.3811, "step": 697 }, { "epoch": 0.44901897716307493, "grad_norm": 2.691427201452633, "learning_rate": 1.212056004335241e-06, "loss": 0.331, "step": 698 }, { "epoch": 0.44966227082663235, "grad_norm": 2.8029144510439576, "learning_rate": 1.210018288872575e-06, "loss": 0.3624, "step": 699 }, { "epoch": 0.45030556449018977, "grad_norm": 2.781326572928353, "learning_rate": 1.2079796607053938e-06, "loss": 0.3763, "step": 700 }, { "epoch": 0.4509488581537472, "grad_norm": 2.5549460443173877, "learning_rate": 1.2059401286932353e-06, "loss": 0.325, "step": 701 }, { "epoch": 0.4515921518173046, "grad_norm": 2.373325548622513, "learning_rate": 1.2038997016995667e-06, "loss": 0.3628, "step": 702 }, { "epoch": 0.45223544548086203, "grad_norm": 2.7456096357124293, "learning_rate": 1.2018583885917433e-06, "loss": 0.3803, "step": 703 }, { "epoch": 0.45287873914441945, "grad_norm": 2.8777051586488187, "learning_rate": 1.1998161982409722e-06, "loss": 0.4223, "step": 704 }, { "epoch": 0.4535220328079768, "grad_norm": 2.568271507527199, "learning_rate": 1.197773139522273e-06, "loss": 0.357, "step": 705 }, { "epoch": 0.45416532647153424, "grad_norm": 2.623979707226975, "learning_rate": 1.1957292213144384e-06, "loss": 0.4761, "step": 706 }, { "epoch": 0.45480862013509166, "grad_norm": 2.553148102970076, "learning_rate": 1.1936844524999966e-06, "loss": 0.399, "step": 707 }, { "epoch": 0.4554519137986491, "grad_norm": 2.4375109214224877, "learning_rate": 1.1916388419651724e-06, "loss": 0.3629, "step": 708 }, { "epoch": 0.4560952074622065, "grad_norm": 2.378437867034685, "learning_rate": 1.1895923985998488e-06, "loss": 0.3629, "step": 709 }, { "epoch": 0.4567385011257639, "grad_norm": 2.6706156270882784, "learning_rate": 1.1875451312975277e-06, "loss": 0.3852, "step": 710 }, { "epoch": 0.45738179478932134, "grad_norm": 2.1993300936340154, "learning_rate": 1.1854970489552922e-06, "loss": 0.382, "step": 711 }, { "epoch": 0.45802508845287876, "grad_norm": 2.3529184832560017, "learning_rate": 1.1834481604737669e-06, "loss": 0.3752, "step": 712 }, { "epoch": 0.4586683821164362, "grad_norm": 2.6174062136168077, "learning_rate": 1.18139847475708e-06, "loss": 0.3672, "step": 713 }, { "epoch": 0.45931167577999354, "grad_norm": 2.2352099062211166, "learning_rate": 1.1793480007128247e-06, "loss": 0.3226, "step": 714 }, { "epoch": 0.45995496944355097, "grad_norm": 2.8202202273178054, "learning_rate": 1.1772967472520195e-06, "loss": 0.4107, "step": 715 }, { "epoch": 0.4605982631071084, "grad_norm": 2.445617501616726, "learning_rate": 1.17524472328907e-06, "loss": 0.3677, "step": 716 }, { "epoch": 0.4612415567706658, "grad_norm": 2.280918842337894, "learning_rate": 1.1731919377417316e-06, "loss": 0.3006, "step": 717 }, { "epoch": 0.4618848504342232, "grad_norm": 2.684109427545211, "learning_rate": 1.171138399531068e-06, "loss": 0.4077, "step": 718 }, { "epoch": 0.46252814409778065, "grad_norm": 2.521073676216543, "learning_rate": 1.1690841175814143e-06, "loss": 0.2649, "step": 719 }, { "epoch": 0.46317143776133807, "grad_norm": 2.0076739027449775, "learning_rate": 1.1670291008203382e-06, "loss": 0.314, "step": 720 }, { "epoch": 0.4638147314248955, "grad_norm": 2.2744806278969687, "learning_rate": 1.1649733581785998e-06, "loss": 0.4018, "step": 721 }, { "epoch": 0.46445802508845285, "grad_norm": 3.0622360434539733, "learning_rate": 1.1629168985901147e-06, "loss": 0.4032, "step": 722 }, { "epoch": 0.4651013187520103, "grad_norm": 2.445286689021327, "learning_rate": 1.160859730991914e-06, "loss": 0.4587, "step": 723 }, { "epoch": 0.4657446124155677, "grad_norm": 2.1997086311633294, "learning_rate": 1.158801864324105e-06, "loss": 0.3457, "step": 724 }, { "epoch": 0.4663879060791251, "grad_norm": 2.750200012788963, "learning_rate": 1.156743307529834e-06, "loss": 0.4649, "step": 725 }, { "epoch": 0.46703119974268253, "grad_norm": 2.793506688368552, "learning_rate": 1.1546840695552465e-06, "loss": 0.4206, "step": 726 }, { "epoch": 0.46767449340623996, "grad_norm": 2.4501434127654007, "learning_rate": 1.1526241593494473e-06, "loss": 0.4031, "step": 727 }, { "epoch": 0.4683177870697974, "grad_norm": 3.1561859231814395, "learning_rate": 1.1505635858644634e-06, "loss": 0.3382, "step": 728 }, { "epoch": 0.4689610807333548, "grad_norm": 2.479855570895971, "learning_rate": 1.1485023580552039e-06, "loss": 0.3672, "step": 729 }, { "epoch": 0.4696043743969122, "grad_norm": 2.5972761084926717, "learning_rate": 1.1464404848794216e-06, "loss": 0.3728, "step": 730 }, { "epoch": 0.4702476680604696, "grad_norm": 2.6546524319677154, "learning_rate": 1.1443779752976743e-06, "loss": 0.4561, "step": 731 }, { "epoch": 0.470890961724027, "grad_norm": 2.501580288616064, "learning_rate": 1.1423148382732853e-06, "loss": 0.384, "step": 732 }, { "epoch": 0.4715342553875844, "grad_norm": 2.7049006113210066, "learning_rate": 1.1402510827723042e-06, "loss": 0.3946, "step": 733 }, { "epoch": 0.47217754905114184, "grad_norm": 2.5383253690315453, "learning_rate": 1.1381867177634693e-06, "loss": 0.3984, "step": 734 }, { "epoch": 0.47282084271469926, "grad_norm": 2.2185039097748636, "learning_rate": 1.1361217522181666e-06, "loss": 0.3626, "step": 735 }, { "epoch": 0.4734641363782567, "grad_norm": 2.8144354428618983, "learning_rate": 1.1340561951103928e-06, "loss": 0.381, "step": 736 }, { "epoch": 0.4741074300418141, "grad_norm": 2.1191682847727935, "learning_rate": 1.1319900554167156e-06, "loss": 0.3088, "step": 737 }, { "epoch": 0.4747507237053715, "grad_norm": 2.738594825411013, "learning_rate": 1.1299233421162338e-06, "loss": 0.4621, "step": 738 }, { "epoch": 0.4753940173689289, "grad_norm": 2.0212472395522187, "learning_rate": 1.1278560641905391e-06, "loss": 0.3369, "step": 739 }, { "epoch": 0.4760373110324863, "grad_norm": 2.521789070766898, "learning_rate": 1.1257882306236775e-06, "loss": 0.3439, "step": 740 }, { "epoch": 0.47668060469604373, "grad_norm": 2.223690445265059, "learning_rate": 1.1237198504021091e-06, "loss": 0.3008, "step": 741 }, { "epoch": 0.47732389835960115, "grad_norm": 2.755649068133421, "learning_rate": 1.12165093251467e-06, "loss": 0.4651, "step": 742 }, { "epoch": 0.47796719202315857, "grad_norm": 2.689275604991562, "learning_rate": 1.119581485952533e-06, "loss": 0.3858, "step": 743 }, { "epoch": 0.478610485686716, "grad_norm": 2.6153647415237247, "learning_rate": 1.117511519709168e-06, "loss": 0.3599, "step": 744 }, { "epoch": 0.4792537793502734, "grad_norm": 2.424683512116255, "learning_rate": 1.115441042780304e-06, "loss": 0.3609, "step": 745 }, { "epoch": 0.47989707301383083, "grad_norm": 2.297479285518035, "learning_rate": 1.113370064163889e-06, "loss": 0.2757, "step": 746 }, { "epoch": 0.48054036667738825, "grad_norm": 2.205321245590632, "learning_rate": 1.111298592860051e-06, "loss": 0.3099, "step": 747 }, { "epoch": 0.4811836603409456, "grad_norm": 2.4060777049859627, "learning_rate": 1.10922663787106e-06, "loss": 0.3403, "step": 748 }, { "epoch": 0.48182695400450304, "grad_norm": 2.5386176661114592, "learning_rate": 1.1071542082012865e-06, "loss": 0.3636, "step": 749 }, { "epoch": 0.48247024766806046, "grad_norm": 2.448840430006797, "learning_rate": 1.1050813128571651e-06, "loss": 0.3711, "step": 750 }, { "epoch": 0.4831135413316179, "grad_norm": 2.4352239001736704, "learning_rate": 1.1030079608471542e-06, "loss": 0.391, "step": 751 }, { "epoch": 0.4837568349951753, "grad_norm": 2.3151553041201702, "learning_rate": 1.1009341611816962e-06, "loss": 0.4223, "step": 752 }, { "epoch": 0.4844001286587327, "grad_norm": 2.7249837148916685, "learning_rate": 1.0988599228731792e-06, "loss": 0.4728, "step": 753 }, { "epoch": 0.48504342232229014, "grad_norm": 2.6963773054590394, "learning_rate": 1.0967852549358974e-06, "loss": 0.4364, "step": 754 }, { "epoch": 0.48568671598584756, "grad_norm": 2.4559926010787962, "learning_rate": 1.094710166386012e-06, "loss": 0.3373, "step": 755 }, { "epoch": 0.4863300096494049, "grad_norm": 2.6891220590027687, "learning_rate": 1.092634666241513e-06, "loss": 0.4349, "step": 756 }, { "epoch": 0.48697330331296235, "grad_norm": 2.604396948863969, "learning_rate": 1.090558763522178e-06, "loss": 0.381, "step": 757 }, { "epoch": 0.48761659697651977, "grad_norm": 2.3980586834681694, "learning_rate": 1.0884824672495345e-06, "loss": 0.4184, "step": 758 }, { "epoch": 0.4882598906400772, "grad_norm": 2.688696431522434, "learning_rate": 1.0864057864468208e-06, "loss": 0.3812, "step": 759 }, { "epoch": 0.4889031843036346, "grad_norm": 2.593638284073519, "learning_rate": 1.0843287301389453e-06, "loss": 0.3903, "step": 760 }, { "epoch": 0.48954647796719203, "grad_norm": 2.840995356117963, "learning_rate": 1.0822513073524496e-06, "loss": 0.429, "step": 761 }, { "epoch": 0.49018977163074945, "grad_norm": 2.910200852855761, "learning_rate": 1.0801735271154667e-06, "loss": 0.4877, "step": 762 }, { "epoch": 0.49083306529430687, "grad_norm": 2.713443326689494, "learning_rate": 1.078095398457684e-06, "loss": 0.3765, "step": 763 }, { "epoch": 0.4914763589578643, "grad_norm": 2.5493063614999034, "learning_rate": 1.076016930410302e-06, "loss": 0.3446, "step": 764 }, { "epoch": 0.49211965262142165, "grad_norm": 2.4681368019467897, "learning_rate": 1.0739381320059977e-06, "loss": 0.3597, "step": 765 }, { "epoch": 0.4927629462849791, "grad_norm": 2.7867105576865163, "learning_rate": 1.071859012278882e-06, "loss": 0.367, "step": 766 }, { "epoch": 0.4934062399485365, "grad_norm": 2.1770704968193653, "learning_rate": 1.0697795802644633e-06, "loss": 0.3119, "step": 767 }, { "epoch": 0.4940495336120939, "grad_norm": 2.411304687061043, "learning_rate": 1.0676998449996073e-06, "loss": 0.3576, "step": 768 }, { "epoch": 0.49469282727565134, "grad_norm": 2.603418794713418, "learning_rate": 1.0656198155224967e-06, "loss": 0.4406, "step": 769 }, { "epoch": 0.49533612093920876, "grad_norm": 2.6988458183791066, "learning_rate": 1.0635395008725936e-06, "loss": 0.3934, "step": 770 }, { "epoch": 0.4959794146027662, "grad_norm": 2.608988906197852, "learning_rate": 1.0614589100905988e-06, "loss": 0.3637, "step": 771 }, { "epoch": 0.4966227082663236, "grad_norm": 2.239747712261335, "learning_rate": 1.0593780522184138e-06, "loss": 0.3533, "step": 772 }, { "epoch": 0.49726600192988096, "grad_norm": 2.3428051195187645, "learning_rate": 1.0572969362990997e-06, "loss": 0.3739, "step": 773 }, { "epoch": 0.4979092955934384, "grad_norm": 2.309028444608846, "learning_rate": 1.0552155713768406e-06, "loss": 0.4109, "step": 774 }, { "epoch": 0.4985525892569958, "grad_norm": 2.1772503359501107, "learning_rate": 1.0531339664969018e-06, "loss": 0.2872, "step": 775 }, { "epoch": 0.4991958829205532, "grad_norm": 2.7580611954684953, "learning_rate": 1.0510521307055913e-06, "loss": 0.3563, "step": 776 }, { "epoch": 0.49983917658411064, "grad_norm": 2.3019873290443216, "learning_rate": 1.0489700730502212e-06, "loss": 0.4332, "step": 777 }, { "epoch": 0.500482470247668, "grad_norm": 2.2860946940833213, "learning_rate": 1.0468878025790668e-06, "loss": 0.3473, "step": 778 }, { "epoch": 0.5011257639112254, "grad_norm": 2.429728090982181, "learning_rate": 1.044805328341329e-06, "loss": 0.3351, "step": 779 }, { "epoch": 0.5017690575747829, "grad_norm": 2.414352078529792, "learning_rate": 1.0427226593870946e-06, "loss": 0.3684, "step": 780 }, { "epoch": 0.5024123512383403, "grad_norm": 2.7549929264995856, "learning_rate": 1.0406398047672956e-06, "loss": 0.3269, "step": 781 }, { "epoch": 0.5030556449018977, "grad_norm": 2.4946450119783097, "learning_rate": 1.038556773533671e-06, "loss": 0.4547, "step": 782 }, { "epoch": 0.5036989385654551, "grad_norm": 2.99296450591758, "learning_rate": 1.036473574738729e-06, "loss": 0.3725, "step": 783 }, { "epoch": 0.5043422322290125, "grad_norm": 2.214695541777331, "learning_rate": 1.0343902174357038e-06, "loss": 0.3234, "step": 784 }, { "epoch": 0.50498552589257, "grad_norm": 2.1214971160749965, "learning_rate": 1.0323067106785198e-06, "loss": 0.3519, "step": 785 }, { "epoch": 0.5056288195561274, "grad_norm": 2.2490236142422098, "learning_rate": 1.0302230635217499e-06, "loss": 0.294, "step": 786 }, { "epoch": 0.5062721132196848, "grad_norm": 2.7234803319516057, "learning_rate": 1.028139285020578e-06, "loss": 0.3678, "step": 787 }, { "epoch": 0.5069154068832422, "grad_norm": 2.3234539202574247, "learning_rate": 1.0260553842307588e-06, "loss": 0.3329, "step": 788 }, { "epoch": 0.5075587005467996, "grad_norm": 2.512243866003769, "learning_rate": 1.023971370208578e-06, "loss": 0.4357, "step": 789 }, { "epoch": 0.508201994210357, "grad_norm": 2.5953227713738305, "learning_rate": 1.0218872520108133e-06, "loss": 0.3739, "step": 790 }, { "epoch": 0.5088452878739145, "grad_norm": 3.0399601195492187, "learning_rate": 1.0198030386946953e-06, "loss": 0.4287, "step": 791 }, { "epoch": 0.5094885815374719, "grad_norm": 2.467726379358667, "learning_rate": 1.0177187393178685e-06, "loss": 0.4141, "step": 792 }, { "epoch": 0.5101318752010293, "grad_norm": 2.86908936876761, "learning_rate": 1.0156343629383507e-06, "loss": 0.3814, "step": 793 }, { "epoch": 0.5107751688645867, "grad_norm": 2.4146917263824372, "learning_rate": 1.013549918614495e-06, "loss": 0.3546, "step": 794 }, { "epoch": 0.511418462528144, "grad_norm": 2.851237288491248, "learning_rate": 1.0114654154049489e-06, "loss": 0.4257, "step": 795 }, { "epoch": 0.5120617561917015, "grad_norm": 2.9863265790947624, "learning_rate": 1.0093808623686165e-06, "loss": 0.3641, "step": 796 }, { "epoch": 0.5127050498552589, "grad_norm": 2.756609023319894, "learning_rate": 1.007296268564618e-06, "loss": 0.3845, "step": 797 }, { "epoch": 0.5133483435188163, "grad_norm": 2.8200111475850798, "learning_rate": 1.0052116430522513e-06, "loss": 0.4278, "step": 798 }, { "epoch": 0.5139916371823737, "grad_norm": 2.8071664243222187, "learning_rate": 1.0031269948909515e-06, "loss": 0.3939, "step": 799 }, { "epoch": 0.5146349308459311, "grad_norm": 2.8642036998212137, "learning_rate": 1.0010423331402527e-06, "loss": 0.3471, "step": 800 }, { "epoch": 0.5152782245094886, "grad_norm": 2.2530723438337183, "learning_rate": 9.989576668597474e-07, "loss": 0.3599, "step": 801 }, { "epoch": 0.515921518173046, "grad_norm": 2.5791219213973897, "learning_rate": 9.968730051090485e-07, "loss": 0.4335, "step": 802 }, { "epoch": 0.5165648118366034, "grad_norm": 2.2395171486902883, "learning_rate": 9.947883569477489e-07, "loss": 0.402, "step": 803 }, { "epoch": 0.5172081055001608, "grad_norm": 2.570361313892079, "learning_rate": 9.92703731435382e-07, "loss": 0.3093, "step": 804 }, { "epoch": 0.5178513991637183, "grad_norm": 2.155293415220677, "learning_rate": 9.906191376313834e-07, "loss": 0.3127, "step": 805 }, { "epoch": 0.5184946928272757, "grad_norm": 2.7460418397121282, "learning_rate": 9.885345845950508e-07, "loss": 0.3963, "step": 806 }, { "epoch": 0.5191379864908331, "grad_norm": 2.3205903174078797, "learning_rate": 9.86450081385505e-07, "loss": 0.4208, "step": 807 }, { "epoch": 0.5197812801543905, "grad_norm": 2.536483393046217, "learning_rate": 9.843656370616488e-07, "loss": 0.3578, "step": 808 }, { "epoch": 0.5204245738179479, "grad_norm": 2.512001355329892, "learning_rate": 9.822812606821314e-07, "loss": 0.3939, "step": 809 }, { "epoch": 0.5210678674815054, "grad_norm": 2.596475389184174, "learning_rate": 9.801969613053046e-07, "loss": 0.4201, "step": 810 }, { "epoch": 0.5217111611450628, "grad_norm": 2.390638095448119, "learning_rate": 9.78112747989187e-07, "loss": 0.364, "step": 811 }, { "epoch": 0.5223544548086201, "grad_norm": 2.577786175268092, "learning_rate": 9.76028629791422e-07, "loss": 0.3731, "step": 812 }, { "epoch": 0.5229977484721775, "grad_norm": 2.7078458844031883, "learning_rate": 9.739446157692415e-07, "loss": 0.3677, "step": 813 }, { "epoch": 0.5236410421357349, "grad_norm": 2.6213990844803847, "learning_rate": 9.718607149794218e-07, "loss": 0.3817, "step": 814 }, { "epoch": 0.5242843357992923, "grad_norm": 2.541928418018708, "learning_rate": 9.6977693647825e-07, "loss": 0.3724, "step": 815 }, { "epoch": 0.5249276294628498, "grad_norm": 2.475653121217444, "learning_rate": 9.676932893214805e-07, "loss": 0.4175, "step": 816 }, { "epoch": 0.5255709231264072, "grad_norm": 2.753704744272801, "learning_rate": 9.65609782564296e-07, "loss": 0.3917, "step": 817 }, { "epoch": 0.5262142167899646, "grad_norm": 2.6209294776278744, "learning_rate": 9.635264252612711e-07, "loss": 0.4126, "step": 818 }, { "epoch": 0.526857510453522, "grad_norm": 2.8113843941653602, "learning_rate": 9.614432264663286e-07, "loss": 0.3256, "step": 819 }, { "epoch": 0.5275008041170794, "grad_norm": 2.374944965980044, "learning_rate": 9.593601952327047e-07, "loss": 0.3466, "step": 820 }, { "epoch": 0.5281440977806369, "grad_norm": 2.363051255657755, "learning_rate": 9.572773406129055e-07, "loss": 0.3788, "step": 821 }, { "epoch": 0.5287873914441943, "grad_norm": 3.3134345100268905, "learning_rate": 9.551946716586713e-07, "loss": 0.4743, "step": 822 }, { "epoch": 0.5294306851077517, "grad_norm": 2.7784235819891334, "learning_rate": 9.531121974209333e-07, "loss": 0.3285, "step": 823 }, { "epoch": 0.5300739787713091, "grad_norm": 2.7705668193767146, "learning_rate": 9.51029926949779e-07, "loss": 0.3475, "step": 824 }, { "epoch": 0.5307172724348665, "grad_norm": 2.6546973220442096, "learning_rate": 9.489478692944086e-07, "loss": 0.372, "step": 825 }, { "epoch": 0.531360566098424, "grad_norm": 2.3933555958078903, "learning_rate": 9.468660335030979e-07, "loss": 0.3565, "step": 826 }, { "epoch": 0.5320038597619814, "grad_norm": 2.3594884114093415, "learning_rate": 9.447844286231594e-07, "loss": 0.3742, "step": 827 }, { "epoch": 0.5326471534255388, "grad_norm": 2.3069626418667757, "learning_rate": 9.427030637009002e-07, "loss": 0.3776, "step": 828 }, { "epoch": 0.5332904470890961, "grad_norm": 3.1620394889233743, "learning_rate": 9.406219477815867e-07, "loss": 0.3705, "step": 829 }, { "epoch": 0.5339337407526535, "grad_norm": 2.7291111837569706, "learning_rate": 9.385410899094013e-07, "loss": 0.3705, "step": 830 }, { "epoch": 0.534577034416211, "grad_norm": 2.2239145694518756, "learning_rate": 9.364604991274068e-07, "loss": 0.3903, "step": 831 }, { "epoch": 0.5352203280797684, "grad_norm": 2.413620395972032, "learning_rate": 9.343801844775033e-07, "loss": 0.358, "step": 832 }, { "epoch": 0.5358636217433258, "grad_norm": 2.8193688074700516, "learning_rate": 9.323001550003929e-07, "loss": 0.3976, "step": 833 }, { "epoch": 0.5365069154068832, "grad_norm": 2.7978317887152877, "learning_rate": 9.302204197355367e-07, "loss": 0.4256, "step": 834 }, { "epoch": 0.5371502090704406, "grad_norm": 2.34009646651624, "learning_rate": 9.281409877211178e-07, "loss": 0.3789, "step": 835 }, { "epoch": 0.5377935027339981, "grad_norm": 2.0591817412611, "learning_rate": 9.260618679940024e-07, "loss": 0.2361, "step": 836 }, { "epoch": 0.5384367963975555, "grad_norm": 2.352183868965661, "learning_rate": 9.239830695896977e-07, "loss": 0.3705, "step": 837 }, { "epoch": 0.5390800900611129, "grad_norm": 2.793811305892386, "learning_rate": 9.219046015423162e-07, "loss": 0.3354, "step": 838 }, { "epoch": 0.5397233837246703, "grad_norm": 2.9046431390920695, "learning_rate": 9.198264728845331e-07, "loss": 0.3656, "step": 839 }, { "epoch": 0.5403666773882277, "grad_norm": 2.693292297026887, "learning_rate": 9.177486926475506e-07, "loss": 0.3864, "step": 840 }, { "epoch": 0.5410099710517852, "grad_norm": 2.657294963349999, "learning_rate": 9.156712698610546e-07, "loss": 0.3966, "step": 841 }, { "epoch": 0.5416532647153426, "grad_norm": 2.9805182409867355, "learning_rate": 9.135942135531795e-07, "loss": 0.3651, "step": 842 }, { "epoch": 0.5422965583789, "grad_norm": 2.529521390486679, "learning_rate": 9.115175327504655e-07, "loss": 0.3839, "step": 843 }, { "epoch": 0.5429398520424574, "grad_norm": 2.7250574537708068, "learning_rate": 9.094412364778221e-07, "loss": 0.3822, "step": 844 }, { "epoch": 0.5435831457060148, "grad_norm": 2.2948512416542393, "learning_rate": 9.07365333758487e-07, "loss": 0.3127, "step": 845 }, { "epoch": 0.5442264393695722, "grad_norm": 2.3185847561089274, "learning_rate": 9.052898336139879e-07, "loss": 0.3795, "step": 846 }, { "epoch": 0.5448697330331296, "grad_norm": 2.611273892070118, "learning_rate": 9.032147450641028e-07, "loss": 0.4263, "step": 847 }, { "epoch": 0.545513026696687, "grad_norm": 2.66621079143366, "learning_rate": 9.011400771268208e-07, "loss": 0.4104, "step": 848 }, { "epoch": 0.5461563203602444, "grad_norm": 2.6854044826083228, "learning_rate": 8.990658388183038e-07, "loss": 0.4012, "step": 849 }, { "epoch": 0.5467996140238018, "grad_norm": 2.7246361034562763, "learning_rate": 8.969920391528457e-07, "loss": 0.342, "step": 850 }, { "epoch": 0.5474429076873593, "grad_norm": 2.45559948090457, "learning_rate": 8.949186871428347e-07, "loss": 0.3862, "step": 851 }, { "epoch": 0.5480862013509167, "grad_norm": 2.45330890850409, "learning_rate": 8.928457917987136e-07, "loss": 0.3961, "step": 852 }, { "epoch": 0.5487294950144741, "grad_norm": 2.387296821061437, "learning_rate": 8.9077336212894e-07, "loss": 0.3583, "step": 853 }, { "epoch": 0.5493727886780315, "grad_norm": 2.4333004411336376, "learning_rate": 8.887014071399488e-07, "loss": 0.4198, "step": 854 }, { "epoch": 0.5500160823415889, "grad_norm": 2.29932627631256, "learning_rate": 8.866299358361107e-07, "loss": 0.4041, "step": 855 }, { "epoch": 0.5506593760051464, "grad_norm": 2.3028456446571606, "learning_rate": 8.845589572196959e-07, "loss": 0.3335, "step": 856 }, { "epoch": 0.5513026696687038, "grad_norm": 2.860004079180836, "learning_rate": 8.824884802908319e-07, "loss": 0.4036, "step": 857 }, { "epoch": 0.5519459633322612, "grad_norm": 2.418895658411742, "learning_rate": 8.804185140474674e-07, "loss": 0.3573, "step": 858 }, { "epoch": 0.5525892569958186, "grad_norm": 2.521302907962032, "learning_rate": 8.783490674853301e-07, "loss": 0.3732, "step": 859 }, { "epoch": 0.553232550659376, "grad_norm": 2.8181982305206024, "learning_rate": 8.762801495978913e-07, "loss": 0.4245, "step": 860 }, { "epoch": 0.5538758443229335, "grad_norm": 2.8904368883478604, "learning_rate": 8.742117693763227e-07, "loss": 0.4289, "step": 861 }, { "epoch": 0.5545191379864909, "grad_norm": 2.732752539252915, "learning_rate": 8.721439358094611e-07, "loss": 0.2899, "step": 862 }, { "epoch": 0.5551624316500482, "grad_norm": 2.520091244112061, "learning_rate": 8.700766578837663e-07, "loss": 0.2917, "step": 863 }, { "epoch": 0.5558057253136056, "grad_norm": 2.2266456390805, "learning_rate": 8.680099445832841e-07, "loss": 0.3544, "step": 864 }, { "epoch": 0.556449018977163, "grad_norm": 2.5562295344024566, "learning_rate": 8.659438048896071e-07, "loss": 0.3962, "step": 865 }, { "epoch": 0.5570923126407205, "grad_norm": 2.506782504822629, "learning_rate": 8.638782477818334e-07, "loss": 0.3868, "step": 866 }, { "epoch": 0.5577356063042779, "grad_norm": 2.4265804046452057, "learning_rate": 8.61813282236531e-07, "loss": 0.3256, "step": 867 }, { "epoch": 0.5583788999678353, "grad_norm": 2.5800057974774124, "learning_rate": 8.597489172276957e-07, "loss": 0.4055, "step": 868 }, { "epoch": 0.5590221936313927, "grad_norm": 2.6746468788670263, "learning_rate": 8.576851617267149e-07, "loss": 0.4431, "step": 869 }, { "epoch": 0.5596654872949501, "grad_norm": 2.368342541515128, "learning_rate": 8.556220247023256e-07, "loss": 0.3746, "step": 870 }, { "epoch": 0.5603087809585076, "grad_norm": 2.373686849477396, "learning_rate": 8.535595151205786e-07, "loss": 0.3088, "step": 871 }, { "epoch": 0.560952074622065, "grad_norm": 2.669830104361449, "learning_rate": 8.514976419447963e-07, "loss": 0.3784, "step": 872 }, { "epoch": 0.5615953682856224, "grad_norm": 2.542812203112294, "learning_rate": 8.494364141355366e-07, "loss": 0.4206, "step": 873 }, { "epoch": 0.5622386619491798, "grad_norm": 2.5729267648708407, "learning_rate": 8.473758406505529e-07, "loss": 0.3736, "step": 874 }, { "epoch": 0.5628819556127372, "grad_norm": 2.5682086273846947, "learning_rate": 8.453159304447533e-07, "loss": 0.3262, "step": 875 }, { "epoch": 0.5635252492762947, "grad_norm": 2.6572938029968136, "learning_rate": 8.432566924701658e-07, "loss": 0.3996, "step": 876 }, { "epoch": 0.5641685429398521, "grad_norm": 2.700244633843925, "learning_rate": 8.411981356758949e-07, "loss": 0.3301, "step": 877 }, { "epoch": 0.5648118366034095, "grad_norm": 2.8517428520508066, "learning_rate": 8.391402690080863e-07, "loss": 0.4153, "step": 878 }, { "epoch": 0.5654551302669669, "grad_norm": 1.9404103953526364, "learning_rate": 8.370831014098853e-07, "loss": 0.2684, "step": 879 }, { "epoch": 0.5660984239305242, "grad_norm": 2.4880493970480804, "learning_rate": 8.350266418214005e-07, "loss": 0.4081, "step": 880 }, { "epoch": 0.5667417175940817, "grad_norm": 2.6578339530762314, "learning_rate": 8.329708991796619e-07, "loss": 0.3333, "step": 881 }, { "epoch": 0.5673850112576391, "grad_norm": 2.2479462389545484, "learning_rate": 8.309158824185853e-07, "loss": 0.3187, "step": 882 }, { "epoch": 0.5680283049211965, "grad_norm": 2.4022771719376665, "learning_rate": 8.288616004689319e-07, "loss": 0.3933, "step": 883 }, { "epoch": 0.5686715985847539, "grad_norm": 2.4147432370461255, "learning_rate": 8.268080622582682e-07, "loss": 0.3718, "step": 884 }, { "epoch": 0.5693148922483113, "grad_norm": 2.6270562687629124, "learning_rate": 8.247552767109299e-07, "loss": 0.394, "step": 885 }, { "epoch": 0.5699581859118688, "grad_norm": 2.7468555349122834, "learning_rate": 8.227032527479806e-07, "loss": 0.3595, "step": 886 }, { "epoch": 0.5706014795754262, "grad_norm": 2.517201466483775, "learning_rate": 8.206519992871756e-07, "loss": 0.3082, "step": 887 }, { "epoch": 0.5712447732389836, "grad_norm": 2.6465288331908514, "learning_rate": 8.1860152524292e-07, "loss": 0.4365, "step": 888 }, { "epoch": 0.571888066902541, "grad_norm": 1.9669141340564429, "learning_rate": 8.165518395262334e-07, "loss": 0.32, "step": 889 }, { "epoch": 0.5725313605660984, "grad_norm": 2.5890472046608055, "learning_rate": 8.145029510447079e-07, "loss": 0.4083, "step": 890 }, { "epoch": 0.5731746542296559, "grad_norm": 2.522303261890387, "learning_rate": 8.124548687024722e-07, "loss": 0.3402, "step": 891 }, { "epoch": 0.5738179478932133, "grad_norm": 2.3455417081611785, "learning_rate": 8.104076014001513e-07, "loss": 0.3479, "step": 892 }, { "epoch": 0.5744612415567707, "grad_norm": 2.7180119873303954, "learning_rate": 8.083611580348277e-07, "loss": 0.3739, "step": 893 }, { "epoch": 0.5751045352203281, "grad_norm": 2.5399675302429072, "learning_rate": 8.063155475000035e-07, "loss": 0.3013, "step": 894 }, { "epoch": 0.5757478288838855, "grad_norm": 2.016940981716488, "learning_rate": 8.042707786855617e-07, "loss": 0.3395, "step": 895 }, { "epoch": 0.576391122547443, "grad_norm": 2.5754773638921367, "learning_rate": 8.022268604777271e-07, "loss": 0.4394, "step": 896 }, { "epoch": 0.5770344162110003, "grad_norm": 2.929183781151791, "learning_rate": 8.001838017590277e-07, "loss": 0.4237, "step": 897 }, { "epoch": 0.5776777098745577, "grad_norm": 2.4041178710122746, "learning_rate": 7.981416114082567e-07, "loss": 0.3341, "step": 898 }, { "epoch": 0.5783210035381151, "grad_norm": 2.181103418609087, "learning_rate": 7.961002983004335e-07, "loss": 0.3491, "step": 899 }, { "epoch": 0.5789642972016725, "grad_norm": 2.6472706940469846, "learning_rate": 7.940598713067646e-07, "loss": 0.3398, "step": 900 }, { "epoch": 0.57960759086523, "grad_norm": 2.4234890114508967, "learning_rate": 7.920203392946063e-07, "loss": 0.3484, "step": 901 }, { "epoch": 0.5802508845287874, "grad_norm": 2.5487526489864827, "learning_rate": 7.899817111274248e-07, "loss": 0.3936, "step": 902 }, { "epoch": 0.5808941781923448, "grad_norm": 2.4305012475592025, "learning_rate": 7.879439956647593e-07, "loss": 0.3807, "step": 903 }, { "epoch": 0.5815374718559022, "grad_norm": 2.6848406121583404, "learning_rate": 7.85907201762181e-07, "loss": 0.3279, "step": 904 }, { "epoch": 0.5821807655194596, "grad_norm": 2.4206666789839453, "learning_rate": 7.838713382712583e-07, "loss": 0.3459, "step": 905 }, { "epoch": 0.582824059183017, "grad_norm": 2.4798771820387473, "learning_rate": 7.818364140395137e-07, "loss": 0.3597, "step": 906 }, { "epoch": 0.5834673528465745, "grad_norm": 2.405035221848343, "learning_rate": 7.798024379103903e-07, "loss": 0.4017, "step": 907 }, { "epoch": 0.5841106465101319, "grad_norm": 3.0245039555932873, "learning_rate": 7.777694187232085e-07, "loss": 0.3419, "step": 908 }, { "epoch": 0.5847539401736893, "grad_norm": 2.2609857947952836, "learning_rate": 7.757373653131322e-07, "loss": 0.415, "step": 909 }, { "epoch": 0.5853972338372467, "grad_norm": 2.425942592631888, "learning_rate": 7.737062865111264e-07, "loss": 0.4043, "step": 910 }, { "epoch": 0.5860405275008042, "grad_norm": 2.6226269835913683, "learning_rate": 7.716761911439216e-07, "loss": 0.3779, "step": 911 }, { "epoch": 0.5866838211643616, "grad_norm": 2.4369598263446854, "learning_rate": 7.69647088033975e-07, "loss": 0.3659, "step": 912 }, { "epoch": 0.587327114827919, "grad_norm": 2.4358705394020954, "learning_rate": 7.676189859994298e-07, "loss": 0.3864, "step": 913 }, { "epoch": 0.5879704084914763, "grad_norm": 2.707379177576887, "learning_rate": 7.655918938540807e-07, "loss": 0.4921, "step": 914 }, { "epoch": 0.5886137021550337, "grad_norm": 2.7493157370564214, "learning_rate": 7.635658204073319e-07, "loss": 0.3908, "step": 915 }, { "epoch": 0.5892569958185911, "grad_norm": 2.6268076052133535, "learning_rate": 7.615407744641618e-07, "loss": 0.3408, "step": 916 }, { "epoch": 0.5899002894821486, "grad_norm": 2.659558090131179, "learning_rate": 7.595167648250824e-07, "loss": 0.4085, "step": 917 }, { "epoch": 0.590543583145706, "grad_norm": 2.420202831435382, "learning_rate": 7.574938002861032e-07, "loss": 0.4021, "step": 918 }, { "epoch": 0.5911868768092634, "grad_norm": 2.413642031480048, "learning_rate": 7.554718896386906e-07, "loss": 0.3504, "step": 919 }, { "epoch": 0.5918301704728208, "grad_norm": 2.8224804237091776, "learning_rate": 7.53451041669732e-07, "loss": 0.3258, "step": 920 }, { "epoch": 0.5924734641363782, "grad_norm": 2.682784798433877, "learning_rate": 7.514312651614964e-07, "loss": 0.3837, "step": 921 }, { "epoch": 0.5931167577999357, "grad_norm": 2.191020836623082, "learning_rate": 7.494125688915959e-07, "loss": 0.2902, "step": 922 }, { "epoch": 0.5937600514634931, "grad_norm": 2.701098354610045, "learning_rate": 7.473949616329487e-07, "loss": 0.4306, "step": 923 }, { "epoch": 0.5944033451270505, "grad_norm": 2.5509139332873842, "learning_rate": 7.453784521537392e-07, "loss": 0.2964, "step": 924 }, { "epoch": 0.5950466387906079, "grad_norm": 2.4041757540198265, "learning_rate": 7.43363049217383e-07, "loss": 0.3707, "step": 925 }, { "epoch": 0.5956899324541653, "grad_norm": 2.453939305561846, "learning_rate": 7.413487615824846e-07, "loss": 0.3514, "step": 926 }, { "epoch": 0.5963332261177228, "grad_norm": 2.0980366082982207, "learning_rate": 7.393355980028038e-07, "loss": 0.3295, "step": 927 }, { "epoch": 0.5969765197812802, "grad_norm": 2.1187843392321524, "learning_rate": 7.373235672272132e-07, "loss": 0.3503, "step": 928 }, { "epoch": 0.5976198134448376, "grad_norm": 2.9025916225193416, "learning_rate": 7.35312677999664e-07, "loss": 0.3959, "step": 929 }, { "epoch": 0.598263107108395, "grad_norm": 2.2714323282043223, "learning_rate": 7.333029390591465e-07, "loss": 0.3502, "step": 930 }, { "epoch": 0.5989064007719523, "grad_norm": 2.525574708513991, "learning_rate": 7.312943591396508e-07, "loss": 0.3937, "step": 931 }, { "epoch": 0.5995496944355098, "grad_norm": 2.699807347027198, "learning_rate": 7.292869469701314e-07, "loss": 0.415, "step": 932 }, { "epoch": 0.6001929880990672, "grad_norm": 2.44693834667427, "learning_rate": 7.272807112744665e-07, "loss": 0.346, "step": 933 }, { "epoch": 0.6008362817626246, "grad_norm": 2.6206243173895705, "learning_rate": 7.252756607714237e-07, "loss": 0.3767, "step": 934 }, { "epoch": 0.601479575426182, "grad_norm": 2.3775113346881143, "learning_rate": 7.232718041746171e-07, "loss": 0.3937, "step": 935 }, { "epoch": 0.6021228690897394, "grad_norm": 3.1401855684125204, "learning_rate": 7.212691501924752e-07, "loss": 0.382, "step": 936 }, { "epoch": 0.6027661627532969, "grad_norm": 2.4491465753332786, "learning_rate": 7.192677075281977e-07, "loss": 0.3129, "step": 937 }, { "epoch": 0.6034094564168543, "grad_norm": 2.6162320129496988, "learning_rate": 7.172674848797217e-07, "loss": 0.3165, "step": 938 }, { "epoch": 0.6040527500804117, "grad_norm": 2.337612678550995, "learning_rate": 7.152684909396823e-07, "loss": 0.3646, "step": 939 }, { "epoch": 0.6046960437439691, "grad_norm": 2.3377444809260406, "learning_rate": 7.132707343953731e-07, "loss": 0.3685, "step": 940 }, { "epoch": 0.6053393374075265, "grad_norm": 2.365930508675933, "learning_rate": 7.112742239287126e-07, "loss": 0.3494, "step": 941 }, { "epoch": 0.605982631071084, "grad_norm": 2.856578082143714, "learning_rate": 7.092789682162024e-07, "loss": 0.3956, "step": 942 }, { "epoch": 0.6066259247346414, "grad_norm": 2.0067347489238236, "learning_rate": 7.072849759288913e-07, "loss": 0.313, "step": 943 }, { "epoch": 0.6072692183981988, "grad_norm": 2.5236317485575275, "learning_rate": 7.052922557323382e-07, "loss": 0.3758, "step": 944 }, { "epoch": 0.6079125120617562, "grad_norm": 2.4197374314783953, "learning_rate": 7.033008162865732e-07, "loss": 0.3566, "step": 945 }, { "epoch": 0.6085558057253136, "grad_norm": 2.616582639655053, "learning_rate": 7.013106662460604e-07, "loss": 0.4657, "step": 946 }, { "epoch": 0.6091990993888711, "grad_norm": 2.4009828216716107, "learning_rate": 6.993218142596608e-07, "loss": 0.383, "step": 947 }, { "epoch": 0.6098423930524285, "grad_norm": 2.441296893373011, "learning_rate": 6.973342689705936e-07, "loss": 0.3788, "step": 948 }, { "epoch": 0.6104856867159858, "grad_norm": 2.3568523406570754, "learning_rate": 6.953480390164e-07, "loss": 0.347, "step": 949 }, { "epoch": 0.6111289803795432, "grad_norm": 2.566703900739419, "learning_rate": 6.933631330289049e-07, "loss": 0.3462, "step": 950 }, { "epoch": 0.6117722740431006, "grad_norm": 2.6239533755569817, "learning_rate": 6.913795596341785e-07, "loss": 0.3908, "step": 951 }, { "epoch": 0.6124155677066581, "grad_norm": 2.630407521688983, "learning_rate": 6.893973274525014e-07, "loss": 0.4559, "step": 952 }, { "epoch": 0.6130588613702155, "grad_norm": 2.536599818072339, "learning_rate": 6.874164450983237e-07, "loss": 0.391, "step": 953 }, { "epoch": 0.6137021550337729, "grad_norm": 2.42317230773736, "learning_rate": 6.854369211802314e-07, "loss": 0.3247, "step": 954 }, { "epoch": 0.6143454486973303, "grad_norm": 2.8394724690593045, "learning_rate": 6.834587643009047e-07, "loss": 0.3701, "step": 955 }, { "epoch": 0.6149887423608877, "grad_norm": 2.512410714347182, "learning_rate": 6.81481983057085e-07, "loss": 0.3267, "step": 956 }, { "epoch": 0.6156320360244452, "grad_norm": 2.9776309777251604, "learning_rate": 6.795065860395334e-07, "loss": 0.4385, "step": 957 }, { "epoch": 0.6162753296880026, "grad_norm": 2.4461836077181105, "learning_rate": 6.775325818329973e-07, "loss": 0.2994, "step": 958 }, { "epoch": 0.61691862335156, "grad_norm": 2.3612697835756116, "learning_rate": 6.755599790161705e-07, "loss": 0.3741, "step": 959 }, { "epoch": 0.6175619170151174, "grad_norm": 2.6831743465447464, "learning_rate": 6.735887861616555e-07, "loss": 0.3739, "step": 960 }, { "epoch": 0.6182052106786748, "grad_norm": 2.577130499886927, "learning_rate": 6.716190118359292e-07, "loss": 0.3444, "step": 961 }, { "epoch": 0.6188485043422323, "grad_norm": 2.4050897324073723, "learning_rate": 6.696506645993018e-07, "loss": 0.3482, "step": 962 }, { "epoch": 0.6194917980057897, "grad_norm": 2.6454004160577154, "learning_rate": 6.676837530058838e-07, "loss": 0.4106, "step": 963 }, { "epoch": 0.6201350916693471, "grad_norm": 3.13370991390132, "learning_rate": 6.657182856035446e-07, "loss": 0.3656, "step": 964 }, { "epoch": 0.6207783853329045, "grad_norm": 2.1850930871688004, "learning_rate": 6.637542709338788e-07, "loss": 0.3282, "step": 965 }, { "epoch": 0.6214216789964618, "grad_norm": 2.676370536547326, "learning_rate": 6.617917175321667e-07, "loss": 0.4169, "step": 966 }, { "epoch": 0.6220649726600193, "grad_norm": 2.617398245380988, "learning_rate": 6.598306339273392e-07, "loss": 0.3848, "step": 967 }, { "epoch": 0.6227082663235767, "grad_norm": 2.801166333188507, "learning_rate": 6.578710286419387e-07, "loss": 0.3254, "step": 968 }, { "epoch": 0.6233515599871341, "grad_norm": 2.63974146561537, "learning_rate": 6.559129101920835e-07, "loss": 0.3913, "step": 969 }, { "epoch": 0.6239948536506915, "grad_norm": 2.5472836084503916, "learning_rate": 6.539562870874309e-07, "loss": 0.3866, "step": 970 }, { "epoch": 0.6246381473142489, "grad_norm": 2.7399294019712, "learning_rate": 6.520011678311381e-07, "loss": 0.4217, "step": 971 }, { "epoch": 0.6252814409778064, "grad_norm": 2.3435138098466086, "learning_rate": 6.50047560919829e-07, "loss": 0.3209, "step": 972 }, { "epoch": 0.6259247346413638, "grad_norm": 2.6072652746840768, "learning_rate": 6.48095474843553e-07, "loss": 0.3779, "step": 973 }, { "epoch": 0.6265680283049212, "grad_norm": 2.841218792044382, "learning_rate": 6.461449180857521e-07, "loss": 0.3664, "step": 974 }, { "epoch": 0.6272113219684786, "grad_norm": 2.4671673849558298, "learning_rate": 6.441958991232202e-07, "loss": 0.3469, "step": 975 }, { "epoch": 0.627854615632036, "grad_norm": 2.5969550450330066, "learning_rate": 6.422484264260697e-07, "loss": 0.3507, "step": 976 }, { "epoch": 0.6284979092955935, "grad_norm": 2.742838386210247, "learning_rate": 6.403025084576932e-07, "loss": 0.3989, "step": 977 }, { "epoch": 0.6291412029591509, "grad_norm": 2.5571503705356746, "learning_rate": 6.383581536747251e-07, "loss": 0.4009, "step": 978 }, { "epoch": 0.6297844966227083, "grad_norm": 2.5560004693936564, "learning_rate": 6.364153705270088e-07, "loss": 0.2938, "step": 979 }, { "epoch": 0.6304277902862657, "grad_norm": 2.3865844584322904, "learning_rate": 6.344741674575553e-07, "loss": 0.329, "step": 980 }, { "epoch": 0.6310710839498231, "grad_norm": 3.1126489796268597, "learning_rate": 6.325345529025106e-07, "loss": 0.4469, "step": 981 }, { "epoch": 0.6317143776133806, "grad_norm": 3.0135128353554914, "learning_rate": 6.305965352911161e-07, "loss": 0.396, "step": 982 }, { "epoch": 0.6323576712769379, "grad_norm": 2.4756066420258196, "learning_rate": 6.286601230456743e-07, "loss": 0.3739, "step": 983 }, { "epoch": 0.6330009649404953, "grad_norm": 2.1917417432156987, "learning_rate": 6.267253245815098e-07, "loss": 0.3415, "step": 984 }, { "epoch": 0.6336442586040527, "grad_norm": 2.5731061749792405, "learning_rate": 6.24792148306935e-07, "loss": 0.4095, "step": 985 }, { "epoch": 0.6342875522676101, "grad_norm": 2.3256684857890284, "learning_rate": 6.228606026232117e-07, "loss": 0.3421, "step": 986 }, { "epoch": 0.6349308459311676, "grad_norm": 2.5627786654566433, "learning_rate": 6.20930695924516e-07, "loss": 0.4079, "step": 987 }, { "epoch": 0.635574139594725, "grad_norm": 2.1823468473345526, "learning_rate": 6.190024365979017e-07, "loss": 0.3281, "step": 988 }, { "epoch": 0.6362174332582824, "grad_norm": 2.438242095475629, "learning_rate": 6.170758330232621e-07, "loss": 0.3187, "step": 989 }, { "epoch": 0.6368607269218398, "grad_norm": 3.4319193263731127, "learning_rate": 6.151508935732955e-07, "loss": 0.4334, "step": 990 }, { "epoch": 0.6375040205853972, "grad_norm": 2.646725386365017, "learning_rate": 6.132276266134687e-07, "loss": 0.3283, "step": 991 }, { "epoch": 0.6381473142489547, "grad_norm": 2.33726710548443, "learning_rate": 6.113060405019793e-07, "loss": 0.3819, "step": 992 }, { "epoch": 0.6387906079125121, "grad_norm": 2.5801273545882513, "learning_rate": 6.093861435897207e-07, "loss": 0.3273, "step": 993 }, { "epoch": 0.6394339015760695, "grad_norm": 2.224943771398719, "learning_rate": 6.074679442202455e-07, "loss": 0.3288, "step": 994 }, { "epoch": 0.6400771952396269, "grad_norm": 2.555282853541804, "learning_rate": 6.055514507297283e-07, "loss": 0.3867, "step": 995 }, { "epoch": 0.6407204889031843, "grad_norm": 2.604321446497435, "learning_rate": 6.03636671446931e-07, "loss": 0.3558, "step": 996 }, { "epoch": 0.6413637825667418, "grad_norm": 2.047997091685558, "learning_rate": 6.017236146931658e-07, "loss": 0.331, "step": 997 }, { "epoch": 0.6420070762302992, "grad_norm": 2.828854072056512, "learning_rate": 5.998122887822582e-07, "loss": 0.3623, "step": 998 }, { "epoch": 0.6426503698938566, "grad_norm": 3.0451012240974915, "learning_rate": 5.979027020205131e-07, "loss": 0.3043, "step": 999 }, { "epoch": 0.6432936635574139, "grad_norm": 2.5117706840048326, "learning_rate": 5.959948627066759e-07, "loss": 0.3872, "step": 1000 }, { "epoch": 0.6439369572209713, "grad_norm": 2.036139837841544, "learning_rate": 5.940887791318992e-07, "loss": 0.3472, "step": 1001 }, { "epoch": 0.6445802508845287, "grad_norm": 2.6212505696177484, "learning_rate": 5.921844595797043e-07, "loss": 0.4086, "step": 1002 }, { "epoch": 0.6452235445480862, "grad_norm": 2.7446949176984834, "learning_rate": 5.902819123259477e-07, "loss": 0.3896, "step": 1003 }, { "epoch": 0.6458668382116436, "grad_norm": 2.4560391710143628, "learning_rate": 5.88381145638782e-07, "loss": 0.2689, "step": 1004 }, { "epoch": 0.646510131875201, "grad_norm": 2.144151226611934, "learning_rate": 5.864821677786234e-07, "loss": 0.3822, "step": 1005 }, { "epoch": 0.6471534255387584, "grad_norm": 2.2916813446390782, "learning_rate": 5.845849869981136e-07, "loss": 0.3412, "step": 1006 }, { "epoch": 0.6477967192023159, "grad_norm": 2.723886076680928, "learning_rate": 5.826896115420839e-07, "loss": 0.4364, "step": 1007 }, { "epoch": 0.6484400128658733, "grad_norm": 2.451680015441699, "learning_rate": 5.807960496475212e-07, "loss": 0.3489, "step": 1008 }, { "epoch": 0.6490833065294307, "grad_norm": 3.4941125388721295, "learning_rate": 5.789043095435289e-07, "loss": 0.5515, "step": 1009 }, { "epoch": 0.6497266001929881, "grad_norm": 2.6074463111252855, "learning_rate": 5.770143994512962e-07, "loss": 0.3321, "step": 1010 }, { "epoch": 0.6503698938565455, "grad_norm": 2.433908968071575, "learning_rate": 5.751263275840561e-07, "loss": 0.3648, "step": 1011 }, { "epoch": 0.651013187520103, "grad_norm": 2.501917535401103, "learning_rate": 5.73240102147056e-07, "loss": 0.373, "step": 1012 }, { "epoch": 0.6516564811836604, "grad_norm": 2.356113740054644, "learning_rate": 5.713557313375161e-07, "loss": 0.3247, "step": 1013 }, { "epoch": 0.6522997748472178, "grad_norm": 2.2164036765722774, "learning_rate": 5.694732233445995e-07, "loss": 0.2759, "step": 1014 }, { "epoch": 0.6529430685107752, "grad_norm": 2.2290270721869043, "learning_rate": 5.67592586349372e-07, "loss": 0.3456, "step": 1015 }, { "epoch": 0.6535863621743326, "grad_norm": 2.5090113098479017, "learning_rate": 5.657138285247687e-07, "loss": 0.3498, "step": 1016 }, { "epoch": 0.65422965583789, "grad_norm": 2.6429360744402883, "learning_rate": 5.638369580355587e-07, "loss": 0.368, "step": 1017 }, { "epoch": 0.6548729495014474, "grad_norm": 2.7116278829868317, "learning_rate": 5.619619830383083e-07, "loss": 0.3672, "step": 1018 }, { "epoch": 0.6555162431650048, "grad_norm": 2.8126162855402574, "learning_rate": 5.600889116813466e-07, "loss": 0.3326, "step": 1019 }, { "epoch": 0.6561595368285622, "grad_norm": 2.4482559112280198, "learning_rate": 5.582177521047301e-07, "loss": 0.3444, "step": 1020 }, { "epoch": 0.6568028304921196, "grad_norm": 2.7043509110051738, "learning_rate": 5.563485124402067e-07, "loss": 0.3996, "step": 1021 }, { "epoch": 0.657446124155677, "grad_norm": 2.35820911306754, "learning_rate": 5.544812008111807e-07, "loss": 0.2653, "step": 1022 }, { "epoch": 0.6580894178192345, "grad_norm": 2.9275320699157383, "learning_rate": 5.526158253326776e-07, "loss": 0.4029, "step": 1023 }, { "epoch": 0.6587327114827919, "grad_norm": 2.6509032665059573, "learning_rate": 5.507523941113086e-07, "loss": 0.393, "step": 1024 }, { "epoch": 0.6593760051463493, "grad_norm": 2.3374047945059235, "learning_rate": 5.488909152452356e-07, "loss": 0.3125, "step": 1025 }, { "epoch": 0.6600192988099067, "grad_norm": 2.377137350167575, "learning_rate": 5.470313968241358e-07, "loss": 0.3063, "step": 1026 }, { "epoch": 0.6606625924734641, "grad_norm": 2.9426989700952695, "learning_rate": 5.451738469291665e-07, "loss": 0.3679, "step": 1027 }, { "epoch": 0.6613058861370216, "grad_norm": 2.9339043310412265, "learning_rate": 5.433182736329305e-07, "loss": 0.3507, "step": 1028 }, { "epoch": 0.661949179800579, "grad_norm": 2.360750970494329, "learning_rate": 5.414646849994403e-07, "loss": 0.3519, "step": 1029 }, { "epoch": 0.6625924734641364, "grad_norm": 2.595324056464516, "learning_rate": 5.396130890840833e-07, "loss": 0.3653, "step": 1030 }, { "epoch": 0.6632357671276938, "grad_norm": 2.4559340826333815, "learning_rate": 5.377634939335873e-07, "loss": 0.3399, "step": 1031 }, { "epoch": 0.6638790607912513, "grad_norm": 2.179561874357909, "learning_rate": 5.359159075859847e-07, "loss": 0.3225, "step": 1032 }, { "epoch": 0.6645223544548087, "grad_norm": 2.525392286864598, "learning_rate": 5.340703380705785e-07, "loss": 0.3605, "step": 1033 }, { "epoch": 0.665165648118366, "grad_norm": 2.802046876500464, "learning_rate": 5.322267934079055e-07, "loss": 0.3353, "step": 1034 }, { "epoch": 0.6658089417819234, "grad_norm": 2.8167595699972914, "learning_rate": 5.303852816097057e-07, "loss": 0.3658, "step": 1035 }, { "epoch": 0.6664522354454808, "grad_norm": 2.4193792323240646, "learning_rate": 5.285458106788808e-07, "loss": 0.4362, "step": 1036 }, { "epoch": 0.6670955291090382, "grad_norm": 2.272038347226993, "learning_rate": 5.267083886094668e-07, "loss": 0.4042, "step": 1037 }, { "epoch": 0.6677388227725957, "grad_norm": 2.4278224806359856, "learning_rate": 5.248730233865925e-07, "loss": 0.4008, "step": 1038 }, { "epoch": 0.6683821164361531, "grad_norm": 2.376761119387891, "learning_rate": 5.230397229864514e-07, "loss": 0.362, "step": 1039 }, { "epoch": 0.6690254100997105, "grad_norm": 2.322539817616314, "learning_rate": 5.212084953762599e-07, "loss": 0.3276, "step": 1040 }, { "epoch": 0.6696687037632679, "grad_norm": 2.8161303700802245, "learning_rate": 5.1937934851423e-07, "loss": 0.3664, "step": 1041 }, { "epoch": 0.6703119974268253, "grad_norm": 2.931313509082492, "learning_rate": 5.17552290349528e-07, "loss": 0.3856, "step": 1042 }, { "epoch": 0.6709552910903828, "grad_norm": 2.52529340521223, "learning_rate": 5.157273288222455e-07, "loss": 0.2968, "step": 1043 }, { "epoch": 0.6715985847539402, "grad_norm": 2.472898267761653, "learning_rate": 5.139044718633609e-07, "loss": 0.3171, "step": 1044 }, { "epoch": 0.6722418784174976, "grad_norm": 2.470623723442538, "learning_rate": 5.120837273947072e-07, "loss": 0.3997, "step": 1045 }, { "epoch": 0.672885172081055, "grad_norm": 2.3697080614542654, "learning_rate": 5.102651033289369e-07, "loss": 0.3794, "step": 1046 }, { "epoch": 0.6735284657446124, "grad_norm": 2.594869617826655, "learning_rate": 5.08448607569487e-07, "loss": 0.3445, "step": 1047 }, { "epoch": 0.6741717594081699, "grad_norm": 2.5148224662589795, "learning_rate": 5.066342480105459e-07, "loss": 0.3674, "step": 1048 }, { "epoch": 0.6748150530717273, "grad_norm": 2.594013003725887, "learning_rate": 5.04822032537018e-07, "loss": 0.343, "step": 1049 }, { "epoch": 0.6754583467352847, "grad_norm": 2.2125487986631156, "learning_rate": 5.030119690244903e-07, "loss": 0.3517, "step": 1050 }, { "epoch": 0.676101640398842, "grad_norm": 2.7822334840135277, "learning_rate": 5.012040653391971e-07, "loss": 0.4163, "step": 1051 }, { "epoch": 0.6767449340623994, "grad_norm": 2.8040152928050928, "learning_rate": 4.99398329337987e-07, "loss": 0.396, "step": 1052 }, { "epoch": 0.6773882277259569, "grad_norm": 2.2357913459062555, "learning_rate": 4.97594768868288e-07, "loss": 0.3425, "step": 1053 }, { "epoch": 0.6780315213895143, "grad_norm": 2.4915438509231387, "learning_rate": 4.957933917680734e-07, "loss": 0.414, "step": 1054 }, { "epoch": 0.6786748150530717, "grad_norm": 2.5847385801363267, "learning_rate": 4.939942058658285e-07, "loss": 0.3802, "step": 1055 }, { "epoch": 0.6793181087166291, "grad_norm": 2.4041679684443755, "learning_rate": 4.921972189805154e-07, "loss": 0.3037, "step": 1056 }, { "epoch": 0.6799614023801865, "grad_norm": 2.6559373624493734, "learning_rate": 4.9040243892154e-07, "loss": 0.4407, "step": 1057 }, { "epoch": 0.680604696043744, "grad_norm": 2.781415202886524, "learning_rate": 4.886098734887177e-07, "loss": 0.4068, "step": 1058 }, { "epoch": 0.6812479897073014, "grad_norm": 2.047192716336921, "learning_rate": 4.868195304722391e-07, "loss": 0.3065, "step": 1059 }, { "epoch": 0.6818912833708588, "grad_norm": 2.428187967162864, "learning_rate": 4.850314176526372e-07, "loss": 0.3147, "step": 1060 }, { "epoch": 0.6825345770344162, "grad_norm": 2.748925784873729, "learning_rate": 4.832455428007525e-07, "loss": 0.3823, "step": 1061 }, { "epoch": 0.6831778706979736, "grad_norm": 2.548479732581523, "learning_rate": 4.814619136776997e-07, "loss": 0.351, "step": 1062 }, { "epoch": 0.6838211643615311, "grad_norm": 2.3557145310265697, "learning_rate": 4.796805380348336e-07, "loss": 0.2871, "step": 1063 }, { "epoch": 0.6844644580250885, "grad_norm": 2.38311578699362, "learning_rate": 4.779014236137173e-07, "loss": 0.4304, "step": 1064 }, { "epoch": 0.6851077516886459, "grad_norm": 2.4091833966936322, "learning_rate": 4.7612457814608453e-07, "loss": 0.3403, "step": 1065 }, { "epoch": 0.6857510453522033, "grad_norm": 2.69197900632577, "learning_rate": 4.743500093538111e-07, "loss": 0.4043, "step": 1066 }, { "epoch": 0.6863943390157607, "grad_norm": 2.3822155422228266, "learning_rate": 4.72577724948876e-07, "loss": 0.3585, "step": 1067 }, { "epoch": 0.687037632679318, "grad_norm": 2.5045212454437236, "learning_rate": 4.7080773263333394e-07, "loss": 0.3272, "step": 1068 }, { "epoch": 0.6876809263428755, "grad_norm": 2.3896808001164382, "learning_rate": 4.6904004009927513e-07, "loss": 0.3216, "step": 1069 }, { "epoch": 0.6883242200064329, "grad_norm": 2.2415405767597014, "learning_rate": 4.6727465502879846e-07, "loss": 0.3681, "step": 1070 }, { "epoch": 0.6889675136699903, "grad_norm": 2.0885330170123577, "learning_rate": 4.6551158509397213e-07, "loss": 0.3373, "step": 1071 }, { "epoch": 0.6896108073335477, "grad_norm": 2.6589910607166725, "learning_rate": 4.637508379568056e-07, "loss": 0.3483, "step": 1072 }, { "epoch": 0.6902541009971052, "grad_norm": 2.7496658259532563, "learning_rate": 4.6199242126921245e-07, "loss": 0.4054, "step": 1073 }, { "epoch": 0.6908973946606626, "grad_norm": 3.054617500579067, "learning_rate": 4.602363426729787e-07, "loss": 0.3799, "step": 1074 }, { "epoch": 0.69154068832422, "grad_norm": 2.5841292596025114, "learning_rate": 4.5848260979972965e-07, "loss": 0.3353, "step": 1075 }, { "epoch": 0.6921839819877774, "grad_norm": 2.5107419907414514, "learning_rate": 4.5673123027089644e-07, "loss": 0.3677, "step": 1076 }, { "epoch": 0.6928272756513348, "grad_norm": 2.269066876309873, "learning_rate": 4.549822116976827e-07, "loss": 0.3195, "step": 1077 }, { "epoch": 0.6934705693148923, "grad_norm": 2.429585568694504, "learning_rate": 4.5323556168103206e-07, "loss": 0.3411, "step": 1078 }, { "epoch": 0.6941138629784497, "grad_norm": 2.214740473941067, "learning_rate": 4.514912878115947e-07, "loss": 0.3712, "step": 1079 }, { "epoch": 0.6947571566420071, "grad_norm": 2.3397551991020813, "learning_rate": 4.497493976696944e-07, "loss": 0.3731, "step": 1080 }, { "epoch": 0.6954004503055645, "grad_norm": 2.7722563630512025, "learning_rate": 4.480098988252957e-07, "loss": 0.4082, "step": 1081 }, { "epoch": 0.6960437439691219, "grad_norm": 2.3447341432997364, "learning_rate": 4.462727988379709e-07, "loss": 0.413, "step": 1082 }, { "epoch": 0.6966870376326794, "grad_norm": 2.5269682478338042, "learning_rate": 4.445381052568674e-07, "loss": 0.3455, "step": 1083 }, { "epoch": 0.6973303312962368, "grad_norm": 2.794629419399601, "learning_rate": 4.428058256206746e-07, "loss": 0.3785, "step": 1084 }, { "epoch": 0.6979736249597941, "grad_norm": 2.5938642876894886, "learning_rate": 4.4107596745759146e-07, "loss": 0.3962, "step": 1085 }, { "epoch": 0.6986169186233515, "grad_norm": 2.8365326564117157, "learning_rate": 4.393485382852935e-07, "loss": 0.3507, "step": 1086 }, { "epoch": 0.6992602122869089, "grad_norm": 2.542794066526858, "learning_rate": 4.376235456109003e-07, "loss": 0.319, "step": 1087 }, { "epoch": 0.6999035059504664, "grad_norm": 2.572780501651036, "learning_rate": 4.3590099693094283e-07, "loss": 0.4048, "step": 1088 }, { "epoch": 0.7005467996140238, "grad_norm": 3.168677047337119, "learning_rate": 4.3418089973133075e-07, "loss": 0.4732, "step": 1089 }, { "epoch": 0.7011900932775812, "grad_norm": 2.3619115320591786, "learning_rate": 4.3246326148731983e-07, "loss": 0.3664, "step": 1090 }, { "epoch": 0.7018333869411386, "grad_norm": 2.526390463778878, "learning_rate": 4.30748089663481e-07, "loss": 0.3805, "step": 1091 }, { "epoch": 0.702476680604696, "grad_norm": 2.601496184955773, "learning_rate": 4.290353917136639e-07, "loss": 0.4501, "step": 1092 }, { "epoch": 0.7031199742682535, "grad_norm": 2.812837803806927, "learning_rate": 4.273251750809701e-07, "loss": 0.329, "step": 1093 }, { "epoch": 0.7037632679318109, "grad_norm": 2.460257411422819, "learning_rate": 4.25617447197715e-07, "loss": 0.3841, "step": 1094 }, { "epoch": 0.7044065615953683, "grad_norm": 2.5156718654595505, "learning_rate": 4.239122154854009e-07, "loss": 0.3637, "step": 1095 }, { "epoch": 0.7050498552589257, "grad_norm": 2.290264376175184, "learning_rate": 4.2220948735467965e-07, "loss": 0.3649, "step": 1096 }, { "epoch": 0.7056931489224831, "grad_norm": 2.45684989224742, "learning_rate": 4.2050927020532565e-07, "loss": 0.3157, "step": 1097 }, { "epoch": 0.7063364425860406, "grad_norm": 2.4171889551656593, "learning_rate": 4.188115714261979e-07, "loss": 0.3922, "step": 1098 }, { "epoch": 0.706979736249598, "grad_norm": 2.2110875004800676, "learning_rate": 4.171163983952144e-07, "loss": 0.3086, "step": 1099 }, { "epoch": 0.7076230299131554, "grad_norm": 2.36574692214889, "learning_rate": 4.154237584793131e-07, "loss": 0.3187, "step": 1100 }, { "epoch": 0.7082663235767128, "grad_norm": 2.2742141555701028, "learning_rate": 4.1373365903442657e-07, "loss": 0.3422, "step": 1101 }, { "epoch": 0.7089096172402701, "grad_norm": 2.137149971601677, "learning_rate": 4.1204610740544545e-07, "loss": 0.3214, "step": 1102 }, { "epoch": 0.7095529109038275, "grad_norm": 2.570475231903908, "learning_rate": 4.103611109261872e-07, "loss": 0.4052, "step": 1103 }, { "epoch": 0.710196204567385, "grad_norm": 2.650432972506939, "learning_rate": 4.0867867691936733e-07, "loss": 0.3497, "step": 1104 }, { "epoch": 0.7108394982309424, "grad_norm": 2.8827398355727114, "learning_rate": 4.069988126965624e-07, "loss": 0.3941, "step": 1105 }, { "epoch": 0.7114827918944998, "grad_norm": 2.7730681472806427, "learning_rate": 4.0532152555818434e-07, "loss": 0.4094, "step": 1106 }, { "epoch": 0.7121260855580572, "grad_norm": 2.6158338111618438, "learning_rate": 4.0364682279344197e-07, "loss": 0.3347, "step": 1107 }, { "epoch": 0.7127693792216147, "grad_norm": 2.4351790325365528, "learning_rate": 4.0197471168031636e-07, "loss": 0.3566, "step": 1108 }, { "epoch": 0.7134126728851721, "grad_norm": 2.44756699444697, "learning_rate": 4.0030519948552253e-07, "loss": 0.3227, "step": 1109 }, { "epoch": 0.7140559665487295, "grad_norm": 2.4773144879402555, "learning_rate": 3.986382934644835e-07, "loss": 0.4545, "step": 1110 }, { "epoch": 0.7146992602122869, "grad_norm": 2.6141198725442836, "learning_rate": 3.9697400086129507e-07, "loss": 0.3771, "step": 1111 }, { "epoch": 0.7153425538758443, "grad_norm": 2.567434810666701, "learning_rate": 3.953123289086957e-07, "loss": 0.3591, "step": 1112 }, { "epoch": 0.7159858475394018, "grad_norm": 2.3693021983875453, "learning_rate": 3.9365328482803504e-07, "loss": 0.3912, "step": 1113 }, { "epoch": 0.7166291412029592, "grad_norm": 2.3901225106151114, "learning_rate": 3.9199687582924246e-07, "loss": 0.2928, "step": 1114 }, { "epoch": 0.7172724348665166, "grad_norm": 2.1864298481289213, "learning_rate": 3.903431091107956e-07, "loss": 0.2965, "step": 1115 }, { "epoch": 0.717915728530074, "grad_norm": 2.353815978535782, "learning_rate": 3.8869199185968936e-07, "loss": 0.3512, "step": 1116 }, { "epoch": 0.7185590221936314, "grad_norm": 2.7181293471811263, "learning_rate": 3.870435312514043e-07, "loss": 0.4327, "step": 1117 }, { "epoch": 0.7192023158571889, "grad_norm": 2.6133224613365296, "learning_rate": 3.85397734449876e-07, "loss": 0.3732, "step": 1118 }, { "epoch": 0.7198456095207462, "grad_norm": 2.2313431430845356, "learning_rate": 3.83754608607463e-07, "loss": 0.3214, "step": 1119 }, { "epoch": 0.7204889031843036, "grad_norm": 2.267687386912498, "learning_rate": 3.8211416086491686e-07, "loss": 0.3354, "step": 1120 }, { "epoch": 0.721132196847861, "grad_norm": 2.271936846902601, "learning_rate": 3.8047639835135026e-07, "loss": 0.3152, "step": 1121 }, { "epoch": 0.7217754905114184, "grad_norm": 2.4545255337684857, "learning_rate": 3.7884132818420645e-07, "loss": 0.3597, "step": 1122 }, { "epoch": 0.7224187841749758, "grad_norm": 2.541293013480265, "learning_rate": 3.772089574692281e-07, "loss": 0.3313, "step": 1123 }, { "epoch": 0.7230620778385333, "grad_norm": 2.557220548092616, "learning_rate": 3.755792933004265e-07, "loss": 0.3814, "step": 1124 }, { "epoch": 0.7237053715020907, "grad_norm": 2.4806338337901455, "learning_rate": 3.739523427600508e-07, "loss": 0.3657, "step": 1125 }, { "epoch": 0.7243486651656481, "grad_norm": 2.6534248275579864, "learning_rate": 3.723281129185574e-07, "loss": 0.4085, "step": 1126 }, { "epoch": 0.7249919588292055, "grad_norm": 2.584294750188816, "learning_rate": 3.7070661083457833e-07, "loss": 0.353, "step": 1127 }, { "epoch": 0.725635252492763, "grad_norm": 2.8279461898469247, "learning_rate": 3.6908784355489164e-07, "loss": 0.3774, "step": 1128 }, { "epoch": 0.7262785461563204, "grad_norm": 2.483488272251326, "learning_rate": 3.6747181811439145e-07, "loss": 0.3481, "step": 1129 }, { "epoch": 0.7269218398198778, "grad_norm": 2.367595027638228, "learning_rate": 3.6585854153605366e-07, "loss": 0.3121, "step": 1130 }, { "epoch": 0.7275651334834352, "grad_norm": 2.0321562317501813, "learning_rate": 3.642480208309112e-07, "loss": 0.2731, "step": 1131 }, { "epoch": 0.7282084271469926, "grad_norm": 2.6847991549958694, "learning_rate": 3.6264026299801754e-07, "loss": 0.3898, "step": 1132 }, { "epoch": 0.72885172081055, "grad_norm": 2.6372423550932176, "learning_rate": 3.610352750244218e-07, "loss": 0.3842, "step": 1133 }, { "epoch": 0.7294950144741075, "grad_norm": 2.5826556933526286, "learning_rate": 3.5943306388513326e-07, "loss": 0.3676, "step": 1134 }, { "epoch": 0.7301383081376649, "grad_norm": 2.3377546200136226, "learning_rate": 3.5783363654309596e-07, "loss": 0.3593, "step": 1135 }, { "epoch": 0.7307816018012222, "grad_norm": 2.3739211773680857, "learning_rate": 3.5623699994915355e-07, "loss": 0.407, "step": 1136 }, { "epoch": 0.7314248954647796, "grad_norm": 2.974362779072703, "learning_rate": 3.54643161042024e-07, "loss": 0.415, "step": 1137 }, { "epoch": 0.732068189128337, "grad_norm": 2.7815441091458983, "learning_rate": 3.5305212674826524e-07, "loss": 0.4145, "step": 1138 }, { "epoch": 0.7327114827918945, "grad_norm": 2.270565832563137, "learning_rate": 3.5146390398224743e-07, "loss": 0.4011, "step": 1139 }, { "epoch": 0.7333547764554519, "grad_norm": 2.181389520842806, "learning_rate": 3.4987849964612246e-07, "loss": 0.342, "step": 1140 }, { "epoch": 0.7339980701190093, "grad_norm": 2.60536865625539, "learning_rate": 3.482959206297935e-07, "loss": 0.3495, "step": 1141 }, { "epoch": 0.7346413637825667, "grad_norm": 2.318003911026614, "learning_rate": 3.4671617381088546e-07, "loss": 0.2979, "step": 1142 }, { "epoch": 0.7352846574461241, "grad_norm": 2.1817598865441865, "learning_rate": 3.45139266054715e-07, "loss": 0.2953, "step": 1143 }, { "epoch": 0.7359279511096816, "grad_norm": 2.1304060488153445, "learning_rate": 3.435652042142606e-07, "loss": 0.3465, "step": 1144 }, { "epoch": 0.736571244773239, "grad_norm": 2.485629102181292, "learning_rate": 3.4199399513013295e-07, "loss": 0.4084, "step": 1145 }, { "epoch": 0.7372145384367964, "grad_norm": 2.574533259577946, "learning_rate": 3.404256456305452e-07, "loss": 0.383, "step": 1146 }, { "epoch": 0.7378578321003538, "grad_norm": 2.4668605417511467, "learning_rate": 3.3886016253128327e-07, "loss": 0.4137, "step": 1147 }, { "epoch": 0.7385011257639112, "grad_norm": 3.1165791352223913, "learning_rate": 3.372975526356757e-07, "loss": 0.423, "step": 1148 }, { "epoch": 0.7391444194274687, "grad_norm": 2.4027490460446077, "learning_rate": 3.357378227345652e-07, "loss": 0.352, "step": 1149 }, { "epoch": 0.7397877130910261, "grad_norm": 2.27746969517431, "learning_rate": 3.3418097960627823e-07, "loss": 0.3317, "step": 1150 }, { "epoch": 0.7404310067545835, "grad_norm": 2.192135106158721, "learning_rate": 3.3262703001659566e-07, "loss": 0.3556, "step": 1151 }, { "epoch": 0.7410743004181409, "grad_norm": 2.476416776293521, "learning_rate": 3.3107598071872376e-07, "loss": 0.3545, "step": 1152 }, { "epoch": 0.7417175940816984, "grad_norm": 2.7253992549433455, "learning_rate": 3.295278384532646e-07, "loss": 0.3682, "step": 1153 }, { "epoch": 0.7423608877452557, "grad_norm": 3.057339585035728, "learning_rate": 3.279826099481866e-07, "loss": 0.3428, "step": 1154 }, { "epoch": 0.7430041814088131, "grad_norm": 2.7073084430231384, "learning_rate": 3.264403019187956e-07, "loss": 0.2943, "step": 1155 }, { "epoch": 0.7436474750723705, "grad_norm": 2.5007419650028933, "learning_rate": 3.2490092106770537e-07, "loss": 0.3723, "step": 1156 }, { "epoch": 0.7442907687359279, "grad_norm": 2.3749868638955274, "learning_rate": 3.2336447408480873e-07, "loss": 0.3445, "step": 1157 }, { "epoch": 0.7449340623994853, "grad_norm": 2.566365739568945, "learning_rate": 3.2183096764724914e-07, "loss": 0.3577, "step": 1158 }, { "epoch": 0.7455773560630428, "grad_norm": 2.4123665163535155, "learning_rate": 3.2030040841938923e-07, "loss": 0.379, "step": 1159 }, { "epoch": 0.7462206497266002, "grad_norm": 2.5336607291283926, "learning_rate": 3.1877280305278566e-07, "loss": 0.2622, "step": 1160 }, { "epoch": 0.7468639433901576, "grad_norm": 2.5000859825486694, "learning_rate": 3.172481581861558e-07, "loss": 0.3903, "step": 1161 }, { "epoch": 0.747507237053715, "grad_norm": 2.309790746995557, "learning_rate": 3.1572648044535364e-07, "loss": 0.3747, "step": 1162 }, { "epoch": 0.7481505307172724, "grad_norm": 2.4901681360707655, "learning_rate": 3.142077764433361e-07, "loss": 0.4165, "step": 1163 }, { "epoch": 0.7487938243808299, "grad_norm": 2.5041462050051506, "learning_rate": 3.12692052780139e-07, "loss": 0.372, "step": 1164 }, { "epoch": 0.7494371180443873, "grad_norm": 2.187999909049904, "learning_rate": 3.1117931604284384e-07, "loss": 0.3211, "step": 1165 }, { "epoch": 0.7500804117079447, "grad_norm": 2.4482092142475858, "learning_rate": 3.096695728055536e-07, "loss": 0.3066, "step": 1166 }, { "epoch": 0.7507237053715021, "grad_norm": 2.2203686532279385, "learning_rate": 3.081628296293608e-07, "loss": 0.373, "step": 1167 }, { "epoch": 0.7513669990350595, "grad_norm": 2.379185155668647, "learning_rate": 3.066590930623202e-07, "loss": 0.3725, "step": 1168 }, { "epoch": 0.752010292698617, "grad_norm": 2.481388262019673, "learning_rate": 3.0515836963942054e-07, "loss": 0.3731, "step": 1169 }, { "epoch": 0.7526535863621744, "grad_norm": 2.485209467257747, "learning_rate": 3.036606658825561e-07, "loss": 0.4218, "step": 1170 }, { "epoch": 0.7532968800257317, "grad_norm": 2.4821701416618054, "learning_rate": 3.02165988300498e-07, "loss": 0.3561, "step": 1171 }, { "epoch": 0.7539401736892891, "grad_norm": 2.7737637669404873, "learning_rate": 3.006743433888662e-07, "loss": 0.4117, "step": 1172 }, { "epoch": 0.7545834673528465, "grad_norm": 2.33947386940943, "learning_rate": 2.99185737630101e-07, "loss": 0.3107, "step": 1173 }, { "epoch": 0.755226761016404, "grad_norm": 2.3477561356823746, "learning_rate": 2.977001774934352e-07, "loss": 0.3411, "step": 1174 }, { "epoch": 0.7558700546799614, "grad_norm": 2.306207575644613, "learning_rate": 2.962176694348659e-07, "loss": 0.2996, "step": 1175 }, { "epoch": 0.7565133483435188, "grad_norm": 2.2651101590735614, "learning_rate": 2.947382198971262e-07, "loss": 0.3446, "step": 1176 }, { "epoch": 0.7571566420070762, "grad_norm": 2.466988459711715, "learning_rate": 2.932618353096575e-07, "loss": 0.377, "step": 1177 }, { "epoch": 0.7577999356706336, "grad_norm": 2.3027724142512214, "learning_rate": 2.917885220885814e-07, "loss": 0.3017, "step": 1178 }, { "epoch": 0.7584432293341911, "grad_norm": 2.529884976732379, "learning_rate": 2.903182866366719e-07, "loss": 0.354, "step": 1179 }, { "epoch": 0.7590865229977485, "grad_norm": 2.306248754106011, "learning_rate": 2.8885113534332737e-07, "loss": 0.3093, "step": 1180 }, { "epoch": 0.7597298166613059, "grad_norm": 2.5972063130149032, "learning_rate": 2.8738707458454327e-07, "loss": 0.4029, "step": 1181 }, { "epoch": 0.7603731103248633, "grad_norm": 2.5651554629048396, "learning_rate": 2.8592611072288353e-07, "loss": 0.3605, "step": 1182 }, { "epoch": 0.7610164039884207, "grad_norm": 3.1017674329582223, "learning_rate": 2.8446825010745424e-07, "loss": 0.4101, "step": 1183 }, { "epoch": 0.7616596976519782, "grad_norm": 2.3962055377280373, "learning_rate": 2.830134990738746e-07, "loss": 0.3735, "step": 1184 }, { "epoch": 0.7623029913155356, "grad_norm": 2.9099014573059567, "learning_rate": 2.8156186394425054e-07, "loss": 0.3741, "step": 1185 }, { "epoch": 0.762946284979093, "grad_norm": 2.5096853693439374, "learning_rate": 2.8011335102714626e-07, "loss": 0.3269, "step": 1186 }, { "epoch": 0.7635895786426504, "grad_norm": 2.7733449319059105, "learning_rate": 2.7866796661755867e-07, "loss": 0.4281, "step": 1187 }, { "epoch": 0.7642328723062077, "grad_norm": 2.386733290902946, "learning_rate": 2.7722571699688655e-07, "loss": 0.4104, "step": 1188 }, { "epoch": 0.7648761659697652, "grad_norm": 2.6020565701878464, "learning_rate": 2.757866084329079e-07, "loss": 0.4354, "step": 1189 }, { "epoch": 0.7655194596333226, "grad_norm": 2.3331747738516118, "learning_rate": 2.743506471797475e-07, "loss": 0.3488, "step": 1190 }, { "epoch": 0.76616275329688, "grad_norm": 2.616309490818197, "learning_rate": 2.729178394778554e-07, "loss": 0.3773, "step": 1191 }, { "epoch": 0.7668060469604374, "grad_norm": 2.5501590199048927, "learning_rate": 2.7148819155397385e-07, "loss": 0.3592, "step": 1192 }, { "epoch": 0.7674493406239948, "grad_norm": 2.4612550975872254, "learning_rate": 2.7006170962111595e-07, "loss": 0.3669, "step": 1193 }, { "epoch": 0.7680926342875523, "grad_norm": 2.2419052962899633, "learning_rate": 2.6863839987853354e-07, "loss": 0.3578, "step": 1194 }, { "epoch": 0.7687359279511097, "grad_norm": 2.4276700763723067, "learning_rate": 2.672182685116946e-07, "loss": 0.4305, "step": 1195 }, { "epoch": 0.7693792216146671, "grad_norm": 2.4959090358045612, "learning_rate": 2.6580132169225333e-07, "loss": 0.3101, "step": 1196 }, { "epoch": 0.7700225152782245, "grad_norm": 3.0893439337875352, "learning_rate": 2.6438756557802477e-07, "loss": 0.3129, "step": 1197 }, { "epoch": 0.7706658089417819, "grad_norm": 2.3917554566331307, "learning_rate": 2.629770063129577e-07, "loss": 0.4038, "step": 1198 }, { "epoch": 0.7713091026053394, "grad_norm": 2.2625177937234495, "learning_rate": 2.615696500271072e-07, "loss": 0.3218, "step": 1199 }, { "epoch": 0.7719523962688968, "grad_norm": 2.513541778811717, "learning_rate": 2.6016550283661033e-07, "loss": 0.334, "step": 1200 }, { "epoch": 0.7725956899324542, "grad_norm": 2.849141912329549, "learning_rate": 2.5876457084365586e-07, "loss": 0.4009, "step": 1201 }, { "epoch": 0.7732389835960116, "grad_norm": 2.5232013592558022, "learning_rate": 2.573668601364623e-07, "loss": 0.3732, "step": 1202 }, { "epoch": 0.773882277259569, "grad_norm": 2.4473137035281205, "learning_rate": 2.559723767892463e-07, "loss": 0.3297, "step": 1203 }, { "epoch": 0.7745255709231265, "grad_norm": 2.546688273789288, "learning_rate": 2.5458112686220135e-07, "loss": 0.3662, "step": 1204 }, { "epoch": 0.7751688645866838, "grad_norm": 2.864710189952921, "learning_rate": 2.5319311640146755e-07, "loss": 0.3591, "step": 1205 }, { "epoch": 0.7758121582502412, "grad_norm": 2.360711183843298, "learning_rate": 2.5180835143910726e-07, "loss": 0.2978, "step": 1206 }, { "epoch": 0.7764554519137986, "grad_norm": 3.062199320751874, "learning_rate": 2.5042683799307865e-07, "loss": 0.3552, "step": 1207 }, { "epoch": 0.777098745577356, "grad_norm": 2.4610437354279586, "learning_rate": 2.490485820672089e-07, "loss": 0.3693, "step": 1208 }, { "epoch": 0.7777420392409135, "grad_norm": 2.385570665932814, "learning_rate": 2.476735896511689e-07, "loss": 0.2876, "step": 1209 }, { "epoch": 0.7783853329044709, "grad_norm": 2.357272174593535, "learning_rate": 2.4630186672044686e-07, "loss": 0.2546, "step": 1210 }, { "epoch": 0.7790286265680283, "grad_norm": 2.850403077574596, "learning_rate": 2.4493341923632195e-07, "loss": 0.3735, "step": 1211 }, { "epoch": 0.7796719202315857, "grad_norm": 2.0024309725797083, "learning_rate": 2.435682531458393e-07, "loss": 0.2811, "step": 1212 }, { "epoch": 0.7803152138951431, "grad_norm": 2.5394691035252275, "learning_rate": 2.4220637438178313e-07, "loss": 0.3015, "step": 1213 }, { "epoch": 0.7809585075587006, "grad_norm": 2.376208344812624, "learning_rate": 2.4084778886265267e-07, "loss": 0.4327, "step": 1214 }, { "epoch": 0.781601801222258, "grad_norm": 2.719368898936071, "learning_rate": 2.394925024926333e-07, "loss": 0.2927, "step": 1215 }, { "epoch": 0.7822450948858154, "grad_norm": 2.4252884431682, "learning_rate": 2.381405211615749e-07, "loss": 0.2957, "step": 1216 }, { "epoch": 0.7828883885493728, "grad_norm": 2.452804899911457, "learning_rate": 2.3679185074496244e-07, "loss": 0.3549, "step": 1217 }, { "epoch": 0.7835316822129302, "grad_norm": 2.0533750404763005, "learning_rate": 2.354464971038933e-07, "loss": 0.2836, "step": 1218 }, { "epoch": 0.7841749758764877, "grad_norm": 2.418521983331092, "learning_rate": 2.341044660850504e-07, "loss": 0.3512, "step": 1219 }, { "epoch": 0.7848182695400451, "grad_norm": 2.6659053645640007, "learning_rate": 2.3276576352067724e-07, "loss": 0.349, "step": 1220 }, { "epoch": 0.7854615632036025, "grad_norm": 2.8387501779823268, "learning_rate": 2.3143039522855202e-07, "loss": 0.3925, "step": 1221 }, { "epoch": 0.7861048568671598, "grad_norm": 2.579123915603667, "learning_rate": 2.3009836701196316e-07, "loss": 0.3353, "step": 1222 }, { "epoch": 0.7867481505307172, "grad_norm": 2.5678630734227346, "learning_rate": 2.287696846596835e-07, "loss": 0.3399, "step": 1223 }, { "epoch": 0.7873914441942746, "grad_norm": 3.6169729019663857, "learning_rate": 2.2744435394594497e-07, "loss": 0.3964, "step": 1224 }, { "epoch": 0.7880347378578321, "grad_norm": 2.2991046853032606, "learning_rate": 2.2612238063041512e-07, "loss": 0.3619, "step": 1225 }, { "epoch": 0.7886780315213895, "grad_norm": 2.632753436592174, "learning_rate": 2.2480377045816856e-07, "loss": 0.3956, "step": 1226 }, { "epoch": 0.7893213251849469, "grad_norm": 2.503267813406814, "learning_rate": 2.23488529159667e-07, "loss": 0.359, "step": 1227 }, { "epoch": 0.7899646188485043, "grad_norm": 2.6773431117656594, "learning_rate": 2.2217666245072863e-07, "loss": 0.3798, "step": 1228 }, { "epoch": 0.7906079125120618, "grad_norm": 2.54995109951605, "learning_rate": 2.2086817603250917e-07, "loss": 0.309, "step": 1229 }, { "epoch": 0.7912512061756192, "grad_norm": 2.5850417964301196, "learning_rate": 2.195630755914716e-07, "loss": 0.3601, "step": 1230 }, { "epoch": 0.7918944998391766, "grad_norm": 2.590381387653309, "learning_rate": 2.18261366799366e-07, "loss": 0.3929, "step": 1231 }, { "epoch": 0.792537793502734, "grad_norm": 2.5285254371392307, "learning_rate": 2.1696305531320115e-07, "loss": 0.3558, "step": 1232 }, { "epoch": 0.7931810871662914, "grad_norm": 2.3268557742520586, "learning_rate": 2.1566814677522337e-07, "loss": 0.318, "step": 1233 }, { "epoch": 0.7938243808298489, "grad_norm": 2.9046399025604166, "learning_rate": 2.1437664681288926e-07, "loss": 0.3459, "step": 1234 }, { "epoch": 0.7944676744934063, "grad_norm": 2.6176729865300277, "learning_rate": 2.1308856103884277e-07, "loss": 0.3331, "step": 1235 }, { "epoch": 0.7951109681569637, "grad_norm": 2.624955472987432, "learning_rate": 2.1180389505089003e-07, "loss": 0.3701, "step": 1236 }, { "epoch": 0.7957542618205211, "grad_norm": 2.106494733034664, "learning_rate": 2.1052265443197558e-07, "loss": 0.3766, "step": 1237 }, { "epoch": 0.7963975554840785, "grad_norm": 2.309511898791735, "learning_rate": 2.092448447501578e-07, "loss": 0.3094, "step": 1238 }, { "epoch": 0.7970408491476358, "grad_norm": 2.5381738036284682, "learning_rate": 2.0797047155858495e-07, "loss": 0.374, "step": 1239 }, { "epoch": 0.7976841428111933, "grad_norm": 2.382072181380231, "learning_rate": 2.0669954039547054e-07, "loss": 0.3696, "step": 1240 }, { "epoch": 0.7983274364747507, "grad_norm": 2.5401348595115425, "learning_rate": 2.0543205678406995e-07, "loss": 0.3238, "step": 1241 }, { "epoch": 0.7989707301383081, "grad_norm": 2.7609423732401672, "learning_rate": 2.0416802623265594e-07, "loss": 0.4252, "step": 1242 }, { "epoch": 0.7996140238018655, "grad_norm": 2.69724596817424, "learning_rate": 2.0290745423449462e-07, "loss": 0.3797, "step": 1243 }, { "epoch": 0.800257317465423, "grad_norm": 2.3356893256902755, "learning_rate": 2.0165034626782218e-07, "loss": 0.3896, "step": 1244 }, { "epoch": 0.8009006111289804, "grad_norm": 2.719800373476501, "learning_rate": 2.0039670779582063e-07, "loss": 0.3744, "step": 1245 }, { "epoch": 0.8015439047925378, "grad_norm": 2.5096882161688616, "learning_rate": 1.991465442665937e-07, "loss": 0.3937, "step": 1246 }, { "epoch": 0.8021871984560952, "grad_norm": 2.5258199349514356, "learning_rate": 1.9789986111314427e-07, "loss": 0.3625, "step": 1247 }, { "epoch": 0.8028304921196526, "grad_norm": 2.5854145031052562, "learning_rate": 1.9665666375334965e-07, "loss": 0.3542, "step": 1248 }, { "epoch": 0.80347378578321, "grad_norm": 2.417047523460248, "learning_rate": 1.9541695758993858e-07, "loss": 0.3288, "step": 1249 }, { "epoch": 0.8041170794467675, "grad_norm": 2.535288682441137, "learning_rate": 1.9418074801046758e-07, "loss": 0.3454, "step": 1250 }, { "epoch": 0.8047603731103249, "grad_norm": 2.409090619655627, "learning_rate": 1.929480403872974e-07, "loss": 0.2749, "step": 1251 }, { "epoch": 0.8054036667738823, "grad_norm": 2.3871269337170813, "learning_rate": 1.9171884007757088e-07, "loss": 0.2913, "step": 1252 }, { "epoch": 0.8060469604374397, "grad_norm": 2.8320744304697563, "learning_rate": 1.9049315242318698e-07, "loss": 0.349, "step": 1253 }, { "epoch": 0.8066902541009972, "grad_norm": 2.528703482084082, "learning_rate": 1.8927098275078111e-07, "loss": 0.3423, "step": 1254 }, { "epoch": 0.8073335477645546, "grad_norm": 2.5079940937263228, "learning_rate": 1.8805233637169825e-07, "loss": 0.311, "step": 1255 }, { "epoch": 0.8079768414281119, "grad_norm": 2.6040043395951726, "learning_rate": 1.8683721858197364e-07, "loss": 0.3486, "step": 1256 }, { "epoch": 0.8086201350916693, "grad_norm": 2.3261198212130316, "learning_rate": 1.8562563466230575e-07, "loss": 0.3978, "step": 1257 }, { "epoch": 0.8092634287552267, "grad_norm": 2.4775312789006843, "learning_rate": 1.8441758987803768e-07, "loss": 0.3164, "step": 1258 }, { "epoch": 0.8099067224187841, "grad_norm": 2.349293474408642, "learning_rate": 1.8321308947912995e-07, "loss": 0.3265, "step": 1259 }, { "epoch": 0.8105500160823416, "grad_norm": 2.257837186272317, "learning_rate": 1.8201213870014166e-07, "loss": 0.355, "step": 1260 }, { "epoch": 0.811193309745899, "grad_norm": 2.600995309517458, "learning_rate": 1.8081474276020404e-07, "loss": 0.354, "step": 1261 }, { "epoch": 0.8118366034094564, "grad_norm": 2.749568065840435, "learning_rate": 1.7962090686300113e-07, "loss": 0.4585, "step": 1262 }, { "epoch": 0.8124798970730138, "grad_norm": 2.210049157134129, "learning_rate": 1.784306361967447e-07, "loss": 0.3592, "step": 1263 }, { "epoch": 0.8131231907365712, "grad_norm": 2.2130569845594152, "learning_rate": 1.7724393593415298e-07, "loss": 0.3709, "step": 1264 }, { "epoch": 0.8137664844001287, "grad_norm": 2.6848504444650545, "learning_rate": 1.7606081123242778e-07, "loss": 0.4041, "step": 1265 }, { "epoch": 0.8144097780636861, "grad_norm": 2.3874489540862096, "learning_rate": 1.7488126723323182e-07, "loss": 0.3385, "step": 1266 }, { "epoch": 0.8150530717272435, "grad_norm": 2.643904295371436, "learning_rate": 1.7370530906266722e-07, "loss": 0.353, "step": 1267 }, { "epoch": 0.8156963653908009, "grad_norm": 2.5982984265749574, "learning_rate": 1.7253294183125222e-07, "loss": 0.3089, "step": 1268 }, { "epoch": 0.8163396590543583, "grad_norm": 2.50808965189434, "learning_rate": 1.7136417063389964e-07, "loss": 0.3133, "step": 1269 }, { "epoch": 0.8169829527179158, "grad_norm": 2.472806702071725, "learning_rate": 1.7019900054989445e-07, "loss": 0.4019, "step": 1270 }, { "epoch": 0.8176262463814732, "grad_norm": 2.1682109813030213, "learning_rate": 1.6903743664287196e-07, "loss": 0.3165, "step": 1271 }, { "epoch": 0.8182695400450306, "grad_norm": 2.53680611125159, "learning_rate": 1.6787948396079542e-07, "loss": 0.4219, "step": 1272 }, { "epoch": 0.8189128337085879, "grad_norm": 2.6356845964299183, "learning_rate": 1.6672514753593435e-07, "loss": 0.3566, "step": 1273 }, { "epoch": 0.8195561273721453, "grad_norm": 2.396721404413659, "learning_rate": 1.655744323848428e-07, "loss": 0.3831, "step": 1274 }, { "epoch": 0.8201994210357028, "grad_norm": 2.575748688609643, "learning_rate": 1.644273435083372e-07, "loss": 0.333, "step": 1275 }, { "epoch": 0.8208427146992602, "grad_norm": 2.5397436206581836, "learning_rate": 1.6328388589147468e-07, "loss": 0.3715, "step": 1276 }, { "epoch": 0.8214860083628176, "grad_norm": 2.2441380997089455, "learning_rate": 1.621440645035318e-07, "loss": 0.3733, "step": 1277 }, { "epoch": 0.822129302026375, "grad_norm": 2.837324125820888, "learning_rate": 1.610078842979824e-07, "loss": 0.404, "step": 1278 }, { "epoch": 0.8227725956899324, "grad_norm": 2.1946717892900667, "learning_rate": 1.5987535021247667e-07, "loss": 0.3284, "step": 1279 }, { "epoch": 0.8234158893534899, "grad_norm": 2.5743902781742585, "learning_rate": 1.5874646716881868e-07, "loss": 0.3746, "step": 1280 }, { "epoch": 0.8240591830170473, "grad_norm": 2.5766499581512314, "learning_rate": 1.576212400729472e-07, "loss": 0.2903, "step": 1281 }, { "epoch": 0.8247024766806047, "grad_norm": 2.227062027105539, "learning_rate": 1.5649967381491058e-07, "loss": 0.2889, "step": 1282 }, { "epoch": 0.8253457703441621, "grad_norm": 2.7186570460244233, "learning_rate": 1.5538177326885027e-07, "loss": 0.4825, "step": 1283 }, { "epoch": 0.8259890640077195, "grad_norm": 2.8853130035000745, "learning_rate": 1.5426754329297475e-07, "loss": 0.4025, "step": 1284 }, { "epoch": 0.826632357671277, "grad_norm": 2.469524369147359, "learning_rate": 1.5315698872954297e-07, "loss": 0.3348, "step": 1285 }, { "epoch": 0.8272756513348344, "grad_norm": 1.8832943030769767, "learning_rate": 1.5205011440483927e-07, "loss": 0.2336, "step": 1286 }, { "epoch": 0.8279189449983918, "grad_norm": 2.8091481817788138, "learning_rate": 1.509469251291562e-07, "loss": 0.3851, "step": 1287 }, { "epoch": 0.8285622386619492, "grad_norm": 2.2570932412896685, "learning_rate": 1.4984742569676955e-07, "loss": 0.3363, "step": 1288 }, { "epoch": 0.8292055323255066, "grad_norm": 2.356833736155564, "learning_rate": 1.4875162088592153e-07, "loss": 0.3157, "step": 1289 }, { "epoch": 0.829848825989064, "grad_norm": 2.1662804372922255, "learning_rate": 1.476595154587973e-07, "loss": 0.349, "step": 1290 }, { "epoch": 0.8304921196526214, "grad_norm": 2.215000077520576, "learning_rate": 1.4657111416150503e-07, "loss": 0.3425, "step": 1291 }, { "epoch": 0.8311354133161788, "grad_norm": 2.5191685625011306, "learning_rate": 1.4548642172405568e-07, "loss": 0.3397, "step": 1292 }, { "epoch": 0.8317787069797362, "grad_norm": 2.1267241589663857, "learning_rate": 1.4440544286034184e-07, "loss": 0.3157, "step": 1293 }, { "epoch": 0.8324220006432936, "grad_norm": 2.63835446719664, "learning_rate": 1.4332818226811806e-07, "loss": 0.37, "step": 1294 }, { "epoch": 0.8330652943068511, "grad_norm": 2.860905110913073, "learning_rate": 1.4225464462897875e-07, "loss": 0.398, "step": 1295 }, { "epoch": 0.8337085879704085, "grad_norm": 3.004768358025291, "learning_rate": 1.4118483460834063e-07, "loss": 0.4138, "step": 1296 }, { "epoch": 0.8343518816339659, "grad_norm": 2.821534557936075, "learning_rate": 1.4011875685541897e-07, "loss": 0.2755, "step": 1297 }, { "epoch": 0.8349951752975233, "grad_norm": 2.505369945569073, "learning_rate": 1.3905641600321129e-07, "loss": 0.3968, "step": 1298 }, { "epoch": 0.8356384689610807, "grad_norm": 2.6156652738937702, "learning_rate": 1.3799781666847299e-07, "loss": 0.4196, "step": 1299 }, { "epoch": 0.8362817626246382, "grad_norm": 2.3774600792492215, "learning_rate": 1.3694296345170119e-07, "loss": 0.3274, "step": 1300 }, { "epoch": 0.8369250562881956, "grad_norm": 2.2618459038978123, "learning_rate": 1.3589186093711223e-07, "loss": 0.3052, "step": 1301 }, { "epoch": 0.837568349951753, "grad_norm": 2.411573802930449, "learning_rate": 1.3484451369262251e-07, "loss": 0.343, "step": 1302 }, { "epoch": 0.8382116436153104, "grad_norm": 2.6656371072338882, "learning_rate": 1.3380092626982887e-07, "loss": 0.4198, "step": 1303 }, { "epoch": 0.8388549372788678, "grad_norm": 2.4614880134148, "learning_rate": 1.3276110320398847e-07, "loss": 0.3212, "step": 1304 }, { "epoch": 0.8394982309424253, "grad_norm": 2.5358956741289513, "learning_rate": 1.3172504901399917e-07, "loss": 0.3661, "step": 1305 }, { "epoch": 0.8401415246059827, "grad_norm": 2.549616701012764, "learning_rate": 1.3069276820237996e-07, "loss": 0.4214, "step": 1306 }, { "epoch": 0.84078481826954, "grad_norm": 2.6043655436134916, "learning_rate": 1.2966426525525142e-07, "loss": 0.3852, "step": 1307 }, { "epoch": 0.8414281119330974, "grad_norm": 2.5174766154346164, "learning_rate": 1.2863954464231618e-07, "loss": 0.3077, "step": 1308 }, { "epoch": 0.8420714055966548, "grad_norm": 2.492989948094338, "learning_rate": 1.2761861081683912e-07, "loss": 0.3646, "step": 1309 }, { "epoch": 0.8427146992602123, "grad_norm": 2.5529281513049584, "learning_rate": 1.2660146821562935e-07, "loss": 0.3736, "step": 1310 }, { "epoch": 0.8433579929237697, "grad_norm": 2.7587349387131055, "learning_rate": 1.255881212590183e-07, "loss": 0.3687, "step": 1311 }, { "epoch": 0.8440012865873271, "grad_norm": 2.353343475320849, "learning_rate": 1.2457857435084407e-07, "loss": 0.3645, "step": 1312 }, { "epoch": 0.8446445802508845, "grad_norm": 2.22694970681727, "learning_rate": 1.235728318784287e-07, "loss": 0.2896, "step": 1313 }, { "epoch": 0.8452878739144419, "grad_norm": 2.6534982931910043, "learning_rate": 1.2257089821256196e-07, "loss": 0.4448, "step": 1314 }, { "epoch": 0.8459311675779994, "grad_norm": 2.5491661098140823, "learning_rate": 1.2157277770748042e-07, "loss": 0.4112, "step": 1315 }, { "epoch": 0.8465744612415568, "grad_norm": 2.046891069706467, "learning_rate": 1.2057847470084992e-07, "loss": 0.3537, "step": 1316 }, { "epoch": 0.8472177549051142, "grad_norm": 2.6689769882731573, "learning_rate": 1.195879935137455e-07, "loss": 0.3468, "step": 1317 }, { "epoch": 0.8478610485686716, "grad_norm": 2.6182170917610335, "learning_rate": 1.1860133845063347e-07, "loss": 0.2811, "step": 1318 }, { "epoch": 0.848504342232229, "grad_norm": 2.480173741488141, "learning_rate": 1.1761851379935306e-07, "loss": 0.3267, "step": 1319 }, { "epoch": 0.8491476358957865, "grad_norm": 2.741769932975622, "learning_rate": 1.1663952383109566e-07, "loss": 0.4328, "step": 1320 }, { "epoch": 0.8497909295593439, "grad_norm": 2.318801933053353, "learning_rate": 1.1566437280038954e-07, "loss": 0.3456, "step": 1321 }, { "epoch": 0.8504342232229013, "grad_norm": 2.4570979844191094, "learning_rate": 1.1469306494507791e-07, "loss": 0.4007, "step": 1322 }, { "epoch": 0.8510775168864587, "grad_norm": 2.3792337223453397, "learning_rate": 1.1372560448630375e-07, "loss": 0.3619, "step": 1323 }, { "epoch": 0.851720810550016, "grad_norm": 2.5116210661711893, "learning_rate": 1.1276199562848842e-07, "loss": 0.3954, "step": 1324 }, { "epoch": 0.8523641042135734, "grad_norm": 2.3084510921915444, "learning_rate": 1.1180224255931625e-07, "loss": 0.3249, "step": 1325 }, { "epoch": 0.8530073978771309, "grad_norm": 2.7507737136109767, "learning_rate": 1.1084634944971349e-07, "loss": 0.3176, "step": 1326 }, { "epoch": 0.8536506915406883, "grad_norm": 2.411023062652163, "learning_rate": 1.0989432045383317e-07, "loss": 0.3681, "step": 1327 }, { "epoch": 0.8542939852042457, "grad_norm": 2.4056146506174825, "learning_rate": 1.0894615970903453e-07, "loss": 0.3871, "step": 1328 }, { "epoch": 0.8549372788678031, "grad_norm": 2.5647891443045543, "learning_rate": 1.0800187133586624e-07, "loss": 0.3193, "step": 1329 }, { "epoch": 0.8555805725313606, "grad_norm": 2.5153290486447304, "learning_rate": 1.0706145943804845e-07, "loss": 0.3478, "step": 1330 }, { "epoch": 0.856223866194918, "grad_norm": 2.267950217400543, "learning_rate": 1.0612492810245465e-07, "loss": 0.3396, "step": 1331 }, { "epoch": 0.8568671598584754, "grad_norm": 2.973885525644176, "learning_rate": 1.0519228139909398e-07, "loss": 0.3949, "step": 1332 }, { "epoch": 0.8575104535220328, "grad_norm": 2.4367152993838603, "learning_rate": 1.042635233810939e-07, "loss": 0.342, "step": 1333 }, { "epoch": 0.8581537471855902, "grad_norm": 2.1869645999092353, "learning_rate": 1.0333865808468201e-07, "loss": 0.2978, "step": 1334 }, { "epoch": 0.8587970408491477, "grad_norm": 2.562239245465279, "learning_rate": 1.0241768952916897e-07, "loss": 0.3572, "step": 1335 }, { "epoch": 0.8594403345127051, "grad_norm": 2.3031514757658886, "learning_rate": 1.0150062171693074e-07, "loss": 0.363, "step": 1336 }, { "epoch": 0.8600836281762625, "grad_norm": 2.3749685374954956, "learning_rate": 1.0058745863339135e-07, "loss": 0.3356, "step": 1337 }, { "epoch": 0.8607269218398199, "grad_norm": 2.3922866123230473, "learning_rate": 9.967820424700546e-08, "loss": 0.3313, "step": 1338 }, { "epoch": 0.8613702155033773, "grad_norm": 2.587511571391405, "learning_rate": 9.877286250924145e-08, "loss": 0.3029, "step": 1339 }, { "epoch": 0.8620135091669348, "grad_norm": 2.6587880711656764, "learning_rate": 9.787143735456349e-08, "loss": 0.4274, "step": 1340 }, { "epoch": 0.8626568028304921, "grad_norm": 2.9210067077890725, "learning_rate": 9.697393270041531e-08, "loss": 0.4137, "step": 1341 }, { "epoch": 0.8633000964940495, "grad_norm": 2.386184283338172, "learning_rate": 9.608035244720292e-08, "loss": 0.3233, "step": 1342 }, { "epoch": 0.8639433901576069, "grad_norm": 2.5639557799502293, "learning_rate": 9.519070047827704e-08, "loss": 0.4508, "step": 1343 }, { "epoch": 0.8645866838211643, "grad_norm": 2.7340492867133532, "learning_rate": 9.43049806599171e-08, "loss": 0.4, "step": 1344 }, { "epoch": 0.8652299774847217, "grad_norm": 2.447906867758782, "learning_rate": 9.342319684131395e-08, "loss": 0.3758, "step": 1345 }, { "epoch": 0.8658732711482792, "grad_norm": 2.650272207596033, "learning_rate": 9.254535285455333e-08, "loss": 0.324, "step": 1346 }, { "epoch": 0.8665165648118366, "grad_norm": 2.5308782002270087, "learning_rate": 9.167145251459873e-08, "loss": 0.3736, "step": 1347 }, { "epoch": 0.867159858475394, "grad_norm": 2.5755175025991304, "learning_rate": 9.08014996192762e-08, "loss": 0.3695, "step": 1348 }, { "epoch": 0.8678031521389514, "grad_norm": 2.3974041938321267, "learning_rate": 8.993549794925537e-08, "loss": 0.3519, "step": 1349 }, { "epoch": 0.8684464458025088, "grad_norm": 2.793246973570418, "learning_rate": 8.907345126803601e-08, "loss": 0.3904, "step": 1350 }, { "epoch": 0.8690897394660663, "grad_norm": 2.3640757689511167, "learning_rate": 8.82153633219288e-08, "loss": 0.4181, "step": 1351 }, { "epoch": 0.8697330331296237, "grad_norm": 2.2726163894814344, "learning_rate": 8.736123784004179e-08, "loss": 0.335, "step": 1352 }, { "epoch": 0.8703763267931811, "grad_norm": 2.794452805678077, "learning_rate": 8.651107853426143e-08, "loss": 0.3377, "step": 1353 }, { "epoch": 0.8710196204567385, "grad_norm": 2.0098865596095474, "learning_rate": 8.566488909923919e-08, "loss": 0.3359, "step": 1354 }, { "epoch": 0.871662914120296, "grad_norm": 2.2446376885575483, "learning_rate": 8.48226732123728e-08, "loss": 0.3303, "step": 1355 }, { "epoch": 0.8723062077838534, "grad_norm": 2.8469047070213866, "learning_rate": 8.398443453379266e-08, "loss": 0.3017, "step": 1356 }, { "epoch": 0.8729495014474108, "grad_norm": 2.2677008176227726, "learning_rate": 8.31501767063445e-08, "loss": 0.3262, "step": 1357 }, { "epoch": 0.8735927951109682, "grad_norm": 2.254518842540354, "learning_rate": 8.231990335557382e-08, "loss": 0.3804, "step": 1358 }, { "epoch": 0.8742360887745255, "grad_norm": 2.392492009237876, "learning_rate": 8.149361808971044e-08, "loss": 0.3223, "step": 1359 }, { "epoch": 0.8748793824380829, "grad_norm": 2.3246110745326205, "learning_rate": 8.067132449965253e-08, "loss": 0.3698, "step": 1360 }, { "epoch": 0.8755226761016404, "grad_norm": 2.250000953774854, "learning_rate": 7.985302615895118e-08, "loss": 0.3518, "step": 1361 }, { "epoch": 0.8761659697651978, "grad_norm": 2.4924882998888065, "learning_rate": 7.903872662379451e-08, "loss": 0.3699, "step": 1362 }, { "epoch": 0.8768092634287552, "grad_norm": 2.7504378653044754, "learning_rate": 7.82284294329929e-08, "loss": 0.34, "step": 1363 }, { "epoch": 0.8774525570923126, "grad_norm": 2.326164064871096, "learning_rate": 7.742213810796306e-08, "loss": 0.3256, "step": 1364 }, { "epoch": 0.87809585075587, "grad_norm": 2.4497859986695394, "learning_rate": 7.661985615271283e-08, "loss": 0.4278, "step": 1365 }, { "epoch": 0.8787391444194275, "grad_norm": 2.335275268413396, "learning_rate": 7.58215870538258e-08, "loss": 0.3967, "step": 1366 }, { "epoch": 0.8793824380829849, "grad_norm": 2.520491618054887, "learning_rate": 7.502733428044683e-08, "loss": 0.2901, "step": 1367 }, { "epoch": 0.8800257317465423, "grad_norm": 2.58560253922545, "learning_rate": 7.423710128426608e-08, "loss": 0.4059, "step": 1368 }, { "epoch": 0.8806690254100997, "grad_norm": 2.390429389374903, "learning_rate": 7.345089149950467e-08, "loss": 0.3664, "step": 1369 }, { "epoch": 0.8813123190736571, "grad_norm": 2.596167848378745, "learning_rate": 7.266870834289929e-08, "loss": 0.4088, "step": 1370 }, { "epoch": 0.8819556127372146, "grad_norm": 2.2817668551488475, "learning_rate": 7.189055521368792e-08, "loss": 0.3304, "step": 1371 }, { "epoch": 0.882598906400772, "grad_norm": 2.901149589826096, "learning_rate": 7.111643549359425e-08, "loss": 0.3874, "step": 1372 }, { "epoch": 0.8832422000643294, "grad_norm": 2.4656226594388624, "learning_rate": 7.034635254681386e-08, "loss": 0.4081, "step": 1373 }, { "epoch": 0.8838854937278868, "grad_norm": 2.792029996875874, "learning_rate": 6.958030971999884e-08, "loss": 0.364, "step": 1374 }, { "epoch": 0.8845287873914442, "grad_norm": 2.38710722380518, "learning_rate": 6.881831034224417e-08, "loss": 0.4031, "step": 1375 }, { "epoch": 0.8851720810550016, "grad_norm": 2.4092285283194133, "learning_rate": 6.806035772507168e-08, "loss": 0.3219, "step": 1376 }, { "epoch": 0.885815374718559, "grad_norm": 2.699979360802778, "learning_rate": 6.730645516241784e-08, "loss": 0.4023, "step": 1377 }, { "epoch": 0.8864586683821164, "grad_norm": 2.3880920896841675, "learning_rate": 6.655660593061718e-08, "loss": 0.4057, "step": 1378 }, { "epoch": 0.8871019620456738, "grad_norm": 2.6558465702761014, "learning_rate": 6.58108132883901e-08, "loss": 0.3795, "step": 1379 }, { "epoch": 0.8877452557092312, "grad_norm": 2.360235064440538, "learning_rate": 6.506908047682669e-08, "loss": 0.3418, "step": 1380 }, { "epoch": 0.8883885493727887, "grad_norm": 2.250009460233644, "learning_rate": 6.433141071937476e-08, "loss": 0.3114, "step": 1381 }, { "epoch": 0.8890318430363461, "grad_norm": 2.692066426488063, "learning_rate": 6.359780722182373e-08, "loss": 0.3826, "step": 1382 }, { "epoch": 0.8896751366999035, "grad_norm": 2.583059798874896, "learning_rate": 6.286827317229294e-08, "loss": 0.3541, "step": 1383 }, { "epoch": 0.8903184303634609, "grad_norm": 2.18124564741357, "learning_rate": 6.214281174121516e-08, "loss": 0.3505, "step": 1384 }, { "epoch": 0.8909617240270183, "grad_norm": 2.492885846253612, "learning_rate": 6.14214260813255e-08, "loss": 0.315, "step": 1385 }, { "epoch": 0.8916050176905758, "grad_norm": 2.387051942811312, "learning_rate": 6.070411932764585e-08, "loss": 0.3834, "step": 1386 }, { "epoch": 0.8922483113541332, "grad_norm": 2.4137110117979685, "learning_rate": 5.999089459747176e-08, "loss": 0.3687, "step": 1387 }, { "epoch": 0.8928916050176906, "grad_norm": 2.4821943000960935, "learning_rate": 5.9281754990359255e-08, "loss": 0.4037, "step": 1388 }, { "epoch": 0.893534898681248, "grad_norm": 2.3020034176499693, "learning_rate": 5.8576703588110953e-08, "loss": 0.4154, "step": 1389 }, { "epoch": 0.8941781923448054, "grad_norm": 2.268681079535142, "learning_rate": 5.787574345476298e-08, "loss": 0.355, "step": 1390 }, { "epoch": 0.8948214860083629, "grad_norm": 2.400381531469322, "learning_rate": 5.717887763657114e-08, "loss": 0.3494, "step": 1391 }, { "epoch": 0.8954647796719203, "grad_norm": 2.6620760278750515, "learning_rate": 5.6486109161998696e-08, "loss": 0.3931, "step": 1392 }, { "epoch": 0.8961080733354776, "grad_norm": 2.37720138899509, "learning_rate": 5.579744104170136e-08, "loss": 0.3841, "step": 1393 }, { "epoch": 0.896751366999035, "grad_norm": 2.6277253761547605, "learning_rate": 5.5112876268516795e-08, "loss": 0.3203, "step": 1394 }, { "epoch": 0.8973946606625924, "grad_norm": 2.683127702044441, "learning_rate": 5.443241781744923e-08, "loss": 0.3441, "step": 1395 }, { "epoch": 0.8980379543261499, "grad_norm": 2.4181832747717826, "learning_rate": 5.375606864565785e-08, "loss": 0.3397, "step": 1396 }, { "epoch": 0.8986812479897073, "grad_norm": 2.2295765456098318, "learning_rate": 5.308383169244346e-08, "loss": 0.3226, "step": 1397 }, { "epoch": 0.8993245416532647, "grad_norm": 2.672408877497257, "learning_rate": 5.241570987923616e-08, "loss": 0.3629, "step": 1398 }, { "epoch": 0.8999678353168221, "grad_norm": 2.7855168743241507, "learning_rate": 5.1751706109581795e-08, "loss": 0.4554, "step": 1399 }, { "epoch": 0.9006111289803795, "grad_norm": 2.334970101752676, "learning_rate": 5.109182326913053e-08, "loss": 0.3498, "step": 1400 }, { "epoch": 0.901254422643937, "grad_norm": 2.3854401403201764, "learning_rate": 5.043606422562308e-08, "loss": 0.3828, "step": 1401 }, { "epoch": 0.9018977163074944, "grad_norm": 2.1977862224221156, "learning_rate": 4.978443182887904e-08, "loss": 0.2837, "step": 1402 }, { "epoch": 0.9025410099710518, "grad_norm": 2.21149202368059, "learning_rate": 4.913692891078436e-08, "loss": 0.307, "step": 1403 }, { "epoch": 0.9031843036346092, "grad_norm": 2.6265249124758543, "learning_rate": 4.8493558285279236e-08, "loss": 0.3649, "step": 1404 }, { "epoch": 0.9038275972981666, "grad_norm": 2.461233914651311, "learning_rate": 4.7854322748344665e-08, "loss": 0.3395, "step": 1405 }, { "epoch": 0.9044708909617241, "grad_norm": 2.5589645827691485, "learning_rate": 4.7219225077992474e-08, "loss": 0.3709, "step": 1406 }, { "epoch": 0.9051141846252815, "grad_norm": 2.754121190959899, "learning_rate": 4.658826803425087e-08, "loss": 0.3619, "step": 1407 }, { "epoch": 0.9057574782888389, "grad_norm": 2.449121943705592, "learning_rate": 4.596145435915466e-08, "loss": 0.3235, "step": 1408 }, { "epoch": 0.9064007719523963, "grad_norm": 2.6464630081211378, "learning_rate": 4.533878677673131e-08, "loss": 0.4528, "step": 1409 }, { "epoch": 0.9070440656159536, "grad_norm": 2.4757900475661905, "learning_rate": 4.4720267992990444e-08, "loss": 0.3763, "step": 1410 }, { "epoch": 0.907687359279511, "grad_norm": 2.504961732908124, "learning_rate": 4.410590069591191e-08, "loss": 0.3277, "step": 1411 }, { "epoch": 0.9083306529430685, "grad_norm": 2.4886777829827977, "learning_rate": 4.3495687555433534e-08, "loss": 0.4323, "step": 1412 }, { "epoch": 0.9089739466066259, "grad_norm": 2.4019413824458966, "learning_rate": 4.288963122344025e-08, "loss": 0.3042, "step": 1413 }, { "epoch": 0.9096172402701833, "grad_norm": 2.562389106321339, "learning_rate": 4.2287734333751657e-08, "loss": 0.3969, "step": 1414 }, { "epoch": 0.9102605339337407, "grad_norm": 2.8226952921438047, "learning_rate": 4.168999950211183e-08, "loss": 0.4255, "step": 1415 }, { "epoch": 0.9109038275972982, "grad_norm": 2.298715152082212, "learning_rate": 4.1096429326176295e-08, "loss": 0.3544, "step": 1416 }, { "epoch": 0.9115471212608556, "grad_norm": 2.621482260205055, "learning_rate": 4.050702638550274e-08, "loss": 0.3443, "step": 1417 }, { "epoch": 0.912190414924413, "grad_norm": 2.5041209172386725, "learning_rate": 3.992179324153766e-08, "loss": 0.3541, "step": 1418 }, { "epoch": 0.9128337085879704, "grad_norm": 2.5909777403016, "learning_rate": 3.934073243760749e-08, "loss": 0.3591, "step": 1419 }, { "epoch": 0.9134770022515278, "grad_norm": 2.546485085430486, "learning_rate": 3.8763846498905185e-08, "loss": 0.263, "step": 1420 }, { "epoch": 0.9141202959150853, "grad_norm": 2.2367380729638646, "learning_rate": 3.819113793248141e-08, "loss": 0.2225, "step": 1421 }, { "epoch": 0.9147635895786427, "grad_norm": 2.3242789926101457, "learning_rate": 3.762260922723182e-08, "loss": 0.3052, "step": 1422 }, { "epoch": 0.9154068832422001, "grad_norm": 2.7417814914723264, "learning_rate": 3.7058262853887804e-08, "loss": 0.3942, "step": 1423 }, { "epoch": 0.9160501769057575, "grad_norm": 2.266512337013008, "learning_rate": 3.649810126500474e-08, "loss": 0.3207, "step": 1424 }, { "epoch": 0.9166934705693149, "grad_norm": 2.1551557075064727, "learning_rate": 3.5942126894951466e-08, "loss": 0.3932, "step": 1425 }, { "epoch": 0.9173367642328724, "grad_norm": 2.718918924378629, "learning_rate": 3.5390342159900224e-08, "loss": 0.3638, "step": 1426 }, { "epoch": 0.9179800578964297, "grad_norm": 2.3472482794173923, "learning_rate": 3.4842749457815404e-08, "loss": 0.3235, "step": 1427 }, { "epoch": 0.9186233515599871, "grad_norm": 3.3154597799629837, "learning_rate": 3.4299351168443736e-08, "loss": 0.3815, "step": 1428 }, { "epoch": 0.9192666452235445, "grad_norm": 2.6312348628681885, "learning_rate": 3.376014965330387e-08, "loss": 0.3699, "step": 1429 }, { "epoch": 0.9199099388871019, "grad_norm": 2.5428476818984547, "learning_rate": 3.322514725567571e-08, "loss": 0.3551, "step": 1430 }, { "epoch": 0.9205532325506594, "grad_norm": 1.9689640089890779, "learning_rate": 3.269434630059054e-08, "loss": 0.2899, "step": 1431 }, { "epoch": 0.9211965262142168, "grad_norm": 2.3385714117435388, "learning_rate": 3.216774909482101e-08, "loss": 0.3252, "step": 1432 }, { "epoch": 0.9218398198777742, "grad_norm": 2.502120398394241, "learning_rate": 3.164535792687095e-08, "loss": 0.3664, "step": 1433 }, { "epoch": 0.9224831135413316, "grad_norm": 2.155938344544481, "learning_rate": 3.112717506696527e-08, "loss": 0.3723, "step": 1434 }, { "epoch": 0.923126407204889, "grad_norm": 2.4970941226718413, "learning_rate": 3.061320276704049e-08, "loss": 0.3677, "step": 1435 }, { "epoch": 0.9237697008684465, "grad_norm": 2.4976632195121478, "learning_rate": 3.010344326073455e-08, "loss": 0.3214, "step": 1436 }, { "epoch": 0.9244129945320039, "grad_norm": 2.6610964535598804, "learning_rate": 2.9597898763377593e-08, "loss": 0.3858, "step": 1437 }, { "epoch": 0.9250562881955613, "grad_norm": 3.056081378506755, "learning_rate": 2.9096571471981636e-08, "loss": 0.4447, "step": 1438 }, { "epoch": 0.9256995818591187, "grad_norm": 2.4891747492857332, "learning_rate": 2.8599463565231685e-08, "loss": 0.3659, "step": 1439 }, { "epoch": 0.9263428755226761, "grad_norm": 2.534345200851124, "learning_rate": 2.8106577203475868e-08, "loss": 0.3459, "step": 1440 }, { "epoch": 0.9269861691862336, "grad_norm": 2.452836440953243, "learning_rate": 2.7617914528716313e-08, "loss": 0.3046, "step": 1441 }, { "epoch": 0.927629462849791, "grad_norm": 2.494973042330855, "learning_rate": 2.7133477664599837e-08, "loss": 0.4229, "step": 1442 }, { "epoch": 0.9282727565133484, "grad_norm": 2.679265643953974, "learning_rate": 2.665326871640794e-08, "loss": 0.3968, "step": 1443 }, { "epoch": 0.9289160501769057, "grad_norm": 2.827529285456862, "learning_rate": 2.617728977104927e-08, "loss": 0.3654, "step": 1444 }, { "epoch": 0.9295593438404631, "grad_norm": 2.2728190310714758, "learning_rate": 2.5705542897048626e-08, "loss": 0.3589, "step": 1445 }, { "epoch": 0.9302026375040205, "grad_norm": 2.4978842539284773, "learning_rate": 2.5238030144539734e-08, "loss": 0.4277, "step": 1446 }, { "epoch": 0.930845931167578, "grad_norm": 2.202402829025042, "learning_rate": 2.4774753545254933e-08, "loss": 0.3136, "step": 1447 }, { "epoch": 0.9314892248311354, "grad_norm": 2.7155003926956542, "learning_rate": 2.431571511251729e-08, "loss": 0.4169, "step": 1448 }, { "epoch": 0.9321325184946928, "grad_norm": 2.341003040553403, "learning_rate": 2.386091684123126e-08, "loss": 0.3971, "step": 1449 }, { "epoch": 0.9327758121582502, "grad_norm": 2.532778765210889, "learning_rate": 2.341036070787472e-08, "loss": 0.2815, "step": 1450 }, { "epoch": 0.9334191058218076, "grad_norm": 2.4034553419638125, "learning_rate": 2.296404867048929e-08, "loss": 0.3712, "step": 1451 }, { "epoch": 0.9340623994853651, "grad_norm": 2.3023501612778166, "learning_rate": 2.2521982668673e-08, "loss": 0.3245, "step": 1452 }, { "epoch": 0.9347056931489225, "grad_norm": 2.31176388878366, "learning_rate": 2.208416462357088e-08, "loss": 0.2917, "step": 1453 }, { "epoch": 0.9353489868124799, "grad_norm": 2.4452297190989767, "learning_rate": 2.1650596437867485e-08, "loss": 0.3445, "step": 1454 }, { "epoch": 0.9359922804760373, "grad_norm": 2.632015188392107, "learning_rate": 2.122127999577783e-08, "loss": 0.4337, "step": 1455 }, { "epoch": 0.9366355741395948, "grad_norm": 2.1318751324718628, "learning_rate": 2.0796217163039588e-08, "loss": 0.304, "step": 1456 }, { "epoch": 0.9372788678031522, "grad_norm": 2.3777946560190673, "learning_rate": 2.0375409786905328e-08, "loss": 0.3156, "step": 1457 }, { "epoch": 0.9379221614667096, "grad_norm": 2.657757233726682, "learning_rate": 1.9958859696133646e-08, "loss": 0.3627, "step": 1458 }, { "epoch": 0.938565455130267, "grad_norm": 2.5373124126691478, "learning_rate": 1.9546568700981926e-08, "loss": 0.3053, "step": 1459 }, { "epoch": 0.9392087487938244, "grad_norm": 2.8613668696368784, "learning_rate": 1.913853859319814e-08, "loss": 0.4259, "step": 1460 }, { "epoch": 0.9398520424573817, "grad_norm": 3.0569446135352445, "learning_rate": 1.8734771146013295e-08, "loss": 0.3313, "step": 1461 }, { "epoch": 0.9404953361209392, "grad_norm": 2.581912321941909, "learning_rate": 1.833526811413344e-08, "loss": 0.3519, "step": 1462 }, { "epoch": 0.9411386297844966, "grad_norm": 2.6029614702021084, "learning_rate": 1.794003123373222e-08, "loss": 0.3898, "step": 1463 }, { "epoch": 0.941781923448054, "grad_norm": 2.3044331543155243, "learning_rate": 1.7549062222443346e-08, "loss": 0.297, "step": 1464 }, { "epoch": 0.9424252171116114, "grad_norm": 2.96915764891259, "learning_rate": 1.716236277935301e-08, "loss": 0.3645, "step": 1465 }, { "epoch": 0.9430685107751688, "grad_norm": 2.5353276653388175, "learning_rate": 1.6779934584992718e-08, "loss": 0.3701, "step": 1466 }, { "epoch": 0.9437118044387263, "grad_norm": 2.479699796171879, "learning_rate": 1.6401779301331464e-08, "loss": 0.391, "step": 1467 }, { "epoch": 0.9443550981022837, "grad_norm": 2.6167212703474383, "learning_rate": 1.6027898571769448e-08, "loss": 0.4215, "step": 1468 }, { "epoch": 0.9449983917658411, "grad_norm": 2.4423314329736554, "learning_rate": 1.5658294021129946e-08, "loss": 0.3484, "step": 1469 }, { "epoch": 0.9456416854293985, "grad_norm": 2.298101404220221, "learning_rate": 1.5292967255652655e-08, "loss": 0.395, "step": 1470 }, { "epoch": 0.946284979092956, "grad_norm": 2.433485844361193, "learning_rate": 1.4931919862987142e-08, "loss": 0.3325, "step": 1471 }, { "epoch": 0.9469282727565134, "grad_norm": 2.352769762830418, "learning_rate": 1.4575153412185071e-08, "loss": 0.3529, "step": 1472 }, { "epoch": 0.9475715664200708, "grad_norm": 2.344857847361463, "learning_rate": 1.4222669453694214e-08, "loss": 0.3566, "step": 1473 }, { "epoch": 0.9482148600836282, "grad_norm": 2.3819859730199084, "learning_rate": 1.3874469519350896e-08, "loss": 0.3999, "step": 1474 }, { "epoch": 0.9488581537471856, "grad_norm": 2.5699721728023865, "learning_rate": 1.3530555122374332e-08, "loss": 0.423, "step": 1475 }, { "epoch": 0.949501447410743, "grad_norm": 2.3191929967674967, "learning_rate": 1.3190927757358972e-08, "loss": 0.3459, "step": 1476 }, { "epoch": 0.9501447410743005, "grad_norm": 2.999542546486663, "learning_rate": 1.2855588900269054e-08, "loss": 0.4068, "step": 1477 }, { "epoch": 0.9507880347378578, "grad_norm": 2.396038297706961, "learning_rate": 1.2524540008431173e-08, "loss": 0.4077, "step": 1478 }, { "epoch": 0.9514313284014152, "grad_norm": 2.3345489542226683, "learning_rate": 1.2197782520528832e-08, "loss": 0.3238, "step": 1479 }, { "epoch": 0.9520746220649726, "grad_norm": 2.337884345025424, "learning_rate": 1.1875317856595679e-08, "loss": 0.378, "step": 1480 }, { "epoch": 0.95271791572853, "grad_norm": 2.542203464591995, "learning_rate": 1.1557147418009394e-08, "loss": 0.2899, "step": 1481 }, { "epoch": 0.9533612093920875, "grad_norm": 2.1308185550096925, "learning_rate": 1.1243272587485809e-08, "loss": 0.331, "step": 1482 }, { "epoch": 0.9540045030556449, "grad_norm": 2.6461209397025267, "learning_rate": 1.0933694729072685e-08, "loss": 0.3485, "step": 1483 }, { "epoch": 0.9546477967192023, "grad_norm": 2.5094705439873812, "learning_rate": 1.0628415188143725e-08, "loss": 0.3381, "step": 1484 }, { "epoch": 0.9552910903827597, "grad_norm": 2.5864164373216325, "learning_rate": 1.0327435291393016e-08, "loss": 0.3576, "step": 1485 }, { "epoch": 0.9559343840463171, "grad_norm": 2.4416960058228505, "learning_rate": 1.003075634682915e-08, "loss": 0.3505, "step": 1486 }, { "epoch": 0.9565776777098746, "grad_norm": 2.4614691889623335, "learning_rate": 9.738379643769446e-09, "loss": 0.3804, "step": 1487 }, { "epoch": 0.957220971373432, "grad_norm": 2.448378638208511, "learning_rate": 9.450306452834178e-09, "loss": 0.3039, "step": 1488 }, { "epoch": 0.9578642650369894, "grad_norm": 2.2609240258981855, "learning_rate": 9.166538025941695e-09, "loss": 0.3677, "step": 1489 }, { "epoch": 0.9585075587005468, "grad_norm": 2.3421682171853244, "learning_rate": 8.887075596302196e-09, "loss": 0.3368, "step": 1490 }, { "epoch": 0.9591508523641042, "grad_norm": 2.0693467212752923, "learning_rate": 8.611920378412962e-09, "loss": 0.2619, "step": 1491 }, { "epoch": 0.9597941460276617, "grad_norm": 2.3663771573935066, "learning_rate": 8.341073568052803e-09, "loss": 0.3384, "step": 1492 }, { "epoch": 0.9604374396912191, "grad_norm": 2.5661602752194494, "learning_rate": 8.074536342276616e-09, "loss": 0.381, "step": 1493 }, { "epoch": 0.9610807333547765, "grad_norm": 2.0447678609307287, "learning_rate": 7.81230985941106e-09, "loss": 0.2703, "step": 1494 }, { "epoch": 0.9617240270183338, "grad_norm": 2.1622850389950723, "learning_rate": 7.554395259048663e-09, "loss": 0.3111, "step": 1495 }, { "epoch": 0.9623673206818912, "grad_norm": 2.48708117026487, "learning_rate": 7.300793662043281e-09, "loss": 0.3613, "step": 1496 }, { "epoch": 0.9630106143454487, "grad_norm": 2.456421058529263, "learning_rate": 7.051506170505317e-09, "loss": 0.3638, "step": 1497 }, { "epoch": 0.9636539080090061, "grad_norm": 2.4431112362381455, "learning_rate": 6.806533867796727e-09, "loss": 0.3848, "step": 1498 }, { "epoch": 0.9642972016725635, "grad_norm": 2.3495917600046385, "learning_rate": 6.565877818526244e-09, "loss": 0.3403, "step": 1499 }, { "epoch": 0.9649404953361209, "grad_norm": 2.106896812733821, "learning_rate": 6.329539068545053e-09, "loss": 0.3725, "step": 1500 }, { "epoch": 0.9655837889996783, "grad_norm": 2.1915041110352163, "learning_rate": 6.097518644942012e-09, "loss": 0.3431, "step": 1501 }, { "epoch": 0.9662270826632358, "grad_norm": 2.4651677327954227, "learning_rate": 5.869817556039325e-09, "loss": 0.4104, "step": 1502 }, { "epoch": 0.9668703763267932, "grad_norm": 2.1229139999195143, "learning_rate": 5.646436791387765e-09, "loss": 0.3257, "step": 1503 }, { "epoch": 0.9675136699903506, "grad_norm": 2.8920458607410797, "learning_rate": 5.427377321763238e-09, "loss": 0.3615, "step": 1504 }, { "epoch": 0.968156963653908, "grad_norm": 2.3661543809866457, "learning_rate": 5.2126400991615585e-09, "loss": 0.3641, "step": 1505 }, { "epoch": 0.9688002573174654, "grad_norm": 2.5782032250694575, "learning_rate": 5.0022260567951225e-09, "loss": 0.3657, "step": 1506 }, { "epoch": 0.9694435509810229, "grad_norm": 2.7823601159390945, "learning_rate": 4.796136109088022e-09, "loss": 0.4049, "step": 1507 }, { "epoch": 0.9700868446445803, "grad_norm": 2.5671463093974523, "learning_rate": 4.5943711516730486e-09, "loss": 0.3661, "step": 1508 }, { "epoch": 0.9707301383081377, "grad_norm": 2.2033150771035928, "learning_rate": 4.396932061387137e-09, "loss": 0.3052, "step": 1509 }, { "epoch": 0.9713734319716951, "grad_norm": 2.5644736204369463, "learning_rate": 4.203819696267485e-09, "loss": 0.3821, "step": 1510 }, { "epoch": 0.9720167256352525, "grad_norm": 2.3702024367667396, "learning_rate": 4.0150348955486635e-09, "loss": 0.4065, "step": 1511 }, { "epoch": 0.9726600192988099, "grad_norm": 2.928661513024743, "learning_rate": 3.8305784796576205e-09, "loss": 0.3341, "step": 1512 }, { "epoch": 0.9733033129623673, "grad_norm": 2.558664645185684, "learning_rate": 3.6504512502113504e-09, "loss": 0.3325, "step": 1513 }, { "epoch": 0.9739466066259247, "grad_norm": 2.2295352443079244, "learning_rate": 3.474653990012899e-09, "loss": 0.3342, "step": 1514 }, { "epoch": 0.9745899002894821, "grad_norm": 2.379974868962145, "learning_rate": 3.3031874630478072e-09, "loss": 0.3341, "step": 1515 }, { "epoch": 0.9752331939530395, "grad_norm": 2.5565945574036113, "learning_rate": 3.1360524144810052e-09, "loss": 0.3987, "step": 1516 }, { "epoch": 0.975876487616597, "grad_norm": 2.6507293917037233, "learning_rate": 2.9732495706535908e-09, "loss": 0.4507, "step": 1517 }, { "epoch": 0.9765197812801544, "grad_norm": 2.195874312298709, "learning_rate": 2.814779639079723e-09, "loss": 0.3066, "step": 1518 }, { "epoch": 0.9771630749437118, "grad_norm": 2.6481444281382593, "learning_rate": 2.6606433084432888e-09, "loss": 0.3979, "step": 1519 }, { "epoch": 0.9778063686072692, "grad_norm": 2.248983235365829, "learning_rate": 2.51084124859513e-09, "loss": 0.2798, "step": 1520 }, { "epoch": 0.9784496622708266, "grad_norm": 2.4300727721829465, "learning_rate": 2.3653741105499336e-09, "loss": 0.3355, "step": 1521 }, { "epoch": 0.9790929559343841, "grad_norm": 2.542608628173163, "learning_rate": 2.2242425264838993e-09, "loss": 0.3364, "step": 1522 }, { "epoch": 0.9797362495979415, "grad_norm": 2.4270255242620014, "learning_rate": 2.0874471097311886e-09, "loss": 0.2992, "step": 1523 }, { "epoch": 0.9803795432614989, "grad_norm": 2.448464374455245, "learning_rate": 1.954988454781925e-09, "loss": 0.4015, "step": 1524 }, { "epoch": 0.9810228369250563, "grad_norm": 2.1983110207485796, "learning_rate": 1.8268671372794197e-09, "loss": 0.3048, "step": 1525 }, { "epoch": 0.9816661305886137, "grad_norm": 2.3857614549975614, "learning_rate": 1.703083714017617e-09, "loss": 0.3192, "step": 1526 }, { "epoch": 0.9823094242521712, "grad_norm": 2.2032052052718223, "learning_rate": 1.5836387229386516e-09, "loss": 0.3456, "step": 1527 }, { "epoch": 0.9829527179157286, "grad_norm": 2.67205810133671, "learning_rate": 1.4685326831305188e-09, "loss": 0.3891, "step": 1528 }, { "epoch": 0.9835960115792859, "grad_norm": 2.6349567082550647, "learning_rate": 1.3577660948249637e-09, "loss": 0.3035, "step": 1529 }, { "epoch": 0.9842393052428433, "grad_norm": 2.2480960628053586, "learning_rate": 1.2513394393950383e-09, "loss": 0.3643, "step": 1530 }, { "epoch": 0.9848825989064007, "grad_norm": 2.4796471401154134, "learning_rate": 1.1492531793534376e-09, "loss": 0.3089, "step": 1531 }, { "epoch": 0.9855258925699582, "grad_norm": 2.9103039325919626, "learning_rate": 1.0515077583498344e-09, "loss": 0.3791, "step": 1532 }, { "epoch": 0.9861691862335156, "grad_norm": 2.4677965930731585, "learning_rate": 9.581036011696575e-10, "loss": 0.3656, "step": 1533 }, { "epoch": 0.986812479897073, "grad_norm": 2.445564088110857, "learning_rate": 8.690411137318721e-10, "loss": 0.3305, "step": 1534 }, { "epoch": 0.9874557735606304, "grad_norm": 2.382794849759407, "learning_rate": 7.843206830872029e-10, "loss": 0.3426, "step": 1535 }, { "epoch": 0.9880990672241878, "grad_norm": 2.6533182251946315, "learning_rate": 7.039426774164692e-10, "loss": 0.415, "step": 1536 }, { "epoch": 0.9887423608877453, "grad_norm": 2.420564339694528, "learning_rate": 6.279074460291412e-10, "loss": 0.3941, "step": 1537 }, { "epoch": 0.9893856545513027, "grad_norm": 2.639473830993184, "learning_rate": 5.562153193616748e-10, "loss": 0.3448, "step": 1538 }, { "epoch": 0.9900289482148601, "grad_norm": 2.5604854688293894, "learning_rate": 4.888666089761795e-10, "loss": 0.4005, "step": 1539 }, { "epoch": 0.9906722418784175, "grad_norm": 2.4798427341261164, "learning_rate": 4.258616075588639e-10, "loss": 0.2729, "step": 1540 }, { "epoch": 0.9913155355419749, "grad_norm": 2.444149523508549, "learning_rate": 3.672005889189256e-10, "loss": 0.3646, "step": 1541 }, { "epoch": 0.9919588292055324, "grad_norm": 2.6380752558173004, "learning_rate": 3.1288380798744075e-10, "loss": 0.4269, "step": 1542 }, { "epoch": 0.9926021228690898, "grad_norm": 2.5646239143178273, "learning_rate": 2.6291150081603207e-10, "loss": 0.4055, "step": 1543 }, { "epoch": 0.9932454165326472, "grad_norm": 2.280393009300122, "learning_rate": 2.172838845762026e-10, "loss": 0.3801, "step": 1544 }, { "epoch": 0.9938887101962046, "grad_norm": 2.346487099561757, "learning_rate": 1.7600115755789235e-10, "loss": 0.3614, "step": 1545 }, { "epoch": 0.9945320038597619, "grad_norm": 2.4610388607360107, "learning_rate": 1.3906349916881223e-10, "loss": 0.3243, "step": 1546 }, { "epoch": 0.9951752975233193, "grad_norm": 2.623411934848152, "learning_rate": 1.0647106993411092e-10, "loss": 0.3832, "step": 1547 }, { "epoch": 0.9958185911868768, "grad_norm": 2.5680420614493324, "learning_rate": 7.822401149482072e-11, "loss": 0.3653, "step": 1548 }, { "epoch": 0.9964618848504342, "grad_norm": 2.707614778116817, "learning_rate": 5.432244660819041e-11, "loss": 0.3361, "step": 1549 }, { "epoch": 0.9971051785139916, "grad_norm": 2.3157263284161376, "learning_rate": 3.476647914624209e-11, "loss": 0.3716, "step": 1550 }, { "epoch": 0.997748472177549, "grad_norm": 2.671035360332632, "learning_rate": 1.9556194096104207e-11, "loss": 0.3843, "step": 1551 }, { "epoch": 0.9983917658411064, "grad_norm": 2.5604965920648377, "learning_rate": 8.691657559012356e-12, "loss": 0.3623, "step": 1552 }, { "epoch": 0.9990350595046639, "grad_norm": 2.5761358324118846, "learning_rate": 2.1729167505313284e-12, "loss": 0.3449, "step": 1553 }, { "epoch": 0.9996783531682213, "grad_norm": 2.431437005973932, "learning_rate": 0.0, "loss": 0.2982, "step": 1554 }, { "epoch": 0.9996783531682213, "step": 1554, "total_flos": 1185522032279552.0, "train_loss": 0.39366521815765776, "train_runtime": 71310.9295, "train_samples_per_second": 1.395, "train_steps_per_second": 0.022 } ], "logging_steps": 1.0, "max_steps": 1554, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1185522032279552.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }