diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,5017 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5781153767193594, + "eval_steps": 88, + "global_step": 704, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.000821186614658181, + "grad_norm": 0.480191707611084, + "learning_rate": 2e-05, + "loss": 1.332, + "step": 1 + }, + { + "epoch": 0.001642373229316362, + "grad_norm": 0.43048733472824097, + "learning_rate": 4e-05, + "loss": 1.2784, + "step": 2 + }, + { + "epoch": 0.002463559843974543, + "grad_norm": 0.33739173412323, + "learning_rate": 6e-05, + "loss": 1.3286, + "step": 3 + }, + { + "epoch": 0.003284746458632724, + "grad_norm": 0.432579904794693, + "learning_rate": 8e-05, + "loss": 1.3079, + "step": 4 + }, + { + "epoch": 0.0041059330732909054, + "grad_norm": 0.3490436375141144, + "learning_rate": 0.0001, + "loss": 1.2182, + "step": 5 + }, + { + "epoch": 0.004927119687949086, + "grad_norm": 0.20206260681152344, + "learning_rate": 9.997257268239166e-05, + "loss": 1.2828, + "step": 6 + }, + { + "epoch": 0.005748306302607267, + "grad_norm": 0.15355628728866577, + "learning_rate": 9.994514536478333e-05, + "loss": 1.1616, + "step": 7 + }, + { + "epoch": 0.006569492917265448, + "grad_norm": 0.16756780445575714, + "learning_rate": 9.9917718047175e-05, + "loss": 1.0935, + "step": 8 + }, + { + "epoch": 0.00739067953192363, + "grad_norm": 0.17429664731025696, + "learning_rate": 9.989029072956665e-05, + "loss": 0.9694, + "step": 9 + }, + { + "epoch": 0.008211866146581811, + "grad_norm": 0.22355175018310547, + "learning_rate": 9.986286341195832e-05, + "loss": 0.9407, + "step": 10 + }, + { + "epoch": 0.009033052761239991, + "grad_norm": 0.33407703042030334, + "learning_rate": 9.983543609434997e-05, + "loss": 0.7717, + "step": 11 + }, + { + "epoch": 0.009854239375898173, + "grad_norm": 0.47473278641700745, + "learning_rate": 9.980800877674164e-05, + "loss": 0.7763, + "step": 12 + }, + { + "epoch": 0.010675425990556354, + "grad_norm": 0.2812059819698334, + "learning_rate": 9.978058145913331e-05, + "loss": 0.6258, + "step": 13 + }, + { + "epoch": 0.011496612605214535, + "grad_norm": 0.23547925055027008, + "learning_rate": 9.975315414152496e-05, + "loss": 0.5968, + "step": 14 + }, + { + "epoch": 0.012317799219872716, + "grad_norm": 0.18453630805015564, + "learning_rate": 9.972572682391662e-05, + "loss": 0.5368, + "step": 15 + }, + { + "epoch": 0.013138985834530896, + "grad_norm": 0.16103577613830566, + "learning_rate": 9.969829950630828e-05, + "loss": 0.4909, + "step": 16 + }, + { + "epoch": 0.013960172449189078, + "grad_norm": 0.18651455640792847, + "learning_rate": 9.967087218869995e-05, + "loss": 0.5135, + "step": 17 + }, + { + "epoch": 0.01478135906384726, + "grad_norm": 0.11300642043352127, + "learning_rate": 9.96434448710916e-05, + "loss": 0.5069, + "step": 18 + }, + { + "epoch": 0.01560254567850544, + "grad_norm": 0.10415703803300858, + "learning_rate": 9.961601755348327e-05, + "loss": 0.4851, + "step": 19 + }, + { + "epoch": 0.016423732293163622, + "grad_norm": 0.11693017929792404, + "learning_rate": 9.958859023587493e-05, + "loss": 0.4625, + "step": 20 + }, + { + "epoch": 0.017244918907821802, + "grad_norm": 0.10035043954849243, + "learning_rate": 9.95611629182666e-05, + "loss": 0.4822, + "step": 21 + }, + { + "epoch": 0.018066105522479982, + "grad_norm": 0.10483390837907791, + "learning_rate": 9.953373560065826e-05, + "loss": 0.4342, + "step": 22 + }, + { + "epoch": 0.018887292137138165, + "grad_norm": 2.8405802249908447, + "learning_rate": 9.950630828304992e-05, + "loss": 0.4664, + "step": 23 + }, + { + "epoch": 0.019708478751796345, + "grad_norm": 0.13821998238563538, + "learning_rate": 9.947888096544159e-05, + "loss": 0.4468, + "step": 24 + }, + { + "epoch": 0.020529665366454525, + "grad_norm": 0.1991378366947174, + "learning_rate": 9.945145364783325e-05, + "loss": 0.4605, + "step": 25 + }, + { + "epoch": 0.02135085198111271, + "grad_norm": 0.07619134336709976, + "learning_rate": 9.942402633022491e-05, + "loss": 0.4597, + "step": 26 + }, + { + "epoch": 0.02217203859577089, + "grad_norm": 0.13373583555221558, + "learning_rate": 9.939659901261658e-05, + "loss": 0.4626, + "step": 27 + }, + { + "epoch": 0.02299322521042907, + "grad_norm": 0.09962721168994904, + "learning_rate": 9.936917169500823e-05, + "loss": 0.4638, + "step": 28 + }, + { + "epoch": 0.023814411825087253, + "grad_norm": 0.09395964443683624, + "learning_rate": 9.93417443773999e-05, + "loss": 0.4569, + "step": 29 + }, + { + "epoch": 0.024635598439745433, + "grad_norm": 0.09109952300786972, + "learning_rate": 9.931431705979157e-05, + "loss": 0.4439, + "step": 30 + }, + { + "epoch": 0.025456785054403613, + "grad_norm": 0.10370515286922455, + "learning_rate": 9.928688974218322e-05, + "loss": 0.4425, + "step": 31 + }, + { + "epoch": 0.026277971669061793, + "grad_norm": 0.2153477966785431, + "learning_rate": 9.925946242457488e-05, + "loss": 0.4503, + "step": 32 + }, + { + "epoch": 0.027099158283719976, + "grad_norm": 0.08772841095924377, + "learning_rate": 9.923203510696654e-05, + "loss": 0.419, + "step": 33 + }, + { + "epoch": 0.027920344898378156, + "grad_norm": 0.10951374471187592, + "learning_rate": 9.920460778935821e-05, + "loss": 0.4353, + "step": 34 + }, + { + "epoch": 0.028741531513036336, + "grad_norm": 0.09190870076417923, + "learning_rate": 9.917718047174987e-05, + "loss": 0.5196, + "step": 35 + }, + { + "epoch": 0.02956271812769452, + "grad_norm": 0.07667124271392822, + "learning_rate": 9.914975315414153e-05, + "loss": 0.4358, + "step": 36 + }, + { + "epoch": 0.0303839047423527, + "grad_norm": 0.1514267474412918, + "learning_rate": 9.912232583653319e-05, + "loss": 0.411, + "step": 37 + }, + { + "epoch": 0.03120509135701088, + "grad_norm": 0.09086549282073975, + "learning_rate": 9.909489851892486e-05, + "loss": 0.4003, + "step": 38 + }, + { + "epoch": 0.032026277971669063, + "grad_norm": 0.2616782486438751, + "learning_rate": 9.906747120131652e-05, + "loss": 0.4842, + "step": 39 + }, + { + "epoch": 0.032847464586327244, + "grad_norm": 0.11908283084630966, + "learning_rate": 9.904004388370818e-05, + "loss": 0.4143, + "step": 40 + }, + { + "epoch": 0.033668651200985424, + "grad_norm": 0.07770542800426483, + "learning_rate": 9.901261656609983e-05, + "loss": 0.3873, + "step": 41 + }, + { + "epoch": 0.034489837815643604, + "grad_norm": 0.08934606611728668, + "learning_rate": 9.89851892484915e-05, + "loss": 0.4235, + "step": 42 + }, + { + "epoch": 0.035311024430301784, + "grad_norm": 0.09303563088178635, + "learning_rate": 9.895776193088317e-05, + "loss": 0.4103, + "step": 43 + }, + { + "epoch": 0.036132211044959964, + "grad_norm": 0.08622181415557861, + "learning_rate": 9.893033461327482e-05, + "loss": 0.448, + "step": 44 + }, + { + "epoch": 0.03695339765961815, + "grad_norm": 0.08822862058877945, + "learning_rate": 9.890290729566649e-05, + "loss": 0.3855, + "step": 45 + }, + { + "epoch": 0.03777458427427633, + "grad_norm": 0.08557698875665665, + "learning_rate": 9.887547997805814e-05, + "loss": 0.3945, + "step": 46 + }, + { + "epoch": 0.03859577088893451, + "grad_norm": 0.07540106773376465, + "learning_rate": 9.884805266044981e-05, + "loss": 0.4192, + "step": 47 + }, + { + "epoch": 0.03941695750359269, + "grad_norm": 0.1023702397942543, + "learning_rate": 9.882062534284148e-05, + "loss": 0.4126, + "step": 48 + }, + { + "epoch": 0.04023814411825087, + "grad_norm": 0.07779772579669952, + "learning_rate": 9.879319802523313e-05, + "loss": 0.4244, + "step": 49 + }, + { + "epoch": 0.04105933073290905, + "grad_norm": 0.08826564252376556, + "learning_rate": 9.876577070762479e-05, + "loss": 0.415, + "step": 50 + }, + { + "epoch": 0.04188051734756724, + "grad_norm": 0.08254576474428177, + "learning_rate": 9.873834339001646e-05, + "loss": 0.4346, + "step": 51 + }, + { + "epoch": 0.04270170396222542, + "grad_norm": 0.08287151902914047, + "learning_rate": 9.871091607240812e-05, + "loss": 0.4142, + "step": 52 + }, + { + "epoch": 0.0435228905768836, + "grad_norm": 0.08196476101875305, + "learning_rate": 9.868348875479978e-05, + "loss": 0.3822, + "step": 53 + }, + { + "epoch": 0.04434407719154178, + "grad_norm": 0.08654092252254486, + "learning_rate": 9.865606143719145e-05, + "loss": 0.3937, + "step": 54 + }, + { + "epoch": 0.04516526380619996, + "grad_norm": 0.1102684736251831, + "learning_rate": 9.86286341195831e-05, + "loss": 0.4508, + "step": 55 + }, + { + "epoch": 0.04598645042085814, + "grad_norm": 0.08240954577922821, + "learning_rate": 9.860120680197478e-05, + "loss": 0.4069, + "step": 56 + }, + { + "epoch": 0.04680763703551632, + "grad_norm": 0.08542217314243317, + "learning_rate": 9.857377948436644e-05, + "loss": 0.5002, + "step": 57 + }, + { + "epoch": 0.047628823650174505, + "grad_norm": 0.08390172570943832, + "learning_rate": 9.854635216675809e-05, + "loss": 0.3851, + "step": 58 + }, + { + "epoch": 0.048450010264832685, + "grad_norm": 0.10694168508052826, + "learning_rate": 9.851892484914976e-05, + "loss": 0.4026, + "step": 59 + }, + { + "epoch": 0.049271196879490865, + "grad_norm": 0.0852806493639946, + "learning_rate": 9.849149753154143e-05, + "loss": 0.424, + "step": 60 + }, + { + "epoch": 0.050092383494149045, + "grad_norm": 0.15425831079483032, + "learning_rate": 9.846407021393308e-05, + "loss": 0.4307, + "step": 61 + }, + { + "epoch": 0.050913570108807225, + "grad_norm": 0.08546218276023865, + "learning_rate": 9.843664289632475e-05, + "loss": 0.4386, + "step": 62 + }, + { + "epoch": 0.051734756723465405, + "grad_norm": 0.08588795363903046, + "learning_rate": 9.84092155787164e-05, + "loss": 0.4433, + "step": 63 + }, + { + "epoch": 0.052555943338123585, + "grad_norm": 0.08901514858007431, + "learning_rate": 9.838178826110807e-05, + "loss": 0.4263, + "step": 64 + }, + { + "epoch": 0.05337712995278177, + "grad_norm": 0.17668095231056213, + "learning_rate": 9.835436094349974e-05, + "loss": 0.407, + "step": 65 + }, + { + "epoch": 0.05419831656743995, + "grad_norm": 0.081763856112957, + "learning_rate": 9.83269336258914e-05, + "loss": 0.374, + "step": 66 + }, + { + "epoch": 0.05501950318209813, + "grad_norm": 0.09063572436571121, + "learning_rate": 9.829950630828305e-05, + "loss": 0.3839, + "step": 67 + }, + { + "epoch": 0.05584068979675631, + "grad_norm": 0.08264392614364624, + "learning_rate": 9.827207899067472e-05, + "loss": 0.4676, + "step": 68 + }, + { + "epoch": 0.05666187641141449, + "grad_norm": 0.0827123150229454, + "learning_rate": 9.824465167306638e-05, + "loss": 0.3801, + "step": 69 + }, + { + "epoch": 0.05748306302607267, + "grad_norm": 0.07972189038991928, + "learning_rate": 9.821722435545804e-05, + "loss": 0.3758, + "step": 70 + }, + { + "epoch": 0.05830424964073085, + "grad_norm": 0.08738942444324493, + "learning_rate": 9.81897970378497e-05, + "loss": 0.3634, + "step": 71 + }, + { + "epoch": 0.05912543625538904, + "grad_norm": 0.08442792296409607, + "learning_rate": 9.816236972024136e-05, + "loss": 0.3803, + "step": 72 + }, + { + "epoch": 0.05994662287004722, + "grad_norm": 0.2751137316226959, + "learning_rate": 9.813494240263303e-05, + "loss": 0.3869, + "step": 73 + }, + { + "epoch": 0.0607678094847054, + "grad_norm": 0.19429296255111694, + "learning_rate": 9.81075150850247e-05, + "loss": 0.3992, + "step": 74 + }, + { + "epoch": 0.06158899609936358, + "grad_norm": 0.08392605930566788, + "learning_rate": 9.808008776741635e-05, + "loss": 0.361, + "step": 75 + }, + { + "epoch": 0.06241018271402176, + "grad_norm": 2.01043963432312, + "learning_rate": 9.8052660449808e-05, + "loss": 0.3787, + "step": 76 + }, + { + "epoch": 0.06323136932867994, + "grad_norm": 0.0808538943529129, + "learning_rate": 9.802523313219967e-05, + "loss": 0.4013, + "step": 77 + }, + { + "epoch": 0.06405255594333813, + "grad_norm": 0.07846518605947495, + "learning_rate": 9.799780581459134e-05, + "loss": 0.3808, + "step": 78 + }, + { + "epoch": 0.0648737425579963, + "grad_norm": 0.08152970671653748, + "learning_rate": 9.7970378496983e-05, + "loss": 0.4076, + "step": 79 + }, + { + "epoch": 0.06569492917265449, + "grad_norm": 0.07745791226625443, + "learning_rate": 9.794295117937466e-05, + "loss": 0.3981, + "step": 80 + }, + { + "epoch": 0.06651611578731266, + "grad_norm": 0.0842173770070076, + "learning_rate": 9.791552386176632e-05, + "loss": 0.4469, + "step": 81 + }, + { + "epoch": 0.06733730240197085, + "grad_norm": 0.07920663058757782, + "learning_rate": 9.788809654415799e-05, + "loss": 0.3988, + "step": 82 + }, + { + "epoch": 0.06815848901662903, + "grad_norm": 0.07899456471204758, + "learning_rate": 9.786066922654965e-05, + "loss": 0.4418, + "step": 83 + }, + { + "epoch": 0.06897967563128721, + "grad_norm": 0.07229727506637573, + "learning_rate": 9.783324190894131e-05, + "loss": 0.4016, + "step": 84 + }, + { + "epoch": 0.0698008622459454, + "grad_norm": 0.08062436431646347, + "learning_rate": 9.780581459133296e-05, + "loss": 0.3762, + "step": 85 + }, + { + "epoch": 0.07062204886060357, + "grad_norm": 0.08440960198640823, + "learning_rate": 9.777838727372464e-05, + "loss": 0.4362, + "step": 86 + }, + { + "epoch": 0.07144323547526175, + "grad_norm": 0.07544733583927155, + "learning_rate": 9.77509599561163e-05, + "loss": 0.4135, + "step": 87 + }, + { + "epoch": 0.07226442208991993, + "grad_norm": 0.1760656237602234, + "learning_rate": 9.772353263850797e-05, + "loss": 0.3887, + "step": 88 + }, + { + "epoch": 0.07226442208991993, + "eval_runtime": 494.3052, + "eval_samples_per_second": 0.399, + "eval_steps_per_second": 0.2, + "step": 88 + }, + { + "epoch": 0.07308560870457811, + "grad_norm": 0.08049104362726212, + "learning_rate": 9.769610532089962e-05, + "loss": 0.4272, + "step": 89 + }, + { + "epoch": 0.0739067953192363, + "grad_norm": 0.07617965340614319, + "learning_rate": 9.766867800329129e-05, + "loss": 0.3873, + "step": 90 + }, + { + "epoch": 0.07472798193389447, + "grad_norm": 0.07975462824106216, + "learning_rate": 9.764125068568296e-05, + "loss": 0.3903, + "step": 91 + }, + { + "epoch": 0.07554916854855266, + "grad_norm": 0.08299189805984497, + "learning_rate": 9.761382336807461e-05, + "loss": 0.404, + "step": 92 + }, + { + "epoch": 0.07637035516321083, + "grad_norm": 0.08578819036483765, + "learning_rate": 9.758639605046626e-05, + "loss": 0.3983, + "step": 93 + }, + { + "epoch": 0.07719154177786902, + "grad_norm": 0.10872071981430054, + "learning_rate": 9.755896873285793e-05, + "loss": 0.3553, + "step": 94 + }, + { + "epoch": 0.07801272839252721, + "grad_norm": 0.08121436834335327, + "learning_rate": 9.75315414152496e-05, + "loss": 0.408, + "step": 95 + }, + { + "epoch": 0.07883391500718538, + "grad_norm": 0.09175996482372284, + "learning_rate": 9.750411409764125e-05, + "loss": 0.4008, + "step": 96 + }, + { + "epoch": 0.07965510162184357, + "grad_norm": 0.08122789114713669, + "learning_rate": 9.747668678003292e-05, + "loss": 0.4045, + "step": 97 + }, + { + "epoch": 0.08047628823650174, + "grad_norm": 0.08402436226606369, + "learning_rate": 9.744925946242458e-05, + "loss": 0.3814, + "step": 98 + }, + { + "epoch": 0.08129747485115993, + "grad_norm": 0.08454861491918564, + "learning_rate": 9.742183214481624e-05, + "loss": 0.3904, + "step": 99 + }, + { + "epoch": 0.0821186614658181, + "grad_norm": 0.08125888556241989, + "learning_rate": 9.739440482720791e-05, + "loss": 0.3681, + "step": 100 + }, + { + "epoch": 0.08293984808047629, + "grad_norm": 0.08544078469276428, + "learning_rate": 9.736697750959957e-05, + "loss": 0.367, + "step": 101 + }, + { + "epoch": 0.08376103469513448, + "grad_norm": 0.07539089769124985, + "learning_rate": 9.733955019199122e-05, + "loss": 0.3821, + "step": 102 + }, + { + "epoch": 0.08458222130979265, + "grad_norm": 0.07529085874557495, + "learning_rate": 9.731212287438289e-05, + "loss": 0.4169, + "step": 103 + }, + { + "epoch": 0.08540340792445084, + "grad_norm": 0.07588034868240356, + "learning_rate": 9.728469555677456e-05, + "loss": 0.4211, + "step": 104 + }, + { + "epoch": 0.08622459453910901, + "grad_norm": 0.08019097149372101, + "learning_rate": 9.725726823916621e-05, + "loss": 0.4033, + "step": 105 + }, + { + "epoch": 0.0870457811537672, + "grad_norm": 0.07878712564706802, + "learning_rate": 9.722984092155788e-05, + "loss": 0.3842, + "step": 106 + }, + { + "epoch": 0.08786696776842537, + "grad_norm": 0.08083963394165039, + "learning_rate": 9.720241360394953e-05, + "loss": 0.3558, + "step": 107 + }, + { + "epoch": 0.08868815438308356, + "grad_norm": 0.08340411633253098, + "learning_rate": 9.71749862863412e-05, + "loss": 0.4037, + "step": 108 + }, + { + "epoch": 0.08950934099774174, + "grad_norm": 0.09317754209041595, + "learning_rate": 9.714755896873287e-05, + "loss": 0.3874, + "step": 109 + }, + { + "epoch": 0.09033052761239992, + "grad_norm": 0.07143125683069229, + "learning_rate": 9.712013165112452e-05, + "loss": 0.3441, + "step": 110 + }, + { + "epoch": 0.0911517142270581, + "grad_norm": 0.0909111350774765, + "learning_rate": 9.709270433351618e-05, + "loss": 0.3552, + "step": 111 + }, + { + "epoch": 0.09197290084171628, + "grad_norm": 0.08041603863239288, + "learning_rate": 9.706527701590785e-05, + "loss": 0.394, + "step": 112 + }, + { + "epoch": 0.09279408745637446, + "grad_norm": 0.08147992193698883, + "learning_rate": 9.703784969829951e-05, + "loss": 0.3916, + "step": 113 + }, + { + "epoch": 0.09361527407103264, + "grad_norm": 0.08365318924188614, + "learning_rate": 9.701042238069117e-05, + "loss": 0.4075, + "step": 114 + }, + { + "epoch": 0.09443646068569082, + "grad_norm": 0.083246149122715, + "learning_rate": 9.698299506308284e-05, + "loss": 0.3566, + "step": 115 + }, + { + "epoch": 0.09525764730034901, + "grad_norm": 0.0942649245262146, + "learning_rate": 9.695556774547449e-05, + "loss": 0.3516, + "step": 116 + }, + { + "epoch": 0.09607883391500718, + "grad_norm": 0.08110091835260391, + "learning_rate": 9.692814042786616e-05, + "loss": 0.3902, + "step": 117 + }, + { + "epoch": 0.09690002052966537, + "grad_norm": 0.43124014139175415, + "learning_rate": 9.690071311025783e-05, + "loss": 0.3956, + "step": 118 + }, + { + "epoch": 0.09772120714432354, + "grad_norm": 0.08204706013202667, + "learning_rate": 9.687328579264948e-05, + "loss": 0.3865, + "step": 119 + }, + { + "epoch": 0.09854239375898173, + "grad_norm": 0.08262762427330017, + "learning_rate": 9.684585847504113e-05, + "loss": 0.3777, + "step": 120 + }, + { + "epoch": 0.0993635803736399, + "grad_norm": 0.08233962953090668, + "learning_rate": 9.681843115743282e-05, + "loss": 0.3976, + "step": 121 + }, + { + "epoch": 0.10018476698829809, + "grad_norm": 0.7428120970726013, + "learning_rate": 9.679100383982447e-05, + "loss": 0.377, + "step": 122 + }, + { + "epoch": 0.10100595360295628, + "grad_norm": 0.07909400761127472, + "learning_rate": 9.676357652221614e-05, + "loss": 0.3654, + "step": 123 + }, + { + "epoch": 0.10182714021761445, + "grad_norm": 0.08714035898447037, + "learning_rate": 9.673614920460779e-05, + "loss": 0.4181, + "step": 124 + }, + { + "epoch": 0.10264832683227264, + "grad_norm": 0.08017311245203018, + "learning_rate": 9.670872188699946e-05, + "loss": 0.4052, + "step": 125 + }, + { + "epoch": 0.10346951344693081, + "grad_norm": 0.13821078836917877, + "learning_rate": 9.668129456939113e-05, + "loss": 0.3733, + "step": 126 + }, + { + "epoch": 0.104290700061589, + "grad_norm": 0.1609969437122345, + "learning_rate": 9.665386725178278e-05, + "loss": 0.3686, + "step": 127 + }, + { + "epoch": 0.10511188667624717, + "grad_norm": 0.1280309557914734, + "learning_rate": 9.662643993417444e-05, + "loss": 0.4036, + "step": 128 + }, + { + "epoch": 0.10593307329090536, + "grad_norm": 0.08747898787260056, + "learning_rate": 9.65990126165661e-05, + "loss": 0.3587, + "step": 129 + }, + { + "epoch": 0.10675425990556354, + "grad_norm": 0.1252209097146988, + "learning_rate": 9.657158529895777e-05, + "loss": 0.3626, + "step": 130 + }, + { + "epoch": 0.10757544652022172, + "grad_norm": 0.09374388307332993, + "learning_rate": 9.654415798134943e-05, + "loss": 0.3824, + "step": 131 + }, + { + "epoch": 0.1083966331348799, + "grad_norm": 0.08730709552764893, + "learning_rate": 9.65167306637411e-05, + "loss": 0.3827, + "step": 132 + }, + { + "epoch": 0.10921781974953808, + "grad_norm": 0.07719024270772934, + "learning_rate": 9.648930334613275e-05, + "loss": 0.3762, + "step": 133 + }, + { + "epoch": 0.11003900636419627, + "grad_norm": 0.08843278139829636, + "learning_rate": 9.646187602852442e-05, + "loss": 0.3843, + "step": 134 + }, + { + "epoch": 0.11086019297885444, + "grad_norm": 0.08568207919597626, + "learning_rate": 9.643444871091608e-05, + "loss": 0.3594, + "step": 135 + }, + { + "epoch": 0.11168137959351263, + "grad_norm": 0.08556952327489853, + "learning_rate": 9.640702139330774e-05, + "loss": 0.3886, + "step": 136 + }, + { + "epoch": 0.11250256620817081, + "grad_norm": 0.09016801416873932, + "learning_rate": 9.63795940756994e-05, + "loss": 0.4023, + "step": 137 + }, + { + "epoch": 0.11332375282282899, + "grad_norm": 0.08133590966463089, + "learning_rate": 9.635216675809106e-05, + "loss": 0.3634, + "step": 138 + }, + { + "epoch": 0.11414493943748717, + "grad_norm": 0.2211730182170868, + "learning_rate": 9.632473944048273e-05, + "loss": 0.3518, + "step": 139 + }, + { + "epoch": 0.11496612605214535, + "grad_norm": 0.08816584199666977, + "learning_rate": 9.629731212287438e-05, + "loss": 0.3727, + "step": 140 + }, + { + "epoch": 0.11578731266680353, + "grad_norm": 0.11618969589471817, + "learning_rate": 9.626988480526605e-05, + "loss": 0.4247, + "step": 141 + }, + { + "epoch": 0.1166084992814617, + "grad_norm": 0.09243030101060867, + "learning_rate": 9.62424574876577e-05, + "loss": 0.396, + "step": 142 + }, + { + "epoch": 0.11742968589611989, + "grad_norm": 0.08566376566886902, + "learning_rate": 9.621503017004937e-05, + "loss": 0.3852, + "step": 143 + }, + { + "epoch": 0.11825087251077808, + "grad_norm": 0.08220973610877991, + "learning_rate": 9.618760285244104e-05, + "loss": 0.3961, + "step": 144 + }, + { + "epoch": 0.11907205912543625, + "grad_norm": 0.08240345865488052, + "learning_rate": 9.61601755348327e-05, + "loss": 0.3518, + "step": 145 + }, + { + "epoch": 0.11989324574009444, + "grad_norm": 0.08472532778978348, + "learning_rate": 9.613274821722435e-05, + "loss": 0.3586, + "step": 146 + }, + { + "epoch": 0.12071443235475261, + "grad_norm": 0.08407485485076904, + "learning_rate": 9.610532089961602e-05, + "loss": 0.3797, + "step": 147 + }, + { + "epoch": 0.1215356189694108, + "grad_norm": 0.09284385293722153, + "learning_rate": 9.607789358200769e-05, + "loss": 0.3499, + "step": 148 + }, + { + "epoch": 0.12235680558406897, + "grad_norm": 0.08499818295240402, + "learning_rate": 9.605046626439934e-05, + "loss": 0.3722, + "step": 149 + }, + { + "epoch": 0.12317799219872716, + "grad_norm": 0.080271415412426, + "learning_rate": 9.602303894679101e-05, + "loss": 0.3871, + "step": 150 + }, + { + "epoch": 0.12399917881338535, + "grad_norm": 0.07850060611963272, + "learning_rate": 9.599561162918266e-05, + "loss": 0.3679, + "step": 151 + }, + { + "epoch": 0.12482036542804352, + "grad_norm": 0.07685016840696335, + "learning_rate": 9.596818431157433e-05, + "loss": 0.4078, + "step": 152 + }, + { + "epoch": 0.1256415520427017, + "grad_norm": 0.09402357786893845, + "learning_rate": 9.5940756993966e-05, + "loss": 0.3996, + "step": 153 + }, + { + "epoch": 0.12646273865735988, + "grad_norm": 0.08445476740598679, + "learning_rate": 9.591332967635765e-05, + "loss": 0.3841, + "step": 154 + }, + { + "epoch": 0.12728392527201807, + "grad_norm": 0.08233911544084549, + "learning_rate": 9.588590235874932e-05, + "loss": 0.3529, + "step": 155 + }, + { + "epoch": 0.12810511188667625, + "grad_norm": 0.07896068692207336, + "learning_rate": 9.585847504114099e-05, + "loss": 0.3585, + "step": 156 + }, + { + "epoch": 0.12892629850133444, + "grad_norm": 0.0822276696562767, + "learning_rate": 9.583104772353264e-05, + "loss": 0.3962, + "step": 157 + }, + { + "epoch": 0.1297474851159926, + "grad_norm": 0.07977598905563354, + "learning_rate": 9.580362040592431e-05, + "loss": 0.454, + "step": 158 + }, + { + "epoch": 0.1305686717306508, + "grad_norm": 0.0857616737484932, + "learning_rate": 9.577619308831597e-05, + "loss": 0.3954, + "step": 159 + }, + { + "epoch": 0.13138985834530897, + "grad_norm": 0.0874355211853981, + "learning_rate": 9.574876577070763e-05, + "loss": 0.3591, + "step": 160 + }, + { + "epoch": 0.13221104495996716, + "grad_norm": 0.07877468317747116, + "learning_rate": 9.57213384530993e-05, + "loss": 0.348, + "step": 161 + }, + { + "epoch": 0.13303223157462532, + "grad_norm": 0.08618593961000443, + "learning_rate": 9.569391113549096e-05, + "loss": 0.3723, + "step": 162 + }, + { + "epoch": 0.1338534181892835, + "grad_norm": 0.08144336938858032, + "learning_rate": 9.566648381788261e-05, + "loss": 0.4146, + "step": 163 + }, + { + "epoch": 0.1346746048039417, + "grad_norm": 0.07322760671377182, + "learning_rate": 9.563905650027428e-05, + "loss": 0.3107, + "step": 164 + }, + { + "epoch": 0.13549579141859988, + "grad_norm": 0.08007095754146576, + "learning_rate": 9.561162918266595e-05, + "loss": 0.4128, + "step": 165 + }, + { + "epoch": 0.13631697803325807, + "grad_norm": 0.09636646509170532, + "learning_rate": 9.55842018650576e-05, + "loss": 0.4089, + "step": 166 + }, + { + "epoch": 0.13713816464791623, + "grad_norm": 0.08381053060293198, + "learning_rate": 9.555677454744927e-05, + "loss": 0.3624, + "step": 167 + }, + { + "epoch": 0.13795935126257441, + "grad_norm": 0.07476504147052765, + "learning_rate": 9.552934722984092e-05, + "loss": 0.3906, + "step": 168 + }, + { + "epoch": 0.1387805378772326, + "grad_norm": 0.0901239663362503, + "learning_rate": 9.550191991223259e-05, + "loss": 0.3378, + "step": 169 + }, + { + "epoch": 0.1396017244918908, + "grad_norm": 0.0813356265425682, + "learning_rate": 9.547449259462426e-05, + "loss": 0.3627, + "step": 170 + }, + { + "epoch": 0.14042291110654898, + "grad_norm": 0.14319093525409698, + "learning_rate": 9.544706527701591e-05, + "loss": 0.3512, + "step": 171 + }, + { + "epoch": 0.14124409772120713, + "grad_norm": 0.13329866528511047, + "learning_rate": 9.541963795940757e-05, + "loss": 0.3809, + "step": 172 + }, + { + "epoch": 0.14206528433586532, + "grad_norm": 0.0815596953034401, + "learning_rate": 9.539221064179923e-05, + "loss": 0.3444, + "step": 173 + }, + { + "epoch": 0.1428864709505235, + "grad_norm": 0.08646956831216812, + "learning_rate": 9.53647833241909e-05, + "loss": 0.3699, + "step": 174 + }, + { + "epoch": 0.1437076575651817, + "grad_norm": 0.09374339133501053, + "learning_rate": 9.533735600658256e-05, + "loss": 0.3569, + "step": 175 + }, + { + "epoch": 0.14452884417983985, + "grad_norm": 0.0834718644618988, + "learning_rate": 9.530992868897422e-05, + "loss": 0.3708, + "step": 176 + }, + { + "epoch": 0.14452884417983985, + "eval_runtime": 493.9539, + "eval_samples_per_second": 0.399, + "eval_steps_per_second": 0.2, + "step": 176 + }, + { + "epoch": 0.14535003079449804, + "grad_norm": 0.08648290485143661, + "learning_rate": 9.528250137136588e-05, + "loss": 0.379, + "step": 177 + }, + { + "epoch": 0.14617121740915623, + "grad_norm": 0.08577203750610352, + "learning_rate": 9.525507405375755e-05, + "loss": 0.4368, + "step": 178 + }, + { + "epoch": 0.14699240402381442, + "grad_norm": 0.1023576483130455, + "learning_rate": 9.522764673614921e-05, + "loss": 0.3553, + "step": 179 + }, + { + "epoch": 0.1478135906384726, + "grad_norm": 0.08062634617090225, + "learning_rate": 9.520021941854087e-05, + "loss": 0.3616, + "step": 180 + }, + { + "epoch": 0.14863477725313076, + "grad_norm": 0.07487751543521881, + "learning_rate": 9.517279210093252e-05, + "loss": 0.4413, + "step": 181 + }, + { + "epoch": 0.14945596386778895, + "grad_norm": 0.07332492619752884, + "learning_rate": 9.514536478332419e-05, + "loss": 0.3166, + "step": 182 + }, + { + "epoch": 0.15027715048244714, + "grad_norm": 0.08658608049154282, + "learning_rate": 9.511793746571586e-05, + "loss": 0.332, + "step": 183 + }, + { + "epoch": 0.15109833709710532, + "grad_norm": 0.10461894422769547, + "learning_rate": 9.509051014810751e-05, + "loss": 0.3653, + "step": 184 + }, + { + "epoch": 0.1519195237117635, + "grad_norm": 0.19619494676589966, + "learning_rate": 9.506308283049918e-05, + "loss": 0.4224, + "step": 185 + }, + { + "epoch": 0.15274071032642167, + "grad_norm": 0.08313202857971191, + "learning_rate": 9.503565551289084e-05, + "loss": 0.3512, + "step": 186 + }, + { + "epoch": 0.15356189694107986, + "grad_norm": 0.08686342090368271, + "learning_rate": 9.50082281952825e-05, + "loss": 0.3352, + "step": 187 + }, + { + "epoch": 0.15438308355573804, + "grad_norm": 0.09020522236824036, + "learning_rate": 9.498080087767417e-05, + "loss": 0.4138, + "step": 188 + }, + { + "epoch": 0.15520427017039623, + "grad_norm": 0.0799839124083519, + "learning_rate": 9.495337356006583e-05, + "loss": 0.3436, + "step": 189 + }, + { + "epoch": 0.15602545678505442, + "grad_norm": 0.08851379156112671, + "learning_rate": 9.49259462424575e-05, + "loss": 0.3933, + "step": 190 + }, + { + "epoch": 0.15684664339971258, + "grad_norm": 0.08521082252264023, + "learning_rate": 9.489851892484916e-05, + "loss": 0.3667, + "step": 191 + }, + { + "epoch": 0.15766783001437076, + "grad_norm": 0.09019312262535095, + "learning_rate": 9.487109160724082e-05, + "loss": 0.3556, + "step": 192 + }, + { + "epoch": 0.15848901662902895, + "grad_norm": 0.08162654936313629, + "learning_rate": 9.484366428963248e-05, + "loss": 0.3864, + "step": 193 + }, + { + "epoch": 0.15931020324368714, + "grad_norm": 0.08963490277528763, + "learning_rate": 9.481623697202414e-05, + "loss": 0.391, + "step": 194 + }, + { + "epoch": 0.1601313898583453, + "grad_norm": 0.10057719051837921, + "learning_rate": 9.47888096544158e-05, + "loss": 0.3599, + "step": 195 + }, + { + "epoch": 0.16095257647300348, + "grad_norm": 0.1694149523973465, + "learning_rate": 9.476138233680747e-05, + "loss": 0.3612, + "step": 196 + }, + { + "epoch": 0.16177376308766167, + "grad_norm": 0.08255323767662048, + "learning_rate": 9.473395501919913e-05, + "loss": 0.3441, + "step": 197 + }, + { + "epoch": 0.16259494970231986, + "grad_norm": 0.09228333085775375, + "learning_rate": 9.470652770159078e-05, + "loss": 0.3432, + "step": 198 + }, + { + "epoch": 0.16341613631697804, + "grad_norm": 0.11736617982387543, + "learning_rate": 9.467910038398245e-05, + "loss": 0.3408, + "step": 199 + }, + { + "epoch": 0.1642373229316362, + "grad_norm": 0.16636626422405243, + "learning_rate": 9.465167306637412e-05, + "loss": 0.3639, + "step": 200 + }, + { + "epoch": 0.1650585095462944, + "grad_norm": 0.0912085771560669, + "learning_rate": 9.462424574876577e-05, + "loss": 0.4055, + "step": 201 + }, + { + "epoch": 0.16587969616095258, + "grad_norm": 0.08452475070953369, + "learning_rate": 9.459681843115744e-05, + "loss": 0.3497, + "step": 202 + }, + { + "epoch": 0.16670088277561076, + "grad_norm": 0.09225429594516754, + "learning_rate": 9.45693911135491e-05, + "loss": 0.3765, + "step": 203 + }, + { + "epoch": 0.16752206939026895, + "grad_norm": 0.08500406891107559, + "learning_rate": 9.454196379594076e-05, + "loss": 0.3622, + "step": 204 + }, + { + "epoch": 0.1683432560049271, + "grad_norm": 0.08277002722024918, + "learning_rate": 9.451453647833243e-05, + "loss": 0.3596, + "step": 205 + }, + { + "epoch": 0.1691644426195853, + "grad_norm": 0.0855122059583664, + "learning_rate": 9.448710916072408e-05, + "loss": 0.3495, + "step": 206 + }, + { + "epoch": 0.16998562923424348, + "grad_norm": 0.11409081518650055, + "learning_rate": 9.445968184311574e-05, + "loss": 0.3123, + "step": 207 + }, + { + "epoch": 0.17080681584890167, + "grad_norm": 0.08898866921663284, + "learning_rate": 9.443225452550741e-05, + "loss": 0.3383, + "step": 208 + }, + { + "epoch": 0.17162800246355983, + "grad_norm": 0.07967101782560349, + "learning_rate": 9.440482720789908e-05, + "loss": 0.3927, + "step": 209 + }, + { + "epoch": 0.17244918907821802, + "grad_norm": 0.08267655968666077, + "learning_rate": 9.437739989029073e-05, + "loss": 0.3503, + "step": 210 + }, + { + "epoch": 0.1732703756928762, + "grad_norm": 0.1024966835975647, + "learning_rate": 9.43499725726824e-05, + "loss": 0.343, + "step": 211 + }, + { + "epoch": 0.1740915623075344, + "grad_norm": 0.11825034767389297, + "learning_rate": 9.432254525507405e-05, + "loss": 0.3694, + "step": 212 + }, + { + "epoch": 0.17491274892219258, + "grad_norm": 0.08225111663341522, + "learning_rate": 9.429511793746572e-05, + "loss": 0.3625, + "step": 213 + }, + { + "epoch": 0.17573393553685074, + "grad_norm": 0.082975834608078, + "learning_rate": 9.426769061985739e-05, + "loss": 0.3692, + "step": 214 + }, + { + "epoch": 0.17655512215150893, + "grad_norm": 0.09659875184297562, + "learning_rate": 9.424026330224904e-05, + "loss": 0.3857, + "step": 215 + }, + { + "epoch": 0.1773763087661671, + "grad_norm": 0.08381886035203934, + "learning_rate": 9.42128359846407e-05, + "loss": 0.3658, + "step": 216 + }, + { + "epoch": 0.1781974953808253, + "grad_norm": 0.08097488433122635, + "learning_rate": 9.418540866703238e-05, + "loss": 0.3556, + "step": 217 + }, + { + "epoch": 0.1790186819954835, + "grad_norm": 0.08661879599094391, + "learning_rate": 9.415798134942403e-05, + "loss": 0.3495, + "step": 218 + }, + { + "epoch": 0.17983986861014165, + "grad_norm": 0.09621778875589371, + "learning_rate": 9.413055403181569e-05, + "loss": 0.3488, + "step": 219 + }, + { + "epoch": 0.18066105522479983, + "grad_norm": 0.0867924615740776, + "learning_rate": 9.410312671420735e-05, + "loss": 0.3496, + "step": 220 + }, + { + "epoch": 0.18148224183945802, + "grad_norm": 0.09928230196237564, + "learning_rate": 9.407569939659901e-05, + "loss": 0.411, + "step": 221 + }, + { + "epoch": 0.1823034284541162, + "grad_norm": 0.08545473217964172, + "learning_rate": 9.404827207899069e-05, + "loss": 0.3707, + "step": 222 + }, + { + "epoch": 0.18312461506877437, + "grad_norm": 0.1317296177148819, + "learning_rate": 9.402084476138234e-05, + "loss": 0.4021, + "step": 223 + }, + { + "epoch": 0.18394580168343255, + "grad_norm": 0.09009065479040146, + "learning_rate": 9.3993417443774e-05, + "loss": 0.3451, + "step": 224 + }, + { + "epoch": 0.18476698829809074, + "grad_norm": 0.08332253247499466, + "learning_rate": 9.396599012616567e-05, + "loss": 0.3457, + "step": 225 + }, + { + "epoch": 0.18558817491274893, + "grad_norm": 0.08737312257289886, + "learning_rate": 9.393856280855733e-05, + "loss": 0.3721, + "step": 226 + }, + { + "epoch": 0.1864093615274071, + "grad_norm": 0.07472239434719086, + "learning_rate": 9.391113549094899e-05, + "loss": 0.3542, + "step": 227 + }, + { + "epoch": 0.18723054814206527, + "grad_norm": 0.08319877088069916, + "learning_rate": 9.388370817334066e-05, + "loss": 0.361, + "step": 228 + }, + { + "epoch": 0.18805173475672346, + "grad_norm": 0.08192326873540878, + "learning_rate": 9.385628085573231e-05, + "loss": 0.344, + "step": 229 + }, + { + "epoch": 0.18887292137138165, + "grad_norm": 0.11502642929553986, + "learning_rate": 9.382885353812398e-05, + "loss": 0.363, + "step": 230 + }, + { + "epoch": 0.18969410798603983, + "grad_norm": 0.08567750453948975, + "learning_rate": 9.380142622051565e-05, + "loss": 0.3565, + "step": 231 + }, + { + "epoch": 0.19051529460069802, + "grad_norm": 0.08059141039848328, + "learning_rate": 9.37739989029073e-05, + "loss": 0.3722, + "step": 232 + }, + { + "epoch": 0.19133648121535618, + "grad_norm": 0.07670270651578903, + "learning_rate": 9.374657158529896e-05, + "loss": 0.3493, + "step": 233 + }, + { + "epoch": 0.19215766783001437, + "grad_norm": 0.08022642135620117, + "learning_rate": 9.371914426769062e-05, + "loss": 0.4142, + "step": 234 + }, + { + "epoch": 0.19297885444467255, + "grad_norm": 0.08015397936105728, + "learning_rate": 9.369171695008229e-05, + "loss": 0.3442, + "step": 235 + }, + { + "epoch": 0.19380004105933074, + "grad_norm": 0.08329442143440247, + "learning_rate": 9.366428963247395e-05, + "loss": 0.3587, + "step": 236 + }, + { + "epoch": 0.1946212276739889, + "grad_norm": 0.08311276882886887, + "learning_rate": 9.363686231486561e-05, + "loss": 0.3792, + "step": 237 + }, + { + "epoch": 0.1954424142886471, + "grad_norm": 0.1332862675189972, + "learning_rate": 9.360943499725727e-05, + "loss": 0.3451, + "step": 238 + }, + { + "epoch": 0.19626360090330527, + "grad_norm": 0.081804558634758, + "learning_rate": 9.358200767964894e-05, + "loss": 0.4212, + "step": 239 + }, + { + "epoch": 0.19708478751796346, + "grad_norm": 0.08397019654512405, + "learning_rate": 9.35545803620406e-05, + "loss": 0.3482, + "step": 240 + }, + { + "epoch": 0.19790597413262165, + "grad_norm": 0.09495637565851212, + "learning_rate": 9.352715304443226e-05, + "loss": 0.3432, + "step": 241 + }, + { + "epoch": 0.1987271607472798, + "grad_norm": 0.09187504649162292, + "learning_rate": 9.349972572682391e-05, + "loss": 0.3938, + "step": 242 + }, + { + "epoch": 0.199548347361938, + "grad_norm": 0.100834921002388, + "learning_rate": 9.347229840921558e-05, + "loss": 0.3567, + "step": 243 + }, + { + "epoch": 0.20036953397659618, + "grad_norm": 0.09292273223400116, + "learning_rate": 9.344487109160725e-05, + "loss": 0.3802, + "step": 244 + }, + { + "epoch": 0.20119072059125437, + "grad_norm": 0.07785986363887787, + "learning_rate": 9.34174437739989e-05, + "loss": 0.3832, + "step": 245 + }, + { + "epoch": 0.20201190720591256, + "grad_norm": 0.081189826130867, + "learning_rate": 9.339001645639057e-05, + "loss": 0.3448, + "step": 246 + }, + { + "epoch": 0.20283309382057071, + "grad_norm": 0.08209879696369171, + "learning_rate": 9.336258913878222e-05, + "loss": 0.3602, + "step": 247 + }, + { + "epoch": 0.2036542804352289, + "grad_norm": 0.0813421905040741, + "learning_rate": 9.333516182117389e-05, + "loss": 0.3477, + "step": 248 + }, + { + "epoch": 0.2044754670498871, + "grad_norm": 0.08188773691654205, + "learning_rate": 9.330773450356556e-05, + "loss": 0.3466, + "step": 249 + }, + { + "epoch": 0.20529665366454528, + "grad_norm": 0.09966633468866348, + "learning_rate": 9.328030718595721e-05, + "loss": 0.3776, + "step": 250 + }, + { + "epoch": 0.20611784027920346, + "grad_norm": 0.08154954016208649, + "learning_rate": 9.325287986834887e-05, + "loss": 0.3727, + "step": 251 + }, + { + "epoch": 0.20693902689386162, + "grad_norm": 0.07942931354045868, + "learning_rate": 9.322545255074055e-05, + "loss": 0.3468, + "step": 252 + }, + { + "epoch": 0.2077602135085198, + "grad_norm": 0.07944433391094208, + "learning_rate": 9.31980252331322e-05, + "loss": 0.3512, + "step": 253 + }, + { + "epoch": 0.208581400123178, + "grad_norm": 0.08637971431016922, + "learning_rate": 9.317059791552387e-05, + "loss": 0.3397, + "step": 254 + }, + { + "epoch": 0.20940258673783618, + "grad_norm": 0.08189195394515991, + "learning_rate": 9.314317059791553e-05, + "loss": 0.4105, + "step": 255 + }, + { + "epoch": 0.21022377335249434, + "grad_norm": 0.08190836012363434, + "learning_rate": 9.31157432803072e-05, + "loss": 0.346, + "step": 256 + }, + { + "epoch": 0.21104495996715253, + "grad_norm": 0.08331865072250366, + "learning_rate": 9.308831596269886e-05, + "loss": 0.3655, + "step": 257 + }, + { + "epoch": 0.21186614658181072, + "grad_norm": 0.0822620540857315, + "learning_rate": 9.306088864509052e-05, + "loss": 0.372, + "step": 258 + }, + { + "epoch": 0.2126873331964689, + "grad_norm": 0.08011777698993683, + "learning_rate": 9.303346132748217e-05, + "loss": 0.3304, + "step": 259 + }, + { + "epoch": 0.2135085198111271, + "grad_norm": 0.08395062386989594, + "learning_rate": 9.300603400987384e-05, + "loss": 0.3903, + "step": 260 + }, + { + "epoch": 0.21432970642578525, + "grad_norm": 0.08101452887058258, + "learning_rate": 9.297860669226551e-05, + "loss": 0.3579, + "step": 261 + }, + { + "epoch": 0.21515089304044344, + "grad_norm": 0.11504925042390823, + "learning_rate": 9.295117937465716e-05, + "loss": 0.3723, + "step": 262 + }, + { + "epoch": 0.21597207965510162, + "grad_norm": 0.0704338401556015, + "learning_rate": 9.292375205704883e-05, + "loss": 0.3341, + "step": 263 + }, + { + "epoch": 0.2167932662697598, + "grad_norm": 0.07944470643997192, + "learning_rate": 9.289632473944048e-05, + "loss": 0.3273, + "step": 264 + }, + { + "epoch": 0.2167932662697598, + "eval_runtime": 507.3366, + "eval_samples_per_second": 0.388, + "eval_steps_per_second": 0.195, + "step": 264 + }, + { + "epoch": 0.217614452884418, + "grad_norm": 0.0778859481215477, + "learning_rate": 9.286889742183215e-05, + "loss": 0.3707, + "step": 265 + }, + { + "epoch": 0.21843563949907616, + "grad_norm": 0.07564377039670944, + "learning_rate": 9.284147010422382e-05, + "loss": 0.334, + "step": 266 + }, + { + "epoch": 0.21925682611373434, + "grad_norm": 0.07744503021240234, + "learning_rate": 9.281404278661547e-05, + "loss": 0.3521, + "step": 267 + }, + { + "epoch": 0.22007801272839253, + "grad_norm": 0.08139525353908539, + "learning_rate": 9.278661546900713e-05, + "loss": 0.4069, + "step": 268 + }, + { + "epoch": 0.22089919934305072, + "grad_norm": 0.08162204176187515, + "learning_rate": 9.27591881513988e-05, + "loss": 0.3292, + "step": 269 + }, + { + "epoch": 0.22172038595770888, + "grad_norm": 0.07798879593610764, + "learning_rate": 9.273176083379046e-05, + "loss": 0.3441, + "step": 270 + }, + { + "epoch": 0.22254157257236706, + "grad_norm": 0.07629597932100296, + "learning_rate": 9.270433351618212e-05, + "loss": 0.3989, + "step": 271 + }, + { + "epoch": 0.22336275918702525, + "grad_norm": 0.07695835083723068, + "learning_rate": 9.267690619857379e-05, + "loss": 0.3506, + "step": 272 + }, + { + "epoch": 0.22418394580168344, + "grad_norm": 0.07938527315855026, + "learning_rate": 9.264947888096544e-05, + "loss": 0.327, + "step": 273 + }, + { + "epoch": 0.22500513241634162, + "grad_norm": 0.08125253766775131, + "learning_rate": 9.262205156335711e-05, + "loss": 0.3749, + "step": 274 + }, + { + "epoch": 0.22582631903099978, + "grad_norm": 0.07831385731697083, + "learning_rate": 9.259462424574878e-05, + "loss": 0.3477, + "step": 275 + }, + { + "epoch": 0.22664750564565797, + "grad_norm": 0.08408527821302414, + "learning_rate": 9.256719692814043e-05, + "loss": 0.3488, + "step": 276 + }, + { + "epoch": 0.22746869226031616, + "grad_norm": 0.0800737589597702, + "learning_rate": 9.253976961053208e-05, + "loss": 0.3607, + "step": 277 + }, + { + "epoch": 0.22828987887497434, + "grad_norm": 0.07969113439321518, + "learning_rate": 9.251234229292375e-05, + "loss": 0.38, + "step": 278 + }, + { + "epoch": 0.22911106548963253, + "grad_norm": 0.08066916465759277, + "learning_rate": 9.248491497531542e-05, + "loss": 0.3469, + "step": 279 + }, + { + "epoch": 0.2299322521042907, + "grad_norm": 0.07686860859394073, + "learning_rate": 9.245748765770708e-05, + "loss": 0.3311, + "step": 280 + }, + { + "epoch": 0.23075343871894888, + "grad_norm": 0.13063722848892212, + "learning_rate": 9.243006034009874e-05, + "loss": 0.3091, + "step": 281 + }, + { + "epoch": 0.23157462533360706, + "grad_norm": 0.08465570211410522, + "learning_rate": 9.24026330224904e-05, + "loss": 0.3548, + "step": 282 + }, + { + "epoch": 0.23239581194826525, + "grad_norm": 0.07824988663196564, + "learning_rate": 9.237520570488207e-05, + "loss": 0.3485, + "step": 283 + }, + { + "epoch": 0.2332169985629234, + "grad_norm": 0.087236687541008, + "learning_rate": 9.234777838727373e-05, + "loss": 0.3564, + "step": 284 + }, + { + "epoch": 0.2340381851775816, + "grad_norm": 0.09531310200691223, + "learning_rate": 9.232035106966539e-05, + "loss": 0.3763, + "step": 285 + }, + { + "epoch": 0.23485937179223978, + "grad_norm": 0.0783536285161972, + "learning_rate": 9.229292375205704e-05, + "loss": 0.3479, + "step": 286 + }, + { + "epoch": 0.23568055840689797, + "grad_norm": 0.08209957927465439, + "learning_rate": 9.226549643444872e-05, + "loss": 0.3563, + "step": 287 + }, + { + "epoch": 0.23650174502155616, + "grad_norm": 0.08195500820875168, + "learning_rate": 9.223806911684038e-05, + "loss": 0.3879, + "step": 288 + }, + { + "epoch": 0.23732293163621432, + "grad_norm": 0.0796385332942009, + "learning_rate": 9.221064179923205e-05, + "loss": 0.3787, + "step": 289 + }, + { + "epoch": 0.2381441182508725, + "grad_norm": 0.08523210138082504, + "learning_rate": 9.21832144816237e-05, + "loss": 0.3608, + "step": 290 + }, + { + "epoch": 0.2389653048655307, + "grad_norm": 0.08573302626609802, + "learning_rate": 9.215578716401537e-05, + "loss": 0.3366, + "step": 291 + }, + { + "epoch": 0.23978649148018888, + "grad_norm": 0.08111650496721268, + "learning_rate": 9.212835984640704e-05, + "loss": 0.3331, + "step": 292 + }, + { + "epoch": 0.24060767809484707, + "grad_norm": 0.07877853512763977, + "learning_rate": 9.210093252879869e-05, + "loss": 0.3346, + "step": 293 + }, + { + "epoch": 0.24142886470950523, + "grad_norm": 0.0814940333366394, + "learning_rate": 9.207350521119034e-05, + "loss": 0.3701, + "step": 294 + }, + { + "epoch": 0.2422500513241634, + "grad_norm": 0.0843539908528328, + "learning_rate": 9.204607789358201e-05, + "loss": 0.3339, + "step": 295 + }, + { + "epoch": 0.2430712379388216, + "grad_norm": 0.0875789001584053, + "learning_rate": 9.201865057597368e-05, + "loss": 0.362, + "step": 296 + }, + { + "epoch": 0.2438924245534798, + "grad_norm": 0.08261071145534515, + "learning_rate": 9.199122325836533e-05, + "loss": 0.3578, + "step": 297 + }, + { + "epoch": 0.24471361116813795, + "grad_norm": 0.09160041064023972, + "learning_rate": 9.1963795940757e-05, + "loss": 0.387, + "step": 298 + }, + { + "epoch": 0.24553479778279613, + "grad_norm": 0.08152728527784348, + "learning_rate": 9.193636862314866e-05, + "loss": 0.3282, + "step": 299 + }, + { + "epoch": 0.24635598439745432, + "grad_norm": 0.079656682908535, + "learning_rate": 9.190894130554032e-05, + "loss": 0.3238, + "step": 300 + }, + { + "epoch": 0.2471771710121125, + "grad_norm": 0.0797642394900322, + "learning_rate": 9.188151398793199e-05, + "loss": 0.3506, + "step": 301 + }, + { + "epoch": 0.2479983576267707, + "grad_norm": 0.08688945323228836, + "learning_rate": 9.185408667032365e-05, + "loss": 0.4177, + "step": 302 + }, + { + "epoch": 0.24881954424142885, + "grad_norm": 0.106851726770401, + "learning_rate": 9.18266593527153e-05, + "loss": 0.3527, + "step": 303 + }, + { + "epoch": 0.24964073085608704, + "grad_norm": 0.08586329221725464, + "learning_rate": 9.179923203510697e-05, + "loss": 0.3763, + "step": 304 + }, + { + "epoch": 0.25046191747074525, + "grad_norm": 0.1141560897231102, + "learning_rate": 9.177180471749864e-05, + "loss": 0.3451, + "step": 305 + }, + { + "epoch": 0.2512831040854034, + "grad_norm": 0.08728976547718048, + "learning_rate": 9.174437739989029e-05, + "loss": 0.3444, + "step": 306 + }, + { + "epoch": 0.2521042907000616, + "grad_norm": 0.07848083972930908, + "learning_rate": 9.171695008228196e-05, + "loss": 0.344, + "step": 307 + }, + { + "epoch": 0.25292547731471976, + "grad_norm": 0.08140011131763458, + "learning_rate": 9.168952276467361e-05, + "loss": 0.3338, + "step": 308 + }, + { + "epoch": 0.25374666392937795, + "grad_norm": 0.07934946566820145, + "learning_rate": 9.166209544706528e-05, + "loss": 0.3869, + "step": 309 + }, + { + "epoch": 0.25456785054403613, + "grad_norm": 0.08028525859117508, + "learning_rate": 9.163466812945695e-05, + "loss": 0.3141, + "step": 310 + }, + { + "epoch": 0.2553890371586943, + "grad_norm": 0.08121496438980103, + "learning_rate": 9.16072408118486e-05, + "loss": 0.4098, + "step": 311 + }, + { + "epoch": 0.2562102237733525, + "grad_norm": 0.09907019138336182, + "learning_rate": 9.157981349424026e-05, + "loss": 0.3594, + "step": 312 + }, + { + "epoch": 0.2570314103880107, + "grad_norm": 0.10270131379365921, + "learning_rate": 9.155238617663193e-05, + "loss": 0.3713, + "step": 313 + }, + { + "epoch": 0.2578525970026689, + "grad_norm": 0.08354981243610382, + "learning_rate": 9.15249588590236e-05, + "loss": 0.3943, + "step": 314 + }, + { + "epoch": 0.258673783617327, + "grad_norm": 0.09153895825147629, + "learning_rate": 9.149753154141525e-05, + "loss": 0.3499, + "step": 315 + }, + { + "epoch": 0.2594949702319852, + "grad_norm": 0.07772937417030334, + "learning_rate": 9.147010422380692e-05, + "loss": 0.3323, + "step": 316 + }, + { + "epoch": 0.2603161568466434, + "grad_norm": 0.08953258395195007, + "learning_rate": 9.144267690619857e-05, + "loss": 0.3684, + "step": 317 + }, + { + "epoch": 0.2611373434613016, + "grad_norm": 0.08204226195812225, + "learning_rate": 9.141524958859024e-05, + "loss": 0.3387, + "step": 318 + }, + { + "epoch": 0.26195853007595976, + "grad_norm": 0.07604125887155533, + "learning_rate": 9.13878222709819e-05, + "loss": 0.3104, + "step": 319 + }, + { + "epoch": 0.26277971669061795, + "grad_norm": 0.07614021003246307, + "learning_rate": 9.136039495337356e-05, + "loss": 0.329, + "step": 320 + }, + { + "epoch": 0.26360090330527614, + "grad_norm": 0.19391578435897827, + "learning_rate": 9.133296763576523e-05, + "loss": 0.349, + "step": 321 + }, + { + "epoch": 0.2644220899199343, + "grad_norm": 0.07927168160676956, + "learning_rate": 9.13055403181569e-05, + "loss": 0.3495, + "step": 322 + }, + { + "epoch": 0.2652432765345925, + "grad_norm": 0.08091975003480911, + "learning_rate": 9.127811300054855e-05, + "loss": 0.3563, + "step": 323 + }, + { + "epoch": 0.26606446314925064, + "grad_norm": 0.07865800708532333, + "learning_rate": 9.125068568294022e-05, + "loss": 0.36, + "step": 324 + }, + { + "epoch": 0.2668856497639088, + "grad_norm": 0.08946891874074936, + "learning_rate": 9.122325836533187e-05, + "loss": 0.3434, + "step": 325 + }, + { + "epoch": 0.267706836378567, + "grad_norm": 0.07987753301858902, + "learning_rate": 9.119583104772354e-05, + "loss": 0.3593, + "step": 326 + }, + { + "epoch": 0.2685280229932252, + "grad_norm": 0.07894265651702881, + "learning_rate": 9.116840373011521e-05, + "loss": 0.3746, + "step": 327 + }, + { + "epoch": 0.2693492096078834, + "grad_norm": 0.08121046423912048, + "learning_rate": 9.114097641250686e-05, + "loss": 0.357, + "step": 328 + }, + { + "epoch": 0.2701703962225416, + "grad_norm": 0.08320184051990509, + "learning_rate": 9.111354909489852e-05, + "loss": 0.3783, + "step": 329 + }, + { + "epoch": 0.27099158283719976, + "grad_norm": 0.10911049693822861, + "learning_rate": 9.108612177729018e-05, + "loss": 0.3699, + "step": 330 + }, + { + "epoch": 0.27181276945185795, + "grad_norm": 0.09739838540554047, + "learning_rate": 9.105869445968185e-05, + "loss": 0.3524, + "step": 331 + }, + { + "epoch": 0.27263395606651614, + "grad_norm": 0.0874953344464302, + "learning_rate": 9.103126714207351e-05, + "loss": 0.3807, + "step": 332 + }, + { + "epoch": 0.2734551426811743, + "grad_norm": 0.09758396446704865, + "learning_rate": 9.100383982446518e-05, + "loss": 0.3876, + "step": 333 + }, + { + "epoch": 0.27427632929583246, + "grad_norm": 0.08628170937299728, + "learning_rate": 9.097641250685683e-05, + "loss": 0.3632, + "step": 334 + }, + { + "epoch": 0.27509751591049064, + "grad_norm": 0.08430638164281845, + "learning_rate": 9.09489851892485e-05, + "loss": 0.3448, + "step": 335 + }, + { + "epoch": 0.27591870252514883, + "grad_norm": 0.0725574865937233, + "learning_rate": 9.092155787164017e-05, + "loss": 0.2914, + "step": 336 + }, + { + "epoch": 0.276739889139807, + "grad_norm": 0.08421092480421066, + "learning_rate": 9.089413055403182e-05, + "loss": 0.3676, + "step": 337 + }, + { + "epoch": 0.2775610757544652, + "grad_norm": 0.08702068030834198, + "learning_rate": 9.086670323642347e-05, + "loss": 0.345, + "step": 338 + }, + { + "epoch": 0.2783822623691234, + "grad_norm": 0.08053667843341827, + "learning_rate": 9.083927591881514e-05, + "loss": 0.324, + "step": 339 + }, + { + "epoch": 0.2792034489837816, + "grad_norm": 0.08294267952442169, + "learning_rate": 9.081184860120681e-05, + "loss": 0.348, + "step": 340 + }, + { + "epoch": 0.28002463559843976, + "grad_norm": 0.18557026982307434, + "learning_rate": 9.078442128359846e-05, + "loss": 0.3483, + "step": 341 + }, + { + "epoch": 0.28084582221309795, + "grad_norm": 0.08436847478151321, + "learning_rate": 9.075699396599013e-05, + "loss": 0.3641, + "step": 342 + }, + { + "epoch": 0.2816670088277561, + "grad_norm": 0.08285505324602127, + "learning_rate": 9.072956664838179e-05, + "loss": 0.3641, + "step": 343 + }, + { + "epoch": 0.28248819544241427, + "grad_norm": 0.08681569993495941, + "learning_rate": 9.070213933077345e-05, + "loss": 0.342, + "step": 344 + }, + { + "epoch": 0.28330938205707246, + "grad_norm": 0.08668892085552216, + "learning_rate": 9.067471201316512e-05, + "loss": 0.3514, + "step": 345 + }, + { + "epoch": 0.28413056867173064, + "grad_norm": 0.08393285423517227, + "learning_rate": 9.064728469555678e-05, + "loss": 0.3531, + "step": 346 + }, + { + "epoch": 0.28495175528638883, + "grad_norm": 0.07783684879541397, + "learning_rate": 9.061985737794843e-05, + "loss": 0.3792, + "step": 347 + }, + { + "epoch": 0.285772941901047, + "grad_norm": 0.08259117603302002, + "learning_rate": 9.059243006034011e-05, + "loss": 0.375, + "step": 348 + }, + { + "epoch": 0.2865941285157052, + "grad_norm": 0.0969453975558281, + "learning_rate": 9.056500274273177e-05, + "loss": 0.3776, + "step": 349 + }, + { + "epoch": 0.2874153151303634, + "grad_norm": 0.08154502511024475, + "learning_rate": 9.053757542512342e-05, + "loss": 0.3577, + "step": 350 + }, + { + "epoch": 0.2882365017450216, + "grad_norm": 0.09322196245193481, + "learning_rate": 9.051014810751509e-05, + "loss": 0.3774, + "step": 351 + }, + { + "epoch": 0.2890576883596797, + "grad_norm": 0.0804852545261383, + "learning_rate": 9.048272078990674e-05, + "loss": 0.333, + "step": 352 + }, + { + "epoch": 0.2890576883596797, + "eval_runtime": 493.668, + "eval_samples_per_second": 0.399, + "eval_steps_per_second": 0.201, + "step": 352 + }, + { + "epoch": 0.2898788749743379, + "grad_norm": 0.08082845062017441, + "learning_rate": 9.045529347229841e-05, + "loss": 0.3111, + "step": 353 + }, + { + "epoch": 0.2907000615889961, + "grad_norm": 0.08396881073713303, + "learning_rate": 9.042786615469008e-05, + "loss": 0.3429, + "step": 354 + }, + { + "epoch": 0.29152124820365427, + "grad_norm": 0.08037351071834564, + "learning_rate": 9.040043883708173e-05, + "loss": 0.349, + "step": 355 + }, + { + "epoch": 0.29234243481831246, + "grad_norm": 0.10421521961688995, + "learning_rate": 9.03730115194734e-05, + "loss": 0.3228, + "step": 356 + }, + { + "epoch": 0.29316362143297064, + "grad_norm": 0.08268436044454575, + "learning_rate": 9.034558420186507e-05, + "loss": 0.3414, + "step": 357 + }, + { + "epoch": 0.29398480804762883, + "grad_norm": 0.08203061670064926, + "learning_rate": 9.031815688425672e-05, + "loss": 0.3596, + "step": 358 + }, + { + "epoch": 0.294805994662287, + "grad_norm": 0.07778481394052505, + "learning_rate": 9.029072956664839e-05, + "loss": 0.3337, + "step": 359 + }, + { + "epoch": 0.2956271812769452, + "grad_norm": 0.14194349944591522, + "learning_rate": 9.026330224904005e-05, + "loss": 0.3613, + "step": 360 + }, + { + "epoch": 0.2964483678916034, + "grad_norm": 0.09854190051555634, + "learning_rate": 9.023587493143171e-05, + "loss": 0.3862, + "step": 361 + }, + { + "epoch": 0.2972695545062615, + "grad_norm": 0.07878783345222473, + "learning_rate": 9.020844761382338e-05, + "loss": 0.3294, + "step": 362 + }, + { + "epoch": 0.2980907411209197, + "grad_norm": 0.07692375034093857, + "learning_rate": 9.018102029621504e-05, + "loss": 0.3083, + "step": 363 + }, + { + "epoch": 0.2989119277355779, + "grad_norm": 0.0772586315870285, + "learning_rate": 9.015359297860669e-05, + "loss": 0.3086, + "step": 364 + }, + { + "epoch": 0.2997331143502361, + "grad_norm": 0.0739215686917305, + "learning_rate": 9.012616566099836e-05, + "loss": 0.3103, + "step": 365 + }, + { + "epoch": 0.30055430096489427, + "grad_norm": 0.08434160053730011, + "learning_rate": 9.009873834339003e-05, + "loss": 0.3485, + "step": 366 + }, + { + "epoch": 0.30137548757955246, + "grad_norm": 0.08038529008626938, + "learning_rate": 9.007131102578168e-05, + "loss": 0.3421, + "step": 367 + }, + { + "epoch": 0.30219667419421065, + "grad_norm": 0.08303964138031006, + "learning_rate": 9.004388370817335e-05, + "loss": 0.3349, + "step": 368 + }, + { + "epoch": 0.30301786080886883, + "grad_norm": 0.08672379702329636, + "learning_rate": 9.0016456390565e-05, + "loss": 0.3228, + "step": 369 + }, + { + "epoch": 0.303839047423527, + "grad_norm": 0.08417553454637527, + "learning_rate": 8.998902907295667e-05, + "loss": 0.3312, + "step": 370 + }, + { + "epoch": 0.30466023403818515, + "grad_norm": 0.0842244029045105, + "learning_rate": 8.996160175534834e-05, + "loss": 0.3545, + "step": 371 + }, + { + "epoch": 0.30548142065284334, + "grad_norm": 0.08479683846235275, + "learning_rate": 8.993417443773999e-05, + "loss": 0.3446, + "step": 372 + }, + { + "epoch": 0.3063026072675015, + "grad_norm": 0.07726123929023743, + "learning_rate": 8.990674712013165e-05, + "loss": 0.3116, + "step": 373 + }, + { + "epoch": 0.3071237938821597, + "grad_norm": 0.08586291968822479, + "learning_rate": 8.987931980252331e-05, + "loss": 0.3617, + "step": 374 + }, + { + "epoch": 0.3079449804968179, + "grad_norm": 0.07475403696298599, + "learning_rate": 8.985189248491498e-05, + "loss": 0.3481, + "step": 375 + }, + { + "epoch": 0.3087661671114761, + "grad_norm": 0.08379320055246353, + "learning_rate": 8.982446516730664e-05, + "loss": 0.3434, + "step": 376 + }, + { + "epoch": 0.3095873537261343, + "grad_norm": 0.07624776661396027, + "learning_rate": 8.97970378496983e-05, + "loss": 0.3262, + "step": 377 + }, + { + "epoch": 0.31040854034079246, + "grad_norm": 0.07878948748111725, + "learning_rate": 8.976961053208996e-05, + "loss": 0.3366, + "step": 378 + }, + { + "epoch": 0.31122972695545065, + "grad_norm": 0.08924887329339981, + "learning_rate": 8.974218321448163e-05, + "loss": 0.3175, + "step": 379 + }, + { + "epoch": 0.31205091357010883, + "grad_norm": 0.08927389234304428, + "learning_rate": 8.97147558968733e-05, + "loss": 0.3338, + "step": 380 + }, + { + "epoch": 0.31287210018476697, + "grad_norm": 0.08250463008880615, + "learning_rate": 8.968732857926495e-05, + "loss": 0.3314, + "step": 381 + }, + { + "epoch": 0.31369328679942515, + "grad_norm": 0.0977792888879776, + "learning_rate": 8.96599012616566e-05, + "loss": 0.3222, + "step": 382 + }, + { + "epoch": 0.31451447341408334, + "grad_norm": 0.08197218924760818, + "learning_rate": 8.963247394404828e-05, + "loss": 0.3288, + "step": 383 + }, + { + "epoch": 0.3153356600287415, + "grad_norm": 0.12502075731754303, + "learning_rate": 8.960504662643994e-05, + "loss": 0.3053, + "step": 384 + }, + { + "epoch": 0.3161568466433997, + "grad_norm": 0.08058564364910126, + "learning_rate": 8.95776193088316e-05, + "loss": 0.3292, + "step": 385 + }, + { + "epoch": 0.3169780332580579, + "grad_norm": 0.08365239202976227, + "learning_rate": 8.955019199122326e-05, + "loss": 0.3747, + "step": 386 + }, + { + "epoch": 0.3177992198727161, + "grad_norm": 0.07539816945791245, + "learning_rate": 8.952276467361493e-05, + "loss": 0.3179, + "step": 387 + }, + { + "epoch": 0.3186204064873743, + "grad_norm": 0.08136752992868423, + "learning_rate": 8.94953373560066e-05, + "loss": 0.3647, + "step": 388 + }, + { + "epoch": 0.31944159310203246, + "grad_norm": 0.09546865522861481, + "learning_rate": 8.946791003839825e-05, + "loss": 0.4056, + "step": 389 + }, + { + "epoch": 0.3202627797166906, + "grad_norm": 0.08654627203941345, + "learning_rate": 8.94404827207899e-05, + "loss": 0.3853, + "step": 390 + }, + { + "epoch": 0.3210839663313488, + "grad_norm": 0.0822821855545044, + "learning_rate": 8.941305540318157e-05, + "loss": 0.327, + "step": 391 + }, + { + "epoch": 0.32190515294600697, + "grad_norm": 0.08066459000110626, + "learning_rate": 8.938562808557324e-05, + "loss": 0.3585, + "step": 392 + }, + { + "epoch": 0.32272633956066515, + "grad_norm": 0.08845409005880356, + "learning_rate": 8.93582007679649e-05, + "loss": 0.3419, + "step": 393 + }, + { + "epoch": 0.32354752617532334, + "grad_norm": 0.08482253551483154, + "learning_rate": 8.933077345035656e-05, + "loss": 0.3263, + "step": 394 + }, + { + "epoch": 0.32436871278998153, + "grad_norm": 0.0801006481051445, + "learning_rate": 8.930334613274822e-05, + "loss": 0.3713, + "step": 395 + }, + { + "epoch": 0.3251898994046397, + "grad_norm": 0.08470006287097931, + "learning_rate": 8.927591881513989e-05, + "loss": 0.3631, + "step": 396 + }, + { + "epoch": 0.3260110860192979, + "grad_norm": 0.08457473665475845, + "learning_rate": 8.924849149753155e-05, + "loss": 0.339, + "step": 397 + }, + { + "epoch": 0.3268322726339561, + "grad_norm": 0.14836351573467255, + "learning_rate": 8.922106417992321e-05, + "loss": 0.3039, + "step": 398 + }, + { + "epoch": 0.3276534592486142, + "grad_norm": 0.08391403406858444, + "learning_rate": 8.919363686231486e-05, + "loss": 0.3415, + "step": 399 + }, + { + "epoch": 0.3284746458632724, + "grad_norm": 0.07998157292604446, + "learning_rate": 8.916620954470653e-05, + "loss": 0.2943, + "step": 400 + }, + { + "epoch": 0.3292958324779306, + "grad_norm": 0.07682473957538605, + "learning_rate": 8.91387822270982e-05, + "loss": 0.3413, + "step": 401 + }, + { + "epoch": 0.3301170190925888, + "grad_norm": 0.09960923343896866, + "learning_rate": 8.911135490948985e-05, + "loss": 0.356, + "step": 402 + }, + { + "epoch": 0.33093820570724697, + "grad_norm": 0.08010439574718475, + "learning_rate": 8.908392759188152e-05, + "loss": 0.3408, + "step": 403 + }, + { + "epoch": 0.33175939232190516, + "grad_norm": 0.0811014398932457, + "learning_rate": 8.905650027427318e-05, + "loss": 0.3159, + "step": 404 + }, + { + "epoch": 0.33258057893656334, + "grad_norm": 0.08086768537759781, + "learning_rate": 8.902907295666484e-05, + "loss": 0.3145, + "step": 405 + }, + { + "epoch": 0.33340176555122153, + "grad_norm": 0.0778844803571701, + "learning_rate": 8.900164563905651e-05, + "loss": 0.3271, + "step": 406 + }, + { + "epoch": 0.3342229521658797, + "grad_norm": 0.08636100590229034, + "learning_rate": 8.897421832144817e-05, + "loss": 0.395, + "step": 407 + }, + { + "epoch": 0.3350441387805379, + "grad_norm": 0.08080980181694031, + "learning_rate": 8.894679100383982e-05, + "loss": 0.3819, + "step": 408 + }, + { + "epoch": 0.33586532539519603, + "grad_norm": 0.08312460780143738, + "learning_rate": 8.891936368623149e-05, + "loss": 0.3413, + "step": 409 + }, + { + "epoch": 0.3366865120098542, + "grad_norm": 0.07980356365442276, + "learning_rate": 8.889193636862316e-05, + "loss": 0.3546, + "step": 410 + }, + { + "epoch": 0.3375076986245124, + "grad_norm": 0.0868874043226242, + "learning_rate": 8.886450905101481e-05, + "loss": 0.3354, + "step": 411 + }, + { + "epoch": 0.3383288852391706, + "grad_norm": 0.08026042580604553, + "learning_rate": 8.883708173340648e-05, + "loss": 0.355, + "step": 412 + }, + { + "epoch": 0.3391500718538288, + "grad_norm": 0.0777151882648468, + "learning_rate": 8.880965441579813e-05, + "loss": 0.2835, + "step": 413 + }, + { + "epoch": 0.33997125846848697, + "grad_norm": 0.08007453382015228, + "learning_rate": 8.87822270981898e-05, + "loss": 0.3042, + "step": 414 + }, + { + "epoch": 0.34079244508314516, + "grad_norm": 0.08239760994911194, + "learning_rate": 8.875479978058147e-05, + "loss": 0.4139, + "step": 415 + }, + { + "epoch": 0.34161363169780334, + "grad_norm": 0.08022072911262512, + "learning_rate": 8.872737246297312e-05, + "loss": 0.319, + "step": 416 + }, + { + "epoch": 0.34243481831246153, + "grad_norm": 0.09205158054828644, + "learning_rate": 8.869994514536478e-05, + "loss": 0.3283, + "step": 417 + }, + { + "epoch": 0.34325600492711966, + "grad_norm": 0.07974108308553696, + "learning_rate": 8.867251782775646e-05, + "loss": 0.3331, + "step": 418 + }, + { + "epoch": 0.34407719154177785, + "grad_norm": 0.0810483992099762, + "learning_rate": 8.864509051014811e-05, + "loss": 0.4005, + "step": 419 + }, + { + "epoch": 0.34489837815643604, + "grad_norm": 0.08448805660009384, + "learning_rate": 8.861766319253978e-05, + "loss": 0.3327, + "step": 420 + }, + { + "epoch": 0.3457195647710942, + "grad_norm": 0.0865749716758728, + "learning_rate": 8.859023587493143e-05, + "loss": 0.4459, + "step": 421 + }, + { + "epoch": 0.3465407513857524, + "grad_norm": 0.08855796605348587, + "learning_rate": 8.85628085573231e-05, + "loss": 0.3566, + "step": 422 + }, + { + "epoch": 0.3473619380004106, + "grad_norm": 0.08421412855386734, + "learning_rate": 8.853538123971477e-05, + "loss": 0.3248, + "step": 423 + }, + { + "epoch": 0.3481831246150688, + "grad_norm": 0.07903390377759933, + "learning_rate": 8.850795392210642e-05, + "loss": 0.3608, + "step": 424 + }, + { + "epoch": 0.34900431122972697, + "grad_norm": 0.09979040920734406, + "learning_rate": 8.848052660449808e-05, + "loss": 0.3079, + "step": 425 + }, + { + "epoch": 0.34982549784438516, + "grad_norm": 0.08713112026453018, + "learning_rate": 8.845309928688975e-05, + "loss": 0.3462, + "step": 426 + }, + { + "epoch": 0.35064668445904335, + "grad_norm": 0.07836464792490005, + "learning_rate": 8.842567196928141e-05, + "loss": 0.3467, + "step": 427 + }, + { + "epoch": 0.3514678710737015, + "grad_norm": 0.07553427666425705, + "learning_rate": 8.839824465167307e-05, + "loss": 0.3325, + "step": 428 + }, + { + "epoch": 0.35228905768835966, + "grad_norm": 0.08221621066331863, + "learning_rate": 8.837081733406474e-05, + "loss": 0.3366, + "step": 429 + }, + { + "epoch": 0.35311024430301785, + "grad_norm": 0.09283772855997086, + "learning_rate": 8.834339001645639e-05, + "loss": 0.3518, + "step": 430 + }, + { + "epoch": 0.35393143091767604, + "grad_norm": 0.08067970722913742, + "learning_rate": 8.831596269884806e-05, + "loss": 0.3521, + "step": 431 + }, + { + "epoch": 0.3547526175323342, + "grad_norm": 0.08192427456378937, + "learning_rate": 8.828853538123973e-05, + "loss": 0.3794, + "step": 432 + }, + { + "epoch": 0.3555738041469924, + "grad_norm": 0.08332204073667526, + "learning_rate": 8.826110806363138e-05, + "loss": 0.3243, + "step": 433 + }, + { + "epoch": 0.3563949907616506, + "grad_norm": 0.08099368214607239, + "learning_rate": 8.823368074602304e-05, + "loss": 0.3409, + "step": 434 + }, + { + "epoch": 0.3572161773763088, + "grad_norm": 0.0801381841301918, + "learning_rate": 8.82062534284147e-05, + "loss": 0.3489, + "step": 435 + }, + { + "epoch": 0.358037363990967, + "grad_norm": 0.0828125849366188, + "learning_rate": 8.817882611080637e-05, + "loss": 0.3332, + "step": 436 + }, + { + "epoch": 0.3588585506056251, + "grad_norm": 0.08463995903730392, + "learning_rate": 8.815139879319803e-05, + "loss": 0.3532, + "step": 437 + }, + { + "epoch": 0.3596797372202833, + "grad_norm": 0.08150382339954376, + "learning_rate": 8.81239714755897e-05, + "loss": 0.3271, + "step": 438 + }, + { + "epoch": 0.3605009238349415, + "grad_norm": 0.08224772661924362, + "learning_rate": 8.809654415798135e-05, + "loss": 0.3847, + "step": 439 + }, + { + "epoch": 0.36132211044959966, + "grad_norm": 0.07509076595306396, + "learning_rate": 8.806911684037302e-05, + "loss": 0.346, + "step": 440 + }, + { + "epoch": 0.36132211044959966, + "eval_runtime": 493.5837, + "eval_samples_per_second": 0.399, + "eval_steps_per_second": 0.201, + "step": 440 + }, + { + "epoch": 0.36214329706425785, + "grad_norm": 0.07801686972379684, + "learning_rate": 8.804168952276468e-05, + "loss": 0.3352, + "step": 441 + }, + { + "epoch": 0.36296448367891604, + "grad_norm": 0.08195506781339645, + "learning_rate": 8.801426220515634e-05, + "loss": 0.3076, + "step": 442 + }, + { + "epoch": 0.3637856702935742, + "grad_norm": 0.08178213983774185, + "learning_rate": 8.798683488754799e-05, + "loss": 0.3565, + "step": 443 + }, + { + "epoch": 0.3646068569082324, + "grad_norm": 0.07905025035142899, + "learning_rate": 8.795940756993966e-05, + "loss": 0.3513, + "step": 444 + }, + { + "epoch": 0.3654280435228906, + "grad_norm": 0.1251908391714096, + "learning_rate": 8.793198025233133e-05, + "loss": 0.3656, + "step": 445 + }, + { + "epoch": 0.36624923013754873, + "grad_norm": 0.08813164383172989, + "learning_rate": 8.790455293472298e-05, + "loss": 0.3352, + "step": 446 + }, + { + "epoch": 0.3670704167522069, + "grad_norm": 0.0773930549621582, + "learning_rate": 8.787712561711465e-05, + "loss": 0.3241, + "step": 447 + }, + { + "epoch": 0.3678916033668651, + "grad_norm": 0.08977091312408447, + "learning_rate": 8.78496982995063e-05, + "loss": 0.3308, + "step": 448 + }, + { + "epoch": 0.3687127899815233, + "grad_norm": 0.0813109427690506, + "learning_rate": 8.782227098189797e-05, + "loss": 0.344, + "step": 449 + }, + { + "epoch": 0.3695339765961815, + "grad_norm": 0.08452528715133667, + "learning_rate": 8.779484366428964e-05, + "loss": 0.2985, + "step": 450 + }, + { + "epoch": 0.37035516321083967, + "grad_norm": 0.08880957961082458, + "learning_rate": 8.77674163466813e-05, + "loss": 0.3674, + "step": 451 + }, + { + "epoch": 0.37117634982549785, + "grad_norm": 0.08020122349262238, + "learning_rate": 8.773998902907295e-05, + "loss": 0.3487, + "step": 452 + }, + { + "epoch": 0.37199753644015604, + "grad_norm": 0.09180288016796112, + "learning_rate": 8.771256171146463e-05, + "loss": 0.364, + "step": 453 + }, + { + "epoch": 0.3728187230548142, + "grad_norm": 0.08132430911064148, + "learning_rate": 8.768513439385628e-05, + "loss": 0.3428, + "step": 454 + }, + { + "epoch": 0.3736399096694724, + "grad_norm": 0.08801679313182831, + "learning_rate": 8.765770707624795e-05, + "loss": 0.3287, + "step": 455 + }, + { + "epoch": 0.37446109628413055, + "grad_norm": 0.088813915848732, + "learning_rate": 8.763027975863961e-05, + "loss": 0.3508, + "step": 456 + }, + { + "epoch": 0.37528228289878873, + "grad_norm": 0.07454930245876312, + "learning_rate": 8.760285244103128e-05, + "loss": 0.3267, + "step": 457 + }, + { + "epoch": 0.3761034695134469, + "grad_norm": 0.0816444680094719, + "learning_rate": 8.757542512342294e-05, + "loss": 0.3237, + "step": 458 + }, + { + "epoch": 0.3769246561281051, + "grad_norm": 0.08916571736335754, + "learning_rate": 8.75479978058146e-05, + "loss": 0.3599, + "step": 459 + }, + { + "epoch": 0.3777458427427633, + "grad_norm": 0.08580256998538971, + "learning_rate": 8.752057048820625e-05, + "loss": 0.351, + "step": 460 + }, + { + "epoch": 0.3785670293574215, + "grad_norm": 0.0849752202630043, + "learning_rate": 8.749314317059792e-05, + "loss": 0.3616, + "step": 461 + }, + { + "epoch": 0.37938821597207967, + "grad_norm": 0.11230983585119247, + "learning_rate": 8.746571585298959e-05, + "loss": 0.3401, + "step": 462 + }, + { + "epoch": 0.38020940258673785, + "grad_norm": 0.08614380657672882, + "learning_rate": 8.743828853538124e-05, + "loss": 0.3603, + "step": 463 + }, + { + "epoch": 0.38103058920139604, + "grad_norm": 0.08322657644748688, + "learning_rate": 8.741086121777291e-05, + "loss": 0.3087, + "step": 464 + }, + { + "epoch": 0.3818517758160542, + "grad_norm": 0.08578125387430191, + "learning_rate": 8.738343390016456e-05, + "loss": 0.3483, + "step": 465 + }, + { + "epoch": 0.38267296243071236, + "grad_norm": 0.08472792059183121, + "learning_rate": 8.735600658255623e-05, + "loss": 0.3445, + "step": 466 + }, + { + "epoch": 0.38349414904537055, + "grad_norm": 0.08300035446882248, + "learning_rate": 8.73285792649479e-05, + "loss": 0.3384, + "step": 467 + }, + { + "epoch": 0.38431533566002873, + "grad_norm": 0.08582165092229843, + "learning_rate": 8.730115194733955e-05, + "loss": 0.3497, + "step": 468 + }, + { + "epoch": 0.3851365222746869, + "grad_norm": 0.28932541608810425, + "learning_rate": 8.727372462973121e-05, + "loss": 0.3091, + "step": 469 + }, + { + "epoch": 0.3859577088893451, + "grad_norm": 0.08396289497613907, + "learning_rate": 8.724629731212288e-05, + "loss": 0.3472, + "step": 470 + }, + { + "epoch": 0.3867788955040033, + "grad_norm": 0.08243526518344879, + "learning_rate": 8.721886999451454e-05, + "loss": 0.3641, + "step": 471 + }, + { + "epoch": 0.3876000821186615, + "grad_norm": 0.08014799654483795, + "learning_rate": 8.71914426769062e-05, + "loss": 0.3389, + "step": 472 + }, + { + "epoch": 0.38842126873331967, + "grad_norm": 0.08125128597021103, + "learning_rate": 8.716401535929787e-05, + "loss": 0.3293, + "step": 473 + }, + { + "epoch": 0.3892424553479778, + "grad_norm": 0.08730563521385193, + "learning_rate": 8.713658804168952e-05, + "loss": 0.3102, + "step": 474 + }, + { + "epoch": 0.390063641962636, + "grad_norm": 0.08945506811141968, + "learning_rate": 8.710916072408119e-05, + "loss": 0.3496, + "step": 475 + }, + { + "epoch": 0.3908848285772942, + "grad_norm": 0.09083441644906998, + "learning_rate": 8.708173340647286e-05, + "loss": 0.3411, + "step": 476 + }, + { + "epoch": 0.39170601519195236, + "grad_norm": 0.08667677640914917, + "learning_rate": 8.705430608886451e-05, + "loss": 0.3289, + "step": 477 + }, + { + "epoch": 0.39252720180661055, + "grad_norm": 0.08793242275714874, + "learning_rate": 8.702687877125617e-05, + "loss": 0.3542, + "step": 478 + }, + { + "epoch": 0.39334838842126874, + "grad_norm": 0.08769497275352478, + "learning_rate": 8.699945145364785e-05, + "loss": 0.3207, + "step": 479 + }, + { + "epoch": 0.3941695750359269, + "grad_norm": 0.08303678780794144, + "learning_rate": 8.69720241360395e-05, + "loss": 0.343, + "step": 480 + }, + { + "epoch": 0.3949907616505851, + "grad_norm": 0.07943850755691528, + "learning_rate": 8.694459681843116e-05, + "loss": 0.3273, + "step": 481 + }, + { + "epoch": 0.3958119482652433, + "grad_norm": 0.08450903743505478, + "learning_rate": 8.691716950082282e-05, + "loss": 0.3895, + "step": 482 + }, + { + "epoch": 0.3966331348799015, + "grad_norm": 0.07759539037942886, + "learning_rate": 8.688974218321448e-05, + "loss": 0.3426, + "step": 483 + }, + { + "epoch": 0.3974543214945596, + "grad_norm": 0.09013447165489197, + "learning_rate": 8.686231486560615e-05, + "loss": 0.336, + "step": 484 + }, + { + "epoch": 0.3982755081092178, + "grad_norm": 0.08396545797586441, + "learning_rate": 8.683488754799781e-05, + "loss": 0.3647, + "step": 485 + }, + { + "epoch": 0.399096694723876, + "grad_norm": 0.08143208920955658, + "learning_rate": 8.680746023038947e-05, + "loss": 0.3239, + "step": 486 + }, + { + "epoch": 0.3999178813385342, + "grad_norm": 0.07628796994686127, + "learning_rate": 8.678003291278114e-05, + "loss": 0.3389, + "step": 487 + }, + { + "epoch": 0.40073906795319236, + "grad_norm": 0.08144954591989517, + "learning_rate": 8.67526055951728e-05, + "loss": 0.3258, + "step": 488 + }, + { + "epoch": 0.40156025456785055, + "grad_norm": 0.08240176737308502, + "learning_rate": 8.672517827756446e-05, + "loss": 0.3316, + "step": 489 + }, + { + "epoch": 0.40238144118250874, + "grad_norm": 0.09237710386514664, + "learning_rate": 8.669775095995613e-05, + "loss": 0.3435, + "step": 490 + }, + { + "epoch": 0.4032026277971669, + "grad_norm": 0.07773042470216751, + "learning_rate": 8.667032364234778e-05, + "loss": 0.3479, + "step": 491 + }, + { + "epoch": 0.4040238144118251, + "grad_norm": 0.08410782366991043, + "learning_rate": 8.664289632473945e-05, + "loss": 0.3792, + "step": 492 + }, + { + "epoch": 0.40484500102648324, + "grad_norm": 0.08251044154167175, + "learning_rate": 8.661546900713112e-05, + "loss": 0.3388, + "step": 493 + }, + { + "epoch": 0.40566618764114143, + "grad_norm": 0.08804575353860855, + "learning_rate": 8.658804168952277e-05, + "loss": 0.3388, + "step": 494 + }, + { + "epoch": 0.4064873742557996, + "grad_norm": 0.10321494936943054, + "learning_rate": 8.656061437191442e-05, + "loss": 0.3348, + "step": 495 + }, + { + "epoch": 0.4073085608704578, + "grad_norm": 0.07642900943756104, + "learning_rate": 8.653318705430609e-05, + "loss": 0.3155, + "step": 496 + }, + { + "epoch": 0.408129747485116, + "grad_norm": 0.08629249781370163, + "learning_rate": 8.650575973669776e-05, + "loss": 0.3303, + "step": 497 + }, + { + "epoch": 0.4089509340997742, + "grad_norm": 0.07933122664690018, + "learning_rate": 8.647833241908941e-05, + "loss": 0.3399, + "step": 498 + }, + { + "epoch": 0.40977212071443236, + "grad_norm": 0.0802338495850563, + "learning_rate": 8.645090510148108e-05, + "loss": 0.3623, + "step": 499 + }, + { + "epoch": 0.41059330732909055, + "grad_norm": 0.07566798478364944, + "learning_rate": 8.642347778387274e-05, + "loss": 0.3195, + "step": 500 + }, + { + "epoch": 0.41141449394374874, + "grad_norm": 0.0790770947933197, + "learning_rate": 8.63960504662644e-05, + "loss": 0.3478, + "step": 501 + }, + { + "epoch": 0.4122356805584069, + "grad_norm": 0.08219192922115326, + "learning_rate": 8.636862314865607e-05, + "loss": 0.4037, + "step": 502 + }, + { + "epoch": 0.41305686717306506, + "grad_norm": 0.09459681063890457, + "learning_rate": 8.634119583104773e-05, + "loss": 0.3604, + "step": 503 + }, + { + "epoch": 0.41387805378772324, + "grad_norm": 0.08201157301664352, + "learning_rate": 8.631376851343938e-05, + "loss": 0.3561, + "step": 504 + }, + { + "epoch": 0.41469924040238143, + "grad_norm": 0.08181356638669968, + "learning_rate": 8.628634119583105e-05, + "loss": 0.3359, + "step": 505 + }, + { + "epoch": 0.4155204270170396, + "grad_norm": 0.08242693543434143, + "learning_rate": 8.625891387822272e-05, + "loss": 0.3501, + "step": 506 + }, + { + "epoch": 0.4163416136316978, + "grad_norm": 0.08045560866594315, + "learning_rate": 8.623148656061437e-05, + "loss": 0.3188, + "step": 507 + }, + { + "epoch": 0.417162800246356, + "grad_norm": 0.08392325043678284, + "learning_rate": 8.620405924300604e-05, + "loss": 0.3607, + "step": 508 + }, + { + "epoch": 0.4179839868610142, + "grad_norm": 0.0920289009809494, + "learning_rate": 8.61766319253977e-05, + "loss": 0.3425, + "step": 509 + }, + { + "epoch": 0.41880517347567237, + "grad_norm": 0.08921059966087341, + "learning_rate": 8.614920460778936e-05, + "loss": 0.3386, + "step": 510 + }, + { + "epoch": 0.41962636009033055, + "grad_norm": 0.08129517734050751, + "learning_rate": 8.612177729018103e-05, + "loss": 0.3631, + "step": 511 + }, + { + "epoch": 0.4204475467049887, + "grad_norm": 0.19061192870140076, + "learning_rate": 8.609434997257268e-05, + "loss": 0.3285, + "step": 512 + }, + { + "epoch": 0.42126873331964687, + "grad_norm": 0.08446541428565979, + "learning_rate": 8.606692265496434e-05, + "loss": 0.4133, + "step": 513 + }, + { + "epoch": 0.42208991993430506, + "grad_norm": 0.07793641835451126, + "learning_rate": 8.603949533735602e-05, + "loss": 0.3166, + "step": 514 + }, + { + "epoch": 0.42291110654896324, + "grad_norm": 0.07862336188554764, + "learning_rate": 8.601206801974767e-05, + "loss": 0.3448, + "step": 515 + }, + { + "epoch": 0.42373229316362143, + "grad_norm": 0.07776626199483871, + "learning_rate": 8.598464070213933e-05, + "loss": 0.3717, + "step": 516 + }, + { + "epoch": 0.4245534797782796, + "grad_norm": 0.08171868324279785, + "learning_rate": 8.5957213384531e-05, + "loss": 0.3384, + "step": 517 + }, + { + "epoch": 0.4253746663929378, + "grad_norm": 0.08469153940677643, + "learning_rate": 8.592978606692266e-05, + "loss": 0.3238, + "step": 518 + }, + { + "epoch": 0.426195853007596, + "grad_norm": 0.0739326924085617, + "learning_rate": 8.590235874931433e-05, + "loss": 0.3033, + "step": 519 + }, + { + "epoch": 0.4270170396222542, + "grad_norm": 0.08084508031606674, + "learning_rate": 8.587493143170599e-05, + "loss": 0.3557, + "step": 520 + }, + { + "epoch": 0.4278382262369123, + "grad_norm": 0.079572893679142, + "learning_rate": 8.584750411409764e-05, + "loss": 0.3369, + "step": 521 + }, + { + "epoch": 0.4286594128515705, + "grad_norm": 0.09056146442890167, + "learning_rate": 8.582007679648931e-05, + "loss": 0.3362, + "step": 522 + }, + { + "epoch": 0.4294805994662287, + "grad_norm": 0.09055227041244507, + "learning_rate": 8.579264947888098e-05, + "loss": 0.3457, + "step": 523 + }, + { + "epoch": 0.43030178608088687, + "grad_norm": 0.11645710468292236, + "learning_rate": 8.576522216127263e-05, + "loss": 0.3378, + "step": 524 + }, + { + "epoch": 0.43112297269554506, + "grad_norm": 0.08169250190258026, + "learning_rate": 8.57377948436643e-05, + "loss": 0.3406, + "step": 525 + }, + { + "epoch": 0.43194415931020325, + "grad_norm": 0.08108235895633698, + "learning_rate": 8.571036752605595e-05, + "loss": 0.3253, + "step": 526 + }, + { + "epoch": 0.43276534592486143, + "grad_norm": 0.07971398532390594, + "learning_rate": 8.568294020844762e-05, + "loss": 0.3389, + "step": 527 + }, + { + "epoch": 0.4335865325395196, + "grad_norm": 0.08281555771827698, + "learning_rate": 8.565551289083929e-05, + "loss": 0.3289, + "step": 528 + }, + { + "epoch": 0.4335865325395196, + "eval_runtime": 493.1791, + "eval_samples_per_second": 0.399, + "eval_steps_per_second": 0.201, + "step": 528 + }, + { + "epoch": 0.4344077191541778, + "grad_norm": 0.08380738645792007, + "learning_rate": 8.562808557323094e-05, + "loss": 0.3445, + "step": 529 + }, + { + "epoch": 0.435228905768836, + "grad_norm": 0.07987751066684723, + "learning_rate": 8.56006582556226e-05, + "loss": 0.3149, + "step": 530 + }, + { + "epoch": 0.4360500923834941, + "grad_norm": 0.08102700114250183, + "learning_rate": 8.557323093801427e-05, + "loss": 0.3309, + "step": 531 + }, + { + "epoch": 0.4368712789981523, + "grad_norm": 0.08056002855300903, + "learning_rate": 8.554580362040593e-05, + "loss": 0.3579, + "step": 532 + }, + { + "epoch": 0.4376924656128105, + "grad_norm": 0.08129940181970596, + "learning_rate": 8.551837630279759e-05, + "loss": 0.3414, + "step": 533 + }, + { + "epoch": 0.4385136522274687, + "grad_norm": 0.07720111310482025, + "learning_rate": 8.549094898518926e-05, + "loss": 0.3066, + "step": 534 + }, + { + "epoch": 0.4393348388421269, + "grad_norm": 0.08578766137361526, + "learning_rate": 8.546352166758091e-05, + "loss": 0.4056, + "step": 535 + }, + { + "epoch": 0.44015602545678506, + "grad_norm": 0.14550408720970154, + "learning_rate": 8.543609434997258e-05, + "loss": 0.353, + "step": 536 + }, + { + "epoch": 0.44097721207144325, + "grad_norm": 0.07624532282352448, + "learning_rate": 8.540866703236425e-05, + "loss": 0.323, + "step": 537 + }, + { + "epoch": 0.44179839868610143, + "grad_norm": 0.07774993777275085, + "learning_rate": 8.53812397147559e-05, + "loss": 0.3244, + "step": 538 + }, + { + "epoch": 0.4426195853007596, + "grad_norm": 0.08061113953590393, + "learning_rate": 8.535381239714755e-05, + "loss": 0.3209, + "step": 539 + }, + { + "epoch": 0.44344077191541775, + "grad_norm": 0.07771355658769608, + "learning_rate": 8.532638507953922e-05, + "loss": 0.3604, + "step": 540 + }, + { + "epoch": 0.44426195853007594, + "grad_norm": 0.08160841464996338, + "learning_rate": 8.529895776193089e-05, + "loss": 0.3373, + "step": 541 + }, + { + "epoch": 0.4450831451447341, + "grad_norm": 0.0819459930062294, + "learning_rate": 8.527153044432254e-05, + "loss": 0.3097, + "step": 542 + }, + { + "epoch": 0.4459043317593923, + "grad_norm": 0.08429809659719467, + "learning_rate": 8.524410312671421e-05, + "loss": 0.3538, + "step": 543 + }, + { + "epoch": 0.4467255183740505, + "grad_norm": 0.08519409596920013, + "learning_rate": 8.521667580910587e-05, + "loss": 0.3486, + "step": 544 + }, + { + "epoch": 0.4475467049887087, + "grad_norm": 0.08900930732488632, + "learning_rate": 8.518924849149753e-05, + "loss": 0.3339, + "step": 545 + }, + { + "epoch": 0.4483678916033669, + "grad_norm": 0.08914072066545486, + "learning_rate": 8.51618211738892e-05, + "loss": 0.2933, + "step": 546 + }, + { + "epoch": 0.44918907821802506, + "grad_norm": 0.08086919039487839, + "learning_rate": 8.513439385628086e-05, + "loss": 0.3367, + "step": 547 + }, + { + "epoch": 0.45001026483268325, + "grad_norm": 0.08358173817396164, + "learning_rate": 8.510696653867251e-05, + "loss": 0.3443, + "step": 548 + }, + { + "epoch": 0.45083145144734144, + "grad_norm": 0.0807274803519249, + "learning_rate": 8.507953922106419e-05, + "loss": 0.3332, + "step": 549 + }, + { + "epoch": 0.45165263806199957, + "grad_norm": 0.08126116544008255, + "learning_rate": 8.505211190345585e-05, + "loss": 0.3485, + "step": 550 + }, + { + "epoch": 0.45247382467665775, + "grad_norm": 0.08835196495056152, + "learning_rate": 8.50246845858475e-05, + "loss": 0.3739, + "step": 551 + }, + { + "epoch": 0.45329501129131594, + "grad_norm": 0.07614720612764359, + "learning_rate": 8.499725726823917e-05, + "loss": 0.3464, + "step": 552 + }, + { + "epoch": 0.45411619790597413, + "grad_norm": 0.07444674521684647, + "learning_rate": 8.496982995063084e-05, + "loss": 0.3553, + "step": 553 + }, + { + "epoch": 0.4549373845206323, + "grad_norm": 0.0741548240184784, + "learning_rate": 8.49424026330225e-05, + "loss": 0.341, + "step": 554 + }, + { + "epoch": 0.4557585711352905, + "grad_norm": 0.0820796862244606, + "learning_rate": 8.491497531541416e-05, + "loss": 0.3359, + "step": 555 + }, + { + "epoch": 0.4565797577499487, + "grad_norm": 0.08371058106422424, + "learning_rate": 8.488754799780581e-05, + "loss": 0.3339, + "step": 556 + }, + { + "epoch": 0.4574009443646069, + "grad_norm": 0.08354955166578293, + "learning_rate": 8.486012068019748e-05, + "loss": 0.3619, + "step": 557 + }, + { + "epoch": 0.45822213097926506, + "grad_norm": 0.08976832777261734, + "learning_rate": 8.483269336258915e-05, + "loss": 0.353, + "step": 558 + }, + { + "epoch": 0.4590433175939232, + "grad_norm": 0.07933076471090317, + "learning_rate": 8.48052660449808e-05, + "loss": 0.3764, + "step": 559 + }, + { + "epoch": 0.4598645042085814, + "grad_norm": 0.08624580502510071, + "learning_rate": 8.477783872737247e-05, + "loss": 0.3372, + "step": 560 + }, + { + "epoch": 0.46068569082323957, + "grad_norm": 0.08301227539777756, + "learning_rate": 8.475041140976413e-05, + "loss": 0.3748, + "step": 561 + }, + { + "epoch": 0.46150687743789776, + "grad_norm": 0.07408706098794937, + "learning_rate": 8.47229840921558e-05, + "loss": 0.2622, + "step": 562 + }, + { + "epoch": 0.46232806405255594, + "grad_norm": 0.07799118012189865, + "learning_rate": 8.469555677454746e-05, + "loss": 0.3614, + "step": 563 + }, + { + "epoch": 0.46314925066721413, + "grad_norm": 0.08185340464115143, + "learning_rate": 8.466812945693912e-05, + "loss": 0.3423, + "step": 564 + }, + { + "epoch": 0.4639704372818723, + "grad_norm": 0.0817733034491539, + "learning_rate": 8.464070213933077e-05, + "loss": 0.3519, + "step": 565 + }, + { + "epoch": 0.4647916238965305, + "grad_norm": 0.08026120066642761, + "learning_rate": 8.461327482172244e-05, + "loss": 0.3372, + "step": 566 + }, + { + "epoch": 0.4656128105111887, + "grad_norm": 0.07413862645626068, + "learning_rate": 8.45858475041141e-05, + "loss": 0.3017, + "step": 567 + }, + { + "epoch": 0.4664339971258468, + "grad_norm": 0.09256873279809952, + "learning_rate": 8.455842018650576e-05, + "loss": 0.3495, + "step": 568 + }, + { + "epoch": 0.467255183740505, + "grad_norm": 0.08319979161024094, + "learning_rate": 8.453099286889743e-05, + "loss": 0.3339, + "step": 569 + }, + { + "epoch": 0.4680763703551632, + "grad_norm": 0.08824891597032547, + "learning_rate": 8.450356555128908e-05, + "loss": 0.3272, + "step": 570 + }, + { + "epoch": 0.4688975569698214, + "grad_norm": 0.0790686085820198, + "learning_rate": 8.447613823368075e-05, + "loss": 0.3446, + "step": 571 + }, + { + "epoch": 0.46971874358447957, + "grad_norm": 0.08730736374855042, + "learning_rate": 8.444871091607242e-05, + "loss": 0.3731, + "step": 572 + }, + { + "epoch": 0.47053993019913776, + "grad_norm": 0.08023872971534729, + "learning_rate": 8.442128359846407e-05, + "loss": 0.3226, + "step": 573 + }, + { + "epoch": 0.47136111681379594, + "grad_norm": 0.09242340922355652, + "learning_rate": 8.439385628085573e-05, + "loss": 0.3229, + "step": 574 + }, + { + "epoch": 0.47218230342845413, + "grad_norm": 0.07871225476264954, + "learning_rate": 8.43664289632474e-05, + "loss": 0.3112, + "step": 575 + }, + { + "epoch": 0.4730034900431123, + "grad_norm": 0.08368773013353348, + "learning_rate": 8.433900164563906e-05, + "loss": 0.3357, + "step": 576 + }, + { + "epoch": 0.4738246766577705, + "grad_norm": 0.07561603933572769, + "learning_rate": 8.431157432803072e-05, + "loss": 0.3125, + "step": 577 + }, + { + "epoch": 0.47464586327242864, + "grad_norm": 0.08441564440727234, + "learning_rate": 8.428414701042238e-05, + "loss": 0.3211, + "step": 578 + }, + { + "epoch": 0.4754670498870868, + "grad_norm": 0.09772761911153793, + "learning_rate": 8.425671969281404e-05, + "loss": 0.3285, + "step": 579 + }, + { + "epoch": 0.476288236501745, + "grad_norm": 0.07900305837392807, + "learning_rate": 8.422929237520571e-05, + "loss": 0.344, + "step": 580 + }, + { + "epoch": 0.4771094231164032, + "grad_norm": 0.13551943004131317, + "learning_rate": 8.420186505759737e-05, + "loss": 0.3515, + "step": 581 + }, + { + "epoch": 0.4779306097310614, + "grad_norm": 0.07592803239822388, + "learning_rate": 8.417443773998903e-05, + "loss": 0.3377, + "step": 582 + }, + { + "epoch": 0.47875179634571957, + "grad_norm": 0.08301271498203278, + "learning_rate": 8.414701042238068e-05, + "loss": 0.365, + "step": 583 + }, + { + "epoch": 0.47957298296037776, + "grad_norm": 0.11750177294015884, + "learning_rate": 8.411958310477237e-05, + "loss": 0.3159, + "step": 584 + }, + { + "epoch": 0.48039416957503595, + "grad_norm": 0.08506017178297043, + "learning_rate": 8.409215578716402e-05, + "loss": 0.3381, + "step": 585 + }, + { + "epoch": 0.48121535618969413, + "grad_norm": 0.0827595517039299, + "learning_rate": 8.406472846955569e-05, + "loss": 0.3497, + "step": 586 + }, + { + "epoch": 0.48203654280435226, + "grad_norm": 0.07703883200883865, + "learning_rate": 8.403730115194734e-05, + "loss": 0.2989, + "step": 587 + }, + { + "epoch": 0.48285772941901045, + "grad_norm": 0.07976412773132324, + "learning_rate": 8.400987383433901e-05, + "loss": 0.3617, + "step": 588 + }, + { + "epoch": 0.48367891603366864, + "grad_norm": 0.09424680471420288, + "learning_rate": 8.398244651673068e-05, + "loss": 0.3623, + "step": 589 + }, + { + "epoch": 0.4845001026483268, + "grad_norm": 0.08831481635570526, + "learning_rate": 8.395501919912233e-05, + "loss": 0.3526, + "step": 590 + }, + { + "epoch": 0.485321289262985, + "grad_norm": 0.07811284810304642, + "learning_rate": 8.392759188151399e-05, + "loss": 0.3721, + "step": 591 + }, + { + "epoch": 0.4861424758776432, + "grad_norm": 0.08139461278915405, + "learning_rate": 8.390016456390565e-05, + "loss": 0.33, + "step": 592 + }, + { + "epoch": 0.4869636624923014, + "grad_norm": 0.0807470753788948, + "learning_rate": 8.387273724629732e-05, + "loss": 0.372, + "step": 593 + }, + { + "epoch": 0.4877848491069596, + "grad_norm": 0.08856850117444992, + "learning_rate": 8.384530992868898e-05, + "loss": 0.3288, + "step": 594 + }, + { + "epoch": 0.48860603572161776, + "grad_norm": 0.08670219779014587, + "learning_rate": 8.381788261108064e-05, + "loss": 0.4113, + "step": 595 + }, + { + "epoch": 0.4894272223362759, + "grad_norm": 0.0814259722828865, + "learning_rate": 8.37904552934723e-05, + "loss": 0.3735, + "step": 596 + }, + { + "epoch": 0.4902484089509341, + "grad_norm": 0.08457124978303909, + "learning_rate": 8.376302797586397e-05, + "loss": 0.3516, + "step": 597 + }, + { + "epoch": 0.49106959556559227, + "grad_norm": 0.0774846151471138, + "learning_rate": 8.373560065825563e-05, + "loss": 0.3604, + "step": 598 + }, + { + "epoch": 0.49189078218025045, + "grad_norm": 0.07951226830482483, + "learning_rate": 8.370817334064729e-05, + "loss": 0.3016, + "step": 599 + }, + { + "epoch": 0.49271196879490864, + "grad_norm": 0.08235177397727966, + "learning_rate": 8.368074602303894e-05, + "loss": 0.3502, + "step": 600 + }, + { + "epoch": 0.4935331554095668, + "grad_norm": 0.08112053573131561, + "learning_rate": 8.365331870543061e-05, + "loss": 0.3122, + "step": 601 + }, + { + "epoch": 0.494354342024225, + "grad_norm": 0.07517421990633011, + "learning_rate": 8.362589138782228e-05, + "loss": 0.3482, + "step": 602 + }, + { + "epoch": 0.4951755286388832, + "grad_norm": 0.08243374526500702, + "learning_rate": 8.359846407021393e-05, + "loss": 0.3455, + "step": 603 + }, + { + "epoch": 0.4959967152535414, + "grad_norm": 0.08028637617826462, + "learning_rate": 8.35710367526056e-05, + "loss": 0.3454, + "step": 604 + }, + { + "epoch": 0.4968179018681996, + "grad_norm": 0.08006370067596436, + "learning_rate": 8.354360943499726e-05, + "loss": 0.3216, + "step": 605 + }, + { + "epoch": 0.4976390884828577, + "grad_norm": 0.1940547525882721, + "learning_rate": 8.351618211738892e-05, + "loss": 0.3017, + "step": 606 + }, + { + "epoch": 0.4984602750975159, + "grad_norm": 0.07875222712755203, + "learning_rate": 8.348875479978059e-05, + "loss": 0.3295, + "step": 607 + }, + { + "epoch": 0.4992814617121741, + "grad_norm": 0.07511032372713089, + "learning_rate": 8.346132748217225e-05, + "loss": 0.328, + "step": 608 + }, + { + "epoch": 0.5001026483268323, + "grad_norm": 0.08197803795337677, + "learning_rate": 8.34339001645639e-05, + "loss": 0.3232, + "step": 609 + }, + { + "epoch": 0.5009238349414905, + "grad_norm": 0.08152435719966888, + "learning_rate": 8.340647284695558e-05, + "loss": 0.3546, + "step": 610 + }, + { + "epoch": 0.5017450215561486, + "grad_norm": 0.0796707347035408, + "learning_rate": 8.337904552934724e-05, + "loss": 0.3518, + "step": 611 + }, + { + "epoch": 0.5025662081708068, + "grad_norm": 0.07969588786363602, + "learning_rate": 8.335161821173889e-05, + "loss": 0.3195, + "step": 612 + }, + { + "epoch": 0.503387394785465, + "grad_norm": 0.08270443230867386, + "learning_rate": 8.332419089413056e-05, + "loss": 0.3244, + "step": 613 + }, + { + "epoch": 0.5042085814001231, + "grad_norm": 0.08303502202033997, + "learning_rate": 8.329676357652221e-05, + "loss": 0.3386, + "step": 614 + }, + { + "epoch": 0.5050297680147814, + "grad_norm": 0.07783270627260208, + "learning_rate": 8.326933625891388e-05, + "loss": 0.3046, + "step": 615 + }, + { + "epoch": 0.5058509546294395, + "grad_norm": 0.08112157136201859, + "learning_rate": 8.324190894130555e-05, + "loss": 0.3244, + "step": 616 + }, + { + "epoch": 0.5058509546294395, + "eval_runtime": 494.4481, + "eval_samples_per_second": 0.398, + "eval_steps_per_second": 0.2, + "step": 616 + }, + { + "epoch": 0.5066721412440978, + "grad_norm": 0.08969878405332565, + "learning_rate": 8.32144816236972e-05, + "loss": 0.3365, + "step": 617 + }, + { + "epoch": 0.5074933278587559, + "grad_norm": 0.08142217993736267, + "learning_rate": 8.318705430608886e-05, + "loss": 0.3413, + "step": 618 + }, + { + "epoch": 0.5083145144734141, + "grad_norm": 0.08319269865751266, + "learning_rate": 8.315962698848054e-05, + "loss": 0.3458, + "step": 619 + }, + { + "epoch": 0.5091357010880723, + "grad_norm": 0.08104688674211502, + "learning_rate": 8.313219967087219e-05, + "loss": 0.3286, + "step": 620 + }, + { + "epoch": 0.5099568877027304, + "grad_norm": 0.08447278290987015, + "learning_rate": 8.310477235326386e-05, + "loss": 0.3485, + "step": 621 + }, + { + "epoch": 0.5107780743173886, + "grad_norm": 0.08879225701093674, + "learning_rate": 8.307734503565551e-05, + "loss": 0.2986, + "step": 622 + }, + { + "epoch": 0.5115992609320468, + "grad_norm": 0.07878706604242325, + "learning_rate": 8.304991771804718e-05, + "loss": 0.3301, + "step": 623 + }, + { + "epoch": 0.512420447546705, + "grad_norm": 0.08089771121740341, + "learning_rate": 8.302249040043885e-05, + "loss": 0.3183, + "step": 624 + }, + { + "epoch": 0.5132416341613631, + "grad_norm": 0.08130493015050888, + "learning_rate": 8.29950630828305e-05, + "loss": 0.3559, + "step": 625 + }, + { + "epoch": 0.5140628207760214, + "grad_norm": 0.0880448967218399, + "learning_rate": 8.296763576522216e-05, + "loss": 0.3567, + "step": 626 + }, + { + "epoch": 0.5148840073906795, + "grad_norm": 0.087735615670681, + "learning_rate": 8.294020844761383e-05, + "loss": 0.3123, + "step": 627 + }, + { + "epoch": 0.5157051940053378, + "grad_norm": 0.08659686893224716, + "learning_rate": 8.29127811300055e-05, + "loss": 0.3212, + "step": 628 + }, + { + "epoch": 0.5165263806199959, + "grad_norm": 0.0853281244635582, + "learning_rate": 8.288535381239715e-05, + "loss": 0.3337, + "step": 629 + }, + { + "epoch": 0.517347567234654, + "grad_norm": 0.07861575484275818, + "learning_rate": 8.285792649478882e-05, + "loss": 0.3666, + "step": 630 + }, + { + "epoch": 0.5181687538493123, + "grad_norm": 0.08911813795566559, + "learning_rate": 8.283049917718047e-05, + "loss": 0.349, + "step": 631 + }, + { + "epoch": 0.5189899404639704, + "grad_norm": 0.08234114944934845, + "learning_rate": 8.280307185957214e-05, + "loss": 0.3354, + "step": 632 + }, + { + "epoch": 0.5198111270786286, + "grad_norm": 0.08694402128458023, + "learning_rate": 8.277564454196381e-05, + "loss": 0.3321, + "step": 633 + }, + { + "epoch": 0.5206323136932868, + "grad_norm": 0.08186987787485123, + "learning_rate": 8.274821722435546e-05, + "loss": 0.3595, + "step": 634 + }, + { + "epoch": 0.521453500307945, + "grad_norm": 0.08425264060497284, + "learning_rate": 8.272078990674712e-05, + "loss": 0.3641, + "step": 635 + }, + { + "epoch": 0.5222746869226031, + "grad_norm": 0.1728411465883255, + "learning_rate": 8.269336258913878e-05, + "loss": 0.3646, + "step": 636 + }, + { + "epoch": 0.5230958735372614, + "grad_norm": 0.09049440920352936, + "learning_rate": 8.266593527153045e-05, + "loss": 0.3368, + "step": 637 + }, + { + "epoch": 0.5239170601519195, + "grad_norm": 0.081216000020504, + "learning_rate": 8.26385079539221e-05, + "loss": 0.3814, + "step": 638 + }, + { + "epoch": 0.5247382467665777, + "grad_norm": 0.08497685194015503, + "learning_rate": 8.261108063631377e-05, + "loss": 0.3315, + "step": 639 + }, + { + "epoch": 0.5255594333812359, + "grad_norm": 0.08426640927791595, + "learning_rate": 8.258365331870543e-05, + "loss": 0.3366, + "step": 640 + }, + { + "epoch": 0.526380619995894, + "grad_norm": 0.08354829251766205, + "learning_rate": 8.25562260010971e-05, + "loss": 0.3439, + "step": 641 + }, + { + "epoch": 0.5272018066105523, + "grad_norm": 0.08015681803226471, + "learning_rate": 8.252879868348876e-05, + "loss": 0.3401, + "step": 642 + }, + { + "epoch": 0.5280229932252104, + "grad_norm": 0.08374017477035522, + "learning_rate": 8.250137136588042e-05, + "loss": 0.3533, + "step": 643 + }, + { + "epoch": 0.5288441798398686, + "grad_norm": 0.09182158857584, + "learning_rate": 8.247394404827207e-05, + "loss": 0.2983, + "step": 644 + }, + { + "epoch": 0.5296653664545268, + "grad_norm": 0.08268212527036667, + "learning_rate": 8.244651673066375e-05, + "loss": 0.331, + "step": 645 + }, + { + "epoch": 0.530486553069185, + "grad_norm": 0.08059857040643692, + "learning_rate": 8.241908941305541e-05, + "loss": 0.3448, + "step": 646 + }, + { + "epoch": 0.5313077396838431, + "grad_norm": 0.08811867982149124, + "learning_rate": 8.239166209544706e-05, + "loss": 0.3678, + "step": 647 + }, + { + "epoch": 0.5321289262985013, + "grad_norm": 0.07894317060709, + "learning_rate": 8.236423477783873e-05, + "loss": 0.3456, + "step": 648 + }, + { + "epoch": 0.5329501129131595, + "grad_norm": 0.07806131988763809, + "learning_rate": 8.23368074602304e-05, + "loss": 0.3286, + "step": 649 + }, + { + "epoch": 0.5337712995278177, + "grad_norm": 0.08501176536083221, + "learning_rate": 8.230938014262205e-05, + "loss": 0.3383, + "step": 650 + }, + { + "epoch": 0.5345924861424759, + "grad_norm": 0.12988948822021484, + "learning_rate": 8.228195282501372e-05, + "loss": 0.3384, + "step": 651 + }, + { + "epoch": 0.535413672757134, + "grad_norm": 0.08005869388580322, + "learning_rate": 8.225452550740537e-05, + "loss": 0.323, + "step": 652 + }, + { + "epoch": 0.5362348593717923, + "grad_norm": 0.07902330160140991, + "learning_rate": 8.222709818979704e-05, + "loss": 0.3248, + "step": 653 + }, + { + "epoch": 0.5370560459864504, + "grad_norm": 0.08319025486707687, + "learning_rate": 8.219967087218871e-05, + "loss": 0.3145, + "step": 654 + }, + { + "epoch": 0.5378772326011086, + "grad_norm": 0.08141138404607773, + "learning_rate": 8.217224355458037e-05, + "loss": 0.3105, + "step": 655 + }, + { + "epoch": 0.5386984192157668, + "grad_norm": 0.07592695206403732, + "learning_rate": 8.214481623697203e-05, + "loss": 0.3184, + "step": 656 + }, + { + "epoch": 0.5395196058304249, + "grad_norm": 0.16522720456123352, + "learning_rate": 8.211738891936369e-05, + "loss": 0.3492, + "step": 657 + }, + { + "epoch": 0.5403407924450832, + "grad_norm": 0.0965307429432869, + "learning_rate": 8.208996160175536e-05, + "loss": 0.3012, + "step": 658 + }, + { + "epoch": 0.5411619790597413, + "grad_norm": 0.08132067322731018, + "learning_rate": 8.206253428414702e-05, + "loss": 0.3256, + "step": 659 + }, + { + "epoch": 0.5419831656743995, + "grad_norm": 0.08183111250400543, + "learning_rate": 8.203510696653868e-05, + "loss": 0.3591, + "step": 660 + }, + { + "epoch": 0.5428043522890577, + "grad_norm": 0.07674401253461838, + "learning_rate": 8.200767964893033e-05, + "loss": 0.3273, + "step": 661 + }, + { + "epoch": 0.5436255389037159, + "grad_norm": 0.07937440276145935, + "learning_rate": 8.1980252331322e-05, + "loss": 0.3028, + "step": 662 + }, + { + "epoch": 0.544446725518374, + "grad_norm": 0.07886291295289993, + "learning_rate": 8.195282501371367e-05, + "loss": 0.3452, + "step": 663 + }, + { + "epoch": 0.5452679121330323, + "grad_norm": 0.078894704580307, + "learning_rate": 8.192539769610532e-05, + "loss": 0.3447, + "step": 664 + }, + { + "epoch": 0.5460890987476904, + "grad_norm": 0.0840119943022728, + "learning_rate": 8.189797037849699e-05, + "loss": 0.3348, + "step": 665 + }, + { + "epoch": 0.5469102853623486, + "grad_norm": 0.08318603038787842, + "learning_rate": 8.187054306088864e-05, + "loss": 0.3523, + "step": 666 + }, + { + "epoch": 0.5477314719770068, + "grad_norm": 0.08060989528894424, + "learning_rate": 8.184311574328031e-05, + "loss": 0.3211, + "step": 667 + }, + { + "epoch": 0.5485526585916649, + "grad_norm": 0.07878737151622772, + "learning_rate": 8.181568842567198e-05, + "loss": 0.3865, + "step": 668 + }, + { + "epoch": 0.5493738452063232, + "grad_norm": 0.08277317881584167, + "learning_rate": 8.178826110806363e-05, + "loss": 0.3507, + "step": 669 + }, + { + "epoch": 0.5501950318209813, + "grad_norm": 0.15883035957813263, + "learning_rate": 8.176083379045529e-05, + "loss": 0.3428, + "step": 670 + }, + { + "epoch": 0.5510162184356395, + "grad_norm": 0.08117318898439407, + "learning_rate": 8.173340647284696e-05, + "loss": 0.3297, + "step": 671 + }, + { + "epoch": 0.5518374050502977, + "grad_norm": 0.07886844873428345, + "learning_rate": 8.170597915523862e-05, + "loss": 0.3489, + "step": 672 + }, + { + "epoch": 0.5526585916649559, + "grad_norm": 0.0774691253900528, + "learning_rate": 8.167855183763028e-05, + "loss": 0.3438, + "step": 673 + }, + { + "epoch": 0.553479778279614, + "grad_norm": 0.07944227755069733, + "learning_rate": 8.165112452002195e-05, + "loss": 0.3183, + "step": 674 + }, + { + "epoch": 0.5543009648942723, + "grad_norm": 0.08308850973844528, + "learning_rate": 8.16236972024136e-05, + "loss": 0.3268, + "step": 675 + }, + { + "epoch": 0.5551221515089304, + "grad_norm": 0.07913679629564285, + "learning_rate": 8.159626988480527e-05, + "loss": 0.3358, + "step": 676 + }, + { + "epoch": 0.5559433381235885, + "grad_norm": 0.08472824096679688, + "learning_rate": 8.156884256719694e-05, + "loss": 0.3619, + "step": 677 + }, + { + "epoch": 0.5567645247382468, + "grad_norm": 0.07870602607727051, + "learning_rate": 8.154141524958859e-05, + "loss": 0.3372, + "step": 678 + }, + { + "epoch": 0.5575857113529049, + "grad_norm": 0.07889708876609802, + "learning_rate": 8.151398793198025e-05, + "loss": 0.34, + "step": 679 + }, + { + "epoch": 0.5584068979675632, + "grad_norm": 0.0856415405869484, + "learning_rate": 8.148656061437193e-05, + "loss": 0.3132, + "step": 680 + }, + { + "epoch": 0.5592280845822213, + "grad_norm": 0.0840810164809227, + "learning_rate": 8.145913329676358e-05, + "loss": 0.3451, + "step": 681 + }, + { + "epoch": 0.5600492711968795, + "grad_norm": 0.07759146392345428, + "learning_rate": 8.143170597915524e-05, + "loss": 0.3452, + "step": 682 + }, + { + "epoch": 0.5608704578115377, + "grad_norm": 0.08038268238306046, + "learning_rate": 8.14042786615469e-05, + "loss": 0.3103, + "step": 683 + }, + { + "epoch": 0.5616916444261959, + "grad_norm": 0.07695576548576355, + "learning_rate": 8.137685134393857e-05, + "loss": 0.342, + "step": 684 + }, + { + "epoch": 0.562512831040854, + "grad_norm": 0.08458397537469864, + "learning_rate": 8.134942402633024e-05, + "loss": 0.344, + "step": 685 + }, + { + "epoch": 0.5633340176555122, + "grad_norm": 0.08124341070652008, + "learning_rate": 8.13219967087219e-05, + "loss": 0.3235, + "step": 686 + }, + { + "epoch": 0.5641552042701704, + "grad_norm": 0.09097409248352051, + "learning_rate": 8.129456939111355e-05, + "loss": 0.3142, + "step": 687 + }, + { + "epoch": 0.5649763908848285, + "grad_norm": 0.07753387093544006, + "learning_rate": 8.126714207350522e-05, + "loss": 0.3295, + "step": 688 + }, + { + "epoch": 0.5657975774994868, + "grad_norm": 0.07960804551839828, + "learning_rate": 8.123971475589688e-05, + "loss": 0.3156, + "step": 689 + }, + { + "epoch": 0.5666187641141449, + "grad_norm": 0.0779690146446228, + "learning_rate": 8.121228743828854e-05, + "loss": 0.3224, + "step": 690 + }, + { + "epoch": 0.5674399507288032, + "grad_norm": 0.07964863628149033, + "learning_rate": 8.11848601206802e-05, + "loss": 0.3846, + "step": 691 + }, + { + "epoch": 0.5682611373434613, + "grad_norm": 0.08143079280853271, + "learning_rate": 8.115743280307186e-05, + "loss": 0.3268, + "step": 692 + }, + { + "epoch": 0.5690823239581195, + "grad_norm": 0.08909327536821365, + "learning_rate": 8.113000548546353e-05, + "loss": 0.3354, + "step": 693 + }, + { + "epoch": 0.5699035105727777, + "grad_norm": 0.07217192649841309, + "learning_rate": 8.11025781678552e-05, + "loss": 0.3077, + "step": 694 + }, + { + "epoch": 0.5707246971874358, + "grad_norm": 0.07756021618843079, + "learning_rate": 8.107515085024685e-05, + "loss": 0.3244, + "step": 695 + }, + { + "epoch": 0.571545883802094, + "grad_norm": 0.08227197080850601, + "learning_rate": 8.10477235326385e-05, + "loss": 0.34, + "step": 696 + }, + { + "epoch": 0.5723670704167522, + "grad_norm": 0.07765581458806992, + "learning_rate": 8.102029621503017e-05, + "loss": 0.2898, + "step": 697 + }, + { + "epoch": 0.5731882570314104, + "grad_norm": 0.07710105925798416, + "learning_rate": 8.099286889742184e-05, + "loss": 0.3056, + "step": 698 + }, + { + "epoch": 0.5740094436460685, + "grad_norm": 0.07832709699869156, + "learning_rate": 8.09654415798135e-05, + "loss": 0.3338, + "step": 699 + }, + { + "epoch": 0.5748306302607268, + "grad_norm": 0.08406300097703934, + "learning_rate": 8.093801426220516e-05, + "loss": 0.3483, + "step": 700 + }, + { + "epoch": 0.5756518168753849, + "grad_norm": 0.07969654351472855, + "learning_rate": 8.091058694459682e-05, + "loss": 0.3645, + "step": 701 + }, + { + "epoch": 0.5764730034900432, + "grad_norm": 0.07744769752025604, + "learning_rate": 8.088315962698848e-05, + "loss": 0.3324, + "step": 702 + }, + { + "epoch": 0.5772941901047013, + "grad_norm": 0.07306936383247375, + "learning_rate": 8.085573230938015e-05, + "loss": 0.3011, + "step": 703 + }, + { + "epoch": 0.5781153767193594, + "grad_norm": 0.09406828135251999, + "learning_rate": 8.082830499177181e-05, + "loss": 0.3518, + "step": 704 + }, + { + "epoch": 0.5781153767193594, + "eval_runtime": 494.5472, + "eval_samples_per_second": 0.398, + "eval_steps_per_second": 0.2, + "step": 704 + } + ], + "logging_steps": 1, + "max_steps": 3651, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 88, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.1068114755114435e+19, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}