{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.14452884417983985, "eval_steps": 88, "global_step": 176, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.000821186614658181, "grad_norm": 0.480191707611084, "learning_rate": 2e-05, "loss": 1.332, "step": 1 }, { "epoch": 0.001642373229316362, "grad_norm": 0.43048733472824097, "learning_rate": 4e-05, "loss": 1.2784, "step": 2 }, { "epoch": 0.002463559843974543, "grad_norm": 0.33739173412323, "learning_rate": 6e-05, "loss": 1.3286, "step": 3 }, { "epoch": 0.003284746458632724, "grad_norm": 0.432579904794693, "learning_rate": 8e-05, "loss": 1.3079, "step": 4 }, { "epoch": 0.0041059330732909054, "grad_norm": 0.3490436375141144, "learning_rate": 0.0001, "loss": 1.2182, "step": 5 }, { "epoch": 0.004927119687949086, "grad_norm": 0.20206260681152344, "learning_rate": 9.997257268239166e-05, "loss": 1.2828, "step": 6 }, { "epoch": 0.005748306302607267, "grad_norm": 0.15355628728866577, "learning_rate": 9.994514536478333e-05, "loss": 1.1616, "step": 7 }, { "epoch": 0.006569492917265448, "grad_norm": 0.16756780445575714, "learning_rate": 9.9917718047175e-05, "loss": 1.0935, "step": 8 }, { "epoch": 0.00739067953192363, "grad_norm": 0.17429664731025696, "learning_rate": 9.989029072956665e-05, "loss": 0.9694, "step": 9 }, { "epoch": 0.008211866146581811, "grad_norm": 0.22355175018310547, "learning_rate": 9.986286341195832e-05, "loss": 0.9407, "step": 10 }, { "epoch": 0.009033052761239991, "grad_norm": 0.33407703042030334, "learning_rate": 9.983543609434997e-05, "loss": 0.7717, "step": 11 }, { "epoch": 0.009854239375898173, "grad_norm": 0.47473278641700745, "learning_rate": 9.980800877674164e-05, "loss": 0.7763, "step": 12 }, { "epoch": 0.010675425990556354, "grad_norm": 0.2812059819698334, "learning_rate": 9.978058145913331e-05, "loss": 0.6258, "step": 13 }, { "epoch": 0.011496612605214535, "grad_norm": 0.23547925055027008, "learning_rate": 9.975315414152496e-05, "loss": 0.5968, "step": 14 }, { "epoch": 0.012317799219872716, "grad_norm": 0.18453630805015564, "learning_rate": 9.972572682391662e-05, "loss": 0.5368, "step": 15 }, { "epoch": 0.013138985834530896, "grad_norm": 0.16103577613830566, "learning_rate": 9.969829950630828e-05, "loss": 0.4909, "step": 16 }, { "epoch": 0.013960172449189078, "grad_norm": 0.18651455640792847, "learning_rate": 9.967087218869995e-05, "loss": 0.5135, "step": 17 }, { "epoch": 0.01478135906384726, "grad_norm": 0.11300642043352127, "learning_rate": 9.96434448710916e-05, "loss": 0.5069, "step": 18 }, { "epoch": 0.01560254567850544, "grad_norm": 0.10415703803300858, "learning_rate": 9.961601755348327e-05, "loss": 0.4851, "step": 19 }, { "epoch": 0.016423732293163622, "grad_norm": 0.11693017929792404, "learning_rate": 9.958859023587493e-05, "loss": 0.4625, "step": 20 }, { "epoch": 0.017244918907821802, "grad_norm": 0.10035043954849243, "learning_rate": 9.95611629182666e-05, "loss": 0.4822, "step": 21 }, { "epoch": 0.018066105522479982, "grad_norm": 0.10483390837907791, "learning_rate": 9.953373560065826e-05, "loss": 0.4342, "step": 22 }, { "epoch": 0.018887292137138165, "grad_norm": 2.8405802249908447, "learning_rate": 9.950630828304992e-05, "loss": 0.4664, "step": 23 }, { "epoch": 0.019708478751796345, "grad_norm": 0.13821998238563538, "learning_rate": 9.947888096544159e-05, "loss": 0.4468, "step": 24 }, { "epoch": 0.020529665366454525, "grad_norm": 0.1991378366947174, "learning_rate": 9.945145364783325e-05, "loss": 0.4605, "step": 25 }, { "epoch": 0.02135085198111271, "grad_norm": 0.07619134336709976, "learning_rate": 9.942402633022491e-05, "loss": 0.4597, "step": 26 }, { "epoch": 0.02217203859577089, "grad_norm": 0.13373583555221558, "learning_rate": 9.939659901261658e-05, "loss": 0.4626, "step": 27 }, { "epoch": 0.02299322521042907, "grad_norm": 0.09962721168994904, "learning_rate": 9.936917169500823e-05, "loss": 0.4638, "step": 28 }, { "epoch": 0.023814411825087253, "grad_norm": 0.09395964443683624, "learning_rate": 9.93417443773999e-05, "loss": 0.4569, "step": 29 }, { "epoch": 0.024635598439745433, "grad_norm": 0.09109952300786972, "learning_rate": 9.931431705979157e-05, "loss": 0.4439, "step": 30 }, { "epoch": 0.025456785054403613, "grad_norm": 0.10370515286922455, "learning_rate": 9.928688974218322e-05, "loss": 0.4425, "step": 31 }, { "epoch": 0.026277971669061793, "grad_norm": 0.2153477966785431, "learning_rate": 9.925946242457488e-05, "loss": 0.4503, "step": 32 }, { "epoch": 0.027099158283719976, "grad_norm": 0.08772841095924377, "learning_rate": 9.923203510696654e-05, "loss": 0.419, "step": 33 }, { "epoch": 0.027920344898378156, "grad_norm": 0.10951374471187592, "learning_rate": 9.920460778935821e-05, "loss": 0.4353, "step": 34 }, { "epoch": 0.028741531513036336, "grad_norm": 0.09190870076417923, "learning_rate": 9.917718047174987e-05, "loss": 0.5196, "step": 35 }, { "epoch": 0.02956271812769452, "grad_norm": 0.07667124271392822, "learning_rate": 9.914975315414153e-05, "loss": 0.4358, "step": 36 }, { "epoch": 0.0303839047423527, "grad_norm": 0.1514267474412918, "learning_rate": 9.912232583653319e-05, "loss": 0.411, "step": 37 }, { "epoch": 0.03120509135701088, "grad_norm": 0.09086549282073975, "learning_rate": 9.909489851892486e-05, "loss": 0.4003, "step": 38 }, { "epoch": 0.032026277971669063, "grad_norm": 0.2616782486438751, "learning_rate": 9.906747120131652e-05, "loss": 0.4842, "step": 39 }, { "epoch": 0.032847464586327244, "grad_norm": 0.11908283084630966, "learning_rate": 9.904004388370818e-05, "loss": 0.4143, "step": 40 }, { "epoch": 0.033668651200985424, "grad_norm": 0.07770542800426483, "learning_rate": 9.901261656609983e-05, "loss": 0.3873, "step": 41 }, { "epoch": 0.034489837815643604, "grad_norm": 0.08934606611728668, "learning_rate": 9.89851892484915e-05, "loss": 0.4235, "step": 42 }, { "epoch": 0.035311024430301784, "grad_norm": 0.09303563088178635, "learning_rate": 9.895776193088317e-05, "loss": 0.4103, "step": 43 }, { "epoch": 0.036132211044959964, "grad_norm": 0.08622181415557861, "learning_rate": 9.893033461327482e-05, "loss": 0.448, "step": 44 }, { "epoch": 0.03695339765961815, "grad_norm": 0.08822862058877945, "learning_rate": 9.890290729566649e-05, "loss": 0.3855, "step": 45 }, { "epoch": 0.03777458427427633, "grad_norm": 0.08557698875665665, "learning_rate": 9.887547997805814e-05, "loss": 0.3945, "step": 46 }, { "epoch": 0.03859577088893451, "grad_norm": 0.07540106773376465, "learning_rate": 9.884805266044981e-05, "loss": 0.4192, "step": 47 }, { "epoch": 0.03941695750359269, "grad_norm": 0.1023702397942543, "learning_rate": 9.882062534284148e-05, "loss": 0.4126, "step": 48 }, { "epoch": 0.04023814411825087, "grad_norm": 0.07779772579669952, "learning_rate": 9.879319802523313e-05, "loss": 0.4244, "step": 49 }, { "epoch": 0.04105933073290905, "grad_norm": 0.08826564252376556, "learning_rate": 9.876577070762479e-05, "loss": 0.415, "step": 50 }, { "epoch": 0.04188051734756724, "grad_norm": 0.08254576474428177, "learning_rate": 9.873834339001646e-05, "loss": 0.4346, "step": 51 }, { "epoch": 0.04270170396222542, "grad_norm": 0.08287151902914047, "learning_rate": 9.871091607240812e-05, "loss": 0.4142, "step": 52 }, { "epoch": 0.0435228905768836, "grad_norm": 0.08196476101875305, "learning_rate": 9.868348875479978e-05, "loss": 0.3822, "step": 53 }, { "epoch": 0.04434407719154178, "grad_norm": 0.08654092252254486, "learning_rate": 9.865606143719145e-05, "loss": 0.3937, "step": 54 }, { "epoch": 0.04516526380619996, "grad_norm": 0.1102684736251831, "learning_rate": 9.86286341195831e-05, "loss": 0.4508, "step": 55 }, { "epoch": 0.04598645042085814, "grad_norm": 0.08240954577922821, "learning_rate": 9.860120680197478e-05, "loss": 0.4069, "step": 56 }, { "epoch": 0.04680763703551632, "grad_norm": 0.08542217314243317, "learning_rate": 9.857377948436644e-05, "loss": 0.5002, "step": 57 }, { "epoch": 0.047628823650174505, "grad_norm": 0.08390172570943832, "learning_rate": 9.854635216675809e-05, "loss": 0.3851, "step": 58 }, { "epoch": 0.048450010264832685, "grad_norm": 0.10694168508052826, "learning_rate": 9.851892484914976e-05, "loss": 0.4026, "step": 59 }, { "epoch": 0.049271196879490865, "grad_norm": 0.0852806493639946, "learning_rate": 9.849149753154143e-05, "loss": 0.424, "step": 60 }, { "epoch": 0.050092383494149045, "grad_norm": 0.15425831079483032, "learning_rate": 9.846407021393308e-05, "loss": 0.4307, "step": 61 }, { "epoch": 0.050913570108807225, "grad_norm": 0.08546218276023865, "learning_rate": 9.843664289632475e-05, "loss": 0.4386, "step": 62 }, { "epoch": 0.051734756723465405, "grad_norm": 0.08588795363903046, "learning_rate": 9.84092155787164e-05, "loss": 0.4433, "step": 63 }, { "epoch": 0.052555943338123585, "grad_norm": 0.08901514858007431, "learning_rate": 9.838178826110807e-05, "loss": 0.4263, "step": 64 }, { "epoch": 0.05337712995278177, "grad_norm": 0.17668095231056213, "learning_rate": 9.835436094349974e-05, "loss": 0.407, "step": 65 }, { "epoch": 0.05419831656743995, "grad_norm": 0.081763856112957, "learning_rate": 9.83269336258914e-05, "loss": 0.374, "step": 66 }, { "epoch": 0.05501950318209813, "grad_norm": 0.09063572436571121, "learning_rate": 9.829950630828305e-05, "loss": 0.3839, "step": 67 }, { "epoch": 0.05584068979675631, "grad_norm": 0.08264392614364624, "learning_rate": 9.827207899067472e-05, "loss": 0.4676, "step": 68 }, { "epoch": 0.05666187641141449, "grad_norm": 0.0827123150229454, "learning_rate": 9.824465167306638e-05, "loss": 0.3801, "step": 69 }, { "epoch": 0.05748306302607267, "grad_norm": 0.07972189038991928, "learning_rate": 9.821722435545804e-05, "loss": 0.3758, "step": 70 }, { "epoch": 0.05830424964073085, "grad_norm": 0.08738942444324493, "learning_rate": 9.81897970378497e-05, "loss": 0.3634, "step": 71 }, { "epoch": 0.05912543625538904, "grad_norm": 0.08442792296409607, "learning_rate": 9.816236972024136e-05, "loss": 0.3803, "step": 72 }, { "epoch": 0.05994662287004722, "grad_norm": 0.2751137316226959, "learning_rate": 9.813494240263303e-05, "loss": 0.3869, "step": 73 }, { "epoch": 0.0607678094847054, "grad_norm": 0.19429296255111694, "learning_rate": 9.81075150850247e-05, "loss": 0.3992, "step": 74 }, { "epoch": 0.06158899609936358, "grad_norm": 0.08392605930566788, "learning_rate": 9.808008776741635e-05, "loss": 0.361, "step": 75 }, { "epoch": 0.06241018271402176, "grad_norm": 2.01043963432312, "learning_rate": 9.8052660449808e-05, "loss": 0.3787, "step": 76 }, { "epoch": 0.06323136932867994, "grad_norm": 0.0808538943529129, "learning_rate": 9.802523313219967e-05, "loss": 0.4013, "step": 77 }, { "epoch": 0.06405255594333813, "grad_norm": 0.07846518605947495, "learning_rate": 9.799780581459134e-05, "loss": 0.3808, "step": 78 }, { "epoch": 0.0648737425579963, "grad_norm": 0.08152970671653748, "learning_rate": 9.7970378496983e-05, "loss": 0.4076, "step": 79 }, { "epoch": 0.06569492917265449, "grad_norm": 0.07745791226625443, "learning_rate": 9.794295117937466e-05, "loss": 0.3981, "step": 80 }, { "epoch": 0.06651611578731266, "grad_norm": 0.0842173770070076, "learning_rate": 9.791552386176632e-05, "loss": 0.4469, "step": 81 }, { "epoch": 0.06733730240197085, "grad_norm": 0.07920663058757782, "learning_rate": 9.788809654415799e-05, "loss": 0.3988, "step": 82 }, { "epoch": 0.06815848901662903, "grad_norm": 0.07899456471204758, "learning_rate": 9.786066922654965e-05, "loss": 0.4418, "step": 83 }, { "epoch": 0.06897967563128721, "grad_norm": 0.07229727506637573, "learning_rate": 9.783324190894131e-05, "loss": 0.4016, "step": 84 }, { "epoch": 0.0698008622459454, "grad_norm": 0.08062436431646347, "learning_rate": 9.780581459133296e-05, "loss": 0.3762, "step": 85 }, { "epoch": 0.07062204886060357, "grad_norm": 0.08440960198640823, "learning_rate": 9.777838727372464e-05, "loss": 0.4362, "step": 86 }, { "epoch": 0.07144323547526175, "grad_norm": 0.07544733583927155, "learning_rate": 9.77509599561163e-05, "loss": 0.4135, "step": 87 }, { "epoch": 0.07226442208991993, "grad_norm": 0.1760656237602234, "learning_rate": 9.772353263850797e-05, "loss": 0.3887, "step": 88 }, { "epoch": 0.07226442208991993, "eval_runtime": 494.3052, "eval_samples_per_second": 0.399, "eval_steps_per_second": 0.2, "step": 88 }, { "epoch": 0.07308560870457811, "grad_norm": 0.08049104362726212, "learning_rate": 9.769610532089962e-05, "loss": 0.4272, "step": 89 }, { "epoch": 0.0739067953192363, "grad_norm": 0.07617965340614319, "learning_rate": 9.766867800329129e-05, "loss": 0.3873, "step": 90 }, { "epoch": 0.07472798193389447, "grad_norm": 0.07975462824106216, "learning_rate": 9.764125068568296e-05, "loss": 0.3903, "step": 91 }, { "epoch": 0.07554916854855266, "grad_norm": 0.08299189805984497, "learning_rate": 9.761382336807461e-05, "loss": 0.404, "step": 92 }, { "epoch": 0.07637035516321083, "grad_norm": 0.08578819036483765, "learning_rate": 9.758639605046626e-05, "loss": 0.3983, "step": 93 }, { "epoch": 0.07719154177786902, "grad_norm": 0.10872071981430054, "learning_rate": 9.755896873285793e-05, "loss": 0.3553, "step": 94 }, { "epoch": 0.07801272839252721, "grad_norm": 0.08121436834335327, "learning_rate": 9.75315414152496e-05, "loss": 0.408, "step": 95 }, { "epoch": 0.07883391500718538, "grad_norm": 0.09175996482372284, "learning_rate": 9.750411409764125e-05, "loss": 0.4008, "step": 96 }, { "epoch": 0.07965510162184357, "grad_norm": 0.08122789114713669, "learning_rate": 9.747668678003292e-05, "loss": 0.4045, "step": 97 }, { "epoch": 0.08047628823650174, "grad_norm": 0.08402436226606369, "learning_rate": 9.744925946242458e-05, "loss": 0.3814, "step": 98 }, { "epoch": 0.08129747485115993, "grad_norm": 0.08454861491918564, "learning_rate": 9.742183214481624e-05, "loss": 0.3904, "step": 99 }, { "epoch": 0.0821186614658181, "grad_norm": 0.08125888556241989, "learning_rate": 9.739440482720791e-05, "loss": 0.3681, "step": 100 }, { "epoch": 0.08293984808047629, "grad_norm": 0.08544078469276428, "learning_rate": 9.736697750959957e-05, "loss": 0.367, "step": 101 }, { "epoch": 0.08376103469513448, "grad_norm": 0.07539089769124985, "learning_rate": 9.733955019199122e-05, "loss": 0.3821, "step": 102 }, { "epoch": 0.08458222130979265, "grad_norm": 0.07529085874557495, "learning_rate": 9.731212287438289e-05, "loss": 0.4169, "step": 103 }, { "epoch": 0.08540340792445084, "grad_norm": 0.07588034868240356, "learning_rate": 9.728469555677456e-05, "loss": 0.4211, "step": 104 }, { "epoch": 0.08622459453910901, "grad_norm": 0.08019097149372101, "learning_rate": 9.725726823916621e-05, "loss": 0.4033, "step": 105 }, { "epoch": 0.0870457811537672, "grad_norm": 0.07878712564706802, "learning_rate": 9.722984092155788e-05, "loss": 0.3842, "step": 106 }, { "epoch": 0.08786696776842537, "grad_norm": 0.08083963394165039, "learning_rate": 9.720241360394953e-05, "loss": 0.3558, "step": 107 }, { "epoch": 0.08868815438308356, "grad_norm": 0.08340411633253098, "learning_rate": 9.71749862863412e-05, "loss": 0.4037, "step": 108 }, { "epoch": 0.08950934099774174, "grad_norm": 0.09317754209041595, "learning_rate": 9.714755896873287e-05, "loss": 0.3874, "step": 109 }, { "epoch": 0.09033052761239992, "grad_norm": 0.07143125683069229, "learning_rate": 9.712013165112452e-05, "loss": 0.3441, "step": 110 }, { "epoch": 0.0911517142270581, "grad_norm": 0.0909111350774765, "learning_rate": 9.709270433351618e-05, "loss": 0.3552, "step": 111 }, { "epoch": 0.09197290084171628, "grad_norm": 0.08041603863239288, "learning_rate": 9.706527701590785e-05, "loss": 0.394, "step": 112 }, { "epoch": 0.09279408745637446, "grad_norm": 0.08147992193698883, "learning_rate": 9.703784969829951e-05, "loss": 0.3916, "step": 113 }, { "epoch": 0.09361527407103264, "grad_norm": 0.08365318924188614, "learning_rate": 9.701042238069117e-05, "loss": 0.4075, "step": 114 }, { "epoch": 0.09443646068569082, "grad_norm": 0.083246149122715, "learning_rate": 9.698299506308284e-05, "loss": 0.3566, "step": 115 }, { "epoch": 0.09525764730034901, "grad_norm": 0.0942649245262146, "learning_rate": 9.695556774547449e-05, "loss": 0.3516, "step": 116 }, { "epoch": 0.09607883391500718, "grad_norm": 0.08110091835260391, "learning_rate": 9.692814042786616e-05, "loss": 0.3902, "step": 117 }, { "epoch": 0.09690002052966537, "grad_norm": 0.43124014139175415, "learning_rate": 9.690071311025783e-05, "loss": 0.3956, "step": 118 }, { "epoch": 0.09772120714432354, "grad_norm": 0.08204706013202667, "learning_rate": 9.687328579264948e-05, "loss": 0.3865, "step": 119 }, { "epoch": 0.09854239375898173, "grad_norm": 0.08262762427330017, "learning_rate": 9.684585847504113e-05, "loss": 0.3777, "step": 120 }, { "epoch": 0.0993635803736399, "grad_norm": 0.08233962953090668, "learning_rate": 9.681843115743282e-05, "loss": 0.3976, "step": 121 }, { "epoch": 0.10018476698829809, "grad_norm": 0.7428120970726013, "learning_rate": 9.679100383982447e-05, "loss": 0.377, "step": 122 }, { "epoch": 0.10100595360295628, "grad_norm": 0.07909400761127472, "learning_rate": 9.676357652221614e-05, "loss": 0.3654, "step": 123 }, { "epoch": 0.10182714021761445, "grad_norm": 0.08714035898447037, "learning_rate": 9.673614920460779e-05, "loss": 0.4181, "step": 124 }, { "epoch": 0.10264832683227264, "grad_norm": 0.08017311245203018, "learning_rate": 9.670872188699946e-05, "loss": 0.4052, "step": 125 }, { "epoch": 0.10346951344693081, "grad_norm": 0.13821078836917877, "learning_rate": 9.668129456939113e-05, "loss": 0.3733, "step": 126 }, { "epoch": 0.104290700061589, "grad_norm": 0.1609969437122345, "learning_rate": 9.665386725178278e-05, "loss": 0.3686, "step": 127 }, { "epoch": 0.10511188667624717, "grad_norm": 0.1280309557914734, "learning_rate": 9.662643993417444e-05, "loss": 0.4036, "step": 128 }, { "epoch": 0.10593307329090536, "grad_norm": 0.08747898787260056, "learning_rate": 9.65990126165661e-05, "loss": 0.3587, "step": 129 }, { "epoch": 0.10675425990556354, "grad_norm": 0.1252209097146988, "learning_rate": 9.657158529895777e-05, "loss": 0.3626, "step": 130 }, { "epoch": 0.10757544652022172, "grad_norm": 0.09374388307332993, "learning_rate": 9.654415798134943e-05, "loss": 0.3824, "step": 131 }, { "epoch": 0.1083966331348799, "grad_norm": 0.08730709552764893, "learning_rate": 9.65167306637411e-05, "loss": 0.3827, "step": 132 }, { "epoch": 0.10921781974953808, "grad_norm": 0.07719024270772934, "learning_rate": 9.648930334613275e-05, "loss": 0.3762, "step": 133 }, { "epoch": 0.11003900636419627, "grad_norm": 0.08843278139829636, "learning_rate": 9.646187602852442e-05, "loss": 0.3843, "step": 134 }, { "epoch": 0.11086019297885444, "grad_norm": 0.08568207919597626, "learning_rate": 9.643444871091608e-05, "loss": 0.3594, "step": 135 }, { "epoch": 0.11168137959351263, "grad_norm": 0.08556952327489853, "learning_rate": 9.640702139330774e-05, "loss": 0.3886, "step": 136 }, { "epoch": 0.11250256620817081, "grad_norm": 0.09016801416873932, "learning_rate": 9.63795940756994e-05, "loss": 0.4023, "step": 137 }, { "epoch": 0.11332375282282899, "grad_norm": 0.08133590966463089, "learning_rate": 9.635216675809106e-05, "loss": 0.3634, "step": 138 }, { "epoch": 0.11414493943748717, "grad_norm": 0.2211730182170868, "learning_rate": 9.632473944048273e-05, "loss": 0.3518, "step": 139 }, { "epoch": 0.11496612605214535, "grad_norm": 0.08816584199666977, "learning_rate": 9.629731212287438e-05, "loss": 0.3727, "step": 140 }, { "epoch": 0.11578731266680353, "grad_norm": 0.11618969589471817, "learning_rate": 9.626988480526605e-05, "loss": 0.4247, "step": 141 }, { "epoch": 0.1166084992814617, "grad_norm": 0.09243030101060867, "learning_rate": 9.62424574876577e-05, "loss": 0.396, "step": 142 }, { "epoch": 0.11742968589611989, "grad_norm": 0.08566376566886902, "learning_rate": 9.621503017004937e-05, "loss": 0.3852, "step": 143 }, { "epoch": 0.11825087251077808, "grad_norm": 0.08220973610877991, "learning_rate": 9.618760285244104e-05, "loss": 0.3961, "step": 144 }, { "epoch": 0.11907205912543625, "grad_norm": 0.08240345865488052, "learning_rate": 9.61601755348327e-05, "loss": 0.3518, "step": 145 }, { "epoch": 0.11989324574009444, "grad_norm": 0.08472532778978348, "learning_rate": 9.613274821722435e-05, "loss": 0.3586, "step": 146 }, { "epoch": 0.12071443235475261, "grad_norm": 0.08407485485076904, "learning_rate": 9.610532089961602e-05, "loss": 0.3797, "step": 147 }, { "epoch": 0.1215356189694108, "grad_norm": 0.09284385293722153, "learning_rate": 9.607789358200769e-05, "loss": 0.3499, "step": 148 }, { "epoch": 0.12235680558406897, "grad_norm": 0.08499818295240402, "learning_rate": 9.605046626439934e-05, "loss": 0.3722, "step": 149 }, { "epoch": 0.12317799219872716, "grad_norm": 0.080271415412426, "learning_rate": 9.602303894679101e-05, "loss": 0.3871, "step": 150 }, { "epoch": 0.12399917881338535, "grad_norm": 0.07850060611963272, "learning_rate": 9.599561162918266e-05, "loss": 0.3679, "step": 151 }, { "epoch": 0.12482036542804352, "grad_norm": 0.07685016840696335, "learning_rate": 9.596818431157433e-05, "loss": 0.4078, "step": 152 }, { "epoch": 0.1256415520427017, "grad_norm": 0.09402357786893845, "learning_rate": 9.5940756993966e-05, "loss": 0.3996, "step": 153 }, { "epoch": 0.12646273865735988, "grad_norm": 0.08445476740598679, "learning_rate": 9.591332967635765e-05, "loss": 0.3841, "step": 154 }, { "epoch": 0.12728392527201807, "grad_norm": 0.08233911544084549, "learning_rate": 9.588590235874932e-05, "loss": 0.3529, "step": 155 }, { "epoch": 0.12810511188667625, "grad_norm": 0.07896068692207336, "learning_rate": 9.585847504114099e-05, "loss": 0.3585, "step": 156 }, { "epoch": 0.12892629850133444, "grad_norm": 0.0822276696562767, "learning_rate": 9.583104772353264e-05, "loss": 0.3962, "step": 157 }, { "epoch": 0.1297474851159926, "grad_norm": 0.07977598905563354, "learning_rate": 9.580362040592431e-05, "loss": 0.454, "step": 158 }, { "epoch": 0.1305686717306508, "grad_norm": 0.0857616737484932, "learning_rate": 9.577619308831597e-05, "loss": 0.3954, "step": 159 }, { "epoch": 0.13138985834530897, "grad_norm": 0.0874355211853981, "learning_rate": 9.574876577070763e-05, "loss": 0.3591, "step": 160 }, { "epoch": 0.13221104495996716, "grad_norm": 0.07877468317747116, "learning_rate": 9.57213384530993e-05, "loss": 0.348, "step": 161 }, { "epoch": 0.13303223157462532, "grad_norm": 0.08618593961000443, "learning_rate": 9.569391113549096e-05, "loss": 0.3723, "step": 162 }, { "epoch": 0.1338534181892835, "grad_norm": 0.08144336938858032, "learning_rate": 9.566648381788261e-05, "loss": 0.4146, "step": 163 }, { "epoch": 0.1346746048039417, "grad_norm": 0.07322760671377182, "learning_rate": 9.563905650027428e-05, "loss": 0.3107, "step": 164 }, { "epoch": 0.13549579141859988, "grad_norm": 0.08007095754146576, "learning_rate": 9.561162918266595e-05, "loss": 0.4128, "step": 165 }, { "epoch": 0.13631697803325807, "grad_norm": 0.09636646509170532, "learning_rate": 9.55842018650576e-05, "loss": 0.4089, "step": 166 }, { "epoch": 0.13713816464791623, "grad_norm": 0.08381053060293198, "learning_rate": 9.555677454744927e-05, "loss": 0.3624, "step": 167 }, { "epoch": 0.13795935126257441, "grad_norm": 0.07476504147052765, "learning_rate": 9.552934722984092e-05, "loss": 0.3906, "step": 168 }, { "epoch": 0.1387805378772326, "grad_norm": 0.0901239663362503, "learning_rate": 9.550191991223259e-05, "loss": 0.3378, "step": 169 }, { "epoch": 0.1396017244918908, "grad_norm": 0.0813356265425682, "learning_rate": 9.547449259462426e-05, "loss": 0.3627, "step": 170 }, { "epoch": 0.14042291110654898, "grad_norm": 0.14319093525409698, "learning_rate": 9.544706527701591e-05, "loss": 0.3512, "step": 171 }, { "epoch": 0.14124409772120713, "grad_norm": 0.13329866528511047, "learning_rate": 9.541963795940757e-05, "loss": 0.3809, "step": 172 }, { "epoch": 0.14206528433586532, "grad_norm": 0.0815596953034401, "learning_rate": 9.539221064179923e-05, "loss": 0.3444, "step": 173 }, { "epoch": 0.1428864709505235, "grad_norm": 0.08646956831216812, "learning_rate": 9.53647833241909e-05, "loss": 0.3699, "step": 174 }, { "epoch": 0.1437076575651817, "grad_norm": 0.09374339133501053, "learning_rate": 9.533735600658256e-05, "loss": 0.3569, "step": 175 }, { "epoch": 0.14452884417983985, "grad_norm": 0.0834718644618988, "learning_rate": 9.530992868897422e-05, "loss": 0.3708, "step": 176 }, { "epoch": 0.14452884417983985, "eval_runtime": 493.9539, "eval_samples_per_second": 0.399, "eval_steps_per_second": 0.2, "step": 176 } ], "logging_steps": 1, "max_steps": 3651, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 88, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 7.760987080135606e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }