{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9985272459499264, "eval_steps": 500, "global_step": 339, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0029455081001472753, "grad_norm": 148176371712.0, "learning_rate": 1.0000000000000002e-06, "loss": 3.5829, "step": 1 }, { "epoch": 0.005891016200294551, "grad_norm": 521838526464.0, "learning_rate": 2.0000000000000003e-06, "loss": 3.7866, "step": 2 }, { "epoch": 0.008836524300441826, "grad_norm": 29312.015625, "learning_rate": 3e-06, "loss": 5.4788, "step": 3 }, { "epoch": 0.011782032400589101, "grad_norm": 4224.009765625, "learning_rate": 4.000000000000001e-06, "loss": 4.9963, "step": 4 }, { "epoch": 0.014727540500736377, "grad_norm": 6816.009765625, "learning_rate": 5e-06, "loss": 6.1696, "step": 5 }, { "epoch": 0.017673048600883652, "grad_norm": 37120.00390625, "learning_rate": 6e-06, "loss": 6.0103, "step": 6 }, { "epoch": 0.020618556701030927, "grad_norm": 16512.0078125, "learning_rate": 7.000000000000001e-06, "loss": 5.2747, "step": 7 }, { "epoch": 0.023564064801178203, "grad_norm": 4544.0126953125, "learning_rate": 8.000000000000001e-06, "loss": 5.0828, "step": 8 }, { "epoch": 0.026509572901325478, "grad_norm": 1360.0126953125, "learning_rate": 9e-06, "loss": 4.4745, "step": 9 }, { "epoch": 0.029455081001472753, "grad_norm": 2528.019287109375, "learning_rate": 1e-05, "loss": 6.0115, "step": 10 }, { "epoch": 0.03240058910162003, "grad_norm": 608.0706787109375, "learning_rate": 1.1000000000000001e-05, "loss": 5.1475, "step": 11 }, { "epoch": 0.035346097201767304, "grad_norm": 1560.0152587890625, "learning_rate": 1.2e-05, "loss": 5.2077, "step": 12 }, { "epoch": 0.03829160530191458, "grad_norm": 1152.024169921875, "learning_rate": 1.3000000000000001e-05, "loss": 4.9248, "step": 13 }, { "epoch": 0.041237113402061855, "grad_norm": 1096.0169677734375, "learning_rate": 1.4000000000000001e-05, "loss": 4.3356, "step": 14 }, { "epoch": 0.044182621502209134, "grad_norm": 2560.037841796875, "learning_rate": 1.5e-05, "loss": 5.9271, "step": 15 }, { "epoch": 0.047128129602356406, "grad_norm": 1312.0172119140625, "learning_rate": 1.6000000000000003e-05, "loss": 4.9467, "step": 16 }, { "epoch": 0.050073637702503684, "grad_norm": 748.0582275390625, "learning_rate": 1.7000000000000003e-05, "loss": 5.5781, "step": 17 }, { "epoch": 0.053019145802650956, "grad_norm": 608.0347290039062, "learning_rate": 1.8e-05, "loss": 4.9313, "step": 18 }, { "epoch": 0.055964653902798235, "grad_norm": 444.0478820800781, "learning_rate": 1.9e-05, "loss": 4.7802, "step": 19 }, { "epoch": 0.05891016200294551, "grad_norm": 400.05474853515625, "learning_rate": 2e-05, "loss": 4.9916, "step": 20 }, { "epoch": 0.061855670103092786, "grad_norm": 324.0529479980469, "learning_rate": 2.1e-05, "loss": 4.5852, "step": 21 }, { "epoch": 0.06480117820324006, "grad_norm": 552.0383911132812, "learning_rate": 2.2000000000000003e-05, "loss": 4.5658, "step": 22 }, { "epoch": 0.06774668630338733, "grad_norm": 780.0543823242188, "learning_rate": 2.3000000000000003e-05, "loss": 4.9509, "step": 23 }, { "epoch": 0.07069219440353461, "grad_norm": 1352.05712890625, "learning_rate": 2.4e-05, "loss": 4.9526, "step": 24 }, { "epoch": 0.07363770250368189, "grad_norm": 692.063720703125, "learning_rate": 2.5e-05, "loss": 5.1079, "step": 25 }, { "epoch": 0.07658321060382917, "grad_norm": 940.0375366210938, "learning_rate": 2.6000000000000002e-05, "loss": 4.3864, "step": 26 }, { "epoch": 0.07952871870397643, "grad_norm": 406.0652770996094, "learning_rate": 2.7000000000000002e-05, "loss": 4.7486, "step": 27 }, { "epoch": 0.08247422680412371, "grad_norm": 736.0764770507812, "learning_rate": 2.8000000000000003e-05, "loss": 5.4081, "step": 28 }, { "epoch": 0.08541973490427099, "grad_norm": 1240.040283203125, "learning_rate": 2.9e-05, "loss": 4.6019, "step": 29 }, { "epoch": 0.08836524300441827, "grad_norm": 1312.10302734375, "learning_rate": 3e-05, "loss": 5.481, "step": 30 }, { "epoch": 0.09131075110456553, "grad_norm": 466.0743103027344, "learning_rate": 3.1e-05, "loss": 4.4326, "step": 31 }, { "epoch": 0.09425625920471281, "grad_norm": 828.0822143554688, "learning_rate": 3.2000000000000005e-05, "loss": 5.0284, "step": 32 }, { "epoch": 0.09720176730486009, "grad_norm": 764.077880859375, "learning_rate": 3.3e-05, "loss": 4.9104, "step": 33 }, { "epoch": 0.10014727540500737, "grad_norm": 231.1255340576172, "learning_rate": 3.4000000000000007e-05, "loss": 3.9872, "step": 34 }, { "epoch": 0.10309278350515463, "grad_norm": 434.1127624511719, "learning_rate": 3.5e-05, "loss": 5.2373, "step": 35 }, { "epoch": 0.10603829160530191, "grad_norm": 253.16815185546875, "learning_rate": 3.6e-05, "loss": 4.8148, "step": 36 }, { "epoch": 0.10898379970544919, "grad_norm": 290.1834716796875, "learning_rate": 3.7e-05, "loss": 3.6852, "step": 37 }, { "epoch": 0.11192930780559647, "grad_norm": 584.0982666015625, "learning_rate": 3.8e-05, "loss": 3.8818, "step": 38 }, { "epoch": 0.11487481590574374, "grad_norm": 904.0762329101562, "learning_rate": 3.9000000000000006e-05, "loss": 4.2878, "step": 39 }, { "epoch": 0.11782032400589101, "grad_norm": 336.28033447265625, "learning_rate": 4e-05, "loss": 4.2046, "step": 40 }, { "epoch": 0.12076583210603829, "grad_norm": 704.1439208984375, "learning_rate": 4.1e-05, "loss": 4.4065, "step": 41 }, { "epoch": 0.12371134020618557, "grad_norm": 804.0842895507812, "learning_rate": 4.2e-05, "loss": 3.9044, "step": 42 }, { "epoch": 0.12665684830633284, "grad_norm": 752.0977172851562, "learning_rate": 4.3e-05, "loss": 3.5964, "step": 43 }, { "epoch": 0.12960235640648013, "grad_norm": 752.181884765625, "learning_rate": 4.4000000000000006e-05, "loss": 4.3488, "step": 44 }, { "epoch": 0.1325478645066274, "grad_norm": 664.2083129882812, "learning_rate": 4.5e-05, "loss": 4.6978, "step": 45 }, { "epoch": 0.13549337260677466, "grad_norm": 1984.197998046875, "learning_rate": 4.600000000000001e-05, "loss": 3.8173, "step": 46 }, { "epoch": 0.13843888070692195, "grad_norm": 1536.0904541015625, "learning_rate": 4.7e-05, "loss": 3.3992, "step": 47 }, { "epoch": 0.14138438880706922, "grad_norm": 616.1571044921875, "learning_rate": 4.8e-05, "loss": 3.5619, "step": 48 }, { "epoch": 0.14432989690721648, "grad_norm": 5408.072265625, "learning_rate": 4.9e-05, "loss": 4.3322, "step": 49 }, { "epoch": 0.14727540500736377, "grad_norm": 3408.068603515625, "learning_rate": 5e-05, "loss": 4.0344, "step": 50 }, { "epoch": 0.15022091310751104, "grad_norm": 724.0775146484375, "learning_rate": 5.1000000000000006e-05, "loss": 3.8171, "step": 51 }, { "epoch": 0.15316642120765833, "grad_norm": 1448.1661376953125, "learning_rate": 5.2000000000000004e-05, "loss": 2.6826, "step": 52 }, { "epoch": 0.1561119293078056, "grad_norm": 304.0506591796875, "learning_rate": 5.300000000000001e-05, "loss": 2.4422, "step": 53 }, { "epoch": 0.15905743740795286, "grad_norm": 396.0224914550781, "learning_rate": 5.4000000000000005e-05, "loss": 2.4475, "step": 54 }, { "epoch": 0.16200294550810015, "grad_norm": 684.0194702148438, "learning_rate": 5.500000000000001e-05, "loss": 2.2932, "step": 55 }, { "epoch": 0.16494845360824742, "grad_norm": 231.03273010253906, "learning_rate": 5.6000000000000006e-05, "loss": 2.5193, "step": 56 }, { "epoch": 0.16789396170839468, "grad_norm": 164.0357208251953, "learning_rate": 5.6999999999999996e-05, "loss": 1.9575, "step": 57 }, { "epoch": 0.17083946980854198, "grad_norm": 1600.0091552734375, "learning_rate": 5.8e-05, "loss": 2.5806, "step": 58 }, { "epoch": 0.17378497790868924, "grad_norm": 50.36320114135742, "learning_rate": 5.9e-05, "loss": 3.651, "step": 59 }, { "epoch": 0.17673048600883653, "grad_norm": 66.08223724365234, "learning_rate": 6e-05, "loss": 3.5998, "step": 60 }, { "epoch": 0.1796759941089838, "grad_norm": 20.88555335998535, "learning_rate": 6.1e-05, "loss": 3.2898, "step": 61 }, { "epoch": 0.18262150220913106, "grad_norm": 33.95121765136719, "learning_rate": 6.2e-05, "loss": 3.5322, "step": 62 }, { "epoch": 0.18556701030927836, "grad_norm": 18.569725036621094, "learning_rate": 6.3e-05, "loss": 2.7009, "step": 63 }, { "epoch": 0.18851251840942562, "grad_norm": 11.800430297851562, "learning_rate": 6.400000000000001e-05, "loss": 3.0148, "step": 64 }, { "epoch": 0.19145802650957292, "grad_norm": 11.598824501037598, "learning_rate": 6.500000000000001e-05, "loss": 2.5356, "step": 65 }, { "epoch": 0.19440353460972018, "grad_norm": 29.798603057861328, "learning_rate": 6.6e-05, "loss": 3.419, "step": 66 }, { "epoch": 0.19734904270986744, "grad_norm": 23.390228271484375, "learning_rate": 6.7e-05, "loss": 3.0823, "step": 67 }, { "epoch": 0.20029455081001474, "grad_norm": 30.04606056213379, "learning_rate": 6.800000000000001e-05, "loss": 2.6675, "step": 68 }, { "epoch": 0.203240058910162, "grad_norm": 59.424739837646484, "learning_rate": 6.9e-05, "loss": 2.6406, "step": 69 }, { "epoch": 0.20618556701030927, "grad_norm": 18.69099235534668, "learning_rate": 7e-05, "loss": 2.8912, "step": 70 }, { "epoch": 0.20913107511045656, "grad_norm": 11.040902137756348, "learning_rate": 7.1e-05, "loss": 2.2423, "step": 71 }, { "epoch": 0.21207658321060383, "grad_norm": 15.589529991149902, "learning_rate": 7.2e-05, "loss": 2.6387, "step": 72 }, { "epoch": 0.21502209131075112, "grad_norm": 8.346650123596191, "learning_rate": 7.3e-05, "loss": 2.7416, "step": 73 }, { "epoch": 0.21796759941089838, "grad_norm": 16.002195358276367, "learning_rate": 7.4e-05, "loss": 2.5981, "step": 74 }, { "epoch": 0.22091310751104565, "grad_norm": 12.781440734863281, "learning_rate": 7.500000000000001e-05, "loss": 2.8455, "step": 75 }, { "epoch": 0.22385861561119294, "grad_norm": 14.14976978302002, "learning_rate": 7.6e-05, "loss": 2.3376, "step": 76 }, { "epoch": 0.2268041237113402, "grad_norm": 14.174376487731934, "learning_rate": 7.7e-05, "loss": 2.2975, "step": 77 }, { "epoch": 0.22974963181148747, "grad_norm": 12.794411659240723, "learning_rate": 7.800000000000001e-05, "loss": 2.2424, "step": 78 }, { "epoch": 0.23269513991163476, "grad_norm": 8.186522483825684, "learning_rate": 7.900000000000001e-05, "loss": 1.9646, "step": 79 }, { "epoch": 0.23564064801178203, "grad_norm": 7.599482536315918, "learning_rate": 8e-05, "loss": 2.2699, "step": 80 }, { "epoch": 0.23858615611192932, "grad_norm": 13.385209083557129, "learning_rate": 8.1e-05, "loss": 2.9279, "step": 81 }, { "epoch": 0.24153166421207659, "grad_norm": 10.520708084106445, "learning_rate": 8.2e-05, "loss": 2.6627, "step": 82 }, { "epoch": 0.24447717231222385, "grad_norm": 10.881275177001953, "learning_rate": 8.3e-05, "loss": 2.6776, "step": 83 }, { "epoch": 0.24742268041237114, "grad_norm": 11.189691543579102, "learning_rate": 8.4e-05, "loss": 2.5101, "step": 84 }, { "epoch": 0.2503681885125184, "grad_norm": 14.185538291931152, "learning_rate": 8.5e-05, "loss": 2.4517, "step": 85 }, { "epoch": 0.2533136966126657, "grad_norm": 11.319371223449707, "learning_rate": 8.6e-05, "loss": 2.3935, "step": 86 }, { "epoch": 0.25625920471281294, "grad_norm": 11.8043212890625, "learning_rate": 8.7e-05, "loss": 2.3776, "step": 87 }, { "epoch": 0.25920471281296026, "grad_norm": 17.876741409301758, "learning_rate": 8.800000000000001e-05, "loss": 2.1173, "step": 88 }, { "epoch": 0.2621502209131075, "grad_norm": 9.818507194519043, "learning_rate": 8.900000000000001e-05, "loss": 2.4566, "step": 89 }, { "epoch": 0.2650957290132548, "grad_norm": 17.97085189819336, "learning_rate": 9e-05, "loss": 2.4889, "step": 90 }, { "epoch": 0.26804123711340205, "grad_norm": 13.525333404541016, "learning_rate": 9.1e-05, "loss": 2.6413, "step": 91 }, { "epoch": 0.2709867452135493, "grad_norm": 11.17566204071045, "learning_rate": 9.200000000000001e-05, "loss": 2.5934, "step": 92 }, { "epoch": 0.27393225331369664, "grad_norm": 10.861861228942871, "learning_rate": 9.300000000000001e-05, "loss": 2.0431, "step": 93 }, { "epoch": 0.2768777614138439, "grad_norm": 10.782200813293457, "learning_rate": 9.4e-05, "loss": 2.2952, "step": 94 }, { "epoch": 0.27982326951399117, "grad_norm": 10.202653884887695, "learning_rate": 9.5e-05, "loss": 1.9, "step": 95 }, { "epoch": 0.28276877761413843, "grad_norm": 11.118406295776367, "learning_rate": 9.6e-05, "loss": 1.8267, "step": 96 }, { "epoch": 0.2857142857142857, "grad_norm": 11.998210906982422, "learning_rate": 9.7e-05, "loss": 2.7094, "step": 97 }, { "epoch": 0.28865979381443296, "grad_norm": 11.203243255615234, "learning_rate": 9.8e-05, "loss": 2.6765, "step": 98 }, { "epoch": 0.2916053019145803, "grad_norm": 11.754383087158203, "learning_rate": 9.900000000000001e-05, "loss": 2.323, "step": 99 }, { "epoch": 0.29455081001472755, "grad_norm": 20.766103744506836, "learning_rate": 0.0001, "loss": 3.2395, "step": 100 }, { "epoch": 0.2974963181148748, "grad_norm": 19.38508415222168, "learning_rate": 9.999568045802217e-05, "loss": 3.1611, "step": 101 }, { "epoch": 0.3004418262150221, "grad_norm": 12.322184562683105, "learning_rate": 9.998272257842641e-05, "loss": 1.8749, "step": 102 }, { "epoch": 0.30338733431516934, "grad_norm": 8.819778442382812, "learning_rate": 9.996112860009688e-05, "loss": 2.1352, "step": 103 }, { "epoch": 0.30633284241531666, "grad_norm": 4.589802265167236, "learning_rate": 9.993090225407743e-05, "loss": 1.9195, "step": 104 }, { "epoch": 0.30927835051546393, "grad_norm": 3.1305477619171143, "learning_rate": 9.989204876292688e-05, "loss": 1.6522, "step": 105 }, { "epoch": 0.3122238586156112, "grad_norm": 2.7143571376800537, "learning_rate": 9.984457483981669e-05, "loss": 1.1763, "step": 106 }, { "epoch": 0.31516936671575846, "grad_norm": 2.8804931640625, "learning_rate": 9.978848868737098e-05, "loss": 1.7974, "step": 107 }, { "epoch": 0.3181148748159057, "grad_norm": 3.790278673171997, "learning_rate": 9.972379999624936e-05, "loss": 2.1482, "step": 108 }, { "epoch": 0.32106038291605304, "grad_norm": 4.674582004547119, "learning_rate": 9.96505199434725e-05, "loss": 2.5313, "step": 109 }, { "epoch": 0.3240058910162003, "grad_norm": 3.20375657081604, "learning_rate": 9.956866119049095e-05, "loss": 2.2333, "step": 110 }, { "epoch": 0.3269513991163476, "grad_norm": 3.65432071685791, "learning_rate": 9.947823788099753e-05, "loss": 2.3037, "step": 111 }, { "epoch": 0.32989690721649484, "grad_norm": 4.25070858001709, "learning_rate": 9.937926563848346e-05, "loss": 2.1421, "step": 112 }, { "epoch": 0.3328424153166421, "grad_norm": 3.4754691123962402, "learning_rate": 9.927176156353899e-05, "loss": 1.6731, "step": 113 }, { "epoch": 0.33578792341678937, "grad_norm": 4.312386989593506, "learning_rate": 9.91557442308987e-05, "loss": 2.2766, "step": 114 }, { "epoch": 0.3387334315169367, "grad_norm": 8.520076751708984, "learning_rate": 9.903123368623216e-05, "loss": 2.3844, "step": 115 }, { "epoch": 0.34167893961708395, "grad_norm": 9.280426979064941, "learning_rate": 9.889825144268029e-05, "loss": 2.5895, "step": 116 }, { "epoch": 0.3446244477172312, "grad_norm": 5.867470741271973, "learning_rate": 9.875682047713846e-05, "loss": 2.1818, "step": 117 }, { "epoch": 0.3475699558173785, "grad_norm": 4.29853630065918, "learning_rate": 9.860696522628639e-05, "loss": 2.1243, "step": 118 }, { "epoch": 0.35051546391752575, "grad_norm": 4.369655132293701, "learning_rate": 9.844871158236591e-05, "loss": 1.9835, "step": 119 }, { "epoch": 0.35346097201767307, "grad_norm": 15.677599906921387, "learning_rate": 9.828208688870735e-05, "loss": 2.3591, "step": 120 }, { "epoch": 0.35640648011782033, "grad_norm": 7.070474624633789, "learning_rate": 9.810711993500507e-05, "loss": 2.1378, "step": 121 }, { "epoch": 0.3593519882179676, "grad_norm": 4.5425004959106445, "learning_rate": 9.792384095234313e-05, "loss": 1.838, "step": 122 }, { "epoch": 0.36229749631811486, "grad_norm": 5.75803804397583, "learning_rate": 9.773228160797188e-05, "loss": 2.045, "step": 123 }, { "epoch": 0.36524300441826213, "grad_norm": 4.905186653137207, "learning_rate": 9.753247499983649e-05, "loss": 2.263, "step": 124 }, { "epoch": 0.36818851251840945, "grad_norm": 5.079438209533691, "learning_rate": 9.732445565085824e-05, "loss": 1.9375, "step": 125 }, { "epoch": 0.3711340206185567, "grad_norm": 5.083024978637695, "learning_rate": 9.71082595029695e-05, "loss": 1.9512, "step": 126 }, { "epoch": 0.374079528718704, "grad_norm": 5.701948642730713, "learning_rate": 9.688392391090373e-05, "loss": 2.0932, "step": 127 }, { "epoch": 0.37702503681885124, "grad_norm": 5.612240314483643, "learning_rate": 9.665148763574123e-05, "loss": 2.1218, "step": 128 }, { "epoch": 0.3799705449189985, "grad_norm": 5.693153381347656, "learning_rate": 9.64109908382119e-05, "loss": 2.4763, "step": 129 }, { "epoch": 0.38291605301914583, "grad_norm": 5.299254417419434, "learning_rate": 9.616247507175623e-05, "loss": 2.3062, "step": 130 }, { "epoch": 0.3858615611192931, "grad_norm": 5.711902618408203, "learning_rate": 9.590598327534564e-05, "loss": 2.4087, "step": 131 }, { "epoch": 0.38880706921944036, "grad_norm": 4.181498050689697, "learning_rate": 9.564155976606339e-05, "loss": 1.7175, "step": 132 }, { "epoch": 0.3917525773195876, "grad_norm": 5.31155252456665, "learning_rate": 9.536925023144742e-05, "loss": 1.6347, "step": 133 }, { "epoch": 0.3946980854197349, "grad_norm": 6.574872970581055, "learning_rate": 9.508910172159635e-05, "loss": 1.8686, "step": 134 }, { "epoch": 0.39764359351988215, "grad_norm": 6.77908992767334, "learning_rate": 9.480116264104011e-05, "loss": 2.1223, "step": 135 }, { "epoch": 0.4005891016200295, "grad_norm": 7.450313091278076, "learning_rate": 9.450548274037653e-05, "loss": 2.3148, "step": 136 }, { "epoch": 0.40353460972017674, "grad_norm": 5.135898113250732, "learning_rate": 9.420211310767533e-05, "loss": 1.386, "step": 137 }, { "epoch": 0.406480117820324, "grad_norm": 8.198874473571777, "learning_rate": 9.389110615965102e-05, "loss": 1.9208, "step": 138 }, { "epoch": 0.40942562592047127, "grad_norm": 10.153027534484863, "learning_rate": 9.35725156326063e-05, "loss": 2.2577, "step": 139 }, { "epoch": 0.41237113402061853, "grad_norm": 8.334872245788574, "learning_rate": 9.324639657314742e-05, "loss": 2.0401, "step": 140 }, { "epoch": 0.41531664212076586, "grad_norm": 6.7754950523376465, "learning_rate": 9.291280532867302e-05, "loss": 1.5202, "step": 141 }, { "epoch": 0.4182621502209131, "grad_norm": 9.433147430419922, "learning_rate": 9.257179953763845e-05, "loss": 2.1687, "step": 142 }, { "epoch": 0.4212076583210604, "grad_norm": 7.446643352508545, "learning_rate": 9.222343811959693e-05, "loss": 1.6996, "step": 143 }, { "epoch": 0.42415316642120765, "grad_norm": 7.048992156982422, "learning_rate": 9.186778126501916e-05, "loss": 1.5279, "step": 144 }, { "epoch": 0.4270986745213549, "grad_norm": 7.340887069702148, "learning_rate": 9.150489042489367e-05, "loss": 1.4406, "step": 145 }, { "epoch": 0.43004418262150224, "grad_norm": 11.633044242858887, "learning_rate": 9.113482830010918e-05, "loss": 1.5187, "step": 146 }, { "epoch": 0.4329896907216495, "grad_norm": 9.241588592529297, "learning_rate": 9.075765883062093e-05, "loss": 1.8564, "step": 147 }, { "epoch": 0.43593519882179677, "grad_norm": 8.745128631591797, "learning_rate": 9.037344718440322e-05, "loss": 1.5419, "step": 148 }, { "epoch": 0.43888070692194403, "grad_norm": 16.220924377441406, "learning_rate": 8.99822597461894e-05, "loss": 2.5006, "step": 149 }, { "epoch": 0.4418262150220913, "grad_norm": 15.44288444519043, "learning_rate": 8.958416410600187e-05, "loss": 2.0001, "step": 150 }, { "epoch": 0.44477172312223856, "grad_norm": 10.727198600769043, "learning_rate": 8.917922904747384e-05, "loss": 2.8376, "step": 151 }, { "epoch": 0.4477172312223859, "grad_norm": 5.658038139343262, "learning_rate": 8.876752453596462e-05, "loss": 1.7435, "step": 152 }, { "epoch": 0.45066273932253315, "grad_norm": 4.7606892585754395, "learning_rate": 8.834912170647101e-05, "loss": 1.3439, "step": 153 }, { "epoch": 0.4536082474226804, "grad_norm": 4.7208757400512695, "learning_rate": 8.792409285133642e-05, "loss": 1.8022, "step": 154 }, { "epoch": 0.4565537555228277, "grad_norm": 3.1722121238708496, "learning_rate": 8.749251140776016e-05, "loss": 0.7509, "step": 155 }, { "epoch": 0.45949926362297494, "grad_norm": 2.528482675552368, "learning_rate": 8.705445194510868e-05, "loss": 1.6068, "step": 156 }, { "epoch": 0.46244477172312226, "grad_norm": 2.8189425468444824, "learning_rate": 8.66099901520315e-05, "loss": 0.4364, "step": 157 }, { "epoch": 0.4653902798232695, "grad_norm": 2.734297752380371, "learning_rate": 8.615920282338355e-05, "loss": 1.6278, "step": 158 }, { "epoch": 0.4683357879234168, "grad_norm": 2.981930732727051, "learning_rate": 8.570216784695637e-05, "loss": 1.9541, "step": 159 }, { "epoch": 0.47128129602356406, "grad_norm": 3.195127248764038, "learning_rate": 8.52389641900206e-05, "loss": 2.0949, "step": 160 }, { "epoch": 0.4742268041237113, "grad_norm": 3.8657867908477783, "learning_rate": 8.476967188568188e-05, "loss": 1.8021, "step": 161 }, { "epoch": 0.47717231222385864, "grad_norm": 3.0095670223236084, "learning_rate": 8.429437201905254e-05, "loss": 1.8119, "step": 162 }, { "epoch": 0.4801178203240059, "grad_norm": 3.025339365005493, "learning_rate": 8.381314671324159e-05, "loss": 1.8117, "step": 163 }, { "epoch": 0.48306332842415317, "grad_norm": 4.113118648529053, "learning_rate": 8.332607911516545e-05, "loss": 1.7912, "step": 164 }, { "epoch": 0.48600883652430044, "grad_norm": 3.183894395828247, "learning_rate": 8.283325338118153e-05, "loss": 1.7364, "step": 165 }, { "epoch": 0.4889543446244477, "grad_norm": 5.081910610198975, "learning_rate": 8.233475466254765e-05, "loss": 2.0706, "step": 166 }, { "epoch": 0.49189985272459497, "grad_norm": 4.033633708953857, "learning_rate": 8.183066909070947e-05, "loss": 2.0481, "step": 167 }, { "epoch": 0.4948453608247423, "grad_norm": 4.505871772766113, "learning_rate": 8.132108376241849e-05, "loss": 2.044, "step": 168 }, { "epoch": 0.49779086892488955, "grad_norm": 6.440969944000244, "learning_rate": 8.08060867246834e-05, "loss": 2.209, "step": 169 }, { "epoch": 0.5007363770250368, "grad_norm": 6.235000133514404, "learning_rate": 8.028576695955711e-05, "loss": 2.0359, "step": 170 }, { "epoch": 0.5036818851251841, "grad_norm": 4.07951021194458, "learning_rate": 7.97602143687623e-05, "loss": 1.6721, "step": 171 }, { "epoch": 0.5066273932253313, "grad_norm": 4.1656413078308105, "learning_rate": 7.922951975815811e-05, "loss": 1.7503, "step": 172 }, { "epoch": 0.5095729013254786, "grad_norm": 3.7077293395996094, "learning_rate": 7.869377482205042e-05, "loss": 1.6881, "step": 173 }, { "epoch": 0.5125184094256259, "grad_norm": 4.08639669418335, "learning_rate": 7.815307212734888e-05, "loss": 1.4451, "step": 174 }, { "epoch": 0.5154639175257731, "grad_norm": 4.255699157714844, "learning_rate": 7.760750509757298e-05, "loss": 1.7914, "step": 175 }, { "epoch": 0.5184094256259205, "grad_norm": 4.119191646575928, "learning_rate": 7.705716799671019e-05, "loss": 1.8201, "step": 176 }, { "epoch": 0.5213549337260678, "grad_norm": 4.7409467697143555, "learning_rate": 7.650215591292888e-05, "loss": 1.8449, "step": 177 }, { "epoch": 0.524300441826215, "grad_norm": 5.348622798919678, "learning_rate": 7.594256474214882e-05, "loss": 1.6767, "step": 178 }, { "epoch": 0.5272459499263623, "grad_norm": 4.13510799407959, "learning_rate": 7.537849117147212e-05, "loss": 1.2513, "step": 179 }, { "epoch": 0.5301914580265096, "grad_norm": 4.210190296173096, "learning_rate": 7.481003266247744e-05, "loss": 1.5867, "step": 180 }, { "epoch": 0.5331369661266568, "grad_norm": 5.271762371063232, "learning_rate": 7.423728743438048e-05, "loss": 2.1488, "step": 181 }, { "epoch": 0.5360824742268041, "grad_norm": 5.800075054168701, "learning_rate": 7.366035444706347e-05, "loss": 1.8385, "step": 182 }, { "epoch": 0.5390279823269514, "grad_norm": 6.030768871307373, "learning_rate": 7.307933338397667e-05, "loss": 1.6634, "step": 183 }, { "epoch": 0.5419734904270986, "grad_norm": 5.144318580627441, "learning_rate": 7.249432463491498e-05, "loss": 1.4081, "step": 184 }, { "epoch": 0.5449189985272459, "grad_norm": 5.144754409790039, "learning_rate": 7.190542927867234e-05, "loss": 1.4496, "step": 185 }, { "epoch": 0.5478645066273933, "grad_norm": 7.374088764190674, "learning_rate": 7.131274906557725e-05, "loss": 1.94, "step": 186 }, { "epoch": 0.5508100147275405, "grad_norm": 6.2576003074646, "learning_rate": 7.071638639991207e-05, "loss": 1.701, "step": 187 }, { "epoch": 0.5537555228276878, "grad_norm": 5.392474174499512, "learning_rate": 7.011644432221958e-05, "loss": 1.474, "step": 188 }, { "epoch": 0.5567010309278351, "grad_norm": 7.743690013885498, "learning_rate": 6.95130264914993e-05, "loss": 2.1458, "step": 189 }, { "epoch": 0.5596465390279823, "grad_norm": 7.468958377838135, "learning_rate": 6.890623716729724e-05, "loss": 2.2593, "step": 190 }, { "epoch": 0.5625920471281296, "grad_norm": 6.144321918487549, "learning_rate": 6.82961811916917e-05, "loss": 1.3348, "step": 191 }, { "epoch": 0.5655375552282769, "grad_norm": 7.852849006652832, "learning_rate": 6.768296397117848e-05, "loss": 1.4751, "step": 192 }, { "epoch": 0.5684830633284241, "grad_norm": 7.6272172927856445, "learning_rate": 6.706669145845863e-05, "loss": 1.7606, "step": 193 }, { "epoch": 0.5714285714285714, "grad_norm": 13.624709129333496, "learning_rate": 6.644747013413168e-05, "loss": 1.9716, "step": 194 }, { "epoch": 0.5743740795287187, "grad_norm": 8.707653999328613, "learning_rate": 6.582540698829781e-05, "loss": 1.8245, "step": 195 }, { "epoch": 0.5773195876288659, "grad_norm": 6.5573039054870605, "learning_rate": 6.520060950207185e-05, "loss": 1.1733, "step": 196 }, { "epoch": 0.5802650957290133, "grad_norm": 8.433490753173828, "learning_rate": 6.457318562901256e-05, "loss": 1.2757, "step": 197 }, { "epoch": 0.5832106038291606, "grad_norm": 9.936591148376465, "learning_rate": 6.394324377647028e-05, "loss": 1.4713, "step": 198 }, { "epoch": 0.5861561119293078, "grad_norm": 12.52322769165039, "learning_rate": 6.331089278685599e-05, "loss": 1.6688, "step": 199 }, { "epoch": 0.5891016200294551, "grad_norm": 16.770870208740234, "learning_rate": 6.26762419188355e-05, "loss": 2.3924, "step": 200 }, { "epoch": 0.5920471281296024, "grad_norm": 3.7220280170440674, "learning_rate": 6.203940082845144e-05, "loss": 1.8733, "step": 201 }, { "epoch": 0.5949926362297496, "grad_norm": 3.919069290161133, "learning_rate": 6.140047955017671e-05, "loss": 0.7927, "step": 202 }, { "epoch": 0.5979381443298969, "grad_norm": 3.081045150756836, "learning_rate": 6.075958847790262e-05, "loss": 1.2878, "step": 203 }, { "epoch": 0.6008836524300442, "grad_norm": 2.82151198387146, "learning_rate": 6.011683834586473e-05, "loss": 1.2722, "step": 204 }, { "epoch": 0.6038291605301914, "grad_norm": 2.1925573348999023, "learning_rate": 5.947234020951015e-05, "loss": 0.7102, "step": 205 }, { "epoch": 0.6067746686303387, "grad_norm": 2.2746307849884033, "learning_rate": 5.882620542630901e-05, "loss": 0.8639, "step": 206 }, { "epoch": 0.6097201767304861, "grad_norm": 2.5907399654388428, "learning_rate": 5.8178545636514145e-05, "loss": 1.3898, "step": 207 }, { "epoch": 0.6126656848306333, "grad_norm": 3.3455467224121094, "learning_rate": 5.752947274387147e-05, "loss": 1.6453, "step": 208 }, { "epoch": 0.6156111929307806, "grad_norm": 3.1584908962249756, "learning_rate": 5.687909889628529e-05, "loss": 1.8621, "step": 209 }, { "epoch": 0.6185567010309279, "grad_norm": 3.1960365772247314, "learning_rate": 5.622753646644102e-05, "loss": 1.6358, "step": 210 }, { "epoch": 0.6215022091310751, "grad_norm": 3.073702573776245, "learning_rate": 5.557489803238933e-05, "loss": 1.5381, "step": 211 }, { "epoch": 0.6244477172312224, "grad_norm": 2.917039632797241, "learning_rate": 5.492129635809473e-05, "loss": 1.6452, "step": 212 }, { "epoch": 0.6273932253313697, "grad_norm": 3.823730707168579, "learning_rate": 5.426684437395196e-05, "loss": 1.7625, "step": 213 }, { "epoch": 0.6303387334315169, "grad_norm": 4.05581521987915, "learning_rate": 5.361165515727374e-05, "loss": 1.9468, "step": 214 }, { "epoch": 0.6332842415316642, "grad_norm": 3.9730498790740967, "learning_rate": 5.295584191275308e-05, "loss": 2.2487, "step": 215 }, { "epoch": 0.6362297496318114, "grad_norm": 3.2518630027770996, "learning_rate": 5.229951795290353e-05, "loss": 1.5587, "step": 216 }, { "epoch": 0.6391752577319587, "grad_norm": 4.549688816070557, "learning_rate": 5.164279667848094e-05, "loss": 1.9062, "step": 217 }, { "epoch": 0.6421207658321061, "grad_norm": 3.6768784523010254, "learning_rate": 5.0985791558889785e-05, "loss": 1.6801, "step": 218 }, { "epoch": 0.6450662739322534, "grad_norm": 3.8859946727752686, "learning_rate": 5.032861611257783e-05, "loss": 1.9869, "step": 219 }, { "epoch": 0.6480117820324006, "grad_norm": 4.19022798538208, "learning_rate": 4.967138388742218e-05, "loss": 2.0944, "step": 220 }, { "epoch": 0.6509572901325479, "grad_norm": 3.7361319065093994, "learning_rate": 4.901420844111021e-05, "loss": 1.3391, "step": 221 }, { "epoch": 0.6539027982326951, "grad_norm": 3.7620272636413574, "learning_rate": 4.835720332151907e-05, "loss": 1.2597, "step": 222 }, { "epoch": 0.6568483063328424, "grad_norm": 3.610649824142456, "learning_rate": 4.770048204709648e-05, "loss": 1.2885, "step": 223 }, { "epoch": 0.6597938144329897, "grad_norm": 4.286924839019775, "learning_rate": 4.7044158087246926e-05, "loss": 1.7704, "step": 224 }, { "epoch": 0.6627393225331369, "grad_norm": 4.295466423034668, "learning_rate": 4.6388344842726264e-05, "loss": 1.3672, "step": 225 }, { "epoch": 0.6656848306332842, "grad_norm": 5.368373870849609, "learning_rate": 4.5733155626048036e-05, "loss": 1.5249, "step": 226 }, { "epoch": 0.6686303387334315, "grad_norm": 5.0956950187683105, "learning_rate": 4.507870364190527e-05, "loss": 1.3973, "step": 227 }, { "epoch": 0.6715758468335787, "grad_norm": 4.604374408721924, "learning_rate": 4.4425101967610674e-05, "loss": 1.3965, "step": 228 }, { "epoch": 0.6745213549337261, "grad_norm": 6.400148391723633, "learning_rate": 4.377246353355899e-05, "loss": 2.1319, "step": 229 }, { "epoch": 0.6774668630338734, "grad_norm": 6.704805850982666, "learning_rate": 4.312090110371473e-05, "loss": 1.9594, "step": 230 }, { "epoch": 0.6804123711340206, "grad_norm": 4.956854343414307, "learning_rate": 4.247052725612852e-05, "loss": 1.3392, "step": 231 }, { "epoch": 0.6833578792341679, "grad_norm": 8.188490867614746, "learning_rate": 4.1821454363485866e-05, "loss": 1.7636, "step": 232 }, { "epoch": 0.6863033873343152, "grad_norm": 5.5076751708984375, "learning_rate": 4.1173794573690996e-05, "loss": 1.3066, "step": 233 }, { "epoch": 0.6892488954344624, "grad_norm": 6.480940341949463, "learning_rate": 4.052765979048986e-05, "loss": 1.8326, "step": 234 }, { "epoch": 0.6921944035346097, "grad_norm": 7.0487284660339355, "learning_rate": 3.988316165413528e-05, "loss": 1.7487, "step": 235 }, { "epoch": 0.695139911634757, "grad_norm": 6.207117557525635, "learning_rate": 3.924041152209739e-05, "loss": 1.4652, "step": 236 }, { "epoch": 0.6980854197349042, "grad_norm": 5.3427205085754395, "learning_rate": 3.859952044982329e-05, "loss": 1.3738, "step": 237 }, { "epoch": 0.7010309278350515, "grad_norm": 7.366813659667969, "learning_rate": 3.7960599171548574e-05, "loss": 1.4897, "step": 238 }, { "epoch": 0.7039764359351989, "grad_norm": 6.943472862243652, "learning_rate": 3.732375808116451e-05, "loss": 1.2526, "step": 239 }, { "epoch": 0.7069219440353461, "grad_norm": 6.768280506134033, "learning_rate": 3.668910721314402e-05, "loss": 1.215, "step": 240 }, { "epoch": 0.7098674521354934, "grad_norm": 8.1182279586792, "learning_rate": 3.605675622352973e-05, "loss": 1.8819, "step": 241 }, { "epoch": 0.7128129602356407, "grad_norm": 6.481826305389404, "learning_rate": 3.542681437098745e-05, "loss": 1.0718, "step": 242 }, { "epoch": 0.7157584683357879, "grad_norm": 6.8479485511779785, "learning_rate": 3.479939049792817e-05, "loss": 1.0558, "step": 243 }, { "epoch": 0.7187039764359352, "grad_norm": 7.331018924713135, "learning_rate": 3.417459301170219e-05, "loss": 1.2658, "step": 244 }, { "epoch": 0.7216494845360825, "grad_norm": 7.68209981918335, "learning_rate": 3.355252986586832e-05, "loss": 1.2073, "step": 245 }, { "epoch": 0.7245949926362297, "grad_norm": 11.420364379882812, "learning_rate": 3.293330854154136e-05, "loss": 1.9316, "step": 246 }, { "epoch": 0.727540500736377, "grad_norm": 7.791138172149658, "learning_rate": 3.2317036028821523e-05, "loss": 1.3763, "step": 247 }, { "epoch": 0.7304860088365243, "grad_norm": 8.121346473693848, "learning_rate": 3.1703818808308324e-05, "loss": 1.2789, "step": 248 }, { "epoch": 0.7334315169366715, "grad_norm": 11.62152099609375, "learning_rate": 3.109376283270277e-05, "loss": 1.9576, "step": 249 }, { "epoch": 0.7363770250368189, "grad_norm": 16.3736515045166, "learning_rate": 3.0486973508500727e-05, "loss": 1.8166, "step": 250 }, { "epoch": 0.7393225331369662, "grad_norm": 2.2238049507141113, "learning_rate": 2.988355567778043e-05, "loss": 1.8472, "step": 251 }, { "epoch": 0.7422680412371134, "grad_norm": 2.6677799224853516, "learning_rate": 2.9283613600087933e-05, "loss": 1.3083, "step": 252 }, { "epoch": 0.7452135493372607, "grad_norm": 2.822364568710327, "learning_rate": 2.8687250934422772e-05, "loss": 0.9437, "step": 253 }, { "epoch": 0.748159057437408, "grad_norm": 2.1095311641693115, "learning_rate": 2.8094570721327662e-05, "loss": 0.9748, "step": 254 }, { "epoch": 0.7511045655375552, "grad_norm": 1.7905985116958618, "learning_rate": 2.750567536508504e-05, "loss": 0.881, "step": 255 }, { "epoch": 0.7540500736377025, "grad_norm": 2.3329577445983887, "learning_rate": 2.6920666616023327e-05, "loss": 1.2698, "step": 256 }, { "epoch": 0.7569955817378498, "grad_norm": 1.4233232736587524, "learning_rate": 2.6339645552936536e-05, "loss": 0.6157, "step": 257 }, { "epoch": 0.759941089837997, "grad_norm": 2.3130457401275635, "learning_rate": 2.5762712565619528e-05, "loss": 1.1048, "step": 258 }, { "epoch": 0.7628865979381443, "grad_norm": 3.3101067543029785, "learning_rate": 2.5189967337522573e-05, "loss": 1.9625, "step": 259 }, { "epoch": 0.7658321060382917, "grad_norm": 3.263514995574951, "learning_rate": 2.46215088285279e-05, "loss": 1.9653, "step": 260 }, { "epoch": 0.7687776141384389, "grad_norm": 4.304958820343018, "learning_rate": 2.4057435257851175e-05, "loss": 1.7951, "step": 261 }, { "epoch": 0.7717231222385862, "grad_norm": 3.644676685333252, "learning_rate": 2.349784408707112e-05, "loss": 1.7257, "step": 262 }, { "epoch": 0.7746686303387335, "grad_norm": 3.691222906112671, "learning_rate": 2.2942832003289823e-05, "loss": 1.3952, "step": 263 }, { "epoch": 0.7776141384388807, "grad_norm": 3.36948823928833, "learning_rate": 2.2392494902427025e-05, "loss": 1.5119, "step": 264 }, { "epoch": 0.780559646539028, "grad_norm": 3.5477828979492188, "learning_rate": 2.1846927872651137e-05, "loss": 1.6785, "step": 265 }, { "epoch": 0.7835051546391752, "grad_norm": 3.44746470451355, "learning_rate": 2.1306225177949585e-05, "loss": 1.3705, "step": 266 }, { "epoch": 0.7864506627393225, "grad_norm": 5.451107501983643, "learning_rate": 2.07704802418419e-05, "loss": 2.1179, "step": 267 }, { "epoch": 0.7893961708394698, "grad_norm": 3.672868251800537, "learning_rate": 2.0239785631237705e-05, "loss": 1.6712, "step": 268 }, { "epoch": 0.792341678939617, "grad_norm": 3.347046375274658, "learning_rate": 1.9714233040442915e-05, "loss": 1.466, "step": 269 }, { "epoch": 0.7952871870397643, "grad_norm": 3.741647481918335, "learning_rate": 1.9193913275316626e-05, "loss": 1.5614, "step": 270 }, { "epoch": 0.7982326951399117, "grad_norm": 3.9401445388793945, "learning_rate": 1.8678916237581522e-05, "loss": 1.3664, "step": 271 }, { "epoch": 0.801178203240059, "grad_norm": 4.1288604736328125, "learning_rate": 1.816933090929055e-05, "loss": 1.5295, "step": 272 }, { "epoch": 0.8041237113402062, "grad_norm": 3.8962388038635254, "learning_rate": 1.7665245337452368e-05, "loss": 1.5394, "step": 273 }, { "epoch": 0.8070692194403535, "grad_norm": 3.2219879627227783, "learning_rate": 1.716674661881848e-05, "loss": 1.1269, "step": 274 }, { "epoch": 0.8100147275405007, "grad_norm": 3.5994858741760254, "learning_rate": 1.667392088483456e-05, "loss": 1.1873, "step": 275 }, { "epoch": 0.812960235640648, "grad_norm": 4.191354274749756, "learning_rate": 1.6186853286758397e-05, "loss": 1.6051, "step": 276 }, { "epoch": 0.8159057437407953, "grad_norm": 3.547515392303467, "learning_rate": 1.570562798094747e-05, "loss": 1.2364, "step": 277 }, { "epoch": 0.8188512518409425, "grad_norm": 7.393482208251953, "learning_rate": 1.5230328114318127e-05, "loss": 1.4475, "step": 278 }, { "epoch": 0.8217967599410898, "grad_norm": 3.847755193710327, "learning_rate": 1.4761035809979395e-05, "loss": 1.1197, "step": 279 }, { "epoch": 0.8247422680412371, "grad_norm": 4.247459888458252, "learning_rate": 1.4297832153043656e-05, "loss": 1.2487, "step": 280 }, { "epoch": 0.8276877761413843, "grad_norm": 5.425104141235352, "learning_rate": 1.3840797176616466e-05, "loss": 1.7611, "step": 281 }, { "epoch": 0.8306332842415317, "grad_norm": 4.583087921142578, "learning_rate": 1.3390009847968504e-05, "loss": 1.4049, "step": 282 }, { "epoch": 0.833578792341679, "grad_norm": 8.560991287231445, "learning_rate": 1.2945548054891321e-05, "loss": 1.2722, "step": 283 }, { "epoch": 0.8365243004418262, "grad_norm": 5.156883716583252, "learning_rate": 1.2507488592239847e-05, "loss": 1.3135, "step": 284 }, { "epoch": 0.8394698085419735, "grad_norm": 4.798750400543213, "learning_rate": 1.2075907148663579e-05, "loss": 1.0881, "step": 285 }, { "epoch": 0.8424153166421208, "grad_norm": 5.219260215759277, "learning_rate": 1.1650878293528994e-05, "loss": 1.3982, "step": 286 }, { "epoch": 0.845360824742268, "grad_norm": 5.036367893218994, "learning_rate": 1.1232475464035385e-05, "loss": 1.2723, "step": 287 }, { "epoch": 0.8483063328424153, "grad_norm": 5.566346645355225, "learning_rate": 1.0820770952526155e-05, "loss": 0.7823, "step": 288 }, { "epoch": 0.8512518409425626, "grad_norm": 4.645596504211426, "learning_rate": 1.0415835893998116e-05, "loss": 0.9811, "step": 289 }, { "epoch": 0.8541973490427098, "grad_norm": 5.099687099456787, "learning_rate": 1.0017740253810609e-05, "loss": 1.0945, "step": 290 }, { "epoch": 0.8571428571428571, "grad_norm": 5.225978374481201, "learning_rate": 9.62655281559679e-06, "loss": 1.0845, "step": 291 }, { "epoch": 0.8600883652430045, "grad_norm": 9.22032356262207, "learning_rate": 9.242341169379076e-06, "loss": 1.221, "step": 292 }, { "epoch": 0.8630338733431517, "grad_norm": 9.09154987335205, "learning_rate": 8.865171699890834e-06, "loss": 1.7676, "step": 293 }, { "epoch": 0.865979381443299, "grad_norm": 6.99169397354126, "learning_rate": 8.49510957510633e-06, "loss": 1.1186, "step": 294 }, { "epoch": 0.8689248895434463, "grad_norm": 7.78825044631958, "learning_rate": 8.132218734980852e-06, "loss": 1.1668, "step": 295 }, { "epoch": 0.8718703976435935, "grad_norm": 6.3908371925354, "learning_rate": 7.776561880403072e-06, "loss": 1.1461, "step": 296 }, { "epoch": 0.8748159057437408, "grad_norm": 8.278850555419922, "learning_rate": 7.4282004623615396e-06, "loss": 1.2128, "step": 297 }, { "epoch": 0.8777614138438881, "grad_norm": 8.332165718078613, "learning_rate": 7.0871946713269856e-06, "loss": 1.016, "step": 298 }, { "epoch": 0.8807069219440353, "grad_norm": 14.03607177734375, "learning_rate": 6.753603426852589e-06, "loss": 1.7378, "step": 299 }, { "epoch": 0.8836524300441826, "grad_norm": 15.087005615234375, "learning_rate": 6.427484367393699e-06, "loss": 1.6779, "step": 300 }, { "epoch": 0.8865979381443299, "grad_norm": 2.469926357269287, "learning_rate": 6.108893840348995e-06, "loss": 1.7045, "step": 301 }, { "epoch": 0.8895434462444771, "grad_norm": 2.2709763050079346, "learning_rate": 5.797886892324694e-06, "loss": 0.9064, "step": 302 }, { "epoch": 0.8924889543446245, "grad_norm": 2.1486294269561768, "learning_rate": 5.494517259623477e-06, "loss": 1.3705, "step": 303 }, { "epoch": 0.8954344624447718, "grad_norm": 1.1387767791748047, "learning_rate": 5.198837358959901e-06, "loss": 0.2727, "step": 304 }, { "epoch": 0.898379970544919, "grad_norm": 4.750706672668457, "learning_rate": 4.910898278403669e-06, "loss": 0.9473, "step": 305 }, { "epoch": 0.9013254786450663, "grad_norm": 3.402461051940918, "learning_rate": 4.630749768552589e-06, "loss": 1.3646, "step": 306 }, { "epoch": 0.9042709867452136, "grad_norm": 3.262354612350464, "learning_rate": 4.358440233936617e-06, "loss": 1.6704, "step": 307 }, { "epoch": 0.9072164948453608, "grad_norm": 2.7272191047668457, "learning_rate": 4.094016724654359e-06, "loss": 1.4253, "step": 308 }, { "epoch": 0.9101620029455081, "grad_norm": 3.3352503776550293, "learning_rate": 3.837524928243774e-06, "loss": 1.5388, "step": 309 }, { "epoch": 0.9131075110456554, "grad_norm": 2.7655768394470215, "learning_rate": 3.589009161788104e-06, "loss": 1.3632, "step": 310 }, { "epoch": 0.9160530191458026, "grad_norm": 3.8515467643737793, "learning_rate": 3.3485123642587658e-06, "loss": 1.822, "step": 311 }, { "epoch": 0.9189985272459499, "grad_norm": 3.5630061626434326, "learning_rate": 3.116076089096265e-06, "loss": 1.6991, "step": 312 }, { "epoch": 0.9219440353460973, "grad_norm": 4.128489017486572, "learning_rate": 2.8917404970305097e-06, "loss": 1.3073, "step": 313 }, { "epoch": 0.9248895434462445, "grad_norm": 4.167806148529053, "learning_rate": 2.675544349141779e-06, "loss": 1.73, "step": 314 }, { "epoch": 0.9278350515463918, "grad_norm": 4.222491264343262, "learning_rate": 2.4675250001635232e-06, "loss": 1.6474, "step": 315 }, { "epoch": 0.930780559646539, "grad_norm": 3.9301912784576416, "learning_rate": 2.2677183920281343e-06, "loss": 1.8446, "step": 316 }, { "epoch": 0.9337260677466863, "grad_norm": 10.473541259765625, "learning_rate": 2.076159047656889e-06, "loss": 1.7129, "step": 317 }, { "epoch": 0.9366715758468336, "grad_norm": 3.969463348388672, "learning_rate": 1.892880064994934e-06, "loss": 1.2407, "step": 318 }, { "epoch": 0.9396170839469808, "grad_norm": 4.088743686676025, "learning_rate": 1.7179131112926627e-06, "loss": 1.5257, "step": 319 }, { "epoch": 0.9425625920471281, "grad_norm": 3.984471559524536, "learning_rate": 1.551288417634106e-06, "loss": 1.2564, "step": 320 }, { "epoch": 0.9455081001472754, "grad_norm": 7.943843841552734, "learning_rate": 1.3930347737136196e-06, "loss": 1.8827, "step": 321 }, { "epoch": 0.9484536082474226, "grad_norm": 5.492068767547607, "learning_rate": 1.2431795228615372e-06, "loss": 1.7856, "step": 322 }, { "epoch": 0.9513991163475699, "grad_norm": 4.9044647216796875, "learning_rate": 1.101748557319715e-06, "loss": 1.4158, "step": 323 }, { "epoch": 0.9543446244477173, "grad_norm": 4.557461261749268, "learning_rate": 9.687663137678604e-07, "loss": 1.1651, "step": 324 }, { "epoch": 0.9572901325478645, "grad_norm": 5.330766201019287, "learning_rate": 8.442557691013043e-07, "loss": 1.3875, "step": 325 }, { "epoch": 0.9602356406480118, "grad_norm": 4.67563009262085, "learning_rate": 7.282384364610206e-07, "loss": 1.0563, "step": 326 }, { "epoch": 0.9631811487481591, "grad_norm": 5.674734592437744, "learning_rate": 6.207343615165561e-07, "loss": 1.1674, "step": 327 }, { "epoch": 0.9661266568483063, "grad_norm": 6.089367866516113, "learning_rate": 5.217621190024779e-07, "loss": 1.1991, "step": 328 }, { "epoch": 0.9690721649484536, "grad_norm": 5.158655643463135, "learning_rate": 4.3133880950905205e-07, "loss": 1.2847, "step": 329 }, { "epoch": 0.9720176730486009, "grad_norm": 5.49578332901001, "learning_rate": 3.494800565275125e-07, "loss": 1.1238, "step": 330 }, { "epoch": 0.9749631811487481, "grad_norm": 5.605154991149902, "learning_rate": 2.762000037506485e-07, "loss": 0.964, "step": 331 }, { "epoch": 0.9779086892488954, "grad_norm": 7.648482799530029, "learning_rate": 2.115113126290258e-07, "loss": 1.4449, "step": 332 }, { "epoch": 0.9808541973490427, "grad_norm": 8.961212158203125, "learning_rate": 1.554251601833201e-07, "loss": 1.5872, "step": 333 }, { "epoch": 0.9837997054491899, "grad_norm": 7.242101669311523, "learning_rate": 1.0795123707312283e-07, "loss": 1.0052, "step": 334 }, { "epoch": 0.9867452135493373, "grad_norm": 7.433496475219727, "learning_rate": 6.909774592258056e-08, "loss": 1.1363, "step": 335 }, { "epoch": 0.9896907216494846, "grad_norm": 7.9218854904174805, "learning_rate": 3.8871399903134265e-08, "loss": 1.1405, "step": 336 }, { "epoch": 0.9926362297496318, "grad_norm": 6.639596462249756, "learning_rate": 1.7277421573608232e-08, "loss": 1.1224, "step": 337 }, { "epoch": 0.9955817378497791, "grad_norm": 9.504895210266113, "learning_rate": 4.319541977831909e-09, "loss": 1.5028, "step": 338 }, { "epoch": 0.9985272459499264, "grad_norm": 8.52180290222168, "learning_rate": 0.0, "loss": 0.8636, "step": 339 } ], "logging_steps": 1, "max_steps": 339, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 239, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7.500002096893133e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }