{ "best_metric": null, "best_model_checkpoint": null, "epoch": 18.46153846153846, "eval_steps": 500, "global_step": 180, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.10256410256410256, "grad_norm": 0.1054198294878006, "learning_rate": 5.555555555555556e-06, "loss": 0.2336, "step": 1 }, { "epoch": 0.20512820512820512, "grad_norm": 0.10983864963054657, "learning_rate": 1.1111111111111112e-05, "loss": 0.2488, "step": 2 }, { "epoch": 0.3076923076923077, "grad_norm": 0.10180113464593887, "learning_rate": 1.6666666666666667e-05, "loss": 0.2371, "step": 3 }, { "epoch": 0.41025641025641024, "grad_norm": 0.10248314589262009, "learning_rate": 2.2222222222222223e-05, "loss": 0.2386, "step": 4 }, { "epoch": 0.5128205128205128, "grad_norm": 0.10735535621643066, "learning_rate": 2.777777777777778e-05, "loss": 0.2428, "step": 5 }, { "epoch": 0.6153846153846154, "grad_norm": 0.10470724105834961, "learning_rate": 3.3333333333333335e-05, "loss": 0.2283, "step": 6 }, { "epoch": 0.717948717948718, "grad_norm": 0.09391864389181137, "learning_rate": 3.888888888888889e-05, "loss": 0.2275, "step": 7 }, { "epoch": 0.8205128205128205, "grad_norm": 0.11098546534776688, "learning_rate": 4.4444444444444447e-05, "loss": 0.2395, "step": 8 }, { "epoch": 0.9230769230769231, "grad_norm": 0.09393182396888733, "learning_rate": 5e-05, "loss": 0.2279, "step": 9 }, { "epoch": 1.0256410256410255, "grad_norm": 0.08448313176631927, "learning_rate": 5.555555555555556e-05, "loss": 0.2339, "step": 10 }, { "epoch": 1.1282051282051282, "grad_norm": 0.06229618564248085, "learning_rate": 6.111111111111112e-05, "loss": 0.2203, "step": 11 }, { "epoch": 1.2307692307692308, "grad_norm": 0.06508181244134903, "learning_rate": 6.666666666666667e-05, "loss": 0.2163, "step": 12 }, { "epoch": 1.3333333333333333, "grad_norm": 0.061453677713871, "learning_rate": 7.222222222222222e-05, "loss": 0.2244, "step": 13 }, { "epoch": 1.435897435897436, "grad_norm": 0.05588587746024132, "learning_rate": 7.777777777777778e-05, "loss": 0.2196, "step": 14 }, { "epoch": 1.5384615384615383, "grad_norm": 0.05503275617957115, "learning_rate": 8.333333333333334e-05, "loss": 0.2193, "step": 15 }, { "epoch": 1.641025641025641, "grad_norm": 0.07127121835947037, "learning_rate": 8.888888888888889e-05, "loss": 0.2086, "step": 16 }, { "epoch": 1.7435897435897436, "grad_norm": 0.06692970544099808, "learning_rate": 9.444444444444444e-05, "loss": 0.2168, "step": 17 }, { "epoch": 1.8461538461538463, "grad_norm": 0.056347738951444626, "learning_rate": 0.0001, "loss": 0.2117, "step": 18 }, { "epoch": 1.9487179487179487, "grad_norm": 0.05091376230120659, "learning_rate": 9.999059852242507e-05, "loss": 0.2096, "step": 19 }, { "epoch": 2.051282051282051, "grad_norm": 0.049117326736450195, "learning_rate": 9.996239762521151e-05, "loss": 0.2066, "step": 20 }, { "epoch": 2.1538461538461537, "grad_norm": 0.04978122562170029, "learning_rate": 9.991540791356342e-05, "loss": 0.2087, "step": 21 }, { "epoch": 2.2564102564102564, "grad_norm": 0.045931246131658554, "learning_rate": 9.98496470583896e-05, "loss": 0.2027, "step": 22 }, { "epoch": 2.358974358974359, "grad_norm": 0.05703091621398926, "learning_rate": 9.976513978965829e-05, "loss": 0.2146, "step": 23 }, { "epoch": 2.4615384615384617, "grad_norm": 0.042917944490909576, "learning_rate": 9.966191788709716e-05, "loss": 0.2067, "step": 24 }, { "epoch": 2.564102564102564, "grad_norm": 0.04547038674354553, "learning_rate": 9.954002016824227e-05, "loss": 0.2069, "step": 25 }, { "epoch": 2.6666666666666665, "grad_norm": 0.046476855874061584, "learning_rate": 9.939949247384046e-05, "loss": 0.2058, "step": 26 }, { "epoch": 2.769230769230769, "grad_norm": 0.05241086706519127, "learning_rate": 9.924038765061042e-05, "loss": 0.2084, "step": 27 }, { "epoch": 2.871794871794872, "grad_norm": 0.04542744159698486, "learning_rate": 9.906276553136923e-05, "loss": 0.2136, "step": 28 }, { "epoch": 2.9743589743589745, "grad_norm": 0.040408626198768616, "learning_rate": 9.88666929125318e-05, "loss": 0.1986, "step": 29 }, { "epoch": 3.076923076923077, "grad_norm": 0.03904923424124718, "learning_rate": 9.865224352899119e-05, "loss": 0.2128, "step": 30 }, { "epoch": 3.1794871794871793, "grad_norm": 0.04009323939681053, "learning_rate": 9.84194980263903e-05, "loss": 0.1921, "step": 31 }, { "epoch": 3.282051282051282, "grad_norm": 0.03925410285592079, "learning_rate": 9.816854393079403e-05, "loss": 0.2041, "step": 32 }, { "epoch": 3.3846153846153846, "grad_norm": 0.043651480227708817, "learning_rate": 9.789947561577445e-05, "loss": 0.2046, "step": 33 }, { "epoch": 3.4871794871794872, "grad_norm": 0.03763292357325554, "learning_rate": 9.761239426692077e-05, "loss": 0.2064, "step": 34 }, { "epoch": 3.58974358974359, "grad_norm": 0.03577704727649689, "learning_rate": 9.730740784378753e-05, "loss": 0.2108, "step": 35 }, { "epoch": 3.6923076923076925, "grad_norm": 0.03381442651152611, "learning_rate": 9.698463103929542e-05, "loss": 0.1996, "step": 36 }, { "epoch": 3.7948717948717947, "grad_norm": 0.05113065987825394, "learning_rate": 9.664418523660004e-05, "loss": 0.1986, "step": 37 }, { "epoch": 3.8974358974358974, "grad_norm": 0.04517243430018425, "learning_rate": 9.628619846344454e-05, "loss": 0.2123, "step": 38 }, { "epoch": 4.0, "grad_norm": 0.04306492581963539, "learning_rate": 9.591080534401371e-05, "loss": 0.1971, "step": 39 }, { "epoch": 4.102564102564102, "grad_norm": 0.030920082703232765, "learning_rate": 9.551814704830734e-05, "loss": 0.1962, "step": 40 }, { "epoch": 4.205128205128205, "grad_norm": 0.040115438401699066, "learning_rate": 9.51083712390519e-05, "loss": 0.2028, "step": 41 }, { "epoch": 4.3076923076923075, "grad_norm": 0.03652967885136604, "learning_rate": 9.468163201617062e-05, "loss": 0.202, "step": 42 }, { "epoch": 4.410256410256411, "grad_norm": 0.043810825794935226, "learning_rate": 9.423808985883289e-05, "loss": 0.2064, "step": 43 }, { "epoch": 4.512820512820513, "grad_norm": 0.04564822092652321, "learning_rate": 9.377791156510455e-05, "loss": 0.2057, "step": 44 }, { "epoch": 4.615384615384615, "grad_norm": 0.030029835179448128, "learning_rate": 9.330127018922194e-05, "loss": 0.1988, "step": 45 }, { "epoch": 4.717948717948718, "grad_norm": 0.03449944779276848, "learning_rate": 9.280834497651334e-05, "loss": 0.1961, "step": 46 }, { "epoch": 4.82051282051282, "grad_norm": 0.038930896669626236, "learning_rate": 9.229932129599205e-05, "loss": 0.2059, "step": 47 }, { "epoch": 4.923076923076923, "grad_norm": 0.050978414714336395, "learning_rate": 9.177439057064683e-05, "loss": 0.1936, "step": 48 }, { "epoch": 5.0256410256410255, "grad_norm": 0.03453517332673073, "learning_rate": 9.123375020545535e-05, "loss": 0.2019, "step": 49 }, { "epoch": 5.128205128205128, "grad_norm": 0.03941568359732628, "learning_rate": 9.067760351314838e-05, "loss": 0.1984, "step": 50 }, { "epoch": 5.230769230769231, "grad_norm": 0.049096424132585526, "learning_rate": 9.01061596377522e-05, "loss": 0.1989, "step": 51 }, { "epoch": 5.333333333333333, "grad_norm": 0.04188678413629532, "learning_rate": 8.951963347593797e-05, "loss": 0.1948, "step": 52 }, { "epoch": 5.435897435897436, "grad_norm": 0.038111768662929535, "learning_rate": 8.891824559620801e-05, "loss": 0.1974, "step": 53 }, { "epoch": 5.538461538461538, "grad_norm": 0.042120371013879776, "learning_rate": 8.83022221559489e-05, "loss": 0.2026, "step": 54 }, { "epoch": 5.641025641025641, "grad_norm": 0.04244457557797432, "learning_rate": 8.767179481638303e-05, "loss": 0.1998, "step": 55 }, { "epoch": 5.743589743589744, "grad_norm": 0.04076429829001427, "learning_rate": 8.702720065545024e-05, "loss": 0.2059, "step": 56 }, { "epoch": 5.846153846153846, "grad_norm": 0.05762708932161331, "learning_rate": 8.636868207865244e-05, "loss": 0.1871, "step": 57 }, { "epoch": 5.948717948717949, "grad_norm": 0.05226736143231392, "learning_rate": 8.569648672789497e-05, "loss": 0.1962, "step": 58 }, { "epoch": 6.051282051282051, "grad_norm": 0.050000473856925964, "learning_rate": 8.501086738835843e-05, "loss": 0.197, "step": 59 }, { "epoch": 6.153846153846154, "grad_norm": 0.04632322117686272, "learning_rate": 8.43120818934367e-05, "loss": 0.1973, "step": 60 }, { "epoch": 6.256410256410256, "grad_norm": 0.05035199597477913, "learning_rate": 8.360039302777612e-05, "loss": 0.198, "step": 61 }, { "epoch": 6.358974358974359, "grad_norm": 0.047098372131586075, "learning_rate": 8.28760684284532e-05, "loss": 0.1955, "step": 62 }, { "epoch": 6.461538461538462, "grad_norm": 0.06358848512172699, "learning_rate": 8.213938048432697e-05, "loss": 0.2009, "step": 63 }, { "epoch": 6.564102564102564, "grad_norm": 0.047331102192401886, "learning_rate": 8.139060623360493e-05, "loss": 0.2012, "step": 64 }, { "epoch": 6.666666666666667, "grad_norm": 0.05147461220622063, "learning_rate": 8.063002725966015e-05, "loss": 0.1969, "step": 65 }, { "epoch": 6.769230769230769, "grad_norm": 0.05776860937476158, "learning_rate": 7.985792958513931e-05, "loss": 0.1934, "step": 66 }, { "epoch": 6.871794871794872, "grad_norm": 0.062228187918663025, "learning_rate": 7.907460356440133e-05, "loss": 0.1949, "step": 67 }, { "epoch": 6.9743589743589745, "grad_norm": 0.0530543327331543, "learning_rate": 7.828034377432693e-05, "loss": 0.1921, "step": 68 }, { "epoch": 7.076923076923077, "grad_norm": 0.05476854741573334, "learning_rate": 7.74754489035403e-05, "loss": 0.2001, "step": 69 }, { "epoch": 7.17948717948718, "grad_norm": 0.05120917409658432, "learning_rate": 7.666022164008457e-05, "loss": 0.1977, "step": 70 }, { "epoch": 7.282051282051282, "grad_norm": 0.051713503897190094, "learning_rate": 7.583496855759316e-05, "loss": 0.2007, "step": 71 }, { "epoch": 7.384615384615385, "grad_norm": 0.05411737784743309, "learning_rate": 7.500000000000001e-05, "loss": 0.1955, "step": 72 }, { "epoch": 7.487179487179487, "grad_norm": 0.058174680918455124, "learning_rate": 7.415562996483192e-05, "loss": 0.1943, "step": 73 }, { "epoch": 7.589743589743589, "grad_norm": 0.05788983032107353, "learning_rate": 7.330217598512695e-05, "loss": 0.1902, "step": 74 }, { "epoch": 7.6923076923076925, "grad_norm": 0.06303142011165619, "learning_rate": 7.243995901002312e-05, "loss": 0.1897, "step": 75 }, { "epoch": 7.794871794871795, "grad_norm": 0.06139795109629631, "learning_rate": 7.156930328406268e-05, "loss": 0.1903, "step": 76 }, { "epoch": 7.897435897435898, "grad_norm": 0.06393580138683319, "learning_rate": 7.069053622525696e-05, "loss": 0.1834, "step": 77 }, { "epoch": 8.0, "grad_norm": 0.07243426889181137, "learning_rate": 6.980398830195785e-05, "loss": 0.1976, "step": 78 }, { "epoch": 8.102564102564102, "grad_norm": 0.06290671974420547, "learning_rate": 6.890999290858214e-05, "loss": 0.1896, "step": 79 }, { "epoch": 8.205128205128204, "grad_norm": 0.05926685035228729, "learning_rate": 6.800888624023553e-05, "loss": 0.1914, "step": 80 }, { "epoch": 8.307692307692308, "grad_norm": 0.06375055015087128, "learning_rate": 6.710100716628344e-05, "loss": 0.1923, "step": 81 }, { "epoch": 8.41025641025641, "grad_norm": 0.058604903519153595, "learning_rate": 6.618669710291606e-05, "loss": 0.1878, "step": 82 }, { "epoch": 8.512820512820513, "grad_norm": 0.06643830239772797, "learning_rate": 6.526629988475567e-05, "loss": 0.1924, "step": 83 }, { "epoch": 8.615384615384615, "grad_norm": 0.06960766762495041, "learning_rate": 6.434016163555452e-05, "loss": 0.193, "step": 84 }, { "epoch": 8.717948717948717, "grad_norm": 0.07951146364212036, "learning_rate": 6.340863063803188e-05, "loss": 0.1894, "step": 85 }, { "epoch": 8.820512820512821, "grad_norm": 0.08499961346387863, "learning_rate": 6.247205720289907e-05, "loss": 0.1946, "step": 86 }, { "epoch": 8.923076923076923, "grad_norm": 0.080729179084301, "learning_rate": 6.153079353712201e-05, "loss": 0.1897, "step": 87 }, { "epoch": 9.025641025641026, "grad_norm": 0.08111605793237686, "learning_rate": 6.058519361147055e-05, "loss": 0.1952, "step": 88 }, { "epoch": 9.128205128205128, "grad_norm": 0.08702465891838074, "learning_rate": 5.963561302740449e-05, "loss": 0.186, "step": 89 }, { "epoch": 9.23076923076923, "grad_norm": 0.08024930208921432, "learning_rate": 5.868240888334653e-05, "loss": 0.1899, "step": 90 }, { "epoch": 9.333333333333334, "grad_norm": 0.07547761499881744, "learning_rate": 5.772593964039203e-05, "loss": 0.1891, "step": 91 }, { "epoch": 9.435897435897436, "grad_norm": 0.0838337242603302, "learning_rate": 5.6766564987506566e-05, "loss": 0.1904, "step": 92 }, { "epoch": 9.538461538461538, "grad_norm": 0.08516307175159454, "learning_rate": 5.5804645706261514e-05, "loss": 0.1895, "step": 93 }, { "epoch": 9.64102564102564, "grad_norm": 0.08381783217191696, "learning_rate": 5.484054353515896e-05, "loss": 0.1878, "step": 94 }, { "epoch": 9.743589743589745, "grad_norm": 0.08215490728616714, "learning_rate": 5.387462103359655e-05, "loss": 0.189, "step": 95 }, { "epoch": 9.846153846153847, "grad_norm": 0.08765217661857605, "learning_rate": 5.290724144552379e-05, "loss": 0.1898, "step": 96 }, { "epoch": 9.948717948717949, "grad_norm": 0.08872053772211075, "learning_rate": 5.193876856284085e-05, "loss": 0.1908, "step": 97 }, { "epoch": 10.051282051282051, "grad_norm": 0.09559184312820435, "learning_rate": 5.096956658859122e-05, "loss": 0.1854, "step": 98 }, { "epoch": 10.153846153846153, "grad_norm": 0.08504116535186768, "learning_rate": 5e-05, "loss": 0.1873, "step": 99 }, { "epoch": 10.256410256410255, "grad_norm": 0.10478716343641281, "learning_rate": 4.903043341140879e-05, "loss": 0.1835, "step": 100 }, { "epoch": 10.35897435897436, "grad_norm": 0.09934550523757935, "learning_rate": 4.806123143715916e-05, "loss": 0.1845, "step": 101 }, { "epoch": 10.461538461538462, "grad_norm": 0.1022539958357811, "learning_rate": 4.709275855447621e-05, "loss": 0.189, "step": 102 }, { "epoch": 10.564102564102564, "grad_norm": 0.10897883027791977, "learning_rate": 4.612537896640346e-05, "loss": 0.1879, "step": 103 }, { "epoch": 10.666666666666666, "grad_norm": 0.11043286323547363, "learning_rate": 4.515945646484105e-05, "loss": 0.1902, "step": 104 }, { "epoch": 10.76923076923077, "grad_norm": 0.11388922482728958, "learning_rate": 4.4195354293738484e-05, "loss": 0.1816, "step": 105 }, { "epoch": 10.871794871794872, "grad_norm": 0.12030429393053055, "learning_rate": 4.323343501249346e-05, "loss": 0.1902, "step": 106 }, { "epoch": 10.974358974358974, "grad_norm": 0.1329633891582489, "learning_rate": 4.227406035960798e-05, "loss": 0.1886, "step": 107 }, { "epoch": 11.076923076923077, "grad_norm": 0.11949360370635986, "learning_rate": 4.131759111665349e-05, "loss": 0.1808, "step": 108 }, { "epoch": 11.179487179487179, "grad_norm": 0.11372341960668564, "learning_rate": 4.036438697259551e-05, "loss": 0.1787, "step": 109 }, { "epoch": 11.282051282051283, "grad_norm": 0.10975624620914459, "learning_rate": 3.941480638852948e-05, "loss": 0.1871, "step": 110 }, { "epoch": 11.384615384615385, "grad_norm": 0.11220697313547134, "learning_rate": 3.846920646287799e-05, "loss": 0.1833, "step": 111 }, { "epoch": 11.487179487179487, "grad_norm": 0.11458546668291092, "learning_rate": 3.752794279710094e-05, "loss": 0.1854, "step": 112 }, { "epoch": 11.58974358974359, "grad_norm": 0.11898455768823624, "learning_rate": 3.6591369361968124e-05, "loss": 0.1875, "step": 113 }, { "epoch": 11.692307692307692, "grad_norm": 0.11748585850000381, "learning_rate": 3.5659838364445505e-05, "loss": 0.1786, "step": 114 }, { "epoch": 11.794871794871796, "grad_norm": 0.14872179925441742, "learning_rate": 3.473370011524435e-05, "loss": 0.1805, "step": 115 }, { "epoch": 11.897435897435898, "grad_norm": 0.13790348172187805, "learning_rate": 3.381330289708396e-05, "loss": 0.1775, "step": 116 }, { "epoch": 12.0, "grad_norm": 0.1334102749824524, "learning_rate": 3.289899283371657e-05, "loss": 0.1868, "step": 117 }, { "epoch": 12.102564102564102, "grad_norm": 0.12504515051841736, "learning_rate": 3.199111375976449e-05, "loss": 0.1802, "step": 118 }, { "epoch": 12.205128205128204, "grad_norm": 0.13375680148601532, "learning_rate": 3.109000709141788e-05, "loss": 0.1775, "step": 119 }, { "epoch": 12.307692307692308, "grad_norm": 0.12553994357585907, "learning_rate": 3.019601169804216e-05, "loss": 0.1818, "step": 120 }, { "epoch": 12.41025641025641, "grad_norm": 0.14915205538272858, "learning_rate": 2.9309463774743046e-05, "loss": 0.1838, "step": 121 }, { "epoch": 12.512820512820513, "grad_norm": 0.13983815908432007, "learning_rate": 2.8430696715937337e-05, "loss": 0.1707, "step": 122 }, { "epoch": 12.615384615384615, "grad_norm": 0.13779574632644653, "learning_rate": 2.7560040989976892e-05, "loss": 0.171, "step": 123 }, { "epoch": 12.717948717948717, "grad_norm": 0.1488693505525589, "learning_rate": 2.6697824014873075e-05, "loss": 0.1811, "step": 124 }, { "epoch": 12.820512820512821, "grad_norm": 0.15547537803649902, "learning_rate": 2.5844370035168073e-05, "loss": 0.1826, "step": 125 }, { "epoch": 12.923076923076923, "grad_norm": 0.15429019927978516, "learning_rate": 2.500000000000001e-05, "loss": 0.1792, "step": 126 }, { "epoch": 13.025641025641026, "grad_norm": 0.17629271745681763, "learning_rate": 2.4165031442406855e-05, "loss": 0.1886, "step": 127 }, { "epoch": 13.128205128205128, "grad_norm": 0.1366981863975525, "learning_rate": 2.333977835991545e-05, "loss": 0.1829, "step": 128 }, { "epoch": 13.23076923076923, "grad_norm": 0.1525011658668518, "learning_rate": 2.25245510964597e-05, "loss": 0.1746, "step": 129 }, { "epoch": 13.333333333333334, "grad_norm": 0.14935700595378876, "learning_rate": 2.171965622567308e-05, "loss": 0.1781, "step": 130 }, { "epoch": 13.435897435897436, "grad_norm": 0.16920122504234314, "learning_rate": 2.0925396435598664e-05, "loss": 0.173, "step": 131 }, { "epoch": 13.538461538461538, "grad_norm": 0.15920865535736084, "learning_rate": 2.0142070414860704e-05, "loss": 0.1705, "step": 132 }, { "epoch": 13.64102564102564, "grad_norm": 0.16385790705680847, "learning_rate": 1.936997274033986e-05, "loss": 0.1761, "step": 133 }, { "epoch": 13.743589743589745, "grad_norm": 0.17888469994068146, "learning_rate": 1.8609393766395085e-05, "loss": 0.1795, "step": 134 }, { "epoch": 13.846153846153847, "grad_norm": 0.19205844402313232, "learning_rate": 1.7860619515673033e-05, "loss": 0.1759, "step": 135 }, { "epoch": 13.948717948717949, "grad_norm": 0.19030755758285522, "learning_rate": 1.7123931571546827e-05, "loss": 0.1804, "step": 136 }, { "epoch": 14.051282051282051, "grad_norm": 0.1694622039794922, "learning_rate": 1.639960697222388e-05, "loss": 0.1675, "step": 137 }, { "epoch": 14.153846153846153, "grad_norm": 0.1634732037782669, "learning_rate": 1.5687918106563326e-05, "loss": 0.1792, "step": 138 }, { "epoch": 14.256410256410255, "grad_norm": 0.1478685885667801, "learning_rate": 1.4989132611641576e-05, "loss": 0.1801, "step": 139 }, { "epoch": 14.35897435897436, "grad_norm": 0.15892252326011658, "learning_rate": 1.4303513272105057e-05, "loss": 0.1707, "step": 140 }, { "epoch": 14.461538461538462, "grad_norm": 0.15700320899486542, "learning_rate": 1.3631317921347563e-05, "loss": 0.1752, "step": 141 }, { "epoch": 14.564102564102564, "grad_norm": 0.16896262764930725, "learning_rate": 1.297279934454978e-05, "loss": 0.1679, "step": 142 }, { "epoch": 14.666666666666666, "grad_norm": 0.17261075973510742, "learning_rate": 1.2328205183616965e-05, "loss": 0.1737, "step": 143 }, { "epoch": 14.76923076923077, "grad_norm": 0.1785310059785843, "learning_rate": 1.1697777844051105e-05, "loss": 0.1806, "step": 144 }, { "epoch": 14.871794871794872, "grad_norm": 0.16977129876613617, "learning_rate": 1.1081754403791999e-05, "loss": 0.1696, "step": 145 }, { "epoch": 14.974358974358974, "grad_norm": 0.1806604415178299, "learning_rate": 1.0480366524062042e-05, "loss": 0.1647, "step": 146 }, { "epoch": 15.076923076923077, "grad_norm": 0.17654891312122345, "learning_rate": 9.893840362247809e-06, "loss": 0.1701, "step": 147 }, { "epoch": 15.179487179487179, "grad_norm": 0.16555620729923248, "learning_rate": 9.322396486851626e-06, "loss": 0.1699, "step": 148 }, { "epoch": 15.282051282051283, "grad_norm": 0.17180851101875305, "learning_rate": 8.766249794544662e-06, "loss": 0.1731, "step": 149 }, { "epoch": 15.384615384615385, "grad_norm": 0.16787266731262207, "learning_rate": 8.225609429353187e-06, "loss": 0.1672, "step": 150 }, { "epoch": 15.487179487179487, "grad_norm": 0.16866062581539154, "learning_rate": 7.700678704007947e-06, "loss": 0.1733, "step": 151 }, { "epoch": 15.58974358974359, "grad_norm": 0.1637592613697052, "learning_rate": 7.191655023486682e-06, "loss": 0.1733, "step": 152 }, { "epoch": 15.692307692307692, "grad_norm": 0.1912868618965149, "learning_rate": 6.698729810778065e-06, "loss": 0.172, "step": 153 }, { "epoch": 15.794871794871796, "grad_norm": 0.1946249157190323, "learning_rate": 6.222088434895462e-06, "loss": 0.1706, "step": 154 }, { "epoch": 15.897435897435898, "grad_norm": 0.1749986857175827, "learning_rate": 5.7619101411671095e-06, "loss": 0.1688, "step": 155 }, { "epoch": 16.0, "grad_norm": 0.19781996309757233, "learning_rate": 5.318367983829392e-06, "loss": 0.1772, "step": 156 }, { "epoch": 16.102564102564102, "grad_norm": 0.17886731028556824, "learning_rate": 4.891628760948114e-06, "loss": 0.1705, "step": 157 }, { "epoch": 16.205128205128204, "grad_norm": 0.1695544272661209, "learning_rate": 4.4818529516926726e-06, "loss": 0.1734, "step": 158 }, { "epoch": 16.307692307692307, "grad_norm": 0.16796259582042694, "learning_rate": 4.089194655986306e-06, "loss": 0.1671, "step": 159 }, { "epoch": 16.41025641025641, "grad_norm": 0.19824908673763275, "learning_rate": 3.7138015365554833e-06, "loss": 0.1745, "step": 160 }, { "epoch": 16.51282051282051, "grad_norm": 0.1771104782819748, "learning_rate": 3.3558147633999728e-06, "loss": 0.176, "step": 161 }, { "epoch": 16.615384615384617, "grad_norm": 0.16765469312667847, "learning_rate": 3.0153689607045845e-06, "loss": 0.1654, "step": 162 }, { "epoch": 16.71794871794872, "grad_norm": 0.17278793454170227, "learning_rate": 2.692592156212487e-06, "loss": 0.1685, "step": 163 }, { "epoch": 16.82051282051282, "grad_norm": 0.17281928658485413, "learning_rate": 2.3876057330792346e-06, "loss": 0.1678, "step": 164 }, { "epoch": 16.923076923076923, "grad_norm": 0.16665330529212952, "learning_rate": 2.100524384225555e-06, "loss": 0.1614, "step": 165 }, { "epoch": 17.025641025641026, "grad_norm": 0.18625028431415558, "learning_rate": 1.8314560692059835e-06, "loss": 0.1716, "step": 166 }, { "epoch": 17.128205128205128, "grad_norm": 0.1805369257926941, "learning_rate": 1.5805019736097104e-06, "loss": 0.1769, "step": 167 }, { "epoch": 17.23076923076923, "grad_norm": 0.16571341454982758, "learning_rate": 1.3477564710088098e-06, "loss": 0.1674, "step": 168 }, { "epoch": 17.333333333333332, "grad_norm": 0.16930599510669708, "learning_rate": 1.1333070874682216e-06, "loss": 0.1656, "step": 169 }, { "epoch": 17.435897435897434, "grad_norm": 0.17679651081562042, "learning_rate": 9.372344686307655e-07, "loss": 0.1734, "step": 170 }, { "epoch": 17.53846153846154, "grad_norm": 0.17293980717658997, "learning_rate": 7.596123493895991e-07, "loss": 0.1678, "step": 171 }, { "epoch": 17.641025641025642, "grad_norm": 0.17545515298843384, "learning_rate": 6.005075261595494e-07, "loss": 0.1662, "step": 172 }, { "epoch": 17.743589743589745, "grad_norm": 0.17302531003952026, "learning_rate": 4.5997983175773417e-07, "loss": 0.1669, "step": 173 }, { "epoch": 17.846153846153847, "grad_norm": 0.1898086816072464, "learning_rate": 3.380821129028489e-07, "loss": 0.1772, "step": 174 }, { "epoch": 17.94871794871795, "grad_norm": 0.1680038422346115, "learning_rate": 2.3486021034170857e-07, "loss": 0.1681, "step": 175 }, { "epoch": 18.05128205128205, "grad_norm": 0.1949233114719391, "learning_rate": 1.503529416103988e-07, "loss": 0.1601, "step": 176 }, { "epoch": 18.153846153846153, "grad_norm": 0.17588374018669128, "learning_rate": 8.459208643659122e-08, "loss": 0.1705, "step": 177 }, { "epoch": 18.256410256410255, "grad_norm": 0.18231520056724548, "learning_rate": 3.760237478849793e-08, "loss": 0.1637, "step": 178 }, { "epoch": 18.358974358974358, "grad_norm": 0.17962919175624847, "learning_rate": 9.401477574932926e-09, "loss": 0.1751, "step": 179 }, { "epoch": 18.46153846153846, "grad_norm": 0.17972521483898163, "learning_rate": 0.0, "loss": 0.1738, "step": 180 } ], "logging_steps": 1.0, "max_steps": 180, "num_input_tokens_seen": 0, "num_train_epochs": 20, "save_steps": 18, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.243916626699223e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }