{ "best_metric": null, "best_model_checkpoint": null, "epoch": 19.08256880733945, "eval_steps": 500, "global_step": 260, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.07339449541284404, "grad_norm": 0.094066321849823, "learning_rate": 3.846153846153847e-06, "loss": 0.2373, "step": 1 }, { "epoch": 0.14678899082568808, "grad_norm": 0.08788317441940308, "learning_rate": 7.692307692307694e-06, "loss": 0.2393, "step": 2 }, { "epoch": 0.22018348623853212, "grad_norm": 0.08641067892313004, "learning_rate": 1.153846153846154e-05, "loss": 0.232, "step": 3 }, { "epoch": 0.29357798165137616, "grad_norm": 0.0884113758802414, "learning_rate": 1.5384615384615387e-05, "loss": 0.2312, "step": 4 }, { "epoch": 0.3669724770642202, "grad_norm": 0.09627583622932434, "learning_rate": 1.923076923076923e-05, "loss": 0.2347, "step": 5 }, { "epoch": 0.44036697247706424, "grad_norm": 0.08840294182300568, "learning_rate": 2.307692307692308e-05, "loss": 0.2379, "step": 6 }, { "epoch": 0.5137614678899083, "grad_norm": 0.08855913579463959, "learning_rate": 2.6923076923076923e-05, "loss": 0.255, "step": 7 }, { "epoch": 0.5871559633027523, "grad_norm": 0.0873115062713623, "learning_rate": 3.0769230769230774e-05, "loss": 0.2322, "step": 8 }, { "epoch": 0.6605504587155964, "grad_norm": 0.0857061967253685, "learning_rate": 3.461538461538462e-05, "loss": 0.239, "step": 9 }, { "epoch": 0.7339449541284404, "grad_norm": 0.08705756813287735, "learning_rate": 3.846153846153846e-05, "loss": 0.2428, "step": 10 }, { "epoch": 0.8073394495412844, "grad_norm": 0.07806955277919769, "learning_rate": 4.230769230769231e-05, "loss": 0.2312, "step": 11 }, { "epoch": 0.8807339449541285, "grad_norm": 0.07232638448476791, "learning_rate": 4.615384615384616e-05, "loss": 0.2254, "step": 12 }, { "epoch": 0.9541284403669725, "grad_norm": 0.061769310384988785, "learning_rate": 5e-05, "loss": 0.2281, "step": 13 }, { "epoch": 1.0275229357798166, "grad_norm": 0.058718591928482056, "learning_rate": 5.384615384615385e-05, "loss": 0.2288, "step": 14 }, { "epoch": 1.1009174311926606, "grad_norm": 0.06115482375025749, "learning_rate": 5.769230769230769e-05, "loss": 0.2279, "step": 15 }, { "epoch": 1.1743119266055047, "grad_norm": 0.064550019800663, "learning_rate": 6.153846153846155e-05, "loss": 0.2235, "step": 16 }, { "epoch": 1.2477064220183487, "grad_norm": 0.062019750475883484, "learning_rate": 6.538461538461539e-05, "loss": 0.2196, "step": 17 }, { "epoch": 1.3211009174311927, "grad_norm": 0.07217342406511307, "learning_rate": 6.923076923076924e-05, "loss": 0.2195, "step": 18 }, { "epoch": 1.3944954128440368, "grad_norm": 0.058537840843200684, "learning_rate": 7.307692307692307e-05, "loss": 0.2209, "step": 19 }, { "epoch": 1.4678899082568808, "grad_norm": 0.04221680015325546, "learning_rate": 7.692307692307693e-05, "loss": 0.2225, "step": 20 }, { "epoch": 1.5412844036697249, "grad_norm": 0.055857084691524506, "learning_rate": 8.076923076923078e-05, "loss": 0.2165, "step": 21 }, { "epoch": 1.614678899082569, "grad_norm": 0.05376669764518738, "learning_rate": 8.461538461538461e-05, "loss": 0.2174, "step": 22 }, { "epoch": 1.688073394495413, "grad_norm": 0.06512868404388428, "learning_rate": 8.846153846153847e-05, "loss": 0.2023, "step": 23 }, { "epoch": 1.761467889908257, "grad_norm": 0.04625726118683815, "learning_rate": 9.230769230769232e-05, "loss": 0.2137, "step": 24 }, { "epoch": 1.834862385321101, "grad_norm": 0.052789829671382904, "learning_rate": 9.615384615384617e-05, "loss": 0.2084, "step": 25 }, { "epoch": 1.908256880733945, "grad_norm": 0.042677175253629684, "learning_rate": 0.0001, "loss": 0.2113, "step": 26 }, { "epoch": 1.981651376146789, "grad_norm": 0.04695236310362816, "learning_rate": 9.999549389283606e-05, "loss": 0.2142, "step": 27 }, { "epoch": 2.055045871559633, "grad_norm": 0.036158736795186996, "learning_rate": 9.998197638354428e-05, "loss": 0.2085, "step": 28 }, { "epoch": 2.128440366972477, "grad_norm": 0.037386395037174225, "learning_rate": 9.995944990857849e-05, "loss": 0.2149, "step": 29 }, { "epoch": 2.2018348623853212, "grad_norm": 0.04186835139989853, "learning_rate": 9.992791852820709e-05, "loss": 0.2107, "step": 30 }, { "epoch": 2.2752293577981653, "grad_norm": 0.05842822790145874, "learning_rate": 9.988738792578126e-05, "loss": 0.194, "step": 31 }, { "epoch": 2.3486238532110093, "grad_norm": 0.04445941001176834, "learning_rate": 9.983786540671051e-05, "loss": 0.2176, "step": 32 }, { "epoch": 2.4220183486238533, "grad_norm": 0.0501200370490551, "learning_rate": 9.977935989714595e-05, "loss": 0.2016, "step": 33 }, { "epoch": 2.4954128440366974, "grad_norm": 0.042810119688510895, "learning_rate": 9.97118819423714e-05, "loss": 0.2091, "step": 34 }, { "epoch": 2.5688073394495414, "grad_norm": 0.03298516198992729, "learning_rate": 9.96354437049027e-05, "loss": 0.2127, "step": 35 }, { "epoch": 2.6422018348623855, "grad_norm": 0.038216155022382736, "learning_rate": 9.955005896229543e-05, "loss": 0.2003, "step": 36 }, { "epoch": 2.7155963302752295, "grad_norm": 0.05605754256248474, "learning_rate": 9.945574310466159e-05, "loss": 0.1987, "step": 37 }, { "epoch": 2.7889908256880735, "grad_norm": 0.04308450222015381, "learning_rate": 9.935251313189564e-05, "loss": 0.2102, "step": 38 }, { "epoch": 2.8623853211009176, "grad_norm": 0.037231989204883575, "learning_rate": 9.924038765061042e-05, "loss": 0.2046, "step": 39 }, { "epoch": 2.9357798165137616, "grad_norm": 0.03728229179978371, "learning_rate": 9.911938687078324e-05, "loss": 0.2103, "step": 40 }, { "epoch": 3.0091743119266057, "grad_norm": 0.034406475722789764, "learning_rate": 9.898953260211338e-05, "loss": 0.2118, "step": 41 }, { "epoch": 3.0825688073394497, "grad_norm": 0.03245306760072708, "learning_rate": 9.885084825009086e-05, "loss": 0.1988, "step": 42 }, { "epoch": 3.1559633027522938, "grad_norm": 0.050477948039770126, "learning_rate": 9.870335881177774e-05, "loss": 0.2027, "step": 43 }, { "epoch": 3.229357798165138, "grad_norm": 0.037724193185567856, "learning_rate": 9.85470908713026e-05, "loss": 0.2019, "step": 44 }, { "epoch": 3.302752293577982, "grad_norm": 0.037339672446250916, "learning_rate": 9.838207259506891e-05, "loss": 0.2106, "step": 45 }, { "epoch": 3.376146788990826, "grad_norm": 0.037933845072984695, "learning_rate": 9.820833372667812e-05, "loss": 0.199, "step": 46 }, { "epoch": 3.44954128440367, "grad_norm": 0.03609978035092354, "learning_rate": 9.802590558156862e-05, "loss": 0.2059, "step": 47 }, { "epoch": 3.522935779816514, "grad_norm": 0.04476189985871315, "learning_rate": 9.783482104137127e-05, "loss": 0.2152, "step": 48 }, { "epoch": 3.5963302752293576, "grad_norm": 0.041969411075115204, "learning_rate": 9.763511454798268e-05, "loss": 0.2117, "step": 49 }, { "epoch": 3.669724770642202, "grad_norm": 0.04141020029783249, "learning_rate": 9.742682209735727e-05, "loss": 0.1912, "step": 50 }, { "epoch": 3.7431192660550456, "grad_norm": 0.03470165282487869, "learning_rate": 9.720998123301923e-05, "loss": 0.1968, "step": 51 }, { "epoch": 3.81651376146789, "grad_norm": 0.03589041531085968, "learning_rate": 9.698463103929542e-05, "loss": 0.2035, "step": 52 }, { "epoch": 3.8899082568807337, "grad_norm": 0.03698645904660225, "learning_rate": 9.675081213427076e-05, "loss": 0.1979, "step": 53 }, { "epoch": 3.963302752293578, "grad_norm": 0.03854459151625633, "learning_rate": 9.650856666246693e-05, "loss": 0.2, "step": 54 }, { "epoch": 4.036697247706422, "grad_norm": 0.037219513207674026, "learning_rate": 9.625793828724618e-05, "loss": 0.1981, "step": 55 }, { "epoch": 4.110091743119266, "grad_norm": 0.038294121623039246, "learning_rate": 9.599897218294122e-05, "loss": 0.1937, "step": 56 }, { "epoch": 4.18348623853211, "grad_norm": 0.04259132966399193, "learning_rate": 9.573171502671273e-05, "loss": 0.1922, "step": 57 }, { "epoch": 4.256880733944954, "grad_norm": 0.04465475678443909, "learning_rate": 9.54562149901362e-05, "loss": 0.1999, "step": 58 }, { "epoch": 4.330275229357798, "grad_norm": 0.041249386966228485, "learning_rate": 9.517252173051911e-05, "loss": 0.1921, "step": 59 }, { "epoch": 4.4036697247706424, "grad_norm": 0.04421278461813927, "learning_rate": 9.48806863819507e-05, "loss": 0.2039, "step": 60 }, { "epoch": 4.477064220183486, "grad_norm": 0.05113544315099716, "learning_rate": 9.458076154608515e-05, "loss": 0.186, "step": 61 }, { "epoch": 4.5504587155963305, "grad_norm": 0.04417051747441292, "learning_rate": 9.42728012826605e-05, "loss": 0.1985, "step": 62 }, { "epoch": 4.623853211009174, "grad_norm": 0.04551684111356735, "learning_rate": 9.395686109975474e-05, "loss": 0.1925, "step": 63 }, { "epoch": 4.697247706422019, "grad_norm": 0.05105452239513397, "learning_rate": 9.363299794378073e-05, "loss": 0.2009, "step": 64 }, { "epoch": 4.770642201834862, "grad_norm": 0.05177849158644676, "learning_rate": 9.330127018922194e-05, "loss": 0.1931, "step": 65 }, { "epoch": 4.844036697247707, "grad_norm": 0.0584421344101429, "learning_rate": 9.296173762811085e-05, "loss": 0.2095, "step": 66 }, { "epoch": 4.91743119266055, "grad_norm": 0.052275680005550385, "learning_rate": 9.261446145925167e-05, "loss": 0.2074, "step": 67 }, { "epoch": 4.990825688073395, "grad_norm": 0.052674662321805954, "learning_rate": 9.225950427718975e-05, "loss": 0.2029, "step": 68 }, { "epoch": 5.064220183486238, "grad_norm": 0.06084563210606575, "learning_rate": 9.189693006092907e-05, "loss": 0.1875, "step": 69 }, { "epoch": 5.137614678899083, "grad_norm": 0.0493173822760582, "learning_rate": 9.152680416240059e-05, "loss": 0.1983, "step": 70 }, { "epoch": 5.2110091743119265, "grad_norm": 0.050358761101961136, "learning_rate": 9.114919329468282e-05, "loss": 0.1989, "step": 71 }, { "epoch": 5.284403669724771, "grad_norm": 0.050705138593912125, "learning_rate": 9.076416551997721e-05, "loss": 0.1929, "step": 72 }, { "epoch": 5.3577981651376145, "grad_norm": 0.06104021519422531, "learning_rate": 9.037179023734035e-05, "loss": 0.1989, "step": 73 }, { "epoch": 5.431192660550459, "grad_norm": 0.05947330221533775, "learning_rate": 8.997213817017507e-05, "loss": 0.2026, "step": 74 }, { "epoch": 5.504587155963303, "grad_norm": 0.05319764092564583, "learning_rate": 8.95652813534831e-05, "loss": 0.2011, "step": 75 }, { "epoch": 5.577981651376147, "grad_norm": 0.05301457270979881, "learning_rate": 8.915129312088112e-05, "loss": 0.1983, "step": 76 }, { "epoch": 5.651376146788991, "grad_norm": 0.058974914252758026, "learning_rate": 8.873024809138272e-05, "loss": 0.1861, "step": 77 }, { "epoch": 5.724770642201835, "grad_norm": 0.0584375262260437, "learning_rate": 8.83022221559489e-05, "loss": 0.1924, "step": 78 }, { "epoch": 5.798165137614679, "grad_norm": 0.06012069061398506, "learning_rate": 8.786729246380901e-05, "loss": 0.1952, "step": 79 }, { "epoch": 5.871559633027523, "grad_norm": 0.06680011004209518, "learning_rate": 8.742553740855506e-05, "loss": 0.1929, "step": 80 }, { "epoch": 5.944954128440367, "grad_norm": 0.06496898084878922, "learning_rate": 8.697703661401186e-05, "loss": 0.1826, "step": 81 }, { "epoch": 6.018348623853211, "grad_norm": 0.06297076493501663, "learning_rate": 8.652187091988517e-05, "loss": 0.1846, "step": 82 }, { "epoch": 6.091743119266055, "grad_norm": 0.06112602353096008, "learning_rate": 8.606012236719073e-05, "loss": 0.1954, "step": 83 }, { "epoch": 6.165137614678899, "grad_norm": 0.06689347326755524, "learning_rate": 8.559187418346703e-05, "loss": 0.1937, "step": 84 }, { "epoch": 6.238532110091743, "grad_norm": 0.06930436193943024, "learning_rate": 8.511721076777389e-05, "loss": 0.1808, "step": 85 }, { "epoch": 6.3119266055045875, "grad_norm": 0.06421869993209839, "learning_rate": 8.463621767547998e-05, "loss": 0.1876, "step": 86 }, { "epoch": 6.385321100917431, "grad_norm": 0.07200642675161362, "learning_rate": 8.414898160284208e-05, "loss": 0.1946, "step": 87 }, { "epoch": 6.458715596330276, "grad_norm": 0.07047700881958008, "learning_rate": 8.365559037137852e-05, "loss": 0.199, "step": 88 }, { "epoch": 6.532110091743119, "grad_norm": 0.07283582538366318, "learning_rate": 8.315613291203976e-05, "loss": 0.1976, "step": 89 }, { "epoch": 6.605504587155964, "grad_norm": 0.07760219275951385, "learning_rate": 8.265069924917925e-05, "loss": 0.1807, "step": 90 }, { "epoch": 6.678899082568807, "grad_norm": 0.08349579572677612, "learning_rate": 8.213938048432697e-05, "loss": 0.1904, "step": 91 }, { "epoch": 6.752293577981652, "grad_norm": 0.08077001571655273, "learning_rate": 8.162226877976887e-05, "loss": 0.1927, "step": 92 }, { "epoch": 6.825688073394495, "grad_norm": 0.07862831652164459, "learning_rate": 8.10994573419352e-05, "loss": 0.189, "step": 93 }, { "epoch": 6.89908256880734, "grad_norm": 0.08872492611408234, "learning_rate": 8.057104040460062e-05, "loss": 0.1907, "step": 94 }, { "epoch": 6.972477064220183, "grad_norm": 0.07660839706659317, "learning_rate": 8.003711321189895e-05, "loss": 0.1878, "step": 95 }, { "epoch": 7.045871559633028, "grad_norm": 0.08058074116706848, "learning_rate": 7.949777200115616e-05, "loss": 0.1946, "step": 96 }, { "epoch": 7.1192660550458715, "grad_norm": 0.08907619118690491, "learning_rate": 7.895311398554395e-05, "loss": 0.2, "step": 97 }, { "epoch": 7.192660550458716, "grad_norm": 0.08919303119182587, "learning_rate": 7.840323733655778e-05, "loss": 0.1741, "step": 98 }, { "epoch": 7.26605504587156, "grad_norm": 0.0940893217921257, "learning_rate": 7.784824116632197e-05, "loss": 0.1919, "step": 99 }, { "epoch": 7.339449541284404, "grad_norm": 0.11027345806360245, "learning_rate": 7.728822550972523e-05, "loss": 0.1921, "step": 100 }, { "epoch": 7.412844036697248, "grad_norm": 0.11393953114748001, "learning_rate": 7.672329130639005e-05, "loss": 0.1892, "step": 101 }, { "epoch": 7.486238532110092, "grad_norm": 0.10183316469192505, "learning_rate": 7.615354038247888e-05, "loss": 0.1852, "step": 102 }, { "epoch": 7.559633027522936, "grad_norm": 0.1067890003323555, "learning_rate": 7.557907543234051e-05, "loss": 0.1891, "step": 103 }, { "epoch": 7.63302752293578, "grad_norm": 0.10177016258239746, "learning_rate": 7.500000000000001e-05, "loss": 0.1897, "step": 104 }, { "epoch": 7.706422018348624, "grad_norm": 0.13205887377262115, "learning_rate": 7.441641846049556e-05, "loss": 0.1751, "step": 105 }, { "epoch": 7.779816513761467, "grad_norm": 0.09557172656059265, "learning_rate": 7.38284360010654e-05, "loss": 0.1808, "step": 106 }, { "epoch": 7.853211009174312, "grad_norm": 0.10306868702173233, "learning_rate": 7.323615860218843e-05, "loss": 0.1828, "step": 107 }, { "epoch": 7.926605504587156, "grad_norm": 0.12413720041513443, "learning_rate": 7.263969301848188e-05, "loss": 0.1848, "step": 108 }, { "epoch": 8.0, "grad_norm": 0.09769123792648315, "learning_rate": 7.203914675945929e-05, "loss": 0.1872, "step": 109 }, { "epoch": 8.073394495412844, "grad_norm": 0.08778978884220123, "learning_rate": 7.143462807015271e-05, "loss": 0.1802, "step": 110 }, { "epoch": 8.146788990825687, "grad_norm": 0.09416276961565018, "learning_rate": 7.082624591160201e-05, "loss": 0.1859, "step": 111 }, { "epoch": 8.220183486238533, "grad_norm": 0.09707635641098022, "learning_rate": 7.021410994121525e-05, "loss": 0.1819, "step": 112 }, { "epoch": 8.293577981651376, "grad_norm": 0.09971417486667633, "learning_rate": 6.959833049300377e-05, "loss": 0.1783, "step": 113 }, { "epoch": 8.36697247706422, "grad_norm": 0.12254350632429123, "learning_rate": 6.897901855769483e-05, "loss": 0.1876, "step": 114 }, { "epoch": 8.440366972477065, "grad_norm": 0.1221860870718956, "learning_rate": 6.835628576272638e-05, "loss": 0.1904, "step": 115 }, { "epoch": 8.513761467889909, "grad_norm": 0.11945543438196182, "learning_rate": 6.773024435212678e-05, "loss": 0.1801, "step": 116 }, { "epoch": 8.587155963302752, "grad_norm": 0.14136166870594025, "learning_rate": 6.710100716628344e-05, "loss": 0.1781, "step": 117 }, { "epoch": 8.660550458715596, "grad_norm": 0.155049666762352, "learning_rate": 6.646868762160399e-05, "loss": 0.1799, "step": 118 }, { "epoch": 8.73394495412844, "grad_norm": 0.17453230917453766, "learning_rate": 6.583339969007363e-05, "loss": 0.1799, "step": 119 }, { "epoch": 8.807339449541285, "grad_norm": 0.14369648694992065, "learning_rate": 6.519525787871235e-05, "loss": 0.1879, "step": 120 }, { "epoch": 8.880733944954128, "grad_norm": 0.12149137258529663, "learning_rate": 6.455437720893564e-05, "loss": 0.1784, "step": 121 }, { "epoch": 8.954128440366972, "grad_norm": 0.13801580667495728, "learning_rate": 6.391087319582264e-05, "loss": 0.1838, "step": 122 }, { "epoch": 9.027522935779816, "grad_norm": 0.11812577396631241, "learning_rate": 6.326486182729504e-05, "loss": 0.1718, "step": 123 }, { "epoch": 9.100917431192661, "grad_norm": 0.12246193736791611, "learning_rate": 6.261645954321109e-05, "loss": 0.1744, "step": 124 }, { "epoch": 9.174311926605505, "grad_norm": 0.11642672121524811, "learning_rate": 6.19657832143779e-05, "loss": 0.1729, "step": 125 }, { "epoch": 9.247706422018348, "grad_norm": 0.11366164684295654, "learning_rate": 6.131295012148612e-05, "loss": 0.1749, "step": 126 }, { "epoch": 9.321100917431192, "grad_norm": 0.12722568213939667, "learning_rate": 6.06580779339708e-05, "loss": 0.1806, "step": 127 }, { "epoch": 9.394495412844037, "grad_norm": 0.13534581661224365, "learning_rate": 6.0001284688802226e-05, "loss": 0.1782, "step": 128 }, { "epoch": 9.46788990825688, "grad_norm": 0.13548819720745087, "learning_rate": 5.9342688769210254e-05, "loss": 0.1673, "step": 129 }, { "epoch": 9.541284403669724, "grad_norm": 0.15632225573062897, "learning_rate": 5.868240888334653e-05, "loss": 0.1834, "step": 130 }, { "epoch": 9.614678899082568, "grad_norm": 0.16248205304145813, "learning_rate": 5.8020564042888015e-05, "loss": 0.1821, "step": 131 }, { "epoch": 9.688073394495413, "grad_norm": 0.16676151752471924, "learning_rate": 5.735727354158581e-05, "loss": 0.1751, "step": 132 }, { "epoch": 9.761467889908257, "grad_norm": 0.17233224213123322, "learning_rate": 5.669265693376309e-05, "loss": 0.1708, "step": 133 }, { "epoch": 9.8348623853211, "grad_norm": 0.15577177703380585, "learning_rate": 5.602683401276615e-05, "loss": 0.1717, "step": 134 }, { "epoch": 9.908256880733944, "grad_norm": 0.17410804331302643, "learning_rate": 5.5359924789372396e-05, "loss": 0.1842, "step": 135 }, { "epoch": 9.98165137614679, "grad_norm": 0.15566076338291168, "learning_rate": 5.469204947015897e-05, "loss": 0.1737, "step": 136 }, { "epoch": 10.055045871559633, "grad_norm": 0.14896251261234283, "learning_rate": 5.402332843583631e-05, "loss": 0.1765, "step": 137 }, { "epoch": 10.128440366972477, "grad_norm": 0.1541670709848404, "learning_rate": 5.335388221955012e-05, "loss": 0.1658, "step": 138 }, { "epoch": 10.20183486238532, "grad_norm": 0.15539778769016266, "learning_rate": 5.2683831485156074e-05, "loss": 0.1656, "step": 139 }, { "epoch": 10.275229357798166, "grad_norm": 0.16468144953250885, "learning_rate": 5.201329700547076e-05, "loss": 0.1783, "step": 140 }, { "epoch": 10.34862385321101, "grad_norm": 0.16573026776313782, "learning_rate": 5.134239964050307e-05, "loss": 0.1603, "step": 141 }, { "epoch": 10.422018348623853, "grad_norm": 0.18577492237091064, "learning_rate": 5.0671260315669875e-05, "loss": 0.1744, "step": 142 }, { "epoch": 10.495412844036696, "grad_norm": 0.1860455870628357, "learning_rate": 5e-05, "loss": 0.1551, "step": 143 }, { "epoch": 10.568807339449542, "grad_norm": 0.18118642270565033, "learning_rate": 4.9328739684330137e-05, "loss": 0.1653, "step": 144 }, { "epoch": 10.642201834862385, "grad_norm": 0.2069067806005478, "learning_rate": 4.865760035949695e-05, "loss": 0.1737, "step": 145 }, { "epoch": 10.715596330275229, "grad_norm": 0.22928771376609802, "learning_rate": 4.798670299452926e-05, "loss": 0.1743, "step": 146 }, { "epoch": 10.788990825688073, "grad_norm": 0.19360584020614624, "learning_rate": 4.731616851484392e-05, "loss": 0.1691, "step": 147 }, { "epoch": 10.862385321100918, "grad_norm": 0.19605328142642975, "learning_rate": 4.6646117780449876e-05, "loss": 0.1752, "step": 148 }, { "epoch": 10.935779816513762, "grad_norm": 0.19604168832302094, "learning_rate": 4.597667156416371e-05, "loss": 0.1801, "step": 149 }, { "epoch": 11.009174311926605, "grad_norm": 0.19816367328166962, "learning_rate": 4.530795052984104e-05, "loss": 0.172, "step": 150 }, { "epoch": 11.082568807339449, "grad_norm": 0.18127745389938354, "learning_rate": 4.4640075210627615e-05, "loss": 0.1683, "step": 151 }, { "epoch": 11.155963302752294, "grad_norm": 0.19333089888095856, "learning_rate": 4.397316598723385e-05, "loss": 0.1593, "step": 152 }, { "epoch": 11.229357798165138, "grad_norm": 0.18506307899951935, "learning_rate": 4.3307343066236935e-05, "loss": 0.155, "step": 153 }, { "epoch": 11.302752293577981, "grad_norm": 0.17834623157978058, "learning_rate": 4.264272645841419e-05, "loss": 0.1619, "step": 154 }, { "epoch": 11.376146788990825, "grad_norm": 0.18679705262184143, "learning_rate": 4.197943595711198e-05, "loss": 0.1565, "step": 155 }, { "epoch": 11.44954128440367, "grad_norm": 0.2451370358467102, "learning_rate": 4.131759111665349e-05, "loss": 0.1666, "step": 156 }, { "epoch": 11.522935779816514, "grad_norm": 0.2141800969839096, "learning_rate": 4.0657311230789764e-05, "loss": 0.1557, "step": 157 }, { "epoch": 11.596330275229358, "grad_norm": 0.24323846399784088, "learning_rate": 3.9998715311197785e-05, "loss": 0.167, "step": 158 }, { "epoch": 11.669724770642201, "grad_norm": 0.2383449822664261, "learning_rate": 3.934192206602921e-05, "loss": 0.1684, "step": 159 }, { "epoch": 11.743119266055047, "grad_norm": 0.22943173348903656, "learning_rate": 3.86870498785139e-05, "loss": 0.158, "step": 160 }, { "epoch": 11.81651376146789, "grad_norm": 0.22590626776218414, "learning_rate": 3.803421678562213e-05, "loss": 0.1653, "step": 161 }, { "epoch": 11.889908256880734, "grad_norm": 0.2344842404127121, "learning_rate": 3.738354045678891e-05, "loss": 0.1611, "step": 162 }, { "epoch": 11.963302752293577, "grad_norm": 0.24516969919204712, "learning_rate": 3.673513817270497e-05, "loss": 0.1691, "step": 163 }, { "epoch": 12.036697247706423, "grad_norm": 0.23740142583847046, "learning_rate": 3.608912680417737e-05, "loss": 0.1614, "step": 164 }, { "epoch": 12.110091743119266, "grad_norm": 0.19846218824386597, "learning_rate": 3.5445622791064356e-05, "loss": 0.1507, "step": 165 }, { "epoch": 12.18348623853211, "grad_norm": 0.2150259017944336, "learning_rate": 3.480474212128766e-05, "loss": 0.1587, "step": 166 }, { "epoch": 12.256880733944953, "grad_norm": 0.2322087585926056, "learning_rate": 3.4166600309926387e-05, "loss": 0.1678, "step": 167 }, { "epoch": 12.330275229357799, "grad_norm": 0.24851596355438232, "learning_rate": 3.3531312378396026e-05, "loss": 0.1563, "step": 168 }, { "epoch": 12.403669724770642, "grad_norm": 0.2380513846874237, "learning_rate": 3.289899283371657e-05, "loss": 0.159, "step": 169 }, { "epoch": 12.477064220183486, "grad_norm": 0.24537961184978485, "learning_rate": 3.226975564787322e-05, "loss": 0.1635, "step": 170 }, { "epoch": 12.55045871559633, "grad_norm": 0.2758086919784546, "learning_rate": 3.164371423727362e-05, "loss": 0.1585, "step": 171 }, { "epoch": 12.623853211009175, "grad_norm": 0.2760733962059021, "learning_rate": 3.1020981442305184e-05, "loss": 0.1582, "step": 172 }, { "epoch": 12.697247706422019, "grad_norm": 0.24997705221176147, "learning_rate": 3.0401669506996256e-05, "loss": 0.1586, "step": 173 }, { "epoch": 12.770642201834862, "grad_norm": 0.23340976238250732, "learning_rate": 2.978589005878476e-05, "loss": 0.1475, "step": 174 }, { "epoch": 12.844036697247706, "grad_norm": 0.24687667191028595, "learning_rate": 2.917375408839803e-05, "loss": 0.1545, "step": 175 }, { "epoch": 12.917431192660551, "grad_norm": 0.2573287785053253, "learning_rate": 2.8565371929847284e-05, "loss": 0.1594, "step": 176 }, { "epoch": 12.990825688073395, "grad_norm": 0.23869659006595612, "learning_rate": 2.79608532405407e-05, "loss": 0.1478, "step": 177 }, { "epoch": 13.064220183486238, "grad_norm": 0.21892689168453217, "learning_rate": 2.7360306981518146e-05, "loss": 0.1537, "step": 178 }, { "epoch": 13.137614678899082, "grad_norm": 0.25160327553749084, "learning_rate": 2.6763841397811573e-05, "loss": 0.1461, "step": 179 }, { "epoch": 13.211009174311927, "grad_norm": 0.22473768889904022, "learning_rate": 2.6171563998934605e-05, "loss": 0.157, "step": 180 }, { "epoch": 13.284403669724771, "grad_norm": 0.271699994802475, "learning_rate": 2.5583581539504464e-05, "loss": 0.1481, "step": 181 }, { "epoch": 13.357798165137615, "grad_norm": 0.2539384961128235, "learning_rate": 2.500000000000001e-05, "loss": 0.1425, "step": 182 }, { "epoch": 13.431192660550458, "grad_norm": 0.25691837072372437, "learning_rate": 2.442092456765951e-05, "loss": 0.1496, "step": 183 }, { "epoch": 13.504587155963304, "grad_norm": 0.27063509821891785, "learning_rate": 2.3846459617521128e-05, "loss": 0.15, "step": 184 }, { "epoch": 13.577981651376147, "grad_norm": 0.2871812582015991, "learning_rate": 2.3276708693609943e-05, "loss": 0.1449, "step": 185 }, { "epoch": 13.65137614678899, "grad_norm": 0.28144845366477966, "learning_rate": 2.2711774490274766e-05, "loss": 0.1488, "step": 186 }, { "epoch": 13.724770642201834, "grad_norm": 0.25536370277404785, "learning_rate": 2.2151758833678045e-05, "loss": 0.1499, "step": 187 }, { "epoch": 13.79816513761468, "grad_norm": 0.276015043258667, "learning_rate": 2.1596762663442218e-05, "loss": 0.16, "step": 188 }, { "epoch": 13.871559633027523, "grad_norm": 0.2981753945350647, "learning_rate": 2.104688601445606e-05, "loss": 0.1619, "step": 189 }, { "epoch": 13.944954128440367, "grad_norm": 0.28739604353904724, "learning_rate": 2.050222799884387e-05, "loss": 0.143, "step": 190 }, { "epoch": 14.01834862385321, "grad_norm": 0.2708568871021271, "learning_rate": 1.996288678810105e-05, "loss": 0.1489, "step": 191 }, { "epoch": 14.091743119266056, "grad_norm": 0.2790358066558838, "learning_rate": 1.942895959539939e-05, "loss": 0.157, "step": 192 }, { "epoch": 14.1651376146789, "grad_norm": 0.2205585092306137, "learning_rate": 1.8900542658064807e-05, "loss": 0.1438, "step": 193 }, { "epoch": 14.238532110091743, "grad_norm": 0.23872052133083344, "learning_rate": 1.837773122023114e-05, "loss": 0.1447, "step": 194 }, { "epoch": 14.311926605504587, "grad_norm": 0.2528257966041565, "learning_rate": 1.7860619515673033e-05, "loss": 0.1404, "step": 195 }, { "epoch": 14.385321100917432, "grad_norm": 0.2546660900115967, "learning_rate": 1.734930075082076e-05, "loss": 0.1505, "step": 196 }, { "epoch": 14.458715596330276, "grad_norm": 0.26501449942588806, "learning_rate": 1.684386708796025e-05, "loss": 0.1382, "step": 197 }, { "epoch": 14.53211009174312, "grad_norm": 0.2743387520313263, "learning_rate": 1.6344409628621484e-05, "loss": 0.1436, "step": 198 }, { "epoch": 14.605504587155963, "grad_norm": 0.26094284653663635, "learning_rate": 1.585101839715792e-05, "loss": 0.1425, "step": 199 }, { "epoch": 14.678899082568808, "grad_norm": 0.2686833441257477, "learning_rate": 1.536378232452003e-05, "loss": 0.1393, "step": 200 }, { "epoch": 14.752293577981652, "grad_norm": 0.29060205817222595, "learning_rate": 1.4882789232226125e-05, "loss": 0.1467, "step": 201 }, { "epoch": 14.825688073394495, "grad_norm": 0.2699175477027893, "learning_rate": 1.440812581653298e-05, "loss": 0.1452, "step": 202 }, { "epoch": 14.899082568807339, "grad_norm": 0.25926584005355835, "learning_rate": 1.3939877632809278e-05, "loss": 0.1425, "step": 203 }, { "epoch": 14.972477064220184, "grad_norm": 0.27713659405708313, "learning_rate": 1.3478129080114848e-05, "loss": 0.1475, "step": 204 }, { "epoch": 15.045871559633028, "grad_norm": 0.25950995087623596, "learning_rate": 1.3022963385988151e-05, "loss": 0.141, "step": 205 }, { "epoch": 15.119266055045872, "grad_norm": 0.2496069371700287, "learning_rate": 1.257446259144494e-05, "loss": 0.1344, "step": 206 }, { "epoch": 15.192660550458715, "grad_norm": 0.24646779894828796, "learning_rate": 1.2132707536191008e-05, "loss": 0.1505, "step": 207 }, { "epoch": 15.26605504587156, "grad_norm": 0.23226536810398102, "learning_rate": 1.1697777844051105e-05, "loss": 0.1347, "step": 208 }, { "epoch": 15.339449541284404, "grad_norm": 0.25484660267829895, "learning_rate": 1.1269751908617277e-05, "loss": 0.1399, "step": 209 }, { "epoch": 15.412844036697248, "grad_norm": 0.25519728660583496, "learning_rate": 1.0848706879118892e-05, "loss": 0.141, "step": 210 }, { "epoch": 15.486238532110091, "grad_norm": 0.2521832585334778, "learning_rate": 1.0434718646516917e-05, "loss": 0.1387, "step": 211 }, { "epoch": 15.559633027522935, "grad_norm": 0.2603716552257538, "learning_rate": 1.0027861829824952e-05, "loss": 0.1401, "step": 212 }, { "epoch": 15.63302752293578, "grad_norm": 0.2601398527622223, "learning_rate": 9.628209762659657e-06, "loss": 0.1288, "step": 213 }, { "epoch": 15.706422018348624, "grad_norm": 0.2938661575317383, "learning_rate": 9.235834480022787e-06, "loss": 0.1454, "step": 214 }, { "epoch": 15.779816513761467, "grad_norm": 0.2794564962387085, "learning_rate": 8.850806705317183e-06, "loss": 0.1555, "step": 215 }, { "epoch": 15.853211009174313, "grad_norm": 0.2902175486087799, "learning_rate": 8.473195837599418e-06, "loss": 0.1436, "step": 216 }, { "epoch": 15.926605504587156, "grad_norm": 0.3273734748363495, "learning_rate": 8.103069939070945e-06, "loss": 0.1355, "step": 217 }, { "epoch": 16.0, "grad_norm": 0.29105839133262634, "learning_rate": 7.740495722810271e-06, "loss": 0.1455, "step": 218 }, { "epoch": 16.073394495412845, "grad_norm": 0.3094259798526764, "learning_rate": 7.385538540748327e-06, "loss": 0.1421, "step": 219 }, { "epoch": 16.146788990825687, "grad_norm": 0.2693122625350952, "learning_rate": 7.038262371889159e-06, "loss": 0.1426, "step": 220 }, { "epoch": 16.220183486238533, "grad_norm": 0.24507200717926025, "learning_rate": 6.698729810778065e-06, "loss": 0.1419, "step": 221 }, { "epoch": 16.293577981651374, "grad_norm": 0.256996750831604, "learning_rate": 6.367002056219284e-06, "loss": 0.1423, "step": 222 }, { "epoch": 16.36697247706422, "grad_norm": 0.23636184632778168, "learning_rate": 6.043138900245277e-06, "loss": 0.134, "step": 223 }, { "epoch": 16.440366972477065, "grad_norm": 0.2329455018043518, "learning_rate": 5.727198717339511e-06, "loss": 0.1313, "step": 224 }, { "epoch": 16.513761467889907, "grad_norm": 0.26152482628822327, "learning_rate": 5.41923845391486e-06, "loss": 0.1371, "step": 225 }, { "epoch": 16.587155963302752, "grad_norm": 0.24304719269275665, "learning_rate": 5.1193136180493095e-06, "loss": 0.1269, "step": 226 }, { "epoch": 16.660550458715598, "grad_norm": 0.2667653262615204, "learning_rate": 4.827478269480895e-06, "loss": 0.1362, "step": 227 }, { "epoch": 16.73394495412844, "grad_norm": 0.2727386951446533, "learning_rate": 4.54378500986381e-06, "loss": 0.1328, "step": 228 }, { "epoch": 16.807339449541285, "grad_norm": 0.28039076924324036, "learning_rate": 4.268284973287273e-06, "loss": 0.14, "step": 229 }, { "epoch": 16.88073394495413, "grad_norm": 0.2691219747066498, "learning_rate": 4.001027817058789e-06, "loss": 0.1424, "step": 230 }, { "epoch": 16.954128440366972, "grad_norm": 0.2565619647502899, "learning_rate": 3.7420617127538248e-06, "loss": 0.1387, "step": 231 }, { "epoch": 17.027522935779817, "grad_norm": 0.2617204189300537, "learning_rate": 3.4914333375330898e-06, "loss": 0.1355, "step": 232 }, { "epoch": 17.10091743119266, "grad_norm": 0.24316667020320892, "learning_rate": 3.249187865729264e-06, "loss": 0.1365, "step": 233 }, { "epoch": 17.174311926605505, "grad_norm": 0.26508602499961853, "learning_rate": 3.0153689607045845e-06, "loss": 0.1346, "step": 234 }, { "epoch": 17.24770642201835, "grad_norm": 0.27173709869384766, "learning_rate": 2.790018766980773e-06, "loss": 0.1359, "step": 235 }, { "epoch": 17.321100917431192, "grad_norm": 0.243478924036026, "learning_rate": 2.573177902642726e-06, "loss": 0.1431, "step": 236 }, { "epoch": 17.394495412844037, "grad_norm": 0.2515186369419098, "learning_rate": 2.3648854520173237e-06, "loss": 0.142, "step": 237 }, { "epoch": 17.46788990825688, "grad_norm": 0.2602452039718628, "learning_rate": 2.1651789586287442e-06, "loss": 0.1402, "step": 238 }, { "epoch": 17.541284403669724, "grad_norm": 0.25888198614120483, "learning_rate": 1.974094418431388e-06, "loss": 0.1312, "step": 239 }, { "epoch": 17.61467889908257, "grad_norm": 0.263540118932724, "learning_rate": 1.7916662733218847e-06, "loss": 0.1424, "step": 240 }, { "epoch": 17.68807339449541, "grad_norm": 0.24854296445846558, "learning_rate": 1.6179274049310966e-06, "loss": 0.1341, "step": 241 }, { "epoch": 17.761467889908257, "grad_norm": 0.24493998289108276, "learning_rate": 1.4529091286973995e-06, "loss": 0.1282, "step": 242 }, { "epoch": 17.834862385321102, "grad_norm": 0.25934794545173645, "learning_rate": 1.2966411882222696e-06, "loss": 0.1438, "step": 243 }, { "epoch": 17.908256880733944, "grad_norm": 0.2634919285774231, "learning_rate": 1.1491517499091498e-06, "loss": 0.1357, "step": 244 }, { "epoch": 17.98165137614679, "grad_norm": 0.25140729546546936, "learning_rate": 1.0104673978866164e-06, "loss": 0.1269, "step": 245 }, { "epoch": 18.05504587155963, "grad_norm": 0.26600712537765503, "learning_rate": 8.806131292167618e-07, "loss": 0.1343, "step": 246 }, { "epoch": 18.128440366972477, "grad_norm": 0.2575834095478058, "learning_rate": 7.596123493895991e-07, "loss": 0.1394, "step": 247 }, { "epoch": 18.201834862385322, "grad_norm": 0.2605387270450592, "learning_rate": 6.474868681043578e-07, "loss": 0.1442, "step": 248 }, { "epoch": 18.275229357798164, "grad_norm": 0.26221585273742676, "learning_rate": 5.442568953384186e-07, "loss": 0.1407, "step": 249 }, { "epoch": 18.34862385321101, "grad_norm": 0.24352163076400757, "learning_rate": 4.4994103770457653e-07, "loss": 0.1383, "step": 250 }, { "epoch": 18.422018348623855, "grad_norm": 0.23631100356578827, "learning_rate": 3.6455629509730136e-07, "loss": 0.1282, "step": 251 }, { "epoch": 18.495412844036696, "grad_norm": 0.24839530885219574, "learning_rate": 2.8811805762860576e-07, "loss": 0.1297, "step": 252 }, { "epoch": 18.568807339449542, "grad_norm": 0.25721636414527893, "learning_rate": 2.206401028540639e-07, "loss": 0.1469, "step": 253 }, { "epoch": 18.642201834862384, "grad_norm": 0.24413825571537018, "learning_rate": 1.6213459328950352e-07, "loss": 0.126, "step": 254 }, { "epoch": 18.71559633027523, "grad_norm": 0.24766796827316284, "learning_rate": 1.1261207421874309e-07, "loss": 0.1397, "step": 255 }, { "epoch": 18.788990825688074, "grad_norm": 0.24844591319561005, "learning_rate": 7.208147179291192e-08, "loss": 0.1355, "step": 256 }, { "epoch": 18.862385321100916, "grad_norm": 0.2460290938615799, "learning_rate": 4.055009142152067e-08, "loss": 0.13, "step": 257 }, { "epoch": 18.93577981651376, "grad_norm": 0.24734196066856384, "learning_rate": 1.802361645573125e-08, "loss": 0.1318, "step": 258 }, { "epoch": 19.009174311926607, "grad_norm": 0.27149155735969543, "learning_rate": 4.506107163948503e-09, "loss": 0.1356, "step": 259 }, { "epoch": 19.08256880733945, "grad_norm": 0.24715793132781982, "learning_rate": 0.0, "loss": 0.1384, "step": 260 } ], "logging_steps": 1.0, "max_steps": 260, "num_input_tokens_seen": 0, "num_train_epochs": 20, "save_steps": 130, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.241583344933929e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }