{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 100, "global_step": 11082, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0027070925825663237, "grad_norm": 4.53125, "learning_rate": 5e-05, "loss": 1.0758, "step": 10 }, { "epoch": 0.005414185165132647, "grad_norm": 0.609375, "learning_rate": 0.0001, "loss": 0.8551, "step": 20 }, { "epoch": 0.008121277747698972, "grad_norm": 1.2578125, "learning_rate": 0.00015, "loss": 0.7724, "step": 30 }, { "epoch": 0.010828370330265295, "grad_norm": 0.267578125, "learning_rate": 0.0002, "loss": 0.7212, "step": 40 }, { "epoch": 0.01353546291283162, "grad_norm": 0.77734375, "learning_rate": 0.00025, "loss": 0.6886, "step": 50 }, { "epoch": 0.016242555495397944, "grad_norm": 0.28125, "learning_rate": 0.0003, "loss": 0.6709, "step": 60 }, { "epoch": 0.018949648077964266, "grad_norm": 0.357421875, "learning_rate": 0.00035, "loss": 0.6294, "step": 70 }, { "epoch": 0.02165674066053059, "grad_norm": 0.326171875, "learning_rate": 0.0004, "loss": 0.5716, "step": 80 }, { "epoch": 0.024363833243096916, "grad_norm": 0.369140625, "learning_rate": 0.00045000000000000004, "loss": 0.4884, "step": 90 }, { "epoch": 0.02707092582566324, "grad_norm": 0.55078125, "learning_rate": 0.0005, "loss": 0.4087, "step": 100 }, { "epoch": 0.02707092582566324, "eval_loss": 0.34339678287506104, "eval_runtime": 44.1506, "eval_samples_per_second": 11.325, "eval_steps_per_second": 0.725, "step": 100 }, { "epoch": 0.02977801840822956, "grad_norm": 0.59765625, "learning_rate": 0.0004995447095246767, "loss": 0.3683, "step": 110 }, { "epoch": 0.03248511099079589, "grad_norm": 0.53515625, "learning_rate": 0.0004990894190493535, "loss": 0.2959, "step": 120 }, { "epoch": 0.03519220357336221, "grad_norm": 0.40625, "learning_rate": 0.0004986341285740302, "loss": 0.3314, "step": 130 }, { "epoch": 0.03789929615592853, "grad_norm": 1.1015625, "learning_rate": 0.000498178838098707, "loss": 0.2978, "step": 140 }, { "epoch": 0.040606388738494856, "grad_norm": 0.296875, "learning_rate": 0.0004977235476233838, "loss": 0.271, "step": 150 }, { "epoch": 0.04331348132106118, "grad_norm": 0.63671875, "learning_rate": 0.0004972682571480605, "loss": 0.2599, "step": 160 }, { "epoch": 0.0460205739036275, "grad_norm": 0.490234375, "learning_rate": 0.0004968129666727372, "loss": 0.2555, "step": 170 }, { "epoch": 0.04872766648619383, "grad_norm": 0.4296875, "learning_rate": 0.0004963576761974139, "loss": 0.2588, "step": 180 }, { "epoch": 0.051434759068760154, "grad_norm": 0.244140625, "learning_rate": 0.0004959023857220907, "loss": 0.278, "step": 190 }, { "epoch": 0.05414185165132648, "grad_norm": 0.4296875, "learning_rate": 0.0004954470952467675, "loss": 0.2379, "step": 200 }, { "epoch": 0.05414185165132648, "eval_loss": 0.20459319651126862, "eval_runtime": 40.1637, "eval_samples_per_second": 12.449, "eval_steps_per_second": 0.797, "step": 200 }, { "epoch": 0.0568489442338928, "grad_norm": 0.3515625, "learning_rate": 0.0004949918047714442, "loss": 0.2851, "step": 210 }, { "epoch": 0.05955603681645912, "grad_norm": 0.45703125, "learning_rate": 0.0004945365142961209, "loss": 0.2532, "step": 220 }, { "epoch": 0.062263129399025445, "grad_norm": 0.248046875, "learning_rate": 0.0004940812238207977, "loss": 0.2604, "step": 230 }, { "epoch": 0.06497022198159177, "grad_norm": 0.53125, "learning_rate": 0.0004936259333454744, "loss": 0.2566, "step": 240 }, { "epoch": 0.0676773145641581, "grad_norm": 0.42578125, "learning_rate": 0.0004931706428701511, "loss": 0.1982, "step": 250 }, { "epoch": 0.07038440714672442, "grad_norm": 0.5859375, "learning_rate": 0.0004927153523948279, "loss": 0.2171, "step": 260 }, { "epoch": 0.07309149972929074, "grad_norm": 0.53515625, "learning_rate": 0.0004922600619195047, "loss": 0.2201, "step": 270 }, { "epoch": 0.07579859231185707, "grad_norm": 0.5, "learning_rate": 0.0004918047714441814, "loss": 0.206, "step": 280 }, { "epoch": 0.07850568489442339, "grad_norm": 0.208984375, "learning_rate": 0.0004913494809688581, "loss": 0.2216, "step": 290 }, { "epoch": 0.08121277747698971, "grad_norm": 0.44140625, "learning_rate": 0.0004908941904935349, "loss": 0.204, "step": 300 }, { "epoch": 0.08121277747698971, "eval_loss": 0.1671503186225891, "eval_runtime": 40.1733, "eval_samples_per_second": 12.446, "eval_steps_per_second": 0.797, "step": 300 }, { "epoch": 0.08391987005955603, "grad_norm": 0.3359375, "learning_rate": 0.0004904389000182116, "loss": 0.2108, "step": 310 }, { "epoch": 0.08662696264212236, "grad_norm": 0.421875, "learning_rate": 0.0004899836095428884, "loss": 0.1954, "step": 320 }, { "epoch": 0.08933405522468868, "grad_norm": 0.267578125, "learning_rate": 0.0004895283190675652, "loss": 0.2326, "step": 330 }, { "epoch": 0.092041147807255, "grad_norm": 0.294921875, "learning_rate": 0.0004890730285922419, "loss": 0.2035, "step": 340 }, { "epoch": 0.09474824038982133, "grad_norm": 0.36328125, "learning_rate": 0.0004886177381169186, "loss": 0.224, "step": 350 }, { "epoch": 0.09745533297238766, "grad_norm": 0.1806640625, "learning_rate": 0.00048816244764159533, "loss": 0.239, "step": 360 }, { "epoch": 0.10016242555495398, "grad_norm": 0.1494140625, "learning_rate": 0.00048770715716627206, "loss": 0.2059, "step": 370 }, { "epoch": 0.10286951813752031, "grad_norm": 0.240234375, "learning_rate": 0.00048725186669094884, "loss": 0.2171, "step": 380 }, { "epoch": 0.10557661072008663, "grad_norm": 0.171875, "learning_rate": 0.00048679657621562557, "loss": 0.1666, "step": 390 }, { "epoch": 0.10828370330265295, "grad_norm": 0.34375, "learning_rate": 0.00048634128574030235, "loss": 0.2287, "step": 400 }, { "epoch": 0.10828370330265295, "eval_loss": 0.1630079299211502, "eval_runtime": 40.2797, "eval_samples_per_second": 12.413, "eval_steps_per_second": 0.794, "step": 400 }, { "epoch": 0.11099079588521928, "grad_norm": 0.3203125, "learning_rate": 0.0004858859952649791, "loss": 0.1821, "step": 410 }, { "epoch": 0.1136978884677856, "grad_norm": 0.49609375, "learning_rate": 0.0004854307047896558, "loss": 0.1857, "step": 420 }, { "epoch": 0.11640498105035192, "grad_norm": 0.4609375, "learning_rate": 0.0004849754143143326, "loss": 0.1873, "step": 430 }, { "epoch": 0.11911207363291824, "grad_norm": 0.24609375, "learning_rate": 0.0004845201238390093, "loss": 0.2127, "step": 440 }, { "epoch": 0.12181916621548457, "grad_norm": 0.328125, "learning_rate": 0.0004840648333636861, "loss": 0.1837, "step": 450 }, { "epoch": 0.12452625879805089, "grad_norm": 0.32421875, "learning_rate": 0.00048360954288836276, "loss": 0.1843, "step": 460 }, { "epoch": 0.12723335138061723, "grad_norm": 0.224609375, "learning_rate": 0.0004831542524130395, "loss": 0.1802, "step": 470 }, { "epoch": 0.12994044396318355, "grad_norm": 0.203125, "learning_rate": 0.00048269896193771627, "loss": 0.2079, "step": 480 }, { "epoch": 0.13264753654574987, "grad_norm": 0.2109375, "learning_rate": 0.000482243671462393, "loss": 0.1781, "step": 490 }, { "epoch": 0.1353546291283162, "grad_norm": 0.3203125, "learning_rate": 0.0004817883809870698, "loss": 0.2368, "step": 500 }, { "epoch": 0.1353546291283162, "eval_loss": 0.15445053577423096, "eval_runtime": 40.1882, "eval_samples_per_second": 12.441, "eval_steps_per_second": 0.796, "step": 500 }, { "epoch": 0.13806172171088252, "grad_norm": 0.376953125, "learning_rate": 0.0004813330905117465, "loss": 0.1727, "step": 510 }, { "epoch": 0.14076881429344884, "grad_norm": 0.4375, "learning_rate": 0.00048087780003642323, "loss": 0.1936, "step": 520 }, { "epoch": 0.14347590687601516, "grad_norm": 0.365234375, "learning_rate": 0.0004804225095611, "loss": 0.2074, "step": 530 }, { "epoch": 0.1461829994585815, "grad_norm": 0.765625, "learning_rate": 0.00047996721908577674, "loss": 0.17, "step": 540 }, { "epoch": 0.1488900920411478, "grad_norm": 0.306640625, "learning_rate": 0.00047951192861045346, "loss": 0.1836, "step": 550 }, { "epoch": 0.15159718462371413, "grad_norm": 0.408203125, "learning_rate": 0.00047905663813513024, "loss": 0.1923, "step": 560 }, { "epoch": 0.15430427720628045, "grad_norm": 0.25390625, "learning_rate": 0.00047860134765980697, "loss": 0.1754, "step": 570 }, { "epoch": 0.15701136978884678, "grad_norm": 0.44140625, "learning_rate": 0.00047814605718448375, "loss": 0.2121, "step": 580 }, { "epoch": 0.1597184623714131, "grad_norm": 0.1845703125, "learning_rate": 0.0004776907667091605, "loss": 0.2001, "step": 590 }, { "epoch": 0.16242555495397942, "grad_norm": 0.201171875, "learning_rate": 0.00047723547623383715, "loss": 0.1309, "step": 600 }, { "epoch": 0.16242555495397942, "eval_loss": 0.14666543900966644, "eval_runtime": 38.741, "eval_samples_per_second": 12.906, "eval_steps_per_second": 0.826, "step": 600 }, { "epoch": 0.16513264753654575, "grad_norm": 0.16015625, "learning_rate": 0.00047678018575851393, "loss": 0.1646, "step": 610 }, { "epoch": 0.16783974011911207, "grad_norm": 0.1279296875, "learning_rate": 0.00047632489528319066, "loss": 0.1928, "step": 620 }, { "epoch": 0.1705468327016784, "grad_norm": 0.310546875, "learning_rate": 0.00047586960480786744, "loss": 0.1708, "step": 630 }, { "epoch": 0.17325392528424471, "grad_norm": 0.216796875, "learning_rate": 0.00047541431433254416, "loss": 0.1991, "step": 640 }, { "epoch": 0.17596101786681104, "grad_norm": 0.400390625, "learning_rate": 0.0004749590238572209, "loss": 0.177, "step": 650 }, { "epoch": 0.17866811044937736, "grad_norm": 0.396484375, "learning_rate": 0.00047450373338189767, "loss": 0.1734, "step": 660 }, { "epoch": 0.18137520303194368, "grad_norm": 0.318359375, "learning_rate": 0.0004740484429065744, "loss": 0.1558, "step": 670 }, { "epoch": 0.18408229561451, "grad_norm": 0.474609375, "learning_rate": 0.0004735931524312512, "loss": 0.2152, "step": 680 }, { "epoch": 0.18678938819707633, "grad_norm": 0.248046875, "learning_rate": 0.0004731378619559279, "loss": 0.1777, "step": 690 }, { "epoch": 0.18949648077964265, "grad_norm": 0.318359375, "learning_rate": 0.00047268257148060463, "loss": 0.1552, "step": 700 }, { "epoch": 0.18949648077964265, "eval_loss": 0.13972881436347961, "eval_runtime": 39.3816, "eval_samples_per_second": 12.696, "eval_steps_per_second": 0.813, "step": 700 }, { "epoch": 0.19220357336220897, "grad_norm": 0.1884765625, "learning_rate": 0.0004722272810052814, "loss": 0.2114, "step": 710 }, { "epoch": 0.19491066594477532, "grad_norm": 0.3359375, "learning_rate": 0.00047177199052995814, "loss": 0.1614, "step": 720 }, { "epoch": 0.19761775852734165, "grad_norm": 0.326171875, "learning_rate": 0.00047131670005463487, "loss": 0.1879, "step": 730 }, { "epoch": 0.20032485110990797, "grad_norm": 0.24609375, "learning_rate": 0.0004708614095793116, "loss": 0.1607, "step": 740 }, { "epoch": 0.2030319436924743, "grad_norm": 0.291015625, "learning_rate": 0.0004704061191039883, "loss": 0.1613, "step": 750 }, { "epoch": 0.20573903627504062, "grad_norm": 0.1982421875, "learning_rate": 0.0004699508286286651, "loss": 0.1906, "step": 760 }, { "epoch": 0.20844612885760694, "grad_norm": 0.37109375, "learning_rate": 0.0004694955381533418, "loss": 0.1471, "step": 770 }, { "epoch": 0.21115322144017326, "grad_norm": 0.28125, "learning_rate": 0.00046904024767801855, "loss": 0.1602, "step": 780 }, { "epoch": 0.21386031402273958, "grad_norm": 0.5390625, "learning_rate": 0.00046858495720269533, "loss": 0.1933, "step": 790 }, { "epoch": 0.2165674066053059, "grad_norm": 0.24609375, "learning_rate": 0.00046812966672737206, "loss": 0.1765, "step": 800 }, { "epoch": 0.2165674066053059, "eval_loss": 0.1443299651145935, "eval_runtime": 40.7468, "eval_samples_per_second": 12.271, "eval_steps_per_second": 0.785, "step": 800 }, { "epoch": 0.21927449918787223, "grad_norm": 0.2578125, "learning_rate": 0.00046767437625204884, "loss": 0.1394, "step": 810 }, { "epoch": 0.22198159177043855, "grad_norm": 0.37890625, "learning_rate": 0.00046721908577672557, "loss": 0.1643, "step": 820 }, { "epoch": 0.22468868435300487, "grad_norm": 0.2138671875, "learning_rate": 0.0004667637953014023, "loss": 0.1483, "step": 830 }, { "epoch": 0.2273957769355712, "grad_norm": 0.2890625, "learning_rate": 0.0004663085048260791, "loss": 0.1532, "step": 840 }, { "epoch": 0.23010286951813752, "grad_norm": 0.30078125, "learning_rate": 0.0004658532143507558, "loss": 0.2081, "step": 850 }, { "epoch": 0.23280996210070384, "grad_norm": 0.298828125, "learning_rate": 0.0004653979238754326, "loss": 0.1947, "step": 860 }, { "epoch": 0.23551705468327017, "grad_norm": 0.12451171875, "learning_rate": 0.0004649426334001093, "loss": 0.1717, "step": 870 }, { "epoch": 0.2382241472658365, "grad_norm": 0.1162109375, "learning_rate": 0.000464487342924786, "loss": 0.1189, "step": 880 }, { "epoch": 0.2409312398484028, "grad_norm": 0.1416015625, "learning_rate": 0.00046403205244946276, "loss": 0.1495, "step": 890 }, { "epoch": 0.24363833243096913, "grad_norm": 0.0849609375, "learning_rate": 0.0004635767619741395, "loss": 0.1301, "step": 900 }, { "epoch": 0.24363833243096913, "eval_loss": 0.14148040115833282, "eval_runtime": 40.7774, "eval_samples_per_second": 12.262, "eval_steps_per_second": 0.785, "step": 900 }, { "epoch": 0.24634542501353546, "grad_norm": 0.359375, "learning_rate": 0.0004631214714988162, "loss": 0.143, "step": 910 }, { "epoch": 0.24905251759610178, "grad_norm": 0.318359375, "learning_rate": 0.000462666181023493, "loss": 0.1708, "step": 920 }, { "epoch": 0.2517596101786681, "grad_norm": 0.291015625, "learning_rate": 0.0004622108905481697, "loss": 0.1803, "step": 930 }, { "epoch": 0.25446670276123445, "grad_norm": 0.224609375, "learning_rate": 0.0004617556000728465, "loss": 0.1415, "step": 940 }, { "epoch": 0.25717379534380075, "grad_norm": 0.1435546875, "learning_rate": 0.00046130030959752323, "loss": 0.1482, "step": 950 }, { "epoch": 0.2598808879263671, "grad_norm": 0.373046875, "learning_rate": 0.00046084501912219995, "loss": 0.1569, "step": 960 }, { "epoch": 0.2625879805089334, "grad_norm": 0.15234375, "learning_rate": 0.00046038972864687674, "loss": 0.1617, "step": 970 }, { "epoch": 0.26529507309149974, "grad_norm": 0.1123046875, "learning_rate": 0.00045993443817155346, "loss": 0.1533, "step": 980 }, { "epoch": 0.26800216567406604, "grad_norm": 0.111328125, "learning_rate": 0.00045947914769623024, "loss": 0.1441, "step": 990 }, { "epoch": 0.2707092582566324, "grad_norm": 0.208984375, "learning_rate": 0.00045902385722090697, "loss": 0.1274, "step": 1000 }, { "epoch": 0.2707092582566324, "eval_loss": 0.12552116811275482, "eval_runtime": 42.5326, "eval_samples_per_second": 11.756, "eval_steps_per_second": 0.752, "step": 1000 }, { "epoch": 0.2734163508391987, "grad_norm": 0.166015625, "learning_rate": 0.0004585685667455837, "loss": 0.1329, "step": 1010 }, { "epoch": 0.27612344342176504, "grad_norm": 0.47265625, "learning_rate": 0.0004581132762702604, "loss": 0.129, "step": 1020 }, { "epoch": 0.27883053600433133, "grad_norm": 0.421875, "learning_rate": 0.00045765798579493715, "loss": 0.142, "step": 1030 }, { "epoch": 0.2815376285868977, "grad_norm": 0.2158203125, "learning_rate": 0.00045720269531961393, "loss": 0.1477, "step": 1040 }, { "epoch": 0.284244721169464, "grad_norm": 0.578125, "learning_rate": 0.00045674740484429066, "loss": 0.1797, "step": 1050 }, { "epoch": 0.2869518137520303, "grad_norm": 0.2060546875, "learning_rate": 0.0004562921143689674, "loss": 0.1634, "step": 1060 }, { "epoch": 0.2896589063345966, "grad_norm": 0.28125, "learning_rate": 0.00045583682389364416, "loss": 0.1546, "step": 1070 }, { "epoch": 0.292365998917163, "grad_norm": 0.291015625, "learning_rate": 0.0004553815334183209, "loss": 0.1368, "step": 1080 }, { "epoch": 0.29507309149972927, "grad_norm": 0.173828125, "learning_rate": 0.0004549262429429976, "loss": 0.166, "step": 1090 }, { "epoch": 0.2977801840822956, "grad_norm": 0.2373046875, "learning_rate": 0.0004544709524676744, "loss": 0.1418, "step": 1100 }, { "epoch": 0.2977801840822956, "eval_loss": 0.12966635823249817, "eval_runtime": 39.9103, "eval_samples_per_second": 12.528, "eval_steps_per_second": 0.802, "step": 1100 }, { "epoch": 0.3004872766648619, "grad_norm": 0.1376953125, "learning_rate": 0.0004540156619923511, "loss": 0.2007, "step": 1110 }, { "epoch": 0.30319436924742826, "grad_norm": 0.1533203125, "learning_rate": 0.0004535603715170279, "loss": 0.1579, "step": 1120 }, { "epoch": 0.30590146182999456, "grad_norm": 0.1357421875, "learning_rate": 0.00045310508104170463, "loss": 0.1262, "step": 1130 }, { "epoch": 0.3086085544125609, "grad_norm": 0.330078125, "learning_rate": 0.00045264979056638136, "loss": 0.1351, "step": 1140 }, { "epoch": 0.31131564699512726, "grad_norm": 0.1376953125, "learning_rate": 0.00045219450009105814, "loss": 0.1318, "step": 1150 }, { "epoch": 0.31402273957769355, "grad_norm": 0.2099609375, "learning_rate": 0.0004517392096157348, "loss": 0.1377, "step": 1160 }, { "epoch": 0.3167298321602599, "grad_norm": 0.419921875, "learning_rate": 0.0004512839191404116, "loss": 0.1786, "step": 1170 }, { "epoch": 0.3194369247428262, "grad_norm": 0.306640625, "learning_rate": 0.0004508286286650883, "loss": 0.1914, "step": 1180 }, { "epoch": 0.32214401732539255, "grad_norm": 0.1865234375, "learning_rate": 0.00045037333818976504, "loss": 0.1458, "step": 1190 }, { "epoch": 0.32485110990795885, "grad_norm": 0.443359375, "learning_rate": 0.0004499180477144418, "loss": 0.1254, "step": 1200 }, { "epoch": 0.32485110990795885, "eval_loss": 0.12671491503715515, "eval_runtime": 39.828, "eval_samples_per_second": 12.554, "eval_steps_per_second": 0.803, "step": 1200 }, { "epoch": 0.3275582024905252, "grad_norm": 0.15625, "learning_rate": 0.00044946275723911855, "loss": 0.1333, "step": 1210 }, { "epoch": 0.3302652950730915, "grad_norm": 0.240234375, "learning_rate": 0.00044900746676379533, "loss": 0.1459, "step": 1220 }, { "epoch": 0.33297238765565784, "grad_norm": 0.31640625, "learning_rate": 0.00044855217628847206, "loss": 0.1877, "step": 1230 }, { "epoch": 0.33567948023822414, "grad_norm": 0.4609375, "learning_rate": 0.0004480968858131488, "loss": 0.1332, "step": 1240 }, { "epoch": 0.3383865728207905, "grad_norm": 0.2294921875, "learning_rate": 0.00044764159533782557, "loss": 0.1538, "step": 1250 }, { "epoch": 0.3410936654033568, "grad_norm": 0.203125, "learning_rate": 0.0004471863048625023, "loss": 0.1322, "step": 1260 }, { "epoch": 0.34380075798592313, "grad_norm": 0.392578125, "learning_rate": 0.000446731014387179, "loss": 0.1662, "step": 1270 }, { "epoch": 0.34650785056848943, "grad_norm": 0.1962890625, "learning_rate": 0.0004462757239118558, "loss": 0.1997, "step": 1280 }, { "epoch": 0.3492149431510558, "grad_norm": 0.255859375, "learning_rate": 0.0004458204334365325, "loss": 0.1633, "step": 1290 }, { "epoch": 0.3519220357336221, "grad_norm": 0.38671875, "learning_rate": 0.0004453651429612093, "loss": 0.179, "step": 1300 }, { "epoch": 0.3519220357336221, "eval_loss": 0.12946651875972748, "eval_runtime": 39.842, "eval_samples_per_second": 12.55, "eval_steps_per_second": 0.803, "step": 1300 }, { "epoch": 0.3546291283161884, "grad_norm": 0.142578125, "learning_rate": 0.000444909852485886, "loss": 0.1617, "step": 1310 }, { "epoch": 0.3573362208987547, "grad_norm": 0.51171875, "learning_rate": 0.0004444545620105627, "loss": 0.164, "step": 1320 }, { "epoch": 0.36004331348132107, "grad_norm": 0.35546875, "learning_rate": 0.0004439992715352395, "loss": 0.1901, "step": 1330 }, { "epoch": 0.36275040606388737, "grad_norm": 0.390625, "learning_rate": 0.0004435439810599162, "loss": 0.1779, "step": 1340 }, { "epoch": 0.3654574986464537, "grad_norm": 0.328125, "learning_rate": 0.000443088690584593, "loss": 0.1609, "step": 1350 }, { "epoch": 0.36816459122902, "grad_norm": 0.1767578125, "learning_rate": 0.0004426334001092697, "loss": 0.1167, "step": 1360 }, { "epoch": 0.37087168381158636, "grad_norm": 0.26171875, "learning_rate": 0.00044217810963394645, "loss": 0.1552, "step": 1370 }, { "epoch": 0.37357877639415266, "grad_norm": 0.2373046875, "learning_rate": 0.0004417228191586232, "loss": 0.1152, "step": 1380 }, { "epoch": 0.376285868976719, "grad_norm": 0.265625, "learning_rate": 0.00044126752868329995, "loss": 0.182, "step": 1390 }, { "epoch": 0.3789929615592853, "grad_norm": 0.1923828125, "learning_rate": 0.00044081223820797673, "loss": 0.1585, "step": 1400 }, { "epoch": 0.3789929615592853, "eval_loss": 0.1260078102350235, "eval_runtime": 40.0713, "eval_samples_per_second": 12.478, "eval_steps_per_second": 0.799, "step": 1400 }, { "epoch": 0.38170005414185165, "grad_norm": 0.265625, "learning_rate": 0.00044035694773265346, "loss": 0.1424, "step": 1410 }, { "epoch": 0.38440714672441795, "grad_norm": 0.2392578125, "learning_rate": 0.0004399016572573302, "loss": 0.1662, "step": 1420 }, { "epoch": 0.3871142393069843, "grad_norm": 0.291015625, "learning_rate": 0.00043944636678200697, "loss": 0.148, "step": 1430 }, { "epoch": 0.38982133188955065, "grad_norm": 0.408203125, "learning_rate": 0.0004389910763066837, "loss": 0.1408, "step": 1440 }, { "epoch": 0.39252842447211694, "grad_norm": 0.07275390625, "learning_rate": 0.00043853578583136037, "loss": 0.1855, "step": 1450 }, { "epoch": 0.3952355170546833, "grad_norm": 0.1767578125, "learning_rate": 0.00043808049535603715, "loss": 0.1475, "step": 1460 }, { "epoch": 0.3979426096372496, "grad_norm": 0.1298828125, "learning_rate": 0.0004376252048807139, "loss": 0.1319, "step": 1470 }, { "epoch": 0.40064970221981594, "grad_norm": 0.138671875, "learning_rate": 0.00043716991440539065, "loss": 0.1863, "step": 1480 }, { "epoch": 0.40335679480238223, "grad_norm": 0.451171875, "learning_rate": 0.0004367146239300674, "loss": 0.15, "step": 1490 }, { "epoch": 0.4060638873849486, "grad_norm": 0.328125, "learning_rate": 0.0004362593334547441, "loss": 0.155, "step": 1500 }, { "epoch": 0.4060638873849486, "eval_loss": 0.11785156279802322, "eval_runtime": 40.5053, "eval_samples_per_second": 12.344, "eval_steps_per_second": 0.79, "step": 1500 }, { "epoch": 0.4087709799675149, "grad_norm": 0.376953125, "learning_rate": 0.0004358040429794209, "loss": 0.1523, "step": 1510 }, { "epoch": 0.41147807255008123, "grad_norm": 0.48828125, "learning_rate": 0.0004353487525040976, "loss": 0.1539, "step": 1520 }, { "epoch": 0.4141851651326475, "grad_norm": 0.296875, "learning_rate": 0.0004348934620287744, "loss": 0.1717, "step": 1530 }, { "epoch": 0.4168922577152139, "grad_norm": 0.208984375, "learning_rate": 0.0004344381715534511, "loss": 0.1356, "step": 1540 }, { "epoch": 0.41959935029778017, "grad_norm": 0.25390625, "learning_rate": 0.00043398288107812785, "loss": 0.1469, "step": 1550 }, { "epoch": 0.4223064428803465, "grad_norm": 0.30078125, "learning_rate": 0.00043352759060280463, "loss": 0.1361, "step": 1560 }, { "epoch": 0.4250135354629128, "grad_norm": 0.15234375, "learning_rate": 0.00043307230012748136, "loss": 0.1636, "step": 1570 }, { "epoch": 0.42772062804547917, "grad_norm": 0.408203125, "learning_rate": 0.00043261700965215814, "loss": 0.174, "step": 1580 }, { "epoch": 0.43042772062804546, "grad_norm": 0.259765625, "learning_rate": 0.0004321617191768348, "loss": 0.1423, "step": 1590 }, { "epoch": 0.4331348132106118, "grad_norm": 0.2412109375, "learning_rate": 0.00043170642870151154, "loss": 0.147, "step": 1600 }, { "epoch": 0.4331348132106118, "eval_loss": 0.12018130719661713, "eval_runtime": 39.6508, "eval_samples_per_second": 12.61, "eval_steps_per_second": 0.807, "step": 1600 }, { "epoch": 0.4358419057931781, "grad_norm": 0.0859375, "learning_rate": 0.0004312511382261883, "loss": 0.1292, "step": 1610 }, { "epoch": 0.43854899837574446, "grad_norm": 0.1337890625, "learning_rate": 0.00043079584775086504, "loss": 0.1546, "step": 1620 }, { "epoch": 0.44125609095831075, "grad_norm": 0.2578125, "learning_rate": 0.00043034055727554177, "loss": 0.133, "step": 1630 }, { "epoch": 0.4439631835408771, "grad_norm": 0.4765625, "learning_rate": 0.00042988526680021855, "loss": 0.1476, "step": 1640 }, { "epoch": 0.4466702761234434, "grad_norm": 0.255859375, "learning_rate": 0.0004294299763248953, "loss": 0.1637, "step": 1650 }, { "epoch": 0.44937736870600975, "grad_norm": 0.09423828125, "learning_rate": 0.00042897468584957206, "loss": 0.1636, "step": 1660 }, { "epoch": 0.45208446128857604, "grad_norm": 0.34375, "learning_rate": 0.0004285193953742488, "loss": 0.1591, "step": 1670 }, { "epoch": 0.4547915538711424, "grad_norm": 0.404296875, "learning_rate": 0.0004280641048989255, "loss": 0.1597, "step": 1680 }, { "epoch": 0.4574986464537087, "grad_norm": 0.0908203125, "learning_rate": 0.0004276088144236023, "loss": 0.1283, "step": 1690 }, { "epoch": 0.46020573903627504, "grad_norm": 0.3359375, "learning_rate": 0.000427153523948279, "loss": 0.1108, "step": 1700 }, { "epoch": 0.46020573903627504, "eval_loss": 0.1205587238073349, "eval_runtime": 40.7397, "eval_samples_per_second": 12.273, "eval_steps_per_second": 0.785, "step": 1700 }, { "epoch": 0.4629128316188414, "grad_norm": 0.3359375, "learning_rate": 0.0004266982334729558, "loss": 0.1419, "step": 1710 }, { "epoch": 0.4656199242014077, "grad_norm": 0.228515625, "learning_rate": 0.0004262429429976325, "loss": 0.1558, "step": 1720 }, { "epoch": 0.46832701678397404, "grad_norm": 0.27734375, "learning_rate": 0.0004257876525223092, "loss": 0.1792, "step": 1730 }, { "epoch": 0.47103410936654033, "grad_norm": 0.07568359375, "learning_rate": 0.000425332362046986, "loss": 0.1341, "step": 1740 }, { "epoch": 0.4737412019491067, "grad_norm": 0.1708984375, "learning_rate": 0.0004248770715716627, "loss": 0.1653, "step": 1750 }, { "epoch": 0.476448294531673, "grad_norm": 0.2041015625, "learning_rate": 0.0004244217810963395, "loss": 0.1651, "step": 1760 }, { "epoch": 0.47915538711423933, "grad_norm": 0.259765625, "learning_rate": 0.0004239664906210162, "loss": 0.1862, "step": 1770 }, { "epoch": 0.4818624796968056, "grad_norm": 0.2041015625, "learning_rate": 0.00042351120014569294, "loss": 0.1757, "step": 1780 }, { "epoch": 0.484569572279372, "grad_norm": 0.3984375, "learning_rate": 0.0004230559096703697, "loss": 0.1238, "step": 1790 }, { "epoch": 0.48727666486193827, "grad_norm": 0.46484375, "learning_rate": 0.00042260061919504645, "loss": 0.1532, "step": 1800 }, { "epoch": 0.48727666486193827, "eval_loss": 0.11244313418865204, "eval_runtime": 37.6449, "eval_samples_per_second": 13.282, "eval_steps_per_second": 0.85, "step": 1800 }, { "epoch": 0.4899837574445046, "grad_norm": 0.26171875, "learning_rate": 0.00042214532871972317, "loss": 0.1312, "step": 1810 }, { "epoch": 0.4926908500270709, "grad_norm": 0.271484375, "learning_rate": 0.00042169003824439995, "loss": 0.1872, "step": 1820 }, { "epoch": 0.49539794260963727, "grad_norm": 0.31640625, "learning_rate": 0.0004212347477690767, "loss": 0.1502, "step": 1830 }, { "epoch": 0.49810503519220356, "grad_norm": 0.1806640625, "learning_rate": 0.00042077945729375346, "loss": 0.1473, "step": 1840 }, { "epoch": 0.5008121277747699, "grad_norm": 0.265625, "learning_rate": 0.0004203241668184302, "loss": 0.1656, "step": 1850 }, { "epoch": 0.5035192203573362, "grad_norm": 0.275390625, "learning_rate": 0.0004198688763431069, "loss": 0.1327, "step": 1860 }, { "epoch": 0.5062263129399025, "grad_norm": 0.095703125, "learning_rate": 0.00041941358586778364, "loss": 0.1291, "step": 1870 }, { "epoch": 0.5089334055224689, "grad_norm": 0.361328125, "learning_rate": 0.00041895829539246037, "loss": 0.1405, "step": 1880 }, { "epoch": 0.5116404981050352, "grad_norm": 0.1767578125, "learning_rate": 0.00041850300491713715, "loss": 0.1347, "step": 1890 }, { "epoch": 0.5143475906876015, "grad_norm": 0.11865234375, "learning_rate": 0.0004180477144418139, "loss": 0.1181, "step": 1900 }, { "epoch": 0.5143475906876015, "eval_loss": 0.11964208632707596, "eval_runtime": 39.8678, "eval_samples_per_second": 12.541, "eval_steps_per_second": 0.803, "step": 1900 }, { "epoch": 0.5170546832701678, "grad_norm": 0.103515625, "learning_rate": 0.0004175924239664906, "loss": 0.1525, "step": 1910 }, { "epoch": 0.5197617758527342, "grad_norm": 0.150390625, "learning_rate": 0.0004171371334911674, "loss": 0.1495, "step": 1920 }, { "epoch": 0.5224688684353005, "grad_norm": 0.72265625, "learning_rate": 0.0004166818430158441, "loss": 0.1399, "step": 1930 }, { "epoch": 0.5251759610178668, "grad_norm": 0.37109375, "learning_rate": 0.0004162265525405209, "loss": 0.156, "step": 1940 }, { "epoch": 0.5278830536004331, "grad_norm": 0.2138671875, "learning_rate": 0.0004157712620651976, "loss": 0.1587, "step": 1950 }, { "epoch": 0.5305901461829995, "grad_norm": 0.44921875, "learning_rate": 0.00041531597158987434, "loss": 0.1364, "step": 1960 }, { "epoch": 0.5332972387655658, "grad_norm": 0.30078125, "learning_rate": 0.0004148606811145511, "loss": 0.1652, "step": 1970 }, { "epoch": 0.5360043313481321, "grad_norm": 0.1328125, "learning_rate": 0.00041440539063922785, "loss": 0.1502, "step": 1980 }, { "epoch": 0.5387114239306985, "grad_norm": 0.28515625, "learning_rate": 0.0004139501001639046, "loss": 0.1634, "step": 1990 }, { "epoch": 0.5414185165132648, "grad_norm": 0.12060546875, "learning_rate": 0.00041349480968858136, "loss": 0.1537, "step": 2000 }, { "epoch": 0.5414185165132648, "eval_loss": 0.12083197385072708, "eval_runtime": 40.7195, "eval_samples_per_second": 12.279, "eval_steps_per_second": 0.786, "step": 2000 }, { "epoch": 0.5441256090958311, "grad_norm": 0.28515625, "learning_rate": 0.00041303951921325803, "loss": 0.1335, "step": 2010 }, { "epoch": 0.5468327016783974, "grad_norm": 0.357421875, "learning_rate": 0.0004125842287379348, "loss": 0.121, "step": 2020 }, { "epoch": 0.5495397942609638, "grad_norm": 0.15234375, "learning_rate": 0.00041212893826261153, "loss": 0.1336, "step": 2030 }, { "epoch": 0.5522468868435301, "grad_norm": 0.1943359375, "learning_rate": 0.00041167364778728826, "loss": 0.1264, "step": 2040 }, { "epoch": 0.5549539794260964, "grad_norm": 0.177734375, "learning_rate": 0.00041121835731196504, "loss": 0.1216, "step": 2050 }, { "epoch": 0.5576610720086627, "grad_norm": 0.20703125, "learning_rate": 0.00041076306683664177, "loss": 0.1247, "step": 2060 }, { "epoch": 0.5603681645912291, "grad_norm": 0.380859375, "learning_rate": 0.00041030777636131855, "loss": 0.1407, "step": 2070 }, { "epoch": 0.5630752571737954, "grad_norm": 0.2197265625, "learning_rate": 0.0004098524858859953, "loss": 0.1428, "step": 2080 }, { "epoch": 0.5657823497563617, "grad_norm": 0.32421875, "learning_rate": 0.000409397195410672, "loss": 0.1547, "step": 2090 }, { "epoch": 0.568489442338928, "grad_norm": 0.236328125, "learning_rate": 0.0004089419049353488, "loss": 0.1366, "step": 2100 }, { "epoch": 0.568489442338928, "eval_loss": 0.11787446588277817, "eval_runtime": 38.4013, "eval_samples_per_second": 13.02, "eval_steps_per_second": 0.833, "step": 2100 }, { "epoch": 0.5711965349214944, "grad_norm": 0.2734375, "learning_rate": 0.0004084866144600255, "loss": 0.1169, "step": 2110 }, { "epoch": 0.5739036275040607, "grad_norm": 0.267578125, "learning_rate": 0.0004080313239847023, "loss": 0.1006, "step": 2120 }, { "epoch": 0.576610720086627, "grad_norm": 0.294921875, "learning_rate": 0.000407576033509379, "loss": 0.1405, "step": 2130 }, { "epoch": 0.5793178126691932, "grad_norm": 0.37890625, "learning_rate": 0.00040712074303405574, "loss": 0.139, "step": 2140 }, { "epoch": 0.5820249052517596, "grad_norm": 0.2236328125, "learning_rate": 0.00040666545255873247, "loss": 0.1465, "step": 2150 }, { "epoch": 0.584731997834326, "grad_norm": 0.2314453125, "learning_rate": 0.0004062101620834092, "loss": 0.1246, "step": 2160 }, { "epoch": 0.5874390904168922, "grad_norm": 0.37890625, "learning_rate": 0.0004057548716080859, "loss": 0.1536, "step": 2170 }, { "epoch": 0.5901461829994585, "grad_norm": 0.17578125, "learning_rate": 0.0004052995811327627, "loss": 0.1126, "step": 2180 }, { "epoch": 0.5928532755820249, "grad_norm": 0.58203125, "learning_rate": 0.00040484429065743943, "loss": 0.1569, "step": 2190 }, { "epoch": 0.5955603681645912, "grad_norm": 0.2021484375, "learning_rate": 0.0004043890001821162, "loss": 0.1199, "step": 2200 }, { "epoch": 0.5955603681645912, "eval_loss": 0.1216907724738121, "eval_runtime": 38.6482, "eval_samples_per_second": 12.937, "eval_steps_per_second": 0.828, "step": 2200 }, { "epoch": 0.5982674607471575, "grad_norm": 0.099609375, "learning_rate": 0.00040393370970679294, "loss": 0.1642, "step": 2210 }, { "epoch": 0.6009745533297238, "grad_norm": 0.146484375, "learning_rate": 0.00040347841923146966, "loss": 0.1083, "step": 2220 }, { "epoch": 0.6036816459122902, "grad_norm": 0.546875, "learning_rate": 0.00040302312875614644, "loss": 0.2072, "step": 2230 }, { "epoch": 0.6063887384948565, "grad_norm": 0.447265625, "learning_rate": 0.00040256783828082317, "loss": 0.1411, "step": 2240 }, { "epoch": 0.6090958310774228, "grad_norm": 0.21484375, "learning_rate": 0.00040211254780549995, "loss": 0.1206, "step": 2250 }, { "epoch": 0.6118029236599891, "grad_norm": 0.2333984375, "learning_rate": 0.0004016572573301767, "loss": 0.1248, "step": 2260 }, { "epoch": 0.6145100162425555, "grad_norm": 0.44921875, "learning_rate": 0.0004012019668548534, "loss": 0.1638, "step": 2270 }, { "epoch": 0.6172171088251218, "grad_norm": 0.3203125, "learning_rate": 0.0004007466763795302, "loss": 0.1537, "step": 2280 }, { "epoch": 0.6199242014076881, "grad_norm": 0.2041015625, "learning_rate": 0.00040029138590420686, "loss": 0.1329, "step": 2290 }, { "epoch": 0.6226312939902545, "grad_norm": 0.2578125, "learning_rate": 0.00039983609542888364, "loss": 0.1315, "step": 2300 }, { "epoch": 0.6226312939902545, "eval_loss": 0.12395916134119034, "eval_runtime": 40.0386, "eval_samples_per_second": 12.488, "eval_steps_per_second": 0.799, "step": 2300 }, { "epoch": 0.6253383865728208, "grad_norm": 0.361328125, "learning_rate": 0.00039938080495356037, "loss": 0.1252, "step": 2310 }, { "epoch": 0.6280454791553871, "grad_norm": 0.328125, "learning_rate": 0.0003989255144782371, "loss": 0.1433, "step": 2320 }, { "epoch": 0.6307525717379534, "grad_norm": 0.400390625, "learning_rate": 0.00039847022400291387, "loss": 0.1239, "step": 2330 }, { "epoch": 0.6334596643205198, "grad_norm": 0.24609375, "learning_rate": 0.0003980149335275906, "loss": 0.1299, "step": 2340 }, { "epoch": 0.6361667569030861, "grad_norm": 0.1376953125, "learning_rate": 0.0003975596430522673, "loss": 0.113, "step": 2350 }, { "epoch": 0.6388738494856524, "grad_norm": 0.30078125, "learning_rate": 0.0003971043525769441, "loss": 0.1001, "step": 2360 }, { "epoch": 0.6415809420682187, "grad_norm": 0.1337890625, "learning_rate": 0.00039664906210162083, "loss": 0.1617, "step": 2370 }, { "epoch": 0.6442880346507851, "grad_norm": 0.546875, "learning_rate": 0.0003961937716262976, "loss": 0.1531, "step": 2380 }, { "epoch": 0.6469951272333514, "grad_norm": 0.1572265625, "learning_rate": 0.00039573848115097434, "loss": 0.1485, "step": 2390 }, { "epoch": 0.6497022198159177, "grad_norm": 0.275390625, "learning_rate": 0.00039528319067565107, "loss": 0.1461, "step": 2400 }, { "epoch": 0.6497022198159177, "eval_loss": 0.12797197699546814, "eval_runtime": 40.7824, "eval_samples_per_second": 12.26, "eval_steps_per_second": 0.785, "step": 2400 }, { "epoch": 0.652409312398484, "grad_norm": 0.373046875, "learning_rate": 0.00039482790020032785, "loss": 0.1564, "step": 2410 }, { "epoch": 0.6551164049810504, "grad_norm": 0.27734375, "learning_rate": 0.0003943726097250046, "loss": 0.1616, "step": 2420 }, { "epoch": 0.6578234975636167, "grad_norm": 0.1923828125, "learning_rate": 0.00039391731924968135, "loss": 0.157, "step": 2430 }, { "epoch": 0.660530590146183, "grad_norm": 0.2109375, "learning_rate": 0.000393462028774358, "loss": 0.1237, "step": 2440 }, { "epoch": 0.6632376827287493, "grad_norm": 0.23828125, "learning_rate": 0.00039300673829903475, "loss": 0.1297, "step": 2450 }, { "epoch": 0.6659447753113157, "grad_norm": 0.349609375, "learning_rate": 0.00039255144782371153, "loss": 0.1133, "step": 2460 }, { "epoch": 0.668651867893882, "grad_norm": 0.2734375, "learning_rate": 0.00039209615734838826, "loss": 0.11, "step": 2470 }, { "epoch": 0.6713589604764483, "grad_norm": 0.373046875, "learning_rate": 0.00039164086687306504, "loss": 0.1271, "step": 2480 }, { "epoch": 0.6740660530590146, "grad_norm": 0.326171875, "learning_rate": 0.00039118557639774177, "loss": 0.1413, "step": 2490 }, { "epoch": 0.676773145641581, "grad_norm": 0.2890625, "learning_rate": 0.0003907302859224185, "loss": 0.1343, "step": 2500 }, { "epoch": 0.676773145641581, "eval_loss": 0.11558062583208084, "eval_runtime": 41.5373, "eval_samples_per_second": 12.037, "eval_steps_per_second": 0.77, "step": 2500 }, { "epoch": 0.6794802382241473, "grad_norm": 0.28125, "learning_rate": 0.0003902749954470953, "loss": 0.1694, "step": 2510 }, { "epoch": 0.6821873308067136, "grad_norm": 0.271484375, "learning_rate": 0.000389819704971772, "loss": 0.1416, "step": 2520 }, { "epoch": 0.6848944233892799, "grad_norm": 0.37109375, "learning_rate": 0.00038936441449644873, "loss": 0.1422, "step": 2530 }, { "epoch": 0.6876015159718463, "grad_norm": 0.16015625, "learning_rate": 0.0003889091240211255, "loss": 0.1212, "step": 2540 }, { "epoch": 0.6903086085544126, "grad_norm": 0.296875, "learning_rate": 0.00038845383354580224, "loss": 0.1567, "step": 2550 }, { "epoch": 0.6930157011369789, "grad_norm": 0.0849609375, "learning_rate": 0.000387998543070479, "loss": 0.115, "step": 2560 }, { "epoch": 0.6957227937195453, "grad_norm": 0.09619140625, "learning_rate": 0.00038754325259515574, "loss": 0.122, "step": 2570 }, { "epoch": 0.6984298863021116, "grad_norm": 0.34375, "learning_rate": 0.0003870879621198324, "loss": 0.0994, "step": 2580 }, { "epoch": 0.7011369788846779, "grad_norm": 0.248046875, "learning_rate": 0.0003866326716445092, "loss": 0.1041, "step": 2590 }, { "epoch": 0.7038440714672441, "grad_norm": 0.287109375, "learning_rate": 0.0003861773811691859, "loss": 0.106, "step": 2600 }, { "epoch": 0.7038440714672441, "eval_loss": 0.11948926001787186, "eval_runtime": 39.4394, "eval_samples_per_second": 12.678, "eval_steps_per_second": 0.811, "step": 2600 }, { "epoch": 0.7065511640498106, "grad_norm": 0.0927734375, "learning_rate": 0.0003857220906938627, "loss": 0.1468, "step": 2610 }, { "epoch": 0.7092582566323768, "grad_norm": 0.2216796875, "learning_rate": 0.00038526680021853943, "loss": 0.1448, "step": 2620 }, { "epoch": 0.7119653492149431, "grad_norm": 0.259765625, "learning_rate": 0.00038481150974321616, "loss": 0.1665, "step": 2630 }, { "epoch": 0.7146724417975094, "grad_norm": 0.2158203125, "learning_rate": 0.00038435621926789294, "loss": 0.1368, "step": 2640 }, { "epoch": 0.7173795343800758, "grad_norm": 0.396484375, "learning_rate": 0.00038390092879256966, "loss": 0.1391, "step": 2650 }, { "epoch": 0.7200866269626421, "grad_norm": 0.259765625, "learning_rate": 0.00038344563831724644, "loss": 0.1338, "step": 2660 }, { "epoch": 0.7227937195452084, "grad_norm": 0.251953125, "learning_rate": 0.00038299034784192317, "loss": 0.141, "step": 2670 }, { "epoch": 0.7255008121277747, "grad_norm": 0.37890625, "learning_rate": 0.0003825350573665999, "loss": 0.1441, "step": 2680 }, { "epoch": 0.7282079047103411, "grad_norm": 0.421875, "learning_rate": 0.0003820797668912767, "loss": 0.1289, "step": 2690 }, { "epoch": 0.7309149972929074, "grad_norm": 0.10693359375, "learning_rate": 0.0003816244764159534, "loss": 0.115, "step": 2700 }, { "epoch": 0.7309149972929074, "eval_loss": 0.12208272516727448, "eval_runtime": 40.4577, "eval_samples_per_second": 12.359, "eval_steps_per_second": 0.791, "step": 2700 }, { "epoch": 0.7336220898754737, "grad_norm": 0.3359375, "learning_rate": 0.0003811691859406301, "loss": 0.1325, "step": 2710 }, { "epoch": 0.73632918245804, "grad_norm": 0.1064453125, "learning_rate": 0.00038071389546530686, "loss": 0.1413, "step": 2720 }, { "epoch": 0.7390362750406064, "grad_norm": 0.25390625, "learning_rate": 0.0003802586049899836, "loss": 0.127, "step": 2730 }, { "epoch": 0.7417433676231727, "grad_norm": 0.287109375, "learning_rate": 0.00037980331451466036, "loss": 0.127, "step": 2740 }, { "epoch": 0.744450460205739, "grad_norm": 0.189453125, "learning_rate": 0.0003793480240393371, "loss": 0.1172, "step": 2750 }, { "epoch": 0.7471575527883053, "grad_norm": 0.474609375, "learning_rate": 0.0003788927335640138, "loss": 0.1429, "step": 2760 }, { "epoch": 0.7498646453708717, "grad_norm": 0.376953125, "learning_rate": 0.0003784374430886906, "loss": 0.1345, "step": 2770 }, { "epoch": 0.752571737953438, "grad_norm": 0.2138671875, "learning_rate": 0.0003779821526133673, "loss": 0.1163, "step": 2780 }, { "epoch": 0.7552788305360043, "grad_norm": 0.103515625, "learning_rate": 0.0003775268621380441, "loss": 0.1324, "step": 2790 }, { "epoch": 0.7579859231185706, "grad_norm": 0.3359375, "learning_rate": 0.00037707157166272083, "loss": 0.1567, "step": 2800 }, { "epoch": 0.7579859231185706, "eval_loss": 0.12093514204025269, "eval_runtime": 42.4616, "eval_samples_per_second": 11.775, "eval_steps_per_second": 0.754, "step": 2800 }, { "epoch": 0.760693015701137, "grad_norm": 0.2236328125, "learning_rate": 0.00037661628118739756, "loss": 0.1366, "step": 2810 }, { "epoch": 0.7634001082837033, "grad_norm": 0.306640625, "learning_rate": 0.00037616099071207434, "loss": 0.1113, "step": 2820 }, { "epoch": 0.7661072008662696, "grad_norm": 0.1845703125, "learning_rate": 0.00037570570023675107, "loss": 0.0958, "step": 2830 }, { "epoch": 0.7688142934488359, "grad_norm": 0.189453125, "learning_rate": 0.00037525040976142785, "loss": 0.1482, "step": 2840 }, { "epoch": 0.7715213860314023, "grad_norm": 0.30859375, "learning_rate": 0.00037479511928610457, "loss": 0.1346, "step": 2850 }, { "epoch": 0.7742284786139686, "grad_norm": 0.33203125, "learning_rate": 0.00037433982881078124, "loss": 0.1138, "step": 2860 }, { "epoch": 0.7769355711965349, "grad_norm": 0.38671875, "learning_rate": 0.000373884538335458, "loss": 0.1257, "step": 2870 }, { "epoch": 0.7796426637791013, "grad_norm": 0.302734375, "learning_rate": 0.00037342924786013475, "loss": 0.1444, "step": 2880 }, { "epoch": 0.7823497563616676, "grad_norm": 0.1318359375, "learning_rate": 0.0003729739573848115, "loss": 0.134, "step": 2890 }, { "epoch": 0.7850568489442339, "grad_norm": 0.232421875, "learning_rate": 0.00037251866690948826, "loss": 0.1032, "step": 2900 }, { "epoch": 0.7850568489442339, "eval_loss": 0.11778505891561508, "eval_runtime": 40.6633, "eval_samples_per_second": 12.296, "eval_steps_per_second": 0.787, "step": 2900 }, { "epoch": 0.7877639415268002, "grad_norm": 0.45703125, "learning_rate": 0.000372063376434165, "loss": 0.1358, "step": 2910 }, { "epoch": 0.7904710341093666, "grad_norm": 0.2275390625, "learning_rate": 0.00037160808595884177, "loss": 0.1341, "step": 2920 }, { "epoch": 0.7931781266919329, "grad_norm": 0.298828125, "learning_rate": 0.0003711527954835185, "loss": 0.131, "step": 2930 }, { "epoch": 0.7958852192744992, "grad_norm": 0.2294921875, "learning_rate": 0.0003706975050081952, "loss": 0.1535, "step": 2940 }, { "epoch": 0.7985923118570655, "grad_norm": 0.267578125, "learning_rate": 0.000370242214532872, "loss": 0.1379, "step": 2950 }, { "epoch": 0.8012994044396319, "grad_norm": 0.1708984375, "learning_rate": 0.0003697869240575487, "loss": 0.1404, "step": 2960 }, { "epoch": 0.8040064970221982, "grad_norm": 0.29296875, "learning_rate": 0.0003693316335822255, "loss": 0.138, "step": 2970 }, { "epoch": 0.8067135896047645, "grad_norm": 0.2431640625, "learning_rate": 0.00036887634310690223, "loss": 0.1347, "step": 2980 }, { "epoch": 0.8094206821873308, "grad_norm": 0.197265625, "learning_rate": 0.00036842105263157896, "loss": 0.1129, "step": 2990 }, { "epoch": 0.8121277747698972, "grad_norm": 0.16015625, "learning_rate": 0.0003679657621562557, "loss": 0.1192, "step": 3000 }, { "epoch": 0.8121277747698972, "eval_loss": 0.11679985374212265, "eval_runtime": 41.3323, "eval_samples_per_second": 12.097, "eval_steps_per_second": 0.774, "step": 3000 }, { "epoch": 0.8148348673524635, "grad_norm": 0.15234375, "learning_rate": 0.0003675104716809324, "loss": 0.1194, "step": 3010 }, { "epoch": 0.8175419599350298, "grad_norm": 0.162109375, "learning_rate": 0.0003670551812056092, "loss": 0.1116, "step": 3020 }, { "epoch": 0.8202490525175961, "grad_norm": 0.05859375, "learning_rate": 0.0003665998907302859, "loss": 0.1497, "step": 3030 }, { "epoch": 0.8229561451001625, "grad_norm": 0.23046875, "learning_rate": 0.00036614460025496265, "loss": 0.1161, "step": 3040 }, { "epoch": 0.8256632376827288, "grad_norm": 0.60546875, "learning_rate": 0.00036568930977963943, "loss": 0.1068, "step": 3050 }, { "epoch": 0.828370330265295, "grad_norm": 0.478515625, "learning_rate": 0.00036523401930431615, "loss": 0.1358, "step": 3060 }, { "epoch": 0.8310774228478613, "grad_norm": 0.1806640625, "learning_rate": 0.0003647787288289929, "loss": 0.1378, "step": 3070 }, { "epoch": 0.8337845154304278, "grad_norm": 0.419921875, "learning_rate": 0.00036432343835366966, "loss": 0.1386, "step": 3080 }, { "epoch": 0.836491608012994, "grad_norm": 0.2890625, "learning_rate": 0.0003638681478783464, "loss": 0.118, "step": 3090 }, { "epoch": 0.8391987005955603, "grad_norm": 0.423828125, "learning_rate": 0.00036341285740302317, "loss": 0.1363, "step": 3100 }, { "epoch": 0.8391987005955603, "eval_loss": 0.12045018374919891, "eval_runtime": 40.3514, "eval_samples_per_second": 12.391, "eval_steps_per_second": 0.793, "step": 3100 }, { "epoch": 0.8419057931781266, "grad_norm": 0.23046875, "learning_rate": 0.0003629575669276999, "loss": 0.1722, "step": 3110 }, { "epoch": 0.844612885760693, "grad_norm": 0.52734375, "learning_rate": 0.0003625022764523766, "loss": 0.1464, "step": 3120 }, { "epoch": 0.8473199783432593, "grad_norm": 0.3984375, "learning_rate": 0.0003620469859770534, "loss": 0.148, "step": 3130 }, { "epoch": 0.8500270709258256, "grad_norm": 0.333984375, "learning_rate": 0.0003615916955017301, "loss": 0.1589, "step": 3140 }, { "epoch": 0.852734163508392, "grad_norm": 0.34765625, "learning_rate": 0.00036113640502640686, "loss": 0.1927, "step": 3150 }, { "epoch": 0.8554412560909583, "grad_norm": 0.328125, "learning_rate": 0.0003606811145510836, "loss": 0.1357, "step": 3160 }, { "epoch": 0.8581483486735246, "grad_norm": 0.333984375, "learning_rate": 0.0003602258240757603, "loss": 0.1667, "step": 3170 }, { "epoch": 0.8608554412560909, "grad_norm": 0.375, "learning_rate": 0.0003597705336004371, "loss": 0.1331, "step": 3180 }, { "epoch": 0.8635625338386573, "grad_norm": 0.275390625, "learning_rate": 0.0003593152431251138, "loss": 0.1457, "step": 3190 }, { "epoch": 0.8662696264212236, "grad_norm": 0.2578125, "learning_rate": 0.0003588599526497906, "loss": 0.1111, "step": 3200 }, { "epoch": 0.8662696264212236, "eval_loss": 0.11618266999721527, "eval_runtime": 40.992, "eval_samples_per_second": 12.198, "eval_steps_per_second": 0.781, "step": 3200 }, { "epoch": 0.8689767190037899, "grad_norm": 0.13671875, "learning_rate": 0.0003584046621744673, "loss": 0.1315, "step": 3210 }, { "epoch": 0.8716838115863562, "grad_norm": 0.21875, "learning_rate": 0.00035794937169914405, "loss": 0.1317, "step": 3220 }, { "epoch": 0.8743909041689226, "grad_norm": 0.203125, "learning_rate": 0.00035749408122382083, "loss": 0.13, "step": 3230 }, { "epoch": 0.8770979967514889, "grad_norm": 0.119140625, "learning_rate": 0.00035703879074849756, "loss": 0.1528, "step": 3240 }, { "epoch": 0.8798050893340552, "grad_norm": 0.19140625, "learning_rate": 0.0003565835002731743, "loss": 0.1249, "step": 3250 }, { "epoch": 0.8825121819166215, "grad_norm": 0.2431640625, "learning_rate": 0.00035612820979785106, "loss": 0.1253, "step": 3260 }, { "epoch": 0.8852192744991879, "grad_norm": 0.1455078125, "learning_rate": 0.0003556729193225278, "loss": 0.1394, "step": 3270 }, { "epoch": 0.8879263670817542, "grad_norm": 0.2890625, "learning_rate": 0.0003552176288472045, "loss": 0.1617, "step": 3280 }, { "epoch": 0.8906334596643205, "grad_norm": 0.150390625, "learning_rate": 0.00035476233837188124, "loss": 0.1261, "step": 3290 }, { "epoch": 0.8933405522468868, "grad_norm": 0.2275390625, "learning_rate": 0.00035430704789655797, "loss": 0.1562, "step": 3300 }, { "epoch": 0.8933405522468868, "eval_loss": 0.11747061461210251, "eval_runtime": 40.263, "eval_samples_per_second": 12.418, "eval_steps_per_second": 0.795, "step": 3300 }, { "epoch": 0.8960476448294532, "grad_norm": 0.298828125, "learning_rate": 0.00035385175742123475, "loss": 0.1247, "step": 3310 }, { "epoch": 0.8987547374120195, "grad_norm": 0.357421875, "learning_rate": 0.0003533964669459115, "loss": 0.0968, "step": 3320 }, { "epoch": 0.9014618299945858, "grad_norm": 0.326171875, "learning_rate": 0.00035294117647058826, "loss": 0.1342, "step": 3330 }, { "epoch": 0.9041689225771521, "grad_norm": 0.09716796875, "learning_rate": 0.000352485885995265, "loss": 0.0842, "step": 3340 }, { "epoch": 0.9068760151597185, "grad_norm": 0.09521484375, "learning_rate": 0.0003520305955199417, "loss": 0.131, "step": 3350 }, { "epoch": 0.9095831077422848, "grad_norm": 0.404296875, "learning_rate": 0.0003515753050446185, "loss": 0.123, "step": 3360 }, { "epoch": 0.9122902003248511, "grad_norm": 0.388671875, "learning_rate": 0.0003511200145692952, "loss": 0.1427, "step": 3370 }, { "epoch": 0.9149972929074174, "grad_norm": 0.240234375, "learning_rate": 0.000350664724093972, "loss": 0.1566, "step": 3380 }, { "epoch": 0.9177043854899838, "grad_norm": 0.154296875, "learning_rate": 0.0003502094336186487, "loss": 0.0815, "step": 3390 }, { "epoch": 0.9204114780725501, "grad_norm": 0.453125, "learning_rate": 0.00034975414314332545, "loss": 0.1302, "step": 3400 }, { "epoch": 0.9204114780725501, "eval_loss": 0.1101117730140686, "eval_runtime": 40.2744, "eval_samples_per_second": 12.415, "eval_steps_per_second": 0.795, "step": 3400 }, { "epoch": 0.9231185706551164, "grad_norm": 0.4765625, "learning_rate": 0.00034929885266800223, "loss": 0.1385, "step": 3410 }, { "epoch": 0.9258256632376828, "grad_norm": 0.1884765625, "learning_rate": 0.0003488435621926789, "loss": 0.1312, "step": 3420 }, { "epoch": 0.9285327558202491, "grad_norm": 0.30859375, "learning_rate": 0.00034838827171735563, "loss": 0.1275, "step": 3430 }, { "epoch": 0.9312398484028154, "grad_norm": 0.2275390625, "learning_rate": 0.0003479329812420324, "loss": 0.1207, "step": 3440 }, { "epoch": 0.9339469409853817, "grad_norm": 0.22265625, "learning_rate": 0.00034747769076670914, "loss": 0.1342, "step": 3450 }, { "epoch": 0.9366540335679481, "grad_norm": 0.173828125, "learning_rate": 0.0003470224002913859, "loss": 0.1489, "step": 3460 }, { "epoch": 0.9393611261505144, "grad_norm": 0.29296875, "learning_rate": 0.00034656710981606265, "loss": 0.1143, "step": 3470 }, { "epoch": 0.9420682187330807, "grad_norm": 0.435546875, "learning_rate": 0.0003461118193407394, "loss": 0.1413, "step": 3480 }, { "epoch": 0.944775311315647, "grad_norm": 0.341796875, "learning_rate": 0.00034565652886541615, "loss": 0.1518, "step": 3490 }, { "epoch": 0.9474824038982134, "grad_norm": 0.07373046875, "learning_rate": 0.0003452012383900929, "loss": 0.1051, "step": 3500 }, { "epoch": 0.9474824038982134, "eval_loss": 0.11577916890382767, "eval_runtime": 40.6893, "eval_samples_per_second": 12.288, "eval_steps_per_second": 0.786, "step": 3500 }, { "epoch": 0.9501894964807797, "grad_norm": 0.578125, "learning_rate": 0.00034474594791476966, "loss": 0.092, "step": 3510 }, { "epoch": 0.952896589063346, "grad_norm": 0.1611328125, "learning_rate": 0.0003442906574394464, "loss": 0.1311, "step": 3520 }, { "epoch": 0.9556036816459123, "grad_norm": 0.251953125, "learning_rate": 0.0003438353669641231, "loss": 0.1054, "step": 3530 }, { "epoch": 0.9583107742284787, "grad_norm": 0.1279296875, "learning_rate": 0.0003433800764887999, "loss": 0.1335, "step": 3540 }, { "epoch": 0.961017866811045, "grad_norm": 0.1845703125, "learning_rate": 0.0003429247860134766, "loss": 0.1036, "step": 3550 }, { "epoch": 0.9637249593936112, "grad_norm": 0.1689453125, "learning_rate": 0.00034246949553815335, "loss": 0.1316, "step": 3560 }, { "epoch": 0.9664320519761775, "grad_norm": 0.23046875, "learning_rate": 0.0003420142050628301, "loss": 0.1546, "step": 3570 }, { "epoch": 0.969139144558744, "grad_norm": 0.384765625, "learning_rate": 0.0003415589145875068, "loss": 0.1433, "step": 3580 }, { "epoch": 0.9718462371413102, "grad_norm": 0.1962890625, "learning_rate": 0.0003411036241121836, "loss": 0.1327, "step": 3590 }, { "epoch": 0.9745533297238765, "grad_norm": 0.166015625, "learning_rate": 0.0003406483336368603, "loss": 0.1188, "step": 3600 }, { "epoch": 0.9745533297238765, "eval_loss": 0.11069635301828384, "eval_runtime": 42.4709, "eval_samples_per_second": 11.773, "eval_steps_per_second": 0.753, "step": 3600 }, { "epoch": 0.9772604223064428, "grad_norm": 0.44921875, "learning_rate": 0.00034019304316153703, "loss": 0.1486, "step": 3610 }, { "epoch": 0.9799675148890092, "grad_norm": 0.341796875, "learning_rate": 0.0003397377526862138, "loss": 0.1184, "step": 3620 }, { "epoch": 0.9826746074715755, "grad_norm": 0.10888671875, "learning_rate": 0.00033928246221089054, "loss": 0.1159, "step": 3630 }, { "epoch": 0.9853817000541418, "grad_norm": 0.1953125, "learning_rate": 0.0003388271717355673, "loss": 0.1541, "step": 3640 }, { "epoch": 0.9880887926367081, "grad_norm": 0.1064453125, "learning_rate": 0.00033837188126024405, "loss": 0.1418, "step": 3650 }, { "epoch": 0.9907958852192745, "grad_norm": 0.384765625, "learning_rate": 0.0003379165907849208, "loss": 0.1309, "step": 3660 }, { "epoch": 0.9935029778018408, "grad_norm": 0.138671875, "learning_rate": 0.00033746130030959756, "loss": 0.1176, "step": 3670 }, { "epoch": 0.9962100703844071, "grad_norm": 0.1728515625, "learning_rate": 0.0003370060098342743, "loss": 0.111, "step": 3680 }, { "epoch": 0.9989171629669734, "grad_norm": 0.1953125, "learning_rate": 0.00033655071935895106, "loss": 0.1059, "step": 3690 }, { "epoch": 1.0016242555495398, "grad_norm": 0.1494140625, "learning_rate": 0.00033609542888362774, "loss": 0.1452, "step": 3700 }, { "epoch": 1.0016242555495398, "eval_loss": 0.10907481610774994, "eval_runtime": 40.91, "eval_samples_per_second": 12.222, "eval_steps_per_second": 0.782, "step": 3700 }, { "epoch": 1.0043313481321061, "grad_norm": 0.154296875, "learning_rate": 0.00033564013840830446, "loss": 0.1131, "step": 3710 }, { "epoch": 1.0070384407146724, "grad_norm": 0.216796875, "learning_rate": 0.00033518484793298124, "loss": 0.1058, "step": 3720 }, { "epoch": 1.0097455332972387, "grad_norm": 0.130859375, "learning_rate": 0.00033472955745765797, "loss": 0.1041, "step": 3730 }, { "epoch": 1.012452625879805, "grad_norm": 0.251953125, "learning_rate": 0.00033427426698233475, "loss": 0.1353, "step": 3740 }, { "epoch": 1.0151597184623715, "grad_norm": 0.228515625, "learning_rate": 0.0003338189765070115, "loss": 0.1098, "step": 3750 }, { "epoch": 1.0178668110449378, "grad_norm": 0.169921875, "learning_rate": 0.0003333636860316882, "loss": 0.1257, "step": 3760 }, { "epoch": 1.020573903627504, "grad_norm": 0.1572265625, "learning_rate": 0.000332908395556365, "loss": 0.111, "step": 3770 }, { "epoch": 1.0232809962100704, "grad_norm": 0.10400390625, "learning_rate": 0.0003324531050810417, "loss": 0.1215, "step": 3780 }, { "epoch": 1.0259880887926367, "grad_norm": 0.296875, "learning_rate": 0.00033199781460571844, "loss": 0.1028, "step": 3790 }, { "epoch": 1.028695181375203, "grad_norm": 0.23046875, "learning_rate": 0.0003315425241303952, "loss": 0.1287, "step": 3800 }, { "epoch": 1.028695181375203, "eval_loss": 0.10907024890184402, "eval_runtime": 38.9399, "eval_samples_per_second": 12.84, "eval_steps_per_second": 0.822, "step": 3800 }, { "epoch": 1.0314022739577693, "grad_norm": 0.166015625, "learning_rate": 0.00033108723365507194, "loss": 0.1275, "step": 3810 }, { "epoch": 1.0341093665403356, "grad_norm": 0.1337890625, "learning_rate": 0.0003306319431797487, "loss": 0.121, "step": 3820 }, { "epoch": 1.036816459122902, "grad_norm": 0.248046875, "learning_rate": 0.00033017665270442545, "loss": 0.105, "step": 3830 }, { "epoch": 1.0395235517054684, "grad_norm": 0.1494140625, "learning_rate": 0.0003297213622291021, "loss": 0.1345, "step": 3840 }, { "epoch": 1.0422306442880347, "grad_norm": 0.16796875, "learning_rate": 0.0003292660717537789, "loss": 0.1052, "step": 3850 }, { "epoch": 1.044937736870601, "grad_norm": 0.26953125, "learning_rate": 0.00032881078127845563, "loss": 0.1181, "step": 3860 }, { "epoch": 1.0476448294531673, "grad_norm": 0.318359375, "learning_rate": 0.0003283554908031324, "loss": 0.1058, "step": 3870 }, { "epoch": 1.0503519220357336, "grad_norm": 0.25390625, "learning_rate": 0.00032790020032780914, "loss": 0.1234, "step": 3880 }, { "epoch": 1.0530590146182999, "grad_norm": 0.1865234375, "learning_rate": 0.00032744490985248587, "loss": 0.1079, "step": 3890 }, { "epoch": 1.0557661072008662, "grad_norm": 0.0986328125, "learning_rate": 0.00032698961937716265, "loss": 0.1325, "step": 3900 }, { "epoch": 1.0557661072008662, "eval_loss": 0.10832607746124268, "eval_runtime": 40.1373, "eval_samples_per_second": 12.457, "eval_steps_per_second": 0.797, "step": 3900 }, { "epoch": 1.0584731997834327, "grad_norm": 0.251953125, "learning_rate": 0.00032653432890183937, "loss": 0.122, "step": 3910 }, { "epoch": 1.061180292365999, "grad_norm": 0.1787109375, "learning_rate": 0.00032607903842651615, "loss": 0.0892, "step": 3920 }, { "epoch": 1.0638873849485653, "grad_norm": 0.185546875, "learning_rate": 0.0003256237479511929, "loss": 0.1102, "step": 3930 }, { "epoch": 1.0665944775311316, "grad_norm": 0.2021484375, "learning_rate": 0.0003251684574758696, "loss": 0.0968, "step": 3940 }, { "epoch": 1.0693015701136979, "grad_norm": 0.162109375, "learning_rate": 0.0003247131670005464, "loss": 0.1264, "step": 3950 }, { "epoch": 1.0720086626962642, "grad_norm": 0.255859375, "learning_rate": 0.0003242578765252231, "loss": 0.1226, "step": 3960 }, { "epoch": 1.0747157552788305, "grad_norm": 0.1572265625, "learning_rate": 0.00032380258604989984, "loss": 0.1388, "step": 3970 }, { "epoch": 1.077422847861397, "grad_norm": 0.361328125, "learning_rate": 0.0003233472955745766, "loss": 0.1343, "step": 3980 }, { "epoch": 1.0801299404439633, "grad_norm": 0.26171875, "learning_rate": 0.0003228920050992533, "loss": 0.1162, "step": 3990 }, { "epoch": 1.0828370330265296, "grad_norm": 0.1669921875, "learning_rate": 0.0003224367146239301, "loss": 0.1029, "step": 4000 }, { "epoch": 1.0828370330265296, "eval_loss": 0.10680222511291504, "eval_runtime": 39.7959, "eval_samples_per_second": 12.564, "eval_steps_per_second": 0.804, "step": 4000 }, { "epoch": 1.0855441256090959, "grad_norm": 0.0908203125, "learning_rate": 0.0003219814241486068, "loss": 0.0978, "step": 4010 }, { "epoch": 1.0882512181916622, "grad_norm": 0.26953125, "learning_rate": 0.0003215261336732835, "loss": 0.1081, "step": 4020 }, { "epoch": 1.0909583107742284, "grad_norm": 0.53125, "learning_rate": 0.0003210708431979603, "loss": 0.1214, "step": 4030 }, { "epoch": 1.0936654033567947, "grad_norm": 0.095703125, "learning_rate": 0.00032061555272263703, "loss": 0.113, "step": 4040 }, { "epoch": 1.096372495939361, "grad_norm": 0.177734375, "learning_rate": 0.0003201602622473138, "loss": 0.1047, "step": 4050 }, { "epoch": 1.0990795885219276, "grad_norm": 0.11376953125, "learning_rate": 0.00031970497177199054, "loss": 0.0707, "step": 4060 }, { "epoch": 1.1017866811044938, "grad_norm": 0.2353515625, "learning_rate": 0.00031924968129666727, "loss": 0.0969, "step": 4070 }, { "epoch": 1.1044937736870601, "grad_norm": 0.03564453125, "learning_rate": 0.00031879439082134405, "loss": 0.1102, "step": 4080 }, { "epoch": 1.1072008662696264, "grad_norm": 0.1689453125, "learning_rate": 0.0003183391003460208, "loss": 0.0815, "step": 4090 }, { "epoch": 1.1099079588521927, "grad_norm": 0.205078125, "learning_rate": 0.00031788380987069756, "loss": 0.1082, "step": 4100 }, { "epoch": 1.1099079588521927, "eval_loss": 0.10858253389596939, "eval_runtime": 36.8542, "eval_samples_per_second": 13.567, "eval_steps_per_second": 0.868, "step": 4100 }, { "epoch": 1.112615051434759, "grad_norm": 0.322265625, "learning_rate": 0.0003174285193953743, "loss": 0.1224, "step": 4110 }, { "epoch": 1.1153221440173253, "grad_norm": 0.1142578125, "learning_rate": 0.000316973228920051, "loss": 0.1146, "step": 4120 }, { "epoch": 1.1180292365998916, "grad_norm": 0.1552734375, "learning_rate": 0.00031651793844472774, "loss": 0.1133, "step": 4130 }, { "epoch": 1.1207363291824581, "grad_norm": 0.314453125, "learning_rate": 0.00031606264796940446, "loss": 0.1162, "step": 4140 }, { "epoch": 1.1234434217650244, "grad_norm": 0.1298828125, "learning_rate": 0.0003156073574940812, "loss": 0.1095, "step": 4150 }, { "epoch": 1.1261505143475907, "grad_norm": 0.1630859375, "learning_rate": 0.00031515206701875797, "loss": 0.1122, "step": 4160 }, { "epoch": 1.128857606930157, "grad_norm": 0.30078125, "learning_rate": 0.0003146967765434347, "loss": 0.1174, "step": 4170 }, { "epoch": 1.1315646995127233, "grad_norm": 0.130859375, "learning_rate": 0.0003142414860681115, "loss": 0.0926, "step": 4180 }, { "epoch": 1.1342717920952896, "grad_norm": 0.13671875, "learning_rate": 0.0003137861955927882, "loss": 0.1124, "step": 4190 }, { "epoch": 1.136978884677856, "grad_norm": 0.146484375, "learning_rate": 0.00031333090511746493, "loss": 0.1213, "step": 4200 }, { "epoch": 1.136978884677856, "eval_loss": 0.11143528670072556, "eval_runtime": 39.9574, "eval_samples_per_second": 12.513, "eval_steps_per_second": 0.801, "step": 4200 }, { "epoch": 1.1396859772604224, "grad_norm": 0.333984375, "learning_rate": 0.0003128756146421417, "loss": 0.1169, "step": 4210 }, { "epoch": 1.1423930698429887, "grad_norm": 0.2490234375, "learning_rate": 0.00031242032416681844, "loss": 0.1002, "step": 4220 }, { "epoch": 1.145100162425555, "grad_norm": 0.216796875, "learning_rate": 0.0003119650336914952, "loss": 0.0917, "step": 4230 }, { "epoch": 1.1478072550081213, "grad_norm": 0.189453125, "learning_rate": 0.00031150974321617194, "loss": 0.0661, "step": 4240 }, { "epoch": 1.1505143475906876, "grad_norm": 0.2470703125, "learning_rate": 0.00031105445274084867, "loss": 0.1264, "step": 4250 }, { "epoch": 1.153221440173254, "grad_norm": 0.267578125, "learning_rate": 0.00031059916226552545, "loss": 0.1003, "step": 4260 }, { "epoch": 1.1559285327558202, "grad_norm": 0.1533203125, "learning_rate": 0.0003101438717902021, "loss": 0.1087, "step": 4270 }, { "epoch": 1.1586356253383865, "grad_norm": 0.328125, "learning_rate": 0.0003096885813148789, "loss": 0.1082, "step": 4280 }, { "epoch": 1.1613427179209528, "grad_norm": 0.1591796875, "learning_rate": 0.00030923329083955563, "loss": 0.1173, "step": 4290 }, { "epoch": 1.1640498105035193, "grad_norm": 0.267578125, "learning_rate": 0.00030877800036423236, "loss": 0.1128, "step": 4300 }, { "epoch": 1.1640498105035193, "eval_loss": 0.11024381965398788, "eval_runtime": 40.8705, "eval_samples_per_second": 12.234, "eval_steps_per_second": 0.783, "step": 4300 }, { "epoch": 1.1667569030860856, "grad_norm": 0.388671875, "learning_rate": 0.00030832270988890914, "loss": 0.117, "step": 4310 }, { "epoch": 1.169463995668652, "grad_norm": 0.1328125, "learning_rate": 0.00030786741941358586, "loss": 0.0951, "step": 4320 }, { "epoch": 1.1721710882512182, "grad_norm": 0.15625, "learning_rate": 0.0003074121289382626, "loss": 0.0936, "step": 4330 }, { "epoch": 1.1748781808337845, "grad_norm": 0.2421875, "learning_rate": 0.00030695683846293937, "loss": 0.1443, "step": 4340 }, { "epoch": 1.1775852734163508, "grad_norm": 0.2119140625, "learning_rate": 0.0003065015479876161, "loss": 0.1025, "step": 4350 }, { "epoch": 1.180292365998917, "grad_norm": 0.154296875, "learning_rate": 0.0003060462575122929, "loss": 0.0865, "step": 4360 }, { "epoch": 1.1829994585814836, "grad_norm": 0.52734375, "learning_rate": 0.0003055909670369696, "loss": 0.1244, "step": 4370 }, { "epoch": 1.1857065511640499, "grad_norm": 0.369140625, "learning_rate": 0.00030513567656164633, "loss": 0.121, "step": 4380 }, { "epoch": 1.1884136437466162, "grad_norm": 0.232421875, "learning_rate": 0.0003046803860863231, "loss": 0.1108, "step": 4390 }, { "epoch": 1.1911207363291825, "grad_norm": 0.19921875, "learning_rate": 0.00030422509561099984, "loss": 0.1155, "step": 4400 }, { "epoch": 1.1911207363291825, "eval_loss": 0.1053057536482811, "eval_runtime": 40.0724, "eval_samples_per_second": 12.477, "eval_steps_per_second": 0.799, "step": 4400 }, { "epoch": 1.1938278289117488, "grad_norm": 0.1494140625, "learning_rate": 0.00030376980513567657, "loss": 0.0727, "step": 4410 }, { "epoch": 1.196534921494315, "grad_norm": 0.1494140625, "learning_rate": 0.0003033145146603533, "loss": 0.1064, "step": 4420 }, { "epoch": 1.1992420140768814, "grad_norm": 0.5625, "learning_rate": 0.00030285922418503, "loss": 0.1187, "step": 4430 }, { "epoch": 1.2019491066594479, "grad_norm": 0.1748046875, "learning_rate": 0.0003024039337097068, "loss": 0.1393, "step": 4440 }, { "epoch": 1.2046561992420142, "grad_norm": 0.357421875, "learning_rate": 0.0003019486432343835, "loss": 0.119, "step": 4450 }, { "epoch": 1.2073632918245805, "grad_norm": 0.302734375, "learning_rate": 0.0003014933527590603, "loss": 0.1157, "step": 4460 }, { "epoch": 1.2100703844071468, "grad_norm": 0.0576171875, "learning_rate": 0.00030103806228373703, "loss": 0.0932, "step": 4470 }, { "epoch": 1.212777476989713, "grad_norm": 0.2197265625, "learning_rate": 0.00030058277180841376, "loss": 0.1134, "step": 4480 }, { "epoch": 1.2154845695722793, "grad_norm": 0.1455078125, "learning_rate": 0.00030012748133309054, "loss": 0.1046, "step": 4490 }, { "epoch": 1.2181916621548456, "grad_norm": 0.2119140625, "learning_rate": 0.00029967219085776727, "loss": 0.0827, "step": 4500 }, { "epoch": 1.2181916621548456, "eval_loss": 0.10633888840675354, "eval_runtime": 41.3016, "eval_samples_per_second": 12.106, "eval_steps_per_second": 0.775, "step": 4500 }, { "epoch": 1.220898754737412, "grad_norm": 0.2421875, "learning_rate": 0.000299216900382444, "loss": 0.1375, "step": 4510 }, { "epoch": 1.2236058473199782, "grad_norm": 0.2421875, "learning_rate": 0.0002987616099071208, "loss": 0.1029, "step": 4520 }, { "epoch": 1.2263129399025448, "grad_norm": 0.2294921875, "learning_rate": 0.0002983063194317975, "loss": 0.1069, "step": 4530 }, { "epoch": 1.229020032485111, "grad_norm": 0.2099609375, "learning_rate": 0.0002978510289564743, "loss": 0.1126, "step": 4540 }, { "epoch": 1.2317271250676773, "grad_norm": 0.16796875, "learning_rate": 0.00029739573848115095, "loss": 0.1001, "step": 4550 }, { "epoch": 1.2344342176502436, "grad_norm": 0.2197265625, "learning_rate": 0.0002969404480058277, "loss": 0.111, "step": 4560 }, { "epoch": 1.23714131023281, "grad_norm": 0.173828125, "learning_rate": 0.00029648515753050446, "loss": 0.0922, "step": 4570 }, { "epoch": 1.2398484028153762, "grad_norm": 0.134765625, "learning_rate": 0.0002960298670551812, "loss": 0.0936, "step": 4580 }, { "epoch": 1.2425554953979425, "grad_norm": 0.318359375, "learning_rate": 0.00029557457657985797, "loss": 0.1092, "step": 4590 }, { "epoch": 1.245262587980509, "grad_norm": 0.12060546875, "learning_rate": 0.0002951192861045347, "loss": 0.1069, "step": 4600 }, { "epoch": 1.245262587980509, "eval_loss": 0.1066039651632309, "eval_runtime": 39.9085, "eval_samples_per_second": 12.529, "eval_steps_per_second": 0.802, "step": 4600 }, { "epoch": 1.2479696805630753, "grad_norm": 0.2099609375, "learning_rate": 0.0002946639956292114, "loss": 0.0893, "step": 4610 }, { "epoch": 1.2506767731456416, "grad_norm": 0.1796875, "learning_rate": 0.0002942087051538882, "loss": 0.0966, "step": 4620 }, { "epoch": 1.253383865728208, "grad_norm": 0.2021484375, "learning_rate": 0.00029375341467856493, "loss": 0.1028, "step": 4630 }, { "epoch": 1.2560909583107742, "grad_norm": 0.125, "learning_rate": 0.0002932981242032417, "loss": 0.1147, "step": 4640 }, { "epoch": 1.2587980508933405, "grad_norm": 0.08056640625, "learning_rate": 0.00029284283372791844, "loss": 0.0878, "step": 4650 }, { "epoch": 1.2615051434759068, "grad_norm": 0.20703125, "learning_rate": 0.00029238754325259516, "loss": 0.1044, "step": 4660 }, { "epoch": 1.2642122360584733, "grad_norm": 0.1015625, "learning_rate": 0.00029193225277727194, "loss": 0.086, "step": 4670 }, { "epoch": 1.2669193286410394, "grad_norm": 0.310546875, "learning_rate": 0.00029147696230194867, "loss": 0.1091, "step": 4680 }, { "epoch": 1.269626421223606, "grad_norm": 0.2470703125, "learning_rate": 0.00029102167182662534, "loss": 0.1043, "step": 4690 }, { "epoch": 1.2723335138061722, "grad_norm": 0.2216796875, "learning_rate": 0.0002905663813513021, "loss": 0.0963, "step": 4700 }, { "epoch": 1.2723335138061722, "eval_loss": 0.10753273218870163, "eval_runtime": 39.8634, "eval_samples_per_second": 12.543, "eval_steps_per_second": 0.803, "step": 4700 }, { "epoch": 1.2750406063887385, "grad_norm": 0.3515625, "learning_rate": 0.00029011109087597885, "loss": 0.1258, "step": 4710 }, { "epoch": 1.2777476989713048, "grad_norm": 0.38671875, "learning_rate": 0.00028965580040065563, "loss": 0.1115, "step": 4720 }, { "epoch": 1.280454791553871, "grad_norm": 0.208984375, "learning_rate": 0.00028920050992533236, "loss": 0.1478, "step": 4730 }, { "epoch": 1.2831618841364374, "grad_norm": 0.244140625, "learning_rate": 0.0002887452194500091, "loss": 0.0923, "step": 4740 }, { "epoch": 1.2858689767190037, "grad_norm": 0.07861328125, "learning_rate": 0.00028828992897468586, "loss": 0.1091, "step": 4750 }, { "epoch": 1.2885760693015702, "grad_norm": 0.34375, "learning_rate": 0.0002878346384993626, "loss": 0.1184, "step": 4760 }, { "epoch": 1.2912831618841365, "grad_norm": 0.1005859375, "learning_rate": 0.00028737934802403937, "loss": 0.1019, "step": 4770 }, { "epoch": 1.2939902544667028, "grad_norm": 0.2177734375, "learning_rate": 0.0002869240575487161, "loss": 0.1096, "step": 4780 }, { "epoch": 1.296697347049269, "grad_norm": 0.287109375, "learning_rate": 0.0002864687670733928, "loss": 0.1325, "step": 4790 }, { "epoch": 1.2994044396318354, "grad_norm": 0.279296875, "learning_rate": 0.0002860134765980696, "loss": 0.1197, "step": 4800 }, { "epoch": 1.2994044396318354, "eval_loss": 0.1108219102025032, "eval_runtime": 39.9454, "eval_samples_per_second": 12.517, "eval_steps_per_second": 0.801, "step": 4800 }, { "epoch": 1.3021115322144017, "grad_norm": 0.11474609375, "learning_rate": 0.00028555818612274633, "loss": 0.0816, "step": 4810 }, { "epoch": 1.304818624796968, "grad_norm": 0.1630859375, "learning_rate": 0.0002851028956474231, "loss": 0.1248, "step": 4820 }, { "epoch": 1.3075257173795345, "grad_norm": 0.15234375, "learning_rate": 0.0002846476051720998, "loss": 0.1093, "step": 4830 }, { "epoch": 1.3102328099621008, "grad_norm": 0.33203125, "learning_rate": 0.0002841923146967765, "loss": 0.0979, "step": 4840 }, { "epoch": 1.312939902544667, "grad_norm": 0.29296875, "learning_rate": 0.0002837370242214533, "loss": 0.1133, "step": 4850 }, { "epoch": 1.3156469951272334, "grad_norm": 0.06640625, "learning_rate": 0.00028328173374613, "loss": 0.1378, "step": 4860 }, { "epoch": 1.3183540877097997, "grad_norm": 0.251953125, "learning_rate": 0.00028282644327080674, "loss": 0.078, "step": 4870 }, { "epoch": 1.321061180292366, "grad_norm": 0.296875, "learning_rate": 0.0002823711527954835, "loss": 0.1035, "step": 4880 }, { "epoch": 1.3237682728749323, "grad_norm": 0.21484375, "learning_rate": 0.00028191586232016025, "loss": 0.1333, "step": 4890 }, { "epoch": 1.3264753654574988, "grad_norm": 0.06591796875, "learning_rate": 0.00028146057184483703, "loss": 0.1611, "step": 4900 }, { "epoch": 1.3264753654574988, "eval_loss": 0.10931263118982315, "eval_runtime": 38.6741, "eval_samples_per_second": 12.929, "eval_steps_per_second": 0.827, "step": 4900 }, { "epoch": 1.3291824580400649, "grad_norm": 0.052978515625, "learning_rate": 0.00028100528136951376, "loss": 0.1067, "step": 4910 }, { "epoch": 1.3318895506226314, "grad_norm": 0.2099609375, "learning_rate": 0.0002805499908941905, "loss": 0.1037, "step": 4920 }, { "epoch": 1.3345966432051977, "grad_norm": 0.41796875, "learning_rate": 0.00028009470041886727, "loss": 0.1169, "step": 4930 }, { "epoch": 1.337303735787764, "grad_norm": 0.31640625, "learning_rate": 0.000279639409943544, "loss": 0.1288, "step": 4940 }, { "epoch": 1.3400108283703303, "grad_norm": 0.341796875, "learning_rate": 0.0002791841194682208, "loss": 0.1023, "step": 4950 }, { "epoch": 1.3427179209528965, "grad_norm": 0.087890625, "learning_rate": 0.0002787288289928975, "loss": 0.1069, "step": 4960 }, { "epoch": 1.3454250135354628, "grad_norm": 0.068359375, "learning_rate": 0.00027827353851757417, "loss": 0.0898, "step": 4970 }, { "epoch": 1.3481321061180291, "grad_norm": 0.1767578125, "learning_rate": 0.00027781824804225095, "loss": 0.1093, "step": 4980 }, { "epoch": 1.3508391987005957, "grad_norm": 0.17578125, "learning_rate": 0.0002773629575669277, "loss": 0.1053, "step": 4990 }, { "epoch": 1.353546291283162, "grad_norm": 0.10888671875, "learning_rate": 0.00027690766709160446, "loss": 0.1079, "step": 5000 }, { "epoch": 1.353546291283162, "eval_loss": 0.11018607765436172, "eval_runtime": 38.2972, "eval_samples_per_second": 13.056, "eval_steps_per_second": 0.836, "step": 5000 }, { "epoch": 1.3562533838657282, "grad_norm": 0.07763671875, "learning_rate": 0.0002764523766162812, "loss": 0.1059, "step": 5010 }, { "epoch": 1.3589604764482945, "grad_norm": 0.177734375, "learning_rate": 0.0002759970861409579, "loss": 0.0937, "step": 5020 }, { "epoch": 1.3616675690308608, "grad_norm": 0.48828125, "learning_rate": 0.0002755417956656347, "loss": 0.1174, "step": 5030 }, { "epoch": 1.3643746616134271, "grad_norm": 0.18359375, "learning_rate": 0.0002750865051903114, "loss": 0.0962, "step": 5040 }, { "epoch": 1.3670817541959934, "grad_norm": 0.1337890625, "learning_rate": 0.00027463121471498815, "loss": 0.0907, "step": 5050 }, { "epoch": 1.36978884677856, "grad_norm": 0.11279296875, "learning_rate": 0.00027417592423966493, "loss": 0.0971, "step": 5060 }, { "epoch": 1.372495939361126, "grad_norm": 0.65625, "learning_rate": 0.00027372063376434165, "loss": 0.1637, "step": 5070 }, { "epoch": 1.3752030319436925, "grad_norm": 0.07958984375, "learning_rate": 0.00027326534328901843, "loss": 0.109, "step": 5080 }, { "epoch": 1.3779101245262588, "grad_norm": 0.271484375, "learning_rate": 0.00027281005281369516, "loss": 0.1077, "step": 5090 }, { "epoch": 1.3806172171088251, "grad_norm": 0.15625, "learning_rate": 0.0002723547623383719, "loss": 0.1204, "step": 5100 }, { "epoch": 1.3806172171088251, "eval_loss": 0.10544871538877487, "eval_runtime": 40.1846, "eval_samples_per_second": 12.443, "eval_steps_per_second": 0.796, "step": 5100 }, { "epoch": 1.3833243096913914, "grad_norm": 0.54296875, "learning_rate": 0.00027189947186304867, "loss": 0.1269, "step": 5110 }, { "epoch": 1.3860314022739577, "grad_norm": 0.1533203125, "learning_rate": 0.00027144418138772534, "loss": 0.1047, "step": 5120 }, { "epoch": 1.3887384948565242, "grad_norm": 0.1650390625, "learning_rate": 0.0002709888909124021, "loss": 0.1127, "step": 5130 }, { "epoch": 1.3914455874390903, "grad_norm": 0.1435546875, "learning_rate": 0.00027053360043707885, "loss": 0.1084, "step": 5140 }, { "epoch": 1.3941526800216568, "grad_norm": 0.267578125, "learning_rate": 0.0002700783099617556, "loss": 0.1351, "step": 5150 }, { "epoch": 1.3968597726042231, "grad_norm": 0.11474609375, "learning_rate": 0.00026962301948643236, "loss": 0.0964, "step": 5160 }, { "epoch": 1.3995668651867894, "grad_norm": 0.3671875, "learning_rate": 0.0002691677290111091, "loss": 0.1052, "step": 5170 }, { "epoch": 1.4022739577693557, "grad_norm": 0.166015625, "learning_rate": 0.00026871243853578586, "loss": 0.1328, "step": 5180 }, { "epoch": 1.404981050351922, "grad_norm": 0.220703125, "learning_rate": 0.0002682571480604626, "loss": 0.1299, "step": 5190 }, { "epoch": 1.4076881429344883, "grad_norm": 0.17578125, "learning_rate": 0.0002678018575851393, "loss": 0.1099, "step": 5200 }, { "epoch": 1.4076881429344883, "eval_loss": 0.1059228852391243, "eval_runtime": 41.1799, "eval_samples_per_second": 12.142, "eval_steps_per_second": 0.777, "step": 5200 }, { "epoch": 1.4103952355170546, "grad_norm": 0.138671875, "learning_rate": 0.0002673465671098161, "loss": 0.1207, "step": 5210 }, { "epoch": 1.413102328099621, "grad_norm": 0.13671875, "learning_rate": 0.0002668912766344928, "loss": 0.1142, "step": 5220 }, { "epoch": 1.4158094206821874, "grad_norm": 0.1416015625, "learning_rate": 0.00026643598615916955, "loss": 0.0989, "step": 5230 }, { "epoch": 1.4185165132647537, "grad_norm": 0.08203125, "learning_rate": 0.00026598069568384633, "loss": 0.1097, "step": 5240 }, { "epoch": 1.42122360584732, "grad_norm": 0.1064453125, "learning_rate": 0.00026552540520852306, "loss": 0.1138, "step": 5250 }, { "epoch": 1.4239306984298863, "grad_norm": 0.31640625, "learning_rate": 0.0002650701147331998, "loss": 0.1002, "step": 5260 }, { "epoch": 1.4266377910124526, "grad_norm": 0.169921875, "learning_rate": 0.0002646148242578765, "loss": 0.0969, "step": 5270 }, { "epoch": 1.4293448835950189, "grad_norm": 0.166015625, "learning_rate": 0.00026415953378255324, "loss": 0.0996, "step": 5280 }, { "epoch": 1.4320519761775854, "grad_norm": 0.271484375, "learning_rate": 0.00026370424330723, "loss": 0.1037, "step": 5290 }, { "epoch": 1.4347590687601515, "grad_norm": 0.26171875, "learning_rate": 0.00026324895283190674, "loss": 0.1022, "step": 5300 }, { "epoch": 1.4347590687601515, "eval_loss": 0.10697787255048752, "eval_runtime": 38.2441, "eval_samples_per_second": 13.074, "eval_steps_per_second": 0.837, "step": 5300 }, { "epoch": 1.437466161342718, "grad_norm": 0.14453125, "learning_rate": 0.0002627936623565835, "loss": 0.0903, "step": 5310 }, { "epoch": 1.4401732539252843, "grad_norm": 0.1796875, "learning_rate": 0.00026233837188126025, "loss": 0.1437, "step": 5320 }, { "epoch": 1.4428803465078506, "grad_norm": 0.37109375, "learning_rate": 0.000261883081405937, "loss": 0.1228, "step": 5330 }, { "epoch": 1.4455874390904169, "grad_norm": 0.197265625, "learning_rate": 0.00026142779093061376, "loss": 0.1094, "step": 5340 }, { "epoch": 1.4482945316729832, "grad_norm": 0.1103515625, "learning_rate": 0.0002609725004552905, "loss": 0.0906, "step": 5350 }, { "epoch": 1.4510016242555495, "grad_norm": 0.15625, "learning_rate": 0.00026051720997996726, "loss": 0.1105, "step": 5360 }, { "epoch": 1.4537087168381158, "grad_norm": 0.0947265625, "learning_rate": 0.000260061919504644, "loss": 0.0987, "step": 5370 }, { "epoch": 1.4564158094206823, "grad_norm": 0.236328125, "learning_rate": 0.0002596066290293207, "loss": 0.0918, "step": 5380 }, { "epoch": 1.4591229020032486, "grad_norm": 0.193359375, "learning_rate": 0.0002591513385539975, "loss": 0.1197, "step": 5390 }, { "epoch": 1.4618299945858149, "grad_norm": 0.162109375, "learning_rate": 0.00025869604807867417, "loss": 0.1066, "step": 5400 }, { "epoch": 1.4618299945858149, "eval_loss": 0.1075124517083168, "eval_runtime": 41.1695, "eval_samples_per_second": 12.145, "eval_steps_per_second": 0.777, "step": 5400 }, { "epoch": 1.4645370871683812, "grad_norm": 0.0859375, "learning_rate": 0.0002582407576033509, "loss": 0.1141, "step": 5410 }, { "epoch": 1.4672441797509475, "grad_norm": 0.099609375, "learning_rate": 0.0002577854671280277, "loss": 0.0731, "step": 5420 }, { "epoch": 1.4699512723335137, "grad_norm": 0.181640625, "learning_rate": 0.0002573301766527044, "loss": 0.1053, "step": 5430 }, { "epoch": 1.47265836491608, "grad_norm": 0.1083984375, "learning_rate": 0.0002568748861773812, "loss": 0.099, "step": 5440 }, { "epoch": 1.4753654574986466, "grad_norm": 0.1455078125, "learning_rate": 0.0002564195957020579, "loss": 0.0835, "step": 5450 }, { "epoch": 1.4780725500812129, "grad_norm": 0.314453125, "learning_rate": 0.00025596430522673464, "loss": 0.1168, "step": 5460 }, { "epoch": 1.4807796426637791, "grad_norm": 0.1533203125, "learning_rate": 0.0002555090147514114, "loss": 0.1494, "step": 5470 }, { "epoch": 1.4834867352463454, "grad_norm": 0.322265625, "learning_rate": 0.00025505372427608815, "loss": 0.119, "step": 5480 }, { "epoch": 1.4861938278289117, "grad_norm": 0.087890625, "learning_rate": 0.0002545984338007649, "loss": 0.0996, "step": 5490 }, { "epoch": 1.488900920411478, "grad_norm": 0.197265625, "learning_rate": 0.00025414314332544165, "loss": 0.1094, "step": 5500 }, { "epoch": 1.488900920411478, "eval_loss": 0.10609368979930878, "eval_runtime": 39.3625, "eval_samples_per_second": 12.702, "eval_steps_per_second": 0.813, "step": 5500 }, { "epoch": 1.4916080129940443, "grad_norm": 0.1259765625, "learning_rate": 0.0002536878528501184, "loss": 0.1082, "step": 5510 }, { "epoch": 1.4943151055766108, "grad_norm": 0.2470703125, "learning_rate": 0.00025323256237479516, "loss": 0.0935, "step": 5520 }, { "epoch": 1.497022198159177, "grad_norm": 0.21875, "learning_rate": 0.0002527772718994719, "loss": 0.121, "step": 5530 }, { "epoch": 1.4997292907417434, "grad_norm": 0.45703125, "learning_rate": 0.0002523219814241486, "loss": 0.1192, "step": 5540 }, { "epoch": 1.5024363833243097, "grad_norm": 0.291015625, "learning_rate": 0.00025186669094882534, "loss": 0.1333, "step": 5550 }, { "epoch": 1.505143475906876, "grad_norm": 0.369140625, "learning_rate": 0.00025141140047350207, "loss": 0.106, "step": 5560 }, { "epoch": 1.5078505684894423, "grad_norm": 0.26171875, "learning_rate": 0.00025095610999817885, "loss": 0.11, "step": 5570 }, { "epoch": 1.5105576610720086, "grad_norm": 0.1279296875, "learning_rate": 0.0002505008195228556, "loss": 0.0981, "step": 5580 }, { "epoch": 1.5132647536545751, "grad_norm": 0.23046875, "learning_rate": 0.0002500455290475323, "loss": 0.1236, "step": 5590 }, { "epoch": 1.5159718462371412, "grad_norm": 0.146484375, "learning_rate": 0.0002495902385722091, "loss": 0.1046, "step": 5600 }, { "epoch": 1.5159718462371412, "eval_loss": 0.1053042858839035, "eval_runtime": 38.9662, "eval_samples_per_second": 12.832, "eval_steps_per_second": 0.821, "step": 5600 }, { "epoch": 1.5186789388197077, "grad_norm": 0.2294921875, "learning_rate": 0.0002491349480968858, "loss": 0.1308, "step": 5610 }, { "epoch": 1.5213860314022738, "grad_norm": 0.283203125, "learning_rate": 0.00024867965762156253, "loss": 0.1105, "step": 5620 }, { "epoch": 1.5240931239848403, "grad_norm": 0.296875, "learning_rate": 0.0002482243671462393, "loss": 0.0983, "step": 5630 }, { "epoch": 1.5268002165674066, "grad_norm": 0.1865234375, "learning_rate": 0.00024776907667091604, "loss": 0.1055, "step": 5640 }, { "epoch": 1.529507309149973, "grad_norm": 0.2333984375, "learning_rate": 0.0002473137861955928, "loss": 0.1007, "step": 5650 }, { "epoch": 1.5322144017325392, "grad_norm": 0.1875, "learning_rate": 0.00024685849572026955, "loss": 0.0937, "step": 5660 }, { "epoch": 1.5349214943151055, "grad_norm": 0.25, "learning_rate": 0.0002464032052449463, "loss": 0.1311, "step": 5670 }, { "epoch": 1.537628586897672, "grad_norm": 0.23046875, "learning_rate": 0.000245947914769623, "loss": 0.0994, "step": 5680 }, { "epoch": 1.540335679480238, "grad_norm": 0.1728515625, "learning_rate": 0.0002454926242942998, "loss": 0.132, "step": 5690 }, { "epoch": 1.5430427720628046, "grad_norm": 0.408203125, "learning_rate": 0.0002450373338189765, "loss": 0.0934, "step": 5700 }, { "epoch": 1.5430427720628046, "eval_loss": 0.10769069194793701, "eval_runtime": 41.1215, "eval_samples_per_second": 12.159, "eval_steps_per_second": 0.778, "step": 5700 }, { "epoch": 1.545749864645371, "grad_norm": 0.115234375, "learning_rate": 0.00024458204334365324, "loss": 0.0925, "step": 5710 }, { "epoch": 1.5484569572279372, "grad_norm": 0.06591796875, "learning_rate": 0.00024412675286833, "loss": 0.082, "step": 5720 }, { "epoch": 1.5511640498105035, "grad_norm": 0.111328125, "learning_rate": 0.00024367146239300674, "loss": 0.1103, "step": 5730 }, { "epoch": 1.5538711423930698, "grad_norm": 0.55859375, "learning_rate": 0.0002432161719176835, "loss": 0.1005, "step": 5740 }, { "epoch": 1.5565782349756363, "grad_norm": 0.19140625, "learning_rate": 0.00024276088144236022, "loss": 0.0869, "step": 5750 }, { "epoch": 1.5592853275582024, "grad_norm": 0.419921875, "learning_rate": 0.00024230559096703698, "loss": 0.1195, "step": 5760 }, { "epoch": 1.561992420140769, "grad_norm": 0.33984375, "learning_rate": 0.00024185030049171373, "loss": 0.1044, "step": 5770 }, { "epoch": 1.5646995127233352, "grad_norm": 0.25390625, "learning_rate": 0.00024139501001639048, "loss": 0.1176, "step": 5780 }, { "epoch": 1.5674066053059015, "grad_norm": 0.26953125, "learning_rate": 0.00024093971954106718, "loss": 0.1241, "step": 5790 }, { "epoch": 1.5701136978884678, "grad_norm": 0.255859375, "learning_rate": 0.00024048442906574394, "loss": 0.1287, "step": 5800 }, { "epoch": 1.5701136978884678, "eval_loss": 0.10391703993082047, "eval_runtime": 37.79, "eval_samples_per_second": 13.231, "eval_steps_per_second": 0.847, "step": 5800 }, { "epoch": 1.572820790471034, "grad_norm": 0.205078125, "learning_rate": 0.0002400291385904207, "loss": 0.0923, "step": 5810 }, { "epoch": 1.5755278830536006, "grad_norm": 0.171875, "learning_rate": 0.00023957384811509744, "loss": 0.1085, "step": 5820 }, { "epoch": 1.5782349756361667, "grad_norm": 0.326171875, "learning_rate": 0.0002391185576397742, "loss": 0.1305, "step": 5830 }, { "epoch": 1.5809420682187332, "grad_norm": 0.302734375, "learning_rate": 0.00023866326716445092, "loss": 0.1224, "step": 5840 }, { "epoch": 1.5836491608012992, "grad_norm": 0.169921875, "learning_rate": 0.00023820797668912768, "loss": 0.0935, "step": 5850 }, { "epoch": 1.5863562533838658, "grad_norm": 0.3984375, "learning_rate": 0.0002377526862138044, "loss": 0.1164, "step": 5860 }, { "epoch": 1.589063345966432, "grad_norm": 0.13671875, "learning_rate": 0.00023729739573848116, "loss": 0.0953, "step": 5870 }, { "epoch": 1.5917704385489984, "grad_norm": 0.294921875, "learning_rate": 0.00023684210526315788, "loss": 0.1077, "step": 5880 }, { "epoch": 1.5944775311315647, "grad_norm": 0.2080078125, "learning_rate": 0.00023638681478783464, "loss": 0.079, "step": 5890 }, { "epoch": 1.597184623714131, "grad_norm": 0.2197265625, "learning_rate": 0.0002359315243125114, "loss": 0.1028, "step": 5900 }, { "epoch": 1.597184623714131, "eval_loss": 0.10197267681360245, "eval_runtime": 41.7148, "eval_samples_per_second": 11.986, "eval_steps_per_second": 0.767, "step": 5900 }, { "epoch": 1.5998917162966975, "grad_norm": 0.2333984375, "learning_rate": 0.00023547623383718814, "loss": 0.1166, "step": 5910 }, { "epoch": 1.6025988088792635, "grad_norm": 0.12451171875, "learning_rate": 0.0002350209433618649, "loss": 0.1147, "step": 5920 }, { "epoch": 1.60530590146183, "grad_norm": 0.1318359375, "learning_rate": 0.0002345656528865416, "loss": 0.1079, "step": 5930 }, { "epoch": 1.6080129940443963, "grad_norm": 0.09375, "learning_rate": 0.00023411036241121835, "loss": 0.1193, "step": 5940 }, { "epoch": 1.6107200866269626, "grad_norm": 0.2138671875, "learning_rate": 0.0002336550719358951, "loss": 0.1086, "step": 5950 }, { "epoch": 1.613427179209529, "grad_norm": 0.326171875, "learning_rate": 0.00023319978146057186, "loss": 0.1169, "step": 5960 }, { "epoch": 1.6161342717920952, "grad_norm": 0.173828125, "learning_rate": 0.00023274449098524859, "loss": 0.1071, "step": 5970 }, { "epoch": 1.6188413643746618, "grad_norm": 0.345703125, "learning_rate": 0.00023228920050992534, "loss": 0.1136, "step": 5980 }, { "epoch": 1.6215484569572278, "grad_norm": 0.0791015625, "learning_rate": 0.0002318339100346021, "loss": 0.0873, "step": 5990 }, { "epoch": 1.6242555495397943, "grad_norm": 0.2451171875, "learning_rate": 0.00023137861955927882, "loss": 0.098, "step": 6000 }, { "epoch": 1.6242555495397943, "eval_loss": 0.10421106219291687, "eval_runtime": 40.4533, "eval_samples_per_second": 12.36, "eval_steps_per_second": 0.791, "step": 6000 }, { "epoch": 1.6269626421223606, "grad_norm": 0.11572265625, "learning_rate": 0.00023092332908395557, "loss": 0.0992, "step": 6010 }, { "epoch": 1.629669734704927, "grad_norm": 0.1181640625, "learning_rate": 0.0002304680386086323, "loss": 0.0827, "step": 6020 }, { "epoch": 1.6323768272874932, "grad_norm": 0.216796875, "learning_rate": 0.00023001274813330905, "loss": 0.1242, "step": 6030 }, { "epoch": 1.6350839198700595, "grad_norm": 0.0849609375, "learning_rate": 0.0002295574576579858, "loss": 0.0938, "step": 6040 }, { "epoch": 1.637791012452626, "grad_norm": 0.1435546875, "learning_rate": 0.00022910216718266256, "loss": 0.1278, "step": 6050 }, { "epoch": 1.6404981050351921, "grad_norm": 0.2001953125, "learning_rate": 0.0002286468767073393, "loss": 0.093, "step": 6060 }, { "epoch": 1.6432051976177586, "grad_norm": 0.10595703125, "learning_rate": 0.000228191586232016, "loss": 0.0862, "step": 6070 }, { "epoch": 1.6459122902003247, "grad_norm": 0.1904296875, "learning_rate": 0.00022773629575669277, "loss": 0.0885, "step": 6080 }, { "epoch": 1.6486193827828912, "grad_norm": 0.3046875, "learning_rate": 0.00022728100528136952, "loss": 0.1219, "step": 6090 }, { "epoch": 1.6513264753654575, "grad_norm": 0.453125, "learning_rate": 0.00022682571480604627, "loss": 0.1073, "step": 6100 }, { "epoch": 1.6513264753654575, "eval_loss": 0.10543978214263916, "eval_runtime": 39.6782, "eval_samples_per_second": 12.601, "eval_steps_per_second": 0.806, "step": 6100 }, { "epoch": 1.6540335679480238, "grad_norm": 0.240234375, "learning_rate": 0.000226370424330723, "loss": 0.1117, "step": 6110 }, { "epoch": 1.65674066053059, "grad_norm": 0.43359375, "learning_rate": 0.00022591513385539975, "loss": 0.1086, "step": 6120 }, { "epoch": 1.6594477531131564, "grad_norm": 0.25, "learning_rate": 0.0002254598433800765, "loss": 0.0925, "step": 6130 }, { "epoch": 1.662154845695723, "grad_norm": 0.1064453125, "learning_rate": 0.00022500455290475323, "loss": 0.0951, "step": 6140 }, { "epoch": 1.664861938278289, "grad_norm": 0.162109375, "learning_rate": 0.00022454926242942996, "loss": 0.0974, "step": 6150 }, { "epoch": 1.6675690308608555, "grad_norm": 0.1943359375, "learning_rate": 0.00022409397195410671, "loss": 0.1169, "step": 6160 }, { "epoch": 1.6702761234434218, "grad_norm": 0.140625, "learning_rate": 0.00022363868147878347, "loss": 0.1166, "step": 6170 }, { "epoch": 1.672983216025988, "grad_norm": 0.236328125, "learning_rate": 0.00022318339100346022, "loss": 0.0993, "step": 6180 }, { "epoch": 1.6756903086085544, "grad_norm": 0.54296875, "learning_rate": 0.00022272810052813698, "loss": 0.1152, "step": 6190 }, { "epoch": 1.6783974011911207, "grad_norm": 0.158203125, "learning_rate": 0.0002222728100528137, "loss": 0.0826, "step": 6200 }, { "epoch": 1.6783974011911207, "eval_loss": 0.10408873856067657, "eval_runtime": 42.4934, "eval_samples_per_second": 11.767, "eval_steps_per_second": 0.753, "step": 6200 }, { "epoch": 1.6811044937736872, "grad_norm": 0.099609375, "learning_rate": 0.00022181751957749043, "loss": 0.1053, "step": 6210 }, { "epoch": 1.6838115863562533, "grad_norm": 0.1015625, "learning_rate": 0.00022136222910216718, "loss": 0.0822, "step": 6220 }, { "epoch": 1.6865186789388198, "grad_norm": 0.48046875, "learning_rate": 0.00022090693862684394, "loss": 0.11, "step": 6230 }, { "epoch": 1.6892257715213859, "grad_norm": 0.205078125, "learning_rate": 0.00022045164815152066, "loss": 0.1102, "step": 6240 }, { "epoch": 1.6919328641039524, "grad_norm": 0.291015625, "learning_rate": 0.00021999635767619742, "loss": 0.0819, "step": 6250 }, { "epoch": 1.6946399566865187, "grad_norm": 0.1240234375, "learning_rate": 0.00021954106720087417, "loss": 0.1069, "step": 6260 }, { "epoch": 1.697347049269085, "grad_norm": 0.35546875, "learning_rate": 0.00021908577672555092, "loss": 0.1166, "step": 6270 }, { "epoch": 1.7000541418516515, "grad_norm": 0.224609375, "learning_rate": 0.00021863048625022765, "loss": 0.1043, "step": 6280 }, { "epoch": 1.7027612344342176, "grad_norm": 0.07470703125, "learning_rate": 0.00021817519577490438, "loss": 0.1131, "step": 6290 }, { "epoch": 1.705468327016784, "grad_norm": 0.23046875, "learning_rate": 0.00021771990529958113, "loss": 0.1017, "step": 6300 }, { "epoch": 1.705468327016784, "eval_loss": 0.10435427725315094, "eval_runtime": 40.3493, "eval_samples_per_second": 12.392, "eval_steps_per_second": 0.793, "step": 6300 }, { "epoch": 1.7081754195993502, "grad_norm": 0.2021484375, "learning_rate": 0.00021726461482425788, "loss": 0.103, "step": 6310 }, { "epoch": 1.7108825121819167, "grad_norm": 0.1884765625, "learning_rate": 0.00021680932434893464, "loss": 0.1322, "step": 6320 }, { "epoch": 1.713589604764483, "grad_norm": 0.1953125, "learning_rate": 0.00021635403387361136, "loss": 0.116, "step": 6330 }, { "epoch": 1.7162966973470493, "grad_norm": 0.482421875, "learning_rate": 0.00021589874339828812, "loss": 0.0846, "step": 6340 }, { "epoch": 1.7190037899296156, "grad_norm": 0.353515625, "learning_rate": 0.00021544345292296484, "loss": 0.0961, "step": 6350 }, { "epoch": 1.7217108825121819, "grad_norm": 0.228515625, "learning_rate": 0.0002149881624476416, "loss": 0.1091, "step": 6360 }, { "epoch": 1.7244179750947484, "grad_norm": 0.18359375, "learning_rate": 0.00021453287197231835, "loss": 0.1077, "step": 6370 }, { "epoch": 1.7271250676773144, "grad_norm": 0.1357421875, "learning_rate": 0.00021407758149699508, "loss": 0.1074, "step": 6380 }, { "epoch": 1.729832160259881, "grad_norm": 0.171875, "learning_rate": 0.00021362229102167183, "loss": 0.0876, "step": 6390 }, { "epoch": 1.7325392528424473, "grad_norm": 0.2177734375, "learning_rate": 0.00021316700054634858, "loss": 0.1018, "step": 6400 }, { "epoch": 1.7325392528424473, "eval_loss": 0.10634534060955048, "eval_runtime": 38.4804, "eval_samples_per_second": 12.994, "eval_steps_per_second": 0.832, "step": 6400 }, { "epoch": 1.7352463454250135, "grad_norm": 0.130859375, "learning_rate": 0.00021271171007102534, "loss": 0.1273, "step": 6410 }, { "epoch": 1.7379534380075798, "grad_norm": 0.55078125, "learning_rate": 0.00021225641959570204, "loss": 0.1279, "step": 6420 }, { "epoch": 1.7406605305901461, "grad_norm": 0.1865234375, "learning_rate": 0.0002118011291203788, "loss": 0.0721, "step": 6430 }, { "epoch": 1.7433676231727127, "grad_norm": 0.06396484375, "learning_rate": 0.00021134583864505554, "loss": 0.0887, "step": 6440 }, { "epoch": 1.7460747157552787, "grad_norm": 0.283203125, "learning_rate": 0.0002108905481697323, "loss": 0.097, "step": 6450 }, { "epoch": 1.7487818083378452, "grad_norm": 0.287109375, "learning_rate": 0.00021043525769440905, "loss": 0.0678, "step": 6460 }, { "epoch": 1.7514889009204113, "grad_norm": 0.1455078125, "learning_rate": 0.00020997996721908578, "loss": 0.1127, "step": 6470 }, { "epoch": 1.7541959935029778, "grad_norm": 0.1767578125, "learning_rate": 0.00020952467674376253, "loss": 0.0835, "step": 6480 }, { "epoch": 1.7569030860855441, "grad_norm": 0.1396484375, "learning_rate": 0.00020906938626843926, "loss": 0.11, "step": 6490 }, { "epoch": 1.7596101786681104, "grad_norm": 0.1728515625, "learning_rate": 0.000208614095793116, "loss": 0.0802, "step": 6500 }, { "epoch": 1.7596101786681104, "eval_loss": 0.10547506809234619, "eval_runtime": 37.3912, "eval_samples_per_second": 13.372, "eval_steps_per_second": 0.856, "step": 6500 }, { "epoch": 1.7623172712506767, "grad_norm": 0.1904296875, "learning_rate": 0.00020815880531779274, "loss": 0.108, "step": 6510 }, { "epoch": 1.765024363833243, "grad_norm": 0.1904296875, "learning_rate": 0.0002077035148424695, "loss": 0.118, "step": 6520 }, { "epoch": 1.7677314564158095, "grad_norm": 0.1005859375, "learning_rate": 0.00020724822436714625, "loss": 0.1029, "step": 6530 }, { "epoch": 1.7704385489983756, "grad_norm": 0.09423828125, "learning_rate": 0.000206792933891823, "loss": 0.0959, "step": 6540 }, { "epoch": 1.7731456415809421, "grad_norm": 0.4375, "learning_rate": 0.00020633764341649975, "loss": 0.1183, "step": 6550 }, { "epoch": 1.7758527341635084, "grad_norm": 0.1328125, "learning_rate": 0.00020588235294117645, "loss": 0.082, "step": 6560 }, { "epoch": 1.7785598267460747, "grad_norm": 0.279296875, "learning_rate": 0.0002054270624658532, "loss": 0.107, "step": 6570 }, { "epoch": 1.781266919328641, "grad_norm": 0.3671875, "learning_rate": 0.00020497177199052996, "loss": 0.1266, "step": 6580 }, { "epoch": 1.7839740119112073, "grad_norm": 0.1259765625, "learning_rate": 0.0002045164815152067, "loss": 0.1184, "step": 6590 }, { "epoch": 1.7866811044937738, "grad_norm": 0.1640625, "learning_rate": 0.00020406119103988344, "loss": 0.1048, "step": 6600 }, { "epoch": 1.7866811044937738, "eval_loss": 0.10778705775737762, "eval_runtime": 41.271, "eval_samples_per_second": 12.115, "eval_steps_per_second": 0.775, "step": 6600 }, { "epoch": 1.78938819707634, "grad_norm": 0.080078125, "learning_rate": 0.0002036059005645602, "loss": 0.0989, "step": 6610 }, { "epoch": 1.7920952896589064, "grad_norm": 0.2890625, "learning_rate": 0.00020315061008923695, "loss": 0.0958, "step": 6620 }, { "epoch": 1.7948023822414727, "grad_norm": 0.5234375, "learning_rate": 0.00020269531961391367, "loss": 0.1318, "step": 6630 }, { "epoch": 1.797509474824039, "grad_norm": 0.134765625, "learning_rate": 0.00020224002913859043, "loss": 0.0899, "step": 6640 }, { "epoch": 1.8002165674066053, "grad_norm": 0.1728515625, "learning_rate": 0.00020178473866326715, "loss": 0.1421, "step": 6650 }, { "epoch": 1.8029236599891716, "grad_norm": 0.259765625, "learning_rate": 0.0002013294481879439, "loss": 0.0998, "step": 6660 }, { "epoch": 1.805630752571738, "grad_norm": 0.1787109375, "learning_rate": 0.00020087415771262066, "loss": 0.1019, "step": 6670 }, { "epoch": 1.8083378451543042, "grad_norm": 0.07470703125, "learning_rate": 0.00020041886723729741, "loss": 0.0975, "step": 6680 }, { "epoch": 1.8110449377368707, "grad_norm": 0.322265625, "learning_rate": 0.00019996357676197414, "loss": 0.1148, "step": 6690 }, { "epoch": 1.8137520303194368, "grad_norm": 0.1396484375, "learning_rate": 0.00019950828628665087, "loss": 0.1015, "step": 6700 }, { "epoch": 1.8137520303194368, "eval_loss": 0.10642816126346588, "eval_runtime": 40.4468, "eval_samples_per_second": 12.362, "eval_steps_per_second": 0.791, "step": 6700 }, { "epoch": 1.8164591229020033, "grad_norm": 0.1611328125, "learning_rate": 0.00019905299581132762, "loss": 0.0996, "step": 6710 }, { "epoch": 1.8191662154845696, "grad_norm": 0.3359375, "learning_rate": 0.00019859770533600437, "loss": 0.1156, "step": 6720 }, { "epoch": 1.8218733080671359, "grad_norm": 0.11572265625, "learning_rate": 0.00019814241486068113, "loss": 0.1069, "step": 6730 }, { "epoch": 1.8245804006497022, "grad_norm": 0.140625, "learning_rate": 0.00019768712438535786, "loss": 0.1159, "step": 6740 }, { "epoch": 1.8272874932322685, "grad_norm": 0.61328125, "learning_rate": 0.0001972318339100346, "loss": 0.0914, "step": 6750 }, { "epoch": 1.829994585814835, "grad_norm": 0.0703125, "learning_rate": 0.00019677654343471136, "loss": 0.0799, "step": 6760 }, { "epoch": 1.832701678397401, "grad_norm": 0.28125, "learning_rate": 0.00019632125295938812, "loss": 0.1058, "step": 6770 }, { "epoch": 1.8354087709799676, "grad_norm": 0.33984375, "learning_rate": 0.00019586596248406482, "loss": 0.0802, "step": 6780 }, { "epoch": 1.8381158635625339, "grad_norm": 0.087890625, "learning_rate": 0.00019541067200874157, "loss": 0.0755, "step": 6790 }, { "epoch": 1.8408229561451002, "grad_norm": 0.08447265625, "learning_rate": 0.00019495538153341832, "loss": 0.1285, "step": 6800 }, { "epoch": 1.8408229561451002, "eval_loss": 0.1069154143333435, "eval_runtime": 40.8303, "eval_samples_per_second": 12.246, "eval_steps_per_second": 0.784, "step": 6800 }, { "epoch": 1.8435300487276665, "grad_norm": 0.15625, "learning_rate": 0.00019450009105809508, "loss": 0.0844, "step": 6810 }, { "epoch": 1.8462371413102328, "grad_norm": 0.111328125, "learning_rate": 0.00019404480058277183, "loss": 0.1021, "step": 6820 }, { "epoch": 1.8489442338927993, "grad_norm": 0.10791015625, "learning_rate": 0.00019358951010744856, "loss": 0.1028, "step": 6830 }, { "epoch": 1.8516513264753653, "grad_norm": 0.25390625, "learning_rate": 0.0001931342196321253, "loss": 0.1174, "step": 6840 }, { "epoch": 1.8543584190579319, "grad_norm": 0.2099609375, "learning_rate": 0.00019267892915680204, "loss": 0.1151, "step": 6850 }, { "epoch": 1.8570655116404982, "grad_norm": 0.17578125, "learning_rate": 0.0001922236386814788, "loss": 0.1221, "step": 6860 }, { "epoch": 1.8597726042230645, "grad_norm": 0.318359375, "learning_rate": 0.00019176834820615552, "loss": 0.1034, "step": 6870 }, { "epoch": 1.8624796968056307, "grad_norm": 0.09326171875, "learning_rate": 0.00019131305773083227, "loss": 0.1071, "step": 6880 }, { "epoch": 1.865186789388197, "grad_norm": 0.216796875, "learning_rate": 0.00019085776725550902, "loss": 0.0957, "step": 6890 }, { "epoch": 1.8678938819707636, "grad_norm": 0.1982421875, "learning_rate": 0.00019040247678018578, "loss": 0.1147, "step": 6900 }, { "epoch": 1.8678938819707636, "eval_loss": 0.10640809684991837, "eval_runtime": 37.6638, "eval_samples_per_second": 13.275, "eval_steps_per_second": 0.85, "step": 6900 }, { "epoch": 1.8706009745533296, "grad_norm": 0.0810546875, "learning_rate": 0.00018994718630486253, "loss": 0.1084, "step": 6910 }, { "epoch": 1.8733080671358961, "grad_norm": 0.259765625, "learning_rate": 0.00018949189582953923, "loss": 0.0882, "step": 6920 }, { "epoch": 1.8760151597184622, "grad_norm": 0.08837890625, "learning_rate": 0.00018903660535421598, "loss": 0.1141, "step": 6930 }, { "epoch": 1.8787222523010287, "grad_norm": 0.24609375, "learning_rate": 0.00018858131487889274, "loss": 0.0978, "step": 6940 }, { "epoch": 1.881429344883595, "grad_norm": 0.333984375, "learning_rate": 0.0001881260244035695, "loss": 0.1493, "step": 6950 }, { "epoch": 1.8841364374661613, "grad_norm": 0.2099609375, "learning_rate": 0.00018767073392824622, "loss": 0.1253, "step": 6960 }, { "epoch": 1.8868435300487276, "grad_norm": 0.291015625, "learning_rate": 0.00018721544345292297, "loss": 0.1289, "step": 6970 }, { "epoch": 1.889550622631294, "grad_norm": 0.216796875, "learning_rate": 0.00018676015297759973, "loss": 0.0927, "step": 6980 }, { "epoch": 1.8922577152138604, "grad_norm": 0.056396484375, "learning_rate": 0.00018630486250227645, "loss": 0.093, "step": 6990 }, { "epoch": 1.8949648077964265, "grad_norm": 0.046875, "learning_rate": 0.0001858495720269532, "loss": 0.0881, "step": 7000 }, { "epoch": 1.8949648077964265, "eval_loss": 0.10610378533601761, "eval_runtime": 38.7918, "eval_samples_per_second": 12.889, "eval_steps_per_second": 0.825, "step": 7000 }, { "epoch": 1.897671900378993, "grad_norm": 0.05419921875, "learning_rate": 0.00018539428155162993, "loss": 0.0789, "step": 7010 }, { "epoch": 1.9003789929615593, "grad_norm": 0.13671875, "learning_rate": 0.00018493899107630669, "loss": 0.0839, "step": 7020 }, { "epoch": 1.9030860855441256, "grad_norm": 0.1796875, "learning_rate": 0.00018448370060098344, "loss": 0.1061, "step": 7030 }, { "epoch": 1.905793178126692, "grad_norm": 0.07470703125, "learning_rate": 0.0001840284101256602, "loss": 0.1098, "step": 7040 }, { "epoch": 1.9085002707092582, "grad_norm": 0.12109375, "learning_rate": 0.00018357311965033692, "loss": 0.108, "step": 7050 }, { "epoch": 1.9112073632918247, "grad_norm": 0.26953125, "learning_rate": 0.00018311782917501365, "loss": 0.1203, "step": 7060 }, { "epoch": 1.9139144558743908, "grad_norm": 0.044921875, "learning_rate": 0.0001826625386996904, "loss": 0.0852, "step": 7070 }, { "epoch": 1.9166215484569573, "grad_norm": 0.51171875, "learning_rate": 0.00018220724822436715, "loss": 0.1248, "step": 7080 }, { "epoch": 1.9193286410395234, "grad_norm": 0.314453125, "learning_rate": 0.0001817519577490439, "loss": 0.1094, "step": 7090 }, { "epoch": 1.92203573362209, "grad_norm": 0.08740234375, "learning_rate": 0.00018129666727372063, "loss": 0.0963, "step": 7100 }, { "epoch": 1.92203573362209, "eval_loss": 0.1067114919424057, "eval_runtime": 41.6638, "eval_samples_per_second": 12.001, "eval_steps_per_second": 0.768, "step": 7100 }, { "epoch": 1.9247428262046562, "grad_norm": 0.376953125, "learning_rate": 0.0001808413767983974, "loss": 0.1009, "step": 7110 }, { "epoch": 1.9274499187872225, "grad_norm": 0.333984375, "learning_rate": 0.00018038608632307414, "loss": 0.094, "step": 7120 }, { "epoch": 1.930157011369789, "grad_norm": 0.06640625, "learning_rate": 0.00017993079584775087, "loss": 0.1064, "step": 7130 }, { "epoch": 1.932864103952355, "grad_norm": 0.259765625, "learning_rate": 0.0001794755053724276, "loss": 0.1009, "step": 7140 }, { "epoch": 1.9355711965349216, "grad_norm": 0.115234375, "learning_rate": 0.00017902021489710435, "loss": 0.0943, "step": 7150 }, { "epoch": 1.9382782891174877, "grad_norm": 0.2236328125, "learning_rate": 0.0001785649244217811, "loss": 0.0998, "step": 7160 }, { "epoch": 1.9409853817000542, "grad_norm": 0.322265625, "learning_rate": 0.00017810963394645785, "loss": 0.0936, "step": 7170 }, { "epoch": 1.9436924742826205, "grad_norm": 0.2021484375, "learning_rate": 0.0001776543434711346, "loss": 0.1144, "step": 7180 }, { "epoch": 1.9463995668651868, "grad_norm": 0.162109375, "learning_rate": 0.00017719905299581133, "loss": 0.118, "step": 7190 }, { "epoch": 1.949106659447753, "grad_norm": 0.2734375, "learning_rate": 0.00017674376252048806, "loss": 0.0792, "step": 7200 }, { "epoch": 1.949106659447753, "eval_loss": 0.10313951969146729, "eval_runtime": 40.7925, "eval_samples_per_second": 12.257, "eval_steps_per_second": 0.784, "step": 7200 }, { "epoch": 1.9518137520303194, "grad_norm": 0.3515625, "learning_rate": 0.00017628847204516481, "loss": 0.0956, "step": 7210 }, { "epoch": 1.9545208446128859, "grad_norm": 0.365234375, "learning_rate": 0.00017583318156984157, "loss": 0.1285, "step": 7220 }, { "epoch": 1.957227937195452, "grad_norm": 0.353515625, "learning_rate": 0.0001753778910945183, "loss": 0.102, "step": 7230 }, { "epoch": 1.9599350297780185, "grad_norm": 0.28515625, "learning_rate": 0.00017492260061919505, "loss": 0.1112, "step": 7240 }, { "epoch": 1.9626421223605848, "grad_norm": 0.1943359375, "learning_rate": 0.0001744673101438718, "loss": 0.0934, "step": 7250 }, { "epoch": 1.965349214943151, "grad_norm": 0.203125, "learning_rate": 0.00017401201966854856, "loss": 0.082, "step": 7260 }, { "epoch": 1.9680563075257174, "grad_norm": 0.1787109375, "learning_rate": 0.00017355672919322528, "loss": 0.1187, "step": 7270 }, { "epoch": 1.9707634001082837, "grad_norm": 0.1513671875, "learning_rate": 0.000173101438717902, "loss": 0.0982, "step": 7280 }, { "epoch": 1.9734704926908502, "grad_norm": 0.380859375, "learning_rate": 0.00017264614824257876, "loss": 0.1336, "step": 7290 }, { "epoch": 1.9761775852734162, "grad_norm": 0.2080078125, "learning_rate": 0.00017219085776725552, "loss": 0.0846, "step": 7300 }, { "epoch": 1.9761775852734162, "eval_loss": 0.10161355882883072, "eval_runtime": 40.0399, "eval_samples_per_second": 12.488, "eval_steps_per_second": 0.799, "step": 7300 }, { "epoch": 1.9788846778559828, "grad_norm": 0.376953125, "learning_rate": 0.00017173556729193227, "loss": 0.0984, "step": 7310 }, { "epoch": 1.9815917704385488, "grad_norm": 0.12109375, "learning_rate": 0.000171280276816609, "loss": 0.0937, "step": 7320 }, { "epoch": 1.9842988630211154, "grad_norm": 0.251953125, "learning_rate": 0.00017082498634128575, "loss": 0.0955, "step": 7330 }, { "epoch": 1.9870059556036817, "grad_norm": 0.0517578125, "learning_rate": 0.00017036969586596248, "loss": 0.0773, "step": 7340 }, { "epoch": 1.989713048186248, "grad_norm": 0.08447265625, "learning_rate": 0.00016991440539063923, "loss": 0.1104, "step": 7350 }, { "epoch": 1.9924201407688142, "grad_norm": 0.059814453125, "learning_rate": 0.00016945911491531598, "loss": 0.1227, "step": 7360 }, { "epoch": 1.9951272333513805, "grad_norm": 0.3359375, "learning_rate": 0.0001690038244399927, "loss": 0.0932, "step": 7370 }, { "epoch": 1.997834325933947, "grad_norm": 0.384765625, "learning_rate": 0.00016854853396466946, "loss": 0.0778, "step": 7380 }, { "epoch": 2.000541418516513, "grad_norm": 0.30859375, "learning_rate": 0.00016809324348934622, "loss": 0.1014, "step": 7390 }, { "epoch": 2.0032485110990796, "grad_norm": 0.1767578125, "learning_rate": 0.00016763795301402297, "loss": 0.0864, "step": 7400 }, { "epoch": 2.0032485110990796, "eval_loss": 0.10161517560482025, "eval_runtime": 38.9047, "eval_samples_per_second": 12.852, "eval_steps_per_second": 0.823, "step": 7400 }, { "epoch": 2.0059556036816457, "grad_norm": 0.0947265625, "learning_rate": 0.00016718266253869967, "loss": 0.0795, "step": 7410 }, { "epoch": 2.0086626962642122, "grad_norm": 0.1474609375, "learning_rate": 0.00016672737206337642, "loss": 0.0696, "step": 7420 }, { "epoch": 2.0113697888467788, "grad_norm": 0.162109375, "learning_rate": 0.00016627208158805318, "loss": 0.117, "step": 7430 }, { "epoch": 2.014076881429345, "grad_norm": 0.189453125, "learning_rate": 0.00016581679111272993, "loss": 0.0766, "step": 7440 }, { "epoch": 2.0167839740119113, "grad_norm": 0.083984375, "learning_rate": 0.00016536150063740668, "loss": 0.085, "step": 7450 }, { "epoch": 2.0194910665944774, "grad_norm": 0.125, "learning_rate": 0.0001649062101620834, "loss": 0.0771, "step": 7460 }, { "epoch": 2.022198159177044, "grad_norm": 0.2392578125, "learning_rate": 0.00016445091968676016, "loss": 0.1123, "step": 7470 }, { "epoch": 2.02490525175961, "grad_norm": 0.2080078125, "learning_rate": 0.0001639956292114369, "loss": 0.0886, "step": 7480 }, { "epoch": 2.0276123443421765, "grad_norm": 0.09130859375, "learning_rate": 0.00016354033873611364, "loss": 0.1066, "step": 7490 }, { "epoch": 2.030319436924743, "grad_norm": 0.1708984375, "learning_rate": 0.00016308504826079037, "loss": 0.0921, "step": 7500 }, { "epoch": 2.030319436924743, "eval_loss": 0.10184778273105621, "eval_runtime": 40.6916, "eval_samples_per_second": 12.288, "eval_steps_per_second": 0.786, "step": 7500 }, { "epoch": 2.033026529507309, "grad_norm": 0.12778520584106445, "learning_rate": 0.00016262975778546712, "loss": 0.0997, "step": 7510 }, { "epoch": 2.0357336220898756, "grad_norm": 0.16627921164035797, "learning_rate": 0.00016217446731014388, "loss": 0.0935, "step": 7520 }, { "epoch": 2.0384407146724417, "grad_norm": 0.42422693967819214, "learning_rate": 0.00016171917683482063, "loss": 0.1118, "step": 7530 }, { "epoch": 2.041147807255008, "grad_norm": 0.27819448709487915, "learning_rate": 0.00016126388635949739, "loss": 0.1049, "step": 7540 }, { "epoch": 2.0438548998375743, "grad_norm": 0.09689724445343018, "learning_rate": 0.00016080859588417409, "loss": 0.1187, "step": 7550 }, { "epoch": 2.046561992420141, "grad_norm": 0.2709833085536957, "learning_rate": 0.00016035330540885084, "loss": 0.0767, "step": 7560 }, { "epoch": 2.049269085002707, "grad_norm": 0.5571317672729492, "learning_rate": 0.0001598980149335276, "loss": 0.1035, "step": 7570 }, { "epoch": 2.0519761775852734, "grad_norm": 0.13689348101615906, "learning_rate": 0.00015944272445820435, "loss": 0.0917, "step": 7580 }, { "epoch": 2.05468327016784, "grad_norm": 0.26327988505363464, "learning_rate": 0.00015898743398288107, "loss": 0.0989, "step": 7590 }, { "epoch": 2.057390362750406, "grad_norm": 0.33440539240837097, "learning_rate": 0.00015853214350755783, "loss": 0.0976, "step": 7600 }, { "epoch": 2.057390362750406, "eval_loss": 0.10394543409347534, "eval_runtime": 48.3601, "eval_samples_per_second": 10.339, "eval_steps_per_second": 0.662, "step": 7600 }, { "epoch": 2.0600974553329725, "grad_norm": 0.24803055822849274, "learning_rate": 0.00015807685303223458, "loss": 0.1202, "step": 7610 }, { "epoch": 2.0628045479155386, "grad_norm": 0.32904544472694397, "learning_rate": 0.0001576215625569113, "loss": 0.1064, "step": 7620 }, { "epoch": 2.065511640498105, "grad_norm": 0.05891219526529312, "learning_rate": 0.00015716627208158806, "loss": 0.0805, "step": 7630 }, { "epoch": 2.068218733080671, "grad_norm": 0.16381780803203583, "learning_rate": 0.0001567109816062648, "loss": 0.1165, "step": 7640 }, { "epoch": 2.0709258256632377, "grad_norm": 0.2032223641872406, "learning_rate": 0.00015625569113094154, "loss": 0.0882, "step": 7650 }, { "epoch": 2.073632918245804, "grad_norm": 0.3922336995601654, "learning_rate": 0.0001558004006556183, "loss": 0.1174, "step": 7660 }, { "epoch": 2.0763400108283703, "grad_norm": 0.13339735567569733, "learning_rate": 0.00015534511018029505, "loss": 0.1047, "step": 7670 }, { "epoch": 2.079047103410937, "grad_norm": 0.2604218125343323, "learning_rate": 0.00015488981970497177, "loss": 0.1168, "step": 7680 }, { "epoch": 2.081754195993503, "grad_norm": 0.1896478235721588, "learning_rate": 0.0001544345292296485, "loss": 0.0953, "step": 7690 }, { "epoch": 2.0844612885760694, "grad_norm": 0.2883220911026001, "learning_rate": 0.00015397923875432525, "loss": 0.0985, "step": 7700 }, { "epoch": 2.0844612885760694, "eval_loss": 0.10269894450902939, "eval_runtime": 39.8398, "eval_samples_per_second": 12.55, "eval_steps_per_second": 0.803, "step": 7700 }, { "epoch": 2.0871683811586355, "grad_norm": 0.29532504081726074, "learning_rate": 0.000153523948279002, "loss": 0.0758, "step": 7710 }, { "epoch": 2.089875473741202, "grad_norm": 0.15288285911083221, "learning_rate": 0.00015306865780367876, "loss": 0.0822, "step": 7720 }, { "epoch": 2.0925825663237685, "grad_norm": 0.4583885967731476, "learning_rate": 0.0001526133673283555, "loss": 0.1121, "step": 7730 }, { "epoch": 2.0952896589063346, "grad_norm": 0.23765157163143158, "learning_rate": 0.00015215807685303224, "loss": 0.0869, "step": 7740 }, { "epoch": 2.097996751488901, "grad_norm": 0.17887142300605774, "learning_rate": 0.000151702786377709, "loss": 0.1067, "step": 7750 }, { "epoch": 2.100703844071467, "grad_norm": 0.3588847517967224, "learning_rate": 0.00015124749590238572, "loss": 0.0897, "step": 7760 }, { "epoch": 2.1034109366540337, "grad_norm": 0.27943047881126404, "learning_rate": 0.00015079220542706245, "loss": 0.1006, "step": 7770 }, { "epoch": 2.1061180292365997, "grad_norm": 0.18609361350536346, "learning_rate": 0.0001503369149517392, "loss": 0.0943, "step": 7780 }, { "epoch": 2.1088251218191663, "grad_norm": 0.1686626672744751, "learning_rate": 0.00014988162447641596, "loss": 0.0874, "step": 7790 }, { "epoch": 2.1115322144017323, "grad_norm": 0.20544451475143433, "learning_rate": 0.0001494263340010927, "loss": 0.1168, "step": 7800 }, { "epoch": 2.1115322144017323, "eval_loss": 0.10231052339076996, "eval_runtime": 42.8106, "eval_samples_per_second": 11.679, "eval_steps_per_second": 0.747, "step": 7800 }, { "epoch": 2.114239306984299, "grad_norm": 0.27586010098457336, "learning_rate": 0.00014897104352576946, "loss": 0.1177, "step": 7810 }, { "epoch": 2.1169463995668654, "grad_norm": 0.21565933525562286, "learning_rate": 0.0001485157530504462, "loss": 0.1019, "step": 7820 }, { "epoch": 2.1196534921494314, "grad_norm": 0.24475212395191193, "learning_rate": 0.00014806046257512292, "loss": 0.1035, "step": 7830 }, { "epoch": 2.122360584731998, "grad_norm": 0.21911804378032684, "learning_rate": 0.00014760517209979967, "loss": 0.0863, "step": 7840 }, { "epoch": 2.125067677314564, "grad_norm": 0.13527126610279083, "learning_rate": 0.00014714988162447642, "loss": 0.0853, "step": 7850 }, { "epoch": 2.1277747698971305, "grad_norm": 0.19728311896324158, "learning_rate": 0.00014669459114915315, "loss": 0.1223, "step": 7860 }, { "epoch": 2.1304818624796966, "grad_norm": 0.083954356610775, "learning_rate": 0.0001462393006738299, "loss": 0.0731, "step": 7870 }, { "epoch": 2.133188955062263, "grad_norm": 0.10076189041137695, "learning_rate": 0.00014578401019850666, "loss": 0.0886, "step": 7880 }, { "epoch": 2.1358960476448297, "grad_norm": 0.07177887111902237, "learning_rate": 0.0001453287197231834, "loss": 0.1017, "step": 7890 }, { "epoch": 2.1386031402273957, "grad_norm": 0.36714619398117065, "learning_rate": 0.00014487342924786014, "loss": 0.1024, "step": 7900 }, { "epoch": 2.1386031402273957, "eval_loss": 0.09967782348394394, "eval_runtime": 40.728, "eval_samples_per_second": 12.277, "eval_steps_per_second": 0.786, "step": 7900 }, { "epoch": 2.1413102328099622, "grad_norm": 0.13318905234336853, "learning_rate": 0.00014441813877253686, "loss": 0.0969, "step": 7910 }, { "epoch": 2.1440173253925283, "grad_norm": 0.18819299340248108, "learning_rate": 0.00014396284829721362, "loss": 0.0979, "step": 7920 }, { "epoch": 2.146724417975095, "grad_norm": 0.23004436492919922, "learning_rate": 0.00014350755782189037, "loss": 0.0989, "step": 7930 }, { "epoch": 2.149431510557661, "grad_norm": 0.11162354052066803, "learning_rate": 0.00014305226734656712, "loss": 0.0679, "step": 7940 }, { "epoch": 2.1521386031402274, "grad_norm": 0.1345404088497162, "learning_rate": 0.00014259697687124385, "loss": 0.1319, "step": 7950 }, { "epoch": 2.154845695722794, "grad_norm": 0.3883691430091858, "learning_rate": 0.0001421416863959206, "loss": 0.0988, "step": 7960 }, { "epoch": 2.15755278830536, "grad_norm": 0.32130658626556396, "learning_rate": 0.00014168639592059733, "loss": 0.0944, "step": 7970 }, { "epoch": 2.1602598808879265, "grad_norm": 0.3231971859931946, "learning_rate": 0.00014123110544527408, "loss": 0.1027, "step": 7980 }, { "epoch": 2.1629669734704926, "grad_norm": 0.22434420883655548, "learning_rate": 0.00014077581496995084, "loss": 0.109, "step": 7990 }, { "epoch": 2.165674066053059, "grad_norm": 0.06923877447843552, "learning_rate": 0.00014032052449462756, "loss": 0.0942, "step": 8000 }, { "epoch": 2.165674066053059, "eval_loss": 0.09789612144231796, "eval_runtime": 41.6904, "eval_samples_per_second": 11.993, "eval_steps_per_second": 0.768, "step": 8000 }, { "epoch": 2.168381158635625, "grad_norm": 0.14302651584148407, "learning_rate": 0.00013986523401930432, "loss": 0.0859, "step": 8010 }, { "epoch": 2.1710882512181917, "grad_norm": 0.05597392097115517, "learning_rate": 0.00013940994354398107, "loss": 0.0935, "step": 8020 }, { "epoch": 2.173795343800758, "grad_norm": 0.20413243770599365, "learning_rate": 0.00013895465306865783, "loss": 0.0884, "step": 8030 }, { "epoch": 2.1765024363833243, "grad_norm": 0.23490768671035767, "learning_rate": 0.00013849936259333452, "loss": 0.0906, "step": 8040 }, { "epoch": 2.179209528965891, "grad_norm": 0.25336843729019165, "learning_rate": 0.00013804407211801128, "loss": 0.0887, "step": 8050 }, { "epoch": 2.181916621548457, "grad_norm": 0.14275997877120972, "learning_rate": 0.00013758878164268803, "loss": 0.0724, "step": 8060 }, { "epoch": 2.1846237141310234, "grad_norm": 0.1846480518579483, "learning_rate": 0.00013713349116736479, "loss": 0.0887, "step": 8070 }, { "epoch": 2.1873308067135895, "grad_norm": 0.05883602797985077, "learning_rate": 0.00013667820069204154, "loss": 0.0862, "step": 8080 }, { "epoch": 2.190037899296156, "grad_norm": 0.25291454792022705, "learning_rate": 0.00013622291021671827, "loss": 0.0977, "step": 8090 }, { "epoch": 2.192744991878722, "grad_norm": 0.1518758088350296, "learning_rate": 0.00013576761974139502, "loss": 0.0942, "step": 8100 }, { "epoch": 2.192744991878722, "eval_loss": 0.09963502734899521, "eval_runtime": 44.7728, "eval_samples_per_second": 11.167, "eval_steps_per_second": 0.715, "step": 8100 }, { "epoch": 2.1954520844612886, "grad_norm": 0.20356951653957367, "learning_rate": 0.00013531232926607177, "loss": 0.0994, "step": 8110 }, { "epoch": 2.198159177043855, "grad_norm": 0.1818905472755432, "learning_rate": 0.0001348570387907485, "loss": 0.0925, "step": 8120 }, { "epoch": 2.200866269626421, "grad_norm": 0.16869580745697021, "learning_rate": 0.00013440174831542523, "loss": 0.0854, "step": 8130 }, { "epoch": 2.2035733622089877, "grad_norm": 0.23541289567947388, "learning_rate": 0.00013394645784010198, "loss": 0.1086, "step": 8140 }, { "epoch": 2.2062804547915538, "grad_norm": 0.31641900539398193, "learning_rate": 0.00013349116736477873, "loss": 0.0939, "step": 8150 }, { "epoch": 2.2089875473741203, "grad_norm": 0.4820684790611267, "learning_rate": 0.0001330358768894555, "loss": 0.0854, "step": 8160 }, { "epoch": 2.2116946399566864, "grad_norm": 0.17261041700839996, "learning_rate": 0.00013258058641413224, "loss": 0.0819, "step": 8170 }, { "epoch": 2.214401732539253, "grad_norm": 0.21738745272159576, "learning_rate": 0.00013212529593880897, "loss": 0.0845, "step": 8180 }, { "epoch": 2.2171088251218194, "grad_norm": 0.23873668909072876, "learning_rate": 0.0001316700054634857, "loss": 0.1089, "step": 8190 }, { "epoch": 2.2198159177043855, "grad_norm": 0.31528007984161377, "learning_rate": 0.00013121471498816245, "loss": 0.0819, "step": 8200 }, { "epoch": 2.2198159177043855, "eval_loss": 0.09538110345602036, "eval_runtime": 42.8223, "eval_samples_per_second": 11.676, "eval_steps_per_second": 0.747, "step": 8200 }, { "epoch": 2.222523010286952, "grad_norm": 0.34557875990867615, "learning_rate": 0.0001307594245128392, "loss": 0.0803, "step": 8210 }, { "epoch": 2.225230102869518, "grad_norm": 0.3071202337741852, "learning_rate": 0.00013030413403751593, "loss": 0.1522, "step": 8220 }, { "epoch": 2.2279371954520846, "grad_norm": 0.32279664278030396, "learning_rate": 0.00012984884356219268, "loss": 0.1022, "step": 8230 }, { "epoch": 2.2306442880346506, "grad_norm": 0.19390374422073364, "learning_rate": 0.00012939355308686943, "loss": 0.0922, "step": 8240 }, { "epoch": 2.233351380617217, "grad_norm": 0.391799658536911, "learning_rate": 0.0001289382626115462, "loss": 0.1113, "step": 8250 }, { "epoch": 2.2360584731997832, "grad_norm": 0.07327993959188461, "learning_rate": 0.00012848297213622291, "loss": 0.1306, "step": 8260 }, { "epoch": 2.2387655657823498, "grad_norm": 0.0734788253903389, "learning_rate": 0.00012802768166089964, "loss": 0.0864, "step": 8270 }, { "epoch": 2.2414726583649163, "grad_norm": 0.024906840175390244, "learning_rate": 0.0001275723911855764, "loss": 0.0814, "step": 8280 }, { "epoch": 2.2441797509474823, "grad_norm": 0.08028047531843185, "learning_rate": 0.00012711710071025315, "loss": 0.0756, "step": 8290 }, { "epoch": 2.246886843530049, "grad_norm": 0.23390653729438782, "learning_rate": 0.0001266618102349299, "loss": 0.0842, "step": 8300 }, { "epoch": 2.246886843530049, "eval_loss": 0.09282638877630234, "eval_runtime": 40.8595, "eval_samples_per_second": 12.237, "eval_steps_per_second": 0.783, "step": 8300 }, { "epoch": 2.249593936112615, "grad_norm": 0.15658049285411835, "learning_rate": 0.00012620651975960663, "loss": 0.0908, "step": 8310 }, { "epoch": 2.2523010286951815, "grad_norm": 0.3025362193584442, "learning_rate": 0.00012575122928428338, "loss": 0.0887, "step": 8320 }, { "epoch": 2.2550081212777475, "grad_norm": 0.08213713765144348, "learning_rate": 0.0001252959388089601, "loss": 0.0897, "step": 8330 }, { "epoch": 2.257715213860314, "grad_norm": 0.18460877239704132, "learning_rate": 0.00012484064833363686, "loss": 0.0837, "step": 8340 }, { "epoch": 2.26042230644288, "grad_norm": 0.1409384310245514, "learning_rate": 0.00012438535785831362, "loss": 0.0662, "step": 8350 }, { "epoch": 2.2631293990254466, "grad_norm": 0.2858680784702301, "learning_rate": 0.00012393006738299034, "loss": 0.0823, "step": 8360 }, { "epoch": 2.265836491608013, "grad_norm": 0.18802529573440552, "learning_rate": 0.0001234747769076671, "loss": 0.0903, "step": 8370 }, { "epoch": 2.268543584190579, "grad_norm": 0.1960514485836029, "learning_rate": 0.00012301948643234385, "loss": 0.1104, "step": 8380 }, { "epoch": 2.2712506767731457, "grad_norm": 0.174361914396286, "learning_rate": 0.00012256419595702058, "loss": 0.1145, "step": 8390 }, { "epoch": 2.273957769355712, "grad_norm": 0.07053990662097931, "learning_rate": 0.00012210890548169733, "loss": 0.111, "step": 8400 }, { "epoch": 2.273957769355712, "eval_loss": 0.09893127530813217, "eval_runtime": 43.1604, "eval_samples_per_second": 11.585, "eval_steps_per_second": 0.741, "step": 8400 }, { "epoch": 2.2766648619382783, "grad_norm": 0.10177663713693619, "learning_rate": 0.00012165361500637407, "loss": 0.0869, "step": 8410 }, { "epoch": 2.279371954520845, "grad_norm": 0.14843662083148956, "learning_rate": 0.00012119832453105081, "loss": 0.1087, "step": 8420 }, { "epoch": 2.282079047103411, "grad_norm": 0.09722846746444702, "learning_rate": 0.00012074303405572756, "loss": 0.0803, "step": 8430 }, { "epoch": 2.2847861396859774, "grad_norm": 0.1296207755804062, "learning_rate": 0.0001202877435804043, "loss": 0.1223, "step": 8440 }, { "epoch": 2.2874932322685435, "grad_norm": 0.11870713531970978, "learning_rate": 0.00011983245310508104, "loss": 0.0859, "step": 8450 }, { "epoch": 2.29020032485111, "grad_norm": 0.198283389210701, "learning_rate": 0.00011937716262975778, "loss": 0.0872, "step": 8460 }, { "epoch": 2.292907417433676, "grad_norm": 0.3522505760192871, "learning_rate": 0.00011892187215443454, "loss": 0.1075, "step": 8470 }, { "epoch": 2.2956145100162426, "grad_norm": 0.09621834754943848, "learning_rate": 0.00011846658167911128, "loss": 0.0789, "step": 8480 }, { "epoch": 2.2983216025988087, "grad_norm": 0.16097977757453918, "learning_rate": 0.00011801129120378802, "loss": 0.0812, "step": 8490 }, { "epoch": 2.301028695181375, "grad_norm": 0.26149022579193115, "learning_rate": 0.00011755600072846476, "loss": 0.104, "step": 8500 }, { "epoch": 2.301028695181375, "eval_loss": 0.09862113744020462, "eval_runtime": 44.4327, "eval_samples_per_second": 11.253, "eval_steps_per_second": 0.72, "step": 8500 }, { "epoch": 2.3037357877639417, "grad_norm": 0.43676215410232544, "learning_rate": 0.00011710071025314151, "loss": 0.0989, "step": 8510 }, { "epoch": 2.306442880346508, "grad_norm": 0.27762916684150696, "learning_rate": 0.00011664541977781825, "loss": 0.1034, "step": 8520 }, { "epoch": 2.3091499729290743, "grad_norm": 0.14587101340293884, "learning_rate": 0.00011619012930249499, "loss": 0.0972, "step": 8530 }, { "epoch": 2.3118570655116404, "grad_norm": 0.13166609406471252, "learning_rate": 0.00011573483882717174, "loss": 0.0934, "step": 8540 }, { "epoch": 2.314564158094207, "grad_norm": 0.24040523171424866, "learning_rate": 0.00011527954835184849, "loss": 0.0872, "step": 8550 }, { "epoch": 2.317271250676773, "grad_norm": 0.2743408977985382, "learning_rate": 0.00011482425787652523, "loss": 0.1008, "step": 8560 }, { "epoch": 2.3199783432593395, "grad_norm": 0.18583521246910095, "learning_rate": 0.00011436896740120197, "loss": 0.1058, "step": 8570 }, { "epoch": 2.3226854358419056, "grad_norm": 0.07435272634029388, "learning_rate": 0.00011391367692587872, "loss": 0.1003, "step": 8580 }, { "epoch": 2.325392528424472, "grad_norm": 0.3827980160713196, "learning_rate": 0.00011345838645055545, "loss": 0.1187, "step": 8590 }, { "epoch": 2.3280996210070386, "grad_norm": 0.14449597895145416, "learning_rate": 0.0001130030959752322, "loss": 0.07, "step": 8600 }, { "epoch": 2.3280996210070386, "eval_loss": 0.09334024041891098, "eval_runtime": 42.4415, "eval_samples_per_second": 11.781, "eval_steps_per_second": 0.754, "step": 8600 }, { "epoch": 2.3308067135896047, "grad_norm": 0.22377613186836243, "learning_rate": 0.00011254780549990895, "loss": 0.1011, "step": 8610 }, { "epoch": 2.333513806172171, "grad_norm": 0.18512682616710663, "learning_rate": 0.00011209251502458569, "loss": 0.0777, "step": 8620 }, { "epoch": 2.3362208987547373, "grad_norm": 0.20986367762088776, "learning_rate": 0.00011163722454926243, "loss": 0.0847, "step": 8630 }, { "epoch": 2.338927991337304, "grad_norm": 0.2790672183036804, "learning_rate": 0.00011118193407393917, "loss": 0.0644, "step": 8640 }, { "epoch": 2.3416350839198703, "grad_norm": 0.08240889012813568, "learning_rate": 0.00011072664359861593, "loss": 0.1102, "step": 8650 }, { "epoch": 2.3443421765024364, "grad_norm": 0.19567365944385529, "learning_rate": 0.00011027135312329265, "loss": 0.0765, "step": 8660 }, { "epoch": 2.347049269085003, "grad_norm": 0.27101942896842957, "learning_rate": 0.0001098160626479694, "loss": 0.0894, "step": 8670 }, { "epoch": 2.349756361667569, "grad_norm": 0.20218369364738464, "learning_rate": 0.00010936077217264615, "loss": 0.0698, "step": 8680 }, { "epoch": 2.3524634542501355, "grad_norm": 0.27786144614219666, "learning_rate": 0.0001089054816973229, "loss": 0.101, "step": 8690 }, { "epoch": 2.3551705468327016, "grad_norm": 0.16329264640808105, "learning_rate": 0.00010845019122199964, "loss": 0.0989, "step": 8700 }, { "epoch": 2.3551705468327016, "eval_loss": 0.09510345757007599, "eval_runtime": 40.3837, "eval_samples_per_second": 12.381, "eval_steps_per_second": 0.792, "step": 8700 }, { "epoch": 2.357877639415268, "grad_norm": 0.1290864795446396, "learning_rate": 0.00010799490074667638, "loss": 0.08, "step": 8710 }, { "epoch": 2.360584731997834, "grad_norm": 0.2313816100358963, "learning_rate": 0.00010753961027135313, "loss": 0.0897, "step": 8720 }, { "epoch": 2.3632918245804007, "grad_norm": 0.159449964761734, "learning_rate": 0.00010708431979602986, "loss": 0.0785, "step": 8730 }, { "epoch": 2.365998917162967, "grad_norm": 0.3195575475692749, "learning_rate": 0.00010662902932070661, "loss": 0.1141, "step": 8740 }, { "epoch": 2.3687060097455332, "grad_norm": 0.30101099610328674, "learning_rate": 0.00010617373884538335, "loss": 0.0899, "step": 8750 }, { "epoch": 2.3714131023280998, "grad_norm": 0.37577658891677856, "learning_rate": 0.00010571844837006011, "loss": 0.1019, "step": 8760 }, { "epoch": 2.374120194910666, "grad_norm": 0.2812362015247345, "learning_rate": 0.00010526315789473683, "loss": 0.0846, "step": 8770 }, { "epoch": 2.3768272874932324, "grad_norm": 0.453045517206192, "learning_rate": 0.00010480786741941359, "loss": 0.0784, "step": 8780 }, { "epoch": 2.3795343800757984, "grad_norm": 0.2965713143348694, "learning_rate": 0.00010435257694409034, "loss": 0.0932, "step": 8790 }, { "epoch": 2.382241472658365, "grad_norm": 0.1862928420305252, "learning_rate": 0.00010389728646876707, "loss": 0.0853, "step": 8800 }, { "epoch": 2.382241472658365, "eval_loss": 0.0960661917924881, "eval_runtime": 44.3311, "eval_samples_per_second": 11.279, "eval_steps_per_second": 0.722, "step": 8800 }, { "epoch": 2.384948565240931, "grad_norm": 0.15719877183437347, "learning_rate": 0.00010344199599344382, "loss": 0.1047, "step": 8810 }, { "epoch": 2.3876556578234975, "grad_norm": 0.21618276834487915, "learning_rate": 0.00010298670551812056, "loss": 0.086, "step": 8820 }, { "epoch": 2.390362750406064, "grad_norm": 0.14036358892917633, "learning_rate": 0.00010253141504279732, "loss": 0.0834, "step": 8830 }, { "epoch": 2.39306984298863, "grad_norm": 0.21179358661174774, "learning_rate": 0.00010207612456747404, "loss": 0.1091, "step": 8840 }, { "epoch": 2.3957769355711966, "grad_norm": 0.15358150005340576, "learning_rate": 0.0001016208340921508, "loss": 0.1075, "step": 8850 }, { "epoch": 2.3984840281537627, "grad_norm": 0.1115250512957573, "learning_rate": 0.00010116554361682754, "loss": 0.0971, "step": 8860 }, { "epoch": 2.4011911207363292, "grad_norm": 0.30404922366142273, "learning_rate": 0.00010071025314150428, "loss": 0.102, "step": 8870 }, { "epoch": 2.4038982133188957, "grad_norm": 0.25525572896003723, "learning_rate": 0.00010025496266618103, "loss": 0.0624, "step": 8880 }, { "epoch": 2.406605305901462, "grad_norm": 0.21130895614624023, "learning_rate": 9.979967219085777e-05, "loss": 0.1141, "step": 8890 }, { "epoch": 2.4093123984840283, "grad_norm": 0.12112224102020264, "learning_rate": 9.934438171553452e-05, "loss": 0.1121, "step": 8900 }, { "epoch": 2.4093123984840283, "eval_loss": 0.09973784536123276, "eval_runtime": 40.7796, "eval_samples_per_second": 12.261, "eval_steps_per_second": 0.785, "step": 8900 }, { "epoch": 2.4120194910665944, "grad_norm": 0.0505850575864315, "learning_rate": 9.888909124021125e-05, "loss": 0.0986, "step": 8910 }, { "epoch": 2.414726583649161, "grad_norm": 0.22455182671546936, "learning_rate": 9.8433800764888e-05, "loss": 0.1028, "step": 8920 }, { "epoch": 2.417433676231727, "grad_norm": 0.25635039806365967, "learning_rate": 9.797851028956474e-05, "loss": 0.0835, "step": 8930 }, { "epoch": 2.4201407688142935, "grad_norm": 0.21484720706939697, "learning_rate": 9.75232198142415e-05, "loss": 0.1197, "step": 8940 }, { "epoch": 2.4228478613968596, "grad_norm": 0.1871686577796936, "learning_rate": 9.706792933891822e-05, "loss": 0.1198, "step": 8950 }, { "epoch": 2.425554953979426, "grad_norm": 0.12417078763246536, "learning_rate": 9.661263886359498e-05, "loss": 0.0766, "step": 8960 }, { "epoch": 2.4282620465619926, "grad_norm": 0.07156313210725784, "learning_rate": 9.615734838827173e-05, "loss": 0.0825, "step": 8970 }, { "epoch": 2.4309691391445587, "grad_norm": 0.17706018686294556, "learning_rate": 9.570205791294846e-05, "loss": 0.0981, "step": 8980 }, { "epoch": 2.433676231727125, "grad_norm": 0.11764348298311234, "learning_rate": 9.524676743762521e-05, "loss": 0.0944, "step": 8990 }, { "epoch": 2.4363833243096913, "grad_norm": 0.20927299559116364, "learning_rate": 9.479147696230195e-05, "loss": 0.1056, "step": 9000 }, { "epoch": 2.4363833243096913, "eval_loss": 0.09689030796289444, "eval_runtime": 43.5061, "eval_samples_per_second": 11.493, "eval_steps_per_second": 0.736, "step": 9000 }, { "epoch": 2.439090416892258, "grad_norm": 0.28628361225128174, "learning_rate": 9.43361864869787e-05, "loss": 0.0948, "step": 9010 }, { "epoch": 2.441797509474824, "grad_norm": 0.08211715519428253, "learning_rate": 9.388089601165543e-05, "loss": 0.0776, "step": 9020 }, { "epoch": 2.4445046020573904, "grad_norm": 0.10449658334255219, "learning_rate": 9.342560553633218e-05, "loss": 0.0957, "step": 9030 }, { "epoch": 2.4472116946399565, "grad_norm": 0.3579693138599396, "learning_rate": 9.297031506100892e-05, "loss": 0.0722, "step": 9040 }, { "epoch": 2.449918787222523, "grad_norm": 0.116954006254673, "learning_rate": 9.251502458568566e-05, "loss": 0.1053, "step": 9050 }, { "epoch": 2.4526258798050895, "grad_norm": 0.14063984155654907, "learning_rate": 9.205973411036242e-05, "loss": 0.0749, "step": 9060 }, { "epoch": 2.4553329723876556, "grad_norm": 0.3165554702281952, "learning_rate": 9.160444363503916e-05, "loss": 0.104, "step": 9070 }, { "epoch": 2.458040064970222, "grad_norm": 0.5529165863990784, "learning_rate": 9.114915315971591e-05, "loss": 0.1231, "step": 9080 }, { "epoch": 2.460747157552788, "grad_norm": 0.06658139824867249, "learning_rate": 9.069386268439264e-05, "loss": 0.1043, "step": 9090 }, { "epoch": 2.4634542501353547, "grad_norm": 0.2607128620147705, "learning_rate": 9.023857220906939e-05, "loss": 0.0863, "step": 9100 }, { "epoch": 2.4634542501353547, "eval_loss": 0.09591835737228394, "eval_runtime": 43.9162, "eval_samples_per_second": 11.385, "eval_steps_per_second": 0.729, "step": 9100 }, { "epoch": 2.466161342717921, "grad_norm": 0.2396080195903778, "learning_rate": 8.978328173374613e-05, "loss": 0.0891, "step": 9110 }, { "epoch": 2.4688684353004873, "grad_norm": 0.18154506385326385, "learning_rate": 8.932799125842287e-05, "loss": 0.0859, "step": 9120 }, { "epoch": 2.471575527883054, "grad_norm": 0.11554720252752304, "learning_rate": 8.887270078309961e-05, "loss": 0.0988, "step": 9130 }, { "epoch": 2.47428262046562, "grad_norm": 0.21819724142551422, "learning_rate": 8.841741030777637e-05, "loss": 0.1001, "step": 9140 }, { "epoch": 2.4769897130481864, "grad_norm": 0.2960222065448761, "learning_rate": 8.796211983245312e-05, "loss": 0.0906, "step": 9150 }, { "epoch": 2.4796968056307525, "grad_norm": 0.05587580054998398, "learning_rate": 8.750682935712985e-05, "loss": 0.1289, "step": 9160 }, { "epoch": 2.482403898213319, "grad_norm": 0.4750611186027527, "learning_rate": 8.70515388818066e-05, "loss": 0.1164, "step": 9170 }, { "epoch": 2.485110990795885, "grad_norm": 0.2813723087310791, "learning_rate": 8.659624840648334e-05, "loss": 0.0937, "step": 9180 }, { "epoch": 2.4878180833784516, "grad_norm": 0.3696264922618866, "learning_rate": 8.614095793116008e-05, "loss": 0.0815, "step": 9190 }, { "epoch": 2.490525175961018, "grad_norm": 0.18263721466064453, "learning_rate": 8.568566745583682e-05, "loss": 0.0814, "step": 9200 }, { "epoch": 2.490525175961018, "eval_loss": 0.09541510045528412, "eval_runtime": 43.3208, "eval_samples_per_second": 11.542, "eval_steps_per_second": 0.739, "step": 9200 }, { "epoch": 2.493232268543584, "grad_norm": 0.23378819227218628, "learning_rate": 8.523037698051357e-05, "loss": 0.0938, "step": 9210 }, { "epoch": 2.4959393611261507, "grad_norm": 0.17593856155872345, "learning_rate": 8.477508650519031e-05, "loss": 0.0978, "step": 9220 }, { "epoch": 2.4986464537087167, "grad_norm": 0.1842566281557083, "learning_rate": 8.431979602986705e-05, "loss": 0.1278, "step": 9230 }, { "epoch": 2.5013535462912833, "grad_norm": 0.4087991714477539, "learning_rate": 8.386450555454381e-05, "loss": 0.0856, "step": 9240 }, { "epoch": 2.5040606388738493, "grad_norm": 0.3823549151420593, "learning_rate": 8.340921507922055e-05, "loss": 0.1025, "step": 9250 }, { "epoch": 2.506767731456416, "grad_norm": 0.33429262042045593, "learning_rate": 8.295392460389729e-05, "loss": 0.0958, "step": 9260 }, { "epoch": 2.509474824038982, "grad_norm": 0.2929178774356842, "learning_rate": 8.249863412857403e-05, "loss": 0.0923, "step": 9270 }, { "epoch": 2.5121819166215484, "grad_norm": 0.22063800692558289, "learning_rate": 8.204334365325078e-05, "loss": 0.0872, "step": 9280 }, { "epoch": 2.514889009204115, "grad_norm": 0.22350943088531494, "learning_rate": 8.158805317792752e-05, "loss": 0.0726, "step": 9290 }, { "epoch": 2.517596101786681, "grad_norm": 0.21482957899570465, "learning_rate": 8.113276270260426e-05, "loss": 0.0947, "step": 9300 }, { "epoch": 2.517596101786681, "eval_loss": 0.0981806218624115, "eval_runtime": 45.1282, "eval_samples_per_second": 11.08, "eval_steps_per_second": 0.709, "step": 9300 }, { "epoch": 2.5203031943692475, "grad_norm": 0.24753856658935547, "learning_rate": 8.0677472227281e-05, "loss": 0.0787, "step": 9310 }, { "epoch": 2.5230102869518136, "grad_norm": 0.18322217464447021, "learning_rate": 8.022218175195775e-05, "loss": 0.0893, "step": 9320 }, { "epoch": 2.52571737953438, "grad_norm": 0.0984969511628151, "learning_rate": 7.97668912766345e-05, "loss": 0.0954, "step": 9330 }, { "epoch": 2.5284244721169467, "grad_norm": 0.4846622347831726, "learning_rate": 7.931160080131124e-05, "loss": 0.0917, "step": 9340 }, { "epoch": 2.5311315646995127, "grad_norm": 0.12956853210926056, "learning_rate": 7.885631032598799e-05, "loss": 0.0884, "step": 9350 }, { "epoch": 2.533838657282079, "grad_norm": 0.3371259570121765, "learning_rate": 7.840101985066473e-05, "loss": 0.1002, "step": 9360 }, { "epoch": 2.5365457498646453, "grad_norm": 0.15814515948295593, "learning_rate": 7.794572937534147e-05, "loss": 0.0636, "step": 9370 }, { "epoch": 2.539252842447212, "grad_norm": 0.2645837664604187, "learning_rate": 7.749043890001821e-05, "loss": 0.0838, "step": 9380 }, { "epoch": 2.541959935029778, "grad_norm": 0.25143322348594666, "learning_rate": 7.703514842469496e-05, "loss": 0.0979, "step": 9390 }, { "epoch": 2.5446670276123444, "grad_norm": 0.10179305821657181, "learning_rate": 7.657985794937169e-05, "loss": 0.0835, "step": 9400 }, { "epoch": 2.5446670276123444, "eval_loss": 0.09655719995498657, "eval_runtime": 43.6504, "eval_samples_per_second": 11.455, "eval_steps_per_second": 0.733, "step": 9400 }, { "epoch": 2.5473741201949105, "grad_norm": 0.07296793162822723, "learning_rate": 7.612456747404844e-05, "loss": 0.0913, "step": 9410 }, { "epoch": 2.550081212777477, "grad_norm": 0.18520580232143402, "learning_rate": 7.56692769987252e-05, "loss": 0.0841, "step": 9420 }, { "epoch": 2.5527883053600435, "grad_norm": 0.14211702346801758, "learning_rate": 7.521398652340194e-05, "loss": 0.0786, "step": 9430 }, { "epoch": 2.5554953979426096, "grad_norm": 0.21824434399604797, "learning_rate": 7.475869604807868e-05, "loss": 0.1022, "step": 9440 }, { "epoch": 2.558202490525176, "grad_norm": 0.22222596406936646, "learning_rate": 7.430340557275542e-05, "loss": 0.0944, "step": 9450 }, { "epoch": 2.560909583107742, "grad_norm": 0.2114141434431076, "learning_rate": 7.384811509743217e-05, "loss": 0.094, "step": 9460 }, { "epoch": 2.5636166756903087, "grad_norm": 0.1213628277182579, "learning_rate": 7.33928246221089e-05, "loss": 0.0817, "step": 9470 }, { "epoch": 2.566323768272875, "grad_norm": 0.6145333051681519, "learning_rate": 7.293753414678565e-05, "loss": 0.0949, "step": 9480 }, { "epoch": 2.5690308608554413, "grad_norm": 0.6199302673339844, "learning_rate": 7.248224367146239e-05, "loss": 0.1174, "step": 9490 }, { "epoch": 2.5717379534380074, "grad_norm": 0.23918431997299194, "learning_rate": 7.202695319613914e-05, "loss": 0.0857, "step": 9500 }, { "epoch": 2.5717379534380074, "eval_loss": 0.09716964513063431, "eval_runtime": 45.0518, "eval_samples_per_second": 11.098, "eval_steps_per_second": 0.71, "step": 9500 }, { "epoch": 2.574445046020574, "grad_norm": 0.1995326578617096, "learning_rate": 7.157166272081588e-05, "loss": 0.0937, "step": 9510 }, { "epoch": 2.5771521386031404, "grad_norm": 0.43306341767311096, "learning_rate": 7.111637224549262e-05, "loss": 0.0862, "step": 9520 }, { "epoch": 2.5798592311857065, "grad_norm": 0.34539929032325745, "learning_rate": 7.066108177016938e-05, "loss": 0.097, "step": 9530 }, { "epoch": 2.582566323768273, "grad_norm": 0.22150956094264984, "learning_rate": 7.02057912948461e-05, "loss": 0.1168, "step": 9540 }, { "epoch": 2.585273416350839, "grad_norm": 0.18889474868774414, "learning_rate": 6.975050081952286e-05, "loss": 0.0886, "step": 9550 }, { "epoch": 2.5879805089334056, "grad_norm": 0.21980567276477814, "learning_rate": 6.92952103441996e-05, "loss": 0.1298, "step": 9560 }, { "epoch": 2.590687601515972, "grad_norm": 0.08730834722518921, "learning_rate": 6.883991986887635e-05, "loss": 0.1387, "step": 9570 }, { "epoch": 2.593394694098538, "grad_norm": 0.2558166980743408, "learning_rate": 6.838462939355308e-05, "loss": 0.0896, "step": 9580 }, { "epoch": 2.5961017866811043, "grad_norm": 0.5545253157615662, "learning_rate": 6.792933891822983e-05, "loss": 0.1323, "step": 9590 }, { "epoch": 2.5988088792636708, "grad_norm": 0.15122897922992706, "learning_rate": 6.747404844290659e-05, "loss": 0.1053, "step": 9600 }, { "epoch": 2.5988088792636708, "eval_loss": 0.09439769387245178, "eval_runtime": 42.026, "eval_samples_per_second": 11.897, "eval_steps_per_second": 0.761, "step": 9600 }, { "epoch": 2.6015159718462373, "grad_norm": 0.2454259693622589, "learning_rate": 6.701875796758333e-05, "loss": 0.0925, "step": 9610 }, { "epoch": 2.6042230644288034, "grad_norm": 0.1580754518508911, "learning_rate": 6.656346749226007e-05, "loss": 0.0797, "step": 9620 }, { "epoch": 2.60693015701137, "grad_norm": 0.19478686153888702, "learning_rate": 6.61081770169368e-05, "loss": 0.0716, "step": 9630 }, { "epoch": 2.609637249593936, "grad_norm": 0.3353942036628723, "learning_rate": 6.565288654161356e-05, "loss": 0.0883, "step": 9640 }, { "epoch": 2.6123443421765025, "grad_norm": 0.2484428584575653, "learning_rate": 6.519759606629029e-05, "loss": 0.1021, "step": 9650 }, { "epoch": 2.615051434759069, "grad_norm": 0.12130456417798996, "learning_rate": 6.474230559096704e-05, "loss": 0.0738, "step": 9660 }, { "epoch": 2.617758527341635, "grad_norm": 0.07446419447660446, "learning_rate": 6.428701511564378e-05, "loss": 0.1052, "step": 9670 }, { "epoch": 2.6204656199242016, "grad_norm": 0.2512432634830475, "learning_rate": 6.383172464032053e-05, "loss": 0.079, "step": 9680 }, { "epoch": 2.6231727125067676, "grad_norm": 0.14929881691932678, "learning_rate": 6.337643416499727e-05, "loss": 0.0666, "step": 9690 }, { "epoch": 2.625879805089334, "grad_norm": 0.38386914134025574, "learning_rate": 6.292114368967401e-05, "loss": 0.1011, "step": 9700 }, { "epoch": 2.625879805089334, "eval_loss": 0.09849968552589417, "eval_runtime": 44.3576, "eval_samples_per_second": 11.272, "eval_steps_per_second": 0.721, "step": 9700 }, { "epoch": 2.6285868976719002, "grad_norm": 0.29416972398757935, "learning_rate": 6.246585321435075e-05, "loss": 0.1057, "step": 9710 }, { "epoch": 2.6312939902544668, "grad_norm": 0.1854691207408905, "learning_rate": 6.201056273902749e-05, "loss": 0.0974, "step": 9720 }, { "epoch": 2.634001082837033, "grad_norm": 0.13421864807605743, "learning_rate": 6.155527226370425e-05, "loss": 0.0971, "step": 9730 }, { "epoch": 2.6367081754195993, "grad_norm": 0.5346465110778809, "learning_rate": 6.109998178838099e-05, "loss": 0.0894, "step": 9740 }, { "epoch": 2.639415268002166, "grad_norm": 0.07187485694885254, "learning_rate": 6.0644691313057734e-05, "loss": 0.0927, "step": 9750 }, { "epoch": 2.642122360584732, "grad_norm": 0.18485724925994873, "learning_rate": 6.018940083773448e-05, "loss": 0.0764, "step": 9760 }, { "epoch": 2.6448294531672984, "grad_norm": 0.25743237137794495, "learning_rate": 5.973411036241122e-05, "loss": 0.0846, "step": 9770 }, { "epoch": 2.6475365457498645, "grad_norm": 0.05086695775389671, "learning_rate": 5.927881988708796e-05, "loss": 0.0916, "step": 9780 }, { "epoch": 2.650243638332431, "grad_norm": 0.2837978005409241, "learning_rate": 5.882352941176471e-05, "loss": 0.1035, "step": 9790 }, { "epoch": 2.6529507309149976, "grad_norm": 0.15418092906475067, "learning_rate": 5.836823893644145e-05, "loss": 0.0806, "step": 9800 }, { "epoch": 2.6529507309149976, "eval_loss": 0.09739308059215546, "eval_runtime": 43.2101, "eval_samples_per_second": 11.571, "eval_steps_per_second": 0.741, "step": 9800 }, { "epoch": 2.6556578234975636, "grad_norm": 0.24253162741661072, "learning_rate": 5.7912948461118194e-05, "loss": 0.0988, "step": 9810 }, { "epoch": 2.6583649160801297, "grad_norm": 0.28691157698631287, "learning_rate": 5.7457657985794934e-05, "loss": 0.0811, "step": 9820 }, { "epoch": 2.661072008662696, "grad_norm": 0.13848303258419037, "learning_rate": 5.700236751047169e-05, "loss": 0.092, "step": 9830 }, { "epoch": 2.6637791012452627, "grad_norm": 0.20538252592086792, "learning_rate": 5.654707703514843e-05, "loss": 0.0924, "step": 9840 }, { "epoch": 2.666486193827829, "grad_norm": 0.34695670008659363, "learning_rate": 5.609178655982517e-05, "loss": 0.0925, "step": 9850 }, { "epoch": 2.6691932864103953, "grad_norm": 0.2103501409292221, "learning_rate": 5.5636496084501915e-05, "loss": 0.0789, "step": 9860 }, { "epoch": 2.6719003789929614, "grad_norm": 0.07302787154912949, "learning_rate": 5.5181205609178655e-05, "loss": 0.0779, "step": 9870 }, { "epoch": 2.674607471575528, "grad_norm": 0.31714409589767456, "learning_rate": 5.47259151338554e-05, "loss": 0.1012, "step": 9880 }, { "epoch": 2.6773145641580944, "grad_norm": 0.4508233070373535, "learning_rate": 5.427062465853214e-05, "loss": 0.0984, "step": 9890 }, { "epoch": 2.6800216567406605, "grad_norm": 0.13542871177196503, "learning_rate": 5.381533418320889e-05, "loss": 0.0996, "step": 9900 }, { "epoch": 2.6800216567406605, "eval_loss": 0.09776340425014496, "eval_runtime": 41.628, "eval_samples_per_second": 12.011, "eval_steps_per_second": 0.769, "step": 9900 }, { "epoch": 2.6827287493232266, "grad_norm": 0.2396474927663803, "learning_rate": 5.336004370788563e-05, "loss": 0.101, "step": 9910 }, { "epoch": 2.685435841905793, "grad_norm": 0.10130654275417328, "learning_rate": 5.2904753232562376e-05, "loss": 0.0643, "step": 9920 }, { "epoch": 2.6881429344883596, "grad_norm": 0.1962793469429016, "learning_rate": 5.244946275723912e-05, "loss": 0.0854, "step": 9930 }, { "epoch": 2.6908500270709257, "grad_norm": 0.1628696620464325, "learning_rate": 5.199417228191586e-05, "loss": 0.0792, "step": 9940 }, { "epoch": 2.693557119653492, "grad_norm": 0.2945195436477661, "learning_rate": 5.153888180659261e-05, "loss": 0.086, "step": 9950 }, { "epoch": 2.6962642122360583, "grad_norm": 0.16473691165447235, "learning_rate": 5.108359133126935e-05, "loss": 0.0696, "step": 9960 }, { "epoch": 2.698971304818625, "grad_norm": 0.12836894392967224, "learning_rate": 5.0628300855946096e-05, "loss": 0.1059, "step": 9970 }, { "epoch": 2.7016783974011913, "grad_norm": 0.10841035097837448, "learning_rate": 5.0173010380622837e-05, "loss": 0.1119, "step": 9980 }, { "epoch": 2.7043854899837574, "grad_norm": 0.7615858912467957, "learning_rate": 4.971771990529958e-05, "loss": 0.1132, "step": 9990 }, { "epoch": 2.707092582566324, "grad_norm": 0.1601744443178177, "learning_rate": 4.9262429429976323e-05, "loss": 0.1068, "step": 10000 }, { "epoch": 2.707092582566324, "eval_loss": 0.09603069722652435, "eval_runtime": 42.1403, "eval_samples_per_second": 11.865, "eval_steps_per_second": 0.759, "step": 10000 }, { "epoch": 2.70979967514889, "grad_norm": 0.3479965627193451, "learning_rate": 4.880713895465307e-05, "loss": 0.0933, "step": 10010 }, { "epoch": 2.7125067677314565, "grad_norm": 0.07585707306861877, "learning_rate": 4.835184847932982e-05, "loss": 0.0942, "step": 10020 }, { "epoch": 2.715213860314023, "grad_norm": 0.22446364164352417, "learning_rate": 4.789655800400656e-05, "loss": 0.0986, "step": 10030 }, { "epoch": 2.717920952896589, "grad_norm": 0.30344006419181824, "learning_rate": 4.7441267528683304e-05, "loss": 0.0891, "step": 10040 }, { "epoch": 2.720628045479155, "grad_norm": 0.23615258932113647, "learning_rate": 4.6985977053360044e-05, "loss": 0.0971, "step": 10050 }, { "epoch": 2.7233351380617217, "grad_norm": 0.31420329213142395, "learning_rate": 4.653068657803679e-05, "loss": 0.1069, "step": 10060 }, { "epoch": 2.726042230644288, "grad_norm": 0.25259941816329956, "learning_rate": 4.607539610271353e-05, "loss": 0.0907, "step": 10070 }, { "epoch": 2.7287493232268543, "grad_norm": 0.12377690523862839, "learning_rate": 4.562010562739027e-05, "loss": 0.1022, "step": 10080 }, { "epoch": 2.731456415809421, "grad_norm": 0.0901675596833229, "learning_rate": 4.516481515206702e-05, "loss": 0.0769, "step": 10090 }, { "epoch": 2.734163508391987, "grad_norm": 0.14518044888973236, "learning_rate": 4.4709524676743765e-05, "loss": 0.0721, "step": 10100 }, { "epoch": 2.734163508391987, "eval_loss": 0.09440316259860992, "eval_runtime": 41.7431, "eval_samples_per_second": 11.978, "eval_steps_per_second": 0.767, "step": 10100 }, { "epoch": 2.7368706009745534, "grad_norm": 0.12626998126506805, "learning_rate": 4.425423420142051e-05, "loss": 0.1052, "step": 10110 }, { "epoch": 2.73957769355712, "grad_norm": 0.33514904975891113, "learning_rate": 4.379894372609725e-05, "loss": 0.0935, "step": 10120 }, { "epoch": 2.742284786139686, "grad_norm": 0.1060992106795311, "learning_rate": 4.3343653250774e-05, "loss": 0.0695, "step": 10130 }, { "epoch": 2.744991878722252, "grad_norm": 0.11455707252025604, "learning_rate": 4.288836277545074e-05, "loss": 0.0745, "step": 10140 }, { "epoch": 2.7476989713048185, "grad_norm": 0.13291935622692108, "learning_rate": 4.243307230012748e-05, "loss": 0.0869, "step": 10150 }, { "epoch": 2.750406063887385, "grad_norm": 0.1432422697544098, "learning_rate": 4.1977781824804226e-05, "loss": 0.0996, "step": 10160 }, { "epoch": 2.753113156469951, "grad_norm": 0.09239622205495834, "learning_rate": 4.1522491349480966e-05, "loss": 0.0648, "step": 10170 }, { "epoch": 2.7558202490525177, "grad_norm": 0.5232021808624268, "learning_rate": 4.106720087415771e-05, "loss": 0.0973, "step": 10180 }, { "epoch": 2.7585273416350837, "grad_norm": 0.2642022967338562, "learning_rate": 4.061191039883446e-05, "loss": 0.0949, "step": 10190 }, { "epoch": 2.7612344342176502, "grad_norm": 0.1363670378923416, "learning_rate": 4.0156619923511206e-05, "loss": 0.0745, "step": 10200 }, { "epoch": 2.7612344342176502, "eval_loss": 0.0950176864862442, "eval_runtime": 43.2646, "eval_samples_per_second": 11.557, "eval_steps_per_second": 0.74, "step": 10200 }, { "epoch": 2.7639415268002168, "grad_norm": 0.14023783802986145, "learning_rate": 3.9701329448187946e-05, "loss": 0.103, "step": 10210 }, { "epoch": 2.766648619382783, "grad_norm": 0.2275523692369461, "learning_rate": 3.9246038972864686e-05, "loss": 0.0832, "step": 10220 }, { "epoch": 2.7693557119653494, "grad_norm": 0.291933536529541, "learning_rate": 3.879074849754143e-05, "loss": 0.0805, "step": 10230 }, { "epoch": 2.7720628045479154, "grad_norm": 0.1608360856771469, "learning_rate": 3.833545802221817e-05, "loss": 0.1007, "step": 10240 }, { "epoch": 2.774769897130482, "grad_norm": 0.1045825183391571, "learning_rate": 3.788016754689492e-05, "loss": 0.068, "step": 10250 }, { "epoch": 2.7774769897130485, "grad_norm": 0.19174662232398987, "learning_rate": 3.742487707157166e-05, "loss": 0.0897, "step": 10260 }, { "epoch": 2.7801840822956145, "grad_norm": 0.19674013555049896, "learning_rate": 3.696958659624841e-05, "loss": 0.0949, "step": 10270 }, { "epoch": 2.7828911748781806, "grad_norm": 0.25490885972976685, "learning_rate": 3.6514296120925154e-05, "loss": 0.0948, "step": 10280 }, { "epoch": 2.785598267460747, "grad_norm": 0.20852668583393097, "learning_rate": 3.6059005645601894e-05, "loss": 0.111, "step": 10290 }, { "epoch": 2.7883053600433136, "grad_norm": 0.08679841458797455, "learning_rate": 3.560371517027864e-05, "loss": 0.1028, "step": 10300 }, { "epoch": 2.7883053600433136, "eval_loss": 0.09747523814439774, "eval_runtime": 41.8271, "eval_samples_per_second": 11.954, "eval_steps_per_second": 0.765, "step": 10300 }, { "epoch": 2.7910124526258797, "grad_norm": 0.2209625244140625, "learning_rate": 3.514842469495538e-05, "loss": 0.0938, "step": 10310 }, { "epoch": 2.7937195452084462, "grad_norm": 0.3107158839702606, "learning_rate": 3.469313421963213e-05, "loss": 0.0834, "step": 10320 }, { "epoch": 2.7964266377910123, "grad_norm": 0.15868419408798218, "learning_rate": 3.423784374430887e-05, "loss": 0.0913, "step": 10330 }, { "epoch": 2.799133730373579, "grad_norm": 0.11937720328569412, "learning_rate": 3.3782553268985615e-05, "loss": 0.1114, "step": 10340 }, { "epoch": 2.8018408229561453, "grad_norm": 0.26096081733703613, "learning_rate": 3.3327262793662355e-05, "loss": 0.0971, "step": 10350 }, { "epoch": 2.8045479155387114, "grad_norm": 0.12076021730899811, "learning_rate": 3.28719723183391e-05, "loss": 0.0876, "step": 10360 }, { "epoch": 2.8072550081212775, "grad_norm": 0.23484507203102112, "learning_rate": 3.241668184301585e-05, "loss": 0.0882, "step": 10370 }, { "epoch": 2.809962100703844, "grad_norm": 0.1609596312046051, "learning_rate": 3.196139136769259e-05, "loss": 0.0886, "step": 10380 }, { "epoch": 2.8126691932864105, "grad_norm": 0.34564417600631714, "learning_rate": 3.1506100892369335e-05, "loss": 0.0708, "step": 10390 }, { "epoch": 2.8153762858689766, "grad_norm": 0.1209503561258316, "learning_rate": 3.1050810417046075e-05, "loss": 0.0738, "step": 10400 }, { "epoch": 2.8153762858689766, "eval_loss": 0.09801511466503143, "eval_runtime": 43.3652, "eval_samples_per_second": 11.53, "eval_steps_per_second": 0.738, "step": 10400 }, { "epoch": 2.818083378451543, "grad_norm": 0.4219966530799866, "learning_rate": 3.059551994172282e-05, "loss": 0.103, "step": 10410 }, { "epoch": 2.820790471034109, "grad_norm": 0.1988462656736374, "learning_rate": 3.0140229466399562e-05, "loss": 0.068, "step": 10420 }, { "epoch": 2.8234975636166757, "grad_norm": 0.2200898379087448, "learning_rate": 2.9684938991076306e-05, "loss": 0.1016, "step": 10430 }, { "epoch": 2.826204656199242, "grad_norm": 0.2532331347465515, "learning_rate": 2.9229648515753052e-05, "loss": 0.09, "step": 10440 }, { "epoch": 2.8289117487818083, "grad_norm": 0.2740263044834137, "learning_rate": 2.8774358040429796e-05, "loss": 0.0714, "step": 10450 }, { "epoch": 2.831618841364375, "grad_norm": 0.16859589517116547, "learning_rate": 2.831906756510654e-05, "loss": 0.0935, "step": 10460 }, { "epoch": 2.834325933946941, "grad_norm": 0.061187438666820526, "learning_rate": 2.7863777089783283e-05, "loss": 0.078, "step": 10470 }, { "epoch": 2.8370330265295074, "grad_norm": 0.07519318908452988, "learning_rate": 2.7408486614460023e-05, "loss": 0.0938, "step": 10480 }, { "epoch": 2.8397401191120735, "grad_norm": 0.33607184886932373, "learning_rate": 2.695319613913677e-05, "loss": 0.1017, "step": 10490 }, { "epoch": 2.84244721169464, "grad_norm": 0.13333143293857574, "learning_rate": 2.6497905663813513e-05, "loss": 0.078, "step": 10500 }, { "epoch": 2.84244721169464, "eval_loss": 0.09900099039077759, "eval_runtime": 41.1162, "eval_samples_per_second": 12.161, "eval_steps_per_second": 0.778, "step": 10500 }, { "epoch": 2.845154304277206, "grad_norm": 0.2760447859764099, "learning_rate": 2.6042615188490257e-05, "loss": 0.0914, "step": 10510 }, { "epoch": 2.8478613968597726, "grad_norm": 0.21297326683998108, "learning_rate": 2.5587324713167e-05, "loss": 0.0859, "step": 10520 }, { "epoch": 2.850568489442339, "grad_norm": 0.3854238986968994, "learning_rate": 2.5132034237843747e-05, "loss": 0.0967, "step": 10530 }, { "epoch": 2.853275582024905, "grad_norm": 0.08490105718374252, "learning_rate": 2.467674376252049e-05, "loss": 0.0861, "step": 10540 }, { "epoch": 2.8559826746074717, "grad_norm": 0.22418226301670074, "learning_rate": 2.4221453287197234e-05, "loss": 0.0631, "step": 10550 }, { "epoch": 2.8586897671900378, "grad_norm": 0.28189709782600403, "learning_rate": 2.3766162811873974e-05, "loss": 0.0874, "step": 10560 }, { "epoch": 2.8613968597726043, "grad_norm": 0.4805290997028351, "learning_rate": 2.3310872336550717e-05, "loss": 0.0954, "step": 10570 }, { "epoch": 2.864103952355171, "grad_norm": 0.19959446787834167, "learning_rate": 2.2855581861227464e-05, "loss": 0.1145, "step": 10580 }, { "epoch": 2.866811044937737, "grad_norm": 0.3269408047199249, "learning_rate": 2.2400291385904208e-05, "loss": 0.0977, "step": 10590 }, { "epoch": 2.869518137520303, "grad_norm": 0.3688705265522003, "learning_rate": 2.194500091058095e-05, "loss": 0.0912, "step": 10600 }, { "epoch": 2.869518137520303, "eval_loss": 0.09780417382717133, "eval_runtime": 43.5267, "eval_samples_per_second": 11.487, "eval_steps_per_second": 0.735, "step": 10600 }, { "epoch": 2.8722252301028695, "grad_norm": 0.23457171022891998, "learning_rate": 2.1489710435257695e-05, "loss": 0.076, "step": 10610 }, { "epoch": 2.874932322685436, "grad_norm": 0.23011524975299835, "learning_rate": 2.103441995993444e-05, "loss": 0.1083, "step": 10620 }, { "epoch": 2.877639415268002, "grad_norm": 0.31096041202545166, "learning_rate": 2.057912948461118e-05, "loss": 0.0881, "step": 10630 }, { "epoch": 2.8803465078505686, "grad_norm": 0.5288337469100952, "learning_rate": 2.0123839009287925e-05, "loss": 0.1015, "step": 10640 }, { "epoch": 2.8830536004331346, "grad_norm": 0.1614571362733841, "learning_rate": 1.966854853396467e-05, "loss": 0.0845, "step": 10650 }, { "epoch": 2.885760693015701, "grad_norm": 0.16663025319576263, "learning_rate": 1.9213258058641412e-05, "loss": 0.083, "step": 10660 }, { "epoch": 2.8884677855982677, "grad_norm": 0.2802322506904602, "learning_rate": 1.875796758331816e-05, "loss": 0.1163, "step": 10670 }, { "epoch": 2.8911748781808337, "grad_norm": 0.07703015208244324, "learning_rate": 1.8302677107994902e-05, "loss": 0.0674, "step": 10680 }, { "epoch": 2.8938819707634003, "grad_norm": 0.23094356060028076, "learning_rate": 1.7847386632671646e-05, "loss": 0.0856, "step": 10690 }, { "epoch": 2.8965890633459663, "grad_norm": 0.2034844607114792, "learning_rate": 1.7392096157348386e-05, "loss": 0.1045, "step": 10700 }, { "epoch": 2.8965890633459663, "eval_loss": 0.09752561897039413, "eval_runtime": 43.2797, "eval_samples_per_second": 11.553, "eval_steps_per_second": 0.739, "step": 10700 }, { "epoch": 2.899296155928533, "grad_norm": 0.20917344093322754, "learning_rate": 1.6936805682025133e-05, "loss": 0.1056, "step": 10710 }, { "epoch": 2.902003248511099, "grad_norm": 0.16512975096702576, "learning_rate": 1.6481515206701876e-05, "loss": 0.0674, "step": 10720 }, { "epoch": 2.9047103410936654, "grad_norm": 0.26610878109931946, "learning_rate": 1.602622473137862e-05, "loss": 0.1196, "step": 10730 }, { "epoch": 2.9074174336762315, "grad_norm": 0.147952139377594, "learning_rate": 1.5570934256055363e-05, "loss": 0.1235, "step": 10740 }, { "epoch": 2.910124526258798, "grad_norm": 0.26650527119636536, "learning_rate": 1.5115643780732108e-05, "loss": 0.1028, "step": 10750 }, { "epoch": 2.9128316188413645, "grad_norm": 0.19182616472244263, "learning_rate": 1.466035330540885e-05, "loss": 0.0816, "step": 10760 }, { "epoch": 2.9155387114239306, "grad_norm": 0.3144613802433014, "learning_rate": 1.4205062830085595e-05, "loss": 0.0916, "step": 10770 }, { "epoch": 2.918245804006497, "grad_norm": 0.1899149864912033, "learning_rate": 1.3749772354762338e-05, "loss": 0.0885, "step": 10780 }, { "epoch": 2.920952896589063, "grad_norm": 0.24909132719039917, "learning_rate": 1.3294481879439084e-05, "loss": 0.1181, "step": 10790 }, { "epoch": 2.9236599891716297, "grad_norm": 0.1766766458749771, "learning_rate": 1.2839191404115825e-05, "loss": 0.0807, "step": 10800 }, { "epoch": 2.9236599891716297, "eval_loss": 0.09685847163200378, "eval_runtime": 41.64, "eval_samples_per_second": 12.008, "eval_steps_per_second": 0.768, "step": 10800 }, { "epoch": 2.9263670817541962, "grad_norm": 0.2989216446876526, "learning_rate": 1.238390092879257e-05, "loss": 0.0767, "step": 10810 }, { "epoch": 2.9290741743367623, "grad_norm": 0.08748666942119598, "learning_rate": 1.1928610453469314e-05, "loss": 0.09, "step": 10820 }, { "epoch": 2.9317812669193284, "grad_norm": 0.15734927356243134, "learning_rate": 1.1473319978146057e-05, "loss": 0.1013, "step": 10830 }, { "epoch": 2.934488359501895, "grad_norm": 0.3059326708316803, "learning_rate": 1.1018029502822801e-05, "loss": 0.1118, "step": 10840 }, { "epoch": 2.9371954520844614, "grad_norm": 0.2867346704006195, "learning_rate": 1.0562739027499544e-05, "loss": 0.1018, "step": 10850 }, { "epoch": 2.9399025446670275, "grad_norm": 0.29118990898132324, "learning_rate": 1.010744855217629e-05, "loss": 0.0947, "step": 10860 }, { "epoch": 2.942609637249594, "grad_norm": 0.11166248470544815, "learning_rate": 9.652158076853031e-06, "loss": 0.1019, "step": 10870 }, { "epoch": 2.94531672983216, "grad_norm": 0.14713464677333832, "learning_rate": 9.196867601529776e-06, "loss": 0.11, "step": 10880 }, { "epoch": 2.9480238224147266, "grad_norm": 0.5256966948509216, "learning_rate": 8.74157712620652e-06, "loss": 0.0949, "step": 10890 }, { "epoch": 2.950730914997293, "grad_norm": 0.13869787752628326, "learning_rate": 8.286286650883265e-06, "loss": 0.0768, "step": 10900 }, { "epoch": 2.950730914997293, "eval_loss": 0.09656227380037308, "eval_runtime": 43.6648, "eval_samples_per_second": 11.451, "eval_steps_per_second": 0.733, "step": 10900 }, { "epoch": 2.953438007579859, "grad_norm": 0.13716360926628113, "learning_rate": 7.830996175560007e-06, "loss": 0.0783, "step": 10910 }, { "epoch": 2.9561451001624257, "grad_norm": 0.19782617688179016, "learning_rate": 7.375705700236751e-06, "loss": 0.1113, "step": 10920 }, { "epoch": 2.958852192744992, "grad_norm": 0.31451940536499023, "learning_rate": 6.920415224913495e-06, "loss": 0.0948, "step": 10930 }, { "epoch": 2.9615592853275583, "grad_norm": 0.3306080996990204, "learning_rate": 6.465124749590239e-06, "loss": 0.0788, "step": 10940 }, { "epoch": 2.9642663779101244, "grad_norm": 0.24199481308460236, "learning_rate": 6.009834274266983e-06, "loss": 0.0933, "step": 10950 }, { "epoch": 2.966973470492691, "grad_norm": 0.2815631330013275, "learning_rate": 5.554543798943726e-06, "loss": 0.1134, "step": 10960 }, { "epoch": 2.969680563075257, "grad_norm": 0.08186428248882294, "learning_rate": 5.09925332362047e-06, "loss": 0.1051, "step": 10970 }, { "epoch": 2.9723876556578235, "grad_norm": 0.2861514091491699, "learning_rate": 4.6439628482972135e-06, "loss": 0.0953, "step": 10980 }, { "epoch": 2.97509474824039, "grad_norm": 0.13638444244861603, "learning_rate": 4.188672372973958e-06, "loss": 0.0813, "step": 10990 }, { "epoch": 2.977801840822956, "grad_norm": 0.10792617499828339, "learning_rate": 3.7333818976507013e-06, "loss": 0.0731, "step": 11000 }, { "epoch": 2.977801840822956, "eval_loss": 0.09705805778503418, "eval_runtime": 42.0021, "eval_samples_per_second": 11.904, "eval_steps_per_second": 0.762, "step": 11000 }, { "epoch": 2.9805089334055226, "grad_norm": 0.4235360622406006, "learning_rate": 3.278091422327445e-06, "loss": 0.1128, "step": 11010 }, { "epoch": 2.9832160259880887, "grad_norm": 0.12676456570625305, "learning_rate": 2.8228009470041886e-06, "loss": 0.0874, "step": 11020 }, { "epoch": 2.985923118570655, "grad_norm": 0.25628066062927246, "learning_rate": 2.3675104716809325e-06, "loss": 0.1014, "step": 11030 }, { "epoch": 2.9886302111532217, "grad_norm": 0.18908998370170593, "learning_rate": 1.9122199963576764e-06, "loss": 0.0964, "step": 11040 }, { "epoch": 2.9913373037357878, "grad_norm": 0.19528800249099731, "learning_rate": 1.4569295210344198e-06, "loss": 0.0932, "step": 11050 }, { "epoch": 2.994044396318354, "grad_norm": 0.2792895436286926, "learning_rate": 1.0016390457111637e-06, "loss": 0.0979, "step": 11060 }, { "epoch": 2.9967514889009204, "grad_norm": 0.11036554723978043, "learning_rate": 5.463485703879075e-07, "loss": 0.0796, "step": 11070 }, { "epoch": 2.999458581483487, "grad_norm": 0.1410888433456421, "learning_rate": 9.105809506465124e-08, "loss": 0.0911, "step": 11080 } ], "logging_steps": 10, "max_steps": 11082, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.037288867508432e+18, "train_batch_size": 32, "trial_name": null, "trial_params": null }