{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 100, "global_step": 11082, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0027070925825663237, "grad_norm": 4.34375, "learning_rate": 5e-05, "loss": 1.3537, "step": 10 }, { "epoch": 0.005414185165132647, "grad_norm": 2.515625, "learning_rate": 0.0001, "loss": 1.0202, "step": 20 }, { "epoch": 0.008121277747698972, "grad_norm": 1.3984375, "learning_rate": 0.00015, "loss": 0.8165, "step": 30 }, { "epoch": 0.010828370330265295, "grad_norm": 0.77734375, "learning_rate": 0.0002, "loss": 0.789, "step": 40 }, { "epoch": 0.01353546291283162, "grad_norm": 1.3359375, "learning_rate": 0.00025, "loss": 0.7576, "step": 50 }, { "epoch": 0.016242555495397944, "grad_norm": 0.8828125, "learning_rate": 0.0003, "loss": 0.7408, "step": 60 }, { "epoch": 0.018949648077964266, "grad_norm": 0.80859375, "learning_rate": 0.00035, "loss": 0.7243, "step": 70 }, { "epoch": 0.02165674066053059, "grad_norm": 0.625, "learning_rate": 0.0004, "loss": 0.708, "step": 80 }, { "epoch": 0.024363833243096916, "grad_norm": 1.0625, "learning_rate": 0.00045000000000000004, "loss": 0.6916, "step": 90 }, { "epoch": 0.02707092582566324, "grad_norm": 0.6796875, "learning_rate": 0.0005, "loss": 0.6795, "step": 100 }, { "epoch": 0.02707092582566324, "eval_loss": 0.7104170918464661, "eval_runtime": 52.5058, "eval_samples_per_second": 9.523, "eval_steps_per_second": 0.305, "step": 100 }, { "epoch": 0.02977801840822956, "grad_norm": 0.59375, "learning_rate": 0.0004995447095246767, "loss": 0.606, "step": 110 }, { "epoch": 0.03248511099079589, "grad_norm": 0.7421875, "learning_rate": 0.0004990894190493535, "loss": 0.492, "step": 120 }, { "epoch": 0.03519220357336221, "grad_norm": 0.98828125, "learning_rate": 0.0004986341285740302, "loss": 0.4467, "step": 130 }, { "epoch": 0.03789929615592853, "grad_norm": 0.625, "learning_rate": 0.000498178838098707, "loss": 0.3712, "step": 140 }, { "epoch": 0.040606388738494856, "grad_norm": 0.353515625, "learning_rate": 0.0004977235476233838, "loss": 0.3499, "step": 150 }, { "epoch": 0.04331348132106118, "grad_norm": 0.6484375, "learning_rate": 0.0004972682571480605, "loss": 0.3513, "step": 160 }, { "epoch": 0.0460205739036275, "grad_norm": 0.59375, "learning_rate": 0.0004968129666727372, "loss": 0.3074, "step": 170 }, { "epoch": 0.04872766648619383, "grad_norm": 0.5859375, "learning_rate": 0.0004963576761974139, "loss": 0.3166, "step": 180 }, { "epoch": 0.051434759068760154, "grad_norm": 0.3671875, "learning_rate": 0.0004959023857220907, "loss": 0.3321, "step": 190 }, { "epoch": 0.05414185165132648, "grad_norm": 0.37890625, "learning_rate": 0.0004954470952467675, "loss": 0.2764, "step": 200 }, { "epoch": 0.05414185165132648, "eval_loss": 0.2984282970428467, "eval_runtime": 51.6076, "eval_samples_per_second": 9.688, "eval_steps_per_second": 0.31, "step": 200 }, { "epoch": 0.0568489442338928, "grad_norm": 0.328125, "learning_rate": 0.0004949918047714442, "loss": 0.3399, "step": 210 }, { "epoch": 0.05955603681645912, "grad_norm": 0.484375, "learning_rate": 0.0004945365142961209, "loss": 0.2881, "step": 220 }, { "epoch": 0.062263129399025445, "grad_norm": 0.625, "learning_rate": 0.0004940812238207977, "loss": 0.2848, "step": 230 }, { "epoch": 0.06497022198159177, "grad_norm": 0.4453125, "learning_rate": 0.0004936259333454744, "loss": 0.2854, "step": 240 }, { "epoch": 0.0676773145641581, "grad_norm": 0.36328125, "learning_rate": 0.0004931706428701511, "loss": 0.2628, "step": 250 }, { "epoch": 0.07038440714672442, "grad_norm": 0.7109375, "learning_rate": 0.0004927153523948279, "loss": 0.2686, "step": 260 }, { "epoch": 0.07309149972929074, "grad_norm": 0.3203125, "learning_rate": 0.0004922600619195047, "loss": 0.2626, "step": 270 }, { "epoch": 0.07579859231185707, "grad_norm": 0.5859375, "learning_rate": 0.0004918047714441814, "loss": 0.2475, "step": 280 }, { "epoch": 0.07850568489442339, "grad_norm": 0.29296875, "learning_rate": 0.0004913494809688581, "loss": 0.2104, "step": 290 }, { "epoch": 0.08121277747698971, "grad_norm": 0.7421875, "learning_rate": 0.0004908941904935349, "loss": 0.2537, "step": 300 }, { "epoch": 0.08121277747698971, "eval_loss": 0.27854180335998535, "eval_runtime": 50.3171, "eval_samples_per_second": 9.937, "eval_steps_per_second": 0.318, "step": 300 }, { "epoch": 0.08391987005955603, "grad_norm": 0.5859375, "learning_rate": 0.0004904389000182116, "loss": 0.2649, "step": 310 }, { "epoch": 0.08662696264212236, "grad_norm": 0.59765625, "learning_rate": 0.0004899836095428884, "loss": 0.2188, "step": 320 }, { "epoch": 0.08933405522468868, "grad_norm": 0.3359375, "learning_rate": 0.0004895283190675652, "loss": 0.2715, "step": 330 }, { "epoch": 0.092041147807255, "grad_norm": 0.349609375, "learning_rate": 0.0004890730285922419, "loss": 0.2425, "step": 340 }, { "epoch": 0.09474824038982133, "grad_norm": 0.451171875, "learning_rate": 0.0004886177381169186, "loss": 0.2697, "step": 350 }, { "epoch": 0.09745533297238766, "grad_norm": 0.51953125, "learning_rate": 0.00048816244764159533, "loss": 0.2589, "step": 360 }, { "epoch": 0.10016242555495398, "grad_norm": 0.279296875, "learning_rate": 0.00048770715716627206, "loss": 0.2436, "step": 370 }, { "epoch": 0.10286951813752031, "grad_norm": 0.26171875, "learning_rate": 0.00048725186669094884, "loss": 0.2405, "step": 380 }, { "epoch": 0.10557661072008663, "grad_norm": 0.1845703125, "learning_rate": 0.00048679657621562557, "loss": 0.1929, "step": 390 }, { "epoch": 0.10828370330265295, "grad_norm": 0.5859375, "learning_rate": 0.00048634128574030235, "loss": 0.2524, "step": 400 }, { "epoch": 0.10828370330265295, "eval_loss": 0.2730386257171631, "eval_runtime": 50.4403, "eval_samples_per_second": 9.913, "eval_steps_per_second": 0.317, "step": 400 }, { "epoch": 0.11099079588521928, "grad_norm": 0.474609375, "learning_rate": 0.0004858859952649791, "loss": 0.2384, "step": 410 }, { "epoch": 0.1136978884677856, "grad_norm": 0.640625, "learning_rate": 0.0004854307047896558, "loss": 0.2345, "step": 420 }, { "epoch": 0.11640498105035192, "grad_norm": 0.490234375, "learning_rate": 0.0004849754143143326, "loss": 0.2222, "step": 430 }, { "epoch": 0.11911207363291824, "grad_norm": 0.2890625, "learning_rate": 0.0004845201238390093, "loss": 0.2448, "step": 440 }, { "epoch": 0.12181916621548457, "grad_norm": 0.306640625, "learning_rate": 0.0004840648333636861, "loss": 0.2293, "step": 450 }, { "epoch": 0.12452625879805089, "grad_norm": 0.3671875, "learning_rate": 0.00048360954288836276, "loss": 0.2342, "step": 460 }, { "epoch": 0.12723335138061723, "grad_norm": 0.181640625, "learning_rate": 0.0004831542524130395, "loss": 0.2127, "step": 470 }, { "epoch": 0.12994044396318355, "grad_norm": 0.2080078125, "learning_rate": 0.00048269896193771627, "loss": 0.2081, "step": 480 }, { "epoch": 0.13264753654574987, "grad_norm": 0.3203125, "learning_rate": 0.000482243671462393, "loss": 0.2002, "step": 490 }, { "epoch": 0.1353546291283162, "grad_norm": 0.4296875, "learning_rate": 0.0004817883809870698, "loss": 0.2682, "step": 500 }, { "epoch": 0.1353546291283162, "eval_loss": 0.2406972348690033, "eval_runtime": 50.5337, "eval_samples_per_second": 9.894, "eval_steps_per_second": 0.317, "step": 500 }, { "epoch": 0.13806172171088252, "grad_norm": 0.453125, "learning_rate": 0.0004813330905117465, "loss": 0.2076, "step": 510 }, { "epoch": 0.14076881429344884, "grad_norm": 0.412109375, "learning_rate": 0.00048087780003642323, "loss": 0.2251, "step": 520 }, { "epoch": 0.14347590687601516, "grad_norm": 0.431640625, "learning_rate": 0.0004804225095611, "loss": 0.2211, "step": 530 }, { "epoch": 0.1461829994585815, "grad_norm": 0.83203125, "learning_rate": 0.00047996721908577674, "loss": 0.2015, "step": 540 }, { "epoch": 0.1488900920411478, "grad_norm": 0.388671875, "learning_rate": 0.00047951192861045346, "loss": 0.2195, "step": 550 }, { "epoch": 0.15159718462371413, "grad_norm": 0.47265625, "learning_rate": 0.00047905663813513024, "loss": 0.2289, "step": 560 }, { "epoch": 0.15430427720628045, "grad_norm": 0.287109375, "learning_rate": 0.00047860134765980697, "loss": 0.2015, "step": 570 }, { "epoch": 0.15701136978884678, "grad_norm": 0.49609375, "learning_rate": 0.00047814605718448375, "loss": 0.2575, "step": 580 }, { "epoch": 0.1597184623714131, "grad_norm": 0.2333984375, "learning_rate": 0.0004776907667091605, "loss": 0.2308, "step": 590 }, { "epoch": 0.16242555495397942, "grad_norm": 0.322265625, "learning_rate": 0.00047723547623383715, "loss": 0.166, "step": 600 }, { "epoch": 0.16242555495397942, "eval_loss": 0.2522674798965454, "eval_runtime": 50.2817, "eval_samples_per_second": 9.944, "eval_steps_per_second": 0.318, "step": 600 }, { "epoch": 0.16513264753654575, "grad_norm": 0.353515625, "learning_rate": 0.00047678018575851393, "loss": 0.2034, "step": 610 }, { "epoch": 0.16783974011911207, "grad_norm": 0.337890625, "learning_rate": 0.00047632489528319066, "loss": 0.2435, "step": 620 }, { "epoch": 0.1705468327016784, "grad_norm": 0.44921875, "learning_rate": 0.00047586960480786744, "loss": 0.1984, "step": 630 }, { "epoch": 0.17325392528424471, "grad_norm": 0.296875, "learning_rate": 0.00047541431433254416, "loss": 0.2178, "step": 640 }, { "epoch": 0.17596101786681104, "grad_norm": 0.3828125, "learning_rate": 0.0004749590238572209, "loss": 0.2117, "step": 650 }, { "epoch": 0.17866811044937736, "grad_norm": 0.359375, "learning_rate": 0.00047450373338189767, "loss": 0.2056, "step": 660 }, { "epoch": 0.18137520303194368, "grad_norm": 0.287109375, "learning_rate": 0.0004740484429065744, "loss": 0.171, "step": 670 }, { "epoch": 0.18408229561451, "grad_norm": 0.6015625, "learning_rate": 0.0004735931524312512, "loss": 0.2604, "step": 680 }, { "epoch": 0.18678938819707633, "grad_norm": 0.306640625, "learning_rate": 0.0004731378619559279, "loss": 0.2011, "step": 690 }, { "epoch": 0.18949648077964265, "grad_norm": 0.90234375, "learning_rate": 0.00047268257148060463, "loss": 0.1928, "step": 700 }, { "epoch": 0.18949648077964265, "eval_loss": 0.25613531470298767, "eval_runtime": 49.6244, "eval_samples_per_second": 10.076, "eval_steps_per_second": 0.322, "step": 700 }, { "epoch": 0.19220357336220897, "grad_norm": 0.56640625, "learning_rate": 0.0004722272810052814, "loss": 0.2571, "step": 710 }, { "epoch": 0.19491066594477532, "grad_norm": 0.341796875, "learning_rate": 0.00047177199052995814, "loss": 0.2028, "step": 720 }, { "epoch": 0.19761775852734165, "grad_norm": 0.1748046875, "learning_rate": 0.00047131670005463487, "loss": 0.2207, "step": 730 }, { "epoch": 0.20032485110990797, "grad_norm": 0.244140625, "learning_rate": 0.0004708614095793116, "loss": 0.1788, "step": 740 }, { "epoch": 0.2030319436924743, "grad_norm": 0.322265625, "learning_rate": 0.0004704061191039883, "loss": 0.1913, "step": 750 }, { "epoch": 0.20573903627504062, "grad_norm": 0.32421875, "learning_rate": 0.0004699508286286651, "loss": 0.2213, "step": 760 }, { "epoch": 0.20844612885760694, "grad_norm": 0.4140625, "learning_rate": 0.0004694955381533418, "loss": 0.1681, "step": 770 }, { "epoch": 0.21115322144017326, "grad_norm": 0.373046875, "learning_rate": 0.00046904024767801855, "loss": 0.1984, "step": 780 }, { "epoch": 0.21386031402273958, "grad_norm": 0.52734375, "learning_rate": 0.00046858495720269533, "loss": 0.2184, "step": 790 }, { "epoch": 0.2165674066053059, "grad_norm": 0.203125, "learning_rate": 0.00046812966672737206, "loss": 0.1834, "step": 800 }, { "epoch": 0.2165674066053059, "eval_loss": 0.229017972946167, "eval_runtime": 50.6851, "eval_samples_per_second": 9.865, "eval_steps_per_second": 0.316, "step": 800 }, { "epoch": 0.21927449918787223, "grad_norm": 0.416015625, "learning_rate": 0.00046767437625204884, "loss": 0.1681, "step": 810 }, { "epoch": 0.22198159177043855, "grad_norm": 0.34375, "learning_rate": 0.00046721908577672557, "loss": 0.1968, "step": 820 }, { "epoch": 0.22468868435300487, "grad_norm": 0.251953125, "learning_rate": 0.0004667637953014023, "loss": 0.1499, "step": 830 }, { "epoch": 0.2273957769355712, "grad_norm": 0.275390625, "learning_rate": 0.0004663085048260791, "loss": 0.1987, "step": 840 }, { "epoch": 0.23010286951813752, "grad_norm": 0.408203125, "learning_rate": 0.0004658532143507558, "loss": 0.2192, "step": 850 }, { "epoch": 0.23280996210070384, "grad_norm": 0.3203125, "learning_rate": 0.0004653979238754326, "loss": 0.1981, "step": 860 }, { "epoch": 0.23551705468327017, "grad_norm": 0.267578125, "learning_rate": 0.0004649426334001093, "loss": 0.1985, "step": 870 }, { "epoch": 0.2382241472658365, "grad_norm": 0.1787109375, "learning_rate": 0.000464487342924786, "loss": 0.1495, "step": 880 }, { "epoch": 0.2409312398484028, "grad_norm": 0.361328125, "learning_rate": 0.00046403205244946276, "loss": 0.1736, "step": 890 }, { "epoch": 0.24363833243096913, "grad_norm": 0.19921875, "learning_rate": 0.0004635767619741395, "loss": 0.1499, "step": 900 }, { "epoch": 0.24363833243096913, "eval_loss": 0.240362286567688, "eval_runtime": 50.0701, "eval_samples_per_second": 9.986, "eval_steps_per_second": 0.32, "step": 900 }, { "epoch": 0.24634542501353546, "grad_norm": 0.484375, "learning_rate": 0.0004631214714988162, "loss": 0.1793, "step": 910 }, { "epoch": 0.24905251759610178, "grad_norm": 0.34765625, "learning_rate": 0.000462666181023493, "loss": 0.2043, "step": 920 }, { "epoch": 0.2517596101786681, "grad_norm": 0.51171875, "learning_rate": 0.0004622108905481697, "loss": 0.211, "step": 930 }, { "epoch": 0.25446670276123445, "grad_norm": 0.53125, "learning_rate": 0.0004617556000728465, "loss": 0.2002, "step": 940 }, { "epoch": 0.25717379534380075, "grad_norm": 0.1435546875, "learning_rate": 0.00046130030959752323, "loss": 0.1758, "step": 950 }, { "epoch": 0.2598808879263671, "grad_norm": 0.263671875, "learning_rate": 0.00046084501912219995, "loss": 0.1695, "step": 960 }, { "epoch": 0.2625879805089334, "grad_norm": 0.29296875, "learning_rate": 0.00046038972864687674, "loss": 0.184, "step": 970 }, { "epoch": 0.26529507309149974, "grad_norm": 0.1103515625, "learning_rate": 0.00045993443817155346, "loss": 0.177, "step": 980 }, { "epoch": 0.26800216567406604, "grad_norm": 0.138671875, "learning_rate": 0.00045947914769623024, "loss": 0.1733, "step": 990 }, { "epoch": 0.2707092582566324, "grad_norm": 0.2080078125, "learning_rate": 0.00045902385722090697, "loss": 0.1469, "step": 1000 }, { "epoch": 0.2707092582566324, "eval_loss": 0.24039389193058014, "eval_runtime": 50.3119, "eval_samples_per_second": 9.938, "eval_steps_per_second": 0.318, "step": 1000 }, { "epoch": 0.2734163508391987, "grad_norm": 0.46875, "learning_rate": 0.0004585685667455837, "loss": 0.1747, "step": 1010 }, { "epoch": 0.27612344342176504, "grad_norm": 0.375, "learning_rate": 0.0004581132762702604, "loss": 0.1451, "step": 1020 }, { "epoch": 0.27883053600433133, "grad_norm": 0.375, "learning_rate": 0.00045765798579493715, "loss": 0.1525, "step": 1030 }, { "epoch": 0.2815376285868977, "grad_norm": 0.185546875, "learning_rate": 0.00045720269531961393, "loss": 0.1606, "step": 1040 }, { "epoch": 0.284244721169464, "grad_norm": 0.56640625, "learning_rate": 0.00045674740484429066, "loss": 0.1908, "step": 1050 }, { "epoch": 0.2869518137520303, "grad_norm": 0.158203125, "learning_rate": 0.0004562921143689674, "loss": 0.1894, "step": 1060 }, { "epoch": 0.2896589063345966, "grad_norm": 0.35546875, "learning_rate": 0.00045583682389364416, "loss": 0.1646, "step": 1070 }, { "epoch": 0.292365998917163, "grad_norm": 0.419921875, "learning_rate": 0.0004553815334183209, "loss": 0.1536, "step": 1080 }, { "epoch": 0.29507309149972927, "grad_norm": 0.2734375, "learning_rate": 0.0004549262429429976, "loss": 0.1654, "step": 1090 }, { "epoch": 0.2977801840822956, "grad_norm": 0.318359375, "learning_rate": 0.0004544709524676744, "loss": 0.1598, "step": 1100 }, { "epoch": 0.2977801840822956, "eval_loss": 0.22400033473968506, "eval_runtime": 50.8906, "eval_samples_per_second": 9.825, "eval_steps_per_second": 0.314, "step": 1100 }, { "epoch": 0.3004872766648619, "grad_norm": 0.24609375, "learning_rate": 0.0004540156619923511, "loss": 0.2339, "step": 1110 }, { "epoch": 0.30319436924742826, "grad_norm": 0.2275390625, "learning_rate": 0.0004535603715170279, "loss": 0.1969, "step": 1120 }, { "epoch": 0.30590146182999456, "grad_norm": 0.18359375, "learning_rate": 0.00045310508104170463, "loss": 0.167, "step": 1130 }, { "epoch": 0.3086085544125609, "grad_norm": 0.4765625, "learning_rate": 0.00045264979056638136, "loss": 0.1763, "step": 1140 }, { "epoch": 0.31131564699512726, "grad_norm": 0.314453125, "learning_rate": 0.00045219450009105814, "loss": 0.14, "step": 1150 }, { "epoch": 0.31402273957769355, "grad_norm": 0.326171875, "learning_rate": 0.0004517392096157348, "loss": 0.1583, "step": 1160 }, { "epoch": 0.3167298321602599, "grad_norm": 0.53515625, "learning_rate": 0.0004512839191404116, "loss": 0.1911, "step": 1170 }, { "epoch": 0.3194369247428262, "grad_norm": 0.384765625, "learning_rate": 0.0004508286286650883, "loss": 0.2127, "step": 1180 }, { "epoch": 0.32214401732539255, "grad_norm": 0.19140625, "learning_rate": 0.00045037333818976504, "loss": 0.1617, "step": 1190 }, { "epoch": 0.32485110990795885, "grad_norm": 0.41015625, "learning_rate": 0.0004499180477144418, "loss": 0.146, "step": 1200 }, { "epoch": 0.32485110990795885, "eval_loss": 0.21456098556518555, "eval_runtime": 51.3811, "eval_samples_per_second": 9.731, "eval_steps_per_second": 0.311, "step": 1200 }, { "epoch": 0.3275582024905252, "grad_norm": 0.224609375, "learning_rate": 0.00044946275723911855, "loss": 0.1529, "step": 1210 }, { "epoch": 0.3302652950730915, "grad_norm": 0.31640625, "learning_rate": 0.00044900746676379533, "loss": 0.1624, "step": 1220 }, { "epoch": 0.33297238765565784, "grad_norm": 0.2275390625, "learning_rate": 0.00044855217628847206, "loss": 0.2105, "step": 1230 }, { "epoch": 0.33567948023822414, "grad_norm": 0.166015625, "learning_rate": 0.0004480968858131488, "loss": 0.1419, "step": 1240 }, { "epoch": 0.3383865728207905, "grad_norm": 0.41015625, "learning_rate": 0.00044764159533782557, "loss": 0.1838, "step": 1250 }, { "epoch": 0.3410936654033568, "grad_norm": 0.234375, "learning_rate": 0.0004471863048625023, "loss": 0.1583, "step": 1260 }, { "epoch": 0.34380075798592313, "grad_norm": 0.30859375, "learning_rate": 0.000446731014387179, "loss": 0.1838, "step": 1270 }, { "epoch": 0.34650785056848943, "grad_norm": 0.23046875, "learning_rate": 0.0004462757239118558, "loss": 0.1976, "step": 1280 }, { "epoch": 0.3492149431510558, "grad_norm": 0.4609375, "learning_rate": 0.0004458204334365325, "loss": 0.1799, "step": 1290 }, { "epoch": 0.3519220357336221, "grad_norm": 0.515625, "learning_rate": 0.0004453651429612093, "loss": 0.2087, "step": 1300 }, { "epoch": 0.3519220357336221, "eval_loss": 0.20530980825424194, "eval_runtime": 51.1083, "eval_samples_per_second": 9.783, "eval_steps_per_second": 0.313, "step": 1300 }, { "epoch": 0.3546291283161884, "grad_norm": 0.1318359375, "learning_rate": 0.000444909852485886, "loss": 0.174, "step": 1310 }, { "epoch": 0.3573362208987547, "grad_norm": 0.390625, "learning_rate": 0.0004444545620105627, "loss": 0.1743, "step": 1320 }, { "epoch": 0.36004331348132107, "grad_norm": 0.36328125, "learning_rate": 0.0004439992715352395, "loss": 0.188, "step": 1330 }, { "epoch": 0.36275040606388737, "grad_norm": 0.330078125, "learning_rate": 0.0004435439810599162, "loss": 0.1952, "step": 1340 }, { "epoch": 0.3654574986464537, "grad_norm": 0.28515625, "learning_rate": 0.000443088690584593, "loss": 0.1742, "step": 1350 }, { "epoch": 0.36816459122902, "grad_norm": 0.2138671875, "learning_rate": 0.0004426334001092697, "loss": 0.1285, "step": 1360 }, { "epoch": 0.37087168381158636, "grad_norm": 0.2578125, "learning_rate": 0.00044217810963394645, "loss": 0.1621, "step": 1370 }, { "epoch": 0.37357877639415266, "grad_norm": 0.38671875, "learning_rate": 0.0004417228191586232, "loss": 0.1448, "step": 1380 }, { "epoch": 0.376285868976719, "grad_norm": 0.28125, "learning_rate": 0.00044126752868329995, "loss": 0.1837, "step": 1390 }, { "epoch": 0.3789929615592853, "grad_norm": 0.16015625, "learning_rate": 0.00044081223820797673, "loss": 0.1831, "step": 1400 }, { "epoch": 0.3789929615592853, "eval_loss": 0.19879956543445587, "eval_runtime": 49.8699, "eval_samples_per_second": 10.026, "eval_steps_per_second": 0.321, "step": 1400 }, { "epoch": 0.38170005414185165, "grad_norm": 0.271484375, "learning_rate": 0.00044035694773265346, "loss": 0.1556, "step": 1410 }, { "epoch": 0.38440714672441795, "grad_norm": 0.2412109375, "learning_rate": 0.0004399016572573302, "loss": 0.1856, "step": 1420 }, { "epoch": 0.3871142393069843, "grad_norm": 0.341796875, "learning_rate": 0.00043944636678200697, "loss": 0.1639, "step": 1430 }, { "epoch": 0.38982133188955065, "grad_norm": 0.2734375, "learning_rate": 0.0004389910763066837, "loss": 0.1609, "step": 1440 }, { "epoch": 0.39252842447211694, "grad_norm": 0.1865234375, "learning_rate": 0.00043853578583136037, "loss": 0.1906, "step": 1450 }, { "epoch": 0.3952355170546833, "grad_norm": 0.25390625, "learning_rate": 0.00043808049535603715, "loss": 0.1423, "step": 1460 }, { "epoch": 0.3979426096372496, "grad_norm": 0.248046875, "learning_rate": 0.0004376252048807139, "loss": 0.1549, "step": 1470 }, { "epoch": 0.40064970221981594, "grad_norm": 0.25, "learning_rate": 0.00043716991440539065, "loss": 0.1808, "step": 1480 }, { "epoch": 0.40335679480238223, "grad_norm": 0.2392578125, "learning_rate": 0.0004367146239300674, "loss": 0.1612, "step": 1490 }, { "epoch": 0.4060638873849486, "grad_norm": 0.1982421875, "learning_rate": 0.0004362593334547441, "loss": 0.1655, "step": 1500 }, { "epoch": 0.4060638873849486, "eval_loss": 0.20820540189743042, "eval_runtime": 49.8508, "eval_samples_per_second": 10.03, "eval_steps_per_second": 0.321, "step": 1500 }, { "epoch": 0.4087709799675149, "grad_norm": 0.40625, "learning_rate": 0.0004358040429794209, "loss": 0.172, "step": 1510 }, { "epoch": 0.41147807255008123, "grad_norm": 0.439453125, "learning_rate": 0.0004353487525040976, "loss": 0.1782, "step": 1520 }, { "epoch": 0.4141851651326475, "grad_norm": 0.4140625, "learning_rate": 0.0004348934620287744, "loss": 0.2182, "step": 1530 }, { "epoch": 0.4168922577152139, "grad_norm": 0.205078125, "learning_rate": 0.0004344381715534511, "loss": 0.1459, "step": 1540 }, { "epoch": 0.41959935029778017, "grad_norm": 0.361328125, "learning_rate": 0.00043398288107812785, "loss": 0.165, "step": 1550 }, { "epoch": 0.4223064428803465, "grad_norm": 0.3203125, "learning_rate": 0.00043352759060280463, "loss": 0.1253, "step": 1560 }, { "epoch": 0.4250135354629128, "grad_norm": 0.17578125, "learning_rate": 0.00043307230012748136, "loss": 0.1682, "step": 1570 }, { "epoch": 0.42772062804547917, "grad_norm": 0.474609375, "learning_rate": 0.00043261700965215814, "loss": 0.1921, "step": 1580 }, { "epoch": 0.43042772062804546, "grad_norm": 0.326171875, "learning_rate": 0.0004321617191768348, "loss": 0.1683, "step": 1590 }, { "epoch": 0.4331348132106118, "grad_norm": 0.1767578125, "learning_rate": 0.00043170642870151154, "loss": 0.1625, "step": 1600 }, { "epoch": 0.4331348132106118, "eval_loss": 0.2006607949733734, "eval_runtime": 50.6005, "eval_samples_per_second": 9.881, "eval_steps_per_second": 0.316, "step": 1600 }, { "epoch": 0.4358419057931781, "grad_norm": 0.11279296875, "learning_rate": 0.0004312511382261883, "loss": 0.1418, "step": 1610 }, { "epoch": 0.43854899837574446, "grad_norm": 0.177734375, "learning_rate": 0.00043079584775086504, "loss": 0.1838, "step": 1620 }, { "epoch": 0.44125609095831075, "grad_norm": 0.2265625, "learning_rate": 0.00043034055727554177, "loss": 0.1515, "step": 1630 }, { "epoch": 0.4439631835408771, "grad_norm": 0.578125, "learning_rate": 0.00042988526680021855, "loss": 0.1601, "step": 1640 }, { "epoch": 0.4466702761234434, "grad_norm": 0.58203125, "learning_rate": 0.0004294299763248953, "loss": 0.1776, "step": 1650 }, { "epoch": 0.44937736870600975, "grad_norm": 0.2314453125, "learning_rate": 0.00042897468584957206, "loss": 0.1732, "step": 1660 }, { "epoch": 0.45208446128857604, "grad_norm": 0.33984375, "learning_rate": 0.0004285193953742488, "loss": 0.1717, "step": 1670 }, { "epoch": 0.4547915538711424, "grad_norm": 0.6015625, "learning_rate": 0.0004280641048989255, "loss": 0.1935, "step": 1680 }, { "epoch": 0.4574986464537087, "grad_norm": 0.154296875, "learning_rate": 0.0004276088144236023, "loss": 0.1489, "step": 1690 }, { "epoch": 0.46020573903627504, "grad_norm": 0.11962890625, "learning_rate": 0.000427153523948279, "loss": 0.1372, "step": 1700 }, { "epoch": 0.46020573903627504, "eval_loss": 0.2127913236618042, "eval_runtime": 51.3425, "eval_samples_per_second": 9.739, "eval_steps_per_second": 0.312, "step": 1700 }, { "epoch": 0.4629128316188414, "grad_norm": 0.298828125, "learning_rate": 0.0004266982334729558, "loss": 0.17, "step": 1710 }, { "epoch": 0.4656199242014077, "grad_norm": 0.224609375, "learning_rate": 0.0004262429429976325, "loss": 0.1759, "step": 1720 }, { "epoch": 0.46832701678397404, "grad_norm": 0.2353515625, "learning_rate": 0.0004257876525223092, "loss": 0.1776, "step": 1730 }, { "epoch": 0.47103410936654033, "grad_norm": 0.2060546875, "learning_rate": 0.000425332362046986, "loss": 0.1677, "step": 1740 }, { "epoch": 0.4737412019491067, "grad_norm": 0.255859375, "learning_rate": 0.0004248770715716627, "loss": 0.1782, "step": 1750 }, { "epoch": 0.476448294531673, "grad_norm": 0.2314453125, "learning_rate": 0.0004244217810963395, "loss": 0.1623, "step": 1760 }, { "epoch": 0.47915538711423933, "grad_norm": 0.322265625, "learning_rate": 0.0004239664906210162, "loss": 0.1823, "step": 1770 }, { "epoch": 0.4818624796968056, "grad_norm": 0.134765625, "learning_rate": 0.00042351120014569294, "loss": 0.1942, "step": 1780 }, { "epoch": 0.484569572279372, "grad_norm": 0.275390625, "learning_rate": 0.0004230559096703697, "loss": 0.1567, "step": 1790 }, { "epoch": 0.48727666486193827, "grad_norm": 0.41796875, "learning_rate": 0.00042260061919504645, "loss": 0.1743, "step": 1800 }, { "epoch": 0.48727666486193827, "eval_loss": 0.19061319530010223, "eval_runtime": 51.1815, "eval_samples_per_second": 9.769, "eval_steps_per_second": 0.313, "step": 1800 }, { "epoch": 0.4899837574445046, "grad_norm": 0.2158203125, "learning_rate": 0.00042214532871972317, "loss": 0.1434, "step": 1810 }, { "epoch": 0.4926908500270709, "grad_norm": 0.283203125, "learning_rate": 0.00042169003824439995, "loss": 0.1843, "step": 1820 }, { "epoch": 0.49539794260963727, "grad_norm": 0.58203125, "learning_rate": 0.0004212347477690767, "loss": 0.1689, "step": 1830 }, { "epoch": 0.49810503519220356, "grad_norm": 0.220703125, "learning_rate": 0.00042077945729375346, "loss": 0.1576, "step": 1840 }, { "epoch": 0.5008121277747699, "grad_norm": 0.287109375, "learning_rate": 0.0004203241668184302, "loss": 0.1912, "step": 1850 }, { "epoch": 0.5035192203573362, "grad_norm": 0.33203125, "learning_rate": 0.0004198688763431069, "loss": 0.17, "step": 1860 }, { "epoch": 0.5062263129399025, "grad_norm": 0.095703125, "learning_rate": 0.00041941358586778364, "loss": 0.1596, "step": 1870 }, { "epoch": 0.5089334055224689, "grad_norm": 0.359375, "learning_rate": 0.00041895829539246037, "loss": 0.1591, "step": 1880 }, { "epoch": 0.5116404981050352, "grad_norm": 0.2392578125, "learning_rate": 0.00041850300491713715, "loss": 0.1489, "step": 1890 }, { "epoch": 0.5143475906876015, "grad_norm": 0.1611328125, "learning_rate": 0.0004180477144418139, "loss": 0.141, "step": 1900 }, { "epoch": 0.5143475906876015, "eval_loss": 0.18286480009555817, "eval_runtime": 51.4361, "eval_samples_per_second": 9.721, "eval_steps_per_second": 0.311, "step": 1900 }, { "epoch": 0.5170546832701678, "grad_norm": 0.1513671875, "learning_rate": 0.0004175924239664906, "loss": 0.1435, "step": 1910 }, { "epoch": 0.5197617758527342, "grad_norm": 0.150390625, "learning_rate": 0.0004171371334911674, "loss": 0.1549, "step": 1920 }, { "epoch": 0.5224688684353005, "grad_norm": 0.5546875, "learning_rate": 0.0004166818430158441, "loss": 0.1548, "step": 1930 }, { "epoch": 0.5251759610178668, "grad_norm": 0.44140625, "learning_rate": 0.0004162265525405209, "loss": 0.1838, "step": 1940 }, { "epoch": 0.5278830536004331, "grad_norm": 0.283203125, "learning_rate": 0.0004157712620651976, "loss": 0.1906, "step": 1950 }, { "epoch": 0.5305901461829995, "grad_norm": 0.462890625, "learning_rate": 0.00041531597158987434, "loss": 0.1393, "step": 1960 }, { "epoch": 0.5332972387655658, "grad_norm": 0.423828125, "learning_rate": 0.0004148606811145511, "loss": 0.167, "step": 1970 }, { "epoch": 0.5360043313481321, "grad_norm": 0.1748046875, "learning_rate": 0.00041440539063922785, "loss": 0.1789, "step": 1980 }, { "epoch": 0.5387114239306985, "grad_norm": 0.2099609375, "learning_rate": 0.0004139501001639046, "loss": 0.1693, "step": 1990 }, { "epoch": 0.5414185165132648, "grad_norm": 0.142578125, "learning_rate": 0.00041349480968858136, "loss": 0.1407, "step": 2000 }, { "epoch": 0.5414185165132648, "eval_loss": 0.1786070615053177, "eval_runtime": 51.1698, "eval_samples_per_second": 9.771, "eval_steps_per_second": 0.313, "step": 2000 }, { "epoch": 0.5441256090958311, "grad_norm": 0.3671875, "learning_rate": 0.00041303951921325803, "loss": 0.1487, "step": 2010 }, { "epoch": 0.5468327016783974, "grad_norm": 0.45703125, "learning_rate": 0.0004125842287379348, "loss": 0.1458, "step": 2020 }, { "epoch": 0.5495397942609638, "grad_norm": 0.2578125, "learning_rate": 0.00041212893826261153, "loss": 0.1763, "step": 2030 }, { "epoch": 0.5522468868435301, "grad_norm": 0.1708984375, "learning_rate": 0.00041167364778728826, "loss": 0.1262, "step": 2040 }, { "epoch": 0.5549539794260964, "grad_norm": 0.341796875, "learning_rate": 0.00041121835731196504, "loss": 0.1468, "step": 2050 }, { "epoch": 0.5576610720086627, "grad_norm": 0.1669921875, "learning_rate": 0.00041076306683664177, "loss": 0.1275, "step": 2060 }, { "epoch": 0.5603681645912291, "grad_norm": 0.4609375, "learning_rate": 0.00041030777636131855, "loss": 0.1663, "step": 2070 }, { "epoch": 0.5630752571737954, "grad_norm": 0.359375, "learning_rate": 0.0004098524858859953, "loss": 0.158, "step": 2080 }, { "epoch": 0.5657823497563617, "grad_norm": 0.2216796875, "learning_rate": 0.000409397195410672, "loss": 0.146, "step": 2090 }, { "epoch": 0.568489442338928, "grad_norm": 0.412109375, "learning_rate": 0.0004089419049353488, "loss": 0.136, "step": 2100 }, { "epoch": 0.568489442338928, "eval_loss": 0.1932000368833542, "eval_runtime": 50.2017, "eval_samples_per_second": 9.96, "eval_steps_per_second": 0.319, "step": 2100 }, { "epoch": 0.5711965349214944, "grad_norm": 0.244140625, "learning_rate": 0.0004084866144600255, "loss": 0.1242, "step": 2110 }, { "epoch": 0.5739036275040607, "grad_norm": 0.2265625, "learning_rate": 0.0004080313239847023, "loss": 0.1162, "step": 2120 }, { "epoch": 0.576610720086627, "grad_norm": 0.2734375, "learning_rate": 0.000407576033509379, "loss": 0.1615, "step": 2130 }, { "epoch": 0.5793178126691932, "grad_norm": 0.349609375, "learning_rate": 0.00040712074303405574, "loss": 0.1513, "step": 2140 }, { "epoch": 0.5820249052517596, "grad_norm": 0.34765625, "learning_rate": 0.00040666545255873247, "loss": 0.1466, "step": 2150 }, { "epoch": 0.584731997834326, "grad_norm": 0.142578125, "learning_rate": 0.0004062101620834092, "loss": 0.1348, "step": 2160 }, { "epoch": 0.5874390904168922, "grad_norm": 0.357421875, "learning_rate": 0.0004057548716080859, "loss": 0.1708, "step": 2170 }, { "epoch": 0.5901461829994585, "grad_norm": 0.263671875, "learning_rate": 0.0004052995811327627, "loss": 0.1304, "step": 2180 }, { "epoch": 0.5928532755820249, "grad_norm": 0.376953125, "learning_rate": 0.00040484429065743943, "loss": 0.1435, "step": 2190 }, { "epoch": 0.5955603681645912, "grad_norm": 0.1689453125, "learning_rate": 0.0004043890001821162, "loss": 0.134, "step": 2200 }, { "epoch": 0.5955603681645912, "eval_loss": 0.18312786519527435, "eval_runtime": 51.6541, "eval_samples_per_second": 9.68, "eval_steps_per_second": 0.31, "step": 2200 }, { "epoch": 0.5982674607471575, "grad_norm": 0.2158203125, "learning_rate": 0.00040393370970679294, "loss": 0.1884, "step": 2210 }, { "epoch": 0.6009745533297238, "grad_norm": 0.2041015625, "learning_rate": 0.00040347841923146966, "loss": 0.1432, "step": 2220 }, { "epoch": 0.6036816459122902, "grad_norm": 0.5390625, "learning_rate": 0.00040302312875614644, "loss": 0.1962, "step": 2230 }, { "epoch": 0.6063887384948565, "grad_norm": 0.3828125, "learning_rate": 0.00040256783828082317, "loss": 0.1512, "step": 2240 }, { "epoch": 0.6090958310774228, "grad_norm": 0.232421875, "learning_rate": 0.00040211254780549995, "loss": 0.1313, "step": 2250 }, { "epoch": 0.6118029236599891, "grad_norm": 0.3203125, "learning_rate": 0.0004016572573301767, "loss": 0.1391, "step": 2260 }, { "epoch": 0.6145100162425555, "grad_norm": 0.625, "learning_rate": 0.0004012019668548534, "loss": 0.1812, "step": 2270 }, { "epoch": 0.6172171088251218, "grad_norm": 0.1875, "learning_rate": 0.0004007466763795302, "loss": 0.1542, "step": 2280 }, { "epoch": 0.6199242014076881, "grad_norm": 0.298828125, "learning_rate": 0.00040029138590420686, "loss": 0.1433, "step": 2290 }, { "epoch": 0.6226312939902545, "grad_norm": 0.419921875, "learning_rate": 0.00039983609542888364, "loss": 0.1541, "step": 2300 }, { "epoch": 0.6226312939902545, "eval_loss": 0.17906835675239563, "eval_runtime": 53.0935, "eval_samples_per_second": 9.417, "eval_steps_per_second": 0.301, "step": 2300 }, { "epoch": 0.6253383865728208, "grad_norm": 0.220703125, "learning_rate": 0.00039938080495356037, "loss": 0.1498, "step": 2310 }, { "epoch": 0.6280454791553871, "grad_norm": 0.41796875, "learning_rate": 0.0003989255144782371, "loss": 0.1688, "step": 2320 }, { "epoch": 0.6307525717379534, "grad_norm": 0.380859375, "learning_rate": 0.00039847022400291387, "loss": 0.1467, "step": 2330 }, { "epoch": 0.6334596643205198, "grad_norm": 0.2060546875, "learning_rate": 0.0003980149335275906, "loss": 0.1255, "step": 2340 }, { "epoch": 0.6361667569030861, "grad_norm": 0.11083984375, "learning_rate": 0.0003975596430522673, "loss": 0.1284, "step": 2350 }, { "epoch": 0.6388738494856524, "grad_norm": 0.33984375, "learning_rate": 0.0003971043525769441, "loss": 0.1225, "step": 2360 }, { "epoch": 0.6415809420682187, "grad_norm": 0.208984375, "learning_rate": 0.00039664906210162083, "loss": 0.1784, "step": 2370 }, { "epoch": 0.6442880346507851, "grad_norm": 0.4453125, "learning_rate": 0.0003961937716262976, "loss": 0.1707, "step": 2380 }, { "epoch": 0.6469951272333514, "grad_norm": 0.1943359375, "learning_rate": 0.00039573848115097434, "loss": 0.1754, "step": 2390 }, { "epoch": 0.6497022198159177, "grad_norm": 0.287109375, "learning_rate": 0.00039528319067565107, "loss": 0.1581, "step": 2400 }, { "epoch": 0.6497022198159177, "eval_loss": 0.1758279949426651, "eval_runtime": 50.8478, "eval_samples_per_second": 9.833, "eval_steps_per_second": 0.315, "step": 2400 }, { "epoch": 0.652409312398484, "grad_norm": 0.474609375, "learning_rate": 0.00039482790020032785, "loss": 0.1885, "step": 2410 }, { "epoch": 0.6551164049810504, "grad_norm": 0.51171875, "learning_rate": 0.0003943726097250046, "loss": 0.167, "step": 2420 }, { "epoch": 0.6578234975636167, "grad_norm": 0.2734375, "learning_rate": 0.00039391731924968135, "loss": 0.1616, "step": 2430 }, { "epoch": 0.660530590146183, "grad_norm": 0.2578125, "learning_rate": 0.000393462028774358, "loss": 0.1327, "step": 2440 }, { "epoch": 0.6632376827287493, "grad_norm": 0.181640625, "learning_rate": 0.00039300673829903475, "loss": 0.155, "step": 2450 }, { "epoch": 0.6659447753113157, "grad_norm": 0.291015625, "learning_rate": 0.00039255144782371153, "loss": 0.125, "step": 2460 }, { "epoch": 0.668651867893882, "grad_norm": 0.23046875, "learning_rate": 0.00039209615734838826, "loss": 0.1478, "step": 2470 }, { "epoch": 0.6713589604764483, "grad_norm": 0.3203125, "learning_rate": 0.00039164086687306504, "loss": 0.1478, "step": 2480 }, { "epoch": 0.6740660530590146, "grad_norm": 0.53125, "learning_rate": 0.00039118557639774177, "loss": 0.1711, "step": 2490 }, { "epoch": 0.676773145641581, "grad_norm": 0.34765625, "learning_rate": 0.0003907302859224185, "loss": 0.1608, "step": 2500 }, { "epoch": 0.676773145641581, "eval_loss": 0.16929227113723755, "eval_runtime": 52.5378, "eval_samples_per_second": 9.517, "eval_steps_per_second": 0.305, "step": 2500 }, { "epoch": 0.6794802382241473, "grad_norm": 0.470703125, "learning_rate": 0.0003902749954470953, "loss": 0.1785, "step": 2510 }, { "epoch": 0.6821873308067136, "grad_norm": 0.2431640625, "learning_rate": 0.000389819704971772, "loss": 0.1459, "step": 2520 }, { "epoch": 0.6848944233892799, "grad_norm": 0.4921875, "learning_rate": 0.00038936441449644873, "loss": 0.1492, "step": 2530 }, { "epoch": 0.6876015159718463, "grad_norm": 0.294921875, "learning_rate": 0.0003889091240211255, "loss": 0.1149, "step": 2540 }, { "epoch": 0.6903086085544126, "grad_norm": 0.283203125, "learning_rate": 0.00038845383354580224, "loss": 0.1742, "step": 2550 }, { "epoch": 0.6930157011369789, "grad_norm": 0.142578125, "learning_rate": 0.000387998543070479, "loss": 0.1301, "step": 2560 }, { "epoch": 0.6957227937195453, "grad_norm": 0.083984375, "learning_rate": 0.00038754325259515574, "loss": 0.1445, "step": 2570 }, { "epoch": 0.6984298863021116, "grad_norm": 0.38671875, "learning_rate": 0.0003870879621198324, "loss": 0.1139, "step": 2580 }, { "epoch": 0.7011369788846779, "grad_norm": 0.326171875, "learning_rate": 0.0003866326716445092, "loss": 0.1363, "step": 2590 }, { "epoch": 0.7038440714672441, "grad_norm": 0.27734375, "learning_rate": 0.0003861773811691859, "loss": 0.1236, "step": 2600 }, { "epoch": 0.7038440714672441, "eval_loss": 0.1715136170387268, "eval_runtime": 50.1305, "eval_samples_per_second": 9.974, "eval_steps_per_second": 0.319, "step": 2600 }, { "epoch": 0.7065511640498106, "grad_norm": 0.3203125, "learning_rate": 0.0003857220906938627, "loss": 0.1507, "step": 2610 }, { "epoch": 0.7092582566323768, "grad_norm": 0.1376953125, "learning_rate": 0.00038526680021853943, "loss": 0.1634, "step": 2620 }, { "epoch": 0.7119653492149431, "grad_norm": 0.1669921875, "learning_rate": 0.00038481150974321616, "loss": 0.1723, "step": 2630 }, { "epoch": 0.7146724417975094, "grad_norm": 0.23046875, "learning_rate": 0.00038435621926789294, "loss": 0.1341, "step": 2640 }, { "epoch": 0.7173795343800758, "grad_norm": 0.2177734375, "learning_rate": 0.00038390092879256966, "loss": 0.1575, "step": 2650 }, { "epoch": 0.7200866269626421, "grad_norm": 0.2314453125, "learning_rate": 0.00038344563831724644, "loss": 0.1506, "step": 2660 }, { "epoch": 0.7227937195452084, "grad_norm": 0.2314453125, "learning_rate": 0.00038299034784192317, "loss": 0.1523, "step": 2670 }, { "epoch": 0.7255008121277747, "grad_norm": 0.453125, "learning_rate": 0.0003825350573665999, "loss": 0.1587, "step": 2680 }, { "epoch": 0.7282079047103411, "grad_norm": 0.2275390625, "learning_rate": 0.0003820797668912767, "loss": 0.1497, "step": 2690 }, { "epoch": 0.7309149972929074, "grad_norm": 0.130859375, "learning_rate": 0.0003816244764159534, "loss": 0.1143, "step": 2700 }, { "epoch": 0.7309149972929074, "eval_loss": 0.17223110795021057, "eval_runtime": 51.7962, "eval_samples_per_second": 9.653, "eval_steps_per_second": 0.309, "step": 2700 }, { "epoch": 0.7336220898754737, "grad_norm": 0.35546875, "learning_rate": 0.0003811691859406301, "loss": 0.1524, "step": 2710 }, { "epoch": 0.73632918245804, "grad_norm": 0.19921875, "learning_rate": 0.00038071389546530686, "loss": 0.1468, "step": 2720 }, { "epoch": 0.7390362750406064, "grad_norm": 0.2373046875, "learning_rate": 0.0003802586049899836, "loss": 0.1432, "step": 2730 }, { "epoch": 0.7417433676231727, "grad_norm": 0.41015625, "learning_rate": 0.00037980331451466036, "loss": 0.1372, "step": 2740 }, { "epoch": 0.744450460205739, "grad_norm": 0.228515625, "learning_rate": 0.0003793480240393371, "loss": 0.1337, "step": 2750 }, { "epoch": 0.7471575527883053, "grad_norm": 0.61328125, "learning_rate": 0.0003788927335640138, "loss": 0.1624, "step": 2760 }, { "epoch": 0.7498646453708717, "grad_norm": 0.2421875, "learning_rate": 0.0003784374430886906, "loss": 0.1501, "step": 2770 }, { "epoch": 0.752571737953438, "grad_norm": 0.345703125, "learning_rate": 0.0003779821526133673, "loss": 0.1482, "step": 2780 }, { "epoch": 0.7552788305360043, "grad_norm": 0.10986328125, "learning_rate": 0.0003775268621380441, "loss": 0.1289, "step": 2790 }, { "epoch": 0.7579859231185706, "grad_norm": 0.35546875, "learning_rate": 0.00037707157166272083, "loss": 0.1625, "step": 2800 }, { "epoch": 0.7579859231185706, "eval_loss": 0.16664518415927887, "eval_runtime": 49.9223, "eval_samples_per_second": 10.016, "eval_steps_per_second": 0.32, "step": 2800 }, { "epoch": 0.760693015701137, "grad_norm": 0.2451171875, "learning_rate": 0.00037661628118739756, "loss": 0.1518, "step": 2810 }, { "epoch": 0.7634001082837033, "grad_norm": 0.326171875, "learning_rate": 0.00037616099071207434, "loss": 0.1197, "step": 2820 }, { "epoch": 0.7661072008662696, "grad_norm": 0.375, "learning_rate": 0.00037570570023675107, "loss": 0.1053, "step": 2830 }, { "epoch": 0.7688142934488359, "grad_norm": 0.283203125, "learning_rate": 0.00037525040976142785, "loss": 0.1734, "step": 2840 }, { "epoch": 0.7715213860314023, "grad_norm": 0.1884765625, "learning_rate": 0.00037479511928610457, "loss": 0.1496, "step": 2850 }, { "epoch": 0.7742284786139686, "grad_norm": 0.50390625, "learning_rate": 0.00037433982881078124, "loss": 0.1551, "step": 2860 }, { "epoch": 0.7769355711965349, "grad_norm": 0.48046875, "learning_rate": 0.000373884538335458, "loss": 0.146, "step": 2870 }, { "epoch": 0.7796426637791013, "grad_norm": 0.423828125, "learning_rate": 0.00037342924786013475, "loss": 0.1618, "step": 2880 }, { "epoch": 0.7823497563616676, "grad_norm": 0.2392578125, "learning_rate": 0.0003729739573848115, "loss": 0.1487, "step": 2890 }, { "epoch": 0.7850568489442339, "grad_norm": 0.1953125, "learning_rate": 0.00037251866690948826, "loss": 0.1107, "step": 2900 }, { "epoch": 0.7850568489442339, "eval_loss": 0.17157955467700958, "eval_runtime": 50.4579, "eval_samples_per_second": 9.909, "eval_steps_per_second": 0.317, "step": 2900 }, { "epoch": 0.7877639415268002, "grad_norm": 0.40625, "learning_rate": 0.000372063376434165, "loss": 0.1585, "step": 2910 }, { "epoch": 0.7904710341093666, "grad_norm": 0.365234375, "learning_rate": 0.00037160808595884177, "loss": 0.1348, "step": 2920 }, { "epoch": 0.7931781266919329, "grad_norm": 0.369140625, "learning_rate": 0.0003711527954835185, "loss": 0.1585, "step": 2930 }, { "epoch": 0.7958852192744992, "grad_norm": 0.1787109375, "learning_rate": 0.0003706975050081952, "loss": 0.1762, "step": 2940 }, { "epoch": 0.7985923118570655, "grad_norm": 0.310546875, "learning_rate": 0.000370242214532872, "loss": 0.1445, "step": 2950 }, { "epoch": 0.8012994044396319, "grad_norm": 0.263671875, "learning_rate": 0.0003697869240575487, "loss": 0.1678, "step": 2960 }, { "epoch": 0.8040064970221982, "grad_norm": 0.265625, "learning_rate": 0.0003693316335822255, "loss": 0.1662, "step": 2970 }, { "epoch": 0.8067135896047645, "grad_norm": 0.1884765625, "learning_rate": 0.00036887634310690223, "loss": 0.1478, "step": 2980 }, { "epoch": 0.8094206821873308, "grad_norm": 0.099609375, "learning_rate": 0.00036842105263157896, "loss": 0.1412, "step": 2990 }, { "epoch": 0.8121277747698972, "grad_norm": 0.251953125, "learning_rate": 0.0003679657621562557, "loss": 0.1212, "step": 3000 }, { "epoch": 0.8121277747698972, "eval_loss": 0.17397309839725494, "eval_runtime": 50.6131, "eval_samples_per_second": 9.879, "eval_steps_per_second": 0.316, "step": 3000 }, { "epoch": 0.8148348673524635, "grad_norm": 0.2275390625, "learning_rate": 0.0003675104716809324, "loss": 0.149, "step": 3010 }, { "epoch": 0.8175419599350298, "grad_norm": 0.275390625, "learning_rate": 0.0003670551812056092, "loss": 0.1344, "step": 3020 }, { "epoch": 0.8202490525175961, "grad_norm": 0.0751953125, "learning_rate": 0.0003665998907302859, "loss": 0.1744, "step": 3030 }, { "epoch": 0.8229561451001625, "grad_norm": 0.259765625, "learning_rate": 0.00036614460025496265, "loss": 0.1394, "step": 3040 }, { "epoch": 0.8256632376827288, "grad_norm": 0.1865234375, "learning_rate": 0.00036568930977963943, "loss": 0.1359, "step": 3050 }, { "epoch": 0.828370330265295, "grad_norm": 0.314453125, "learning_rate": 0.00036523401930431615, "loss": 0.1631, "step": 3060 }, { "epoch": 0.8310774228478613, "grad_norm": 0.349609375, "learning_rate": 0.0003647787288289929, "loss": 0.1509, "step": 3070 }, { "epoch": 0.8337845154304278, "grad_norm": 0.2119140625, "learning_rate": 0.00036432343835366966, "loss": 0.1534, "step": 3080 }, { "epoch": 0.836491608012994, "grad_norm": 0.359375, "learning_rate": 0.0003638681478783464, "loss": 0.1473, "step": 3090 }, { "epoch": 0.8391987005955603, "grad_norm": 0.486328125, "learning_rate": 0.00036341285740302317, "loss": 0.1406, "step": 3100 }, { "epoch": 0.8391987005955603, "eval_loss": 0.1714637130498886, "eval_runtime": 50.3998, "eval_samples_per_second": 9.921, "eval_steps_per_second": 0.317, "step": 3100 }, { "epoch": 0.8419057931781266, "grad_norm": 0.3828125, "learning_rate": 0.0003629575669276999, "loss": 0.1712, "step": 3110 }, { "epoch": 0.844612885760693, "grad_norm": 0.439453125, "learning_rate": 0.0003625022764523766, "loss": 0.162, "step": 3120 }, { "epoch": 0.8473199783432593, "grad_norm": 0.3828125, "learning_rate": 0.0003620469859770534, "loss": 0.1609, "step": 3130 }, { "epoch": 0.8500270709258256, "grad_norm": 0.2470703125, "learning_rate": 0.0003615916955017301, "loss": 0.1701, "step": 3140 }, { "epoch": 0.852734163508392, "grad_norm": 0.349609375, "learning_rate": 0.00036113640502640686, "loss": 0.2107, "step": 3150 }, { "epoch": 0.8554412560909583, "grad_norm": 0.41015625, "learning_rate": 0.0003606811145510836, "loss": 0.1407, "step": 3160 }, { "epoch": 0.8581483486735246, "grad_norm": 0.2373046875, "learning_rate": 0.0003602258240757603, "loss": 0.1761, "step": 3170 }, { "epoch": 0.8608554412560909, "grad_norm": 0.392578125, "learning_rate": 0.0003597705336004371, "loss": 0.1434, "step": 3180 }, { "epoch": 0.8635625338386573, "grad_norm": 0.5625, "learning_rate": 0.0003593152431251138, "loss": 0.1506, "step": 3190 }, { "epoch": 0.8662696264212236, "grad_norm": 0.3359375, "learning_rate": 0.0003588599526497906, "loss": 0.1245, "step": 3200 }, { "epoch": 0.8662696264212236, "eval_loss": 0.16608840227127075, "eval_runtime": 50.1258, "eval_samples_per_second": 9.975, "eval_steps_per_second": 0.319, "step": 3200 }, { "epoch": 0.8689767190037899, "grad_norm": 0.2265625, "learning_rate": 0.0003584046621744673, "loss": 0.1403, "step": 3210 }, { "epoch": 0.8716838115863562, "grad_norm": 0.12353515625, "learning_rate": 0.00035794937169914405, "loss": 0.1356, "step": 3220 }, { "epoch": 0.8743909041689226, "grad_norm": 0.34765625, "learning_rate": 0.00035749408122382083, "loss": 0.1398, "step": 3230 }, { "epoch": 0.8770979967514889, "grad_norm": 0.279296875, "learning_rate": 0.00035703879074849756, "loss": 0.1641, "step": 3240 }, { "epoch": 0.8798050893340552, "grad_norm": 0.13671875, "learning_rate": 0.0003565835002731743, "loss": 0.1326, "step": 3250 }, { "epoch": 0.8825121819166215, "grad_norm": 0.26953125, "learning_rate": 0.00035612820979785106, "loss": 0.1267, "step": 3260 }, { "epoch": 0.8852192744991879, "grad_norm": 0.1474609375, "learning_rate": 0.0003556729193225278, "loss": 0.1514, "step": 3270 }, { "epoch": 0.8879263670817542, "grad_norm": 0.45703125, "learning_rate": 0.0003552176288472045, "loss": 0.1558, "step": 3280 }, { "epoch": 0.8906334596643205, "grad_norm": 0.357421875, "learning_rate": 0.00035476233837188124, "loss": 0.1496, "step": 3290 }, { "epoch": 0.8933405522468868, "grad_norm": 0.1767578125, "learning_rate": 0.00035430704789655797, "loss": 0.1574, "step": 3300 }, { "epoch": 0.8933405522468868, "eval_loss": 0.1689441204071045, "eval_runtime": 50.2047, "eval_samples_per_second": 9.959, "eval_steps_per_second": 0.319, "step": 3300 }, { "epoch": 0.8960476448294532, "grad_norm": 0.4375, "learning_rate": 0.00035385175742123475, "loss": 0.1382, "step": 3310 }, { "epoch": 0.8987547374120195, "grad_norm": 0.30859375, "learning_rate": 0.0003533964669459115, "loss": 0.1095, "step": 3320 }, { "epoch": 0.9014618299945858, "grad_norm": 0.3828125, "learning_rate": 0.00035294117647058826, "loss": 0.1643, "step": 3330 }, { "epoch": 0.9041689225771521, "grad_norm": 0.279296875, "learning_rate": 0.000352485885995265, "loss": 0.1082, "step": 3340 }, { "epoch": 0.9068760151597185, "grad_norm": 0.1904296875, "learning_rate": 0.0003520305955199417, "loss": 0.1377, "step": 3350 }, { "epoch": 0.9095831077422848, "grad_norm": 0.2392578125, "learning_rate": 0.0003515753050446185, "loss": 0.1264, "step": 3360 }, { "epoch": 0.9122902003248511, "grad_norm": 0.392578125, "learning_rate": 0.0003511200145692952, "loss": 0.1663, "step": 3370 }, { "epoch": 0.9149972929074174, "grad_norm": 0.150390625, "learning_rate": 0.000350664724093972, "loss": 0.1463, "step": 3380 }, { "epoch": 0.9177043854899838, "grad_norm": 0.1533203125, "learning_rate": 0.0003502094336186487, "loss": 0.1004, "step": 3390 }, { "epoch": 0.9204114780725501, "grad_norm": 0.1806640625, "learning_rate": 0.00034975414314332545, "loss": 0.1394, "step": 3400 }, { "epoch": 0.9204114780725501, "eval_loss": 0.16821011900901794, "eval_runtime": 51.9302, "eval_samples_per_second": 9.628, "eval_steps_per_second": 0.308, "step": 3400 }, { "epoch": 0.9231185706551164, "grad_norm": 0.455078125, "learning_rate": 0.00034929885266800223, "loss": 0.1351, "step": 3410 }, { "epoch": 0.9258256632376828, "grad_norm": 0.310546875, "learning_rate": 0.0003488435621926789, "loss": 0.1314, "step": 3420 }, { "epoch": 0.9285327558202491, "grad_norm": 0.2353515625, "learning_rate": 0.00034838827171735563, "loss": 0.1377, "step": 3430 }, { "epoch": 0.9312398484028154, "grad_norm": 0.318359375, "learning_rate": 0.0003479329812420324, "loss": 0.143, "step": 3440 }, { "epoch": 0.9339469409853817, "grad_norm": 0.34375, "learning_rate": 0.00034747769076670914, "loss": 0.1408, "step": 3450 }, { "epoch": 0.9366540335679481, "grad_norm": 0.328125, "learning_rate": 0.0003470224002913859, "loss": 0.1777, "step": 3460 }, { "epoch": 0.9393611261505144, "grad_norm": 0.16015625, "learning_rate": 0.00034656710981606265, "loss": 0.1183, "step": 3470 }, { "epoch": 0.9420682187330807, "grad_norm": 0.6875, "learning_rate": 0.0003461118193407394, "loss": 0.1524, "step": 3480 }, { "epoch": 0.944775311315647, "grad_norm": 0.265625, "learning_rate": 0.00034565652886541615, "loss": 0.155, "step": 3490 }, { "epoch": 0.9474824038982134, "grad_norm": 0.2275390625, "learning_rate": 0.0003452012383900929, "loss": 0.145, "step": 3500 }, { "epoch": 0.9474824038982134, "eval_loss": 0.16947884857654572, "eval_runtime": 51.2584, "eval_samples_per_second": 9.755, "eval_steps_per_second": 0.312, "step": 3500 }, { "epoch": 0.9501894964807797, "grad_norm": 0.71875, "learning_rate": 0.00034474594791476966, "loss": 0.1241, "step": 3510 }, { "epoch": 0.952896589063346, "grad_norm": 0.2314453125, "learning_rate": 0.0003442906574394464, "loss": 0.1459, "step": 3520 }, { "epoch": 0.9556036816459123, "grad_norm": 0.23046875, "learning_rate": 0.0003438353669641231, "loss": 0.1171, "step": 3530 }, { "epoch": 0.9583107742284787, "grad_norm": 0.2109375, "learning_rate": 0.0003433800764887999, "loss": 0.1398, "step": 3540 }, { "epoch": 0.961017866811045, "grad_norm": 0.302734375, "learning_rate": 0.0003429247860134766, "loss": 0.1196, "step": 3550 }, { "epoch": 0.9637249593936112, "grad_norm": 0.17578125, "learning_rate": 0.00034246949553815335, "loss": 0.1478, "step": 3560 }, { "epoch": 0.9664320519761775, "grad_norm": 0.23828125, "learning_rate": 0.0003420142050628301, "loss": 0.1532, "step": 3570 }, { "epoch": 0.969139144558744, "grad_norm": 0.455078125, "learning_rate": 0.0003415589145875068, "loss": 0.1341, "step": 3580 }, { "epoch": 0.9718462371413102, "grad_norm": 0.201171875, "learning_rate": 0.0003411036241121836, "loss": 0.1649, "step": 3590 }, { "epoch": 0.9745533297238765, "grad_norm": 0.16796875, "learning_rate": 0.0003406483336368603, "loss": 0.1501, "step": 3600 }, { "epoch": 0.9745533297238765, "eval_loss": 0.16499453783035278, "eval_runtime": 52.0838, "eval_samples_per_second": 9.6, "eval_steps_per_second": 0.307, "step": 3600 }, { "epoch": 0.9772604223064428, "grad_norm": 0.400390625, "learning_rate": 0.00034019304316153703, "loss": 0.1629, "step": 3610 }, { "epoch": 0.9799675148890092, "grad_norm": 0.419921875, "learning_rate": 0.0003397377526862138, "loss": 0.1329, "step": 3620 }, { "epoch": 0.9826746074715755, "grad_norm": 0.1728515625, "learning_rate": 0.00033928246221089054, "loss": 0.1379, "step": 3630 }, { "epoch": 0.9853817000541418, "grad_norm": 0.1552734375, "learning_rate": 0.0003388271717355673, "loss": 0.1488, "step": 3640 }, { "epoch": 0.9880887926367081, "grad_norm": 0.140625, "learning_rate": 0.00033837188126024405, "loss": 0.1597, "step": 3650 }, { "epoch": 0.9907958852192745, "grad_norm": 0.45703125, "learning_rate": 0.0003379165907849208, "loss": 0.1496, "step": 3660 }, { "epoch": 0.9935029778018408, "grad_norm": 0.140625, "learning_rate": 0.00033746130030959756, "loss": 0.1262, "step": 3670 }, { "epoch": 0.9962100703844071, "grad_norm": 0.359375, "learning_rate": 0.0003370060098342743, "loss": 0.1194, "step": 3680 }, { "epoch": 0.9989171629669734, "grad_norm": 0.1904296875, "learning_rate": 0.00033655071935895106, "loss": 0.1285, "step": 3690 }, { "epoch": 1.0016242555495398, "grad_norm": 0.302734375, "learning_rate": 0.00033609542888362774, "loss": 0.1643, "step": 3700 }, { "epoch": 1.0016242555495398, "eval_loss": 0.1689431369304657, "eval_runtime": 50.9236, "eval_samples_per_second": 9.819, "eval_steps_per_second": 0.314, "step": 3700 }, { "epoch": 1.0043313481321061, "grad_norm": 0.15625, "learning_rate": 0.00033564013840830446, "loss": 0.1191, "step": 3710 }, { "epoch": 1.0070384407146724, "grad_norm": 0.234375, "learning_rate": 0.00033518484793298124, "loss": 0.116, "step": 3720 }, { "epoch": 1.0097455332972387, "grad_norm": 0.1611328125, "learning_rate": 0.00033472955745765797, "loss": 0.1083, "step": 3730 }, { "epoch": 1.012452625879805, "grad_norm": 0.1220703125, "learning_rate": 0.00033427426698233475, "loss": 0.1558, "step": 3740 }, { "epoch": 1.0151597184623715, "grad_norm": 0.2265625, "learning_rate": 0.0003338189765070115, "loss": 0.1273, "step": 3750 }, { "epoch": 1.0178668110449378, "grad_norm": 0.32421875, "learning_rate": 0.0003333636860316882, "loss": 0.145, "step": 3760 }, { "epoch": 1.020573903627504, "grad_norm": 0.2333984375, "learning_rate": 0.000332908395556365, "loss": 0.1308, "step": 3770 }, { "epoch": 1.0232809962100704, "grad_norm": 0.1611328125, "learning_rate": 0.0003324531050810417, "loss": 0.1483, "step": 3780 }, { "epoch": 1.0259880887926367, "grad_norm": 0.3125, "learning_rate": 0.00033199781460571844, "loss": 0.1164, "step": 3790 }, { "epoch": 1.028695181375203, "grad_norm": 0.384765625, "learning_rate": 0.0003315425241303952, "loss": 0.1587, "step": 3800 }, { "epoch": 1.028695181375203, "eval_loss": 0.1659206598997116, "eval_runtime": 50.5276, "eval_samples_per_second": 9.896, "eval_steps_per_second": 0.317, "step": 3800 }, { "epoch": 1.0314022739577693, "grad_norm": 0.294921875, "learning_rate": 0.00033108723365507194, "loss": 0.125, "step": 3810 }, { "epoch": 1.0341093665403356, "grad_norm": 0.1005859375, "learning_rate": 0.0003306319431797487, "loss": 0.1409, "step": 3820 }, { "epoch": 1.036816459122902, "grad_norm": 0.134765625, "learning_rate": 0.00033017665270442545, "loss": 0.1236, "step": 3830 }, { "epoch": 1.0395235517054684, "grad_norm": 0.134765625, "learning_rate": 0.0003297213622291021, "loss": 0.1327, "step": 3840 }, { "epoch": 1.0422306442880347, "grad_norm": 0.162109375, "learning_rate": 0.0003292660717537789, "loss": 0.1083, "step": 3850 }, { "epoch": 1.044937736870601, "grad_norm": 0.2431640625, "learning_rate": 0.00032881078127845563, "loss": 0.1372, "step": 3860 }, { "epoch": 1.0476448294531673, "grad_norm": 0.412109375, "learning_rate": 0.0003283554908031324, "loss": 0.1111, "step": 3870 }, { "epoch": 1.0503519220357336, "grad_norm": 0.298828125, "learning_rate": 0.00032790020032780914, "loss": 0.137, "step": 3880 }, { "epoch": 1.0530590146182999, "grad_norm": 0.4765625, "learning_rate": 0.00032744490985248587, "loss": 0.1124, "step": 3890 }, { "epoch": 1.0557661072008662, "grad_norm": 0.1650390625, "learning_rate": 0.00032698961937716265, "loss": 0.137, "step": 3900 }, { "epoch": 1.0557661072008662, "eval_loss": 0.16376519203186035, "eval_runtime": 50.6812, "eval_samples_per_second": 9.866, "eval_steps_per_second": 0.316, "step": 3900 }, { "epoch": 1.0584731997834327, "grad_norm": 0.111328125, "learning_rate": 0.00032653432890183937, "loss": 0.1337, "step": 3910 }, { "epoch": 1.061180292365999, "grad_norm": 0.2490234375, "learning_rate": 0.00032607903842651615, "loss": 0.0989, "step": 3920 }, { "epoch": 1.0638873849485653, "grad_norm": 0.203125, "learning_rate": 0.0003256237479511929, "loss": 0.128, "step": 3930 }, { "epoch": 1.0665944775311316, "grad_norm": 0.25390625, "learning_rate": 0.0003251684574758696, "loss": 0.1188, "step": 3940 }, { "epoch": 1.0693015701136979, "grad_norm": 0.1982421875, "learning_rate": 0.0003247131670005464, "loss": 0.1374, "step": 3950 }, { "epoch": 1.0720086626962642, "grad_norm": 0.177734375, "learning_rate": 0.0003242578765252231, "loss": 0.1486, "step": 3960 }, { "epoch": 1.0747157552788305, "grad_norm": 0.12890625, "learning_rate": 0.00032380258604989984, "loss": 0.15, "step": 3970 }, { "epoch": 1.077422847861397, "grad_norm": 0.365234375, "learning_rate": 0.0003233472955745766, "loss": 0.148, "step": 3980 }, { "epoch": 1.0801299404439633, "grad_norm": 0.271484375, "learning_rate": 0.0003228920050992533, "loss": 0.1351, "step": 3990 }, { "epoch": 1.0828370330265296, "grad_norm": 0.1640625, "learning_rate": 0.0003224367146239301, "loss": 0.1109, "step": 4000 }, { "epoch": 1.0828370330265296, "eval_loss": 0.1612129956483841, "eval_runtime": 50.3507, "eval_samples_per_second": 9.93, "eval_steps_per_second": 0.318, "step": 4000 }, { "epoch": 1.0855441256090959, "grad_norm": 0.1376953125, "learning_rate": 0.0003219814241486068, "loss": 0.105, "step": 4010 }, { "epoch": 1.0882512181916622, "grad_norm": 0.359375, "learning_rate": 0.0003215261336732835, "loss": 0.1322, "step": 4020 }, { "epoch": 1.0909583107742284, "grad_norm": 0.173828125, "learning_rate": 0.0003210708431979603, "loss": 0.127, "step": 4030 }, { "epoch": 1.0936654033567947, "grad_norm": 0.2099609375, "learning_rate": 0.00032061555272263703, "loss": 0.1185, "step": 4040 }, { "epoch": 1.096372495939361, "grad_norm": 0.3359375, "learning_rate": 0.0003201602622473138, "loss": 0.129, "step": 4050 }, { "epoch": 1.0990795885219276, "grad_norm": 0.10107421875, "learning_rate": 0.00031970497177199054, "loss": 0.0838, "step": 4060 }, { "epoch": 1.1017866811044938, "grad_norm": 0.4140625, "learning_rate": 0.00031924968129666727, "loss": 0.1047, "step": 4070 }, { "epoch": 1.1044937736870601, "grad_norm": 0.123046875, "learning_rate": 0.00031879439082134405, "loss": 0.1305, "step": 4080 }, { "epoch": 1.1072008662696264, "grad_norm": 0.279296875, "learning_rate": 0.0003183391003460208, "loss": 0.1006, "step": 4090 }, { "epoch": 1.1099079588521927, "grad_norm": 0.2392578125, "learning_rate": 0.00031788380987069756, "loss": 0.1227, "step": 4100 }, { "epoch": 1.1099079588521927, "eval_loss": 0.16056661307811737, "eval_runtime": 51.1636, "eval_samples_per_second": 9.773, "eval_steps_per_second": 0.313, "step": 4100 }, { "epoch": 1.112615051434759, "grad_norm": 0.365234375, "learning_rate": 0.0003174285193953743, "loss": 0.1338, "step": 4110 }, { "epoch": 1.1153221440173253, "grad_norm": 0.189453125, "learning_rate": 0.000316973228920051, "loss": 0.1076, "step": 4120 }, { "epoch": 1.1180292365998916, "grad_norm": 0.1328125, "learning_rate": 0.00031651793844472774, "loss": 0.1242, "step": 4130 }, { "epoch": 1.1207363291824581, "grad_norm": 0.41015625, "learning_rate": 0.00031606264796940446, "loss": 0.1359, "step": 4140 }, { "epoch": 1.1234434217650244, "grad_norm": 0.134765625, "learning_rate": 0.0003156073574940812, "loss": 0.1229, "step": 4150 }, { "epoch": 1.1261505143475907, "grad_norm": 0.326171875, "learning_rate": 0.00031515206701875797, "loss": 0.1236, "step": 4160 }, { "epoch": 1.128857606930157, "grad_norm": 0.271484375, "learning_rate": 0.0003146967765434347, "loss": 0.1342, "step": 4170 }, { "epoch": 1.1315646995127233, "grad_norm": 0.19140625, "learning_rate": 0.0003142414860681115, "loss": 0.0986, "step": 4180 }, { "epoch": 1.1342717920952896, "grad_norm": 0.345703125, "learning_rate": 0.0003137861955927882, "loss": 0.1257, "step": 4190 }, { "epoch": 1.136978884677856, "grad_norm": 0.1904296875, "learning_rate": 0.00031333090511746493, "loss": 0.1356, "step": 4200 }, { "epoch": 1.136978884677856, "eval_loss": 0.1527666598558426, "eval_runtime": 50.0731, "eval_samples_per_second": 9.985, "eval_steps_per_second": 0.32, "step": 4200 }, { "epoch": 1.1396859772604224, "grad_norm": 0.33984375, "learning_rate": 0.0003128756146421417, "loss": 0.1214, "step": 4210 }, { "epoch": 1.1423930698429887, "grad_norm": 0.1982421875, "learning_rate": 0.00031242032416681844, "loss": 0.1158, "step": 4220 }, { "epoch": 1.145100162425555, "grad_norm": 0.296875, "learning_rate": 0.0003119650336914952, "loss": 0.1128, "step": 4230 }, { "epoch": 1.1478072550081213, "grad_norm": 0.1259765625, "learning_rate": 0.00031150974321617194, "loss": 0.0777, "step": 4240 }, { "epoch": 1.1505143475906876, "grad_norm": 0.294921875, "learning_rate": 0.00031105445274084867, "loss": 0.1316, "step": 4250 }, { "epoch": 1.153221440173254, "grad_norm": 0.3671875, "learning_rate": 0.00031059916226552545, "loss": 0.1176, "step": 4260 }, { "epoch": 1.1559285327558202, "grad_norm": 0.1826171875, "learning_rate": 0.0003101438717902021, "loss": 0.1292, "step": 4270 }, { "epoch": 1.1586356253383865, "grad_norm": 0.294921875, "learning_rate": 0.0003096885813148789, "loss": 0.1368, "step": 4280 }, { "epoch": 1.1613427179209528, "grad_norm": 0.25, "learning_rate": 0.00030923329083955563, "loss": 0.1283, "step": 4290 }, { "epoch": 1.1640498105035193, "grad_norm": 0.283203125, "learning_rate": 0.00030877800036423236, "loss": 0.1383, "step": 4300 }, { "epoch": 1.1640498105035193, "eval_loss": 0.1612272709608078, "eval_runtime": 50.8695, "eval_samples_per_second": 9.829, "eval_steps_per_second": 0.315, "step": 4300 }, { "epoch": 1.1667569030860856, "grad_norm": 0.326171875, "learning_rate": 0.00030832270988890914, "loss": 0.1211, "step": 4310 }, { "epoch": 1.169463995668652, "grad_norm": 0.1552734375, "learning_rate": 0.00030786741941358586, "loss": 0.1308, "step": 4320 }, { "epoch": 1.1721710882512182, "grad_norm": 0.1845703125, "learning_rate": 0.0003074121289382626, "loss": 0.1196, "step": 4330 }, { "epoch": 1.1748781808337845, "grad_norm": 0.23046875, "learning_rate": 0.00030695683846293937, "loss": 0.1596, "step": 4340 }, { "epoch": 1.1775852734163508, "grad_norm": 0.2216796875, "learning_rate": 0.0003065015479876161, "loss": 0.1029, "step": 4350 }, { "epoch": 1.180292365998917, "grad_norm": 0.1220703125, "learning_rate": 0.0003060462575122929, "loss": 0.0975, "step": 4360 }, { "epoch": 1.1829994585814836, "grad_norm": 0.5390625, "learning_rate": 0.0003055909670369696, "loss": 0.1485, "step": 4370 }, { "epoch": 1.1857065511640499, "grad_norm": 0.451171875, "learning_rate": 0.00030513567656164633, "loss": 0.1329, "step": 4380 }, { "epoch": 1.1884136437466162, "grad_norm": 0.2353515625, "learning_rate": 0.0003046803860863231, "loss": 0.1443, "step": 4390 }, { "epoch": 1.1911207363291825, "grad_norm": 0.2197265625, "learning_rate": 0.00030422509561099984, "loss": 0.1266, "step": 4400 }, { "epoch": 1.1911207363291825, "eval_loss": 0.15496616065502167, "eval_runtime": 50.2857, "eval_samples_per_second": 9.943, "eval_steps_per_second": 0.318, "step": 4400 }, { "epoch": 1.1938278289117488, "grad_norm": 0.119140625, "learning_rate": 0.00030376980513567657, "loss": 0.0877, "step": 4410 }, { "epoch": 1.196534921494315, "grad_norm": 0.470703125, "learning_rate": 0.0003033145146603533, "loss": 0.1178, "step": 4420 }, { "epoch": 1.1992420140768814, "grad_norm": 0.43359375, "learning_rate": 0.00030285922418503, "loss": 0.1191, "step": 4430 }, { "epoch": 1.2019491066594479, "grad_norm": 0.205078125, "learning_rate": 0.0003024039337097068, "loss": 0.1465, "step": 4440 }, { "epoch": 1.2046561992420142, "grad_norm": 0.46875, "learning_rate": 0.0003019486432343835, "loss": 0.1411, "step": 4450 }, { "epoch": 1.2073632918245805, "grad_norm": 0.34375, "learning_rate": 0.0003014933527590603, "loss": 0.139, "step": 4460 }, { "epoch": 1.2100703844071468, "grad_norm": 0.0703125, "learning_rate": 0.00030103806228373703, "loss": 0.1043, "step": 4470 }, { "epoch": 1.212777476989713, "grad_norm": 0.19921875, "learning_rate": 0.00030058277180841376, "loss": 0.133, "step": 4480 }, { "epoch": 1.2154845695722793, "grad_norm": 0.0830078125, "learning_rate": 0.00030012748133309054, "loss": 0.1323, "step": 4490 }, { "epoch": 1.2181916621548456, "grad_norm": 0.376953125, "learning_rate": 0.00029967219085776727, "loss": 0.0916, "step": 4500 }, { "epoch": 1.2181916621548456, "eval_loss": 0.15184161067008972, "eval_runtime": 51.497, "eval_samples_per_second": 9.709, "eval_steps_per_second": 0.311, "step": 4500 }, { "epoch": 1.220898754737412, "grad_norm": 0.34765625, "learning_rate": 0.000299216900382444, "loss": 0.1554, "step": 4510 }, { "epoch": 1.2236058473199782, "grad_norm": 0.291015625, "learning_rate": 0.0002987616099071208, "loss": 0.119, "step": 4520 }, { "epoch": 1.2263129399025448, "grad_norm": 0.3984375, "learning_rate": 0.0002983063194317975, "loss": 0.1219, "step": 4530 }, { "epoch": 1.229020032485111, "grad_norm": 0.2109375, "learning_rate": 0.0002978510289564743, "loss": 0.1145, "step": 4540 }, { "epoch": 1.2317271250676773, "grad_norm": 0.189453125, "learning_rate": 0.00029739573848115095, "loss": 0.1175, "step": 4550 }, { "epoch": 1.2344342176502436, "grad_norm": 0.4453125, "learning_rate": 0.0002969404480058277, "loss": 0.1393, "step": 4560 }, { "epoch": 1.23714131023281, "grad_norm": 0.2470703125, "learning_rate": 0.00029648515753050446, "loss": 0.1182, "step": 4570 }, { "epoch": 1.2398484028153762, "grad_norm": 0.275390625, "learning_rate": 0.0002960298670551812, "loss": 0.1182, "step": 4580 }, { "epoch": 1.2425554953979425, "grad_norm": 0.2333984375, "learning_rate": 0.00029557457657985797, "loss": 0.1228, "step": 4590 }, { "epoch": 1.245262587980509, "grad_norm": 0.16796875, "learning_rate": 0.0002951192861045347, "loss": 0.1389, "step": 4600 }, { "epoch": 1.245262587980509, "eval_loss": 0.15244832634925842, "eval_runtime": 48.8981, "eval_samples_per_second": 10.225, "eval_steps_per_second": 0.327, "step": 4600 }, { "epoch": 1.2479696805630753, "grad_norm": 0.2578125, "learning_rate": 0.0002946639956292114, "loss": 0.116, "step": 4610 }, { "epoch": 1.2506767731456416, "grad_norm": 0.16796875, "learning_rate": 0.0002942087051538882, "loss": 0.1196, "step": 4620 }, { "epoch": 1.253383865728208, "grad_norm": 0.2275390625, "learning_rate": 0.00029375341467856493, "loss": 0.1146, "step": 4630 }, { "epoch": 1.2560909583107742, "grad_norm": 0.10791015625, "learning_rate": 0.0002932981242032417, "loss": 0.1242, "step": 4640 }, { "epoch": 1.2587980508933405, "grad_norm": 0.1728515625, "learning_rate": 0.00029284283372791844, "loss": 0.0968, "step": 4650 }, { "epoch": 1.2615051434759068, "grad_norm": 0.13671875, "learning_rate": 0.00029238754325259516, "loss": 0.1109, "step": 4660 }, { "epoch": 1.2642122360584733, "grad_norm": 0.2197265625, "learning_rate": 0.00029193225277727194, "loss": 0.108, "step": 4670 }, { "epoch": 1.2669193286410394, "grad_norm": 0.439453125, "learning_rate": 0.00029147696230194867, "loss": 0.1303, "step": 4680 }, { "epoch": 1.269626421223606, "grad_norm": 0.185546875, "learning_rate": 0.00029102167182662534, "loss": 0.1217, "step": 4690 }, { "epoch": 1.2723335138061722, "grad_norm": 0.294921875, "learning_rate": 0.0002905663813513021, "loss": 0.1103, "step": 4700 }, { "epoch": 1.2723335138061722, "eval_loss": 0.15599898993968964, "eval_runtime": 49.5398, "eval_samples_per_second": 10.093, "eval_steps_per_second": 0.323, "step": 4700 }, { "epoch": 1.2750406063887385, "grad_norm": 0.458984375, "learning_rate": 0.00029011109087597885, "loss": 0.1386, "step": 4710 }, { "epoch": 1.2777476989713048, "grad_norm": 0.29296875, "learning_rate": 0.00028965580040065563, "loss": 0.1032, "step": 4720 }, { "epoch": 1.280454791553871, "grad_norm": 0.27734375, "learning_rate": 0.00028920050992533236, "loss": 0.1613, "step": 4730 }, { "epoch": 1.2831618841364374, "grad_norm": 0.30859375, "learning_rate": 0.0002887452194500091, "loss": 0.1049, "step": 4740 }, { "epoch": 1.2858689767190037, "grad_norm": 0.1259765625, "learning_rate": 0.00028828992897468586, "loss": 0.1108, "step": 4750 }, { "epoch": 1.2885760693015702, "grad_norm": 0.19921875, "learning_rate": 0.0002878346384993626, "loss": 0.1154, "step": 4760 }, { "epoch": 1.2912831618841365, "grad_norm": 0.1259765625, "learning_rate": 0.00028737934802403937, "loss": 0.1158, "step": 4770 }, { "epoch": 1.2939902544667028, "grad_norm": 0.384765625, "learning_rate": 0.0002869240575487161, "loss": 0.1171, "step": 4780 }, { "epoch": 1.296697347049269, "grad_norm": 0.244140625, "learning_rate": 0.0002864687670733928, "loss": 0.1393, "step": 4790 }, { "epoch": 1.2994044396318354, "grad_norm": 0.466796875, "learning_rate": 0.0002860134765980696, "loss": 0.1372, "step": 4800 }, { "epoch": 1.2994044396318354, "eval_loss": 0.15471972525119781, "eval_runtime": 52.6076, "eval_samples_per_second": 9.504, "eval_steps_per_second": 0.304, "step": 4800 }, { "epoch": 1.3021115322144017, "grad_norm": 0.1083984375, "learning_rate": 0.00028555818612274633, "loss": 0.0928, "step": 4810 }, { "epoch": 1.304818624796968, "grad_norm": 0.2197265625, "learning_rate": 0.0002851028956474231, "loss": 0.129, "step": 4820 }, { "epoch": 1.3075257173795345, "grad_norm": 0.11083984375, "learning_rate": 0.0002846476051720998, "loss": 0.1089, "step": 4830 }, { "epoch": 1.3102328099621008, "grad_norm": 0.63671875, "learning_rate": 0.0002841923146967765, "loss": 0.1102, "step": 4840 }, { "epoch": 1.312939902544667, "grad_norm": 0.255859375, "learning_rate": 0.0002837370242214533, "loss": 0.119, "step": 4850 }, { "epoch": 1.3156469951272334, "grad_norm": 0.1201171875, "learning_rate": 0.00028328173374613, "loss": 0.1385, "step": 4860 }, { "epoch": 1.3183540877097997, "grad_norm": 0.333984375, "learning_rate": 0.00028282644327080674, "loss": 0.0835, "step": 4870 }, { "epoch": 1.321061180292366, "grad_norm": 0.45703125, "learning_rate": 0.0002823711527954835, "loss": 0.1346, "step": 4880 }, { "epoch": 1.3237682728749323, "grad_norm": 0.40625, "learning_rate": 0.00028191586232016025, "loss": 0.1573, "step": 4890 }, { "epoch": 1.3264753654574988, "grad_norm": 0.07275390625, "learning_rate": 0.00028146057184483703, "loss": 0.1716, "step": 4900 }, { "epoch": 1.3264753654574988, "eval_loss": 0.1505041867494583, "eval_runtime": 51.3184, "eval_samples_per_second": 9.743, "eval_steps_per_second": 0.312, "step": 4900 }, { "epoch": 1.3291824580400649, "grad_norm": 0.059326171875, "learning_rate": 0.00028100528136951376, "loss": 0.1112, "step": 4910 }, { "epoch": 1.3318895506226314, "grad_norm": 0.2265625, "learning_rate": 0.0002805499908941905, "loss": 0.1306, "step": 4920 }, { "epoch": 1.3345966432051977, "grad_norm": 0.326171875, "learning_rate": 0.00028009470041886727, "loss": 0.1261, "step": 4930 }, { "epoch": 1.337303735787764, "grad_norm": 0.3515625, "learning_rate": 0.000279639409943544, "loss": 0.1284, "step": 4940 }, { "epoch": 1.3400108283703303, "grad_norm": 0.326171875, "learning_rate": 0.0002791841194682208, "loss": 0.1275, "step": 4950 }, { "epoch": 1.3427179209528965, "grad_norm": 0.1318359375, "learning_rate": 0.0002787288289928975, "loss": 0.1327, "step": 4960 }, { "epoch": 1.3454250135354628, "grad_norm": 0.0859375, "learning_rate": 0.00027827353851757417, "loss": 0.1023, "step": 4970 }, { "epoch": 1.3481321061180291, "grad_norm": 0.173828125, "learning_rate": 0.00027781824804225095, "loss": 0.1184, "step": 4980 }, { "epoch": 1.3508391987005957, "grad_norm": 0.26953125, "learning_rate": 0.0002773629575669277, "loss": 0.1258, "step": 4990 }, { "epoch": 1.353546291283162, "grad_norm": 0.26953125, "learning_rate": 0.00027690766709160446, "loss": 0.1334, "step": 5000 }, { "epoch": 1.353546291283162, "eval_loss": 0.15792174637317657, "eval_runtime": 51.3475, "eval_samples_per_second": 9.738, "eval_steps_per_second": 0.312, "step": 5000 }, { "epoch": 1.3562533838657282, "grad_norm": 0.0556640625, "learning_rate": 0.0002764523766162812, "loss": 0.1221, "step": 5010 }, { "epoch": 1.3589604764482945, "grad_norm": 0.1650390625, "learning_rate": 0.0002759970861409579, "loss": 0.1017, "step": 5020 }, { "epoch": 1.3616675690308608, "grad_norm": 0.49609375, "learning_rate": 0.0002755417956656347, "loss": 0.1287, "step": 5030 }, { "epoch": 1.3643746616134271, "grad_norm": 0.1923828125, "learning_rate": 0.0002750865051903114, "loss": 0.1031, "step": 5040 }, { "epoch": 1.3670817541959934, "grad_norm": 0.2197265625, "learning_rate": 0.00027463121471498815, "loss": 0.1034, "step": 5050 }, { "epoch": 1.36978884677856, "grad_norm": 0.267578125, "learning_rate": 0.00027417592423966493, "loss": 0.1253, "step": 5060 }, { "epoch": 1.372495939361126, "grad_norm": 0.6328125, "learning_rate": 0.00027372063376434165, "loss": 0.1762, "step": 5070 }, { "epoch": 1.3752030319436925, "grad_norm": 0.2060546875, "learning_rate": 0.00027326534328901843, "loss": 0.1169, "step": 5080 }, { "epoch": 1.3779101245262588, "grad_norm": 0.51171875, "learning_rate": 0.00027281005281369516, "loss": 0.1402, "step": 5090 }, { "epoch": 1.3806172171088251, "grad_norm": 0.314453125, "learning_rate": 0.0002723547623383719, "loss": 0.1372, "step": 5100 }, { "epoch": 1.3806172171088251, "eval_loss": 0.1598336547613144, "eval_runtime": 50.3768, "eval_samples_per_second": 9.925, "eval_steps_per_second": 0.318, "step": 5100 }, { "epoch": 1.3833243096913914, "grad_norm": 0.54296875, "learning_rate": 0.00027189947186304867, "loss": 0.1452, "step": 5110 }, { "epoch": 1.3860314022739577, "grad_norm": 0.2001953125, "learning_rate": 0.00027144418138772534, "loss": 0.1236, "step": 5120 }, { "epoch": 1.3887384948565242, "grad_norm": 0.13671875, "learning_rate": 0.0002709888909124021, "loss": 0.1157, "step": 5130 }, { "epoch": 1.3914455874390903, "grad_norm": 0.326171875, "learning_rate": 0.00027053360043707885, "loss": 0.1247, "step": 5140 }, { "epoch": 1.3941526800216568, "grad_norm": 0.2490234375, "learning_rate": 0.0002700783099617556, "loss": 0.1463, "step": 5150 }, { "epoch": 1.3968597726042231, "grad_norm": 0.1494140625, "learning_rate": 0.00026962301948643236, "loss": 0.1092, "step": 5160 }, { "epoch": 1.3995668651867894, "grad_norm": 0.35546875, "learning_rate": 0.0002691677290111091, "loss": 0.1224, "step": 5170 }, { "epoch": 1.4022739577693557, "grad_norm": 0.2431640625, "learning_rate": 0.00026871243853578586, "loss": 0.1487, "step": 5180 }, { "epoch": 1.404981050351922, "grad_norm": 0.2060546875, "learning_rate": 0.0002682571480604626, "loss": 0.1434, "step": 5190 }, { "epoch": 1.4076881429344883, "grad_norm": 0.263671875, "learning_rate": 0.0002678018575851393, "loss": 0.1347, "step": 5200 }, { "epoch": 1.4076881429344883, "eval_loss": 0.15952111780643463, "eval_runtime": 49.7815, "eval_samples_per_second": 10.044, "eval_steps_per_second": 0.321, "step": 5200 }, { "epoch": 1.4103952355170546, "grad_norm": 0.1572265625, "learning_rate": 0.0002673465671098161, "loss": 0.1377, "step": 5210 }, { "epoch": 1.413102328099621, "grad_norm": 0.1279296875, "learning_rate": 0.0002668912766344928, "loss": 0.1274, "step": 5220 }, { "epoch": 1.4158094206821874, "grad_norm": 0.25, "learning_rate": 0.00026643598615916955, "loss": 0.133, "step": 5230 }, { "epoch": 1.4185165132647537, "grad_norm": 0.1611328125, "learning_rate": 0.00026598069568384633, "loss": 0.1196, "step": 5240 }, { "epoch": 1.42122360584732, "grad_norm": 0.1787109375, "learning_rate": 0.00026552540520852306, "loss": 0.1445, "step": 5250 }, { "epoch": 1.4239306984298863, "grad_norm": 0.2197265625, "learning_rate": 0.0002650701147331998, "loss": 0.1009, "step": 5260 }, { "epoch": 1.4266377910124526, "grad_norm": 0.1298828125, "learning_rate": 0.0002646148242578765, "loss": 0.1107, "step": 5270 }, { "epoch": 1.4293448835950189, "grad_norm": 0.2333984375, "learning_rate": 0.00026415953378255324, "loss": 0.1158, "step": 5280 }, { "epoch": 1.4320519761775854, "grad_norm": 0.2119140625, "learning_rate": 0.00026370424330723, "loss": 0.1222, "step": 5290 }, { "epoch": 1.4347590687601515, "grad_norm": 0.296875, "learning_rate": 0.00026324895283190674, "loss": 0.1167, "step": 5300 }, { "epoch": 1.4347590687601515, "eval_loss": 0.15868441760540009, "eval_runtime": 50.4922, "eval_samples_per_second": 9.903, "eval_steps_per_second": 0.317, "step": 5300 }, { "epoch": 1.437466161342718, "grad_norm": 0.1630859375, "learning_rate": 0.0002627936623565835, "loss": 0.0958, "step": 5310 }, { "epoch": 1.4401732539252843, "grad_norm": 0.4375, "learning_rate": 0.00026233837188126025, "loss": 0.1646, "step": 5320 }, { "epoch": 1.4428803465078506, "grad_norm": 0.337890625, "learning_rate": 0.000261883081405937, "loss": 0.1436, "step": 5330 }, { "epoch": 1.4455874390904169, "grad_norm": 0.1474609375, "learning_rate": 0.00026142779093061376, "loss": 0.1172, "step": 5340 }, { "epoch": 1.4482945316729832, "grad_norm": 0.1943359375, "learning_rate": 0.0002609725004552905, "loss": 0.112, "step": 5350 }, { "epoch": 1.4510016242555495, "grad_norm": 0.12255859375, "learning_rate": 0.00026051720997996726, "loss": 0.1186, "step": 5360 }, { "epoch": 1.4537087168381158, "grad_norm": 0.119140625, "learning_rate": 0.000260061919504644, "loss": 0.0967, "step": 5370 }, { "epoch": 1.4564158094206823, "grad_norm": 0.28515625, "learning_rate": 0.0002596066290293207, "loss": 0.1182, "step": 5380 }, { "epoch": 1.4591229020032486, "grad_norm": 0.1845703125, "learning_rate": 0.0002591513385539975, "loss": 0.1395, "step": 5390 }, { "epoch": 1.4618299945858149, "grad_norm": 0.1826171875, "learning_rate": 0.00025869604807867417, "loss": 0.1237, "step": 5400 }, { "epoch": 1.4618299945858149, "eval_loss": 0.1595367193222046, "eval_runtime": 50.3465, "eval_samples_per_second": 9.931, "eval_steps_per_second": 0.318, "step": 5400 }, { "epoch": 1.4645370871683812, "grad_norm": 0.11181640625, "learning_rate": 0.0002582407576033509, "loss": 0.118, "step": 5410 }, { "epoch": 1.4672441797509475, "grad_norm": 0.1904296875, "learning_rate": 0.0002577854671280277, "loss": 0.0953, "step": 5420 }, { "epoch": 1.4699512723335137, "grad_norm": 0.22265625, "learning_rate": 0.0002573301766527044, "loss": 0.1219, "step": 5430 }, { "epoch": 1.47265836491608, "grad_norm": 0.140625, "learning_rate": 0.0002568748861773812, "loss": 0.11, "step": 5440 }, { "epoch": 1.4753654574986466, "grad_norm": 0.23828125, "learning_rate": 0.0002564195957020579, "loss": 0.1108, "step": 5450 }, { "epoch": 1.4780725500812129, "grad_norm": 0.455078125, "learning_rate": 0.00025596430522673464, "loss": 0.1313, "step": 5460 }, { "epoch": 1.4807796426637791, "grad_norm": 0.2734375, "learning_rate": 0.0002555090147514114, "loss": 0.163, "step": 5470 }, { "epoch": 1.4834867352463454, "grad_norm": 0.291015625, "learning_rate": 0.00025505372427608815, "loss": 0.138, "step": 5480 }, { "epoch": 1.4861938278289117, "grad_norm": 0.103515625, "learning_rate": 0.0002545984338007649, "loss": 0.1257, "step": 5490 }, { "epoch": 1.488900920411478, "grad_norm": 0.150390625, "learning_rate": 0.00025414314332544165, "loss": 0.1304, "step": 5500 }, { "epoch": 1.488900920411478, "eval_loss": 0.16379128396511078, "eval_runtime": 51.0248, "eval_samples_per_second": 9.799, "eval_steps_per_second": 0.314, "step": 5500 }, { "epoch": 1.4916080129940443, "grad_norm": 0.2197265625, "learning_rate": 0.0002536878528501184, "loss": 0.1124, "step": 5510 }, { "epoch": 1.4943151055766108, "grad_norm": 0.298828125, "learning_rate": 0.00025323256237479516, "loss": 0.1048, "step": 5520 }, { "epoch": 1.497022198159177, "grad_norm": 0.2578125, "learning_rate": 0.0002527772718994719, "loss": 0.1251, "step": 5530 }, { "epoch": 1.4997292907417434, "grad_norm": 0.51953125, "learning_rate": 0.0002523219814241486, "loss": 0.1434, "step": 5540 }, { "epoch": 1.5024363833243097, "grad_norm": 0.35546875, "learning_rate": 0.00025186669094882534, "loss": 0.1254, "step": 5550 }, { "epoch": 1.505143475906876, "grad_norm": 0.287109375, "learning_rate": 0.00025141140047350207, "loss": 0.0985, "step": 5560 }, { "epoch": 1.5078505684894423, "grad_norm": 0.287109375, "learning_rate": 0.00025095610999817885, "loss": 0.1109, "step": 5570 }, { "epoch": 1.5105576610720086, "grad_norm": 0.171875, "learning_rate": 0.0002505008195228556, "loss": 0.116, "step": 5580 }, { "epoch": 1.5132647536545751, "grad_norm": 0.396484375, "learning_rate": 0.0002500455290475323, "loss": 0.1411, "step": 5590 }, { "epoch": 1.5159718462371412, "grad_norm": 0.154296875, "learning_rate": 0.0002495902385722091, "loss": 0.1172, "step": 5600 }, { "epoch": 1.5159718462371412, "eval_loss": 0.15981900691986084, "eval_runtime": 51.7233, "eval_samples_per_second": 9.667, "eval_steps_per_second": 0.309, "step": 5600 }, { "epoch": 1.5186789388197077, "grad_norm": 0.306640625, "learning_rate": 0.0002491349480968858, "loss": 0.1516, "step": 5610 }, { "epoch": 1.5213860314022738, "grad_norm": 0.2255859375, "learning_rate": 0.00024867965762156253, "loss": 0.1095, "step": 5620 }, { "epoch": 1.5240931239848403, "grad_norm": 0.54296875, "learning_rate": 0.0002482243671462393, "loss": 0.125, "step": 5630 }, { "epoch": 1.5268002165674066, "grad_norm": 0.16796875, "learning_rate": 0.00024776907667091604, "loss": 0.1146, "step": 5640 }, { "epoch": 1.529507309149973, "grad_norm": 0.28125, "learning_rate": 0.0002473137861955928, "loss": 0.1103, "step": 5650 }, { "epoch": 1.5322144017325392, "grad_norm": 0.2734375, "learning_rate": 0.00024685849572026955, "loss": 0.1014, "step": 5660 }, { "epoch": 1.5349214943151055, "grad_norm": 0.203125, "learning_rate": 0.0002464032052449463, "loss": 0.1246, "step": 5670 }, { "epoch": 1.537628586897672, "grad_norm": 0.298828125, "learning_rate": 0.000245947914769623, "loss": 0.1121, "step": 5680 }, { "epoch": 1.540335679480238, "grad_norm": 0.255859375, "learning_rate": 0.0002454926242942998, "loss": 0.1614, "step": 5690 }, { "epoch": 1.5430427720628046, "grad_norm": 0.259765625, "learning_rate": 0.0002450373338189765, "loss": 0.1091, "step": 5700 }, { "epoch": 1.5430427720628046, "eval_loss": 0.16020576655864716, "eval_runtime": 51.2639, "eval_samples_per_second": 9.753, "eval_steps_per_second": 0.312, "step": 5700 }, { "epoch": 1.545749864645371, "grad_norm": 0.1533203125, "learning_rate": 0.00024458204334365324, "loss": 0.1055, "step": 5710 }, { "epoch": 1.5484569572279372, "grad_norm": 0.1044921875, "learning_rate": 0.00024412675286833, "loss": 0.0968, "step": 5720 }, { "epoch": 1.5511640498105035, "grad_norm": 0.11767578125, "learning_rate": 0.00024367146239300674, "loss": 0.1163, "step": 5730 }, { "epoch": 1.5538711423930698, "grad_norm": 0.498046875, "learning_rate": 0.0002432161719176835, "loss": 0.1126, "step": 5740 }, { "epoch": 1.5565782349756363, "grad_norm": 0.193359375, "learning_rate": 0.00024276088144236022, "loss": 0.1062, "step": 5750 }, { "epoch": 1.5592853275582024, "grad_norm": 0.484375, "learning_rate": 0.00024230559096703698, "loss": 0.1405, "step": 5760 }, { "epoch": 1.561992420140769, "grad_norm": 0.25390625, "learning_rate": 0.00024185030049171373, "loss": 0.1097, "step": 5770 }, { "epoch": 1.5646995127233352, "grad_norm": 0.203125, "learning_rate": 0.00024139501001639048, "loss": 0.1447, "step": 5780 }, { "epoch": 1.5674066053059015, "grad_norm": 0.2734375, "learning_rate": 0.00024093971954106718, "loss": 0.1413, "step": 5790 }, { "epoch": 1.5701136978884678, "grad_norm": 0.267578125, "learning_rate": 0.00024048442906574394, "loss": 0.1561, "step": 5800 }, { "epoch": 1.5701136978884678, "eval_loss": 0.16192464530467987, "eval_runtime": 49.5357, "eval_samples_per_second": 10.094, "eval_steps_per_second": 0.323, "step": 5800 }, { "epoch": 1.572820790471034, "grad_norm": 0.330078125, "learning_rate": 0.0002400291385904207, "loss": 0.1137, "step": 5810 }, { "epoch": 1.5755278830536006, "grad_norm": 0.2216796875, "learning_rate": 0.00023957384811509744, "loss": 0.1074, "step": 5820 }, { "epoch": 1.5782349756361667, "grad_norm": 0.30078125, "learning_rate": 0.0002391185576397742, "loss": 0.1498, "step": 5830 }, { "epoch": 1.5809420682187332, "grad_norm": 0.2109375, "learning_rate": 0.00023866326716445092, "loss": 0.1271, "step": 5840 }, { "epoch": 1.5836491608012992, "grad_norm": 0.28515625, "learning_rate": 0.00023820797668912768, "loss": 0.106, "step": 5850 }, { "epoch": 1.5863562533838658, "grad_norm": 0.359375, "learning_rate": 0.0002377526862138044, "loss": 0.1288, "step": 5860 }, { "epoch": 1.589063345966432, "grad_norm": 0.1455078125, "learning_rate": 0.00023729739573848116, "loss": 0.1038, "step": 5870 }, { "epoch": 1.5917704385489984, "grad_norm": 0.431640625, "learning_rate": 0.00023684210526315788, "loss": 0.127, "step": 5880 }, { "epoch": 1.5944775311315647, "grad_norm": 0.353515625, "learning_rate": 0.00023638681478783464, "loss": 0.093, "step": 5890 }, { "epoch": 1.597184623714131, "grad_norm": 0.32421875, "learning_rate": 0.0002359315243125114, "loss": 0.111, "step": 5900 }, { "epoch": 1.597184623714131, "eval_loss": 0.16383513808250427, "eval_runtime": 50.6949, "eval_samples_per_second": 9.863, "eval_steps_per_second": 0.316, "step": 5900 }, { "epoch": 1.5998917162966975, "grad_norm": 0.40625, "learning_rate": 0.00023547623383718814, "loss": 0.1285, "step": 5910 }, { "epoch": 1.6025988088792635, "grad_norm": 0.3125, "learning_rate": 0.0002350209433618649, "loss": 0.1296, "step": 5920 }, { "epoch": 1.60530590146183, "grad_norm": 0.22265625, "learning_rate": 0.0002345656528865416, "loss": 0.1152, "step": 5930 }, { "epoch": 1.6080129940443963, "grad_norm": 0.10546875, "learning_rate": 0.00023411036241121835, "loss": 0.1388, "step": 5940 }, { "epoch": 1.6107200866269626, "grad_norm": 0.35546875, "learning_rate": 0.0002336550719358951, "loss": 0.1083, "step": 5950 }, { "epoch": 1.613427179209529, "grad_norm": 0.455078125, "learning_rate": 0.00023319978146057186, "loss": 0.1388, "step": 5960 }, { "epoch": 1.6161342717920952, "grad_norm": 0.26171875, "learning_rate": 0.00023274449098524859, "loss": 0.1116, "step": 5970 }, { "epoch": 1.6188413643746618, "grad_norm": 0.484375, "learning_rate": 0.00023228920050992534, "loss": 0.128, "step": 5980 }, { "epoch": 1.6215484569572278, "grad_norm": 0.08984375, "learning_rate": 0.0002318339100346021, "loss": 0.0834, "step": 5990 }, { "epoch": 1.6242555495397943, "grad_norm": 0.296875, "learning_rate": 0.00023137861955927882, "loss": 0.1207, "step": 6000 }, { "epoch": 1.6242555495397943, "eval_loss": 0.15791058540344238, "eval_runtime": 50.2164, "eval_samples_per_second": 9.957, "eval_steps_per_second": 0.319, "step": 6000 }, { "epoch": 1.6269626421223606, "grad_norm": 0.1484375, "learning_rate": 0.00023092332908395557, "loss": 0.119, "step": 6010 }, { "epoch": 1.629669734704927, "grad_norm": 0.125, "learning_rate": 0.0002304680386086323, "loss": 0.1005, "step": 6020 }, { "epoch": 1.6323768272874932, "grad_norm": 0.37109375, "learning_rate": 0.00023001274813330905, "loss": 0.1397, "step": 6030 }, { "epoch": 1.6350839198700595, "grad_norm": 0.1337890625, "learning_rate": 0.0002295574576579858, "loss": 0.1025, "step": 6040 }, { "epoch": 1.637791012452626, "grad_norm": 0.1416015625, "learning_rate": 0.00022910216718266256, "loss": 0.1437, "step": 6050 }, { "epoch": 1.6404981050351921, "grad_norm": 0.271484375, "learning_rate": 0.0002286468767073393, "loss": 0.1037, "step": 6060 }, { "epoch": 1.6432051976177586, "grad_norm": 0.078125, "learning_rate": 0.000228191586232016, "loss": 0.0911, "step": 6070 }, { "epoch": 1.6459122902003247, "grad_norm": 0.2294921875, "learning_rate": 0.00022773629575669277, "loss": 0.0996, "step": 6080 }, { "epoch": 1.6486193827828912, "grad_norm": 0.439453125, "learning_rate": 0.00022728100528136952, "loss": 0.1274, "step": 6090 }, { "epoch": 1.6513264753654575, "grad_norm": 0.29296875, "learning_rate": 0.00022682571480604627, "loss": 0.1214, "step": 6100 }, { "epoch": 1.6513264753654575, "eval_loss": 0.15438494086265564, "eval_runtime": 49.8084, "eval_samples_per_second": 10.038, "eval_steps_per_second": 0.321, "step": 6100 }, { "epoch": 1.6540335679480238, "grad_norm": 0.115234375, "learning_rate": 0.000226370424330723, "loss": 0.1349, "step": 6110 }, { "epoch": 1.65674066053059, "grad_norm": 0.328125, "learning_rate": 0.00022591513385539975, "loss": 0.1153, "step": 6120 }, { "epoch": 1.6594477531131564, "grad_norm": 0.380859375, "learning_rate": 0.0002254598433800765, "loss": 0.1108, "step": 6130 }, { "epoch": 1.662154845695723, "grad_norm": 0.1748046875, "learning_rate": 0.00022500455290475323, "loss": 0.115, "step": 6140 }, { "epoch": 1.664861938278289, "grad_norm": 0.35546875, "learning_rate": 0.00022454926242942996, "loss": 0.1185, "step": 6150 }, { "epoch": 1.6675690308608555, "grad_norm": 0.2109375, "learning_rate": 0.00022409397195410671, "loss": 0.1193, "step": 6160 }, { "epoch": 1.6702761234434218, "grad_norm": 0.16796875, "learning_rate": 0.00022363868147878347, "loss": 0.1319, "step": 6170 }, { "epoch": 1.672983216025988, "grad_norm": 0.1796875, "learning_rate": 0.00022318339100346022, "loss": 0.1, "step": 6180 }, { "epoch": 1.6756903086085544, "grad_norm": 0.455078125, "learning_rate": 0.00022272810052813698, "loss": 0.1286, "step": 6190 }, { "epoch": 1.6783974011911207, "grad_norm": 0.271484375, "learning_rate": 0.0002222728100528137, "loss": 0.1062, "step": 6200 }, { "epoch": 1.6783974011911207, "eval_loss": 0.1509903371334076, "eval_runtime": 50.1548, "eval_samples_per_second": 9.969, "eval_steps_per_second": 0.319, "step": 6200 }, { "epoch": 1.6811044937736872, "grad_norm": 0.12109375, "learning_rate": 0.00022181751957749043, "loss": 0.1115, "step": 6210 }, { "epoch": 1.6838115863562533, "grad_norm": 0.2275390625, "learning_rate": 0.00022136222910216718, "loss": 0.0958, "step": 6220 }, { "epoch": 1.6865186789388198, "grad_norm": 0.419921875, "learning_rate": 0.00022090693862684394, "loss": 0.1072, "step": 6230 }, { "epoch": 1.6892257715213859, "grad_norm": 0.490234375, "learning_rate": 0.00022045164815152066, "loss": 0.1338, "step": 6240 }, { "epoch": 1.6919328641039524, "grad_norm": 0.384765625, "learning_rate": 0.00021999635767619742, "loss": 0.1025, "step": 6250 }, { "epoch": 1.6946399566865187, "grad_norm": 0.31640625, "learning_rate": 0.00021954106720087417, "loss": 0.1323, "step": 6260 }, { "epoch": 1.697347049269085, "grad_norm": 0.47265625, "learning_rate": 0.00021908577672555092, "loss": 0.1312, "step": 6270 }, { "epoch": 1.7000541418516515, "grad_norm": 0.248046875, "learning_rate": 0.00021863048625022765, "loss": 0.1137, "step": 6280 }, { "epoch": 1.7027612344342176, "grad_norm": 0.09375, "learning_rate": 0.00021817519577490438, "loss": 0.1201, "step": 6290 }, { "epoch": 1.705468327016784, "grad_norm": 0.298828125, "learning_rate": 0.00021771990529958113, "loss": 0.1161, "step": 6300 }, { "epoch": 1.705468327016784, "eval_loss": 0.1535915732383728, "eval_runtime": 51.7815, "eval_samples_per_second": 9.656, "eval_steps_per_second": 0.309, "step": 6300 }, { "epoch": 1.7081754195993502, "grad_norm": 0.1884765625, "learning_rate": 0.00021726461482425788, "loss": 0.109, "step": 6310 }, { "epoch": 1.7108825121819167, "grad_norm": 0.13671875, "learning_rate": 0.00021680932434893464, "loss": 0.1241, "step": 6320 }, { "epoch": 1.713589604764483, "grad_norm": 0.111328125, "learning_rate": 0.00021635403387361136, "loss": 0.1272, "step": 6330 }, { "epoch": 1.7162966973470493, "grad_norm": 0.5390625, "learning_rate": 0.00021589874339828812, "loss": 0.1066, "step": 6340 }, { "epoch": 1.7190037899296156, "grad_norm": 0.494140625, "learning_rate": 0.00021544345292296484, "loss": 0.1083, "step": 6350 }, { "epoch": 1.7217108825121819, "grad_norm": 0.228515625, "learning_rate": 0.0002149881624476416, "loss": 0.1119, "step": 6360 }, { "epoch": 1.7244179750947484, "grad_norm": 0.1494140625, "learning_rate": 0.00021453287197231835, "loss": 0.1142, "step": 6370 }, { "epoch": 1.7271250676773144, "grad_norm": 0.15234375, "learning_rate": 0.00021407758149699508, "loss": 0.1237, "step": 6380 }, { "epoch": 1.729832160259881, "grad_norm": 0.283203125, "learning_rate": 0.00021362229102167183, "loss": 0.1137, "step": 6390 }, { "epoch": 1.7325392528424473, "grad_norm": 0.224609375, "learning_rate": 0.00021316700054634858, "loss": 0.115, "step": 6400 }, { "epoch": 1.7325392528424473, "eval_loss": 0.15494155883789062, "eval_runtime": 51.5495, "eval_samples_per_second": 9.699, "eval_steps_per_second": 0.31, "step": 6400 }, { "epoch": 1.7352463454250135, "grad_norm": 0.07666015625, "learning_rate": 0.00021271171007102534, "loss": 0.1437, "step": 6410 }, { "epoch": 1.7379534380075798, "grad_norm": 0.55078125, "learning_rate": 0.00021225641959570204, "loss": 0.1266, "step": 6420 }, { "epoch": 1.7406605305901461, "grad_norm": 0.1025390625, "learning_rate": 0.0002118011291203788, "loss": 0.0967, "step": 6430 }, { "epoch": 1.7433676231727127, "grad_norm": 0.126953125, "learning_rate": 0.00021134583864505554, "loss": 0.1034, "step": 6440 }, { "epoch": 1.7460747157552787, "grad_norm": 0.27734375, "learning_rate": 0.0002108905481697323, "loss": 0.1016, "step": 6450 }, { "epoch": 1.7487818083378452, "grad_norm": 0.1806640625, "learning_rate": 0.00021043525769440905, "loss": 0.0758, "step": 6460 }, { "epoch": 1.7514889009204113, "grad_norm": 0.2275390625, "learning_rate": 0.00020997996721908578, "loss": 0.1228, "step": 6470 }, { "epoch": 1.7541959935029778, "grad_norm": 0.1240234375, "learning_rate": 0.00020952467674376253, "loss": 0.0879, "step": 6480 }, { "epoch": 1.7569030860855441, "grad_norm": 0.140625, "learning_rate": 0.00020906938626843926, "loss": 0.1235, "step": 6490 }, { "epoch": 1.7596101786681104, "grad_norm": 0.25390625, "learning_rate": 0.000208614095793116, "loss": 0.0861, "step": 6500 }, { "epoch": 1.7596101786681104, "eval_loss": 0.15266965329647064, "eval_runtime": 49.6548, "eval_samples_per_second": 10.07, "eval_steps_per_second": 0.322, "step": 6500 }, { "epoch": 1.7623172712506767, "grad_norm": 0.419921875, "learning_rate": 0.00020815880531779274, "loss": 0.1356, "step": 6510 }, { "epoch": 1.765024363833243, "grad_norm": 0.2294921875, "learning_rate": 0.0002077035148424695, "loss": 0.1153, "step": 6520 }, { "epoch": 1.7677314564158095, "grad_norm": 0.1455078125, "learning_rate": 0.00020724822436714625, "loss": 0.1144, "step": 6530 }, { "epoch": 1.7704385489983756, "grad_norm": 0.1904296875, "learning_rate": 0.000206792933891823, "loss": 0.1145, "step": 6540 }, { "epoch": 1.7731456415809421, "grad_norm": 0.42578125, "learning_rate": 0.00020633764341649975, "loss": 0.1313, "step": 6550 }, { "epoch": 1.7758527341635084, "grad_norm": 0.26171875, "learning_rate": 0.00020588235294117645, "loss": 0.0949, "step": 6560 }, { "epoch": 1.7785598267460747, "grad_norm": 0.337890625, "learning_rate": 0.0002054270624658532, "loss": 0.1124, "step": 6570 }, { "epoch": 1.781266919328641, "grad_norm": 0.2578125, "learning_rate": 0.00020497177199052996, "loss": 0.1421, "step": 6580 }, { "epoch": 1.7839740119112073, "grad_norm": 0.154296875, "learning_rate": 0.0002045164815152067, "loss": 0.1376, "step": 6590 }, { "epoch": 1.7866811044937738, "grad_norm": 0.1484375, "learning_rate": 0.00020406119103988344, "loss": 0.1102, "step": 6600 }, { "epoch": 1.7866811044937738, "eval_loss": 0.1520686000585556, "eval_runtime": 52.089, "eval_samples_per_second": 9.599, "eval_steps_per_second": 0.307, "step": 6600 }, { "epoch": 1.78938819707634, "grad_norm": 0.193359375, "learning_rate": 0.0002036059005645602, "loss": 0.1197, "step": 6610 }, { "epoch": 1.7920952896589064, "grad_norm": 0.201171875, "learning_rate": 0.00020315061008923695, "loss": 0.1083, "step": 6620 }, { "epoch": 1.7948023822414727, "grad_norm": 0.30859375, "learning_rate": 0.00020269531961391367, "loss": 0.1413, "step": 6630 }, { "epoch": 1.797509474824039, "grad_norm": 0.169921875, "learning_rate": 0.00020224002913859043, "loss": 0.1038, "step": 6640 }, { "epoch": 1.8002165674066053, "grad_norm": 0.154296875, "learning_rate": 0.00020178473866326715, "loss": 0.1429, "step": 6650 }, { "epoch": 1.8029236599891716, "grad_norm": 0.173828125, "learning_rate": 0.0002013294481879439, "loss": 0.1182, "step": 6660 }, { "epoch": 1.805630752571738, "grad_norm": 0.1044921875, "learning_rate": 0.00020087415771262066, "loss": 0.109, "step": 6670 }, { "epoch": 1.8083378451543042, "grad_norm": 0.115234375, "learning_rate": 0.00020041886723729741, "loss": 0.1167, "step": 6680 }, { "epoch": 1.8110449377368707, "grad_norm": 0.2177734375, "learning_rate": 0.00019996357676197414, "loss": 0.1191, "step": 6690 }, { "epoch": 1.8137520303194368, "grad_norm": 0.248046875, "learning_rate": 0.00019950828628665087, "loss": 0.1249, "step": 6700 }, { "epoch": 1.8137520303194368, "eval_loss": 0.14892610907554626, "eval_runtime": 50.9547, "eval_samples_per_second": 9.813, "eval_steps_per_second": 0.314, "step": 6700 }, { "epoch": 1.8164591229020033, "grad_norm": 0.39453125, "learning_rate": 0.00019905299581132762, "loss": 0.1164, "step": 6710 }, { "epoch": 1.8191662154845696, "grad_norm": 0.287109375, "learning_rate": 0.00019859770533600437, "loss": 0.1257, "step": 6720 }, { "epoch": 1.8218733080671359, "grad_norm": 0.25390625, "learning_rate": 0.00019814241486068113, "loss": 0.1283, "step": 6730 }, { "epoch": 1.8245804006497022, "grad_norm": 0.1650390625, "learning_rate": 0.00019768712438535786, "loss": 0.1121, "step": 6740 }, { "epoch": 1.8272874932322685, "grad_norm": 0.5546875, "learning_rate": 0.0001972318339100346, "loss": 0.0988, "step": 6750 }, { "epoch": 1.829994585814835, "grad_norm": 0.275390625, "learning_rate": 0.00019677654343471136, "loss": 0.0845, "step": 6760 }, { "epoch": 1.832701678397401, "grad_norm": 0.30078125, "learning_rate": 0.00019632125295938812, "loss": 0.1241, "step": 6770 }, { "epoch": 1.8354087709799676, "grad_norm": 0.2236328125, "learning_rate": 0.00019586596248406482, "loss": 0.1042, "step": 6780 }, { "epoch": 1.8381158635625339, "grad_norm": 0.173828125, "learning_rate": 0.00019541067200874157, "loss": 0.0877, "step": 6790 }, { "epoch": 1.8408229561451002, "grad_norm": 0.150390625, "learning_rate": 0.00019495538153341832, "loss": 0.1287, "step": 6800 }, { "epoch": 1.8408229561451002, "eval_loss": 0.1503976583480835, "eval_runtime": 50.5525, "eval_samples_per_second": 9.891, "eval_steps_per_second": 0.317, "step": 6800 }, { "epoch": 1.8435300487276665, "grad_norm": 0.248046875, "learning_rate": 0.00019450009105809508, "loss": 0.1022, "step": 6810 }, { "epoch": 1.8462371413102328, "grad_norm": 0.11572265625, "learning_rate": 0.00019404480058277183, "loss": 0.1109, "step": 6820 }, { "epoch": 1.8489442338927993, "grad_norm": 0.123046875, "learning_rate": 0.00019358951010744856, "loss": 0.1186, "step": 6830 }, { "epoch": 1.8516513264753653, "grad_norm": 0.267578125, "learning_rate": 0.0001931342196321253, "loss": 0.1238, "step": 6840 }, { "epoch": 1.8543584190579319, "grad_norm": 0.373046875, "learning_rate": 0.00019267892915680204, "loss": 0.1492, "step": 6850 }, { "epoch": 1.8570655116404982, "grad_norm": 0.1806640625, "learning_rate": 0.0001922236386814788, "loss": 0.1294, "step": 6860 }, { "epoch": 1.8597726042230645, "grad_norm": 0.2109375, "learning_rate": 0.00019176834820615552, "loss": 0.1256, "step": 6870 }, { "epoch": 1.8624796968056307, "grad_norm": 0.140625, "learning_rate": 0.00019131305773083227, "loss": 0.1055, "step": 6880 }, { "epoch": 1.865186789388197, "grad_norm": 0.255859375, "learning_rate": 0.00019085776725550902, "loss": 0.1017, "step": 6890 }, { "epoch": 1.8678938819707636, "grad_norm": 0.375, "learning_rate": 0.00019040247678018578, "loss": 0.1311, "step": 6900 }, { "epoch": 1.8678938819707636, "eval_loss": 0.15065474808216095, "eval_runtime": 51.105, "eval_samples_per_second": 9.784, "eval_steps_per_second": 0.313, "step": 6900 }, { "epoch": 1.8706009745533296, "grad_norm": 0.19140625, "learning_rate": 0.00018994718630486253, "loss": 0.1184, "step": 6910 }, { "epoch": 1.8733080671358961, "grad_norm": 0.275390625, "learning_rate": 0.00018949189582953923, "loss": 0.0929, "step": 6920 }, { "epoch": 1.8760151597184622, "grad_norm": 0.099609375, "learning_rate": 0.00018903660535421598, "loss": 0.1177, "step": 6930 }, { "epoch": 1.8787222523010287, "grad_norm": 0.150390625, "learning_rate": 0.00018858131487889274, "loss": 0.1116, "step": 6940 }, { "epoch": 1.881429344883595, "grad_norm": 0.33203125, "learning_rate": 0.0001881260244035695, "loss": 0.1492, "step": 6950 }, { "epoch": 1.8841364374661613, "grad_norm": 0.234375, "learning_rate": 0.00018767073392824622, "loss": 0.1335, "step": 6960 }, { "epoch": 1.8868435300487276, "grad_norm": 0.3984375, "learning_rate": 0.00018721544345292297, "loss": 0.1445, "step": 6970 }, { "epoch": 1.889550622631294, "grad_norm": 0.2177734375, "learning_rate": 0.00018676015297759973, "loss": 0.1011, "step": 6980 }, { "epoch": 1.8922577152138604, "grad_norm": 0.059326171875, "learning_rate": 0.00018630486250227645, "loss": 0.1133, "step": 6990 }, { "epoch": 1.8949648077964265, "grad_norm": 0.0654296875, "learning_rate": 0.0001858495720269532, "loss": 0.0976, "step": 7000 }, { "epoch": 1.8949648077964265, "eval_loss": 0.15026648342609406, "eval_runtime": 50.4333, "eval_samples_per_second": 9.914, "eval_steps_per_second": 0.317, "step": 7000 }, { "epoch": 1.897671900378993, "grad_norm": 0.08056640625, "learning_rate": 0.00018539428155162993, "loss": 0.092, "step": 7010 }, { "epoch": 1.9003789929615593, "grad_norm": 0.1416015625, "learning_rate": 0.00018493899107630669, "loss": 0.0998, "step": 7020 }, { "epoch": 1.9030860855441256, "grad_norm": 0.267578125, "learning_rate": 0.00018448370060098344, "loss": 0.1095, "step": 7030 }, { "epoch": 1.905793178126692, "grad_norm": 0.1474609375, "learning_rate": 0.0001840284101256602, "loss": 0.119, "step": 7040 }, { "epoch": 1.9085002707092582, "grad_norm": 0.2421875, "learning_rate": 0.00018357311965033692, "loss": 0.1162, "step": 7050 }, { "epoch": 1.9112073632918247, "grad_norm": 0.3515625, "learning_rate": 0.00018311782917501365, "loss": 0.117, "step": 7060 }, { "epoch": 1.9139144558743908, "grad_norm": 0.181640625, "learning_rate": 0.0001826625386996904, "loss": 0.1154, "step": 7070 }, { "epoch": 1.9166215484569573, "grad_norm": 0.322265625, "learning_rate": 0.00018220724822436715, "loss": 0.1305, "step": 7080 }, { "epoch": 1.9193286410395234, "grad_norm": 0.388671875, "learning_rate": 0.0001817519577490439, "loss": 0.1165, "step": 7090 }, { "epoch": 1.92203573362209, "grad_norm": 0.0986328125, "learning_rate": 0.00018129666727372063, "loss": 0.1134, "step": 7100 }, { "epoch": 1.92203573362209, "eval_loss": 0.15066738426685333, "eval_runtime": 50.4193, "eval_samples_per_second": 9.917, "eval_steps_per_second": 0.317, "step": 7100 }, { "epoch": 1.9247428262046562, "grad_norm": 0.322265625, "learning_rate": 0.0001808413767983974, "loss": 0.1113, "step": 7110 }, { "epoch": 1.9274499187872225, "grad_norm": 0.212890625, "learning_rate": 0.00018038608632307414, "loss": 0.0985, "step": 7120 }, { "epoch": 1.930157011369789, "grad_norm": 0.185546875, "learning_rate": 0.00017993079584775087, "loss": 0.1138, "step": 7130 }, { "epoch": 1.932864103952355, "grad_norm": 0.29296875, "learning_rate": 0.0001794755053724276, "loss": 0.1113, "step": 7140 }, { "epoch": 1.9355711965349216, "grad_norm": 0.1259765625, "learning_rate": 0.00017902021489710435, "loss": 0.1194, "step": 7150 }, { "epoch": 1.9382782891174877, "grad_norm": 0.353515625, "learning_rate": 0.0001785649244217811, "loss": 0.1109, "step": 7160 }, { "epoch": 1.9409853817000542, "grad_norm": 0.32421875, "learning_rate": 0.00017810963394645785, "loss": 0.1055, "step": 7170 }, { "epoch": 1.9436924742826205, "grad_norm": 0.197265625, "learning_rate": 0.0001776543434711346, "loss": 0.1217, "step": 7180 }, { "epoch": 1.9463995668651868, "grad_norm": 0.34765625, "learning_rate": 0.00017719905299581133, "loss": 0.1275, "step": 7190 }, { "epoch": 1.949106659447753, "grad_norm": 0.0810546875, "learning_rate": 0.00017674376252048806, "loss": 0.0839, "step": 7200 }, { "epoch": 1.949106659447753, "eval_loss": 0.15255558490753174, "eval_runtime": 50.1937, "eval_samples_per_second": 9.961, "eval_steps_per_second": 0.319, "step": 7200 }, { "epoch": 1.9518137520303194, "grad_norm": 0.29296875, "learning_rate": 0.00017628847204516481, "loss": 0.1094, "step": 7210 }, { "epoch": 1.9545208446128859, "grad_norm": 0.482421875, "learning_rate": 0.00017583318156984157, "loss": 0.1362, "step": 7220 }, { "epoch": 1.957227937195452, "grad_norm": 0.3984375, "learning_rate": 0.0001753778910945183, "loss": 0.115, "step": 7230 }, { "epoch": 1.9599350297780185, "grad_norm": 0.49609375, "learning_rate": 0.00017492260061919505, "loss": 0.1188, "step": 7240 }, { "epoch": 1.9626421223605848, "grad_norm": 0.16796875, "learning_rate": 0.0001744673101438718, "loss": 0.1114, "step": 7250 }, { "epoch": 1.965349214943151, "grad_norm": 0.26953125, "learning_rate": 0.00017401201966854856, "loss": 0.0983, "step": 7260 }, { "epoch": 1.9680563075257174, "grad_norm": 0.216796875, "learning_rate": 0.00017355672919322528, "loss": 0.135, "step": 7270 }, { "epoch": 1.9707634001082837, "grad_norm": 0.19140625, "learning_rate": 0.000173101438717902, "loss": 0.1057, "step": 7280 }, { "epoch": 1.9734704926908502, "grad_norm": 0.2734375, "learning_rate": 0.00017264614824257876, "loss": 0.1421, "step": 7290 }, { "epoch": 1.9761775852734162, "grad_norm": 0.298828125, "learning_rate": 0.00017219085776725552, "loss": 0.0989, "step": 7300 }, { "epoch": 1.9761775852734162, "eval_loss": 0.15032486617565155, "eval_runtime": 50.1976, "eval_samples_per_second": 9.961, "eval_steps_per_second": 0.319, "step": 7300 }, { "epoch": 1.9788846778559828, "grad_norm": 0.2490234375, "learning_rate": 0.00017173556729193227, "loss": 0.1003, "step": 7310 }, { "epoch": 1.9815917704385488, "grad_norm": 0.111328125, "learning_rate": 0.000171280276816609, "loss": 0.106, "step": 7320 }, { "epoch": 1.9842988630211154, "grad_norm": 0.251953125, "learning_rate": 0.00017082498634128575, "loss": 0.1063, "step": 7330 }, { "epoch": 1.9870059556036817, "grad_norm": 0.1875, "learning_rate": 0.00017036969586596248, "loss": 0.0799, "step": 7340 }, { "epoch": 1.989713048186248, "grad_norm": 0.171875, "learning_rate": 0.00016991440539063923, "loss": 0.1114, "step": 7350 }, { "epoch": 1.9924201407688142, "grad_norm": 0.205078125, "learning_rate": 0.00016945911491531598, "loss": 0.1166, "step": 7360 }, { "epoch": 1.9951272333513805, "grad_norm": 0.369140625, "learning_rate": 0.0001690038244399927, "loss": 0.1113, "step": 7370 }, { "epoch": 1.997834325933947, "grad_norm": 0.37890625, "learning_rate": 0.00016854853396466946, "loss": 0.084, "step": 7380 }, { "epoch": 2.000541418516513, "grad_norm": 0.173828125, "learning_rate": 0.00016809324348934622, "loss": 0.1154, "step": 7390 }, { "epoch": 2.0032485110990796, "grad_norm": 0.24609375, "learning_rate": 0.00016763795301402297, "loss": 0.0998, "step": 7400 }, { "epoch": 2.0032485110990796, "eval_loss": 0.1515250951051712, "eval_runtime": 51.4661, "eval_samples_per_second": 9.715, "eval_steps_per_second": 0.311, "step": 7400 }, { "epoch": 2.0059556036816457, "grad_norm": 0.234375, "learning_rate": 0.00016718266253869967, "loss": 0.0854, "step": 7410 }, { "epoch": 2.0086626962642122, "grad_norm": 0.1845703125, "learning_rate": 0.00016672737206337642, "loss": 0.0835, "step": 7420 }, { "epoch": 2.0113697888467788, "grad_norm": 0.234375, "learning_rate": 0.00016627208158805318, "loss": 0.1461, "step": 7430 }, { "epoch": 2.014076881429345, "grad_norm": 0.30078125, "learning_rate": 0.00016581679111272993, "loss": 0.0953, "step": 7440 }, { "epoch": 2.0167839740119113, "grad_norm": 0.0830078125, "learning_rate": 0.00016536150063740668, "loss": 0.0911, "step": 7450 }, { "epoch": 2.0194910665944774, "grad_norm": 0.1630859375, "learning_rate": 0.0001649062101620834, "loss": 0.0795, "step": 7460 }, { "epoch": 2.022198159177044, "grad_norm": 0.2890625, "learning_rate": 0.00016445091968676016, "loss": 0.1355, "step": 7470 }, { "epoch": 2.02490525175961, "grad_norm": 0.1904296875, "learning_rate": 0.0001639956292114369, "loss": 0.0898, "step": 7480 }, { "epoch": 2.0276123443421765, "grad_norm": 0.1064453125, "learning_rate": 0.00016354033873611364, "loss": 0.1181, "step": 7490 }, { "epoch": 2.030319436924743, "grad_norm": 0.138671875, "learning_rate": 0.00016308504826079037, "loss": 0.1093, "step": 7500 }, { "epoch": 2.030319436924743, "eval_loss": 0.15121085941791534, "eval_runtime": 50.3881, "eval_samples_per_second": 9.923, "eval_steps_per_second": 0.318, "step": 7500 }, { "epoch": 2.033026529507309, "grad_norm": 0.16410693526268005, "learning_rate": 0.00016262975778546712, "loss": 0.1239, "step": 7510 }, { "epoch": 2.0357336220898756, "grad_norm": 0.2050447016954422, "learning_rate": 0.00016217446731014388, "loss": 0.1089, "step": 7520 }, { "epoch": 2.0384407146724417, "grad_norm": 0.3048134446144104, "learning_rate": 0.00016171917683482063, "loss": 0.1338, "step": 7530 }, { "epoch": 2.041147807255008, "grad_norm": 0.2749716639518738, "learning_rate": 0.00016126388635949739, "loss": 0.125, "step": 7540 }, { "epoch": 2.0438548998375743, "grad_norm": 0.2698230445384979, "learning_rate": 0.00016080859588417409, "loss": 0.1449, "step": 7550 }, { "epoch": 2.046561992420141, "grad_norm": 0.13137570023536682, "learning_rate": 0.00016035330540885084, "loss": 0.0914, "step": 7560 }, { "epoch": 2.049269085002707, "grad_norm": 0.5193370580673218, "learning_rate": 0.0001598980149335276, "loss": 0.1134, "step": 7570 }, { "epoch": 2.0519761775852734, "grad_norm": 0.19694364070892334, "learning_rate": 0.00015944272445820435, "loss": 0.1058, "step": 7580 }, { "epoch": 2.05468327016784, "grad_norm": 0.35678553581237793, "learning_rate": 0.00015898743398288107, "loss": 0.1209, "step": 7590 }, { "epoch": 2.057390362750406, "grad_norm": 0.26627665758132935, "learning_rate": 0.00015853214350755783, "loss": 0.1166, "step": 7600 }, { "epoch": 2.057390362750406, "eval_loss": 0.15223285555839539, "eval_runtime": 54.6423, "eval_samples_per_second": 9.15, "eval_steps_per_second": 0.293, "step": 7600 }, { "epoch": 2.0600974553329725, "grad_norm": 0.28212130069732666, "learning_rate": 0.00015807685303223458, "loss": 0.1307, "step": 7610 }, { "epoch": 2.0628045479155386, "grad_norm": 0.2795688807964325, "learning_rate": 0.0001576215625569113, "loss": 0.1183, "step": 7620 }, { "epoch": 2.065511640498105, "grad_norm": 0.17099031805992126, "learning_rate": 0.00015716627208158806, "loss": 0.0937, "step": 7630 }, { "epoch": 2.068218733080671, "grad_norm": 0.3722110092639923, "learning_rate": 0.0001567109816062648, "loss": 0.1389, "step": 7640 }, { "epoch": 2.0709258256632377, "grad_norm": 0.3928683400154114, "learning_rate": 0.00015625569113094154, "loss": 0.1033, "step": 7650 }, { "epoch": 2.073632918245804, "grad_norm": 0.4039820432662964, "learning_rate": 0.0001558004006556183, "loss": 0.1348, "step": 7660 }, { "epoch": 2.0763400108283703, "grad_norm": 0.18146918714046478, "learning_rate": 0.00015534511018029505, "loss": 0.1063, "step": 7670 }, { "epoch": 2.079047103410937, "grad_norm": 0.27164018154144287, "learning_rate": 0.00015488981970497177, "loss": 0.1314, "step": 7680 }, { "epoch": 2.081754195993503, "grad_norm": 0.15179571509361267, "learning_rate": 0.0001544345292296485, "loss": 0.1215, "step": 7690 }, { "epoch": 2.0844612885760694, "grad_norm": 0.317647784948349, "learning_rate": 0.00015397923875432525, "loss": 0.1121, "step": 7700 }, { "epoch": 2.0844612885760694, "eval_loss": 0.1496654599905014, "eval_runtime": 51.3749, "eval_samples_per_second": 9.732, "eval_steps_per_second": 0.311, "step": 7700 }, { "epoch": 2.0871683811586355, "grad_norm": 0.20976316928863525, "learning_rate": 0.000153523948279002, "loss": 0.0827, "step": 7710 }, { "epoch": 2.089875473741202, "grad_norm": 0.36998632550239563, "learning_rate": 0.00015306865780367876, "loss": 0.0985, "step": 7720 }, { "epoch": 2.0925825663237685, "grad_norm": 0.3467310070991516, "learning_rate": 0.0001526133673283555, "loss": 0.123, "step": 7730 }, { "epoch": 2.0952896589063346, "grad_norm": 0.24463088810443878, "learning_rate": 0.00015215807685303224, "loss": 0.0928, "step": 7740 }, { "epoch": 2.097996751488901, "grad_norm": 0.19121113419532776, "learning_rate": 0.000151702786377709, "loss": 0.1227, "step": 7750 }, { "epoch": 2.100703844071467, "grad_norm": 0.3805799186229706, "learning_rate": 0.00015124749590238572, "loss": 0.1059, "step": 7760 }, { "epoch": 2.1034109366540337, "grad_norm": 0.21060994267463684, "learning_rate": 0.00015079220542706245, "loss": 0.1047, "step": 7770 }, { "epoch": 2.1061180292365997, "grad_norm": 0.27969229221343994, "learning_rate": 0.0001503369149517392, "loss": 0.1097, "step": 7780 }, { "epoch": 2.1088251218191663, "grad_norm": 0.17218650877475739, "learning_rate": 0.00014988162447641596, "loss": 0.1041, "step": 7790 }, { "epoch": 2.1115322144017323, "grad_norm": 0.16784489154815674, "learning_rate": 0.0001494263340010927, "loss": 0.1231, "step": 7800 }, { "epoch": 2.1115322144017323, "eval_loss": 0.15185900032520294, "eval_runtime": 51.7594, "eval_samples_per_second": 9.66, "eval_steps_per_second": 0.309, "step": 7800 }, { "epoch": 2.114239306984299, "grad_norm": 0.15567447245121002, "learning_rate": 0.00014897104352576946, "loss": 0.1209, "step": 7810 }, { "epoch": 2.1169463995668654, "grad_norm": 0.3253188729286194, "learning_rate": 0.0001485157530504462, "loss": 0.1214, "step": 7820 }, { "epoch": 2.1196534921494314, "grad_norm": 0.20111972093582153, "learning_rate": 0.00014806046257512292, "loss": 0.1132, "step": 7830 }, { "epoch": 2.122360584731998, "grad_norm": 0.26392456889152527, "learning_rate": 0.00014760517209979967, "loss": 0.1035, "step": 7840 }, { "epoch": 2.125067677314564, "grad_norm": 0.21585513651371002, "learning_rate": 0.00014714988162447642, "loss": 0.1051, "step": 7850 }, { "epoch": 2.1277747698971305, "grad_norm": 0.2535959482192993, "learning_rate": 0.00014669459114915315, "loss": 0.1261, "step": 7860 }, { "epoch": 2.1304818624796966, "grad_norm": 0.1298813670873642, "learning_rate": 0.0001462393006738299, "loss": 0.0847, "step": 7870 }, { "epoch": 2.133188955062263, "grad_norm": 0.25601625442504883, "learning_rate": 0.00014578401019850666, "loss": 0.1023, "step": 7880 }, { "epoch": 2.1358960476448297, "grad_norm": 0.11331689357757568, "learning_rate": 0.0001453287197231834, "loss": 0.1191, "step": 7890 }, { "epoch": 2.1386031402273957, "grad_norm": 0.28952756524086, "learning_rate": 0.00014487342924786014, "loss": 0.1102, "step": 7900 }, { "epoch": 2.1386031402273957, "eval_loss": 0.14867374300956726, "eval_runtime": 51.7768, "eval_samples_per_second": 9.657, "eval_steps_per_second": 0.309, "step": 7900 }, { "epoch": 2.1413102328099622, "grad_norm": 0.19193996489048004, "learning_rate": 0.00014441813877253686, "loss": 0.1217, "step": 7910 }, { "epoch": 2.1440173253925283, "grad_norm": 0.2836934030056, "learning_rate": 0.00014396284829721362, "loss": 0.1214, "step": 7920 }, { "epoch": 2.146724417975095, "grad_norm": 0.17073260247707367, "learning_rate": 0.00014350755782189037, "loss": 0.1124, "step": 7930 }, { "epoch": 2.149431510557661, "grad_norm": 0.18840056657791138, "learning_rate": 0.00014305226734656712, "loss": 0.0757, "step": 7940 }, { "epoch": 2.1521386031402274, "grad_norm": 0.2775890529155731, "learning_rate": 0.00014259697687124385, "loss": 0.1388, "step": 7950 }, { "epoch": 2.154845695722794, "grad_norm": 0.44438081979751587, "learning_rate": 0.0001421416863959206, "loss": 0.1055, "step": 7960 }, { "epoch": 2.15755278830536, "grad_norm": 0.29212743043899536, "learning_rate": 0.00014168639592059733, "loss": 0.0971, "step": 7970 }, { "epoch": 2.1602598808879265, "grad_norm": 0.4168376326560974, "learning_rate": 0.00014123110544527408, "loss": 0.109, "step": 7980 }, { "epoch": 2.1629669734704926, "grad_norm": 0.3299972414970398, "learning_rate": 0.00014077581496995084, "loss": 0.1135, "step": 7990 }, { "epoch": 2.165674066053059, "grad_norm": 0.06806978583335876, "learning_rate": 0.00014032052449462756, "loss": 0.1063, "step": 8000 }, { "epoch": 2.165674066053059, "eval_loss": 0.14812108874320984, "eval_runtime": 53.3268, "eval_samples_per_second": 9.376, "eval_steps_per_second": 0.3, "step": 8000 }, { "epoch": 2.168381158635625, "grad_norm": 0.31759846210479736, "learning_rate": 0.00013986523401930432, "loss": 0.094, "step": 8010 }, { "epoch": 2.1710882512181917, "grad_norm": 0.07891532778739929, "learning_rate": 0.00013940994354398107, "loss": 0.1148, "step": 8020 }, { "epoch": 2.173795343800758, "grad_norm": 0.19877280294895172, "learning_rate": 0.00013895465306865783, "loss": 0.1104, "step": 8030 }, { "epoch": 2.1765024363833243, "grad_norm": 0.2574431300163269, "learning_rate": 0.00013849936259333452, "loss": 0.0945, "step": 8040 }, { "epoch": 2.179209528965891, "grad_norm": 0.31091296672821045, "learning_rate": 0.00013804407211801128, "loss": 0.1113, "step": 8050 }, { "epoch": 2.181916621548457, "grad_norm": 0.2589088976383209, "learning_rate": 0.00013758878164268803, "loss": 0.0903, "step": 8060 }, { "epoch": 2.1846237141310234, "grad_norm": 0.1652253419160843, "learning_rate": 0.00013713349116736479, "loss": 0.106, "step": 8070 }, { "epoch": 2.1873308067135895, "grad_norm": 0.11054456979036331, "learning_rate": 0.00013667820069204154, "loss": 0.0956, "step": 8080 }, { "epoch": 2.190037899296156, "grad_norm": 0.13949480652809143, "learning_rate": 0.00013622291021671827, "loss": 0.1021, "step": 8090 }, { "epoch": 2.192744991878722, "grad_norm": 0.1612575799226761, "learning_rate": 0.00013576761974139502, "loss": 0.0975, "step": 8100 }, { "epoch": 2.192744991878722, "eval_loss": 0.15088878571987152, "eval_runtime": 52.3858, "eval_samples_per_second": 9.545, "eval_steps_per_second": 0.305, "step": 8100 }, { "epoch": 2.1954520844612886, "grad_norm": 0.17387332022190094, "learning_rate": 0.00013531232926607177, "loss": 0.1086, "step": 8110 }, { "epoch": 2.198159177043855, "grad_norm": 0.17473745346069336, "learning_rate": 0.0001348570387907485, "loss": 0.1015, "step": 8120 }, { "epoch": 2.200866269626421, "grad_norm": 0.26711127161979675, "learning_rate": 0.00013440174831542523, "loss": 0.0953, "step": 8130 }, { "epoch": 2.2035733622089877, "grad_norm": 0.407000869512558, "learning_rate": 0.00013394645784010198, "loss": 0.1357, "step": 8140 }, { "epoch": 2.2062804547915538, "grad_norm": 0.4457509517669678, "learning_rate": 0.00013349116736477873, "loss": 0.1154, "step": 8150 }, { "epoch": 2.2089875473741203, "grad_norm": 0.38976404070854187, "learning_rate": 0.0001330358768894555, "loss": 0.0974, "step": 8160 }, { "epoch": 2.2116946399566864, "grad_norm": 0.15152068436145782, "learning_rate": 0.00013258058641413224, "loss": 0.1106, "step": 8170 }, { "epoch": 2.214401732539253, "grad_norm": 0.31610003113746643, "learning_rate": 0.00013212529593880897, "loss": 0.1105, "step": 8180 }, { "epoch": 2.2171088251218194, "grad_norm": 0.3985111117362976, "learning_rate": 0.0001316700054634857, "loss": 0.1184, "step": 8190 }, { "epoch": 2.2198159177043855, "grad_norm": 0.28766173124313354, "learning_rate": 0.00013121471498816245, "loss": 0.0931, "step": 8200 }, { "epoch": 2.2198159177043855, "eval_loss": 0.14935404062271118, "eval_runtime": 52.889, "eval_samples_per_second": 9.454, "eval_steps_per_second": 0.303, "step": 8200 }, { "epoch": 2.222523010286952, "grad_norm": 0.3414205312728882, "learning_rate": 0.0001307594245128392, "loss": 0.104, "step": 8210 }, { "epoch": 2.225230102869518, "grad_norm": 0.49229031801223755, "learning_rate": 0.00013030413403751593, "loss": 0.1446, "step": 8220 }, { "epoch": 2.2279371954520846, "grad_norm": 0.3560900092124939, "learning_rate": 0.00012984884356219268, "loss": 0.1134, "step": 8230 }, { "epoch": 2.2306442880346506, "grad_norm": 0.17200881242752075, "learning_rate": 0.00012939355308686943, "loss": 0.1069, "step": 8240 }, { "epoch": 2.233351380617217, "grad_norm": 0.3682667016983032, "learning_rate": 0.0001289382626115462, "loss": 0.1188, "step": 8250 }, { "epoch": 2.2360584731997832, "grad_norm": 0.16727127134799957, "learning_rate": 0.00012848297213622291, "loss": 0.1427, "step": 8260 }, { "epoch": 2.2387655657823498, "grad_norm": 0.14669375121593475, "learning_rate": 0.00012802768166089964, "loss": 0.0982, "step": 8270 }, { "epoch": 2.2414726583649163, "grad_norm": 0.04038509353995323, "learning_rate": 0.0001275723911855764, "loss": 0.0908, "step": 8280 }, { "epoch": 2.2441797509474823, "grad_norm": 0.10155865550041199, "learning_rate": 0.00012711710071025315, "loss": 0.0877, "step": 8290 }, { "epoch": 2.246886843530049, "grad_norm": 0.24377843737602234, "learning_rate": 0.0001266618102349299, "loss": 0.0991, "step": 8300 }, { "epoch": 2.246886843530049, "eval_loss": 0.15127049386501312, "eval_runtime": 52.9728, "eval_samples_per_second": 9.439, "eval_steps_per_second": 0.302, "step": 8300 }, { "epoch": 2.249593936112615, "grad_norm": 0.23823745548725128, "learning_rate": 0.00012620651975960663, "loss": 0.1043, "step": 8310 }, { "epoch": 2.2523010286951815, "grad_norm": 0.227587029337883, "learning_rate": 0.00012575122928428338, "loss": 0.1056, "step": 8320 }, { "epoch": 2.2550081212777475, "grad_norm": 0.14021603763103485, "learning_rate": 0.0001252959388089601, "loss": 0.0994, "step": 8330 }, { "epoch": 2.257715213860314, "grad_norm": 0.23243439197540283, "learning_rate": 0.00012484064833363686, "loss": 0.1024, "step": 8340 }, { "epoch": 2.26042230644288, "grad_norm": 0.1790105700492859, "learning_rate": 0.00012438535785831362, "loss": 0.084, "step": 8350 }, { "epoch": 2.2631293990254466, "grad_norm": 0.2110361009836197, "learning_rate": 0.00012393006738299034, "loss": 0.0895, "step": 8360 }, { "epoch": 2.265836491608013, "grad_norm": 0.24758604168891907, "learning_rate": 0.0001234747769076671, "loss": 0.1154, "step": 8370 }, { "epoch": 2.268543584190579, "grad_norm": 0.16578835248947144, "learning_rate": 0.00012301948643234385, "loss": 0.1193, "step": 8380 }, { "epoch": 2.2712506767731457, "grad_norm": 0.20040668547153473, "learning_rate": 0.00012256419595702058, "loss": 0.1194, "step": 8390 }, { "epoch": 2.273957769355712, "grad_norm": 0.08023863285779953, "learning_rate": 0.00012210890548169733, "loss": 0.1105, "step": 8400 }, { "epoch": 2.273957769355712, "eval_loss": 0.15107358992099762, "eval_runtime": 51.9198, "eval_samples_per_second": 9.63, "eval_steps_per_second": 0.308, "step": 8400 }, { "epoch": 2.2766648619382783, "grad_norm": 0.11227234452962875, "learning_rate": 0.00012165361500637407, "loss": 0.1002, "step": 8410 }, { "epoch": 2.279371954520845, "grad_norm": 0.20993013679981232, "learning_rate": 0.00012119832453105081, "loss": 0.1062, "step": 8420 }, { "epoch": 2.282079047103411, "grad_norm": 0.2064633071422577, "learning_rate": 0.00012074303405572756, "loss": 0.0985, "step": 8430 }, { "epoch": 2.2847861396859774, "grad_norm": 0.19416238367557526, "learning_rate": 0.0001202877435804043, "loss": 0.1329, "step": 8440 }, { "epoch": 2.2874932322685435, "grad_norm": 0.1850809007883072, "learning_rate": 0.00011983245310508104, "loss": 0.0988, "step": 8450 }, { "epoch": 2.29020032485111, "grad_norm": 0.17581935226917267, "learning_rate": 0.00011937716262975778, "loss": 0.1033, "step": 8460 }, { "epoch": 2.292907417433676, "grad_norm": 0.38977089524269104, "learning_rate": 0.00011892187215443454, "loss": 0.1167, "step": 8470 }, { "epoch": 2.2956145100162426, "grad_norm": 0.1589565873146057, "learning_rate": 0.00011846658167911128, "loss": 0.0888, "step": 8480 }, { "epoch": 2.2983216025988087, "grad_norm": 0.2188950628042221, "learning_rate": 0.00011801129120378802, "loss": 0.0936, "step": 8490 }, { "epoch": 2.301028695181375, "grad_norm": 0.15488547086715698, "learning_rate": 0.00011755600072846476, "loss": 0.1217, "step": 8500 }, { "epoch": 2.301028695181375, "eval_loss": 0.15261313319206238, "eval_runtime": 53.1666, "eval_samples_per_second": 9.404, "eval_steps_per_second": 0.301, "step": 8500 }, { "epoch": 2.3037357877639417, "grad_norm": 0.448257178068161, "learning_rate": 0.00011710071025314151, "loss": 0.1098, "step": 8510 }, { "epoch": 2.306442880346508, "grad_norm": 0.2177487164735794, "learning_rate": 0.00011664541977781825, "loss": 0.1096, "step": 8520 }, { "epoch": 2.3091499729290743, "grad_norm": 0.1455925852060318, "learning_rate": 0.00011619012930249499, "loss": 0.1125, "step": 8530 }, { "epoch": 2.3118570655116404, "grad_norm": 0.22162719070911407, "learning_rate": 0.00011573483882717174, "loss": 0.1104, "step": 8540 }, { "epoch": 2.314564158094207, "grad_norm": 0.31285014748573303, "learning_rate": 0.00011527954835184849, "loss": 0.1086, "step": 8550 }, { "epoch": 2.317271250676773, "grad_norm": 0.3933559060096741, "learning_rate": 0.00011482425787652523, "loss": 0.1179, "step": 8560 }, { "epoch": 2.3199783432593395, "grad_norm": 0.2300388514995575, "learning_rate": 0.00011436896740120197, "loss": 0.117, "step": 8570 }, { "epoch": 2.3226854358419056, "grad_norm": 0.23163248598575592, "learning_rate": 0.00011391367692587872, "loss": 0.1183, "step": 8580 }, { "epoch": 2.325392528424472, "grad_norm": 0.5895636677742004, "learning_rate": 0.00011345838645055545, "loss": 0.1439, "step": 8590 }, { "epoch": 2.3280996210070386, "grad_norm": 0.2042539417743683, "learning_rate": 0.0001130030959752322, "loss": 0.0834, "step": 8600 }, { "epoch": 2.3280996210070386, "eval_loss": 0.15408456325531006, "eval_runtime": 51.8771, "eval_samples_per_second": 9.638, "eval_steps_per_second": 0.308, "step": 8600 }, { "epoch": 2.3308067135896047, "grad_norm": 0.2250804603099823, "learning_rate": 0.00011254780549990895, "loss": 0.1125, "step": 8610 }, { "epoch": 2.333513806172171, "grad_norm": 0.19512999057769775, "learning_rate": 0.00011209251502458569, "loss": 0.0954, "step": 8620 }, { "epoch": 2.3362208987547373, "grad_norm": 0.2871667146682739, "learning_rate": 0.00011163722454926243, "loss": 0.1123, "step": 8630 }, { "epoch": 2.338927991337304, "grad_norm": 0.5765364766120911, "learning_rate": 0.00011118193407393917, "loss": 0.0942, "step": 8640 }, { "epoch": 2.3416350839198703, "grad_norm": 0.26905539631843567, "learning_rate": 0.00011072664359861593, "loss": 0.1314, "step": 8650 }, { "epoch": 2.3443421765024364, "grad_norm": 0.40762558579444885, "learning_rate": 0.00011027135312329265, "loss": 0.0969, "step": 8660 }, { "epoch": 2.347049269085003, "grad_norm": 0.2780246436595917, "learning_rate": 0.0001098160626479694, "loss": 0.1091, "step": 8670 }, { "epoch": 2.349756361667569, "grad_norm": 0.2559877634048462, "learning_rate": 0.00010936077217264615, "loss": 0.0946, "step": 8680 }, { "epoch": 2.3524634542501355, "grad_norm": 0.31101638078689575, "learning_rate": 0.0001089054816973229, "loss": 0.1132, "step": 8690 }, { "epoch": 2.3551705468327016, "grad_norm": 0.2569589912891388, "learning_rate": 0.00010845019122199964, "loss": 0.1025, "step": 8700 }, { "epoch": 2.3551705468327016, "eval_loss": 0.15364421904087067, "eval_runtime": 53.3286, "eval_samples_per_second": 9.376, "eval_steps_per_second": 0.3, "step": 8700 }, { "epoch": 2.357877639415268, "grad_norm": 0.11930614709854126, "learning_rate": 0.00010799490074667638, "loss": 0.1002, "step": 8710 }, { "epoch": 2.360584731997834, "grad_norm": 0.1807277500629425, "learning_rate": 0.00010753961027135313, "loss": 0.107, "step": 8720 }, { "epoch": 2.3632918245804007, "grad_norm": 0.12084764242172241, "learning_rate": 0.00010708431979602986, "loss": 0.0863, "step": 8730 }, { "epoch": 2.365998917162967, "grad_norm": 0.5520146489143372, "learning_rate": 0.00010662902932070661, "loss": 0.1199, "step": 8740 }, { "epoch": 2.3687060097455332, "grad_norm": 0.3508378267288208, "learning_rate": 0.00010617373884538335, "loss": 0.1079, "step": 8750 }, { "epoch": 2.3714131023280998, "grad_norm": 0.414885550737381, "learning_rate": 0.00010571844837006011, "loss": 0.1308, "step": 8760 }, { "epoch": 2.374120194910666, "grad_norm": 0.2651754319667816, "learning_rate": 0.00010526315789473683, "loss": 0.109, "step": 8770 }, { "epoch": 2.3768272874932324, "grad_norm": 0.23027944564819336, "learning_rate": 0.00010480786741941359, "loss": 0.0785, "step": 8780 }, { "epoch": 2.3795343800757984, "grad_norm": 0.2035895586013794, "learning_rate": 0.00010435257694409034, "loss": 0.1022, "step": 8790 }, { "epoch": 2.382241472658365, "grad_norm": 0.3900894522666931, "learning_rate": 0.00010389728646876707, "loss": 0.1025, "step": 8800 }, { "epoch": 2.382241472658365, "eval_loss": 0.15135331451892853, "eval_runtime": 52.1508, "eval_samples_per_second": 9.588, "eval_steps_per_second": 0.307, "step": 8800 }, { "epoch": 2.384948565240931, "grad_norm": 0.25562533736228943, "learning_rate": 0.00010344199599344382, "loss": 0.1165, "step": 8810 }, { "epoch": 2.3876556578234975, "grad_norm": 0.3861636519432068, "learning_rate": 0.00010298670551812056, "loss": 0.1022, "step": 8820 }, { "epoch": 2.390362750406064, "grad_norm": 0.11617624759674072, "learning_rate": 0.00010253141504279732, "loss": 0.0868, "step": 8830 }, { "epoch": 2.39306984298863, "grad_norm": 0.21198932826519012, "learning_rate": 0.00010207612456747404, "loss": 0.1252, "step": 8840 }, { "epoch": 2.3957769355711966, "grad_norm": 0.17055420577526093, "learning_rate": 0.0001016208340921508, "loss": 0.1306, "step": 8850 }, { "epoch": 2.3984840281537627, "grad_norm": 0.11652148514986038, "learning_rate": 0.00010116554361682754, "loss": 0.1009, "step": 8860 }, { "epoch": 2.4011911207363292, "grad_norm": 0.13691246509552002, "learning_rate": 0.00010071025314150428, "loss": 0.1085, "step": 8870 }, { "epoch": 2.4038982133188957, "grad_norm": 0.07319015264511108, "learning_rate": 0.00010025496266618103, "loss": 0.0754, "step": 8880 }, { "epoch": 2.406605305901462, "grad_norm": 0.3480643630027771, "learning_rate": 9.979967219085777e-05, "loss": 0.1308, "step": 8890 }, { "epoch": 2.4093123984840283, "grad_norm": 0.15720431506633759, "learning_rate": 9.934438171553452e-05, "loss": 0.1385, "step": 8900 }, { "epoch": 2.4093123984840283, "eval_loss": 0.14918014407157898, "eval_runtime": 53.1709, "eval_samples_per_second": 9.404, "eval_steps_per_second": 0.301, "step": 8900 }, { "epoch": 2.4120194910665944, "grad_norm": 0.0895061045885086, "learning_rate": 9.888909124021125e-05, "loss": 0.1012, "step": 8910 }, { "epoch": 2.414726583649161, "grad_norm": 0.3535843789577484, "learning_rate": 9.8433800764888e-05, "loss": 0.1049, "step": 8920 }, { "epoch": 2.417433676231727, "grad_norm": 0.303932785987854, "learning_rate": 9.797851028956474e-05, "loss": 0.0859, "step": 8930 }, { "epoch": 2.4201407688142935, "grad_norm": 0.3384048640727997, "learning_rate": 9.75232198142415e-05, "loss": 0.1232, "step": 8940 }, { "epoch": 2.4228478613968596, "grad_norm": 0.44046324491500854, "learning_rate": 9.706792933891822e-05, "loss": 0.1284, "step": 8950 }, { "epoch": 2.425554953979426, "grad_norm": 0.13039818406105042, "learning_rate": 9.661263886359498e-05, "loss": 0.092, "step": 8960 }, { "epoch": 2.4282620465619926, "grad_norm": 0.07619010657072067, "learning_rate": 9.615734838827173e-05, "loss": 0.0939, "step": 8970 }, { "epoch": 2.4309691391445587, "grad_norm": 0.23351195454597473, "learning_rate": 9.570205791294846e-05, "loss": 0.1092, "step": 8980 }, { "epoch": 2.433676231727125, "grad_norm": 0.14060206711292267, "learning_rate": 9.524676743762521e-05, "loss": 0.0893, "step": 8990 }, { "epoch": 2.4363833243096913, "grad_norm": 0.18231533467769623, "learning_rate": 9.479147696230195e-05, "loss": 0.1229, "step": 9000 }, { "epoch": 2.4363833243096913, "eval_loss": 0.14696462452411652, "eval_runtime": 52.2474, "eval_samples_per_second": 9.57, "eval_steps_per_second": 0.306, "step": 9000 }, { "epoch": 2.439090416892258, "grad_norm": 0.26201844215393066, "learning_rate": 9.43361864869787e-05, "loss": 0.1107, "step": 9010 }, { "epoch": 2.441797509474824, "grad_norm": 0.16013525426387787, "learning_rate": 9.388089601165543e-05, "loss": 0.1007, "step": 9020 }, { "epoch": 2.4445046020573904, "grad_norm": 0.3495844006538391, "learning_rate": 9.342560553633218e-05, "loss": 0.1171, "step": 9030 }, { "epoch": 2.4472116946399565, "grad_norm": 0.323893666267395, "learning_rate": 9.297031506100892e-05, "loss": 0.1002, "step": 9040 }, { "epoch": 2.449918787222523, "grad_norm": 0.27720922231674194, "learning_rate": 9.251502458568566e-05, "loss": 0.1335, "step": 9050 }, { "epoch": 2.4526258798050895, "grad_norm": 0.20315824449062347, "learning_rate": 9.205973411036242e-05, "loss": 0.0839, "step": 9060 }, { "epoch": 2.4553329723876556, "grad_norm": 0.18524716794490814, "learning_rate": 9.160444363503916e-05, "loss": 0.1173, "step": 9070 }, { "epoch": 2.458040064970222, "grad_norm": 0.5791286826133728, "learning_rate": 9.114915315971591e-05, "loss": 0.1349, "step": 9080 }, { "epoch": 2.460747157552788, "grad_norm": 0.0970366969704628, "learning_rate": 9.069386268439264e-05, "loss": 0.1141, "step": 9090 }, { "epoch": 2.4634542501353547, "grad_norm": 0.425823837518692, "learning_rate": 9.023857220906939e-05, "loss": 0.1014, "step": 9100 }, { "epoch": 2.4634542501353547, "eval_loss": 0.15250267088413239, "eval_runtime": 52.0809, "eval_samples_per_second": 9.6, "eval_steps_per_second": 0.307, "step": 9100 }, { "epoch": 2.466161342717921, "grad_norm": 0.2796113193035126, "learning_rate": 8.978328173374613e-05, "loss": 0.1077, "step": 9110 }, { "epoch": 2.4688684353004873, "grad_norm": 0.32091084122657776, "learning_rate": 8.932799125842287e-05, "loss": 0.0973, "step": 9120 }, { "epoch": 2.471575527883054, "grad_norm": 0.17554207146167755, "learning_rate": 8.887270078309961e-05, "loss": 0.1089, "step": 9130 }, { "epoch": 2.47428262046562, "grad_norm": 0.2353290617465973, "learning_rate": 8.841741030777637e-05, "loss": 0.1037, "step": 9140 }, { "epoch": 2.4769897130481864, "grad_norm": 0.2735700309276581, "learning_rate": 8.796211983245312e-05, "loss": 0.1067, "step": 9150 }, { "epoch": 2.4796968056307525, "grad_norm": 0.07648351788520813, "learning_rate": 8.750682935712985e-05, "loss": 0.1258, "step": 9160 }, { "epoch": 2.482403898213319, "grad_norm": 0.2713570296764374, "learning_rate": 8.70515388818066e-05, "loss": 0.1134, "step": 9170 }, { "epoch": 2.485110990795885, "grad_norm": 0.2194133698940277, "learning_rate": 8.659624840648334e-05, "loss": 0.1103, "step": 9180 }, { "epoch": 2.4878180833784516, "grad_norm": 0.34865009784698486, "learning_rate": 8.614095793116008e-05, "loss": 0.0919, "step": 9190 }, { "epoch": 2.490525175961018, "grad_norm": 0.22364215552806854, "learning_rate": 8.568566745583682e-05, "loss": 0.1001, "step": 9200 }, { "epoch": 2.490525175961018, "eval_loss": 0.15467584133148193, "eval_runtime": 53.1037, "eval_samples_per_second": 9.416, "eval_steps_per_second": 0.301, "step": 9200 }, { "epoch": 2.493232268543584, "grad_norm": 0.3470965623855591, "learning_rate": 8.523037698051357e-05, "loss": 0.1028, "step": 9210 }, { "epoch": 2.4959393611261507, "grad_norm": 0.23215217888355255, "learning_rate": 8.477508650519031e-05, "loss": 0.0985, "step": 9220 }, { "epoch": 2.4986464537087167, "grad_norm": 0.16627094149589539, "learning_rate": 8.431979602986705e-05, "loss": 0.1572, "step": 9230 }, { "epoch": 2.5013535462912833, "grad_norm": 0.34447458386421204, "learning_rate": 8.386450555454381e-05, "loss": 0.1048, "step": 9240 }, { "epoch": 2.5040606388738493, "grad_norm": 0.5015201568603516, "learning_rate": 8.340921507922055e-05, "loss": 0.1147, "step": 9250 }, { "epoch": 2.506767731456416, "grad_norm": 0.5202590227127075, "learning_rate": 8.295392460389729e-05, "loss": 0.1112, "step": 9260 }, { "epoch": 2.509474824038982, "grad_norm": 0.22088183462619781, "learning_rate": 8.249863412857403e-05, "loss": 0.1044, "step": 9270 }, { "epoch": 2.5121819166215484, "grad_norm": 0.269815593957901, "learning_rate": 8.204334365325078e-05, "loss": 0.0893, "step": 9280 }, { "epoch": 2.514889009204115, "grad_norm": 0.3005484938621521, "learning_rate": 8.158805317792752e-05, "loss": 0.0742, "step": 9290 }, { "epoch": 2.517596101786681, "grad_norm": 0.2150215059518814, "learning_rate": 8.113276270260426e-05, "loss": 0.1216, "step": 9300 }, { "epoch": 2.517596101786681, "eval_loss": 0.15590840578079224, "eval_runtime": 53.3113, "eval_samples_per_second": 9.379, "eval_steps_per_second": 0.3, "step": 9300 }, { "epoch": 2.5203031943692475, "grad_norm": 0.3834645748138428, "learning_rate": 8.0677472227281e-05, "loss": 0.1044, "step": 9310 }, { "epoch": 2.5230102869518136, "grad_norm": 0.22386111319065094, "learning_rate": 8.022218175195775e-05, "loss": 0.1021, "step": 9320 }, { "epoch": 2.52571737953438, "grad_norm": 0.3309231996536255, "learning_rate": 7.97668912766345e-05, "loss": 0.1251, "step": 9330 }, { "epoch": 2.5284244721169467, "grad_norm": 0.6254246830940247, "learning_rate": 7.931160080131124e-05, "loss": 0.1014, "step": 9340 }, { "epoch": 2.5311315646995127, "grad_norm": 0.08925973623991013, "learning_rate": 7.885631032598799e-05, "loss": 0.0984, "step": 9350 }, { "epoch": 2.533838657282079, "grad_norm": 0.2096097618341446, "learning_rate": 7.840101985066473e-05, "loss": 0.1046, "step": 9360 }, { "epoch": 2.5365457498646453, "grad_norm": 0.3621312379837036, "learning_rate": 7.794572937534147e-05, "loss": 0.0862, "step": 9370 }, { "epoch": 2.539252842447212, "grad_norm": 0.40978795289993286, "learning_rate": 7.749043890001821e-05, "loss": 0.0962, "step": 9380 }, { "epoch": 2.541959935029778, "grad_norm": 0.40507039427757263, "learning_rate": 7.703514842469496e-05, "loss": 0.115, "step": 9390 }, { "epoch": 2.5446670276123444, "grad_norm": 0.09742297232151031, "learning_rate": 7.657985794937169e-05, "loss": 0.0933, "step": 9400 }, { "epoch": 2.5446670276123444, "eval_loss": 0.1495400369167328, "eval_runtime": 52.305, "eval_samples_per_second": 9.559, "eval_steps_per_second": 0.306, "step": 9400 }, { "epoch": 2.5473741201949105, "grad_norm": 0.24087636172771454, "learning_rate": 7.612456747404844e-05, "loss": 0.1054, "step": 9410 }, { "epoch": 2.550081212777477, "grad_norm": 0.25820058584213257, "learning_rate": 7.56692769987252e-05, "loss": 0.0947, "step": 9420 }, { "epoch": 2.5527883053600435, "grad_norm": 0.1936085969209671, "learning_rate": 7.521398652340194e-05, "loss": 0.0821, "step": 9430 }, { "epoch": 2.5554953979426096, "grad_norm": 0.3259497284889221, "learning_rate": 7.475869604807868e-05, "loss": 0.1147, "step": 9440 }, { "epoch": 2.558202490525176, "grad_norm": 0.22895482182502747, "learning_rate": 7.430340557275542e-05, "loss": 0.0919, "step": 9450 }, { "epoch": 2.560909583107742, "grad_norm": 0.17221614718437195, "learning_rate": 7.384811509743217e-05, "loss": 0.1029, "step": 9460 }, { "epoch": 2.5636166756903087, "grad_norm": 0.46346572041511536, "learning_rate": 7.33928246221089e-05, "loss": 0.1073, "step": 9470 }, { "epoch": 2.566323768272875, "grad_norm": 0.5658268332481384, "learning_rate": 7.293753414678565e-05, "loss": 0.119, "step": 9480 }, { "epoch": 2.5690308608554413, "grad_norm": 0.25370046496391296, "learning_rate": 7.248224367146239e-05, "loss": 0.1091, "step": 9490 }, { "epoch": 2.5717379534380074, "grad_norm": 0.26131343841552734, "learning_rate": 7.202695319613914e-05, "loss": 0.0881, "step": 9500 }, { "epoch": 2.5717379534380074, "eval_loss": 0.15294285118579865, "eval_runtime": 51.764, "eval_samples_per_second": 9.659, "eval_steps_per_second": 0.309, "step": 9500 }, { "epoch": 2.574445046020574, "grad_norm": 0.17354480922222137, "learning_rate": 7.157166272081588e-05, "loss": 0.1093, "step": 9510 }, { "epoch": 2.5771521386031404, "grad_norm": 0.3243185877799988, "learning_rate": 7.111637224549262e-05, "loss": 0.0885, "step": 9520 }, { "epoch": 2.5798592311857065, "grad_norm": 0.3836454451084137, "learning_rate": 7.066108177016938e-05, "loss": 0.1073, "step": 9530 }, { "epoch": 2.582566323768273, "grad_norm": 0.3406403660774231, "learning_rate": 7.02057912948461e-05, "loss": 0.1115, "step": 9540 }, { "epoch": 2.585273416350839, "grad_norm": 0.2781931161880493, "learning_rate": 6.975050081952286e-05, "loss": 0.1074, "step": 9550 }, { "epoch": 2.5879805089334056, "grad_norm": 0.16563768684864044, "learning_rate": 6.92952103441996e-05, "loss": 0.1395, "step": 9560 }, { "epoch": 2.590687601515972, "grad_norm": 0.12540295720100403, "learning_rate": 6.883991986887635e-05, "loss": 0.1409, "step": 9570 }, { "epoch": 2.593394694098538, "grad_norm": 0.22774340212345123, "learning_rate": 6.838462939355308e-05, "loss": 0.1045, "step": 9580 }, { "epoch": 2.5961017866811043, "grad_norm": 0.6437894701957703, "learning_rate": 6.792933891822983e-05, "loss": 0.1695, "step": 9590 }, { "epoch": 2.5988088792636708, "grad_norm": 0.08183804899454117, "learning_rate": 6.747404844290659e-05, "loss": 0.1047, "step": 9600 }, { "epoch": 2.5988088792636708, "eval_loss": 0.15261416137218475, "eval_runtime": 51.5572, "eval_samples_per_second": 9.698, "eval_steps_per_second": 0.31, "step": 9600 }, { "epoch": 2.6015159718462373, "grad_norm": 0.34598779678344727, "learning_rate": 6.701875796758333e-05, "loss": 0.1092, "step": 9610 }, { "epoch": 2.6042230644288034, "grad_norm": 0.36952805519104004, "learning_rate": 6.656346749226007e-05, "loss": 0.0929, "step": 9620 }, { "epoch": 2.60693015701137, "grad_norm": 0.09631340950727463, "learning_rate": 6.61081770169368e-05, "loss": 0.0918, "step": 9630 }, { "epoch": 2.609637249593936, "grad_norm": 0.3857302963733673, "learning_rate": 6.565288654161356e-05, "loss": 0.1054, "step": 9640 }, { "epoch": 2.6123443421765025, "grad_norm": 0.38102293014526367, "learning_rate": 6.519759606629029e-05, "loss": 0.117, "step": 9650 }, { "epoch": 2.615051434759069, "grad_norm": 0.18504104018211365, "learning_rate": 6.474230559096704e-05, "loss": 0.0875, "step": 9660 }, { "epoch": 2.617758527341635, "grad_norm": 0.12376754730939865, "learning_rate": 6.428701511564378e-05, "loss": 0.1076, "step": 9670 }, { "epoch": 2.6204656199242016, "grad_norm": 0.3745812475681305, "learning_rate": 6.383172464032053e-05, "loss": 0.0932, "step": 9680 }, { "epoch": 2.6231727125067676, "grad_norm": 0.17206531763076782, "learning_rate": 6.337643416499727e-05, "loss": 0.0923, "step": 9690 }, { "epoch": 2.625879805089334, "grad_norm": 0.4507389962673187, "learning_rate": 6.292114368967401e-05, "loss": 0.1159, "step": 9700 }, { "epoch": 2.625879805089334, "eval_loss": 0.15273411571979523, "eval_runtime": 51.4115, "eval_samples_per_second": 9.725, "eval_steps_per_second": 0.311, "step": 9700 }, { "epoch": 2.6285868976719002, "grad_norm": 0.22679084539413452, "learning_rate": 6.246585321435075e-05, "loss": 0.1191, "step": 9710 }, { "epoch": 2.6312939902544668, "grad_norm": 0.15608219802379608, "learning_rate": 6.201056273902749e-05, "loss": 0.113, "step": 9720 }, { "epoch": 2.634001082837033, "grad_norm": 0.20562225580215454, "learning_rate": 6.155527226370425e-05, "loss": 0.1299, "step": 9730 }, { "epoch": 2.6367081754195993, "grad_norm": 0.5659207701683044, "learning_rate": 6.109998178838099e-05, "loss": 0.1035, "step": 9740 }, { "epoch": 2.639415268002166, "grad_norm": 0.17303211987018585, "learning_rate": 6.0644691313057734e-05, "loss": 0.1127, "step": 9750 }, { "epoch": 2.642122360584732, "grad_norm": 0.17225438356399536, "learning_rate": 6.018940083773448e-05, "loss": 0.0961, "step": 9760 }, { "epoch": 2.6448294531672984, "grad_norm": 0.2602575123310089, "learning_rate": 5.973411036241122e-05, "loss": 0.0971, "step": 9770 }, { "epoch": 2.6475365457498645, "grad_norm": 0.17937231063842773, "learning_rate": 5.927881988708796e-05, "loss": 0.09, "step": 9780 }, { "epoch": 2.650243638332431, "grad_norm": 0.35192254185676575, "learning_rate": 5.882352941176471e-05, "loss": 0.1198, "step": 9790 }, { "epoch": 2.6529507309149976, "grad_norm": 0.1931726187467575, "learning_rate": 5.836823893644145e-05, "loss": 0.0925, "step": 9800 }, { "epoch": 2.6529507309149976, "eval_loss": 0.1525072157382965, "eval_runtime": 54.8545, "eval_samples_per_second": 9.115, "eval_steps_per_second": 0.292, "step": 9800 }, { "epoch": 2.6556578234975636, "grad_norm": 0.24726568162441254, "learning_rate": 5.7912948461118194e-05, "loss": 0.1037, "step": 9810 }, { "epoch": 2.6583649160801297, "grad_norm": 0.1900894194841385, "learning_rate": 5.7457657985794934e-05, "loss": 0.0859, "step": 9820 }, { "epoch": 2.661072008662696, "grad_norm": 0.15850365161895752, "learning_rate": 5.700236751047169e-05, "loss": 0.104, "step": 9830 }, { "epoch": 2.6637791012452627, "grad_norm": 0.3096344769001007, "learning_rate": 5.654707703514843e-05, "loss": 0.1046, "step": 9840 }, { "epoch": 2.666486193827829, "grad_norm": 0.16559633612632751, "learning_rate": 5.609178655982517e-05, "loss": 0.103, "step": 9850 }, { "epoch": 2.6691932864103953, "grad_norm": 0.19194342195987701, "learning_rate": 5.5636496084501915e-05, "loss": 0.0991, "step": 9860 }, { "epoch": 2.6719003789929614, "grad_norm": 0.18227119743824005, "learning_rate": 5.5181205609178655e-05, "loss": 0.0961, "step": 9870 }, { "epoch": 2.674607471575528, "grad_norm": 0.2713763117790222, "learning_rate": 5.47259151338554e-05, "loss": 0.1147, "step": 9880 }, { "epoch": 2.6773145641580944, "grad_norm": 0.36927032470703125, "learning_rate": 5.427062465853214e-05, "loss": 0.1079, "step": 9890 }, { "epoch": 2.6800216567406605, "grad_norm": 0.13315054774284363, "learning_rate": 5.381533418320889e-05, "loss": 0.1017, "step": 9900 }, { "epoch": 2.6800216567406605, "eval_loss": 0.15056326985359192, "eval_runtime": 51.794, "eval_samples_per_second": 9.654, "eval_steps_per_second": 0.309, "step": 9900 }, { "epoch": 2.6827287493232266, "grad_norm": 0.24048204720020294, "learning_rate": 5.336004370788563e-05, "loss": 0.1147, "step": 9910 }, { "epoch": 2.685435841905793, "grad_norm": 0.20163370668888092, "learning_rate": 5.2904753232562376e-05, "loss": 0.0801, "step": 9920 }, { "epoch": 2.6881429344883596, "grad_norm": 0.26496630907058716, "learning_rate": 5.244946275723912e-05, "loss": 0.0981, "step": 9930 }, { "epoch": 2.6908500270709257, "grad_norm": 0.176680788397789, "learning_rate": 5.199417228191586e-05, "loss": 0.0975, "step": 9940 }, { "epoch": 2.693557119653492, "grad_norm": 0.2778788208961487, "learning_rate": 5.153888180659261e-05, "loss": 0.0963, "step": 9950 }, { "epoch": 2.6962642122360583, "grad_norm": 0.31696370244026184, "learning_rate": 5.108359133126935e-05, "loss": 0.0943, "step": 9960 }, { "epoch": 2.698971304818625, "grad_norm": 0.16855411231517792, "learning_rate": 5.0628300855946096e-05, "loss": 0.1076, "step": 9970 }, { "epoch": 2.7016783974011913, "grad_norm": 0.1627056747674942, "learning_rate": 5.0173010380622837e-05, "loss": 0.1268, "step": 9980 }, { "epoch": 2.7043854899837574, "grad_norm": 0.4440973103046417, "learning_rate": 4.971771990529958e-05, "loss": 0.1233, "step": 9990 }, { "epoch": 2.707092582566324, "grad_norm": 0.11012829840183258, "learning_rate": 4.9262429429976323e-05, "loss": 0.1123, "step": 10000 }, { "epoch": 2.707092582566324, "eval_loss": 0.14893101155757904, "eval_runtime": 53.2999, "eval_samples_per_second": 9.381, "eval_steps_per_second": 0.3, "step": 10000 }, { "epoch": 2.70979967514889, "grad_norm": 0.2912381589412689, "learning_rate": 4.880713895465307e-05, "loss": 0.1033, "step": 10010 }, { "epoch": 2.7125067677314565, "grad_norm": 0.15744605660438538, "learning_rate": 4.835184847932982e-05, "loss": 0.1073, "step": 10020 }, { "epoch": 2.715213860314023, "grad_norm": 0.1352146863937378, "learning_rate": 4.789655800400656e-05, "loss": 0.0889, "step": 10030 }, { "epoch": 2.717920952896589, "grad_norm": 0.30048781633377075, "learning_rate": 4.7441267528683304e-05, "loss": 0.1009, "step": 10040 }, { "epoch": 2.720628045479155, "grad_norm": 0.21028059720993042, "learning_rate": 4.6985977053360044e-05, "loss": 0.1017, "step": 10050 }, { "epoch": 2.7233351380617217, "grad_norm": 0.3342764973640442, "learning_rate": 4.653068657803679e-05, "loss": 0.1127, "step": 10060 }, { "epoch": 2.726042230644288, "grad_norm": 0.23139092326164246, "learning_rate": 4.607539610271353e-05, "loss": 0.1009, "step": 10070 }, { "epoch": 2.7287493232268543, "grad_norm": 0.18197013437747955, "learning_rate": 4.562010562739027e-05, "loss": 0.0933, "step": 10080 }, { "epoch": 2.731456415809421, "grad_norm": 0.1409190595149994, "learning_rate": 4.516481515206702e-05, "loss": 0.0856, "step": 10090 }, { "epoch": 2.734163508391987, "grad_norm": 0.1978885680437088, "learning_rate": 4.4709524676743765e-05, "loss": 0.0888, "step": 10100 }, { "epoch": 2.734163508391987, "eval_loss": 0.15170806646347046, "eval_runtime": 53.4463, "eval_samples_per_second": 9.355, "eval_steps_per_second": 0.299, "step": 10100 }, { "epoch": 2.7368706009745534, "grad_norm": 0.18989257514476776, "learning_rate": 4.425423420142051e-05, "loss": 0.106, "step": 10110 }, { "epoch": 2.73957769355712, "grad_norm": 0.15426640212535858, "learning_rate": 4.379894372609725e-05, "loss": 0.0973, "step": 10120 }, { "epoch": 2.742284786139686, "grad_norm": 0.17095568776130676, "learning_rate": 4.3343653250774e-05, "loss": 0.0904, "step": 10130 }, { "epoch": 2.744991878722252, "grad_norm": 0.19205018877983093, "learning_rate": 4.288836277545074e-05, "loss": 0.0936, "step": 10140 }, { "epoch": 2.7476989713048185, "grad_norm": 0.20674671232700348, "learning_rate": 4.243307230012748e-05, "loss": 0.1101, "step": 10150 }, { "epoch": 2.750406063887385, "grad_norm": 0.1812015026807785, "learning_rate": 4.1977781824804226e-05, "loss": 0.1288, "step": 10160 }, { "epoch": 2.753113156469951, "grad_norm": 0.28594598174095154, "learning_rate": 4.1522491349480966e-05, "loss": 0.0758, "step": 10170 }, { "epoch": 2.7558202490525177, "grad_norm": 0.4622754752635956, "learning_rate": 4.106720087415771e-05, "loss": 0.1008, "step": 10180 }, { "epoch": 2.7585273416350837, "grad_norm": 0.32906344532966614, "learning_rate": 4.061191039883446e-05, "loss": 0.1031, "step": 10190 }, { "epoch": 2.7612344342176502, "grad_norm": 0.131916344165802, "learning_rate": 4.0156619923511206e-05, "loss": 0.0938, "step": 10200 }, { "epoch": 2.7612344342176502, "eval_loss": 0.14903075993061066, "eval_runtime": 53.7121, "eval_samples_per_second": 9.309, "eval_steps_per_second": 0.298, "step": 10200 }, { "epoch": 2.7639415268002168, "grad_norm": 0.14141401648521423, "learning_rate": 3.9701329448187946e-05, "loss": 0.1089, "step": 10210 }, { "epoch": 2.766648619382783, "grad_norm": 0.20000173151493073, "learning_rate": 3.9246038972864686e-05, "loss": 0.0864, "step": 10220 }, { "epoch": 2.7693557119653494, "grad_norm": 0.5456509590148926, "learning_rate": 3.879074849754143e-05, "loss": 0.0859, "step": 10230 }, { "epoch": 2.7720628045479154, "grad_norm": 0.27678969502449036, "learning_rate": 3.833545802221817e-05, "loss": 0.1197, "step": 10240 }, { "epoch": 2.774769897130482, "grad_norm": 0.253469318151474, "learning_rate": 3.788016754689492e-05, "loss": 0.0762, "step": 10250 }, { "epoch": 2.7774769897130485, "grad_norm": 0.1845778226852417, "learning_rate": 3.742487707157166e-05, "loss": 0.1043, "step": 10260 }, { "epoch": 2.7801840822956145, "grad_norm": 0.27909618616104126, "learning_rate": 3.696958659624841e-05, "loss": 0.1042, "step": 10270 }, { "epoch": 2.7828911748781806, "grad_norm": 0.39091622829437256, "learning_rate": 3.6514296120925154e-05, "loss": 0.1029, "step": 10280 }, { "epoch": 2.785598267460747, "grad_norm": 0.38871899247169495, "learning_rate": 3.6059005645601894e-05, "loss": 0.125, "step": 10290 }, { "epoch": 2.7883053600433136, "grad_norm": 0.1890731155872345, "learning_rate": 3.560371517027864e-05, "loss": 0.1166, "step": 10300 }, { "epoch": 2.7883053600433136, "eval_loss": 0.14813275635242462, "eval_runtime": 53.2518, "eval_samples_per_second": 9.389, "eval_steps_per_second": 0.3, "step": 10300 }, { "epoch": 2.7910124526258797, "grad_norm": 0.24631626904010773, "learning_rate": 3.514842469495538e-05, "loss": 0.1058, "step": 10310 }, { "epoch": 2.7937195452084462, "grad_norm": 0.23233796656131744, "learning_rate": 3.469313421963213e-05, "loss": 0.0929, "step": 10320 }, { "epoch": 2.7964266377910123, "grad_norm": 0.17788082361221313, "learning_rate": 3.423784374430887e-05, "loss": 0.1115, "step": 10330 }, { "epoch": 2.799133730373579, "grad_norm": 0.10910706222057343, "learning_rate": 3.3782553268985615e-05, "loss": 0.122, "step": 10340 }, { "epoch": 2.8018408229561453, "grad_norm": 0.37555086612701416, "learning_rate": 3.3327262793662355e-05, "loss": 0.1072, "step": 10350 }, { "epoch": 2.8045479155387114, "grad_norm": 0.057445164769887924, "learning_rate": 3.28719723183391e-05, "loss": 0.0953, "step": 10360 }, { "epoch": 2.8072550081212775, "grad_norm": 0.16641780734062195, "learning_rate": 3.241668184301585e-05, "loss": 0.1032, "step": 10370 }, { "epoch": 2.809962100703844, "grad_norm": 0.29210391640663147, "learning_rate": 3.196139136769259e-05, "loss": 0.0918, "step": 10380 }, { "epoch": 2.8126691932864105, "grad_norm": 0.23352032899856567, "learning_rate": 3.1506100892369335e-05, "loss": 0.0713, "step": 10390 }, { "epoch": 2.8153762858689766, "grad_norm": 0.1475362628698349, "learning_rate": 3.1050810417046075e-05, "loss": 0.0904, "step": 10400 }, { "epoch": 2.8153762858689766, "eval_loss": 0.1505148857831955, "eval_runtime": 52.203, "eval_samples_per_second": 9.578, "eval_steps_per_second": 0.306, "step": 10400 }, { "epoch": 2.818083378451543, "grad_norm": 0.4008180797100067, "learning_rate": 3.059551994172282e-05, "loss": 0.1166, "step": 10410 }, { "epoch": 2.820790471034109, "grad_norm": 0.11473873257637024, "learning_rate": 3.0140229466399562e-05, "loss": 0.0805, "step": 10420 }, { "epoch": 2.8234975636166757, "grad_norm": 0.24119402468204498, "learning_rate": 2.9684938991076306e-05, "loss": 0.1163, "step": 10430 }, { "epoch": 2.826204656199242, "grad_norm": 0.37018585205078125, "learning_rate": 2.9229648515753052e-05, "loss": 0.1032, "step": 10440 }, { "epoch": 2.8289117487818083, "grad_norm": 0.29082244634628296, "learning_rate": 2.8774358040429796e-05, "loss": 0.0882, "step": 10450 }, { "epoch": 2.831618841364375, "grad_norm": 0.18865616619586945, "learning_rate": 2.831906756510654e-05, "loss": 0.1011, "step": 10460 }, { "epoch": 2.834325933946941, "grad_norm": 0.1716868132352829, "learning_rate": 2.7863777089783283e-05, "loss": 0.0936, "step": 10470 }, { "epoch": 2.8370330265295074, "grad_norm": 0.15654022991657257, "learning_rate": 2.7408486614460023e-05, "loss": 0.0926, "step": 10480 }, { "epoch": 2.8397401191120735, "grad_norm": 0.31513598561286926, "learning_rate": 2.695319613913677e-05, "loss": 0.1046, "step": 10490 }, { "epoch": 2.84244721169464, "grad_norm": 0.32550135254859924, "learning_rate": 2.6497905663813513e-05, "loss": 0.1064, "step": 10500 }, { "epoch": 2.84244721169464, "eval_loss": 0.14921054244041443, "eval_runtime": 53.2325, "eval_samples_per_second": 9.393, "eval_steps_per_second": 0.301, "step": 10500 }, { "epoch": 2.845154304277206, "grad_norm": 0.1897037923336029, "learning_rate": 2.6042615188490257e-05, "loss": 0.1141, "step": 10510 }, { "epoch": 2.8478613968597726, "grad_norm": 0.21038039028644562, "learning_rate": 2.5587324713167e-05, "loss": 0.1005, "step": 10520 }, { "epoch": 2.850568489442339, "grad_norm": 0.32643935084342957, "learning_rate": 2.5132034237843747e-05, "loss": 0.0984, "step": 10530 }, { "epoch": 2.853275582024905, "grad_norm": 0.1864829808473587, "learning_rate": 2.467674376252049e-05, "loss": 0.0901, "step": 10540 }, { "epoch": 2.8559826746074717, "grad_norm": 0.2076212465763092, "learning_rate": 2.4221453287197234e-05, "loss": 0.0794, "step": 10550 }, { "epoch": 2.8586897671900378, "grad_norm": 0.26182305812835693, "learning_rate": 2.3766162811873974e-05, "loss": 0.0986, "step": 10560 }, { "epoch": 2.8613968597726043, "grad_norm": 0.42842212319374084, "learning_rate": 2.3310872336550717e-05, "loss": 0.0984, "step": 10570 }, { "epoch": 2.864103952355171, "grad_norm": 0.3149246871471405, "learning_rate": 2.2855581861227464e-05, "loss": 0.1205, "step": 10580 }, { "epoch": 2.866811044937737, "grad_norm": 0.3597836494445801, "learning_rate": 2.2400291385904208e-05, "loss": 0.1264, "step": 10590 }, { "epoch": 2.869518137520303, "grad_norm": 0.2727559208869934, "learning_rate": 2.194500091058095e-05, "loss": 0.1158, "step": 10600 }, { "epoch": 2.869518137520303, "eval_loss": 0.1474287360906601, "eval_runtime": 52.781, "eval_samples_per_second": 9.473, "eval_steps_per_second": 0.303, "step": 10600 }, { "epoch": 2.8722252301028695, "grad_norm": 0.22489579021930695, "learning_rate": 2.1489710435257695e-05, "loss": 0.0899, "step": 10610 }, { "epoch": 2.874932322685436, "grad_norm": 0.21127289533615112, "learning_rate": 2.103441995993444e-05, "loss": 0.1077, "step": 10620 }, { "epoch": 2.877639415268002, "grad_norm": 0.27591705322265625, "learning_rate": 2.057912948461118e-05, "loss": 0.0933, "step": 10630 }, { "epoch": 2.8803465078505686, "grad_norm": 0.3215999901294708, "learning_rate": 2.0123839009287925e-05, "loss": 0.1045, "step": 10640 }, { "epoch": 2.8830536004331346, "grad_norm": 0.2435692399740219, "learning_rate": 1.966854853396467e-05, "loss": 0.0967, "step": 10650 }, { "epoch": 2.885760693015701, "grad_norm": 0.2958237826824188, "learning_rate": 1.9213258058641412e-05, "loss": 0.0843, "step": 10660 }, { "epoch": 2.8884677855982677, "grad_norm": 0.4848542809486389, "learning_rate": 1.875796758331816e-05, "loss": 0.1119, "step": 10670 }, { "epoch": 2.8911748781808337, "grad_norm": 0.09771529585123062, "learning_rate": 1.8302677107994902e-05, "loss": 0.0851, "step": 10680 }, { "epoch": 2.8938819707634003, "grad_norm": 0.30730119347572327, "learning_rate": 1.7847386632671646e-05, "loss": 0.0978, "step": 10690 }, { "epoch": 2.8965890633459663, "grad_norm": 0.16735531389713287, "learning_rate": 1.7392096157348386e-05, "loss": 0.1139, "step": 10700 }, { "epoch": 2.8965890633459663, "eval_loss": 0.14749093353748322, "eval_runtime": 52.8044, "eval_samples_per_second": 9.469, "eval_steps_per_second": 0.303, "step": 10700 }, { "epoch": 2.899296155928533, "grad_norm": 0.18200255930423737, "learning_rate": 1.6936805682025133e-05, "loss": 0.1289, "step": 10710 }, { "epoch": 2.902003248511099, "grad_norm": 0.2688741683959961, "learning_rate": 1.6481515206701876e-05, "loss": 0.077, "step": 10720 }, { "epoch": 2.9047103410936654, "grad_norm": 0.34718960523605347, "learning_rate": 1.602622473137862e-05, "loss": 0.136, "step": 10730 }, { "epoch": 2.9074174336762315, "grad_norm": 0.21658600866794586, "learning_rate": 1.5570934256055363e-05, "loss": 0.1305, "step": 10740 }, { "epoch": 2.910124526258798, "grad_norm": 0.31314098834991455, "learning_rate": 1.5115643780732108e-05, "loss": 0.1157, "step": 10750 }, { "epoch": 2.9128316188413645, "grad_norm": 0.10926242917776108, "learning_rate": 1.466035330540885e-05, "loss": 0.096, "step": 10760 }, { "epoch": 2.9155387114239306, "grad_norm": 0.3475625216960907, "learning_rate": 1.4205062830085595e-05, "loss": 0.1047, "step": 10770 }, { "epoch": 2.918245804006497, "grad_norm": 0.2349124401807785, "learning_rate": 1.3749772354762338e-05, "loss": 0.1017, "step": 10780 }, { "epoch": 2.920952896589063, "grad_norm": 0.29219192266464233, "learning_rate": 1.3294481879439084e-05, "loss": 0.1327, "step": 10790 }, { "epoch": 2.9236599891716297, "grad_norm": 0.26028120517730713, "learning_rate": 1.2839191404115825e-05, "loss": 0.0921, "step": 10800 }, { "epoch": 2.9236599891716297, "eval_loss": 0.1457306444644928, "eval_runtime": 53.5264, "eval_samples_per_second": 9.341, "eval_steps_per_second": 0.299, "step": 10800 }, { "epoch": 2.9263670817541962, "grad_norm": 0.2346615195274353, "learning_rate": 1.238390092879257e-05, "loss": 0.1017, "step": 10810 }, { "epoch": 2.9290741743367623, "grad_norm": 0.1553301364183426, "learning_rate": 1.1928610453469314e-05, "loss": 0.1024, "step": 10820 }, { "epoch": 2.9317812669193284, "grad_norm": 0.18650998175144196, "learning_rate": 1.1473319978146057e-05, "loss": 0.1278, "step": 10830 }, { "epoch": 2.934488359501895, "grad_norm": 0.3397215008735657, "learning_rate": 1.1018029502822801e-05, "loss": 0.1161, "step": 10840 }, { "epoch": 2.9371954520844614, "grad_norm": 0.3774276077747345, "learning_rate": 1.0562739027499544e-05, "loss": 0.1111, "step": 10850 }, { "epoch": 2.9399025446670275, "grad_norm": 0.06833741813898087, "learning_rate": 1.010744855217629e-05, "loss": 0.1067, "step": 10860 }, { "epoch": 2.942609637249594, "grad_norm": 0.11081808060407639, "learning_rate": 9.652158076853031e-06, "loss": 0.1156, "step": 10870 }, { "epoch": 2.94531672983216, "grad_norm": 0.21258531510829926, "learning_rate": 9.196867601529776e-06, "loss": 0.1178, "step": 10880 }, { "epoch": 2.9480238224147266, "grad_norm": 0.4802851676940918, "learning_rate": 8.74157712620652e-06, "loss": 0.1103, "step": 10890 }, { "epoch": 2.950730914997293, "grad_norm": 0.3009449541568756, "learning_rate": 8.286286650883265e-06, "loss": 0.0918, "step": 10900 }, { "epoch": 2.950730914997293, "eval_loss": 0.1471521258354187, "eval_runtime": 52.5879, "eval_samples_per_second": 9.508, "eval_steps_per_second": 0.304, "step": 10900 }, { "epoch": 2.953438007579859, "grad_norm": 0.19883263111114502, "learning_rate": 7.830996175560007e-06, "loss": 0.0901, "step": 10910 }, { "epoch": 2.9561451001624257, "grad_norm": 0.16847321391105652, "learning_rate": 7.375705700236751e-06, "loss": 0.1249, "step": 10920 }, { "epoch": 2.958852192744992, "grad_norm": 0.4061388373374939, "learning_rate": 6.920415224913495e-06, "loss": 0.1278, "step": 10930 }, { "epoch": 2.9615592853275583, "grad_norm": 0.22170685231685638, "learning_rate": 6.465124749590239e-06, "loss": 0.0946, "step": 10940 }, { "epoch": 2.9642663779101244, "grad_norm": 0.41052505373954773, "learning_rate": 6.009834274266983e-06, "loss": 0.1146, "step": 10950 }, { "epoch": 2.966973470492691, "grad_norm": 0.3749537765979767, "learning_rate": 5.554543798943726e-06, "loss": 0.1297, "step": 10960 }, { "epoch": 2.969680563075257, "grad_norm": 0.13162532448768616, "learning_rate": 5.09925332362047e-06, "loss": 0.1492, "step": 10970 }, { "epoch": 2.9723876556578235, "grad_norm": 0.32849496603012085, "learning_rate": 4.6439628482972135e-06, "loss": 0.1038, "step": 10980 }, { "epoch": 2.97509474824039, "grad_norm": 0.15441325306892395, "learning_rate": 4.188672372973958e-06, "loss": 0.0905, "step": 10990 }, { "epoch": 2.977801840822956, "grad_norm": 0.14825201034545898, "learning_rate": 3.7333818976507013e-06, "loss": 0.0941, "step": 11000 }, { "epoch": 2.977801840822956, "eval_loss": 0.1465151160955429, "eval_runtime": 51.5309, "eval_samples_per_second": 9.703, "eval_steps_per_second": 0.31, "step": 11000 }, { "epoch": 2.9805089334055226, "grad_norm": 0.5358628630638123, "learning_rate": 3.278091422327445e-06, "loss": 0.118, "step": 11010 }, { "epoch": 2.9832160259880887, "grad_norm": 0.27287015318870544, "learning_rate": 2.8228009470041886e-06, "loss": 0.1072, "step": 11020 }, { "epoch": 2.985923118570655, "grad_norm": 0.17518705129623413, "learning_rate": 2.3675104716809325e-06, "loss": 0.1033, "step": 11030 }, { "epoch": 2.9886302111532217, "grad_norm": 0.27911001443862915, "learning_rate": 1.9122199963576764e-06, "loss": 0.0967, "step": 11040 }, { "epoch": 2.9913373037357878, "grad_norm": 0.17551441490650177, "learning_rate": 1.4569295210344198e-06, "loss": 0.1081, "step": 11050 }, { "epoch": 2.994044396318354, "grad_norm": 0.2103782594203949, "learning_rate": 1.0016390457111637e-06, "loss": 0.1113, "step": 11060 }, { "epoch": 2.9967514889009204, "grad_norm": 0.13261333107948303, "learning_rate": 5.463485703879075e-07, "loss": 0.0912, "step": 11070 }, { "epoch": 2.999458581483487, "grad_norm": 0.20483067631721497, "learning_rate": 9.105809506465124e-08, "loss": 0.0981, "step": 11080 } ], "logging_steps": 10, "max_steps": 11082, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.041237140368352e+17, "train_batch_size": 32, "trial_name": null, "trial_params": null }