{ "best_metric": null, "best_model_checkpoint": null, "epoch": 9.98120873539868, "eval_steps": 500, "global_step": 4920, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.02031488065007618, "grad_norm": 7.584718704223633, "learning_rate": 1.999990002577244e-05, "loss": 3.2828, "step": 10 }, { "epoch": 0.04062976130015236, "grad_norm": 6.282063007354736, "learning_rate": 1.999941036089858e-05, "loss": 3.084, "step": 20 }, { "epoch": 0.06094464195022854, "grad_norm": 4.432021617889404, "learning_rate": 1.9998512662721364e-05, "loss": 2.9074, "step": 30 }, { "epoch": 0.08125952260030472, "grad_norm": 4.4949116706848145, "learning_rate": 1.9997206967872025e-05, "loss": 2.7432, "step": 40 }, { "epoch": 0.1015744032503809, "grad_norm": 3.794994592666626, "learning_rate": 1.99954933296304e-05, "loss": 2.5308, "step": 50 }, { "epoch": 0.12188928390045708, "grad_norm": 3.7512996196746826, "learning_rate": 1.999337181792276e-05, "loss": 2.3957, "step": 60 }, { "epoch": 0.14220416455053325, "grad_norm": 3.3922431468963623, "learning_rate": 1.999084251931896e-05, "loss": 2.2018, "step": 70 }, { "epoch": 0.16251904520060945, "grad_norm": 2.835357666015625, "learning_rate": 1.998790553702888e-05, "loss": 2.0505, "step": 80 }, { "epoch": 0.1828339258506856, "grad_norm": 2.4077184200286865, "learning_rate": 1.998456099089825e-05, "loss": 1.8802, "step": 90 }, { "epoch": 0.2031488065007618, "grad_norm": 2.9254958629608154, "learning_rate": 1.998080901740374e-05, "loss": 1.772, "step": 100 }, { "epoch": 0.22346368715083798, "grad_norm": 2.102215051651001, "learning_rate": 1.9976649769647392e-05, "loss": 1.7375, "step": 110 }, { "epoch": 0.24377856780091417, "grad_norm": 2.1416189670562744, "learning_rate": 1.997208341735037e-05, "loss": 1.6127, "step": 120 }, { "epoch": 0.26409344845099036, "grad_norm": 1.9121668338775635, "learning_rate": 1.9967110146846044e-05, "loss": 1.5663, "step": 130 }, { "epoch": 0.2844083291010665, "grad_norm": 1.7665735483169556, "learning_rate": 1.9961730161072383e-05, "loss": 1.5384, "step": 140 }, { "epoch": 0.3047232097511427, "grad_norm": 2.695359468460083, "learning_rate": 1.9955943679563655e-05, "loss": 1.4752, "step": 150 }, { "epoch": 0.3250380904012189, "grad_norm": 1.7461923360824585, "learning_rate": 1.9949750938441517e-05, "loss": 1.4581, "step": 160 }, { "epoch": 0.3453529710512951, "grad_norm": 1.856467366218567, "learning_rate": 1.994315219040532e-05, "loss": 1.4243, "step": 170 }, { "epoch": 0.3656678517013712, "grad_norm": 1.6817877292633057, "learning_rate": 1.9936147704721836e-05, "loss": 1.4437, "step": 180 }, { "epoch": 0.3859827323514474, "grad_norm": 1.6875474452972412, "learning_rate": 1.9928737767214267e-05, "loss": 1.4146, "step": 190 }, { "epoch": 0.4062976130015236, "grad_norm": 2.079127311706543, "learning_rate": 1.992092268025057e-05, "loss": 1.3857, "step": 200 }, { "epoch": 0.4266124936515998, "grad_norm": 1.4948805570602417, "learning_rate": 1.9912702762731118e-05, "loss": 1.3748, "step": 210 }, { "epoch": 0.44692737430167595, "grad_norm": 1.6643149852752686, "learning_rate": 1.9904078350075703e-05, "loss": 1.3904, "step": 220 }, { "epoch": 0.46724225495175215, "grad_norm": 2.044997215270996, "learning_rate": 1.9895049794209834e-05, "loss": 1.3519, "step": 230 }, { "epoch": 0.48755713560182834, "grad_norm": 1.9969228506088257, "learning_rate": 1.9885617463550376e-05, "loss": 1.3637, "step": 240 }, { "epoch": 0.5078720162519045, "grad_norm": 2.5120513439178467, "learning_rate": 1.987678345628239e-05, "loss": 1.3516, "step": 250 }, { "epoch": 0.5281868969019807, "grad_norm": 2.049866199493408, "learning_rate": 1.9866585027500926e-05, "loss": 1.3149, "step": 260 }, { "epoch": 0.5485017775520569, "grad_norm": 2.5176775455474854, "learning_rate": 1.9855983985451647e-05, "loss": 1.2744, "step": 270 }, { "epoch": 0.568816658202133, "grad_norm": 2.2515692710876465, "learning_rate": 1.9844980762717877e-05, "loss": 1.2986, "step": 280 }, { "epoch": 0.5891315388522093, "grad_norm": 1.9790077209472656, "learning_rate": 1.983357580829422e-05, "loss": 1.2556, "step": 290 }, { "epoch": 0.6094464195022854, "grad_norm": 2.0562939643859863, "learning_rate": 1.982176958756825e-05, "loss": 1.2565, "step": 300 }, { "epoch": 0.6297613001523616, "grad_norm": 2.895200252532959, "learning_rate": 1.9809562582301487e-05, "loss": 1.2217, "step": 310 }, { "epoch": 0.6500761808024378, "grad_norm": 2.379382371902466, "learning_rate": 1.9796955290609775e-05, "loss": 1.2186, "step": 320 }, { "epoch": 0.6703910614525139, "grad_norm": 1.5008665323257446, "learning_rate": 1.9783948226942936e-05, "loss": 1.2093, "step": 330 }, { "epoch": 0.6907059421025902, "grad_norm": 2.294449806213379, "learning_rate": 1.9770541922063778e-05, "loss": 1.228, "step": 340 }, { "epoch": 0.7110208227526663, "grad_norm": 1.5949811935424805, "learning_rate": 1.975673692302645e-05, "loss": 1.211, "step": 350 }, { "epoch": 0.7313357034027425, "grad_norm": 2.0285587310791016, "learning_rate": 1.97425337931541e-05, "loss": 1.2224, "step": 360 }, { "epoch": 0.7516505840528187, "grad_norm": 1.7120383977890015, "learning_rate": 1.97279331120159e-05, "loss": 1.2181, "step": 370 }, { "epoch": 0.7719654647028948, "grad_norm": 2.0460448265075684, "learning_rate": 1.97129354754034e-05, "loss": 1.206, "step": 380 }, { "epoch": 0.7922803453529711, "grad_norm": 2.479123592376709, "learning_rate": 1.96975414953062e-05, "loss": 1.1987, "step": 390 }, { "epoch": 0.8125952260030472, "grad_norm": 2.1566030979156494, "learning_rate": 1.9681751799887e-05, "loss": 1.2081, "step": 400 }, { "epoch": 0.8329101066531234, "grad_norm": 1.7905540466308594, "learning_rate": 1.9665567033455948e-05, "loss": 1.205, "step": 410 }, { "epoch": 0.8532249873031996, "grad_norm": 1.5104587078094482, "learning_rate": 1.9648987856444356e-05, "loss": 1.1808, "step": 420 }, { "epoch": 0.8735398679532758, "grad_norm": 1.601035714149475, "learning_rate": 1.9632014945377748e-05, "loss": 1.2039, "step": 430 }, { "epoch": 0.8938547486033519, "grad_norm": 1.4743677377700806, "learning_rate": 1.9614648992848265e-05, "loss": 1.1923, "step": 440 }, { "epoch": 0.9141696292534282, "grad_norm": 2.2900314331054688, "learning_rate": 1.959689070748638e-05, "loss": 1.19, "step": 450 }, { "epoch": 0.9344845099035043, "grad_norm": 2.500079393386841, "learning_rate": 1.9578740813932017e-05, "loss": 1.1826, "step": 460 }, { "epoch": 0.9547993905535805, "grad_norm": 1.704733967781067, "learning_rate": 1.9560200052804935e-05, "loss": 1.1811, "step": 470 }, { "epoch": 0.9751142712036567, "grad_norm": 2.3521337509155273, "learning_rate": 1.954126918067455e-05, "loss": 1.1868, "step": 480 }, { "epoch": 0.9954291518537328, "grad_norm": 2.1910970211029053, "learning_rate": 1.9521948970029045e-05, "loss": 1.2032, "step": 490 }, { "epoch": 1.0, "eval_loss": 1.1905763149261475, "eval_runtime": 57.9944, "eval_samples_per_second": 33.969, "eval_steps_per_second": 8.501, "step": 493 }, { "epoch": 1.0142204164550532, "grad_norm": 1.6047745943069458, "learning_rate": 1.9502240209243834e-05, "loss": 1.1649, "step": 500 }, { "epoch": 1.0345352971051296, "grad_norm": 2.1489083766937256, "learning_rate": 1.948214370254942e-05, "loss": 1.1811, "step": 510 }, { "epoch": 1.0548501777552057, "grad_norm": 1.674709677696228, "learning_rate": 1.9461660269998556e-05, "loss": 1.1681, "step": 520 }, { "epoch": 1.0751650584052819, "grad_norm": 1.839033603668213, "learning_rate": 1.9440790747432797e-05, "loss": 1.1683, "step": 530 }, { "epoch": 1.095479939055358, "grad_norm": 2.0029451847076416, "learning_rate": 1.941953598644837e-05, "loss": 1.1471, "step": 540 }, { "epoch": 1.1157948197054341, "grad_norm": 2.1408133506774902, "learning_rate": 1.9397896854361454e-05, "loss": 1.1865, "step": 550 }, { "epoch": 1.1361097003555103, "grad_norm": 1.6449344158172607, "learning_rate": 1.937587423417277e-05, "loss": 1.1616, "step": 560 }, { "epoch": 1.1564245810055866, "grad_norm": 2.530738592147827, "learning_rate": 1.935346902453155e-05, "loss": 1.1847, "step": 570 }, { "epoch": 1.1767394616556628, "grad_norm": 1.7173138856887817, "learning_rate": 1.9330682139698872e-05, "loss": 1.1836, "step": 580 }, { "epoch": 1.197054342305739, "grad_norm": 1.9412487745285034, "learning_rate": 1.9307514509510358e-05, "loss": 1.1745, "step": 590 }, { "epoch": 1.217369222955815, "grad_norm": 1.9713300466537476, "learning_rate": 1.928396707933821e-05, "loss": 1.1498, "step": 600 }, { "epoch": 1.2376841036058912, "grad_norm": 2.253971815109253, "learning_rate": 1.926004081005266e-05, "loss": 1.1554, "step": 610 }, { "epoch": 1.2579989842559676, "grad_norm": 1.9793624877929688, "learning_rate": 1.9235736677982747e-05, "loss": 1.1445, "step": 620 }, { "epoch": 1.2783138649060437, "grad_norm": 2.292801856994629, "learning_rate": 1.9211055674876475e-05, "loss": 1.1373, "step": 630 }, { "epoch": 1.2986287455561198, "grad_norm": 1.8089720010757446, "learning_rate": 1.9185998807860357e-05, "loss": 1.1324, "step": 640 }, { "epoch": 1.318943626206196, "grad_norm": 2.003289222717285, "learning_rate": 1.9160567099398294e-05, "loss": 1.173, "step": 650 }, { "epoch": 1.3392585068562721, "grad_norm": 2.1809918880462646, "learning_rate": 1.9134761587249886e-05, "loss": 1.1408, "step": 660 }, { "epoch": 1.3595733875063485, "grad_norm": 2.0070571899414062, "learning_rate": 1.9108583324428054e-05, "loss": 1.1811, "step": 670 }, { "epoch": 1.3798882681564246, "grad_norm": 1.6615877151489258, "learning_rate": 1.9082033379156096e-05, "loss": 1.164, "step": 680 }, { "epoch": 1.4002031488065008, "grad_norm": 2.217325210571289, "learning_rate": 1.9055112834824075e-05, "loss": 1.1534, "step": 690 }, { "epoch": 1.420518029456577, "grad_norm": 1.8769993782043457, "learning_rate": 1.9027822789944625e-05, "loss": 1.1476, "step": 700 }, { "epoch": 1.440832910106653, "grad_norm": 1.8755251169204712, "learning_rate": 1.900016435810813e-05, "loss": 1.1773, "step": 710 }, { "epoch": 1.4611477907567294, "grad_norm": 2.059292793273926, "learning_rate": 1.8972138667937268e-05, "loss": 1.1512, "step": 720 }, { "epoch": 1.4814626714068055, "grad_norm": 2.2438113689422607, "learning_rate": 1.894374686304096e-05, "loss": 1.1313, "step": 730 }, { "epoch": 1.5017775520568817, "grad_norm": 1.6689549684524536, "learning_rate": 1.8914990101967717e-05, "loss": 1.1506, "step": 740 }, { "epoch": 1.5220924327069578, "grad_norm": 1.9220091104507446, "learning_rate": 1.8885869558158348e-05, "loss": 1.1389, "step": 750 }, { "epoch": 1.542407313357034, "grad_norm": 1.970219373703003, "learning_rate": 1.885638641989809e-05, "loss": 1.1535, "step": 760 }, { "epoch": 1.5627221940071103, "grad_norm": 2.3392066955566406, "learning_rate": 1.882654189026811e-05, "loss": 1.1467, "step": 770 }, { "epoch": 1.5830370746571862, "grad_norm": 1.9110764265060425, "learning_rate": 1.8796337187096415e-05, "loss": 1.1457, "step": 780 }, { "epoch": 1.6033519553072626, "grad_norm": 3.4612278938293457, "learning_rate": 1.8765773542908165e-05, "loss": 1.1443, "step": 790 }, { "epoch": 1.6236668359573387, "grad_norm": 1.8444197177886963, "learning_rate": 1.8734852204875363e-05, "loss": 1.1375, "step": 800 }, { "epoch": 1.6439817166074149, "grad_norm": 1.7353184223175049, "learning_rate": 1.8703574434765986e-05, "loss": 1.1802, "step": 810 }, { "epoch": 1.6642965972574912, "grad_norm": 2.283703327178955, "learning_rate": 1.8671941508892467e-05, "loss": 1.1457, "step": 820 }, { "epoch": 1.6846114779075672, "grad_norm": 2.3982231616973877, "learning_rate": 1.8639954718059647e-05, "loss": 1.134, "step": 830 }, { "epoch": 1.7049263585576435, "grad_norm": 2.6923394203186035, "learning_rate": 1.8607615367512077e-05, "loss": 1.1387, "step": 840 }, { "epoch": 1.7252412392077197, "grad_norm": 1.870726466178894, "learning_rate": 1.8574924776880777e-05, "loss": 1.1479, "step": 850 }, { "epoch": 1.7455561198577958, "grad_norm": 2.014890432357788, "learning_rate": 1.8541884280129367e-05, "loss": 1.1724, "step": 860 }, { "epoch": 1.7658710005078722, "grad_norm": 1.8836543560028076, "learning_rate": 1.8508495225499643e-05, "loss": 1.1483, "step": 870 }, { "epoch": 1.786185881157948, "grad_norm": 1.8277486562728882, "learning_rate": 1.8474758975456576e-05, "loss": 1.121, "step": 880 }, { "epoch": 1.8065007618080244, "grad_norm": 1.8098642826080322, "learning_rate": 1.8440676906632688e-05, "loss": 1.1404, "step": 890 }, { "epoch": 1.8268156424581006, "grad_norm": 2.0172536373138428, "learning_rate": 1.8406250409771894e-05, "loss": 1.1424, "step": 900 }, { "epoch": 1.8471305231081767, "grad_norm": 2.2588918209075928, "learning_rate": 1.8371480889672756e-05, "loss": 1.1394, "step": 910 }, { "epoch": 1.867445403758253, "grad_norm": 2.9289000034332275, "learning_rate": 1.8336369765131137e-05, "loss": 1.1273, "step": 920 }, { "epoch": 1.887760284408329, "grad_norm": 1.9783216714859009, "learning_rate": 1.8300918468882337e-05, "loss": 1.1316, "step": 930 }, { "epoch": 1.9080751650584054, "grad_norm": 2.6187262535095215, "learning_rate": 1.8265128447542606e-05, "loss": 1.1347, "step": 940 }, { "epoch": 1.9283900457084815, "grad_norm": 1.7901383638381958, "learning_rate": 1.822900116155012e-05, "loss": 1.1419, "step": 950 }, { "epoch": 1.9487049263585576, "grad_norm": 2.4337475299835205, "learning_rate": 1.819253808510538e-05, "loss": 1.1253, "step": 960 }, { "epoch": 1.9690198070086338, "grad_norm": 1.88914155960083, "learning_rate": 1.8155740706111086e-05, "loss": 1.1278, "step": 970 }, { "epoch": 1.98933468765871, "grad_norm": 2.087872266769409, "learning_rate": 1.811861052611137e-05, "loss": 1.1239, "step": 980 }, { "epoch": 2.0, "eval_loss": 1.1426180601119995, "eval_runtime": 56.1438, "eval_samples_per_second": 35.088, "eval_steps_per_second": 8.781, "step": 986 }, { "epoch": 2.0081259522600305, "grad_norm": 2.3936855792999268, "learning_rate": 1.8081149060230584e-05, "loss": 1.1307, "step": 990 }, { "epoch": 2.0284408329101065, "grad_norm": 2.97658371925354, "learning_rate": 1.804335783711142e-05, "loss": 1.1115, "step": 1000 }, { "epoch": 2.048755713560183, "grad_norm": 1.7273203134536743, "learning_rate": 1.8005238398852567e-05, "loss": 1.1145, "step": 1010 }, { "epoch": 2.069070594210259, "grad_norm": 2.1559698581695557, "learning_rate": 1.7966792300945782e-05, "loss": 1.1333, "step": 1020 }, { "epoch": 2.089385474860335, "grad_norm": 1.8381649255752563, "learning_rate": 1.7928021112212397e-05, "loss": 1.0982, "step": 1030 }, { "epoch": 2.1097003555104115, "grad_norm": 2.330662727355957, "learning_rate": 1.788892641473932e-05, "loss": 1.1246, "step": 1040 }, { "epoch": 2.1300152361604874, "grad_norm": 2.1321730613708496, "learning_rate": 1.784950980381447e-05, "loss": 1.1182, "step": 1050 }, { "epoch": 2.1503301168105637, "grad_norm": 2.3504936695098877, "learning_rate": 1.7809772887861684e-05, "loss": 1.1449, "step": 1060 }, { "epoch": 2.17064499746064, "grad_norm": 1.9942519664764404, "learning_rate": 1.7769717288375078e-05, "loss": 1.1355, "step": 1070 }, { "epoch": 2.190959878110716, "grad_norm": 2.6634910106658936, "learning_rate": 1.7729344639852886e-05, "loss": 1.1317, "step": 1080 }, { "epoch": 2.2112747587607924, "grad_norm": 2.0692458152770996, "learning_rate": 1.768865658973076e-05, "loss": 1.1349, "step": 1090 }, { "epoch": 2.2315896394108683, "grad_norm": 2.0810651779174805, "learning_rate": 1.7647654798314554e-05, "loss": 1.1084, "step": 1100 }, { "epoch": 2.2519045200609447, "grad_norm": 1.9821966886520386, "learning_rate": 1.760634093871256e-05, "loss": 1.1274, "step": 1110 }, { "epoch": 2.2722194007110206, "grad_norm": 2.893826484680176, "learning_rate": 1.756471669676723e-05, "loss": 1.1262, "step": 1120 }, { "epoch": 2.292534281361097, "grad_norm": 1.9358024597167969, "learning_rate": 1.7522783770986425e-05, "loss": 1.1138, "step": 1130 }, { "epoch": 2.3128491620111733, "grad_norm": 1.746727466583252, "learning_rate": 1.748054387247405e-05, "loss": 1.1313, "step": 1140 }, { "epoch": 2.333164042661249, "grad_norm": 1.878308892250061, "learning_rate": 1.743799872486028e-05, "loss": 1.1313, "step": 1150 }, { "epoch": 2.3534789233113256, "grad_norm": 2.322291135787964, "learning_rate": 1.739515006423118e-05, "loss": 1.1173, "step": 1160 }, { "epoch": 2.3737938039614015, "grad_norm": 1.9874950647354126, "learning_rate": 1.7351999639057915e-05, "loss": 1.1023, "step": 1170 }, { "epoch": 2.394108684611478, "grad_norm": 2.070664882659912, "learning_rate": 1.7308549210125346e-05, "loss": 1.1208, "step": 1180 }, { "epoch": 2.414423565261554, "grad_norm": 1.8656115531921387, "learning_rate": 1.7264800550460235e-05, "loss": 1.1324, "step": 1190 }, { "epoch": 2.43473844591163, "grad_norm": 1.7181984186172485, "learning_rate": 1.7220755445258847e-05, "loss": 1.1299, "step": 1200 }, { "epoch": 2.4550533265617065, "grad_norm": 2.0644876956939697, "learning_rate": 1.717641569181414e-05, "loss": 1.1057, "step": 1210 }, { "epoch": 2.4753682072117824, "grad_norm": 2.54202938079834, "learning_rate": 1.7131783099442408e-05, "loss": 1.1371, "step": 1220 }, { "epoch": 2.4956830878618588, "grad_norm": 1.8502358198165894, "learning_rate": 1.708685948940944e-05, "loss": 1.0938, "step": 1230 }, { "epoch": 2.515997968511935, "grad_norm": 2.2226638793945312, "learning_rate": 1.7041646694856237e-05, "loss": 1.1306, "step": 1240 }, { "epoch": 2.536312849162011, "grad_norm": 2.0658068656921387, "learning_rate": 1.6996146560724163e-05, "loss": 1.1465, "step": 1250 }, { "epoch": 2.5566277298120874, "grad_norm": 2.187943458557129, "learning_rate": 1.6950360943679708e-05, "loss": 1.1071, "step": 1260 }, { "epoch": 2.5769426104621633, "grad_norm": 2.104330062866211, "learning_rate": 1.690429171203869e-05, "loss": 1.1247, "step": 1270 }, { "epoch": 2.5972574911122397, "grad_norm": 2.1676745414733887, "learning_rate": 1.685794074569003e-05, "loss": 1.112, "step": 1280 }, { "epoch": 2.617572371762316, "grad_norm": 1.8994890451431274, "learning_rate": 1.6811309936019034e-05, "loss": 1.1129, "step": 1290 }, { "epoch": 2.637887252412392, "grad_norm": 1.9885021448135376, "learning_rate": 1.6764401185830248e-05, "loss": 1.1074, "step": 1300 }, { "epoch": 2.6582021330624683, "grad_norm": 1.8062092065811157, "learning_rate": 1.6717216409269756e-05, "loss": 1.1083, "step": 1310 }, { "epoch": 2.6785170137125442, "grad_norm": 2.118276834487915, "learning_rate": 1.666975753174711e-05, "loss": 1.0967, "step": 1320 }, { "epoch": 2.6988318943626206, "grad_norm": 2.219618082046509, "learning_rate": 1.6622026489856767e-05, "loss": 1.1294, "step": 1330 }, { "epoch": 2.719146775012697, "grad_norm": 2.3651201725006104, "learning_rate": 1.657402523129903e-05, "loss": 1.1303, "step": 1340 }, { "epoch": 2.739461655662773, "grad_norm": 2.132640838623047, "learning_rate": 1.6525755714800605e-05, "loss": 1.0982, "step": 1350 }, { "epoch": 2.7597765363128492, "grad_norm": 2.0165741443634033, "learning_rate": 1.6477219910034662e-05, "loss": 1.1215, "step": 1360 }, { "epoch": 2.780091416962925, "grad_norm": 2.050801992416382, "learning_rate": 1.6428419797540455e-05, "loss": 1.0938, "step": 1370 }, { "epoch": 2.8004062976130015, "grad_norm": 2.0775163173675537, "learning_rate": 1.637935736864251e-05, "loss": 1.0976, "step": 1380 }, { "epoch": 2.820721178263078, "grad_norm": 1.9527651071548462, "learning_rate": 1.6330034625369366e-05, "loss": 1.108, "step": 1390 }, { "epoch": 2.841036058913154, "grad_norm": 2.5815951824188232, "learning_rate": 1.6280453580371887e-05, "loss": 1.1187, "step": 1400 }, { "epoch": 2.86135093956323, "grad_norm": 2.0379745960235596, "learning_rate": 1.6230616256841122e-05, "loss": 1.1263, "step": 1410 }, { "epoch": 2.881665820213306, "grad_norm": 2.2415878772735596, "learning_rate": 1.618052468842576e-05, "loss": 1.13, "step": 1420 }, { "epoch": 2.9019807008633824, "grad_norm": 1.951768159866333, "learning_rate": 1.6130180919149145e-05, "loss": 1.1179, "step": 1430 }, { "epoch": 2.922295581513459, "grad_norm": 2.3024680614471436, "learning_rate": 1.607958700332584e-05, "loss": 1.0993, "step": 1440 }, { "epoch": 2.9426104621635347, "grad_norm": 1.86294686794281, "learning_rate": 1.602874500547785e-05, "loss": 1.1071, "step": 1450 }, { "epoch": 2.962925342813611, "grad_norm": 2.225184202194214, "learning_rate": 1.5977657000250332e-05, "loss": 1.1189, "step": 1460 }, { "epoch": 2.983240223463687, "grad_norm": 2.0135252475738525, "learning_rate": 1.5926325072326963e-05, "loss": 1.1019, "step": 1470 }, { "epoch": 3.0, "eval_loss": 1.1209979057312012, "eval_runtime": 55.7989, "eval_samples_per_second": 35.305, "eval_steps_per_second": 8.835, "step": 1479 }, { "epoch": 3.0020314880650076, "grad_norm": 2.6400208473205566, "learning_rate": 1.587475131634486e-05, "loss": 1.1055, "step": 1480 }, { "epoch": 3.022346368715084, "grad_norm": 1.8572973012924194, "learning_rate": 1.582293783680912e-05, "loss": 1.1214, "step": 1490 }, { "epoch": 3.04266124936516, "grad_norm": 2.232374668121338, "learning_rate": 1.5770886748006925e-05, "loss": 1.1016, "step": 1500 }, { "epoch": 3.0629761300152363, "grad_norm": 1.8693956136703491, "learning_rate": 1.5718600173921287e-05, "loss": 1.1011, "step": 1510 }, { "epoch": 3.083291010665312, "grad_norm": 1.9302908182144165, "learning_rate": 1.5666080248144364e-05, "loss": 1.0913, "step": 1520 }, { "epoch": 3.1036058913153886, "grad_norm": 2.3789591789245605, "learning_rate": 1.5613329113790407e-05, "loss": 1.1153, "step": 1530 }, { "epoch": 3.123920771965465, "grad_norm": 2.4649009704589844, "learning_rate": 1.5560348923408297e-05, "loss": 1.1046, "step": 1540 }, { "epoch": 3.144235652615541, "grad_norm": 1.7256394624710083, "learning_rate": 1.5507141838893714e-05, "loss": 1.1111, "step": 1550 }, { "epoch": 3.164550533265617, "grad_norm": 2.8762755393981934, "learning_rate": 1.545371003140093e-05, "loss": 1.1196, "step": 1560 }, { "epoch": 3.184865413915693, "grad_norm": 1.8774479627609253, "learning_rate": 1.540005568125419e-05, "loss": 1.0758, "step": 1570 }, { "epoch": 3.2051802945657695, "grad_norm": 2.3003156185150146, "learning_rate": 1.5346180977858766e-05, "loss": 1.1024, "step": 1580 }, { "epoch": 3.225495175215846, "grad_norm": 1.9327248334884644, "learning_rate": 1.529208811961161e-05, "loss": 1.1217, "step": 1590 }, { "epoch": 3.2458100558659218, "grad_norm": 2.096296548843384, "learning_rate": 1.5237779313811622e-05, "loss": 1.0966, "step": 1600 }, { "epoch": 3.266124936515998, "grad_norm": 1.896620750427246, "learning_rate": 1.5183256776569625e-05, "loss": 1.1019, "step": 1610 }, { "epoch": 3.286439817166074, "grad_norm": 2.3992044925689697, "learning_rate": 1.5128522732717907e-05, "loss": 1.1048, "step": 1620 }, { "epoch": 3.3067546978161504, "grad_norm": 2.170178174972534, "learning_rate": 1.5073579415719422e-05, "loss": 1.0886, "step": 1630 }, { "epoch": 3.3270695784662268, "grad_norm": 2.299849510192871, "learning_rate": 1.5018429067576691e-05, "loss": 1.093, "step": 1640 }, { "epoch": 3.3473844591163027, "grad_norm": 2.0816946029663086, "learning_rate": 1.4963073938740276e-05, "loss": 1.1009, "step": 1650 }, { "epoch": 3.367699339766379, "grad_norm": 2.180161237716675, "learning_rate": 1.4907516288016972e-05, "loss": 1.1239, "step": 1660 }, { "epoch": 3.388014220416455, "grad_norm": 2.5566606521606445, "learning_rate": 1.4851758382477632e-05, "loss": 1.1168, "step": 1670 }, { "epoch": 3.4083291010665313, "grad_norm": 2.167325019836426, "learning_rate": 1.4795802497364655e-05, "loss": 1.0941, "step": 1680 }, { "epoch": 3.4286439817166072, "grad_norm": 2.132948875427246, "learning_rate": 1.4739650915999136e-05, "loss": 1.0935, "step": 1690 }, { "epoch": 3.4489588623666836, "grad_norm": 2.1108901500701904, "learning_rate": 1.46833059296877e-05, "loss": 1.0963, "step": 1700 }, { "epoch": 3.46927374301676, "grad_norm": 2.1430904865264893, "learning_rate": 1.4626769837629012e-05, "loss": 1.1058, "step": 1710 }, { "epoch": 3.489588623666836, "grad_norm": 2.2873406410217285, "learning_rate": 1.4570044946819937e-05, "loss": 1.0925, "step": 1720 }, { "epoch": 3.5099035043169122, "grad_norm": 2.4914746284484863, "learning_rate": 1.4513133571961415e-05, "loss": 1.1094, "step": 1730 }, { "epoch": 3.5302183849669886, "grad_norm": 2.074937582015991, "learning_rate": 1.4456038035364014e-05, "loss": 1.1123, "step": 1740 }, { "epoch": 3.5505332656170645, "grad_norm": 2.091817855834961, "learning_rate": 1.4398760666853143e-05, "loss": 1.0965, "step": 1750 }, { "epoch": 3.5708481462671404, "grad_norm": 1.8256065845489502, "learning_rate": 1.434130380367401e-05, "loss": 1.1179, "step": 1760 }, { "epoch": 3.591163026917217, "grad_norm": 2.202068567276001, "learning_rate": 1.428366979039622e-05, "loss": 1.1102, "step": 1770 }, { "epoch": 3.611477907567293, "grad_norm": 1.8671810626983643, "learning_rate": 1.4225860978818134e-05, "loss": 1.089, "step": 1780 }, { "epoch": 3.631792788217369, "grad_norm": 2.394472599029541, "learning_rate": 1.4167879727870873e-05, "loss": 1.0968, "step": 1790 }, { "epoch": 3.6521076688674454, "grad_norm": 1.918837308883667, "learning_rate": 1.4109728403522088e-05, "loss": 1.0971, "step": 1800 }, { "epoch": 3.6724225495175213, "grad_norm": 2.170398235321045, "learning_rate": 1.405140937867938e-05, "loss": 1.0953, "step": 1810 }, { "epoch": 3.6927374301675977, "grad_norm": 1.9695782661437988, "learning_rate": 1.3992925033093514e-05, "loss": 1.0954, "step": 1820 }, { "epoch": 3.713052310817674, "grad_norm": 2.0776660442352295, "learning_rate": 1.3934277753261274e-05, "loss": 1.0989, "step": 1830 }, { "epoch": 3.73336719146775, "grad_norm": 2.3341407775878906, "learning_rate": 1.3875469932328104e-05, "loss": 1.1108, "step": 1840 }, { "epoch": 3.7536820721178263, "grad_norm": 2.07391095161438, "learning_rate": 1.3816503969990444e-05, "loss": 1.1194, "step": 1850 }, { "epoch": 3.7739969527679023, "grad_norm": 1.9829392433166504, "learning_rate": 1.3757382272397805e-05, "loss": 1.0695, "step": 1860 }, { "epoch": 3.7943118334179786, "grad_norm": 2.744811534881592, "learning_rate": 1.3698107252054604e-05, "loss": 1.0825, "step": 1870 }, { "epoch": 3.814626714068055, "grad_norm": 2.293558120727539, "learning_rate": 1.3638681327721686e-05, "loss": 1.0942, "step": 1880 }, { "epoch": 3.834941594718131, "grad_norm": 2.2765190601348877, "learning_rate": 1.357910692431766e-05, "loss": 1.0907, "step": 1890 }, { "epoch": 3.8552564753682073, "grad_norm": 2.5928955078125, "learning_rate": 1.3519386472819927e-05, "loss": 1.0911, "step": 1900 }, { "epoch": 3.875571356018283, "grad_norm": 2.018359661102295, "learning_rate": 1.3459522410165488e-05, "loss": 1.0965, "step": 1910 }, { "epoch": 3.8958862366683595, "grad_norm": 3.01343035697937, "learning_rate": 1.3399517179151504e-05, "loss": 1.0946, "step": 1920 }, { "epoch": 3.916201117318436, "grad_norm": 2.592369794845581, "learning_rate": 1.3339373228335614e-05, "loss": 1.0901, "step": 1930 }, { "epoch": 3.936515997968512, "grad_norm": 2.4442763328552246, "learning_rate": 1.3279093011936021e-05, "loss": 1.0906, "step": 1940 }, { "epoch": 3.956830878618588, "grad_norm": 2.2360129356384277, "learning_rate": 1.3218678989731352e-05, "loss": 1.0988, "step": 1950 }, { "epoch": 3.977145759268664, "grad_norm": 2.1769418716430664, "learning_rate": 1.3158133626960261e-05, "loss": 1.0978, "step": 1960 }, { "epoch": 3.9974606399187405, "grad_norm": 2.4375362396240234, "learning_rate": 1.3097459394220867e-05, "loss": 1.0874, "step": 1970 }, { "epoch": 4.0, "eval_loss": 1.1085847616195679, "eval_runtime": 55.3659, "eval_samples_per_second": 35.581, "eval_steps_per_second": 8.904, "step": 1972 }, { "epoch": 4.016251904520061, "grad_norm": 2.620096445083618, "learning_rate": 1.3036658767369919e-05, "loss": 1.1006, "step": 1980 }, { "epoch": 4.0365667851701375, "grad_norm": 2.652712345123291, "learning_rate": 1.2975734227421768e-05, "loss": 1.0797, "step": 1990 }, { "epoch": 4.056881665820213, "grad_norm": 2.4038774967193604, "learning_rate": 1.291468826044713e-05, "loss": 1.1085, "step": 2000 }, { "epoch": 4.077196546470289, "grad_norm": 2.6778948307037354, "learning_rate": 1.2853523357471644e-05, "loss": 1.0818, "step": 2010 }, { "epoch": 4.097511427120366, "grad_norm": 2.2972545623779297, "learning_rate": 1.2792242014374216e-05, "loss": 1.0921, "step": 2020 }, { "epoch": 4.117826307770442, "grad_norm": 2.1259145736694336, "learning_rate": 1.273084673178518e-05, "loss": 1.0911, "step": 2030 }, { "epoch": 4.138141188420518, "grad_norm": 2.6703901290893555, "learning_rate": 1.2669340014984252e-05, "loss": 1.0939, "step": 2040 }, { "epoch": 4.158456069070594, "grad_norm": 2.7287003993988037, "learning_rate": 1.260772437379831e-05, "loss": 1.0896, "step": 2050 }, { "epoch": 4.17877094972067, "grad_norm": 2.5945608615875244, "learning_rate": 1.254600232249896e-05, "loss": 1.0973, "step": 2060 }, { "epoch": 4.199085830370747, "grad_norm": 2.0844831466674805, "learning_rate": 1.248417637969996e-05, "loss": 1.0883, "step": 2070 }, { "epoch": 4.219400711020823, "grad_norm": 2.237929105758667, "learning_rate": 1.2422249068254433e-05, "loss": 1.0963, "step": 2080 }, { "epoch": 4.239715591670899, "grad_norm": 2.147749662399292, "learning_rate": 1.2360222915151929e-05, "loss": 1.0939, "step": 2090 }, { "epoch": 4.260030472320975, "grad_norm": 2.0512821674346924, "learning_rate": 1.22981004514153e-05, "loss": 1.0984, "step": 2100 }, { "epoch": 4.280345352971051, "grad_norm": 2.3941662311553955, "learning_rate": 1.223588421199743e-05, "loss": 1.1019, "step": 2110 }, { "epoch": 4.3006602336211275, "grad_norm": 2.5357017517089844, "learning_rate": 1.2173576735677775e-05, "loss": 1.1074, "step": 2120 }, { "epoch": 4.320975114271204, "grad_norm": 2.3033499717712402, "learning_rate": 1.2111180564958796e-05, "loss": 1.0841, "step": 2130 }, { "epoch": 4.34128999492128, "grad_norm": 2.249016046524048, "learning_rate": 1.2048698245962184e-05, "loss": 1.0746, "step": 2140 }, { "epoch": 4.361604875571356, "grad_norm": 2.2466604709625244, "learning_rate": 1.1986132328324986e-05, "loss": 1.105, "step": 2150 }, { "epoch": 4.381919756221432, "grad_norm": 2.7963321208953857, "learning_rate": 1.192348536509554e-05, "loss": 1.0876, "step": 2160 }, { "epoch": 4.402234636871508, "grad_norm": 2.159446954727173, "learning_rate": 1.1860759912629311e-05, "loss": 1.1072, "step": 2170 }, { "epoch": 4.422549517521585, "grad_norm": 2.521923542022705, "learning_rate": 1.179795853048458e-05, "loss": 1.1047, "step": 2180 }, { "epoch": 4.442864398171661, "grad_norm": 2.2701151371002197, "learning_rate": 1.1735083781317992e-05, "loss": 1.0879, "step": 2190 }, { "epoch": 4.463179278821737, "grad_norm": 2.706636667251587, "learning_rate": 1.167213823077999e-05, "loss": 1.0793, "step": 2200 }, { "epoch": 4.483494159471813, "grad_norm": 2.425516128540039, "learning_rate": 1.1609124447410114e-05, "loss": 1.093, "step": 2210 }, { "epoch": 4.503809040121889, "grad_norm": 2.4822216033935547, "learning_rate": 1.1546045002532202e-05, "loss": 1.0961, "step": 2220 }, { "epoch": 4.524123920771966, "grad_norm": 2.3154852390289307, "learning_rate": 1.148290247014945e-05, "loss": 1.0861, "step": 2230 }, { "epoch": 4.544438801422041, "grad_norm": 2.016235589981079, "learning_rate": 1.1419699426839393e-05, "loss": 1.0913, "step": 2240 }, { "epoch": 4.5647536820721175, "grad_norm": 2.701549530029297, "learning_rate": 1.1356438451648754e-05, "loss": 1.0918, "step": 2250 }, { "epoch": 4.585068562722194, "grad_norm": 2.02785587310791, "learning_rate": 1.1293122125988211e-05, "loss": 1.1, "step": 2260 }, { "epoch": 4.60538344337227, "grad_norm": 2.196712017059326, "learning_rate": 1.122975303352706e-05, "loss": 1.0657, "step": 2270 }, { "epoch": 4.625698324022347, "grad_norm": 2.8437767028808594, "learning_rate": 1.1166333760087784e-05, "loss": 1.1043, "step": 2280 }, { "epoch": 4.646013204672423, "grad_norm": 2.5000803470611572, "learning_rate": 1.1102866893540537e-05, "loss": 1.0855, "step": 2290 }, { "epoch": 4.666328085322498, "grad_norm": 2.488802194595337, "learning_rate": 1.1039355023697547e-05, "loss": 1.0792, "step": 2300 }, { "epoch": 4.686642965972575, "grad_norm": 2.04819655418396, "learning_rate": 1.097580074220743e-05, "loss": 1.0746, "step": 2310 }, { "epoch": 4.706957846622651, "grad_norm": 2.619755506515503, "learning_rate": 1.091220664244945e-05, "loss": 1.0807, "step": 2320 }, { "epoch": 4.7272727272727275, "grad_norm": 2.009214401245117, "learning_rate": 1.0848575319427673e-05, "loss": 1.0881, "step": 2330 }, { "epoch": 4.747587607922803, "grad_norm": 2.155071258544922, "learning_rate": 1.0784909369665092e-05, "loss": 1.0879, "step": 2340 }, { "epoch": 4.767902488572879, "grad_norm": 2.4336414337158203, "learning_rate": 1.0721211391097675e-05, "loss": 1.0915, "step": 2350 }, { "epoch": 4.788217369222956, "grad_norm": 2.942394733428955, "learning_rate": 1.0657483982968341e-05, "loss": 1.0753, "step": 2360 }, { "epoch": 4.808532249873032, "grad_norm": 2.1195554733276367, "learning_rate": 1.0593729745720913e-05, "loss": 1.067, "step": 2370 }, { "epoch": 4.828847130523108, "grad_norm": 2.387273073196411, "learning_rate": 1.0529951280893986e-05, "loss": 1.0828, "step": 2380 }, { "epoch": 4.849162011173185, "grad_norm": 2.598560094833374, "learning_rate": 1.0466151191014784e-05, "loss": 1.0905, "step": 2390 }, { "epoch": 4.86947689182326, "grad_norm": 2.338967800140381, "learning_rate": 1.0402332079492959e-05, "loss": 1.0888, "step": 2400 }, { "epoch": 4.889791772473337, "grad_norm": 2.038918972015381, "learning_rate": 1.033849655051435e-05, "loss": 1.0713, "step": 2410 }, { "epoch": 4.910106653123413, "grad_norm": 2.3864147663116455, "learning_rate": 1.0274647208934732e-05, "loss": 1.0997, "step": 2420 }, { "epoch": 4.930421533773489, "grad_norm": 2.313906669616699, "learning_rate": 1.02107866601735e-05, "loss": 1.088, "step": 2430 }, { "epoch": 4.950736414423565, "grad_norm": 1.9792237281799316, "learning_rate": 1.0146917510107377e-05, "loss": 1.0792, "step": 2440 }, { "epoch": 4.971051295073641, "grad_norm": 2.2212326526641846, "learning_rate": 1.0083042364964062e-05, "loss": 1.0804, "step": 2450 }, { "epoch": 4.9913661757237175, "grad_norm": 2.190819501876831, "learning_rate": 1.0019163831215894e-05, "loss": 1.0815, "step": 2460 }, { "epoch": 5.0, "eval_loss": 1.1006184816360474, "eval_runtime": 54.9569, "eval_samples_per_second": 35.846, "eval_steps_per_second": 8.971, "step": 2465 }, { "epoch": 5.010157440325038, "grad_norm": 2.375826597213745, "learning_rate": 9.955284515473478e-06, "loss": 1.0816, "step": 2470 }, { "epoch": 5.0304723209751145, "grad_norm": 2.3722262382507324, "learning_rate": 9.89140702437934e-06, "loss": 1.081, "step": 2480 }, { "epoch": 5.05078720162519, "grad_norm": 2.717057943344116, "learning_rate": 9.827533964501537e-06, "loss": 1.1021, "step": 2490 }, { "epoch": 5.071102082275266, "grad_norm": 2.2786624431610107, "learning_rate": 9.763667942227323e-06, "loss": 1.0983, "step": 2500 }, { "epoch": 5.091416962925343, "grad_norm": 2.0201127529144287, "learning_rate": 9.699811563656758e-06, "loss": 1.0694, "step": 2510 }, { "epoch": 5.111731843575419, "grad_norm": 2.4694550037384033, "learning_rate": 9.635967434496396e-06, "loss": 1.0899, "step": 2520 }, { "epoch": 5.1320467242254955, "grad_norm": 2.8942248821258545, "learning_rate": 9.572138159952945e-06, "loss": 1.099, "step": 2530 }, { "epoch": 5.152361604875571, "grad_norm": 2.079746723175049, "learning_rate": 9.508326344626955e-06, "loss": 1.0789, "step": 2540 }, { "epoch": 5.172676485525647, "grad_norm": 2.178121566772461, "learning_rate": 9.444534592406541e-06, "loss": 1.0763, "step": 2550 }, { "epoch": 5.192991366175724, "grad_norm": 2.4157040119171143, "learning_rate": 9.380765506361132e-06, "loss": 1.0735, "step": 2560 }, { "epoch": 5.2133062468258, "grad_norm": 2.3019464015960693, "learning_rate": 9.317021688635236e-06, "loss": 1.0746, "step": 2570 }, { "epoch": 5.233621127475876, "grad_norm": 2.6220922470092773, "learning_rate": 9.253305740342287e-06, "loss": 1.0663, "step": 2580 }, { "epoch": 5.253936008125952, "grad_norm": 2.5131311416625977, "learning_rate": 9.189620261458467e-06, "loss": 1.0616, "step": 2590 }, { "epoch": 5.274250888776028, "grad_norm": 2.5169596672058105, "learning_rate": 9.125967850716648e-06, "loss": 1.0709, "step": 2600 }, { "epoch": 5.294565769426105, "grad_norm": 2.525052309036255, "learning_rate": 9.062351105500316e-06, "loss": 1.07, "step": 2610 }, { "epoch": 5.314880650076181, "grad_norm": 2.4557714462280273, "learning_rate": 8.998772621737609e-06, "loss": 1.0802, "step": 2620 }, { "epoch": 5.335195530726257, "grad_norm": 2.4241530895233154, "learning_rate": 8.935234993795378e-06, "loss": 1.0765, "step": 2630 }, { "epoch": 5.355510411376333, "grad_norm": 2.6684749126434326, "learning_rate": 8.87174081437332e-06, "loss": 1.0793, "step": 2640 }, { "epoch": 5.375825292026409, "grad_norm": 2.0009050369262695, "learning_rate": 8.808292674398186e-06, "loss": 1.0584, "step": 2650 }, { "epoch": 5.3961401726764855, "grad_norm": 2.189467668533325, "learning_rate": 8.744893162918045e-06, "loss": 1.1002, "step": 2660 }, { "epoch": 5.416455053326562, "grad_norm": 2.2791178226470947, "learning_rate": 8.681544866996657e-06, "loss": 1.0928, "step": 2670 }, { "epoch": 5.436769933976638, "grad_norm": 2.332437753677368, "learning_rate": 8.618250371607879e-06, "loss": 1.0715, "step": 2680 }, { "epoch": 5.457084814626714, "grad_norm": 2.518786907196045, "learning_rate": 8.555012259530215e-06, "loss": 1.074, "step": 2690 }, { "epoch": 5.47739969527679, "grad_norm": 2.394573211669922, "learning_rate": 8.491833111241392e-06, "loss": 1.0846, "step": 2700 }, { "epoch": 5.497714575926866, "grad_norm": 2.2103583812713623, "learning_rate": 8.428715504813084e-06, "loss": 1.0952, "step": 2710 }, { "epoch": 5.518029456576943, "grad_norm": 2.51759934425354, "learning_rate": 8.365662015805694e-06, "loss": 1.0918, "step": 2720 }, { "epoch": 5.538344337227019, "grad_norm": 2.085672616958618, "learning_rate": 8.302675217163279e-06, "loss": 1.0726, "step": 2730 }, { "epoch": 5.558659217877095, "grad_norm": 2.341189384460449, "learning_rate": 8.23975767910854e-06, "loss": 1.0762, "step": 2740 }, { "epoch": 5.578974098527171, "grad_norm": 2.3695602416992188, "learning_rate": 8.176911969037951e-06, "loss": 1.0794, "step": 2750 }, { "epoch": 5.599288979177247, "grad_norm": 2.8153765201568604, "learning_rate": 8.11414065141699e-06, "loss": 1.0823, "step": 2760 }, { "epoch": 5.619603859827324, "grad_norm": 2.3020527362823486, "learning_rate": 8.051446287675499e-06, "loss": 1.064, "step": 2770 }, { "epoch": 5.6399187404774, "grad_norm": 2.436936378479004, "learning_rate": 7.988831436103157e-06, "loss": 1.0911, "step": 2780 }, { "epoch": 5.6602336211274755, "grad_norm": 2.706698417663574, "learning_rate": 7.926298651745095e-06, "loss": 1.1105, "step": 2790 }, { "epoch": 5.680548501777552, "grad_norm": 2.290712594985962, "learning_rate": 7.863850486297622e-06, "loss": 1.0668, "step": 2800 }, { "epoch": 5.700863382427628, "grad_norm": 2.3723671436309814, "learning_rate": 7.801489488004124e-06, "loss": 1.0701, "step": 2810 }, { "epoch": 5.721178263077705, "grad_norm": 2.1805179119110107, "learning_rate": 7.739218201551054e-06, "loss": 1.0718, "step": 2820 }, { "epoch": 5.741493143727781, "grad_norm": 2.39038348197937, "learning_rate": 7.677039167964114e-06, "loss": 1.075, "step": 2830 }, { "epoch": 5.761808024377856, "grad_norm": 2.164767265319824, "learning_rate": 7.614954924504553e-06, "loss": 1.0857, "step": 2840 }, { "epoch": 5.782122905027933, "grad_norm": 2.3338029384613037, "learning_rate": 7.552968004565653e-06, "loss": 1.0692, "step": 2850 }, { "epoch": 5.802437785678009, "grad_norm": 2.186400890350342, "learning_rate": 7.491080937569321e-06, "loss": 1.0759, "step": 2860 }, { "epoch": 5.8227526663280855, "grad_norm": 2.548473834991455, "learning_rate": 7.429296248862907e-06, "loss": 1.0799, "step": 2870 }, { "epoch": 5.843067546978162, "grad_norm": 2.4101195335388184, "learning_rate": 7.367616459616127e-06, "loss": 1.0772, "step": 2880 }, { "epoch": 5.863382427628237, "grad_norm": 2.5707526206970215, "learning_rate": 7.306044086718207e-06, "loss": 1.0772, "step": 2890 }, { "epoch": 5.883697308278314, "grad_norm": 2.447829246520996, "learning_rate": 7.244581642675159e-06, "loss": 1.0823, "step": 2900 }, { "epoch": 5.90401218892839, "grad_norm": 2.3447105884552, "learning_rate": 7.1832316355072785e-06, "loss": 1.0927, "step": 2910 }, { "epoch": 5.924327069578466, "grad_norm": 2.4636731147766113, "learning_rate": 7.121996568646786e-06, "loss": 1.0963, "step": 2920 }, { "epoch": 5.944641950228542, "grad_norm": 2.2715587615966797, "learning_rate": 7.060878940835674e-06, "loss": 1.1112, "step": 2930 }, { "epoch": 5.964956830878618, "grad_norm": 2.37288236618042, "learning_rate": 6.999881246023754e-06, "loss": 1.0959, "step": 2940 }, { "epoch": 5.985271711528695, "grad_norm": 2.28794264793396, "learning_rate": 6.9390059732668835e-06, "loss": 1.0642, "step": 2950 }, { "epoch": 6.0, "eval_loss": 1.0949872732162476, "eval_runtime": 54.7572, "eval_samples_per_second": 35.977, "eval_steps_per_second": 9.003, "step": 2958 }, { "epoch": 6.004062976130015, "grad_norm": 2.5767059326171875, "learning_rate": 6.878255606625395e-06, "loss": 1.0996, "step": 2960 }, { "epoch": 6.024377856780092, "grad_norm": 2.55753755569458, "learning_rate": 6.817632625062735e-06, "loss": 1.0868, "step": 2970 }, { "epoch": 6.044692737430168, "grad_norm": 2.5268876552581787, "learning_rate": 6.757139502344312e-06, "loss": 1.0957, "step": 2980 }, { "epoch": 6.0650076180802435, "grad_norm": 2.596231698989868, "learning_rate": 6.696778706936542e-06, "loss": 1.078, "step": 2990 }, { "epoch": 6.08532249873032, "grad_norm": 2.4920766353607178, "learning_rate": 6.636552701906135e-06, "loss": 1.0653, "step": 3000 }, { "epoch": 6.105637379380396, "grad_norm": 2.75101900100708, "learning_rate": 6.576463944819576e-06, "loss": 1.1024, "step": 3010 }, { "epoch": 6.1259522600304726, "grad_norm": 2.3427553176879883, "learning_rate": 6.516514887642846e-06, "loss": 1.0864, "step": 3020 }, { "epoch": 6.146267140680549, "grad_norm": 2.1833336353302, "learning_rate": 6.456707976641368e-06, "loss": 1.066, "step": 3030 }, { "epoch": 6.166582021330624, "grad_norm": 2.4288933277130127, "learning_rate": 6.397045652280192e-06, "loss": 1.0787, "step": 3040 }, { "epoch": 6.186896901980701, "grad_norm": 2.3094112873077393, "learning_rate": 6.3375303491243925e-06, "loss": 1.0703, "step": 3050 }, { "epoch": 6.207211782630777, "grad_norm": 2.3378853797912598, "learning_rate": 6.278164495739749e-06, "loss": 1.1107, "step": 3060 }, { "epoch": 6.2275266632808535, "grad_norm": 2.429354667663574, "learning_rate": 6.21895051459362e-06, "loss": 1.0632, "step": 3070 }, { "epoch": 6.24784154393093, "grad_norm": 2.5030083656311035, "learning_rate": 6.159890821956118e-06, "loss": 1.0794, "step": 3080 }, { "epoch": 6.268156424581005, "grad_norm": 2.2454118728637695, "learning_rate": 6.100987827801486e-06, "loss": 1.0555, "step": 3090 }, { "epoch": 6.288471305231082, "grad_norm": 2.3601973056793213, "learning_rate": 6.042243935709782e-06, "loss": 1.0758, "step": 3100 }, { "epoch": 6.308786185881158, "grad_norm": 1.9250520467758179, "learning_rate": 5.983661542768777e-06, "loss": 1.066, "step": 3110 }, { "epoch": 6.329101066531234, "grad_norm": 2.4363834857940674, "learning_rate": 5.9252430394761625e-06, "loss": 1.089, "step": 3120 }, { "epoch": 6.34941594718131, "grad_norm": 2.1766016483306885, "learning_rate": 5.866990809641978e-06, "loss": 1.0695, "step": 3130 }, { "epoch": 6.369730827831386, "grad_norm": 2.575039863586426, "learning_rate": 5.808907230291359e-06, "loss": 1.0874, "step": 3140 }, { "epoch": 6.390045708481463, "grad_norm": 2.4749789237976074, "learning_rate": 5.75099467156754e-06, "loss": 1.1004, "step": 3150 }, { "epoch": 6.410360589131539, "grad_norm": 2.669945240020752, "learning_rate": 5.6932554966351185e-06, "loss": 1.0756, "step": 3160 }, { "epoch": 6.430675469781615, "grad_norm": 2.250049114227295, "learning_rate": 5.635692061583651e-06, "loss": 1.074, "step": 3170 }, { "epoch": 6.450990350431692, "grad_norm": 2.271474599838257, "learning_rate": 5.578306715331494e-06, "loss": 1.0881, "step": 3180 }, { "epoch": 6.471305231081767, "grad_norm": 2.2663955688476562, "learning_rate": 5.521101799529965e-06, "loss": 1.0601, "step": 3190 }, { "epoch": 6.4916201117318435, "grad_norm": 2.209956407546997, "learning_rate": 5.464079648467769e-06, "loss": 1.0569, "step": 3200 }, { "epoch": 6.51193499238192, "grad_norm": 2.395031213760376, "learning_rate": 5.407242588975777e-06, "loss": 1.0687, "step": 3210 }, { "epoch": 6.532249873031996, "grad_norm": 2.570211410522461, "learning_rate": 5.3505929403320575e-06, "loss": 1.064, "step": 3220 }, { "epoch": 6.552564753682072, "grad_norm": 2.569349765777588, "learning_rate": 5.294133014167242e-06, "loss": 1.0752, "step": 3230 }, { "epoch": 6.572879634332148, "grad_norm": 2.717036008834839, "learning_rate": 5.237865114370192e-06, "loss": 1.06, "step": 3240 }, { "epoch": 6.593194514982224, "grad_norm": 2.299722671508789, "learning_rate": 5.18179153699399e-06, "loss": 1.0828, "step": 3250 }, { "epoch": 6.613509395632301, "grad_norm": 2.2542483806610107, "learning_rate": 5.125914570162259e-06, "loss": 1.0829, "step": 3260 }, { "epoch": 6.633824276282377, "grad_norm": 2.26954984664917, "learning_rate": 5.070236493975777e-06, "loss": 1.0877, "step": 3270 }, { "epoch": 6.6541391569324535, "grad_norm": 2.2126009464263916, "learning_rate": 5.014759580419448e-06, "loss": 1.0676, "step": 3280 }, { "epoch": 6.674454037582529, "grad_norm": 2.658097505569458, "learning_rate": 4.959486093269577e-06, "loss": 1.0746, "step": 3290 }, { "epoch": 6.694768918232605, "grad_norm": 2.2925167083740234, "learning_rate": 4.904418288001511e-06, "loss": 1.0689, "step": 3300 }, { "epoch": 6.715083798882682, "grad_norm": 2.2561416625976562, "learning_rate": 4.849558411697604e-06, "loss": 1.0787, "step": 3310 }, { "epoch": 6.735398679532758, "grad_norm": 2.590914726257324, "learning_rate": 4.794908702955496e-06, "loss": 1.0673, "step": 3320 }, { "epoch": 6.7557135601828335, "grad_norm": 2.577042579650879, "learning_rate": 4.740471391796807e-06, "loss": 1.0841, "step": 3330 }, { "epoch": 6.77602844083291, "grad_norm": 2.53352689743042, "learning_rate": 4.686248699576098e-06, "loss": 1.0603, "step": 3340 }, { "epoch": 6.796343321482986, "grad_norm": 2.2674384117126465, "learning_rate": 4.632242838890257e-06, "loss": 1.0651, "step": 3350 }, { "epoch": 6.816658202133063, "grad_norm": 2.618176221847534, "learning_rate": 4.578456013488206e-06, "loss": 1.0714, "step": 3360 }, { "epoch": 6.836973082783139, "grad_norm": 2.610429286956787, "learning_rate": 4.524890418180954e-06, "loss": 1.1131, "step": 3370 }, { "epoch": 6.8572879634332145, "grad_norm": 2.254110813140869, "learning_rate": 4.471548238752068e-06, "loss": 1.0697, "step": 3380 }, { "epoch": 6.877602844083291, "grad_norm": 2.650230884552002, "learning_rate": 4.4184316518684625e-06, "loss": 1.0892, "step": 3390 }, { "epoch": 6.897917724733367, "grad_norm": 2.45845890045166, "learning_rate": 4.365542824991581e-06, "loss": 1.0798, "step": 3400 }, { "epoch": 6.9182326053834435, "grad_norm": 2.395923376083374, "learning_rate": 4.312883916288947e-06, "loss": 1.0596, "step": 3410 }, { "epoch": 6.93854748603352, "grad_norm": 2.744338274002075, "learning_rate": 4.26045707454611e-06, "loss": 1.0624, "step": 3420 }, { "epoch": 6.958862366683595, "grad_norm": 2.431194543838501, "learning_rate": 4.2082644390789565e-06, "loss": 1.0762, "step": 3430 }, { "epoch": 6.979177247333672, "grad_norm": 2.344907522201538, "learning_rate": 4.156308139646417e-06, "loss": 1.0716, "step": 3440 }, { "epoch": 6.999492127983748, "grad_norm": 2.289046049118042, "learning_rate": 4.1045902963635355e-06, "loss": 1.0749, "step": 3450 }, { "epoch": 7.0, "eval_loss": 1.0920889377593994, "eval_runtime": 54.5848, "eval_samples_per_second": 36.091, "eval_steps_per_second": 9.032, "step": 3451 }, { "epoch": 7.018283392585069, "grad_norm": 2.9955337047576904, "learning_rate": 4.053113019615001e-06, "loss": 1.0685, "step": 3460 }, { "epoch": 7.038598273235145, "grad_norm": 2.43476939201355, "learning_rate": 4.001878409969e-06, "loss": 1.0741, "step": 3470 }, { "epoch": 7.058913153885221, "grad_norm": 2.4879932403564453, "learning_rate": 3.95088855809151e-06, "loss": 1.0906, "step": 3480 }, { "epoch": 7.079228034535297, "grad_norm": 2.2968192100524902, "learning_rate": 3.900145544660981e-06, "loss": 1.0646, "step": 3490 }, { "epoch": 7.099542915185373, "grad_norm": 2.6439547538757324, "learning_rate": 3.849651440283443e-06, "loss": 1.0802, "step": 3500 }, { "epoch": 7.11985779583545, "grad_norm": 2.3785431385040283, "learning_rate": 3.7994083054080123e-06, "loss": 1.067, "step": 3510 }, { "epoch": 7.140172676485526, "grad_norm": 2.7672765254974365, "learning_rate": 3.7494181902428094e-06, "loss": 1.082, "step": 3520 }, { "epoch": 7.1604875571356015, "grad_norm": 2.634470224380493, "learning_rate": 3.6996831346712892e-06, "loss": 1.0749, "step": 3530 }, { "epoch": 7.180802437785678, "grad_norm": 2.251033067703247, "learning_rate": 3.6502051681690266e-06, "loss": 1.0831, "step": 3540 }, { "epoch": 7.201117318435754, "grad_norm": 2.855700969696045, "learning_rate": 3.600986309720873e-06, "loss": 1.0765, "step": 3550 }, { "epoch": 7.221432199085831, "grad_norm": 2.3868894577026367, "learning_rate": 3.552028567738596e-06, "loss": 1.0614, "step": 3560 }, { "epoch": 7.241747079735907, "grad_norm": 2.3770251274108887, "learning_rate": 3.5033339399789125e-06, "loss": 1.0764, "step": 3570 }, { "epoch": 7.262061960385982, "grad_norm": 2.417048215866089, "learning_rate": 3.4549044134619603e-06, "loss": 1.0607, "step": 3580 }, { "epoch": 7.282376841036059, "grad_norm": 2.5006279945373535, "learning_rate": 3.4067419643902367e-06, "loss": 1.0481, "step": 3590 }, { "epoch": 7.302691721686135, "grad_norm": 2.4641923904418945, "learning_rate": 3.3588485580679465e-06, "loss": 1.0755, "step": 3600 }, { "epoch": 7.3230066023362115, "grad_norm": 2.6034882068634033, "learning_rate": 3.3112261488207962e-06, "loss": 1.068, "step": 3610 }, { "epoch": 7.343321482986288, "grad_norm": 2.483919858932495, "learning_rate": 3.263876679916267e-06, "loss": 1.0659, "step": 3620 }, { "epoch": 7.363636363636363, "grad_norm": 2.473473072052002, "learning_rate": 3.216802083484307e-06, "loss": 1.0953, "step": 3630 }, { "epoch": 7.38395124428644, "grad_norm": 2.57886004447937, "learning_rate": 3.1700042804384924e-06, "loss": 1.0537, "step": 3640 }, { "epoch": 7.404266124936516, "grad_norm": 2.4494612216949463, "learning_rate": 3.1234851803976352e-06, "loss": 1.0834, "step": 3650 }, { "epoch": 7.424581005586592, "grad_norm": 2.679572105407715, "learning_rate": 3.077246681607865e-06, "loss": 1.0538, "step": 3660 }, { "epoch": 7.444895886236669, "grad_norm": 2.955970287322998, "learning_rate": 3.03129067086518e-06, "loss": 1.0855, "step": 3670 }, { "epoch": 7.465210766886744, "grad_norm": 2.511077404022217, "learning_rate": 2.98561902343844e-06, "loss": 1.094, "step": 3680 }, { "epoch": 7.485525647536821, "grad_norm": 3.1838951110839844, "learning_rate": 2.9402336029928526e-06, "loss": 1.0813, "step": 3690 }, { "epoch": 7.505840528186897, "grad_norm": 2.397426128387451, "learning_rate": 2.8951362615139167e-06, "loss": 1.0666, "step": 3700 }, { "epoch": 7.526155408836973, "grad_norm": 2.655226230621338, "learning_rate": 2.8503288392318627e-06, "loss": 1.0586, "step": 3710 }, { "epoch": 7.546470289487049, "grad_norm": 2.5811400413513184, "learning_rate": 2.8058131645465546e-06, "loss": 1.0763, "step": 3720 }, { "epoch": 7.566785170137125, "grad_norm": 2.3530173301696777, "learning_rate": 2.7615910539528824e-06, "loss": 1.069, "step": 3730 }, { "epoch": 7.5871000507872015, "grad_norm": 2.2900733947753906, "learning_rate": 2.717664311966627e-06, "loss": 1.0621, "step": 3740 }, { "epoch": 7.607414931437278, "grad_norm": 2.464092493057251, "learning_rate": 2.6740347310508487e-06, "loss": 1.0622, "step": 3750 }, { "epoch": 7.627729812087354, "grad_norm": 2.5965607166290283, "learning_rate": 2.6307040915427196e-06, "loss": 1.0682, "step": 3760 }, { "epoch": 7.648044692737431, "grad_norm": 2.3087034225463867, "learning_rate": 2.5876741615809007e-06, "loss": 1.0653, "step": 3770 }, { "epoch": 7.668359573387506, "grad_norm": 2.2895405292510986, "learning_rate": 2.5449466970333658e-06, "loss": 1.0716, "step": 3780 }, { "epoch": 7.688674454037582, "grad_norm": 2.888216972351074, "learning_rate": 2.5025234414257717e-06, "loss": 1.0812, "step": 3790 }, { "epoch": 7.708989334687659, "grad_norm": 2.578557014465332, "learning_rate": 2.460406125870305e-06, "loss": 1.0694, "step": 3800 }, { "epoch": 7.729304215337735, "grad_norm": 2.9586007595062256, "learning_rate": 2.418596468995044e-06, "loss": 1.0819, "step": 3810 }, { "epoch": 7.749619095987811, "grad_norm": 2.469660520553589, "learning_rate": 2.3770961768738222e-06, "loss": 1.0815, "step": 3820 }, { "epoch": 7.769933976637887, "grad_norm": 2.530141592025757, "learning_rate": 2.3359069429566216e-06, "loss": 1.0876, "step": 3830 }, { "epoch": 7.790248857287963, "grad_norm": 2.477205514907837, "learning_rate": 2.295030448000466e-06, "loss": 1.0533, "step": 3840 }, { "epoch": 7.81056373793804, "grad_norm": 2.493910789489746, "learning_rate": 2.2544683600008345e-06, "loss": 1.0663, "step": 3850 }, { "epoch": 7.830878618588116, "grad_norm": 2.667234182357788, "learning_rate": 2.214222334123596e-06, "loss": 1.0802, "step": 3860 }, { "epoch": 7.851193499238192, "grad_norm": 2.600020408630371, "learning_rate": 2.17429401263747e-06, "loss": 1.1002, "step": 3870 }, { "epoch": 7.871508379888268, "grad_norm": 2.2591912746429443, "learning_rate": 2.134685024847023e-06, "loss": 1.0694, "step": 3880 }, { "epoch": 7.891823260538344, "grad_norm": 2.7257189750671387, "learning_rate": 2.0953969870261704e-06, "loss": 1.0684, "step": 3890 }, { "epoch": 7.912138141188421, "grad_norm": 2.2915289402008057, "learning_rate": 2.0564315023522218e-06, "loss": 1.0601, "step": 3900 }, { "epoch": 7.932453021838497, "grad_norm": 2.529210090637207, "learning_rate": 2.017790160840478e-06, "loss": 1.0902, "step": 3910 }, { "epoch": 7.9527679024885725, "grad_norm": 2.3367769718170166, "learning_rate": 1.9794745392793344e-06, "loss": 1.0629, "step": 3920 }, { "epoch": 7.973082783138649, "grad_norm": 2.5318503379821777, "learning_rate": 1.9414862011659462e-06, "loss": 1.0851, "step": 3930 }, { "epoch": 7.993397663788725, "grad_norm": 2.2982535362243652, "learning_rate": 1.9038266966424213e-06, "loss": 1.0725, "step": 3940 }, { "epoch": 8.0, "eval_loss": 1.0901824235916138, "eval_runtime": 55.0806, "eval_samples_per_second": 35.766, "eval_steps_per_second": 8.951, "step": 3944 }, { "epoch": 8.012188928390046, "grad_norm": 2.5969467163085938, "learning_rate": 1.8664975624325743e-06, "loss": 1.0801, "step": 3950 }, { "epoch": 8.032503809040122, "grad_norm": 2.4681787490844727, "learning_rate": 1.8295003217792206e-06, "loss": 1.0732, "step": 3960 }, { "epoch": 8.052818689690199, "grad_norm": 2.256989002227783, "learning_rate": 1.7928364843820034e-06, "loss": 1.0615, "step": 3970 }, { "epoch": 8.073133570340275, "grad_norm": 2.6549007892608643, "learning_rate": 1.756507546335814e-06, "loss": 1.0692, "step": 3980 }, { "epoch": 8.093448450990351, "grad_norm": 2.311284065246582, "learning_rate": 1.7205149900697138e-06, "loss": 1.0669, "step": 3990 }, { "epoch": 8.113763331640426, "grad_norm": 2.461761236190796, "learning_rate": 1.6848602842864726e-06, "loss": 1.0617, "step": 4000 }, { "epoch": 8.134078212290502, "grad_norm": 2.347125291824341, "learning_rate": 1.6495448839026128e-06, "loss": 1.063, "step": 4010 }, { "epoch": 8.154393092940579, "grad_norm": 2.2828891277313232, "learning_rate": 1.6145702299890576e-06, "loss": 1.1011, "step": 4020 }, { "epoch": 8.174707973590655, "grad_norm": 2.536144256591797, "learning_rate": 1.5799377497123104e-06, "loss": 1.0696, "step": 4030 }, { "epoch": 8.195022854240731, "grad_norm": 2.7368884086608887, "learning_rate": 1.545648856276234e-06, "loss": 1.0463, "step": 4040 }, { "epoch": 8.215337734890808, "grad_norm": 2.709357976913452, "learning_rate": 1.5117049488643787e-06, "loss": 1.0894, "step": 4050 }, { "epoch": 8.235652615540884, "grad_norm": 2.5076632499694824, "learning_rate": 1.4781074125828799e-06, "loss": 1.0951, "step": 4060 }, { "epoch": 8.25596749619096, "grad_norm": 2.656967878341675, "learning_rate": 1.4448576184039442e-06, "loss": 1.0818, "step": 4070 }, { "epoch": 8.276282376841037, "grad_norm": 2.170839786529541, "learning_rate": 1.4119569231099107e-06, "loss": 1.0555, "step": 4080 }, { "epoch": 8.296597257491113, "grad_norm": 2.704313278198242, "learning_rate": 1.3794066692378793e-06, "loss": 1.0575, "step": 4090 }, { "epoch": 8.316912138141188, "grad_norm": 2.512399911880493, "learning_rate": 1.347208185024932e-06, "loss": 1.067, "step": 4100 }, { "epoch": 8.337227018791264, "grad_norm": 2.5209548473358154, "learning_rate": 1.315362784353923e-06, "loss": 1.0529, "step": 4110 }, { "epoch": 8.35754189944134, "grad_norm": 2.6369845867156982, "learning_rate": 1.2838717666998802e-06, "loss": 1.0723, "step": 4120 }, { "epoch": 8.377856780091417, "grad_norm": 3.272552013397217, "learning_rate": 1.2527364170769686e-06, "loss": 1.0733, "step": 4130 }, { "epoch": 8.398171660741493, "grad_norm": 2.4088523387908936, "learning_rate": 1.2219580059860602e-06, "loss": 1.0554, "step": 4140 }, { "epoch": 8.41848654139157, "grad_norm": 2.388200044631958, "learning_rate": 1.191537789362881e-06, "loss": 1.0691, "step": 4150 }, { "epoch": 8.438801422041646, "grad_norm": 2.4707677364349365, "learning_rate": 1.1614770085267724e-06, "loss": 1.0583, "step": 4160 }, { "epoch": 8.459116302691722, "grad_norm": 2.6969428062438965, "learning_rate": 1.1317768901300375e-06, "loss": 1.0633, "step": 4170 }, { "epoch": 8.479431183341799, "grad_norm": 2.359264850616455, "learning_rate": 1.1024386461078752e-06, "loss": 1.0503, "step": 4180 }, { "epoch": 8.499746063991875, "grad_norm": 2.645862579345703, "learning_rate": 1.073463473628935e-06, "loss": 1.0953, "step": 4190 }, { "epoch": 8.52006094464195, "grad_norm": 2.5734922885894775, "learning_rate": 1.0448525550464704e-06, "loss": 1.0529, "step": 4200 }, { "epoch": 8.540375825292026, "grad_norm": 2.237166404724121, "learning_rate": 1.0166070578500842e-06, "loss": 1.0684, "step": 4210 }, { "epoch": 8.560690705942102, "grad_norm": 2.320554733276367, "learning_rate": 9.887281346180888e-07, "loss": 1.0654, "step": 4220 }, { "epoch": 8.581005586592179, "grad_norm": 2.724879503250122, "learning_rate": 9.612169229704748e-07, "loss": 1.0667, "step": 4230 }, { "epoch": 8.601320467242255, "grad_norm": 2.951643705368042, "learning_rate": 9.340745455224943e-07, "loss": 1.0892, "step": 4240 }, { "epoch": 8.621635347892331, "grad_norm": 2.4628891944885254, "learning_rate": 9.073021098388468e-07, "loss": 1.0776, "step": 4250 }, { "epoch": 8.641950228542408, "grad_norm": 2.5746138095855713, "learning_rate": 8.809007083884868e-07, "loss": 1.0644, "step": 4260 }, { "epoch": 8.662265109192484, "grad_norm": 2.6229591369628906, "learning_rate": 8.548714185000429e-07, "loss": 1.0795, "step": 4270 }, { "epoch": 8.68257998984256, "grad_norm": 2.9902350902557373, "learning_rate": 8.292153023178517e-07, "loss": 1.0791, "step": 4280 }, { "epoch": 8.702894870492635, "grad_norm": 2.228269577026367, "learning_rate": 8.039334067586325e-07, "loss": 1.0862, "step": 4290 }, { "epoch": 8.723209751142711, "grad_norm": 2.6434853076934814, "learning_rate": 7.790267634687487e-07, "loss": 1.0752, "step": 4300 }, { "epoch": 8.743524631792788, "grad_norm": 2.551692008972168, "learning_rate": 7.544963887821233e-07, "loss": 1.0675, "step": 4310 }, { "epoch": 8.763839512442864, "grad_norm": 2.4782984256744385, "learning_rate": 7.303432836787572e-07, "loss": 1.0768, "step": 4320 }, { "epoch": 8.78415439309294, "grad_norm": 2.706620931625366, "learning_rate": 7.065684337438916e-07, "loss": 1.0881, "step": 4330 }, { "epoch": 8.804469273743017, "grad_norm": 2.345564365386963, "learning_rate": 6.831728091277867e-07, "loss": 1.0771, "step": 4340 }, { "epoch": 8.824784154393093, "grad_norm": 2.543198347091675, "learning_rate": 6.601573645061321e-07, "loss": 1.049, "step": 4350 }, { "epoch": 8.84509903504317, "grad_norm": 2.7005691528320312, "learning_rate": 6.375230390410947e-07, "loss": 1.0871, "step": 4360 }, { "epoch": 8.865413915693246, "grad_norm": 2.9565064907073975, "learning_rate": 6.152707563429949e-07, "loss": 1.0691, "step": 4370 }, { "epoch": 8.885728796343322, "grad_norm": 2.2274577617645264, "learning_rate": 5.934014244326114e-07, "loss": 1.0697, "step": 4380 }, { "epoch": 8.906043676993399, "grad_norm": 2.464477062225342, "learning_rate": 5.719159357041415e-07, "loss": 1.0937, "step": 4390 }, { "epoch": 8.926358557643473, "grad_norm": 2.860454797744751, "learning_rate": 5.508151668887696e-07, "loss": 1.065, "step": 4400 }, { "epoch": 8.94667343829355, "grad_norm": 2.2630953788757324, "learning_rate": 5.300999790189088e-07, "loss": 1.0709, "step": 4410 }, { "epoch": 8.966988318943626, "grad_norm": 2.171909809112549, "learning_rate": 5.11786680609313e-07, "loss": 1.0732, "step": 4420 }, { "epoch": 8.987303199593702, "grad_norm": 2.525136947631836, "learning_rate": 4.918064123013932e-07, "loss": 1.0703, "step": 4430 }, { "epoch": 9.0, "eval_loss": 1.0895577669143677, "eval_runtime": 54.6524, "eval_samples_per_second": 36.046, "eval_steps_per_second": 9.021, "step": 4437 }, { "epoch": 9.006094464195023, "grad_norm": 2.3463656902313232, "learning_rate": 4.722141328346708e-07, "loss": 1.0563, "step": 4440 }, { "epoch": 9.0264093448451, "grad_norm": 2.3120319843292236, "learning_rate": 4.530106416865332e-07, "loss": 1.0576, "step": 4450 }, { "epoch": 9.046724225495176, "grad_norm": 2.431831121444702, "learning_rate": 4.3419672246958313e-07, "loss": 1.0689, "step": 4460 }, { "epoch": 9.067039106145252, "grad_norm": 2.210456609725952, "learning_rate": 4.1577314289964474e-07, "loss": 1.0576, "step": 4470 }, { "epoch": 9.087353986795328, "grad_norm": 2.559464931488037, "learning_rate": 3.977406547644513e-07, "loss": 1.0773, "step": 4480 }, { "epoch": 9.107668867445403, "grad_norm": 2.405012845993042, "learning_rate": 3.800999938929606e-07, "loss": 1.0851, "step": 4490 }, { "epoch": 9.12798374809548, "grad_norm": 2.8481688499450684, "learning_rate": 3.628518801253311e-07, "loss": 1.0486, "step": 4500 }, { "epoch": 9.148298628745556, "grad_norm": 2.405825138092041, "learning_rate": 3.4599701728354474e-07, "loss": 1.0776, "step": 4510 }, { "epoch": 9.168613509395632, "grad_norm": 2.5540664196014404, "learning_rate": 3.2953609314269163e-07, "loss": 1.0623, "step": 4520 }, { "epoch": 9.188928390045708, "grad_norm": 2.9651260375976562, "learning_rate": 3.1346977940290404e-07, "loss": 1.0583, "step": 4530 }, { "epoch": 9.209243270695785, "grad_norm": 2.4424266815185547, "learning_rate": 2.9779873166194686e-07, "loss": 1.0869, "step": 4540 }, { "epoch": 9.229558151345861, "grad_norm": 2.3093559741973877, "learning_rate": 2.8252358938846145e-07, "loss": 1.0667, "step": 4550 }, { "epoch": 9.249873031995937, "grad_norm": 2.28373646736145, "learning_rate": 2.6764497589587544e-07, "loss": 1.0634, "step": 4560 }, { "epoch": 9.270187912646014, "grad_norm": 2.4830288887023926, "learning_rate": 2.5316349831697175e-07, "loss": 1.0964, "step": 4570 }, { "epoch": 9.29050279329609, "grad_norm": 2.0313925743103027, "learning_rate": 2.3907974757910735e-07, "loss": 1.0625, "step": 4580 }, { "epoch": 9.310817673946165, "grad_norm": 2.334379196166992, "learning_rate": 2.253942983800994e-07, "loss": 1.0654, "step": 4590 }, { "epoch": 9.331132554596241, "grad_norm": 2.6533384323120117, "learning_rate": 2.1210770916478384e-07, "loss": 1.0769, "step": 4600 }, { "epoch": 9.351447435246317, "grad_norm": 2.715524435043335, "learning_rate": 1.9922052210221477e-07, "loss": 1.0784, "step": 4610 }, { "epoch": 9.371762315896394, "grad_norm": 2.238746404647827, "learning_rate": 1.8673326306355432e-07, "loss": 1.0694, "step": 4620 }, { "epoch": 9.39207719654647, "grad_norm": 2.3140628337860107, "learning_rate": 1.7464644160059885e-07, "loss": 1.0733, "step": 4630 }, { "epoch": 9.412392077196547, "grad_norm": 2.2057929039001465, "learning_rate": 1.6296055092500207e-07, "loss": 1.0837, "step": 4640 }, { "epoch": 9.432706957846623, "grad_norm": 2.212195873260498, "learning_rate": 1.5167606788813794e-07, "loss": 1.0748, "step": 4650 }, { "epoch": 9.4530218384967, "grad_norm": 2.6805107593536377, "learning_rate": 1.4079345296164504e-07, "loss": 1.0793, "step": 4660 }, { "epoch": 9.473336719146776, "grad_norm": 2.494554042816162, "learning_rate": 1.3031315021864054e-07, "loss": 1.0564, "step": 4670 }, { "epoch": 9.493651599796852, "grad_norm": 2.6305577754974365, "learning_rate": 1.2023558731559582e-07, "loss": 1.0756, "step": 4680 }, { "epoch": 9.513966480446927, "grad_norm": 2.3631041049957275, "learning_rate": 1.1056117547488588e-07, "loss": 1.0828, "step": 4690 }, { "epoch": 9.534281361097003, "grad_norm": 2.3476672172546387, "learning_rate": 1.0129030946801066e-07, "loss": 1.0885, "step": 4700 }, { "epoch": 9.55459624174708, "grad_norm": 2.513430595397949, "learning_rate": 9.242336759948457e-08, "loss": 1.0698, "step": 4710 }, { "epoch": 9.574911122397156, "grad_norm": 2.3962695598602295, "learning_rate": 8.396071169139986e-08, "loss": 1.082, "step": 4720 }, { "epoch": 9.595226003047232, "grad_norm": 2.5314340591430664, "learning_rate": 7.590268706866188e-08, "loss": 1.075, "step": 4730 }, { "epoch": 9.615540883697308, "grad_norm": 2.3048431873321533, "learning_rate": 6.824962254490141e-08, "loss": 1.0527, "step": 4740 }, { "epoch": 9.635855764347385, "grad_norm": 2.2387807369232178, "learning_rate": 6.100183040905095e-08, "loss": 1.0573, "step": 4750 }, { "epoch": 9.656170644997461, "grad_norm": 2.7502102851867676, "learning_rate": 5.415960641260487e-08, "loss": 1.0662, "step": 4760 }, { "epoch": 9.676485525647538, "grad_norm": 2.980741024017334, "learning_rate": 4.772322975755361e-08, "loss": 1.099, "step": 4770 }, { "epoch": 9.696800406297612, "grad_norm": 2.4962992668151855, "learning_rate": 4.169296308498494e-08, "loss": 1.065, "step": 4780 }, { "epoch": 9.717115286947688, "grad_norm": 2.1819846630096436, "learning_rate": 3.606905246437253e-08, "loss": 1.0641, "step": 4790 }, { "epoch": 9.737430167597765, "grad_norm": 2.336205244064331, "learning_rate": 3.0851727383529596e-08, "loss": 1.052, "step": 4800 }, { "epoch": 9.757745048247841, "grad_norm": 2.4870941638946533, "learning_rate": 2.604120073924965e-08, "loss": 1.0729, "step": 4810 }, { "epoch": 9.778059928897918, "grad_norm": 2.8149468898773193, "learning_rate": 2.16376688286124e-08, "loss": 1.0905, "step": 4820 }, { "epoch": 9.798374809547994, "grad_norm": 2.3567795753479004, "learning_rate": 1.7641311340982346e-08, "loss": 1.0628, "step": 4830 }, { "epoch": 9.81868969019807, "grad_norm": 2.4957797527313232, "learning_rate": 1.40522913506691e-08, "loss": 1.071, "step": 4840 }, { "epoch": 9.839004570848147, "grad_norm": 2.4365956783294678, "learning_rate": 1.0870755310276038e-08, "loss": 1.0788, "step": 4850 }, { "epoch": 9.859319451498223, "grad_norm": 2.4922902584075928, "learning_rate": 8.096833044723973e-09, "loss": 1.0908, "step": 4860 }, { "epoch": 9.8796343321483, "grad_norm": 2.4978227615356445, "learning_rate": 5.730637745954282e-09, "loss": 1.0725, "step": 4870 }, { "epoch": 9.899949212798376, "grad_norm": 2.3648996353149414, "learning_rate": 3.772265968309263e-09, "loss": 1.0646, "step": 4880 }, { "epoch": 9.92026409344845, "grad_norm": 2.4374351501464844, "learning_rate": 2.2217976245886286e-09, "loss": 1.0571, "step": 4890 }, { "epoch": 9.940578974098527, "grad_norm": 2.5737831592559814, "learning_rate": 1.0792959827987671e-09, "loss": 1.0577, "step": 4900 }, { "epoch": 9.960893854748603, "grad_norm": 2.2821946144104004, "learning_rate": 3.448076635570452e-10, "loss": 1.0729, "step": 4910 }, { "epoch": 9.98120873539868, "grad_norm": 2.4394965171813965, "learning_rate": 1.8362638203317517e-11, "loss": 1.068, "step": 4920 }, { "epoch": 9.98120873539868, "eval_loss": 1.0894742012023926, "eval_runtime": 55.0248, "eval_samples_per_second": 35.802, "eval_steps_per_second": 8.96, "step": 4920 }, { "epoch": 9.98120873539868, "step": 4920, "total_flos": 2.0218089447773184e+16, "train_loss": 1.1380672057469685, "train_runtime": 8182.6752, "train_samples_per_second": 9.625, "train_steps_per_second": 0.601 } ], "logging_steps": 10, "max_steps": 4920, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.0218089447773184e+16, "train_batch_size": 4, "trial_name": null, "trial_params": null }