{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.99979354483518, "eval_steps": 200, "global_step": 3632, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00027527355309338655, "grad_norm": 47.970261255811366, "learning_rate": 2.7472527472527476e-08, "loss": 2.9286, "step": 1 }, { "epoch": 0.0013763677654669328, "grad_norm": 197.0659220078, "learning_rate": 1.3736263736263737e-07, "loss": 3.3156, "step": 5 }, { "epoch": 0.0027527355309338655, "grad_norm": 155.2306778030855, "learning_rate": 2.7472527472527475e-07, "loss": 3.1021, "step": 10 }, { "epoch": 0.004129103296400798, "grad_norm": 40.67122345119242, "learning_rate": 4.120879120879121e-07, "loss": 2.9314, "step": 15 }, { "epoch": 0.005505471061867731, "grad_norm": 59.59101654287074, "learning_rate": 5.494505494505495e-07, "loss": 3.0831, "step": 20 }, { "epoch": 0.006881838827334664, "grad_norm": 105.45909774967997, "learning_rate": 6.868131868131869e-07, "loss": 3.0525, "step": 25 }, { "epoch": 0.008258206592801597, "grad_norm": 34.27227049204782, "learning_rate": 8.241758241758242e-07, "loss": 3.1182, "step": 30 }, { "epoch": 0.00963457435826853, "grad_norm": 127.02820304897055, "learning_rate": 9.615384615384617e-07, "loss": 2.9352, "step": 35 }, { "epoch": 0.011010942123735462, "grad_norm": 64.04507215334816, "learning_rate": 1.098901098901099e-06, "loss": 2.9873, "step": 40 }, { "epoch": 0.012387309889202395, "grad_norm": 23.999491442433914, "learning_rate": 1.2362637362637365e-06, "loss": 2.667, "step": 45 }, { "epoch": 0.013763677654669328, "grad_norm": 34.50645980021284, "learning_rate": 1.3736263736263738e-06, "loss": 2.6091, "step": 50 }, { "epoch": 0.01514004542013626, "grad_norm": 46.64865828445111, "learning_rate": 1.510989010989011e-06, "loss": 2.4922, "step": 55 }, { "epoch": 0.016516413185603193, "grad_norm": 21.789171957916576, "learning_rate": 1.6483516483516484e-06, "loss": 2.4031, "step": 60 }, { "epoch": 0.017892780951070126, "grad_norm": 38.24383608079909, "learning_rate": 1.7857142857142859e-06, "loss": 2.3918, "step": 65 }, { "epoch": 0.01926914871653706, "grad_norm": 15.503348119198089, "learning_rate": 1.9230769230769234e-06, "loss": 2.2549, "step": 70 }, { "epoch": 0.02064551648200399, "grad_norm": 12.557848365372337, "learning_rate": 2.0604395604395607e-06, "loss": 2.254, "step": 75 }, { "epoch": 0.022021884247470924, "grad_norm": 13.411610689976131, "learning_rate": 2.197802197802198e-06, "loss": 2.2895, "step": 80 }, { "epoch": 0.023398252012937857, "grad_norm": 11.512956200014314, "learning_rate": 2.3351648351648353e-06, "loss": 1.9875, "step": 85 }, { "epoch": 0.02477461977840479, "grad_norm": 10.834444585244098, "learning_rate": 2.472527472527473e-06, "loss": 1.9547, "step": 90 }, { "epoch": 0.026150987543871723, "grad_norm": 10.311358732225472, "learning_rate": 2.6098901098901103e-06, "loss": 1.9957, "step": 95 }, { "epoch": 0.027527355309338655, "grad_norm": 8.779165691485517, "learning_rate": 2.7472527472527476e-06, "loss": 1.8935, "step": 100 }, { "epoch": 0.028903723074805588, "grad_norm": 7.066737844316085, "learning_rate": 2.8846153846153845e-06, "loss": 1.7438, "step": 105 }, { "epoch": 0.03028009084027252, "grad_norm": 7.678771227526979, "learning_rate": 3.021978021978022e-06, "loss": 1.6582, "step": 110 }, { "epoch": 0.03165645860573946, "grad_norm": 6.076422896577125, "learning_rate": 3.1593406593406595e-06, "loss": 1.6652, "step": 115 }, { "epoch": 0.033032826371206386, "grad_norm": 6.807996481097765, "learning_rate": 3.2967032967032968e-06, "loss": 1.6854, "step": 120 }, { "epoch": 0.03440919413667332, "grad_norm": 5.621259165131704, "learning_rate": 3.4340659340659345e-06, "loss": 1.6461, "step": 125 }, { "epoch": 0.03578556190214025, "grad_norm": 4.18571348105954, "learning_rate": 3.5714285714285718e-06, "loss": 1.5098, "step": 130 }, { "epoch": 0.03716192966760719, "grad_norm": 4.80488403899634, "learning_rate": 3.708791208791209e-06, "loss": 1.5253, "step": 135 }, { "epoch": 0.03853829743307412, "grad_norm": 5.012572093969651, "learning_rate": 3.846153846153847e-06, "loss": 1.5984, "step": 140 }, { "epoch": 0.039914665198541054, "grad_norm": 4.552751574433519, "learning_rate": 3.983516483516483e-06, "loss": 1.5606, "step": 145 }, { "epoch": 0.04129103296400798, "grad_norm": 4.3599791639751535, "learning_rate": 4.120879120879121e-06, "loss": 1.4992, "step": 150 }, { "epoch": 0.04266740072947492, "grad_norm": 3.488042700729171, "learning_rate": 4.258241758241759e-06, "loss": 1.4261, "step": 155 }, { "epoch": 0.04404376849494185, "grad_norm": 3.8333617793891577, "learning_rate": 4.395604395604396e-06, "loss": 1.4617, "step": 160 }, { "epoch": 0.045420136260408785, "grad_norm": 3.0548962932241848, "learning_rate": 4.532967032967033e-06, "loss": 1.4781, "step": 165 }, { "epoch": 0.046796504025875714, "grad_norm": 5.5346794342223, "learning_rate": 4.6703296703296706e-06, "loss": 1.4395, "step": 170 }, { "epoch": 0.04817287179134265, "grad_norm": 3.685964787103759, "learning_rate": 4.807692307692308e-06, "loss": 1.3802, "step": 175 }, { "epoch": 0.04954923955680958, "grad_norm": 4.058757049675143, "learning_rate": 4.945054945054946e-06, "loss": 1.4061, "step": 180 }, { "epoch": 0.050925607322276516, "grad_norm": 2.9156243613759965, "learning_rate": 4.999990671457219e-06, "loss": 1.3884, "step": 185 }, { "epoch": 0.052301975087743445, "grad_norm": 2.8044219675737225, "learning_rate": 4.999933663947887e-06, "loss": 1.3079, "step": 190 }, { "epoch": 0.05367834285321038, "grad_norm": 2.5580737825571354, "learning_rate": 4.999824832633327e-06, "loss": 1.2732, "step": 195 }, { "epoch": 0.05505471061867731, "grad_norm": 2.5474355608388106, "learning_rate": 4.999664179769621e-06, "loss": 1.3348, "step": 200 }, { "epoch": 0.05505471061867731, "eval_loss": 1.2703925371170044, "eval_runtime": 37.5835, "eval_samples_per_second": 133.037, "eval_steps_per_second": 2.102, "step": 200 }, { "epoch": 0.05643107838414425, "grad_norm": 2.5400210842014985, "learning_rate": 4.999451708687114e-06, "loss": 1.2628, "step": 205 }, { "epoch": 0.057807446149611176, "grad_norm": 2.9134202135126803, "learning_rate": 4.999187423790347e-06, "loss": 1.3461, "step": 210 }, { "epoch": 0.05918381391507811, "grad_norm": 2.979982911881612, "learning_rate": 4.9988713305579665e-06, "loss": 1.279, "step": 215 }, { "epoch": 0.06056018168054504, "grad_norm": 2.3341235353360754, "learning_rate": 4.998503435542605e-06, "loss": 1.2791, "step": 220 }, { "epoch": 0.06193654944601198, "grad_norm": 2.553595077194149, "learning_rate": 4.9980837463707545e-06, "loss": 1.2591, "step": 225 }, { "epoch": 0.06331291721147891, "grad_norm": 2.1559374955549675, "learning_rate": 4.997612271742601e-06, "loss": 1.2184, "step": 230 }, { "epoch": 0.06468928497694584, "grad_norm": 2.143686811503016, "learning_rate": 4.9970890214318494e-06, "loss": 1.2676, "step": 235 }, { "epoch": 0.06606565274241277, "grad_norm": 1.75568515745194, "learning_rate": 4.996514006285514e-06, "loss": 1.2013, "step": 240 }, { "epoch": 0.0674420205078797, "grad_norm": 2.4736934108040027, "learning_rate": 4.995887238223703e-06, "loss": 1.2801, "step": 245 }, { "epoch": 0.06881838827334665, "grad_norm": 1.9847695711171895, "learning_rate": 4.99520873023936e-06, "loss": 1.2026, "step": 250 }, { "epoch": 0.07019475603881357, "grad_norm": 2.3258055796641592, "learning_rate": 4.994478496398007e-06, "loss": 1.2115, "step": 255 }, { "epoch": 0.0715711238042805, "grad_norm": 1.8207346260417532, "learning_rate": 4.993696551837444e-06, "loss": 1.2521, "step": 260 }, { "epoch": 0.07294749156974743, "grad_norm": 1.901002680506845, "learning_rate": 4.9928629127674375e-06, "loss": 1.1437, "step": 265 }, { "epoch": 0.07432385933521438, "grad_norm": 2.1411204432050925, "learning_rate": 4.991977596469385e-06, "loss": 1.1638, "step": 270 }, { "epoch": 0.0757002271006813, "grad_norm": 1.6695878470807146, "learning_rate": 4.991040621295959e-06, "loss": 1.1406, "step": 275 }, { "epoch": 0.07707659486614823, "grad_norm": 1.9670297948191877, "learning_rate": 4.990052006670722e-06, "loss": 1.1152, "step": 280 }, { "epoch": 0.07845296263161516, "grad_norm": 1.8164273632468562, "learning_rate": 4.989011773087725e-06, "loss": 1.154, "step": 285 }, { "epoch": 0.07982933039708211, "grad_norm": 1.4090788234057618, "learning_rate": 4.9879199421110865e-06, "loss": 1.0789, "step": 290 }, { "epoch": 0.08120569816254904, "grad_norm": 1.5318431537024289, "learning_rate": 4.9867765363745426e-06, "loss": 1.156, "step": 295 }, { "epoch": 0.08258206592801597, "grad_norm": 1.5662411802833258, "learning_rate": 4.9855815795809735e-06, "loss": 1.2253, "step": 300 }, { "epoch": 0.0839584336934829, "grad_norm": 1.487588263963499, "learning_rate": 4.984335096501922e-06, "loss": 1.1697, "step": 305 }, { "epoch": 0.08533480145894984, "grad_norm": 1.5795690517983507, "learning_rate": 4.983037112977072e-06, "loss": 1.1747, "step": 310 }, { "epoch": 0.08671116922441677, "grad_norm": 1.5801325920576266, "learning_rate": 4.981687655913716e-06, "loss": 1.1812, "step": 315 }, { "epoch": 0.0880875369898837, "grad_norm": 1.5881269944242014, "learning_rate": 4.980286753286196e-06, "loss": 1.1158, "step": 320 }, { "epoch": 0.08946390475535063, "grad_norm": 1.3649187456591194, "learning_rate": 4.978834434135323e-06, "loss": 1.0911, "step": 325 }, { "epoch": 0.09084027252081757, "grad_norm": 1.3003473744310392, "learning_rate": 4.977330728567778e-06, "loss": 1.0947, "step": 330 }, { "epoch": 0.0922166402862845, "grad_norm": 2.0679161756500832, "learning_rate": 4.975775667755489e-06, "loss": 1.1364, "step": 335 }, { "epoch": 0.09359300805175143, "grad_norm": 1.5739073785267514, "learning_rate": 4.974169283934976e-06, "loss": 1.172, "step": 340 }, { "epoch": 0.09496937581721836, "grad_norm": 1.7213917188161827, "learning_rate": 4.972511610406693e-06, "loss": 1.1608, "step": 345 }, { "epoch": 0.0963457435826853, "grad_norm": 1.4537784803209515, "learning_rate": 4.970802681534331e-06, "loss": 1.1647, "step": 350 }, { "epoch": 0.09772211134815223, "grad_norm": 1.3731655227058528, "learning_rate": 4.969042532744109e-06, "loss": 1.0853, "step": 355 }, { "epoch": 0.09909847911361916, "grad_norm": 1.2363319744461196, "learning_rate": 4.967231200524037e-06, "loss": 1.0066, "step": 360 }, { "epoch": 0.10047484687908609, "grad_norm": 1.5759080996624, "learning_rate": 4.965368722423166e-06, "loss": 1.1516, "step": 365 }, { "epoch": 0.10185121464455303, "grad_norm": 1.3507882575141734, "learning_rate": 4.9634551370507985e-06, "loss": 1.1073, "step": 370 }, { "epoch": 0.10322758241001996, "grad_norm": 1.277730572175811, "learning_rate": 4.961490484075698e-06, "loss": 1.1298, "step": 375 }, { "epoch": 0.10460395017548689, "grad_norm": 1.4535170267483315, "learning_rate": 4.9594748042252635e-06, "loss": 1.1084, "step": 380 }, { "epoch": 0.10598031794095382, "grad_norm": 1.6186734387061241, "learning_rate": 4.957408139284682e-06, "loss": 1.1102, "step": 385 }, { "epoch": 0.10735668570642076, "grad_norm": 1.132537102297945, "learning_rate": 4.9552905320960685e-06, "loss": 1.065, "step": 390 }, { "epoch": 0.10873305347188769, "grad_norm": 1.3026196221634236, "learning_rate": 4.9531220265575714e-06, "loss": 1.1021, "step": 395 }, { "epoch": 0.11010942123735462, "grad_norm": 1.3077867242379508, "learning_rate": 4.950902667622468e-06, "loss": 1.0411, "step": 400 }, { "epoch": 0.11010942123735462, "eval_loss": 1.0434951782226562, "eval_runtime": 37.5694, "eval_samples_per_second": 133.087, "eval_steps_per_second": 2.103, "step": 400 }, { "epoch": 0.11148578900282155, "grad_norm": 1.6148360944145506, "learning_rate": 4.948632501298228e-06, "loss": 1.0545, "step": 405 }, { "epoch": 0.1128621567682885, "grad_norm": 1.316099506488652, "learning_rate": 4.9463115746455656e-06, "loss": 1.0593, "step": 410 }, { "epoch": 0.11423852453375542, "grad_norm": 1.4454673179124138, "learning_rate": 4.943939935777455e-06, "loss": 1.0217, "step": 415 }, { "epoch": 0.11561489229922235, "grad_norm": 1.2512294045462884, "learning_rate": 4.941517633858142e-06, "loss": 1.1085, "step": 420 }, { "epoch": 0.11699126006468928, "grad_norm": 1.0101964296414214, "learning_rate": 4.93904471910212e-06, "loss": 1.0003, "step": 425 }, { "epoch": 0.11836762783015622, "grad_norm": 1.1944857398005146, "learning_rate": 4.936521242773091e-06, "loss": 1.1296, "step": 430 }, { "epoch": 0.11974399559562315, "grad_norm": 1.2974835007262053, "learning_rate": 4.933947257182901e-06, "loss": 1.1402, "step": 435 }, { "epoch": 0.12112036336109008, "grad_norm": 1.6118257673114658, "learning_rate": 4.931322815690457e-06, "loss": 1.0763, "step": 440 }, { "epoch": 0.12249673112655701, "grad_norm": 1.224531658843626, "learning_rate": 4.92864797270062e-06, "loss": 1.1138, "step": 445 }, { "epoch": 0.12387309889202396, "grad_norm": 1.1659984082065094, "learning_rate": 4.925922783663079e-06, "loss": 1.0189, "step": 450 }, { "epoch": 0.12524946665749087, "grad_norm": 8.709307393389583, "learning_rate": 4.923147305071199e-06, "loss": 1.0822, "step": 455 }, { "epoch": 0.12662583442295783, "grad_norm": 1.1509719961863358, "learning_rate": 4.9203215944608515e-06, "loss": 1.0373, "step": 460 }, { "epoch": 0.12800220218842476, "grad_norm": 1.0234613424480405, "learning_rate": 4.917445710409221e-06, "loss": 1.037, "step": 465 }, { "epoch": 0.1293785699538917, "grad_norm": 1.1018659479000739, "learning_rate": 4.914519712533592e-06, "loss": 1.088, "step": 470 }, { "epoch": 0.13075493771935862, "grad_norm": 0.9721625793315171, "learning_rate": 4.911543661490111e-06, "loss": 1.05, "step": 475 }, { "epoch": 0.13213130548482555, "grad_norm": 0.9710335336779444, "learning_rate": 4.908517618972532e-06, "loss": 1.0123, "step": 480 }, { "epoch": 0.13350767325029247, "grad_norm": 1.5166549245825005, "learning_rate": 4.905441647710932e-06, "loss": 1.05, "step": 485 }, { "epoch": 0.1348840410157594, "grad_norm": 0.8785700675801744, "learning_rate": 4.90231581147042e-06, "loss": 1.032, "step": 490 }, { "epoch": 0.13626040878122633, "grad_norm": 1.5060433665040545, "learning_rate": 4.899140175049806e-06, "loss": 1.0196, "step": 495 }, { "epoch": 0.1376367765466933, "grad_norm": 1.0487460758379659, "learning_rate": 4.895914804280262e-06, "loss": 1.089, "step": 500 }, { "epoch": 0.13901314431216022, "grad_norm": 1.3023717508397745, "learning_rate": 4.892639766023957e-06, "loss": 1.022, "step": 505 }, { "epoch": 0.14038951207762715, "grad_norm": 1.0972179140074843, "learning_rate": 4.889315128172669e-06, "loss": 1.0049, "step": 510 }, { "epoch": 0.14176587984309408, "grad_norm": 0.8796434076141977, "learning_rate": 4.885940959646383e-06, "loss": 0.9685, "step": 515 }, { "epoch": 0.143142247608561, "grad_norm": 0.9834949701746455, "learning_rate": 4.882517330391854e-06, "loss": 1.0246, "step": 520 }, { "epoch": 0.14451861537402794, "grad_norm": 1.138039557393203, "learning_rate": 4.879044311381164e-06, "loss": 1.1077, "step": 525 }, { "epoch": 0.14589498313949487, "grad_norm": 1.085162105922894, "learning_rate": 4.875521974610247e-06, "loss": 1.0675, "step": 530 }, { "epoch": 0.1472713509049618, "grad_norm": 0.871775704668988, "learning_rate": 4.8719503930973995e-06, "loss": 1.019, "step": 535 }, { "epoch": 0.14864771867042875, "grad_norm": 0.9961249615759942, "learning_rate": 4.868329640881764e-06, "loss": 1.0749, "step": 540 }, { "epoch": 0.15002408643589568, "grad_norm": 1.1040648660006491, "learning_rate": 4.864659793021795e-06, "loss": 1.0435, "step": 545 }, { "epoch": 0.1514004542013626, "grad_norm": 0.9631823018123004, "learning_rate": 4.860940925593703e-06, "loss": 0.9509, "step": 550 }, { "epoch": 0.15277682196682954, "grad_norm": 1.0645503004444963, "learning_rate": 4.8571731156898785e-06, "loss": 1.0142, "step": 555 }, { "epoch": 0.15415318973229647, "grad_norm": 1.0599242806058304, "learning_rate": 4.8533564414172915e-06, "loss": 0.9796, "step": 560 }, { "epoch": 0.1555295574977634, "grad_norm": 1.0803032824855876, "learning_rate": 4.849490981895877e-06, "loss": 0.9508, "step": 565 }, { "epoch": 0.15690592526323033, "grad_norm": 0.8429127532715366, "learning_rate": 4.845576817256888e-06, "loss": 0.9975, "step": 570 }, { "epoch": 0.15828229302869726, "grad_norm": 0.8648413729480695, "learning_rate": 4.841614028641241e-06, "loss": 1.0446, "step": 575 }, { "epoch": 0.15965866079416421, "grad_norm": 0.766164996990341, "learning_rate": 4.83760269819783e-06, "loss": 1.0017, "step": 580 }, { "epoch": 0.16103502855963114, "grad_norm": 1.1710332971275654, "learning_rate": 4.833542909081824e-06, "loss": 0.977, "step": 585 }, { "epoch": 0.16241139632509807, "grad_norm": 0.8805076425003365, "learning_rate": 4.829434745452944e-06, "loss": 1.0282, "step": 590 }, { "epoch": 0.163787764090565, "grad_norm": 0.9917312981640594, "learning_rate": 4.82527829247372e-06, "loss": 1.0007, "step": 595 }, { "epoch": 0.16516413185603193, "grad_norm": 0.9418229318080311, "learning_rate": 4.821073636307719e-06, "loss": 1.0483, "step": 600 }, { "epoch": 0.16516413185603193, "eval_loss": 0.9694015383720398, "eval_runtime": 37.5636, "eval_samples_per_second": 133.108, "eval_steps_per_second": 2.103, "step": 600 }, { "epoch": 0.16654049962149886, "grad_norm": 1.151675240339901, "learning_rate": 4.81682086411777e-06, "loss": 1.0735, "step": 605 }, { "epoch": 0.1679168673869658, "grad_norm": 0.8785348232118754, "learning_rate": 4.812520064064146e-06, "loss": 0.9803, "step": 610 }, { "epoch": 0.16929323515243272, "grad_norm": 0.7903326312503011, "learning_rate": 4.8081713253027415e-06, "loss": 1.0074, "step": 615 }, { "epoch": 0.17066960291789968, "grad_norm": 1.1239070075461262, "learning_rate": 4.803774737983226e-06, "loss": 0.9774, "step": 620 }, { "epoch": 0.1720459706833666, "grad_norm": 0.8575748601821118, "learning_rate": 4.799330393247173e-06, "loss": 0.9554, "step": 625 }, { "epoch": 0.17342233844883354, "grad_norm": 0.7681124502085931, "learning_rate": 4.7948383832261665e-06, "loss": 0.9925, "step": 630 }, { "epoch": 0.17479870621430046, "grad_norm": 0.9533830559317458, "learning_rate": 4.790298801039901e-06, "loss": 0.9942, "step": 635 }, { "epoch": 0.1761750739797674, "grad_norm": 0.9451172948449857, "learning_rate": 4.785711740794241e-06, "loss": 1.0296, "step": 640 }, { "epoch": 0.17755144174523432, "grad_norm": 0.9522113767917231, "learning_rate": 4.781077297579278e-06, "loss": 0.9792, "step": 645 }, { "epoch": 0.17892780951070125, "grad_norm": 0.671435197412033, "learning_rate": 4.776395567467353e-06, "loss": 0.967, "step": 650 }, { "epoch": 0.18030417727616818, "grad_norm": 0.775668527281842, "learning_rate": 4.7716666475110686e-06, "loss": 1.0187, "step": 655 }, { "epoch": 0.18168054504163514, "grad_norm": 0.8387050001267089, "learning_rate": 4.766890635741278e-06, "loss": 1.0319, "step": 660 }, { "epoch": 0.18305691280710207, "grad_norm": 0.824748710334662, "learning_rate": 4.762067631165049e-06, "loss": 0.9293, "step": 665 }, { "epoch": 0.184433280572569, "grad_norm": 0.9799352793752129, "learning_rate": 4.757197733763615e-06, "loss": 0.9157, "step": 670 }, { "epoch": 0.18580964833803593, "grad_norm": 0.8362870057769561, "learning_rate": 4.7522810444903004e-06, "loss": 0.949, "step": 675 }, { "epoch": 0.18718601610350286, "grad_norm": 0.9321760596367203, "learning_rate": 4.7473176652684276e-06, "loss": 0.9901, "step": 680 }, { "epoch": 0.18856238386896979, "grad_norm": 0.7717121636862594, "learning_rate": 4.742307698989207e-06, "loss": 1.0114, "step": 685 }, { "epoch": 0.18993875163443671, "grad_norm": 1.0300652464671185, "learning_rate": 4.7372512495096005e-06, "loss": 1.0247, "step": 690 }, { "epoch": 0.19131511939990364, "grad_norm": 0.7755223484828447, "learning_rate": 4.732148421650171e-06, "loss": 0.9337, "step": 695 }, { "epoch": 0.1926914871653706, "grad_norm": 1.0381026277362817, "learning_rate": 4.7269993211929086e-06, "loss": 0.9709, "step": 700 }, { "epoch": 0.19406785493083753, "grad_norm": 0.7702377174214647, "learning_rate": 4.721804054879036e-06, "loss": 0.9726, "step": 705 }, { "epoch": 0.19544422269630446, "grad_norm": 0.9053621913087226, "learning_rate": 4.7165627304068e-06, "loss": 0.953, "step": 710 }, { "epoch": 0.1968205904617714, "grad_norm": 1.003117563447582, "learning_rate": 4.711275456429235e-06, "loss": 0.9849, "step": 715 }, { "epoch": 0.19819695822723832, "grad_norm": 0.9099542081572461, "learning_rate": 4.70594234255191e-06, "loss": 0.9901, "step": 720 }, { "epoch": 0.19957332599270525, "grad_norm": 1.0432447959538196, "learning_rate": 4.700563499330664e-06, "loss": 0.9535, "step": 725 }, { "epoch": 0.20094969375817218, "grad_norm": 0.8237571322857683, "learning_rate": 4.695139038269303e-06, "loss": 0.9535, "step": 730 }, { "epoch": 0.2023260615236391, "grad_norm": 0.9877142384272304, "learning_rate": 4.689669071817296e-06, "loss": 0.9509, "step": 735 }, { "epoch": 0.20370242928910606, "grad_norm": 0.681443730837864, "learning_rate": 4.684153713367442e-06, "loss": 0.917, "step": 740 }, { "epoch": 0.205078797054573, "grad_norm": 1.0414452790533368, "learning_rate": 4.678593077253521e-06, "loss": 0.9662, "step": 745 }, { "epoch": 0.20645516482003992, "grad_norm": 0.7632284159752581, "learning_rate": 4.672987278747919e-06, "loss": 0.9588, "step": 750 }, { "epoch": 0.20783153258550685, "grad_norm": 0.7404164880344434, "learning_rate": 4.667336434059246e-06, "loss": 0.9426, "step": 755 }, { "epoch": 0.20920790035097378, "grad_norm": 0.7658934106898954, "learning_rate": 4.661640660329918e-06, "loss": 0.9787, "step": 760 }, { "epoch": 0.2105842681164407, "grad_norm": 0.7794575989249981, "learning_rate": 4.655900075633736e-06, "loss": 0.9341, "step": 765 }, { "epoch": 0.21196063588190764, "grad_norm": 0.681238788669416, "learning_rate": 4.650114798973434e-06, "loss": 0.9734, "step": 770 }, { "epoch": 0.21333700364737457, "grad_norm": 1.0301580037782345, "learning_rate": 4.644284950278217e-06, "loss": 0.9438, "step": 775 }, { "epoch": 0.21471337141284152, "grad_norm": 1.3078635807263586, "learning_rate": 4.638410650401267e-06, "loss": 0.9335, "step": 780 }, { "epoch": 0.21608973917830845, "grad_norm": 0.730305470715918, "learning_rate": 4.632492021117245e-06, "loss": 0.9164, "step": 785 }, { "epoch": 0.21746610694377538, "grad_norm": 1.1199081512447784, "learning_rate": 4.626529185119763e-06, "loss": 0.9451, "step": 790 }, { "epoch": 0.2188424747092423, "grad_norm": 1.0679579821267784, "learning_rate": 4.620522266018841e-06, "loss": 0.9914, "step": 795 }, { "epoch": 0.22021884247470924, "grad_norm": 0.6094802947859647, "learning_rate": 4.614471388338346e-06, "loss": 0.8801, "step": 800 }, { "epoch": 0.22021884247470924, "eval_loss": 0.9227399230003357, "eval_runtime": 37.5694, "eval_samples_per_second": 133.087, "eval_steps_per_second": 2.103, "step": 800 }, { "epoch": 0.22159521024017617, "grad_norm": 0.8418029822067341, "learning_rate": 4.60837667751341e-06, "loss": 0.8924, "step": 805 }, { "epoch": 0.2229715780056431, "grad_norm": 0.822305082880161, "learning_rate": 4.602238259887825e-06, "loss": 0.9395, "step": 810 }, { "epoch": 0.22434794577111003, "grad_norm": 0.6380622331066487, "learning_rate": 4.596056262711434e-06, "loss": 0.9366, "step": 815 }, { "epoch": 0.225724313536577, "grad_norm": 0.816860005626295, "learning_rate": 4.5898308141374835e-06, "loss": 0.9472, "step": 820 }, { "epoch": 0.22710068130204392, "grad_norm": 0.6422659239711008, "learning_rate": 4.583562043219972e-06, "loss": 0.9558, "step": 825 }, { "epoch": 0.22847704906751085, "grad_norm": 0.7575538911525173, "learning_rate": 4.577250079910973e-06, "loss": 0.933, "step": 830 }, { "epoch": 0.22985341683297777, "grad_norm": 0.72595045850496, "learning_rate": 4.57089505505794e-06, "loss": 0.9754, "step": 835 }, { "epoch": 0.2312297845984447, "grad_norm": 0.9692773154744749, "learning_rate": 4.564497100400998e-06, "loss": 0.9833, "step": 840 }, { "epoch": 0.23260615236391163, "grad_norm": 0.8045647953099883, "learning_rate": 4.558056348570209e-06, "loss": 0.918, "step": 845 }, { "epoch": 0.23398252012937856, "grad_norm": 0.8493707730777622, "learning_rate": 4.551572933082823e-06, "loss": 0.9389, "step": 850 }, { "epoch": 0.2353588878948455, "grad_norm": 0.7008145118663581, "learning_rate": 4.545046988340509e-06, "loss": 0.8909, "step": 855 }, { "epoch": 0.23673525566031245, "grad_norm": 0.9052644910175894, "learning_rate": 4.538478649626575e-06, "loss": 0.9574, "step": 860 }, { "epoch": 0.23811162342577938, "grad_norm": 0.785016020446548, "learning_rate": 4.531868053103153e-06, "loss": 1.0396, "step": 865 }, { "epoch": 0.2394879911912463, "grad_norm": 0.8758044752350663, "learning_rate": 4.52521533580839e-06, "loss": 0.8471, "step": 870 }, { "epoch": 0.24086435895671324, "grad_norm": 0.7552178534994997, "learning_rate": 4.518520635653594e-06, "loss": 0.973, "step": 875 }, { "epoch": 0.24224072672218017, "grad_norm": 0.6477878269303148, "learning_rate": 4.5117840914203805e-06, "loss": 0.93, "step": 880 }, { "epoch": 0.2436170944876471, "grad_norm": 0.7740483767304198, "learning_rate": 4.5050058427578e-06, "loss": 0.8919, "step": 885 }, { "epoch": 0.24499346225311402, "grad_norm": 0.5175079894302275, "learning_rate": 4.498186030179434e-06, "loss": 0.9334, "step": 890 }, { "epoch": 0.24636983001858095, "grad_norm": 0.7470864740423165, "learning_rate": 4.491324795060491e-06, "loss": 0.9059, "step": 895 }, { "epoch": 0.2477461977840479, "grad_norm": 0.5912054028857261, "learning_rate": 4.4844222796348705e-06, "loss": 0.9406, "step": 900 }, { "epoch": 0.24912256554951484, "grad_norm": 0.7632257634965951, "learning_rate": 4.477478626992214e-06, "loss": 0.9365, "step": 905 }, { "epoch": 0.25049893331498174, "grad_norm": 0.938994254636935, "learning_rate": 4.47049398107494e-06, "loss": 0.8971, "step": 910 }, { "epoch": 0.2518753010804487, "grad_norm": 0.5444477069055039, "learning_rate": 4.4634684866752665e-06, "loss": 0.9098, "step": 915 }, { "epoch": 0.25325166884591566, "grad_norm": 0.9370323101958166, "learning_rate": 4.456402289432196e-06, "loss": 0.988, "step": 920 }, { "epoch": 0.25462803661138256, "grad_norm": 0.656323325746701, "learning_rate": 4.44929553582851e-06, "loss": 0.9647, "step": 925 }, { "epoch": 0.2560044043768495, "grad_norm": 0.8566812639937329, "learning_rate": 4.442148373187722e-06, "loss": 0.9587, "step": 930 }, { "epoch": 0.2573807721423164, "grad_norm": 0.623784946912734, "learning_rate": 4.434960949671028e-06, "loss": 0.8996, "step": 935 }, { "epoch": 0.2587571399077834, "grad_norm": 0.556551458145588, "learning_rate": 4.427733414274238e-06, "loss": 0.8582, "step": 940 }, { "epoch": 0.2601335076732503, "grad_norm": 0.6165188495979379, "learning_rate": 4.420465916824681e-06, "loss": 0.9263, "step": 945 }, { "epoch": 0.26150987543871723, "grad_norm": 0.5388586078056022, "learning_rate": 4.413158607978104e-06, "loss": 0.8803, "step": 950 }, { "epoch": 0.26288624320418413, "grad_norm": 0.6313491960276545, "learning_rate": 4.405811639215547e-06, "loss": 0.9321, "step": 955 }, { "epoch": 0.2642626109696511, "grad_norm": 0.7175326284846463, "learning_rate": 4.398425162840202e-06, "loss": 0.921, "step": 960 }, { "epoch": 0.26563897873511805, "grad_norm": 0.7423069055487866, "learning_rate": 4.390999331974257e-06, "loss": 0.9461, "step": 965 }, { "epoch": 0.26701534650058495, "grad_norm": 0.8331321575994248, "learning_rate": 4.383534300555722e-06, "loss": 0.962, "step": 970 }, { "epoch": 0.2683917142660519, "grad_norm": 0.7921869343480575, "learning_rate": 4.376030223335237e-06, "loss": 0.8739, "step": 975 }, { "epoch": 0.2697680820315188, "grad_norm": 0.8214762995889284, "learning_rate": 4.368487255872864e-06, "loss": 0.9187, "step": 980 }, { "epoch": 0.27114444979698576, "grad_norm": 0.5484449386313469, "learning_rate": 4.360905554534864e-06, "loss": 0.8698, "step": 985 }, { "epoch": 0.27252081756245267, "grad_norm": 0.7491545097860447, "learning_rate": 4.35328527649045e-06, "loss": 0.865, "step": 990 }, { "epoch": 0.2738971853279196, "grad_norm": 0.7804679526519547, "learning_rate": 4.3456265797085375e-06, "loss": 0.9351, "step": 995 }, { "epoch": 0.2752735530933866, "grad_norm": 0.5938120679599327, "learning_rate": 4.3379296229544635e-06, "loss": 0.8996, "step": 1000 }, { "epoch": 0.2752735530933866, "eval_loss": 0.8887820243835449, "eval_runtime": 37.567, "eval_samples_per_second": 133.095, "eval_steps_per_second": 2.103, "step": 1000 }, { "epoch": 0.2766499208588535, "grad_norm": 0.5199466760472583, "learning_rate": 4.330194565786696e-06, "loss": 0.9159, "step": 1005 }, { "epoch": 0.27802628862432044, "grad_norm": 0.5983605277568362, "learning_rate": 4.322421568553529e-06, "loss": 0.9187, "step": 1010 }, { "epoch": 0.27940265638978734, "grad_norm": 0.6616302612956438, "learning_rate": 4.314610792389757e-06, "loss": 0.958, "step": 1015 }, { "epoch": 0.2807790241552543, "grad_norm": 0.4539525897659009, "learning_rate": 4.30676239921333e-06, "loss": 0.8607, "step": 1020 }, { "epoch": 0.2821553919207212, "grad_norm": 0.665268425928804, "learning_rate": 4.298876551722007e-06, "loss": 0.8738, "step": 1025 }, { "epoch": 0.28353175968618816, "grad_norm": 0.5882893400505045, "learning_rate": 4.290953413389977e-06, "loss": 0.8947, "step": 1030 }, { "epoch": 0.28490812745165506, "grad_norm": 0.5736299385687077, "learning_rate": 4.282993148464467e-06, "loss": 0.9378, "step": 1035 }, { "epoch": 0.286284495217122, "grad_norm": 0.7145463884016067, "learning_rate": 4.2749959219623434e-06, "loss": 0.9029, "step": 1040 }, { "epoch": 0.28766086298258897, "grad_norm": 0.5441170060136393, "learning_rate": 4.266961899666689e-06, "loss": 0.9119, "step": 1045 }, { "epoch": 0.2890372307480559, "grad_norm": 0.7902074547549074, "learning_rate": 4.2588912481233666e-06, "loss": 0.9143, "step": 1050 }, { "epoch": 0.29041359851352283, "grad_norm": 0.6019716832259343, "learning_rate": 4.250784134637564e-06, "loss": 0.8692, "step": 1055 }, { "epoch": 0.29178996627898973, "grad_norm": 0.660623434550896, "learning_rate": 4.242640727270329e-06, "loss": 0.935, "step": 1060 }, { "epoch": 0.2931663340444567, "grad_norm": 0.5545068926773359, "learning_rate": 4.234461194835083e-06, "loss": 0.9124, "step": 1065 }, { "epoch": 0.2945427018099236, "grad_norm": 0.6474889256130907, "learning_rate": 4.2262457068941245e-06, "loss": 0.9003, "step": 1070 }, { "epoch": 0.29591906957539055, "grad_norm": 0.8379201054192673, "learning_rate": 4.217994433755112e-06, "loss": 0.8946, "step": 1075 }, { "epoch": 0.2972954373408575, "grad_norm": 0.5968271233441237, "learning_rate": 4.209707546467531e-06, "loss": 0.906, "step": 1080 }, { "epoch": 0.2986718051063244, "grad_norm": 0.5971995063991833, "learning_rate": 4.201385216819155e-06, "loss": 0.9148, "step": 1085 }, { "epoch": 0.30004817287179136, "grad_norm": 0.6868895981296808, "learning_rate": 4.193027617332476e-06, "loss": 0.8785, "step": 1090 }, { "epoch": 0.30142454063725826, "grad_norm": 0.5652277468935024, "learning_rate": 4.184634921261136e-06, "loss": 0.9108, "step": 1095 }, { "epoch": 0.3028009084027252, "grad_norm": 0.5880419121074434, "learning_rate": 4.176207302586329e-06, "loss": 0.8955, "step": 1100 }, { "epoch": 0.3041772761681921, "grad_norm": 0.7043218507759722, "learning_rate": 4.1677449360132e-06, "loss": 0.9431, "step": 1105 }, { "epoch": 0.3055536439336591, "grad_norm": 0.615367412623154, "learning_rate": 4.159247996967216e-06, "loss": 0.9234, "step": 1110 }, { "epoch": 0.306930011699126, "grad_norm": 0.7427039646128769, "learning_rate": 4.150716661590538e-06, "loss": 0.8887, "step": 1115 }, { "epoch": 0.30830637946459294, "grad_norm": 0.7310077258792612, "learning_rate": 4.142151106738364e-06, "loss": 0.8959, "step": 1120 }, { "epoch": 0.3096827472300599, "grad_norm": 0.5946178494604679, "learning_rate": 4.133551509975264e-06, "loss": 0.8957, "step": 1125 }, { "epoch": 0.3110591149955268, "grad_norm": 0.4040536498559825, "learning_rate": 4.124918049571499e-06, "loss": 0.8815, "step": 1130 }, { "epoch": 0.31243548276099375, "grad_norm": 0.510452525326286, "learning_rate": 4.1162509044993264e-06, "loss": 0.8413, "step": 1135 }, { "epoch": 0.31381185052646066, "grad_norm": 1.095591544246511, "learning_rate": 4.107550254429289e-06, "loss": 0.8945, "step": 1140 }, { "epoch": 0.3151882182919276, "grad_norm": 0.6346832255515471, "learning_rate": 4.09881627972649e-06, "loss": 0.8879, "step": 1145 }, { "epoch": 0.3165645860573945, "grad_norm": 0.6680761432639554, "learning_rate": 4.090049161446855e-06, "loss": 0.9161, "step": 1150 }, { "epoch": 0.31794095382286147, "grad_norm": 0.696714218990553, "learning_rate": 4.081249081333381e-06, "loss": 0.9182, "step": 1155 }, { "epoch": 0.31931732158832843, "grad_norm": 1.0720208443645471, "learning_rate": 4.07241622181236e-06, "loss": 0.9112, "step": 1160 }, { "epoch": 0.32069368935379533, "grad_norm": 0.6189592447847784, "learning_rate": 4.063550765989609e-06, "loss": 0.9185, "step": 1165 }, { "epoch": 0.3220700571192623, "grad_norm": 0.7117706954107574, "learning_rate": 4.054652897646666e-06, "loss": 0.8858, "step": 1170 }, { "epoch": 0.3234464248847292, "grad_norm": 0.6362209727557948, "learning_rate": 4.0457228012369855e-06, "loss": 0.8753, "step": 1175 }, { "epoch": 0.32482279265019615, "grad_norm": 0.817603348475789, "learning_rate": 4.036760661882109e-06, "loss": 0.8376, "step": 1180 }, { "epoch": 0.32619916041566305, "grad_norm": 0.7216789765837882, "learning_rate": 4.027766665367833e-06, "loss": 0.9097, "step": 1185 }, { "epoch": 0.32757552818113, "grad_norm": 0.7604270143327692, "learning_rate": 4.0187409981403525e-06, "loss": 0.8924, "step": 1190 }, { "epoch": 0.3289518959465969, "grad_norm": 0.5264426392037226, "learning_rate": 4.009683847302401e-06, "loss": 0.8908, "step": 1195 }, { "epoch": 0.33032826371206386, "grad_norm": 0.6167807303641967, "learning_rate": 4.00059540060937e-06, "loss": 0.8682, "step": 1200 }, { "epoch": 0.33032826371206386, "eval_loss": 0.8648103475570679, "eval_runtime": 37.566, "eval_samples_per_second": 133.099, "eval_steps_per_second": 2.103, "step": 1200 }, { "epoch": 0.3317046314775308, "grad_norm": 0.47612017363637654, "learning_rate": 3.991475846465415e-06, "loss": 0.8904, "step": 1205 }, { "epoch": 0.3330809992429977, "grad_norm": 0.724998856357164, "learning_rate": 3.982325373919549e-06, "loss": 0.9, "step": 1210 }, { "epoch": 0.3344573670084647, "grad_norm": 0.4223709120139991, "learning_rate": 3.973144172661731e-06, "loss": 0.8838, "step": 1215 }, { "epoch": 0.3358337347739316, "grad_norm": 0.5379099635161511, "learning_rate": 3.963932433018924e-06, "loss": 0.9138, "step": 1220 }, { "epoch": 0.33721010253939854, "grad_norm": 0.6682983804147051, "learning_rate": 3.954690345951156e-06, "loss": 0.8771, "step": 1225 }, { "epoch": 0.33858647030486544, "grad_norm": 0.7454287773417096, "learning_rate": 3.945418103047558e-06, "loss": 0.8805, "step": 1230 }, { "epoch": 0.3399628380703324, "grad_norm": 0.7063066134977648, "learning_rate": 3.936115896522395e-06, "loss": 0.8563, "step": 1235 }, { "epoch": 0.34133920583579935, "grad_norm": 0.6645576290102445, "learning_rate": 3.92678391921108e-06, "loss": 0.9031, "step": 1240 }, { "epoch": 0.34271557360126625, "grad_norm": 0.37691206208242306, "learning_rate": 3.917422364566175e-06, "loss": 0.8369, "step": 1245 }, { "epoch": 0.3440919413667332, "grad_norm": 0.41960300904843606, "learning_rate": 3.908031426653383e-06, "loss": 0.9235, "step": 1250 }, { "epoch": 0.3454683091322001, "grad_norm": 0.5867963483036928, "learning_rate": 3.898611300147525e-06, "loss": 0.8511, "step": 1255 }, { "epoch": 0.34684467689766707, "grad_norm": 0.7099329212172958, "learning_rate": 3.889162180328504e-06, "loss": 0.9318, "step": 1260 }, { "epoch": 0.34822104466313397, "grad_norm": 0.615124184951399, "learning_rate": 3.879684263077255e-06, "loss": 0.8774, "step": 1265 }, { "epoch": 0.34959741242860093, "grad_norm": 0.5769333758225725, "learning_rate": 3.870177744871686e-06, "loss": 0.8878, "step": 1270 }, { "epoch": 0.35097378019406783, "grad_norm": 0.7797713242735278, "learning_rate": 3.860642822782605e-06, "loss": 0.8559, "step": 1275 }, { "epoch": 0.3523501479595348, "grad_norm": 0.5476182116265297, "learning_rate": 3.851079694469636e-06, "loss": 0.8503, "step": 1280 }, { "epoch": 0.35372651572500174, "grad_norm": 0.5593250703560667, "learning_rate": 3.841488558177118e-06, "loss": 0.8666, "step": 1285 }, { "epoch": 0.35510288349046865, "grad_norm": 0.5174533433417139, "learning_rate": 3.831869612729999e-06, "loss": 0.88, "step": 1290 }, { "epoch": 0.3564792512559356, "grad_norm": 0.5259246199350045, "learning_rate": 3.822223057529712e-06, "loss": 0.8522, "step": 1295 }, { "epoch": 0.3578556190214025, "grad_norm": 0.9561871466566957, "learning_rate": 3.8125490925500426e-06, "loss": 0.8947, "step": 1300 }, { "epoch": 0.35923198678686946, "grad_norm": 0.5703094998753527, "learning_rate": 3.8028479183329816e-06, "loss": 0.8721, "step": 1305 }, { "epoch": 0.36060835455233636, "grad_norm": 0.7959291859516844, "learning_rate": 3.793119735984572e-06, "loss": 0.903, "step": 1310 }, { "epoch": 0.3619847223178033, "grad_norm": 0.6289787758596613, "learning_rate": 3.7833647471707345e-06, "loss": 0.8642, "step": 1315 }, { "epoch": 0.3633610900832703, "grad_norm": 0.5823311703497488, "learning_rate": 3.773583154113092e-06, "loss": 0.8812, "step": 1320 }, { "epoch": 0.3647374578487372, "grad_norm": 0.5458593624626218, "learning_rate": 3.7637751595847734e-06, "loss": 0.8848, "step": 1325 }, { "epoch": 0.36611382561420414, "grad_norm": 0.5422663450938033, "learning_rate": 3.7539409669062138e-06, "loss": 0.8546, "step": 1330 }, { "epoch": 0.36749019337967104, "grad_norm": 0.5470752416021339, "learning_rate": 3.744080779940937e-06, "loss": 0.8803, "step": 1335 }, { "epoch": 0.368866561145138, "grad_norm": 0.6784499815024887, "learning_rate": 3.7341948030913293e-06, "loss": 0.8431, "step": 1340 }, { "epoch": 0.3702429289106049, "grad_norm": 0.5173768738501396, "learning_rate": 3.7242832412944047e-06, "loss": 0.923, "step": 1345 }, { "epoch": 0.37161929667607185, "grad_norm": 0.5139718864020647, "learning_rate": 3.714346300017555e-06, "loss": 0.925, "step": 1350 }, { "epoch": 0.37299566444153875, "grad_norm": 0.5535987985827241, "learning_rate": 3.7043841852542884e-06, "loss": 0.816, "step": 1355 }, { "epoch": 0.3743720322070057, "grad_norm": 0.6220599394236812, "learning_rate": 3.6943971035199642e-06, "loss": 0.8975, "step": 1360 }, { "epoch": 0.37574839997247267, "grad_norm": 0.6424558337857292, "learning_rate": 3.684385261847506e-06, "loss": 0.8696, "step": 1365 }, { "epoch": 0.37712476773793957, "grad_norm": 0.6193220187952029, "learning_rate": 3.674348867783115e-06, "loss": 0.9187, "step": 1370 }, { "epoch": 0.3785011355034065, "grad_norm": 0.7948870412022867, "learning_rate": 3.6642881293819643e-06, "loss": 0.8794, "step": 1375 }, { "epoch": 0.37987750326887343, "grad_norm": 0.4277830946620168, "learning_rate": 3.654203255203886e-06, "loss": 0.8369, "step": 1380 }, { "epoch": 0.3812538710343404, "grad_norm": 0.5346386829006233, "learning_rate": 3.6440944543090505e-06, "loss": 0.8175, "step": 1385 }, { "epoch": 0.3826302387998073, "grad_norm": 0.5567956320600921, "learning_rate": 3.633961936253628e-06, "loss": 0.9047, "step": 1390 }, { "epoch": 0.38400660656527424, "grad_norm": 0.48101573175427903, "learning_rate": 3.623805911085452e-06, "loss": 0.8312, "step": 1395 }, { "epoch": 0.3853829743307412, "grad_norm": 0.5357232714738431, "learning_rate": 3.613626589339653e-06, "loss": 0.8757, "step": 1400 }, { "epoch": 0.3853829743307412, "eval_loss": 0.846756637096405, "eval_runtime": 37.5777, "eval_samples_per_second": 133.058, "eval_steps_per_second": 2.102, "step": 1400 }, { "epoch": 0.3867593420962081, "grad_norm": 0.4727877531161341, "learning_rate": 3.6034241820343086e-06, "loss": 0.8599, "step": 1405 }, { "epoch": 0.38813570986167506, "grad_norm": 0.4454834707158075, "learning_rate": 3.5931989006660567e-06, "loss": 0.9158, "step": 1410 }, { "epoch": 0.38951207762714196, "grad_norm": 0.6273941979012451, "learning_rate": 3.582950957205718e-06, "loss": 0.8325, "step": 1415 }, { "epoch": 0.3908884453926089, "grad_norm": 0.4418868761207735, "learning_rate": 3.5726805640939e-06, "loss": 0.8455, "step": 1420 }, { "epoch": 0.3922648131580758, "grad_norm": 0.459223577643327, "learning_rate": 3.562387934236593e-06, "loss": 0.8554, "step": 1425 }, { "epoch": 0.3936411809235428, "grad_norm": 0.6308918701832376, "learning_rate": 3.552073281000757e-06, "loss": 0.905, "step": 1430 }, { "epoch": 0.3950175486890097, "grad_norm": 0.5375942052449636, "learning_rate": 3.541736818209897e-06, "loss": 0.8989, "step": 1435 }, { "epoch": 0.39639391645447664, "grad_norm": 0.45898548069155615, "learning_rate": 3.5313787601396328e-06, "loss": 0.8568, "step": 1440 }, { "epoch": 0.3977702842199436, "grad_norm": 0.6621205621192929, "learning_rate": 3.5209993215132556e-06, "loss": 0.8988, "step": 1445 }, { "epoch": 0.3991466519854105, "grad_norm": 0.5810221477718872, "learning_rate": 3.510598717497276e-06, "loss": 0.8574, "step": 1450 }, { "epoch": 0.40052301975087745, "grad_norm": 0.49897461838543894, "learning_rate": 3.5001771636969677e-06, "loss": 0.8677, "step": 1455 }, { "epoch": 0.40189938751634435, "grad_norm": 0.343456686019272, "learning_rate": 3.4897348761518913e-06, "loss": 0.8568, "step": 1460 }, { "epoch": 0.4032757552818113, "grad_norm": 0.5890854774221135, "learning_rate": 3.4792720713314223e-06, "loss": 0.8084, "step": 1465 }, { "epoch": 0.4046521230472782, "grad_norm": 0.5050249705187615, "learning_rate": 3.4687889661302577e-06, "loss": 0.822, "step": 1470 }, { "epoch": 0.40602849081274517, "grad_norm": 0.6261040635433304, "learning_rate": 3.458285777863926e-06, "loss": 0.8983, "step": 1475 }, { "epoch": 0.4074048585782121, "grad_norm": 0.6179599241399432, "learning_rate": 3.4477627242642782e-06, "loss": 0.8186, "step": 1480 }, { "epoch": 0.408781226343679, "grad_norm": 0.4118673045686313, "learning_rate": 3.4372200234749735e-06, "loss": 0.8005, "step": 1485 }, { "epoch": 0.410157594109146, "grad_norm": 0.5523708195732413, "learning_rate": 3.4266578940469605e-06, "loss": 0.8231, "step": 1490 }, { "epoch": 0.4115339618746129, "grad_norm": 0.6262009345044368, "learning_rate": 3.416076554933944e-06, "loss": 0.8134, "step": 1495 }, { "epoch": 0.41291032964007984, "grad_norm": 0.5855645975272682, "learning_rate": 3.4054762254878477e-06, "loss": 0.8583, "step": 1500 }, { "epoch": 0.41428669740554674, "grad_norm": 0.5148476566821822, "learning_rate": 3.394857125454267e-06, "loss": 0.8362, "step": 1505 }, { "epoch": 0.4156630651710137, "grad_norm": 0.5372672904747088, "learning_rate": 3.3842194749679086e-06, "loss": 0.8381, "step": 1510 }, { "epoch": 0.4170394329364806, "grad_norm": 0.6533529330850166, "learning_rate": 3.373563494548037e-06, "loss": 0.8884, "step": 1515 }, { "epoch": 0.41841580070194756, "grad_norm": 0.7855188602038081, "learning_rate": 3.3628894050938945e-06, "loss": 0.8554, "step": 1520 }, { "epoch": 0.4197921684674145, "grad_norm": 0.7164029448250899, "learning_rate": 3.352197427880126e-06, "loss": 0.8902, "step": 1525 }, { "epoch": 0.4211685362328814, "grad_norm": 0.5654733251747366, "learning_rate": 3.3414877845521904e-06, "loss": 0.8858, "step": 1530 }, { "epoch": 0.4225449039983484, "grad_norm": 0.7161901522114019, "learning_rate": 3.3307606971217665e-06, "loss": 0.8793, "step": 1535 }, { "epoch": 0.4239212717638153, "grad_norm": 0.37921076195825976, "learning_rate": 3.320016387962151e-06, "loss": 0.8133, "step": 1540 }, { "epoch": 0.42529763952928223, "grad_norm": 0.57460942583143, "learning_rate": 3.309255079803647e-06, "loss": 0.8308, "step": 1545 }, { "epoch": 0.42667400729474914, "grad_norm": 0.6237565822649527, "learning_rate": 3.29847699572895e-06, "loss": 0.9122, "step": 1550 }, { "epoch": 0.4280503750602161, "grad_norm": 0.6604400360810607, "learning_rate": 3.2876823591685214e-06, "loss": 0.7869, "step": 1555 }, { "epoch": 0.42942674282568305, "grad_norm": 0.7964686696906034, "learning_rate": 3.276871393895954e-06, "loss": 0.8302, "step": 1560 }, { "epoch": 0.43080311059114995, "grad_norm": 0.7353573467891227, "learning_rate": 3.2660443240233387e-06, "loss": 0.8878, "step": 1565 }, { "epoch": 0.4321794783566169, "grad_norm": 0.45353882448336647, "learning_rate": 3.2552013739966147e-06, "loss": 0.8555, "step": 1570 }, { "epoch": 0.4335558461220838, "grad_norm": 0.600983398689409, "learning_rate": 3.24434276859092e-06, "loss": 0.8003, "step": 1575 }, { "epoch": 0.43493221388755077, "grad_norm": 0.33468339832513494, "learning_rate": 3.233468732905927e-06, "loss": 0.7919, "step": 1580 }, { "epoch": 0.43630858165301767, "grad_norm": 0.664686119618901, "learning_rate": 3.222579492361179e-06, "loss": 0.8585, "step": 1585 }, { "epoch": 0.4376849494184846, "grad_norm": 0.580736295078617, "learning_rate": 3.21167527269142e-06, "loss": 0.8537, "step": 1590 }, { "epoch": 0.4390613171839515, "grad_norm": 0.4795429342164793, "learning_rate": 3.2007562999419094e-06, "loss": 0.8691, "step": 1595 }, { "epoch": 0.4404376849494185, "grad_norm": 0.5533419912378857, "learning_rate": 3.189822800463742e-06, "loss": 0.8441, "step": 1600 }, { "epoch": 0.4404376849494185, "eval_loss": 0.8311466574668884, "eval_runtime": 37.5777, "eval_samples_per_second": 133.058, "eval_steps_per_second": 2.102, "step": 1600 }, { "epoch": 0.44181405271488544, "grad_norm": 0.44901727418791887, "learning_rate": 3.1788750009091473e-06, "loss": 0.8785, "step": 1605 }, { "epoch": 0.44319042048035234, "grad_norm": 0.7054761131009049, "learning_rate": 3.167913128226803e-06, "loss": 0.8442, "step": 1610 }, { "epoch": 0.4445667882458193, "grad_norm": 0.6338208711460487, "learning_rate": 3.156937409657119e-06, "loss": 0.8968, "step": 1615 }, { "epoch": 0.4459431560112862, "grad_norm": 0.6103638077202322, "learning_rate": 3.145948072727535e-06, "loss": 0.8823, "step": 1620 }, { "epoch": 0.44731952377675316, "grad_norm": 0.5814180130346099, "learning_rate": 3.134945345247797e-06, "loss": 0.8224, "step": 1625 }, { "epoch": 0.44869589154222006, "grad_norm": 0.6704683844032799, "learning_rate": 3.123929455305239e-06, "loss": 0.8797, "step": 1630 }, { "epoch": 0.450072259307687, "grad_norm": 0.4798892663438211, "learning_rate": 3.1129006312600558e-06, "loss": 0.8386, "step": 1635 }, { "epoch": 0.451448627073154, "grad_norm": 0.5497411331632864, "learning_rate": 3.101859101740565e-06, "loss": 0.858, "step": 1640 }, { "epoch": 0.4528249948386209, "grad_norm": 0.5856897243913328, "learning_rate": 3.09080509563847e-06, "loss": 0.8904, "step": 1645 }, { "epoch": 0.45420136260408783, "grad_norm": 0.47574657706442885, "learning_rate": 3.079738842104115e-06, "loss": 0.831, "step": 1650 }, { "epoch": 0.45557773036955473, "grad_norm": 0.7356917244396622, "learning_rate": 3.0686605705417337e-06, "loss": 0.8638, "step": 1655 }, { "epoch": 0.4569540981350217, "grad_norm": 0.4609027088459466, "learning_rate": 3.057570510604696e-06, "loss": 0.8342, "step": 1660 }, { "epoch": 0.4583304659004886, "grad_norm": 0.649733338307217, "learning_rate": 3.0464688921907436e-06, "loss": 0.844, "step": 1665 }, { "epoch": 0.45970683366595555, "grad_norm": 0.6452557057054518, "learning_rate": 3.035355945437228e-06, "loss": 0.901, "step": 1670 }, { "epoch": 0.46108320143142245, "grad_norm": 0.48053889170013253, "learning_rate": 3.0242319007163373e-06, "loss": 0.8237, "step": 1675 }, { "epoch": 0.4624595691968894, "grad_norm": 0.5156070687409189, "learning_rate": 3.01309698863032e-06, "loss": 0.7832, "step": 1680 }, { "epoch": 0.46383593696235637, "grad_norm": 0.6487171872307684, "learning_rate": 3.001951440006708e-06, "loss": 0.8302, "step": 1685 }, { "epoch": 0.46521230472782327, "grad_norm": 0.4088236693992105, "learning_rate": 2.9907954858935277e-06, "loss": 0.7978, "step": 1690 }, { "epoch": 0.4665886724932902, "grad_norm": 0.6166137125554756, "learning_rate": 2.9796293575545143e-06, "loss": 0.8327, "step": 1695 }, { "epoch": 0.4679650402587571, "grad_norm": 0.5324800896120733, "learning_rate": 2.9684532864643123e-06, "loss": 0.8497, "step": 1700 }, { "epoch": 0.4693414080242241, "grad_norm": 0.5330587838888724, "learning_rate": 2.957267504303682e-06, "loss": 0.8318, "step": 1705 }, { "epoch": 0.470717775789691, "grad_norm": 0.3066813832965278, "learning_rate": 2.946072242954695e-06, "loss": 0.7959, "step": 1710 }, { "epoch": 0.47209414355515794, "grad_norm": 0.4174047766855664, "learning_rate": 2.934867734495927e-06, "loss": 0.8157, "step": 1715 }, { "epoch": 0.4734705113206249, "grad_norm": 0.5513234180318742, "learning_rate": 2.9236542111976468e-06, "loss": 0.8657, "step": 1720 }, { "epoch": 0.4748468790860918, "grad_norm": 0.36186756893473565, "learning_rate": 2.9124319055170012e-06, "loss": 0.8108, "step": 1725 }, { "epoch": 0.47622324685155876, "grad_norm": 0.48483961573426937, "learning_rate": 2.9012010500931966e-06, "loss": 0.8532, "step": 1730 }, { "epoch": 0.47759961461702566, "grad_norm": 0.656595154954433, "learning_rate": 2.8899618777426763e-06, "loss": 0.8186, "step": 1735 }, { "epoch": 0.4789759823824926, "grad_norm": 0.45682292991009654, "learning_rate": 2.878714621454294e-06, "loss": 0.8507, "step": 1740 }, { "epoch": 0.4803523501479595, "grad_norm": 0.4456324138614082, "learning_rate": 2.867459514384485e-06, "loss": 0.8809, "step": 1745 }, { "epoch": 0.4817287179134265, "grad_norm": 0.5388546054688201, "learning_rate": 2.856196789852429e-06, "loss": 0.8236, "step": 1750 }, { "epoch": 0.4831050856788934, "grad_norm": 0.660462229832185, "learning_rate": 2.84492668133522e-06, "loss": 0.8338, "step": 1755 }, { "epoch": 0.48448145344436033, "grad_norm": 0.5278146557932175, "learning_rate": 2.833649422463019e-06, "loss": 0.814, "step": 1760 }, { "epoch": 0.4858578212098273, "grad_norm": 0.5470942669154103, "learning_rate": 2.8223652470142184e-06, "loss": 0.8183, "step": 1765 }, { "epoch": 0.4872341889752942, "grad_norm": 0.505713995433775, "learning_rate": 2.8110743889105874e-06, "loss": 0.8387, "step": 1770 }, { "epoch": 0.48861055674076115, "grad_norm": 0.4875452733503727, "learning_rate": 2.79977708221243e-06, "loss": 0.7981, "step": 1775 }, { "epoch": 0.48998692450622805, "grad_norm": 0.43916802590802495, "learning_rate": 2.7884735611137288e-06, "loss": 0.8532, "step": 1780 }, { "epoch": 0.491363292271695, "grad_norm": 0.33369528684284744, "learning_rate": 2.777164059937292e-06, "loss": 0.8408, "step": 1785 }, { "epoch": 0.4927396600371619, "grad_norm": 0.7202736685154576, "learning_rate": 2.765848813129895e-06, "loss": 0.8532, "step": 1790 }, { "epoch": 0.49411602780262887, "grad_norm": 0.42511937449619386, "learning_rate": 2.7545280552574204e-06, "loss": 0.8224, "step": 1795 }, { "epoch": 0.4954923955680958, "grad_norm": 0.545135520242412, "learning_rate": 2.7432020209999956e-06, "loss": 0.8197, "step": 1800 }, { "epoch": 0.4954923955680958, "eval_loss": 0.8205735087394714, "eval_runtime": 37.5781, "eval_samples_per_second": 133.056, "eval_steps_per_second": 2.102, "step": 1800 }, { "epoch": 0.4968687633335627, "grad_norm": 0.48066576255326976, "learning_rate": 2.7318709451471288e-06, "loss": 0.8239, "step": 1805 }, { "epoch": 0.4982451310990297, "grad_norm": 0.7847734357041681, "learning_rate": 2.7205350625928383e-06, "loss": 0.9108, "step": 1810 }, { "epoch": 0.4996214988644966, "grad_norm": 0.7016473795626382, "learning_rate": 2.70919460833079e-06, "loss": 0.8367, "step": 1815 }, { "epoch": 0.5009978666299635, "grad_norm": 0.6178351624669732, "learning_rate": 2.697849817449415e-06, "loss": 0.8282, "step": 1820 }, { "epoch": 0.5023742343954305, "grad_norm": 0.4091504823861622, "learning_rate": 2.6865009251270506e-06, "loss": 0.8526, "step": 1825 }, { "epoch": 0.5037506021608974, "grad_norm": 0.6183413041352451, "learning_rate": 2.6751481666270513e-06, "loss": 0.8473, "step": 1830 }, { "epoch": 0.5051269699263643, "grad_norm": 0.606173243841225, "learning_rate": 2.6637917772929213e-06, "loss": 0.8567, "step": 1835 }, { "epoch": 0.5065033376918313, "grad_norm": 0.6132486268789019, "learning_rate": 2.65243199254343e-06, "loss": 0.8325, "step": 1840 }, { "epoch": 0.5078797054572982, "grad_norm": 0.4854570121858536, "learning_rate": 2.6410690478677353e-06, "loss": 0.7892, "step": 1845 }, { "epoch": 0.5092560732227651, "grad_norm": 0.37412561464355937, "learning_rate": 2.6297031788205004e-06, "loss": 0.8094, "step": 1850 }, { "epoch": 0.510632440988232, "grad_norm": 0.6203401035441314, "learning_rate": 2.618334621017009e-06, "loss": 0.822, "step": 1855 }, { "epoch": 0.512008808753699, "grad_norm": 0.653767694616689, "learning_rate": 2.6069636101282862e-06, "loss": 0.8367, "step": 1860 }, { "epoch": 0.5133851765191659, "grad_norm": 0.6245289959361092, "learning_rate": 2.595590381876209e-06, "loss": 0.8328, "step": 1865 }, { "epoch": 0.5147615442846328, "grad_norm": 0.4457307365521416, "learning_rate": 2.584215172028618e-06, "loss": 0.8312, "step": 1870 }, { "epoch": 0.5161379120500997, "grad_norm": 0.4815075541618031, "learning_rate": 2.572838216394434e-06, "loss": 0.8686, "step": 1875 }, { "epoch": 0.5175142798155667, "grad_norm": 0.5702137712078795, "learning_rate": 2.561459750818769e-06, "loss": 0.8347, "step": 1880 }, { "epoch": 0.5188906475810336, "grad_norm": 0.48523555192405865, "learning_rate": 2.5500800111780357e-06, "loss": 0.8036, "step": 1885 }, { "epoch": 0.5202670153465005, "grad_norm": 0.3430136628045643, "learning_rate": 2.5386992333750565e-06, "loss": 0.8291, "step": 1890 }, { "epoch": 0.5216433831119676, "grad_norm": 0.470576052419782, "learning_rate": 2.5273176533341777e-06, "loss": 0.77, "step": 1895 }, { "epoch": 0.5230197508774345, "grad_norm": 0.4479209182378322, "learning_rate": 2.5159355069963744e-06, "loss": 0.8091, "step": 1900 }, { "epoch": 0.5243961186429014, "grad_norm": 0.5763895685381267, "learning_rate": 2.5045530303143604e-06, "loss": 0.863, "step": 1905 }, { "epoch": 0.5257724864083683, "grad_norm": 0.6080627036981748, "learning_rate": 2.4931704592477e-06, "loss": 0.8713, "step": 1910 }, { "epoch": 0.5271488541738353, "grad_norm": 0.5799585453878151, "learning_rate": 2.4817880297579134e-06, "loss": 0.7895, "step": 1915 }, { "epoch": 0.5285252219393022, "grad_norm": 0.6516081569612333, "learning_rate": 2.4704059778035823e-06, "loss": 0.8062, "step": 1920 }, { "epoch": 0.5299015897047691, "grad_norm": 0.5284447836149975, "learning_rate": 2.459024539335467e-06, "loss": 0.8549, "step": 1925 }, { "epoch": 0.5312779574702361, "grad_norm": 0.5029144016269022, "learning_rate": 2.447643950291608e-06, "loss": 0.8279, "step": 1930 }, { "epoch": 0.532654325235703, "grad_norm": 0.6658834516119427, "learning_rate": 2.4362644465924367e-06, "loss": 0.8335, "step": 1935 }, { "epoch": 0.5340306930011699, "grad_norm": 0.5953435916222078, "learning_rate": 2.4248862641358865e-06, "loss": 0.7918, "step": 1940 }, { "epoch": 0.5354070607666368, "grad_norm": 0.37670098774520216, "learning_rate": 2.4135096387925e-06, "loss": 0.8638, "step": 1945 }, { "epoch": 0.5367834285321038, "grad_norm": 0.511664128571869, "learning_rate": 2.4021348064005417e-06, "loss": 0.8377, "step": 1950 }, { "epoch": 0.5381597962975707, "grad_norm": 0.42161706759354106, "learning_rate": 2.3907620027611083e-06, "loss": 0.83, "step": 1955 }, { "epoch": 0.5395361640630376, "grad_norm": 0.3040178118533278, "learning_rate": 2.3793914636332394e-06, "loss": 0.8746, "step": 1960 }, { "epoch": 0.5409125318285046, "grad_norm": 0.689149159117422, "learning_rate": 2.3680234247290305e-06, "loss": 0.8247, "step": 1965 }, { "epoch": 0.5422888995939715, "grad_norm": 0.44356157035761207, "learning_rate": 2.3566581217087496e-06, "loss": 0.8277, "step": 1970 }, { "epoch": 0.5436652673594384, "grad_norm": 0.5918309406851234, "learning_rate": 2.3452957901759486e-06, "loss": 0.8025, "step": 1975 }, { "epoch": 0.5450416351249053, "grad_norm": 0.4997809391017281, "learning_rate": 2.333936665672579e-06, "loss": 0.835, "step": 1980 }, { "epoch": 0.5464180028903723, "grad_norm": 0.49766628433257887, "learning_rate": 2.3225809836741118e-06, "loss": 0.7756, "step": 1985 }, { "epoch": 0.5477943706558392, "grad_norm": 0.40407922865092843, "learning_rate": 2.3112289795846537e-06, "loss": 0.7967, "step": 1990 }, { "epoch": 0.5491707384213061, "grad_norm": 0.6308739524640976, "learning_rate": 2.2998808887320697e-06, "loss": 0.781, "step": 1995 }, { "epoch": 0.5505471061867732, "grad_norm": 0.5743558152124677, "learning_rate": 2.2885369463631003e-06, "loss": 0.7807, "step": 2000 }, { "epoch": 0.5505471061867732, "eval_loss": 0.8089554905891418, "eval_runtime": 37.5623, "eval_samples_per_second": 133.112, "eval_steps_per_second": 2.103, "step": 2000 }, { "epoch": 0.5519234739522401, "grad_norm": 0.5814715096200206, "learning_rate": 2.277197387638491e-06, "loss": 0.8329, "step": 2005 }, { "epoch": 0.553299841717707, "grad_norm": 0.5479788170588639, "learning_rate": 2.265862447628111e-06, "loss": 0.8742, "step": 2010 }, { "epoch": 0.5546762094831739, "grad_norm": 0.3062291452983948, "learning_rate": 2.254532361306085e-06, "loss": 0.7671, "step": 2015 }, { "epoch": 0.5560525772486409, "grad_norm": 0.5763488126780687, "learning_rate": 2.2432073635459196e-06, "loss": 0.8437, "step": 2020 }, { "epoch": 0.5574289450141078, "grad_norm": 0.668939310724625, "learning_rate": 2.2318876891156356e-06, "loss": 0.8973, "step": 2025 }, { "epoch": 0.5588053127795747, "grad_norm": 0.37776371512529605, "learning_rate": 2.2205735726729023e-06, "loss": 0.8345, "step": 2030 }, { "epoch": 0.5601816805450416, "grad_norm": 0.41280705089503705, "learning_rate": 2.2092652487601675e-06, "loss": 0.8323, "step": 2035 }, { "epoch": 0.5615580483105086, "grad_norm": 0.4632509161576438, "learning_rate": 2.1979629517998027e-06, "loss": 0.8282, "step": 2040 }, { "epoch": 0.5629344160759755, "grad_norm": 0.7200713467674419, "learning_rate": 2.186666916089239e-06, "loss": 0.8379, "step": 2045 }, { "epoch": 0.5643107838414424, "grad_norm": 0.5178576662970729, "learning_rate": 2.1753773757961137e-06, "loss": 0.8261, "step": 2050 }, { "epoch": 0.5656871516069094, "grad_norm": 0.5610321336212082, "learning_rate": 2.1640945649534096e-06, "loss": 0.8309, "step": 2055 }, { "epoch": 0.5670635193723763, "grad_norm": 0.5892891551665735, "learning_rate": 2.1528187174546093e-06, "loss": 0.8297, "step": 2060 }, { "epoch": 0.5684398871378432, "grad_norm": 0.4486760776076786, "learning_rate": 2.141550067048846e-06, "loss": 0.8389, "step": 2065 }, { "epoch": 0.5698162549033101, "grad_norm": 0.7907237469372685, "learning_rate": 2.1302888473360566e-06, "loss": 0.8321, "step": 2070 }, { "epoch": 0.5711926226687771, "grad_norm": 0.566706200199042, "learning_rate": 2.119035291762136e-06, "loss": 0.8212, "step": 2075 }, { "epoch": 0.572568990434244, "grad_norm": 0.5263130247632009, "learning_rate": 2.1077896336141043e-06, "loss": 0.8042, "step": 2080 }, { "epoch": 0.5739453581997109, "grad_norm": 0.5308066511740118, "learning_rate": 2.096552106015266e-06, "loss": 0.815, "step": 2085 }, { "epoch": 0.5753217259651779, "grad_norm": 0.6957541746253699, "learning_rate": 2.0853229419203808e-06, "loss": 0.8261, "step": 2090 }, { "epoch": 0.5766980937306448, "grad_norm": 0.5966904887807215, "learning_rate": 2.0741023741108276e-06, "loss": 0.827, "step": 2095 }, { "epoch": 0.5780744614961117, "grad_norm": 0.4539597770375397, "learning_rate": 2.0628906351897885e-06, "loss": 0.8182, "step": 2100 }, { "epoch": 0.5794508292615786, "grad_norm": 0.6834026151474581, "learning_rate": 2.0516879575774203e-06, "loss": 0.8303, "step": 2105 }, { "epoch": 0.5808271970270457, "grad_norm": 0.5239200524956313, "learning_rate": 2.040494573506038e-06, "loss": 0.7741, "step": 2110 }, { "epoch": 0.5822035647925126, "grad_norm": 0.45753694417604984, "learning_rate": 2.0293107150153006e-06, "loss": 0.8397, "step": 2115 }, { "epoch": 0.5835799325579795, "grad_norm": 0.46224504028065294, "learning_rate": 2.018136613947401e-06, "loss": 0.8244, "step": 2120 }, { "epoch": 0.5849563003234465, "grad_norm": 0.4341362951474634, "learning_rate": 2.0069725019422624e-06, "loss": 0.8009, "step": 2125 }, { "epoch": 0.5863326680889134, "grad_norm": 0.4900348861826574, "learning_rate": 1.9958186104327317e-06, "loss": 0.8483, "step": 2130 }, { "epoch": 0.5877090358543803, "grad_norm": 0.6257220799254032, "learning_rate": 1.9846751706397832e-06, "loss": 0.8405, "step": 2135 }, { "epoch": 0.5890854036198472, "grad_norm": 0.522323526016304, "learning_rate": 1.9735424135677283e-06, "loss": 0.8322, "step": 2140 }, { "epoch": 0.5904617713853142, "grad_norm": 0.6328347166830222, "learning_rate": 1.9624205699994256e-06, "loss": 0.8607, "step": 2145 }, { "epoch": 0.5918381391507811, "grad_norm": 0.52757076025723, "learning_rate": 1.951309870491494e-06, "loss": 0.8003, "step": 2150 }, { "epoch": 0.593214506916248, "grad_norm": 0.4530698938004871, "learning_rate": 1.9402105453695356e-06, "loss": 0.843, "step": 2155 }, { "epoch": 0.594590874681715, "grad_norm": 0.5275287498766006, "learning_rate": 1.9291228247233607e-06, "loss": 0.7959, "step": 2160 }, { "epoch": 0.5959672424471819, "grad_norm": 0.28183154522813586, "learning_rate": 1.9180469384022203e-06, "loss": 0.7799, "step": 2165 }, { "epoch": 0.5973436102126488, "grad_norm": 0.5336877667445091, "learning_rate": 1.9069831160100338e-06, "loss": 0.7979, "step": 2170 }, { "epoch": 0.5987199779781157, "grad_norm": 0.6532051021851727, "learning_rate": 1.8959315869006405e-06, "loss": 0.8359, "step": 2175 }, { "epoch": 0.6000963457435827, "grad_norm": 0.5307514597676389, "learning_rate": 1.8848925801730344e-06, "loss": 0.8937, "step": 2180 }, { "epoch": 0.6014727135090496, "grad_norm": 0.4042895440437914, "learning_rate": 1.8738663246666234e-06, "loss": 0.8252, "step": 2185 }, { "epoch": 0.6028490812745165, "grad_norm": 0.44894176165695887, "learning_rate": 1.8628530489564771e-06, "loss": 0.7835, "step": 2190 }, { "epoch": 0.6042254490399834, "grad_norm": 0.4991539092087521, "learning_rate": 1.8518529813485973e-06, "loss": 0.814, "step": 2195 }, { "epoch": 0.6056018168054504, "grad_norm": 0.3709613327877024, "learning_rate": 1.8408663498751788e-06, "loss": 0.7757, "step": 2200 }, { "epoch": 0.6056018168054504, "eval_loss": 0.8015441298484802, "eval_runtime": 37.566, "eval_samples_per_second": 133.099, "eval_steps_per_second": 2.103, "step": 2200 }, { "epoch": 0.6069781845709173, "grad_norm": 0.4683158255043096, "learning_rate": 1.829893382289886e-06, "loss": 0.8097, "step": 2205 }, { "epoch": 0.6083545523363842, "grad_norm": 0.5627576336667334, "learning_rate": 1.818934306063126e-06, "loss": 0.806, "step": 2210 }, { "epoch": 0.6097309201018513, "grad_norm": 0.5488618788671827, "learning_rate": 1.8079893483773413e-06, "loss": 0.8185, "step": 2215 }, { "epoch": 0.6111072878673182, "grad_norm": 0.44918917109293605, "learning_rate": 1.7970587361222946e-06, "loss": 0.8271, "step": 2220 }, { "epoch": 0.6124836556327851, "grad_norm": 0.32979924519766196, "learning_rate": 1.786142695890367e-06, "loss": 0.7828, "step": 2225 }, { "epoch": 0.613860023398252, "grad_norm": 0.35131817470857274, "learning_rate": 1.7752414539718582e-06, "loss": 0.8191, "step": 2230 }, { "epoch": 0.615236391163719, "grad_norm": 0.4656658318567486, "learning_rate": 1.7643552363503009e-06, "loss": 0.8358, "step": 2235 }, { "epoch": 0.6166127589291859, "grad_norm": 0.570964532177242, "learning_rate": 1.7534842686977721e-06, "loss": 0.8596, "step": 2240 }, { "epoch": 0.6179891266946528, "grad_norm": 0.34212198519333364, "learning_rate": 1.742628776370216e-06, "loss": 0.818, "step": 2245 }, { "epoch": 0.6193654944601198, "grad_norm": 0.48535112289382754, "learning_rate": 1.7317889844027707e-06, "loss": 0.8623, "step": 2250 }, { "epoch": 0.6207418622255867, "grad_norm": 0.49166416763933013, "learning_rate": 1.7209651175051056e-06, "loss": 0.8468, "step": 2255 }, { "epoch": 0.6221182299910536, "grad_norm": 0.4583562391030405, "learning_rate": 1.7101574000567633e-06, "loss": 0.822, "step": 2260 }, { "epoch": 0.6234945977565205, "grad_norm": 0.46794730423054154, "learning_rate": 1.6993660561025072e-06, "loss": 0.8562, "step": 2265 }, { "epoch": 0.6248709655219875, "grad_norm": 0.6099197600468227, "learning_rate": 1.6885913093476741e-06, "loss": 0.8078, "step": 2270 }, { "epoch": 0.6262473332874544, "grad_norm": 0.5249915623303978, "learning_rate": 1.677833383153542e-06, "loss": 0.8219, "step": 2275 }, { "epoch": 0.6276237010529213, "grad_norm": 0.4377376254057894, "learning_rate": 1.6670925005326977e-06, "loss": 0.8179, "step": 2280 }, { "epoch": 0.6290000688183883, "grad_norm": 0.42122297813103243, "learning_rate": 1.6563688841444137e-06, "loss": 0.8418, "step": 2285 }, { "epoch": 0.6303764365838552, "grad_norm": 0.5683008577226283, "learning_rate": 1.6456627562900296e-06, "loss": 0.7891, "step": 2290 }, { "epoch": 0.6317528043493221, "grad_norm": 0.6037662603976885, "learning_rate": 1.63497433890835e-06, "loss": 0.832, "step": 2295 }, { "epoch": 0.633129172114789, "grad_norm": 0.5247870050565184, "learning_rate": 1.6243038535710365e-06, "loss": 0.8076, "step": 2300 }, { "epoch": 0.634505539880256, "grad_norm": 0.41197503482402553, "learning_rate": 1.6136515214780227e-06, "loss": 0.7596, "step": 2305 }, { "epoch": 0.6358819076457229, "grad_norm": 0.4791099245291028, "learning_rate": 1.603017563452919e-06, "loss": 0.8107, "step": 2310 }, { "epoch": 0.6372582754111898, "grad_norm": 0.43592237438522047, "learning_rate": 1.592402199938443e-06, "loss": 0.8185, "step": 2315 }, { "epoch": 0.6386346431766569, "grad_norm": 0.5940623980499528, "learning_rate": 1.5818056509918478e-06, "loss": 0.8004, "step": 2320 }, { "epoch": 0.6400110109421238, "grad_norm": 0.5250401511659805, "learning_rate": 1.5712281362803561e-06, "loss": 0.802, "step": 2325 }, { "epoch": 0.6413873787075907, "grad_norm": 0.5915893010847437, "learning_rate": 1.5606698750766108e-06, "loss": 0.8642, "step": 2330 }, { "epoch": 0.6427637464730576, "grad_norm": 0.46778493168617225, "learning_rate": 1.550131086254129e-06, "loss": 0.8092, "step": 2335 }, { "epoch": 0.6441401142385246, "grad_norm": 0.491189650305899, "learning_rate": 1.5396119882827651e-06, "loss": 0.8026, "step": 2340 }, { "epoch": 0.6455164820039915, "grad_norm": 0.43582541174722744, "learning_rate": 1.5291127992241766e-06, "loss": 0.8141, "step": 2345 }, { "epoch": 0.6468928497694584, "grad_norm": 0.6142246346723809, "learning_rate": 1.5186337367273105e-06, "loss": 0.8008, "step": 2350 }, { "epoch": 0.6482692175349253, "grad_norm": 0.35462943795501145, "learning_rate": 1.5081750180238891e-06, "loss": 0.7667, "step": 2355 }, { "epoch": 0.6496455853003923, "grad_norm": 0.41180209198737644, "learning_rate": 1.4977368599239061e-06, "loss": 0.8028, "step": 2360 }, { "epoch": 0.6510219530658592, "grad_norm": 0.7189067391117583, "learning_rate": 1.487319478811131e-06, "loss": 0.8339, "step": 2365 }, { "epoch": 0.6523983208313261, "grad_norm": 0.3806482298690338, "learning_rate": 1.4769230906386272e-06, "loss": 0.8151, "step": 2370 }, { "epoch": 0.6537746885967931, "grad_norm": 0.4983576378974423, "learning_rate": 1.4665479109242696e-06, "loss": 0.7939, "step": 2375 }, { "epoch": 0.65515105636226, "grad_norm": 0.33177024863231713, "learning_rate": 1.4561941547462855e-06, "loss": 0.8009, "step": 2380 }, { "epoch": 0.6565274241277269, "grad_norm": 0.538297674107413, "learning_rate": 1.4458620367387838e-06, "loss": 0.8025, "step": 2385 }, { "epoch": 0.6579037918931938, "grad_norm": 0.4901494849475784, "learning_rate": 1.4355517710873184e-06, "loss": 0.7845, "step": 2390 }, { "epoch": 0.6592801596586608, "grad_norm": 0.5486200463100078, "learning_rate": 1.4252635715244394e-06, "loss": 0.8208, "step": 2395 }, { "epoch": 0.6606565274241277, "grad_norm": 0.4897253532401449, "learning_rate": 1.4149976513252677e-06, "loss": 0.7818, "step": 2400 }, { "epoch": 0.6606565274241277, "eval_loss": 0.7957330942153931, "eval_runtime": 37.5749, "eval_samples_per_second": 133.068, "eval_steps_per_second": 2.102, "step": 2400 }, { "epoch": 0.6620328951895946, "grad_norm": 0.46111910295697045, "learning_rate": 1.4047542233030683e-06, "loss": 0.8258, "step": 2405 }, { "epoch": 0.6634092629550616, "grad_norm": 0.5784849194355406, "learning_rate": 1.3945334998048425e-06, "loss": 0.8157, "step": 2410 }, { "epoch": 0.6647856307205285, "grad_norm": 0.5455404395159691, "learning_rate": 1.3843356927069266e-06, "loss": 0.8155, "step": 2415 }, { "epoch": 0.6661619984859954, "grad_norm": 0.4060432121547348, "learning_rate": 1.3741610134105984e-06, "loss": 0.7862, "step": 2420 }, { "epoch": 0.6675383662514623, "grad_norm": 0.5175461237539263, "learning_rate": 1.3640096728376922e-06, "loss": 0.796, "step": 2425 }, { "epoch": 0.6689147340169294, "grad_norm": 0.5579421958353933, "learning_rate": 1.353881881426231e-06, "loss": 0.8159, "step": 2430 }, { "epoch": 0.6702911017823963, "grad_norm": 0.3938550567000424, "learning_rate": 1.3437778491260626e-06, "loss": 0.7888, "step": 2435 }, { "epoch": 0.6716674695478632, "grad_norm": 0.3621816668127261, "learning_rate": 1.3336977853945055e-06, "loss": 0.7831, "step": 2440 }, { "epoch": 0.6730438373133302, "grad_norm": 0.6046109892992472, "learning_rate": 1.3236418991920065e-06, "loss": 0.7899, "step": 2445 }, { "epoch": 0.6744202050787971, "grad_norm": 0.52626524074172, "learning_rate": 1.3136103989778138e-06, "loss": 0.7591, "step": 2450 }, { "epoch": 0.675796572844264, "grad_norm": 0.5540405574756208, "learning_rate": 1.303603492705649e-06, "loss": 0.8, "step": 2455 }, { "epoch": 0.6771729406097309, "grad_norm": 0.3755183771069488, "learning_rate": 1.2936213878194031e-06, "loss": 0.819, "step": 2460 }, { "epoch": 0.6785493083751979, "grad_norm": 0.585894250903544, "learning_rate": 1.2836642912488287e-06, "loss": 0.8327, "step": 2465 }, { "epoch": 0.6799256761406648, "grad_norm": 0.5007223315080989, "learning_rate": 1.2737324094052569e-06, "loss": 0.8055, "step": 2470 }, { "epoch": 0.6813020439061317, "grad_norm": 0.4543771383459902, "learning_rate": 1.2638259481773164e-06, "loss": 0.7892, "step": 2475 }, { "epoch": 0.6826784116715987, "grad_norm": 0.45325256663521224, "learning_rate": 1.2539451129266603e-06, "loss": 0.7904, "step": 2480 }, { "epoch": 0.6840547794370656, "grad_norm": 0.42755072695822427, "learning_rate": 1.244090108483718e-06, "loss": 0.8696, "step": 2485 }, { "epoch": 0.6854311472025325, "grad_norm": 0.5749380186174909, "learning_rate": 1.2342611391434424e-06, "loss": 0.7695, "step": 2490 }, { "epoch": 0.6868075149679994, "grad_norm": 0.47696517661731086, "learning_rate": 1.2244584086610783e-06, "loss": 0.8061, "step": 2495 }, { "epoch": 0.6881838827334664, "grad_norm": 0.47628535730003985, "learning_rate": 1.2146821202479347e-06, "loss": 0.8252, "step": 2500 }, { "epoch": 0.6895602504989333, "grad_norm": 0.4918566967586663, "learning_rate": 1.204932476567175e-06, "loss": 0.8306, "step": 2505 }, { "epoch": 0.6909366182644002, "grad_norm": 0.45157773288125097, "learning_rate": 1.1952096797296167e-06, "loss": 0.7911, "step": 2510 }, { "epoch": 0.6923129860298671, "grad_norm": 0.5449035155568759, "learning_rate": 1.1855139312895412e-06, "loss": 0.8297, "step": 2515 }, { "epoch": 0.6936893537953341, "grad_norm": 0.539632313493914, "learning_rate": 1.175845432240511e-06, "loss": 0.7938, "step": 2520 }, { "epoch": 0.695065721560801, "grad_norm": 0.267579658107954, "learning_rate": 1.16620438301121e-06, "loss": 0.8326, "step": 2525 }, { "epoch": 0.6964420893262679, "grad_norm": 0.45094192311072506, "learning_rate": 1.1565909834612843e-06, "loss": 0.8183, "step": 2530 }, { "epoch": 0.697818457091735, "grad_norm": 0.45018310356306585, "learning_rate": 1.1470054328772015e-06, "loss": 0.8312, "step": 2535 }, { "epoch": 0.6991948248572019, "grad_norm": 0.5424670154260764, "learning_rate": 1.1374479299681144e-06, "loss": 0.8547, "step": 2540 }, { "epoch": 0.7005711926226688, "grad_norm": 0.3929689914659013, "learning_rate": 1.12791867286175e-06, "loss": 0.7631, "step": 2545 }, { "epoch": 0.7019475603881357, "grad_norm": 0.5189958974232449, "learning_rate": 1.1184178591002936e-06, "loss": 0.7974, "step": 2550 }, { "epoch": 0.7033239281536027, "grad_norm": 0.4940843745776494, "learning_rate": 1.1089456856363023e-06, "loss": 0.7859, "step": 2555 }, { "epoch": 0.7047002959190696, "grad_norm": 0.3903233975311801, "learning_rate": 1.0995023488286132e-06, "loss": 0.7555, "step": 2560 }, { "epoch": 0.7060766636845365, "grad_norm": 0.44367526110003797, "learning_rate": 1.090088044438281e-06, "loss": 0.83, "step": 2565 }, { "epoch": 0.7074530314500035, "grad_norm": 0.3625988730276094, "learning_rate": 1.0807029676245146e-06, "loss": 0.772, "step": 2570 }, { "epoch": 0.7088293992154704, "grad_norm": 0.4378410267181844, "learning_rate": 1.0713473129406342e-06, "loss": 0.7913, "step": 2575 }, { "epoch": 0.7102057669809373, "grad_norm": 0.5659553149700242, "learning_rate": 1.062021274330035e-06, "loss": 0.8333, "step": 2580 }, { "epoch": 0.7115821347464042, "grad_norm": 0.480279896191206, "learning_rate": 1.0527250451221714e-06, "loss": 0.7924, "step": 2585 }, { "epoch": 0.7129585025118712, "grad_norm": 0.5794313371061417, "learning_rate": 1.043458818028546e-06, "loss": 0.8025, "step": 2590 }, { "epoch": 0.7143348702773381, "grad_norm": 0.577692103807798, "learning_rate": 1.0342227851387132e-06, "loss": 0.8102, "step": 2595 }, { "epoch": 0.715711238042805, "grad_norm": 0.7048819923650593, "learning_rate": 1.0250171379163035e-06, "loss": 0.8235, "step": 2600 }, { "epoch": 0.715711238042805, "eval_loss": 0.7914655208587646, "eval_runtime": 37.5636, "eval_samples_per_second": 133.108, "eval_steps_per_second": 2.103, "step": 2600 }, { "epoch": 0.717087605808272, "grad_norm": 0.5135405438687206, "learning_rate": 1.0158420671950458e-06, "loss": 0.8354, "step": 2605 }, { "epoch": 0.7184639735737389, "grad_norm": 0.49088601989364183, "learning_rate": 1.0066977631748192e-06, "loss": 0.8243, "step": 2610 }, { "epoch": 0.7198403413392058, "grad_norm": 0.43340747956612413, "learning_rate": 9.975844154177068e-07, "loss": 0.8082, "step": 2615 }, { "epoch": 0.7212167091046727, "grad_norm": 0.4929677260928928, "learning_rate": 9.88502212844063e-07, "loss": 0.8259, "step": 2620 }, { "epoch": 0.7225930768701397, "grad_norm": 0.5110702548169729, "learning_rate": 9.794513437286039e-07, "loss": 0.8231, "step": 2625 }, { "epoch": 0.7239694446356066, "grad_norm": 0.43992378048573655, "learning_rate": 9.704319956964997e-07, "loss": 0.7803, "step": 2630 }, { "epoch": 0.7253458124010735, "grad_norm": 0.3073181934194982, "learning_rate": 9.61444355719484e-07, "loss": 0.7606, "step": 2635 }, { "epoch": 0.7267221801665406, "grad_norm": 0.5324585928575014, "learning_rate": 9.524886101119846e-07, "loss": 0.8537, "step": 2640 }, { "epoch": 0.7280985479320075, "grad_norm": 0.45921892305675543, "learning_rate": 9.435649445272516e-07, "loss": 0.8069, "step": 2645 }, { "epoch": 0.7294749156974744, "grad_norm": 0.48054038898023743, "learning_rate": 9.346735439535182e-07, "loss": 0.8097, "step": 2650 }, { "epoch": 0.7308512834629413, "grad_norm": 0.4380511249962659, "learning_rate": 9.25814592710158e-07, "loss": 0.7914, "step": 2655 }, { "epoch": 0.7322276512284083, "grad_norm": 0.337588637619449, "learning_rate": 9.16988274443871e-07, "loss": 0.7967, "step": 2660 }, { "epoch": 0.7336040189938752, "grad_norm": 0.4171603673327812, "learning_rate": 9.08194772124871e-07, "loss": 0.7909, "step": 2665 }, { "epoch": 0.7349803867593421, "grad_norm": 0.5340434480071784, "learning_rate": 8.994342680430971e-07, "loss": 0.7702, "step": 2670 }, { "epoch": 0.736356754524809, "grad_norm": 0.45146796076237466, "learning_rate": 8.907069438044283e-07, "loss": 0.8057, "step": 2675 }, { "epoch": 0.737733122290276, "grad_norm": 0.4216880628061979, "learning_rate": 8.820129803269272e-07, "loss": 0.8074, "step": 2680 }, { "epoch": 0.7391094900557429, "grad_norm": 0.5979240606977037, "learning_rate": 8.733525578370849e-07, "loss": 0.8162, "step": 2685 }, { "epoch": 0.7404858578212098, "grad_norm": 0.490528351797546, "learning_rate": 8.647258558660829e-07, "loss": 0.8103, "step": 2690 }, { "epoch": 0.7418622255866768, "grad_norm": 0.6084011328958213, "learning_rate": 8.561330532460765e-07, "loss": 0.8821, "step": 2695 }, { "epoch": 0.7432385933521437, "grad_norm": 0.3812636095028781, "learning_rate": 8.47574328106483e-07, "loss": 0.7365, "step": 2700 }, { "epoch": 0.7446149611176106, "grad_norm": 0.4418699971833215, "learning_rate": 8.390498578702924e-07, "loss": 0.8175, "step": 2705 }, { "epoch": 0.7459913288830775, "grad_norm": 0.3665034778827065, "learning_rate": 8.305598192503892e-07, "loss": 0.7635, "step": 2710 }, { "epoch": 0.7473676966485445, "grad_norm": 0.39085090708364434, "learning_rate": 8.22104388245884e-07, "loss": 0.7682, "step": 2715 }, { "epoch": 0.7487440644140114, "grad_norm": 0.515141732505318, "learning_rate": 8.136837401384734e-07, "loss": 0.8256, "step": 2720 }, { "epoch": 0.7501204321794783, "grad_norm": 0.4873699702775256, "learning_rate": 8.052980494887996e-07, "loss": 0.8079, "step": 2725 }, { "epoch": 0.7514967999449453, "grad_norm": 0.3657787697288285, "learning_rate": 7.969474901328359e-07, "loss": 0.78, "step": 2730 }, { "epoch": 0.7528731677104122, "grad_norm": 0.514544078198768, "learning_rate": 7.886322351782782e-07, "loss": 0.821, "step": 2735 }, { "epoch": 0.7542495354758791, "grad_norm": 0.31173151058089976, "learning_rate": 7.803524570009638e-07, "loss": 0.793, "step": 2740 }, { "epoch": 0.755625903241346, "grad_norm": 0.42241508592582616, "learning_rate": 7.7210832724129e-07, "loss": 0.7798, "step": 2745 }, { "epoch": 0.757002271006813, "grad_norm": 0.3071764331948748, "learning_rate": 7.63900016800663e-07, "loss": 0.7698, "step": 2750 }, { "epoch": 0.75837863877228, "grad_norm": 0.5439311482897147, "learning_rate": 7.55727695837949e-07, "loss": 0.8452, "step": 2755 }, { "epoch": 0.7597550065377469, "grad_norm": 0.3563361621522225, "learning_rate": 7.475915337659517e-07, "loss": 0.7901, "step": 2760 }, { "epoch": 0.7611313743032139, "grad_norm": 0.5754448075856651, "learning_rate": 7.394916992478982e-07, "loss": 0.7638, "step": 2765 }, { "epoch": 0.7625077420686808, "grad_norm": 0.511147390789392, "learning_rate": 7.314283601939432e-07, "loss": 0.7966, "step": 2770 }, { "epoch": 0.7638841098341477, "grad_norm": 0.5082660532801647, "learning_rate": 7.234016837576855e-07, "loss": 0.7977, "step": 2775 }, { "epoch": 0.7652604775996146, "grad_norm": 0.4037699385261041, "learning_rate": 7.154118363327076e-07, "loss": 0.8714, "step": 2780 }, { "epoch": 0.7666368453650816, "grad_norm": 0.4667815625876788, "learning_rate": 7.074589835491236e-07, "loss": 0.797, "step": 2785 }, { "epoch": 0.7680132131305485, "grad_norm": 0.4480832075891221, "learning_rate": 6.995432902701452e-07, "loss": 0.8327, "step": 2790 }, { "epoch": 0.7693895808960154, "grad_norm": 0.42380996565690365, "learning_rate": 6.916649205886639e-07, "loss": 0.7462, "step": 2795 }, { "epoch": 0.7707659486614824, "grad_norm": 0.33453237656086077, "learning_rate": 6.838240378238528e-07, "loss": 0.7854, "step": 2800 }, { "epoch": 0.7707659486614824, "eval_loss": 0.788250744342804, "eval_runtime": 37.5666, "eval_samples_per_second": 133.097, "eval_steps_per_second": 2.103, "step": 2800 }, { "epoch": 0.7721423164269493, "grad_norm": 0.7936762263185101, "learning_rate": 6.760208045177777e-07, "loss": 0.8265, "step": 2805 }, { "epoch": 0.7735186841924162, "grad_norm": 0.5449658294351013, "learning_rate": 6.68255382432027e-07, "loss": 0.8289, "step": 2810 }, { "epoch": 0.7748950519578831, "grad_norm": 0.5162826155892138, "learning_rate": 6.605279325443615e-07, "loss": 0.7767, "step": 2815 }, { "epoch": 0.7762714197233501, "grad_norm": 0.36993986629866255, "learning_rate": 6.528386150453747e-07, "loss": 0.7914, "step": 2820 }, { "epoch": 0.777647787488817, "grad_norm": 0.4025106600408673, "learning_rate": 6.451875893351742e-07, "loss": 0.8094, "step": 2825 }, { "epoch": 0.7790241552542839, "grad_norm": 0.3803752081384692, "learning_rate": 6.375750140200729e-07, "loss": 0.7834, "step": 2830 }, { "epoch": 0.7804005230197508, "grad_norm": 0.4647690834948723, "learning_rate": 6.300010469093085e-07, "loss": 0.7677, "step": 2835 }, { "epoch": 0.7817768907852178, "grad_norm": 0.4360851526308311, "learning_rate": 6.224658450117638e-07, "loss": 0.8241, "step": 2840 }, { "epoch": 0.7831532585506847, "grad_norm": 0.40784724834132835, "learning_rate": 6.149695645327197e-07, "loss": 0.7794, "step": 2845 }, { "epoch": 0.7845296263161516, "grad_norm": 0.3733748290573589, "learning_rate": 6.075123608706093e-07, "loss": 0.7934, "step": 2850 }, { "epoch": 0.7859059940816187, "grad_norm": 0.46257419791313253, "learning_rate": 6.000943886138039e-07, "loss": 0.8197, "step": 2855 }, { "epoch": 0.7872823618470856, "grad_norm": 0.3846998032798624, "learning_rate": 5.927158015374032e-07, "loss": 0.7601, "step": 2860 }, { "epoch": 0.7886587296125525, "grad_norm": 0.32739463317156486, "learning_rate": 5.853767526000506e-07, "loss": 0.7976, "step": 2865 }, { "epoch": 0.7900350973780194, "grad_norm": 0.38672592567017716, "learning_rate": 5.780773939407586e-07, "loss": 0.8075, "step": 2870 }, { "epoch": 0.7914114651434864, "grad_norm": 0.4386867822178414, "learning_rate": 5.708178768757594e-07, "loss": 0.8151, "step": 2875 }, { "epoch": 0.7927878329089533, "grad_norm": 0.3283397967295564, "learning_rate": 5.635983518953664e-07, "loss": 0.8467, "step": 2880 }, { "epoch": 0.7941642006744202, "grad_norm": 0.4837636826148133, "learning_rate": 5.564189686608528e-07, "loss": 0.829, "step": 2885 }, { "epoch": 0.7955405684398872, "grad_norm": 0.7164604908817186, "learning_rate": 5.492798760013504e-07, "loss": 0.8363, "step": 2890 }, { "epoch": 0.7969169362053541, "grad_norm": 0.5246720628848824, "learning_rate": 5.421812219107652e-07, "loss": 0.7728, "step": 2895 }, { "epoch": 0.798293303970821, "grad_norm": 0.48145772973666245, "learning_rate": 5.351231535447096e-07, "loss": 0.8351, "step": 2900 }, { "epoch": 0.7996696717362879, "grad_norm": 0.6428159033099584, "learning_rate": 5.2810581721745e-07, "loss": 0.7936, "step": 2905 }, { "epoch": 0.8010460395017549, "grad_norm": 0.42644687973421025, "learning_rate": 5.211293583988736e-07, "loss": 0.7612, "step": 2910 }, { "epoch": 0.8024224072672218, "grad_norm": 0.49844315374917036, "learning_rate": 5.141939217114761e-07, "loss": 0.9081, "step": 2915 }, { "epoch": 0.8037987750326887, "grad_norm": 0.45961744850293057, "learning_rate": 5.072996509273597e-07, "loss": 0.7703, "step": 2920 }, { "epoch": 0.8051751427981557, "grad_norm": 0.46944692269351923, "learning_rate": 5.004466889652568e-07, "loss": 0.8183, "step": 2925 }, { "epoch": 0.8065515105636226, "grad_norm": 0.43054493887650536, "learning_rate": 4.93635177887562e-07, "loss": 0.8182, "step": 2930 }, { "epoch": 0.8079278783290895, "grad_norm": 0.40134994197247525, "learning_rate": 4.86865258897391e-07, "loss": 0.7662, "step": 2935 }, { "epoch": 0.8093042460945564, "grad_norm": 0.3679569701658888, "learning_rate": 4.801370723356533e-07, "loss": 0.7397, "step": 2940 }, { "epoch": 0.8106806138600234, "grad_norm": 0.4996662512653364, "learning_rate": 4.7345075767814277e-07, "loss": 0.7655, "step": 2945 }, { "epoch": 0.8120569816254903, "grad_norm": 0.33008573990481305, "learning_rate": 4.668064535326433e-07, "loss": 0.7733, "step": 2950 }, { "epoch": 0.8134333493909572, "grad_norm": 0.5216225381160015, "learning_rate": 4.602042976360596e-07, "loss": 0.8131, "step": 2955 }, { "epoch": 0.8148097171564243, "grad_norm": 0.3009799921127595, "learning_rate": 4.536444268515608e-07, "loss": 0.761, "step": 2960 }, { "epoch": 0.8161860849218912, "grad_norm": 0.4808517772368335, "learning_rate": 4.4712697716573994e-07, "loss": 0.7887, "step": 2965 }, { "epoch": 0.817562452687358, "grad_norm": 0.31056324010954034, "learning_rate": 4.406520836858003e-07, "loss": 0.7373, "step": 2970 }, { "epoch": 0.818938820452825, "grad_norm": 0.4026317051978566, "learning_rate": 4.342198806367512e-07, "loss": 0.8102, "step": 2975 }, { "epoch": 0.820315188218292, "grad_norm": 0.571522212386199, "learning_rate": 4.2783050135862454e-07, "loss": 0.8232, "step": 2980 }, { "epoch": 0.8216915559837589, "grad_norm": 0.6020021292977741, "learning_rate": 4.2148407830371553e-07, "loss": 0.8423, "step": 2985 }, { "epoch": 0.8230679237492258, "grad_norm": 0.31515971826344896, "learning_rate": 4.1518074303383006e-07, "loss": 0.7556, "step": 2990 }, { "epoch": 0.8244442915146927, "grad_norm": 0.4964415865676406, "learning_rate": 4.0892062621756436e-07, "loss": 0.8106, "step": 2995 }, { "epoch": 0.8258206592801597, "grad_norm": 0.5295609606861289, "learning_rate": 4.027038576275921e-07, "loss": 0.7958, "step": 3000 }, { "epoch": 0.8258206592801597, "eval_loss": 0.7862712144851685, "eval_runtime": 37.5789, "eval_samples_per_second": 133.054, "eval_steps_per_second": 2.102, "step": 3000 }, { "epoch": 0.8271970270456266, "grad_norm": 0.5235293567756518, "learning_rate": 3.9653056613797315e-07, "loss": 0.8119, "step": 3005 }, { "epoch": 0.8285733948110935, "grad_norm": 0.46096657495011545, "learning_rate": 3.904008797214867e-07, "loss": 0.7939, "step": 3010 }, { "epoch": 0.8299497625765605, "grad_norm": 0.4583761081561358, "learning_rate": 3.8431492544697384e-07, "loss": 0.8206, "step": 3015 }, { "epoch": 0.8313261303420274, "grad_norm": 0.4307989854498663, "learning_rate": 3.7827282947670686e-07, "loss": 0.805, "step": 3020 }, { "epoch": 0.8327024981074943, "grad_norm": 0.6458058398139984, "learning_rate": 3.722747170637703e-07, "loss": 0.8272, "step": 3025 }, { "epoch": 0.8340788658729612, "grad_norm": 0.27246801971613765, "learning_rate": 3.663207125494667e-07, "loss": 0.7188, "step": 3030 }, { "epoch": 0.8354552336384282, "grad_norm": 0.31316968826402736, "learning_rate": 3.604109393607397e-07, "loss": 0.7771, "step": 3035 }, { "epoch": 0.8368316014038951, "grad_norm": 0.46558859230251337, "learning_rate": 3.545455200076148e-07, "loss": 0.7697, "step": 3040 }, { "epoch": 0.838207969169362, "grad_norm": 0.4510507655849259, "learning_rate": 3.4872457608065706e-07, "loss": 0.7729, "step": 3045 }, { "epoch": 0.839584336934829, "grad_norm": 0.43470367010584865, "learning_rate": 3.4294822824845447e-07, "loss": 0.8024, "step": 3050 }, { "epoch": 0.8409607047002959, "grad_norm": 0.5457490936555786, "learning_rate": 3.3721659625511466e-07, "loss": 0.8288, "step": 3055 }, { "epoch": 0.8423370724657628, "grad_norm": 0.4460830059047153, "learning_rate": 3.315297989177829e-07, "loss": 0.7704, "step": 3060 }, { "epoch": 0.8437134402312297, "grad_norm": 0.4017748109392074, "learning_rate": 3.2588795412417715e-07, "loss": 0.8081, "step": 3065 }, { "epoch": 0.8450898079966968, "grad_norm": 0.43509526046571056, "learning_rate": 3.20291178830148e-07, "loss": 0.7696, "step": 3070 }, { "epoch": 0.8464661757621637, "grad_norm": 0.4968191608476648, "learning_rate": 3.1473958905725023e-07, "loss": 0.8007, "step": 3075 }, { "epoch": 0.8478425435276306, "grad_norm": 0.35779440235114646, "learning_rate": 3.092332998903416e-07, "loss": 0.7844, "step": 3080 }, { "epoch": 0.8492189112930976, "grad_norm": 0.41824405953159766, "learning_rate": 3.0377242547519224e-07, "loss": 0.8119, "step": 3085 }, { "epoch": 0.8505952790585645, "grad_norm": 0.3627881737322641, "learning_rate": 2.983570790161236e-07, "loss": 0.7926, "step": 3090 }, { "epoch": 0.8519716468240314, "grad_norm": 0.6401049982386239, "learning_rate": 2.9298737277365875e-07, "loss": 0.7957, "step": 3095 }, { "epoch": 0.8533480145894983, "grad_norm": 0.4518026617163562, "learning_rate": 2.8766341806219565e-07, "loss": 0.8071, "step": 3100 }, { "epoch": 0.8547243823549653, "grad_norm": 0.431837948798665, "learning_rate": 2.823853252476988e-07, "loss": 0.8007, "step": 3105 }, { "epoch": 0.8561007501204322, "grad_norm": 0.347884381355328, "learning_rate": 2.771532037454136e-07, "loss": 0.8173, "step": 3110 }, { "epoch": 0.8574771178858991, "grad_norm": 0.4330221328925689, "learning_rate": 2.719671620175968e-07, "loss": 0.7266, "step": 3115 }, { "epoch": 0.8588534856513661, "grad_norm": 0.3637332745725778, "learning_rate": 2.6682730757126627e-07, "loss": 0.8076, "step": 3120 }, { "epoch": 0.860229853416833, "grad_norm": 0.6109148107559905, "learning_rate": 2.6173374695597693e-07, "loss": 0.8339, "step": 3125 }, { "epoch": 0.8616062211822999, "grad_norm": 0.5076827526837399, "learning_rate": 2.566865857616066e-07, "loss": 0.8432, "step": 3130 }, { "epoch": 0.8629825889477668, "grad_norm": 0.5457670411835718, "learning_rate": 2.5168592861617216e-07, "loss": 0.7928, "step": 3135 }, { "epoch": 0.8643589567132338, "grad_norm": 0.41906024544255965, "learning_rate": 2.4673187918365593e-07, "loss": 0.7741, "step": 3140 }, { "epoch": 0.8657353244787007, "grad_norm": 0.5161588655724126, "learning_rate": 2.4182454016186046e-07, "loss": 0.8115, "step": 3145 }, { "epoch": 0.8671116922441676, "grad_norm": 0.3816011931932575, "learning_rate": 2.3696401328027806e-07, "loss": 0.7693, "step": 3150 }, { "epoch": 0.8684880600096345, "grad_norm": 0.4654707569379619, "learning_rate": 2.3215039929798205e-07, "loss": 0.8122, "step": 3155 }, { "epoch": 0.8698644277751015, "grad_norm": 0.306667830180016, "learning_rate": 2.2738379800153641e-07, "loss": 0.7612, "step": 3160 }, { "epoch": 0.8712407955405684, "grad_norm": 0.46545229964609797, "learning_rate": 2.226643082029309e-07, "loss": 0.7892, "step": 3165 }, { "epoch": 0.8726171633060353, "grad_norm": 0.1953075510516031, "learning_rate": 2.1799202773752943e-07, "loss": 0.7521, "step": 3170 }, { "epoch": 0.8739935310715023, "grad_norm": 0.381786701995952, "learning_rate": 2.1336705346204301e-07, "loss": 0.8512, "step": 3175 }, { "epoch": 0.8753698988369693, "grad_norm": 0.43933009705251314, "learning_rate": 2.087894812525218e-07, "loss": 0.8025, "step": 3180 }, { "epoch": 0.8767462666024362, "grad_norm": 0.44005358531346456, "learning_rate": 2.042594060023681e-07, "loss": 0.7756, "step": 3185 }, { "epoch": 0.878122634367903, "grad_norm": 0.49824718850948474, "learning_rate": 1.9977692162036876e-07, "loss": 0.7978, "step": 3190 }, { "epoch": 0.8794990021333701, "grad_norm": 0.3227198829491034, "learning_rate": 1.95342121028749e-07, "loss": 0.7738, "step": 3195 }, { "epoch": 0.880875369898837, "grad_norm": 0.34555908654114686, "learning_rate": 1.9095509616124385e-07, "loss": 0.8192, "step": 3200 }, { "epoch": 0.880875369898837, "eval_loss": 0.7828695774078369, "eval_runtime": 37.5732, "eval_samples_per_second": 133.074, "eval_steps_per_second": 2.103, "step": 3200 }, { "epoch": 0.8822517376643039, "grad_norm": 0.4058193409462779, "learning_rate": 1.866159379611965e-07, "loss": 0.7827, "step": 3205 }, { "epoch": 0.8836281054297709, "grad_norm": 0.634249713668319, "learning_rate": 1.8232473637966874e-07, "loss": 0.8316, "step": 3210 }, { "epoch": 0.8850044731952378, "grad_norm": 0.47086197421634446, "learning_rate": 1.7808158037357997e-07, "loss": 0.8106, "step": 3215 }, { "epoch": 0.8863808409607047, "grad_norm": 0.4923463916845965, "learning_rate": 1.7388655790385928e-07, "loss": 0.7618, "step": 3220 }, { "epoch": 0.8877572087261716, "grad_norm": 0.5706429066921409, "learning_rate": 1.6973975593362557e-07, "loss": 0.8026, "step": 3225 }, { "epoch": 0.8891335764916386, "grad_norm": 0.40654338758220415, "learning_rate": 1.656412604263824e-07, "loss": 0.805, "step": 3230 }, { "epoch": 0.8905099442571055, "grad_norm": 0.36239810433911784, "learning_rate": 1.615911563442385e-07, "loss": 0.7901, "step": 3235 }, { "epoch": 0.8918863120225724, "grad_norm": 0.3902839217508603, "learning_rate": 1.5758952764614254e-07, "loss": 0.772, "step": 3240 }, { "epoch": 0.8932626797880394, "grad_norm": 0.4730030214965539, "learning_rate": 1.536364572861465e-07, "loss": 0.7981, "step": 3245 }, { "epoch": 0.8946390475535063, "grad_norm": 0.5222497792049411, "learning_rate": 1.4973202721168452e-07, "loss": 0.774, "step": 3250 }, { "epoch": 0.8960154153189732, "grad_norm": 0.457253030971323, "learning_rate": 1.4587631836187362e-07, "loss": 0.7762, "step": 3255 }, { "epoch": 0.8973917830844401, "grad_norm": 0.35084396615645064, "learning_rate": 1.420694106658363e-07, "loss": 0.7956, "step": 3260 }, { "epoch": 0.8987681508499071, "grad_norm": 0.46947234334867594, "learning_rate": 1.3831138304104374e-07, "loss": 0.7488, "step": 3265 }, { "epoch": 0.900144518615374, "grad_norm": 0.4211703728724711, "learning_rate": 1.3460231339168018e-07, "loss": 0.7594, "step": 3270 }, { "epoch": 0.9015208863808409, "grad_norm": 0.4318480909616356, "learning_rate": 1.3094227860702636e-07, "loss": 0.7535, "step": 3275 }, { "epoch": 0.902897254146308, "grad_norm": 0.5304843144531105, "learning_rate": 1.2733135455986755e-07, "loss": 0.7631, "step": 3280 }, { "epoch": 0.9042736219117748, "grad_norm": 0.39712712425481755, "learning_rate": 1.237696161049201e-07, "loss": 0.7967, "step": 3285 }, { "epoch": 0.9056499896772418, "grad_norm": 0.3607138196755753, "learning_rate": 1.2025713707727954e-07, "loss": 0.7673, "step": 3290 }, { "epoch": 0.9070263574427087, "grad_norm": 0.36526857232256044, "learning_rate": 1.1679399029088878e-07, "loss": 0.8021, "step": 3295 }, { "epoch": 0.9084027252081757, "grad_norm": 0.4307690645765424, "learning_rate": 1.1338024753703076e-07, "loss": 0.7855, "step": 3300 }, { "epoch": 0.9097790929736426, "grad_norm": 0.5370242540693829, "learning_rate": 1.1001597958283927e-07, "loss": 0.7942, "step": 3305 }, { "epoch": 0.9111554607391095, "grad_norm": 0.762967780858564, "learning_rate": 1.067012561698319e-07, "loss": 0.7809, "step": 3310 }, { "epoch": 0.9125318285045764, "grad_norm": 0.38146623134983393, "learning_rate": 1.0343614601246388e-07, "loss": 0.8512, "step": 3315 }, { "epoch": 0.9139081962700434, "grad_norm": 0.483635357463634, "learning_rate": 1.0022071679670426e-07, "loss": 0.8334, "step": 3320 }, { "epoch": 0.9152845640355103, "grad_norm": 0.45987532559150957, "learning_rate": 9.705503517863286e-08, "loss": 0.7624, "step": 3325 }, { "epoch": 0.9166609318009772, "grad_norm": 0.5146306345364831, "learning_rate": 9.393916678305831e-08, "loss": 0.781, "step": 3330 }, { "epoch": 0.9180372995664442, "grad_norm": 0.4353941113364763, "learning_rate": 9.087317620215642e-08, "loss": 0.7926, "step": 3335 }, { "epoch": 0.9194136673319111, "grad_norm": 0.5238188017998141, "learning_rate": 8.78571269941339e-08, "loss": 0.7944, "step": 3340 }, { "epoch": 0.920790035097378, "grad_norm": 0.39784081936662746, "learning_rate": 8.48910816819079e-08, "loss": 0.7456, "step": 3345 }, { "epoch": 0.9221664028628449, "grad_norm": 0.4021648902631367, "learning_rate": 8.197510175181279e-08, "loss": 0.7972, "step": 3350 }, { "epoch": 0.9235427706283119, "grad_norm": 0.4535198430400062, "learning_rate": 7.910924765232169e-08, "loss": 0.7716, "step": 3355 }, { "epoch": 0.9249191383937788, "grad_norm": 0.4679043443981073, "learning_rate": 7.629357879279764e-08, "loss": 0.8151, "step": 3360 }, { "epoch": 0.9262955061592457, "grad_norm": 0.4138512577922224, "learning_rate": 7.352815354225856e-08, "loss": 0.7778, "step": 3365 }, { "epoch": 0.9276718739247127, "grad_norm": 0.4959390023966221, "learning_rate": 7.08130292281703e-08, "loss": 0.7654, "step": 3370 }, { "epoch": 0.9290482416901796, "grad_norm": 0.3635744148121797, "learning_rate": 6.8148262135255e-08, "loss": 0.7713, "step": 3375 }, { "epoch": 0.9304246094556465, "grad_norm": 0.3880051004435765, "learning_rate": 6.553390750432709e-08, "loss": 0.797, "step": 3380 }, { "epoch": 0.9318009772211134, "grad_norm": 0.376848176957449, "learning_rate": 6.297001953114696e-08, "loss": 0.7915, "step": 3385 }, { "epoch": 0.9331773449865804, "grad_norm": 0.5229520565025577, "learning_rate": 6.045665136529683e-08, "loss": 0.7831, "step": 3390 }, { "epoch": 0.9345537127520473, "grad_norm": 0.3749002422429637, "learning_rate": 5.799385510908029e-08, "loss": 0.813, "step": 3395 }, { "epoch": 0.9359300805175143, "grad_norm": 0.538377010361361, "learning_rate": 5.558168181644147e-08, "loss": 0.765, "step": 3400 }, { "epoch": 0.9359300805175143, "eval_loss": 0.7824124693870544, "eval_runtime": 37.5729, "eval_samples_per_second": 133.075, "eval_steps_per_second": 2.103, "step": 3400 }, { "epoch": 0.9373064482829813, "grad_norm": 0.40542132743231607, "learning_rate": 5.3220181491906997e-08, "loss": 0.7939, "step": 3405 }, { "epoch": 0.9386828160484482, "grad_norm": 0.45653448883349285, "learning_rate": 5.0909403089548504e-08, "loss": 0.7683, "step": 3410 }, { "epoch": 0.9400591838139151, "grad_norm": 0.43498449283812496, "learning_rate": 4.864939451196926e-08, "loss": 0.7706, "step": 3415 }, { "epoch": 0.941435551579382, "grad_norm": 0.4632642432067583, "learning_rate": 4.6440202609309983e-08, "loss": 0.847, "step": 3420 }, { "epoch": 0.942811919344849, "grad_norm": 0.398573024905695, "learning_rate": 4.428187317827848e-08, "loss": 0.8004, "step": 3425 }, { "epoch": 0.9441882871103159, "grad_norm": 0.3520961675944394, "learning_rate": 4.217445096119932e-08, "loss": 0.7768, "step": 3430 }, { "epoch": 0.9455646548757828, "grad_norm": 0.2659907213915567, "learning_rate": 4.011797964508707e-08, "loss": 0.8068, "step": 3435 }, { "epoch": 0.9469410226412498, "grad_norm": 0.35721378611449023, "learning_rate": 3.8112501860740893e-08, "loss": 0.7761, "step": 3440 }, { "epoch": 0.9483173904067167, "grad_norm": 0.5662605277154725, "learning_rate": 3.615805918185999e-08, "loss": 0.7956, "step": 3445 }, { "epoch": 0.9496937581721836, "grad_norm": 0.4708997292827784, "learning_rate": 3.4254692124181256e-08, "loss": 0.781, "step": 3450 }, { "epoch": 0.9510701259376505, "grad_norm": 0.3463983685723411, "learning_rate": 3.240244014464211e-08, "loss": 0.8038, "step": 3455 }, { "epoch": 0.9524464937031175, "grad_norm": 0.30914041030379646, "learning_rate": 3.060134164055928e-08, "loss": 0.7855, "step": 3460 }, { "epoch": 0.9538228614685844, "grad_norm": 0.4094929093409069, "learning_rate": 2.885143394883466e-08, "loss": 0.7922, "step": 3465 }, { "epoch": 0.9551992292340513, "grad_norm": 0.3550357942305495, "learning_rate": 2.7152753345181248e-08, "loss": 0.7488, "step": 3470 }, { "epoch": 0.9565755969995182, "grad_norm": 0.4969121322892503, "learning_rate": 2.5505335043370105e-08, "loss": 0.8235, "step": 3475 }, { "epoch": 0.9579519647649852, "grad_norm": 0.5594870015775546, "learning_rate": 2.3909213194501513e-08, "loss": 0.8019, "step": 3480 }, { "epoch": 0.9593283325304521, "grad_norm": 0.3982458733337497, "learning_rate": 2.2364420886297202e-08, "loss": 0.7931, "step": 3485 }, { "epoch": 0.960704700295919, "grad_norm": 0.40600940068470787, "learning_rate": 2.087099014241256e-08, "loss": 0.7751, "step": 3490 }, { "epoch": 0.962081068061386, "grad_norm": 0.5688667017637077, "learning_rate": 1.9428951921774687e-08, "loss": 0.8253, "step": 3495 }, { "epoch": 0.963457435826853, "grad_norm": 0.41507791635586866, "learning_rate": 1.8038336117940368e-08, "loss": 0.7615, "step": 3500 }, { "epoch": 0.9648338035923198, "grad_norm": 0.3974167729618538, "learning_rate": 1.6699171558474946e-08, "loss": 0.7943, "step": 3505 }, { "epoch": 0.9662101713577868, "grad_norm": 0.6940727253100508, "learning_rate": 1.541148600435721e-08, "loss": 0.8198, "step": 3510 }, { "epoch": 0.9675865391232538, "grad_norm": 0.46568158788586206, "learning_rate": 1.4175306149400715e-08, "loss": 0.8164, "step": 3515 }, { "epoch": 0.9689629068887207, "grad_norm": 0.5038186258746858, "learning_rate": 1.2990657619703361e-08, "loss": 0.755, "step": 3520 }, { "epoch": 0.9703392746541876, "grad_norm": 0.44963940045587836, "learning_rate": 1.1857564973114798e-08, "loss": 0.8276, "step": 3525 }, { "epoch": 0.9717156424196546, "grad_norm": 0.3869453345256143, "learning_rate": 1.0776051698727363e-08, "loss": 0.7643, "step": 3530 }, { "epoch": 0.9730920101851215, "grad_norm": 0.4625561038516912, "learning_rate": 9.746140216388978e-09, "loss": 0.7961, "step": 3535 }, { "epoch": 0.9744683779505884, "grad_norm": 0.43657476353887914, "learning_rate": 8.767851876239075e-09, "loss": 0.785, "step": 3540 }, { "epoch": 0.9758447457160553, "grad_norm": 0.36886304870486564, "learning_rate": 7.841206958265901e-09, "loss": 0.8109, "step": 3545 }, { "epoch": 0.9772211134815223, "grad_norm": 0.48365435933116596, "learning_rate": 6.9662246718849025e-09, "loss": 0.805, "step": 3550 }, { "epoch": 0.9785974812469892, "grad_norm": 0.5676220156571874, "learning_rate": 6.142923155542379e-09, "loss": 0.8249, "step": 3555 }, { "epoch": 0.9799738490124561, "grad_norm": 0.5777045577708868, "learning_rate": 5.371319476338288e-09, "loss": 0.8371, "step": 3560 }, { "epoch": 0.9813502167779231, "grad_norm": 0.4285639278318182, "learning_rate": 4.651429629672077e-09, "loss": 0.8493, "step": 3565 }, { "epoch": 0.98272658454339, "grad_norm": 0.5788587404914071, "learning_rate": 3.9832685389123995e-09, "loss": 0.8533, "step": 3570 }, { "epoch": 0.9841029523088569, "grad_norm": 0.504857407057922, "learning_rate": 3.3668500550870787e-09, "loss": 0.8482, "step": 3575 }, { "epoch": 0.9854793200743238, "grad_norm": 0.48362838745693276, "learning_rate": 2.8021869565958427e-09, "loss": 0.7877, "step": 3580 }, { "epoch": 0.9868556878397908, "grad_norm": 0.4381509362003844, "learning_rate": 2.289290948944978e-09, "loss": 0.8337, "step": 3585 }, { "epoch": 0.9882320556052577, "grad_norm": 0.35883074278050137, "learning_rate": 1.8281726645061338e-09, "loss": 0.8103, "step": 3590 }, { "epoch": 0.9896084233707246, "grad_norm": 0.30791028488533356, "learning_rate": 1.4188416622945566e-09, "loss": 0.7772, "step": 3595 }, { "epoch": 0.9909847911361916, "grad_norm": 0.519345784734355, "learning_rate": 1.0613064277711916e-09, "loss": 0.7939, "step": 3600 }, { "epoch": 0.9909847911361916, "eval_loss": 0.7823675870895386, "eval_runtime": 37.5664, "eval_samples_per_second": 133.098, "eval_steps_per_second": 2.103, "step": 3600 }, { "epoch": 0.9923611589016585, "grad_norm": 0.5511524759279022, "learning_rate": 7.555743726675446e-10, "loss": 0.7623, "step": 3605 }, { "epoch": 0.9937375266671254, "grad_norm": 0.46879838306977284, "learning_rate": 5.01651834831085e-10, "loss": 0.7906, "step": 3610 }, { "epoch": 0.9951138944325923, "grad_norm": 0.5200935954838051, "learning_rate": 2.9954407809423823e-10, "loss": 0.788, "step": 3615 }, { "epoch": 0.9964902621980594, "grad_norm": 0.47789258781015914, "learning_rate": 1.4925529216558432e-10, "loss": 0.7885, "step": 3620 }, { "epoch": 0.9978666299635263, "grad_norm": 0.2942434423794945, "learning_rate": 5.078859254242785e-11, "loss": 0.8001, "step": 3625 }, { "epoch": 0.9992429977289932, "grad_norm": 0.49341815424912977, "learning_rate": 4.1460204466825526e-12, "loss": 0.8235, "step": 3630 }, { "epoch": 0.99979354483518, "step": 3632, "total_flos": 1258708224212992.0, "train_loss": 0.935843440076328, "train_runtime": 26049.5256, "train_samples_per_second": 35.699, "train_steps_per_second": 0.139 } ], "logging_steps": 5, "max_steps": 3632, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1258708224212992.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }