{ "best_metric": 1.1752163171768188, "best_model_checkpoint": "./bert_mlm_finetuned/checkpoint-1600", "epoch": 0.2048, "eval_steps": 100, "global_step": 1600, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00064, "grad_norm": 7.650606632232666, "learning_rate": 1.0000000000000002e-06, "loss": 6.29, "step": 5 }, { "epoch": 0.00128, "grad_norm": 4.541823387145996, "learning_rate": 2.0000000000000003e-06, "loss": 6.3815, "step": 10 }, { "epoch": 0.00192, "grad_norm": 4.245054721832275, "learning_rate": 3e-06, "loss": 6.2854, "step": 15 }, { "epoch": 0.00256, "grad_norm": 4.5587897300720215, "learning_rate": 4.000000000000001e-06, "loss": 6.0674, "step": 20 }, { "epoch": 0.0032, "grad_norm": 3.7703804969787598, "learning_rate": 4.800000000000001e-06, "loss": 6.2961, "step": 25 }, { "epoch": 0.00384, "grad_norm": 3.8425862789154053, "learning_rate": 5.8e-06, "loss": 6.3326, "step": 30 }, { "epoch": 0.00448, "grad_norm": 4.413463115692139, "learning_rate": 6.800000000000001e-06, "loss": 6.183, "step": 35 }, { "epoch": 0.00512, "grad_norm": 4.1980509757995605, "learning_rate": 7.800000000000002e-06, "loss": 6.2654, "step": 40 }, { "epoch": 0.00576, "grad_norm": 3.9166719913482666, "learning_rate": 8.8e-06, "loss": 6.0916, "step": 45 }, { "epoch": 0.0064, "grad_norm": 3.4706904888153076, "learning_rate": 9.800000000000001e-06, "loss": 6.103, "step": 50 }, { "epoch": 0.00704, "grad_norm": 5.138203144073486, "learning_rate": 9.999998372356185e-06, "loss": 6.2379, "step": 55 }, { "epoch": 0.00768, "grad_norm": 3.7806520462036133, "learning_rate": 9.999991760055e-06, "loss": 6.1776, "step": 60 }, { "epoch": 0.00832, "grad_norm": 3.5731871128082275, "learning_rate": 9.999980061375427e-06, "loss": 6.2082, "step": 65 }, { "epoch": 0.00896, "grad_norm": 3.661797285079956, "learning_rate": 9.999963276329369e-06, "loss": 6.0704, "step": 70 }, { "epoch": 0.0096, "grad_norm": 3.6181113719940186, "learning_rate": 9.999941404933902e-06, "loss": 6.2081, "step": 75 }, { "epoch": 0.01024, "grad_norm": 3.3162803649902344, "learning_rate": 9.99991444721127e-06, "loss": 5.8807, "step": 80 }, { "epoch": 0.01088, "grad_norm": 3.6022472381591797, "learning_rate": 9.999882403188902e-06, "loss": 6.1092, "step": 85 }, { "epoch": 0.01152, "grad_norm": 7.291418552398682, "learning_rate": 9.999845272899393e-06, "loss": 5.7668, "step": 90 }, { "epoch": 0.01216, "grad_norm": 3.522437810897827, "learning_rate": 9.999803056380517e-06, "loss": 6.1621, "step": 95 }, { "epoch": 0.0128, "grad_norm": 3.9014439582824707, "learning_rate": 9.999755753675216e-06, "loss": 6.0573, "step": 100 }, { "epoch": 0.0128, "eval_loss": 1.5072969198226929, "eval_runtime": 11.1161, "eval_samples_per_second": 89.96, "eval_steps_per_second": 11.245, "step": 100 }, { "epoch": 0.01344, "grad_norm": 3.7579081058502197, "learning_rate": 9.999703364831614e-06, "loss": 6.1671, "step": 105 }, { "epoch": 0.01408, "grad_norm": 3.7058262825012207, "learning_rate": 9.999645889903002e-06, "loss": 6.1348, "step": 110 }, { "epoch": 0.01472, "grad_norm": 5.018667697906494, "learning_rate": 9.99958332894785e-06, "loss": 5.9376, "step": 115 }, { "epoch": 0.01536, "grad_norm": 3.5420188903808594, "learning_rate": 9.999515682029798e-06, "loss": 5.9961, "step": 120 }, { "epoch": 0.016, "grad_norm": 3.5725393295288086, "learning_rate": 9.999442949217663e-06, "loss": 5.8439, "step": 125 }, { "epoch": 0.01664, "grad_norm": 3.8440959453582764, "learning_rate": 9.999365130585435e-06, "loss": 5.7857, "step": 130 }, { "epoch": 0.01728, "grad_norm": 3.4371285438537598, "learning_rate": 9.999282226212276e-06, "loss": 5.799, "step": 135 }, { "epoch": 0.01792, "grad_norm": 3.996847152709961, "learning_rate": 9.999194236182523e-06, "loss": 6.0022, "step": 140 }, { "epoch": 0.01856, "grad_norm": 3.720330238342285, "learning_rate": 9.999101160585687e-06, "loss": 5.925, "step": 145 }, { "epoch": 0.0192, "grad_norm": 3.8822953701019287, "learning_rate": 9.99900299951645e-06, "loss": 5.8085, "step": 150 }, { "epoch": 0.01984, "grad_norm": 3.599283456802368, "learning_rate": 9.99889975307467e-06, "loss": 5.6533, "step": 155 }, { "epoch": 0.02048, "grad_norm": 3.4847381114959717, "learning_rate": 9.998791421365376e-06, "loss": 5.9021, "step": 160 }, { "epoch": 0.02112, "grad_norm": 3.4302055835723877, "learning_rate": 9.998678004498774e-06, "loss": 5.962, "step": 165 }, { "epoch": 0.02176, "grad_norm": 4.561929702758789, "learning_rate": 9.99855950259024e-06, "loss": 5.9011, "step": 170 }, { "epoch": 0.0224, "grad_norm": 4.069271087646484, "learning_rate": 9.998435915760323e-06, "loss": 5.6782, "step": 175 }, { "epoch": 0.02304, "grad_norm": 3.5959055423736572, "learning_rate": 9.998307244134741e-06, "loss": 5.8107, "step": 180 }, { "epoch": 0.02368, "grad_norm": 3.5477242469787598, "learning_rate": 9.998173487844396e-06, "loss": 5.8335, "step": 185 }, { "epoch": 0.02432, "grad_norm": 4.488218307495117, "learning_rate": 9.998034647025349e-06, "loss": 5.8285, "step": 190 }, { "epoch": 0.02496, "grad_norm": 3.555074691772461, "learning_rate": 9.997890721818844e-06, "loss": 5.817, "step": 195 }, { "epoch": 0.0256, "grad_norm": 3.6248419284820557, "learning_rate": 9.99774171237129e-06, "loss": 5.8368, "step": 200 }, { "epoch": 0.0256, "eval_loss": 1.440572738647461, "eval_runtime": 6.6468, "eval_samples_per_second": 150.448, "eval_steps_per_second": 18.806, "step": 200 }, { "epoch": 0.02624, "grad_norm": 3.432421922683716, "learning_rate": 9.997587618834272e-06, "loss": 5.7842, "step": 205 }, { "epoch": 0.02688, "grad_norm": 3.333038806915283, "learning_rate": 9.997428441364546e-06, "loss": 5.7173, "step": 210 }, { "epoch": 0.02752, "grad_norm": 3.7716541290283203, "learning_rate": 9.997264180124038e-06, "loss": 5.719, "step": 215 }, { "epoch": 0.02816, "grad_norm": 3.345600128173828, "learning_rate": 9.99709483527985e-06, "loss": 5.8428, "step": 220 }, { "epoch": 0.0288, "grad_norm": 3.7677502632141113, "learning_rate": 9.99692040700425e-06, "loss": 5.7393, "step": 225 }, { "epoch": 0.02944, "grad_norm": 11.996383666992188, "learning_rate": 9.996740895474682e-06, "loss": 5.5566, "step": 230 }, { "epoch": 0.03008, "grad_norm": 3.6089084148406982, "learning_rate": 9.996556300873758e-06, "loss": 5.6939, "step": 235 }, { "epoch": 0.03072, "grad_norm": 3.834825038909912, "learning_rate": 9.996366623389263e-06, "loss": 5.8123, "step": 240 }, { "epoch": 0.03136, "grad_norm": 3.570263147354126, "learning_rate": 9.99617186321415e-06, "loss": 5.6839, "step": 245 }, { "epoch": 0.032, "grad_norm": 3.5728812217712402, "learning_rate": 9.995972020546545e-06, "loss": 5.7764, "step": 250 }, { "epoch": 0.03264, "grad_norm": 3.4725637435913086, "learning_rate": 9.995767095589743e-06, "loss": 5.6879, "step": 255 }, { "epoch": 0.03328, "grad_norm": 3.811537742614746, "learning_rate": 9.99555708855221e-06, "loss": 5.6418, "step": 260 }, { "epoch": 0.03392, "grad_norm": 3.494992971420288, "learning_rate": 9.99534199964758e-06, "loss": 5.6927, "step": 265 }, { "epoch": 0.03456, "grad_norm": 3.8107383251190186, "learning_rate": 9.995121829094662e-06, "loss": 5.5658, "step": 270 }, { "epoch": 0.0352, "grad_norm": 3.570551633834839, "learning_rate": 9.994896577117425e-06, "loss": 5.8131, "step": 275 }, { "epoch": 0.03584, "grad_norm": 3.540811538696289, "learning_rate": 9.994666243945018e-06, "loss": 5.6009, "step": 280 }, { "epoch": 0.03648, "grad_norm": 3.7275819778442383, "learning_rate": 9.99443082981175e-06, "loss": 5.6407, "step": 285 }, { "epoch": 0.03712, "grad_norm": 4.194495677947998, "learning_rate": 9.994190334957103e-06, "loss": 5.8319, "step": 290 }, { "epoch": 0.03776, "grad_norm": 3.5107626914978027, "learning_rate": 9.993944759625728e-06, "loss": 5.5765, "step": 295 }, { "epoch": 0.0384, "grad_norm": 3.4100208282470703, "learning_rate": 9.993694104067444e-06, "loss": 5.7473, "step": 300 }, { "epoch": 0.0384, "eval_loss": 1.407908320426941, "eval_runtime": 6.6542, "eval_samples_per_second": 150.281, "eval_steps_per_second": 18.785, "step": 300 }, { "epoch": 0.03904, "grad_norm": 3.7727818489074707, "learning_rate": 9.993438368537236e-06, "loss": 5.6802, "step": 305 }, { "epoch": 0.03968, "grad_norm": 3.445909023284912, "learning_rate": 9.993177553295258e-06, "loss": 5.7484, "step": 310 }, { "epoch": 0.04032, "grad_norm": 3.4199888706207275, "learning_rate": 9.992911658606832e-06, "loss": 5.7648, "step": 315 }, { "epoch": 0.04096, "grad_norm": 4.9640655517578125, "learning_rate": 9.992640684742445e-06, "loss": 5.7922, "step": 320 }, { "epoch": 0.0416, "grad_norm": 3.3730976581573486, "learning_rate": 9.992364631977754e-06, "loss": 5.677, "step": 325 }, { "epoch": 0.04224, "grad_norm": 3.540597915649414, "learning_rate": 9.99208350059358e-06, "loss": 5.5495, "step": 330 }, { "epoch": 0.04288, "grad_norm": 3.6853768825531006, "learning_rate": 9.991797290875915e-06, "loss": 5.4089, "step": 335 }, { "epoch": 0.04352, "grad_norm": 3.6380045413970947, "learning_rate": 9.991506003115911e-06, "loss": 5.4849, "step": 340 }, { "epoch": 0.04416, "grad_norm": 3.265488862991333, "learning_rate": 9.991209637609887e-06, "loss": 5.523, "step": 345 }, { "epoch": 0.0448, "grad_norm": 3.2634189128875732, "learning_rate": 9.990908194659332e-06, "loss": 5.5664, "step": 350 }, { "epoch": 0.04544, "grad_norm": 3.569810152053833, "learning_rate": 9.990601674570895e-06, "loss": 5.5059, "step": 355 }, { "epoch": 0.04608, "grad_norm": 3.580211877822876, "learning_rate": 9.990290077656393e-06, "loss": 5.4079, "step": 360 }, { "epoch": 0.04672, "grad_norm": 3.4860317707061768, "learning_rate": 9.989973404232805e-06, "loss": 5.6858, "step": 365 }, { "epoch": 0.04736, "grad_norm": 4.026730060577393, "learning_rate": 9.989651654622277e-06, "loss": 5.5662, "step": 370 }, { "epoch": 0.048, "grad_norm": 3.364692449569702, "learning_rate": 9.989324829152119e-06, "loss": 5.5304, "step": 375 }, { "epoch": 0.04864, "grad_norm": 3.611964464187622, "learning_rate": 9.9889929281548e-06, "loss": 5.3911, "step": 380 }, { "epoch": 0.04928, "grad_norm": 3.2946035861968994, "learning_rate": 9.988655951967958e-06, "loss": 5.4102, "step": 385 }, { "epoch": 0.04992, "grad_norm": 3.963909864425659, "learning_rate": 9.98831390093439e-06, "loss": 5.549, "step": 390 }, { "epoch": 0.05056, "grad_norm": 3.2876341342926025, "learning_rate": 9.987966775402056e-06, "loss": 5.5388, "step": 395 }, { "epoch": 0.0512, "grad_norm": 3.8467471599578857, "learning_rate": 9.98761457572408e-06, "loss": 5.454, "step": 400 }, { "epoch": 0.0512, "eval_loss": 1.3826359510421753, "eval_runtime": 7.0199, "eval_samples_per_second": 142.452, "eval_steps_per_second": 17.807, "step": 400 }, { "epoch": 0.05184, "grad_norm": 3.675231695175171, "learning_rate": 9.987257302258748e-06, "loss": 5.674, "step": 405 }, { "epoch": 0.05248, "grad_norm": 3.787940263748169, "learning_rate": 9.986894955369504e-06, "loss": 5.5466, "step": 410 }, { "epoch": 0.05312, "grad_norm": 3.677966833114624, "learning_rate": 9.986527535424956e-06, "loss": 5.4762, "step": 415 }, { "epoch": 0.05376, "grad_norm": 3.5083606243133545, "learning_rate": 9.986155042798874e-06, "loss": 5.3145, "step": 420 }, { "epoch": 0.0544, "grad_norm": 3.536379098892212, "learning_rate": 9.98577747787018e-06, "loss": 5.3769, "step": 425 }, { "epoch": 0.05504, "grad_norm": 3.5448412895202637, "learning_rate": 9.98539484102297e-06, "loss": 5.3996, "step": 430 }, { "epoch": 0.05568, "grad_norm": 3.359647274017334, "learning_rate": 9.985007132646489e-06, "loss": 5.3114, "step": 435 }, { "epoch": 0.05632, "grad_norm": 3.3419110774993896, "learning_rate": 9.984614353135143e-06, "loss": 5.4383, "step": 440 }, { "epoch": 0.05696, "grad_norm": 3.558025360107422, "learning_rate": 9.984216502888496e-06, "loss": 5.5239, "step": 445 }, { "epoch": 0.0576, "grad_norm": 3.6349422931671143, "learning_rate": 9.983813582311277e-06, "loss": 5.5639, "step": 450 }, { "epoch": 0.05824, "grad_norm": 3.2916922569274902, "learning_rate": 9.983405591813362e-06, "loss": 5.3886, "step": 455 }, { "epoch": 0.05888, "grad_norm": 3.32891845703125, "learning_rate": 9.982992531809796e-06, "loss": 5.526, "step": 460 }, { "epoch": 0.05952, "grad_norm": 3.8752880096435547, "learning_rate": 9.982574402720773e-06, "loss": 5.6599, "step": 465 }, { "epoch": 0.06016, "grad_norm": 3.604433536529541, "learning_rate": 9.982151204971646e-06, "loss": 5.4567, "step": 470 }, { "epoch": 0.0608, "grad_norm": 3.3058159351348877, "learning_rate": 9.981722938992926e-06, "loss": 5.4981, "step": 475 }, { "epoch": 0.06144, "grad_norm": 3.7341926097869873, "learning_rate": 9.981289605220276e-06, "loss": 5.3278, "step": 480 }, { "epoch": 0.06208, "grad_norm": 3.51798415184021, "learning_rate": 9.980851204094519e-06, "loss": 5.5029, "step": 485 }, { "epoch": 0.06272, "grad_norm": 3.6541428565979004, "learning_rate": 9.980407736061629e-06, "loss": 5.3987, "step": 490 }, { "epoch": 0.06336, "grad_norm": 3.420767307281494, "learning_rate": 9.979959201572736e-06, "loss": 5.405, "step": 495 }, { "epoch": 0.064, "grad_norm": 3.7169559001922607, "learning_rate": 9.979505601084124e-06, "loss": 5.498, "step": 500 }, { "epoch": 0.064, "eval_loss": 1.3493109941482544, "eval_runtime": 7.1309, "eval_samples_per_second": 140.234, "eval_steps_per_second": 17.529, "step": 500 }, { "epoch": 0.06464, "grad_norm": 4.536627769470215, "learning_rate": 9.97904693505723e-06, "loss": 5.5237, "step": 505 }, { "epoch": 0.06528, "grad_norm": 3.204948902130127, "learning_rate": 9.978583203958649e-06, "loss": 5.3746, "step": 510 }, { "epoch": 0.06592, "grad_norm": 3.4658005237579346, "learning_rate": 9.978114408260118e-06, "loss": 5.4567, "step": 515 }, { "epoch": 0.06656, "grad_norm": 4.932333469390869, "learning_rate": 9.977640548438534e-06, "loss": 5.1959, "step": 520 }, { "epoch": 0.0672, "grad_norm": 3.4697563648223877, "learning_rate": 9.977161624975948e-06, "loss": 5.4013, "step": 525 }, { "epoch": 0.06784, "grad_norm": 3.441819667816162, "learning_rate": 9.976677638359553e-06, "loss": 5.4899, "step": 530 }, { "epoch": 0.06848, "grad_norm": 3.4293930530548096, "learning_rate": 9.9761885890817e-06, "loss": 5.3569, "step": 535 }, { "epoch": 0.06912, "grad_norm": 3.5388574600219727, "learning_rate": 9.975694477639885e-06, "loss": 5.2739, "step": 540 }, { "epoch": 0.06976, "grad_norm": 3.735548973083496, "learning_rate": 9.97519530453676e-06, "loss": 5.4253, "step": 545 }, { "epoch": 0.0704, "grad_norm": 3.33503794670105, "learning_rate": 9.974691070280121e-06, "loss": 5.1569, "step": 550 }, { "epoch": 0.07104, "grad_norm": 3.5171401500701904, "learning_rate": 9.974181775382915e-06, "loss": 5.3242, "step": 555 }, { "epoch": 0.07168, "grad_norm": 3.565356969833374, "learning_rate": 9.973667420363233e-06, "loss": 5.3893, "step": 560 }, { "epoch": 0.07232, "grad_norm": 3.172163248062134, "learning_rate": 9.973148005744319e-06, "loss": 5.3824, "step": 565 }, { "epoch": 0.07296, "grad_norm": 3.517838716506958, "learning_rate": 9.972623532054564e-06, "loss": 5.2673, "step": 570 }, { "epoch": 0.0736, "grad_norm": 3.328416585922241, "learning_rate": 9.9720939998275e-06, "loss": 5.2649, "step": 575 }, { "epoch": 0.07424, "grad_norm": 3.475539445877075, "learning_rate": 9.971559409601807e-06, "loss": 5.3318, "step": 580 }, { "epoch": 0.07488, "grad_norm": 3.492013692855835, "learning_rate": 9.971019761921317e-06, "loss": 5.2735, "step": 585 }, { "epoch": 0.07552, "grad_norm": 3.474803924560547, "learning_rate": 9.970475057334997e-06, "loss": 5.3722, "step": 590 }, { "epoch": 0.07616, "grad_norm": 3.4162726402282715, "learning_rate": 9.96992529639696e-06, "loss": 5.3901, "step": 595 }, { "epoch": 0.0768, "grad_norm": 3.3643155097961426, "learning_rate": 9.969370479666473e-06, "loss": 5.2384, "step": 600 }, { "epoch": 0.0768, "eval_loss": 1.3373793363571167, "eval_runtime": 6.5847, "eval_samples_per_second": 151.867, "eval_steps_per_second": 18.983, "step": 600 }, { "epoch": 0.07744, "grad_norm": 3.44301176071167, "learning_rate": 9.968810607707933e-06, "loss": 5.2322, "step": 605 }, { "epoch": 0.07808, "grad_norm": 3.422262668609619, "learning_rate": 9.968245681090887e-06, "loss": 5.1708, "step": 610 }, { "epoch": 0.07872, "grad_norm": 3.2879252433776855, "learning_rate": 9.96767570039002e-06, "loss": 5.2291, "step": 615 }, { "epoch": 0.07936, "grad_norm": 3.6026480197906494, "learning_rate": 9.967100666185163e-06, "loss": 5.4241, "step": 620 }, { "epoch": 0.08, "grad_norm": 3.3642101287841797, "learning_rate": 9.966520579061286e-06, "loss": 5.4473, "step": 625 }, { "epoch": 0.08064, "grad_norm": 3.5968470573425293, "learning_rate": 9.965935439608493e-06, "loss": 5.3982, "step": 630 }, { "epoch": 0.08128, "grad_norm": 3.352083206176758, "learning_rate": 9.96534524842204e-06, "loss": 5.3953, "step": 635 }, { "epoch": 0.08192, "grad_norm": 3.3571720123291016, "learning_rate": 9.964750006102311e-06, "loss": 5.3159, "step": 640 }, { "epoch": 0.08256, "grad_norm": 3.486246109008789, "learning_rate": 9.964149713254833e-06, "loss": 5.211, "step": 645 }, { "epoch": 0.0832, "grad_norm": 3.674906015396118, "learning_rate": 9.96354437049027e-06, "loss": 5.3374, "step": 650 }, { "epoch": 0.08384, "grad_norm": 3.590810537338257, "learning_rate": 9.962933978424426e-06, "loss": 5.2194, "step": 655 }, { "epoch": 0.08448, "grad_norm": 3.551786184310913, "learning_rate": 9.962318537678238e-06, "loss": 5.1187, "step": 660 }, { "epoch": 0.08512, "grad_norm": 3.5391581058502197, "learning_rate": 9.961698048877776e-06, "loss": 5.2001, "step": 665 }, { "epoch": 0.08576, "grad_norm": 3.6105592250823975, "learning_rate": 9.961072512654255e-06, "loss": 5.2758, "step": 670 }, { "epoch": 0.0864, "grad_norm": 3.7463858127593994, "learning_rate": 9.960441929644017e-06, "loss": 5.2137, "step": 675 }, { "epoch": 0.08704, "grad_norm": 3.9237470626831055, "learning_rate": 9.959806300488538e-06, "loss": 5.2047, "step": 680 }, { "epoch": 0.08768, "grad_norm": 3.392827272415161, "learning_rate": 9.95916562583443e-06, "loss": 5.3071, "step": 685 }, { "epoch": 0.08832, "grad_norm": 3.221484661102295, "learning_rate": 9.958519906333438e-06, "loss": 5.183, "step": 690 }, { "epoch": 0.08896, "grad_norm": 3.5143983364105225, "learning_rate": 9.957869142642437e-06, "loss": 5.3171, "step": 695 }, { "epoch": 0.0896, "grad_norm": 3.497072696685791, "learning_rate": 9.957213335423433e-06, "loss": 5.1784, "step": 700 }, { "epoch": 0.0896, "eval_loss": 1.2988511323928833, "eval_runtime": 6.9763, "eval_samples_per_second": 143.342, "eval_steps_per_second": 17.918, "step": 700 }, { "epoch": 0.09024, "grad_norm": 3.3822438716888428, "learning_rate": 9.956552485343566e-06, "loss": 5.1732, "step": 705 }, { "epoch": 0.09088, "grad_norm": 3.3949694633483887, "learning_rate": 9.955886593075101e-06, "loss": 5.2725, "step": 710 }, { "epoch": 0.09152, "grad_norm": 3.2577288150787354, "learning_rate": 9.955215659295438e-06, "loss": 5.2207, "step": 715 }, { "epoch": 0.09216, "grad_norm": 3.769519567489624, "learning_rate": 9.954539684687103e-06, "loss": 5.2152, "step": 720 }, { "epoch": 0.0928, "grad_norm": 3.3824892044067383, "learning_rate": 9.953858669937746e-06, "loss": 5.2085, "step": 725 }, { "epoch": 0.09344, "grad_norm": 3.771742105484009, "learning_rate": 9.953172615740152e-06, "loss": 5.1575, "step": 730 }, { "epoch": 0.09408, "grad_norm": 3.7706689834594727, "learning_rate": 9.952481522792226e-06, "loss": 4.9608, "step": 735 }, { "epoch": 0.09472, "grad_norm": 3.8110334873199463, "learning_rate": 9.951785391797001e-06, "loss": 5.21, "step": 740 }, { "epoch": 0.09536, "grad_norm": 3.3012993335723877, "learning_rate": 9.951084223462636e-06, "loss": 5.2475, "step": 745 }, { "epoch": 0.096, "grad_norm": 3.6353518962860107, "learning_rate": 9.950378018502415e-06, "loss": 5.0985, "step": 750 }, { "epoch": 0.09664, "grad_norm": 3.369378089904785, "learning_rate": 9.949666777634743e-06, "loss": 5.1986, "step": 755 }, { "epoch": 0.09728, "grad_norm": 3.2247676849365234, "learning_rate": 9.948950501583147e-06, "loss": 5.3192, "step": 760 }, { "epoch": 0.09792, "grad_norm": 3.6966888904571533, "learning_rate": 9.948229191076284e-06, "loss": 5.1654, "step": 765 }, { "epoch": 0.09856, "grad_norm": 3.5823962688446045, "learning_rate": 9.947502846847921e-06, "loss": 5.1351, "step": 770 }, { "epoch": 0.0992, "grad_norm": 3.5258729457855225, "learning_rate": 9.946771469636955e-06, "loss": 5.1745, "step": 775 }, { "epoch": 0.09984, "grad_norm": 3.42067813873291, "learning_rate": 9.946035060187398e-06, "loss": 5.1569, "step": 780 }, { "epoch": 0.10048, "grad_norm": 3.9832825660705566, "learning_rate": 9.945293619248383e-06, "loss": 4.9796, "step": 785 }, { "epoch": 0.10112, "grad_norm": 3.742013692855835, "learning_rate": 9.944547147574162e-06, "loss": 5.1625, "step": 790 }, { "epoch": 0.10176, "grad_norm": 3.3150367736816406, "learning_rate": 9.943795645924104e-06, "loss": 5.099, "step": 795 }, { "epoch": 0.1024, "grad_norm": 3.359069585800171, "learning_rate": 9.943039115062691e-06, "loss": 5.1877, "step": 800 }, { "epoch": 0.1024, "eval_loss": 1.2946017980575562, "eval_runtime": 7.4306, "eval_samples_per_second": 134.579, "eval_steps_per_second": 16.822, "step": 800 }, { "epoch": 0.10304, "grad_norm": 3.703000545501709, "learning_rate": 9.94227755575953e-06, "loss": 5.1581, "step": 805 }, { "epoch": 0.10368, "grad_norm": 3.5370070934295654, "learning_rate": 9.941510968789334e-06, "loss": 5.2402, "step": 810 }, { "epoch": 0.10432, "grad_norm": 3.5010828971862793, "learning_rate": 9.940739354931936e-06, "loss": 5.1828, "step": 815 }, { "epoch": 0.10496, "grad_norm": 3.4637820720672607, "learning_rate": 9.93996271497228e-06, "loss": 5.1792, "step": 820 }, { "epoch": 0.1056, "grad_norm": 3.409712076187134, "learning_rate": 9.939181049700427e-06, "loss": 5.0721, "step": 825 }, { "epoch": 0.10624, "grad_norm": 3.589414596557617, "learning_rate": 9.938394359911545e-06, "loss": 5.234, "step": 830 }, { "epoch": 0.10688, "grad_norm": 3.444977045059204, "learning_rate": 9.937602646405918e-06, "loss": 4.9763, "step": 835 }, { "epoch": 0.10752, "grad_norm": 3.3560900688171387, "learning_rate": 9.936805909988935e-06, "loss": 5.2006, "step": 840 }, { "epoch": 0.10816, "grad_norm": 3.345703601837158, "learning_rate": 9.9360041514711e-06, "loss": 5.0287, "step": 845 }, { "epoch": 0.1088, "grad_norm": 3.492363691329956, "learning_rate": 9.935197371668024e-06, "loss": 5.0908, "step": 850 }, { "epoch": 0.10944, "grad_norm": 7.459951400756836, "learning_rate": 9.934385571400425e-06, "loss": 5.1735, "step": 855 }, { "epoch": 0.11008, "grad_norm": 3.5033841133117676, "learning_rate": 9.933568751494131e-06, "loss": 5.053, "step": 860 }, { "epoch": 0.11072, "grad_norm": 3.5542259216308594, "learning_rate": 9.93274691278007e-06, "loss": 5.1463, "step": 865 }, { "epoch": 0.11136, "grad_norm": 3.3819243907928467, "learning_rate": 9.931920056094285e-06, "loss": 5.0397, "step": 870 }, { "epoch": 0.112, "grad_norm": 3.406768798828125, "learning_rate": 9.931088182277915e-06, "loss": 5.179, "step": 875 }, { "epoch": 0.11264, "grad_norm": 5.960773944854736, "learning_rate": 9.930251292177206e-06, "loss": 5.217, "step": 880 }, { "epoch": 0.11328, "grad_norm": 3.5821049213409424, "learning_rate": 9.929409386643511e-06, "loss": 5.0374, "step": 885 }, { "epoch": 0.11392, "grad_norm": 3.3204903602600098, "learning_rate": 9.928562466533279e-06, "loss": 5.1856, "step": 890 }, { "epoch": 0.11456, "grad_norm": 4.022350788116455, "learning_rate": 9.927710532708064e-06, "loss": 5.1051, "step": 895 }, { "epoch": 0.1152, "grad_norm": 3.3810718059539795, "learning_rate": 9.926853586034515e-06, "loss": 5.1691, "step": 900 }, { "epoch": 0.1152, "eval_loss": 1.2660380601882935, "eval_runtime": 6.8853, "eval_samples_per_second": 145.238, "eval_steps_per_second": 18.155, "step": 900 }, { "epoch": 0.11584, "grad_norm": 3.5757713317871094, "learning_rate": 9.92599162738439e-06, "loss": 5.1505, "step": 905 }, { "epoch": 0.11648, "grad_norm": 3.38582706451416, "learning_rate": 9.925124657634537e-06, "loss": 5.0915, "step": 910 }, { "epoch": 0.11712, "grad_norm": 3.4189300537109375, "learning_rate": 9.924252677666905e-06, "loss": 5.1992, "step": 915 }, { "epoch": 0.11776, "grad_norm": 3.4118812084198, "learning_rate": 9.92337568836854e-06, "loss": 5.1334, "step": 920 }, { "epoch": 0.1184, "grad_norm": 3.5167789459228516, "learning_rate": 9.922493690631583e-06, "loss": 5.1003, "step": 925 }, { "epoch": 0.11904, "grad_norm": 3.546893358230591, "learning_rate": 9.921606685353268e-06, "loss": 5.1346, "step": 930 }, { "epoch": 0.11968, "grad_norm": 3.1576385498046875, "learning_rate": 9.920714673435931e-06, "loss": 4.9601, "step": 935 }, { "epoch": 0.12032, "grad_norm": 3.4227495193481445, "learning_rate": 9.91981765578699e-06, "loss": 5.0087, "step": 940 }, { "epoch": 0.12096, "grad_norm": 3.4890694618225098, "learning_rate": 9.918915633318964e-06, "loss": 5.1319, "step": 945 }, { "epoch": 0.1216, "grad_norm": 3.7377865314483643, "learning_rate": 9.918008606949459e-06, "loss": 5.0618, "step": 950 }, { "epoch": 0.12224, "grad_norm": 3.793402671813965, "learning_rate": 9.917096577601172e-06, "loss": 4.9998, "step": 955 }, { "epoch": 0.12288, "grad_norm": 3.404918909072876, "learning_rate": 9.916179546201889e-06, "loss": 5.0865, "step": 960 }, { "epoch": 0.12352, "grad_norm": 3.6076908111572266, "learning_rate": 9.915257513684488e-06, "loss": 5.0004, "step": 965 }, { "epoch": 0.12416, "grad_norm": 3.631777286529541, "learning_rate": 9.914330480986932e-06, "loss": 5.2806, "step": 970 }, { "epoch": 0.1248, "grad_norm": 3.323333501815796, "learning_rate": 9.913398449052266e-06, "loss": 5.07, "step": 975 }, { "epoch": 0.12544, "grad_norm": 3.6380035877227783, "learning_rate": 9.912461418828628e-06, "loss": 5.0559, "step": 980 }, { "epoch": 0.12608, "grad_norm": 3.7685458660125732, "learning_rate": 9.911519391269238e-06, "loss": 5.0497, "step": 985 }, { "epoch": 0.12672, "grad_norm": 3.4882941246032715, "learning_rate": 9.910572367332397e-06, "loss": 5.0388, "step": 990 }, { "epoch": 0.12736, "grad_norm": 3.27787184715271, "learning_rate": 9.909620347981493e-06, "loss": 5.0285, "step": 995 }, { "epoch": 0.128, "grad_norm": 3.388284921646118, "learning_rate": 9.908663334184994e-06, "loss": 5.1426, "step": 1000 }, { "epoch": 0.128, "eval_loss": 1.2478246688842773, "eval_runtime": 9.3123, "eval_samples_per_second": 107.384, "eval_steps_per_second": 13.423, "step": 1000 }, { "epoch": 0.12864, "grad_norm": 3.4602177143096924, "learning_rate": 9.907701326916448e-06, "loss": 4.8852, "step": 1005 }, { "epoch": 0.12928, "grad_norm": 3.7464816570281982, "learning_rate": 9.906734327154481e-06, "loss": 4.9129, "step": 1010 }, { "epoch": 0.12992, "grad_norm": 6.138649940490723, "learning_rate": 9.905762335882804e-06, "loss": 5.1037, "step": 1015 }, { "epoch": 0.13056, "grad_norm": 3.5933375358581543, "learning_rate": 9.904785354090198e-06, "loss": 4.9644, "step": 1020 }, { "epoch": 0.1312, "grad_norm": 3.6777257919311523, "learning_rate": 9.903803382770528e-06, "loss": 5.0575, "step": 1025 }, { "epoch": 0.13184, "grad_norm": 3.4429285526275635, "learning_rate": 9.902816422922727e-06, "loss": 4.8722, "step": 1030 }, { "epoch": 0.13248, "grad_norm": 3.7400121688842773, "learning_rate": 9.90182447555081e-06, "loss": 4.9521, "step": 1035 }, { "epoch": 0.13312, "grad_norm": 3.2183690071105957, "learning_rate": 9.900827541663862e-06, "loss": 5.0314, "step": 1040 }, { "epoch": 0.13376, "grad_norm": 3.563539505004883, "learning_rate": 9.899825622276041e-06, "loss": 4.9471, "step": 1045 }, { "epoch": 0.1344, "grad_norm": 3.3289413452148438, "learning_rate": 9.898818718406578e-06, "loss": 5.0223, "step": 1050 }, { "epoch": 0.13504, "grad_norm": 3.3363258838653564, "learning_rate": 9.89780683107977e-06, "loss": 4.8883, "step": 1055 }, { "epoch": 0.13568, "grad_norm": 3.5950427055358887, "learning_rate": 9.896789961324991e-06, "loss": 4.9488, "step": 1060 }, { "epoch": 0.13632, "grad_norm": 3.2444112300872803, "learning_rate": 9.895768110176677e-06, "loss": 4.9408, "step": 1065 }, { "epoch": 0.13696, "grad_norm": 3.2985880374908447, "learning_rate": 9.894741278674337e-06, "loss": 4.9875, "step": 1070 }, { "epoch": 0.1376, "grad_norm": 3.474818229675293, "learning_rate": 9.89370946786254e-06, "loss": 5.0526, "step": 1075 }, { "epoch": 0.13824, "grad_norm": 4.721025466918945, "learning_rate": 9.892672678790926e-06, "loss": 5.1362, "step": 1080 }, { "epoch": 0.13888, "grad_norm": 3.84086012840271, "learning_rate": 9.891630912514197e-06, "loss": 4.9631, "step": 1085 }, { "epoch": 0.13952, "grad_norm": 3.487732172012329, "learning_rate": 9.890584170092115e-06, "loss": 4.9211, "step": 1090 }, { "epoch": 0.14016, "grad_norm": 3.398810625076294, "learning_rate": 9.889532452589512e-06, "loss": 4.9814, "step": 1095 }, { "epoch": 0.1408, "grad_norm": 3.3263680934906006, "learning_rate": 9.888475761076273e-06, "loss": 4.9985, "step": 1100 }, { "epoch": 0.1408, "eval_loss": 1.2442607879638672, "eval_runtime": 6.5582, "eval_samples_per_second": 152.481, "eval_steps_per_second": 19.06, "step": 1100 }, { "epoch": 0.14144, "grad_norm": 3.4481613636016846, "learning_rate": 9.887414096627348e-06, "loss": 5.0169, "step": 1105 }, { "epoch": 0.14208, "grad_norm": 3.2736401557922363, "learning_rate": 9.886347460322744e-06, "loss": 5.0703, "step": 1110 }, { "epoch": 0.14272, "grad_norm": 3.2973997592926025, "learning_rate": 9.885275853247526e-06, "loss": 4.9957, "step": 1115 }, { "epoch": 0.14336, "grad_norm": 3.6516940593719482, "learning_rate": 9.884199276491817e-06, "loss": 5.0162, "step": 1120 }, { "epoch": 0.144, "grad_norm": 3.1835155487060547, "learning_rate": 9.883117731150792e-06, "loss": 4.9765, "step": 1125 }, { "epoch": 0.14464, "grad_norm": 3.21928334236145, "learning_rate": 9.882031218324681e-06, "loss": 5.0611, "step": 1130 }, { "epoch": 0.14528, "grad_norm": 4.601723670959473, "learning_rate": 9.880939739118772e-06, "loss": 5.0637, "step": 1135 }, { "epoch": 0.14592, "grad_norm": 3.2973368167877197, "learning_rate": 9.879843294643402e-06, "loss": 4.9621, "step": 1140 }, { "epoch": 0.14656, "grad_norm": 3.4781899452209473, "learning_rate": 9.878741886013959e-06, "loss": 4.9482, "step": 1145 }, { "epoch": 0.1472, "grad_norm": 3.5175704956054688, "learning_rate": 9.877635514350878e-06, "loss": 4.8594, "step": 1150 }, { "epoch": 0.14784, "grad_norm": 3.4302468299865723, "learning_rate": 9.87652418077965e-06, "loss": 4.8865, "step": 1155 }, { "epoch": 0.14848, "grad_norm": 3.464651346206665, "learning_rate": 9.875407886430806e-06, "loss": 4.9922, "step": 1160 }, { "epoch": 0.14912, "grad_norm": 4.064827919006348, "learning_rate": 9.87428663243993e-06, "loss": 4.9592, "step": 1165 }, { "epoch": 0.14976, "grad_norm": 3.654902458190918, "learning_rate": 9.873160419947645e-06, "loss": 4.9286, "step": 1170 }, { "epoch": 0.1504, "grad_norm": 3.395596981048584, "learning_rate": 9.872029250099626e-06, "loss": 5.0057, "step": 1175 }, { "epoch": 0.15104, "grad_norm": 3.745281457901001, "learning_rate": 9.870893124046582e-06, "loss": 4.8671, "step": 1180 }, { "epoch": 0.15168, "grad_norm": 3.449518918991089, "learning_rate": 9.869752042944271e-06, "loss": 4.8306, "step": 1185 }, { "epoch": 0.15232, "grad_norm": 3.1926662921905518, "learning_rate": 9.868606007953487e-06, "loss": 5.0347, "step": 1190 }, { "epoch": 0.15296, "grad_norm": 3.4620425701141357, "learning_rate": 9.86745502024007e-06, "loss": 4.857, "step": 1195 }, { "epoch": 0.1536, "grad_norm": 3.5597681999206543, "learning_rate": 9.866299080974886e-06, "loss": 4.9225, "step": 1200 }, { "epoch": 0.1536, "eval_loss": 1.2185124158859253, "eval_runtime": 7.9383, "eval_samples_per_second": 125.972, "eval_steps_per_second": 15.746, "step": 1200 }, { "epoch": 0.15424, "grad_norm": 3.5934455394744873, "learning_rate": 9.865138191333852e-06, "loss": 4.7654, "step": 1205 }, { "epoch": 0.15488, "grad_norm": 3.8588831424713135, "learning_rate": 9.863972352497912e-06, "loss": 4.9993, "step": 1210 }, { "epoch": 0.15552, "grad_norm": 3.58868408203125, "learning_rate": 9.86280156565305e-06, "loss": 4.8217, "step": 1215 }, { "epoch": 0.15616, "grad_norm": 3.5407521724700928, "learning_rate": 9.861625831990278e-06, "loss": 4.875, "step": 1220 }, { "epoch": 0.1568, "grad_norm": 3.4974656105041504, "learning_rate": 9.860445152705644e-06, "loss": 5.0627, "step": 1225 }, { "epoch": 0.15744, "grad_norm": 3.655677556991577, "learning_rate": 9.859259529000228e-06, "loss": 4.8015, "step": 1230 }, { "epoch": 0.15808, "grad_norm": 3.55148983001709, "learning_rate": 9.858068962080136e-06, "loss": 5.1209, "step": 1235 }, { "epoch": 0.15872, "grad_norm": 3.4331536293029785, "learning_rate": 9.856873453156506e-06, "loss": 4.9739, "step": 1240 }, { "epoch": 0.15936, "grad_norm": 3.374394655227661, "learning_rate": 9.855673003445502e-06, "loss": 4.8138, "step": 1245 }, { "epoch": 0.16, "grad_norm": 3.5296385288238525, "learning_rate": 9.854467614168315e-06, "loss": 5.0274, "step": 1250 }, { "epoch": 0.16064, "grad_norm": 3.6533989906311035, "learning_rate": 9.85325728655116e-06, "loss": 4.9979, "step": 1255 }, { "epoch": 0.16128, "grad_norm": 3.3504199981689453, "learning_rate": 9.852042021825272e-06, "loss": 4.8317, "step": 1260 }, { "epoch": 0.16192, "grad_norm": 3.614529609680176, "learning_rate": 9.850821821226918e-06, "loss": 4.9413, "step": 1265 }, { "epoch": 0.16256, "grad_norm": 3.4821839332580566, "learning_rate": 9.849596685997376e-06, "loss": 4.904, "step": 1270 }, { "epoch": 0.1632, "grad_norm": 3.3400087356567383, "learning_rate": 9.848366617382951e-06, "loss": 4.9039, "step": 1275 }, { "epoch": 0.16384, "grad_norm": 4.062397003173828, "learning_rate": 9.847131616634963e-06, "loss": 4.7378, "step": 1280 }, { "epoch": 0.16448, "grad_norm": 3.689796209335327, "learning_rate": 9.845891685009751e-06, "loss": 4.8799, "step": 1285 }, { "epoch": 0.16512, "grad_norm": 3.509657621383667, "learning_rate": 9.84464682376867e-06, "loss": 4.8513, "step": 1290 }, { "epoch": 0.16576, "grad_norm": 3.4828646183013916, "learning_rate": 9.843397034178088e-06, "loss": 5.0151, "step": 1295 }, { "epoch": 0.1664, "grad_norm": 3.394510507583618, "learning_rate": 9.842142317509387e-06, "loss": 4.7585, "step": 1300 }, { "epoch": 0.1664, "eval_loss": 1.2179418802261353, "eval_runtime": 6.7952, "eval_samples_per_second": 147.163, "eval_steps_per_second": 18.395, "step": 1300 }, { "epoch": 0.16704, "grad_norm": 3.4089293479919434, "learning_rate": 9.840882675038962e-06, "loss": 4.7646, "step": 1305 }, { "epoch": 0.16768, "grad_norm": 3.1607353687286377, "learning_rate": 9.83961810804822e-06, "loss": 4.9528, "step": 1310 }, { "epoch": 0.16832, "grad_norm": 3.30869197845459, "learning_rate": 9.838348617823573e-06, "loss": 5.0086, "step": 1315 }, { "epoch": 0.16896, "grad_norm": 3.6550564765930176, "learning_rate": 9.837074205656452e-06, "loss": 4.8675, "step": 1320 }, { "epoch": 0.1696, "grad_norm": 3.6141419410705566, "learning_rate": 9.835794872843281e-06, "loss": 4.8885, "step": 1325 }, { "epoch": 0.17024, "grad_norm": 3.4006361961364746, "learning_rate": 9.834510620685497e-06, "loss": 4.7784, "step": 1330 }, { "epoch": 0.17088, "grad_norm": 3.4397149085998535, "learning_rate": 9.833221450489543e-06, "loss": 4.929, "step": 1335 }, { "epoch": 0.17152, "grad_norm": 3.613502025604248, "learning_rate": 9.83192736356686e-06, "loss": 4.8763, "step": 1340 }, { "epoch": 0.17216, "grad_norm": 3.613837957382202, "learning_rate": 9.830628361233896e-06, "loss": 4.8765, "step": 1345 }, { "epoch": 0.1728, "grad_norm": 3.775621175765991, "learning_rate": 9.829324444812096e-06, "loss": 4.8103, "step": 1350 }, { "epoch": 0.17344, "grad_norm": 3.6856908798217773, "learning_rate": 9.828015615627904e-06, "loss": 4.8867, "step": 1355 }, { "epoch": 0.17408, "grad_norm": 3.3510427474975586, "learning_rate": 9.826701875012763e-06, "loss": 4.7708, "step": 1360 }, { "epoch": 0.17472, "grad_norm": 3.342366933822632, "learning_rate": 9.82538322430311e-06, "loss": 4.8404, "step": 1365 }, { "epoch": 0.17536, "grad_norm": 3.5898385047912598, "learning_rate": 9.824059664840378e-06, "loss": 4.8205, "step": 1370 }, { "epoch": 0.176, "grad_norm": 3.1588313579559326, "learning_rate": 9.822731197970998e-06, "loss": 4.7214, "step": 1375 }, { "epoch": 0.17664, "grad_norm": 3.431478261947632, "learning_rate": 9.821397825046387e-06, "loss": 4.8892, "step": 1380 }, { "epoch": 0.17728, "grad_norm": 3.7104616165161133, "learning_rate": 9.820059547422952e-06, "loss": 4.8027, "step": 1385 }, { "epoch": 0.17792, "grad_norm": 3.189239263534546, "learning_rate": 9.818716366462098e-06, "loss": 4.8692, "step": 1390 }, { "epoch": 0.17856, "grad_norm": 3.3543105125427246, "learning_rate": 9.81736828353021e-06, "loss": 4.9076, "step": 1395 }, { "epoch": 0.1792, "grad_norm": 3.2962117195129395, "learning_rate": 9.816015299998663e-06, "loss": 4.93, "step": 1400 }, { "epoch": 0.1792, "eval_loss": 1.2212570905685425, "eval_runtime": 7.0462, "eval_samples_per_second": 141.92, "eval_steps_per_second": 17.74, "step": 1400 }, { "epoch": 0.17984, "grad_norm": 3.2857654094696045, "learning_rate": 9.814657417243814e-06, "loss": 4.7544, "step": 1405 }, { "epoch": 0.18048, "grad_norm": 3.31211256980896, "learning_rate": 9.813294636647009e-06, "loss": 4.9007, "step": 1410 }, { "epoch": 0.18112, "grad_norm": 3.3026342391967773, "learning_rate": 9.81192695959457e-06, "loss": 4.8136, "step": 1415 }, { "epoch": 0.18176, "grad_norm": 3.6015031337738037, "learning_rate": 9.810554387477812e-06, "loss": 4.8296, "step": 1420 }, { "epoch": 0.1824, "grad_norm": 3.5558950901031494, "learning_rate": 9.809176921693013e-06, "loss": 4.9049, "step": 1425 }, { "epoch": 0.18304, "grad_norm": 3.272860288619995, "learning_rate": 9.807794563641442e-06, "loss": 4.868, "step": 1430 }, { "epoch": 0.18368, "grad_norm": 3.427809715270996, "learning_rate": 9.806407314729341e-06, "loss": 4.7899, "step": 1435 }, { "epoch": 0.18432, "grad_norm": 3.545553207397461, "learning_rate": 9.805015176367924e-06, "loss": 4.9774, "step": 1440 }, { "epoch": 0.18496, "grad_norm": 3.5434036254882812, "learning_rate": 9.803618149973383e-06, "loss": 4.8174, "step": 1445 }, { "epoch": 0.1856, "grad_norm": 3.5401341915130615, "learning_rate": 9.802216236966882e-06, "loss": 4.8138, "step": 1450 }, { "epoch": 0.18624, "grad_norm": 3.339459180831909, "learning_rate": 9.800809438774557e-06, "loss": 4.9385, "step": 1455 }, { "epoch": 0.18688, "grad_norm": 3.541703224182129, "learning_rate": 9.799397756827508e-06, "loss": 4.8764, "step": 1460 }, { "epoch": 0.18752, "grad_norm": 3.3053269386291504, "learning_rate": 9.79798119256181e-06, "loss": 4.5765, "step": 1465 }, { "epoch": 0.18816, "grad_norm": 3.461660146713257, "learning_rate": 9.7965597474185e-06, "loss": 4.6125, "step": 1470 }, { "epoch": 0.1888, "grad_norm": 3.564030885696411, "learning_rate": 9.795133422843583e-06, "loss": 4.8758, "step": 1475 }, { "epoch": 0.18944, "grad_norm": 3.635293483734131, "learning_rate": 9.793702220288028e-06, "loss": 4.7954, "step": 1480 }, { "epoch": 0.19008, "grad_norm": 3.4663326740264893, "learning_rate": 9.792266141207763e-06, "loss": 4.8442, "step": 1485 }, { "epoch": 0.19072, "grad_norm": 3.556608200073242, "learning_rate": 9.790825187063677e-06, "loss": 4.8431, "step": 1490 }, { "epoch": 0.19136, "grad_norm": 3.726987838745117, "learning_rate": 9.789379359321624e-06, "loss": 4.8309, "step": 1495 }, { "epoch": 0.192, "grad_norm": 3.535627603530884, "learning_rate": 9.78792865945241e-06, "loss": 4.8779, "step": 1500 }, { "epoch": 0.192, "eval_loss": 1.2059489488601685, "eval_runtime": 6.8452, "eval_samples_per_second": 146.087, "eval_steps_per_second": 18.261, "step": 1500 }, { "epoch": 0.19264, "grad_norm": 3.3409645557403564, "learning_rate": 9.7864730889318e-06, "loss": 4.8398, "step": 1505 }, { "epoch": 0.19328, "grad_norm": 3.240247964859009, "learning_rate": 9.78501264924051e-06, "loss": 4.689, "step": 1510 }, { "epoch": 0.19392, "grad_norm": 3.6355326175689697, "learning_rate": 9.783547341864216e-06, "loss": 4.7737, "step": 1515 }, { "epoch": 0.19456, "grad_norm": 3.4650771617889404, "learning_rate": 9.78207716829354e-06, "loss": 4.7844, "step": 1520 }, { "epoch": 0.1952, "grad_norm": 3.281463146209717, "learning_rate": 9.780602130024055e-06, "loss": 4.6872, "step": 1525 }, { "epoch": 0.19584, "grad_norm": 3.264622926712036, "learning_rate": 9.779122228556289e-06, "loss": 4.7438, "step": 1530 }, { "epoch": 0.19648, "grad_norm": 3.598848342895508, "learning_rate": 9.777637465395706e-06, "loss": 4.6983, "step": 1535 }, { "epoch": 0.19712, "grad_norm": 3.3951942920684814, "learning_rate": 9.776147842052725e-06, "loss": 4.8429, "step": 1540 }, { "epoch": 0.19776, "grad_norm": 3.088014841079712, "learning_rate": 9.774653360042706e-06, "loss": 4.8207, "step": 1545 }, { "epoch": 0.1984, "grad_norm": 3.4452457427978516, "learning_rate": 9.773154020885953e-06, "loss": 4.8426, "step": 1550 }, { "epoch": 0.19904, "grad_norm": 3.3782291412353516, "learning_rate": 9.771649826107707e-06, "loss": 4.9081, "step": 1555 }, { "epoch": 0.19968, "grad_norm": 3.420620918273926, "learning_rate": 9.770140777238153e-06, "loss": 4.8296, "step": 1560 }, { "epoch": 0.20032, "grad_norm": 3.3439509868621826, "learning_rate": 9.76862687581241e-06, "loss": 4.626, "step": 1565 }, { "epoch": 0.20096, "grad_norm": 3.2657105922698975, "learning_rate": 9.76710812337054e-06, "loss": 4.8035, "step": 1570 }, { "epoch": 0.2016, "grad_norm": 3.4240477085113525, "learning_rate": 9.765584521457533e-06, "loss": 4.776, "step": 1575 }, { "epoch": 0.20224, "grad_norm": 3.7116453647613525, "learning_rate": 9.764056071623314e-06, "loss": 4.8099, "step": 1580 }, { "epoch": 0.20288, "grad_norm": 3.3470919132232666, "learning_rate": 9.762522775422741e-06, "loss": 4.6686, "step": 1585 }, { "epoch": 0.20352, "grad_norm": 3.552156925201416, "learning_rate": 9.760984634415602e-06, "loss": 4.7256, "step": 1590 }, { "epoch": 0.20416, "grad_norm": 3.144547939300537, "learning_rate": 9.759441650166612e-06, "loss": 4.6914, "step": 1595 }, { "epoch": 0.2048, "grad_norm": 3.3078038692474365, "learning_rate": 9.757893824245414e-06, "loss": 4.7828, "step": 1600 }, { "epoch": 0.2048, "eval_loss": 1.1752163171768188, "eval_runtime": 6.6642, "eval_samples_per_second": 150.056, "eval_steps_per_second": 18.757, "step": 1600 } ], "logging_steps": 5, "max_steps": 15624, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 5.585434243497984e+17, "train_batch_size": 32, "trial_name": null, "trial_params": null }