|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.97561622051418, |
|
"eval_steps": 0, |
|
"global_step": 90000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.010943206027150154, |
|
"learning_rate": 1.6563604240282684e-06, |
|
"loss": 0.0168, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.010196206159889698, |
|
"learning_rate": 3.312720848056537e-06, |
|
"loss": 0.0139, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.009697528555989265, |
|
"learning_rate": 4.969081272084806e-06, |
|
"loss": 0.0142, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.011516646482050419, |
|
"learning_rate": 6.625441696113074e-06, |
|
"loss": 0.0139, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.01910693757236004, |
|
"learning_rate": 8.281802120141344e-06, |
|
"loss": 0.0137, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.004923399072140455, |
|
"learning_rate": 9.938162544169612e-06, |
|
"loss": 0.0139, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.007313380483537912, |
|
"learning_rate": 1.159452296819788e-05, |
|
"loss": 0.0137, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.003835548646748066, |
|
"learning_rate": 1.3250883392226147e-05, |
|
"loss": 0.0139, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.007223702035844326, |
|
"learning_rate": 1.4907243816254417e-05, |
|
"loss": 0.0136, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.010532427579164505, |
|
"learning_rate": 1.6563604240282687e-05, |
|
"loss": 0.0135, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.0022224171552807093, |
|
"learning_rate": 1.8219964664310956e-05, |
|
"loss": 0.0137, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.0046217963099479675, |
|
"learning_rate": 1.9876325088339224e-05, |
|
"loss": 0.0138, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.007814540527760983, |
|
"learning_rate": 2.1532685512367493e-05, |
|
"loss": 0.0136, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.011144978925585747, |
|
"learning_rate": 2.318904593639576e-05, |
|
"loss": 0.0137, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.0014108135364949703, |
|
"learning_rate": 2.484540636042403e-05, |
|
"loss": 0.0138, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.0039382693357765675, |
|
"learning_rate": 2.6501766784452294e-05, |
|
"loss": 0.0135, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.014407187700271606, |
|
"learning_rate": 2.8158127208480566e-05, |
|
"loss": 0.0138, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.003938812296837568, |
|
"learning_rate": 2.9814487632508834e-05, |
|
"loss": 0.0138, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.0075125317089259624, |
|
"learning_rate": 2.983655639540591e-05, |
|
"loss": 0.0141, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.004186388570815325, |
|
"learning_rate": 2.9652498282124278e-05, |
|
"loss": 0.0139, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.004882327280938625, |
|
"learning_rate": 2.9468440168842645e-05, |
|
"loss": 0.0139, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.002360760699957609, |
|
"learning_rate": 2.9284382055561012e-05, |
|
"loss": 0.0138, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.003932056948542595, |
|
"learning_rate": 2.910032394227938e-05, |
|
"loss": 0.0141, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.015781084075570107, |
|
"learning_rate": 2.8916265828997743e-05, |
|
"loss": 0.0138, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.002987222746014595, |
|
"learning_rate": 2.873220771571611e-05, |
|
"loss": 0.0138, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.002478894544765353, |
|
"learning_rate": 2.8548149602434473e-05, |
|
"loss": 0.0138, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.00281110149808228, |
|
"learning_rate": 2.836409148915284e-05, |
|
"loss": 0.0138, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.0016416048165410757, |
|
"learning_rate": 2.8180033375871207e-05, |
|
"loss": 0.0136, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.003159256186336279, |
|
"learning_rate": 2.7995975262589574e-05, |
|
"loss": 0.0139, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.03868310526013374, |
|
"learning_rate": 2.781191714930794e-05, |
|
"loss": 0.0151, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.026466334238648415, |
|
"learning_rate": 2.7627859036026308e-05, |
|
"loss": 0.019, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.014158655889332294, |
|
"learning_rate": 2.7443800922744675e-05, |
|
"loss": 0.0184, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.0531819723546505, |
|
"learning_rate": 2.7259742809463042e-05, |
|
"loss": 0.018, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.031152892857789993, |
|
"learning_rate": 2.707568469618141e-05, |
|
"loss": 0.0163, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.045258529484272, |
|
"learning_rate": 2.6891626582899776e-05, |
|
"loss": 0.0164, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.08072955161333084, |
|
"learning_rate": 2.6707568469618143e-05, |
|
"loss": 0.0158, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.06747995316982269, |
|
"learning_rate": 2.652351035633651e-05, |
|
"loss": 0.0155, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.029322072863578796, |
|
"learning_rate": 2.6339452243054877e-05, |
|
"loss": 0.0151, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.03149860352277756, |
|
"learning_rate": 2.615539412977324e-05, |
|
"loss": 0.0151, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.03845517709851265, |
|
"learning_rate": 2.5971336016491608e-05, |
|
"loss": 0.0152, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.02114521712064743, |
|
"learning_rate": 2.578727790320997e-05, |
|
"loss": 0.0152, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.04469970241189003, |
|
"learning_rate": 2.560321978992834e-05, |
|
"loss": 0.0151, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.03439483791589737, |
|
"learning_rate": 2.5419161676646705e-05, |
|
"loss": 0.0148, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.016261784359812737, |
|
"learning_rate": 2.5235103563365072e-05, |
|
"loss": 0.015, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.04864068329334259, |
|
"learning_rate": 2.505104545008344e-05, |
|
"loss": 0.0147, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 0.024570118635892868, |
|
"learning_rate": 2.4866987336801806e-05, |
|
"loss": 0.0149, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"grad_norm": 0.015043354593217373, |
|
"learning_rate": 2.4682929223520173e-05, |
|
"loss": 0.0151, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"grad_norm": 0.038648445159196854, |
|
"learning_rate": 2.449887111023854e-05, |
|
"loss": 0.015, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"grad_norm": 0.2623123824596405, |
|
"learning_rate": 2.4314812996956907e-05, |
|
"loss": 0.0148, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 0.02235906571149826, |
|
"learning_rate": 2.4130754883675274e-05, |
|
"loss": 0.0147, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 0.005854467861354351, |
|
"learning_rate": 2.394669677039364e-05, |
|
"loss": 0.0149, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"grad_norm": 0.011547247879207134, |
|
"learning_rate": 2.376263865711201e-05, |
|
"loss": 0.0147, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 0.03119933046400547, |
|
"learning_rate": 2.3578580543830375e-05, |
|
"loss": 0.015, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 0.047728102654218674, |
|
"learning_rate": 2.339452243054874e-05, |
|
"loss": 0.0146, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 0.04931659996509552, |
|
"learning_rate": 2.3210464317267106e-05, |
|
"loss": 0.0145, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 0.22793345153331757, |
|
"learning_rate": 2.3026406203985473e-05, |
|
"loss": 0.0147, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 0.046943288296461105, |
|
"learning_rate": 2.2842348090703837e-05, |
|
"loss": 0.0149, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.07938718795776367, |
|
"learning_rate": 2.2658289977422203e-05, |
|
"loss": 0.0147, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 0.02574564516544342, |
|
"learning_rate": 2.247423186414057e-05, |
|
"loss": 0.0144, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 0.011776907369494438, |
|
"learning_rate": 2.2290173750858937e-05, |
|
"loss": 0.0147, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 0.0066869258880615234, |
|
"learning_rate": 2.2106115637577304e-05, |
|
"loss": 0.0147, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 0.010923570953309536, |
|
"learning_rate": 2.192205752429567e-05, |
|
"loss": 0.0145, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"grad_norm": 0.03816843405365944, |
|
"learning_rate": 2.173799941101404e-05, |
|
"loss": 0.0149, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 0.04863005876541138, |
|
"learning_rate": 2.1553941297732405e-05, |
|
"loss": 0.0147, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 0.09232094883918762, |
|
"learning_rate": 2.1369883184450772e-05, |
|
"loss": 0.0148, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 0.006221645046025515, |
|
"learning_rate": 2.118582507116914e-05, |
|
"loss": 0.0148, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 0.008674496784806252, |
|
"learning_rate": 2.1001766957887506e-05, |
|
"loss": 0.0145, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.013610797002911568, |
|
"learning_rate": 2.081770884460587e-05, |
|
"loss": 0.0149, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 0.020487351343035698, |
|
"learning_rate": 2.0633650731324237e-05, |
|
"loss": 0.0147, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 0.004388992674648762, |
|
"learning_rate": 2.0449592618042604e-05, |
|
"loss": 0.0146, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 0.029407095164060593, |
|
"learning_rate": 2.026553450476097e-05, |
|
"loss": 0.0144, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 0.04079248011112213, |
|
"learning_rate": 2.0081476391479335e-05, |
|
"loss": 0.0146, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 0.033315982669591904, |
|
"learning_rate": 1.98974182781977e-05, |
|
"loss": 0.0143, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 0.00864444486796856, |
|
"learning_rate": 1.971336016491607e-05, |
|
"loss": 0.0145, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 0.03435393422842026, |
|
"learning_rate": 1.9529302051634436e-05, |
|
"loss": 0.0145, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 0.008053929544985294, |
|
"learning_rate": 1.9345243938352803e-05, |
|
"loss": 0.0146, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 0.004771470092236996, |
|
"learning_rate": 1.916118582507117e-05, |
|
"loss": 0.0143, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 0.016594666987657547, |
|
"learning_rate": 1.8977127711789537e-05, |
|
"loss": 0.0149, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 0.01181800477206707, |
|
"learning_rate": 1.8793069598507904e-05, |
|
"loss": 0.0143, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 0.03508065640926361, |
|
"learning_rate": 1.860901148522627e-05, |
|
"loss": 0.0146, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 0.028093870729207993, |
|
"learning_rate": 1.8424953371944638e-05, |
|
"loss": 0.0146, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 0.05029403790831566, |
|
"learning_rate": 1.8240895258663005e-05, |
|
"loss": 0.0146, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 0.006528925616294146, |
|
"learning_rate": 1.8056837145381368e-05, |
|
"loss": 0.0142, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 0.02557162009179592, |
|
"learning_rate": 1.7872779032099735e-05, |
|
"loss": 0.0144, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 0.025360217317938805, |
|
"learning_rate": 1.7688720918818102e-05, |
|
"loss": 0.0146, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 0.04788580909371376, |
|
"learning_rate": 1.750466280553647e-05, |
|
"loss": 0.0147, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.02906920574605465, |
|
"learning_rate": 1.7320604692254836e-05, |
|
"loss": 0.0144, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 0.012823808006942272, |
|
"learning_rate": 1.71365465789732e-05, |
|
"loss": 0.0145, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 0.008996455930173397, |
|
"learning_rate": 1.6952488465691567e-05, |
|
"loss": 0.0143, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": 0.010119748301804066, |
|
"learning_rate": 1.6768430352409934e-05, |
|
"loss": 0.0146, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 0.02591855265200138, |
|
"learning_rate": 1.65843722391283e-05, |
|
"loss": 0.0143, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 0.013729671947658062, |
|
"learning_rate": 1.6400314125846668e-05, |
|
"loss": 0.0147, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 0.0771203562617302, |
|
"learning_rate": 1.6216256012565035e-05, |
|
"loss": 0.0146, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 0.04501279070973396, |
|
"learning_rate": 1.60321978992834e-05, |
|
"loss": 0.0144, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 0.03493111953139305, |
|
"learning_rate": 1.584813978600177e-05, |
|
"loss": 0.0144, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 0.01472916454076767, |
|
"learning_rate": 1.5664081672720136e-05, |
|
"loss": 0.0144, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"grad_norm": 0.04763146862387657, |
|
"learning_rate": 1.54800235594385e-05, |
|
"loss": 0.0145, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 0.024467509239912033, |
|
"learning_rate": 1.5295965446156866e-05, |
|
"loss": 0.0144, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 0.01768341101706028, |
|
"learning_rate": 1.5111907332875235e-05, |
|
"loss": 0.0144, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"grad_norm": 0.06102894991636276, |
|
"learning_rate": 1.49278492195936e-05, |
|
"loss": 0.0141, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 0.01851697266101837, |
|
"learning_rate": 1.4743791106311966e-05, |
|
"loss": 0.0142, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 0.015444310382008553, |
|
"learning_rate": 1.4559732993030333e-05, |
|
"loss": 0.0145, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 0.013120009563863277, |
|
"learning_rate": 1.43756748797487e-05, |
|
"loss": 0.0143, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"grad_norm": 0.01589464209973812, |
|
"learning_rate": 1.4191616766467067e-05, |
|
"loss": 0.0141, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 0.040490709245204926, |
|
"learning_rate": 1.4007558653185433e-05, |
|
"loss": 0.0144, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"grad_norm": 0.025874989107251167, |
|
"learning_rate": 1.38235005399038e-05, |
|
"loss": 0.0143, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 0.022394156083464622, |
|
"learning_rate": 1.3639442426622166e-05, |
|
"loss": 0.0144, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"grad_norm": 0.010273805819451809, |
|
"learning_rate": 1.3455384313340531e-05, |
|
"loss": 0.0144, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"grad_norm": 0.028374383226037025, |
|
"learning_rate": 1.3271326200058898e-05, |
|
"loss": 0.0144, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 0.018441613763570786, |
|
"learning_rate": 1.3087268086777265e-05, |
|
"loss": 0.0145, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 0.006460436619818211, |
|
"learning_rate": 1.2903209973495632e-05, |
|
"loss": 0.0145, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 0.00770485308021307, |
|
"learning_rate": 1.2719151860213999e-05, |
|
"loss": 0.0144, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.00849447026848793, |
|
"learning_rate": 1.2535093746932366e-05, |
|
"loss": 0.0147, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 0.035621609538793564, |
|
"learning_rate": 1.2351035633650733e-05, |
|
"loss": 0.0145, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"grad_norm": 0.008923010900616646, |
|
"learning_rate": 1.2166977520369098e-05, |
|
"loss": 0.0144, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 0.0056849876418709755, |
|
"learning_rate": 1.1982919407087464e-05, |
|
"loss": 0.0143, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 0.0071659935638308525, |
|
"learning_rate": 1.179886129380583e-05, |
|
"loss": 0.0144, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 0.021617043763399124, |
|
"learning_rate": 1.1614803180524198e-05, |
|
"loss": 0.0143, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"grad_norm": 0.011144719086587429, |
|
"learning_rate": 1.1430745067242565e-05, |
|
"loss": 0.0142, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"grad_norm": 0.010943782515823841, |
|
"learning_rate": 1.1246686953960932e-05, |
|
"loss": 0.0143, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"grad_norm": 0.010286700911819935, |
|
"learning_rate": 1.1062628840679299e-05, |
|
"loss": 0.0143, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"grad_norm": 0.010169615969061852, |
|
"learning_rate": 1.0878570727397666e-05, |
|
"loss": 0.0143, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"grad_norm": 0.032067082822322845, |
|
"learning_rate": 1.069451261411603e-05, |
|
"loss": 0.0144, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"grad_norm": 0.008680183440446854, |
|
"learning_rate": 1.0510454500834396e-05, |
|
"loss": 0.0143, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"grad_norm": 0.01648719422519207, |
|
"learning_rate": 1.0326396387552763e-05, |
|
"loss": 0.0143, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"grad_norm": 0.0210120789706707, |
|
"learning_rate": 1.014233827427113e-05, |
|
"loss": 0.0146, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 0.034336596727371216, |
|
"learning_rate": 9.958280160989497e-06, |
|
"loss": 0.0144, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 0.03138417750597, |
|
"learning_rate": 9.774222047707864e-06, |
|
"loss": 0.0141, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 0.01799875684082508, |
|
"learning_rate": 9.590163934426231e-06, |
|
"loss": 0.0142, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 0.02960127592086792, |
|
"learning_rate": 9.406105821144595e-06, |
|
"loss": 0.0143, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"grad_norm": 0.012712860479950905, |
|
"learning_rate": 9.222047707862962e-06, |
|
"loss": 0.0146, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 0.009180006571114063, |
|
"learning_rate": 9.037989594581329e-06, |
|
"loss": 0.0143, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 0.0106426402926445, |
|
"learning_rate": 8.853931481299696e-06, |
|
"loss": 0.0143, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 0.03638075664639473, |
|
"learning_rate": 8.669873368018063e-06, |
|
"loss": 0.0141, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 0.02028089202940464, |
|
"learning_rate": 8.48581525473643e-06, |
|
"loss": 0.0144, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 0.004987742286175489, |
|
"learning_rate": 8.301757141454797e-06, |
|
"loss": 0.0143, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"grad_norm": 0.012421207502484322, |
|
"learning_rate": 8.117699028173162e-06, |
|
"loss": 0.0145, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"grad_norm": 0.05489884316921234, |
|
"learning_rate": 7.933640914891527e-06, |
|
"loss": 0.0142, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"grad_norm": 0.007833471521735191, |
|
"learning_rate": 7.749582801609894e-06, |
|
"loss": 0.0145, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"grad_norm": 0.014776123687624931, |
|
"learning_rate": 7.565524688328261e-06, |
|
"loss": 0.0142, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"grad_norm": 0.015590249560773373, |
|
"learning_rate": 7.381466575046628e-06, |
|
"loss": 0.0143, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"grad_norm": 0.018214261159300804, |
|
"learning_rate": 7.197408461764995e-06, |
|
"loss": 0.0142, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"grad_norm": 0.029773008078336716, |
|
"learning_rate": 7.013350348483361e-06, |
|
"loss": 0.0143, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"grad_norm": 0.028139958158135414, |
|
"learning_rate": 6.8292922352017276e-06, |
|
"loss": 0.0143, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"grad_norm": 0.024558302015066147, |
|
"learning_rate": 6.6452341219200945e-06, |
|
"loss": 0.014, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"grad_norm": 0.05188705772161484, |
|
"learning_rate": 6.461176008638461e-06, |
|
"loss": 0.0141, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"grad_norm": 0.02240253984928131, |
|
"learning_rate": 6.277117895356828e-06, |
|
"loss": 0.0142, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"grad_norm": 0.015994379296898842, |
|
"learning_rate": 6.093059782075194e-06, |
|
"loss": 0.0143, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"grad_norm": 0.014095323160290718, |
|
"learning_rate": 5.909001668793561e-06, |
|
"loss": 0.0141, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"grad_norm": 0.0076615894213318825, |
|
"learning_rate": 5.724943555511927e-06, |
|
"loss": 0.0141, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"grad_norm": 0.023330098018050194, |
|
"learning_rate": 5.540885442230294e-06, |
|
"loss": 0.0143, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"grad_norm": 0.022397508844733238, |
|
"learning_rate": 5.35682732894866e-06, |
|
"loss": 0.0141, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"grad_norm": 0.01998765394091606, |
|
"learning_rate": 5.172769215667027e-06, |
|
"loss": 0.0143, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"grad_norm": 0.07193479686975479, |
|
"learning_rate": 4.988711102385393e-06, |
|
"loss": 0.0143, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"grad_norm": 0.030124777927994728, |
|
"learning_rate": 4.80465298910376e-06, |
|
"loss": 0.0146, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"grad_norm": 0.0762249082326889, |
|
"learning_rate": 4.620594875822126e-06, |
|
"loss": 0.0144, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"grad_norm": 0.030013220384716988, |
|
"learning_rate": 4.4365367625404925e-06, |
|
"loss": 0.0143, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"grad_norm": 0.013210024684667587, |
|
"learning_rate": 4.2524786492588595e-06, |
|
"loss": 0.0144, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"grad_norm": 0.021476522088050842, |
|
"learning_rate": 4.068420535977226e-06, |
|
"loss": 0.0145, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"grad_norm": 0.005120801739394665, |
|
"learning_rate": 3.884362422695593e-06, |
|
"loss": 0.0142, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"grad_norm": 0.03378542512655258, |
|
"learning_rate": 3.7003043094139592e-06, |
|
"loss": 0.0144, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"grad_norm": 0.004559422377496958, |
|
"learning_rate": 3.5162461961323254e-06, |
|
"loss": 0.0143, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"grad_norm": 0.022087154909968376, |
|
"learning_rate": 3.3321880828506924e-06, |
|
"loss": 0.0142, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"grad_norm": 0.027302134782075882, |
|
"learning_rate": 3.1481299695690585e-06, |
|
"loss": 0.0142, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"grad_norm": 0.0070667564868927, |
|
"learning_rate": 2.964071856287425e-06, |
|
"loss": 0.0142, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"grad_norm": 0.004392644390463829, |
|
"learning_rate": 2.7800137430057916e-06, |
|
"loss": 0.014, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"grad_norm": 0.004756265785545111, |
|
"learning_rate": 2.595955629724158e-06, |
|
"loss": 0.0144, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"grad_norm": 0.028167065232992172, |
|
"learning_rate": 2.4118975164425248e-06, |
|
"loss": 0.0141, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"grad_norm": 0.04241223633289337, |
|
"learning_rate": 2.2278394031608913e-06, |
|
"loss": 0.0143, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"grad_norm": 0.0073333000764250755, |
|
"learning_rate": 2.043781289879258e-06, |
|
"loss": 0.014, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"grad_norm": 0.022960776463150978, |
|
"learning_rate": 1.8597231765976245e-06, |
|
"loss": 0.0146, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"grad_norm": 0.009491208009421825, |
|
"learning_rate": 1.675665063315991e-06, |
|
"loss": 0.0143, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"grad_norm": 0.04244249686598778, |
|
"learning_rate": 1.4916069500343576e-06, |
|
"loss": 0.0142, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"grad_norm": 0.0845978856086731, |
|
"learning_rate": 1.307548836752724e-06, |
|
"loss": 0.0139, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"grad_norm": 0.012684383429586887, |
|
"learning_rate": 1.1234907234710905e-06, |
|
"loss": 0.0143, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"grad_norm": 0.010290221311151981, |
|
"learning_rate": 9.394326101894571e-07, |
|
"loss": 0.0144, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"grad_norm": 0.009122644551098347, |
|
"learning_rate": 7.553744969078238e-07, |
|
"loss": 0.0143, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"grad_norm": 0.02026693895459175, |
|
"learning_rate": 5.713163836261902e-07, |
|
"loss": 0.0142, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"grad_norm": 0.03241865336894989, |
|
"learning_rate": 3.8725827034455676e-07, |
|
"loss": 0.0142, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"grad_norm": 0.04753628000617027, |
|
"learning_rate": 2.0320015706292333e-07, |
|
"loss": 0.0142, |
|
"step": 90000 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 90552, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 10000, |
|
"total_flos": 0.0, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|