{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.03375171395422424, "eval_steps": 100, "global_step": 400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 8.437928488556059e-05, "eval_loss": 0.9957674145698547, "eval_runtime": 419.6878, "eval_samples_per_second": 11.89, "eval_steps_per_second": 5.945, "step": 1 }, { "epoch": 0.00042189642442780296, "grad_norm": 0.4106890559196472, "learning_rate": 1.6666666666666667e-05, "loss": 0.8825, "step": 5 }, { "epoch": 0.0008437928488556059, "grad_norm": 0.5586373209953308, "learning_rate": 3.3333333333333335e-05, "loss": 0.9305, "step": 10 }, { "epoch": 0.0012656892732834089, "grad_norm": 0.5272752046585083, "learning_rate": 5e-05, "loss": 0.9003, "step": 15 }, { "epoch": 0.0016875856977112118, "grad_norm": 0.5173702836036682, "learning_rate": 6.666666666666667e-05, "loss": 0.7796, "step": 20 }, { "epoch": 0.002109482122139015, "grad_norm": 0.5352601408958435, "learning_rate": 8.333333333333334e-05, "loss": 0.8669, "step": 25 }, { "epoch": 0.0025313785465668178, "grad_norm": 0.48157429695129395, "learning_rate": 0.0001, "loss": 0.835, "step": 30 }, { "epoch": 0.0029532749709946207, "grad_norm": 0.5189157724380493, "learning_rate": 9.995494831023409e-05, "loss": 0.7987, "step": 35 }, { "epoch": 0.0033751713954224237, "grad_norm": 0.5436115860939026, "learning_rate": 9.981987442712633e-05, "loss": 0.9009, "step": 40 }, { "epoch": 0.0037970678198502267, "grad_norm": 0.6219505071640015, "learning_rate": 9.959502176294383e-05, "loss": 0.8412, "step": 45 }, { "epoch": 0.00421896424427803, "grad_norm": 1.6684471368789673, "learning_rate": 9.928079551738543e-05, "loss": 0.9159, "step": 50 }, { "epoch": 0.004640860668705833, "grad_norm": 0.5012656450271606, "learning_rate": 9.887776194738432e-05, "loss": 0.7569, "step": 55 }, { "epoch": 0.0050627570931336355, "grad_norm": 0.35097789764404297, "learning_rate": 9.838664734667495e-05, "loss": 0.7247, "step": 60 }, { "epoch": 0.005484653517561439, "grad_norm": 0.4114624261856079, "learning_rate": 9.780833673696254e-05, "loss": 0.78, "step": 65 }, { "epoch": 0.0059065499419892415, "grad_norm": 0.39524176716804504, "learning_rate": 9.714387227305422e-05, "loss": 0.7716, "step": 70 }, { "epoch": 0.006328446366417045, "grad_norm": 0.39941051602363586, "learning_rate": 9.639445136482548e-05, "loss": 0.8346, "step": 75 }, { "epoch": 0.006750342790844847, "grad_norm": 0.38641512393951416, "learning_rate": 9.55614245194068e-05, "loss": 0.7287, "step": 80 }, { "epoch": 0.007172239215272651, "grad_norm": 0.5262604355812073, "learning_rate": 9.464629290747842e-05, "loss": 0.7636, "step": 85 }, { "epoch": 0.007594135639700453, "grad_norm": 0.605768084526062, "learning_rate": 9.365070565805941e-05, "loss": 0.8944, "step": 90 }, { "epoch": 0.008016032064128256, "grad_norm": 0.6336283683776855, "learning_rate": 9.257645688666556e-05, "loss": 0.8311, "step": 95 }, { "epoch": 0.00843792848855606, "grad_norm": 1.458014965057373, "learning_rate": 9.142548246219212e-05, "loss": 0.8093, "step": 100 }, { "epoch": 0.00843792848855606, "eval_loss": 0.7453532814979553, "eval_runtime": 421.5346, "eval_samples_per_second": 11.838, "eval_steps_per_second": 5.919, "step": 100 }, { "epoch": 0.008859824912983863, "grad_norm": 0.3544303774833679, "learning_rate": 9.019985651834703e-05, "loss": 0.6757, "step": 105 }, { "epoch": 0.009281721337411666, "grad_norm": 0.31981468200683594, "learning_rate": 8.890178771592199e-05, "loss": 0.7116, "step": 110 }, { "epoch": 0.009703617761839468, "grad_norm": 0.3641217350959778, "learning_rate": 8.753361526263621e-05, "loss": 0.704, "step": 115 }, { "epoch": 0.010125514186267271, "grad_norm": 0.41695696115493774, "learning_rate": 8.609780469772623e-05, "loss": 0.6933, "step": 120 }, { "epoch": 0.010547410610695074, "grad_norm": 0.40811964869499207, "learning_rate": 8.459694344887732e-05, "loss": 0.7293, "step": 125 }, { "epoch": 0.010969307035122878, "grad_norm": 0.4640657901763916, "learning_rate": 8.303373616950408e-05, "loss": 0.8195, "step": 130 }, { "epoch": 0.01139120345955068, "grad_norm": 0.4916837811470032, "learning_rate": 8.141099986478212e-05, "loss": 0.7061, "step": 135 }, { "epoch": 0.011813099883978483, "grad_norm": 0.43793752789497375, "learning_rate": 7.973165881521434e-05, "loss": 0.8276, "step": 140 }, { "epoch": 0.012234996308406286, "grad_norm": 0.6117736101150513, "learning_rate": 7.799873930687978e-05, "loss": 0.893, "step": 145 }, { "epoch": 0.01265689273283409, "grad_norm": 1.3102891445159912, "learning_rate": 7.621536417786159e-05, "loss": 0.8688, "step": 150 }, { "epoch": 0.013078789157261891, "grad_norm": 0.35164037346839905, "learning_rate": 7.438474719068173e-05, "loss": 0.5983, "step": 155 }, { "epoch": 0.013500685581689695, "grad_norm": 0.32440951466560364, "learning_rate": 7.251018724088367e-05, "loss": 0.6319, "step": 160 }, { "epoch": 0.013922582006117498, "grad_norm": 0.36062324047088623, "learning_rate": 7.059506241219965e-05, "loss": 0.6391, "step": 165 }, { "epoch": 0.014344478430545302, "grad_norm": 0.38565579056739807, "learning_rate": 6.864282388901544e-05, "loss": 0.6643, "step": 170 }, { "epoch": 0.014766374854973103, "grad_norm": 0.411285400390625, "learning_rate": 6.665698973710288e-05, "loss": 0.7558, "step": 175 }, { "epoch": 0.015188271279400907, "grad_norm": 0.3739734888076782, "learning_rate": 6.464113856382752e-05, "loss": 0.6523, "step": 180 }, { "epoch": 0.01561016770382871, "grad_norm": 0.4831923842430115, "learning_rate": 6.259890306925627e-05, "loss": 0.7802, "step": 185 }, { "epoch": 0.01603206412825651, "grad_norm": 0.5024823546409607, "learning_rate": 6.0533963499786314e-05, "loss": 0.6872, "step": 190 }, { "epoch": 0.016453960552684315, "grad_norm": 0.6557715535163879, "learning_rate": 5.8450041016092464e-05, "loss": 0.7473, "step": 195 }, { "epoch": 0.01687585697711212, "grad_norm": 1.3833485841751099, "learning_rate": 5.6350890987343944e-05, "loss": 0.8261, "step": 200 }, { "epoch": 0.01687585697711212, "eval_loss": 0.7295480370521545, "eval_runtime": 421.3959, "eval_samples_per_second": 11.842, "eval_steps_per_second": 5.921, "step": 200 }, { "epoch": 0.017297753401539922, "grad_norm": 0.4051777720451355, "learning_rate": 5.4240296223775465e-05, "loss": 0.6449, "step": 205 }, { "epoch": 0.017719649825967725, "grad_norm": 0.36194536089897156, "learning_rate": 5.212206015980742e-05, "loss": 0.7147, "step": 210 }, { "epoch": 0.01814154625039553, "grad_norm": 0.37661212682724, "learning_rate": 5e-05, "loss": 0.734, "step": 215 }, { "epoch": 0.018563442674823332, "grad_norm": 0.37295767664909363, "learning_rate": 4.78779398401926e-05, "loss": 0.6716, "step": 220 }, { "epoch": 0.018985339099251135, "grad_norm": 0.474931001663208, "learning_rate": 4.575970377622456e-05, "loss": 0.749, "step": 225 }, { "epoch": 0.019407235523678935, "grad_norm": 0.5339241027832031, "learning_rate": 4.364910901265606e-05, "loss": 0.8035, "step": 230 }, { "epoch": 0.01982913194810674, "grad_norm": 0.546904981136322, "learning_rate": 4.1549958983907555e-05, "loss": 0.7336, "step": 235 }, { "epoch": 0.020251028372534542, "grad_norm": 0.5156819820404053, "learning_rate": 3.94660365002137e-05, "loss": 0.7587, "step": 240 }, { "epoch": 0.020672924796962346, "grad_norm": 0.5965439677238464, "learning_rate": 3.740109693074375e-05, "loss": 0.9158, "step": 245 }, { "epoch": 0.02109482122139015, "grad_norm": 1.0562564134597778, "learning_rate": 3.5358861436172485e-05, "loss": 1.0234, "step": 250 }, { "epoch": 0.021516717645817952, "grad_norm": 0.33975425362586975, "learning_rate": 3.334301026289712e-05, "loss": 0.6781, "step": 255 }, { "epoch": 0.021938614070245756, "grad_norm": 0.31108027696609497, "learning_rate": 3.135717611098458e-05, "loss": 0.6653, "step": 260 }, { "epoch": 0.02236051049467356, "grad_norm": 0.3663179874420166, "learning_rate": 2.9404937587800375e-05, "loss": 0.6752, "step": 265 }, { "epoch": 0.02278240691910136, "grad_norm": 0.5073758363723755, "learning_rate": 2.748981275911633e-05, "loss": 0.7357, "step": 270 }, { "epoch": 0.023204303343529162, "grad_norm": 0.41087105870246887, "learning_rate": 2.5615252809318284e-05, "loss": 0.6715, "step": 275 }, { "epoch": 0.023626199767956966, "grad_norm": 0.4009557068347931, "learning_rate": 2.3784635822138424e-05, "loss": 0.6759, "step": 280 }, { "epoch": 0.02404809619238477, "grad_norm": 0.46906349062919617, "learning_rate": 2.2001260693120233e-05, "loss": 0.7395, "step": 285 }, { "epoch": 0.024469992616812573, "grad_norm": 0.6131988763809204, "learning_rate": 2.026834118478567e-05, "loss": 0.8374, "step": 290 }, { "epoch": 0.024891889041240376, "grad_norm": 0.735552191734314, "learning_rate": 1.858900013521788e-05, "loss": 0.9337, "step": 295 }, { "epoch": 0.02531378546566818, "grad_norm": 1.2397476434707642, "learning_rate": 1.6966263830495936e-05, "loss": 0.8693, "step": 300 }, { "epoch": 0.02531378546566818, "eval_loss": 0.718722939491272, "eval_runtime": 421.5469, "eval_samples_per_second": 11.837, "eval_steps_per_second": 5.919, "step": 300 }, { "epoch": 0.025735681890095983, "grad_norm": 0.2928362786769867, "learning_rate": 1.5403056551122697e-05, "loss": 0.6769, "step": 305 }, { "epoch": 0.026157578314523783, "grad_norm": 0.35365402698516846, "learning_rate": 1.3902195302273779e-05, "loss": 0.6254, "step": 310 }, { "epoch": 0.026579474738951586, "grad_norm": 0.3873966634273529, "learning_rate": 1.246638473736378e-05, "loss": 0.7157, "step": 315 }, { "epoch": 0.02700137116337939, "grad_norm": 0.42512309551239014, "learning_rate": 1.1098212284078036e-05, "loss": 0.7382, "step": 320 }, { "epoch": 0.027423267587807193, "grad_norm": 0.45126789808273315, "learning_rate": 9.800143481652979e-06, "loss": 0.6671, "step": 325 }, { "epoch": 0.027845164012234996, "grad_norm": 0.4876130223274231, "learning_rate": 8.574517537807897e-06, "loss": 0.7558, "step": 330 }, { "epoch": 0.0282670604366628, "grad_norm": 0.4966984689235687, "learning_rate": 7.423543113334436e-06, "loss": 0.7365, "step": 335 }, { "epoch": 0.028688956861090603, "grad_norm": 0.4963556230068207, "learning_rate": 6.349294341940593e-06, "loss": 0.7839, "step": 340 }, { "epoch": 0.029110853285518407, "grad_norm": 0.5703970193862915, "learning_rate": 5.353707092521582e-06, "loss": 0.6884, "step": 345 }, { "epoch": 0.029532749709946206, "grad_norm": 1.9548819065093994, "learning_rate": 4.43857548059321e-06, "loss": 1.012, "step": 350 }, { "epoch": 0.02995464613437401, "grad_norm": 0.31155750155448914, "learning_rate": 3.605548635174533e-06, "loss": 0.6154, "step": 355 }, { "epoch": 0.030376542558801813, "grad_norm": 0.2996324896812439, "learning_rate": 2.85612772694579e-06, "loss": 0.733, "step": 360 }, { "epoch": 0.030798438983229617, "grad_norm": 0.37856683135032654, "learning_rate": 2.191663263037458e-06, "loss": 0.6717, "step": 365 }, { "epoch": 0.03122033540765742, "grad_norm": 0.37820345163345337, "learning_rate": 1.6133526533250565e-06, "loss": 0.6931, "step": 370 }, { "epoch": 0.03164223183208522, "grad_norm": 0.39253878593444824, "learning_rate": 1.1222380526156928e-06, "loss": 0.6896, "step": 375 }, { "epoch": 0.03206412825651302, "grad_norm": 0.4378221035003662, "learning_rate": 7.192044826145771e-07, "loss": 0.752, "step": 380 }, { "epoch": 0.03248602468094083, "grad_norm": 0.5840653777122498, "learning_rate": 4.049782370561583e-07, "loss": 0.6758, "step": 385 }, { "epoch": 0.03290792110536863, "grad_norm": 0.6441401839256287, "learning_rate": 1.8012557287367392e-07, "loss": 0.7952, "step": 390 }, { "epoch": 0.033329817529796434, "grad_norm": 0.6340252161026001, "learning_rate": 4.5051689765929214e-08, "loss": 0.779, "step": 395 }, { "epoch": 0.03375171395422424, "grad_norm": 1.275892734527588, "learning_rate": 0.0, "loss": 0.9144, "step": 400 }, { "epoch": 0.03375171395422424, "eval_loss": 0.7176097631454468, "eval_runtime": 421.8457, "eval_samples_per_second": 11.829, "eval_steps_per_second": 5.914, "step": 400 } ], "logging_steps": 5, "max_steps": 400, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.5129435549833626e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }