{ "best_metric": 1.2658852338790894, "best_model_checkpoint": "data/Llama-31-8B_task-1_120-samples_config-3/checkpoint-198", "epoch": 25.0, "eval_steps": 500, "global_step": 275, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.09090909090909091, "grad_norm": 1.9471415281295776, "learning_rate": 6.060606060606061e-08, "loss": 2.2569, "step": 1 }, { "epoch": 0.18181818181818182, "grad_norm": 1.9496530294418335, "learning_rate": 1.2121212121212122e-07, "loss": 2.3228, "step": 2 }, { "epoch": 0.36363636363636365, "grad_norm": 1.717644453048706, "learning_rate": 2.4242424242424244e-07, "loss": 2.0831, "step": 4 }, { "epoch": 0.5454545454545454, "grad_norm": 1.5924347639083862, "learning_rate": 3.6363636363636366e-07, "loss": 2.0746, "step": 6 }, { "epoch": 0.7272727272727273, "grad_norm": 1.8148092031478882, "learning_rate": 4.848484848484849e-07, "loss": 2.1661, "step": 8 }, { "epoch": 0.9090909090909091, "grad_norm": 1.6654292345046997, "learning_rate": 6.060606060606061e-07, "loss": 1.9625, "step": 10 }, { "epoch": 1.0, "eval_loss": 2.0993740558624268, "eval_runtime": 9.6328, "eval_samples_per_second": 2.491, "eval_steps_per_second": 2.491, "step": 11 }, { "epoch": 1.0909090909090908, "grad_norm": 1.543703556060791, "learning_rate": 7.272727272727273e-07, "loss": 1.9298, "step": 12 }, { "epoch": 1.2727272727272727, "grad_norm": 1.8203456401824951, "learning_rate": 8.484848484848486e-07, "loss": 2.1289, "step": 14 }, { "epoch": 1.4545454545454546, "grad_norm": 1.7504804134368896, "learning_rate": 9.696969696969698e-07, "loss": 2.153, "step": 16 }, { "epoch": 1.6363636363636362, "grad_norm": 1.895839810371399, "learning_rate": 1.090909090909091e-06, "loss": 2.0811, "step": 18 }, { "epoch": 1.8181818181818183, "grad_norm": 1.7807648181915283, "learning_rate": 1.2121212121212122e-06, "loss": 2.0446, "step": 20 }, { "epoch": 2.0, "grad_norm": 1.8615657091140747, "learning_rate": 1.3333333333333334e-06, "loss": 2.1365, "step": 22 }, { "epoch": 2.0, "eval_loss": 2.0816335678100586, "eval_runtime": 9.6405, "eval_samples_per_second": 2.49, "eval_steps_per_second": 2.49, "step": 22 }, { "epoch": 2.1818181818181817, "grad_norm": 1.8219581842422485, "learning_rate": 1.4545454545454546e-06, "loss": 2.1017, "step": 24 }, { "epoch": 2.3636363636363638, "grad_norm": 1.6915614604949951, "learning_rate": 1.5757575757575759e-06, "loss": 2.0744, "step": 26 }, { "epoch": 2.5454545454545454, "grad_norm": 1.8501787185668945, "learning_rate": 1.6969696969696973e-06, "loss": 2.0474, "step": 28 }, { "epoch": 2.7272727272727275, "grad_norm": 1.6685246229171753, "learning_rate": 1.8181818181818183e-06, "loss": 1.9084, "step": 30 }, { "epoch": 2.909090909090909, "grad_norm": 2.073890209197998, "learning_rate": 1.9393939393939395e-06, "loss": 2.1371, "step": 32 }, { "epoch": 3.0, "eval_loss": 2.0466740131378174, "eval_runtime": 9.6342, "eval_samples_per_second": 2.491, "eval_steps_per_second": 2.491, "step": 33 }, { "epoch": 3.090909090909091, "grad_norm": 1.7652820348739624, "learning_rate": 2.0606060606060607e-06, "loss": 2.1285, "step": 34 }, { "epoch": 3.2727272727272725, "grad_norm": 2.0258305072784424, "learning_rate": 2.181818181818182e-06, "loss": 2.0643, "step": 36 }, { "epoch": 3.4545454545454546, "grad_norm": 1.8539706468582153, "learning_rate": 2.303030303030303e-06, "loss": 1.9892, "step": 38 }, { "epoch": 3.6363636363636362, "grad_norm": 1.8331531286239624, "learning_rate": 2.4242424242424244e-06, "loss": 1.9928, "step": 40 }, { "epoch": 3.8181818181818183, "grad_norm": 1.7377302646636963, "learning_rate": 2.5454545454545456e-06, "loss": 1.9357, "step": 42 }, { "epoch": 4.0, "grad_norm": 2.0487308502197266, "learning_rate": 2.666666666666667e-06, "loss": 2.0536, "step": 44 }, { "epoch": 4.0, "eval_loss": 1.9862397909164429, "eval_runtime": 9.6288, "eval_samples_per_second": 2.493, "eval_steps_per_second": 2.493, "step": 44 }, { "epoch": 4.181818181818182, "grad_norm": 1.935196876525879, "learning_rate": 2.7878787878787885e-06, "loss": 1.9591, "step": 46 }, { "epoch": 4.363636363636363, "grad_norm": 1.7882719039916992, "learning_rate": 2.9090909090909093e-06, "loss": 1.9446, "step": 48 }, { "epoch": 4.545454545454545, "grad_norm": 1.80415940284729, "learning_rate": 3.0303030303030305e-06, "loss": 1.9487, "step": 50 }, { "epoch": 4.7272727272727275, "grad_norm": 1.7832120656967163, "learning_rate": 3.1515151515151517e-06, "loss": 1.9723, "step": 52 }, { "epoch": 4.909090909090909, "grad_norm": 1.6389567852020264, "learning_rate": 3.272727272727273e-06, "loss": 1.8317, "step": 54 }, { "epoch": 5.0, "eval_loss": 1.8956037759780884, "eval_runtime": 9.6308, "eval_samples_per_second": 2.492, "eval_steps_per_second": 2.492, "step": 55 }, { "epoch": 5.090909090909091, "grad_norm": 1.7195862531661987, "learning_rate": 3.3939393939393946e-06, "loss": 1.9008, "step": 56 }, { "epoch": 5.2727272727272725, "grad_norm": 1.8558402061462402, "learning_rate": 3.5151515151515154e-06, "loss": 1.916, "step": 58 }, { "epoch": 5.454545454545454, "grad_norm": 1.9300419092178345, "learning_rate": 3.6363636363636366e-06, "loss": 1.8262, "step": 60 }, { "epoch": 5.636363636363637, "grad_norm": 1.8043519258499146, "learning_rate": 3.757575757575758e-06, "loss": 1.8013, "step": 62 }, { "epoch": 5.818181818181818, "grad_norm": 1.5544039011001587, "learning_rate": 3.878787878787879e-06, "loss": 1.7996, "step": 64 }, { "epoch": 6.0, "grad_norm": 1.6001099348068237, "learning_rate": 4.000000000000001e-06, "loss": 1.7607, "step": 66 }, { "epoch": 6.0, "eval_loss": 1.7668356895446777, "eval_runtime": 9.6388, "eval_samples_per_second": 2.49, "eval_steps_per_second": 2.49, "step": 66 }, { "epoch": 6.181818181818182, "grad_norm": 1.52090322971344, "learning_rate": 4.1212121212121215e-06, "loss": 1.7673, "step": 68 }, { "epoch": 6.363636363636363, "grad_norm": 1.4516123533248901, "learning_rate": 4.242424242424243e-06, "loss": 1.782, "step": 70 }, { "epoch": 6.545454545454545, "grad_norm": 1.318569540977478, "learning_rate": 4.363636363636364e-06, "loss": 1.6176, "step": 72 }, { "epoch": 6.7272727272727275, "grad_norm": 1.3474187850952148, "learning_rate": 4.4848484848484855e-06, "loss": 1.6543, "step": 74 }, { "epoch": 6.909090909090909, "grad_norm": 1.1612062454223633, "learning_rate": 4.606060606060606e-06, "loss": 1.6452, "step": 76 }, { "epoch": 7.0, "eval_loss": 1.6453262567520142, "eval_runtime": 9.6333, "eval_samples_per_second": 2.491, "eval_steps_per_second": 2.491, "step": 77 }, { "epoch": 7.090909090909091, "grad_norm": 1.062233328819275, "learning_rate": 4.727272727272728e-06, "loss": 1.5258, "step": 78 }, { "epoch": 7.2727272727272725, "grad_norm": 0.9276080131530762, "learning_rate": 4.848484848484849e-06, "loss": 1.7018, "step": 80 }, { "epoch": 7.454545454545454, "grad_norm": 0.7713845372200012, "learning_rate": 4.9696969696969696e-06, "loss": 1.5366, "step": 82 }, { "epoch": 7.636363636363637, "grad_norm": 0.8294752240180969, "learning_rate": 5.090909090909091e-06, "loss": 1.5103, "step": 84 }, { "epoch": 7.818181818181818, "grad_norm": 0.7149582505226135, "learning_rate": 5.212121212121213e-06, "loss": 1.5768, "step": 86 }, { "epoch": 8.0, "grad_norm": 0.7764841914176941, "learning_rate": 5.333333333333334e-06, "loss": 1.548, "step": 88 }, { "epoch": 8.0, "eval_loss": 1.5728096961975098, "eval_runtime": 9.6306, "eval_samples_per_second": 2.492, "eval_steps_per_second": 2.492, "step": 88 }, { "epoch": 8.181818181818182, "grad_norm": 0.7410159111022949, "learning_rate": 5.4545454545454545e-06, "loss": 1.4837, "step": 90 }, { "epoch": 8.363636363636363, "grad_norm": 0.7778185606002808, "learning_rate": 5.575757575757577e-06, "loss": 1.518, "step": 92 }, { "epoch": 8.545454545454545, "grad_norm": 0.7047359943389893, "learning_rate": 5.696969696969698e-06, "loss": 1.5698, "step": 94 }, { "epoch": 8.727272727272727, "grad_norm": 0.8407279253005981, "learning_rate": 5.8181818181818185e-06, "loss": 1.5277, "step": 96 }, { "epoch": 8.909090909090908, "grad_norm": 0.6757096648216248, "learning_rate": 5.93939393939394e-06, "loss": 1.4631, "step": 98 }, { "epoch": 9.0, "eval_loss": 1.5217022895812988, "eval_runtime": 9.631, "eval_samples_per_second": 2.492, "eval_steps_per_second": 2.492, "step": 99 }, { "epoch": 9.090909090909092, "grad_norm": 0.6718485355377197, "learning_rate": 6.060606060606061e-06, "loss": 1.4975, "step": 100 }, { "epoch": 9.272727272727273, "grad_norm": 0.657922089099884, "learning_rate": 6.181818181818182e-06, "loss": 1.457, "step": 102 }, { "epoch": 9.454545454545455, "grad_norm": 0.8073777556419373, "learning_rate": 6.303030303030303e-06, "loss": 1.5072, "step": 104 }, { "epoch": 9.636363636363637, "grad_norm": 0.7699378132820129, "learning_rate": 6.424242424242425e-06, "loss": 1.4021, "step": 106 }, { "epoch": 9.818181818181818, "grad_norm": 0.7344899773597717, "learning_rate": 6.545454545454546e-06, "loss": 1.4539, "step": 108 }, { "epoch": 10.0, "grad_norm": 0.6962461471557617, "learning_rate": 6.666666666666667e-06, "loss": 1.4126, "step": 110 }, { "epoch": 10.0, "eval_loss": 1.471091866493225, "eval_runtime": 9.628, "eval_samples_per_second": 2.493, "eval_steps_per_second": 2.493, "step": 110 }, { "epoch": 10.181818181818182, "grad_norm": 0.756142258644104, "learning_rate": 6.787878787878789e-06, "loss": 1.4482, "step": 112 }, { "epoch": 10.363636363636363, "grad_norm": 0.7393659353256226, "learning_rate": 6.90909090909091e-06, "loss": 1.45, "step": 114 }, { "epoch": 10.545454545454545, "grad_norm": 0.7522909641265869, "learning_rate": 7.030303030303031e-06, "loss": 1.4183, "step": 116 }, { "epoch": 10.727272727272727, "grad_norm": 0.632103681564331, "learning_rate": 7.151515151515152e-06, "loss": 1.4051, "step": 118 }, { "epoch": 10.909090909090908, "grad_norm": 0.6698535680770874, "learning_rate": 7.272727272727273e-06, "loss": 1.3079, "step": 120 }, { "epoch": 11.0, "eval_loss": 1.4176009893417358, "eval_runtime": 9.6277, "eval_samples_per_second": 2.493, "eval_steps_per_second": 2.493, "step": 121 }, { "epoch": 11.090909090909092, "grad_norm": 0.7236391305923462, "learning_rate": 7.393939393939395e-06, "loss": 1.3409, "step": 122 }, { "epoch": 11.272727272727273, "grad_norm": 0.7769017219543457, "learning_rate": 7.515151515151516e-06, "loss": 1.3852, "step": 124 }, { "epoch": 11.454545454545455, "grad_norm": 0.7625659704208374, "learning_rate": 7.636363636363638e-06, "loss": 1.4075, "step": 126 }, { "epoch": 11.636363636363637, "grad_norm": 0.7391623258590698, "learning_rate": 7.757575757575758e-06, "loss": 1.3438, "step": 128 }, { "epoch": 11.818181818181818, "grad_norm": 0.7174716591835022, "learning_rate": 7.87878787878788e-06, "loss": 1.2229, "step": 130 }, { "epoch": 12.0, "grad_norm": 0.6933201551437378, "learning_rate": 8.000000000000001e-06, "loss": 1.3012, "step": 132 }, { "epoch": 12.0, "eval_loss": 1.3768787384033203, "eval_runtime": 9.6401, "eval_samples_per_second": 2.49, "eval_steps_per_second": 2.49, "step": 132 }, { "epoch": 12.181818181818182, "grad_norm": 0.7602589130401611, "learning_rate": 8.121212121212121e-06, "loss": 1.3075, "step": 134 }, { "epoch": 12.363636363636363, "grad_norm": 0.7896175384521484, "learning_rate": 8.242424242424243e-06, "loss": 1.2531, "step": 136 }, { "epoch": 12.545454545454545, "grad_norm": 0.7628654837608337, "learning_rate": 8.363636363636365e-06, "loss": 1.2487, "step": 138 }, { "epoch": 12.727272727272727, "grad_norm": 0.7553589940071106, "learning_rate": 8.484848484848486e-06, "loss": 1.3375, "step": 140 }, { "epoch": 12.909090909090908, "grad_norm": 0.8193318247795105, "learning_rate": 8.606060606060606e-06, "loss": 1.2575, "step": 142 }, { "epoch": 13.0, "eval_loss": 1.34230375289917, "eval_runtime": 9.6362, "eval_samples_per_second": 2.491, "eval_steps_per_second": 2.491, "step": 143 }, { "epoch": 13.090909090909092, "grad_norm": 0.7942626476287842, "learning_rate": 8.727272727272728e-06, "loss": 1.2746, "step": 144 }, { "epoch": 13.272727272727273, "grad_norm": 0.7456735372543335, "learning_rate": 8.84848484848485e-06, "loss": 1.2148, "step": 146 }, { "epoch": 13.454545454545455, "grad_norm": 0.8296701908111572, "learning_rate": 8.969696969696971e-06, "loss": 1.2372, "step": 148 }, { "epoch": 13.636363636363637, "grad_norm": 0.8604351282119751, "learning_rate": 9.090909090909091e-06, "loss": 1.1917, "step": 150 }, { "epoch": 13.818181818181818, "grad_norm": 0.8084425330162048, "learning_rate": 9.212121212121213e-06, "loss": 1.2257, "step": 152 }, { "epoch": 14.0, "grad_norm": 0.8843783140182495, "learning_rate": 9.333333333333334e-06, "loss": 1.2537, "step": 154 }, { "epoch": 14.0, "eval_loss": 1.3097882270812988, "eval_runtime": 9.642, "eval_samples_per_second": 2.489, "eval_steps_per_second": 2.489, "step": 154 }, { "epoch": 14.181818181818182, "grad_norm": 0.9855999946594238, "learning_rate": 9.454545454545456e-06, "loss": 1.1807, "step": 156 }, { "epoch": 14.363636363636363, "grad_norm": 0.7823004722595215, "learning_rate": 9.575757575757576e-06, "loss": 1.1481, "step": 158 }, { "epoch": 14.545454545454545, "grad_norm": 0.8486412167549133, "learning_rate": 9.696969696969698e-06, "loss": 1.1524, "step": 160 }, { "epoch": 14.727272727272727, "grad_norm": 0.9296090006828308, "learning_rate": 9.81818181818182e-06, "loss": 1.223, "step": 162 }, { "epoch": 14.909090909090908, "grad_norm": 0.962130069732666, "learning_rate": 9.939393939393939e-06, "loss": 1.1994, "step": 164 }, { "epoch": 15.0, "eval_loss": 1.2874382734298706, "eval_runtime": 9.6369, "eval_samples_per_second": 2.49, "eval_steps_per_second": 2.49, "step": 165 }, { "epoch": 15.090909090909092, "grad_norm": 1.0040925741195679, "learning_rate": 9.999988811118232e-06, "loss": 1.0752, "step": 166 }, { "epoch": 15.272727272727273, "grad_norm": 0.8788989782333374, "learning_rate": 9.999899300364534e-06, "loss": 1.1945, "step": 168 }, { "epoch": 15.454545454545455, "grad_norm": 0.8372966647148132, "learning_rate": 9.999720280459576e-06, "loss": 1.1612, "step": 170 }, { "epoch": 15.636363636363637, "grad_norm": 0.9981568455696106, "learning_rate": 9.999451754608208e-06, "loss": 1.0675, "step": 172 }, { "epoch": 15.818181818181818, "grad_norm": 0.989941418170929, "learning_rate": 9.99909372761763e-06, "loss": 1.1123, "step": 174 }, { "epoch": 16.0, "grad_norm": 1.071035623550415, "learning_rate": 9.99864620589731e-06, "loss": 1.1054, "step": 176 }, { "epoch": 16.0, "eval_loss": 1.2712817192077637, "eval_runtime": 9.6277, "eval_samples_per_second": 2.493, "eval_steps_per_second": 2.493, "step": 176 }, { "epoch": 16.181818181818183, "grad_norm": 1.0144329071044922, "learning_rate": 9.998109197458865e-06, "loss": 1.1681, "step": 178 }, { "epoch": 16.363636363636363, "grad_norm": 1.0573251247406006, "learning_rate": 9.997482711915926e-06, "loss": 1.06, "step": 180 }, { "epoch": 16.545454545454547, "grad_norm": 1.000741720199585, "learning_rate": 9.996766760483955e-06, "loss": 1.0508, "step": 182 }, { "epoch": 16.727272727272727, "grad_norm": 1.2005733251571655, "learning_rate": 9.995961355980052e-06, "loss": 1.0129, "step": 184 }, { "epoch": 16.90909090909091, "grad_norm": 1.195275902748108, "learning_rate": 9.99506651282272e-06, "loss": 1.0452, "step": 186 }, { "epoch": 17.0, "eval_loss": 1.2680319547653198, "eval_runtime": 9.6377, "eval_samples_per_second": 2.49, "eval_steps_per_second": 2.49, "step": 187 }, { "epoch": 17.09090909090909, "grad_norm": 1.2155280113220215, "learning_rate": 9.994082247031613e-06, "loss": 1.0263, "step": 188 }, { "epoch": 17.272727272727273, "grad_norm": 1.143007516860962, "learning_rate": 9.993008576227248e-06, "loss": 0.9197, "step": 190 }, { "epoch": 17.454545454545453, "grad_norm": 1.260664939880371, "learning_rate": 9.991845519630679e-06, "loss": 0.9681, "step": 192 }, { "epoch": 17.636363636363637, "grad_norm": 1.3235067129135132, "learning_rate": 9.99059309806317e-06, "loss": 1.0343, "step": 194 }, { "epoch": 17.818181818181817, "grad_norm": 1.4046649932861328, "learning_rate": 9.989251333945813e-06, "loss": 1.0078, "step": 196 }, { "epoch": 18.0, "grad_norm": 1.2338204383850098, "learning_rate": 9.987820251299121e-06, "loss": 1.0716, "step": 198 }, { "epoch": 18.0, "eval_loss": 1.2658852338790894, "eval_runtime": 9.6422, "eval_samples_per_second": 2.489, "eval_steps_per_second": 2.489, "step": 198 }, { "epoch": 18.181818181818183, "grad_norm": 1.3795467615127563, "learning_rate": 9.986299875742612e-06, "loss": 0.9914, "step": 200 }, { "epoch": 18.363636363636363, "grad_norm": 1.2211111783981323, "learning_rate": 9.984690234494338e-06, "loss": 0.9788, "step": 202 }, { "epoch": 18.545454545454547, "grad_norm": 1.4873826503753662, "learning_rate": 9.982991356370404e-06, "loss": 0.9206, "step": 204 }, { "epoch": 18.727272727272727, "grad_norm": 1.1939488649368286, "learning_rate": 9.98120327178445e-06, "loss": 0.9713, "step": 206 }, { "epoch": 18.90909090909091, "grad_norm": 1.4163975715637207, "learning_rate": 9.979326012747106e-06, "loss": 0.9207, "step": 208 }, { "epoch": 19.0, "eval_loss": 1.2754586935043335, "eval_runtime": 9.6506, "eval_samples_per_second": 2.487, "eval_steps_per_second": 2.487, "step": 209 }, { "epoch": 19.09090909090909, "grad_norm": 1.57706880569458, "learning_rate": 9.977359612865424e-06, "loss": 0.847, "step": 210 }, { "epoch": 19.272727272727273, "grad_norm": 1.3526296615600586, "learning_rate": 9.975304107342268e-06, "loss": 0.83, "step": 212 }, { "epoch": 19.454545454545453, "grad_norm": 1.3322209119796753, "learning_rate": 9.973159532975691e-06, "loss": 0.9059, "step": 214 }, { "epoch": 19.636363636363637, "grad_norm": 1.7349506616592407, "learning_rate": 9.970925928158275e-06, "loss": 0.9101, "step": 216 }, { "epoch": 19.818181818181817, "grad_norm": 1.7136040925979614, "learning_rate": 9.968603332876435e-06, "loss": 0.9291, "step": 218 }, { "epoch": 20.0, "grad_norm": 1.6715441942214966, "learning_rate": 9.966191788709716e-06, "loss": 0.8712, "step": 220 }, { "epoch": 20.0, "eval_loss": 1.291751503944397, "eval_runtime": 9.6384, "eval_samples_per_second": 2.49, "eval_steps_per_second": 2.49, "step": 220 }, { "epoch": 20.181818181818183, "grad_norm": 1.7461074590682983, "learning_rate": 9.963691338830045e-06, "loss": 0.8728, "step": 222 }, { "epoch": 20.363636363636363, "grad_norm": 1.5934933423995972, "learning_rate": 9.961102028000948e-06, "loss": 0.7777, "step": 224 }, { "epoch": 20.545454545454547, "grad_norm": 1.8316841125488281, "learning_rate": 9.958423902576764e-06, "loss": 0.9123, "step": 226 }, { "epoch": 20.727272727272727, "grad_norm": 1.6827653646469116, "learning_rate": 9.955657010501807e-06, "loss": 0.741, "step": 228 }, { "epoch": 20.90909090909091, "grad_norm": 1.7905974388122559, "learning_rate": 9.952801401309504e-06, "loss": 0.8179, "step": 230 }, { "epoch": 21.0, "eval_loss": 1.337079644203186, "eval_runtime": 9.6484, "eval_samples_per_second": 2.487, "eval_steps_per_second": 2.487, "step": 231 }, { "epoch": 21.09090909090909, "grad_norm": 1.711150884628296, "learning_rate": 9.949857126121519e-06, "loss": 0.7859, "step": 232 }, { "epoch": 21.272727272727273, "grad_norm": 1.8518304824829102, "learning_rate": 9.946824237646823e-06, "loss": 0.7743, "step": 234 }, { "epoch": 21.454545454545453, "grad_norm": 2.065861701965332, "learning_rate": 9.94370279018077e-06, "loss": 0.7725, "step": 236 }, { "epoch": 21.636363636363637, "grad_norm": 1.9285728931427002, "learning_rate": 9.940492839604103e-06, "loss": 0.7043, "step": 238 }, { "epoch": 21.818181818181817, "grad_norm": 1.9804288148880005, "learning_rate": 9.937194443381972e-06, "loss": 0.8355, "step": 240 }, { "epoch": 22.0, "grad_norm": 1.9706605672836304, "learning_rate": 9.933807660562898e-06, "loss": 0.6485, "step": 242 }, { "epoch": 22.0, "eval_loss": 1.3560718297958374, "eval_runtime": 9.6266, "eval_samples_per_second": 2.493, "eval_steps_per_second": 2.493, "step": 242 }, { "epoch": 22.181818181818183, "grad_norm": 1.8734219074249268, "learning_rate": 9.930332551777709e-06, "loss": 0.6863, "step": 244 }, { "epoch": 22.363636363636363, "grad_norm": 2.110110282897949, "learning_rate": 9.926769179238467e-06, "loss": 0.6799, "step": 246 }, { "epoch": 22.545454545454547, "grad_norm": 2.442317008972168, "learning_rate": 9.923117606737347e-06, "loss": 0.6703, "step": 248 }, { "epoch": 22.727272727272727, "grad_norm": 2.216526508331299, "learning_rate": 9.919377899645497e-06, "loss": 0.687, "step": 250 }, { "epoch": 22.90909090909091, "grad_norm": 2.3219642639160156, "learning_rate": 9.915550124911866e-06, "loss": 0.6958, "step": 252 }, { "epoch": 23.0, "eval_loss": 1.4414390325546265, "eval_runtime": 9.6326, "eval_samples_per_second": 2.492, "eval_steps_per_second": 2.492, "step": 253 }, { "epoch": 23.09090909090909, "grad_norm": 2.1852304935455322, "learning_rate": 9.91163435106201e-06, "loss": 0.5976, "step": 254 }, { "epoch": 23.272727272727273, "grad_norm": 2.0445878505706787, "learning_rate": 9.907630648196857e-06, "loss": 0.687, "step": 256 }, { "epoch": 23.454545454545453, "grad_norm": 2.2275197505950928, "learning_rate": 9.903539087991462e-06, "loss": 0.62, "step": 258 }, { "epoch": 23.636363636363637, "grad_norm": 2.462937593460083, "learning_rate": 9.899359743693715e-06, "loss": 0.5359, "step": 260 }, { "epoch": 23.818181818181817, "grad_norm": 2.7242631912231445, "learning_rate": 9.895092690123036e-06, "loss": 0.6217, "step": 262 }, { "epoch": 24.0, "grad_norm": 2.585885763168335, "learning_rate": 9.890738003669029e-06, "loss": 0.5845, "step": 264 }, { "epoch": 24.0, "eval_loss": 1.5147289037704468, "eval_runtime": 9.6323, "eval_samples_per_second": 2.492, "eval_steps_per_second": 2.492, "step": 264 }, { "epoch": 24.181818181818183, "grad_norm": 2.6411542892456055, "learning_rate": 9.886295762290125e-06, "loss": 0.5457, "step": 266 }, { "epoch": 24.363636363636363, "grad_norm": 2.4892776012420654, "learning_rate": 9.881766045512176e-06, "loss": 0.5765, "step": 268 }, { "epoch": 24.545454545454547, "grad_norm": 2.6362385749816895, "learning_rate": 9.877148934427037e-06, "loss": 0.5477, "step": 270 }, { "epoch": 24.727272727272727, "grad_norm": 2.6221413612365723, "learning_rate": 9.872444511691108e-06, "loss": 0.4876, "step": 272 }, { "epoch": 24.90909090909091, "grad_norm": 3.3728928565979004, "learning_rate": 9.867652861523866e-06, "loss": 0.5274, "step": 274 }, { "epoch": 25.0, "eval_loss": 1.5911802053451538, "eval_runtime": 9.6269, "eval_samples_per_second": 2.493, "eval_steps_per_second": 2.493, "step": 275 }, { "epoch": 25.0, "step": 275, "total_flos": 3.3769591557586944e+16, "train_loss": 1.3221848821640014, "train_runtime": 2580.936, "train_samples_per_second": 5.114, "train_steps_per_second": 0.639 } ], "logging_steps": 2, "max_steps": 1650, "num_input_tokens_seen": 0, "num_train_epochs": 150, "save_steps": 25, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 7, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.3769591557586944e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }