|
{ |
|
"best_metric": 0.9015458226203918, |
|
"best_model_checkpoint": "data/Llama-31-8B_task-1_120-samples_config-3_full/checkpoint-396", |
|
"epoch": 43.0, |
|
"eval_steps": 500, |
|
"global_step": 473, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.09090909090909091, |
|
"grad_norm": 2.021212100982666, |
|
"learning_rate": 6.060606060606061e-08, |
|
"loss": 2.4844, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.18181818181818182, |
|
"grad_norm": 1.8868809938430786, |
|
"learning_rate": 1.2121212121212122e-07, |
|
"loss": 2.5082, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.36363636363636365, |
|
"grad_norm": 1.9227168560028076, |
|
"learning_rate": 2.4242424242424244e-07, |
|
"loss": 2.5149, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.5454545454545454, |
|
"grad_norm": 2.164538860321045, |
|
"learning_rate": 3.6363636363636366e-07, |
|
"loss": 2.4938, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.7272727272727273, |
|
"grad_norm": 1.9525693655014038, |
|
"learning_rate": 4.848484848484849e-07, |
|
"loss": 2.4402, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.9090909090909091, |
|
"grad_norm": 1.973608136177063, |
|
"learning_rate": 6.060606060606061e-07, |
|
"loss": 2.4681, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 2.453854560852051, |
|
"eval_runtime": 9.6392, |
|
"eval_samples_per_second": 2.49, |
|
"eval_steps_per_second": 2.49, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 1.0909090909090908, |
|
"grad_norm": 1.9462313652038574, |
|
"learning_rate": 7.272727272727273e-07, |
|
"loss": 2.4751, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 1.2727272727272727, |
|
"grad_norm": 1.8247989416122437, |
|
"learning_rate": 8.484848484848486e-07, |
|
"loss": 2.523, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 1.4545454545454546, |
|
"grad_norm": 1.8403159379959106, |
|
"learning_rate": 9.696969696969698e-07, |
|
"loss": 2.4192, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 1.6363636363636362, |
|
"grad_norm": 1.6670445203781128, |
|
"learning_rate": 1.090909090909091e-06, |
|
"loss": 2.5069, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 1.8181818181818183, |
|
"grad_norm": 1.802819013595581, |
|
"learning_rate": 1.2121212121212122e-06, |
|
"loss": 2.4854, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 1.4742677211761475, |
|
"learning_rate": 1.3333333333333334e-06, |
|
"loss": 2.3894, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 2.4259960651397705, |
|
"eval_runtime": 9.6188, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 2.1818181818181817, |
|
"grad_norm": 1.6892441511154175, |
|
"learning_rate": 1.4545454545454546e-06, |
|
"loss": 2.4004, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 2.3636363636363638, |
|
"grad_norm": 1.5024837255477905, |
|
"learning_rate": 1.5757575757575759e-06, |
|
"loss": 2.4339, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 2.5454545454545454, |
|
"grad_norm": 1.6032872200012207, |
|
"learning_rate": 1.6969696969696973e-06, |
|
"loss": 2.4351, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 2.7272727272727275, |
|
"grad_norm": 1.526031732559204, |
|
"learning_rate": 1.8181818181818183e-06, |
|
"loss": 2.3962, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 2.909090909090909, |
|
"grad_norm": 1.6687572002410889, |
|
"learning_rate": 1.9393939393939395e-06, |
|
"loss": 2.4746, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 2.382659912109375, |
|
"eval_runtime": 9.6201, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 3.090909090909091, |
|
"grad_norm": 1.5408605337142944, |
|
"learning_rate": 2.0606060606060607e-06, |
|
"loss": 2.4125, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 3.2727272727272725, |
|
"grad_norm": 1.6628339290618896, |
|
"learning_rate": 2.181818181818182e-06, |
|
"loss": 2.3574, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 3.4545454545454546, |
|
"grad_norm": 1.6864051818847656, |
|
"learning_rate": 2.303030303030303e-06, |
|
"loss": 2.3033, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 3.6363636363636362, |
|
"grad_norm": 1.705551266670227, |
|
"learning_rate": 2.4242424242424244e-06, |
|
"loss": 2.4227, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 3.8181818181818183, |
|
"grad_norm": 1.6931661367416382, |
|
"learning_rate": 2.5454545454545456e-06, |
|
"loss": 2.3626, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 1.910327434539795, |
|
"learning_rate": 2.666666666666667e-06, |
|
"loss": 2.4177, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 2.313791036605835, |
|
"eval_runtime": 9.6408, |
|
"eval_samples_per_second": 2.489, |
|
"eval_steps_per_second": 2.489, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 4.181818181818182, |
|
"grad_norm": 1.9009881019592285, |
|
"learning_rate": 2.7878787878787885e-06, |
|
"loss": 2.2818, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 4.363636363636363, |
|
"grad_norm": 1.9625577926635742, |
|
"learning_rate": 2.9090909090909093e-06, |
|
"loss": 2.4051, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 4.545454545454545, |
|
"grad_norm": 1.4953864812850952, |
|
"learning_rate": 3.0303030303030305e-06, |
|
"loss": 2.2829, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 4.7272727272727275, |
|
"grad_norm": 1.5853980779647827, |
|
"learning_rate": 3.1515151515151517e-06, |
|
"loss": 2.2705, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 4.909090909090909, |
|
"grad_norm": 1.5361964702606201, |
|
"learning_rate": 3.272727272727273e-06, |
|
"loss": 2.1959, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 2.226909875869751, |
|
"eval_runtime": 9.6204, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 5.090909090909091, |
|
"grad_norm": 1.919047236442566, |
|
"learning_rate": 3.3939393939393946e-06, |
|
"loss": 2.2901, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 5.2727272727272725, |
|
"grad_norm": 2.0125391483306885, |
|
"learning_rate": 3.5151515151515154e-06, |
|
"loss": 2.2584, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 5.454545454545454, |
|
"grad_norm": 1.8322199583053589, |
|
"learning_rate": 3.6363636363636366e-06, |
|
"loss": 2.1565, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 5.636363636363637, |
|
"grad_norm": 1.318441390991211, |
|
"learning_rate": 3.757575757575758e-06, |
|
"loss": 2.2066, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 5.818181818181818, |
|
"grad_norm": 1.2064168453216553, |
|
"learning_rate": 3.878787878787879e-06, |
|
"loss": 2.1475, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"grad_norm": 1.1312469244003296, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 2.16, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 2.1177315711975098, |
|
"eval_runtime": 9.6288, |
|
"eval_samples_per_second": 2.493, |
|
"eval_steps_per_second": 2.493, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 6.181818181818182, |
|
"grad_norm": 1.0877413749694824, |
|
"learning_rate": 4.1212121212121215e-06, |
|
"loss": 2.0577, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 6.363636363636363, |
|
"grad_norm": 1.1328318119049072, |
|
"learning_rate": 4.242424242424243e-06, |
|
"loss": 2.1584, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 6.545454545454545, |
|
"grad_norm": 1.1114661693572998, |
|
"learning_rate": 4.363636363636364e-06, |
|
"loss": 2.1214, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 6.7272727272727275, |
|
"grad_norm": 1.0736863613128662, |
|
"learning_rate": 4.4848484848484855e-06, |
|
"loss": 2.0338, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 6.909090909090909, |
|
"grad_norm": 1.1103681325912476, |
|
"learning_rate": 4.606060606060606e-06, |
|
"loss": 2.0388, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 1.9843875169754028, |
|
"eval_runtime": 9.6304, |
|
"eval_samples_per_second": 2.492, |
|
"eval_steps_per_second": 2.492, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 7.090909090909091, |
|
"grad_norm": 1.0429500341415405, |
|
"learning_rate": 4.727272727272728e-06, |
|
"loss": 1.9515, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 7.2727272727272725, |
|
"grad_norm": 1.0388323068618774, |
|
"learning_rate": 4.848484848484849e-06, |
|
"loss": 2.0436, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 7.454545454545454, |
|
"grad_norm": 0.9600175023078918, |
|
"learning_rate": 4.9696969696969696e-06, |
|
"loss": 1.8726, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 7.636363636363637, |
|
"grad_norm": 0.9380725026130676, |
|
"learning_rate": 5.090909090909091e-06, |
|
"loss": 1.9307, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 7.818181818181818, |
|
"grad_norm": 0.8361033797264099, |
|
"learning_rate": 5.212121212121213e-06, |
|
"loss": 1.869, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 0.8874518275260925, |
|
"learning_rate": 5.333333333333334e-06, |
|
"loss": 1.8932, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 1.8441652059555054, |
|
"eval_runtime": 9.6269, |
|
"eval_samples_per_second": 2.493, |
|
"eval_steps_per_second": 2.493, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 8.181818181818182, |
|
"grad_norm": 0.8483957052230835, |
|
"learning_rate": 5.4545454545454545e-06, |
|
"loss": 1.8148, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 8.363636363636363, |
|
"grad_norm": 0.9277289509773254, |
|
"learning_rate": 5.575757575757577e-06, |
|
"loss": 1.8286, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 8.545454545454545, |
|
"grad_norm": 0.9386701583862305, |
|
"learning_rate": 5.696969696969698e-06, |
|
"loss": 1.7935, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 8.727272727272727, |
|
"grad_norm": 0.9653158783912659, |
|
"learning_rate": 5.8181818181818185e-06, |
|
"loss": 1.7714, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 8.909090909090908, |
|
"grad_norm": 0.929076075553894, |
|
"learning_rate": 5.93939393939394e-06, |
|
"loss": 1.7199, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 1.6830488443374634, |
|
"eval_runtime": 9.6281, |
|
"eval_samples_per_second": 2.493, |
|
"eval_steps_per_second": 2.493, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 9.090909090909092, |
|
"grad_norm": 0.9527233839035034, |
|
"learning_rate": 6.060606060606061e-06, |
|
"loss": 1.7054, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 9.272727272727273, |
|
"grad_norm": 1.1190400123596191, |
|
"learning_rate": 6.181818181818182e-06, |
|
"loss": 1.623, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 9.454545454545455, |
|
"grad_norm": 1.1004133224487305, |
|
"learning_rate": 6.303030303030303e-06, |
|
"loss": 1.6696, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 9.636363636363637, |
|
"grad_norm": 1.036831259727478, |
|
"learning_rate": 6.424242424242425e-06, |
|
"loss": 1.5955, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 9.818181818181818, |
|
"grad_norm": 0.9210329055786133, |
|
"learning_rate": 6.545454545454546e-06, |
|
"loss": 1.5252, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 1.0149990320205688, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 1.4973, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 1.4929395914077759, |
|
"eval_runtime": 9.6201, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 10.181818181818182, |
|
"grad_norm": 0.9690227508544922, |
|
"learning_rate": 6.787878787878789e-06, |
|
"loss": 1.4956, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 10.363636363636363, |
|
"grad_norm": 0.8670191168785095, |
|
"learning_rate": 6.90909090909091e-06, |
|
"loss": 1.4899, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 10.545454545454545, |
|
"grad_norm": 0.9516677260398865, |
|
"learning_rate": 7.030303030303031e-06, |
|
"loss": 1.3702, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 10.727272727272727, |
|
"grad_norm": 0.987469494342804, |
|
"learning_rate": 7.151515151515152e-06, |
|
"loss": 1.4109, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 10.909090909090908, |
|
"grad_norm": 0.9726764559745789, |
|
"learning_rate": 7.272727272727273e-06, |
|
"loss": 1.2726, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_loss": 1.2980190515518188, |
|
"eval_runtime": 9.6227, |
|
"eval_samples_per_second": 2.494, |
|
"eval_steps_per_second": 2.494, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 11.090909090909092, |
|
"grad_norm": 0.925747811794281, |
|
"learning_rate": 7.393939393939395e-06, |
|
"loss": 1.2873, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 11.272727272727273, |
|
"grad_norm": 1.0416390895843506, |
|
"learning_rate": 7.515151515151516e-06, |
|
"loss": 1.2906, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 11.454545454545455, |
|
"grad_norm": 0.8796542286872864, |
|
"learning_rate": 7.636363636363638e-06, |
|
"loss": 1.2497, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 11.636363636363637, |
|
"grad_norm": 0.7352049350738525, |
|
"learning_rate": 7.757575757575758e-06, |
|
"loss": 1.2023, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 11.818181818181818, |
|
"grad_norm": 0.7811641693115234, |
|
"learning_rate": 7.87878787878788e-06, |
|
"loss": 1.1205, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"grad_norm": 0.6574164032936096, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 1.204, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_loss": 1.1553701162338257, |
|
"eval_runtime": 9.6262, |
|
"eval_samples_per_second": 2.493, |
|
"eval_steps_per_second": 2.493, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 12.181818181818182, |
|
"grad_norm": 0.7204536199569702, |
|
"learning_rate": 8.121212121212121e-06, |
|
"loss": 1.134, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 12.363636363636363, |
|
"grad_norm": 0.7305138111114502, |
|
"learning_rate": 8.242424242424243e-06, |
|
"loss": 1.046, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 12.545454545454545, |
|
"grad_norm": 0.7755628228187561, |
|
"learning_rate": 8.363636363636365e-06, |
|
"loss": 1.1237, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 12.727272727272727, |
|
"grad_norm": 0.8240987062454224, |
|
"learning_rate": 8.484848484848486e-06, |
|
"loss": 1.1215, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 12.909090909090908, |
|
"grad_norm": 0.716643750667572, |
|
"learning_rate": 8.606060606060606e-06, |
|
"loss": 1.0597, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_loss": 1.0772147178649902, |
|
"eval_runtime": 9.6246, |
|
"eval_samples_per_second": 2.494, |
|
"eval_steps_per_second": 2.494, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 13.090909090909092, |
|
"grad_norm": 0.6686062812805176, |
|
"learning_rate": 8.727272727272728e-06, |
|
"loss": 1.09, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 13.272727272727273, |
|
"grad_norm": 0.5461896061897278, |
|
"learning_rate": 8.84848484848485e-06, |
|
"loss": 1.0405, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 13.454545454545455, |
|
"grad_norm": 0.5121557712554932, |
|
"learning_rate": 8.969696969696971e-06, |
|
"loss": 1.0367, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 13.636363636363637, |
|
"grad_norm": 0.5350603461265564, |
|
"learning_rate": 9.090909090909091e-06, |
|
"loss": 1.0327, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 13.818181818181818, |
|
"grad_norm": 0.48038652539253235, |
|
"learning_rate": 9.212121212121213e-06, |
|
"loss": 1.0409, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"grad_norm": 0.5158300399780273, |
|
"learning_rate": 9.333333333333334e-06, |
|
"loss": 1.0642, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_loss": 1.0425277948379517, |
|
"eval_runtime": 9.6218, |
|
"eval_samples_per_second": 2.494, |
|
"eval_steps_per_second": 2.494, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 14.181818181818182, |
|
"grad_norm": 0.5284631252288818, |
|
"learning_rate": 9.454545454545456e-06, |
|
"loss": 1.0258, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 14.363636363636363, |
|
"grad_norm": 0.5030417442321777, |
|
"learning_rate": 9.575757575757576e-06, |
|
"loss": 1.0207, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 14.545454545454545, |
|
"grad_norm": 0.5023478269577026, |
|
"learning_rate": 9.696969696969698e-06, |
|
"loss": 0.9857, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 14.727272727272727, |
|
"grad_norm": 0.4412921071052551, |
|
"learning_rate": 9.81818181818182e-06, |
|
"loss": 1.006, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 14.909090909090908, |
|
"grad_norm": 0.4654785096645355, |
|
"learning_rate": 9.939393939393939e-06, |
|
"loss": 1.0466, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_loss": 1.020145297050476, |
|
"eval_runtime": 9.6194, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 15.090909090909092, |
|
"grad_norm": 0.4788985848426819, |
|
"learning_rate": 9.999988811118232e-06, |
|
"loss": 0.9711, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 15.272727272727273, |
|
"grad_norm": 0.505489706993103, |
|
"learning_rate": 9.999899300364534e-06, |
|
"loss": 0.9704, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 15.454545454545455, |
|
"grad_norm": 0.5644559860229492, |
|
"learning_rate": 9.999720280459576e-06, |
|
"loss": 1.0248, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 15.636363636363637, |
|
"grad_norm": 0.4749113619327545, |
|
"learning_rate": 9.999451754608208e-06, |
|
"loss": 0.9284, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 15.818181818181818, |
|
"grad_norm": 0.658306896686554, |
|
"learning_rate": 9.99909372761763e-06, |
|
"loss": 1.0245, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"grad_norm": 0.5014739036560059, |
|
"learning_rate": 9.99864620589731e-06, |
|
"loss": 1.0044, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_loss": 1.0009642839431763, |
|
"eval_runtime": 9.6199, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 16.181818181818183, |
|
"grad_norm": 0.5191582441329956, |
|
"learning_rate": 9.998109197458865e-06, |
|
"loss": 0.9723, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 16.363636363636363, |
|
"grad_norm": 0.5656840205192566, |
|
"learning_rate": 9.997482711915926e-06, |
|
"loss": 0.9901, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 16.545454545454547, |
|
"grad_norm": 0.49254581332206726, |
|
"learning_rate": 9.996766760483955e-06, |
|
"loss": 0.9321, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 16.727272727272727, |
|
"grad_norm": 0.500410795211792, |
|
"learning_rate": 9.995961355980052e-06, |
|
"loss": 0.9564, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 16.90909090909091, |
|
"grad_norm": 0.6280571222305298, |
|
"learning_rate": 9.99506651282272e-06, |
|
"loss": 0.9967, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_loss": 0.9866144061088562, |
|
"eval_runtime": 9.621, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 17.09090909090909, |
|
"grad_norm": 0.578583300113678, |
|
"learning_rate": 9.994082247031613e-06, |
|
"loss": 0.9784, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 17.272727272727273, |
|
"grad_norm": 0.5218169093132019, |
|
"learning_rate": 9.993008576227248e-06, |
|
"loss": 0.8817, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 17.454545454545453, |
|
"grad_norm": 0.6009635329246521, |
|
"learning_rate": 9.991845519630679e-06, |
|
"loss": 0.9515, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 17.636363636363637, |
|
"grad_norm": 0.5707380771636963, |
|
"learning_rate": 9.99059309806317e-06, |
|
"loss": 0.9694, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 17.818181818181817, |
|
"grad_norm": 0.6348333358764648, |
|
"learning_rate": 9.989251333945813e-06, |
|
"loss": 0.9498, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"grad_norm": 0.5231234431266785, |
|
"learning_rate": 9.987820251299121e-06, |
|
"loss": 0.9863, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_loss": 0.9735579490661621, |
|
"eval_runtime": 9.6192, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 18.181818181818183, |
|
"grad_norm": 0.6517296433448792, |
|
"learning_rate": 9.986299875742612e-06, |
|
"loss": 0.9591, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 18.363636363636363, |
|
"grad_norm": 0.6254110932350159, |
|
"learning_rate": 9.984690234494338e-06, |
|
"loss": 0.9296, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 18.545454545454547, |
|
"grad_norm": 0.7348714470863342, |
|
"learning_rate": 9.982991356370404e-06, |
|
"loss": 0.9482, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 18.727272727272727, |
|
"grad_norm": 0.5867448449134827, |
|
"learning_rate": 9.98120327178445e-06, |
|
"loss": 0.9276, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 18.90909090909091, |
|
"grad_norm": 0.6433852910995483, |
|
"learning_rate": 9.979326012747106e-06, |
|
"loss": 0.9065, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_loss": 0.9643786549568176, |
|
"eval_runtime": 9.647, |
|
"eval_samples_per_second": 2.488, |
|
"eval_steps_per_second": 2.488, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 19.09090909090909, |
|
"grad_norm": 0.650622546672821, |
|
"learning_rate": 9.977359612865424e-06, |
|
"loss": 0.8947, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 19.272727272727273, |
|
"grad_norm": 0.7716985940933228, |
|
"learning_rate": 9.975304107342268e-06, |
|
"loss": 0.929, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 19.454545454545453, |
|
"grad_norm": 0.5945254564285278, |
|
"learning_rate": 9.973159532975691e-06, |
|
"loss": 0.9743, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 19.636363636363637, |
|
"grad_norm": 0.6686264872550964, |
|
"learning_rate": 9.970925928158275e-06, |
|
"loss": 0.9398, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 19.818181818181817, |
|
"grad_norm": 0.6576296091079712, |
|
"learning_rate": 9.968603332876435e-06, |
|
"loss": 0.9251, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"grad_norm": 0.5890394449234009, |
|
"learning_rate": 9.966191788709716e-06, |
|
"loss": 0.8669, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_loss": 0.9538940787315369, |
|
"eval_runtime": 9.6225, |
|
"eval_samples_per_second": 2.494, |
|
"eval_steps_per_second": 2.494, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 20.181818181818183, |
|
"grad_norm": 0.7149389386177063, |
|
"learning_rate": 9.963691338830045e-06, |
|
"loss": 0.9322, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 20.363636363636363, |
|
"grad_norm": 0.6404098272323608, |
|
"learning_rate": 9.961102028000948e-06, |
|
"loss": 0.8914, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 20.545454545454547, |
|
"grad_norm": 0.6958956122398376, |
|
"learning_rate": 9.958423902576764e-06, |
|
"loss": 0.9489, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 20.727272727272727, |
|
"grad_norm": 0.6515837907791138, |
|
"learning_rate": 9.955657010501807e-06, |
|
"loss": 0.8477, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 20.90909090909091, |
|
"grad_norm": 0.8464462161064148, |
|
"learning_rate": 9.952801401309504e-06, |
|
"loss": 0.9253, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_loss": 0.9454106688499451, |
|
"eval_runtime": 9.6279, |
|
"eval_samples_per_second": 2.493, |
|
"eval_steps_per_second": 2.493, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 21.09090909090909, |
|
"grad_norm": 0.7701188325881958, |
|
"learning_rate": 9.949857126121519e-06, |
|
"loss": 0.9027, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 21.272727272727273, |
|
"grad_norm": 0.6819570660591125, |
|
"learning_rate": 9.946824237646823e-06, |
|
"loss": 0.8545, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 21.454545454545453, |
|
"grad_norm": 0.7400511503219604, |
|
"learning_rate": 9.94370279018077e-06, |
|
"loss": 0.9229, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 21.636363636363637, |
|
"grad_norm": 0.6731954216957092, |
|
"learning_rate": 9.940492839604103e-06, |
|
"loss": 0.8596, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 21.818181818181817, |
|
"grad_norm": 0.7251449227333069, |
|
"learning_rate": 9.937194443381972e-06, |
|
"loss": 0.9052, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"grad_norm": 0.7023078799247742, |
|
"learning_rate": 9.933807660562898e-06, |
|
"loss": 0.872, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_loss": 0.9398066997528076, |
|
"eval_runtime": 9.6205, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 22.181818181818183, |
|
"grad_norm": 0.6639079451560974, |
|
"learning_rate": 9.930332551777709e-06, |
|
"loss": 0.8602, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 22.363636363636363, |
|
"grad_norm": 0.8378622531890869, |
|
"learning_rate": 9.926769179238467e-06, |
|
"loss": 0.9034, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 22.545454545454547, |
|
"grad_norm": 0.8481977581977844, |
|
"learning_rate": 9.923117606737347e-06, |
|
"loss": 0.8493, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 22.727272727272727, |
|
"grad_norm": 0.8623536825180054, |
|
"learning_rate": 9.919377899645497e-06, |
|
"loss": 0.8883, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 22.90909090909091, |
|
"grad_norm": 0.8115867376327515, |
|
"learning_rate": 9.915550124911866e-06, |
|
"loss": 0.8824, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_loss": 0.9327845573425293, |
|
"eval_runtime": 9.6274, |
|
"eval_samples_per_second": 2.493, |
|
"eval_steps_per_second": 2.493, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 23.09090909090909, |
|
"grad_norm": 0.7263345122337341, |
|
"learning_rate": 9.91163435106201e-06, |
|
"loss": 0.8292, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 23.272727272727273, |
|
"grad_norm": 0.8343206644058228, |
|
"learning_rate": 9.907630648196857e-06, |
|
"loss": 0.9044, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 23.454545454545453, |
|
"grad_norm": 0.7706195116043091, |
|
"learning_rate": 9.903539087991462e-06, |
|
"loss": 0.8363, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 23.636363636363637, |
|
"grad_norm": 0.7953319549560547, |
|
"learning_rate": 9.899359743693715e-06, |
|
"loss": 0.8528, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 23.818181818181817, |
|
"grad_norm": 0.8244236707687378, |
|
"learning_rate": 9.895092690123036e-06, |
|
"loss": 0.8787, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"grad_norm": 0.9236502647399902, |
|
"learning_rate": 9.890738003669029e-06, |
|
"loss": 0.8582, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_loss": 0.9282767176628113, |
|
"eval_runtime": 9.6299, |
|
"eval_samples_per_second": 2.492, |
|
"eval_steps_per_second": 2.492, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 24.181818181818183, |
|
"grad_norm": 0.7912314534187317, |
|
"learning_rate": 9.886295762290125e-06, |
|
"loss": 0.8291, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 24.363636363636363, |
|
"grad_norm": 0.8443674445152283, |
|
"learning_rate": 9.881766045512176e-06, |
|
"loss": 0.8599, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 24.545454545454547, |
|
"grad_norm": 0.9810454249382019, |
|
"learning_rate": 9.877148934427037e-06, |
|
"loss": 0.844, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 24.727272727272727, |
|
"grad_norm": 0.827754020690918, |
|
"learning_rate": 9.872444511691108e-06, |
|
"loss": 0.8447, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 24.90909090909091, |
|
"grad_norm": 0.9015039801597595, |
|
"learning_rate": 9.867652861523866e-06, |
|
"loss": 0.8763, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_loss": 0.9221246838569641, |
|
"eval_runtime": 9.6209, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 25.09090909090909, |
|
"grad_norm": 0.8841392993927002, |
|
"learning_rate": 9.862774069706346e-06, |
|
"loss": 0.8184, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 25.272727272727273, |
|
"grad_norm": 0.8848661780357361, |
|
"learning_rate": 9.85780822357961e-06, |
|
"loss": 0.8496, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 25.454545454545453, |
|
"grad_norm": 0.7846701145172119, |
|
"learning_rate": 9.85275541204318e-06, |
|
"loss": 0.8165, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 25.636363636363637, |
|
"grad_norm": 1.0641032457351685, |
|
"learning_rate": 9.847615725553457e-06, |
|
"loss": 0.8485, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 25.818181818181817, |
|
"grad_norm": 0.9293933510780334, |
|
"learning_rate": 9.842389256122086e-06, |
|
"loss": 0.829, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"grad_norm": 0.8131471872329712, |
|
"learning_rate": 9.83707609731432e-06, |
|
"loss": 0.8199, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_loss": 0.917729377746582, |
|
"eval_runtime": 9.6285, |
|
"eval_samples_per_second": 2.493, |
|
"eval_steps_per_second": 2.493, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 26.181818181818183, |
|
"grad_norm": 0.8052296042442322, |
|
"learning_rate": 9.831676344247343e-06, |
|
"loss": 0.8356, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 26.363636363636363, |
|
"grad_norm": 0.9019801020622253, |
|
"learning_rate": 9.826190093588564e-06, |
|
"loss": 0.8427, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 26.545454545454547, |
|
"grad_norm": 0.8782145380973816, |
|
"learning_rate": 9.820617443553889e-06, |
|
"loss": 0.8175, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 26.727272727272727, |
|
"grad_norm": 1.0153470039367676, |
|
"learning_rate": 9.814958493905962e-06, |
|
"loss": 0.8059, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 26.90909090909091, |
|
"grad_norm": 0.9501240253448486, |
|
"learning_rate": 9.80921334595238e-06, |
|
"loss": 0.7986, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_loss": 0.9146122336387634, |
|
"eval_runtime": 9.6275, |
|
"eval_samples_per_second": 2.493, |
|
"eval_steps_per_second": 2.493, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 27.09090909090909, |
|
"grad_norm": 0.8864769339561462, |
|
"learning_rate": 9.80338210254388e-06, |
|
"loss": 0.7819, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 27.272727272727273, |
|
"grad_norm": 0.8944150805473328, |
|
"learning_rate": 9.797464868072489e-06, |
|
"loss": 0.7875, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 27.454545454545453, |
|
"grad_norm": 0.9001815915107727, |
|
"learning_rate": 9.791461748469669e-06, |
|
"loss": 0.8734, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 27.636363636363637, |
|
"grad_norm": 0.9148157835006714, |
|
"learning_rate": 9.785372851204415e-06, |
|
"loss": 0.8076, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 27.818181818181817, |
|
"grad_norm": 1.0519267320632935, |
|
"learning_rate": 9.779198285281326e-06, |
|
"loss": 0.7918, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"grad_norm": 0.8914878964424133, |
|
"learning_rate": 9.77293816123866e-06, |
|
"loss": 0.7754, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_loss": 0.9141943454742432, |
|
"eval_runtime": 9.6378, |
|
"eval_samples_per_second": 2.49, |
|
"eval_steps_per_second": 2.49, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 28.181818181818183, |
|
"grad_norm": 0.9089798331260681, |
|
"learning_rate": 9.766592591146353e-06, |
|
"loss": 0.7795, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 28.363636363636363, |
|
"grad_norm": 1.054513931274414, |
|
"learning_rate": 9.760161688604008e-06, |
|
"loss": 0.7935, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 28.545454545454547, |
|
"grad_norm": 0.9446949362754822, |
|
"learning_rate": 9.753645568738872e-06, |
|
"loss": 0.797, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 28.727272727272727, |
|
"grad_norm": 0.9865514039993286, |
|
"learning_rate": 9.747044348203766e-06, |
|
"loss": 0.7824, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 28.90909090909091, |
|
"grad_norm": 1.0658093690872192, |
|
"learning_rate": 9.740358145174999e-06, |
|
"loss": 0.7893, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_loss": 0.9086329340934753, |
|
"eval_runtime": 9.6391, |
|
"eval_samples_per_second": 2.49, |
|
"eval_steps_per_second": 2.49, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 29.09090909090909, |
|
"grad_norm": 0.9681562185287476, |
|
"learning_rate": 9.733587079350254e-06, |
|
"loss": 0.7736, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 29.272727272727273, |
|
"grad_norm": 0.9801912903785706, |
|
"learning_rate": 9.72673127194644e-06, |
|
"loss": 0.8747, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 29.454545454545453, |
|
"grad_norm": 1.076412558555603, |
|
"learning_rate": 9.719790845697534e-06, |
|
"loss": 0.7737, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 29.636363636363637, |
|
"grad_norm": 1.0102134943008423, |
|
"learning_rate": 9.71276592485237e-06, |
|
"loss": 0.7632, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 29.818181818181817, |
|
"grad_norm": 1.1170337200164795, |
|
"learning_rate": 9.705656635172418e-06, |
|
"loss": 0.7417, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"grad_norm": 1.0282917022705078, |
|
"learning_rate": 9.698463103929542e-06, |
|
"loss": 0.7312, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_loss": 0.9086909294128418, |
|
"eval_runtime": 9.6293, |
|
"eval_samples_per_second": 2.492, |
|
"eval_steps_per_second": 2.492, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 30.181818181818183, |
|
"grad_norm": 1.050240397453308, |
|
"learning_rate": 9.69118545990371e-06, |
|
"loss": 0.7415, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 30.363636363636363, |
|
"grad_norm": 1.090142011642456, |
|
"learning_rate": 9.683823833380692e-06, |
|
"loss": 0.7346, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 30.545454545454547, |
|
"grad_norm": 1.1241227388381958, |
|
"learning_rate": 9.676378356149733e-06, |
|
"loss": 0.7492, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 30.727272727272727, |
|
"grad_norm": 1.0831716060638428, |
|
"learning_rate": 9.668849161501186e-06, |
|
"loss": 0.7854, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 30.90909090909091, |
|
"grad_norm": 0.9916213154792786, |
|
"learning_rate": 9.66123638422413e-06, |
|
"loss": 0.7431, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_loss": 0.9049713611602783, |
|
"eval_runtime": 9.6277, |
|
"eval_samples_per_second": 2.493, |
|
"eval_steps_per_second": 2.493, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 31.09090909090909, |
|
"grad_norm": 0.9737184047698975, |
|
"learning_rate": 9.653540160603956e-06, |
|
"loss": 0.8, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 31.272727272727273, |
|
"grad_norm": 1.002685308456421, |
|
"learning_rate": 9.64576062841993e-06, |
|
"loss": 0.7644, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 31.454545454545453, |
|
"grad_norm": 1.2203502655029297, |
|
"learning_rate": 9.637897926942716e-06, |
|
"loss": 0.7335, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 31.636363636363637, |
|
"grad_norm": 1.2266589403152466, |
|
"learning_rate": 9.629952196931902e-06, |
|
"loss": 0.7724, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 31.818181818181817, |
|
"grad_norm": 1.3024228811264038, |
|
"learning_rate": 9.621923580633462e-06, |
|
"loss": 0.7164, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"grad_norm": 1.1638389825820923, |
|
"learning_rate": 9.613812221777212e-06, |
|
"loss": 0.7103, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_loss": 0.9036614298820496, |
|
"eval_runtime": 9.6223, |
|
"eval_samples_per_second": 2.494, |
|
"eval_steps_per_second": 2.494, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 32.18181818181818, |
|
"grad_norm": 1.1456729173660278, |
|
"learning_rate": 9.60561826557425e-06, |
|
"loss": 0.7032, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 32.36363636363637, |
|
"grad_norm": 1.0486174821853638, |
|
"learning_rate": 9.597341858714344e-06, |
|
"loss": 0.7216, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 32.54545454545455, |
|
"grad_norm": 1.2181857824325562, |
|
"learning_rate": 9.588983149363307e-06, |
|
"loss": 0.7591, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 32.72727272727273, |
|
"grad_norm": 1.0791609287261963, |
|
"learning_rate": 9.580542287160348e-06, |
|
"loss": 0.7719, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 32.90909090909091, |
|
"grad_norm": 1.1914441585540771, |
|
"learning_rate": 9.572019423215395e-06, |
|
"loss": 0.6967, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_loss": 0.9091615080833435, |
|
"eval_runtime": 9.619, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 33.09090909090909, |
|
"grad_norm": 1.206595778465271, |
|
"learning_rate": 9.563414710106382e-06, |
|
"loss": 0.6713, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 33.27272727272727, |
|
"grad_norm": 1.2539595365524292, |
|
"learning_rate": 9.554728301876525e-06, |
|
"loss": 0.7171, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 33.45454545454545, |
|
"grad_norm": 1.135501503944397, |
|
"learning_rate": 9.545960354031564e-06, |
|
"loss": 0.7324, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 33.63636363636363, |
|
"grad_norm": 1.2056708335876465, |
|
"learning_rate": 9.537111023536973e-06, |
|
"loss": 0.7101, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 33.81818181818182, |
|
"grad_norm": 1.1787803173065186, |
|
"learning_rate": 9.528180468815155e-06, |
|
"loss": 0.7468, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"grad_norm": 1.490050196647644, |
|
"learning_rate": 9.519168849742603e-06, |
|
"loss": 0.6502, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_loss": 0.9070794582366943, |
|
"eval_runtime": 9.63, |
|
"eval_samples_per_second": 2.492, |
|
"eval_steps_per_second": 2.492, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 34.18181818181818, |
|
"grad_norm": 1.2995131015777588, |
|
"learning_rate": 9.510076327647043e-06, |
|
"loss": 0.6144, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 34.36363636363637, |
|
"grad_norm": 1.3441383838653564, |
|
"learning_rate": 9.50090306530454e-06, |
|
"loss": 0.7332, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 34.54545454545455, |
|
"grad_norm": 1.1878114938735962, |
|
"learning_rate": 9.491649226936586e-06, |
|
"loss": 0.7056, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 34.72727272727273, |
|
"grad_norm": 1.3715388774871826, |
|
"learning_rate": 9.48231497820716e-06, |
|
"loss": 0.67, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 34.90909090909091, |
|
"grad_norm": 1.466930866241455, |
|
"learning_rate": 9.47290048621977e-06, |
|
"loss": 0.6659, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_loss": 0.9019081592559814, |
|
"eval_runtime": 9.6208, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 35.09090909090909, |
|
"grad_norm": 1.4354052543640137, |
|
"learning_rate": 9.46340591951444e-06, |
|
"loss": 0.7589, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 35.27272727272727, |
|
"grad_norm": 1.2188067436218262, |
|
"learning_rate": 9.453831448064717e-06, |
|
"loss": 0.6215, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 35.45454545454545, |
|
"grad_norm": 1.3039608001708984, |
|
"learning_rate": 9.444177243274619e-06, |
|
"loss": 0.6502, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 35.63636363636363, |
|
"grad_norm": 1.3067864179611206, |
|
"learning_rate": 9.434443477975557e-06, |
|
"loss": 0.6884, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 35.81818181818182, |
|
"grad_norm": 1.33348548412323, |
|
"learning_rate": 9.42463032642326e-06, |
|
"loss": 0.6852, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"grad_norm": 1.5278242826461792, |
|
"learning_rate": 9.414737964294636e-06, |
|
"loss": 0.7003, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_loss": 0.9015458226203918, |
|
"eval_runtime": 9.6251, |
|
"eval_samples_per_second": 2.493, |
|
"eval_steps_per_second": 2.493, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 36.18181818181818, |
|
"grad_norm": 1.3526606559753418, |
|
"learning_rate": 9.40476656868464e-06, |
|
"loss": 0.6329, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 36.36363636363637, |
|
"grad_norm": 1.6108198165893555, |
|
"learning_rate": 9.394716318103098e-06, |
|
"loss": 0.6626, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 36.54545454545455, |
|
"grad_norm": 1.5193760395050049, |
|
"learning_rate": 9.384587392471516e-06, |
|
"loss": 0.6816, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 36.72727272727273, |
|
"grad_norm": 1.2977081537246704, |
|
"learning_rate": 9.37437997311985e-06, |
|
"loss": 0.6422, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 36.90909090909091, |
|
"grad_norm": 1.4797639846801758, |
|
"learning_rate": 9.364094242783272e-06, |
|
"loss": 0.629, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_loss": 0.9018394351005554, |
|
"eval_runtime": 9.6212, |
|
"eval_samples_per_second": 2.494, |
|
"eval_steps_per_second": 2.494, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 37.09090909090909, |
|
"grad_norm": 1.2966375350952148, |
|
"learning_rate": 9.353730385598887e-06, |
|
"loss": 0.6654, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 37.27272727272727, |
|
"grad_norm": 1.3821908235549927, |
|
"learning_rate": 9.343288587102444e-06, |
|
"loss": 0.6054, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 37.45454545454545, |
|
"grad_norm": 1.3437285423278809, |
|
"learning_rate": 9.332769034225012e-06, |
|
"loss": 0.6449, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 37.63636363636363, |
|
"grad_norm": 1.4815727472305298, |
|
"learning_rate": 9.322171915289635e-06, |
|
"loss": 0.658, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 37.81818181818182, |
|
"grad_norm": 1.273409366607666, |
|
"learning_rate": 9.311497420007955e-06, |
|
"loss": 0.6629, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"grad_norm": 1.4128069877624512, |
|
"learning_rate": 9.30074573947683e-06, |
|
"loss": 0.6299, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_loss": 0.9080850481987, |
|
"eval_runtime": 9.6337, |
|
"eval_samples_per_second": 2.491, |
|
"eval_steps_per_second": 2.491, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 38.18181818181818, |
|
"grad_norm": 1.467628836631775, |
|
"learning_rate": 9.289917066174887e-06, |
|
"loss": 0.6516, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 38.36363636363637, |
|
"grad_norm": 1.4847270250320435, |
|
"learning_rate": 9.279011593959107e-06, |
|
"loss": 0.6093, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 38.54545454545455, |
|
"grad_norm": 1.7236661911010742, |
|
"learning_rate": 9.268029518061335e-06, |
|
"loss": 0.6272, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 38.72727272727273, |
|
"grad_norm": 1.563481330871582, |
|
"learning_rate": 9.256971035084786e-06, |
|
"loss": 0.6293, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 38.90909090909091, |
|
"grad_norm": 1.4896105527877808, |
|
"learning_rate": 9.245836343000534e-06, |
|
"loss": 0.6259, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_loss": 0.9161927103996277, |
|
"eval_runtime": 9.6218, |
|
"eval_samples_per_second": 2.494, |
|
"eval_steps_per_second": 2.494, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 39.09090909090909, |
|
"grad_norm": 1.36057448387146, |
|
"learning_rate": 9.234625641143962e-06, |
|
"loss": 0.5801, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 39.27272727272727, |
|
"grad_norm": 1.5864324569702148, |
|
"learning_rate": 9.223339130211194e-06, |
|
"loss": 0.5785, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 39.45454545454545, |
|
"grad_norm": 1.609424352645874, |
|
"learning_rate": 9.211977012255497e-06, |
|
"loss": 0.5777, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 39.63636363636363, |
|
"grad_norm": 1.4554574489593506, |
|
"learning_rate": 9.200539490683682e-06, |
|
"loss": 0.5927, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 39.81818181818182, |
|
"grad_norm": 1.5687090158462524, |
|
"learning_rate": 9.189026770252437e-06, |
|
"loss": 0.6421, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"grad_norm": 1.4782568216323853, |
|
"learning_rate": 9.177439057064684e-06, |
|
"loss": 0.6262, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_loss": 0.9212110638618469, |
|
"eval_runtime": 9.6258, |
|
"eval_samples_per_second": 2.493, |
|
"eval_steps_per_second": 2.493, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 40.18181818181818, |
|
"grad_norm": 1.4512465000152588, |
|
"learning_rate": 9.16577655856587e-06, |
|
"loss": 0.5827, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 40.36363636363637, |
|
"grad_norm": 2.0493810176849365, |
|
"learning_rate": 9.154039483540273e-06, |
|
"loss": 0.5539, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 40.54545454545455, |
|
"grad_norm": 1.6594229936599731, |
|
"learning_rate": 9.142228042107248e-06, |
|
"loss": 0.5907, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 40.72727272727273, |
|
"grad_norm": 1.4589492082595825, |
|
"learning_rate": 9.130342445717474e-06, |
|
"loss": 0.6161, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 40.90909090909091, |
|
"grad_norm": 1.7103596925735474, |
|
"learning_rate": 9.118382907149164e-06, |
|
"loss": 0.5707, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"eval_loss": 0.9212350845336914, |
|
"eval_runtime": 9.6173, |
|
"eval_samples_per_second": 2.495, |
|
"eval_steps_per_second": 2.495, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 41.09090909090909, |
|
"grad_norm": 1.3275471925735474, |
|
"learning_rate": 9.10634964050426e-06, |
|
"loss": 0.6536, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 41.27272727272727, |
|
"grad_norm": 1.6759672164916992, |
|
"learning_rate": 9.094242861204598e-06, |
|
"loss": 0.5257, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 41.45454545454545, |
|
"grad_norm": 1.9791761636734009, |
|
"learning_rate": 9.08206278598805e-06, |
|
"loss": 0.5604, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 41.63636363636363, |
|
"grad_norm": 1.740043044090271, |
|
"learning_rate": 9.069809632904647e-06, |
|
"loss": 0.544, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 41.81818181818182, |
|
"grad_norm": 1.7312424182891846, |
|
"learning_rate": 9.057483621312671e-06, |
|
"loss": 0.5987, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"grad_norm": 1.5688552856445312, |
|
"learning_rate": 9.045084971874738e-06, |
|
"loss": 0.5749, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"eval_loss": 0.927391529083252, |
|
"eval_runtime": 9.6244, |
|
"eval_samples_per_second": 2.494, |
|
"eval_steps_per_second": 2.494, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 42.18181818181818, |
|
"grad_norm": 1.69509756565094, |
|
"learning_rate": 9.032613906553833e-06, |
|
"loss": 0.6091, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 42.36363636363637, |
|
"grad_norm": 2.032390832901001, |
|
"learning_rate": 9.020070648609347e-06, |
|
"loss": 0.5022, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 42.54545454545455, |
|
"grad_norm": 1.9728387594223022, |
|
"learning_rate": 9.007455422593077e-06, |
|
"loss": 0.5395, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 42.72727272727273, |
|
"grad_norm": 1.7745451927185059, |
|
"learning_rate": 8.994768454345207e-06, |
|
"loss": 0.5662, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 42.90909090909091, |
|
"grad_norm": 1.9252541065216064, |
|
"learning_rate": 8.982009970990262e-06, |
|
"loss": 0.533, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"eval_loss": 0.9368504881858826, |
|
"eval_runtime": 9.6216, |
|
"eval_samples_per_second": 2.494, |
|
"eval_steps_per_second": 2.494, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"step": 473, |
|
"total_flos": 5.808369757165978e+16, |
|
"train_loss": 1.1408403902678863, |
|
"train_runtime": 4431.3714, |
|
"train_samples_per_second": 2.979, |
|
"train_steps_per_second": 0.372 |
|
} |
|
], |
|
"logging_steps": 2, |
|
"max_steps": 1650, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 150, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 7, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5.808369757165978e+16, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|