|
{ |
|
"best_metric": 0.8382187147688839, |
|
"best_model_checkpoint": "vit-base-patch32-224-in21k-finetuned-galaxy10-decals/checkpoint-1871", |
|
"epoch": 29.819639278557116, |
|
"eval_steps": 500, |
|
"global_step": 3720, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 2.688172043010753e-06, |
|
"loss": 2.3046, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 5.376344086021506e-06, |
|
"loss": 2.2944, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.064516129032258e-06, |
|
"loss": 2.2619, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.0752688172043012e-05, |
|
"loss": 2.2004, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.3440860215053763e-05, |
|
"loss": 2.1433, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.6129032258064517e-05, |
|
"loss": 2.0364, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.881720430107527e-05, |
|
"loss": 1.9357, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.1505376344086024e-05, |
|
"loss": 1.8297, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.4193548387096777e-05, |
|
"loss": 1.7391, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.6881720430107527e-05, |
|
"loss": 1.638, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.9569892473118284e-05, |
|
"loss": 1.5664, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.2258064516129034e-05, |
|
"loss": 1.4731, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"eval_accuracy": 0.6110484780157835, |
|
"eval_f1": 0.5796642871948662, |
|
"eval_loss": 1.385038137435913, |
|
"eval_precision": 0.5791356846200767, |
|
"eval_recall": 0.6110484780157835, |
|
"eval_runtime": 11.3228, |
|
"eval_samples_per_second": 156.675, |
|
"eval_steps_per_second": 4.946, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 3.494623655913979e-05, |
|
"loss": 1.4448, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 3.763440860215054e-05, |
|
"loss": 1.3659, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 4.032258064516129e-05, |
|
"loss": 1.2679, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 4.301075268817205e-05, |
|
"loss": 1.2652, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 4.56989247311828e-05, |
|
"loss": 1.1719, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 4.8387096774193554e-05, |
|
"loss": 1.1752, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 5.1075268817204304e-05, |
|
"loss": 1.0939, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 5.3763440860215054e-05, |
|
"loss": 1.1118, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 5.645161290322582e-05, |
|
"loss": 1.0168, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 5.913978494623657e-05, |
|
"loss": 1.0283, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 6.182795698924732e-05, |
|
"loss": 1.0155, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 6.451612903225807e-05, |
|
"loss": 0.9858, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.7508455467869222, |
|
"eval_f1": 0.7336567894074051, |
|
"eval_loss": 0.8899810910224915, |
|
"eval_precision": 0.757814640272845, |
|
"eval_recall": 0.7508455467869222, |
|
"eval_runtime": 11.3092, |
|
"eval_samples_per_second": 156.864, |
|
"eval_steps_per_second": 4.952, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 6.720430107526882e-05, |
|
"loss": 1.0038, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 6.989247311827958e-05, |
|
"loss": 0.9533, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 7.258064516129033e-05, |
|
"loss": 0.9765, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 7.526881720430108e-05, |
|
"loss": 0.9596, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 7.795698924731183e-05, |
|
"loss": 0.9075, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 8.064516129032258e-05, |
|
"loss": 0.9168, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 8.333333333333334e-05, |
|
"loss": 0.9543, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 8.60215053763441e-05, |
|
"loss": 0.9058, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 8.870967741935484e-05, |
|
"loss": 0.8467, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 9.13978494623656e-05, |
|
"loss": 0.9296, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 9.408602150537636e-05, |
|
"loss": 0.8902, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 9.677419354838711e-05, |
|
"loss": 0.8669, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 9.946236559139786e-05, |
|
"loss": 0.9475, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.7598647125140925, |
|
"eval_f1": 0.7558619248229688, |
|
"eval_loss": 0.7799468636512756, |
|
"eval_precision": 0.7666628356432069, |
|
"eval_recall": 0.7598647125140925, |
|
"eval_runtime": 11.311, |
|
"eval_samples_per_second": 156.839, |
|
"eval_steps_per_second": 4.951, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 9.97610513739546e-05, |
|
"loss": 0.8879, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 9.946236559139786e-05, |
|
"loss": 0.8372, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 9.916367980884111e-05, |
|
"loss": 0.8123, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 9.886499402628435e-05, |
|
"loss": 0.8235, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 9.85663082437276e-05, |
|
"loss": 0.8051, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 9.826762246117085e-05, |
|
"loss": 0.8667, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 9.79689366786141e-05, |
|
"loss": 0.8102, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 9.767025089605735e-05, |
|
"loss": 0.7919, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 9.73715651135006e-05, |
|
"loss": 0.7995, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 9.707287933094386e-05, |
|
"loss": 0.8217, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 9.677419354838711e-05, |
|
"loss": 0.783, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 9.647550776583036e-05, |
|
"loss": 0.7778, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.7779030439684329, |
|
"eval_f1": 0.7728962693294915, |
|
"eval_loss": 0.6798376441001892, |
|
"eval_precision": 0.7825312914978793, |
|
"eval_recall": 0.7779030439684329, |
|
"eval_runtime": 11.2463, |
|
"eval_samples_per_second": 157.741, |
|
"eval_steps_per_second": 4.979, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 9.61768219832736e-05, |
|
"loss": 0.7619, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 9.587813620071685e-05, |
|
"loss": 0.8295, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 9.55794504181601e-05, |
|
"loss": 0.7672, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 9.528076463560335e-05, |
|
"loss": 0.7742, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 9.49820788530466e-05, |
|
"loss": 0.7288, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 9.468339307048984e-05, |
|
"loss": 0.7885, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 9.438470728793309e-05, |
|
"loss": 0.7272, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 9.408602150537636e-05, |
|
"loss": 0.7187, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 9.378733572281961e-05, |
|
"loss": 0.6938, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 9.348864994026285e-05, |
|
"loss": 0.6948, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 9.31899641577061e-05, |
|
"loss": 0.7753, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 9.289127837514935e-05, |
|
"loss": 0.6739, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 9.25925925925926e-05, |
|
"loss": 0.6831, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"eval_accuracy": 0.7914317925591883, |
|
"eval_f1": 0.7888654936777544, |
|
"eval_loss": 0.6352207660675049, |
|
"eval_precision": 0.7915751085761195, |
|
"eval_recall": 0.7914317925591883, |
|
"eval_runtime": 11.2858, |
|
"eval_samples_per_second": 157.189, |
|
"eval_steps_per_second": 4.962, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 9.229390681003584e-05, |
|
"loss": 0.7109, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"learning_rate": 9.199522102747909e-05, |
|
"loss": 0.7, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 9.169653524492234e-05, |
|
"loss": 0.7144, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 5.29, |
|
"learning_rate": 9.13978494623656e-05, |
|
"loss": 0.7142, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 9.109916367980885e-05, |
|
"loss": 0.6764, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"learning_rate": 9.080047789725208e-05, |
|
"loss": 0.6753, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 9.050179211469535e-05, |
|
"loss": 0.6559, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 5.61, |
|
"learning_rate": 9.02031063321386e-05, |
|
"loss": 0.7044, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 8.990442054958185e-05, |
|
"loss": 0.6263, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 5.77, |
|
"learning_rate": 8.960573476702509e-05, |
|
"loss": 0.6242, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 5.85, |
|
"learning_rate": 8.930704898446834e-05, |
|
"loss": 0.7016, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 8.900836320191159e-05, |
|
"loss": 0.6953, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_accuracy": 0.8043968432919955, |
|
"eval_f1": 0.8022664787526015, |
|
"eval_loss": 0.5931026339530945, |
|
"eval_precision": 0.8076356387187705, |
|
"eval_recall": 0.8043968432919955, |
|
"eval_runtime": 11.3344, |
|
"eval_samples_per_second": 156.515, |
|
"eval_steps_per_second": 4.941, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 8.870967741935484e-05, |
|
"loss": 0.6853, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 8.84109916367981e-05, |
|
"loss": 0.6455, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 6.17, |
|
"learning_rate": 8.811230585424133e-05, |
|
"loss": 0.6514, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 8.781362007168459e-05, |
|
"loss": 0.6451, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"learning_rate": 8.751493428912784e-05, |
|
"loss": 0.6141, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 6.41, |
|
"learning_rate": 8.72162485065711e-05, |
|
"loss": 0.6374, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 6.49, |
|
"learning_rate": 8.691756272401434e-05, |
|
"loss": 0.6375, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 6.57, |
|
"learning_rate": 8.661887694145759e-05, |
|
"loss": 0.6502, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 8.632019115890084e-05, |
|
"loss": 0.6361, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 8.60215053763441e-05, |
|
"loss": 0.6389, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 8.572281959378735e-05, |
|
"loss": 0.6508, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 6.89, |
|
"learning_rate": 8.542413381123058e-05, |
|
"loss": 0.6195, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 8.512544802867384e-05, |
|
"loss": 0.6725, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_accuracy": 0.7536640360766629, |
|
"eval_f1": 0.7519458839777032, |
|
"eval_loss": 0.730423629283905, |
|
"eval_precision": 0.7670637522757748, |
|
"eval_recall": 0.7536640360766629, |
|
"eval_runtime": 11.4098, |
|
"eval_samples_per_second": 155.481, |
|
"eval_steps_per_second": 4.908, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"learning_rate": 8.482676224611709e-05, |
|
"loss": 0.6617, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"learning_rate": 8.452807646356034e-05, |
|
"loss": 0.6533, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 7.21, |
|
"learning_rate": 8.422939068100359e-05, |
|
"loss": 0.6591, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 8.393070489844683e-05, |
|
"loss": 0.5751, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 8.363201911589009e-05, |
|
"loss": 0.6242, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 7.45, |
|
"learning_rate": 8.333333333333334e-05, |
|
"loss": 0.6356, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 7.54, |
|
"learning_rate": 8.30346475507766e-05, |
|
"loss": 0.6538, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 7.62, |
|
"learning_rate": 8.273596176821983e-05, |
|
"loss": 0.6228, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 8.243727598566309e-05, |
|
"loss": 0.5677, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 8.213859020310634e-05, |
|
"loss": 0.5748, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 8.183990442054959e-05, |
|
"loss": 0.5663, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 7.94, |
|
"learning_rate": 8.154121863799284e-05, |
|
"loss": 0.5648, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_accuracy": 0.7908680947012401, |
|
"eval_f1": 0.786803166085158, |
|
"eval_loss": 0.6351723670959473, |
|
"eval_precision": 0.7961365198849726, |
|
"eval_recall": 0.7908680947012401, |
|
"eval_runtime": 11.3517, |
|
"eval_samples_per_second": 156.276, |
|
"eval_steps_per_second": 4.933, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 8.124253285543608e-05, |
|
"loss": 0.5992, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 8.1, |
|
"learning_rate": 8.094384707287933e-05, |
|
"loss": 0.5262, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 8.18, |
|
"learning_rate": 8.064516129032258e-05, |
|
"loss": 0.5595, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 8.26, |
|
"learning_rate": 8.034647550776585e-05, |
|
"loss": 0.5926, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 8.34, |
|
"learning_rate": 8.004778972520908e-05, |
|
"loss": 0.5816, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 8.42, |
|
"learning_rate": 7.974910394265234e-05, |
|
"loss": 0.5985, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 7.945041816009559e-05, |
|
"loss": 0.629, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 8.58, |
|
"learning_rate": 7.915173237753884e-05, |
|
"loss": 0.6118, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 8.66, |
|
"learning_rate": 7.885304659498209e-05, |
|
"loss": 0.586, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 8.74, |
|
"learning_rate": 7.855436081242533e-05, |
|
"loss": 0.6238, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 8.82, |
|
"learning_rate": 7.825567502986858e-05, |
|
"loss": 0.5759, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 8.9, |
|
"learning_rate": 7.795698924731183e-05, |
|
"loss": 0.6095, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 8.98, |
|
"learning_rate": 7.765830346475508e-05, |
|
"loss": 0.6127, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 8.99, |
|
"eval_accuracy": 0.7857948139797069, |
|
"eval_f1": 0.7819663913155567, |
|
"eval_loss": 0.6087137460708618, |
|
"eval_precision": 0.7878518240325064, |
|
"eval_recall": 0.7857948139797069, |
|
"eval_runtime": 11.3142, |
|
"eval_samples_per_second": 156.794, |
|
"eval_steps_per_second": 4.95, |
|
"step": 1122 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 7.735961768219832e-05, |
|
"loss": 0.5814, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 7.706093189964157e-05, |
|
"loss": 0.5711, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 9.22, |
|
"learning_rate": 7.676224611708484e-05, |
|
"loss": 0.5915, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 9.3, |
|
"learning_rate": 7.646356033452809e-05, |
|
"loss": 0.5142, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 9.38, |
|
"learning_rate": 7.616487455197133e-05, |
|
"loss": 0.5525, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 7.586618876941458e-05, |
|
"loss": 0.566, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 9.54, |
|
"learning_rate": 7.556750298685783e-05, |
|
"loss": 0.5837, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 9.62, |
|
"learning_rate": 7.526881720430108e-05, |
|
"loss": 0.5202, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 9.7, |
|
"learning_rate": 7.497013142174433e-05, |
|
"loss": 0.5232, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 9.78, |
|
"learning_rate": 7.467144563918757e-05, |
|
"loss": 0.5185, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 7.437275985663082e-05, |
|
"loss": 0.5759, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 9.94, |
|
"learning_rate": 7.407407407407407e-05, |
|
"loss": 0.529, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_accuracy": 0.8072153325817362, |
|
"eval_f1": 0.8040653133107246, |
|
"eval_loss": 0.5826550126075745, |
|
"eval_precision": 0.8074015727126425, |
|
"eval_recall": 0.8072153325817362, |
|
"eval_runtime": 11.2959, |
|
"eval_samples_per_second": 157.048, |
|
"eval_steps_per_second": 4.958, |
|
"step": 1247 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 7.377538829151732e-05, |
|
"loss": 0.5447, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 10.1, |
|
"learning_rate": 7.347670250896058e-05, |
|
"loss": 0.529, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 10.18, |
|
"learning_rate": 7.317801672640383e-05, |
|
"loss": 0.5399, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 10.26, |
|
"learning_rate": 7.287933094384708e-05, |
|
"loss": 0.5181, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 10.34, |
|
"learning_rate": 7.258064516129033e-05, |
|
"loss": 0.5674, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 10.42, |
|
"learning_rate": 7.228195937873358e-05, |
|
"loss": 0.5519, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 10.5, |
|
"learning_rate": 7.198327359617682e-05, |
|
"loss": 0.5313, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 10.58, |
|
"learning_rate": 7.168458781362007e-05, |
|
"loss": 0.5416, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 10.66, |
|
"learning_rate": 7.138590203106332e-05, |
|
"loss": 0.517, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 10.74, |
|
"learning_rate": 7.108721624850657e-05, |
|
"loss": 0.524, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 10.82, |
|
"learning_rate": 7.078853046594983e-05, |
|
"loss": 0.5312, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 10.9, |
|
"learning_rate": 7.048984468339306e-05, |
|
"loss": 0.4881, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 10.98, |
|
"learning_rate": 7.019115890083633e-05, |
|
"loss": 0.5212, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_accuracy": 0.8179255918827508, |
|
"eval_f1": 0.8107896552767364, |
|
"eval_loss": 0.5787181258201599, |
|
"eval_precision": 0.8177313788238276, |
|
"eval_recall": 0.8179255918827508, |
|
"eval_runtime": 11.3993, |
|
"eval_samples_per_second": 155.624, |
|
"eval_steps_per_second": 4.913, |
|
"step": 1372 |
|
}, |
|
{ |
|
"epoch": 11.06, |
|
"learning_rate": 6.989247311827958e-05, |
|
"loss": 0.5224, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 11.14, |
|
"learning_rate": 6.959378733572283e-05, |
|
"loss": 0.4974, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 11.22, |
|
"learning_rate": 6.929510155316607e-05, |
|
"loss": 0.5288, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 11.3, |
|
"learning_rate": 6.899641577060932e-05, |
|
"loss": 0.4737, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 11.38, |
|
"learning_rate": 6.869772998805257e-05, |
|
"loss": 0.5013, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 11.46, |
|
"learning_rate": 6.839904420549582e-05, |
|
"loss": 0.483, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 11.54, |
|
"learning_rate": 6.810035842293908e-05, |
|
"loss": 0.5051, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 11.62, |
|
"learning_rate": 6.780167264038231e-05, |
|
"loss": 0.5138, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 11.7, |
|
"learning_rate": 6.750298685782557e-05, |
|
"loss": 0.5337, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 11.78, |
|
"learning_rate": 6.720430107526882e-05, |
|
"loss": 0.4946, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 11.86, |
|
"learning_rate": 6.690561529271207e-05, |
|
"loss": 0.4811, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 11.94, |
|
"learning_rate": 6.660692951015532e-05, |
|
"loss": 0.4665, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_accuracy": 0.8167981961668546, |
|
"eval_f1": 0.8134368902882168, |
|
"eval_loss": 0.5597419142723083, |
|
"eval_precision": 0.8213118777363821, |
|
"eval_recall": 0.8167981961668546, |
|
"eval_runtime": 11.3339, |
|
"eval_samples_per_second": 156.521, |
|
"eval_steps_per_second": 4.941, |
|
"step": 1497 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 6.630824372759857e-05, |
|
"loss": 0.5099, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 12.1, |
|
"learning_rate": 6.600955794504182e-05, |
|
"loss": 0.4931, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 12.18, |
|
"learning_rate": 6.571087216248507e-05, |
|
"loss": 0.4414, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 12.26, |
|
"learning_rate": 6.541218637992833e-05, |
|
"loss": 0.4794, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 12.34, |
|
"learning_rate": 6.511350059737156e-05, |
|
"loss": 0.4742, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 12.42, |
|
"learning_rate": 6.481481481481482e-05, |
|
"loss": 0.4989, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 12.51, |
|
"learning_rate": 6.451612903225807e-05, |
|
"loss": 0.4863, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 12.59, |
|
"learning_rate": 6.421744324970132e-05, |
|
"loss": 0.5107, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 12.67, |
|
"learning_rate": 6.391875746714456e-05, |
|
"loss": 0.5382, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 12.75, |
|
"learning_rate": 6.362007168458781e-05, |
|
"loss": 0.5014, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 12.83, |
|
"learning_rate": 6.332138590203107e-05, |
|
"loss": 0.5203, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 12.91, |
|
"learning_rate": 6.302270011947432e-05, |
|
"loss": 0.4916, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 12.99, |
|
"learning_rate": 6.272401433691756e-05, |
|
"loss": 0.5123, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 12.99, |
|
"eval_accuracy": 0.8043968432919955, |
|
"eval_f1": 0.8044262875941992, |
|
"eval_loss": 0.5839796662330627, |
|
"eval_precision": 0.8162621736409407, |
|
"eval_recall": 0.8043968432919955, |
|
"eval_runtime": 11.3115, |
|
"eval_samples_per_second": 156.832, |
|
"eval_steps_per_second": 4.951, |
|
"step": 1621 |
|
}, |
|
{ |
|
"epoch": 13.07, |
|
"learning_rate": 6.242532855436081e-05, |
|
"loss": 0.4817, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 13.15, |
|
"learning_rate": 6.212664277180407e-05, |
|
"loss": 0.4517, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 13.23, |
|
"learning_rate": 6.182795698924732e-05, |
|
"loss": 0.4581, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 13.31, |
|
"learning_rate": 6.152927120669057e-05, |
|
"loss": 0.547, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 13.39, |
|
"learning_rate": 6.12305854241338e-05, |
|
"loss": 0.503, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 13.47, |
|
"learning_rate": 6.093189964157706e-05, |
|
"loss": 0.4547, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 13.55, |
|
"learning_rate": 6.063321385902031e-05, |
|
"loss": 0.4799, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 13.63, |
|
"learning_rate": 6.033452807646356e-05, |
|
"loss": 0.458, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 13.71, |
|
"learning_rate": 6.0035842293906806e-05, |
|
"loss": 0.4721, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 13.79, |
|
"learning_rate": 5.9737156511350064e-05, |
|
"loss": 0.4568, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 13.87, |
|
"learning_rate": 5.9438470728793316e-05, |
|
"loss": 0.4924, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 13.95, |
|
"learning_rate": 5.913978494623657e-05, |
|
"loss": 0.4918, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_accuracy": 0.8218714768883878, |
|
"eval_f1": 0.8195381684600302, |
|
"eval_loss": 0.5591737031936646, |
|
"eval_precision": 0.822102163318251, |
|
"eval_recall": 0.8218714768883878, |
|
"eval_runtime": 11.2939, |
|
"eval_samples_per_second": 157.076, |
|
"eval_steps_per_second": 4.958, |
|
"step": 1746 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 5.884109916367981e-05, |
|
"loss": 0.4604, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 14.11, |
|
"learning_rate": 5.8542413381123063e-05, |
|
"loss": 0.4672, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 14.19, |
|
"learning_rate": 5.824372759856631e-05, |
|
"loss": 0.4585, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 14.27, |
|
"learning_rate": 5.794504181600956e-05, |
|
"loss": 0.445, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 14.35, |
|
"learning_rate": 5.764635603345281e-05, |
|
"loss": 0.4283, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 14.43, |
|
"learning_rate": 5.7347670250896056e-05, |
|
"loss": 0.4702, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 14.51, |
|
"learning_rate": 5.704898446833931e-05, |
|
"loss": 0.4795, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 14.59, |
|
"learning_rate": 5.675029868578255e-05, |
|
"loss": 0.4968, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 14.67, |
|
"learning_rate": 5.645161290322582e-05, |
|
"loss": 0.487, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 14.75, |
|
"learning_rate": 5.615292712066906e-05, |
|
"loss": 0.433, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 14.83, |
|
"learning_rate": 5.5854241338112313e-05, |
|
"loss": 0.4273, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 14.91, |
|
"learning_rate": 5.555555555555556e-05, |
|
"loss": 0.4672, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 14.99, |
|
"learning_rate": 5.525686977299881e-05, |
|
"loss": 0.4733, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_accuracy": 0.8382187147688839, |
|
"eval_f1": 0.8346153815083935, |
|
"eval_loss": 0.5180031657218933, |
|
"eval_precision": 0.8362802583563215, |
|
"eval_recall": 0.8382187147688839, |
|
"eval_runtime": 11.3151, |
|
"eval_samples_per_second": 156.781, |
|
"eval_steps_per_second": 4.949, |
|
"step": 1871 |
|
}, |
|
{ |
|
"epoch": 15.07, |
|
"learning_rate": 5.495818399044206e-05, |
|
"loss": 0.4284, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 15.15, |
|
"learning_rate": 5.4659498207885306e-05, |
|
"loss": 0.4298, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 15.23, |
|
"learning_rate": 5.436081242532856e-05, |
|
"loss": 0.4165, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 15.31, |
|
"learning_rate": 5.40621266427718e-05, |
|
"loss": 0.4462, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 15.39, |
|
"learning_rate": 5.3763440860215054e-05, |
|
"loss": 0.4293, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 15.47, |
|
"learning_rate": 5.34647550776583e-05, |
|
"loss": 0.4903, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 15.55, |
|
"learning_rate": 5.316606929510155e-05, |
|
"loss": 0.4258, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 15.63, |
|
"learning_rate": 5.286738351254481e-05, |
|
"loss": 0.4424, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 15.71, |
|
"learning_rate": 5.256869772998806e-05, |
|
"loss": 0.426, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 15.79, |
|
"learning_rate": 5.2270011947431304e-05, |
|
"loss": 0.4478, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 15.87, |
|
"learning_rate": 5.1971326164874556e-05, |
|
"loss": 0.4513, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 15.95, |
|
"learning_rate": 5.167264038231781e-05, |
|
"loss": 0.4552, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_accuracy": 0.8173618940248027, |
|
"eval_f1": 0.8153338114415309, |
|
"eval_loss": 0.5673156380653381, |
|
"eval_precision": 0.8181218548788447, |
|
"eval_recall": 0.8173618940248027, |
|
"eval_runtime": 11.3073, |
|
"eval_samples_per_second": 156.89, |
|
"eval_steps_per_second": 4.953, |
|
"step": 1996 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 5.137395459976105e-05, |
|
"loss": 0.4004, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 16.11, |
|
"learning_rate": 5.1075268817204304e-05, |
|
"loss": 0.387, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 16.19, |
|
"learning_rate": 5.077658303464755e-05, |
|
"loss": 0.4145, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 16.27, |
|
"learning_rate": 5.04778972520908e-05, |
|
"loss": 0.4012, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 16.35, |
|
"learning_rate": 5.017921146953405e-05, |
|
"loss": 0.3978, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 16.43, |
|
"learning_rate": 4.98805256869773e-05, |
|
"loss": 0.4254, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 16.51, |
|
"learning_rate": 4.9581839904420555e-05, |
|
"loss": 0.4249, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 16.59, |
|
"learning_rate": 4.92831541218638e-05, |
|
"loss": 0.4594, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 16.67, |
|
"learning_rate": 4.898446833930705e-05, |
|
"loss": 0.4115, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 16.75, |
|
"learning_rate": 4.86857825567503e-05, |
|
"loss": 0.4144, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 16.83, |
|
"learning_rate": 4.8387096774193554e-05, |
|
"loss": 0.4057, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 16.91, |
|
"learning_rate": 4.80884109916368e-05, |
|
"loss": 0.3639, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 16.99, |
|
"learning_rate": 4.778972520908005e-05, |
|
"loss": 0.4004, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 16.99, |
|
"eval_accuracy": 0.822435174746336, |
|
"eval_f1": 0.8199245109924258, |
|
"eval_loss": 0.5711179971694946, |
|
"eval_precision": 0.8239497086292771, |
|
"eval_recall": 0.822435174746336, |
|
"eval_runtime": 11.3062, |
|
"eval_samples_per_second": 156.905, |
|
"eval_steps_per_second": 4.953, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 17.07, |
|
"learning_rate": 4.74910394265233e-05, |
|
"loss": 0.4155, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 17.15, |
|
"learning_rate": 4.7192353643966546e-05, |
|
"loss": 0.3781, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 17.23, |
|
"learning_rate": 4.6893667861409805e-05, |
|
"loss": 0.3771, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 17.31, |
|
"learning_rate": 4.659498207885305e-05, |
|
"loss": 0.4308, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 17.39, |
|
"learning_rate": 4.62962962962963e-05, |
|
"loss": 0.4381, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 17.47, |
|
"learning_rate": 4.5997610513739546e-05, |
|
"loss": 0.3867, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 17.56, |
|
"learning_rate": 4.56989247311828e-05, |
|
"loss": 0.4103, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 17.64, |
|
"learning_rate": 4.540023894862604e-05, |
|
"loss": 0.3991, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 17.72, |
|
"learning_rate": 4.51015531660693e-05, |
|
"loss": 0.421, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 17.8, |
|
"learning_rate": 4.4802867383512545e-05, |
|
"loss": 0.3761, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 17.88, |
|
"learning_rate": 4.4504181600955796e-05, |
|
"loss": 0.4077, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 17.96, |
|
"learning_rate": 4.420549581839905e-05, |
|
"loss": 0.3359, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_accuracy": 0.8167981961668546, |
|
"eval_f1": 0.8147039132954932, |
|
"eval_loss": 0.581308901309967, |
|
"eval_precision": 0.8152621935521966, |
|
"eval_recall": 0.8167981961668546, |
|
"eval_runtime": 11.3377, |
|
"eval_samples_per_second": 156.47, |
|
"eval_steps_per_second": 4.939, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 18.04, |
|
"learning_rate": 4.390681003584229e-05, |
|
"loss": 0.4214, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 18.12, |
|
"learning_rate": 4.360812425328555e-05, |
|
"loss": 0.3672, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 18.2, |
|
"learning_rate": 4.3309438470728796e-05, |
|
"loss": 0.3448, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 18.28, |
|
"learning_rate": 4.301075268817205e-05, |
|
"loss": 0.37, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 18.36, |
|
"learning_rate": 4.271206690561529e-05, |
|
"loss": 0.3992, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 18.44, |
|
"learning_rate": 4.241338112305854e-05, |
|
"loss": 0.3518, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 18.52, |
|
"learning_rate": 4.2114695340501795e-05, |
|
"loss": 0.3397, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 18.6, |
|
"learning_rate": 4.1816009557945046e-05, |
|
"loss": 0.3999, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 18.68, |
|
"learning_rate": 4.15173237753883e-05, |
|
"loss": 0.3643, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 18.76, |
|
"learning_rate": 4.121863799283154e-05, |
|
"loss": 0.3577, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 18.84, |
|
"learning_rate": 4.0919952210274794e-05, |
|
"loss": 0.3637, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 18.92, |
|
"learning_rate": 4.062126642771804e-05, |
|
"loss": 0.3573, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"learning_rate": 4.032258064516129e-05, |
|
"loss": 0.4069, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_accuracy": 0.8342728297632469, |
|
"eval_f1": 0.8307296947323864, |
|
"eval_loss": 0.5482383370399475, |
|
"eval_precision": 0.8352289890978287, |
|
"eval_recall": 0.8342728297632469, |
|
"eval_runtime": 11.3975, |
|
"eval_samples_per_second": 155.649, |
|
"eval_steps_per_second": 4.913, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 19.08, |
|
"learning_rate": 4.002389486260454e-05, |
|
"loss": 0.3956, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 19.16, |
|
"learning_rate": 3.972520908004779e-05, |
|
"loss": 0.362, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 19.24, |
|
"learning_rate": 3.9426523297491045e-05, |
|
"loss": 0.3233, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 19.32, |
|
"learning_rate": 3.912783751493429e-05, |
|
"loss": 0.3481, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 19.4, |
|
"learning_rate": 3.882915173237754e-05, |
|
"loss": 0.3574, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 19.48, |
|
"learning_rate": 3.8530465949820786e-05, |
|
"loss": 0.4001, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 19.56, |
|
"learning_rate": 3.8231780167264044e-05, |
|
"loss": 0.348, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 19.64, |
|
"learning_rate": 3.793309438470729e-05, |
|
"loss": 0.3783, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 19.72, |
|
"learning_rate": 3.763440860215054e-05, |
|
"loss": 0.3542, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 19.8, |
|
"learning_rate": 3.7335722819593785e-05, |
|
"loss": 0.3482, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 19.88, |
|
"learning_rate": 3.7037037037037037e-05, |
|
"loss": 0.3996, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 19.96, |
|
"learning_rate": 3.673835125448029e-05, |
|
"loss": 0.3783, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_accuracy": 0.8179255918827508, |
|
"eval_f1": 0.8150463159197603, |
|
"eval_loss": 0.5658373236656189, |
|
"eval_precision": 0.8168528878369955, |
|
"eval_recall": 0.8179255918827508, |
|
"eval_runtime": 11.4436, |
|
"eval_samples_per_second": 155.021, |
|
"eval_steps_per_second": 4.894, |
|
"step": 2495 |
|
}, |
|
{ |
|
"epoch": 20.04, |
|
"learning_rate": 3.643966547192354e-05, |
|
"loss": 0.3675, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 20.12, |
|
"learning_rate": 3.614097968936679e-05, |
|
"loss": 0.3249, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 20.2, |
|
"learning_rate": 3.5842293906810036e-05, |
|
"loss": 0.357, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 20.28, |
|
"learning_rate": 3.554360812425329e-05, |
|
"loss": 0.4023, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 20.36, |
|
"learning_rate": 3.524492234169653e-05, |
|
"loss": 0.3384, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 20.44, |
|
"learning_rate": 3.494623655913979e-05, |
|
"loss": 0.3366, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 20.52, |
|
"learning_rate": 3.4647550776583035e-05, |
|
"loss": 0.3467, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 20.6, |
|
"learning_rate": 3.4348864994026287e-05, |
|
"loss": 0.338, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 20.68, |
|
"learning_rate": 3.405017921146954e-05, |
|
"loss": 0.3224, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 20.76, |
|
"learning_rate": 3.375149342891278e-05, |
|
"loss": 0.3647, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 20.84, |
|
"learning_rate": 3.3452807646356034e-05, |
|
"loss": 0.395, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 20.92, |
|
"learning_rate": 3.3154121863799286e-05, |
|
"loss": 0.3293, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 20.99, |
|
"eval_accuracy": 0.8246899661781285, |
|
"eval_f1": 0.8229604972103561, |
|
"eval_loss": 0.564728319644928, |
|
"eval_precision": 0.8234334027561456, |
|
"eval_recall": 0.8246899661781285, |
|
"eval_runtime": 11.3492, |
|
"eval_samples_per_second": 156.31, |
|
"eval_steps_per_second": 4.934, |
|
"step": 2619 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"learning_rate": 3.285543608124254e-05, |
|
"loss": 0.3223, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 21.08, |
|
"learning_rate": 3.255675029868578e-05, |
|
"loss": 0.2962, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 21.16, |
|
"learning_rate": 3.2258064516129034e-05, |
|
"loss": 0.3448, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 21.24, |
|
"learning_rate": 3.195937873357228e-05, |
|
"loss": 0.3352, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 21.32, |
|
"learning_rate": 3.1660692951015537e-05, |
|
"loss": 0.3237, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 21.4, |
|
"learning_rate": 3.136200716845878e-05, |
|
"loss": 0.3541, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 21.48, |
|
"learning_rate": 3.106332138590203e-05, |
|
"loss": 0.3474, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 21.56, |
|
"learning_rate": 3.0764635603345284e-05, |
|
"loss": 0.3285, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 21.64, |
|
"learning_rate": 3.046594982078853e-05, |
|
"loss": 0.3186, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 21.72, |
|
"learning_rate": 3.016726403823178e-05, |
|
"loss": 0.3284, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 21.8, |
|
"learning_rate": 2.9868578255675032e-05, |
|
"loss": 0.3627, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 21.88, |
|
"learning_rate": 2.9569892473118284e-05, |
|
"loss": 0.3636, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 21.96, |
|
"learning_rate": 2.9271206690561532e-05, |
|
"loss": 0.3214, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_accuracy": 0.830890642615558, |
|
"eval_f1": 0.8293042643812296, |
|
"eval_loss": 0.5654408931732178, |
|
"eval_precision": 0.8288839649071558, |
|
"eval_recall": 0.830890642615558, |
|
"eval_runtime": 11.3586, |
|
"eval_samples_per_second": 156.181, |
|
"eval_steps_per_second": 4.93, |
|
"step": 2744 |
|
}, |
|
{ |
|
"epoch": 22.04, |
|
"learning_rate": 2.897252090800478e-05, |
|
"loss": 0.3317, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 22.12, |
|
"learning_rate": 2.8673835125448028e-05, |
|
"loss": 0.3225, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 22.2, |
|
"learning_rate": 2.8375149342891276e-05, |
|
"loss": 0.2913, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 22.28, |
|
"learning_rate": 2.807646356033453e-05, |
|
"loss": 0.3137, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 22.36, |
|
"learning_rate": 2.777777777777778e-05, |
|
"loss": 0.3353, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 22.44, |
|
"learning_rate": 2.747909199522103e-05, |
|
"loss": 0.3328, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 22.53, |
|
"learning_rate": 2.718040621266428e-05, |
|
"loss": 0.2779, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 22.61, |
|
"learning_rate": 2.6881720430107527e-05, |
|
"loss": 0.2953, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 22.69, |
|
"learning_rate": 2.6583034647550775e-05, |
|
"loss": 0.3272, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 22.77, |
|
"learning_rate": 2.628434886499403e-05, |
|
"loss": 0.3443, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 22.85, |
|
"learning_rate": 2.5985663082437278e-05, |
|
"loss": 0.3481, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 22.93, |
|
"learning_rate": 2.5686977299880526e-05, |
|
"loss": 0.3285, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_accuracy": 0.8213077790304397, |
|
"eval_f1": 0.8201485238469571, |
|
"eval_loss": 0.5942615270614624, |
|
"eval_precision": 0.8226255291354058, |
|
"eval_recall": 0.8213077790304397, |
|
"eval_runtime": 11.322, |
|
"eval_samples_per_second": 156.685, |
|
"eval_steps_per_second": 4.946, |
|
"step": 2869 |
|
}, |
|
{ |
|
"epoch": 23.01, |
|
"learning_rate": 2.5388291517323774e-05, |
|
"loss": 0.3354, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 23.09, |
|
"learning_rate": 2.5089605734767026e-05, |
|
"loss": 0.3264, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 23.17, |
|
"learning_rate": 2.4790919952210277e-05, |
|
"loss": 0.2951, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 23.25, |
|
"learning_rate": 2.4492234169653525e-05, |
|
"loss": 0.2826, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 23.33, |
|
"learning_rate": 2.4193548387096777e-05, |
|
"loss": 0.313, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 23.41, |
|
"learning_rate": 2.3894862604540025e-05, |
|
"loss": 0.2853, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 23.49, |
|
"learning_rate": 2.3596176821983273e-05, |
|
"loss": 0.2893, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 23.57, |
|
"learning_rate": 2.3297491039426525e-05, |
|
"loss": 0.3209, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 23.65, |
|
"learning_rate": 2.2998805256869773e-05, |
|
"loss": 0.3135, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 23.73, |
|
"learning_rate": 2.270011947431302e-05, |
|
"loss": 0.2946, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 23.81, |
|
"learning_rate": 2.2401433691756272e-05, |
|
"loss": 0.2743, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 23.89, |
|
"learning_rate": 2.2102747909199524e-05, |
|
"loss": 0.3148, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 23.97, |
|
"learning_rate": 2.1804062126642775e-05, |
|
"loss": 0.2934, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_accuracy": 0.826381059751973, |
|
"eval_f1": 0.8259189314518771, |
|
"eval_loss": 0.5931342244148254, |
|
"eval_precision": 0.8286601748663491, |
|
"eval_recall": 0.826381059751973, |
|
"eval_runtime": 11.2876, |
|
"eval_samples_per_second": 157.164, |
|
"eval_steps_per_second": 4.961, |
|
"step": 2994 |
|
}, |
|
{ |
|
"epoch": 24.05, |
|
"learning_rate": 2.1505376344086024e-05, |
|
"loss": 0.3438, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 24.13, |
|
"learning_rate": 2.120669056152927e-05, |
|
"loss": 0.3135, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 24.21, |
|
"learning_rate": 2.0908004778972523e-05, |
|
"loss": 0.2954, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 24.29, |
|
"learning_rate": 2.060931899641577e-05, |
|
"loss": 0.3392, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 24.37, |
|
"learning_rate": 2.031063321385902e-05, |
|
"loss": 0.2939, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 24.45, |
|
"learning_rate": 2.001194743130227e-05, |
|
"loss": 0.312, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 24.53, |
|
"learning_rate": 1.9713261648745522e-05, |
|
"loss": 0.2761, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 24.61, |
|
"learning_rate": 1.941457586618877e-05, |
|
"loss": 0.2775, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 24.69, |
|
"learning_rate": 1.9115890083632022e-05, |
|
"loss": 0.3379, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 24.77, |
|
"learning_rate": 1.881720430107527e-05, |
|
"loss": 0.2699, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 24.85, |
|
"learning_rate": 1.8518518518518518e-05, |
|
"loss": 0.322, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 24.93, |
|
"learning_rate": 1.821983273596177e-05, |
|
"loss": 0.3051, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 24.99, |
|
"eval_accuracy": 0.830890642615558, |
|
"eval_f1": 0.8302639937505762, |
|
"eval_loss": 0.5788145661354065, |
|
"eval_precision": 0.8324873371475154, |
|
"eval_recall": 0.830890642615558, |
|
"eval_runtime": 11.2955, |
|
"eval_samples_per_second": 157.054, |
|
"eval_steps_per_second": 4.958, |
|
"step": 3118 |
|
}, |
|
{ |
|
"epoch": 25.01, |
|
"learning_rate": 1.7921146953405018e-05, |
|
"loss": 0.2746, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 25.09, |
|
"learning_rate": 1.7622461170848266e-05, |
|
"loss": 0.2837, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 25.17, |
|
"learning_rate": 1.7323775388291518e-05, |
|
"loss": 0.3028, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 25.25, |
|
"learning_rate": 1.702508960573477e-05, |
|
"loss": 0.2583, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 25.33, |
|
"learning_rate": 1.6726403823178017e-05, |
|
"loss": 0.2869, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 25.41, |
|
"learning_rate": 1.642771804062127e-05, |
|
"loss": 0.2706, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 25.49, |
|
"learning_rate": 1.6129032258064517e-05, |
|
"loss": 0.3174, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 25.57, |
|
"learning_rate": 1.5830346475507768e-05, |
|
"loss": 0.2755, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 25.65, |
|
"learning_rate": 1.5531660692951016e-05, |
|
"loss": 0.2819, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 25.73, |
|
"learning_rate": 1.5232974910394265e-05, |
|
"loss": 0.2791, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 25.81, |
|
"learning_rate": 1.4934289127837516e-05, |
|
"loss": 0.2345, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 25.89, |
|
"learning_rate": 1.4635603345280766e-05, |
|
"loss": 0.281, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 25.97, |
|
"learning_rate": 1.4336917562724014e-05, |
|
"loss": 0.2911, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_accuracy": 0.8376550169109357, |
|
"eval_f1": 0.8358147014612041, |
|
"eval_loss": 0.5699869394302368, |
|
"eval_precision": 0.8353641528998437, |
|
"eval_recall": 0.8376550169109357, |
|
"eval_runtime": 11.3089, |
|
"eval_samples_per_second": 156.868, |
|
"eval_steps_per_second": 4.952, |
|
"step": 3243 |
|
}, |
|
{ |
|
"epoch": 26.05, |
|
"learning_rate": 1.4038231780167265e-05, |
|
"loss": 0.2839, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 26.13, |
|
"learning_rate": 1.3739545997610515e-05, |
|
"loss": 0.3274, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 26.21, |
|
"learning_rate": 1.3440860215053763e-05, |
|
"loss": 0.277, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 26.29, |
|
"learning_rate": 1.3142174432497015e-05, |
|
"loss": 0.2593, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 26.37, |
|
"learning_rate": 1.2843488649940263e-05, |
|
"loss": 0.2967, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 26.45, |
|
"learning_rate": 1.2544802867383513e-05, |
|
"loss": 0.2681, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 26.53, |
|
"learning_rate": 1.2246117084826763e-05, |
|
"loss": 0.2953, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 26.61, |
|
"learning_rate": 1.1947431302270013e-05, |
|
"loss": 0.2709, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 26.69, |
|
"learning_rate": 1.1648745519713262e-05, |
|
"loss": 0.2654, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 26.77, |
|
"learning_rate": 1.135005973715651e-05, |
|
"loss": 0.2703, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 26.85, |
|
"learning_rate": 1.1051373954599762e-05, |
|
"loss": 0.276, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 26.93, |
|
"learning_rate": 1.0752688172043012e-05, |
|
"loss": 0.2893, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_accuracy": 0.8286358511837655, |
|
"eval_f1": 0.8290920717416481, |
|
"eval_loss": 0.5970751643180847, |
|
"eval_precision": 0.8319741830782457, |
|
"eval_recall": 0.8286358511837655, |
|
"eval_runtime": 11.344, |
|
"eval_samples_per_second": 156.383, |
|
"eval_steps_per_second": 4.937, |
|
"step": 3368 |
|
}, |
|
{ |
|
"epoch": 27.01, |
|
"learning_rate": 1.0454002389486262e-05, |
|
"loss": 0.276, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 27.09, |
|
"learning_rate": 1.015531660692951e-05, |
|
"loss": 0.2726, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 27.17, |
|
"learning_rate": 9.856630824372761e-06, |
|
"loss": 0.2567, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 27.25, |
|
"learning_rate": 9.557945041816011e-06, |
|
"loss": 0.3261, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 27.33, |
|
"learning_rate": 9.259259259259259e-06, |
|
"loss": 0.2735, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 27.41, |
|
"learning_rate": 8.960573476702509e-06, |
|
"loss": 0.2772, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 27.49, |
|
"learning_rate": 8.661887694145759e-06, |
|
"loss": 0.2716, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 27.58, |
|
"learning_rate": 8.363201911589009e-06, |
|
"loss": 0.2577, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 27.66, |
|
"learning_rate": 8.064516129032258e-06, |
|
"loss": 0.2456, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 27.74, |
|
"learning_rate": 7.765830346475508e-06, |
|
"loss": 0.2653, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 27.82, |
|
"learning_rate": 7.467144563918758e-06, |
|
"loss": 0.2599, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 27.9, |
|
"learning_rate": 7.168458781362007e-06, |
|
"loss": 0.2674, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 27.98, |
|
"learning_rate": 6.869772998805258e-06, |
|
"loss": 0.2794, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_accuracy": 0.8314543404735062, |
|
"eval_f1": 0.8303304998912855, |
|
"eval_loss": 0.59084552526474, |
|
"eval_precision": 0.8307246017121835, |
|
"eval_recall": 0.8314543404735062, |
|
"eval_runtime": 11.3258, |
|
"eval_samples_per_second": 156.633, |
|
"eval_steps_per_second": 4.944, |
|
"step": 3493 |
|
}, |
|
{ |
|
"epoch": 28.06, |
|
"learning_rate": 6.5710872162485075e-06, |
|
"loss": 0.3034, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 28.14, |
|
"learning_rate": 6.2724014336917564e-06, |
|
"loss": 0.2656, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 28.22, |
|
"learning_rate": 5.973715651135006e-06, |
|
"loss": 0.2968, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 28.3, |
|
"learning_rate": 5.675029868578255e-06, |
|
"loss": 0.2493, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 28.38, |
|
"learning_rate": 5.376344086021506e-06, |
|
"loss": 0.278, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 28.46, |
|
"learning_rate": 5.077658303464755e-06, |
|
"loss": 0.2541, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 28.54, |
|
"learning_rate": 4.7789725209080055e-06, |
|
"loss": 0.2532, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 28.62, |
|
"learning_rate": 4.4802867383512545e-06, |
|
"loss": 0.286, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 28.7, |
|
"learning_rate": 4.181600955794504e-06, |
|
"loss": 0.259, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 28.78, |
|
"learning_rate": 3.882915173237754e-06, |
|
"loss": 0.2654, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 28.86, |
|
"learning_rate": 3.5842293906810035e-06, |
|
"loss": 0.2729, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 28.94, |
|
"learning_rate": 3.2855436081242537e-06, |
|
"loss": 0.2506, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 28.99, |
|
"eval_accuracy": 0.830890642615558, |
|
"eval_f1": 0.8306383082164391, |
|
"eval_loss": 0.591434121131897, |
|
"eval_precision": 0.8314261484426553, |
|
"eval_recall": 0.830890642615558, |
|
"eval_runtime": 11.3065, |
|
"eval_samples_per_second": 156.901, |
|
"eval_steps_per_second": 4.953, |
|
"step": 3617 |
|
}, |
|
{ |
|
"epoch": 29.02, |
|
"learning_rate": 2.986857825567503e-06, |
|
"loss": 0.2733, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 29.1, |
|
"learning_rate": 2.688172043010753e-06, |
|
"loss": 0.2596, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 29.18, |
|
"learning_rate": 2.3894862604540028e-06, |
|
"loss": 0.2374, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 29.26, |
|
"learning_rate": 2.090800477897252e-06, |
|
"loss": 0.2406, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 29.34, |
|
"learning_rate": 1.7921146953405017e-06, |
|
"loss": 0.241, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 29.42, |
|
"learning_rate": 1.4934289127837516e-06, |
|
"loss": 0.2652, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 29.5, |
|
"learning_rate": 1.1947431302270014e-06, |
|
"loss": 0.2586, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 29.58, |
|
"learning_rate": 8.960573476702509e-07, |
|
"loss": 0.2763, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 29.66, |
|
"learning_rate": 5.973715651135007e-07, |
|
"loss": 0.2471, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 29.74, |
|
"learning_rate": 2.9868578255675034e-07, |
|
"loss": 0.2533, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 29.82, |
|
"learning_rate": 0.0, |
|
"loss": 0.2421, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 29.82, |
|
"eval_accuracy": 0.8365276211950394, |
|
"eval_f1": 0.8359075671561115, |
|
"eval_loss": 0.5861225724220276, |
|
"eval_precision": 0.8365853817017682, |
|
"eval_recall": 0.8365276211950394, |
|
"eval_runtime": 11.3024, |
|
"eval_samples_per_second": 156.957, |
|
"eval_steps_per_second": 4.955, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 29.82, |
|
"step": 3720, |
|
"total_flos": 3.759990116038978e+19, |
|
"train_loss": 0.5362014834598828, |
|
"train_runtime": 4396.4695, |
|
"train_samples_per_second": 108.919, |
|
"train_steps_per_second": 0.846 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 3720, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 30, |
|
"save_steps": 500, |
|
"total_flos": 3.759990116038978e+19, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|