|
{ |
|
"best_metric": 0.998719590268886, |
|
"best_model_checkpoint": "PlantDiseaseDetectorV2/checkpoint-1533", |
|
"epoch": 6.976109215017065, |
|
"eval_steps": 500, |
|
"global_step": 1533, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 3.2467532467532465e-06, |
|
"loss": 3.6354, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 6.493506493506493e-06, |
|
"loss": 3.6146, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.740259740259742e-06, |
|
"loss": 3.5784, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.2987012987012986e-05, |
|
"loss": 3.5183, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.6233766233766234e-05, |
|
"loss": 3.435, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.9480519480519483e-05, |
|
"loss": 3.3236, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 2.272727272727273e-05, |
|
"loss": 3.1774, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 2.5974025974025972e-05, |
|
"loss": 3.0042, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 2.922077922077922e-05, |
|
"loss": 2.8095, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 3.246753246753247e-05, |
|
"loss": 2.6034, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 3.571428571428572e-05, |
|
"loss": 2.3817, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 3.8961038961038966e-05, |
|
"loss": 2.1751, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.220779220779221e-05, |
|
"loss": 1.9796, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 4.545454545454546e-05, |
|
"loss": 1.7805, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.87012987012987e-05, |
|
"loss": 1.6209, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.978245105148659e-05, |
|
"loss": 1.4515, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.941986947063089e-05, |
|
"loss": 1.3197, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 4.90572878897752e-05, |
|
"loss": 1.1848, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.8694706308919505e-05, |
|
"loss": 1.0976, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.833212472806382e-05, |
|
"loss": 0.9857, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.7969543147208126e-05, |
|
"loss": 0.9051, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.986057760705648, |
|
"eval_loss": 0.8024668097496033, |
|
"eval_runtime": 70.6659, |
|
"eval_samples_per_second": 99.468, |
|
"eval_steps_per_second": 1.557, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 4.760696156635243e-05, |
|
"loss": 0.8431, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 4.724437998549674e-05, |
|
"loss": 0.7784, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 4.6881798404641046e-05, |
|
"loss": 0.7198, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 4.651921682378535e-05, |
|
"loss": 0.6712, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 4.6156635242929666e-05, |
|
"loss": 0.6365, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 4.579405366207397e-05, |
|
"loss": 0.5892, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 4.543147208121827e-05, |
|
"loss": 0.5563, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 4.506889050036258e-05, |
|
"loss": 0.5254, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 4.4706308919506886e-05, |
|
"loss": 0.4982, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 4.43437273386512e-05, |
|
"loss": 0.4667, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 4.398114575779551e-05, |
|
"loss": 0.4481, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 4.3618564176939813e-05, |
|
"loss": 0.4199, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.325598259608412e-05, |
|
"loss": 0.4117, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.289340101522843e-05, |
|
"loss": 0.3867, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 4.253081943437274e-05, |
|
"loss": 0.3736, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.216823785351705e-05, |
|
"loss": 0.3668, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 4.1805656272661354e-05, |
|
"loss": 0.3369, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 4.1443074691805654e-05, |
|
"loss": 0.3416, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.108049311094996e-05, |
|
"loss": 0.3287, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 4.0717911530094274e-05, |
|
"loss": 0.2989, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 4.035532994923858e-05, |
|
"loss": 0.2934, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 3.999274836838289e-05, |
|
"loss": 0.2801, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.9958742353108551, |
|
"eval_loss": 0.26061177253723145, |
|
"eval_runtime": 71.3214, |
|
"eval_samples_per_second": 98.554, |
|
"eval_steps_per_second": 1.542, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 3.9630166787527195e-05, |
|
"loss": 0.2742, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 3.92675852066715e-05, |
|
"loss": 0.2615, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 3.890500362581581e-05, |
|
"loss": 0.2512, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 3.854242204496012e-05, |
|
"loss": 0.2442, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 3.817984046410443e-05, |
|
"loss": 0.2369, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 3.7817258883248735e-05, |
|
"loss": 0.2288, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 3.7454677302393035e-05, |
|
"loss": 0.2226, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 3.709209572153735e-05, |
|
"loss": 0.2162, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 3.6729514140681655e-05, |
|
"loss": 0.2055, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 3.636693255982596e-05, |
|
"loss": 0.199, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 3.600435097897027e-05, |
|
"loss": 0.2128, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.5641769398114576e-05, |
|
"loss": 0.1885, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.527918781725888e-05, |
|
"loss": 0.1854, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 3.4916606236403196e-05, |
|
"loss": 0.1845, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 3.45540246555475e-05, |
|
"loss": 0.1814, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 3.419144307469181e-05, |
|
"loss": 0.1816, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 3.3828861493836116e-05, |
|
"loss": 0.1689, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 3.3466279912980416e-05, |
|
"loss": 0.1658, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 3.310369833212473e-05, |
|
"loss": 0.1612, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 3.2741116751269036e-05, |
|
"loss": 0.1572, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 3.237853517041334e-05, |
|
"loss": 0.1552, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 3.201595358955765e-05, |
|
"loss": 0.1455, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.9972969127898705, |
|
"eval_loss": 0.14022748172283173, |
|
"eval_runtime": 71.3997, |
|
"eval_samples_per_second": 98.446, |
|
"eval_steps_per_second": 1.541, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 3.165337200870196e-05, |
|
"loss": 0.1446, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 3.129079042784627e-05, |
|
"loss": 0.1422, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 3.092820884699058e-05, |
|
"loss": 0.1411, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 3.0565627266134884e-05, |
|
"loss": 0.1334, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 3.020304568527919e-05, |
|
"loss": 0.1333, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 2.98404641044235e-05, |
|
"loss": 0.1325, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 2.9477882523567807e-05, |
|
"loss": 0.1278, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 2.911530094271211e-05, |
|
"loss": 0.127, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 2.8752719361856417e-05, |
|
"loss": 0.1272, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 2.8390137781000724e-05, |
|
"loss": 0.1289, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 2.8027556200145034e-05, |
|
"loss": 0.1235, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 2.766497461928934e-05, |
|
"loss": 0.1194, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 2.7302393038433648e-05, |
|
"loss": 0.1179, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 2.6939811457577958e-05, |
|
"loss": 0.11, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 2.6577229876722265e-05, |
|
"loss": 0.1126, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 2.621464829586657e-05, |
|
"loss": 0.1106, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 2.585206671501088e-05, |
|
"loss": 0.1073, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 2.548948513415519e-05, |
|
"loss": 0.1024, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 2.5126903553299492e-05, |
|
"loss": 0.1001, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 2.4764321972443802e-05, |
|
"loss": 0.1013, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 2.440174039158811e-05, |
|
"loss": 0.1003, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 2.4039158810732415e-05, |
|
"loss": 0.0949, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.9985773225209845, |
|
"eval_loss": 0.09417087584733963, |
|
"eval_runtime": 71.2514, |
|
"eval_samples_per_second": 98.651, |
|
"eval_steps_per_second": 1.544, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 2.3676577229876722e-05, |
|
"loss": 0.0965, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 2.3313995649021032e-05, |
|
"loss": 0.0936, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 2.295141406816534e-05, |
|
"loss": 0.0938, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 2.2588832487309646e-05, |
|
"loss": 0.0907, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 2.2226250906453953e-05, |
|
"loss": 0.0889, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 2.186366932559826e-05, |
|
"loss": 0.0889, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 2.150108774474257e-05, |
|
"loss": 0.0905, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 2.1138506163886876e-05, |
|
"loss": 0.0888, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 2.0775924583031183e-05, |
|
"loss": 0.0849, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 2.0413343002175493e-05, |
|
"loss": 0.0852, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 2.0050761421319797e-05, |
|
"loss": 0.0843, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 1.9688179840464103e-05, |
|
"loss": 0.0851, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 1.9325598259608413e-05, |
|
"loss": 0.0884, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 1.896301667875272e-05, |
|
"loss": 0.0899, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 1.860043509789703e-05, |
|
"loss": 0.0848, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"learning_rate": 1.8237853517041334e-05, |
|
"loss": 0.0794, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 1.787527193618564e-05, |
|
"loss": 0.0802, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 1.751269035532995e-05, |
|
"loss": 0.0808, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 1.7150108774474257e-05, |
|
"loss": 0.0763, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 1.6787527193618564e-05, |
|
"loss": 0.0801, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 1.6424945612762874e-05, |
|
"loss": 0.0745, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 1.6062364031907178e-05, |
|
"loss": 0.0741, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_accuracy": 0.9984350547730829, |
|
"eval_loss": 0.07485666126012802, |
|
"eval_runtime": 71.7237, |
|
"eval_samples_per_second": 98.001, |
|
"eval_steps_per_second": 1.534, |
|
"step": 1098 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 1.5699782451051488e-05, |
|
"loss": 0.0738, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 1.5337200870195794e-05, |
|
"loss": 0.0729, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 1.4974619289340103e-05, |
|
"loss": 0.0748, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 1.4612037708484411e-05, |
|
"loss": 0.0698, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 5.19, |
|
"learning_rate": 1.4249456127628716e-05, |
|
"loss": 0.0721, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 5.23, |
|
"learning_rate": 1.3886874546773023e-05, |
|
"loss": 0.0748, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 1.3524292965917332e-05, |
|
"loss": 0.0695, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 1.316171138506164e-05, |
|
"loss": 0.0688, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 1.2799129804205947e-05, |
|
"loss": 0.0723, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 1.2436548223350254e-05, |
|
"loss": 0.0683, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 5.46, |
|
"learning_rate": 1.2073966642494562e-05, |
|
"loss": 0.0709, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"learning_rate": 1.1711385061638869e-05, |
|
"loss": 0.0674, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 1.1348803480783177e-05, |
|
"loss": 0.0685, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 1.0986221899927484e-05, |
|
"loss": 0.0679, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 5.64, |
|
"learning_rate": 1.062364031907179e-05, |
|
"loss": 0.0672, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 1.02610587382161e-05, |
|
"loss": 0.0648, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 9.898477157360408e-06, |
|
"loss": 0.0659, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 5.78, |
|
"learning_rate": 9.535895576504714e-06, |
|
"loss": 0.0629, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 9.173313995649021e-06, |
|
"loss": 0.0627, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 5.87, |
|
"learning_rate": 8.81073241479333e-06, |
|
"loss": 0.0643, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 5.92, |
|
"learning_rate": 8.448150833937636e-06, |
|
"loss": 0.065, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 8.085569253081945e-06, |
|
"loss": 0.0623, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_accuracy": 0.9984350547730829, |
|
"eval_loss": 0.06421820819377899, |
|
"eval_runtime": 71.8803, |
|
"eval_samples_per_second": 97.788, |
|
"eval_steps_per_second": 1.53, |
|
"step": 1318 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.72298767222625e-06, |
|
"loss": 0.0634, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 6.05, |
|
"learning_rate": 7.360406091370558e-06, |
|
"loss": 0.0625, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 6.1, |
|
"learning_rate": 6.997824510514867e-06, |
|
"loss": 0.0612, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 6.14, |
|
"learning_rate": 6.635242929659173e-06, |
|
"loss": 0.0607, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"learning_rate": 6.272661348803481e-06, |
|
"loss": 0.0595, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 6.23, |
|
"learning_rate": 5.910079767947789e-06, |
|
"loss": 0.0602, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 6.28, |
|
"learning_rate": 5.547498187092096e-06, |
|
"loss": 0.0606, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"learning_rate": 5.184916606236404e-06, |
|
"loss": 0.0608, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 6.37, |
|
"learning_rate": 4.822335025380711e-06, |
|
"loss": 0.0621, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 6.42, |
|
"learning_rate": 4.459753444525018e-06, |
|
"loss": 0.0608, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 4.097171863669326e-06, |
|
"loss": 0.0615, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 6.51, |
|
"learning_rate": 3.734590282813633e-06, |
|
"loss": 0.0604, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 6.55, |
|
"learning_rate": 3.3720087019579407e-06, |
|
"loss": 0.0612, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 3.0094271211022483e-06, |
|
"loss": 0.058, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 6.64, |
|
"learning_rate": 2.646845540246556e-06, |
|
"loss": 0.0598, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 6.69, |
|
"learning_rate": 2.284263959390863e-06, |
|
"loss": 0.0592, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 1.9216823785351706e-06, |
|
"loss": 0.058, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"learning_rate": 1.559100797679478e-06, |
|
"loss": 0.0573, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 6.83, |
|
"learning_rate": 1.1965192168237854e-06, |
|
"loss": 0.058, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"learning_rate": 8.339376359680928e-07, |
|
"loss": 0.0603, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 6.92, |
|
"learning_rate": 4.713560551124003e-07, |
|
"loss": 0.0576, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 6.96, |
|
"learning_rate": 1.0877447425670776e-07, |
|
"loss": 0.0586, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"eval_accuracy": 0.998719590268886, |
|
"eval_loss": 0.06100322678685188, |
|
"eval_runtime": 70.9668, |
|
"eval_samples_per_second": 99.046, |
|
"eval_steps_per_second": 1.55, |
|
"step": 1533 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"step": 1533, |
|
"total_flos": 3.041208679733851e+19, |
|
"train_loss": 0.46931519971240676, |
|
"train_runtime": 11436.94, |
|
"train_samples_per_second": 34.419, |
|
"train_steps_per_second": 0.134 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1533, |
|
"num_train_epochs": 7, |
|
"save_steps": 500, |
|
"total_flos": 3.041208679733851e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|