|
{ |
|
"best_metric": 0.8028383549429728, |
|
"best_model_checkpoint": "convnextv2-tiny-1k-224-finetuned-combinedSpiders/checkpoint-13456", |
|
"epoch": 3.9991084032989077, |
|
"eval_steps": 500, |
|
"global_step": 13456, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3.7147102526002973e-07, |
|
"loss": 5.4605, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.429420505200595e-07, |
|
"loss": 5.4325, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.1144130757800891e-06, |
|
"loss": 5.4352, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.485884101040119e-06, |
|
"loss": 5.4319, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.8573551263001487e-06, |
|
"loss": 5.4179, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 2.2288261515601783e-06, |
|
"loss": 5.4016, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 2.600297176820208e-06, |
|
"loss": 5.3692, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 2.971768202080238e-06, |
|
"loss": 5.35, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 3.3432392273402677e-06, |
|
"loss": 5.315, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 3.7147102526002974e-06, |
|
"loss": 5.291, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.086181277860327e-06, |
|
"loss": 5.2526, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.457652303120357e-06, |
|
"loss": 5.2126, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.829123328380386e-06, |
|
"loss": 5.166, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 5.200594353640416e-06, |
|
"loss": 5.1061, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 5.572065378900446e-06, |
|
"loss": 5.0615, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 5.943536404160476e-06, |
|
"loss": 4.9806, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 6.3150074294205055e-06, |
|
"loss": 4.8978, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 6.686478454680535e-06, |
|
"loss": 4.7853, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 7.057949479940565e-06, |
|
"loss": 4.6225, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 7.429420505200595e-06, |
|
"loss": 4.5904, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 7.800891530460625e-06, |
|
"loss": 4.396, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.172362555720654e-06, |
|
"loss": 4.2786, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.543833580980684e-06, |
|
"loss": 4.2328, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 8.915304606240713e-06, |
|
"loss": 4.23, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.286775631500744e-06, |
|
"loss": 4.0821, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.658246656760773e-06, |
|
"loss": 4.0607, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.0029717682020803e-05, |
|
"loss": 3.9432, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.0401188707280832e-05, |
|
"loss": 3.9038, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.0772659732540863e-05, |
|
"loss": 3.9666, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.1144130757800892e-05, |
|
"loss": 3.9316, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.1515601783060921e-05, |
|
"loss": 3.7636, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.1887072808320951e-05, |
|
"loss": 3.7484, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.225854383358098e-05, |
|
"loss": 3.6972, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.2630014858841011e-05, |
|
"loss": 3.5604, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.3001485884101042e-05, |
|
"loss": 3.7302, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.337295690936107e-05, |
|
"loss": 3.4564, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.3744427934621101e-05, |
|
"loss": 3.675, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.411589895988113e-05, |
|
"loss": 3.5408, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.4487369985141161e-05, |
|
"loss": 3.5079, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.485884101040119e-05, |
|
"loss": 3.6402, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.5230312035661217e-05, |
|
"loss": 3.4247, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.560178306092125e-05, |
|
"loss": 3.4156, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.5973254086181278e-05, |
|
"loss": 3.4463, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.6344725111441307e-05, |
|
"loss": 3.4605, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.6716196136701336e-05, |
|
"loss": 3.3595, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.708766716196137e-05, |
|
"loss": 3.3557, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.7459138187221397e-05, |
|
"loss": 3.3645, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.7830609212481426e-05, |
|
"loss": 3.3692, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.8202080237741455e-05, |
|
"loss": 3.3335, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.8573551263001488e-05, |
|
"loss": 3.2619, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.8945022288261517e-05, |
|
"loss": 3.175, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9316493313521546e-05, |
|
"loss": 3.2085, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9687964338781574e-05, |
|
"loss": 3.0792, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 2.0059435364041607e-05, |
|
"loss": 2.988, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 2.0430906389301636e-05, |
|
"loss": 3.1117, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 2.0802377414561665e-05, |
|
"loss": 3.1018, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 2.1173848439821694e-05, |
|
"loss": 3.0903, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 2.1545319465081726e-05, |
|
"loss": 3.0333, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 2.1916790490341755e-05, |
|
"loss": 3.0295, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 2.2288261515601784e-05, |
|
"loss": 3.0337, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 2.2659732540861813e-05, |
|
"loss": 3.0827, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 2.3031203566121842e-05, |
|
"loss": 3.0105, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 2.3402674591381874e-05, |
|
"loss": 2.9386, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 2.3774145616641903e-05, |
|
"loss": 2.9191, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 2.4145616641901932e-05, |
|
"loss": 3.0147, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 2.451708766716196e-05, |
|
"loss": 2.9571, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 2.4888558692421993e-05, |
|
"loss": 2.9609, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 2.5260029717682022e-05, |
|
"loss": 2.9023, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 2.563150074294205e-05, |
|
"loss": 2.9555, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 2.6002971768202083e-05, |
|
"loss": 2.9386, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 2.6374442793462112e-05, |
|
"loss": 2.8448, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 2.674591381872214e-05, |
|
"loss": 2.9019, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 2.711738484398217e-05, |
|
"loss": 2.7546, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 2.7488855869242203e-05, |
|
"loss": 2.9225, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 2.786032689450223e-05, |
|
"loss": 2.7843, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 2.823179791976226e-05, |
|
"loss": 2.7189, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 2.860326894502229e-05, |
|
"loss": 2.8452, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 2.8974739970282322e-05, |
|
"loss": 2.8528, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 2.934621099554235e-05, |
|
"loss": 2.9453, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 2.971768202080238e-05, |
|
"loss": 2.7887, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 3.008915304606241e-05, |
|
"loss": 2.6366, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 3.0460624071322434e-05, |
|
"loss": 2.6173, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 3.083209509658247e-05, |
|
"loss": 2.6045, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 3.12035661218425e-05, |
|
"loss": 2.6325, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 3.157503714710253e-05, |
|
"loss": 2.6887, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 3.1946508172362557e-05, |
|
"loss": 2.5826, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 3.2317979197622586e-05, |
|
"loss": 2.7811, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 3.2689450222882614e-05, |
|
"loss": 2.7793, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 3.3060921248142643e-05, |
|
"loss": 2.5181, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 3.343239227340267e-05, |
|
"loss": 2.5493, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 3.380386329866271e-05, |
|
"loss": 2.4256, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 3.417533432392274e-05, |
|
"loss": 2.6975, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 3.4546805349182766e-05, |
|
"loss": 2.6269, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 3.4918276374442795e-05, |
|
"loss": 2.6296, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 3.5289747399702824e-05, |
|
"loss": 2.5659, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 3.566121842496285e-05, |
|
"loss": 2.658, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 3.603268945022288e-05, |
|
"loss": 2.4835, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 3.640416047548291e-05, |
|
"loss": 2.4395, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 3.6775631500742946e-05, |
|
"loss": 2.5997, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 3.7147102526002975e-05, |
|
"loss": 2.5819, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 3.7518573551263004e-05, |
|
"loss": 2.4744, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 3.789004457652303e-05, |
|
"loss": 2.4781, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 3.826151560178306e-05, |
|
"loss": 2.5338, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 3.863298662704309e-05, |
|
"loss": 2.4157, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 3.900445765230312e-05, |
|
"loss": 2.4724, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 3.937592867756315e-05, |
|
"loss": 2.4876, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 3.974739970282318e-05, |
|
"loss": 2.3551, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.0118870728083214e-05, |
|
"loss": 2.5717, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.049034175334324e-05, |
|
"loss": 2.5295, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.086181277860327e-05, |
|
"loss": 2.4617, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.12332838038633e-05, |
|
"loss": 2.3018, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.160475482912333e-05, |
|
"loss": 2.4669, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.197622585438336e-05, |
|
"loss": 2.3105, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.234769687964339e-05, |
|
"loss": 2.3296, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.2719167904903416e-05, |
|
"loss": 2.3018, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.309063893016345e-05, |
|
"loss": 2.2635, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.346210995542348e-05, |
|
"loss": 2.3177, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.383358098068351e-05, |
|
"loss": 2.3303, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.420505200594354e-05, |
|
"loss": 2.3017, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.457652303120357e-05, |
|
"loss": 2.2928, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.49479940564636e-05, |
|
"loss": 2.323, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.5319465081723626e-05, |
|
"loss": 2.2759, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.5690936106983654e-05, |
|
"loss": 2.3482, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.6062407132243683e-05, |
|
"loss": 2.4109, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.643387815750372e-05, |
|
"loss": 2.0446, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.680534918276375e-05, |
|
"loss": 2.1124, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.717682020802378e-05, |
|
"loss": 2.2935, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.7548291233283806e-05, |
|
"loss": 2.218, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.7919762258543835e-05, |
|
"loss": 2.1742, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.8291233283803864e-05, |
|
"loss": 2.3684, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.866270430906389e-05, |
|
"loss": 2.1898, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.903417533432392e-05, |
|
"loss": 2.1509, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.940564635958396e-05, |
|
"loss": 2.2106, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.9777117384843986e-05, |
|
"loss": 2.2716, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.9983484723369115e-05, |
|
"loss": 2.1564, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.994219653179191e-05, |
|
"loss": 2.1979, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.9900908340214704e-05, |
|
"loss": 2.1508, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.985962014863749e-05, |
|
"loss": 2.0062, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.981833195706028e-05, |
|
"loss": 2.0443, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.9777043765483074e-05, |
|
"loss": 2.1045, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.973575557390587e-05, |
|
"loss": 2.1194, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.9694467382328656e-05, |
|
"loss": 2.1251, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.965317919075145e-05, |
|
"loss": 2.1424, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.961189099917424e-05, |
|
"loss": 2.1282, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.9570602807597026e-05, |
|
"loss": 2.0335, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.952931461601982e-05, |
|
"loss": 2.1543, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.9488026424442615e-05, |
|
"loss": 2.1021, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.94467382328654e-05, |
|
"loss": 2.0869, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.940545004128819e-05, |
|
"loss": 2.1499, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.9364161849710985e-05, |
|
"loss": 2.0766, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.932287365813378e-05, |
|
"loss": 2.139, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.928158546655657e-05, |
|
"loss": 2.0513, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.9240297274979355e-05, |
|
"loss": 2.1771, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.919900908340215e-05, |
|
"loss": 2.2364, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.915772089182494e-05, |
|
"loss": 2.0404, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.911643270024773e-05, |
|
"loss": 1.9915, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.9075144508670526e-05, |
|
"loss": 2.0518, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.9033856317093314e-05, |
|
"loss": 2.0724, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.89925681255161e-05, |
|
"loss": 1.9577, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.8951279933938896e-05, |
|
"loss": 2.1348, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.890999174236169e-05, |
|
"loss": 2.037, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.886870355078448e-05, |
|
"loss": 2.093, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.8827415359207266e-05, |
|
"loss": 1.9399, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.878612716763006e-05, |
|
"loss": 2.0158, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.8744838976052855e-05, |
|
"loss": 1.8691, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.870355078447564e-05, |
|
"loss": 1.8896, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.866226259289843e-05, |
|
"loss": 2.0427, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.8620974401321225e-05, |
|
"loss": 2.0428, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.857968620974401e-05, |
|
"loss": 1.9077, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.853839801816681e-05, |
|
"loss": 2.0966, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.8497109826589595e-05, |
|
"loss": 2.0306, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.845582163501239e-05, |
|
"loss": 1.9391, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.841453344343518e-05, |
|
"loss": 1.9048, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.8373245251857965e-05, |
|
"loss": 1.9258, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.8331957060280766e-05, |
|
"loss": 1.9493, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.8290668868703554e-05, |
|
"loss": 1.8874, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.824938067712634e-05, |
|
"loss": 1.9082, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.8208092485549136e-05, |
|
"loss": 1.9093, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.816680429397193e-05, |
|
"loss": 1.8379, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.812551610239472e-05, |
|
"loss": 1.9901, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.8084227910817506e-05, |
|
"loss": 1.9216, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.80429397192403e-05, |
|
"loss": 1.9569, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.800165152766309e-05, |
|
"loss": 1.9444, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.796036333608588e-05, |
|
"loss": 1.9036, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.791907514450867e-05, |
|
"loss": 1.8946, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.7877786952931465e-05, |
|
"loss": 1.931, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.783649876135425e-05, |
|
"loss": 1.8613, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.779521056977704e-05, |
|
"loss": 1.9037, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.775392237819984e-05, |
|
"loss": 1.7337, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.771263418662263e-05, |
|
"loss": 2.0291, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.767134599504542e-05, |
|
"loss": 1.7789, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.763005780346821e-05, |
|
"loss": 1.8012, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.7588769611891e-05, |
|
"loss": 1.8637, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.7547481420313794e-05, |
|
"loss": 1.8404, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.750619322873658e-05, |
|
"loss": 2.0005, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.7464905037159377e-05, |
|
"loss": 1.8318, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.7423616845582164e-05, |
|
"loss": 1.6325, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.738232865400496e-05, |
|
"loss": 1.8505, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.7341040462427747e-05, |
|
"loss": 1.8743, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.729975227085054e-05, |
|
"loss": 1.7277, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.725846407927333e-05, |
|
"loss": 1.8526, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.7217175887696117e-05, |
|
"loss": 1.7623, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.717588769611892e-05, |
|
"loss": 1.7366, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 4.7134599504541706e-05, |
|
"loss": 1.5675, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 4.709331131296449e-05, |
|
"loss": 1.7807, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 4.705202312138728e-05, |
|
"loss": 1.8787, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 4.7010734929810076e-05, |
|
"loss": 1.8578, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 4.696944673823287e-05, |
|
"loss": 1.8138, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 4.692815854665566e-05, |
|
"loss": 1.8116, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 4.688687035507845e-05, |
|
"loss": 1.7444, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 4.684558216350124e-05, |
|
"loss": 1.8627, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 4.680429397192403e-05, |
|
"loss": 1.8269, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 4.676300578034682e-05, |
|
"loss": 1.6736, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 4.672171758876962e-05, |
|
"loss": 1.9007, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 4.6680429397192405e-05, |
|
"loss": 1.6758, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 4.663914120561519e-05, |
|
"loss": 1.8044, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 4.659785301403799e-05, |
|
"loss": 1.7804, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 4.655656482246078e-05, |
|
"loss": 1.8263, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 4.651527663088357e-05, |
|
"loss": 1.6985, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 4.647398843930636e-05, |
|
"loss": 1.6104, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 4.643270024772915e-05, |
|
"loss": 1.815, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 4.6391412056151946e-05, |
|
"loss": 1.6286, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 4.6350123864574733e-05, |
|
"loss": 1.6516, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 4.630883567299753e-05, |
|
"loss": 1.6558, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 4.6267547481420316e-05, |
|
"loss": 1.7759, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 4.6226259289843103e-05, |
|
"loss": 1.7525, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 4.61849710982659e-05, |
|
"loss": 1.8402, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.614368290668869e-05, |
|
"loss": 1.7702, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.610239471511148e-05, |
|
"loss": 1.6884, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.606110652353427e-05, |
|
"loss": 1.8021, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.601981833195706e-05, |
|
"loss": 1.9284, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.597853014037986e-05, |
|
"loss": 1.644, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.5937241948802645e-05, |
|
"loss": 1.7579, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.589595375722543e-05, |
|
"loss": 1.6644, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.585466556564823e-05, |
|
"loss": 1.6751, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.581337737407102e-05, |
|
"loss": 1.6732, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.577208918249381e-05, |
|
"loss": 1.7098, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.57308009909166e-05, |
|
"loss": 1.8091, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.568951279933939e-05, |
|
"loss": 1.6788, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.564822460776218e-05, |
|
"loss": 1.7283, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.5606936416184974e-05, |
|
"loss": 1.6439, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.556564822460777e-05, |
|
"loss": 1.6698, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.5524360033030556e-05, |
|
"loss": 1.739, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.5483071841453344e-05, |
|
"loss": 1.6923, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.544178364987614e-05, |
|
"loss": 1.7966, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.540049545829893e-05, |
|
"loss": 1.6849, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 4.535920726672172e-05, |
|
"loss": 1.6037, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 4.531791907514451e-05, |
|
"loss": 1.7647, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 4.52766308835673e-05, |
|
"loss": 1.7864, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 4.523534269199009e-05, |
|
"loss": 1.8, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.5194054500412885e-05, |
|
"loss": 1.6583, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.515276630883567e-05, |
|
"loss": 1.6698, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.511147811725847e-05, |
|
"loss": 1.7684, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 4.5070189925681255e-05, |
|
"loss": 1.5858, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.502890173410405e-05, |
|
"loss": 1.6393, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.4987613542526844e-05, |
|
"loss": 1.65, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.494632535094963e-05, |
|
"loss": 1.5802, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.490503715937242e-05, |
|
"loss": 1.6065, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.4863748967795214e-05, |
|
"loss": 1.5884, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.482246077621801e-05, |
|
"loss": 1.7489, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.4781172584640796e-05, |
|
"loss": 1.5268, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.4739884393063584e-05, |
|
"loss": 1.6435, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.469859620148638e-05, |
|
"loss": 1.5229, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 4.4657308009909166e-05, |
|
"loss": 1.5382, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.461601981833196e-05, |
|
"loss": 1.6764, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.457473162675475e-05, |
|
"loss": 1.7127, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.453344343517754e-05, |
|
"loss": 1.5444, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.449215524360033e-05, |
|
"loss": 1.6341, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.4450867052023125e-05, |
|
"loss": 1.6411, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.440957886044591e-05, |
|
"loss": 1.5477, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 4.436829066886871e-05, |
|
"loss": 1.5743, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 4.4327002477291495e-05, |
|
"loss": 1.6598, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 4.428571428571428e-05, |
|
"loss": 1.7103, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 4.4244426094137084e-05, |
|
"loss": 1.5787, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 4.420313790255987e-05, |
|
"loss": 1.7142, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 4.416184971098266e-05, |
|
"loss": 1.5171, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 4.4120561519405454e-05, |
|
"loss": 1.6006, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 4.407927332782824e-05, |
|
"loss": 1.488, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 4.4037985136251036e-05, |
|
"loss": 1.7033, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 4.3996696944673824e-05, |
|
"loss": 1.6683, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 4.395540875309662e-05, |
|
"loss": 1.6554, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 4.3914120561519406e-05, |
|
"loss": 1.6176, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 4.3872832369942194e-05, |
|
"loss": 1.4423, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 4.383154417836499e-05, |
|
"loss": 1.5564, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 4.379025598678778e-05, |
|
"loss": 1.6337, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 4.374896779521057e-05, |
|
"loss": 1.7138, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 4.370767960363336e-05, |
|
"loss": 1.5507, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.366639141205616e-05, |
|
"loss": 1.5643, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.362510322047895e-05, |
|
"loss": 1.6057, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.3583815028901735e-05, |
|
"loss": 1.5235, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.354252683732453e-05, |
|
"loss": 1.5131, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.350123864574732e-05, |
|
"loss": 1.5029, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.345995045417011e-05, |
|
"loss": 1.5114, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.34186622625929e-05, |
|
"loss": 1.4857, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 4.3377374071015694e-05, |
|
"loss": 1.6208, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 4.333608587943848e-05, |
|
"loss": 1.5734, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 4.329479768786127e-05, |
|
"loss": 1.6537, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 4.3253509496284064e-05, |
|
"loss": 1.385, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 4.321222130470686e-05, |
|
"loss": 1.4797, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 4.3170933113129646e-05, |
|
"loss": 1.4878, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 4.3129644921552434e-05, |
|
"loss": 1.6333, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.308835672997523e-05, |
|
"loss": 1.4782, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.304706853839802e-05, |
|
"loss": 1.4026, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.300578034682081e-05, |
|
"loss": 1.6673, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.29644921552436e-05, |
|
"loss": 1.5342, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.292320396366639e-05, |
|
"loss": 1.6623, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.288191577208919e-05, |
|
"loss": 1.5922, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.2840627580511975e-05, |
|
"loss": 1.4948, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.279933938893477e-05, |
|
"loss": 1.5139, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.275805119735756e-05, |
|
"loss": 1.4698, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.2716763005780345e-05, |
|
"loss": 1.4643, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 4.267547481420314e-05, |
|
"loss": 1.5716, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 4.2634186622625934e-05, |
|
"loss": 1.4673, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 4.259289843104872e-05, |
|
"loss": 1.5823, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 4.255161023947151e-05, |
|
"loss": 1.4593, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 4.2510322047894304e-05, |
|
"loss": 1.5401, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 4.24690338563171e-05, |
|
"loss": 1.4931, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 4.242774566473989e-05, |
|
"loss": 1.5893, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 4.2386457473162674e-05, |
|
"loss": 1.7503, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 4.234516928158547e-05, |
|
"loss": 1.3982, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 4.230388109000826e-05, |
|
"loss": 1.5719, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.226259289843105e-05, |
|
"loss": 1.4949, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.2221304706853846e-05, |
|
"loss": 1.5559, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.2180016515276633e-05, |
|
"loss": 1.5806, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.213872832369942e-05, |
|
"loss": 1.5158, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.2097440132122216e-05, |
|
"loss": 1.4574, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.205615194054501e-05, |
|
"loss": 1.387, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.20148637489678e-05, |
|
"loss": 1.5243, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.1973575557390586e-05, |
|
"loss": 1.4882, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.193228736581338e-05, |
|
"loss": 1.4553, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.1890999174236175e-05, |
|
"loss": 1.4018, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.184971098265896e-05, |
|
"loss": 1.5547, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.180842279108175e-05, |
|
"loss": 1.6648, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.1767134599504545e-05, |
|
"loss": 1.4174, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 4.172584640792733e-05, |
|
"loss": 1.4215, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 4.168455821635013e-05, |
|
"loss": 1.5949, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_accuracy": 0.6898985771074042, |
|
"eval_f1": 0.2088199260520037, |
|
"eval_loss": 1.2680630683898926, |
|
"eval_precision": 0.29172505046989156, |
|
"eval_recall": 0.20532387928216864, |
|
"eval_runtime": 115.9675, |
|
"eval_samples_per_second": 232.108, |
|
"eval_steps_per_second": 14.513, |
|
"step": 3364 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 4.1643270024772915e-05, |
|
"loss": 1.5191, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 4.160198183319571e-05, |
|
"loss": 1.4042, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.15606936416185e-05, |
|
"loss": 1.3943, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.1519405450041285e-05, |
|
"loss": 1.5074, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.1478117258464086e-05, |
|
"loss": 1.4612, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 4.1436829066886874e-05, |
|
"loss": 1.2996, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 4.139554087530966e-05, |
|
"loss": 1.4856, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 4.1354252683732456e-05, |
|
"loss": 1.5183, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 4.131296449215525e-05, |
|
"loss": 1.3999, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 4.127167630057804e-05, |
|
"loss": 1.4908, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 4.1230388109000826e-05, |
|
"loss": 1.3749, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 4.118909991742362e-05, |
|
"loss": 1.3539, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 4.114781172584641e-05, |
|
"loss": 1.46, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 4.11065235342692e-05, |
|
"loss": 1.3909, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 4.106523534269199e-05, |
|
"loss": 1.3695, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 4.1023947151114785e-05, |
|
"loss": 1.3741, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 4.098265895953757e-05, |
|
"loss": 1.6322, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 4.094137076796036e-05, |
|
"loss": 1.4463, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 4.090008257638316e-05, |
|
"loss": 1.3569, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 4.085879438480595e-05, |
|
"loss": 1.4292, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 4.081750619322874e-05, |
|
"loss": 1.3239, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 4.077621800165153e-05, |
|
"loss": 1.3519, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 4.073492981007432e-05, |
|
"loss": 1.4845, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 4.0693641618497114e-05, |
|
"loss": 1.2886, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 4.06523534269199e-05, |
|
"loss": 1.4691, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 4.0611065235342696e-05, |
|
"loss": 1.4813, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 4.0569777043765484e-05, |
|
"loss": 1.3047, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 4.052848885218828e-05, |
|
"loss": 1.3411, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 4.0487200660611066e-05, |
|
"loss": 1.3753, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 4.044591246903386e-05, |
|
"loss": 1.4322, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 4.040462427745665e-05, |
|
"loss": 1.4688, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 4.0363336085879436e-05, |
|
"loss": 1.3789, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 4.032204789430224e-05, |
|
"loss": 1.2859, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 4.0280759702725025e-05, |
|
"loss": 1.3198, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 4.023947151114781e-05, |
|
"loss": 1.174, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 4.01981833195706e-05, |
|
"loss": 1.2963, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 4.0156895127993395e-05, |
|
"loss": 1.4789, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 4.011560693641619e-05, |
|
"loss": 1.3821, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 4.007431874483898e-05, |
|
"loss": 1.3983, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 4.003303055326177e-05, |
|
"loss": 1.3409, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 3.999174236168456e-05, |
|
"loss": 1.3337, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 3.995045417010735e-05, |
|
"loss": 1.2793, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 3.990916597853014e-05, |
|
"loss": 1.389, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 3.9867877786952936e-05, |
|
"loss": 1.3747, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 3.9826589595375724e-05, |
|
"loss": 1.3534, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.978530140379851e-05, |
|
"loss": 1.4509, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.9744013212221306e-05, |
|
"loss": 1.3881, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.97027250206441e-05, |
|
"loss": 1.433, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.966143682906689e-05, |
|
"loss": 1.3735, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 3.9620148637489676e-05, |
|
"loss": 1.3591, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 3.957886044591247e-05, |
|
"loss": 1.3889, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 3.9537572254335265e-05, |
|
"loss": 1.4983, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 3.949628406275805e-05, |
|
"loss": 1.388, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 3.945499587118085e-05, |
|
"loss": 1.4386, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 3.9413707679603635e-05, |
|
"loss": 1.2915, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 3.937241948802642e-05, |
|
"loss": 1.4085, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 3.933113129644922e-05, |
|
"loss": 1.3631, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 3.928984310487201e-05, |
|
"loss": 1.3878, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 3.92485549132948e-05, |
|
"loss": 1.2828, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 3.920726672171759e-05, |
|
"loss": 1.1466, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 3.916597853014038e-05, |
|
"loss": 1.3189, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 3.9124690338563176e-05, |
|
"loss": 1.3828, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.9083402146985964e-05, |
|
"loss": 1.4046, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.904211395540875e-05, |
|
"loss": 1.4652, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.9000825763831546e-05, |
|
"loss": 1.3855, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.895953757225434e-05, |
|
"loss": 1.2254, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 3.891824938067713e-05, |
|
"loss": 1.4688, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 3.8876961189099916e-05, |
|
"loss": 1.3672, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 3.883567299752271e-05, |
|
"loss": 1.3091, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 3.87943848059455e-05, |
|
"loss": 1.3738, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 3.875309661436829e-05, |
|
"loss": 1.311, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 3.871180842279109e-05, |
|
"loss": 1.392, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.8670520231213875e-05, |
|
"loss": 1.2262, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.862923203963666e-05, |
|
"loss": 1.3561, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.858794384805946e-05, |
|
"loss": 1.5167, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.854665565648225e-05, |
|
"loss": 1.3112, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 3.850536746490504e-05, |
|
"loss": 1.4052, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 3.846407927332783e-05, |
|
"loss": 1.2428, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 3.842279108175062e-05, |
|
"loss": 1.2908, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 3.838150289017341e-05, |
|
"loss": 1.3457, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 3.8340214698596204e-05, |
|
"loss": 1.3791, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 3.829892650701899e-05, |
|
"loss": 1.405, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 3.8257638315441787e-05, |
|
"loss": 1.4444, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 3.8216350123864574e-05, |
|
"loss": 1.3148, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 3.817506193228737e-05, |
|
"loss": 1.2789, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 3.813377374071016e-05, |
|
"loss": 1.3725, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 3.809248554913295e-05, |
|
"loss": 1.2572, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 3.805119735755574e-05, |
|
"loss": 1.3122, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 3.800990916597853e-05, |
|
"loss": 1.3316, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 3.796862097440133e-05, |
|
"loss": 1.3186, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 3.7927332782824116e-05, |
|
"loss": 1.4219, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 3.78860445912469e-05, |
|
"loss": 1.337, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 3.78447563996697e-05, |
|
"loss": 1.2039, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 3.7803468208092486e-05, |
|
"loss": 1.3506, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 3.776218001651528e-05, |
|
"loss": 1.316, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 3.772089182493807e-05, |
|
"loss": 1.3707, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 3.767960363336086e-05, |
|
"loss": 1.3445, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 3.763831544178365e-05, |
|
"loss": 1.2905, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 3.759702725020644e-05, |
|
"loss": 1.2962, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 3.755573905862924e-05, |
|
"loss": 1.34, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 3.751445086705203e-05, |
|
"loss": 1.2235, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 3.7473162675474815e-05, |
|
"loss": 1.325, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 3.74318744838976e-05, |
|
"loss": 1.4342, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 3.7390586292320404e-05, |
|
"loss": 1.2154, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 3.734929810074319e-05, |
|
"loss": 1.3157, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 3.730800990916598e-05, |
|
"loss": 1.2812, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 3.7266721717588773e-05, |
|
"loss": 1.2912, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 3.722543352601156e-05, |
|
"loss": 1.3025, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 3.7184145334434356e-05, |
|
"loss": 1.3333, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 3.7142857142857143e-05, |
|
"loss": 1.4011, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 3.710156895127994e-05, |
|
"loss": 1.2194, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 3.7060280759702726e-05, |
|
"loss": 1.3489, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 3.7018992568125513e-05, |
|
"loss": 1.2511, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 3.697770437654831e-05, |
|
"loss": 1.211, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 3.69364161849711e-05, |
|
"loss": 1.2259, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 3.689512799339389e-05, |
|
"loss": 1.1592, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 3.685383980181668e-05, |
|
"loss": 1.242, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 3.681255161023947e-05, |
|
"loss": 1.3463, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 3.677126341866227e-05, |
|
"loss": 1.266, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 3.6729975227085055e-05, |
|
"loss": 1.1763, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 3.668868703550785e-05, |
|
"loss": 1.3541, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 3.664739884393064e-05, |
|
"loss": 1.369, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 3.660611065235343e-05, |
|
"loss": 1.2973, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 3.656482246077622e-05, |
|
"loss": 1.3156, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 3.6523534269199014e-05, |
|
"loss": 1.2911, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 3.64822460776218e-05, |
|
"loss": 1.3744, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 3.644095788604459e-05, |
|
"loss": 1.2752, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 3.6399669694467384e-05, |
|
"loss": 1.2033, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 3.635838150289018e-05, |
|
"loss": 1.2002, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 3.6317093311312966e-05, |
|
"loss": 1.2473, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 3.6275805119735754e-05, |
|
"loss": 1.4017, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 3.623451692815855e-05, |
|
"loss": 1.1695, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 3.619322873658134e-05, |
|
"loss": 1.3059, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 3.615194054500413e-05, |
|
"loss": 1.3189, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 3.611065235342692e-05, |
|
"loss": 1.2395, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 3.606936416184971e-05, |
|
"loss": 1.4092, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 3.60280759702725e-05, |
|
"loss": 1.3118, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 3.5986787778695295e-05, |
|
"loss": 1.3575, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 3.594549958711809e-05, |
|
"loss": 1.2792, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 3.590421139554088e-05, |
|
"loss": 1.1715, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 3.5862923203963665e-05, |
|
"loss": 1.1266, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 3.582163501238646e-05, |
|
"loss": 1.2639, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 3.5780346820809254e-05, |
|
"loss": 1.1947, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 3.573905862923204e-05, |
|
"loss": 1.3212, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 3.569777043765483e-05, |
|
"loss": 1.2242, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 3.5656482246077624e-05, |
|
"loss": 1.3444, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 3.561519405450042e-05, |
|
"loss": 1.3461, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 3.5573905862923206e-05, |
|
"loss": 1.2348, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 3.5532617671345994e-05, |
|
"loss": 1.3249, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 3.549132947976879e-05, |
|
"loss": 1.2492, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 3.5450041288191576e-05, |
|
"loss": 1.2767, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 3.540875309661437e-05, |
|
"loss": 1.3091, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 3.5367464905037165e-05, |
|
"loss": 1.2656, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 3.532617671345995e-05, |
|
"loss": 1.1787, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 3.528488852188274e-05, |
|
"loss": 1.2166, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 3.5243600330305535e-05, |
|
"loss": 1.1642, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 3.520231213872833e-05, |
|
"loss": 1.1641, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 3.516102394715112e-05, |
|
"loss": 1.2701, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 3.5119735755573905e-05, |
|
"loss": 1.2755, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 3.50784475639967e-05, |
|
"loss": 1.2654, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 3.5037159372419494e-05, |
|
"loss": 1.2874, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 3.499587118084228e-05, |
|
"loss": 1.4094, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 3.495458298926507e-05, |
|
"loss": 1.2861, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 3.4913294797687864e-05, |
|
"loss": 1.3304, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 3.487200660611065e-05, |
|
"loss": 1.2916, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 3.4830718414533446e-05, |
|
"loss": 1.3441, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 3.478943022295624e-05, |
|
"loss": 1.2485, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 3.474814203137903e-05, |
|
"loss": 1.3803, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 3.4706853839801816e-05, |
|
"loss": 1.1951, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 3.4665565648224604e-05, |
|
"loss": 1.3139, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 3.4624277456647405e-05, |
|
"loss": 1.3378, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 3.458298926507019e-05, |
|
"loss": 1.2094, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 3.454170107349298e-05, |
|
"loss": 1.2898, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 3.4500412881915775e-05, |
|
"loss": 1.2546, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 3.445912469033856e-05, |
|
"loss": 1.324, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 3.441783649876136e-05, |
|
"loss": 1.2459, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 3.4376548307184145e-05, |
|
"loss": 1.2124, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 3.433526011560694e-05, |
|
"loss": 1.2847, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 3.429397192402973e-05, |
|
"loss": 1.2943, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 3.425268373245252e-05, |
|
"loss": 1.2969, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 3.421139554087531e-05, |
|
"loss": 1.2125, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 3.4170107349298104e-05, |
|
"loss": 1.2847, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 3.412881915772089e-05, |
|
"loss": 1.1312, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 3.408753096614368e-05, |
|
"loss": 1.3225, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 3.404624277456648e-05, |
|
"loss": 1.2969, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 3.400495458298927e-05, |
|
"loss": 1.2287, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 3.3963666391412056e-05, |
|
"loss": 1.2308, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 3.392237819983485e-05, |
|
"loss": 1.1808, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 3.388109000825764e-05, |
|
"loss": 1.172, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 3.383980181668043e-05, |
|
"loss": 1.1626, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 3.379851362510322e-05, |
|
"loss": 1.3131, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 3.3757225433526015e-05, |
|
"loss": 1.2337, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 3.37159372419488e-05, |
|
"loss": 1.287, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 3.367464905037159e-05, |
|
"loss": 1.2529, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 3.3633360858794385e-05, |
|
"loss": 1.2119, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 3.359207266721718e-05, |
|
"loss": 1.1748, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 3.355078447563997e-05, |
|
"loss": 1.2432, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 3.3509496284062755e-05, |
|
"loss": 1.2848, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 3.346820809248556e-05, |
|
"loss": 1.1996, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 3.3426919900908344e-05, |
|
"loss": 1.3023, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.338563170933113e-05, |
|
"loss": 1.2872, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.334434351775392e-05, |
|
"loss": 1.3858, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.3303055326176714e-05, |
|
"loss": 1.3018, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.326176713459951e-05, |
|
"loss": 1.3543, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.32204789430223e-05, |
|
"loss": 1.1292, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.317919075144509e-05, |
|
"loss": 1.184, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.313790255986788e-05, |
|
"loss": 1.1552, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.309661436829067e-05, |
|
"loss": 1.202, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.305532617671346e-05, |
|
"loss": 1.1616, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.3014037985136256e-05, |
|
"loss": 1.2804, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.297274979355904e-05, |
|
"loss": 1.2441, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.293146160198183e-05, |
|
"loss": 1.0989, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.2890173410404626e-05, |
|
"loss": 1.3184, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.284888521882742e-05, |
|
"loss": 1.1883, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.280759702725021e-05, |
|
"loss": 1.2348, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.2766308835672996e-05, |
|
"loss": 1.1964, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.272502064409579e-05, |
|
"loss": 1.2519, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.2683732452518585e-05, |
|
"loss": 1.169, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.264244426094137e-05, |
|
"loss": 1.1751, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.260115606936417e-05, |
|
"loss": 1.3166, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.2559867877786955e-05, |
|
"loss": 1.2843, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.251857968620974e-05, |
|
"loss": 1.2293, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.247729149463254e-05, |
|
"loss": 1.2106, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.243600330305533e-05, |
|
"loss": 1.1989, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.239471511147812e-05, |
|
"loss": 1.193, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.235342691990091e-05, |
|
"loss": 1.146, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.23121387283237e-05, |
|
"loss": 1.1873, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.2270850536746496e-05, |
|
"loss": 1.076, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.2229562345169284e-05, |
|
"loss": 1.2097, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.218827415359207e-05, |
|
"loss": 1.3275, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.2146985962014866e-05, |
|
"loss": 1.251, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.2105697770437654e-05, |
|
"loss": 1.2234, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.206440957886045e-05, |
|
"loss": 1.1994, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.202312138728324e-05, |
|
"loss": 1.1733, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 3.198183319570603e-05, |
|
"loss": 1.1403, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 3.194054500412882e-05, |
|
"loss": 1.1013, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 3.189925681255161e-05, |
|
"loss": 1.165, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 3.185796862097441e-05, |
|
"loss": 1.1219, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 3.1816680429397195e-05, |
|
"loss": 1.2251, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 3.177539223781998e-05, |
|
"loss": 1.1818, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 3.173410404624278e-05, |
|
"loss": 1.1326, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 3.169281585466557e-05, |
|
"loss": 1.1989, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 3.165152766308836e-05, |
|
"loss": 1.1413, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 3.161023947151115e-05, |
|
"loss": 1.1865, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.156895127993394e-05, |
|
"loss": 1.1335, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.152766308835673e-05, |
|
"loss": 1.1335, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.1486374896779524e-05, |
|
"loss": 1.1906, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.144508670520231e-05, |
|
"loss": 1.1254, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.1403798513625106e-05, |
|
"loss": 1.186, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.1362510322047894e-05, |
|
"loss": 1.1186, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.132122213047068e-05, |
|
"loss": 1.1418, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.127993393889348e-05, |
|
"loss": 1.2051, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.123864574731627e-05, |
|
"loss": 1.1497, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.119735755573906e-05, |
|
"loss": 1.3156, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.115606936416185e-05, |
|
"loss": 1.1628, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.111478117258465e-05, |
|
"loss": 1.2643, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.1073492981007435e-05, |
|
"loss": 1.1373, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.103220478943022e-05, |
|
"loss": 1.2689, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.099091659785302e-05, |
|
"loss": 1.1297, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.0949628406275805e-05, |
|
"loss": 1.2303, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.09083402146986e-05, |
|
"loss": 1.1538, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.086705202312139e-05, |
|
"loss": 1.1264, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.082576383154418e-05, |
|
"loss": 1.1043, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.078447563996697e-05, |
|
"loss": 1.2112, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.074318744838976e-05, |
|
"loss": 1.2051, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.070189925681256e-05, |
|
"loss": 1.2237, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.0660611065235346e-05, |
|
"loss": 1.1619, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.0619322873658134e-05, |
|
"loss": 1.2211, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.057803468208092e-05, |
|
"loss": 1.3253, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.0536746490503716e-05, |
|
"loss": 1.1918, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.049545829892651e-05, |
|
"loss": 1.2563, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 3.0454170107349302e-05, |
|
"loss": 1.102, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 3.041288191577209e-05, |
|
"loss": 1.2306, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 3.037159372419488e-05, |
|
"loss": 1.2248, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 3.0330305532617675e-05, |
|
"loss": 1.1809, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 3.0289017341040466e-05, |
|
"loss": 1.1658, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 3.0247729149463254e-05, |
|
"loss": 1.2236, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 3.0206440957886045e-05, |
|
"loss": 1.1089, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 3.0165152766308836e-05, |
|
"loss": 1.226, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 3.012386457473163e-05, |
|
"loss": 1.22, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 3.0082576383154422e-05, |
|
"loss": 1.1888, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 3.004128819157721e-05, |
|
"loss": 1.2021, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 3e-05, |
|
"loss": 1.2183, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 2.9958711808422792e-05, |
|
"loss": 1.2773, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.9917423616845586e-05, |
|
"loss": 1.1104, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.9876135425268374e-05, |
|
"loss": 1.2823, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.9834847233691165e-05, |
|
"loss": 1.1548, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.9793559042113956e-05, |
|
"loss": 1.2115, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.9752270850536744e-05, |
|
"loss": 1.2013, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.9710982658959542e-05, |
|
"loss": 1.1177, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.966969446738233e-05, |
|
"loss": 1.1704, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 2.962840627580512e-05, |
|
"loss": 1.1511, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 2.9587118084227912e-05, |
|
"loss": 1.1418, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 2.9545829892650706e-05, |
|
"loss": 1.127, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 2.9504541701073494e-05, |
|
"loss": 1.1554, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 2.9463253509496285e-05, |
|
"loss": 1.0901, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 2.9421965317919076e-05, |
|
"loss": 1.2227, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 2.9380677126341864e-05, |
|
"loss": 1.0584, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 2.9339388934764662e-05, |
|
"loss": 1.1485, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 2.929810074318745e-05, |
|
"loss": 1.1717, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 2.925681255161024e-05, |
|
"loss": 1.0988, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 2.9215524360033032e-05, |
|
"loss": 1.2547, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 2.917423616845582e-05, |
|
"loss": 1.22, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 2.9132947976878618e-05, |
|
"loss": 1.1028, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 2.9091659785301405e-05, |
|
"loss": 1.1602, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 2.9050371593724197e-05, |
|
"loss": 1.2626, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 2.9009083402146984e-05, |
|
"loss": 1.1956, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 2.8967795210569775e-05, |
|
"loss": 1.1146, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 2.892650701899257e-05, |
|
"loss": 1.243, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 2.888521882741536e-05, |
|
"loss": 1.0613, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 2.8843930635838152e-05, |
|
"loss": 1.2575, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 2.880264244426094e-05, |
|
"loss": 1.0758, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 2.8761354252683738e-05, |
|
"loss": 1.0957, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 2.8720066061106526e-05, |
|
"loss": 1.223, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 2.8678777869529317e-05, |
|
"loss": 1.2306, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 2.8637489677952108e-05, |
|
"loss": 1.149, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 2.8596201486374896e-05, |
|
"loss": 1.1652, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 2.855491329479769e-05, |
|
"loss": 1.0307, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 2.851362510322048e-05, |
|
"loss": 1.1703, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 2.8472336911643272e-05, |
|
"loss": 1.0903, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 2.843104872006606e-05, |
|
"loss": 1.1752, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 2.838976052848885e-05, |
|
"loss": 1.1753, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 2.8348472336911646e-05, |
|
"loss": 1.1358, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 2.8307184145334437e-05, |
|
"loss": 1.0608, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 2.8265895953757228e-05, |
|
"loss": 1.1006, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.8224607762180016e-05, |
|
"loss": 1.191, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.8183319570602807e-05, |
|
"loss": 1.1456, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.81420313790256e-05, |
|
"loss": 1.2327, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.8100743187448392e-05, |
|
"loss": 1.1942, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.805945499587118e-05, |
|
"loss": 1.1548, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.801816680429397e-05, |
|
"loss": 1.0672, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 2.7976878612716766e-05, |
|
"loss": 1.1671, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 2.7935590421139557e-05, |
|
"loss": 1.2069, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 2.7894302229562348e-05, |
|
"loss": 1.1314, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 2.7853014037985136e-05, |
|
"loss": 1.0567, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.7811725846407927e-05, |
|
"loss": 1.1529, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_accuracy": 0.7594085522160716, |
|
"eval_f1": 0.391139702780038, |
|
"eval_loss": 0.9176002740859985, |
|
"eval_precision": 0.5332888650055155, |
|
"eval_recall": 0.3667394560847271, |
|
"eval_runtime": 114.5822, |
|
"eval_samples_per_second": 234.914, |
|
"eval_steps_per_second": 14.688, |
|
"step": 6729 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.777043765483072e-05, |
|
"loss": 1.1577, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.7729149463253512e-05, |
|
"loss": 1.0793, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 2.7687861271676304e-05, |
|
"loss": 0.97, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 2.764657308009909e-05, |
|
"loss": 1.1013, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 2.7605284888521882e-05, |
|
"loss": 1.1128, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 2.7563996696944677e-05, |
|
"loss": 1.0391, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 2.7522708505367468e-05, |
|
"loss": 1.0868, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 2.7481420313790256e-05, |
|
"loss": 1.0598, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 2.7440132122213047e-05, |
|
"loss": 1.0763, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 2.7398843930635838e-05, |
|
"loss": 1.0376, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 2.7357555739058633e-05, |
|
"loss": 1.1522, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 2.7316267547481424e-05, |
|
"loss": 1.0975, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 2.727497935590421e-05, |
|
"loss": 1.1273, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 2.7233691164327003e-05, |
|
"loss": 1.0447, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 2.7192402972749797e-05, |
|
"loss": 0.9637, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 2.7151114781172588e-05, |
|
"loss": 1.0614, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 2.7109826589595376e-05, |
|
"loss": 0.9819, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 2.7068538398018167e-05, |
|
"loss": 1.065, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 2.7027250206440958e-05, |
|
"loss": 1.1279, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 2.6985962014863753e-05, |
|
"loss": 1.1818, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 2.6944673823286544e-05, |
|
"loss": 1.0979, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 2.690338563170933e-05, |
|
"loss": 0.9635, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 2.6862097440132123e-05, |
|
"loss": 1.1322, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 2.6820809248554914e-05, |
|
"loss": 1.0714, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 2.6779521056977708e-05, |
|
"loss": 1.0064, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 2.6738232865400496e-05, |
|
"loss": 1.1475, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 2.6696944673823287e-05, |
|
"loss": 1.0754, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 2.6655656482246078e-05, |
|
"loss": 1.0984, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 2.6614368290668866e-05, |
|
"loss": 1.0474, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 2.6573080099091664e-05, |
|
"loss": 1.065, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 2.653179190751445e-05, |
|
"loss": 1.0154, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 2.6490503715937243e-05, |
|
"loss": 1.1216, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 2.6449215524360034e-05, |
|
"loss": 0.9945, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 2.640792733278283e-05, |
|
"loss": 0.9421, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 2.636663914120562e-05, |
|
"loss": 1.1399, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 2.6325350949628407e-05, |
|
"loss": 1.0276, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 2.62840627580512e-05, |
|
"loss": 1.0576, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 2.6242774566473986e-05, |
|
"loss": 1.0006, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 2.6201486374896784e-05, |
|
"loss": 1.1287, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 2.616019818331957e-05, |
|
"loss": 1.0196, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 2.6118909991742363e-05, |
|
"loss": 1.039, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 2.6077621800165154e-05, |
|
"loss": 0.997, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 2.603633360858794e-05, |
|
"loss": 1.1505, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 2.599504541701074e-05, |
|
"loss": 1.0114, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 2.5953757225433527e-05, |
|
"loss": 1.1124, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 2.591246903385632e-05, |
|
"loss": 1.0978, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 2.587118084227911e-05, |
|
"loss": 1.1963, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 2.5829892650701897e-05, |
|
"loss": 1.075, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 2.5788604459124692e-05, |
|
"loss": 1.0681, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 2.5747316267547483e-05, |
|
"loss": 1.0323, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 2.5706028075970274e-05, |
|
"loss": 1.0897, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 2.5664739884393062e-05, |
|
"loss": 1.1528, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 2.562345169281586e-05, |
|
"loss": 1.037, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 2.5582163501238647e-05, |
|
"loss": 1.0544, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 2.554087530966144e-05, |
|
"loss": 1.1794, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 2.549958711808423e-05, |
|
"loss": 1.05, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 2.5458298926507017e-05, |
|
"loss": 1.0657, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 2.5417010734929815e-05, |
|
"loss": 1.104, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 2.5375722543352603e-05, |
|
"loss": 1.1565, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 2.5334434351775394e-05, |
|
"loss": 1.0604, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 2.5293146160198182e-05, |
|
"loss": 1.1646, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 2.5251857968620973e-05, |
|
"loss": 1.0713, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 2.5210569777043767e-05, |
|
"loss": 0.9944, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 2.516928158546656e-05, |
|
"loss": 1.1764, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 2.512799339388935e-05, |
|
"loss": 1.0989, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 2.5086705202312137e-05, |
|
"loss": 1.0236, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 2.504541701073493e-05, |
|
"loss": 1.0349, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 2.5004128819157723e-05, |
|
"loss": 1.0024, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 2.4962840627580514e-05, |
|
"loss": 1.0735, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 2.4921552436003305e-05, |
|
"loss": 0.9657, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 2.4880264244426096e-05, |
|
"loss": 0.9673, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 2.4838976052848888e-05, |
|
"loss": 1.0233, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 2.4797687861271675e-05, |
|
"loss": 1.0292, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 2.475639966969447e-05, |
|
"loss": 1.1065, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 2.4715111478117258e-05, |
|
"loss": 1.0382, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 2.4673823286540052e-05, |
|
"loss": 1.0011, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 2.463253509496284e-05, |
|
"loss": 1.1076, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 2.4591246903385634e-05, |
|
"loss": 0.9976, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 2.4549958711808425e-05, |
|
"loss": 1.0667, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 2.4508670520231213e-05, |
|
"loss": 0.9816, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 2.4467382328654008e-05, |
|
"loss": 1.0732, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 2.4426094137076795e-05, |
|
"loss": 1.1308, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 2.438480594549959e-05, |
|
"loss": 1.1325, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 2.4343517753922378e-05, |
|
"loss": 1.0117, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 2.4302229562345172e-05, |
|
"loss": 1.05, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 2.4260941370767963e-05, |
|
"loss": 1.0774, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 2.421965317919075e-05, |
|
"loss": 1.1242, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 2.4178364987613546e-05, |
|
"loss": 1.0643, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 2.4137076796036333e-05, |
|
"loss": 1.02, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 2.4095788604459128e-05, |
|
"loss": 1.0318, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 2.4054500412881916e-05, |
|
"loss": 1.0233, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 2.4013212221304707e-05, |
|
"loss": 1.1628, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 2.3971924029727498e-05, |
|
"loss": 0.977, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 2.393063583815029e-05, |
|
"loss": 1.0223, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 2.3889347646573083e-05, |
|
"loss": 1.0534, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 2.384805945499587e-05, |
|
"loss": 1.0421, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 2.3806771263418666e-05, |
|
"loss": 0.9778, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 2.3765483071841453e-05, |
|
"loss": 1.0498, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 2.3724194880264244e-05, |
|
"loss": 0.9477, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 2.3682906688687036e-05, |
|
"loss": 1.0464, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 2.3641618497109827e-05, |
|
"loss": 1.0557, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 2.360033030553262e-05, |
|
"loss": 1.1357, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 2.355904211395541e-05, |
|
"loss": 1.0043, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 2.3517753922378203e-05, |
|
"loss": 0.9995, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 2.347646573080099e-05, |
|
"loss": 0.9967, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 2.3435177539223782e-05, |
|
"loss": 1.1144, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 2.3393889347646573e-05, |
|
"loss": 0.9845, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 2.3352601156069365e-05, |
|
"loss": 0.9468, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 2.331131296449216e-05, |
|
"loss": 1.0404, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 2.3270024772914947e-05, |
|
"loss": 0.926, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 2.3228736581337738e-05, |
|
"loss": 1.1061, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 2.318744838976053e-05, |
|
"loss": 1.0672, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 2.314616019818332e-05, |
|
"loss": 1.0285, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 2.310487200660611e-05, |
|
"loss": 1.0303, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 2.3063583815028902e-05, |
|
"loss": 1.0394, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 2.3022295623451694e-05, |
|
"loss": 1.0357, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 2.2981007431874485e-05, |
|
"loss": 1.0837, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 2.2939719240297276e-05, |
|
"loss": 0.9979, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 2.2898431048720067e-05, |
|
"loss": 0.9874, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 2.2857142857142858e-05, |
|
"loss": 0.935, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 2.281585466556565e-05, |
|
"loss": 1.0546, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 2.277456647398844e-05, |
|
"loss": 0.8943, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 2.273327828241123e-05, |
|
"loss": 0.9844, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 2.2691990090834023e-05, |
|
"loss": 1.0761, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 2.2650701899256814e-05, |
|
"loss": 0.9882, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 2.2609413707679605e-05, |
|
"loss": 1.0094, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 2.2568125516102396e-05, |
|
"loss": 1.0059, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 2.2526837324525187e-05, |
|
"loss": 0.9582, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 2.2485549132947978e-05, |
|
"loss": 0.9667, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 2.244426094137077e-05, |
|
"loss": 0.9865, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 2.240297274979356e-05, |
|
"loss": 1.1066, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 2.236168455821635e-05, |
|
"loss": 1.0651, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 2.2320396366639143e-05, |
|
"loss": 1.1177, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 2.2279108175061934e-05, |
|
"loss": 1.0275, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 2.2237819983484725e-05, |
|
"loss": 0.9847, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 2.2196531791907516e-05, |
|
"loss": 0.9942, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 2.2155243600330307e-05, |
|
"loss": 1.0896, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 2.2113955408753098e-05, |
|
"loss": 1.0107, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 2.207266721717589e-05, |
|
"loss": 1.0275, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 2.203137902559868e-05, |
|
"loss": 1.1083, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 2.199009083402147e-05, |
|
"loss": 1.0743, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 2.1948802642444263e-05, |
|
"loss": 1.0205, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 2.1907514450867054e-05, |
|
"loss": 1.0024, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 2.186622625928984e-05, |
|
"loss": 1.0004, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 2.1824938067712636e-05, |
|
"loss": 1.0878, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 2.1783649876135427e-05, |
|
"loss": 1.1676, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 2.174236168455822e-05, |
|
"loss": 1.0535, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 2.170107349298101e-05, |
|
"loss": 0.9921, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 2.1659785301403797e-05, |
|
"loss": 1.0698, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 2.1618497109826592e-05, |
|
"loss": 1.0949, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 2.157720891824938e-05, |
|
"loss": 1.0792, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 2.1535920726672174e-05, |
|
"loss": 0.9902, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 2.1494632535094965e-05, |
|
"loss": 0.9935, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 2.1453344343517756e-05, |
|
"loss": 1.1441, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 2.1412056151940547e-05, |
|
"loss": 1.0967, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 2.1370767960363335e-05, |
|
"loss": 1.0815, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 2.132947976878613e-05, |
|
"loss": 0.9774, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 2.1288191577208917e-05, |
|
"loss": 0.9863, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 2.1246903385631712e-05, |
|
"loss": 0.9236, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 2.12056151940545e-05, |
|
"loss": 1.1055, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 2.1164327002477294e-05, |
|
"loss": 1.0841, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 2.1123038810900085e-05, |
|
"loss": 1.0685, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 2.1081750619322873e-05, |
|
"loss": 1.0606, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 2.1040462427745667e-05, |
|
"loss": 1.0472, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 2.0999174236168455e-05, |
|
"loss": 1.0021, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 2.095788604459125e-05, |
|
"loss": 1.0177, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 2.0916597853014037e-05, |
|
"loss": 0.9739, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 2.087530966143683e-05, |
|
"loss": 1.1024, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 2.0834021469859623e-05, |
|
"loss": 1.0497, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 2.079273327828241e-05, |
|
"loss": 1.0258, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 2.0751445086705205e-05, |
|
"loss": 1.0898, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 2.0710156895127993e-05, |
|
"loss": 1.0222, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 2.0668868703550787e-05, |
|
"loss": 1.0016, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 2.0627580511973575e-05, |
|
"loss": 0.8966, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 2.0586292320396366e-05, |
|
"loss": 1.0537, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 2.0545004128819157e-05, |
|
"loss": 0.9702, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 2.050371593724195e-05, |
|
"loss": 1.021, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 2.0462427745664743e-05, |
|
"loss": 1.0565, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 2.042113955408753e-05, |
|
"loss": 1.0511, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 2.0379851362510325e-05, |
|
"loss": 0.9871, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 2.0338563170933113e-05, |
|
"loss": 1.0572, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 2.0297274979355904e-05, |
|
"loss": 1.008, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 2.0255986787778695e-05, |
|
"loss": 1.0466, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 2.0214698596201486e-05, |
|
"loss": 0.9638, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 2.017341040462428e-05, |
|
"loss": 1.0415, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 2.013212221304707e-05, |
|
"loss": 1.0746, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 2.009083402146986e-05, |
|
"loss": 0.968, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 2.004954582989265e-05, |
|
"loss": 1.1031, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 2.0008257638315442e-05, |
|
"loss": 0.9651, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 1.9966969446738233e-05, |
|
"loss": 0.9085, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 1.9925681255161024e-05, |
|
"loss": 0.9675, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 1.988439306358382e-05, |
|
"loss": 1.0355, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 1.9843104872006607e-05, |
|
"loss": 0.9327, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 1.9801816680429398e-05, |
|
"loss": 1.0593, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 1.976052848885219e-05, |
|
"loss": 0.9641, |
|
"step": 8670 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 1.971924029727498e-05, |
|
"loss": 0.9795, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 1.967795210569777e-05, |
|
"loss": 1.1296, |
|
"step": 8690 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 1.9636663914120562e-05, |
|
"loss": 1.0139, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 1.9595375722543353e-05, |
|
"loss": 1.0541, |
|
"step": 8710 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 1.9554087530966144e-05, |
|
"loss": 0.9802, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 1.9512799339388936e-05, |
|
"loss": 1.0617, |
|
"step": 8730 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 1.9471511147811727e-05, |
|
"loss": 1.0648, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 1.9430222956234518e-05, |
|
"loss": 0.9572, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 1.938893476465731e-05, |
|
"loss": 1.0266, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 1.93476465730801e-05, |
|
"loss": 0.9396, |
|
"step": 8770 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 1.930635838150289e-05, |
|
"loss": 1.0307, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 1.9265070189925682e-05, |
|
"loss": 1.0705, |
|
"step": 8790 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 1.9223781998348473e-05, |
|
"loss": 1.1392, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 1.9182493806771264e-05, |
|
"loss": 0.993, |
|
"step": 8810 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 1.9141205615194056e-05, |
|
"loss": 1.076, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 1.9099917423616847e-05, |
|
"loss": 0.9515, |
|
"step": 8830 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 1.9058629232039638e-05, |
|
"loss": 0.8663, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 1.901734104046243e-05, |
|
"loss": 1.0044, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 1.897605284888522e-05, |
|
"loss": 1.0138, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 1.893476465730801e-05, |
|
"loss": 0.9632, |
|
"step": 8870 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 1.8893476465730802e-05, |
|
"loss": 1.1827, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 1.8852188274153593e-05, |
|
"loss": 0.877, |
|
"step": 8890 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 1.8810900082576385e-05, |
|
"loss": 1.0811, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 1.8769611890999176e-05, |
|
"loss": 1.0753, |
|
"step": 8910 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 1.8728323699421967e-05, |
|
"loss": 0.9152, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 1.8687035507844758e-05, |
|
"loss": 1.0065, |
|
"step": 8930 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 1.864574731626755e-05, |
|
"loss": 1.0118, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 1.860445912469034e-05, |
|
"loss": 0.9865, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 1.856317093311313e-05, |
|
"loss": 1.0144, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 1.852188274153592e-05, |
|
"loss": 1.1353, |
|
"step": 8970 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 1.8480594549958714e-05, |
|
"loss": 1.1403, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 1.84393063583815e-05, |
|
"loss": 0.9343, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 1.8398018166804296e-05, |
|
"loss": 1.0659, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 1.8356729975227087e-05, |
|
"loss": 1.0746, |
|
"step": 9010 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 1.8315441783649878e-05, |
|
"loss": 0.9841, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 1.827415359207267e-05, |
|
"loss": 1.0243, |
|
"step": 9030 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 1.8232865400495457e-05, |
|
"loss": 0.9991, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 1.819157720891825e-05, |
|
"loss": 1.0827, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 1.815028901734104e-05, |
|
"loss": 0.9928, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 1.8109000825763834e-05, |
|
"loss": 1.1273, |
|
"step": 9070 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 1.8067712634186625e-05, |
|
"loss": 0.9417, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 1.8026424442609416e-05, |
|
"loss": 1.0517, |
|
"step": 9090 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 1.7985136251032207e-05, |
|
"loss": 0.9739, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 1.7943848059454995e-05, |
|
"loss": 0.9638, |
|
"step": 9110 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 1.790255986787779e-05, |
|
"loss": 0.914, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 1.7861271676300577e-05, |
|
"loss": 0.9318, |
|
"step": 9130 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.781998348472337e-05, |
|
"loss": 0.9152, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.777869529314616e-05, |
|
"loss": 1.0002, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.773740710156895e-05, |
|
"loss": 0.937, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.7696118909991745e-05, |
|
"loss": 1.0005, |
|
"step": 9170 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.7654830718414533e-05, |
|
"loss": 1.0194, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.7613542526837327e-05, |
|
"loss": 1.016, |
|
"step": 9190 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.7572254335260115e-05, |
|
"loss": 1.0235, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.753096614368291e-05, |
|
"loss": 0.8933, |
|
"step": 9210 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.7489677952105697e-05, |
|
"loss": 0.9867, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.7448389760528488e-05, |
|
"loss": 0.968, |
|
"step": 9230 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.7407101568951283e-05, |
|
"loss": 1.116, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.736581337737407e-05, |
|
"loss": 0.9769, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.7324525185796865e-05, |
|
"loss": 1.0424, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.7283236994219653e-05, |
|
"loss": 1.0752, |
|
"step": 9270 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.7241948802642447e-05, |
|
"loss": 0.972, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.7200660611065235e-05, |
|
"loss": 1.0728, |
|
"step": 9290 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.7159372419488026e-05, |
|
"loss": 0.9349, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.711808422791082e-05, |
|
"loss": 0.9323, |
|
"step": 9310 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.707679603633361e-05, |
|
"loss": 1.0655, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.7035507844756403e-05, |
|
"loss": 1.0919, |
|
"step": 9330 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.699421965317919e-05, |
|
"loss": 1.0599, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.695293146160198e-05, |
|
"loss": 1.0199, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.6911643270024773e-05, |
|
"loss": 1.0026, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.6870355078447564e-05, |
|
"loss": 1.0486, |
|
"step": 9370 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.6829066886870355e-05, |
|
"loss": 1.124, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.6787778695293146e-05, |
|
"loss": 1.0031, |
|
"step": 9390 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.674649050371594e-05, |
|
"loss": 1.0062, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.670520231213873e-05, |
|
"loss": 1.1425, |
|
"step": 9410 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.666391412056152e-05, |
|
"loss": 0.9444, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.662262592898431e-05, |
|
"loss": 1.0659, |
|
"step": 9430 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.6581337737407102e-05, |
|
"loss": 1.0006, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.6540049545829893e-05, |
|
"loss": 1.0085, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.6498761354252684e-05, |
|
"loss": 1.0337, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.645747316267548e-05, |
|
"loss": 0.9573, |
|
"step": 9470 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.6416184971098266e-05, |
|
"loss": 1.0845, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.6374896779521057e-05, |
|
"loss": 1.0099, |
|
"step": 9490 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.633360858794385e-05, |
|
"loss": 0.9384, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.629232039636664e-05, |
|
"loss": 1.0056, |
|
"step": 9510 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.625103220478943e-05, |
|
"loss": 1.0884, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.6209744013212222e-05, |
|
"loss": 0.943, |
|
"step": 9530 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.6168455821635013e-05, |
|
"loss": 0.9475, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.6127167630057804e-05, |
|
"loss": 0.9335, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.6085879438480595e-05, |
|
"loss": 1.0512, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.6044591246903386e-05, |
|
"loss": 0.8913, |
|
"step": 9570 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.6003303055326177e-05, |
|
"loss": 0.9363, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.596201486374897e-05, |
|
"loss": 1.0665, |
|
"step": 9590 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.592072667217176e-05, |
|
"loss": 1.0039, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.587943848059455e-05, |
|
"loss": 0.9873, |
|
"step": 9610 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.5838150289017342e-05, |
|
"loss": 0.9473, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.5796862097440133e-05, |
|
"loss": 1.0315, |
|
"step": 9630 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.5755573905862924e-05, |
|
"loss": 1.0087, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.5714285714285715e-05, |
|
"loss": 0.9808, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.5672997522708506e-05, |
|
"loss": 1.0034, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.5631709331131298e-05, |
|
"loss": 1.0216, |
|
"step": 9670 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.559042113955409e-05, |
|
"loss": 1.0435, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.554913294797688e-05, |
|
"loss": 1.0322, |
|
"step": 9690 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.550784475639967e-05, |
|
"loss": 1.0437, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.5466556564822462e-05, |
|
"loss": 0.9632, |
|
"step": 9710 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.5425268373245253e-05, |
|
"loss": 1.0344, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.538398018166804e-05, |
|
"loss": 1.0011, |
|
"step": 9730 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.5342691990090835e-05, |
|
"loss": 1.0548, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.5301403798513627e-05, |
|
"loss": 0.8887, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.5260115606936418e-05, |
|
"loss": 1.0574, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.5218827415359207e-05, |
|
"loss": 1.0122, |
|
"step": 9770 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.5177539223782e-05, |
|
"loss": 0.9831, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.513625103220479e-05, |
|
"loss": 0.9642, |
|
"step": 9790 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.509496284062758e-05, |
|
"loss": 1.0055, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.5053674649050373e-05, |
|
"loss": 1.0261, |
|
"step": 9810 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.5012386457473163e-05, |
|
"loss": 1.1469, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.4971098265895956e-05, |
|
"loss": 1.0866, |
|
"step": 9830 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.4929810074318745e-05, |
|
"loss": 0.9159, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.4888521882741538e-05, |
|
"loss": 0.9805, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.4847233691164327e-05, |
|
"loss": 1.0301, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.4805945499587118e-05, |
|
"loss": 0.8966, |
|
"step": 9870 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.4764657308009911e-05, |
|
"loss": 0.9712, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.47233691164327e-05, |
|
"loss": 0.908, |
|
"step": 9890 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.4682080924855493e-05, |
|
"loss": 1.0038, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.4640792733278283e-05, |
|
"loss": 0.9793, |
|
"step": 9910 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.4599504541701072e-05, |
|
"loss": 0.9966, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.4558216350123865e-05, |
|
"loss": 1.0065, |
|
"step": 9930 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.4516928158546656e-05, |
|
"loss": 1.0938, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.4475639966969447e-05, |
|
"loss": 0.9796, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.4434351775392238e-05, |
|
"loss": 1.0492, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.4393063583815031e-05, |
|
"loss": 0.893, |
|
"step": 9970 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.435177539223782e-05, |
|
"loss": 0.9077, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.431048720066061e-05, |
|
"loss": 0.9832, |
|
"step": 9990 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.4269199009083403e-05, |
|
"loss": 0.9928, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.4227910817506192e-05, |
|
"loss": 0.942, |
|
"step": 10010 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.4186622625928985e-05, |
|
"loss": 1.0488, |
|
"step": 10020 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.4145334434351776e-05, |
|
"loss": 1.0297, |
|
"step": 10030 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.4104046242774569e-05, |
|
"loss": 0.9424, |
|
"step": 10040 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.4062758051197359e-05, |
|
"loss": 0.9437, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.4021469859620148e-05, |
|
"loss": 1.0469, |
|
"step": 10060 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.398018166804294e-05, |
|
"loss": 0.8718, |
|
"step": 10070 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.393889347646573e-05, |
|
"loss": 1.0406, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.3897605284888523e-05, |
|
"loss": 1.0071, |
|
"step": 10090 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_accuracy": 0.7909128060333618, |
|
"eval_f1": 0.48598154603607574, |
|
"eval_loss": 0.7875941395759583, |
|
"eval_precision": 0.6262905869925817, |
|
"eval_recall": 0.4524727278747925, |
|
"eval_runtime": 114.6581, |
|
"eval_samples_per_second": 234.759, |
|
"eval_steps_per_second": 14.678, |
|
"step": 10094 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.3856317093311314e-05, |
|
"loss": 0.8953, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.3815028901734104e-05, |
|
"loss": 0.9918, |
|
"step": 10110 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 1.3773740710156896e-05, |
|
"loss": 0.992, |
|
"step": 10120 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 1.3732452518579686e-05, |
|
"loss": 0.8583, |
|
"step": 10130 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 1.3691164327002479e-05, |
|
"loss": 0.8438, |
|
"step": 10140 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 1.3649876135425268e-05, |
|
"loss": 0.8838, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 1.360858794384806e-05, |
|
"loss": 0.9325, |
|
"step": 10160 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 1.3567299752270852e-05, |
|
"loss": 0.9193, |
|
"step": 10170 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 1.3526011560693641e-05, |
|
"loss": 0.8802, |
|
"step": 10180 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 1.3484723369116434e-05, |
|
"loss": 0.9594, |
|
"step": 10190 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 1.3443435177539224e-05, |
|
"loss": 0.8835, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 1.3402146985962016e-05, |
|
"loss": 0.9174, |
|
"step": 10210 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 1.3360858794384806e-05, |
|
"loss": 0.9595, |
|
"step": 10220 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 1.3319570602807599e-05, |
|
"loss": 0.9409, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 1.3278282411230388e-05, |
|
"loss": 0.987, |
|
"step": 10240 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.323699421965318e-05, |
|
"loss": 0.8553, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.3195706028075972e-05, |
|
"loss": 0.9893, |
|
"step": 10260 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.3154417836498761e-05, |
|
"loss": 0.973, |
|
"step": 10270 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 1.3113129644921554e-05, |
|
"loss": 0.8118, |
|
"step": 10280 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 1.3071841453344344e-05, |
|
"loss": 0.9595, |
|
"step": 10290 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 1.3030553261767133e-05, |
|
"loss": 0.9936, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 1.2989265070189926e-05, |
|
"loss": 1.0295, |
|
"step": 10310 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 1.2947976878612717e-05, |
|
"loss": 1.0863, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 1.290668868703551e-05, |
|
"loss": 0.9424, |
|
"step": 10330 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 1.28654004954583e-05, |
|
"loss": 0.9466, |
|
"step": 10340 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 1.2824112303881092e-05, |
|
"loss": 0.9604, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 1.2782824112303882e-05, |
|
"loss": 1.0032, |
|
"step": 10360 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 1.2741535920726671e-05, |
|
"loss": 0.904, |
|
"step": 10370 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 1.2700247729149464e-05, |
|
"loss": 0.9557, |
|
"step": 10380 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.2658959537572255e-05, |
|
"loss": 0.9056, |
|
"step": 10390 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.2617671345995046e-05, |
|
"loss": 0.963, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.2576383154417837e-05, |
|
"loss": 0.9466, |
|
"step": 10410 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 1.253509496284063e-05, |
|
"loss": 0.8427, |
|
"step": 10420 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 1.249380677126342e-05, |
|
"loss": 1.0001, |
|
"step": 10430 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 1.245251857968621e-05, |
|
"loss": 1.0051, |
|
"step": 10440 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 1.2411230388109002e-05, |
|
"loss": 0.8966, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 1.2369942196531791e-05, |
|
"loss": 0.8635, |
|
"step": 10460 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 1.2328654004954584e-05, |
|
"loss": 0.973, |
|
"step": 10470 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 1.2287365813377375e-05, |
|
"loss": 0.953, |
|
"step": 10480 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 1.2246077621800166e-05, |
|
"loss": 0.8928, |
|
"step": 10490 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 1.2204789430222957e-05, |
|
"loss": 1.0, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 1.2163501238645748e-05, |
|
"loss": 0.9181, |
|
"step": 10510 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 1.212221304706854e-05, |
|
"loss": 0.9614, |
|
"step": 10520 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 1.2080924855491329e-05, |
|
"loss": 0.912, |
|
"step": 10530 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 1.203963666391412e-05, |
|
"loss": 1.0034, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 1.1998348472336913e-05, |
|
"loss": 0.9272, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 1.1957060280759704e-05, |
|
"loss": 0.9857, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 1.1915772089182495e-05, |
|
"loss": 0.8547, |
|
"step": 10570 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 1.1874483897605286e-05, |
|
"loss": 0.9533, |
|
"step": 10580 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 1.1833195706028076e-05, |
|
"loss": 0.9289, |
|
"step": 10590 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 1.1791907514450867e-05, |
|
"loss": 1.0137, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 1.1750619322873658e-05, |
|
"loss": 0.9359, |
|
"step": 10610 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 1.1709331131296449e-05, |
|
"loss": 0.8753, |
|
"step": 10620 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 1.1668042939719242e-05, |
|
"loss": 0.9577, |
|
"step": 10630 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 1.1626754748142033e-05, |
|
"loss": 1.0264, |
|
"step": 10640 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 1.1585466556564822e-05, |
|
"loss": 0.9486, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 1.1544178364987614e-05, |
|
"loss": 0.8917, |
|
"step": 10660 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 1.1502890173410405e-05, |
|
"loss": 0.8931, |
|
"step": 10670 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 1.1461601981833196e-05, |
|
"loss": 1.0185, |
|
"step": 10680 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 1.1420313790255987e-05, |
|
"loss": 1.0049, |
|
"step": 10690 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 1.1379025598678778e-05, |
|
"loss": 0.9114, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 1.133773740710157e-05, |
|
"loss": 0.9465, |
|
"step": 10710 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 1.129644921552436e-05, |
|
"loss": 0.9741, |
|
"step": 10720 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 1.1255161023947151e-05, |
|
"loss": 0.9789, |
|
"step": 10730 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 1.1213872832369943e-05, |
|
"loss": 0.9447, |
|
"step": 10740 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 1.1172584640792734e-05, |
|
"loss": 0.8526, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 1.1131296449215525e-05, |
|
"loss": 0.9653, |
|
"step": 10760 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 1.1090008257638316e-05, |
|
"loss": 0.8994, |
|
"step": 10770 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 1.1048720066061107e-05, |
|
"loss": 0.9167, |
|
"step": 10780 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 1.1007431874483898e-05, |
|
"loss": 0.9848, |
|
"step": 10790 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 1.096614368290669e-05, |
|
"loss": 0.9352, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 1.092485549132948e-05, |
|
"loss": 0.8854, |
|
"step": 10810 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 1.0883567299752271e-05, |
|
"loss": 0.9074, |
|
"step": 10820 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 1.0842279108175063e-05, |
|
"loss": 0.9669, |
|
"step": 10830 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 1.0800990916597854e-05, |
|
"loss": 1.0194, |
|
"step": 10840 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 1.0759702725020645e-05, |
|
"loss": 1.0115, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 1.0718414533443436e-05, |
|
"loss": 0.9864, |
|
"step": 10860 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 1.0677126341866227e-05, |
|
"loss": 0.9659, |
|
"step": 10870 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 1.0635838150289018e-05, |
|
"loss": 0.8672, |
|
"step": 10880 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 1.059454995871181e-05, |
|
"loss": 0.9295, |
|
"step": 10890 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 1.05532617671346e-05, |
|
"loss": 1.0901, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 1.051197357555739e-05, |
|
"loss": 0.9348, |
|
"step": 10910 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 1.0470685383980183e-05, |
|
"loss": 0.8586, |
|
"step": 10920 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 1.0429397192402974e-05, |
|
"loss": 0.9332, |
|
"step": 10930 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 1.0388109000825765e-05, |
|
"loss": 0.9389, |
|
"step": 10940 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 1.0346820809248556e-05, |
|
"loss": 0.9745, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 1.0305532617671347e-05, |
|
"loss": 1.0033, |
|
"step": 10960 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 1.0264244426094137e-05, |
|
"loss": 0.9185, |
|
"step": 10970 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 1.0222956234516928e-05, |
|
"loss": 0.841, |
|
"step": 10980 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 1.0181668042939719e-05, |
|
"loss": 0.9474, |
|
"step": 10990 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 1.0140379851362512e-05, |
|
"loss": 1.0044, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 1.0099091659785303e-05, |
|
"loss": 0.9491, |
|
"step": 11010 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 1.0057803468208094e-05, |
|
"loss": 0.8805, |
|
"step": 11020 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 1.0016515276630883e-05, |
|
"loss": 0.9112, |
|
"step": 11030 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 9.975227085053674e-06, |
|
"loss": 0.9389, |
|
"step": 11040 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 9.933938893476466e-06, |
|
"loss": 0.8873, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 9.892650701899257e-06, |
|
"loss": 0.9235, |
|
"step": 11060 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 9.851362510322048e-06, |
|
"loss": 0.8426, |
|
"step": 11070 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 9.81007431874484e-06, |
|
"loss": 0.883, |
|
"step": 11080 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 9.768786127167632e-06, |
|
"loss": 0.853, |
|
"step": 11090 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 9.727497935590421e-06, |
|
"loss": 0.9797, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 9.686209744013212e-06, |
|
"loss": 0.8857, |
|
"step": 11110 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 9.644921552436003e-06, |
|
"loss": 0.9336, |
|
"step": 11120 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 9.603633360858795e-06, |
|
"loss": 0.8257, |
|
"step": 11130 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 9.562345169281586e-06, |
|
"loss": 1.057, |
|
"step": 11140 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 9.521056977704377e-06, |
|
"loss": 0.9302, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 9.479768786127168e-06, |
|
"loss": 0.8687, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 9.438480594549959e-06, |
|
"loss": 0.8862, |
|
"step": 11170 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 9.39719240297275e-06, |
|
"loss": 1.0411, |
|
"step": 11180 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 9.355904211395541e-06, |
|
"loss": 0.8978, |
|
"step": 11190 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 9.314616019818332e-06, |
|
"loss": 0.8987, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 9.273327828241124e-06, |
|
"loss": 0.9941, |
|
"step": 11210 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 9.232039636663915e-06, |
|
"loss": 0.9068, |
|
"step": 11220 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 9.190751445086706e-06, |
|
"loss": 0.876, |
|
"step": 11230 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 9.149463253509497e-06, |
|
"loss": 0.8567, |
|
"step": 11240 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 9.108175061932288e-06, |
|
"loss": 0.8855, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 9.066886870355079e-06, |
|
"loss": 0.9531, |
|
"step": 11260 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 9.02559867877787e-06, |
|
"loss": 0.9153, |
|
"step": 11270 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 8.984310487200661e-06, |
|
"loss": 0.8551, |
|
"step": 11280 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 8.94302229562345e-06, |
|
"loss": 0.9453, |
|
"step": 11290 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 8.901734104046244e-06, |
|
"loss": 0.8633, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 8.860445912469035e-06, |
|
"loss": 0.8984, |
|
"step": 11310 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 8.819157720891826e-06, |
|
"loss": 0.8235, |
|
"step": 11320 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 8.777869529314617e-06, |
|
"loss": 0.9257, |
|
"step": 11330 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 8.736581337737408e-06, |
|
"loss": 0.8852, |
|
"step": 11340 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 8.695293146160198e-06, |
|
"loss": 0.9009, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 8.654004954582989e-06, |
|
"loss": 0.9958, |
|
"step": 11360 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 8.61271676300578e-06, |
|
"loss": 0.9091, |
|
"step": 11370 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 8.571428571428573e-06, |
|
"loss": 0.8763, |
|
"step": 11380 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 8.530140379851364e-06, |
|
"loss": 0.9321, |
|
"step": 11390 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 8.488852188274155e-06, |
|
"loss": 0.8947, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 8.447563996696944e-06, |
|
"loss": 0.8589, |
|
"step": 11410 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 8.406275805119735e-06, |
|
"loss": 0.9435, |
|
"step": 11420 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 8.364987613542527e-06, |
|
"loss": 0.9077, |
|
"step": 11430 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 8.323699421965318e-06, |
|
"loss": 0.8764, |
|
"step": 11440 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 8.28241123038811e-06, |
|
"loss": 1.009, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 8.241123038810902e-06, |
|
"loss": 0.9208, |
|
"step": 11460 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 8.199834847233693e-06, |
|
"loss": 0.9348, |
|
"step": 11470 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 8.158546655656482e-06, |
|
"loss": 0.8944, |
|
"step": 11480 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 8.117258464079273e-06, |
|
"loss": 0.9075, |
|
"step": 11490 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 8.075970272502064e-06, |
|
"loss": 0.9387, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 8.034682080924856e-06, |
|
"loss": 0.9026, |
|
"step": 11510 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 7.993393889347647e-06, |
|
"loss": 0.9384, |
|
"step": 11520 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 7.95210569777044e-06, |
|
"loss": 0.9407, |
|
"step": 11530 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 7.910817506193229e-06, |
|
"loss": 0.9137, |
|
"step": 11540 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 7.86952931461602e-06, |
|
"loss": 0.8788, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 7.828241123038811e-06, |
|
"loss": 0.9236, |
|
"step": 11560 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 7.786952931461602e-06, |
|
"loss": 0.9725, |
|
"step": 11570 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 7.745664739884393e-06, |
|
"loss": 0.9734, |
|
"step": 11580 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 7.704376548307184e-06, |
|
"loss": 0.9422, |
|
"step": 11590 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 7.663088356729976e-06, |
|
"loss": 0.9497, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 7.621800165152766e-06, |
|
"loss": 0.9108, |
|
"step": 11610 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 7.580511973575558e-06, |
|
"loss": 0.8832, |
|
"step": 11620 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 7.539223781998349e-06, |
|
"loss": 0.9141, |
|
"step": 11630 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 7.49793559042114e-06, |
|
"loss": 0.8465, |
|
"step": 11640 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 7.456647398843931e-06, |
|
"loss": 0.8659, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 7.415359207266722e-06, |
|
"loss": 0.8838, |
|
"step": 11660 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 7.374071015689513e-06, |
|
"loss": 0.942, |
|
"step": 11670 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 7.332782824112304e-06, |
|
"loss": 0.9328, |
|
"step": 11680 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 7.291494632535095e-06, |
|
"loss": 0.9015, |
|
"step": 11690 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 7.250206440957887e-06, |
|
"loss": 0.9307, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 7.208918249380678e-06, |
|
"loss": 0.9011, |
|
"step": 11710 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 7.167630057803469e-06, |
|
"loss": 0.9057, |
|
"step": 11720 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 7.126341866226259e-06, |
|
"loss": 0.9029, |
|
"step": 11730 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 7.0850536746490505e-06, |
|
"loss": 0.8731, |
|
"step": 11740 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 7.043765483071842e-06, |
|
"loss": 0.7972, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 7.002477291494633e-06, |
|
"loss": 0.9547, |
|
"step": 11760 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 6.961189099917424e-06, |
|
"loss": 0.8425, |
|
"step": 11770 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 6.919900908340216e-06, |
|
"loss": 0.9347, |
|
"step": 11780 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 6.878612716763005e-06, |
|
"loss": 0.9269, |
|
"step": 11790 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 6.837324525185796e-06, |
|
"loss": 0.8175, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 6.796036333608588e-06, |
|
"loss": 0.922, |
|
"step": 11810 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 6.7547481420313794e-06, |
|
"loss": 0.879, |
|
"step": 11820 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 6.7134599504541706e-06, |
|
"loss": 0.9067, |
|
"step": 11830 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 6.672171758876962e-06, |
|
"loss": 0.8772, |
|
"step": 11840 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 6.630883567299754e-06, |
|
"loss": 0.9466, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 6.589595375722543e-06, |
|
"loss": 0.8492, |
|
"step": 11860 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 6.548307184145334e-06, |
|
"loss": 0.8936, |
|
"step": 11870 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 6.507018992568126e-06, |
|
"loss": 0.8145, |
|
"step": 11880 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 6.465730800990917e-06, |
|
"loss": 0.7936, |
|
"step": 11890 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 6.424442609413708e-06, |
|
"loss": 0.9378, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 6.3831544178364995e-06, |
|
"loss": 0.8736, |
|
"step": 11910 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 6.34186622625929e-06, |
|
"loss": 0.8803, |
|
"step": 11920 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 6.300578034682081e-06, |
|
"loss": 0.9858, |
|
"step": 11930 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 6.259289843104872e-06, |
|
"loss": 0.8679, |
|
"step": 11940 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 6.218001651527663e-06, |
|
"loss": 0.9454, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 6.176713459950455e-06, |
|
"loss": 1.0106, |
|
"step": 11960 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 6.135425268373245e-06, |
|
"loss": 0.8688, |
|
"step": 11970 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 6.0941370767960365e-06, |
|
"loss": 0.896, |
|
"step": 11980 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 6.052848885218828e-06, |
|
"loss": 0.8965, |
|
"step": 11990 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 6.011560693641619e-06, |
|
"loss": 0.9319, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 5.97027250206441e-06, |
|
"loss": 0.9851, |
|
"step": 12010 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 5.928984310487201e-06, |
|
"loss": 0.9143, |
|
"step": 12020 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 5.887696118909992e-06, |
|
"loss": 0.9021, |
|
"step": 12030 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 5.846407927332783e-06, |
|
"loss": 0.9096, |
|
"step": 12040 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 5.805119735755574e-06, |
|
"loss": 0.9316, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 5.763831544178365e-06, |
|
"loss": 0.9772, |
|
"step": 12060 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 5.722543352601157e-06, |
|
"loss": 0.8342, |
|
"step": 12070 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 5.681255161023948e-06, |
|
"loss": 0.9234, |
|
"step": 12080 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 5.639966969446738e-06, |
|
"loss": 0.8931, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 5.598678777869529e-06, |
|
"loss": 0.8829, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 5.557390586292321e-06, |
|
"loss": 0.9263, |
|
"step": 12110 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 5.516102394715112e-06, |
|
"loss": 1.031, |
|
"step": 12120 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 5.4748142031379025e-06, |
|
"loss": 0.9329, |
|
"step": 12130 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 5.433526011560694e-06, |
|
"loss": 1.0245, |
|
"step": 12140 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 5.392237819983486e-06, |
|
"loss": 0.8366, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 5.350949628406276e-06, |
|
"loss": 0.8749, |
|
"step": 12160 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 5.309661436829067e-06, |
|
"loss": 0.8166, |
|
"step": 12170 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 5.268373245251858e-06, |
|
"loss": 0.8904, |
|
"step": 12180 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 5.227085053674649e-06, |
|
"loss": 0.8515, |
|
"step": 12190 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 5.18579686209744e-06, |
|
"loss": 0.8638, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 5.1445086705202315e-06, |
|
"loss": 0.843, |
|
"step": 12210 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 5.103220478943023e-06, |
|
"loss": 0.9534, |
|
"step": 12220 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 5.061932287365814e-06, |
|
"loss": 0.9772, |
|
"step": 12230 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 5.020644095788605e-06, |
|
"loss": 0.9394, |
|
"step": 12240 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 4.979355904211395e-06, |
|
"loss": 0.8932, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 4.938067712634187e-06, |
|
"loss": 0.947, |
|
"step": 12260 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 4.896779521056978e-06, |
|
"loss": 0.9598, |
|
"step": 12270 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 4.8554913294797685e-06, |
|
"loss": 0.891, |
|
"step": 12280 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 4.81420313790256e-06, |
|
"loss": 0.8912, |
|
"step": 12290 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 4.772914946325352e-06, |
|
"loss": 0.8975, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 4.731626754748143e-06, |
|
"loss": 0.8922, |
|
"step": 12310 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 4.690338563170933e-06, |
|
"loss": 0.9523, |
|
"step": 12320 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 4.649050371593724e-06, |
|
"loss": 0.89, |
|
"step": 12330 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 4.607762180016516e-06, |
|
"loss": 0.9059, |
|
"step": 12340 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 4.566473988439306e-06, |
|
"loss": 0.9347, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 4.5251857968620975e-06, |
|
"loss": 0.8894, |
|
"step": 12360 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 4.483897605284889e-06, |
|
"loss": 0.8675, |
|
"step": 12370 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 4.44260941370768e-06, |
|
"loss": 1.0249, |
|
"step": 12380 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 4.401321222130471e-06, |
|
"loss": 0.8599, |
|
"step": 12390 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 4.360033030553262e-06, |
|
"loss": 0.9422, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 4.318744838976053e-06, |
|
"loss": 0.9763, |
|
"step": 12410 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 4.277456647398844e-06, |
|
"loss": 0.8395, |
|
"step": 12420 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 4.236168455821635e-06, |
|
"loss": 0.8698, |
|
"step": 12430 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 4.194880264244426e-06, |
|
"loss": 0.8642, |
|
"step": 12440 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 4.1535920726672176e-06, |
|
"loss": 0.9101, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 4.112303881090009e-06, |
|
"loss": 0.9178, |
|
"step": 12460 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 4.0710156895128e-06, |
|
"loss": 0.8701, |
|
"step": 12470 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 4.02972749793559e-06, |
|
"loss": 0.8926, |
|
"step": 12480 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 3.988439306358382e-06, |
|
"loss": 0.92, |
|
"step": 12490 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 3.947151114781173e-06, |
|
"loss": 0.8252, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 3.9058629232039634e-06, |
|
"loss": 0.8395, |
|
"step": 12510 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 3.8645747316267546e-06, |
|
"loss": 0.9389, |
|
"step": 12520 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 3.8232865400495465e-06, |
|
"loss": 0.7671, |
|
"step": 12530 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 3.781998348472337e-06, |
|
"loss": 0.8992, |
|
"step": 12540 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 3.7407101568951284e-06, |
|
"loss": 0.8744, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 3.6994219653179195e-06, |
|
"loss": 0.8293, |
|
"step": 12560 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 3.65813377374071e-06, |
|
"loss": 0.9531, |
|
"step": 12570 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 3.6168455821635013e-06, |
|
"loss": 1.0111, |
|
"step": 12580 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 3.575557390586293e-06, |
|
"loss": 0.9563, |
|
"step": 12590 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 3.534269199009083e-06, |
|
"loss": 0.8381, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 3.4929810074318747e-06, |
|
"loss": 0.918, |
|
"step": 12610 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 3.4516928158546658e-06, |
|
"loss": 0.9229, |
|
"step": 12620 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 3.4104046242774565e-06, |
|
"loss": 0.9257, |
|
"step": 12630 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 3.3691164327002476e-06, |
|
"loss": 0.9155, |
|
"step": 12640 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 3.327828241123039e-06, |
|
"loss": 0.8525, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 3.2865400495458303e-06, |
|
"loss": 0.9185, |
|
"step": 12660 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 3.245251857968621e-06, |
|
"loss": 0.9616, |
|
"step": 12670 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 3.203963666391412e-06, |
|
"loss": 0.9456, |
|
"step": 12680 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 3.1626754748142036e-06, |
|
"loss": 0.948, |
|
"step": 12690 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 3.1213872832369943e-06, |
|
"loss": 0.8063, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 3.0800990916597855e-06, |
|
"loss": 0.8692, |
|
"step": 12710 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 3.0388109000825766e-06, |
|
"loss": 0.8543, |
|
"step": 12720 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 2.9975227085053677e-06, |
|
"loss": 0.9614, |
|
"step": 12730 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 2.956234516928159e-06, |
|
"loss": 0.8755, |
|
"step": 12740 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 2.9149463253509495e-06, |
|
"loss": 1.0031, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 2.873658133773741e-06, |
|
"loss": 0.9364, |
|
"step": 12760 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 2.8323699421965318e-06, |
|
"loss": 0.9223, |
|
"step": 12770 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 2.791081750619323e-06, |
|
"loss": 0.9554, |
|
"step": 12780 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 2.749793559042114e-06, |
|
"loss": 0.9022, |
|
"step": 12790 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 2.708505367464905e-06, |
|
"loss": 0.9029, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 2.6672171758876962e-06, |
|
"loss": 0.9376, |
|
"step": 12810 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 2.6259289843104874e-06, |
|
"loss": 0.9986, |
|
"step": 12820 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 2.5846407927332785e-06, |
|
"loss": 0.9003, |
|
"step": 12830 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 2.5433526011560696e-06, |
|
"loss": 0.9685, |
|
"step": 12840 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 2.5020644095788607e-06, |
|
"loss": 0.9416, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 2.4607762180016514e-06, |
|
"loss": 0.8018, |
|
"step": 12860 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 2.419488026424443e-06, |
|
"loss": 0.9066, |
|
"step": 12870 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 2.3781998348472337e-06, |
|
"loss": 0.9611, |
|
"step": 12880 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 2.3369116432700252e-06, |
|
"loss": 0.8978, |
|
"step": 12890 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 2.295623451692816e-06, |
|
"loss": 0.8674, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 2.254335260115607e-06, |
|
"loss": 0.8783, |
|
"step": 12910 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 2.213047068538398e-06, |
|
"loss": 1.0783, |
|
"step": 12920 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 2.1717588769611893e-06, |
|
"loss": 0.8873, |
|
"step": 12930 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 2.13047068538398e-06, |
|
"loss": 0.9271, |
|
"step": 12940 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 2.0891824938067715e-06, |
|
"loss": 0.95, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 2.0478943022295622e-06, |
|
"loss": 1.0039, |
|
"step": 12960 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 2.0066061106523533e-06, |
|
"loss": 1.0292, |
|
"step": 12970 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 1.9653179190751445e-06, |
|
"loss": 0.8818, |
|
"step": 12980 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 1.9240297274979356e-06, |
|
"loss": 0.9431, |
|
"step": 12990 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 1.882741535920727e-06, |
|
"loss": 0.8331, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 1.8414533443435178e-06, |
|
"loss": 0.9331, |
|
"step": 13010 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 1.8001651527663087e-06, |
|
"loss": 0.9859, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 1.7588769611891e-06, |
|
"loss": 0.8269, |
|
"step": 13030 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 1.717588769611891e-06, |
|
"loss": 0.8372, |
|
"step": 13040 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 1.6763005780346821e-06, |
|
"loss": 0.8566, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 1.6350123864574732e-06, |
|
"loss": 0.9604, |
|
"step": 13060 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 1.5937241948802644e-06, |
|
"loss": 0.8972, |
|
"step": 13070 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 1.5524360033030555e-06, |
|
"loss": 0.9291, |
|
"step": 13080 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 1.5111478117258466e-06, |
|
"loss": 0.8829, |
|
"step": 13090 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 1.4698596201486377e-06, |
|
"loss": 0.906, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 1.4285714285714286e-06, |
|
"loss": 0.937, |
|
"step": 13110 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 1.3872832369942197e-06, |
|
"loss": 0.88, |
|
"step": 13120 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 1.3459950454170109e-06, |
|
"loss": 0.9529, |
|
"step": 13130 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 1.3047068538398018e-06, |
|
"loss": 0.8807, |
|
"step": 13140 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 1.263418662262593e-06, |
|
"loss": 0.9028, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 1.222130470685384e-06, |
|
"loss": 0.9342, |
|
"step": 13160 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 1.1808422791081751e-06, |
|
"loss": 0.9117, |
|
"step": 13170 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 1.139554087530966e-06, |
|
"loss": 0.869, |
|
"step": 13180 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 1.0982658959537572e-06, |
|
"loss": 1.0468, |
|
"step": 13190 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 1.0569777043765483e-06, |
|
"loss": 0.9269, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 1.0156895127993394e-06, |
|
"loss": 0.9193, |
|
"step": 13210 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 9.744013212221305e-07, |
|
"loss": 0.9165, |
|
"step": 13220 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 9.331131296449216e-07, |
|
"loss": 0.9507, |
|
"step": 13230 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 8.918249380677127e-07, |
|
"loss": 0.9153, |
|
"step": 13240 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 8.505367464905038e-07, |
|
"loss": 0.953, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 8.092485549132948e-07, |
|
"loss": 0.9218, |
|
"step": 13260 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 7.679603633360859e-07, |
|
"loss": 0.9418, |
|
"step": 13270 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 7.266721717588771e-07, |
|
"loss": 0.9154, |
|
"step": 13280 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 6.853839801816681e-07, |
|
"loss": 0.8085, |
|
"step": 13290 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 6.440957886044592e-07, |
|
"loss": 0.9172, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 6.028075970272502e-07, |
|
"loss": 0.8716, |
|
"step": 13310 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 5.615194054500413e-07, |
|
"loss": 0.9145, |
|
"step": 13320 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 5.202312138728324e-07, |
|
"loss": 0.9514, |
|
"step": 13330 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 4.789430222956235e-07, |
|
"loss": 0.8823, |
|
"step": 13340 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 4.3765483071841454e-07, |
|
"loss": 1.0048, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 3.963666391412056e-07, |
|
"loss": 0.8191, |
|
"step": 13360 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 3.5507844756399673e-07, |
|
"loss": 0.9084, |
|
"step": 13370 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 3.137902559867878e-07, |
|
"loss": 0.9743, |
|
"step": 13380 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 2.7250206440957887e-07, |
|
"loss": 0.9524, |
|
"step": 13390 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 2.3121387283236997e-07, |
|
"loss": 0.8666, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 1.8992568125516104e-07, |
|
"loss": 0.9419, |
|
"step": 13410 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 1.486374896779521e-07, |
|
"loss": 0.8898, |
|
"step": 13420 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 1.0734929810074319e-07, |
|
"loss": 0.8142, |
|
"step": 13430 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 6.606110652353427e-08, |
|
"loss": 0.8921, |
|
"step": 13440 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 2.477291494632535e-08, |
|
"loss": 0.9761, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.8028383549429728, |
|
"eval_f1": 0.5170105123405551, |
|
"eval_loss": 0.7372918128967285, |
|
"eval_precision": 0.6434174024938317, |
|
"eval_recall": 0.48293329058965584, |
|
"eval_runtime": 113.7791, |
|
"eval_samples_per_second": 236.572, |
|
"eval_steps_per_second": 14.792, |
|
"step": 13456 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"step": 13456, |
|
"total_flos": 2.1807238425425895e+19, |
|
"train_loss": 1.3893684443976735, |
|
"train_runtime": 7393.9648, |
|
"train_samples_per_second": 116.493, |
|
"train_steps_per_second": 1.82 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 13456, |
|
"num_train_epochs": 4, |
|
"save_steps": 500, |
|
"total_flos": 2.1807238425425895e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|