|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.3713417039942442, |
|
"global_step": 6000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.278350515463919e-08, |
|
"loss": 2.5166, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.9587628865979384e-07, |
|
"loss": 1.9059, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.989690721649485e-07, |
|
"loss": 1.7869, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.0206185567010316e-07, |
|
"loss": 1.7863, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 5.051546391752578e-07, |
|
"loss": 1.6946, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 6.082474226804124e-07, |
|
"loss": 1.7419, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 7.11340206185567e-07, |
|
"loss": 1.645, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 8.144329896907217e-07, |
|
"loss": 1.6611, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.175257731958763e-07, |
|
"loss": 1.5985, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.020618556701031e-06, |
|
"loss": 1.6198, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.1237113402061856e-06, |
|
"loss": 1.6259, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.2268041237113403e-06, |
|
"loss": 1.5973, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.329896907216495e-06, |
|
"loss": 1.5941, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.4329896907216496e-06, |
|
"loss": 1.5597, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.5360824742268042e-06, |
|
"loss": 1.5672, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.639175257731959e-06, |
|
"loss": 1.5372, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.7422680412371134e-06, |
|
"loss": 1.5715, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.8453608247422682e-06, |
|
"loss": 1.5389, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.948453608247423e-06, |
|
"loss": 1.5525, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.0515463917525773e-06, |
|
"loss": 1.5871, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.1546391752577322e-06, |
|
"loss": 1.5442, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.2577319587628867e-06, |
|
"loss": 1.5335, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.3608247422680415e-06, |
|
"loss": 1.5103, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.463917525773196e-06, |
|
"loss": 1.5016, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 2.5670103092783504e-06, |
|
"loss": 1.5101, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 2.6701030927835053e-06, |
|
"loss": 1.5323, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 2.77319587628866e-06, |
|
"loss": 1.4785, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 2.8762886597938146e-06, |
|
"loss": 1.4132, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 2.979381443298969e-06, |
|
"loss": 1.4733, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.082474226804124e-06, |
|
"loss": 1.4258, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.1855670103092784e-06, |
|
"loss": 1.4557, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.2886597938144333e-06, |
|
"loss": 1.44, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.391752577319588e-06, |
|
"loss": 1.4348, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.494845360824742e-06, |
|
"loss": 1.406, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.597938144329897e-06, |
|
"loss": 1.4239, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.701030927835052e-06, |
|
"loss": 1.4185, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.8041237113402064e-06, |
|
"loss": 1.3954, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.907216494845361e-06, |
|
"loss": 1.3759, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.010309278350516e-06, |
|
"loss": 1.3231, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.11340206185567e-06, |
|
"loss": 1.3059, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.216494845360825e-06, |
|
"loss": 1.3258, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.31958762886598e-06, |
|
"loss": 1.3722, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.422680412371134e-06, |
|
"loss": 1.2917, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.525773195876289e-06, |
|
"loss": 1.2893, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.628865979381444e-06, |
|
"loss": 1.2865, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.731958762886599e-06, |
|
"loss": 1.28, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.835051546391753e-06, |
|
"loss": 1.299, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.9381443298969075e-06, |
|
"loss": 1.2949, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.041237113402062e-06, |
|
"loss": 1.3078, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.144329896907216e-06, |
|
"loss": 1.2242, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.247422680412372e-06, |
|
"loss": 1.2714, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.350515463917526e-06, |
|
"loss": 1.2482, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.45360824742268e-06, |
|
"loss": 1.2703, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.556701030927836e-06, |
|
"loss": 1.2578, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.65979381443299e-06, |
|
"loss": 1.2989, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.762886597938144e-06, |
|
"loss": 1.2853, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 5.8659793814433e-06, |
|
"loss": 1.3022, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 5.969072164948454e-06, |
|
"loss": 1.2871, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.0721649484536086e-06, |
|
"loss": 1.2679, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.1752577319587634e-06, |
|
"loss": 1.2732, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.278350515463918e-06, |
|
"loss": 1.2641, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.381443298969072e-06, |
|
"loss": 1.2919, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.484536082474227e-06, |
|
"loss": 1.2594, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.587628865979382e-06, |
|
"loss": 1.2592, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.690721649484536e-06, |
|
"loss": 1.2652, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.793814432989692e-06, |
|
"loss": 1.3274, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.896907216494846e-06, |
|
"loss": 1.2155, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7e-06, |
|
"loss": 1.2837, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.103092783505156e-06, |
|
"loss": 1.265, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.20618556701031e-06, |
|
"loss": 1.2427, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.309278350515464e-06, |
|
"loss": 1.2688, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.412371134020619e-06, |
|
"loss": 1.3071, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 7.515463917525773e-06, |
|
"loss": 1.2346, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 7.618556701030928e-06, |
|
"loss": 1.2246, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 7.721649484536083e-06, |
|
"loss": 1.2604, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 7.824742268041238e-06, |
|
"loss": 1.2589, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 7.927835051546391e-06, |
|
"loss": 1.2512, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.030927835051548e-06, |
|
"loss": 1.2229, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.134020618556701e-06, |
|
"loss": 1.2326, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.237113402061856e-06, |
|
"loss": 1.3097, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.34020618556701e-06, |
|
"loss": 1.2358, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.443298969072166e-06, |
|
"loss": 1.2746, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.54639175257732e-06, |
|
"loss": 1.3063, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.649484536082475e-06, |
|
"loss": 1.2702, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.75257731958763e-06, |
|
"loss": 1.2421, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.855670103092783e-06, |
|
"loss": 1.3047, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.95876288659794e-06, |
|
"loss": 1.2452, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.061855670103093e-06, |
|
"loss": 1.2753, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.164948453608248e-06, |
|
"loss": 1.2532, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.268041237113403e-06, |
|
"loss": 1.2379, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.371134020618558e-06, |
|
"loss": 1.2891, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.474226804123711e-06, |
|
"loss": 1.2773, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.577319587628868e-06, |
|
"loss": 1.2753, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.68041237113402e-06, |
|
"loss": 1.1976, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.783505154639176e-06, |
|
"loss": 1.2721, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.88659793814433e-06, |
|
"loss": 1.3058, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.989690721649485e-06, |
|
"loss": 1.2383, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.998105303046253e-06, |
|
"loss": 1.2356, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.996000084208755e-06, |
|
"loss": 1.2639, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.993894865371256e-06, |
|
"loss": 1.2689, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.991789646533758e-06, |
|
"loss": 1.2764, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.989684427696261e-06, |
|
"loss": 1.2602, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.987579208858762e-06, |
|
"loss": 1.3034, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.985473990021264e-06, |
|
"loss": 1.2318, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.983368771183765e-06, |
|
"loss": 1.252, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.981263552346267e-06, |
|
"loss": 1.3111, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.97915833350877e-06, |
|
"loss": 1.255, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.97705311467127e-06, |
|
"loss": 1.2736, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.974947895833773e-06, |
|
"loss": 1.3128, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.972842676996276e-06, |
|
"loss": 1.2349, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.970737458158777e-06, |
|
"loss": 1.2797, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.968632239321279e-06, |
|
"loss": 1.2698, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.96652702048378e-06, |
|
"loss": 1.2714, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.964421801646282e-06, |
|
"loss": 1.2965, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.962316582808785e-06, |
|
"loss": 1.2764, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.960211363971286e-06, |
|
"loss": 1.2294, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.958106145133788e-06, |
|
"loss": 1.2623, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.956000926296289e-06, |
|
"loss": 1.2814, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.953895707458791e-06, |
|
"loss": 1.2793, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.951790488621294e-06, |
|
"loss": 1.2749, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.949685269783795e-06, |
|
"loss": 1.2689, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.947580050946297e-06, |
|
"loss": 1.2913, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.9454748321088e-06, |
|
"loss": 1.2386, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.9433696132713e-06, |
|
"loss": 1.2519, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.941264394433803e-06, |
|
"loss": 1.2893, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.939159175596304e-06, |
|
"loss": 1.2398, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.937053956758805e-06, |
|
"loss": 1.2836, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.934948737921307e-06, |
|
"loss": 1.2326, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.93284351908381e-06, |
|
"loss": 1.2768, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.93073830024631e-06, |
|
"loss": 1.2584, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.928633081408813e-06, |
|
"loss": 1.2738, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.926527862571314e-06, |
|
"loss": 1.2329, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.924422643733816e-06, |
|
"loss": 1.2623, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.922317424896319e-06, |
|
"loss": 1.2839, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.92021220605882e-06, |
|
"loss": 1.2806, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.918106987221322e-06, |
|
"loss": 1.232, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.916001768383825e-06, |
|
"loss": 1.256, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.913896549546325e-06, |
|
"loss": 1.225, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.911791330708828e-06, |
|
"loss": 1.2247, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.909686111871329e-06, |
|
"loss": 1.2422, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.907580893033831e-06, |
|
"loss": 1.2499, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.905475674196334e-06, |
|
"loss": 1.2569, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.903370455358834e-06, |
|
"loss": 1.2771, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.901265236521337e-06, |
|
"loss": 1.2789, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.89916001768384e-06, |
|
"loss": 1.2213, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.89705479884634e-06, |
|
"loss": 1.2393, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.894949580008843e-06, |
|
"loss": 1.257, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.892844361171344e-06, |
|
"loss": 1.2572, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.890739142333846e-06, |
|
"loss": 1.2503, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.888633923496349e-06, |
|
"loss": 1.2404, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.88652870465885e-06, |
|
"loss": 1.2847, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.884423485821352e-06, |
|
"loss": 1.2551, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.882318266983854e-06, |
|
"loss": 1.2741, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.880213048146355e-06, |
|
"loss": 1.2735, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.878107829308858e-06, |
|
"loss": 1.273, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.876002610471358e-06, |
|
"loss": 1.2357, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.873897391633861e-06, |
|
"loss": 1.1864, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.871792172796363e-06, |
|
"loss": 1.277, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.869686953958864e-06, |
|
"loss": 1.2375, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.867581735121367e-06, |
|
"loss": 1.2754, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.865476516283868e-06, |
|
"loss": 1.2051, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.86337129744637e-06, |
|
"loss": 1.2579, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.861266078608873e-06, |
|
"loss": 1.2216, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.859160859771373e-06, |
|
"loss": 1.2529, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.857055640933876e-06, |
|
"loss": 1.207, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.854950422096378e-06, |
|
"loss": 1.2275, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.85284520325888e-06, |
|
"loss": 1.2769, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.850739984421382e-06, |
|
"loss": 1.2165, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.848634765583883e-06, |
|
"loss": 1.2903, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.846529546746385e-06, |
|
"loss": 1.2548, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.844424327908888e-06, |
|
"loss": 1.2652, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.842319109071388e-06, |
|
"loss": 1.2718, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.84021389023389e-06, |
|
"loss": 1.269, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.838108671396393e-06, |
|
"loss": 1.2362, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.836003452558894e-06, |
|
"loss": 1.205, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.833898233721397e-06, |
|
"loss": 1.2649, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.831793014883897e-06, |
|
"loss": 1.2517, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.8296877960464e-06, |
|
"loss": 1.2015, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.827582577208902e-06, |
|
"loss": 1.2382, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.825477358371403e-06, |
|
"loss": 1.2236, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.823372139533906e-06, |
|
"loss": 1.2503, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.821266920696407e-06, |
|
"loss": 1.2409, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.819161701858909e-06, |
|
"loss": 1.2236, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.817056483021412e-06, |
|
"loss": 1.2246, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.814951264183912e-06, |
|
"loss": 1.234, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.812846045346415e-06, |
|
"loss": 1.2384, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.810740826508917e-06, |
|
"loss": 1.2103, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.808635607671418e-06, |
|
"loss": 1.2398, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.80653038883392e-06, |
|
"loss": 1.2246, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.804425169996421e-06, |
|
"loss": 1.1894, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.802319951158924e-06, |
|
"loss": 1.262, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.800214732321426e-06, |
|
"loss": 1.2595, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.798109513483927e-06, |
|
"loss": 1.1954, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.79600429464643e-06, |
|
"loss": 1.2578, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.793899075808932e-06, |
|
"loss": 1.2286, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.791793856971433e-06, |
|
"loss": 1.2455, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.789688638133936e-06, |
|
"loss": 1.214, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.787583419296436e-06, |
|
"loss": 1.2488, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.785478200458939e-06, |
|
"loss": 1.266, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.783372981621441e-06, |
|
"loss": 1.2139, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.781267762783942e-06, |
|
"loss": 1.2516, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.779162543946445e-06, |
|
"loss": 1.2878, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.777057325108945e-06, |
|
"loss": 1.2494, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.774952106271448e-06, |
|
"loss": 1.207, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.77284688743395e-06, |
|
"loss": 1.2796, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.770741668596451e-06, |
|
"loss": 1.2285, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.768636449758954e-06, |
|
"loss": 1.2467, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.766531230921456e-06, |
|
"loss": 1.1801, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.764426012083957e-06, |
|
"loss": 1.2399, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.76232079324646e-06, |
|
"loss": 1.2359, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.76021557440896e-06, |
|
"loss": 1.2074, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.758110355571463e-06, |
|
"loss": 1.2601, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.756005136733965e-06, |
|
"loss": 1.2456, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.753899917896466e-06, |
|
"loss": 1.2479, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.751794699058969e-06, |
|
"loss": 1.2593, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.749689480221471e-06, |
|
"loss": 1.1856, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.747584261383972e-06, |
|
"loss": 1.2634, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.745479042546474e-06, |
|
"loss": 1.2046, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.743373823708975e-06, |
|
"loss": 1.2753, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.741268604871478e-06, |
|
"loss": 1.2393, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.739163386033979e-06, |
|
"loss": 1.224, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.737058167196481e-06, |
|
"loss": 1.2767, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.734952948358982e-06, |
|
"loss": 1.2584, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.732847729521484e-06, |
|
"loss": 1.2717, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.730742510683985e-06, |
|
"loss": 1.2364, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.728637291846488e-06, |
|
"loss": 1.2354, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.72653207300899e-06, |
|
"loss": 1.2347, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.724426854171491e-06, |
|
"loss": 1.2537, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.722321635333993e-06, |
|
"loss": 1.2186, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.720216416496496e-06, |
|
"loss": 1.2351, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.718111197658997e-06, |
|
"loss": 1.2325, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.7160059788215e-06, |
|
"loss": 1.1996, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.713900759984e-06, |
|
"loss": 1.2023, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.711795541146503e-06, |
|
"loss": 1.2527, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.709690322309005e-06, |
|
"loss": 1.2281, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.707585103471506e-06, |
|
"loss": 1.2382, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.705479884634008e-06, |
|
"loss": 1.2405, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.70337466579651e-06, |
|
"loss": 1.248, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.701269446959012e-06, |
|
"loss": 1.224, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.699164228121514e-06, |
|
"loss": 1.22, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.697059009284015e-06, |
|
"loss": 1.219, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.694953790446517e-06, |
|
"loss": 1.2518, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.69284857160902e-06, |
|
"loss": 1.219, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.69074335277152e-06, |
|
"loss": 1.2168, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.688638133934023e-06, |
|
"loss": 1.2469, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.686532915096524e-06, |
|
"loss": 1.2381, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.684427696259027e-06, |
|
"loss": 1.2001, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.682322477421529e-06, |
|
"loss": 1.2004, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.68021725858403e-06, |
|
"loss": 1.2409, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.678112039746532e-06, |
|
"loss": 1.2389, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.676006820909035e-06, |
|
"loss": 1.242, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.673901602071536e-06, |
|
"loss": 1.2372, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.671796383234038e-06, |
|
"loss": 1.2223, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.669691164396539e-06, |
|
"loss": 1.2506, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.667585945559041e-06, |
|
"loss": 1.2093, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.665480726721544e-06, |
|
"loss": 1.2171, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.663375507884045e-06, |
|
"loss": 1.2363, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.661270289046547e-06, |
|
"loss": 1.2978, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.65916507020905e-06, |
|
"loss": 1.2216, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.65705985137155e-06, |
|
"loss": 1.1937, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.654954632534053e-06, |
|
"loss": 1.2366, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.652849413696554e-06, |
|
"loss": 1.2465, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.650744194859056e-06, |
|
"loss": 1.2704, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.648638976021559e-06, |
|
"loss": 1.2113, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.64653375718406e-06, |
|
"loss": 1.2679, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.644428538346562e-06, |
|
"loss": 1.2005, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.642323319509063e-06, |
|
"loss": 1.2474, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.640218100671565e-06, |
|
"loss": 1.2308, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.638112881834068e-06, |
|
"loss": 1.2391, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.636007662996569e-06, |
|
"loss": 1.1968, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.633902444159071e-06, |
|
"loss": 1.2001, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.631797225321574e-06, |
|
"loss": 1.2688, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.629692006484075e-06, |
|
"loss": 1.2646, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.627586787646577e-06, |
|
"loss": 1.2606, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.625481568809078e-06, |
|
"loss": 1.1915, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.62337634997158e-06, |
|
"loss": 1.204, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.621271131134083e-06, |
|
"loss": 1.2128, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.619165912296584e-06, |
|
"loss": 1.2116, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.617060693459086e-06, |
|
"loss": 1.2287, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.614955474621589e-06, |
|
"loss": 1.2443, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.61285025578409e-06, |
|
"loss": 1.2926, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.610745036946592e-06, |
|
"loss": 1.2195, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.608639818109093e-06, |
|
"loss": 1.2345, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.606534599271595e-06, |
|
"loss": 1.2588, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.604429380434098e-06, |
|
"loss": 1.2392, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.602324161596599e-06, |
|
"loss": 1.2529, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.600218942759101e-06, |
|
"loss": 1.2119, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.598113723921602e-06, |
|
"loss": 1.2416, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.596008505084104e-06, |
|
"loss": 1.2111, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.593903286246607e-06, |
|
"loss": 1.2493, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.591798067409108e-06, |
|
"loss": 1.2481, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.58969284857161e-06, |
|
"loss": 1.2265, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.587587629734113e-06, |
|
"loss": 1.2549, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.585482410896613e-06, |
|
"loss": 1.2474, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.583377192059116e-06, |
|
"loss": 1.1773, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.581271973221617e-06, |
|
"loss": 1.2612, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.57916675438412e-06, |
|
"loss": 1.2247, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.577061535546622e-06, |
|
"loss": 1.2075, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.574956316709123e-06, |
|
"loss": 1.1812, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.572851097871625e-06, |
|
"loss": 1.2058, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.570745879034128e-06, |
|
"loss": 1.2781, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.568640660196628e-06, |
|
"loss": 1.2572, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.566535441359131e-06, |
|
"loss": 1.2794, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.564430222521632e-06, |
|
"loss": 1.2136, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.562325003684134e-06, |
|
"loss": 1.2632, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.560219784846637e-06, |
|
"loss": 1.2584, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.558114566009137e-06, |
|
"loss": 1.286, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.55600934717164e-06, |
|
"loss": 1.247, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.55390412833414e-06, |
|
"loss": 1.2715, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.551798909496643e-06, |
|
"loss": 1.2184, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.549693690659146e-06, |
|
"loss": 1.261, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.547588471821647e-06, |
|
"loss": 1.2183, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.545483252984149e-06, |
|
"loss": 1.1887, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.543378034146652e-06, |
|
"loss": 1.2405, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.541272815309152e-06, |
|
"loss": 1.2499, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.539167596471653e-06, |
|
"loss": 1.2164, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.537062377634156e-06, |
|
"loss": 1.2614, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.534957158796657e-06, |
|
"loss": 1.2475, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.532851939959159e-06, |
|
"loss": 1.2559, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.530746721121662e-06, |
|
"loss": 1.2457, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.528641502284162e-06, |
|
"loss": 1.2228, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.526536283446665e-06, |
|
"loss": 1.219, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.524431064609166e-06, |
|
"loss": 1.2255, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.522325845771668e-06, |
|
"loss": 1.1923, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.52022062693417e-06, |
|
"loss": 1.1996, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.518115408096671e-06, |
|
"loss": 1.2186, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.516010189259174e-06, |
|
"loss": 1.2384, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.513904970421676e-06, |
|
"loss": 1.2119, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.511799751584177e-06, |
|
"loss": 1.2455, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.50969453274668e-06, |
|
"loss": 1.2314, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.50758931390918e-06, |
|
"loss": 1.1995, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.505484095071683e-06, |
|
"loss": 1.2308, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.503378876234186e-06, |
|
"loss": 1.1957, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.501273657396686e-06, |
|
"loss": 1.2557, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.499168438559189e-06, |
|
"loss": 1.2351, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.497063219721691e-06, |
|
"loss": 1.2085, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.494958000884192e-06, |
|
"loss": 1.2241, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.492852782046695e-06, |
|
"loss": 1.1909, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.490747563209195e-06, |
|
"loss": 1.1886, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.488642344371698e-06, |
|
"loss": 1.2161, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.4865371255342e-06, |
|
"loss": 1.2718, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.484431906696701e-06, |
|
"loss": 1.2007, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.482326687859204e-06, |
|
"loss": 1.2038, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.480221469021705e-06, |
|
"loss": 1.2153, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.478116250184207e-06, |
|
"loss": 1.1932, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.47601103134671e-06, |
|
"loss": 1.255, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.47390581250921e-06, |
|
"loss": 1.2738, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.471800593671713e-06, |
|
"loss": 1.2533, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.469695374834215e-06, |
|
"loss": 1.2408, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.467590155996716e-06, |
|
"loss": 1.2331, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.465484937159219e-06, |
|
"loss": 1.2205, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.46337971832172e-06, |
|
"loss": 1.2569, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.461274499484222e-06, |
|
"loss": 1.2245, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.459169280646724e-06, |
|
"loss": 1.2192, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.457064061809225e-06, |
|
"loss": 1.2518, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.454958842971728e-06, |
|
"loss": 1.2619, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.45285362413423e-06, |
|
"loss": 1.1859, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.450748405296731e-06, |
|
"loss": 1.2216, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.448643186459234e-06, |
|
"loss": 1.221, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.446537967621734e-06, |
|
"loss": 1.2224, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.444432748784237e-06, |
|
"loss": 1.2324, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.44232752994674e-06, |
|
"loss": 1.1747, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.44022231110924e-06, |
|
"loss": 1.1962, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.438117092271743e-06, |
|
"loss": 1.2235, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.436011873434245e-06, |
|
"loss": 1.2081, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.433906654596746e-06, |
|
"loss": 1.2398, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.431801435759248e-06, |
|
"loss": 1.248, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.42969621692175e-06, |
|
"loss": 1.1936, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.427590998084252e-06, |
|
"loss": 1.2317, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.425485779246754e-06, |
|
"loss": 1.2139, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.423380560409255e-06, |
|
"loss": 1.1601, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.421275341571758e-06, |
|
"loss": 1.2127, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.419170122734258e-06, |
|
"loss": 1.2082, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.41706490389676e-06, |
|
"loss": 1.1971, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.414959685059263e-06, |
|
"loss": 1.2289, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.412854466221764e-06, |
|
"loss": 1.2133, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.410749247384267e-06, |
|
"loss": 1.2111, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.408644028546769e-06, |
|
"loss": 1.2342, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.40653880970927e-06, |
|
"loss": 1.217, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.404433590871772e-06, |
|
"loss": 1.2651, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.402328372034273e-06, |
|
"loss": 1.2259, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.400223153196776e-06, |
|
"loss": 1.2434, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.398117934359278e-06, |
|
"loss": 1.2199, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.396012715521779e-06, |
|
"loss": 1.2299, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.393907496684282e-06, |
|
"loss": 1.2156, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.391802277846784e-06, |
|
"loss": 1.2402, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.389697059009285e-06, |
|
"loss": 1.237, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.387591840171787e-06, |
|
"loss": 1.2141, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.385486621334288e-06, |
|
"loss": 1.2253, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.38338140249679e-06, |
|
"loss": 1.217, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.381276183659293e-06, |
|
"loss": 1.1919, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.379170964821794e-06, |
|
"loss": 1.1663, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.377065745984296e-06, |
|
"loss": 1.2338, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.374960527146797e-06, |
|
"loss": 1.2399, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.3728553083093e-06, |
|
"loss": 1.1608, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.370750089471802e-06, |
|
"loss": 1.1752, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.368644870634303e-06, |
|
"loss": 1.2364, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.366539651796806e-06, |
|
"loss": 1.2053, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.364434432959308e-06, |
|
"loss": 1.2431, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.362329214121809e-06, |
|
"loss": 1.1948, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.360223995284311e-06, |
|
"loss": 1.2248, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.358118776446812e-06, |
|
"loss": 1.2057, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.356013557609315e-06, |
|
"loss": 1.2373, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.353908338771817e-06, |
|
"loss": 1.1993, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.351803119934318e-06, |
|
"loss": 1.1474, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.34969790109682e-06, |
|
"loss": 1.2084, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.347592682259323e-06, |
|
"loss": 1.224, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.345487463421824e-06, |
|
"loss": 1.206, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.343382244584326e-06, |
|
"loss": 1.2225, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.341277025746827e-06, |
|
"loss": 1.2189, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.339171806909328e-06, |
|
"loss": 1.25, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.33706658807183e-06, |
|
"loss": 1.251, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.334961369234333e-06, |
|
"loss": 1.2048, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.332856150396834e-06, |
|
"loss": 1.2369, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.330750931559336e-06, |
|
"loss": 1.2427, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.328645712721837e-06, |
|
"loss": 1.2873, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.32654049388434e-06, |
|
"loss": 1.1579, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.324435275046842e-06, |
|
"loss": 1.2025, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.322330056209343e-06, |
|
"loss": 1.209, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.320224837371845e-06, |
|
"loss": 1.2015, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.318119618534348e-06, |
|
"loss": 1.2509, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.316014399696849e-06, |
|
"loss": 1.2696, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.313909180859351e-06, |
|
"loss": 1.2281, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.311803962021852e-06, |
|
"loss": 1.2089, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.309698743184354e-06, |
|
"loss": 1.2831, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.307593524346857e-06, |
|
"loss": 1.2757, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.305488305509358e-06, |
|
"loss": 1.2267, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.30338308667186e-06, |
|
"loss": 1.1837, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.301277867834361e-06, |
|
"loss": 1.2245, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.299172648996863e-06, |
|
"loss": 1.157, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.297067430159366e-06, |
|
"loss": 1.215, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.294962211321867e-06, |
|
"loss": 1.2421, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.29285699248437e-06, |
|
"loss": 1.2581, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.290751773646872e-06, |
|
"loss": 1.1966, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.288646554809373e-06, |
|
"loss": 1.2494, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.286541335971875e-06, |
|
"loss": 1.1633, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.284436117134376e-06, |
|
"loss": 1.2258, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.282330898296878e-06, |
|
"loss": 1.2703, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.280225679459381e-06, |
|
"loss": 1.1973, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.278120460621882e-06, |
|
"loss": 1.2614, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.276015241784384e-06, |
|
"loss": 1.243, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.273910022946887e-06, |
|
"loss": 1.2473, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.271804804109387e-06, |
|
"loss": 1.2269, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.26969958527189e-06, |
|
"loss": 1.2466, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.26759436643439e-06, |
|
"loss": 1.2362, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.265489147596893e-06, |
|
"loss": 1.2277, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.263383928759396e-06, |
|
"loss": 1.1939, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.261278709921897e-06, |
|
"loss": 1.2013, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.259173491084399e-06, |
|
"loss": 1.2057, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.257068272246902e-06, |
|
"loss": 1.2276, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.254963053409402e-06, |
|
"loss": 1.2029, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.252857834571905e-06, |
|
"loss": 1.2285, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.250752615734406e-06, |
|
"loss": 1.2078, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.248647396896908e-06, |
|
"loss": 1.2317, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.24654217805941e-06, |
|
"loss": 1.2266, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.244436959221911e-06, |
|
"loss": 1.212, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.242331740384414e-06, |
|
"loss": 1.1849, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.240226521546915e-06, |
|
"loss": 1.2238, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.238121302709417e-06, |
|
"loss": 1.221, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.23601608387192e-06, |
|
"loss": 1.2356, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.23391086503442e-06, |
|
"loss": 1.2287, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.231805646196923e-06, |
|
"loss": 1.2226, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.229700427359426e-06, |
|
"loss": 1.159, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.227595208521926e-06, |
|
"loss": 1.2239, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.225489989684429e-06, |
|
"loss": 1.2547, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.22338477084693e-06, |
|
"loss": 1.1689, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.221279552009432e-06, |
|
"loss": 1.1546, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.219174333171935e-06, |
|
"loss": 1.2197, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.217069114334436e-06, |
|
"loss": 1.1826, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.214963895496938e-06, |
|
"loss": 1.2543, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.21285867665944e-06, |
|
"loss": 1.1747, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.210753457821941e-06, |
|
"loss": 1.2486, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.208648238984444e-06, |
|
"loss": 1.2506, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.206543020146945e-06, |
|
"loss": 1.2257, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.204437801309447e-06, |
|
"loss": 1.183, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.20233258247195e-06, |
|
"loss": 1.2092, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.20022736363445e-06, |
|
"loss": 1.1907, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.198122144796953e-06, |
|
"loss": 1.1898, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.196016925959454e-06, |
|
"loss": 1.1834, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.193911707121956e-06, |
|
"loss": 1.2316, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.191806488284459e-06, |
|
"loss": 1.2018, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.18970126944696e-06, |
|
"loss": 1.2066, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.187596050609462e-06, |
|
"loss": 1.182, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.185490831771965e-06, |
|
"loss": 1.2089, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.183385612934465e-06, |
|
"loss": 1.1919, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.181280394096968e-06, |
|
"loss": 1.2734, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.179175175259469e-06, |
|
"loss": 1.25, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.177069956421971e-06, |
|
"loss": 1.213, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.174964737584474e-06, |
|
"loss": 1.2126, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.172859518746974e-06, |
|
"loss": 1.203, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.170754299909477e-06, |
|
"loss": 1.2718, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.16864908107198e-06, |
|
"loss": 1.1845, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.16654386223448e-06, |
|
"loss": 1.1991, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.164438643396983e-06, |
|
"loss": 1.2049, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.162333424559484e-06, |
|
"loss": 1.2345, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.160228205721986e-06, |
|
"loss": 1.2284, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.158122986884489e-06, |
|
"loss": 1.2253, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.15601776804699e-06, |
|
"loss": 1.2038, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.153912549209492e-06, |
|
"loss": 1.2353, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.151807330371993e-06, |
|
"loss": 1.2029, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.149702111534495e-06, |
|
"loss": 1.1451, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.147596892696998e-06, |
|
"loss": 1.186, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.145491673859498e-06, |
|
"loss": 1.2406, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.143386455022e-06, |
|
"loss": 1.1957, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.141281236184502e-06, |
|
"loss": 1.19, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.139176017347004e-06, |
|
"loss": 1.2007, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.137070798509505e-06, |
|
"loss": 1.2259, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.134965579672008e-06, |
|
"loss": 1.2204, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.132860360834508e-06, |
|
"loss": 1.2021, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.13075514199701e-06, |
|
"loss": 1.2208, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.128649923159513e-06, |
|
"loss": 1.1702, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.126544704322014e-06, |
|
"loss": 1.2513, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.124439485484517e-06, |
|
"loss": 1.1855, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.122334266647017e-06, |
|
"loss": 1.1868, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.12022904780952e-06, |
|
"loss": 1.1861, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.118123828972022e-06, |
|
"loss": 1.184, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.116018610134523e-06, |
|
"loss": 1.2227, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.113913391297026e-06, |
|
"loss": 1.1739, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.111808172459528e-06, |
|
"loss": 1.1705, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.109702953622029e-06, |
|
"loss": 1.1681, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.107597734784532e-06, |
|
"loss": 1.2116, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.105492515947032e-06, |
|
"loss": 1.2158, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.103387297109535e-06, |
|
"loss": 1.1452, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.101282078272037e-06, |
|
"loss": 1.1695, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.099176859434538e-06, |
|
"loss": 1.2099, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.09707164059704e-06, |
|
"loss": 1.2359, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.094966421759543e-06, |
|
"loss": 1.207, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.092861202922044e-06, |
|
"loss": 1.2119, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.090755984084546e-06, |
|
"loss": 1.1408, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.088650765247047e-06, |
|
"loss": 1.1842, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.08654554640955e-06, |
|
"loss": 1.1786, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.084440327572052e-06, |
|
"loss": 1.2016, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.082335108734553e-06, |
|
"loss": 1.2011, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.080229889897056e-06, |
|
"loss": 1.235, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.078124671059556e-06, |
|
"loss": 1.2116, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.076019452222059e-06, |
|
"loss": 1.2083, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.073914233384561e-06, |
|
"loss": 1.2152, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.071809014547062e-06, |
|
"loss": 1.239, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.069703795709565e-06, |
|
"loss": 1.1989, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.067598576872067e-06, |
|
"loss": 1.2193, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.065493358034568e-06, |
|
"loss": 1.2051, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.06338813919707e-06, |
|
"loss": 1.1981, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.061282920359571e-06, |
|
"loss": 1.2218, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.059177701522074e-06, |
|
"loss": 1.2669, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.057072482684576e-06, |
|
"loss": 1.2127, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.054967263847077e-06, |
|
"loss": 1.1538, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.05286204500958e-06, |
|
"loss": 1.1664, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.050756826172082e-06, |
|
"loss": 1.1976, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.048651607334583e-06, |
|
"loss": 1.2223, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.046546388497085e-06, |
|
"loss": 1.201, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.044441169659586e-06, |
|
"loss": 1.2372, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.042335950822089e-06, |
|
"loss": 1.1758, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.040230731984591e-06, |
|
"loss": 1.1793, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.038125513147092e-06, |
|
"loss": 1.1706, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.036020294309594e-06, |
|
"loss": 1.2267, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.033915075472097e-06, |
|
"loss": 1.218, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.031809856634598e-06, |
|
"loss": 1.1789, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.0297046377971e-06, |
|
"loss": 1.2317, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.027599418959601e-06, |
|
"loss": 1.23, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.025494200122104e-06, |
|
"loss": 1.2058, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.023388981284606e-06, |
|
"loss": 1.276, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.021283762447107e-06, |
|
"loss": 1.1758, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.01917854360961e-06, |
|
"loss": 1.182, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.01707332477211e-06, |
|
"loss": 1.2027, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.014968105934613e-06, |
|
"loss": 1.2442, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.012862887097115e-06, |
|
"loss": 1.1509, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.010757668259616e-06, |
|
"loss": 1.2369, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.008652449422118e-06, |
|
"loss": 1.225, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.006547230584621e-06, |
|
"loss": 1.2575, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.004442011747122e-06, |
|
"loss": 1.1801, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.002336792909624e-06, |
|
"loss": 1.1817, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.000231574072125e-06, |
|
"loss": 1.2392, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 8.998126355234628e-06, |
|
"loss": 1.1718, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.99602113639713e-06, |
|
"loss": 1.2155, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.993915917559631e-06, |
|
"loss": 1.1976, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.991810698722133e-06, |
|
"loss": 1.1715, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.989705479884636e-06, |
|
"loss": 1.1555, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.987600261047137e-06, |
|
"loss": 1.2071, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.98549504220964e-06, |
|
"loss": 1.2062, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.98338982337214e-06, |
|
"loss": 1.1978, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.981284604534642e-06, |
|
"loss": 1.2125, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.979179385697145e-06, |
|
"loss": 1.1887, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.977074166859646e-06, |
|
"loss": 1.2384, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.974968948022148e-06, |
|
"loss": 1.2708, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.972863729184649e-06, |
|
"loss": 1.1985, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.970758510347152e-06, |
|
"loss": 1.2202, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.968653291509654e-06, |
|
"loss": 1.2281, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.966548072672155e-06, |
|
"loss": 1.2158, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.964442853834657e-06, |
|
"loss": 1.1912, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.96233763499716e-06, |
|
"loss": 1.1822, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.96023241615966e-06, |
|
"loss": 1.2181, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.958127197322163e-06, |
|
"loss": 1.1901, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.956021978484664e-06, |
|
"loss": 1.1742, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.953916759647166e-06, |
|
"loss": 1.1675, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.951811540809669e-06, |
|
"loss": 1.249, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.94970632197217e-06, |
|
"loss": 1.1683, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.947601103134672e-06, |
|
"loss": 1.1558, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.945495884297175e-06, |
|
"loss": 1.1685, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.943390665459674e-06, |
|
"loss": 1.1734, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.941285446622176e-06, |
|
"loss": 1.1478, |
|
"step": 6000 |
|
} |
|
], |
|
"max_steps": 48471, |
|
"num_train_epochs": 3, |
|
"total_flos": 3.0400555051175117e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|