End of training
Browse files- all_results.json +10 -10
- eval_results.json +6 -6
- train_results.json +5 -5
- trainer_state.json +2588 -50
all_results.json
CHANGED
@@ -1,14 +1,14 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
-
"eval_loss": 0.
|
4 |
-
"eval_runtime":
|
5 |
"eval_samples": 7110,
|
6 |
-
"eval_samples_per_second":
|
7 |
-
"eval_steps_per_second": 1.
|
8 |
-
"eval_wer": 0.
|
9 |
-
"train_loss": 0.
|
10 |
-
"train_runtime":
|
11 |
"train_samples": 19531,
|
12 |
-
"train_samples_per_second":
|
13 |
-
"train_steps_per_second": 0.
|
14 |
}
|
|
|
1 |
{
|
2 |
+
"epoch": 150.0,
|
3 |
+
"eval_loss": 0.34484201669692993,
|
4 |
+
"eval_runtime": 317.3231,
|
5 |
"eval_samples": 7110,
|
6 |
+
"eval_samples_per_second": 22.406,
|
7 |
+
"eval_steps_per_second": 1.402,
|
8 |
+
"eval_wer": 0.347837027000871,
|
9 |
+
"train_loss": 0.16461168266179269,
|
10 |
+
"train_runtime": 24582.5742,
|
11 |
"train_samples": 19531,
|
12 |
+
"train_samples_per_second": 119.176,
|
13 |
+
"train_steps_per_second": 0.927
|
14 |
}
|
eval_results.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
-
"eval_loss": 0.
|
4 |
-
"eval_runtime":
|
5 |
"eval_samples": 7110,
|
6 |
-
"eval_samples_per_second":
|
7 |
-
"eval_steps_per_second": 1.
|
8 |
-
"eval_wer": 0.
|
9 |
}
|
|
|
1 |
{
|
2 |
+
"epoch": 150.0,
|
3 |
+
"eval_loss": 0.34484201669692993,
|
4 |
+
"eval_runtime": 317.3231,
|
5 |
"eval_samples": 7110,
|
6 |
+
"eval_samples_per_second": 22.406,
|
7 |
+
"eval_steps_per_second": 1.402,
|
8 |
+
"eval_wer": 0.347837027000871
|
9 |
}
|
train_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
-
"train_loss": 0.
|
4 |
-
"train_runtime":
|
5 |
"train_samples": 19531,
|
6 |
-
"train_samples_per_second":
|
7 |
-
"train_steps_per_second": 0.
|
8 |
}
|
|
|
1 |
{
|
2 |
+
"epoch": 150.0,
|
3 |
+
"train_loss": 0.16461168266179269,
|
4 |
+
"train_runtime": 24582.5742,
|
5 |
"train_samples": 19531,
|
6 |
+
"train_samples_per_second": 119.176,
|
7 |
+
"train_steps_per_second": 0.927
|
8 |
}
|
trainer_state.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -11442,131 +11442,2669 @@
|
|
11442 |
},
|
11443 |
{
|
11444 |
"epoch": 121.77,
|
11445 |
-
"learning_rate":
|
11446 |
-
"loss": 0.
|
11447 |
"step": 18510
|
11448 |
},
|
11449 |
{
|
11450 |
"epoch": 121.84,
|
11451 |
-
"learning_rate":
|
11452 |
-
"loss": 0.
|
11453 |
"step": 18520
|
11454 |
},
|
11455 |
{
|
11456 |
"epoch": 121.9,
|
11457 |
-
"learning_rate":
|
11458 |
-
"loss": 0.
|
11459 |
"step": 18530
|
11460 |
},
|
11461 |
{
|
11462 |
"epoch": 121.97,
|
11463 |
-
"learning_rate":
|
11464 |
-
"loss": 0.
|
11465 |
"step": 18540
|
11466 |
},
|
11467 |
{
|
11468 |
"epoch": 122.04,
|
11469 |
-
"learning_rate":
|
11470 |
-
"loss": 0.
|
11471 |
"step": 18550
|
11472 |
},
|
11473 |
{
|
11474 |
"epoch": 122.1,
|
11475 |
-
"learning_rate":
|
11476 |
-
"loss": 0.
|
11477 |
"step": 18560
|
11478 |
},
|
11479 |
{
|
11480 |
"epoch": 122.17,
|
11481 |
-
"learning_rate":
|
11482 |
-
"loss": 0.
|
11483 |
"step": 18570
|
11484 |
},
|
11485 |
{
|
11486 |
"epoch": 122.24,
|
11487 |
-
"learning_rate":
|
11488 |
-
"loss": 0.
|
11489 |
"step": 18580
|
11490 |
},
|
11491 |
{
|
11492 |
"epoch": 122.3,
|
11493 |
-
"learning_rate":
|
11494 |
-
"loss": 0.
|
11495 |
"step": 18590
|
11496 |
},
|
11497 |
{
|
11498 |
"epoch": 122.37,
|
11499 |
-
"learning_rate":
|
11500 |
-
"loss": 0.
|
11501 |
"step": 18600
|
11502 |
},
|
11503 |
{
|
11504 |
"epoch": 122.43,
|
11505 |
-
"learning_rate":
|
11506 |
-
"loss": 0.
|
11507 |
"step": 18610
|
11508 |
},
|
11509 |
{
|
11510 |
"epoch": 122.5,
|
11511 |
-
"learning_rate":
|
11512 |
-
"loss": 0.
|
11513 |
"step": 18620
|
11514 |
},
|
11515 |
{
|
11516 |
"epoch": 122.56,
|
11517 |
-
"learning_rate":
|
11518 |
-
"loss": 0.
|
11519 |
"step": 18630
|
11520 |
},
|
11521 |
{
|
11522 |
"epoch": 122.63,
|
11523 |
-
"learning_rate":
|
11524 |
-
"loss": 0.
|
11525 |
"step": 18640
|
11526 |
},
|
11527 |
{
|
11528 |
"epoch": 122.69,
|
11529 |
-
"learning_rate": 1.
|
11530 |
-
"loss": 0.
|
11531 |
"step": 18650
|
11532 |
},
|
11533 |
{
|
11534 |
"epoch": 122.76,
|
11535 |
-
"learning_rate": 1.
|
11536 |
-
"loss": 0.
|
11537 |
"step": 18660
|
11538 |
},
|
11539 |
{
|
11540 |
"epoch": 122.82,
|
11541 |
-
"learning_rate": 1.
|
11542 |
-
"loss": 0.
|
11543 |
"step": 18670
|
11544 |
},
|
11545 |
{
|
11546 |
"epoch": 122.89,
|
11547 |
-
"learning_rate":
|
11548 |
-
"loss": 0.
|
11549 |
"step": 18680
|
11550 |
},
|
11551 |
{
|
11552 |
"epoch": 122.96,
|
11553 |
-
"learning_rate":
|
11554 |
-
"loss": 0.
|
11555 |
"step": 18690
|
11556 |
},
|
11557 |
{
|
11558 |
-
"epoch": 123.
|
11559 |
-
"
|
11560 |
-
"
|
11561 |
-
"
|
11562 |
-
|
11563 |
-
|
11564 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11565 |
}
|
11566 |
],
|
11567 |
-
"max_steps":
|
11568 |
-
"num_train_epochs":
|
11569 |
-
"total_flos":
|
11570 |
"trial_name": null,
|
11571 |
"trial_params": null
|
11572 |
}
|
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
+
"epoch": 149.99509001636662,
|
5 |
+
"global_step": 22800,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
11442 |
},
|
11443 |
{
|
11444 |
"epoch": 121.77,
|
11445 |
+
"learning_rate": 1.1362995594713658e-05,
|
11446 |
+
"loss": 0.9705,
|
11447 |
"step": 18510
|
11448 |
},
|
11449 |
{
|
11450 |
"epoch": 121.84,
|
11451 |
+
"learning_rate": 1.1336563876651981e-05,
|
11452 |
+
"loss": 0.8461,
|
11453 |
"step": 18520
|
11454 |
},
|
11455 |
{
|
11456 |
"epoch": 121.9,
|
11457 |
+
"learning_rate": 1.1310132158590309e-05,
|
11458 |
+
"loss": 0.8587,
|
11459 |
"step": 18530
|
11460 |
},
|
11461 |
{
|
11462 |
"epoch": 121.97,
|
11463 |
+
"learning_rate": 1.1283700440528634e-05,
|
11464 |
+
"loss": 0.8934,
|
11465 |
"step": 18540
|
11466 |
},
|
11467 |
{
|
11468 |
"epoch": 122.04,
|
11469 |
+
"learning_rate": 1.1257268722466961e-05,
|
11470 |
+
"loss": 0.9975,
|
11471 |
"step": 18550
|
11472 |
},
|
11473 |
{
|
11474 |
"epoch": 122.1,
|
11475 |
+
"learning_rate": 1.1230837004405287e-05,
|
11476 |
+
"loss": 0.8517,
|
11477 |
"step": 18560
|
11478 |
},
|
11479 |
{
|
11480 |
"epoch": 122.17,
|
11481 |
+
"learning_rate": 1.1204405286343612e-05,
|
11482 |
+
"loss": 0.8799,
|
11483 |
"step": 18570
|
11484 |
},
|
11485 |
{
|
11486 |
"epoch": 122.24,
|
11487 |
+
"learning_rate": 1.117797356828194e-05,
|
11488 |
+
"loss": 0.9021,
|
11489 |
"step": 18580
|
11490 |
},
|
11491 |
{
|
11492 |
"epoch": 122.3,
|
11493 |
+
"learning_rate": 1.1151541850220265e-05,
|
11494 |
+
"loss": 0.8569,
|
11495 |
"step": 18590
|
11496 |
},
|
11497 |
{
|
11498 |
"epoch": 122.37,
|
11499 |
+
"learning_rate": 1.112511013215859e-05,
|
11500 |
+
"loss": 0.8779,
|
11501 |
"step": 18600
|
11502 |
},
|
11503 |
{
|
11504 |
"epoch": 122.43,
|
11505 |
+
"learning_rate": 1.1098678414096916e-05,
|
11506 |
+
"loss": 0.88,
|
11507 |
"step": 18610
|
11508 |
},
|
11509 |
{
|
11510 |
"epoch": 122.5,
|
11511 |
+
"learning_rate": 1.1072246696035243e-05,
|
11512 |
+
"loss": 0.9219,
|
11513 |
"step": 18620
|
11514 |
},
|
11515 |
{
|
11516 |
"epoch": 122.56,
|
11517 |
+
"learning_rate": 1.1045814977973568e-05,
|
11518 |
+
"loss": 0.8598,
|
11519 |
"step": 18630
|
11520 |
},
|
11521 |
{
|
11522 |
"epoch": 122.63,
|
11523 |
+
"learning_rate": 1.1019383259911894e-05,
|
11524 |
+
"loss": 0.8648,
|
11525 |
"step": 18640
|
11526 |
},
|
11527 |
{
|
11528 |
"epoch": 122.69,
|
11529 |
+
"learning_rate": 1.0992951541850221e-05,
|
11530 |
+
"loss": 0.8566,
|
11531 |
"step": 18650
|
11532 |
},
|
11533 |
{
|
11534 |
"epoch": 122.76,
|
11535 |
+
"learning_rate": 1.0966519823788547e-05,
|
11536 |
+
"loss": 0.9491,
|
11537 |
"step": 18660
|
11538 |
},
|
11539 |
{
|
11540 |
"epoch": 122.82,
|
11541 |
+
"learning_rate": 1.0940088105726872e-05,
|
11542 |
+
"loss": 0.8215,
|
11543 |
"step": 18670
|
11544 |
},
|
11545 |
{
|
11546 |
"epoch": 122.89,
|
11547 |
+
"learning_rate": 1.09136563876652e-05,
|
11548 |
+
"loss": 0.8726,
|
11549 |
"step": 18680
|
11550 |
},
|
11551 |
{
|
11552 |
"epoch": 122.96,
|
11553 |
+
"learning_rate": 1.0887224669603525e-05,
|
11554 |
+
"loss": 0.8909,
|
11555 |
"step": 18690
|
11556 |
},
|
11557 |
{
|
11558 |
+
"epoch": 123.03,
|
11559 |
+
"learning_rate": 1.086079295154185e-05,
|
11560 |
+
"loss": 0.9692,
|
11561 |
+
"step": 18700
|
11562 |
+
},
|
11563 |
+
{
|
11564 |
+
"epoch": 123.09,
|
11565 |
+
"learning_rate": 1.0834361233480176e-05,
|
11566 |
+
"loss": 0.85,
|
11567 |
+
"step": 18710
|
11568 |
+
},
|
11569 |
+
{
|
11570 |
+
"epoch": 123.16,
|
11571 |
+
"learning_rate": 1.0807929515418503e-05,
|
11572 |
+
"loss": 0.8889,
|
11573 |
+
"step": 18720
|
11574 |
+
},
|
11575 |
+
{
|
11576 |
+
"epoch": 123.22,
|
11577 |
+
"learning_rate": 1.0781497797356828e-05,
|
11578 |
+
"loss": 0.8586,
|
11579 |
+
"step": 18730
|
11580 |
+
},
|
11581 |
+
{
|
11582 |
+
"epoch": 123.29,
|
11583 |
+
"learning_rate": 1.0755066079295154e-05,
|
11584 |
+
"loss": 0.9315,
|
11585 |
+
"step": 18740
|
11586 |
+
},
|
11587 |
+
{
|
11588 |
+
"epoch": 123.35,
|
11589 |
+
"learning_rate": 1.0728634361233481e-05,
|
11590 |
+
"loss": 0.8433,
|
11591 |
+
"step": 18750
|
11592 |
+
},
|
11593 |
+
{
|
11594 |
+
"epoch": 123.42,
|
11595 |
+
"learning_rate": 1.0702202643171806e-05,
|
11596 |
+
"loss": 0.8922,
|
11597 |
+
"step": 18760
|
11598 |
+
},
|
11599 |
+
{
|
11600 |
+
"epoch": 123.48,
|
11601 |
+
"learning_rate": 1.0675770925110134e-05,
|
11602 |
+
"loss": 0.9134,
|
11603 |
+
"step": 18770
|
11604 |
+
},
|
11605 |
+
{
|
11606 |
+
"epoch": 123.55,
|
11607 |
+
"learning_rate": 1.0649339207048457e-05,
|
11608 |
+
"loss": 0.8996,
|
11609 |
+
"step": 18780
|
11610 |
+
},
|
11611 |
+
{
|
11612 |
+
"epoch": 123.62,
|
11613 |
+
"learning_rate": 1.0622907488986785e-05,
|
11614 |
+
"loss": 0.8622,
|
11615 |
+
"step": 18790
|
11616 |
+
},
|
11617 |
+
{
|
11618 |
+
"epoch": 123.68,
|
11619 |
+
"learning_rate": 1.0596475770925112e-05,
|
11620 |
+
"loss": 0.8691,
|
11621 |
+
"step": 18800
|
11622 |
+
},
|
11623 |
+
{
|
11624 |
+
"epoch": 123.75,
|
11625 |
+
"learning_rate": 1.0570044052863435e-05,
|
11626 |
+
"loss": 0.9627,
|
11627 |
+
"step": 18810
|
11628 |
+
},
|
11629 |
+
{
|
11630 |
+
"epoch": 123.81,
|
11631 |
+
"learning_rate": 1.0543612334801763e-05,
|
11632 |
+
"loss": 0.8812,
|
11633 |
+
"step": 18820
|
11634 |
+
},
|
11635 |
+
{
|
11636 |
+
"epoch": 123.88,
|
11637 |
+
"learning_rate": 1.0517180616740088e-05,
|
11638 |
+
"loss": 0.8543,
|
11639 |
+
"step": 18830
|
11640 |
+
},
|
11641 |
+
{
|
11642 |
+
"epoch": 123.94,
|
11643 |
+
"learning_rate": 1.0490748898678415e-05,
|
11644 |
+
"loss": 0.8661,
|
11645 |
+
"step": 18840
|
11646 |
+
},
|
11647 |
+
{
|
11648 |
+
"epoch": 124.01,
|
11649 |
+
"learning_rate": 1.0464317180616739e-05,
|
11650 |
+
"loss": 1.0126,
|
11651 |
+
"step": 18850
|
11652 |
+
},
|
11653 |
+
{
|
11654 |
+
"epoch": 124.08,
|
11655 |
+
"learning_rate": 1.0437885462555066e-05,
|
11656 |
+
"loss": 0.8392,
|
11657 |
+
"step": 18860
|
11658 |
+
},
|
11659 |
+
{
|
11660 |
+
"epoch": 124.14,
|
11661 |
+
"learning_rate": 1.0411453744493393e-05,
|
11662 |
+
"loss": 0.8753,
|
11663 |
+
"step": 18870
|
11664 |
+
},
|
11665 |
+
{
|
11666 |
+
"epoch": 124.21,
|
11667 |
+
"learning_rate": 1.0385022026431717e-05,
|
11668 |
+
"loss": 0.8561,
|
11669 |
+
"step": 18880
|
11670 |
+
},
|
11671 |
+
{
|
11672 |
+
"epoch": 124.27,
|
11673 |
+
"learning_rate": 1.0358590308370044e-05,
|
11674 |
+
"loss": 0.9043,
|
11675 |
+
"step": 18890
|
11676 |
+
},
|
11677 |
+
{
|
11678 |
+
"epoch": 124.34,
|
11679 |
+
"learning_rate": 1.033215859030837e-05,
|
11680 |
+
"loss": 0.8524,
|
11681 |
+
"step": 18900
|
11682 |
+
},
|
11683 |
+
{
|
11684 |
+
"epoch": 124.41,
|
11685 |
+
"learning_rate": 1.0305726872246697e-05,
|
11686 |
+
"loss": 0.8642,
|
11687 |
+
"step": 18910
|
11688 |
+
},
|
11689 |
+
{
|
11690 |
+
"epoch": 124.47,
|
11691 |
+
"learning_rate": 1.0279295154185022e-05,
|
11692 |
+
"loss": 0.8889,
|
11693 |
+
"step": 18920
|
11694 |
+
},
|
11695 |
+
{
|
11696 |
+
"epoch": 124.54,
|
11697 |
+
"learning_rate": 1.0252863436123348e-05,
|
11698 |
+
"loss": 0.9359,
|
11699 |
+
"step": 18930
|
11700 |
+
},
|
11701 |
+
{
|
11702 |
+
"epoch": 124.6,
|
11703 |
+
"learning_rate": 1.0226431718061675e-05,
|
11704 |
+
"loss": 0.8471,
|
11705 |
+
"step": 18940
|
11706 |
+
},
|
11707 |
+
{
|
11708 |
+
"epoch": 124.67,
|
11709 |
+
"learning_rate": 1.02e-05,
|
11710 |
+
"loss": 0.859,
|
11711 |
+
"step": 18950
|
11712 |
+
},
|
11713 |
+
{
|
11714 |
+
"epoch": 124.73,
|
11715 |
+
"learning_rate": 1.0173568281938326e-05,
|
11716 |
+
"loss": 0.8839,
|
11717 |
+
"step": 18960
|
11718 |
+
},
|
11719 |
+
{
|
11720 |
+
"epoch": 124.8,
|
11721 |
+
"learning_rate": 1.0147136563876652e-05,
|
11722 |
+
"loss": 0.8672,
|
11723 |
+
"step": 18970
|
11724 |
+
},
|
11725 |
+
{
|
11726 |
+
"epoch": 124.86,
|
11727 |
+
"learning_rate": 1.0120704845814979e-05,
|
11728 |
+
"loss": 0.8499,
|
11729 |
+
"step": 18980
|
11730 |
+
},
|
11731 |
+
{
|
11732 |
+
"epoch": 124.93,
|
11733 |
+
"learning_rate": 1.0094273127753304e-05,
|
11734 |
+
"loss": 0.9104,
|
11735 |
+
"step": 18990
|
11736 |
+
},
|
11737 |
+
{
|
11738 |
+
"epoch": 125.0,
|
11739 |
+
"learning_rate": 1.006784140969163e-05,
|
11740 |
+
"loss": 0.9453,
|
11741 |
+
"step": 19000
|
11742 |
+
},
|
11743 |
+
{
|
11744 |
+
"epoch": 125.0,
|
11745 |
+
"eval_loss": 0.3414785861968994,
|
11746 |
+
"eval_runtime": 320.2801,
|
11747 |
+
"eval_samples_per_second": 22.199,
|
11748 |
+
"eval_steps_per_second": 1.389,
|
11749 |
+
"eval_wer": 0.34803058163166556,
|
11750 |
+
"step": 19000
|
11751 |
+
},
|
11752 |
+
{
|
11753 |
+
"epoch": 125.07,
|
11754 |
+
"learning_rate": 1.0041409691629957e-05,
|
11755 |
+
"loss": 0.922,
|
11756 |
+
"step": 19010
|
11757 |
+
},
|
11758 |
+
{
|
11759 |
+
"epoch": 125.13,
|
11760 |
+
"learning_rate": 1.0014977973568282e-05,
|
11761 |
+
"loss": 0.8491,
|
11762 |
+
"step": 19020
|
11763 |
+
},
|
11764 |
+
{
|
11765 |
+
"epoch": 125.2,
|
11766 |
+
"learning_rate": 9.988546255506608e-06,
|
11767 |
+
"loss": 0.8899,
|
11768 |
+
"step": 19030
|
11769 |
+
},
|
11770 |
+
{
|
11771 |
+
"epoch": 125.26,
|
11772 |
+
"learning_rate": 9.962114537444933e-06,
|
11773 |
+
"loss": 0.9163,
|
11774 |
+
"step": 19040
|
11775 |
+
},
|
11776 |
+
{
|
11777 |
+
"epoch": 125.33,
|
11778 |
+
"learning_rate": 9.93568281938326e-06,
|
11779 |
+
"loss": 0.853,
|
11780 |
+
"step": 19050
|
11781 |
+
},
|
11782 |
+
{
|
11783 |
+
"epoch": 125.39,
|
11784 |
+
"learning_rate": 9.909251101321588e-06,
|
11785 |
+
"loss": 0.866,
|
11786 |
+
"step": 19060
|
11787 |
+
},
|
11788 |
+
{
|
11789 |
+
"epoch": 125.46,
|
11790 |
+
"learning_rate": 9.882819383259911e-06,
|
11791 |
+
"loss": 0.8817,
|
11792 |
+
"step": 19070
|
11793 |
+
},
|
11794 |
+
{
|
11795 |
+
"epoch": 125.52,
|
11796 |
+
"learning_rate": 9.856387665198239e-06,
|
11797 |
+
"loss": 0.9513,
|
11798 |
+
"step": 19080
|
11799 |
+
},
|
11800 |
+
{
|
11801 |
+
"epoch": 125.59,
|
11802 |
+
"learning_rate": 9.829955947136564e-06,
|
11803 |
+
"loss": 0.8217,
|
11804 |
+
"step": 19090
|
11805 |
+
},
|
11806 |
+
{
|
11807 |
+
"epoch": 125.65,
|
11808 |
+
"learning_rate": 9.80352422907489e-06,
|
11809 |
+
"loss": 0.8926,
|
11810 |
+
"step": 19100
|
11811 |
+
},
|
11812 |
+
{
|
11813 |
+
"epoch": 125.72,
|
11814 |
+
"learning_rate": 9.777092511013217e-06,
|
11815 |
+
"loss": 0.8642,
|
11816 |
+
"step": 19110
|
11817 |
+
},
|
11818 |
+
{
|
11819 |
+
"epoch": 125.79,
|
11820 |
+
"learning_rate": 9.750660792951542e-06,
|
11821 |
+
"loss": 0.9165,
|
11822 |
+
"step": 19120
|
11823 |
+
},
|
11824 |
+
{
|
11825 |
+
"epoch": 125.85,
|
11826 |
+
"learning_rate": 9.72422907488987e-06,
|
11827 |
+
"loss": 0.8646,
|
11828 |
+
"step": 19130
|
11829 |
+
},
|
11830 |
+
{
|
11831 |
+
"epoch": 125.92,
|
11832 |
+
"learning_rate": 9.697797356828193e-06,
|
11833 |
+
"loss": 0.8568,
|
11834 |
+
"step": 19140
|
11835 |
+
},
|
11836 |
+
{
|
11837 |
+
"epoch": 125.98,
|
11838 |
+
"learning_rate": 9.67136563876652e-06,
|
11839 |
+
"loss": 0.8925,
|
11840 |
+
"step": 19150
|
11841 |
+
},
|
11842 |
+
{
|
11843 |
+
"epoch": 126.05,
|
11844 |
+
"learning_rate": 9.644933920704846e-06,
|
11845 |
+
"loss": 0.9546,
|
11846 |
+
"step": 19160
|
11847 |
+
},
|
11848 |
+
{
|
11849 |
+
"epoch": 126.12,
|
11850 |
+
"learning_rate": 9.618502202643171e-06,
|
11851 |
+
"loss": 0.8381,
|
11852 |
+
"step": 19170
|
11853 |
+
},
|
11854 |
+
{
|
11855 |
+
"epoch": 126.18,
|
11856 |
+
"learning_rate": 9.592070484581498e-06,
|
11857 |
+
"loss": 0.8962,
|
11858 |
+
"step": 19180
|
11859 |
+
},
|
11860 |
+
{
|
11861 |
+
"epoch": 126.25,
|
11862 |
+
"learning_rate": 9.565638766519824e-06,
|
11863 |
+
"loss": 0.9426,
|
11864 |
+
"step": 19190
|
11865 |
+
},
|
11866 |
+
{
|
11867 |
+
"epoch": 126.31,
|
11868 |
+
"learning_rate": 9.539207048458151e-06,
|
11869 |
+
"loss": 0.8578,
|
11870 |
+
"step": 19200
|
11871 |
+
},
|
11872 |
+
{
|
11873 |
+
"epoch": 126.38,
|
11874 |
+
"learning_rate": 9.512775330396475e-06,
|
11875 |
+
"loss": 0.8628,
|
11876 |
+
"step": 19210
|
11877 |
+
},
|
11878 |
+
{
|
11879 |
+
"epoch": 126.45,
|
11880 |
+
"learning_rate": 9.486343612334802e-06,
|
11881 |
+
"loss": 0.8734,
|
11882 |
+
"step": 19220
|
11883 |
+
},
|
11884 |
+
{
|
11885 |
+
"epoch": 126.51,
|
11886 |
+
"learning_rate": 9.459911894273127e-06,
|
11887 |
+
"loss": 0.9106,
|
11888 |
+
"step": 19230
|
11889 |
+
},
|
11890 |
+
{
|
11891 |
+
"epoch": 126.58,
|
11892 |
+
"learning_rate": 9.433480176211455e-06,
|
11893 |
+
"loss": 0.8388,
|
11894 |
+
"step": 19240
|
11895 |
+
},
|
11896 |
+
{
|
11897 |
+
"epoch": 126.64,
|
11898 |
+
"learning_rate": 9.40704845814978e-06,
|
11899 |
+
"loss": 0.8349,
|
11900 |
+
"step": 19250
|
11901 |
+
},
|
11902 |
+
{
|
11903 |
+
"epoch": 126.71,
|
11904 |
+
"learning_rate": 9.380616740088106e-06,
|
11905 |
+
"loss": 0.8748,
|
11906 |
+
"step": 19260
|
11907 |
+
},
|
11908 |
+
{
|
11909 |
+
"epoch": 126.77,
|
11910 |
+
"learning_rate": 9.354185022026433e-06,
|
11911 |
+
"loss": 0.9127,
|
11912 |
+
"step": 19270
|
11913 |
+
},
|
11914 |
+
{
|
11915 |
+
"epoch": 126.84,
|
11916 |
+
"learning_rate": 9.327753303964757e-06,
|
11917 |
+
"loss": 0.8584,
|
11918 |
+
"step": 19280
|
11919 |
+
},
|
11920 |
+
{
|
11921 |
+
"epoch": 126.9,
|
11922 |
+
"learning_rate": 9.301321585903084e-06,
|
11923 |
+
"loss": 0.8903,
|
11924 |
+
"step": 19290
|
11925 |
+
},
|
11926 |
+
{
|
11927 |
+
"epoch": 126.97,
|
11928 |
+
"learning_rate": 9.274889867841411e-06,
|
11929 |
+
"loss": 0.8449,
|
11930 |
+
"step": 19300
|
11931 |
+
},
|
11932 |
+
{
|
11933 |
+
"epoch": 127.04,
|
11934 |
+
"learning_rate": 9.248458149779736e-06,
|
11935 |
+
"loss": 0.9825,
|
11936 |
+
"step": 19310
|
11937 |
+
},
|
11938 |
+
{
|
11939 |
+
"epoch": 127.1,
|
11940 |
+
"learning_rate": 9.222026431718062e-06,
|
11941 |
+
"loss": 0.8495,
|
11942 |
+
"step": 19320
|
11943 |
+
},
|
11944 |
+
{
|
11945 |
+
"epoch": 127.17,
|
11946 |
+
"learning_rate": 9.195594713656387e-06,
|
11947 |
+
"loss": 0.8666,
|
11948 |
+
"step": 19330
|
11949 |
+
},
|
11950 |
+
{
|
11951 |
+
"epoch": 127.24,
|
11952 |
+
"learning_rate": 9.169162995594714e-06,
|
11953 |
+
"loss": 0.8931,
|
11954 |
+
"step": 19340
|
11955 |
+
},
|
11956 |
+
{
|
11957 |
+
"epoch": 127.3,
|
11958 |
+
"learning_rate": 9.142731277533038e-06,
|
11959 |
+
"loss": 0.8946,
|
11960 |
+
"step": 19350
|
11961 |
+
},
|
11962 |
+
{
|
11963 |
+
"epoch": 127.37,
|
11964 |
+
"learning_rate": 9.116299559471365e-06,
|
11965 |
+
"loss": 0.8899,
|
11966 |
+
"step": 19360
|
11967 |
+
},
|
11968 |
+
{
|
11969 |
+
"epoch": 127.43,
|
11970 |
+
"learning_rate": 9.089867841409693e-06,
|
11971 |
+
"loss": 0.8799,
|
11972 |
+
"step": 19370
|
11973 |
+
},
|
11974 |
+
{
|
11975 |
+
"epoch": 127.5,
|
11976 |
+
"learning_rate": 9.063436123348018e-06,
|
11977 |
+
"loss": 0.9319,
|
11978 |
+
"step": 19380
|
11979 |
+
},
|
11980 |
+
{
|
11981 |
+
"epoch": 127.56,
|
11982 |
+
"learning_rate": 9.037004405286344e-06,
|
11983 |
+
"loss": 0.8407,
|
11984 |
+
"step": 19390
|
11985 |
+
},
|
11986 |
+
{
|
11987 |
+
"epoch": 127.63,
|
11988 |
+
"learning_rate": 9.010572687224669e-06,
|
11989 |
+
"loss": 0.8338,
|
11990 |
+
"step": 19400
|
11991 |
+
},
|
11992 |
+
{
|
11993 |
+
"epoch": 127.69,
|
11994 |
+
"learning_rate": 8.984140969162996e-06,
|
11995 |
+
"loss": 0.8834,
|
11996 |
+
"step": 19410
|
11997 |
+
},
|
11998 |
+
{
|
11999 |
+
"epoch": 127.76,
|
12000 |
+
"learning_rate": 8.957709251101323e-06,
|
12001 |
+
"loss": 0.9279,
|
12002 |
+
"step": 19420
|
12003 |
+
},
|
12004 |
+
{
|
12005 |
+
"epoch": 127.82,
|
12006 |
+
"learning_rate": 8.931277533039647e-06,
|
12007 |
+
"loss": 0.8409,
|
12008 |
+
"step": 19430
|
12009 |
+
},
|
12010 |
+
{
|
12011 |
+
"epoch": 127.89,
|
12012 |
+
"learning_rate": 8.904845814977974e-06,
|
12013 |
+
"loss": 0.871,
|
12014 |
+
"step": 19440
|
12015 |
+
},
|
12016 |
+
{
|
12017 |
+
"epoch": 127.96,
|
12018 |
+
"learning_rate": 8.8784140969163e-06,
|
12019 |
+
"loss": 0.8758,
|
12020 |
+
"step": 19450
|
12021 |
+
},
|
12022 |
+
{
|
12023 |
+
"epoch": 128.03,
|
12024 |
+
"learning_rate": 8.851982378854625e-06,
|
12025 |
+
"loss": 0.9881,
|
12026 |
+
"step": 19460
|
12027 |
+
},
|
12028 |
+
{
|
12029 |
+
"epoch": 128.09,
|
12030 |
+
"learning_rate": 8.82555066079295e-06,
|
12031 |
+
"loss": 0.8283,
|
12032 |
+
"step": 19470
|
12033 |
+
},
|
12034 |
+
{
|
12035 |
+
"epoch": 128.16,
|
12036 |
+
"learning_rate": 8.799118942731278e-06,
|
12037 |
+
"loss": 0.8865,
|
12038 |
+
"step": 19480
|
12039 |
+
},
|
12040 |
+
{
|
12041 |
+
"epoch": 128.22,
|
12042 |
+
"learning_rate": 8.772687224669605e-06,
|
12043 |
+
"loss": 0.8691,
|
12044 |
+
"step": 19490
|
12045 |
+
},
|
12046 |
+
{
|
12047 |
+
"epoch": 128.29,
|
12048 |
+
"learning_rate": 8.746255506607929e-06,
|
12049 |
+
"loss": 0.9267,
|
12050 |
+
"step": 19500
|
12051 |
+
},
|
12052 |
+
{
|
12053 |
+
"epoch": 128.29,
|
12054 |
+
"eval_loss": 0.3477088212966919,
|
12055 |
+
"eval_runtime": 318.7167,
|
12056 |
+
"eval_samples_per_second": 22.308,
|
12057 |
+
"eval_steps_per_second": 1.396,
|
12058 |
+
"eval_wer": 0.35029517081196165,
|
12059 |
+
"step": 19500
|
12060 |
+
},
|
12061 |
+
{
|
12062 |
+
"epoch": 128.35,
|
12063 |
+
"learning_rate": 8.719823788546256e-06,
|
12064 |
+
"loss": 0.8531,
|
12065 |
+
"step": 19510
|
12066 |
+
},
|
12067 |
+
{
|
12068 |
+
"epoch": 128.42,
|
12069 |
+
"learning_rate": 8.693392070484582e-06,
|
12070 |
+
"loss": 0.8753,
|
12071 |
+
"step": 19520
|
12072 |
+
},
|
12073 |
+
{
|
12074 |
+
"epoch": 128.48,
|
12075 |
+
"learning_rate": 8.666960352422909e-06,
|
12076 |
+
"loss": 0.8862,
|
12077 |
+
"step": 19530
|
12078 |
+
},
|
12079 |
+
{
|
12080 |
+
"epoch": 128.55,
|
12081 |
+
"learning_rate": 8.640528634361234e-06,
|
12082 |
+
"loss": 0.8788,
|
12083 |
+
"step": 19540
|
12084 |
+
},
|
12085 |
+
{
|
12086 |
+
"epoch": 128.62,
|
12087 |
+
"learning_rate": 8.61409691629956e-06,
|
12088 |
+
"loss": 0.8425,
|
12089 |
+
"step": 19550
|
12090 |
+
},
|
12091 |
+
{
|
12092 |
+
"epoch": 128.68,
|
12093 |
+
"learning_rate": 8.587665198237887e-06,
|
12094 |
+
"loss": 0.8931,
|
12095 |
+
"step": 19560
|
12096 |
+
},
|
12097 |
+
{
|
12098 |
+
"epoch": 128.75,
|
12099 |
+
"learning_rate": 8.56123348017621e-06,
|
12100 |
+
"loss": 0.8952,
|
12101 |
+
"step": 19570
|
12102 |
+
},
|
12103 |
+
{
|
12104 |
+
"epoch": 128.81,
|
12105 |
+
"learning_rate": 8.534801762114538e-06,
|
12106 |
+
"loss": 0.8496,
|
12107 |
+
"step": 19580
|
12108 |
+
},
|
12109 |
+
{
|
12110 |
+
"epoch": 128.88,
|
12111 |
+
"learning_rate": 8.508370044052863e-06,
|
12112 |
+
"loss": 0.8719,
|
12113 |
+
"step": 19590
|
12114 |
+
},
|
12115 |
+
{
|
12116 |
+
"epoch": 128.94,
|
12117 |
+
"learning_rate": 8.48193832599119e-06,
|
12118 |
+
"loss": 0.8714,
|
12119 |
+
"step": 19600
|
12120 |
+
},
|
12121 |
+
{
|
12122 |
+
"epoch": 129.01,
|
12123 |
+
"learning_rate": 8.455506607929516e-06,
|
12124 |
+
"loss": 0.9905,
|
12125 |
+
"step": 19610
|
12126 |
+
},
|
12127 |
+
{
|
12128 |
+
"epoch": 129.08,
|
12129 |
+
"learning_rate": 8.429074889867841e-06,
|
12130 |
+
"loss": 0.8359,
|
12131 |
+
"step": 19620
|
12132 |
+
},
|
12133 |
+
{
|
12134 |
+
"epoch": 129.14,
|
12135 |
+
"learning_rate": 8.402643171806169e-06,
|
12136 |
+
"loss": 0.8713,
|
12137 |
+
"step": 19630
|
12138 |
+
},
|
12139 |
+
{
|
12140 |
+
"epoch": 129.21,
|
12141 |
+
"learning_rate": 8.376211453744492e-06,
|
12142 |
+
"loss": 0.8612,
|
12143 |
+
"step": 19640
|
12144 |
+
},
|
12145 |
+
{
|
12146 |
+
"epoch": 129.27,
|
12147 |
+
"learning_rate": 8.34977973568282e-06,
|
12148 |
+
"loss": 0.9565,
|
12149 |
+
"step": 19650
|
12150 |
+
},
|
12151 |
+
{
|
12152 |
+
"epoch": 129.34,
|
12153 |
+
"learning_rate": 8.323348017621145e-06,
|
12154 |
+
"loss": 0.8521,
|
12155 |
+
"step": 19660
|
12156 |
+
},
|
12157 |
+
{
|
12158 |
+
"epoch": 129.41,
|
12159 |
+
"learning_rate": 8.296916299559472e-06,
|
12160 |
+
"loss": 0.8548,
|
12161 |
+
"step": 19670
|
12162 |
+
},
|
12163 |
+
{
|
12164 |
+
"epoch": 129.47,
|
12165 |
+
"learning_rate": 8.270484581497798e-06,
|
12166 |
+
"loss": 0.861,
|
12167 |
+
"step": 19680
|
12168 |
+
},
|
12169 |
+
{
|
12170 |
+
"epoch": 129.54,
|
12171 |
+
"learning_rate": 8.244052863436123e-06,
|
12172 |
+
"loss": 0.9403,
|
12173 |
+
"step": 19690
|
12174 |
+
},
|
12175 |
+
{
|
12176 |
+
"epoch": 129.6,
|
12177 |
+
"learning_rate": 8.21762114537445e-06,
|
12178 |
+
"loss": 0.8556,
|
12179 |
+
"step": 19700
|
12180 |
+
},
|
12181 |
+
{
|
12182 |
+
"epoch": 129.67,
|
12183 |
+
"learning_rate": 8.191189427312776e-06,
|
12184 |
+
"loss": 0.853,
|
12185 |
+
"step": 19710
|
12186 |
+
},
|
12187 |
+
{
|
12188 |
+
"epoch": 129.73,
|
12189 |
+
"learning_rate": 8.164757709251101e-06,
|
12190 |
+
"loss": 0.8882,
|
12191 |
+
"step": 19720
|
12192 |
+
},
|
12193 |
+
{
|
12194 |
+
"epoch": 129.8,
|
12195 |
+
"learning_rate": 8.138325991189428e-06,
|
12196 |
+
"loss": 0.889,
|
12197 |
+
"step": 19730
|
12198 |
+
},
|
12199 |
+
{
|
12200 |
+
"epoch": 129.86,
|
12201 |
+
"learning_rate": 8.111894273127754e-06,
|
12202 |
+
"loss": 0.8306,
|
12203 |
+
"step": 19740
|
12204 |
+
},
|
12205 |
+
{
|
12206 |
+
"epoch": 129.93,
|
12207 |
+
"learning_rate": 8.08546255506608e-06,
|
12208 |
+
"loss": 0.8692,
|
12209 |
+
"step": 19750
|
12210 |
+
},
|
12211 |
+
{
|
12212 |
+
"epoch": 130.0,
|
12213 |
+
"learning_rate": 8.059030837004405e-06,
|
12214 |
+
"loss": 0.894,
|
12215 |
+
"step": 19760
|
12216 |
+
},
|
12217 |
+
{
|
12218 |
+
"epoch": 130.07,
|
12219 |
+
"learning_rate": 8.032599118942732e-06,
|
12220 |
+
"loss": 0.9069,
|
12221 |
+
"step": 19770
|
12222 |
+
},
|
12223 |
+
{
|
12224 |
+
"epoch": 130.13,
|
12225 |
+
"learning_rate": 8.006167400881057e-06,
|
12226 |
+
"loss": 0.8602,
|
12227 |
+
"step": 19780
|
12228 |
+
},
|
12229 |
+
{
|
12230 |
+
"epoch": 130.2,
|
12231 |
+
"learning_rate": 7.979735682819383e-06,
|
12232 |
+
"loss": 0.8758,
|
12233 |
+
"step": 19790
|
12234 |
+
},
|
12235 |
+
{
|
12236 |
+
"epoch": 130.26,
|
12237 |
+
"learning_rate": 7.95330396475771e-06,
|
12238 |
+
"loss": 0.9363,
|
12239 |
+
"step": 19800
|
12240 |
+
},
|
12241 |
+
{
|
12242 |
+
"epoch": 130.33,
|
12243 |
+
"learning_rate": 7.926872246696036e-06,
|
12244 |
+
"loss": 0.8534,
|
12245 |
+
"step": 19810
|
12246 |
+
},
|
12247 |
+
{
|
12248 |
+
"epoch": 130.39,
|
12249 |
+
"learning_rate": 7.900440528634361e-06,
|
12250 |
+
"loss": 0.8481,
|
12251 |
+
"step": 19820
|
12252 |
+
},
|
12253 |
+
{
|
12254 |
+
"epoch": 130.46,
|
12255 |
+
"learning_rate": 7.874008810572686e-06,
|
12256 |
+
"loss": 0.869,
|
12257 |
+
"step": 19830
|
12258 |
+
},
|
12259 |
+
{
|
12260 |
+
"epoch": 130.52,
|
12261 |
+
"learning_rate": 7.847577092511014e-06,
|
12262 |
+
"loss": 0.9053,
|
12263 |
+
"step": 19840
|
12264 |
+
},
|
12265 |
+
{
|
12266 |
+
"epoch": 130.59,
|
12267 |
+
"learning_rate": 7.821145374449339e-06,
|
12268 |
+
"loss": 0.8458,
|
12269 |
+
"step": 19850
|
12270 |
+
},
|
12271 |
+
{
|
12272 |
+
"epoch": 130.65,
|
12273 |
+
"learning_rate": 7.794713656387665e-06,
|
12274 |
+
"loss": 0.8722,
|
12275 |
+
"step": 19860
|
12276 |
+
},
|
12277 |
+
{
|
12278 |
+
"epoch": 130.72,
|
12279 |
+
"learning_rate": 7.768281938325992e-06,
|
12280 |
+
"loss": 0.8498,
|
12281 |
+
"step": 19870
|
12282 |
+
},
|
12283 |
+
{
|
12284 |
+
"epoch": 130.79,
|
12285 |
+
"learning_rate": 7.741850220264317e-06,
|
12286 |
+
"loss": 0.9224,
|
12287 |
+
"step": 19880
|
12288 |
+
},
|
12289 |
+
{
|
12290 |
+
"epoch": 130.85,
|
12291 |
+
"learning_rate": 7.715418502202644e-06,
|
12292 |
+
"loss": 0.8419,
|
12293 |
+
"step": 19890
|
12294 |
+
},
|
12295 |
+
{
|
12296 |
+
"epoch": 130.92,
|
12297 |
+
"learning_rate": 7.688986784140968e-06,
|
12298 |
+
"loss": 0.8453,
|
12299 |
+
"step": 19900
|
12300 |
+
},
|
12301 |
+
{
|
12302 |
+
"epoch": 130.98,
|
12303 |
+
"learning_rate": 7.662555066079295e-06,
|
12304 |
+
"loss": 0.8761,
|
12305 |
+
"step": 19910
|
12306 |
+
},
|
12307 |
+
{
|
12308 |
+
"epoch": 131.05,
|
12309 |
+
"learning_rate": 7.636123348017623e-06,
|
12310 |
+
"loss": 0.9359,
|
12311 |
+
"step": 19920
|
12312 |
+
},
|
12313 |
+
{
|
12314 |
+
"epoch": 131.12,
|
12315 |
+
"learning_rate": 7.609691629955946e-06,
|
12316 |
+
"loss": 0.8415,
|
12317 |
+
"step": 19930
|
12318 |
+
},
|
12319 |
+
{
|
12320 |
+
"epoch": 131.18,
|
12321 |
+
"learning_rate": 7.583259911894273e-06,
|
12322 |
+
"loss": 0.8569,
|
12323 |
+
"step": 19940
|
12324 |
+
},
|
12325 |
+
{
|
12326 |
+
"epoch": 131.25,
|
12327 |
+
"learning_rate": 7.5568281938326e-06,
|
12328 |
+
"loss": 0.9054,
|
12329 |
+
"step": 19950
|
12330 |
+
},
|
12331 |
+
{
|
12332 |
+
"epoch": 131.31,
|
12333 |
+
"learning_rate": 7.530396475770926e-06,
|
12334 |
+
"loss": 0.8698,
|
12335 |
+
"step": 19960
|
12336 |
+
},
|
12337 |
+
{
|
12338 |
+
"epoch": 131.38,
|
12339 |
+
"learning_rate": 7.503964757709251e-06,
|
12340 |
+
"loss": 0.8543,
|
12341 |
+
"step": 19970
|
12342 |
+
},
|
12343 |
+
{
|
12344 |
+
"epoch": 131.45,
|
12345 |
+
"learning_rate": 7.477533039647577e-06,
|
12346 |
+
"loss": 0.8847,
|
12347 |
+
"step": 19980
|
12348 |
+
},
|
12349 |
+
{
|
12350 |
+
"epoch": 131.51,
|
12351 |
+
"learning_rate": 7.451101321585903e-06,
|
12352 |
+
"loss": 0.9421,
|
12353 |
+
"step": 19990
|
12354 |
+
},
|
12355 |
+
{
|
12356 |
+
"epoch": 131.58,
|
12357 |
+
"learning_rate": 7.424669603524229e-06,
|
12358 |
+
"loss": 0.8315,
|
12359 |
+
"step": 20000
|
12360 |
+
},
|
12361 |
+
{
|
12362 |
+
"epoch": 131.58,
|
12363 |
+
"eval_loss": 0.3476375639438629,
|
12364 |
+
"eval_runtime": 318.7177,
|
12365 |
+
"eval_samples_per_second": 22.308,
|
12366 |
+
"eval_steps_per_second": 1.396,
|
12367 |
+
"eval_wer": 0.3504887254427562,
|
12368 |
+
"step": 20000
|
12369 |
+
},
|
12370 |
+
{
|
12371 |
+
"epoch": 131.64,
|
12372 |
+
"learning_rate": 7.398237885462555e-06,
|
12373 |
+
"loss": 0.8478,
|
12374 |
+
"step": 20010
|
12375 |
+
},
|
12376 |
+
{
|
12377 |
+
"epoch": 131.71,
|
12378 |
+
"learning_rate": 7.3718061674008815e-06,
|
12379 |
+
"loss": 0.8467,
|
12380 |
+
"step": 20020
|
12381 |
+
},
|
12382 |
+
{
|
12383 |
+
"epoch": 131.77,
|
12384 |
+
"learning_rate": 7.345374449339207e-06,
|
12385 |
+
"loss": 0.9429,
|
12386 |
+
"step": 20030
|
12387 |
+
},
|
12388 |
+
{
|
12389 |
+
"epoch": 131.84,
|
12390 |
+
"learning_rate": 7.318942731277533e-06,
|
12391 |
+
"loss": 0.8449,
|
12392 |
+
"step": 20040
|
12393 |
+
},
|
12394 |
+
{
|
12395 |
+
"epoch": 131.9,
|
12396 |
+
"learning_rate": 7.292511013215859e-06,
|
12397 |
+
"loss": 0.8877,
|
12398 |
+
"step": 20050
|
12399 |
+
},
|
12400 |
+
{
|
12401 |
+
"epoch": 131.97,
|
12402 |
+
"learning_rate": 7.266079295154185e-06,
|
12403 |
+
"loss": 0.8775,
|
12404 |
+
"step": 20060
|
12405 |
+
},
|
12406 |
+
{
|
12407 |
+
"epoch": 132.04,
|
12408 |
+
"learning_rate": 7.2396475770925115e-06,
|
12409 |
+
"loss": 0.956,
|
12410 |
+
"step": 20070
|
12411 |
+
},
|
12412 |
+
{
|
12413 |
+
"epoch": 132.1,
|
12414 |
+
"learning_rate": 7.213215859030838e-06,
|
12415 |
+
"loss": 0.8259,
|
12416 |
+
"step": 20080
|
12417 |
+
},
|
12418 |
+
{
|
12419 |
+
"epoch": 132.17,
|
12420 |
+
"learning_rate": 7.186784140969163e-06,
|
12421 |
+
"loss": 0.8629,
|
12422 |
+
"step": 20090
|
12423 |
+
},
|
12424 |
+
{
|
12425 |
+
"epoch": 132.24,
|
12426 |
+
"learning_rate": 7.160352422907489e-06,
|
12427 |
+
"loss": 0.8964,
|
12428 |
+
"step": 20100
|
12429 |
+
},
|
12430 |
+
{
|
12431 |
+
"epoch": 132.3,
|
12432 |
+
"learning_rate": 7.133920704845815e-06,
|
12433 |
+
"loss": 0.9075,
|
12434 |
+
"step": 20110
|
12435 |
+
},
|
12436 |
+
{
|
12437 |
+
"epoch": 132.37,
|
12438 |
+
"learning_rate": 7.1074889867841405e-06,
|
12439 |
+
"loss": 0.8127,
|
12440 |
+
"step": 20120
|
12441 |
+
},
|
12442 |
+
{
|
12443 |
+
"epoch": 132.43,
|
12444 |
+
"learning_rate": 7.081057268722468e-06,
|
12445 |
+
"loss": 0.8789,
|
12446 |
+
"step": 20130
|
12447 |
+
},
|
12448 |
+
{
|
12449 |
+
"epoch": 132.5,
|
12450 |
+
"learning_rate": 7.054625550660793e-06,
|
12451 |
+
"loss": 0.9102,
|
12452 |
+
"step": 20140
|
12453 |
+
},
|
12454 |
+
{
|
12455 |
+
"epoch": 132.56,
|
12456 |
+
"learning_rate": 7.0281938325991195e-06,
|
12457 |
+
"loss": 0.8588,
|
12458 |
+
"step": 20150
|
12459 |
+
},
|
12460 |
+
{
|
12461 |
+
"epoch": 132.63,
|
12462 |
+
"learning_rate": 7.001762114537445e-06,
|
12463 |
+
"loss": 0.8515,
|
12464 |
+
"step": 20160
|
12465 |
+
},
|
12466 |
+
{
|
12467 |
+
"epoch": 132.69,
|
12468 |
+
"learning_rate": 6.975330396475771e-06,
|
12469 |
+
"loss": 0.8963,
|
12470 |
+
"step": 20170
|
12471 |
+
},
|
12472 |
+
{
|
12473 |
+
"epoch": 132.76,
|
12474 |
+
"learning_rate": 6.948898678414097e-06,
|
12475 |
+
"loss": 0.9224,
|
12476 |
+
"step": 20180
|
12477 |
+
},
|
12478 |
+
{
|
12479 |
+
"epoch": 132.82,
|
12480 |
+
"learning_rate": 6.922466960352423e-06,
|
12481 |
+
"loss": 0.8314,
|
12482 |
+
"step": 20190
|
12483 |
+
},
|
12484 |
+
{
|
12485 |
+
"epoch": 132.89,
|
12486 |
+
"learning_rate": 6.896035242290749e-06,
|
12487 |
+
"loss": 0.857,
|
12488 |
+
"step": 20200
|
12489 |
+
},
|
12490 |
+
{
|
12491 |
+
"epoch": 132.96,
|
12492 |
+
"learning_rate": 6.869603524229075e-06,
|
12493 |
+
"loss": 0.8458,
|
12494 |
+
"step": 20210
|
12495 |
+
},
|
12496 |
+
{
|
12497 |
+
"epoch": 133.03,
|
12498 |
+
"learning_rate": 6.843171806167401e-06,
|
12499 |
+
"loss": 0.9865,
|
12500 |
+
"step": 20220
|
12501 |
+
},
|
12502 |
+
{
|
12503 |
+
"epoch": 133.09,
|
12504 |
+
"learning_rate": 6.816740088105727e-06,
|
12505 |
+
"loss": 0.8427,
|
12506 |
+
"step": 20230
|
12507 |
+
},
|
12508 |
+
{
|
12509 |
+
"epoch": 133.16,
|
12510 |
+
"learning_rate": 6.790308370044053e-06,
|
12511 |
+
"loss": 0.8474,
|
12512 |
+
"step": 20240
|
12513 |
+
},
|
12514 |
+
{
|
12515 |
+
"epoch": 133.22,
|
12516 |
+
"learning_rate": 6.7638766519823785e-06,
|
12517 |
+
"loss": 0.8718,
|
12518 |
+
"step": 20250
|
12519 |
+
},
|
12520 |
+
{
|
12521 |
+
"epoch": 133.29,
|
12522 |
+
"learning_rate": 6.737444933920706e-06,
|
12523 |
+
"loss": 0.9156,
|
12524 |
+
"step": 20260
|
12525 |
+
},
|
12526 |
+
{
|
12527 |
+
"epoch": 133.35,
|
12528 |
+
"learning_rate": 6.711013215859031e-06,
|
12529 |
+
"loss": 0.8345,
|
12530 |
+
"step": 20270
|
12531 |
+
},
|
12532 |
+
{
|
12533 |
+
"epoch": 133.42,
|
12534 |
+
"learning_rate": 6.6845814977973575e-06,
|
12535 |
+
"loss": 0.8664,
|
12536 |
+
"step": 20280
|
12537 |
+
},
|
12538 |
+
{
|
12539 |
+
"epoch": 133.48,
|
12540 |
+
"learning_rate": 6.658149779735683e-06,
|
12541 |
+
"loss": 0.9086,
|
12542 |
+
"step": 20290
|
12543 |
+
},
|
12544 |
+
{
|
12545 |
+
"epoch": 133.55,
|
12546 |
+
"learning_rate": 6.631718061674008e-06,
|
12547 |
+
"loss": 0.8815,
|
12548 |
+
"step": 20300
|
12549 |
+
},
|
12550 |
+
{
|
12551 |
+
"epoch": 133.62,
|
12552 |
+
"learning_rate": 6.605286343612335e-06,
|
12553 |
+
"loss": 0.8388,
|
12554 |
+
"step": 20310
|
12555 |
+
},
|
12556 |
+
{
|
12557 |
+
"epoch": 133.68,
|
12558 |
+
"learning_rate": 6.578854625550661e-06,
|
12559 |
+
"loss": 0.8464,
|
12560 |
+
"step": 20320
|
12561 |
+
},
|
12562 |
+
{
|
12563 |
+
"epoch": 133.75,
|
12564 |
+
"learning_rate": 6.552422907488987e-06,
|
12565 |
+
"loss": 0.9118,
|
12566 |
+
"step": 20330
|
12567 |
+
},
|
12568 |
+
{
|
12569 |
+
"epoch": 133.81,
|
12570 |
+
"learning_rate": 6.525991189427313e-06,
|
12571 |
+
"loss": 0.836,
|
12572 |
+
"step": 20340
|
12573 |
+
},
|
12574 |
+
{
|
12575 |
+
"epoch": 133.88,
|
12576 |
+
"learning_rate": 6.499559471365639e-06,
|
12577 |
+
"loss": 0.8364,
|
12578 |
+
"step": 20350
|
12579 |
+
},
|
12580 |
+
{
|
12581 |
+
"epoch": 133.94,
|
12582 |
+
"learning_rate": 6.473127753303965e-06,
|
12583 |
+
"loss": 0.8733,
|
12584 |
+
"step": 20360
|
12585 |
+
},
|
12586 |
+
{
|
12587 |
+
"epoch": 134.01,
|
12588 |
+
"learning_rate": 6.446696035242291e-06,
|
12589 |
+
"loss": 0.9697,
|
12590 |
+
"step": 20370
|
12591 |
+
},
|
12592 |
+
{
|
12593 |
+
"epoch": 134.08,
|
12594 |
+
"learning_rate": 6.420264317180617e-06,
|
12595 |
+
"loss": 0.8447,
|
12596 |
+
"step": 20380
|
12597 |
+
},
|
12598 |
+
{
|
12599 |
+
"epoch": 134.14,
|
12600 |
+
"learning_rate": 6.393832599118943e-06,
|
12601 |
+
"loss": 0.8811,
|
12602 |
+
"step": 20390
|
12603 |
+
},
|
12604 |
+
{
|
12605 |
+
"epoch": 134.21,
|
12606 |
+
"learning_rate": 6.367400881057269e-06,
|
12607 |
+
"loss": 0.8537,
|
12608 |
+
"step": 20400
|
12609 |
+
},
|
12610 |
+
{
|
12611 |
+
"epoch": 134.27,
|
12612 |
+
"learning_rate": 6.3409691629955946e-06,
|
12613 |
+
"loss": 0.9016,
|
12614 |
+
"step": 20410
|
12615 |
+
},
|
12616 |
+
{
|
12617 |
+
"epoch": 134.34,
|
12618 |
+
"learning_rate": 6.314537444933921e-06,
|
12619 |
+
"loss": 0.8262,
|
12620 |
+
"step": 20420
|
12621 |
+
},
|
12622 |
+
{
|
12623 |
+
"epoch": 134.41,
|
12624 |
+
"learning_rate": 6.288105726872246e-06,
|
12625 |
+
"loss": 0.8514,
|
12626 |
+
"step": 20430
|
12627 |
+
},
|
12628 |
+
{
|
12629 |
+
"epoch": 134.47,
|
12630 |
+
"learning_rate": 6.261674008810573e-06,
|
12631 |
+
"loss": 0.8821,
|
12632 |
+
"step": 20440
|
12633 |
+
},
|
12634 |
+
{
|
12635 |
+
"epoch": 134.54,
|
12636 |
+
"learning_rate": 6.235242290748899e-06,
|
12637 |
+
"loss": 0.8956,
|
12638 |
+
"step": 20450
|
12639 |
+
},
|
12640 |
+
{
|
12641 |
+
"epoch": 134.6,
|
12642 |
+
"learning_rate": 6.208810572687225e-06,
|
12643 |
+
"loss": 0.8119,
|
12644 |
+
"step": 20460
|
12645 |
+
},
|
12646 |
+
{
|
12647 |
+
"epoch": 134.67,
|
12648 |
+
"learning_rate": 6.182378854625551e-06,
|
12649 |
+
"loss": 0.8541,
|
12650 |
+
"step": 20470
|
12651 |
+
},
|
12652 |
+
{
|
12653 |
+
"epoch": 134.73,
|
12654 |
+
"learning_rate": 6.155947136563876e-06,
|
12655 |
+
"loss": 0.8941,
|
12656 |
+
"step": 20480
|
12657 |
+
},
|
12658 |
+
{
|
12659 |
+
"epoch": 134.8,
|
12660 |
+
"learning_rate": 6.129515418502203e-06,
|
12661 |
+
"loss": 0.8826,
|
12662 |
+
"step": 20490
|
12663 |
+
},
|
12664 |
+
{
|
12665 |
+
"epoch": 134.86,
|
12666 |
+
"learning_rate": 6.103083700440528e-06,
|
12667 |
+
"loss": 0.8542,
|
12668 |
+
"step": 20500
|
12669 |
+
},
|
12670 |
+
{
|
12671 |
+
"epoch": 134.86,
|
12672 |
+
"eval_loss": 0.34747639298439026,
|
12673 |
+
"eval_runtime": 318.584,
|
12674 |
+
"eval_samples_per_second": 22.318,
|
12675 |
+
"eval_steps_per_second": 1.397,
|
12676 |
+
"eval_wer": 0.3505855027581535,
|
12677 |
+
"step": 20500
|
12678 |
+
},
|
12679 |
+
{
|
12680 |
+
"epoch": 134.93,
|
12681 |
+
"learning_rate": 6.076651982378855e-06,
|
12682 |
+
"loss": 0.8579,
|
12683 |
+
"step": 20510
|
12684 |
+
},
|
12685 |
+
{
|
12686 |
+
"epoch": 135.0,
|
12687 |
+
"learning_rate": 6.050220264317181e-06,
|
12688 |
+
"loss": 0.9107,
|
12689 |
+
"step": 20520
|
12690 |
+
},
|
12691 |
+
{
|
12692 |
+
"epoch": 135.07,
|
12693 |
+
"learning_rate": 6.023788546255507e-06,
|
12694 |
+
"loss": 0.9237,
|
12695 |
+
"step": 20530
|
12696 |
+
},
|
12697 |
+
{
|
12698 |
+
"epoch": 135.13,
|
12699 |
+
"learning_rate": 5.9973568281938325e-06,
|
12700 |
+
"loss": 0.8477,
|
12701 |
+
"step": 20540
|
12702 |
+
},
|
12703 |
+
{
|
12704 |
+
"epoch": 135.2,
|
12705 |
+
"learning_rate": 5.970925110132159e-06,
|
12706 |
+
"loss": 0.8807,
|
12707 |
+
"step": 20550
|
12708 |
+
},
|
12709 |
+
{
|
12710 |
+
"epoch": 135.26,
|
12711 |
+
"learning_rate": 5.944493392070484e-06,
|
12712 |
+
"loss": 0.9002,
|
12713 |
+
"step": 20560
|
12714 |
+
},
|
12715 |
+
{
|
12716 |
+
"epoch": 135.33,
|
12717 |
+
"learning_rate": 5.918061674008811e-06,
|
12718 |
+
"loss": 0.8213,
|
12719 |
+
"step": 20570
|
12720 |
+
},
|
12721 |
+
{
|
12722 |
+
"epoch": 135.39,
|
12723 |
+
"learning_rate": 5.891629955947137e-06,
|
12724 |
+
"loss": 0.8608,
|
12725 |
+
"step": 20580
|
12726 |
+
},
|
12727 |
+
{
|
12728 |
+
"epoch": 135.46,
|
12729 |
+
"learning_rate": 5.8651982378854624e-06,
|
12730 |
+
"loss": 0.8611,
|
12731 |
+
"step": 20590
|
12732 |
+
},
|
12733 |
+
{
|
12734 |
+
"epoch": 135.52,
|
12735 |
+
"learning_rate": 5.838766519823789e-06,
|
12736 |
+
"loss": 0.9184,
|
12737 |
+
"step": 20600
|
12738 |
+
},
|
12739 |
+
{
|
12740 |
+
"epoch": 135.59,
|
12741 |
+
"learning_rate": 5.812334801762114e-06,
|
12742 |
+
"loss": 0.8341,
|
12743 |
+
"step": 20610
|
12744 |
+
},
|
12745 |
+
{
|
12746 |
+
"epoch": 135.65,
|
12747 |
+
"learning_rate": 5.7859030837004406e-06,
|
12748 |
+
"loss": 0.8371,
|
12749 |
+
"step": 20620
|
12750 |
+
},
|
12751 |
+
{
|
12752 |
+
"epoch": 135.72,
|
12753 |
+
"learning_rate": 5.759471365638767e-06,
|
12754 |
+
"loss": 0.866,
|
12755 |
+
"step": 20630
|
12756 |
+
},
|
12757 |
+
{
|
12758 |
+
"epoch": 135.79,
|
12759 |
+
"learning_rate": 5.733039647577093e-06,
|
12760 |
+
"loss": 0.9034,
|
12761 |
+
"step": 20640
|
12762 |
+
},
|
12763 |
+
{
|
12764 |
+
"epoch": 135.85,
|
12765 |
+
"learning_rate": 5.706607929515419e-06,
|
12766 |
+
"loss": 0.85,
|
12767 |
+
"step": 20650
|
12768 |
+
},
|
12769 |
+
{
|
12770 |
+
"epoch": 135.92,
|
12771 |
+
"learning_rate": 5.680176211453745e-06,
|
12772 |
+
"loss": 0.8806,
|
12773 |
+
"step": 20660
|
12774 |
+
},
|
12775 |
+
{
|
12776 |
+
"epoch": 135.98,
|
12777 |
+
"learning_rate": 5.656387665198238e-06,
|
12778 |
+
"loss": 0.8653,
|
12779 |
+
"step": 20670
|
12780 |
+
},
|
12781 |
+
{
|
12782 |
+
"epoch": 136.05,
|
12783 |
+
"learning_rate": 5.6299559471365644e-06,
|
12784 |
+
"loss": 0.9354,
|
12785 |
+
"step": 20680
|
12786 |
+
},
|
12787 |
+
{
|
12788 |
+
"epoch": 136.12,
|
12789 |
+
"learning_rate": 5.60352422907489e-06,
|
12790 |
+
"loss": 0.8336,
|
12791 |
+
"step": 20690
|
12792 |
+
},
|
12793 |
+
{
|
12794 |
+
"epoch": 136.18,
|
12795 |
+
"learning_rate": 5.577092511013216e-06,
|
12796 |
+
"loss": 0.856,
|
12797 |
+
"step": 20700
|
12798 |
+
},
|
12799 |
+
{
|
12800 |
+
"epoch": 136.25,
|
12801 |
+
"learning_rate": 5.550660792951542e-06,
|
12802 |
+
"loss": 0.8877,
|
12803 |
+
"step": 20710
|
12804 |
+
},
|
12805 |
+
{
|
12806 |
+
"epoch": 136.31,
|
12807 |
+
"learning_rate": 5.524229074889868e-06,
|
12808 |
+
"loss": 0.8517,
|
12809 |
+
"step": 20720
|
12810 |
+
},
|
12811 |
+
{
|
12812 |
+
"epoch": 136.38,
|
12813 |
+
"learning_rate": 5.4977973568281935e-06,
|
12814 |
+
"loss": 0.8718,
|
12815 |
+
"step": 20730
|
12816 |
+
},
|
12817 |
+
{
|
12818 |
+
"epoch": 136.45,
|
12819 |
+
"learning_rate": 5.471365638766521e-06,
|
12820 |
+
"loss": 0.8488,
|
12821 |
+
"step": 20740
|
12822 |
+
},
|
12823 |
+
{
|
12824 |
+
"epoch": 136.51,
|
12825 |
+
"learning_rate": 5.444933920704846e-06,
|
12826 |
+
"loss": 0.8924,
|
12827 |
+
"step": 20750
|
12828 |
+
},
|
12829 |
+
{
|
12830 |
+
"epoch": 136.58,
|
12831 |
+
"learning_rate": 5.418502202643172e-06,
|
12832 |
+
"loss": 0.8292,
|
12833 |
+
"step": 20760
|
12834 |
+
},
|
12835 |
+
{
|
12836 |
+
"epoch": 136.64,
|
12837 |
+
"learning_rate": 5.392070484581498e-06,
|
12838 |
+
"loss": 0.8556,
|
12839 |
+
"step": 20770
|
12840 |
+
},
|
12841 |
+
{
|
12842 |
+
"epoch": 136.71,
|
12843 |
+
"learning_rate": 5.3656387665198234e-06,
|
12844 |
+
"loss": 0.8545,
|
12845 |
+
"step": 20780
|
12846 |
+
},
|
12847 |
+
{
|
12848 |
+
"epoch": 136.77,
|
12849 |
+
"learning_rate": 5.33920704845815e-06,
|
12850 |
+
"loss": 0.9007,
|
12851 |
+
"step": 20790
|
12852 |
+
},
|
12853 |
+
{
|
12854 |
+
"epoch": 136.84,
|
12855 |
+
"learning_rate": 5.312775330396476e-06,
|
12856 |
+
"loss": 0.8256,
|
12857 |
+
"step": 20800
|
12858 |
+
},
|
12859 |
+
{
|
12860 |
+
"epoch": 136.9,
|
12861 |
+
"learning_rate": 5.286343612334802e-06,
|
12862 |
+
"loss": 0.8613,
|
12863 |
+
"step": 20810
|
12864 |
+
},
|
12865 |
+
{
|
12866 |
+
"epoch": 136.97,
|
12867 |
+
"learning_rate": 5.259911894273128e-06,
|
12868 |
+
"loss": 0.8568,
|
12869 |
+
"step": 20820
|
12870 |
+
},
|
12871 |
+
{
|
12872 |
+
"epoch": 137.04,
|
12873 |
+
"learning_rate": 5.233480176211454e-06,
|
12874 |
+
"loss": 0.9363,
|
12875 |
+
"step": 20830
|
12876 |
+
},
|
12877 |
+
{
|
12878 |
+
"epoch": 137.1,
|
12879 |
+
"learning_rate": 5.20704845814978e-06,
|
12880 |
+
"loss": 0.8349,
|
12881 |
+
"step": 20840
|
12882 |
+
},
|
12883 |
+
{
|
12884 |
+
"epoch": 137.17,
|
12885 |
+
"learning_rate": 5.180616740088105e-06,
|
12886 |
+
"loss": 0.8585,
|
12887 |
+
"step": 20850
|
12888 |
+
},
|
12889 |
+
{
|
12890 |
+
"epoch": 137.24,
|
12891 |
+
"learning_rate": 5.154185022026432e-06,
|
12892 |
+
"loss": 0.906,
|
12893 |
+
"step": 20860
|
12894 |
+
},
|
12895 |
+
{
|
12896 |
+
"epoch": 137.3,
|
12897 |
+
"learning_rate": 5.127753303964758e-06,
|
12898 |
+
"loss": 0.8841,
|
12899 |
+
"step": 20870
|
12900 |
+
},
|
12901 |
+
{
|
12902 |
+
"epoch": 137.37,
|
12903 |
+
"learning_rate": 5.101321585903084e-06,
|
12904 |
+
"loss": 0.8308,
|
12905 |
+
"step": 20880
|
12906 |
+
},
|
12907 |
+
{
|
12908 |
+
"epoch": 137.43,
|
12909 |
+
"learning_rate": 5.07488986784141e-06,
|
12910 |
+
"loss": 0.868,
|
12911 |
+
"step": 20890
|
12912 |
+
},
|
12913 |
+
{
|
12914 |
+
"epoch": 137.5,
|
12915 |
+
"learning_rate": 5.048458149779736e-06,
|
12916 |
+
"loss": 0.8993,
|
12917 |
+
"step": 20900
|
12918 |
+
},
|
12919 |
+
{
|
12920 |
+
"epoch": 137.56,
|
12921 |
+
"learning_rate": 5.022026431718061e-06,
|
12922 |
+
"loss": 0.8606,
|
12923 |
+
"step": 20910
|
12924 |
+
},
|
12925 |
+
{
|
12926 |
+
"epoch": 137.63,
|
12927 |
+
"learning_rate": 4.9955947136563886e-06,
|
12928 |
+
"loss": 0.8182,
|
12929 |
+
"step": 20920
|
12930 |
+
},
|
12931 |
+
{
|
12932 |
+
"epoch": 137.69,
|
12933 |
+
"learning_rate": 4.969162995594714e-06,
|
12934 |
+
"loss": 0.8791,
|
12935 |
+
"step": 20930
|
12936 |
+
},
|
12937 |
+
{
|
12938 |
+
"epoch": 137.76,
|
12939 |
+
"learning_rate": 4.9427312775330395e-06,
|
12940 |
+
"loss": 0.9186,
|
12941 |
+
"step": 20940
|
12942 |
+
},
|
12943 |
+
{
|
12944 |
+
"epoch": 137.82,
|
12945 |
+
"learning_rate": 4.916299559471366e-06,
|
12946 |
+
"loss": 0.8232,
|
12947 |
+
"step": 20950
|
12948 |
+
},
|
12949 |
+
{
|
12950 |
+
"epoch": 137.89,
|
12951 |
+
"learning_rate": 4.889867841409691e-06,
|
12952 |
+
"loss": 0.8488,
|
12953 |
+
"step": 20960
|
12954 |
+
},
|
12955 |
+
{
|
12956 |
+
"epoch": 137.96,
|
12957 |
+
"learning_rate": 4.863436123348018e-06,
|
12958 |
+
"loss": 0.8555,
|
12959 |
+
"step": 20970
|
12960 |
+
},
|
12961 |
+
{
|
12962 |
+
"epoch": 138.03,
|
12963 |
+
"learning_rate": 4.837004405286343e-06,
|
12964 |
+
"loss": 0.9995,
|
12965 |
+
"step": 20980
|
12966 |
+
},
|
12967 |
+
{
|
12968 |
+
"epoch": 138.09,
|
12969 |
+
"learning_rate": 4.81057268722467e-06,
|
12970 |
+
"loss": 0.8541,
|
12971 |
+
"step": 20990
|
12972 |
+
},
|
12973 |
+
{
|
12974 |
+
"epoch": 138.16,
|
12975 |
+
"learning_rate": 4.784140969162996e-06,
|
12976 |
+
"loss": 0.8478,
|
12977 |
+
"step": 21000
|
12978 |
+
},
|
12979 |
+
{
|
12980 |
+
"epoch": 138.16,
|
12981 |
+
"eval_loss": 0.34300604462623596,
|
12982 |
+
"eval_runtime": 316.2575,
|
12983 |
+
"eval_samples_per_second": 22.482,
|
12984 |
+
"eval_steps_per_second": 1.407,
|
12985 |
+
"eval_wer": 0.3481467144101423,
|
12986 |
+
"step": 21000
|
12987 |
+
},
|
12988 |
+
{
|
12989 |
+
"epoch": 138.22,
|
12990 |
+
"learning_rate": 4.757709251101322e-06,
|
12991 |
+
"loss": 0.8544,
|
12992 |
+
"step": 21010
|
12993 |
+
},
|
12994 |
+
{
|
12995 |
+
"epoch": 138.29,
|
12996 |
+
"learning_rate": 4.7312775330396475e-06,
|
12997 |
+
"loss": 0.8809,
|
12998 |
+
"step": 21020
|
12999 |
+
},
|
13000 |
+
{
|
13001 |
+
"epoch": 138.35,
|
13002 |
+
"learning_rate": 4.704845814977973e-06,
|
13003 |
+
"loss": 0.8406,
|
13004 |
+
"step": 21030
|
13005 |
+
},
|
13006 |
+
{
|
13007 |
+
"epoch": 138.42,
|
13008 |
+
"learning_rate": 4.678414096916299e-06,
|
13009 |
+
"loss": 0.8689,
|
13010 |
+
"step": 21040
|
13011 |
+
},
|
13012 |
+
{
|
13013 |
+
"epoch": 138.48,
|
13014 |
+
"learning_rate": 4.651982378854626e-06,
|
13015 |
+
"loss": 0.8715,
|
13016 |
+
"step": 21050
|
13017 |
+
},
|
13018 |
+
{
|
13019 |
+
"epoch": 138.55,
|
13020 |
+
"learning_rate": 4.625550660792952e-06,
|
13021 |
+
"loss": 0.8716,
|
13022 |
+
"step": 21060
|
13023 |
+
},
|
13024 |
+
{
|
13025 |
+
"epoch": 138.62,
|
13026 |
+
"learning_rate": 4.5991189427312775e-06,
|
13027 |
+
"loss": 0.8309,
|
13028 |
+
"step": 21070
|
13029 |
+
},
|
13030 |
+
{
|
13031 |
+
"epoch": 138.68,
|
13032 |
+
"learning_rate": 4.572687224669604e-06,
|
13033 |
+
"loss": 0.8757,
|
13034 |
+
"step": 21080
|
13035 |
+
},
|
13036 |
+
{
|
13037 |
+
"epoch": 138.75,
|
13038 |
+
"learning_rate": 4.546255506607929e-06,
|
13039 |
+
"loss": 0.9188,
|
13040 |
+
"step": 21090
|
13041 |
+
},
|
13042 |
+
{
|
13043 |
+
"epoch": 138.81,
|
13044 |
+
"learning_rate": 4.519823788546256e-06,
|
13045 |
+
"loss": 0.8356,
|
13046 |
+
"step": 21100
|
13047 |
+
},
|
13048 |
+
{
|
13049 |
+
"epoch": 138.88,
|
13050 |
+
"learning_rate": 4.493392070484582e-06,
|
13051 |
+
"loss": 0.8031,
|
13052 |
+
"step": 21110
|
13053 |
+
},
|
13054 |
+
{
|
13055 |
+
"epoch": 138.94,
|
13056 |
+
"learning_rate": 4.466960352422908e-06,
|
13057 |
+
"loss": 0.8519,
|
13058 |
+
"step": 21120
|
13059 |
+
},
|
13060 |
+
{
|
13061 |
+
"epoch": 139.01,
|
13062 |
+
"learning_rate": 4.440528634361234e-06,
|
13063 |
+
"loss": 0.9894,
|
13064 |
+
"step": 21130
|
13065 |
+
},
|
13066 |
+
{
|
13067 |
+
"epoch": 139.08,
|
13068 |
+
"learning_rate": 4.414096916299559e-06,
|
13069 |
+
"loss": 0.835,
|
13070 |
+
"step": 21140
|
13071 |
+
},
|
13072 |
+
{
|
13073 |
+
"epoch": 139.14,
|
13074 |
+
"learning_rate": 4.3876651982378855e-06,
|
13075 |
+
"loss": 0.8651,
|
13076 |
+
"step": 21150
|
13077 |
+
},
|
13078 |
+
{
|
13079 |
+
"epoch": 139.21,
|
13080 |
+
"learning_rate": 4.361233480176211e-06,
|
13081 |
+
"loss": 0.8675,
|
13082 |
+
"step": 21160
|
13083 |
+
},
|
13084 |
+
{
|
13085 |
+
"epoch": 139.27,
|
13086 |
+
"learning_rate": 4.334801762114538e-06,
|
13087 |
+
"loss": 0.9105,
|
13088 |
+
"step": 21170
|
13089 |
+
},
|
13090 |
+
{
|
13091 |
+
"epoch": 139.34,
|
13092 |
+
"learning_rate": 4.308370044052864e-06,
|
13093 |
+
"loss": 0.8082,
|
13094 |
+
"step": 21180
|
13095 |
+
},
|
13096 |
+
{
|
13097 |
+
"epoch": 139.41,
|
13098 |
+
"learning_rate": 4.28193832599119e-06,
|
13099 |
+
"loss": 0.8535,
|
13100 |
+
"step": 21190
|
13101 |
+
},
|
13102 |
+
{
|
13103 |
+
"epoch": 139.47,
|
13104 |
+
"learning_rate": 4.255506607929515e-06,
|
13105 |
+
"loss": 0.8652,
|
13106 |
+
"step": 21200
|
13107 |
+
},
|
13108 |
+
{
|
13109 |
+
"epoch": 139.54,
|
13110 |
+
"learning_rate": 4.229074889867842e-06,
|
13111 |
+
"loss": 0.9299,
|
13112 |
+
"step": 21210
|
13113 |
+
},
|
13114 |
+
{
|
13115 |
+
"epoch": 139.6,
|
13116 |
+
"learning_rate": 4.202643171806167e-06,
|
13117 |
+
"loss": 0.8279,
|
13118 |
+
"step": 21220
|
13119 |
+
},
|
13120 |
+
{
|
13121 |
+
"epoch": 139.67,
|
13122 |
+
"learning_rate": 4.1762114537444935e-06,
|
13123 |
+
"loss": 0.8695,
|
13124 |
+
"step": 21230
|
13125 |
+
},
|
13126 |
+
{
|
13127 |
+
"epoch": 139.73,
|
13128 |
+
"learning_rate": 4.14977973568282e-06,
|
13129 |
+
"loss": 0.9049,
|
13130 |
+
"step": 21240
|
13131 |
+
},
|
13132 |
+
{
|
13133 |
+
"epoch": 139.8,
|
13134 |
+
"learning_rate": 4.123348017621145e-06,
|
13135 |
+
"loss": 0.8715,
|
13136 |
+
"step": 21250
|
13137 |
+
},
|
13138 |
+
{
|
13139 |
+
"epoch": 139.86,
|
13140 |
+
"learning_rate": 4.096916299559472e-06,
|
13141 |
+
"loss": 0.8472,
|
13142 |
+
"step": 21260
|
13143 |
+
},
|
13144 |
+
{
|
13145 |
+
"epoch": 139.93,
|
13146 |
+
"learning_rate": 4.070484581497797e-06,
|
13147 |
+
"loss": 0.8788,
|
13148 |
+
"step": 21270
|
13149 |
+
},
|
13150 |
+
{
|
13151 |
+
"epoch": 140.0,
|
13152 |
+
"learning_rate": 4.0440528634361235e-06,
|
13153 |
+
"loss": 0.8908,
|
13154 |
+
"step": 21280
|
13155 |
+
},
|
13156 |
+
{
|
13157 |
+
"epoch": 140.07,
|
13158 |
+
"learning_rate": 4.017621145374449e-06,
|
13159 |
+
"loss": 0.9117,
|
13160 |
+
"step": 21290
|
13161 |
+
},
|
13162 |
+
{
|
13163 |
+
"epoch": 140.13,
|
13164 |
+
"learning_rate": 3.991189427312776e-06,
|
13165 |
+
"loss": 0.8214,
|
13166 |
+
"step": 21300
|
13167 |
+
},
|
13168 |
+
{
|
13169 |
+
"epoch": 140.2,
|
13170 |
+
"learning_rate": 3.964757709251102e-06,
|
13171 |
+
"loss": 0.8663,
|
13172 |
+
"step": 21310
|
13173 |
+
},
|
13174 |
+
{
|
13175 |
+
"epoch": 140.26,
|
13176 |
+
"learning_rate": 3.938325991189427e-06,
|
13177 |
+
"loss": 0.9228,
|
13178 |
+
"step": 21320
|
13179 |
+
},
|
13180 |
+
{
|
13181 |
+
"epoch": 140.33,
|
13182 |
+
"learning_rate": 3.911894273127753e-06,
|
13183 |
+
"loss": 0.8202,
|
13184 |
+
"step": 21330
|
13185 |
+
},
|
13186 |
+
{
|
13187 |
+
"epoch": 140.39,
|
13188 |
+
"learning_rate": 3.885462555066079e-06,
|
13189 |
+
"loss": 0.8349,
|
13190 |
+
"step": 21340
|
13191 |
+
},
|
13192 |
+
{
|
13193 |
+
"epoch": 140.46,
|
13194 |
+
"learning_rate": 3.859030837004405e-06,
|
13195 |
+
"loss": 0.8695,
|
13196 |
+
"step": 21350
|
13197 |
+
},
|
13198 |
+
{
|
13199 |
+
"epoch": 140.52,
|
13200 |
+
"learning_rate": 3.8325991189427315e-06,
|
13201 |
+
"loss": 0.9028,
|
13202 |
+
"step": 21360
|
13203 |
+
},
|
13204 |
+
{
|
13205 |
+
"epoch": 140.59,
|
13206 |
+
"learning_rate": 3.806167400881058e-06,
|
13207 |
+
"loss": 0.8398,
|
13208 |
+
"step": 21370
|
13209 |
+
},
|
13210 |
+
{
|
13211 |
+
"epoch": 140.65,
|
13212 |
+
"learning_rate": 3.7797356828193833e-06,
|
13213 |
+
"loss": 0.8379,
|
13214 |
+
"step": 21380
|
13215 |
+
},
|
13216 |
+
{
|
13217 |
+
"epoch": 140.72,
|
13218 |
+
"learning_rate": 3.7533039647577096e-06,
|
13219 |
+
"loss": 0.8491,
|
13220 |
+
"step": 21390
|
13221 |
+
},
|
13222 |
+
{
|
13223 |
+
"epoch": 140.79,
|
13224 |
+
"learning_rate": 3.7268722466960355e-06,
|
13225 |
+
"loss": 0.9013,
|
13226 |
+
"step": 21400
|
13227 |
+
},
|
13228 |
+
{
|
13229 |
+
"epoch": 140.85,
|
13230 |
+
"learning_rate": 3.7004405286343614e-06,
|
13231 |
+
"loss": 0.845,
|
13232 |
+
"step": 21410
|
13233 |
+
},
|
13234 |
+
{
|
13235 |
+
"epoch": 140.92,
|
13236 |
+
"learning_rate": 3.6740088105726873e-06,
|
13237 |
+
"loss": 0.8546,
|
13238 |
+
"step": 21420
|
13239 |
+
},
|
13240 |
+
{
|
13241 |
+
"epoch": 140.98,
|
13242 |
+
"learning_rate": 3.6475770925110132e-06,
|
13243 |
+
"loss": 0.8963,
|
13244 |
+
"step": 21430
|
13245 |
+
},
|
13246 |
+
{
|
13247 |
+
"epoch": 141.05,
|
13248 |
+
"learning_rate": 3.6211453744493395e-06,
|
13249 |
+
"loss": 0.9693,
|
13250 |
+
"step": 21440
|
13251 |
+
},
|
13252 |
+
{
|
13253 |
+
"epoch": 141.12,
|
13254 |
+
"learning_rate": 3.594713656387665e-06,
|
13255 |
+
"loss": 0.8501,
|
13256 |
+
"step": 21450
|
13257 |
+
},
|
13258 |
+
{
|
13259 |
+
"epoch": 141.18,
|
13260 |
+
"learning_rate": 3.568281938325991e-06,
|
13261 |
+
"loss": 0.8597,
|
13262 |
+
"step": 21460
|
13263 |
+
},
|
13264 |
+
{
|
13265 |
+
"epoch": 141.25,
|
13266 |
+
"learning_rate": 3.5418502202643172e-06,
|
13267 |
+
"loss": 0.8948,
|
13268 |
+
"step": 21470
|
13269 |
+
},
|
13270 |
+
{
|
13271 |
+
"epoch": 141.31,
|
13272 |
+
"learning_rate": 3.515418502202643e-06,
|
13273 |
+
"loss": 0.844,
|
13274 |
+
"step": 21480
|
13275 |
+
},
|
13276 |
+
{
|
13277 |
+
"epoch": 141.38,
|
13278 |
+
"learning_rate": 3.488986784140969e-06,
|
13279 |
+
"loss": 0.8161,
|
13280 |
+
"step": 21490
|
13281 |
+
},
|
13282 |
+
{
|
13283 |
+
"epoch": 141.45,
|
13284 |
+
"learning_rate": 3.4625550660792954e-06,
|
13285 |
+
"loss": 0.8643,
|
13286 |
+
"step": 21500
|
13287 |
+
},
|
13288 |
+
{
|
13289 |
+
"epoch": 141.45,
|
13290 |
+
"eval_loss": 0.3450528085231781,
|
13291 |
+
"eval_runtime": 320.1699,
|
13292 |
+
"eval_samples_per_second": 22.207,
|
13293 |
+
"eval_steps_per_second": 1.39,
|
13294 |
+
"eval_wer": 0.3484564018194135,
|
13295 |
+
"step": 21500
|
13296 |
+
},
|
13297 |
+
{
|
13298 |
+
"epoch": 141.51,
|
13299 |
+
"learning_rate": 3.4361233480176213e-06,
|
13300 |
+
"loss": 0.9,
|
13301 |
+
"step": 21510
|
13302 |
+
},
|
13303 |
+
{
|
13304 |
+
"epoch": 141.58,
|
13305 |
+
"learning_rate": 3.409691629955947e-06,
|
13306 |
+
"loss": 0.8424,
|
13307 |
+
"step": 21520
|
13308 |
+
},
|
13309 |
+
{
|
13310 |
+
"epoch": 141.64,
|
13311 |
+
"learning_rate": 3.3832599118942735e-06,
|
13312 |
+
"loss": 0.8523,
|
13313 |
+
"step": 21530
|
13314 |
+
},
|
13315 |
+
{
|
13316 |
+
"epoch": 141.71,
|
13317 |
+
"learning_rate": 3.3568281938325994e-06,
|
13318 |
+
"loss": 0.8772,
|
13319 |
+
"step": 21540
|
13320 |
+
},
|
13321 |
+
{
|
13322 |
+
"epoch": 141.77,
|
13323 |
+
"learning_rate": 3.330396475770925e-06,
|
13324 |
+
"loss": 0.9149,
|
13325 |
+
"step": 21550
|
13326 |
+
},
|
13327 |
+
{
|
13328 |
+
"epoch": 141.84,
|
13329 |
+
"learning_rate": 3.303964757709251e-06,
|
13330 |
+
"loss": 0.8252,
|
13331 |
+
"step": 21560
|
13332 |
+
},
|
13333 |
+
{
|
13334 |
+
"epoch": 141.9,
|
13335 |
+
"learning_rate": 3.277533039647577e-06,
|
13336 |
+
"loss": 0.8308,
|
13337 |
+
"step": 21570
|
13338 |
+
},
|
13339 |
+
{
|
13340 |
+
"epoch": 141.97,
|
13341 |
+
"learning_rate": 3.251101321585903e-06,
|
13342 |
+
"loss": 0.8585,
|
13343 |
+
"step": 21580
|
13344 |
+
},
|
13345 |
+
{
|
13346 |
+
"epoch": 142.04,
|
13347 |
+
"learning_rate": 3.2246696035242293e-06,
|
13348 |
+
"loss": 0.9729,
|
13349 |
+
"step": 21590
|
13350 |
+
},
|
13351 |
+
{
|
13352 |
+
"epoch": 142.1,
|
13353 |
+
"learning_rate": 3.198237885462555e-06,
|
13354 |
+
"loss": 0.8169,
|
13355 |
+
"step": 21600
|
13356 |
+
},
|
13357 |
+
{
|
13358 |
+
"epoch": 142.17,
|
13359 |
+
"learning_rate": 3.171806167400881e-06,
|
13360 |
+
"loss": 0.8536,
|
13361 |
+
"step": 21610
|
13362 |
+
},
|
13363 |
+
{
|
13364 |
+
"epoch": 142.24,
|
13365 |
+
"learning_rate": 3.1453744493392074e-06,
|
13366 |
+
"loss": 0.8764,
|
13367 |
+
"step": 21620
|
13368 |
+
},
|
13369 |
+
{
|
13370 |
+
"epoch": 142.3,
|
13371 |
+
"learning_rate": 3.1189427312775333e-06,
|
13372 |
+
"loss": 0.8731,
|
13373 |
+
"step": 21630
|
13374 |
+
},
|
13375 |
+
{
|
13376 |
+
"epoch": 142.37,
|
13377 |
+
"learning_rate": 3.092511013215859e-06,
|
13378 |
+
"loss": 0.8372,
|
13379 |
+
"step": 21640
|
13380 |
+
},
|
13381 |
+
{
|
13382 |
+
"epoch": 142.43,
|
13383 |
+
"learning_rate": 3.066079295154185e-06,
|
13384 |
+
"loss": 0.8622,
|
13385 |
+
"step": 21650
|
13386 |
+
},
|
13387 |
+
{
|
13388 |
+
"epoch": 142.5,
|
13389 |
+
"learning_rate": 3.039647577092511e-06,
|
13390 |
+
"loss": 0.8759,
|
13391 |
+
"step": 21660
|
13392 |
+
},
|
13393 |
+
{
|
13394 |
+
"epoch": 142.56,
|
13395 |
+
"learning_rate": 3.013215859030837e-06,
|
13396 |
+
"loss": 0.8428,
|
13397 |
+
"step": 21670
|
13398 |
+
},
|
13399 |
+
{
|
13400 |
+
"epoch": 142.63,
|
13401 |
+
"learning_rate": 2.9867841409691632e-06,
|
13402 |
+
"loss": 0.849,
|
13403 |
+
"step": 21680
|
13404 |
+
},
|
13405 |
+
{
|
13406 |
+
"epoch": 142.69,
|
13407 |
+
"learning_rate": 2.960352422907489e-06,
|
13408 |
+
"loss": 0.8815,
|
13409 |
+
"step": 21690
|
13410 |
+
},
|
13411 |
+
{
|
13412 |
+
"epoch": 142.76,
|
13413 |
+
"learning_rate": 2.933920704845815e-06,
|
13414 |
+
"loss": 0.9493,
|
13415 |
+
"step": 21700
|
13416 |
+
},
|
13417 |
+
{
|
13418 |
+
"epoch": 142.82,
|
13419 |
+
"learning_rate": 2.9074889867841414e-06,
|
13420 |
+
"loss": 0.814,
|
13421 |
+
"step": 21710
|
13422 |
+
},
|
13423 |
+
{
|
13424 |
+
"epoch": 142.89,
|
13425 |
+
"learning_rate": 2.8810572687224673e-06,
|
13426 |
+
"loss": 0.8421,
|
13427 |
+
"step": 21720
|
13428 |
+
},
|
13429 |
+
{
|
13430 |
+
"epoch": 142.96,
|
13431 |
+
"learning_rate": 2.854625550660793e-06,
|
13432 |
+
"loss": 0.853,
|
13433 |
+
"step": 21730
|
13434 |
+
},
|
13435 |
+
{
|
13436 |
+
"epoch": 143.03,
|
13437 |
+
"learning_rate": 2.828193832599119e-06,
|
13438 |
+
"loss": 1.0032,
|
13439 |
+
"step": 21740
|
13440 |
+
},
|
13441 |
+
{
|
13442 |
+
"epoch": 143.09,
|
13443 |
+
"learning_rate": 2.801762114537445e-06,
|
13444 |
+
"loss": 0.826,
|
13445 |
+
"step": 21750
|
13446 |
+
},
|
13447 |
+
{
|
13448 |
+
"epoch": 143.16,
|
13449 |
+
"learning_rate": 2.775330396475771e-06,
|
13450 |
+
"loss": 0.8516,
|
13451 |
+
"step": 21760
|
13452 |
+
},
|
13453 |
+
{
|
13454 |
+
"epoch": 143.22,
|
13455 |
+
"learning_rate": 2.7488986784140968e-06,
|
13456 |
+
"loss": 0.8449,
|
13457 |
+
"step": 21770
|
13458 |
+
},
|
13459 |
+
{
|
13460 |
+
"epoch": 143.29,
|
13461 |
+
"learning_rate": 2.722466960352423e-06,
|
13462 |
+
"loss": 0.9136,
|
13463 |
+
"step": 21780
|
13464 |
+
},
|
13465 |
+
{
|
13466 |
+
"epoch": 143.35,
|
13467 |
+
"learning_rate": 2.696035242290749e-06,
|
13468 |
+
"loss": 0.8496,
|
13469 |
+
"step": 21790
|
13470 |
+
},
|
13471 |
+
{
|
13472 |
+
"epoch": 143.42,
|
13473 |
+
"learning_rate": 2.669603524229075e-06,
|
13474 |
+
"loss": 0.8406,
|
13475 |
+
"step": 21800
|
13476 |
+
},
|
13477 |
+
{
|
13478 |
+
"epoch": 143.48,
|
13479 |
+
"learning_rate": 2.643171806167401e-06,
|
13480 |
+
"loss": 0.8884,
|
13481 |
+
"step": 21810
|
13482 |
+
},
|
13483 |
+
{
|
13484 |
+
"epoch": 143.55,
|
13485 |
+
"learning_rate": 2.616740088105727e-06,
|
13486 |
+
"loss": 0.8791,
|
13487 |
+
"step": 21820
|
13488 |
+
},
|
13489 |
+
{
|
13490 |
+
"epoch": 143.62,
|
13491 |
+
"learning_rate": 2.5903083700440526e-06,
|
13492 |
+
"loss": 0.8221,
|
13493 |
+
"step": 21830
|
13494 |
+
},
|
13495 |
+
{
|
13496 |
+
"epoch": 143.68,
|
13497 |
+
"learning_rate": 2.563876651982379e-06,
|
13498 |
+
"loss": 0.8445,
|
13499 |
+
"step": 21840
|
13500 |
+
},
|
13501 |
+
{
|
13502 |
+
"epoch": 143.75,
|
13503 |
+
"learning_rate": 2.537444933920705e-06,
|
13504 |
+
"loss": 0.9059,
|
13505 |
+
"step": 21850
|
13506 |
+
},
|
13507 |
+
{
|
13508 |
+
"epoch": 143.81,
|
13509 |
+
"learning_rate": 2.5110132158590307e-06,
|
13510 |
+
"loss": 0.8484,
|
13511 |
+
"step": 21860
|
13512 |
+
},
|
13513 |
+
{
|
13514 |
+
"epoch": 143.88,
|
13515 |
+
"learning_rate": 2.484581497797357e-06,
|
13516 |
+
"loss": 0.8594,
|
13517 |
+
"step": 21870
|
13518 |
+
},
|
13519 |
+
{
|
13520 |
+
"epoch": 143.94,
|
13521 |
+
"learning_rate": 2.458149779735683e-06,
|
13522 |
+
"loss": 0.8631,
|
13523 |
+
"step": 21880
|
13524 |
+
},
|
13525 |
+
{
|
13526 |
+
"epoch": 144.01,
|
13527 |
+
"learning_rate": 2.431718061674009e-06,
|
13528 |
+
"loss": 0.9621,
|
13529 |
+
"step": 21890
|
13530 |
+
},
|
13531 |
+
{
|
13532 |
+
"epoch": 144.08,
|
13533 |
+
"learning_rate": 2.405286343612335e-06,
|
13534 |
+
"loss": 0.8186,
|
13535 |
+
"step": 21900
|
13536 |
+
},
|
13537 |
+
{
|
13538 |
+
"epoch": 144.14,
|
13539 |
+
"learning_rate": 2.378854625550661e-06,
|
13540 |
+
"loss": 0.8318,
|
13541 |
+
"step": 21910
|
13542 |
+
},
|
13543 |
+
{
|
13544 |
+
"epoch": 144.21,
|
13545 |
+
"learning_rate": 2.3524229074889865e-06,
|
13546 |
+
"loss": 0.8535,
|
13547 |
+
"step": 21920
|
13548 |
+
},
|
13549 |
+
{
|
13550 |
+
"epoch": 144.27,
|
13551 |
+
"learning_rate": 2.325991189427313e-06,
|
13552 |
+
"loss": 0.9028,
|
13553 |
+
"step": 21930
|
13554 |
+
},
|
13555 |
+
{
|
13556 |
+
"epoch": 144.34,
|
13557 |
+
"learning_rate": 2.2995594713656387e-06,
|
13558 |
+
"loss": 0.8314,
|
13559 |
+
"step": 21940
|
13560 |
+
},
|
13561 |
+
{
|
13562 |
+
"epoch": 144.41,
|
13563 |
+
"learning_rate": 2.2731277533039646e-06,
|
13564 |
+
"loss": 0.8417,
|
13565 |
+
"step": 21950
|
13566 |
+
},
|
13567 |
+
{
|
13568 |
+
"epoch": 144.47,
|
13569 |
+
"learning_rate": 2.249339207048458e-06,
|
13570 |
+
"loss": 0.849,
|
13571 |
+
"step": 21960
|
13572 |
+
},
|
13573 |
+
{
|
13574 |
+
"epoch": 144.54,
|
13575 |
+
"learning_rate": 2.222907488986784e-06,
|
13576 |
+
"loss": 0.9065,
|
13577 |
+
"step": 21970
|
13578 |
+
},
|
13579 |
+
{
|
13580 |
+
"epoch": 144.6,
|
13581 |
+
"learning_rate": 2.1964757709251104e-06,
|
13582 |
+
"loss": 0.8296,
|
13583 |
+
"step": 21980
|
13584 |
+
},
|
13585 |
+
{
|
13586 |
+
"epoch": 144.67,
|
13587 |
+
"learning_rate": 2.1700440528634363e-06,
|
13588 |
+
"loss": 0.8638,
|
13589 |
+
"step": 21990
|
13590 |
+
},
|
13591 |
+
{
|
13592 |
+
"epoch": 144.73,
|
13593 |
+
"learning_rate": 2.143612334801762e-06,
|
13594 |
+
"loss": 0.8705,
|
13595 |
+
"step": 22000
|
13596 |
+
},
|
13597 |
+
{
|
13598 |
+
"epoch": 144.73,
|
13599 |
+
"eval_loss": 0.34444093704223633,
|
13600 |
+
"eval_runtime": 318.8078,
|
13601 |
+
"eval_samples_per_second": 22.302,
|
13602 |
+
"eval_steps_per_second": 1.396,
|
13603 |
+
"eval_wer": 0.34743056227620245,
|
13604 |
+
"step": 22000
|
13605 |
+
},
|
13606 |
+
{
|
13607 |
+
"epoch": 144.8,
|
13608 |
+
"learning_rate": 2.117180616740088e-06,
|
13609 |
+
"loss": 0.8493,
|
13610 |
+
"step": 22010
|
13611 |
+
},
|
13612 |
+
{
|
13613 |
+
"epoch": 144.86,
|
13614 |
+
"learning_rate": 2.090748898678414e-06,
|
13615 |
+
"loss": 0.8297,
|
13616 |
+
"step": 22020
|
13617 |
+
},
|
13618 |
+
{
|
13619 |
+
"epoch": 144.93,
|
13620 |
+
"learning_rate": 2.06431718061674e-06,
|
13621 |
+
"loss": 0.8733,
|
13622 |
+
"step": 22030
|
13623 |
+
},
|
13624 |
+
{
|
13625 |
+
"epoch": 145.0,
|
13626 |
+
"learning_rate": 2.037885462555066e-06,
|
13627 |
+
"loss": 0.8968,
|
13628 |
+
"step": 22040
|
13629 |
+
},
|
13630 |
+
{
|
13631 |
+
"epoch": 145.07,
|
13632 |
+
"learning_rate": 2.011453744493392e-06,
|
13633 |
+
"loss": 0.9201,
|
13634 |
+
"step": 22050
|
13635 |
+
},
|
13636 |
+
{
|
13637 |
+
"epoch": 145.13,
|
13638 |
+
"learning_rate": 1.985022026431718e-06,
|
13639 |
+
"loss": 0.8446,
|
13640 |
+
"step": 22060
|
13641 |
+
},
|
13642 |
+
{
|
13643 |
+
"epoch": 145.2,
|
13644 |
+
"learning_rate": 1.9585903083700443e-06,
|
13645 |
+
"loss": 0.849,
|
13646 |
+
"step": 22070
|
13647 |
+
},
|
13648 |
+
{
|
13649 |
+
"epoch": 145.26,
|
13650 |
+
"learning_rate": 1.93215859030837e-06,
|
13651 |
+
"loss": 0.91,
|
13652 |
+
"step": 22080
|
13653 |
+
},
|
13654 |
+
{
|
13655 |
+
"epoch": 145.33,
|
13656 |
+
"learning_rate": 1.9057268722466963e-06,
|
13657 |
+
"loss": 0.8113,
|
13658 |
+
"step": 22090
|
13659 |
+
},
|
13660 |
+
{
|
13661 |
+
"epoch": 145.39,
|
13662 |
+
"learning_rate": 1.8792951541850218e-06,
|
13663 |
+
"loss": 0.8432,
|
13664 |
+
"step": 22100
|
13665 |
+
},
|
13666 |
+
{
|
13667 |
+
"epoch": 145.46,
|
13668 |
+
"learning_rate": 1.8528634361233481e-06,
|
13669 |
+
"loss": 0.8635,
|
13670 |
+
"step": 22110
|
13671 |
+
},
|
13672 |
+
{
|
13673 |
+
"epoch": 145.52,
|
13674 |
+
"learning_rate": 1.8264317180616742e-06,
|
13675 |
+
"loss": 0.8872,
|
13676 |
+
"step": 22120
|
13677 |
+
},
|
13678 |
+
{
|
13679 |
+
"epoch": 145.59,
|
13680 |
+
"learning_rate": 1.8e-06,
|
13681 |
+
"loss": 0.8099,
|
13682 |
+
"step": 22130
|
13683 |
+
},
|
13684 |
+
{
|
13685 |
+
"epoch": 145.65,
|
13686 |
+
"learning_rate": 1.773568281938326e-06,
|
13687 |
+
"loss": 0.862,
|
13688 |
+
"step": 22140
|
13689 |
+
},
|
13690 |
+
{
|
13691 |
+
"epoch": 145.72,
|
13692 |
+
"learning_rate": 1.7471365638766521e-06,
|
13693 |
+
"loss": 0.8395,
|
13694 |
+
"step": 22150
|
13695 |
+
},
|
13696 |
+
{
|
13697 |
+
"epoch": 145.79,
|
13698 |
+
"learning_rate": 1.720704845814978e-06,
|
13699 |
+
"loss": 0.9084,
|
13700 |
+
"step": 22160
|
13701 |
+
},
|
13702 |
+
{
|
13703 |
+
"epoch": 145.85,
|
13704 |
+
"learning_rate": 1.694273127753304e-06,
|
13705 |
+
"loss": 0.8402,
|
13706 |
+
"step": 22170
|
13707 |
+
},
|
13708 |
+
{
|
13709 |
+
"epoch": 145.92,
|
13710 |
+
"learning_rate": 1.66784140969163e-06,
|
13711 |
+
"loss": 0.8627,
|
13712 |
+
"step": 22180
|
13713 |
+
},
|
13714 |
+
{
|
13715 |
+
"epoch": 145.98,
|
13716 |
+
"learning_rate": 1.641409691629956e-06,
|
13717 |
+
"loss": 0.8402,
|
13718 |
+
"step": 22190
|
13719 |
+
},
|
13720 |
+
{
|
13721 |
+
"epoch": 146.05,
|
13722 |
+
"learning_rate": 1.614977973568282e-06,
|
13723 |
+
"loss": 0.9459,
|
13724 |
+
"step": 22200
|
13725 |
+
},
|
13726 |
+
{
|
13727 |
+
"epoch": 146.12,
|
13728 |
+
"learning_rate": 1.588546255506608e-06,
|
13729 |
+
"loss": 0.8286,
|
13730 |
+
"step": 22210
|
13731 |
+
},
|
13732 |
+
{
|
13733 |
+
"epoch": 146.18,
|
13734 |
+
"learning_rate": 1.5621145374449339e-06,
|
13735 |
+
"loss": 0.8557,
|
13736 |
+
"step": 22220
|
13737 |
+
},
|
13738 |
+
{
|
13739 |
+
"epoch": 146.25,
|
13740 |
+
"learning_rate": 1.53568281938326e-06,
|
13741 |
+
"loss": 0.8967,
|
13742 |
+
"step": 22230
|
13743 |
+
},
|
13744 |
+
{
|
13745 |
+
"epoch": 146.31,
|
13746 |
+
"learning_rate": 1.5092511013215859e-06,
|
13747 |
+
"loss": 0.8315,
|
13748 |
+
"step": 22240
|
13749 |
+
},
|
13750 |
+
{
|
13751 |
+
"epoch": 146.38,
|
13752 |
+
"learning_rate": 1.482819383259912e-06,
|
13753 |
+
"loss": 0.8259,
|
13754 |
+
"step": 22250
|
13755 |
+
},
|
13756 |
+
{
|
13757 |
+
"epoch": 146.45,
|
13758 |
+
"learning_rate": 1.456387665198238e-06,
|
13759 |
+
"loss": 0.8549,
|
13760 |
+
"step": 22260
|
13761 |
+
},
|
13762 |
+
{
|
13763 |
+
"epoch": 146.51,
|
13764 |
+
"learning_rate": 1.4299559471365638e-06,
|
13765 |
+
"loss": 0.8958,
|
13766 |
+
"step": 22270
|
13767 |
+
},
|
13768 |
+
{
|
13769 |
+
"epoch": 146.58,
|
13770 |
+
"learning_rate": 1.4035242290748899e-06,
|
13771 |
+
"loss": 0.8303,
|
13772 |
+
"step": 22280
|
13773 |
+
},
|
13774 |
+
{
|
13775 |
+
"epoch": 146.64,
|
13776 |
+
"learning_rate": 1.377092511013216e-06,
|
13777 |
+
"loss": 0.8444,
|
13778 |
+
"step": 22290
|
13779 |
+
},
|
13780 |
+
{
|
13781 |
+
"epoch": 146.71,
|
13782 |
+
"learning_rate": 1.350660792951542e-06,
|
13783 |
+
"loss": 0.8439,
|
13784 |
+
"step": 22300
|
13785 |
+
},
|
13786 |
+
{
|
13787 |
+
"epoch": 146.77,
|
13788 |
+
"learning_rate": 1.3242290748898678e-06,
|
13789 |
+
"loss": 0.9093,
|
13790 |
+
"step": 22310
|
13791 |
+
},
|
13792 |
+
{
|
13793 |
+
"epoch": 146.84,
|
13794 |
+
"learning_rate": 1.297797356828194e-06,
|
13795 |
+
"loss": 0.8584,
|
13796 |
+
"step": 22320
|
13797 |
+
},
|
13798 |
+
{
|
13799 |
+
"epoch": 146.9,
|
13800 |
+
"learning_rate": 1.2713656387665198e-06,
|
13801 |
+
"loss": 0.8541,
|
13802 |
+
"step": 22330
|
13803 |
+
},
|
13804 |
+
{
|
13805 |
+
"epoch": 146.97,
|
13806 |
+
"learning_rate": 1.244933920704846e-06,
|
13807 |
+
"loss": 0.8791,
|
13808 |
+
"step": 22340
|
13809 |
+
},
|
13810 |
+
{
|
13811 |
+
"epoch": 147.04,
|
13812 |
+
"learning_rate": 1.2185022026431718e-06,
|
13813 |
+
"loss": 0.9583,
|
13814 |
+
"step": 22350
|
13815 |
+
},
|
13816 |
+
{
|
13817 |
+
"epoch": 147.1,
|
13818 |
+
"learning_rate": 1.1920704845814977e-06,
|
13819 |
+
"loss": 0.7998,
|
13820 |
+
"step": 22360
|
13821 |
+
},
|
13822 |
+
{
|
13823 |
+
"epoch": 147.17,
|
13824 |
+
"learning_rate": 1.1656387665198238e-06,
|
13825 |
+
"loss": 0.8725,
|
13826 |
+
"step": 22370
|
13827 |
+
},
|
13828 |
+
{
|
13829 |
+
"epoch": 147.24,
|
13830 |
+
"learning_rate": 1.1392070484581497e-06,
|
13831 |
+
"loss": 0.8524,
|
13832 |
+
"step": 22380
|
13833 |
+
},
|
13834 |
+
{
|
13835 |
+
"epoch": 147.3,
|
13836 |
+
"learning_rate": 1.1127753303964758e-06,
|
13837 |
+
"loss": 0.8648,
|
13838 |
+
"step": 22390
|
13839 |
+
},
|
13840 |
+
{
|
13841 |
+
"epoch": 147.37,
|
13842 |
+
"learning_rate": 1.086343612334802e-06,
|
13843 |
+
"loss": 0.8385,
|
13844 |
+
"step": 22400
|
13845 |
+
},
|
13846 |
+
{
|
13847 |
+
"epoch": 147.43,
|
13848 |
+
"learning_rate": 1.0599118942731276e-06,
|
13849 |
+
"loss": 0.8608,
|
13850 |
+
"step": 22410
|
13851 |
+
},
|
13852 |
+
{
|
13853 |
+
"epoch": 147.5,
|
13854 |
+
"learning_rate": 1.0334801762114537e-06,
|
13855 |
+
"loss": 0.8992,
|
13856 |
+
"step": 22420
|
13857 |
+
},
|
13858 |
+
{
|
13859 |
+
"epoch": 147.56,
|
13860 |
+
"learning_rate": 1.0070484581497799e-06,
|
13861 |
+
"loss": 0.8481,
|
13862 |
+
"step": 22430
|
13863 |
+
},
|
13864 |
+
{
|
13865 |
+
"epoch": 147.63,
|
13866 |
+
"learning_rate": 9.806167400881058e-07,
|
13867 |
+
"loss": 0.8521,
|
13868 |
+
"step": 22440
|
13869 |
+
},
|
13870 |
+
{
|
13871 |
+
"epoch": 147.69,
|
13872 |
+
"learning_rate": 9.541850220264317e-07,
|
13873 |
+
"loss": 0.8477,
|
13874 |
+
"step": 22450
|
13875 |
+
},
|
13876 |
+
{
|
13877 |
+
"epoch": 147.76,
|
13878 |
+
"learning_rate": 9.277533039647578e-07,
|
13879 |
+
"loss": 0.9214,
|
13880 |
+
"step": 22460
|
13881 |
+
},
|
13882 |
+
{
|
13883 |
+
"epoch": 147.82,
|
13884 |
+
"learning_rate": 9.013215859030837e-07,
|
13885 |
+
"loss": 0.8459,
|
13886 |
+
"step": 22470
|
13887 |
+
},
|
13888 |
+
{
|
13889 |
+
"epoch": 147.89,
|
13890 |
+
"learning_rate": 8.748898678414098e-07,
|
13891 |
+
"loss": 0.8296,
|
13892 |
+
"step": 22480
|
13893 |
+
},
|
13894 |
+
{
|
13895 |
+
"epoch": 147.96,
|
13896 |
+
"learning_rate": 8.484581497797357e-07,
|
13897 |
+
"loss": 0.8857,
|
13898 |
+
"step": 22490
|
13899 |
+
},
|
13900 |
+
{
|
13901 |
+
"epoch": 148.03,
|
13902 |
+
"learning_rate": 8.220264317180617e-07,
|
13903 |
+
"loss": 0.9869,
|
13904 |
+
"step": 22500
|
13905 |
+
},
|
13906 |
+
{
|
13907 |
+
"epoch": 148.03,
|
13908 |
+
"eval_loss": 0.3441205620765686,
|
13909 |
+
"eval_runtime": 320.0073,
|
13910 |
+
"eval_samples_per_second": 22.218,
|
13911 |
+
"eval_steps_per_second": 1.391,
|
13912 |
+
"eval_wer": 0.3493080421949095,
|
13913 |
+
"step": 22500
|
13914 |
+
},
|
13915 |
+
{
|
13916 |
+
"epoch": 148.09,
|
13917 |
+
"learning_rate": 7.955947136563877e-07,
|
13918 |
+
"loss": 0.8326,
|
13919 |
+
"step": 22510
|
13920 |
+
},
|
13921 |
+
{
|
13922 |
+
"epoch": 148.16,
|
13923 |
+
"learning_rate": 7.691629955947137e-07,
|
13924 |
+
"loss": 0.8538,
|
13925 |
+
"step": 22520
|
13926 |
+
},
|
13927 |
+
{
|
13928 |
+
"epoch": 148.22,
|
13929 |
+
"learning_rate": 7.427312775330397e-07,
|
13930 |
+
"loss": 0.8947,
|
13931 |
+
"step": 22530
|
13932 |
+
},
|
13933 |
+
{
|
13934 |
+
"epoch": 148.29,
|
13935 |
+
"learning_rate": 7.162995594713656e-07,
|
13936 |
+
"loss": 0.9051,
|
13937 |
+
"step": 22540
|
13938 |
+
},
|
13939 |
+
{
|
13940 |
+
"epoch": 148.35,
|
13941 |
+
"learning_rate": 6.898678414096917e-07,
|
13942 |
+
"loss": 0.8197,
|
13943 |
+
"step": 22550
|
13944 |
+
},
|
13945 |
+
{
|
13946 |
+
"epoch": 148.42,
|
13947 |
+
"learning_rate": 6.634361233480176e-07,
|
13948 |
+
"loss": 0.8446,
|
13949 |
+
"step": 22560
|
13950 |
+
},
|
13951 |
+
{
|
13952 |
+
"epoch": 148.48,
|
13953 |
+
"learning_rate": 6.370044052863436e-07,
|
13954 |
+
"loss": 0.9035,
|
13955 |
+
"step": 22570
|
13956 |
+
},
|
13957 |
+
{
|
13958 |
+
"epoch": 148.55,
|
13959 |
+
"learning_rate": 6.105726872246696e-07,
|
13960 |
+
"loss": 0.861,
|
13961 |
+
"step": 22580
|
13962 |
+
},
|
13963 |
+
{
|
13964 |
+
"epoch": 148.62,
|
13965 |
+
"learning_rate": 5.841409691629956e-07,
|
13966 |
+
"loss": 0.8269,
|
13967 |
+
"step": 22590
|
13968 |
+
},
|
13969 |
+
{
|
13970 |
+
"epoch": 148.68,
|
13971 |
+
"learning_rate": 5.577092511013216e-07,
|
13972 |
+
"loss": 0.8597,
|
13973 |
+
"step": 22600
|
13974 |
+
},
|
13975 |
+
{
|
13976 |
+
"epoch": 148.75,
|
13977 |
+
"learning_rate": 5.312775330396475e-07,
|
13978 |
+
"loss": 0.8928,
|
13979 |
+
"step": 22610
|
13980 |
+
},
|
13981 |
+
{
|
13982 |
+
"epoch": 148.81,
|
13983 |
+
"learning_rate": 5.048458149779736e-07,
|
13984 |
+
"loss": 0.8236,
|
13985 |
+
"step": 22620
|
13986 |
+
},
|
13987 |
+
{
|
13988 |
+
"epoch": 148.88,
|
13989 |
+
"learning_rate": 4.784140969162995e-07,
|
13990 |
+
"loss": 0.8326,
|
13991 |
+
"step": 22630
|
13992 |
+
},
|
13993 |
+
{
|
13994 |
+
"epoch": 148.94,
|
13995 |
+
"learning_rate": 4.519823788546256e-07,
|
13996 |
+
"loss": 0.858,
|
13997 |
+
"step": 22640
|
13998 |
+
},
|
13999 |
+
{
|
14000 |
+
"epoch": 149.01,
|
14001 |
+
"learning_rate": 4.2555066079295154e-07,
|
14002 |
+
"loss": 0.9665,
|
14003 |
+
"step": 22650
|
14004 |
+
},
|
14005 |
+
{
|
14006 |
+
"epoch": 149.08,
|
14007 |
+
"learning_rate": 3.9911894273127755e-07,
|
14008 |
+
"loss": 0.8098,
|
14009 |
+
"step": 22660
|
14010 |
+
},
|
14011 |
+
{
|
14012 |
+
"epoch": 149.14,
|
14013 |
+
"learning_rate": 3.726872246696035e-07,
|
14014 |
+
"loss": 0.8435,
|
14015 |
+
"step": 22670
|
14016 |
+
},
|
14017 |
+
{
|
14018 |
+
"epoch": 149.21,
|
14019 |
+
"learning_rate": 3.462555066079295e-07,
|
14020 |
+
"loss": 0.8414,
|
14021 |
+
"step": 22680
|
14022 |
+
},
|
14023 |
+
{
|
14024 |
+
"epoch": 149.27,
|
14025 |
+
"learning_rate": 3.1982378854625556e-07,
|
14026 |
+
"loss": 0.9122,
|
14027 |
+
"step": 22690
|
14028 |
+
},
|
14029 |
+
{
|
14030 |
+
"epoch": 149.34,
|
14031 |
+
"learning_rate": 2.933920704845815e-07,
|
14032 |
+
"loss": 0.8136,
|
14033 |
+
"step": 22700
|
14034 |
+
},
|
14035 |
+
{
|
14036 |
+
"epoch": 149.41,
|
14037 |
+
"learning_rate": 2.669603524229075e-07,
|
14038 |
+
"loss": 0.856,
|
14039 |
+
"step": 22710
|
14040 |
+
},
|
14041 |
+
{
|
14042 |
+
"epoch": 149.47,
|
14043 |
+
"learning_rate": 2.4052863436123347e-07,
|
14044 |
+
"loss": 0.8759,
|
14045 |
+
"step": 22720
|
14046 |
+
},
|
14047 |
+
{
|
14048 |
+
"epoch": 149.54,
|
14049 |
+
"learning_rate": 2.1409691629955948e-07,
|
14050 |
+
"loss": 0.884,
|
14051 |
+
"step": 22730
|
14052 |
+
},
|
14053 |
+
{
|
14054 |
+
"epoch": 149.6,
|
14055 |
+
"learning_rate": 1.8766519823788548e-07,
|
14056 |
+
"loss": 0.8404,
|
14057 |
+
"step": 22740
|
14058 |
+
},
|
14059 |
+
{
|
14060 |
+
"epoch": 149.67,
|
14061 |
+
"learning_rate": 1.6123348017621146e-07,
|
14062 |
+
"loss": 0.8286,
|
14063 |
+
"step": 22750
|
14064 |
+
},
|
14065 |
+
{
|
14066 |
+
"epoch": 149.73,
|
14067 |
+
"learning_rate": 1.3480176211453744e-07,
|
14068 |
+
"loss": 0.8571,
|
14069 |
+
"step": 22760
|
14070 |
+
},
|
14071 |
+
{
|
14072 |
+
"epoch": 149.8,
|
14073 |
+
"learning_rate": 1.0837004405286344e-07,
|
14074 |
+
"loss": 0.8685,
|
14075 |
+
"step": 22770
|
14076 |
+
},
|
14077 |
+
{
|
14078 |
+
"epoch": 149.86,
|
14079 |
+
"learning_rate": 8.193832599118942e-08,
|
14080 |
+
"loss": 0.8315,
|
14081 |
+
"step": 22780
|
14082 |
+
},
|
14083 |
+
{
|
14084 |
+
"epoch": 149.93,
|
14085 |
+
"learning_rate": 5.550660792951542e-08,
|
14086 |
+
"loss": 0.8556,
|
14087 |
+
"step": 22790
|
14088 |
+
},
|
14089 |
+
{
|
14090 |
+
"epoch": 150.0,
|
14091 |
+
"learning_rate": 2.907488986784141e-08,
|
14092 |
+
"loss": 0.8774,
|
14093 |
+
"step": 22800
|
14094 |
+
},
|
14095 |
+
{
|
14096 |
+
"epoch": 150.0,
|
14097 |
+
"step": 22800,
|
14098 |
+
"total_flos": 3.150278594946483e+20,
|
14099 |
+
"train_loss": 0.16461168266179269,
|
14100 |
+
"train_runtime": 24582.5742,
|
14101 |
+
"train_samples_per_second": 119.176,
|
14102 |
+
"train_steps_per_second": 0.927
|
14103 |
}
|
14104 |
],
|
14105 |
+
"max_steps": 22800,
|
14106 |
+
"num_train_epochs": 150,
|
14107 |
+
"total_flos": 3.150278594946483e+20,
|
14108 |
"trial_name": null,
|
14109 |
"trial_params": null
|
14110 |
}
|