Training in progress, epoch 111, checkpoint
Browse files
last-checkpoint/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 166496880
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:292c73693098875692abcb4fe84b6b84cc28407036bb27a962300192d5d9f048
|
3 |
size 166496880
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 330495866
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c70174351bbc61e80037e803568d9a6f4338016d31c52d7c41f2228630dfa92a
|
3 |
size 330495866
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0c387bfd9a993f978958650fcc5a25ab904bd56418ed7491f4791553eb1382ae
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:91165b4467d36cadf91977e383e5a88c4e25b217f011b6ddd68eb9dee41f87cb
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": 0.2962397336959839,
|
3 |
"best_model_checkpoint": "chickens-composite-201616161616-150-epochs-wo-transform/checkpoint-45000",
|
4 |
-
"epoch":
|
5 |
"eval_steps": 500,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -15698,6 +15698,151 @@
|
|
15698 |
"eval_samples_per_second": 7.28,
|
15699 |
"eval_steps_per_second": 0.946,
|
15700 |
"step": 55000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15701 |
}
|
15702 |
],
|
15703 |
"logging_steps": 30,
|
@@ -15717,7 +15862,7 @@
|
|
15717 |
"attributes": {}
|
15718 |
}
|
15719 |
},
|
15720 |
-
"total_flos": 1.
|
15721 |
"train_batch_size": 2,
|
15722 |
"trial_name": null,
|
15723 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": 0.2962397336959839,
|
3 |
"best_model_checkpoint": "chickens-composite-201616161616-150-epochs-wo-transform/checkpoint-45000",
|
4 |
+
"epoch": 111.0,
|
5 |
"eval_steps": 500,
|
6 |
+
"global_step": 55500,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
15698 |
"eval_samples_per_second": 7.28,
|
15699 |
"eval_steps_per_second": 0.946,
|
15700 |
"step": 55000
|
15701 |
+
},
|
15702 |
+
{
|
15703 |
+
"epoch": 110.04,
|
15704 |
+
"grad_norm": 35.87388610839844,
|
15705 |
+
"learning_rate": 1.651235264858177e-06,
|
15706 |
+
"loss": 0.2443,
|
15707 |
+
"step": 55020
|
15708 |
+
},
|
15709 |
+
{
|
15710 |
+
"epoch": 110.1,
|
15711 |
+
"grad_norm": 47.754180908203125,
|
15712 |
+
"learning_rate": 1.6465721173164e-06,
|
15713 |
+
"loss": 0.2129,
|
15714 |
+
"step": 55050
|
15715 |
+
},
|
15716 |
+
{
|
15717 |
+
"epoch": 110.16,
|
15718 |
+
"grad_norm": 26.22478675842285,
|
15719 |
+
"learning_rate": 1.6419142652949793e-06,
|
15720 |
+
"loss": 0.1971,
|
15721 |
+
"step": 55080
|
15722 |
+
},
|
15723 |
+
{
|
15724 |
+
"epoch": 110.22,
|
15725 |
+
"grad_norm": 51.385337829589844,
|
15726 |
+
"learning_rate": 1.6372617161493014e-06,
|
15727 |
+
"loss": 0.2324,
|
15728 |
+
"step": 55110
|
15729 |
+
},
|
15730 |
+
{
|
15731 |
+
"epoch": 110.28,
|
15732 |
+
"grad_norm": 54.496002197265625,
|
15733 |
+
"learning_rate": 1.6326144772263752e-06,
|
15734 |
+
"loss": 0.2315,
|
15735 |
+
"step": 55140
|
15736 |
+
},
|
15737 |
+
{
|
15738 |
+
"epoch": 110.34,
|
15739 |
+
"grad_norm": 73.95658111572266,
|
15740 |
+
"learning_rate": 1.627972555864824e-06,
|
15741 |
+
"loss": 0.2268,
|
15742 |
+
"step": 55170
|
15743 |
+
},
|
15744 |
+
{
|
15745 |
+
"epoch": 110.4,
|
15746 |
+
"grad_norm": 56.15229415893555,
|
15747 |
+
"learning_rate": 1.6233359593948777e-06,
|
15748 |
+
"loss": 0.1911,
|
15749 |
+
"step": 55200
|
15750 |
+
},
|
15751 |
+
{
|
15752 |
+
"epoch": 110.46,
|
15753 |
+
"grad_norm": 26.1507511138916,
|
15754 |
+
"learning_rate": 1.618704695138353e-06,
|
15755 |
+
"loss": 0.2455,
|
15756 |
+
"step": 55230
|
15757 |
+
},
|
15758 |
+
{
|
15759 |
+
"epoch": 110.52,
|
15760 |
+
"grad_norm": 73.8873291015625,
|
15761 |
+
"learning_rate": 1.6140787704086502e-06,
|
15762 |
+
"loss": 0.1911,
|
15763 |
+
"step": 55260
|
15764 |
+
},
|
15765 |
+
{
|
15766 |
+
"epoch": 110.58,
|
15767 |
+
"grad_norm": 101.59705352783203,
|
15768 |
+
"learning_rate": 1.6094581925107356e-06,
|
15769 |
+
"loss": 0.2273,
|
15770 |
+
"step": 55290
|
15771 |
+
},
|
15772 |
+
{
|
15773 |
+
"epoch": 110.64,
|
15774 |
+
"grad_norm": 31.966201782226562,
|
15775 |
+
"learning_rate": 1.6048429687411294e-06,
|
15776 |
+
"loss": 0.2295,
|
15777 |
+
"step": 55320
|
15778 |
+
},
|
15779 |
+
{
|
15780 |
+
"epoch": 110.7,
|
15781 |
+
"grad_norm": 60.9480094909668,
|
15782 |
+
"learning_rate": 1.600233106387904e-06,
|
15783 |
+
"loss": 0.2284,
|
15784 |
+
"step": 55350
|
15785 |
+
},
|
15786 |
+
{
|
15787 |
+
"epoch": 110.76,
|
15788 |
+
"grad_norm": 65.5320816040039,
|
15789 |
+
"learning_rate": 1.5956286127306591e-06,
|
15790 |
+
"loss": 0.2274,
|
15791 |
+
"step": 55380
|
15792 |
+
},
|
15793 |
+
{
|
15794 |
+
"epoch": 110.82,
|
15795 |
+
"grad_norm": 67.60034942626953,
|
15796 |
+
"learning_rate": 1.591029495040518e-06,
|
15797 |
+
"loss": 0.2207,
|
15798 |
+
"step": 55410
|
15799 |
+
},
|
15800 |
+
{
|
15801 |
+
"epoch": 110.88,
|
15802 |
+
"grad_norm": 134.33175659179688,
|
15803 |
+
"learning_rate": 1.586435760580118e-06,
|
15804 |
+
"loss": 0.2187,
|
15805 |
+
"step": 55440
|
15806 |
+
},
|
15807 |
+
{
|
15808 |
+
"epoch": 110.94,
|
15809 |
+
"grad_norm": 66.95919799804688,
|
15810 |
+
"learning_rate": 1.5818474166035907e-06,
|
15811 |
+
"loss": 0.2397,
|
15812 |
+
"step": 55470
|
15813 |
+
},
|
15814 |
+
{
|
15815 |
+
"epoch": 111.0,
|
15816 |
+
"grad_norm": 59.430965423583984,
|
15817 |
+
"learning_rate": 1.5772644703565564e-06,
|
15818 |
+
"loss": 0.2212,
|
15819 |
+
"step": 55500
|
15820 |
+
},
|
15821 |
+
{
|
15822 |
+
"epoch": 111.0,
|
15823 |
+
"eval_loss": 0.31890785694122314,
|
15824 |
+
"eval_map": 0.8013,
|
15825 |
+
"eval_map_50": 0.9453,
|
15826 |
+
"eval_map_75": 0.8992,
|
15827 |
+
"eval_map_chicken": 0.8086,
|
15828 |
+
"eval_map_duck": 0.7247,
|
15829 |
+
"eval_map_large": 0.7895,
|
15830 |
+
"eval_map_medium": 0.8062,
|
15831 |
+
"eval_map_plant": 0.8707,
|
15832 |
+
"eval_map_small": 0.3336,
|
15833 |
+
"eval_mar_1": 0.3229,
|
15834 |
+
"eval_mar_10": 0.8357,
|
15835 |
+
"eval_mar_100": 0.8399,
|
15836 |
+
"eval_mar_100_chicken": 0.8488,
|
15837 |
+
"eval_mar_100_duck": 0.7742,
|
15838 |
+
"eval_mar_100_plant": 0.8967,
|
15839 |
+
"eval_mar_large": 0.8245,
|
15840 |
+
"eval_mar_medium": 0.8465,
|
15841 |
+
"eval_mar_small": 0.3881,
|
15842 |
+
"eval_runtime": 11.8276,
|
15843 |
+
"eval_samples_per_second": 8.455,
|
15844 |
+
"eval_steps_per_second": 1.099,
|
15845 |
+
"step": 55500
|
15846 |
}
|
15847 |
],
|
15848 |
"logging_steps": 30,
|
|
|
15862 |
"attributes": {}
|
15863 |
}
|
15864 |
},
|
15865 |
+
"total_flos": 1.90932827664384e+19,
|
15866 |
"train_batch_size": 2,
|
15867 |
"trial_name": null,
|
15868 |
"trial_params": null
|