|
{ |
|
"best_metric": 0.9886363636363636, |
|
"best_model_checkpoint": "swin-large-patch4-window12-384-in22k-finetuned-batch8/checkpoint-98", |
|
"epoch": 2.984771573604061, |
|
"eval_steps": 500, |
|
"global_step": 147, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.20304568527918782, |
|
"grad_norm": 6.88034200668335, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 0.643, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.40609137055837563, |
|
"grad_norm": 19.021066665649414, |
|
"learning_rate": 4.810606060606061e-05, |
|
"loss": 0.3831, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.6091370558375635, |
|
"grad_norm": 9.288772583007812, |
|
"learning_rate": 4.431818181818182e-05, |
|
"loss": 0.1721, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.8121827411167513, |
|
"grad_norm": 17.481571197509766, |
|
"learning_rate": 4.053030303030303e-05, |
|
"loss": 0.2123, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.9949238578680203, |
|
"eval_accuracy": 0.9659090909090909, |
|
"eval_loss": 0.09163513779640198, |
|
"eval_runtime": 15.1203, |
|
"eval_samples_per_second": 11.64, |
|
"eval_steps_per_second": 1.455, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 1.015228426395939, |
|
"grad_norm": 8.028657913208008, |
|
"learning_rate": 3.6742424242424246e-05, |
|
"loss": 0.2193, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.218274111675127, |
|
"grad_norm": 14.979452133178711, |
|
"learning_rate": 3.295454545454545e-05, |
|
"loss": 0.2239, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.4213197969543148, |
|
"grad_norm": 8.288687705993652, |
|
"learning_rate": 2.916666666666667e-05, |
|
"loss": 0.1636, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.6243654822335025, |
|
"grad_norm": 9.035058975219727, |
|
"learning_rate": 2.537878787878788e-05, |
|
"loss": 0.1848, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.8274111675126905, |
|
"grad_norm": 18.96419906616211, |
|
"learning_rate": 2.1590909090909093e-05, |
|
"loss": 0.1613, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.9898477157360406, |
|
"eval_accuracy": 0.9886363636363636, |
|
"eval_loss": 0.04295675456523895, |
|
"eval_runtime": 15.0722, |
|
"eval_samples_per_second": 11.677, |
|
"eval_steps_per_second": 1.46, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 2.030456852791878, |
|
"grad_norm": 5.229619979858398, |
|
"learning_rate": 1.7803030303030303e-05, |
|
"loss": 0.1375, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.233502538071066, |
|
"grad_norm": 6.713926792144775, |
|
"learning_rate": 1.4015151515151515e-05, |
|
"loss": 0.1352, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 2.436548223350254, |
|
"grad_norm": 14.589942932128906, |
|
"learning_rate": 1.0227272727272729e-05, |
|
"loss": 0.1208, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.6395939086294415, |
|
"grad_norm": 1.9984550476074219, |
|
"learning_rate": 6.43939393939394e-06, |
|
"loss": 0.107, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.8426395939086295, |
|
"grad_norm": 10.917881965637207, |
|
"learning_rate": 2.651515151515152e-06, |
|
"loss": 0.116, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.984771573604061, |
|
"eval_accuracy": 0.9886363636363636, |
|
"eval_loss": 0.034594591706991196, |
|
"eval_runtime": 15.0628, |
|
"eval_samples_per_second": 11.684, |
|
"eval_steps_per_second": 1.461, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 2.984771573604061, |
|
"step": 147, |
|
"total_flos": 2.437168844903547e+18, |
|
"train_loss": 0.20754414107523808, |
|
"train_runtime": 1195.5632, |
|
"train_samples_per_second": 3.955, |
|
"train_steps_per_second": 0.123 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 147, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"total_flos": 2.437168844903547e+18, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|