|
{ |
|
"best_metric": 2.22774338722229, |
|
"best_model_checkpoint": "./outputs/checkpoint-2400", |
|
"epoch": 1.748633879781421, |
|
"eval_steps": 100, |
|
"global_step": 2400, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0002, |
|
"loss": 2.8179, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 2.723731517791748, |
|
"eval_runtime": 208.4619, |
|
"eval_samples_per_second": 30.097, |
|
"eval_steps_per_second": 3.766, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0002, |
|
"loss": 2.6952, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 2.6728808879852295, |
|
"eval_runtime": 204.3784, |
|
"eval_samples_per_second": 30.698, |
|
"eval_steps_per_second": 3.841, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0002, |
|
"loss": 2.6508, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"eval_loss": 2.6387226581573486, |
|
"eval_runtime": 205.5464, |
|
"eval_samples_per_second": 30.524, |
|
"eval_steps_per_second": 3.819, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0002, |
|
"loss": 2.6247, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"eval_loss": 2.6044600009918213, |
|
"eval_runtime": 204.5115, |
|
"eval_samples_per_second": 30.678, |
|
"eval_steps_per_second": 3.838, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002, |
|
"loss": 2.5814, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_loss": 2.5784292221069336, |
|
"eval_runtime": 204.5265, |
|
"eval_samples_per_second": 30.676, |
|
"eval_steps_per_second": 3.838, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002, |
|
"loss": 2.5538, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 2.5523250102996826, |
|
"eval_runtime": 205.1078, |
|
"eval_samples_per_second": 30.589, |
|
"eval_steps_per_second": 3.827, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0002, |
|
"loss": 2.5382, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"eval_loss": 2.5274579524993896, |
|
"eval_runtime": 204.8352, |
|
"eval_samples_per_second": 30.63, |
|
"eval_steps_per_second": 3.832, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.0002, |
|
"loss": 2.5141, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"eval_loss": 2.506831645965576, |
|
"eval_runtime": 204.4886, |
|
"eval_samples_per_second": 30.681, |
|
"eval_steps_per_second": 3.839, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.0002, |
|
"loss": 2.4765, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"eval_loss": 2.484386920928955, |
|
"eval_runtime": 204.734, |
|
"eval_samples_per_second": 30.645, |
|
"eval_steps_per_second": 3.834, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.0002, |
|
"loss": 2.4729, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"eval_loss": 2.4648232460021973, |
|
"eval_runtime": 204.5514, |
|
"eval_samples_per_second": 30.672, |
|
"eval_steps_per_second": 3.838, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.0002, |
|
"loss": 2.468, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"eval_loss": 2.442662477493286, |
|
"eval_runtime": 204.6558, |
|
"eval_samples_per_second": 30.656, |
|
"eval_steps_per_second": 3.836, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.0002, |
|
"loss": 2.4198, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"eval_loss": 2.4222779273986816, |
|
"eval_runtime": 204.5507, |
|
"eval_samples_per_second": 30.672, |
|
"eval_steps_per_second": 3.838, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.0002, |
|
"loss": 2.4326, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"eval_loss": 2.405611753463745, |
|
"eval_runtime": 204.6008, |
|
"eval_samples_per_second": 30.665, |
|
"eval_steps_per_second": 3.837, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.0002, |
|
"loss": 2.3782, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"eval_loss": 2.381305694580078, |
|
"eval_runtime": 204.4744, |
|
"eval_samples_per_second": 30.684, |
|
"eval_steps_per_second": 3.839, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.0002, |
|
"loss": 2.3396, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"eval_loss": 2.3646771907806396, |
|
"eval_runtime": 204.6348, |
|
"eval_samples_per_second": 30.659, |
|
"eval_steps_per_second": 3.836, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.0002, |
|
"loss": 2.3327, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"eval_loss": 2.3456737995147705, |
|
"eval_runtime": 204.4826, |
|
"eval_samples_per_second": 30.682, |
|
"eval_steps_per_second": 3.839, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.0002, |
|
"loss": 2.3308, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"eval_loss": 2.329716920852661, |
|
"eval_runtime": 204.6937, |
|
"eval_samples_per_second": 30.651, |
|
"eval_steps_per_second": 3.835, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.0002, |
|
"loss": 2.2938, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"eval_loss": 2.313258409500122, |
|
"eval_runtime": 204.5658, |
|
"eval_samples_per_second": 30.67, |
|
"eval_steps_per_second": 3.837, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 0.0002, |
|
"loss": 2.2756, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"eval_loss": 2.298874855041504, |
|
"eval_runtime": 204.84, |
|
"eval_samples_per_second": 30.629, |
|
"eval_steps_per_second": 3.832, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 0.0002, |
|
"loss": 2.2724, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"eval_loss": 2.2830755710601807, |
|
"eval_runtime": 204.6478, |
|
"eval_samples_per_second": 30.658, |
|
"eval_steps_per_second": 3.836, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 0.0002, |
|
"loss": 2.252, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"eval_loss": 2.2688376903533936, |
|
"eval_runtime": 204.8236, |
|
"eval_samples_per_second": 30.631, |
|
"eval_steps_per_second": 3.833, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 0.0002, |
|
"loss": 2.2357, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"eval_loss": 2.2520592212677, |
|
"eval_runtime": 204.8048, |
|
"eval_samples_per_second": 30.634, |
|
"eval_steps_per_second": 3.833, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 0.0002, |
|
"loss": 2.2266, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"eval_loss": 2.240290880203247, |
|
"eval_runtime": 204.7035, |
|
"eval_samples_per_second": 30.649, |
|
"eval_steps_per_second": 3.835, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 0.0002, |
|
"loss": 2.2316, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"eval_loss": 2.22774338722229, |
|
"eval_runtime": 204.8482, |
|
"eval_samples_per_second": 30.628, |
|
"eval_steps_per_second": 3.832, |
|
"step": 2400 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 4116, |
|
"num_train_epochs": 3, |
|
"save_steps": 100, |
|
"total_flos": 6.841646122235904e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|