|
{ |
|
"best_metric": 12.700243949890137, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-50", |
|
"epoch": 0.0931098696461825, |
|
"eval_steps": 50, |
|
"global_step": 50, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00186219739292365, |
|
"grad_norm": 7.960118770599365, |
|
"learning_rate": 1.5000000000000002e-07, |
|
"loss": 14.0369, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.00186219739292365, |
|
"eval_loss": 13.924460411071777, |
|
"eval_runtime": 12.7592, |
|
"eval_samples_per_second": 70.929, |
|
"eval_steps_per_second": 8.935, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0037243947858473, |
|
"grad_norm": 8.014508247375488, |
|
"learning_rate": 3.0000000000000004e-07, |
|
"loss": 13.7099, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.00558659217877095, |
|
"grad_norm": 8.808537483215332, |
|
"learning_rate": 4.5e-07, |
|
"loss": 13.8903, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0074487895716946, |
|
"grad_norm": 8.004558563232422, |
|
"learning_rate": 6.000000000000001e-07, |
|
"loss": 14.0816, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.00931098696461825, |
|
"grad_norm": 7.907912254333496, |
|
"learning_rate": 7.5e-07, |
|
"loss": 14.4216, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0111731843575419, |
|
"grad_norm": 7.714488506317139, |
|
"learning_rate": 9e-07, |
|
"loss": 13.8048, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01303538175046555, |
|
"grad_norm": 8.773198127746582, |
|
"learning_rate": 1.0500000000000001e-06, |
|
"loss": 13.7771, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0148975791433892, |
|
"grad_norm": 8.656310081481934, |
|
"learning_rate": 1.2000000000000002e-06, |
|
"loss": 13.9204, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.01675977653631285, |
|
"grad_norm": 7.340457439422607, |
|
"learning_rate": 1.35e-06, |
|
"loss": 14.1696, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.0186219739292365, |
|
"grad_norm": 7.3591461181640625, |
|
"learning_rate": 1.5e-06, |
|
"loss": 13.6311, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.020484171322160148, |
|
"grad_norm": 7.934558868408203, |
|
"learning_rate": 1.65e-06, |
|
"loss": 14.0909, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.0223463687150838, |
|
"grad_norm": 8.731240272521973, |
|
"learning_rate": 1.8e-06, |
|
"loss": 13.8947, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.024208566108007448, |
|
"grad_norm": 8.52802562713623, |
|
"learning_rate": 1.95e-06, |
|
"loss": 13.941, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0260707635009311, |
|
"grad_norm": 8.355961799621582, |
|
"learning_rate": 2.1000000000000002e-06, |
|
"loss": 13.9408, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.027932960893854747, |
|
"grad_norm": 8.493719100952148, |
|
"learning_rate": 2.25e-06, |
|
"loss": 13.8167, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0297951582867784, |
|
"grad_norm": 7.872385025024414, |
|
"learning_rate": 2.4000000000000003e-06, |
|
"loss": 13.8842, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.03165735567970205, |
|
"grad_norm": 7.678893566131592, |
|
"learning_rate": 2.55e-06, |
|
"loss": 13.8789, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0335195530726257, |
|
"grad_norm": 8.681936264038086, |
|
"learning_rate": 2.7e-06, |
|
"loss": 13.9825, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.035381750465549346, |
|
"grad_norm": 8.17630672454834, |
|
"learning_rate": 2.8500000000000002e-06, |
|
"loss": 13.7963, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.037243947858473, |
|
"grad_norm": 8.064035415649414, |
|
"learning_rate": 3e-06, |
|
"loss": 13.941, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03910614525139665, |
|
"grad_norm": 8.132279396057129, |
|
"learning_rate": 3.15e-06, |
|
"loss": 13.8393, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.040968342644320296, |
|
"grad_norm": 8.09195613861084, |
|
"learning_rate": 3.3e-06, |
|
"loss": 13.9476, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.04283054003724395, |
|
"grad_norm": 9.464410781860352, |
|
"learning_rate": 3.4500000000000004e-06, |
|
"loss": 13.9283, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.0446927374301676, |
|
"grad_norm": 8.223386764526367, |
|
"learning_rate": 3.6e-06, |
|
"loss": 13.8445, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.04655493482309125, |
|
"grad_norm": 8.496528625488281, |
|
"learning_rate": 3.75e-06, |
|
"loss": 13.9996, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.048417132216014895, |
|
"grad_norm": 7.405032157897949, |
|
"learning_rate": 3.9e-06, |
|
"loss": 13.4262, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.05027932960893855, |
|
"grad_norm": 7.917985916137695, |
|
"learning_rate": 4.05e-06, |
|
"loss": 13.7422, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.0521415270018622, |
|
"grad_norm": 7.779364585876465, |
|
"learning_rate": 4.2000000000000004e-06, |
|
"loss": 13.8355, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.054003724394785846, |
|
"grad_norm": 7.747602462768555, |
|
"learning_rate": 4.35e-06, |
|
"loss": 13.6872, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.055865921787709494, |
|
"grad_norm": 8.12415599822998, |
|
"learning_rate": 4.5e-06, |
|
"loss": 13.8575, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.05772811918063315, |
|
"grad_norm": 7.393126487731934, |
|
"learning_rate": 4.65e-06, |
|
"loss": 13.6684, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.0595903165735568, |
|
"grad_norm": 7.366344451904297, |
|
"learning_rate": 4.800000000000001e-06, |
|
"loss": 13.2535, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.061452513966480445, |
|
"grad_norm": 7.462878227233887, |
|
"learning_rate": 4.95e-06, |
|
"loss": 13.6117, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.0633147113594041, |
|
"grad_norm": 8.132235527038574, |
|
"learning_rate": 5.1e-06, |
|
"loss": 14.0737, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.06517690875232775, |
|
"grad_norm": 8.085715293884277, |
|
"learning_rate": 5.25e-06, |
|
"loss": 13.7491, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0670391061452514, |
|
"grad_norm": 7.527085304260254, |
|
"learning_rate": 5.4e-06, |
|
"loss": 13.4391, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.06890130353817504, |
|
"grad_norm": 7.162932872772217, |
|
"learning_rate": 5.55e-06, |
|
"loss": 13.2994, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.07076350093109869, |
|
"grad_norm": 6.486966133117676, |
|
"learning_rate": 5.7000000000000005e-06, |
|
"loss": 13.5081, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.07262569832402235, |
|
"grad_norm": 7.425045967102051, |
|
"learning_rate": 5.850000000000001e-06, |
|
"loss": 13.1145, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.074487895716946, |
|
"grad_norm": 7.909907817840576, |
|
"learning_rate": 6e-06, |
|
"loss": 13.3482, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07635009310986965, |
|
"grad_norm": 7.439389228820801, |
|
"learning_rate": 6.1499999999999996e-06, |
|
"loss": 13.4463, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.0782122905027933, |
|
"grad_norm": 7.182775974273682, |
|
"learning_rate": 6.3e-06, |
|
"loss": 13.2005, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.08007448789571694, |
|
"grad_norm": 7.072144031524658, |
|
"learning_rate": 6.45e-06, |
|
"loss": 12.8393, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.08193668528864059, |
|
"grad_norm": 7.034002780914307, |
|
"learning_rate": 6.6e-06, |
|
"loss": 12.8302, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.08379888268156424, |
|
"grad_norm": 6.952661037445068, |
|
"learning_rate": 6.750000000000001e-06, |
|
"loss": 13.1515, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.0856610800744879, |
|
"grad_norm": 7.125630855560303, |
|
"learning_rate": 6.900000000000001e-06, |
|
"loss": 13.0473, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.08752327746741155, |
|
"grad_norm": 7.452700138092041, |
|
"learning_rate": 7.049999999999999e-06, |
|
"loss": 13.0773, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.0893854748603352, |
|
"grad_norm": 7.248137474060059, |
|
"learning_rate": 7.2e-06, |
|
"loss": 13.0051, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.09124767225325885, |
|
"grad_norm": 6.84269905090332, |
|
"learning_rate": 7.35e-06, |
|
"loss": 12.6788, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.0931098696461825, |
|
"grad_norm": 6.934720993041992, |
|
"learning_rate": 7.5e-06, |
|
"loss": 12.7921, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0931098696461825, |
|
"eval_loss": 12.700243949890137, |
|
"eval_runtime": 12.0083, |
|
"eval_samples_per_second": 75.365, |
|
"eval_steps_per_second": 9.493, |
|
"step": 50 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 100, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 2, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 0 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3095861416427520.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|