bbytxt's picture
Training in progress, step 50, checkpoint
69287c6 verified
raw
history blame
9.61 kB
{
"best_metric": 12.700243949890137,
"best_model_checkpoint": "miner_id_24/checkpoint-50",
"epoch": 0.0931098696461825,
"eval_steps": 50,
"global_step": 50,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00186219739292365,
"grad_norm": 7.960118770599365,
"learning_rate": 1.5000000000000002e-07,
"loss": 14.0369,
"step": 1
},
{
"epoch": 0.00186219739292365,
"eval_loss": 13.924460411071777,
"eval_runtime": 12.7592,
"eval_samples_per_second": 70.929,
"eval_steps_per_second": 8.935,
"step": 1
},
{
"epoch": 0.0037243947858473,
"grad_norm": 8.014508247375488,
"learning_rate": 3.0000000000000004e-07,
"loss": 13.7099,
"step": 2
},
{
"epoch": 0.00558659217877095,
"grad_norm": 8.808537483215332,
"learning_rate": 4.5e-07,
"loss": 13.8903,
"step": 3
},
{
"epoch": 0.0074487895716946,
"grad_norm": 8.004558563232422,
"learning_rate": 6.000000000000001e-07,
"loss": 14.0816,
"step": 4
},
{
"epoch": 0.00931098696461825,
"grad_norm": 7.907912254333496,
"learning_rate": 7.5e-07,
"loss": 14.4216,
"step": 5
},
{
"epoch": 0.0111731843575419,
"grad_norm": 7.714488506317139,
"learning_rate": 9e-07,
"loss": 13.8048,
"step": 6
},
{
"epoch": 0.01303538175046555,
"grad_norm": 8.773198127746582,
"learning_rate": 1.0500000000000001e-06,
"loss": 13.7771,
"step": 7
},
{
"epoch": 0.0148975791433892,
"grad_norm": 8.656310081481934,
"learning_rate": 1.2000000000000002e-06,
"loss": 13.9204,
"step": 8
},
{
"epoch": 0.01675977653631285,
"grad_norm": 7.340457439422607,
"learning_rate": 1.35e-06,
"loss": 14.1696,
"step": 9
},
{
"epoch": 0.0186219739292365,
"grad_norm": 7.3591461181640625,
"learning_rate": 1.5e-06,
"loss": 13.6311,
"step": 10
},
{
"epoch": 0.020484171322160148,
"grad_norm": 7.934558868408203,
"learning_rate": 1.65e-06,
"loss": 14.0909,
"step": 11
},
{
"epoch": 0.0223463687150838,
"grad_norm": 8.731240272521973,
"learning_rate": 1.8e-06,
"loss": 13.8947,
"step": 12
},
{
"epoch": 0.024208566108007448,
"grad_norm": 8.52802562713623,
"learning_rate": 1.95e-06,
"loss": 13.941,
"step": 13
},
{
"epoch": 0.0260707635009311,
"grad_norm": 8.355961799621582,
"learning_rate": 2.1000000000000002e-06,
"loss": 13.9408,
"step": 14
},
{
"epoch": 0.027932960893854747,
"grad_norm": 8.493719100952148,
"learning_rate": 2.25e-06,
"loss": 13.8167,
"step": 15
},
{
"epoch": 0.0297951582867784,
"grad_norm": 7.872385025024414,
"learning_rate": 2.4000000000000003e-06,
"loss": 13.8842,
"step": 16
},
{
"epoch": 0.03165735567970205,
"grad_norm": 7.678893566131592,
"learning_rate": 2.55e-06,
"loss": 13.8789,
"step": 17
},
{
"epoch": 0.0335195530726257,
"grad_norm": 8.681936264038086,
"learning_rate": 2.7e-06,
"loss": 13.9825,
"step": 18
},
{
"epoch": 0.035381750465549346,
"grad_norm": 8.17630672454834,
"learning_rate": 2.8500000000000002e-06,
"loss": 13.7963,
"step": 19
},
{
"epoch": 0.037243947858473,
"grad_norm": 8.064035415649414,
"learning_rate": 3e-06,
"loss": 13.941,
"step": 20
},
{
"epoch": 0.03910614525139665,
"grad_norm": 8.132279396057129,
"learning_rate": 3.15e-06,
"loss": 13.8393,
"step": 21
},
{
"epoch": 0.040968342644320296,
"grad_norm": 8.09195613861084,
"learning_rate": 3.3e-06,
"loss": 13.9476,
"step": 22
},
{
"epoch": 0.04283054003724395,
"grad_norm": 9.464410781860352,
"learning_rate": 3.4500000000000004e-06,
"loss": 13.9283,
"step": 23
},
{
"epoch": 0.0446927374301676,
"grad_norm": 8.223386764526367,
"learning_rate": 3.6e-06,
"loss": 13.8445,
"step": 24
},
{
"epoch": 0.04655493482309125,
"grad_norm": 8.496528625488281,
"learning_rate": 3.75e-06,
"loss": 13.9996,
"step": 25
},
{
"epoch": 0.048417132216014895,
"grad_norm": 7.405032157897949,
"learning_rate": 3.9e-06,
"loss": 13.4262,
"step": 26
},
{
"epoch": 0.05027932960893855,
"grad_norm": 7.917985916137695,
"learning_rate": 4.05e-06,
"loss": 13.7422,
"step": 27
},
{
"epoch": 0.0521415270018622,
"grad_norm": 7.779364585876465,
"learning_rate": 4.2000000000000004e-06,
"loss": 13.8355,
"step": 28
},
{
"epoch": 0.054003724394785846,
"grad_norm": 7.747602462768555,
"learning_rate": 4.35e-06,
"loss": 13.6872,
"step": 29
},
{
"epoch": 0.055865921787709494,
"grad_norm": 8.12415599822998,
"learning_rate": 4.5e-06,
"loss": 13.8575,
"step": 30
},
{
"epoch": 0.05772811918063315,
"grad_norm": 7.393126487731934,
"learning_rate": 4.65e-06,
"loss": 13.6684,
"step": 31
},
{
"epoch": 0.0595903165735568,
"grad_norm": 7.366344451904297,
"learning_rate": 4.800000000000001e-06,
"loss": 13.2535,
"step": 32
},
{
"epoch": 0.061452513966480445,
"grad_norm": 7.462878227233887,
"learning_rate": 4.95e-06,
"loss": 13.6117,
"step": 33
},
{
"epoch": 0.0633147113594041,
"grad_norm": 8.132235527038574,
"learning_rate": 5.1e-06,
"loss": 14.0737,
"step": 34
},
{
"epoch": 0.06517690875232775,
"grad_norm": 8.085715293884277,
"learning_rate": 5.25e-06,
"loss": 13.7491,
"step": 35
},
{
"epoch": 0.0670391061452514,
"grad_norm": 7.527085304260254,
"learning_rate": 5.4e-06,
"loss": 13.4391,
"step": 36
},
{
"epoch": 0.06890130353817504,
"grad_norm": 7.162932872772217,
"learning_rate": 5.55e-06,
"loss": 13.2994,
"step": 37
},
{
"epoch": 0.07076350093109869,
"grad_norm": 6.486966133117676,
"learning_rate": 5.7000000000000005e-06,
"loss": 13.5081,
"step": 38
},
{
"epoch": 0.07262569832402235,
"grad_norm": 7.425045967102051,
"learning_rate": 5.850000000000001e-06,
"loss": 13.1145,
"step": 39
},
{
"epoch": 0.074487895716946,
"grad_norm": 7.909907817840576,
"learning_rate": 6e-06,
"loss": 13.3482,
"step": 40
},
{
"epoch": 0.07635009310986965,
"grad_norm": 7.439389228820801,
"learning_rate": 6.1499999999999996e-06,
"loss": 13.4463,
"step": 41
},
{
"epoch": 0.0782122905027933,
"grad_norm": 7.182775974273682,
"learning_rate": 6.3e-06,
"loss": 13.2005,
"step": 42
},
{
"epoch": 0.08007448789571694,
"grad_norm": 7.072144031524658,
"learning_rate": 6.45e-06,
"loss": 12.8393,
"step": 43
},
{
"epoch": 0.08193668528864059,
"grad_norm": 7.034002780914307,
"learning_rate": 6.6e-06,
"loss": 12.8302,
"step": 44
},
{
"epoch": 0.08379888268156424,
"grad_norm": 6.952661037445068,
"learning_rate": 6.750000000000001e-06,
"loss": 13.1515,
"step": 45
},
{
"epoch": 0.0856610800744879,
"grad_norm": 7.125630855560303,
"learning_rate": 6.900000000000001e-06,
"loss": 13.0473,
"step": 46
},
{
"epoch": 0.08752327746741155,
"grad_norm": 7.452700138092041,
"learning_rate": 7.049999999999999e-06,
"loss": 13.0773,
"step": 47
},
{
"epoch": 0.0893854748603352,
"grad_norm": 7.248137474060059,
"learning_rate": 7.2e-06,
"loss": 13.0051,
"step": 48
},
{
"epoch": 0.09124767225325885,
"grad_norm": 6.84269905090332,
"learning_rate": 7.35e-06,
"loss": 12.6788,
"step": 49
},
{
"epoch": 0.0931098696461825,
"grad_norm": 6.934720993041992,
"learning_rate": 7.5e-06,
"loss": 12.7921,
"step": 50
},
{
"epoch": 0.0931098696461825,
"eval_loss": 12.700243949890137,
"eval_runtime": 12.0083,
"eval_samples_per_second": 75.365,
"eval_steps_per_second": 9.493,
"step": 50
}
],
"logging_steps": 1,
"max_steps": 100,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 50,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 2,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3095861416427520.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}