|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 244, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.020491803278688523, |
|
"grad_norm": 0.3614635057701486, |
|
"learning_rate": 2e-05, |
|
"loss": 1.3385, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.040983606557377046, |
|
"grad_norm": 0.2489719842782885, |
|
"learning_rate": 4e-05, |
|
"loss": 1.3036, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.06147540983606557, |
|
"grad_norm": 0.21147248994437587, |
|
"learning_rate": 6e-05, |
|
"loss": 1.2625, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.08196721311475409, |
|
"grad_norm": 0.19373427885281047, |
|
"learning_rate": 8e-05, |
|
"loss": 1.1662, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.10245901639344263, |
|
"grad_norm": 0.16443335321911356, |
|
"learning_rate": 0.0001, |
|
"loss": 1.2303, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.12295081967213115, |
|
"grad_norm": 0.14174512674576367, |
|
"learning_rate": 9.987144015627809e-05, |
|
"loss": 1.1398, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.14344262295081966, |
|
"grad_norm": 0.1381398528306756, |
|
"learning_rate": 9.948642173044905e-05, |
|
"loss": 1.3659, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.16393442622950818, |
|
"grad_norm": 19.814446058738557, |
|
"learning_rate": 9.88469246388591e-05, |
|
"loss": 1.2665, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.18442622950819673, |
|
"grad_norm": 0.15246937379745587, |
|
"learning_rate": 9.79562374273544e-05, |
|
"loss": 1.2956, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.20491803278688525, |
|
"grad_norm": 0.11304889245590244, |
|
"learning_rate": 9.681894036028365e-05, |
|
"loss": 1.3647, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.22540983606557377, |
|
"grad_norm": 0.11799961373671249, |
|
"learning_rate": 9.544088186697515e-05, |
|
"loss": 1.2491, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.2459016393442623, |
|
"grad_norm": 0.0886669617443345, |
|
"learning_rate": 9.382914846681047e-05, |
|
"loss": 1.1422, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.26639344262295084, |
|
"grad_norm": 0.13383163861854966, |
|
"learning_rate": 9.19920283275515e-05, |
|
"loss": 1.217, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.28688524590163933, |
|
"grad_norm": 0.12664816060876666, |
|
"learning_rate": 8.993896864431826e-05, |
|
"loss": 1.1728, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.3073770491803279, |
|
"grad_norm": 0.11280543596286248, |
|
"learning_rate": 8.76805270583919e-05, |
|
"loss": 1.2914, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.32786885245901637, |
|
"grad_norm": 0.11457176824246548, |
|
"learning_rate": 8.522831736566607e-05, |
|
"loss": 1.2975, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.3483606557377049, |
|
"grad_norm": 0.125952078460287, |
|
"learning_rate": 8.259494979393563e-05, |
|
"loss": 1.2325, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.36885245901639346, |
|
"grad_norm": 0.14141125617004555, |
|
"learning_rate": 7.97939661561399e-05, |
|
"loss": 1.3628, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.38934426229508196, |
|
"grad_norm": 0.12898427772056165, |
|
"learning_rate": 7.68397702130286e-05, |
|
"loss": 1.3272, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.4098360655737705, |
|
"grad_norm": 0.12462484863790337, |
|
"learning_rate": 7.374755360335253e-05, |
|
"loss": 1.2025, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.430327868852459, |
|
"grad_norm": 0.15405301651774622, |
|
"learning_rate": 7.053321772247545e-05, |
|
"loss": 1.3946, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.45081967213114754, |
|
"grad_norm": 0.11850088757256459, |
|
"learning_rate": 6.721329195113801e-05, |
|
"loss": 1.0897, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.4713114754098361, |
|
"grad_norm": 0.10161954341550462, |
|
"learning_rate": 6.380484865487347e-05, |
|
"loss": 1.3683, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.4918032786885246, |
|
"grad_norm": 0.1338711959037296, |
|
"learning_rate": 6.032541539118187e-05, |
|
"loss": 1.0908, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.5122950819672131, |
|
"grad_norm": 0.12226354104013447, |
|
"learning_rate": 5.679288477592815e-05, |
|
"loss": 1.301, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.5327868852459017, |
|
"grad_norm": 0.11741161974948867, |
|
"learning_rate": 5.3225422472465824e-05, |
|
"loss": 1.2055, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.5532786885245902, |
|
"grad_norm": 0.1220585245683397, |
|
"learning_rate": 4.9641373776643616e-05, |
|
"loss": 1.2117, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.5737704918032787, |
|
"grad_norm": 0.1283897077062444, |
|
"learning_rate": 4.605916927807057e-05, |
|
"loss": 1.1798, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.5942622950819673, |
|
"grad_norm": 0.13195983815349108, |
|
"learning_rate": 4.249723008276737e-05, |
|
"loss": 1.2929, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.6147540983606558, |
|
"grad_norm": 0.10033441692536059, |
|
"learning_rate": 3.89738730845858e-05, |
|
"loss": 1.1865, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.6352459016393442, |
|
"grad_norm": 0.10526603636165197, |
|
"learning_rate": 3.550721677252839e-05, |
|
"loss": 1.1911, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.6557377049180327, |
|
"grad_norm": 0.11769891562161476, |
|
"learning_rate": 3.2115088058343725e-05, |
|
"loss": 1.1442, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.6762295081967213, |
|
"grad_norm": 0.09808797452553064, |
|
"learning_rate": 2.8814930603527068e-05, |
|
"loss": 1.2862, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.6967213114754098, |
|
"grad_norm": 0.25170420435929075, |
|
"learning_rate": 2.5623715117144336e-05, |
|
"loss": 1.2449, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.7172131147540983, |
|
"grad_norm": 0.14162820792284236, |
|
"learning_rate": 2.2557852085764053e-05, |
|
"loss": 1.2093, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.7377049180327869, |
|
"grad_norm": 0.09769546802692396, |
|
"learning_rate": 1.963310738427367e-05, |
|
"loss": 1.1962, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.7581967213114754, |
|
"grad_norm": 0.10477577716382458, |
|
"learning_rate": 1.6864521201543197e-05, |
|
"loss": 1.1101, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.7786885245901639, |
|
"grad_norm": 0.1197892608117282, |
|
"learning_rate": 1.4266330697851954e-05, |
|
"loss": 1.3805, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.7991803278688525, |
|
"grad_norm": 0.10767908097089735, |
|
"learning_rate": 1.1851896791804507e-05, |
|
"loss": 1.3342, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.819672131147541, |
|
"grad_norm": 0.10976930143838647, |
|
"learning_rate": 9.633635453226376e-06, |
|
"loss": 1.3511, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.8401639344262295, |
|
"grad_norm": 0.10305982170288747, |
|
"learning_rate": 7.6229538553584556e-06, |
|
"loss": 1.265, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.860655737704918, |
|
"grad_norm": 0.09095323069274693, |
|
"learning_rate": 5.830191714680577e-06, |
|
"loss": 1.2826, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.8811475409836066, |
|
"grad_norm": 0.12664820610601704, |
|
"learning_rate": 4.26456812001822e-06, |
|
"loss": 1.2256, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.9016393442622951, |
|
"grad_norm": 0.11352528001951419, |
|
"learning_rate": 2.934134124357646e-06, |
|
"loss": 1.2583, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.9221311475409836, |
|
"grad_norm": 0.10147400081855255, |
|
"learning_rate": 1.8457313431614498e-06, |
|
"loss": 1.2593, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.9426229508196722, |
|
"grad_norm": 0.11813110730985737, |
|
"learning_rate": 1.004956772087895e-06, |
|
"loss": 1.2518, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.9631147540983607, |
|
"grad_norm": 0.10427451227791518, |
|
"learning_rate": 4.1613400503550114e-07, |
|
"loss": 1.3182, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.9836065573770492, |
|
"grad_norm": 0.10936990134938639, |
|
"learning_rate": 8.229100052074556e-08, |
|
"loss": 1.2333, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.1466403007507324, |
|
"eval_runtime": 27.0287, |
|
"eval_samples_per_second": 11.802, |
|
"eval_steps_per_second": 0.37, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 244, |
|
"total_flos": 84368078012416.0, |
|
"train_loss": 0.0, |
|
"train_runtime": 1.4481, |
|
"train_samples_per_second": 10777.998, |
|
"train_steps_per_second": 168.493 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 244, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 84368078012416.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|