|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 12.991322457480042, |
|
"eval_steps": 500, |
|
"global_step": 18720, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.3471017007983339, |
|
"grad_norm": 9.175567626953125, |
|
"learning_rate": 2e-05, |
|
"loss": 10.9573, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.6942034015966678, |
|
"grad_norm": 6.246857166290283, |
|
"learning_rate": 1.9451152579582877e-05, |
|
"loss": 2.2463, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.4781490564346313, |
|
"eval_runtime": 78.7792, |
|
"eval_samples_per_second": 146.282, |
|
"eval_steps_per_second": 4.582, |
|
"step": 1441 |
|
}, |
|
{ |
|
"epoch": 1.0409580006942034, |
|
"grad_norm": 0.7370556592941284, |
|
"learning_rate": 1.890230515916575e-05, |
|
"loss": 1.8168, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.3880597014925373, |
|
"grad_norm": 600.9544677734375, |
|
"learning_rate": 1.835345773874863e-05, |
|
"loss": 2.8979, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.7351614022908712, |
|
"grad_norm": 0.528658390045166, |
|
"learning_rate": 1.7804610318331505e-05, |
|
"loss": 1.701, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 1.3760780096054077, |
|
"eval_runtime": 78.8326, |
|
"eval_samples_per_second": 146.183, |
|
"eval_steps_per_second": 4.579, |
|
"step": 2882 |
|
}, |
|
{ |
|
"epoch": 2.081916001388407, |
|
"grad_norm": 0.45837515592575073, |
|
"learning_rate": 1.725576289791438e-05, |
|
"loss": 1.5956, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.4290177021867407, |
|
"grad_norm": 0.41563984751701355, |
|
"learning_rate": 1.6706915477497255e-05, |
|
"loss": 1.5615, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.7761194029850746, |
|
"grad_norm": 0.4553076922893524, |
|
"learning_rate": 1.6158068057080134e-05, |
|
"loss": 1.5261, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 1.3131805658340454, |
|
"eval_runtime": 78.7183, |
|
"eval_samples_per_second": 146.396, |
|
"eval_steps_per_second": 4.586, |
|
"step": 4323 |
|
}, |
|
{ |
|
"epoch": 3.12287400208261, |
|
"grad_norm": 0.4429566562175751, |
|
"learning_rate": 1.560922063666301e-05, |
|
"loss": 1.4933, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.469975702880944, |
|
"grad_norm": 0.4545910656452179, |
|
"learning_rate": 1.5060373216245885e-05, |
|
"loss": 1.4766, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 3.817077403679278, |
|
"grad_norm": 0.4629240930080414, |
|
"learning_rate": 1.451152579582876e-05, |
|
"loss": 1.4626, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 1.2775077819824219, |
|
"eval_runtime": 78.6454, |
|
"eval_samples_per_second": 146.531, |
|
"eval_steps_per_second": 4.59, |
|
"step": 5764 |
|
}, |
|
{ |
|
"epoch": 4.163832002776814, |
|
"grad_norm": 0.3974866271018982, |
|
"learning_rate": 1.3962678375411635e-05, |
|
"loss": 1.448, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 4.510933703575148, |
|
"grad_norm": 0.4241460859775543, |
|
"learning_rate": 1.3413830954994514e-05, |
|
"loss": 1.4241, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 4.8580354043734815, |
|
"grad_norm": 0.4666702151298523, |
|
"learning_rate": 1.2864983534577389e-05, |
|
"loss": 1.4085, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 1.254936695098877, |
|
"eval_runtime": 78.5765, |
|
"eval_samples_per_second": 146.66, |
|
"eval_steps_per_second": 4.594, |
|
"step": 7205 |
|
}, |
|
{ |
|
"epoch": 5.204790003471017, |
|
"grad_norm": 0.4225795567035675, |
|
"learning_rate": 1.2316136114160264e-05, |
|
"loss": 1.395, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 5.551891704269351, |
|
"grad_norm": 0.42448151111602783, |
|
"learning_rate": 1.176728869374314e-05, |
|
"loss": 1.3938, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 5.898993405067685, |
|
"grad_norm": 0.47224199771881104, |
|
"learning_rate": 1.1218441273326017e-05, |
|
"loss": 1.3844, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 1.2364274263381958, |
|
"eval_runtime": 78.7665, |
|
"eval_samples_per_second": 146.306, |
|
"eval_steps_per_second": 4.583, |
|
"step": 8646 |
|
}, |
|
{ |
|
"epoch": 6.24574800416522, |
|
"grad_norm": 0.3676798939704895, |
|
"learning_rate": 1.0669593852908892e-05, |
|
"loss": 1.3668, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 6.592849704963554, |
|
"grad_norm": 0.37000757455825806, |
|
"learning_rate": 1.0120746432491769e-05, |
|
"loss": 1.3629, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 6.939951405761888, |
|
"grad_norm": 0.45095133781433105, |
|
"learning_rate": 9.571899012074644e-06, |
|
"loss": 1.3647, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 1.222563624382019, |
|
"eval_runtime": 78.5655, |
|
"eval_samples_per_second": 146.68, |
|
"eval_steps_per_second": 4.595, |
|
"step": 10087 |
|
}, |
|
{ |
|
"epoch": 7.286706004859424, |
|
"grad_norm": 0.4932011365890503, |
|
"learning_rate": 9.02305159165752e-06, |
|
"loss": 1.358, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 7.633807705657758, |
|
"grad_norm": 0.36681583523750305, |
|
"learning_rate": 8.474204171240396e-06, |
|
"loss": 1.3422, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 7.9809094064560915, |
|
"grad_norm": 0.39713332056999207, |
|
"learning_rate": 7.925356750823272e-06, |
|
"loss": 1.3442, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 1.214061975479126, |
|
"eval_runtime": 78.596, |
|
"eval_samples_per_second": 146.623, |
|
"eval_steps_per_second": 4.593, |
|
"step": 11528 |
|
}, |
|
{ |
|
"epoch": 8.327664005553627, |
|
"grad_norm": 0.4920157194137573, |
|
"learning_rate": 7.376509330406147e-06, |
|
"loss": 1.3358, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 8.67476570635196, |
|
"grad_norm": 0.4292038083076477, |
|
"learning_rate": 6.827661909989024e-06, |
|
"loss": 1.3373, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 1.2043074369430542, |
|
"eval_runtime": 78.6255, |
|
"eval_samples_per_second": 146.568, |
|
"eval_steps_per_second": 4.591, |
|
"step": 12969 |
|
}, |
|
{ |
|
"epoch": 9.021520305449497, |
|
"grad_norm": 0.390511691570282, |
|
"learning_rate": 6.278814489571899e-06, |
|
"loss": 1.3305, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 9.368622006247831, |
|
"grad_norm": 0.42470699548721313, |
|
"learning_rate": 5.729967069154776e-06, |
|
"loss": 1.3294, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 9.715723707046164, |
|
"grad_norm": 0.4358290135860443, |
|
"learning_rate": 5.181119648737652e-06, |
|
"loss": 1.3182, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 1.1995124816894531, |
|
"eval_runtime": 78.7492, |
|
"eval_samples_per_second": 146.338, |
|
"eval_steps_per_second": 4.584, |
|
"step": 14410 |
|
}, |
|
{ |
|
"epoch": 10.0624783061437, |
|
"grad_norm": 0.37242305278778076, |
|
"learning_rate": 4.632272228320528e-06, |
|
"loss": 1.3192, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 10.409580006942035, |
|
"grad_norm": 0.4692925810813904, |
|
"learning_rate": 4.0834248079034035e-06, |
|
"loss": 1.3177, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 10.756681707740368, |
|
"grad_norm": 0.42067646980285645, |
|
"learning_rate": 3.5345773874862793e-06, |
|
"loss": 1.3169, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_loss": 1.1953924894332886, |
|
"eval_runtime": 78.6673, |
|
"eval_samples_per_second": 146.49, |
|
"eval_steps_per_second": 4.589, |
|
"step": 15851 |
|
}, |
|
{ |
|
"epoch": 11.103436306837903, |
|
"grad_norm": 0.4010460674762726, |
|
"learning_rate": 2.9857299670691548e-06, |
|
"loss": 1.3145, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 11.450538007636238, |
|
"grad_norm": 0.4176557660102844, |
|
"learning_rate": 2.436882546652031e-06, |
|
"loss": 1.3063, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 11.797639708434572, |
|
"grad_norm": 0.3621806800365448, |
|
"learning_rate": 1.888035126234907e-06, |
|
"loss": 1.3195, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_loss": 1.1941872835159302, |
|
"eval_runtime": 78.6747, |
|
"eval_samples_per_second": 146.477, |
|
"eval_steps_per_second": 4.589, |
|
"step": 17292 |
|
}, |
|
{ |
|
"epoch": 12.144394307532107, |
|
"grad_norm": 0.41595444083213806, |
|
"learning_rate": 1.3391877058177828e-06, |
|
"loss": 1.3222, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 12.49149600833044, |
|
"grad_norm": 0.40309804677963257, |
|
"learning_rate": 7.903402854006587e-07, |
|
"loss": 1.3122, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 12.838597709128775, |
|
"grad_norm": 0.4489055871963501, |
|
"learning_rate": 2.414928649835346e-07, |
|
"loss": 1.3016, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 12.991322457480042, |
|
"eval_loss": 1.19288170337677, |
|
"eval_runtime": 78.6526, |
|
"eval_samples_per_second": 146.518, |
|
"eval_steps_per_second": 4.59, |
|
"step": 18720 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 18720, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 13, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.070590055763149e+17, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|