|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.905320108205591, |
|
"eval_steps": 9, |
|
"global_step": 100, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.028854824165915238, |
|
"eval_loss": 2.2234888076782227, |
|
"eval_runtime": 63.3246, |
|
"eval_samples_per_second": 1.848, |
|
"eval_steps_per_second": 0.932, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.08656447249774572, |
|
"grad_norm": 11.650296211242676, |
|
"learning_rate": 3e-05, |
|
"loss": 71.8161, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.17312894499549145, |
|
"grad_norm": 12.471053123474121, |
|
"learning_rate": 6e-05, |
|
"loss": 72.0251, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.25969341749323716, |
|
"grad_norm": 11.917525291442871, |
|
"learning_rate": 9e-05, |
|
"loss": 66.3847, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.25969341749323716, |
|
"eval_loss": 1.9188385009765625, |
|
"eval_runtime": 63.2986, |
|
"eval_samples_per_second": 1.848, |
|
"eval_steps_per_second": 0.932, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.3462578899909829, |
|
"grad_norm": 11.10000991821289, |
|
"learning_rate": 0.00012, |
|
"loss": 59.632, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.4328223624887286, |
|
"grad_norm": 12.129708290100098, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 55.7109, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.5193868349864743, |
|
"grad_norm": 7.577877044677734, |
|
"learning_rate": 0.00018, |
|
"loss": 51.9351, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.5193868349864743, |
|
"eval_loss": 1.5988303422927856, |
|
"eval_runtime": 63.413, |
|
"eval_samples_per_second": 1.845, |
|
"eval_steps_per_second": 0.93, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.60595130748422, |
|
"grad_norm": 6.87202262878418, |
|
"learning_rate": 0.0001999229036240723, |
|
"loss": 51.0237, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.6925157799819658, |
|
"grad_norm": 6.496087551116943, |
|
"learning_rate": 0.00019876883405951377, |
|
"loss": 48.9338, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.7790802524797115, |
|
"grad_norm": 6.911287784576416, |
|
"learning_rate": 0.00019624552364536473, |
|
"loss": 47.7303, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.7790802524797115, |
|
"eval_loss": 1.4810194969177246, |
|
"eval_runtime": 63.4002, |
|
"eval_samples_per_second": 1.845, |
|
"eval_steps_per_second": 0.931, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.8656447249774571, |
|
"grad_norm": 6.416966915130615, |
|
"learning_rate": 0.0001923879532511287, |
|
"loss": 47.3634, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.9522091974752029, |
|
"grad_norm": 6.382626056671143, |
|
"learning_rate": 0.00018724960070727972, |
|
"loss": 46.6147, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 1.048692515779982, |
|
"grad_norm": 7.635010719299316, |
|
"learning_rate": 0.00018090169943749476, |
|
"loss": 46.168, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 1.048692515779982, |
|
"eval_loss": 1.4284162521362305, |
|
"eval_runtime": 63.4232, |
|
"eval_samples_per_second": 1.845, |
|
"eval_steps_per_second": 0.93, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 1.1352569882777277, |
|
"grad_norm": 7.26214075088501, |
|
"learning_rate": 0.00017343225094356855, |
|
"loss": 42.9364, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 1.2218214607754734, |
|
"grad_norm": 8.307271003723145, |
|
"learning_rate": 0.00016494480483301836, |
|
"loss": 43.8462, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 1.308385933273219, |
|
"grad_norm": 8.332019805908203, |
|
"learning_rate": 0.00015555702330196023, |
|
"loss": 43.2469, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 1.308385933273219, |
|
"eval_loss": 1.39437735080719, |
|
"eval_runtime": 63.2944, |
|
"eval_samples_per_second": 1.849, |
|
"eval_steps_per_second": 0.932, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 1.394950405770965, |
|
"grad_norm": 8.096892356872559, |
|
"learning_rate": 0.00014539904997395468, |
|
"loss": 41.2373, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 1.4815148782687104, |
|
"grad_norm": 8.60307502746582, |
|
"learning_rate": 0.0001346117057077493, |
|
"loss": 42.4838, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 1.5680793507664563, |
|
"grad_norm": 8.435927391052246, |
|
"learning_rate": 0.00012334453638559057, |
|
"loss": 43.469, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 1.5680793507664563, |
|
"eval_loss": 1.376704454421997, |
|
"eval_runtime": 63.2953, |
|
"eval_samples_per_second": 1.848, |
|
"eval_steps_per_second": 0.932, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 1.654643823264202, |
|
"grad_norm": 9.683271408081055, |
|
"learning_rate": 0.00011175373974578378, |
|
"loss": 41.1976, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 1.7412082957619477, |
|
"grad_norm": 8.922101974487305, |
|
"learning_rate": 0.0001, |
|
"loss": 41.2615, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.8277727682596934, |
|
"grad_norm": 8.805673599243164, |
|
"learning_rate": 8.824626025421626e-05, |
|
"loss": 40.953, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 1.8277727682596934, |
|
"eval_loss": 1.3633700609207153, |
|
"eval_runtime": 63.3102, |
|
"eval_samples_per_second": 1.848, |
|
"eval_steps_per_second": 0.932, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 1.914337240757439, |
|
"grad_norm": 9.028271675109863, |
|
"learning_rate": 7.66554636144095e-05, |
|
"loss": 40.7417, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 2.010820559062218, |
|
"grad_norm": 9.204501152038574, |
|
"learning_rate": 6.538829429225069e-05, |
|
"loss": 41.2274, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 2.097385031559964, |
|
"grad_norm": 9.566389083862305, |
|
"learning_rate": 5.4600950026045326e-05, |
|
"loss": 37.3547, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 2.097385031559964, |
|
"eval_loss": 1.3545383214950562, |
|
"eval_runtime": 63.2718, |
|
"eval_samples_per_second": 1.849, |
|
"eval_steps_per_second": 0.932, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 2.1839495040577095, |
|
"grad_norm": 10.605365753173828, |
|
"learning_rate": 4.444297669803981e-05, |
|
"loss": 36.6109, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 2.2705139765554554, |
|
"grad_norm": 11.574414253234863, |
|
"learning_rate": 3.5055195166981645e-05, |
|
"loss": 37.0909, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 2.3570784490532013, |
|
"grad_norm": 11.285858154296875, |
|
"learning_rate": 2.6567749056431467e-05, |
|
"loss": 38.6229, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 2.3570784490532013, |
|
"eval_loss": 1.3696061372756958, |
|
"eval_runtime": 63.2953, |
|
"eval_samples_per_second": 1.848, |
|
"eval_steps_per_second": 0.932, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 2.4436429215509468, |
|
"grad_norm": 11.047082901000977, |
|
"learning_rate": 1.9098300562505266e-05, |
|
"loss": 37.311, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 2.5302073940486927, |
|
"grad_norm": 11.326862335205078, |
|
"learning_rate": 1.2750399292720283e-05, |
|
"loss": 35.5301, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 2.616771866546438, |
|
"grad_norm": 10.88636302947998, |
|
"learning_rate": 7.612046748871327e-06, |
|
"loss": 36.4226, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 2.616771866546438, |
|
"eval_loss": 1.3591108322143555, |
|
"eval_runtime": 63.3023, |
|
"eval_samples_per_second": 1.848, |
|
"eval_steps_per_second": 0.932, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 2.703336339044184, |
|
"grad_norm": 10.809871673583984, |
|
"learning_rate": 3.7544763546352834e-06, |
|
"loss": 37.0722, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 2.78990081154193, |
|
"grad_norm": 11.037607192993164, |
|
"learning_rate": 1.231165940486234e-06, |
|
"loss": 35.8129, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 2.8764652840396754, |
|
"grad_norm": 10.600300788879395, |
|
"learning_rate": 7.709637592770991e-08, |
|
"loss": 35.964, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 2.8764652840396754, |
|
"eval_loss": 1.3585889339447021, |
|
"eval_runtime": 63.2827, |
|
"eval_samples_per_second": 1.849, |
|
"eval_steps_per_second": 0.932, |
|
"step": 99 |
|
} |
|
], |
|
"logging_steps": 3, |
|
"max_steps": 100, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 9, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.020574826496e+18, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|