|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 6.0, |
|
"eval_steps": 500, |
|
"global_step": 7662, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.07830853563038372, |
|
"grad_norm": 9.886259078979492, |
|
"learning_rate": 9.874706342991386e-05, |
|
"loss": 2.2961, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.15661707126076743, |
|
"grad_norm": 9.754347801208496, |
|
"learning_rate": 9.744192116940747e-05, |
|
"loss": 2.1652, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.23492560689115113, |
|
"grad_norm": 10.011685371398926, |
|
"learning_rate": 9.613677890890107e-05, |
|
"loss": 2.2701, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.31323414252153486, |
|
"grad_norm": 10.389019966125488, |
|
"learning_rate": 9.483163664839469e-05, |
|
"loss": 2.1357, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.39154267815191857, |
|
"grad_norm": 8.717257499694824, |
|
"learning_rate": 9.352649438788829e-05, |
|
"loss": 2.0979, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.46985121378230227, |
|
"grad_norm": 10.2243070602417, |
|
"learning_rate": 9.222135212738188e-05, |
|
"loss": 2.0926, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.548159749412686, |
|
"grad_norm": 9.24715518951416, |
|
"learning_rate": 9.091620986687549e-05, |
|
"loss": 2.0199, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6264682850430697, |
|
"grad_norm": 8.67659854888916, |
|
"learning_rate": 8.96110676063691e-05, |
|
"loss": 2.071, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.7047768206734534, |
|
"grad_norm": 6.2473015785217285, |
|
"learning_rate": 8.830592534586271e-05, |
|
"loss": 2.0342, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7830853563038371, |
|
"grad_norm": 9.428214073181152, |
|
"learning_rate": 8.700078308535631e-05, |
|
"loss": 1.9497, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.8613938919342208, |
|
"grad_norm": 7.909013271331787, |
|
"learning_rate": 8.569564082484992e-05, |
|
"loss": 2.006, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.9397024275646045, |
|
"grad_norm": 7.305349826812744, |
|
"learning_rate": 8.43904985643435e-05, |
|
"loss": 1.9751, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.0180109631949883, |
|
"grad_norm": 6.404293060302734, |
|
"learning_rate": 8.308535630383712e-05, |
|
"loss": 1.7841, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.096319498825372, |
|
"grad_norm": 6.121842384338379, |
|
"learning_rate": 8.178021404333073e-05, |
|
"loss": 1.2646, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.1746280344557556, |
|
"grad_norm": 5.871735572814941, |
|
"learning_rate": 8.047507178282433e-05, |
|
"loss": 1.2666, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.2529365700861395, |
|
"grad_norm": 6.930913925170898, |
|
"learning_rate": 7.916992952231794e-05, |
|
"loss": 1.295, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.331245105716523, |
|
"grad_norm": 5.334323406219482, |
|
"learning_rate": 7.786478726181154e-05, |
|
"loss": 1.3333, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.4095536413469067, |
|
"grad_norm": 7.161041259765625, |
|
"learning_rate": 7.655964500130514e-05, |
|
"loss": 1.3355, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.4878621769772904, |
|
"grad_norm": 7.43231725692749, |
|
"learning_rate": 7.525450274079875e-05, |
|
"loss": 1.284, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.5661707126076743, |
|
"grad_norm": 6.2656354904174805, |
|
"learning_rate": 7.394936048029235e-05, |
|
"loss": 1.287, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.644479248238058, |
|
"grad_norm": 5.97381067276001, |
|
"learning_rate": 7.264421821978597e-05, |
|
"loss": 1.3196, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.7227877838684416, |
|
"grad_norm": 7.828345775604248, |
|
"learning_rate": 7.133907595927957e-05, |
|
"loss": 1.2884, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.8010963194988254, |
|
"grad_norm": 7.866464614868164, |
|
"learning_rate": 7.003393369877316e-05, |
|
"loss": 1.3136, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.879404855129209, |
|
"grad_norm": 5.8095598220825195, |
|
"learning_rate": 6.872879143826677e-05, |
|
"loss": 1.2181, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.9577133907595927, |
|
"grad_norm": 6.666238784790039, |
|
"learning_rate": 6.742364917776038e-05, |
|
"loss": 1.2461, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.0360219263899766, |
|
"grad_norm": 7.129757881164551, |
|
"learning_rate": 6.611850691725399e-05, |
|
"loss": 0.9798, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.11433046202036, |
|
"grad_norm": 6.111073017120361, |
|
"learning_rate": 6.481336465674759e-05, |
|
"loss": 0.6804, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.192638997650744, |
|
"grad_norm": 5.344665050506592, |
|
"learning_rate": 6.35082223962412e-05, |
|
"loss": 0.6946, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.2709475332811278, |
|
"grad_norm": 5.611790657043457, |
|
"learning_rate": 6.220308013573479e-05, |
|
"loss": 0.6516, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.349256068911511, |
|
"grad_norm": 7.892216682434082, |
|
"learning_rate": 6.0897937875228404e-05, |
|
"loss": 0.6718, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.427564604541895, |
|
"grad_norm": 8.479430198669434, |
|
"learning_rate": 5.959279561472201e-05, |
|
"loss": 0.6601, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.505873140172279, |
|
"grad_norm": 6.140246391296387, |
|
"learning_rate": 5.828765335421561e-05, |
|
"loss": 0.6619, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.5841816758026623, |
|
"grad_norm": 6.076893329620361, |
|
"learning_rate": 5.6982511093709215e-05, |
|
"loss": 0.69, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.662490211433046, |
|
"grad_norm": 6.6155900955200195, |
|
"learning_rate": 5.5677368833202826e-05, |
|
"loss": 0.6866, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.74079874706343, |
|
"grad_norm": 6.685502052307129, |
|
"learning_rate": 5.437222657269643e-05, |
|
"loss": 0.6615, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.8191072826938135, |
|
"grad_norm": 8.586551666259766, |
|
"learning_rate": 5.306708431219003e-05, |
|
"loss": 0.7118, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.8974158183241974, |
|
"grad_norm": 5.001748085021973, |
|
"learning_rate": 5.176194205168363e-05, |
|
"loss": 0.7117, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 2.975724353954581, |
|
"grad_norm": 5.666826248168945, |
|
"learning_rate": 5.045679979117724e-05, |
|
"loss": 0.6868, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 3.0540328895849647, |
|
"grad_norm": 5.463041305541992, |
|
"learning_rate": 4.9151657530670846e-05, |
|
"loss": 0.4208, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 3.1323414252153485, |
|
"grad_norm": 6.563859939575195, |
|
"learning_rate": 4.784651527016445e-05, |
|
"loss": 0.2744, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.210649960845732, |
|
"grad_norm": 5.918147087097168, |
|
"learning_rate": 4.654137300965806e-05, |
|
"loss": 0.309, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 3.288958496476116, |
|
"grad_norm": 4.2588958740234375, |
|
"learning_rate": 4.523623074915166e-05, |
|
"loss": 0.3063, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 3.3672670321064997, |
|
"grad_norm": 5.194436073303223, |
|
"learning_rate": 4.393108848864527e-05, |
|
"loss": 0.3002, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 3.445575567736883, |
|
"grad_norm": 4.340348720550537, |
|
"learning_rate": 4.262594622813887e-05, |
|
"loss": 0.2957, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 3.523884103367267, |
|
"grad_norm": 3.6878747940063477, |
|
"learning_rate": 4.1320803967632476e-05, |
|
"loss": 0.2968, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.602192638997651, |
|
"grad_norm": 4.565743923187256, |
|
"learning_rate": 4.001566170712608e-05, |
|
"loss": 0.2921, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 3.6805011746280343, |
|
"grad_norm": 4.834578514099121, |
|
"learning_rate": 3.8710519446619684e-05, |
|
"loss": 0.2937, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 3.758809710258418, |
|
"grad_norm": 4.321404933929443, |
|
"learning_rate": 3.740537718611329e-05, |
|
"loss": 0.2927, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 3.837118245888802, |
|
"grad_norm": 5.987460136413574, |
|
"learning_rate": 3.610023492560689e-05, |
|
"loss": 0.2813, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 3.9154267815191854, |
|
"grad_norm": 4.690079689025879, |
|
"learning_rate": 3.4795092665100495e-05, |
|
"loss": 0.2848, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 3.9937353171495693, |
|
"grad_norm": 4.9851298332214355, |
|
"learning_rate": 3.34899504045941e-05, |
|
"loss": 0.2776, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 4.072043852779953, |
|
"grad_norm": 4.890774726867676, |
|
"learning_rate": 3.218480814408771e-05, |
|
"loss": 0.1309, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 4.150352388410337, |
|
"grad_norm": 3.3844189643859863, |
|
"learning_rate": 3.087966588358131e-05, |
|
"loss": 0.1197, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 4.22866092404072, |
|
"grad_norm": 3.3200528621673584, |
|
"learning_rate": 2.9574523623074918e-05, |
|
"loss": 0.1252, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 4.306969459671104, |
|
"grad_norm": 2.3074514865875244, |
|
"learning_rate": 2.826938136256852e-05, |
|
"loss": 0.1181, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 4.385277995301488, |
|
"grad_norm": 4.1699724197387695, |
|
"learning_rate": 2.696423910206213e-05, |
|
"loss": 0.1234, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 4.463586530931871, |
|
"grad_norm": 3.833683729171753, |
|
"learning_rate": 2.565909684155573e-05, |
|
"loss": 0.1161, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 4.5418950665622555, |
|
"grad_norm": 2.7541275024414062, |
|
"learning_rate": 2.4353954581049333e-05, |
|
"loss": 0.1101, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 4.620203602192639, |
|
"grad_norm": 4.193445205688477, |
|
"learning_rate": 2.304881232054294e-05, |
|
"loss": 0.1118, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 4.698512137823022, |
|
"grad_norm": 3.851712465286255, |
|
"learning_rate": 2.1743670060036544e-05, |
|
"loss": 0.1144, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 4.776820673453407, |
|
"grad_norm": 3.333712100982666, |
|
"learning_rate": 2.043852779953015e-05, |
|
"loss": 0.1117, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 4.85512920908379, |
|
"grad_norm": 3.7569446563720703, |
|
"learning_rate": 1.9133385539023756e-05, |
|
"loss": 0.1049, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 4.9334377447141735, |
|
"grad_norm": 3.2152910232543945, |
|
"learning_rate": 1.782824327851736e-05, |
|
"loss": 0.1064, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 5.011746280344558, |
|
"grad_norm": 1.593434453010559, |
|
"learning_rate": 1.6523101018010963e-05, |
|
"loss": 0.0936, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 5.090054815974941, |
|
"grad_norm": 1.8612021207809448, |
|
"learning_rate": 1.5217958757504569e-05, |
|
"loss": 0.047, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 5.168363351605325, |
|
"grad_norm": 2.482856035232544, |
|
"learning_rate": 1.3912816496998175e-05, |
|
"loss": 0.0448, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 5.246671887235709, |
|
"grad_norm": 1.4229090213775635, |
|
"learning_rate": 1.2607674236491779e-05, |
|
"loss": 0.044, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 5.324980422866092, |
|
"grad_norm": 1.7034039497375488, |
|
"learning_rate": 1.1302531975985382e-05, |
|
"loss": 0.0459, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 5.403288958496476, |
|
"grad_norm": 2.2560577392578125, |
|
"learning_rate": 9.997389715478986e-06, |
|
"loss": 0.0453, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 5.48159749412686, |
|
"grad_norm": 2.826206922531128, |
|
"learning_rate": 8.692247454972592e-06, |
|
"loss": 0.0439, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 5.559906029757244, |
|
"grad_norm": 1.7038235664367676, |
|
"learning_rate": 7.387105194466197e-06, |
|
"loss": 0.0421, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 5.638214565387627, |
|
"grad_norm": 2.713496685028076, |
|
"learning_rate": 6.081962933959802e-06, |
|
"loss": 0.0417, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 5.716523101018011, |
|
"grad_norm": 0.9614657163619995, |
|
"learning_rate": 4.776820673453407e-06, |
|
"loss": 0.0403, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 5.794831636648395, |
|
"grad_norm": 1.6873942613601685, |
|
"learning_rate": 3.471678412947011e-06, |
|
"loss": 0.0407, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 5.873140172278778, |
|
"grad_norm": 2.548095464706421, |
|
"learning_rate": 2.166536152440616e-06, |
|
"loss": 0.0397, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 5.9514487079091625, |
|
"grad_norm": 1.1459206342697144, |
|
"learning_rate": 8.613938919342209e-07, |
|
"loss": 0.0366, |
|
"step": 7600 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 7662, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 6, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 8424085506490368.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|