|
{ |
|
"best_metric": 1.1385369300842285, |
|
"best_model_checkpoint": "miner_id_24/checkpoint-50", |
|
"epoch": 3.004016064257028, |
|
"eval_steps": 50, |
|
"global_step": 187, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01606425702811245, |
|
"grad_norm": 8.358396530151367, |
|
"learning_rate": 1e-05, |
|
"loss": 10.6499, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01606425702811245, |
|
"eval_loss": 2.549973487854004, |
|
"eval_runtime": 5.012, |
|
"eval_samples_per_second": 20.95, |
|
"eval_steps_per_second": 5.387, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0321285140562249, |
|
"grad_norm": 8.561280250549316, |
|
"learning_rate": 2e-05, |
|
"loss": 10.1041, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.04819277108433735, |
|
"grad_norm": 7.631792068481445, |
|
"learning_rate": 3e-05, |
|
"loss": 9.1649, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0642570281124498, |
|
"grad_norm": 8.794489860534668, |
|
"learning_rate": 4e-05, |
|
"loss": 10.3223, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.08032128514056225, |
|
"grad_norm": 8.187786102294922, |
|
"learning_rate": 5e-05, |
|
"loss": 8.9209, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0963855421686747, |
|
"grad_norm": 9.42801284790039, |
|
"learning_rate": 6e-05, |
|
"loss": 8.9984, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.11244979919678715, |
|
"grad_norm": 8.824488639831543, |
|
"learning_rate": 7e-05, |
|
"loss": 9.4921, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.1285140562248996, |
|
"grad_norm": 7.4875640869140625, |
|
"learning_rate": 8e-05, |
|
"loss": 7.2293, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.14457831325301204, |
|
"grad_norm": 5.650693416595459, |
|
"learning_rate": 9e-05, |
|
"loss": 6.9636, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.1606425702811245, |
|
"grad_norm": 4.513960838317871, |
|
"learning_rate": 0.0001, |
|
"loss": 6.1581, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.17670682730923695, |
|
"grad_norm": 4.6525983810424805, |
|
"learning_rate": 9.99921244331919e-05, |
|
"loss": 5.6187, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.1927710843373494, |
|
"grad_norm": 4.981886863708496, |
|
"learning_rate": 9.996850021374968e-05, |
|
"loss": 5.8963, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.20883534136546184, |
|
"grad_norm": 4.619255065917969, |
|
"learning_rate": 9.99291347838381e-05, |
|
"loss": 6.3916, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.2248995983935743, |
|
"grad_norm": 42.78565979003906, |
|
"learning_rate": 9.987404054446008e-05, |
|
"loss": 5.2134, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.24096385542168675, |
|
"grad_norm": 4.6155290603637695, |
|
"learning_rate": 9.980323485155013e-05, |
|
"loss": 3.8579, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.2570281124497992, |
|
"grad_norm": 4.810256004333496, |
|
"learning_rate": 9.971674001050686e-05, |
|
"loss": 6.0338, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.27309236947791166, |
|
"grad_norm": 4.097535610198975, |
|
"learning_rate": 9.961458326916624e-05, |
|
"loss": 5.6122, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.2891566265060241, |
|
"grad_norm": 4.028859615325928, |
|
"learning_rate": 9.94967968092179e-05, |
|
"loss": 5.251, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.30522088353413657, |
|
"grad_norm": 3.8736844062805176, |
|
"learning_rate": 9.936341773606723e-05, |
|
"loss": 5.9792, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.321285140562249, |
|
"grad_norm": 3.783942937850952, |
|
"learning_rate": 9.921448806714631e-05, |
|
"loss": 5.6756, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.3373493975903614, |
|
"grad_norm": 3.129595994949341, |
|
"learning_rate": 9.905005471867739e-05, |
|
"loss": 4.8383, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.3534136546184739, |
|
"grad_norm": 3.2784130573272705, |
|
"learning_rate": 9.887016949089333e-05, |
|
"loss": 4.6347, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.36947791164658633, |
|
"grad_norm": 3.4912328720092773, |
|
"learning_rate": 9.867488905171933e-05, |
|
"loss": 4.8285, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.3855421686746988, |
|
"grad_norm": 4.122783184051514, |
|
"learning_rate": 9.846427491892118e-05, |
|
"loss": 4.9993, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.40160642570281124, |
|
"grad_norm": 3.454385757446289, |
|
"learning_rate": 9.82383934407258e-05, |
|
"loss": 4.7978, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.41767068273092367, |
|
"grad_norm": 3.439769744873047, |
|
"learning_rate": 9.79973157749201e-05, |
|
"loss": 4.8947, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.43373493975903615, |
|
"grad_norm": 3.4278244972229004, |
|
"learning_rate": 9.77411178664346e-05, |
|
"loss": 4.5172, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.4497991967871486, |
|
"grad_norm": 3.5005717277526855, |
|
"learning_rate": 9.746988042341906e-05, |
|
"loss": 4.7726, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.46586345381526106, |
|
"grad_norm": 3.3018925189971924, |
|
"learning_rate": 9.718368889181764e-05, |
|
"loss": 4.3067, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.4819277108433735, |
|
"grad_norm": 3.5539093017578125, |
|
"learning_rate": 9.68826334284514e-05, |
|
"loss": 3.7747, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.4979919678714859, |
|
"grad_norm": 3.4804933071136475, |
|
"learning_rate": 9.656680887261693e-05, |
|
"loss": 5.1528, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.5140562248995983, |
|
"grad_norm": 3.2824409008026123, |
|
"learning_rate": 9.62363147162098e-05, |
|
"loss": 4.9029, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.5301204819277109, |
|
"grad_norm": 3.1446473598480225, |
|
"learning_rate": 9.589125507238233e-05, |
|
"loss": 4.8181, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.5461847389558233, |
|
"grad_norm": 3.476330041885376, |
|
"learning_rate": 9.553173864274567e-05, |
|
"loss": 5.1613, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.5622489959839357, |
|
"grad_norm": 3.2551815509796143, |
|
"learning_rate": 9.515787868312619e-05, |
|
"loss": 4.7212, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.5783132530120482, |
|
"grad_norm": 3.1978161334991455, |
|
"learning_rate": 9.476979296788747e-05, |
|
"loss": 4.7767, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.5943775100401606, |
|
"grad_norm": 2.728731632232666, |
|
"learning_rate": 9.436760375282859e-05, |
|
"loss": 4.4009, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.6104417670682731, |
|
"grad_norm": 2.96553373336792, |
|
"learning_rate": 9.395143773667088e-05, |
|
"loss": 5.6028, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.6265060240963856, |
|
"grad_norm": 3.6579477787017822, |
|
"learning_rate": 9.352142602114486e-05, |
|
"loss": 4.9442, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.642570281124498, |
|
"grad_norm": 2.903290271759033, |
|
"learning_rate": 9.30777040696903e-05, |
|
"loss": 4.5713, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.6586345381526104, |
|
"grad_norm": 3.4129068851470947, |
|
"learning_rate": 9.262041166478214e-05, |
|
"loss": 4.6581, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.6746987951807228, |
|
"grad_norm": 2.9026341438293457, |
|
"learning_rate": 9.214969286389576e-05, |
|
"loss": 3.8248, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.6907630522088354, |
|
"grad_norm": 2.8055315017700195, |
|
"learning_rate": 9.166569595412575e-05, |
|
"loss": 3.7736, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.7068273092369478, |
|
"grad_norm": 3.1325647830963135, |
|
"learning_rate": 9.116857340547202e-05, |
|
"loss": 4.526, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.7228915662650602, |
|
"grad_norm": 3.3222157955169678, |
|
"learning_rate": 9.065848182280833e-05, |
|
"loss": 3.4443, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.7389558232931727, |
|
"grad_norm": 9.711248397827148, |
|
"learning_rate": 9.013558189654819e-05, |
|
"loss": 4.5834, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.7550200803212851, |
|
"grad_norm": 3.783061981201172, |
|
"learning_rate": 8.96000383520237e-05, |
|
"loss": 4.9376, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.7710843373493976, |
|
"grad_norm": 3.666478157043457, |
|
"learning_rate": 8.905201989759341e-05, |
|
"loss": 4.9693, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.7871485943775101, |
|
"grad_norm": 2.7563154697418213, |
|
"learning_rate": 8.849169917149531e-05, |
|
"loss": 4.0881, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.8032128514056225, |
|
"grad_norm": 3.036672353744507, |
|
"learning_rate": 8.791925268746193e-05, |
|
"loss": 4.4524, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.8032128514056225, |
|
"eval_loss": 1.1385369300842285, |
|
"eval_runtime": 5.1628, |
|
"eval_samples_per_second": 20.338, |
|
"eval_steps_per_second": 5.23, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.8192771084337349, |
|
"grad_norm": 2.9426279067993164, |
|
"learning_rate": 8.73348607791144e-05, |
|
"loss": 4.6989, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.8353413654618473, |
|
"grad_norm": 3.0048654079437256, |
|
"learning_rate": 8.673870754315336e-05, |
|
"loss": 4.2927, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.8514056224899599, |
|
"grad_norm": 3.0875372886657715, |
|
"learning_rate": 8.613098078136437e-05, |
|
"loss": 4.8515, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.8674698795180723, |
|
"grad_norm": 3.7591512203216553, |
|
"learning_rate": 8.551187194145592e-05, |
|
"loss": 4.032, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.8835341365461847, |
|
"grad_norm": 2.896423816680908, |
|
"learning_rate": 8.488157605674925e-05, |
|
"loss": 4.0854, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.8995983935742972, |
|
"grad_norm": 2.782850503921509, |
|
"learning_rate": 8.424029168473829e-05, |
|
"loss": 4.6205, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.9156626506024096, |
|
"grad_norm": 2.9316649436950684, |
|
"learning_rate": 8.358822084453965e-05, |
|
"loss": 3.9744, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.9317269076305221, |
|
"grad_norm": 2.7768986225128174, |
|
"learning_rate": 8.292556895325194e-05, |
|
"loss": 4.2567, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.9477911646586346, |
|
"grad_norm": 3.6404924392700195, |
|
"learning_rate": 8.225254476124478e-05, |
|
"loss": 4.2149, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.963855421686747, |
|
"grad_norm": 3.508265972137451, |
|
"learning_rate": 8.156936028639767e-05, |
|
"loss": 3.9642, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.9799196787148594, |
|
"grad_norm": 3.133049726486206, |
|
"learning_rate": 8.08762307473096e-05, |
|
"loss": 4.7128, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.9959839357429718, |
|
"grad_norm": 3.1079959869384766, |
|
"learning_rate": 8.01733744955002e-05, |
|
"loss": 4.2072, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 1.0120481927710843, |
|
"grad_norm": 5.682511806488037, |
|
"learning_rate": 7.946101294662418e-05, |
|
"loss": 5.271, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 1.0281124497991967, |
|
"grad_norm": 2.7712900638580322, |
|
"learning_rate": 7.873937051072035e-05, |
|
"loss": 4.2853, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 1.0441767068273093, |
|
"grad_norm": 2.630441188812256, |
|
"learning_rate": 7.80086745215173e-05, |
|
"loss": 3.7277, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.0602409638554218, |
|
"grad_norm": 2.798489570617676, |
|
"learning_rate": 7.726915516481824e-05, |
|
"loss": 4.2442, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 1.0763052208835342, |
|
"grad_norm": 2.5500569343566895, |
|
"learning_rate": 7.652104540598712e-05, |
|
"loss": 3.378, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 1.0923694779116466, |
|
"grad_norm": 2.5407848358154297, |
|
"learning_rate": 7.57645809165594e-05, |
|
"loss": 4.2423, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 1.108433734939759, |
|
"grad_norm": 2.9551448822021484, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 4.8987, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 1.1244979919678715, |
|
"grad_norm": 2.51187801361084, |
|
"learning_rate": 7.422754351663252e-05, |
|
"loss": 3.8333, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.140562248995984, |
|
"grad_norm": 2.5731523036956787, |
|
"learning_rate": 7.344745480776257e-05, |
|
"loss": 3.8188, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 1.1566265060240963, |
|
"grad_norm": 2.586987018585205, |
|
"learning_rate": 7.265997961901987e-05, |
|
"loss": 3.5334, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 1.1726907630522088, |
|
"grad_norm": 2.6580755710601807, |
|
"learning_rate": 7.186536602294278e-05, |
|
"loss": 3.6089, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 1.1887550200803212, |
|
"grad_norm": 2.7322309017181396, |
|
"learning_rate": 7.106386434082979e-05, |
|
"loss": 3.5912, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 1.2048192771084336, |
|
"grad_norm": 2.6883673667907715, |
|
"learning_rate": 7.025572706388268e-05, |
|
"loss": 3.4219, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.2208835341365463, |
|
"grad_norm": 3.3504085540771484, |
|
"learning_rate": 6.944120877366604e-05, |
|
"loss": 3.5701, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 1.2369477911646587, |
|
"grad_norm": 2.829219102859497, |
|
"learning_rate": 6.86205660619083e-05, |
|
"loss": 2.6253, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 1.2530120481927711, |
|
"grad_norm": 3.903244733810425, |
|
"learning_rate": 6.779405744966954e-05, |
|
"loss": 3.8675, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 1.2690763052208835, |
|
"grad_norm": 3.95448637008667, |
|
"learning_rate": 6.696194330590151e-05, |
|
"loss": 4.0529, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 1.285140562248996, |
|
"grad_norm": 2.8472986221313477, |
|
"learning_rate": 6.612448576542545e-05, |
|
"loss": 3.1193, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.3012048192771084, |
|
"grad_norm": 3.7642226219177246, |
|
"learning_rate": 6.528194864635369e-05, |
|
"loss": 3.6889, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 1.3172690763052208, |
|
"grad_norm": 3.163255214691162, |
|
"learning_rate": 6.443459736698105e-05, |
|
"loss": 3.8981, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 1.3333333333333333, |
|
"grad_norm": 3.7267022132873535, |
|
"learning_rate": 6.358269886217194e-05, |
|
"loss": 3.7673, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 1.3493975903614457, |
|
"grad_norm": 3.128511428833008, |
|
"learning_rate": 6.272652149926988e-05, |
|
"loss": 3.1023, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 1.3654618473895583, |
|
"grad_norm": 3.0362024307250977, |
|
"learning_rate": 6.186633499355576e-05, |
|
"loss": 3.2074, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.3815261044176708, |
|
"grad_norm": 3.2987911701202393, |
|
"learning_rate": 6.100241032328124e-05, |
|
"loss": 3.4199, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 1.3975903614457832, |
|
"grad_norm": 3.5199403762817383, |
|
"learning_rate": 6.013501964430468e-05, |
|
"loss": 3.5126, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 1.4136546184738956, |
|
"grad_norm": 3.5599286556243896, |
|
"learning_rate": 5.9264436204355724e-05, |
|
"loss": 3.6378, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 1.429718875502008, |
|
"grad_norm": 3.366896152496338, |
|
"learning_rate": 5.839093425695609e-05, |
|
"loss": 3.1565, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 1.4457831325301205, |
|
"grad_norm": 3.9521143436431885, |
|
"learning_rate": 5.751478897502352e-05, |
|
"loss": 3.5207, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.461847389558233, |
|
"grad_norm": 3.8751003742218018, |
|
"learning_rate": 5.6636276364186105e-05, |
|
"loss": 3.0394, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 1.4779116465863453, |
|
"grad_norm": 4.066254615783691, |
|
"learning_rate": 5.5755673175834145e-05, |
|
"loss": 3.2113, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 1.4939759036144578, |
|
"grad_norm": 4.048666954040527, |
|
"learning_rate": 5.487325681993733e-05, |
|
"loss": 3.5889, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 1.5100401606425704, |
|
"grad_norm": 4.316978931427002, |
|
"learning_rate": 5.3989305277654156e-05, |
|
"loss": 3.9679, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 1.5261044176706826, |
|
"grad_norm": 3.883434534072876, |
|
"learning_rate": 5.31040970137617e-05, |
|
"loss": 3.3537, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.5421686746987953, |
|
"grad_norm": 4.3469672203063965, |
|
"learning_rate": 5.221791088893282e-05, |
|
"loss": 4.7528, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 1.5582329317269075, |
|
"grad_norm": 3.930917739868164, |
|
"learning_rate": 5.133102607188874e-05, |
|
"loss": 3.8965, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 1.5742971887550201, |
|
"grad_norm": 3.873133420944214, |
|
"learning_rate": 5.044372195145455e-05, |
|
"loss": 3.3711, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 1.5903614457831325, |
|
"grad_norm": 3.6565146446228027, |
|
"learning_rate": 4.955627804854545e-05, |
|
"loss": 3.6828, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 1.606425702811245, |
|
"grad_norm": 3.9164047241210938, |
|
"learning_rate": 4.866897392811126e-05, |
|
"loss": 4.1462, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.606425702811245, |
|
"eval_loss": 1.1504853963851929, |
|
"eval_runtime": 5.1386, |
|
"eval_samples_per_second": 20.434, |
|
"eval_steps_per_second": 5.254, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.6224899598393574, |
|
"grad_norm": 4.047707557678223, |
|
"learning_rate": 4.7782089111067176e-05, |
|
"loss": 3.7656, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 1.6385542168674698, |
|
"grad_norm": 3.6224374771118164, |
|
"learning_rate": 4.6895902986238304e-05, |
|
"loss": 3.7808, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 1.6546184738955825, |
|
"grad_norm": 3.4380745887756348, |
|
"learning_rate": 4.601069472234584e-05, |
|
"loss": 2.9976, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 1.6706827309236947, |
|
"grad_norm": 3.690455436706543, |
|
"learning_rate": 4.512674318006268e-05, |
|
"loss": 2.969, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 1.6867469879518073, |
|
"grad_norm": 3.4021284580230713, |
|
"learning_rate": 4.424432682416585e-05, |
|
"loss": 3.0345, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 1.7028112449799195, |
|
"grad_norm": 3.821765422821045, |
|
"learning_rate": 4.336372363581391e-05, |
|
"loss": 3.0546, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 1.7188755020080322, |
|
"grad_norm": 3.7544493675231934, |
|
"learning_rate": 4.2485211024976496e-05, |
|
"loss": 2.6392, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 1.7349397590361446, |
|
"grad_norm": 3.898599863052368, |
|
"learning_rate": 4.160906574304392e-05, |
|
"loss": 3.675, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 1.751004016064257, |
|
"grad_norm": 3.801405429840088, |
|
"learning_rate": 4.0735563795644294e-05, |
|
"loss": 3.8661, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 1.7670682730923695, |
|
"grad_norm": 4.313454627990723, |
|
"learning_rate": 3.986498035569532e-05, |
|
"loss": 4.2648, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.783132530120482, |
|
"grad_norm": 3.6857495307922363, |
|
"learning_rate": 3.899758967671878e-05, |
|
"loss": 3.2822, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 1.7991967871485943, |
|
"grad_norm": 3.5725905895233154, |
|
"learning_rate": 3.8133665006444255e-05, |
|
"loss": 3.4181, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 1.8152610441767068, |
|
"grad_norm": 3.708075761795044, |
|
"learning_rate": 3.727347850073012e-05, |
|
"loss": 3.0178, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 1.8313253012048194, |
|
"grad_norm": 4.1876220703125, |
|
"learning_rate": 3.641730113782807e-05, |
|
"loss": 3.5986, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 1.8473895582329316, |
|
"grad_norm": 3.863704204559326, |
|
"learning_rate": 3.556540263301896e-05, |
|
"loss": 3.5571, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.8634538152610443, |
|
"grad_norm": 3.584493398666382, |
|
"learning_rate": 3.47180513536463e-05, |
|
"loss": 3.3766, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 1.8795180722891565, |
|
"grad_norm": 3.944830894470215, |
|
"learning_rate": 3.3875514234574556e-05, |
|
"loss": 3.5299, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 1.895582329317269, |
|
"grad_norm": 3.59722900390625, |
|
"learning_rate": 3.303805669409848e-05, |
|
"loss": 3.1769, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 1.9116465863453815, |
|
"grad_norm": 3.9105989933013916, |
|
"learning_rate": 3.2205942550330456e-05, |
|
"loss": 3.3873, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 1.927710843373494, |
|
"grad_norm": 3.794104814529419, |
|
"learning_rate": 3.1379433938091696e-05, |
|
"loss": 3.3131, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.9437751004016064, |
|
"grad_norm": 3.642608165740967, |
|
"learning_rate": 3.055879122633397e-05, |
|
"loss": 2.8884, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 1.9598393574297188, |
|
"grad_norm": 4.019187927246094, |
|
"learning_rate": 2.9744272936117323e-05, |
|
"loss": 2.927, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 1.9759036144578315, |
|
"grad_norm": 4.073461532592773, |
|
"learning_rate": 2.8936135659170216e-05, |
|
"loss": 3.2049, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 1.9919678714859437, |
|
"grad_norm": 3.866515874862671, |
|
"learning_rate": 2.8134633977057235e-05, |
|
"loss": 3.0155, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 2.0080321285140563, |
|
"grad_norm": 6.784224033355713, |
|
"learning_rate": 2.7340020380980146e-05, |
|
"loss": 2.7591, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.0240963855421685, |
|
"grad_norm": 3.676565408706665, |
|
"learning_rate": 2.655254519223746e-05, |
|
"loss": 3.2366, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 2.040160642570281, |
|
"grad_norm": 3.3340706825256348, |
|
"learning_rate": 2.5772456483367497e-05, |
|
"loss": 2.8017, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 2.0562248995983934, |
|
"grad_norm": 3.5175349712371826, |
|
"learning_rate": 2.500000000000001e-05, |
|
"loss": 2.9206, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 2.072289156626506, |
|
"grad_norm": 3.3527610301971436, |
|
"learning_rate": 2.4235419083440613e-05, |
|
"loss": 2.5938, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 2.0883534136546187, |
|
"grad_norm": 3.2494187355041504, |
|
"learning_rate": 2.347895459401288e-05, |
|
"loss": 2.2693, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.104417670682731, |
|
"grad_norm": 4.090683460235596, |
|
"learning_rate": 2.2730844835181757e-05, |
|
"loss": 2.9655, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 2.1204819277108435, |
|
"grad_norm": 3.746830701828003, |
|
"learning_rate": 2.1991325478482694e-05, |
|
"loss": 2.6929, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 2.1365461847389557, |
|
"grad_norm": 3.2926580905914307, |
|
"learning_rate": 2.126062948927966e-05, |
|
"loss": 2.395, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 2.1526104417670684, |
|
"grad_norm": 4.051050662994385, |
|
"learning_rate": 2.053898705337583e-05, |
|
"loss": 2.8629, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 2.1686746987951806, |
|
"grad_norm": 3.782912254333496, |
|
"learning_rate": 1.9826625504499806e-05, |
|
"loss": 2.7832, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 2.1847389558232932, |
|
"grad_norm": 3.7877721786499023, |
|
"learning_rate": 1.912376925269041e-05, |
|
"loss": 1.9438, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 2.2008032128514055, |
|
"grad_norm": 3.5854597091674805, |
|
"learning_rate": 1.8430639713602316e-05, |
|
"loss": 1.9395, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 2.216867469879518, |
|
"grad_norm": 4.118504524230957, |
|
"learning_rate": 1.7747455238755223e-05, |
|
"loss": 2.7666, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 2.2329317269076308, |
|
"grad_norm": 4.0326924324035645, |
|
"learning_rate": 1.7074431046748075e-05, |
|
"loss": 1.8387, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 2.248995983935743, |
|
"grad_norm": 4.268743991851807, |
|
"learning_rate": 1.641177915546036e-05, |
|
"loss": 2.5107, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.2650602409638556, |
|
"grad_norm": 4.364184856414795, |
|
"learning_rate": 1.5759708315261722e-05, |
|
"loss": 2.4838, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 2.281124497991968, |
|
"grad_norm": 4.482365131378174, |
|
"learning_rate": 1.5118423943250771e-05, |
|
"loss": 2.5102, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 2.2971887550200805, |
|
"grad_norm": 4.34283447265625, |
|
"learning_rate": 1.4488128058544098e-05, |
|
"loss": 2.5356, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 2.3132530120481927, |
|
"grad_norm": 4.787532329559326, |
|
"learning_rate": 1.3869019218635642e-05, |
|
"loss": 3.0075, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 2.3293172690763053, |
|
"grad_norm": 4.763615608215332, |
|
"learning_rate": 1.3261292456846647e-05, |
|
"loss": 2.6659, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 2.3453815261044175, |
|
"grad_norm": 4.660789489746094, |
|
"learning_rate": 1.2665139220885613e-05, |
|
"loss": 2.6328, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 2.36144578313253, |
|
"grad_norm": 4.864875793457031, |
|
"learning_rate": 1.2080747312538083e-05, |
|
"loss": 2.6357, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 2.3775100401606424, |
|
"grad_norm": 5.13457727432251, |
|
"learning_rate": 1.150830082850468e-05, |
|
"loss": 2.601, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 2.393574297188755, |
|
"grad_norm": 4.843163967132568, |
|
"learning_rate": 1.0947980102406596e-05, |
|
"loss": 2.1365, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 2.4096385542168672, |
|
"grad_norm": 4.978252410888672, |
|
"learning_rate": 1.0399961647976314e-05, |
|
"loss": 2.4167, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.4096385542168672, |
|
"eval_loss": 1.3296469449996948, |
|
"eval_runtime": 5.1411, |
|
"eval_samples_per_second": 20.423, |
|
"eval_steps_per_second": 5.252, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.42570281124498, |
|
"grad_norm": 5.148977279663086, |
|
"learning_rate": 9.864418103451828e-06, |
|
"loss": 2.2449, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 2.4417670682730925, |
|
"grad_norm": 4.7471466064453125, |
|
"learning_rate": 9.34151817719166e-06, |
|
"loss": 1.7333, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 2.4578313253012047, |
|
"grad_norm": 5.056032180786133, |
|
"learning_rate": 8.831426594527975e-06, |
|
"loss": 1.7489, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 2.4738955823293174, |
|
"grad_norm": 5.261003017425537, |
|
"learning_rate": 8.334304045874247e-06, |
|
"loss": 1.6044, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 2.4899598393574296, |
|
"grad_norm": 5.535795211791992, |
|
"learning_rate": 7.850307136104247e-06, |
|
"loss": 2.2367, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 2.5060240963855422, |
|
"grad_norm": 5.856411933898926, |
|
"learning_rate": 7.379588335217874e-06, |
|
"loss": 3.0989, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 2.522088353413655, |
|
"grad_norm": 5.084694862365723, |
|
"learning_rate": 6.922295930309691e-06, |
|
"loss": 2.7171, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 2.538152610441767, |
|
"grad_norm": 6.212282657623291, |
|
"learning_rate": 6.478573978855146e-06, |
|
"loss": 2.8528, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 2.5542168674698793, |
|
"grad_norm": 4.591296195983887, |
|
"learning_rate": 6.048562263329138e-06, |
|
"loss": 2.0066, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 2.570281124497992, |
|
"grad_norm": 5.471847057342529, |
|
"learning_rate": 5.6323962471714286e-06, |
|
"loss": 2.6712, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.5863453815261046, |
|
"grad_norm": 6.344399929046631, |
|
"learning_rate": 5.23020703211255e-06, |
|
"loss": 2.3738, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 2.602409638554217, |
|
"grad_norm": 4.858112335205078, |
|
"learning_rate": 4.842121316873821e-06, |
|
"loss": 2.2417, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 2.6184738955823295, |
|
"grad_norm": 5.603279113769531, |
|
"learning_rate": 4.468261357254339e-06, |
|
"loss": 2.6957, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 2.6345381526104417, |
|
"grad_norm": 5.545329570770264, |
|
"learning_rate": 4.108744927617669e-06, |
|
"loss": 2.2776, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 2.6506024096385543, |
|
"grad_norm": 4.7836222648620605, |
|
"learning_rate": 3.763685283790208e-06, |
|
"loss": 2.0821, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 2.6666666666666665, |
|
"grad_norm": 4.492027282714844, |
|
"learning_rate": 3.4331911273830784e-06, |
|
"loss": 2.031, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 2.682730923694779, |
|
"grad_norm": 6.321730136871338, |
|
"learning_rate": 3.117366571548608e-06, |
|
"loss": 2.5016, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 2.6987951807228914, |
|
"grad_norm": 8.664602279663086, |
|
"learning_rate": 2.816311108182368e-06, |
|
"loss": 2.1106, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 2.714859437751004, |
|
"grad_norm": 5.606003284454346, |
|
"learning_rate": 2.530119576580936e-06, |
|
"loss": 2.185, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 2.7309236947791167, |
|
"grad_norm": 5.336299896240234, |
|
"learning_rate": 2.258882133565404e-06, |
|
"loss": 2.0902, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.746987951807229, |
|
"grad_norm": 5.943612575531006, |
|
"learning_rate": 2.0026842250799038e-06, |
|
"loss": 3.1001, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 2.7630522088353415, |
|
"grad_norm": 5.235433101654053, |
|
"learning_rate": 1.7616065592742038e-06, |
|
"loss": 2.7069, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 2.7791164658634537, |
|
"grad_norm": 4.886165142059326, |
|
"learning_rate": 1.5357250810788314e-06, |
|
"loss": 2.3469, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 2.7951807228915664, |
|
"grad_norm": 5.185632228851318, |
|
"learning_rate": 1.3251109482806666e-06, |
|
"loss": 2.7937, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 2.8112449799196786, |
|
"grad_norm": 5.091357231140137, |
|
"learning_rate": 1.1298305091066664e-06, |
|
"loss": 2.5354, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.8273092369477912, |
|
"grad_norm": 4.6926774978637695, |
|
"learning_rate": 9.499452813226284e-07, |
|
"loss": 2.2616, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 2.8433734939759034, |
|
"grad_norm": 4.782034397125244, |
|
"learning_rate": 7.855119328537109e-07, |
|
"loss": 2.0711, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 2.859437751004016, |
|
"grad_norm": 5.071019172668457, |
|
"learning_rate": 6.365822639327723e-07, |
|
"loss": 2.1635, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 2.8755020080321287, |
|
"grad_norm": 4.6255974769592285, |
|
"learning_rate": 5.032031907821089e-07, |
|
"loss": 2.1709, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 2.891566265060241, |
|
"grad_norm": 5.228384494781494, |
|
"learning_rate": 3.854167308337708e-07, |
|
"loss": 2.1617, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 2.907630522088353, |
|
"grad_norm": 5.127053260803223, |
|
"learning_rate": 2.8325998949314536e-07, |
|
"loss": 2.1208, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 2.923694779116466, |
|
"grad_norm": 4.4402055740356445, |
|
"learning_rate": 1.9676514844987337e-07, |
|
"loss": 1.9813, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 2.9397590361445785, |
|
"grad_norm": 6.084360122680664, |
|
"learning_rate": 1.2595945553992573e-07, |
|
"loss": 2.1314, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 2.9558232931726907, |
|
"grad_norm": 4.875538349151611, |
|
"learning_rate": 7.086521616190279e-08, |
|
"loss": 1.9403, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 2.9718875502008033, |
|
"grad_norm": 4.680423736572266, |
|
"learning_rate": 3.149978625032191e-08, |
|
"loss": 1.7237, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 2.9879518072289155, |
|
"grad_norm": 5.582780838012695, |
|
"learning_rate": 7.875566808107637e-09, |
|
"loss": 2.8136, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 3.004016064257028, |
|
"grad_norm": 5.710657596588135, |
|
"learning_rate": 0.0, |
|
"loss": 1.9331, |
|
"step": 187 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 187, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 50, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 2 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.4076236595304858e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|