|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.01241304905619208, |
|
"eval_steps": 500, |
|
"global_step": 192, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 6.465129716766708e-05, |
|
"grad_norm": 52.735591888427734, |
|
"learning_rate": 6.451612903225807e-08, |
|
"loss": 1.9276, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.00012930259433533415, |
|
"grad_norm": 54.205406188964844, |
|
"learning_rate": 1.2903225806451614e-07, |
|
"loss": 1.8498, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.00019395389150300124, |
|
"grad_norm": 55.422950744628906, |
|
"learning_rate": 1.935483870967742e-07, |
|
"loss": 2.0204, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0002586051886706683, |
|
"grad_norm": 52.57908248901367, |
|
"learning_rate": 2.580645161290323e-07, |
|
"loss": 1.9851, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0003232564858383354, |
|
"grad_norm": 53.102073669433594, |
|
"learning_rate": 3.2258064516129035e-07, |
|
"loss": 1.9575, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0003879077830060025, |
|
"grad_norm": 39.96001052856445, |
|
"learning_rate": 3.870967741935484e-07, |
|
"loss": 1.9575, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.00045255908017366957, |
|
"grad_norm": 32.277435302734375, |
|
"learning_rate": 4.5161290322580644e-07, |
|
"loss": 1.8748, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0005172103773413366, |
|
"grad_norm": 35.69306945800781, |
|
"learning_rate": 5.161290322580646e-07, |
|
"loss": 1.8319, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.0005818616745090037, |
|
"grad_norm": 32.267822265625, |
|
"learning_rate": 5.806451612903227e-07, |
|
"loss": 1.8492, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.0006465129716766708, |
|
"grad_norm": 15.390063285827637, |
|
"learning_rate": 6.451612903225807e-07, |
|
"loss": 1.8046, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0007111642688443379, |
|
"grad_norm": 25.763904571533203, |
|
"learning_rate": 7.096774193548388e-07, |
|
"loss": 1.7649, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.000775815566012005, |
|
"grad_norm": 29.1771297454834, |
|
"learning_rate": 7.741935483870968e-07, |
|
"loss": 1.7684, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.000840466863179672, |
|
"grad_norm": 19.140987396240234, |
|
"learning_rate": 8.38709677419355e-07, |
|
"loss": 1.7782, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0009051181603473391, |
|
"grad_norm": 13.840346336364746, |
|
"learning_rate": 9.032258064516129e-07, |
|
"loss": 1.8303, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.0009697694575150061, |
|
"grad_norm": 26.080778121948242, |
|
"learning_rate": 9.67741935483871e-07, |
|
"loss": 1.776, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0010344207546826732, |
|
"grad_norm": 34.16382598876953, |
|
"learning_rate": 1.0322580645161291e-06, |
|
"loss": 1.6359, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.0010990720518503404, |
|
"grad_norm": 25.75543785095215, |
|
"learning_rate": 1.0967741935483872e-06, |
|
"loss": 1.6014, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0011637233490180074, |
|
"grad_norm": 11.885198593139648, |
|
"learning_rate": 1.1612903225806454e-06, |
|
"loss": 1.6831, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.0012283746461856746, |
|
"grad_norm": 14.924345016479492, |
|
"learning_rate": 1.2258064516129033e-06, |
|
"loss": 1.7223, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.0012930259433533416, |
|
"grad_norm": 18.01887321472168, |
|
"learning_rate": 1.2903225806451614e-06, |
|
"loss": 1.6811, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0013576772405210086, |
|
"grad_norm": 12.34964656829834, |
|
"learning_rate": 1.3548387096774195e-06, |
|
"loss": 1.6169, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.0014223285376886757, |
|
"grad_norm": 11.162965774536133, |
|
"learning_rate": 1.4193548387096776e-06, |
|
"loss": 1.6814, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.0014869798348563427, |
|
"grad_norm": 8.694445610046387, |
|
"learning_rate": 1.4838709677419356e-06, |
|
"loss": 1.5999, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.00155163113202401, |
|
"grad_norm": 9.510068893432617, |
|
"learning_rate": 1.5483870967741937e-06, |
|
"loss": 1.6327, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.001616282429191677, |
|
"grad_norm": 9.788911819458008, |
|
"learning_rate": 1.6129032258064516e-06, |
|
"loss": 1.6985, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.001680933726359344, |
|
"grad_norm": 8.04429817199707, |
|
"learning_rate": 1.67741935483871e-06, |
|
"loss": 1.6618, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.001745585023527011, |
|
"grad_norm": 7.545624256134033, |
|
"learning_rate": 1.7419354838709678e-06, |
|
"loss": 1.5472, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.0018102363206946783, |
|
"grad_norm": 7.069305419921875, |
|
"learning_rate": 1.8064516129032258e-06, |
|
"loss": 1.643, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.0018748876178623453, |
|
"grad_norm": 9.195314407348633, |
|
"learning_rate": 1.870967741935484e-06, |
|
"loss": 1.5956, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.0019395389150300122, |
|
"grad_norm": 6.884894371032715, |
|
"learning_rate": 1.935483870967742e-06, |
|
"loss": 1.5752, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.0020041902121976794, |
|
"grad_norm": 7.165380477905273, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 1.5812, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.0020688415093653464, |
|
"grad_norm": 6.563286781311035, |
|
"learning_rate": 2.0645161290322582e-06, |
|
"loss": 1.5938, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.0021334928065330134, |
|
"grad_norm": 5.998967170715332, |
|
"learning_rate": 2.129032258064516e-06, |
|
"loss": 1.5716, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.002198144103700681, |
|
"grad_norm": 6.407791614532471, |
|
"learning_rate": 2.1935483870967745e-06, |
|
"loss": 1.5848, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.002262795400868348, |
|
"grad_norm": 6.110191822052002, |
|
"learning_rate": 2.2580645161290324e-06, |
|
"loss": 1.643, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0023274466980360148, |
|
"grad_norm": 6.058048248291016, |
|
"learning_rate": 2.3225806451612907e-06, |
|
"loss": 1.5242, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.0023920979952036818, |
|
"grad_norm": 7.348267078399658, |
|
"learning_rate": 2.3870967741935486e-06, |
|
"loss": 1.5112, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.002456749292371349, |
|
"grad_norm": 6.676770210266113, |
|
"learning_rate": 2.4516129032258066e-06, |
|
"loss": 1.5184, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.002521400589539016, |
|
"grad_norm": 9.655611038208008, |
|
"learning_rate": 2.5161290322580645e-06, |
|
"loss": 1.4867, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.002586051886706683, |
|
"grad_norm": 10.357527732849121, |
|
"learning_rate": 2.580645161290323e-06, |
|
"loss": 1.5255, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.00265070318387435, |
|
"grad_norm": 6.523240566253662, |
|
"learning_rate": 2.645161290322581e-06, |
|
"loss": 1.5326, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.002715354481042017, |
|
"grad_norm": 5.7459282875061035, |
|
"learning_rate": 2.709677419354839e-06, |
|
"loss": 1.5533, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.0027800057782096845, |
|
"grad_norm": 5.776258945465088, |
|
"learning_rate": 2.774193548387097e-06, |
|
"loss": 1.5261, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.0028446570753773515, |
|
"grad_norm": 7.176516056060791, |
|
"learning_rate": 2.8387096774193553e-06, |
|
"loss": 1.4434, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.0029093083725450185, |
|
"grad_norm": 6.083931922912598, |
|
"learning_rate": 2.903225806451613e-06, |
|
"loss": 1.5127, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.0029739596697126855, |
|
"grad_norm": 8.212278366088867, |
|
"learning_rate": 2.967741935483871e-06, |
|
"loss": 1.569, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.003038610966880353, |
|
"grad_norm": 6.138173580169678, |
|
"learning_rate": 3.0322580645161295e-06, |
|
"loss": 1.4985, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.00310326226404802, |
|
"grad_norm": 9.2051362991333, |
|
"learning_rate": 3.0967741935483874e-06, |
|
"loss": 1.4528, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.003167913561215687, |
|
"grad_norm": 9.6658353805542, |
|
"learning_rate": 3.1612903225806453e-06, |
|
"loss": 1.4432, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.003232564858383354, |
|
"grad_norm": 6.775389671325684, |
|
"learning_rate": 3.225806451612903e-06, |
|
"loss": 1.4649, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.003297216155551021, |
|
"grad_norm": 11.512835502624512, |
|
"learning_rate": 3.2903225806451615e-06, |
|
"loss": 1.4106, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.003361867452718688, |
|
"grad_norm": 5.44810676574707, |
|
"learning_rate": 3.35483870967742e-06, |
|
"loss": 1.4714, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.003426518749886355, |
|
"grad_norm": 6.4585185050964355, |
|
"learning_rate": 3.4193548387096773e-06, |
|
"loss": 1.5064, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.003491170047054022, |
|
"grad_norm": 8.523478507995605, |
|
"learning_rate": 3.4838709677419357e-06, |
|
"loss": 1.3392, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.003555821344221689, |
|
"grad_norm": 5.496897220611572, |
|
"learning_rate": 3.548387096774194e-06, |
|
"loss": 1.5616, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.0036204726413893566, |
|
"grad_norm": 8.096810340881348, |
|
"learning_rate": 3.6129032258064515e-06, |
|
"loss": 1.4941, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.0036851239385570236, |
|
"grad_norm": 6.6727705001831055, |
|
"learning_rate": 3.67741935483871e-06, |
|
"loss": 1.5466, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.0037497752357246905, |
|
"grad_norm": 7.022820949554443, |
|
"learning_rate": 3.741935483870968e-06, |
|
"loss": 1.4893, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.0038144265328923575, |
|
"grad_norm": 6.383000373840332, |
|
"learning_rate": 3.8064516129032257e-06, |
|
"loss": 1.4271, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.0038790778300600245, |
|
"grad_norm": 5.412594318389893, |
|
"learning_rate": 3.870967741935484e-06, |
|
"loss": 1.5204, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.003943729127227692, |
|
"grad_norm": 6.230757713317871, |
|
"learning_rate": 3.935483870967742e-06, |
|
"loss": 1.4508, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.004008380424395359, |
|
"grad_norm": 7.456300258636475, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.4439, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.004073031721563026, |
|
"grad_norm": 5.5062432289123535, |
|
"learning_rate": 4.064516129032259e-06, |
|
"loss": 1.4589, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.004137683018730693, |
|
"grad_norm": 6.719184398651123, |
|
"learning_rate": 4.1290322580645165e-06, |
|
"loss": 1.4779, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.00420233431589836, |
|
"grad_norm": 5.637788772583008, |
|
"learning_rate": 4.193548387096774e-06, |
|
"loss": 1.6323, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.004266985613066027, |
|
"grad_norm": 5.063558578491211, |
|
"learning_rate": 4.258064516129032e-06, |
|
"loss": 1.5131, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.004331636910233695, |
|
"grad_norm": 6.299156188964844, |
|
"learning_rate": 4.32258064516129e-06, |
|
"loss": 1.4999, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.004396288207401362, |
|
"grad_norm": 5.770033836364746, |
|
"learning_rate": 4.387096774193549e-06, |
|
"loss": 1.3813, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.004460939504569029, |
|
"grad_norm": 7.366373062133789, |
|
"learning_rate": 4.451612903225807e-06, |
|
"loss": 1.5073, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.004525590801736696, |
|
"grad_norm": 5.583267688751221, |
|
"learning_rate": 4.516129032258065e-06, |
|
"loss": 1.4769, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.004590242098904363, |
|
"grad_norm": 9.454716682434082, |
|
"learning_rate": 4.580645161290323e-06, |
|
"loss": 1.5203, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.0046548933960720296, |
|
"grad_norm": 6.360778331756592, |
|
"learning_rate": 4.6451612903225815e-06, |
|
"loss": 1.4848, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.0047195446932396965, |
|
"grad_norm": 4.729072093963623, |
|
"learning_rate": 4.7096774193548385e-06, |
|
"loss": 1.4475, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.0047841959904073635, |
|
"grad_norm": 6.975382328033447, |
|
"learning_rate": 4.774193548387097e-06, |
|
"loss": 1.5431, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.0048488472875750305, |
|
"grad_norm": 6.034017562866211, |
|
"learning_rate": 4.838709677419355e-06, |
|
"loss": 1.4028, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.004913498584742698, |
|
"grad_norm": 5.913821697235107, |
|
"learning_rate": 4.903225806451613e-06, |
|
"loss": 1.4364, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.004978149881910365, |
|
"grad_norm": 9.405481338500977, |
|
"learning_rate": 4.967741935483871e-06, |
|
"loss": 1.5207, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.005042801179078032, |
|
"grad_norm": 6.983378887176514, |
|
"learning_rate": 5.032258064516129e-06, |
|
"loss": 1.4487, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.005107452476245699, |
|
"grad_norm": 5.04365348815918, |
|
"learning_rate": 5.096774193548387e-06, |
|
"loss": 1.4282, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.005172103773413366, |
|
"grad_norm": 7.815653324127197, |
|
"learning_rate": 5.161290322580646e-06, |
|
"loss": 1.3231, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.005236755070581033, |
|
"grad_norm": 6.0287041664123535, |
|
"learning_rate": 5.2258064516129035e-06, |
|
"loss": 1.4106, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.0053014063677487, |
|
"grad_norm": 5.727312088012695, |
|
"learning_rate": 5.290322580645162e-06, |
|
"loss": 1.52, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.005366057664916367, |
|
"grad_norm": 4.75112771987915, |
|
"learning_rate": 5.35483870967742e-06, |
|
"loss": 1.434, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.005430708962084034, |
|
"grad_norm": 5.614027500152588, |
|
"learning_rate": 5.419354838709678e-06, |
|
"loss": 1.4501, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.005495360259251702, |
|
"grad_norm": 6.246868133544922, |
|
"learning_rate": 5.483870967741935e-06, |
|
"loss": 1.4065, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.005560011556419369, |
|
"grad_norm": 4.8930559158325195, |
|
"learning_rate": 5.548387096774194e-06, |
|
"loss": 1.4059, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.005624662853587036, |
|
"grad_norm": 9.081551551818848, |
|
"learning_rate": 5.612903225806452e-06, |
|
"loss": 1.4045, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.005689314150754703, |
|
"grad_norm": 6.593941688537598, |
|
"learning_rate": 5.677419354838711e-06, |
|
"loss": 1.4229, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.00575396544792237, |
|
"grad_norm": 4.863624095916748, |
|
"learning_rate": 5.7419354838709685e-06, |
|
"loss": 1.4073, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.005818616745090037, |
|
"grad_norm": 5.167389392852783, |
|
"learning_rate": 5.806451612903226e-06, |
|
"loss": 1.5046, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.005883268042257704, |
|
"grad_norm": 4.816722869873047, |
|
"learning_rate": 5.8709677419354835e-06, |
|
"loss": 1.4358, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.005947919339425371, |
|
"grad_norm": 5.505555629730225, |
|
"learning_rate": 5.935483870967742e-06, |
|
"loss": 1.5727, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.006012570636593038, |
|
"grad_norm": 5.764698505401611, |
|
"learning_rate": 6e-06, |
|
"loss": 1.4799, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.006077221933760706, |
|
"grad_norm": 5.312406063079834, |
|
"learning_rate": 6.064516129032259e-06, |
|
"loss": 1.5233, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.006141873230928373, |
|
"grad_norm": 6.0215253829956055, |
|
"learning_rate": 6.129032258064517e-06, |
|
"loss": 1.3764, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.00620652452809604, |
|
"grad_norm": 6.582176208496094, |
|
"learning_rate": 6.193548387096775e-06, |
|
"loss": 1.3687, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.006271175825263707, |
|
"grad_norm": 4.946031093597412, |
|
"learning_rate": 6.2580645161290335e-06, |
|
"loss": 1.4276, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.006335827122431374, |
|
"grad_norm": 5.586654186248779, |
|
"learning_rate": 6.3225806451612906e-06, |
|
"loss": 1.4366, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.006400478419599041, |
|
"grad_norm": 6.844956874847412, |
|
"learning_rate": 6.3870967741935485e-06, |
|
"loss": 1.3953, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.006465129716766708, |
|
"grad_norm": 5.709033012390137, |
|
"learning_rate": 6.451612903225806e-06, |
|
"loss": 1.441, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.006529781013934375, |
|
"grad_norm": 5.266170501708984, |
|
"learning_rate": 6.516129032258065e-06, |
|
"loss": 1.5336, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.006594432311102042, |
|
"grad_norm": 5.025949954986572, |
|
"learning_rate": 6.580645161290323e-06, |
|
"loss": 1.3602, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.0066590836082697094, |
|
"grad_norm": 6.4935126304626465, |
|
"learning_rate": 6.645161290322582e-06, |
|
"loss": 1.4612, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.006723734905437376, |
|
"grad_norm": 5.3272881507873535, |
|
"learning_rate": 6.70967741935484e-06, |
|
"loss": 1.4056, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.006788386202605043, |
|
"grad_norm": 5.2499213218688965, |
|
"learning_rate": 6.774193548387097e-06, |
|
"loss": 1.4908, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.00685303749977271, |
|
"grad_norm": 5.568053722381592, |
|
"learning_rate": 6.838709677419355e-06, |
|
"loss": 1.3711, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.006917688796940377, |
|
"grad_norm": 5.129876613616943, |
|
"learning_rate": 6.9032258064516135e-06, |
|
"loss": 1.3819, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.006982340094108044, |
|
"grad_norm": 6.102237224578857, |
|
"learning_rate": 6.967741935483871e-06, |
|
"loss": 1.4534, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.007046991391275711, |
|
"grad_norm": 6.690826892852783, |
|
"learning_rate": 7.03225806451613e-06, |
|
"loss": 1.3944, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.007111642688443378, |
|
"grad_norm": 4.548076629638672, |
|
"learning_rate": 7.096774193548388e-06, |
|
"loss": 1.5244, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.007176293985611045, |
|
"grad_norm": 4.901518821716309, |
|
"learning_rate": 7.161290322580646e-06, |
|
"loss": 1.4677, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.007240945282778713, |
|
"grad_norm": 5.2007155418396, |
|
"learning_rate": 7.225806451612903e-06, |
|
"loss": 1.4194, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.00730559657994638, |
|
"grad_norm": 5.50141716003418, |
|
"learning_rate": 7.290322580645162e-06, |
|
"loss": 1.4837, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.007370247877114047, |
|
"grad_norm": 5.740527153015137, |
|
"learning_rate": 7.35483870967742e-06, |
|
"loss": 1.5181, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.007434899174281714, |
|
"grad_norm": 6.510746002197266, |
|
"learning_rate": 7.4193548387096784e-06, |
|
"loss": 1.3614, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.007499550471449381, |
|
"grad_norm": 6.607003211975098, |
|
"learning_rate": 7.483870967741936e-06, |
|
"loss": 1.4396, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.007564201768617048, |
|
"grad_norm": 6.828821182250977, |
|
"learning_rate": 7.548387096774194e-06, |
|
"loss": 1.4888, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.007628853065784715, |
|
"grad_norm": 5.51243782043457, |
|
"learning_rate": 7.612903225806451e-06, |
|
"loss": 1.4666, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.007693504362952382, |
|
"grad_norm": 5.797337532043457, |
|
"learning_rate": 7.67741935483871e-06, |
|
"loss": 1.3761, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.007758155660120049, |
|
"grad_norm": 5.460038185119629, |
|
"learning_rate": 7.741935483870968e-06, |
|
"loss": 1.4361, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.007822806957287716, |
|
"grad_norm": 5.366038799285889, |
|
"learning_rate": 7.806451612903227e-06, |
|
"loss": 1.4087, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.007887458254455384, |
|
"grad_norm": 6.333535194396973, |
|
"learning_rate": 7.870967741935484e-06, |
|
"loss": 1.4527, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.00795210955162305, |
|
"grad_norm": 5.908946514129639, |
|
"learning_rate": 7.935483870967743e-06, |
|
"loss": 1.4128, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.008016760848790718, |
|
"grad_norm": 5.050029754638672, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 1.4311, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.008081412145958386, |
|
"grad_norm": 6.2919816970825195, |
|
"learning_rate": 8.064516129032258e-06, |
|
"loss": 1.3892, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.008146063443126052, |
|
"grad_norm": 5.789970874786377, |
|
"learning_rate": 8.129032258064517e-06, |
|
"loss": 1.4024, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.00821071474029372, |
|
"grad_norm": 5.259674549102783, |
|
"learning_rate": 8.193548387096774e-06, |
|
"loss": 1.3323, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.008275366037461386, |
|
"grad_norm": 5.545688152313232, |
|
"learning_rate": 8.258064516129033e-06, |
|
"loss": 1.332, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.008340017334629054, |
|
"grad_norm": 6.5449676513671875, |
|
"learning_rate": 8.322580645161292e-06, |
|
"loss": 1.3696, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.00840466863179672, |
|
"grad_norm": 4.6496806144714355, |
|
"learning_rate": 8.387096774193549e-06, |
|
"loss": 1.4208, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.008469319928964388, |
|
"grad_norm": 5.263716220855713, |
|
"learning_rate": 8.451612903225808e-06, |
|
"loss": 1.3825, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.008533971226132054, |
|
"grad_norm": 5.719542980194092, |
|
"learning_rate": 8.516129032258065e-06, |
|
"loss": 1.445, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.008598622523299721, |
|
"grad_norm": 4.548699855804443, |
|
"learning_rate": 8.580645161290323e-06, |
|
"loss": 1.3844, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.00866327382046739, |
|
"grad_norm": 4.696458339691162, |
|
"learning_rate": 8.64516129032258e-06, |
|
"loss": 1.3721, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.008727925117635055, |
|
"grad_norm": 5.142691612243652, |
|
"learning_rate": 8.70967741935484e-06, |
|
"loss": 1.3952, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.008792576414802723, |
|
"grad_norm": 4.870302677154541, |
|
"learning_rate": 8.774193548387098e-06, |
|
"loss": 1.462, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.00885722771197039, |
|
"grad_norm": 5.207381248474121, |
|
"learning_rate": 8.838709677419357e-06, |
|
"loss": 1.4555, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.008921879009138057, |
|
"grad_norm": 4.8738226890563965, |
|
"learning_rate": 8.903225806451614e-06, |
|
"loss": 1.3776, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.008986530306305723, |
|
"grad_norm": 5.064065456390381, |
|
"learning_rate": 8.967741935483871e-06, |
|
"loss": 1.4316, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.009051181603473391, |
|
"grad_norm": 5.246875286102295, |
|
"learning_rate": 9.03225806451613e-06, |
|
"loss": 1.5043, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.009115832900641057, |
|
"grad_norm": 5.119837284088135, |
|
"learning_rate": 9.096774193548388e-06, |
|
"loss": 1.4734, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.009180484197808725, |
|
"grad_norm": 6.316291332244873, |
|
"learning_rate": 9.161290322580645e-06, |
|
"loss": 1.4854, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.009245135494976393, |
|
"grad_norm": 4.688175201416016, |
|
"learning_rate": 9.225806451612904e-06, |
|
"loss": 1.4162, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.009309786792144059, |
|
"grad_norm": 5.010703086853027, |
|
"learning_rate": 9.290322580645163e-06, |
|
"loss": 1.3963, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.009374438089311727, |
|
"grad_norm": 4.959914207458496, |
|
"learning_rate": 9.35483870967742e-06, |
|
"loss": 1.3585, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.009439089386479393, |
|
"grad_norm": 4.275010585784912, |
|
"learning_rate": 9.419354838709677e-06, |
|
"loss": 1.4689, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.009503740683647061, |
|
"grad_norm": 4.852663040161133, |
|
"learning_rate": 9.483870967741936e-06, |
|
"loss": 1.4583, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.009568391980814727, |
|
"grad_norm": 4.935046195983887, |
|
"learning_rate": 9.548387096774195e-06, |
|
"loss": 1.4283, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.009633043277982395, |
|
"grad_norm": 4.662842273712158, |
|
"learning_rate": 9.612903225806453e-06, |
|
"loss": 1.4022, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.009697694575150061, |
|
"grad_norm": 6.32338285446167, |
|
"learning_rate": 9.67741935483871e-06, |
|
"loss": 1.4224, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.009762345872317729, |
|
"grad_norm": 4.785338401794434, |
|
"learning_rate": 9.74193548387097e-06, |
|
"loss": 1.4596, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.009826997169485397, |
|
"grad_norm": 4.482601642608643, |
|
"learning_rate": 9.806451612903226e-06, |
|
"loss": 1.4124, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.009891648466653063, |
|
"grad_norm": 4.771188259124756, |
|
"learning_rate": 9.870967741935485e-06, |
|
"loss": 1.4576, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.00995629976382073, |
|
"grad_norm": 4.497616291046143, |
|
"learning_rate": 9.935483870967742e-06, |
|
"loss": 1.4631, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.010020951060988397, |
|
"grad_norm": 4.109068870544434, |
|
"learning_rate": 1e-05, |
|
"loss": 1.403, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.010085602358156065, |
|
"grad_norm": 4.779379367828369, |
|
"learning_rate": 9.999999894761194e-06, |
|
"loss": 1.3295, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.01015025365532373, |
|
"grad_norm": 4.87537145614624, |
|
"learning_rate": 9.999999579044782e-06, |
|
"loss": 1.3468, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.010214904952491399, |
|
"grad_norm": 4.218890190124512, |
|
"learning_rate": 9.999999052850775e-06, |
|
"loss": 1.3863, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.010279556249659065, |
|
"grad_norm": 5.191406726837158, |
|
"learning_rate": 9.999998316179195e-06, |
|
"loss": 1.4638, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.010344207546826733, |
|
"grad_norm": 4.713414669036865, |
|
"learning_rate": 9.999997369030074e-06, |
|
"loss": 1.3629, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.0104088588439944, |
|
"grad_norm": 4.200235366821289, |
|
"learning_rate": 9.999996211403454e-06, |
|
"loss": 1.4429, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.010473510141162067, |
|
"grad_norm": 4.418542385101318, |
|
"learning_rate": 9.999994843299381e-06, |
|
"loss": 1.3971, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.010538161438329734, |
|
"grad_norm": 4.672099590301514, |
|
"learning_rate": 9.999993264717911e-06, |
|
"loss": 1.3352, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.0106028127354974, |
|
"grad_norm": 4.304332256317139, |
|
"learning_rate": 9.999991475659115e-06, |
|
"loss": 1.3116, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.010667464032665068, |
|
"grad_norm": 4.34376335144043, |
|
"learning_rate": 9.999989476123067e-06, |
|
"loss": 1.3822, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.010732115329832734, |
|
"grad_norm": 4.60739278793335, |
|
"learning_rate": 9.999987266109848e-06, |
|
"loss": 1.3655, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.010796766627000402, |
|
"grad_norm": 6.895850658416748, |
|
"learning_rate": 9.999984845619553e-06, |
|
"loss": 1.3699, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.010861417924168068, |
|
"grad_norm": 3.9377450942993164, |
|
"learning_rate": 9.999982214652286e-06, |
|
"loss": 1.4356, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.010926069221335736, |
|
"grad_norm": 4.687614440917969, |
|
"learning_rate": 9.999979373208155e-06, |
|
"loss": 1.3916, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.010990720518503404, |
|
"grad_norm": 4.686786651611328, |
|
"learning_rate": 9.99997632128728e-06, |
|
"loss": 1.3448, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.01105537181567107, |
|
"grad_norm": 4.827507495880127, |
|
"learning_rate": 9.999973058889791e-06, |
|
"loss": 1.3514, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.011120023112838738, |
|
"grad_norm": 4.3625617027282715, |
|
"learning_rate": 9.999969586015824e-06, |
|
"loss": 1.4232, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.011184674410006404, |
|
"grad_norm": 4.434966087341309, |
|
"learning_rate": 9.999965902665524e-06, |
|
"loss": 1.4217, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.011249325707174072, |
|
"grad_norm": 5.053067207336426, |
|
"learning_rate": 9.99996200883905e-06, |
|
"loss": 1.4709, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.011313977004341738, |
|
"grad_norm": 4.309473514556885, |
|
"learning_rate": 9.999957904536562e-06, |
|
"loss": 1.3668, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.011378628301509406, |
|
"grad_norm": 4.460648536682129, |
|
"learning_rate": 9.999953589758235e-06, |
|
"loss": 1.4512, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.011443279598677072, |
|
"grad_norm": 4.450823783874512, |
|
"learning_rate": 9.99994906450425e-06, |
|
"loss": 1.2986, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.01150793089584474, |
|
"grad_norm": 4.192991256713867, |
|
"learning_rate": 9.999944328774796e-06, |
|
"loss": 1.413, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.011572582193012408, |
|
"grad_norm": 4.157929420471191, |
|
"learning_rate": 9.999939382570075e-06, |
|
"loss": 1.4509, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.011637233490180074, |
|
"grad_norm": 3.6802518367767334, |
|
"learning_rate": 9.999934225890294e-06, |
|
"loss": 1.3918, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.011701884787347742, |
|
"grad_norm": 5.291732311248779, |
|
"learning_rate": 9.999928858735668e-06, |
|
"loss": 1.4045, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.011766536084515408, |
|
"grad_norm": 4.054734230041504, |
|
"learning_rate": 9.999923281106427e-06, |
|
"loss": 1.3101, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.011831187381683076, |
|
"grad_norm": 4.11862325668335, |
|
"learning_rate": 9.999917493002803e-06, |
|
"loss": 1.4398, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.011895838678850742, |
|
"grad_norm": 4.4882307052612305, |
|
"learning_rate": 9.999911494425041e-06, |
|
"loss": 1.3585, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.01196048997601841, |
|
"grad_norm": 4.65657377243042, |
|
"learning_rate": 9.999905285373392e-06, |
|
"loss": 1.3673, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.012025141273186076, |
|
"grad_norm": 3.629178047180176, |
|
"learning_rate": 9.999898865848119e-06, |
|
"loss": 1.4998, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.012089792570353744, |
|
"grad_norm": 4.499670028686523, |
|
"learning_rate": 9.99989223584949e-06, |
|
"loss": 1.4083, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.012154443867521411, |
|
"grad_norm": 4.342660427093506, |
|
"learning_rate": 9.999885395377788e-06, |
|
"loss": 1.3362, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.012219095164689078, |
|
"grad_norm": 4.276759147644043, |
|
"learning_rate": 9.999878344433297e-06, |
|
"loss": 1.3912, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.012283746461856745, |
|
"grad_norm": 3.7648189067840576, |
|
"learning_rate": 9.999871083016316e-06, |
|
"loss": 1.3583, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.012348397759024412, |
|
"grad_norm": 4.836668491363525, |
|
"learning_rate": 9.999863611127149e-06, |
|
"loss": 1.3465, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.01241304905619208, |
|
"grad_norm": 4.41436767578125, |
|
"learning_rate": 9.999855928766113e-06, |
|
"loss": 1.4314, |
|
"step": 192 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 15467, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 16, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.232522107274281e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|