{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 0, "global_step": 2391, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0012547051442910915, "grad_norm": 0.482421875, "learning_rate": 9.995817649519031e-06, "loss": 1.8793, "step": 1 }, { "epoch": 0.002509410288582183, "grad_norm": 0.486328125, "learning_rate": 9.99163529903806e-06, "loss": 1.8797, "step": 2 }, { "epoch": 0.0037641154328732747, "grad_norm": 0.46875, "learning_rate": 9.98745294855709e-06, "loss": 1.8632, "step": 3 }, { "epoch": 0.005018820577164366, "grad_norm": 0.466796875, "learning_rate": 9.98327059807612e-06, "loss": 1.9315, "step": 4 }, { "epoch": 0.006273525721455458, "grad_norm": 0.5, "learning_rate": 9.979088247595148e-06, "loss": 1.917, "step": 5 }, { "epoch": 0.0075282308657465494, "grad_norm": 0.431640625, "learning_rate": 9.974905897114179e-06, "loss": 1.8117, "step": 6 }, { "epoch": 0.00878293601003764, "grad_norm": 0.419921875, "learning_rate": 9.97072354663321e-06, "loss": 1.8632, "step": 7 }, { "epoch": 0.010037641154328732, "grad_norm": 0.40234375, "learning_rate": 9.966541196152238e-06, "loss": 1.7053, "step": 8 }, { "epoch": 0.011292346298619825, "grad_norm": 0.390625, "learning_rate": 9.962358845671269e-06, "loss": 1.8161, "step": 9 }, { "epoch": 0.012547051442910916, "grad_norm": 0.380859375, "learning_rate": 9.958176495190298e-06, "loss": 1.8387, "step": 10 }, { "epoch": 0.013801756587202008, "grad_norm": 0.41015625, "learning_rate": 9.953994144709328e-06, "loss": 1.8818, "step": 11 }, { "epoch": 0.015056461731493099, "grad_norm": 0.333984375, "learning_rate": 9.949811794228357e-06, "loss": 1.6265, "step": 12 }, { "epoch": 0.01631116687578419, "grad_norm": 0.357421875, "learning_rate": 9.945629443747386e-06, "loss": 1.693, "step": 13 }, { "epoch": 0.01756587202007528, "grad_norm": 0.318359375, "learning_rate": 9.941447093266416e-06, "loss": 1.5874, "step": 14 }, { "epoch": 0.018820577164366373, "grad_norm": 0.373046875, "learning_rate": 9.937264742785447e-06, "loss": 1.8216, "step": 15 }, { "epoch": 0.020075282308657464, "grad_norm": 0.337890625, "learning_rate": 9.933082392304476e-06, "loss": 1.6679, "step": 16 }, { "epoch": 0.02132998745294856, "grad_norm": 0.3125, "learning_rate": 9.928900041823506e-06, "loss": 1.6552, "step": 17 }, { "epoch": 0.02258469259723965, "grad_norm": 0.337890625, "learning_rate": 9.924717691342535e-06, "loss": 1.6848, "step": 18 }, { "epoch": 0.02383939774153074, "grad_norm": 0.330078125, "learning_rate": 9.920535340861564e-06, "loss": 1.6751, "step": 19 }, { "epoch": 0.025094102885821833, "grad_norm": 0.31640625, "learning_rate": 9.916352990380594e-06, "loss": 1.6425, "step": 20 }, { "epoch": 0.026348808030112924, "grad_norm": 0.30078125, "learning_rate": 9.912170639899625e-06, "loss": 1.5976, "step": 21 }, { "epoch": 0.027603513174404015, "grad_norm": 0.283203125, "learning_rate": 9.907988289418654e-06, "loss": 1.5458, "step": 22 }, { "epoch": 0.028858218318695106, "grad_norm": 0.2734375, "learning_rate": 9.903805938937684e-06, "loss": 1.5979, "step": 23 }, { "epoch": 0.030112923462986198, "grad_norm": 0.39453125, "learning_rate": 9.899623588456713e-06, "loss": 1.4864, "step": 24 }, { "epoch": 0.03136762860727729, "grad_norm": 0.25390625, "learning_rate": 9.895441237975744e-06, "loss": 1.5406, "step": 25 }, { "epoch": 0.03262233375156838, "grad_norm": 0.2373046875, "learning_rate": 9.891258887494772e-06, "loss": 1.4942, "step": 26 }, { "epoch": 0.033877038895859475, "grad_norm": 0.2431640625, "learning_rate": 9.887076537013803e-06, "loss": 1.5209, "step": 27 }, { "epoch": 0.03513174404015056, "grad_norm": 0.349609375, "learning_rate": 9.882894186532832e-06, "loss": 1.5269, "step": 28 }, { "epoch": 0.03638644918444166, "grad_norm": 0.24609375, "learning_rate": 9.878711836051862e-06, "loss": 1.471, "step": 29 }, { "epoch": 0.037641154328732745, "grad_norm": 0.228515625, "learning_rate": 9.874529485570891e-06, "loss": 1.4801, "step": 30 }, { "epoch": 0.03889585947302384, "grad_norm": 0.236328125, "learning_rate": 9.870347135089922e-06, "loss": 1.5134, "step": 31 }, { "epoch": 0.04015056461731493, "grad_norm": 0.21875, "learning_rate": 9.866164784608952e-06, "loss": 1.4466, "step": 32 }, { "epoch": 0.04140526976160602, "grad_norm": 0.2255859375, "learning_rate": 9.861982434127981e-06, "loss": 1.4458, "step": 33 }, { "epoch": 0.04265997490589712, "grad_norm": 0.205078125, "learning_rate": 9.85780008364701e-06, "loss": 1.4475, "step": 34 }, { "epoch": 0.043914680050188205, "grad_norm": 0.22265625, "learning_rate": 9.85361773316604e-06, "loss": 1.4664, "step": 35 }, { "epoch": 0.0451693851944793, "grad_norm": 0.1982421875, "learning_rate": 9.849435382685069e-06, "loss": 1.4043, "step": 36 }, { "epoch": 0.04642409033877039, "grad_norm": 0.205078125, "learning_rate": 9.8452530322041e-06, "loss": 1.3909, "step": 37 }, { "epoch": 0.04767879548306148, "grad_norm": 0.2021484375, "learning_rate": 9.84107068172313e-06, "loss": 1.4434, "step": 38 }, { "epoch": 0.04893350062735257, "grad_norm": 0.2021484375, "learning_rate": 9.836888331242159e-06, "loss": 1.4309, "step": 39 }, { "epoch": 0.050188205771643665, "grad_norm": 0.2158203125, "learning_rate": 9.83270598076119e-06, "loss": 1.4785, "step": 40 }, { "epoch": 0.05144291091593475, "grad_norm": 0.212890625, "learning_rate": 9.828523630280218e-06, "loss": 1.4453, "step": 41 }, { "epoch": 0.05269761606022585, "grad_norm": 0.1865234375, "learning_rate": 9.824341279799247e-06, "loss": 1.2693, "step": 42 }, { "epoch": 0.053952321204516936, "grad_norm": 0.19921875, "learning_rate": 9.820158929318278e-06, "loss": 1.3958, "step": 43 }, { "epoch": 0.05520702634880803, "grad_norm": 0.1943359375, "learning_rate": 9.815976578837307e-06, "loss": 1.3629, "step": 44 }, { "epoch": 0.056461731493099125, "grad_norm": 0.1669921875, "learning_rate": 9.811794228356337e-06, "loss": 1.3598, "step": 45 }, { "epoch": 0.05771643663739021, "grad_norm": 0.2177734375, "learning_rate": 9.807611877875368e-06, "loss": 1.3242, "step": 46 }, { "epoch": 0.05897114178168131, "grad_norm": 0.1806640625, "learning_rate": 9.803429527394396e-06, "loss": 1.3912, "step": 47 }, { "epoch": 0.060225846925972396, "grad_norm": 0.162109375, "learning_rate": 9.799247176913425e-06, "loss": 1.3282, "step": 48 }, { "epoch": 0.06148055207026349, "grad_norm": 0.18359375, "learning_rate": 9.795064826432456e-06, "loss": 1.3295, "step": 49 }, { "epoch": 0.06273525721455459, "grad_norm": 0.16796875, "learning_rate": 9.790882475951485e-06, "loss": 1.3488, "step": 50 }, { "epoch": 0.06398996235884567, "grad_norm": 0.173828125, "learning_rate": 9.786700125470515e-06, "loss": 1.3561, "step": 51 }, { "epoch": 0.06524466750313676, "grad_norm": 0.173828125, "learning_rate": 9.782517774989546e-06, "loss": 1.3696, "step": 52 }, { "epoch": 0.06649937264742785, "grad_norm": 0.16796875, "learning_rate": 9.778335424508574e-06, "loss": 1.2524, "step": 53 }, { "epoch": 0.06775407779171895, "grad_norm": 0.1689453125, "learning_rate": 9.774153074027605e-06, "loss": 1.2812, "step": 54 }, { "epoch": 0.06900878293601004, "grad_norm": 0.193359375, "learning_rate": 9.769970723546634e-06, "loss": 1.3113, "step": 55 }, { "epoch": 0.07026348808030113, "grad_norm": 0.2109375, "learning_rate": 9.765788373065663e-06, "loss": 1.3429, "step": 56 }, { "epoch": 0.07151819322459223, "grad_norm": 0.16015625, "learning_rate": 9.761606022584693e-06, "loss": 1.3745, "step": 57 }, { "epoch": 0.07277289836888332, "grad_norm": 0.158203125, "learning_rate": 9.757423672103724e-06, "loss": 1.2872, "step": 58 }, { "epoch": 0.0740276035131744, "grad_norm": 0.201171875, "learning_rate": 9.753241321622753e-06, "loss": 1.3621, "step": 59 }, { "epoch": 0.07528230865746549, "grad_norm": 0.1591796875, "learning_rate": 9.749058971141783e-06, "loss": 1.2819, "step": 60 }, { "epoch": 0.07653701380175659, "grad_norm": 0.1904296875, "learning_rate": 9.744876620660812e-06, "loss": 1.2682, "step": 61 }, { "epoch": 0.07779171894604768, "grad_norm": 0.177734375, "learning_rate": 9.740694270179842e-06, "loss": 1.3453, "step": 62 }, { "epoch": 0.07904642409033877, "grad_norm": 0.1650390625, "learning_rate": 9.736511919698871e-06, "loss": 1.2721, "step": 63 }, { "epoch": 0.08030112923462986, "grad_norm": 0.15625, "learning_rate": 9.732329569217902e-06, "loss": 1.2533, "step": 64 }, { "epoch": 0.08155583437892096, "grad_norm": 0.1552734375, "learning_rate": 9.72814721873693e-06, "loss": 1.2264, "step": 65 }, { "epoch": 0.08281053952321205, "grad_norm": 0.1669921875, "learning_rate": 9.723964868255961e-06, "loss": 1.2851, "step": 66 }, { "epoch": 0.08406524466750313, "grad_norm": 0.1875, "learning_rate": 9.71978251777499e-06, "loss": 1.384, "step": 67 }, { "epoch": 0.08531994981179424, "grad_norm": 0.2060546875, "learning_rate": 9.71560016729402e-06, "loss": 1.2909, "step": 68 }, { "epoch": 0.08657465495608532, "grad_norm": 0.1904296875, "learning_rate": 9.711417816813051e-06, "loss": 1.2598, "step": 69 }, { "epoch": 0.08782936010037641, "grad_norm": 0.181640625, "learning_rate": 9.707235466332078e-06, "loss": 1.2528, "step": 70 }, { "epoch": 0.0890840652446675, "grad_norm": 0.1650390625, "learning_rate": 9.703053115851109e-06, "loss": 1.2491, "step": 71 }, { "epoch": 0.0903387703889586, "grad_norm": 0.1689453125, "learning_rate": 9.69887076537014e-06, "loss": 1.297, "step": 72 }, { "epoch": 0.09159347553324969, "grad_norm": 0.1533203125, "learning_rate": 9.694688414889168e-06, "loss": 1.2628, "step": 73 }, { "epoch": 0.09284818067754078, "grad_norm": 0.1513671875, "learning_rate": 9.690506064408199e-06, "loss": 1.2604, "step": 74 }, { "epoch": 0.09410288582183186, "grad_norm": 0.208984375, "learning_rate": 9.686323713927227e-06, "loss": 1.1958, "step": 75 }, { "epoch": 0.09535759096612297, "grad_norm": 0.15234375, "learning_rate": 9.682141363446258e-06, "loss": 1.2598, "step": 76 }, { "epoch": 0.09661229611041405, "grad_norm": 0.15625, "learning_rate": 9.677959012965288e-06, "loss": 1.2724, "step": 77 }, { "epoch": 0.09786700125470514, "grad_norm": 0.1787109375, "learning_rate": 9.673776662484317e-06, "loss": 1.2388, "step": 78 }, { "epoch": 0.09912170639899624, "grad_norm": 0.1513671875, "learning_rate": 9.669594312003346e-06, "loss": 1.2165, "step": 79 }, { "epoch": 0.10037641154328733, "grad_norm": 0.345703125, "learning_rate": 9.665411961522377e-06, "loss": 1.2197, "step": 80 }, { "epoch": 0.10163111668757842, "grad_norm": 0.1650390625, "learning_rate": 9.661229611041405e-06, "loss": 1.2683, "step": 81 }, { "epoch": 0.1028858218318695, "grad_norm": 0.1708984375, "learning_rate": 9.657047260560436e-06, "loss": 1.1695, "step": 82 }, { "epoch": 0.10414052697616061, "grad_norm": 0.671875, "learning_rate": 9.652864910079466e-06, "loss": 1.2201, "step": 83 }, { "epoch": 0.1053952321204517, "grad_norm": 0.154296875, "learning_rate": 9.648682559598495e-06, "loss": 1.2606, "step": 84 }, { "epoch": 0.10664993726474278, "grad_norm": 0.1630859375, "learning_rate": 9.644500209117524e-06, "loss": 1.1966, "step": 85 }, { "epoch": 0.10790464240903387, "grad_norm": 0.16796875, "learning_rate": 9.640317858636555e-06, "loss": 1.2279, "step": 86 }, { "epoch": 0.10915934755332497, "grad_norm": 0.1875, "learning_rate": 9.636135508155583e-06, "loss": 1.197, "step": 87 }, { "epoch": 0.11041405269761606, "grad_norm": 0.154296875, "learning_rate": 9.631953157674614e-06, "loss": 1.2272, "step": 88 }, { "epoch": 0.11166875784190715, "grad_norm": 0.30859375, "learning_rate": 9.627770807193644e-06, "loss": 1.2726, "step": 89 }, { "epoch": 0.11292346298619825, "grad_norm": 0.1669921875, "learning_rate": 9.623588456712673e-06, "loss": 1.214, "step": 90 }, { "epoch": 0.11417816813048934, "grad_norm": 0.162109375, "learning_rate": 9.619406106231704e-06, "loss": 1.2265, "step": 91 }, { "epoch": 0.11543287327478043, "grad_norm": 0.1494140625, "learning_rate": 9.615223755750733e-06, "loss": 1.1881, "step": 92 }, { "epoch": 0.11668757841907151, "grad_norm": 0.177734375, "learning_rate": 9.611041405269761e-06, "loss": 1.1777, "step": 93 }, { "epoch": 0.11794228356336262, "grad_norm": 0.1533203125, "learning_rate": 9.606859054788792e-06, "loss": 1.2021, "step": 94 }, { "epoch": 0.1191969887076537, "grad_norm": 0.15625, "learning_rate": 9.602676704307823e-06, "loss": 1.1822, "step": 95 }, { "epoch": 0.12045169385194479, "grad_norm": 0.1767578125, "learning_rate": 9.598494353826851e-06, "loss": 1.2182, "step": 96 }, { "epoch": 0.12170639899623588, "grad_norm": 0.15234375, "learning_rate": 9.594312003345882e-06, "loss": 1.1972, "step": 97 }, { "epoch": 0.12296110414052698, "grad_norm": 0.1552734375, "learning_rate": 9.59012965286491e-06, "loss": 1.1837, "step": 98 }, { "epoch": 0.12421580928481807, "grad_norm": 0.1904296875, "learning_rate": 9.585947302383941e-06, "loss": 1.2148, "step": 99 }, { "epoch": 0.12547051442910917, "grad_norm": 0.1630859375, "learning_rate": 9.58176495190297e-06, "loss": 1.1449, "step": 100 }, { "epoch": 0.12672521957340024, "grad_norm": 0.17578125, "learning_rate": 9.577582601421999e-06, "loss": 1.1683, "step": 101 }, { "epoch": 0.12797992471769135, "grad_norm": 0.17578125, "learning_rate": 9.57340025094103e-06, "loss": 1.213, "step": 102 }, { "epoch": 0.12923462986198245, "grad_norm": 0.154296875, "learning_rate": 9.56921790046006e-06, "loss": 1.1446, "step": 103 }, { "epoch": 0.13048933500627352, "grad_norm": 0.158203125, "learning_rate": 9.565035549979089e-06, "loss": 1.2061, "step": 104 }, { "epoch": 0.13174404015056462, "grad_norm": 0.1708984375, "learning_rate": 9.56085319949812e-06, "loss": 1.2273, "step": 105 }, { "epoch": 0.1329987452948557, "grad_norm": 0.1640625, "learning_rate": 9.556670849017148e-06, "loss": 1.1453, "step": 106 }, { "epoch": 0.1342534504391468, "grad_norm": 0.333984375, "learning_rate": 9.552488498536177e-06, "loss": 1.1272, "step": 107 }, { "epoch": 0.1355081555834379, "grad_norm": 0.1767578125, "learning_rate": 9.548306148055207e-06, "loss": 1.2182, "step": 108 }, { "epoch": 0.13676286072772897, "grad_norm": 0.1787109375, "learning_rate": 9.544123797574238e-06, "loss": 1.2351, "step": 109 }, { "epoch": 0.13801756587202008, "grad_norm": 0.171875, "learning_rate": 9.539941447093267e-06, "loss": 1.2538, "step": 110 }, { "epoch": 0.13927227101631118, "grad_norm": 0.1650390625, "learning_rate": 9.535759096612297e-06, "loss": 1.1764, "step": 111 }, { "epoch": 0.14052697616060225, "grad_norm": 0.251953125, "learning_rate": 9.531576746131326e-06, "loss": 1.1249, "step": 112 }, { "epoch": 0.14178168130489335, "grad_norm": 0.23828125, "learning_rate": 9.527394395650357e-06, "loss": 1.1586, "step": 113 }, { "epoch": 0.14303638644918445, "grad_norm": 0.1728515625, "learning_rate": 9.523212045169386e-06, "loss": 1.1776, "step": 114 }, { "epoch": 0.14429109159347553, "grad_norm": 0.166015625, "learning_rate": 9.519029694688416e-06, "loss": 1.1833, "step": 115 }, { "epoch": 0.14554579673776663, "grad_norm": 0.1611328125, "learning_rate": 9.514847344207445e-06, "loss": 1.1692, "step": 116 }, { "epoch": 0.1468005018820577, "grad_norm": 0.22265625, "learning_rate": 9.510664993726475e-06, "loss": 1.1319, "step": 117 }, { "epoch": 0.1480552070263488, "grad_norm": 0.162109375, "learning_rate": 9.506482643245504e-06, "loss": 1.2387, "step": 118 }, { "epoch": 0.1493099121706399, "grad_norm": 0.16015625, "learning_rate": 9.502300292764535e-06, "loss": 1.2227, "step": 119 }, { "epoch": 0.15056461731493098, "grad_norm": 0.1611328125, "learning_rate": 9.498117942283565e-06, "loss": 1.1967, "step": 120 }, { "epoch": 0.15181932245922208, "grad_norm": 0.1748046875, "learning_rate": 9.493935591802594e-06, "loss": 1.1522, "step": 121 }, { "epoch": 0.15307402760351319, "grad_norm": 0.1943359375, "learning_rate": 9.489753241321623e-06, "loss": 1.1818, "step": 122 }, { "epoch": 0.15432873274780426, "grad_norm": 0.1689453125, "learning_rate": 9.485570890840653e-06, "loss": 1.2108, "step": 123 }, { "epoch": 0.15558343789209536, "grad_norm": 0.1640625, "learning_rate": 9.481388540359682e-06, "loss": 1.1483, "step": 124 }, { "epoch": 0.15683814303638646, "grad_norm": 0.1650390625, "learning_rate": 9.477206189878713e-06, "loss": 1.1516, "step": 125 }, { "epoch": 0.15809284818067754, "grad_norm": 0.16015625, "learning_rate": 9.473023839397743e-06, "loss": 1.1278, "step": 126 }, { "epoch": 0.15934755332496864, "grad_norm": 0.1572265625, "learning_rate": 9.468841488916772e-06, "loss": 1.1631, "step": 127 }, { "epoch": 0.1606022584692597, "grad_norm": 0.171875, "learning_rate": 9.464659138435803e-06, "loss": 1.169, "step": 128 }, { "epoch": 0.1618569636135508, "grad_norm": 0.21875, "learning_rate": 9.460476787954832e-06, "loss": 1.2488, "step": 129 }, { "epoch": 0.16311166875784192, "grad_norm": 0.2109375, "learning_rate": 9.45629443747386e-06, "loss": 1.1211, "step": 130 }, { "epoch": 0.164366373902133, "grad_norm": 0.1689453125, "learning_rate": 9.452112086992891e-06, "loss": 1.1642, "step": 131 }, { "epoch": 0.1656210790464241, "grad_norm": 0.1611328125, "learning_rate": 9.44792973651192e-06, "loss": 1.1381, "step": 132 }, { "epoch": 0.1668757841907152, "grad_norm": 0.171875, "learning_rate": 9.44374738603095e-06, "loss": 1.1505, "step": 133 }, { "epoch": 0.16813048933500627, "grad_norm": 0.1611328125, "learning_rate": 9.43956503554998e-06, "loss": 1.1378, "step": 134 }, { "epoch": 0.16938519447929737, "grad_norm": 0.166015625, "learning_rate": 9.43538268506901e-06, "loss": 1.14, "step": 135 }, { "epoch": 0.17063989962358847, "grad_norm": 0.1689453125, "learning_rate": 9.431200334588038e-06, "loss": 1.1292, "step": 136 }, { "epoch": 0.17189460476787954, "grad_norm": 0.169921875, "learning_rate": 9.427017984107069e-06, "loss": 1.141, "step": 137 }, { "epoch": 0.17314930991217065, "grad_norm": 0.1708984375, "learning_rate": 9.422835633626098e-06, "loss": 1.1932, "step": 138 }, { "epoch": 0.17440401505646172, "grad_norm": 0.1708984375, "learning_rate": 9.418653283145128e-06, "loss": 1.1448, "step": 139 }, { "epoch": 0.17565872020075282, "grad_norm": 0.1650390625, "learning_rate": 9.414470932664159e-06, "loss": 1.1395, "step": 140 }, { "epoch": 0.17691342534504392, "grad_norm": 0.162109375, "learning_rate": 9.410288582183188e-06, "loss": 1.1307, "step": 141 }, { "epoch": 0.178168130489335, "grad_norm": 0.16796875, "learning_rate": 9.406106231702218e-06, "loss": 1.1613, "step": 142 }, { "epoch": 0.1794228356336261, "grad_norm": 0.169921875, "learning_rate": 9.401923881221247e-06, "loss": 1.175, "step": 143 }, { "epoch": 0.1806775407779172, "grad_norm": 0.1650390625, "learning_rate": 9.397741530740276e-06, "loss": 1.1123, "step": 144 }, { "epoch": 0.18193224592220827, "grad_norm": 0.208984375, "learning_rate": 9.393559180259306e-06, "loss": 1.1361, "step": 145 }, { "epoch": 0.18318695106649938, "grad_norm": 0.1767578125, "learning_rate": 9.389376829778337e-06, "loss": 1.168, "step": 146 }, { "epoch": 0.18444165621079048, "grad_norm": 0.1689453125, "learning_rate": 9.385194479297366e-06, "loss": 1.1046, "step": 147 }, { "epoch": 0.18569636135508155, "grad_norm": 0.1767578125, "learning_rate": 9.381012128816396e-06, "loss": 1.1286, "step": 148 }, { "epoch": 0.18695106649937265, "grad_norm": 0.1640625, "learning_rate": 9.376829778335425e-06, "loss": 1.1003, "step": 149 }, { "epoch": 0.18820577164366373, "grad_norm": 0.181640625, "learning_rate": 9.372647427854456e-06, "loss": 1.1881, "step": 150 }, { "epoch": 0.18946047678795483, "grad_norm": 0.1748046875, "learning_rate": 9.368465077373484e-06, "loss": 1.158, "step": 151 }, { "epoch": 0.19071518193224593, "grad_norm": 0.2080078125, "learning_rate": 9.364282726892515e-06, "loss": 1.1083, "step": 152 }, { "epoch": 0.191969887076537, "grad_norm": 0.16015625, "learning_rate": 9.360100376411544e-06, "loss": 1.1151, "step": 153 }, { "epoch": 0.1932245922208281, "grad_norm": 0.19140625, "learning_rate": 9.355918025930574e-06, "loss": 1.1106, "step": 154 }, { "epoch": 0.1944792973651192, "grad_norm": 0.1728515625, "learning_rate": 9.351735675449603e-06, "loss": 1.1226, "step": 155 }, { "epoch": 0.19573400250941028, "grad_norm": 0.185546875, "learning_rate": 9.347553324968634e-06, "loss": 1.1263, "step": 156 }, { "epoch": 0.19698870765370138, "grad_norm": 0.1689453125, "learning_rate": 9.343370974487664e-06, "loss": 1.1033, "step": 157 }, { "epoch": 0.19824341279799249, "grad_norm": 0.1845703125, "learning_rate": 9.339188624006691e-06, "loss": 1.094, "step": 158 }, { "epoch": 0.19949811794228356, "grad_norm": 0.1845703125, "learning_rate": 9.335006273525722e-06, "loss": 1.0974, "step": 159 }, { "epoch": 0.20075282308657466, "grad_norm": 0.189453125, "learning_rate": 9.330823923044752e-06, "loss": 1.1983, "step": 160 }, { "epoch": 0.20200752823086573, "grad_norm": 0.18359375, "learning_rate": 9.326641572563781e-06, "loss": 1.1021, "step": 161 }, { "epoch": 0.20326223337515684, "grad_norm": 0.1748046875, "learning_rate": 9.322459222082812e-06, "loss": 1.1062, "step": 162 }, { "epoch": 0.20451693851944794, "grad_norm": 0.1875, "learning_rate": 9.31827687160184e-06, "loss": 1.1019, "step": 163 }, { "epoch": 0.205771643663739, "grad_norm": 0.2099609375, "learning_rate": 9.314094521120871e-06, "loss": 1.2144, "step": 164 }, { "epoch": 0.20702634880803011, "grad_norm": 0.1728515625, "learning_rate": 9.309912170639902e-06, "loss": 1.1522, "step": 165 }, { "epoch": 0.20828105395232122, "grad_norm": 0.1806640625, "learning_rate": 9.30572982015893e-06, "loss": 1.1447, "step": 166 }, { "epoch": 0.2095357590966123, "grad_norm": 0.1806640625, "learning_rate": 9.30154746967796e-06, "loss": 1.177, "step": 167 }, { "epoch": 0.2107904642409034, "grad_norm": 0.189453125, "learning_rate": 9.29736511919699e-06, "loss": 1.1043, "step": 168 }, { "epoch": 0.2120451693851945, "grad_norm": 0.1787109375, "learning_rate": 9.293182768716019e-06, "loss": 1.1697, "step": 169 }, { "epoch": 0.21329987452948557, "grad_norm": 0.1865234375, "learning_rate": 9.289000418235049e-06, "loss": 1.1634, "step": 170 }, { "epoch": 0.21455457967377667, "grad_norm": 0.166015625, "learning_rate": 9.28481806775408e-06, "loss": 1.1265, "step": 171 }, { "epoch": 0.21580928481806774, "grad_norm": 0.2353515625, "learning_rate": 9.280635717273108e-06, "loss": 1.107, "step": 172 }, { "epoch": 0.21706398996235884, "grad_norm": 0.18359375, "learning_rate": 9.276453366792137e-06, "loss": 1.0814, "step": 173 }, { "epoch": 0.21831869510664995, "grad_norm": 0.171875, "learning_rate": 9.272271016311168e-06, "loss": 1.1062, "step": 174 }, { "epoch": 0.21957340025094102, "grad_norm": 0.28515625, "learning_rate": 9.268088665830197e-06, "loss": 1.1546, "step": 175 }, { "epoch": 0.22082810539523212, "grad_norm": 0.189453125, "learning_rate": 9.263906315349227e-06, "loss": 1.0584, "step": 176 }, { "epoch": 0.22208281053952322, "grad_norm": 0.19140625, "learning_rate": 9.259723964868258e-06, "loss": 1.1411, "step": 177 }, { "epoch": 0.2233375156838143, "grad_norm": 0.220703125, "learning_rate": 9.255541614387286e-06, "loss": 1.1248, "step": 178 }, { "epoch": 0.2245922208281054, "grad_norm": 0.185546875, "learning_rate": 9.251359263906317e-06, "loss": 1.1178, "step": 179 }, { "epoch": 0.2258469259723965, "grad_norm": 0.2021484375, "learning_rate": 9.247176913425346e-06, "loss": 1.1662, "step": 180 }, { "epoch": 0.22710163111668757, "grad_norm": 0.1982421875, "learning_rate": 9.242994562944375e-06, "loss": 1.1226, "step": 181 }, { "epoch": 0.22835633626097868, "grad_norm": 0.1953125, "learning_rate": 9.238812212463405e-06, "loss": 1.0666, "step": 182 }, { "epoch": 0.22961104140526975, "grad_norm": 0.1787109375, "learning_rate": 9.234629861982434e-06, "loss": 1.0842, "step": 183 }, { "epoch": 0.23086574654956085, "grad_norm": 0.1767578125, "learning_rate": 9.230447511501465e-06, "loss": 1.1085, "step": 184 }, { "epoch": 0.23212045169385195, "grad_norm": 0.171875, "learning_rate": 9.226265161020495e-06, "loss": 1.1006, "step": 185 }, { "epoch": 0.23337515683814303, "grad_norm": 0.1826171875, "learning_rate": 9.222082810539524e-06, "loss": 1.061, "step": 186 }, { "epoch": 0.23462986198243413, "grad_norm": 0.2001953125, "learning_rate": 9.217900460058554e-06, "loss": 1.1097, "step": 187 }, { "epoch": 0.23588456712672523, "grad_norm": 0.193359375, "learning_rate": 9.213718109577583e-06, "loss": 1.0882, "step": 188 }, { "epoch": 0.2371392722710163, "grad_norm": 0.216796875, "learning_rate": 9.209535759096612e-06, "loss": 1.0467, "step": 189 }, { "epoch": 0.2383939774153074, "grad_norm": 0.18359375, "learning_rate": 9.205353408615643e-06, "loss": 1.1215, "step": 190 }, { "epoch": 0.2396486825595985, "grad_norm": 0.2177734375, "learning_rate": 9.201171058134673e-06, "loss": 1.0859, "step": 191 }, { "epoch": 0.24090338770388958, "grad_norm": 0.1884765625, "learning_rate": 9.196988707653702e-06, "loss": 1.1244, "step": 192 }, { "epoch": 0.24215809284818068, "grad_norm": 0.1845703125, "learning_rate": 9.192806357172732e-06, "loss": 1.0775, "step": 193 }, { "epoch": 0.24341279799247176, "grad_norm": 0.220703125, "learning_rate": 9.188624006691761e-06, "loss": 1.0903, "step": 194 }, { "epoch": 0.24466750313676286, "grad_norm": 0.1875, "learning_rate": 9.18444165621079e-06, "loss": 1.0903, "step": 195 }, { "epoch": 0.24592220828105396, "grad_norm": 0.18359375, "learning_rate": 9.18025930572982e-06, "loss": 1.0504, "step": 196 }, { "epoch": 0.24717691342534504, "grad_norm": 0.1826171875, "learning_rate": 9.176076955248851e-06, "loss": 1.0891, "step": 197 }, { "epoch": 0.24843161856963614, "grad_norm": 0.1796875, "learning_rate": 9.17189460476788e-06, "loss": 1.0932, "step": 198 }, { "epoch": 0.24968632371392724, "grad_norm": 0.1826171875, "learning_rate": 9.16771225428691e-06, "loss": 1.0785, "step": 199 }, { "epoch": 0.25094102885821834, "grad_norm": 0.1865234375, "learning_rate": 9.16352990380594e-06, "loss": 1.1318, "step": 200 }, { "epoch": 0.2521957340025094, "grad_norm": 0.181640625, "learning_rate": 9.15934755332497e-06, "loss": 1.0685, "step": 201 }, { "epoch": 0.2534504391468005, "grad_norm": 0.31640625, "learning_rate": 9.155165202843999e-06, "loss": 1.1644, "step": 202 }, { "epoch": 0.2547051442910916, "grad_norm": 0.1923828125, "learning_rate": 9.15098285236303e-06, "loss": 1.0946, "step": 203 }, { "epoch": 0.2559598494353827, "grad_norm": 0.189453125, "learning_rate": 9.146800501882058e-06, "loss": 1.0887, "step": 204 }, { "epoch": 0.2572145545796738, "grad_norm": 0.185546875, "learning_rate": 9.142618151401089e-06, "loss": 1.0714, "step": 205 }, { "epoch": 0.2584692597239649, "grad_norm": 0.1953125, "learning_rate": 9.138435800920117e-06, "loss": 1.0805, "step": 206 }, { "epoch": 0.25972396486825594, "grad_norm": 0.25, "learning_rate": 9.134253450439148e-06, "loss": 1.0687, "step": 207 }, { "epoch": 0.26097867001254704, "grad_norm": 0.2001953125, "learning_rate": 9.130071099958178e-06, "loss": 1.1066, "step": 208 }, { "epoch": 0.26223337515683814, "grad_norm": 0.18359375, "learning_rate": 9.125888749477207e-06, "loss": 1.0733, "step": 209 }, { "epoch": 0.26348808030112925, "grad_norm": 0.271484375, "learning_rate": 9.121706398996236e-06, "loss": 1.0688, "step": 210 }, { "epoch": 0.26474278544542035, "grad_norm": 0.185546875, "learning_rate": 9.117524048515267e-06, "loss": 1.1061, "step": 211 }, { "epoch": 0.2659974905897114, "grad_norm": 0.2021484375, "learning_rate": 9.113341698034295e-06, "loss": 1.084, "step": 212 }, { "epoch": 0.2672521957340025, "grad_norm": 0.203125, "learning_rate": 9.109159347553326e-06, "loss": 1.1238, "step": 213 }, { "epoch": 0.2685069008782936, "grad_norm": 0.1904296875, "learning_rate": 9.104976997072355e-06, "loss": 1.09, "step": 214 }, { "epoch": 0.2697616060225847, "grad_norm": 0.2099609375, "learning_rate": 9.100794646591385e-06, "loss": 1.1027, "step": 215 }, { "epoch": 0.2710163111668758, "grad_norm": 0.1845703125, "learning_rate": 9.096612296110416e-06, "loss": 1.096, "step": 216 }, { "epoch": 0.2722710163111669, "grad_norm": 0.201171875, "learning_rate": 9.092429945629445e-06, "loss": 1.0193, "step": 217 }, { "epoch": 0.27352572145545795, "grad_norm": 0.185546875, "learning_rate": 9.088247595148473e-06, "loss": 1.0826, "step": 218 }, { "epoch": 0.27478042659974905, "grad_norm": 0.1865234375, "learning_rate": 9.084065244667504e-06, "loss": 1.056, "step": 219 }, { "epoch": 0.27603513174404015, "grad_norm": 0.1962890625, "learning_rate": 9.079882894186533e-06, "loss": 1.1055, "step": 220 }, { "epoch": 0.27728983688833125, "grad_norm": 0.2158203125, "learning_rate": 9.075700543705563e-06, "loss": 1.089, "step": 221 }, { "epoch": 0.27854454203262236, "grad_norm": 0.185546875, "learning_rate": 9.071518193224594e-06, "loss": 1.0795, "step": 222 }, { "epoch": 0.2797992471769134, "grad_norm": 0.208984375, "learning_rate": 9.067335842743623e-06, "loss": 1.1617, "step": 223 }, { "epoch": 0.2810539523212045, "grad_norm": 0.26953125, "learning_rate": 9.063153492262652e-06, "loss": 1.0552, "step": 224 }, { "epoch": 0.2823086574654956, "grad_norm": 0.1884765625, "learning_rate": 9.058971141781682e-06, "loss": 1.0942, "step": 225 }, { "epoch": 0.2835633626097867, "grad_norm": 0.1943359375, "learning_rate": 9.054788791300711e-06, "loss": 1.0745, "step": 226 }, { "epoch": 0.2848180677540778, "grad_norm": 0.20703125, "learning_rate": 9.050606440819741e-06, "loss": 1.031, "step": 227 }, { "epoch": 0.2860727728983689, "grad_norm": 0.19140625, "learning_rate": 9.046424090338772e-06, "loss": 1.0866, "step": 228 }, { "epoch": 0.28732747804265996, "grad_norm": 0.185546875, "learning_rate": 9.0422417398578e-06, "loss": 1.0929, "step": 229 }, { "epoch": 0.28858218318695106, "grad_norm": 0.203125, "learning_rate": 9.038059389376831e-06, "loss": 1.0584, "step": 230 }, { "epoch": 0.28983688833124216, "grad_norm": 0.2021484375, "learning_rate": 9.03387703889586e-06, "loss": 1.064, "step": 231 }, { "epoch": 0.29109159347553326, "grad_norm": 0.2080078125, "learning_rate": 9.029694688414889e-06, "loss": 1.0727, "step": 232 }, { "epoch": 0.29234629861982436, "grad_norm": 0.205078125, "learning_rate": 9.02551233793392e-06, "loss": 1.135, "step": 233 }, { "epoch": 0.2936010037641154, "grad_norm": 0.244140625, "learning_rate": 9.02132998745295e-06, "loss": 1.0518, "step": 234 }, { "epoch": 0.2948557089084065, "grad_norm": 0.205078125, "learning_rate": 9.017147636971979e-06, "loss": 1.0559, "step": 235 }, { "epoch": 0.2961104140526976, "grad_norm": 0.1865234375, "learning_rate": 9.01296528649101e-06, "loss": 1.044, "step": 236 }, { "epoch": 0.2973651191969887, "grad_norm": 0.193359375, "learning_rate": 9.008782936010038e-06, "loss": 1.0754, "step": 237 }, { "epoch": 0.2986198243412798, "grad_norm": 0.1923828125, "learning_rate": 9.004600585529069e-06, "loss": 1.0651, "step": 238 }, { "epoch": 0.2998745294855709, "grad_norm": 0.2041015625, "learning_rate": 9.000418235048098e-06, "loss": 1.187, "step": 239 }, { "epoch": 0.30112923462986196, "grad_norm": 0.1806640625, "learning_rate": 8.996235884567126e-06, "loss": 1.0554, "step": 240 }, { "epoch": 0.30238393977415307, "grad_norm": 0.1904296875, "learning_rate": 8.992053534086157e-06, "loss": 1.102, "step": 241 }, { "epoch": 0.30363864491844417, "grad_norm": 0.189453125, "learning_rate": 8.987871183605187e-06, "loss": 1.0383, "step": 242 }, { "epoch": 0.30489335006273527, "grad_norm": 0.1904296875, "learning_rate": 8.983688833124216e-06, "loss": 1.0935, "step": 243 }, { "epoch": 0.30614805520702637, "grad_norm": 0.208984375, "learning_rate": 8.979506482643247e-06, "loss": 1.059, "step": 244 }, { "epoch": 0.3074027603513174, "grad_norm": 0.203125, "learning_rate": 8.975324132162276e-06, "loss": 1.0791, "step": 245 }, { "epoch": 0.3086574654956085, "grad_norm": 0.2021484375, "learning_rate": 8.971141781681304e-06, "loss": 1.0281, "step": 246 }, { "epoch": 0.3099121706398996, "grad_norm": 0.1875, "learning_rate": 8.966959431200335e-06, "loss": 1.0391, "step": 247 }, { "epoch": 0.3111668757841907, "grad_norm": 0.1845703125, "learning_rate": 8.962777080719365e-06, "loss": 1.0516, "step": 248 }, { "epoch": 0.3124215809284818, "grad_norm": 0.208984375, "learning_rate": 8.958594730238394e-06, "loss": 1.0821, "step": 249 }, { "epoch": 0.3136762860727729, "grad_norm": 0.265625, "learning_rate": 8.954412379757425e-06, "loss": 1.0265, "step": 250 }, { "epoch": 0.31493099121706397, "grad_norm": 0.1904296875, "learning_rate": 8.950230029276454e-06, "loss": 1.0784, "step": 251 }, { "epoch": 0.3161856963613551, "grad_norm": 0.2138671875, "learning_rate": 8.946047678795484e-06, "loss": 1.0757, "step": 252 }, { "epoch": 0.3174404015056462, "grad_norm": 0.1865234375, "learning_rate": 8.941865328314515e-06, "loss": 1.0652, "step": 253 }, { "epoch": 0.3186951066499373, "grad_norm": 0.193359375, "learning_rate": 8.937682977833543e-06, "loss": 1.0766, "step": 254 }, { "epoch": 0.3199498117942284, "grad_norm": 0.2021484375, "learning_rate": 8.933500627352572e-06, "loss": 1.0424, "step": 255 }, { "epoch": 0.3212045169385194, "grad_norm": 0.1982421875, "learning_rate": 8.929318276871603e-06, "loss": 1.0786, "step": 256 }, { "epoch": 0.3224592220828105, "grad_norm": 0.2041015625, "learning_rate": 8.925135926390632e-06, "loss": 1.079, "step": 257 }, { "epoch": 0.3237139272271016, "grad_norm": 0.1953125, "learning_rate": 8.920953575909662e-06, "loss": 1.0653, "step": 258 }, { "epoch": 0.32496863237139273, "grad_norm": 0.2060546875, "learning_rate": 8.916771225428693e-06, "loss": 1.0922, "step": 259 }, { "epoch": 0.32622333751568383, "grad_norm": 0.2099609375, "learning_rate": 8.912588874947722e-06, "loss": 1.0685, "step": 260 }, { "epoch": 0.32747804265997493, "grad_norm": 0.19921875, "learning_rate": 8.90840652446675e-06, "loss": 1.0257, "step": 261 }, { "epoch": 0.328732747804266, "grad_norm": 0.203125, "learning_rate": 8.904224173985781e-06, "loss": 1.1041, "step": 262 }, { "epoch": 0.3299874529485571, "grad_norm": 0.1875, "learning_rate": 8.90004182350481e-06, "loss": 1.0283, "step": 263 }, { "epoch": 0.3312421580928482, "grad_norm": 0.1962890625, "learning_rate": 8.89585947302384e-06, "loss": 1.0779, "step": 264 }, { "epoch": 0.3324968632371393, "grad_norm": 0.203125, "learning_rate": 8.89167712254287e-06, "loss": 1.0597, "step": 265 }, { "epoch": 0.3337515683814304, "grad_norm": 0.1845703125, "learning_rate": 8.8874947720619e-06, "loss": 1.0699, "step": 266 }, { "epoch": 0.33500627352572143, "grad_norm": 0.201171875, "learning_rate": 8.88331242158093e-06, "loss": 1.0437, "step": 267 }, { "epoch": 0.33626097867001253, "grad_norm": 0.1962890625, "learning_rate": 8.879130071099959e-06, "loss": 1.0681, "step": 268 }, { "epoch": 0.33751568381430364, "grad_norm": 0.1962890625, "learning_rate": 8.874947720618988e-06, "loss": 1.0635, "step": 269 }, { "epoch": 0.33877038895859474, "grad_norm": 0.19921875, "learning_rate": 8.870765370138018e-06, "loss": 1.0072, "step": 270 }, { "epoch": 0.34002509410288584, "grad_norm": 0.1953125, "learning_rate": 8.866583019657047e-06, "loss": 1.0594, "step": 271 }, { "epoch": 0.34127979924717694, "grad_norm": 0.193359375, "learning_rate": 8.862400669176078e-06, "loss": 1.0456, "step": 272 }, { "epoch": 0.342534504391468, "grad_norm": 0.2060546875, "learning_rate": 8.858218318695108e-06, "loss": 1.1048, "step": 273 }, { "epoch": 0.3437892095357591, "grad_norm": 0.1865234375, "learning_rate": 8.854035968214137e-06, "loss": 1.0407, "step": 274 }, { "epoch": 0.3450439146800502, "grad_norm": 0.19921875, "learning_rate": 8.849853617733168e-06, "loss": 1.0505, "step": 275 }, { "epoch": 0.3462986198243413, "grad_norm": 0.2001953125, "learning_rate": 8.845671267252196e-06, "loss": 1.0145, "step": 276 }, { "epoch": 0.3475533249686324, "grad_norm": 0.2109375, "learning_rate": 8.841488916771225e-06, "loss": 1.058, "step": 277 }, { "epoch": 0.34880803011292344, "grad_norm": 0.19140625, "learning_rate": 8.837306566290256e-06, "loss": 1.0548, "step": 278 }, { "epoch": 0.35006273525721454, "grad_norm": 0.1904296875, "learning_rate": 8.833124215809286e-06, "loss": 1.0582, "step": 279 }, { "epoch": 0.35131744040150564, "grad_norm": 0.1982421875, "learning_rate": 8.828941865328315e-06, "loss": 1.082, "step": 280 }, { "epoch": 0.35257214554579674, "grad_norm": 0.2373046875, "learning_rate": 8.824759514847346e-06, "loss": 1.0462, "step": 281 }, { "epoch": 0.35382685069008785, "grad_norm": 0.1904296875, "learning_rate": 8.820577164366374e-06, "loss": 1.0427, "step": 282 }, { "epoch": 0.35508155583437895, "grad_norm": 0.19140625, "learning_rate": 8.816394813885403e-06, "loss": 1.0209, "step": 283 }, { "epoch": 0.35633626097867, "grad_norm": 0.201171875, "learning_rate": 8.812212463404434e-06, "loss": 1.0903, "step": 284 }, { "epoch": 0.3575909661229611, "grad_norm": 0.1982421875, "learning_rate": 8.808030112923464e-06, "loss": 1.0617, "step": 285 }, { "epoch": 0.3588456712672522, "grad_norm": 0.1943359375, "learning_rate": 8.803847762442493e-06, "loss": 0.9971, "step": 286 }, { "epoch": 0.3601003764115433, "grad_norm": 0.1943359375, "learning_rate": 8.799665411961524e-06, "loss": 1.0575, "step": 287 }, { "epoch": 0.3613550815558344, "grad_norm": 0.24609375, "learning_rate": 8.795483061480552e-06, "loss": 1.082, "step": 288 }, { "epoch": 0.36260978670012545, "grad_norm": 0.234375, "learning_rate": 8.791300710999583e-06, "loss": 1.0523, "step": 289 }, { "epoch": 0.36386449184441655, "grad_norm": 0.1904296875, "learning_rate": 8.787118360518612e-06, "loss": 1.0174, "step": 290 }, { "epoch": 0.36511919698870765, "grad_norm": 0.2578125, "learning_rate": 8.782936010037642e-06, "loss": 1.0369, "step": 291 }, { "epoch": 0.36637390213299875, "grad_norm": 0.26171875, "learning_rate": 8.778753659556671e-06, "loss": 1.0544, "step": 292 }, { "epoch": 0.36762860727728985, "grad_norm": 0.2001953125, "learning_rate": 8.774571309075702e-06, "loss": 1.0456, "step": 293 }, { "epoch": 0.36888331242158096, "grad_norm": 0.19921875, "learning_rate": 8.77038895859473e-06, "loss": 1.1067, "step": 294 }, { "epoch": 0.370138017565872, "grad_norm": 0.205078125, "learning_rate": 8.766206608113761e-06, "loss": 1.0571, "step": 295 }, { "epoch": 0.3713927227101631, "grad_norm": 0.201171875, "learning_rate": 8.762024257632792e-06, "loss": 1.0328, "step": 296 }, { "epoch": 0.3726474278544542, "grad_norm": 0.2353515625, "learning_rate": 8.75784190715182e-06, "loss": 1.0467, "step": 297 }, { "epoch": 0.3739021329987453, "grad_norm": 0.2041015625, "learning_rate": 8.75365955667085e-06, "loss": 1.0196, "step": 298 }, { "epoch": 0.3751568381430364, "grad_norm": 0.203125, "learning_rate": 8.74947720618988e-06, "loss": 1.0675, "step": 299 }, { "epoch": 0.37641154328732745, "grad_norm": 0.2080078125, "learning_rate": 8.745294855708909e-06, "loss": 1.0341, "step": 300 }, { "epoch": 0.37766624843161856, "grad_norm": 0.203125, "learning_rate": 8.741112505227939e-06, "loss": 1.0226, "step": 301 }, { "epoch": 0.37892095357590966, "grad_norm": 0.2021484375, "learning_rate": 8.736930154746968e-06, "loss": 1.0879, "step": 302 }, { "epoch": 0.38017565872020076, "grad_norm": 0.2158203125, "learning_rate": 8.732747804265998e-06, "loss": 1.0703, "step": 303 }, { "epoch": 0.38143036386449186, "grad_norm": 0.216796875, "learning_rate": 8.728565453785029e-06, "loss": 1.0548, "step": 304 }, { "epoch": 0.38268506900878296, "grad_norm": 0.2041015625, "learning_rate": 8.724383103304058e-06, "loss": 1.027, "step": 305 }, { "epoch": 0.383939774153074, "grad_norm": 0.279296875, "learning_rate": 8.720200752823087e-06, "loss": 1.0274, "step": 306 }, { "epoch": 0.3851944792973651, "grad_norm": 0.2119140625, "learning_rate": 8.716018402342117e-06, "loss": 1.0309, "step": 307 }, { "epoch": 0.3864491844416562, "grad_norm": 0.19921875, "learning_rate": 8.711836051861146e-06, "loss": 1.0398, "step": 308 }, { "epoch": 0.3877038895859473, "grad_norm": 0.208984375, "learning_rate": 8.707653701380176e-06, "loss": 1.0056, "step": 309 }, { "epoch": 0.3889585947302384, "grad_norm": 0.224609375, "learning_rate": 8.703471350899207e-06, "loss": 1.0319, "step": 310 }, { "epoch": 0.39021329987452946, "grad_norm": 0.21875, "learning_rate": 8.699289000418236e-06, "loss": 1.0308, "step": 311 }, { "epoch": 0.39146800501882056, "grad_norm": 0.2353515625, "learning_rate": 8.695106649937265e-06, "loss": 1.0706, "step": 312 }, { "epoch": 0.39272271016311167, "grad_norm": 0.361328125, "learning_rate": 8.690924299456295e-06, "loss": 1.0563, "step": 313 }, { "epoch": 0.39397741530740277, "grad_norm": 0.2099609375, "learning_rate": 8.686741948975324e-06, "loss": 1.0513, "step": 314 }, { "epoch": 0.39523212045169387, "grad_norm": 0.19921875, "learning_rate": 8.682559598494355e-06, "loss": 1.0475, "step": 315 }, { "epoch": 0.39648682559598497, "grad_norm": 0.2001953125, "learning_rate": 8.678377248013385e-06, "loss": 1.039, "step": 316 }, { "epoch": 0.397741530740276, "grad_norm": 0.208984375, "learning_rate": 8.674194897532414e-06, "loss": 1.0672, "step": 317 }, { "epoch": 0.3989962358845671, "grad_norm": 0.2080078125, "learning_rate": 8.670012547051444e-06, "loss": 1.0172, "step": 318 }, { "epoch": 0.4002509410288582, "grad_norm": 0.2060546875, "learning_rate": 8.665830196570473e-06, "loss": 1.0869, "step": 319 }, { "epoch": 0.4015056461731493, "grad_norm": 0.1962890625, "learning_rate": 8.661647846089502e-06, "loss": 1.0653, "step": 320 }, { "epoch": 0.4027603513174404, "grad_norm": 0.19921875, "learning_rate": 8.657465495608533e-06, "loss": 1.0832, "step": 321 }, { "epoch": 0.40401505646173147, "grad_norm": 0.21875, "learning_rate": 8.653283145127563e-06, "loss": 1.0474, "step": 322 }, { "epoch": 0.40526976160602257, "grad_norm": 0.23828125, "learning_rate": 8.649100794646592e-06, "loss": 1.0287, "step": 323 }, { "epoch": 0.4065244667503137, "grad_norm": 0.2255859375, "learning_rate": 8.644918444165622e-06, "loss": 1.0567, "step": 324 }, { "epoch": 0.4077791718946048, "grad_norm": 0.205078125, "learning_rate": 8.640736093684651e-06, "loss": 0.9904, "step": 325 }, { "epoch": 0.4090338770388959, "grad_norm": 0.2001953125, "learning_rate": 8.636553743203682e-06, "loss": 1.0025, "step": 326 }, { "epoch": 0.410288582183187, "grad_norm": 0.2138671875, "learning_rate": 8.63237139272271e-06, "loss": 1.0497, "step": 327 }, { "epoch": 0.411543287327478, "grad_norm": 0.2060546875, "learning_rate": 8.62818904224174e-06, "loss": 1.095, "step": 328 }, { "epoch": 0.4127979924717691, "grad_norm": 0.232421875, "learning_rate": 8.62400669176077e-06, "loss": 1.0771, "step": 329 }, { "epoch": 0.41405269761606023, "grad_norm": 0.2158203125, "learning_rate": 8.6198243412798e-06, "loss": 1.0596, "step": 330 }, { "epoch": 0.41530740276035133, "grad_norm": 0.240234375, "learning_rate": 8.61564199079883e-06, "loss": 1.045, "step": 331 }, { "epoch": 0.41656210790464243, "grad_norm": 0.216796875, "learning_rate": 8.61145964031786e-06, "loss": 1.0481, "step": 332 }, { "epoch": 0.4178168130489335, "grad_norm": 0.205078125, "learning_rate": 8.607277289836889e-06, "loss": 1.0339, "step": 333 }, { "epoch": 0.4190715181932246, "grad_norm": 0.208984375, "learning_rate": 8.603094939355918e-06, "loss": 1.0101, "step": 334 }, { "epoch": 0.4203262233375157, "grad_norm": 0.208984375, "learning_rate": 8.598912588874948e-06, "loss": 1.0582, "step": 335 }, { "epoch": 0.4215809284818068, "grad_norm": 0.2236328125, "learning_rate": 8.594730238393979e-06, "loss": 1.0242, "step": 336 }, { "epoch": 0.4228356336260979, "grad_norm": 0.20703125, "learning_rate": 8.590547887913007e-06, "loss": 1.056, "step": 337 }, { "epoch": 0.424090338770389, "grad_norm": 0.220703125, "learning_rate": 8.586365537432038e-06, "loss": 1.0404, "step": 338 }, { "epoch": 0.42534504391468003, "grad_norm": 0.25390625, "learning_rate": 8.582183186951067e-06, "loss": 1.0304, "step": 339 }, { "epoch": 0.42659974905897113, "grad_norm": 0.2412109375, "learning_rate": 8.578000836470097e-06, "loss": 1.0384, "step": 340 }, { "epoch": 0.42785445420326224, "grad_norm": 0.2119140625, "learning_rate": 8.573818485989128e-06, "loss": 1.04, "step": 341 }, { "epoch": 0.42910915934755334, "grad_norm": 0.208984375, "learning_rate": 8.569636135508157e-06, "loss": 1.0517, "step": 342 }, { "epoch": 0.43036386449184444, "grad_norm": 0.2041015625, "learning_rate": 8.565453785027185e-06, "loss": 1.0191, "step": 343 }, { "epoch": 0.4316185696361355, "grad_norm": 0.2099609375, "learning_rate": 8.561271434546216e-06, "loss": 1.0595, "step": 344 }, { "epoch": 0.4328732747804266, "grad_norm": 0.2060546875, "learning_rate": 8.557089084065245e-06, "loss": 1.0822, "step": 345 }, { "epoch": 0.4341279799247177, "grad_norm": 0.2119140625, "learning_rate": 8.552906733584275e-06, "loss": 1.0439, "step": 346 }, { "epoch": 0.4353826850690088, "grad_norm": 0.2158203125, "learning_rate": 8.548724383103306e-06, "loss": 1.0513, "step": 347 }, { "epoch": 0.4366373902132999, "grad_norm": 0.2314453125, "learning_rate": 8.544542032622335e-06, "loss": 0.9979, "step": 348 }, { "epoch": 0.437892095357591, "grad_norm": 0.251953125, "learning_rate": 8.540359682141364e-06, "loss": 1.0278, "step": 349 }, { "epoch": 0.43914680050188204, "grad_norm": 0.2197265625, "learning_rate": 8.536177331660394e-06, "loss": 1.0181, "step": 350 }, { "epoch": 0.44040150564617314, "grad_norm": 0.228515625, "learning_rate": 8.531994981179423e-06, "loss": 1.0416, "step": 351 }, { "epoch": 0.44165621079046424, "grad_norm": 0.2255859375, "learning_rate": 8.527812630698453e-06, "loss": 1.0266, "step": 352 }, { "epoch": 0.44291091593475534, "grad_norm": 0.2138671875, "learning_rate": 8.523630280217482e-06, "loss": 1.0429, "step": 353 }, { "epoch": 0.44416562107904645, "grad_norm": 0.2158203125, "learning_rate": 8.519447929736513e-06, "loss": 0.971, "step": 354 }, { "epoch": 0.4454203262233375, "grad_norm": 0.259765625, "learning_rate": 8.515265579255543e-06, "loss": 1.0833, "step": 355 }, { "epoch": 0.4466750313676286, "grad_norm": 0.283203125, "learning_rate": 8.511083228774572e-06, "loss": 1.0253, "step": 356 }, { "epoch": 0.4479297365119197, "grad_norm": 0.21875, "learning_rate": 8.506900878293601e-06, "loss": 1.0332, "step": 357 }, { "epoch": 0.4491844416562108, "grad_norm": 0.2578125, "learning_rate": 8.502718527812631e-06, "loss": 1.0672, "step": 358 }, { "epoch": 0.4504391468005019, "grad_norm": 0.212890625, "learning_rate": 8.49853617733166e-06, "loss": 1.0325, "step": 359 }, { "epoch": 0.451693851944793, "grad_norm": 0.2138671875, "learning_rate": 8.49435382685069e-06, "loss": 0.9798, "step": 360 }, { "epoch": 0.45294855708908405, "grad_norm": 0.294921875, "learning_rate": 8.490171476369721e-06, "loss": 1.0098, "step": 361 }, { "epoch": 0.45420326223337515, "grad_norm": 0.2314453125, "learning_rate": 8.48598912588875e-06, "loss": 1.0338, "step": 362 }, { "epoch": 0.45545796737766625, "grad_norm": 0.2451171875, "learning_rate": 8.48180677540778e-06, "loss": 1.0384, "step": 363 }, { "epoch": 0.45671267252195735, "grad_norm": 0.21875, "learning_rate": 8.47762442492681e-06, "loss": 1.0436, "step": 364 }, { "epoch": 0.45796737766624845, "grad_norm": 0.2333984375, "learning_rate": 8.473442074445838e-06, "loss": 1.0358, "step": 365 }, { "epoch": 0.4592220828105395, "grad_norm": 0.2373046875, "learning_rate": 8.469259723964869e-06, "loss": 1.1077, "step": 366 }, { "epoch": 0.4604767879548306, "grad_norm": 0.20703125, "learning_rate": 8.4650773734839e-06, "loss": 1.0196, "step": 367 }, { "epoch": 0.4617314930991217, "grad_norm": 0.2099609375, "learning_rate": 8.460895023002928e-06, "loss": 0.9875, "step": 368 }, { "epoch": 0.4629861982434128, "grad_norm": 0.2119140625, "learning_rate": 8.456712672521959e-06, "loss": 1.0033, "step": 369 }, { "epoch": 0.4642409033877039, "grad_norm": 0.21484375, "learning_rate": 8.452530322040988e-06, "loss": 1.0246, "step": 370 }, { "epoch": 0.465495608531995, "grad_norm": 0.2119140625, "learning_rate": 8.448347971560016e-06, "loss": 1.0066, "step": 371 }, { "epoch": 0.46675031367628605, "grad_norm": 0.208984375, "learning_rate": 8.444165621079047e-06, "loss": 1.0539, "step": 372 }, { "epoch": 0.46800501882057716, "grad_norm": 0.234375, "learning_rate": 8.439983270598077e-06, "loss": 1.056, "step": 373 }, { "epoch": 0.46925972396486826, "grad_norm": 0.2158203125, "learning_rate": 8.435800920117106e-06, "loss": 1.0904, "step": 374 }, { "epoch": 0.47051442910915936, "grad_norm": 0.216796875, "learning_rate": 8.431618569636137e-06, "loss": 0.9608, "step": 375 }, { "epoch": 0.47176913425345046, "grad_norm": 0.2197265625, "learning_rate": 8.427436219155166e-06, "loss": 1.0258, "step": 376 }, { "epoch": 0.4730238393977415, "grad_norm": 0.216796875, "learning_rate": 8.423253868674196e-06, "loss": 1.0339, "step": 377 }, { "epoch": 0.4742785445420326, "grad_norm": 0.267578125, "learning_rate": 8.419071518193225e-06, "loss": 1.0044, "step": 378 }, { "epoch": 0.4755332496863237, "grad_norm": 0.2060546875, "learning_rate": 8.414889167712254e-06, "loss": 1.002, "step": 379 }, { "epoch": 0.4767879548306148, "grad_norm": 0.2275390625, "learning_rate": 8.410706817231284e-06, "loss": 1.0398, "step": 380 }, { "epoch": 0.4780426599749059, "grad_norm": 0.212890625, "learning_rate": 8.406524466750315e-06, "loss": 1.0015, "step": 381 }, { "epoch": 0.479297365119197, "grad_norm": 0.6015625, "learning_rate": 8.402342116269344e-06, "loss": 1.0573, "step": 382 }, { "epoch": 0.48055207026348806, "grad_norm": 0.216796875, "learning_rate": 8.398159765788374e-06, "loss": 0.9718, "step": 383 }, { "epoch": 0.48180677540777916, "grad_norm": 0.2177734375, "learning_rate": 8.393977415307403e-06, "loss": 1.006, "step": 384 }, { "epoch": 0.48306148055207027, "grad_norm": 0.2470703125, "learning_rate": 8.389795064826434e-06, "loss": 1.0714, "step": 385 }, { "epoch": 0.48431618569636137, "grad_norm": 0.2109375, "learning_rate": 8.385612714345462e-06, "loss": 0.9695, "step": 386 }, { "epoch": 0.48557089084065247, "grad_norm": 0.23046875, "learning_rate": 8.381430363864493e-06, "loss": 1.0223, "step": 387 }, { "epoch": 0.4868255959849435, "grad_norm": 0.212890625, "learning_rate": 8.377248013383522e-06, "loss": 0.9794, "step": 388 }, { "epoch": 0.4880803011292346, "grad_norm": 0.2109375, "learning_rate": 8.373065662902552e-06, "loss": 0.9704, "step": 389 }, { "epoch": 0.4893350062735257, "grad_norm": 0.216796875, "learning_rate": 8.368883312421581e-06, "loss": 1.0395, "step": 390 }, { "epoch": 0.4905897114178168, "grad_norm": 0.21875, "learning_rate": 8.364700961940612e-06, "loss": 0.9873, "step": 391 }, { "epoch": 0.4918444165621079, "grad_norm": 0.2216796875, "learning_rate": 8.360518611459642e-06, "loss": 0.9835, "step": 392 }, { "epoch": 0.493099121706399, "grad_norm": 0.232421875, "learning_rate": 8.356336260978671e-06, "loss": 1.0159, "step": 393 }, { "epoch": 0.49435382685069007, "grad_norm": 0.21484375, "learning_rate": 8.3521539104977e-06, "loss": 1.0066, "step": 394 }, { "epoch": 0.49560853199498117, "grad_norm": 0.2177734375, "learning_rate": 8.34797156001673e-06, "loss": 0.9811, "step": 395 }, { "epoch": 0.4968632371392723, "grad_norm": 0.42578125, "learning_rate": 8.343789209535759e-06, "loss": 0.9903, "step": 396 }, { "epoch": 0.4981179422835634, "grad_norm": 0.2333984375, "learning_rate": 8.33960685905479e-06, "loss": 1.0178, "step": 397 }, { "epoch": 0.4993726474278545, "grad_norm": 0.33203125, "learning_rate": 8.33542450857382e-06, "loss": 1.0212, "step": 398 }, { "epoch": 0.5006273525721455, "grad_norm": 0.224609375, "learning_rate": 8.331242158092849e-06, "loss": 1.0137, "step": 399 }, { "epoch": 0.5018820577164367, "grad_norm": 0.2333984375, "learning_rate": 8.327059807611878e-06, "loss": 1.0656, "step": 400 }, { "epoch": 0.5031367628607277, "grad_norm": 0.2333984375, "learning_rate": 8.322877457130908e-06, "loss": 0.9897, "step": 401 }, { "epoch": 0.5043914680050188, "grad_norm": 0.310546875, "learning_rate": 8.318695106649937e-06, "loss": 1.0711, "step": 402 }, { "epoch": 0.5056461731493099, "grad_norm": 0.220703125, "learning_rate": 8.314512756168968e-06, "loss": 1.0218, "step": 403 }, { "epoch": 0.506900878293601, "grad_norm": 0.2373046875, "learning_rate": 8.310330405687998e-06, "loss": 1.0379, "step": 404 }, { "epoch": 0.5081555834378921, "grad_norm": 0.2412109375, "learning_rate": 8.306148055207027e-06, "loss": 1.0484, "step": 405 }, { "epoch": 0.5094102885821832, "grad_norm": 0.220703125, "learning_rate": 8.301965704726058e-06, "loss": 0.9861, "step": 406 }, { "epoch": 0.5106649937264742, "grad_norm": 0.2158203125, "learning_rate": 8.297783354245086e-06, "loss": 0.9804, "step": 407 }, { "epoch": 0.5119196988707654, "grad_norm": 0.212890625, "learning_rate": 8.293601003764115e-06, "loss": 1.0532, "step": 408 }, { "epoch": 0.5131744040150564, "grad_norm": 0.234375, "learning_rate": 8.289418653283146e-06, "loss": 0.9992, "step": 409 }, { "epoch": 0.5144291091593476, "grad_norm": 0.228515625, "learning_rate": 8.285236302802175e-06, "loss": 1.0196, "step": 410 }, { "epoch": 0.5156838143036386, "grad_norm": 0.23828125, "learning_rate": 8.281053952321205e-06, "loss": 1.0008, "step": 411 }, { "epoch": 0.5169385194479298, "grad_norm": 0.21484375, "learning_rate": 8.276871601840236e-06, "loss": 1.0156, "step": 412 }, { "epoch": 0.5181932245922208, "grad_norm": 0.234375, "learning_rate": 8.272689251359264e-06, "loss": 1.0251, "step": 413 }, { "epoch": 0.5194479297365119, "grad_norm": 0.21875, "learning_rate": 8.268506900878295e-06, "loss": 0.9977, "step": 414 }, { "epoch": 0.520702634880803, "grad_norm": 0.212890625, "learning_rate": 8.264324550397324e-06, "loss": 1.0198, "step": 415 }, { "epoch": 0.5219573400250941, "grad_norm": 0.26953125, "learning_rate": 8.260142199916353e-06, "loss": 1.016, "step": 416 }, { "epoch": 0.5232120451693852, "grad_norm": 0.306640625, "learning_rate": 8.255959849435383e-06, "loss": 1.0187, "step": 417 }, { "epoch": 0.5244667503136763, "grad_norm": 0.2255859375, "learning_rate": 8.251777498954414e-06, "loss": 0.9804, "step": 418 }, { "epoch": 0.5257214554579673, "grad_norm": 0.2021484375, "learning_rate": 8.247595148473443e-06, "loss": 1.0048, "step": 419 }, { "epoch": 0.5269761606022585, "grad_norm": 0.2294921875, "learning_rate": 8.243412797992473e-06, "loss": 1.0547, "step": 420 }, { "epoch": 0.5282308657465495, "grad_norm": 0.2099609375, "learning_rate": 8.239230447511502e-06, "loss": 0.9922, "step": 421 }, { "epoch": 0.5294855708908407, "grad_norm": 0.2197265625, "learning_rate": 8.23504809703053e-06, "loss": 1.0304, "step": 422 }, { "epoch": 0.5307402760351317, "grad_norm": 0.2158203125, "learning_rate": 8.230865746549561e-06, "loss": 0.9792, "step": 423 }, { "epoch": 0.5319949811794228, "grad_norm": 0.220703125, "learning_rate": 8.226683396068592e-06, "loss": 1.0434, "step": 424 }, { "epoch": 0.533249686323714, "grad_norm": 0.234375, "learning_rate": 8.22250104558762e-06, "loss": 1.0205, "step": 425 }, { "epoch": 0.534504391468005, "grad_norm": 0.2294921875, "learning_rate": 8.218318695106651e-06, "loss": 1.0201, "step": 426 }, { "epoch": 0.5357590966122961, "grad_norm": 0.224609375, "learning_rate": 8.21413634462568e-06, "loss": 1.0631, "step": 427 }, { "epoch": 0.5370138017565872, "grad_norm": 0.25390625, "learning_rate": 8.20995399414471e-06, "loss": 1.0009, "step": 428 }, { "epoch": 0.5382685069008782, "grad_norm": 0.234375, "learning_rate": 8.205771643663741e-06, "loss": 1.0001, "step": 429 }, { "epoch": 0.5395232120451694, "grad_norm": 0.21875, "learning_rate": 8.20158929318277e-06, "loss": 1.0417, "step": 430 }, { "epoch": 0.5407779171894604, "grad_norm": 0.2294921875, "learning_rate": 8.197406942701799e-06, "loss": 1.0221, "step": 431 }, { "epoch": 0.5420326223337516, "grad_norm": 0.220703125, "learning_rate": 8.193224592220829e-06, "loss": 1.0829, "step": 432 }, { "epoch": 0.5432873274780426, "grad_norm": 0.2275390625, "learning_rate": 8.189042241739858e-06, "loss": 1.0319, "step": 433 }, { "epoch": 0.5445420326223338, "grad_norm": 0.2060546875, "learning_rate": 8.184859891258888e-06, "loss": 1.0119, "step": 434 }, { "epoch": 0.5457967377666249, "grad_norm": 0.2109375, "learning_rate": 8.180677540777919e-06, "loss": 0.996, "step": 435 }, { "epoch": 0.5470514429109159, "grad_norm": 0.2265625, "learning_rate": 8.176495190296948e-06, "loss": 1.0364, "step": 436 }, { "epoch": 0.548306148055207, "grad_norm": 0.2314453125, "learning_rate": 8.172312839815977e-06, "loss": 1.0173, "step": 437 }, { "epoch": 0.5495608531994981, "grad_norm": 0.220703125, "learning_rate": 8.168130489335007e-06, "loss": 1.0503, "step": 438 }, { "epoch": 0.5508155583437893, "grad_norm": 0.220703125, "learning_rate": 8.163948138854036e-06, "loss": 0.9819, "step": 439 }, { "epoch": 0.5520702634880803, "grad_norm": 0.2490234375, "learning_rate": 8.159765788373067e-06, "loss": 1.0331, "step": 440 }, { "epoch": 0.5533249686323714, "grad_norm": 0.228515625, "learning_rate": 8.155583437892095e-06, "loss": 1.0458, "step": 441 }, { "epoch": 0.5545796737766625, "grad_norm": 0.271484375, "learning_rate": 8.151401087411126e-06, "loss": 0.9841, "step": 442 }, { "epoch": 0.5558343789209536, "grad_norm": 0.296875, "learning_rate": 8.147218736930156e-06, "loss": 1.0274, "step": 443 }, { "epoch": 0.5570890840652447, "grad_norm": 0.2490234375, "learning_rate": 8.143036386449185e-06, "loss": 0.9981, "step": 444 }, { "epoch": 0.5583437892095358, "grad_norm": 0.228515625, "learning_rate": 8.138854035968214e-06, "loss": 1.0116, "step": 445 }, { "epoch": 0.5595984943538268, "grad_norm": 0.234375, "learning_rate": 8.134671685487245e-06, "loss": 1.0133, "step": 446 }, { "epoch": 0.560853199498118, "grad_norm": 0.2314453125, "learning_rate": 8.130489335006273e-06, "loss": 1.0012, "step": 447 }, { "epoch": 0.562107904642409, "grad_norm": 0.2353515625, "learning_rate": 8.126306984525304e-06, "loss": 1.0523, "step": 448 }, { "epoch": 0.5633626097867002, "grad_norm": 0.2333984375, "learning_rate": 8.122124634044334e-06, "loss": 0.9821, "step": 449 }, { "epoch": 0.5646173149309912, "grad_norm": 0.220703125, "learning_rate": 8.117942283563363e-06, "loss": 1.0288, "step": 450 }, { "epoch": 0.5658720200752823, "grad_norm": 0.216796875, "learning_rate": 8.113759933082394e-06, "loss": 1.0134, "step": 451 }, { "epoch": 0.5671267252195734, "grad_norm": 0.2431640625, "learning_rate": 8.109577582601423e-06, "loss": 1.0146, "step": 452 }, { "epoch": 0.5683814303638645, "grad_norm": 0.220703125, "learning_rate": 8.105395232120451e-06, "loss": 1.012, "step": 453 }, { "epoch": 0.5696361355081556, "grad_norm": 0.23828125, "learning_rate": 8.101212881639482e-06, "loss": 0.9946, "step": 454 }, { "epoch": 0.5708908406524467, "grad_norm": 0.2275390625, "learning_rate": 8.097030531158513e-06, "loss": 1.0417, "step": 455 }, { "epoch": 0.5721455457967378, "grad_norm": 0.294921875, "learning_rate": 8.092848180677541e-06, "loss": 1.0155, "step": 456 }, { "epoch": 0.5734002509410289, "grad_norm": 0.248046875, "learning_rate": 8.088665830196572e-06, "loss": 1.006, "step": 457 }, { "epoch": 0.5746549560853199, "grad_norm": 0.236328125, "learning_rate": 8.0844834797156e-06, "loss": 0.9814, "step": 458 }, { "epoch": 0.5759096612296111, "grad_norm": 0.26171875, "learning_rate": 8.08030112923463e-06, "loss": 1.0201, "step": 459 }, { "epoch": 0.5771643663739021, "grad_norm": 0.236328125, "learning_rate": 8.07611877875366e-06, "loss": 0.9739, "step": 460 }, { "epoch": 0.5784190715181933, "grad_norm": 0.259765625, "learning_rate": 8.07193642827269e-06, "loss": 0.9555, "step": 461 }, { "epoch": 0.5796737766624843, "grad_norm": 0.232421875, "learning_rate": 8.06775407779172e-06, "loss": 1.0299, "step": 462 }, { "epoch": 0.5809284818067754, "grad_norm": 0.2333984375, "learning_rate": 8.06357172731075e-06, "loss": 1.0246, "step": 463 }, { "epoch": 0.5821831869510665, "grad_norm": 0.2138671875, "learning_rate": 8.059389376829779e-06, "loss": 0.9699, "step": 464 }, { "epoch": 0.5834378920953576, "grad_norm": 0.310546875, "learning_rate": 8.05520702634881e-06, "loss": 1.0191, "step": 465 }, { "epoch": 0.5846925972396487, "grad_norm": 0.224609375, "learning_rate": 8.051024675867838e-06, "loss": 1.0183, "step": 466 }, { "epoch": 0.5859473023839398, "grad_norm": 0.2265625, "learning_rate": 8.046842325386867e-06, "loss": 1.0084, "step": 467 }, { "epoch": 0.5872020075282308, "grad_norm": 0.251953125, "learning_rate": 8.042659974905897e-06, "loss": 0.9922, "step": 468 }, { "epoch": 0.588456712672522, "grad_norm": 0.22265625, "learning_rate": 8.038477624424928e-06, "loss": 0.9887, "step": 469 }, { "epoch": 0.589711417816813, "grad_norm": 0.259765625, "learning_rate": 8.034295273943957e-06, "loss": 0.9409, "step": 470 }, { "epoch": 0.5909661229611042, "grad_norm": 0.275390625, "learning_rate": 8.030112923462987e-06, "loss": 0.9773, "step": 471 }, { "epoch": 0.5922208281053952, "grad_norm": 0.240234375, "learning_rate": 8.025930572982016e-06, "loss": 0.9887, "step": 472 }, { "epoch": 0.5934755332496863, "grad_norm": 0.25390625, "learning_rate": 8.021748222501047e-06, "loss": 0.9971, "step": 473 }, { "epoch": 0.5947302383939774, "grad_norm": 0.2109375, "learning_rate": 8.017565872020076e-06, "loss": 1.0071, "step": 474 }, { "epoch": 0.5959849435382685, "grad_norm": 0.220703125, "learning_rate": 8.013383521539106e-06, "loss": 0.9746, "step": 475 }, { "epoch": 0.5972396486825596, "grad_norm": 0.25, "learning_rate": 8.009201171058135e-06, "loss": 1.0038, "step": 476 }, { "epoch": 0.5984943538268507, "grad_norm": 0.44921875, "learning_rate": 8.005018820577165e-06, "loss": 0.9755, "step": 477 }, { "epoch": 0.5997490589711418, "grad_norm": 0.2197265625, "learning_rate": 8.000836470096194e-06, "loss": 0.9784, "step": 478 }, { "epoch": 0.6010037641154329, "grad_norm": 0.28515625, "learning_rate": 7.996654119615225e-06, "loss": 1.0621, "step": 479 }, { "epoch": 0.6022584692597239, "grad_norm": 0.2451171875, "learning_rate": 7.992471769134255e-06, "loss": 0.9513, "step": 480 }, { "epoch": 0.6035131744040151, "grad_norm": 0.294921875, "learning_rate": 7.988289418653284e-06, "loss": 0.986, "step": 481 }, { "epoch": 0.6047678795483061, "grad_norm": 0.263671875, "learning_rate": 7.984107068172313e-06, "loss": 1.0217, "step": 482 }, { "epoch": 0.6060225846925973, "grad_norm": 0.27734375, "learning_rate": 7.979924717691343e-06, "loss": 1.0034, "step": 483 }, { "epoch": 0.6072772898368883, "grad_norm": 0.291015625, "learning_rate": 7.975742367210372e-06, "loss": 1.0114, "step": 484 }, { "epoch": 0.6085319949811794, "grad_norm": 0.2421875, "learning_rate": 7.971560016729403e-06, "loss": 1.0674, "step": 485 }, { "epoch": 0.6097867001254705, "grad_norm": 0.232421875, "learning_rate": 7.967377666248433e-06, "loss": 0.9976, "step": 486 }, { "epoch": 0.6110414052697616, "grad_norm": 0.2451171875, "learning_rate": 7.963195315767462e-06, "loss": 1.021, "step": 487 }, { "epoch": 0.6122961104140527, "grad_norm": 0.2890625, "learning_rate": 7.959012965286491e-06, "loss": 0.9923, "step": 488 }, { "epoch": 0.6135508155583438, "grad_norm": 0.228515625, "learning_rate": 7.954830614805521e-06, "loss": 1.0084, "step": 489 }, { "epoch": 0.6148055207026348, "grad_norm": 0.30859375, "learning_rate": 7.95064826432455e-06, "loss": 0.9933, "step": 490 }, { "epoch": 0.616060225846926, "grad_norm": 0.232421875, "learning_rate": 7.94646591384358e-06, "loss": 1.0392, "step": 491 }, { "epoch": 0.617314930991217, "grad_norm": 0.412109375, "learning_rate": 7.942283563362611e-06, "loss": 0.9691, "step": 492 }, { "epoch": 0.6185696361355082, "grad_norm": 0.2392578125, "learning_rate": 7.93810121288164e-06, "loss": 1.0176, "step": 493 }, { "epoch": 0.6198243412797992, "grad_norm": 0.23828125, "learning_rate": 7.93391886240067e-06, "loss": 0.9932, "step": 494 }, { "epoch": 0.6210790464240903, "grad_norm": 0.25, "learning_rate": 7.9297365119197e-06, "loss": 0.959, "step": 495 }, { "epoch": 0.6223337515683814, "grad_norm": 0.2451171875, "learning_rate": 7.925554161438728e-06, "loss": 0.9965, "step": 496 }, { "epoch": 0.6235884567126725, "grad_norm": 0.255859375, "learning_rate": 7.921371810957759e-06, "loss": 1.0306, "step": 497 }, { "epoch": 0.6248431618569636, "grad_norm": 0.2373046875, "learning_rate": 7.917189460476788e-06, "loss": 0.9842, "step": 498 }, { "epoch": 0.6260978670012547, "grad_norm": 0.234375, "learning_rate": 7.913007109995818e-06, "loss": 0.9957, "step": 499 }, { "epoch": 0.6273525721455459, "grad_norm": 0.21484375, "learning_rate": 7.908824759514849e-06, "loss": 0.9837, "step": 500 }, { "epoch": 0.6286072772898369, "grad_norm": 0.23046875, "learning_rate": 7.904642409033878e-06, "loss": 0.9882, "step": 501 }, { "epoch": 0.6298619824341279, "grad_norm": 0.251953125, "learning_rate": 7.900460058552908e-06, "loss": 0.9403, "step": 502 }, { "epoch": 0.6311166875784191, "grad_norm": 0.244140625, "learning_rate": 7.896277708071937e-06, "loss": 0.9933, "step": 503 }, { "epoch": 0.6323713927227101, "grad_norm": 0.25, "learning_rate": 7.892095357590966e-06, "loss": 1.0015, "step": 504 }, { "epoch": 0.6336260978670013, "grad_norm": 0.24609375, "learning_rate": 7.887913007109996e-06, "loss": 1.0183, "step": 505 }, { "epoch": 0.6348808030112923, "grad_norm": 0.240234375, "learning_rate": 7.883730656629027e-06, "loss": 0.9862, "step": 506 }, { "epoch": 0.6361355081555834, "grad_norm": 0.224609375, "learning_rate": 7.879548306148056e-06, "loss": 1.0231, "step": 507 }, { "epoch": 0.6373902132998746, "grad_norm": 0.2431640625, "learning_rate": 7.875365955667086e-06, "loss": 0.9403, "step": 508 }, { "epoch": 0.6386449184441656, "grad_norm": 0.2353515625, "learning_rate": 7.871183605186115e-06, "loss": 0.9958, "step": 509 }, { "epoch": 0.6398996235884568, "grad_norm": 0.23828125, "learning_rate": 7.867001254705144e-06, "loss": 0.9763, "step": 510 }, { "epoch": 0.6411543287327478, "grad_norm": 0.333984375, "learning_rate": 7.862818904224174e-06, "loss": 0.9952, "step": 511 }, { "epoch": 0.6424090338770388, "grad_norm": 0.263671875, "learning_rate": 7.858636553743205e-06, "loss": 0.9733, "step": 512 }, { "epoch": 0.64366373902133, "grad_norm": 0.251953125, "learning_rate": 7.854454203262234e-06, "loss": 1.0052, "step": 513 }, { "epoch": 0.644918444165621, "grad_norm": 0.2275390625, "learning_rate": 7.850271852781264e-06, "loss": 0.9415, "step": 514 }, { "epoch": 0.6461731493099122, "grad_norm": 0.2333984375, "learning_rate": 7.846089502300293e-06, "loss": 1.0217, "step": 515 }, { "epoch": 0.6474278544542033, "grad_norm": 0.2421875, "learning_rate": 7.841907151819324e-06, "loss": 0.9833, "step": 516 }, { "epoch": 0.6486825595984943, "grad_norm": 0.2412109375, "learning_rate": 7.837724801338354e-06, "loss": 1.0216, "step": 517 }, { "epoch": 0.6499372647427855, "grad_norm": 0.2265625, "learning_rate": 7.833542450857381e-06, "loss": 1.003, "step": 518 }, { "epoch": 0.6511919698870765, "grad_norm": 0.22265625, "learning_rate": 7.829360100376412e-06, "loss": 1.0005, "step": 519 }, { "epoch": 0.6524466750313677, "grad_norm": 0.22265625, "learning_rate": 7.825177749895442e-06, "loss": 1.0068, "step": 520 }, { "epoch": 0.6537013801756587, "grad_norm": 0.220703125, "learning_rate": 7.820995399414471e-06, "loss": 0.9909, "step": 521 }, { "epoch": 0.6549560853199499, "grad_norm": 0.2451171875, "learning_rate": 7.816813048933502e-06, "loss": 0.9888, "step": 522 }, { "epoch": 0.6562107904642409, "grad_norm": 0.248046875, "learning_rate": 7.81263069845253e-06, "loss": 0.9709, "step": 523 }, { "epoch": 0.657465495608532, "grad_norm": 0.2275390625, "learning_rate": 7.808448347971561e-06, "loss": 0.9443, "step": 524 }, { "epoch": 0.6587202007528231, "grad_norm": 0.2412109375, "learning_rate": 7.80426599749059e-06, "loss": 1.0158, "step": 525 }, { "epoch": 0.6599749058971142, "grad_norm": 0.2197265625, "learning_rate": 7.80008364700962e-06, "loss": 0.989, "step": 526 }, { "epoch": 0.6612296110414053, "grad_norm": 0.24609375, "learning_rate": 7.79590129652865e-06, "loss": 0.9851, "step": 527 }, { "epoch": 0.6624843161856964, "grad_norm": 0.23828125, "learning_rate": 7.79171894604768e-06, "loss": 1.0343, "step": 528 }, { "epoch": 0.6637390213299874, "grad_norm": 0.28515625, "learning_rate": 7.787536595566709e-06, "loss": 1.0069, "step": 529 }, { "epoch": 0.6649937264742786, "grad_norm": 0.24609375, "learning_rate": 7.783354245085739e-06, "loss": 1.0138, "step": 530 }, { "epoch": 0.6662484316185696, "grad_norm": 0.2431640625, "learning_rate": 7.77917189460477e-06, "loss": 1.0085, "step": 531 }, { "epoch": 0.6675031367628608, "grad_norm": 0.25390625, "learning_rate": 7.774989544123798e-06, "loss": 0.9774, "step": 532 }, { "epoch": 0.6687578419071518, "grad_norm": 0.22265625, "learning_rate": 7.770807193642827e-06, "loss": 0.9935, "step": 533 }, { "epoch": 0.6700125470514429, "grad_norm": 0.2578125, "learning_rate": 7.766624843161858e-06, "loss": 0.9728, "step": 534 }, { "epoch": 0.671267252195734, "grad_norm": 0.2236328125, "learning_rate": 7.762442492680887e-06, "loss": 0.997, "step": 535 }, { "epoch": 0.6725219573400251, "grad_norm": 0.23046875, "learning_rate": 7.758260142199917e-06, "loss": 1.0129, "step": 536 }, { "epoch": 0.6737766624843162, "grad_norm": 0.224609375, "learning_rate": 7.754077791718948e-06, "loss": 0.9845, "step": 537 }, { "epoch": 0.6750313676286073, "grad_norm": 0.232421875, "learning_rate": 7.749895441237976e-06, "loss": 1.0239, "step": 538 }, { "epoch": 0.6762860727728983, "grad_norm": 0.2431640625, "learning_rate": 7.745713090757007e-06, "loss": 1.0143, "step": 539 }, { "epoch": 0.6775407779171895, "grad_norm": 0.2421875, "learning_rate": 7.741530740276036e-06, "loss": 1.0314, "step": 540 }, { "epoch": 0.6787954830614805, "grad_norm": 0.2216796875, "learning_rate": 7.737348389795065e-06, "loss": 0.9991, "step": 541 }, { "epoch": 0.6800501882057717, "grad_norm": 0.2314453125, "learning_rate": 7.733166039314095e-06, "loss": 0.9787, "step": 542 }, { "epoch": 0.6813048933500627, "grad_norm": 0.2373046875, "learning_rate": 7.728983688833126e-06, "loss": 1.0142, "step": 543 }, { "epoch": 0.6825595984943539, "grad_norm": 0.2255859375, "learning_rate": 7.724801338352154e-06, "loss": 0.9687, "step": 544 }, { "epoch": 0.6838143036386449, "grad_norm": 0.23046875, "learning_rate": 7.720618987871185e-06, "loss": 1.0039, "step": 545 }, { "epoch": 0.685069008782936, "grad_norm": 0.23046875, "learning_rate": 7.716436637390214e-06, "loss": 1.0017, "step": 546 }, { "epoch": 0.6863237139272271, "grad_norm": 0.294921875, "learning_rate": 7.712254286909243e-06, "loss": 0.9462, "step": 547 }, { "epoch": 0.6875784190715182, "grad_norm": 0.26953125, "learning_rate": 7.708071936428273e-06, "loss": 0.9834, "step": 548 }, { "epoch": 0.6888331242158093, "grad_norm": 0.2421875, "learning_rate": 7.703889585947302e-06, "loss": 0.9929, "step": 549 }, { "epoch": 0.6900878293601004, "grad_norm": 0.2490234375, "learning_rate": 7.699707235466333e-06, "loss": 1.0571, "step": 550 }, { "epoch": 0.6913425345043914, "grad_norm": 0.259765625, "learning_rate": 7.695524884985363e-06, "loss": 1.042, "step": 551 }, { "epoch": 0.6925972396486826, "grad_norm": 0.30078125, "learning_rate": 7.691342534504392e-06, "loss": 0.9764, "step": 552 }, { "epoch": 0.6938519447929736, "grad_norm": 0.26953125, "learning_rate": 7.687160184023422e-06, "loss": 0.9694, "step": 553 }, { "epoch": 0.6951066499372648, "grad_norm": 0.279296875, "learning_rate": 7.682977833542451e-06, "loss": 0.9885, "step": 554 }, { "epoch": 0.6963613550815558, "grad_norm": 0.236328125, "learning_rate": 7.67879548306148e-06, "loss": 1.0018, "step": 555 }, { "epoch": 0.6976160602258469, "grad_norm": 0.2353515625, "learning_rate": 7.67461313258051e-06, "loss": 1.021, "step": 556 }, { "epoch": 0.698870765370138, "grad_norm": 0.248046875, "learning_rate": 7.670430782099541e-06, "loss": 0.9756, "step": 557 }, { "epoch": 0.7001254705144291, "grad_norm": 0.298828125, "learning_rate": 7.66624843161857e-06, "loss": 0.9966, "step": 558 }, { "epoch": 0.7013801756587202, "grad_norm": 0.275390625, "learning_rate": 7.6620660811376e-06, "loss": 1.0418, "step": 559 }, { "epoch": 0.7026348808030113, "grad_norm": 0.2490234375, "learning_rate": 7.65788373065663e-06, "loss": 0.9837, "step": 560 }, { "epoch": 0.7038895859473023, "grad_norm": 0.263671875, "learning_rate": 7.65370138017566e-06, "loss": 0.9657, "step": 561 }, { "epoch": 0.7051442910915935, "grad_norm": 1.640625, "learning_rate": 7.649519029694689e-06, "loss": 0.9667, "step": 562 }, { "epoch": 0.7063989962358845, "grad_norm": 0.26171875, "learning_rate": 7.64533667921372e-06, "loss": 1.0314, "step": 563 }, { "epoch": 0.7076537013801757, "grad_norm": 0.2421875, "learning_rate": 7.641154328732748e-06, "loss": 1.0065, "step": 564 }, { "epoch": 0.7089084065244667, "grad_norm": 0.267578125, "learning_rate": 7.636971978251779e-06, "loss": 0.993, "step": 565 }, { "epoch": 0.7101631116687579, "grad_norm": 0.240234375, "learning_rate": 7.632789627770807e-06, "loss": 0.9817, "step": 566 }, { "epoch": 0.7114178168130489, "grad_norm": 0.27734375, "learning_rate": 7.628607277289838e-06, "loss": 1.0113, "step": 567 }, { "epoch": 0.71267252195734, "grad_norm": 0.2294921875, "learning_rate": 7.6244249268088676e-06, "loss": 0.9671, "step": 568 }, { "epoch": 0.7139272271016311, "grad_norm": 0.2255859375, "learning_rate": 7.620242576327896e-06, "loss": 0.9764, "step": 569 }, { "epoch": 0.7151819322459222, "grad_norm": 0.23828125, "learning_rate": 7.616060225846926e-06, "loss": 1.005, "step": 570 }, { "epoch": 0.7164366373902133, "grad_norm": 0.24609375, "learning_rate": 7.611877875365956e-06, "loss": 0.9808, "step": 571 }, { "epoch": 0.7176913425345044, "grad_norm": 0.259765625, "learning_rate": 7.607695524884986e-06, "loss": 0.9815, "step": 572 }, { "epoch": 0.7189460476787954, "grad_norm": 0.25, "learning_rate": 7.603513174404016e-06, "loss": 0.9592, "step": 573 }, { "epoch": 0.7202007528230866, "grad_norm": 0.2373046875, "learning_rate": 7.599330823923046e-06, "loss": 1.0006, "step": 574 }, { "epoch": 0.7214554579673776, "grad_norm": 0.2431640625, "learning_rate": 7.595148473442075e-06, "loss": 0.9763, "step": 575 }, { "epoch": 0.7227101631116688, "grad_norm": 0.248046875, "learning_rate": 7.590966122961104e-06, "loss": 0.9814, "step": 576 }, { "epoch": 0.7239648682559598, "grad_norm": 0.263671875, "learning_rate": 7.586783772480134e-06, "loss": 0.9268, "step": 577 }, { "epoch": 0.7252195734002509, "grad_norm": 0.25390625, "learning_rate": 7.582601421999164e-06, "loss": 0.9824, "step": 578 }, { "epoch": 0.726474278544542, "grad_norm": 0.2431640625, "learning_rate": 7.578419071518194e-06, "loss": 0.9992, "step": 579 }, { "epoch": 0.7277289836888331, "grad_norm": 0.2451171875, "learning_rate": 7.574236721037224e-06, "loss": 1.0384, "step": 580 }, { "epoch": 0.7289836888331243, "grad_norm": 0.25, "learning_rate": 7.570054370556253e-06, "loss": 1.0103, "step": 581 }, { "epoch": 0.7302383939774153, "grad_norm": 0.2451171875, "learning_rate": 7.565872020075283e-06, "loss": 0.9842, "step": 582 }, { "epoch": 0.7314930991217063, "grad_norm": 0.30078125, "learning_rate": 7.5616896695943135e-06, "loss": 1.0319, "step": 583 }, { "epoch": 0.7327478042659975, "grad_norm": 0.2392578125, "learning_rate": 7.5575073191133415e-06, "loss": 1.0063, "step": 584 }, { "epoch": 0.7340025094102886, "grad_norm": 0.2412109375, "learning_rate": 7.553324968632372e-06, "loss": 1.026, "step": 585 }, { "epoch": 0.7352572145545797, "grad_norm": 0.23828125, "learning_rate": 7.549142618151402e-06, "loss": 0.9935, "step": 586 }, { "epoch": 0.7365119196988708, "grad_norm": 0.2392578125, "learning_rate": 7.544960267670431e-06, "loss": 0.9443, "step": 587 }, { "epoch": 0.7377666248431619, "grad_norm": 0.25390625, "learning_rate": 7.540777917189461e-06, "loss": 0.9663, "step": 588 }, { "epoch": 0.739021329987453, "grad_norm": 0.255859375, "learning_rate": 7.536595566708491e-06, "loss": 1.0124, "step": 589 }, { "epoch": 0.740276035131744, "grad_norm": 0.2890625, "learning_rate": 7.532413216227521e-06, "loss": 1.0193, "step": 590 }, { "epoch": 0.7415307402760352, "grad_norm": 0.2314453125, "learning_rate": 7.52823086574655e-06, "loss": 0.9373, "step": 591 }, { "epoch": 0.7427854454203262, "grad_norm": 0.240234375, "learning_rate": 7.52404851526558e-06, "loss": 1.0066, "step": 592 }, { "epoch": 0.7440401505646174, "grad_norm": 0.2421875, "learning_rate": 7.5198661647846094e-06, "loss": 1.0164, "step": 593 }, { "epoch": 0.7452948557089084, "grad_norm": 0.28125, "learning_rate": 7.515683814303639e-06, "loss": 1.0045, "step": 594 }, { "epoch": 0.7465495608531995, "grad_norm": 0.255859375, "learning_rate": 7.511501463822669e-06, "loss": 1.0181, "step": 595 }, { "epoch": 0.7478042659974906, "grad_norm": 0.26171875, "learning_rate": 7.507319113341699e-06, "loss": 0.9871, "step": 596 }, { "epoch": 0.7490589711417817, "grad_norm": 0.2353515625, "learning_rate": 7.503136762860729e-06, "loss": 0.9447, "step": 597 }, { "epoch": 0.7503136762860728, "grad_norm": 0.255859375, "learning_rate": 7.498954412379758e-06, "loss": 0.9238, "step": 598 }, { "epoch": 0.7515683814303639, "grad_norm": 0.2451171875, "learning_rate": 7.4947720618987875e-06, "loss": 1.0066, "step": 599 }, { "epoch": 0.7528230865746549, "grad_norm": 0.228515625, "learning_rate": 7.490589711417817e-06, "loss": 0.9654, "step": 600 }, { "epoch": 0.7540777917189461, "grad_norm": 0.2734375, "learning_rate": 7.486407360936847e-06, "loss": 0.9177, "step": 601 }, { "epoch": 0.7553324968632371, "grad_norm": 0.2392578125, "learning_rate": 7.4822250104558765e-06, "loss": 1.0188, "step": 602 }, { "epoch": 0.7565872020075283, "grad_norm": 0.23828125, "learning_rate": 7.478042659974907e-06, "loss": 0.9891, "step": 603 }, { "epoch": 0.7578419071518193, "grad_norm": 0.26171875, "learning_rate": 7.473860309493937e-06, "loss": 1.004, "step": 604 }, { "epoch": 0.7590966122961104, "grad_norm": 0.228515625, "learning_rate": 7.469677959012966e-06, "loss": 0.9574, "step": 605 }, { "epoch": 0.7603513174404015, "grad_norm": 0.2578125, "learning_rate": 7.465495608531995e-06, "loss": 1.0011, "step": 606 }, { "epoch": 0.7616060225846926, "grad_norm": 0.35546875, "learning_rate": 7.461313258051025e-06, "loss": 1.0357, "step": 607 }, { "epoch": 0.7628607277289837, "grad_norm": 0.23828125, "learning_rate": 7.4571309075700546e-06, "loss": 0.954, "step": 608 }, { "epoch": 0.7641154328732748, "grad_norm": 0.298828125, "learning_rate": 7.452948557089085e-06, "loss": 0.9904, "step": 609 }, { "epoch": 0.7653701380175659, "grad_norm": 0.25390625, "learning_rate": 7.448766206608115e-06, "loss": 1.0058, "step": 610 }, { "epoch": 0.766624843161857, "grad_norm": 0.267578125, "learning_rate": 7.4445838561271445e-06, "loss": 1.0108, "step": 611 }, { "epoch": 0.767879548306148, "grad_norm": 0.341796875, "learning_rate": 7.440401505646174e-06, "loss": 1.0107, "step": 612 }, { "epoch": 0.7691342534504392, "grad_norm": 0.2431640625, "learning_rate": 7.436219155165203e-06, "loss": 1.0126, "step": 613 }, { "epoch": 0.7703889585947302, "grad_norm": 0.25, "learning_rate": 7.432036804684233e-06, "loss": 0.9537, "step": 614 }, { "epoch": 0.7716436637390214, "grad_norm": 0.2265625, "learning_rate": 7.427854454203262e-06, "loss": 0.9351, "step": 615 }, { "epoch": 0.7728983688833124, "grad_norm": 0.25390625, "learning_rate": 7.423672103722293e-06, "loss": 1.0069, "step": 616 }, { "epoch": 0.7741530740276035, "grad_norm": 0.2490234375, "learning_rate": 7.4194897532413225e-06, "loss": 0.9449, "step": 617 }, { "epoch": 0.7754077791718946, "grad_norm": 0.251953125, "learning_rate": 7.415307402760352e-06, "loss": 0.9476, "step": 618 }, { "epoch": 0.7766624843161857, "grad_norm": 0.2578125, "learning_rate": 7.411125052279382e-06, "loss": 1.0401, "step": 619 }, { "epoch": 0.7779171894604768, "grad_norm": 0.271484375, "learning_rate": 7.406942701798411e-06, "loss": 0.9515, "step": 620 }, { "epoch": 0.7791718946047679, "grad_norm": 0.265625, "learning_rate": 7.40276035131744e-06, "loss": 0.9915, "step": 621 }, { "epoch": 0.7804265997490589, "grad_norm": 0.255859375, "learning_rate": 7.398578000836471e-06, "loss": 1.0144, "step": 622 }, { "epoch": 0.7816813048933501, "grad_norm": 0.2314453125, "learning_rate": 7.3943956503555006e-06, "loss": 0.9698, "step": 623 }, { "epoch": 0.7829360100376411, "grad_norm": 0.2314453125, "learning_rate": 7.39021329987453e-06, "loss": 0.9902, "step": 624 }, { "epoch": 0.7841907151819323, "grad_norm": 0.2314453125, "learning_rate": 7.38603094939356e-06, "loss": 0.9666, "step": 625 }, { "epoch": 0.7854454203262233, "grad_norm": 0.2421875, "learning_rate": 7.38184859891259e-06, "loss": 0.9647, "step": 626 }, { "epoch": 0.7867001254705144, "grad_norm": 0.251953125, "learning_rate": 7.37766624843162e-06, "loss": 1.0111, "step": 627 }, { "epoch": 0.7879548306148055, "grad_norm": 0.2578125, "learning_rate": 7.373483897950648e-06, "loss": 0.9626, "step": 628 }, { "epoch": 0.7892095357590966, "grad_norm": 0.259765625, "learning_rate": 7.369301547469679e-06, "loss": 1.0255, "step": 629 }, { "epoch": 0.7904642409033877, "grad_norm": 0.25, "learning_rate": 7.365119196988708e-06, "loss": 0.973, "step": 630 }, { "epoch": 0.7917189460476788, "grad_norm": 0.24609375, "learning_rate": 7.360936846507738e-06, "loss": 0.9947, "step": 631 }, { "epoch": 0.7929736511919699, "grad_norm": 0.25390625, "learning_rate": 7.356754496026768e-06, "loss": 0.9986, "step": 632 }, { "epoch": 0.794228356336261, "grad_norm": 0.25, "learning_rate": 7.352572145545797e-06, "loss": 0.9833, "step": 633 }, { "epoch": 0.795483061480552, "grad_norm": 0.25, "learning_rate": 7.348389795064828e-06, "loss": 0.9664, "step": 634 }, { "epoch": 0.7967377666248432, "grad_norm": 0.236328125, "learning_rate": 7.344207444583857e-06, "loss": 0.9478, "step": 635 }, { "epoch": 0.7979924717691342, "grad_norm": 0.23828125, "learning_rate": 7.340025094102886e-06, "loss": 1.0257, "step": 636 }, { "epoch": 0.7992471769134254, "grad_norm": 0.265625, "learning_rate": 7.335842743621916e-06, "loss": 0.9644, "step": 637 }, { "epoch": 0.8005018820577164, "grad_norm": 0.2412109375, "learning_rate": 7.331660393140946e-06, "loss": 0.9687, "step": 638 }, { "epoch": 0.8017565872020075, "grad_norm": 0.259765625, "learning_rate": 7.327478042659975e-06, "loss": 1.0024, "step": 639 }, { "epoch": 0.8030112923462986, "grad_norm": 0.232421875, "learning_rate": 7.323295692179006e-06, "loss": 0.9872, "step": 640 }, { "epoch": 0.8042659974905897, "grad_norm": 0.248046875, "learning_rate": 7.3191133416980356e-06, "loss": 0.965, "step": 641 }, { "epoch": 0.8055207026348808, "grad_norm": 0.314453125, "learning_rate": 7.314930991217064e-06, "loss": 0.9634, "step": 642 }, { "epoch": 0.8067754077791719, "grad_norm": 0.236328125, "learning_rate": 7.310748640736094e-06, "loss": 0.967, "step": 643 }, { "epoch": 0.8080301129234629, "grad_norm": 0.24609375, "learning_rate": 7.306566290255124e-06, "loss": 0.969, "step": 644 }, { "epoch": 0.8092848180677541, "grad_norm": 0.2431640625, "learning_rate": 7.302383939774153e-06, "loss": 0.9747, "step": 645 }, { "epoch": 0.8105395232120451, "grad_norm": 0.25, "learning_rate": 7.298201589293183e-06, "loss": 1.0123, "step": 646 }, { "epoch": 0.8117942283563363, "grad_norm": 0.2392578125, "learning_rate": 7.294019238812214e-06, "loss": 0.9774, "step": 647 }, { "epoch": 0.8130489335006273, "grad_norm": 0.26171875, "learning_rate": 7.289836888331243e-06, "loss": 1.0019, "step": 648 }, { "epoch": 0.8143036386449184, "grad_norm": 0.25390625, "learning_rate": 7.285654537850273e-06, "loss": 1.0455, "step": 649 }, { "epoch": 0.8155583437892095, "grad_norm": 0.2421875, "learning_rate": 7.281472187369302e-06, "loss": 1.0004, "step": 650 }, { "epoch": 0.8168130489335006, "grad_norm": 0.25, "learning_rate": 7.2772898368883315e-06, "loss": 1.0068, "step": 651 }, { "epoch": 0.8180677540777918, "grad_norm": 0.2451171875, "learning_rate": 7.273107486407361e-06, "loss": 0.9283, "step": 652 }, { "epoch": 0.8193224592220828, "grad_norm": 0.251953125, "learning_rate": 7.268925135926392e-06, "loss": 0.9339, "step": 653 }, { "epoch": 0.820577164366374, "grad_norm": 0.2373046875, "learning_rate": 7.264742785445421e-06, "loss": 1.0247, "step": 654 }, { "epoch": 0.821831869510665, "grad_norm": 0.2431640625, "learning_rate": 7.260560434964451e-06, "loss": 0.9598, "step": 655 }, { "epoch": 0.823086574654956, "grad_norm": 0.23046875, "learning_rate": 7.256378084483481e-06, "loss": 0.951, "step": 656 }, { "epoch": 0.8243412797992472, "grad_norm": 0.271484375, "learning_rate": 7.2521957340025095e-06, "loss": 0.9663, "step": 657 }, { "epoch": 0.8255959849435383, "grad_norm": 0.24609375, "learning_rate": 7.248013383521539e-06, "loss": 0.9792, "step": 658 }, { "epoch": 0.8268506900878294, "grad_norm": 0.234375, "learning_rate": 7.243831033040569e-06, "loss": 0.9126, "step": 659 }, { "epoch": 0.8281053952321205, "grad_norm": 0.27734375, "learning_rate": 7.239648682559599e-06, "loss": 0.9875, "step": 660 }, { "epoch": 0.8293601003764115, "grad_norm": 0.30859375, "learning_rate": 7.235466332078629e-06, "loss": 0.9696, "step": 661 }, { "epoch": 0.8306148055207027, "grad_norm": 0.228515625, "learning_rate": 7.231283981597659e-06, "loss": 0.937, "step": 662 }, { "epoch": 0.8318695106649937, "grad_norm": 0.248046875, "learning_rate": 7.2271016311166884e-06, "loss": 0.9761, "step": 663 }, { "epoch": 0.8331242158092849, "grad_norm": 0.251953125, "learning_rate": 7.222919280635717e-06, "loss": 1.0537, "step": 664 }, { "epoch": 0.8343789209535759, "grad_norm": 0.2373046875, "learning_rate": 7.218736930154747e-06, "loss": 0.9682, "step": 665 }, { "epoch": 0.835633626097867, "grad_norm": 0.302734375, "learning_rate": 7.2145545796737775e-06, "loss": 0.9608, "step": 666 }, { "epoch": 0.8368883312421581, "grad_norm": 0.2734375, "learning_rate": 7.210372229192807e-06, "loss": 0.9631, "step": 667 }, { "epoch": 0.8381430363864492, "grad_norm": 0.24609375, "learning_rate": 7.206189878711837e-06, "loss": 0.955, "step": 668 }, { "epoch": 0.8393977415307403, "grad_norm": 0.234375, "learning_rate": 7.2020075282308665e-06, "loss": 0.9982, "step": 669 }, { "epoch": 0.8406524466750314, "grad_norm": 0.255859375, "learning_rate": 7.197825177749896e-06, "loss": 1.0204, "step": 670 }, { "epoch": 0.8419071518193224, "grad_norm": 0.265625, "learning_rate": 7.193642827268927e-06, "loss": 0.9727, "step": 671 }, { "epoch": 0.8431618569636136, "grad_norm": 0.2578125, "learning_rate": 7.189460476787955e-06, "loss": 1.0109, "step": 672 }, { "epoch": 0.8444165621079046, "grad_norm": 0.25, "learning_rate": 7.185278126306985e-06, "loss": 0.9816, "step": 673 }, { "epoch": 0.8456712672521958, "grad_norm": 0.263671875, "learning_rate": 7.181095775826015e-06, "loss": 1.017, "step": 674 }, { "epoch": 0.8469259723964868, "grad_norm": 0.251953125, "learning_rate": 7.1769134253450445e-06, "loss": 0.9983, "step": 675 }, { "epoch": 0.848180677540778, "grad_norm": 0.2373046875, "learning_rate": 7.172731074864074e-06, "loss": 0.9558, "step": 676 }, { "epoch": 0.849435382685069, "grad_norm": 0.3515625, "learning_rate": 7.168548724383104e-06, "loss": 0.9782, "step": 677 }, { "epoch": 0.8506900878293601, "grad_norm": 0.24609375, "learning_rate": 7.164366373902134e-06, "loss": 0.9384, "step": 678 }, { "epoch": 0.8519447929736512, "grad_norm": 0.2470703125, "learning_rate": 7.160184023421162e-06, "loss": 0.9683, "step": 679 }, { "epoch": 0.8531994981179423, "grad_norm": 0.259765625, "learning_rate": 7.156001672940193e-06, "loss": 1.0279, "step": 680 }, { "epoch": 0.8544542032622334, "grad_norm": 0.30078125, "learning_rate": 7.151819322459223e-06, "loss": 0.9925, "step": 681 }, { "epoch": 0.8557089084065245, "grad_norm": 0.25, "learning_rate": 7.147636971978252e-06, "loss": 0.9527, "step": 682 }, { "epoch": 0.8569636135508155, "grad_norm": 0.25390625, "learning_rate": 7.143454621497282e-06, "loss": 1.0443, "step": 683 }, { "epoch": 0.8582183186951067, "grad_norm": 0.26171875, "learning_rate": 7.1392722710163125e-06, "loss": 1.0051, "step": 684 }, { "epoch": 0.8594730238393977, "grad_norm": 0.244140625, "learning_rate": 7.135089920535342e-06, "loss": 0.9797, "step": 685 }, { "epoch": 0.8607277289836889, "grad_norm": 0.37109375, "learning_rate": 7.130907570054371e-06, "loss": 0.9537, "step": 686 }, { "epoch": 0.8619824341279799, "grad_norm": 0.259765625, "learning_rate": 7.126725219573401e-06, "loss": 0.9671, "step": 687 }, { "epoch": 0.863237139272271, "grad_norm": 0.2373046875, "learning_rate": 7.12254286909243e-06, "loss": 0.966, "step": 688 }, { "epoch": 0.8644918444165621, "grad_norm": 0.251953125, "learning_rate": 7.11836051861146e-06, "loss": 0.9556, "step": 689 }, { "epoch": 0.8657465495608532, "grad_norm": 0.248046875, "learning_rate": 7.11417816813049e-06, "loss": 0.9796, "step": 690 }, { "epoch": 0.8670012547051443, "grad_norm": 0.2373046875, "learning_rate": 7.10999581764952e-06, "loss": 0.9604, "step": 691 }, { "epoch": 0.8682559598494354, "grad_norm": 0.2421875, "learning_rate": 7.10581346716855e-06, "loss": 0.9847, "step": 692 }, { "epoch": 0.8695106649937264, "grad_norm": 0.2578125, "learning_rate": 7.1016311166875795e-06, "loss": 1.0696, "step": 693 }, { "epoch": 0.8707653701380176, "grad_norm": 0.2353515625, "learning_rate": 7.097448766206608e-06, "loss": 0.98, "step": 694 }, { "epoch": 0.8720200752823086, "grad_norm": 0.31640625, "learning_rate": 7.093266415725638e-06, "loss": 0.9398, "step": 695 }, { "epoch": 0.8732747804265998, "grad_norm": 0.240234375, "learning_rate": 7.089084065244668e-06, "loss": 0.9584, "step": 696 }, { "epoch": 0.8745294855708908, "grad_norm": 0.3359375, "learning_rate": 7.084901714763697e-06, "loss": 0.9398, "step": 697 }, { "epoch": 0.875784190715182, "grad_norm": 0.2392578125, "learning_rate": 7.080719364282728e-06, "loss": 0.9863, "step": 698 }, { "epoch": 0.877038895859473, "grad_norm": 0.248046875, "learning_rate": 7.076537013801758e-06, "loss": 1.0106, "step": 699 }, { "epoch": 0.8782936010037641, "grad_norm": 0.255859375, "learning_rate": 7.072354663320787e-06, "loss": 0.9878, "step": 700 }, { "epoch": 0.8795483061480552, "grad_norm": 0.275390625, "learning_rate": 7.068172312839816e-06, "loss": 0.9684, "step": 701 }, { "epoch": 0.8808030112923463, "grad_norm": 0.255859375, "learning_rate": 7.063989962358846e-06, "loss": 0.9889, "step": 702 }, { "epoch": 0.8820577164366374, "grad_norm": 0.2734375, "learning_rate": 7.0598076118778754e-06, "loss": 0.9339, "step": 703 }, { "epoch": 0.8833124215809285, "grad_norm": 0.251953125, "learning_rate": 7.055625261396906e-06, "loss": 0.9748, "step": 704 }, { "epoch": 0.8845671267252195, "grad_norm": 0.2392578125, "learning_rate": 7.051442910915936e-06, "loss": 0.9648, "step": 705 }, { "epoch": 0.8858218318695107, "grad_norm": 0.25390625, "learning_rate": 7.047260560434965e-06, "loss": 0.9576, "step": 706 }, { "epoch": 0.8870765370138017, "grad_norm": 0.26171875, "learning_rate": 7.043078209953995e-06, "loss": 0.9451, "step": 707 }, { "epoch": 0.8883312421580929, "grad_norm": 0.23828125, "learning_rate": 7.038895859473024e-06, "loss": 0.9448, "step": 708 }, { "epoch": 0.8895859473023839, "grad_norm": 0.251953125, "learning_rate": 7.0347135089920535e-06, "loss": 0.9245, "step": 709 }, { "epoch": 0.890840652446675, "grad_norm": 0.27734375, "learning_rate": 7.030531158511083e-06, "loss": 1.0323, "step": 710 }, { "epoch": 0.8920953575909661, "grad_norm": 0.279296875, "learning_rate": 7.026348808030114e-06, "loss": 0.9723, "step": 711 }, { "epoch": 0.8933500627352572, "grad_norm": 0.294921875, "learning_rate": 7.022166457549143e-06, "loss": 0.926, "step": 712 }, { "epoch": 0.8946047678795483, "grad_norm": 0.2890625, "learning_rate": 7.017984107068173e-06, "loss": 1.001, "step": 713 }, { "epoch": 0.8958594730238394, "grad_norm": 0.2421875, "learning_rate": 7.013801756587203e-06, "loss": 0.9585, "step": 714 }, { "epoch": 0.8971141781681304, "grad_norm": 0.244140625, "learning_rate": 7.009619406106232e-06, "loss": 0.9695, "step": 715 }, { "epoch": 0.8983688833124216, "grad_norm": 0.25, "learning_rate": 7.005437055625261e-06, "loss": 0.9906, "step": 716 }, { "epoch": 0.8996235884567126, "grad_norm": 0.248046875, "learning_rate": 7.001254705144292e-06, "loss": 0.9556, "step": 717 }, { "epoch": 0.9008782936010038, "grad_norm": 0.267578125, "learning_rate": 6.9970723546633214e-06, "loss": 1.0158, "step": 718 }, { "epoch": 0.9021329987452948, "grad_norm": 0.2431640625, "learning_rate": 6.992890004182351e-06, "loss": 0.9688, "step": 719 }, { "epoch": 0.903387703889586, "grad_norm": 0.240234375, "learning_rate": 6.988707653701381e-06, "loss": 0.9522, "step": 720 }, { "epoch": 0.904642409033877, "grad_norm": 0.24609375, "learning_rate": 6.9845253032204105e-06, "loss": 0.9699, "step": 721 }, { "epoch": 0.9058971141781681, "grad_norm": 0.2392578125, "learning_rate": 6.980342952739441e-06, "loss": 0.9313, "step": 722 }, { "epoch": 0.9071518193224593, "grad_norm": 0.255859375, "learning_rate": 6.976160602258469e-06, "loss": 0.9804, "step": 723 }, { "epoch": 0.9084065244667503, "grad_norm": 0.2578125, "learning_rate": 6.9719782517774995e-06, "loss": 0.9639, "step": 724 }, { "epoch": 0.9096612296110415, "grad_norm": 0.287109375, "learning_rate": 6.967795901296529e-06, "loss": 0.9788, "step": 725 }, { "epoch": 0.9109159347553325, "grad_norm": 0.271484375, "learning_rate": 6.963613550815559e-06, "loss": 0.9912, "step": 726 }, { "epoch": 0.9121706398996235, "grad_norm": 0.248046875, "learning_rate": 6.9594312003345885e-06, "loss": 0.9294, "step": 727 }, { "epoch": 0.9134253450439147, "grad_norm": 0.28515625, "learning_rate": 6.955248849853618e-06, "loss": 0.9283, "step": 728 }, { "epoch": 0.9146800501882058, "grad_norm": 0.23828125, "learning_rate": 6.951066499372649e-06, "loss": 0.9862, "step": 729 }, { "epoch": 0.9159347553324969, "grad_norm": 0.251953125, "learning_rate": 6.9468841488916775e-06, "loss": 0.9911, "step": 730 }, { "epoch": 0.917189460476788, "grad_norm": 0.267578125, "learning_rate": 6.942701798410707e-06, "loss": 0.9994, "step": 731 }, { "epoch": 0.918444165621079, "grad_norm": 0.2431640625, "learning_rate": 6.938519447929737e-06, "loss": 0.9083, "step": 732 }, { "epoch": 0.9196988707653702, "grad_norm": 0.26953125, "learning_rate": 6.9343370974487666e-06, "loss": 0.9723, "step": 733 }, { "epoch": 0.9209535759096612, "grad_norm": 0.2490234375, "learning_rate": 6.930154746967796e-06, "loss": 0.9541, "step": 734 }, { "epoch": 0.9222082810539524, "grad_norm": 0.28515625, "learning_rate": 6.925972396486827e-06, "loss": 0.9161, "step": 735 }, { "epoch": 0.9234629861982434, "grad_norm": 0.248046875, "learning_rate": 6.9217900460058564e-06, "loss": 0.9756, "step": 736 }, { "epoch": 0.9247176913425345, "grad_norm": 0.255859375, "learning_rate": 6.917607695524886e-06, "loss": 0.9605, "step": 737 }, { "epoch": 0.9259723964868256, "grad_norm": 0.5078125, "learning_rate": 6.913425345043915e-06, "loss": 0.9749, "step": 738 }, { "epoch": 0.9272271016311167, "grad_norm": 0.287109375, "learning_rate": 6.909242994562945e-06, "loss": 0.9749, "step": 739 }, { "epoch": 0.9284818067754078, "grad_norm": 0.248046875, "learning_rate": 6.905060644081974e-06, "loss": 0.9503, "step": 740 }, { "epoch": 0.9297365119196989, "grad_norm": 0.25390625, "learning_rate": 6.900878293601004e-06, "loss": 0.9943, "step": 741 }, { "epoch": 0.93099121706399, "grad_norm": 0.2470703125, "learning_rate": 6.8966959431200345e-06, "loss": 0.9807, "step": 742 }, { "epoch": 0.9322459222082811, "grad_norm": 0.298828125, "learning_rate": 6.892513592639064e-06, "loss": 1.0096, "step": 743 }, { "epoch": 0.9335006273525721, "grad_norm": 0.25, "learning_rate": 6.888331242158094e-06, "loss": 0.9472, "step": 744 }, { "epoch": 0.9347553324968633, "grad_norm": 0.263671875, "learning_rate": 6.884148891677123e-06, "loss": 0.9674, "step": 745 }, { "epoch": 0.9360100376411543, "grad_norm": 0.2373046875, "learning_rate": 6.879966541196152e-06, "loss": 0.9846, "step": 746 }, { "epoch": 0.9372647427854455, "grad_norm": 0.26171875, "learning_rate": 6.875784190715182e-06, "loss": 0.9981, "step": 747 }, { "epoch": 0.9385194479297365, "grad_norm": 0.255859375, "learning_rate": 6.8716018402342125e-06, "loss": 1.06, "step": 748 }, { "epoch": 0.9397741530740276, "grad_norm": 0.25, "learning_rate": 6.867419489753242e-06, "loss": 0.9554, "step": 749 }, { "epoch": 0.9410288582183187, "grad_norm": 0.27734375, "learning_rate": 6.863237139272272e-06, "loss": 1.0162, "step": 750 }, { "epoch": 0.9422835633626098, "grad_norm": 0.2412109375, "learning_rate": 6.8590547887913016e-06, "loss": 0.9535, "step": 751 }, { "epoch": 0.9435382685069009, "grad_norm": 0.287109375, "learning_rate": 6.85487243831033e-06, "loss": 0.9817, "step": 752 }, { "epoch": 0.944792973651192, "grad_norm": 0.25390625, "learning_rate": 6.85069008782936e-06, "loss": 0.9809, "step": 753 }, { "epoch": 0.946047678795483, "grad_norm": 0.263671875, "learning_rate": 6.84650773734839e-06, "loss": 0.9686, "step": 754 }, { "epoch": 0.9473023839397742, "grad_norm": 0.255859375, "learning_rate": 6.84232538686742e-06, "loss": 0.9964, "step": 755 }, { "epoch": 0.9485570890840652, "grad_norm": 0.306640625, "learning_rate": 6.83814303638645e-06, "loss": 0.9806, "step": 756 }, { "epoch": 0.9498117942283564, "grad_norm": 0.279296875, "learning_rate": 6.83396068590548e-06, "loss": 0.9909, "step": 757 }, { "epoch": 0.9510664993726474, "grad_norm": 0.263671875, "learning_rate": 6.829778335424509e-06, "loss": 0.9812, "step": 758 }, { "epoch": 0.9523212045169385, "grad_norm": 0.26953125, "learning_rate": 6.825595984943539e-06, "loss": 0.9691, "step": 759 }, { "epoch": 0.9535759096612296, "grad_norm": 0.265625, "learning_rate": 6.821413634462568e-06, "loss": 0.9782, "step": 760 }, { "epoch": 0.9548306148055207, "grad_norm": 0.2451171875, "learning_rate": 6.817231283981598e-06, "loss": 0.9326, "step": 761 }, { "epoch": 0.9560853199498118, "grad_norm": 0.2578125, "learning_rate": 6.813048933500628e-06, "loss": 1.0102, "step": 762 }, { "epoch": 0.9573400250941029, "grad_norm": 0.251953125, "learning_rate": 6.808866583019658e-06, "loss": 1.0099, "step": 763 }, { "epoch": 0.958594730238394, "grad_norm": 0.236328125, "learning_rate": 6.804684232538687e-06, "loss": 0.9749, "step": 764 }, { "epoch": 0.9598494353826851, "grad_norm": 0.271484375, "learning_rate": 6.800501882057717e-06, "loss": 0.989, "step": 765 }, { "epoch": 0.9611041405269761, "grad_norm": 0.23046875, "learning_rate": 6.7963195315767476e-06, "loss": 0.9656, "step": 766 }, { "epoch": 0.9623588456712673, "grad_norm": 0.251953125, "learning_rate": 6.7921371810957755e-06, "loss": 0.954, "step": 767 }, { "epoch": 0.9636135508155583, "grad_norm": 0.2431640625, "learning_rate": 6.787954830614806e-06, "loss": 0.9587, "step": 768 }, { "epoch": 0.9648682559598495, "grad_norm": 0.255859375, "learning_rate": 6.783772480133836e-06, "loss": 0.9598, "step": 769 }, { "epoch": 0.9661229611041405, "grad_norm": 0.259765625, "learning_rate": 6.779590129652865e-06, "loss": 0.9435, "step": 770 }, { "epoch": 0.9673776662484316, "grad_norm": 0.23828125, "learning_rate": 6.775407779171895e-06, "loss": 0.9629, "step": 771 }, { "epoch": 0.9686323713927227, "grad_norm": 0.26171875, "learning_rate": 6.771225428690925e-06, "loss": 1.0219, "step": 772 }, { "epoch": 0.9698870765370138, "grad_norm": 0.2392578125, "learning_rate": 6.767043078209955e-06, "loss": 0.9506, "step": 773 }, { "epoch": 0.9711417816813049, "grad_norm": 0.259765625, "learning_rate": 6.762860727728984e-06, "loss": 0.9623, "step": 774 }, { "epoch": 0.972396486825596, "grad_norm": 0.24609375, "learning_rate": 6.758678377248014e-06, "loss": 0.9421, "step": 775 }, { "epoch": 0.973651191969887, "grad_norm": 0.2412109375, "learning_rate": 6.7544960267670435e-06, "loss": 0.9569, "step": 776 }, { "epoch": 0.9749058971141782, "grad_norm": 0.267578125, "learning_rate": 6.750313676286073e-06, "loss": 0.9911, "step": 777 }, { "epoch": 0.9761606022584692, "grad_norm": 0.25390625, "learning_rate": 6.746131325805103e-06, "loss": 0.9812, "step": 778 }, { "epoch": 0.9774153074027604, "grad_norm": 0.251953125, "learning_rate": 6.741948975324133e-06, "loss": 0.9462, "step": 779 }, { "epoch": 0.9786700125470514, "grad_norm": 0.341796875, "learning_rate": 6.737766624843163e-06, "loss": 1.0152, "step": 780 }, { "epoch": 0.9799247176913425, "grad_norm": 0.240234375, "learning_rate": 6.733584274362193e-06, "loss": 0.9435, "step": 781 }, { "epoch": 0.9811794228356336, "grad_norm": 0.259765625, "learning_rate": 6.7294019238812215e-06, "loss": 0.9925, "step": 782 }, { "epoch": 0.9824341279799247, "grad_norm": 0.263671875, "learning_rate": 6.725219573400251e-06, "loss": 1.0352, "step": 783 }, { "epoch": 0.9836888331242158, "grad_norm": 0.294921875, "learning_rate": 6.721037222919281e-06, "loss": 0.9998, "step": 784 }, { "epoch": 0.9849435382685069, "grad_norm": 0.259765625, "learning_rate": 6.7168548724383105e-06, "loss": 0.9899, "step": 785 }, { "epoch": 0.986198243412798, "grad_norm": 0.263671875, "learning_rate": 6.712672521957341e-06, "loss": 0.9759, "step": 786 }, { "epoch": 0.9874529485570891, "grad_norm": 0.26171875, "learning_rate": 6.708490171476371e-06, "loss": 0.965, "step": 787 }, { "epoch": 0.9887076537013801, "grad_norm": 0.287109375, "learning_rate": 6.7043078209954e-06, "loss": 0.9651, "step": 788 }, { "epoch": 0.9899623588456713, "grad_norm": 0.255859375, "learning_rate": 6.700125470514429e-06, "loss": 0.9914, "step": 789 }, { "epoch": 0.9912170639899623, "grad_norm": 0.25390625, "learning_rate": 6.695943120033459e-06, "loss": 0.9674, "step": 790 }, { "epoch": 0.9924717691342535, "grad_norm": 0.263671875, "learning_rate": 6.691760769552489e-06, "loss": 0.9315, "step": 791 }, { "epoch": 0.9937264742785445, "grad_norm": 0.2431640625, "learning_rate": 6.687578419071519e-06, "loss": 0.9695, "step": 792 }, { "epoch": 0.9949811794228356, "grad_norm": 0.2421875, "learning_rate": 6.683396068590549e-06, "loss": 0.9753, "step": 793 }, { "epoch": 0.9962358845671268, "grad_norm": 0.306640625, "learning_rate": 6.6792137181095785e-06, "loss": 0.9729, "step": 794 }, { "epoch": 0.9974905897114178, "grad_norm": 0.302734375, "learning_rate": 6.675031367628608e-06, "loss": 0.9817, "step": 795 }, { "epoch": 0.998745294855709, "grad_norm": 0.2451171875, "learning_rate": 6.670849017147637e-06, "loss": 0.9345, "step": 796 }, { "epoch": 1.0, "grad_norm": 0.328125, "learning_rate": 6.666666666666667e-06, "loss": 0.9586, "step": 797 }, { "epoch": 1.001254705144291, "grad_norm": 0.2490234375, "learning_rate": 6.662484316185696e-06, "loss": 0.9758, "step": 798 }, { "epoch": 1.002509410288582, "grad_norm": 0.271484375, "learning_rate": 6.658301965704727e-06, "loss": 1.0353, "step": 799 }, { "epoch": 1.0037641154328734, "grad_norm": 0.251953125, "learning_rate": 6.6541196152237565e-06, "loss": 0.924, "step": 800 }, { "epoch": 1.0050188205771644, "grad_norm": 0.271484375, "learning_rate": 6.649937264742786e-06, "loss": 0.9406, "step": 801 }, { "epoch": 1.0062735257214555, "grad_norm": 0.255859375, "learning_rate": 6.645754914261816e-06, "loss": 0.93, "step": 802 }, { "epoch": 1.0075282308657465, "grad_norm": 0.2470703125, "learning_rate": 6.6415725637808455e-06, "loss": 0.938, "step": 803 }, { "epoch": 1.0087829360100375, "grad_norm": 0.28515625, "learning_rate": 6.637390213299874e-06, "loss": 1.0221, "step": 804 }, { "epoch": 1.0100376411543288, "grad_norm": 0.265625, "learning_rate": 6.633207862818905e-06, "loss": 0.9356, "step": 805 }, { "epoch": 1.0112923462986199, "grad_norm": 0.25390625, "learning_rate": 6.6290255123379346e-06, "loss": 0.9653, "step": 806 }, { "epoch": 1.012547051442911, "grad_norm": 0.26953125, "learning_rate": 6.624843161856964e-06, "loss": 0.966, "step": 807 }, { "epoch": 1.013801756587202, "grad_norm": 0.275390625, "learning_rate": 6.620660811375994e-06, "loss": 1.0125, "step": 808 }, { "epoch": 1.015056461731493, "grad_norm": 0.291015625, "learning_rate": 6.616478460895024e-06, "loss": 0.9374, "step": 809 }, { "epoch": 1.0163111668757843, "grad_norm": 0.27734375, "learning_rate": 6.612296110414054e-06, "loss": 0.9708, "step": 810 }, { "epoch": 1.0175658720200753, "grad_norm": 0.248046875, "learning_rate": 6.608113759933082e-06, "loss": 0.9595, "step": 811 }, { "epoch": 1.0188205771643664, "grad_norm": 0.2490234375, "learning_rate": 6.603931409452113e-06, "loss": 0.9731, "step": 812 }, { "epoch": 1.0200752823086574, "grad_norm": 0.244140625, "learning_rate": 6.599749058971142e-06, "loss": 0.9809, "step": 813 }, { "epoch": 1.0213299874529485, "grad_norm": 0.267578125, "learning_rate": 6.595566708490172e-06, "loss": 0.9495, "step": 814 }, { "epoch": 1.0225846925972397, "grad_norm": 0.267578125, "learning_rate": 6.591384358009202e-06, "loss": 0.9622, "step": 815 }, { "epoch": 1.0238393977415308, "grad_norm": 0.26171875, "learning_rate": 6.587202007528231e-06, "loss": 0.9726, "step": 816 }, { "epoch": 1.0250941028858218, "grad_norm": 0.2490234375, "learning_rate": 6.583019657047262e-06, "loss": 0.9942, "step": 817 }, { "epoch": 1.0263488080301129, "grad_norm": 0.328125, "learning_rate": 6.578837306566291e-06, "loss": 0.9851, "step": 818 }, { "epoch": 1.027603513174404, "grad_norm": 0.26953125, "learning_rate": 6.57465495608532e-06, "loss": 0.9933, "step": 819 }, { "epoch": 1.0288582183186952, "grad_norm": 0.26953125, "learning_rate": 6.57047260560435e-06, "loss": 0.9904, "step": 820 }, { "epoch": 1.0301129234629862, "grad_norm": 0.2490234375, "learning_rate": 6.56629025512338e-06, "loss": 0.9683, "step": 821 }, { "epoch": 1.0313676286072773, "grad_norm": 0.275390625, "learning_rate": 6.562107904642409e-06, "loss": 0.9661, "step": 822 }, { "epoch": 1.0326223337515683, "grad_norm": 0.251953125, "learning_rate": 6.55792555416144e-06, "loss": 0.9583, "step": 823 }, { "epoch": 1.0338770388958596, "grad_norm": 0.3359375, "learning_rate": 6.55374320368047e-06, "loss": 0.9291, "step": 824 }, { "epoch": 1.0351317440401506, "grad_norm": 0.259765625, "learning_rate": 6.549560853199499e-06, "loss": 0.9889, "step": 825 }, { "epoch": 1.0363864491844417, "grad_norm": 0.259765625, "learning_rate": 6.545378502718528e-06, "loss": 0.9373, "step": 826 }, { "epoch": 1.0376411543287327, "grad_norm": 0.251953125, "learning_rate": 6.541196152237558e-06, "loss": 0.9603, "step": 827 }, { "epoch": 1.0388958594730238, "grad_norm": 0.33984375, "learning_rate": 6.5370138017565874e-06, "loss": 0.9184, "step": 828 }, { "epoch": 1.040150564617315, "grad_norm": 0.255859375, "learning_rate": 6.532831451275617e-06, "loss": 0.971, "step": 829 }, { "epoch": 1.041405269761606, "grad_norm": 0.48828125, "learning_rate": 6.528649100794648e-06, "loss": 0.9286, "step": 830 }, { "epoch": 1.0426599749058971, "grad_norm": 0.2470703125, "learning_rate": 6.524466750313677e-06, "loss": 0.9541, "step": 831 }, { "epoch": 1.0439146800501882, "grad_norm": 0.34765625, "learning_rate": 6.520284399832707e-06, "loss": 1.0076, "step": 832 }, { "epoch": 1.0451693851944792, "grad_norm": 0.26953125, "learning_rate": 6.516102049351736e-06, "loss": 0.9637, "step": 833 }, { "epoch": 1.0464240903387705, "grad_norm": 0.259765625, "learning_rate": 6.5119196988707655e-06, "loss": 0.9519, "step": 834 }, { "epoch": 1.0476787954830615, "grad_norm": 0.400390625, "learning_rate": 6.507737348389795e-06, "loss": 0.9541, "step": 835 }, { "epoch": 1.0489335006273526, "grad_norm": 0.2578125, "learning_rate": 6.503554997908826e-06, "loss": 0.9587, "step": 836 }, { "epoch": 1.0501882057716436, "grad_norm": 0.263671875, "learning_rate": 6.499372647427855e-06, "loss": 0.9725, "step": 837 }, { "epoch": 1.0514429109159347, "grad_norm": 0.255859375, "learning_rate": 6.495190296946885e-06, "loss": 0.9499, "step": 838 }, { "epoch": 1.052697616060226, "grad_norm": 0.314453125, "learning_rate": 6.491007946465915e-06, "loss": 1.0076, "step": 839 }, { "epoch": 1.053952321204517, "grad_norm": 0.248046875, "learning_rate": 6.4868255959849435e-06, "loss": 0.9338, "step": 840 }, { "epoch": 1.055207026348808, "grad_norm": 0.2578125, "learning_rate": 6.482643245503973e-06, "loss": 0.9176, "step": 841 }, { "epoch": 1.056461731493099, "grad_norm": 0.28125, "learning_rate": 6.478460895023003e-06, "loss": 0.9645, "step": 842 }, { "epoch": 1.0577164366373901, "grad_norm": 0.25390625, "learning_rate": 6.474278544542033e-06, "loss": 0.9087, "step": 843 }, { "epoch": 1.0589711417816814, "grad_norm": 0.251953125, "learning_rate": 6.470096194061063e-06, "loss": 0.9699, "step": 844 }, { "epoch": 1.0602258469259724, "grad_norm": 0.271484375, "learning_rate": 6.465913843580093e-06, "loss": 1.0, "step": 845 }, { "epoch": 1.0614805520702635, "grad_norm": 0.26171875, "learning_rate": 6.4617314930991224e-06, "loss": 0.9292, "step": 846 }, { "epoch": 1.0627352572145545, "grad_norm": 0.263671875, "learning_rate": 6.457549142618152e-06, "loss": 0.9312, "step": 847 }, { "epoch": 1.0639899623588456, "grad_norm": 0.306640625, "learning_rate": 6.453366792137181e-06, "loss": 0.9646, "step": 848 }, { "epoch": 1.0652446675031368, "grad_norm": 0.2734375, "learning_rate": 6.4491844416562115e-06, "loss": 0.9867, "step": 849 }, { "epoch": 1.066499372647428, "grad_norm": 0.25, "learning_rate": 6.445002091175241e-06, "loss": 0.9259, "step": 850 }, { "epoch": 1.067754077791719, "grad_norm": 0.255859375, "learning_rate": 6.440819740694271e-06, "loss": 0.9714, "step": 851 }, { "epoch": 1.06900878293601, "grad_norm": 0.2734375, "learning_rate": 6.4366373902133005e-06, "loss": 0.9463, "step": 852 }, { "epoch": 1.070263488080301, "grad_norm": 0.279296875, "learning_rate": 6.43245503973233e-06, "loss": 0.9388, "step": 853 }, { "epoch": 1.0715181932245923, "grad_norm": 0.26953125, "learning_rate": 6.428272689251361e-06, "loss": 0.9989, "step": 854 }, { "epoch": 1.0727728983688833, "grad_norm": 0.2470703125, "learning_rate": 6.424090338770389e-06, "loss": 0.9238, "step": 855 }, { "epoch": 1.0740276035131744, "grad_norm": 0.265625, "learning_rate": 6.419907988289419e-06, "loss": 0.9903, "step": 856 }, { "epoch": 1.0752823086574654, "grad_norm": 0.2890625, "learning_rate": 6.415725637808449e-06, "loss": 0.9916, "step": 857 }, { "epoch": 1.0765370138017567, "grad_norm": 0.244140625, "learning_rate": 6.4115432873274786e-06, "loss": 0.9268, "step": 858 }, { "epoch": 1.0777917189460477, "grad_norm": 0.263671875, "learning_rate": 6.407360936846508e-06, "loss": 0.9599, "step": 859 }, { "epoch": 1.0790464240903388, "grad_norm": 0.259765625, "learning_rate": 6.403178586365538e-06, "loss": 0.9606, "step": 860 }, { "epoch": 1.0803011292346298, "grad_norm": 0.2578125, "learning_rate": 6.3989962358845684e-06, "loss": 0.9584, "step": 861 }, { "epoch": 1.0815558343789209, "grad_norm": 0.265625, "learning_rate": 6.394813885403596e-06, "loss": 0.9961, "step": 862 }, { "epoch": 1.0828105395232122, "grad_norm": 0.25, "learning_rate": 6.390631534922627e-06, "loss": 0.9544, "step": 863 }, { "epoch": 1.0840652446675032, "grad_norm": 0.2578125, "learning_rate": 6.386449184441657e-06, "loss": 0.9822, "step": 864 }, { "epoch": 1.0853199498117942, "grad_norm": 0.251953125, "learning_rate": 6.382266833960686e-06, "loss": 0.9537, "step": 865 }, { "epoch": 1.0865746549560853, "grad_norm": 0.2734375, "learning_rate": 6.378084483479716e-06, "loss": 1.004, "step": 866 }, { "epoch": 1.0878293601003763, "grad_norm": 0.28125, "learning_rate": 6.373902132998746e-06, "loss": 0.9317, "step": 867 }, { "epoch": 1.0890840652446676, "grad_norm": 0.26171875, "learning_rate": 6.369719782517776e-06, "loss": 0.9808, "step": 868 }, { "epoch": 1.0903387703889587, "grad_norm": 0.267578125, "learning_rate": 6.365537432036806e-06, "loss": 1.0081, "step": 869 }, { "epoch": 1.0915934755332497, "grad_norm": 0.2490234375, "learning_rate": 6.361355081555835e-06, "loss": 0.9104, "step": 870 }, { "epoch": 1.0928481806775407, "grad_norm": 0.26171875, "learning_rate": 6.357172731074864e-06, "loss": 0.978, "step": 871 }, { "epoch": 1.0941028858218318, "grad_norm": 0.251953125, "learning_rate": 6.352990380593894e-06, "loss": 0.9666, "step": 872 }, { "epoch": 1.095357590966123, "grad_norm": 0.271484375, "learning_rate": 6.348808030112924e-06, "loss": 0.9802, "step": 873 }, { "epoch": 1.096612296110414, "grad_norm": 0.263671875, "learning_rate": 6.344625679631954e-06, "loss": 0.9693, "step": 874 }, { "epoch": 1.0978670012547052, "grad_norm": 0.259765625, "learning_rate": 6.340443329150984e-06, "loss": 0.987, "step": 875 }, { "epoch": 1.0991217063989962, "grad_norm": 0.263671875, "learning_rate": 6.3362609786700136e-06, "loss": 0.919, "step": 876 }, { "epoch": 1.1003764115432872, "grad_norm": 0.26171875, "learning_rate": 6.332078628189042e-06, "loss": 1.0096, "step": 877 }, { "epoch": 1.1016311166875785, "grad_norm": 0.2431640625, "learning_rate": 6.327896277708072e-06, "loss": 0.9614, "step": 878 }, { "epoch": 1.1028858218318696, "grad_norm": 0.271484375, "learning_rate": 6.323713927227102e-06, "loss": 0.9754, "step": 879 }, { "epoch": 1.1041405269761606, "grad_norm": 0.2578125, "learning_rate": 6.319531576746131e-06, "loss": 0.9655, "step": 880 }, { "epoch": 1.1053952321204517, "grad_norm": 0.267578125, "learning_rate": 6.315349226265162e-06, "loss": 1.0026, "step": 881 }, { "epoch": 1.1066499372647427, "grad_norm": 0.251953125, "learning_rate": 6.311166875784192e-06, "loss": 0.9265, "step": 882 }, { "epoch": 1.107904642409034, "grad_norm": 0.267578125, "learning_rate": 6.306984525303221e-06, "loss": 0.9752, "step": 883 }, { "epoch": 1.109159347553325, "grad_norm": 0.275390625, "learning_rate": 6.30280217482225e-06, "loss": 0.9387, "step": 884 }, { "epoch": 1.110414052697616, "grad_norm": 0.259765625, "learning_rate": 6.29861982434128e-06, "loss": 0.9027, "step": 885 }, { "epoch": 1.111668757841907, "grad_norm": 0.271484375, "learning_rate": 6.2944374738603095e-06, "loss": 0.9745, "step": 886 }, { "epoch": 1.1129234629861982, "grad_norm": 0.25, "learning_rate": 6.29025512337934e-06, "loss": 0.9873, "step": 887 }, { "epoch": 1.1141781681304894, "grad_norm": 0.271484375, "learning_rate": 6.28607277289837e-06, "loss": 0.9567, "step": 888 }, { "epoch": 1.1154328732747805, "grad_norm": 0.275390625, "learning_rate": 6.281890422417399e-06, "loss": 0.9988, "step": 889 }, { "epoch": 1.1166875784190715, "grad_norm": 0.26171875, "learning_rate": 6.277708071936429e-06, "loss": 0.9405, "step": 890 }, { "epoch": 1.1179422835633626, "grad_norm": 0.255859375, "learning_rate": 6.273525721455459e-06, "loss": 0.9605, "step": 891 }, { "epoch": 1.1191969887076536, "grad_norm": 0.251953125, "learning_rate": 6.2693433709744875e-06, "loss": 0.9346, "step": 892 }, { "epoch": 1.1204516938519449, "grad_norm": 0.251953125, "learning_rate": 6.265161020493517e-06, "loss": 0.968, "step": 893 }, { "epoch": 1.121706398996236, "grad_norm": 0.283203125, "learning_rate": 6.260978670012548e-06, "loss": 0.9778, "step": 894 }, { "epoch": 1.122961104140527, "grad_norm": 0.287109375, "learning_rate": 6.256796319531577e-06, "loss": 0.9561, "step": 895 }, { "epoch": 1.124215809284818, "grad_norm": 0.30859375, "learning_rate": 6.252613969050607e-06, "loss": 0.9241, "step": 896 }, { "epoch": 1.125470514429109, "grad_norm": 0.26953125, "learning_rate": 6.248431618569637e-06, "loss": 0.9595, "step": 897 }, { "epoch": 1.1267252195734003, "grad_norm": 0.29296875, "learning_rate": 6.244249268088666e-06, "loss": 0.9282, "step": 898 }, { "epoch": 1.1279799247176914, "grad_norm": 0.271484375, "learning_rate": 6.240066917607695e-06, "loss": 0.9568, "step": 899 }, { "epoch": 1.1292346298619824, "grad_norm": 0.279296875, "learning_rate": 6.235884567126726e-06, "loss": 0.9257, "step": 900 }, { "epoch": 1.1304893350062735, "grad_norm": 0.287109375, "learning_rate": 6.2317022166457554e-06, "loss": 0.9664, "step": 901 }, { "epoch": 1.1317440401505645, "grad_norm": 0.26953125, "learning_rate": 6.227519866164785e-06, "loss": 0.9586, "step": 902 }, { "epoch": 1.1329987452948558, "grad_norm": 0.28125, "learning_rate": 6.223337515683815e-06, "loss": 1.0192, "step": 903 }, { "epoch": 1.1342534504391468, "grad_norm": 0.2734375, "learning_rate": 6.2191551652028445e-06, "loss": 0.9509, "step": 904 }, { "epoch": 1.1355081555834379, "grad_norm": 0.390625, "learning_rate": 6.214972814721875e-06, "loss": 0.9536, "step": 905 }, { "epoch": 1.136762860727729, "grad_norm": 0.26171875, "learning_rate": 6.210790464240903e-06, "loss": 0.9397, "step": 906 }, { "epoch": 1.13801756587202, "grad_norm": 0.42578125, "learning_rate": 6.2066081137599335e-06, "loss": 0.9344, "step": 907 }, { "epoch": 1.1392722710163112, "grad_norm": 0.267578125, "learning_rate": 6.202425763278963e-06, "loss": 0.9806, "step": 908 }, { "epoch": 1.1405269761606023, "grad_norm": 0.259765625, "learning_rate": 6.198243412797993e-06, "loss": 0.9652, "step": 909 }, { "epoch": 1.1417816813048933, "grad_norm": 0.251953125, "learning_rate": 6.1940610623170225e-06, "loss": 0.9737, "step": 910 }, { "epoch": 1.1430363864491844, "grad_norm": 0.265625, "learning_rate": 6.189878711836052e-06, "loss": 0.94, "step": 911 }, { "epoch": 1.1442910915934754, "grad_norm": 0.3515625, "learning_rate": 6.185696361355083e-06, "loss": 0.9272, "step": 912 }, { "epoch": 1.1455457967377667, "grad_norm": 0.26953125, "learning_rate": 6.181514010874112e-06, "loss": 0.9257, "step": 913 }, { "epoch": 1.1468005018820577, "grad_norm": 0.26171875, "learning_rate": 6.177331660393141e-06, "loss": 0.9781, "step": 914 }, { "epoch": 1.1480552070263488, "grad_norm": 0.341796875, "learning_rate": 6.173149309912171e-06, "loss": 0.9776, "step": 915 }, { "epoch": 1.1493099121706398, "grad_norm": 0.283203125, "learning_rate": 6.168966959431201e-06, "loss": 0.9598, "step": 916 }, { "epoch": 1.1505646173149309, "grad_norm": 0.306640625, "learning_rate": 6.16478460895023e-06, "loss": 0.943, "step": 917 }, { "epoch": 1.1518193224592221, "grad_norm": 0.255859375, "learning_rate": 6.160602258469261e-06, "loss": 0.9462, "step": 918 }, { "epoch": 1.1530740276035132, "grad_norm": 0.353515625, "learning_rate": 6.1564199079882905e-06, "loss": 0.9702, "step": 919 }, { "epoch": 1.1543287327478042, "grad_norm": 0.26953125, "learning_rate": 6.15223755750732e-06, "loss": 0.9405, "step": 920 }, { "epoch": 1.1555834378920953, "grad_norm": 0.2578125, "learning_rate": 6.148055207026349e-06, "loss": 0.9466, "step": 921 }, { "epoch": 1.1568381430363865, "grad_norm": 0.25390625, "learning_rate": 6.143872856545379e-06, "loss": 0.9899, "step": 922 }, { "epoch": 1.1580928481806776, "grad_norm": 0.29296875, "learning_rate": 6.139690506064408e-06, "loss": 0.9675, "step": 923 }, { "epoch": 1.1593475533249686, "grad_norm": 0.26953125, "learning_rate": 6.135508155583438e-06, "loss": 0.9322, "step": 924 }, { "epoch": 1.1606022584692597, "grad_norm": 0.267578125, "learning_rate": 6.1313258051024685e-06, "loss": 0.9686, "step": 925 }, { "epoch": 1.1618569636135507, "grad_norm": 0.28515625, "learning_rate": 6.127143454621498e-06, "loss": 1.0023, "step": 926 }, { "epoch": 1.163111668757842, "grad_norm": 0.263671875, "learning_rate": 6.122961104140528e-06, "loss": 0.9435, "step": 927 }, { "epoch": 1.164366373902133, "grad_norm": 0.265625, "learning_rate": 6.118778753659557e-06, "loss": 0.9655, "step": 928 }, { "epoch": 1.165621079046424, "grad_norm": 0.31640625, "learning_rate": 6.114596403178586e-06, "loss": 0.9688, "step": 929 }, { "epoch": 1.1668757841907151, "grad_norm": 0.26953125, "learning_rate": 6.110414052697616e-06, "loss": 0.9564, "step": 930 }, { "epoch": 1.1681304893350062, "grad_norm": 0.28125, "learning_rate": 6.1062317022166466e-06, "loss": 0.9775, "step": 931 }, { "epoch": 1.1693851944792975, "grad_norm": 0.283203125, "learning_rate": 6.102049351735676e-06, "loss": 0.9824, "step": 932 }, { "epoch": 1.1706398996235885, "grad_norm": 0.26953125, "learning_rate": 6.097867001254706e-06, "loss": 0.9666, "step": 933 }, { "epoch": 1.1718946047678795, "grad_norm": 0.267578125, "learning_rate": 6.093684650773736e-06, "loss": 0.9655, "step": 934 }, { "epoch": 1.1731493099121706, "grad_norm": 0.26171875, "learning_rate": 6.089502300292765e-06, "loss": 0.9811, "step": 935 }, { "epoch": 1.1744040150564616, "grad_norm": 0.28515625, "learning_rate": 6.085319949811794e-06, "loss": 0.9522, "step": 936 }, { "epoch": 1.175658720200753, "grad_norm": 0.365234375, "learning_rate": 6.081137599330824e-06, "loss": 1.0026, "step": 937 }, { "epoch": 1.176913425345044, "grad_norm": 0.263671875, "learning_rate": 6.076955248849854e-06, "loss": 0.9679, "step": 938 }, { "epoch": 1.178168130489335, "grad_norm": 0.34375, "learning_rate": 6.072772898368884e-06, "loss": 0.9308, "step": 939 }, { "epoch": 1.179422835633626, "grad_norm": 0.265625, "learning_rate": 6.068590547887914e-06, "loss": 0.9915, "step": 940 }, { "epoch": 1.1806775407779173, "grad_norm": 0.259765625, "learning_rate": 6.064408197406943e-06, "loss": 0.9608, "step": 941 }, { "epoch": 1.1819322459222084, "grad_norm": 0.255859375, "learning_rate": 6.060225846925973e-06, "loss": 0.9461, "step": 942 }, { "epoch": 1.1831869510664994, "grad_norm": 0.2890625, "learning_rate": 6.056043496445002e-06, "loss": 0.945, "step": 943 }, { "epoch": 1.1844416562107905, "grad_norm": 0.271484375, "learning_rate": 6.051861145964032e-06, "loss": 0.9687, "step": 944 }, { "epoch": 1.1856963613550815, "grad_norm": 0.3125, "learning_rate": 6.047678795483062e-06, "loss": 1.0105, "step": 945 }, { "epoch": 1.1869510664993728, "grad_norm": 0.267578125, "learning_rate": 6.043496445002092e-06, "loss": 0.9839, "step": 946 }, { "epoch": 1.1882057716436638, "grad_norm": 0.279296875, "learning_rate": 6.039314094521121e-06, "loss": 0.9121, "step": 947 }, { "epoch": 1.1894604767879549, "grad_norm": 0.275390625, "learning_rate": 6.035131744040151e-06, "loss": 0.9587, "step": 948 }, { "epoch": 1.190715181932246, "grad_norm": 0.28125, "learning_rate": 6.0309493935591816e-06, "loss": 0.939, "step": 949 }, { "epoch": 1.191969887076537, "grad_norm": 0.267578125, "learning_rate": 6.0267670430782096e-06, "loss": 0.9742, "step": 950 }, { "epoch": 1.1932245922208282, "grad_norm": 0.2578125, "learning_rate": 6.02258469259724e-06, "loss": 0.9172, "step": 951 }, { "epoch": 1.1944792973651193, "grad_norm": 0.263671875, "learning_rate": 6.01840234211627e-06, "loss": 0.9679, "step": 952 }, { "epoch": 1.1957340025094103, "grad_norm": 0.267578125, "learning_rate": 6.0142199916352994e-06, "loss": 0.9417, "step": 953 }, { "epoch": 1.1969887076537014, "grad_norm": 0.263671875, "learning_rate": 6.010037641154329e-06, "loss": 0.9738, "step": 954 }, { "epoch": 1.1982434127979924, "grad_norm": 0.322265625, "learning_rate": 6.005855290673359e-06, "loss": 0.9289, "step": 955 }, { "epoch": 1.1994981179422837, "grad_norm": 0.2578125, "learning_rate": 6.001672940192389e-06, "loss": 0.9471, "step": 956 }, { "epoch": 1.2007528230865747, "grad_norm": 0.255859375, "learning_rate": 5.997490589711419e-06, "loss": 0.968, "step": 957 }, { "epoch": 1.2020075282308658, "grad_norm": 0.2734375, "learning_rate": 5.993308239230448e-06, "loss": 0.9827, "step": 958 }, { "epoch": 1.2032622333751568, "grad_norm": 0.26953125, "learning_rate": 5.9891258887494775e-06, "loss": 0.9882, "step": 959 }, { "epoch": 1.2045169385194479, "grad_norm": 0.263671875, "learning_rate": 5.984943538268507e-06, "loss": 0.9564, "step": 960 }, { "epoch": 1.2057716436637391, "grad_norm": 0.26171875, "learning_rate": 5.980761187787537e-06, "loss": 0.9196, "step": 961 }, { "epoch": 1.2070263488080302, "grad_norm": 0.3359375, "learning_rate": 5.976578837306567e-06, "loss": 0.9519, "step": 962 }, { "epoch": 1.2082810539523212, "grad_norm": 0.265625, "learning_rate": 5.972396486825597e-06, "loss": 0.964, "step": 963 }, { "epoch": 1.2095357590966123, "grad_norm": 0.271484375, "learning_rate": 5.968214136344627e-06, "loss": 0.9225, "step": 964 }, { "epoch": 1.2107904642409033, "grad_norm": 0.26171875, "learning_rate": 5.9640317858636555e-06, "loss": 0.9475, "step": 965 }, { "epoch": 1.2120451693851946, "grad_norm": 0.263671875, "learning_rate": 5.959849435382685e-06, "loss": 0.9348, "step": 966 }, { "epoch": 1.2132998745294856, "grad_norm": 0.265625, "learning_rate": 5.955667084901715e-06, "loss": 0.9484, "step": 967 }, { "epoch": 1.2145545796737767, "grad_norm": 0.267578125, "learning_rate": 5.9514847344207446e-06, "loss": 0.9328, "step": 968 }, { "epoch": 1.2158092848180677, "grad_norm": 0.2734375, "learning_rate": 5.947302383939775e-06, "loss": 0.9723, "step": 969 }, { "epoch": 1.2170639899623588, "grad_norm": 0.259765625, "learning_rate": 5.943120033458805e-06, "loss": 0.9172, "step": 970 }, { "epoch": 1.21831869510665, "grad_norm": 0.3125, "learning_rate": 5.9389376829778344e-06, "loss": 0.9041, "step": 971 }, { "epoch": 1.219573400250941, "grad_norm": 0.271484375, "learning_rate": 5.934755332496863e-06, "loss": 0.917, "step": 972 }, { "epoch": 1.2208281053952321, "grad_norm": 0.296875, "learning_rate": 5.930572982015893e-06, "loss": 1.0325, "step": 973 }, { "epoch": 1.2220828105395232, "grad_norm": 0.3125, "learning_rate": 5.926390631534923e-06, "loss": 0.9772, "step": 974 }, { "epoch": 1.2233375156838142, "grad_norm": 0.251953125, "learning_rate": 5.922208281053953e-06, "loss": 0.9308, "step": 975 }, { "epoch": 1.2245922208281055, "grad_norm": 0.353515625, "learning_rate": 5.918025930572983e-06, "loss": 0.9738, "step": 976 }, { "epoch": 1.2258469259723965, "grad_norm": 0.26171875, "learning_rate": 5.9138435800920125e-06, "loss": 0.9397, "step": 977 }, { "epoch": 1.2271016311166876, "grad_norm": 0.5234375, "learning_rate": 5.909661229611042e-06, "loss": 0.9594, "step": 978 }, { "epoch": 1.2283563362609786, "grad_norm": 0.263671875, "learning_rate": 5.905478879130072e-06, "loss": 0.9205, "step": 979 }, { "epoch": 1.2296110414052697, "grad_norm": 0.267578125, "learning_rate": 5.901296528649101e-06, "loss": 1.0001, "step": 980 }, { "epoch": 1.230865746549561, "grad_norm": 0.259765625, "learning_rate": 5.89711417816813e-06, "loss": 0.9309, "step": 981 }, { "epoch": 1.232120451693852, "grad_norm": 0.255859375, "learning_rate": 5.892931827687161e-06, "loss": 0.9438, "step": 982 }, { "epoch": 1.233375156838143, "grad_norm": 0.271484375, "learning_rate": 5.8887494772061905e-06, "loss": 0.9447, "step": 983 }, { "epoch": 1.234629861982434, "grad_norm": 0.27734375, "learning_rate": 5.88456712672522e-06, "loss": 0.933, "step": 984 }, { "epoch": 1.2358845671267251, "grad_norm": 0.296875, "learning_rate": 5.88038477624425e-06, "loss": 0.9966, "step": 985 }, { "epoch": 1.2371392722710164, "grad_norm": 0.27734375, "learning_rate": 5.8762024257632796e-06, "loss": 0.9072, "step": 986 }, { "epoch": 1.2383939774153074, "grad_norm": 0.28125, "learning_rate": 5.872020075282308e-06, "loss": 0.9398, "step": 987 }, { "epoch": 1.2396486825595985, "grad_norm": 0.349609375, "learning_rate": 5.867837724801339e-06, "loss": 0.9175, "step": 988 }, { "epoch": 1.2409033877038895, "grad_norm": 0.29296875, "learning_rate": 5.863655374320369e-06, "loss": 0.9474, "step": 989 }, { "epoch": 1.2421580928481806, "grad_norm": 0.251953125, "learning_rate": 5.859473023839398e-06, "loss": 0.9169, "step": 990 }, { "epoch": 1.2434127979924718, "grad_norm": 0.2578125, "learning_rate": 5.855290673358428e-06, "loss": 0.8959, "step": 991 }, { "epoch": 1.2446675031367629, "grad_norm": 0.29296875, "learning_rate": 5.851108322877458e-06, "loss": 0.9651, "step": 992 }, { "epoch": 1.245922208281054, "grad_norm": 0.2578125, "learning_rate": 5.846925972396488e-06, "loss": 0.9187, "step": 993 }, { "epoch": 1.247176913425345, "grad_norm": 0.259765625, "learning_rate": 5.842743621915516e-06, "loss": 0.8922, "step": 994 }, { "epoch": 1.248431618569636, "grad_norm": 0.259765625, "learning_rate": 5.838561271434547e-06, "loss": 0.9347, "step": 995 }, { "epoch": 1.2496863237139273, "grad_norm": 0.2890625, "learning_rate": 5.834378920953576e-06, "loss": 1.0081, "step": 996 }, { "epoch": 1.2509410288582183, "grad_norm": 0.263671875, "learning_rate": 5.830196570472606e-06, "loss": 0.9368, "step": 997 }, { "epoch": 1.2521957340025094, "grad_norm": 0.29296875, "learning_rate": 5.826014219991636e-06, "loss": 0.9596, "step": 998 }, { "epoch": 1.2534504391468004, "grad_norm": 0.298828125, "learning_rate": 5.821831869510665e-06, "loss": 0.9642, "step": 999 }, { "epoch": 1.2547051442910915, "grad_norm": 0.37109375, "learning_rate": 5.817649519029696e-06, "loss": 0.9273, "step": 1000 }, { "epoch": 1.2559598494353827, "grad_norm": 0.251953125, "learning_rate": 5.8134671685487255e-06, "loss": 0.9305, "step": 1001 }, { "epoch": 1.2572145545796738, "grad_norm": 0.26171875, "learning_rate": 5.809284818067754e-06, "loss": 0.9637, "step": 1002 }, { "epoch": 1.2584692597239648, "grad_norm": 0.296875, "learning_rate": 5.805102467586784e-06, "loss": 0.8979, "step": 1003 }, { "epoch": 1.2597239648682559, "grad_norm": 0.263671875, "learning_rate": 5.800920117105814e-06, "loss": 0.9348, "step": 1004 }, { "epoch": 1.260978670012547, "grad_norm": 0.255859375, "learning_rate": 5.796737766624843e-06, "loss": 0.9299, "step": 1005 }, { "epoch": 1.2622333751568382, "grad_norm": 0.3046875, "learning_rate": 5.792555416143874e-06, "loss": 0.9425, "step": 1006 }, { "epoch": 1.2634880803011292, "grad_norm": 0.267578125, "learning_rate": 5.788373065662904e-06, "loss": 0.9076, "step": 1007 }, { "epoch": 1.2647427854454203, "grad_norm": 0.2734375, "learning_rate": 5.784190715181933e-06, "loss": 0.9531, "step": 1008 }, { "epoch": 1.2659974905897113, "grad_norm": 0.30078125, "learning_rate": 5.780008364700962e-06, "loss": 0.9222, "step": 1009 }, { "epoch": 1.2672521957340024, "grad_norm": 0.263671875, "learning_rate": 5.775826014219992e-06, "loss": 0.9394, "step": 1010 }, { "epoch": 1.2685069008782937, "grad_norm": 0.28515625, "learning_rate": 5.7716436637390215e-06, "loss": 0.9772, "step": 1011 }, { "epoch": 1.2697616060225847, "grad_norm": 0.267578125, "learning_rate": 5.767461313258051e-06, "loss": 0.9609, "step": 1012 }, { "epoch": 1.2710163111668757, "grad_norm": 0.259765625, "learning_rate": 5.763278962777082e-06, "loss": 0.9245, "step": 1013 }, { "epoch": 1.272271016311167, "grad_norm": 0.279296875, "learning_rate": 5.759096612296111e-06, "loss": 1.0299, "step": 1014 }, { "epoch": 1.2735257214554578, "grad_norm": 0.287109375, "learning_rate": 5.754914261815141e-06, "loss": 0.9677, "step": 1015 }, { "epoch": 1.274780426599749, "grad_norm": 0.265625, "learning_rate": 5.75073191133417e-06, "loss": 0.955, "step": 1016 }, { "epoch": 1.2760351317440402, "grad_norm": 0.26953125, "learning_rate": 5.7465495608531995e-06, "loss": 0.9218, "step": 1017 }, { "epoch": 1.2772898368883312, "grad_norm": 0.275390625, "learning_rate": 5.742367210372229e-06, "loss": 0.967, "step": 1018 }, { "epoch": 1.2785445420326225, "grad_norm": 0.3125, "learning_rate": 5.73818485989126e-06, "loss": 0.9605, "step": 1019 }, { "epoch": 1.2797992471769133, "grad_norm": 0.28125, "learning_rate": 5.734002509410289e-06, "loss": 0.9734, "step": 1020 }, { "epoch": 1.2810539523212046, "grad_norm": 0.271484375, "learning_rate": 5.729820158929319e-06, "loss": 0.9443, "step": 1021 }, { "epoch": 1.2823086574654956, "grad_norm": 0.265625, "learning_rate": 5.725637808448349e-06, "loss": 0.9433, "step": 1022 }, { "epoch": 1.2835633626097867, "grad_norm": 0.27734375, "learning_rate": 5.721455457967378e-06, "loss": 0.9543, "step": 1023 }, { "epoch": 1.284818067754078, "grad_norm": 0.279296875, "learning_rate": 5.717273107486407e-06, "loss": 0.9451, "step": 1024 }, { "epoch": 1.286072772898369, "grad_norm": 0.28515625, "learning_rate": 5.713090757005437e-06, "loss": 1.0009, "step": 1025 }, { "epoch": 1.28732747804266, "grad_norm": 0.259765625, "learning_rate": 5.7089084065244674e-06, "loss": 0.9438, "step": 1026 }, { "epoch": 1.288582183186951, "grad_norm": 0.30859375, "learning_rate": 5.704726056043497e-06, "loss": 0.9334, "step": 1027 }, { "epoch": 1.289836888331242, "grad_norm": 0.265625, "learning_rate": 5.700543705562527e-06, "loss": 0.9435, "step": 1028 }, { "epoch": 1.2910915934755334, "grad_norm": 0.26953125, "learning_rate": 5.6963613550815565e-06, "loss": 0.955, "step": 1029 }, { "epoch": 1.2923462986198244, "grad_norm": 0.26953125, "learning_rate": 5.692179004600586e-06, "loss": 0.9598, "step": 1030 }, { "epoch": 1.2936010037641155, "grad_norm": 0.263671875, "learning_rate": 5.687996654119615e-06, "loss": 0.9603, "step": 1031 }, { "epoch": 1.2948557089084065, "grad_norm": 0.314453125, "learning_rate": 5.683814303638645e-06, "loss": 0.8706, "step": 1032 }, { "epoch": 1.2961104140526976, "grad_norm": 0.265625, "learning_rate": 5.679631953157675e-06, "loss": 0.972, "step": 1033 }, { "epoch": 1.2973651191969888, "grad_norm": 0.271484375, "learning_rate": 5.675449602676705e-06, "loss": 0.9304, "step": 1034 }, { "epoch": 1.2986198243412799, "grad_norm": 0.275390625, "learning_rate": 5.6712672521957345e-06, "loss": 0.9677, "step": 1035 }, { "epoch": 1.299874529485571, "grad_norm": 0.259765625, "learning_rate": 5.667084901714764e-06, "loss": 0.9314, "step": 1036 }, { "epoch": 1.301129234629862, "grad_norm": 0.283203125, "learning_rate": 5.662902551233794e-06, "loss": 0.9489, "step": 1037 }, { "epoch": 1.302383939774153, "grad_norm": 0.29296875, "learning_rate": 5.658720200752823e-06, "loss": 0.9176, "step": 1038 }, { "epoch": 1.3036386449184443, "grad_norm": 0.271484375, "learning_rate": 5.654537850271853e-06, "loss": 0.926, "step": 1039 }, { "epoch": 1.3048933500627353, "grad_norm": 0.2578125, "learning_rate": 5.650355499790883e-06, "loss": 0.9537, "step": 1040 }, { "epoch": 1.3061480552070264, "grad_norm": 0.33984375, "learning_rate": 5.6461731493099126e-06, "loss": 0.9543, "step": 1041 }, { "epoch": 1.3074027603513174, "grad_norm": 0.27734375, "learning_rate": 5.641990798828942e-06, "loss": 0.9602, "step": 1042 }, { "epoch": 1.3086574654956085, "grad_norm": 0.267578125, "learning_rate": 5.637808448347972e-06, "loss": 0.92, "step": 1043 }, { "epoch": 1.3099121706398997, "grad_norm": 0.259765625, "learning_rate": 5.6336260978670024e-06, "loss": 0.9533, "step": 1044 }, { "epoch": 1.3111668757841908, "grad_norm": 0.2734375, "learning_rate": 5.629443747386032e-06, "loss": 0.9007, "step": 1045 }, { "epoch": 1.3124215809284818, "grad_norm": 0.2578125, "learning_rate": 5.625261396905061e-06, "loss": 0.9483, "step": 1046 }, { "epoch": 1.3136762860727729, "grad_norm": 0.265625, "learning_rate": 5.621079046424091e-06, "loss": 0.9323, "step": 1047 }, { "epoch": 1.314930991217064, "grad_norm": 0.25390625, "learning_rate": 5.61689669594312e-06, "loss": 0.9476, "step": 1048 }, { "epoch": 1.3161856963613552, "grad_norm": 0.267578125, "learning_rate": 5.61271434546215e-06, "loss": 0.9715, "step": 1049 }, { "epoch": 1.3174404015056462, "grad_norm": 0.2734375, "learning_rate": 5.60853199498118e-06, "loss": 0.9206, "step": 1050 }, { "epoch": 1.3186951066499373, "grad_norm": 0.259765625, "learning_rate": 5.60434964450021e-06, "loss": 0.9493, "step": 1051 }, { "epoch": 1.3199498117942283, "grad_norm": 0.28125, "learning_rate": 5.60016729401924e-06, "loss": 1.0005, "step": 1052 }, { "epoch": 1.3212045169385194, "grad_norm": 0.275390625, "learning_rate": 5.595984943538269e-06, "loss": 0.9554, "step": 1053 }, { "epoch": 1.3224592220828106, "grad_norm": 0.283203125, "learning_rate": 5.591802593057298e-06, "loss": 0.9394, "step": 1054 }, { "epoch": 1.3237139272271017, "grad_norm": 0.27734375, "learning_rate": 5.587620242576328e-06, "loss": 0.9073, "step": 1055 }, { "epoch": 1.3249686323713927, "grad_norm": 0.404296875, "learning_rate": 5.583437892095358e-06, "loss": 1.0202, "step": 1056 }, { "epoch": 1.3262233375156838, "grad_norm": 0.263671875, "learning_rate": 5.579255541614388e-06, "loss": 0.935, "step": 1057 }, { "epoch": 1.3274780426599748, "grad_norm": 0.275390625, "learning_rate": 5.575073191133418e-06, "loss": 1.0099, "step": 1058 }, { "epoch": 1.328732747804266, "grad_norm": 0.306640625, "learning_rate": 5.5708908406524476e-06, "loss": 1.0049, "step": 1059 }, { "epoch": 1.3299874529485571, "grad_norm": 0.294921875, "learning_rate": 5.566708490171476e-06, "loss": 0.9454, "step": 1060 }, { "epoch": 1.3312421580928482, "grad_norm": 0.3046875, "learning_rate": 5.562526139690506e-06, "loss": 0.9347, "step": 1061 }, { "epoch": 1.3324968632371392, "grad_norm": 0.28515625, "learning_rate": 5.558343789209536e-06, "loss": 0.9024, "step": 1062 }, { "epoch": 1.3337515683814303, "grad_norm": 0.259765625, "learning_rate": 5.5541614387285654e-06, "loss": 0.9527, "step": 1063 }, { "epoch": 1.3350062735257215, "grad_norm": 0.2734375, "learning_rate": 5.549979088247596e-06, "loss": 0.9513, "step": 1064 }, { "epoch": 1.3362609786700126, "grad_norm": 0.26171875, "learning_rate": 5.545796737766626e-06, "loss": 0.9088, "step": 1065 }, { "epoch": 1.3375156838143036, "grad_norm": 0.267578125, "learning_rate": 5.541614387285655e-06, "loss": 0.9458, "step": 1066 }, { "epoch": 1.3387703889585947, "grad_norm": 0.27734375, "learning_rate": 5.537432036804685e-06, "loss": 0.9752, "step": 1067 }, { "epoch": 1.3400250941028857, "grad_norm": 0.287109375, "learning_rate": 5.533249686323714e-06, "loss": 0.9426, "step": 1068 }, { "epoch": 1.341279799247177, "grad_norm": 0.2890625, "learning_rate": 5.5290673358427435e-06, "loss": 1.0002, "step": 1069 }, { "epoch": 1.342534504391468, "grad_norm": 0.283203125, "learning_rate": 5.524884985361774e-06, "loss": 0.9783, "step": 1070 }, { "epoch": 1.343789209535759, "grad_norm": 0.263671875, "learning_rate": 5.520702634880804e-06, "loss": 0.9462, "step": 1071 }, { "epoch": 1.3450439146800501, "grad_norm": 0.2734375, "learning_rate": 5.516520284399833e-06, "loss": 0.9521, "step": 1072 }, { "epoch": 1.3462986198243412, "grad_norm": 0.275390625, "learning_rate": 5.512337933918863e-06, "loss": 0.9071, "step": 1073 }, { "epoch": 1.3475533249686324, "grad_norm": 0.3515625, "learning_rate": 5.508155583437893e-06, "loss": 0.9716, "step": 1074 }, { "epoch": 1.3488080301129235, "grad_norm": 0.2734375, "learning_rate": 5.5039732329569215e-06, "loss": 0.9251, "step": 1075 }, { "epoch": 1.3500627352572145, "grad_norm": 0.26953125, "learning_rate": 5.499790882475951e-06, "loss": 0.9876, "step": 1076 }, { "epoch": 1.3513174404015056, "grad_norm": 0.271484375, "learning_rate": 5.495608531994982e-06, "loss": 0.9954, "step": 1077 }, { "epoch": 1.3525721455457966, "grad_norm": 0.25390625, "learning_rate": 5.491426181514011e-06, "loss": 0.9293, "step": 1078 }, { "epoch": 1.353826850690088, "grad_norm": 0.271484375, "learning_rate": 5.487243831033041e-06, "loss": 0.9695, "step": 1079 }, { "epoch": 1.355081555834379, "grad_norm": 0.26171875, "learning_rate": 5.483061480552071e-06, "loss": 0.9301, "step": 1080 }, { "epoch": 1.35633626097867, "grad_norm": 0.2734375, "learning_rate": 5.4788791300711004e-06, "loss": 0.9153, "step": 1081 }, { "epoch": 1.357590966122961, "grad_norm": 0.2734375, "learning_rate": 5.474696779590129e-06, "loss": 0.9341, "step": 1082 }, { "epoch": 1.358845671267252, "grad_norm": 0.26953125, "learning_rate": 5.47051442910916e-06, "loss": 0.9526, "step": 1083 }, { "epoch": 1.3601003764115434, "grad_norm": 0.267578125, "learning_rate": 5.4663320786281895e-06, "loss": 0.9179, "step": 1084 }, { "epoch": 1.3613550815558344, "grad_norm": 0.26171875, "learning_rate": 5.462149728147219e-06, "loss": 0.9597, "step": 1085 }, { "epoch": 1.3626097867001254, "grad_norm": 0.283203125, "learning_rate": 5.457967377666249e-06, "loss": 1.0219, "step": 1086 }, { "epoch": 1.3638644918444165, "grad_norm": 0.271484375, "learning_rate": 5.4537850271852785e-06, "loss": 0.9421, "step": 1087 }, { "epoch": 1.3651191969887075, "grad_norm": 0.263671875, "learning_rate": 5.449602676704309e-06, "loss": 0.9706, "step": 1088 }, { "epoch": 1.3663739021329988, "grad_norm": 0.275390625, "learning_rate": 5.445420326223339e-06, "loss": 0.9287, "step": 1089 }, { "epoch": 1.3676286072772899, "grad_norm": 0.263671875, "learning_rate": 5.4412379757423675e-06, "loss": 0.9222, "step": 1090 }, { "epoch": 1.368883312421581, "grad_norm": 0.25390625, "learning_rate": 5.437055625261397e-06, "loss": 0.9484, "step": 1091 }, { "epoch": 1.370138017565872, "grad_norm": 0.265625, "learning_rate": 5.432873274780427e-06, "loss": 0.9506, "step": 1092 }, { "epoch": 1.371392722710163, "grad_norm": 0.33984375, "learning_rate": 5.4286909242994565e-06, "loss": 0.9595, "step": 1093 }, { "epoch": 1.3726474278544543, "grad_norm": 0.298828125, "learning_rate": 5.424508573818486e-06, "loss": 0.977, "step": 1094 }, { "epoch": 1.3739021329987453, "grad_norm": 0.265625, "learning_rate": 5.420326223337517e-06, "loss": 0.9309, "step": 1095 }, { "epoch": 1.3751568381430364, "grad_norm": 0.341796875, "learning_rate": 5.416143872856546e-06, "loss": 0.8713, "step": 1096 }, { "epoch": 1.3764115432873274, "grad_norm": 0.2734375, "learning_rate": 5.411961522375575e-06, "loss": 0.9795, "step": 1097 }, { "epoch": 1.3776662484316184, "grad_norm": 0.26171875, "learning_rate": 5.407779171894605e-06, "loss": 0.9464, "step": 1098 }, { "epoch": 1.3789209535759097, "grad_norm": 0.271484375, "learning_rate": 5.403596821413635e-06, "loss": 0.962, "step": 1099 }, { "epoch": 1.3801756587202008, "grad_norm": 0.271484375, "learning_rate": 5.399414470932664e-06, "loss": 0.9211, "step": 1100 }, { "epoch": 1.3814303638644918, "grad_norm": 0.271484375, "learning_rate": 5.395232120451695e-06, "loss": 0.9477, "step": 1101 }, { "epoch": 1.382685069008783, "grad_norm": 0.271484375, "learning_rate": 5.3910497699707245e-06, "loss": 0.9108, "step": 1102 }, { "epoch": 1.383939774153074, "grad_norm": 0.28125, "learning_rate": 5.386867419489754e-06, "loss": 0.9815, "step": 1103 }, { "epoch": 1.3851944792973652, "grad_norm": 0.26953125, "learning_rate": 5.382685069008783e-06, "loss": 0.9433, "step": 1104 }, { "epoch": 1.3864491844416562, "grad_norm": 0.306640625, "learning_rate": 5.378502718527813e-06, "loss": 0.9226, "step": 1105 }, { "epoch": 1.3877038895859473, "grad_norm": 0.283203125, "learning_rate": 5.374320368046842e-06, "loss": 0.9652, "step": 1106 }, { "epoch": 1.3889585947302385, "grad_norm": 0.26953125, "learning_rate": 5.370138017565872e-06, "loss": 0.923, "step": 1107 }, { "epoch": 1.3902132998745294, "grad_norm": 0.310546875, "learning_rate": 5.3659556670849025e-06, "loss": 0.8805, "step": 1108 }, { "epoch": 1.3914680050188206, "grad_norm": 0.287109375, "learning_rate": 5.361773316603932e-06, "loss": 0.9781, "step": 1109 }, { "epoch": 1.3927227101631117, "grad_norm": 0.28125, "learning_rate": 5.357590966122962e-06, "loss": 0.9818, "step": 1110 }, { "epoch": 1.3939774153074027, "grad_norm": 0.2734375, "learning_rate": 5.3534086156419916e-06, "loss": 0.9601, "step": 1111 }, { "epoch": 1.395232120451694, "grad_norm": 0.279296875, "learning_rate": 5.34922626516102e-06, "loss": 0.9528, "step": 1112 }, { "epoch": 1.396486825595985, "grad_norm": 0.271484375, "learning_rate": 5.34504391468005e-06, "loss": 0.9401, "step": 1113 }, { "epoch": 1.397741530740276, "grad_norm": 0.263671875, "learning_rate": 5.340861564199081e-06, "loss": 0.9548, "step": 1114 }, { "epoch": 1.3989962358845671, "grad_norm": 0.267578125, "learning_rate": 5.33667921371811e-06, "loss": 0.9816, "step": 1115 }, { "epoch": 1.4002509410288582, "grad_norm": 0.3046875, "learning_rate": 5.33249686323714e-06, "loss": 0.9528, "step": 1116 }, { "epoch": 1.4015056461731494, "grad_norm": 0.28515625, "learning_rate": 5.32831451275617e-06, "loss": 0.8897, "step": 1117 }, { "epoch": 1.4027603513174405, "grad_norm": 0.2734375, "learning_rate": 5.324132162275199e-06, "loss": 0.9115, "step": 1118 }, { "epoch": 1.4040150564617315, "grad_norm": 0.263671875, "learning_rate": 5.319949811794228e-06, "loss": 0.9558, "step": 1119 }, { "epoch": 1.4052697616060226, "grad_norm": 0.26953125, "learning_rate": 5.315767461313258e-06, "loss": 0.9363, "step": 1120 }, { "epoch": 1.4065244667503136, "grad_norm": 0.28515625, "learning_rate": 5.311585110832288e-06, "loss": 0.9689, "step": 1121 }, { "epoch": 1.4077791718946049, "grad_norm": 0.283203125, "learning_rate": 5.307402760351318e-06, "loss": 0.9148, "step": 1122 }, { "epoch": 1.409033877038896, "grad_norm": 0.271484375, "learning_rate": 5.303220409870348e-06, "loss": 0.9468, "step": 1123 }, { "epoch": 1.410288582183187, "grad_norm": 0.279296875, "learning_rate": 5.299038059389377e-06, "loss": 0.9059, "step": 1124 }, { "epoch": 1.411543287327478, "grad_norm": 0.279296875, "learning_rate": 5.294855708908407e-06, "loss": 0.9253, "step": 1125 }, { "epoch": 1.412797992471769, "grad_norm": 0.28125, "learning_rate": 5.290673358427436e-06, "loss": 0.9735, "step": 1126 }, { "epoch": 1.4140526976160603, "grad_norm": 0.263671875, "learning_rate": 5.286491007946466e-06, "loss": 0.9346, "step": 1127 }, { "epoch": 1.4153074027603514, "grad_norm": 0.30859375, "learning_rate": 5.282308657465496e-06, "loss": 0.9296, "step": 1128 }, { "epoch": 1.4165621079046424, "grad_norm": 0.279296875, "learning_rate": 5.278126306984526e-06, "loss": 0.9275, "step": 1129 }, { "epoch": 1.4178168130489335, "grad_norm": 0.263671875, "learning_rate": 5.273943956503555e-06, "loss": 0.952, "step": 1130 }, { "epoch": 1.4190715181932245, "grad_norm": 0.296875, "learning_rate": 5.269761606022585e-06, "loss": 0.98, "step": 1131 }, { "epoch": 1.4203262233375158, "grad_norm": 0.2734375, "learning_rate": 5.265579255541616e-06, "loss": 0.9604, "step": 1132 }, { "epoch": 1.4215809284818068, "grad_norm": 0.294921875, "learning_rate": 5.261396905060645e-06, "loss": 1.0178, "step": 1133 }, { "epoch": 1.4228356336260979, "grad_norm": 0.27734375, "learning_rate": 5.257214554579674e-06, "loss": 0.9243, "step": 1134 }, { "epoch": 1.424090338770389, "grad_norm": 0.2734375, "learning_rate": 5.253032204098704e-06, "loss": 0.9194, "step": 1135 }, { "epoch": 1.42534504391468, "grad_norm": 0.263671875, "learning_rate": 5.2488498536177334e-06, "loss": 0.9544, "step": 1136 }, { "epoch": 1.4265997490589712, "grad_norm": 0.275390625, "learning_rate": 5.244667503136763e-06, "loss": 0.962, "step": 1137 }, { "epoch": 1.4278544542032623, "grad_norm": 0.271484375, "learning_rate": 5.240485152655793e-06, "loss": 0.9459, "step": 1138 }, { "epoch": 1.4291091593475533, "grad_norm": 0.322265625, "learning_rate": 5.236302802174823e-06, "loss": 0.9132, "step": 1139 }, { "epoch": 1.4303638644918444, "grad_norm": 0.267578125, "learning_rate": 5.232120451693853e-06, "loss": 0.9104, "step": 1140 }, { "epoch": 1.4316185696361354, "grad_norm": 0.2578125, "learning_rate": 5.227938101212882e-06, "loss": 0.9253, "step": 1141 }, { "epoch": 1.4328732747804267, "grad_norm": 0.2734375, "learning_rate": 5.2237557507319115e-06, "loss": 0.9746, "step": 1142 }, { "epoch": 1.4341279799247177, "grad_norm": 0.275390625, "learning_rate": 5.219573400250941e-06, "loss": 0.9523, "step": 1143 }, { "epoch": 1.4353826850690088, "grad_norm": 0.275390625, "learning_rate": 5.215391049769971e-06, "loss": 0.9546, "step": 1144 }, { "epoch": 1.4366373902132998, "grad_norm": 0.28125, "learning_rate": 5.211208699289001e-06, "loss": 0.9473, "step": 1145 }, { "epoch": 1.4378920953575909, "grad_norm": 0.26171875, "learning_rate": 5.207026348808031e-06, "loss": 0.9403, "step": 1146 }, { "epoch": 1.4391468005018822, "grad_norm": 0.255859375, "learning_rate": 5.202843998327061e-06, "loss": 0.9644, "step": 1147 }, { "epoch": 1.4404015056461732, "grad_norm": 0.28515625, "learning_rate": 5.1986616478460895e-06, "loss": 0.9353, "step": 1148 }, { "epoch": 1.4416562107904642, "grad_norm": 0.291015625, "learning_rate": 5.194479297365119e-06, "loss": 0.9876, "step": 1149 }, { "epoch": 1.4429109159347553, "grad_norm": 0.28515625, "learning_rate": 5.190296946884149e-06, "loss": 0.8985, "step": 1150 }, { "epoch": 1.4441656210790463, "grad_norm": 0.2734375, "learning_rate": 5.1861145964031786e-06, "loss": 0.9307, "step": 1151 }, { "epoch": 1.4454203262233376, "grad_norm": 0.28515625, "learning_rate": 5.181932245922209e-06, "loss": 0.9642, "step": 1152 }, { "epoch": 1.4466750313676286, "grad_norm": 0.287109375, "learning_rate": 5.177749895441239e-06, "loss": 0.9919, "step": 1153 }, { "epoch": 1.4479297365119197, "grad_norm": 0.30078125, "learning_rate": 5.1735675449602684e-06, "loss": 0.942, "step": 1154 }, { "epoch": 1.4491844416562107, "grad_norm": 0.271484375, "learning_rate": 5.169385194479298e-06, "loss": 0.9205, "step": 1155 }, { "epoch": 1.4504391468005018, "grad_norm": 0.267578125, "learning_rate": 5.165202843998327e-06, "loss": 0.9397, "step": 1156 }, { "epoch": 1.451693851944793, "grad_norm": 0.32421875, "learning_rate": 5.161020493517357e-06, "loss": 1.0375, "step": 1157 }, { "epoch": 1.452948557089084, "grad_norm": 0.271484375, "learning_rate": 5.156838143036387e-06, "loss": 0.9516, "step": 1158 }, { "epoch": 1.4542032622333751, "grad_norm": 0.294921875, "learning_rate": 5.152655792555417e-06, "loss": 0.9762, "step": 1159 }, { "epoch": 1.4554579673776662, "grad_norm": 0.26171875, "learning_rate": 5.1484734420744465e-06, "loss": 0.9364, "step": 1160 }, { "epoch": 1.4567126725219572, "grad_norm": 0.26953125, "learning_rate": 5.144291091593476e-06, "loss": 0.9545, "step": 1161 }, { "epoch": 1.4579673776662485, "grad_norm": 0.263671875, "learning_rate": 5.140108741112506e-06, "loss": 0.9822, "step": 1162 }, { "epoch": 1.4592220828105396, "grad_norm": 0.2734375, "learning_rate": 5.135926390631535e-06, "loss": 0.9215, "step": 1163 }, { "epoch": 1.4604767879548306, "grad_norm": 0.26953125, "learning_rate": 5.131744040150564e-06, "loss": 0.9745, "step": 1164 }, { "epoch": 1.4617314930991216, "grad_norm": 0.39453125, "learning_rate": 5.127561689669595e-06, "loss": 0.9888, "step": 1165 }, { "epoch": 1.4629861982434127, "grad_norm": 0.2734375, "learning_rate": 5.1233793391886246e-06, "loss": 0.992, "step": 1166 }, { "epoch": 1.464240903387704, "grad_norm": 0.296875, "learning_rate": 5.119196988707654e-06, "loss": 0.9344, "step": 1167 }, { "epoch": 1.465495608531995, "grad_norm": 0.275390625, "learning_rate": 5.115014638226684e-06, "loss": 0.8846, "step": 1168 }, { "epoch": 1.466750313676286, "grad_norm": 0.263671875, "learning_rate": 5.110832287745714e-06, "loss": 0.9298, "step": 1169 }, { "epoch": 1.468005018820577, "grad_norm": 0.267578125, "learning_rate": 5.106649937264742e-06, "loss": 0.912, "step": 1170 }, { "epoch": 1.4692597239648681, "grad_norm": 0.271484375, "learning_rate": 5.102467586783773e-06, "loss": 0.9758, "step": 1171 }, { "epoch": 1.4705144291091594, "grad_norm": 0.265625, "learning_rate": 5.098285236302803e-06, "loss": 0.9392, "step": 1172 }, { "epoch": 1.4717691342534505, "grad_norm": 0.2734375, "learning_rate": 5.094102885821832e-06, "loss": 0.9406, "step": 1173 }, { "epoch": 1.4730238393977415, "grad_norm": 0.26953125, "learning_rate": 5.089920535340862e-06, "loss": 0.9547, "step": 1174 }, { "epoch": 1.4742785445420326, "grad_norm": 0.263671875, "learning_rate": 5.085738184859892e-06, "loss": 0.9054, "step": 1175 }, { "epoch": 1.4755332496863236, "grad_norm": 0.275390625, "learning_rate": 5.081555834378922e-06, "loss": 0.9352, "step": 1176 }, { "epoch": 1.4767879548306149, "grad_norm": 0.27734375, "learning_rate": 5.077373483897952e-06, "loss": 0.9859, "step": 1177 }, { "epoch": 1.478042659974906, "grad_norm": 0.259765625, "learning_rate": 5.073191133416981e-06, "loss": 0.9451, "step": 1178 }, { "epoch": 1.479297365119197, "grad_norm": 0.263671875, "learning_rate": 5.06900878293601e-06, "loss": 0.9172, "step": 1179 }, { "epoch": 1.480552070263488, "grad_norm": 0.26171875, "learning_rate": 5.06482643245504e-06, "loss": 0.8976, "step": 1180 }, { "epoch": 1.481806775407779, "grad_norm": 0.296875, "learning_rate": 5.06064408197407e-06, "loss": 0.9777, "step": 1181 }, { "epoch": 1.4830614805520703, "grad_norm": 0.291015625, "learning_rate": 5.056461731493099e-06, "loss": 0.9549, "step": 1182 }, { "epoch": 1.4843161856963614, "grad_norm": 0.263671875, "learning_rate": 5.05227938101213e-06, "loss": 0.9571, "step": 1183 }, { "epoch": 1.4855708908406524, "grad_norm": 0.349609375, "learning_rate": 5.0480970305311596e-06, "loss": 0.9152, "step": 1184 }, { "epoch": 1.4868255959849435, "grad_norm": 0.265625, "learning_rate": 5.043914680050188e-06, "loss": 0.9269, "step": 1185 }, { "epoch": 1.4880803011292345, "grad_norm": 0.283203125, "learning_rate": 5.039732329569218e-06, "loss": 0.9348, "step": 1186 }, { "epoch": 1.4893350062735258, "grad_norm": 0.28515625, "learning_rate": 5.035549979088248e-06, "loss": 0.9351, "step": 1187 }, { "epoch": 1.4905897114178168, "grad_norm": 0.294921875, "learning_rate": 5.031367628607277e-06, "loss": 0.9636, "step": 1188 }, { "epoch": 1.4918444165621079, "grad_norm": 0.2578125, "learning_rate": 5.027185278126308e-06, "loss": 0.9301, "step": 1189 }, { "epoch": 1.4930991217063991, "grad_norm": 0.26171875, "learning_rate": 5.023002927645338e-06, "loss": 0.9736, "step": 1190 }, { "epoch": 1.49435382685069, "grad_norm": 0.26953125, "learning_rate": 5.018820577164367e-06, "loss": 0.9472, "step": 1191 }, { "epoch": 1.4956085319949812, "grad_norm": 0.263671875, "learning_rate": 5.014638226683396e-06, "loss": 0.9551, "step": 1192 }, { "epoch": 1.4968632371392723, "grad_norm": 0.283203125, "learning_rate": 5.010455876202426e-06, "loss": 0.9479, "step": 1193 }, { "epoch": 1.4981179422835633, "grad_norm": 0.26953125, "learning_rate": 5.0062735257214555e-06, "loss": 0.95, "step": 1194 }, { "epoch": 1.4993726474278546, "grad_norm": 0.25390625, "learning_rate": 5.002091175240485e-06, "loss": 0.9427, "step": 1195 }, { "epoch": 1.5006273525721454, "grad_norm": 0.3046875, "learning_rate": 4.997908824759516e-06, "loss": 0.9577, "step": 1196 }, { "epoch": 1.5018820577164367, "grad_norm": 0.263671875, "learning_rate": 4.993726474278545e-06, "loss": 0.9464, "step": 1197 }, { "epoch": 1.5031367628607277, "grad_norm": 0.29296875, "learning_rate": 4.989544123797574e-06, "loss": 0.9761, "step": 1198 }, { "epoch": 1.5043914680050188, "grad_norm": 0.328125, "learning_rate": 4.985361773316605e-06, "loss": 0.8915, "step": 1199 }, { "epoch": 1.50564617314931, "grad_norm": 0.29296875, "learning_rate": 4.981179422835634e-06, "loss": 0.9799, "step": 1200 }, { "epoch": 1.5069008782936009, "grad_norm": 0.2734375, "learning_rate": 4.976997072354664e-06, "loss": 0.9539, "step": 1201 }, { "epoch": 1.5081555834378921, "grad_norm": 0.291015625, "learning_rate": 4.972814721873693e-06, "loss": 0.9595, "step": 1202 }, { "epoch": 1.5094102885821832, "grad_norm": 0.2734375, "learning_rate": 4.968632371392723e-06, "loss": 0.9595, "step": 1203 }, { "epoch": 1.5106649937264742, "grad_norm": 0.265625, "learning_rate": 4.964450020911753e-06, "loss": 0.9848, "step": 1204 }, { "epoch": 1.5119196988707655, "grad_norm": 0.263671875, "learning_rate": 4.960267670430782e-06, "loss": 0.9608, "step": 1205 }, { "epoch": 1.5131744040150563, "grad_norm": 0.2734375, "learning_rate": 4.9560853199498124e-06, "loss": 0.9928, "step": 1206 }, { "epoch": 1.5144291091593476, "grad_norm": 0.2890625, "learning_rate": 4.951902969468842e-06, "loss": 0.9115, "step": 1207 }, { "epoch": 1.5156838143036386, "grad_norm": 0.294921875, "learning_rate": 4.947720618987872e-06, "loss": 0.9372, "step": 1208 }, { "epoch": 1.5169385194479297, "grad_norm": 0.28125, "learning_rate": 4.9435382685069015e-06, "loss": 0.9042, "step": 1209 }, { "epoch": 1.518193224592221, "grad_norm": 0.275390625, "learning_rate": 4.939355918025931e-06, "loss": 0.9727, "step": 1210 }, { "epoch": 1.5194479297365118, "grad_norm": 0.337890625, "learning_rate": 4.935173567544961e-06, "loss": 0.9465, "step": 1211 }, { "epoch": 1.520702634880803, "grad_norm": 0.37890625, "learning_rate": 4.9309912170639905e-06, "loss": 0.9008, "step": 1212 }, { "epoch": 1.521957340025094, "grad_norm": 0.255859375, "learning_rate": 4.92680886658302e-06, "loss": 0.9045, "step": 1213 }, { "epoch": 1.5232120451693851, "grad_norm": 0.2734375, "learning_rate": 4.92262651610205e-06, "loss": 0.9477, "step": 1214 }, { "epoch": 1.5244667503136764, "grad_norm": 0.279296875, "learning_rate": 4.9184441656210795e-06, "loss": 0.9701, "step": 1215 }, { "epoch": 1.5257214554579672, "grad_norm": 0.28125, "learning_rate": 4.914261815140109e-06, "loss": 0.9812, "step": 1216 }, { "epoch": 1.5269761606022585, "grad_norm": 0.263671875, "learning_rate": 4.910079464659139e-06, "loss": 0.9219, "step": 1217 }, { "epoch": 1.5282308657465495, "grad_norm": 0.275390625, "learning_rate": 4.9058971141781685e-06, "loss": 0.9632, "step": 1218 }, { "epoch": 1.5294855708908406, "grad_norm": 0.294921875, "learning_rate": 4.901714763697198e-06, "loss": 0.9788, "step": 1219 }, { "epoch": 1.5307402760351319, "grad_norm": 0.306640625, "learning_rate": 4.897532413216228e-06, "loss": 0.9848, "step": 1220 }, { "epoch": 1.5319949811794227, "grad_norm": 0.265625, "learning_rate": 4.8933500627352576e-06, "loss": 0.9165, "step": 1221 }, { "epoch": 1.533249686323714, "grad_norm": 0.32421875, "learning_rate": 4.889167712254287e-06, "loss": 0.9631, "step": 1222 }, { "epoch": 1.534504391468005, "grad_norm": 0.39453125, "learning_rate": 4.884985361773317e-06, "loss": 0.9519, "step": 1223 }, { "epoch": 1.535759096612296, "grad_norm": 0.26953125, "learning_rate": 4.880803011292347e-06, "loss": 0.9831, "step": 1224 }, { "epoch": 1.5370138017565873, "grad_norm": 0.271484375, "learning_rate": 4.876620660811376e-06, "loss": 0.967, "step": 1225 }, { "epoch": 1.5382685069008781, "grad_norm": 0.2890625, "learning_rate": 4.872438310330406e-06, "loss": 0.9682, "step": 1226 }, { "epoch": 1.5395232120451694, "grad_norm": 0.279296875, "learning_rate": 4.868255959849436e-06, "loss": 0.9308, "step": 1227 }, { "epoch": 1.5407779171894604, "grad_norm": 0.294921875, "learning_rate": 4.864073609368465e-06, "loss": 0.9805, "step": 1228 }, { "epoch": 1.5420326223337515, "grad_norm": 0.3203125, "learning_rate": 4.859891258887495e-06, "loss": 0.962, "step": 1229 }, { "epoch": 1.5432873274780428, "grad_norm": 0.392578125, "learning_rate": 4.8557089084065255e-06, "loss": 0.8751, "step": 1230 }, { "epoch": 1.5445420326223338, "grad_norm": 0.296875, "learning_rate": 4.851526557925554e-06, "loss": 0.905, "step": 1231 }, { "epoch": 1.5457967377666249, "grad_norm": 0.27734375, "learning_rate": 4.847344207444584e-06, "loss": 0.9749, "step": 1232 }, { "epoch": 1.547051442910916, "grad_norm": 0.267578125, "learning_rate": 4.843161856963614e-06, "loss": 0.9355, "step": 1233 }, { "epoch": 1.548306148055207, "grad_norm": 0.2734375, "learning_rate": 4.838979506482644e-06, "loss": 0.9625, "step": 1234 }, { "epoch": 1.5495608531994982, "grad_norm": 0.283203125, "learning_rate": 4.834797156001673e-06, "loss": 0.9241, "step": 1235 }, { "epoch": 1.5508155583437893, "grad_norm": 0.2578125, "learning_rate": 4.830614805520703e-06, "loss": 0.9606, "step": 1236 }, { "epoch": 1.5520702634880803, "grad_norm": 0.263671875, "learning_rate": 4.826432455039733e-06, "loss": 0.9562, "step": 1237 }, { "epoch": 1.5533249686323714, "grad_norm": 0.265625, "learning_rate": 4.822250104558762e-06, "loss": 0.9602, "step": 1238 }, { "epoch": 1.5545796737766624, "grad_norm": 0.2578125, "learning_rate": 4.818067754077792e-06, "loss": 0.9224, "step": 1239 }, { "epoch": 1.5558343789209537, "grad_norm": 0.298828125, "learning_rate": 4.813885403596822e-06, "loss": 0.9634, "step": 1240 }, { "epoch": 1.5570890840652447, "grad_norm": 0.271484375, "learning_rate": 4.809703053115852e-06, "loss": 0.9895, "step": 1241 }, { "epoch": 1.5583437892095358, "grad_norm": 0.26953125, "learning_rate": 4.805520702634881e-06, "loss": 0.9085, "step": 1242 }, { "epoch": 1.5595984943538268, "grad_norm": 0.255859375, "learning_rate": 4.801338352153911e-06, "loss": 0.9516, "step": 1243 }, { "epoch": 1.5608531994981178, "grad_norm": 0.279296875, "learning_rate": 4.797156001672941e-06, "loss": 0.9317, "step": 1244 }, { "epoch": 1.5621079046424091, "grad_norm": 0.271484375, "learning_rate": 4.792973651191971e-06, "loss": 0.954, "step": 1245 }, { "epoch": 1.5633626097867002, "grad_norm": 0.265625, "learning_rate": 4.7887913007109994e-06, "loss": 0.9488, "step": 1246 }, { "epoch": 1.5646173149309912, "grad_norm": 0.29296875, "learning_rate": 4.78460895023003e-06, "loss": 0.9439, "step": 1247 }, { "epoch": 1.5658720200752823, "grad_norm": 0.259765625, "learning_rate": 4.78042659974906e-06, "loss": 0.9706, "step": 1248 }, { "epoch": 1.5671267252195733, "grad_norm": 0.271484375, "learning_rate": 4.7762442492680885e-06, "loss": 0.9174, "step": 1249 }, { "epoch": 1.5683814303638646, "grad_norm": 0.349609375, "learning_rate": 4.772061898787119e-06, "loss": 0.9758, "step": 1250 }, { "epoch": 1.5696361355081556, "grad_norm": 0.267578125, "learning_rate": 4.767879548306149e-06, "loss": 0.9522, "step": 1251 }, { "epoch": 1.5708908406524467, "grad_norm": 0.25390625, "learning_rate": 4.763697197825178e-06, "loss": 0.914, "step": 1252 }, { "epoch": 1.572145545796738, "grad_norm": 0.287109375, "learning_rate": 4.759514847344208e-06, "loss": 0.984, "step": 1253 }, { "epoch": 1.5734002509410288, "grad_norm": 0.32421875, "learning_rate": 4.755332496863238e-06, "loss": 0.9355, "step": 1254 }, { "epoch": 1.57465495608532, "grad_norm": 0.265625, "learning_rate": 4.751150146382267e-06, "loss": 0.9155, "step": 1255 }, { "epoch": 1.575909661229611, "grad_norm": 0.279296875, "learning_rate": 4.746967795901297e-06, "loss": 0.9669, "step": 1256 }, { "epoch": 1.5771643663739021, "grad_norm": 0.302734375, "learning_rate": 4.742785445420327e-06, "loss": 0.9345, "step": 1257 }, { "epoch": 1.5784190715181934, "grad_norm": 0.267578125, "learning_rate": 4.738603094939356e-06, "loss": 0.9698, "step": 1258 }, { "epoch": 1.5796737766624842, "grad_norm": 0.263671875, "learning_rate": 4.734420744458386e-06, "loss": 0.9387, "step": 1259 }, { "epoch": 1.5809284818067755, "grad_norm": 0.42578125, "learning_rate": 4.730238393977416e-06, "loss": 0.9509, "step": 1260 }, { "epoch": 1.5821831869510665, "grad_norm": 0.275390625, "learning_rate": 4.7260560434964454e-06, "loss": 0.9456, "step": 1261 }, { "epoch": 1.5834378920953576, "grad_norm": 0.29296875, "learning_rate": 4.721873693015475e-06, "loss": 0.9672, "step": 1262 }, { "epoch": 1.5846925972396488, "grad_norm": 0.2734375, "learning_rate": 4.717691342534505e-06, "loss": 0.9118, "step": 1263 }, { "epoch": 1.5859473023839397, "grad_norm": 0.267578125, "learning_rate": 4.7135089920535345e-06, "loss": 0.9105, "step": 1264 }, { "epoch": 1.587202007528231, "grad_norm": 0.265625, "learning_rate": 4.709326641572564e-06, "loss": 0.9289, "step": 1265 }, { "epoch": 1.588456712672522, "grad_norm": 0.28125, "learning_rate": 4.705144291091594e-06, "loss": 1.0065, "step": 1266 }, { "epoch": 1.589711417816813, "grad_norm": 0.267578125, "learning_rate": 4.7009619406106235e-06, "loss": 0.9846, "step": 1267 }, { "epoch": 1.5909661229611043, "grad_norm": 0.2578125, "learning_rate": 4.696779590129653e-06, "loss": 0.9267, "step": 1268 }, { "epoch": 1.5922208281053951, "grad_norm": 0.306640625, "learning_rate": 4.692597239648683e-06, "loss": 0.9622, "step": 1269 }, { "epoch": 1.5934755332496864, "grad_norm": 0.28515625, "learning_rate": 4.6884148891677125e-06, "loss": 0.8962, "step": 1270 }, { "epoch": 1.5947302383939774, "grad_norm": 0.26953125, "learning_rate": 4.684232538686742e-06, "loss": 0.9209, "step": 1271 }, { "epoch": 1.5959849435382685, "grad_norm": 0.294921875, "learning_rate": 4.680050188205772e-06, "loss": 0.931, "step": 1272 }, { "epoch": 1.5972396486825597, "grad_norm": 0.2734375, "learning_rate": 4.6758678377248015e-06, "loss": 0.9195, "step": 1273 }, { "epoch": 1.5984943538268506, "grad_norm": 0.271484375, "learning_rate": 4.671685487243832e-06, "loss": 0.9705, "step": 1274 }, { "epoch": 1.5997490589711418, "grad_norm": 0.2734375, "learning_rate": 4.667503136762861e-06, "loss": 0.9807, "step": 1275 }, { "epoch": 1.6010037641154329, "grad_norm": 0.314453125, "learning_rate": 4.6633207862818906e-06, "loss": 0.9327, "step": 1276 }, { "epoch": 1.602258469259724, "grad_norm": 0.28515625, "learning_rate": 4.65913843580092e-06, "loss": 0.9309, "step": 1277 }, { "epoch": 1.6035131744040152, "grad_norm": 0.267578125, "learning_rate": 4.654956085319951e-06, "loss": 0.9493, "step": 1278 }, { "epoch": 1.604767879548306, "grad_norm": 0.26171875, "learning_rate": 4.65077373483898e-06, "loss": 0.9744, "step": 1279 }, { "epoch": 1.6060225846925973, "grad_norm": 0.26171875, "learning_rate": 4.646591384358009e-06, "loss": 0.8877, "step": 1280 }, { "epoch": 1.6072772898368883, "grad_norm": 0.259765625, "learning_rate": 4.64240903387704e-06, "loss": 0.9054, "step": 1281 }, { "epoch": 1.6085319949811794, "grad_norm": 0.2734375, "learning_rate": 4.638226683396069e-06, "loss": 0.9757, "step": 1282 }, { "epoch": 1.6097867001254706, "grad_norm": 0.263671875, "learning_rate": 4.634044332915098e-06, "loss": 0.9387, "step": 1283 }, { "epoch": 1.6110414052697615, "grad_norm": 0.31640625, "learning_rate": 4.629861982434129e-06, "loss": 0.9306, "step": 1284 }, { "epoch": 1.6122961104140527, "grad_norm": 0.30859375, "learning_rate": 4.6256796319531585e-06, "loss": 0.9574, "step": 1285 }, { "epoch": 1.6135508155583438, "grad_norm": 0.267578125, "learning_rate": 4.621497281472187e-06, "loss": 0.9492, "step": 1286 }, { "epoch": 1.6148055207026348, "grad_norm": 0.26953125, "learning_rate": 4.617314930991217e-06, "loss": 0.9678, "step": 1287 }, { "epoch": 1.616060225846926, "grad_norm": 0.259765625, "learning_rate": 4.6131325805102475e-06, "loss": 0.9534, "step": 1288 }, { "epoch": 1.617314930991217, "grad_norm": 0.29296875, "learning_rate": 4.608950230029277e-06, "loss": 0.9672, "step": 1289 }, { "epoch": 1.6185696361355082, "grad_norm": 0.287109375, "learning_rate": 4.604767879548306e-06, "loss": 0.9158, "step": 1290 }, { "epoch": 1.6198243412797992, "grad_norm": 0.2734375, "learning_rate": 4.6005855290673365e-06, "loss": 0.9476, "step": 1291 }, { "epoch": 1.6210790464240903, "grad_norm": 0.2734375, "learning_rate": 4.596403178586366e-06, "loss": 0.9695, "step": 1292 }, { "epoch": 1.6223337515683816, "grad_norm": 0.28515625, "learning_rate": 4.592220828105395e-06, "loss": 0.9929, "step": 1293 }, { "epoch": 1.6235884567126724, "grad_norm": 0.392578125, "learning_rate": 4.5880384776244256e-06, "loss": 0.9409, "step": 1294 }, { "epoch": 1.6248431618569636, "grad_norm": 0.265625, "learning_rate": 4.583856127143455e-06, "loss": 0.9224, "step": 1295 }, { "epoch": 1.6260978670012547, "grad_norm": 0.2734375, "learning_rate": 4.579673776662485e-06, "loss": 0.9455, "step": 1296 }, { "epoch": 1.6273525721455457, "grad_norm": 0.279296875, "learning_rate": 4.575491426181515e-06, "loss": 0.9145, "step": 1297 }, { "epoch": 1.628607277289837, "grad_norm": 0.275390625, "learning_rate": 4.571309075700544e-06, "loss": 0.9183, "step": 1298 }, { "epoch": 1.6298619824341278, "grad_norm": 0.26953125, "learning_rate": 4.567126725219574e-06, "loss": 0.9301, "step": 1299 }, { "epoch": 1.631116687578419, "grad_norm": 0.263671875, "learning_rate": 4.562944374738604e-06, "loss": 0.8999, "step": 1300 }, { "epoch": 1.6323713927227101, "grad_norm": 0.373046875, "learning_rate": 4.558762024257633e-06, "loss": 0.9773, "step": 1301 }, { "epoch": 1.6336260978670012, "grad_norm": 0.36328125, "learning_rate": 4.554579673776663e-06, "loss": 0.906, "step": 1302 }, { "epoch": 1.6348808030112925, "grad_norm": 0.265625, "learning_rate": 4.550397323295693e-06, "loss": 0.9539, "step": 1303 }, { "epoch": 1.6361355081555833, "grad_norm": 0.26953125, "learning_rate": 4.546214972814722e-06, "loss": 0.9563, "step": 1304 }, { "epoch": 1.6373902132998746, "grad_norm": 0.283203125, "learning_rate": 4.542032622333752e-06, "loss": 0.9738, "step": 1305 }, { "epoch": 1.6386449184441656, "grad_norm": 0.28125, "learning_rate": 4.537850271852782e-06, "loss": 0.9674, "step": 1306 }, { "epoch": 1.6398996235884566, "grad_norm": 0.302734375, "learning_rate": 4.533667921371811e-06, "loss": 0.919, "step": 1307 }, { "epoch": 1.641154328732748, "grad_norm": 0.275390625, "learning_rate": 4.529485570890841e-06, "loss": 0.9623, "step": 1308 }, { "epoch": 1.6424090338770387, "grad_norm": 0.26953125, "learning_rate": 4.525303220409871e-06, "loss": 0.8997, "step": 1309 }, { "epoch": 1.64366373902133, "grad_norm": 0.255859375, "learning_rate": 4.5211208699289e-06, "loss": 0.9259, "step": 1310 }, { "epoch": 1.644918444165621, "grad_norm": 0.27734375, "learning_rate": 4.51693851944793e-06, "loss": 0.9539, "step": 1311 }, { "epoch": 1.646173149309912, "grad_norm": 0.298828125, "learning_rate": 4.51275616896696e-06, "loss": 0.93, "step": 1312 }, { "epoch": 1.6474278544542034, "grad_norm": 0.26953125, "learning_rate": 4.508573818485989e-06, "loss": 0.8979, "step": 1313 }, { "epoch": 1.6486825595984942, "grad_norm": 0.263671875, "learning_rate": 4.504391468005019e-06, "loss": 0.9336, "step": 1314 }, { "epoch": 1.6499372647427855, "grad_norm": 0.349609375, "learning_rate": 4.500209117524049e-06, "loss": 0.9369, "step": 1315 }, { "epoch": 1.6511919698870765, "grad_norm": 0.28125, "learning_rate": 4.4960267670430784e-06, "loss": 0.9423, "step": 1316 }, { "epoch": 1.6524466750313676, "grad_norm": 0.283203125, "learning_rate": 4.491844416562108e-06, "loss": 0.9258, "step": 1317 }, { "epoch": 1.6537013801756588, "grad_norm": 0.271484375, "learning_rate": 4.487662066081138e-06, "loss": 0.9234, "step": 1318 }, { "epoch": 1.6549560853199499, "grad_norm": 0.259765625, "learning_rate": 4.4834797156001675e-06, "loss": 0.881, "step": 1319 }, { "epoch": 1.656210790464241, "grad_norm": 0.28125, "learning_rate": 4.479297365119197e-06, "loss": 0.9464, "step": 1320 }, { "epoch": 1.657465495608532, "grad_norm": 0.271484375, "learning_rate": 4.475115014638227e-06, "loss": 0.9544, "step": 1321 }, { "epoch": 1.658720200752823, "grad_norm": 0.291015625, "learning_rate": 4.470932664157257e-06, "loss": 0.9198, "step": 1322 }, { "epoch": 1.6599749058971143, "grad_norm": 0.33984375, "learning_rate": 4.466750313676286e-06, "loss": 0.9037, "step": 1323 }, { "epoch": 1.6612296110414053, "grad_norm": 0.267578125, "learning_rate": 4.462567963195316e-06, "loss": 0.9327, "step": 1324 }, { "epoch": 1.6624843161856964, "grad_norm": 0.369140625, "learning_rate": 4.458385612714346e-06, "loss": 0.9092, "step": 1325 }, { "epoch": 1.6637390213299874, "grad_norm": 0.37890625, "learning_rate": 4.454203262233375e-06, "loss": 0.9954, "step": 1326 }, { "epoch": 1.6649937264742785, "grad_norm": 0.28125, "learning_rate": 4.450020911752405e-06, "loss": 0.9692, "step": 1327 }, { "epoch": 1.6662484316185697, "grad_norm": 0.27734375, "learning_rate": 4.445838561271435e-06, "loss": 0.9387, "step": 1328 }, { "epoch": 1.6675031367628608, "grad_norm": 0.275390625, "learning_rate": 4.441656210790465e-06, "loss": 0.9114, "step": 1329 }, { "epoch": 1.6687578419071518, "grad_norm": 0.294921875, "learning_rate": 4.437473860309494e-06, "loss": 0.9418, "step": 1330 }, { "epoch": 1.6700125470514429, "grad_norm": 0.79296875, "learning_rate": 4.4332915098285236e-06, "loss": 0.9102, "step": 1331 }, { "epoch": 1.671267252195734, "grad_norm": 0.28515625, "learning_rate": 4.429109159347554e-06, "loss": 0.9516, "step": 1332 }, { "epoch": 1.6725219573400252, "grad_norm": 0.29296875, "learning_rate": 4.424926808866584e-06, "loss": 0.9996, "step": 1333 }, { "epoch": 1.6737766624843162, "grad_norm": 0.267578125, "learning_rate": 4.420744458385613e-06, "loss": 0.9275, "step": 1334 }, { "epoch": 1.6750313676286073, "grad_norm": 0.275390625, "learning_rate": 4.416562107904643e-06, "loss": 0.9577, "step": 1335 }, { "epoch": 1.6762860727728983, "grad_norm": 0.2890625, "learning_rate": 4.412379757423673e-06, "loss": 0.9605, "step": 1336 }, { "epoch": 1.6775407779171894, "grad_norm": 0.27734375, "learning_rate": 4.408197406942702e-06, "loss": 0.9233, "step": 1337 }, { "epoch": 1.6787954830614806, "grad_norm": 0.28515625, "learning_rate": 4.404015056461732e-06, "loss": 0.9751, "step": 1338 }, { "epoch": 1.6800501882057717, "grad_norm": 0.271484375, "learning_rate": 4.399832705980762e-06, "loss": 0.9206, "step": 1339 }, { "epoch": 1.6813048933500627, "grad_norm": 0.263671875, "learning_rate": 4.3956503554997915e-06, "loss": 0.923, "step": 1340 }, { "epoch": 1.682559598494354, "grad_norm": 0.28125, "learning_rate": 4.391468005018821e-06, "loss": 0.9212, "step": 1341 }, { "epoch": 1.6838143036386448, "grad_norm": 0.26953125, "learning_rate": 4.387285654537851e-06, "loss": 0.9006, "step": 1342 }, { "epoch": 1.685069008782936, "grad_norm": 0.267578125, "learning_rate": 4.3831033040568805e-06, "loss": 0.9491, "step": 1343 }, { "epoch": 1.6863237139272271, "grad_norm": 0.267578125, "learning_rate": 4.37892095357591e-06, "loss": 0.9436, "step": 1344 }, { "epoch": 1.6875784190715182, "grad_norm": 0.30078125, "learning_rate": 4.37473860309494e-06, "loss": 0.9649, "step": 1345 }, { "epoch": 1.6888331242158094, "grad_norm": 0.26953125, "learning_rate": 4.3705562526139695e-06, "loss": 0.9452, "step": 1346 }, { "epoch": 1.6900878293601003, "grad_norm": 0.26171875, "learning_rate": 4.366373902132999e-06, "loss": 0.9352, "step": 1347 }, { "epoch": 1.6913425345043915, "grad_norm": 0.279296875, "learning_rate": 4.362191551652029e-06, "loss": 0.9965, "step": 1348 }, { "epoch": 1.6925972396486826, "grad_norm": 0.37109375, "learning_rate": 4.3580092011710586e-06, "loss": 0.9574, "step": 1349 }, { "epoch": 1.6938519447929736, "grad_norm": 0.294921875, "learning_rate": 4.353826850690088e-06, "loss": 0.9535, "step": 1350 }, { "epoch": 1.695106649937265, "grad_norm": 0.287109375, "learning_rate": 4.349644500209118e-06, "loss": 0.957, "step": 1351 }, { "epoch": 1.6963613550815557, "grad_norm": 0.296875, "learning_rate": 4.345462149728148e-06, "loss": 1.0158, "step": 1352 }, { "epoch": 1.697616060225847, "grad_norm": 0.361328125, "learning_rate": 4.341279799247177e-06, "loss": 0.9637, "step": 1353 }, { "epoch": 1.698870765370138, "grad_norm": 0.298828125, "learning_rate": 4.337097448766207e-06, "loss": 0.9243, "step": 1354 }, { "epoch": 1.700125470514429, "grad_norm": 0.283203125, "learning_rate": 4.332915098285237e-06, "loss": 0.9615, "step": 1355 }, { "epoch": 1.7013801756587204, "grad_norm": 0.341796875, "learning_rate": 4.328732747804266e-06, "loss": 0.955, "step": 1356 }, { "epoch": 1.7026348808030112, "grad_norm": 0.29296875, "learning_rate": 4.324550397323296e-06, "loss": 0.9827, "step": 1357 }, { "epoch": 1.7038895859473024, "grad_norm": 0.265625, "learning_rate": 4.320368046842326e-06, "loss": 0.9033, "step": 1358 }, { "epoch": 1.7051442910915935, "grad_norm": 0.291015625, "learning_rate": 4.316185696361355e-06, "loss": 0.9374, "step": 1359 }, { "epoch": 1.7063989962358845, "grad_norm": 0.2734375, "learning_rate": 4.312003345880385e-06, "loss": 0.9504, "step": 1360 }, { "epoch": 1.7076537013801758, "grad_norm": 0.267578125, "learning_rate": 4.307820995399415e-06, "loss": 0.9622, "step": 1361 }, { "epoch": 1.7089084065244666, "grad_norm": 0.259765625, "learning_rate": 4.303638644918444e-06, "loss": 0.9282, "step": 1362 }, { "epoch": 1.710163111668758, "grad_norm": 0.265625, "learning_rate": 4.299456294437474e-06, "loss": 0.9667, "step": 1363 }, { "epoch": 1.711417816813049, "grad_norm": 0.34765625, "learning_rate": 4.295273943956504e-06, "loss": 0.992, "step": 1364 }, { "epoch": 1.71267252195734, "grad_norm": 0.267578125, "learning_rate": 4.291091593475533e-06, "loss": 0.8975, "step": 1365 }, { "epoch": 1.7139272271016313, "grad_norm": 0.267578125, "learning_rate": 4.286909242994564e-06, "loss": 0.9614, "step": 1366 }, { "epoch": 1.715181932245922, "grad_norm": 0.28515625, "learning_rate": 4.282726892513593e-06, "loss": 0.9852, "step": 1367 }, { "epoch": 1.7164366373902133, "grad_norm": 0.271484375, "learning_rate": 4.278544542032622e-06, "loss": 0.9321, "step": 1368 }, { "epoch": 1.7176913425345044, "grad_norm": 0.296875, "learning_rate": 4.274362191551653e-06, "loss": 0.9904, "step": 1369 }, { "epoch": 1.7189460476787954, "grad_norm": 0.267578125, "learning_rate": 4.270179841070682e-06, "loss": 0.9517, "step": 1370 }, { "epoch": 1.7202007528230867, "grad_norm": 0.310546875, "learning_rate": 4.2659974905897114e-06, "loss": 0.9122, "step": 1371 }, { "epoch": 1.7214554579673775, "grad_norm": 0.263671875, "learning_rate": 4.261815140108741e-06, "loss": 0.9245, "step": 1372 }, { "epoch": 1.7227101631116688, "grad_norm": 0.279296875, "learning_rate": 4.257632789627772e-06, "loss": 0.9261, "step": 1373 }, { "epoch": 1.7239648682559598, "grad_norm": 0.26953125, "learning_rate": 4.2534504391468005e-06, "loss": 0.9266, "step": 1374 }, { "epoch": 1.725219573400251, "grad_norm": 0.283203125, "learning_rate": 4.24926808866583e-06, "loss": 0.9108, "step": 1375 }, { "epoch": 1.7264742785445422, "grad_norm": 0.384765625, "learning_rate": 4.245085738184861e-06, "loss": 0.9527, "step": 1376 }, { "epoch": 1.727728983688833, "grad_norm": 0.287109375, "learning_rate": 4.24090338770389e-06, "loss": 0.9464, "step": 1377 }, { "epoch": 1.7289836888331243, "grad_norm": 0.271484375, "learning_rate": 4.236721037222919e-06, "loss": 0.9277, "step": 1378 }, { "epoch": 1.7302383939774153, "grad_norm": 0.306640625, "learning_rate": 4.23253868674195e-06, "loss": 0.9852, "step": 1379 }, { "epoch": 1.7314930991217063, "grad_norm": 0.294921875, "learning_rate": 4.228356336260979e-06, "loss": 0.9597, "step": 1380 }, { "epoch": 1.7327478042659976, "grad_norm": 0.291015625, "learning_rate": 4.224173985780008e-06, "loss": 0.9281, "step": 1381 }, { "epoch": 1.7340025094102884, "grad_norm": 0.279296875, "learning_rate": 4.219991635299039e-06, "loss": 0.9199, "step": 1382 }, { "epoch": 1.7352572145545797, "grad_norm": 0.265625, "learning_rate": 4.215809284818068e-06, "loss": 0.9329, "step": 1383 }, { "epoch": 1.7365119196988708, "grad_norm": 0.27734375, "learning_rate": 4.211626934337098e-06, "loss": 0.9763, "step": 1384 }, { "epoch": 1.7377666248431618, "grad_norm": 0.283203125, "learning_rate": 4.207444583856127e-06, "loss": 0.9691, "step": 1385 }, { "epoch": 1.739021329987453, "grad_norm": 0.2734375, "learning_rate": 4.203262233375157e-06, "loss": 1.0113, "step": 1386 }, { "epoch": 1.740276035131744, "grad_norm": 0.255859375, "learning_rate": 4.199079882894187e-06, "loss": 0.9266, "step": 1387 }, { "epoch": 1.7415307402760352, "grad_norm": 0.2890625, "learning_rate": 4.194897532413217e-06, "loss": 0.9745, "step": 1388 }, { "epoch": 1.7427854454203262, "grad_norm": 0.271484375, "learning_rate": 4.1907151819322464e-06, "loss": 0.9197, "step": 1389 }, { "epoch": 1.7440401505646173, "grad_norm": 0.26953125, "learning_rate": 4.186532831451276e-06, "loss": 0.9426, "step": 1390 }, { "epoch": 1.7452948557089085, "grad_norm": 0.61328125, "learning_rate": 4.182350480970306e-06, "loss": 0.982, "step": 1391 }, { "epoch": 1.7465495608531993, "grad_norm": 0.263671875, "learning_rate": 4.1781681304893355e-06, "loss": 0.9701, "step": 1392 }, { "epoch": 1.7478042659974906, "grad_norm": 0.2734375, "learning_rate": 4.173985780008365e-06, "loss": 0.9677, "step": 1393 }, { "epoch": 1.7490589711417817, "grad_norm": 0.322265625, "learning_rate": 4.169803429527395e-06, "loss": 0.9274, "step": 1394 }, { "epoch": 1.7503136762860727, "grad_norm": 0.275390625, "learning_rate": 4.1656210790464245e-06, "loss": 0.9453, "step": 1395 }, { "epoch": 1.751568381430364, "grad_norm": 0.27734375, "learning_rate": 4.161438728565454e-06, "loss": 0.9297, "step": 1396 }, { "epoch": 1.7528230865746548, "grad_norm": 0.29296875, "learning_rate": 4.157256378084484e-06, "loss": 0.8825, "step": 1397 }, { "epoch": 1.754077791718946, "grad_norm": 0.275390625, "learning_rate": 4.1530740276035135e-06, "loss": 0.9003, "step": 1398 }, { "epoch": 1.7553324968632371, "grad_norm": 0.515625, "learning_rate": 4.148891677122543e-06, "loss": 0.9064, "step": 1399 }, { "epoch": 1.7565872020075282, "grad_norm": 0.283203125, "learning_rate": 4.144709326641573e-06, "loss": 0.9746, "step": 1400 }, { "epoch": 1.7578419071518194, "grad_norm": 0.306640625, "learning_rate": 4.1405269761606026e-06, "loss": 0.9684, "step": 1401 }, { "epoch": 1.7590966122961103, "grad_norm": 0.298828125, "learning_rate": 4.136344625679632e-06, "loss": 0.928, "step": 1402 }, { "epoch": 1.7603513174404015, "grad_norm": 0.275390625, "learning_rate": 4.132162275198662e-06, "loss": 0.9324, "step": 1403 }, { "epoch": 1.7616060225846926, "grad_norm": 0.373046875, "learning_rate": 4.127979924717692e-06, "loss": 1.0713, "step": 1404 }, { "epoch": 1.7628607277289836, "grad_norm": 0.283203125, "learning_rate": 4.123797574236721e-06, "loss": 0.9834, "step": 1405 }, { "epoch": 1.7641154328732749, "grad_norm": 0.2890625, "learning_rate": 4.119615223755751e-06, "loss": 0.9332, "step": 1406 }, { "epoch": 1.765370138017566, "grad_norm": 0.357421875, "learning_rate": 4.115432873274781e-06, "loss": 0.974, "step": 1407 }, { "epoch": 1.766624843161857, "grad_norm": 0.275390625, "learning_rate": 4.11125052279381e-06, "loss": 0.932, "step": 1408 }, { "epoch": 1.767879548306148, "grad_norm": 0.275390625, "learning_rate": 4.10706817231284e-06, "loss": 0.9463, "step": 1409 }, { "epoch": 1.769134253450439, "grad_norm": 0.291015625, "learning_rate": 4.1028858218318705e-06, "loss": 0.9227, "step": 1410 }, { "epoch": 1.7703889585947303, "grad_norm": 0.28515625, "learning_rate": 4.098703471350899e-06, "loss": 0.9355, "step": 1411 }, { "epoch": 1.7716436637390214, "grad_norm": 0.283203125, "learning_rate": 4.094521120869929e-06, "loss": 0.9328, "step": 1412 }, { "epoch": 1.7728983688833124, "grad_norm": 0.28125, "learning_rate": 4.0903387703889595e-06, "loss": 0.9376, "step": 1413 }, { "epoch": 1.7741530740276035, "grad_norm": 0.29296875, "learning_rate": 4.086156419907988e-06, "loss": 0.9094, "step": 1414 }, { "epoch": 1.7754077791718945, "grad_norm": 0.2734375, "learning_rate": 4.081974069427018e-06, "loss": 0.8908, "step": 1415 }, { "epoch": 1.7766624843161858, "grad_norm": 0.2734375, "learning_rate": 4.077791718946048e-06, "loss": 0.9396, "step": 1416 }, { "epoch": 1.7779171894604768, "grad_norm": 0.2734375, "learning_rate": 4.073609368465078e-06, "loss": 0.9524, "step": 1417 }, { "epoch": 1.7791718946047679, "grad_norm": 0.271484375, "learning_rate": 4.069427017984107e-06, "loss": 0.8994, "step": 1418 }, { "epoch": 1.780426599749059, "grad_norm": 0.283203125, "learning_rate": 4.065244667503137e-06, "loss": 0.9339, "step": 1419 }, { "epoch": 1.78168130489335, "grad_norm": 0.326171875, "learning_rate": 4.061062317022167e-06, "loss": 0.9122, "step": 1420 }, { "epoch": 1.7829360100376412, "grad_norm": 0.28515625, "learning_rate": 4.056879966541197e-06, "loss": 0.8963, "step": 1421 }, { "epoch": 1.7841907151819323, "grad_norm": 0.287109375, "learning_rate": 4.052697616060226e-06, "loss": 0.9333, "step": 1422 }, { "epoch": 1.7854454203262233, "grad_norm": 0.27734375, "learning_rate": 4.048515265579256e-06, "loss": 0.9702, "step": 1423 }, { "epoch": 1.7867001254705144, "grad_norm": 0.271484375, "learning_rate": 4.044332915098286e-06, "loss": 0.9184, "step": 1424 }, { "epoch": 1.7879548306148054, "grad_norm": 0.30078125, "learning_rate": 4.040150564617315e-06, "loss": 1.0158, "step": 1425 }, { "epoch": 1.7892095357590967, "grad_norm": 0.33203125, "learning_rate": 4.035968214136345e-06, "loss": 0.9422, "step": 1426 }, { "epoch": 1.7904642409033877, "grad_norm": 0.271484375, "learning_rate": 4.031785863655375e-06, "loss": 0.8994, "step": 1427 }, { "epoch": 1.7917189460476788, "grad_norm": 0.28125, "learning_rate": 4.027603513174405e-06, "loss": 0.9893, "step": 1428 }, { "epoch": 1.79297365119197, "grad_norm": 0.3203125, "learning_rate": 4.0234211626934335e-06, "loss": 0.9217, "step": 1429 }, { "epoch": 1.7942283563362609, "grad_norm": 0.271484375, "learning_rate": 4.019238812212464e-06, "loss": 0.9228, "step": 1430 }, { "epoch": 1.7954830614805521, "grad_norm": 0.271484375, "learning_rate": 4.015056461731494e-06, "loss": 0.931, "step": 1431 }, { "epoch": 1.7967377666248432, "grad_norm": 0.283203125, "learning_rate": 4.010874111250523e-06, "loss": 0.9653, "step": 1432 }, { "epoch": 1.7979924717691342, "grad_norm": 0.2734375, "learning_rate": 4.006691760769553e-06, "loss": 0.8854, "step": 1433 }, { "epoch": 1.7992471769134255, "grad_norm": 0.283203125, "learning_rate": 4.002509410288583e-06, "loss": 0.9516, "step": 1434 }, { "epoch": 1.8005018820577163, "grad_norm": 0.271484375, "learning_rate": 3.998327059807612e-06, "loss": 0.8874, "step": 1435 }, { "epoch": 1.8017565872020076, "grad_norm": 0.26171875, "learning_rate": 3.994144709326642e-06, "loss": 0.9442, "step": 1436 }, { "epoch": 1.8030112923462986, "grad_norm": 0.2890625, "learning_rate": 3.989962358845672e-06, "loss": 0.967, "step": 1437 }, { "epoch": 1.8042659974905897, "grad_norm": 0.271484375, "learning_rate": 3.985780008364701e-06, "loss": 0.9374, "step": 1438 }, { "epoch": 1.805520702634881, "grad_norm": 0.259765625, "learning_rate": 3.981597657883731e-06, "loss": 0.9022, "step": 1439 }, { "epoch": 1.8067754077791718, "grad_norm": 0.279296875, "learning_rate": 3.977415307402761e-06, "loss": 0.9464, "step": 1440 }, { "epoch": 1.808030112923463, "grad_norm": 0.28125, "learning_rate": 3.97323295692179e-06, "loss": 0.9199, "step": 1441 }, { "epoch": 1.809284818067754, "grad_norm": 0.279296875, "learning_rate": 3.96905060644082e-06, "loss": 0.9508, "step": 1442 }, { "epoch": 1.8105395232120451, "grad_norm": 0.265625, "learning_rate": 3.96486825595985e-06, "loss": 0.93, "step": 1443 }, { "epoch": 1.8117942283563364, "grad_norm": 0.28515625, "learning_rate": 3.9606859054788794e-06, "loss": 0.9715, "step": 1444 }, { "epoch": 1.8130489335006272, "grad_norm": 0.287109375, "learning_rate": 3.956503554997909e-06, "loss": 0.9506, "step": 1445 }, { "epoch": 1.8143036386449185, "grad_norm": 0.306640625, "learning_rate": 3.952321204516939e-06, "loss": 0.938, "step": 1446 }, { "epoch": 1.8155583437892095, "grad_norm": 0.306640625, "learning_rate": 3.9481388540359685e-06, "loss": 0.9365, "step": 1447 }, { "epoch": 1.8168130489335006, "grad_norm": 0.30078125, "learning_rate": 3.943956503554998e-06, "loss": 0.915, "step": 1448 }, { "epoch": 1.8180677540777919, "grad_norm": 0.2734375, "learning_rate": 3.939774153074028e-06, "loss": 0.9406, "step": 1449 }, { "epoch": 1.8193224592220827, "grad_norm": 0.27734375, "learning_rate": 3.9355918025930575e-06, "loss": 0.9703, "step": 1450 }, { "epoch": 1.820577164366374, "grad_norm": 0.26953125, "learning_rate": 3.931409452112087e-06, "loss": 0.9479, "step": 1451 }, { "epoch": 1.821831869510665, "grad_norm": 0.279296875, "learning_rate": 3.927227101631117e-06, "loss": 0.9284, "step": 1452 }, { "epoch": 1.823086574654956, "grad_norm": 0.271484375, "learning_rate": 3.9230447511501465e-06, "loss": 0.8826, "step": 1453 }, { "epoch": 1.8243412797992473, "grad_norm": 0.2890625, "learning_rate": 3.918862400669177e-06, "loss": 1.0005, "step": 1454 }, { "epoch": 1.8255959849435381, "grad_norm": 0.279296875, "learning_rate": 3.914680050188206e-06, "loss": 0.9269, "step": 1455 }, { "epoch": 1.8268506900878294, "grad_norm": 0.279296875, "learning_rate": 3.9104976997072356e-06, "loss": 0.9706, "step": 1456 }, { "epoch": 1.8281053952321205, "grad_norm": 0.34375, "learning_rate": 3.906315349226265e-06, "loss": 0.921, "step": 1457 }, { "epoch": 1.8293601003764115, "grad_norm": 0.275390625, "learning_rate": 3.902132998745295e-06, "loss": 0.9356, "step": 1458 }, { "epoch": 1.8306148055207028, "grad_norm": 0.267578125, "learning_rate": 3.897950648264325e-06, "loss": 0.8792, "step": 1459 }, { "epoch": 1.8318695106649936, "grad_norm": 0.26953125, "learning_rate": 3.893768297783354e-06, "loss": 0.987, "step": 1460 }, { "epoch": 1.8331242158092849, "grad_norm": 0.265625, "learning_rate": 3.889585947302385e-06, "loss": 0.9256, "step": 1461 }, { "epoch": 1.834378920953576, "grad_norm": 0.26953125, "learning_rate": 3.885403596821414e-06, "loss": 0.9213, "step": 1462 }, { "epoch": 1.835633626097867, "grad_norm": 0.27734375, "learning_rate": 3.881221246340443e-06, "loss": 0.9403, "step": 1463 }, { "epoch": 1.8368883312421582, "grad_norm": 0.271484375, "learning_rate": 3.877038895859474e-06, "loss": 0.9117, "step": 1464 }, { "epoch": 1.838143036386449, "grad_norm": 0.2890625, "learning_rate": 3.8728565453785035e-06, "loss": 0.9553, "step": 1465 }, { "epoch": 1.8393977415307403, "grad_norm": 0.263671875, "learning_rate": 3.868674194897532e-06, "loss": 0.9168, "step": 1466 }, { "epoch": 1.8406524466750314, "grad_norm": 0.291015625, "learning_rate": 3.864491844416563e-06, "loss": 0.9413, "step": 1467 }, { "epoch": 1.8419071518193224, "grad_norm": 0.2734375, "learning_rate": 3.8603094939355925e-06, "loss": 0.9393, "step": 1468 }, { "epoch": 1.8431618569636137, "grad_norm": 0.283203125, "learning_rate": 3.856127143454621e-06, "loss": 0.9713, "step": 1469 }, { "epoch": 1.8444165621079045, "grad_norm": 0.271484375, "learning_rate": 3.851944792973651e-06, "loss": 0.8793, "step": 1470 }, { "epoch": 1.8456712672521958, "grad_norm": 0.265625, "learning_rate": 3.8477624424926815e-06, "loss": 0.9354, "step": 1471 }, { "epoch": 1.8469259723964868, "grad_norm": 0.265625, "learning_rate": 3.843580092011711e-06, "loss": 0.9466, "step": 1472 }, { "epoch": 1.8481806775407779, "grad_norm": 0.275390625, "learning_rate": 3.83939774153074e-06, "loss": 0.9761, "step": 1473 }, { "epoch": 1.8494353826850691, "grad_norm": 0.271484375, "learning_rate": 3.8352153910497706e-06, "loss": 0.9271, "step": 1474 }, { "epoch": 1.85069008782936, "grad_norm": 0.318359375, "learning_rate": 3.8310330405688e-06, "loss": 0.9344, "step": 1475 }, { "epoch": 1.8519447929736512, "grad_norm": 0.310546875, "learning_rate": 3.82685069008783e-06, "loss": 0.9382, "step": 1476 }, { "epoch": 1.8531994981179423, "grad_norm": 0.294921875, "learning_rate": 3.82266833960686e-06, "loss": 0.8989, "step": 1477 }, { "epoch": 1.8544542032622333, "grad_norm": 0.2734375, "learning_rate": 3.818485989125889e-06, "loss": 0.9743, "step": 1478 }, { "epoch": 1.8557089084065246, "grad_norm": 0.28125, "learning_rate": 3.814303638644919e-06, "loss": 0.9673, "step": 1479 }, { "epoch": 1.8569636135508154, "grad_norm": 0.26953125, "learning_rate": 3.810121288163948e-06, "loss": 0.9257, "step": 1480 }, { "epoch": 1.8582183186951067, "grad_norm": 0.2734375, "learning_rate": 3.805938937682978e-06, "loss": 0.9229, "step": 1481 }, { "epoch": 1.8594730238393977, "grad_norm": 0.283203125, "learning_rate": 3.801756587202008e-06, "loss": 0.9306, "step": 1482 }, { "epoch": 1.8607277289836888, "grad_norm": 0.283203125, "learning_rate": 3.7975742367210376e-06, "loss": 0.9952, "step": 1483 }, { "epoch": 1.86198243412798, "grad_norm": 0.263671875, "learning_rate": 3.793391886240067e-06, "loss": 0.9307, "step": 1484 }, { "epoch": 1.8632371392722709, "grad_norm": 0.294921875, "learning_rate": 3.789209535759097e-06, "loss": 0.8737, "step": 1485 }, { "epoch": 1.8644918444165621, "grad_norm": 0.28125, "learning_rate": 3.7850271852781267e-06, "loss": 0.9631, "step": 1486 }, { "epoch": 1.8657465495608532, "grad_norm": 0.26953125, "learning_rate": 3.7808448347971568e-06, "loss": 0.9198, "step": 1487 }, { "epoch": 1.8670012547051442, "grad_norm": 0.30078125, "learning_rate": 3.776662484316186e-06, "loss": 0.9251, "step": 1488 }, { "epoch": 1.8682559598494355, "grad_norm": 0.359375, "learning_rate": 3.7724801338352157e-06, "loss": 0.9101, "step": 1489 }, { "epoch": 1.8695106649937263, "grad_norm": 0.3125, "learning_rate": 3.7682977833542454e-06, "loss": 0.9503, "step": 1490 }, { "epoch": 1.8707653701380176, "grad_norm": 0.2890625, "learning_rate": 3.764115432873275e-06, "loss": 1.0051, "step": 1491 }, { "epoch": 1.8720200752823086, "grad_norm": 0.28125, "learning_rate": 3.7599330823923047e-06, "loss": 0.8889, "step": 1492 }, { "epoch": 1.8732747804265997, "grad_norm": 0.3125, "learning_rate": 3.7557507319113344e-06, "loss": 0.9419, "step": 1493 }, { "epoch": 1.874529485570891, "grad_norm": 0.267578125, "learning_rate": 3.7515683814303645e-06, "loss": 0.9166, "step": 1494 }, { "epoch": 1.875784190715182, "grad_norm": 0.2890625, "learning_rate": 3.7473860309493937e-06, "loss": 1.0235, "step": 1495 }, { "epoch": 1.877038895859473, "grad_norm": 0.279296875, "learning_rate": 3.7432036804684234e-06, "loss": 0.9373, "step": 1496 }, { "epoch": 1.878293601003764, "grad_norm": 0.283203125, "learning_rate": 3.7390213299874535e-06, "loss": 0.9332, "step": 1497 }, { "epoch": 1.8795483061480551, "grad_norm": 0.271484375, "learning_rate": 3.734838979506483e-06, "loss": 0.9157, "step": 1498 }, { "epoch": 1.8808030112923464, "grad_norm": 0.330078125, "learning_rate": 3.7306566290255125e-06, "loss": 0.9528, "step": 1499 }, { "epoch": 1.8820577164366374, "grad_norm": 0.26953125, "learning_rate": 3.7264742785445425e-06, "loss": 0.9361, "step": 1500 }, { "epoch": 1.8833124215809285, "grad_norm": 0.283203125, "learning_rate": 3.7222919280635722e-06, "loss": 0.9125, "step": 1501 }, { "epoch": 1.8845671267252195, "grad_norm": 0.27734375, "learning_rate": 3.7181095775826015e-06, "loss": 0.9544, "step": 1502 }, { "epoch": 1.8858218318695106, "grad_norm": 0.26953125, "learning_rate": 3.713927227101631e-06, "loss": 0.9149, "step": 1503 }, { "epoch": 1.8870765370138018, "grad_norm": 0.2734375, "learning_rate": 3.7097448766206613e-06, "loss": 0.939, "step": 1504 }, { "epoch": 1.888331242158093, "grad_norm": 0.267578125, "learning_rate": 3.705562526139691e-06, "loss": 0.9155, "step": 1505 }, { "epoch": 1.889585947302384, "grad_norm": 0.279296875, "learning_rate": 3.70138017565872e-06, "loss": 0.9411, "step": 1506 }, { "epoch": 1.890840652446675, "grad_norm": 0.271484375, "learning_rate": 3.6971978251777503e-06, "loss": 0.9604, "step": 1507 }, { "epoch": 1.892095357590966, "grad_norm": 0.3515625, "learning_rate": 3.69301547469678e-06, "loss": 0.8965, "step": 1508 }, { "epoch": 1.8933500627352573, "grad_norm": 0.26953125, "learning_rate": 3.68883312421581e-06, "loss": 0.9021, "step": 1509 }, { "epoch": 1.8946047678795483, "grad_norm": 0.265625, "learning_rate": 3.6846507737348393e-06, "loss": 0.9475, "step": 1510 }, { "epoch": 1.8958594730238394, "grad_norm": 0.30078125, "learning_rate": 3.680468423253869e-06, "loss": 0.9988, "step": 1511 }, { "epoch": 1.8971141781681304, "grad_norm": 0.267578125, "learning_rate": 3.6762860727728987e-06, "loss": 0.897, "step": 1512 }, { "epoch": 1.8983688833124215, "grad_norm": 0.27734375, "learning_rate": 3.6721037222919283e-06, "loss": 0.9427, "step": 1513 }, { "epoch": 1.8996235884567128, "grad_norm": 0.267578125, "learning_rate": 3.667921371810958e-06, "loss": 0.9603, "step": 1514 }, { "epoch": 1.9008782936010038, "grad_norm": 0.2734375, "learning_rate": 3.6637390213299877e-06, "loss": 0.9369, "step": 1515 }, { "epoch": 1.9021329987452948, "grad_norm": 0.279296875, "learning_rate": 3.6595566708490178e-06, "loss": 0.8856, "step": 1516 }, { "epoch": 1.9033877038895861, "grad_norm": 0.275390625, "learning_rate": 3.655374320368047e-06, "loss": 0.9776, "step": 1517 }, { "epoch": 1.904642409033877, "grad_norm": 0.26953125, "learning_rate": 3.6511919698870767e-06, "loss": 0.9528, "step": 1518 }, { "epoch": 1.9058971141781682, "grad_norm": 0.267578125, "learning_rate": 3.647009619406107e-06, "loss": 0.934, "step": 1519 }, { "epoch": 1.9071518193224593, "grad_norm": 0.26953125, "learning_rate": 3.6428272689251365e-06, "loss": 0.9267, "step": 1520 }, { "epoch": 1.9084065244667503, "grad_norm": 0.271484375, "learning_rate": 3.6386449184441657e-06, "loss": 0.9586, "step": 1521 }, { "epoch": 1.9096612296110416, "grad_norm": 0.271484375, "learning_rate": 3.634462567963196e-06, "loss": 0.9031, "step": 1522 }, { "epoch": 1.9109159347553324, "grad_norm": 0.275390625, "learning_rate": 3.6302802174822255e-06, "loss": 0.9004, "step": 1523 }, { "epoch": 1.9121706398996237, "grad_norm": 0.271484375, "learning_rate": 3.6260978670012548e-06, "loss": 0.9467, "step": 1524 }, { "epoch": 1.9134253450439147, "grad_norm": 0.2734375, "learning_rate": 3.6219155165202844e-06, "loss": 0.9402, "step": 1525 }, { "epoch": 1.9146800501882058, "grad_norm": 0.4453125, "learning_rate": 3.6177331660393145e-06, "loss": 0.9974, "step": 1526 }, { "epoch": 1.915934755332497, "grad_norm": 0.267578125, "learning_rate": 3.6135508155583442e-06, "loss": 0.8698, "step": 1527 }, { "epoch": 1.9171894604767878, "grad_norm": 0.58203125, "learning_rate": 3.6093684650773735e-06, "loss": 0.9852, "step": 1528 }, { "epoch": 1.9184441656210791, "grad_norm": 0.296875, "learning_rate": 3.6051861145964036e-06, "loss": 0.943, "step": 1529 }, { "epoch": 1.9196988707653702, "grad_norm": 0.287109375, "learning_rate": 3.6010037641154332e-06, "loss": 0.9233, "step": 1530 }, { "epoch": 1.9209535759096612, "grad_norm": 0.2734375, "learning_rate": 3.5968214136344633e-06, "loss": 0.9579, "step": 1531 }, { "epoch": 1.9222082810539525, "grad_norm": 0.275390625, "learning_rate": 3.5926390631534926e-06, "loss": 0.9481, "step": 1532 }, { "epoch": 1.9234629861982433, "grad_norm": 0.287109375, "learning_rate": 3.5884567126725223e-06, "loss": 0.9663, "step": 1533 }, { "epoch": 1.9247176913425346, "grad_norm": 0.26953125, "learning_rate": 3.584274362191552e-06, "loss": 0.9591, "step": 1534 }, { "epoch": 1.9259723964868256, "grad_norm": 0.28125, "learning_rate": 3.580092011710581e-06, "loss": 0.9308, "step": 1535 }, { "epoch": 1.9272271016311167, "grad_norm": 0.283203125, "learning_rate": 3.5759096612296113e-06, "loss": 0.964, "step": 1536 }, { "epoch": 1.928481806775408, "grad_norm": 0.294921875, "learning_rate": 3.571727310748641e-06, "loss": 0.9608, "step": 1537 }, { "epoch": 1.9297365119196987, "grad_norm": 0.2734375, "learning_rate": 3.567544960267671e-06, "loss": 0.9336, "step": 1538 }, { "epoch": 1.93099121706399, "grad_norm": 0.283203125, "learning_rate": 3.5633626097867003e-06, "loss": 0.9125, "step": 1539 }, { "epoch": 1.932245922208281, "grad_norm": 0.2734375, "learning_rate": 3.55918025930573e-06, "loss": 0.9746, "step": 1540 }, { "epoch": 1.933500627352572, "grad_norm": 0.267578125, "learning_rate": 3.55499790882476e-06, "loss": 0.9203, "step": 1541 }, { "epoch": 1.9347553324968634, "grad_norm": 0.267578125, "learning_rate": 3.5508155583437898e-06, "loss": 0.9139, "step": 1542 }, { "epoch": 1.9360100376411542, "grad_norm": 0.271484375, "learning_rate": 3.546633207862819e-06, "loss": 0.9569, "step": 1543 }, { "epoch": 1.9372647427854455, "grad_norm": 0.27734375, "learning_rate": 3.5424508573818487e-06, "loss": 0.9864, "step": 1544 }, { "epoch": 1.9385194479297365, "grad_norm": 0.267578125, "learning_rate": 3.538268506900879e-06, "loss": 0.917, "step": 1545 }, { "epoch": 1.9397741530740276, "grad_norm": 0.27734375, "learning_rate": 3.534086156419908e-06, "loss": 0.947, "step": 1546 }, { "epoch": 1.9410288582183188, "grad_norm": 0.279296875, "learning_rate": 3.5299038059389377e-06, "loss": 0.9044, "step": 1547 }, { "epoch": 1.9422835633626097, "grad_norm": 0.314453125, "learning_rate": 3.525721455457968e-06, "loss": 0.9698, "step": 1548 }, { "epoch": 1.943538268506901, "grad_norm": 0.349609375, "learning_rate": 3.5215391049769975e-06, "loss": 0.9457, "step": 1549 }, { "epoch": 1.944792973651192, "grad_norm": 0.28515625, "learning_rate": 3.5173567544960268e-06, "loss": 0.9858, "step": 1550 }, { "epoch": 1.946047678795483, "grad_norm": 0.28125, "learning_rate": 3.513174404015057e-06, "loss": 0.9111, "step": 1551 }, { "epoch": 1.9473023839397743, "grad_norm": 0.2734375, "learning_rate": 3.5089920535340865e-06, "loss": 0.9864, "step": 1552 }, { "epoch": 1.948557089084065, "grad_norm": 0.265625, "learning_rate": 3.504809703053116e-06, "loss": 0.9221, "step": 1553 }, { "epoch": 1.9498117942283564, "grad_norm": 0.36328125, "learning_rate": 3.500627352572146e-06, "loss": 0.9919, "step": 1554 }, { "epoch": 1.9510664993726474, "grad_norm": 0.27734375, "learning_rate": 3.4964450020911756e-06, "loss": 0.904, "step": 1555 }, { "epoch": 1.9523212045169385, "grad_norm": 0.275390625, "learning_rate": 3.4922626516102052e-06, "loss": 0.9585, "step": 1556 }, { "epoch": 1.9535759096612297, "grad_norm": 0.279296875, "learning_rate": 3.4880803011292345e-06, "loss": 0.9008, "step": 1557 }, { "epoch": 1.9548306148055206, "grad_norm": 0.2734375, "learning_rate": 3.4838979506482646e-06, "loss": 0.9285, "step": 1558 }, { "epoch": 1.9560853199498118, "grad_norm": 0.279296875, "learning_rate": 3.4797156001672943e-06, "loss": 0.9014, "step": 1559 }, { "epoch": 1.9573400250941029, "grad_norm": 0.28125, "learning_rate": 3.4755332496863244e-06, "loss": 0.8824, "step": 1560 }, { "epoch": 1.958594730238394, "grad_norm": 0.27734375, "learning_rate": 3.4713508992053536e-06, "loss": 0.9654, "step": 1561 }, { "epoch": 1.9598494353826852, "grad_norm": 0.26953125, "learning_rate": 3.4671685487243833e-06, "loss": 0.915, "step": 1562 }, { "epoch": 1.961104140526976, "grad_norm": 0.2734375, "learning_rate": 3.4629861982434134e-06, "loss": 0.9742, "step": 1563 }, { "epoch": 1.9623588456712673, "grad_norm": 0.36328125, "learning_rate": 3.458803847762443e-06, "loss": 0.9617, "step": 1564 }, { "epoch": 1.9636135508155583, "grad_norm": 0.28515625, "learning_rate": 3.4546214972814723e-06, "loss": 0.9559, "step": 1565 }, { "epoch": 1.9648682559598494, "grad_norm": 0.275390625, "learning_rate": 3.450439146800502e-06, "loss": 0.942, "step": 1566 }, { "epoch": 1.9661229611041406, "grad_norm": 0.2890625, "learning_rate": 3.446256796319532e-06, "loss": 0.9331, "step": 1567 }, { "epoch": 1.9673776662484315, "grad_norm": 0.275390625, "learning_rate": 3.4420744458385613e-06, "loss": 0.9179, "step": 1568 }, { "epoch": 1.9686323713927227, "grad_norm": 0.27734375, "learning_rate": 3.437892095357591e-06, "loss": 0.9025, "step": 1569 }, { "epoch": 1.9698870765370138, "grad_norm": 0.30078125, "learning_rate": 3.433709744876621e-06, "loss": 0.9602, "step": 1570 }, { "epoch": 1.9711417816813048, "grad_norm": 0.310546875, "learning_rate": 3.4295273943956508e-06, "loss": 0.9266, "step": 1571 }, { "epoch": 1.972396486825596, "grad_norm": 0.283203125, "learning_rate": 3.42534504391468e-06, "loss": 0.974, "step": 1572 }, { "epoch": 1.973651191969887, "grad_norm": 0.271484375, "learning_rate": 3.42116269343371e-06, "loss": 0.9219, "step": 1573 }, { "epoch": 1.9749058971141782, "grad_norm": 0.275390625, "learning_rate": 3.41698034295274e-06, "loss": 0.9559, "step": 1574 }, { "epoch": 1.9761606022584692, "grad_norm": 0.3046875, "learning_rate": 3.4127979924717695e-06, "loss": 0.9924, "step": 1575 }, { "epoch": 1.9774153074027603, "grad_norm": 0.275390625, "learning_rate": 3.408615641990799e-06, "loss": 0.968, "step": 1576 }, { "epoch": 1.9786700125470515, "grad_norm": 0.291015625, "learning_rate": 3.404433291509829e-06, "loss": 0.9398, "step": 1577 }, { "epoch": 1.9799247176913424, "grad_norm": 0.275390625, "learning_rate": 3.4002509410288585e-06, "loss": 0.9453, "step": 1578 }, { "epoch": 1.9811794228356336, "grad_norm": 0.2734375, "learning_rate": 3.3960685905478878e-06, "loss": 0.8941, "step": 1579 }, { "epoch": 1.9824341279799247, "grad_norm": 0.32421875, "learning_rate": 3.391886240066918e-06, "loss": 0.9539, "step": 1580 }, { "epoch": 1.9836888331242157, "grad_norm": 0.306640625, "learning_rate": 3.3877038895859475e-06, "loss": 0.921, "step": 1581 }, { "epoch": 1.984943538268507, "grad_norm": 0.271484375, "learning_rate": 3.3835215391049776e-06, "loss": 0.9377, "step": 1582 }, { "epoch": 1.986198243412798, "grad_norm": 0.263671875, "learning_rate": 3.379339188624007e-06, "loss": 0.8842, "step": 1583 }, { "epoch": 1.987452948557089, "grad_norm": 0.2734375, "learning_rate": 3.3751568381430366e-06, "loss": 0.9436, "step": 1584 }, { "epoch": 1.9887076537013801, "grad_norm": 0.275390625, "learning_rate": 3.3709744876620667e-06, "loss": 0.9278, "step": 1585 }, { "epoch": 1.9899623588456712, "grad_norm": 0.26953125, "learning_rate": 3.3667921371810963e-06, "loss": 0.9401, "step": 1586 }, { "epoch": 1.9912170639899625, "grad_norm": 0.30078125, "learning_rate": 3.3626097867001256e-06, "loss": 0.9362, "step": 1587 }, { "epoch": 1.9924717691342535, "grad_norm": 0.28125, "learning_rate": 3.3584274362191553e-06, "loss": 0.9098, "step": 1588 }, { "epoch": 1.9937264742785445, "grad_norm": 0.27734375, "learning_rate": 3.3542450857381854e-06, "loss": 0.9551, "step": 1589 }, { "epoch": 1.9949811794228356, "grad_norm": 0.29296875, "learning_rate": 3.3500627352572146e-06, "loss": 1.0057, "step": 1590 }, { "epoch": 1.9962358845671266, "grad_norm": 0.267578125, "learning_rate": 3.3458803847762443e-06, "loss": 0.9282, "step": 1591 }, { "epoch": 1.997490589711418, "grad_norm": 0.267578125, "learning_rate": 3.3416980342952744e-06, "loss": 0.9065, "step": 1592 }, { "epoch": 1.998745294855709, "grad_norm": 0.365234375, "learning_rate": 3.337515683814304e-06, "loss": 0.8831, "step": 1593 }, { "epoch": 2.0, "grad_norm": 0.2734375, "learning_rate": 3.3333333333333333e-06, "loss": 0.9529, "step": 1594 }, { "epoch": 2.0012547051442913, "grad_norm": 0.275390625, "learning_rate": 3.3291509828523634e-06, "loss": 0.9623, "step": 1595 }, { "epoch": 2.002509410288582, "grad_norm": 0.296875, "learning_rate": 3.324968632371393e-06, "loss": 0.9784, "step": 1596 }, { "epoch": 2.0037641154328734, "grad_norm": 0.27734375, "learning_rate": 3.3207862818904228e-06, "loss": 0.938, "step": 1597 }, { "epoch": 2.005018820577164, "grad_norm": 0.263671875, "learning_rate": 3.3166039314094524e-06, "loss": 0.9365, "step": 1598 }, { "epoch": 2.0062735257214555, "grad_norm": 0.267578125, "learning_rate": 3.312421580928482e-06, "loss": 0.9186, "step": 1599 }, { "epoch": 2.0075282308657467, "grad_norm": 0.287109375, "learning_rate": 3.308239230447512e-06, "loss": 0.96, "step": 1600 }, { "epoch": 2.0087829360100375, "grad_norm": 0.267578125, "learning_rate": 3.304056879966541e-06, "loss": 0.936, "step": 1601 }, { "epoch": 2.010037641154329, "grad_norm": 0.27734375, "learning_rate": 3.299874529485571e-06, "loss": 0.9302, "step": 1602 }, { "epoch": 2.0112923462986196, "grad_norm": 0.263671875, "learning_rate": 3.295692179004601e-06, "loss": 0.9432, "step": 1603 }, { "epoch": 2.012547051442911, "grad_norm": 0.275390625, "learning_rate": 3.291509828523631e-06, "loss": 0.8779, "step": 1604 }, { "epoch": 2.013801756587202, "grad_norm": 0.265625, "learning_rate": 3.28732747804266e-06, "loss": 0.9348, "step": 1605 }, { "epoch": 2.015056461731493, "grad_norm": 0.2734375, "learning_rate": 3.28314512756169e-06, "loss": 0.9632, "step": 1606 }, { "epoch": 2.0163111668757843, "grad_norm": 0.271484375, "learning_rate": 3.27896277708072e-06, "loss": 0.9308, "step": 1607 }, { "epoch": 2.017565872020075, "grad_norm": 0.267578125, "learning_rate": 3.2747804265997496e-06, "loss": 0.916, "step": 1608 }, { "epoch": 2.0188205771643664, "grad_norm": 0.26953125, "learning_rate": 3.270598076118779e-06, "loss": 0.9117, "step": 1609 }, { "epoch": 2.0200752823086576, "grad_norm": 0.28125, "learning_rate": 3.2664157256378086e-06, "loss": 0.9095, "step": 1610 }, { "epoch": 2.0213299874529485, "grad_norm": 0.26953125, "learning_rate": 3.2622333751568387e-06, "loss": 0.9232, "step": 1611 }, { "epoch": 2.0225846925972397, "grad_norm": 0.291015625, "learning_rate": 3.258051024675868e-06, "loss": 0.942, "step": 1612 }, { "epoch": 2.0238393977415305, "grad_norm": 0.3125, "learning_rate": 3.2538686741948976e-06, "loss": 0.9023, "step": 1613 }, { "epoch": 2.025094102885822, "grad_norm": 0.279296875, "learning_rate": 3.2496863237139277e-06, "loss": 0.9309, "step": 1614 }, { "epoch": 2.026348808030113, "grad_norm": 0.28125, "learning_rate": 3.2455039732329574e-06, "loss": 0.9008, "step": 1615 }, { "epoch": 2.027603513174404, "grad_norm": 0.28125, "learning_rate": 3.2413216227519866e-06, "loss": 0.9268, "step": 1616 }, { "epoch": 2.028858218318695, "grad_norm": 0.30859375, "learning_rate": 3.2371392722710167e-06, "loss": 0.9224, "step": 1617 }, { "epoch": 2.030112923462986, "grad_norm": 0.2734375, "learning_rate": 3.2329569217900464e-06, "loss": 0.9689, "step": 1618 }, { "epoch": 2.0313676286072773, "grad_norm": 0.27734375, "learning_rate": 3.228774571309076e-06, "loss": 0.9157, "step": 1619 }, { "epoch": 2.0326223337515685, "grad_norm": 0.2890625, "learning_rate": 3.2245922208281057e-06, "loss": 0.9584, "step": 1620 }, { "epoch": 2.0338770388958594, "grad_norm": 0.310546875, "learning_rate": 3.2204098703471354e-06, "loss": 0.9611, "step": 1621 }, { "epoch": 2.0351317440401506, "grad_norm": 0.279296875, "learning_rate": 3.216227519866165e-06, "loss": 0.9673, "step": 1622 }, { "epoch": 2.0363864491844414, "grad_norm": 0.28125, "learning_rate": 3.2120451693851943e-06, "loss": 0.938, "step": 1623 }, { "epoch": 2.0376411543287327, "grad_norm": 0.267578125, "learning_rate": 3.2078628189042244e-06, "loss": 0.915, "step": 1624 }, { "epoch": 2.038895859473024, "grad_norm": 0.283203125, "learning_rate": 3.203680468423254e-06, "loss": 0.8942, "step": 1625 }, { "epoch": 2.040150564617315, "grad_norm": 0.267578125, "learning_rate": 3.1994981179422842e-06, "loss": 0.9241, "step": 1626 }, { "epoch": 2.041405269761606, "grad_norm": 0.27734375, "learning_rate": 3.1953157674613135e-06, "loss": 0.9322, "step": 1627 }, { "epoch": 2.042659974905897, "grad_norm": 0.275390625, "learning_rate": 3.191133416980343e-06, "loss": 0.9256, "step": 1628 }, { "epoch": 2.043914680050188, "grad_norm": 0.306640625, "learning_rate": 3.186951066499373e-06, "loss": 0.9672, "step": 1629 }, { "epoch": 2.0451693851944794, "grad_norm": 0.283203125, "learning_rate": 3.182768716018403e-06, "loss": 0.9277, "step": 1630 }, { "epoch": 2.0464240903387703, "grad_norm": 0.279296875, "learning_rate": 3.178586365537432e-06, "loss": 0.9673, "step": 1631 }, { "epoch": 2.0476787954830615, "grad_norm": 0.279296875, "learning_rate": 3.174404015056462e-06, "loss": 0.9482, "step": 1632 }, { "epoch": 2.0489335006273524, "grad_norm": 0.2734375, "learning_rate": 3.170221664575492e-06, "loss": 0.8869, "step": 1633 }, { "epoch": 2.0501882057716436, "grad_norm": 0.275390625, "learning_rate": 3.166039314094521e-06, "loss": 0.9291, "step": 1634 }, { "epoch": 2.051442910915935, "grad_norm": 0.3046875, "learning_rate": 3.161856963613551e-06, "loss": 0.8963, "step": 1635 }, { "epoch": 2.0526976160602257, "grad_norm": 0.302734375, "learning_rate": 3.157674613132581e-06, "loss": 0.9212, "step": 1636 }, { "epoch": 2.053952321204517, "grad_norm": 0.26953125, "learning_rate": 3.1534922626516106e-06, "loss": 0.92, "step": 1637 }, { "epoch": 2.055207026348808, "grad_norm": 0.291015625, "learning_rate": 3.14930991217064e-06, "loss": 0.9675, "step": 1638 }, { "epoch": 2.056461731493099, "grad_norm": 0.376953125, "learning_rate": 3.14512756168967e-06, "loss": 0.954, "step": 1639 }, { "epoch": 2.0577164366373903, "grad_norm": 0.27734375, "learning_rate": 3.1409452112086997e-06, "loss": 0.9341, "step": 1640 }, { "epoch": 2.058971141781681, "grad_norm": 0.2734375, "learning_rate": 3.1367628607277293e-06, "loss": 0.9201, "step": 1641 }, { "epoch": 2.0602258469259724, "grad_norm": 0.31640625, "learning_rate": 3.1325805102467586e-06, "loss": 0.992, "step": 1642 }, { "epoch": 2.0614805520702637, "grad_norm": 0.3359375, "learning_rate": 3.1283981597657887e-06, "loss": 0.8934, "step": 1643 }, { "epoch": 2.0627352572145545, "grad_norm": 0.267578125, "learning_rate": 3.1242158092848184e-06, "loss": 0.9337, "step": 1644 }, { "epoch": 2.063989962358846, "grad_norm": 0.3515625, "learning_rate": 3.1200334588038476e-06, "loss": 0.8815, "step": 1645 }, { "epoch": 2.0652446675031366, "grad_norm": 0.271484375, "learning_rate": 3.1158511083228777e-06, "loss": 0.9087, "step": 1646 }, { "epoch": 2.066499372647428, "grad_norm": 0.27734375, "learning_rate": 3.1116687578419074e-06, "loss": 0.962, "step": 1647 }, { "epoch": 2.067754077791719, "grad_norm": 0.275390625, "learning_rate": 3.1074864073609375e-06, "loss": 0.935, "step": 1648 }, { "epoch": 2.06900878293601, "grad_norm": 0.275390625, "learning_rate": 3.1033040568799668e-06, "loss": 0.9767, "step": 1649 }, { "epoch": 2.0702634880803013, "grad_norm": 0.283203125, "learning_rate": 3.0991217063989964e-06, "loss": 0.9157, "step": 1650 }, { "epoch": 2.071518193224592, "grad_norm": 0.26953125, "learning_rate": 3.094939355918026e-06, "loss": 0.935, "step": 1651 }, { "epoch": 2.0727728983688833, "grad_norm": 0.27734375, "learning_rate": 3.090757005437056e-06, "loss": 0.9686, "step": 1652 }, { "epoch": 2.0740276035131746, "grad_norm": 0.2734375, "learning_rate": 3.0865746549560855e-06, "loss": 0.9041, "step": 1653 }, { "epoch": 2.0752823086574654, "grad_norm": 0.287109375, "learning_rate": 3.082392304475115e-06, "loss": 0.9478, "step": 1654 }, { "epoch": 2.0765370138017567, "grad_norm": 0.271484375, "learning_rate": 3.0782099539941452e-06, "loss": 0.9406, "step": 1655 }, { "epoch": 2.0777917189460475, "grad_norm": 0.419921875, "learning_rate": 3.0740276035131745e-06, "loss": 0.8974, "step": 1656 }, { "epoch": 2.079046424090339, "grad_norm": 0.283203125, "learning_rate": 3.069845253032204e-06, "loss": 0.948, "step": 1657 }, { "epoch": 2.08030112923463, "grad_norm": 0.369140625, "learning_rate": 3.0656629025512343e-06, "loss": 0.931, "step": 1658 }, { "epoch": 2.081555834378921, "grad_norm": 0.26953125, "learning_rate": 3.061480552070264e-06, "loss": 0.9175, "step": 1659 }, { "epoch": 2.082810539523212, "grad_norm": 0.279296875, "learning_rate": 3.057298201589293e-06, "loss": 0.9309, "step": 1660 }, { "epoch": 2.084065244667503, "grad_norm": 0.27734375, "learning_rate": 3.0531158511083233e-06, "loss": 0.912, "step": 1661 }, { "epoch": 2.0853199498117942, "grad_norm": 0.275390625, "learning_rate": 3.048933500627353e-06, "loss": 0.9304, "step": 1662 }, { "epoch": 2.0865746549560855, "grad_norm": 0.287109375, "learning_rate": 3.0447511501463826e-06, "loss": 0.9188, "step": 1663 }, { "epoch": 2.0878293601003763, "grad_norm": 0.27734375, "learning_rate": 3.040568799665412e-06, "loss": 0.9196, "step": 1664 }, { "epoch": 2.0890840652446676, "grad_norm": 0.2734375, "learning_rate": 3.036386449184442e-06, "loss": 0.8954, "step": 1665 }, { "epoch": 2.0903387703889584, "grad_norm": 0.2890625, "learning_rate": 3.0322040987034717e-06, "loss": 0.9835, "step": 1666 }, { "epoch": 2.0915934755332497, "grad_norm": 0.2734375, "learning_rate": 3.028021748222501e-06, "loss": 0.9238, "step": 1667 }, { "epoch": 2.092848180677541, "grad_norm": 0.2890625, "learning_rate": 3.023839397741531e-06, "loss": 0.9483, "step": 1668 }, { "epoch": 2.094102885821832, "grad_norm": 0.275390625, "learning_rate": 3.0196570472605607e-06, "loss": 0.942, "step": 1669 }, { "epoch": 2.095357590966123, "grad_norm": 0.287109375, "learning_rate": 3.0154746967795908e-06, "loss": 0.9421, "step": 1670 }, { "epoch": 2.096612296110414, "grad_norm": 0.2890625, "learning_rate": 3.01129234629862e-06, "loss": 0.8682, "step": 1671 }, { "epoch": 2.097867001254705, "grad_norm": 0.302734375, "learning_rate": 3.0071099958176497e-06, "loss": 0.9263, "step": 1672 }, { "epoch": 2.0991217063989964, "grad_norm": 0.302734375, "learning_rate": 3.0029276453366794e-06, "loss": 0.9136, "step": 1673 }, { "epoch": 2.1003764115432872, "grad_norm": 0.333984375, "learning_rate": 2.9987452948557095e-06, "loss": 0.9032, "step": 1674 }, { "epoch": 2.1016311166875785, "grad_norm": 0.265625, "learning_rate": 2.9945629443747387e-06, "loss": 0.8901, "step": 1675 }, { "epoch": 2.1028858218318693, "grad_norm": 0.294921875, "learning_rate": 2.9903805938937684e-06, "loss": 0.9397, "step": 1676 }, { "epoch": 2.1041405269761606, "grad_norm": 0.275390625, "learning_rate": 2.9861982434127985e-06, "loss": 0.87, "step": 1677 }, { "epoch": 2.105395232120452, "grad_norm": 0.2890625, "learning_rate": 2.9820158929318278e-06, "loss": 0.9571, "step": 1678 }, { "epoch": 2.1066499372647427, "grad_norm": 0.283203125, "learning_rate": 2.9778335424508574e-06, "loss": 0.9552, "step": 1679 }, { "epoch": 2.107904642409034, "grad_norm": 0.287109375, "learning_rate": 2.9736511919698875e-06, "loss": 0.9776, "step": 1680 }, { "epoch": 2.109159347553325, "grad_norm": 0.2734375, "learning_rate": 2.9694688414889172e-06, "loss": 0.8905, "step": 1681 }, { "epoch": 2.110414052697616, "grad_norm": 0.2734375, "learning_rate": 2.9652864910079465e-06, "loss": 0.9547, "step": 1682 }, { "epoch": 2.1116687578419073, "grad_norm": 0.318359375, "learning_rate": 2.9611041405269766e-06, "loss": 0.9479, "step": 1683 }, { "epoch": 2.112923462986198, "grad_norm": 0.271484375, "learning_rate": 2.9569217900460062e-06, "loss": 0.9305, "step": 1684 }, { "epoch": 2.1141781681304894, "grad_norm": 0.29296875, "learning_rate": 2.952739439565036e-06, "loss": 0.98, "step": 1685 }, { "epoch": 2.1154328732747802, "grad_norm": 0.26953125, "learning_rate": 2.948557089084065e-06, "loss": 0.9233, "step": 1686 }, { "epoch": 2.1166875784190715, "grad_norm": 0.365234375, "learning_rate": 2.9443747386030953e-06, "loss": 0.9626, "step": 1687 }, { "epoch": 2.117942283563363, "grad_norm": 0.271484375, "learning_rate": 2.940192388122125e-06, "loss": 0.9446, "step": 1688 }, { "epoch": 2.1191969887076536, "grad_norm": 0.287109375, "learning_rate": 2.936010037641154e-06, "loss": 0.9927, "step": 1689 }, { "epoch": 2.120451693851945, "grad_norm": 0.287109375, "learning_rate": 2.9318276871601843e-06, "loss": 0.9906, "step": 1690 }, { "epoch": 2.1217063989962357, "grad_norm": 0.279296875, "learning_rate": 2.927645336679214e-06, "loss": 0.9387, "step": 1691 }, { "epoch": 2.122961104140527, "grad_norm": 0.294921875, "learning_rate": 2.923462986198244e-06, "loss": 0.9276, "step": 1692 }, { "epoch": 2.1242158092848182, "grad_norm": 0.29296875, "learning_rate": 2.9192806357172733e-06, "loss": 1.0133, "step": 1693 }, { "epoch": 2.125470514429109, "grad_norm": 0.314453125, "learning_rate": 2.915098285236303e-06, "loss": 0.9509, "step": 1694 }, { "epoch": 2.1267252195734003, "grad_norm": 0.294921875, "learning_rate": 2.9109159347553327e-06, "loss": 0.9625, "step": 1695 }, { "epoch": 2.127979924717691, "grad_norm": 0.283203125, "learning_rate": 2.9067335842743628e-06, "loss": 0.9183, "step": 1696 }, { "epoch": 2.1292346298619824, "grad_norm": 0.314453125, "learning_rate": 2.902551233793392e-06, "loss": 0.959, "step": 1697 }, { "epoch": 2.1304893350062737, "grad_norm": 0.283203125, "learning_rate": 2.8983688833124217e-06, "loss": 0.9793, "step": 1698 }, { "epoch": 2.1317440401505645, "grad_norm": 0.291015625, "learning_rate": 2.894186532831452e-06, "loss": 0.9179, "step": 1699 }, { "epoch": 2.132998745294856, "grad_norm": 0.302734375, "learning_rate": 2.890004182350481e-06, "loss": 0.9562, "step": 1700 }, { "epoch": 2.1342534504391466, "grad_norm": 0.30078125, "learning_rate": 2.8858218318695107e-06, "loss": 0.9402, "step": 1701 }, { "epoch": 2.135508155583438, "grad_norm": 0.28125, "learning_rate": 2.881639481388541e-06, "loss": 0.9333, "step": 1702 }, { "epoch": 2.136762860727729, "grad_norm": 0.26953125, "learning_rate": 2.8774571309075705e-06, "loss": 0.9143, "step": 1703 }, { "epoch": 2.13801756587202, "grad_norm": 0.287109375, "learning_rate": 2.8732747804265998e-06, "loss": 0.9106, "step": 1704 }, { "epoch": 2.1392722710163112, "grad_norm": 0.279296875, "learning_rate": 2.86909242994563e-06, "loss": 0.9647, "step": 1705 }, { "epoch": 2.140526976160602, "grad_norm": 0.287109375, "learning_rate": 2.8649100794646595e-06, "loss": 0.9578, "step": 1706 }, { "epoch": 2.1417816813048933, "grad_norm": 0.27734375, "learning_rate": 2.860727728983689e-06, "loss": 0.9455, "step": 1707 }, { "epoch": 2.1430363864491846, "grad_norm": 0.28125, "learning_rate": 2.8565453785027185e-06, "loss": 0.9641, "step": 1708 }, { "epoch": 2.1442910915934754, "grad_norm": 0.361328125, "learning_rate": 2.8523630280217486e-06, "loss": 0.959, "step": 1709 }, { "epoch": 2.1455457967377667, "grad_norm": 0.265625, "learning_rate": 2.8481806775407782e-06, "loss": 0.9086, "step": 1710 }, { "epoch": 2.1468005018820575, "grad_norm": 0.271484375, "learning_rate": 2.8439983270598075e-06, "loss": 0.9155, "step": 1711 }, { "epoch": 2.148055207026349, "grad_norm": 0.365234375, "learning_rate": 2.8398159765788376e-06, "loss": 0.9632, "step": 1712 }, { "epoch": 2.14930991217064, "grad_norm": 0.30078125, "learning_rate": 2.8356336260978673e-06, "loss": 0.9826, "step": 1713 }, { "epoch": 2.150564617314931, "grad_norm": 0.2890625, "learning_rate": 2.831451275616897e-06, "loss": 0.9573, "step": 1714 }, { "epoch": 2.151819322459222, "grad_norm": 0.26953125, "learning_rate": 2.8272689251359266e-06, "loss": 0.9681, "step": 1715 }, { "epoch": 2.1530740276035134, "grad_norm": 0.275390625, "learning_rate": 2.8230865746549563e-06, "loss": 0.9355, "step": 1716 }, { "epoch": 2.1543287327478042, "grad_norm": 0.271484375, "learning_rate": 2.818904224173986e-06, "loss": 0.9079, "step": 1717 }, { "epoch": 2.1555834378920955, "grad_norm": 0.27734375, "learning_rate": 2.814721873693016e-06, "loss": 0.9586, "step": 1718 }, { "epoch": 2.1568381430363863, "grad_norm": 0.373046875, "learning_rate": 2.8105395232120453e-06, "loss": 0.9779, "step": 1719 }, { "epoch": 2.1580928481806776, "grad_norm": 0.275390625, "learning_rate": 2.806357172731075e-06, "loss": 0.9956, "step": 1720 }, { "epoch": 2.1593475533249684, "grad_norm": 0.3203125, "learning_rate": 2.802174822250105e-06, "loss": 0.8954, "step": 1721 }, { "epoch": 2.1606022584692597, "grad_norm": 0.310546875, "learning_rate": 2.7979924717691343e-06, "loss": 0.9473, "step": 1722 }, { "epoch": 2.161856963613551, "grad_norm": 0.2890625, "learning_rate": 2.793810121288164e-06, "loss": 0.8861, "step": 1723 }, { "epoch": 2.1631116687578418, "grad_norm": 0.2890625, "learning_rate": 2.789627770807194e-06, "loss": 0.9597, "step": 1724 }, { "epoch": 2.164366373902133, "grad_norm": 0.283203125, "learning_rate": 2.7854454203262238e-06, "loss": 0.9409, "step": 1725 }, { "epoch": 2.1656210790464243, "grad_norm": 0.53125, "learning_rate": 2.781263069845253e-06, "loss": 0.9116, "step": 1726 }, { "epoch": 2.166875784190715, "grad_norm": 0.28125, "learning_rate": 2.7770807193642827e-06, "loss": 0.9064, "step": 1727 }, { "epoch": 2.1681304893350064, "grad_norm": 0.267578125, "learning_rate": 2.772898368883313e-06, "loss": 0.9434, "step": 1728 }, { "epoch": 2.1693851944792972, "grad_norm": 0.275390625, "learning_rate": 2.7687160184023425e-06, "loss": 0.9333, "step": 1729 }, { "epoch": 2.1706398996235885, "grad_norm": 0.267578125, "learning_rate": 2.7645336679213717e-06, "loss": 0.9013, "step": 1730 }, { "epoch": 2.1718946047678793, "grad_norm": 0.29296875, "learning_rate": 2.760351317440402e-06, "loss": 0.953, "step": 1731 }, { "epoch": 2.1731493099121706, "grad_norm": 0.2734375, "learning_rate": 2.7561689669594315e-06, "loss": 0.8803, "step": 1732 }, { "epoch": 2.174404015056462, "grad_norm": 0.29296875, "learning_rate": 2.7519866164784608e-06, "loss": 0.8976, "step": 1733 }, { "epoch": 2.1756587202007527, "grad_norm": 0.296875, "learning_rate": 2.747804265997491e-06, "loss": 0.915, "step": 1734 }, { "epoch": 2.176913425345044, "grad_norm": 0.2734375, "learning_rate": 2.7436219155165205e-06, "loss": 0.9622, "step": 1735 }, { "epoch": 2.178168130489335, "grad_norm": 0.275390625, "learning_rate": 2.7394395650355502e-06, "loss": 0.9238, "step": 1736 }, { "epoch": 2.179422835633626, "grad_norm": 0.287109375, "learning_rate": 2.73525721455458e-06, "loss": 0.9622, "step": 1737 }, { "epoch": 2.1806775407779173, "grad_norm": 0.296875, "learning_rate": 2.7310748640736096e-06, "loss": 0.9085, "step": 1738 }, { "epoch": 2.181932245922208, "grad_norm": 0.279296875, "learning_rate": 2.7268925135926392e-06, "loss": 0.9357, "step": 1739 }, { "epoch": 2.1831869510664994, "grad_norm": 0.2734375, "learning_rate": 2.7227101631116693e-06, "loss": 0.9109, "step": 1740 }, { "epoch": 2.1844416562107907, "grad_norm": 0.271484375, "learning_rate": 2.7185278126306986e-06, "loss": 0.9066, "step": 1741 }, { "epoch": 2.1856963613550815, "grad_norm": 0.265625, "learning_rate": 2.7143454621497283e-06, "loss": 0.9038, "step": 1742 }, { "epoch": 2.1869510664993728, "grad_norm": 0.275390625, "learning_rate": 2.7101631116687584e-06, "loss": 0.9488, "step": 1743 }, { "epoch": 2.1882057716436636, "grad_norm": 0.28515625, "learning_rate": 2.7059807611877876e-06, "loss": 0.9644, "step": 1744 }, { "epoch": 2.189460476787955, "grad_norm": 0.291015625, "learning_rate": 2.7017984107068173e-06, "loss": 0.9226, "step": 1745 }, { "epoch": 2.190715181932246, "grad_norm": 0.265625, "learning_rate": 2.6976160602258474e-06, "loss": 0.9249, "step": 1746 }, { "epoch": 2.191969887076537, "grad_norm": 0.279296875, "learning_rate": 2.693433709744877e-06, "loss": 0.8941, "step": 1747 }, { "epoch": 2.193224592220828, "grad_norm": 0.26953125, "learning_rate": 2.6892513592639063e-06, "loss": 0.9283, "step": 1748 }, { "epoch": 2.194479297365119, "grad_norm": 0.3046875, "learning_rate": 2.685069008782936e-06, "loss": 0.987, "step": 1749 }, { "epoch": 2.1957340025094103, "grad_norm": 0.267578125, "learning_rate": 2.680886658301966e-06, "loss": 0.9127, "step": 1750 }, { "epoch": 2.1969887076537016, "grad_norm": 0.263671875, "learning_rate": 2.6767043078209958e-06, "loss": 0.8816, "step": 1751 }, { "epoch": 2.1982434127979924, "grad_norm": 0.27734375, "learning_rate": 2.672521957340025e-06, "loss": 0.9872, "step": 1752 }, { "epoch": 2.1994981179422837, "grad_norm": 0.271484375, "learning_rate": 2.668339606859055e-06, "loss": 0.9045, "step": 1753 }, { "epoch": 2.2007528230865745, "grad_norm": 0.267578125, "learning_rate": 2.664157256378085e-06, "loss": 0.9379, "step": 1754 }, { "epoch": 2.2020075282308658, "grad_norm": 0.28125, "learning_rate": 2.659974905897114e-06, "loss": 0.9453, "step": 1755 }, { "epoch": 2.203262233375157, "grad_norm": 0.294921875, "learning_rate": 2.655792555416144e-06, "loss": 0.9315, "step": 1756 }, { "epoch": 2.204516938519448, "grad_norm": 0.33984375, "learning_rate": 2.651610204935174e-06, "loss": 0.8946, "step": 1757 }, { "epoch": 2.205771643663739, "grad_norm": 0.279296875, "learning_rate": 2.6474278544542035e-06, "loss": 0.9246, "step": 1758 }, { "epoch": 2.20702634880803, "grad_norm": 0.28125, "learning_rate": 2.643245503973233e-06, "loss": 0.9434, "step": 1759 }, { "epoch": 2.208281053952321, "grad_norm": 0.341796875, "learning_rate": 2.639063153492263e-06, "loss": 0.9049, "step": 1760 }, { "epoch": 2.2095357590966125, "grad_norm": 0.279296875, "learning_rate": 2.6348808030112925e-06, "loss": 0.9273, "step": 1761 }, { "epoch": 2.2107904642409033, "grad_norm": 0.341796875, "learning_rate": 2.6306984525303226e-06, "loss": 0.9302, "step": 1762 }, { "epoch": 2.2120451693851946, "grad_norm": 0.287109375, "learning_rate": 2.626516102049352e-06, "loss": 0.9285, "step": 1763 }, { "epoch": 2.2132998745294854, "grad_norm": 0.291015625, "learning_rate": 2.6223337515683816e-06, "loss": 0.9215, "step": 1764 }, { "epoch": 2.2145545796737767, "grad_norm": 0.2890625, "learning_rate": 2.6181514010874117e-06, "loss": 0.9251, "step": 1765 }, { "epoch": 2.215809284818068, "grad_norm": 0.2890625, "learning_rate": 2.613969050606441e-06, "loss": 0.9497, "step": 1766 }, { "epoch": 2.2170639899623588, "grad_norm": 0.291015625, "learning_rate": 2.6097867001254706e-06, "loss": 0.9756, "step": 1767 }, { "epoch": 2.21831869510665, "grad_norm": 0.27734375, "learning_rate": 2.6056043496445007e-06, "loss": 0.9307, "step": 1768 }, { "epoch": 2.219573400250941, "grad_norm": 0.294921875, "learning_rate": 2.6014219991635304e-06, "loss": 0.9246, "step": 1769 }, { "epoch": 2.220828105395232, "grad_norm": 0.271484375, "learning_rate": 2.5972396486825596e-06, "loss": 0.9226, "step": 1770 }, { "epoch": 2.2220828105395234, "grad_norm": 0.27734375, "learning_rate": 2.5930572982015893e-06, "loss": 0.8948, "step": 1771 }, { "epoch": 2.223337515683814, "grad_norm": 0.287109375, "learning_rate": 2.5888749477206194e-06, "loss": 0.9677, "step": 1772 }, { "epoch": 2.2245922208281055, "grad_norm": 0.283203125, "learning_rate": 2.584692597239649e-06, "loss": 0.9573, "step": 1773 }, { "epoch": 2.2258469259723963, "grad_norm": 0.27734375, "learning_rate": 2.5805102467586783e-06, "loss": 0.9526, "step": 1774 }, { "epoch": 2.2271016311166876, "grad_norm": 0.283203125, "learning_rate": 2.5763278962777084e-06, "loss": 0.9517, "step": 1775 }, { "epoch": 2.228356336260979, "grad_norm": 0.28125, "learning_rate": 2.572145545796738e-06, "loss": 0.9239, "step": 1776 }, { "epoch": 2.2296110414052697, "grad_norm": 0.275390625, "learning_rate": 2.5679631953157673e-06, "loss": 0.9313, "step": 1777 }, { "epoch": 2.230865746549561, "grad_norm": 0.279296875, "learning_rate": 2.5637808448347974e-06, "loss": 0.9438, "step": 1778 }, { "epoch": 2.2321204516938518, "grad_norm": 0.28125, "learning_rate": 2.559598494353827e-06, "loss": 0.9051, "step": 1779 }, { "epoch": 2.233375156838143, "grad_norm": 0.298828125, "learning_rate": 2.555416143872857e-06, "loss": 0.918, "step": 1780 }, { "epoch": 2.2346298619824343, "grad_norm": 0.275390625, "learning_rate": 2.5512337933918865e-06, "loss": 0.8974, "step": 1781 }, { "epoch": 2.235884567126725, "grad_norm": 0.267578125, "learning_rate": 2.547051442910916e-06, "loss": 0.9255, "step": 1782 }, { "epoch": 2.2371392722710164, "grad_norm": 0.275390625, "learning_rate": 2.542869092429946e-06, "loss": 0.9672, "step": 1783 }, { "epoch": 2.238393977415307, "grad_norm": 0.3125, "learning_rate": 2.538686741948976e-06, "loss": 0.9708, "step": 1784 }, { "epoch": 2.2396486825595985, "grad_norm": 0.279296875, "learning_rate": 2.534504391468005e-06, "loss": 0.9588, "step": 1785 }, { "epoch": 2.2409033877038897, "grad_norm": 0.365234375, "learning_rate": 2.530322040987035e-06, "loss": 0.9044, "step": 1786 }, { "epoch": 2.2421580928481806, "grad_norm": 0.2890625, "learning_rate": 2.526139690506065e-06, "loss": 0.9614, "step": 1787 }, { "epoch": 2.243412797992472, "grad_norm": 0.2890625, "learning_rate": 2.521957340025094e-06, "loss": 0.8977, "step": 1788 }, { "epoch": 2.2446675031367627, "grad_norm": 0.267578125, "learning_rate": 2.517774989544124e-06, "loss": 0.9389, "step": 1789 }, { "epoch": 2.245922208281054, "grad_norm": 0.28515625, "learning_rate": 2.513592639063154e-06, "loss": 0.9916, "step": 1790 }, { "epoch": 2.247176913425345, "grad_norm": 0.28125, "learning_rate": 2.5094102885821836e-06, "loss": 0.9302, "step": 1791 }, { "epoch": 2.248431618569636, "grad_norm": 0.380859375, "learning_rate": 2.505227938101213e-06, "loss": 0.9193, "step": 1792 }, { "epoch": 2.2496863237139273, "grad_norm": 0.28125, "learning_rate": 2.5010455876202426e-06, "loss": 0.9268, "step": 1793 }, { "epoch": 2.250941028858218, "grad_norm": 0.28125, "learning_rate": 2.4968632371392727e-06, "loss": 0.916, "step": 1794 }, { "epoch": 2.2521957340025094, "grad_norm": 0.302734375, "learning_rate": 2.4926808866583023e-06, "loss": 0.938, "step": 1795 }, { "epoch": 2.2534504391468007, "grad_norm": 0.43359375, "learning_rate": 2.488498536177332e-06, "loss": 0.9763, "step": 1796 }, { "epoch": 2.2547051442910915, "grad_norm": 0.287109375, "learning_rate": 2.4843161856963617e-06, "loss": 0.9685, "step": 1797 }, { "epoch": 2.2559598494353827, "grad_norm": 0.314453125, "learning_rate": 2.480133835215391e-06, "loss": 0.9228, "step": 1798 }, { "epoch": 2.257214554579674, "grad_norm": 0.294921875, "learning_rate": 2.475951484734421e-06, "loss": 0.9871, "step": 1799 }, { "epoch": 2.258469259723965, "grad_norm": 0.271484375, "learning_rate": 2.4717691342534507e-06, "loss": 0.9164, "step": 1800 }, { "epoch": 2.259723964868256, "grad_norm": 0.271484375, "learning_rate": 2.4675867837724804e-06, "loss": 0.9992, "step": 1801 }, { "epoch": 2.260978670012547, "grad_norm": 0.27734375, "learning_rate": 2.46340443329151e-06, "loss": 0.935, "step": 1802 }, { "epoch": 2.262233375156838, "grad_norm": 0.283203125, "learning_rate": 2.4592220828105398e-06, "loss": 0.936, "step": 1803 }, { "epoch": 2.263488080301129, "grad_norm": 0.27734375, "learning_rate": 2.4550397323295694e-06, "loss": 0.9259, "step": 1804 }, { "epoch": 2.2647427854454203, "grad_norm": 0.279296875, "learning_rate": 2.450857381848599e-06, "loss": 0.9479, "step": 1805 }, { "epoch": 2.2659974905897116, "grad_norm": 0.298828125, "learning_rate": 2.4466750313676288e-06, "loss": 0.9463, "step": 1806 }, { "epoch": 2.2672521957340024, "grad_norm": 0.28125, "learning_rate": 2.4424926808866585e-06, "loss": 0.9386, "step": 1807 }, { "epoch": 2.2685069008782937, "grad_norm": 0.88671875, "learning_rate": 2.438310330405688e-06, "loss": 0.9524, "step": 1808 }, { "epoch": 2.269761606022585, "grad_norm": 0.287109375, "learning_rate": 2.434127979924718e-06, "loss": 0.9271, "step": 1809 }, { "epoch": 2.2710163111668757, "grad_norm": 0.333984375, "learning_rate": 2.4299456294437475e-06, "loss": 0.9278, "step": 1810 }, { "epoch": 2.272271016311167, "grad_norm": 0.306640625, "learning_rate": 2.425763278962777e-06, "loss": 0.9081, "step": 1811 }, { "epoch": 2.273525721455458, "grad_norm": 0.267578125, "learning_rate": 2.421580928481807e-06, "loss": 0.8893, "step": 1812 }, { "epoch": 2.274780426599749, "grad_norm": 0.306640625, "learning_rate": 2.4173985780008365e-06, "loss": 0.963, "step": 1813 }, { "epoch": 2.27603513174404, "grad_norm": 0.2890625, "learning_rate": 2.4132162275198666e-06, "loss": 0.893, "step": 1814 }, { "epoch": 2.277289836888331, "grad_norm": 0.35546875, "learning_rate": 2.409033877038896e-06, "loss": 0.9495, "step": 1815 }, { "epoch": 2.2785445420326225, "grad_norm": 0.3203125, "learning_rate": 2.404851526557926e-06, "loss": 0.9419, "step": 1816 }, { "epoch": 2.2797992471769133, "grad_norm": 0.275390625, "learning_rate": 2.4006691760769556e-06, "loss": 0.9343, "step": 1817 }, { "epoch": 2.2810539523212046, "grad_norm": 0.28515625, "learning_rate": 2.3964868255959853e-06, "loss": 0.9064, "step": 1818 }, { "epoch": 2.282308657465496, "grad_norm": 0.271484375, "learning_rate": 2.392304475115015e-06, "loss": 0.9346, "step": 1819 }, { "epoch": 2.2835633626097867, "grad_norm": 0.271484375, "learning_rate": 2.3881221246340442e-06, "loss": 0.9277, "step": 1820 }, { "epoch": 2.284818067754078, "grad_norm": 0.267578125, "learning_rate": 2.3839397741530743e-06, "loss": 0.9559, "step": 1821 }, { "epoch": 2.2860727728983687, "grad_norm": 0.2734375, "learning_rate": 2.379757423672104e-06, "loss": 0.945, "step": 1822 }, { "epoch": 2.28732747804266, "grad_norm": 0.287109375, "learning_rate": 2.3755750731911337e-06, "loss": 0.9913, "step": 1823 }, { "epoch": 2.288582183186951, "grad_norm": 0.291015625, "learning_rate": 2.3713927227101634e-06, "loss": 0.9205, "step": 1824 }, { "epoch": 2.289836888331242, "grad_norm": 0.26953125, "learning_rate": 2.367210372229193e-06, "loss": 0.9279, "step": 1825 }, { "epoch": 2.2910915934755334, "grad_norm": 0.279296875, "learning_rate": 2.3630280217482227e-06, "loss": 0.9656, "step": 1826 }, { "epoch": 2.292346298619824, "grad_norm": 0.2890625, "learning_rate": 2.3588456712672524e-06, "loss": 0.9482, "step": 1827 }, { "epoch": 2.2936010037641155, "grad_norm": 0.474609375, "learning_rate": 2.354663320786282e-06, "loss": 0.9065, "step": 1828 }, { "epoch": 2.2948557089084067, "grad_norm": 0.298828125, "learning_rate": 2.3504809703053117e-06, "loss": 0.9418, "step": 1829 }, { "epoch": 2.2961104140526976, "grad_norm": 0.29296875, "learning_rate": 2.3462986198243414e-06, "loss": 0.914, "step": 1830 }, { "epoch": 2.297365119196989, "grad_norm": 0.298828125, "learning_rate": 2.342116269343371e-06, "loss": 0.9714, "step": 1831 }, { "epoch": 2.2986198243412796, "grad_norm": 0.302734375, "learning_rate": 2.3379339188624008e-06, "loss": 0.9899, "step": 1832 }, { "epoch": 2.299874529485571, "grad_norm": 0.275390625, "learning_rate": 2.3337515683814304e-06, "loss": 0.9197, "step": 1833 }, { "epoch": 2.3011292346298617, "grad_norm": 0.28515625, "learning_rate": 2.32956921790046e-06, "loss": 0.8987, "step": 1834 }, { "epoch": 2.302383939774153, "grad_norm": 0.275390625, "learning_rate": 2.32538686741949e-06, "loss": 0.9258, "step": 1835 }, { "epoch": 2.3036386449184443, "grad_norm": 0.283203125, "learning_rate": 2.32120451693852e-06, "loss": 0.9757, "step": 1836 }, { "epoch": 2.304893350062735, "grad_norm": 0.275390625, "learning_rate": 2.317022166457549e-06, "loss": 0.9738, "step": 1837 }, { "epoch": 2.3061480552070264, "grad_norm": 0.287109375, "learning_rate": 2.3128398159765792e-06, "loss": 0.9575, "step": 1838 }, { "epoch": 2.3074027603513176, "grad_norm": 0.287109375, "learning_rate": 2.3086574654956085e-06, "loss": 0.8912, "step": 1839 }, { "epoch": 2.3086574654956085, "grad_norm": 0.2734375, "learning_rate": 2.3044751150146386e-06, "loss": 0.9522, "step": 1840 }, { "epoch": 2.3099121706398997, "grad_norm": 0.28125, "learning_rate": 2.3002927645336683e-06, "loss": 0.8911, "step": 1841 }, { "epoch": 2.3111668757841906, "grad_norm": 0.2734375, "learning_rate": 2.2961104140526975e-06, "loss": 0.9284, "step": 1842 }, { "epoch": 2.312421580928482, "grad_norm": 0.26953125, "learning_rate": 2.2919280635717276e-06, "loss": 0.9353, "step": 1843 }, { "epoch": 2.313676286072773, "grad_norm": 0.27734375, "learning_rate": 2.2877457130907573e-06, "loss": 0.9653, "step": 1844 }, { "epoch": 2.314930991217064, "grad_norm": 0.265625, "learning_rate": 2.283563362609787e-06, "loss": 0.9335, "step": 1845 }, { "epoch": 2.316185696361355, "grad_norm": 0.2578125, "learning_rate": 2.2793810121288166e-06, "loss": 0.9482, "step": 1846 }, { "epoch": 2.317440401505646, "grad_norm": 0.287109375, "learning_rate": 2.2751986616478463e-06, "loss": 0.9681, "step": 1847 }, { "epoch": 2.3186951066499373, "grad_norm": 0.27734375, "learning_rate": 2.271016311166876e-06, "loss": 0.9445, "step": 1848 }, { "epoch": 2.3199498117942285, "grad_norm": 0.267578125, "learning_rate": 2.2668339606859057e-06, "loss": 0.913, "step": 1849 }, { "epoch": 2.3212045169385194, "grad_norm": 0.298828125, "learning_rate": 2.2626516102049354e-06, "loss": 0.9831, "step": 1850 }, { "epoch": 2.3224592220828106, "grad_norm": 0.29296875, "learning_rate": 2.258469259723965e-06, "loss": 0.8941, "step": 1851 }, { "epoch": 2.3237139272271015, "grad_norm": 0.396484375, "learning_rate": 2.2542869092429947e-06, "loss": 0.9203, "step": 1852 }, { "epoch": 2.3249686323713927, "grad_norm": 0.2734375, "learning_rate": 2.2501045587620244e-06, "loss": 0.9328, "step": 1853 }, { "epoch": 2.326223337515684, "grad_norm": 0.28515625, "learning_rate": 2.245922208281054e-06, "loss": 0.9496, "step": 1854 }, { "epoch": 2.327478042659975, "grad_norm": 0.26953125, "learning_rate": 2.2417398578000837e-06, "loss": 0.9147, "step": 1855 }, { "epoch": 2.328732747804266, "grad_norm": 0.279296875, "learning_rate": 2.2375575073191134e-06, "loss": 0.9321, "step": 1856 }, { "epoch": 2.329987452948557, "grad_norm": 0.26953125, "learning_rate": 2.233375156838143e-06, "loss": 0.9222, "step": 1857 }, { "epoch": 2.331242158092848, "grad_norm": 0.404296875, "learning_rate": 2.229192806357173e-06, "loss": 0.9707, "step": 1858 }, { "epoch": 2.3324968632371395, "grad_norm": 0.2734375, "learning_rate": 2.2250104558762024e-06, "loss": 0.9289, "step": 1859 }, { "epoch": 2.3337515683814303, "grad_norm": 0.28125, "learning_rate": 2.2208281053952325e-06, "loss": 0.9249, "step": 1860 }, { "epoch": 2.3350062735257215, "grad_norm": 0.287109375, "learning_rate": 2.2166457549142618e-06, "loss": 0.9125, "step": 1861 }, { "epoch": 2.3362609786700124, "grad_norm": 0.265625, "learning_rate": 2.212463404433292e-06, "loss": 0.9463, "step": 1862 }, { "epoch": 2.3375156838143036, "grad_norm": 0.267578125, "learning_rate": 2.2082810539523216e-06, "loss": 0.8802, "step": 1863 }, { "epoch": 2.338770388958595, "grad_norm": 0.28125, "learning_rate": 2.204098703471351e-06, "loss": 0.9851, "step": 1864 }, { "epoch": 2.3400250941028857, "grad_norm": 0.2734375, "learning_rate": 2.199916352990381e-06, "loss": 0.8841, "step": 1865 }, { "epoch": 2.341279799247177, "grad_norm": 0.2734375, "learning_rate": 2.1957340025094106e-06, "loss": 0.9573, "step": 1866 }, { "epoch": 2.342534504391468, "grad_norm": 0.263671875, "learning_rate": 2.1915516520284403e-06, "loss": 0.9058, "step": 1867 }, { "epoch": 2.343789209535759, "grad_norm": 0.306640625, "learning_rate": 2.18736930154747e-06, "loss": 0.928, "step": 1868 }, { "epoch": 2.3450439146800504, "grad_norm": 0.271484375, "learning_rate": 2.1831869510664996e-06, "loss": 0.9325, "step": 1869 }, { "epoch": 2.346298619824341, "grad_norm": 0.28125, "learning_rate": 2.1790046005855293e-06, "loss": 0.9629, "step": 1870 }, { "epoch": 2.3475533249686324, "grad_norm": 0.306640625, "learning_rate": 2.174822250104559e-06, "loss": 0.9305, "step": 1871 }, { "epoch": 2.3488080301129233, "grad_norm": 0.275390625, "learning_rate": 2.1706398996235886e-06, "loss": 0.9478, "step": 1872 }, { "epoch": 2.3500627352572145, "grad_norm": 0.287109375, "learning_rate": 2.1664575491426183e-06, "loss": 0.962, "step": 1873 }, { "epoch": 2.351317440401506, "grad_norm": 0.291015625, "learning_rate": 2.162275198661648e-06, "loss": 0.9477, "step": 1874 }, { "epoch": 2.3525721455457966, "grad_norm": 0.28125, "learning_rate": 2.1580928481806777e-06, "loss": 0.9758, "step": 1875 }, { "epoch": 2.353826850690088, "grad_norm": 0.27734375, "learning_rate": 2.1539104976997073e-06, "loss": 0.8833, "step": 1876 }, { "epoch": 2.3550815558343787, "grad_norm": 0.3359375, "learning_rate": 2.149728147218737e-06, "loss": 0.9449, "step": 1877 }, { "epoch": 2.35633626097867, "grad_norm": 0.314453125, "learning_rate": 2.1455457967377667e-06, "loss": 0.9723, "step": 1878 }, { "epoch": 2.3575909661229613, "grad_norm": 0.2734375, "learning_rate": 2.1413634462567964e-06, "loss": 0.8931, "step": 1879 }, { "epoch": 2.358845671267252, "grad_norm": 0.271484375, "learning_rate": 2.1371810957758265e-06, "loss": 0.9044, "step": 1880 }, { "epoch": 2.3601003764115434, "grad_norm": 0.29296875, "learning_rate": 2.1329987452948557e-06, "loss": 0.9513, "step": 1881 }, { "epoch": 2.3613550815558346, "grad_norm": 0.30078125, "learning_rate": 2.128816394813886e-06, "loss": 0.9774, "step": 1882 }, { "epoch": 2.3626097867001254, "grad_norm": 0.279296875, "learning_rate": 2.124634044332915e-06, "loss": 0.9583, "step": 1883 }, { "epoch": 2.3638644918444167, "grad_norm": 0.294921875, "learning_rate": 2.120451693851945e-06, "loss": 0.9913, "step": 1884 }, { "epoch": 2.3651191969887075, "grad_norm": 0.279296875, "learning_rate": 2.116269343370975e-06, "loss": 0.8872, "step": 1885 }, { "epoch": 2.366373902132999, "grad_norm": 0.2734375, "learning_rate": 2.112086992890004e-06, "loss": 0.9689, "step": 1886 }, { "epoch": 2.3676286072772896, "grad_norm": 0.28125, "learning_rate": 2.107904642409034e-06, "loss": 0.9841, "step": 1887 }, { "epoch": 2.368883312421581, "grad_norm": 0.271484375, "learning_rate": 2.1037222919280634e-06, "loss": 0.9662, "step": 1888 }, { "epoch": 2.370138017565872, "grad_norm": 0.275390625, "learning_rate": 2.0995399414470935e-06, "loss": 0.9465, "step": 1889 }, { "epoch": 2.371392722710163, "grad_norm": 0.271484375, "learning_rate": 2.0953575909661232e-06, "loss": 0.9346, "step": 1890 }, { "epoch": 2.3726474278544543, "grad_norm": 0.283203125, "learning_rate": 2.091175240485153e-06, "loss": 0.9228, "step": 1891 }, { "epoch": 2.3739021329987455, "grad_norm": 0.287109375, "learning_rate": 2.0869928900041826e-06, "loss": 0.9587, "step": 1892 }, { "epoch": 2.3751568381430364, "grad_norm": 0.27734375, "learning_rate": 2.0828105395232122e-06, "loss": 0.8764, "step": 1893 }, { "epoch": 2.3764115432873276, "grad_norm": 0.27734375, "learning_rate": 2.078628189042242e-06, "loss": 0.9071, "step": 1894 }, { "epoch": 2.3776662484316184, "grad_norm": 0.326171875, "learning_rate": 2.0744458385612716e-06, "loss": 0.9156, "step": 1895 }, { "epoch": 2.3789209535759097, "grad_norm": 0.298828125, "learning_rate": 2.0702634880803013e-06, "loss": 0.9037, "step": 1896 }, { "epoch": 2.3801756587202005, "grad_norm": 0.59765625, "learning_rate": 2.066081137599331e-06, "loss": 0.9105, "step": 1897 }, { "epoch": 2.381430363864492, "grad_norm": 0.275390625, "learning_rate": 2.0618987871183606e-06, "loss": 0.9395, "step": 1898 }, { "epoch": 2.382685069008783, "grad_norm": 0.337890625, "learning_rate": 2.0577164366373903e-06, "loss": 0.9432, "step": 1899 }, { "epoch": 2.383939774153074, "grad_norm": 0.283203125, "learning_rate": 2.05353408615642e-06, "loss": 0.8961, "step": 1900 }, { "epoch": 2.385194479297365, "grad_norm": 0.29296875, "learning_rate": 2.0493517356754497e-06, "loss": 0.9667, "step": 1901 }, { "epoch": 2.3864491844416564, "grad_norm": 0.283203125, "learning_rate": 2.0451693851944798e-06, "loss": 0.8919, "step": 1902 }, { "epoch": 2.3877038895859473, "grad_norm": 0.275390625, "learning_rate": 2.040987034713509e-06, "loss": 0.9273, "step": 1903 }, { "epoch": 2.3889585947302385, "grad_norm": 0.26953125, "learning_rate": 2.036804684232539e-06, "loss": 0.9258, "step": 1904 }, { "epoch": 2.3902132998745294, "grad_norm": 0.2734375, "learning_rate": 2.0326223337515684e-06, "loss": 0.8952, "step": 1905 }, { "epoch": 2.3914680050188206, "grad_norm": 0.28515625, "learning_rate": 2.0284399832705985e-06, "loss": 0.9239, "step": 1906 }, { "epoch": 2.3927227101631114, "grad_norm": 0.302734375, "learning_rate": 2.024257632789628e-06, "loss": 0.9707, "step": 1907 }, { "epoch": 2.3939774153074027, "grad_norm": 0.287109375, "learning_rate": 2.0200752823086574e-06, "loss": 0.935, "step": 1908 }, { "epoch": 2.395232120451694, "grad_norm": 0.275390625, "learning_rate": 2.0158929318276875e-06, "loss": 0.9087, "step": 1909 }, { "epoch": 2.396486825595985, "grad_norm": 0.271484375, "learning_rate": 2.0117105813467167e-06, "loss": 0.9081, "step": 1910 }, { "epoch": 2.397741530740276, "grad_norm": 0.28515625, "learning_rate": 2.007528230865747e-06, "loss": 0.9593, "step": 1911 }, { "epoch": 2.3989962358845673, "grad_norm": 0.291015625, "learning_rate": 2.0033458803847765e-06, "loss": 0.9648, "step": 1912 }, { "epoch": 2.400250941028858, "grad_norm": 0.28515625, "learning_rate": 1.999163529903806e-06, "loss": 0.897, "step": 1913 }, { "epoch": 2.4015056461731494, "grad_norm": 0.283203125, "learning_rate": 1.994981179422836e-06, "loss": 0.9462, "step": 1914 }, { "epoch": 2.4027603513174403, "grad_norm": 0.283203125, "learning_rate": 1.9907988289418655e-06, "loss": 0.9639, "step": 1915 }, { "epoch": 2.4040150564617315, "grad_norm": 0.2734375, "learning_rate": 1.986616478460895e-06, "loss": 0.9112, "step": 1916 }, { "epoch": 2.4052697616060223, "grad_norm": 0.279296875, "learning_rate": 1.982434127979925e-06, "loss": 0.939, "step": 1917 }, { "epoch": 2.4065244667503136, "grad_norm": 0.271484375, "learning_rate": 1.9782517774989546e-06, "loss": 0.9382, "step": 1918 }, { "epoch": 2.407779171894605, "grad_norm": 0.279296875, "learning_rate": 1.9740694270179842e-06, "loss": 0.9645, "step": 1919 }, { "epoch": 2.4090338770388957, "grad_norm": 0.28125, "learning_rate": 1.969887076537014e-06, "loss": 0.945, "step": 1920 }, { "epoch": 2.410288582183187, "grad_norm": 0.283203125, "learning_rate": 1.9657047260560436e-06, "loss": 0.9621, "step": 1921 }, { "epoch": 2.4115432873274782, "grad_norm": 0.28125, "learning_rate": 1.9615223755750733e-06, "loss": 0.8866, "step": 1922 }, { "epoch": 2.412797992471769, "grad_norm": 0.27734375, "learning_rate": 1.957340025094103e-06, "loss": 0.922, "step": 1923 }, { "epoch": 2.4140526976160603, "grad_norm": 0.314453125, "learning_rate": 1.9531576746131326e-06, "loss": 0.908, "step": 1924 }, { "epoch": 2.415307402760351, "grad_norm": 0.275390625, "learning_rate": 1.9489753241321623e-06, "loss": 0.8975, "step": 1925 }, { "epoch": 2.4165621079046424, "grad_norm": 0.26953125, "learning_rate": 1.9447929736511924e-06, "loss": 0.9213, "step": 1926 }, { "epoch": 2.4178168130489333, "grad_norm": 0.27734375, "learning_rate": 1.9406106231702216e-06, "loss": 0.9423, "step": 1927 }, { "epoch": 2.4190715181932245, "grad_norm": 0.318359375, "learning_rate": 1.9364282726892517e-06, "loss": 0.8901, "step": 1928 }, { "epoch": 2.420326223337516, "grad_norm": 0.28515625, "learning_rate": 1.9322459222082814e-06, "loss": 0.9656, "step": 1929 }, { "epoch": 2.4215809284818066, "grad_norm": 0.271484375, "learning_rate": 1.9280635717273107e-06, "loss": 0.9449, "step": 1930 }, { "epoch": 2.422835633626098, "grad_norm": 0.275390625, "learning_rate": 1.9238812212463408e-06, "loss": 0.9196, "step": 1931 }, { "epoch": 2.424090338770389, "grad_norm": 0.26171875, "learning_rate": 1.91969887076537e-06, "loss": 0.9011, "step": 1932 }, { "epoch": 2.42534504391468, "grad_norm": 0.279296875, "learning_rate": 1.9155165202844e-06, "loss": 0.9441, "step": 1933 }, { "epoch": 2.4265997490589712, "grad_norm": 0.34375, "learning_rate": 1.91133416980343e-06, "loss": 0.8917, "step": 1934 }, { "epoch": 2.427854454203262, "grad_norm": 0.265625, "learning_rate": 1.9071518193224595e-06, "loss": 0.9168, "step": 1935 }, { "epoch": 2.4291091593475533, "grad_norm": 0.283203125, "learning_rate": 1.902969468841489e-06, "loss": 0.9435, "step": 1936 }, { "epoch": 2.4303638644918446, "grad_norm": 0.271484375, "learning_rate": 1.8987871183605188e-06, "loss": 0.9429, "step": 1937 }, { "epoch": 2.4316185696361354, "grad_norm": 0.283203125, "learning_rate": 1.8946047678795485e-06, "loss": 0.9094, "step": 1938 }, { "epoch": 2.4328732747804267, "grad_norm": 0.291015625, "learning_rate": 1.8904224173985784e-06, "loss": 0.9579, "step": 1939 }, { "epoch": 2.4341279799247175, "grad_norm": 0.29296875, "learning_rate": 1.8862400669176078e-06, "loss": 0.9522, "step": 1940 }, { "epoch": 2.435382685069009, "grad_norm": 0.28125, "learning_rate": 1.8820577164366375e-06, "loss": 0.9683, "step": 1941 }, { "epoch": 2.4366373902133, "grad_norm": 0.283203125, "learning_rate": 1.8778753659556672e-06, "loss": 0.9547, "step": 1942 }, { "epoch": 2.437892095357591, "grad_norm": 0.30078125, "learning_rate": 1.8736930154746969e-06, "loss": 0.9507, "step": 1943 }, { "epoch": 2.439146800501882, "grad_norm": 0.267578125, "learning_rate": 1.8695106649937268e-06, "loss": 0.937, "step": 1944 }, { "epoch": 2.440401505646173, "grad_norm": 0.2578125, "learning_rate": 1.8653283145127562e-06, "loss": 0.8865, "step": 1945 }, { "epoch": 2.4416562107904642, "grad_norm": 0.26953125, "learning_rate": 1.8611459640317861e-06, "loss": 0.9063, "step": 1946 }, { "epoch": 2.4429109159347555, "grad_norm": 0.263671875, "learning_rate": 1.8569636135508156e-06, "loss": 0.9276, "step": 1947 }, { "epoch": 2.4441656210790463, "grad_norm": 0.28125, "learning_rate": 1.8527812630698455e-06, "loss": 0.9367, "step": 1948 }, { "epoch": 2.4454203262233376, "grad_norm": 0.34375, "learning_rate": 1.8485989125888751e-06, "loss": 0.9107, "step": 1949 }, { "epoch": 2.4466750313676284, "grad_norm": 0.328125, "learning_rate": 1.844416562107905e-06, "loss": 0.9456, "step": 1950 }, { "epoch": 2.4479297365119197, "grad_norm": 0.275390625, "learning_rate": 1.8402342116269345e-06, "loss": 0.9224, "step": 1951 }, { "epoch": 2.449184441656211, "grad_norm": 0.283203125, "learning_rate": 1.8360518611459642e-06, "loss": 0.9204, "step": 1952 }, { "epoch": 2.450439146800502, "grad_norm": 0.279296875, "learning_rate": 1.8318695106649938e-06, "loss": 0.9525, "step": 1953 }, { "epoch": 2.451693851944793, "grad_norm": 0.265625, "learning_rate": 1.8276871601840235e-06, "loss": 0.8975, "step": 1954 }, { "epoch": 2.452948557089084, "grad_norm": 0.275390625, "learning_rate": 1.8235048097030534e-06, "loss": 0.9297, "step": 1955 }, { "epoch": 2.454203262233375, "grad_norm": 0.275390625, "learning_rate": 1.8193224592220829e-06, "loss": 0.9325, "step": 1956 }, { "epoch": 2.4554579673776664, "grad_norm": 0.2890625, "learning_rate": 1.8151401087411128e-06, "loss": 0.9283, "step": 1957 }, { "epoch": 2.4567126725219572, "grad_norm": 0.357421875, "learning_rate": 1.8109577582601422e-06, "loss": 0.9371, "step": 1958 }, { "epoch": 2.4579673776662485, "grad_norm": 0.3046875, "learning_rate": 1.8067754077791721e-06, "loss": 0.9549, "step": 1959 }, { "epoch": 2.4592220828105393, "grad_norm": 0.271484375, "learning_rate": 1.8025930572982018e-06, "loss": 0.9441, "step": 1960 }, { "epoch": 2.4604767879548306, "grad_norm": 0.27734375, "learning_rate": 1.7984107068172317e-06, "loss": 0.9675, "step": 1961 }, { "epoch": 2.461731493099122, "grad_norm": 0.28515625, "learning_rate": 1.7942283563362611e-06, "loss": 0.9618, "step": 1962 }, { "epoch": 2.4629861982434127, "grad_norm": 0.279296875, "learning_rate": 1.7900460058552906e-06, "loss": 0.9661, "step": 1963 }, { "epoch": 2.464240903387704, "grad_norm": 0.4609375, "learning_rate": 1.7858636553743205e-06, "loss": 0.9309, "step": 1964 }, { "epoch": 2.4654956085319952, "grad_norm": 0.271484375, "learning_rate": 1.7816813048933502e-06, "loss": 0.9166, "step": 1965 }, { "epoch": 2.466750313676286, "grad_norm": 0.33203125, "learning_rate": 1.77749895441238e-06, "loss": 0.9279, "step": 1966 }, { "epoch": 2.4680050188205773, "grad_norm": 0.271484375, "learning_rate": 1.7733166039314095e-06, "loss": 0.8878, "step": 1967 }, { "epoch": 2.469259723964868, "grad_norm": 0.3125, "learning_rate": 1.7691342534504394e-06, "loss": 0.923, "step": 1968 }, { "epoch": 2.4705144291091594, "grad_norm": 0.359375, "learning_rate": 1.7649519029694689e-06, "loss": 0.9735, "step": 1969 }, { "epoch": 2.4717691342534502, "grad_norm": 0.283203125, "learning_rate": 1.7607695524884987e-06, "loss": 0.9488, "step": 1970 }, { "epoch": 2.4730238393977415, "grad_norm": 0.271484375, "learning_rate": 1.7565872020075284e-06, "loss": 0.9304, "step": 1971 }, { "epoch": 2.4742785445420328, "grad_norm": 0.287109375, "learning_rate": 1.752404851526558e-06, "loss": 0.9526, "step": 1972 }, { "epoch": 2.4755332496863236, "grad_norm": 0.294921875, "learning_rate": 1.7482225010455878e-06, "loss": 0.9844, "step": 1973 }, { "epoch": 2.476787954830615, "grad_norm": 0.27734375, "learning_rate": 1.7440401505646172e-06, "loss": 0.9407, "step": 1974 }, { "epoch": 2.478042659974906, "grad_norm": 0.279296875, "learning_rate": 1.7398578000836471e-06, "loss": 0.9473, "step": 1975 }, { "epoch": 2.479297365119197, "grad_norm": 0.265625, "learning_rate": 1.7356754496026768e-06, "loss": 0.9189, "step": 1976 }, { "epoch": 2.4805520702634882, "grad_norm": 0.275390625, "learning_rate": 1.7314930991217067e-06, "loss": 0.9433, "step": 1977 }, { "epoch": 2.481806775407779, "grad_norm": 0.314453125, "learning_rate": 1.7273107486407362e-06, "loss": 0.9218, "step": 1978 }, { "epoch": 2.4830614805520703, "grad_norm": 0.28515625, "learning_rate": 1.723128398159766e-06, "loss": 0.9463, "step": 1979 }, { "epoch": 2.484316185696361, "grad_norm": 0.267578125, "learning_rate": 1.7189460476787955e-06, "loss": 0.9367, "step": 1980 }, { "epoch": 2.4855708908406524, "grad_norm": 0.27734375, "learning_rate": 1.7147636971978254e-06, "loss": 0.9424, "step": 1981 }, { "epoch": 2.4868255959849437, "grad_norm": 0.302734375, "learning_rate": 1.710581346716855e-06, "loss": 0.9589, "step": 1982 }, { "epoch": 2.4880803011292345, "grad_norm": 0.3984375, "learning_rate": 1.7063989962358847e-06, "loss": 0.9237, "step": 1983 }, { "epoch": 2.4893350062735258, "grad_norm": 0.2734375, "learning_rate": 1.7022166457549144e-06, "loss": 0.9092, "step": 1984 }, { "epoch": 2.490589711417817, "grad_norm": 0.36328125, "learning_rate": 1.6980342952739439e-06, "loss": 0.9224, "step": 1985 }, { "epoch": 2.491844416562108, "grad_norm": 0.271484375, "learning_rate": 1.6938519447929738e-06, "loss": 0.923, "step": 1986 }, { "epoch": 2.493099121706399, "grad_norm": 0.408203125, "learning_rate": 1.6896695943120034e-06, "loss": 0.959, "step": 1987 }, { "epoch": 2.49435382685069, "grad_norm": 0.4453125, "learning_rate": 1.6854872438310333e-06, "loss": 0.9727, "step": 1988 }, { "epoch": 2.4956085319949812, "grad_norm": 0.2890625, "learning_rate": 1.6813048933500628e-06, "loss": 0.8949, "step": 1989 }, { "epoch": 2.496863237139272, "grad_norm": 0.275390625, "learning_rate": 1.6771225428690927e-06, "loss": 0.9687, "step": 1990 }, { "epoch": 2.4981179422835633, "grad_norm": 0.2734375, "learning_rate": 1.6729401923881221e-06, "loss": 0.9723, "step": 1991 }, { "epoch": 2.4993726474278546, "grad_norm": 0.271484375, "learning_rate": 1.668757841907152e-06, "loss": 0.9369, "step": 1992 }, { "epoch": 2.5006273525721454, "grad_norm": 0.314453125, "learning_rate": 1.6645754914261817e-06, "loss": 0.9067, "step": 1993 }, { "epoch": 2.5018820577164367, "grad_norm": 0.271484375, "learning_rate": 1.6603931409452114e-06, "loss": 0.9531, "step": 1994 }, { "epoch": 2.503136762860728, "grad_norm": 0.30078125, "learning_rate": 1.656210790464241e-06, "loss": 0.9512, "step": 1995 }, { "epoch": 2.5043914680050188, "grad_norm": 0.28125, "learning_rate": 1.6520284399832705e-06, "loss": 0.9771, "step": 1996 }, { "epoch": 2.50564617314931, "grad_norm": 0.2734375, "learning_rate": 1.6478460895023004e-06, "loss": 0.9149, "step": 1997 }, { "epoch": 2.506900878293601, "grad_norm": 0.275390625, "learning_rate": 1.64366373902133e-06, "loss": 0.89, "step": 1998 }, { "epoch": 2.508155583437892, "grad_norm": 0.2890625, "learning_rate": 1.63948138854036e-06, "loss": 0.9522, "step": 1999 }, { "epoch": 2.509410288582183, "grad_norm": 0.287109375, "learning_rate": 1.6352990380593894e-06, "loss": 0.9459, "step": 2000 }, { "epoch": 2.5106649937264742, "grad_norm": 0.29296875, "learning_rate": 1.6311166875784193e-06, "loss": 0.985, "step": 2001 }, { "epoch": 2.5119196988707655, "grad_norm": 0.275390625, "learning_rate": 1.6269343370974488e-06, "loss": 0.893, "step": 2002 }, { "epoch": 2.5131744040150563, "grad_norm": 0.27734375, "learning_rate": 1.6227519866164787e-06, "loss": 0.948, "step": 2003 }, { "epoch": 2.5144291091593476, "grad_norm": 0.326171875, "learning_rate": 1.6185696361355084e-06, "loss": 0.8847, "step": 2004 }, { "epoch": 2.515683814303639, "grad_norm": 0.279296875, "learning_rate": 1.614387285654538e-06, "loss": 0.9261, "step": 2005 }, { "epoch": 2.5169385194479297, "grad_norm": 0.28125, "learning_rate": 1.6102049351735677e-06, "loss": 0.8711, "step": 2006 }, { "epoch": 2.518193224592221, "grad_norm": 0.259765625, "learning_rate": 1.6060225846925972e-06, "loss": 0.9046, "step": 2007 }, { "epoch": 2.5194479297365118, "grad_norm": 0.287109375, "learning_rate": 1.601840234211627e-06, "loss": 0.9758, "step": 2008 }, { "epoch": 2.520702634880803, "grad_norm": 0.2890625, "learning_rate": 1.5976578837306567e-06, "loss": 0.9414, "step": 2009 }, { "epoch": 2.521957340025094, "grad_norm": 0.2734375, "learning_rate": 1.5934755332496864e-06, "loss": 0.924, "step": 2010 }, { "epoch": 2.523212045169385, "grad_norm": 0.357421875, "learning_rate": 1.589293182768716e-06, "loss": 0.9119, "step": 2011 }, { "epoch": 2.5244667503136764, "grad_norm": 0.291015625, "learning_rate": 1.585110832287746e-06, "loss": 0.9934, "step": 2012 }, { "epoch": 2.5257214554579672, "grad_norm": 0.26953125, "learning_rate": 1.5809284818067754e-06, "loss": 0.9157, "step": 2013 }, { "epoch": 2.5269761606022585, "grad_norm": 0.26953125, "learning_rate": 1.5767461313258053e-06, "loss": 0.9031, "step": 2014 }, { "epoch": 2.5282308657465498, "grad_norm": 0.287109375, "learning_rate": 1.572563780844835e-06, "loss": 0.9552, "step": 2015 }, { "epoch": 2.5294855708908406, "grad_norm": 0.26953125, "learning_rate": 1.5683814303638647e-06, "loss": 0.9303, "step": 2016 }, { "epoch": 2.530740276035132, "grad_norm": 0.279296875, "learning_rate": 1.5641990798828943e-06, "loss": 0.9404, "step": 2017 }, { "epoch": 2.5319949811794227, "grad_norm": 0.26953125, "learning_rate": 1.5600167294019238e-06, "loss": 0.9084, "step": 2018 }, { "epoch": 2.533249686323714, "grad_norm": 0.298828125, "learning_rate": 1.5558343789209537e-06, "loss": 0.9038, "step": 2019 }, { "epoch": 2.5345043914680048, "grad_norm": 0.26953125, "learning_rate": 1.5516520284399834e-06, "loss": 0.9128, "step": 2020 }, { "epoch": 2.535759096612296, "grad_norm": 0.275390625, "learning_rate": 1.547469677959013e-06, "loss": 0.9446, "step": 2021 }, { "epoch": 2.5370138017565873, "grad_norm": 0.279296875, "learning_rate": 1.5432873274780427e-06, "loss": 0.9292, "step": 2022 }, { "epoch": 2.538268506900878, "grad_norm": 0.283203125, "learning_rate": 1.5391049769970726e-06, "loss": 0.9771, "step": 2023 }, { "epoch": 2.5395232120451694, "grad_norm": 0.265625, "learning_rate": 1.534922626516102e-06, "loss": 0.8933, "step": 2024 }, { "epoch": 2.5407779171894607, "grad_norm": 0.26953125, "learning_rate": 1.530740276035132e-06, "loss": 0.8693, "step": 2025 }, { "epoch": 2.5420326223337515, "grad_norm": 0.27734375, "learning_rate": 1.5265579255541616e-06, "loss": 0.9603, "step": 2026 }, { "epoch": 2.5432873274780428, "grad_norm": 0.28515625, "learning_rate": 1.5223755750731913e-06, "loss": 0.9172, "step": 2027 }, { "epoch": 2.544542032622334, "grad_norm": 0.28515625, "learning_rate": 1.518193224592221e-06, "loss": 0.9315, "step": 2028 }, { "epoch": 2.545796737766625, "grad_norm": 0.37890625, "learning_rate": 1.5140108741112505e-06, "loss": 0.8993, "step": 2029 }, { "epoch": 2.5470514429109157, "grad_norm": 0.26953125, "learning_rate": 1.5098285236302803e-06, "loss": 0.9352, "step": 2030 }, { "epoch": 2.548306148055207, "grad_norm": 0.28515625, "learning_rate": 1.50564617314931e-06, "loss": 0.9218, "step": 2031 }, { "epoch": 2.549560853199498, "grad_norm": 0.28125, "learning_rate": 1.5014638226683397e-06, "loss": 0.9594, "step": 2032 }, { "epoch": 2.550815558343789, "grad_norm": 0.2734375, "learning_rate": 1.4972814721873694e-06, "loss": 0.9627, "step": 2033 }, { "epoch": 2.5520702634880803, "grad_norm": 0.26953125, "learning_rate": 1.4930991217063993e-06, "loss": 0.9376, "step": 2034 }, { "epoch": 2.5533249686323716, "grad_norm": 0.287109375, "learning_rate": 1.4889167712254287e-06, "loss": 0.9657, "step": 2035 }, { "epoch": 2.5545796737766624, "grad_norm": 0.267578125, "learning_rate": 1.4847344207444586e-06, "loss": 0.9016, "step": 2036 }, { "epoch": 2.5558343789209537, "grad_norm": 0.2890625, "learning_rate": 1.4805520702634883e-06, "loss": 0.9215, "step": 2037 }, { "epoch": 2.557089084065245, "grad_norm": 0.296875, "learning_rate": 1.476369719782518e-06, "loss": 0.9031, "step": 2038 }, { "epoch": 2.5583437892095358, "grad_norm": 0.306640625, "learning_rate": 1.4721873693015476e-06, "loss": 0.8872, "step": 2039 }, { "epoch": 2.5595984943538266, "grad_norm": 0.275390625, "learning_rate": 1.468005018820577e-06, "loss": 0.9402, "step": 2040 }, { "epoch": 2.560853199498118, "grad_norm": 0.27734375, "learning_rate": 1.463822668339607e-06, "loss": 0.9282, "step": 2041 }, { "epoch": 2.562107904642409, "grad_norm": 0.28125, "learning_rate": 1.4596403178586367e-06, "loss": 0.9521, "step": 2042 }, { "epoch": 2.5633626097867, "grad_norm": 0.271484375, "learning_rate": 1.4554579673776663e-06, "loss": 0.9344, "step": 2043 }, { "epoch": 2.564617314930991, "grad_norm": 0.298828125, "learning_rate": 1.451275616896696e-06, "loss": 0.9392, "step": 2044 }, { "epoch": 2.5658720200752825, "grad_norm": 0.267578125, "learning_rate": 1.447093266415726e-06, "loss": 0.9375, "step": 2045 }, { "epoch": 2.5671267252195733, "grad_norm": 0.275390625, "learning_rate": 1.4429109159347554e-06, "loss": 0.9411, "step": 2046 }, { "epoch": 2.5683814303638646, "grad_norm": 0.28515625, "learning_rate": 1.4387285654537853e-06, "loss": 0.9463, "step": 2047 }, { "epoch": 2.569636135508156, "grad_norm": 0.294921875, "learning_rate": 1.434546214972815e-06, "loss": 0.9609, "step": 2048 }, { "epoch": 2.5708908406524467, "grad_norm": 0.279296875, "learning_rate": 1.4303638644918446e-06, "loss": 0.9124, "step": 2049 }, { "epoch": 2.572145545796738, "grad_norm": 0.27734375, "learning_rate": 1.4261815140108743e-06, "loss": 0.9846, "step": 2050 }, { "epoch": 2.5734002509410288, "grad_norm": 0.267578125, "learning_rate": 1.4219991635299037e-06, "loss": 0.9131, "step": 2051 }, { "epoch": 2.57465495608532, "grad_norm": 0.26953125, "learning_rate": 1.4178168130489336e-06, "loss": 0.9258, "step": 2052 }, { "epoch": 2.575909661229611, "grad_norm": 0.265625, "learning_rate": 1.4136344625679633e-06, "loss": 0.937, "step": 2053 }, { "epoch": 2.577164366373902, "grad_norm": 0.275390625, "learning_rate": 1.409452112086993e-06, "loss": 0.9261, "step": 2054 }, { "epoch": 2.5784190715181934, "grad_norm": 0.26953125, "learning_rate": 1.4052697616060227e-06, "loss": 0.8979, "step": 2055 }, { "epoch": 2.579673776662484, "grad_norm": 0.2734375, "learning_rate": 1.4010874111250525e-06, "loss": 0.9095, "step": 2056 }, { "epoch": 2.5809284818067755, "grad_norm": 0.314453125, "learning_rate": 1.396905060644082e-06, "loss": 0.9065, "step": 2057 }, { "epoch": 2.5821831869510667, "grad_norm": 0.275390625, "learning_rate": 1.3927227101631119e-06, "loss": 0.9094, "step": 2058 }, { "epoch": 2.5834378920953576, "grad_norm": 0.279296875, "learning_rate": 1.3885403596821414e-06, "loss": 0.94, "step": 2059 }, { "epoch": 2.584692597239649, "grad_norm": 0.279296875, "learning_rate": 1.3843580092011712e-06, "loss": 0.9809, "step": 2060 }, { "epoch": 2.5859473023839397, "grad_norm": 0.279296875, "learning_rate": 1.380175658720201e-06, "loss": 0.8781, "step": 2061 }, { "epoch": 2.587202007528231, "grad_norm": 0.333984375, "learning_rate": 1.3759933082392304e-06, "loss": 0.9916, "step": 2062 }, { "epoch": 2.5884567126725218, "grad_norm": 0.2734375, "learning_rate": 1.3718109577582603e-06, "loss": 0.9345, "step": 2063 }, { "epoch": 2.589711417816813, "grad_norm": 0.64453125, "learning_rate": 1.36762860727729e-06, "loss": 0.9482, "step": 2064 }, { "epoch": 2.5909661229611043, "grad_norm": 0.26171875, "learning_rate": 1.3634462567963196e-06, "loss": 0.8968, "step": 2065 }, { "epoch": 2.592220828105395, "grad_norm": 0.3046875, "learning_rate": 1.3592639063153493e-06, "loss": 0.9595, "step": 2066 }, { "epoch": 2.5934755332496864, "grad_norm": 0.275390625, "learning_rate": 1.3550815558343792e-06, "loss": 0.9458, "step": 2067 }, { "epoch": 2.5947302383939777, "grad_norm": 0.40625, "learning_rate": 1.3508992053534086e-06, "loss": 0.9143, "step": 2068 }, { "epoch": 2.5959849435382685, "grad_norm": 0.2890625, "learning_rate": 1.3467168548724385e-06, "loss": 0.9851, "step": 2069 }, { "epoch": 2.5972396486825597, "grad_norm": 0.2890625, "learning_rate": 1.342534504391468e-06, "loss": 0.9363, "step": 2070 }, { "epoch": 2.5984943538268506, "grad_norm": 0.275390625, "learning_rate": 1.3383521539104979e-06, "loss": 0.9332, "step": 2071 }, { "epoch": 2.599749058971142, "grad_norm": 0.287109375, "learning_rate": 1.3341698034295276e-06, "loss": 0.9306, "step": 2072 }, { "epoch": 2.6010037641154327, "grad_norm": 0.27734375, "learning_rate": 1.329987452948557e-06, "loss": 0.9456, "step": 2073 }, { "epoch": 2.602258469259724, "grad_norm": 0.27734375, "learning_rate": 1.325805102467587e-06, "loss": 0.9359, "step": 2074 }, { "epoch": 2.603513174404015, "grad_norm": 0.298828125, "learning_rate": 1.3216227519866166e-06, "loss": 0.9337, "step": 2075 }, { "epoch": 2.604767879548306, "grad_norm": 0.294921875, "learning_rate": 1.3174404015056463e-06, "loss": 0.9267, "step": 2076 }, { "epoch": 2.6060225846925973, "grad_norm": 0.287109375, "learning_rate": 1.313258051024676e-06, "loss": 0.9524, "step": 2077 }, { "epoch": 2.6072772898368886, "grad_norm": 0.271484375, "learning_rate": 1.3090757005437058e-06, "loss": 0.9204, "step": 2078 }, { "epoch": 2.6085319949811794, "grad_norm": 0.28125, "learning_rate": 1.3048933500627353e-06, "loss": 0.9037, "step": 2079 }, { "epoch": 2.6097867001254706, "grad_norm": 0.2734375, "learning_rate": 1.3007109995817652e-06, "loss": 0.9835, "step": 2080 }, { "epoch": 2.6110414052697615, "grad_norm": 0.275390625, "learning_rate": 1.2965286491007946e-06, "loss": 0.8875, "step": 2081 }, { "epoch": 2.6122961104140527, "grad_norm": 0.2734375, "learning_rate": 1.2923462986198245e-06, "loss": 0.8999, "step": 2082 }, { "epoch": 2.6135508155583436, "grad_norm": 0.27734375, "learning_rate": 1.2881639481388542e-06, "loss": 0.9731, "step": 2083 }, { "epoch": 2.614805520702635, "grad_norm": 0.287109375, "learning_rate": 1.2839815976578837e-06, "loss": 1.0183, "step": 2084 }, { "epoch": 2.616060225846926, "grad_norm": 0.33203125, "learning_rate": 1.2797992471769136e-06, "loss": 0.9179, "step": 2085 }, { "epoch": 2.617314930991217, "grad_norm": 0.275390625, "learning_rate": 1.2756168966959432e-06, "loss": 0.8909, "step": 2086 }, { "epoch": 2.618569636135508, "grad_norm": 0.271484375, "learning_rate": 1.271434546214973e-06, "loss": 0.9504, "step": 2087 }, { "epoch": 2.6198243412797995, "grad_norm": 0.267578125, "learning_rate": 1.2672521957340026e-06, "loss": 0.9086, "step": 2088 }, { "epoch": 2.6210790464240903, "grad_norm": 0.2734375, "learning_rate": 1.2630698452530325e-06, "loss": 0.93, "step": 2089 }, { "epoch": 2.6223337515683816, "grad_norm": 0.296875, "learning_rate": 1.258887494772062e-06, "loss": 0.9288, "step": 2090 }, { "epoch": 2.6235884567126724, "grad_norm": 0.283203125, "learning_rate": 1.2547051442910918e-06, "loss": 0.9553, "step": 2091 }, { "epoch": 2.6248431618569636, "grad_norm": 0.29296875, "learning_rate": 1.2505227938101213e-06, "loss": 0.9534, "step": 2092 }, { "epoch": 2.6260978670012545, "grad_norm": 0.28125, "learning_rate": 1.2463404433291512e-06, "loss": 0.9268, "step": 2093 }, { "epoch": 2.6273525721455457, "grad_norm": 0.265625, "learning_rate": 1.2421580928481808e-06, "loss": 0.927, "step": 2094 }, { "epoch": 2.628607277289837, "grad_norm": 0.408203125, "learning_rate": 1.2379757423672105e-06, "loss": 0.9137, "step": 2095 }, { "epoch": 2.629861982434128, "grad_norm": 0.26953125, "learning_rate": 1.2337933918862402e-06, "loss": 0.9283, "step": 2096 }, { "epoch": 2.631116687578419, "grad_norm": 0.271484375, "learning_rate": 1.2296110414052699e-06, "loss": 0.9184, "step": 2097 }, { "epoch": 2.6323713927227104, "grad_norm": 0.267578125, "learning_rate": 1.2254286909242996e-06, "loss": 0.9209, "step": 2098 }, { "epoch": 2.633626097867001, "grad_norm": 0.279296875, "learning_rate": 1.2212463404433292e-06, "loss": 0.9093, "step": 2099 }, { "epoch": 2.6348808030112925, "grad_norm": 0.28125, "learning_rate": 1.217063989962359e-06, "loss": 0.9423, "step": 2100 }, { "epoch": 2.6361355081555833, "grad_norm": 0.28515625, "learning_rate": 1.2128816394813886e-06, "loss": 1.0115, "step": 2101 }, { "epoch": 2.6373902132998746, "grad_norm": 0.263671875, "learning_rate": 1.2086992890004183e-06, "loss": 0.9021, "step": 2102 }, { "epoch": 2.6386449184441654, "grad_norm": 0.275390625, "learning_rate": 1.204516938519448e-06, "loss": 0.9063, "step": 2103 }, { "epoch": 2.6398996235884566, "grad_norm": 0.271484375, "learning_rate": 1.2003345880384778e-06, "loss": 0.9358, "step": 2104 }, { "epoch": 2.641154328732748, "grad_norm": 0.27734375, "learning_rate": 1.1961522375575075e-06, "loss": 0.92, "step": 2105 }, { "epoch": 2.6424090338770387, "grad_norm": 0.29296875, "learning_rate": 1.1919698870765372e-06, "loss": 0.9505, "step": 2106 }, { "epoch": 2.64366373902133, "grad_norm": 0.2734375, "learning_rate": 1.1877875365955668e-06, "loss": 0.9286, "step": 2107 }, { "epoch": 2.6449184441656213, "grad_norm": 0.279296875, "learning_rate": 1.1836051861145965e-06, "loss": 0.8938, "step": 2108 }, { "epoch": 2.646173149309912, "grad_norm": 0.267578125, "learning_rate": 1.1794228356336262e-06, "loss": 0.9245, "step": 2109 }, { "epoch": 2.6474278544542034, "grad_norm": 0.29296875, "learning_rate": 1.1752404851526559e-06, "loss": 0.932, "step": 2110 }, { "epoch": 2.648682559598494, "grad_norm": 0.26953125, "learning_rate": 1.1710581346716855e-06, "loss": 0.8881, "step": 2111 }, { "epoch": 2.6499372647427855, "grad_norm": 0.283203125, "learning_rate": 1.1668757841907152e-06, "loss": 0.9708, "step": 2112 }, { "epoch": 2.6511919698870763, "grad_norm": 0.330078125, "learning_rate": 1.162693433709745e-06, "loss": 0.9489, "step": 2113 }, { "epoch": 2.6524466750313676, "grad_norm": 0.28125, "learning_rate": 1.1585110832287746e-06, "loss": 0.9455, "step": 2114 }, { "epoch": 2.653701380175659, "grad_norm": 0.2734375, "learning_rate": 1.1543287327478042e-06, "loss": 0.9174, "step": 2115 }, { "epoch": 2.6549560853199496, "grad_norm": 0.296875, "learning_rate": 1.1501463822668341e-06, "loss": 0.9963, "step": 2116 }, { "epoch": 2.656210790464241, "grad_norm": 0.283203125, "learning_rate": 1.1459640317858638e-06, "loss": 0.9364, "step": 2117 }, { "epoch": 2.657465495608532, "grad_norm": 0.341796875, "learning_rate": 1.1417816813048935e-06, "loss": 0.9811, "step": 2118 }, { "epoch": 2.658720200752823, "grad_norm": 0.271484375, "learning_rate": 1.1375993308239232e-06, "loss": 0.9274, "step": 2119 }, { "epoch": 2.6599749058971143, "grad_norm": 0.275390625, "learning_rate": 1.1334169803429528e-06, "loss": 0.909, "step": 2120 }, { "epoch": 2.6612296110414055, "grad_norm": 0.275390625, "learning_rate": 1.1292346298619825e-06, "loss": 0.9159, "step": 2121 }, { "epoch": 2.6624843161856964, "grad_norm": 0.359375, "learning_rate": 1.1250522793810122e-06, "loss": 0.8821, "step": 2122 }, { "epoch": 2.663739021329987, "grad_norm": 0.29296875, "learning_rate": 1.1208699289000419e-06, "loss": 0.9224, "step": 2123 }, { "epoch": 2.6649937264742785, "grad_norm": 0.279296875, "learning_rate": 1.1166875784190715e-06, "loss": 0.9099, "step": 2124 }, { "epoch": 2.6662484316185697, "grad_norm": 0.373046875, "learning_rate": 1.1125052279381012e-06, "loss": 0.9275, "step": 2125 }, { "epoch": 2.6675031367628605, "grad_norm": 0.2734375, "learning_rate": 1.1083228774571309e-06, "loss": 0.9203, "step": 2126 }, { "epoch": 2.668757841907152, "grad_norm": 0.271484375, "learning_rate": 1.1041405269761608e-06, "loss": 0.8892, "step": 2127 }, { "epoch": 2.670012547051443, "grad_norm": 0.26953125, "learning_rate": 1.0999581764951905e-06, "loss": 0.8857, "step": 2128 }, { "epoch": 2.671267252195734, "grad_norm": 0.2734375, "learning_rate": 1.0957758260142201e-06, "loss": 0.9331, "step": 2129 }, { "epoch": 2.672521957340025, "grad_norm": 0.28125, "learning_rate": 1.0915934755332498e-06, "loss": 0.9514, "step": 2130 }, { "epoch": 2.6737766624843164, "grad_norm": 0.275390625, "learning_rate": 1.0874111250522795e-06, "loss": 0.9628, "step": 2131 }, { "epoch": 2.6750313676286073, "grad_norm": 0.26953125, "learning_rate": 1.0832287745713092e-06, "loss": 0.9198, "step": 2132 }, { "epoch": 2.676286072772898, "grad_norm": 0.29296875, "learning_rate": 1.0790464240903388e-06, "loss": 0.9225, "step": 2133 }, { "epoch": 2.6775407779171894, "grad_norm": 0.265625, "learning_rate": 1.0748640736093685e-06, "loss": 0.9027, "step": 2134 }, { "epoch": 2.6787954830614806, "grad_norm": 0.26953125, "learning_rate": 1.0706817231283982e-06, "loss": 0.945, "step": 2135 }, { "epoch": 2.6800501882057715, "grad_norm": 0.275390625, "learning_rate": 1.0664993726474279e-06, "loss": 0.9212, "step": 2136 }, { "epoch": 2.6813048933500627, "grad_norm": 0.275390625, "learning_rate": 1.0623170221664575e-06, "loss": 0.8945, "step": 2137 }, { "epoch": 2.682559598494354, "grad_norm": 0.263671875, "learning_rate": 1.0581346716854874e-06, "loss": 0.9185, "step": 2138 }, { "epoch": 2.683814303638645, "grad_norm": 0.314453125, "learning_rate": 1.053952321204517e-06, "loss": 0.9318, "step": 2139 }, { "epoch": 2.685069008782936, "grad_norm": 0.275390625, "learning_rate": 1.0497699707235468e-06, "loss": 0.9204, "step": 2140 }, { "epoch": 2.6863237139272274, "grad_norm": 0.26953125, "learning_rate": 1.0455876202425764e-06, "loss": 0.9491, "step": 2141 }, { "epoch": 2.687578419071518, "grad_norm": 0.26953125, "learning_rate": 1.0414052697616061e-06, "loss": 0.9751, "step": 2142 }, { "epoch": 2.6888331242158094, "grad_norm": 0.2890625, "learning_rate": 1.0372229192806358e-06, "loss": 0.9215, "step": 2143 }, { "epoch": 2.6900878293601003, "grad_norm": 0.298828125, "learning_rate": 1.0330405687996655e-06, "loss": 0.9811, "step": 2144 }, { "epoch": 2.6913425345043915, "grad_norm": 0.275390625, "learning_rate": 1.0288582183186952e-06, "loss": 0.9425, "step": 2145 }, { "epoch": 2.6925972396486824, "grad_norm": 0.26953125, "learning_rate": 1.0246758678377248e-06, "loss": 0.8885, "step": 2146 }, { "epoch": 2.6938519447929736, "grad_norm": 0.296875, "learning_rate": 1.0204935173567545e-06, "loss": 0.9412, "step": 2147 }, { "epoch": 2.695106649937265, "grad_norm": 0.291015625, "learning_rate": 1.0163111668757842e-06, "loss": 0.91, "step": 2148 }, { "epoch": 2.6963613550815557, "grad_norm": 0.265625, "learning_rate": 1.012128816394814e-06, "loss": 0.944, "step": 2149 }, { "epoch": 2.697616060225847, "grad_norm": 0.28125, "learning_rate": 1.0079464659138437e-06, "loss": 0.9842, "step": 2150 }, { "epoch": 2.6988707653701383, "grad_norm": 0.283203125, "learning_rate": 1.0037641154328734e-06, "loss": 0.9036, "step": 2151 }, { "epoch": 2.700125470514429, "grad_norm": 0.275390625, "learning_rate": 9.99581764951903e-07, "loss": 0.9235, "step": 2152 }, { "epoch": 2.7013801756587204, "grad_norm": 0.275390625, "learning_rate": 9.953994144709328e-07, "loss": 0.9648, "step": 2153 }, { "epoch": 2.702634880803011, "grad_norm": 0.35546875, "learning_rate": 9.912170639899624e-07, "loss": 0.8707, "step": 2154 }, { "epoch": 2.7038895859473024, "grad_norm": 0.263671875, "learning_rate": 9.870347135089921e-07, "loss": 0.9022, "step": 2155 }, { "epoch": 2.7051442910915933, "grad_norm": 0.291015625, "learning_rate": 9.828523630280218e-07, "loss": 0.9194, "step": 2156 }, { "epoch": 2.7063989962358845, "grad_norm": 0.279296875, "learning_rate": 9.786700125470515e-07, "loss": 0.8878, "step": 2157 }, { "epoch": 2.707653701380176, "grad_norm": 0.283203125, "learning_rate": 9.744876620660811e-07, "loss": 0.9852, "step": 2158 }, { "epoch": 2.7089084065244666, "grad_norm": 0.27734375, "learning_rate": 9.703053115851108e-07, "loss": 0.9316, "step": 2159 }, { "epoch": 2.710163111668758, "grad_norm": 0.275390625, "learning_rate": 9.661229611041407e-07, "loss": 0.9701, "step": 2160 }, { "epoch": 2.711417816813049, "grad_norm": 0.294921875, "learning_rate": 9.619406106231704e-07, "loss": 0.9119, "step": 2161 }, { "epoch": 2.71267252195734, "grad_norm": 0.28515625, "learning_rate": 9.577582601422e-07, "loss": 0.9529, "step": 2162 }, { "epoch": 2.7139272271016313, "grad_norm": 0.328125, "learning_rate": 9.535759096612297e-07, "loss": 0.9056, "step": 2163 }, { "epoch": 2.715181932245922, "grad_norm": 0.28125, "learning_rate": 9.493935591802594e-07, "loss": 0.9989, "step": 2164 }, { "epoch": 2.7164366373902133, "grad_norm": 0.279296875, "learning_rate": 9.452112086992892e-07, "loss": 0.9028, "step": 2165 }, { "epoch": 2.717691342534504, "grad_norm": 0.275390625, "learning_rate": 9.410288582183188e-07, "loss": 0.9499, "step": 2166 }, { "epoch": 2.7189460476787954, "grad_norm": 0.283203125, "learning_rate": 9.368465077373484e-07, "loss": 0.9628, "step": 2167 }, { "epoch": 2.7202007528230867, "grad_norm": 0.283203125, "learning_rate": 9.326641572563781e-07, "loss": 0.9506, "step": 2168 }, { "epoch": 2.7214554579673775, "grad_norm": 0.275390625, "learning_rate": 9.284818067754078e-07, "loss": 0.9082, "step": 2169 }, { "epoch": 2.722710163111669, "grad_norm": 0.2734375, "learning_rate": 9.242994562944376e-07, "loss": 0.9489, "step": 2170 }, { "epoch": 2.72396486825596, "grad_norm": 0.275390625, "learning_rate": 9.201171058134672e-07, "loss": 0.9698, "step": 2171 }, { "epoch": 2.725219573400251, "grad_norm": 0.2890625, "learning_rate": 9.159347553324969e-07, "loss": 0.9496, "step": 2172 }, { "epoch": 2.726474278544542, "grad_norm": 0.302734375, "learning_rate": 9.117524048515267e-07, "loss": 0.9838, "step": 2173 }, { "epoch": 2.727728983688833, "grad_norm": 0.3671875, "learning_rate": 9.075700543705564e-07, "loss": 0.9112, "step": 2174 }, { "epoch": 2.7289836888331243, "grad_norm": 0.271484375, "learning_rate": 9.033877038895861e-07, "loss": 0.9517, "step": 2175 }, { "epoch": 2.730238393977415, "grad_norm": 0.298828125, "learning_rate": 8.992053534086158e-07, "loss": 0.9741, "step": 2176 }, { "epoch": 2.7314930991217063, "grad_norm": 0.283203125, "learning_rate": 8.950230029276453e-07, "loss": 0.919, "step": 2177 }, { "epoch": 2.7327478042659976, "grad_norm": 0.283203125, "learning_rate": 8.908406524466751e-07, "loss": 0.9735, "step": 2178 }, { "epoch": 2.7340025094102884, "grad_norm": 0.275390625, "learning_rate": 8.866583019657048e-07, "loss": 0.9824, "step": 2179 }, { "epoch": 2.7352572145545797, "grad_norm": 0.275390625, "learning_rate": 8.824759514847344e-07, "loss": 0.9458, "step": 2180 }, { "epoch": 2.736511919698871, "grad_norm": 0.271484375, "learning_rate": 8.782936010037642e-07, "loss": 0.9131, "step": 2181 }, { "epoch": 2.737766624843162, "grad_norm": 0.26953125, "learning_rate": 8.741112505227939e-07, "loss": 0.9191, "step": 2182 }, { "epoch": 2.739021329987453, "grad_norm": 0.275390625, "learning_rate": 8.699289000418236e-07, "loss": 0.9655, "step": 2183 }, { "epoch": 2.740276035131744, "grad_norm": 0.326171875, "learning_rate": 8.657465495608533e-07, "loss": 0.9029, "step": 2184 }, { "epoch": 2.741530740276035, "grad_norm": 0.283203125, "learning_rate": 8.61564199079883e-07, "loss": 1.0018, "step": 2185 }, { "epoch": 2.742785445420326, "grad_norm": 0.271484375, "learning_rate": 8.573818485989127e-07, "loss": 0.9067, "step": 2186 }, { "epoch": 2.7440401505646173, "grad_norm": 0.267578125, "learning_rate": 8.531994981179424e-07, "loss": 0.9514, "step": 2187 }, { "epoch": 2.7452948557089085, "grad_norm": 0.275390625, "learning_rate": 8.490171476369719e-07, "loss": 0.9137, "step": 2188 }, { "epoch": 2.7465495608531993, "grad_norm": 0.271484375, "learning_rate": 8.448347971560017e-07, "loss": 0.9494, "step": 2189 }, { "epoch": 2.7478042659974906, "grad_norm": 0.267578125, "learning_rate": 8.406524466750314e-07, "loss": 0.9267, "step": 2190 }, { "epoch": 2.749058971141782, "grad_norm": 0.279296875, "learning_rate": 8.364700961940611e-07, "loss": 0.9321, "step": 2191 }, { "epoch": 2.7503136762860727, "grad_norm": 0.314453125, "learning_rate": 8.322877457130909e-07, "loss": 0.9336, "step": 2192 }, { "epoch": 2.751568381430364, "grad_norm": 0.26953125, "learning_rate": 8.281053952321205e-07, "loss": 0.9381, "step": 2193 }, { "epoch": 2.752823086574655, "grad_norm": 0.271484375, "learning_rate": 8.239230447511502e-07, "loss": 0.9339, "step": 2194 }, { "epoch": 2.754077791718946, "grad_norm": 0.271484375, "learning_rate": 8.1974069427018e-07, "loss": 0.9682, "step": 2195 }, { "epoch": 2.755332496863237, "grad_norm": 0.283203125, "learning_rate": 8.155583437892097e-07, "loss": 0.9544, "step": 2196 }, { "epoch": 2.756587202007528, "grad_norm": 0.265625, "learning_rate": 8.113759933082393e-07, "loss": 0.9066, "step": 2197 }, { "epoch": 2.7578419071518194, "grad_norm": 0.279296875, "learning_rate": 8.07193642827269e-07, "loss": 0.9109, "step": 2198 }, { "epoch": 2.7590966122961103, "grad_norm": 0.265625, "learning_rate": 8.030112923462986e-07, "loss": 0.8929, "step": 2199 }, { "epoch": 2.7603513174404015, "grad_norm": 0.3671875, "learning_rate": 7.988289418653284e-07, "loss": 0.9335, "step": 2200 }, { "epoch": 2.761606022584693, "grad_norm": 0.283203125, "learning_rate": 7.94646591384358e-07, "loss": 0.9349, "step": 2201 }, { "epoch": 2.7628607277289836, "grad_norm": 0.279296875, "learning_rate": 7.904642409033877e-07, "loss": 0.9292, "step": 2202 }, { "epoch": 2.764115432873275, "grad_norm": 0.2734375, "learning_rate": 7.862818904224175e-07, "loss": 0.9164, "step": 2203 }, { "epoch": 2.765370138017566, "grad_norm": 0.275390625, "learning_rate": 7.820995399414472e-07, "loss": 0.9526, "step": 2204 }, { "epoch": 2.766624843161857, "grad_norm": 0.26953125, "learning_rate": 7.779171894604768e-07, "loss": 0.9068, "step": 2205 }, { "epoch": 2.767879548306148, "grad_norm": 0.2890625, "learning_rate": 7.737348389795065e-07, "loss": 0.9028, "step": 2206 }, { "epoch": 2.769134253450439, "grad_norm": 0.28125, "learning_rate": 7.695524884985363e-07, "loss": 0.9025, "step": 2207 }, { "epoch": 2.7703889585947303, "grad_norm": 0.2734375, "learning_rate": 7.65370138017566e-07, "loss": 0.9765, "step": 2208 }, { "epoch": 2.771643663739021, "grad_norm": 0.29296875, "learning_rate": 7.611877875365957e-07, "loss": 0.9404, "step": 2209 }, { "epoch": 2.7728983688833124, "grad_norm": 0.28125, "learning_rate": 7.570054370556252e-07, "loss": 0.9315, "step": 2210 }, { "epoch": 2.7741530740276037, "grad_norm": 0.27734375, "learning_rate": 7.52823086574655e-07, "loss": 0.9556, "step": 2211 }, { "epoch": 2.7754077791718945, "grad_norm": 0.412109375, "learning_rate": 7.486407360936847e-07, "loss": 0.8987, "step": 2212 }, { "epoch": 2.776662484316186, "grad_norm": 0.267578125, "learning_rate": 7.444583856127144e-07, "loss": 0.8878, "step": 2213 }, { "epoch": 2.777917189460477, "grad_norm": 0.26953125, "learning_rate": 7.402760351317441e-07, "loss": 0.9204, "step": 2214 }, { "epoch": 2.779171894604768, "grad_norm": 0.275390625, "learning_rate": 7.360936846507738e-07, "loss": 0.9311, "step": 2215 }, { "epoch": 2.7804265997490587, "grad_norm": 0.345703125, "learning_rate": 7.319113341698035e-07, "loss": 0.9749, "step": 2216 }, { "epoch": 2.78168130489335, "grad_norm": 0.2734375, "learning_rate": 7.277289836888332e-07, "loss": 0.946, "step": 2217 }, { "epoch": 2.7829360100376412, "grad_norm": 0.361328125, "learning_rate": 7.23546633207863e-07, "loss": 0.9437, "step": 2218 }, { "epoch": 2.784190715181932, "grad_norm": 0.287109375, "learning_rate": 7.193642827268926e-07, "loss": 0.927, "step": 2219 }, { "epoch": 2.7854454203262233, "grad_norm": 0.263671875, "learning_rate": 7.151819322459223e-07, "loss": 0.9082, "step": 2220 }, { "epoch": 2.7867001254705146, "grad_norm": 0.283203125, "learning_rate": 7.109995817649519e-07, "loss": 0.9415, "step": 2221 }, { "epoch": 2.7879548306148054, "grad_norm": 0.318359375, "learning_rate": 7.068172312839817e-07, "loss": 0.9161, "step": 2222 }, { "epoch": 2.7892095357590967, "grad_norm": 0.26953125, "learning_rate": 7.026348808030113e-07, "loss": 0.9419, "step": 2223 }, { "epoch": 2.790464240903388, "grad_norm": 0.28125, "learning_rate": 6.98452530322041e-07, "loss": 0.9717, "step": 2224 }, { "epoch": 2.791718946047679, "grad_norm": 0.279296875, "learning_rate": 6.942701798410707e-07, "loss": 0.9818, "step": 2225 }, { "epoch": 2.79297365119197, "grad_norm": 0.27734375, "learning_rate": 6.900878293601005e-07, "loss": 0.9518, "step": 2226 }, { "epoch": 2.794228356336261, "grad_norm": 0.271484375, "learning_rate": 6.859054788791301e-07, "loss": 0.9224, "step": 2227 }, { "epoch": 2.795483061480552, "grad_norm": 0.296875, "learning_rate": 6.817231283981598e-07, "loss": 0.9384, "step": 2228 }, { "epoch": 2.796737766624843, "grad_norm": 0.30078125, "learning_rate": 6.775407779171896e-07, "loss": 0.9166, "step": 2229 }, { "epoch": 2.7979924717691342, "grad_norm": 0.271484375, "learning_rate": 6.733584274362193e-07, "loss": 0.9337, "step": 2230 }, { "epoch": 2.7992471769134255, "grad_norm": 0.291015625, "learning_rate": 6.691760769552489e-07, "loss": 0.9116, "step": 2231 }, { "epoch": 2.8005018820577163, "grad_norm": 0.287109375, "learning_rate": 6.649937264742785e-07, "loss": 0.9278, "step": 2232 }, { "epoch": 2.8017565872020076, "grad_norm": 0.265625, "learning_rate": 6.608113759933083e-07, "loss": 0.9171, "step": 2233 }, { "epoch": 2.803011292346299, "grad_norm": 0.283203125, "learning_rate": 6.56629025512338e-07, "loss": 0.9688, "step": 2234 }, { "epoch": 2.8042659974905897, "grad_norm": 0.27734375, "learning_rate": 6.524466750313676e-07, "loss": 0.8943, "step": 2235 }, { "epoch": 2.805520702634881, "grad_norm": 0.3046875, "learning_rate": 6.482643245503973e-07, "loss": 0.9008, "step": 2236 }, { "epoch": 2.806775407779172, "grad_norm": 0.3515625, "learning_rate": 6.440819740694271e-07, "loss": 0.9344, "step": 2237 }, { "epoch": 2.808030112923463, "grad_norm": 0.287109375, "learning_rate": 6.398996235884568e-07, "loss": 0.9681, "step": 2238 }, { "epoch": 2.809284818067754, "grad_norm": 0.2734375, "learning_rate": 6.357172731074865e-07, "loss": 0.9022, "step": 2239 }, { "epoch": 2.810539523212045, "grad_norm": 0.29296875, "learning_rate": 6.315349226265162e-07, "loss": 0.9756, "step": 2240 }, { "epoch": 2.8117942283563364, "grad_norm": 0.2890625, "learning_rate": 6.273525721455459e-07, "loss": 0.9407, "step": 2241 }, { "epoch": 2.8130489335006272, "grad_norm": 0.3671875, "learning_rate": 6.231702216645756e-07, "loss": 0.9302, "step": 2242 }, { "epoch": 2.8143036386449185, "grad_norm": 0.27734375, "learning_rate": 6.189878711836053e-07, "loss": 0.9227, "step": 2243 }, { "epoch": 2.8155583437892098, "grad_norm": 0.27734375, "learning_rate": 6.148055207026349e-07, "loss": 0.9253, "step": 2244 }, { "epoch": 2.8168130489335006, "grad_norm": 0.26953125, "learning_rate": 6.106231702216646e-07, "loss": 0.9536, "step": 2245 }, { "epoch": 2.818067754077792, "grad_norm": 0.26953125, "learning_rate": 6.064408197406943e-07, "loss": 0.9391, "step": 2246 }, { "epoch": 2.8193224592220827, "grad_norm": 0.28515625, "learning_rate": 6.02258469259724e-07, "loss": 0.8954, "step": 2247 }, { "epoch": 2.820577164366374, "grad_norm": 0.30078125, "learning_rate": 5.980761187787537e-07, "loss": 0.9091, "step": 2248 }, { "epoch": 2.821831869510665, "grad_norm": 0.2734375, "learning_rate": 5.938937682977834e-07, "loss": 0.9142, "step": 2249 }, { "epoch": 2.823086574654956, "grad_norm": 0.27734375, "learning_rate": 5.897114178168131e-07, "loss": 0.9652, "step": 2250 }, { "epoch": 2.8243412797992473, "grad_norm": 0.28125, "learning_rate": 5.855290673358428e-07, "loss": 0.9405, "step": 2251 }, { "epoch": 2.825595984943538, "grad_norm": 0.28515625, "learning_rate": 5.813467168548724e-07, "loss": 0.9485, "step": 2252 }, { "epoch": 2.8268506900878294, "grad_norm": 0.271484375, "learning_rate": 5.771643663739021e-07, "loss": 0.9341, "step": 2253 }, { "epoch": 2.8281053952321207, "grad_norm": 0.279296875, "learning_rate": 5.729820158929319e-07, "loss": 0.95, "step": 2254 }, { "epoch": 2.8293601003764115, "grad_norm": 0.275390625, "learning_rate": 5.687996654119616e-07, "loss": 0.9477, "step": 2255 }, { "epoch": 2.8306148055207028, "grad_norm": 0.296875, "learning_rate": 5.646173149309913e-07, "loss": 0.925, "step": 2256 }, { "epoch": 2.8318695106649936, "grad_norm": 0.283203125, "learning_rate": 5.604349644500209e-07, "loss": 0.9665, "step": 2257 }, { "epoch": 2.833124215809285, "grad_norm": 0.291015625, "learning_rate": 5.562526139690506e-07, "loss": 0.9504, "step": 2258 }, { "epoch": 2.8343789209535757, "grad_norm": 0.28125, "learning_rate": 5.520702634880804e-07, "loss": 0.9573, "step": 2259 }, { "epoch": 2.835633626097867, "grad_norm": 0.28125, "learning_rate": 5.478879130071101e-07, "loss": 0.9414, "step": 2260 }, { "epoch": 2.8368883312421582, "grad_norm": 0.27734375, "learning_rate": 5.437055625261397e-07, "loss": 0.9369, "step": 2261 }, { "epoch": 2.838143036386449, "grad_norm": 0.267578125, "learning_rate": 5.395232120451694e-07, "loss": 0.9277, "step": 2262 }, { "epoch": 2.8393977415307403, "grad_norm": 0.283203125, "learning_rate": 5.353408615641991e-07, "loss": 0.9431, "step": 2263 }, { "epoch": 2.8406524466750316, "grad_norm": 0.271484375, "learning_rate": 5.311585110832288e-07, "loss": 0.905, "step": 2264 }, { "epoch": 2.8419071518193224, "grad_norm": 0.265625, "learning_rate": 5.269761606022585e-07, "loss": 0.8929, "step": 2265 }, { "epoch": 2.8431618569636137, "grad_norm": 0.28515625, "learning_rate": 5.227938101212882e-07, "loss": 0.9712, "step": 2266 }, { "epoch": 2.8444165621079045, "grad_norm": 0.326171875, "learning_rate": 5.186114596403179e-07, "loss": 0.8926, "step": 2267 }, { "epoch": 2.8456712672521958, "grad_norm": 0.271484375, "learning_rate": 5.144291091593476e-07, "loss": 0.9065, "step": 2268 }, { "epoch": 2.8469259723964866, "grad_norm": 0.271484375, "learning_rate": 5.102467586783773e-07, "loss": 0.9164, "step": 2269 }, { "epoch": 2.848180677540778, "grad_norm": 0.26953125, "learning_rate": 5.06064408197407e-07, "loss": 0.9337, "step": 2270 }, { "epoch": 2.849435382685069, "grad_norm": 0.271484375, "learning_rate": 5.018820577164367e-07, "loss": 0.9492, "step": 2271 }, { "epoch": 2.85069008782936, "grad_norm": 0.296875, "learning_rate": 4.976997072354664e-07, "loss": 0.9189, "step": 2272 }, { "epoch": 2.851944792973651, "grad_norm": 0.27734375, "learning_rate": 4.935173567544961e-07, "loss": 0.9291, "step": 2273 }, { "epoch": 2.8531994981179425, "grad_norm": 0.279296875, "learning_rate": 4.893350062735257e-07, "loss": 0.9298, "step": 2274 }, { "epoch": 2.8544542032622333, "grad_norm": 0.2734375, "learning_rate": 4.851526557925554e-07, "loss": 0.9488, "step": 2275 }, { "epoch": 2.8557089084065246, "grad_norm": 0.2734375, "learning_rate": 4.809703053115852e-07, "loss": 0.9419, "step": 2276 }, { "epoch": 2.8569636135508154, "grad_norm": 0.26953125, "learning_rate": 4.7678795483061487e-07, "loss": 0.9021, "step": 2277 }, { "epoch": 2.8582183186951067, "grad_norm": 0.287109375, "learning_rate": 4.726056043496446e-07, "loss": 0.9516, "step": 2278 }, { "epoch": 2.8594730238393975, "grad_norm": 0.275390625, "learning_rate": 4.684232538686742e-07, "loss": 0.9134, "step": 2279 }, { "epoch": 2.8607277289836888, "grad_norm": 0.31640625, "learning_rate": 4.642409033877039e-07, "loss": 0.971, "step": 2280 }, { "epoch": 2.86198243412798, "grad_norm": 0.265625, "learning_rate": 4.600585529067336e-07, "loss": 0.921, "step": 2281 }, { "epoch": 2.863237139272271, "grad_norm": 0.271484375, "learning_rate": 4.5587620242576335e-07, "loss": 0.9135, "step": 2282 }, { "epoch": 2.864491844416562, "grad_norm": 0.28125, "learning_rate": 4.5169385194479303e-07, "loss": 0.9776, "step": 2283 }, { "epoch": 2.8657465495608534, "grad_norm": 0.275390625, "learning_rate": 4.4751150146382265e-07, "loss": 0.9733, "step": 2284 }, { "epoch": 2.867001254705144, "grad_norm": 0.283203125, "learning_rate": 4.433291509828524e-07, "loss": 0.997, "step": 2285 }, { "epoch": 2.8682559598494355, "grad_norm": 0.27734375, "learning_rate": 4.391468005018821e-07, "loss": 0.9538, "step": 2286 }, { "epoch": 2.8695106649937263, "grad_norm": 0.28515625, "learning_rate": 4.349644500209118e-07, "loss": 0.9604, "step": 2287 }, { "epoch": 2.8707653701380176, "grad_norm": 0.337890625, "learning_rate": 4.307820995399415e-07, "loss": 0.9054, "step": 2288 }, { "epoch": 2.8720200752823084, "grad_norm": 0.330078125, "learning_rate": 4.265997490589712e-07, "loss": 1.045, "step": 2289 }, { "epoch": 2.8732747804265997, "grad_norm": 0.27734375, "learning_rate": 4.2241739857800086e-07, "loss": 0.9048, "step": 2290 }, { "epoch": 2.874529485570891, "grad_norm": 0.302734375, "learning_rate": 4.1823504809703054e-07, "loss": 0.9704, "step": 2291 }, { "epoch": 2.8757841907151818, "grad_norm": 0.279296875, "learning_rate": 4.1405269761606027e-07, "loss": 0.9772, "step": 2292 }, { "epoch": 2.877038895859473, "grad_norm": 0.333984375, "learning_rate": 4.0987034713509e-07, "loss": 0.9174, "step": 2293 }, { "epoch": 2.8782936010037643, "grad_norm": 0.279296875, "learning_rate": 4.0568799665411967e-07, "loss": 0.9103, "step": 2294 }, { "epoch": 2.879548306148055, "grad_norm": 0.26953125, "learning_rate": 4.015056461731493e-07, "loss": 0.9249, "step": 2295 }, { "epoch": 2.8808030112923464, "grad_norm": 0.275390625, "learning_rate": 3.97323295692179e-07, "loss": 0.9691, "step": 2296 }, { "epoch": 2.8820577164366377, "grad_norm": 0.291015625, "learning_rate": 3.9314094521120875e-07, "loss": 0.9395, "step": 2297 }, { "epoch": 2.8833124215809285, "grad_norm": 0.279296875, "learning_rate": 3.889585947302384e-07, "loss": 0.9584, "step": 2298 }, { "epoch": 2.8845671267252193, "grad_norm": 0.28515625, "learning_rate": 3.8477624424926815e-07, "loss": 0.912, "step": 2299 }, { "epoch": 2.8858218318695106, "grad_norm": 0.27734375, "learning_rate": 3.8059389376829783e-07, "loss": 0.9355, "step": 2300 }, { "epoch": 2.887076537013802, "grad_norm": 0.2734375, "learning_rate": 3.764115432873275e-07, "loss": 0.9262, "step": 2301 }, { "epoch": 2.8883312421580927, "grad_norm": 0.28125, "learning_rate": 3.722291928063572e-07, "loss": 0.9455, "step": 2302 }, { "epoch": 2.889585947302384, "grad_norm": 0.265625, "learning_rate": 3.680468423253869e-07, "loss": 0.915, "step": 2303 }, { "epoch": 2.890840652446675, "grad_norm": 0.29296875, "learning_rate": 3.638644918444166e-07, "loss": 0.9607, "step": 2304 }, { "epoch": 2.892095357590966, "grad_norm": 0.326171875, "learning_rate": 3.596821413634463e-07, "loss": 0.8943, "step": 2305 }, { "epoch": 2.8933500627352573, "grad_norm": 0.2890625, "learning_rate": 3.5549979088247594e-07, "loss": 0.9591, "step": 2306 }, { "epoch": 2.8946047678795486, "grad_norm": 0.28125, "learning_rate": 3.5131744040150566e-07, "loss": 0.9482, "step": 2307 }, { "epoch": 2.8958594730238394, "grad_norm": 0.287109375, "learning_rate": 3.4713508992053534e-07, "loss": 0.8961, "step": 2308 }, { "epoch": 2.89711417816813, "grad_norm": 0.296875, "learning_rate": 3.4295273943956507e-07, "loss": 0.9621, "step": 2309 }, { "epoch": 2.8983688833124215, "grad_norm": 0.28125, "learning_rate": 3.387703889585948e-07, "loss": 0.9417, "step": 2310 }, { "epoch": 2.8996235884567128, "grad_norm": 0.275390625, "learning_rate": 3.3458803847762447e-07, "loss": 0.9142, "step": 2311 }, { "epoch": 2.9008782936010036, "grad_norm": 0.283203125, "learning_rate": 3.3040568799665415e-07, "loss": 0.9075, "step": 2312 }, { "epoch": 2.902132998745295, "grad_norm": 0.279296875, "learning_rate": 3.262233375156838e-07, "loss": 0.9089, "step": 2313 }, { "epoch": 2.903387703889586, "grad_norm": 0.28125, "learning_rate": 3.2204098703471355e-07, "loss": 0.9534, "step": 2314 }, { "epoch": 2.904642409033877, "grad_norm": 0.27734375, "learning_rate": 3.1785863655374323e-07, "loss": 0.9169, "step": 2315 }, { "epoch": 2.905897114178168, "grad_norm": 0.283203125, "learning_rate": 3.1367628607277296e-07, "loss": 0.92, "step": 2316 }, { "epoch": 2.9071518193224595, "grad_norm": 0.271484375, "learning_rate": 3.0949393559180263e-07, "loss": 0.8963, "step": 2317 }, { "epoch": 2.9084065244667503, "grad_norm": 0.2734375, "learning_rate": 3.053115851108323e-07, "loss": 0.9457, "step": 2318 }, { "epoch": 2.9096612296110416, "grad_norm": 0.267578125, "learning_rate": 3.01129234629862e-07, "loss": 0.9618, "step": 2319 }, { "epoch": 2.9109159347553324, "grad_norm": 0.41015625, "learning_rate": 2.969468841488917e-07, "loss": 0.9677, "step": 2320 }, { "epoch": 2.9121706398996237, "grad_norm": 0.271484375, "learning_rate": 2.927645336679214e-07, "loss": 0.945, "step": 2321 }, { "epoch": 2.9134253450439145, "grad_norm": 0.298828125, "learning_rate": 2.8858218318695106e-07, "loss": 0.9639, "step": 2322 }, { "epoch": 2.9146800501882058, "grad_norm": 0.298828125, "learning_rate": 2.843998327059808e-07, "loss": 0.9382, "step": 2323 }, { "epoch": 2.915934755332497, "grad_norm": 0.291015625, "learning_rate": 2.8021748222501047e-07, "loss": 0.8885, "step": 2324 }, { "epoch": 2.917189460476788, "grad_norm": 0.267578125, "learning_rate": 2.760351317440402e-07, "loss": 0.9141, "step": 2325 }, { "epoch": 2.918444165621079, "grad_norm": 0.265625, "learning_rate": 2.7185278126306987e-07, "loss": 0.9067, "step": 2326 }, { "epoch": 2.9196988707653704, "grad_norm": 0.361328125, "learning_rate": 2.6767043078209955e-07, "loss": 0.9234, "step": 2327 }, { "epoch": 2.920953575909661, "grad_norm": 0.259765625, "learning_rate": 2.634880803011293e-07, "loss": 0.9012, "step": 2328 }, { "epoch": 2.9222082810539525, "grad_norm": 0.279296875, "learning_rate": 2.5930572982015895e-07, "loss": 0.96, "step": 2329 }, { "epoch": 2.9234629861982433, "grad_norm": 0.318359375, "learning_rate": 2.551233793391886e-07, "loss": 0.9008, "step": 2330 }, { "epoch": 2.9247176913425346, "grad_norm": 0.275390625, "learning_rate": 2.5094102885821835e-07, "loss": 0.9554, "step": 2331 }, { "epoch": 2.9259723964868254, "grad_norm": 0.279296875, "learning_rate": 2.4675867837724803e-07, "loss": 0.9419, "step": 2332 }, { "epoch": 2.9272271016311167, "grad_norm": 0.2734375, "learning_rate": 2.425763278962777e-07, "loss": 0.9343, "step": 2333 }, { "epoch": 2.928481806775408, "grad_norm": 0.28515625, "learning_rate": 2.3839397741530743e-07, "loss": 0.9601, "step": 2334 }, { "epoch": 2.9297365119196987, "grad_norm": 0.279296875, "learning_rate": 2.342116269343371e-07, "loss": 0.9207, "step": 2335 }, { "epoch": 2.93099121706399, "grad_norm": 0.271484375, "learning_rate": 2.300292764533668e-07, "loss": 0.9389, "step": 2336 }, { "epoch": 2.9322459222082813, "grad_norm": 0.283203125, "learning_rate": 2.2584692597239651e-07, "loss": 0.9008, "step": 2337 }, { "epoch": 2.933500627352572, "grad_norm": 0.26953125, "learning_rate": 2.216645754914262e-07, "loss": 0.9548, "step": 2338 }, { "epoch": 2.9347553324968634, "grad_norm": 0.287109375, "learning_rate": 2.174822250104559e-07, "loss": 0.941, "step": 2339 }, { "epoch": 2.936010037641154, "grad_norm": 0.27734375, "learning_rate": 2.132998745294856e-07, "loss": 0.9351, "step": 2340 }, { "epoch": 2.9372647427854455, "grad_norm": 0.279296875, "learning_rate": 2.0911752404851527e-07, "loss": 0.9306, "step": 2341 }, { "epoch": 2.9385194479297363, "grad_norm": 0.2890625, "learning_rate": 2.04935173567545e-07, "loss": 0.9248, "step": 2342 }, { "epoch": 2.9397741530740276, "grad_norm": 0.275390625, "learning_rate": 2.0075282308657465e-07, "loss": 0.9182, "step": 2343 }, { "epoch": 2.941028858218319, "grad_norm": 0.28515625, "learning_rate": 1.9657047260560437e-07, "loss": 0.9667, "step": 2344 }, { "epoch": 2.9422835633626097, "grad_norm": 0.275390625, "learning_rate": 1.9238812212463408e-07, "loss": 0.9448, "step": 2345 }, { "epoch": 2.943538268506901, "grad_norm": 0.287109375, "learning_rate": 1.8820577164366375e-07, "loss": 0.9342, "step": 2346 }, { "epoch": 2.944792973651192, "grad_norm": 0.2734375, "learning_rate": 1.8402342116269345e-07, "loss": 0.9463, "step": 2347 }, { "epoch": 2.946047678795483, "grad_norm": 0.2734375, "learning_rate": 1.7984107068172316e-07, "loss": 0.9347, "step": 2348 }, { "epoch": 2.9473023839397743, "grad_norm": 0.28125, "learning_rate": 1.7565872020075283e-07, "loss": 0.914, "step": 2349 }, { "epoch": 2.948557089084065, "grad_norm": 0.271484375, "learning_rate": 1.7147636971978253e-07, "loss": 0.9427, "step": 2350 }, { "epoch": 2.9498117942283564, "grad_norm": 0.275390625, "learning_rate": 1.6729401923881224e-07, "loss": 0.9051, "step": 2351 }, { "epoch": 2.951066499372647, "grad_norm": 0.271484375, "learning_rate": 1.631116687578419e-07, "loss": 0.9228, "step": 2352 }, { "epoch": 2.9523212045169385, "grad_norm": 0.26953125, "learning_rate": 1.5892931827687161e-07, "loss": 0.8927, "step": 2353 }, { "epoch": 2.9535759096612297, "grad_norm": 0.27734375, "learning_rate": 1.5474696779590132e-07, "loss": 0.9175, "step": 2354 }, { "epoch": 2.9548306148055206, "grad_norm": 0.283203125, "learning_rate": 1.50564617314931e-07, "loss": 0.9215, "step": 2355 }, { "epoch": 2.956085319949812, "grad_norm": 0.283203125, "learning_rate": 1.463822668339607e-07, "loss": 0.9054, "step": 2356 }, { "epoch": 2.957340025094103, "grad_norm": 0.275390625, "learning_rate": 1.421999163529904e-07, "loss": 0.8834, "step": 2357 }, { "epoch": 2.958594730238394, "grad_norm": 0.2734375, "learning_rate": 1.380175658720201e-07, "loss": 0.9043, "step": 2358 }, { "epoch": 2.959849435382685, "grad_norm": 0.2734375, "learning_rate": 1.3383521539104977e-07, "loss": 0.9147, "step": 2359 }, { "epoch": 2.961104140526976, "grad_norm": 0.28125, "learning_rate": 1.2965286491007947e-07, "loss": 0.9578, "step": 2360 }, { "epoch": 2.9623588456712673, "grad_norm": 0.283203125, "learning_rate": 1.2547051442910918e-07, "loss": 0.9625, "step": 2361 }, { "epoch": 2.963613550815558, "grad_norm": 0.26953125, "learning_rate": 1.2128816394813885e-07, "loss": 0.922, "step": 2362 }, { "epoch": 2.9648682559598494, "grad_norm": 0.28125, "learning_rate": 1.1710581346716855e-07, "loss": 0.9337, "step": 2363 }, { "epoch": 2.9661229611041406, "grad_norm": 0.275390625, "learning_rate": 1.1292346298619826e-07, "loss": 0.9468, "step": 2364 }, { "epoch": 2.9673776662484315, "grad_norm": 0.287109375, "learning_rate": 1.0874111250522795e-07, "loss": 0.9574, "step": 2365 }, { "epoch": 2.9686323713927227, "grad_norm": 0.3984375, "learning_rate": 1.0455876202425763e-07, "loss": 1.0053, "step": 2366 }, { "epoch": 2.969887076537014, "grad_norm": 0.294921875, "learning_rate": 1.0037641154328732e-07, "loss": 0.8973, "step": 2367 }, { "epoch": 2.971141781681305, "grad_norm": 0.302734375, "learning_rate": 9.619406106231704e-08, "loss": 0.9581, "step": 2368 }, { "epoch": 2.972396486825596, "grad_norm": 0.291015625, "learning_rate": 9.201171058134673e-08, "loss": 0.984, "step": 2369 }, { "epoch": 2.973651191969887, "grad_norm": 0.27734375, "learning_rate": 8.782936010037642e-08, "loss": 0.9351, "step": 2370 }, { "epoch": 2.974905897114178, "grad_norm": 0.279296875, "learning_rate": 8.364700961940612e-08, "loss": 0.9003, "step": 2371 }, { "epoch": 2.976160602258469, "grad_norm": 0.265625, "learning_rate": 7.946465913843581e-08, "loss": 0.9248, "step": 2372 }, { "epoch": 2.9774153074027603, "grad_norm": 0.26953125, "learning_rate": 7.52823086574655e-08, "loss": 0.8858, "step": 2373 }, { "epoch": 2.9786700125470515, "grad_norm": 0.27734375, "learning_rate": 7.10999581764952e-08, "loss": 0.9311, "step": 2374 }, { "epoch": 2.9799247176913424, "grad_norm": 0.3125, "learning_rate": 6.691760769552489e-08, "loss": 0.9075, "step": 2375 }, { "epoch": 2.9811794228356336, "grad_norm": 0.28125, "learning_rate": 6.273525721455459e-08, "loss": 0.8862, "step": 2376 }, { "epoch": 2.982434127979925, "grad_norm": 0.26953125, "learning_rate": 5.855290673358428e-08, "loss": 0.8948, "step": 2377 }, { "epoch": 2.9836888331242157, "grad_norm": 0.275390625, "learning_rate": 5.437055625261397e-08, "loss": 0.9584, "step": 2378 }, { "epoch": 2.984943538268507, "grad_norm": 0.271484375, "learning_rate": 5.018820577164366e-08, "loss": 0.932, "step": 2379 }, { "epoch": 2.9861982434127983, "grad_norm": 0.2890625, "learning_rate": 4.6005855290673364e-08, "loss": 0.9452, "step": 2380 }, { "epoch": 2.987452948557089, "grad_norm": 0.267578125, "learning_rate": 4.182350480970306e-08, "loss": 0.908, "step": 2381 }, { "epoch": 2.98870765370138, "grad_norm": 0.283203125, "learning_rate": 3.764115432873275e-08, "loss": 0.9475, "step": 2382 }, { "epoch": 2.989962358845671, "grad_norm": 0.279296875, "learning_rate": 3.345880384776244e-08, "loss": 0.9028, "step": 2383 }, { "epoch": 2.9912170639899625, "grad_norm": 0.279296875, "learning_rate": 2.927645336679214e-08, "loss": 0.8816, "step": 2384 }, { "epoch": 2.9924717691342533, "grad_norm": 0.283203125, "learning_rate": 2.509410288582183e-08, "loss": 0.9463, "step": 2385 }, { "epoch": 2.9937264742785445, "grad_norm": 0.2890625, "learning_rate": 2.091175240485153e-08, "loss": 0.9191, "step": 2386 }, { "epoch": 2.994981179422836, "grad_norm": 0.2734375, "learning_rate": 1.672940192388122e-08, "loss": 0.9308, "step": 2387 }, { "epoch": 2.9962358845671266, "grad_norm": 0.2734375, "learning_rate": 1.2547051442910915e-08, "loss": 0.9233, "step": 2388 }, { "epoch": 2.997490589711418, "grad_norm": 0.28515625, "learning_rate": 8.36470096194061e-09, "loss": 0.9911, "step": 2389 }, { "epoch": 2.998745294855709, "grad_norm": 0.2890625, "learning_rate": 4.182350480970305e-09, "loss": 0.9596, "step": 2390 }, { "epoch": 3.0, "grad_norm": 0.283203125, "learning_rate": 0.0, "loss": 0.9366, "step": 2391 } ], "logging_steps": 1.0, "max_steps": 2391, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7.677047380121747e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }