diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,9833 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.236421725239617, + "eval_steps": 500, + "global_step": 1400, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.001597444089456869, + "grad_norm": 128.6180419921875, + "learning_rate": 0.0, + "loss": 1.9747, + "step": 1 + }, + { + "epoch": 0.003194888178913738, + "grad_norm": 127.44181060791016, + "learning_rate": 7.628243530057963e-08, + "loss": 1.9176, + "step": 2 + }, + { + "epoch": 0.004792332268370607, + "grad_norm": 130.29476928710938, + "learning_rate": 1.2090479941510652e-07, + "loss": 1.9397, + "step": 3 + }, + { + "epoch": 0.006389776357827476, + "grad_norm": 128.3740234375, + "learning_rate": 1.5256487060115926e-07, + "loss": 1.9046, + "step": 4 + }, + { + "epoch": 0.007987220447284345, + "grad_norm": 136.06153869628906, + "learning_rate": 1.7712232967084336e-07, + "loss": 1.947, + "step": 5 + }, + { + "epoch": 0.009584664536741214, + "grad_norm": 134.7310333251953, + "learning_rate": 1.9718723471568615e-07, + "loss": 1.9161, + "step": 6 + }, + { + "epoch": 0.011182108626198083, + "grad_norm": 136.66058349609375, + "learning_rate": 2.14151870207623e-07, + "loss": 1.779, + "step": 7 + }, + { + "epoch": 0.012779552715654952, + "grad_norm": 142.6785888671875, + "learning_rate": 2.288473059017389e-07, + "loss": 1.7793, + "step": 8 + }, + { + "epoch": 0.01437699680511182, + "grad_norm": 139.76708984375, + "learning_rate": 2.4180959883021303e-07, + "loss": 1.7003, + "step": 9 + }, + { + "epoch": 0.01597444089456869, + "grad_norm": 147.7816162109375, + "learning_rate": 2.53404764971423e-07, + "loss": 1.742, + "step": 10 + }, + { + "epoch": 0.01757188498402556, + "grad_norm": 159.0968780517578, + "learning_rate": 2.638938686254791e-07, + "loss": 1.6079, + "step": 11 + }, + { + "epoch": 0.019169329073482427, + "grad_norm": 138.03765869140625, + "learning_rate": 2.7346967001626583e-07, + "loss": 1.1859, + "step": 12 + }, + { + "epoch": 0.020766773162939296, + "grad_norm": 130.79318237304688, + "learning_rate": 2.8227855338279304e-07, + "loss": 1.1544, + "step": 13 + }, + { + "epoch": 0.022364217252396165, + "grad_norm": 128.29359436035156, + "learning_rate": 2.904343055082026e-07, + "loss": 1.0924, + "step": 14 + }, + { + "epoch": 0.023961661341853034, + "grad_norm": 118.21534729003906, + "learning_rate": 2.9802712908594985e-07, + "loss": 1.0143, + "step": 15 + }, + { + "epoch": 0.025559105431309903, + "grad_norm": 110.80689239501953, + "learning_rate": 3.051297412023185e-07, + "loss": 0.941, + "step": 16 + }, + { + "epoch": 0.027156549520766772, + "grad_norm": 112.3552474975586, + "learning_rate": 3.1180161973120243e-07, + "loss": 0.9826, + "step": 17 + }, + { + "epoch": 0.02875399361022364, + "grad_norm": 59.84565353393555, + "learning_rate": 3.1809203413079266e-07, + "loss": 0.6646, + "step": 18 + }, + { + "epoch": 0.03035143769968051, + "grad_norm": 61.63838195800781, + "learning_rate": 3.2404225570581245e-07, + "loss": 0.535, + "step": 19 + }, + { + "epoch": 0.03194888178913738, + "grad_norm": 57.68476486206055, + "learning_rate": 3.2968720027200265e-07, + "loss": 0.451, + "step": 20 + }, + { + "epoch": 0.03354632587859425, + "grad_norm": 58.70195388793945, + "learning_rate": 3.350566696227295e-07, + "loss": 0.4398, + "step": 21 + }, + { + "epoch": 0.03514376996805112, + "grad_norm": 60.67943572998047, + "learning_rate": 3.401763039260588e-07, + "loss": 0.3542, + "step": 22 + }, + { + "epoch": 0.036741214057507986, + "grad_norm": 54.30119705200195, + "learning_rate": 3.450683222410826e-07, + "loss": 0.3362, + "step": 23 + }, + { + "epoch": 0.038338658146964855, + "grad_norm": 27.211286544799805, + "learning_rate": 3.4975210531684546e-07, + "loss": 0.3206, + "step": 24 + }, + { + "epoch": 0.039936102236421724, + "grad_norm": 20.16046905517578, + "learning_rate": 3.542446593416867e-07, + "loss": 0.2747, + "step": 25 + }, + { + "epoch": 0.04153354632587859, + "grad_norm": 18.191621780395508, + "learning_rate": 3.585609886833727e-07, + "loss": 0.2485, + "step": 26 + }, + { + "epoch": 0.04313099041533546, + "grad_norm": 17.691118240356445, + "learning_rate": 3.627143982453195e-07, + "loss": 0.2113, + "step": 27 + }, + { + "epoch": 0.04472843450479233, + "grad_norm": 16.775856018066406, + "learning_rate": 3.6671674080878223e-07, + "loss": 0.2395, + "step": 28 + }, + { + "epoch": 0.0463258785942492, + "grad_norm": 11.414810180664062, + "learning_rate": 3.7057862095226456e-07, + "loss": 0.1754, + "step": 29 + }, + { + "epoch": 0.04792332268370607, + "grad_norm": 11.763614654541016, + "learning_rate": 3.7430956438652954e-07, + "loss": 0.171, + "step": 30 + }, + { + "epoch": 0.04952076677316294, + "grad_norm": 11.317436218261719, + "learning_rate": 3.7791815951345716e-07, + "loss": 0.1304, + "step": 31 + }, + { + "epoch": 0.051118210862619806, + "grad_norm": 11.866049766540527, + "learning_rate": 3.814121765028982e-07, + "loss": 0.1429, + "step": 32 + }, + { + "epoch": 0.052715654952076675, + "grad_norm": 7.949328899383545, + "learning_rate": 3.8479866804058563e-07, + "loss": 0.1008, + "step": 33 + }, + { + "epoch": 0.054313099041533544, + "grad_norm": 6.994161605834961, + "learning_rate": 3.880840550317821e-07, + "loss": 0.0818, + "step": 34 + }, + { + "epoch": 0.05591054313099041, + "grad_norm": 8.409013748168945, + "learning_rate": 3.912741998784663e-07, + "loss": 0.1117, + "step": 35 + }, + { + "epoch": 0.05750798722044728, + "grad_norm": 5.039752006530762, + "learning_rate": 3.943744694313723e-07, + "loss": 0.0769, + "step": 36 + }, + { + "epoch": 0.05910543130990415, + "grad_norm": 5.305444717407227, + "learning_rate": 3.973897893149772e-07, + "loss": 0.0547, + "step": 37 + }, + { + "epoch": 0.06070287539936102, + "grad_norm": 6.901062488555908, + "learning_rate": 4.003246910063921e-07, + "loss": 0.0888, + "step": 38 + }, + { + "epoch": 0.06230031948881789, + "grad_norm": 5.630770683288574, + "learning_rate": 4.031833527978995e-07, + "loss": 0.0664, + "step": 39 + }, + { + "epoch": 0.06389776357827476, + "grad_norm": 5.812962532043457, + "learning_rate": 4.059696355725823e-07, + "loss": 0.0643, + "step": 40 + }, + { + "epoch": 0.06549520766773163, + "grad_norm": 4.684011936187744, + "learning_rate": 4.0868711416176974e-07, + "loss": 0.0563, + "step": 41 + }, + { + "epoch": 0.0670926517571885, + "grad_norm": 5.2407050132751465, + "learning_rate": 4.113391049233091e-07, + "loss": 0.0546, + "step": 42 + }, + { + "epoch": 0.06869009584664537, + "grad_norm": 5.062054634094238, + "learning_rate": 4.139286900743784e-07, + "loss": 0.0488, + "step": 43 + }, + { + "epoch": 0.07028753993610223, + "grad_norm": 4.678827285766602, + "learning_rate": 4.1645873922663837e-07, + "loss": 0.039, + "step": 44 + }, + { + "epoch": 0.07188498402555911, + "grad_norm": 4.474348068237305, + "learning_rate": 4.1893192850105637e-07, + "loss": 0.0327, + "step": 45 + }, + { + "epoch": 0.07348242811501597, + "grad_norm": 5.580406665802002, + "learning_rate": 4.213507575416622e-07, + "loss": 0.0265, + "step": 46 + }, + { + "epoch": 0.07507987220447285, + "grad_norm": 4.145895957946777, + "learning_rate": 4.2371756469942035e-07, + "loss": 0.034, + "step": 47 + }, + { + "epoch": 0.07667731629392971, + "grad_norm": 2.864013671875, + "learning_rate": 4.260345406174251e-07, + "loss": 0.0195, + "step": 48 + }, + { + "epoch": 0.07827476038338659, + "grad_norm": 3.931077718734741, + "learning_rate": 4.28303740415246e-07, + "loss": 0.028, + "step": 49 + }, + { + "epoch": 0.07987220447284345, + "grad_norm": 3.2536861896514893, + "learning_rate": 4.305270946422664e-07, + "loss": 0.0167, + "step": 50 + }, + { + "epoch": 0.08146964856230032, + "grad_norm": 2.361112356185913, + "learning_rate": 4.3270641914630895e-07, + "loss": 0.0137, + "step": 51 + }, + { + "epoch": 0.08306709265175719, + "grad_norm": 3.177391529083252, + "learning_rate": 4.3484342398395235e-07, + "loss": 0.0168, + "step": 52 + }, + { + "epoch": 0.08466453674121406, + "grad_norm": 3.647589921951294, + "learning_rate": 4.3693972148208397e-07, + "loss": 0.0171, + "step": 53 + }, + { + "epoch": 0.08626198083067092, + "grad_norm": 3.1123926639556885, + "learning_rate": 4.389968335458992e-07, + "loss": 0.0096, + "step": 54 + }, + { + "epoch": 0.0878594249201278, + "grad_norm": 3.218924045562744, + "learning_rate": 4.4101619829632255e-07, + "loss": 0.0128, + "step": 55 + }, + { + "epoch": 0.08945686900958466, + "grad_norm": 3.422558307647705, + "learning_rate": 4.429991761093619e-07, + "loss": 0.0075, + "step": 56 + }, + { + "epoch": 0.09105431309904154, + "grad_norm": 4.4138288497924805, + "learning_rate": 4.44947055120919e-07, + "loss": 0.0106, + "step": 57 + }, + { + "epoch": 0.0926517571884984, + "grad_norm": 4.1317219734191895, + "learning_rate": 4.4686105625284414e-07, + "loss": 0.0147, + "step": 58 + }, + { + "epoch": 0.09424920127795527, + "grad_norm": 1.8419156074523926, + "learning_rate": 4.4874233780934923e-07, + "loss": 0.002, + "step": 59 + }, + { + "epoch": 0.09584664536741214, + "grad_norm": 5.771890640258789, + "learning_rate": 4.5059199968710917e-07, + "loss": 0.0132, + "step": 60 + }, + { + "epoch": 0.09744408945686901, + "grad_norm": 3.2066822052001953, + "learning_rate": 4.5241108723737287e-07, + "loss": 0.0121, + "step": 61 + }, + { + "epoch": 0.09904153354632587, + "grad_norm": 2.5449273586273193, + "learning_rate": 4.5420059481403685e-07, + "loss": 0.0112, + "step": 62 + }, + { + "epoch": 0.10063897763578275, + "grad_norm": 1.833320140838623, + "learning_rate": 4.55961469037836e-07, + "loss": 0.0057, + "step": 63 + }, + { + "epoch": 0.10223642172523961, + "grad_norm": 0.9052779674530029, + "learning_rate": 4.576946118034778e-07, + "loss": 0.0029, + "step": 64 + }, + { + "epoch": 0.10383386581469649, + "grad_norm": 1.7721703052520752, + "learning_rate": 4.594008830536363e-07, + "loss": 0.0073, + "step": 65 + }, + { + "epoch": 0.10543130990415335, + "grad_norm": 1.3743584156036377, + "learning_rate": 4.6108110334116526e-07, + "loss": 0.0042, + "step": 66 + }, + { + "epoch": 0.10702875399361023, + "grad_norm": 5.354936599731445, + "learning_rate": 4.627360561986405e-07, + "loss": 0.0067, + "step": 67 + }, + { + "epoch": 0.10862619808306709, + "grad_norm": 0.8433341979980469, + "learning_rate": 4.6436649033236175e-07, + "loss": 0.0036, + "step": 68 + }, + { + "epoch": 0.11022364217252396, + "grad_norm": 1.8066205978393555, + "learning_rate": 4.6597312165618915e-07, + "loss": 0.0065, + "step": 69 + }, + { + "epoch": 0.11182108626198083, + "grad_norm": 2.4936585426330566, + "learning_rate": 4.6755663517904604e-07, + "loss": 0.013, + "step": 70 + }, + { + "epoch": 0.1134185303514377, + "grad_norm": 1.3142043352127075, + "learning_rate": 4.691176867585419e-07, + "loss": 0.0106, + "step": 71 + }, + { + "epoch": 0.11501597444089456, + "grad_norm": 1.3923885822296143, + "learning_rate": 4.70656904731952e-07, + "loss": 0.0033, + "step": 72 + }, + { + "epoch": 0.11661341853035144, + "grad_norm": 1.4490530490875244, + "learning_rate": 4.721748914347038e-07, + "loss": 0.0079, + "step": 73 + }, + { + "epoch": 0.1182108626198083, + "grad_norm": 0.8172028660774231, + "learning_rate": 4.736722246155569e-07, + "loss": 0.0039, + "step": 74 + }, + { + "epoch": 0.11980830670926518, + "grad_norm": 1.627395510673523, + "learning_rate": 4.751494587567932e-07, + "loss": 0.0038, + "step": 75 + }, + { + "epoch": 0.12140575079872204, + "grad_norm": 1.3568391799926758, + "learning_rate": 4.7660712630697176e-07, + "loss": 0.0067, + "step": 76 + }, + { + "epoch": 0.12300319488817892, + "grad_norm": 1.5196893215179443, + "learning_rate": 4.780457388331021e-07, + "loss": 0.0064, + "step": 77 + }, + { + "epoch": 0.12460063897763578, + "grad_norm": 1.7342876195907593, + "learning_rate": 4.794657880984791e-07, + "loss": 0.0112, + "step": 78 + }, + { + "epoch": 0.12619808306709265, + "grad_norm": 1.1238106489181519, + "learning_rate": 4.808677470718594e-07, + "loss": 0.008, + "step": 79 + }, + { + "epoch": 0.12779552715654952, + "grad_norm": 0.6927244663238525, + "learning_rate": 4.822520708731619e-07, + "loss": 0.0028, + "step": 80 + }, + { + "epoch": 0.12939297124600638, + "grad_norm": 0.710464596748352, + "learning_rate": 4.836191976604261e-07, + "loss": 0.0046, + "step": 81 + }, + { + "epoch": 0.13099041533546327, + "grad_norm": 1.29885995388031, + "learning_rate": 4.849695494623494e-07, + "loss": 0.0028, + "step": 82 + }, + { + "epoch": 0.13258785942492013, + "grad_norm": 0.7128888964653015, + "learning_rate": 4.863035329603659e-07, + "loss": 0.0017, + "step": 83 + }, + { + "epoch": 0.134185303514377, + "grad_norm": 1.3970448970794678, + "learning_rate": 4.876215402238887e-07, + "loss": 0.007, + "step": 84 + }, + { + "epoch": 0.13578274760383385, + "grad_norm": 1.8840289115905762, + "learning_rate": 4.889239494020458e-07, + "loss": 0.0073, + "step": 85 + }, + { + "epoch": 0.13738019169329074, + "grad_norm": 0.9983999729156494, + "learning_rate": 4.902111253749581e-07, + "loss": 0.0008, + "step": 86 + }, + { + "epoch": 0.1389776357827476, + "grad_norm": 1.001712441444397, + "learning_rate": 4.91483420367371e-07, + "loss": 0.0045, + "step": 87 + }, + { + "epoch": 0.14057507987220447, + "grad_norm": 0.8022336959838867, + "learning_rate": 4.927411745272181e-07, + "loss": 0.005, + "step": 88 + }, + { + "epoch": 0.14217252396166133, + "grad_norm": 1.9497357606887817, + "learning_rate": 4.939847164714948e-07, + "loss": 0.0132, + "step": 89 + }, + { + "epoch": 0.14376996805111822, + "grad_norm": 0.4873954951763153, + "learning_rate": 4.952143638016361e-07, + "loss": 0.0036, + "step": 90 + }, + { + "epoch": 0.14536741214057508, + "grad_norm": 1.6238665580749512, + "learning_rate": 4.96430423590416e-07, + "loss": 0.0079, + "step": 91 + }, + { + "epoch": 0.14696485623003194, + "grad_norm": 2.2722995281219482, + "learning_rate": 4.976331928422418e-07, + "loss": 0.0112, + "step": 92 + }, + { + "epoch": 0.1485623003194888, + "grad_norm": 1.2474795579910278, + "learning_rate": 4.988229589285637e-07, + "loss": 0.0063, + "step": 93 + }, + { + "epoch": 0.1501597444089457, + "grad_norm": 1.0009785890579224, + "learning_rate": 5e-07, + "loss": 0.0085, + "step": 94 + }, + { + "epoch": 0.15175718849840256, + "grad_norm": 0.6343518495559692, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 95 + }, + { + "epoch": 0.15335463258785942, + "grad_norm": 0.4515174329280853, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 96 + }, + { + "epoch": 0.15495207667731628, + "grad_norm": 0.7897706627845764, + "learning_rate": 5e-07, + "loss": 0.0057, + "step": 97 + }, + { + "epoch": 0.15654952076677317, + "grad_norm": 1.2340161800384521, + "learning_rate": 5e-07, + "loss": 0.006, + "step": 98 + }, + { + "epoch": 0.15814696485623003, + "grad_norm": 0.5955281257629395, + "learning_rate": 5e-07, + "loss": 0.0043, + "step": 99 + }, + { + "epoch": 0.1597444089456869, + "grad_norm": 1.140373706817627, + "learning_rate": 5e-07, + "loss": 0.0094, + "step": 100 + }, + { + "epoch": 0.16134185303514376, + "grad_norm": 0.5656710267066956, + "learning_rate": 5e-07, + "loss": 0.0049, + "step": 101 + }, + { + "epoch": 0.16293929712460065, + "grad_norm": 1.3329702615737915, + "learning_rate": 5e-07, + "loss": 0.0102, + "step": 102 + }, + { + "epoch": 0.1645367412140575, + "grad_norm": 0.42556625604629517, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 103 + }, + { + "epoch": 0.16613418530351437, + "grad_norm": 0.6244504451751709, + "learning_rate": 5e-07, + "loss": 0.0046, + "step": 104 + }, + { + "epoch": 0.16773162939297126, + "grad_norm": 0.6920517086982727, + "learning_rate": 5e-07, + "loss": 0.0041, + "step": 105 + }, + { + "epoch": 0.16932907348242812, + "grad_norm": 0.660217821598053, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 106 + }, + { + "epoch": 0.17092651757188498, + "grad_norm": 1.0536694526672363, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 107 + }, + { + "epoch": 0.17252396166134185, + "grad_norm": 0.862185537815094, + "learning_rate": 5e-07, + "loss": 0.005, + "step": 108 + }, + { + "epoch": 0.17412140575079874, + "grad_norm": 0.33221954107284546, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 109 + }, + { + "epoch": 0.1757188498402556, + "grad_norm": 3.744462013244629, + "learning_rate": 5e-07, + "loss": 0.0047, + "step": 110 + }, + { + "epoch": 0.17731629392971246, + "grad_norm": 1.6666382551193237, + "learning_rate": 5e-07, + "loss": 0.0067, + "step": 111 + }, + { + "epoch": 0.17891373801916932, + "grad_norm": 0.3559662103652954, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 112 + }, + { + "epoch": 0.1805111821086262, + "grad_norm": 0.3662293255329132, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 113 + }, + { + "epoch": 0.18210862619808307, + "grad_norm": 0.49407243728637695, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 114 + }, + { + "epoch": 0.18370607028753994, + "grad_norm": 1.0715949535369873, + "learning_rate": 5e-07, + "loss": 0.0054, + "step": 115 + }, + { + "epoch": 0.1853035143769968, + "grad_norm": 0.618974506855011, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 116 + }, + { + "epoch": 0.1869009584664537, + "grad_norm": 1.6139345169067383, + "learning_rate": 5e-07, + "loss": 0.0089, + "step": 117 + }, + { + "epoch": 0.18849840255591055, + "grad_norm": 0.9241402745246887, + "learning_rate": 5e-07, + "loss": 0.0041, + "step": 118 + }, + { + "epoch": 0.1900958466453674, + "grad_norm": 0.7571372389793396, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 119 + }, + { + "epoch": 0.19169329073482427, + "grad_norm": 0.5674700736999512, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 120 + }, + { + "epoch": 0.19329073482428116, + "grad_norm": 2.3434317111968994, + "learning_rate": 5e-07, + "loss": 0.0113, + "step": 121 + }, + { + "epoch": 0.19488817891373802, + "grad_norm": 0.5285120010375977, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 122 + }, + { + "epoch": 0.1964856230031949, + "grad_norm": 0.3353559076786041, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 123 + }, + { + "epoch": 0.19808306709265175, + "grad_norm": 0.6090297698974609, + "learning_rate": 5e-07, + "loss": 0.0052, + "step": 124 + }, + { + "epoch": 0.19968051118210864, + "grad_norm": 0.6662172079086304, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 125 + }, + { + "epoch": 0.2012779552715655, + "grad_norm": 0.6109351515769958, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 126 + }, + { + "epoch": 0.20287539936102236, + "grad_norm": 1.2161667346954346, + "learning_rate": 5e-07, + "loss": 0.0077, + "step": 127 + }, + { + "epoch": 0.20447284345047922, + "grad_norm": 0.5375782251358032, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 128 + }, + { + "epoch": 0.20607028753993611, + "grad_norm": 0.8255560398101807, + "learning_rate": 5e-07, + "loss": 0.0046, + "step": 129 + }, + { + "epoch": 0.20766773162939298, + "grad_norm": 1.2899589538574219, + "learning_rate": 5e-07, + "loss": 0.0073, + "step": 130 + }, + { + "epoch": 0.20926517571884984, + "grad_norm": 0.4597526490688324, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 131 + }, + { + "epoch": 0.2108626198083067, + "grad_norm": 0.2904457151889801, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 132 + }, + { + "epoch": 0.2124600638977636, + "grad_norm": 0.2778950035572052, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 133 + }, + { + "epoch": 0.21405750798722045, + "grad_norm": 0.45779383182525635, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 134 + }, + { + "epoch": 0.21565495207667731, + "grad_norm": 1.7134373188018799, + "learning_rate": 5e-07, + "loss": 0.0073, + "step": 135 + }, + { + "epoch": 0.21725239616613418, + "grad_norm": 0.8295688033103943, + "learning_rate": 5e-07, + "loss": 0.005, + "step": 136 + }, + { + "epoch": 0.21884984025559107, + "grad_norm": 1.5231693983078003, + "learning_rate": 5e-07, + "loss": 0.0087, + "step": 137 + }, + { + "epoch": 0.22044728434504793, + "grad_norm": 0.6921976804733276, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 138 + }, + { + "epoch": 0.2220447284345048, + "grad_norm": 0.613254964351654, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 139 + }, + { + "epoch": 0.22364217252396165, + "grad_norm": 0.9046704769134521, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 140 + }, + { + "epoch": 0.22523961661341854, + "grad_norm": 0.827433705329895, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 141 + }, + { + "epoch": 0.2268370607028754, + "grad_norm": 0.42438846826553345, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 142 + }, + { + "epoch": 0.22843450479233227, + "grad_norm": 0.6912664771080017, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 143 + }, + { + "epoch": 0.23003194888178913, + "grad_norm": 0.6359020471572876, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 144 + }, + { + "epoch": 0.23162939297124602, + "grad_norm": 0.45392975211143494, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 145 + }, + { + "epoch": 0.23322683706070288, + "grad_norm": 0.2925805151462555, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 146 + }, + { + "epoch": 0.23482428115015974, + "grad_norm": 0.6216511130332947, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 147 + }, + { + "epoch": 0.2364217252396166, + "grad_norm": 1.4257566928863525, + "learning_rate": 5e-07, + "loss": 0.0049, + "step": 148 + }, + { + "epoch": 0.2380191693290735, + "grad_norm": 0.5238153338432312, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 149 + }, + { + "epoch": 0.23961661341853036, + "grad_norm": 0.4455612301826477, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 150 + }, + { + "epoch": 0.24121405750798722, + "grad_norm": 0.23266546428203583, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 151 + }, + { + "epoch": 0.24281150159744408, + "grad_norm": 0.3799627125263214, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 152 + }, + { + "epoch": 0.24440894568690097, + "grad_norm": 0.48497480154037476, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 153 + }, + { + "epoch": 0.24600638977635783, + "grad_norm": 0.6147518157958984, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 154 + }, + { + "epoch": 0.2476038338658147, + "grad_norm": 0.6456011533737183, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 155 + }, + { + "epoch": 0.24920127795527156, + "grad_norm": 0.5384286642074585, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 156 + }, + { + "epoch": 0.2507987220447284, + "grad_norm": 0.8868528008460999, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 157 + }, + { + "epoch": 0.2523961661341853, + "grad_norm": 0.25091347098350525, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 158 + }, + { + "epoch": 0.2539936102236422, + "grad_norm": 0.5789719820022583, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 159 + }, + { + "epoch": 0.25559105431309903, + "grad_norm": 1.5586203336715698, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 160 + }, + { + "epoch": 0.2571884984025559, + "grad_norm": 0.705531895160675, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 161 + }, + { + "epoch": 0.25878594249201275, + "grad_norm": 0.7628393173217773, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 162 + }, + { + "epoch": 0.26038338658146964, + "grad_norm": 0.8703659176826477, + "learning_rate": 5e-07, + "loss": 0.004, + "step": 163 + }, + { + "epoch": 0.26198083067092653, + "grad_norm": 0.3005010783672333, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 164 + }, + { + "epoch": 0.26357827476038337, + "grad_norm": 0.19402694702148438, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 165 + }, + { + "epoch": 0.26517571884984026, + "grad_norm": 2.271894931793213, + "learning_rate": 5e-07, + "loss": 0.006, + "step": 166 + }, + { + "epoch": 0.26677316293929715, + "grad_norm": 0.3555186092853546, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 167 + }, + { + "epoch": 0.268370607028754, + "grad_norm": 1.9556201696395874, + "learning_rate": 5e-07, + "loss": 0.0045, + "step": 168 + }, + { + "epoch": 0.26996805111821087, + "grad_norm": 1.7738531827926636, + "learning_rate": 5e-07, + "loss": 0.0072, + "step": 169 + }, + { + "epoch": 0.2715654952076677, + "grad_norm": 2.2895989418029785, + "learning_rate": 5e-07, + "loss": 0.0049, + "step": 170 + }, + { + "epoch": 0.2731629392971246, + "grad_norm": 1.7725859880447388, + "learning_rate": 5e-07, + "loss": 0.009, + "step": 171 + }, + { + "epoch": 0.2747603833865815, + "grad_norm": 0.5722166895866394, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 172 + }, + { + "epoch": 0.2763578274760383, + "grad_norm": 1.063871145248413, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 173 + }, + { + "epoch": 0.2779552715654952, + "grad_norm": 0.3389774560928345, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 174 + }, + { + "epoch": 0.2795527156549521, + "grad_norm": 0.45647579431533813, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 175 + }, + { + "epoch": 0.28115015974440893, + "grad_norm": 0.7398928999900818, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 176 + }, + { + "epoch": 0.2827476038338658, + "grad_norm": 0.7850615382194519, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 177 + }, + { + "epoch": 0.28434504792332266, + "grad_norm": 3.63692045211792, + "learning_rate": 5e-07, + "loss": 0.006, + "step": 178 + }, + { + "epoch": 0.28594249201277955, + "grad_norm": 0.44707363843917847, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 179 + }, + { + "epoch": 0.28753993610223644, + "grad_norm": 0.9111068248748779, + "learning_rate": 5e-07, + "loss": 0.0041, + "step": 180 + }, + { + "epoch": 0.28913738019169327, + "grad_norm": 0.6590073704719543, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 181 + }, + { + "epoch": 0.29073482428115016, + "grad_norm": 0.5553420782089233, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 182 + }, + { + "epoch": 0.29233226837060705, + "grad_norm": 0.47674331068992615, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 183 + }, + { + "epoch": 0.2939297124600639, + "grad_norm": 0.29876044392585754, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 184 + }, + { + "epoch": 0.2955271565495208, + "grad_norm": 0.6842983365058899, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 185 + }, + { + "epoch": 0.2971246006389776, + "grad_norm": 0.4935712516307831, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 186 + }, + { + "epoch": 0.2987220447284345, + "grad_norm": 1.2657763957977295, + "learning_rate": 5e-07, + "loss": 0.0076, + "step": 187 + }, + { + "epoch": 0.3003194888178914, + "grad_norm": 0.48839613795280457, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 188 + }, + { + "epoch": 0.3019169329073482, + "grad_norm": 1.063694715499878, + "learning_rate": 5e-07, + "loss": 0.0037, + "step": 189 + }, + { + "epoch": 0.3035143769968051, + "grad_norm": 0.7980586290359497, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 190 + }, + { + "epoch": 0.305111821086262, + "grad_norm": 0.45484256744384766, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 191 + }, + { + "epoch": 0.30670926517571884, + "grad_norm": 1.688433051109314, + "learning_rate": 5e-07, + "loss": 0.0056, + "step": 192 + }, + { + "epoch": 0.3083067092651757, + "grad_norm": 0.7312079071998596, + "learning_rate": 5e-07, + "loss": 0.0037, + "step": 193 + }, + { + "epoch": 0.30990415335463256, + "grad_norm": 0.4801974594593048, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 194 + }, + { + "epoch": 0.31150159744408945, + "grad_norm": 0.4754345118999481, + "learning_rate": 5e-07, + "loss": 0.0058, + "step": 195 + }, + { + "epoch": 0.31309904153354634, + "grad_norm": 2.1157238483428955, + "learning_rate": 5e-07, + "loss": 0.0091, + "step": 196 + }, + { + "epoch": 0.3146964856230032, + "grad_norm": 0.6343661546707153, + "learning_rate": 5e-07, + "loss": 0.0037, + "step": 197 + }, + { + "epoch": 0.31629392971246006, + "grad_norm": 0.5099918842315674, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 198 + }, + { + "epoch": 0.31789137380191695, + "grad_norm": 0.333951860666275, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 199 + }, + { + "epoch": 0.3194888178913738, + "grad_norm": 0.7018607258796692, + "learning_rate": 5e-07, + "loss": 0.0044, + "step": 200 + }, + { + "epoch": 0.3210862619808307, + "grad_norm": 0.6244291067123413, + "learning_rate": 5e-07, + "loss": 0.0043, + "step": 201 + }, + { + "epoch": 0.3226837060702875, + "grad_norm": 0.8400171399116516, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 202 + }, + { + "epoch": 0.3242811501597444, + "grad_norm": 1.1659806966781616, + "learning_rate": 5e-07, + "loss": 0.0047, + "step": 203 + }, + { + "epoch": 0.3258785942492013, + "grad_norm": 0.9981520175933838, + "learning_rate": 5e-07, + "loss": 0.0056, + "step": 204 + }, + { + "epoch": 0.3274760383386581, + "grad_norm": 0.32153695821762085, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 205 + }, + { + "epoch": 0.329073482428115, + "grad_norm": 3.8695359230041504, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 206 + }, + { + "epoch": 0.3306709265175719, + "grad_norm": 0.6529371738433838, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 207 + }, + { + "epoch": 0.33226837060702874, + "grad_norm": 0.7385216951370239, + "learning_rate": 5e-07, + "loss": 0.0048, + "step": 208 + }, + { + "epoch": 0.33386581469648563, + "grad_norm": 0.2696060836315155, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 209 + }, + { + "epoch": 0.3354632587859425, + "grad_norm": 2.1431872844696045, + "learning_rate": 5e-07, + "loss": 0.0104, + "step": 210 + }, + { + "epoch": 0.33706070287539935, + "grad_norm": 0.45158419013023376, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 211 + }, + { + "epoch": 0.33865814696485624, + "grad_norm": 0.28861358761787415, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 212 + }, + { + "epoch": 0.3402555910543131, + "grad_norm": 0.8536441326141357, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 213 + }, + { + "epoch": 0.34185303514376997, + "grad_norm": 0.3299103081226349, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 214 + }, + { + "epoch": 0.34345047923322686, + "grad_norm": 0.6404473185539246, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 215 + }, + { + "epoch": 0.3450479233226837, + "grad_norm": 0.2772029638290405, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 216 + }, + { + "epoch": 0.3466453674121406, + "grad_norm": 3.0106074810028076, + "learning_rate": 5e-07, + "loss": 0.0051, + "step": 217 + }, + { + "epoch": 0.34824281150159747, + "grad_norm": 1.7244184017181396, + "learning_rate": 5e-07, + "loss": 0.0066, + "step": 218 + }, + { + "epoch": 0.3498402555910543, + "grad_norm": 0.25184711813926697, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 219 + }, + { + "epoch": 0.3514376996805112, + "grad_norm": 0.6831691861152649, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 220 + }, + { + "epoch": 0.35303514376996803, + "grad_norm": 0.309230774641037, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 221 + }, + { + "epoch": 0.3546325878594249, + "grad_norm": 0.5623671412467957, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 222 + }, + { + "epoch": 0.3562300319488818, + "grad_norm": 0.31381672620773315, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 223 + }, + { + "epoch": 0.35782747603833864, + "grad_norm": 0.6246241927146912, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 224 + }, + { + "epoch": 0.35942492012779553, + "grad_norm": 0.9443689584732056, + "learning_rate": 5e-07, + "loss": 0.0056, + "step": 225 + }, + { + "epoch": 0.3610223642172524, + "grad_norm": 0.8880484104156494, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 226 + }, + { + "epoch": 0.36261980830670926, + "grad_norm": 0.9324328899383545, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 227 + }, + { + "epoch": 0.36421725239616615, + "grad_norm": 0.7231195569038391, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 228 + }, + { + "epoch": 0.365814696485623, + "grad_norm": 0.5298541188240051, + "learning_rate": 5e-07, + "loss": 0.0047, + "step": 229 + }, + { + "epoch": 0.36741214057507987, + "grad_norm": 1.5357270240783691, + "learning_rate": 5e-07, + "loss": 0.0054, + "step": 230 + }, + { + "epoch": 0.36900958466453676, + "grad_norm": 0.4374171495437622, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 231 + }, + { + "epoch": 0.3706070287539936, + "grad_norm": 0.2250860631465912, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 232 + }, + { + "epoch": 0.3722044728434505, + "grad_norm": 1.8229856491088867, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 233 + }, + { + "epoch": 0.3738019169329074, + "grad_norm": 1.0502337217330933, + "learning_rate": 5e-07, + "loss": 0.005, + "step": 234 + }, + { + "epoch": 0.3753993610223642, + "grad_norm": 1.1811085939407349, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 235 + }, + { + "epoch": 0.3769968051118211, + "grad_norm": 0.5202884674072266, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 236 + }, + { + "epoch": 0.37859424920127793, + "grad_norm": 0.29556524753570557, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 237 + }, + { + "epoch": 0.3801916932907348, + "grad_norm": 0.7253068685531616, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 238 + }, + { + "epoch": 0.3817891373801917, + "grad_norm": 0.9315080642700195, + "learning_rate": 5e-07, + "loss": 0.0044, + "step": 239 + }, + { + "epoch": 0.38338658146964855, + "grad_norm": 0.5284143090248108, + "learning_rate": 5e-07, + "loss": 0.0043, + "step": 240 + }, + { + "epoch": 0.38498402555910544, + "grad_norm": 0.15872539579868317, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 241 + }, + { + "epoch": 0.3865814696485623, + "grad_norm": 0.3570381999015808, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 242 + }, + { + "epoch": 0.38817891373801916, + "grad_norm": 1.1983716487884521, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 243 + }, + { + "epoch": 0.38977635782747605, + "grad_norm": 0.4511491656303406, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 244 + }, + { + "epoch": 0.3913738019169329, + "grad_norm": 0.21462838351726532, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 245 + }, + { + "epoch": 0.3929712460063898, + "grad_norm": 3.4422619342803955, + "learning_rate": 5e-07, + "loss": 0.0036, + "step": 246 + }, + { + "epoch": 0.39456869009584666, + "grad_norm": 2.029860496520996, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 247 + }, + { + "epoch": 0.3961661341853035, + "grad_norm": 0.4887297749519348, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 248 + }, + { + "epoch": 0.3977635782747604, + "grad_norm": 3.5348706245422363, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 249 + }, + { + "epoch": 0.3993610223642173, + "grad_norm": 0.12261940538883209, + "learning_rate": 5e-07, + "loss": 0.0006, + "step": 250 + }, + { + "epoch": 0.4009584664536741, + "grad_norm": 1.2506657838821411, + "learning_rate": 5e-07, + "loss": 0.0044, + "step": 251 + }, + { + "epoch": 0.402555910543131, + "grad_norm": 1.385932445526123, + "learning_rate": 5e-07, + "loss": 0.0071, + "step": 252 + }, + { + "epoch": 0.40415335463258784, + "grad_norm": 0.6065950393676758, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 253 + }, + { + "epoch": 0.4057507987220447, + "grad_norm": 0.3198873698711395, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 254 + }, + { + "epoch": 0.4073482428115016, + "grad_norm": 0.21003501117229462, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 255 + }, + { + "epoch": 0.40894568690095845, + "grad_norm": 0.70599764585495, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 256 + }, + { + "epoch": 0.41054313099041534, + "grad_norm": 0.27071261405944824, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 257 + }, + { + "epoch": 0.41214057507987223, + "grad_norm": 0.4544181525707245, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 258 + }, + { + "epoch": 0.41373801916932906, + "grad_norm": 0.3489130437374115, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 259 + }, + { + "epoch": 0.41533546325878595, + "grad_norm": 2.232909679412842, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 260 + }, + { + "epoch": 0.4169329073482428, + "grad_norm": 0.24653400480747223, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 261 + }, + { + "epoch": 0.4185303514376997, + "grad_norm": 0.70169597864151, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 262 + }, + { + "epoch": 0.42012779552715657, + "grad_norm": 0.26590874791145325, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 263 + }, + { + "epoch": 0.4217252396166134, + "grad_norm": 1.3357137441635132, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 264 + }, + { + "epoch": 0.4233226837060703, + "grad_norm": 0.17348721623420715, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 265 + }, + { + "epoch": 0.4249201277955272, + "grad_norm": 0.21971674263477325, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 266 + }, + { + "epoch": 0.426517571884984, + "grad_norm": 0.3568759262561798, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 267 + }, + { + "epoch": 0.4281150159744409, + "grad_norm": 0.22301608324050903, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 268 + }, + { + "epoch": 0.42971246006389774, + "grad_norm": 0.2780276834964752, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 269 + }, + { + "epoch": 0.43130990415335463, + "grad_norm": 0.399155855178833, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 270 + }, + { + "epoch": 0.4329073482428115, + "grad_norm": 0.35904622077941895, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 271 + }, + { + "epoch": 0.43450479233226835, + "grad_norm": 0.14762169122695923, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 272 + }, + { + "epoch": 0.43610223642172524, + "grad_norm": 0.26977628469467163, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 273 + }, + { + "epoch": 0.43769968051118213, + "grad_norm": 0.9291149377822876, + "learning_rate": 5e-07, + "loss": 0.004, + "step": 274 + }, + { + "epoch": 0.43929712460063897, + "grad_norm": 0.6625785827636719, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 275 + }, + { + "epoch": 0.44089456869009586, + "grad_norm": 0.8231006860733032, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 276 + }, + { + "epoch": 0.4424920127795527, + "grad_norm": 0.44783130288124084, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 277 + }, + { + "epoch": 0.4440894568690096, + "grad_norm": 0.8134899139404297, + "learning_rate": 5e-07, + "loss": 0.0036, + "step": 278 + }, + { + "epoch": 0.44568690095846647, + "grad_norm": 0.18782544136047363, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 279 + }, + { + "epoch": 0.4472843450479233, + "grad_norm": 2.3220651149749756, + "learning_rate": 5e-07, + "loss": 0.0057, + "step": 280 + }, + { + "epoch": 0.4488817891373802, + "grad_norm": 3.4821548461914062, + "learning_rate": 5e-07, + "loss": 0.0034, + "step": 281 + }, + { + "epoch": 0.4504792332268371, + "grad_norm": 1.2853020429611206, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 282 + }, + { + "epoch": 0.4520766773162939, + "grad_norm": 2.967637777328491, + "learning_rate": 5e-07, + "loss": 0.0045, + "step": 283 + }, + { + "epoch": 0.4536741214057508, + "grad_norm": 1.1847063302993774, + "learning_rate": 5e-07, + "loss": 0.0066, + "step": 284 + }, + { + "epoch": 0.45527156549520764, + "grad_norm": 0.6564436554908752, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 285 + }, + { + "epoch": 0.45686900958466453, + "grad_norm": 1.7592343091964722, + "learning_rate": 5e-07, + "loss": 0.0043, + "step": 286 + }, + { + "epoch": 0.4584664536741214, + "grad_norm": 0.21900290250778198, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 287 + }, + { + "epoch": 0.46006389776357826, + "grad_norm": 0.31253165006637573, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 288 + }, + { + "epoch": 0.46166134185303515, + "grad_norm": 0.7273770570755005, + "learning_rate": 5e-07, + "loss": 0.004, + "step": 289 + }, + { + "epoch": 0.46325878594249204, + "grad_norm": 0.9572885036468506, + "learning_rate": 5e-07, + "loss": 0.005, + "step": 290 + }, + { + "epoch": 0.46485623003194887, + "grad_norm": 4.609229564666748, + "learning_rate": 5e-07, + "loss": 0.0105, + "step": 291 + }, + { + "epoch": 0.46645367412140576, + "grad_norm": 0.7994480729103088, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 292 + }, + { + "epoch": 0.4680511182108626, + "grad_norm": 0.512992262840271, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 293 + }, + { + "epoch": 0.4696485623003195, + "grad_norm": 0.7132886052131653, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 294 + }, + { + "epoch": 0.4712460063897764, + "grad_norm": 0.4453645348548889, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 295 + }, + { + "epoch": 0.4728434504792332, + "grad_norm": 0.32878679037094116, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 296 + }, + { + "epoch": 0.4744408945686901, + "grad_norm": 0.4557294249534607, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 297 + }, + { + "epoch": 0.476038338658147, + "grad_norm": 0.7930683493614197, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 298 + }, + { + "epoch": 0.4776357827476038, + "grad_norm": 0.16142797470092773, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 299 + }, + { + "epoch": 0.4792332268370607, + "grad_norm": 0.16838717460632324, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 300 + }, + { + "epoch": 0.48083067092651754, + "grad_norm": 1.0744272470474243, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 301 + }, + { + "epoch": 0.48242811501597443, + "grad_norm": 1.177935004234314, + "learning_rate": 5e-07, + "loss": 0.0043, + "step": 302 + }, + { + "epoch": 0.4840255591054313, + "grad_norm": 0.33150434494018555, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 303 + }, + { + "epoch": 0.48562300319488816, + "grad_norm": 0.14007702469825745, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 304 + }, + { + "epoch": 0.48722044728434505, + "grad_norm": 0.23931963741779327, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 305 + }, + { + "epoch": 0.48881789137380194, + "grad_norm": 1.721514344215393, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 306 + }, + { + "epoch": 0.4904153354632588, + "grad_norm": 1.3192071914672852, + "learning_rate": 5e-07, + "loss": 0.0095, + "step": 307 + }, + { + "epoch": 0.49201277955271566, + "grad_norm": 6.429077625274658, + "learning_rate": 5e-07, + "loss": 0.011, + "step": 308 + }, + { + "epoch": 0.4936102236421725, + "grad_norm": 0.919262707233429, + "learning_rate": 5e-07, + "loss": 0.0066, + "step": 309 + }, + { + "epoch": 0.4952076677316294, + "grad_norm": 0.6547032594680786, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 310 + }, + { + "epoch": 0.4968051118210863, + "grad_norm": 0.487663209438324, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 311 + }, + { + "epoch": 0.4984025559105431, + "grad_norm": 0.33844229578971863, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 312 + }, + { + "epoch": 0.5, + "grad_norm": 0.686505138874054, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 313 + }, + { + "epoch": 0.5015974440894568, + "grad_norm": 0.2250794619321823, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 314 + }, + { + "epoch": 0.5031948881789138, + "grad_norm": 0.40359318256378174, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 315 + }, + { + "epoch": 0.5047923322683706, + "grad_norm": 0.3062373697757721, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 316 + }, + { + "epoch": 0.5063897763578274, + "grad_norm": 0.3083527386188507, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 317 + }, + { + "epoch": 0.5079872204472844, + "grad_norm": 0.4790252447128296, + "learning_rate": 5e-07, + "loss": 0.0045, + "step": 318 + }, + { + "epoch": 0.5095846645367412, + "grad_norm": 0.4106828570365906, + "learning_rate": 5e-07, + "loss": 0.0045, + "step": 319 + }, + { + "epoch": 0.5111821086261981, + "grad_norm": 0.5916593670845032, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 320 + }, + { + "epoch": 0.512779552715655, + "grad_norm": 0.29794517159461975, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 321 + }, + { + "epoch": 0.5143769968051118, + "grad_norm": 0.3094945251941681, + "learning_rate": 5e-07, + "loss": 0.004, + "step": 322 + }, + { + "epoch": 0.5159744408945687, + "grad_norm": 0.8037744760513306, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 323 + }, + { + "epoch": 0.5175718849840255, + "grad_norm": 1.0918713808059692, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 324 + }, + { + "epoch": 0.5191693290734825, + "grad_norm": 0.1457972228527069, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 325 + }, + { + "epoch": 0.5207667731629393, + "grad_norm": 0.9666080474853516, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 326 + }, + { + "epoch": 0.5223642172523961, + "grad_norm": 0.5684059262275696, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 327 + }, + { + "epoch": 0.5239616613418531, + "grad_norm": 0.5098831057548523, + "learning_rate": 5e-07, + "loss": 0.0034, + "step": 328 + }, + { + "epoch": 0.5255591054313099, + "grad_norm": 0.23925842344760895, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 329 + }, + { + "epoch": 0.5271565495207667, + "grad_norm": 0.17939390242099762, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 330 + }, + { + "epoch": 0.5287539936102237, + "grad_norm": 0.16178075969219208, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 331 + }, + { + "epoch": 0.5303514376996805, + "grad_norm": 0.671220064163208, + "learning_rate": 5e-07, + "loss": 0.0046, + "step": 332 + }, + { + "epoch": 0.5319488817891374, + "grad_norm": 0.816387414932251, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 333 + }, + { + "epoch": 0.5335463258785943, + "grad_norm": 0.24421027302742004, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 334 + }, + { + "epoch": 0.5351437699680511, + "grad_norm": 0.557515025138855, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 335 + }, + { + "epoch": 0.536741214057508, + "grad_norm": 1.0557501316070557, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 336 + }, + { + "epoch": 0.5383386581469649, + "grad_norm": 0.959538459777832, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 337 + }, + { + "epoch": 0.5399361022364217, + "grad_norm": 0.3274182677268982, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 338 + }, + { + "epoch": 0.5415335463258786, + "grad_norm": 0.6471118927001953, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 339 + }, + { + "epoch": 0.5431309904153354, + "grad_norm": 0.38304996490478516, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 340 + }, + { + "epoch": 0.5447284345047924, + "grad_norm": 1.0141305923461914, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 341 + }, + { + "epoch": 0.5463258785942492, + "grad_norm": 0.9558109641075134, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 342 + }, + { + "epoch": 0.547923322683706, + "grad_norm": 0.20122268795967102, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 343 + }, + { + "epoch": 0.549520766773163, + "grad_norm": 0.4421844780445099, + "learning_rate": 5e-07, + "loss": 0.0037, + "step": 344 + }, + { + "epoch": 0.5511182108626198, + "grad_norm": 0.4365698993206024, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 345 + }, + { + "epoch": 0.5527156549520766, + "grad_norm": 1.578316330909729, + "learning_rate": 5e-07, + "loss": 0.0053, + "step": 346 + }, + { + "epoch": 0.5543130990415336, + "grad_norm": 0.21290868520736694, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 347 + }, + { + "epoch": 0.5559105431309904, + "grad_norm": 0.6181979179382324, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 348 + }, + { + "epoch": 0.5575079872204473, + "grad_norm": 0.20121750235557556, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 349 + }, + { + "epoch": 0.5591054313099042, + "grad_norm": 0.5325523018836975, + "learning_rate": 5e-07, + "loss": 0.0036, + "step": 350 + }, + { + "epoch": 0.560702875399361, + "grad_norm": 0.6079647541046143, + "learning_rate": 5e-07, + "loss": 0.0052, + "step": 351 + }, + { + "epoch": 0.5623003194888179, + "grad_norm": 0.2826935946941376, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 352 + }, + { + "epoch": 0.5638977635782748, + "grad_norm": 0.15892595052719116, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 353 + }, + { + "epoch": 0.5654952076677316, + "grad_norm": 0.3881969749927521, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 354 + }, + { + "epoch": 0.5670926517571885, + "grad_norm": 0.3390671908855438, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 355 + }, + { + "epoch": 0.5686900958466453, + "grad_norm": 0.2904258668422699, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 356 + }, + { + "epoch": 0.5702875399361023, + "grad_norm": 0.46645137667655945, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 357 + }, + { + "epoch": 0.5718849840255591, + "grad_norm": 0.6814522743225098, + "learning_rate": 5e-07, + "loss": 0.0034, + "step": 358 + }, + { + "epoch": 0.5734824281150159, + "grad_norm": 0.9146270751953125, + "learning_rate": 5e-07, + "loss": 0.0044, + "step": 359 + }, + { + "epoch": 0.5750798722044729, + "grad_norm": 0.21922031044960022, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 360 + }, + { + "epoch": 0.5766773162939297, + "grad_norm": 0.939836859703064, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 361 + }, + { + "epoch": 0.5782747603833865, + "grad_norm": 0.6049763560295105, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 362 + }, + { + "epoch": 0.5798722044728435, + "grad_norm": 0.25285854935646057, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 363 + }, + { + "epoch": 0.5814696485623003, + "grad_norm": 0.3022539019584656, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 364 + }, + { + "epoch": 0.5830670926517572, + "grad_norm": 0.25078529119491577, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 365 + }, + { + "epoch": 0.5846645367412141, + "grad_norm": 0.3507131040096283, + "learning_rate": 5e-07, + "loss": 0.0043, + "step": 366 + }, + { + "epoch": 0.5862619808306709, + "grad_norm": 0.6294477581977844, + "learning_rate": 5e-07, + "loss": 0.0058, + "step": 367 + }, + { + "epoch": 0.5878594249201278, + "grad_norm": 0.48274868726730347, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 368 + }, + { + "epoch": 0.5894568690095847, + "grad_norm": 0.39043542742729187, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 369 + }, + { + "epoch": 0.5910543130990416, + "grad_norm": 0.19649553298950195, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 370 + }, + { + "epoch": 0.5926517571884984, + "grad_norm": 0.12885907292366028, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 371 + }, + { + "epoch": 0.5942492012779552, + "grad_norm": 0.1488872468471527, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 372 + }, + { + "epoch": 0.5958466453674122, + "grad_norm": 0.3566787540912628, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 373 + }, + { + "epoch": 0.597444089456869, + "grad_norm": 0.986865758895874, + "learning_rate": 5e-07, + "loss": 0.0047, + "step": 374 + }, + { + "epoch": 0.5990415335463258, + "grad_norm": 0.1579206883907318, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 375 + }, + { + "epoch": 0.6006389776357828, + "grad_norm": 0.13746462762355804, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 376 + }, + { + "epoch": 0.6022364217252396, + "grad_norm": 0.25995907187461853, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 377 + }, + { + "epoch": 0.6038338658146964, + "grad_norm": 0.38791167736053467, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 378 + }, + { + "epoch": 0.6054313099041534, + "grad_norm": 1.6591804027557373, + "learning_rate": 5e-07, + "loss": 0.0034, + "step": 379 + }, + { + "epoch": 0.6070287539936102, + "grad_norm": 0.11244110763072968, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 380 + }, + { + "epoch": 0.6086261980830671, + "grad_norm": 0.28090763092041016, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 381 + }, + { + "epoch": 0.610223642172524, + "grad_norm": 0.8654282689094543, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 382 + }, + { + "epoch": 0.6118210862619808, + "grad_norm": 0.3197280764579773, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 383 + }, + { + "epoch": 0.6134185303514377, + "grad_norm": 0.13376931846141815, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 384 + }, + { + "epoch": 0.6150159744408946, + "grad_norm": 0.3771628439426422, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 385 + }, + { + "epoch": 0.6166134185303515, + "grad_norm": 3.0465352535247803, + "learning_rate": 5e-07, + "loss": 0.0057, + "step": 386 + }, + { + "epoch": 0.6182108626198083, + "grad_norm": 0.6134064793586731, + "learning_rate": 5e-07, + "loss": 0.0045, + "step": 387 + }, + { + "epoch": 0.6198083067092651, + "grad_norm": 1.984861969947815, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 388 + }, + { + "epoch": 0.6214057507987221, + "grad_norm": 0.7715848684310913, + "learning_rate": 5e-07, + "loss": 0.0034, + "step": 389 + }, + { + "epoch": 0.6230031948881789, + "grad_norm": 1.4465832710266113, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 390 + }, + { + "epoch": 0.6246006389776357, + "grad_norm": 0.4437814950942993, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 391 + }, + { + "epoch": 0.6261980830670927, + "grad_norm": 0.7803134322166443, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 392 + }, + { + "epoch": 0.6277955271565495, + "grad_norm": 0.4495057761669159, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 393 + }, + { + "epoch": 0.6293929712460063, + "grad_norm": 0.9235338568687439, + "learning_rate": 5e-07, + "loss": 0.0062, + "step": 394 + }, + { + "epoch": 0.6309904153354633, + "grad_norm": 1.1879374980926514, + "learning_rate": 5e-07, + "loss": 0.0046, + "step": 395 + }, + { + "epoch": 0.6325878594249201, + "grad_norm": 0.22155530750751495, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 396 + }, + { + "epoch": 0.634185303514377, + "grad_norm": 0.6639124155044556, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 397 + }, + { + "epoch": 0.6357827476038339, + "grad_norm": 0.24812071025371552, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 398 + }, + { + "epoch": 0.6373801916932907, + "grad_norm": 0.24182091653347015, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 399 + }, + { + "epoch": 0.6389776357827476, + "grad_norm": 0.25194039940834045, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 400 + }, + { + "epoch": 0.6405750798722045, + "grad_norm": 1.6056197881698608, + "learning_rate": 5e-07, + "loss": 0.006, + "step": 401 + }, + { + "epoch": 0.6421725239616614, + "grad_norm": 3.634866952896118, + "learning_rate": 5e-07, + "loss": 0.0048, + "step": 402 + }, + { + "epoch": 0.6437699680511182, + "grad_norm": 1.0231901407241821, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 403 + }, + { + "epoch": 0.645367412140575, + "grad_norm": 0.27629023790359497, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 404 + }, + { + "epoch": 0.646964856230032, + "grad_norm": 0.1949998289346695, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 405 + }, + { + "epoch": 0.6485623003194888, + "grad_norm": 0.34071701765060425, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 406 + }, + { + "epoch": 0.6501597444089456, + "grad_norm": 0.21671850979328156, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 407 + }, + { + "epoch": 0.6517571884984026, + "grad_norm": 0.27686911821365356, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 408 + }, + { + "epoch": 0.6533546325878594, + "grad_norm": 0.3393467366695404, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 409 + }, + { + "epoch": 0.6549520766773163, + "grad_norm": 0.5887823700904846, + "learning_rate": 5e-07, + "loss": 0.004, + "step": 410 + }, + { + "epoch": 0.6565495207667732, + "grad_norm": 0.22454427182674408, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 411 + }, + { + "epoch": 0.65814696485623, + "grad_norm": 0.13433396816253662, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 412 + }, + { + "epoch": 0.6597444089456869, + "grad_norm": 0.14173242449760437, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 413 + }, + { + "epoch": 0.6613418530351438, + "grad_norm": 0.30982744693756104, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 414 + }, + { + "epoch": 0.6629392971246006, + "grad_norm": 0.47411805391311646, + "learning_rate": 5e-07, + "loss": 0.004, + "step": 415 + }, + { + "epoch": 0.6645367412140575, + "grad_norm": 0.19577734172344208, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 416 + }, + { + "epoch": 0.6661341853035144, + "grad_norm": 0.6344949007034302, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 417 + }, + { + "epoch": 0.6677316293929713, + "grad_norm": 0.45841068029403687, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 418 + }, + { + "epoch": 0.6693290734824281, + "grad_norm": 0.4367104768753052, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 419 + }, + { + "epoch": 0.670926517571885, + "grad_norm": 0.363915354013443, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 420 + }, + { + "epoch": 0.6725239616613419, + "grad_norm": 0.18759390711784363, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 421 + }, + { + "epoch": 0.6741214057507987, + "grad_norm": 0.15988245606422424, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 422 + }, + { + "epoch": 0.6757188498402555, + "grad_norm": 0.1806800663471222, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 423 + }, + { + "epoch": 0.6773162939297125, + "grad_norm": 0.21876202523708344, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 424 + }, + { + "epoch": 0.6789137380191693, + "grad_norm": 0.13135991990566254, + "learning_rate": 5e-07, + "loss": 0.0005, + "step": 425 + }, + { + "epoch": 0.6805111821086262, + "grad_norm": 1.3671129941940308, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 426 + }, + { + "epoch": 0.6821086261980831, + "grad_norm": 0.1489122062921524, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 427 + }, + { + "epoch": 0.6837060702875399, + "grad_norm": 0.17866690456867218, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 428 + }, + { + "epoch": 0.6853035143769968, + "grad_norm": 0.655277669429779, + "learning_rate": 5e-07, + "loss": 0.0045, + "step": 429 + }, + { + "epoch": 0.6869009584664537, + "grad_norm": 0.17315036058425903, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 430 + }, + { + "epoch": 0.6884984025559105, + "grad_norm": 0.20632481575012207, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 431 + }, + { + "epoch": 0.6900958466453674, + "grad_norm": 0.2345188409090042, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 432 + }, + { + "epoch": 0.6916932907348243, + "grad_norm": 3.234980583190918, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 433 + }, + { + "epoch": 0.6932907348242812, + "grad_norm": 0.5852277874946594, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 434 + }, + { + "epoch": 0.694888178913738, + "grad_norm": 2.7778074741363525, + "learning_rate": 5e-07, + "loss": 0.0047, + "step": 435 + }, + { + "epoch": 0.6964856230031949, + "grad_norm": 0.13274647295475006, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 436 + }, + { + "epoch": 0.6980830670926518, + "grad_norm": 0.3051118850708008, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 437 + }, + { + "epoch": 0.6996805111821086, + "grad_norm": 0.1823217123746872, + "learning_rate": 5e-07, + "loss": 0.0034, + "step": 438 + }, + { + "epoch": 0.7012779552715654, + "grad_norm": 0.24809595942497253, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 439 + }, + { + "epoch": 0.7028753993610224, + "grad_norm": 0.191731795668602, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 440 + }, + { + "epoch": 0.7044728434504792, + "grad_norm": 0.32459452748298645, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 441 + }, + { + "epoch": 0.7060702875399361, + "grad_norm": 0.46944916248321533, + "learning_rate": 5e-07, + "loss": 0.0037, + "step": 442 + }, + { + "epoch": 0.707667731629393, + "grad_norm": 0.28335535526275635, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 443 + }, + { + "epoch": 0.7092651757188498, + "grad_norm": 0.4613920748233795, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 444 + }, + { + "epoch": 0.7108626198083067, + "grad_norm": 0.25062379240989685, + "learning_rate": 5e-07, + "loss": 0.0037, + "step": 445 + }, + { + "epoch": 0.7124600638977636, + "grad_norm": 0.18467797338962555, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 446 + }, + { + "epoch": 0.7140575079872205, + "grad_norm": 0.11972685158252716, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 447 + }, + { + "epoch": 0.7156549520766773, + "grad_norm": 0.42692673206329346, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 448 + }, + { + "epoch": 0.7172523961661342, + "grad_norm": 0.24261529743671417, + "learning_rate": 5e-07, + "loss": 0.0036, + "step": 449 + }, + { + "epoch": 0.7188498402555911, + "grad_norm": 0.5654745101928711, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 450 + }, + { + "epoch": 0.7204472843450479, + "grad_norm": 1.1742912530899048, + "learning_rate": 5e-07, + "loss": 0.005, + "step": 451 + }, + { + "epoch": 0.7220447284345048, + "grad_norm": 0.4380282163619995, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 452 + }, + { + "epoch": 0.7236421725239617, + "grad_norm": 0.10869846493005753, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 453 + }, + { + "epoch": 0.7252396166134185, + "grad_norm": 0.24093355238437653, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 454 + }, + { + "epoch": 0.7268370607028753, + "grad_norm": 0.10973268747329712, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 455 + }, + { + "epoch": 0.7284345047923323, + "grad_norm": 0.3785342276096344, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 456 + }, + { + "epoch": 0.7300319488817891, + "grad_norm": 0.9914066791534424, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 457 + }, + { + "epoch": 0.731629392971246, + "grad_norm": 0.4453500211238861, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 458 + }, + { + "epoch": 0.7332268370607029, + "grad_norm": 0.2674676477909088, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 459 + }, + { + "epoch": 0.7348242811501597, + "grad_norm": 0.9725813269615173, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 460 + }, + { + "epoch": 0.7364217252396166, + "grad_norm": 0.14569059014320374, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 461 + }, + { + "epoch": 0.7380191693290735, + "grad_norm": 0.25622859597206116, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 462 + }, + { + "epoch": 0.7396166134185304, + "grad_norm": 0.33373990654945374, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 463 + }, + { + "epoch": 0.7412140575079872, + "grad_norm": 0.28780269622802734, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 464 + }, + { + "epoch": 0.7428115015974441, + "grad_norm": 0.3068907856941223, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 465 + }, + { + "epoch": 0.744408945686901, + "grad_norm": 0.1721941977739334, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 466 + }, + { + "epoch": 0.7460063897763578, + "grad_norm": 0.08544081449508667, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 467 + }, + { + "epoch": 0.7476038338658147, + "grad_norm": 0.5271486043930054, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 468 + }, + { + "epoch": 0.7492012779552716, + "grad_norm": 1.1334803104400635, + "learning_rate": 5e-07, + "loss": 0.0045, + "step": 469 + }, + { + "epoch": 0.7507987220447284, + "grad_norm": 0.7498302459716797, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 470 + }, + { + "epoch": 0.7523961661341853, + "grad_norm": 0.2573578357696533, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 471 + }, + { + "epoch": 0.7539936102236422, + "grad_norm": 0.658050537109375, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 472 + }, + { + "epoch": 0.755591054313099, + "grad_norm": 0.9222865104675293, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 473 + }, + { + "epoch": 0.7571884984025559, + "grad_norm": 0.15267683565616608, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 474 + }, + { + "epoch": 0.7587859424920128, + "grad_norm": 1.6186271905899048, + "learning_rate": 5e-07, + "loss": 0.0042, + "step": 475 + }, + { + "epoch": 0.7603833865814696, + "grad_norm": 0.27657464146614075, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 476 + }, + { + "epoch": 0.7619808306709265, + "grad_norm": 1.7262461185455322, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 477 + }, + { + "epoch": 0.7635782747603834, + "grad_norm": 0.2438318431377411, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 478 + }, + { + "epoch": 0.7651757188498403, + "grad_norm": 1.9847776889801025, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 479 + }, + { + "epoch": 0.7667731629392971, + "grad_norm": 1.081231713294983, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 480 + }, + { + "epoch": 0.768370607028754, + "grad_norm": 0.23922041058540344, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 481 + }, + { + "epoch": 0.7699680511182109, + "grad_norm": 0.9737673401832581, + "learning_rate": 5e-07, + "loss": 0.0065, + "step": 482 + }, + { + "epoch": 0.7715654952076677, + "grad_norm": 0.14736409485340118, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 483 + }, + { + "epoch": 0.7731629392971247, + "grad_norm": 0.2849075198173523, + "learning_rate": 5e-07, + "loss": 0.0034, + "step": 484 + }, + { + "epoch": 0.7747603833865815, + "grad_norm": 0.7391708493232727, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 485 + }, + { + "epoch": 0.7763578274760383, + "grad_norm": 0.24760426580905914, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 486 + }, + { + "epoch": 0.7779552715654952, + "grad_norm": 0.1888248175382614, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 487 + }, + { + "epoch": 0.7795527156549521, + "grad_norm": 0.27008160948753357, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 488 + }, + { + "epoch": 0.7811501597444089, + "grad_norm": 0.17157332599163055, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 489 + }, + { + "epoch": 0.7827476038338658, + "grad_norm": 0.2907833755016327, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 490 + }, + { + "epoch": 0.7843450479233227, + "grad_norm": 0.4900013506412506, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 491 + }, + { + "epoch": 0.7859424920127795, + "grad_norm": 0.1433720886707306, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 492 + }, + { + "epoch": 0.7875399361022364, + "grad_norm": 0.09725961834192276, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 493 + }, + { + "epoch": 0.7891373801916933, + "grad_norm": 0.8984055519104004, + "learning_rate": 5e-07, + "loss": 0.0048, + "step": 494 + }, + { + "epoch": 0.7907348242811502, + "grad_norm": 0.22446835041046143, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 495 + }, + { + "epoch": 0.792332268370607, + "grad_norm": 0.2916003167629242, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 496 + }, + { + "epoch": 0.7939297124600639, + "grad_norm": 0.42406606674194336, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 497 + }, + { + "epoch": 0.7955271565495208, + "grad_norm": 0.8635863661766052, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 498 + }, + { + "epoch": 0.7971246006389776, + "grad_norm": 0.4940405488014221, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 499 + }, + { + "epoch": 0.7987220447284346, + "grad_norm": 0.6001198291778564, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 500 + }, + { + "epoch": 0.8003194888178914, + "grad_norm": 0.4194888174533844, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 501 + }, + { + "epoch": 0.8019169329073482, + "grad_norm": 0.1558961719274521, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 502 + }, + { + "epoch": 0.8035143769968051, + "grad_norm": 0.1700737476348877, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 503 + }, + { + "epoch": 0.805111821086262, + "grad_norm": 0.15276680886745453, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 504 + }, + { + "epoch": 0.8067092651757188, + "grad_norm": 0.21490108966827393, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 505 + }, + { + "epoch": 0.8083067092651757, + "grad_norm": 0.48414719104766846, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 506 + }, + { + "epoch": 0.8099041533546326, + "grad_norm": 0.14232124388217926, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 507 + }, + { + "epoch": 0.8115015974440895, + "grad_norm": 0.13643090426921844, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 508 + }, + { + "epoch": 0.8130990415335463, + "grad_norm": 0.4378473162651062, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 509 + }, + { + "epoch": 0.8146964856230032, + "grad_norm": 0.5300367474555969, + "learning_rate": 5e-07, + "loss": 0.0036, + "step": 510 + }, + { + "epoch": 0.8162939297124601, + "grad_norm": 0.3394503593444824, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 511 + }, + { + "epoch": 0.8178913738019169, + "grad_norm": 0.15926502645015717, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 512 + }, + { + "epoch": 0.8194888178913738, + "grad_norm": 0.14627420902252197, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 513 + }, + { + "epoch": 0.8210862619808307, + "grad_norm": 0.2294456660747528, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 514 + }, + { + "epoch": 0.8226837060702875, + "grad_norm": 0.2830236554145813, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 515 + }, + { + "epoch": 0.8242811501597445, + "grad_norm": 0.49337977170944214, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 516 + }, + { + "epoch": 0.8258785942492013, + "grad_norm": 0.12382300198078156, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 517 + }, + { + "epoch": 0.8274760383386581, + "grad_norm": 0.5072153210639954, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 518 + }, + { + "epoch": 0.829073482428115, + "grad_norm": 0.6549058556556702, + "learning_rate": 5e-07, + "loss": 0.0044, + "step": 519 + }, + { + "epoch": 0.8306709265175719, + "grad_norm": 0.21236664056777954, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 520 + }, + { + "epoch": 0.8322683706070287, + "grad_norm": 0.6704789996147156, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 521 + }, + { + "epoch": 0.8338658146964856, + "grad_norm": 0.1254369020462036, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 522 + }, + { + "epoch": 0.8354632587859425, + "grad_norm": 2.3178699016571045, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 523 + }, + { + "epoch": 0.8370607028753994, + "grad_norm": 3.5792298316955566, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 524 + }, + { + "epoch": 0.8386581469648562, + "grad_norm": 0.7073982357978821, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 525 + }, + { + "epoch": 0.8402555910543131, + "grad_norm": 0.5759369134902954, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 526 + }, + { + "epoch": 0.84185303514377, + "grad_norm": 1.6669633388519287, + "learning_rate": 5e-07, + "loss": 0.0043, + "step": 527 + }, + { + "epoch": 0.8434504792332268, + "grad_norm": 0.6380763649940491, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 528 + }, + { + "epoch": 0.8450479233226837, + "grad_norm": 0.19805264472961426, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 529 + }, + { + "epoch": 0.8466453674121406, + "grad_norm": 0.6069849133491516, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 530 + }, + { + "epoch": 0.8482428115015974, + "grad_norm": 0.1687164306640625, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 531 + }, + { + "epoch": 0.8498402555910544, + "grad_norm": 8.362351417541504, + "learning_rate": 5e-07, + "loss": 0.0054, + "step": 532 + }, + { + "epoch": 0.8514376996805112, + "grad_norm": 0.2093094438314438, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 533 + }, + { + "epoch": 0.853035143769968, + "grad_norm": 0.34365883469581604, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 534 + }, + { + "epoch": 0.854632587859425, + "grad_norm": 0.12474914640188217, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 535 + }, + { + "epoch": 0.8562300319488818, + "grad_norm": 0.510906994342804, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 536 + }, + { + "epoch": 0.8578274760383386, + "grad_norm": 0.587397038936615, + "learning_rate": 5e-07, + "loss": 0.0036, + "step": 537 + }, + { + "epoch": 0.8594249201277955, + "grad_norm": 0.844524621963501, + "learning_rate": 5e-07, + "loss": 0.0052, + "step": 538 + }, + { + "epoch": 0.8610223642172524, + "grad_norm": 0.13176505267620087, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 539 + }, + { + "epoch": 0.8626198083067093, + "grad_norm": 0.16751867532730103, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 540 + }, + { + "epoch": 0.8642172523961661, + "grad_norm": 0.41158822178840637, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 541 + }, + { + "epoch": 0.865814696485623, + "grad_norm": 0.12536601722240448, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 542 + }, + { + "epoch": 0.8674121405750799, + "grad_norm": 0.08342156559228897, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 543 + }, + { + "epoch": 0.8690095846645367, + "grad_norm": 1.5413963794708252, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 544 + }, + { + "epoch": 0.8706070287539937, + "grad_norm": 0.24465155601501465, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 545 + }, + { + "epoch": 0.8722044728434505, + "grad_norm": 0.1531795710325241, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 546 + }, + { + "epoch": 0.8738019169329073, + "grad_norm": 0.1711510270833969, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 547 + }, + { + "epoch": 0.8753993610223643, + "grad_norm": 0.10270589590072632, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 548 + }, + { + "epoch": 0.8769968051118211, + "grad_norm": 0.6564199328422546, + "learning_rate": 5e-07, + "loss": 0.004, + "step": 549 + }, + { + "epoch": 0.8785942492012779, + "grad_norm": 0.15586939454078674, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 550 + }, + { + "epoch": 0.8801916932907349, + "grad_norm": 0.11600361764431, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 551 + }, + { + "epoch": 0.8817891373801917, + "grad_norm": 0.10974587500095367, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 552 + }, + { + "epoch": 0.8833865814696485, + "grad_norm": 0.2567409574985504, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 553 + }, + { + "epoch": 0.8849840255591054, + "grad_norm": 0.27306219935417175, + "learning_rate": 5e-07, + "loss": 0.0043, + "step": 554 + }, + { + "epoch": 0.8865814696485623, + "grad_norm": 0.3767613470554352, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 555 + }, + { + "epoch": 0.8881789137380192, + "grad_norm": 1.1817584037780762, + "learning_rate": 5e-07, + "loss": 0.0057, + "step": 556 + }, + { + "epoch": 0.889776357827476, + "grad_norm": 0.3905037045478821, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 557 + }, + { + "epoch": 0.8913738019169329, + "grad_norm": 0.23763906955718994, + "learning_rate": 5e-07, + "loss": 0.0006, + "step": 558 + }, + { + "epoch": 0.8929712460063898, + "grad_norm": 0.3837161362171173, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 559 + }, + { + "epoch": 0.8945686900958466, + "grad_norm": 0.6008383631706238, + "learning_rate": 5e-07, + "loss": 0.0041, + "step": 560 + }, + { + "epoch": 0.8961661341853036, + "grad_norm": 0.1975802332162857, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 561 + }, + { + "epoch": 0.8977635782747604, + "grad_norm": 0.15433986485004425, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 562 + }, + { + "epoch": 0.8993610223642172, + "grad_norm": 0.4065248966217041, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 563 + }, + { + "epoch": 0.9009584664536742, + "grad_norm": 0.25533992052078247, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 564 + }, + { + "epoch": 0.902555910543131, + "grad_norm": 0.46665945649147034, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 565 + }, + { + "epoch": 0.9041533546325878, + "grad_norm": 0.16390737891197205, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 566 + }, + { + "epoch": 0.9057507987220448, + "grad_norm": 0.06191791966557503, + "learning_rate": 5e-07, + "loss": 0.0005, + "step": 567 + }, + { + "epoch": 0.9073482428115016, + "grad_norm": 1.692974328994751, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 568 + }, + { + "epoch": 0.9089456869009584, + "grad_norm": 0.21359291672706604, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 569 + }, + { + "epoch": 0.9105431309904153, + "grad_norm": 0.09417448937892914, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 570 + }, + { + "epoch": 0.9121405750798722, + "grad_norm": 0.22416774928569794, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 571 + }, + { + "epoch": 0.9137380191693291, + "grad_norm": 0.15862461924552917, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 572 + }, + { + "epoch": 0.9153354632587859, + "grad_norm": 0.325225293636322, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 573 + }, + { + "epoch": 0.9169329073482428, + "grad_norm": 0.4235207140445709, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 574 + }, + { + "epoch": 0.9185303514376997, + "grad_norm": 0.17048902809619904, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 575 + }, + { + "epoch": 0.9201277955271565, + "grad_norm": 0.28779876232147217, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 576 + }, + { + "epoch": 0.9217252396166135, + "grad_norm": 0.16058605909347534, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 577 + }, + { + "epoch": 0.9233226837060703, + "grad_norm": 0.08717779070138931, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 578 + }, + { + "epoch": 0.9249201277955271, + "grad_norm": 0.852260172367096, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 579 + }, + { + "epoch": 0.9265175718849841, + "grad_norm": 0.34097403287887573, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 580 + }, + { + "epoch": 0.9281150159744409, + "grad_norm": 0.3408906161785126, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 581 + }, + { + "epoch": 0.9297124600638977, + "grad_norm": 0.4731082618236542, + "learning_rate": 5e-07, + "loss": 0.0037, + "step": 582 + }, + { + "epoch": 0.9313099041533547, + "grad_norm": 0.49347537755966187, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 583 + }, + { + "epoch": 0.9329073482428115, + "grad_norm": 0.17723320424556732, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 584 + }, + { + "epoch": 0.9345047923322684, + "grad_norm": 0.16061758995056152, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 585 + }, + { + "epoch": 0.9361022364217252, + "grad_norm": 0.11341333389282227, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 586 + }, + { + "epoch": 0.9376996805111821, + "grad_norm": 0.12614861130714417, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 587 + }, + { + "epoch": 0.939297124600639, + "grad_norm": 0.1919727921485901, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 588 + }, + { + "epoch": 0.9408945686900958, + "grad_norm": 0.5518249273300171, + "learning_rate": 5e-07, + "loss": 0.0034, + "step": 589 + }, + { + "epoch": 0.9424920127795527, + "grad_norm": 0.11890781670808792, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 590 + }, + { + "epoch": 0.9440894568690096, + "grad_norm": 0.5765843391418457, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 591 + }, + { + "epoch": 0.9456869009584664, + "grad_norm": 0.17196467518806458, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 592 + }, + { + "epoch": 0.9472843450479234, + "grad_norm": 0.2513171136379242, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 593 + }, + { + "epoch": 0.9488817891373802, + "grad_norm": 1.7397063970565796, + "learning_rate": 5e-07, + "loss": 0.0108, + "step": 594 + }, + { + "epoch": 0.950479233226837, + "grad_norm": 0.2406977415084839, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 595 + }, + { + "epoch": 0.952076677316294, + "grad_norm": 0.23791703581809998, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 596 + }, + { + "epoch": 0.9536741214057508, + "grad_norm": 0.22479400038719177, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 597 + }, + { + "epoch": 0.9552715654952076, + "grad_norm": 0.23164299130439758, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 598 + }, + { + "epoch": 0.9568690095846646, + "grad_norm": 0.1600014865398407, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 599 + }, + { + "epoch": 0.9584664536741214, + "grad_norm": 0.2001635879278183, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 600 + }, + { + "epoch": 0.9600638977635783, + "grad_norm": 0.2946987450122833, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 601 + }, + { + "epoch": 0.9616613418530351, + "grad_norm": 0.2080981582403183, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 602 + }, + { + "epoch": 0.963258785942492, + "grad_norm": 0.6526065468788147, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 603 + }, + { + "epoch": 0.9648562300319489, + "grad_norm": 0.3620392382144928, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 604 + }, + { + "epoch": 0.9664536741214057, + "grad_norm": 0.21057778596878052, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 605 + }, + { + "epoch": 0.9680511182108626, + "grad_norm": 0.11293834447860718, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 606 + }, + { + "epoch": 0.9696485623003195, + "grad_norm": 1.2960716485977173, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 607 + }, + { + "epoch": 0.9712460063897763, + "grad_norm": 0.1437765210866928, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 608 + }, + { + "epoch": 0.9728434504792333, + "grad_norm": 0.6358578205108643, + "learning_rate": 5e-07, + "loss": 0.0049, + "step": 609 + }, + { + "epoch": 0.9744408945686901, + "grad_norm": 0.14403465390205383, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 610 + }, + { + "epoch": 0.9760383386581469, + "grad_norm": 0.4725677967071533, + "learning_rate": 5e-07, + "loss": 0.0045, + "step": 611 + }, + { + "epoch": 0.9776357827476039, + "grad_norm": 0.8543768525123596, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 612 + }, + { + "epoch": 0.9792332268370607, + "grad_norm": 0.27371206879615784, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 613 + }, + { + "epoch": 0.9808306709265175, + "grad_norm": 0.23279932141304016, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 614 + }, + { + "epoch": 0.9824281150159745, + "grad_norm": 0.8952144384384155, + "learning_rate": 5e-07, + "loss": 0.0037, + "step": 615 + }, + { + "epoch": 0.9840255591054313, + "grad_norm": 0.11720918118953705, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 616 + }, + { + "epoch": 0.9856230031948882, + "grad_norm": 0.9963851571083069, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 617 + }, + { + "epoch": 0.987220447284345, + "grad_norm": 0.08971522003412247, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 618 + }, + { + "epoch": 0.9888178913738019, + "grad_norm": 0.1214229092001915, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 619 + }, + { + "epoch": 0.9904153354632588, + "grad_norm": 0.2924940288066864, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 620 + }, + { + "epoch": 0.9920127795527156, + "grad_norm": 1.189788579940796, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 621 + }, + { + "epoch": 0.9936102236421726, + "grad_norm": 0.5802843570709229, + "learning_rate": 5e-07, + "loss": 0.0044, + "step": 622 + }, + { + "epoch": 0.9952076677316294, + "grad_norm": 0.11823207885026932, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 623 + }, + { + "epoch": 0.9968051118210862, + "grad_norm": 0.16288286447525024, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 624 + }, + { + "epoch": 0.9984025559105432, + "grad_norm": 0.13317349553108215, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 625 + }, + { + "epoch": 1.0, + "grad_norm": 0.1848553717136383, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 626 + }, + { + "epoch": 1.0015974440894568, + "grad_norm": 0.5851624608039856, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 627 + }, + { + "epoch": 1.0031948881789137, + "grad_norm": 0.09685046225786209, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 628 + }, + { + "epoch": 1.0047923322683705, + "grad_norm": 0.11342930048704147, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 629 + }, + { + "epoch": 1.0063897763578276, + "grad_norm": 0.46440932154655457, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 630 + }, + { + "epoch": 1.0079872204472844, + "grad_norm": 0.19794267416000366, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 631 + }, + { + "epoch": 1.0095846645367412, + "grad_norm": 0.3643769919872284, + "learning_rate": 5e-07, + "loss": 0.0041, + "step": 632 + }, + { + "epoch": 1.011182108626198, + "grad_norm": 0.08710943907499313, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 633 + }, + { + "epoch": 1.012779552715655, + "grad_norm": 0.4915333688259125, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 634 + }, + { + "epoch": 1.0143769968051117, + "grad_norm": 0.23040416836738586, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 635 + }, + { + "epoch": 1.0159744408945688, + "grad_norm": 0.24800904095172882, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 636 + }, + { + "epoch": 1.0175718849840256, + "grad_norm": 0.29919004440307617, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 637 + }, + { + "epoch": 1.0191693290734825, + "grad_norm": 0.11212212592363358, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 638 + }, + { + "epoch": 1.0207667731629393, + "grad_norm": 0.16430723667144775, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 639 + }, + { + "epoch": 1.0223642172523961, + "grad_norm": 0.32202014327049255, + "learning_rate": 5e-07, + "loss": 0.0046, + "step": 640 + }, + { + "epoch": 1.023961661341853, + "grad_norm": 0.3237399458885193, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 641 + }, + { + "epoch": 1.0255591054313098, + "grad_norm": 0.1953740268945694, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 642 + }, + { + "epoch": 1.0271565495207668, + "grad_norm": 0.12679028511047363, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 643 + }, + { + "epoch": 1.0287539936102237, + "grad_norm": 0.10133963078260422, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 644 + }, + { + "epoch": 1.0303514376996805, + "grad_norm": 0.0973275899887085, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 645 + }, + { + "epoch": 1.0319488817891374, + "grad_norm": 0.5217294096946716, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 646 + }, + { + "epoch": 1.0335463258785942, + "grad_norm": 0.15706409513950348, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 647 + }, + { + "epoch": 1.035143769968051, + "grad_norm": 0.10915224999189377, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 648 + }, + { + "epoch": 1.036741214057508, + "grad_norm": 0.4759727716445923, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 649 + }, + { + "epoch": 1.038338658146965, + "grad_norm": 0.11390835046768188, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 650 + }, + { + "epoch": 1.0399361022364217, + "grad_norm": 1.093390703201294, + "learning_rate": 5e-07, + "loss": 0.0046, + "step": 651 + }, + { + "epoch": 1.0415335463258786, + "grad_norm": 0.07437872141599655, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 652 + }, + { + "epoch": 1.0431309904153354, + "grad_norm": 0.21833601593971252, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 653 + }, + { + "epoch": 1.0447284345047922, + "grad_norm": 0.21499225497245789, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 654 + }, + { + "epoch": 1.0463258785942493, + "grad_norm": 0.12159962207078934, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 655 + }, + { + "epoch": 1.0479233226837061, + "grad_norm": 0.3396832346916199, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 656 + }, + { + "epoch": 1.049520766773163, + "grad_norm": 0.21546514332294464, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 657 + }, + { + "epoch": 1.0511182108626198, + "grad_norm": 0.19969788193702698, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 658 + }, + { + "epoch": 1.0527156549520766, + "grad_norm": 0.08988461643457413, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 659 + }, + { + "epoch": 1.0543130990415335, + "grad_norm": 0.26418858766555786, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 660 + }, + { + "epoch": 1.0559105431309903, + "grad_norm": 0.2093944102525711, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 661 + }, + { + "epoch": 1.0575079872204474, + "grad_norm": 0.36498188972473145, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 662 + }, + { + "epoch": 1.0591054313099042, + "grad_norm": 0.07945135235786438, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 663 + }, + { + "epoch": 1.060702875399361, + "grad_norm": 0.12991008162498474, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 664 + }, + { + "epoch": 1.0623003194888179, + "grad_norm": 0.2589464783668518, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 665 + }, + { + "epoch": 1.0638977635782747, + "grad_norm": 0.10802029818296432, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 666 + }, + { + "epoch": 1.0654952076677315, + "grad_norm": 0.4823663532733917, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 667 + }, + { + "epoch": 1.0670926517571886, + "grad_norm": 0.44493567943573, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 668 + }, + { + "epoch": 1.0686900958466454, + "grad_norm": 0.23464570939540863, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 669 + }, + { + "epoch": 1.0702875399361023, + "grad_norm": 0.28489264845848083, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 670 + }, + { + "epoch": 1.071884984025559, + "grad_norm": 0.2952616214752197, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 671 + }, + { + "epoch": 1.073482428115016, + "grad_norm": 0.20882096886634827, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 672 + }, + { + "epoch": 1.0750798722044728, + "grad_norm": 0.12549445033073425, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 673 + }, + { + "epoch": 1.0766773162939298, + "grad_norm": 0.22115099430084229, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 674 + }, + { + "epoch": 1.0782747603833867, + "grad_norm": 0.15710076689720154, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 675 + }, + { + "epoch": 1.0798722044728435, + "grad_norm": 0.15250109136104584, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 676 + }, + { + "epoch": 1.0814696485623003, + "grad_norm": 0.1936073750257492, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 677 + }, + { + "epoch": 1.0830670926517572, + "grad_norm": 0.08779247850179672, + "learning_rate": 5e-07, + "loss": 0.0006, + "step": 678 + }, + { + "epoch": 1.084664536741214, + "grad_norm": 0.1437022089958191, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 679 + }, + { + "epoch": 1.0862619808306708, + "grad_norm": 0.2850930392742157, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 680 + }, + { + "epoch": 1.0878594249201279, + "grad_norm": 0.3779478967189789, + "learning_rate": 5e-07, + "loss": 0.0051, + "step": 681 + }, + { + "epoch": 1.0894568690095847, + "grad_norm": 0.20855233073234558, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 682 + }, + { + "epoch": 1.0910543130990416, + "grad_norm": 0.11478544771671295, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 683 + }, + { + "epoch": 1.0926517571884984, + "grad_norm": 0.3283560276031494, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 684 + }, + { + "epoch": 1.0942492012779552, + "grad_norm": 0.35279855132102966, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 685 + }, + { + "epoch": 1.095846645367412, + "grad_norm": 0.166551411151886, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 686 + }, + { + "epoch": 1.097444089456869, + "grad_norm": 0.10785111039876938, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 687 + }, + { + "epoch": 1.099041533546326, + "grad_norm": 0.1150643453001976, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 688 + }, + { + "epoch": 1.1006389776357828, + "grad_norm": 0.10543517768383026, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 689 + }, + { + "epoch": 1.1022364217252396, + "grad_norm": 0.12300563603639603, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 690 + }, + { + "epoch": 1.1038338658146964, + "grad_norm": 0.15775950253009796, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 691 + }, + { + "epoch": 1.1054313099041533, + "grad_norm": 0.052068062126636505, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 692 + }, + { + "epoch": 1.1070287539936103, + "grad_norm": 0.1371004432439804, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 693 + }, + { + "epoch": 1.1086261980830672, + "grad_norm": 0.11871246248483658, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 694 + }, + { + "epoch": 1.110223642172524, + "grad_norm": 0.106499083340168, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 695 + }, + { + "epoch": 1.1118210862619808, + "grad_norm": 0.05005495250225067, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 696 + }, + { + "epoch": 1.1134185303514377, + "grad_norm": 0.6656812429428101, + "learning_rate": 5e-07, + "loss": 0.0036, + "step": 697 + }, + { + "epoch": 1.1150159744408945, + "grad_norm": 0.14445379376411438, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 698 + }, + { + "epoch": 1.1166134185303513, + "grad_norm": 0.268867164850235, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 699 + }, + { + "epoch": 1.1182108626198084, + "grad_norm": 0.16358579695224762, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 700 + }, + { + "epoch": 1.1198083067092652, + "grad_norm": 0.11881609261035919, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 701 + }, + { + "epoch": 1.121405750798722, + "grad_norm": 0.10697460919618607, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 702 + }, + { + "epoch": 1.123003194888179, + "grad_norm": 0.1314931958913803, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 703 + }, + { + "epoch": 1.1246006389776357, + "grad_norm": 0.13953609764575958, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 704 + }, + { + "epoch": 1.1261980830670926, + "grad_norm": 0.30124762654304504, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 705 + }, + { + "epoch": 1.1277955271565494, + "grad_norm": 0.2047244906425476, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 706 + }, + { + "epoch": 1.1293929712460065, + "grad_norm": 0.28825926780700684, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 707 + }, + { + "epoch": 1.1309904153354633, + "grad_norm": 0.1030370369553566, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 708 + }, + { + "epoch": 1.1325878594249201, + "grad_norm": 0.08416575193405151, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 709 + }, + { + "epoch": 1.134185303514377, + "grad_norm": 0.6813434958457947, + "learning_rate": 5e-07, + "loss": 0.0046, + "step": 710 + }, + { + "epoch": 1.1357827476038338, + "grad_norm": 0.7300125360488892, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 711 + }, + { + "epoch": 1.1373801916932909, + "grad_norm": 0.21185138821601868, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 712 + }, + { + "epoch": 1.1389776357827477, + "grad_norm": 0.3095090389251709, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 713 + }, + { + "epoch": 1.1405750798722045, + "grad_norm": 0.13913819193840027, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 714 + }, + { + "epoch": 1.1421725239616614, + "grad_norm": 0.2016637921333313, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 715 + }, + { + "epoch": 1.1437699680511182, + "grad_norm": 0.12468226253986359, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 716 + }, + { + "epoch": 1.145367412140575, + "grad_norm": 0.1510785073041916, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 717 + }, + { + "epoch": 1.1469648562300319, + "grad_norm": 0.11004221439361572, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 718 + }, + { + "epoch": 1.148562300319489, + "grad_norm": 0.12669166922569275, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 719 + }, + { + "epoch": 1.1501597444089458, + "grad_norm": 0.11249173432588577, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 720 + }, + { + "epoch": 1.1517571884984026, + "grad_norm": 0.21114954352378845, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 721 + }, + { + "epoch": 1.1533546325878594, + "grad_norm": 0.3067387342453003, + "learning_rate": 5e-07, + "loss": 0.0037, + "step": 722 + }, + { + "epoch": 1.1549520766773163, + "grad_norm": 0.11942733824253082, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 723 + }, + { + "epoch": 1.156549520766773, + "grad_norm": 0.2498818337917328, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 724 + }, + { + "epoch": 1.15814696485623, + "grad_norm": 0.589302122592926, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 725 + }, + { + "epoch": 1.159744408945687, + "grad_norm": 0.11040360480546951, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 726 + }, + { + "epoch": 1.1613418530351438, + "grad_norm": 0.1865539252758026, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 727 + }, + { + "epoch": 1.1629392971246006, + "grad_norm": 0.09886899590492249, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 728 + }, + { + "epoch": 1.1645367412140575, + "grad_norm": 0.10603392124176025, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 729 + }, + { + "epoch": 1.1661341853035143, + "grad_norm": 0.17114412784576416, + "learning_rate": 5e-07, + "loss": 0.0034, + "step": 730 + }, + { + "epoch": 1.1677316293929714, + "grad_norm": 0.11507996916770935, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 731 + }, + { + "epoch": 1.1693290734824282, + "grad_norm": 0.10117539763450623, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 732 + }, + { + "epoch": 1.170926517571885, + "grad_norm": 0.18606479465961456, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 733 + }, + { + "epoch": 1.1725239616613419, + "grad_norm": 0.1562090963125229, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 734 + }, + { + "epoch": 1.1741214057507987, + "grad_norm": 0.2511482834815979, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 735 + }, + { + "epoch": 1.1757188498402555, + "grad_norm": 0.131776362657547, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 736 + }, + { + "epoch": 1.1773162939297124, + "grad_norm": 0.2350272536277771, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 737 + }, + { + "epoch": 1.1789137380191694, + "grad_norm": 0.5341865420341492, + "learning_rate": 5e-07, + "loss": 0.0045, + "step": 738 + }, + { + "epoch": 1.1805111821086263, + "grad_norm": 0.18602542579174042, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 739 + }, + { + "epoch": 1.182108626198083, + "grad_norm": 0.08645334094762802, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 740 + }, + { + "epoch": 1.18370607028754, + "grad_norm": 0.1708468645811081, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 741 + }, + { + "epoch": 1.1853035143769968, + "grad_norm": 0.11233728379011154, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 742 + }, + { + "epoch": 1.1869009584664536, + "grad_norm": 0.20240649580955505, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 743 + }, + { + "epoch": 1.1884984025559104, + "grad_norm": 0.14982427656650543, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 744 + }, + { + "epoch": 1.1900958466453675, + "grad_norm": 0.13709725439548492, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 745 + }, + { + "epoch": 1.1916932907348243, + "grad_norm": 0.18215787410736084, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 746 + }, + { + "epoch": 1.1932907348242812, + "grad_norm": 0.12986968457698822, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 747 + }, + { + "epoch": 1.194888178913738, + "grad_norm": 0.10665430873632431, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 748 + }, + { + "epoch": 1.1964856230031948, + "grad_norm": 0.11768273264169693, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 749 + }, + { + "epoch": 1.1980830670926517, + "grad_norm": 0.6945746541023254, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 750 + }, + { + "epoch": 1.1996805111821087, + "grad_norm": 0.25220727920532227, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 751 + }, + { + "epoch": 1.2012779552715656, + "grad_norm": 0.07079087197780609, + "learning_rate": 5e-07, + "loss": 0.0005, + "step": 752 + }, + { + "epoch": 1.2028753993610224, + "grad_norm": 0.16933268308639526, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 753 + }, + { + "epoch": 1.2044728434504792, + "grad_norm": 0.15669192373752594, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 754 + }, + { + "epoch": 1.206070287539936, + "grad_norm": 6.026957035064697, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 755 + }, + { + "epoch": 1.207667731629393, + "grad_norm": 0.5998969674110413, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 756 + }, + { + "epoch": 1.20926517571885, + "grad_norm": 0.06143517419695854, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 757 + }, + { + "epoch": 1.2108626198083068, + "grad_norm": 0.2319423258304596, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 758 + }, + { + "epoch": 1.2124600638977636, + "grad_norm": 0.701130211353302, + "learning_rate": 5e-07, + "loss": 0.0056, + "step": 759 + }, + { + "epoch": 1.2140575079872205, + "grad_norm": 0.3482913076877594, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 760 + }, + { + "epoch": 1.2156549520766773, + "grad_norm": 0.21112686395645142, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 761 + }, + { + "epoch": 1.2172523961661341, + "grad_norm": 5.14704704284668, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 762 + }, + { + "epoch": 1.218849840255591, + "grad_norm": 0.41259634494781494, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 763 + }, + { + "epoch": 1.220447284345048, + "grad_norm": 0.2882460951805115, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 764 + }, + { + "epoch": 1.2220447284345048, + "grad_norm": 0.22736166417598724, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 765 + }, + { + "epoch": 1.2236421725239617, + "grad_norm": 0.2859097719192505, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 766 + }, + { + "epoch": 1.2252396166134185, + "grad_norm": 1.8469651937484741, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 767 + }, + { + "epoch": 1.2268370607028753, + "grad_norm": 0.15771770477294922, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 768 + }, + { + "epoch": 1.2284345047923322, + "grad_norm": 0.10112886875867844, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 769 + }, + { + "epoch": 1.230031948881789, + "grad_norm": 0.7125353813171387, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 770 + }, + { + "epoch": 1.231629392971246, + "grad_norm": 0.10188297927379608, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 771 + }, + { + "epoch": 1.233226837060703, + "grad_norm": 0.9515169262886047, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 772 + }, + { + "epoch": 1.2348242811501597, + "grad_norm": 0.06836796551942825, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 773 + }, + { + "epoch": 1.2364217252396166, + "grad_norm": 0.13632221519947052, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 774 + }, + { + "epoch": 1.2380191693290734, + "grad_norm": 0.572127103805542, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 775 + }, + { + "epoch": 1.2396166134185305, + "grad_norm": 0.11243536323308945, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 776 + }, + { + "epoch": 1.2412140575079873, + "grad_norm": 0.47509658336639404, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 777 + }, + { + "epoch": 1.2428115015974441, + "grad_norm": 3.768028974533081, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 778 + }, + { + "epoch": 1.244408945686901, + "grad_norm": 0.346608966588974, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 779 + }, + { + "epoch": 1.2460063897763578, + "grad_norm": 0.23147884011268616, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 780 + }, + { + "epoch": 1.2476038338658146, + "grad_norm": 0.14679104089736938, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 781 + }, + { + "epoch": 1.2492012779552715, + "grad_norm": 0.23181918263435364, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 782 + }, + { + "epoch": 1.2507987220447285, + "grad_norm": 0.16475149989128113, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 783 + }, + { + "epoch": 1.2523961661341854, + "grad_norm": 0.08807860314846039, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 784 + }, + { + "epoch": 1.2539936102236422, + "grad_norm": 0.16376349329948425, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 785 + }, + { + "epoch": 1.255591054313099, + "grad_norm": 0.19524359703063965, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 786 + }, + { + "epoch": 1.2571884984025559, + "grad_norm": 0.06623630225658417, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 787 + }, + { + "epoch": 1.2587859424920127, + "grad_norm": 0.03929225727915764, + "learning_rate": 5e-07, + "loss": 0.0003, + "step": 788 + }, + { + "epoch": 1.2603833865814695, + "grad_norm": 0.11669357866048813, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 789 + }, + { + "epoch": 1.2619808306709266, + "grad_norm": 0.00831072498112917, + "learning_rate": 5e-07, + "loss": 0.0, + "step": 790 + }, + { + "epoch": 1.2635782747603834, + "grad_norm": 0.13926006853580475, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 791 + }, + { + "epoch": 1.2651757188498403, + "grad_norm": 4.567328929901123, + "learning_rate": 5e-07, + "loss": 0.0034, + "step": 792 + }, + { + "epoch": 1.266773162939297, + "grad_norm": 1.018916368484497, + "learning_rate": 5e-07, + "loss": 0.0067, + "step": 793 + }, + { + "epoch": 1.268370607028754, + "grad_norm": 0.2978529632091522, + "learning_rate": 5e-07, + "loss": 0.0034, + "step": 794 + }, + { + "epoch": 1.269968051118211, + "grad_norm": 0.5247751474380493, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 795 + }, + { + "epoch": 1.2715654952076676, + "grad_norm": 0.06592018157243729, + "learning_rate": 5e-07, + "loss": 0.0006, + "step": 796 + }, + { + "epoch": 1.2731629392971247, + "grad_norm": 0.1426411271095276, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 797 + }, + { + "epoch": 1.2747603833865815, + "grad_norm": 0.14098455011844635, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 798 + }, + { + "epoch": 1.2763578274760383, + "grad_norm": 0.11731888353824615, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 799 + }, + { + "epoch": 1.2779552715654952, + "grad_norm": 0.8000310063362122, + "learning_rate": 5e-07, + "loss": 0.0044, + "step": 800 + }, + { + "epoch": 1.279552715654952, + "grad_norm": 0.8283488154411316, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 801 + }, + { + "epoch": 1.281150159744409, + "grad_norm": 0.6855292916297913, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 802 + }, + { + "epoch": 1.2827476038338659, + "grad_norm": 0.12922891974449158, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 803 + }, + { + "epoch": 1.2843450479233227, + "grad_norm": 0.8001663088798523, + "learning_rate": 5e-07, + "loss": 0.0055, + "step": 804 + }, + { + "epoch": 1.2859424920127795, + "grad_norm": 0.09436812996864319, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 805 + }, + { + "epoch": 1.2875399361022364, + "grad_norm": 0.3059661388397217, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 806 + }, + { + "epoch": 1.2891373801916932, + "grad_norm": 0.22149628400802612, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 807 + }, + { + "epoch": 1.29073482428115, + "grad_norm": 0.1476268321275711, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 808 + }, + { + "epoch": 1.292332268370607, + "grad_norm": 0.11144911497831345, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 809 + }, + { + "epoch": 1.293929712460064, + "grad_norm": 0.2204524427652359, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 810 + }, + { + "epoch": 1.2955271565495208, + "grad_norm": 0.07029737532138824, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 811 + }, + { + "epoch": 1.2971246006389776, + "grad_norm": 0.09939467906951904, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 812 + }, + { + "epoch": 1.2987220447284344, + "grad_norm": 0.1769058108329773, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 813 + }, + { + "epoch": 1.3003194888178915, + "grad_norm": 0.3592926561832428, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 814 + }, + { + "epoch": 1.3019169329073481, + "grad_norm": 0.2275543361902237, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 815 + }, + { + "epoch": 1.3035143769968052, + "grad_norm": 0.07945606857538223, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 816 + }, + { + "epoch": 1.305111821086262, + "grad_norm": 0.08711861073970795, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 817 + }, + { + "epoch": 1.3067092651757188, + "grad_norm": 0.28836479783058167, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 818 + }, + { + "epoch": 1.3083067092651757, + "grad_norm": 0.08159520477056503, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 819 + }, + { + "epoch": 1.3099041533546325, + "grad_norm": 0.07466080784797668, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 820 + }, + { + "epoch": 1.3115015974440896, + "grad_norm": 0.5451674461364746, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 821 + }, + { + "epoch": 1.3130990415335464, + "grad_norm": 0.14322105050086975, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 822 + }, + { + "epoch": 1.3146964856230032, + "grad_norm": 0.12718692421913147, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 823 + }, + { + "epoch": 1.31629392971246, + "grad_norm": 0.13329675793647766, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 824 + }, + { + "epoch": 1.317891373801917, + "grad_norm": 0.22774100303649902, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 825 + }, + { + "epoch": 1.3194888178913737, + "grad_norm": 0.8411527276039124, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 826 + }, + { + "epoch": 1.3210862619808306, + "grad_norm": 0.18316972255706787, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 827 + }, + { + "epoch": 1.3226837060702876, + "grad_norm": 0.07081547379493713, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 828 + }, + { + "epoch": 1.3242811501597445, + "grad_norm": 0.3037576973438263, + "learning_rate": 5e-07, + "loss": 0.0036, + "step": 829 + }, + { + "epoch": 1.3258785942492013, + "grad_norm": 0.10193005204200745, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 830 + }, + { + "epoch": 1.3274760383386581, + "grad_norm": 0.10602962970733643, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 831 + }, + { + "epoch": 1.329073482428115, + "grad_norm": 0.2908097505569458, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 832 + }, + { + "epoch": 1.330670926517572, + "grad_norm": 0.3724125027656555, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 833 + }, + { + "epoch": 1.3322683706070286, + "grad_norm": 0.11064999550580978, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 834 + }, + { + "epoch": 1.3338658146964857, + "grad_norm": 0.228192538022995, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 835 + }, + { + "epoch": 1.3354632587859425, + "grad_norm": 0.7090662717819214, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 836 + }, + { + "epoch": 1.3370607028753994, + "grad_norm": 0.14330582320690155, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 837 + }, + { + "epoch": 1.3386581469648562, + "grad_norm": 0.22641916573047638, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 838 + }, + { + "epoch": 1.340255591054313, + "grad_norm": 0.21619798243045807, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 839 + }, + { + "epoch": 1.34185303514377, + "grad_norm": 0.10568831115961075, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 840 + }, + { + "epoch": 1.343450479233227, + "grad_norm": 0.1081472709774971, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 841 + }, + { + "epoch": 1.3450479233226837, + "grad_norm": 0.10540477186441422, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 842 + }, + { + "epoch": 1.3466453674121406, + "grad_norm": 0.11737114936113358, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 843 + }, + { + "epoch": 1.3482428115015974, + "grad_norm": 0.17713451385498047, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 844 + }, + { + "epoch": 1.3498402555910542, + "grad_norm": 0.2664765417575836, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 845 + }, + { + "epoch": 1.351437699680511, + "grad_norm": 0.11203097552061081, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 846 + }, + { + "epoch": 1.3530351437699681, + "grad_norm": 0.24115419387817383, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 847 + }, + { + "epoch": 1.354632587859425, + "grad_norm": 0.13034223020076752, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 848 + }, + { + "epoch": 1.3562300319488818, + "grad_norm": 1.2168869972229004, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 849 + }, + { + "epoch": 1.3578274760383386, + "grad_norm": 0.04777536913752556, + "learning_rate": 5e-07, + "loss": 0.0005, + "step": 850 + }, + { + "epoch": 1.3594249201277955, + "grad_norm": 0.14492247998714447, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 851 + }, + { + "epoch": 1.3610223642172525, + "grad_norm": 0.11650210618972778, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 852 + }, + { + "epoch": 1.3626198083067091, + "grad_norm": 0.13418716192245483, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 853 + }, + { + "epoch": 1.3642172523961662, + "grad_norm": 1.2964794635772705, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 854 + }, + { + "epoch": 1.365814696485623, + "grad_norm": 0.17427192628383636, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 855 + }, + { + "epoch": 1.3674121405750799, + "grad_norm": 0.8252497911453247, + "learning_rate": 5e-07, + "loss": 0.0049, + "step": 856 + }, + { + "epoch": 1.3690095846645367, + "grad_norm": 0.3598852753639221, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 857 + }, + { + "epoch": 1.3706070287539935, + "grad_norm": 0.23161448538303375, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 858 + }, + { + "epoch": 1.3722044728434506, + "grad_norm": 0.17750093340873718, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 859 + }, + { + "epoch": 1.3738019169329074, + "grad_norm": 0.07043975591659546, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 860 + }, + { + "epoch": 1.3753993610223643, + "grad_norm": 0.06552740186452866, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 861 + }, + { + "epoch": 1.376996805111821, + "grad_norm": 0.167598158121109, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 862 + }, + { + "epoch": 1.378594249201278, + "grad_norm": 0.8627744913101196, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 863 + }, + { + "epoch": 1.3801916932907348, + "grad_norm": 0.23403562605381012, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 864 + }, + { + "epoch": 1.3817891373801916, + "grad_norm": 1.3149168491363525, + "learning_rate": 5e-07, + "loss": 0.0043, + "step": 865 + }, + { + "epoch": 1.3833865814696487, + "grad_norm": 0.16277176141738892, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 866 + }, + { + "epoch": 1.3849840255591055, + "grad_norm": 0.1406593918800354, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 867 + }, + { + "epoch": 1.3865814696485623, + "grad_norm": 0.24820701777935028, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 868 + }, + { + "epoch": 1.3881789137380192, + "grad_norm": 0.10332539677619934, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 869 + }, + { + "epoch": 1.389776357827476, + "grad_norm": 0.16070209443569183, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 870 + }, + { + "epoch": 1.3913738019169328, + "grad_norm": 0.16537906229496002, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 871 + }, + { + "epoch": 1.3929712460063897, + "grad_norm": 0.2067955881357193, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 872 + }, + { + "epoch": 1.3945686900958467, + "grad_norm": 0.1595505326986313, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 873 + }, + { + "epoch": 1.3961661341853036, + "grad_norm": 0.7071959972381592, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 874 + }, + { + "epoch": 1.3977635782747604, + "grad_norm": 0.08548534661531448, + "learning_rate": 5e-07, + "loss": 0.0005, + "step": 875 + }, + { + "epoch": 1.3993610223642172, + "grad_norm": 1.5511912107467651, + "learning_rate": 5e-07, + "loss": 0.005, + "step": 876 + }, + { + "epoch": 1.400958466453674, + "grad_norm": 0.622124969959259, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 877 + }, + { + "epoch": 1.4025559105431311, + "grad_norm": 3.853699207305908, + "learning_rate": 5e-07, + "loss": 0.0054, + "step": 878 + }, + { + "epoch": 1.4041533546325877, + "grad_norm": 0.07736141234636307, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 879 + }, + { + "epoch": 1.4057507987220448, + "grad_norm": 0.2597537338733673, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 880 + }, + { + "epoch": 1.4073482428115016, + "grad_norm": 0.09090615063905716, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 881 + }, + { + "epoch": 1.4089456869009584, + "grad_norm": 0.4754495620727539, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 882 + }, + { + "epoch": 1.4105431309904153, + "grad_norm": 0.4713440537452698, + "learning_rate": 5e-07, + "loss": 0.0045, + "step": 883 + }, + { + "epoch": 1.4121405750798721, + "grad_norm": 0.3018096387386322, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 884 + }, + { + "epoch": 1.4137380191693292, + "grad_norm": 0.21582446992397308, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 885 + }, + { + "epoch": 1.415335463258786, + "grad_norm": 0.28057220578193665, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 886 + }, + { + "epoch": 1.4169329073482428, + "grad_norm": 0.25230610370635986, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 887 + }, + { + "epoch": 1.4185303514376997, + "grad_norm": 0.3251979947090149, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 888 + }, + { + "epoch": 1.4201277955271565, + "grad_norm": 1.2839736938476562, + "learning_rate": 5e-07, + "loss": 0.0045, + "step": 889 + }, + { + "epoch": 1.4217252396166133, + "grad_norm": 0.6521425247192383, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 890 + }, + { + "epoch": 1.4233226837060702, + "grad_norm": 0.19354148209095, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 891 + }, + { + "epoch": 1.4249201277955272, + "grad_norm": 0.26791778206825256, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 892 + }, + { + "epoch": 1.426517571884984, + "grad_norm": 0.9978561997413635, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 893 + }, + { + "epoch": 1.428115015974441, + "grad_norm": 1.0886672735214233, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 894 + }, + { + "epoch": 1.4297124600638977, + "grad_norm": 0.1294880360364914, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 895 + }, + { + "epoch": 1.4313099041533546, + "grad_norm": 0.1127467229962349, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 896 + }, + { + "epoch": 1.4329073482428116, + "grad_norm": 0.14215292036533356, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 897 + }, + { + "epoch": 1.4345047923322682, + "grad_norm": 0.11055054515600204, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 898 + }, + { + "epoch": 1.4361022364217253, + "grad_norm": 0.8130955100059509, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 899 + }, + { + "epoch": 1.4376996805111821, + "grad_norm": 0.6208978295326233, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 900 + }, + { + "epoch": 1.439297124600639, + "grad_norm": 0.10077293962240219, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 901 + }, + { + "epoch": 1.4408945686900958, + "grad_norm": 0.15302683413028717, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 902 + }, + { + "epoch": 1.4424920127795526, + "grad_norm": 0.2711629867553711, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 903 + }, + { + "epoch": 1.4440894568690097, + "grad_norm": 0.6740185022354126, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 904 + }, + { + "epoch": 1.4456869009584665, + "grad_norm": 0.09207923710346222, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 905 + }, + { + "epoch": 1.4472843450479234, + "grad_norm": 0.13100707530975342, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 906 + }, + { + "epoch": 1.4488817891373802, + "grad_norm": 0.22404153645038605, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 907 + }, + { + "epoch": 1.450479233226837, + "grad_norm": 0.19576634466648102, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 908 + }, + { + "epoch": 1.4520766773162939, + "grad_norm": 0.3745575547218323, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 909 + }, + { + "epoch": 1.4536741214057507, + "grad_norm": 0.3619185984134674, + "learning_rate": 5e-07, + "loss": 0.0049, + "step": 910 + }, + { + "epoch": 1.4552715654952078, + "grad_norm": 0.11963102966547012, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 911 + }, + { + "epoch": 1.4568690095846646, + "grad_norm": 0.6312240958213806, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 912 + }, + { + "epoch": 1.4584664536741214, + "grad_norm": 0.44361862540245056, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 913 + }, + { + "epoch": 1.4600638977635783, + "grad_norm": 0.40353521704673767, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 914 + }, + { + "epoch": 1.461661341853035, + "grad_norm": 0.14119647443294525, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 915 + }, + { + "epoch": 1.4632587859424921, + "grad_norm": 0.07256528735160828, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 916 + }, + { + "epoch": 1.4648562300319488, + "grad_norm": 0.4794658422470093, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 917 + }, + { + "epoch": 1.4664536741214058, + "grad_norm": 0.2629548907279968, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 918 + }, + { + "epoch": 1.4680511182108626, + "grad_norm": 0.6776370406150818, + "learning_rate": 5e-07, + "loss": 0.0041, + "step": 919 + }, + { + "epoch": 1.4696485623003195, + "grad_norm": 0.14644776284694672, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 920 + }, + { + "epoch": 1.4712460063897763, + "grad_norm": 0.09278970211744308, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 921 + }, + { + "epoch": 1.4728434504792332, + "grad_norm": 0.08292179554700851, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 922 + }, + { + "epoch": 1.4744408945686902, + "grad_norm": 0.9622533321380615, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 923 + }, + { + "epoch": 1.476038338658147, + "grad_norm": 0.17527416348457336, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 924 + }, + { + "epoch": 1.4776357827476039, + "grad_norm": 0.07049129158258438, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 925 + }, + { + "epoch": 1.4792332268370607, + "grad_norm": 0.07232940942049026, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 926 + }, + { + "epoch": 1.4808306709265175, + "grad_norm": 0.06436185538768768, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 927 + }, + { + "epoch": 1.4824281150159744, + "grad_norm": 0.24670840799808502, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 928 + }, + { + "epoch": 1.4840255591054312, + "grad_norm": 0.7605392932891846, + "learning_rate": 5e-07, + "loss": 0.0041, + "step": 929 + }, + { + "epoch": 1.4856230031948883, + "grad_norm": 0.06602557003498077, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 930 + }, + { + "epoch": 1.487220447284345, + "grad_norm": 0.15429018437862396, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 931 + }, + { + "epoch": 1.488817891373802, + "grad_norm": 0.155389204621315, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 932 + }, + { + "epoch": 1.4904153354632588, + "grad_norm": 0.3137715458869934, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 933 + }, + { + "epoch": 1.4920127795527156, + "grad_norm": 0.13918346166610718, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 934 + }, + { + "epoch": 1.4936102236421724, + "grad_norm": 0.161741241812706, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 935 + }, + { + "epoch": 1.4952076677316293, + "grad_norm": 0.2871551513671875, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 936 + }, + { + "epoch": 1.4968051118210863, + "grad_norm": 0.3078411817550659, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 937 + }, + { + "epoch": 1.4984025559105432, + "grad_norm": 0.1792358011007309, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 938 + }, + { + "epoch": 1.5, + "grad_norm": 0.08885855227708817, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 939 + }, + { + "epoch": 1.5015974440894568, + "grad_norm": 0.10388941317796707, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 940 + }, + { + "epoch": 1.5031948881789137, + "grad_norm": 0.1879977583885193, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 941 + }, + { + "epoch": 1.5047923322683707, + "grad_norm": 0.08109968155622482, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 942 + }, + { + "epoch": 1.5063897763578273, + "grad_norm": 0.14684544503688812, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 943 + }, + { + "epoch": 1.5079872204472844, + "grad_norm": 0.250420480966568, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 944 + }, + { + "epoch": 1.5095846645367412, + "grad_norm": 0.0925571396946907, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 945 + }, + { + "epoch": 1.511182108626198, + "grad_norm": 0.1837424635887146, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 946 + }, + { + "epoch": 1.5127795527156551, + "grad_norm": 0.17538028955459595, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 947 + }, + { + "epoch": 1.5143769968051117, + "grad_norm": 0.08512210845947266, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 948 + }, + { + "epoch": 1.5159744408945688, + "grad_norm": 0.08973109722137451, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 949 + }, + { + "epoch": 1.5175718849840254, + "grad_norm": 0.1241975948214531, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 950 + }, + { + "epoch": 1.5191693290734825, + "grad_norm": 3.0702695846557617, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 951 + }, + { + "epoch": 1.5207667731629393, + "grad_norm": 0.07310276478528976, + "learning_rate": 5e-07, + "loss": 0.0006, + "step": 952 + }, + { + "epoch": 1.5223642172523961, + "grad_norm": 0.2920989692211151, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 953 + }, + { + "epoch": 1.5239616613418532, + "grad_norm": 0.5082775950431824, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 954 + }, + { + "epoch": 1.5255591054313098, + "grad_norm": 0.1549331098794937, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 955 + }, + { + "epoch": 1.5271565495207668, + "grad_norm": 0.1703399121761322, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 956 + }, + { + "epoch": 1.5287539936102237, + "grad_norm": 0.21957652270793915, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 957 + }, + { + "epoch": 1.5303514376996805, + "grad_norm": 0.3131657838821411, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 958 + }, + { + "epoch": 1.5319488817891374, + "grad_norm": 0.3629818260669708, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 959 + }, + { + "epoch": 1.5335463258785942, + "grad_norm": 0.2978671193122864, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 960 + }, + { + "epoch": 1.5351437699680512, + "grad_norm": 0.26085761189460754, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 961 + }, + { + "epoch": 1.5367412140575079, + "grad_norm": 0.13067355751991272, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 962 + }, + { + "epoch": 1.538338658146965, + "grad_norm": 1.0939770936965942, + "learning_rate": 5e-07, + "loss": 0.0075, + "step": 963 + }, + { + "epoch": 1.5399361022364217, + "grad_norm": 0.4009106755256653, + "learning_rate": 5e-07, + "loss": 0.0041, + "step": 964 + }, + { + "epoch": 1.5415335463258786, + "grad_norm": 0.69194495677948, + "learning_rate": 5e-07, + "loss": 0.0071, + "step": 965 + }, + { + "epoch": 1.5431309904153354, + "grad_norm": 0.15541456639766693, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 966 + }, + { + "epoch": 1.5447284345047922, + "grad_norm": 0.1687278002500534, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 967 + }, + { + "epoch": 1.5463258785942493, + "grad_norm": 0.08637956529855728, + "learning_rate": 5e-07, + "loss": 0.0004, + "step": 968 + }, + { + "epoch": 1.547923322683706, + "grad_norm": 0.49107223749160767, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 969 + }, + { + "epoch": 1.549520766773163, + "grad_norm": 0.40780824422836304, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 970 + }, + { + "epoch": 1.5511182108626198, + "grad_norm": 0.14083029329776764, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 971 + }, + { + "epoch": 1.5527156549520766, + "grad_norm": 0.1509598046541214, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 972 + }, + { + "epoch": 1.5543130990415337, + "grad_norm": 0.22124463319778442, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 973 + }, + { + "epoch": 1.5559105431309903, + "grad_norm": 0.29571980237960815, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 974 + }, + { + "epoch": 1.5575079872204474, + "grad_norm": 0.12444217503070831, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 975 + }, + { + "epoch": 1.5591054313099042, + "grad_norm": 0.19605819880962372, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 976 + }, + { + "epoch": 1.560702875399361, + "grad_norm": 0.36710816621780396, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 977 + }, + { + "epoch": 1.5623003194888179, + "grad_norm": 0.4394298493862152, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 978 + }, + { + "epoch": 1.5638977635782747, + "grad_norm": 3.3451671600341797, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 979 + }, + { + "epoch": 1.5654952076677318, + "grad_norm": 0.0950264111161232, + "learning_rate": 5e-07, + "loss": 0.0004, + "step": 980 + }, + { + "epoch": 1.5670926517571884, + "grad_norm": 0.5253363847732544, + "learning_rate": 5e-07, + "loss": 0.004, + "step": 981 + }, + { + "epoch": 1.5686900958466454, + "grad_norm": 0.08567991107702255, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 982 + }, + { + "epoch": 1.5702875399361023, + "grad_norm": 0.21886466443538666, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 983 + }, + { + "epoch": 1.571884984025559, + "grad_norm": 0.12182791531085968, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 984 + }, + { + "epoch": 1.573482428115016, + "grad_norm": 0.10883834213018417, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 985 + }, + { + "epoch": 1.5750798722044728, + "grad_norm": 0.2287680059671402, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 986 + }, + { + "epoch": 1.5766773162939298, + "grad_norm": 0.7266914248466492, + "learning_rate": 5e-07, + "loss": 0.0046, + "step": 987 + }, + { + "epoch": 1.5782747603833864, + "grad_norm": 0.11474244296550751, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 988 + }, + { + "epoch": 1.5798722044728435, + "grad_norm": 0.23904937505722046, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 989 + }, + { + "epoch": 1.5814696485623003, + "grad_norm": 0.055184707045555115, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 990 + }, + { + "epoch": 1.5830670926517572, + "grad_norm": 0.22353950142860413, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 991 + }, + { + "epoch": 1.5846645367412142, + "grad_norm": 0.19132664799690247, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 992 + }, + { + "epoch": 1.5862619808306708, + "grad_norm": 0.2778601348400116, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 993 + }, + { + "epoch": 1.5878594249201279, + "grad_norm": 0.05275161191821098, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 994 + }, + { + "epoch": 1.5894568690095847, + "grad_norm": 0.2460019290447235, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 995 + }, + { + "epoch": 1.5910543130990416, + "grad_norm": 0.3010371923446655, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 996 + }, + { + "epoch": 1.5926517571884984, + "grad_norm": 0.10904088616371155, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 997 + }, + { + "epoch": 1.5942492012779552, + "grad_norm": 0.05686507746577263, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 998 + }, + { + "epoch": 1.5958466453674123, + "grad_norm": 0.20314623415470123, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 999 + }, + { + "epoch": 1.5974440894568689, + "grad_norm": 1.588057279586792, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1000 + }, + { + "epoch": 1.599041533546326, + "grad_norm": 0.48463931679725647, + "learning_rate": 5e-07, + "loss": 0.004, + "step": 1001 + }, + { + "epoch": 1.6006389776357828, + "grad_norm": 0.6624436974525452, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 1002 + }, + { + "epoch": 1.6022364217252396, + "grad_norm": 0.14726407825946808, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1003 + }, + { + "epoch": 1.6038338658146964, + "grad_norm": 0.7131458520889282, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1004 + }, + { + "epoch": 1.6054313099041533, + "grad_norm": 0.46038198471069336, + "learning_rate": 5e-07, + "loss": 0.0043, + "step": 1005 + }, + { + "epoch": 1.6070287539936103, + "grad_norm": 0.2227628529071808, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 1006 + }, + { + "epoch": 1.608626198083067, + "grad_norm": 0.13341805338859558, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 1007 + }, + { + "epoch": 1.610223642172524, + "grad_norm": 0.07892493903636932, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 1008 + }, + { + "epoch": 1.6118210862619808, + "grad_norm": 0.12278909236192703, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 1009 + }, + { + "epoch": 1.6134185303514377, + "grad_norm": 0.5151563882827759, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 1010 + }, + { + "epoch": 1.6150159744408947, + "grad_norm": 0.23579691350460052, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 1011 + }, + { + "epoch": 1.6166134185303513, + "grad_norm": 0.05294935405254364, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1012 + }, + { + "epoch": 1.6182108626198084, + "grad_norm": 0.12360315769910812, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 1013 + }, + { + "epoch": 1.619808306709265, + "grad_norm": 0.35527291893959045, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1014 + }, + { + "epoch": 1.621405750798722, + "grad_norm": 0.25084754824638367, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 1015 + }, + { + "epoch": 1.623003194888179, + "grad_norm": 0.13500788807868958, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 1016 + }, + { + "epoch": 1.6246006389776357, + "grad_norm": 0.17942559719085693, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 1017 + }, + { + "epoch": 1.6261980830670928, + "grad_norm": 0.30760741233825684, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1018 + }, + { + "epoch": 1.6277955271565494, + "grad_norm": 0.12108216434717178, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1019 + }, + { + "epoch": 1.6293929712460065, + "grad_norm": 0.36486342549324036, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 1020 + }, + { + "epoch": 1.6309904153354633, + "grad_norm": 0.21493099629878998, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1021 + }, + { + "epoch": 1.6325878594249201, + "grad_norm": 0.2136039435863495, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1022 + }, + { + "epoch": 1.634185303514377, + "grad_norm": 0.19852106273174286, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 1023 + }, + { + "epoch": 1.6357827476038338, + "grad_norm": 0.07942666113376617, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 1024 + }, + { + "epoch": 1.6373801916932909, + "grad_norm": 0.14488880336284637, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 1025 + }, + { + "epoch": 1.6389776357827475, + "grad_norm": 0.07575450092554092, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1026 + }, + { + "epoch": 1.6405750798722045, + "grad_norm": 0.18136127293109894, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1027 + }, + { + "epoch": 1.6421725239616614, + "grad_norm": 0.09243427962064743, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1028 + }, + { + "epoch": 1.6437699680511182, + "grad_norm": 0.15622855722904205, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 1029 + }, + { + "epoch": 1.645367412140575, + "grad_norm": 0.20699208974838257, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1030 + }, + { + "epoch": 1.6469648562300319, + "grad_norm": 0.17613987624645233, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1031 + }, + { + "epoch": 1.648562300319489, + "grad_norm": 0.10338564217090607, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1032 + }, + { + "epoch": 1.6501597444089455, + "grad_norm": 0.07882916927337646, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1033 + }, + { + "epoch": 1.6517571884984026, + "grad_norm": 0.14468686282634735, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1034 + }, + { + "epoch": 1.6533546325878594, + "grad_norm": 0.33097293972969055, + "learning_rate": 5e-07, + "loss": 0.0037, + "step": 1035 + }, + { + "epoch": 1.6549520766773163, + "grad_norm": 0.46003833413124084, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 1036 + }, + { + "epoch": 1.6565495207667733, + "grad_norm": 0.38343682885169983, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 1037 + }, + { + "epoch": 1.65814696485623, + "grad_norm": 0.22435209155082703, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1038 + }, + { + "epoch": 1.659744408945687, + "grad_norm": 0.10138783603906631, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1039 + }, + { + "epoch": 1.6613418530351438, + "grad_norm": 0.42318132519721985, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 1040 + }, + { + "epoch": 1.6629392971246006, + "grad_norm": 0.10508925467729568, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1041 + }, + { + "epoch": 1.6645367412140575, + "grad_norm": 0.3548614978790283, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 1042 + }, + { + "epoch": 1.6661341853035143, + "grad_norm": 0.22214291989803314, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1043 + }, + { + "epoch": 1.6677316293929714, + "grad_norm": 0.23011279106140137, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1044 + }, + { + "epoch": 1.669329073482428, + "grad_norm": 0.08691424876451492, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 1045 + }, + { + "epoch": 1.670926517571885, + "grad_norm": 0.6495136618614197, + "learning_rate": 5e-07, + "loss": 0.005, + "step": 1046 + }, + { + "epoch": 1.6725239616613419, + "grad_norm": 0.20477008819580078, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1047 + }, + { + "epoch": 1.6741214057507987, + "grad_norm": 0.3025433123111725, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1048 + }, + { + "epoch": 1.6757188498402555, + "grad_norm": 0.1731894165277481, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1049 + }, + { + "epoch": 1.6773162939297124, + "grad_norm": 0.06371040642261505, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1050 + }, + { + "epoch": 1.6789137380191694, + "grad_norm": 0.14911451935768127, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1051 + }, + { + "epoch": 1.680511182108626, + "grad_norm": 0.09349862486124039, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1052 + }, + { + "epoch": 1.682108626198083, + "grad_norm": 0.21330159902572632, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 1053 + }, + { + "epoch": 1.68370607028754, + "grad_norm": 0.13860629498958588, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1054 + }, + { + "epoch": 1.6853035143769968, + "grad_norm": 0.1365477740764618, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1055 + }, + { + "epoch": 1.6869009584664538, + "grad_norm": 0.17503461241722107, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1056 + }, + { + "epoch": 1.6884984025559104, + "grad_norm": 0.111385278403759, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 1057 + }, + { + "epoch": 1.6900958466453675, + "grad_norm": 0.06377507001161575, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1058 + }, + { + "epoch": 1.6916932907348243, + "grad_norm": 0.0641062781214714, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1059 + }, + { + "epoch": 1.6932907348242812, + "grad_norm": 0.09682592004537582, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1060 + }, + { + "epoch": 1.694888178913738, + "grad_norm": 0.1440698206424713, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1061 + }, + { + "epoch": 1.6964856230031948, + "grad_norm": 0.04743053764104843, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1062 + }, + { + "epoch": 1.6980830670926519, + "grad_norm": 0.18199731409549713, + "learning_rate": 5e-07, + "loss": 0.0036, + "step": 1063 + }, + { + "epoch": 1.6996805111821085, + "grad_norm": 0.053704094141721725, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 1064 + }, + { + "epoch": 1.7012779552715656, + "grad_norm": 0.10925064235925674, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1065 + }, + { + "epoch": 1.7028753993610224, + "grad_norm": 0.09475322812795639, + "learning_rate": 5e-07, + "loss": 0.0005, + "step": 1066 + }, + { + "epoch": 1.7044728434504792, + "grad_norm": 0.14424358308315277, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1067 + }, + { + "epoch": 1.706070287539936, + "grad_norm": 0.07309069484472275, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1068 + }, + { + "epoch": 1.707667731629393, + "grad_norm": 0.17733772099018097, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 1069 + }, + { + "epoch": 1.70926517571885, + "grad_norm": 0.9875695705413818, + "learning_rate": 5e-07, + "loss": 0.0086, + "step": 1070 + }, + { + "epoch": 1.7108626198083066, + "grad_norm": 0.408796101808548, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1071 + }, + { + "epoch": 1.7124600638977636, + "grad_norm": 0.17582127451896667, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1072 + }, + { + "epoch": 1.7140575079872205, + "grad_norm": 0.1303548812866211, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1073 + }, + { + "epoch": 1.7156549520766773, + "grad_norm": 0.10945620387792587, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1074 + }, + { + "epoch": 1.7172523961661343, + "grad_norm": 0.07630528509616852, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1075 + }, + { + "epoch": 1.718849840255591, + "grad_norm": 0.07517647743225098, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1076 + }, + { + "epoch": 1.720447284345048, + "grad_norm": 0.0669432058930397, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1077 + }, + { + "epoch": 1.7220447284345048, + "grad_norm": 0.1877792626619339, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 1078 + }, + { + "epoch": 1.7236421725239617, + "grad_norm": 0.04336528107523918, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 1079 + }, + { + "epoch": 1.7252396166134185, + "grad_norm": 0.06632347404956818, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1080 + }, + { + "epoch": 1.7268370607028753, + "grad_norm": 0.09459614008665085, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1081 + }, + { + "epoch": 1.7284345047923324, + "grad_norm": 0.5509624481201172, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1082 + }, + { + "epoch": 1.730031948881789, + "grad_norm": 0.2771284878253937, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1083 + }, + { + "epoch": 1.731629392971246, + "grad_norm": 0.05486772954463959, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1084 + }, + { + "epoch": 1.733226837060703, + "grad_norm": 0.2757831811904907, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 1085 + }, + { + "epoch": 1.7348242811501597, + "grad_norm": 0.04002672806382179, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 1086 + }, + { + "epoch": 1.7364217252396166, + "grad_norm": 0.06649244576692581, + "learning_rate": 5e-07, + "loss": 0.0003, + "step": 1087 + }, + { + "epoch": 1.7380191693290734, + "grad_norm": 0.11669522523880005, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1088 + }, + { + "epoch": 1.7396166134185305, + "grad_norm": 0.12358154356479645, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1089 + }, + { + "epoch": 1.741214057507987, + "grad_norm": 0.0829407200217247, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1090 + }, + { + "epoch": 1.7428115015974441, + "grad_norm": 0.09355901926755905, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 1091 + }, + { + "epoch": 1.744408945686901, + "grad_norm": 0.0956120640039444, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1092 + }, + { + "epoch": 1.7460063897763578, + "grad_norm": 0.4703699052333832, + "learning_rate": 5e-07, + "loss": 0.0055, + "step": 1093 + }, + { + "epoch": 1.7476038338658149, + "grad_norm": 0.3924386203289032, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 1094 + }, + { + "epoch": 1.7492012779552715, + "grad_norm": 0.16191145777702332, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1095 + }, + { + "epoch": 1.7507987220447285, + "grad_norm": 0.06510366499423981, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1096 + }, + { + "epoch": 1.7523961661341851, + "grad_norm": 0.19621287286281586, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1097 + }, + { + "epoch": 1.7539936102236422, + "grad_norm": 0.1870349496603012, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1098 + }, + { + "epoch": 1.755591054313099, + "grad_norm": 0.14898639917373657, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1099 + }, + { + "epoch": 1.7571884984025559, + "grad_norm": 0.15484674274921417, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1100 + }, + { + "epoch": 1.758785942492013, + "grad_norm": 0.1420629471540451, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1101 + }, + { + "epoch": 1.7603833865814695, + "grad_norm": 0.12516824901103973, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1102 + }, + { + "epoch": 1.7619808306709266, + "grad_norm": 0.09688828140497208, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1103 + }, + { + "epoch": 1.7635782747603834, + "grad_norm": 0.09412126988172531, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1104 + }, + { + "epoch": 1.7651757188498403, + "grad_norm": 1.9070310592651367, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1105 + }, + { + "epoch": 1.766773162939297, + "grad_norm": 0.16800744831562042, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1106 + }, + { + "epoch": 1.768370607028754, + "grad_norm": 0.10442492365837097, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1107 + }, + { + "epoch": 1.769968051118211, + "grad_norm": 0.16451716423034668, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1108 + }, + { + "epoch": 1.7715654952076676, + "grad_norm": 0.8484131097793579, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 1109 + }, + { + "epoch": 1.7731629392971247, + "grad_norm": 0.3258044719696045, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 1110 + }, + { + "epoch": 1.7747603833865815, + "grad_norm": 0.17745210230350494, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1111 + }, + { + "epoch": 1.7763578274760383, + "grad_norm": 0.07777510583400726, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1112 + }, + { + "epoch": 1.7779552715654952, + "grad_norm": 0.08588631451129913, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1113 + }, + { + "epoch": 1.779552715654952, + "grad_norm": 0.6624420881271362, + "learning_rate": 5e-07, + "loss": 0.005, + "step": 1114 + }, + { + "epoch": 1.781150159744409, + "grad_norm": 0.11576593667268753, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1115 + }, + { + "epoch": 1.7827476038338657, + "grad_norm": 0.859341561794281, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1116 + }, + { + "epoch": 1.7843450479233227, + "grad_norm": 0.10600177943706512, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1117 + }, + { + "epoch": 1.7859424920127795, + "grad_norm": 0.14881321787834167, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1118 + }, + { + "epoch": 1.7875399361022364, + "grad_norm": 0.05584081634879112, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1119 + }, + { + "epoch": 1.7891373801916934, + "grad_norm": 0.06437966227531433, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1120 + }, + { + "epoch": 1.79073482428115, + "grad_norm": 0.2528286874294281, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1121 + }, + { + "epoch": 1.792332268370607, + "grad_norm": 0.18848662078380585, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1122 + }, + { + "epoch": 1.793929712460064, + "grad_norm": 0.14006567001342773, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 1123 + }, + { + "epoch": 1.7955271565495208, + "grad_norm": 0.07045773416757584, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 1124 + }, + { + "epoch": 1.7971246006389776, + "grad_norm": 0.0998934954404831, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1125 + }, + { + "epoch": 1.7987220447284344, + "grad_norm": 0.09063876420259476, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1126 + }, + { + "epoch": 1.8003194888178915, + "grad_norm": 0.2747619152069092, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 1127 + }, + { + "epoch": 1.8019169329073481, + "grad_norm": 0.0671829804778099, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1128 + }, + { + "epoch": 1.8035143769968052, + "grad_norm": 0.23768068850040436, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 1129 + }, + { + "epoch": 1.805111821086262, + "grad_norm": 0.14285339415073395, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1130 + }, + { + "epoch": 1.8067092651757188, + "grad_norm": 3.77717661857605, + "learning_rate": 5e-07, + "loss": 0.005, + "step": 1131 + }, + { + "epoch": 1.8083067092651757, + "grad_norm": 0.24507032334804535, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 1132 + }, + { + "epoch": 1.8099041533546325, + "grad_norm": 0.15166565775871277, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1133 + }, + { + "epoch": 1.8115015974440896, + "grad_norm": 0.1067359670996666, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1134 + }, + { + "epoch": 1.8130990415335462, + "grad_norm": 0.18612955510616302, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1135 + }, + { + "epoch": 1.8146964856230032, + "grad_norm": 0.17414437234401703, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1136 + }, + { + "epoch": 1.81629392971246, + "grad_norm": 0.3064412474632263, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 1137 + }, + { + "epoch": 1.817891373801917, + "grad_norm": 0.071932353079319, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1138 + }, + { + "epoch": 1.819488817891374, + "grad_norm": 0.1119665578007698, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1139 + }, + { + "epoch": 1.8210862619808306, + "grad_norm": 0.20450662076473236, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1140 + }, + { + "epoch": 1.8226837060702876, + "grad_norm": 0.13636766374111176, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1141 + }, + { + "epoch": 1.8242811501597445, + "grad_norm": 0.07295648753643036, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1142 + }, + { + "epoch": 1.8258785942492013, + "grad_norm": 0.17274846136569977, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 1143 + }, + { + "epoch": 1.8274760383386581, + "grad_norm": 0.14178737998008728, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 1144 + }, + { + "epoch": 1.829073482428115, + "grad_norm": 0.2096203863620758, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1145 + }, + { + "epoch": 1.830670926517572, + "grad_norm": 0.09562284499406815, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1146 + }, + { + "epoch": 1.8322683706070286, + "grad_norm": 0.10589710623025894, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 1147 + }, + { + "epoch": 1.8338658146964857, + "grad_norm": 0.2623519003391266, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 1148 + }, + { + "epoch": 1.8354632587859425, + "grad_norm": 0.06434428691864014, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 1149 + }, + { + "epoch": 1.8370607028753994, + "grad_norm": 0.31623294949531555, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 1150 + }, + { + "epoch": 1.8386581469648562, + "grad_norm": 0.11535608768463135, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1151 + }, + { + "epoch": 1.840255591054313, + "grad_norm": 0.25482088327407837, + "learning_rate": 5e-07, + "loss": 0.0006, + "step": 1152 + }, + { + "epoch": 1.84185303514377, + "grad_norm": 0.1523568332195282, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1153 + }, + { + "epoch": 1.8434504792332267, + "grad_norm": 0.09644993394613266, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1154 + }, + { + "epoch": 1.8450479233226837, + "grad_norm": 0.30026814341545105, + "learning_rate": 5e-07, + "loss": 0.0036, + "step": 1155 + }, + { + "epoch": 1.8466453674121406, + "grad_norm": 5.034646511077881, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1156 + }, + { + "epoch": 1.8482428115015974, + "grad_norm": 0.1875925064086914, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1157 + }, + { + "epoch": 1.8498402555910545, + "grad_norm": 0.15650227665901184, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1158 + }, + { + "epoch": 1.851437699680511, + "grad_norm": 0.17901848256587982, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1159 + }, + { + "epoch": 1.8530351437699681, + "grad_norm": 0.27782970666885376, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1160 + }, + { + "epoch": 1.854632587859425, + "grad_norm": 0.11091984808444977, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1161 + }, + { + "epoch": 1.8562300319488818, + "grad_norm": 0.3836122453212738, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1162 + }, + { + "epoch": 1.8578274760383386, + "grad_norm": 2.26542067527771, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 1163 + }, + { + "epoch": 1.8594249201277955, + "grad_norm": 0.8616245985031128, + "learning_rate": 5e-07, + "loss": 0.0051, + "step": 1164 + }, + { + "epoch": 1.8610223642172525, + "grad_norm": 0.1418013572692871, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 1165 + }, + { + "epoch": 1.8626198083067091, + "grad_norm": 0.20582883059978485, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 1166 + }, + { + "epoch": 1.8642172523961662, + "grad_norm": 0.05279695987701416, + "learning_rate": 5e-07, + "loss": 0.0006, + "step": 1167 + }, + { + "epoch": 1.865814696485623, + "grad_norm": 0.30152246356010437, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1168 + }, + { + "epoch": 1.8674121405750799, + "grad_norm": 1.215949296951294, + "learning_rate": 5e-07, + "loss": 0.0042, + "step": 1169 + }, + { + "epoch": 1.8690095846645367, + "grad_norm": 0.1583871841430664, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 1170 + }, + { + "epoch": 1.8706070287539935, + "grad_norm": 0.18552114069461823, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 1171 + }, + { + "epoch": 1.8722044728434506, + "grad_norm": 0.25958728790283203, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1172 + }, + { + "epoch": 1.8738019169329072, + "grad_norm": 0.3198787271976471, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1173 + }, + { + "epoch": 1.8753993610223643, + "grad_norm": 0.4293941557407379, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 1174 + }, + { + "epoch": 1.876996805111821, + "grad_norm": 0.12936322391033173, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1175 + }, + { + "epoch": 1.878594249201278, + "grad_norm": 0.26641571521759033, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 1176 + }, + { + "epoch": 1.880191693290735, + "grad_norm": 1.098074197769165, + "learning_rate": 5e-07, + "loss": 0.0036, + "step": 1177 + }, + { + "epoch": 1.8817891373801916, + "grad_norm": 0.11431043595075607, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1178 + }, + { + "epoch": 1.8833865814696487, + "grad_norm": 0.37071993947029114, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1179 + }, + { + "epoch": 1.8849840255591053, + "grad_norm": 0.08628468960523605, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1180 + }, + { + "epoch": 1.8865814696485623, + "grad_norm": 0.14398378133773804, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1181 + }, + { + "epoch": 1.8881789137380192, + "grad_norm": 0.35993340611457825, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 1182 + }, + { + "epoch": 1.889776357827476, + "grad_norm": 0.1559559851884842, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1183 + }, + { + "epoch": 1.891373801916933, + "grad_norm": 0.2564754784107208, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1184 + }, + { + "epoch": 1.8929712460063897, + "grad_norm": 0.11279148608446121, + "learning_rate": 5e-07, + "loss": 0.0005, + "step": 1185 + }, + { + "epoch": 1.8945686900958467, + "grad_norm": 0.08808083087205887, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1186 + }, + { + "epoch": 1.8961661341853036, + "grad_norm": 0.10382703691720963, + "learning_rate": 5e-07, + "loss": 0.0004, + "step": 1187 + }, + { + "epoch": 1.8977635782747604, + "grad_norm": 0.15366709232330322, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 1188 + }, + { + "epoch": 1.8993610223642172, + "grad_norm": 0.10865357518196106, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1189 + }, + { + "epoch": 1.900958466453674, + "grad_norm": 0.24182426929473877, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1190 + }, + { + "epoch": 1.9025559105431311, + "grad_norm": 0.4804140627384186, + "learning_rate": 5e-07, + "loss": 0.0048, + "step": 1191 + }, + { + "epoch": 1.9041533546325877, + "grad_norm": 0.23404286801815033, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1192 + }, + { + "epoch": 1.9057507987220448, + "grad_norm": 0.36086928844451904, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1193 + }, + { + "epoch": 1.9073482428115016, + "grad_norm": 0.5354902744293213, + "learning_rate": 5e-07, + "loss": 0.004, + "step": 1194 + }, + { + "epoch": 1.9089456869009584, + "grad_norm": 0.11621209979057312, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1195 + }, + { + "epoch": 1.9105431309904153, + "grad_norm": 0.1943303495645523, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1196 + }, + { + "epoch": 1.9121405750798721, + "grad_norm": 0.2628275454044342, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 1197 + }, + { + "epoch": 1.9137380191693292, + "grad_norm": 0.20775623619556427, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1198 + }, + { + "epoch": 1.9153354632587858, + "grad_norm": 0.12529809772968292, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1199 + }, + { + "epoch": 1.9169329073482428, + "grad_norm": 0.09771095961332321, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1200 + }, + { + "epoch": 1.9185303514376997, + "grad_norm": 0.10013385117053986, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1201 + }, + { + "epoch": 1.9201277955271565, + "grad_norm": 0.29045844078063965, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 1202 + }, + { + "epoch": 1.9217252396166136, + "grad_norm": 0.15383721888065338, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 1203 + }, + { + "epoch": 1.9233226837060702, + "grad_norm": 0.6897266507148743, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1204 + }, + { + "epoch": 1.9249201277955272, + "grad_norm": 0.15652583539485931, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1205 + }, + { + "epoch": 1.926517571884984, + "grad_norm": 0.0853152722120285, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1206 + }, + { + "epoch": 1.928115015974441, + "grad_norm": 0.11881982535123825, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1207 + }, + { + "epoch": 1.9297124600638977, + "grad_norm": 0.10427961498498917, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1208 + }, + { + "epoch": 1.9313099041533546, + "grad_norm": 0.1068776324391365, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1209 + }, + { + "epoch": 1.9329073482428116, + "grad_norm": 0.13248351216316223, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 1210 + }, + { + "epoch": 1.9345047923322682, + "grad_norm": 0.25468289852142334, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1211 + }, + { + "epoch": 1.9361022364217253, + "grad_norm": 0.16791756451129913, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 1212 + }, + { + "epoch": 1.9376996805111821, + "grad_norm": 0.13099579513072968, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1213 + }, + { + "epoch": 1.939297124600639, + "grad_norm": 0.26930350065231323, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 1214 + }, + { + "epoch": 1.9408945686900958, + "grad_norm": 2.077728271484375, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1215 + }, + { + "epoch": 1.9424920127795526, + "grad_norm": 0.31897902488708496, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1216 + }, + { + "epoch": 1.9440894568690097, + "grad_norm": 0.07247374951839447, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1217 + }, + { + "epoch": 1.9456869009584663, + "grad_norm": 0.11995717883110046, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1218 + }, + { + "epoch": 1.9472843450479234, + "grad_norm": 0.2621629536151886, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1219 + }, + { + "epoch": 1.9488817891373802, + "grad_norm": 0.1726490557193756, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 1220 + }, + { + "epoch": 1.950479233226837, + "grad_norm": 0.16272880136966705, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1221 + }, + { + "epoch": 1.952076677316294, + "grad_norm": 0.10036841779947281, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 1222 + }, + { + "epoch": 1.9536741214057507, + "grad_norm": 0.7988063097000122, + "learning_rate": 5e-07, + "loss": 0.0046, + "step": 1223 + }, + { + "epoch": 1.9552715654952078, + "grad_norm": 0.2782542407512665, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1224 + }, + { + "epoch": 1.9568690095846646, + "grad_norm": 0.2055627703666687, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 1225 + }, + { + "epoch": 1.9584664536741214, + "grad_norm": 0.19908927381038666, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1226 + }, + { + "epoch": 1.9600638977635783, + "grad_norm": 0.21738676726818085, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1227 + }, + { + "epoch": 1.961661341853035, + "grad_norm": 0.16997866332530975, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1228 + }, + { + "epoch": 1.9632587859424921, + "grad_norm": 0.10853756219148636, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1229 + }, + { + "epoch": 1.9648562300319488, + "grad_norm": 0.2967356741428375, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 1230 + }, + { + "epoch": 1.9664536741214058, + "grad_norm": 0.24150174856185913, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 1231 + }, + { + "epoch": 1.9680511182108626, + "grad_norm": 0.09346190094947815, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1232 + }, + { + "epoch": 1.9696485623003195, + "grad_norm": 0.16779378056526184, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1233 + }, + { + "epoch": 1.9712460063897763, + "grad_norm": 0.1114976555109024, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1234 + }, + { + "epoch": 1.9728434504792332, + "grad_norm": 0.45826414227485657, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 1235 + }, + { + "epoch": 1.9744408945686902, + "grad_norm": 0.4743911921977997, + "learning_rate": 5e-07, + "loss": 0.0034, + "step": 1236 + }, + { + "epoch": 1.9760383386581468, + "grad_norm": 0.144659623503685, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 1237 + }, + { + "epoch": 1.9776357827476039, + "grad_norm": 0.9392029643058777, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 1238 + }, + { + "epoch": 1.9792332268370607, + "grad_norm": 0.5866590738296509, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 1239 + }, + { + "epoch": 1.9808306709265175, + "grad_norm": 0.2669076919555664, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 1240 + }, + { + "epoch": 1.9824281150159746, + "grad_norm": 0.34604692459106445, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1241 + }, + { + "epoch": 1.9840255591054312, + "grad_norm": 0.27668237686157227, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1242 + }, + { + "epoch": 1.9856230031948883, + "grad_norm": 0.1560208648443222, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1243 + }, + { + "epoch": 1.9872204472843449, + "grad_norm": 0.35375985503196716, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1244 + }, + { + "epoch": 1.988817891373802, + "grad_norm": 0.134053036570549, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1245 + }, + { + "epoch": 1.9904153354632588, + "grad_norm": 0.10250476002693176, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1246 + }, + { + "epoch": 1.9920127795527156, + "grad_norm": 0.13041843473911285, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1247 + }, + { + "epoch": 1.9936102236421727, + "grad_norm": 0.062465403228998184, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1248 + }, + { + "epoch": 1.9952076677316293, + "grad_norm": 0.10093759000301361, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1249 + }, + { + "epoch": 1.9968051118210863, + "grad_norm": 0.08476297557353973, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1250 + }, + { + "epoch": 1.9984025559105432, + "grad_norm": 0.22871960699558258, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1251 + }, + { + "epoch": 2.0, + "grad_norm": 0.19058428704738617, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1252 + }, + { + "epoch": 2.001597444089457, + "grad_norm": 0.2544306218624115, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1253 + }, + { + "epoch": 2.0031948881789137, + "grad_norm": 0.14264468848705292, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 1254 + }, + { + "epoch": 2.0047923322683707, + "grad_norm": 0.18681064248085022, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1255 + }, + { + "epoch": 2.0063897763578273, + "grad_norm": 0.17501874268054962, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1256 + }, + { + "epoch": 2.0079872204472844, + "grad_norm": 0.1236346885561943, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1257 + }, + { + "epoch": 2.009584664536741, + "grad_norm": 0.08623294532299042, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 1258 + }, + { + "epoch": 2.011182108626198, + "grad_norm": 0.12931625545024872, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1259 + }, + { + "epoch": 2.012779552715655, + "grad_norm": 0.10413195192813873, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1260 + }, + { + "epoch": 2.0143769968051117, + "grad_norm": 0.23485882580280304, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1261 + }, + { + "epoch": 2.015974440894569, + "grad_norm": 0.07863974571228027, + "learning_rate": 5e-07, + "loss": 0.0005, + "step": 1262 + }, + { + "epoch": 2.0175718849840254, + "grad_norm": 0.33484408259391785, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 1263 + }, + { + "epoch": 2.0191693290734825, + "grad_norm": 0.31939253211021423, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 1264 + }, + { + "epoch": 2.0207667731629395, + "grad_norm": 0.10941127687692642, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 1265 + }, + { + "epoch": 2.022364217252396, + "grad_norm": 0.10784590244293213, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1266 + }, + { + "epoch": 2.023961661341853, + "grad_norm": 0.16436795890331268, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 1267 + }, + { + "epoch": 2.02555910543131, + "grad_norm": 0.09606944024562836, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1268 + }, + { + "epoch": 2.027156549520767, + "grad_norm": 0.17230142652988434, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1269 + }, + { + "epoch": 2.0287539936102235, + "grad_norm": 0.5120985507965088, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 1270 + }, + { + "epoch": 2.0303514376996805, + "grad_norm": 0.06792975217103958, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 1271 + }, + { + "epoch": 2.0319488817891376, + "grad_norm": 0.12373680621385574, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1272 + }, + { + "epoch": 2.033546325878594, + "grad_norm": 0.10998155921697617, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1273 + }, + { + "epoch": 2.0351437699680512, + "grad_norm": 0.15938100218772888, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 1274 + }, + { + "epoch": 2.036741214057508, + "grad_norm": 0.07384390383958817, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 1275 + }, + { + "epoch": 2.038338658146965, + "grad_norm": 0.23505450785160065, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1276 + }, + { + "epoch": 2.0399361022364215, + "grad_norm": 1.6043593883514404, + "learning_rate": 5e-07, + "loss": 0.0044, + "step": 1277 + }, + { + "epoch": 2.0415335463258786, + "grad_norm": 0.08369333297014236, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 1278 + }, + { + "epoch": 2.0431309904153356, + "grad_norm": 0.15266691148281097, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1279 + }, + { + "epoch": 2.0447284345047922, + "grad_norm": 0.1948157548904419, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1280 + }, + { + "epoch": 2.0463258785942493, + "grad_norm": 0.33584949374198914, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 1281 + }, + { + "epoch": 2.047923322683706, + "grad_norm": 0.6691318154335022, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1282 + }, + { + "epoch": 2.049520766773163, + "grad_norm": 0.1382388472557068, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1283 + }, + { + "epoch": 2.0511182108626196, + "grad_norm": 0.12497832626104355, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 1284 + }, + { + "epoch": 2.0527156549520766, + "grad_norm": 0.27157703042030334, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1285 + }, + { + "epoch": 2.0543130990415337, + "grad_norm": 0.2626846134662628, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 1286 + }, + { + "epoch": 2.0559105431309903, + "grad_norm": 0.16861018538475037, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1287 + }, + { + "epoch": 2.0575079872204474, + "grad_norm": 0.049949146807193756, + "learning_rate": 5e-07, + "loss": 0.0004, + "step": 1288 + }, + { + "epoch": 2.059105431309904, + "grad_norm": 0.0886000543832779, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 1289 + }, + { + "epoch": 2.060702875399361, + "grad_norm": 0.20434832572937012, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1290 + }, + { + "epoch": 2.062300319488818, + "grad_norm": 0.11731091886758804, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1291 + }, + { + "epoch": 2.0638977635782747, + "grad_norm": 0.20026597380638123, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1292 + }, + { + "epoch": 2.0654952076677318, + "grad_norm": 0.27669885754585266, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 1293 + }, + { + "epoch": 2.0670926517571884, + "grad_norm": 0.2785263657569885, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 1294 + }, + { + "epoch": 2.0686900958466454, + "grad_norm": 0.15540768206119537, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1295 + }, + { + "epoch": 2.070287539936102, + "grad_norm": 0.09532786905765533, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 1296 + }, + { + "epoch": 2.071884984025559, + "grad_norm": 0.1427212655544281, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1297 + }, + { + "epoch": 2.073482428115016, + "grad_norm": 0.3582955300807953, + "learning_rate": 5e-07, + "loss": 0.003, + "step": 1298 + }, + { + "epoch": 2.0750798722044728, + "grad_norm": 0.3676643669605255, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 1299 + }, + { + "epoch": 2.07667731629393, + "grad_norm": 0.15875279903411865, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1300 + }, + { + "epoch": 2.0782747603833864, + "grad_norm": 0.17880000174045563, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1301 + }, + { + "epoch": 2.0798722044728435, + "grad_norm": 0.09481951594352722, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1302 + }, + { + "epoch": 2.0814696485623, + "grad_norm": 0.0720115378499031, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 1303 + }, + { + "epoch": 2.083067092651757, + "grad_norm": 0.1885913908481598, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1304 + }, + { + "epoch": 2.084664536741214, + "grad_norm": 0.07471724599599838, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 1305 + }, + { + "epoch": 2.086261980830671, + "grad_norm": 0.10295652598142624, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1306 + }, + { + "epoch": 2.087859424920128, + "grad_norm": 0.48676663637161255, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1307 + }, + { + "epoch": 2.0894568690095845, + "grad_norm": 0.06689424067735672, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1308 + }, + { + "epoch": 2.0910543130990416, + "grad_norm": 0.3771530091762543, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 1309 + }, + { + "epoch": 2.0926517571884986, + "grad_norm": 0.37652865052223206, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1310 + }, + { + "epoch": 2.094249201277955, + "grad_norm": 0.0743250697851181, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1311 + }, + { + "epoch": 2.0958466453674123, + "grad_norm": 0.1055615022778511, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 1312 + }, + { + "epoch": 2.097444089456869, + "grad_norm": 0.25487279891967773, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1313 + }, + { + "epoch": 2.099041533546326, + "grad_norm": 0.13495591282844543, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1314 + }, + { + "epoch": 2.1006389776357826, + "grad_norm": 0.17297884821891785, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1315 + }, + { + "epoch": 2.1022364217252396, + "grad_norm": 0.14259670674800873, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1316 + }, + { + "epoch": 2.1038338658146967, + "grad_norm": 0.06441140174865723, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 1317 + }, + { + "epoch": 2.1054313099041533, + "grad_norm": 0.0832604393362999, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1318 + }, + { + "epoch": 2.1070287539936103, + "grad_norm": 0.6112750172615051, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1319 + }, + { + "epoch": 2.108626198083067, + "grad_norm": 0.14800317585468292, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1320 + }, + { + "epoch": 2.110223642172524, + "grad_norm": 0.389436811208725, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 1321 + }, + { + "epoch": 2.1118210862619806, + "grad_norm": 0.1899157464504242, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1322 + }, + { + "epoch": 2.1134185303514377, + "grad_norm": 0.08389966934919357, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1323 + }, + { + "epoch": 2.1150159744408947, + "grad_norm": 0.10899113118648529, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1324 + }, + { + "epoch": 2.1166134185303513, + "grad_norm": 0.10662740468978882, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1325 + }, + { + "epoch": 2.1182108626198084, + "grad_norm": 0.11692536622285843, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1326 + }, + { + "epoch": 2.119808306709265, + "grad_norm": 0.09900128096342087, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1327 + }, + { + "epoch": 2.121405750798722, + "grad_norm": 0.08334943652153015, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 1328 + }, + { + "epoch": 2.123003194888179, + "grad_norm": 0.10552877187728882, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1329 + }, + { + "epoch": 2.1246006389776357, + "grad_norm": 0.8137688040733337, + "learning_rate": 5e-07, + "loss": 0.0014, + "step": 1330 + }, + { + "epoch": 2.126198083067093, + "grad_norm": 0.06990483403205872, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 1331 + }, + { + "epoch": 2.1277955271565494, + "grad_norm": 0.1675483137369156, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1332 + }, + { + "epoch": 2.1293929712460065, + "grad_norm": 0.2427297979593277, + "learning_rate": 5e-07, + "loss": 0.0024, + "step": 1333 + }, + { + "epoch": 2.130990415335463, + "grad_norm": 0.1828804910182953, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1334 + }, + { + "epoch": 2.13258785942492, + "grad_norm": 0.35362717509269714, + "learning_rate": 5e-07, + "loss": 0.0029, + "step": 1335 + }, + { + "epoch": 2.134185303514377, + "grad_norm": 0.21310366690158844, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1336 + }, + { + "epoch": 2.135782747603834, + "grad_norm": 0.14100836217403412, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1337 + }, + { + "epoch": 2.137380191693291, + "grad_norm": 0.10196174681186676, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 1338 + }, + { + "epoch": 2.1389776357827475, + "grad_norm": 0.2148902863264084, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1339 + }, + { + "epoch": 2.1405750798722045, + "grad_norm": 0.05017423257231712, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 1340 + }, + { + "epoch": 2.142172523961661, + "grad_norm": 2.62032151222229, + "learning_rate": 5e-07, + "loss": 0.0035, + "step": 1341 + }, + { + "epoch": 2.143769968051118, + "grad_norm": 0.23405258357524872, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1342 + }, + { + "epoch": 2.1453674121405752, + "grad_norm": 0.05571659281849861, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1343 + }, + { + "epoch": 2.146964856230032, + "grad_norm": 0.17637008428573608, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1344 + }, + { + "epoch": 2.148562300319489, + "grad_norm": 0.10391382873058319, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1345 + }, + { + "epoch": 2.1501597444089455, + "grad_norm": 0.5443282723426819, + "learning_rate": 5e-07, + "loss": 0.0027, + "step": 1346 + }, + { + "epoch": 2.1517571884984026, + "grad_norm": 0.09875518828630447, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1347 + }, + { + "epoch": 2.1533546325878596, + "grad_norm": 0.3345401883125305, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1348 + }, + { + "epoch": 2.1549520766773163, + "grad_norm": 0.14103494584560394, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1349 + }, + { + "epoch": 2.1565495207667733, + "grad_norm": 0.1581650972366333, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1350 + }, + { + "epoch": 2.15814696485623, + "grad_norm": 0.1288004070520401, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 1351 + }, + { + "epoch": 2.159744408945687, + "grad_norm": 0.19039283692836761, + "learning_rate": 5e-07, + "loss": 0.0026, + "step": 1352 + }, + { + "epoch": 2.1613418530351436, + "grad_norm": 0.15048496425151825, + "learning_rate": 5e-07, + "loss": 0.0009, + "step": 1353 + }, + { + "epoch": 2.1629392971246006, + "grad_norm": 0.14754633605480194, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1354 + }, + { + "epoch": 2.1645367412140577, + "grad_norm": 0.20571894943714142, + "learning_rate": 5e-07, + "loss": 0.0033, + "step": 1355 + }, + { + "epoch": 2.1661341853035143, + "grad_norm": 0.12298233062028885, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1356 + }, + { + "epoch": 2.1677316293929714, + "grad_norm": 0.060658592730760574, + "learning_rate": 5e-07, + "loss": 0.0007, + "step": 1357 + }, + { + "epoch": 2.169329073482428, + "grad_norm": 0.2829779088497162, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1358 + }, + { + "epoch": 2.170926517571885, + "grad_norm": 0.1882810741662979, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1359 + }, + { + "epoch": 2.1725239616613417, + "grad_norm": 0.5703380703926086, + "learning_rate": 5e-07, + "loss": 0.0039, + "step": 1360 + }, + { + "epoch": 2.1741214057507987, + "grad_norm": 0.7452375292778015, + "learning_rate": 5e-07, + "loss": 0.0052, + "step": 1361 + }, + { + "epoch": 2.1757188498402558, + "grad_norm": 0.593784749507904, + "learning_rate": 5e-07, + "loss": 0.0041, + "step": 1362 + }, + { + "epoch": 2.1773162939297124, + "grad_norm": 0.472791850566864, + "learning_rate": 5e-07, + "loss": 0.0041, + "step": 1363 + }, + { + "epoch": 2.1789137380191694, + "grad_norm": 0.2769353687763214, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1364 + }, + { + "epoch": 2.180511182108626, + "grad_norm": 0.5703862309455872, + "learning_rate": 5e-07, + "loss": 0.0031, + "step": 1365 + }, + { + "epoch": 2.182108626198083, + "grad_norm": 0.19744770228862762, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1366 + }, + { + "epoch": 2.18370607028754, + "grad_norm": 0.7886673212051392, + "learning_rate": 5e-07, + "loss": 0.0043, + "step": 1367 + }, + { + "epoch": 2.1853035143769968, + "grad_norm": 0.13506360352039337, + "learning_rate": 5e-07, + "loss": 0.001, + "step": 1368 + }, + { + "epoch": 2.186900958466454, + "grad_norm": 0.2344270497560501, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1369 + }, + { + "epoch": 2.1884984025559104, + "grad_norm": 0.19468432664871216, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1370 + }, + { + "epoch": 2.1900958466453675, + "grad_norm": 0.1934957057237625, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1371 + }, + { + "epoch": 2.191693290734824, + "grad_norm": 0.2196214497089386, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1372 + }, + { + "epoch": 2.193290734824281, + "grad_norm": 0.13796208798885345, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 1373 + }, + { + "epoch": 2.194888178913738, + "grad_norm": 0.27561473846435547, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 1374 + }, + { + "epoch": 2.196485623003195, + "grad_norm": 0.12160798162221909, + "learning_rate": 5e-07, + "loss": 0.0013, + "step": 1375 + }, + { + "epoch": 2.198083067092652, + "grad_norm": 0.10912802815437317, + "learning_rate": 5e-07, + "loss": 0.0022, + "step": 1376 + }, + { + "epoch": 2.1996805111821085, + "grad_norm": 0.09718818962574005, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 1377 + }, + { + "epoch": 2.2012779552715656, + "grad_norm": 0.11051057279109955, + "learning_rate": 5e-07, + "loss": 0.0006, + "step": 1378 + }, + { + "epoch": 2.202875399361022, + "grad_norm": 0.12758323550224304, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1379 + }, + { + "epoch": 2.2044728434504792, + "grad_norm": 0.08249010145664215, + "learning_rate": 5e-07, + "loss": 0.0012, + "step": 1380 + }, + { + "epoch": 2.2060702875399363, + "grad_norm": 0.1392274647951126, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1381 + }, + { + "epoch": 2.207667731629393, + "grad_norm": 0.10520713776350021, + "learning_rate": 5e-07, + "loss": 0.0023, + "step": 1382 + }, + { + "epoch": 2.20926517571885, + "grad_norm": 0.10856632888317108, + "learning_rate": 5e-07, + "loss": 0.0038, + "step": 1383 + }, + { + "epoch": 2.2108626198083066, + "grad_norm": 0.08818315714597702, + "learning_rate": 5e-07, + "loss": 0.002, + "step": 1384 + }, + { + "epoch": 2.2124600638977636, + "grad_norm": 0.13421356678009033, + "learning_rate": 5e-07, + "loss": 0.0016, + "step": 1385 + }, + { + "epoch": 2.2140575079872207, + "grad_norm": 0.32427483797073364, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 1386 + }, + { + "epoch": 2.2156549520766773, + "grad_norm": 0.2926160991191864, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1387 + }, + { + "epoch": 2.2172523961661343, + "grad_norm": 0.2298290878534317, + "learning_rate": 5e-07, + "loss": 0.0032, + "step": 1388 + }, + { + "epoch": 2.218849840255591, + "grad_norm": 0.09151104092597961, + "learning_rate": 5e-07, + "loss": 0.0015, + "step": 1389 + }, + { + "epoch": 2.220447284345048, + "grad_norm": 0.10674551874399185, + "learning_rate": 5e-07, + "loss": 0.0021, + "step": 1390 + }, + { + "epoch": 2.2220447284345046, + "grad_norm": 0.2997196614742279, + "learning_rate": 5e-07, + "loss": 0.0025, + "step": 1391 + }, + { + "epoch": 2.2236421725239617, + "grad_norm": 0.11480361223220825, + "learning_rate": 5e-07, + "loss": 0.0028, + "step": 1392 + }, + { + "epoch": 2.2252396166134187, + "grad_norm": 0.07671017944812775, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1393 + }, + { + "epoch": 2.2268370607028753, + "grad_norm": 0.1539941430091858, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1394 + }, + { + "epoch": 2.2284345047923324, + "grad_norm": 0.12852609157562256, + "learning_rate": 5e-07, + "loss": 0.0008, + "step": 1395 + }, + { + "epoch": 2.230031948881789, + "grad_norm": 0.13677293062210083, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1396 + }, + { + "epoch": 2.231629392971246, + "grad_norm": 0.05409352481365204, + "learning_rate": 5e-07, + "loss": 0.0011, + "step": 1397 + }, + { + "epoch": 2.2332268370607027, + "grad_norm": 0.07923945039510727, + "learning_rate": 5e-07, + "loss": 0.0018, + "step": 1398 + }, + { + "epoch": 2.2348242811501597, + "grad_norm": 0.10353945195674896, + "learning_rate": 5e-07, + "loss": 0.0017, + "step": 1399 + }, + { + "epoch": 2.236421725239617, + "grad_norm": 0.24928894639015198, + "learning_rate": 5e-07, + "loss": 0.0019, + "step": 1400 + } + ], + "logging_steps": 1.0, + "max_steps": 3130, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 387658364567552.0, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +}