{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.0778410631950095, "eval_steps": 500, "global_step": 3687, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00027122321670735016, "grad_norm": 33.32280731201172, "learning_rate": 5.405405405405406e-09, "loss": 1.4933, "num_input_tokens_seen": 2097152, "step": 1 }, { "epoch": 0.0005424464334147003, "grad_norm": 35.11781692504883, "learning_rate": 1.0810810810810811e-08, "loss": 1.6295, "num_input_tokens_seen": 4194304, "step": 2 }, { "epoch": 0.0008136696501220504, "grad_norm": 39.06733703613281, "learning_rate": 1.6216216216216218e-08, "loss": 1.7417, "num_input_tokens_seen": 6291456, "step": 3 }, { "epoch": 0.0010848928668294006, "grad_norm": 32.58023452758789, "learning_rate": 2.1621621621621623e-08, "loss": 1.4688, "num_input_tokens_seen": 8388608, "step": 4 }, { "epoch": 0.0013561160835367507, "grad_norm": 40.15693664550781, "learning_rate": 2.7027027027027028e-08, "loss": 2.078, "num_input_tokens_seen": 10485760, "step": 5 }, { "epoch": 0.0016273393002441008, "grad_norm": 33.44337463378906, "learning_rate": 3.2432432432432436e-08, "loss": 1.6623, "num_input_tokens_seen": 12582912, "step": 6 }, { "epoch": 0.001898562516951451, "grad_norm": 31.053829193115234, "learning_rate": 3.783783783783784e-08, "loss": 1.5643, "num_input_tokens_seen": 14680064, "step": 7 }, { "epoch": 0.0021697857336588013, "grad_norm": 34.127193450927734, "learning_rate": 4.3243243243243246e-08, "loss": 1.7671, "num_input_tokens_seen": 16777216, "step": 8 }, { "epoch": 0.0024410089503661514, "grad_norm": 52.89729309082031, "learning_rate": 4.864864864864865e-08, "loss": 2.3356, "num_input_tokens_seen": 18874368, "step": 9 }, { "epoch": 0.0027122321670735015, "grad_norm": 44.45952606201172, "learning_rate": 5.4054054054054056e-08, "loss": 2.0531, "num_input_tokens_seen": 20971520, "step": 10 }, { "epoch": 0.0029834553837808516, "grad_norm": 43.71432113647461, "learning_rate": 5.945945945945946e-08, "loss": 2.2434, "num_input_tokens_seen": 23068672, "step": 11 }, { "epoch": 0.0032546786004882017, "grad_norm": 35.20856857299805, "learning_rate": 6.486486486486487e-08, "loss": 1.5967, "num_input_tokens_seen": 25165824, "step": 12 }, { "epoch": 0.003525901817195552, "grad_norm": 32.52849578857422, "learning_rate": 7.027027027027027e-08, "loss": 1.622, "num_input_tokens_seen": 27262976, "step": 13 }, { "epoch": 0.003797125033902902, "grad_norm": 42.55111312866211, "learning_rate": 7.567567567567568e-08, "loss": 2.1261, "num_input_tokens_seen": 29360128, "step": 14 }, { "epoch": 0.0040683482506102524, "grad_norm": 25.502803802490234, "learning_rate": 8.108108108108108e-08, "loss": 1.102, "num_input_tokens_seen": 31457280, "step": 15 }, { "epoch": 0.0043395714673176026, "grad_norm": 41.81550216674805, "learning_rate": 8.648648648648649e-08, "loss": 1.6793, "num_input_tokens_seen": 33554432, "step": 16 }, { "epoch": 0.004610794684024953, "grad_norm": 51.921199798583984, "learning_rate": 9.189189189189189e-08, "loss": 2.5367, "num_input_tokens_seen": 35651584, "step": 17 }, { "epoch": 0.004882017900732303, "grad_norm": 39.028221130371094, "learning_rate": 9.72972972972973e-08, "loss": 2.173, "num_input_tokens_seen": 37748736, "step": 18 }, { "epoch": 0.005153241117439653, "grad_norm": 27.318889617919922, "learning_rate": 1.0270270270270271e-07, "loss": 1.2697, "num_input_tokens_seen": 39845888, "step": 19 }, { "epoch": 0.005424464334147003, "grad_norm": 24.04189682006836, "learning_rate": 1.0810810810810811e-07, "loss": 1.1464, "num_input_tokens_seen": 41943040, "step": 20 }, { "epoch": 0.005695687550854353, "grad_norm": 29.57497787475586, "learning_rate": 1.1351351351351351e-07, "loss": 1.2522, "num_input_tokens_seen": 44040192, "step": 21 }, { "epoch": 0.005966910767561703, "grad_norm": 38.306819915771484, "learning_rate": 1.1891891891891891e-07, "loss": 2.1454, "num_input_tokens_seen": 46137344, "step": 22 }, { "epoch": 0.006238133984269053, "grad_norm": 39.06193542480469, "learning_rate": 1.2432432432432432e-07, "loss": 2.2203, "num_input_tokens_seen": 48234496, "step": 23 }, { "epoch": 0.006509357200976403, "grad_norm": 40.80998229980469, "learning_rate": 1.2972972972972974e-07, "loss": 2.4594, "num_input_tokens_seen": 50331648, "step": 24 }, { "epoch": 0.0067805804176837535, "grad_norm": 33.7674560546875, "learning_rate": 1.3513513513513515e-07, "loss": 1.7983, "num_input_tokens_seen": 52428800, "step": 25 }, { "epoch": 0.007051803634391104, "grad_norm": 27.341468811035156, "learning_rate": 1.4054054054054055e-07, "loss": 1.2909, "num_input_tokens_seen": 54525952, "step": 26 }, { "epoch": 0.007323026851098454, "grad_norm": 26.22152328491211, "learning_rate": 1.4594594594594595e-07, "loss": 1.0993, "num_input_tokens_seen": 56623104, "step": 27 }, { "epoch": 0.007594250067805804, "grad_norm": 26.978374481201172, "learning_rate": 1.5135135135135135e-07, "loss": 1.2144, "num_input_tokens_seen": 58720256, "step": 28 }, { "epoch": 0.007865473284513154, "grad_norm": 27.742353439331055, "learning_rate": 1.5675675675675675e-07, "loss": 1.3153, "num_input_tokens_seen": 60817408, "step": 29 }, { "epoch": 0.008136696501220505, "grad_norm": 36.712318420410156, "learning_rate": 1.6216216216216215e-07, "loss": 1.9966, "num_input_tokens_seen": 62914560, "step": 30 }, { "epoch": 0.008407919717927854, "grad_norm": 25.96659278869629, "learning_rate": 1.6756756756756758e-07, "loss": 1.226, "num_input_tokens_seen": 65011712, "step": 31 }, { "epoch": 0.008679142934635205, "grad_norm": 27.041015625, "learning_rate": 1.7297297297297298e-07, "loss": 1.3386, "num_input_tokens_seen": 67108864, "step": 32 }, { "epoch": 0.008950366151342554, "grad_norm": 21.146181106567383, "learning_rate": 1.7837837837837838e-07, "loss": 0.9403, "num_input_tokens_seen": 69206016, "step": 33 }, { "epoch": 0.009221589368049905, "grad_norm": 25.72652244567871, "learning_rate": 1.8378378378378379e-07, "loss": 1.4175, "num_input_tokens_seen": 71303168, "step": 34 }, { "epoch": 0.009492812584757255, "grad_norm": 42.69353485107422, "learning_rate": 1.891891891891892e-07, "loss": 2.6364, "num_input_tokens_seen": 73400320, "step": 35 }, { "epoch": 0.009764035801464606, "grad_norm": 24.534645080566406, "learning_rate": 1.945945945945946e-07, "loss": 1.4978, "num_input_tokens_seen": 75497472, "step": 36 }, { "epoch": 0.010035259018171955, "grad_norm": 22.76692771911621, "learning_rate": 2e-07, "loss": 1.2377, "num_input_tokens_seen": 77594624, "step": 37 }, { "epoch": 0.010306482234879306, "grad_norm": 22.586381912231445, "learning_rate": 2.0540540540540542e-07, "loss": 1.2021, "num_input_tokens_seen": 79691776, "step": 38 }, { "epoch": 0.010577705451586655, "grad_norm": 24.025348663330078, "learning_rate": 2.1081081081081082e-07, "loss": 1.0536, "num_input_tokens_seen": 81788928, "step": 39 }, { "epoch": 0.010848928668294006, "grad_norm": 33.128868103027344, "learning_rate": 2.1621621621621622e-07, "loss": 1.8307, "num_input_tokens_seen": 83886080, "step": 40 }, { "epoch": 0.011120151885001357, "grad_norm": 34.85871887207031, "learning_rate": 2.2162162162162162e-07, "loss": 2.1082, "num_input_tokens_seen": 85983232, "step": 41 }, { "epoch": 0.011391375101708706, "grad_norm": 22.261695861816406, "learning_rate": 2.2702702702702703e-07, "loss": 0.9698, "num_input_tokens_seen": 88080384, "step": 42 }, { "epoch": 0.011662598318416057, "grad_norm": 28.77890396118164, "learning_rate": 2.3243243243243243e-07, "loss": 1.2482, "num_input_tokens_seen": 90177536, "step": 43 }, { "epoch": 0.011933821535123406, "grad_norm": 22.587844848632812, "learning_rate": 2.3783783783783783e-07, "loss": 1.1466, "num_input_tokens_seen": 92274688, "step": 44 }, { "epoch": 0.012205044751830757, "grad_norm": 18.199480056762695, "learning_rate": 2.4324324324324326e-07, "loss": 0.9606, "num_input_tokens_seen": 94371840, "step": 45 }, { "epoch": 0.012476267968538107, "grad_norm": 27.911603927612305, "learning_rate": 2.4864864864864863e-07, "loss": 1.6163, "num_input_tokens_seen": 96468992, "step": 46 }, { "epoch": 0.012747491185245458, "grad_norm": 34.57044982910156, "learning_rate": 2.5405405405405406e-07, "loss": 1.4498, "num_input_tokens_seen": 98566144, "step": 47 }, { "epoch": 0.013018714401952807, "grad_norm": 21.867979049682617, "learning_rate": 2.594594594594595e-07, "loss": 1.3337, "num_input_tokens_seen": 100663296, "step": 48 }, { "epoch": 0.013289937618660158, "grad_norm": 19.342687606811523, "learning_rate": 2.6486486486486486e-07, "loss": 1.0142, "num_input_tokens_seen": 102760448, "step": 49 }, { "epoch": 0.013561160835367507, "grad_norm": 24.752653121948242, "learning_rate": 2.702702702702703e-07, "loss": 1.0282, "num_input_tokens_seen": 104857600, "step": 50 }, { "epoch": 0.013832384052074858, "grad_norm": 20.70758628845215, "learning_rate": 2.7567567567567567e-07, "loss": 0.8941, "num_input_tokens_seen": 106954752, "step": 51 }, { "epoch": 0.014103607268782207, "grad_norm": 22.01601219177246, "learning_rate": 2.810810810810811e-07, "loss": 0.8012, "num_input_tokens_seen": 109051904, "step": 52 }, { "epoch": 0.014374830485489558, "grad_norm": 34.209354400634766, "learning_rate": 2.8648648648648647e-07, "loss": 2.5913, "num_input_tokens_seen": 111149056, "step": 53 }, { "epoch": 0.014646053702196907, "grad_norm": 19.16649627685547, "learning_rate": 2.918918918918919e-07, "loss": 0.9785, "num_input_tokens_seen": 113246208, "step": 54 }, { "epoch": 0.014917276918904258, "grad_norm": 25.85431480407715, "learning_rate": 2.972972972972973e-07, "loss": 1.7306, "num_input_tokens_seen": 115343360, "step": 55 }, { "epoch": 0.015188500135611608, "grad_norm": 19.06846809387207, "learning_rate": 3.027027027027027e-07, "loss": 0.9716, "num_input_tokens_seen": 117440512, "step": 56 }, { "epoch": 0.015459723352318959, "grad_norm": 23.541532516479492, "learning_rate": 3.0810810810810813e-07, "loss": 1.4204, "num_input_tokens_seen": 119537664, "step": 57 }, { "epoch": 0.015730946569026308, "grad_norm": 16.523822784423828, "learning_rate": 3.135135135135135e-07, "loss": 0.7799, "num_input_tokens_seen": 121634816, "step": 58 }, { "epoch": 0.01600216978573366, "grad_norm": 21.27098274230957, "learning_rate": 3.1891891891891893e-07, "loss": 0.9598, "num_input_tokens_seen": 123731968, "step": 59 }, { "epoch": 0.01627339300244101, "grad_norm": 22.250896453857422, "learning_rate": 3.243243243243243e-07, "loss": 0.847, "num_input_tokens_seen": 125829120, "step": 60 }, { "epoch": 0.01654461621914836, "grad_norm": 37.0035514831543, "learning_rate": 3.2972972972972973e-07, "loss": 1.2704, "num_input_tokens_seen": 127926272, "step": 61 }, { "epoch": 0.016815839435855708, "grad_norm": 20.06436538696289, "learning_rate": 3.3513513513513516e-07, "loss": 0.9768, "num_input_tokens_seen": 130023424, "step": 62 }, { "epoch": 0.01708706265256306, "grad_norm": 20.020830154418945, "learning_rate": 3.4054054054054054e-07, "loss": 1.1647, "num_input_tokens_seen": 132120576, "step": 63 }, { "epoch": 0.01735828586927041, "grad_norm": 19.906391143798828, "learning_rate": 3.4594594594594597e-07, "loss": 0.8014, "num_input_tokens_seen": 134217728, "step": 64 }, { "epoch": 0.01762950908597776, "grad_norm": 18.05296516418457, "learning_rate": 3.5135135135135134e-07, "loss": 0.9887, "num_input_tokens_seen": 136314880, "step": 65 }, { "epoch": 0.01790073230268511, "grad_norm": 18.5097599029541, "learning_rate": 3.5675675675675677e-07, "loss": 0.9417, "num_input_tokens_seen": 138412032, "step": 66 }, { "epoch": 0.01817195551939246, "grad_norm": 21.897937774658203, "learning_rate": 3.6216216216216214e-07, "loss": 1.2979, "num_input_tokens_seen": 140509184, "step": 67 }, { "epoch": 0.01844317873609981, "grad_norm": 23.45905113220215, "learning_rate": 3.6756756756756757e-07, "loss": 1.2426, "num_input_tokens_seen": 142606336, "step": 68 }, { "epoch": 0.01871440195280716, "grad_norm": 21.97277069091797, "learning_rate": 3.72972972972973e-07, "loss": 1.4745, "num_input_tokens_seen": 144703488, "step": 69 }, { "epoch": 0.01898562516951451, "grad_norm": 24.767820358276367, "learning_rate": 3.783783783783784e-07, "loss": 0.6843, "num_input_tokens_seen": 146800640, "step": 70 }, { "epoch": 0.01925684838622186, "grad_norm": 20.10371971130371, "learning_rate": 3.837837837837838e-07, "loss": 1.1348, "num_input_tokens_seen": 148897792, "step": 71 }, { "epoch": 0.01952807160292921, "grad_norm": 24.476518630981445, "learning_rate": 3.891891891891892e-07, "loss": 1.7115, "num_input_tokens_seen": 150994944, "step": 72 }, { "epoch": 0.019799294819636562, "grad_norm": 19.302143096923828, "learning_rate": 3.945945945945946e-07, "loss": 0.7886, "num_input_tokens_seen": 153092096, "step": 73 }, { "epoch": 0.02007051803634391, "grad_norm": 14.634087562561035, "learning_rate": 4e-07, "loss": 0.5778, "num_input_tokens_seen": 155189248, "step": 74 }, { "epoch": 0.02034174125305126, "grad_norm": 14.836756706237793, "learning_rate": 4.054054054054054e-07, "loss": 0.6952, "num_input_tokens_seen": 157286400, "step": 75 }, { "epoch": 0.02061296446975861, "grad_norm": 15.128747940063477, "learning_rate": 4.1081081081081084e-07, "loss": 0.7274, "num_input_tokens_seen": 159383552, "step": 76 }, { "epoch": 0.020884187686465962, "grad_norm": 14.638301849365234, "learning_rate": 4.162162162162162e-07, "loss": 0.634, "num_input_tokens_seen": 161480704, "step": 77 }, { "epoch": 0.02115541090317331, "grad_norm": 23.832578659057617, "learning_rate": 4.2162162162162164e-07, "loss": 1.0147, "num_input_tokens_seen": 163577856, "step": 78 }, { "epoch": 0.02142663411988066, "grad_norm": 16.2668514251709, "learning_rate": 4.27027027027027e-07, "loss": 0.7479, "num_input_tokens_seen": 165675008, "step": 79 }, { "epoch": 0.021697857336588012, "grad_norm": 14.96450424194336, "learning_rate": 4.3243243243243244e-07, "loss": 0.6993, "num_input_tokens_seen": 167772160, "step": 80 }, { "epoch": 0.021969080553295363, "grad_norm": 19.316436767578125, "learning_rate": 4.378378378378378e-07, "loss": 0.6489, "num_input_tokens_seen": 169869312, "step": 81 }, { "epoch": 0.022240303770002714, "grad_norm": 14.418977737426758, "learning_rate": 4.4324324324324325e-07, "loss": 0.592, "num_input_tokens_seen": 171966464, "step": 82 }, { "epoch": 0.02251152698671006, "grad_norm": 19.933923721313477, "learning_rate": 4.486486486486487e-07, "loss": 0.7344, "num_input_tokens_seen": 174063616, "step": 83 }, { "epoch": 0.022782750203417412, "grad_norm": 20.08726692199707, "learning_rate": 4.5405405405405405e-07, "loss": 0.916, "num_input_tokens_seen": 176160768, "step": 84 }, { "epoch": 0.023053973420124763, "grad_norm": 18.082517623901367, "learning_rate": 4.594594594594595e-07, "loss": 0.9287, "num_input_tokens_seen": 178257920, "step": 85 }, { "epoch": 0.023325196636832114, "grad_norm": 15.847949981689453, "learning_rate": 4.6486486486486485e-07, "loss": 0.8382, "num_input_tokens_seen": 180355072, "step": 86 }, { "epoch": 0.023596419853539462, "grad_norm": 22.586856842041016, "learning_rate": 4.702702702702703e-07, "loss": 1.0967, "num_input_tokens_seen": 182452224, "step": 87 }, { "epoch": 0.023867643070246813, "grad_norm": 16.663034439086914, "learning_rate": 4.7567567567567566e-07, "loss": 0.675, "num_input_tokens_seen": 184549376, "step": 88 }, { "epoch": 0.024138866286954164, "grad_norm": 19.878616333007812, "learning_rate": 4.810810810810811e-07, "loss": 1.1732, "num_input_tokens_seen": 186646528, "step": 89 }, { "epoch": 0.024410089503661515, "grad_norm": 16.984296798706055, "learning_rate": 4.864864864864865e-07, "loss": 0.8503, "num_input_tokens_seen": 188743680, "step": 90 }, { "epoch": 0.024681312720368862, "grad_norm": 14.563621520996094, "learning_rate": 4.918918918918919e-07, "loss": 0.4946, "num_input_tokens_seen": 190840832, "step": 91 }, { "epoch": 0.024952535937076213, "grad_norm": 15.319222450256348, "learning_rate": 4.972972972972973e-07, "loss": 0.7149, "num_input_tokens_seen": 192937984, "step": 92 }, { "epoch": 0.025223759153783564, "grad_norm": 14.346334457397461, "learning_rate": 5.027027027027027e-07, "loss": 0.3925, "num_input_tokens_seen": 195035136, "step": 93 }, { "epoch": 0.025494982370490915, "grad_norm": 19.126708984375, "learning_rate": 5.081081081081081e-07, "loss": 0.9146, "num_input_tokens_seen": 197132288, "step": 94 }, { "epoch": 0.025766205587198263, "grad_norm": 18.894250869750977, "learning_rate": 5.135135135135134e-07, "loss": 0.8788, "num_input_tokens_seen": 199229440, "step": 95 }, { "epoch": 0.026037428803905614, "grad_norm": 19.886512756347656, "learning_rate": 5.18918918918919e-07, "loss": 0.9447, "num_input_tokens_seen": 201326592, "step": 96 }, { "epoch": 0.026308652020612965, "grad_norm": 16.53607940673828, "learning_rate": 5.243243243243243e-07, "loss": 0.7017, "num_input_tokens_seen": 203423744, "step": 97 }, { "epoch": 0.026579875237320316, "grad_norm": 21.681087493896484, "learning_rate": 5.297297297297297e-07, "loss": 1.0761, "num_input_tokens_seen": 205520896, "step": 98 }, { "epoch": 0.026851098454027666, "grad_norm": 28.16925048828125, "learning_rate": 5.35135135135135e-07, "loss": 1.9542, "num_input_tokens_seen": 207618048, "step": 99 }, { "epoch": 0.027122321670735014, "grad_norm": 16.13709831237793, "learning_rate": 5.405405405405406e-07, "loss": 0.5452, "num_input_tokens_seen": 209715200, "step": 100 }, { "epoch": 0.027393544887442365, "grad_norm": 12.437922477722168, "learning_rate": 5.459459459459459e-07, "loss": 0.4683, "num_input_tokens_seen": 211812352, "step": 101 }, { "epoch": 0.027664768104149716, "grad_norm": 15.048489570617676, "learning_rate": 5.513513513513513e-07, "loss": 0.6694, "num_input_tokens_seen": 213909504, "step": 102 }, { "epoch": 0.027935991320857067, "grad_norm": 20.845455169677734, "learning_rate": 5.567567567567567e-07, "loss": 0.9275, "num_input_tokens_seen": 216006656, "step": 103 }, { "epoch": 0.028207214537564414, "grad_norm": 15.5338716506958, "learning_rate": 5.621621621621622e-07, "loss": 0.6088, "num_input_tokens_seen": 218103808, "step": 104 }, { "epoch": 0.028478437754271765, "grad_norm": 12.675715446472168, "learning_rate": 5.675675675675675e-07, "loss": 0.6012, "num_input_tokens_seen": 220200960, "step": 105 }, { "epoch": 0.028749660970979116, "grad_norm": 16.053550720214844, "learning_rate": 5.729729729729729e-07, "loss": 0.5751, "num_input_tokens_seen": 222298112, "step": 106 }, { "epoch": 0.029020884187686467, "grad_norm": 18.570253372192383, "learning_rate": 5.783783783783784e-07, "loss": 0.8626, "num_input_tokens_seen": 224395264, "step": 107 }, { "epoch": 0.029292107404393815, "grad_norm": 15.664776802062988, "learning_rate": 5.837837837837838e-07, "loss": 0.6016, "num_input_tokens_seen": 226492416, "step": 108 }, { "epoch": 0.029563330621101166, "grad_norm": 17.67211151123047, "learning_rate": 5.891891891891891e-07, "loss": 0.8534, "num_input_tokens_seen": 228589568, "step": 109 }, { "epoch": 0.029834553837808517, "grad_norm": 23.127309799194336, "learning_rate": 5.945945945945947e-07, "loss": 1.1914, "num_input_tokens_seen": 230686720, "step": 110 }, { "epoch": 0.030105777054515868, "grad_norm": 18.636947631835938, "learning_rate": 6e-07, "loss": 0.9078, "num_input_tokens_seen": 232783872, "step": 111 }, { "epoch": 0.030377000271223215, "grad_norm": 20.506677627563477, "learning_rate": 6.054054054054054e-07, "loss": 1.1244, "num_input_tokens_seen": 234881024, "step": 112 }, { "epoch": 0.030648223487930566, "grad_norm": 15.10786247253418, "learning_rate": 6.108108108108107e-07, "loss": 0.5326, "num_input_tokens_seen": 236978176, "step": 113 }, { "epoch": 0.030919446704637917, "grad_norm": 14.648728370666504, "learning_rate": 6.162162162162163e-07, "loss": 0.468, "num_input_tokens_seen": 239075328, "step": 114 }, { "epoch": 0.031190669921345268, "grad_norm": 16.831241607666016, "learning_rate": 6.216216216216216e-07, "loss": 0.6454, "num_input_tokens_seen": 241172480, "step": 115 }, { "epoch": 0.031461893138052616, "grad_norm": 13.11506175994873, "learning_rate": 6.27027027027027e-07, "loss": 0.4922, "num_input_tokens_seen": 243269632, "step": 116 }, { "epoch": 0.03173311635475997, "grad_norm": 18.388965606689453, "learning_rate": 6.324324324324324e-07, "loss": 0.6721, "num_input_tokens_seen": 245366784, "step": 117 }, { "epoch": 0.03200433957146732, "grad_norm": 19.308448791503906, "learning_rate": 6.378378378378379e-07, "loss": 0.8231, "num_input_tokens_seen": 247463936, "step": 118 }, { "epoch": 0.032275562788174665, "grad_norm": 19.583574295043945, "learning_rate": 6.432432432432432e-07, "loss": 1.0588, "num_input_tokens_seen": 249561088, "step": 119 }, { "epoch": 0.03254678600488202, "grad_norm": 15.67451000213623, "learning_rate": 6.486486486486486e-07, "loss": 0.5752, "num_input_tokens_seen": 251658240, "step": 120 }, { "epoch": 0.03281800922158937, "grad_norm": 18.46150779724121, "learning_rate": 6.54054054054054e-07, "loss": 0.8636, "num_input_tokens_seen": 253755392, "step": 121 }, { "epoch": 0.03308923243829672, "grad_norm": 18.50663948059082, "learning_rate": 6.594594594594595e-07, "loss": 0.9201, "num_input_tokens_seen": 255852544, "step": 122 }, { "epoch": 0.03336045565500407, "grad_norm": 18.124380111694336, "learning_rate": 6.648648648648648e-07, "loss": 0.7413, "num_input_tokens_seen": 257949696, "step": 123 }, { "epoch": 0.033631678871711417, "grad_norm": 20.973876953125, "learning_rate": 6.702702702702703e-07, "loss": 0.9747, "num_input_tokens_seen": 260046848, "step": 124 }, { "epoch": 0.03390290208841877, "grad_norm": 15.687307357788086, "learning_rate": 6.756756756756756e-07, "loss": 0.487, "num_input_tokens_seen": 262144000, "step": 125 }, { "epoch": 0.03417412530512612, "grad_norm": 17.15902328491211, "learning_rate": 6.810810810810811e-07, "loss": 0.8454, "num_input_tokens_seen": 264241152, "step": 126 }, { "epoch": 0.034445348521833466, "grad_norm": 14.142715454101562, "learning_rate": 6.864864864864864e-07, "loss": 0.5951, "num_input_tokens_seen": 266338304, "step": 127 }, { "epoch": 0.03471657173854082, "grad_norm": 20.538616180419922, "learning_rate": 6.918918918918919e-07, "loss": 0.9529, "num_input_tokens_seen": 268435456, "step": 128 }, { "epoch": 0.03498779495524817, "grad_norm": 22.79062843322754, "learning_rate": 6.972972972972973e-07, "loss": 1.0202, "num_input_tokens_seen": 270532608, "step": 129 }, { "epoch": 0.03525901817195552, "grad_norm": 16.5878963470459, "learning_rate": 7.027027027027027e-07, "loss": 0.7667, "num_input_tokens_seen": 272629760, "step": 130 }, { "epoch": 0.03553024138866287, "grad_norm": 16.780086517333984, "learning_rate": 7.081081081081081e-07, "loss": 0.6279, "num_input_tokens_seen": 274726912, "step": 131 }, { "epoch": 0.03580146460537022, "grad_norm": 23.6799259185791, "learning_rate": 7.135135135135135e-07, "loss": 1.2914, "num_input_tokens_seen": 276824064, "step": 132 }, { "epoch": 0.03607268782207757, "grad_norm": 17.561779022216797, "learning_rate": 7.189189189189189e-07, "loss": 0.5839, "num_input_tokens_seen": 278921216, "step": 133 }, { "epoch": 0.03634391103878492, "grad_norm": 21.706342697143555, "learning_rate": 7.243243243243243e-07, "loss": 0.7747, "num_input_tokens_seen": 281018368, "step": 134 }, { "epoch": 0.03661513425549227, "grad_norm": 25.93421173095703, "learning_rate": 7.297297297297297e-07, "loss": 1.2564, "num_input_tokens_seen": 283115520, "step": 135 }, { "epoch": 0.03688635747219962, "grad_norm": 26.61385154724121, "learning_rate": 7.351351351351351e-07, "loss": 0.4117, "num_input_tokens_seen": 285212672, "step": 136 }, { "epoch": 0.03715758068890697, "grad_norm": 19.071443557739258, "learning_rate": 7.405405405405405e-07, "loss": 0.6143, "num_input_tokens_seen": 287309824, "step": 137 }, { "epoch": 0.03742880390561432, "grad_norm": 15.275123596191406, "learning_rate": 7.45945945945946e-07, "loss": 0.6617, "num_input_tokens_seen": 289406976, "step": 138 }, { "epoch": 0.03770002712232167, "grad_norm": 13.6419095993042, "learning_rate": 7.513513513513513e-07, "loss": 0.4291, "num_input_tokens_seen": 291504128, "step": 139 }, { "epoch": 0.03797125033902902, "grad_norm": 14.96066951751709, "learning_rate": 7.567567567567568e-07, "loss": 0.5393, "num_input_tokens_seen": 293601280, "step": 140 }, { "epoch": 0.03824247355573637, "grad_norm": 12.42337703704834, "learning_rate": 7.621621621621621e-07, "loss": 0.3799, "num_input_tokens_seen": 295698432, "step": 141 }, { "epoch": 0.03851369677244372, "grad_norm": 13.157722473144531, "learning_rate": 7.675675675675676e-07, "loss": 0.3632, "num_input_tokens_seen": 297795584, "step": 142 }, { "epoch": 0.038784919989151075, "grad_norm": 16.27894401550293, "learning_rate": 7.729729729729729e-07, "loss": 0.7346, "num_input_tokens_seen": 299892736, "step": 143 }, { "epoch": 0.03905614320585842, "grad_norm": 20.3560848236084, "learning_rate": 7.783783783783784e-07, "loss": 0.7771, "num_input_tokens_seen": 301989888, "step": 144 }, { "epoch": 0.03932736642256577, "grad_norm": 21.640209197998047, "learning_rate": 7.837837837837838e-07, "loss": 1.1799, "num_input_tokens_seen": 304087040, "step": 145 }, { "epoch": 0.039598589639273124, "grad_norm": 14.982572555541992, "learning_rate": 7.891891891891892e-07, "loss": 0.5518, "num_input_tokens_seen": 306184192, "step": 146 }, { "epoch": 0.03986981285598047, "grad_norm": 18.2089900970459, "learning_rate": 7.945945945945945e-07, "loss": 0.624, "num_input_tokens_seen": 308281344, "step": 147 }, { "epoch": 0.04014103607268782, "grad_norm": 19.922109603881836, "learning_rate": 8e-07, "loss": 0.8285, "num_input_tokens_seen": 310378496, "step": 148 }, { "epoch": 0.040412259289395173, "grad_norm": 15.228759765625, "learning_rate": 8.054054054054054e-07, "loss": 0.5316, "num_input_tokens_seen": 312475648, "step": 149 }, { "epoch": 0.04068348250610252, "grad_norm": 17.04909324645996, "learning_rate": 8.108108108108108e-07, "loss": 0.6752, "num_input_tokens_seen": 314572800, "step": 150 }, { "epoch": 0.040954705722809875, "grad_norm": 18.965408325195312, "learning_rate": 8.162162162162161e-07, "loss": 0.8268, "num_input_tokens_seen": 316669952, "step": 151 }, { "epoch": 0.04122592893951722, "grad_norm": 18.962045669555664, "learning_rate": 8.216216216216217e-07, "loss": 0.747, "num_input_tokens_seen": 318767104, "step": 152 }, { "epoch": 0.04149715215622457, "grad_norm": 19.22246551513672, "learning_rate": 8.27027027027027e-07, "loss": 0.6027, "num_input_tokens_seen": 320864256, "step": 153 }, { "epoch": 0.041768375372931925, "grad_norm": 15.074151039123535, "learning_rate": 8.324324324324324e-07, "loss": 0.5226, "num_input_tokens_seen": 322961408, "step": 154 }, { "epoch": 0.04203959858963927, "grad_norm": 21.02098274230957, "learning_rate": 8.378378378378377e-07, "loss": 0.9997, "num_input_tokens_seen": 325058560, "step": 155 }, { "epoch": 0.04231082180634662, "grad_norm": 18.864036560058594, "learning_rate": 8.432432432432433e-07, "loss": 0.7729, "num_input_tokens_seen": 327155712, "step": 156 }, { "epoch": 0.042582045023053974, "grad_norm": 18.284622192382812, "learning_rate": 8.486486486486486e-07, "loss": 0.7608, "num_input_tokens_seen": 329252864, "step": 157 }, { "epoch": 0.04285326823976132, "grad_norm": 16.468151092529297, "learning_rate": 8.54054054054054e-07, "loss": 0.7438, "num_input_tokens_seen": 331350016, "step": 158 }, { "epoch": 0.043124491456468676, "grad_norm": 15.429784774780273, "learning_rate": 8.594594594594595e-07, "loss": 0.6437, "num_input_tokens_seen": 333447168, "step": 159 }, { "epoch": 0.043395714673176024, "grad_norm": 24.898374557495117, "learning_rate": 8.648648648648649e-07, "loss": 1.0021, "num_input_tokens_seen": 335544320, "step": 160 }, { "epoch": 0.04366693788988337, "grad_norm": 15.75402545928955, "learning_rate": 8.702702702702702e-07, "loss": 0.5328, "num_input_tokens_seen": 337641472, "step": 161 }, { "epoch": 0.043938161106590726, "grad_norm": 21.882734298706055, "learning_rate": 8.756756756756756e-07, "loss": 0.9156, "num_input_tokens_seen": 339738624, "step": 162 }, { "epoch": 0.04420938432329807, "grad_norm": 21.429712295532227, "learning_rate": 8.810810810810811e-07, "loss": 1.2029, "num_input_tokens_seen": 341835776, "step": 163 }, { "epoch": 0.04448060754000543, "grad_norm": 12.5687837600708, "learning_rate": 8.864864864864865e-07, "loss": 0.3537, "num_input_tokens_seen": 343932928, "step": 164 }, { "epoch": 0.044751830756712775, "grad_norm": 19.376272201538086, "learning_rate": 8.918918918918918e-07, "loss": 0.646, "num_input_tokens_seen": 346030080, "step": 165 }, { "epoch": 0.04502305397342012, "grad_norm": 17.1505184173584, "learning_rate": 8.972972972972974e-07, "loss": 0.6591, "num_input_tokens_seen": 348127232, "step": 166 }, { "epoch": 0.04529427719012748, "grad_norm": 15.413045883178711, "learning_rate": 9.027027027027027e-07, "loss": 0.4214, "num_input_tokens_seen": 350224384, "step": 167 }, { "epoch": 0.045565500406834825, "grad_norm": 17.270416259765625, "learning_rate": 9.081081081081081e-07, "loss": 0.6298, "num_input_tokens_seen": 352321536, "step": 168 }, { "epoch": 0.04583672362354217, "grad_norm": 14.51281452178955, "learning_rate": 9.135135135135134e-07, "loss": 0.5393, "num_input_tokens_seen": 354418688, "step": 169 }, { "epoch": 0.04610794684024953, "grad_norm": 18.907087326049805, "learning_rate": 9.18918918918919e-07, "loss": 0.6491, "num_input_tokens_seen": 356515840, "step": 170 }, { "epoch": 0.046379170056956874, "grad_norm": 17.48542594909668, "learning_rate": 9.243243243243243e-07, "loss": 0.5932, "num_input_tokens_seen": 358612992, "step": 171 }, { "epoch": 0.04665039327366423, "grad_norm": 15.82155704498291, "learning_rate": 9.297297297297297e-07, "loss": 0.5997, "num_input_tokens_seen": 360710144, "step": 172 }, { "epoch": 0.046921616490371576, "grad_norm": 54.783302307128906, "learning_rate": 9.351351351351351e-07, "loss": 0.5147, "num_input_tokens_seen": 362807296, "step": 173 }, { "epoch": 0.047192839707078924, "grad_norm": 15.447076797485352, "learning_rate": 9.405405405405406e-07, "loss": 0.5626, "num_input_tokens_seen": 364904448, "step": 174 }, { "epoch": 0.04746406292378628, "grad_norm": 21.660140991210938, "learning_rate": 9.459459459459459e-07, "loss": 0.5531, "num_input_tokens_seen": 367001600, "step": 175 }, { "epoch": 0.047735286140493625, "grad_norm": 19.583106994628906, "learning_rate": 9.513513513513513e-07, "loss": 0.7348, "num_input_tokens_seen": 369098752, "step": 176 }, { "epoch": 0.04800650935720097, "grad_norm": 19.693716049194336, "learning_rate": 9.567567567567567e-07, "loss": 0.7629, "num_input_tokens_seen": 371195904, "step": 177 }, { "epoch": 0.04827773257390833, "grad_norm": 21.72786521911621, "learning_rate": 9.621621621621622e-07, "loss": 0.7271, "num_input_tokens_seen": 373293056, "step": 178 }, { "epoch": 0.048548955790615675, "grad_norm": 20.64727210998535, "learning_rate": 9.675675675675676e-07, "loss": 0.944, "num_input_tokens_seen": 375390208, "step": 179 }, { "epoch": 0.04882017900732303, "grad_norm": 15.752840042114258, "learning_rate": 9.72972972972973e-07, "loss": 0.4082, "num_input_tokens_seen": 377487360, "step": 180 }, { "epoch": 0.04909140222403038, "grad_norm": 16.297456741333008, "learning_rate": 9.783783783783782e-07, "loss": 0.5716, "num_input_tokens_seen": 379584512, "step": 181 }, { "epoch": 0.049362625440737724, "grad_norm": 16.411787033081055, "learning_rate": 9.837837837837839e-07, "loss": 0.3719, "num_input_tokens_seen": 381681664, "step": 182 }, { "epoch": 0.04963384865744508, "grad_norm": 17.700828552246094, "learning_rate": 9.89189189189189e-07, "loss": 0.6982, "num_input_tokens_seen": 383778816, "step": 183 }, { "epoch": 0.049905071874152426, "grad_norm": 13.723708152770996, "learning_rate": 9.945945945945945e-07, "loss": 0.4343, "num_input_tokens_seen": 385875968, "step": 184 }, { "epoch": 0.05017629509085978, "grad_norm": 18.31014633178711, "learning_rate": 1e-06, "loss": 0.511, "num_input_tokens_seen": 387973120, "step": 185 }, { "epoch": 0.05044751830756713, "grad_norm": 16.061124801635742, "learning_rate": 9.99999818928562e-07, "loss": 0.5106, "num_input_tokens_seen": 390070272, "step": 186 }, { "epoch": 0.050718741524274476, "grad_norm": 17.339305877685547, "learning_rate": 9.999992757143933e-07, "loss": 0.5723, "num_input_tokens_seen": 392167424, "step": 187 }, { "epoch": 0.05098996474098183, "grad_norm": 15.1113920211792, "learning_rate": 9.999983703579313e-07, "loss": 0.5508, "num_input_tokens_seen": 394264576, "step": 188 }, { "epoch": 0.05126118795768918, "grad_norm": 18.564332962036133, "learning_rate": 9.999971028599045e-07, "loss": 0.7198, "num_input_tokens_seen": 396361728, "step": 189 }, { "epoch": 0.051532411174396525, "grad_norm": 14.539531707763672, "learning_rate": 9.99995473221333e-07, "loss": 0.4645, "num_input_tokens_seen": 398458880, "step": 190 }, { "epoch": 0.05180363439110388, "grad_norm": 16.04762077331543, "learning_rate": 9.999934814435284e-07, "loss": 0.6121, "num_input_tokens_seen": 400556032, "step": 191 }, { "epoch": 0.05207485760781123, "grad_norm": 17.753110885620117, "learning_rate": 9.999911275280933e-07, "loss": 0.6252, "num_input_tokens_seen": 402653184, "step": 192 }, { "epoch": 0.05234608082451858, "grad_norm": 23.137664794921875, "learning_rate": 9.999884114769223e-07, "loss": 1.1105, "num_input_tokens_seen": 404750336, "step": 193 }, { "epoch": 0.05261730404122593, "grad_norm": 17.16234588623047, "learning_rate": 9.99985333292201e-07, "loss": 0.693, "num_input_tokens_seen": 406847488, "step": 194 }, { "epoch": 0.05288852725793328, "grad_norm": 25.552888870239258, "learning_rate": 9.999818929764068e-07, "loss": 1.1388, "num_input_tokens_seen": 408944640, "step": 195 }, { "epoch": 0.05315975047464063, "grad_norm": 17.101301193237305, "learning_rate": 9.99978090532308e-07, "loss": 0.5531, "num_input_tokens_seen": 411041792, "step": 196 }, { "epoch": 0.05343097369134798, "grad_norm": 16.60688591003418, "learning_rate": 9.99973925962965e-07, "loss": 0.7798, "num_input_tokens_seen": 413138944, "step": 197 }, { "epoch": 0.05370219690805533, "grad_norm": 12.576007843017578, "learning_rate": 9.999693992717292e-07, "loss": 0.4118, "num_input_tokens_seen": 415236096, "step": 198 }, { "epoch": 0.05397342012476268, "grad_norm": 12.9966402053833, "learning_rate": 9.999645104622434e-07, "loss": 0.4088, "num_input_tokens_seen": 417333248, "step": 199 }, { "epoch": 0.05424464334147003, "grad_norm": 16.39759635925293, "learning_rate": 9.99959259538442e-07, "loss": 0.5526, "num_input_tokens_seen": 419430400, "step": 200 }, { "epoch": 0.05451586655817738, "grad_norm": 14.523542404174805, "learning_rate": 9.99953646504551e-07, "loss": 0.4645, "num_input_tokens_seen": 421527552, "step": 201 }, { "epoch": 0.05478708977488473, "grad_norm": 17.604217529296875, "learning_rate": 9.99947671365087e-07, "loss": 0.794, "num_input_tokens_seen": 423624704, "step": 202 }, { "epoch": 0.05505831299159208, "grad_norm": 16.45903778076172, "learning_rate": 9.99941334124859e-07, "loss": 0.6843, "num_input_tokens_seen": 425721856, "step": 203 }, { "epoch": 0.05532953620829943, "grad_norm": 17.023611068725586, "learning_rate": 9.999346347889667e-07, "loss": 0.6436, "num_input_tokens_seen": 427819008, "step": 204 }, { "epoch": 0.05560075942500678, "grad_norm": 17.66231346130371, "learning_rate": 9.999275733628017e-07, "loss": 0.707, "num_input_tokens_seen": 429916160, "step": 205 }, { "epoch": 0.055871982641714134, "grad_norm": 14.247184753417969, "learning_rate": 9.999201498520466e-07, "loss": 0.5772, "num_input_tokens_seen": 432013312, "step": 206 }, { "epoch": 0.05614320585842148, "grad_norm": 16.290454864501953, "learning_rate": 9.999123642626758e-07, "loss": 0.5432, "num_input_tokens_seen": 434110464, "step": 207 }, { "epoch": 0.05641442907512883, "grad_norm": 16.772056579589844, "learning_rate": 9.999042166009544e-07, "loss": 0.5976, "num_input_tokens_seen": 436207616, "step": 208 }, { "epoch": 0.05668565229183618, "grad_norm": 10.778910636901855, "learning_rate": 9.998957068734399e-07, "loss": 0.3159, "num_input_tokens_seen": 438304768, "step": 209 }, { "epoch": 0.05695687550854353, "grad_norm": 21.58819580078125, "learning_rate": 9.9988683508698e-07, "loss": 0.9222, "num_input_tokens_seen": 440401920, "step": 210 }, { "epoch": 0.05722809872525088, "grad_norm": 21.694862365722656, "learning_rate": 9.99877601248715e-07, "loss": 1.0365, "num_input_tokens_seen": 442499072, "step": 211 }, { "epoch": 0.05749932194195823, "grad_norm": 20.887956619262695, "learning_rate": 9.998680053660756e-07, "loss": 0.5279, "num_input_tokens_seen": 444596224, "step": 212 }, { "epoch": 0.05777054515866558, "grad_norm": 20.843788146972656, "learning_rate": 9.998580474467842e-07, "loss": 0.7565, "num_input_tokens_seen": 446693376, "step": 213 }, { "epoch": 0.058041768375372935, "grad_norm": 13.617545127868652, "learning_rate": 9.998477274988545e-07, "loss": 0.4137, "num_input_tokens_seen": 448790528, "step": 214 }, { "epoch": 0.05831299159208028, "grad_norm": 11.583077430725098, "learning_rate": 9.998370455305918e-07, "loss": 0.4458, "num_input_tokens_seen": 450887680, "step": 215 }, { "epoch": 0.05858421480878763, "grad_norm": 16.68201446533203, "learning_rate": 9.998260015505923e-07, "loss": 0.6287, "num_input_tokens_seen": 452984832, "step": 216 }, { "epoch": 0.058855438025494984, "grad_norm": 13.695748329162598, "learning_rate": 9.998145955677438e-07, "loss": 0.4641, "num_input_tokens_seen": 455081984, "step": 217 }, { "epoch": 0.05912666124220233, "grad_norm": 12.062211990356445, "learning_rate": 9.998028275912257e-07, "loss": 0.3928, "num_input_tokens_seen": 457179136, "step": 218 }, { "epoch": 0.059397884458909686, "grad_norm": 19.77566146850586, "learning_rate": 9.997906976305082e-07, "loss": 0.5814, "num_input_tokens_seen": 459276288, "step": 219 }, { "epoch": 0.059669107675617034, "grad_norm": 19.376760482788086, "learning_rate": 9.99778205695353e-07, "loss": 0.8212, "num_input_tokens_seen": 461373440, "step": 220 }, { "epoch": 0.05994033089232438, "grad_norm": 13.375431060791016, "learning_rate": 9.997653517958132e-07, "loss": 0.4862, "num_input_tokens_seen": 463470592, "step": 221 }, { "epoch": 0.060211554109031735, "grad_norm": 16.914608001708984, "learning_rate": 9.997521359422332e-07, "loss": 0.6401, "num_input_tokens_seen": 465567744, "step": 222 }, { "epoch": 0.06048277732573908, "grad_norm": 13.095565795898438, "learning_rate": 9.997385581452484e-07, "loss": 0.4319, "num_input_tokens_seen": 467664896, "step": 223 }, { "epoch": 0.06075400054244643, "grad_norm": 15.450024604797363, "learning_rate": 9.99724618415786e-07, "loss": 0.3974, "num_input_tokens_seen": 469762048, "step": 224 }, { "epoch": 0.061025223759153785, "grad_norm": 15.416088104248047, "learning_rate": 9.997103167650637e-07, "loss": 0.7345, "num_input_tokens_seen": 471859200, "step": 225 }, { "epoch": 0.06129644697586113, "grad_norm": 12.959712982177734, "learning_rate": 9.996956532045914e-07, "loss": 0.4602, "num_input_tokens_seen": 473956352, "step": 226 }, { "epoch": 0.06156767019256849, "grad_norm": 24.262161254882812, "learning_rate": 9.996806277461696e-07, "loss": 0.9696, "num_input_tokens_seen": 476053504, "step": 227 }, { "epoch": 0.061838893409275834, "grad_norm": 21.882034301757812, "learning_rate": 9.9966524040189e-07, "loss": 0.7863, "num_input_tokens_seen": 478150656, "step": 228 }, { "epoch": 0.06211011662598318, "grad_norm": 17.270530700683594, "learning_rate": 9.996494911841363e-07, "loss": 0.6471, "num_input_tokens_seen": 480247808, "step": 229 }, { "epoch": 0.062381339842690536, "grad_norm": 13.247891426086426, "learning_rate": 9.996333801055823e-07, "loss": 0.3576, "num_input_tokens_seen": 482344960, "step": 230 }, { "epoch": 0.06265256305939788, "grad_norm": 13.374452590942383, "learning_rate": 9.99616907179194e-07, "loss": 0.4596, "num_input_tokens_seen": 484442112, "step": 231 }, { "epoch": 0.06292378627610523, "grad_norm": 18.631214141845703, "learning_rate": 9.996000724182278e-07, "loss": 0.7839, "num_input_tokens_seen": 486539264, "step": 232 }, { "epoch": 0.06319500949281258, "grad_norm": 18.132368087768555, "learning_rate": 9.99582875836232e-07, "loss": 0.636, "num_input_tokens_seen": 488636416, "step": 233 }, { "epoch": 0.06346623270951994, "grad_norm": 15.695167541503906, "learning_rate": 9.995653174470456e-07, "loss": 0.6983, "num_input_tokens_seen": 490733568, "step": 234 }, { "epoch": 0.06373745592622729, "grad_norm": 13.286776542663574, "learning_rate": 9.99547397264799e-07, "loss": 0.4567, "num_input_tokens_seen": 492830720, "step": 235 }, { "epoch": 0.06400867914293464, "grad_norm": 11.679216384887695, "learning_rate": 9.995291153039135e-07, "loss": 0.359, "num_input_tokens_seen": 494927872, "step": 236 }, { "epoch": 0.06427990235964198, "grad_norm": 16.616655349731445, "learning_rate": 9.99510471579102e-07, "loss": 0.6919, "num_input_tokens_seen": 497025024, "step": 237 }, { "epoch": 0.06455112557634933, "grad_norm": 15.427544593811035, "learning_rate": 9.99491466105368e-07, "loss": 0.5059, "num_input_tokens_seen": 499122176, "step": 238 }, { "epoch": 0.06482234879305669, "grad_norm": 38.47232437133789, "learning_rate": 9.994720988980065e-07, "loss": 0.6509, "num_input_tokens_seen": 501219328, "step": 239 }, { "epoch": 0.06509357200976404, "grad_norm": 12.464300155639648, "learning_rate": 9.994523699726035e-07, "loss": 0.3664, "num_input_tokens_seen": 503316480, "step": 240 }, { "epoch": 0.06536479522647139, "grad_norm": 25.51685905456543, "learning_rate": 9.994322793450361e-07, "loss": 0.9145, "num_input_tokens_seen": 505413632, "step": 241 }, { "epoch": 0.06563601844317873, "grad_norm": 18.906505584716797, "learning_rate": 9.994118270314725e-07, "loss": 0.7751, "num_input_tokens_seen": 507510784, "step": 242 }, { "epoch": 0.06590724165988608, "grad_norm": 22.564781188964844, "learning_rate": 9.993910130483717e-07, "loss": 0.8191, "num_input_tokens_seen": 509607936, "step": 243 }, { "epoch": 0.06617846487659344, "grad_norm": 12.493294715881348, "learning_rate": 9.993698374124844e-07, "loss": 0.4122, "num_input_tokens_seen": 511705088, "step": 244 }, { "epoch": 0.06644968809330079, "grad_norm": 16.386438369750977, "learning_rate": 9.993483001408516e-07, "loss": 0.6446, "num_input_tokens_seen": 513802240, "step": 245 }, { "epoch": 0.06672091131000814, "grad_norm": 20.80731773376465, "learning_rate": 9.99326401250806e-07, "loss": 0.7367, "num_input_tokens_seen": 515899392, "step": 246 }, { "epoch": 0.06699213452671549, "grad_norm": 21.987316131591797, "learning_rate": 9.993041407599708e-07, "loss": 0.5971, "num_input_tokens_seen": 517996544, "step": 247 }, { "epoch": 0.06726335774342283, "grad_norm": 14.191584587097168, "learning_rate": 9.992815186862602e-07, "loss": 0.468, "num_input_tokens_seen": 520093696, "step": 248 }, { "epoch": 0.06753458096013018, "grad_norm": 19.399492263793945, "learning_rate": 9.9925853504788e-07, "loss": 0.7622, "num_input_tokens_seen": 522190848, "step": 249 }, { "epoch": 0.06780580417683754, "grad_norm": 15.56145191192627, "learning_rate": 9.992351898633262e-07, "loss": 0.6956, "num_input_tokens_seen": 524288000, "step": 250 }, { "epoch": 0.06807702739354489, "grad_norm": 10.482128143310547, "learning_rate": 9.992114831513863e-07, "loss": 0.2901, "num_input_tokens_seen": 526385152, "step": 251 }, { "epoch": 0.06834825061025224, "grad_norm": 14.247305870056152, "learning_rate": 9.991874149311386e-07, "loss": 0.5265, "num_input_tokens_seen": 528482304, "step": 252 }, { "epoch": 0.06861947382695958, "grad_norm": 23.847684860229492, "learning_rate": 9.991629852219523e-07, "loss": 1.0187, "num_input_tokens_seen": 530579456, "step": 253 }, { "epoch": 0.06889069704366693, "grad_norm": 19.322067260742188, "learning_rate": 9.991381940434873e-07, "loss": 0.7402, "num_input_tokens_seen": 532676608, "step": 254 }, { "epoch": 0.0691619202603743, "grad_norm": 19.398937225341797, "learning_rate": 9.991130414156946e-07, "loss": 0.8914, "num_input_tokens_seen": 534773760, "step": 255 }, { "epoch": 0.06943314347708164, "grad_norm": 14.539494514465332, "learning_rate": 9.990875273588161e-07, "loss": 0.6505, "num_input_tokens_seen": 536870912, "step": 256 }, { "epoch": 0.06970436669378899, "grad_norm": 15.979605674743652, "learning_rate": 9.99061651893385e-07, "loss": 0.6311, "num_input_tokens_seen": 538968064, "step": 257 }, { "epoch": 0.06997558991049634, "grad_norm": 13.959792137145996, "learning_rate": 9.990354150402242e-07, "loss": 0.4779, "num_input_tokens_seen": 541065216, "step": 258 }, { "epoch": 0.07024681312720368, "grad_norm": 14.552385330200195, "learning_rate": 9.990088168204487e-07, "loss": 0.4992, "num_input_tokens_seen": 543162368, "step": 259 }, { "epoch": 0.07051803634391104, "grad_norm": 18.06981086730957, "learning_rate": 9.989818572554633e-07, "loss": 0.8364, "num_input_tokens_seen": 545259520, "step": 260 }, { "epoch": 0.07078925956061839, "grad_norm": 16.378019332885742, "learning_rate": 9.989545363669644e-07, "loss": 0.7462, "num_input_tokens_seen": 547356672, "step": 261 }, { "epoch": 0.07106048277732574, "grad_norm": 12.575121879577637, "learning_rate": 9.989268541769383e-07, "loss": 0.4064, "num_input_tokens_seen": 549453824, "step": 262 }, { "epoch": 0.07133170599403309, "grad_norm": 14.232921600341797, "learning_rate": 9.988988107076632e-07, "loss": 0.6201, "num_input_tokens_seen": 551550976, "step": 263 }, { "epoch": 0.07160292921074043, "grad_norm": 14.675073623657227, "learning_rate": 9.98870405981707e-07, "loss": 0.4924, "num_input_tokens_seen": 553648128, "step": 264 }, { "epoch": 0.0718741524274478, "grad_norm": 15.136611938476562, "learning_rate": 9.988416400219288e-07, "loss": 0.6525, "num_input_tokens_seen": 555745280, "step": 265 }, { "epoch": 0.07214537564415514, "grad_norm": 18.230609893798828, "learning_rate": 9.988125128514785e-07, "loss": 0.8809, "num_input_tokens_seen": 557842432, "step": 266 }, { "epoch": 0.07241659886086249, "grad_norm": 15.418828964233398, "learning_rate": 9.987830244937964e-07, "loss": 0.8042, "num_input_tokens_seen": 559939584, "step": 267 }, { "epoch": 0.07268782207756984, "grad_norm": 20.395606994628906, "learning_rate": 9.987531749726137e-07, "loss": 0.5648, "num_input_tokens_seen": 562036736, "step": 268 }, { "epoch": 0.07295904529427719, "grad_norm": 16.41619873046875, "learning_rate": 9.98722964311952e-07, "loss": 0.6025, "num_input_tokens_seen": 564133888, "step": 269 }, { "epoch": 0.07323026851098453, "grad_norm": 15.053420066833496, "learning_rate": 9.986923925361238e-07, "loss": 0.6662, "num_input_tokens_seen": 566231040, "step": 270 }, { "epoch": 0.0735014917276919, "grad_norm": 19.180273056030273, "learning_rate": 9.98661459669732e-07, "loss": 0.7407, "num_input_tokens_seen": 568328192, "step": 271 }, { "epoch": 0.07377271494439924, "grad_norm": 19.252796173095703, "learning_rate": 9.986301657376705e-07, "loss": 0.7506, "num_input_tokens_seen": 570425344, "step": 272 }, { "epoch": 0.07404393816110659, "grad_norm": 12.588926315307617, "learning_rate": 9.985985107651231e-07, "loss": 0.418, "num_input_tokens_seen": 572522496, "step": 273 }, { "epoch": 0.07431516137781394, "grad_norm": 19.830280303955078, "learning_rate": 9.985664947775649e-07, "loss": 0.9066, "num_input_tokens_seen": 574619648, "step": 274 }, { "epoch": 0.07458638459452128, "grad_norm": 12.818648338317871, "learning_rate": 9.985341178007608e-07, "loss": 0.5215, "num_input_tokens_seen": 576716800, "step": 275 }, { "epoch": 0.07485760781122865, "grad_norm": 18.200252532958984, "learning_rate": 9.985013798607666e-07, "loss": 0.8381, "num_input_tokens_seen": 578813952, "step": 276 }, { "epoch": 0.075128831027936, "grad_norm": 17.148433685302734, "learning_rate": 9.98468280983929e-07, "loss": 0.6161, "num_input_tokens_seen": 580911104, "step": 277 }, { "epoch": 0.07540005424464334, "grad_norm": 22.672407150268555, "learning_rate": 9.984348211968837e-07, "loss": 0.8064, "num_input_tokens_seen": 583008256, "step": 278 }, { "epoch": 0.07567127746135069, "grad_norm": 17.892045974731445, "learning_rate": 9.984010005265592e-07, "loss": 0.6633, "num_input_tokens_seen": 585105408, "step": 279 }, { "epoch": 0.07594250067805804, "grad_norm": 16.25893211364746, "learning_rate": 9.98366819000172e-07, "loss": 0.6054, "num_input_tokens_seen": 587202560, "step": 280 }, { "epoch": 0.0762137238947654, "grad_norm": 13.858470916748047, "learning_rate": 9.983322766452305e-07, "loss": 0.4507, "num_input_tokens_seen": 589299712, "step": 281 }, { "epoch": 0.07648494711147275, "grad_norm": 12.110980033874512, "learning_rate": 9.98297373489533e-07, "loss": 0.3055, "num_input_tokens_seen": 591396864, "step": 282 }, { "epoch": 0.07675617032818009, "grad_norm": 18.724361419677734, "learning_rate": 9.982621095611686e-07, "loss": 0.6864, "num_input_tokens_seen": 593494016, "step": 283 }, { "epoch": 0.07702739354488744, "grad_norm": 13.076481819152832, "learning_rate": 9.98226484888516e-07, "loss": 0.4906, "num_input_tokens_seen": 595591168, "step": 284 }, { "epoch": 0.07729861676159479, "grad_norm": 22.247615814208984, "learning_rate": 9.981904995002443e-07, "loss": 0.447, "num_input_tokens_seen": 597688320, "step": 285 }, { "epoch": 0.07756983997830215, "grad_norm": 15.656414031982422, "learning_rate": 9.98154153425314e-07, "loss": 0.5142, "num_input_tokens_seen": 599785472, "step": 286 }, { "epoch": 0.0778410631950095, "grad_norm": 16.987443923950195, "learning_rate": 9.981174466929742e-07, "loss": 0.6084, "num_input_tokens_seen": 601882624, "step": 287 }, { "epoch": 0.07811228641171684, "grad_norm": 15.571359634399414, "learning_rate": 9.980803793327655e-07, "loss": 0.5127, "num_input_tokens_seen": 603979776, "step": 288 }, { "epoch": 0.07838350962842419, "grad_norm": 17.951066970825195, "learning_rate": 9.980429513745182e-07, "loss": 0.8033, "num_input_tokens_seen": 606076928, "step": 289 }, { "epoch": 0.07865473284513154, "grad_norm": 24.38446617126465, "learning_rate": 9.980051628483532e-07, "loss": 1.0212, "num_input_tokens_seen": 608174080, "step": 290 }, { "epoch": 0.07892595606183889, "grad_norm": 14.629419326782227, "learning_rate": 9.979670137846806e-07, "loss": 0.6336, "num_input_tokens_seen": 610271232, "step": 291 }, { "epoch": 0.07919717927854625, "grad_norm": 23.52570343017578, "learning_rate": 9.97928504214202e-07, "loss": 1.1455, "num_input_tokens_seen": 612368384, "step": 292 }, { "epoch": 0.0794684024952536, "grad_norm": 20.963306427001953, "learning_rate": 9.97889634167908e-07, "loss": 0.9892, "num_input_tokens_seen": 614465536, "step": 293 }, { "epoch": 0.07973962571196094, "grad_norm": 16.687395095825195, "learning_rate": 9.978504036770802e-07, "loss": 0.3016, "num_input_tokens_seen": 616562688, "step": 294 }, { "epoch": 0.08001084892866829, "grad_norm": 11.436803817749023, "learning_rate": 9.978108127732892e-07, "loss": 0.378, "num_input_tokens_seen": 618659840, "step": 295 }, { "epoch": 0.08028207214537564, "grad_norm": 16.03774642944336, "learning_rate": 9.977708614883965e-07, "loss": 0.5094, "num_input_tokens_seen": 620756992, "step": 296 }, { "epoch": 0.080553295362083, "grad_norm": 25.998807907104492, "learning_rate": 9.977305498545537e-07, "loss": 0.7122, "num_input_tokens_seen": 622854144, "step": 297 }, { "epoch": 0.08082451857879035, "grad_norm": 12.087435722351074, "learning_rate": 9.976898779042018e-07, "loss": 0.455, "num_input_tokens_seen": 624951296, "step": 298 }, { "epoch": 0.0810957417954977, "grad_norm": 19.255043029785156, "learning_rate": 9.976488456700717e-07, "loss": 0.7966, "num_input_tokens_seen": 627048448, "step": 299 }, { "epoch": 0.08136696501220504, "grad_norm": 15.657422065734863, "learning_rate": 9.97607453185185e-07, "loss": 0.61, "num_input_tokens_seen": 629145600, "step": 300 }, { "epoch": 0.08163818822891239, "grad_norm": 17.275453567504883, "learning_rate": 9.97565700482853e-07, "loss": 0.7554, "num_input_tokens_seen": 631242752, "step": 301 }, { "epoch": 0.08190941144561975, "grad_norm": 16.451929092407227, "learning_rate": 9.97523587596676e-07, "loss": 0.5496, "num_input_tokens_seen": 633339904, "step": 302 }, { "epoch": 0.0821806346623271, "grad_norm": 11.221899032592773, "learning_rate": 9.974811145605453e-07, "loss": 0.2887, "num_input_tokens_seen": 635437056, "step": 303 }, { "epoch": 0.08245185787903445, "grad_norm": 15.957961082458496, "learning_rate": 9.974382814086418e-07, "loss": 0.3712, "num_input_tokens_seen": 637534208, "step": 304 }, { "epoch": 0.0827230810957418, "grad_norm": 14.90071964263916, "learning_rate": 9.973950881754353e-07, "loss": 0.4835, "num_input_tokens_seen": 639631360, "step": 305 }, { "epoch": 0.08299430431244914, "grad_norm": 17.702465057373047, "learning_rate": 9.973515348956869e-07, "loss": 0.6427, "num_input_tokens_seen": 641728512, "step": 306 }, { "epoch": 0.0832655275291565, "grad_norm": 13.002301216125488, "learning_rate": 9.97307621604446e-07, "loss": 0.5525, "num_input_tokens_seen": 643825664, "step": 307 }, { "epoch": 0.08353675074586385, "grad_norm": 12.444464683532715, "learning_rate": 9.972633483370526e-07, "loss": 0.4219, "num_input_tokens_seen": 645922816, "step": 308 }, { "epoch": 0.0838079739625712, "grad_norm": 34.6700439453125, "learning_rate": 9.97218715129136e-07, "loss": 0.8305, "num_input_tokens_seen": 648019968, "step": 309 }, { "epoch": 0.08407919717927854, "grad_norm": 13.552191734313965, "learning_rate": 9.971737220166155e-07, "loss": 0.4636, "num_input_tokens_seen": 650117120, "step": 310 }, { "epoch": 0.08435042039598589, "grad_norm": 15.92861270904541, "learning_rate": 9.971283690356997e-07, "loss": 0.68, "num_input_tokens_seen": 652214272, "step": 311 }, { "epoch": 0.08462164361269324, "grad_norm": 10.955565452575684, "learning_rate": 9.97082656222887e-07, "loss": 0.292, "num_input_tokens_seen": 654311424, "step": 312 }, { "epoch": 0.0848928668294006, "grad_norm": 16.476566314697266, "learning_rate": 9.970365836149654e-07, "loss": 0.6708, "num_input_tokens_seen": 656408576, "step": 313 }, { "epoch": 0.08516409004610795, "grad_norm": 21.01806640625, "learning_rate": 9.969901512490121e-07, "loss": 1.0652, "num_input_tokens_seen": 658505728, "step": 314 }, { "epoch": 0.0854353132628153, "grad_norm": 16.38351058959961, "learning_rate": 9.969433591623946e-07, "loss": 0.6264, "num_input_tokens_seen": 660602880, "step": 315 }, { "epoch": 0.08570653647952264, "grad_norm": 23.092439651489258, "learning_rate": 9.96896207392769e-07, "loss": 1.0512, "num_input_tokens_seen": 662700032, "step": 316 }, { "epoch": 0.08597775969622999, "grad_norm": 14.13205337524414, "learning_rate": 9.968486959780813e-07, "loss": 0.5465, "num_input_tokens_seen": 664797184, "step": 317 }, { "epoch": 0.08624898291293735, "grad_norm": 16.44294548034668, "learning_rate": 9.96800824956567e-07, "loss": 0.6146, "num_input_tokens_seen": 666894336, "step": 318 }, { "epoch": 0.0865202061296447, "grad_norm": 16.18340301513672, "learning_rate": 9.967525943667506e-07, "loss": 0.6067, "num_input_tokens_seen": 668991488, "step": 319 }, { "epoch": 0.08679142934635205, "grad_norm": 14.490288734436035, "learning_rate": 9.967040042474467e-07, "loss": 0.5758, "num_input_tokens_seen": 671088640, "step": 320 }, { "epoch": 0.0870626525630594, "grad_norm": 20.6240234375, "learning_rate": 9.966550546377586e-07, "loss": 0.8391, "num_input_tokens_seen": 673185792, "step": 321 }, { "epoch": 0.08733387577976674, "grad_norm": 12.437811851501465, "learning_rate": 9.966057455770788e-07, "loss": 0.3815, "num_input_tokens_seen": 675282944, "step": 322 }, { "epoch": 0.0876050989964741, "grad_norm": 17.535430908203125, "learning_rate": 9.965560771050896e-07, "loss": 0.7666, "num_input_tokens_seen": 677380096, "step": 323 }, { "epoch": 0.08787632221318145, "grad_norm": 17.357746124267578, "learning_rate": 9.965060492617623e-07, "loss": 0.6267, "num_input_tokens_seen": 679477248, "step": 324 }, { "epoch": 0.0881475454298888, "grad_norm": 19.566974639892578, "learning_rate": 9.964556620873573e-07, "loss": 0.6236, "num_input_tokens_seen": 681574400, "step": 325 }, { "epoch": 0.08841876864659615, "grad_norm": 14.313620567321777, "learning_rate": 9.964049156224244e-07, "loss": 0.5585, "num_input_tokens_seen": 683671552, "step": 326 }, { "epoch": 0.0886899918633035, "grad_norm": 13.445440292358398, "learning_rate": 9.963538099078024e-07, "loss": 0.4307, "num_input_tokens_seen": 685768704, "step": 327 }, { "epoch": 0.08896121508001086, "grad_norm": 17.229440689086914, "learning_rate": 9.963023449846194e-07, "loss": 0.7462, "num_input_tokens_seen": 687865856, "step": 328 }, { "epoch": 0.0892324382967182, "grad_norm": 20.201749801635742, "learning_rate": 9.962505208942919e-07, "loss": 0.9177, "num_input_tokens_seen": 689963008, "step": 329 }, { "epoch": 0.08950366151342555, "grad_norm": 13.015423774719238, "learning_rate": 9.961983376785264e-07, "loss": 0.4933, "num_input_tokens_seen": 692060160, "step": 330 }, { "epoch": 0.0897748847301329, "grad_norm": 11.634929656982422, "learning_rate": 9.96145795379318e-07, "loss": 0.3818, "num_input_tokens_seen": 694157312, "step": 331 }, { "epoch": 0.09004610794684025, "grad_norm": 18.479869842529297, "learning_rate": 9.960928940389503e-07, "loss": 0.8161, "num_input_tokens_seen": 696254464, "step": 332 }, { "epoch": 0.09031733116354759, "grad_norm": 12.271891593933105, "learning_rate": 9.960396336999967e-07, "loss": 0.451, "num_input_tokens_seen": 698351616, "step": 333 }, { "epoch": 0.09058855438025495, "grad_norm": 10.953863143920898, "learning_rate": 9.95986014405319e-07, "loss": 0.3789, "num_input_tokens_seen": 700448768, "step": 334 }, { "epoch": 0.0908597775969623, "grad_norm": 9.575960159301758, "learning_rate": 9.959320361980679e-07, "loss": 0.3936, "num_input_tokens_seen": 702545920, "step": 335 }, { "epoch": 0.09113100081366965, "grad_norm": 13.222553253173828, "learning_rate": 9.95877699121683e-07, "loss": 0.256, "num_input_tokens_seen": 704643072, "step": 336 }, { "epoch": 0.091402224030377, "grad_norm": 20.804859161376953, "learning_rate": 9.95823003219893e-07, "loss": 1.0641, "num_input_tokens_seen": 706740224, "step": 337 }, { "epoch": 0.09167344724708434, "grad_norm": 11.381640434265137, "learning_rate": 9.957679485367144e-07, "loss": 0.3291, "num_input_tokens_seen": 708837376, "step": 338 }, { "epoch": 0.0919446704637917, "grad_norm": 13.033456802368164, "learning_rate": 9.95712535116454e-07, "loss": 0.4397, "num_input_tokens_seen": 710934528, "step": 339 }, { "epoch": 0.09221589368049905, "grad_norm": 21.973115921020508, "learning_rate": 9.956567630037058e-07, "loss": 0.6069, "num_input_tokens_seen": 713031680, "step": 340 }, { "epoch": 0.0924871168972064, "grad_norm": 20.099735260009766, "learning_rate": 9.95600632243353e-07, "loss": 0.8007, "num_input_tokens_seen": 715128832, "step": 341 }, { "epoch": 0.09275834011391375, "grad_norm": 19.394807815551758, "learning_rate": 9.95544142880568e-07, "loss": 0.7435, "num_input_tokens_seen": 717225984, "step": 342 }, { "epoch": 0.0930295633306211, "grad_norm": 17.3151798248291, "learning_rate": 9.954872949608108e-07, "loss": 0.5055, "num_input_tokens_seen": 719323136, "step": 343 }, { "epoch": 0.09330078654732846, "grad_norm": 19.800769805908203, "learning_rate": 9.954300885298309e-07, "loss": 0.6907, "num_input_tokens_seen": 721420288, "step": 344 }, { "epoch": 0.0935720097640358, "grad_norm": 11.735955238342285, "learning_rate": 9.953725236336653e-07, "loss": 0.3009, "num_input_tokens_seen": 723517440, "step": 345 }, { "epoch": 0.09384323298074315, "grad_norm": 15.41151237487793, "learning_rate": 9.953146003186407e-07, "loss": 0.577, "num_input_tokens_seen": 725614592, "step": 346 }, { "epoch": 0.0941144561974505, "grad_norm": 21.002737045288086, "learning_rate": 9.952563186313711e-07, "loss": 0.9646, "num_input_tokens_seen": 727711744, "step": 347 }, { "epoch": 0.09438567941415785, "grad_norm": 8.514850616455078, "learning_rate": 9.951976786187598e-07, "loss": 0.1761, "num_input_tokens_seen": 729808896, "step": 348 }, { "epoch": 0.09465690263086521, "grad_norm": 17.795473098754883, "learning_rate": 9.951386803279973e-07, "loss": 0.5064, "num_input_tokens_seen": 731906048, "step": 349 }, { "epoch": 0.09492812584757256, "grad_norm": 15.562695503234863, "learning_rate": 9.95079323806564e-07, "loss": 0.5842, "num_input_tokens_seen": 734003200, "step": 350 }, { "epoch": 0.0951993490642799, "grad_norm": 14.850616455078125, "learning_rate": 9.950196091022274e-07, "loss": 0.4142, "num_input_tokens_seen": 736100352, "step": 351 }, { "epoch": 0.09547057228098725, "grad_norm": 13.248444557189941, "learning_rate": 9.949595362630435e-07, "loss": 0.3965, "num_input_tokens_seen": 738197504, "step": 352 }, { "epoch": 0.0957417954976946, "grad_norm": 19.31031608581543, "learning_rate": 9.948991053373567e-07, "loss": 0.6689, "num_input_tokens_seen": 740294656, "step": 353 }, { "epoch": 0.09601301871440195, "grad_norm": 18.403900146484375, "learning_rate": 9.948383163738e-07, "loss": 0.6139, "num_input_tokens_seen": 742391808, "step": 354 }, { "epoch": 0.09628424193110931, "grad_norm": 16.723413467407227, "learning_rate": 9.947771694212933e-07, "loss": 0.5709, "num_input_tokens_seen": 744488960, "step": 355 }, { "epoch": 0.09655546514781665, "grad_norm": 14.325742721557617, "learning_rate": 9.947156645290456e-07, "loss": 0.6557, "num_input_tokens_seen": 746586112, "step": 356 }, { "epoch": 0.096826688364524, "grad_norm": 19.636194229125977, "learning_rate": 9.94653801746554e-07, "loss": 0.8035, "num_input_tokens_seen": 748683264, "step": 357 }, { "epoch": 0.09709791158123135, "grad_norm": 13.98243236541748, "learning_rate": 9.945915811236029e-07, "loss": 0.508, "num_input_tokens_seen": 750780416, "step": 358 }, { "epoch": 0.0973691347979387, "grad_norm": 14.497775077819824, "learning_rate": 9.945290027102654e-07, "loss": 0.4487, "num_input_tokens_seen": 752877568, "step": 359 }, { "epoch": 0.09764035801464606, "grad_norm": 15.805448532104492, "learning_rate": 9.944660665569023e-07, "loss": 0.6489, "num_input_tokens_seen": 754974720, "step": 360 }, { "epoch": 0.0979115812313534, "grad_norm": 14.843584060668945, "learning_rate": 9.944027727141617e-07, "loss": 0.5395, "num_input_tokens_seen": 757071872, "step": 361 }, { "epoch": 0.09818280444806075, "grad_norm": 15.000771522521973, "learning_rate": 9.943391212329805e-07, "loss": 0.4855, "num_input_tokens_seen": 759169024, "step": 362 }, { "epoch": 0.0984540276647681, "grad_norm": 26.341798782348633, "learning_rate": 9.942751121645828e-07, "loss": 1.1353, "num_input_tokens_seen": 761266176, "step": 363 }, { "epoch": 0.09872525088147545, "grad_norm": 14.345602989196777, "learning_rate": 9.94210745560481e-07, "loss": 0.5821, "num_input_tokens_seen": 763363328, "step": 364 }, { "epoch": 0.09899647409818281, "grad_norm": 13.088194847106934, "learning_rate": 9.941460214724747e-07, "loss": 0.3652, "num_input_tokens_seen": 765460480, "step": 365 }, { "epoch": 0.09926769731489016, "grad_norm": 25.987003326416016, "learning_rate": 9.94080939952651e-07, "loss": 0.8981, "num_input_tokens_seen": 767557632, "step": 366 }, { "epoch": 0.0995389205315975, "grad_norm": 19.704673767089844, "learning_rate": 9.940155010533855e-07, "loss": 0.721, "num_input_tokens_seen": 769654784, "step": 367 }, { "epoch": 0.09981014374830485, "grad_norm": 15.577407836914062, "learning_rate": 9.939497048273407e-07, "loss": 0.741, "num_input_tokens_seen": 771751936, "step": 368 }, { "epoch": 0.1000813669650122, "grad_norm": 23.83604621887207, "learning_rate": 9.938835513274672e-07, "loss": 0.7845, "num_input_tokens_seen": 773849088, "step": 369 }, { "epoch": 0.10035259018171956, "grad_norm": 13.335168838500977, "learning_rate": 9.938170406070025e-07, "loss": 0.4799, "num_input_tokens_seen": 775946240, "step": 370 }, { "epoch": 0.10062381339842691, "grad_norm": 12.84589672088623, "learning_rate": 9.937501727194721e-07, "loss": 0.4814, "num_input_tokens_seen": 778043392, "step": 371 }, { "epoch": 0.10089503661513426, "grad_norm": 17.722841262817383, "learning_rate": 9.936829477186884e-07, "loss": 0.6229, "num_input_tokens_seen": 780140544, "step": 372 }, { "epoch": 0.1011662598318416, "grad_norm": 11.619221687316895, "learning_rate": 9.93615365658752e-07, "loss": 0.3687, "num_input_tokens_seen": 782237696, "step": 373 }, { "epoch": 0.10143748304854895, "grad_norm": 17.260656356811523, "learning_rate": 9.9354742659405e-07, "loss": 0.7215, "num_input_tokens_seen": 784334848, "step": 374 }, { "epoch": 0.10170870626525631, "grad_norm": 19.690866470336914, "learning_rate": 9.934791305792575e-07, "loss": 1.015, "num_input_tokens_seen": 786432000, "step": 375 }, { "epoch": 0.10197992948196366, "grad_norm": 22.19734001159668, "learning_rate": 9.934104776693363e-07, "loss": 1.135, "num_input_tokens_seen": 788529152, "step": 376 }, { "epoch": 0.10225115269867101, "grad_norm": 12.495186805725098, "learning_rate": 9.933414679195354e-07, "loss": 0.4608, "num_input_tokens_seen": 790626304, "step": 377 }, { "epoch": 0.10252237591537836, "grad_norm": 13.882384300231934, "learning_rate": 9.932721013853917e-07, "loss": 0.4727, "num_input_tokens_seen": 792723456, "step": 378 }, { "epoch": 0.1027935991320857, "grad_norm": 17.025941848754883, "learning_rate": 9.932023781227287e-07, "loss": 0.3467, "num_input_tokens_seen": 794820608, "step": 379 }, { "epoch": 0.10306482234879305, "grad_norm": 9.603739738464355, "learning_rate": 9.931322981876567e-07, "loss": 0.3002, "num_input_tokens_seen": 796917760, "step": 380 }, { "epoch": 0.10333604556550041, "grad_norm": 9.856173515319824, "learning_rate": 9.930618616365737e-07, "loss": 0.2998, "num_input_tokens_seen": 799014912, "step": 381 }, { "epoch": 0.10360726878220776, "grad_norm": 16.00568199157715, "learning_rate": 9.92991068526164e-07, "loss": 0.6395, "num_input_tokens_seen": 801112064, "step": 382 }, { "epoch": 0.1038784919989151, "grad_norm": 13.31304931640625, "learning_rate": 9.929199189133996e-07, "loss": 0.4683, "num_input_tokens_seen": 803209216, "step": 383 }, { "epoch": 0.10414971521562245, "grad_norm": 22.845258712768555, "learning_rate": 9.928484128555388e-07, "loss": 0.9509, "num_input_tokens_seen": 805306368, "step": 384 }, { "epoch": 0.1044209384323298, "grad_norm": 18.43478012084961, "learning_rate": 9.92776550410127e-07, "loss": 0.6683, "num_input_tokens_seen": 807403520, "step": 385 }, { "epoch": 0.10469216164903716, "grad_norm": 19.82917022705078, "learning_rate": 9.927043316349962e-07, "loss": 0.375, "num_input_tokens_seen": 809500672, "step": 386 }, { "epoch": 0.10496338486574451, "grad_norm": 17.08930206298828, "learning_rate": 9.926317565882657e-07, "loss": 0.4983, "num_input_tokens_seen": 811597824, "step": 387 }, { "epoch": 0.10523460808245186, "grad_norm": 13.114651679992676, "learning_rate": 9.925588253283407e-07, "loss": 0.433, "num_input_tokens_seen": 813694976, "step": 388 }, { "epoch": 0.1055058312991592, "grad_norm": 12.93115234375, "learning_rate": 9.924855379139136e-07, "loss": 0.4286, "num_input_tokens_seen": 815792128, "step": 389 }, { "epoch": 0.10577705451586655, "grad_norm": 16.018659591674805, "learning_rate": 9.924118944039635e-07, "loss": 0.5672, "num_input_tokens_seen": 817889280, "step": 390 }, { "epoch": 0.10604827773257391, "grad_norm": 17.725582122802734, "learning_rate": 9.923378948577558e-07, "loss": 0.7503, "num_input_tokens_seen": 819986432, "step": 391 }, { "epoch": 0.10631950094928126, "grad_norm": 12.33151912689209, "learning_rate": 9.922635393348425e-07, "loss": 0.4547, "num_input_tokens_seen": 822083584, "step": 392 }, { "epoch": 0.10659072416598861, "grad_norm": 15.34462833404541, "learning_rate": 9.92188827895062e-07, "loss": 0.4688, "num_input_tokens_seen": 824180736, "step": 393 }, { "epoch": 0.10686194738269596, "grad_norm": 20.61101722717285, "learning_rate": 9.921137605985397e-07, "loss": 0.8941, "num_input_tokens_seen": 826277888, "step": 394 }, { "epoch": 0.1071331705994033, "grad_norm": 14.963687896728516, "learning_rate": 9.920383375056863e-07, "loss": 0.584, "num_input_tokens_seen": 828375040, "step": 395 }, { "epoch": 0.10740439381611067, "grad_norm": 17.79062843322754, "learning_rate": 9.919625586771998e-07, "loss": 0.7246, "num_input_tokens_seen": 830472192, "step": 396 }, { "epoch": 0.10767561703281801, "grad_norm": 16.45888328552246, "learning_rate": 9.918864241740639e-07, "loss": 0.6275, "num_input_tokens_seen": 832569344, "step": 397 }, { "epoch": 0.10794684024952536, "grad_norm": 17.214998245239258, "learning_rate": 9.918099340575487e-07, "loss": 0.53, "num_input_tokens_seen": 834666496, "step": 398 }, { "epoch": 0.10821806346623271, "grad_norm": 18.772640228271484, "learning_rate": 9.91733088389211e-07, "loss": 0.5958, "num_input_tokens_seen": 836763648, "step": 399 }, { "epoch": 0.10848928668294006, "grad_norm": 13.285908699035645, "learning_rate": 9.916558872308929e-07, "loss": 0.3651, "num_input_tokens_seen": 838860800, "step": 400 }, { "epoch": 0.1087605098996474, "grad_norm": 14.000696182250977, "learning_rate": 9.915783306447229e-07, "loss": 0.5324, "num_input_tokens_seen": 840957952, "step": 401 }, { "epoch": 0.10903173311635476, "grad_norm": 14.921085357666016, "learning_rate": 9.915004186931156e-07, "loss": 0.5363, "num_input_tokens_seen": 843055104, "step": 402 }, { "epoch": 0.10930295633306211, "grad_norm": 13.49205493927002, "learning_rate": 9.91422151438772e-07, "loss": 0.3414, "num_input_tokens_seen": 845152256, "step": 403 }, { "epoch": 0.10957417954976946, "grad_norm": 20.21406364440918, "learning_rate": 9.91343528944678e-07, "loss": 0.8933, "num_input_tokens_seen": 847249408, "step": 404 }, { "epoch": 0.10984540276647681, "grad_norm": 11.765609741210938, "learning_rate": 9.912645512741064e-07, "loss": 0.3803, "num_input_tokens_seen": 849346560, "step": 405 }, { "epoch": 0.11011662598318415, "grad_norm": 22.0726318359375, "learning_rate": 9.911852184906151e-07, "loss": 0.5149, "num_input_tokens_seen": 851443712, "step": 406 }, { "epoch": 0.11038784919989152, "grad_norm": 14.043606758117676, "learning_rate": 9.911055306580485e-07, "loss": 0.5114, "num_input_tokens_seen": 853540864, "step": 407 }, { "epoch": 0.11065907241659886, "grad_norm": 12.334258079528809, "learning_rate": 9.910254878405361e-07, "loss": 0.3898, "num_input_tokens_seen": 855638016, "step": 408 }, { "epoch": 0.11093029563330621, "grad_norm": 11.179019927978516, "learning_rate": 9.909450901024935e-07, "loss": 0.3856, "num_input_tokens_seen": 857735168, "step": 409 }, { "epoch": 0.11120151885001356, "grad_norm": 17.625316619873047, "learning_rate": 9.908643375086213e-07, "loss": 0.7504, "num_input_tokens_seen": 859832320, "step": 410 }, { "epoch": 0.1114727420667209, "grad_norm": 11.948003768920898, "learning_rate": 9.907832301239066e-07, "loss": 0.4077, "num_input_tokens_seen": 861929472, "step": 411 }, { "epoch": 0.11174396528342827, "grad_norm": 22.031713485717773, "learning_rate": 9.907017680136213e-07, "loss": 0.8141, "num_input_tokens_seen": 864026624, "step": 412 }, { "epoch": 0.11201518850013562, "grad_norm": 14.142679214477539, "learning_rate": 9.90619951243323e-07, "loss": 0.5477, "num_input_tokens_seen": 866123776, "step": 413 }, { "epoch": 0.11228641171684296, "grad_norm": 16.43646240234375, "learning_rate": 9.905377798788547e-07, "loss": 0.6771, "num_input_tokens_seen": 868220928, "step": 414 }, { "epoch": 0.11255763493355031, "grad_norm": 14.17488956451416, "learning_rate": 9.904552539863452e-07, "loss": 0.4444, "num_input_tokens_seen": 870318080, "step": 415 }, { "epoch": 0.11282885815025766, "grad_norm": 17.454513549804688, "learning_rate": 9.903723736322075e-07, "loss": 0.7704, "num_input_tokens_seen": 872415232, "step": 416 }, { "epoch": 0.11310008136696502, "grad_norm": 18.859848022460938, "learning_rate": 9.90289138883141e-07, "loss": 0.813, "num_input_tokens_seen": 874512384, "step": 417 }, { "epoch": 0.11337130458367237, "grad_norm": 21.147764205932617, "learning_rate": 9.9020554980613e-07, "loss": 0.6002, "num_input_tokens_seen": 876609536, "step": 418 }, { "epoch": 0.11364252780037971, "grad_norm": 20.489925384521484, "learning_rate": 9.901216064684434e-07, "loss": 0.9423, "num_input_tokens_seen": 878706688, "step": 419 }, { "epoch": 0.11391375101708706, "grad_norm": 12.808001518249512, "learning_rate": 9.900373089376357e-07, "loss": 0.4228, "num_input_tokens_seen": 880803840, "step": 420 }, { "epoch": 0.11418497423379441, "grad_norm": 12.175069808959961, "learning_rate": 9.899526572815465e-07, "loss": 0.3785, "num_input_tokens_seen": 882900992, "step": 421 }, { "epoch": 0.11445619745050176, "grad_norm": 12.485739707946777, "learning_rate": 9.898676515683001e-07, "loss": 0.4116, "num_input_tokens_seen": 884998144, "step": 422 }, { "epoch": 0.11472742066720912, "grad_norm": 16.668554306030273, "learning_rate": 9.897822918663062e-07, "loss": 0.5853, "num_input_tokens_seen": 887095296, "step": 423 }, { "epoch": 0.11499864388391647, "grad_norm": 20.581716537475586, "learning_rate": 9.896965782442584e-07, "loss": 0.6277, "num_input_tokens_seen": 889192448, "step": 424 }, { "epoch": 0.11526986710062381, "grad_norm": 17.985069274902344, "learning_rate": 9.896105107711365e-07, "loss": 0.5649, "num_input_tokens_seen": 891289600, "step": 425 }, { "epoch": 0.11554109031733116, "grad_norm": 16.7568302154541, "learning_rate": 9.895240895162037e-07, "loss": 0.5382, "num_input_tokens_seen": 893386752, "step": 426 }, { "epoch": 0.11581231353403851, "grad_norm": 14.392632484436035, "learning_rate": 9.89437314549009e-07, "loss": 0.493, "num_input_tokens_seen": 895483904, "step": 427 }, { "epoch": 0.11608353675074587, "grad_norm": 11.781503677368164, "learning_rate": 9.89350185939385e-07, "loss": 0.2819, "num_input_tokens_seen": 897581056, "step": 428 }, { "epoch": 0.11635475996745322, "grad_norm": 18.502288818359375, "learning_rate": 9.8926270375745e-07, "loss": 0.8497, "num_input_tokens_seen": 899678208, "step": 429 }, { "epoch": 0.11662598318416056, "grad_norm": 13.56667709350586, "learning_rate": 9.891748680736064e-07, "loss": 0.4923, "num_input_tokens_seen": 901775360, "step": 430 }, { "epoch": 0.11689720640086791, "grad_norm": 19.49355125427246, "learning_rate": 9.890866789585407e-07, "loss": 0.6438, "num_input_tokens_seen": 903872512, "step": 431 }, { "epoch": 0.11716842961757526, "grad_norm": 11.730159759521484, "learning_rate": 9.889981364832245e-07, "loss": 0.3834, "num_input_tokens_seen": 905969664, "step": 432 }, { "epoch": 0.11743965283428262, "grad_norm": 10.967796325683594, "learning_rate": 9.889092407189129e-07, "loss": 0.3682, "num_input_tokens_seen": 908066816, "step": 433 }, { "epoch": 0.11771087605098997, "grad_norm": 16.377155303955078, "learning_rate": 9.88819991737146e-07, "loss": 0.454, "num_input_tokens_seen": 910163968, "step": 434 }, { "epoch": 0.11798209926769732, "grad_norm": 15.872443199157715, "learning_rate": 9.887303896097483e-07, "loss": 0.6401, "num_input_tokens_seen": 912261120, "step": 435 }, { "epoch": 0.11825332248440466, "grad_norm": 15.673075675964355, "learning_rate": 9.88640434408828e-07, "loss": 0.7127, "num_input_tokens_seen": 914358272, "step": 436 }, { "epoch": 0.11852454570111201, "grad_norm": 18.29340934753418, "learning_rate": 9.885501262067776e-07, "loss": 0.4442, "num_input_tokens_seen": 916455424, "step": 437 }, { "epoch": 0.11879576891781937, "grad_norm": 21.140331268310547, "learning_rate": 9.884594650762734e-07, "loss": 1.0107, "num_input_tokens_seen": 918552576, "step": 438 }, { "epoch": 0.11906699213452672, "grad_norm": 21.628915786743164, "learning_rate": 9.883684510902767e-07, "loss": 0.4715, "num_input_tokens_seen": 920649728, "step": 439 }, { "epoch": 0.11933821535123407, "grad_norm": 17.046772003173828, "learning_rate": 9.882770843220316e-07, "loss": 0.7166, "num_input_tokens_seen": 922746880, "step": 440 }, { "epoch": 0.11960943856794141, "grad_norm": 13.360857963562012, "learning_rate": 9.881853648450667e-07, "loss": 0.4021, "num_input_tokens_seen": 924844032, "step": 441 }, { "epoch": 0.11988066178464876, "grad_norm": 12.711376190185547, "learning_rate": 9.880932927331942e-07, "loss": 0.3842, "num_input_tokens_seen": 926941184, "step": 442 }, { "epoch": 0.12015188500135611, "grad_norm": 13.425957679748535, "learning_rate": 9.880008680605104e-07, "loss": 0.5039, "num_input_tokens_seen": 929038336, "step": 443 }, { "epoch": 0.12042310821806347, "grad_norm": 14.570152282714844, "learning_rate": 9.879080909013955e-07, "loss": 0.3667, "num_input_tokens_seen": 931135488, "step": 444 }, { "epoch": 0.12069433143477082, "grad_norm": 15.635329246520996, "learning_rate": 9.878149613305125e-07, "loss": 0.6197, "num_input_tokens_seen": 933232640, "step": 445 }, { "epoch": 0.12096555465147817, "grad_norm": 10.609899520874023, "learning_rate": 9.877214794228087e-07, "loss": 0.3233, "num_input_tokens_seen": 935329792, "step": 446 }, { "epoch": 0.12123677786818551, "grad_norm": 18.10693359375, "learning_rate": 9.876276452535147e-07, "loss": 0.664, "num_input_tokens_seen": 937426944, "step": 447 }, { "epoch": 0.12150800108489286, "grad_norm": 12.761764526367188, "learning_rate": 9.87533458898145e-07, "loss": 0.3728, "num_input_tokens_seen": 939524096, "step": 448 }, { "epoch": 0.12177922430160022, "grad_norm": 11.233665466308594, "learning_rate": 9.874389204324967e-07, "loss": 0.4342, "num_input_tokens_seen": 941621248, "step": 449 }, { "epoch": 0.12205044751830757, "grad_norm": 14.0928955078125, "learning_rate": 9.873440299326513e-07, "loss": 0.5682, "num_input_tokens_seen": 943718400, "step": 450 }, { "epoch": 0.12232167073501492, "grad_norm": 14.01526927947998, "learning_rate": 9.872487874749726e-07, "loss": 0.5666, "num_input_tokens_seen": 945815552, "step": 451 }, { "epoch": 0.12259289395172226, "grad_norm": 15.093655586242676, "learning_rate": 9.871531931361084e-07, "loss": 0.4609, "num_input_tokens_seen": 947912704, "step": 452 }, { "epoch": 0.12286411716842961, "grad_norm": 16.862146377563477, "learning_rate": 9.870572469929892e-07, "loss": 0.4889, "num_input_tokens_seen": 950009856, "step": 453 }, { "epoch": 0.12313534038513697, "grad_norm": 14.488612174987793, "learning_rate": 9.869609491228288e-07, "loss": 0.4386, "num_input_tokens_seen": 952107008, "step": 454 }, { "epoch": 0.12340656360184432, "grad_norm": 10.171273231506348, "learning_rate": 9.868642996031243e-07, "loss": 0.3064, "num_input_tokens_seen": 954204160, "step": 455 }, { "epoch": 0.12367778681855167, "grad_norm": 13.183239936828613, "learning_rate": 9.867672985116553e-07, "loss": 0.4111, "num_input_tokens_seen": 956301312, "step": 456 }, { "epoch": 0.12394901003525902, "grad_norm": 13.565240859985352, "learning_rate": 9.866699459264846e-07, "loss": 0.5021, "num_input_tokens_seen": 958398464, "step": 457 }, { "epoch": 0.12422023325196636, "grad_norm": 14.678339958190918, "learning_rate": 9.865722419259582e-07, "loss": 0.4652, "num_input_tokens_seen": 960495616, "step": 458 }, { "epoch": 0.12449145646867373, "grad_norm": 20.48042869567871, "learning_rate": 9.864741865887042e-07, "loss": 0.91, "num_input_tokens_seen": 962592768, "step": 459 }, { "epoch": 0.12476267968538107, "grad_norm": 20.1795711517334, "learning_rate": 9.86375779993634e-07, "loss": 0.6582, "num_input_tokens_seen": 964689920, "step": 460 }, { "epoch": 0.1250339029020884, "grad_norm": 28.812326431274414, "learning_rate": 9.86277022219941e-07, "loss": 0.6452, "num_input_tokens_seen": 966787072, "step": 461 }, { "epoch": 0.12530512611879577, "grad_norm": 21.05647850036621, "learning_rate": 9.861779133471025e-07, "loss": 0.9777, "num_input_tokens_seen": 968884224, "step": 462 }, { "epoch": 0.12557634933550313, "grad_norm": 18.095561981201172, "learning_rate": 9.86078453454877e-07, "loss": 0.3998, "num_input_tokens_seen": 970981376, "step": 463 }, { "epoch": 0.12584757255221046, "grad_norm": 12.681943893432617, "learning_rate": 9.859786426233061e-07, "loss": 0.4588, "num_input_tokens_seen": 973078528, "step": 464 }, { "epoch": 0.12611879576891782, "grad_norm": 11.929615020751953, "learning_rate": 9.85878480932714e-07, "loss": 0.3087, "num_input_tokens_seen": 975175680, "step": 465 }, { "epoch": 0.12639001898562516, "grad_norm": 20.00129508972168, "learning_rate": 9.857779684637068e-07, "loss": 0.8953, "num_input_tokens_seen": 977272832, "step": 466 }, { "epoch": 0.12666124220233252, "grad_norm": 12.261861801147461, "learning_rate": 9.856771052971733e-07, "loss": 0.3192, "num_input_tokens_seen": 979369984, "step": 467 }, { "epoch": 0.12693246541903988, "grad_norm": 15.92088794708252, "learning_rate": 9.85575891514284e-07, "loss": 0.5931, "num_input_tokens_seen": 981467136, "step": 468 }, { "epoch": 0.12720368863574721, "grad_norm": 12.561171531677246, "learning_rate": 9.85474327196492e-07, "loss": 0.3394, "num_input_tokens_seen": 983564288, "step": 469 }, { "epoch": 0.12747491185245458, "grad_norm": 16.519527435302734, "learning_rate": 9.853724124255328e-07, "loss": 0.4335, "num_input_tokens_seen": 985661440, "step": 470 }, { "epoch": 0.1277461350691619, "grad_norm": 23.729719161987305, "learning_rate": 9.85270147283423e-07, "loss": 0.6265, "num_input_tokens_seen": 987758592, "step": 471 }, { "epoch": 0.12801735828586927, "grad_norm": 20.1112003326416, "learning_rate": 9.85167531852462e-07, "loss": 0.8864, "num_input_tokens_seen": 989855744, "step": 472 }, { "epoch": 0.12828858150257663, "grad_norm": 11.423717498779297, "learning_rate": 9.850645662152308e-07, "loss": 0.3044, "num_input_tokens_seen": 991952896, "step": 473 }, { "epoch": 0.12855980471928397, "grad_norm": 15.77081298828125, "learning_rate": 9.84961250454592e-07, "loss": 0.4784, "num_input_tokens_seen": 994050048, "step": 474 }, { "epoch": 0.12883102793599133, "grad_norm": 14.118724822998047, "learning_rate": 9.848575846536902e-07, "loss": 0.4869, "num_input_tokens_seen": 996147200, "step": 475 }, { "epoch": 0.12910225115269866, "grad_norm": 14.419587135314941, "learning_rate": 9.847535688959523e-07, "loss": 0.5372, "num_input_tokens_seen": 998244352, "step": 476 }, { "epoch": 0.12937347436940602, "grad_norm": 17.4903507232666, "learning_rate": 9.846492032650855e-07, "loss": 0.5932, "num_input_tokens_seen": 1000341504, "step": 477 }, { "epoch": 0.12964469758611338, "grad_norm": 13.698707580566406, "learning_rate": 9.845444878450794e-07, "loss": 0.4332, "num_input_tokens_seen": 1002438656, "step": 478 }, { "epoch": 0.12991592080282072, "grad_norm": 20.30847930908203, "learning_rate": 9.844394227202053e-07, "loss": 0.5934, "num_input_tokens_seen": 1004535808, "step": 479 }, { "epoch": 0.13018714401952808, "grad_norm": 14.15876293182373, "learning_rate": 9.843340079750154e-07, "loss": 0.4459, "num_input_tokens_seen": 1006632960, "step": 480 }, { "epoch": 0.1304583672362354, "grad_norm": 12.466696739196777, "learning_rate": 9.842282436943435e-07, "loss": 0.3993, "num_input_tokens_seen": 1008730112, "step": 481 }, { "epoch": 0.13072959045294277, "grad_norm": 17.355510711669922, "learning_rate": 9.841221299633049e-07, "loss": 0.517, "num_input_tokens_seen": 1010827264, "step": 482 }, { "epoch": 0.13100081366965013, "grad_norm": 23.515899658203125, "learning_rate": 9.840156668672953e-07, "loss": 0.7, "num_input_tokens_seen": 1012924416, "step": 483 }, { "epoch": 0.13127203688635747, "grad_norm": 8.689852714538574, "learning_rate": 9.839088544919927e-07, "loss": 0.3053, "num_input_tokens_seen": 1015021568, "step": 484 }, { "epoch": 0.13154326010306483, "grad_norm": 16.73533058166504, "learning_rate": 9.838016929233555e-07, "loss": 0.4763, "num_input_tokens_seen": 1017118720, "step": 485 }, { "epoch": 0.13181448331977216, "grad_norm": 16.722810745239258, "learning_rate": 9.836941822476232e-07, "loss": 0.3931, "num_input_tokens_seen": 1019215872, "step": 486 }, { "epoch": 0.13208570653647952, "grad_norm": 16.654083251953125, "learning_rate": 9.835863225513163e-07, "loss": 0.63, "num_input_tokens_seen": 1021313024, "step": 487 }, { "epoch": 0.13235692975318689, "grad_norm": 17.620363235473633, "learning_rate": 9.83478113921236e-07, "loss": 0.8202, "num_input_tokens_seen": 1023410176, "step": 488 }, { "epoch": 0.13262815296989422, "grad_norm": 14.268228530883789, "learning_rate": 9.833695564444652e-07, "loss": 0.4924, "num_input_tokens_seen": 1025507328, "step": 489 }, { "epoch": 0.13289937618660158, "grad_norm": 12.854048728942871, "learning_rate": 9.832606502083658e-07, "loss": 0.4732, "num_input_tokens_seen": 1027604480, "step": 490 }, { "epoch": 0.13317059940330891, "grad_norm": 17.436908721923828, "learning_rate": 9.83151395300582e-07, "loss": 0.5987, "num_input_tokens_seen": 1029701632, "step": 491 }, { "epoch": 0.13344182262001628, "grad_norm": 13.574828147888184, "learning_rate": 9.83041791809038e-07, "loss": 0.4847, "num_input_tokens_seen": 1031798784, "step": 492 }, { "epoch": 0.13371304583672364, "grad_norm": 18.534931182861328, "learning_rate": 9.829318398219385e-07, "loss": 0.5305, "num_input_tokens_seen": 1033895936, "step": 493 }, { "epoch": 0.13398426905343097, "grad_norm": 13.970446586608887, "learning_rate": 9.828215394277686e-07, "loss": 0.3694, "num_input_tokens_seen": 1035993088, "step": 494 }, { "epoch": 0.13425549227013833, "grad_norm": 15.922163009643555, "learning_rate": 9.827108907152937e-07, "loss": 0.5241, "num_input_tokens_seen": 1038090240, "step": 495 }, { "epoch": 0.13452671548684567, "grad_norm": 15.5534029006958, "learning_rate": 9.825998937735599e-07, "loss": 0.6606, "num_input_tokens_seen": 1040187392, "step": 496 }, { "epoch": 0.13479793870355303, "grad_norm": 18.83726692199707, "learning_rate": 9.824885486918932e-07, "loss": 0.7119, "num_input_tokens_seen": 1042284544, "step": 497 }, { "epoch": 0.13506916192026036, "grad_norm": 19.1335506439209, "learning_rate": 9.823768555599e-07, "loss": 0.8065, "num_input_tokens_seen": 1044381696, "step": 498 }, { "epoch": 0.13534038513696772, "grad_norm": 13.459641456604004, "learning_rate": 9.822648144674664e-07, "loss": 0.324, "num_input_tokens_seen": 1046478848, "step": 499 }, { "epoch": 0.13561160835367508, "grad_norm": 11.397218704223633, "learning_rate": 9.821524255047592e-07, "loss": 0.3856, "num_input_tokens_seen": 1048576000, "step": 500 }, { "epoch": 0.13588283157038242, "grad_norm": 13.145438194274902, "learning_rate": 9.820396887622245e-07, "loss": 0.4602, "num_input_tokens_seen": 1050673152, "step": 501 }, { "epoch": 0.13615405478708978, "grad_norm": 12.989845275878906, "learning_rate": 9.819266043305887e-07, "loss": 0.3918, "num_input_tokens_seen": 1052770304, "step": 502 }, { "epoch": 0.1364252780037971, "grad_norm": 18.566017150878906, "learning_rate": 9.818131723008576e-07, "loss": 0.5483, "num_input_tokens_seen": 1054867456, "step": 503 }, { "epoch": 0.13669650122050447, "grad_norm": 20.60564613342285, "learning_rate": 9.816993927643174e-07, "loss": 0.8955, "num_input_tokens_seen": 1056964608, "step": 504 }, { "epoch": 0.13696772443721184, "grad_norm": 13.428114891052246, "learning_rate": 9.815852658125332e-07, "loss": 0.4461, "num_input_tokens_seen": 1059061760, "step": 505 }, { "epoch": 0.13723894765391917, "grad_norm": 15.920856475830078, "learning_rate": 9.8147079153735e-07, "loss": 0.6438, "num_input_tokens_seen": 1061158912, "step": 506 }, { "epoch": 0.13751017087062653, "grad_norm": 11.576249122619629, "learning_rate": 9.813559700308925e-07, "loss": 0.3101, "num_input_tokens_seen": 1063256064, "step": 507 }, { "epoch": 0.13778139408733386, "grad_norm": 20.885211944580078, "learning_rate": 9.812408013855646e-07, "loss": 0.7386, "num_input_tokens_seen": 1065353216, "step": 508 }, { "epoch": 0.13805261730404123, "grad_norm": 16.579875946044922, "learning_rate": 9.811252856940496e-07, "loss": 0.6981, "num_input_tokens_seen": 1067450368, "step": 509 }, { "epoch": 0.1383238405207486, "grad_norm": 10.057287216186523, "learning_rate": 9.810094230493104e-07, "loss": 0.2932, "num_input_tokens_seen": 1069547520, "step": 510 }, { "epoch": 0.13859506373745592, "grad_norm": 17.37000846862793, "learning_rate": 9.808932135445885e-07, "loss": 0.5647, "num_input_tokens_seen": 1071644672, "step": 511 }, { "epoch": 0.13886628695416328, "grad_norm": 15.917064666748047, "learning_rate": 9.807766572734052e-07, "loss": 0.7026, "num_input_tokens_seen": 1073741824, "step": 512 }, { "epoch": 0.13913751017087062, "grad_norm": 11.753175735473633, "learning_rate": 9.806597543295603e-07, "loss": 0.416, "num_input_tokens_seen": 1075838976, "step": 513 }, { "epoch": 0.13940873338757798, "grad_norm": 18.979516983032227, "learning_rate": 9.80542504807133e-07, "loss": 0.5972, "num_input_tokens_seen": 1077936128, "step": 514 }, { "epoch": 0.13967995660428534, "grad_norm": 28.44046401977539, "learning_rate": 9.804249088004812e-07, "loss": 0.4001, "num_input_tokens_seen": 1080033280, "step": 515 }, { "epoch": 0.13995117982099267, "grad_norm": 12.617019653320312, "learning_rate": 9.803069664042416e-07, "loss": 0.3327, "num_input_tokens_seen": 1082130432, "step": 516 }, { "epoch": 0.14022240303770003, "grad_norm": 14.718525886535645, "learning_rate": 9.801886777133297e-07, "loss": 0.5243, "num_input_tokens_seen": 1084227584, "step": 517 }, { "epoch": 0.14049362625440737, "grad_norm": 17.5297794342041, "learning_rate": 9.8007004282294e-07, "loss": 0.5279, "num_input_tokens_seen": 1086324736, "step": 518 }, { "epoch": 0.14076484947111473, "grad_norm": 13.261266708374023, "learning_rate": 9.799510618285454e-07, "loss": 0.4663, "num_input_tokens_seen": 1088421888, "step": 519 }, { "epoch": 0.1410360726878221, "grad_norm": 22.933935165405273, "learning_rate": 9.79831734825897e-07, "loss": 0.7519, "num_input_tokens_seen": 1090519040, "step": 520 }, { "epoch": 0.14130729590452942, "grad_norm": 20.257604598999023, "learning_rate": 9.797120619110245e-07, "loss": 0.7251, "num_input_tokens_seen": 1092616192, "step": 521 }, { "epoch": 0.14157851912123678, "grad_norm": 19.605735778808594, "learning_rate": 9.795920431802365e-07, "loss": 0.7534, "num_input_tokens_seen": 1094713344, "step": 522 }, { "epoch": 0.14184974233794412, "grad_norm": 13.443857192993164, "learning_rate": 9.794716787301194e-07, "loss": 0.5247, "num_input_tokens_seen": 1096810496, "step": 523 }, { "epoch": 0.14212096555465148, "grad_norm": 21.489839553833008, "learning_rate": 9.793509686575378e-07, "loss": 0.7223, "num_input_tokens_seen": 1098907648, "step": 524 }, { "epoch": 0.14239218877135884, "grad_norm": 15.88458251953125, "learning_rate": 9.792299130596346e-07, "loss": 0.6552, "num_input_tokens_seen": 1101004800, "step": 525 }, { "epoch": 0.14266341198806617, "grad_norm": 20.70093536376953, "learning_rate": 9.79108512033831e-07, "loss": 0.8404, "num_input_tokens_seen": 1103101952, "step": 526 }, { "epoch": 0.14293463520477354, "grad_norm": 17.892988204956055, "learning_rate": 9.789867656778254e-07, "loss": 0.6306, "num_input_tokens_seen": 1105199104, "step": 527 }, { "epoch": 0.14320585842148087, "grad_norm": 16.286060333251953, "learning_rate": 9.788646740895952e-07, "loss": 0.6665, "num_input_tokens_seen": 1107296256, "step": 528 }, { "epoch": 0.14347708163818823, "grad_norm": 14.654132843017578, "learning_rate": 9.787422373673945e-07, "loss": 0.5283, "num_input_tokens_seen": 1109393408, "step": 529 }, { "epoch": 0.1437483048548956, "grad_norm": 16.374067306518555, "learning_rate": 9.786194556097564e-07, "loss": 0.6534, "num_input_tokens_seen": 1111490560, "step": 530 }, { "epoch": 0.14401952807160293, "grad_norm": 19.840892791748047, "learning_rate": 9.784963289154902e-07, "loss": 0.928, "num_input_tokens_seen": 1113587712, "step": 531 }, { "epoch": 0.1442907512883103, "grad_norm": 12.525459289550781, "learning_rate": 9.783728573836843e-07, "loss": 0.4473, "num_input_tokens_seen": 1115684864, "step": 532 }, { "epoch": 0.14456197450501762, "grad_norm": 15.423880577087402, "learning_rate": 9.782490411137035e-07, "loss": 0.387, "num_input_tokens_seen": 1117782016, "step": 533 }, { "epoch": 0.14483319772172498, "grad_norm": 14.608306884765625, "learning_rate": 9.781248802051904e-07, "loss": 0.5885, "num_input_tokens_seen": 1119879168, "step": 534 }, { "epoch": 0.14510442093843234, "grad_norm": 17.79987144470215, "learning_rate": 9.780003747580651e-07, "loss": 0.674, "num_input_tokens_seen": 1121976320, "step": 535 }, { "epoch": 0.14537564415513968, "grad_norm": 11.311891555786133, "learning_rate": 9.778755248725248e-07, "loss": 0.4104, "num_input_tokens_seen": 1124073472, "step": 536 }, { "epoch": 0.14564686737184704, "grad_norm": 13.756656646728516, "learning_rate": 9.77750330649044e-07, "loss": 0.5469, "num_input_tokens_seen": 1126170624, "step": 537 }, { "epoch": 0.14591809058855437, "grad_norm": 19.89360809326172, "learning_rate": 9.776247921883743e-07, "loss": 0.9223, "num_input_tokens_seen": 1128267776, "step": 538 }, { "epoch": 0.14618931380526173, "grad_norm": 14.971158027648926, "learning_rate": 9.774989095915442e-07, "loss": 0.491, "num_input_tokens_seen": 1130364928, "step": 539 }, { "epoch": 0.14646053702196907, "grad_norm": 15.824216842651367, "learning_rate": 9.77372682959859e-07, "loss": 0.6727, "num_input_tokens_seen": 1132462080, "step": 540 }, { "epoch": 0.14673176023867643, "grad_norm": 16.35426902770996, "learning_rate": 9.772461123949015e-07, "loss": 0.7117, "num_input_tokens_seen": 1134559232, "step": 541 }, { "epoch": 0.1470029834553838, "grad_norm": 17.143768310546875, "learning_rate": 9.771191979985303e-07, "loss": 0.772, "num_input_tokens_seen": 1136656384, "step": 542 }, { "epoch": 0.14727420667209112, "grad_norm": 11.403589248657227, "learning_rate": 9.76991939872882e-07, "loss": 0.3995, "num_input_tokens_seen": 1138753536, "step": 543 }, { "epoch": 0.14754542988879848, "grad_norm": 13.465624809265137, "learning_rate": 9.768643381203686e-07, "loss": 0.5924, "num_input_tokens_seen": 1140850688, "step": 544 }, { "epoch": 0.14781665310550582, "grad_norm": 15.813559532165527, "learning_rate": 9.767363928436793e-07, "loss": 0.5763, "num_input_tokens_seen": 1142947840, "step": 545 }, { "epoch": 0.14808787632221318, "grad_norm": 17.59358787536621, "learning_rate": 9.766081041457795e-07, "loss": 0.6967, "num_input_tokens_seen": 1145044992, "step": 546 }, { "epoch": 0.14835909953892054, "grad_norm": 15.673956871032715, "learning_rate": 9.764794721299113e-07, "loss": 0.7236, "num_input_tokens_seen": 1147142144, "step": 547 }, { "epoch": 0.14863032275562787, "grad_norm": 13.424758911132812, "learning_rate": 9.763504968995927e-07, "loss": 0.4046, "num_input_tokens_seen": 1149239296, "step": 548 }, { "epoch": 0.14890154597233524, "grad_norm": 18.995798110961914, "learning_rate": 9.762211785586178e-07, "loss": 0.781, "num_input_tokens_seen": 1151336448, "step": 549 }, { "epoch": 0.14917276918904257, "grad_norm": 12.103052139282227, "learning_rate": 9.760915172110574e-07, "loss": 0.3885, "num_input_tokens_seen": 1153433600, "step": 550 }, { "epoch": 0.14944399240574993, "grad_norm": 12.997087478637695, "learning_rate": 9.759615129612579e-07, "loss": 0.3381, "num_input_tokens_seen": 1155530752, "step": 551 }, { "epoch": 0.1497152156224573, "grad_norm": 15.175359725952148, "learning_rate": 9.75831165913842e-07, "loss": 0.5939, "num_input_tokens_seen": 1157627904, "step": 552 }, { "epoch": 0.14998643883916463, "grad_norm": 15.981290817260742, "learning_rate": 9.757004761737077e-07, "loss": 0.5293, "num_input_tokens_seen": 1159725056, "step": 553 }, { "epoch": 0.150257662055872, "grad_norm": 11.93864631652832, "learning_rate": 9.755694438460293e-07, "loss": 0.4189, "num_input_tokens_seen": 1161822208, "step": 554 }, { "epoch": 0.15052888527257932, "grad_norm": 15.042703628540039, "learning_rate": 9.754380690362565e-07, "loss": 0.4993, "num_input_tokens_seen": 1163919360, "step": 555 }, { "epoch": 0.15080010848928668, "grad_norm": 14.069710731506348, "learning_rate": 9.75306351850115e-07, "loss": 0.472, "num_input_tokens_seen": 1166016512, "step": 556 }, { "epoch": 0.15107133170599404, "grad_norm": 15.970101356506348, "learning_rate": 9.751742923936055e-07, "loss": 0.4091, "num_input_tokens_seen": 1168113664, "step": 557 }, { "epoch": 0.15134255492270138, "grad_norm": 16.250965118408203, "learning_rate": 9.75041890773005e-07, "loss": 0.5406, "num_input_tokens_seen": 1170210816, "step": 558 }, { "epoch": 0.15161377813940874, "grad_norm": 17.741085052490234, "learning_rate": 9.749091470948643e-07, "loss": 0.4236, "num_input_tokens_seen": 1172307968, "step": 559 }, { "epoch": 0.15188500135611607, "grad_norm": 13.735379219055176, "learning_rate": 9.747760614660111e-07, "loss": 0.4524, "num_input_tokens_seen": 1174405120, "step": 560 }, { "epoch": 0.15215622457282343, "grad_norm": 14.853401184082031, "learning_rate": 9.746426339935477e-07, "loss": 0.4694, "num_input_tokens_seen": 1176502272, "step": 561 }, { "epoch": 0.1524274477895308, "grad_norm": 11.256040573120117, "learning_rate": 9.745088647848515e-07, "loss": 0.32, "num_input_tokens_seen": 1178599424, "step": 562 }, { "epoch": 0.15269867100623813, "grad_norm": 21.761676788330078, "learning_rate": 9.743747539475744e-07, "loss": 0.7391, "num_input_tokens_seen": 1180696576, "step": 563 }, { "epoch": 0.1529698942229455, "grad_norm": 15.350510597229004, "learning_rate": 9.74240301589644e-07, "loss": 0.5598, "num_input_tokens_seen": 1182793728, "step": 564 }, { "epoch": 0.15324111743965282, "grad_norm": 12.007080078125, "learning_rate": 9.741055078192626e-07, "loss": 0.4444, "num_input_tokens_seen": 1184890880, "step": 565 }, { "epoch": 0.15351234065636019, "grad_norm": 12.715327262878418, "learning_rate": 9.739703727449068e-07, "loss": 0.4789, "num_input_tokens_seen": 1186988032, "step": 566 }, { "epoch": 0.15378356387306755, "grad_norm": 10.164214134216309, "learning_rate": 9.738348964753283e-07, "loss": 0.3289, "num_input_tokens_seen": 1189085184, "step": 567 }, { "epoch": 0.15405478708977488, "grad_norm": 15.66612720489502, "learning_rate": 9.736990791195532e-07, "loss": 0.4524, "num_input_tokens_seen": 1191182336, "step": 568 }, { "epoch": 0.15432601030648224, "grad_norm": 12.935588836669922, "learning_rate": 9.735629207868824e-07, "loss": 0.3797, "num_input_tokens_seen": 1193279488, "step": 569 }, { "epoch": 0.15459723352318958, "grad_norm": 14.650870323181152, "learning_rate": 9.734264215868904e-07, "loss": 0.5074, "num_input_tokens_seen": 1195376640, "step": 570 }, { "epoch": 0.15486845673989694, "grad_norm": 16.69452667236328, "learning_rate": 9.73289581629427e-07, "loss": 0.4658, "num_input_tokens_seen": 1197473792, "step": 571 }, { "epoch": 0.1551396799566043, "grad_norm": 19.8432674407959, "learning_rate": 9.73152401024616e-07, "loss": 0.7099, "num_input_tokens_seen": 1199570944, "step": 572 }, { "epoch": 0.15541090317331163, "grad_norm": 17.770639419555664, "learning_rate": 9.730148798828543e-07, "loss": 0.5642, "num_input_tokens_seen": 1201668096, "step": 573 }, { "epoch": 0.155682126390019, "grad_norm": 20.076339721679688, "learning_rate": 9.728770183148143e-07, "loss": 0.7079, "num_input_tokens_seen": 1203765248, "step": 574 }, { "epoch": 0.15595334960672633, "grad_norm": 18.26179313659668, "learning_rate": 9.727388164314415e-07, "loss": 0.721, "num_input_tokens_seen": 1205862400, "step": 575 }, { "epoch": 0.1562245728234337, "grad_norm": 21.209247589111328, "learning_rate": 9.72600274343956e-07, "loss": 0.5755, "num_input_tokens_seen": 1207959552, "step": 576 }, { "epoch": 0.15649579604014105, "grad_norm": 16.366296768188477, "learning_rate": 9.724613921638506e-07, "loss": 0.3086, "num_input_tokens_seen": 1210056704, "step": 577 }, { "epoch": 0.15676701925684838, "grad_norm": 16.17742919921875, "learning_rate": 9.723221700028928e-07, "loss": 0.5038, "num_input_tokens_seen": 1212153856, "step": 578 }, { "epoch": 0.15703824247355574, "grad_norm": 19.471908569335938, "learning_rate": 9.72182607973123e-07, "loss": 0.6945, "num_input_tokens_seen": 1214251008, "step": 579 }, { "epoch": 0.15730946569026308, "grad_norm": 23.296817779541016, "learning_rate": 9.720427061868558e-07, "loss": 0.6556, "num_input_tokens_seen": 1216348160, "step": 580 }, { "epoch": 0.15758068890697044, "grad_norm": 11.207878112792969, "learning_rate": 9.71902464756678e-07, "loss": 0.3241, "num_input_tokens_seen": 1218445312, "step": 581 }, { "epoch": 0.15785191212367777, "grad_norm": 19.07700538635254, "learning_rate": 9.717618837954517e-07, "loss": 0.5486, "num_input_tokens_seen": 1220542464, "step": 582 }, { "epoch": 0.15812313534038513, "grad_norm": 13.274455070495605, "learning_rate": 9.716209634163102e-07, "loss": 0.5149, "num_input_tokens_seen": 1222639616, "step": 583 }, { "epoch": 0.1583943585570925, "grad_norm": 13.982866287231445, "learning_rate": 9.714797037326616e-07, "loss": 0.4192, "num_input_tokens_seen": 1224736768, "step": 584 }, { "epoch": 0.15866558177379983, "grad_norm": 17.50905990600586, "learning_rate": 9.713381048581855e-07, "loss": 0.7549, "num_input_tokens_seen": 1226833920, "step": 585 }, { "epoch": 0.1589368049905072, "grad_norm": 12.565037727355957, "learning_rate": 9.71196166906836e-07, "loss": 0.4246, "num_input_tokens_seen": 1228931072, "step": 586 }, { "epoch": 0.15920802820721452, "grad_norm": 14.987252235412598, "learning_rate": 9.71053889992839e-07, "loss": 0.6191, "num_input_tokens_seen": 1231028224, "step": 587 }, { "epoch": 0.1594792514239219, "grad_norm": 11.8529691696167, "learning_rate": 9.709112742306936e-07, "loss": 0.4308, "num_input_tokens_seen": 1233125376, "step": 588 }, { "epoch": 0.15975047464062925, "grad_norm": 11.237396240234375, "learning_rate": 9.707683197351715e-07, "loss": 0.375, "num_input_tokens_seen": 1235222528, "step": 589 }, { "epoch": 0.16002169785733658, "grad_norm": 11.192814826965332, "learning_rate": 9.706250266213173e-07, "loss": 0.2555, "num_input_tokens_seen": 1237319680, "step": 590 }, { "epoch": 0.16029292107404394, "grad_norm": 11.35294246673584, "learning_rate": 9.704813950044476e-07, "loss": 0.3542, "num_input_tokens_seen": 1239416832, "step": 591 }, { "epoch": 0.16056414429075128, "grad_norm": 17.085115432739258, "learning_rate": 9.703374250001516e-07, "loss": 0.7244, "num_input_tokens_seen": 1241513984, "step": 592 }, { "epoch": 0.16083536750745864, "grad_norm": 15.55121898651123, "learning_rate": 9.70193116724291e-07, "loss": 0.6381, "num_input_tokens_seen": 1243611136, "step": 593 }, { "epoch": 0.161106590724166, "grad_norm": 17.19528579711914, "learning_rate": 9.700484702929996e-07, "loss": 0.5695, "num_input_tokens_seen": 1245708288, "step": 594 }, { "epoch": 0.16137781394087333, "grad_norm": 14.350486755371094, "learning_rate": 9.699034858226834e-07, "loss": 0.5896, "num_input_tokens_seen": 1247805440, "step": 595 }, { "epoch": 0.1616490371575807, "grad_norm": 11.893250465393066, "learning_rate": 9.697581634300202e-07, "loss": 0.3935, "num_input_tokens_seen": 1249902592, "step": 596 }, { "epoch": 0.16192026037428803, "grad_norm": 17.247047424316406, "learning_rate": 9.6961250323196e-07, "loss": 0.8572, "num_input_tokens_seen": 1251999744, "step": 597 }, { "epoch": 0.1621914835909954, "grad_norm": 22.222002029418945, "learning_rate": 9.69466505345725e-07, "loss": 0.8848, "num_input_tokens_seen": 1254096896, "step": 598 }, { "epoch": 0.16246270680770275, "grad_norm": 17.794937133789062, "learning_rate": 9.69320169888808e-07, "loss": 0.7104, "num_input_tokens_seen": 1256194048, "step": 599 }, { "epoch": 0.16273393002441008, "grad_norm": 13.776861190795898, "learning_rate": 9.691734969789746e-07, "loss": 0.5435, "num_input_tokens_seen": 1258291200, "step": 600 }, { "epoch": 0.16300515324111745, "grad_norm": 15.114326477050781, "learning_rate": 9.690264867342618e-07, "loss": 0.4268, "num_input_tokens_seen": 1260388352, "step": 601 }, { "epoch": 0.16327637645782478, "grad_norm": 14.888233184814453, "learning_rate": 9.688791392729775e-07, "loss": 0.5672, "num_input_tokens_seen": 1262485504, "step": 602 }, { "epoch": 0.16354759967453214, "grad_norm": 12.982342720031738, "learning_rate": 9.687314547137016e-07, "loss": 0.4089, "num_input_tokens_seen": 1264582656, "step": 603 }, { "epoch": 0.1638188228912395, "grad_norm": 11.086604118347168, "learning_rate": 9.685834331752846e-07, "loss": 0.2905, "num_input_tokens_seen": 1266679808, "step": 604 }, { "epoch": 0.16409004610794684, "grad_norm": 20.390928268432617, "learning_rate": 9.684350747768492e-07, "loss": 0.9503, "num_input_tokens_seen": 1268776960, "step": 605 }, { "epoch": 0.1643612693246542, "grad_norm": 35.44593048095703, "learning_rate": 9.68286379637788e-07, "loss": 0.5335, "num_input_tokens_seen": 1270874112, "step": 606 }, { "epoch": 0.16463249254136153, "grad_norm": 11.213769912719727, "learning_rate": 9.681373478777654e-07, "loss": 0.3713, "num_input_tokens_seen": 1272971264, "step": 607 }, { "epoch": 0.1649037157580689, "grad_norm": 11.05787181854248, "learning_rate": 9.679879796167166e-07, "loss": 0.4049, "num_input_tokens_seen": 1275068416, "step": 608 }, { "epoch": 0.16517493897477625, "grad_norm": 15.489423751831055, "learning_rate": 9.678382749748477e-07, "loss": 0.5591, "num_input_tokens_seen": 1277165568, "step": 609 }, { "epoch": 0.1654461621914836, "grad_norm": 16.684755325317383, "learning_rate": 9.676882340726345e-07, "loss": 0.592, "num_input_tokens_seen": 1279262720, "step": 610 }, { "epoch": 0.16571738540819095, "grad_norm": 17.940027236938477, "learning_rate": 9.675378570308253e-07, "loss": 0.6475, "num_input_tokens_seen": 1281359872, "step": 611 }, { "epoch": 0.16598860862489828, "grad_norm": 11.783378601074219, "learning_rate": 9.673871439704369e-07, "loss": 0.3884, "num_input_tokens_seen": 1283457024, "step": 612 }, { "epoch": 0.16625983184160564, "grad_norm": 16.390104293823242, "learning_rate": 9.672360950127578e-07, "loss": 0.5195, "num_input_tokens_seen": 1285554176, "step": 613 }, { "epoch": 0.166531055058313, "grad_norm": 15.546297073364258, "learning_rate": 9.670847102793464e-07, "loss": 0.3534, "num_input_tokens_seen": 1287651328, "step": 614 }, { "epoch": 0.16680227827502034, "grad_norm": 15.699833869934082, "learning_rate": 9.669329898920317e-07, "loss": 0.5498, "num_input_tokens_seen": 1289748480, "step": 615 }, { "epoch": 0.1670735014917277, "grad_norm": 19.23993682861328, "learning_rate": 9.66780933972912e-07, "loss": 0.6282, "num_input_tokens_seen": 1291845632, "step": 616 }, { "epoch": 0.16734472470843503, "grad_norm": 9.65346908569336, "learning_rate": 9.666285426443564e-07, "loss": 0.2954, "num_input_tokens_seen": 1293942784, "step": 617 }, { "epoch": 0.1676159479251424, "grad_norm": 11.71711254119873, "learning_rate": 9.664758160290036e-07, "loss": 0.444, "num_input_tokens_seen": 1296039936, "step": 618 }, { "epoch": 0.16788717114184976, "grad_norm": 20.66088104248047, "learning_rate": 9.66322754249762e-07, "loss": 0.6955, "num_input_tokens_seen": 1298137088, "step": 619 }, { "epoch": 0.1681583943585571, "grad_norm": 20.292829513549805, "learning_rate": 9.661693574298102e-07, "loss": 0.8379, "num_input_tokens_seen": 1300234240, "step": 620 }, { "epoch": 0.16842961757526445, "grad_norm": 12.26496410369873, "learning_rate": 9.66015625692596e-07, "loss": 0.3796, "num_input_tokens_seen": 1302331392, "step": 621 }, { "epoch": 0.16870084079197178, "grad_norm": 13.123128890991211, "learning_rate": 9.658615591618366e-07, "loss": 0.2908, "num_input_tokens_seen": 1304428544, "step": 622 }, { "epoch": 0.16897206400867915, "grad_norm": 13.454907417297363, "learning_rate": 9.657071579615191e-07, "loss": 0.4747, "num_input_tokens_seen": 1306525696, "step": 623 }, { "epoch": 0.16924328722538648, "grad_norm": 17.779613494873047, "learning_rate": 9.655524222159e-07, "loss": 0.745, "num_input_tokens_seen": 1308622848, "step": 624 }, { "epoch": 0.16951451044209384, "grad_norm": 18.326885223388672, "learning_rate": 9.653973520495042e-07, "loss": 0.6889, "num_input_tokens_seen": 1310720000, "step": 625 }, { "epoch": 0.1697857336588012, "grad_norm": 15.276704788208008, "learning_rate": 9.652419475871267e-07, "loss": 0.4917, "num_input_tokens_seen": 1312817152, "step": 626 }, { "epoch": 0.17005695687550854, "grad_norm": 15.219459533691406, "learning_rate": 9.650862089538307e-07, "loss": 0.4999, "num_input_tokens_seen": 1314914304, "step": 627 }, { "epoch": 0.1703281800922159, "grad_norm": 16.05649757385254, "learning_rate": 9.64930136274949e-07, "loss": 0.5888, "num_input_tokens_seen": 1317011456, "step": 628 }, { "epoch": 0.17059940330892323, "grad_norm": 16.75012969970703, "learning_rate": 9.647737296760828e-07, "loss": 0.6463, "num_input_tokens_seen": 1319108608, "step": 629 }, { "epoch": 0.1708706265256306, "grad_norm": 14.331846237182617, "learning_rate": 9.646169892831025e-07, "loss": 0.4548, "num_input_tokens_seen": 1321205760, "step": 630 }, { "epoch": 0.17114184974233795, "grad_norm": 14.530692100524902, "learning_rate": 9.644599152221465e-07, "loss": 0.5452, "num_input_tokens_seen": 1323302912, "step": 631 }, { "epoch": 0.1714130729590453, "grad_norm": 13.921281814575195, "learning_rate": 9.643025076196219e-07, "loss": 0.599, "num_input_tokens_seen": 1325400064, "step": 632 }, { "epoch": 0.17168429617575265, "grad_norm": 14.689400672912598, "learning_rate": 9.641447666022048e-07, "loss": 0.5125, "num_input_tokens_seen": 1327497216, "step": 633 }, { "epoch": 0.17195551939245998, "grad_norm": 21.08767318725586, "learning_rate": 9.639866922968387e-07, "loss": 0.7141, "num_input_tokens_seen": 1329594368, "step": 634 }, { "epoch": 0.17222674260916734, "grad_norm": 21.691171646118164, "learning_rate": 9.638282848307361e-07, "loss": 0.7668, "num_input_tokens_seen": 1331691520, "step": 635 }, { "epoch": 0.1724979658258747, "grad_norm": 18.770170211791992, "learning_rate": 9.636695443313773e-07, "loss": 0.6828, "num_input_tokens_seen": 1333788672, "step": 636 }, { "epoch": 0.17276918904258204, "grad_norm": 18.908126831054688, "learning_rate": 9.635104709265103e-07, "loss": 0.8134, "num_input_tokens_seen": 1335885824, "step": 637 }, { "epoch": 0.1730404122592894, "grad_norm": 11.843361854553223, "learning_rate": 9.633510647441518e-07, "loss": 0.3847, "num_input_tokens_seen": 1337982976, "step": 638 }, { "epoch": 0.17331163547599673, "grad_norm": 19.700693130493164, "learning_rate": 9.631913259125854e-07, "loss": 0.9532, "num_input_tokens_seen": 1340080128, "step": 639 }, { "epoch": 0.1735828586927041, "grad_norm": 10.508553504943848, "learning_rate": 9.630312545603631e-07, "loss": 0.3257, "num_input_tokens_seen": 1342177280, "step": 640 }, { "epoch": 0.17385408190941146, "grad_norm": 19.769977569580078, "learning_rate": 9.628708508163041e-07, "loss": 0.7283, "num_input_tokens_seen": 1344274432, "step": 641 }, { "epoch": 0.1741253051261188, "grad_norm": 10.588939666748047, "learning_rate": 9.627101148094952e-07, "loss": 0.3603, "num_input_tokens_seen": 1346371584, "step": 642 }, { "epoch": 0.17439652834282615, "grad_norm": 16.81378936767578, "learning_rate": 9.625490466692906e-07, "loss": 0.7399, "num_input_tokens_seen": 1348468736, "step": 643 }, { "epoch": 0.17466775155953349, "grad_norm": 13.483939170837402, "learning_rate": 9.623876465253122e-07, "loss": 0.4786, "num_input_tokens_seen": 1350565888, "step": 644 }, { "epoch": 0.17493897477624085, "grad_norm": 16.27707290649414, "learning_rate": 9.622259145074482e-07, "loss": 0.4953, "num_input_tokens_seen": 1352663040, "step": 645 }, { "epoch": 0.1752101979929482, "grad_norm": 15.7435941696167, "learning_rate": 9.620638507458547e-07, "loss": 0.6285, "num_input_tokens_seen": 1354760192, "step": 646 }, { "epoch": 0.17548142120965554, "grad_norm": 20.336198806762695, "learning_rate": 9.619014553709542e-07, "loss": 0.8913, "num_input_tokens_seen": 1356857344, "step": 647 }, { "epoch": 0.1757526444263629, "grad_norm": 25.84923553466797, "learning_rate": 9.617387285134364e-07, "loss": 0.9168, "num_input_tokens_seen": 1358954496, "step": 648 }, { "epoch": 0.17602386764307024, "grad_norm": 12.170639991760254, "learning_rate": 9.615756703042575e-07, "loss": 0.4383, "num_input_tokens_seen": 1361051648, "step": 649 }, { "epoch": 0.1762950908597776, "grad_norm": 16.723989486694336, "learning_rate": 9.61412280874641e-07, "loss": 0.4632, "num_input_tokens_seen": 1363148800, "step": 650 }, { "epoch": 0.17656631407648496, "grad_norm": 16.861356735229492, "learning_rate": 9.612485603560763e-07, "loss": 0.8342, "num_input_tokens_seen": 1365245952, "step": 651 }, { "epoch": 0.1768375372931923, "grad_norm": 19.289173126220703, "learning_rate": 9.610845088803194e-07, "loss": 0.5916, "num_input_tokens_seen": 1367343104, "step": 652 }, { "epoch": 0.17710876050989965, "grad_norm": 17.8167781829834, "learning_rate": 9.609201265793927e-07, "loss": 0.5876, "num_input_tokens_seen": 1369440256, "step": 653 }, { "epoch": 0.177379983726607, "grad_norm": 13.8448486328125, "learning_rate": 9.607554135855847e-07, "loss": 0.524, "num_input_tokens_seen": 1371537408, "step": 654 }, { "epoch": 0.17765120694331435, "grad_norm": 14.017175674438477, "learning_rate": 9.605903700314503e-07, "loss": 0.4775, "num_input_tokens_seen": 1373634560, "step": 655 }, { "epoch": 0.1779224301600217, "grad_norm": 11.310769081115723, "learning_rate": 9.604249960498102e-07, "loss": 0.359, "num_input_tokens_seen": 1375731712, "step": 656 }, { "epoch": 0.17819365337672904, "grad_norm": 13.998455047607422, "learning_rate": 9.602592917737512e-07, "loss": 0.4158, "num_input_tokens_seen": 1377828864, "step": 657 }, { "epoch": 0.1784648765934364, "grad_norm": 21.207416534423828, "learning_rate": 9.600932573366254e-07, "loss": 0.8973, "num_input_tokens_seen": 1379926016, "step": 658 }, { "epoch": 0.17873609981014374, "grad_norm": 12.346809387207031, "learning_rate": 9.599268928720518e-07, "loss": 0.3974, "num_input_tokens_seen": 1382023168, "step": 659 }, { "epoch": 0.1790073230268511, "grad_norm": 16.258766174316406, "learning_rate": 9.597601985139132e-07, "loss": 0.3364, "num_input_tokens_seen": 1384120320, "step": 660 }, { "epoch": 0.17927854624355846, "grad_norm": 16.505359649658203, "learning_rate": 9.595931743963596e-07, "loss": 0.5018, "num_input_tokens_seen": 1386217472, "step": 661 }, { "epoch": 0.1795497694602658, "grad_norm": 15.472549438476562, "learning_rate": 9.594258206538054e-07, "loss": 0.4393, "num_input_tokens_seen": 1388314624, "step": 662 }, { "epoch": 0.17982099267697316, "grad_norm": 12.654108047485352, "learning_rate": 9.592581374209306e-07, "loss": 0.3907, "num_input_tokens_seen": 1390411776, "step": 663 }, { "epoch": 0.1800922158936805, "grad_norm": 15.311179161071777, "learning_rate": 9.590901248326802e-07, "loss": 0.5991, "num_input_tokens_seen": 1392508928, "step": 664 }, { "epoch": 0.18036343911038785, "grad_norm": 13.300477981567383, "learning_rate": 9.589217830242645e-07, "loss": 0.3853, "num_input_tokens_seen": 1394606080, "step": 665 }, { "epoch": 0.18063466232709519, "grad_norm": 14.14307975769043, "learning_rate": 9.587531121311582e-07, "loss": 0.3291, "num_input_tokens_seen": 1396703232, "step": 666 }, { "epoch": 0.18090588554380255, "grad_norm": 19.638458251953125, "learning_rate": 9.585841122891016e-07, "loss": 0.8024, "num_input_tokens_seen": 1398800384, "step": 667 }, { "epoch": 0.1811771087605099, "grad_norm": 18.304893493652344, "learning_rate": 9.584147836340992e-07, "loss": 0.6279, "num_input_tokens_seen": 1400897536, "step": 668 }, { "epoch": 0.18144833197721724, "grad_norm": 13.643768310546875, "learning_rate": 9.582451263024202e-07, "loss": 0.5066, "num_input_tokens_seen": 1402994688, "step": 669 }, { "epoch": 0.1817195551939246, "grad_norm": 17.128938674926758, "learning_rate": 9.580751404305985e-07, "loss": 0.7491, "num_input_tokens_seen": 1405091840, "step": 670 }, { "epoch": 0.18199077841063194, "grad_norm": 18.64154052734375, "learning_rate": 9.579048261554321e-07, "loss": 0.8889, "num_input_tokens_seen": 1407188992, "step": 671 }, { "epoch": 0.1822620016273393, "grad_norm": 14.319369316101074, "learning_rate": 9.577341836139837e-07, "loss": 0.4579, "num_input_tokens_seen": 1409286144, "step": 672 }, { "epoch": 0.18253322484404666, "grad_norm": 15.52761459350586, "learning_rate": 9.575632129435796e-07, "loss": 0.6158, "num_input_tokens_seen": 1411383296, "step": 673 }, { "epoch": 0.182804448060754, "grad_norm": 11.78895092010498, "learning_rate": 9.573919142818109e-07, "loss": 0.3947, "num_input_tokens_seen": 1413480448, "step": 674 }, { "epoch": 0.18307567127746135, "grad_norm": 17.20797348022461, "learning_rate": 9.572202877665317e-07, "loss": 0.6729, "num_input_tokens_seen": 1415577600, "step": 675 }, { "epoch": 0.1833468944941687, "grad_norm": 10.157732009887695, "learning_rate": 9.57048333535861e-07, "loss": 0.2969, "num_input_tokens_seen": 1417674752, "step": 676 }, { "epoch": 0.18361811771087605, "grad_norm": 20.86275863647461, "learning_rate": 9.568760517281808e-07, "loss": 0.9544, "num_input_tokens_seen": 1419771904, "step": 677 }, { "epoch": 0.1838893409275834, "grad_norm": 14.223817825317383, "learning_rate": 9.56703442482137e-07, "loss": 0.3624, "num_input_tokens_seen": 1421869056, "step": 678 }, { "epoch": 0.18416056414429074, "grad_norm": 16.383438110351562, "learning_rate": 9.565305059366385e-07, "loss": 0.8115, "num_input_tokens_seen": 1423966208, "step": 679 }, { "epoch": 0.1844317873609981, "grad_norm": 12.147848129272461, "learning_rate": 9.563572422308588e-07, "loss": 0.3638, "num_input_tokens_seen": 1426063360, "step": 680 }, { "epoch": 0.18470301057770544, "grad_norm": 15.152525901794434, "learning_rate": 9.561836515042336e-07, "loss": 0.5524, "num_input_tokens_seen": 1428160512, "step": 681 }, { "epoch": 0.1849742337944128, "grad_norm": 13.729914665222168, "learning_rate": 9.56009733896462e-07, "loss": 0.4642, "num_input_tokens_seen": 1430257664, "step": 682 }, { "epoch": 0.18524545701112016, "grad_norm": 16.161523818969727, "learning_rate": 9.558354895475065e-07, "loss": 0.6551, "num_input_tokens_seen": 1432354816, "step": 683 }, { "epoch": 0.1855166802278275, "grad_norm": 22.390674591064453, "learning_rate": 9.55660918597592e-07, "loss": 0.9867, "num_input_tokens_seen": 1434451968, "step": 684 }, { "epoch": 0.18578790344453486, "grad_norm": 17.674570083618164, "learning_rate": 9.55486021187207e-07, "loss": 0.6112, "num_input_tokens_seen": 1436549120, "step": 685 }, { "epoch": 0.1860591266612422, "grad_norm": 11.00310230255127, "learning_rate": 9.553107974571018e-07, "loss": 0.3261, "num_input_tokens_seen": 1438646272, "step": 686 }, { "epoch": 0.18633034987794955, "grad_norm": 18.365558624267578, "learning_rate": 9.551352475482902e-07, "loss": 0.7754, "num_input_tokens_seen": 1440743424, "step": 687 }, { "epoch": 0.18660157309465691, "grad_norm": 13.15377140045166, "learning_rate": 9.549593716020478e-07, "loss": 0.4411, "num_input_tokens_seen": 1442840576, "step": 688 }, { "epoch": 0.18687279631136425, "grad_norm": 11.530652046203613, "learning_rate": 9.54783169759913e-07, "loss": 0.4043, "num_input_tokens_seen": 1444937728, "step": 689 }, { "epoch": 0.1871440195280716, "grad_norm": 14.537562370300293, "learning_rate": 9.546066421636867e-07, "loss": 0.389, "num_input_tokens_seen": 1447034880, "step": 690 }, { "epoch": 0.18741524274477894, "grad_norm": 9.434741020202637, "learning_rate": 9.54429788955431e-07, "loss": 0.2483, "num_input_tokens_seen": 1449132032, "step": 691 }, { "epoch": 0.1876864659614863, "grad_norm": 15.039995193481445, "learning_rate": 9.542526102774704e-07, "loss": 0.5747, "num_input_tokens_seen": 1451229184, "step": 692 }, { "epoch": 0.18795768917819367, "grad_norm": 16.6597900390625, "learning_rate": 9.540751062723923e-07, "loss": 0.7014, "num_input_tokens_seen": 1453326336, "step": 693 }, { "epoch": 0.188228912394901, "grad_norm": 17.40535545349121, "learning_rate": 9.53897277083045e-07, "loss": 0.7324, "num_input_tokens_seen": 1455423488, "step": 694 }, { "epoch": 0.18850013561160836, "grad_norm": 14.920904159545898, "learning_rate": 9.537191228525382e-07, "loss": 0.4232, "num_input_tokens_seen": 1457520640, "step": 695 }, { "epoch": 0.1887713588283157, "grad_norm": 16.280458450317383, "learning_rate": 9.535406437242444e-07, "loss": 0.5262, "num_input_tokens_seen": 1459617792, "step": 696 }, { "epoch": 0.18904258204502306, "grad_norm": 13.120879173278809, "learning_rate": 9.533618398417962e-07, "loss": 0.534, "num_input_tokens_seen": 1461714944, "step": 697 }, { "epoch": 0.18931380526173042, "grad_norm": 16.71864891052246, "learning_rate": 9.531827113490883e-07, "loss": 0.6104, "num_input_tokens_seen": 1463812096, "step": 698 }, { "epoch": 0.18958502847843775, "grad_norm": 14.800333023071289, "learning_rate": 9.530032583902767e-07, "loss": 0.5911, "num_input_tokens_seen": 1465909248, "step": 699 }, { "epoch": 0.1898562516951451, "grad_norm": 15.830343246459961, "learning_rate": 9.528234811097781e-07, "loss": 0.6818, "num_input_tokens_seen": 1468006400, "step": 700 }, { "epoch": 0.19012747491185245, "grad_norm": 20.223512649536133, "learning_rate": 9.526433796522702e-07, "loss": 0.6553, "num_input_tokens_seen": 1470103552, "step": 701 }, { "epoch": 0.1903986981285598, "grad_norm": 12.725762367248535, "learning_rate": 9.524629541626925e-07, "loss": 0.4537, "num_input_tokens_seen": 1472200704, "step": 702 }, { "epoch": 0.19066992134526717, "grad_norm": 18.15487289428711, "learning_rate": 9.522822047862438e-07, "loss": 0.6253, "num_input_tokens_seen": 1474297856, "step": 703 }, { "epoch": 0.1909411445619745, "grad_norm": 17.46477699279785, "learning_rate": 9.521011316683849e-07, "loss": 0.6778, "num_input_tokens_seen": 1476395008, "step": 704 }, { "epoch": 0.19121236777868186, "grad_norm": 16.60508155822754, "learning_rate": 9.519197349548364e-07, "loss": 0.6406, "num_input_tokens_seen": 1478492160, "step": 705 }, { "epoch": 0.1914835909953892, "grad_norm": 14.466193199157715, "learning_rate": 9.517380147915791e-07, "loss": 0.4449, "num_input_tokens_seen": 1480589312, "step": 706 }, { "epoch": 0.19175481421209656, "grad_norm": 16.949872970581055, "learning_rate": 9.515559713248549e-07, "loss": 0.6814, "num_input_tokens_seen": 1482686464, "step": 707 }, { "epoch": 0.1920260374288039, "grad_norm": 9.922392845153809, "learning_rate": 9.513736047011653e-07, "loss": 0.3183, "num_input_tokens_seen": 1484783616, "step": 708 }, { "epoch": 0.19229726064551125, "grad_norm": 15.374136924743652, "learning_rate": 9.511909150672721e-07, "loss": 0.6533, "num_input_tokens_seen": 1486880768, "step": 709 }, { "epoch": 0.19256848386221861, "grad_norm": 16.63404083251953, "learning_rate": 9.510079025701967e-07, "loss": 0.5225, "num_input_tokens_seen": 1488977920, "step": 710 }, { "epoch": 0.19283970707892595, "grad_norm": 13.690201759338379, "learning_rate": 9.508245673572209e-07, "loss": 0.3678, "num_input_tokens_seen": 1491075072, "step": 711 }, { "epoch": 0.1931109302956333, "grad_norm": 12.991211891174316, "learning_rate": 9.50640909575886e-07, "loss": 0.352, "num_input_tokens_seen": 1493172224, "step": 712 }, { "epoch": 0.19338215351234064, "grad_norm": 18.690780639648438, "learning_rate": 9.504569293739923e-07, "loss": 0.7666, "num_input_tokens_seen": 1495269376, "step": 713 }, { "epoch": 0.193653376729048, "grad_norm": 15.247132301330566, "learning_rate": 9.502726268996005e-07, "loss": 0.5814, "num_input_tokens_seen": 1497366528, "step": 714 }, { "epoch": 0.19392459994575537, "grad_norm": 10.226189613342285, "learning_rate": 9.5008800230103e-07, "loss": 0.2609, "num_input_tokens_seen": 1499463680, "step": 715 }, { "epoch": 0.1941958231624627, "grad_norm": 19.02218246459961, "learning_rate": 9.499030557268599e-07, "loss": 0.6701, "num_input_tokens_seen": 1501560832, "step": 716 }, { "epoch": 0.19446704637917006, "grad_norm": 14.307744979858398, "learning_rate": 9.497177873259279e-07, "loss": 0.4341, "num_input_tokens_seen": 1503657984, "step": 717 }, { "epoch": 0.1947382695958774, "grad_norm": 15.487143516540527, "learning_rate": 9.495321972473311e-07, "loss": 0.3923, "num_input_tokens_seen": 1505755136, "step": 718 }, { "epoch": 0.19500949281258476, "grad_norm": 11.34209156036377, "learning_rate": 9.493462856404251e-07, "loss": 0.4535, "num_input_tokens_seen": 1507852288, "step": 719 }, { "epoch": 0.19528071602929212, "grad_norm": 13.599776268005371, "learning_rate": 9.491600526548247e-07, "loss": 0.4414, "num_input_tokens_seen": 1509949440, "step": 720 }, { "epoch": 0.19555193924599945, "grad_norm": 12.897449493408203, "learning_rate": 9.489734984404033e-07, "loss": 0.4073, "num_input_tokens_seen": 1512046592, "step": 721 }, { "epoch": 0.1958231624627068, "grad_norm": 16.54181480407715, "learning_rate": 9.487866231472922e-07, "loss": 0.5141, "num_input_tokens_seen": 1514143744, "step": 722 }, { "epoch": 0.19609438567941415, "grad_norm": 9.380181312561035, "learning_rate": 9.48599426925882e-07, "loss": 0.2837, "num_input_tokens_seen": 1516240896, "step": 723 }, { "epoch": 0.1963656088961215, "grad_norm": 17.064441680908203, "learning_rate": 9.484119099268206e-07, "loss": 0.6393, "num_input_tokens_seen": 1518338048, "step": 724 }, { "epoch": 0.19663683211282887, "grad_norm": 12.202215194702148, "learning_rate": 9.482240723010148e-07, "loss": 0.3783, "num_input_tokens_seen": 1520435200, "step": 725 }, { "epoch": 0.1969080553295362, "grad_norm": 20.838542938232422, "learning_rate": 9.480359141996295e-07, "loss": 0.7739, "num_input_tokens_seen": 1522532352, "step": 726 }, { "epoch": 0.19717927854624356, "grad_norm": 13.571410179138184, "learning_rate": 9.478474357740864e-07, "loss": 0.4288, "num_input_tokens_seen": 1524629504, "step": 727 }, { "epoch": 0.1974505017629509, "grad_norm": 13.46505355834961, "learning_rate": 9.476586371760665e-07, "loss": 0.4699, "num_input_tokens_seen": 1526726656, "step": 728 }, { "epoch": 0.19772172497965826, "grad_norm": 16.68222427368164, "learning_rate": 9.474695185575072e-07, "loss": 0.782, "num_input_tokens_seen": 1528823808, "step": 729 }, { "epoch": 0.19799294819636562, "grad_norm": 11.398141860961914, "learning_rate": 9.472800800706044e-07, "loss": 0.3531, "num_input_tokens_seen": 1530920960, "step": 730 }, { "epoch": 0.19826417141307295, "grad_norm": 17.724573135375977, "learning_rate": 9.470903218678108e-07, "loss": 0.8683, "num_input_tokens_seen": 1533018112, "step": 731 }, { "epoch": 0.19853539462978032, "grad_norm": 20.757295608520508, "learning_rate": 9.469002441018366e-07, "loss": 0.8129, "num_input_tokens_seen": 1535115264, "step": 732 }, { "epoch": 0.19880661784648765, "grad_norm": 16.623191833496094, "learning_rate": 9.467098469256488e-07, "loss": 0.5387, "num_input_tokens_seen": 1537212416, "step": 733 }, { "epoch": 0.199077841063195, "grad_norm": 11.821105003356934, "learning_rate": 9.465191304924725e-07, "loss": 0.3834, "num_input_tokens_seen": 1539309568, "step": 734 }, { "epoch": 0.19934906427990237, "grad_norm": 16.763580322265625, "learning_rate": 9.463280949557885e-07, "loss": 0.5618, "num_input_tokens_seen": 1541406720, "step": 735 }, { "epoch": 0.1996202874966097, "grad_norm": 14.250548362731934, "learning_rate": 9.46136740469335e-07, "loss": 0.5418, "num_input_tokens_seen": 1543503872, "step": 736 }, { "epoch": 0.19989151071331707, "grad_norm": 22.913976669311523, "learning_rate": 9.45945067187107e-07, "loss": 0.4692, "num_input_tokens_seen": 1545601024, "step": 737 }, { "epoch": 0.2001627339300244, "grad_norm": 15.227788925170898, "learning_rate": 9.457530752633557e-07, "loss": 0.522, "num_input_tokens_seen": 1547698176, "step": 738 }, { "epoch": 0.20043395714673176, "grad_norm": 12.005767822265625, "learning_rate": 9.455607648525889e-07, "loss": 0.2608, "num_input_tokens_seen": 1549795328, "step": 739 }, { "epoch": 0.20070518036343912, "grad_norm": 11.50250244140625, "learning_rate": 9.45368136109571e-07, "loss": 0.3001, "num_input_tokens_seen": 1551892480, "step": 740 }, { "epoch": 0.20097640358014646, "grad_norm": 19.086029052734375, "learning_rate": 9.451751891893217e-07, "loss": 0.6888, "num_input_tokens_seen": 1553989632, "step": 741 }, { "epoch": 0.20124762679685382, "grad_norm": 14.965021133422852, "learning_rate": 9.449819242471179e-07, "loss": 0.6411, "num_input_tokens_seen": 1556086784, "step": 742 }, { "epoch": 0.20151885001356115, "grad_norm": 12.905089378356934, "learning_rate": 9.447883414384916e-07, "loss": 0.4511, "num_input_tokens_seen": 1558183936, "step": 743 }, { "epoch": 0.2017900732302685, "grad_norm": 15.355229377746582, "learning_rate": 9.445944409192308e-07, "loss": 0.3277, "num_input_tokens_seen": 1560281088, "step": 744 }, { "epoch": 0.20206129644697587, "grad_norm": 12.57699203491211, "learning_rate": 9.444002228453796e-07, "loss": 0.3726, "num_input_tokens_seen": 1562378240, "step": 745 }, { "epoch": 0.2023325196636832, "grad_norm": 17.249740600585938, "learning_rate": 9.442056873732369e-07, "loss": 0.7487, "num_input_tokens_seen": 1564475392, "step": 746 }, { "epoch": 0.20260374288039057, "grad_norm": 19.854642868041992, "learning_rate": 9.440108346593579e-07, "loss": 0.6115, "num_input_tokens_seen": 1566572544, "step": 747 }, { "epoch": 0.2028749660970979, "grad_norm": 21.570199966430664, "learning_rate": 9.438156648605521e-07, "loss": 0.8311, "num_input_tokens_seen": 1568669696, "step": 748 }, { "epoch": 0.20314618931380526, "grad_norm": 14.75474739074707, "learning_rate": 9.436201781338852e-07, "loss": 0.4656, "num_input_tokens_seen": 1570766848, "step": 749 }, { "epoch": 0.20341741253051263, "grad_norm": 11.912012100219727, "learning_rate": 9.434243746366771e-07, "loss": 0.4637, "num_input_tokens_seen": 1572864000, "step": 750 }, { "epoch": 0.20368863574721996, "grad_norm": 18.71771240234375, "learning_rate": 9.432282545265034e-07, "loss": 0.5026, "num_input_tokens_seen": 1574961152, "step": 751 }, { "epoch": 0.20395985896392732, "grad_norm": 11.189618110656738, "learning_rate": 9.430318179611938e-07, "loss": 0.3925, "num_input_tokens_seen": 1577058304, "step": 752 }, { "epoch": 0.20423108218063465, "grad_norm": 15.765860557556152, "learning_rate": 9.42835065098833e-07, "loss": 0.6212, "num_input_tokens_seen": 1579155456, "step": 753 }, { "epoch": 0.20450230539734202, "grad_norm": 19.61184310913086, "learning_rate": 9.426379960977605e-07, "loss": 0.582, "num_input_tokens_seen": 1581252608, "step": 754 }, { "epoch": 0.20477352861404935, "grad_norm": 9.36821174621582, "learning_rate": 9.424406111165697e-07, "loss": 0.2286, "num_input_tokens_seen": 1583349760, "step": 755 }, { "epoch": 0.2050447518307567, "grad_norm": 25.251680374145508, "learning_rate": 9.422429103141084e-07, "loss": 1.0384, "num_input_tokens_seen": 1585446912, "step": 756 }, { "epoch": 0.20531597504746407, "grad_norm": 14.59071159362793, "learning_rate": 9.42044893849479e-07, "loss": 0.3857, "num_input_tokens_seen": 1587544064, "step": 757 }, { "epoch": 0.2055871982641714, "grad_norm": 16.961990356445312, "learning_rate": 9.418465618820374e-07, "loss": 0.6815, "num_input_tokens_seen": 1589641216, "step": 758 }, { "epoch": 0.20585842148087877, "grad_norm": 10.979452133178711, "learning_rate": 9.416479145713936e-07, "loss": 0.3136, "num_input_tokens_seen": 1591738368, "step": 759 }, { "epoch": 0.2061296446975861, "grad_norm": 14.783207893371582, "learning_rate": 9.414489520774114e-07, "loss": 0.5799, "num_input_tokens_seen": 1593835520, "step": 760 }, { "epoch": 0.20640086791429346, "grad_norm": 19.04448890686035, "learning_rate": 9.412496745602084e-07, "loss": 0.6763, "num_input_tokens_seen": 1595932672, "step": 761 }, { "epoch": 0.20667209113100082, "grad_norm": 24.7120418548584, "learning_rate": 9.410500821801556e-07, "loss": 0.6357, "num_input_tokens_seen": 1598029824, "step": 762 }, { "epoch": 0.20694331434770816, "grad_norm": 18.664295196533203, "learning_rate": 9.408501750978769e-07, "loss": 0.7047, "num_input_tokens_seen": 1600126976, "step": 763 }, { "epoch": 0.20721453756441552, "grad_norm": 12.567000389099121, "learning_rate": 9.406499534742503e-07, "loss": 0.3672, "num_input_tokens_seen": 1602224128, "step": 764 }, { "epoch": 0.20748576078112285, "grad_norm": 16.22792625427246, "learning_rate": 9.404494174704068e-07, "loss": 0.4399, "num_input_tokens_seen": 1604321280, "step": 765 }, { "epoch": 0.2077569839978302, "grad_norm": 11.791084289550781, "learning_rate": 9.402485672477296e-07, "loss": 0.3927, "num_input_tokens_seen": 1606418432, "step": 766 }, { "epoch": 0.20802820721453757, "grad_norm": 12.438189506530762, "learning_rate": 9.400474029678555e-07, "loss": 0.3697, "num_input_tokens_seen": 1608515584, "step": 767 }, { "epoch": 0.2082994304312449, "grad_norm": 14.400805473327637, "learning_rate": 9.39845924792674e-07, "loss": 0.5527, "num_input_tokens_seen": 1610612736, "step": 768 }, { "epoch": 0.20857065364795227, "grad_norm": 21.340023040771484, "learning_rate": 9.396441328843268e-07, "loss": 0.7031, "num_input_tokens_seen": 1612709888, "step": 769 }, { "epoch": 0.2088418768646596, "grad_norm": 18.95412254333496, "learning_rate": 9.394420274052088e-07, "loss": 0.6709, "num_input_tokens_seen": 1614807040, "step": 770 }, { "epoch": 0.20911310008136696, "grad_norm": 15.32834243774414, "learning_rate": 9.392396085179662e-07, "loss": 0.4891, "num_input_tokens_seen": 1616904192, "step": 771 }, { "epoch": 0.20938432329807433, "grad_norm": 14.922652244567871, "learning_rate": 9.390368763854985e-07, "loss": 0.5442, "num_input_tokens_seen": 1619001344, "step": 772 }, { "epoch": 0.20965554651478166, "grad_norm": 15.146223068237305, "learning_rate": 9.388338311709566e-07, "loss": 0.4359, "num_input_tokens_seen": 1621098496, "step": 773 }, { "epoch": 0.20992676973148902, "grad_norm": 18.45183753967285, "learning_rate": 9.386304730377437e-07, "loss": 0.7346, "num_input_tokens_seen": 1623195648, "step": 774 }, { "epoch": 0.21019799294819635, "grad_norm": 15.511632919311523, "learning_rate": 9.384268021495145e-07, "loss": 0.4655, "num_input_tokens_seen": 1625292800, "step": 775 }, { "epoch": 0.21046921616490372, "grad_norm": 12.674488067626953, "learning_rate": 9.382228186701756e-07, "loss": 0.4528, "num_input_tokens_seen": 1627389952, "step": 776 }, { "epoch": 0.21074043938161108, "grad_norm": 17.0650634765625, "learning_rate": 9.380185227638854e-07, "loss": 0.4335, "num_input_tokens_seen": 1629487104, "step": 777 }, { "epoch": 0.2110116625983184, "grad_norm": 13.060535430908203, "learning_rate": 9.378139145950532e-07, "loss": 0.4389, "num_input_tokens_seen": 1631584256, "step": 778 }, { "epoch": 0.21128288581502577, "grad_norm": 11.724872589111328, "learning_rate": 9.376089943283398e-07, "loss": 0.326, "num_input_tokens_seen": 1633681408, "step": 779 }, { "epoch": 0.2115541090317331, "grad_norm": 20.703407287597656, "learning_rate": 9.374037621286574e-07, "loss": 0.579, "num_input_tokens_seen": 1635778560, "step": 780 }, { "epoch": 0.21182533224844047, "grad_norm": 13.414408683776855, "learning_rate": 9.371982181611692e-07, "loss": 0.4883, "num_input_tokens_seen": 1637875712, "step": 781 }, { "epoch": 0.21209655546514783, "grad_norm": 15.019268989562988, "learning_rate": 9.369923625912888e-07, "loss": 0.495, "num_input_tokens_seen": 1639972864, "step": 782 }, { "epoch": 0.21236777868185516, "grad_norm": 11.590291976928711, "learning_rate": 9.367861955846813e-07, "loss": 0.3606, "num_input_tokens_seen": 1642070016, "step": 783 }, { "epoch": 0.21263900189856252, "grad_norm": 13.894659996032715, "learning_rate": 9.365797173072619e-07, "loss": 0.4129, "num_input_tokens_seen": 1644167168, "step": 784 }, { "epoch": 0.21291022511526986, "grad_norm": 17.303585052490234, "learning_rate": 9.363729279251965e-07, "loss": 0.634, "num_input_tokens_seen": 1646264320, "step": 785 }, { "epoch": 0.21318144833197722, "grad_norm": 15.712759017944336, "learning_rate": 9.361658276049012e-07, "loss": 0.6141, "num_input_tokens_seen": 1648361472, "step": 786 }, { "epoch": 0.21345267154868458, "grad_norm": 12.570011138916016, "learning_rate": 9.359584165130426e-07, "loss": 0.3697, "num_input_tokens_seen": 1650458624, "step": 787 }, { "epoch": 0.21372389476539191, "grad_norm": 17.42540740966797, "learning_rate": 9.357506948165372e-07, "loss": 0.6994, "num_input_tokens_seen": 1652555776, "step": 788 }, { "epoch": 0.21399511798209928, "grad_norm": 13.891875267028809, "learning_rate": 9.355426626825516e-07, "loss": 0.3642, "num_input_tokens_seen": 1654652928, "step": 789 }, { "epoch": 0.2142663411988066, "grad_norm": 14.662867546081543, "learning_rate": 9.353343202785019e-07, "loss": 0.4218, "num_input_tokens_seen": 1656750080, "step": 790 }, { "epoch": 0.21453756441551397, "grad_norm": 16.362796783447266, "learning_rate": 9.351256677720542e-07, "loss": 0.5427, "num_input_tokens_seen": 1658847232, "step": 791 }, { "epoch": 0.21480878763222133, "grad_norm": 16.475540161132812, "learning_rate": 9.349167053311245e-07, "loss": 0.6179, "num_input_tokens_seen": 1660944384, "step": 792 }, { "epoch": 0.21508001084892867, "grad_norm": 11.967279434204102, "learning_rate": 9.347074331238774e-07, "loss": 0.4276, "num_input_tokens_seen": 1663041536, "step": 793 }, { "epoch": 0.21535123406563603, "grad_norm": 10.945496559143066, "learning_rate": 9.344978513187271e-07, "loss": 0.2828, "num_input_tokens_seen": 1665138688, "step": 794 }, { "epoch": 0.21562245728234336, "grad_norm": 21.34172821044922, "learning_rate": 9.342879600843376e-07, "loss": 0.9658, "num_input_tokens_seen": 1667235840, "step": 795 }, { "epoch": 0.21589368049905072, "grad_norm": 13.481067657470703, "learning_rate": 9.34077759589621e-07, "loss": 0.4159, "num_input_tokens_seen": 1669332992, "step": 796 }, { "epoch": 0.21616490371575806, "grad_norm": 21.213205337524414, "learning_rate": 9.338672500037387e-07, "loss": 0.8269, "num_input_tokens_seen": 1671430144, "step": 797 }, { "epoch": 0.21643612693246542, "grad_norm": 12.533496856689453, "learning_rate": 9.336564314961008e-07, "loss": 0.4823, "num_input_tokens_seen": 1673527296, "step": 798 }, { "epoch": 0.21670735014917278, "grad_norm": 13.096843719482422, "learning_rate": 9.334453042363661e-07, "loss": 0.3558, "num_input_tokens_seen": 1675624448, "step": 799 }, { "epoch": 0.2169785733658801, "grad_norm": 21.356908798217773, "learning_rate": 9.332338683944415e-07, "loss": 0.8749, "num_input_tokens_seen": 1677721600, "step": 800 }, { "epoch": 0.21724979658258747, "grad_norm": 11.876173973083496, "learning_rate": 9.33022124140483e-07, "loss": 0.3332, "num_input_tokens_seen": 1679818752, "step": 801 }, { "epoch": 0.2175210197992948, "grad_norm": 10.453437805175781, "learning_rate": 9.32810071644894e-07, "loss": 0.3932, "num_input_tokens_seen": 1681915904, "step": 802 }, { "epoch": 0.21779224301600217, "grad_norm": 10.991945266723633, "learning_rate": 9.325977110783263e-07, "loss": 0.323, "num_input_tokens_seen": 1684013056, "step": 803 }, { "epoch": 0.21806346623270953, "grad_norm": 15.58512020111084, "learning_rate": 9.323850426116797e-07, "loss": 0.5063, "num_input_tokens_seen": 1686110208, "step": 804 }, { "epoch": 0.21833468944941686, "grad_norm": 18.722299575805664, "learning_rate": 9.321720664161017e-07, "loss": 0.5303, "num_input_tokens_seen": 1688207360, "step": 805 }, { "epoch": 0.21860591266612422, "grad_norm": 17.243942260742188, "learning_rate": 9.319587826629872e-07, "loss": 0.4814, "num_input_tokens_seen": 1690304512, "step": 806 }, { "epoch": 0.21887713588283156, "grad_norm": 14.880188941955566, "learning_rate": 9.317451915239792e-07, "loss": 0.6408, "num_input_tokens_seen": 1692401664, "step": 807 }, { "epoch": 0.21914835909953892, "grad_norm": 14.305965423583984, "learning_rate": 9.315312931709674e-07, "loss": 0.5187, "num_input_tokens_seen": 1694498816, "step": 808 }, { "epoch": 0.21941958231624628, "grad_norm": 18.705963134765625, "learning_rate": 9.313170877760892e-07, "loss": 0.6711, "num_input_tokens_seen": 1696595968, "step": 809 }, { "epoch": 0.21969080553295361, "grad_norm": 14.972651481628418, "learning_rate": 9.311025755117291e-07, "loss": 0.4127, "num_input_tokens_seen": 1698693120, "step": 810 }, { "epoch": 0.21996202874966098, "grad_norm": 35.52400207519531, "learning_rate": 9.308877565505181e-07, "loss": 0.7297, "num_input_tokens_seen": 1700790272, "step": 811 }, { "epoch": 0.2202332519663683, "grad_norm": 11.86293888092041, "learning_rate": 9.306726310653346e-07, "loss": 0.3399, "num_input_tokens_seen": 1702887424, "step": 812 }, { "epoch": 0.22050447518307567, "grad_norm": 15.077362060546875, "learning_rate": 9.304571992293032e-07, "loss": 0.574, "num_input_tokens_seen": 1704984576, "step": 813 }, { "epoch": 0.22077569839978303, "grad_norm": 24.58783531188965, "learning_rate": 9.302414612157954e-07, "loss": 0.935, "num_input_tokens_seen": 1707081728, "step": 814 }, { "epoch": 0.22104692161649037, "grad_norm": 18.85881996154785, "learning_rate": 9.300254171984289e-07, "loss": 0.5214, "num_input_tokens_seen": 1709178880, "step": 815 }, { "epoch": 0.22131814483319773, "grad_norm": 23.3087158203125, "learning_rate": 9.298090673510677e-07, "loss": 0.6932, "num_input_tokens_seen": 1711276032, "step": 816 }, { "epoch": 0.22158936804990506, "grad_norm": 15.919090270996094, "learning_rate": 9.295924118478218e-07, "loss": 0.5998, "num_input_tokens_seen": 1713373184, "step": 817 }, { "epoch": 0.22186059126661242, "grad_norm": 19.33287239074707, "learning_rate": 9.293754508630473e-07, "loss": 0.7676, "num_input_tokens_seen": 1715470336, "step": 818 }, { "epoch": 0.22213181448331978, "grad_norm": 13.031168937683105, "learning_rate": 9.291581845713466e-07, "loss": 0.4777, "num_input_tokens_seen": 1717567488, "step": 819 }, { "epoch": 0.22240303770002712, "grad_norm": 13.561074256896973, "learning_rate": 9.289406131475665e-07, "loss": 0.4002, "num_input_tokens_seen": 1719664640, "step": 820 }, { "epoch": 0.22267426091673448, "grad_norm": 16.580875396728516, "learning_rate": 9.287227367668012e-07, "loss": 0.6063, "num_input_tokens_seen": 1721761792, "step": 821 }, { "epoch": 0.2229454841334418, "grad_norm": 11.243143081665039, "learning_rate": 9.285045556043885e-07, "loss": 0.3497, "num_input_tokens_seen": 1723858944, "step": 822 }, { "epoch": 0.22321670735014917, "grad_norm": 16.73326873779297, "learning_rate": 9.282860698359128e-07, "loss": 0.7562, "num_input_tokens_seen": 1725956096, "step": 823 }, { "epoch": 0.22348793056685654, "grad_norm": 19.721914291381836, "learning_rate": 9.280672796372029e-07, "loss": 0.9825, "num_input_tokens_seen": 1728053248, "step": 824 }, { "epoch": 0.22375915378356387, "grad_norm": 18.111730575561523, "learning_rate": 9.278481851843327e-07, "loss": 0.7322, "num_input_tokens_seen": 1730150400, "step": 825 }, { "epoch": 0.22403037700027123, "grad_norm": 20.524099349975586, "learning_rate": 9.276287866536215e-07, "loss": 0.7254, "num_input_tokens_seen": 1732247552, "step": 826 }, { "epoch": 0.22430160021697856, "grad_norm": 13.87217903137207, "learning_rate": 9.274090842216326e-07, "loss": 0.3707, "num_input_tokens_seen": 1734344704, "step": 827 }, { "epoch": 0.22457282343368593, "grad_norm": 19.11530113220215, "learning_rate": 9.271890780651741e-07, "loss": 0.7017, "num_input_tokens_seen": 1736441856, "step": 828 }, { "epoch": 0.2248440466503933, "grad_norm": 9.669900894165039, "learning_rate": 9.269687683612987e-07, "loss": 0.2848, "num_input_tokens_seen": 1738539008, "step": 829 }, { "epoch": 0.22511526986710062, "grad_norm": 14.931427955627441, "learning_rate": 9.267481552873033e-07, "loss": 0.4288, "num_input_tokens_seen": 1740636160, "step": 830 }, { "epoch": 0.22538649308380798, "grad_norm": 16.54037094116211, "learning_rate": 9.265272390207289e-07, "loss": 0.5282, "num_input_tokens_seen": 1742733312, "step": 831 }, { "epoch": 0.22565771630051532, "grad_norm": 25.891746520996094, "learning_rate": 9.263060197393603e-07, "loss": 0.5665, "num_input_tokens_seen": 1744830464, "step": 832 }, { "epoch": 0.22592893951722268, "grad_norm": 18.841882705688477, "learning_rate": 9.260844976212268e-07, "loss": 0.8732, "num_input_tokens_seen": 1746927616, "step": 833 }, { "epoch": 0.22620016273393004, "grad_norm": 12.339317321777344, "learning_rate": 9.258626728446004e-07, "loss": 0.4302, "num_input_tokens_seen": 1749024768, "step": 834 }, { "epoch": 0.22647138595063737, "grad_norm": 18.42291831970215, "learning_rate": 9.256405455879977e-07, "loss": 0.7744, "num_input_tokens_seen": 1751121920, "step": 835 }, { "epoch": 0.22674260916734473, "grad_norm": 22.923080444335938, "learning_rate": 9.25418116030178e-07, "loss": 0.726, "num_input_tokens_seen": 1753219072, "step": 836 }, { "epoch": 0.22701383238405207, "grad_norm": 12.434412956237793, "learning_rate": 9.251953843501443e-07, "loss": 0.3722, "num_input_tokens_seen": 1755316224, "step": 837 }, { "epoch": 0.22728505560075943, "grad_norm": 18.6683292388916, "learning_rate": 9.249723507271425e-07, "loss": 0.5697, "num_input_tokens_seen": 1757413376, "step": 838 }, { "epoch": 0.22755627881746676, "grad_norm": 14.937889099121094, "learning_rate": 9.247490153406617e-07, "loss": 0.5072, "num_input_tokens_seen": 1759510528, "step": 839 }, { "epoch": 0.22782750203417412, "grad_norm": 12.447941780090332, "learning_rate": 9.245253783704334e-07, "loss": 0.3661, "num_input_tokens_seen": 1761607680, "step": 840 }, { "epoch": 0.22809872525088148, "grad_norm": 11.856642723083496, "learning_rate": 9.243014399964324e-07, "loss": 0.3809, "num_input_tokens_seen": 1763704832, "step": 841 }, { "epoch": 0.22836994846758882, "grad_norm": 10.702392578125, "learning_rate": 9.240772003988758e-07, "loss": 0.3321, "num_input_tokens_seen": 1765801984, "step": 842 }, { "epoch": 0.22864117168429618, "grad_norm": 13.526901245117188, "learning_rate": 9.238526597582229e-07, "loss": 0.5027, "num_input_tokens_seen": 1767899136, "step": 843 }, { "epoch": 0.2289123949010035, "grad_norm": 22.885881423950195, "learning_rate": 9.236278182551758e-07, "loss": 0.7197, "num_input_tokens_seen": 1769996288, "step": 844 }, { "epoch": 0.22918361811771087, "grad_norm": 20.40297508239746, "learning_rate": 9.23402676070678e-07, "loss": 0.5759, "num_input_tokens_seen": 1772093440, "step": 845 }, { "epoch": 0.22945484133441824, "grad_norm": 19.0419921875, "learning_rate": 9.231772333859154e-07, "loss": 0.6163, "num_input_tokens_seen": 1774190592, "step": 846 }, { "epoch": 0.22972606455112557, "grad_norm": 12.630133628845215, "learning_rate": 9.22951490382316e-07, "loss": 0.3334, "num_input_tokens_seen": 1776287744, "step": 847 }, { "epoch": 0.22999728776783293, "grad_norm": 11.405631065368652, "learning_rate": 9.22725447241549e-07, "loss": 0.3219, "num_input_tokens_seen": 1778384896, "step": 848 }, { "epoch": 0.23026851098454026, "grad_norm": 14.75662899017334, "learning_rate": 9.224991041455252e-07, "loss": 0.4205, "num_input_tokens_seen": 1780482048, "step": 849 }, { "epoch": 0.23053973420124763, "grad_norm": 16.446325302124023, "learning_rate": 9.222724612763971e-07, "loss": 0.6833, "num_input_tokens_seen": 1782579200, "step": 850 }, { "epoch": 0.230810957417955, "grad_norm": 14.62348747253418, "learning_rate": 9.220455188165582e-07, "loss": 0.5999, "num_input_tokens_seen": 1784676352, "step": 851 }, { "epoch": 0.23108218063466232, "grad_norm": 20.784975051879883, "learning_rate": 9.218182769486433e-07, "loss": 0.7337, "num_input_tokens_seen": 1786773504, "step": 852 }, { "epoch": 0.23135340385136968, "grad_norm": 16.6190128326416, "learning_rate": 9.215907358555276e-07, "loss": 0.6725, "num_input_tokens_seen": 1788870656, "step": 853 }, { "epoch": 0.23162462706807702, "grad_norm": 19.758882522583008, "learning_rate": 9.213628957203277e-07, "loss": 0.8521, "num_input_tokens_seen": 1790967808, "step": 854 }, { "epoch": 0.23189585028478438, "grad_norm": 21.532390594482422, "learning_rate": 9.21134756726401e-07, "loss": 0.7921, "num_input_tokens_seen": 1793064960, "step": 855 }, { "epoch": 0.23216707350149174, "grad_norm": 19.4443302154541, "learning_rate": 9.209063190573445e-07, "loss": 0.673, "num_input_tokens_seen": 1795162112, "step": 856 }, { "epoch": 0.23243829671819907, "grad_norm": 14.834383964538574, "learning_rate": 9.206775828969967e-07, "loss": 0.3932, "num_input_tokens_seen": 1797259264, "step": 857 }, { "epoch": 0.23270951993490643, "grad_norm": 12.486455917358398, "learning_rate": 9.204485484294355e-07, "loss": 0.37, "num_input_tokens_seen": 1799356416, "step": 858 }, { "epoch": 0.23298074315161377, "grad_norm": 15.540726661682129, "learning_rate": 9.202192158389791e-07, "loss": 0.6087, "num_input_tokens_seen": 1801453568, "step": 859 }, { "epoch": 0.23325196636832113, "grad_norm": 16.4538631439209, "learning_rate": 9.199895853101856e-07, "loss": 0.7186, "num_input_tokens_seen": 1803550720, "step": 860 }, { "epoch": 0.2335231895850285, "grad_norm": 18.737592697143555, "learning_rate": 9.197596570278529e-07, "loss": 0.7831, "num_input_tokens_seen": 1805647872, "step": 861 }, { "epoch": 0.23379441280173582, "grad_norm": 13.706647872924805, "learning_rate": 9.19529431177019e-07, "loss": 0.3793, "num_input_tokens_seen": 1807745024, "step": 862 }, { "epoch": 0.23406563601844319, "grad_norm": 11.725102424621582, "learning_rate": 9.192989079429603e-07, "loss": 0.1998, "num_input_tokens_seen": 1809842176, "step": 863 }, { "epoch": 0.23433685923515052, "grad_norm": 15.088400840759277, "learning_rate": 9.190680875111934e-07, "loss": 0.4601, "num_input_tokens_seen": 1811939328, "step": 864 }, { "epoch": 0.23460808245185788, "grad_norm": 14.075953483581543, "learning_rate": 9.188369700674735e-07, "loss": 0.4754, "num_input_tokens_seen": 1814036480, "step": 865 }, { "epoch": 0.23487930566856524, "grad_norm": 19.2708740234375, "learning_rate": 9.186055557977957e-07, "loss": 0.5167, "num_input_tokens_seen": 1816133632, "step": 866 }, { "epoch": 0.23515052888527258, "grad_norm": 18.750097274780273, "learning_rate": 9.18373844888393e-07, "loss": 0.7142, "num_input_tokens_seen": 1818230784, "step": 867 }, { "epoch": 0.23542175210197994, "grad_norm": 13.886237144470215, "learning_rate": 9.181418375257374e-07, "loss": 0.4226, "num_input_tokens_seen": 1820327936, "step": 868 }, { "epoch": 0.23569297531868727, "grad_norm": 18.76662254333496, "learning_rate": 9.179095338965401e-07, "loss": 0.6679, "num_input_tokens_seen": 1822425088, "step": 869 }, { "epoch": 0.23596419853539463, "grad_norm": 13.589776039123535, "learning_rate": 9.176769341877497e-07, "loss": 0.4632, "num_input_tokens_seen": 1824522240, "step": 870 }, { "epoch": 0.236235421752102, "grad_norm": 16.83132553100586, "learning_rate": 9.17444038586554e-07, "loss": 0.5446, "num_input_tokens_seen": 1826619392, "step": 871 }, { "epoch": 0.23650664496880933, "grad_norm": 26.304426193237305, "learning_rate": 9.172108472803782e-07, "loss": 0.6821, "num_input_tokens_seen": 1828716544, "step": 872 }, { "epoch": 0.2367778681855167, "grad_norm": 18.7716121673584, "learning_rate": 9.16977360456886e-07, "loss": 0.4784, "num_input_tokens_seen": 1830813696, "step": 873 }, { "epoch": 0.23704909140222402, "grad_norm": 11.739683151245117, "learning_rate": 9.167435783039786e-07, "loss": 0.3455, "num_input_tokens_seen": 1832910848, "step": 874 }, { "epoch": 0.23732031461893138, "grad_norm": 14.229680061340332, "learning_rate": 9.165095010097949e-07, "loss": 0.6126, "num_input_tokens_seen": 1835008000, "step": 875 }, { "epoch": 0.23759153783563874, "grad_norm": 15.295489311218262, "learning_rate": 9.162751287627116e-07, "loss": 0.6215, "num_input_tokens_seen": 1837105152, "step": 876 }, { "epoch": 0.23786276105234608, "grad_norm": 13.08266830444336, "learning_rate": 9.160404617513424e-07, "loss": 0.4853, "num_input_tokens_seen": 1839202304, "step": 877 }, { "epoch": 0.23813398426905344, "grad_norm": 18.74358558654785, "learning_rate": 9.158055001645385e-07, "loss": 0.7992, "num_input_tokens_seen": 1841299456, "step": 878 }, { "epoch": 0.23840520748576077, "grad_norm": 14.171740531921387, "learning_rate": 9.155702441913881e-07, "loss": 0.4748, "num_input_tokens_seen": 1843396608, "step": 879 }, { "epoch": 0.23867643070246813, "grad_norm": 15.582724571228027, "learning_rate": 9.15334694021216e-07, "loss": 0.5375, "num_input_tokens_seen": 1845493760, "step": 880 }, { "epoch": 0.23894765391917547, "grad_norm": 16.704227447509766, "learning_rate": 9.150988498435843e-07, "loss": 0.5576, "num_input_tokens_seen": 1847590912, "step": 881 }, { "epoch": 0.23921887713588283, "grad_norm": 12.607769966125488, "learning_rate": 9.148627118482912e-07, "loss": 0.287, "num_input_tokens_seen": 1849688064, "step": 882 }, { "epoch": 0.2394901003525902, "grad_norm": 16.634292602539062, "learning_rate": 9.146262802253717e-07, "loss": 0.752, "num_input_tokens_seen": 1851785216, "step": 883 }, { "epoch": 0.23976132356929752, "grad_norm": 13.200092315673828, "learning_rate": 9.14389555165097e-07, "loss": 0.3663, "num_input_tokens_seen": 1853882368, "step": 884 }, { "epoch": 0.24003254678600489, "grad_norm": 15.857352256774902, "learning_rate": 9.141525368579742e-07, "loss": 0.507, "num_input_tokens_seen": 1855979520, "step": 885 }, { "epoch": 0.24030377000271222, "grad_norm": 15.330340385437012, "learning_rate": 9.139152254947469e-07, "loss": 0.4418, "num_input_tokens_seen": 1858076672, "step": 886 }, { "epoch": 0.24057499321941958, "grad_norm": 12.231500625610352, "learning_rate": 9.136776212663942e-07, "loss": 0.4606, "num_input_tokens_seen": 1860173824, "step": 887 }, { "epoch": 0.24084621643612694, "grad_norm": 19.30156898498535, "learning_rate": 9.134397243641307e-07, "loss": 0.6261, "num_input_tokens_seen": 1862270976, "step": 888 }, { "epoch": 0.24111743965283428, "grad_norm": 23.425487518310547, "learning_rate": 9.132015349794069e-07, "loss": 0.4225, "num_input_tokens_seen": 1864368128, "step": 889 }, { "epoch": 0.24138866286954164, "grad_norm": 9.727029800415039, "learning_rate": 9.129630533039086e-07, "loss": 0.2508, "num_input_tokens_seen": 1866465280, "step": 890 }, { "epoch": 0.24165988608624897, "grad_norm": 20.487241744995117, "learning_rate": 9.127242795295569e-07, "loss": 0.7279, "num_input_tokens_seen": 1868562432, "step": 891 }, { "epoch": 0.24193110930295633, "grad_norm": 16.180301666259766, "learning_rate": 9.124852138485076e-07, "loss": 0.5699, "num_input_tokens_seen": 1870659584, "step": 892 }, { "epoch": 0.2422023325196637, "grad_norm": 20.350139617919922, "learning_rate": 9.12245856453152e-07, "loss": 0.856, "num_input_tokens_seen": 1872756736, "step": 893 }, { "epoch": 0.24247355573637103, "grad_norm": 18.013940811157227, "learning_rate": 9.120062075361155e-07, "loss": 0.6445, "num_input_tokens_seen": 1874853888, "step": 894 }, { "epoch": 0.2427447789530784, "grad_norm": 16.004846572875977, "learning_rate": 9.117662672902584e-07, "loss": 0.6282, "num_input_tokens_seen": 1876951040, "step": 895 }, { "epoch": 0.24301600216978572, "grad_norm": 14.721831321716309, "learning_rate": 9.115260359086757e-07, "loss": 0.3772, "num_input_tokens_seen": 1879048192, "step": 896 }, { "epoch": 0.24328722538649308, "grad_norm": 19.249744415283203, "learning_rate": 9.112855135846964e-07, "loss": 0.5968, "num_input_tokens_seen": 1881145344, "step": 897 }, { "epoch": 0.24355844860320044, "grad_norm": 12.756824493408203, "learning_rate": 9.110447005118836e-07, "loss": 0.4154, "num_input_tokens_seen": 1883242496, "step": 898 }, { "epoch": 0.24382967181990778, "grad_norm": 11.557883262634277, "learning_rate": 9.108035968840348e-07, "loss": 0.3552, "num_input_tokens_seen": 1885339648, "step": 899 }, { "epoch": 0.24410089503661514, "grad_norm": 19.659439086914062, "learning_rate": 9.105622028951806e-07, "loss": 0.6714, "num_input_tokens_seen": 1887436800, "step": 900 }, { "epoch": 0.24437211825332247, "grad_norm": 15.646737098693848, "learning_rate": 9.103205187395861e-07, "loss": 0.367, "num_input_tokens_seen": 1889533952, "step": 901 }, { "epoch": 0.24464334147002983, "grad_norm": 13.878365516662598, "learning_rate": 9.100785446117493e-07, "loss": 0.4032, "num_input_tokens_seen": 1891631104, "step": 902 }, { "epoch": 0.2449145646867372, "grad_norm": 16.992267608642578, "learning_rate": 9.098362807064017e-07, "loss": 0.5326, "num_input_tokens_seen": 1893728256, "step": 903 }, { "epoch": 0.24518578790344453, "grad_norm": 12.592440605163574, "learning_rate": 9.095937272185083e-07, "loss": 0.3813, "num_input_tokens_seen": 1895825408, "step": 904 }, { "epoch": 0.2454570111201519, "grad_norm": 14.522198677062988, "learning_rate": 9.093508843432667e-07, "loss": 0.449, "num_input_tokens_seen": 1897922560, "step": 905 }, { "epoch": 0.24572823433685922, "grad_norm": 19.913555145263672, "learning_rate": 9.091077522761078e-07, "loss": 0.6901, "num_input_tokens_seen": 1900019712, "step": 906 }, { "epoch": 0.2459994575535666, "grad_norm": 18.202465057373047, "learning_rate": 9.088643312126948e-07, "loss": 0.7029, "num_input_tokens_seen": 1902116864, "step": 907 }, { "epoch": 0.24627068077027395, "grad_norm": 13.385407447814941, "learning_rate": 9.086206213489239e-07, "loss": 0.4506, "num_input_tokens_seen": 1904214016, "step": 908 }, { "epoch": 0.24654190398698128, "grad_norm": 17.774667739868164, "learning_rate": 9.083766228809234e-07, "loss": 0.6725, "num_input_tokens_seen": 1906311168, "step": 909 }, { "epoch": 0.24681312720368864, "grad_norm": 14.268959999084473, "learning_rate": 9.081323360050543e-07, "loss": 0.4936, "num_input_tokens_seen": 1908408320, "step": 910 }, { "epoch": 0.24708435042039598, "grad_norm": 15.435829162597656, "learning_rate": 9.078877609179088e-07, "loss": 0.4626, "num_input_tokens_seen": 1910505472, "step": 911 }, { "epoch": 0.24735557363710334, "grad_norm": 11.141834259033203, "learning_rate": 9.076428978163121e-07, "loss": 0.337, "num_input_tokens_seen": 1912602624, "step": 912 }, { "epoch": 0.2476267968538107, "grad_norm": 21.919893264770508, "learning_rate": 9.073977468973206e-07, "loss": 0.993, "num_input_tokens_seen": 1914699776, "step": 913 }, { "epoch": 0.24789802007051803, "grad_norm": 15.084315299987793, "learning_rate": 9.071523083582223e-07, "loss": 0.3735, "num_input_tokens_seen": 1916796928, "step": 914 }, { "epoch": 0.2481692432872254, "grad_norm": 19.048635482788086, "learning_rate": 9.06906582396537e-07, "loss": 0.7275, "num_input_tokens_seen": 1918894080, "step": 915 }, { "epoch": 0.24844046650393273, "grad_norm": 11.731395721435547, "learning_rate": 9.066605692100155e-07, "loss": 0.388, "num_input_tokens_seen": 1920991232, "step": 916 }, { "epoch": 0.2487116897206401, "grad_norm": 11.196650505065918, "learning_rate": 9.064142689966397e-07, "loss": 0.3481, "num_input_tokens_seen": 1923088384, "step": 917 }, { "epoch": 0.24898291293734745, "grad_norm": 21.21782684326172, "learning_rate": 9.061676819546229e-07, "loss": 0.863, "num_input_tokens_seen": 1925185536, "step": 918 }, { "epoch": 0.24925413615405478, "grad_norm": 14.912806510925293, "learning_rate": 9.059208082824087e-07, "loss": 0.5098, "num_input_tokens_seen": 1927282688, "step": 919 }, { "epoch": 0.24952535937076215, "grad_norm": 18.304712295532227, "learning_rate": 9.05673648178672e-07, "loss": 0.6095, "num_input_tokens_seen": 1929379840, "step": 920 }, { "epoch": 0.24979658258746948, "grad_norm": 11.65080738067627, "learning_rate": 9.054262018423175e-07, "loss": 0.3207, "num_input_tokens_seen": 1931476992, "step": 921 }, { "epoch": 0.2500678058041768, "grad_norm": 21.233457565307617, "learning_rate": 9.051784694724808e-07, "loss": 0.8062, "num_input_tokens_seen": 1933574144, "step": 922 }, { "epoch": 0.2503390290208842, "grad_norm": 9.506631851196289, "learning_rate": 9.049304512685274e-07, "loss": 0.2804, "num_input_tokens_seen": 1935671296, "step": 923 }, { "epoch": 0.25061025223759154, "grad_norm": 10.630436897277832, "learning_rate": 9.046821474300527e-07, "loss": 0.3033, "num_input_tokens_seen": 1937768448, "step": 924 }, { "epoch": 0.2508814754542989, "grad_norm": 18.508195877075195, "learning_rate": 9.044335581568827e-07, "loss": 0.6053, "num_input_tokens_seen": 1939865600, "step": 925 }, { "epoch": 0.25115269867100626, "grad_norm": 11.025355339050293, "learning_rate": 9.041846836490723e-07, "loss": 0.3725, "num_input_tokens_seen": 1941962752, "step": 926 }, { "epoch": 0.25142392188771356, "grad_norm": 13.680339813232422, "learning_rate": 9.03935524106906e-07, "loss": 0.4114, "num_input_tokens_seen": 1944059904, "step": 927 }, { "epoch": 0.2516951451044209, "grad_norm": 10.082228660583496, "learning_rate": 9.036860797308984e-07, "loss": 0.2863, "num_input_tokens_seen": 1946157056, "step": 928 }, { "epoch": 0.2519663683211283, "grad_norm": 15.916239738464355, "learning_rate": 9.034363507217925e-07, "loss": 0.452, "num_input_tokens_seen": 1948254208, "step": 929 }, { "epoch": 0.25223759153783565, "grad_norm": 20.79644012451172, "learning_rate": 9.031863372805606e-07, "loss": 0.6497, "num_input_tokens_seen": 1950351360, "step": 930 }, { "epoch": 0.252508814754543, "grad_norm": 15.72608757019043, "learning_rate": 9.029360396084043e-07, "loss": 0.3006, "num_input_tokens_seen": 1952448512, "step": 931 }, { "epoch": 0.2527800379712503, "grad_norm": 13.770834922790527, "learning_rate": 9.026854579067537e-07, "loss": 0.5086, "num_input_tokens_seen": 1954545664, "step": 932 }, { "epoch": 0.2530512611879577, "grad_norm": 18.707565307617188, "learning_rate": 9.024345923772671e-07, "loss": 0.513, "num_input_tokens_seen": 1956642816, "step": 933 }, { "epoch": 0.25332248440466504, "grad_norm": 11.008857727050781, "learning_rate": 9.021834432218317e-07, "loss": 0.336, "num_input_tokens_seen": 1958739968, "step": 934 }, { "epoch": 0.2535937076213724, "grad_norm": 13.265803337097168, "learning_rate": 9.019320106425629e-07, "loss": 0.3631, "num_input_tokens_seen": 1960837120, "step": 935 }, { "epoch": 0.25386493083807976, "grad_norm": 11.919256210327148, "learning_rate": 9.016802948418038e-07, "loss": 0.2662, "num_input_tokens_seen": 1962934272, "step": 936 }, { "epoch": 0.25413615405478707, "grad_norm": 13.845582962036133, "learning_rate": 9.014282960221257e-07, "loss": 0.5253, "num_input_tokens_seen": 1965031424, "step": 937 }, { "epoch": 0.25440737727149443, "grad_norm": 14.255606651306152, "learning_rate": 9.01176014386328e-07, "loss": 0.444, "num_input_tokens_seen": 1967128576, "step": 938 }, { "epoch": 0.2546786004882018, "grad_norm": 19.263900756835938, "learning_rate": 9.009234501374371e-07, "loss": 0.6345, "num_input_tokens_seen": 1969225728, "step": 939 }, { "epoch": 0.25494982370490915, "grad_norm": 21.17646026611328, "learning_rate": 9.006706034787071e-07, "loss": 0.5709, "num_input_tokens_seen": 1971322880, "step": 940 }, { "epoch": 0.2552210469216165, "grad_norm": 16.962255477905273, "learning_rate": 9.004174746136196e-07, "loss": 0.6367, "num_input_tokens_seen": 1973420032, "step": 941 }, { "epoch": 0.2554922701383238, "grad_norm": 12.953526496887207, "learning_rate": 9.001640637458829e-07, "loss": 0.4449, "num_input_tokens_seen": 1975517184, "step": 942 }, { "epoch": 0.2557634933550312, "grad_norm": 19.991397857666016, "learning_rate": 8.999103710794323e-07, "loss": 0.9314, "num_input_tokens_seen": 1977614336, "step": 943 }, { "epoch": 0.25603471657173854, "grad_norm": 12.952409744262695, "learning_rate": 8.996563968184302e-07, "loss": 0.3151, "num_input_tokens_seen": 1979711488, "step": 944 }, { "epoch": 0.2563059397884459, "grad_norm": 18.413883209228516, "learning_rate": 8.994021411672653e-07, "loss": 0.7246, "num_input_tokens_seen": 1981808640, "step": 945 }, { "epoch": 0.25657716300515326, "grad_norm": 15.939887046813965, "learning_rate": 8.99147604330553e-07, "loss": 0.5387, "num_input_tokens_seen": 1983905792, "step": 946 }, { "epoch": 0.25684838622186057, "grad_norm": 18.361181259155273, "learning_rate": 8.988927865131347e-07, "loss": 0.5397, "num_input_tokens_seen": 1986002944, "step": 947 }, { "epoch": 0.25711960943856793, "grad_norm": 16.950855255126953, "learning_rate": 8.986376879200783e-07, "loss": 0.5404, "num_input_tokens_seen": 1988100096, "step": 948 }, { "epoch": 0.2573908326552753, "grad_norm": 13.930158615112305, "learning_rate": 8.983823087566772e-07, "loss": 0.3202, "num_input_tokens_seen": 1990197248, "step": 949 }, { "epoch": 0.25766205587198265, "grad_norm": 18.551612854003906, "learning_rate": 8.981266492284511e-07, "loss": 0.5814, "num_input_tokens_seen": 1992294400, "step": 950 }, { "epoch": 0.25793327908869, "grad_norm": 17.812480926513672, "learning_rate": 8.978707095411446e-07, "loss": 0.6843, "num_input_tokens_seen": 1994391552, "step": 951 }, { "epoch": 0.2582045023053973, "grad_norm": 22.378063201904297, "learning_rate": 8.976144899007288e-07, "loss": 1.0731, "num_input_tokens_seen": 1996488704, "step": 952 }, { "epoch": 0.2584757255221047, "grad_norm": 12.730658531188965, "learning_rate": 8.973579905133991e-07, "loss": 0.4466, "num_input_tokens_seen": 1998585856, "step": 953 }, { "epoch": 0.25874694873881204, "grad_norm": 9.290279388427734, "learning_rate": 8.971012115855766e-07, "loss": 0.2807, "num_input_tokens_seen": 2000683008, "step": 954 }, { "epoch": 0.2590181719555194, "grad_norm": 14.961581230163574, "learning_rate": 8.968441533239073e-07, "loss": 0.5023, "num_input_tokens_seen": 2002780160, "step": 955 }, { "epoch": 0.25928939517222677, "grad_norm": 11.520310401916504, "learning_rate": 8.965868159352616e-07, "loss": 0.3769, "num_input_tokens_seen": 2004877312, "step": 956 }, { "epoch": 0.2595606183889341, "grad_norm": 13.574871063232422, "learning_rate": 8.963291996267354e-07, "loss": 0.4165, "num_input_tokens_seen": 2006974464, "step": 957 }, { "epoch": 0.25983184160564143, "grad_norm": 10.172213554382324, "learning_rate": 8.960713046056478e-07, "loss": 0.3482, "num_input_tokens_seen": 2009071616, "step": 958 }, { "epoch": 0.2601030648223488, "grad_norm": 15.749048233032227, "learning_rate": 8.958131310795434e-07, "loss": 0.4703, "num_input_tokens_seen": 2011168768, "step": 959 }, { "epoch": 0.26037428803905616, "grad_norm": 13.04376220703125, "learning_rate": 8.955546792561902e-07, "loss": 0.3705, "num_input_tokens_seen": 2013265920, "step": 960 }, { "epoch": 0.2606455112557635, "grad_norm": 13.472237586975098, "learning_rate": 8.952959493435806e-07, "loss": 0.4021, "num_input_tokens_seen": 2015363072, "step": 961 }, { "epoch": 0.2609167344724708, "grad_norm": 15.37488079071045, "learning_rate": 8.950369415499304e-07, "loss": 0.5569, "num_input_tokens_seen": 2017460224, "step": 962 }, { "epoch": 0.2611879576891782, "grad_norm": 18.64729881286621, "learning_rate": 8.947776560836793e-07, "loss": 0.7467, "num_input_tokens_seen": 2019557376, "step": 963 }, { "epoch": 0.26145918090588555, "grad_norm": 13.231241226196289, "learning_rate": 8.945180931534902e-07, "loss": 0.4492, "num_input_tokens_seen": 2021654528, "step": 964 }, { "epoch": 0.2617304041225929, "grad_norm": 13.122672080993652, "learning_rate": 8.942582529682496e-07, "loss": 0.387, "num_input_tokens_seen": 2023751680, "step": 965 }, { "epoch": 0.26200162733930027, "grad_norm": 12.181504249572754, "learning_rate": 8.939981357370672e-07, "loss": 0.3468, "num_input_tokens_seen": 2025848832, "step": 966 }, { "epoch": 0.2622728505560076, "grad_norm": 13.75951862335205, "learning_rate": 8.937377416692752e-07, "loss": 0.4132, "num_input_tokens_seen": 2027945984, "step": 967 }, { "epoch": 0.26254407377271494, "grad_norm": 15.17529582977295, "learning_rate": 8.934770709744289e-07, "loss": 0.5208, "num_input_tokens_seen": 2030043136, "step": 968 }, { "epoch": 0.2628152969894223, "grad_norm": 12.855225563049316, "learning_rate": 8.93216123862306e-07, "loss": 0.3606, "num_input_tokens_seen": 2032140288, "step": 969 }, { "epoch": 0.26308652020612966, "grad_norm": 23.8533992767334, "learning_rate": 8.929549005429071e-07, "loss": 1.1691, "num_input_tokens_seen": 2034237440, "step": 970 }, { "epoch": 0.263357743422837, "grad_norm": 12.444233894348145, "learning_rate": 8.926934012264546e-07, "loss": 0.4991, "num_input_tokens_seen": 2036334592, "step": 971 }, { "epoch": 0.2636289666395443, "grad_norm": 19.38515853881836, "learning_rate": 8.924316261233933e-07, "loss": 0.603, "num_input_tokens_seen": 2038431744, "step": 972 }, { "epoch": 0.2639001898562517, "grad_norm": 25.431591033935547, "learning_rate": 8.921695754443898e-07, "loss": 0.4341, "num_input_tokens_seen": 2040528896, "step": 973 }, { "epoch": 0.26417141307295905, "grad_norm": 14.830904960632324, "learning_rate": 8.919072494003325e-07, "loss": 0.4873, "num_input_tokens_seen": 2042626048, "step": 974 }, { "epoch": 0.2644426362896664, "grad_norm": 10.834352493286133, "learning_rate": 8.916446482023313e-07, "loss": 0.386, "num_input_tokens_seen": 2044723200, "step": 975 }, { "epoch": 0.26471385950637377, "grad_norm": 18.31397819519043, "learning_rate": 8.913817720617178e-07, "loss": 0.5712, "num_input_tokens_seen": 2046820352, "step": 976 }, { "epoch": 0.2649850827230811, "grad_norm": 14.418971061706543, "learning_rate": 8.911186211900448e-07, "loss": 0.3854, "num_input_tokens_seen": 2048917504, "step": 977 }, { "epoch": 0.26525630593978844, "grad_norm": 15.985840797424316, "learning_rate": 8.908551957990858e-07, "loss": 0.6277, "num_input_tokens_seen": 2051014656, "step": 978 }, { "epoch": 0.2655275291564958, "grad_norm": 20.523096084594727, "learning_rate": 8.90591496100836e-07, "loss": 0.6733, "num_input_tokens_seen": 2053111808, "step": 979 }, { "epoch": 0.26579875237320316, "grad_norm": 21.447572708129883, "learning_rate": 8.903275223075104e-07, "loss": 0.5786, "num_input_tokens_seen": 2055208960, "step": 980 }, { "epoch": 0.2660699755899105, "grad_norm": 15.837870597839355, "learning_rate": 8.900632746315454e-07, "loss": 0.6692, "num_input_tokens_seen": 2057306112, "step": 981 }, { "epoch": 0.26634119880661783, "grad_norm": 25.582645416259766, "learning_rate": 8.897987532855972e-07, "loss": 1.0872, "num_input_tokens_seen": 2059403264, "step": 982 }, { "epoch": 0.2666124220233252, "grad_norm": 17.301437377929688, "learning_rate": 8.895339584825429e-07, "loss": 0.5819, "num_input_tokens_seen": 2061500416, "step": 983 }, { "epoch": 0.26688364524003255, "grad_norm": 14.604111671447754, "learning_rate": 8.892688904354787e-07, "loss": 0.4918, "num_input_tokens_seen": 2063597568, "step": 984 }, { "epoch": 0.2671548684567399, "grad_norm": 12.52351188659668, "learning_rate": 8.890035493577219e-07, "loss": 0.388, "num_input_tokens_seen": 2065694720, "step": 985 }, { "epoch": 0.2674260916734473, "grad_norm": 21.226011276245117, "learning_rate": 8.887379354628085e-07, "loss": 0.9703, "num_input_tokens_seen": 2067791872, "step": 986 }, { "epoch": 0.2676973148901546, "grad_norm": 12.156089782714844, "learning_rate": 8.884720489644945e-07, "loss": 0.36, "num_input_tokens_seen": 2069889024, "step": 987 }, { "epoch": 0.26796853810686194, "grad_norm": 16.55057144165039, "learning_rate": 8.882058900767555e-07, "loss": 0.6824, "num_input_tokens_seen": 2071986176, "step": 988 }, { "epoch": 0.2682397613235693, "grad_norm": 19.046653747558594, "learning_rate": 8.879394590137857e-07, "loss": 0.6588, "num_input_tokens_seen": 2074083328, "step": 989 }, { "epoch": 0.26851098454027666, "grad_norm": 13.83050537109375, "learning_rate": 8.876727559899989e-07, "loss": 0.3678, "num_input_tokens_seen": 2076180480, "step": 990 }, { "epoch": 0.26878220775698397, "grad_norm": 23.640356063842773, "learning_rate": 8.874057812200274e-07, "loss": 0.9572, "num_input_tokens_seen": 2078277632, "step": 991 }, { "epoch": 0.26905343097369133, "grad_norm": 15.057960510253906, "learning_rate": 8.871385349187225e-07, "loss": 0.722, "num_input_tokens_seen": 2080374784, "step": 992 }, { "epoch": 0.2693246541903987, "grad_norm": 18.069377899169922, "learning_rate": 8.868710173011538e-07, "loss": 0.7448, "num_input_tokens_seen": 2082471936, "step": 993 }, { "epoch": 0.26959587740710605, "grad_norm": 13.327810287475586, "learning_rate": 8.866032285826091e-07, "loss": 0.5023, "num_input_tokens_seen": 2084569088, "step": 994 }, { "epoch": 0.2698671006238134, "grad_norm": 11.229975700378418, "learning_rate": 8.86335168978595e-07, "loss": 0.357, "num_input_tokens_seen": 2086666240, "step": 995 }, { "epoch": 0.2701383238405207, "grad_norm": 10.154101371765137, "learning_rate": 8.860668387048353e-07, "loss": 0.2903, "num_input_tokens_seen": 2088763392, "step": 996 }, { "epoch": 0.2704095470572281, "grad_norm": 12.260966300964355, "learning_rate": 8.85798237977272e-07, "loss": 0.3106, "num_input_tokens_seen": 2090860544, "step": 997 }, { "epoch": 0.27068077027393544, "grad_norm": 18.005229949951172, "learning_rate": 8.85529367012065e-07, "loss": 0.6471, "num_input_tokens_seen": 2092957696, "step": 998 }, { "epoch": 0.2709519934906428, "grad_norm": 13.38134765625, "learning_rate": 8.852602260255911e-07, "loss": 0.4971, "num_input_tokens_seen": 2095054848, "step": 999 }, { "epoch": 0.27122321670735017, "grad_norm": 17.00292205810547, "learning_rate": 8.849908152344451e-07, "loss": 0.7469, "num_input_tokens_seen": 2097152000, "step": 1000 }, { "epoch": 0.00027122321670735016, "grad_norm": 21.211387634277344, "learning_rate": 8.847211348554382e-07, "loss": 0.5452, "num_input_tokens_seen": 2099249152, "step": 1001 }, { "epoch": 0.0005424464334147003, "grad_norm": 15.134449005126953, "learning_rate": 8.844511851055991e-07, "loss": 0.494, "num_input_tokens_seen": 2101346304, "step": 1002 }, { "epoch": 0.0008136696501220504, "grad_norm": 15.855742454528809, "learning_rate": 8.841809662021731e-07, "loss": 0.5585, "num_input_tokens_seen": 2103443456, "step": 1003 }, { "epoch": 0.0010848928668294006, "grad_norm": 25.325538635253906, "learning_rate": 8.839104783626219e-07, "loss": 0.9896, "num_input_tokens_seen": 2105540608, "step": 1004 }, { "epoch": 0.0013561160835367507, "grad_norm": 12.955252647399902, "learning_rate": 8.836397218046239e-07, "loss": 0.423, "num_input_tokens_seen": 2107637760, "step": 1005 }, { "epoch": 0.0016273393002441008, "grad_norm": 18.2934513092041, "learning_rate": 8.83368696746074e-07, "loss": 0.5536, "num_input_tokens_seen": 2109734912, "step": 1006 }, { "epoch": 0.001898562516951451, "grad_norm": 19.734365463256836, "learning_rate": 8.830974034050824e-07, "loss": 0.8425, "num_input_tokens_seen": 2111832064, "step": 1007 }, { "epoch": 0.0021697857336588013, "grad_norm": 9.506952285766602, "learning_rate": 8.828258419999759e-07, "loss": 0.2779, "num_input_tokens_seen": 2113929216, "step": 1008 }, { "epoch": 0.0024410089503661514, "grad_norm": 10.620465278625488, "learning_rate": 8.825540127492965e-07, "loss": 0.3105, "num_input_tokens_seen": 2116026368, "step": 1009 }, { "epoch": 0.0027122321670735015, "grad_norm": 20.96041488647461, "learning_rate": 8.822819158718026e-07, "loss": 0.775, "num_input_tokens_seen": 2118123520, "step": 1010 }, { "epoch": 0.0029834553837808516, "grad_norm": 15.075592994689941, "learning_rate": 8.820095515864669e-07, "loss": 0.505, "num_input_tokens_seen": 2120220672, "step": 1011 }, { "epoch": 0.0032546786004882017, "grad_norm": 19.47187042236328, "learning_rate": 8.81736920112478e-07, "loss": 0.6136, "num_input_tokens_seen": 2122317824, "step": 1012 }, { "epoch": 0.003525901817195552, "grad_norm": 17.397031784057617, "learning_rate": 8.814640216692391e-07, "loss": 0.4918, "num_input_tokens_seen": 2124414976, "step": 1013 }, { "epoch": 0.003797125033902902, "grad_norm": 23.4683895111084, "learning_rate": 8.81190856476369e-07, "loss": 0.8976, "num_input_tokens_seen": 2126512128, "step": 1014 }, { "epoch": 0.0040683482506102524, "grad_norm": 21.26357078552246, "learning_rate": 8.809174247537003e-07, "loss": 0.6695, "num_input_tokens_seen": 2128609280, "step": 1015 }, { "epoch": 0.0043395714673176026, "grad_norm": 9.867420196533203, "learning_rate": 8.806437267212805e-07, "loss": 0.2329, "num_input_tokens_seen": 2130706432, "step": 1016 }, { "epoch": 0.004610794684024953, "grad_norm": 26.58327293395996, "learning_rate": 8.803697625993713e-07, "loss": 0.8958, "num_input_tokens_seen": 2132803584, "step": 1017 }, { "epoch": 0.004882017900732303, "grad_norm": 15.50483226776123, "learning_rate": 8.800955326084487e-07, "loss": 0.62, "num_input_tokens_seen": 2134900736, "step": 1018 }, { "epoch": 0.005153241117439653, "grad_norm": 9.97493839263916, "learning_rate": 8.798210369692025e-07, "loss": 0.255, "num_input_tokens_seen": 2136997888, "step": 1019 }, { "epoch": 0.005424464334147003, "grad_norm": 13.728446960449219, "learning_rate": 8.795462759025364e-07, "loss": 0.5253, "num_input_tokens_seen": 2139095040, "step": 1020 }, { "epoch": 0.005695687550854353, "grad_norm": 13.042479515075684, "learning_rate": 8.792712496295677e-07, "loss": 0.3554, "num_input_tokens_seen": 2141192192, "step": 1021 }, { "epoch": 0.005966910767561703, "grad_norm": 14.499848365783691, "learning_rate": 8.789959583716268e-07, "loss": 0.4398, "num_input_tokens_seen": 2143289344, "step": 1022 }, { "epoch": 0.006238133984269053, "grad_norm": 12.071355819702148, "learning_rate": 8.787204023502579e-07, "loss": 0.3154, "num_input_tokens_seen": 2145386496, "step": 1023 }, { "epoch": 0.006509357200976403, "grad_norm": 16.29543113708496, "learning_rate": 8.78444581787218e-07, "loss": 0.4772, "num_input_tokens_seen": 2147483648, "step": 1024 }, { "epoch": 0.0067805804176837535, "grad_norm": 14.60418701171875, "learning_rate": 8.781684969044769e-07, "loss": 0.5242, "num_input_tokens_seen": 2149580800, "step": 1025 }, { "epoch": 0.007051803634391104, "grad_norm": 16.060916900634766, "learning_rate": 8.778921479242173e-07, "loss": 0.6014, "num_input_tokens_seen": 2151677952, "step": 1026 }, { "epoch": 0.007323026851098454, "grad_norm": 11.258378028869629, "learning_rate": 8.776155350688342e-07, "loss": 0.2425, "num_input_tokens_seen": 2153775104, "step": 1027 }, { "epoch": 0.007594250067805804, "grad_norm": 13.625775337219238, "learning_rate": 8.773386585609352e-07, "loss": 0.3129, "num_input_tokens_seen": 2155872256, "step": 1028 }, { "epoch": 0.007865473284513154, "grad_norm": 14.161125183105469, "learning_rate": 8.770615186233398e-07, "loss": 0.4769, "num_input_tokens_seen": 2157969408, "step": 1029 }, { "epoch": 0.008136696501220505, "grad_norm": 14.752073287963867, "learning_rate": 8.7678411547908e-07, "loss": 0.4059, "num_input_tokens_seen": 2160066560, "step": 1030 }, { "epoch": 0.008407919717927854, "grad_norm": 19.5262451171875, "learning_rate": 8.76506449351399e-07, "loss": 0.9232, "num_input_tokens_seen": 2162163712, "step": 1031 }, { "epoch": 0.008679142934635205, "grad_norm": 23.627025604248047, "learning_rate": 8.762285204637522e-07, "loss": 0.3903, "num_input_tokens_seen": 2164260864, "step": 1032 }, { "epoch": 0.008950366151342554, "grad_norm": 22.0424861907959, "learning_rate": 8.75950329039806e-07, "loss": 0.9677, "num_input_tokens_seen": 2166358016, "step": 1033 }, { "epoch": 0.009221589368049905, "grad_norm": 16.031373977661133, "learning_rate": 8.756718753034381e-07, "loss": 0.4666, "num_input_tokens_seen": 2168455168, "step": 1034 }, { "epoch": 0.009492812584757255, "grad_norm": 12.060325622558594, "learning_rate": 8.75393159478738e-07, "loss": 0.4239, "num_input_tokens_seen": 2170552320, "step": 1035 }, { "epoch": 0.009764035801464606, "grad_norm": 17.0775089263916, "learning_rate": 8.751141817900052e-07, "loss": 0.5839, "num_input_tokens_seen": 2172649472, "step": 1036 }, { "epoch": 0.010035259018171955, "grad_norm": 12.101827621459961, "learning_rate": 8.748349424617504e-07, "loss": 0.3156, "num_input_tokens_seen": 2174746624, "step": 1037 }, { "epoch": 0.010306482234879306, "grad_norm": 19.348678588867188, "learning_rate": 8.745554417186946e-07, "loss": 0.7375, "num_input_tokens_seen": 2176843776, "step": 1038 }, { "epoch": 0.010577705451586655, "grad_norm": 20.518571853637695, "learning_rate": 8.742756797857698e-07, "loss": 0.7549, "num_input_tokens_seen": 2178940928, "step": 1039 }, { "epoch": 0.010848928668294006, "grad_norm": 29.709476470947266, "learning_rate": 8.739956568881174e-07, "loss": 0.5175, "num_input_tokens_seen": 2181038080, "step": 1040 }, { "epoch": 0.011120151885001357, "grad_norm": 16.73153305053711, "learning_rate": 8.737153732510894e-07, "loss": 0.5272, "num_input_tokens_seen": 2183135232, "step": 1041 }, { "epoch": 0.011391375101708706, "grad_norm": 11.584559440612793, "learning_rate": 8.734348291002472e-07, "loss": 0.4049, "num_input_tokens_seen": 2185232384, "step": 1042 }, { "epoch": 0.011662598318416057, "grad_norm": 16.299226760864258, "learning_rate": 8.731540246613621e-07, "loss": 0.6874, "num_input_tokens_seen": 2187329536, "step": 1043 }, { "epoch": 0.011933821535123406, "grad_norm": 12.125165939331055, "learning_rate": 8.728729601604149e-07, "loss": 0.2929, "num_input_tokens_seen": 2189426688, "step": 1044 }, { "epoch": 0.012205044751830757, "grad_norm": 15.17818546295166, "learning_rate": 8.725916358235956e-07, "loss": 0.5507, "num_input_tokens_seen": 2191523840, "step": 1045 }, { "epoch": 0.012476267968538107, "grad_norm": 15.40369987487793, "learning_rate": 8.723100518773034e-07, "loss": 0.7183, "num_input_tokens_seen": 2193620992, "step": 1046 }, { "epoch": 0.012747491185245458, "grad_norm": 13.198980331420898, "learning_rate": 8.720282085481463e-07, "loss": 0.3474, "num_input_tokens_seen": 2195718144, "step": 1047 }, { "epoch": 0.013018714401952807, "grad_norm": 13.224248886108398, "learning_rate": 8.717461060629408e-07, "loss": 0.4623, "num_input_tokens_seen": 2197815296, "step": 1048 }, { "epoch": 0.013289937618660158, "grad_norm": 21.415082931518555, "learning_rate": 8.714637446487127e-07, "loss": 0.6044, "num_input_tokens_seen": 2199912448, "step": 1049 }, { "epoch": 0.013561160835367507, "grad_norm": 17.107685089111328, "learning_rate": 8.711811245326955e-07, "loss": 0.4799, "num_input_tokens_seen": 2202009600, "step": 1050 }, { "epoch": 0.013832384052074858, "grad_norm": 16.153474807739258, "learning_rate": 8.70898245942331e-07, "loss": 0.5492, "num_input_tokens_seen": 2204106752, "step": 1051 }, { "epoch": 0.014103607268782207, "grad_norm": 14.7495756149292, "learning_rate": 8.706151091052693e-07, "loss": 0.6168, "num_input_tokens_seen": 2206203904, "step": 1052 }, { "epoch": 0.014374830485489558, "grad_norm": 15.120250701904297, "learning_rate": 8.703317142493681e-07, "loss": 0.5481, "num_input_tokens_seen": 2208301056, "step": 1053 }, { "epoch": 0.014646053702196907, "grad_norm": 18.195310592651367, "learning_rate": 8.700480616026928e-07, "loss": 0.6981, "num_input_tokens_seen": 2210398208, "step": 1054 }, { "epoch": 0.014917276918904258, "grad_norm": 15.834187507629395, "learning_rate": 8.697641513935164e-07, "loss": 0.5543, "num_input_tokens_seen": 2212495360, "step": 1055 }, { "epoch": 0.015188500135611608, "grad_norm": 13.692312240600586, "learning_rate": 8.694799838503186e-07, "loss": 0.5255, "num_input_tokens_seen": 2214592512, "step": 1056 }, { "epoch": 0.015459723352318959, "grad_norm": 11.233868598937988, "learning_rate": 8.691955592017872e-07, "loss": 0.3443, "num_input_tokens_seen": 2216689664, "step": 1057 }, { "epoch": 0.015730946569026308, "grad_norm": 20.017642974853516, "learning_rate": 8.689108776768159e-07, "loss": 0.7494, "num_input_tokens_seen": 2218786816, "step": 1058 }, { "epoch": 0.01600216978573366, "grad_norm": 14.512331008911133, "learning_rate": 8.686259395045056e-07, "loss": 0.6063, "num_input_tokens_seen": 2220883968, "step": 1059 }, { "epoch": 0.01627339300244101, "grad_norm": 15.613311767578125, "learning_rate": 8.68340744914164e-07, "loss": 0.4511, "num_input_tokens_seen": 2222981120, "step": 1060 }, { "epoch": 0.01654461621914836, "grad_norm": 13.627348899841309, "learning_rate": 8.680552941353045e-07, "loss": 0.3019, "num_input_tokens_seen": 2225078272, "step": 1061 }, { "epoch": 0.016815839435855708, "grad_norm": 16.624229431152344, "learning_rate": 8.677695873976473e-07, "loss": 0.701, "num_input_tokens_seen": 2227175424, "step": 1062 }, { "epoch": 0.01708706265256306, "grad_norm": 11.855605125427246, "learning_rate": 8.674836249311182e-07, "loss": 0.4085, "num_input_tokens_seen": 2229272576, "step": 1063 }, { "epoch": 0.01735828586927041, "grad_norm": 16.31574249267578, "learning_rate": 8.671974069658488e-07, "loss": 0.6829, "num_input_tokens_seen": 2231369728, "step": 1064 }, { "epoch": 0.01762950908597776, "grad_norm": 16.633039474487305, "learning_rate": 8.669109337321767e-07, "loss": 0.603, "num_input_tokens_seen": 2233466880, "step": 1065 }, { "epoch": 0.01790073230268511, "grad_norm": 12.263994216918945, "learning_rate": 8.666242054606444e-07, "loss": 0.4269, "num_input_tokens_seen": 2235564032, "step": 1066 }, { "epoch": 0.01817195551939246, "grad_norm": 13.086897850036621, "learning_rate": 8.66337222382e-07, "loss": 0.5071, "num_input_tokens_seen": 2237661184, "step": 1067 }, { "epoch": 0.01844317873609981, "grad_norm": 18.24944305419922, "learning_rate": 8.660499847271965e-07, "loss": 0.7171, "num_input_tokens_seen": 2239758336, "step": 1068 }, { "epoch": 0.01871440195280716, "grad_norm": 12.880998611450195, "learning_rate": 8.657624927273919e-07, "loss": 0.3733, "num_input_tokens_seen": 2241855488, "step": 1069 }, { "epoch": 0.01898562516951451, "grad_norm": 11.316634178161621, "learning_rate": 8.654747466139488e-07, "loss": 0.417, "num_input_tokens_seen": 2243952640, "step": 1070 }, { "epoch": 0.01925684838622186, "grad_norm": 16.127735137939453, "learning_rate": 8.651867466184344e-07, "loss": 0.515, "num_input_tokens_seen": 2246049792, "step": 1071 }, { "epoch": 0.01952807160292921, "grad_norm": 13.648701667785645, "learning_rate": 8.6489849297262e-07, "loss": 0.4598, "num_input_tokens_seen": 2248146944, "step": 1072 }, { "epoch": 0.019799294819636562, "grad_norm": 18.77680778503418, "learning_rate": 8.646099859084812e-07, "loss": 0.4691, "num_input_tokens_seen": 2250244096, "step": 1073 }, { "epoch": 0.02007051803634391, "grad_norm": 14.230118751525879, "learning_rate": 8.643212256581978e-07, "loss": 0.4609, "num_input_tokens_seen": 2252341248, "step": 1074 }, { "epoch": 0.02034174125305126, "grad_norm": 20.32296371459961, "learning_rate": 8.640322124541525e-07, "loss": 0.4302, "num_input_tokens_seen": 2254438400, "step": 1075 }, { "epoch": 0.02061296446975861, "grad_norm": 19.631397247314453, "learning_rate": 8.637429465289324e-07, "loss": 0.7439, "num_input_tokens_seen": 2256535552, "step": 1076 }, { "epoch": 0.020884187686465962, "grad_norm": 20.95248794555664, "learning_rate": 8.63453428115328e-07, "loss": 0.9491, "num_input_tokens_seen": 2258632704, "step": 1077 }, { "epoch": 0.02115541090317331, "grad_norm": 23.594327926635742, "learning_rate": 8.631636574463321e-07, "loss": 1.0095, "num_input_tokens_seen": 2260729856, "step": 1078 }, { "epoch": 0.02142663411988066, "grad_norm": 13.66223430633545, "learning_rate": 8.628736347551417e-07, "loss": 0.4976, "num_input_tokens_seen": 2262827008, "step": 1079 }, { "epoch": 0.021697857336588012, "grad_norm": 14.790290832519531, "learning_rate": 8.625833602751559e-07, "loss": 0.4426, "num_input_tokens_seen": 2264924160, "step": 1080 }, { "epoch": 0.021969080553295363, "grad_norm": 22.70588493347168, "learning_rate": 8.622928342399762e-07, "loss": 0.9969, "num_input_tokens_seen": 2267021312, "step": 1081 }, { "epoch": 0.022240303770002714, "grad_norm": 15.693807601928711, "learning_rate": 8.620020568834072e-07, "loss": 0.6167, "num_input_tokens_seen": 2269118464, "step": 1082 }, { "epoch": 0.02251152698671006, "grad_norm": 14.573871612548828, "learning_rate": 8.617110284394553e-07, "loss": 0.5203, "num_input_tokens_seen": 2271215616, "step": 1083 }, { "epoch": 0.022782750203417412, "grad_norm": 12.033546447753906, "learning_rate": 8.614197491423293e-07, "loss": 0.3329, "num_input_tokens_seen": 2273312768, "step": 1084 }, { "epoch": 0.023053973420124763, "grad_norm": 21.395675659179688, "learning_rate": 8.611282192264396e-07, "loss": 0.8502, "num_input_tokens_seen": 2275409920, "step": 1085 }, { "epoch": 0.023325196636832114, "grad_norm": 10.336832046508789, "learning_rate": 8.608364389263984e-07, "loss": 0.3252, "num_input_tokens_seen": 2277507072, "step": 1086 }, { "epoch": 0.023596419853539462, "grad_norm": 14.438992500305176, "learning_rate": 8.605444084770192e-07, "loss": 0.4895, "num_input_tokens_seen": 2279604224, "step": 1087 }, { "epoch": 0.023867643070246813, "grad_norm": 12.839303970336914, "learning_rate": 8.602521281133173e-07, "loss": 0.4126, "num_input_tokens_seen": 2281701376, "step": 1088 }, { "epoch": 0.024138866286954164, "grad_norm": 15.832900047302246, "learning_rate": 8.599595980705085e-07, "loss": 0.298, "num_input_tokens_seen": 2283798528, "step": 1089 }, { "epoch": 0.024410089503661515, "grad_norm": 16.126102447509766, "learning_rate": 8.596668185840102e-07, "loss": 0.5055, "num_input_tokens_seen": 2285895680, "step": 1090 }, { "epoch": 0.024681312720368862, "grad_norm": 13.678728103637695, "learning_rate": 8.593737898894398e-07, "loss": 0.5144, "num_input_tokens_seen": 2287992832, "step": 1091 }, { "epoch": 0.024952535937076213, "grad_norm": 13.563033103942871, "learning_rate": 8.59080512222616e-07, "loss": 0.5175, "num_input_tokens_seen": 2290089984, "step": 1092 }, { "epoch": 0.025223759153783564, "grad_norm": 22.787578582763672, "learning_rate": 8.587869858195574e-07, "loss": 0.662, "num_input_tokens_seen": 2292187136, "step": 1093 }, { "epoch": 0.025494982370490915, "grad_norm": 12.561062812805176, "learning_rate": 8.584932109164826e-07, "loss": 0.4915, "num_input_tokens_seen": 2294284288, "step": 1094 }, { "epoch": 0.025766205587198263, "grad_norm": 10.324007034301758, "learning_rate": 8.581991877498109e-07, "loss": 0.3419, "num_input_tokens_seen": 2296381440, "step": 1095 }, { "epoch": 0.026037428803905614, "grad_norm": 16.793432235717773, "learning_rate": 8.579049165561607e-07, "loss": 0.5506, "num_input_tokens_seen": 2298478592, "step": 1096 }, { "epoch": 0.026308652020612965, "grad_norm": 18.08753776550293, "learning_rate": 8.576103975723502e-07, "loss": 0.6403, "num_input_tokens_seen": 2300575744, "step": 1097 }, { "epoch": 0.026579875237320316, "grad_norm": 14.949502944946289, "learning_rate": 8.573156310353974e-07, "loss": 0.5685, "num_input_tokens_seen": 2302672896, "step": 1098 }, { "epoch": 0.026851098454027666, "grad_norm": 14.36096477508545, "learning_rate": 8.570206171825188e-07, "loss": 0.5644, "num_input_tokens_seen": 2304770048, "step": 1099 }, { "epoch": 0.027122321670735014, "grad_norm": 13.551562309265137, "learning_rate": 8.567253562511306e-07, "loss": 0.439, "num_input_tokens_seen": 2306867200, "step": 1100 }, { "epoch": 0.027393544887442365, "grad_norm": 18.999597549438477, "learning_rate": 8.564298484788472e-07, "loss": 0.4937, "num_input_tokens_seen": 2308964352, "step": 1101 }, { "epoch": 0.027664768104149716, "grad_norm": 10.107467651367188, "learning_rate": 8.561340941034825e-07, "loss": 0.2738, "num_input_tokens_seen": 2311061504, "step": 1102 }, { "epoch": 0.027935991320857067, "grad_norm": 19.000905990600586, "learning_rate": 8.55838093363048e-07, "loss": 0.6454, "num_input_tokens_seen": 2313158656, "step": 1103 }, { "epoch": 0.028207214537564414, "grad_norm": 31.94487190246582, "learning_rate": 8.555418464957542e-07, "loss": 0.6641, "num_input_tokens_seen": 2315255808, "step": 1104 }, { "epoch": 0.028478437754271765, "grad_norm": 16.993053436279297, "learning_rate": 8.552453537400089e-07, "loss": 0.4975, "num_input_tokens_seen": 2317352960, "step": 1105 }, { "epoch": 0.028749660970979116, "grad_norm": 21.718708038330078, "learning_rate": 8.549486153344183e-07, "loss": 0.7325, "num_input_tokens_seen": 2319450112, "step": 1106 }, { "epoch": 0.029020884187686467, "grad_norm": 16.96457862854004, "learning_rate": 8.546516315177863e-07, "loss": 0.5915, "num_input_tokens_seen": 2321547264, "step": 1107 }, { "epoch": 0.029292107404393815, "grad_norm": 15.942675590515137, "learning_rate": 8.543544025291143e-07, "loss": 0.5334, "num_input_tokens_seen": 2323644416, "step": 1108 }, { "epoch": 0.029563330621101166, "grad_norm": 14.432857513427734, "learning_rate": 8.540569286076004e-07, "loss": 0.5069, "num_input_tokens_seen": 2325741568, "step": 1109 }, { "epoch": 0.029834553837808517, "grad_norm": 16.693790435791016, "learning_rate": 8.537592099926407e-07, "loss": 0.6163, "num_input_tokens_seen": 2327838720, "step": 1110 }, { "epoch": 0.030105777054515868, "grad_norm": 13.320590019226074, "learning_rate": 8.534612469238278e-07, "loss": 0.3806, "num_input_tokens_seen": 2329935872, "step": 1111 }, { "epoch": 0.030377000271223215, "grad_norm": 20.88117790222168, "learning_rate": 8.531630396409507e-07, "loss": 0.6614, "num_input_tokens_seen": 2332033024, "step": 1112 }, { "epoch": 0.030648223487930566, "grad_norm": 12.577048301696777, "learning_rate": 8.528645883839956e-07, "loss": 0.489, "num_input_tokens_seen": 2334130176, "step": 1113 }, { "epoch": 0.030919446704637917, "grad_norm": 16.97111701965332, "learning_rate": 8.525658933931448e-07, "loss": 0.5687, "num_input_tokens_seen": 2336227328, "step": 1114 }, { "epoch": 0.031190669921345268, "grad_norm": 15.8145112991333, "learning_rate": 8.522669549087762e-07, "loss": 0.556, "num_input_tokens_seen": 2338324480, "step": 1115 }, { "epoch": 0.031461893138052616, "grad_norm": 16.058319091796875, "learning_rate": 8.519677731714645e-07, "loss": 0.7402, "num_input_tokens_seen": 2340421632, "step": 1116 }, { "epoch": 0.03173311635475997, "grad_norm": 11.520198822021484, "learning_rate": 8.516683484219797e-07, "loss": 0.4213, "num_input_tokens_seen": 2342518784, "step": 1117 }, { "epoch": 0.03200433957146732, "grad_norm": 18.839998245239258, "learning_rate": 8.513686809012875e-07, "loss": 0.5229, "num_input_tokens_seen": 2344615936, "step": 1118 }, { "epoch": 0.032275562788174665, "grad_norm": 16.22575569152832, "learning_rate": 8.510687708505489e-07, "loss": 0.5792, "num_input_tokens_seen": 2346713088, "step": 1119 }, { "epoch": 0.03254678600488202, "grad_norm": 15.674661636352539, "learning_rate": 8.507686185111199e-07, "loss": 0.5386, "num_input_tokens_seen": 2348810240, "step": 1120 }, { "epoch": 0.03281800922158937, "grad_norm": 18.9416446685791, "learning_rate": 8.504682241245516e-07, "loss": 0.3716, "num_input_tokens_seen": 2350907392, "step": 1121 }, { "epoch": 0.03308923243829672, "grad_norm": 10.184718132019043, "learning_rate": 8.501675879325906e-07, "loss": 0.2198, "num_input_tokens_seen": 2353004544, "step": 1122 }, { "epoch": 0.03336045565500407, "grad_norm": 12.775616645812988, "learning_rate": 8.498667101771769e-07, "loss": 0.4381, "num_input_tokens_seen": 2355101696, "step": 1123 }, { "epoch": 0.033631678871711417, "grad_norm": 19.200225830078125, "learning_rate": 8.495655911004456e-07, "loss": 0.6905, "num_input_tokens_seen": 2357198848, "step": 1124 }, { "epoch": 0.03390290208841877, "grad_norm": 16.312833786010742, "learning_rate": 8.492642309447257e-07, "loss": 0.551, "num_input_tokens_seen": 2359296000, "step": 1125 }, { "epoch": 0.03417412530512612, "grad_norm": 12.200955390930176, "learning_rate": 8.489626299525409e-07, "loss": 0.3892, "num_input_tokens_seen": 2361393152, "step": 1126 }, { "epoch": 0.034445348521833466, "grad_norm": 12.41970157623291, "learning_rate": 8.486607883666077e-07, "loss": 0.4058, "num_input_tokens_seen": 2363490304, "step": 1127 }, { "epoch": 0.03471657173854082, "grad_norm": 17.284793853759766, "learning_rate": 8.483587064298372e-07, "loss": 0.5063, "num_input_tokens_seen": 2365587456, "step": 1128 }, { "epoch": 0.03498779495524817, "grad_norm": 12.992026329040527, "learning_rate": 8.480563843853328e-07, "loss": 0.3901, "num_input_tokens_seen": 2367684608, "step": 1129 }, { "epoch": 0.03525901817195552, "grad_norm": 25.536312103271484, "learning_rate": 8.477538224763923e-07, "loss": 0.6989, "num_input_tokens_seen": 2369781760, "step": 1130 }, { "epoch": 0.03553024138866287, "grad_norm": 14.612037658691406, "learning_rate": 8.474510209465058e-07, "loss": 0.4448, "num_input_tokens_seen": 2371878912, "step": 1131 }, { "epoch": 0.03580146460537022, "grad_norm": 17.197641372680664, "learning_rate": 8.471479800393565e-07, "loss": 0.5952, "num_input_tokens_seen": 2373976064, "step": 1132 }, { "epoch": 0.03607268782207757, "grad_norm": 15.149774551391602, "learning_rate": 8.468446999988202e-07, "loss": 0.5679, "num_input_tokens_seen": 2376073216, "step": 1133 }, { "epoch": 0.03634391103878492, "grad_norm": 20.161657333374023, "learning_rate": 8.465411810689653e-07, "loss": 0.9363, "num_input_tokens_seen": 2378170368, "step": 1134 }, { "epoch": 0.03661513425549227, "grad_norm": 13.918910026550293, "learning_rate": 8.462374234940517e-07, "loss": 0.4595, "num_input_tokens_seen": 2380267520, "step": 1135 }, { "epoch": 0.03688635747219962, "grad_norm": 11.941801071166992, "learning_rate": 8.459334275185325e-07, "loss": 0.3769, "num_input_tokens_seen": 2382364672, "step": 1136 }, { "epoch": 0.03715758068890697, "grad_norm": 23.785158157348633, "learning_rate": 8.456291933870521e-07, "loss": 1.0131, "num_input_tokens_seen": 2384461824, "step": 1137 }, { "epoch": 0.03742880390561432, "grad_norm": 12.919363021850586, "learning_rate": 8.453247213444463e-07, "loss": 0.2438, "num_input_tokens_seen": 2386558976, "step": 1138 }, { "epoch": 0.03770002712232167, "grad_norm": 14.640851020812988, "learning_rate": 8.450200116357428e-07, "loss": 0.4749, "num_input_tokens_seen": 2388656128, "step": 1139 }, { "epoch": 0.03797125033902902, "grad_norm": 10.543968200683594, "learning_rate": 8.4471506450616e-07, "loss": 0.3616, "num_input_tokens_seen": 2390753280, "step": 1140 }, { "epoch": 0.03824247355573637, "grad_norm": 15.393962860107422, "learning_rate": 8.444098802011083e-07, "loss": 0.4972, "num_input_tokens_seen": 2392850432, "step": 1141 }, { "epoch": 0.03851369677244372, "grad_norm": 79.32931518554688, "learning_rate": 8.441044589661881e-07, "loss": 0.5945, "num_input_tokens_seen": 2394947584, "step": 1142 }, { "epoch": 0.038784919989151075, "grad_norm": 13.988775253295898, "learning_rate": 8.437988010471907e-07, "loss": 0.5593, "num_input_tokens_seen": 2397044736, "step": 1143 }, { "epoch": 0.03905614320585842, "grad_norm": 14.631237983703613, "learning_rate": 8.434929066900982e-07, "loss": 0.5196, "num_input_tokens_seen": 2399141888, "step": 1144 }, { "epoch": 0.03932736642256577, "grad_norm": 23.943798065185547, "learning_rate": 8.431867761410826e-07, "loss": 0.8291, "num_input_tokens_seen": 2401239040, "step": 1145 }, { "epoch": 0.039598589639273124, "grad_norm": 13.034646034240723, "learning_rate": 8.42880409646506e-07, "loss": 0.4124, "num_input_tokens_seen": 2403336192, "step": 1146 }, { "epoch": 0.03986981285598047, "grad_norm": 16.411720275878906, "learning_rate": 8.42573807452921e-07, "loss": 0.4623, "num_input_tokens_seen": 2405433344, "step": 1147 }, { "epoch": 0.04014103607268782, "grad_norm": 18.51169204711914, "learning_rate": 8.422669698070687e-07, "loss": 0.5727, "num_input_tokens_seen": 2407530496, "step": 1148 }, { "epoch": 0.040412259289395173, "grad_norm": 10.572891235351562, "learning_rate": 8.419598969558808e-07, "loss": 0.2515, "num_input_tokens_seen": 2409627648, "step": 1149 }, { "epoch": 0.04068348250610252, "grad_norm": 10.220544815063477, "learning_rate": 8.416525891464776e-07, "loss": 0.3125, "num_input_tokens_seen": 2411724800, "step": 1150 }, { "epoch": 0.040954705722809875, "grad_norm": 14.300618171691895, "learning_rate": 8.413450466261691e-07, "loss": 0.429, "num_input_tokens_seen": 2413821952, "step": 1151 }, { "epoch": 0.04122592893951722, "grad_norm": 14.614937782287598, "learning_rate": 8.410372696424535e-07, "loss": 0.3731, "num_input_tokens_seen": 2415919104, "step": 1152 }, { "epoch": 0.04149715215622457, "grad_norm": 12.267914772033691, "learning_rate": 8.40729258443018e-07, "loss": 0.4024, "num_input_tokens_seen": 2418016256, "step": 1153 }, { "epoch": 0.041768375372931925, "grad_norm": 12.720690727233887, "learning_rate": 8.404210132757385e-07, "loss": 0.3267, "num_input_tokens_seen": 2420113408, "step": 1154 }, { "epoch": 0.04203959858963927, "grad_norm": 15.737774848937988, "learning_rate": 8.401125343886787e-07, "loss": 0.5473, "num_input_tokens_seen": 2422210560, "step": 1155 }, { "epoch": 0.04231082180634662, "grad_norm": 9.714899063110352, "learning_rate": 8.398038220300908e-07, "loss": 0.3099, "num_input_tokens_seen": 2424307712, "step": 1156 }, { "epoch": 0.042582045023053974, "grad_norm": 18.72726058959961, "learning_rate": 8.39494876448415e-07, "loss": 0.7204, "num_input_tokens_seen": 2426404864, "step": 1157 }, { "epoch": 0.04285326823976132, "grad_norm": 19.449329376220703, "learning_rate": 8.391856978922785e-07, "loss": 0.8302, "num_input_tokens_seen": 2428502016, "step": 1158 }, { "epoch": 0.043124491456468676, "grad_norm": 12.25051212310791, "learning_rate": 8.38876286610497e-07, "loss": 0.2353, "num_input_tokens_seen": 2430599168, "step": 1159 }, { "epoch": 0.043395714673176024, "grad_norm": 17.40892791748047, "learning_rate": 8.385666428520723e-07, "loss": 0.6826, "num_input_tokens_seen": 2432696320, "step": 1160 }, { "epoch": 0.04366693788988337, "grad_norm": 20.962051391601562, "learning_rate": 8.382567668661943e-07, "loss": 0.694, "num_input_tokens_seen": 2434793472, "step": 1161 }, { "epoch": 0.043938161106590726, "grad_norm": 17.564403533935547, "learning_rate": 8.379466589022393e-07, "loss": 0.6455, "num_input_tokens_seen": 2436890624, "step": 1162 }, { "epoch": 0.04420938432329807, "grad_norm": 16.586353302001953, "learning_rate": 8.376363192097703e-07, "loss": 0.5219, "num_input_tokens_seen": 2438987776, "step": 1163 }, { "epoch": 0.04448060754000543, "grad_norm": 20.66847801208496, "learning_rate": 8.37325748038537e-07, "loss": 0.6066, "num_input_tokens_seen": 2441084928, "step": 1164 }, { "epoch": 0.044751830756712775, "grad_norm": 17.191225051879883, "learning_rate": 8.370149456384754e-07, "loss": 0.4234, "num_input_tokens_seen": 2443182080, "step": 1165 }, { "epoch": 0.04502305397342012, "grad_norm": 13.013045310974121, "learning_rate": 8.36703912259707e-07, "loss": 0.3087, "num_input_tokens_seen": 2445279232, "step": 1166 }, { "epoch": 0.04529427719012748, "grad_norm": 22.083459854125977, "learning_rate": 8.363926481525402e-07, "loss": 0.7541, "num_input_tokens_seen": 2447376384, "step": 1167 }, { "epoch": 0.045565500406834825, "grad_norm": 14.312165260314941, "learning_rate": 8.360811535674682e-07, "loss": 0.4852, "num_input_tokens_seen": 2449473536, "step": 1168 }, { "epoch": 0.04583672362354217, "grad_norm": 20.598003387451172, "learning_rate": 8.357694287551698e-07, "loss": 0.7828, "num_input_tokens_seen": 2451570688, "step": 1169 }, { "epoch": 0.04610794684024953, "grad_norm": 19.325916290283203, "learning_rate": 8.354574739665096e-07, "loss": 0.4868, "num_input_tokens_seen": 2453667840, "step": 1170 }, { "epoch": 0.046379170056956874, "grad_norm": 12.34057903289795, "learning_rate": 8.351452894525368e-07, "loss": 0.3321, "num_input_tokens_seen": 2455764992, "step": 1171 }, { "epoch": 0.04665039327366423, "grad_norm": 12.274550437927246, "learning_rate": 8.348328754644855e-07, "loss": 0.4123, "num_input_tokens_seen": 2457862144, "step": 1172 }, { "epoch": 0.046921616490371576, "grad_norm": 14.192599296569824, "learning_rate": 8.34520232253775e-07, "loss": 0.5802, "num_input_tokens_seen": 2459959296, "step": 1173 }, { "epoch": 0.047192839707078924, "grad_norm": 20.330703735351562, "learning_rate": 8.342073600720082e-07, "loss": 0.6071, "num_input_tokens_seen": 2462056448, "step": 1174 }, { "epoch": 0.04746406292378628, "grad_norm": 16.746747970581055, "learning_rate": 8.33894259170973e-07, "loss": 0.4686, "num_input_tokens_seen": 2464153600, "step": 1175 }, { "epoch": 0.047735286140493625, "grad_norm": 13.190831184387207, "learning_rate": 8.335809298026409e-07, "loss": 0.329, "num_input_tokens_seen": 2466250752, "step": 1176 }, { "epoch": 0.04800650935720097, "grad_norm": 14.298337936401367, "learning_rate": 8.332673722191677e-07, "loss": 0.396, "num_input_tokens_seen": 2468347904, "step": 1177 }, { "epoch": 0.04827773257390833, "grad_norm": 14.075898170471191, "learning_rate": 8.329535866728922e-07, "loss": 0.4024, "num_input_tokens_seen": 2470445056, "step": 1178 }, { "epoch": 0.048548955790615675, "grad_norm": 15.5481595993042, "learning_rate": 8.326395734163375e-07, "loss": 0.6025, "num_input_tokens_seen": 2472542208, "step": 1179 }, { "epoch": 0.04882017900732303, "grad_norm": 12.739563941955566, "learning_rate": 8.323253327022094e-07, "loss": 0.3926, "num_input_tokens_seen": 2474639360, "step": 1180 }, { "epoch": 0.04909140222403038, "grad_norm": 10.208687782287598, "learning_rate": 8.320108647833967e-07, "loss": 0.3176, "num_input_tokens_seen": 2476736512, "step": 1181 }, { "epoch": 0.049362625440737724, "grad_norm": 9.83712387084961, "learning_rate": 8.316961699129714e-07, "loss": 0.247, "num_input_tokens_seen": 2478833664, "step": 1182 }, { "epoch": 0.04963384865744508, "grad_norm": 19.233394622802734, "learning_rate": 8.313812483441879e-07, "loss": 0.7097, "num_input_tokens_seen": 2480930816, "step": 1183 }, { "epoch": 0.049905071874152426, "grad_norm": 12.595343589782715, "learning_rate": 8.310661003304829e-07, "loss": 0.33, "num_input_tokens_seen": 2483027968, "step": 1184 }, { "epoch": 0.05017629509085978, "grad_norm": 14.636686325073242, "learning_rate": 8.30750726125476e-07, "loss": 0.4683, "num_input_tokens_seen": 2485125120, "step": 1185 }, { "epoch": 0.05044751830756713, "grad_norm": 13.301069259643555, "learning_rate": 8.304351259829678e-07, "loss": 0.4655, "num_input_tokens_seen": 2487222272, "step": 1186 }, { "epoch": 0.050718741524274476, "grad_norm": 13.354597091674805, "learning_rate": 8.301193001569418e-07, "loss": 0.3759, "num_input_tokens_seen": 2489319424, "step": 1187 }, { "epoch": 0.05098996474098183, "grad_norm": 12.58713150024414, "learning_rate": 8.298032489015623e-07, "loss": 0.3448, "num_input_tokens_seen": 2491416576, "step": 1188 }, { "epoch": 0.05126118795768918, "grad_norm": 19.054805755615234, "learning_rate": 8.294869724711752e-07, "loss": 0.586, "num_input_tokens_seen": 2493513728, "step": 1189 }, { "epoch": 0.051532411174396525, "grad_norm": 11.936301231384277, "learning_rate": 8.291704711203082e-07, "loss": 0.3833, "num_input_tokens_seen": 2495610880, "step": 1190 }, { "epoch": 0.05180363439110388, "grad_norm": 18.23284912109375, "learning_rate": 8.288537451036691e-07, "loss": 0.4227, "num_input_tokens_seen": 2497708032, "step": 1191 }, { "epoch": 0.05207485760781123, "grad_norm": 14.435386657714844, "learning_rate": 8.28536794676147e-07, "loss": 0.3232, "num_input_tokens_seen": 2499805184, "step": 1192 }, { "epoch": 0.05234608082451858, "grad_norm": 13.897747039794922, "learning_rate": 8.282196200928119e-07, "loss": 0.4033, "num_input_tokens_seen": 2501902336, "step": 1193 }, { "epoch": 0.05261730404122593, "grad_norm": 11.174321174621582, "learning_rate": 8.279022216089135e-07, "loss": 0.3174, "num_input_tokens_seen": 2503999488, "step": 1194 }, { "epoch": 0.05288852725793328, "grad_norm": 15.450333595275879, "learning_rate": 8.275845994798821e-07, "loss": 0.4879, "num_input_tokens_seen": 2506096640, "step": 1195 }, { "epoch": 0.05315975047464063, "grad_norm": 12.975201606750488, "learning_rate": 8.272667539613281e-07, "loss": 0.3927, "num_input_tokens_seen": 2508193792, "step": 1196 }, { "epoch": 0.05343097369134798, "grad_norm": 20.552017211914062, "learning_rate": 8.26948685309041e-07, "loss": 0.6859, "num_input_tokens_seen": 2510290944, "step": 1197 }, { "epoch": 0.05370219690805533, "grad_norm": 14.239463806152344, "learning_rate": 8.266303937789908e-07, "loss": 0.4479, "num_input_tokens_seen": 2512388096, "step": 1198 }, { "epoch": 0.05397342012476268, "grad_norm": 13.604122161865234, "learning_rate": 8.263118796273263e-07, "loss": 0.5931, "num_input_tokens_seen": 2514485248, "step": 1199 }, { "epoch": 0.05424464334147003, "grad_norm": 22.870668411254883, "learning_rate": 8.259931431103754e-07, "loss": 0.4736, "num_input_tokens_seen": 2516582400, "step": 1200 }, { "epoch": 0.00027122321670735016, "grad_norm": 16.450984954833984, "learning_rate": 8.256741844846452e-07, "loss": 0.5646, "num_input_tokens_seen": 2518679552, "step": 1201 }, { "epoch": 0.0005424464334147003, "grad_norm": 15.368687629699707, "learning_rate": 8.253550040068216e-07, "loss": 0.4454, "num_input_tokens_seen": 2520776704, "step": 1202 }, { "epoch": 0.0008136696501220504, "grad_norm": 15.554755210876465, "learning_rate": 8.250356019337688e-07, "loss": 0.4901, "num_input_tokens_seen": 2522873856, "step": 1203 }, { "epoch": 0.0010848928668294006, "grad_norm": 19.45119285583496, "learning_rate": 8.247159785225295e-07, "loss": 0.8747, "num_input_tokens_seen": 2524971008, "step": 1204 }, { "epoch": 0.0013561160835367507, "grad_norm": 19.58152198791504, "learning_rate": 8.243961340303245e-07, "loss": 0.6606, "num_input_tokens_seen": 2527068160, "step": 1205 }, { "epoch": 0.0016273393002441008, "grad_norm": 7.020671367645264, "learning_rate": 8.240760687145521e-07, "loss": 0.1893, "num_input_tokens_seen": 2529165312, "step": 1206 }, { "epoch": 0.001898562516951451, "grad_norm": 14.452476501464844, "learning_rate": 8.237557828327891e-07, "loss": 0.5641, "num_input_tokens_seen": 2531262464, "step": 1207 }, { "epoch": 0.0021697857336588013, "grad_norm": 17.215316772460938, "learning_rate": 8.234352766427894e-07, "loss": 0.6351, "num_input_tokens_seen": 2533359616, "step": 1208 }, { "epoch": 0.0024410089503661514, "grad_norm": 16.62022590637207, "learning_rate": 8.231145504024838e-07, "loss": 0.6488, "num_input_tokens_seen": 2535456768, "step": 1209 }, { "epoch": 0.0027122321670735015, "grad_norm": 9.606282234191895, "learning_rate": 8.22793604369981e-07, "loss": 0.353, "num_input_tokens_seen": 2537553920, "step": 1210 }, { "epoch": 0.0029834553837808516, "grad_norm": 12.205081939697266, "learning_rate": 8.224724388035659e-07, "loss": 0.3574, "num_input_tokens_seen": 2539651072, "step": 1211 }, { "epoch": 0.0032546786004882017, "grad_norm": 16.890460968017578, "learning_rate": 8.221510539617003e-07, "loss": 0.4982, "num_input_tokens_seen": 2541748224, "step": 1212 }, { "epoch": 0.003525901817195552, "grad_norm": 11.31159496307373, "learning_rate": 8.218294501030226e-07, "loss": 0.3511, "num_input_tokens_seen": 2543845376, "step": 1213 }, { "epoch": 0.003797125033902902, "grad_norm": 14.827753067016602, "learning_rate": 8.215076274863476e-07, "loss": 0.5236, "num_input_tokens_seen": 2545942528, "step": 1214 }, { "epoch": 0.0040683482506102524, "grad_norm": 13.192427635192871, "learning_rate": 8.211855863706654e-07, "loss": 0.4119, "num_input_tokens_seen": 2548039680, "step": 1215 }, { "epoch": 0.0043395714673176026, "grad_norm": 14.75649356842041, "learning_rate": 8.208633270151426e-07, "loss": 0.5348, "num_input_tokens_seen": 2550136832, "step": 1216 }, { "epoch": 0.004610794684024953, "grad_norm": 20.833091735839844, "learning_rate": 8.205408496791216e-07, "loss": 0.5757, "num_input_tokens_seen": 2552233984, "step": 1217 }, { "epoch": 0.004882017900732303, "grad_norm": 11.868510246276855, "learning_rate": 8.202181546221193e-07, "loss": 0.3579, "num_input_tokens_seen": 2554331136, "step": 1218 }, { "epoch": 0.005153241117439653, "grad_norm": 17.338224411010742, "learning_rate": 8.19895242103829e-07, "loss": 0.6951, "num_input_tokens_seen": 2556428288, "step": 1219 }, { "epoch": 0.005424464334147003, "grad_norm": 23.777769088745117, "learning_rate": 8.19572112384118e-07, "loss": 0.6578, "num_input_tokens_seen": 2558525440, "step": 1220 }, { "epoch": 0.005695687550854353, "grad_norm": 13.37986946105957, "learning_rate": 8.192487657230288e-07, "loss": 0.3095, "num_input_tokens_seen": 2560622592, "step": 1221 }, { "epoch": 0.005966910767561703, "grad_norm": 12.245566368103027, "learning_rate": 8.18925202380779e-07, "loss": 0.3178, "num_input_tokens_seen": 2562719744, "step": 1222 }, { "epoch": 0.006238133984269053, "grad_norm": 13.670166969299316, "learning_rate": 8.186014226177594e-07, "loss": 0.491, "num_input_tokens_seen": 2564816896, "step": 1223 }, { "epoch": 0.006509357200976403, "grad_norm": 13.172407150268555, "learning_rate": 8.18277426694536e-07, "loss": 0.4929, "num_input_tokens_seen": 2566914048, "step": 1224 }, { "epoch": 0.0067805804176837535, "grad_norm": 14.171710014343262, "learning_rate": 8.179532148718483e-07, "loss": 0.5016, "num_input_tokens_seen": 2569011200, "step": 1225 }, { "epoch": 0.007051803634391104, "grad_norm": 25.743188858032227, "learning_rate": 8.176287874106097e-07, "loss": 0.6181, "num_input_tokens_seen": 2571108352, "step": 1226 }, { "epoch": 0.007323026851098454, "grad_norm": 16.81450653076172, "learning_rate": 8.173041445719069e-07, "loss": 0.6927, "num_input_tokens_seen": 2573205504, "step": 1227 }, { "epoch": 0.007594250067805804, "grad_norm": 21.463520050048828, "learning_rate": 8.169792866170003e-07, "loss": 0.817, "num_input_tokens_seen": 2575302656, "step": 1228 }, { "epoch": 0.007865473284513154, "grad_norm": 12.594096183776855, "learning_rate": 8.166542138073232e-07, "loss": 0.5051, "num_input_tokens_seen": 2577399808, "step": 1229 }, { "epoch": 0.008136696501220505, "grad_norm": 14.551346778869629, "learning_rate": 8.163289264044817e-07, "loss": 0.5802, "num_input_tokens_seen": 2579496960, "step": 1230 }, { "epoch": 0.008407919717927854, "grad_norm": 14.674225807189941, "learning_rate": 8.160034246702548e-07, "loss": 0.5542, "num_input_tokens_seen": 2581594112, "step": 1231 }, { "epoch": 0.008679142934635205, "grad_norm": 14.190738677978516, "learning_rate": 8.156777088665939e-07, "loss": 0.44, "num_input_tokens_seen": 2583691264, "step": 1232 }, { "epoch": 0.008950366151342554, "grad_norm": 11.013334274291992, "learning_rate": 8.153517792556226e-07, "loss": 0.2932, "num_input_tokens_seen": 2585788416, "step": 1233 }, { "epoch": 0.009221589368049905, "grad_norm": 16.620725631713867, "learning_rate": 8.15025636099637e-07, "loss": 0.6058, "num_input_tokens_seen": 2587885568, "step": 1234 }, { "epoch": 0.009492812584757255, "grad_norm": 13.847301483154297, "learning_rate": 8.146992796611042e-07, "loss": 0.3527, "num_input_tokens_seen": 2589982720, "step": 1235 }, { "epoch": 0.009764035801464606, "grad_norm": 18.135631561279297, "learning_rate": 8.143727102026638e-07, "loss": 0.852, "num_input_tokens_seen": 2592079872, "step": 1236 }, { "epoch": 0.010035259018171955, "grad_norm": 11.433444023132324, "learning_rate": 8.140459279871264e-07, "loss": 0.2981, "num_input_tokens_seen": 2594177024, "step": 1237 }, { "epoch": 0.010306482234879306, "grad_norm": 20.82098960876465, "learning_rate": 8.137189332774738e-07, "loss": 0.7318, "num_input_tokens_seen": 2596274176, "step": 1238 }, { "epoch": 0.010577705451586655, "grad_norm": 12.118500709533691, "learning_rate": 8.133917263368589e-07, "loss": 0.3293, "num_input_tokens_seen": 2598371328, "step": 1239 }, { "epoch": 0.010848928668294006, "grad_norm": 19.532363891601562, "learning_rate": 8.130643074286056e-07, "loss": 0.458, "num_input_tokens_seen": 2600468480, "step": 1240 }, { "epoch": 0.011120151885001357, "grad_norm": 16.201066970825195, "learning_rate": 8.127366768162077e-07, "loss": 0.4731, "num_input_tokens_seen": 2602565632, "step": 1241 }, { "epoch": 0.011391375101708706, "grad_norm": 13.628470420837402, "learning_rate": 8.124088347633304e-07, "loss": 0.4137, "num_input_tokens_seen": 2604662784, "step": 1242 }, { "epoch": 0.011662598318416057, "grad_norm": 12.295647621154785, "learning_rate": 8.120807815338083e-07, "loss": 0.3935, "num_input_tokens_seen": 2606759936, "step": 1243 }, { "epoch": 0.011933821535123406, "grad_norm": 12.40659236907959, "learning_rate": 8.11752517391646e-07, "loss": 0.3719, "num_input_tokens_seen": 2608857088, "step": 1244 }, { "epoch": 0.012205044751830757, "grad_norm": 12.558025360107422, "learning_rate": 8.114240426010183e-07, "loss": 0.3506, "num_input_tokens_seen": 2610954240, "step": 1245 }, { "epoch": 0.012476267968538107, "grad_norm": 18.552087783813477, "learning_rate": 8.11095357426269e-07, "loss": 0.6331, "num_input_tokens_seen": 2613051392, "step": 1246 }, { "epoch": 0.012747491185245458, "grad_norm": 14.407464981079102, "learning_rate": 8.107664621319113e-07, "loss": 0.5219, "num_input_tokens_seen": 2615148544, "step": 1247 }, { "epoch": 0.013018714401952807, "grad_norm": 14.350251197814941, "learning_rate": 8.10437356982628e-07, "loss": 0.4026, "num_input_tokens_seen": 2617245696, "step": 1248 }, { "epoch": 0.013289937618660158, "grad_norm": 13.879180908203125, "learning_rate": 8.1010804224327e-07, "loss": 0.4596, "num_input_tokens_seen": 2619342848, "step": 1249 }, { "epoch": 0.013561160835367507, "grad_norm": 12.347127914428711, "learning_rate": 8.097785181788574e-07, "loss": 0.3675, "num_input_tokens_seen": 2621440000, "step": 1250 }, { "epoch": 0.013832384052074858, "grad_norm": 13.129162788391113, "learning_rate": 8.09448785054579e-07, "loss": 0.3485, "num_input_tokens_seen": 2623537152, "step": 1251 }, { "epoch": 0.014103607268782207, "grad_norm": 19.645524978637695, "learning_rate": 8.091188431357908e-07, "loss": 0.9547, "num_input_tokens_seen": 2625634304, "step": 1252 }, { "epoch": 0.014374830485489558, "grad_norm": 19.11149787902832, "learning_rate": 8.087886926880181e-07, "loss": 0.8976, "num_input_tokens_seen": 2627731456, "step": 1253 }, { "epoch": 0.014646053702196907, "grad_norm": 12.85779094696045, "learning_rate": 8.084583339769531e-07, "loss": 0.4815, "num_input_tokens_seen": 2629828608, "step": 1254 }, { "epoch": 0.014917276918904258, "grad_norm": 13.494819641113281, "learning_rate": 8.081277672684557e-07, "loss": 0.4236, "num_input_tokens_seen": 2631925760, "step": 1255 }, { "epoch": 0.015188500135611608, "grad_norm": 13.848788261413574, "learning_rate": 8.077969928285541e-07, "loss": 0.5823, "num_input_tokens_seen": 2634022912, "step": 1256 }, { "epoch": 0.015459723352318959, "grad_norm": 15.541146278381348, "learning_rate": 8.074660109234424e-07, "loss": 0.456, "num_input_tokens_seen": 2636120064, "step": 1257 }, { "epoch": 0.015730946569026308, "grad_norm": 18.899259567260742, "learning_rate": 8.071348218194823e-07, "loss": 0.5888, "num_input_tokens_seen": 2638217216, "step": 1258 }, { "epoch": 0.01600216978573366, "grad_norm": 17.18483543395996, "learning_rate": 8.068034257832026e-07, "loss": 0.7188, "num_input_tokens_seen": 2640314368, "step": 1259 }, { "epoch": 0.01627339300244101, "grad_norm": 15.19528579711914, "learning_rate": 8.064718230812976e-07, "loss": 0.4918, "num_input_tokens_seen": 2642411520, "step": 1260 }, { "epoch": 0.01654461621914836, "grad_norm": 16.64455795288086, "learning_rate": 8.06140013980629e-07, "loss": 0.7336, "num_input_tokens_seen": 2644508672, "step": 1261 }, { "epoch": 0.016815839435855708, "grad_norm": 10.2529296875, "learning_rate": 8.05807998748224e-07, "loss": 0.3127, "num_input_tokens_seen": 2646605824, "step": 1262 }, { "epoch": 0.01708706265256306, "grad_norm": 10.849603652954102, "learning_rate": 8.05475777651276e-07, "loss": 0.3275, "num_input_tokens_seen": 2648702976, "step": 1263 }, { "epoch": 0.01735828586927041, "grad_norm": 18.223087310791016, "learning_rate": 8.051433509571435e-07, "loss": 0.6497, "num_input_tokens_seen": 2650800128, "step": 1264 }, { "epoch": 0.01762950908597776, "grad_norm": 16.752042770385742, "learning_rate": 8.04810718933351e-07, "loss": 0.5918, "num_input_tokens_seen": 2652897280, "step": 1265 }, { "epoch": 0.01790073230268511, "grad_norm": 12.526599884033203, "learning_rate": 8.044778818475884e-07, "loss": 0.3929, "num_input_tokens_seen": 2654994432, "step": 1266 }, { "epoch": 0.01817195551939246, "grad_norm": 15.401561737060547, "learning_rate": 8.0414483996771e-07, "loss": 0.6053, "num_input_tokens_seen": 2657091584, "step": 1267 }, { "epoch": 0.01844317873609981, "grad_norm": 18.10499382019043, "learning_rate": 8.038115935617355e-07, "loss": 0.8571, "num_input_tokens_seen": 2659188736, "step": 1268 }, { "epoch": 0.01871440195280716, "grad_norm": 11.17979621887207, "learning_rate": 8.034781428978484e-07, "loss": 0.3619, "num_input_tokens_seen": 2661285888, "step": 1269 }, { "epoch": 0.01898562516951451, "grad_norm": 17.378992080688477, "learning_rate": 8.031444882443976e-07, "loss": 0.231, "num_input_tokens_seen": 2663383040, "step": 1270 }, { "epoch": 0.01925684838622186, "grad_norm": 17.07963752746582, "learning_rate": 8.028106298698957e-07, "loss": 0.6219, "num_input_tokens_seen": 2665480192, "step": 1271 }, { "epoch": 0.01952807160292921, "grad_norm": 15.965147972106934, "learning_rate": 8.024765680430188e-07, "loss": 0.3175, "num_input_tokens_seen": 2667577344, "step": 1272 }, { "epoch": 0.019799294819636562, "grad_norm": 13.383700370788574, "learning_rate": 8.021423030326075e-07, "loss": 0.5492, "num_input_tokens_seen": 2669674496, "step": 1273 }, { "epoch": 0.02007051803634391, "grad_norm": 8.79860782623291, "learning_rate": 8.018078351076653e-07, "loss": 0.2271, "num_input_tokens_seen": 2671771648, "step": 1274 }, { "epoch": 0.02034174125305126, "grad_norm": 10.821599006652832, "learning_rate": 8.014731645373595e-07, "loss": 0.3977, "num_input_tokens_seen": 2673868800, "step": 1275 }, { "epoch": 0.02061296446975861, "grad_norm": 15.333173751831055, "learning_rate": 8.011382915910203e-07, "loss": 0.4958, "num_input_tokens_seen": 2675965952, "step": 1276 }, { "epoch": 0.020884187686465962, "grad_norm": 10.879945755004883, "learning_rate": 8.008032165381403e-07, "loss": 0.3322, "num_input_tokens_seen": 2678063104, "step": 1277 }, { "epoch": 0.02115541090317331, "grad_norm": 16.250761032104492, "learning_rate": 8.004679396483756e-07, "loss": 0.266, "num_input_tokens_seen": 2680160256, "step": 1278 }, { "epoch": 0.02142663411988066, "grad_norm": 20.67441749572754, "learning_rate": 8.001324611915441e-07, "loss": 0.82, "num_input_tokens_seen": 2682257408, "step": 1279 }, { "epoch": 0.021697857336588012, "grad_norm": 14.254646301269531, "learning_rate": 7.99796781437626e-07, "loss": 0.4862, "num_input_tokens_seen": 2684354560, "step": 1280 }, { "epoch": 0.021969080553295363, "grad_norm": 11.306888580322266, "learning_rate": 7.994609006567635e-07, "loss": 0.287, "num_input_tokens_seen": 2686451712, "step": 1281 }, { "epoch": 0.022240303770002714, "grad_norm": 17.26317596435547, "learning_rate": 7.99124819119261e-07, "loss": 0.4729, "num_input_tokens_seen": 2688548864, "step": 1282 }, { "epoch": 0.02251152698671006, "grad_norm": 19.987943649291992, "learning_rate": 7.987885370955841e-07, "loss": 0.6808, "num_input_tokens_seen": 2690646016, "step": 1283 }, { "epoch": 0.022782750203417412, "grad_norm": 16.031089782714844, "learning_rate": 7.984520548563594e-07, "loss": 0.6981, "num_input_tokens_seen": 2692743168, "step": 1284 }, { "epoch": 0.023053973420124763, "grad_norm": 14.085775375366211, "learning_rate": 7.981153726723755e-07, "loss": 0.4403, "num_input_tokens_seen": 2694840320, "step": 1285 }, { "epoch": 0.023325196636832114, "grad_norm": 9.78898811340332, "learning_rate": 7.977784908145809e-07, "loss": 0.3044, "num_input_tokens_seen": 2696937472, "step": 1286 }, { "epoch": 0.023596419853539462, "grad_norm": 9.197432518005371, "learning_rate": 7.974414095540858e-07, "loss": 0.1947, "num_input_tokens_seen": 2699034624, "step": 1287 }, { "epoch": 0.023867643070246813, "grad_norm": 9.063929557800293, "learning_rate": 7.971041291621601e-07, "loss": 0.3043, "num_input_tokens_seen": 2701131776, "step": 1288 }, { "epoch": 0.024138866286954164, "grad_norm": 15.241850852966309, "learning_rate": 7.967666499102341e-07, "loss": 0.4811, "num_input_tokens_seen": 2703228928, "step": 1289 }, { "epoch": 0.024410089503661515, "grad_norm": 12.951624870300293, "learning_rate": 7.964289720698986e-07, "loss": 0.375, "num_input_tokens_seen": 2705326080, "step": 1290 }, { "epoch": 0.024681312720368862, "grad_norm": 15.193502426147461, "learning_rate": 7.960910959129037e-07, "loss": 0.6131, "num_input_tokens_seen": 2707423232, "step": 1291 }, { "epoch": 0.024952535937076213, "grad_norm": 16.782381057739258, "learning_rate": 7.957530217111591e-07, "loss": 0.5336, "num_input_tokens_seen": 2709520384, "step": 1292 }, { "epoch": 0.025223759153783564, "grad_norm": 19.65116310119629, "learning_rate": 7.954147497367343e-07, "loss": 0.8438, "num_input_tokens_seen": 2711617536, "step": 1293 }, { "epoch": 0.025494982370490915, "grad_norm": 13.590276718139648, "learning_rate": 7.950762802618576e-07, "loss": 0.4738, "num_input_tokens_seen": 2713714688, "step": 1294 }, { "epoch": 0.025766205587198263, "grad_norm": 15.773996353149414, "learning_rate": 7.947376135589164e-07, "loss": 0.595, "num_input_tokens_seen": 2715811840, "step": 1295 }, { "epoch": 0.026037428803905614, "grad_norm": 23.006914138793945, "learning_rate": 7.943987499004569e-07, "loss": 0.7875, "num_input_tokens_seen": 2717908992, "step": 1296 }, { "epoch": 0.026308652020612965, "grad_norm": 18.493175506591797, "learning_rate": 7.940596895591835e-07, "loss": 0.6582, "num_input_tokens_seen": 2720006144, "step": 1297 }, { "epoch": 0.026579875237320316, "grad_norm": 17.077844619750977, "learning_rate": 7.93720432807959e-07, "loss": 0.3572, "num_input_tokens_seen": 2722103296, "step": 1298 }, { "epoch": 0.026851098454027666, "grad_norm": 17.862571716308594, "learning_rate": 7.933809799198045e-07, "loss": 0.6255, "num_input_tokens_seen": 2724200448, "step": 1299 }, { "epoch": 0.027122321670735014, "grad_norm": 12.72701644897461, "learning_rate": 7.930413311678987e-07, "loss": 0.4388, "num_input_tokens_seen": 2726297600, "step": 1300 }, { "epoch": 0.027393544887442365, "grad_norm": 27.155799865722656, "learning_rate": 7.92701486825578e-07, "loss": 0.3302, "num_input_tokens_seen": 2728394752, "step": 1301 }, { "epoch": 0.027664768104149716, "grad_norm": 17.29124641418457, "learning_rate": 7.923614471663361e-07, "loss": 0.5452, "num_input_tokens_seen": 2730491904, "step": 1302 }, { "epoch": 0.027935991320857067, "grad_norm": 10.107339859008789, "learning_rate": 7.920212124638241e-07, "loss": 0.2633, "num_input_tokens_seen": 2732589056, "step": 1303 }, { "epoch": 0.028207214537564414, "grad_norm": 12.023981094360352, "learning_rate": 7.916807829918499e-07, "loss": 0.3122, "num_input_tokens_seen": 2734686208, "step": 1304 }, { "epoch": 0.028478437754271765, "grad_norm": 22.489961624145508, "learning_rate": 7.913401590243781e-07, "loss": 0.3229, "num_input_tokens_seen": 2736783360, "step": 1305 }, { "epoch": 0.028749660970979116, "grad_norm": 28.63505744934082, "learning_rate": 7.909993408355302e-07, "loss": 0.5431, "num_input_tokens_seen": 2738880512, "step": 1306 }, { "epoch": 0.029020884187686467, "grad_norm": 17.687850952148438, "learning_rate": 7.906583286995834e-07, "loss": 0.6771, "num_input_tokens_seen": 2740977664, "step": 1307 }, { "epoch": 0.029292107404393815, "grad_norm": 10.240657806396484, "learning_rate": 7.903171228909714e-07, "loss": 0.276, "num_input_tokens_seen": 2743074816, "step": 1308 }, { "epoch": 0.029563330621101166, "grad_norm": 12.936287879943848, "learning_rate": 7.899757236842836e-07, "loss": 0.3849, "num_input_tokens_seen": 2745171968, "step": 1309 }, { "epoch": 0.029834553837808517, "grad_norm": 19.912282943725586, "learning_rate": 7.89634131354265e-07, "loss": 0.7212, "num_input_tokens_seen": 2747269120, "step": 1310 }, { "epoch": 0.030105777054515868, "grad_norm": 16.74636459350586, "learning_rate": 7.892923461758165e-07, "loss": 0.5772, "num_input_tokens_seen": 2749366272, "step": 1311 }, { "epoch": 0.030377000271223215, "grad_norm": 13.208150863647461, "learning_rate": 7.889503684239933e-07, "loss": 0.4566, "num_input_tokens_seen": 2751463424, "step": 1312 }, { "epoch": 0.030648223487930566, "grad_norm": 9.817733764648438, "learning_rate": 7.886081983740066e-07, "loss": 0.2727, "num_input_tokens_seen": 2753560576, "step": 1313 }, { "epoch": 0.030919446704637917, "grad_norm": 15.745170593261719, "learning_rate": 7.882658363012214e-07, "loss": 0.5364, "num_input_tokens_seen": 2755657728, "step": 1314 }, { "epoch": 0.031190669921345268, "grad_norm": 19.3269100189209, "learning_rate": 7.879232824811579e-07, "loss": 0.4603, "num_input_tokens_seen": 2757754880, "step": 1315 }, { "epoch": 0.031461893138052616, "grad_norm": 21.108278274536133, "learning_rate": 7.875805371894904e-07, "loss": 0.8861, "num_input_tokens_seen": 2759852032, "step": 1316 }, { "epoch": 0.03173311635475997, "grad_norm": 21.515796661376953, "learning_rate": 7.872376007020469e-07, "loss": 0.5236, "num_input_tokens_seen": 2761949184, "step": 1317 }, { "epoch": 0.03200433957146732, "grad_norm": 11.72025203704834, "learning_rate": 7.8689447329481e-07, "loss": 0.3716, "num_input_tokens_seen": 2764046336, "step": 1318 }, { "epoch": 0.032275562788174665, "grad_norm": 19.262054443359375, "learning_rate": 7.865511552439156e-07, "loss": 0.5643, "num_input_tokens_seen": 2766143488, "step": 1319 }, { "epoch": 0.03254678600488202, "grad_norm": 13.74807357788086, "learning_rate": 7.862076468256529e-07, "loss": 0.5372, "num_input_tokens_seen": 2768240640, "step": 1320 }, { "epoch": 0.03281800922158937, "grad_norm": 15.6185941696167, "learning_rate": 7.858639483164643e-07, "loss": 0.4638, "num_input_tokens_seen": 2770337792, "step": 1321 }, { "epoch": 0.03308923243829672, "grad_norm": 14.039484024047852, "learning_rate": 7.855200599929456e-07, "loss": 0.5236, "num_input_tokens_seen": 2772434944, "step": 1322 }, { "epoch": 0.03336045565500407, "grad_norm": 14.383334159851074, "learning_rate": 7.851759821318447e-07, "loss": 0.4256, "num_input_tokens_seen": 2774532096, "step": 1323 }, { "epoch": 0.033631678871711417, "grad_norm": 18.412899017333984, "learning_rate": 7.848317150100624e-07, "loss": 0.5748, "num_input_tokens_seen": 2776629248, "step": 1324 }, { "epoch": 0.03390290208841877, "grad_norm": 12.141471862792969, "learning_rate": 7.84487258904652e-07, "loss": 0.4111, "num_input_tokens_seen": 2778726400, "step": 1325 }, { "epoch": 0.03417412530512612, "grad_norm": 22.77025032043457, "learning_rate": 7.841426140928184e-07, "loss": 0.8071, "num_input_tokens_seen": 2780823552, "step": 1326 }, { "epoch": 0.034445348521833466, "grad_norm": 27.863340377807617, "learning_rate": 7.837977808519189e-07, "loss": 0.7943, "num_input_tokens_seen": 2782920704, "step": 1327 }, { "epoch": 0.03471657173854082, "grad_norm": 19.592632293701172, "learning_rate": 7.834527594594618e-07, "loss": 0.6998, "num_input_tokens_seen": 2785017856, "step": 1328 }, { "epoch": 0.03498779495524817, "grad_norm": 19.130634307861328, "learning_rate": 7.831075501931078e-07, "loss": 0.5628, "num_input_tokens_seen": 2787115008, "step": 1329 }, { "epoch": 0.03525901817195552, "grad_norm": 16.67599105834961, "learning_rate": 7.827621533306677e-07, "loss": 0.6865, "num_input_tokens_seen": 2789212160, "step": 1330 }, { "epoch": 0.03553024138866287, "grad_norm": 13.267627716064453, "learning_rate": 7.824165691501036e-07, "loss": 0.4898, "num_input_tokens_seen": 2791309312, "step": 1331 }, { "epoch": 0.03580146460537022, "grad_norm": 17.262802124023438, "learning_rate": 7.820707979295289e-07, "loss": 0.3767, "num_input_tokens_seen": 2793406464, "step": 1332 }, { "epoch": 0.03607268782207757, "grad_norm": 21.728710174560547, "learning_rate": 7.81724839947207e-07, "loss": 0.7031, "num_input_tokens_seen": 2795503616, "step": 1333 }, { "epoch": 0.03634391103878492, "grad_norm": 20.991073608398438, "learning_rate": 7.813786954815517e-07, "loss": 0.4953, "num_input_tokens_seen": 2797600768, "step": 1334 }, { "epoch": 0.03661513425549227, "grad_norm": 14.685553550720215, "learning_rate": 7.810323648111269e-07, "loss": 0.3939, "num_input_tokens_seen": 2799697920, "step": 1335 }, { "epoch": 0.03688635747219962, "grad_norm": 14.160144805908203, "learning_rate": 7.806858482146463e-07, "loss": 0.4822, "num_input_tokens_seen": 2801795072, "step": 1336 }, { "epoch": 0.03715758068890697, "grad_norm": 16.465970993041992, "learning_rate": 7.803391459709733e-07, "loss": 0.5157, "num_input_tokens_seen": 2803892224, "step": 1337 }, { "epoch": 0.03742880390561432, "grad_norm": 12.798428535461426, "learning_rate": 7.799922583591206e-07, "loss": 0.4207, "num_input_tokens_seen": 2805989376, "step": 1338 }, { "epoch": 0.03770002712232167, "grad_norm": 12.751097679138184, "learning_rate": 7.796451856582504e-07, "loss": 0.4589, "num_input_tokens_seen": 2808086528, "step": 1339 }, { "epoch": 0.03797125033902902, "grad_norm": 15.334579467773438, "learning_rate": 7.792979281476734e-07, "loss": 0.3559, "num_input_tokens_seen": 2810183680, "step": 1340 }, { "epoch": 0.03824247355573637, "grad_norm": 13.238780975341797, "learning_rate": 7.789504861068492e-07, "loss": 0.4655, "num_input_tokens_seen": 2812280832, "step": 1341 }, { "epoch": 0.03851369677244372, "grad_norm": 15.215901374816895, "learning_rate": 7.78602859815386e-07, "loss": 0.5126, "num_input_tokens_seen": 2814377984, "step": 1342 }, { "epoch": 0.038784919989151075, "grad_norm": 13.011480331420898, "learning_rate": 7.782550495530402e-07, "loss": 0.327, "num_input_tokens_seen": 2816475136, "step": 1343 }, { "epoch": 0.03905614320585842, "grad_norm": 18.614702224731445, "learning_rate": 7.779070555997162e-07, "loss": 0.7977, "num_input_tokens_seen": 2818572288, "step": 1344 }, { "epoch": 0.03932736642256577, "grad_norm": 18.841096878051758, "learning_rate": 7.775588782354666e-07, "loss": 0.7646, "num_input_tokens_seen": 2820669440, "step": 1345 }, { "epoch": 0.039598589639273124, "grad_norm": 20.057369232177734, "learning_rate": 7.77210517740491e-07, "loss": 0.7922, "num_input_tokens_seen": 2822766592, "step": 1346 }, { "epoch": 0.03986981285598047, "grad_norm": 10.478140830993652, "learning_rate": 7.768619743951366e-07, "loss": 0.2467, "num_input_tokens_seen": 2824863744, "step": 1347 }, { "epoch": 0.04014103607268782, "grad_norm": 10.633049964904785, "learning_rate": 7.765132484798978e-07, "loss": 0.3628, "num_input_tokens_seen": 2826960896, "step": 1348 }, { "epoch": 0.040412259289395173, "grad_norm": 18.010910034179688, "learning_rate": 7.761643402754163e-07, "loss": 0.6482, "num_input_tokens_seen": 2829058048, "step": 1349 }, { "epoch": 0.04068348250610252, "grad_norm": 20.25141716003418, "learning_rate": 7.7581525006248e-07, "loss": 0.5458, "num_input_tokens_seen": 2831155200, "step": 1350 }, { "epoch": 0.040954705722809875, "grad_norm": 16.61979866027832, "learning_rate": 7.754659781220232e-07, "loss": 0.5326, "num_input_tokens_seen": 2833252352, "step": 1351 }, { "epoch": 0.04122592893951722, "grad_norm": 11.642721176147461, "learning_rate": 7.751165247351269e-07, "loss": 0.3478, "num_input_tokens_seen": 2835349504, "step": 1352 }, { "epoch": 0.04149715215622457, "grad_norm": 15.440996170043945, "learning_rate": 7.747668901830178e-07, "loss": 0.559, "num_input_tokens_seen": 2837446656, "step": 1353 }, { "epoch": 0.041768375372931925, "grad_norm": 15.072626113891602, "learning_rate": 7.744170747470685e-07, "loss": 0.5729, "num_input_tokens_seen": 2839543808, "step": 1354 }, { "epoch": 0.04203959858963927, "grad_norm": 19.414575576782227, "learning_rate": 7.740670787087972e-07, "loss": 0.9754, "num_input_tokens_seen": 2841640960, "step": 1355 }, { "epoch": 0.04231082180634662, "grad_norm": 14.357555389404297, "learning_rate": 7.737169023498672e-07, "loss": 0.3421, "num_input_tokens_seen": 2843738112, "step": 1356 }, { "epoch": 0.042582045023053974, "grad_norm": 13.045211791992188, "learning_rate": 7.733665459520873e-07, "loss": 0.3127, "num_input_tokens_seen": 2845835264, "step": 1357 }, { "epoch": 0.04285326823976132, "grad_norm": 11.419096946716309, "learning_rate": 7.730160097974109e-07, "loss": 0.3253, "num_input_tokens_seen": 2847932416, "step": 1358 }, { "epoch": 0.043124491456468676, "grad_norm": 14.921306610107422, "learning_rate": 7.726652941679365e-07, "loss": 0.46, "num_input_tokens_seen": 2850029568, "step": 1359 }, { "epoch": 0.043395714673176024, "grad_norm": 12.84131908416748, "learning_rate": 7.723143993459061e-07, "loss": 0.4562, "num_input_tokens_seen": 2852126720, "step": 1360 }, { "epoch": 0.04366693788988337, "grad_norm": 15.976582527160645, "learning_rate": 7.719633256137067e-07, "loss": 0.5137, "num_input_tokens_seen": 2854223872, "step": 1361 }, { "epoch": 0.043938161106590726, "grad_norm": 10.385905265808105, "learning_rate": 7.716120732538696e-07, "loss": 0.2903, "num_input_tokens_seen": 2856321024, "step": 1362 }, { "epoch": 0.04420938432329807, "grad_norm": 17.1414737701416, "learning_rate": 7.712606425490687e-07, "loss": 0.7363, "num_input_tokens_seen": 2858418176, "step": 1363 }, { "epoch": 0.04448060754000543, "grad_norm": 20.92815399169922, "learning_rate": 7.70909033782122e-07, "loss": 0.5892, "num_input_tokens_seen": 2860515328, "step": 1364 }, { "epoch": 0.044751830756712775, "grad_norm": 15.580632209777832, "learning_rate": 7.705572472359913e-07, "loss": 0.5212, "num_input_tokens_seen": 2862612480, "step": 1365 }, { "epoch": 0.04502305397342012, "grad_norm": 13.09688949584961, "learning_rate": 7.702052831937811e-07, "loss": 0.5206, "num_input_tokens_seen": 2864709632, "step": 1366 }, { "epoch": 0.04529427719012748, "grad_norm": 17.451766967773438, "learning_rate": 7.698531419387382e-07, "loss": 0.6266, "num_input_tokens_seen": 2866806784, "step": 1367 }, { "epoch": 0.045565500406834825, "grad_norm": 16.42080307006836, "learning_rate": 7.695008237542526e-07, "loss": 0.3435, "num_input_tokens_seen": 2868903936, "step": 1368 }, { "epoch": 0.04583672362354217, "grad_norm": 21.22711753845215, "learning_rate": 7.691483289238569e-07, "loss": 0.9744, "num_input_tokens_seen": 2871001088, "step": 1369 }, { "epoch": 0.04610794684024953, "grad_norm": 14.438827514648438, "learning_rate": 7.687956577312251e-07, "loss": 0.5828, "num_input_tokens_seen": 2873098240, "step": 1370 }, { "epoch": 0.046379170056956874, "grad_norm": 13.411489486694336, "learning_rate": 7.684428104601739e-07, "loss": 0.3964, "num_input_tokens_seen": 2875195392, "step": 1371 }, { "epoch": 0.04665039327366423, "grad_norm": 11.179100036621094, "learning_rate": 7.680897873946611e-07, "loss": 0.3965, "num_input_tokens_seen": 2877292544, "step": 1372 }, { "epoch": 0.046921616490371576, "grad_norm": 13.811481475830078, "learning_rate": 7.677365888187864e-07, "loss": 0.472, "num_input_tokens_seen": 2879389696, "step": 1373 }, { "epoch": 0.047192839707078924, "grad_norm": 13.83006763458252, "learning_rate": 7.673832150167906e-07, "loss": 0.4331, "num_input_tokens_seen": 2881486848, "step": 1374 }, { "epoch": 0.04746406292378628, "grad_norm": 11.234846115112305, "learning_rate": 7.670296662730552e-07, "loss": 0.3489, "num_input_tokens_seen": 2883584000, "step": 1375 }, { "epoch": 0.047735286140493625, "grad_norm": 15.245635986328125, "learning_rate": 7.66675942872103e-07, "loss": 0.5866, "num_input_tokens_seen": 2885681152, "step": 1376 }, { "epoch": 0.04800650935720097, "grad_norm": 16.57020378112793, "learning_rate": 7.663220450985973e-07, "loss": 0.4715, "num_input_tokens_seen": 2887778304, "step": 1377 }, { "epoch": 0.04827773257390833, "grad_norm": 18.864973068237305, "learning_rate": 7.659679732373413e-07, "loss": 0.6713, "num_input_tokens_seen": 2889875456, "step": 1378 }, { "epoch": 0.048548955790615675, "grad_norm": 10.371908187866211, "learning_rate": 7.656137275732786e-07, "loss": 0.3513, "num_input_tokens_seen": 2891972608, "step": 1379 }, { "epoch": 0.04882017900732303, "grad_norm": 18.576017379760742, "learning_rate": 7.652593083914927e-07, "loss": 0.628, "num_input_tokens_seen": 2894069760, "step": 1380 }, { "epoch": 0.04909140222403038, "grad_norm": 17.365699768066406, "learning_rate": 7.649047159772064e-07, "loss": 0.6301, "num_input_tokens_seen": 2896166912, "step": 1381 }, { "epoch": 0.049362625440737724, "grad_norm": 13.766050338745117, "learning_rate": 7.645499506157827e-07, "loss": 0.432, "num_input_tokens_seen": 2898264064, "step": 1382 }, { "epoch": 0.04963384865744508, "grad_norm": 15.873371124267578, "learning_rate": 7.641950125927228e-07, "loss": 0.613, "num_input_tokens_seen": 2900361216, "step": 1383 }, { "epoch": 0.049905071874152426, "grad_norm": 10.771862983703613, "learning_rate": 7.638399021936675e-07, "loss": 0.3301, "num_input_tokens_seen": 2902458368, "step": 1384 }, { "epoch": 0.05017629509085978, "grad_norm": 26.106536865234375, "learning_rate": 7.634846197043963e-07, "loss": 0.4127, "num_input_tokens_seen": 2904555520, "step": 1385 }, { "epoch": 0.05044751830756713, "grad_norm": 16.403791427612305, "learning_rate": 7.631291654108266e-07, "loss": 0.6469, "num_input_tokens_seen": 2906652672, "step": 1386 }, { "epoch": 0.050718741524274476, "grad_norm": 25.416994094848633, "learning_rate": 7.627735395990149e-07, "loss": 0.8975, "num_input_tokens_seen": 2908749824, "step": 1387 }, { "epoch": 0.05098996474098183, "grad_norm": 26.157794952392578, "learning_rate": 7.624177425551552e-07, "loss": 1.0145, "num_input_tokens_seen": 2910846976, "step": 1388 }, { "epoch": 0.05126118795768918, "grad_norm": 20.74803924560547, "learning_rate": 7.620617745655793e-07, "loss": 0.7406, "num_input_tokens_seen": 2912944128, "step": 1389 }, { "epoch": 0.051532411174396525, "grad_norm": 15.389482498168945, "learning_rate": 7.617056359167568e-07, "loss": 0.4884, "num_input_tokens_seen": 2915041280, "step": 1390 }, { "epoch": 0.05180363439110388, "grad_norm": 15.435873985290527, "learning_rate": 7.613493268952947e-07, "loss": 0.5215, "num_input_tokens_seen": 2917138432, "step": 1391 }, { "epoch": 0.05207485760781123, "grad_norm": 12.54256534576416, "learning_rate": 7.609928477879365e-07, "loss": 0.3087, "num_input_tokens_seen": 2919235584, "step": 1392 }, { "epoch": 0.05234608082451858, "grad_norm": 15.502516746520996, "learning_rate": 7.606361988815633e-07, "loss": 0.5513, "num_input_tokens_seen": 2921332736, "step": 1393 }, { "epoch": 0.05261730404122593, "grad_norm": 13.708436965942383, "learning_rate": 7.602793804631927e-07, "loss": 0.273, "num_input_tokens_seen": 2923429888, "step": 1394 }, { "epoch": 0.05288852725793328, "grad_norm": 15.411025047302246, "learning_rate": 7.599223928199781e-07, "loss": 0.4805, "num_input_tokens_seen": 2925527040, "step": 1395 }, { "epoch": 0.05315975047464063, "grad_norm": 16.42882537841797, "learning_rate": 7.595652362392103e-07, "loss": 0.6015, "num_input_tokens_seen": 2927624192, "step": 1396 }, { "epoch": 0.05343097369134798, "grad_norm": 17.059511184692383, "learning_rate": 7.592079110083146e-07, "loss": 0.711, "num_input_tokens_seen": 2929721344, "step": 1397 }, { "epoch": 0.05370219690805533, "grad_norm": 15.640413284301758, "learning_rate": 7.588504174148532e-07, "loss": 0.3908, "num_input_tokens_seen": 2931818496, "step": 1398 }, { "epoch": 0.05397342012476268, "grad_norm": 18.543855667114258, "learning_rate": 7.584927557465233e-07, "loss": 0.7099, "num_input_tokens_seen": 2933915648, "step": 1399 }, { "epoch": 0.05424464334147003, "grad_norm": 11.331782341003418, "learning_rate": 7.581349262911573e-07, "loss": 0.3008, "num_input_tokens_seen": 2936012800, "step": 1400 }, { "epoch": 0.05451586655817738, "grad_norm": 13.517926216125488, "learning_rate": 7.577769293367226e-07, "loss": 0.451, "num_input_tokens_seen": 2938109952, "step": 1401 }, { "epoch": 0.05478708977488473, "grad_norm": 15.007146835327148, "learning_rate": 7.574187651713218e-07, "loss": 0.4981, "num_input_tokens_seen": 2940207104, "step": 1402 }, { "epoch": 0.05505831299159208, "grad_norm": 13.51999282836914, "learning_rate": 7.570604340831916e-07, "loss": 0.4342, "num_input_tokens_seen": 2942304256, "step": 1403 }, { "epoch": 0.05532953620829943, "grad_norm": 9.16441822052002, "learning_rate": 7.567019363607032e-07, "loss": 0.2665, "num_input_tokens_seen": 2944401408, "step": 1404 }, { "epoch": 0.05560075942500678, "grad_norm": 13.39521598815918, "learning_rate": 7.563432722923621e-07, "loss": 0.4227, "num_input_tokens_seen": 2946498560, "step": 1405 }, { "epoch": 0.055871982641714134, "grad_norm": 12.353023529052734, "learning_rate": 7.559844421668074e-07, "loss": 0.3066, "num_input_tokens_seen": 2948595712, "step": 1406 }, { "epoch": 0.05614320585842148, "grad_norm": 14.080292701721191, "learning_rate": 7.556254462728122e-07, "loss": 0.5427, "num_input_tokens_seen": 2950692864, "step": 1407 }, { "epoch": 0.05641442907512883, "grad_norm": 14.238213539123535, "learning_rate": 7.552662848992822e-07, "loss": 0.4679, "num_input_tokens_seen": 2952790016, "step": 1408 }, { "epoch": 0.05668565229183618, "grad_norm": 13.333026885986328, "learning_rate": 7.54906958335257e-07, "loss": 0.4319, "num_input_tokens_seen": 2954887168, "step": 1409 }, { "epoch": 0.05695687550854353, "grad_norm": 10.981861114501953, "learning_rate": 7.545474668699091e-07, "loss": 0.3214, "num_input_tokens_seen": 2956984320, "step": 1410 }, { "epoch": 0.05722809872525088, "grad_norm": 10.62180233001709, "learning_rate": 7.541878107925435e-07, "loss": 0.3711, "num_input_tokens_seen": 2959081472, "step": 1411 }, { "epoch": 0.05749932194195823, "grad_norm": 13.643486976623535, "learning_rate": 7.538279903925977e-07, "loss": 0.2983, "num_input_tokens_seen": 2961178624, "step": 1412 }, { "epoch": 0.05777054515866558, "grad_norm": 15.046182632446289, "learning_rate": 7.534680059596414e-07, "loss": 0.5167, "num_input_tokens_seen": 2963275776, "step": 1413 }, { "epoch": 0.058041768375372935, "grad_norm": 22.363344192504883, "learning_rate": 7.531078577833765e-07, "loss": 0.6654, "num_input_tokens_seen": 2965372928, "step": 1414 }, { "epoch": 0.05831299159208028, "grad_norm": 12.185877799987793, "learning_rate": 7.527475461536363e-07, "loss": 0.3195, "num_input_tokens_seen": 2967470080, "step": 1415 }, { "epoch": 0.05858421480878763, "grad_norm": 18.2763614654541, "learning_rate": 7.523870713603864e-07, "loss": 0.3883, "num_input_tokens_seen": 2969567232, "step": 1416 }, { "epoch": 0.058855438025494984, "grad_norm": 12.003350257873535, "learning_rate": 7.520264336937227e-07, "loss": 0.4257, "num_input_tokens_seen": 2971664384, "step": 1417 }, { "epoch": 0.05912666124220233, "grad_norm": 10.348063468933105, "learning_rate": 7.516656334438727e-07, "loss": 0.3473, "num_input_tokens_seen": 2973761536, "step": 1418 }, { "epoch": 0.059397884458909686, "grad_norm": 18.062877655029297, "learning_rate": 7.513046709011951e-07, "loss": 0.5485, "num_input_tokens_seen": 2975858688, "step": 1419 }, { "epoch": 0.059669107675617034, "grad_norm": 11.781852722167969, "learning_rate": 7.509435463561785e-07, "loss": 0.3139, "num_input_tokens_seen": 2977955840, "step": 1420 }, { "epoch": 0.05994033089232438, "grad_norm": 17.999149322509766, "learning_rate": 7.505822600994423e-07, "loss": 0.8586, "num_input_tokens_seen": 2980052992, "step": 1421 }, { "epoch": 0.060211554109031735, "grad_norm": 19.359094619750977, "learning_rate": 7.502208124217357e-07, "loss": 0.5113, "num_input_tokens_seen": 2982150144, "step": 1422 }, { "epoch": 0.06048277732573908, "grad_norm": 18.1299991607666, "learning_rate": 7.498592036139383e-07, "loss": 0.6946, "num_input_tokens_seen": 2984247296, "step": 1423 }, { "epoch": 0.06075400054244643, "grad_norm": 12.68889045715332, "learning_rate": 7.494974339670591e-07, "loss": 0.3701, "num_input_tokens_seen": 2986344448, "step": 1424 }, { "epoch": 0.061025223759153785, "grad_norm": 12.457084655761719, "learning_rate": 7.491355037722365e-07, "loss": 0.3567, "num_input_tokens_seen": 2988441600, "step": 1425 }, { "epoch": 0.06129644697586113, "grad_norm": 21.13746452331543, "learning_rate": 7.487734133207382e-07, "loss": 0.6114, "num_input_tokens_seen": 2990538752, "step": 1426 }, { "epoch": 0.06156767019256849, "grad_norm": 10.113631248474121, "learning_rate": 7.484111629039607e-07, "loss": 0.2361, "num_input_tokens_seen": 2992635904, "step": 1427 }, { "epoch": 0.061838893409275834, "grad_norm": 18.88080406188965, "learning_rate": 7.480487528134292e-07, "loss": 0.7382, "num_input_tokens_seen": 2994733056, "step": 1428 }, { "epoch": 0.06211011662598318, "grad_norm": 13.798151969909668, "learning_rate": 7.47686183340798e-07, "loss": 0.3842, "num_input_tokens_seen": 2996830208, "step": 1429 }, { "epoch": 0.062381339842690536, "grad_norm": 17.79098892211914, "learning_rate": 7.473234547778489e-07, "loss": 0.6524, "num_input_tokens_seen": 2998927360, "step": 1430 }, { "epoch": 0.06265256305939788, "grad_norm": 14.120518684387207, "learning_rate": 7.46960567416492e-07, "loss": 0.476, "num_input_tokens_seen": 3001024512, "step": 1431 }, { "epoch": 0.06292378627610523, "grad_norm": 12.636813163757324, "learning_rate": 7.465975215487655e-07, "loss": 0.3001, "num_input_tokens_seen": 3003121664, "step": 1432 }, { "epoch": 0.06319500949281258, "grad_norm": 18.47852897644043, "learning_rate": 7.462343174668346e-07, "loss": 0.6856, "num_input_tokens_seen": 3005218816, "step": 1433 }, { "epoch": 0.06346623270951994, "grad_norm": 15.712651252746582, "learning_rate": 7.458709554629924e-07, "loss": 0.569, "num_input_tokens_seen": 3007315968, "step": 1434 }, { "epoch": 0.06373745592622729, "grad_norm": 20.598899841308594, "learning_rate": 7.455074358296586e-07, "loss": 0.6715, "num_input_tokens_seen": 3009413120, "step": 1435 }, { "epoch": 0.06400867914293464, "grad_norm": 11.868573188781738, "learning_rate": 7.451437588593802e-07, "loss": 0.4039, "num_input_tokens_seen": 3011510272, "step": 1436 }, { "epoch": 0.06427990235964198, "grad_norm": 16.7470703125, "learning_rate": 7.447799248448303e-07, "loss": 0.5598, "num_input_tokens_seen": 3013607424, "step": 1437 }, { "epoch": 0.06455112557634933, "grad_norm": 16.11475944519043, "learning_rate": 7.444159340788088e-07, "loss": 0.5864, "num_input_tokens_seen": 3015704576, "step": 1438 }, { "epoch": 0.06482234879305669, "grad_norm": 12.091106414794922, "learning_rate": 7.440517868542417e-07, "loss": 0.394, "num_input_tokens_seen": 3017801728, "step": 1439 }, { "epoch": 0.06509357200976404, "grad_norm": 13.48214340209961, "learning_rate": 7.436874834641807e-07, "loss": 0.365, "num_input_tokens_seen": 3019898880, "step": 1440 }, { "epoch": 0.06536479522647139, "grad_norm": 18.32723045349121, "learning_rate": 7.433230242018035e-07, "loss": 0.6249, "num_input_tokens_seen": 3021996032, "step": 1441 }, { "epoch": 0.06563601844317873, "grad_norm": 13.360453605651855, "learning_rate": 7.429584093604128e-07, "loss": 0.4254, "num_input_tokens_seen": 3024093184, "step": 1442 }, { "epoch": 0.06590724165988608, "grad_norm": 16.321001052856445, "learning_rate": 7.425936392334368e-07, "loss": 0.6535, "num_input_tokens_seen": 3026190336, "step": 1443 }, { "epoch": 0.06617846487659344, "grad_norm": 15.212583541870117, "learning_rate": 7.422287141144287e-07, "loss": 0.6856, "num_input_tokens_seen": 3028287488, "step": 1444 }, { "epoch": 0.06644968809330079, "grad_norm": 20.179534912109375, "learning_rate": 7.418636342970665e-07, "loss": 0.8147, "num_input_tokens_seen": 3030384640, "step": 1445 }, { "epoch": 0.06672091131000814, "grad_norm": 11.90196418762207, "learning_rate": 7.414984000751521e-07, "loss": 0.3257, "num_input_tokens_seen": 3032481792, "step": 1446 }, { "epoch": 0.06699213452671549, "grad_norm": 12.35843563079834, "learning_rate": 7.411330117426125e-07, "loss": 0.3769, "num_input_tokens_seen": 3034578944, "step": 1447 }, { "epoch": 0.06726335774342283, "grad_norm": 9.548952102661133, "learning_rate": 7.407674695934983e-07, "loss": 0.2638, "num_input_tokens_seen": 3036676096, "step": 1448 }, { "epoch": 0.06753458096013018, "grad_norm": 14.05826187133789, "learning_rate": 7.404017739219836e-07, "loss": 0.4847, "num_input_tokens_seen": 3038773248, "step": 1449 }, { "epoch": 0.06780580417683754, "grad_norm": 13.904762268066406, "learning_rate": 7.400359250223667e-07, "loss": 0.3108, "num_input_tokens_seen": 3040870400, "step": 1450 }, { "epoch": 0.06807702739354489, "grad_norm": 15.42446517944336, "learning_rate": 7.396699231890689e-07, "loss": 0.59, "num_input_tokens_seen": 3042967552, "step": 1451 }, { "epoch": 0.06834825061025224, "grad_norm": 16.7812442779541, "learning_rate": 7.393037687166342e-07, "loss": 0.591, "num_input_tokens_seen": 3045064704, "step": 1452 }, { "epoch": 0.06861947382695958, "grad_norm": 14.538595199584961, "learning_rate": 7.3893746189973e-07, "loss": 0.4907, "num_input_tokens_seen": 3047161856, "step": 1453 }, { "epoch": 0.06889069704366693, "grad_norm": 15.920239448547363, "learning_rate": 7.385710030331461e-07, "loss": 0.4527, "num_input_tokens_seen": 3049259008, "step": 1454 }, { "epoch": 0.0691619202603743, "grad_norm": 11.364521026611328, "learning_rate": 7.382043924117945e-07, "loss": 0.4068, "num_input_tokens_seen": 3051356160, "step": 1455 }, { "epoch": 0.06943314347708164, "grad_norm": 15.396533012390137, "learning_rate": 7.378376303307099e-07, "loss": 0.285, "num_input_tokens_seen": 3053453312, "step": 1456 }, { "epoch": 0.06970436669378899, "grad_norm": 14.151115417480469, "learning_rate": 7.374707170850479e-07, "loss": 0.4752, "num_input_tokens_seen": 3055550464, "step": 1457 }, { "epoch": 0.06997558991049634, "grad_norm": 10.992759704589844, "learning_rate": 7.371036529700866e-07, "loss": 0.4462, "num_input_tokens_seen": 3057647616, "step": 1458 }, { "epoch": 0.07024681312720368, "grad_norm": 12.24362564086914, "learning_rate": 7.367364382812253e-07, "loss": 0.3646, "num_input_tokens_seen": 3059744768, "step": 1459 }, { "epoch": 0.07051803634391104, "grad_norm": 16.037370681762695, "learning_rate": 7.363690733139842e-07, "loss": 0.6663, "num_input_tokens_seen": 3061841920, "step": 1460 }, { "epoch": 0.07078925956061839, "grad_norm": 18.999717712402344, "learning_rate": 7.360015583640049e-07, "loss": 0.7606, "num_input_tokens_seen": 3063939072, "step": 1461 }, { "epoch": 0.07106048277732574, "grad_norm": 10.200159072875977, "learning_rate": 7.356338937270492e-07, "loss": 0.3097, "num_input_tokens_seen": 3066036224, "step": 1462 }, { "epoch": 0.07133170599403309, "grad_norm": 20.75838279724121, "learning_rate": 7.352660796989999e-07, "loss": 0.3529, "num_input_tokens_seen": 3068133376, "step": 1463 }, { "epoch": 0.07160292921074043, "grad_norm": 15.318017959594727, "learning_rate": 7.348981165758595e-07, "loss": 0.4486, "num_input_tokens_seen": 3070230528, "step": 1464 }, { "epoch": 0.0718741524274478, "grad_norm": 11.934765815734863, "learning_rate": 7.345300046537507e-07, "loss": 0.411, "num_input_tokens_seen": 3072327680, "step": 1465 }, { "epoch": 0.07214537564415514, "grad_norm": 16.084491729736328, "learning_rate": 7.341617442289159e-07, "loss": 0.5918, "num_input_tokens_seen": 3074424832, "step": 1466 }, { "epoch": 0.07241659886086249, "grad_norm": 15.078495979309082, "learning_rate": 7.337933355977175e-07, "loss": 0.495, "num_input_tokens_seen": 3076521984, "step": 1467 }, { "epoch": 0.07268782207756984, "grad_norm": 12.67852783203125, "learning_rate": 7.334247790566364e-07, "loss": 0.3726, "num_input_tokens_seen": 3078619136, "step": 1468 }, { "epoch": 0.07295904529427719, "grad_norm": 12.588438034057617, "learning_rate": 7.330560749022728e-07, "loss": 0.364, "num_input_tokens_seen": 3080716288, "step": 1469 }, { "epoch": 0.07323026851098453, "grad_norm": 31.776931762695312, "learning_rate": 7.326872234313459e-07, "loss": 0.5082, "num_input_tokens_seen": 3082813440, "step": 1470 }, { "epoch": 0.0735014917276919, "grad_norm": 16.448945999145508, "learning_rate": 7.323182249406936e-07, "loss": 0.4378, "num_input_tokens_seen": 3084910592, "step": 1471 }, { "epoch": 0.07377271494439924, "grad_norm": 19.21479606628418, "learning_rate": 7.319490797272714e-07, "loss": 0.6422, "num_input_tokens_seen": 3087007744, "step": 1472 }, { "epoch": 0.07404393816110659, "grad_norm": 12.941019058227539, "learning_rate": 7.315797880881535e-07, "loss": 0.4457, "num_input_tokens_seen": 3089104896, "step": 1473 }, { "epoch": 0.07431516137781394, "grad_norm": 11.186482429504395, "learning_rate": 7.312103503205318e-07, "loss": 0.4441, "num_input_tokens_seen": 3091202048, "step": 1474 }, { "epoch": 0.07458638459452128, "grad_norm": 9.005460739135742, "learning_rate": 7.308407667217158e-07, "loss": 0.1681, "num_input_tokens_seen": 3093299200, "step": 1475 }, { "epoch": 0.07485760781122865, "grad_norm": 15.578516960144043, "learning_rate": 7.304710375891323e-07, "loss": 0.5316, "num_input_tokens_seen": 3095396352, "step": 1476 }, { "epoch": 0.075128831027936, "grad_norm": 22.51955795288086, "learning_rate": 7.30101163220325e-07, "loss": 0.798, "num_input_tokens_seen": 3097493504, "step": 1477 }, { "epoch": 0.07540005424464334, "grad_norm": 14.802974700927734, "learning_rate": 7.29731143912955e-07, "loss": 0.6361, "num_input_tokens_seen": 3099590656, "step": 1478 }, { "epoch": 0.07567127746135069, "grad_norm": 13.981009483337402, "learning_rate": 7.293609799647996e-07, "loss": 0.507, "num_input_tokens_seen": 3101687808, "step": 1479 }, { "epoch": 0.07594250067805804, "grad_norm": 17.046981811523438, "learning_rate": 7.289906716737528e-07, "loss": 0.5821, "num_input_tokens_seen": 3103784960, "step": 1480 }, { "epoch": 0.0762137238947654, "grad_norm": 11.836614608764648, "learning_rate": 7.286202193378244e-07, "loss": 0.3149, "num_input_tokens_seen": 3105882112, "step": 1481 }, { "epoch": 0.07648494711147275, "grad_norm": 14.480379104614258, "learning_rate": 7.282496232551406e-07, "loss": 0.5075, "num_input_tokens_seen": 3107979264, "step": 1482 }, { "epoch": 0.07675617032818009, "grad_norm": 17.524093627929688, "learning_rate": 7.278788837239429e-07, "loss": 0.5825, "num_input_tokens_seen": 3110076416, "step": 1483 }, { "epoch": 0.07702739354488744, "grad_norm": 10.272031784057617, "learning_rate": 7.275080010425883e-07, "loss": 0.3495, "num_input_tokens_seen": 3112173568, "step": 1484 }, { "epoch": 0.07729861676159479, "grad_norm": 14.482696533203125, "learning_rate": 7.271369755095494e-07, "loss": 0.5638, "num_input_tokens_seen": 3114270720, "step": 1485 }, { "epoch": 0.07756983997830215, "grad_norm": 17.09143829345703, "learning_rate": 7.267658074234128e-07, "loss": 0.7527, "num_input_tokens_seen": 3116367872, "step": 1486 }, { "epoch": 0.0778410631950095, "grad_norm": 15.939448356628418, "learning_rate": 7.26394497082881e-07, "loss": 0.5614, "num_input_tokens_seen": 3118465024, "step": 1487 }, { "epoch": 0.07811228641171684, "grad_norm": 13.810276985168457, "learning_rate": 7.260230447867703e-07, "loss": 0.4629, "num_input_tokens_seen": 3120562176, "step": 1488 }, { "epoch": 0.07838350962842419, "grad_norm": 14.860830307006836, "learning_rate": 7.256514508340114e-07, "loss": 0.5521, "num_input_tokens_seen": 3122659328, "step": 1489 }, { "epoch": 0.07865473284513154, "grad_norm": 17.182228088378906, "learning_rate": 7.252797155236488e-07, "loss": 0.624, "num_input_tokens_seen": 3124756480, "step": 1490 }, { "epoch": 0.07892595606183889, "grad_norm": 15.805337905883789, "learning_rate": 7.249078391548409e-07, "loss": 0.596, "num_input_tokens_seen": 3126853632, "step": 1491 }, { "epoch": 0.07919717927854625, "grad_norm": 9.75827407836914, "learning_rate": 7.245358220268599e-07, "loss": 0.2837, "num_input_tokens_seen": 3128950784, "step": 1492 }, { "epoch": 0.0794684024952536, "grad_norm": 22.41292953491211, "learning_rate": 7.24163664439091e-07, "loss": 0.6901, "num_input_tokens_seen": 3131047936, "step": 1493 }, { "epoch": 0.07973962571196094, "grad_norm": 9.785002708435059, "learning_rate": 7.237913666910322e-07, "loss": 0.3265, "num_input_tokens_seen": 3133145088, "step": 1494 }, { "epoch": 0.08001084892866829, "grad_norm": 14.571457862854004, "learning_rate": 7.234189290822947e-07, "loss": 0.6036, "num_input_tokens_seen": 3135242240, "step": 1495 }, { "epoch": 0.08028207214537564, "grad_norm": 13.217930793762207, "learning_rate": 7.230463519126024e-07, "loss": 0.4217, "num_input_tokens_seen": 3137339392, "step": 1496 }, { "epoch": 0.080553295362083, "grad_norm": 12.922200202941895, "learning_rate": 7.226736354817908e-07, "loss": 0.4309, "num_input_tokens_seen": 3139436544, "step": 1497 }, { "epoch": 0.08082451857879035, "grad_norm": 26.469648361206055, "learning_rate": 7.223007800898082e-07, "loss": 1.4187, "num_input_tokens_seen": 3141533696, "step": 1498 }, { "epoch": 0.0810957417954977, "grad_norm": 16.948135375976562, "learning_rate": 7.219277860367143e-07, "loss": 0.6953, "num_input_tokens_seen": 3143630848, "step": 1499 }, { "epoch": 0.08136696501220504, "grad_norm": 28.913721084594727, "learning_rate": 7.215546536226805e-07, "loss": 0.7974, "num_input_tokens_seen": 3145728000, "step": 1500 }, { "epoch": 0.08163818822891239, "grad_norm": 12.902350425720215, "learning_rate": 7.211813831479896e-07, "loss": 0.4603, "num_input_tokens_seen": 3147825152, "step": 1501 }, { "epoch": 0.08190941144561975, "grad_norm": 13.105647087097168, "learning_rate": 7.208079749130356e-07, "loss": 0.4464, "num_input_tokens_seen": 3149922304, "step": 1502 }, { "epoch": 0.0821806346623271, "grad_norm": 22.866016387939453, "learning_rate": 7.204344292183228e-07, "loss": 0.5586, "num_input_tokens_seen": 3152019456, "step": 1503 }, { "epoch": 0.08245185787903445, "grad_norm": 12.577240943908691, "learning_rate": 7.200607463644673e-07, "loss": 0.3706, "num_input_tokens_seen": 3154116608, "step": 1504 }, { "epoch": 0.0827230810957418, "grad_norm": 16.995012283325195, "learning_rate": 7.196869266521941e-07, "loss": 0.25, "num_input_tokens_seen": 3156213760, "step": 1505 }, { "epoch": 0.08299430431244914, "grad_norm": 13.941088676452637, "learning_rate": 7.193129703823395e-07, "loss": 0.5905, "num_input_tokens_seen": 3158310912, "step": 1506 }, { "epoch": 0.0832655275291565, "grad_norm": 17.264551162719727, "learning_rate": 7.189388778558491e-07, "loss": 0.5209, "num_input_tokens_seen": 3160408064, "step": 1507 }, { "epoch": 0.08353675074586385, "grad_norm": 9.845773696899414, "learning_rate": 7.185646493737785e-07, "loss": 0.306, "num_input_tokens_seen": 3162505216, "step": 1508 }, { "epoch": 0.0838079739625712, "grad_norm": 12.871875762939453, "learning_rate": 7.181902852372924e-07, "loss": 0.3667, "num_input_tokens_seen": 3164602368, "step": 1509 }, { "epoch": 0.08407919717927854, "grad_norm": 14.597813606262207, "learning_rate": 7.17815785747665e-07, "loss": 0.5319, "num_input_tokens_seen": 3166699520, "step": 1510 }, { "epoch": 0.08435042039598589, "grad_norm": 15.713737487792969, "learning_rate": 7.174411512062789e-07, "loss": 0.4813, "num_input_tokens_seen": 3168796672, "step": 1511 }, { "epoch": 0.08462164361269324, "grad_norm": 15.4694242477417, "learning_rate": 7.170663819146259e-07, "loss": 0.5327, "num_input_tokens_seen": 3170893824, "step": 1512 }, { "epoch": 0.0848928668294006, "grad_norm": 16.36274528503418, "learning_rate": 7.166914781743062e-07, "loss": 0.7225, "num_input_tokens_seen": 3172990976, "step": 1513 }, { "epoch": 0.08516409004610795, "grad_norm": 10.62022876739502, "learning_rate": 7.16316440287028e-07, "loss": 0.3582, "num_input_tokens_seen": 3175088128, "step": 1514 }, { "epoch": 0.0854353132628153, "grad_norm": 19.60194206237793, "learning_rate": 7.159412685546073e-07, "loss": 0.5609, "num_input_tokens_seen": 3177185280, "step": 1515 }, { "epoch": 0.08570653647952264, "grad_norm": 16.77193260192871, "learning_rate": 7.155659632789683e-07, "loss": 0.7256, "num_input_tokens_seen": 3179282432, "step": 1516 }, { "epoch": 0.08597775969622999, "grad_norm": 15.905506134033203, "learning_rate": 7.151905247621422e-07, "loss": 0.6207, "num_input_tokens_seen": 3181379584, "step": 1517 }, { "epoch": 0.08624898291293735, "grad_norm": 13.54134750366211, "learning_rate": 7.148149533062678e-07, "loss": 0.4583, "num_input_tokens_seen": 3183476736, "step": 1518 }, { "epoch": 0.0865202061296447, "grad_norm": 16.33243751525879, "learning_rate": 7.144392492135908e-07, "loss": 0.4972, "num_input_tokens_seen": 3185573888, "step": 1519 }, { "epoch": 0.08679142934635205, "grad_norm": 17.10910987854004, "learning_rate": 7.140634127864632e-07, "loss": 0.6943, "num_input_tokens_seen": 3187671040, "step": 1520 }, { "epoch": 0.0870626525630594, "grad_norm": 19.709762573242188, "learning_rate": 7.136874443273442e-07, "loss": 0.4614, "num_input_tokens_seen": 3189768192, "step": 1521 }, { "epoch": 0.08733387577976674, "grad_norm": 18.73311424255371, "learning_rate": 7.133113441387988e-07, "loss": 0.7051, "num_input_tokens_seen": 3191865344, "step": 1522 }, { "epoch": 0.0876050989964741, "grad_norm": 9.619464874267578, "learning_rate": 7.129351125234979e-07, "loss": 0.2741, "num_input_tokens_seen": 3193962496, "step": 1523 }, { "epoch": 0.08787632221318145, "grad_norm": 16.603302001953125, "learning_rate": 7.125587497842189e-07, "loss": 0.666, "num_input_tokens_seen": 3196059648, "step": 1524 }, { "epoch": 0.0881475454298888, "grad_norm": 10.391305923461914, "learning_rate": 7.121822562238436e-07, "loss": 0.282, "num_input_tokens_seen": 3198156800, "step": 1525 }, { "epoch": 0.08841876864659615, "grad_norm": 15.561429977416992, "learning_rate": 7.118056321453601e-07, "loss": 0.5204, "num_input_tokens_seen": 3200253952, "step": 1526 }, { "epoch": 0.0886899918633035, "grad_norm": 15.614664077758789, "learning_rate": 7.11428877851861e-07, "loss": 0.5045, "num_input_tokens_seen": 3202351104, "step": 1527 }, { "epoch": 0.08896121508001086, "grad_norm": 13.624761581420898, "learning_rate": 7.110519936465438e-07, "loss": 0.4001, "num_input_tokens_seen": 3204448256, "step": 1528 }, { "epoch": 0.0892324382967182, "grad_norm": 15.247365951538086, "learning_rate": 7.106749798327106e-07, "loss": 0.3207, "num_input_tokens_seen": 3206545408, "step": 1529 }, { "epoch": 0.08950366151342555, "grad_norm": 12.790914535522461, "learning_rate": 7.102978367137679e-07, "loss": 0.403, "num_input_tokens_seen": 3208642560, "step": 1530 }, { "epoch": 0.0897748847301329, "grad_norm": 16.698665618896484, "learning_rate": 7.099205645932258e-07, "loss": 0.4211, "num_input_tokens_seen": 3210739712, "step": 1531 }, { "epoch": 0.09004610794684025, "grad_norm": 16.94194221496582, "learning_rate": 7.095431637746988e-07, "loss": 0.6433, "num_input_tokens_seen": 3212836864, "step": 1532 }, { "epoch": 0.09031733116354759, "grad_norm": 12.014008522033691, "learning_rate": 7.091656345619047e-07, "loss": 0.2988, "num_input_tokens_seen": 3214934016, "step": 1533 }, { "epoch": 0.09058855438025495, "grad_norm": 14.939805030822754, "learning_rate": 7.087879772586647e-07, "loss": 0.523, "num_input_tokens_seen": 3217031168, "step": 1534 }, { "epoch": 0.0908597775969623, "grad_norm": 21.286935806274414, "learning_rate": 7.084101921689029e-07, "loss": 0.5719, "num_input_tokens_seen": 3219128320, "step": 1535 }, { "epoch": 0.09113100081366965, "grad_norm": 15.480658531188965, "learning_rate": 7.080322795966462e-07, "loss": 0.536, "num_input_tokens_seen": 3221225472, "step": 1536 }, { "epoch": 0.091402224030377, "grad_norm": 14.61371898651123, "learning_rate": 7.076542398460247e-07, "loss": 0.48, "num_input_tokens_seen": 3223322624, "step": 1537 }, { "epoch": 0.09167344724708434, "grad_norm": 15.25113296508789, "learning_rate": 7.0727607322127e-07, "loss": 0.4631, "num_input_tokens_seen": 3225419776, "step": 1538 }, { "epoch": 0.0919446704637917, "grad_norm": 16.593374252319336, "learning_rate": 7.068977800267164e-07, "loss": 0.4729, "num_input_tokens_seen": 3227516928, "step": 1539 }, { "epoch": 0.09221589368049905, "grad_norm": 11.405816078186035, "learning_rate": 7.065193605667999e-07, "loss": 0.3467, "num_input_tokens_seen": 3229614080, "step": 1540 }, { "epoch": 0.0924871168972064, "grad_norm": 15.258930206298828, "learning_rate": 7.06140815146058e-07, "loss": 0.3913, "num_input_tokens_seen": 3231711232, "step": 1541 }, { "epoch": 0.09275834011391375, "grad_norm": 19.230300903320312, "learning_rate": 7.057621440691296e-07, "loss": 0.6633, "num_input_tokens_seen": 3233808384, "step": 1542 }, { "epoch": 0.0930295633306211, "grad_norm": 17.410680770874023, "learning_rate": 7.053833476407549e-07, "loss": 0.6296, "num_input_tokens_seen": 3235905536, "step": 1543 }, { "epoch": 0.09330078654732846, "grad_norm": 12.020039558410645, "learning_rate": 7.050044261657748e-07, "loss": 0.319, "num_input_tokens_seen": 3238002688, "step": 1544 }, { "epoch": 0.0935720097640358, "grad_norm": 10.420241355895996, "learning_rate": 7.04625379949131e-07, "loss": 0.2475, "num_input_tokens_seen": 3240099840, "step": 1545 }, { "epoch": 0.09384323298074315, "grad_norm": 14.750914573669434, "learning_rate": 7.042462092958651e-07, "loss": 0.4473, "num_input_tokens_seen": 3242196992, "step": 1546 }, { "epoch": 0.0941144561974505, "grad_norm": 12.780601501464844, "learning_rate": 7.038669145111195e-07, "loss": 0.409, "num_input_tokens_seen": 3244294144, "step": 1547 }, { "epoch": 0.09438567941415785, "grad_norm": 16.467487335205078, "learning_rate": 7.034874959001363e-07, "loss": 0.4919, "num_input_tokens_seen": 3246391296, "step": 1548 }, { "epoch": 0.09465690263086521, "grad_norm": 13.055885314941406, "learning_rate": 7.031079537682569e-07, "loss": 0.3439, "num_input_tokens_seen": 3248488448, "step": 1549 }, { "epoch": 0.09492812584757256, "grad_norm": 17.769779205322266, "learning_rate": 7.027282884209227e-07, "loss": 0.6011, "num_input_tokens_seen": 3250585600, "step": 1550 }, { "epoch": 0.0951993490642799, "grad_norm": 15.251014709472656, "learning_rate": 7.023485001636737e-07, "loss": 0.3822, "num_input_tokens_seen": 3252682752, "step": 1551 }, { "epoch": 0.09547057228098725, "grad_norm": 14.227032661437988, "learning_rate": 7.019685893021488e-07, "loss": 0.5338, "num_input_tokens_seen": 3254779904, "step": 1552 }, { "epoch": 0.0957417954976946, "grad_norm": 10.746809005737305, "learning_rate": 7.015885561420863e-07, "loss": 0.3099, "num_input_tokens_seen": 3256877056, "step": 1553 }, { "epoch": 0.09601301871440195, "grad_norm": 10.69054889678955, "learning_rate": 7.012084009893221e-07, "loss": 0.273, "num_input_tokens_seen": 3258974208, "step": 1554 }, { "epoch": 0.09628424193110931, "grad_norm": 15.015727996826172, "learning_rate": 7.008281241497908e-07, "loss": 0.6222, "num_input_tokens_seen": 3261071360, "step": 1555 }, { "epoch": 0.09655546514781665, "grad_norm": 15.404815673828125, "learning_rate": 7.004477259295244e-07, "loss": 0.4674, "num_input_tokens_seen": 3263168512, "step": 1556 }, { "epoch": 0.096826688364524, "grad_norm": 7.422001838684082, "learning_rate": 7.000672066346532e-07, "loss": 0.1853, "num_input_tokens_seen": 3265265664, "step": 1557 }, { "epoch": 0.09709791158123135, "grad_norm": 11.59667682647705, "learning_rate": 6.996865665714046e-07, "loss": 0.3281, "num_input_tokens_seen": 3267362816, "step": 1558 }, { "epoch": 0.0973691347979387, "grad_norm": 17.11145782470703, "learning_rate": 6.993058060461035e-07, "loss": 0.6195, "num_input_tokens_seen": 3269459968, "step": 1559 }, { "epoch": 0.09764035801464606, "grad_norm": 18.34768295288086, "learning_rate": 6.989249253651708e-07, "loss": 0.5665, "num_input_tokens_seen": 3271557120, "step": 1560 }, { "epoch": 0.0979115812313534, "grad_norm": 11.051552772521973, "learning_rate": 6.985439248351254e-07, "loss": 0.2874, "num_input_tokens_seen": 3273654272, "step": 1561 }, { "epoch": 0.09818280444806075, "grad_norm": 23.903244018554688, "learning_rate": 6.981628047625818e-07, "loss": 0.925, "num_input_tokens_seen": 3275751424, "step": 1562 }, { "epoch": 0.0984540276647681, "grad_norm": 16.90290641784668, "learning_rate": 6.977815654542508e-07, "loss": 0.6217, "num_input_tokens_seen": 3277848576, "step": 1563 }, { "epoch": 0.09872525088147545, "grad_norm": 9.315507888793945, "learning_rate": 6.974002072169395e-07, "loss": 0.1993, "num_input_tokens_seen": 3279945728, "step": 1564 }, { "epoch": 0.09899647409818281, "grad_norm": 15.298419952392578, "learning_rate": 6.970187303575505e-07, "loss": 0.336, "num_input_tokens_seen": 3282042880, "step": 1565 }, { "epoch": 0.09926769731489016, "grad_norm": 20.24376106262207, "learning_rate": 6.966371351830818e-07, "loss": 0.8491, "num_input_tokens_seen": 3284140032, "step": 1566 }, { "epoch": 0.0995389205315975, "grad_norm": 11.215019226074219, "learning_rate": 6.962554220006265e-07, "loss": 0.3051, "num_input_tokens_seen": 3286237184, "step": 1567 }, { "epoch": 0.09981014374830485, "grad_norm": 16.658477783203125, "learning_rate": 6.958735911173729e-07, "loss": 0.5136, "num_input_tokens_seen": 3288334336, "step": 1568 }, { "epoch": 0.1000813669650122, "grad_norm": 19.485260009765625, "learning_rate": 6.954916428406045e-07, "loss": 0.6946, "num_input_tokens_seen": 3290431488, "step": 1569 }, { "epoch": 0.10035259018171956, "grad_norm": 15.814196586608887, "learning_rate": 6.95109577477698e-07, "loss": 0.5309, "num_input_tokens_seen": 3292528640, "step": 1570 }, { "epoch": 0.10062381339842691, "grad_norm": 14.682167053222656, "learning_rate": 6.947273953361255e-07, "loss": 0.5357, "num_input_tokens_seen": 3294625792, "step": 1571 }, { "epoch": 0.10089503661513426, "grad_norm": 16.301332473754883, "learning_rate": 6.943450967234524e-07, "loss": 0.388, "num_input_tokens_seen": 3296722944, "step": 1572 }, { "epoch": 0.1011662598318416, "grad_norm": 11.630681991577148, "learning_rate": 6.939626819473384e-07, "loss": 0.3335, "num_input_tokens_seen": 3298820096, "step": 1573 }, { "epoch": 0.10143748304854895, "grad_norm": 21.778146743774414, "learning_rate": 6.93580151315536e-07, "loss": 0.791, "num_input_tokens_seen": 3300917248, "step": 1574 }, { "epoch": 0.10170870626525631, "grad_norm": 16.302976608276367, "learning_rate": 6.931975051358914e-07, "loss": 0.3302, "num_input_tokens_seen": 3303014400, "step": 1575 }, { "epoch": 0.10197992948196366, "grad_norm": 14.719386100769043, "learning_rate": 6.928147437163439e-07, "loss": 0.5707, "num_input_tokens_seen": 3305111552, "step": 1576 }, { "epoch": 0.10225115269867101, "grad_norm": 10.73705768585205, "learning_rate": 6.92431867364925e-07, "loss": 0.3245, "num_input_tokens_seen": 3307208704, "step": 1577 }, { "epoch": 0.10252237591537836, "grad_norm": 21.941381454467773, "learning_rate": 6.920488763897593e-07, "loss": 0.8567, "num_input_tokens_seen": 3309305856, "step": 1578 }, { "epoch": 0.1027935991320857, "grad_norm": 20.246137619018555, "learning_rate": 6.916657710990632e-07, "loss": 0.5603, "num_input_tokens_seen": 3311403008, "step": 1579 }, { "epoch": 0.10306482234879305, "grad_norm": 15.607077598571777, "learning_rate": 6.912825518011452e-07, "loss": 0.6866, "num_input_tokens_seen": 3313500160, "step": 1580 }, { "epoch": 0.10333604556550041, "grad_norm": 13.166605949401855, "learning_rate": 6.90899218804406e-07, "loss": 0.4176, "num_input_tokens_seen": 3315597312, "step": 1581 }, { "epoch": 0.10360726878220776, "grad_norm": 14.276864051818848, "learning_rate": 6.905157724173369e-07, "loss": 0.3567, "num_input_tokens_seen": 3317694464, "step": 1582 }, { "epoch": 0.1038784919989151, "grad_norm": 15.237847328186035, "learning_rate": 6.901322129485212e-07, "loss": 0.5268, "num_input_tokens_seen": 3319791616, "step": 1583 }, { "epoch": 0.10414971521562245, "grad_norm": 10.99821949005127, "learning_rate": 6.897485407066329e-07, "loss": 0.3016, "num_input_tokens_seen": 3321888768, "step": 1584 }, { "epoch": 0.1044209384323298, "grad_norm": 13.16862964630127, "learning_rate": 6.893647560004369e-07, "loss": 0.5617, "num_input_tokens_seen": 3323985920, "step": 1585 }, { "epoch": 0.10469216164903716, "grad_norm": 15.644988059997559, "learning_rate": 6.889808591387885e-07, "loss": 0.4087, "num_input_tokens_seen": 3326083072, "step": 1586 }, { "epoch": 0.10496338486574451, "grad_norm": 14.807315826416016, "learning_rate": 6.885968504306334e-07, "loss": 0.4284, "num_input_tokens_seen": 3328180224, "step": 1587 }, { "epoch": 0.10523460808245186, "grad_norm": 18.928443908691406, "learning_rate": 6.882127301850069e-07, "loss": 0.7086, "num_input_tokens_seen": 3330277376, "step": 1588 }, { "epoch": 0.1055058312991592, "grad_norm": 17.395771026611328, "learning_rate": 6.878284987110345e-07, "loss": 0.5846, "num_input_tokens_seen": 3332374528, "step": 1589 }, { "epoch": 0.10577705451586655, "grad_norm": 10.824260711669922, "learning_rate": 6.874441563179313e-07, "loss": 0.3061, "num_input_tokens_seen": 3334471680, "step": 1590 }, { "epoch": 0.10604827773257391, "grad_norm": 18.152671813964844, "learning_rate": 6.870597033150012e-07, "loss": 0.664, "num_input_tokens_seen": 3336568832, "step": 1591 }, { "epoch": 0.10631950094928126, "grad_norm": 13.247352600097656, "learning_rate": 6.866751400116374e-07, "loss": 0.4636, "num_input_tokens_seen": 3338665984, "step": 1592 }, { "epoch": 0.10659072416598861, "grad_norm": 26.162593841552734, "learning_rate": 6.862904667173216e-07, "loss": 0.7328, "num_input_tokens_seen": 3340763136, "step": 1593 }, { "epoch": 0.10686194738269596, "grad_norm": 22.150165557861328, "learning_rate": 6.859056837416245e-07, "loss": 0.9154, "num_input_tokens_seen": 3342860288, "step": 1594 }, { "epoch": 0.1071331705994033, "grad_norm": 10.474020957946777, "learning_rate": 6.855207913942048e-07, "loss": 0.388, "num_input_tokens_seen": 3344957440, "step": 1595 }, { "epoch": 0.10740439381611067, "grad_norm": 13.631202697753906, "learning_rate": 6.85135789984809e-07, "loss": 0.4211, "num_input_tokens_seen": 3347054592, "step": 1596 }, { "epoch": 0.10767561703281801, "grad_norm": 14.622827529907227, "learning_rate": 6.847506798232719e-07, "loss": 0.5837, "num_input_tokens_seen": 3349151744, "step": 1597 }, { "epoch": 0.10794684024952536, "grad_norm": 16.3840389251709, "learning_rate": 6.843654612195152e-07, "loss": 0.4376, "num_input_tokens_seen": 3351248896, "step": 1598 }, { "epoch": 0.10821806346623271, "grad_norm": 16.240615844726562, "learning_rate": 6.839801344835484e-07, "loss": 0.6028, "num_input_tokens_seen": 3353346048, "step": 1599 }, { "epoch": 0.10848928668294006, "grad_norm": 13.917694091796875, "learning_rate": 6.835946999254677e-07, "loss": 0.6392, "num_input_tokens_seen": 3355443200, "step": 1600 }, { "epoch": 0.1087605098996474, "grad_norm": 11.366427421569824, "learning_rate": 6.83209157855456e-07, "loss": 0.3998, "num_input_tokens_seen": 3357540352, "step": 1601 }, { "epoch": 0.10903173311635476, "grad_norm": 14.33171272277832, "learning_rate": 6.828235085837831e-07, "loss": 0.4276, "num_input_tokens_seen": 3359637504, "step": 1602 }, { "epoch": 0.10930295633306211, "grad_norm": 12.968931198120117, "learning_rate": 6.824377524208047e-07, "loss": 0.489, "num_input_tokens_seen": 3361734656, "step": 1603 }, { "epoch": 0.10957417954976946, "grad_norm": 15.882153511047363, "learning_rate": 6.820518896769629e-07, "loss": 0.5021, "num_input_tokens_seen": 3363831808, "step": 1604 }, { "epoch": 0.10984540276647681, "grad_norm": 10.402983665466309, "learning_rate": 6.816659206627853e-07, "loss": 0.2281, "num_input_tokens_seen": 3365928960, "step": 1605 }, { "epoch": 0.11011662598318415, "grad_norm": 14.085593223571777, "learning_rate": 6.812798456888849e-07, "loss": 0.4662, "num_input_tokens_seen": 3368026112, "step": 1606 }, { "epoch": 0.11038784919989152, "grad_norm": 14.953266143798828, "learning_rate": 6.808936650659605e-07, "loss": 0.6349, "num_input_tokens_seen": 3370123264, "step": 1607 }, { "epoch": 0.11065907241659886, "grad_norm": 15.086980819702148, "learning_rate": 6.805073791047951e-07, "loss": 0.6972, "num_input_tokens_seen": 3372220416, "step": 1608 }, { "epoch": 0.11093029563330621, "grad_norm": 14.327410697937012, "learning_rate": 6.80120988116257e-07, "loss": 0.3387, "num_input_tokens_seen": 3374317568, "step": 1609 }, { "epoch": 0.11120151885001356, "grad_norm": 13.357967376708984, "learning_rate": 6.797344924112995e-07, "loss": 0.3773, "num_input_tokens_seen": 3376414720, "step": 1610 }, { "epoch": 0.1114727420667209, "grad_norm": 16.17928695678711, "learning_rate": 6.793478923009592e-07, "loss": 0.6131, "num_input_tokens_seen": 3378511872, "step": 1611 }, { "epoch": 0.11174396528342827, "grad_norm": 9.867973327636719, "learning_rate": 6.789611880963569e-07, "loss": 0.2377, "num_input_tokens_seen": 3380609024, "step": 1612 }, { "epoch": 0.11201518850013562, "grad_norm": 13.116447448730469, "learning_rate": 6.78574380108698e-07, "loss": 0.4105, "num_input_tokens_seen": 3382706176, "step": 1613 }, { "epoch": 0.11228641171684296, "grad_norm": 13.960227012634277, "learning_rate": 6.781874686492706e-07, "loss": 0.475, "num_input_tokens_seen": 3384803328, "step": 1614 }, { "epoch": 0.11255763493355031, "grad_norm": 11.566934585571289, "learning_rate": 6.778004540294464e-07, "loss": 0.3708, "num_input_tokens_seen": 3386900480, "step": 1615 }, { "epoch": 0.11282885815025766, "grad_norm": 21.169614791870117, "learning_rate": 6.774133365606801e-07, "loss": 0.8485, "num_input_tokens_seen": 3388997632, "step": 1616 }, { "epoch": 0.11310008136696502, "grad_norm": 20.027353286743164, "learning_rate": 6.770261165545085e-07, "loss": 0.8033, "num_input_tokens_seen": 3391094784, "step": 1617 }, { "epoch": 0.11337130458367237, "grad_norm": 19.385282516479492, "learning_rate": 6.766387943225524e-07, "loss": 0.7747, "num_input_tokens_seen": 3393191936, "step": 1618 }, { "epoch": 0.11364252780037971, "grad_norm": 11.849488258361816, "learning_rate": 6.762513701765135e-07, "loss": 0.4182, "num_input_tokens_seen": 3395289088, "step": 1619 }, { "epoch": 0.11391375101708706, "grad_norm": 11.74028205871582, "learning_rate": 6.75863844428176e-07, "loss": 0.3773, "num_input_tokens_seen": 3397386240, "step": 1620 }, { "epoch": 0.11418497423379441, "grad_norm": 16.50984001159668, "learning_rate": 6.754762173894061e-07, "loss": 0.5555, "num_input_tokens_seen": 3399483392, "step": 1621 }, { "epoch": 0.11445619745050176, "grad_norm": 10.819229125976562, "learning_rate": 6.750884893721511e-07, "loss": 0.2958, "num_input_tokens_seen": 3401580544, "step": 1622 }, { "epoch": 0.11472742066720912, "grad_norm": 14.737079620361328, "learning_rate": 6.747006606884398e-07, "loss": 0.5203, "num_input_tokens_seen": 3403677696, "step": 1623 }, { "epoch": 0.11499864388391647, "grad_norm": 19.5377197265625, "learning_rate": 6.74312731650382e-07, "loss": 0.5925, "num_input_tokens_seen": 3405774848, "step": 1624 }, { "epoch": 0.11526986710062381, "grad_norm": 17.311817169189453, "learning_rate": 6.739247025701683e-07, "loss": 0.5197, "num_input_tokens_seen": 3407872000, "step": 1625 }, { "epoch": 0.11554109031733116, "grad_norm": 33.43289566040039, "learning_rate": 6.735365737600695e-07, "loss": 0.8992, "num_input_tokens_seen": 3409969152, "step": 1626 }, { "epoch": 0.11581231353403851, "grad_norm": 12.524608612060547, "learning_rate": 6.731483455324374e-07, "loss": 0.4096, "num_input_tokens_seen": 3412066304, "step": 1627 }, { "epoch": 0.11608353675074587, "grad_norm": 9.463518142700195, "learning_rate": 6.727600181997026e-07, "loss": 0.2309, "num_input_tokens_seen": 3414163456, "step": 1628 }, { "epoch": 0.11635475996745322, "grad_norm": 16.343246459960938, "learning_rate": 6.723715920743767e-07, "loss": 0.5599, "num_input_tokens_seen": 3416260608, "step": 1629 }, { "epoch": 0.11662598318416056, "grad_norm": 12.191635131835938, "learning_rate": 6.7198306746905e-07, "loss": 0.3698, "num_input_tokens_seen": 3418357760, "step": 1630 }, { "epoch": 0.11689720640086791, "grad_norm": 16.117708206176758, "learning_rate": 6.715944446963924e-07, "loss": 0.576, "num_input_tokens_seen": 3420454912, "step": 1631 }, { "epoch": 0.11716842961757526, "grad_norm": 19.668415069580078, "learning_rate": 6.712057240691527e-07, "loss": 0.9028, "num_input_tokens_seen": 3422552064, "step": 1632 }, { "epoch": 0.11743965283428262, "grad_norm": 11.864884376525879, "learning_rate": 6.708169059001586e-07, "loss": 0.4027, "num_input_tokens_seen": 3424649216, "step": 1633 }, { "epoch": 0.11771087605098997, "grad_norm": 15.043469429016113, "learning_rate": 6.704279905023159e-07, "loss": 0.579, "num_input_tokens_seen": 3426746368, "step": 1634 }, { "epoch": 0.11798209926769732, "grad_norm": 21.475412368774414, "learning_rate": 6.700389781886091e-07, "loss": 0.6607, "num_input_tokens_seen": 3428843520, "step": 1635 }, { "epoch": 0.11825332248440466, "grad_norm": 13.802813529968262, "learning_rate": 6.696498692721006e-07, "loss": 0.3898, "num_input_tokens_seen": 3430940672, "step": 1636 }, { "epoch": 0.11852454570111201, "grad_norm": 15.97105598449707, "learning_rate": 6.692606640659302e-07, "loss": 0.5449, "num_input_tokens_seen": 3433037824, "step": 1637 }, { "epoch": 0.11879576891781937, "grad_norm": 18.830066680908203, "learning_rate": 6.688713628833157e-07, "loss": 0.7314, "num_input_tokens_seen": 3435134976, "step": 1638 }, { "epoch": 0.11906699213452672, "grad_norm": 28.016523361206055, "learning_rate": 6.684819660375516e-07, "loss": 0.4921, "num_input_tokens_seen": 3437232128, "step": 1639 }, { "epoch": 0.11933821535123407, "grad_norm": 20.086233139038086, "learning_rate": 6.6809247384201e-07, "loss": 0.6695, "num_input_tokens_seen": 3439329280, "step": 1640 }, { "epoch": 0.11960943856794141, "grad_norm": 12.381033897399902, "learning_rate": 6.67702886610139e-07, "loss": 0.433, "num_input_tokens_seen": 3441426432, "step": 1641 }, { "epoch": 0.11988066178464876, "grad_norm": 16.39443016052246, "learning_rate": 6.673132046554639e-07, "loss": 0.6587, "num_input_tokens_seen": 3443523584, "step": 1642 }, { "epoch": 0.12015188500135611, "grad_norm": 9.596854209899902, "learning_rate": 6.669234282915857e-07, "loss": 0.2475, "num_input_tokens_seen": 3445620736, "step": 1643 }, { "epoch": 0.12042310821806347, "grad_norm": 14.690909385681152, "learning_rate": 6.665335578321819e-07, "loss": 0.4947, "num_input_tokens_seen": 3447717888, "step": 1644 }, { "epoch": 0.12069433143477082, "grad_norm": 19.749242782592773, "learning_rate": 6.661435935910048e-07, "loss": 0.6977, "num_input_tokens_seen": 3449815040, "step": 1645 }, { "epoch": 0.12096555465147817, "grad_norm": 11.498983383178711, "learning_rate": 6.657535358818833e-07, "loss": 0.3786, "num_input_tokens_seen": 3451912192, "step": 1646 }, { "epoch": 0.12123677786818551, "grad_norm": 18.334880828857422, "learning_rate": 6.653633850187211e-07, "loss": 0.7039, "num_input_tokens_seen": 3454009344, "step": 1647 }, { "epoch": 0.12150800108489286, "grad_norm": 14.584370613098145, "learning_rate": 6.649731413154964e-07, "loss": 0.4845, "num_input_tokens_seen": 3456106496, "step": 1648 }, { "epoch": 0.12177922430160022, "grad_norm": 14.279857635498047, "learning_rate": 6.645828050862626e-07, "loss": 0.5301, "num_input_tokens_seen": 3458203648, "step": 1649 }, { "epoch": 0.12205044751830757, "grad_norm": 9.879186630249023, "learning_rate": 6.641923766451475e-07, "loss": 0.2123, "num_input_tokens_seen": 3460300800, "step": 1650 }, { "epoch": 0.12232167073501492, "grad_norm": 11.519014358520508, "learning_rate": 6.63801856306353e-07, "loss": 0.4437, "num_input_tokens_seen": 3462397952, "step": 1651 }, { "epoch": 0.12259289395172226, "grad_norm": 15.039155960083008, "learning_rate": 6.634112443841551e-07, "loss": 0.5951, "num_input_tokens_seen": 3464495104, "step": 1652 }, { "epoch": 0.12286411716842961, "grad_norm": 14.848783493041992, "learning_rate": 6.630205411929032e-07, "loss": 0.5456, "num_input_tokens_seen": 3466592256, "step": 1653 }, { "epoch": 0.12313534038513697, "grad_norm": 10.93878173828125, "learning_rate": 6.626297470470205e-07, "loss": 0.3535, "num_input_tokens_seen": 3468689408, "step": 1654 }, { "epoch": 0.12340656360184432, "grad_norm": 11.482619285583496, "learning_rate": 6.622388622610034e-07, "loss": 0.2725, "num_input_tokens_seen": 3470786560, "step": 1655 }, { "epoch": 0.12367778681855167, "grad_norm": 12.639951705932617, "learning_rate": 6.618478871494209e-07, "loss": 0.3996, "num_input_tokens_seen": 3472883712, "step": 1656 }, { "epoch": 0.12394901003525902, "grad_norm": 12.591337203979492, "learning_rate": 6.61456822026915e-07, "loss": 0.3903, "num_input_tokens_seen": 3474980864, "step": 1657 }, { "epoch": 0.12422023325196636, "grad_norm": 18.513185501098633, "learning_rate": 6.610656672081999e-07, "loss": 0.4721, "num_input_tokens_seen": 3477078016, "step": 1658 }, { "epoch": 0.12449145646867373, "grad_norm": 16.166847229003906, "learning_rate": 6.606744230080622e-07, "loss": 0.5814, "num_input_tokens_seen": 3479175168, "step": 1659 }, { "epoch": 0.12476267968538107, "grad_norm": 19.309938430786133, "learning_rate": 6.602830897413603e-07, "loss": 0.7281, "num_input_tokens_seen": 3481272320, "step": 1660 }, { "epoch": 0.1250339029020884, "grad_norm": 20.90107536315918, "learning_rate": 6.598916677230243e-07, "loss": 0.9278, "num_input_tokens_seen": 3483369472, "step": 1661 }, { "epoch": 0.12530512611879577, "grad_norm": 14.196072578430176, "learning_rate": 6.59500157268056e-07, "loss": 0.488, "num_input_tokens_seen": 3485466624, "step": 1662 }, { "epoch": 0.12557634933550313, "grad_norm": 14.922718048095703, "learning_rate": 6.591085586915279e-07, "loss": 0.4756, "num_input_tokens_seen": 3487563776, "step": 1663 }, { "epoch": 0.12584757255221046, "grad_norm": 19.60527992248535, "learning_rate": 6.587168723085836e-07, "loss": 0.7925, "num_input_tokens_seen": 3489660928, "step": 1664 }, { "epoch": 0.12611879576891782, "grad_norm": 13.591885566711426, "learning_rate": 6.583250984344374e-07, "loss": 0.4419, "num_input_tokens_seen": 3491758080, "step": 1665 }, { "epoch": 0.12639001898562516, "grad_norm": 13.59085464477539, "learning_rate": 6.57933237384374e-07, "loss": 0.4151, "num_input_tokens_seen": 3493855232, "step": 1666 }, { "epoch": 0.12666124220233252, "grad_norm": 18.338226318359375, "learning_rate": 6.575412894737484e-07, "loss": 0.6672, "num_input_tokens_seen": 3495952384, "step": 1667 }, { "epoch": 0.12693246541903988, "grad_norm": 19.015380859375, "learning_rate": 6.571492550179853e-07, "loss": 0.9037, "num_input_tokens_seen": 3498049536, "step": 1668 }, { "epoch": 0.12720368863574721, "grad_norm": 15.90075969696045, "learning_rate": 6.567571343325791e-07, "loss": 0.6357, "num_input_tokens_seen": 3500146688, "step": 1669 }, { "epoch": 0.12747491185245458, "grad_norm": 17.481428146362305, "learning_rate": 6.563649277330935e-07, "loss": 0.5654, "num_input_tokens_seen": 3502243840, "step": 1670 }, { "epoch": 0.1277461350691619, "grad_norm": 13.026409149169922, "learning_rate": 6.559726355351617e-07, "loss": 0.4402, "num_input_tokens_seen": 3504340992, "step": 1671 }, { "epoch": 0.12801735828586927, "grad_norm": 12.516473770141602, "learning_rate": 6.555802580544853e-07, "loss": 0.3762, "num_input_tokens_seen": 3506438144, "step": 1672 }, { "epoch": 0.12828858150257663, "grad_norm": 16.42319107055664, "learning_rate": 6.551877956068349e-07, "loss": 0.4072, "num_input_tokens_seen": 3508535296, "step": 1673 }, { "epoch": 0.12855980471928397, "grad_norm": 23.695953369140625, "learning_rate": 6.547952485080491e-07, "loss": 0.6364, "num_input_tokens_seen": 3510632448, "step": 1674 }, { "epoch": 0.12883102793599133, "grad_norm": 11.105734825134277, "learning_rate": 6.544026170740352e-07, "loss": 0.3398, "num_input_tokens_seen": 3512729600, "step": 1675 }, { "epoch": 0.12910225115269866, "grad_norm": 12.451509475708008, "learning_rate": 6.540099016207679e-07, "loss": 0.42, "num_input_tokens_seen": 3514826752, "step": 1676 }, { "epoch": 0.12937347436940602, "grad_norm": 12.851999282836914, "learning_rate": 6.536171024642896e-07, "loss": 0.4315, "num_input_tokens_seen": 3516923904, "step": 1677 }, { "epoch": 0.12964469758611338, "grad_norm": 14.279557228088379, "learning_rate": 6.532242199207103e-07, "loss": 0.4199, "num_input_tokens_seen": 3519021056, "step": 1678 }, { "epoch": 0.12991592080282072, "grad_norm": 12.546601295471191, "learning_rate": 6.528312543062066e-07, "loss": 0.2372, "num_input_tokens_seen": 3521118208, "step": 1679 }, { "epoch": 0.13018714401952808, "grad_norm": 17.787132263183594, "learning_rate": 6.524382059370226e-07, "loss": 0.6352, "num_input_tokens_seen": 3523215360, "step": 1680 }, { "epoch": 0.1304583672362354, "grad_norm": 14.964465141296387, "learning_rate": 6.520450751294685e-07, "loss": 0.5617, "num_input_tokens_seen": 3525312512, "step": 1681 }, { "epoch": 0.13072959045294277, "grad_norm": 24.104333877563477, "learning_rate": 6.516518621999209e-07, "loss": 0.7431, "num_input_tokens_seen": 3527409664, "step": 1682 }, { "epoch": 0.13100081366965013, "grad_norm": 15.481319427490234, "learning_rate": 6.512585674648227e-07, "loss": 0.5991, "num_input_tokens_seen": 3529506816, "step": 1683 }, { "epoch": 0.13127203688635747, "grad_norm": 9.043676376342773, "learning_rate": 6.50865191240683e-07, "loss": 0.2865, "num_input_tokens_seen": 3531603968, "step": 1684 }, { "epoch": 0.13154326010306483, "grad_norm": 15.140732765197754, "learning_rate": 6.504717338440751e-07, "loss": 0.5084, "num_input_tokens_seen": 3533701120, "step": 1685 }, { "epoch": 0.13181448331977216, "grad_norm": 10.798766136169434, "learning_rate": 6.500781955916393e-07, "loss": 0.3529, "num_input_tokens_seen": 3535798272, "step": 1686 }, { "epoch": 0.13208570653647952, "grad_norm": 15.705323219299316, "learning_rate": 6.4968457680008e-07, "loss": 0.5974, "num_input_tokens_seen": 3537895424, "step": 1687 }, { "epoch": 0.13235692975318689, "grad_norm": 12.985349655151367, "learning_rate": 6.492908777861664e-07, "loss": 0.3212, "num_input_tokens_seen": 3539992576, "step": 1688 }, { "epoch": 0.13262815296989422, "grad_norm": 17.304018020629883, "learning_rate": 6.488970988667327e-07, "loss": 0.7088, "num_input_tokens_seen": 3542089728, "step": 1689 }, { "epoch": 0.13289937618660158, "grad_norm": 24.675233840942383, "learning_rate": 6.485032403586772e-07, "loss": 0.9613, "num_input_tokens_seen": 3544186880, "step": 1690 }, { "epoch": 0.13317059940330891, "grad_norm": 14.140911102294922, "learning_rate": 6.481093025789621e-07, "loss": 0.3862, "num_input_tokens_seen": 3546284032, "step": 1691 }, { "epoch": 0.13344182262001628, "grad_norm": 17.960968017578125, "learning_rate": 6.477152858446136e-07, "loss": 0.5059, "num_input_tokens_seen": 3548381184, "step": 1692 }, { "epoch": 0.13371304583672364, "grad_norm": 14.560555458068848, "learning_rate": 6.473211904727216e-07, "loss": 0.5256, "num_input_tokens_seen": 3550478336, "step": 1693 }, { "epoch": 0.13398426905343097, "grad_norm": 15.947834014892578, "learning_rate": 6.469270167804386e-07, "loss": 0.5166, "num_input_tokens_seen": 3552575488, "step": 1694 }, { "epoch": 0.13425549227013833, "grad_norm": 13.651203155517578, "learning_rate": 6.46532765084981e-07, "loss": 0.4082, "num_input_tokens_seen": 3554672640, "step": 1695 }, { "epoch": 0.13452671548684567, "grad_norm": 11.05508041381836, "learning_rate": 6.461384357036274e-07, "loss": 0.3031, "num_input_tokens_seen": 3556769792, "step": 1696 }, { "epoch": 0.13479793870355303, "grad_norm": 18.280160903930664, "learning_rate": 6.457440289537191e-07, "loss": 0.5375, "num_input_tokens_seen": 3558866944, "step": 1697 }, { "epoch": 0.13506916192026036, "grad_norm": 10.277115821838379, "learning_rate": 6.453495451526595e-07, "loss": 0.2785, "num_input_tokens_seen": 3560964096, "step": 1698 }, { "epoch": 0.13534038513696772, "grad_norm": 13.42917537689209, "learning_rate": 6.449549846179144e-07, "loss": 0.3343, "num_input_tokens_seen": 3563061248, "step": 1699 }, { "epoch": 0.13561160835367508, "grad_norm": 20.96129035949707, "learning_rate": 6.445603476670109e-07, "loss": 0.5684, "num_input_tokens_seen": 3565158400, "step": 1700 }, { "epoch": 0.13588283157038242, "grad_norm": 18.72563934326172, "learning_rate": 6.44165634617538e-07, "loss": 0.8015, "num_input_tokens_seen": 3567255552, "step": 1701 }, { "epoch": 0.13615405478708978, "grad_norm": 12.687376022338867, "learning_rate": 6.437708457871455e-07, "loss": 0.4409, "num_input_tokens_seen": 3569352704, "step": 1702 }, { "epoch": 0.1364252780037971, "grad_norm": 17.298492431640625, "learning_rate": 6.433759814935447e-07, "loss": 0.56, "num_input_tokens_seen": 3571449856, "step": 1703 }, { "epoch": 0.13669650122050447, "grad_norm": 10.709501266479492, "learning_rate": 6.429810420545072e-07, "loss": 0.3482, "num_input_tokens_seen": 3573547008, "step": 1704 }, { "epoch": 0.13696772443721184, "grad_norm": 15.926192283630371, "learning_rate": 6.425860277878651e-07, "loss": 0.6368, "num_input_tokens_seen": 3575644160, "step": 1705 }, { "epoch": 0.13723894765391917, "grad_norm": 13.588170051574707, "learning_rate": 6.42190939011511e-07, "loss": 0.4412, "num_input_tokens_seen": 3577741312, "step": 1706 }, { "epoch": 0.13751017087062653, "grad_norm": 13.829466819763184, "learning_rate": 6.417957760433974e-07, "loss": 0.4696, "num_input_tokens_seen": 3579838464, "step": 1707 }, { "epoch": 0.13778139408733386, "grad_norm": 17.291702270507812, "learning_rate": 6.414005392015364e-07, "loss": 0.4773, "num_input_tokens_seen": 3581935616, "step": 1708 }, { "epoch": 0.13805261730404123, "grad_norm": 15.568994522094727, "learning_rate": 6.410052288039994e-07, "loss": 0.631, "num_input_tokens_seen": 3584032768, "step": 1709 }, { "epoch": 0.1383238405207486, "grad_norm": 13.949945449829102, "learning_rate": 6.406098451689171e-07, "loss": 0.3946, "num_input_tokens_seen": 3586129920, "step": 1710 }, { "epoch": 0.13859506373745592, "grad_norm": 18.95940589904785, "learning_rate": 6.402143886144797e-07, "loss": 0.453, "num_input_tokens_seen": 3588227072, "step": 1711 }, { "epoch": 0.13886628695416328, "grad_norm": 17.928314208984375, "learning_rate": 6.39818859458935e-07, "loss": 0.4933, "num_input_tokens_seen": 3590324224, "step": 1712 }, { "epoch": 0.13913751017087062, "grad_norm": 24.465429306030273, "learning_rate": 6.394232580205903e-07, "loss": 0.9862, "num_input_tokens_seen": 3592421376, "step": 1713 }, { "epoch": 0.13940873338757798, "grad_norm": 16.95507049560547, "learning_rate": 6.390275846178102e-07, "loss": 0.6797, "num_input_tokens_seen": 3594518528, "step": 1714 }, { "epoch": 0.13967995660428534, "grad_norm": 12.937496185302734, "learning_rate": 6.386318395690178e-07, "loss": 0.296, "num_input_tokens_seen": 3596615680, "step": 1715 }, { "epoch": 0.13995117982099267, "grad_norm": 16.558284759521484, "learning_rate": 6.382360231926935e-07, "loss": 0.4922, "num_input_tokens_seen": 3598712832, "step": 1716 }, { "epoch": 0.14022240303770003, "grad_norm": 15.192840576171875, "learning_rate": 6.378401358073754e-07, "loss": 0.4696, "num_input_tokens_seen": 3600809984, "step": 1717 }, { "epoch": 0.14049362625440737, "grad_norm": 13.925806045532227, "learning_rate": 6.374441777316587e-07, "loss": 0.539, "num_input_tokens_seen": 3602907136, "step": 1718 }, { "epoch": 0.14076484947111473, "grad_norm": 12.96522331237793, "learning_rate": 6.370481492841952e-07, "loss": 0.438, "num_input_tokens_seen": 3605004288, "step": 1719 }, { "epoch": 0.1410360726878221, "grad_norm": 13.147494316101074, "learning_rate": 6.366520507836934e-07, "loss": 0.4557, "num_input_tokens_seen": 3607101440, "step": 1720 }, { "epoch": 0.14130729590452942, "grad_norm": 14.254191398620605, "learning_rate": 6.362558825489187e-07, "loss": 0.337, "num_input_tokens_seen": 3609198592, "step": 1721 }, { "epoch": 0.14157851912123678, "grad_norm": 16.945711135864258, "learning_rate": 6.35859644898692e-07, "loss": 0.565, "num_input_tokens_seen": 3611295744, "step": 1722 }, { "epoch": 0.14184974233794412, "grad_norm": 16.973342895507812, "learning_rate": 6.354633381518901e-07, "loss": 0.6576, "num_input_tokens_seen": 3613392896, "step": 1723 }, { "epoch": 0.14212096555465148, "grad_norm": 15.234251022338867, "learning_rate": 6.350669626274461e-07, "loss": 0.4715, "num_input_tokens_seen": 3615490048, "step": 1724 }, { "epoch": 0.14239218877135884, "grad_norm": 14.848320007324219, "learning_rate": 6.346705186443474e-07, "loss": 0.4606, "num_input_tokens_seen": 3617587200, "step": 1725 }, { "epoch": 0.14266341198806617, "grad_norm": 14.895830154418945, "learning_rate": 6.342740065216371e-07, "loss": 0.5635, "num_input_tokens_seen": 3619684352, "step": 1726 }, { "epoch": 0.14293463520477354, "grad_norm": 12.153023719787598, "learning_rate": 6.338774265784134e-07, "loss": 0.3342, "num_input_tokens_seen": 3621781504, "step": 1727 }, { "epoch": 0.14320585842148087, "grad_norm": 14.42286205291748, "learning_rate": 6.334807791338286e-07, "loss": 0.4809, "num_input_tokens_seen": 3623878656, "step": 1728 }, { "epoch": 0.14347708163818823, "grad_norm": 18.29535675048828, "learning_rate": 6.330840645070894e-07, "loss": 0.6635, "num_input_tokens_seen": 3625975808, "step": 1729 }, { "epoch": 0.1437483048548956, "grad_norm": 14.65046215057373, "learning_rate": 6.326872830174566e-07, "loss": 0.5119, "num_input_tokens_seen": 3628072960, "step": 1730 }, { "epoch": 0.14401952807160293, "grad_norm": 15.046059608459473, "learning_rate": 6.32290434984245e-07, "loss": 0.4347, "num_input_tokens_seen": 3630170112, "step": 1731 }, { "epoch": 0.1442907512883103, "grad_norm": 15.064492225646973, "learning_rate": 6.318935207268227e-07, "loss": 0.5375, "num_input_tokens_seen": 3632267264, "step": 1732 }, { "epoch": 0.14456197450501762, "grad_norm": 12.767688751220703, "learning_rate": 6.314965405646113e-07, "loss": 0.5119, "num_input_tokens_seen": 3634364416, "step": 1733 }, { "epoch": 0.14483319772172498, "grad_norm": 22.676115036010742, "learning_rate": 6.31099494817085e-07, "loss": 0.7369, "num_input_tokens_seen": 3636461568, "step": 1734 }, { "epoch": 0.14510442093843234, "grad_norm": 12.052305221557617, "learning_rate": 6.307023838037714e-07, "loss": 0.4235, "num_input_tokens_seen": 3638558720, "step": 1735 }, { "epoch": 0.14537564415513968, "grad_norm": 21.096866607666016, "learning_rate": 6.303052078442503e-07, "loss": 0.6735, "num_input_tokens_seen": 3640655872, "step": 1736 }, { "epoch": 0.14564686737184704, "grad_norm": 13.492481231689453, "learning_rate": 6.299079672581537e-07, "loss": 0.4236, "num_input_tokens_seen": 3642753024, "step": 1737 }, { "epoch": 0.14591809058855437, "grad_norm": 15.035043716430664, "learning_rate": 6.295106623651661e-07, "loss": 0.5539, "num_input_tokens_seen": 3644850176, "step": 1738 }, { "epoch": 0.14618931380526173, "grad_norm": 11.176956176757812, "learning_rate": 6.291132934850225e-07, "loss": 0.3057, "num_input_tokens_seen": 3646947328, "step": 1739 }, { "epoch": 0.14646053702196907, "grad_norm": 19.135766983032227, "learning_rate": 6.28715860937511e-07, "loss": 0.8421, "num_input_tokens_seen": 3649044480, "step": 1740 }, { "epoch": 0.14673176023867643, "grad_norm": 16.260631561279297, "learning_rate": 6.283183650424701e-07, "loss": 0.528, "num_input_tokens_seen": 3651141632, "step": 1741 }, { "epoch": 0.1470029834553838, "grad_norm": 18.478513717651367, "learning_rate": 6.279208061197892e-07, "loss": 0.7138, "num_input_tokens_seen": 3653238784, "step": 1742 }, { "epoch": 0.14727420667209112, "grad_norm": 13.567277908325195, "learning_rate": 6.275231844894086e-07, "loss": 0.43, "num_input_tokens_seen": 3655335936, "step": 1743 }, { "epoch": 0.14754542988879848, "grad_norm": 16.15982437133789, "learning_rate": 6.271255004713192e-07, "loss": 0.494, "num_input_tokens_seen": 3657433088, "step": 1744 }, { "epoch": 0.14781665310550582, "grad_norm": 10.626032829284668, "learning_rate": 6.267277543855618e-07, "loss": 0.3169, "num_input_tokens_seen": 3659530240, "step": 1745 }, { "epoch": 0.14808787632221318, "grad_norm": 16.266632080078125, "learning_rate": 6.263299465522274e-07, "loss": 0.5772, "num_input_tokens_seen": 3661627392, "step": 1746 }, { "epoch": 0.14835909953892054, "grad_norm": 12.792734146118164, "learning_rate": 6.259320772914566e-07, "loss": 0.4696, "num_input_tokens_seen": 3663724544, "step": 1747 }, { "epoch": 0.14863032275562787, "grad_norm": 18.667470932006836, "learning_rate": 6.255341469234393e-07, "loss": 0.8274, "num_input_tokens_seen": 3665821696, "step": 1748 }, { "epoch": 0.14890154597233524, "grad_norm": 15.332829475402832, "learning_rate": 6.25136155768415e-07, "loss": 0.6095, "num_input_tokens_seen": 3667918848, "step": 1749 }, { "epoch": 0.14917276918904257, "grad_norm": 12.824024200439453, "learning_rate": 6.247381041466716e-07, "loss": 0.4388, "num_input_tokens_seen": 3670016000, "step": 1750 }, { "epoch": 0.14944399240574993, "grad_norm": 9.92809772491455, "learning_rate": 6.243399923785459e-07, "loss": 0.2921, "num_input_tokens_seen": 3672113152, "step": 1751 }, { "epoch": 0.1497152156224573, "grad_norm": 15.665802001953125, "learning_rate": 6.239418207844232e-07, "loss": 0.6068, "num_input_tokens_seen": 3674210304, "step": 1752 }, { "epoch": 0.14998643883916463, "grad_norm": 15.602997779846191, "learning_rate": 6.23543589684737e-07, "loss": 0.5003, "num_input_tokens_seen": 3676307456, "step": 1753 }, { "epoch": 0.150257662055872, "grad_norm": 15.580018043518066, "learning_rate": 6.231452993999683e-07, "loss": 0.5314, "num_input_tokens_seen": 3678404608, "step": 1754 }, { "epoch": 0.15052888527257932, "grad_norm": 18.052650451660156, "learning_rate": 6.22746950250646e-07, "loss": 0.5794, "num_input_tokens_seen": 3680501760, "step": 1755 }, { "epoch": 0.15080010848928668, "grad_norm": 14.942179679870605, "learning_rate": 6.223485425573463e-07, "loss": 0.4202, "num_input_tokens_seen": 3682598912, "step": 1756 }, { "epoch": 0.15107133170599404, "grad_norm": 12.076883316040039, "learning_rate": 6.219500766406926e-07, "loss": 0.385, "num_input_tokens_seen": 3684696064, "step": 1757 }, { "epoch": 0.15134255492270138, "grad_norm": 16.620975494384766, "learning_rate": 6.215515528213553e-07, "loss": 0.5812, "num_input_tokens_seen": 3686793216, "step": 1758 }, { "epoch": 0.15161377813940874, "grad_norm": 11.731767654418945, "learning_rate": 6.211529714200509e-07, "loss": 0.376, "num_input_tokens_seen": 3688890368, "step": 1759 }, { "epoch": 0.15188500135611607, "grad_norm": 24.166730880737305, "learning_rate": 6.207543327575426e-07, "loss": 0.5209, "num_input_tokens_seen": 3690987520, "step": 1760 }, { "epoch": 0.15215622457282343, "grad_norm": 14.723969459533691, "learning_rate": 6.203556371546399e-07, "loss": 0.4827, "num_input_tokens_seen": 3693084672, "step": 1761 }, { "epoch": 0.1524274477895308, "grad_norm": 14.81787395477295, "learning_rate": 6.199568849321975e-07, "loss": 0.4707, "num_input_tokens_seen": 3695181824, "step": 1762 }, { "epoch": 0.15269867100623813, "grad_norm": 15.406253814697266, "learning_rate": 6.195580764111163e-07, "loss": 0.4198, "num_input_tokens_seen": 3697278976, "step": 1763 }, { "epoch": 0.1529698942229455, "grad_norm": 12.230108261108398, "learning_rate": 6.191592119123419e-07, "loss": 0.2609, "num_input_tokens_seen": 3699376128, "step": 1764 }, { "epoch": 0.15324111743965282, "grad_norm": 7.24866247177124, "learning_rate": 6.187602917568655e-07, "loss": 0.1833, "num_input_tokens_seen": 3701473280, "step": 1765 }, { "epoch": 0.15351234065636019, "grad_norm": 10.991303443908691, "learning_rate": 6.183613162657229e-07, "loss": 0.2878, "num_input_tokens_seen": 3703570432, "step": 1766 }, { "epoch": 0.15378356387306755, "grad_norm": 13.555655479431152, "learning_rate": 6.179622857599942e-07, "loss": 0.4992, "num_input_tokens_seen": 3705667584, "step": 1767 }, { "epoch": 0.15405478708977488, "grad_norm": 21.7625675201416, "learning_rate": 6.175632005608043e-07, "loss": 0.6617, "num_input_tokens_seen": 3707764736, "step": 1768 }, { "epoch": 0.15432601030648224, "grad_norm": 16.64663314819336, "learning_rate": 6.171640609893213e-07, "loss": 0.4283, "num_input_tokens_seen": 3709861888, "step": 1769 }, { "epoch": 0.15459723352318958, "grad_norm": 14.173835754394531, "learning_rate": 6.167648673667581e-07, "loss": 0.4511, "num_input_tokens_seen": 3711959040, "step": 1770 }, { "epoch": 0.15486845673989694, "grad_norm": 19.849355697631836, "learning_rate": 6.1636562001437e-07, "loss": 0.619, "num_input_tokens_seen": 3714056192, "step": 1771 }, { "epoch": 0.1551396799566043, "grad_norm": 13.180706024169922, "learning_rate": 6.159663192534561e-07, "loss": 0.486, "num_input_tokens_seen": 3716153344, "step": 1772 }, { "epoch": 0.15541090317331163, "grad_norm": 10.8189697265625, "learning_rate": 6.155669654053592e-07, "loss": 0.3495, "num_input_tokens_seen": 3718250496, "step": 1773 }, { "epoch": 0.155682126390019, "grad_norm": 28.293182373046875, "learning_rate": 6.151675587914631e-07, "loss": 0.7353, "num_input_tokens_seen": 3720347648, "step": 1774 }, { "epoch": 0.15595334960672633, "grad_norm": 21.1544132232666, "learning_rate": 6.147680997331958e-07, "loss": 0.8318, "num_input_tokens_seen": 3722444800, "step": 1775 }, { "epoch": 0.1562245728234337, "grad_norm": 16.06313705444336, "learning_rate": 6.143685885520263e-07, "loss": 0.5683, "num_input_tokens_seen": 3724541952, "step": 1776 }, { "epoch": 0.15649579604014105, "grad_norm": 14.972505569458008, "learning_rate": 6.13969025569466e-07, "loss": 0.3172, "num_input_tokens_seen": 3726639104, "step": 1777 }, { "epoch": 0.15676701925684838, "grad_norm": 11.951261520385742, "learning_rate": 6.135694111070684e-07, "loss": 0.3382, "num_input_tokens_seen": 3728736256, "step": 1778 }, { "epoch": 0.15703824247355574, "grad_norm": 15.1442289352417, "learning_rate": 6.131697454864277e-07, "loss": 0.433, "num_input_tokens_seen": 3730833408, "step": 1779 }, { "epoch": 0.15730946569026308, "grad_norm": 18.318092346191406, "learning_rate": 6.127700290291794e-07, "loss": 0.6716, "num_input_tokens_seen": 3732930560, "step": 1780 }, { "epoch": 0.15758068890697044, "grad_norm": 17.144010543823242, "learning_rate": 6.123702620570005e-07, "loss": 0.5492, "num_input_tokens_seen": 3735027712, "step": 1781 }, { "epoch": 0.15785191212367777, "grad_norm": 18.71530532836914, "learning_rate": 6.11970444891608e-07, "loss": 0.5079, "num_input_tokens_seen": 3737124864, "step": 1782 }, { "epoch": 0.15812313534038513, "grad_norm": 11.793743133544922, "learning_rate": 6.115705778547597e-07, "loss": 0.258, "num_input_tokens_seen": 3739222016, "step": 1783 }, { "epoch": 0.1583943585570925, "grad_norm": 18.33780288696289, "learning_rate": 6.111706612682532e-07, "loss": 0.6809, "num_input_tokens_seen": 3741319168, "step": 1784 }, { "epoch": 0.15866558177379983, "grad_norm": 12.799018859863281, "learning_rate": 6.107706954539261e-07, "loss": 0.3735, "num_input_tokens_seen": 3743416320, "step": 1785 }, { "epoch": 0.1589368049905072, "grad_norm": 14.640045166015625, "learning_rate": 6.103706807336559e-07, "loss": 0.5625, "num_input_tokens_seen": 3745513472, "step": 1786 }, { "epoch": 0.15920802820721452, "grad_norm": 11.989145278930664, "learning_rate": 6.099706174293592e-07, "loss": 0.4499, "num_input_tokens_seen": 3747610624, "step": 1787 }, { "epoch": 0.1594792514239219, "grad_norm": 14.482452392578125, "learning_rate": 6.095705058629914e-07, "loss": 0.3148, "num_input_tokens_seen": 3749707776, "step": 1788 }, { "epoch": 0.15975047464062925, "grad_norm": 12.022076606750488, "learning_rate": 6.091703463565475e-07, "loss": 0.3346, "num_input_tokens_seen": 3751804928, "step": 1789 }, { "epoch": 0.16002169785733658, "grad_norm": 14.825411796569824, "learning_rate": 6.087701392320606e-07, "loss": 0.5093, "num_input_tokens_seen": 3753902080, "step": 1790 }, { "epoch": 0.16029292107404394, "grad_norm": 12.915584564208984, "learning_rate": 6.083698848116018e-07, "loss": 0.3152, "num_input_tokens_seen": 3755999232, "step": 1791 }, { "epoch": 0.16056414429075128, "grad_norm": 21.649930953979492, "learning_rate": 6.079695834172808e-07, "loss": 0.9817, "num_input_tokens_seen": 3758096384, "step": 1792 }, { "epoch": 0.16083536750745864, "grad_norm": 15.43675422668457, "learning_rate": 6.075692353712451e-07, "loss": 0.4618, "num_input_tokens_seen": 3760193536, "step": 1793 }, { "epoch": 0.161106590724166, "grad_norm": 16.996694564819336, "learning_rate": 6.071688409956793e-07, "loss": 0.5975, "num_input_tokens_seen": 3762290688, "step": 1794 }, { "epoch": 0.16137781394087333, "grad_norm": 16.773155212402344, "learning_rate": 6.06768400612806e-07, "loss": 0.5768, "num_input_tokens_seen": 3764387840, "step": 1795 }, { "epoch": 0.1616490371575807, "grad_norm": 22.683000564575195, "learning_rate": 6.063679145448838e-07, "loss": 0.7318, "num_input_tokens_seen": 3766484992, "step": 1796 }, { "epoch": 0.16192026037428803, "grad_norm": 13.12799072265625, "learning_rate": 6.05967383114209e-07, "loss": 0.3581, "num_input_tokens_seen": 3768582144, "step": 1797 }, { "epoch": 0.1621914835909954, "grad_norm": 18.787900924682617, "learning_rate": 6.055668066431142e-07, "loss": 0.6844, "num_input_tokens_seen": 3770679296, "step": 1798 }, { "epoch": 0.16246270680770275, "grad_norm": 10.89427661895752, "learning_rate": 6.051661854539677e-07, "loss": 0.3149, "num_input_tokens_seen": 3772776448, "step": 1799 }, { "epoch": 0.16273393002441008, "grad_norm": 18.79866600036621, "learning_rate": 6.047655198691742e-07, "loss": 0.8429, "num_input_tokens_seen": 3774873600, "step": 1800 }, { "epoch": 0.16300515324111745, "grad_norm": 13.372797966003418, "learning_rate": 6.043648102111745e-07, "loss": 0.2438, "num_input_tokens_seen": 3776970752, "step": 1801 }, { "epoch": 0.16327637645782478, "grad_norm": 18.644472122192383, "learning_rate": 6.039640568024443e-07, "loss": 0.6407, "num_input_tokens_seen": 3779067904, "step": 1802 }, { "epoch": 0.16354759967453214, "grad_norm": 15.050209045410156, "learning_rate": 6.035632599654946e-07, "loss": 0.5603, "num_input_tokens_seen": 3781165056, "step": 1803 }, { "epoch": 0.1638188228912395, "grad_norm": 12.185089111328125, "learning_rate": 6.031624200228715e-07, "loss": 0.3552, "num_input_tokens_seen": 3783262208, "step": 1804 }, { "epoch": 0.16409004610794684, "grad_norm": 13.226753234863281, "learning_rate": 6.027615372971558e-07, "loss": 0.4319, "num_input_tokens_seen": 3785359360, "step": 1805 }, { "epoch": 0.1643612693246542, "grad_norm": 13.992337226867676, "learning_rate": 6.023606121109626e-07, "loss": 0.4704, "num_input_tokens_seen": 3787456512, "step": 1806 }, { "epoch": 0.16463249254136153, "grad_norm": 17.444171905517578, "learning_rate": 6.019596447869413e-07, "loss": 0.4482, "num_input_tokens_seen": 3789553664, "step": 1807 }, { "epoch": 0.1649037157580689, "grad_norm": 14.175294876098633, "learning_rate": 6.015586356477749e-07, "loss": 0.4118, "num_input_tokens_seen": 3791650816, "step": 1808 }, { "epoch": 0.16517493897477625, "grad_norm": 18.26668930053711, "learning_rate": 6.011575850161805e-07, "loss": 0.5852, "num_input_tokens_seen": 3793747968, "step": 1809 }, { "epoch": 0.1654461621914836, "grad_norm": 15.502846717834473, "learning_rate": 6.007564932149086e-07, "loss": 0.5226, "num_input_tokens_seen": 3795845120, "step": 1810 }, { "epoch": 0.16571738540819095, "grad_norm": 13.446937561035156, "learning_rate": 6.003553605667423e-07, "loss": 0.4124, "num_input_tokens_seen": 3797942272, "step": 1811 }, { "epoch": 0.16598860862489828, "grad_norm": 13.046964645385742, "learning_rate": 5.999541873944979e-07, "loss": 0.4353, "num_input_tokens_seen": 3800039424, "step": 1812 }, { "epoch": 0.16625983184160564, "grad_norm": 11.632257461547852, "learning_rate": 5.995529740210244e-07, "loss": 0.3998, "num_input_tokens_seen": 3802136576, "step": 1813 }, { "epoch": 0.166531055058313, "grad_norm": 12.484574317932129, "learning_rate": 5.99151720769203e-07, "loss": 0.318, "num_input_tokens_seen": 3804233728, "step": 1814 }, { "epoch": 0.16680227827502034, "grad_norm": 11.363998413085938, "learning_rate": 5.987504279619473e-07, "loss": 0.4028, "num_input_tokens_seen": 3806330880, "step": 1815 }, { "epoch": 0.1670735014917277, "grad_norm": 18.80327033996582, "learning_rate": 5.98349095922202e-07, "loss": 0.3596, "num_input_tokens_seen": 3808428032, "step": 1816 }, { "epoch": 0.16734472470843503, "grad_norm": 27.840991973876953, "learning_rate": 5.979477249729442e-07, "loss": 1.0105, "num_input_tokens_seen": 3810525184, "step": 1817 }, { "epoch": 0.1676159479251424, "grad_norm": 13.593608856201172, "learning_rate": 5.975463154371822e-07, "loss": 0.5061, "num_input_tokens_seen": 3812622336, "step": 1818 }, { "epoch": 0.16788717114184976, "grad_norm": 14.841533660888672, "learning_rate": 5.971448676379544e-07, "loss": 0.4594, "num_input_tokens_seen": 3814719488, "step": 1819 }, { "epoch": 0.1681583943585571, "grad_norm": 10.198492050170898, "learning_rate": 5.967433818983311e-07, "loss": 0.3427, "num_input_tokens_seen": 3816816640, "step": 1820 }, { "epoch": 0.16842961757526445, "grad_norm": 12.751827239990234, "learning_rate": 5.963418585414129e-07, "loss": 0.3955, "num_input_tokens_seen": 3818913792, "step": 1821 }, { "epoch": 0.16870084079197178, "grad_norm": 11.35025405883789, "learning_rate": 5.959402978903306e-07, "loss": 0.2443, "num_input_tokens_seen": 3821010944, "step": 1822 }, { "epoch": 0.16897206400867915, "grad_norm": 16.358675003051758, "learning_rate": 5.955387002682445e-07, "loss": 0.7063, "num_input_tokens_seen": 3823108096, "step": 1823 }, { "epoch": 0.16924328722538648, "grad_norm": 10.499540328979492, "learning_rate": 5.951370659983452e-07, "loss": 0.3716, "num_input_tokens_seen": 3825205248, "step": 1824 }, { "epoch": 0.16951451044209384, "grad_norm": 9.75558853149414, "learning_rate": 5.94735395403853e-07, "loss": 0.2681, "num_input_tokens_seen": 3827302400, "step": 1825 }, { "epoch": 0.1697857336588012, "grad_norm": 21.415407180786133, "learning_rate": 5.943336888080167e-07, "loss": 0.7333, "num_input_tokens_seen": 3829399552, "step": 1826 }, { "epoch": 0.17005695687550854, "grad_norm": 16.07489585876465, "learning_rate": 5.939319465341148e-07, "loss": 0.5555, "num_input_tokens_seen": 3831496704, "step": 1827 }, { "epoch": 0.1703281800922159, "grad_norm": 16.00339126586914, "learning_rate": 5.93530168905454e-07, "loss": 0.4852, "num_input_tokens_seen": 3833593856, "step": 1828 }, { "epoch": 0.17059940330892323, "grad_norm": 15.944879531860352, "learning_rate": 5.931283562453696e-07, "loss": 0.511, "num_input_tokens_seen": 3835691008, "step": 1829 }, { "epoch": 0.1708706265256306, "grad_norm": 12.453446388244629, "learning_rate": 5.927265088772255e-07, "loss": 0.4054, "num_input_tokens_seen": 3837788160, "step": 1830 }, { "epoch": 0.17114184974233795, "grad_norm": 16.834558486938477, "learning_rate": 5.923246271244127e-07, "loss": 0.6094, "num_input_tokens_seen": 3839885312, "step": 1831 }, { "epoch": 0.1714130729590453, "grad_norm": 20.197141647338867, "learning_rate": 5.919227113103508e-07, "loss": 0.4895, "num_input_tokens_seen": 3841982464, "step": 1832 }, { "epoch": 0.17168429617575265, "grad_norm": 10.21462631225586, "learning_rate": 5.915207617584858e-07, "loss": 0.2921, "num_input_tokens_seen": 3844079616, "step": 1833 }, { "epoch": 0.17195551939245998, "grad_norm": 15.528749465942383, "learning_rate": 5.911187787922918e-07, "loss": 0.5211, "num_input_tokens_seen": 3846176768, "step": 1834 }, { "epoch": 0.17222674260916734, "grad_norm": 13.571805000305176, "learning_rate": 5.907167627352697e-07, "loss": 0.3398, "num_input_tokens_seen": 3848273920, "step": 1835 }, { "epoch": 0.1724979658258747, "grad_norm": 12.700458526611328, "learning_rate": 5.90314713910946e-07, "loss": 0.4671, "num_input_tokens_seen": 3850371072, "step": 1836 }, { "epoch": 0.17276918904258204, "grad_norm": 12.876029968261719, "learning_rate": 5.899126326428746e-07, "loss": 0.4635, "num_input_tokens_seen": 3852468224, "step": 1837 }, { "epoch": 0.1730404122592894, "grad_norm": 19.769418716430664, "learning_rate": 5.895105192546353e-07, "loss": 0.7521, "num_input_tokens_seen": 3854565376, "step": 1838 }, { "epoch": 0.17331163547599673, "grad_norm": 10.773146629333496, "learning_rate": 5.891083740698337e-07, "loss": 0.2714, "num_input_tokens_seen": 3856662528, "step": 1839 }, { "epoch": 0.1735828586927041, "grad_norm": 20.43178939819336, "learning_rate": 5.887061974121007e-07, "loss": 0.4969, "num_input_tokens_seen": 3858759680, "step": 1840 }, { "epoch": 0.17385408190941146, "grad_norm": 16.35590934753418, "learning_rate": 5.883039896050928e-07, "loss": 0.6751, "num_input_tokens_seen": 3860856832, "step": 1841 }, { "epoch": 0.1741253051261188, "grad_norm": 13.783025741577148, "learning_rate": 5.879017509724919e-07, "loss": 0.4219, "num_input_tokens_seen": 3862953984, "step": 1842 }, { "epoch": 0.17439652834282615, "grad_norm": 16.210859298706055, "learning_rate": 5.874994818380039e-07, "loss": 0.5322, "num_input_tokens_seen": 3865051136, "step": 1843 }, { "epoch": 0.17466775155953349, "grad_norm": 14.429267883300781, "learning_rate": 5.8709718252536e-07, "loss": 0.4487, "num_input_tokens_seen": 3867148288, "step": 1844 }, { "epoch": 0.17493897477624085, "grad_norm": 13.477580070495605, "learning_rate": 5.866948533583152e-07, "loss": 0.3541, "num_input_tokens_seen": 3869245440, "step": 1845 }, { "epoch": 0.1752101979929482, "grad_norm": 10.946706771850586, "learning_rate": 5.862924946606487e-07, "loss": 0.247, "num_input_tokens_seen": 3871342592, "step": 1846 }, { "epoch": 0.17548142120965554, "grad_norm": 17.450336456298828, "learning_rate": 5.858901067561637e-07, "loss": 0.5306, "num_input_tokens_seen": 3873439744, "step": 1847 }, { "epoch": 0.1757526444263629, "grad_norm": 18.952194213867188, "learning_rate": 5.854876899686864e-07, "loss": 0.8947, "num_input_tokens_seen": 3875536896, "step": 1848 }, { "epoch": 0.17602386764307024, "grad_norm": 16.47836685180664, "learning_rate": 5.850852446220666e-07, "loss": 0.6694, "num_input_tokens_seen": 3877634048, "step": 1849 }, { "epoch": 0.1762950908597776, "grad_norm": 14.285144805908203, "learning_rate": 5.84682771040177e-07, "loss": 0.3916, "num_input_tokens_seen": 3879731200, "step": 1850 }, { "epoch": 0.17656631407648496, "grad_norm": 11.325536727905273, "learning_rate": 5.842802695469131e-07, "loss": 0.3398, "num_input_tokens_seen": 3881828352, "step": 1851 }, { "epoch": 0.1768375372931923, "grad_norm": 14.129776954650879, "learning_rate": 5.838777404661927e-07, "loss": 0.3623, "num_input_tokens_seen": 3883925504, "step": 1852 }, { "epoch": 0.17710876050989965, "grad_norm": 16.82501792907715, "learning_rate": 5.83475184121956e-07, "loss": 0.707, "num_input_tokens_seen": 3886022656, "step": 1853 }, { "epoch": 0.177379983726607, "grad_norm": 21.06279182434082, "learning_rate": 5.830726008381648e-07, "loss": 0.704, "num_input_tokens_seen": 3888119808, "step": 1854 }, { "epoch": 0.17765120694331435, "grad_norm": 14.380115509033203, "learning_rate": 5.826699909388031e-07, "loss": 0.4179, "num_input_tokens_seen": 3890216960, "step": 1855 }, { "epoch": 0.1779224301600217, "grad_norm": 15.511592864990234, "learning_rate": 5.822673547478757e-07, "loss": 0.652, "num_input_tokens_seen": 3892314112, "step": 1856 }, { "epoch": 0.17819365337672904, "grad_norm": 16.734813690185547, "learning_rate": 5.818646925894092e-07, "loss": 0.6038, "num_input_tokens_seen": 3894411264, "step": 1857 }, { "epoch": 0.1784648765934364, "grad_norm": 9.239023208618164, "learning_rate": 5.814620047874505e-07, "loss": 0.2753, "num_input_tokens_seen": 3896508416, "step": 1858 }, { "epoch": 0.17873609981014374, "grad_norm": 14.336617469787598, "learning_rate": 5.810592916660677e-07, "loss": 0.4859, "num_input_tokens_seen": 3898605568, "step": 1859 }, { "epoch": 0.1790073230268511, "grad_norm": 13.082498550415039, "learning_rate": 5.806565535493489e-07, "loss": 0.3663, "num_input_tokens_seen": 3900702720, "step": 1860 }, { "epoch": 0.17927854624355846, "grad_norm": 12.747309684753418, "learning_rate": 5.802537907614023e-07, "loss": 0.3627, "num_input_tokens_seen": 3902799872, "step": 1861 }, { "epoch": 0.1795497694602658, "grad_norm": 19.392642974853516, "learning_rate": 5.798510036263561e-07, "loss": 0.8228, "num_input_tokens_seen": 3904897024, "step": 1862 }, { "epoch": 0.17982099267697316, "grad_norm": 14.767632484436035, "learning_rate": 5.794481924683581e-07, "loss": 0.5703, "num_input_tokens_seen": 3906994176, "step": 1863 }, { "epoch": 0.1800922158936805, "grad_norm": 17.085071563720703, "learning_rate": 5.790453576115756e-07, "loss": 0.4947, "num_input_tokens_seen": 3909091328, "step": 1864 }, { "epoch": 0.18036343911038785, "grad_norm": 10.42067813873291, "learning_rate": 5.786424993801942e-07, "loss": 0.2922, "num_input_tokens_seen": 3911188480, "step": 1865 }, { "epoch": 0.18063466232709519, "grad_norm": 12.602887153625488, "learning_rate": 5.782396180984194e-07, "loss": 0.353, "num_input_tokens_seen": 3913285632, "step": 1866 }, { "epoch": 0.18090588554380255, "grad_norm": 13.061318397521973, "learning_rate": 5.778367140904746e-07, "loss": 0.4512, "num_input_tokens_seen": 3915382784, "step": 1867 }, { "epoch": 0.1811771087605099, "grad_norm": 18.99317741394043, "learning_rate": 5.774337876806016e-07, "loss": 0.6254, "num_input_tokens_seen": 3917479936, "step": 1868 }, { "epoch": 0.18144833197721724, "grad_norm": 16.320486068725586, "learning_rate": 5.770308391930601e-07, "loss": 0.5455, "num_input_tokens_seen": 3919577088, "step": 1869 }, { "epoch": 0.1817195551939246, "grad_norm": 14.826481819152832, "learning_rate": 5.766278689521278e-07, "loss": 0.4508, "num_input_tokens_seen": 3921674240, "step": 1870 }, { "epoch": 0.18199077841063194, "grad_norm": 16.660938262939453, "learning_rate": 5.762248772820999e-07, "loss": 0.3395, "num_input_tokens_seen": 3923771392, "step": 1871 }, { "epoch": 0.1822620016273393, "grad_norm": 11.43323040008545, "learning_rate": 5.758218645072887e-07, "loss": 0.3886, "num_input_tokens_seen": 3925868544, "step": 1872 }, { "epoch": 0.18253322484404666, "grad_norm": 19.8907470703125, "learning_rate": 5.754188309520235e-07, "loss": 0.6173, "num_input_tokens_seen": 3927965696, "step": 1873 }, { "epoch": 0.182804448060754, "grad_norm": 26.984996795654297, "learning_rate": 5.750157769406504e-07, "loss": 0.6616, "num_input_tokens_seen": 3930062848, "step": 1874 }, { "epoch": 0.18307567127746135, "grad_norm": 15.848214149475098, "learning_rate": 5.74612702797532e-07, "loss": 0.7792, "num_input_tokens_seen": 3932160000, "step": 1875 }, { "epoch": 0.1833468944941687, "grad_norm": 17.739269256591797, "learning_rate": 5.74209608847047e-07, "loss": 0.6638, "num_input_tokens_seen": 3934257152, "step": 1876 }, { "epoch": 0.18361811771087605, "grad_norm": 21.376651763916016, "learning_rate": 5.7380649541359e-07, "loss": 0.7258, "num_input_tokens_seen": 3936354304, "step": 1877 }, { "epoch": 0.1838893409275834, "grad_norm": 11.857672691345215, "learning_rate": 5.734033628215714e-07, "loss": 0.4125, "num_input_tokens_seen": 3938451456, "step": 1878 }, { "epoch": 0.18416056414429074, "grad_norm": 17.930675506591797, "learning_rate": 5.730002113954169e-07, "loss": 0.6946, "num_input_tokens_seen": 3940548608, "step": 1879 }, { "epoch": 0.1844317873609981, "grad_norm": 18.330848693847656, "learning_rate": 5.725970414595675e-07, "loss": 0.6827, "num_input_tokens_seen": 3942645760, "step": 1880 }, { "epoch": 0.18470301057770544, "grad_norm": 16.09734535217285, "learning_rate": 5.721938533384791e-07, "loss": 0.6227, "num_input_tokens_seen": 3944742912, "step": 1881 }, { "epoch": 0.1849742337944128, "grad_norm": 14.484498023986816, "learning_rate": 5.717906473566219e-07, "loss": 0.4096, "num_input_tokens_seen": 3946840064, "step": 1882 }, { "epoch": 0.18524545701112016, "grad_norm": 21.29253578186035, "learning_rate": 5.71387423838481e-07, "loss": 0.658, "num_input_tokens_seen": 3948937216, "step": 1883 }, { "epoch": 0.1855166802278275, "grad_norm": 12.313188552856445, "learning_rate": 5.709841831085553e-07, "loss": 0.3509, "num_input_tokens_seen": 3951034368, "step": 1884 }, { "epoch": 0.18578790344453486, "grad_norm": 12.124674797058105, "learning_rate": 5.705809254913576e-07, "loss": 0.3977, "num_input_tokens_seen": 3953131520, "step": 1885 }, { "epoch": 0.1860591266612422, "grad_norm": 19.086606979370117, "learning_rate": 5.70177651311414e-07, "loss": 0.7218, "num_input_tokens_seen": 3955228672, "step": 1886 }, { "epoch": 0.18633034987794955, "grad_norm": 11.468111991882324, "learning_rate": 5.697743608932646e-07, "loss": 0.3488, "num_input_tokens_seen": 3957325824, "step": 1887 }, { "epoch": 0.18660157309465691, "grad_norm": 22.346908569335938, "learning_rate": 5.693710545614621e-07, "loss": 0.9688, "num_input_tokens_seen": 3959422976, "step": 1888 }, { "epoch": 0.18687279631136425, "grad_norm": 18.068498611450195, "learning_rate": 5.689677326405719e-07, "loss": 0.7224, "num_input_tokens_seen": 3961520128, "step": 1889 }, { "epoch": 0.1871440195280716, "grad_norm": 21.589921951293945, "learning_rate": 5.685643954551722e-07, "loss": 0.8734, "num_input_tokens_seen": 3963617280, "step": 1890 }, { "epoch": 0.18741524274477894, "grad_norm": 15.714263916015625, "learning_rate": 5.681610433298535e-07, "loss": 0.5104, "num_input_tokens_seen": 3965714432, "step": 1891 }, { "epoch": 0.1876864659614863, "grad_norm": 12.886143684387207, "learning_rate": 5.677576765892182e-07, "loss": 0.4624, "num_input_tokens_seen": 3967811584, "step": 1892 }, { "epoch": 0.18795768917819367, "grad_norm": 20.087732315063477, "learning_rate": 5.673542955578806e-07, "loss": 0.4943, "num_input_tokens_seen": 3969908736, "step": 1893 }, { "epoch": 0.188228912394901, "grad_norm": 12.997749328613281, "learning_rate": 5.669509005604663e-07, "loss": 0.3788, "num_input_tokens_seen": 3972005888, "step": 1894 }, { "epoch": 0.18850013561160836, "grad_norm": 18.928436279296875, "learning_rate": 5.665474919216122e-07, "loss": 0.7576, "num_input_tokens_seen": 3974103040, "step": 1895 }, { "epoch": 0.1887713588283157, "grad_norm": 12.904560089111328, "learning_rate": 5.661440699659663e-07, "loss": 0.4746, "num_input_tokens_seen": 3976200192, "step": 1896 }, { "epoch": 0.18904258204502306, "grad_norm": 13.534509658813477, "learning_rate": 5.657406350181872e-07, "loss": 0.4028, "num_input_tokens_seen": 3978297344, "step": 1897 }, { "epoch": 0.18931380526173042, "grad_norm": 18.946544647216797, "learning_rate": 5.65337187402944e-07, "loss": 0.7937, "num_input_tokens_seen": 3980394496, "step": 1898 }, { "epoch": 0.18958502847843775, "grad_norm": 21.76073455810547, "learning_rate": 5.64933727444916e-07, "loss": 0.5819, "num_input_tokens_seen": 3982491648, "step": 1899 }, { "epoch": 0.1898562516951451, "grad_norm": 13.195499420166016, "learning_rate": 5.645302554687925e-07, "loss": 0.4144, "num_input_tokens_seen": 3984588800, "step": 1900 }, { "epoch": 0.19012747491185245, "grad_norm": 17.136215209960938, "learning_rate": 5.641267717992723e-07, "loss": 0.4328, "num_input_tokens_seen": 3986685952, "step": 1901 }, { "epoch": 0.1903986981285598, "grad_norm": 13.509635925292969, "learning_rate": 5.637232767610637e-07, "loss": 0.4364, "num_input_tokens_seen": 3988783104, "step": 1902 }, { "epoch": 0.19066992134526717, "grad_norm": 16.542097091674805, "learning_rate": 5.63319770678884e-07, "loss": 0.4641, "num_input_tokens_seen": 3990880256, "step": 1903 }, { "epoch": 0.1909411445619745, "grad_norm": 13.608541488647461, "learning_rate": 5.6291625387746e-07, "loss": 0.403, "num_input_tokens_seen": 3992977408, "step": 1904 }, { "epoch": 0.19121236777868186, "grad_norm": 15.277917861938477, "learning_rate": 5.625127266815263e-07, "loss": 0.3942, "num_input_tokens_seen": 3995074560, "step": 1905 }, { "epoch": 0.1914835909953892, "grad_norm": 20.041263580322266, "learning_rate": 5.621091894158261e-07, "loss": 0.3556, "num_input_tokens_seen": 3997171712, "step": 1906 }, { "epoch": 0.19175481421209656, "grad_norm": 17.10890769958496, "learning_rate": 5.617056424051113e-07, "loss": 0.5611, "num_input_tokens_seen": 3999268864, "step": 1907 }, { "epoch": 0.1920260374288039, "grad_norm": 19.5421199798584, "learning_rate": 5.613020859741408e-07, "loss": 0.5567, "num_input_tokens_seen": 4001366016, "step": 1908 }, { "epoch": 0.19229726064551125, "grad_norm": 20.52210235595703, "learning_rate": 5.608985204476817e-07, "loss": 0.7141, "num_input_tokens_seen": 4003463168, "step": 1909 }, { "epoch": 0.19256848386221861, "grad_norm": 26.15365982055664, "learning_rate": 5.60494946150508e-07, "loss": 0.5505, "num_input_tokens_seen": 4005560320, "step": 1910 }, { "epoch": 0.19283970707892595, "grad_norm": 14.202539443969727, "learning_rate": 5.600913634074009e-07, "loss": 0.4245, "num_input_tokens_seen": 4007657472, "step": 1911 }, { "epoch": 0.1931109302956333, "grad_norm": 17.88532257080078, "learning_rate": 5.596877725431487e-07, "loss": 0.4943, "num_input_tokens_seen": 4009754624, "step": 1912 }, { "epoch": 0.19338215351234064, "grad_norm": 10.894256591796875, "learning_rate": 5.592841738825457e-07, "loss": 0.3248, "num_input_tokens_seen": 4011851776, "step": 1913 }, { "epoch": 0.193653376729048, "grad_norm": 14.625932693481445, "learning_rate": 5.588805677503928e-07, "loss": 0.4324, "num_input_tokens_seen": 4013948928, "step": 1914 }, { "epoch": 0.19392459994575537, "grad_norm": 12.210338592529297, "learning_rate": 5.58476954471497e-07, "loss": 0.4082, "num_input_tokens_seen": 4016046080, "step": 1915 }, { "epoch": 0.1941958231624627, "grad_norm": 14.13292407989502, "learning_rate": 5.580733343706708e-07, "loss": 0.6738, "num_input_tokens_seen": 4018143232, "step": 1916 }, { "epoch": 0.19446704637917006, "grad_norm": 17.628400802612305, "learning_rate": 5.576697077727323e-07, "loss": 0.529, "num_input_tokens_seen": 4020240384, "step": 1917 }, { "epoch": 0.1947382695958774, "grad_norm": 16.035490036010742, "learning_rate": 5.572660750025049e-07, "loss": 0.4977, "num_input_tokens_seen": 4022337536, "step": 1918 }, { "epoch": 0.19500949281258476, "grad_norm": 12.913204193115234, "learning_rate": 5.568624363848166e-07, "loss": 0.3954, "num_input_tokens_seen": 4024434688, "step": 1919 }, { "epoch": 0.19528071602929212, "grad_norm": 10.610041618347168, "learning_rate": 5.564587922445008e-07, "loss": 0.3049, "num_input_tokens_seen": 4026531840, "step": 1920 }, { "epoch": 0.19555193924599945, "grad_norm": 12.461469650268555, "learning_rate": 5.560551429063949e-07, "loss": 0.3567, "num_input_tokens_seen": 4028628992, "step": 1921 }, { "epoch": 0.1958231624627068, "grad_norm": 19.98131561279297, "learning_rate": 5.556514886953403e-07, "loss": 0.7164, "num_input_tokens_seen": 4030726144, "step": 1922 }, { "epoch": 0.19609438567941415, "grad_norm": 14.228310585021973, "learning_rate": 5.552478299361826e-07, "loss": 0.4572, "num_input_tokens_seen": 4032823296, "step": 1923 }, { "epoch": 0.1963656088961215, "grad_norm": 14.937055587768555, "learning_rate": 5.548441669537712e-07, "loss": 0.6819, "num_input_tokens_seen": 4034920448, "step": 1924 }, { "epoch": 0.19663683211282887, "grad_norm": 18.690763473510742, "learning_rate": 5.544405000729584e-07, "loss": 0.6, "num_input_tokens_seen": 4037017600, "step": 1925 }, { "epoch": 0.1969080553295362, "grad_norm": 18.642099380493164, "learning_rate": 5.540368296186002e-07, "loss": 0.4619, "num_input_tokens_seen": 4039114752, "step": 1926 }, { "epoch": 0.19717927854624356, "grad_norm": 11.240645408630371, "learning_rate": 5.53633155915555e-07, "loss": 0.3568, "num_input_tokens_seen": 4041211904, "step": 1927 }, { "epoch": 0.1974505017629509, "grad_norm": 19.68428611755371, "learning_rate": 5.532294792886843e-07, "loss": 0.8633, "num_input_tokens_seen": 4043309056, "step": 1928 }, { "epoch": 0.19772172497965826, "grad_norm": 15.801280975341797, "learning_rate": 5.528258000628518e-07, "loss": 0.4886, "num_input_tokens_seen": 4045406208, "step": 1929 }, { "epoch": 0.19799294819636562, "grad_norm": 14.434005737304688, "learning_rate": 5.524221185629224e-07, "loss": 0.5922, "num_input_tokens_seen": 4047503360, "step": 1930 }, { "epoch": 0.19826417141307295, "grad_norm": 12.741336822509766, "learning_rate": 5.520184351137646e-07, "loss": 0.4185, "num_input_tokens_seen": 4049600512, "step": 1931 }, { "epoch": 0.19853539462978032, "grad_norm": 11.868021965026855, "learning_rate": 5.51614750040247e-07, "loss": 0.3508, "num_input_tokens_seen": 4051697664, "step": 1932 }, { "epoch": 0.19880661784648765, "grad_norm": 18.497217178344727, "learning_rate": 5.512110636672396e-07, "loss": 0.5296, "num_input_tokens_seen": 4053794816, "step": 1933 }, { "epoch": 0.199077841063195, "grad_norm": 16.076515197753906, "learning_rate": 5.508073763196146e-07, "loss": 0.3934, "num_input_tokens_seen": 4055891968, "step": 1934 }, { "epoch": 0.19934906427990237, "grad_norm": 14.377509117126465, "learning_rate": 5.504036883222438e-07, "loss": 0.4811, "num_input_tokens_seen": 4057989120, "step": 1935 }, { "epoch": 0.1996202874966097, "grad_norm": 20.239578247070312, "learning_rate": 5.5e-07, "loss": 0.7046, "num_input_tokens_seen": 4060086272, "step": 1936 }, { "epoch": 0.19989151071331707, "grad_norm": 15.798903465270996, "learning_rate": 5.495963116777562e-07, "loss": 0.3267, "num_input_tokens_seen": 4062183424, "step": 1937 }, { "epoch": 0.2001627339300244, "grad_norm": 17.59996795654297, "learning_rate": 5.491926236803854e-07, "loss": 0.726, "num_input_tokens_seen": 4064280576, "step": 1938 }, { "epoch": 0.20043395714673176, "grad_norm": 14.18375301361084, "learning_rate": 5.487889363327603e-07, "loss": 0.5171, "num_input_tokens_seen": 4066377728, "step": 1939 }, { "epoch": 0.20070518036343912, "grad_norm": 12.889824867248535, "learning_rate": 5.483852499597532e-07, "loss": 0.3304, "num_input_tokens_seen": 4068474880, "step": 1940 }, { "epoch": 0.20097640358014646, "grad_norm": 18.258941650390625, "learning_rate": 5.479815648862355e-07, "loss": 0.6463, "num_input_tokens_seen": 4070572032, "step": 1941 }, { "epoch": 0.20124762679685382, "grad_norm": 9.00694465637207, "learning_rate": 5.475778814370776e-07, "loss": 0.2284, "num_input_tokens_seen": 4072669184, "step": 1942 }, { "epoch": 0.20151885001356115, "grad_norm": 11.947250366210938, "learning_rate": 5.471741999371483e-07, "loss": 0.3377, "num_input_tokens_seen": 4074766336, "step": 1943 }, { "epoch": 0.2017900732302685, "grad_norm": 14.319008827209473, "learning_rate": 5.467705207113156e-07, "loss": 0.3758, "num_input_tokens_seen": 4076863488, "step": 1944 }, { "epoch": 0.20206129644697587, "grad_norm": 15.36997127532959, "learning_rate": 5.46366844084445e-07, "loss": 0.5119, "num_input_tokens_seen": 4078960640, "step": 1945 }, { "epoch": 0.2023325196636832, "grad_norm": 16.01044273376465, "learning_rate": 5.459631703813998e-07, "loss": 0.4673, "num_input_tokens_seen": 4081057792, "step": 1946 }, { "epoch": 0.20260374288039057, "grad_norm": 12.556026458740234, "learning_rate": 5.455594999270416e-07, "loss": 0.3059, "num_input_tokens_seen": 4083154944, "step": 1947 }, { "epoch": 0.2028749660970979, "grad_norm": 17.68223762512207, "learning_rate": 5.451558330462289e-07, "loss": 0.6006, "num_input_tokens_seen": 4085252096, "step": 1948 }, { "epoch": 0.20314618931380526, "grad_norm": 15.037524223327637, "learning_rate": 5.447521700638174e-07, "loss": 0.4331, "num_input_tokens_seen": 4087349248, "step": 1949 }, { "epoch": 0.20341741253051263, "grad_norm": 13.928112030029297, "learning_rate": 5.443485113046597e-07, "loss": 0.4327, "num_input_tokens_seen": 4089446400, "step": 1950 }, { "epoch": 0.20368863574721996, "grad_norm": 11.044869422912598, "learning_rate": 5.439448570936052e-07, "loss": 0.2896, "num_input_tokens_seen": 4091543552, "step": 1951 }, { "epoch": 0.20395985896392732, "grad_norm": 14.348036766052246, "learning_rate": 5.435412077554991e-07, "loss": 0.3981, "num_input_tokens_seen": 4093640704, "step": 1952 }, { "epoch": 0.20423108218063465, "grad_norm": 14.003972053527832, "learning_rate": 5.431375636151833e-07, "loss": 0.3766, "num_input_tokens_seen": 4095737856, "step": 1953 }, { "epoch": 0.20450230539734202, "grad_norm": 18.595943450927734, "learning_rate": 5.427339249974952e-07, "loss": 0.7953, "num_input_tokens_seen": 4097835008, "step": 1954 }, { "epoch": 0.20477352861404935, "grad_norm": 12.756117820739746, "learning_rate": 5.423302922272677e-07, "loss": 0.4171, "num_input_tokens_seen": 4099932160, "step": 1955 }, { "epoch": 0.2050447518307567, "grad_norm": 17.62868881225586, "learning_rate": 5.419266656293293e-07, "loss": 0.5959, "num_input_tokens_seen": 4102029312, "step": 1956 }, { "epoch": 0.20531597504746407, "grad_norm": 14.314393997192383, "learning_rate": 5.415230455285031e-07, "loss": 0.4849, "num_input_tokens_seen": 4104126464, "step": 1957 }, { "epoch": 0.2055871982641714, "grad_norm": 11.385714530944824, "learning_rate": 5.411194322496073e-07, "loss": 0.341, "num_input_tokens_seen": 4106223616, "step": 1958 }, { "epoch": 0.20585842148087877, "grad_norm": 23.866220474243164, "learning_rate": 5.407158261174543e-07, "loss": 0.3017, "num_input_tokens_seen": 4108320768, "step": 1959 }, { "epoch": 0.2061296446975861, "grad_norm": 12.497730255126953, "learning_rate": 5.403122274568514e-07, "loss": 0.2759, "num_input_tokens_seen": 4110417920, "step": 1960 }, { "epoch": 0.20640086791429346, "grad_norm": 35.96780776977539, "learning_rate": 5.399086365925991e-07, "loss": 0.9278, "num_input_tokens_seen": 4112515072, "step": 1961 }, { "epoch": 0.20667209113100082, "grad_norm": 13.677523612976074, "learning_rate": 5.395050538494922e-07, "loss": 0.2937, "num_input_tokens_seen": 4114612224, "step": 1962 }, { "epoch": 0.20694331434770816, "grad_norm": 9.891523361206055, "learning_rate": 5.391014795523184e-07, "loss": 0.2558, "num_input_tokens_seen": 4116709376, "step": 1963 }, { "epoch": 0.20721453756441552, "grad_norm": 15.449033737182617, "learning_rate": 5.386979140258592e-07, "loss": 0.5304, "num_input_tokens_seen": 4118806528, "step": 1964 }, { "epoch": 0.20748576078112285, "grad_norm": 12.77426815032959, "learning_rate": 5.382943575948888e-07, "loss": 0.3884, "num_input_tokens_seen": 4120903680, "step": 1965 }, { "epoch": 0.2077569839978302, "grad_norm": 27.787797927856445, "learning_rate": 5.378908105841738e-07, "loss": 0.5596, "num_input_tokens_seen": 4123000832, "step": 1966 }, { "epoch": 0.20802820721453757, "grad_norm": 17.83061408996582, "learning_rate": 5.374872733184737e-07, "loss": 0.5323, "num_input_tokens_seen": 4125097984, "step": 1967 }, { "epoch": 0.2082994304312449, "grad_norm": 16.810670852661133, "learning_rate": 5.370837461225402e-07, "loss": 0.6982, "num_input_tokens_seen": 4127195136, "step": 1968 }, { "epoch": 0.20857065364795227, "grad_norm": 16.809240341186523, "learning_rate": 5.36680229321116e-07, "loss": 0.4564, "num_input_tokens_seen": 4129292288, "step": 1969 }, { "epoch": 0.2088418768646596, "grad_norm": 16.903656005859375, "learning_rate": 5.362767232389365e-07, "loss": 0.5688, "num_input_tokens_seen": 4131389440, "step": 1970 }, { "epoch": 0.20911310008136696, "grad_norm": 11.472921371459961, "learning_rate": 5.358732282007278e-07, "loss": 0.3912, "num_input_tokens_seen": 4133486592, "step": 1971 }, { "epoch": 0.20938432329807433, "grad_norm": 15.253931999206543, "learning_rate": 5.354697445312074e-07, "loss": 0.5295, "num_input_tokens_seen": 4135583744, "step": 1972 }, { "epoch": 0.20965554651478166, "grad_norm": 14.575647354125977, "learning_rate": 5.35066272555084e-07, "loss": 0.6172, "num_input_tokens_seen": 4137680896, "step": 1973 }, { "epoch": 0.20992676973148902, "grad_norm": 15.925987243652344, "learning_rate": 5.346628125970562e-07, "loss": 0.4215, "num_input_tokens_seen": 4139778048, "step": 1974 }, { "epoch": 0.21019799294819635, "grad_norm": 21.621906280517578, "learning_rate": 5.342593649818129e-07, "loss": 0.6736, "num_input_tokens_seen": 4141875200, "step": 1975 }, { "epoch": 0.21046921616490372, "grad_norm": 13.804909706115723, "learning_rate": 5.338559300340338e-07, "loss": 0.4555, "num_input_tokens_seen": 4143972352, "step": 1976 }, { "epoch": 0.21074043938161108, "grad_norm": 15.145031929016113, "learning_rate": 5.334525080783879e-07, "loss": 0.485, "num_input_tokens_seen": 4146069504, "step": 1977 }, { "epoch": 0.2110116625983184, "grad_norm": 12.25986099243164, "learning_rate": 5.330490994395338e-07, "loss": 0.3228, "num_input_tokens_seen": 4148166656, "step": 1978 }, { "epoch": 0.21128288581502577, "grad_norm": 15.601759910583496, "learning_rate": 5.326457044421195e-07, "loss": 0.5354, "num_input_tokens_seen": 4150263808, "step": 1979 }, { "epoch": 0.2115541090317331, "grad_norm": 12.419604301452637, "learning_rate": 5.322423234107818e-07, "loss": 0.4416, "num_input_tokens_seen": 4152360960, "step": 1980 }, { "epoch": 0.21182533224844047, "grad_norm": 17.01136589050293, "learning_rate": 5.318389566701465e-07, "loss": 0.5834, "num_input_tokens_seen": 4154458112, "step": 1981 }, { "epoch": 0.21209655546514783, "grad_norm": 13.978080749511719, "learning_rate": 5.314356045448278e-07, "loss": 0.388, "num_input_tokens_seen": 4156555264, "step": 1982 }, { "epoch": 0.21236777868185516, "grad_norm": 18.458215713500977, "learning_rate": 5.310322673594282e-07, "loss": 0.4498, "num_input_tokens_seen": 4158652416, "step": 1983 }, { "epoch": 0.21263900189856252, "grad_norm": 13.268664360046387, "learning_rate": 5.306289454385379e-07, "loss": 0.2379, "num_input_tokens_seen": 4160749568, "step": 1984 }, { "epoch": 0.21291022511526986, "grad_norm": 14.612197875976562, "learning_rate": 5.302256391067354e-07, "loss": 0.5217, "num_input_tokens_seen": 4162846720, "step": 1985 }, { "epoch": 0.21318144833197722, "grad_norm": 14.71393871307373, "learning_rate": 5.29822348688586e-07, "loss": 0.4559, "num_input_tokens_seen": 4164943872, "step": 1986 }, { "epoch": 0.21345267154868458, "grad_norm": 18.54011344909668, "learning_rate": 5.294190745086426e-07, "loss": 0.5078, "num_input_tokens_seen": 4167041024, "step": 1987 }, { "epoch": 0.21372389476539191, "grad_norm": 22.229833602905273, "learning_rate": 5.290158168914447e-07, "loss": 0.6627, "num_input_tokens_seen": 4169138176, "step": 1988 }, { "epoch": 0.21399511798209928, "grad_norm": 15.733107566833496, "learning_rate": 5.28612576161519e-07, "loss": 0.6356, "num_input_tokens_seen": 4171235328, "step": 1989 }, { "epoch": 0.2142663411988066, "grad_norm": 16.396329879760742, "learning_rate": 5.282093526433781e-07, "loss": 0.5705, "num_input_tokens_seen": 4173332480, "step": 1990 }, { "epoch": 0.21453756441551397, "grad_norm": 10.583121299743652, "learning_rate": 5.27806146661521e-07, "loss": 0.2994, "num_input_tokens_seen": 4175429632, "step": 1991 }, { "epoch": 0.21480878763222133, "grad_norm": 14.593506813049316, "learning_rate": 5.274029585404326e-07, "loss": 0.3533, "num_input_tokens_seen": 4177526784, "step": 1992 }, { "epoch": 0.21508001084892867, "grad_norm": 12.540976524353027, "learning_rate": 5.269997886045833e-07, "loss": 0.4252, "num_input_tokens_seen": 4179623936, "step": 1993 }, { "epoch": 0.21535123406563603, "grad_norm": 15.491146087646484, "learning_rate": 5.265966371784287e-07, "loss": 0.5759, "num_input_tokens_seen": 4181721088, "step": 1994 }, { "epoch": 0.21562245728234336, "grad_norm": 13.454513549804688, "learning_rate": 5.261935045864101e-07, "loss": 0.3348, "num_input_tokens_seen": 4183818240, "step": 1995 }, { "epoch": 0.21589368049905072, "grad_norm": 13.391287803649902, "learning_rate": 5.257903911529532e-07, "loss": 0.416, "num_input_tokens_seen": 4185915392, "step": 1996 }, { "epoch": 0.21616490371575806, "grad_norm": 11.039448738098145, "learning_rate": 5.253872972024681e-07, "loss": 0.2622, "num_input_tokens_seen": 4188012544, "step": 1997 }, { "epoch": 0.21643612693246542, "grad_norm": 12.495340347290039, "learning_rate": 5.249842230593497e-07, "loss": 0.3805, "num_input_tokens_seen": 4190109696, "step": 1998 }, { "epoch": 0.21670735014917278, "grad_norm": 18.3435001373291, "learning_rate": 5.245811690479766e-07, "loss": 0.4074, "num_input_tokens_seen": 4192206848, "step": 1999 }, { "epoch": 0.2169785733658801, "grad_norm": 13.659356117248535, "learning_rate": 5.241781354927113e-07, "loss": 0.4192, "num_input_tokens_seen": 4194304000, "step": 2000 }, { "epoch": 0.00027122321670735016, "grad_norm": 23.389102935791016, "learning_rate": 5.237751227179001e-07, "loss": 0.5657, "num_input_tokens_seen": 4196401152, "step": 2001 }, { "epoch": 0.0005424464334147003, "grad_norm": 17.303232192993164, "learning_rate": 5.233721310478722e-07, "loss": 0.5492, "num_input_tokens_seen": 4198498304, "step": 2002 }, { "epoch": 0.0008136696501220504, "grad_norm": 9.405117988586426, "learning_rate": 5.2296916080694e-07, "loss": 0.2159, "num_input_tokens_seen": 4200595456, "step": 2003 }, { "epoch": 0.0010848928668294006, "grad_norm": 17.866546630859375, "learning_rate": 5.225662123193985e-07, "loss": 0.4617, "num_input_tokens_seen": 4202692608, "step": 2004 }, { "epoch": 0.0013561160835367507, "grad_norm": 17.711912155151367, "learning_rate": 5.221632859095254e-07, "loss": 0.375, "num_input_tokens_seen": 4204789760, "step": 2005 }, { "epoch": 0.0016273393002441008, "grad_norm": 17.354616165161133, "learning_rate": 5.217603819015805e-07, "loss": 0.5621, "num_input_tokens_seen": 4206886912, "step": 2006 }, { "epoch": 0.001898562516951451, "grad_norm": 18.786190032958984, "learning_rate": 5.213575006198057e-07, "loss": 0.7147, "num_input_tokens_seen": 4208984064, "step": 2007 }, { "epoch": 0.0021697857336588013, "grad_norm": 16.878746032714844, "learning_rate": 5.209546423884246e-07, "loss": 0.5038, "num_input_tokens_seen": 4211081216, "step": 2008 }, { "epoch": 0.0024410089503661514, "grad_norm": 13.854147911071777, "learning_rate": 5.20551807531642e-07, "loss": 0.3361, "num_input_tokens_seen": 4213178368, "step": 2009 }, { "epoch": 0.0027122321670735015, "grad_norm": 16.524831771850586, "learning_rate": 5.20148996373644e-07, "loss": 0.5854, "num_input_tokens_seen": 4215275520, "step": 2010 }, { "epoch": 0.0029834553837808516, "grad_norm": 11.64826488494873, "learning_rate": 5.197462092385978e-07, "loss": 0.3409, "num_input_tokens_seen": 4217372672, "step": 2011 }, { "epoch": 0.0032546786004882017, "grad_norm": 9.127813339233398, "learning_rate": 5.193434464506511e-07, "loss": 0.2305, "num_input_tokens_seen": 4219469824, "step": 2012 }, { "epoch": 0.003525901817195552, "grad_norm": 15.029864311218262, "learning_rate": 5.189407083339324e-07, "loss": 0.4268, "num_input_tokens_seen": 4221566976, "step": 2013 }, { "epoch": 0.003797125033902902, "grad_norm": 9.265511512756348, "learning_rate": 5.185379952125494e-07, "loss": 0.2192, "num_input_tokens_seen": 4223664128, "step": 2014 }, { "epoch": 0.0040683482506102524, "grad_norm": 15.287160873413086, "learning_rate": 5.18135307410591e-07, "loss": 0.4158, "num_input_tokens_seen": 4225761280, "step": 2015 }, { "epoch": 0.0043395714673176026, "grad_norm": 20.016700744628906, "learning_rate": 5.177326452521242e-07, "loss": 0.3771, "num_input_tokens_seen": 4227858432, "step": 2016 }, { "epoch": 0.004610794684024953, "grad_norm": 19.859949111938477, "learning_rate": 5.17330009061197e-07, "loss": 0.467, "num_input_tokens_seen": 4229955584, "step": 2017 }, { "epoch": 0.004882017900732303, "grad_norm": 16.03641700744629, "learning_rate": 5.169273991618351e-07, "loss": 0.43, "num_input_tokens_seen": 4232052736, "step": 2018 }, { "epoch": 0.005153241117439653, "grad_norm": 19.55535316467285, "learning_rate": 5.165248158780441e-07, "loss": 0.4449, "num_input_tokens_seen": 4234149888, "step": 2019 }, { "epoch": 0.005424464334147003, "grad_norm": 19.658199310302734, "learning_rate": 5.161222595338073e-07, "loss": 0.842, "num_input_tokens_seen": 4236247040, "step": 2020 }, { "epoch": 0.005695687550854353, "grad_norm": 14.803008079528809, "learning_rate": 5.157197304530869e-07, "loss": 0.4334, "num_input_tokens_seen": 4238344192, "step": 2021 }, { "epoch": 0.005966910767561703, "grad_norm": 16.178064346313477, "learning_rate": 5.15317228959823e-07, "loss": 0.4884, "num_input_tokens_seen": 4240441344, "step": 2022 }, { "epoch": 0.006238133984269053, "grad_norm": 20.011199951171875, "learning_rate": 5.149147553779335e-07, "loss": 0.7486, "num_input_tokens_seen": 4242538496, "step": 2023 }, { "epoch": 0.006509357200976403, "grad_norm": 12.02757740020752, "learning_rate": 5.145123100313137e-07, "loss": 0.3126, "num_input_tokens_seen": 4244635648, "step": 2024 }, { "epoch": 0.0067805804176837535, "grad_norm": 18.81563377380371, "learning_rate": 5.141098932438365e-07, "loss": 0.4878, "num_input_tokens_seen": 4246732800, "step": 2025 }, { "epoch": 0.007051803634391104, "grad_norm": 19.94472312927246, "learning_rate": 5.137075053393512e-07, "loss": 0.5076, "num_input_tokens_seen": 4248829952, "step": 2026 }, { "epoch": 0.007323026851098454, "grad_norm": 23.192970275878906, "learning_rate": 5.133051466416849e-07, "loss": 0.7827, "num_input_tokens_seen": 4250927104, "step": 2027 }, { "epoch": 0.007594250067805804, "grad_norm": 14.543439865112305, "learning_rate": 5.129028174746399e-07, "loss": 0.3696, "num_input_tokens_seen": 4253024256, "step": 2028 }, { "epoch": 0.007865473284513154, "grad_norm": 21.16613006591797, "learning_rate": 5.12500518161996e-07, "loss": 1.0285, "num_input_tokens_seen": 4255121408, "step": 2029 }, { "epoch": 0.008136696501220505, "grad_norm": 15.31861686706543, "learning_rate": 5.120982490275081e-07, "loss": 0.468, "num_input_tokens_seen": 4257218560, "step": 2030 }, { "epoch": 0.008407919717927854, "grad_norm": 17.255081176757812, "learning_rate": 5.116960103949071e-07, "loss": 0.7309, "num_input_tokens_seen": 4259315712, "step": 2031 }, { "epoch": 0.008679142934635205, "grad_norm": 15.113636016845703, "learning_rate": 5.112938025878994e-07, "loss": 0.3611, "num_input_tokens_seen": 4261412864, "step": 2032 }, { "epoch": 0.008950366151342554, "grad_norm": 9.508875846862793, "learning_rate": 5.108916259301663e-07, "loss": 0.2398, "num_input_tokens_seen": 4263510016, "step": 2033 }, { "epoch": 0.009221589368049905, "grad_norm": 10.597322463989258, "learning_rate": 5.104894807453647e-07, "loss": 0.3246, "num_input_tokens_seen": 4265607168, "step": 2034 }, { "epoch": 0.009492812584757255, "grad_norm": 17.956418991088867, "learning_rate": 5.100873673571253e-07, "loss": 0.5213, "num_input_tokens_seen": 4267704320, "step": 2035 }, { "epoch": 0.009764035801464606, "grad_norm": 16.9884090423584, "learning_rate": 5.096852860890541e-07, "loss": 0.3541, "num_input_tokens_seen": 4269801472, "step": 2036 }, { "epoch": 0.010035259018171955, "grad_norm": 22.198923110961914, "learning_rate": 5.092832372647304e-07, "loss": 0.5872, "num_input_tokens_seen": 4271898624, "step": 2037 }, { "epoch": 0.010306482234879306, "grad_norm": 16.18609046936035, "learning_rate": 5.08881221207708e-07, "loss": 0.5318, "num_input_tokens_seen": 4273995776, "step": 2038 }, { "epoch": 0.010577705451586655, "grad_norm": 17.77985191345215, "learning_rate": 5.084792382415141e-07, "loss": 0.2986, "num_input_tokens_seen": 4276092928, "step": 2039 }, { "epoch": 0.010848928668294006, "grad_norm": 17.145601272583008, "learning_rate": 5.080772886896493e-07, "loss": 0.4883, "num_input_tokens_seen": 4278190080, "step": 2040 }, { "epoch": 0.011120151885001357, "grad_norm": 13.083086013793945, "learning_rate": 5.076753728755871e-07, "loss": 0.3112, "num_input_tokens_seen": 4280287232, "step": 2041 }, { "epoch": 0.011391375101708706, "grad_norm": 17.41132354736328, "learning_rate": 5.072734911227746e-07, "loss": 0.4163, "num_input_tokens_seen": 4282384384, "step": 2042 }, { "epoch": 0.011662598318416057, "grad_norm": 8.932369232177734, "learning_rate": 5.068716437546305e-07, "loss": 0.2173, "num_input_tokens_seen": 4284481536, "step": 2043 }, { "epoch": 0.011933821535123406, "grad_norm": 16.91132164001465, "learning_rate": 5.06469831094546e-07, "loss": 0.6219, "num_input_tokens_seen": 4286578688, "step": 2044 }, { "epoch": 0.012205044751830757, "grad_norm": 15.44660758972168, "learning_rate": 5.060680534658852e-07, "loss": 0.4811, "num_input_tokens_seen": 4288675840, "step": 2045 }, { "epoch": 0.012476267968538107, "grad_norm": 11.51047134399414, "learning_rate": 5.056663111919833e-07, "loss": 0.4082, "num_input_tokens_seen": 4290772992, "step": 2046 }, { "epoch": 0.012747491185245458, "grad_norm": 9.88651180267334, "learning_rate": 5.052646045961471e-07, "loss": 0.2301, "num_input_tokens_seen": 4292870144, "step": 2047 }, { "epoch": 0.013018714401952807, "grad_norm": 14.237520217895508, "learning_rate": 5.048629340016548e-07, "loss": 0.4273, "num_input_tokens_seen": 4294967296, "step": 2048 }, { "epoch": 0.013289937618660158, "grad_norm": 14.176645278930664, "learning_rate": 5.044612997317557e-07, "loss": 0.3895, "num_input_tokens_seen": 4297064448, "step": 2049 }, { "epoch": 0.013561160835367507, "grad_norm": 18.345462799072266, "learning_rate": 5.040597021096695e-07, "loss": 0.7524, "num_input_tokens_seen": 4299161600, "step": 2050 }, { "epoch": 0.013832384052074858, "grad_norm": 15.396586418151855, "learning_rate": 5.03658141458587e-07, "loss": 0.4844, "num_input_tokens_seen": 4301258752, "step": 2051 }, { "epoch": 0.014103607268782207, "grad_norm": 17.37771224975586, "learning_rate": 5.032566181016688e-07, "loss": 0.381, "num_input_tokens_seen": 4303355904, "step": 2052 }, { "epoch": 0.014374830485489558, "grad_norm": 15.787535667419434, "learning_rate": 5.028551323620458e-07, "loss": 0.6099, "num_input_tokens_seen": 4305453056, "step": 2053 }, { "epoch": 0.014646053702196907, "grad_norm": 14.97440242767334, "learning_rate": 5.024536845628181e-07, "loss": 0.4472, "num_input_tokens_seen": 4307550208, "step": 2054 }, { "epoch": 0.014917276918904258, "grad_norm": 26.290592193603516, "learning_rate": 5.020522750270558e-07, "loss": 0.7158, "num_input_tokens_seen": 4309647360, "step": 2055 }, { "epoch": 0.015188500135611608, "grad_norm": 14.337723731994629, "learning_rate": 5.01650904077798e-07, "loss": 0.3129, "num_input_tokens_seen": 4311744512, "step": 2056 }, { "epoch": 0.015459723352318959, "grad_norm": 20.845989227294922, "learning_rate": 5.012495720380527e-07, "loss": 0.7573, "num_input_tokens_seen": 4313841664, "step": 2057 }, { "epoch": 0.015730946569026308, "grad_norm": 13.03821086883545, "learning_rate": 5.008482792307968e-07, "loss": 0.3595, "num_input_tokens_seen": 4315938816, "step": 2058 }, { "epoch": 0.01600216978573366, "grad_norm": 11.9432954788208, "learning_rate": 5.004470259789757e-07, "loss": 0.3451, "num_input_tokens_seen": 4318035968, "step": 2059 }, { "epoch": 0.01627339300244101, "grad_norm": 11.00033950805664, "learning_rate": 5.000458126055021e-07, "loss": 0.3564, "num_input_tokens_seen": 4320133120, "step": 2060 }, { "epoch": 0.01654461621914836, "grad_norm": 18.752838134765625, "learning_rate": 4.996446394332578e-07, "loss": 0.5661, "num_input_tokens_seen": 4322230272, "step": 2061 }, { "epoch": 0.016815839435855708, "grad_norm": 12.34859848022461, "learning_rate": 4.992435067850914e-07, "loss": 0.3662, "num_input_tokens_seen": 4324327424, "step": 2062 }, { "epoch": 0.01708706265256306, "grad_norm": 17.15289306640625, "learning_rate": 4.988424149838192e-07, "loss": 0.508, "num_input_tokens_seen": 4326424576, "step": 2063 }, { "epoch": 0.01735828586927041, "grad_norm": 31.281986236572266, "learning_rate": 4.984413643522251e-07, "loss": 0.8106, "num_input_tokens_seen": 4328521728, "step": 2064 }, { "epoch": 0.01762950908597776, "grad_norm": 12.578858375549316, "learning_rate": 4.980403552130589e-07, "loss": 0.3856, "num_input_tokens_seen": 4330618880, "step": 2065 }, { "epoch": 0.01790073230268511, "grad_norm": 14.403166770935059, "learning_rate": 4.976393878890375e-07, "loss": 0.4615, "num_input_tokens_seen": 4332716032, "step": 2066 }, { "epoch": 0.01817195551939246, "grad_norm": 14.294715881347656, "learning_rate": 4.972384627028442e-07, "loss": 0.4769, "num_input_tokens_seen": 4334813184, "step": 2067 }, { "epoch": 0.01844317873609981, "grad_norm": 16.856159210205078, "learning_rate": 4.968375799771285e-07, "loss": 0.3869, "num_input_tokens_seen": 4336910336, "step": 2068 }, { "epoch": 0.01871440195280716, "grad_norm": 11.613068580627441, "learning_rate": 4.964367400345053e-07, "loss": 0.3629, "num_input_tokens_seen": 4339007488, "step": 2069 }, { "epoch": 0.01898562516951451, "grad_norm": 17.622570037841797, "learning_rate": 4.960359431975555e-07, "loss": 0.738, "num_input_tokens_seen": 4341104640, "step": 2070 }, { "epoch": 0.01925684838622186, "grad_norm": 21.49909782409668, "learning_rate": 4.956351897888256e-07, "loss": 0.7802, "num_input_tokens_seen": 4343201792, "step": 2071 }, { "epoch": 0.01952807160292921, "grad_norm": 17.356836318969727, "learning_rate": 4.952344801308258e-07, "loss": 0.6596, "num_input_tokens_seen": 4345298944, "step": 2072 }, { "epoch": 0.019799294819636562, "grad_norm": 12.582341194152832, "learning_rate": 4.948338145460324e-07, "loss": 0.373, "num_input_tokens_seen": 4347396096, "step": 2073 }, { "epoch": 0.02007051803634391, "grad_norm": 18.529993057250977, "learning_rate": 4.944331933568858e-07, "loss": 0.7155, "num_input_tokens_seen": 4349493248, "step": 2074 }, { "epoch": 0.02034174125305126, "grad_norm": 16.39361572265625, "learning_rate": 4.94032616885791e-07, "loss": 0.6209, "num_input_tokens_seen": 4351590400, "step": 2075 }, { "epoch": 0.02061296446975861, "grad_norm": 12.521406173706055, "learning_rate": 4.936320854551163e-07, "loss": 0.4043, "num_input_tokens_seen": 4353687552, "step": 2076 }, { "epoch": 0.020884187686465962, "grad_norm": 9.356000900268555, "learning_rate": 4.932315993871942e-07, "loss": 0.2696, "num_input_tokens_seen": 4355784704, "step": 2077 }, { "epoch": 0.02115541090317331, "grad_norm": 13.908049583435059, "learning_rate": 4.928311590043206e-07, "loss": 0.3287, "num_input_tokens_seen": 4357881856, "step": 2078 }, { "epoch": 0.02142663411988066, "grad_norm": 17.52366065979004, "learning_rate": 4.924307646287549e-07, "loss": 0.7094, "num_input_tokens_seen": 4359979008, "step": 2079 }, { "epoch": 0.021697857336588012, "grad_norm": 21.215076446533203, "learning_rate": 4.920304165827192e-07, "loss": 0.7868, "num_input_tokens_seen": 4362076160, "step": 2080 }, { "epoch": 0.021969080553295363, "grad_norm": 24.048057556152344, "learning_rate": 4.916301151883983e-07, "loss": 0.4977, "num_input_tokens_seen": 4364173312, "step": 2081 }, { "epoch": 0.022240303770002714, "grad_norm": 18.432392120361328, "learning_rate": 4.912298607679396e-07, "loss": 0.6461, "num_input_tokens_seen": 4366270464, "step": 2082 }, { "epoch": 0.02251152698671006, "grad_norm": 16.7590389251709, "learning_rate": 4.908296536434525e-07, "loss": 0.4585, "num_input_tokens_seen": 4368367616, "step": 2083 }, { "epoch": 0.022782750203417412, "grad_norm": 21.218637466430664, "learning_rate": 4.904294941370086e-07, "loss": 0.7689, "num_input_tokens_seen": 4370464768, "step": 2084 }, { "epoch": 0.023053973420124763, "grad_norm": 13.70450210571289, "learning_rate": 4.900293825706408e-07, "loss": 0.3011, "num_input_tokens_seen": 4372561920, "step": 2085 }, { "epoch": 0.023325196636832114, "grad_norm": 12.731541633605957, "learning_rate": 4.89629319266344e-07, "loss": 0.444, "num_input_tokens_seen": 4374659072, "step": 2086 }, { "epoch": 0.023596419853539462, "grad_norm": 13.77962589263916, "learning_rate": 4.892293045460738e-07, "loss": 0.4021, "num_input_tokens_seen": 4376756224, "step": 2087 }, { "epoch": 0.023867643070246813, "grad_norm": 14.275106430053711, "learning_rate": 4.88829338731747e-07, "loss": 0.3847, "num_input_tokens_seen": 4378853376, "step": 2088 }, { "epoch": 0.024138866286954164, "grad_norm": 11.474467277526855, "learning_rate": 4.884294221452405e-07, "loss": 0.3539, "num_input_tokens_seen": 4380950528, "step": 2089 }, { "epoch": 0.024410089503661515, "grad_norm": 17.160314559936523, "learning_rate": 4.88029555108392e-07, "loss": 0.6452, "num_input_tokens_seen": 4383047680, "step": 2090 }, { "epoch": 0.024681312720368862, "grad_norm": 10.082201957702637, "learning_rate": 4.876297379429995e-07, "loss": 0.2754, "num_input_tokens_seen": 4385144832, "step": 2091 }, { "epoch": 0.024952535937076213, "grad_norm": 16.620037078857422, "learning_rate": 4.872299709708206e-07, "loss": 0.4956, "num_input_tokens_seen": 4387241984, "step": 2092 }, { "epoch": 0.025223759153783564, "grad_norm": 13.162679672241211, "learning_rate": 4.868302545135725e-07, "loss": 0.3437, "num_input_tokens_seen": 4389339136, "step": 2093 }, { "epoch": 0.025494982370490915, "grad_norm": 19.05023765563965, "learning_rate": 4.864305888929318e-07, "loss": 0.9092, "num_input_tokens_seen": 4391436288, "step": 2094 }, { "epoch": 0.025766205587198263, "grad_norm": 19.761568069458008, "learning_rate": 4.860309744305339e-07, "loss": 0.8281, "num_input_tokens_seen": 4393533440, "step": 2095 }, { "epoch": 0.026037428803905614, "grad_norm": 18.267724990844727, "learning_rate": 4.856314114479738e-07, "loss": 0.7678, "num_input_tokens_seen": 4395630592, "step": 2096 }, { "epoch": 0.026308652020612965, "grad_norm": 12.300085067749023, "learning_rate": 4.852319002668044e-07, "loss": 0.3752, "num_input_tokens_seen": 4397727744, "step": 2097 }, { "epoch": 0.026579875237320316, "grad_norm": 15.5980806350708, "learning_rate": 4.848324412085367e-07, "loss": 0.5343, "num_input_tokens_seen": 4399824896, "step": 2098 }, { "epoch": 0.026851098454027666, "grad_norm": 17.482070922851562, "learning_rate": 4.84433034594641e-07, "loss": 0.5354, "num_input_tokens_seen": 4401922048, "step": 2099 }, { "epoch": 0.027122321670735014, "grad_norm": 14.687858581542969, "learning_rate": 4.840336807465439e-07, "loss": 0.4802, "num_input_tokens_seen": 4404019200, "step": 2100 }, { "epoch": 0.027393544887442365, "grad_norm": 13.101119995117188, "learning_rate": 4.836343799856302e-07, "loss": 0.4752, "num_input_tokens_seen": 4406116352, "step": 2101 }, { "epoch": 0.027664768104149716, "grad_norm": 8.815567970275879, "learning_rate": 4.83235132633242e-07, "loss": 0.2088, "num_input_tokens_seen": 4408213504, "step": 2102 }, { "epoch": 0.027935991320857067, "grad_norm": 9.5147066116333, "learning_rate": 4.828359390106786e-07, "loss": 0.2341, "num_input_tokens_seen": 4410310656, "step": 2103 }, { "epoch": 0.028207214537564414, "grad_norm": 12.234147071838379, "learning_rate": 4.824367994391958e-07, "loss": 0.4088, "num_input_tokens_seen": 4412407808, "step": 2104 }, { "epoch": 0.028478437754271765, "grad_norm": 11.71864128112793, "learning_rate": 4.820377142400058e-07, "loss": 0.2418, "num_input_tokens_seen": 4414504960, "step": 2105 }, { "epoch": 0.028749660970979116, "grad_norm": 22.507688522338867, "learning_rate": 4.816386837342771e-07, "loss": 0.5335, "num_input_tokens_seen": 4416602112, "step": 2106 }, { "epoch": 0.029020884187686467, "grad_norm": 15.12206745147705, "learning_rate": 4.812397082431345e-07, "loss": 0.5299, "num_input_tokens_seen": 4418699264, "step": 2107 }, { "epoch": 0.029292107404393815, "grad_norm": 23.719499588012695, "learning_rate": 4.808407880876581e-07, "loss": 0.5769, "num_input_tokens_seen": 4420796416, "step": 2108 }, { "epoch": 0.029563330621101166, "grad_norm": 15.224874496459961, "learning_rate": 4.804419235888838e-07, "loss": 0.5696, "num_input_tokens_seen": 4422893568, "step": 2109 }, { "epoch": 0.029834553837808517, "grad_norm": 15.45009994506836, "learning_rate": 4.800431150678026e-07, "loss": 0.4187, "num_input_tokens_seen": 4424990720, "step": 2110 }, { "epoch": 0.030105777054515868, "grad_norm": 18.44724464416504, "learning_rate": 4.796443628453603e-07, "loss": 0.6679, "num_input_tokens_seen": 4427087872, "step": 2111 }, { "epoch": 0.030377000271223215, "grad_norm": 15.730889320373535, "learning_rate": 4.792456672424574e-07, "loss": 0.4933, "num_input_tokens_seen": 4429185024, "step": 2112 }, { "epoch": 0.030648223487930566, "grad_norm": 10.117815971374512, "learning_rate": 4.788470285799492e-07, "loss": 0.2543, "num_input_tokens_seen": 4431282176, "step": 2113 }, { "epoch": 0.030919446704637917, "grad_norm": 14.274503707885742, "learning_rate": 4.784484471786447e-07, "loss": 0.4122, "num_input_tokens_seen": 4433379328, "step": 2114 }, { "epoch": 0.031190669921345268, "grad_norm": 17.943662643432617, "learning_rate": 4.780499233593073e-07, "loss": 0.5615, "num_input_tokens_seen": 4435476480, "step": 2115 }, { "epoch": 0.031461893138052616, "grad_norm": 12.883108139038086, "learning_rate": 4.776514574426538e-07, "loss": 0.3805, "num_input_tokens_seen": 4437573632, "step": 2116 }, { "epoch": 0.03173311635475997, "grad_norm": 11.446172714233398, "learning_rate": 4.772530497493541e-07, "loss": 0.3671, "num_input_tokens_seen": 4439670784, "step": 2117 }, { "epoch": 0.03200433957146732, "grad_norm": 15.992230415344238, "learning_rate": 4.768547006000317e-07, "loss": 0.5018, "num_input_tokens_seen": 4441767936, "step": 2118 }, { "epoch": 0.032275562788174665, "grad_norm": 15.411177635192871, "learning_rate": 4.76456410315263e-07, "loss": 0.5656, "num_input_tokens_seen": 4443865088, "step": 2119 }, { "epoch": 0.03254678600488202, "grad_norm": 13.412214279174805, "learning_rate": 4.7605817921557666e-07, "loss": 0.3278, "num_input_tokens_seen": 4445962240, "step": 2120 }, { "epoch": 0.03281800922158937, "grad_norm": 10.45663833618164, "learning_rate": 4.756600076214541e-07, "loss": 0.3316, "num_input_tokens_seen": 4448059392, "step": 2121 }, { "epoch": 0.03308923243829672, "grad_norm": 17.893962860107422, "learning_rate": 4.7526189585332855e-07, "loss": 0.5041, "num_input_tokens_seen": 4450156544, "step": 2122 }, { "epoch": 0.03336045565500407, "grad_norm": 18.3114013671875, "learning_rate": 4.748638442315851e-07, "loss": 0.7515, "num_input_tokens_seen": 4452253696, "step": 2123 }, { "epoch": 0.033631678871711417, "grad_norm": 19.698787689208984, "learning_rate": 4.7446585307656074e-07, "loss": 0.425, "num_input_tokens_seen": 4454350848, "step": 2124 }, { "epoch": 0.03390290208841877, "grad_norm": 15.37483024597168, "learning_rate": 4.740679227085436e-07, "loss": 0.6293, "num_input_tokens_seen": 4456448000, "step": 2125 }, { "epoch": 0.03417412530512612, "grad_norm": 36.20682907104492, "learning_rate": 4.7367005344777255e-07, "loss": 0.5484, "num_input_tokens_seen": 4458545152, "step": 2126 }, { "epoch": 0.034445348521833466, "grad_norm": 14.473278045654297, "learning_rate": 4.732722456144381e-07, "loss": 0.4536, "num_input_tokens_seen": 4460642304, "step": 2127 }, { "epoch": 0.03471657173854082, "grad_norm": 13.409770965576172, "learning_rate": 4.7287449952868084e-07, "loss": 0.445, "num_input_tokens_seen": 4462739456, "step": 2128 }, { "epoch": 0.03498779495524817, "grad_norm": 11.895949363708496, "learning_rate": 4.724768155105914e-07, "loss": 0.2948, "num_input_tokens_seen": 4464836608, "step": 2129 }, { "epoch": 0.03525901817195552, "grad_norm": 18.184831619262695, "learning_rate": 4.7207919388021076e-07, "loss": 0.6784, "num_input_tokens_seen": 4466933760, "step": 2130 }, { "epoch": 0.03553024138866287, "grad_norm": 15.693188667297363, "learning_rate": 4.7168163495752977e-07, "loss": 0.4952, "num_input_tokens_seen": 4469030912, "step": 2131 }, { "epoch": 0.03580146460537022, "grad_norm": 14.902560234069824, "learning_rate": 4.7128413906248885e-07, "loss": 0.5021, "num_input_tokens_seen": 4471128064, "step": 2132 }, { "epoch": 0.03607268782207757, "grad_norm": 13.470413208007812, "learning_rate": 4.708867065149775e-07, "loss": 0.4083, "num_input_tokens_seen": 4473225216, "step": 2133 }, { "epoch": 0.03634391103878492, "grad_norm": 18.67334747314453, "learning_rate": 4.704893376348342e-07, "loss": 0.6536, "num_input_tokens_seen": 4475322368, "step": 2134 }, { "epoch": 0.03661513425549227, "grad_norm": 16.05646324157715, "learning_rate": 4.700920327418463e-07, "loss": 0.6247, "num_input_tokens_seen": 4477419520, "step": 2135 }, { "epoch": 0.03688635747219962, "grad_norm": 12.6869535446167, "learning_rate": 4.696947921557498e-07, "loss": 0.3876, "num_input_tokens_seen": 4479516672, "step": 2136 }, { "epoch": 0.03715758068890697, "grad_norm": 13.598705291748047, "learning_rate": 4.6929761619622866e-07, "loss": 0.4517, "num_input_tokens_seen": 4481613824, "step": 2137 }, { "epoch": 0.03742880390561432, "grad_norm": 12.73837661743164, "learning_rate": 4.68900505182915e-07, "loss": 0.3866, "num_input_tokens_seen": 4483710976, "step": 2138 }, { "epoch": 0.03770002712232167, "grad_norm": 10.884099006652832, "learning_rate": 4.6850345943538896e-07, "loss": 0.2201, "num_input_tokens_seen": 4485808128, "step": 2139 }, { "epoch": 0.03797125033902902, "grad_norm": 18.487733840942383, "learning_rate": 4.6810647927317735e-07, "loss": 0.5549, "num_input_tokens_seen": 4487905280, "step": 2140 }, { "epoch": 0.03824247355573637, "grad_norm": 22.46426010131836, "learning_rate": 4.677095650157551e-07, "loss": 0.4995, "num_input_tokens_seen": 4490002432, "step": 2141 }, { "epoch": 0.03851369677244372, "grad_norm": 16.194440841674805, "learning_rate": 4.6731271698254326e-07, "loss": 0.5513, "num_input_tokens_seen": 4492099584, "step": 2142 }, { "epoch": 0.038784919989151075, "grad_norm": 20.5705623626709, "learning_rate": 4.669159354929105e-07, "loss": 0.9542, "num_input_tokens_seen": 4494196736, "step": 2143 }, { "epoch": 0.03905614320585842, "grad_norm": 9.620034217834473, "learning_rate": 4.6651922086617134e-07, "loss": 0.2924, "num_input_tokens_seen": 4496293888, "step": 2144 }, { "epoch": 0.03932736642256577, "grad_norm": 15.673210144042969, "learning_rate": 4.661225734215867e-07, "loss": 0.4557, "num_input_tokens_seen": 4498391040, "step": 2145 }, { "epoch": 0.039598589639273124, "grad_norm": 10.837203025817871, "learning_rate": 4.657259934783628e-07, "loss": 0.3044, "num_input_tokens_seen": 4500488192, "step": 2146 }, { "epoch": 0.03986981285598047, "grad_norm": 15.021329879760742, "learning_rate": 4.6532948135565264e-07, "loss": 0.5526, "num_input_tokens_seen": 4502585344, "step": 2147 }, { "epoch": 0.04014103607268782, "grad_norm": 15.685761451721191, "learning_rate": 4.6493303737255397e-07, "loss": 0.5141, "num_input_tokens_seen": 4504682496, "step": 2148 }, { "epoch": 0.040412259289395173, "grad_norm": 17.80036163330078, "learning_rate": 4.645366618481098e-07, "loss": 0.6464, "num_input_tokens_seen": 4506779648, "step": 2149 }, { "epoch": 0.04068348250610252, "grad_norm": 17.086820602416992, "learning_rate": 4.641403551013081e-07, "loss": 0.8128, "num_input_tokens_seen": 4508876800, "step": 2150 }, { "epoch": 0.040954705722809875, "grad_norm": 13.269697189331055, "learning_rate": 4.637441174510813e-07, "loss": 0.4149, "num_input_tokens_seen": 4510973952, "step": 2151 }, { "epoch": 0.04122592893951722, "grad_norm": 14.31224250793457, "learning_rate": 4.633479492163066e-07, "loss": 0.4536, "num_input_tokens_seen": 4513071104, "step": 2152 }, { "epoch": 0.04149715215622457, "grad_norm": 13.99878215789795, "learning_rate": 4.6295185071580487e-07, "loss": 0.4298, "num_input_tokens_seen": 4515168256, "step": 2153 }, { "epoch": 0.041768375372931925, "grad_norm": 13.95372200012207, "learning_rate": 4.6255582226834133e-07, "loss": 0.4801, "num_input_tokens_seen": 4517265408, "step": 2154 }, { "epoch": 0.04203959858963927, "grad_norm": 19.383800506591797, "learning_rate": 4.6215986419262444e-07, "loss": 0.6272, "num_input_tokens_seen": 4519362560, "step": 2155 }, { "epoch": 0.04231082180634662, "grad_norm": 22.379093170166016, "learning_rate": 4.617639768073066e-07, "loss": 0.9261, "num_input_tokens_seen": 4521459712, "step": 2156 }, { "epoch": 0.042582045023053974, "grad_norm": 9.793371200561523, "learning_rate": 4.613681604309824e-07, "loss": 0.3081, "num_input_tokens_seen": 4523556864, "step": 2157 }, { "epoch": 0.04285326823976132, "grad_norm": 16.70575714111328, "learning_rate": 4.609724153821898e-07, "loss": 0.5864, "num_input_tokens_seen": 4525654016, "step": 2158 }, { "epoch": 0.043124491456468676, "grad_norm": 15.424320220947266, "learning_rate": 4.6057674197940974e-07, "loss": 0.5138, "num_input_tokens_seen": 4527751168, "step": 2159 }, { "epoch": 0.043395714673176024, "grad_norm": 12.337757110595703, "learning_rate": 4.6018114054106494e-07, "loss": 0.3479, "num_input_tokens_seen": 4529848320, "step": 2160 }, { "epoch": 0.04366693788988337, "grad_norm": 12.336735725402832, "learning_rate": 4.597856113855203e-07, "loss": 0.3803, "num_input_tokens_seen": 4531945472, "step": 2161 }, { "epoch": 0.043938161106590726, "grad_norm": 18.751192092895508, "learning_rate": 4.593901548310828e-07, "loss": 0.6854, "num_input_tokens_seen": 4534042624, "step": 2162 }, { "epoch": 0.04420938432329807, "grad_norm": 13.754616737365723, "learning_rate": 4.5899477119600073e-07, "loss": 0.483, "num_input_tokens_seen": 4536139776, "step": 2163 }, { "epoch": 0.04448060754000543, "grad_norm": 22.812694549560547, "learning_rate": 4.585994607984637e-07, "loss": 0.5172, "num_input_tokens_seen": 4538236928, "step": 2164 }, { "epoch": 0.044751830756712775, "grad_norm": 13.95494270324707, "learning_rate": 4.582042239566026e-07, "loss": 0.5708, "num_input_tokens_seen": 4540334080, "step": 2165 }, { "epoch": 0.04502305397342012, "grad_norm": 14.064531326293945, "learning_rate": 4.578090609884889e-07, "loss": 0.3845, "num_input_tokens_seen": 4542431232, "step": 2166 }, { "epoch": 0.04529427719012748, "grad_norm": 13.075774192810059, "learning_rate": 4.57413972212135e-07, "loss": 0.3608, "num_input_tokens_seen": 4544528384, "step": 2167 }, { "epoch": 0.045565500406834825, "grad_norm": 12.075623512268066, "learning_rate": 4.5701895794549293e-07, "loss": 0.3836, "num_input_tokens_seen": 4546625536, "step": 2168 }, { "epoch": 0.04583672362354217, "grad_norm": 14.203977584838867, "learning_rate": 4.566240185064554e-07, "loss": 0.3646, "num_input_tokens_seen": 4548722688, "step": 2169 }, { "epoch": 0.04610794684024953, "grad_norm": 16.41777229309082, "learning_rate": 4.5622915421285446e-07, "loss": 0.5869, "num_input_tokens_seen": 4550819840, "step": 2170 }, { "epoch": 0.046379170056956874, "grad_norm": 5.759540557861328, "learning_rate": 4.558343653824619e-07, "loss": 0.1035, "num_input_tokens_seen": 4552916992, "step": 2171 }, { "epoch": 0.04665039327366423, "grad_norm": 11.012629508972168, "learning_rate": 4.55439652332989e-07, "loss": 0.3469, "num_input_tokens_seen": 4555014144, "step": 2172 }, { "epoch": 0.046921616490371576, "grad_norm": 16.583019256591797, "learning_rate": 4.550450153820857e-07, "loss": 0.4748, "num_input_tokens_seen": 4557111296, "step": 2173 }, { "epoch": 0.047192839707078924, "grad_norm": 12.1753511428833, "learning_rate": 4.5465045484734044e-07, "loss": 0.4094, "num_input_tokens_seen": 4559208448, "step": 2174 }, { "epoch": 0.04746406292378628, "grad_norm": 12.30431842803955, "learning_rate": 4.5425597104628086e-07, "loss": 0.3467, "num_input_tokens_seen": 4561305600, "step": 2175 }, { "epoch": 0.047735286140493625, "grad_norm": 14.555771827697754, "learning_rate": 4.5386156429637256e-07, "loss": 0.4869, "num_input_tokens_seen": 4563402752, "step": 2176 }, { "epoch": 0.04800650935720097, "grad_norm": 14.621318817138672, "learning_rate": 4.534672349150189e-07, "loss": 0.3947, "num_input_tokens_seen": 4565499904, "step": 2177 }, { "epoch": 0.04827773257390833, "grad_norm": 14.621907234191895, "learning_rate": 4.530729832195612e-07, "loss": 0.3238, "num_input_tokens_seen": 4567597056, "step": 2178 }, { "epoch": 0.048548955790615675, "grad_norm": 17.090391159057617, "learning_rate": 4.5267880952727845e-07, "loss": 0.4632, "num_input_tokens_seen": 4569694208, "step": 2179 }, { "epoch": 0.04882017900732303, "grad_norm": 22.62020492553711, "learning_rate": 4.5228471415538637e-07, "loss": 0.9961, "num_input_tokens_seen": 4571791360, "step": 2180 }, { "epoch": 0.04909140222403038, "grad_norm": 15.037042617797852, "learning_rate": 4.518906974210379e-07, "loss": 0.4256, "num_input_tokens_seen": 4573888512, "step": 2181 }, { "epoch": 0.049362625440737724, "grad_norm": 13.915372848510742, "learning_rate": 4.514967596413228e-07, "loss": 0.4098, "num_input_tokens_seen": 4575985664, "step": 2182 }, { "epoch": 0.04963384865744508, "grad_norm": 14.323426246643066, "learning_rate": 4.511029011332672e-07, "loss": 0.3687, "num_input_tokens_seen": 4578082816, "step": 2183 }, { "epoch": 0.049905071874152426, "grad_norm": 16.720897674560547, "learning_rate": 4.507091222138335e-07, "loss": 0.6297, "num_input_tokens_seen": 4580179968, "step": 2184 }, { "epoch": 0.05017629509085978, "grad_norm": 12.606710433959961, "learning_rate": 4.503154231999202e-07, "loss": 0.2422, "num_input_tokens_seen": 4582277120, "step": 2185 }, { "epoch": 0.05044751830756713, "grad_norm": 19.402252197265625, "learning_rate": 4.499218044083608e-07, "loss": 0.7964, "num_input_tokens_seen": 4584374272, "step": 2186 }, { "epoch": 0.050718741524274476, "grad_norm": 13.136670112609863, "learning_rate": 4.495282661559248e-07, "loss": 0.3877, "num_input_tokens_seen": 4586471424, "step": 2187 }, { "epoch": 0.05098996474098183, "grad_norm": 16.292800903320312, "learning_rate": 4.491348087593171e-07, "loss": 0.6262, "num_input_tokens_seen": 4588568576, "step": 2188 }, { "epoch": 0.05126118795768918, "grad_norm": 9.792939186096191, "learning_rate": 4.4874143253517706e-07, "loss": 0.2886, "num_input_tokens_seen": 4590665728, "step": 2189 }, { "epoch": 0.051532411174396525, "grad_norm": 20.708683013916016, "learning_rate": 4.483481378000791e-07, "loss": 0.8719, "num_input_tokens_seen": 4592762880, "step": 2190 }, { "epoch": 0.05180363439110388, "grad_norm": 21.062313079833984, "learning_rate": 4.4795492487053155e-07, "loss": 0.5302, "num_input_tokens_seen": 4594860032, "step": 2191 }, { "epoch": 0.05207485760781123, "grad_norm": 11.393677711486816, "learning_rate": 4.4756179406297744e-07, "loss": 0.3503, "num_input_tokens_seen": 4596957184, "step": 2192 }, { "epoch": 0.05234608082451858, "grad_norm": 13.3013334274292, "learning_rate": 4.4716874569379336e-07, "loss": 0.4068, "num_input_tokens_seen": 4599054336, "step": 2193 }, { "epoch": 0.05261730404122593, "grad_norm": 14.940890312194824, "learning_rate": 4.467757800792897e-07, "loss": 0.3933, "num_input_tokens_seen": 4601151488, "step": 2194 }, { "epoch": 0.05288852725793328, "grad_norm": 18.109926223754883, "learning_rate": 4.4638289753571025e-07, "loss": 0.5994, "num_input_tokens_seen": 4603248640, "step": 2195 }, { "epoch": 0.05315975047464063, "grad_norm": 11.895615577697754, "learning_rate": 4.459900983792321e-07, "loss": 0.355, "num_input_tokens_seen": 4605345792, "step": 2196 }, { "epoch": 0.05343097369134798, "grad_norm": 12.7399320602417, "learning_rate": 4.455973829259648e-07, "loss": 0.3251, "num_input_tokens_seen": 4607442944, "step": 2197 }, { "epoch": 0.05370219690805533, "grad_norm": 14.225614547729492, "learning_rate": 4.4520475149195093e-07, "loss": 0.4504, "num_input_tokens_seen": 4609540096, "step": 2198 }, { "epoch": 0.05397342012476268, "grad_norm": 12.06645679473877, "learning_rate": 4.4481220439316514e-07, "loss": 0.3057, "num_input_tokens_seen": 4611637248, "step": 2199 }, { "epoch": 0.05424464334147003, "grad_norm": 17.66695785522461, "learning_rate": 4.444197419455147e-07, "loss": 0.7025, "num_input_tokens_seen": 4613734400, "step": 2200 }, { "epoch": 0.05451586655817738, "grad_norm": 14.438081741333008, "learning_rate": 4.4402736446483824e-07, "loss": 0.5549, "num_input_tokens_seen": 4615831552, "step": 2201 }, { "epoch": 0.05478708977488473, "grad_norm": 21.305463790893555, "learning_rate": 4.436350722669065e-07, "loss": 0.4533, "num_input_tokens_seen": 4617928704, "step": 2202 }, { "epoch": 0.05505831299159208, "grad_norm": 17.13565444946289, "learning_rate": 4.43242865667421e-07, "loss": 0.5579, "num_input_tokens_seen": 4620025856, "step": 2203 }, { "epoch": 0.05532953620829943, "grad_norm": 16.301301956176758, "learning_rate": 4.428507449820147e-07, "loss": 0.5478, "num_input_tokens_seen": 4622123008, "step": 2204 }, { "epoch": 0.05560075942500678, "grad_norm": 17.26921272277832, "learning_rate": 4.424587105262516e-07, "loss": 0.6332, "num_input_tokens_seen": 4624220160, "step": 2205 }, { "epoch": 0.055871982641714134, "grad_norm": 16.908565521240234, "learning_rate": 4.4206676261562603e-07, "loss": 0.6461, "num_input_tokens_seen": 4626317312, "step": 2206 }, { "epoch": 0.05614320585842148, "grad_norm": 15.162359237670898, "learning_rate": 4.4167490156556274e-07, "loss": 0.4447, "num_input_tokens_seen": 4628414464, "step": 2207 }, { "epoch": 0.05641442907512883, "grad_norm": 12.857962608337402, "learning_rate": 4.4128312769141664e-07, "loss": 0.4044, "num_input_tokens_seen": 4630511616, "step": 2208 }, { "epoch": 0.05668565229183618, "grad_norm": 12.184791564941406, "learning_rate": 4.408914413084722e-07, "loss": 0.3242, "num_input_tokens_seen": 4632608768, "step": 2209 }, { "epoch": 0.05695687550854353, "grad_norm": 19.207063674926758, "learning_rate": 4.4049984273194406e-07, "loss": 0.7085, "num_input_tokens_seen": 4634705920, "step": 2210 }, { "epoch": 0.05722809872525088, "grad_norm": 19.66663932800293, "learning_rate": 4.401083322769756e-07, "loss": 0.7906, "num_input_tokens_seen": 4636803072, "step": 2211 }, { "epoch": 0.05749932194195823, "grad_norm": 13.727479934692383, "learning_rate": 4.3971691025863964e-07, "loss": 0.4326, "num_input_tokens_seen": 4638900224, "step": 2212 }, { "epoch": 0.05777054515866558, "grad_norm": 15.684118270874023, "learning_rate": 4.3932557699193794e-07, "loss": 0.5288, "num_input_tokens_seen": 4640997376, "step": 2213 }, { "epoch": 0.058041768375372935, "grad_norm": 12.542720794677734, "learning_rate": 4.389343327918004e-07, "loss": 0.401, "num_input_tokens_seen": 4643094528, "step": 2214 }, { "epoch": 0.05831299159208028, "grad_norm": 13.282032012939453, "learning_rate": 4.3854317797308515e-07, "loss": 0.4052, "num_input_tokens_seen": 4645191680, "step": 2215 }, { "epoch": 0.05858421480878763, "grad_norm": 12.997306823730469, "learning_rate": 4.381521128505791e-07, "loss": 0.3949, "num_input_tokens_seen": 4647288832, "step": 2216 }, { "epoch": 0.058855438025494984, "grad_norm": 10.497172355651855, "learning_rate": 4.377611377389966e-07, "loss": 0.2387, "num_input_tokens_seen": 4649385984, "step": 2217 }, { "epoch": 0.05912666124220233, "grad_norm": 11.194908142089844, "learning_rate": 4.3737025295297937e-07, "loss": 0.3324, "num_input_tokens_seen": 4651483136, "step": 2218 }, { "epoch": 0.059397884458909686, "grad_norm": 16.6368408203125, "learning_rate": 4.3697945880709686e-07, "loss": 0.5526, "num_input_tokens_seen": 4653580288, "step": 2219 }, { "epoch": 0.059669107675617034, "grad_norm": 17.438602447509766, "learning_rate": 4.3658875561584494e-07, "loss": 0.5616, "num_input_tokens_seen": 4655677440, "step": 2220 }, { "epoch": 0.05994033089232438, "grad_norm": 10.89105224609375, "learning_rate": 4.36198143693647e-07, "loss": 0.2007, "num_input_tokens_seen": 4657774592, "step": 2221 }, { "epoch": 0.060211554109031735, "grad_norm": 10.294246673583984, "learning_rate": 4.3580762335485253e-07, "loss": 0.2698, "num_input_tokens_seen": 4659871744, "step": 2222 }, { "epoch": 0.06048277732573908, "grad_norm": 22.007549285888672, "learning_rate": 4.3541719491373743e-07, "loss": 0.7437, "num_input_tokens_seen": 4661968896, "step": 2223 }, { "epoch": 0.06075400054244643, "grad_norm": 18.49791717529297, "learning_rate": 4.350268586845035e-07, "loss": 0.6865, "num_input_tokens_seen": 4664066048, "step": 2224 }, { "epoch": 0.061025223759153785, "grad_norm": 13.56779956817627, "learning_rate": 4.34636614981279e-07, "loss": 0.4945, "num_input_tokens_seen": 4666163200, "step": 2225 }, { "epoch": 0.06129644697586113, "grad_norm": 14.960966110229492, "learning_rate": 4.342464641181166e-07, "loss": 0.4721, "num_input_tokens_seen": 4668260352, "step": 2226 }, { "epoch": 0.06156767019256849, "grad_norm": 11.218042373657227, "learning_rate": 4.3385640640899524e-07, "loss": 0.3769, "num_input_tokens_seen": 4670357504, "step": 2227 }, { "epoch": 0.061838893409275834, "grad_norm": 13.593111991882324, "learning_rate": 4.3346644216781823e-07, "loss": 0.4218, "num_input_tokens_seen": 4672454656, "step": 2228 }, { "epoch": 0.06211011662598318, "grad_norm": 11.54797077178955, "learning_rate": 4.3307657170841417e-07, "loss": 0.2934, "num_input_tokens_seen": 4674551808, "step": 2229 }, { "epoch": 0.062381339842690536, "grad_norm": 17.747589111328125, "learning_rate": 4.326867953445362e-07, "loss": 0.6918, "num_input_tokens_seen": 4676648960, "step": 2230 }, { "epoch": 0.06265256305939788, "grad_norm": 14.534211158752441, "learning_rate": 4.322971133898611e-07, "loss": 0.4678, "num_input_tokens_seen": 4678746112, "step": 2231 }, { "epoch": 0.06292378627610523, "grad_norm": 15.299826622009277, "learning_rate": 4.319075261579901e-07, "loss": 0.3943, "num_input_tokens_seen": 4680843264, "step": 2232 }, { "epoch": 0.06319500949281258, "grad_norm": 14.455315589904785, "learning_rate": 4.3151803396244833e-07, "loss": 0.5782, "num_input_tokens_seen": 4682940416, "step": 2233 }, { "epoch": 0.06346623270951994, "grad_norm": 18.758607864379883, "learning_rate": 4.3112863711668435e-07, "loss": 0.3796, "num_input_tokens_seen": 4685037568, "step": 2234 }, { "epoch": 0.06373745592622729, "grad_norm": 19.549890518188477, "learning_rate": 4.3073933593406974e-07, "loss": 0.9138, "num_input_tokens_seen": 4687134720, "step": 2235 }, { "epoch": 0.06400867914293464, "grad_norm": 11.473661422729492, "learning_rate": 4.303501307278994e-07, "loss": 0.3049, "num_input_tokens_seen": 4689231872, "step": 2236 }, { "epoch": 0.06427990235964198, "grad_norm": 12.786169052124023, "learning_rate": 4.299610218113908e-07, "loss": 0.4731, "num_input_tokens_seen": 4691329024, "step": 2237 }, { "epoch": 0.06455112557634933, "grad_norm": 15.072799682617188, "learning_rate": 4.2957200949768414e-07, "loss": 0.3859, "num_input_tokens_seen": 4693426176, "step": 2238 }, { "epoch": 0.06482234879305669, "grad_norm": 20.06634521484375, "learning_rate": 4.2918309409984145e-07, "loss": 0.72, "num_input_tokens_seen": 4695523328, "step": 2239 }, { "epoch": 0.06509357200976404, "grad_norm": 8.56005573272705, "learning_rate": 4.2879427593084714e-07, "loss": 0.1702, "num_input_tokens_seen": 4697620480, "step": 2240 }, { "epoch": 0.06536479522647139, "grad_norm": 14.993252754211426, "learning_rate": 4.2840555530360756e-07, "loss": 0.403, "num_input_tokens_seen": 4699717632, "step": 2241 }, { "epoch": 0.06563601844317873, "grad_norm": 13.477447509765625, "learning_rate": 4.280169325309502e-07, "loss": 0.3717, "num_input_tokens_seen": 4701814784, "step": 2242 }, { "epoch": 0.06590724165988608, "grad_norm": 16.231098175048828, "learning_rate": 4.276284079256235e-07, "loss": 0.3061, "num_input_tokens_seen": 4703911936, "step": 2243 }, { "epoch": 0.06617846487659344, "grad_norm": 16.42155647277832, "learning_rate": 4.272399818002974e-07, "loss": 0.561, "num_input_tokens_seen": 4706009088, "step": 2244 }, { "epoch": 0.06644968809330079, "grad_norm": 14.12633228302002, "learning_rate": 4.268516544675628e-07, "loss": 0.293, "num_input_tokens_seen": 4708106240, "step": 2245 }, { "epoch": 0.06672091131000814, "grad_norm": 14.42664623260498, "learning_rate": 4.2646342623993035e-07, "loss": 0.3539, "num_input_tokens_seen": 4710203392, "step": 2246 }, { "epoch": 0.06699213452671549, "grad_norm": 11.090810775756836, "learning_rate": 4.2607529742983174e-07, "loss": 0.3227, "num_input_tokens_seen": 4712300544, "step": 2247 }, { "epoch": 0.06726335774342283, "grad_norm": 21.97173500061035, "learning_rate": 4.2568726834961797e-07, "loss": 0.7627, "num_input_tokens_seen": 4714397696, "step": 2248 }, { "epoch": 0.06753458096013018, "grad_norm": 10.479046821594238, "learning_rate": 4.252993393115601e-07, "loss": 0.2715, "num_input_tokens_seen": 4716494848, "step": 2249 }, { "epoch": 0.06780580417683754, "grad_norm": 20.145353317260742, "learning_rate": 4.249115106278489e-07, "loss": 0.4898, "num_input_tokens_seen": 4718592000, "step": 2250 }, { "epoch": 0.06807702739354489, "grad_norm": 18.290271759033203, "learning_rate": 4.2452378261059397e-07, "loss": 0.5795, "num_input_tokens_seen": 4720689152, "step": 2251 }, { "epoch": 0.06834825061025224, "grad_norm": 16.21881103515625, "learning_rate": 4.24136155571824e-07, "loss": 0.48, "num_input_tokens_seen": 4722786304, "step": 2252 }, { "epoch": 0.06861947382695958, "grad_norm": 14.017351150512695, "learning_rate": 4.2374862982348657e-07, "loss": 0.431, "num_input_tokens_seen": 4724883456, "step": 2253 }, { "epoch": 0.06889069704366693, "grad_norm": 20.59170150756836, "learning_rate": 4.233612056774477e-07, "loss": 0.776, "num_input_tokens_seen": 4726980608, "step": 2254 }, { "epoch": 0.0691619202603743, "grad_norm": 13.373947143554688, "learning_rate": 4.2297388344549146e-07, "loss": 0.3047, "num_input_tokens_seen": 4729077760, "step": 2255 }, { "epoch": 0.06943314347708164, "grad_norm": 15.071645736694336, "learning_rate": 4.2258666343932004e-07, "loss": 0.4074, "num_input_tokens_seen": 4731174912, "step": 2256 }, { "epoch": 0.06970436669378899, "grad_norm": 14.832840919494629, "learning_rate": 4.2219954597055354e-07, "loss": 0.3752, "num_input_tokens_seen": 4733272064, "step": 2257 }, { "epoch": 0.06997558991049634, "grad_norm": 11.459256172180176, "learning_rate": 4.2181253135072925e-07, "loss": 0.3155, "num_input_tokens_seen": 4735369216, "step": 2258 }, { "epoch": 0.07024681312720368, "grad_norm": 22.15343475341797, "learning_rate": 4.21425619891302e-07, "loss": 0.6153, "num_input_tokens_seen": 4737466368, "step": 2259 }, { "epoch": 0.07051803634391104, "grad_norm": 21.2708740234375, "learning_rate": 4.21038811903643e-07, "loss": 0.6954, "num_input_tokens_seen": 4739563520, "step": 2260 }, { "epoch": 0.07078925956061839, "grad_norm": 17.88982582092285, "learning_rate": 4.206521076990409e-07, "loss": 0.6006, "num_input_tokens_seen": 4741660672, "step": 2261 }, { "epoch": 0.07106048277732574, "grad_norm": 14.392377853393555, "learning_rate": 4.202655075887005e-07, "loss": 0.451, "num_input_tokens_seen": 4743757824, "step": 2262 }, { "epoch": 0.07133170599403309, "grad_norm": 12.109110832214355, "learning_rate": 4.1987901188374286e-07, "loss": 0.2934, "num_input_tokens_seen": 4745854976, "step": 2263 }, { "epoch": 0.07160292921074043, "grad_norm": 14.256464958190918, "learning_rate": 4.194926208952051e-07, "loss": 0.4602, "num_input_tokens_seen": 4747952128, "step": 2264 }, { "epoch": 0.0718741524274478, "grad_norm": 11.486831665039062, "learning_rate": 4.191063349340397e-07, "loss": 0.3412, "num_input_tokens_seen": 4750049280, "step": 2265 }, { "epoch": 0.07214537564415514, "grad_norm": 12.454242706298828, "learning_rate": 4.1872015431111505e-07, "loss": 0.3429, "num_input_tokens_seen": 4752146432, "step": 2266 }, { "epoch": 0.07241659886086249, "grad_norm": 17.0006103515625, "learning_rate": 4.1833407933721476e-07, "loss": 0.6483, "num_input_tokens_seen": 4754243584, "step": 2267 }, { "epoch": 0.07268782207756984, "grad_norm": 15.69262981414795, "learning_rate": 4.17948110323037e-07, "loss": 0.3751, "num_input_tokens_seen": 4756340736, "step": 2268 }, { "epoch": 0.07295904529427719, "grad_norm": 12.206964492797852, "learning_rate": 4.1756224757919513e-07, "loss": 0.3653, "num_input_tokens_seen": 4758437888, "step": 2269 }, { "epoch": 0.07323026851098453, "grad_norm": 16.216537475585938, "learning_rate": 4.17176491416217e-07, "loss": 0.5839, "num_input_tokens_seen": 4760535040, "step": 2270 }, { "epoch": 0.0735014917276919, "grad_norm": 11.075773239135742, "learning_rate": 4.1679084214454405e-07, "loss": 0.3098, "num_input_tokens_seen": 4762632192, "step": 2271 }, { "epoch": 0.07377271494439924, "grad_norm": 16.252365112304688, "learning_rate": 4.1640530007453245e-07, "loss": 0.4551, "num_input_tokens_seen": 4764729344, "step": 2272 }, { "epoch": 0.07404393816110659, "grad_norm": 11.240346908569336, "learning_rate": 4.1601986551645163e-07, "loss": 0.2972, "num_input_tokens_seen": 4766826496, "step": 2273 }, { "epoch": 0.07431516137781394, "grad_norm": 13.240141868591309, "learning_rate": 4.156345387804847e-07, "loss": 0.3682, "num_input_tokens_seen": 4768923648, "step": 2274 }, { "epoch": 0.07458638459452128, "grad_norm": 10.888014793395996, "learning_rate": 4.152493201767281e-07, "loss": 0.336, "num_input_tokens_seen": 4771020800, "step": 2275 }, { "epoch": 0.07485760781122865, "grad_norm": 18.147985458374023, "learning_rate": 4.1486421001519087e-07, "loss": 0.6925, "num_input_tokens_seen": 4773117952, "step": 2276 }, { "epoch": 0.075128831027936, "grad_norm": 13.516504287719727, "learning_rate": 4.1447920860579524e-07, "loss": 0.3624, "num_input_tokens_seen": 4775215104, "step": 2277 }, { "epoch": 0.07540005424464334, "grad_norm": 16.765531539916992, "learning_rate": 4.1409431625837545e-07, "loss": 0.5945, "num_input_tokens_seen": 4777312256, "step": 2278 }, { "epoch": 0.07567127746135069, "grad_norm": 15.376947402954102, "learning_rate": 4.137095332826784e-07, "loss": 0.4172, "num_input_tokens_seen": 4779409408, "step": 2279 }, { "epoch": 0.07594250067805804, "grad_norm": 15.102494239807129, "learning_rate": 4.1332485998836277e-07, "loss": 0.4686, "num_input_tokens_seen": 4781506560, "step": 2280 }, { "epoch": 0.0762137238947654, "grad_norm": 18.478492736816406, "learning_rate": 4.129402966849987e-07, "loss": 0.2808, "num_input_tokens_seen": 4783603712, "step": 2281 }, { "epoch": 0.07648494711147275, "grad_norm": 16.129724502563477, "learning_rate": 4.1255584368206877e-07, "loss": 0.5027, "num_input_tokens_seen": 4785700864, "step": 2282 }, { "epoch": 0.07675617032818009, "grad_norm": 16.364654541015625, "learning_rate": 4.121715012889655e-07, "loss": 0.4003, "num_input_tokens_seen": 4787798016, "step": 2283 }, { "epoch": 0.07702739354488744, "grad_norm": 12.427045822143555, "learning_rate": 4.1178726981499313e-07, "loss": 0.393, "num_input_tokens_seen": 4789895168, "step": 2284 }, { "epoch": 0.07729861676159479, "grad_norm": 27.002498626708984, "learning_rate": 4.1140314956936663e-07, "loss": 0.6929, "num_input_tokens_seen": 4791992320, "step": 2285 }, { "epoch": 0.07756983997830215, "grad_norm": 11.09440803527832, "learning_rate": 4.1101914086121137e-07, "loss": 0.2603, "num_input_tokens_seen": 4794089472, "step": 2286 }, { "epoch": 0.0778410631950095, "grad_norm": 15.960718154907227, "learning_rate": 4.106352439995632e-07, "loss": 0.482, "num_input_tokens_seen": 4796186624, "step": 2287 }, { "epoch": 0.07811228641171684, "grad_norm": 14.623363494873047, "learning_rate": 4.102514592933671e-07, "loss": 0.3698, "num_input_tokens_seen": 4798283776, "step": 2288 }, { "epoch": 0.07838350962842419, "grad_norm": 21.368053436279297, "learning_rate": 4.098677870514788e-07, "loss": 0.4633, "num_input_tokens_seen": 4800380928, "step": 2289 }, { "epoch": 0.07865473284513154, "grad_norm": 21.473129272460938, "learning_rate": 4.094842275826631e-07, "loss": 0.4671, "num_input_tokens_seen": 4802478080, "step": 2290 }, { "epoch": 0.07892595606183889, "grad_norm": 18.82076644897461, "learning_rate": 4.091007811955941e-07, "loss": 0.4475, "num_input_tokens_seen": 4804575232, "step": 2291 }, { "epoch": 0.07919717927854625, "grad_norm": 15.747272491455078, "learning_rate": 4.0871744819885467e-07, "loss": 0.5627, "num_input_tokens_seen": 4806672384, "step": 2292 }, { "epoch": 0.0794684024952536, "grad_norm": 18.276941299438477, "learning_rate": 4.0833422890093684e-07, "loss": 0.6141, "num_input_tokens_seen": 4808769536, "step": 2293 }, { "epoch": 0.07973962571196094, "grad_norm": 18.258466720581055, "learning_rate": 4.0795112361024075e-07, "loss": 0.5932, "num_input_tokens_seen": 4810866688, "step": 2294 }, { "epoch": 0.08001084892866829, "grad_norm": 12.13038158416748, "learning_rate": 4.0756813263507496e-07, "loss": 0.3242, "num_input_tokens_seen": 4812963840, "step": 2295 }, { "epoch": 0.08028207214537564, "grad_norm": 13.339944839477539, "learning_rate": 4.0718525628365617e-07, "loss": 0.3595, "num_input_tokens_seen": 4815060992, "step": 2296 }, { "epoch": 0.080553295362083, "grad_norm": 24.59098243713379, "learning_rate": 4.0680249486410845e-07, "loss": 0.4553, "num_input_tokens_seen": 4817158144, "step": 2297 }, { "epoch": 0.08082451857879035, "grad_norm": 18.36080551147461, "learning_rate": 4.0641984868446386e-07, "loss": 0.5813, "num_input_tokens_seen": 4819255296, "step": 2298 }, { "epoch": 0.0810957417954977, "grad_norm": 12.506386756896973, "learning_rate": 4.0603731805266175e-07, "loss": 0.369, "num_input_tokens_seen": 4821352448, "step": 2299 }, { "epoch": 0.08136696501220504, "grad_norm": 14.215474128723145, "learning_rate": 4.056549032765476e-07, "loss": 0.4783, "num_input_tokens_seen": 4823449600, "step": 2300 }, { "epoch": 0.08163818822891239, "grad_norm": 25.43838119506836, "learning_rate": 4.0527260466387446e-07, "loss": 0.8514, "num_input_tokens_seen": 4825546752, "step": 2301 }, { "epoch": 0.08190941144561975, "grad_norm": 17.24186134338379, "learning_rate": 4.0489042252230197e-07, "loss": 0.5077, "num_input_tokens_seen": 4827643904, "step": 2302 }, { "epoch": 0.0821806346623271, "grad_norm": 13.061026573181152, "learning_rate": 4.045083571593955e-07, "loss": 0.3865, "num_input_tokens_seen": 4829741056, "step": 2303 }, { "epoch": 0.08245185787903445, "grad_norm": 19.54876136779785, "learning_rate": 4.0412640888262697e-07, "loss": 0.7823, "num_input_tokens_seen": 4831838208, "step": 2304 }, { "epoch": 0.0827230810957418, "grad_norm": 13.487531661987305, "learning_rate": 4.0374457799937354e-07, "loss": 0.4054, "num_input_tokens_seen": 4833935360, "step": 2305 }, { "epoch": 0.08299430431244914, "grad_norm": 15.34504222869873, "learning_rate": 4.033628648169184e-07, "loss": 0.3901, "num_input_tokens_seen": 4836032512, "step": 2306 }, { "epoch": 0.0832655275291565, "grad_norm": 23.76385498046875, "learning_rate": 4.0298126964244947e-07, "loss": 0.5777, "num_input_tokens_seen": 4838129664, "step": 2307 }, { "epoch": 0.08353675074586385, "grad_norm": 11.67712116241455, "learning_rate": 4.025997927830604e-07, "loss": 0.3415, "num_input_tokens_seen": 4840226816, "step": 2308 }, { "epoch": 0.0838079739625712, "grad_norm": 18.445024490356445, "learning_rate": 4.022184345457492e-07, "loss": 0.3676, "num_input_tokens_seen": 4842323968, "step": 2309 }, { "epoch": 0.08407919717927854, "grad_norm": 15.196893692016602, "learning_rate": 4.0183719523741846e-07, "loss": 0.4861, "num_input_tokens_seen": 4844421120, "step": 2310 }, { "epoch": 0.08435042039598589, "grad_norm": 16.10106658935547, "learning_rate": 4.014560751648747e-07, "loss": 0.4878, "num_input_tokens_seen": 4846518272, "step": 2311 }, { "epoch": 0.08462164361269324, "grad_norm": 17.08901023864746, "learning_rate": 4.0107507463482924e-07, "loss": 0.4657, "num_input_tokens_seen": 4848615424, "step": 2312 }, { "epoch": 0.0848928668294006, "grad_norm": 16.41490364074707, "learning_rate": 4.0069419395389657e-07, "loss": 0.5807, "num_input_tokens_seen": 4850712576, "step": 2313 }, { "epoch": 0.08516409004610795, "grad_norm": 15.139748573303223, "learning_rate": 4.0031343342859526e-07, "loss": 0.5088, "num_input_tokens_seen": 4852809728, "step": 2314 }, { "epoch": 0.0854353132628153, "grad_norm": 16.5777645111084, "learning_rate": 3.9993279336534667e-07, "loss": 0.6122, "num_input_tokens_seen": 4854906880, "step": 2315 }, { "epoch": 0.08570653647952264, "grad_norm": 14.561357498168945, "learning_rate": 3.995522740704758e-07, "loss": 0.5766, "num_input_tokens_seen": 4857004032, "step": 2316 }, { "epoch": 0.08597775969622999, "grad_norm": 17.497167587280273, "learning_rate": 3.991718758502094e-07, "loss": 0.4089, "num_input_tokens_seen": 4859101184, "step": 2317 }, { "epoch": 0.08624898291293735, "grad_norm": 19.5342960357666, "learning_rate": 3.987915990106779e-07, "loss": 0.5453, "num_input_tokens_seen": 4861198336, "step": 2318 }, { "epoch": 0.0865202061296447, "grad_norm": 14.03099250793457, "learning_rate": 3.984114438579137e-07, "loss": 0.4075, "num_input_tokens_seen": 4863295488, "step": 2319 }, { "epoch": 0.08679142934635205, "grad_norm": 12.443893432617188, "learning_rate": 3.980314106978512e-07, "loss": 0.4093, "num_input_tokens_seen": 4865392640, "step": 2320 }, { "epoch": 0.0870626525630594, "grad_norm": 11.860504150390625, "learning_rate": 3.976514998363265e-07, "loss": 0.3936, "num_input_tokens_seen": 4867489792, "step": 2321 }, { "epoch": 0.08733387577976674, "grad_norm": 14.789926528930664, "learning_rate": 3.972717115790773e-07, "loss": 0.372, "num_input_tokens_seen": 4869586944, "step": 2322 }, { "epoch": 0.0876050989964741, "grad_norm": 18.579681396484375, "learning_rate": 3.968920462317431e-07, "loss": 0.5335, "num_input_tokens_seen": 4871684096, "step": 2323 }, { "epoch": 0.08787632221318145, "grad_norm": 12.13874340057373, "learning_rate": 3.965125040998637e-07, "loss": 0.3787, "num_input_tokens_seen": 4873781248, "step": 2324 }, { "epoch": 0.0881475454298888, "grad_norm": 14.869010925292969, "learning_rate": 3.961330854888805e-07, "loss": 0.4379, "num_input_tokens_seen": 4875878400, "step": 2325 }, { "epoch": 0.08841876864659615, "grad_norm": 10.25628662109375, "learning_rate": 3.9575379070413485e-07, "loss": 0.2783, "num_input_tokens_seen": 4877975552, "step": 2326 }, { "epoch": 0.0886899918633035, "grad_norm": 26.221960067749023, "learning_rate": 3.953746200508693e-07, "loss": 0.7503, "num_input_tokens_seen": 4880072704, "step": 2327 }, { "epoch": 0.08896121508001086, "grad_norm": 17.73170280456543, "learning_rate": 3.9499557383422534e-07, "loss": 0.6836, "num_input_tokens_seen": 4882169856, "step": 2328 }, { "epoch": 0.0892324382967182, "grad_norm": 14.907843589782715, "learning_rate": 3.94616652359245e-07, "loss": 0.3363, "num_input_tokens_seen": 4884267008, "step": 2329 }, { "epoch": 0.08950366151342555, "grad_norm": 8.722765922546387, "learning_rate": 3.942378559308703e-07, "loss": 0.2435, "num_input_tokens_seen": 4886364160, "step": 2330 }, { "epoch": 0.0897748847301329, "grad_norm": 17.718971252441406, "learning_rate": 3.93859184853942e-07, "loss": 0.4271, "num_input_tokens_seen": 4888461312, "step": 2331 }, { "epoch": 0.09004610794684025, "grad_norm": 12.278681755065918, "learning_rate": 3.934806394332001e-07, "loss": 0.3989, "num_input_tokens_seen": 4890558464, "step": 2332 }, { "epoch": 0.09031733116354759, "grad_norm": 19.766408920288086, "learning_rate": 3.9310221997328363e-07, "loss": 0.5485, "num_input_tokens_seen": 4892655616, "step": 2333 }, { "epoch": 0.09058855438025495, "grad_norm": 18.075057983398438, "learning_rate": 3.9272392677873e-07, "loss": 0.6252, "num_input_tokens_seen": 4894752768, "step": 2334 }, { "epoch": 0.0908597775969623, "grad_norm": 21.87603187561035, "learning_rate": 3.9234576015397536e-07, "loss": 0.8567, "num_input_tokens_seen": 4896849920, "step": 2335 }, { "epoch": 0.09113100081366965, "grad_norm": 11.228300094604492, "learning_rate": 3.9196772040335367e-07, "loss": 0.3036, "num_input_tokens_seen": 4898947072, "step": 2336 }, { "epoch": 0.091402224030377, "grad_norm": 25.12662696838379, "learning_rate": 3.915898078310972e-07, "loss": 0.674, "num_input_tokens_seen": 4901044224, "step": 2337 }, { "epoch": 0.09167344724708434, "grad_norm": 19.210277557373047, "learning_rate": 3.9121202274133525e-07, "loss": 0.6185, "num_input_tokens_seen": 4903141376, "step": 2338 }, { "epoch": 0.0919446704637917, "grad_norm": 15.579242706298828, "learning_rate": 3.9083436543809536e-07, "loss": 0.625, "num_input_tokens_seen": 4905238528, "step": 2339 }, { "epoch": 0.09221589368049905, "grad_norm": 15.06516170501709, "learning_rate": 3.904568362253011e-07, "loss": 0.3804, "num_input_tokens_seen": 4907335680, "step": 2340 }, { "epoch": 0.0924871168972064, "grad_norm": 15.726207733154297, "learning_rate": 3.9007943540677426e-07, "loss": 0.48, "num_input_tokens_seen": 4909432832, "step": 2341 }, { "epoch": 0.09275834011391375, "grad_norm": 16.687849044799805, "learning_rate": 3.897021632862321e-07, "loss": 0.4229, "num_input_tokens_seen": 4911529984, "step": 2342 }, { "epoch": 0.0930295633306211, "grad_norm": 13.342137336730957, "learning_rate": 3.893250201672893e-07, "loss": 0.4569, "num_input_tokens_seen": 4913627136, "step": 2343 }, { "epoch": 0.09330078654732846, "grad_norm": 14.77047061920166, "learning_rate": 3.889480063534563e-07, "loss": 0.4165, "num_input_tokens_seen": 4915724288, "step": 2344 }, { "epoch": 0.0935720097640358, "grad_norm": 10.642462730407715, "learning_rate": 3.88571122148139e-07, "loss": 0.2877, "num_input_tokens_seen": 4917821440, "step": 2345 }, { "epoch": 0.09384323298074315, "grad_norm": 13.166287422180176, "learning_rate": 3.881943678546399e-07, "loss": 0.4986, "num_input_tokens_seen": 4919918592, "step": 2346 }, { "epoch": 0.0941144561974505, "grad_norm": 25.655261993408203, "learning_rate": 3.878177437761564e-07, "loss": 0.9066, "num_input_tokens_seen": 4922015744, "step": 2347 }, { "epoch": 0.09438567941415785, "grad_norm": 17.33159828186035, "learning_rate": 3.8744125021578123e-07, "loss": 0.4746, "num_input_tokens_seen": 4924112896, "step": 2348 }, { "epoch": 0.09465690263086521, "grad_norm": 18.616378784179688, "learning_rate": 3.87064887476502e-07, "loss": 0.6, "num_input_tokens_seen": 4926210048, "step": 2349 }, { "epoch": 0.09492812584757256, "grad_norm": 16.331174850463867, "learning_rate": 3.8668865586120124e-07, "loss": 0.4109, "num_input_tokens_seen": 4928307200, "step": 2350 }, { "epoch": 0.0951993490642799, "grad_norm": 23.782737731933594, "learning_rate": 3.8631255567265573e-07, "loss": 0.5772, "num_input_tokens_seen": 4930404352, "step": 2351 }, { "epoch": 0.09547057228098725, "grad_norm": 14.399580001831055, "learning_rate": 3.859365872135367e-07, "loss": 0.408, "num_input_tokens_seen": 4932501504, "step": 2352 }, { "epoch": 0.0957417954976946, "grad_norm": 13.540172576904297, "learning_rate": 3.8556075078640925e-07, "loss": 0.4367, "num_input_tokens_seen": 4934598656, "step": 2353 }, { "epoch": 0.09601301871440195, "grad_norm": 16.12992286682129, "learning_rate": 3.85185046693732e-07, "loss": 0.5215, "num_input_tokens_seen": 4936695808, "step": 2354 }, { "epoch": 0.09628424193110931, "grad_norm": 13.96000862121582, "learning_rate": 3.8480947523785767e-07, "loss": 0.4359, "num_input_tokens_seen": 4938792960, "step": 2355 }, { "epoch": 0.09655546514781665, "grad_norm": 16.476837158203125, "learning_rate": 3.844340367210318e-07, "loss": 0.518, "num_input_tokens_seen": 4940890112, "step": 2356 }, { "epoch": 0.096826688364524, "grad_norm": 11.905656814575195, "learning_rate": 3.840587314453928e-07, "loss": 0.3644, "num_input_tokens_seen": 4942987264, "step": 2357 }, { "epoch": 0.09709791158123135, "grad_norm": 16.627532958984375, "learning_rate": 3.8368355971297204e-07, "loss": 0.3056, "num_input_tokens_seen": 4945084416, "step": 2358 }, { "epoch": 0.0973691347979387, "grad_norm": 14.9150972366333, "learning_rate": 3.8330852182569374e-07, "loss": 0.3175, "num_input_tokens_seen": 4947181568, "step": 2359 }, { "epoch": 0.09764035801464606, "grad_norm": 15.67542839050293, "learning_rate": 3.8293361808537404e-07, "loss": 0.5347, "num_input_tokens_seen": 4949278720, "step": 2360 }, { "epoch": 0.0979115812313534, "grad_norm": 18.2048397064209, "learning_rate": 3.825588487937211e-07, "loss": 0.5031, "num_input_tokens_seen": 4951375872, "step": 2361 }, { "epoch": 0.09818280444806075, "grad_norm": 11.449779510498047, "learning_rate": 3.821842142523352e-07, "loss": 0.266, "num_input_tokens_seen": 4953473024, "step": 2362 }, { "epoch": 0.0984540276647681, "grad_norm": 16.407081604003906, "learning_rate": 3.818097147627076e-07, "loss": 0.509, "num_input_tokens_seen": 4955570176, "step": 2363 }, { "epoch": 0.09872525088147545, "grad_norm": 19.54366683959961, "learning_rate": 3.8143535062622154e-07, "loss": 0.6122, "num_input_tokens_seen": 4957667328, "step": 2364 }, { "epoch": 0.09899647409818281, "grad_norm": 19.319889068603516, "learning_rate": 3.8106112214415087e-07, "loss": 0.6662, "num_input_tokens_seen": 4959764480, "step": 2365 }, { "epoch": 0.09926769731489016, "grad_norm": 14.861392974853516, "learning_rate": 3.8068702961766053e-07, "loss": 0.479, "num_input_tokens_seen": 4961861632, "step": 2366 }, { "epoch": 0.0995389205315975, "grad_norm": 8.038253784179688, "learning_rate": 3.8031307334780594e-07, "loss": 0.1862, "num_input_tokens_seen": 4963958784, "step": 2367 }, { "epoch": 0.09981014374830485, "grad_norm": 13.285365104675293, "learning_rate": 3.7993925363553294e-07, "loss": 0.3774, "num_input_tokens_seen": 4966055936, "step": 2368 }, { "epoch": 0.1000813669650122, "grad_norm": 15.662388801574707, "learning_rate": 3.795655707816772e-07, "loss": 0.5155, "num_input_tokens_seen": 4968153088, "step": 2369 }, { "epoch": 0.10035259018171956, "grad_norm": 15.881659507751465, "learning_rate": 3.791920250869645e-07, "loss": 0.5164, "num_input_tokens_seen": 4970250240, "step": 2370 }, { "epoch": 0.10062381339842691, "grad_norm": 23.03855323791504, "learning_rate": 3.7881861685201033e-07, "loss": 1.0028, "num_input_tokens_seen": 4972347392, "step": 2371 }, { "epoch": 0.10089503661513426, "grad_norm": 11.0081148147583, "learning_rate": 3.784453463773194e-07, "loss": 0.2919, "num_input_tokens_seen": 4974444544, "step": 2372 }, { "epoch": 0.1011662598318416, "grad_norm": 20.63175392150879, "learning_rate": 3.780722139632858e-07, "loss": 0.5551, "num_input_tokens_seen": 4976541696, "step": 2373 }, { "epoch": 0.10143748304854895, "grad_norm": 20.467164993286133, "learning_rate": 3.776992199101918e-07, "loss": 0.7329, "num_input_tokens_seen": 4978638848, "step": 2374 }, { "epoch": 0.10170870626525631, "grad_norm": 17.156503677368164, "learning_rate": 3.773263645182091e-07, "loss": 0.5547, "num_input_tokens_seen": 4980736000, "step": 2375 }, { "epoch": 0.10197992948196366, "grad_norm": 18.2172794342041, "learning_rate": 3.769536480873976e-07, "loss": 0.3412, "num_input_tokens_seen": 4982833152, "step": 2376 }, { "epoch": 0.10225115269867101, "grad_norm": 19.269289016723633, "learning_rate": 3.765810709177052e-07, "loss": 0.6615, "num_input_tokens_seen": 4984930304, "step": 2377 }, { "epoch": 0.10252237591537836, "grad_norm": 18.64261817932129, "learning_rate": 3.762086333089678e-07, "loss": 0.7011, "num_input_tokens_seen": 4987027456, "step": 2378 }, { "epoch": 0.1027935991320857, "grad_norm": 18.141159057617188, "learning_rate": 3.758363355609092e-07, "loss": 0.4578, "num_input_tokens_seen": 4989124608, "step": 2379 }, { "epoch": 0.10306482234879305, "grad_norm": 12.744918823242188, "learning_rate": 3.7546417797314023e-07, "loss": 0.3624, "num_input_tokens_seen": 4991221760, "step": 2380 }, { "epoch": 0.10333604556550041, "grad_norm": 17.12135887145996, "learning_rate": 3.7509216084515916e-07, "loss": 0.427, "num_input_tokens_seen": 4993318912, "step": 2381 }, { "epoch": 0.10360726878220776, "grad_norm": 16.627965927124023, "learning_rate": 3.747202844763514e-07, "loss": 0.5558, "num_input_tokens_seen": 4995416064, "step": 2382 }, { "epoch": 0.1038784919989151, "grad_norm": 19.756608963012695, "learning_rate": 3.743485491659887e-07, "loss": 0.6773, "num_input_tokens_seen": 4997513216, "step": 2383 }, { "epoch": 0.10414971521562245, "grad_norm": 14.47923755645752, "learning_rate": 3.739769552132298e-07, "loss": 0.4862, "num_input_tokens_seen": 4999610368, "step": 2384 }, { "epoch": 0.1044209384323298, "grad_norm": 12.582110404968262, "learning_rate": 3.73605502917119e-07, "loss": 0.2835, "num_input_tokens_seen": 5001707520, "step": 2385 }, { "epoch": 0.10469216164903716, "grad_norm": 15.337152481079102, "learning_rate": 3.7323419257658716e-07, "loss": 0.5735, "num_input_tokens_seen": 5003804672, "step": 2386 }, { "epoch": 0.10496338486574451, "grad_norm": 12.261470794677734, "learning_rate": 3.728630244904507e-07, "loss": 0.3295, "num_input_tokens_seen": 5005901824, "step": 2387 }, { "epoch": 0.10523460808245186, "grad_norm": 14.014972686767578, "learning_rate": 3.724919989574116e-07, "loss": 0.4343, "num_input_tokens_seen": 5007998976, "step": 2388 }, { "epoch": 0.1055058312991592, "grad_norm": 19.110898971557617, "learning_rate": 3.7212111627605704e-07, "loss": 0.783, "num_input_tokens_seen": 5010096128, "step": 2389 }, { "epoch": 0.10577705451586655, "grad_norm": 15.27743148803711, "learning_rate": 3.717503767448593e-07, "loss": 0.5481, "num_input_tokens_seen": 5012193280, "step": 2390 }, { "epoch": 0.10604827773257391, "grad_norm": 20.981355667114258, "learning_rate": 3.7137978066217555e-07, "loss": 0.5404, "num_input_tokens_seen": 5014290432, "step": 2391 }, { "epoch": 0.10631950094928126, "grad_norm": 19.111934661865234, "learning_rate": 3.710093283262472e-07, "loss": 0.7735, "num_input_tokens_seen": 5016387584, "step": 2392 }, { "epoch": 0.10659072416598861, "grad_norm": 13.102584838867188, "learning_rate": 3.706390200352003e-07, "loss": 0.3887, "num_input_tokens_seen": 5018484736, "step": 2393 }, { "epoch": 0.10686194738269596, "grad_norm": 23.91544532775879, "learning_rate": 3.7026885608704494e-07, "loss": 0.5647, "num_input_tokens_seen": 5020581888, "step": 2394 }, { "epoch": 0.1071331705994033, "grad_norm": 14.230181694030762, "learning_rate": 3.6989883677967483e-07, "loss": 0.3571, "num_input_tokens_seen": 5022679040, "step": 2395 }, { "epoch": 0.10740439381611067, "grad_norm": 14.798707008361816, "learning_rate": 3.6952896241086783e-07, "loss": 0.4492, "num_input_tokens_seen": 5024776192, "step": 2396 }, { "epoch": 0.10767561703281801, "grad_norm": 13.679465293884277, "learning_rate": 3.6915923327828423e-07, "loss": 0.3954, "num_input_tokens_seen": 5026873344, "step": 2397 }, { "epoch": 0.10794684024952536, "grad_norm": 12.650426864624023, "learning_rate": 3.6878964967946813e-07, "loss": 0.3741, "num_input_tokens_seen": 5028970496, "step": 2398 }, { "epoch": 0.10821806346623271, "grad_norm": 15.329103469848633, "learning_rate": 3.6842021191184636e-07, "loss": 0.4692, "num_input_tokens_seen": 5031067648, "step": 2399 }, { "epoch": 0.10848928668294006, "grad_norm": 20.871158599853516, "learning_rate": 3.6805092027272853e-07, "loss": 0.8217, "num_input_tokens_seen": 5033164800, "step": 2400 }, { "epoch": 0.00027122321670735016, "grad_norm": 14.983903884887695, "learning_rate": 3.6768177505930655e-07, "loss": 0.4378, "num_input_tokens_seen": 5035261952, "step": 2401 }, { "epoch": 0.0005424464334147003, "grad_norm": 17.2396240234375, "learning_rate": 3.67312776568654e-07, "loss": 0.5583, "num_input_tokens_seen": 5037359104, "step": 2402 }, { "epoch": 0.0008136696501220504, "grad_norm": 18.06374740600586, "learning_rate": 3.669439250977272e-07, "loss": 0.6734, "num_input_tokens_seen": 5039456256, "step": 2403 }, { "epoch": 0.0010848928668294006, "grad_norm": 16.376707077026367, "learning_rate": 3.665752209433637e-07, "loss": 0.5364, "num_input_tokens_seen": 5041553408, "step": 2404 }, { "epoch": 0.0013561160835367507, "grad_norm": 12.664093017578125, "learning_rate": 3.6620666440228254e-07, "loss": 0.3744, "num_input_tokens_seen": 5043650560, "step": 2405 }, { "epoch": 0.0016273393002441008, "grad_norm": 11.83484935760498, "learning_rate": 3.6583825577108397e-07, "loss": 0.2879, "num_input_tokens_seen": 5045747712, "step": 2406 }, { "epoch": 0.001898562516951451, "grad_norm": 30.771093368530273, "learning_rate": 3.654699953462494e-07, "loss": 0.5672, "num_input_tokens_seen": 5047844864, "step": 2407 }, { "epoch": 0.0021697857336588013, "grad_norm": 12.490880012512207, "learning_rate": 3.651018834241406e-07, "loss": 0.352, "num_input_tokens_seen": 5049942016, "step": 2408 }, { "epoch": 0.0024410089503661514, "grad_norm": 12.38934326171875, "learning_rate": 3.6473392030100014e-07, "loss": 0.377, "num_input_tokens_seen": 5052039168, "step": 2409 }, { "epoch": 0.0027122321670735015, "grad_norm": 15.052350044250488, "learning_rate": 3.6436610627295074e-07, "loss": 0.4124, "num_input_tokens_seen": 5054136320, "step": 2410 }, { "epoch": 0.0029834553837808516, "grad_norm": 16.288164138793945, "learning_rate": 3.639984416359949e-07, "loss": 0.6888, "num_input_tokens_seen": 5056233472, "step": 2411 }, { "epoch": 0.0032546786004882017, "grad_norm": 10.213361740112305, "learning_rate": 3.636309266860156e-07, "loss": 0.3183, "num_input_tokens_seen": 5058330624, "step": 2412 }, { "epoch": 0.003525901817195552, "grad_norm": 15.612173080444336, "learning_rate": 3.6326356171877483e-07, "loss": 0.5927, "num_input_tokens_seen": 5060427776, "step": 2413 }, { "epoch": 0.003797125033902902, "grad_norm": 15.306659698486328, "learning_rate": 3.6289634702991343e-07, "loss": 0.3537, "num_input_tokens_seen": 5062524928, "step": 2414 }, { "epoch": 0.0040683482506102524, "grad_norm": 13.14424991607666, "learning_rate": 3.625292829149521e-07, "loss": 0.2652, "num_input_tokens_seen": 5064622080, "step": 2415 }, { "epoch": 0.0043395714673176026, "grad_norm": 22.333417892456055, "learning_rate": 3.6216236966929015e-07, "loss": 0.8254, "num_input_tokens_seen": 5066719232, "step": 2416 }, { "epoch": 0.004610794684024953, "grad_norm": 15.718981742858887, "learning_rate": 3.6179560758820527e-07, "loss": 0.3925, "num_input_tokens_seen": 5068816384, "step": 2417 }, { "epoch": 0.004882017900732303, "grad_norm": 16.70728302001953, "learning_rate": 3.61428996966854e-07, "loss": 0.589, "num_input_tokens_seen": 5070913536, "step": 2418 }, { "epoch": 0.005153241117439653, "grad_norm": 12.905607223510742, "learning_rate": 3.610625381002701e-07, "loss": 0.3304, "num_input_tokens_seen": 5073010688, "step": 2419 }, { "epoch": 0.005424464334147003, "grad_norm": 15.081153869628906, "learning_rate": 3.606962312833659e-07, "loss": 0.4949, "num_input_tokens_seen": 5075107840, "step": 2420 }, { "epoch": 0.005695687550854353, "grad_norm": 13.203760147094727, "learning_rate": 3.603300768109311e-07, "loss": 0.3444, "num_input_tokens_seen": 5077204992, "step": 2421 }, { "epoch": 0.005966910767561703, "grad_norm": 21.240129470825195, "learning_rate": 3.5996407497763305e-07, "loss": 0.6882, "num_input_tokens_seen": 5079302144, "step": 2422 }, { "epoch": 0.006238133984269053, "grad_norm": 13.102483749389648, "learning_rate": 3.5959822607801617e-07, "loss": 0.3807, "num_input_tokens_seen": 5081399296, "step": 2423 }, { "epoch": 0.006509357200976403, "grad_norm": 17.356565475463867, "learning_rate": 3.592325304065018e-07, "loss": 0.5024, "num_input_tokens_seen": 5083496448, "step": 2424 }, { "epoch": 0.0067805804176837535, "grad_norm": 38.09556198120117, "learning_rate": 3.588669882573875e-07, "loss": 0.5346, "num_input_tokens_seen": 5085593600, "step": 2425 }, { "epoch": 0.007051803634391104, "grad_norm": 24.49935531616211, "learning_rate": 3.5850159992484787e-07, "loss": 0.9886, "num_input_tokens_seen": 5087690752, "step": 2426 }, { "epoch": 0.007323026851098454, "grad_norm": 15.095088005065918, "learning_rate": 3.581363657029336e-07, "loss": 0.3287, "num_input_tokens_seen": 5089787904, "step": 2427 }, { "epoch": 0.007594250067805804, "grad_norm": 8.799110412597656, "learning_rate": 3.5777128588557126e-07, "loss": 0.2119, "num_input_tokens_seen": 5091885056, "step": 2428 }, { "epoch": 0.007865473284513154, "grad_norm": 15.774507522583008, "learning_rate": 3.574063607665633e-07, "loss": 0.6325, "num_input_tokens_seen": 5093982208, "step": 2429 }, { "epoch": 0.008136696501220505, "grad_norm": 20.946643829345703, "learning_rate": 3.570415906395873e-07, "loss": 0.5039, "num_input_tokens_seen": 5096079360, "step": 2430 }, { "epoch": 0.008407919717927854, "grad_norm": 14.345913887023926, "learning_rate": 3.5667697579819655e-07, "loss": 0.4575, "num_input_tokens_seen": 5098176512, "step": 2431 }, { "epoch": 0.008679142934635205, "grad_norm": 14.178055763244629, "learning_rate": 3.563125165358193e-07, "loss": 0.5042, "num_input_tokens_seen": 5100273664, "step": 2432 }, { "epoch": 0.008950366151342554, "grad_norm": 17.350387573242188, "learning_rate": 3.559482131457583e-07, "loss": 0.5089, "num_input_tokens_seen": 5102370816, "step": 2433 }, { "epoch": 0.009221589368049905, "grad_norm": 18.10607147216797, "learning_rate": 3.5558406592119115e-07, "loss": 0.4904, "num_input_tokens_seen": 5104467968, "step": 2434 }, { "epoch": 0.009492812584757255, "grad_norm": 15.926411628723145, "learning_rate": 3.552200751551697e-07, "loss": 0.5084, "num_input_tokens_seen": 5106565120, "step": 2435 }, { "epoch": 0.009764035801464606, "grad_norm": 17.589426040649414, "learning_rate": 3.548562411406201e-07, "loss": 0.4039, "num_input_tokens_seen": 5108662272, "step": 2436 }, { "epoch": 0.010035259018171955, "grad_norm": 15.575359344482422, "learning_rate": 3.544925641703413e-07, "loss": 0.4175, "num_input_tokens_seen": 5110759424, "step": 2437 }, { "epoch": 0.010306482234879306, "grad_norm": 21.171030044555664, "learning_rate": 3.5412904453700754e-07, "loss": 0.6827, "num_input_tokens_seen": 5112856576, "step": 2438 }, { "epoch": 0.010577705451586655, "grad_norm": 18.639638900756836, "learning_rate": 3.537656825331653e-07, "loss": 0.6128, "num_input_tokens_seen": 5114953728, "step": 2439 }, { "epoch": 0.010848928668294006, "grad_norm": 19.739368438720703, "learning_rate": 3.534024784512345e-07, "loss": 0.5211, "num_input_tokens_seen": 5117050880, "step": 2440 }, { "epoch": 0.011120151885001357, "grad_norm": 14.70162296295166, "learning_rate": 3.5303943258350813e-07, "loss": 0.3884, "num_input_tokens_seen": 5119148032, "step": 2441 }, { "epoch": 0.011391375101708706, "grad_norm": 15.77631950378418, "learning_rate": 3.526765452221512e-07, "loss": 0.3623, "num_input_tokens_seen": 5121245184, "step": 2442 }, { "epoch": 0.011662598318416057, "grad_norm": 16.529884338378906, "learning_rate": 3.523138166592021e-07, "loss": 0.4567, "num_input_tokens_seen": 5123342336, "step": 2443 }, { "epoch": 0.011933821535123406, "grad_norm": 19.871736526489258, "learning_rate": 3.5195124718657075e-07, "loss": 0.7492, "num_input_tokens_seen": 5125439488, "step": 2444 }, { "epoch": 0.012205044751830757, "grad_norm": 9.847620964050293, "learning_rate": 3.5158883709603946e-07, "loss": 0.2599, "num_input_tokens_seen": 5127536640, "step": 2445 }, { "epoch": 0.012476267968538107, "grad_norm": 14.648107528686523, "learning_rate": 3.5122658667926177e-07, "loss": 0.5326, "num_input_tokens_seen": 5129633792, "step": 2446 }, { "epoch": 0.012747491185245458, "grad_norm": 14.89171314239502, "learning_rate": 3.5086449622776346e-07, "loss": 0.583, "num_input_tokens_seen": 5131730944, "step": 2447 }, { "epoch": 0.013018714401952807, "grad_norm": 14.507434844970703, "learning_rate": 3.505025660329408e-07, "loss": 0.4423, "num_input_tokens_seen": 5133828096, "step": 2448 }, { "epoch": 0.013289937618660158, "grad_norm": 13.03704833984375, "learning_rate": 3.5014079638606164e-07, "loss": 0.3409, "num_input_tokens_seen": 5135925248, "step": 2449 }, { "epoch": 0.013561160835367507, "grad_norm": 12.08354377746582, "learning_rate": 3.497791875782643e-07, "loss": 0.3119, "num_input_tokens_seen": 5138022400, "step": 2450 }, { "epoch": 0.013832384052074858, "grad_norm": 14.133995056152344, "learning_rate": 3.4941773990055777e-07, "loss": 0.404, "num_input_tokens_seen": 5140119552, "step": 2451 }, { "epoch": 0.014103607268782207, "grad_norm": 19.439821243286133, "learning_rate": 3.490564536438215e-07, "loss": 0.6819, "num_input_tokens_seen": 5142216704, "step": 2452 }, { "epoch": 0.014374830485489558, "grad_norm": 14.30314826965332, "learning_rate": 3.4869532909880485e-07, "loss": 0.4549, "num_input_tokens_seen": 5144313856, "step": 2453 }, { "epoch": 0.014646053702196907, "grad_norm": 15.834941864013672, "learning_rate": 3.483343665561271e-07, "loss": 0.3616, "num_input_tokens_seen": 5146411008, "step": 2454 }, { "epoch": 0.014917276918904258, "grad_norm": 17.570751190185547, "learning_rate": 3.479735663062773e-07, "loss": 0.4955, "num_input_tokens_seen": 5148508160, "step": 2455 }, { "epoch": 0.015188500135611608, "grad_norm": 11.63994026184082, "learning_rate": 3.4761292863961354e-07, "loss": 0.357, "num_input_tokens_seen": 5150605312, "step": 2456 }, { "epoch": 0.015459723352318959, "grad_norm": 13.828446388244629, "learning_rate": 3.4725245384636347e-07, "loss": 0.3693, "num_input_tokens_seen": 5152702464, "step": 2457 }, { "epoch": 0.015730946569026308, "grad_norm": 13.007326126098633, "learning_rate": 3.4689214221662364e-07, "loss": 0.2658, "num_input_tokens_seen": 5154799616, "step": 2458 }, { "epoch": 0.01600216978573366, "grad_norm": 15.117969512939453, "learning_rate": 3.465319940403587e-07, "loss": 0.5309, "num_input_tokens_seen": 5156896768, "step": 2459 }, { "epoch": 0.01627339300244101, "grad_norm": 12.923035621643066, "learning_rate": 3.4617200960740247e-07, "loss": 0.3563, "num_input_tokens_seen": 5158993920, "step": 2460 }, { "epoch": 0.01654461621914836, "grad_norm": 10.716797828674316, "learning_rate": 3.4581218920745663e-07, "loss": 0.3419, "num_input_tokens_seen": 5161091072, "step": 2461 }, { "epoch": 0.016815839435855708, "grad_norm": 14.895060539245605, "learning_rate": 3.454525331300908e-07, "loss": 0.3487, "num_input_tokens_seen": 5163188224, "step": 2462 }, { "epoch": 0.01708706265256306, "grad_norm": 17.449199676513672, "learning_rate": 3.450930416647429e-07, "loss": 0.4316, "num_input_tokens_seen": 5165285376, "step": 2463 }, { "epoch": 0.01735828586927041, "grad_norm": 13.589836120605469, "learning_rate": 3.4473371510071795e-07, "loss": 0.458, "num_input_tokens_seen": 5167382528, "step": 2464 }, { "epoch": 0.01762950908597776, "grad_norm": 17.81792640686035, "learning_rate": 3.4437455372718795e-07, "loss": 0.6118, "num_input_tokens_seen": 5169479680, "step": 2465 }, { "epoch": 0.01790073230268511, "grad_norm": 15.582667350769043, "learning_rate": 3.440155578331925e-07, "loss": 0.4143, "num_input_tokens_seen": 5171576832, "step": 2466 }, { "epoch": 0.01817195551939246, "grad_norm": 13.768962860107422, "learning_rate": 3.4365672770763783e-07, "loss": 0.3377, "num_input_tokens_seen": 5173673984, "step": 2467 }, { "epoch": 0.01844317873609981, "grad_norm": 14.987922668457031, "learning_rate": 3.432980636392967e-07, "loss": 0.4133, "num_input_tokens_seen": 5175771136, "step": 2468 }, { "epoch": 0.01871440195280716, "grad_norm": 14.550777435302734, "learning_rate": 3.429395659168084e-07, "loss": 0.3833, "num_input_tokens_seen": 5177868288, "step": 2469 }, { "epoch": 0.01898562516951451, "grad_norm": 17.40359115600586, "learning_rate": 3.425812348286782e-07, "loss": 0.5748, "num_input_tokens_seen": 5179965440, "step": 2470 }, { "epoch": 0.01925684838622186, "grad_norm": 21.7893123626709, "learning_rate": 3.422230706632774e-07, "loss": 0.8634, "num_input_tokens_seen": 5182062592, "step": 2471 }, { "epoch": 0.01952807160292921, "grad_norm": 17.623117446899414, "learning_rate": 3.418650737088427e-07, "loss": 0.4959, "num_input_tokens_seen": 5184159744, "step": 2472 }, { "epoch": 0.019799294819636562, "grad_norm": 15.7328519821167, "learning_rate": 3.415072442534767e-07, "loss": 0.3269, "num_input_tokens_seen": 5186256896, "step": 2473 }, { "epoch": 0.02007051803634391, "grad_norm": 14.757015228271484, "learning_rate": 3.411495825851467e-07, "loss": 0.3758, "num_input_tokens_seen": 5188354048, "step": 2474 }, { "epoch": 0.02034174125305126, "grad_norm": 12.561208724975586, "learning_rate": 3.4079208899168545e-07, "loss": 0.3777, "num_input_tokens_seen": 5190451200, "step": 2475 }, { "epoch": 0.02061296446975861, "grad_norm": 20.02168083190918, "learning_rate": 3.404347637607899e-07, "loss": 0.4442, "num_input_tokens_seen": 5192548352, "step": 2476 }, { "epoch": 0.020884187686465962, "grad_norm": 16.987884521484375, "learning_rate": 3.400776071800219e-07, "loss": 0.5238, "num_input_tokens_seen": 5194645504, "step": 2477 }, { "epoch": 0.02115541090317331, "grad_norm": 16.009906768798828, "learning_rate": 3.3972061953680734e-07, "loss": 0.2866, "num_input_tokens_seen": 5196742656, "step": 2478 }, { "epoch": 0.02142663411988066, "grad_norm": 15.513143539428711, "learning_rate": 3.3936380111843666e-07, "loss": 0.4394, "num_input_tokens_seen": 5198839808, "step": 2479 }, { "epoch": 0.021697857336588012, "grad_norm": 12.317339897155762, "learning_rate": 3.390071522120635e-07, "loss": 0.3811, "num_input_tokens_seen": 5200936960, "step": 2480 }, { "epoch": 0.021969080553295363, "grad_norm": 13.320518493652344, "learning_rate": 3.3865067310470554e-07, "loss": 0.3729, "num_input_tokens_seen": 5203034112, "step": 2481 }, { "epoch": 0.022240303770002714, "grad_norm": 11.673955917358398, "learning_rate": 3.3829436408324316e-07, "loss": 0.3577, "num_input_tokens_seen": 5205131264, "step": 2482 }, { "epoch": 0.02251152698671006, "grad_norm": 15.189332962036133, "learning_rate": 3.3793822543442074e-07, "loss": 0.4539, "num_input_tokens_seen": 5207228416, "step": 2483 }, { "epoch": 0.022782750203417412, "grad_norm": 13.745153427124023, "learning_rate": 3.3758225744484483e-07, "loss": 0.3885, "num_input_tokens_seen": 5209325568, "step": 2484 }, { "epoch": 0.023053973420124763, "grad_norm": 13.338912963867188, "learning_rate": 3.372264604009851e-07, "loss": 0.4417, "num_input_tokens_seen": 5211422720, "step": 2485 }, { "epoch": 0.023325196636832114, "grad_norm": 15.356630325317383, "learning_rate": 3.3687083458917344e-07, "loss": 0.4507, "num_input_tokens_seen": 5213519872, "step": 2486 }, { "epoch": 0.023596419853539462, "grad_norm": 16.816343307495117, "learning_rate": 3.3651538029560377e-07, "loss": 0.4607, "num_input_tokens_seen": 5215617024, "step": 2487 }, { "epoch": 0.023867643070246813, "grad_norm": 13.717020034790039, "learning_rate": 3.361600978063325e-07, "loss": 0.4352, "num_input_tokens_seen": 5217714176, "step": 2488 }, { "epoch": 0.024138866286954164, "grad_norm": 17.89053726196289, "learning_rate": 3.358049874072771e-07, "loss": 0.5094, "num_input_tokens_seen": 5219811328, "step": 2489 }, { "epoch": 0.024410089503661515, "grad_norm": 13.972254753112793, "learning_rate": 3.3545004938421734e-07, "loss": 0.4116, "num_input_tokens_seen": 5221908480, "step": 2490 }, { "epoch": 0.024681312720368862, "grad_norm": 20.813188552856445, "learning_rate": 3.3509528402279357e-07, "loss": 0.7909, "num_input_tokens_seen": 5224005632, "step": 2491 }, { "epoch": 0.024952535937076213, "grad_norm": 23.339984893798828, "learning_rate": 3.347406916085074e-07, "loss": 0.9871, "num_input_tokens_seen": 5226102784, "step": 2492 }, { "epoch": 0.025223759153783564, "grad_norm": 19.53771209716797, "learning_rate": 3.3438627242672164e-07, "loss": 0.7506, "num_input_tokens_seen": 5228199936, "step": 2493 }, { "epoch": 0.025494982370490915, "grad_norm": 13.463472366333008, "learning_rate": 3.3403202676265875e-07, "loss": 0.4105, "num_input_tokens_seen": 5230297088, "step": 2494 }, { "epoch": 0.025766205587198263, "grad_norm": 11.80983829498291, "learning_rate": 3.336779549014026e-07, "loss": 0.2789, "num_input_tokens_seen": 5232394240, "step": 2495 }, { "epoch": 0.026037428803905614, "grad_norm": 11.199295997619629, "learning_rate": 3.333240571278968e-07, "loss": 0.3236, "num_input_tokens_seen": 5234491392, "step": 2496 }, { "epoch": 0.026308652020612965, "grad_norm": 11.20341968536377, "learning_rate": 3.3297033372694473e-07, "loss": 0.2513, "num_input_tokens_seen": 5236588544, "step": 2497 }, { "epoch": 0.026579875237320316, "grad_norm": 19.574689865112305, "learning_rate": 3.3261678498320954e-07, "loss": 0.8343, "num_input_tokens_seen": 5238685696, "step": 2498 }, { "epoch": 0.026851098454027666, "grad_norm": 17.769323348999023, "learning_rate": 3.3226341118121367e-07, "loss": 0.6071, "num_input_tokens_seen": 5240782848, "step": 2499 }, { "epoch": 0.027122321670735014, "grad_norm": 10.659870147705078, "learning_rate": 3.319102126053389e-07, "loss": 0.3469, "num_input_tokens_seen": 5242880000, "step": 2500 }, { "epoch": 0.027393544887442365, "grad_norm": 14.309873580932617, "learning_rate": 3.315571895398261e-07, "loss": 0.2899, "num_input_tokens_seen": 5244977152, "step": 2501 }, { "epoch": 0.027664768104149716, "grad_norm": 12.785883903503418, "learning_rate": 3.312043422687749e-07, "loss": 0.3751, "num_input_tokens_seen": 5247074304, "step": 2502 }, { "epoch": 0.027935991320857067, "grad_norm": 9.752639770507812, "learning_rate": 3.3085167107614297e-07, "loss": 0.2094, "num_input_tokens_seen": 5249171456, "step": 2503 }, { "epoch": 0.028207214537564414, "grad_norm": 20.031442642211914, "learning_rate": 3.3049917624574737e-07, "loss": 0.6936, "num_input_tokens_seen": 5251268608, "step": 2504 }, { "epoch": 0.028478437754271765, "grad_norm": 12.938445091247559, "learning_rate": 3.301468580612619e-07, "loss": 0.3843, "num_input_tokens_seen": 5253365760, "step": 2505 }, { "epoch": 0.028749660970979116, "grad_norm": 15.744908332824707, "learning_rate": 3.2979471680621903e-07, "loss": 0.3618, "num_input_tokens_seen": 5255462912, "step": 2506 }, { "epoch": 0.029020884187686467, "grad_norm": 18.101648330688477, "learning_rate": 3.2944275276400857e-07, "loss": 0.6241, "num_input_tokens_seen": 5257560064, "step": 2507 }, { "epoch": 0.029292107404393815, "grad_norm": 15.829540252685547, "learning_rate": 3.290909662178779e-07, "loss": 0.4732, "num_input_tokens_seen": 5259657216, "step": 2508 }, { "epoch": 0.029563330621101166, "grad_norm": 10.271330833435059, "learning_rate": 3.2873935745093145e-07, "loss": 0.1894, "num_input_tokens_seen": 5261754368, "step": 2509 }, { "epoch": 0.029834553837808517, "grad_norm": 14.364543914794922, "learning_rate": 3.283879267461305e-07, "loss": 0.4071, "num_input_tokens_seen": 5263851520, "step": 2510 }, { "epoch": 0.030105777054515868, "grad_norm": 14.449692726135254, "learning_rate": 3.280366743862931e-07, "loss": 0.2794, "num_input_tokens_seen": 5265948672, "step": 2511 }, { "epoch": 0.030377000271223215, "grad_norm": 14.511991500854492, "learning_rate": 3.276856006540939e-07, "loss": 0.4112, "num_input_tokens_seen": 5268045824, "step": 2512 }, { "epoch": 0.030648223487930566, "grad_norm": 19.19399642944336, "learning_rate": 3.2733470583206357e-07, "loss": 0.7822, "num_input_tokens_seen": 5270142976, "step": 2513 }, { "epoch": 0.030919446704637917, "grad_norm": 23.72905158996582, "learning_rate": 3.2698399020258895e-07, "loss": 0.7289, "num_input_tokens_seen": 5272240128, "step": 2514 }, { "epoch": 0.031190669921345268, "grad_norm": 12.964669227600098, "learning_rate": 3.266334540479128e-07, "loss": 0.3558, "num_input_tokens_seen": 5274337280, "step": 2515 }, { "epoch": 0.031461893138052616, "grad_norm": 16.30192756652832, "learning_rate": 3.262830976501329e-07, "loss": 0.4661, "num_input_tokens_seen": 5276434432, "step": 2516 }, { "epoch": 0.03173311635475997, "grad_norm": 17.11813735961914, "learning_rate": 3.2593292129120295e-07, "loss": 0.4023, "num_input_tokens_seen": 5278531584, "step": 2517 }, { "epoch": 0.03200433957146732, "grad_norm": 13.747211456298828, "learning_rate": 3.2558292525293156e-07, "loss": 0.396, "num_input_tokens_seen": 5280628736, "step": 2518 }, { "epoch": 0.032275562788174665, "grad_norm": 15.1743803024292, "learning_rate": 3.2523310981698213e-07, "loss": 0.4891, "num_input_tokens_seen": 5282725888, "step": 2519 }, { "epoch": 0.03254678600488202, "grad_norm": 21.062551498413086, "learning_rate": 3.24883475264873e-07, "loss": 0.7359, "num_input_tokens_seen": 5284823040, "step": 2520 }, { "epoch": 0.03281800922158937, "grad_norm": 12.402223587036133, "learning_rate": 3.2453402187797684e-07, "loss": 0.5071, "num_input_tokens_seen": 5286920192, "step": 2521 }, { "epoch": 0.03308923243829672, "grad_norm": 16.109939575195312, "learning_rate": 3.241847499375201e-07, "loss": 0.4946, "num_input_tokens_seen": 5289017344, "step": 2522 }, { "epoch": 0.03336045565500407, "grad_norm": 10.402777671813965, "learning_rate": 3.238356597245837e-07, "loss": 0.2449, "num_input_tokens_seen": 5291114496, "step": 2523 }, { "epoch": 0.033631678871711417, "grad_norm": 11.929893493652344, "learning_rate": 3.2348675152010217e-07, "loss": 0.4098, "num_input_tokens_seen": 5293211648, "step": 2524 }, { "epoch": 0.03390290208841877, "grad_norm": 16.611042022705078, "learning_rate": 3.2313802560486353e-07, "loss": 0.5525, "num_input_tokens_seen": 5295308800, "step": 2525 }, { "epoch": 0.03417412530512612, "grad_norm": 18.180625915527344, "learning_rate": 3.2278948225950916e-07, "loss": 0.654, "num_input_tokens_seen": 5297405952, "step": 2526 }, { "epoch": 0.034445348521833466, "grad_norm": 13.363505363464355, "learning_rate": 3.2244112176453343e-07, "loss": 0.4729, "num_input_tokens_seen": 5299503104, "step": 2527 }, { "epoch": 0.03471657173854082, "grad_norm": 20.741127014160156, "learning_rate": 3.2209294440028366e-07, "loss": 0.7143, "num_input_tokens_seen": 5301600256, "step": 2528 }, { "epoch": 0.03498779495524817, "grad_norm": 13.14611530303955, "learning_rate": 3.2174495044695973e-07, "loss": 0.3495, "num_input_tokens_seen": 5303697408, "step": 2529 }, { "epoch": 0.03525901817195552, "grad_norm": 16.342422485351562, "learning_rate": 3.2139714018461396e-07, "loss": 0.5414, "num_input_tokens_seen": 5305794560, "step": 2530 }, { "epoch": 0.03553024138866287, "grad_norm": 16.03512191772461, "learning_rate": 3.2104951389315073e-07, "loss": 0.5638, "num_input_tokens_seen": 5307891712, "step": 2531 }, { "epoch": 0.03580146460537022, "grad_norm": 17.44066047668457, "learning_rate": 3.207020718523266e-07, "loss": 0.6175, "num_input_tokens_seen": 5309988864, "step": 2532 }, { "epoch": 0.03607268782207757, "grad_norm": 13.835807800292969, "learning_rate": 3.2035481434174966e-07, "loss": 0.4456, "num_input_tokens_seen": 5312086016, "step": 2533 }, { "epoch": 0.03634391103878492, "grad_norm": 16.608179092407227, "learning_rate": 3.200077416408794e-07, "loss": 0.4694, "num_input_tokens_seen": 5314183168, "step": 2534 }, { "epoch": 0.03661513425549227, "grad_norm": 23.50665855407715, "learning_rate": 3.196608540290266e-07, "loss": 0.597, "num_input_tokens_seen": 5316280320, "step": 2535 }, { "epoch": 0.03688635747219962, "grad_norm": 19.46283721923828, "learning_rate": 3.193141517853536e-07, "loss": 0.5605, "num_input_tokens_seen": 5318377472, "step": 2536 }, { "epoch": 0.03715758068890697, "grad_norm": 14.471436500549316, "learning_rate": 3.1896763518887305e-07, "loss": 0.4885, "num_input_tokens_seen": 5320474624, "step": 2537 }, { "epoch": 0.03742880390561432, "grad_norm": 22.120628356933594, "learning_rate": 3.186213045184484e-07, "loss": 0.8826, "num_input_tokens_seen": 5322571776, "step": 2538 }, { "epoch": 0.03770002712232167, "grad_norm": 11.51758861541748, "learning_rate": 3.1827516005279306e-07, "loss": 0.3493, "num_input_tokens_seen": 5324668928, "step": 2539 }, { "epoch": 0.03797125033902902, "grad_norm": 14.106184959411621, "learning_rate": 3.1792920207047114e-07, "loss": 0.4086, "num_input_tokens_seen": 5326766080, "step": 2540 }, { "epoch": 0.03824247355573637, "grad_norm": 15.168310165405273, "learning_rate": 3.175834308498964e-07, "loss": 0.4476, "num_input_tokens_seen": 5328863232, "step": 2541 }, { "epoch": 0.03851369677244372, "grad_norm": 16.02938461303711, "learning_rate": 3.172378466693325e-07, "loss": 0.3896, "num_input_tokens_seen": 5330960384, "step": 2542 }, { "epoch": 0.038784919989151075, "grad_norm": 21.418581008911133, "learning_rate": 3.168924498068923e-07, "loss": 0.9796, "num_input_tokens_seen": 5333057536, "step": 2543 }, { "epoch": 0.03905614320585842, "grad_norm": 16.576860427856445, "learning_rate": 3.1654724054053805e-07, "loss": 0.4524, "num_input_tokens_seen": 5335154688, "step": 2544 }, { "epoch": 0.03932736642256577, "grad_norm": 14.02090072631836, "learning_rate": 3.1620221914808115e-07, "loss": 0.4474, "num_input_tokens_seen": 5337251840, "step": 2545 }, { "epoch": 0.039598589639273124, "grad_norm": 14.15399169921875, "learning_rate": 3.1585738590718157e-07, "loss": 0.3879, "num_input_tokens_seen": 5339348992, "step": 2546 }, { "epoch": 0.03986981285598047, "grad_norm": 21.25953483581543, "learning_rate": 3.1551274109534805e-07, "loss": 0.6324, "num_input_tokens_seen": 5341446144, "step": 2547 }, { "epoch": 0.04014103607268782, "grad_norm": 17.72227668762207, "learning_rate": 3.151682849899376e-07, "loss": 0.7032, "num_input_tokens_seen": 5343543296, "step": 2548 }, { "epoch": 0.040412259289395173, "grad_norm": 14.34379768371582, "learning_rate": 3.148240178681553e-07, "loss": 0.44, "num_input_tokens_seen": 5345640448, "step": 2549 }, { "epoch": 0.04068348250610252, "grad_norm": 18.565998077392578, "learning_rate": 3.1447994000705456e-07, "loss": 0.6165, "num_input_tokens_seen": 5347737600, "step": 2550 }, { "epoch": 0.040954705722809875, "grad_norm": 11.777302742004395, "learning_rate": 3.141360516835356e-07, "loss": 0.3904, "num_input_tokens_seen": 5349834752, "step": 2551 }, { "epoch": 0.04122592893951722, "grad_norm": 7.871033191680908, "learning_rate": 3.1379235317434703e-07, "loss": 0.1931, "num_input_tokens_seen": 5351931904, "step": 2552 }, { "epoch": 0.04149715215622457, "grad_norm": 13.070982933044434, "learning_rate": 3.134488447560843e-07, "loss": 0.3965, "num_input_tokens_seen": 5354029056, "step": 2553 }, { "epoch": 0.041768375372931925, "grad_norm": 19.49721908569336, "learning_rate": 3.1310552670518987e-07, "loss": 0.8075, "num_input_tokens_seen": 5356126208, "step": 2554 }, { "epoch": 0.04203959858963927, "grad_norm": 15.62477970123291, "learning_rate": 3.127623992979532e-07, "loss": 0.4351, "num_input_tokens_seen": 5358223360, "step": 2555 }, { "epoch": 0.04231082180634662, "grad_norm": 13.880343437194824, "learning_rate": 3.124194628105098e-07, "loss": 0.3508, "num_input_tokens_seen": 5360320512, "step": 2556 }, { "epoch": 0.042582045023053974, "grad_norm": 17.306838989257812, "learning_rate": 3.120767175188422e-07, "loss": 0.5777, "num_input_tokens_seen": 5362417664, "step": 2557 }, { "epoch": 0.04285326823976132, "grad_norm": 18.195343017578125, "learning_rate": 3.1173416369877864e-07, "loss": 0.7024, "num_input_tokens_seen": 5364514816, "step": 2558 }, { "epoch": 0.043124491456468676, "grad_norm": 16.49274444580078, "learning_rate": 3.1139180162599346e-07, "loss": 0.519, "num_input_tokens_seen": 5366611968, "step": 2559 }, { "epoch": 0.043395714673176024, "grad_norm": 12.536412239074707, "learning_rate": 3.110496315760065e-07, "loss": 0.3714, "num_input_tokens_seen": 5368709120, "step": 2560 }, { "epoch": 0.04366693788988337, "grad_norm": 14.73096752166748, "learning_rate": 3.107076538241835e-07, "loss": 0.5109, "num_input_tokens_seen": 5370806272, "step": 2561 }, { "epoch": 0.043938161106590726, "grad_norm": 12.924612998962402, "learning_rate": 3.103658686457349e-07, "loss": 0.4013, "num_input_tokens_seen": 5372903424, "step": 2562 }, { "epoch": 0.04420938432329807, "grad_norm": 13.50550365447998, "learning_rate": 3.1002427631571646e-07, "loss": 0.3039, "num_input_tokens_seen": 5375000576, "step": 2563 }, { "epoch": 0.04448060754000543, "grad_norm": 12.640801429748535, "learning_rate": 3.0968287710902866e-07, "loss": 0.3334, "num_input_tokens_seen": 5377097728, "step": 2564 }, { "epoch": 0.044751830756712775, "grad_norm": 15.517276763916016, "learning_rate": 3.0934167130041666e-07, "loss": 0.6337, "num_input_tokens_seen": 5379194880, "step": 2565 }, { "epoch": 0.04502305397342012, "grad_norm": 16.423845291137695, "learning_rate": 3.090006591644698e-07, "loss": 0.637, "num_input_tokens_seen": 5381292032, "step": 2566 }, { "epoch": 0.04529427719012748, "grad_norm": 13.07593822479248, "learning_rate": 3.0865984097562183e-07, "loss": 0.2506, "num_input_tokens_seen": 5383389184, "step": 2567 }, { "epoch": 0.045565500406834825, "grad_norm": 21.440753936767578, "learning_rate": 3.083192170081501e-07, "loss": 0.5519, "num_input_tokens_seen": 5385486336, "step": 2568 }, { "epoch": 0.04583672362354217, "grad_norm": 14.802675247192383, "learning_rate": 3.079787875361759e-07, "loss": 0.476, "num_input_tokens_seen": 5387583488, "step": 2569 }, { "epoch": 0.04610794684024953, "grad_norm": 17.272769927978516, "learning_rate": 3.0763855283366386e-07, "loss": 0.4442, "num_input_tokens_seen": 5389680640, "step": 2570 }, { "epoch": 0.046379170056956874, "grad_norm": 10.331268310546875, "learning_rate": 3.07298513174422e-07, "loss": 0.3056, "num_input_tokens_seen": 5391777792, "step": 2571 }, { "epoch": 0.04665039327366423, "grad_norm": 15.530325889587402, "learning_rate": 3.0695866883210143e-07, "loss": 0.504, "num_input_tokens_seen": 5393874944, "step": 2572 }, { "epoch": 0.046921616490371576, "grad_norm": 20.427730560302734, "learning_rate": 3.0661902008019556e-07, "loss": 0.7343, "num_input_tokens_seen": 5395972096, "step": 2573 }, { "epoch": 0.047192839707078924, "grad_norm": 22.61469078063965, "learning_rate": 3.062795671920411e-07, "loss": 0.8057, "num_input_tokens_seen": 5398069248, "step": 2574 }, { "epoch": 0.04746406292378628, "grad_norm": 11.421932220458984, "learning_rate": 3.059403104408166e-07, "loss": 0.3017, "num_input_tokens_seen": 5400166400, "step": 2575 }, { "epoch": 0.047735286140493625, "grad_norm": 11.474665641784668, "learning_rate": 3.0560125009954296e-07, "loss": 0.2996, "num_input_tokens_seen": 5402263552, "step": 2576 }, { "epoch": 0.04800650935720097, "grad_norm": 26.75455665588379, "learning_rate": 3.0526238644108335e-07, "loss": 0.5857, "num_input_tokens_seen": 5404360704, "step": 2577 }, { "epoch": 0.04827773257390833, "grad_norm": 15.098592758178711, "learning_rate": 3.049237197381424e-07, "loss": 0.5197, "num_input_tokens_seen": 5406457856, "step": 2578 }, { "epoch": 0.048548955790615675, "grad_norm": 15.30495548248291, "learning_rate": 3.045852502632657e-07, "loss": 0.5844, "num_input_tokens_seen": 5408555008, "step": 2579 }, { "epoch": 0.04882017900732303, "grad_norm": 8.980935096740723, "learning_rate": 3.042469782888409e-07, "loss": 0.216, "num_input_tokens_seen": 5410652160, "step": 2580 }, { "epoch": 0.04909140222403038, "grad_norm": 20.313331604003906, "learning_rate": 3.0390890408709645e-07, "loss": 0.4768, "num_input_tokens_seen": 5412749312, "step": 2581 }, { "epoch": 0.049362625440737724, "grad_norm": 11.710707664489746, "learning_rate": 3.0357102793010145e-07, "loss": 0.3278, "num_input_tokens_seen": 5414846464, "step": 2582 }, { "epoch": 0.04963384865744508, "grad_norm": 11.84463119506836, "learning_rate": 3.032333500897659e-07, "loss": 0.209, "num_input_tokens_seen": 5416943616, "step": 2583 }, { "epoch": 0.049905071874152426, "grad_norm": 15.556648254394531, "learning_rate": 3.0289587083784e-07, "loss": 0.4434, "num_input_tokens_seen": 5419040768, "step": 2584 }, { "epoch": 0.05017629509085978, "grad_norm": 10.980393409729004, "learning_rate": 3.0255859044591425e-07, "loss": 0.2918, "num_input_tokens_seen": 5421137920, "step": 2585 }, { "epoch": 0.05044751830756713, "grad_norm": 9.970746994018555, "learning_rate": 3.02221509185419e-07, "loss": 0.2593, "num_input_tokens_seen": 5423235072, "step": 2586 }, { "epoch": 0.050718741524274476, "grad_norm": 12.755905151367188, "learning_rate": 3.0188462732762457e-07, "loss": 0.4138, "num_input_tokens_seen": 5425332224, "step": 2587 }, { "epoch": 0.05098996474098183, "grad_norm": 15.958523750305176, "learning_rate": 3.015479451436406e-07, "loss": 0.4157, "num_input_tokens_seen": 5427429376, "step": 2588 }, { "epoch": 0.05126118795768918, "grad_norm": 16.09610366821289, "learning_rate": 3.0121146290441593e-07, "loss": 0.453, "num_input_tokens_seen": 5429526528, "step": 2589 }, { "epoch": 0.051532411174396525, "grad_norm": 18.44511604309082, "learning_rate": 3.0087518088073905e-07, "loss": 0.675, "num_input_tokens_seen": 5431623680, "step": 2590 }, { "epoch": 0.05180363439110388, "grad_norm": 15.760098457336426, "learning_rate": 3.005390993432366e-07, "loss": 0.4527, "num_input_tokens_seen": 5433720832, "step": 2591 }, { "epoch": 0.05207485760781123, "grad_norm": 25.40867805480957, "learning_rate": 3.00203218562374e-07, "loss": 0.398, "num_input_tokens_seen": 5435817984, "step": 2592 }, { "epoch": 0.05234608082451858, "grad_norm": 11.827055931091309, "learning_rate": 2.9986753880845596e-07, "loss": 0.2695, "num_input_tokens_seen": 5437915136, "step": 2593 }, { "epoch": 0.05261730404122593, "grad_norm": 11.77489185333252, "learning_rate": 2.9953206035162433e-07, "loss": 0.2848, "num_input_tokens_seen": 5440012288, "step": 2594 }, { "epoch": 0.05288852725793328, "grad_norm": 13.155770301818848, "learning_rate": 2.991967834618597e-07, "loss": 0.3632, "num_input_tokens_seen": 5442109440, "step": 2595 }, { "epoch": 0.05315975047464063, "grad_norm": 10.677423477172852, "learning_rate": 2.9886170840897977e-07, "loss": 0.3251, "num_input_tokens_seen": 5444206592, "step": 2596 }, { "epoch": 0.05343097369134798, "grad_norm": 15.714378356933594, "learning_rate": 2.9852683546264047e-07, "loss": 0.4266, "num_input_tokens_seen": 5446303744, "step": 2597 }, { "epoch": 0.05370219690805533, "grad_norm": 21.693063735961914, "learning_rate": 2.9819216489233467e-07, "loss": 0.7394, "num_input_tokens_seen": 5448400896, "step": 2598 }, { "epoch": 0.05397342012476268, "grad_norm": 11.83186149597168, "learning_rate": 2.978576969673926e-07, "loss": 0.3142, "num_input_tokens_seen": 5450498048, "step": 2599 }, { "epoch": 0.05424464334147003, "grad_norm": 14.470967292785645, "learning_rate": 2.9752343195698125e-07, "loss": 0.4464, "num_input_tokens_seen": 5452595200, "step": 2600 }, { "epoch": 0.05451586655817738, "grad_norm": 14.670758247375488, "learning_rate": 2.9718937013010444e-07, "loss": 0.354, "num_input_tokens_seen": 5454692352, "step": 2601 }, { "epoch": 0.05478708977488473, "grad_norm": 7.69711971282959, "learning_rate": 2.9685551175560235e-07, "loss": 0.2038, "num_input_tokens_seen": 5456789504, "step": 2602 }, { "epoch": 0.05505831299159208, "grad_norm": 17.91798973083496, "learning_rate": 2.965218571021516e-07, "loss": 0.7645, "num_input_tokens_seen": 5458886656, "step": 2603 }, { "epoch": 0.05532953620829943, "grad_norm": 14.40876579284668, "learning_rate": 2.9618840643826464e-07, "loss": 0.4033, "num_input_tokens_seen": 5460983808, "step": 2604 }, { "epoch": 0.05560075942500678, "grad_norm": 17.744657516479492, "learning_rate": 2.958551600322899e-07, "loss": 0.5836, "num_input_tokens_seen": 5463080960, "step": 2605 }, { "epoch": 0.055871982641714134, "grad_norm": 14.996540069580078, "learning_rate": 2.9552211815241156e-07, "loss": 0.4728, "num_input_tokens_seen": 5465178112, "step": 2606 }, { "epoch": 0.05614320585842148, "grad_norm": 13.542115211486816, "learning_rate": 2.9518928106664897e-07, "loss": 0.3398, "num_input_tokens_seen": 5467275264, "step": 2607 }, { "epoch": 0.05641442907512883, "grad_norm": 17.0628719329834, "learning_rate": 2.9485664904285643e-07, "loss": 0.5308, "num_input_tokens_seen": 5469372416, "step": 2608 }, { "epoch": 0.05668565229183618, "grad_norm": 16.092180252075195, "learning_rate": 2.94524222348724e-07, "loss": 0.4883, "num_input_tokens_seen": 5471469568, "step": 2609 }, { "epoch": 0.05695687550854353, "grad_norm": 16.60764503479004, "learning_rate": 2.9419200125177585e-07, "loss": 0.4317, "num_input_tokens_seen": 5473566720, "step": 2610 }, { "epoch": 0.05722809872525088, "grad_norm": 9.960092544555664, "learning_rate": 2.938599860193709e-07, "loss": 0.2807, "num_input_tokens_seen": 5475663872, "step": 2611 }, { "epoch": 0.05749932194195823, "grad_norm": 8.7996187210083, "learning_rate": 2.935281769187025e-07, "loss": 0.198, "num_input_tokens_seen": 5477761024, "step": 2612 }, { "epoch": 0.05777054515866558, "grad_norm": 16.473575592041016, "learning_rate": 2.9319657421679757e-07, "loss": 0.4382, "num_input_tokens_seen": 5479858176, "step": 2613 }, { "epoch": 0.058041768375372935, "grad_norm": 13.767109870910645, "learning_rate": 2.928651781805177e-07, "loss": 0.3479, "num_input_tokens_seen": 5481955328, "step": 2614 }, { "epoch": 0.05831299159208028, "grad_norm": 15.12543773651123, "learning_rate": 2.9253398907655775e-07, "loss": 0.2618, "num_input_tokens_seen": 5484052480, "step": 2615 }, { "epoch": 0.05858421480878763, "grad_norm": 10.026293754577637, "learning_rate": 2.9220300717144597e-07, "loss": 0.2574, "num_input_tokens_seen": 5486149632, "step": 2616 }, { "epoch": 0.058855438025494984, "grad_norm": 10.266706466674805, "learning_rate": 2.9187223273154406e-07, "loss": 0.2729, "num_input_tokens_seen": 5488246784, "step": 2617 }, { "epoch": 0.05912666124220233, "grad_norm": 16.52318000793457, "learning_rate": 2.91541666023047e-07, "loss": 0.5708, "num_input_tokens_seen": 5490343936, "step": 2618 }, { "epoch": 0.059397884458909686, "grad_norm": 23.888343811035156, "learning_rate": 2.9121130731198204e-07, "loss": 0.5555, "num_input_tokens_seen": 5492441088, "step": 2619 }, { "epoch": 0.059669107675617034, "grad_norm": 12.785634994506836, "learning_rate": 2.9088115686420917e-07, "loss": 0.4039, "num_input_tokens_seen": 5494538240, "step": 2620 }, { "epoch": 0.05994033089232438, "grad_norm": 15.309110641479492, "learning_rate": 2.9055121494542115e-07, "loss": 0.3993, "num_input_tokens_seen": 5496635392, "step": 2621 }, { "epoch": 0.060211554109031735, "grad_norm": 12.312765121459961, "learning_rate": 2.9022148182114247e-07, "loss": 0.3103, "num_input_tokens_seen": 5498732544, "step": 2622 }, { "epoch": 0.06048277732573908, "grad_norm": 20.542030334472656, "learning_rate": 2.8989195775673e-07, "loss": 0.5541, "num_input_tokens_seen": 5500829696, "step": 2623 }, { "epoch": 0.06075400054244643, "grad_norm": 10.995097160339355, "learning_rate": 2.89562643017372e-07, "loss": 0.3218, "num_input_tokens_seen": 5502926848, "step": 2624 }, { "epoch": 0.061025223759153785, "grad_norm": 20.725732803344727, "learning_rate": 2.8923353786808857e-07, "loss": 0.4933, "num_input_tokens_seen": 5505024000, "step": 2625 }, { "epoch": 0.06129644697586113, "grad_norm": 20.948585510253906, "learning_rate": 2.8890464257373105e-07, "loss": 0.8481, "num_input_tokens_seen": 5507121152, "step": 2626 }, { "epoch": 0.06156767019256849, "grad_norm": 12.35352611541748, "learning_rate": 2.8857595739898164e-07, "loss": 0.3946, "num_input_tokens_seen": 5509218304, "step": 2627 }, { "epoch": 0.061838893409275834, "grad_norm": 15.794177055358887, "learning_rate": 2.8824748260835386e-07, "loss": 0.6322, "num_input_tokens_seen": 5511315456, "step": 2628 }, { "epoch": 0.06211011662598318, "grad_norm": 20.326683044433594, "learning_rate": 2.879192184661918e-07, "loss": 0.6344, "num_input_tokens_seen": 5513412608, "step": 2629 }, { "epoch": 0.062381339842690536, "grad_norm": 16.63013458251953, "learning_rate": 2.8759116523666973e-07, "loss": 0.3775, "num_input_tokens_seen": 5515509760, "step": 2630 }, { "epoch": 0.06265256305939788, "grad_norm": 17.79449462890625, "learning_rate": 2.872633231837923e-07, "loss": 0.6498, "num_input_tokens_seen": 5517606912, "step": 2631 }, { "epoch": 0.06292378627610523, "grad_norm": 14.09510612487793, "learning_rate": 2.869356925713946e-07, "loss": 0.3053, "num_input_tokens_seen": 5519704064, "step": 2632 }, { "epoch": 0.06319500949281258, "grad_norm": 22.424938201904297, "learning_rate": 2.86608273663141e-07, "loss": 0.7264, "num_input_tokens_seen": 5521801216, "step": 2633 }, { "epoch": 0.06346623270951994, "grad_norm": 19.868343353271484, "learning_rate": 2.8628106672252614e-07, "loss": 0.7189, "num_input_tokens_seen": 5523898368, "step": 2634 }, { "epoch": 0.06373745592622729, "grad_norm": 11.038687705993652, "learning_rate": 2.859540720128737e-07, "loss": 0.316, "num_input_tokens_seen": 5525995520, "step": 2635 }, { "epoch": 0.06400867914293464, "grad_norm": 20.234725952148438, "learning_rate": 2.856272897973362e-07, "loss": 0.4332, "num_input_tokens_seen": 5528092672, "step": 2636 }, { "epoch": 0.06427990235964198, "grad_norm": 17.736494064331055, "learning_rate": 2.853007203388958e-07, "loss": 0.5984, "num_input_tokens_seen": 5530189824, "step": 2637 }, { "epoch": 0.06455112557634933, "grad_norm": 13.119430541992188, "learning_rate": 2.849743639003631e-07, "loss": 0.318, "num_input_tokens_seen": 5532286976, "step": 2638 }, { "epoch": 0.06482234879305669, "grad_norm": 17.339921951293945, "learning_rate": 2.846482207443773e-07, "loss": 0.4915, "num_input_tokens_seen": 5534384128, "step": 2639 }, { "epoch": 0.06509357200976404, "grad_norm": 12.775398254394531, "learning_rate": 2.843222911334061e-07, "loss": 0.3488, "num_input_tokens_seen": 5536481280, "step": 2640 }, { "epoch": 0.06536479522647139, "grad_norm": 19.59604263305664, "learning_rate": 2.839965753297452e-07, "loss": 0.6288, "num_input_tokens_seen": 5538578432, "step": 2641 }, { "epoch": 0.06563601844317873, "grad_norm": 11.156400680541992, "learning_rate": 2.8367107359551835e-07, "loss": 0.3287, "num_input_tokens_seen": 5540675584, "step": 2642 }, { "epoch": 0.06590724165988608, "grad_norm": 17.5681095123291, "learning_rate": 2.8334578619267683e-07, "loss": 0.6211, "num_input_tokens_seen": 5542772736, "step": 2643 }, { "epoch": 0.06617846487659344, "grad_norm": 11.513365745544434, "learning_rate": 2.830207133829997e-07, "loss": 0.2376, "num_input_tokens_seen": 5544869888, "step": 2644 }, { "epoch": 0.06644968809330079, "grad_norm": 18.175596237182617, "learning_rate": 2.8269585542809305e-07, "loss": 0.5273, "num_input_tokens_seen": 5546967040, "step": 2645 }, { "epoch": 0.06672091131000814, "grad_norm": 15.030982971191406, "learning_rate": 2.8237121258939036e-07, "loss": 0.3895, "num_input_tokens_seen": 5549064192, "step": 2646 }, { "epoch": 0.06699213452671549, "grad_norm": 23.305875778198242, "learning_rate": 2.8204678512815185e-07, "loss": 0.5009, "num_input_tokens_seen": 5551161344, "step": 2647 }, { "epoch": 0.06726335774342283, "grad_norm": 19.74225425720215, "learning_rate": 2.8172257330546414e-07, "loss": 0.5924, "num_input_tokens_seen": 5553258496, "step": 2648 }, { "epoch": 0.06753458096013018, "grad_norm": 13.259350776672363, "learning_rate": 2.8139857738224055e-07, "loss": 0.3439, "num_input_tokens_seen": 5555355648, "step": 2649 }, { "epoch": 0.06780580417683754, "grad_norm": 13.237799644470215, "learning_rate": 2.81074797619221e-07, "loss": 0.378, "num_input_tokens_seen": 5557452800, "step": 2650 }, { "epoch": 0.06807702739354489, "grad_norm": 14.329866409301758, "learning_rate": 2.8075123427697093e-07, "loss": 0.3658, "num_input_tokens_seen": 5559549952, "step": 2651 }, { "epoch": 0.06834825061025224, "grad_norm": 9.912088394165039, "learning_rate": 2.8042788761588204e-07, "loss": 0.2397, "num_input_tokens_seen": 5561647104, "step": 2652 }, { "epoch": 0.06861947382695958, "grad_norm": 19.02084732055664, "learning_rate": 2.8010475789617105e-07, "loss": 0.6075, "num_input_tokens_seen": 5563744256, "step": 2653 }, { "epoch": 0.06889069704366693, "grad_norm": 15.462594032287598, "learning_rate": 2.797818453778806e-07, "loss": 0.5206, "num_input_tokens_seen": 5565841408, "step": 2654 }, { "epoch": 0.0691619202603743, "grad_norm": 12.81130313873291, "learning_rate": 2.794591503208785e-07, "loss": 0.4167, "num_input_tokens_seen": 5567938560, "step": 2655 }, { "epoch": 0.06943314347708164, "grad_norm": 21.61235809326172, "learning_rate": 2.791366729848574e-07, "loss": 0.665, "num_input_tokens_seen": 5570035712, "step": 2656 }, { "epoch": 0.06970436669378899, "grad_norm": 27.101455688476562, "learning_rate": 2.7881441362933464e-07, "loss": 0.6858, "num_input_tokens_seen": 5572132864, "step": 2657 }, { "epoch": 0.06997558991049634, "grad_norm": 17.959209442138672, "learning_rate": 2.784923725136525e-07, "loss": 0.5518, "num_input_tokens_seen": 5574230016, "step": 2658 }, { "epoch": 0.07024681312720368, "grad_norm": 10.713489532470703, "learning_rate": 2.781705498969773e-07, "loss": 0.2727, "num_input_tokens_seen": 5576327168, "step": 2659 }, { "epoch": 0.07051803634391104, "grad_norm": 13.588273048400879, "learning_rate": 2.7784894603829966e-07, "loss": 0.4162, "num_input_tokens_seen": 5578424320, "step": 2660 }, { "epoch": 0.07078925956061839, "grad_norm": 14.655363082885742, "learning_rate": 2.7752756119643416e-07, "loss": 0.4192, "num_input_tokens_seen": 5580521472, "step": 2661 }, { "epoch": 0.07106048277732574, "grad_norm": 11.51174545288086, "learning_rate": 2.77206395630019e-07, "loss": 0.2364, "num_input_tokens_seen": 5582618624, "step": 2662 }, { "epoch": 0.07133170599403309, "grad_norm": 14.545519828796387, "learning_rate": 2.7688544959751615e-07, "loss": 0.4139, "num_input_tokens_seen": 5584715776, "step": 2663 }, { "epoch": 0.07160292921074043, "grad_norm": 18.516597747802734, "learning_rate": 2.765647233572108e-07, "loss": 0.4624, "num_input_tokens_seen": 5586812928, "step": 2664 }, { "epoch": 0.0718741524274478, "grad_norm": 16.15486717224121, "learning_rate": 2.7624421716721086e-07, "loss": 0.6008, "num_input_tokens_seen": 5588910080, "step": 2665 }, { "epoch": 0.07214537564415514, "grad_norm": 14.004398345947266, "learning_rate": 2.7592393128544784e-07, "loss": 0.4258, "num_input_tokens_seen": 5591007232, "step": 2666 }, { "epoch": 0.07241659886086249, "grad_norm": 18.64406967163086, "learning_rate": 2.7560386596967553e-07, "loss": 0.3099, "num_input_tokens_seen": 5593104384, "step": 2667 }, { "epoch": 0.07268782207756984, "grad_norm": 23.5781307220459, "learning_rate": 2.7528402147747045e-07, "loss": 0.8708, "num_input_tokens_seen": 5595201536, "step": 2668 }, { "epoch": 0.07295904529427719, "grad_norm": 16.206514358520508, "learning_rate": 2.7496439806623124e-07, "loss": 0.5205, "num_input_tokens_seen": 5597298688, "step": 2669 }, { "epoch": 0.07323026851098453, "grad_norm": 13.137125968933105, "learning_rate": 2.746449959931784e-07, "loss": 0.3308, "num_input_tokens_seen": 5599395840, "step": 2670 }, { "epoch": 0.0735014917276919, "grad_norm": 16.53045082092285, "learning_rate": 2.743258155153548e-07, "loss": 0.3276, "num_input_tokens_seen": 5601492992, "step": 2671 }, { "epoch": 0.07377271494439924, "grad_norm": 17.230728149414062, "learning_rate": 2.740068568896247e-07, "loss": 0.662, "num_input_tokens_seen": 5603590144, "step": 2672 }, { "epoch": 0.07404393816110659, "grad_norm": 16.93472671508789, "learning_rate": 2.7368812037267387e-07, "loss": 0.6562, "num_input_tokens_seen": 5605687296, "step": 2673 }, { "epoch": 0.07431516137781394, "grad_norm": 24.3737735748291, "learning_rate": 2.7336960622100907e-07, "loss": 0.8715, "num_input_tokens_seen": 5607784448, "step": 2674 }, { "epoch": 0.07458638459452128, "grad_norm": 10.38357162475586, "learning_rate": 2.7305131469095906e-07, "loss": 0.2184, "num_input_tokens_seen": 5609881600, "step": 2675 }, { "epoch": 0.07485760781122865, "grad_norm": 17.325672149658203, "learning_rate": 2.7273324603867203e-07, "loss": 0.5331, "num_input_tokens_seen": 5611978752, "step": 2676 }, { "epoch": 0.075128831027936, "grad_norm": 17.097993850708008, "learning_rate": 2.7241540052011787e-07, "loss": 0.6643, "num_input_tokens_seen": 5614075904, "step": 2677 }, { "epoch": 0.07540005424464334, "grad_norm": 9.489489555358887, "learning_rate": 2.720977783910865e-07, "loss": 0.2313, "num_input_tokens_seen": 5616173056, "step": 2678 }, { "epoch": 0.07567127746135069, "grad_norm": 16.426868438720703, "learning_rate": 2.717803799071881e-07, "loss": 0.4908, "num_input_tokens_seen": 5618270208, "step": 2679 }, { "epoch": 0.07594250067805804, "grad_norm": 16.419797897338867, "learning_rate": 2.714632053238529e-07, "loss": 0.6232, "num_input_tokens_seen": 5620367360, "step": 2680 }, { "epoch": 0.0762137238947654, "grad_norm": 14.052704811096191, "learning_rate": 2.711462548963309e-07, "loss": 0.3795, "num_input_tokens_seen": 5622464512, "step": 2681 }, { "epoch": 0.07648494711147275, "grad_norm": 14.086715698242188, "learning_rate": 2.708295288796918e-07, "loss": 0.4398, "num_input_tokens_seen": 5624561664, "step": 2682 }, { "epoch": 0.07675617032818009, "grad_norm": 17.156911849975586, "learning_rate": 2.7051302752882467e-07, "loss": 0.5515, "num_input_tokens_seen": 5626658816, "step": 2683 }, { "epoch": 0.07702739354488744, "grad_norm": 14.56786823272705, "learning_rate": 2.7019675109843777e-07, "loss": 0.3963, "num_input_tokens_seen": 5628755968, "step": 2684 }, { "epoch": 0.07729861676159479, "grad_norm": 19.659238815307617, "learning_rate": 2.6988069984305817e-07, "loss": 0.4958, "num_input_tokens_seen": 5630853120, "step": 2685 }, { "epoch": 0.07756983997830215, "grad_norm": 17.3908634185791, "learning_rate": 2.6956487401703207e-07, "loss": 0.5915, "num_input_tokens_seen": 5632950272, "step": 2686 }, { "epoch": 0.0778410631950095, "grad_norm": 24.59274673461914, "learning_rate": 2.692492738745241e-07, "loss": 0.6446, "num_input_tokens_seen": 5635047424, "step": 2687 }, { "epoch": 0.07811228641171684, "grad_norm": 21.10965347290039, "learning_rate": 2.689338996695171e-07, "loss": 0.7802, "num_input_tokens_seen": 5637144576, "step": 2688 }, { "epoch": 0.07838350962842419, "grad_norm": 12.290288925170898, "learning_rate": 2.686187516558122e-07, "loss": 0.3851, "num_input_tokens_seen": 5639241728, "step": 2689 }, { "epoch": 0.07865473284513154, "grad_norm": 17.741069793701172, "learning_rate": 2.683038300870285e-07, "loss": 0.5609, "num_input_tokens_seen": 5641338880, "step": 2690 }, { "epoch": 0.07892595606183889, "grad_norm": 11.127710342407227, "learning_rate": 2.679891352166032e-07, "loss": 0.3566, "num_input_tokens_seen": 5643436032, "step": 2691 }, { "epoch": 0.07919717927854625, "grad_norm": 17.833637237548828, "learning_rate": 2.6767466729779074e-07, "loss": 0.5932, "num_input_tokens_seen": 5645533184, "step": 2692 }, { "epoch": 0.0794684024952536, "grad_norm": 18.04448127746582, "learning_rate": 2.6736042658366255e-07, "loss": 0.5263, "num_input_tokens_seen": 5647630336, "step": 2693 }, { "epoch": 0.07973962571196094, "grad_norm": 16.75208282470703, "learning_rate": 2.6704641332710783e-07, "loss": 0.5939, "num_input_tokens_seen": 5649727488, "step": 2694 }, { "epoch": 0.08001084892866829, "grad_norm": 18.136194229125977, "learning_rate": 2.6673262778083246e-07, "loss": 0.5117, "num_input_tokens_seen": 5651824640, "step": 2695 }, { "epoch": 0.08028207214537564, "grad_norm": 14.582008361816406, "learning_rate": 2.6641907019735914e-07, "loss": 0.4515, "num_input_tokens_seen": 5653921792, "step": 2696 }, { "epoch": 0.080553295362083, "grad_norm": 9.552046775817871, "learning_rate": 2.6610574082902704e-07, "loss": 0.2099, "num_input_tokens_seen": 5656018944, "step": 2697 }, { "epoch": 0.08082451857879035, "grad_norm": 15.047245979309082, "learning_rate": 2.657926399279918e-07, "loss": 0.5305, "num_input_tokens_seen": 5658116096, "step": 2698 }, { "epoch": 0.0810957417954977, "grad_norm": 11.793010711669922, "learning_rate": 2.65479767746225e-07, "loss": 0.2842, "num_input_tokens_seen": 5660213248, "step": 2699 }, { "epoch": 0.08136696501220504, "grad_norm": 11.611842155456543, "learning_rate": 2.651671245355144e-07, "loss": 0.2676, "num_input_tokens_seen": 5662310400, "step": 2700 }, { "epoch": 0.08163818822891239, "grad_norm": 14.877463340759277, "learning_rate": 2.6485471054746315e-07, "loss": 0.4741, "num_input_tokens_seen": 5664407552, "step": 2701 }, { "epoch": 0.08190941144561975, "grad_norm": 16.487375259399414, "learning_rate": 2.645425260334904e-07, "loss": 0.4078, "num_input_tokens_seen": 5666504704, "step": 2702 }, { "epoch": 0.0821806346623271, "grad_norm": 21.23032569885254, "learning_rate": 2.6423057124483015e-07, "loss": 0.7662, "num_input_tokens_seen": 5668601856, "step": 2703 }, { "epoch": 0.08245185787903445, "grad_norm": 9.899909973144531, "learning_rate": 2.6391884643253197e-07, "loss": 0.2434, "num_input_tokens_seen": 5670699008, "step": 2704 }, { "epoch": 0.0827230810957418, "grad_norm": 13.111746788024902, "learning_rate": 2.6360735184745984e-07, "loss": 0.3323, "num_input_tokens_seen": 5672796160, "step": 2705 }, { "epoch": 0.08299430431244914, "grad_norm": 13.60911750793457, "learning_rate": 2.6329608774029285e-07, "loss": 0.4452, "num_input_tokens_seen": 5674893312, "step": 2706 }, { "epoch": 0.0832655275291565, "grad_norm": 19.949071884155273, "learning_rate": 2.6298505436152457e-07, "loss": 0.6172, "num_input_tokens_seen": 5676990464, "step": 2707 }, { "epoch": 0.08353675074586385, "grad_norm": 24.41508674621582, "learning_rate": 2.626742519614629e-07, "loss": 0.3937, "num_input_tokens_seen": 5679087616, "step": 2708 }, { "epoch": 0.0838079739625712, "grad_norm": 14.442510604858398, "learning_rate": 2.623636807902298e-07, "loss": 0.4276, "num_input_tokens_seen": 5681184768, "step": 2709 }, { "epoch": 0.08407919717927854, "grad_norm": 20.97626495361328, "learning_rate": 2.620533410977609e-07, "loss": 0.3205, "num_input_tokens_seen": 5683281920, "step": 2710 }, { "epoch": 0.08435042039598589, "grad_norm": 17.24020004272461, "learning_rate": 2.617432331338059e-07, "loss": 0.4976, "num_input_tokens_seen": 5685379072, "step": 2711 }, { "epoch": 0.08462164361269324, "grad_norm": 25.12828826904297, "learning_rate": 2.614333571479279e-07, "loss": 0.7029, "num_input_tokens_seen": 5687476224, "step": 2712 }, { "epoch": 0.0848928668294006, "grad_norm": 10.766886711120605, "learning_rate": 2.6112371338950325e-07, "loss": 0.2783, "num_input_tokens_seen": 5689573376, "step": 2713 }, { "epoch": 0.08516409004610795, "grad_norm": 19.415145874023438, "learning_rate": 2.608143021077215e-07, "loss": 0.6848, "num_input_tokens_seen": 5691670528, "step": 2714 }, { "epoch": 0.0854353132628153, "grad_norm": 11.980889320373535, "learning_rate": 2.6050512355158503e-07, "loss": 0.35, "num_input_tokens_seen": 5693767680, "step": 2715 }, { "epoch": 0.08570653647952264, "grad_norm": 18.769763946533203, "learning_rate": 2.601961779699091e-07, "loss": 0.6817, "num_input_tokens_seen": 5695864832, "step": 2716 }, { "epoch": 0.08597775969622999, "grad_norm": 11.969738960266113, "learning_rate": 2.5988746561132136e-07, "loss": 0.3525, "num_input_tokens_seen": 5697961984, "step": 2717 }, { "epoch": 0.08624898291293735, "grad_norm": 13.127973556518555, "learning_rate": 2.595789867242616e-07, "loss": 0.3985, "num_input_tokens_seen": 5700059136, "step": 2718 }, { "epoch": 0.0865202061296447, "grad_norm": 14.819026947021484, "learning_rate": 2.5927074155698203e-07, "loss": 0.3679, "num_input_tokens_seen": 5702156288, "step": 2719 }, { "epoch": 0.08679142934635205, "grad_norm": 19.240089416503906, "learning_rate": 2.589627303575465e-07, "loss": 0.7484, "num_input_tokens_seen": 5704253440, "step": 2720 }, { "epoch": 0.0870626525630594, "grad_norm": 15.054210662841797, "learning_rate": 2.5865495337383103e-07, "loss": 0.4892, "num_input_tokens_seen": 5706350592, "step": 2721 }, { "epoch": 0.08733387577976674, "grad_norm": 20.892154693603516, "learning_rate": 2.5834741085352223e-07, "loss": 0.4666, "num_input_tokens_seen": 5708447744, "step": 2722 }, { "epoch": 0.0876050989964741, "grad_norm": 17.323524475097656, "learning_rate": 2.5804010304411914e-07, "loss": 0.7108, "num_input_tokens_seen": 5710544896, "step": 2723 }, { "epoch": 0.08787632221318145, "grad_norm": 9.577323913574219, "learning_rate": 2.5773303019293123e-07, "loss": 0.2308, "num_input_tokens_seen": 5712642048, "step": 2724 }, { "epoch": 0.0881475454298888, "grad_norm": 8.77543830871582, "learning_rate": 2.5742619254707905e-07, "loss": 0.1873, "num_input_tokens_seen": 5714739200, "step": 2725 }, { "epoch": 0.08841876864659615, "grad_norm": 16.472463607788086, "learning_rate": 2.5711959035349396e-07, "loss": 0.5128, "num_input_tokens_seen": 5716836352, "step": 2726 }, { "epoch": 0.0886899918633035, "grad_norm": 10.295055389404297, "learning_rate": 2.568132238589175e-07, "loss": 0.2232, "num_input_tokens_seen": 5718933504, "step": 2727 }, { "epoch": 0.08896121508001086, "grad_norm": 17.552204132080078, "learning_rate": 2.565070933099019e-07, "loss": 0.4109, "num_input_tokens_seen": 5721030656, "step": 2728 }, { "epoch": 0.0892324382967182, "grad_norm": 14.377561569213867, "learning_rate": 2.5620119895280935e-07, "loss": 0.3614, "num_input_tokens_seen": 5723127808, "step": 2729 }, { "epoch": 0.08950366151342555, "grad_norm": 14.117269515991211, "learning_rate": 2.5589554103381195e-07, "loss": 0.3619, "num_input_tokens_seen": 5725224960, "step": 2730 }, { "epoch": 0.0897748847301329, "grad_norm": 17.120271682739258, "learning_rate": 2.5559011979889155e-07, "loss": 0.4312, "num_input_tokens_seen": 5727322112, "step": 2731 }, { "epoch": 0.09004610794684025, "grad_norm": 16.064838409423828, "learning_rate": 2.552849354938399e-07, "loss": 0.3673, "num_input_tokens_seen": 5729419264, "step": 2732 }, { "epoch": 0.09031733116354759, "grad_norm": 12.923369407653809, "learning_rate": 2.549799883642573e-07, "loss": 0.3846, "num_input_tokens_seen": 5731516416, "step": 2733 }, { "epoch": 0.09058855438025495, "grad_norm": 10.39802074432373, "learning_rate": 2.5467527865555366e-07, "loss": 0.2633, "num_input_tokens_seen": 5733613568, "step": 2734 }, { "epoch": 0.0908597775969623, "grad_norm": 21.36480140686035, "learning_rate": 2.5437080661294785e-07, "loss": 0.627, "num_input_tokens_seen": 5735710720, "step": 2735 }, { "epoch": 0.09113100081366965, "grad_norm": 18.12044906616211, "learning_rate": 2.5406657248146735e-07, "loss": 0.5679, "num_input_tokens_seen": 5737807872, "step": 2736 }, { "epoch": 0.091402224030377, "grad_norm": 17.808435440063477, "learning_rate": 2.5376257650594823e-07, "loss": 0.5542, "num_input_tokens_seen": 5739905024, "step": 2737 }, { "epoch": 0.09167344724708434, "grad_norm": 11.723854064941406, "learning_rate": 2.5345881893103484e-07, "loss": 0.2498, "num_input_tokens_seen": 5742002176, "step": 2738 }, { "epoch": 0.0919446704637917, "grad_norm": 21.123031616210938, "learning_rate": 2.5315530000117973e-07, "loss": 0.7246, "num_input_tokens_seen": 5744099328, "step": 2739 }, { "epoch": 0.09221589368049905, "grad_norm": 21.011804580688477, "learning_rate": 2.528520199606434e-07, "loss": 0.7669, "num_input_tokens_seen": 5746196480, "step": 2740 }, { "epoch": 0.0924871168972064, "grad_norm": 16.76370620727539, "learning_rate": 2.525489790534941e-07, "loss": 0.6061, "num_input_tokens_seen": 5748293632, "step": 2741 }, { "epoch": 0.09275834011391375, "grad_norm": 20.120906829833984, "learning_rate": 2.5224617752360766e-07, "loss": 0.7128, "num_input_tokens_seen": 5750390784, "step": 2742 }, { "epoch": 0.0930295633306211, "grad_norm": 12.05733871459961, "learning_rate": 2.519436156146671e-07, "loss": 0.4059, "num_input_tokens_seen": 5752487936, "step": 2743 }, { "epoch": 0.09330078654732846, "grad_norm": 14.433719635009766, "learning_rate": 2.51641293570163e-07, "loss": 0.5306, "num_input_tokens_seen": 5754585088, "step": 2744 }, { "epoch": 0.0935720097640358, "grad_norm": 10.016216278076172, "learning_rate": 2.513392116333922e-07, "loss": 0.2541, "num_input_tokens_seen": 5756682240, "step": 2745 }, { "epoch": 0.09384323298074315, "grad_norm": 15.06049919128418, "learning_rate": 2.510373700474592e-07, "loss": 0.4363, "num_input_tokens_seen": 5758779392, "step": 2746 }, { "epoch": 0.0941144561974505, "grad_norm": 12.102774620056152, "learning_rate": 2.5073576905527407e-07, "loss": 0.3181, "num_input_tokens_seen": 5760876544, "step": 2747 }, { "epoch": 0.09438567941415785, "grad_norm": 16.28287124633789, "learning_rate": 2.5043440889955434e-07, "loss": 0.5829, "num_input_tokens_seen": 5762973696, "step": 2748 }, { "epoch": 0.09465690263086521, "grad_norm": 17.23906135559082, "learning_rate": 2.501332898228232e-07, "loss": 0.4747, "num_input_tokens_seen": 5765070848, "step": 2749 }, { "epoch": 0.09492812584757256, "grad_norm": 16.309682846069336, "learning_rate": 2.4983241206740945e-07, "loss": 0.3862, "num_input_tokens_seen": 5767168000, "step": 2750 }, { "epoch": 0.0951993490642799, "grad_norm": 16.106428146362305, "learning_rate": 2.495317758754483e-07, "loss": 0.5324, "num_input_tokens_seen": 5769265152, "step": 2751 }, { "epoch": 0.09547057228098725, "grad_norm": 17.912614822387695, "learning_rate": 2.492313814888802e-07, "loss": 0.5044, "num_input_tokens_seen": 5771362304, "step": 2752 }, { "epoch": 0.0957417954976946, "grad_norm": 14.133460998535156, "learning_rate": 2.4893122914945124e-07, "loss": 0.3362, "num_input_tokens_seen": 5773459456, "step": 2753 }, { "epoch": 0.09601301871440195, "grad_norm": 15.399311065673828, "learning_rate": 2.4863131909871247e-07, "loss": 0.3883, "num_input_tokens_seen": 5775556608, "step": 2754 }, { "epoch": 0.09628424193110931, "grad_norm": 15.49337100982666, "learning_rate": 2.483316515780202e-07, "loss": 0.5335, "num_input_tokens_seen": 5777653760, "step": 2755 }, { "epoch": 0.09655546514781665, "grad_norm": 9.771089553833008, "learning_rate": 2.480322268285354e-07, "loss": 0.2648, "num_input_tokens_seen": 5779750912, "step": 2756 }, { "epoch": 0.096826688364524, "grad_norm": 12.803630828857422, "learning_rate": 2.4773304509122374e-07, "loss": 0.3456, "num_input_tokens_seen": 5781848064, "step": 2757 }, { "epoch": 0.09709791158123135, "grad_norm": 14.358229637145996, "learning_rate": 2.474341066068553e-07, "loss": 0.4647, "num_input_tokens_seen": 5783945216, "step": 2758 }, { "epoch": 0.0973691347979387, "grad_norm": 19.96444320678711, "learning_rate": 2.4713541161600434e-07, "loss": 0.7291, "num_input_tokens_seen": 5786042368, "step": 2759 }, { "epoch": 0.09764035801464606, "grad_norm": 15.138256072998047, "learning_rate": 2.4683696035904926e-07, "loss": 0.4123, "num_input_tokens_seen": 5788139520, "step": 2760 }, { "epoch": 0.0979115812313534, "grad_norm": 14.141944885253906, "learning_rate": 2.465387530761724e-07, "loss": 0.3199, "num_input_tokens_seen": 5790236672, "step": 2761 }, { "epoch": 0.09818280444806075, "grad_norm": 14.20753288269043, "learning_rate": 2.462407900073594e-07, "loss": 0.3559, "num_input_tokens_seen": 5792333824, "step": 2762 }, { "epoch": 0.0984540276647681, "grad_norm": 17.01785659790039, "learning_rate": 2.459430713923995e-07, "loss": 0.6968, "num_input_tokens_seen": 5794430976, "step": 2763 }, { "epoch": 0.09872525088147545, "grad_norm": 9.875651359558105, "learning_rate": 2.4564559747088573e-07, "loss": 0.2179, "num_input_tokens_seen": 5796528128, "step": 2764 }, { "epoch": 0.09899647409818281, "grad_norm": 16.090343475341797, "learning_rate": 2.4534836848221355e-07, "loss": 0.4962, "num_input_tokens_seen": 5798625280, "step": 2765 }, { "epoch": 0.09926769731489016, "grad_norm": 13.723416328430176, "learning_rate": 2.450513846655817e-07, "loss": 0.4518, "num_input_tokens_seen": 5800722432, "step": 2766 }, { "epoch": 0.0995389205315975, "grad_norm": 13.636754989624023, "learning_rate": 2.4475464625999113e-07, "loss": 0.4715, "num_input_tokens_seen": 5802819584, "step": 2767 }, { "epoch": 0.09981014374830485, "grad_norm": 16.98908233642578, "learning_rate": 2.444581535042459e-07, "loss": 0.4008, "num_input_tokens_seen": 5804916736, "step": 2768 }, { "epoch": 0.1000813669650122, "grad_norm": 23.043598175048828, "learning_rate": 2.441619066369519e-07, "loss": 0.988, "num_input_tokens_seen": 5807013888, "step": 2769 }, { "epoch": 0.10035259018171956, "grad_norm": 16.22281837463379, "learning_rate": 2.438659058965175e-07, "loss": 0.585, "num_input_tokens_seen": 5809111040, "step": 2770 }, { "epoch": 0.10062381339842691, "grad_norm": 13.878887176513672, "learning_rate": 2.435701515211527e-07, "loss": 0.3956, "num_input_tokens_seen": 5811208192, "step": 2771 }, { "epoch": 0.10089503661513426, "grad_norm": 15.073182106018066, "learning_rate": 2.4327464374886955e-07, "loss": 0.3474, "num_input_tokens_seen": 5813305344, "step": 2772 }, { "epoch": 0.1011662598318416, "grad_norm": 37.12259292602539, "learning_rate": 2.4297938281748124e-07, "loss": 0.4542, "num_input_tokens_seen": 5815402496, "step": 2773 }, { "epoch": 0.10143748304854895, "grad_norm": 21.452543258666992, "learning_rate": 2.4268436896460267e-07, "loss": 0.5917, "num_input_tokens_seen": 5817499648, "step": 2774 }, { "epoch": 0.10170870626525631, "grad_norm": 18.589229583740234, "learning_rate": 2.423896024276497e-07, "loss": 0.6739, "num_input_tokens_seen": 5819596800, "step": 2775 }, { "epoch": 0.10197992948196366, "grad_norm": 15.261178970336914, "learning_rate": 2.4209508344383926e-07, "loss": 0.5353, "num_input_tokens_seen": 5821693952, "step": 2776 }, { "epoch": 0.10225115269867101, "grad_norm": 19.2561092376709, "learning_rate": 2.4180081225018906e-07, "loss": 0.6213, "num_input_tokens_seen": 5823791104, "step": 2777 }, { "epoch": 0.10252237591537836, "grad_norm": 17.913654327392578, "learning_rate": 2.4150678908351744e-07, "loss": 0.748, "num_input_tokens_seen": 5825888256, "step": 2778 }, { "epoch": 0.1027935991320857, "grad_norm": 11.78269100189209, "learning_rate": 2.4121301418044264e-07, "loss": 0.3224, "num_input_tokens_seen": 5827985408, "step": 2779 }, { "epoch": 0.10306482234879305, "grad_norm": 11.825636863708496, "learning_rate": 2.409194877773839e-07, "loss": 0.2892, "num_input_tokens_seen": 5830082560, "step": 2780 }, { "epoch": 0.10333604556550041, "grad_norm": 21.131614685058594, "learning_rate": 2.4062621011056006e-07, "loss": 0.4957, "num_input_tokens_seen": 5832179712, "step": 2781 }, { "epoch": 0.10360726878220776, "grad_norm": 16.387815475463867, "learning_rate": 2.4033318141598977e-07, "loss": 0.3793, "num_input_tokens_seen": 5834276864, "step": 2782 }, { "epoch": 0.1038784919989151, "grad_norm": 15.578926086425781, "learning_rate": 2.4004040192949155e-07, "loss": 0.5131, "num_input_tokens_seen": 5836374016, "step": 2783 }, { "epoch": 0.10414971521562245, "grad_norm": 18.233552932739258, "learning_rate": 2.3974787188668284e-07, "loss": 0.3963, "num_input_tokens_seen": 5838471168, "step": 2784 }, { "epoch": 0.1044209384323298, "grad_norm": 16.372631072998047, "learning_rate": 2.3945559152298085e-07, "loss": 0.5632, "num_input_tokens_seen": 5840568320, "step": 2785 }, { "epoch": 0.10469216164903716, "grad_norm": 13.143184661865234, "learning_rate": 2.3916356107360174e-07, "loss": 0.4031, "num_input_tokens_seen": 5842665472, "step": 2786 }, { "epoch": 0.10496338486574451, "grad_norm": 17.36909294128418, "learning_rate": 2.388717807735605e-07, "loss": 0.5266, "num_input_tokens_seen": 5844762624, "step": 2787 }, { "epoch": 0.10523460808245186, "grad_norm": 11.10837173461914, "learning_rate": 2.385802508576706e-07, "loss": 0.2815, "num_input_tokens_seen": 5846859776, "step": 2788 }, { "epoch": 0.1055058312991592, "grad_norm": 16.8798885345459, "learning_rate": 2.382889715605447e-07, "loss": 0.2858, "num_input_tokens_seen": 5848956928, "step": 2789 }, { "epoch": 0.10577705451586655, "grad_norm": 17.28868293762207, "learning_rate": 2.3799794311659286e-07, "loss": 0.2763, "num_input_tokens_seen": 5851054080, "step": 2790 }, { "epoch": 0.10604827773257391, "grad_norm": 20.567678451538086, "learning_rate": 2.3770716576002383e-07, "loss": 0.5661, "num_input_tokens_seen": 5853151232, "step": 2791 }, { "epoch": 0.10631950094928126, "grad_norm": 15.973060607910156, "learning_rate": 2.3741663972484416e-07, "loss": 0.5525, "num_input_tokens_seen": 5855248384, "step": 2792 }, { "epoch": 0.10659072416598861, "grad_norm": 12.698427200317383, "learning_rate": 2.3712636524485817e-07, "loss": 0.2953, "num_input_tokens_seen": 5857345536, "step": 2793 }, { "epoch": 0.10686194738269596, "grad_norm": 11.15018081665039, "learning_rate": 2.3683634255366773e-07, "loss": 0.246, "num_input_tokens_seen": 5859442688, "step": 2794 }, { "epoch": 0.1071331705994033, "grad_norm": 18.49885368347168, "learning_rate": 2.3654657188467203e-07, "loss": 0.6083, "num_input_tokens_seen": 5861539840, "step": 2795 }, { "epoch": 0.10740439381611067, "grad_norm": 30.425281524658203, "learning_rate": 2.3625705347106748e-07, "loss": 0.5344, "num_input_tokens_seen": 5863636992, "step": 2796 }, { "epoch": 0.10767561703281801, "grad_norm": 14.430277824401855, "learning_rate": 2.3596778754584752e-07, "loss": 0.4491, "num_input_tokens_seen": 5865734144, "step": 2797 }, { "epoch": 0.10794684024952536, "grad_norm": 16.26645278930664, "learning_rate": 2.356787743418023e-07, "loss": 0.402, "num_input_tokens_seen": 5867831296, "step": 2798 }, { "epoch": 0.10821806346623271, "grad_norm": 9.986083030700684, "learning_rate": 2.3539001409151867e-07, "loss": 0.2545, "num_input_tokens_seen": 5869928448, "step": 2799 }, { "epoch": 0.10848928668294006, "grad_norm": 13.592569351196289, "learning_rate": 2.351015070273799e-07, "loss": 0.4131, "num_input_tokens_seen": 5872025600, "step": 2800 }, { "epoch": 0.1087605098996474, "grad_norm": 18.985036849975586, "learning_rate": 2.3481325338156568e-07, "loss": 0.7892, "num_input_tokens_seen": 5874122752, "step": 2801 }, { "epoch": 0.10903173311635476, "grad_norm": 10.7304048538208, "learning_rate": 2.3452525338605126e-07, "loss": 0.2678, "num_input_tokens_seen": 5876219904, "step": 2802 }, { "epoch": 0.10930295633306211, "grad_norm": 16.625381469726562, "learning_rate": 2.3423750727260813e-07, "loss": 0.3209, "num_input_tokens_seen": 5878317056, "step": 2803 }, { "epoch": 0.10957417954976946, "grad_norm": 13.496749877929688, "learning_rate": 2.3395001527280343e-07, "loss": 0.2586, "num_input_tokens_seen": 5880414208, "step": 2804 }, { "epoch": 0.10984540276647681, "grad_norm": 17.611549377441406, "learning_rate": 2.3366277761799993e-07, "loss": 0.6402, "num_input_tokens_seen": 5882511360, "step": 2805 }, { "epoch": 0.11011662598318415, "grad_norm": 16.595611572265625, "learning_rate": 2.333757945393557e-07, "loss": 0.5554, "num_input_tokens_seen": 5884608512, "step": 2806 }, { "epoch": 0.11038784919989152, "grad_norm": 14.515238761901855, "learning_rate": 2.3308906626782333e-07, "loss": 0.4586, "num_input_tokens_seen": 5886705664, "step": 2807 }, { "epoch": 0.11065907241659886, "grad_norm": 23.05605697631836, "learning_rate": 2.3280259303415115e-07, "loss": 0.6299, "num_input_tokens_seen": 5888802816, "step": 2808 }, { "epoch": 0.11093029563330621, "grad_norm": 10.770841598510742, "learning_rate": 2.3251637506888178e-07, "loss": 0.191, "num_input_tokens_seen": 5890899968, "step": 2809 }, { "epoch": 0.11120151885001356, "grad_norm": 12.446057319641113, "learning_rate": 2.3223041260235265e-07, "loss": 0.3576, "num_input_tokens_seen": 5892997120, "step": 2810 }, { "epoch": 0.1114727420667209, "grad_norm": 8.817853927612305, "learning_rate": 2.3194470586469543e-07, "loss": 0.2229, "num_input_tokens_seen": 5895094272, "step": 2811 }, { "epoch": 0.11174396528342827, "grad_norm": 16.82341957092285, "learning_rate": 2.3165925508583598e-07, "loss": 0.372, "num_input_tokens_seen": 5897191424, "step": 2812 }, { "epoch": 0.11201518850013562, "grad_norm": 19.0521183013916, "learning_rate": 2.313740604954943e-07, "loss": 0.4898, "num_input_tokens_seen": 5899288576, "step": 2813 }, { "epoch": 0.11228641171684296, "grad_norm": 16.593416213989258, "learning_rate": 2.3108912232318413e-07, "loss": 0.4401, "num_input_tokens_seen": 5901385728, "step": 2814 }, { "epoch": 0.11255763493355031, "grad_norm": 12.823646545410156, "learning_rate": 2.3080444079821284e-07, "loss": 0.3815, "num_input_tokens_seen": 5903482880, "step": 2815 }, { "epoch": 0.11282885815025766, "grad_norm": 26.622953414916992, "learning_rate": 2.305200161496813e-07, "loss": 0.9118, "num_input_tokens_seen": 5905580032, "step": 2816 }, { "epoch": 0.11310008136696502, "grad_norm": 16.885168075561523, "learning_rate": 2.3023584860648364e-07, "loss": 0.7591, "num_input_tokens_seen": 5907677184, "step": 2817 }, { "epoch": 0.11337130458367237, "grad_norm": 15.160221099853516, "learning_rate": 2.2995193839730727e-07, "loss": 0.4651, "num_input_tokens_seen": 5909774336, "step": 2818 }, { "epoch": 0.11364252780037971, "grad_norm": 14.561100959777832, "learning_rate": 2.2966828575063196e-07, "loss": 0.4091, "num_input_tokens_seen": 5911871488, "step": 2819 }, { "epoch": 0.11391375101708706, "grad_norm": 16.13934898376465, "learning_rate": 2.2938489089473065e-07, "loss": 0.483, "num_input_tokens_seen": 5913968640, "step": 2820 }, { "epoch": 0.11418497423379441, "grad_norm": 20.38286590576172, "learning_rate": 2.2910175405766896e-07, "loss": 0.7366, "num_input_tokens_seen": 5916065792, "step": 2821 }, { "epoch": 0.11445619745050176, "grad_norm": 18.793577194213867, "learning_rate": 2.288188754673045e-07, "loss": 0.4723, "num_input_tokens_seen": 5918162944, "step": 2822 }, { "epoch": 0.11472742066720912, "grad_norm": 18.476377487182617, "learning_rate": 2.2853625535128735e-07, "loss": 0.7003, "num_input_tokens_seen": 5920260096, "step": 2823 }, { "epoch": 0.11499864388391647, "grad_norm": 11.239076614379883, "learning_rate": 2.2825389393705922e-07, "loss": 0.2991, "num_input_tokens_seen": 5922357248, "step": 2824 }, { "epoch": 0.11526986710062381, "grad_norm": 17.184232711791992, "learning_rate": 2.2797179145185384e-07, "loss": 0.6045, "num_input_tokens_seen": 5924454400, "step": 2825 }, { "epoch": 0.11554109031733116, "grad_norm": 12.037182807922363, "learning_rate": 2.2768994812269666e-07, "loss": 0.2699, "num_input_tokens_seen": 5926551552, "step": 2826 }, { "epoch": 0.11581231353403851, "grad_norm": 12.547082901000977, "learning_rate": 2.2740836417640432e-07, "loss": 0.3864, "num_input_tokens_seen": 5928648704, "step": 2827 }, { "epoch": 0.11608353675074587, "grad_norm": 16.02481460571289, "learning_rate": 2.271270398395851e-07, "loss": 0.6835, "num_input_tokens_seen": 5930745856, "step": 2828 }, { "epoch": 0.11635475996745322, "grad_norm": 11.960121154785156, "learning_rate": 2.2684597533863793e-07, "loss": 0.3943, "num_input_tokens_seen": 5932843008, "step": 2829 }, { "epoch": 0.11662598318416056, "grad_norm": 11.560256958007812, "learning_rate": 2.2656517089975286e-07, "loss": 0.3184, "num_input_tokens_seen": 5934940160, "step": 2830 }, { "epoch": 0.11689720640086791, "grad_norm": 15.655224800109863, "learning_rate": 2.2628462674891062e-07, "loss": 0.5279, "num_input_tokens_seen": 5937037312, "step": 2831 }, { "epoch": 0.11716842961757526, "grad_norm": 15.328285217285156, "learning_rate": 2.2600434311188253e-07, "loss": 0.4113, "num_input_tokens_seen": 5939134464, "step": 2832 }, { "epoch": 0.11743965283428262, "grad_norm": 20.01653289794922, "learning_rate": 2.2572432021423018e-07, "loss": 0.5133, "num_input_tokens_seen": 5941231616, "step": 2833 }, { "epoch": 0.11771087605098997, "grad_norm": 18.01945686340332, "learning_rate": 2.2544455828130528e-07, "loss": 0.7168, "num_input_tokens_seen": 5943328768, "step": 2834 }, { "epoch": 0.11798209926769732, "grad_norm": 14.951272964477539, "learning_rate": 2.251650575382498e-07, "loss": 0.5018, "num_input_tokens_seen": 5945425920, "step": 2835 }, { "epoch": 0.11825332248440466, "grad_norm": 20.382293701171875, "learning_rate": 2.2488581820999481e-07, "loss": 0.5145, "num_input_tokens_seen": 5947523072, "step": 2836 }, { "epoch": 0.11852454570111201, "grad_norm": 17.990644454956055, "learning_rate": 2.2460684052126195e-07, "loss": 0.6625, "num_input_tokens_seen": 5949620224, "step": 2837 }, { "epoch": 0.11879576891781937, "grad_norm": 14.664972305297852, "learning_rate": 2.243281246965617e-07, "loss": 0.5635, "num_input_tokens_seen": 5951717376, "step": 2838 }, { "epoch": 0.11906699213452672, "grad_norm": 12.643186569213867, "learning_rate": 2.24049670960194e-07, "loss": 0.2588, "num_input_tokens_seen": 5953814528, "step": 2839 }, { "epoch": 0.11933821535123407, "grad_norm": 12.24855899810791, "learning_rate": 2.2377147953624776e-07, "loss": 0.3325, "num_input_tokens_seen": 5955911680, "step": 2840 }, { "epoch": 0.11960943856794141, "grad_norm": 18.2128849029541, "learning_rate": 2.2349355064860104e-07, "loss": 0.5419, "num_input_tokens_seen": 5958008832, "step": 2841 }, { "epoch": 0.11988066178464876, "grad_norm": 18.80755043029785, "learning_rate": 2.2321588452092006e-07, "loss": 0.7344, "num_input_tokens_seen": 5960105984, "step": 2842 }, { "epoch": 0.12015188500135611, "grad_norm": 9.964282035827637, "learning_rate": 2.229384813766602e-07, "loss": 0.3019, "num_input_tokens_seen": 5962203136, "step": 2843 }, { "epoch": 0.12042310821806347, "grad_norm": 14.523286819458008, "learning_rate": 2.2266134143906496e-07, "loss": 0.3197, "num_input_tokens_seen": 5964300288, "step": 2844 }, { "epoch": 0.12069433143477082, "grad_norm": 14.210477828979492, "learning_rate": 2.2238446493116572e-07, "loss": 0.378, "num_input_tokens_seen": 5966397440, "step": 2845 }, { "epoch": 0.12096555465147817, "grad_norm": 16.729429244995117, "learning_rate": 2.2210785207578275e-07, "loss": 0.5794, "num_input_tokens_seen": 5968494592, "step": 2846 }, { "epoch": 0.12123677786818551, "grad_norm": 11.647602081298828, "learning_rate": 2.2183150309552308e-07, "loss": 0.2484, "num_input_tokens_seen": 5970591744, "step": 2847 }, { "epoch": 0.12150800108489286, "grad_norm": 18.91074562072754, "learning_rate": 2.2155541821278196e-07, "loss": 0.4457, "num_input_tokens_seen": 5972688896, "step": 2848 }, { "epoch": 0.12177922430160022, "grad_norm": 13.002408027648926, "learning_rate": 2.2127959764974203e-07, "loss": 0.3013, "num_input_tokens_seen": 5974786048, "step": 2849 }, { "epoch": 0.12205044751830757, "grad_norm": 17.768815994262695, "learning_rate": 2.2100404162837317e-07, "loss": 0.5265, "num_input_tokens_seen": 5976883200, "step": 2850 }, { "epoch": 0.12232167073501492, "grad_norm": 17.112438201904297, "learning_rate": 2.2072875037043232e-07, "loss": 0.5274, "num_input_tokens_seen": 5978980352, "step": 2851 }, { "epoch": 0.12259289395172226, "grad_norm": 11.445368766784668, "learning_rate": 2.204537240974636e-07, "loss": 0.3618, "num_input_tokens_seen": 5981077504, "step": 2852 }, { "epoch": 0.12286411716842961, "grad_norm": 17.361818313598633, "learning_rate": 2.2017896303079743e-07, "loss": 0.5425, "num_input_tokens_seen": 5983174656, "step": 2853 }, { "epoch": 0.12313534038513697, "grad_norm": 19.054218292236328, "learning_rate": 2.1990446739155128e-07, "loss": 0.3945, "num_input_tokens_seen": 5985271808, "step": 2854 }, { "epoch": 0.12340656360184432, "grad_norm": 11.030587196350098, "learning_rate": 2.1963023740062864e-07, "loss": 0.2286, "num_input_tokens_seen": 5987368960, "step": 2855 }, { "epoch": 0.12367778681855167, "grad_norm": 12.42334270477295, "learning_rate": 2.1935627327871948e-07, "loss": 0.2749, "num_input_tokens_seen": 5989466112, "step": 2856 }, { "epoch": 0.12394901003525902, "grad_norm": 10.632824897766113, "learning_rate": 2.1908257524629963e-07, "loss": 0.4374, "num_input_tokens_seen": 5991563264, "step": 2857 }, { "epoch": 0.12422023325196636, "grad_norm": 19.065475463867188, "learning_rate": 2.1880914352363106e-07, "loss": 0.581, "num_input_tokens_seen": 5993660416, "step": 2858 }, { "epoch": 0.12449145646867373, "grad_norm": 16.23573112487793, "learning_rate": 2.1853597833076088e-07, "loss": 0.5547, "num_input_tokens_seen": 5995757568, "step": 2859 }, { "epoch": 0.12476267968538107, "grad_norm": 12.071977615356445, "learning_rate": 2.1826307988752212e-07, "loss": 0.4157, "num_input_tokens_seen": 5997854720, "step": 2860 }, { "epoch": 0.1250339029020884, "grad_norm": 14.335684776306152, "learning_rate": 2.179904484135331e-07, "loss": 0.3816, "num_input_tokens_seen": 5999951872, "step": 2861 }, { "epoch": 0.12530512611879577, "grad_norm": 14.643807411193848, "learning_rate": 2.177180841281974e-07, "loss": 0.4579, "num_input_tokens_seen": 6002049024, "step": 2862 }, { "epoch": 0.12557634933550313, "grad_norm": 12.312231063842773, "learning_rate": 2.1744598725070347e-07, "loss": 0.4013, "num_input_tokens_seen": 6004146176, "step": 2863 }, { "epoch": 0.12584757255221046, "grad_norm": 17.71323013305664, "learning_rate": 2.1717415800002425e-07, "loss": 0.5469, "num_input_tokens_seen": 6006243328, "step": 2864 }, { "epoch": 0.12611879576891782, "grad_norm": 17.643123626708984, "learning_rate": 2.1690259659491768e-07, "loss": 0.5235, "num_input_tokens_seen": 6008340480, "step": 2865 }, { "epoch": 0.12639001898562516, "grad_norm": 13.04069995880127, "learning_rate": 2.166313032539261e-07, "loss": 0.3721, "num_input_tokens_seen": 6010437632, "step": 2866 }, { "epoch": 0.12666124220233252, "grad_norm": 15.557741165161133, "learning_rate": 2.1636027819537605e-07, "loss": 0.5084, "num_input_tokens_seen": 6012534784, "step": 2867 }, { "epoch": 0.12693246541903988, "grad_norm": 13.02405071258545, "learning_rate": 2.160895216373781e-07, "loss": 0.2844, "num_input_tokens_seen": 6014631936, "step": 2868 }, { "epoch": 0.12720368863574721, "grad_norm": 7.915323257446289, "learning_rate": 2.15819033797827e-07, "loss": 0.1902, "num_input_tokens_seen": 6016729088, "step": 2869 }, { "epoch": 0.12747491185245458, "grad_norm": 13.274603843688965, "learning_rate": 2.1554881489440092e-07, "loss": 0.4097, "num_input_tokens_seen": 6018826240, "step": 2870 }, { "epoch": 0.1277461350691619, "grad_norm": 14.94298267364502, "learning_rate": 2.1527886514456178e-07, "loss": 0.5581, "num_input_tokens_seen": 6020923392, "step": 2871 }, { "epoch": 0.12801735828586927, "grad_norm": 11.724287033081055, "learning_rate": 2.150091847655549e-07, "loss": 0.3464, "num_input_tokens_seen": 6023020544, "step": 2872 }, { "epoch": 0.12828858150257663, "grad_norm": 8.7376127243042, "learning_rate": 2.1473977397440878e-07, "loss": 0.1962, "num_input_tokens_seen": 6025117696, "step": 2873 }, { "epoch": 0.12855980471928397, "grad_norm": 13.356949806213379, "learning_rate": 2.14470632987935e-07, "loss": 0.5469, "num_input_tokens_seen": 6027214848, "step": 2874 }, { "epoch": 0.12883102793599133, "grad_norm": 11.348970413208008, "learning_rate": 2.1420176202272805e-07, "loss": 0.2122, "num_input_tokens_seen": 6029312000, "step": 2875 }, { "epoch": 0.12910225115269866, "grad_norm": 12.910688400268555, "learning_rate": 2.1393316129516482e-07, "loss": 0.4277, "num_input_tokens_seen": 6031409152, "step": 2876 }, { "epoch": 0.12937347436940602, "grad_norm": 18.22507667541504, "learning_rate": 2.1366483102140497e-07, "loss": 0.3707, "num_input_tokens_seen": 6033506304, "step": 2877 }, { "epoch": 0.12964469758611338, "grad_norm": 12.903253555297852, "learning_rate": 2.1339677141739074e-07, "loss": 0.3805, "num_input_tokens_seen": 6035603456, "step": 2878 }, { "epoch": 0.12991592080282072, "grad_norm": 15.44893741607666, "learning_rate": 2.1312898269884616e-07, "loss": 0.5398, "num_input_tokens_seen": 6037700608, "step": 2879 }, { "epoch": 0.13018714401952808, "grad_norm": 9.813923835754395, "learning_rate": 2.128614650812775e-07, "loss": 0.2239, "num_input_tokens_seen": 6039797760, "step": 2880 }, { "epoch": 0.1304583672362354, "grad_norm": 23.024755477905273, "learning_rate": 2.125942187799726e-07, "loss": 0.667, "num_input_tokens_seen": 6041894912, "step": 2881 }, { "epoch": 0.13072959045294277, "grad_norm": 21.844196319580078, "learning_rate": 2.123272440100012e-07, "loss": 0.7983, "num_input_tokens_seen": 6043992064, "step": 2882 }, { "epoch": 0.13100081366965013, "grad_norm": 16.548011779785156, "learning_rate": 2.1206054098621434e-07, "loss": 0.7041, "num_input_tokens_seen": 6046089216, "step": 2883 }, { "epoch": 0.13127203688635747, "grad_norm": 21.630081176757812, "learning_rate": 2.117941099232446e-07, "loss": 0.713, "num_input_tokens_seen": 6048186368, "step": 2884 }, { "epoch": 0.13154326010306483, "grad_norm": 12.728462219238281, "learning_rate": 2.1152795103550547e-07, "loss": 0.294, "num_input_tokens_seen": 6050283520, "step": 2885 }, { "epoch": 0.13181448331977216, "grad_norm": 33.04969787597656, "learning_rate": 2.1126206453719157e-07, "loss": 0.6548, "num_input_tokens_seen": 6052380672, "step": 2886 }, { "epoch": 0.13208570653647952, "grad_norm": 12.736434936523438, "learning_rate": 2.1099645064227817e-07, "loss": 0.3612, "num_input_tokens_seen": 6054477824, "step": 2887 }, { "epoch": 0.13235692975318689, "grad_norm": 13.483380317687988, "learning_rate": 2.1073110956452126e-07, "loss": 0.3295, "num_input_tokens_seen": 6056574976, "step": 2888 }, { "epoch": 0.13262815296989422, "grad_norm": 12.762460708618164, "learning_rate": 2.1046604151745723e-07, "loss": 0.3028, "num_input_tokens_seen": 6058672128, "step": 2889 }, { "epoch": 0.13289937618660158, "grad_norm": 12.392841339111328, "learning_rate": 2.1020124671440274e-07, "loss": 0.3086, "num_input_tokens_seen": 6060769280, "step": 2890 }, { "epoch": 0.13317059940330891, "grad_norm": 15.763758659362793, "learning_rate": 2.0993672536845458e-07, "loss": 0.3993, "num_input_tokens_seen": 6062866432, "step": 2891 }, { "epoch": 0.13344182262001628, "grad_norm": 14.49795150756836, "learning_rate": 2.0967247769248968e-07, "loss": 0.458, "num_input_tokens_seen": 6064963584, "step": 2892 }, { "epoch": 0.13371304583672364, "grad_norm": 17.557239532470703, "learning_rate": 2.0940850389916398e-07, "loss": 0.2814, "num_input_tokens_seen": 6067060736, "step": 2893 }, { "epoch": 0.13398426905343097, "grad_norm": 9.931197166442871, "learning_rate": 2.09144804200914e-07, "loss": 0.1937, "num_input_tokens_seen": 6069157888, "step": 2894 }, { "epoch": 0.13425549227013833, "grad_norm": 10.081891059875488, "learning_rate": 2.0888137880995514e-07, "loss": 0.2684, "num_input_tokens_seen": 6071255040, "step": 2895 }, { "epoch": 0.13452671548684567, "grad_norm": 18.273103713989258, "learning_rate": 2.086182279382821e-07, "loss": 0.6209, "num_input_tokens_seen": 6073352192, "step": 2896 }, { "epoch": 0.13479793870355303, "grad_norm": 16.057680130004883, "learning_rate": 2.083553517976686e-07, "loss": 0.4178, "num_input_tokens_seen": 6075449344, "step": 2897 }, { "epoch": 0.13506916192026036, "grad_norm": 14.120661735534668, "learning_rate": 2.0809275059966764e-07, "loss": 0.477, "num_input_tokens_seen": 6077546496, "step": 2898 }, { "epoch": 0.13534038513696772, "grad_norm": 10.962662696838379, "learning_rate": 2.0783042455561023e-07, "loss": 0.3011, "num_input_tokens_seen": 6079643648, "step": 2899 }, { "epoch": 0.13561160835367508, "grad_norm": 12.212385177612305, "learning_rate": 2.0756837387660676e-07, "loss": 0.361, "num_input_tokens_seen": 6081740800, "step": 2900 }, { "epoch": 0.13588283157038242, "grad_norm": 15.496916770935059, "learning_rate": 2.073065987735454e-07, "loss": 0.3815, "num_input_tokens_seen": 6083837952, "step": 2901 }, { "epoch": 0.13615405478708978, "grad_norm": 13.568929672241211, "learning_rate": 2.070450994570928e-07, "loss": 0.4771, "num_input_tokens_seen": 6085935104, "step": 2902 }, { "epoch": 0.1364252780037971, "grad_norm": 15.480619430541992, "learning_rate": 2.0678387613769397e-07, "loss": 0.4991, "num_input_tokens_seen": 6088032256, "step": 2903 }, { "epoch": 0.13669650122050447, "grad_norm": 17.832595825195312, "learning_rate": 2.0652292902557117e-07, "loss": 0.5667, "num_input_tokens_seen": 6090129408, "step": 2904 }, { "epoch": 0.13696772443721184, "grad_norm": 19.86845588684082, "learning_rate": 2.0626225833072487e-07, "loss": 0.8251, "num_input_tokens_seen": 6092226560, "step": 2905 }, { "epoch": 0.13723894765391917, "grad_norm": 14.052447319030762, "learning_rate": 2.0600186426293282e-07, "loss": 0.4079, "num_input_tokens_seen": 6094323712, "step": 2906 }, { "epoch": 0.13751017087062653, "grad_norm": 13.390406608581543, "learning_rate": 2.0574174703175034e-07, "loss": 0.2496, "num_input_tokens_seen": 6096420864, "step": 2907 }, { "epoch": 0.13778139408733386, "grad_norm": 16.510652542114258, "learning_rate": 2.0548190684650981e-07, "loss": 0.483, "num_input_tokens_seen": 6098518016, "step": 2908 }, { "epoch": 0.13805261730404123, "grad_norm": 14.933103561401367, "learning_rate": 2.052223439163208e-07, "loss": 0.5226, "num_input_tokens_seen": 6100615168, "step": 2909 }, { "epoch": 0.1383238405207486, "grad_norm": 10.2757568359375, "learning_rate": 2.049630584500696e-07, "loss": 0.3106, "num_input_tokens_seen": 6102712320, "step": 2910 }, { "epoch": 0.13859506373745592, "grad_norm": 21.213306427001953, "learning_rate": 2.0470405065641938e-07, "loss": 0.6228, "num_input_tokens_seen": 6104809472, "step": 2911 }, { "epoch": 0.13886628695416328, "grad_norm": 17.050752639770508, "learning_rate": 2.0444532074380973e-07, "loss": 0.5536, "num_input_tokens_seen": 6106906624, "step": 2912 }, { "epoch": 0.13913751017087062, "grad_norm": 11.27585506439209, "learning_rate": 2.0418686892045654e-07, "loss": 0.2669, "num_input_tokens_seen": 6109003776, "step": 2913 }, { "epoch": 0.13940873338757798, "grad_norm": 16.672374725341797, "learning_rate": 2.039286953943521e-07, "loss": 0.5225, "num_input_tokens_seen": 6111100928, "step": 2914 }, { "epoch": 0.13967995660428534, "grad_norm": 13.986382484436035, "learning_rate": 2.0367080037326472e-07, "loss": 0.3568, "num_input_tokens_seen": 6113198080, "step": 2915 }, { "epoch": 0.13995117982099267, "grad_norm": 12.307029724121094, "learning_rate": 2.0341318406473833e-07, "loss": 0.3359, "num_input_tokens_seen": 6115295232, "step": 2916 }, { "epoch": 0.14022240303770003, "grad_norm": 16.73017120361328, "learning_rate": 2.031558466760927e-07, "loss": 0.4768, "num_input_tokens_seen": 6117392384, "step": 2917 }, { "epoch": 0.14049362625440737, "grad_norm": 18.181156158447266, "learning_rate": 2.0289878841442325e-07, "loss": 0.5648, "num_input_tokens_seen": 6119489536, "step": 2918 }, { "epoch": 0.14076484947111473, "grad_norm": 14.489134788513184, "learning_rate": 2.0264200948660076e-07, "loss": 0.3855, "num_input_tokens_seen": 6121586688, "step": 2919 }, { "epoch": 0.1410360726878221, "grad_norm": 15.638324737548828, "learning_rate": 2.0238551009927125e-07, "loss": 0.4921, "num_input_tokens_seen": 6123683840, "step": 2920 }, { "epoch": 0.14130729590452942, "grad_norm": 13.121326446533203, "learning_rate": 2.021292904588553e-07, "loss": 0.369, "num_input_tokens_seen": 6125780992, "step": 2921 }, { "epoch": 0.14157851912123678, "grad_norm": 16.779006958007812, "learning_rate": 2.0187335077154905e-07, "loss": 0.5329, "num_input_tokens_seen": 6127878144, "step": 2922 }, { "epoch": 0.14184974233794412, "grad_norm": 13.377449035644531, "learning_rate": 2.0161769124332278e-07, "loss": 0.4986, "num_input_tokens_seen": 6129975296, "step": 2923 }, { "epoch": 0.14212096555465148, "grad_norm": 11.84175968170166, "learning_rate": 2.013623120799217e-07, "loss": 0.349, "num_input_tokens_seen": 6132072448, "step": 2924 }, { "epoch": 0.14239218877135884, "grad_norm": 15.083674430847168, "learning_rate": 2.0110721348686523e-07, "loss": 0.4303, "num_input_tokens_seen": 6134169600, "step": 2925 }, { "epoch": 0.14266341198806617, "grad_norm": 13.313584327697754, "learning_rate": 2.0085239566944702e-07, "loss": 0.1782, "num_input_tokens_seen": 6136266752, "step": 2926 }, { "epoch": 0.14293463520477354, "grad_norm": 15.516534805297852, "learning_rate": 2.0059785883273463e-07, "loss": 0.4328, "num_input_tokens_seen": 6138363904, "step": 2927 }, { "epoch": 0.14320585842148087, "grad_norm": 15.41818618774414, "learning_rate": 2.003436031815698e-07, "loss": 0.4351, "num_input_tokens_seen": 6140461056, "step": 2928 }, { "epoch": 0.14347708163818823, "grad_norm": 11.988521575927734, "learning_rate": 2.0008962892056762e-07, "loss": 0.3346, "num_input_tokens_seen": 6142558208, "step": 2929 }, { "epoch": 0.1437483048548956, "grad_norm": 15.152950286865234, "learning_rate": 1.998359362541171e-07, "loss": 0.5238, "num_input_tokens_seen": 6144655360, "step": 2930 }, { "epoch": 0.14401952807160293, "grad_norm": 14.549203872680664, "learning_rate": 1.9958252538638032e-07, "loss": 0.4193, "num_input_tokens_seen": 6146752512, "step": 2931 }, { "epoch": 0.1442907512883103, "grad_norm": 13.985295295715332, "learning_rate": 1.9932939652129283e-07, "loss": 0.2768, "num_input_tokens_seen": 6148849664, "step": 2932 }, { "epoch": 0.14456197450501762, "grad_norm": 16.646543502807617, "learning_rate": 1.990765498625629e-07, "loss": 0.6969, "num_input_tokens_seen": 6150946816, "step": 2933 }, { "epoch": 0.14483319772172498, "grad_norm": 12.219400405883789, "learning_rate": 1.988239856136719e-07, "loss": 0.3376, "num_input_tokens_seen": 6153043968, "step": 2934 }, { "epoch": 0.14510442093843234, "grad_norm": 14.528848648071289, "learning_rate": 1.9857170397787415e-07, "loss": 0.4568, "num_input_tokens_seen": 6155141120, "step": 2935 }, { "epoch": 0.14537564415513968, "grad_norm": 12.83976936340332, "learning_rate": 1.9831970515819625e-07, "loss": 0.3684, "num_input_tokens_seen": 6157238272, "step": 2936 }, { "epoch": 0.14564686737184704, "grad_norm": 11.202268600463867, "learning_rate": 1.980679893574373e-07, "loss": 0.2945, "num_input_tokens_seen": 6159335424, "step": 2937 }, { "epoch": 0.14591809058855437, "grad_norm": 22.032058715820312, "learning_rate": 1.9781655677816838e-07, "loss": 0.4701, "num_input_tokens_seen": 6161432576, "step": 2938 }, { "epoch": 0.14618931380526173, "grad_norm": 14.524810791015625, "learning_rate": 1.97565407622733e-07, "loss": 0.3425, "num_input_tokens_seen": 6163529728, "step": 2939 }, { "epoch": 0.14646053702196907, "grad_norm": 17.290315628051758, "learning_rate": 1.9731454209324644e-07, "loss": 0.6485, "num_input_tokens_seen": 6165626880, "step": 2940 }, { "epoch": 0.14673176023867643, "grad_norm": 25.62712860107422, "learning_rate": 1.9706396039159568e-07, "loss": 0.5547, "num_input_tokens_seen": 6167724032, "step": 2941 }, { "epoch": 0.1470029834553838, "grad_norm": 29.259971618652344, "learning_rate": 1.9681366271943945e-07, "loss": 0.3389, "num_input_tokens_seen": 6169821184, "step": 2942 }, { "epoch": 0.14727420667209112, "grad_norm": 24.600215911865234, "learning_rate": 1.965636492782077e-07, "loss": 0.2235, "num_input_tokens_seen": 6171918336, "step": 2943 }, { "epoch": 0.14754542988879848, "grad_norm": 20.362010955810547, "learning_rate": 1.963139202691017e-07, "loss": 0.7951, "num_input_tokens_seen": 6174015488, "step": 2944 }, { "epoch": 0.14781665310550582, "grad_norm": 22.120241165161133, "learning_rate": 1.9606447589309397e-07, "loss": 0.7222, "num_input_tokens_seen": 6176112640, "step": 2945 }, { "epoch": 0.14808787632221318, "grad_norm": 9.955594062805176, "learning_rate": 1.9581531635092773e-07, "loss": 0.2238, "num_input_tokens_seen": 6178209792, "step": 2946 }, { "epoch": 0.14835909953892054, "grad_norm": 13.511630058288574, "learning_rate": 1.9556644184311728e-07, "loss": 0.3301, "num_input_tokens_seen": 6180306944, "step": 2947 }, { "epoch": 0.14863032275562787, "grad_norm": 14.704965591430664, "learning_rate": 1.9531785256994716e-07, "loss": 0.4311, "num_input_tokens_seen": 6182404096, "step": 2948 }, { "epoch": 0.14890154597233524, "grad_norm": 16.59981918334961, "learning_rate": 1.9506954873147276e-07, "loss": 0.5568, "num_input_tokens_seen": 6184501248, "step": 2949 }, { "epoch": 0.14917276918904257, "grad_norm": 12.132022857666016, "learning_rate": 1.9482153052751921e-07, "loss": 0.4439, "num_input_tokens_seen": 6186598400, "step": 2950 }, { "epoch": 0.14944399240574993, "grad_norm": 22.458187103271484, "learning_rate": 1.9457379815768245e-07, "loss": 0.543, "num_input_tokens_seen": 6188695552, "step": 2951 }, { "epoch": 0.1497152156224573, "grad_norm": 18.78033447265625, "learning_rate": 1.9432635182132795e-07, "loss": 0.4977, "num_input_tokens_seen": 6190792704, "step": 2952 }, { "epoch": 0.14998643883916463, "grad_norm": 22.32543182373047, "learning_rate": 1.9407919171759112e-07, "loss": 0.345, "num_input_tokens_seen": 6192889856, "step": 2953 }, { "epoch": 0.150257662055872, "grad_norm": 18.59827423095703, "learning_rate": 1.9383231804537704e-07, "loss": 0.4047, "num_input_tokens_seen": 6194987008, "step": 2954 }, { "epoch": 0.15052888527257932, "grad_norm": 12.72734260559082, "learning_rate": 1.9358573100336034e-07, "loss": 0.347, "num_input_tokens_seen": 6197084160, "step": 2955 }, { "epoch": 0.15080010848928668, "grad_norm": 14.508316040039062, "learning_rate": 1.933394307899846e-07, "loss": 0.4512, "num_input_tokens_seen": 6199181312, "step": 2956 }, { "epoch": 0.15107133170599404, "grad_norm": 19.344667434692383, "learning_rate": 1.9309341760346304e-07, "loss": 0.7429, "num_input_tokens_seen": 6201278464, "step": 2957 }, { "epoch": 0.15134255492270138, "grad_norm": 11.575425148010254, "learning_rate": 1.9284769164177767e-07, "loss": 0.2508, "num_input_tokens_seen": 6203375616, "step": 2958 }, { "epoch": 0.15161377813940874, "grad_norm": 19.255504608154297, "learning_rate": 1.9260225310267925e-07, "loss": 0.5238, "num_input_tokens_seen": 6205472768, "step": 2959 }, { "epoch": 0.15188500135611607, "grad_norm": 18.295150756835938, "learning_rate": 1.9235710218368784e-07, "loss": 0.5423, "num_input_tokens_seen": 6207569920, "step": 2960 }, { "epoch": 0.15215622457282343, "grad_norm": 16.604528427124023, "learning_rate": 1.9211223908209114e-07, "loss": 0.4414, "num_input_tokens_seen": 6209667072, "step": 2961 }, { "epoch": 0.1524274477895308, "grad_norm": 28.41756248474121, "learning_rate": 1.9186766399494581e-07, "loss": 0.4075, "num_input_tokens_seen": 6211764224, "step": 2962 }, { "epoch": 0.15269867100623813, "grad_norm": 13.66691780090332, "learning_rate": 1.9162337711907657e-07, "loss": 0.4251, "num_input_tokens_seen": 6213861376, "step": 2963 }, { "epoch": 0.1529698942229455, "grad_norm": 14.04883098602295, "learning_rate": 1.9137937865107606e-07, "loss": 0.19, "num_input_tokens_seen": 6215958528, "step": 2964 }, { "epoch": 0.15324111743965282, "grad_norm": 12.723421096801758, "learning_rate": 1.9113566878730515e-07, "loss": 0.3229, "num_input_tokens_seen": 6218055680, "step": 2965 }, { "epoch": 0.15351234065636019, "grad_norm": 28.06492042541504, "learning_rate": 1.9089224772389223e-07, "loss": 0.9522, "num_input_tokens_seen": 6220152832, "step": 2966 }, { "epoch": 0.15378356387306755, "grad_norm": 13.24654483795166, "learning_rate": 1.9064911565673328e-07, "loss": 0.3005, "num_input_tokens_seen": 6222249984, "step": 2967 }, { "epoch": 0.15405478708977488, "grad_norm": 15.379100799560547, "learning_rate": 1.9040627278149168e-07, "loss": 0.4348, "num_input_tokens_seen": 6224347136, "step": 2968 }, { "epoch": 0.15432601030648224, "grad_norm": 23.722576141357422, "learning_rate": 1.9016371929359824e-07, "loss": 0.9692, "num_input_tokens_seen": 6226444288, "step": 2969 }, { "epoch": 0.15459723352318958, "grad_norm": 19.0823917388916, "learning_rate": 1.8992145538825066e-07, "loss": 0.6268, "num_input_tokens_seen": 6228541440, "step": 2970 }, { "epoch": 0.15486845673989694, "grad_norm": 15.874011993408203, "learning_rate": 1.8967948126041383e-07, "loss": 0.4841, "num_input_tokens_seen": 6230638592, "step": 2971 }, { "epoch": 0.1551396799566043, "grad_norm": 14.604007720947266, "learning_rate": 1.8943779710481938e-07, "loss": 0.5069, "num_input_tokens_seen": 6232735744, "step": 2972 }, { "epoch": 0.15541090317331163, "grad_norm": 20.07442283630371, "learning_rate": 1.891964031159653e-07, "loss": 0.7613, "num_input_tokens_seen": 6234832896, "step": 2973 }, { "epoch": 0.155682126390019, "grad_norm": 18.36520767211914, "learning_rate": 1.8895529948811638e-07, "loss": 0.6046, "num_input_tokens_seen": 6236930048, "step": 2974 }, { "epoch": 0.15595334960672633, "grad_norm": 15.209831237792969, "learning_rate": 1.8871448641530353e-07, "loss": 0.4244, "num_input_tokens_seen": 6239027200, "step": 2975 }, { "epoch": 0.1562245728234337, "grad_norm": 11.920095443725586, "learning_rate": 1.8847396409132423e-07, "loss": 0.3894, "num_input_tokens_seen": 6241124352, "step": 2976 }, { "epoch": 0.15649579604014105, "grad_norm": 13.935537338256836, "learning_rate": 1.8823373270974164e-07, "loss": 0.4652, "num_input_tokens_seen": 6243221504, "step": 2977 }, { "epoch": 0.15676701925684838, "grad_norm": 11.79638957977295, "learning_rate": 1.8799379246388463e-07, "loss": 0.3007, "num_input_tokens_seen": 6245318656, "step": 2978 }, { "epoch": 0.15703824247355574, "grad_norm": 20.775779724121094, "learning_rate": 1.8775414354684804e-07, "loss": 0.8416, "num_input_tokens_seen": 6247415808, "step": 2979 }, { "epoch": 0.15730946569026308, "grad_norm": 9.75412368774414, "learning_rate": 1.875147861514923e-07, "loss": 0.2322, "num_input_tokens_seen": 6249512960, "step": 2980 }, { "epoch": 0.15758068890697044, "grad_norm": 11.689945220947266, "learning_rate": 1.8727572047044308e-07, "loss": 0.3366, "num_input_tokens_seen": 6251610112, "step": 2981 }, { "epoch": 0.15785191212367777, "grad_norm": 12.28046703338623, "learning_rate": 1.8703694669609133e-07, "loss": 0.3674, "num_input_tokens_seen": 6253707264, "step": 2982 }, { "epoch": 0.15812313534038513, "grad_norm": 14.490015029907227, "learning_rate": 1.8679846502059306e-07, "loss": 0.4055, "num_input_tokens_seen": 6255804416, "step": 2983 }, { "epoch": 0.1583943585570925, "grad_norm": 14.019413948059082, "learning_rate": 1.8656027563586934e-07, "loss": 0.4061, "num_input_tokens_seen": 6257901568, "step": 2984 }, { "epoch": 0.15866558177379983, "grad_norm": 10.86307144165039, "learning_rate": 1.863223787336059e-07, "loss": 0.1615, "num_input_tokens_seen": 6259998720, "step": 2985 }, { "epoch": 0.1589368049905072, "grad_norm": 16.09626007080078, "learning_rate": 1.8608477450525308e-07, "loss": 0.5691, "num_input_tokens_seen": 6262095872, "step": 2986 }, { "epoch": 0.15920802820721452, "grad_norm": 11.85105037689209, "learning_rate": 1.8584746314202574e-07, "loss": 0.3223, "num_input_tokens_seen": 6264193024, "step": 2987 }, { "epoch": 0.1594792514239219, "grad_norm": 14.974807739257812, "learning_rate": 1.8561044483490301e-07, "loss": 0.5907, "num_input_tokens_seen": 6266290176, "step": 2988 }, { "epoch": 0.15975047464062925, "grad_norm": 13.03945255279541, "learning_rate": 1.8537371977462835e-07, "loss": 0.4826, "num_input_tokens_seen": 6268387328, "step": 2989 }, { "epoch": 0.16002169785733658, "grad_norm": 19.678932189941406, "learning_rate": 1.8513728815170885e-07, "loss": 0.6468, "num_input_tokens_seen": 6270484480, "step": 2990 }, { "epoch": 0.16029292107404394, "grad_norm": 22.712543487548828, "learning_rate": 1.8490115015641566e-07, "loss": 0.5548, "num_input_tokens_seen": 6272581632, "step": 2991 }, { "epoch": 0.16056414429075128, "grad_norm": 22.17213249206543, "learning_rate": 1.846653059787839e-07, "loss": 0.8835, "num_input_tokens_seen": 6274678784, "step": 2992 }, { "epoch": 0.16083536750745864, "grad_norm": 22.142024993896484, "learning_rate": 1.8442975580861185e-07, "loss": 0.791, "num_input_tokens_seen": 6276775936, "step": 2993 }, { "epoch": 0.161106590724166, "grad_norm": 16.552152633666992, "learning_rate": 1.8419449983546136e-07, "loss": 0.3504, "num_input_tokens_seen": 6278873088, "step": 2994 }, { "epoch": 0.16137781394087333, "grad_norm": 12.928352355957031, "learning_rate": 1.8395953824865762e-07, "loss": 0.3077, "num_input_tokens_seen": 6280970240, "step": 2995 }, { "epoch": 0.1616490371575807, "grad_norm": 18.581602096557617, "learning_rate": 1.8372487123728844e-07, "loss": 0.5233, "num_input_tokens_seen": 6283067392, "step": 2996 }, { "epoch": 0.16192026037428803, "grad_norm": 14.512155532836914, "learning_rate": 1.8349049899020514e-07, "loss": 0.4539, "num_input_tokens_seen": 6285164544, "step": 2997 }, { "epoch": 0.1621914835909954, "grad_norm": 11.53394603729248, "learning_rate": 1.832564216960215e-07, "loss": 0.3514, "num_input_tokens_seen": 6287261696, "step": 2998 }, { "epoch": 0.16246270680770275, "grad_norm": 13.360198020935059, "learning_rate": 1.8302263954311408e-07, "loss": 0.3348, "num_input_tokens_seen": 6289358848, "step": 2999 }, { "epoch": 0.16273393002441008, "grad_norm": 17.085603713989258, "learning_rate": 1.827891527196218e-07, "loss": 0.7366, "num_input_tokens_seen": 6291456000, "step": 3000 }, { "epoch": 0.16300515324111745, "grad_norm": 14.20055866241455, "learning_rate": 1.8255596141344605e-07, "loss": 0.2508, "num_input_tokens_seen": 6293553152, "step": 3001 }, { "epoch": 0.16327637645782478, "grad_norm": 14.382533073425293, "learning_rate": 1.823230658122502e-07, "loss": 0.4234, "num_input_tokens_seen": 6295650304, "step": 3002 }, { "epoch": 0.16354759967453214, "grad_norm": 19.175596237182617, "learning_rate": 1.820904661034599e-07, "loss": 0.4568, "num_input_tokens_seen": 6297747456, "step": 3003 }, { "epoch": 0.1638188228912395, "grad_norm": 13.114893913269043, "learning_rate": 1.8185816247426245e-07, "loss": 0.3285, "num_input_tokens_seen": 6299844608, "step": 3004 }, { "epoch": 0.16409004610794684, "grad_norm": 19.748310089111328, "learning_rate": 1.8162615511160701e-07, "loss": 0.6501, "num_input_tokens_seen": 6301941760, "step": 3005 }, { "epoch": 0.1643612693246542, "grad_norm": 16.454687118530273, "learning_rate": 1.8139444420220438e-07, "loss": 0.332, "num_input_tokens_seen": 6304038912, "step": 3006 }, { "epoch": 0.16463249254136153, "grad_norm": 19.270048141479492, "learning_rate": 1.8116302993252636e-07, "loss": 0.8047, "num_input_tokens_seen": 6306136064, "step": 3007 }, { "epoch": 0.1649037157580689, "grad_norm": 10.644708633422852, "learning_rate": 1.809319124888067e-07, "loss": 0.2935, "num_input_tokens_seen": 6308233216, "step": 3008 }, { "epoch": 0.16517493897477625, "grad_norm": 10.366632461547852, "learning_rate": 1.8070109205703972e-07, "loss": 0.2606, "num_input_tokens_seen": 6310330368, "step": 3009 }, { "epoch": 0.1654461621914836, "grad_norm": 16.61554527282715, "learning_rate": 1.80470568822981e-07, "loss": 0.5537, "num_input_tokens_seen": 6312427520, "step": 3010 }, { "epoch": 0.16571738540819095, "grad_norm": 20.610681533813477, "learning_rate": 1.8024034297214686e-07, "loss": 0.86, "num_input_tokens_seen": 6314524672, "step": 3011 }, { "epoch": 0.16598860862489828, "grad_norm": 15.841955184936523, "learning_rate": 1.8001041468981442e-07, "loss": 0.4544, "num_input_tokens_seen": 6316621824, "step": 3012 }, { "epoch": 0.16625983184160564, "grad_norm": 16.40627670288086, "learning_rate": 1.79780784161021e-07, "loss": 0.4465, "num_input_tokens_seen": 6318718976, "step": 3013 }, { "epoch": 0.166531055058313, "grad_norm": 18.409828186035156, "learning_rate": 1.7955145157056456e-07, "loss": 0.6821, "num_input_tokens_seen": 6320816128, "step": 3014 }, { "epoch": 0.16680227827502034, "grad_norm": 17.844806671142578, "learning_rate": 1.793224171030033e-07, "loss": 0.6167, "num_input_tokens_seen": 6322913280, "step": 3015 }, { "epoch": 0.1670735014917277, "grad_norm": 12.6576566696167, "learning_rate": 1.7909368094265532e-07, "loss": 0.3874, "num_input_tokens_seen": 6325010432, "step": 3016 }, { "epoch": 0.16734472470843503, "grad_norm": 14.177213668823242, "learning_rate": 1.788652432735991e-07, "loss": 0.3264, "num_input_tokens_seen": 6327107584, "step": 3017 }, { "epoch": 0.1676159479251424, "grad_norm": 16.917213439941406, "learning_rate": 1.7863710427967222e-07, "loss": 0.2328, "num_input_tokens_seen": 6329204736, "step": 3018 }, { "epoch": 0.16788717114184976, "grad_norm": 21.851543426513672, "learning_rate": 1.784092641444725e-07, "loss": 0.5402, "num_input_tokens_seen": 6331301888, "step": 3019 }, { "epoch": 0.1681583943585571, "grad_norm": 21.96310043334961, "learning_rate": 1.7818172305135683e-07, "loss": 0.5764, "num_input_tokens_seen": 6333399040, "step": 3020 }, { "epoch": 0.16842961757526445, "grad_norm": 16.358335494995117, "learning_rate": 1.7795448118344174e-07, "loss": 0.4508, "num_input_tokens_seen": 6335496192, "step": 3021 }, { "epoch": 0.16870084079197178, "grad_norm": 19.15180206298828, "learning_rate": 1.7772753872360286e-07, "loss": 0.5401, "num_input_tokens_seen": 6337593344, "step": 3022 }, { "epoch": 0.16897206400867915, "grad_norm": 15.562478065490723, "learning_rate": 1.7750089585447473e-07, "loss": 0.4368, "num_input_tokens_seen": 6339690496, "step": 3023 }, { "epoch": 0.16924328722538648, "grad_norm": 14.692290306091309, "learning_rate": 1.77274552758451e-07, "loss": 0.3316, "num_input_tokens_seen": 6341787648, "step": 3024 }, { "epoch": 0.16951451044209384, "grad_norm": 12.466604232788086, "learning_rate": 1.770485096176839e-07, "loss": 0.3341, "num_input_tokens_seen": 6343884800, "step": 3025 }, { "epoch": 0.1697857336588012, "grad_norm": 15.91280746459961, "learning_rate": 1.7682276661408448e-07, "loss": 0.6699, "num_input_tokens_seen": 6345981952, "step": 3026 }, { "epoch": 0.17005695687550854, "grad_norm": 17.506284713745117, "learning_rate": 1.7659732392932202e-07, "loss": 0.4563, "num_input_tokens_seen": 6348079104, "step": 3027 }, { "epoch": 0.1703281800922159, "grad_norm": 11.954529762268066, "learning_rate": 1.7637218174482422e-07, "loss": 0.3093, "num_input_tokens_seen": 6350176256, "step": 3028 }, { "epoch": 0.17059940330892323, "grad_norm": 12.980586051940918, "learning_rate": 1.7614734024177707e-07, "loss": 0.4789, "num_input_tokens_seen": 6352273408, "step": 3029 }, { "epoch": 0.1708706265256306, "grad_norm": 16.40062713623047, "learning_rate": 1.7592279960112424e-07, "loss": 0.6716, "num_input_tokens_seen": 6354370560, "step": 3030 }, { "epoch": 0.17114184974233795, "grad_norm": 17.116609573364258, "learning_rate": 1.7569856000356758e-07, "loss": 0.3462, "num_input_tokens_seen": 6356467712, "step": 3031 }, { "epoch": 0.1714130729590453, "grad_norm": 17.028945922851562, "learning_rate": 1.754746216295665e-07, "loss": 0.64, "num_input_tokens_seen": 6358564864, "step": 3032 }, { "epoch": 0.17168429617575265, "grad_norm": 21.552492141723633, "learning_rate": 1.7525098465933834e-07, "loss": 0.7373, "num_input_tokens_seen": 6360662016, "step": 3033 }, { "epoch": 0.17195551939245998, "grad_norm": 15.610780715942383, "learning_rate": 1.750276492728575e-07, "loss": 0.2999, "num_input_tokens_seen": 6362759168, "step": 3034 }, { "epoch": 0.17222674260916734, "grad_norm": 11.220949172973633, "learning_rate": 1.748046156498557e-07, "loss": 0.3368, "num_input_tokens_seen": 6364856320, "step": 3035 }, { "epoch": 0.1724979658258747, "grad_norm": 16.892784118652344, "learning_rate": 1.74581883969822e-07, "loss": 0.571, "num_input_tokens_seen": 6366953472, "step": 3036 }, { "epoch": 0.17276918904258204, "grad_norm": 14.306039810180664, "learning_rate": 1.7435945441200232e-07, "loss": 0.4861, "num_input_tokens_seen": 6369050624, "step": 3037 }, { "epoch": 0.1730404122592894, "grad_norm": 16.46154022216797, "learning_rate": 1.7413732715539954e-07, "loss": 0.3976, "num_input_tokens_seen": 6371147776, "step": 3038 }, { "epoch": 0.17331163547599673, "grad_norm": 18.606840133666992, "learning_rate": 1.7391550237877326e-07, "loss": 0.6797, "num_input_tokens_seen": 6373244928, "step": 3039 }, { "epoch": 0.1735828586927041, "grad_norm": 15.219120979309082, "learning_rate": 1.7369398026063958e-07, "loss": 0.4798, "num_input_tokens_seen": 6375342080, "step": 3040 }, { "epoch": 0.17385408190941146, "grad_norm": 15.45386791229248, "learning_rate": 1.7347276097927105e-07, "loss": 0.6304, "num_input_tokens_seen": 6377439232, "step": 3041 }, { "epoch": 0.1741253051261188, "grad_norm": 15.114662170410156, "learning_rate": 1.732518447126966e-07, "loss": 0.4832, "num_input_tokens_seen": 6379536384, "step": 3042 }, { "epoch": 0.17439652834282615, "grad_norm": 20.997955322265625, "learning_rate": 1.730312316387012e-07, "loss": 0.4316, "num_input_tokens_seen": 6381633536, "step": 3043 }, { "epoch": 0.17466775155953349, "grad_norm": 13.511618614196777, "learning_rate": 1.7281092193482582e-07, "loss": 0.4517, "num_input_tokens_seen": 6383730688, "step": 3044 }, { "epoch": 0.17493897477624085, "grad_norm": 17.852827072143555, "learning_rate": 1.7259091577836737e-07, "loss": 0.5438, "num_input_tokens_seen": 6385827840, "step": 3045 }, { "epoch": 0.1752101979929482, "grad_norm": 15.084151268005371, "learning_rate": 1.723712133463785e-07, "loss": 0.5421, "num_input_tokens_seen": 6387924992, "step": 3046 }, { "epoch": 0.17548142120965554, "grad_norm": 12.10033893585205, "learning_rate": 1.7215181481566727e-07, "loss": 0.304, "num_input_tokens_seen": 6390022144, "step": 3047 }, { "epoch": 0.1757526444263629, "grad_norm": 12.1007719039917, "learning_rate": 1.719327203627971e-07, "loss": 0.2765, "num_input_tokens_seen": 6392119296, "step": 3048 }, { "epoch": 0.17602386764307024, "grad_norm": 14.346907615661621, "learning_rate": 1.7171393016408715e-07, "loss": 0.3953, "num_input_tokens_seen": 6394216448, "step": 3049 }, { "epoch": 0.1762950908597776, "grad_norm": 13.626975059509277, "learning_rate": 1.7149544439561135e-07, "loss": 0.3358, "num_input_tokens_seen": 6396313600, "step": 3050 }, { "epoch": 0.17656631407648496, "grad_norm": 15.245820999145508, "learning_rate": 1.7127726323319874e-07, "loss": 0.3938, "num_input_tokens_seen": 6398410752, "step": 3051 }, { "epoch": 0.1768375372931923, "grad_norm": 9.872340202331543, "learning_rate": 1.7105938685243339e-07, "loss": 0.2623, "num_input_tokens_seen": 6400507904, "step": 3052 }, { "epoch": 0.17710876050989965, "grad_norm": 15.143594741821289, "learning_rate": 1.7084181542865355e-07, "loss": 0.5046, "num_input_tokens_seen": 6402605056, "step": 3053 }, { "epoch": 0.177379983726607, "grad_norm": 17.209291458129883, "learning_rate": 1.7062454913695264e-07, "loss": 0.5746, "num_input_tokens_seen": 6404702208, "step": 3054 }, { "epoch": 0.17765120694331435, "grad_norm": 14.591382026672363, "learning_rate": 1.7040758815217827e-07, "loss": 0.3467, "num_input_tokens_seen": 6406799360, "step": 3055 }, { "epoch": 0.1779224301600217, "grad_norm": 16.17945098876953, "learning_rate": 1.7019093264893236e-07, "loss": 0.6641, "num_input_tokens_seen": 6408896512, "step": 3056 }, { "epoch": 0.17819365337672904, "grad_norm": 11.522351264953613, "learning_rate": 1.6997458280157113e-07, "loss": 0.3645, "num_input_tokens_seen": 6410993664, "step": 3057 }, { "epoch": 0.1784648765934364, "grad_norm": 19.618732452392578, "learning_rate": 1.6975853878420457e-07, "loss": 0.6604, "num_input_tokens_seen": 6413090816, "step": 3058 }, { "epoch": 0.17873609981014374, "grad_norm": 15.74427604675293, "learning_rate": 1.6954280077069676e-07, "loss": 0.3712, "num_input_tokens_seen": 6415187968, "step": 3059 }, { "epoch": 0.1790073230268511, "grad_norm": 16.00514793395996, "learning_rate": 1.6932736893466536e-07, "loss": 0.4623, "num_input_tokens_seen": 6417285120, "step": 3060 }, { "epoch": 0.17927854624355846, "grad_norm": 12.272290229797363, "learning_rate": 1.691122434494818e-07, "loss": 0.3603, "num_input_tokens_seen": 6419382272, "step": 3061 }, { "epoch": 0.1795497694602658, "grad_norm": 15.91757583618164, "learning_rate": 1.688974244882709e-07, "loss": 0.5306, "num_input_tokens_seen": 6421479424, "step": 3062 }, { "epoch": 0.17982099267697316, "grad_norm": 15.56859302520752, "learning_rate": 1.686829122239108e-07, "loss": 0.4703, "num_input_tokens_seen": 6423576576, "step": 3063 }, { "epoch": 0.1800922158936805, "grad_norm": 18.83671760559082, "learning_rate": 1.684687068290326e-07, "loss": 0.6076, "num_input_tokens_seen": 6425673728, "step": 3064 }, { "epoch": 0.18036343911038785, "grad_norm": 14.046809196472168, "learning_rate": 1.682548084760208e-07, "loss": 0.3576, "num_input_tokens_seen": 6427770880, "step": 3065 }, { "epoch": 0.18063466232709519, "grad_norm": 12.50328540802002, "learning_rate": 1.6804121733701277e-07, "loss": 0.3457, "num_input_tokens_seen": 6429868032, "step": 3066 }, { "epoch": 0.18090588554380255, "grad_norm": 22.34259605407715, "learning_rate": 1.678279335838983e-07, "loss": 1.0324, "num_input_tokens_seen": 6431965184, "step": 3067 }, { "epoch": 0.1811771087605099, "grad_norm": 12.437803268432617, "learning_rate": 1.676149573883202e-07, "loss": 0.3294, "num_input_tokens_seen": 6434062336, "step": 3068 }, { "epoch": 0.18144833197721724, "grad_norm": 19.77998161315918, "learning_rate": 1.674022889216737e-07, "loss": 0.4922, "num_input_tokens_seen": 6436159488, "step": 3069 }, { "epoch": 0.1817195551939246, "grad_norm": 16.024463653564453, "learning_rate": 1.67189928355106e-07, "loss": 0.3612, "num_input_tokens_seen": 6438256640, "step": 3070 }, { "epoch": 0.18199077841063194, "grad_norm": 18.659027099609375, "learning_rate": 1.66977875859517e-07, "loss": 0.5333, "num_input_tokens_seen": 6440353792, "step": 3071 }, { "epoch": 0.1822620016273393, "grad_norm": 12.65479850769043, "learning_rate": 1.6676613160555846e-07, "loss": 0.3801, "num_input_tokens_seen": 6442450944, "step": 3072 }, { "epoch": 0.18253322484404666, "grad_norm": 16.552722930908203, "learning_rate": 1.665546957636339e-07, "loss": 0.6134, "num_input_tokens_seen": 6444548096, "step": 3073 }, { "epoch": 0.182804448060754, "grad_norm": 11.289351463317871, "learning_rate": 1.6634356850389926e-07, "loss": 0.3346, "num_input_tokens_seen": 6446645248, "step": 3074 }, { "epoch": 0.18307567127746135, "grad_norm": 13.140058517456055, "learning_rate": 1.6613274999626134e-07, "loss": 0.4239, "num_input_tokens_seen": 6448742400, "step": 3075 }, { "epoch": 0.1833468944941687, "grad_norm": 16.135107040405273, "learning_rate": 1.65922240410379e-07, "loss": 0.5723, "num_input_tokens_seen": 6450839552, "step": 3076 }, { "epoch": 0.18361811771087605, "grad_norm": 12.902073860168457, "learning_rate": 1.6571203991566234e-07, "loss": 0.3716, "num_input_tokens_seen": 6452936704, "step": 3077 }, { "epoch": 0.1838893409275834, "grad_norm": 13.520687103271484, "learning_rate": 1.6550214868127276e-07, "loss": 0.272, "num_input_tokens_seen": 6455033856, "step": 3078 }, { "epoch": 0.18416056414429074, "grad_norm": 15.4642333984375, "learning_rate": 1.6529256687612264e-07, "loss": 0.4952, "num_input_tokens_seen": 6457131008, "step": 3079 }, { "epoch": 0.1844317873609981, "grad_norm": 11.49740982055664, "learning_rate": 1.6508329466887548e-07, "loss": 0.321, "num_input_tokens_seen": 6459228160, "step": 3080 }, { "epoch": 0.18470301057770544, "grad_norm": 19.20050048828125, "learning_rate": 1.6487433222794566e-07, "loss": 0.5862, "num_input_tokens_seen": 6461325312, "step": 3081 }, { "epoch": 0.1849742337944128, "grad_norm": 9.488527297973633, "learning_rate": 1.6466567972149806e-07, "loss": 0.2648, "num_input_tokens_seen": 6463422464, "step": 3082 }, { "epoch": 0.18524545701112016, "grad_norm": 12.057193756103516, "learning_rate": 1.6445733731744842e-07, "loss": 0.3546, "num_input_tokens_seen": 6465519616, "step": 3083 }, { "epoch": 0.1855166802278275, "grad_norm": 10.167229652404785, "learning_rate": 1.642493051834627e-07, "loss": 0.2825, "num_input_tokens_seen": 6467616768, "step": 3084 }, { "epoch": 0.18578790344453486, "grad_norm": 15.535594940185547, "learning_rate": 1.6404158348695729e-07, "loss": 0.4639, "num_input_tokens_seen": 6469713920, "step": 3085 }, { "epoch": 0.1860591266612422, "grad_norm": 14.652868270874023, "learning_rate": 1.6383417239509878e-07, "loss": 0.5855, "num_input_tokens_seen": 6471811072, "step": 3086 }, { "epoch": 0.18633034987794955, "grad_norm": 16.557931900024414, "learning_rate": 1.6362707207480347e-07, "loss": 0.729, "num_input_tokens_seen": 6473908224, "step": 3087 }, { "epoch": 0.18660157309465691, "grad_norm": 20.1795654296875, "learning_rate": 1.6342028269273802e-07, "loss": 0.694, "num_input_tokens_seen": 6476005376, "step": 3088 }, { "epoch": 0.18687279631136425, "grad_norm": 16.401769638061523, "learning_rate": 1.6321380441531852e-07, "loss": 0.6733, "num_input_tokens_seen": 6478102528, "step": 3089 }, { "epoch": 0.1871440195280716, "grad_norm": 12.176185607910156, "learning_rate": 1.63007637408711e-07, "loss": 0.3886, "num_input_tokens_seen": 6480199680, "step": 3090 }, { "epoch": 0.18741524274477894, "grad_norm": 14.493315696716309, "learning_rate": 1.6280178183883081e-07, "loss": 0.4193, "num_input_tokens_seen": 6482296832, "step": 3091 }, { "epoch": 0.1876864659614863, "grad_norm": 13.448966979980469, "learning_rate": 1.6259623787134258e-07, "loss": 0.3855, "num_input_tokens_seen": 6484393984, "step": 3092 }, { "epoch": 0.18795768917819367, "grad_norm": 14.371041297912598, "learning_rate": 1.6239100567166026e-07, "loss": 0.3938, "num_input_tokens_seen": 6486491136, "step": 3093 }, { "epoch": 0.188228912394901, "grad_norm": 11.803318977355957, "learning_rate": 1.6218608540494693e-07, "loss": 0.3518, "num_input_tokens_seen": 6488588288, "step": 3094 }, { "epoch": 0.18850013561160836, "grad_norm": 16.85552978515625, "learning_rate": 1.619814772361147e-07, "loss": 0.3771, "num_input_tokens_seen": 6490685440, "step": 3095 }, { "epoch": 0.1887713588283157, "grad_norm": 15.620574951171875, "learning_rate": 1.6177718132982441e-07, "loss": 0.6079, "num_input_tokens_seen": 6492782592, "step": 3096 }, { "epoch": 0.18904258204502306, "grad_norm": 20.06139373779297, "learning_rate": 1.6157319785048555e-07, "loss": 0.5848, "num_input_tokens_seen": 6494879744, "step": 3097 }, { "epoch": 0.18931380526173042, "grad_norm": 14.808846473693848, "learning_rate": 1.6136952696225634e-07, "loss": 0.3535, "num_input_tokens_seen": 6496976896, "step": 3098 }, { "epoch": 0.18958502847843775, "grad_norm": 16.144725799560547, "learning_rate": 1.6116616882904332e-07, "loss": 0.5276, "num_input_tokens_seen": 6499074048, "step": 3099 }, { "epoch": 0.1898562516951451, "grad_norm": 14.656734466552734, "learning_rate": 1.6096312361450142e-07, "loss": 0.4783, "num_input_tokens_seen": 6501171200, "step": 3100 }, { "epoch": 0.19012747491185245, "grad_norm": 16.5491943359375, "learning_rate": 1.6076039148203373e-07, "loss": 0.453, "num_input_tokens_seen": 6503268352, "step": 3101 }, { "epoch": 0.1903986981285598, "grad_norm": 9.266581535339355, "learning_rate": 1.6055797259479125e-07, "loss": 0.2459, "num_input_tokens_seen": 6505365504, "step": 3102 }, { "epoch": 0.19066992134526717, "grad_norm": 22.371381759643555, "learning_rate": 1.6035586711567318e-07, "loss": 0.6326, "num_input_tokens_seen": 6507462656, "step": 3103 }, { "epoch": 0.1909411445619745, "grad_norm": 17.789241790771484, "learning_rate": 1.601540752073261e-07, "loss": 0.6873, "num_input_tokens_seen": 6509559808, "step": 3104 }, { "epoch": 0.19121236777868186, "grad_norm": 17.25879669189453, "learning_rate": 1.5995259703214445e-07, "loss": 0.284, "num_input_tokens_seen": 6511656960, "step": 3105 }, { "epoch": 0.1914835909953892, "grad_norm": 12.322155952453613, "learning_rate": 1.5975143275227039e-07, "loss": 0.3976, "num_input_tokens_seen": 6513754112, "step": 3106 }, { "epoch": 0.19175481421209656, "grad_norm": 18.380308151245117, "learning_rate": 1.5955058252959318e-07, "loss": 0.4834, "num_input_tokens_seen": 6515851264, "step": 3107 }, { "epoch": 0.1920260374288039, "grad_norm": 17.78908920288086, "learning_rate": 1.5935004652574947e-07, "loss": 0.5889, "num_input_tokens_seen": 6517948416, "step": 3108 }, { "epoch": 0.19229726064551125, "grad_norm": 11.990644454956055, "learning_rate": 1.591498249021231e-07, "loss": 0.2529, "num_input_tokens_seen": 6520045568, "step": 3109 }, { "epoch": 0.19256848386221861, "grad_norm": 20.534889221191406, "learning_rate": 1.5894991781984456e-07, "loss": 0.6632, "num_input_tokens_seen": 6522142720, "step": 3110 }, { "epoch": 0.19283970707892595, "grad_norm": 7.436341285705566, "learning_rate": 1.587503254397916e-07, "loss": 0.1396, "num_input_tokens_seen": 6524239872, "step": 3111 }, { "epoch": 0.1931109302956333, "grad_norm": 11.600310325622559, "learning_rate": 1.585510479225886e-07, "loss": 0.242, "num_input_tokens_seen": 6526337024, "step": 3112 }, { "epoch": 0.19338215351234064, "grad_norm": 13.862635612487793, "learning_rate": 1.5835208542860648e-07, "loss": 0.3614, "num_input_tokens_seen": 6528434176, "step": 3113 }, { "epoch": 0.193653376729048, "grad_norm": 10.956169128417969, "learning_rate": 1.581534381179627e-07, "loss": 0.3421, "num_input_tokens_seen": 6530531328, "step": 3114 }, { "epoch": 0.19392459994575537, "grad_norm": 17.733055114746094, "learning_rate": 1.5795510615052104e-07, "loss": 0.5629, "num_input_tokens_seen": 6532628480, "step": 3115 }, { "epoch": 0.1941958231624627, "grad_norm": 17.17679214477539, "learning_rate": 1.5775708968589155e-07, "loss": 0.4017, "num_input_tokens_seen": 6534725632, "step": 3116 }, { "epoch": 0.19446704637917006, "grad_norm": 16.183635711669922, "learning_rate": 1.575593888834303e-07, "loss": 0.4437, "num_input_tokens_seen": 6536822784, "step": 3117 }, { "epoch": 0.1947382695958774, "grad_norm": 25.411428451538086, "learning_rate": 1.5736200390223942e-07, "loss": 0.8358, "num_input_tokens_seen": 6538919936, "step": 3118 }, { "epoch": 0.19500949281258476, "grad_norm": 28.588600158691406, "learning_rate": 1.5716493490116684e-07, "loss": 0.3588, "num_input_tokens_seen": 6541017088, "step": 3119 }, { "epoch": 0.19528071602929212, "grad_norm": 15.308911323547363, "learning_rate": 1.5696818203880624e-07, "loss": 0.4436, "num_input_tokens_seen": 6543114240, "step": 3120 }, { "epoch": 0.19555193924599945, "grad_norm": 14.383072853088379, "learning_rate": 1.5677174547349655e-07, "loss": 0.5685, "num_input_tokens_seen": 6545211392, "step": 3121 }, { "epoch": 0.1958231624627068, "grad_norm": 21.179271697998047, "learning_rate": 1.565756253633228e-07, "loss": 0.5654, "num_input_tokens_seen": 6547308544, "step": 3122 }, { "epoch": 0.19609438567941415, "grad_norm": 17.321855545043945, "learning_rate": 1.5637982186611481e-07, "loss": 0.4894, "num_input_tokens_seen": 6549405696, "step": 3123 }, { "epoch": 0.1963656088961215, "grad_norm": 13.507481575012207, "learning_rate": 1.5618433513944787e-07, "loss": 0.3423, "num_input_tokens_seen": 6551502848, "step": 3124 }, { "epoch": 0.19663683211282887, "grad_norm": 18.613161087036133, "learning_rate": 1.5598916534064216e-07, "loss": 0.5254, "num_input_tokens_seen": 6553600000, "step": 3125 }, { "epoch": 0.1969080553295362, "grad_norm": 13.173487663269043, "learning_rate": 1.557943126267631e-07, "loss": 0.4205, "num_input_tokens_seen": 6555697152, "step": 3126 }, { "epoch": 0.19717927854624356, "grad_norm": 13.874591827392578, "learning_rate": 1.5559977715462044e-07, "loss": 0.3249, "num_input_tokens_seen": 6557794304, "step": 3127 }, { "epoch": 0.1974505017629509, "grad_norm": 16.8161678314209, "learning_rate": 1.5540555908076914e-07, "loss": 0.6233, "num_input_tokens_seen": 6559891456, "step": 3128 }, { "epoch": 0.19772172497965826, "grad_norm": 16.189739227294922, "learning_rate": 1.5521165856150841e-07, "loss": 0.4502, "num_input_tokens_seen": 6561988608, "step": 3129 }, { "epoch": 0.19799294819636562, "grad_norm": 15.414895057678223, "learning_rate": 1.55018075752882e-07, "loss": 0.5046, "num_input_tokens_seen": 6564085760, "step": 3130 }, { "epoch": 0.19826417141307295, "grad_norm": 13.006385803222656, "learning_rate": 1.5482481081067825e-07, "loss": 0.3974, "num_input_tokens_seen": 6566182912, "step": 3131 }, { "epoch": 0.19853539462978032, "grad_norm": 15.883471488952637, "learning_rate": 1.546318638904291e-07, "loss": 0.4731, "num_input_tokens_seen": 6568280064, "step": 3132 }, { "epoch": 0.19880661784648765, "grad_norm": 11.235620498657227, "learning_rate": 1.5443923514741103e-07, "loss": 0.2018, "num_input_tokens_seen": 6570377216, "step": 3133 }, { "epoch": 0.199077841063195, "grad_norm": 15.52456283569336, "learning_rate": 1.5424692473664429e-07, "loss": 0.3148, "num_input_tokens_seen": 6572474368, "step": 3134 }, { "epoch": 0.19934906427990237, "grad_norm": 19.32282066345215, "learning_rate": 1.5405493281289302e-07, "loss": 0.6533, "num_input_tokens_seen": 6574571520, "step": 3135 }, { "epoch": 0.1996202874966097, "grad_norm": 18.52271270751953, "learning_rate": 1.5386325953066494e-07, "loss": 0.6963, "num_input_tokens_seen": 6576668672, "step": 3136 }, { "epoch": 0.19989151071331707, "grad_norm": 12.21414852142334, "learning_rate": 1.5367190504421157e-07, "loss": 0.3308, "num_input_tokens_seen": 6578765824, "step": 3137 }, { "epoch": 0.2001627339300244, "grad_norm": 18.998111724853516, "learning_rate": 1.5348086950752753e-07, "loss": 0.3845, "num_input_tokens_seen": 6580862976, "step": 3138 }, { "epoch": 0.20043395714673176, "grad_norm": 16.331748962402344, "learning_rate": 1.532901530743511e-07, "loss": 0.452, "num_input_tokens_seen": 6582960128, "step": 3139 }, { "epoch": 0.20070518036343912, "grad_norm": 13.402908325195312, "learning_rate": 1.530997558981635e-07, "loss": 0.4053, "num_input_tokens_seen": 6585057280, "step": 3140 }, { "epoch": 0.20097640358014646, "grad_norm": 17.369251251220703, "learning_rate": 1.5290967813218919e-07, "loss": 0.7292, "num_input_tokens_seen": 6587154432, "step": 3141 }, { "epoch": 0.20124762679685382, "grad_norm": 14.586211204528809, "learning_rate": 1.527199199293955e-07, "loss": 0.3989, "num_input_tokens_seen": 6589251584, "step": 3142 }, { "epoch": 0.20151885001356115, "grad_norm": 26.158672332763672, "learning_rate": 1.525304814424927e-07, "loss": 0.8601, "num_input_tokens_seen": 6591348736, "step": 3143 }, { "epoch": 0.2017900732302685, "grad_norm": 12.1050386428833, "learning_rate": 1.5234136282393356e-07, "loss": 0.3601, "num_input_tokens_seen": 6593445888, "step": 3144 }, { "epoch": 0.20206129644697587, "grad_norm": 19.57971954345703, "learning_rate": 1.5215256422591357e-07, "loss": 0.5809, "num_input_tokens_seen": 6595543040, "step": 3145 }, { "epoch": 0.2023325196636832, "grad_norm": 23.03449821472168, "learning_rate": 1.5196408580037058e-07, "loss": 0.2546, "num_input_tokens_seen": 6597640192, "step": 3146 }, { "epoch": 0.20260374288039057, "grad_norm": 16.772308349609375, "learning_rate": 1.51775927698985e-07, "loss": 0.2833, "num_input_tokens_seen": 6599737344, "step": 3147 }, { "epoch": 0.2028749660970979, "grad_norm": 12.067276954650879, "learning_rate": 1.5158809007317926e-07, "loss": 0.4126, "num_input_tokens_seen": 6601834496, "step": 3148 }, { "epoch": 0.20314618931380526, "grad_norm": 23.149063110351562, "learning_rate": 1.514005730741181e-07, "loss": 0.7943, "num_input_tokens_seen": 6603931648, "step": 3149 }, { "epoch": 0.20341741253051263, "grad_norm": 15.860441207885742, "learning_rate": 1.512133768527077e-07, "loss": 0.4636, "num_input_tokens_seen": 6606028800, "step": 3150 }, { "epoch": 0.20368863574721996, "grad_norm": 16.582843780517578, "learning_rate": 1.510265015595967e-07, "loss": 0.3957, "num_input_tokens_seen": 6608125952, "step": 3151 }, { "epoch": 0.20395985896392732, "grad_norm": 20.899213790893555, "learning_rate": 1.5083994734517518e-07, "loss": 0.6574, "num_input_tokens_seen": 6610223104, "step": 3152 }, { "epoch": 0.20423108218063465, "grad_norm": 12.626710891723633, "learning_rate": 1.5065371435957484e-07, "loss": 0.4247, "num_input_tokens_seen": 6612320256, "step": 3153 }, { "epoch": 0.20450230539734202, "grad_norm": 28.32889175415039, "learning_rate": 1.5046780275266897e-07, "loss": 0.7678, "num_input_tokens_seen": 6614417408, "step": 3154 }, { "epoch": 0.20477352861404935, "grad_norm": 16.073993682861328, "learning_rate": 1.5028221267407207e-07, "loss": 0.4877, "num_input_tokens_seen": 6616514560, "step": 3155 }, { "epoch": 0.2050447518307567, "grad_norm": 9.331436157226562, "learning_rate": 1.5009694427314007e-07, "loss": 0.2676, "num_input_tokens_seen": 6618611712, "step": 3156 }, { "epoch": 0.20531597504746407, "grad_norm": 13.70032787322998, "learning_rate": 1.4991199769896983e-07, "loss": 0.3683, "num_input_tokens_seen": 6620708864, "step": 3157 }, { "epoch": 0.2055871982641714, "grad_norm": 12.181900024414062, "learning_rate": 1.497273731003994e-07, "loss": 0.3804, "num_input_tokens_seen": 6622806016, "step": 3158 }, { "epoch": 0.20585842148087877, "grad_norm": 11.273839950561523, "learning_rate": 1.4954307062600758e-07, "loss": 0.3363, "num_input_tokens_seen": 6624903168, "step": 3159 }, { "epoch": 0.2061296446975861, "grad_norm": 18.64375114440918, "learning_rate": 1.4935909042411412e-07, "loss": 0.7101, "num_input_tokens_seen": 6627000320, "step": 3160 }, { "epoch": 0.20640086791429346, "grad_norm": 17.58132553100586, "learning_rate": 1.4917543264277901e-07, "loss": 0.5021, "num_input_tokens_seen": 6629097472, "step": 3161 }, { "epoch": 0.20667209113100082, "grad_norm": 11.847007751464844, "learning_rate": 1.4899209742980317e-07, "loss": 0.4044, "num_input_tokens_seen": 6631194624, "step": 3162 }, { "epoch": 0.20694331434770816, "grad_norm": 17.20060157775879, "learning_rate": 1.488090849327279e-07, "loss": 0.5288, "num_input_tokens_seen": 6633291776, "step": 3163 }, { "epoch": 0.20721453756441552, "grad_norm": 19.338298797607422, "learning_rate": 1.4862639529883463e-07, "loss": 0.3903, "num_input_tokens_seen": 6635388928, "step": 3164 }, { "epoch": 0.20748576078112285, "grad_norm": 13.957944869995117, "learning_rate": 1.4844402867514503e-07, "loss": 0.4285, "num_input_tokens_seen": 6637486080, "step": 3165 }, { "epoch": 0.2077569839978302, "grad_norm": 10.898870468139648, "learning_rate": 1.4826198520842093e-07, "loss": 0.3047, "num_input_tokens_seen": 6639583232, "step": 3166 }, { "epoch": 0.20802820721453757, "grad_norm": 12.25651741027832, "learning_rate": 1.4808026504516374e-07, "loss": 0.435, "num_input_tokens_seen": 6641680384, "step": 3167 }, { "epoch": 0.2082994304312449, "grad_norm": 17.442764282226562, "learning_rate": 1.4789886833161506e-07, "loss": 0.582, "num_input_tokens_seen": 6643777536, "step": 3168 }, { "epoch": 0.20857065364795227, "grad_norm": 10.577066421508789, "learning_rate": 1.477177952137561e-07, "loss": 0.3007, "num_input_tokens_seen": 6645874688, "step": 3169 }, { "epoch": 0.2088418768646596, "grad_norm": 18.119701385498047, "learning_rate": 1.4753704583730754e-07, "loss": 0.5214, "num_input_tokens_seen": 6647971840, "step": 3170 }, { "epoch": 0.20911310008136696, "grad_norm": 13.43950366973877, "learning_rate": 1.4735662034772962e-07, "loss": 0.4102, "num_input_tokens_seen": 6650068992, "step": 3171 }, { "epoch": 0.20938432329807433, "grad_norm": 16.050125122070312, "learning_rate": 1.47176518890222e-07, "loss": 0.3653, "num_input_tokens_seen": 6652166144, "step": 3172 }, { "epoch": 0.20965554651478166, "grad_norm": 10.848344802856445, "learning_rate": 1.4699674160972337e-07, "loss": 0.3367, "num_input_tokens_seen": 6654263296, "step": 3173 }, { "epoch": 0.20992676973148902, "grad_norm": 11.283642768859863, "learning_rate": 1.4681728865091165e-07, "loss": 0.2537, "num_input_tokens_seen": 6656360448, "step": 3174 }, { "epoch": 0.21019799294819635, "grad_norm": 19.65658187866211, "learning_rate": 1.466381601582038e-07, "loss": 0.521, "num_input_tokens_seen": 6658457600, "step": 3175 }, { "epoch": 0.21046921616490372, "grad_norm": 12.568547248840332, "learning_rate": 1.4645935627575562e-07, "loss": 0.2631, "num_input_tokens_seen": 6660554752, "step": 3176 }, { "epoch": 0.21074043938161108, "grad_norm": 17.255279541015625, "learning_rate": 1.462808771474617e-07, "loss": 0.6249, "num_input_tokens_seen": 6662651904, "step": 3177 }, { "epoch": 0.2110116625983184, "grad_norm": 10.227863311767578, "learning_rate": 1.4610272291695503e-07, "loss": 0.2266, "num_input_tokens_seen": 6664749056, "step": 3178 }, { "epoch": 0.21128288581502577, "grad_norm": 17.062204360961914, "learning_rate": 1.4592489372760764e-07, "loss": 0.4047, "num_input_tokens_seen": 6666846208, "step": 3179 }, { "epoch": 0.2115541090317331, "grad_norm": 13.168872833251953, "learning_rate": 1.4574738972252953e-07, "loss": 0.3819, "num_input_tokens_seen": 6668943360, "step": 3180 }, { "epoch": 0.21182533224844047, "grad_norm": 14.085567474365234, "learning_rate": 1.4557021104456923e-07, "loss": 0.4032, "num_input_tokens_seen": 6671040512, "step": 3181 }, { "epoch": 0.21209655546514783, "grad_norm": 14.73873519897461, "learning_rate": 1.4539335783631346e-07, "loss": 0.415, "num_input_tokens_seen": 6673137664, "step": 3182 }, { "epoch": 0.21236777868185516, "grad_norm": 12.23835277557373, "learning_rate": 1.4521683024008687e-07, "loss": 0.4222, "num_input_tokens_seen": 6675234816, "step": 3183 }, { "epoch": 0.21263900189856252, "grad_norm": 15.95200252532959, "learning_rate": 1.450406283979521e-07, "loss": 0.4986, "num_input_tokens_seen": 6677331968, "step": 3184 }, { "epoch": 0.21291022511526986, "grad_norm": 10.582295417785645, "learning_rate": 1.4486475245170977e-07, "loss": 0.2614, "num_input_tokens_seen": 6679429120, "step": 3185 }, { "epoch": 0.21318144833197722, "grad_norm": 20.6806583404541, "learning_rate": 1.446892025428981e-07, "loss": 0.4077, "num_input_tokens_seen": 6681526272, "step": 3186 }, { "epoch": 0.21345267154868458, "grad_norm": 10.347526550292969, "learning_rate": 1.4451397881279298e-07, "loss": 0.1838, "num_input_tokens_seen": 6683623424, "step": 3187 }, { "epoch": 0.21372389476539191, "grad_norm": 15.546320915222168, "learning_rate": 1.4433908140240792e-07, "loss": 0.4314, "num_input_tokens_seen": 6685720576, "step": 3188 }, { "epoch": 0.21399511798209928, "grad_norm": 14.439404487609863, "learning_rate": 1.4416451045249354e-07, "loss": 0.3794, "num_input_tokens_seen": 6687817728, "step": 3189 }, { "epoch": 0.2142663411988066, "grad_norm": 14.810487747192383, "learning_rate": 1.43990266103538e-07, "loss": 0.3653, "num_input_tokens_seen": 6689914880, "step": 3190 }, { "epoch": 0.21453756441551397, "grad_norm": 13.329632759094238, "learning_rate": 1.4381634849576644e-07, "loss": 0.3458, "num_input_tokens_seen": 6692012032, "step": 3191 }, { "epoch": 0.21480878763222133, "grad_norm": 11.149251937866211, "learning_rate": 1.4364275776914112e-07, "loss": 0.3134, "num_input_tokens_seen": 6694109184, "step": 3192 }, { "epoch": 0.21508001084892867, "grad_norm": 11.901870727539062, "learning_rate": 1.4346949406336136e-07, "loss": 0.3335, "num_input_tokens_seen": 6696206336, "step": 3193 }, { "epoch": 0.21535123406563603, "grad_norm": 10.92190933227539, "learning_rate": 1.4329655751786318e-07, "loss": 0.3104, "num_input_tokens_seen": 6698303488, "step": 3194 }, { "epoch": 0.21562245728234336, "grad_norm": 13.854144096374512, "learning_rate": 1.4312394827181925e-07, "loss": 0.4487, "num_input_tokens_seen": 6700400640, "step": 3195 }, { "epoch": 0.21589368049905072, "grad_norm": 13.08950138092041, "learning_rate": 1.4295166646413898e-07, "loss": 0.3734, "num_input_tokens_seen": 6702497792, "step": 3196 }, { "epoch": 0.21616490371575806, "grad_norm": 13.053932189941406, "learning_rate": 1.4277971223346825e-07, "loss": 0.3552, "num_input_tokens_seen": 6704594944, "step": 3197 }, { "epoch": 0.21643612693246542, "grad_norm": 17.404874801635742, "learning_rate": 1.4260808571818917e-07, "loss": 0.3981, "num_input_tokens_seen": 6706692096, "step": 3198 }, { "epoch": 0.21670735014917278, "grad_norm": 16.526578903198242, "learning_rate": 1.4243678705642027e-07, "loss": 0.3884, "num_input_tokens_seen": 6708789248, "step": 3199 }, { "epoch": 0.2169785733658801, "grad_norm": 16.661752700805664, "learning_rate": 1.4226581638601635e-07, "loss": 0.4724, "num_input_tokens_seen": 6710886400, "step": 3200 }, { "epoch": 0.00027122321670735016, "grad_norm": 15.5323486328125, "learning_rate": 1.4209517384456786e-07, "loss": 0.505, "num_input_tokens_seen": 6712983552, "step": 3201 }, { "epoch": 0.0005424464334147003, "grad_norm": 9.104025840759277, "learning_rate": 1.4192485956940153e-07, "loss": 0.2839, "num_input_tokens_seen": 6715080704, "step": 3202 }, { "epoch": 0.0008136696501220504, "grad_norm": 9.672213554382324, "learning_rate": 1.4175487369757973e-07, "loss": 0.2322, "num_input_tokens_seen": 6717177856, "step": 3203 }, { "epoch": 0.0010848928668294006, "grad_norm": 15.418540954589844, "learning_rate": 1.4158521636590079e-07, "loss": 0.5406, "num_input_tokens_seen": 6719275008, "step": 3204 }, { "epoch": 0.0013561160835367507, "grad_norm": 19.670473098754883, "learning_rate": 1.4141588771089838e-07, "loss": 0.7279, "num_input_tokens_seen": 6721372160, "step": 3205 }, { "epoch": 0.0016273393002441008, "grad_norm": 9.864188194274902, "learning_rate": 1.412468878688418e-07, "loss": 0.2668, "num_input_tokens_seen": 6723469312, "step": 3206 }, { "epoch": 0.001898562516951451, "grad_norm": 9.84155559539795, "learning_rate": 1.410782169757356e-07, "loss": 0.2388, "num_input_tokens_seen": 6725566464, "step": 3207 }, { "epoch": 0.0021697857336588013, "grad_norm": 12.789325714111328, "learning_rate": 1.4090987516731977e-07, "loss": 0.4207, "num_input_tokens_seen": 6727663616, "step": 3208 }, { "epoch": 0.0024410089503661514, "grad_norm": 13.97022819519043, "learning_rate": 1.4074186257906934e-07, "loss": 0.3951, "num_input_tokens_seen": 6729760768, "step": 3209 }, { "epoch": 0.0027122321670735015, "grad_norm": 20.8309268951416, "learning_rate": 1.4057417934619453e-07, "loss": 0.5764, "num_input_tokens_seen": 6731857920, "step": 3210 }, { "epoch": 0.0029834553837808516, "grad_norm": 17.302791595458984, "learning_rate": 1.404068256036403e-07, "loss": 0.5127, "num_input_tokens_seen": 6733955072, "step": 3211 }, { "epoch": 0.0032546786004882017, "grad_norm": 18.93463134765625, "learning_rate": 1.4023980148608667e-07, "loss": 0.4078, "num_input_tokens_seen": 6736052224, "step": 3212 }, { "epoch": 0.003525901817195552, "grad_norm": 9.369474411010742, "learning_rate": 1.4007310712794827e-07, "loss": 0.1545, "num_input_tokens_seen": 6738149376, "step": 3213 }, { "epoch": 0.003797125033902902, "grad_norm": 17.984560012817383, "learning_rate": 1.3990674266337442e-07, "loss": 0.5359, "num_input_tokens_seen": 6740246528, "step": 3214 }, { "epoch": 0.0040683482506102524, "grad_norm": 11.723960876464844, "learning_rate": 1.3974070822624883e-07, "loss": 0.3113, "num_input_tokens_seen": 6742343680, "step": 3215 }, { "epoch": 0.0043395714673176026, "grad_norm": 15.782196998596191, "learning_rate": 1.3957500395018977e-07, "loss": 0.5187, "num_input_tokens_seen": 6744440832, "step": 3216 }, { "epoch": 0.004610794684024953, "grad_norm": 17.8033390045166, "learning_rate": 1.394096299685497e-07, "loss": 0.5859, "num_input_tokens_seen": 6746537984, "step": 3217 }, { "epoch": 0.004882017900732303, "grad_norm": 12.155229568481445, "learning_rate": 1.3924458641441532e-07, "loss": 0.3767, "num_input_tokens_seen": 6748635136, "step": 3218 }, { "epoch": 0.005153241117439653, "grad_norm": 14.897374153137207, "learning_rate": 1.3907987342060725e-07, "loss": 0.5853, "num_input_tokens_seen": 6750732288, "step": 3219 }, { "epoch": 0.005424464334147003, "grad_norm": 18.14958381652832, "learning_rate": 1.389154911196805e-07, "loss": 0.5894, "num_input_tokens_seen": 6752829440, "step": 3220 }, { "epoch": 0.005695687550854353, "grad_norm": 12.628942489624023, "learning_rate": 1.3875143964392355e-07, "loss": 0.2947, "num_input_tokens_seen": 6754926592, "step": 3221 }, { "epoch": 0.005966910767561703, "grad_norm": 22.452190399169922, "learning_rate": 1.3858771912535877e-07, "loss": 0.8263, "num_input_tokens_seen": 6757023744, "step": 3222 }, { "epoch": 0.006238133984269053, "grad_norm": 9.461743354797363, "learning_rate": 1.3842432969574238e-07, "loss": 0.1619, "num_input_tokens_seen": 6759120896, "step": 3223 }, { "epoch": 0.006509357200976403, "grad_norm": 12.965723037719727, "learning_rate": 1.382612714865637e-07, "loss": 0.3498, "num_input_tokens_seen": 6761218048, "step": 3224 }, { "epoch": 0.0067805804176837535, "grad_norm": 20.17690086364746, "learning_rate": 1.3809854462904596e-07, "loss": 0.5391, "num_input_tokens_seen": 6763315200, "step": 3225 }, { "epoch": 0.007051803634391104, "grad_norm": 10.085182189941406, "learning_rate": 1.3793614925414542e-07, "loss": 0.1799, "num_input_tokens_seen": 6765412352, "step": 3226 }, { "epoch": 0.007323026851098454, "grad_norm": 11.336246490478516, "learning_rate": 1.3777408549255182e-07, "loss": 0.3041, "num_input_tokens_seen": 6767509504, "step": 3227 }, { "epoch": 0.007594250067805804, "grad_norm": 7.451137542724609, "learning_rate": 1.3761235347468784e-07, "loss": 0.1651, "num_input_tokens_seen": 6769606656, "step": 3228 }, { "epoch": 0.007865473284513154, "grad_norm": 11.728433609008789, "learning_rate": 1.3745095333070926e-07, "loss": 0.2284, "num_input_tokens_seen": 6771703808, "step": 3229 }, { "epoch": 0.008136696501220505, "grad_norm": 21.89021873474121, "learning_rate": 1.3728988519050476e-07, "loss": 0.9454, "num_input_tokens_seen": 6773800960, "step": 3230 }, { "epoch": 0.008407919717927854, "grad_norm": 13.923346519470215, "learning_rate": 1.3712914918369595e-07, "loss": 0.4593, "num_input_tokens_seen": 6775898112, "step": 3231 }, { "epoch": 0.008679142934635205, "grad_norm": 17.16434097290039, "learning_rate": 1.369687454396369e-07, "loss": 0.5116, "num_input_tokens_seen": 6777995264, "step": 3232 }, { "epoch": 0.008950366151342554, "grad_norm": 13.330866813659668, "learning_rate": 1.3680867408741457e-07, "loss": 0.3037, "num_input_tokens_seen": 6780092416, "step": 3233 }, { "epoch": 0.009221589368049905, "grad_norm": 28.39195442199707, "learning_rate": 1.366489352558483e-07, "loss": 0.6069, "num_input_tokens_seen": 6782189568, "step": 3234 }, { "epoch": 0.009492812584757255, "grad_norm": 23.18076515197754, "learning_rate": 1.3648952907348957e-07, "loss": 0.4345, "num_input_tokens_seen": 6784286720, "step": 3235 }, { "epoch": 0.009764035801464606, "grad_norm": 13.762514114379883, "learning_rate": 1.3633045566862268e-07, "loss": 0.4396, "num_input_tokens_seen": 6786383872, "step": 3236 }, { "epoch": 0.010035259018171955, "grad_norm": 13.543777465820312, "learning_rate": 1.3617171516926382e-07, "loss": 0.37, "num_input_tokens_seen": 6788481024, "step": 3237 }, { "epoch": 0.010306482234879306, "grad_norm": 12.794722557067871, "learning_rate": 1.3601330770316122e-07, "loss": 0.2349, "num_input_tokens_seen": 6790578176, "step": 3238 }, { "epoch": 0.010577705451586655, "grad_norm": 17.2214298248291, "learning_rate": 1.3585523339779523e-07, "loss": 0.3966, "num_input_tokens_seen": 6792675328, "step": 3239 }, { "epoch": 0.010848928668294006, "grad_norm": 21.573457717895508, "learning_rate": 1.356974923803781e-07, "loss": 0.7812, "num_input_tokens_seen": 6794772480, "step": 3240 }, { "epoch": 0.011120151885001357, "grad_norm": 15.258363723754883, "learning_rate": 1.3554008477785367e-07, "loss": 0.4462, "num_input_tokens_seen": 6796869632, "step": 3241 }, { "epoch": 0.011391375101708706, "grad_norm": 15.593113899230957, "learning_rate": 1.3538301071689756e-07, "loss": 0.3842, "num_input_tokens_seen": 6798966784, "step": 3242 }, { "epoch": 0.011662598318416057, "grad_norm": 15.24464225769043, "learning_rate": 1.3522627032391715e-07, "loss": 0.4702, "num_input_tokens_seen": 6801063936, "step": 3243 }, { "epoch": 0.011933821535123406, "grad_norm": 11.781341552734375, "learning_rate": 1.3506986372505098e-07, "loss": 0.3587, "num_input_tokens_seen": 6803161088, "step": 3244 }, { "epoch": 0.012205044751830757, "grad_norm": 19.120126724243164, "learning_rate": 1.3491379104616938e-07, "loss": 0.5466, "num_input_tokens_seen": 6805258240, "step": 3245 }, { "epoch": 0.012476267968538107, "grad_norm": 10.356958389282227, "learning_rate": 1.3475805241287345e-07, "loss": 0.2264, "num_input_tokens_seen": 6807355392, "step": 3246 }, { "epoch": 0.012747491185245458, "grad_norm": 17.21544075012207, "learning_rate": 1.3460264795049577e-07, "loss": 0.6428, "num_input_tokens_seen": 6809452544, "step": 3247 }, { "epoch": 0.013018714401952807, "grad_norm": 16.766895294189453, "learning_rate": 1.344475777841e-07, "loss": 0.3485, "num_input_tokens_seen": 6811549696, "step": 3248 }, { "epoch": 0.013289937618660158, "grad_norm": 11.51309585571289, "learning_rate": 1.3429284203848074e-07, "loss": 0.2704, "num_input_tokens_seen": 6813646848, "step": 3249 }, { "epoch": 0.013561160835367507, "grad_norm": 26.110618591308594, "learning_rate": 1.3413844083816334e-07, "loss": 1.218, "num_input_tokens_seen": 6815744000, "step": 3250 }, { "epoch": 0.013832384052074858, "grad_norm": 14.552084922790527, "learning_rate": 1.3398437430740403e-07, "loss": 0.3287, "num_input_tokens_seen": 6817841152, "step": 3251 }, { "epoch": 0.014103607268782207, "grad_norm": 11.300999641418457, "learning_rate": 1.3383064257018978e-07, "loss": 0.2987, "num_input_tokens_seen": 6819938304, "step": 3252 }, { "epoch": 0.014374830485489558, "grad_norm": 12.652029991149902, "learning_rate": 1.3367724575023786e-07, "loss": 0.3022, "num_input_tokens_seen": 6822035456, "step": 3253 }, { "epoch": 0.014646053702196907, "grad_norm": 18.908370971679688, "learning_rate": 1.3352418397099638e-07, "loss": 0.6609, "num_input_tokens_seen": 6824132608, "step": 3254 }, { "epoch": 0.014917276918904258, "grad_norm": 16.611713409423828, "learning_rate": 1.333714573556435e-07, "loss": 0.4249, "num_input_tokens_seen": 6826229760, "step": 3255 }, { "epoch": 0.015188500135611608, "grad_norm": 15.014352798461914, "learning_rate": 1.332190660270879e-07, "loss": 0.4945, "num_input_tokens_seen": 6828326912, "step": 3256 }, { "epoch": 0.015459723352318959, "grad_norm": 14.692705154418945, "learning_rate": 1.3306701010796832e-07, "loss": 0.4748, "num_input_tokens_seen": 6830424064, "step": 3257 }, { "epoch": 0.015730946569026308, "grad_norm": 15.815617561340332, "learning_rate": 1.3291528972065347e-07, "loss": 0.559, "num_input_tokens_seen": 6832521216, "step": 3258 }, { "epoch": 0.01600216978573366, "grad_norm": 13.448586463928223, "learning_rate": 1.327639049872422e-07, "loss": 0.3937, "num_input_tokens_seen": 6834618368, "step": 3259 }, { "epoch": 0.01627339300244101, "grad_norm": 14.515972137451172, "learning_rate": 1.326128560295631e-07, "loss": 0.4, "num_input_tokens_seen": 6836715520, "step": 3260 }, { "epoch": 0.01654461621914836, "grad_norm": 14.561549186706543, "learning_rate": 1.324621429691748e-07, "loss": 0.4807, "num_input_tokens_seen": 6838812672, "step": 3261 }, { "epoch": 0.016815839435855708, "grad_norm": 22.775854110717773, "learning_rate": 1.3231176592736528e-07, "loss": 0.545, "num_input_tokens_seen": 6840909824, "step": 3262 }, { "epoch": 0.01708706265256306, "grad_norm": 17.545738220214844, "learning_rate": 1.3216172502515248e-07, "loss": 0.6031, "num_input_tokens_seen": 6843006976, "step": 3263 }, { "epoch": 0.01735828586927041, "grad_norm": 19.9493350982666, "learning_rate": 1.320120203832833e-07, "loss": 0.5962, "num_input_tokens_seen": 6845104128, "step": 3264 }, { "epoch": 0.01762950908597776, "grad_norm": 16.441688537597656, "learning_rate": 1.3186265212223459e-07, "loss": 0.5388, "num_input_tokens_seen": 6847201280, "step": 3265 }, { "epoch": 0.01790073230268511, "grad_norm": 17.45583724975586, "learning_rate": 1.3171362036221202e-07, "loss": 0.394, "num_input_tokens_seen": 6849298432, "step": 3266 }, { "epoch": 0.01817195551939246, "grad_norm": 21.072965621948242, "learning_rate": 1.3156492522315088e-07, "loss": 0.7281, "num_input_tokens_seen": 6851395584, "step": 3267 }, { "epoch": 0.01844317873609981, "grad_norm": 12.47169303894043, "learning_rate": 1.314165668247153e-07, "loss": 0.4182, "num_input_tokens_seen": 6853492736, "step": 3268 }, { "epoch": 0.01871440195280716, "grad_norm": 17.800216674804688, "learning_rate": 1.3126854528629843e-07, "loss": 0.5971, "num_input_tokens_seen": 6855589888, "step": 3269 }, { "epoch": 0.01898562516951451, "grad_norm": 13.492378234863281, "learning_rate": 1.3112086072702238e-07, "loss": 0.3652, "num_input_tokens_seen": 6857687040, "step": 3270 }, { "epoch": 0.01925684838622186, "grad_norm": 16.468965530395508, "learning_rate": 1.3097351326573814e-07, "loss": 0.4573, "num_input_tokens_seen": 6859784192, "step": 3271 }, { "epoch": 0.01952807160292921, "grad_norm": 19.35951042175293, "learning_rate": 1.3082650302102524e-07, "loss": 0.6076, "num_input_tokens_seen": 6861881344, "step": 3272 }, { "epoch": 0.019799294819636562, "grad_norm": 8.793999671936035, "learning_rate": 1.30679830111192e-07, "loss": 0.184, "num_input_tokens_seen": 6863978496, "step": 3273 }, { "epoch": 0.02007051803634391, "grad_norm": 14.717864990234375, "learning_rate": 1.3053349465427518e-07, "loss": 0.4201, "num_input_tokens_seen": 6866075648, "step": 3274 }, { "epoch": 0.02034174125305126, "grad_norm": 13.672134399414062, "learning_rate": 1.303874967680399e-07, "loss": 0.3217, "num_input_tokens_seen": 6868172800, "step": 3275 }, { "epoch": 0.02061296446975861, "grad_norm": 13.778663635253906, "learning_rate": 1.302418365699798e-07, "loss": 0.2786, "num_input_tokens_seen": 6870269952, "step": 3276 }, { "epoch": 0.020884187686465962, "grad_norm": 24.35689353942871, "learning_rate": 1.300965141773167e-07, "loss": 0.935, "num_input_tokens_seen": 6872367104, "step": 3277 }, { "epoch": 0.02115541090317331, "grad_norm": 12.533529281616211, "learning_rate": 1.2995152970700044e-07, "loss": 0.3299, "num_input_tokens_seen": 6874464256, "step": 3278 }, { "epoch": 0.02142663411988066, "grad_norm": 15.271020889282227, "learning_rate": 1.2980688327570906e-07, "loss": 0.5584, "num_input_tokens_seen": 6876561408, "step": 3279 }, { "epoch": 0.021697857336588012, "grad_norm": 12.875490188598633, "learning_rate": 1.296625749998485e-07, "loss": 0.3324, "num_input_tokens_seen": 6878658560, "step": 3280 }, { "epoch": 0.021969080553295363, "grad_norm": 13.227240562438965, "learning_rate": 1.2951860499555253e-07, "loss": 0.2358, "num_input_tokens_seen": 6880755712, "step": 3281 }, { "epoch": 0.022240303770002714, "grad_norm": 16.106595993041992, "learning_rate": 1.2937497337868274e-07, "loss": 0.5393, "num_input_tokens_seen": 6882852864, "step": 3282 }, { "epoch": 0.02251152698671006, "grad_norm": 18.33268928527832, "learning_rate": 1.2923168026482843e-07, "loss": 0.5663, "num_input_tokens_seen": 6884950016, "step": 3283 }, { "epoch": 0.022782750203417412, "grad_norm": 13.891176223754883, "learning_rate": 1.290887257693064e-07, "loss": 0.3277, "num_input_tokens_seen": 6887047168, "step": 3284 }, { "epoch": 0.023053973420124763, "grad_norm": 22.6708984375, "learning_rate": 1.2894611000716103e-07, "loss": 0.539, "num_input_tokens_seen": 6889144320, "step": 3285 }, { "epoch": 0.023325196636832114, "grad_norm": 17.467336654663086, "learning_rate": 1.28803833093164e-07, "loss": 0.4937, "num_input_tokens_seen": 6891241472, "step": 3286 }, { "epoch": 0.023596419853539462, "grad_norm": 14.677574157714844, "learning_rate": 1.286618951418144e-07, "loss": 0.3119, "num_input_tokens_seen": 6893338624, "step": 3287 }, { "epoch": 0.023867643070246813, "grad_norm": 14.258770942687988, "learning_rate": 1.285202962673385e-07, "loss": 0.4642, "num_input_tokens_seen": 6895435776, "step": 3288 }, { "epoch": 0.024138866286954164, "grad_norm": 19.215715408325195, "learning_rate": 1.2837903658368965e-07, "loss": 0.6045, "num_input_tokens_seen": 6897532928, "step": 3289 }, { "epoch": 0.024410089503661515, "grad_norm": 16.469133377075195, "learning_rate": 1.2823811620454833e-07, "loss": 0.3608, "num_input_tokens_seen": 6899630080, "step": 3290 }, { "epoch": 0.024681312720368862, "grad_norm": 19.01442527770996, "learning_rate": 1.2809753524332188e-07, "loss": 0.5565, "num_input_tokens_seen": 6901727232, "step": 3291 }, { "epoch": 0.024952535937076213, "grad_norm": 18.02033042907715, "learning_rate": 1.2795729381314434e-07, "loss": 0.5867, "num_input_tokens_seen": 6903824384, "step": 3292 }, { "epoch": 0.025223759153783564, "grad_norm": 14.180427551269531, "learning_rate": 1.2781739202687694e-07, "loss": 0.4978, "num_input_tokens_seen": 6905921536, "step": 3293 }, { "epoch": 0.025494982370490915, "grad_norm": 18.889726638793945, "learning_rate": 1.2767782999710715e-07, "loss": 0.8013, "num_input_tokens_seen": 6908018688, "step": 3294 }, { "epoch": 0.025766205587198263, "grad_norm": 17.951160430908203, "learning_rate": 1.2753860783614925e-07, "loss": 0.6829, "num_input_tokens_seen": 6910115840, "step": 3295 }, { "epoch": 0.026037428803905614, "grad_norm": 13.510412216186523, "learning_rate": 1.2739972565604396e-07, "loss": 0.3965, "num_input_tokens_seen": 6912212992, "step": 3296 }, { "epoch": 0.026308652020612965, "grad_norm": 15.73293399810791, "learning_rate": 1.2726118356855834e-07, "loss": 0.6381, "num_input_tokens_seen": 6914310144, "step": 3297 }, { "epoch": 0.026579875237320316, "grad_norm": 16.247882843017578, "learning_rate": 1.271229816851857e-07, "loss": 0.3781, "num_input_tokens_seen": 6916407296, "step": 3298 }, { "epoch": 0.026851098454027666, "grad_norm": 15.698039054870605, "learning_rate": 1.2698512011714578e-07, "loss": 0.4067, "num_input_tokens_seen": 6918504448, "step": 3299 }, { "epoch": 0.027122321670735014, "grad_norm": 18.372669219970703, "learning_rate": 1.2684759897538422e-07, "loss": 0.4066, "num_input_tokens_seen": 6920601600, "step": 3300 }, { "epoch": 0.027393544887442365, "grad_norm": 17.756454467773438, "learning_rate": 1.2671041837057287e-07, "loss": 0.5158, "num_input_tokens_seen": 6922698752, "step": 3301 }, { "epoch": 0.027664768104149716, "grad_norm": 18.298004150390625, "learning_rate": 1.2657357841310941e-07, "loss": 0.449, "num_input_tokens_seen": 6924795904, "step": 3302 }, { "epoch": 0.027935991320857067, "grad_norm": 14.342238426208496, "learning_rate": 1.2643707921311765e-07, "loss": 0.341, "num_input_tokens_seen": 6926893056, "step": 3303 }, { "epoch": 0.028207214537564414, "grad_norm": 10.686249732971191, "learning_rate": 1.2630092088044664e-07, "loss": 0.2287, "num_input_tokens_seen": 6928990208, "step": 3304 }, { "epoch": 0.028478437754271765, "grad_norm": 13.416088104248047, "learning_rate": 1.2616510352467158e-07, "loss": 0.3559, "num_input_tokens_seen": 6931087360, "step": 3305 }, { "epoch": 0.028749660970979116, "grad_norm": 21.23432731628418, "learning_rate": 1.2602962725509307e-07, "loss": 0.7772, "num_input_tokens_seen": 6933184512, "step": 3306 }, { "epoch": 0.029020884187686467, "grad_norm": 16.567541122436523, "learning_rate": 1.2589449218073728e-07, "loss": 0.4244, "num_input_tokens_seen": 6935281664, "step": 3307 }, { "epoch": 0.029292107404393815, "grad_norm": 10.55752944946289, "learning_rate": 1.2575969841035578e-07, "loss": 0.3227, "num_input_tokens_seen": 6937378816, "step": 3308 }, { "epoch": 0.029563330621101166, "grad_norm": 17.855588912963867, "learning_rate": 1.256252460524255e-07, "loss": 0.63, "num_input_tokens_seen": 6939475968, "step": 3309 }, { "epoch": 0.029834553837808517, "grad_norm": 20.68912124633789, "learning_rate": 1.2549113521514844e-07, "loss": 0.5901, "num_input_tokens_seen": 6941573120, "step": 3310 }, { "epoch": 0.030105777054515868, "grad_norm": 11.569897651672363, "learning_rate": 1.2535736600645212e-07, "loss": 0.3302, "num_input_tokens_seen": 6943670272, "step": 3311 }, { "epoch": 0.030377000271223215, "grad_norm": 16.52699851989746, "learning_rate": 1.2522393853398868e-07, "loss": 0.4751, "num_input_tokens_seen": 6945767424, "step": 3312 }, { "epoch": 0.030648223487930566, "grad_norm": 15.689116477966309, "learning_rate": 1.2509085290513563e-07, "loss": 0.2153, "num_input_tokens_seen": 6947864576, "step": 3313 }, { "epoch": 0.030919446704637917, "grad_norm": 14.995532035827637, "learning_rate": 1.2495810922699517e-07, "loss": 0.4613, "num_input_tokens_seen": 6949961728, "step": 3314 }, { "epoch": 0.031190669921345268, "grad_norm": 12.99809455871582, "learning_rate": 1.2482570760639439e-07, "loss": 0.3528, "num_input_tokens_seen": 6952058880, "step": 3315 }, { "epoch": 0.031461893138052616, "grad_norm": 11.95545482635498, "learning_rate": 1.24693648149885e-07, "loss": 0.2344, "num_input_tokens_seen": 6954156032, "step": 3316 }, { "epoch": 0.03173311635475997, "grad_norm": 9.108811378479004, "learning_rate": 1.2456193096374338e-07, "loss": 0.2519, "num_input_tokens_seen": 6956253184, "step": 3317 }, { "epoch": 0.03200433957146732, "grad_norm": 31.258995056152344, "learning_rate": 1.244305561539707e-07, "loss": 0.6474, "num_input_tokens_seen": 6958350336, "step": 3318 }, { "epoch": 0.032275562788174665, "grad_norm": 16.379648208618164, "learning_rate": 1.242995238262923e-07, "loss": 0.496, "num_input_tokens_seen": 6960447488, "step": 3319 }, { "epoch": 0.03254678600488202, "grad_norm": 15.703734397888184, "learning_rate": 1.2416883408615809e-07, "loss": 0.4296, "num_input_tokens_seen": 6962544640, "step": 3320 }, { "epoch": 0.03281800922158937, "grad_norm": 14.436956405639648, "learning_rate": 1.2403848703874202e-07, "loss": 0.3409, "num_input_tokens_seen": 6964641792, "step": 3321 }, { "epoch": 0.03308923243829672, "grad_norm": 13.694087028503418, "learning_rate": 1.2390848278894264e-07, "loss": 0.4741, "num_input_tokens_seen": 6966738944, "step": 3322 }, { "epoch": 0.03336045565500407, "grad_norm": 15.317477226257324, "learning_rate": 1.2377882144138222e-07, "loss": 0.4036, "num_input_tokens_seen": 6968836096, "step": 3323 }, { "epoch": 0.033631678871711417, "grad_norm": 14.88959789276123, "learning_rate": 1.2364950310040744e-07, "loss": 0.5187, "num_input_tokens_seen": 6970933248, "step": 3324 }, { "epoch": 0.03390290208841877, "grad_norm": 10.9374361038208, "learning_rate": 1.235205278700887e-07, "loss": 0.2573, "num_input_tokens_seen": 6973030400, "step": 3325 }, { "epoch": 0.03417412530512612, "grad_norm": 20.69586181640625, "learning_rate": 1.233918958542204e-07, "loss": 0.7775, "num_input_tokens_seen": 6975127552, "step": 3326 }, { "epoch": 0.034445348521833466, "grad_norm": 16.079696655273438, "learning_rate": 1.2326360715632069e-07, "loss": 0.4014, "num_input_tokens_seen": 6977224704, "step": 3327 }, { "epoch": 0.03471657173854082, "grad_norm": 21.355201721191406, "learning_rate": 1.231356618796314e-07, "loss": 0.6694, "num_input_tokens_seen": 6979321856, "step": 3328 }, { "epoch": 0.03498779495524817, "grad_norm": 18.189098358154297, "learning_rate": 1.2300806012711798e-07, "loss": 0.6481, "num_input_tokens_seen": 6981419008, "step": 3329 }, { "epoch": 0.03525901817195552, "grad_norm": 8.602594375610352, "learning_rate": 1.228808020014696e-07, "loss": 0.1969, "num_input_tokens_seen": 6983516160, "step": 3330 }, { "epoch": 0.03553024138866287, "grad_norm": 14.284547805786133, "learning_rate": 1.227538876050987e-07, "loss": 0.4433, "num_input_tokens_seen": 6985613312, "step": 3331 }, { "epoch": 0.03580146460537022, "grad_norm": 15.345396995544434, "learning_rate": 1.226273170401411e-07, "loss": 0.3931, "num_input_tokens_seen": 6987710464, "step": 3332 }, { "epoch": 0.03607268782207757, "grad_norm": 19.23891258239746, "learning_rate": 1.2250109040845589e-07, "loss": 0.4526, "num_input_tokens_seen": 6989807616, "step": 3333 }, { "epoch": 0.03634391103878492, "grad_norm": 13.172293663024902, "learning_rate": 1.2237520781162567e-07, "loss": 0.251, "num_input_tokens_seen": 6991904768, "step": 3334 }, { "epoch": 0.03661513425549227, "grad_norm": 14.895088195800781, "learning_rate": 1.222496693509559e-07, "loss": 0.4675, "num_input_tokens_seen": 6994001920, "step": 3335 }, { "epoch": 0.03688635747219962, "grad_norm": 17.735687255859375, "learning_rate": 1.2212447512747506e-07, "loss": 0.4232, "num_input_tokens_seen": 6996099072, "step": 3336 }, { "epoch": 0.03715758068890697, "grad_norm": 10.56796646118164, "learning_rate": 1.2199962524193492e-07, "loss": 0.2311, "num_input_tokens_seen": 6998196224, "step": 3337 }, { "epoch": 0.03742880390561432, "grad_norm": 17.399538040161133, "learning_rate": 1.218751197948096e-07, "loss": 0.4626, "num_input_tokens_seen": 7000293376, "step": 3338 }, { "epoch": 0.03770002712232167, "grad_norm": 21.85934066772461, "learning_rate": 1.2175095888629657e-07, "loss": 0.7143, "num_input_tokens_seen": 7002390528, "step": 3339 }, { "epoch": 0.03797125033902902, "grad_norm": 20.307708740234375, "learning_rate": 1.2162714261631575e-07, "loss": 0.6817, "num_input_tokens_seen": 7004487680, "step": 3340 }, { "epoch": 0.03824247355573637, "grad_norm": 12.21007251739502, "learning_rate": 1.215036710845097e-07, "loss": 0.3089, "num_input_tokens_seen": 7006584832, "step": 3341 }, { "epoch": 0.03851369677244372, "grad_norm": 17.77985954284668, "learning_rate": 1.213805443902437e-07, "loss": 0.3941, "num_input_tokens_seen": 7008681984, "step": 3342 }, { "epoch": 0.038784919989151075, "grad_norm": 13.20972728729248, "learning_rate": 1.212577626326054e-07, "loss": 0.3183, "num_input_tokens_seen": 7010779136, "step": 3343 }, { "epoch": 0.03905614320585842, "grad_norm": 13.693172454833984, "learning_rate": 1.2113532591040488e-07, "loss": 0.3986, "num_input_tokens_seen": 7012876288, "step": 3344 }, { "epoch": 0.03932736642256577, "grad_norm": 18.79658317565918, "learning_rate": 1.2101323432217454e-07, "loss": 0.438, "num_input_tokens_seen": 7014973440, "step": 3345 }, { "epoch": 0.039598589639273124, "grad_norm": 15.080113410949707, "learning_rate": 1.208914879661691e-07, "loss": 0.3442, "num_input_tokens_seen": 7017070592, "step": 3346 }, { "epoch": 0.03986981285598047, "grad_norm": 10.353347778320312, "learning_rate": 1.2077008694036527e-07, "loss": 0.2322, "num_input_tokens_seen": 7019167744, "step": 3347 }, { "epoch": 0.04014103607268782, "grad_norm": 14.034728050231934, "learning_rate": 1.2064903134246221e-07, "loss": 0.4263, "num_input_tokens_seen": 7021264896, "step": 3348 }, { "epoch": 0.040412259289395173, "grad_norm": 16.963212966918945, "learning_rate": 1.2052832126988053e-07, "loss": 0.7119, "num_input_tokens_seen": 7023362048, "step": 3349 }, { "epoch": 0.04068348250610252, "grad_norm": 16.336538314819336, "learning_rate": 1.2040795681976338e-07, "loss": 0.5812, "num_input_tokens_seen": 7025459200, "step": 3350 }, { "epoch": 0.040954705722809875, "grad_norm": 17.866945266723633, "learning_rate": 1.2028793808897537e-07, "loss": 0.6111, "num_input_tokens_seen": 7027556352, "step": 3351 }, { "epoch": 0.04122592893951722, "grad_norm": 19.273788452148438, "learning_rate": 1.20168265174103e-07, "loss": 0.3194, "num_input_tokens_seen": 7029653504, "step": 3352 }, { "epoch": 0.04149715215622457, "grad_norm": 19.322425842285156, "learning_rate": 1.2004893817145456e-07, "loss": 0.3223, "num_input_tokens_seen": 7031750656, "step": 3353 }, { "epoch": 0.041768375372931925, "grad_norm": 19.382596969604492, "learning_rate": 1.1992995717705982e-07, "loss": 0.7825, "num_input_tokens_seen": 7033847808, "step": 3354 }, { "epoch": 0.04203959858963927, "grad_norm": 12.71596908569336, "learning_rate": 1.198113222866702e-07, "loss": 0.326, "num_input_tokens_seen": 7035944960, "step": 3355 }, { "epoch": 0.04231082180634662, "grad_norm": 12.268967628479004, "learning_rate": 1.1969303359575845e-07, "loss": 0.3764, "num_input_tokens_seen": 7038042112, "step": 3356 }, { "epoch": 0.042582045023053974, "grad_norm": 12.389196395874023, "learning_rate": 1.1957509119951885e-07, "loss": 0.3302, "num_input_tokens_seen": 7040139264, "step": 3357 }, { "epoch": 0.04285326823976132, "grad_norm": 13.899662971496582, "learning_rate": 1.1945749519286694e-07, "loss": 0.3105, "num_input_tokens_seen": 7042236416, "step": 3358 }, { "epoch": 0.043124491456468676, "grad_norm": 16.53765106201172, "learning_rate": 1.193402456704396e-07, "loss": 0.4031, "num_input_tokens_seen": 7044333568, "step": 3359 }, { "epoch": 0.043395714673176024, "grad_norm": 17.577390670776367, "learning_rate": 1.1922334272659474e-07, "loss": 0.5253, "num_input_tokens_seen": 7046430720, "step": 3360 }, { "epoch": 0.04366693788988337, "grad_norm": 14.368865966796875, "learning_rate": 1.1910678645541137e-07, "loss": 0.4683, "num_input_tokens_seen": 7048527872, "step": 3361 }, { "epoch": 0.043938161106590726, "grad_norm": 10.524388313293457, "learning_rate": 1.1899057695068954e-07, "loss": 0.2427, "num_input_tokens_seen": 7050625024, "step": 3362 }, { "epoch": 0.04420938432329807, "grad_norm": 18.85384178161621, "learning_rate": 1.1887471430595028e-07, "loss": 0.5473, "num_input_tokens_seen": 7052722176, "step": 3363 }, { "epoch": 0.04448060754000543, "grad_norm": 18.354103088378906, "learning_rate": 1.187591986144354e-07, "loss": 0.5499, "num_input_tokens_seen": 7054819328, "step": 3364 }, { "epoch": 0.044751830756712775, "grad_norm": 17.2979736328125, "learning_rate": 1.1864402996910748e-07, "loss": 0.7023, "num_input_tokens_seen": 7056916480, "step": 3365 }, { "epoch": 0.04502305397342012, "grad_norm": 14.632923126220703, "learning_rate": 1.1852920846265002e-07, "loss": 0.4373, "num_input_tokens_seen": 7059013632, "step": 3366 }, { "epoch": 0.04529427719012748, "grad_norm": 17.784364700317383, "learning_rate": 1.1841473418746685e-07, "loss": 0.6175, "num_input_tokens_seen": 7061110784, "step": 3367 }, { "epoch": 0.045565500406834825, "grad_norm": 13.559286117553711, "learning_rate": 1.1830060723568256e-07, "loss": 0.3332, "num_input_tokens_seen": 7063207936, "step": 3368 }, { "epoch": 0.04583672362354217, "grad_norm": 14.777471542358398, "learning_rate": 1.1818682769914226e-07, "loss": 0.5035, "num_input_tokens_seen": 7065305088, "step": 3369 }, { "epoch": 0.04610794684024953, "grad_norm": 18.318246841430664, "learning_rate": 1.1807339566941123e-07, "loss": 0.6274, "num_input_tokens_seen": 7067402240, "step": 3370 }, { "epoch": 0.046379170056956874, "grad_norm": 16.257017135620117, "learning_rate": 1.1796031123777546e-07, "loss": 0.6656, "num_input_tokens_seen": 7069499392, "step": 3371 }, { "epoch": 0.04665039327366423, "grad_norm": 16.286823272705078, "learning_rate": 1.178475744952408e-07, "loss": 0.3738, "num_input_tokens_seen": 7071596544, "step": 3372 }, { "epoch": 0.046921616490371576, "grad_norm": 13.061100959777832, "learning_rate": 1.1773518553253352e-07, "loss": 0.3498, "num_input_tokens_seen": 7073693696, "step": 3373 }, { "epoch": 0.047192839707078924, "grad_norm": 19.254472732543945, "learning_rate": 1.1762314444010002e-07, "loss": 0.5665, "num_input_tokens_seen": 7075790848, "step": 3374 }, { "epoch": 0.04746406292378628, "grad_norm": 9.76313591003418, "learning_rate": 1.1751145130810675e-07, "loss": 0.1893, "num_input_tokens_seen": 7077888000, "step": 3375 }, { "epoch": 0.047735286140493625, "grad_norm": 11.822184562683105, "learning_rate": 1.1740010622644008e-07, "loss": 0.3819, "num_input_tokens_seen": 7079985152, "step": 3376 }, { "epoch": 0.04800650935720097, "grad_norm": 19.319860458374023, "learning_rate": 1.1728910928470629e-07, "loss": 0.4923, "num_input_tokens_seen": 7082082304, "step": 3377 }, { "epoch": 0.04827773257390833, "grad_norm": 10.68764877319336, "learning_rate": 1.1717846057223143e-07, "loss": 0.2708, "num_input_tokens_seen": 7084179456, "step": 3378 }, { "epoch": 0.048548955790615675, "grad_norm": 18.42941665649414, "learning_rate": 1.1706816017806142e-07, "loss": 0.5531, "num_input_tokens_seen": 7086276608, "step": 3379 }, { "epoch": 0.04882017900732303, "grad_norm": 19.539121627807617, "learning_rate": 1.1695820819096186e-07, "loss": 0.5045, "num_input_tokens_seen": 7088373760, "step": 3380 }, { "epoch": 0.04909140222403038, "grad_norm": 22.56864356994629, "learning_rate": 1.1684860469941785e-07, "loss": 0.8717, "num_input_tokens_seen": 7090470912, "step": 3381 }, { "epoch": 0.049362625440737724, "grad_norm": 18.064945220947266, "learning_rate": 1.1673934979163417e-07, "loss": 0.5734, "num_input_tokens_seen": 7092568064, "step": 3382 }, { "epoch": 0.04963384865744508, "grad_norm": 12.753329277038574, "learning_rate": 1.1663044355553493e-07, "loss": 0.2754, "num_input_tokens_seen": 7094665216, "step": 3383 }, { "epoch": 0.049905071874152426, "grad_norm": 16.793869018554688, "learning_rate": 1.1652188607876379e-07, "loss": 0.4608, "num_input_tokens_seen": 7096762368, "step": 3384 }, { "epoch": 0.05017629509085978, "grad_norm": 18.119853973388672, "learning_rate": 1.1641367744868372e-07, "loss": 0.7662, "num_input_tokens_seen": 7098859520, "step": 3385 }, { "epoch": 0.05044751830756713, "grad_norm": 18.567729949951172, "learning_rate": 1.1630581775237681e-07, "loss": 0.3328, "num_input_tokens_seen": 7100956672, "step": 3386 }, { "epoch": 0.050718741524274476, "grad_norm": 10.636219024658203, "learning_rate": 1.1619830707664446e-07, "loss": 0.2226, "num_input_tokens_seen": 7103053824, "step": 3387 }, { "epoch": 0.05098996474098183, "grad_norm": 15.37312126159668, "learning_rate": 1.160911455080073e-07, "loss": 0.3721, "num_input_tokens_seen": 7105150976, "step": 3388 }, { "epoch": 0.05126118795768918, "grad_norm": 20.098434448242188, "learning_rate": 1.1598433313270472e-07, "loss": 0.5492, "num_input_tokens_seen": 7107248128, "step": 3389 }, { "epoch": 0.051532411174396525, "grad_norm": 17.885168075561523, "learning_rate": 1.1587787003669523e-07, "loss": 0.4829, "num_input_tokens_seen": 7109345280, "step": 3390 }, { "epoch": 0.05180363439110388, "grad_norm": 15.35325813293457, "learning_rate": 1.1577175630565646e-07, "loss": 0.4408, "num_input_tokens_seen": 7111442432, "step": 3391 }, { "epoch": 0.05207485760781123, "grad_norm": 29.616416931152344, "learning_rate": 1.1566599202498456e-07, "loss": 0.7706, "num_input_tokens_seen": 7113539584, "step": 3392 }, { "epoch": 0.05234608082451858, "grad_norm": 10.501668930053711, "learning_rate": 1.1556057727979468e-07, "loss": 0.296, "num_input_tokens_seen": 7115636736, "step": 3393 }, { "epoch": 0.05261730404122593, "grad_norm": 17.35365867614746, "learning_rate": 1.1545551215492055e-07, "loss": 0.6134, "num_input_tokens_seen": 7117733888, "step": 3394 }, { "epoch": 0.05288852725793328, "grad_norm": 22.813182830810547, "learning_rate": 1.1535079673491458e-07, "loss": 0.4376, "num_input_tokens_seen": 7119831040, "step": 3395 }, { "epoch": 0.05315975047464063, "grad_norm": 11.890691757202148, "learning_rate": 1.1524643110404778e-07, "loss": 0.3133, "num_input_tokens_seen": 7121928192, "step": 3396 }, { "epoch": 0.05343097369134798, "grad_norm": 13.04570484161377, "learning_rate": 1.1514241534630963e-07, "loss": 0.3544, "num_input_tokens_seen": 7124025344, "step": 3397 }, { "epoch": 0.05370219690805533, "grad_norm": 15.982794761657715, "learning_rate": 1.1503874954540804e-07, "loss": 0.4587, "num_input_tokens_seen": 7126122496, "step": 3398 }, { "epoch": 0.05397342012476268, "grad_norm": 17.98113441467285, "learning_rate": 1.1493543378476927e-07, "loss": 0.3568, "num_input_tokens_seen": 7128219648, "step": 3399 }, { "epoch": 0.05424464334147003, "grad_norm": 14.370156288146973, "learning_rate": 1.1483246814753798e-07, "loss": 0.464, "num_input_tokens_seen": 7130316800, "step": 3400 }, { "epoch": 0.00027122321670735016, "grad_norm": 12.226541519165039, "learning_rate": 1.1472985271657697e-07, "loss": 0.35, "num_input_tokens_seen": 7132413952, "step": 3401 }, { "epoch": 0.0005424464334147003, "grad_norm": 15.276947975158691, "learning_rate": 1.1462758757446728e-07, "loss": 0.5296, "num_input_tokens_seen": 7134511104, "step": 3402 }, { "epoch": 0.0008136696501220504, "grad_norm": 15.25047492980957, "learning_rate": 1.1452567280350789e-07, "loss": 0.3315, "num_input_tokens_seen": 7136608256, "step": 3403 }, { "epoch": 0.0010848928668294006, "grad_norm": 18.95487403869629, "learning_rate": 1.1442410848571602e-07, "loss": 0.369, "num_input_tokens_seen": 7138705408, "step": 3404 }, { "epoch": 0.0013561160835367507, "grad_norm": 15.813984870910645, "learning_rate": 1.1432289470282683e-07, "loss": 0.4243, "num_input_tokens_seen": 7140802560, "step": 3405 }, { "epoch": 0.0016273393002441008, "grad_norm": 15.477920532226562, "learning_rate": 1.1422203153629312e-07, "loss": 0.3994, "num_input_tokens_seen": 7142899712, "step": 3406 }, { "epoch": 0.001898562516951451, "grad_norm": 15.906928062438965, "learning_rate": 1.1412151906728589e-07, "loss": 0.4129, "num_input_tokens_seen": 7144996864, "step": 3407 }, { "epoch": 0.0021697857336588013, "grad_norm": 19.251590728759766, "learning_rate": 1.1402135737669372e-07, "loss": 0.7718, "num_input_tokens_seen": 7147094016, "step": 3408 }, { "epoch": 0.0024410089503661514, "grad_norm": 14.729852676391602, "learning_rate": 1.1392154654512289e-07, "loss": 0.4258, "num_input_tokens_seen": 7149191168, "step": 3409 }, { "epoch": 0.0027122321670735015, "grad_norm": 16.52750587463379, "learning_rate": 1.1382208665289742e-07, "loss": 0.6094, "num_input_tokens_seen": 7151288320, "step": 3410 }, { "epoch": 0.0029834553837808516, "grad_norm": 30.927867889404297, "learning_rate": 1.1372297778005883e-07, "loss": 0.7976, "num_input_tokens_seen": 7153385472, "step": 3411 }, { "epoch": 0.0032546786004882017, "grad_norm": 14.352428436279297, "learning_rate": 1.1362422000636609e-07, "loss": 0.3816, "num_input_tokens_seen": 7155482624, "step": 3412 }, { "epoch": 0.003525901817195552, "grad_norm": 18.680057525634766, "learning_rate": 1.135258134112958e-07, "loss": 0.4478, "num_input_tokens_seen": 7157579776, "step": 3413 }, { "epoch": 0.003797125033902902, "grad_norm": 13.85073471069336, "learning_rate": 1.1342775807404177e-07, "loss": 0.4313, "num_input_tokens_seen": 7159676928, "step": 3414 }, { "epoch": 0.0040683482506102524, "grad_norm": 12.328075408935547, "learning_rate": 1.1333005407351516e-07, "loss": 0.301, "num_input_tokens_seen": 7161774080, "step": 3415 }, { "epoch": 0.0043395714673176026, "grad_norm": 15.454139709472656, "learning_rate": 1.1323270148834461e-07, "loss": 0.6079, "num_input_tokens_seen": 7163871232, "step": 3416 }, { "epoch": 0.004610794684024953, "grad_norm": 17.02078628540039, "learning_rate": 1.1313570039687571e-07, "loss": 0.5509, "num_input_tokens_seen": 7165968384, "step": 3417 }, { "epoch": 0.004882017900732303, "grad_norm": 16.010780334472656, "learning_rate": 1.1303905087717111e-07, "loss": 0.5081, "num_input_tokens_seen": 7168065536, "step": 3418 }, { "epoch": 0.005153241117439653, "grad_norm": 12.919666290283203, "learning_rate": 1.1294275300701085e-07, "loss": 0.3705, "num_input_tokens_seen": 7170162688, "step": 3419 }, { "epoch": 0.005424464334147003, "grad_norm": 14.903799057006836, "learning_rate": 1.1284680686389163e-07, "loss": 0.5029, "num_input_tokens_seen": 7172259840, "step": 3420 }, { "epoch": 0.005695687550854353, "grad_norm": 18.086734771728516, "learning_rate": 1.1275121252502738e-07, "loss": 0.408, "num_input_tokens_seen": 7174356992, "step": 3421 }, { "epoch": 0.005966910767561703, "grad_norm": 13.25420093536377, "learning_rate": 1.1265597006734872e-07, "loss": 0.3967, "num_input_tokens_seen": 7176454144, "step": 3422 }, { "epoch": 0.006238133984269053, "grad_norm": 16.535417556762695, "learning_rate": 1.1256107956750319e-07, "loss": 0.4737, "num_input_tokens_seen": 7178551296, "step": 3423 }, { "epoch": 0.006509357200976403, "grad_norm": 13.090187072753906, "learning_rate": 1.1246654110185501e-07, "loss": 0.2047, "num_input_tokens_seen": 7180648448, "step": 3424 }, { "epoch": 0.0067805804176837535, "grad_norm": 21.00216293334961, "learning_rate": 1.1237235474648516e-07, "loss": 0.8687, "num_input_tokens_seen": 7182745600, "step": 3425 }, { "epoch": 0.007051803634391104, "grad_norm": 18.46866226196289, "learning_rate": 1.1227852057719125e-07, "loss": 0.5996, "num_input_tokens_seen": 7184842752, "step": 3426 }, { "epoch": 0.007323026851098454, "grad_norm": 21.20566749572754, "learning_rate": 1.121850386694875e-07, "loss": 0.4136, "num_input_tokens_seen": 7186939904, "step": 3427 }, { "epoch": 0.007594250067805804, "grad_norm": 15.836469650268555, "learning_rate": 1.1209190909860453e-07, "loss": 0.5273, "num_input_tokens_seen": 7189037056, "step": 3428 }, { "epoch": 0.007865473284513154, "grad_norm": 17.805950164794922, "learning_rate": 1.119991319394894e-07, "loss": 0.4879, "num_input_tokens_seen": 7191134208, "step": 3429 }, { "epoch": 0.008136696501220505, "grad_norm": 21.2362117767334, "learning_rate": 1.1190670726680579e-07, "loss": 0.6706, "num_input_tokens_seen": 7193231360, "step": 3430 }, { "epoch": 0.008407919717927854, "grad_norm": 12.916275024414062, "learning_rate": 1.1181463515493336e-07, "loss": 0.2483, "num_input_tokens_seen": 7195328512, "step": 3431 }, { "epoch": 0.008679142934635205, "grad_norm": 14.223495483398438, "learning_rate": 1.1172291567796846e-07, "loss": 0.3723, "num_input_tokens_seen": 7197425664, "step": 3432 }, { "epoch": 0.008950366151342554, "grad_norm": 12.331622123718262, "learning_rate": 1.1163154890972333e-07, "loss": 0.3586, "num_input_tokens_seen": 7199522816, "step": 3433 }, { "epoch": 0.009221589368049905, "grad_norm": 9.547661781311035, "learning_rate": 1.1154053492372654e-07, "loss": 0.2378, "num_input_tokens_seen": 7201619968, "step": 3434 }, { "epoch": 0.009492812584757255, "grad_norm": 19.91844367980957, "learning_rate": 1.1144987379322254e-07, "loss": 0.5829, "num_input_tokens_seen": 7203717120, "step": 3435 }, { "epoch": 0.009764035801464606, "grad_norm": 23.920061111450195, "learning_rate": 1.1135956559117207e-07, "loss": 0.7757, "num_input_tokens_seen": 7205814272, "step": 3436 }, { "epoch": 0.010035259018171955, "grad_norm": 9.220373153686523, "learning_rate": 1.1126961039025168e-07, "loss": 0.2022, "num_input_tokens_seen": 7207911424, "step": 3437 }, { "epoch": 0.010306482234879306, "grad_norm": 14.202072143554688, "learning_rate": 1.111800082628539e-07, "loss": 0.3518, "num_input_tokens_seen": 7210008576, "step": 3438 }, { "epoch": 0.010577705451586655, "grad_norm": 16.335010528564453, "learning_rate": 1.1109075928108715e-07, "loss": 0.5367, "num_input_tokens_seen": 7212105728, "step": 3439 }, { "epoch": 0.010848928668294006, "grad_norm": 20.81877326965332, "learning_rate": 1.1100186351677567e-07, "loss": 0.8455, "num_input_tokens_seen": 7214202880, "step": 3440 }, { "epoch": 0.011120151885001357, "grad_norm": 12.109017372131348, "learning_rate": 1.1091332104145921e-07, "loss": 0.404, "num_input_tokens_seen": 7216300032, "step": 3441 }, { "epoch": 0.011391375101708706, "grad_norm": 51.72409439086914, "learning_rate": 1.1082513192639353e-07, "loss": 0.4581, "num_input_tokens_seen": 7218397184, "step": 3442 }, { "epoch": 0.011662598318416057, "grad_norm": 11.072718620300293, "learning_rate": 1.1073729624254984e-07, "loss": 0.2449, "num_input_tokens_seen": 7220494336, "step": 3443 }, { "epoch": 0.011933821535123406, "grad_norm": 20.055789947509766, "learning_rate": 1.1064981406061494e-07, "loss": 0.6404, "num_input_tokens_seen": 7222591488, "step": 3444 }, { "epoch": 0.012205044751830757, "grad_norm": 12.380927085876465, "learning_rate": 1.1056268545099117e-07, "loss": 0.4006, "num_input_tokens_seen": 7224688640, "step": 3445 }, { "epoch": 0.012476267968538107, "grad_norm": 13.249361038208008, "learning_rate": 1.1047591048379635e-07, "loss": 0.2531, "num_input_tokens_seen": 7226785792, "step": 3446 }, { "epoch": 0.012747491185245458, "grad_norm": 16.427322387695312, "learning_rate": 1.1038948922886355e-07, "loss": 0.5931, "num_input_tokens_seen": 7228882944, "step": 3447 }, { "epoch": 0.013018714401952807, "grad_norm": 16.59769630432129, "learning_rate": 1.1030342175574144e-07, "loss": 0.4725, "num_input_tokens_seen": 7230980096, "step": 3448 }, { "epoch": 0.013289937618660158, "grad_norm": 25.384323120117188, "learning_rate": 1.1021770813369378e-07, "loss": 0.6082, "num_input_tokens_seen": 7233077248, "step": 3449 }, { "epoch": 0.013561160835367507, "grad_norm": 14.333030700683594, "learning_rate": 1.1013234843169967e-07, "loss": 0.2975, "num_input_tokens_seen": 7235174400, "step": 3450 }, { "epoch": 0.013832384052074858, "grad_norm": 11.298495292663574, "learning_rate": 1.100473427184534e-07, "loss": 0.2593, "num_input_tokens_seen": 7237271552, "step": 3451 }, { "epoch": 0.014103607268782207, "grad_norm": 16.831668853759766, "learning_rate": 1.0996269106236425e-07, "loss": 0.5843, "num_input_tokens_seen": 7239368704, "step": 3452 }, { "epoch": 0.014374830485489558, "grad_norm": 12.40096664428711, "learning_rate": 1.0987839353155661e-07, "loss": 0.3489, "num_input_tokens_seen": 7241465856, "step": 3453 }, { "epoch": 0.014646053702196907, "grad_norm": 14.534159660339355, "learning_rate": 1.0979445019387e-07, "loss": 0.4209, "num_input_tokens_seen": 7243563008, "step": 3454 }, { "epoch": 0.014917276918904258, "grad_norm": 14.342657089233398, "learning_rate": 1.0971086111685883e-07, "loss": 0.3707, "num_input_tokens_seen": 7245660160, "step": 3455 }, { "epoch": 0.015188500135611608, "grad_norm": 19.60882568359375, "learning_rate": 1.0962762636779235e-07, "loss": 0.5742, "num_input_tokens_seen": 7247757312, "step": 3456 }, { "epoch": 0.015459723352318959, "grad_norm": 18.226106643676758, "learning_rate": 1.0954474601365482e-07, "loss": 0.6129, "num_input_tokens_seen": 7249854464, "step": 3457 }, { "epoch": 0.015730946569026308, "grad_norm": 15.364544868469238, "learning_rate": 1.094622201211451e-07, "loss": 0.438, "num_input_tokens_seen": 7251951616, "step": 3458 }, { "epoch": 0.01600216978573366, "grad_norm": 13.685408592224121, "learning_rate": 1.0938004875667689e-07, "loss": 0.4399, "num_input_tokens_seen": 7254048768, "step": 3459 }, { "epoch": 0.01627339300244101, "grad_norm": 13.642942428588867, "learning_rate": 1.0929823198637866e-07, "loss": 0.3581, "num_input_tokens_seen": 7256145920, "step": 3460 }, { "epoch": 0.01654461621914836, "grad_norm": 15.478311538696289, "learning_rate": 1.0921676987609335e-07, "loss": 0.5739, "num_input_tokens_seen": 7258243072, "step": 3461 }, { "epoch": 0.016815839435855708, "grad_norm": 15.546080589294434, "learning_rate": 1.0913566249137865e-07, "loss": 0.6112, "num_input_tokens_seen": 7260340224, "step": 3462 }, { "epoch": 0.01708706265256306, "grad_norm": 14.256020545959473, "learning_rate": 1.0905490989750656e-07, "loss": 0.532, "num_input_tokens_seen": 7262437376, "step": 3463 }, { "epoch": 0.01735828586927041, "grad_norm": 18.604835510253906, "learning_rate": 1.0897451215946378e-07, "loss": 0.673, "num_input_tokens_seen": 7264534528, "step": 3464 }, { "epoch": 0.01762950908597776, "grad_norm": 17.011693954467773, "learning_rate": 1.0889446934195141e-07, "loss": 0.5489, "num_input_tokens_seen": 7266631680, "step": 3465 }, { "epoch": 0.01790073230268511, "grad_norm": 13.888276100158691, "learning_rate": 1.0881478150938475e-07, "loss": 0.4176, "num_input_tokens_seen": 7268728832, "step": 3466 }, { "epoch": 0.01817195551939246, "grad_norm": 15.176491737365723, "learning_rate": 1.0873544872589361e-07, "loss": 0.396, "num_input_tokens_seen": 7270825984, "step": 3467 }, { "epoch": 0.01844317873609981, "grad_norm": 14.927526473999023, "learning_rate": 1.08656471055322e-07, "loss": 0.4393, "num_input_tokens_seen": 7272923136, "step": 3468 }, { "epoch": 0.01871440195280716, "grad_norm": 21.812307357788086, "learning_rate": 1.0857784856122812e-07, "loss": 0.8493, "num_input_tokens_seen": 7275020288, "step": 3469 }, { "epoch": 0.01898562516951451, "grad_norm": 14.81186580657959, "learning_rate": 1.084995813068843e-07, "loss": 0.3828, "num_input_tokens_seen": 7277117440, "step": 3470 }, { "epoch": 0.01925684838622186, "grad_norm": 16.475910186767578, "learning_rate": 1.0842166935527716e-07, "loss": 0.5323, "num_input_tokens_seen": 7279214592, "step": 3471 }, { "epoch": 0.01952807160292921, "grad_norm": 14.052254676818848, "learning_rate": 1.0834411276910715e-07, "loss": 0.3918, "num_input_tokens_seen": 7281311744, "step": 3472 }, { "epoch": 0.019799294819636562, "grad_norm": 17.08545684814453, "learning_rate": 1.0826691161078895e-07, "loss": 0.5653, "num_input_tokens_seen": 7283408896, "step": 3473 }, { "epoch": 0.02007051803634391, "grad_norm": 13.46341323852539, "learning_rate": 1.0819006594245114e-07, "loss": 0.3184, "num_input_tokens_seen": 7285506048, "step": 3474 }, { "epoch": 0.02034174125305126, "grad_norm": 18.920207977294922, "learning_rate": 1.0811357582593613e-07, "loss": 0.6962, "num_input_tokens_seen": 7287603200, "step": 3475 }, { "epoch": 0.02061296446975861, "grad_norm": 18.384702682495117, "learning_rate": 1.0803744132280025e-07, "loss": 0.8108, "num_input_tokens_seen": 7289700352, "step": 3476 }, { "epoch": 0.020884187686465962, "grad_norm": 19.510459899902344, "learning_rate": 1.0796166249431371e-07, "loss": 0.5891, "num_input_tokens_seen": 7291797504, "step": 3477 }, { "epoch": 0.02115541090317331, "grad_norm": 18.225069046020508, "learning_rate": 1.0788623940146032e-07, "loss": 0.5444, "num_input_tokens_seen": 7293894656, "step": 3478 }, { "epoch": 0.02142663411988066, "grad_norm": 13.333983421325684, "learning_rate": 1.0781117210493781e-07, "loss": 0.3763, "num_input_tokens_seen": 7295991808, "step": 3479 }, { "epoch": 0.021697857336588012, "grad_norm": 18.025068283081055, "learning_rate": 1.0773646066515748e-07, "loss": 0.6043, "num_input_tokens_seen": 7298088960, "step": 3480 }, { "epoch": 0.021969080553295363, "grad_norm": 11.711277961730957, "learning_rate": 1.0766210514224419e-07, "loss": 0.2852, "num_input_tokens_seen": 7300186112, "step": 3481 }, { "epoch": 0.022240303770002714, "grad_norm": 17.692590713500977, "learning_rate": 1.0758810559603651e-07, "loss": 0.4692, "num_input_tokens_seen": 7302283264, "step": 3482 }, { "epoch": 0.02251152698671006, "grad_norm": 13.046731948852539, "learning_rate": 1.0751446208608642e-07, "loss": 0.3152, "num_input_tokens_seen": 7304380416, "step": 3483 }, { "epoch": 0.022782750203417412, "grad_norm": 11.570886611938477, "learning_rate": 1.0744117467165938e-07, "loss": 0.4343, "num_input_tokens_seen": 7306477568, "step": 3484 }, { "epoch": 0.023053973420124763, "grad_norm": 13.106870651245117, "learning_rate": 1.0736824341173442e-07, "loss": 0.3972, "num_input_tokens_seen": 7308574720, "step": 3485 }, { "epoch": 0.023325196636832114, "grad_norm": 16.621774673461914, "learning_rate": 1.0729566836500373e-07, "loss": 0.4051, "num_input_tokens_seen": 7310671872, "step": 3486 }, { "epoch": 0.023596419853539462, "grad_norm": 18.08404541015625, "learning_rate": 1.07223449589873e-07, "loss": 0.4515, "num_input_tokens_seen": 7312769024, "step": 3487 }, { "epoch": 0.023867643070246813, "grad_norm": 10.743152618408203, "learning_rate": 1.0715158714446109e-07, "loss": 0.3258, "num_input_tokens_seen": 7314866176, "step": 3488 }, { "epoch": 0.024138866286954164, "grad_norm": 16.302248001098633, "learning_rate": 1.0708008108660026e-07, "loss": 0.4079, "num_input_tokens_seen": 7316963328, "step": 3489 }, { "epoch": 0.024410089503661515, "grad_norm": 10.805963516235352, "learning_rate": 1.0700893147383582e-07, "loss": 0.2726, "num_input_tokens_seen": 7319060480, "step": 3490 }, { "epoch": 0.024681312720368862, "grad_norm": 15.196581840515137, "learning_rate": 1.069381383634263e-07, "loss": 0.5156, "num_input_tokens_seen": 7321157632, "step": 3491 }, { "epoch": 0.024952535937076213, "grad_norm": 12.510798454284668, "learning_rate": 1.0686770181234322e-07, "loss": 0.4177, "num_input_tokens_seen": 7323254784, "step": 3492 }, { "epoch": 0.025223759153783564, "grad_norm": 14.379700660705566, "learning_rate": 1.0679762187727129e-07, "loss": 0.356, "num_input_tokens_seen": 7325351936, "step": 3493 }, { "epoch": 0.025494982370490915, "grad_norm": 10.159677505493164, "learning_rate": 1.0672789861460818e-07, "loss": 0.2734, "num_input_tokens_seen": 7327449088, "step": 3494 }, { "epoch": 0.025766205587198263, "grad_norm": 14.783156394958496, "learning_rate": 1.0665853208046449e-07, "loss": 0.5266, "num_input_tokens_seen": 7329546240, "step": 3495 }, { "epoch": 0.026037428803905614, "grad_norm": 16.891870498657227, "learning_rate": 1.0658952233066381e-07, "loss": 0.3847, "num_input_tokens_seen": 7331643392, "step": 3496 }, { "epoch": 0.026308652020612965, "grad_norm": 16.215871810913086, "learning_rate": 1.0652086942074255e-07, "loss": 0.5542, "num_input_tokens_seen": 7333740544, "step": 3497 }, { "epoch": 0.026579875237320316, "grad_norm": 11.77064323425293, "learning_rate": 1.0645257340594988e-07, "loss": 0.2651, "num_input_tokens_seen": 7335837696, "step": 3498 }, { "epoch": 0.026851098454027666, "grad_norm": 11.754627227783203, "learning_rate": 1.06384634341248e-07, "loss": 0.3247, "num_input_tokens_seen": 7337934848, "step": 3499 }, { "epoch": 0.027122321670735014, "grad_norm": 12.69980239868164, "learning_rate": 1.0631705228131149e-07, "loss": 0.3777, "num_input_tokens_seen": 7340032000, "step": 3500 }, { "epoch": 0.027393544887442365, "grad_norm": 13.112504959106445, "learning_rate": 1.0624982728052795e-07, "loss": 0.2684, "num_input_tokens_seen": 7342129152, "step": 3501 }, { "epoch": 0.027664768104149716, "grad_norm": 13.424850463867188, "learning_rate": 1.0618295939299752e-07, "loss": 0.3104, "num_input_tokens_seen": 7344226304, "step": 3502 }, { "epoch": 0.027935991320857067, "grad_norm": 17.421981811523438, "learning_rate": 1.0611644867253284e-07, "loss": 0.6812, "num_input_tokens_seen": 7346323456, "step": 3503 }, { "epoch": 0.028207214537564414, "grad_norm": 18.970033645629883, "learning_rate": 1.0605029517265918e-07, "loss": 0.6363, "num_input_tokens_seen": 7348420608, "step": 3504 }, { "epoch": 0.028478437754271765, "grad_norm": 35.321075439453125, "learning_rate": 1.0598449894661445e-07, "loss": 0.3609, "num_input_tokens_seen": 7350517760, "step": 3505 }, { "epoch": 0.028749660970979116, "grad_norm": 16.84616470336914, "learning_rate": 1.0591906004734895e-07, "loss": 0.6572, "num_input_tokens_seen": 7352614912, "step": 3506 }, { "epoch": 0.029020884187686467, "grad_norm": 18.14307403564453, "learning_rate": 1.0585397852752544e-07, "loss": 0.87, "num_input_tokens_seen": 7354712064, "step": 3507 }, { "epoch": 0.029292107404393815, "grad_norm": 16.01613426208496, "learning_rate": 1.0578925443951895e-07, "loss": 0.5537, "num_input_tokens_seen": 7356809216, "step": 3508 }, { "epoch": 0.029563330621101166, "grad_norm": 8.767593383789062, "learning_rate": 1.0572488783541702e-07, "loss": 0.2193, "num_input_tokens_seen": 7358906368, "step": 3509 }, { "epoch": 0.029834553837808517, "grad_norm": 11.045705795288086, "learning_rate": 1.0566087876701941e-07, "loss": 0.3346, "num_input_tokens_seen": 7361003520, "step": 3510 }, { "epoch": 0.030105777054515868, "grad_norm": 17.116188049316406, "learning_rate": 1.0559722728583825e-07, "loss": 0.5335, "num_input_tokens_seen": 7363100672, "step": 3511 }, { "epoch": 0.030377000271223215, "grad_norm": 10.3762845993042, "learning_rate": 1.0553393344309775e-07, "loss": 0.3105, "num_input_tokens_seen": 7365197824, "step": 3512 }, { "epoch": 0.030648223487930566, "grad_norm": 13.67977237701416, "learning_rate": 1.054709972897344e-07, "loss": 0.3106, "num_input_tokens_seen": 7367294976, "step": 3513 }, { "epoch": 0.030919446704637917, "grad_norm": 17.335601806640625, "learning_rate": 1.0540841887639698e-07, "loss": 0.6964, "num_input_tokens_seen": 7369392128, "step": 3514 }, { "epoch": 0.031190669921345268, "grad_norm": 15.248687744140625, "learning_rate": 1.0534619825344596e-07, "loss": 0.3862, "num_input_tokens_seen": 7371489280, "step": 3515 }, { "epoch": 0.031461893138052616, "grad_norm": 12.66186809539795, "learning_rate": 1.052843354709543e-07, "loss": 0.3237, "num_input_tokens_seen": 7373586432, "step": 3516 }, { "epoch": 0.03173311635475997, "grad_norm": 13.880027770996094, "learning_rate": 1.0522283057870675e-07, "loss": 0.3875, "num_input_tokens_seen": 7375683584, "step": 3517 }, { "epoch": 0.03200433957146732, "grad_norm": 18.82491683959961, "learning_rate": 1.0516168362620013e-07, "loss": 0.4796, "num_input_tokens_seen": 7377780736, "step": 3518 }, { "epoch": 0.032275562788174665, "grad_norm": 16.534194946289062, "learning_rate": 1.0510089466264321e-07, "loss": 0.4197, "num_input_tokens_seen": 7379877888, "step": 3519 }, { "epoch": 0.03254678600488202, "grad_norm": 10.575153350830078, "learning_rate": 1.0504046373695648e-07, "loss": 0.283, "num_input_tokens_seen": 7381975040, "step": 3520 }, { "epoch": 0.03281800922158937, "grad_norm": 16.936920166015625, "learning_rate": 1.0498039089777265e-07, "loss": 0.4884, "num_input_tokens_seen": 7384072192, "step": 3521 }, { "epoch": 0.03308923243829672, "grad_norm": 13.413094520568848, "learning_rate": 1.0492067619343594e-07, "loss": 0.4334, "num_input_tokens_seen": 7386169344, "step": 3522 }, { "epoch": 0.03336045565500407, "grad_norm": 9.609505653381348, "learning_rate": 1.0486131967200254e-07, "loss": 0.2049, "num_input_tokens_seen": 7388266496, "step": 3523 }, { "epoch": 0.033631678871711417, "grad_norm": 13.24195384979248, "learning_rate": 1.048023213812403e-07, "loss": 0.296, "num_input_tokens_seen": 7390363648, "step": 3524 }, { "epoch": 0.03390290208841877, "grad_norm": 15.955248832702637, "learning_rate": 1.0474368136862876e-07, "loss": 0.4047, "num_input_tokens_seen": 7392460800, "step": 3525 }, { "epoch": 0.03417412530512612, "grad_norm": 18.489961624145508, "learning_rate": 1.0468539968135922e-07, "loss": 0.6169, "num_input_tokens_seen": 7394557952, "step": 3526 }, { "epoch": 0.034445348521833466, "grad_norm": 15.667023658752441, "learning_rate": 1.046274763663345e-07, "loss": 0.4134, "num_input_tokens_seen": 7396655104, "step": 3527 }, { "epoch": 0.03471657173854082, "grad_norm": 20.19091796875, "learning_rate": 1.045699114701691e-07, "loss": 0.6246, "num_input_tokens_seen": 7398752256, "step": 3528 }, { "epoch": 0.03498779495524817, "grad_norm": 18.28851890563965, "learning_rate": 1.0451270503918906e-07, "loss": 0.5747, "num_input_tokens_seen": 7400849408, "step": 3529 }, { "epoch": 0.03525901817195552, "grad_norm": 20.15885353088379, "learning_rate": 1.0445585711943205e-07, "loss": 0.6841, "num_input_tokens_seen": 7402946560, "step": 3530 }, { "epoch": 0.03553024138866287, "grad_norm": 15.178775787353516, "learning_rate": 1.0439936775664699e-07, "loss": 0.5073, "num_input_tokens_seen": 7405043712, "step": 3531 }, { "epoch": 0.03580146460537022, "grad_norm": 13.089923858642578, "learning_rate": 1.043432369962943e-07, "loss": 0.3552, "num_input_tokens_seen": 7407140864, "step": 3532 }, { "epoch": 0.03607268782207757, "grad_norm": 12.704035758972168, "learning_rate": 1.0428746488354606e-07, "loss": 0.4504, "num_input_tokens_seen": 7409238016, "step": 3533 }, { "epoch": 0.03634391103878492, "grad_norm": 13.723061561584473, "learning_rate": 1.0423205146328548e-07, "loss": 0.3546, "num_input_tokens_seen": 7411335168, "step": 3534 }, { "epoch": 0.03661513425549227, "grad_norm": 15.719633102416992, "learning_rate": 1.0417699678010708e-07, "loss": 0.5327, "num_input_tokens_seen": 7413432320, "step": 3535 }, { "epoch": 0.03688635747219962, "grad_norm": 18.25145721435547, "learning_rate": 1.0412230087831689e-07, "loss": 0.607, "num_input_tokens_seen": 7415529472, "step": 3536 }, { "epoch": 0.03715758068890697, "grad_norm": 16.473398208618164, "learning_rate": 1.0406796380193203e-07, "loss": 0.4334, "num_input_tokens_seen": 7417626624, "step": 3537 }, { "epoch": 0.03742880390561432, "grad_norm": 14.576959609985352, "learning_rate": 1.0401398559468098e-07, "loss": 0.5355, "num_input_tokens_seen": 7419723776, "step": 3538 }, { "epoch": 0.03770002712232167, "grad_norm": 23.102252960205078, "learning_rate": 1.0396036630000324e-07, "loss": 0.4434, "num_input_tokens_seen": 7421820928, "step": 3539 }, { "epoch": 0.03797125033902902, "grad_norm": 14.231484413146973, "learning_rate": 1.0390710596104965e-07, "loss": 0.3967, "num_input_tokens_seen": 7423918080, "step": 3540 }, { "epoch": 0.03824247355573637, "grad_norm": 20.063039779663086, "learning_rate": 1.0385420462068206e-07, "loss": 0.556, "num_input_tokens_seen": 7426015232, "step": 3541 }, { "epoch": 0.03851369677244372, "grad_norm": 11.724411010742188, "learning_rate": 1.0380166232147354e-07, "loss": 0.308, "num_input_tokens_seen": 7428112384, "step": 3542 }, { "epoch": 0.038784919989151075, "grad_norm": 11.57480239868164, "learning_rate": 1.0374947910570805e-07, "loss": 0.3655, "num_input_tokens_seen": 7430209536, "step": 3543 }, { "epoch": 0.03905614320585842, "grad_norm": 13.075111389160156, "learning_rate": 1.0369765501538067e-07, "loss": 0.3675, "num_input_tokens_seen": 7432306688, "step": 3544 }, { "epoch": 0.03932736642256577, "grad_norm": 15.664528846740723, "learning_rate": 1.0364619009219743e-07, "loss": 0.3647, "num_input_tokens_seen": 7434403840, "step": 3545 }, { "epoch": 0.039598589639273124, "grad_norm": 16.314237594604492, "learning_rate": 1.0359508437757544e-07, "loss": 0.3922, "num_input_tokens_seen": 7436500992, "step": 3546 }, { "epoch": 0.03986981285598047, "grad_norm": 15.859907150268555, "learning_rate": 1.0354433791264255e-07, "loss": 0.4337, "num_input_tokens_seen": 7438598144, "step": 3547 }, { "epoch": 0.04014103607268782, "grad_norm": 21.789993286132812, "learning_rate": 1.0349395073823768e-07, "loss": 0.639, "num_input_tokens_seen": 7440695296, "step": 3548 }, { "epoch": 0.040412259289395173, "grad_norm": 13.827730178833008, "learning_rate": 1.0344392289491038e-07, "loss": 0.3828, "num_input_tokens_seen": 7442792448, "step": 3549 }, { "epoch": 0.04068348250610252, "grad_norm": 18.37379264831543, "learning_rate": 1.0339425442292118e-07, "loss": 0.7685, "num_input_tokens_seen": 7444889600, "step": 3550 }, { "epoch": 0.040954705722809875, "grad_norm": 14.437710762023926, "learning_rate": 1.0334494536224146e-07, "loss": 0.508, "num_input_tokens_seen": 7446986752, "step": 3551 }, { "epoch": 0.04122592893951722, "grad_norm": 15.604809761047363, "learning_rate": 1.0329599575255321e-07, "loss": 0.5133, "num_input_tokens_seen": 7449083904, "step": 3552 }, { "epoch": 0.04149715215622457, "grad_norm": 11.595867156982422, "learning_rate": 1.0324740563324923e-07, "loss": 0.311, "num_input_tokens_seen": 7451181056, "step": 3553 }, { "epoch": 0.041768375372931925, "grad_norm": 20.499563217163086, "learning_rate": 1.0319917504343297e-07, "loss": 0.8097, "num_input_tokens_seen": 7453278208, "step": 3554 }, { "epoch": 0.04203959858963927, "grad_norm": 15.316178321838379, "learning_rate": 1.0315130402191866e-07, "loss": 0.3584, "num_input_tokens_seen": 7455375360, "step": 3555 }, { "epoch": 0.04231082180634662, "grad_norm": 13.93494701385498, "learning_rate": 1.0310379260723094e-07, "loss": 0.4657, "num_input_tokens_seen": 7457472512, "step": 3556 }, { "epoch": 0.042582045023053974, "grad_norm": 18.21516227722168, "learning_rate": 1.0305664083760532e-07, "loss": 0.6825, "num_input_tokens_seen": 7459569664, "step": 3557 }, { "epoch": 0.04285326823976132, "grad_norm": 15.643766403198242, "learning_rate": 1.0300984875098772e-07, "loss": 0.5517, "num_input_tokens_seen": 7461666816, "step": 3558 }, { "epoch": 0.043124491456468676, "grad_norm": 19.415977478027344, "learning_rate": 1.0296341638503458e-07, "loss": 0.4861, "num_input_tokens_seen": 7463763968, "step": 3559 }, { "epoch": 0.043395714673176024, "grad_norm": 14.686422348022461, "learning_rate": 1.029173437771129e-07, "loss": 0.3778, "num_input_tokens_seen": 7465861120, "step": 3560 }, { "epoch": 0.04366693788988337, "grad_norm": 18.66650390625, "learning_rate": 1.0287163096430024e-07, "loss": 0.7157, "num_input_tokens_seen": 7467958272, "step": 3561 }, { "epoch": 0.043938161106590726, "grad_norm": 12.286962509155273, "learning_rate": 1.0282627798338444e-07, "loss": 0.331, "num_input_tokens_seen": 7470055424, "step": 3562 }, { "epoch": 0.04420938432329807, "grad_norm": 14.363128662109375, "learning_rate": 1.0278128487086387e-07, "loss": 0.5657, "num_input_tokens_seen": 7472152576, "step": 3563 }, { "epoch": 0.04448060754000543, "grad_norm": 15.838930130004883, "learning_rate": 1.0273665166294735e-07, "loss": 0.3759, "num_input_tokens_seen": 7474249728, "step": 3564 }, { "epoch": 0.044751830756712775, "grad_norm": 12.090499877929688, "learning_rate": 1.0269237839555398e-07, "loss": 0.3603, "num_input_tokens_seen": 7476346880, "step": 3565 }, { "epoch": 0.04502305397342012, "grad_norm": 11.031499862670898, "learning_rate": 1.0264846510431307e-07, "loss": 0.2938, "num_input_tokens_seen": 7478444032, "step": 3566 }, { "epoch": 0.04529427719012748, "grad_norm": 18.495397567749023, "learning_rate": 1.0260491182456453e-07, "loss": 0.8536, "num_input_tokens_seen": 7480541184, "step": 3567 }, { "epoch": 0.045565500406834825, "grad_norm": 12.062615394592285, "learning_rate": 1.0256171859135826e-07, "loss": 0.3868, "num_input_tokens_seen": 7482638336, "step": 3568 }, { "epoch": 0.04583672362354217, "grad_norm": 11.398265838623047, "learning_rate": 1.0251888543945458e-07, "loss": 0.3781, "num_input_tokens_seen": 7484735488, "step": 3569 }, { "epoch": 0.04610794684024953, "grad_norm": 11.258602142333984, "learning_rate": 1.0247641240332397e-07, "loss": 0.368, "num_input_tokens_seen": 7486832640, "step": 3570 }, { "epoch": 0.046379170056956874, "grad_norm": 12.560171127319336, "learning_rate": 1.0243429951714714e-07, "loss": 0.389, "num_input_tokens_seen": 7488929792, "step": 3571 }, { "epoch": 0.04665039327366423, "grad_norm": 13.790288925170898, "learning_rate": 1.023925468148149e-07, "loss": 0.4665, "num_input_tokens_seen": 7491026944, "step": 3572 }, { "epoch": 0.046921616490371576, "grad_norm": 8.38277530670166, "learning_rate": 1.023511543299283e-07, "loss": 0.1851, "num_input_tokens_seen": 7493124096, "step": 3573 }, { "epoch": 0.047192839707078924, "grad_norm": 13.927973747253418, "learning_rate": 1.0231012209579831e-07, "loss": 0.4549, "num_input_tokens_seen": 7495221248, "step": 3574 }, { "epoch": 0.04746406292378628, "grad_norm": 23.091039657592773, "learning_rate": 1.0226945014544624e-07, "loss": 0.6259, "num_input_tokens_seen": 7497318400, "step": 3575 }, { "epoch": 0.047735286140493625, "grad_norm": 17.690340042114258, "learning_rate": 1.0222913851160335e-07, "loss": 0.6754, "num_input_tokens_seen": 7499415552, "step": 3576 }, { "epoch": 0.04800650935720097, "grad_norm": 13.187427520751953, "learning_rate": 1.0218918722671074e-07, "loss": 0.4299, "num_input_tokens_seen": 7501512704, "step": 3577 }, { "epoch": 0.04827773257390833, "grad_norm": 15.680404663085938, "learning_rate": 1.0214959632291984e-07, "loss": 0.5009, "num_input_tokens_seen": 7503609856, "step": 3578 }, { "epoch": 0.048548955790615675, "grad_norm": 13.153005599975586, "learning_rate": 1.0211036583209181e-07, "loss": 0.3405, "num_input_tokens_seen": 7505707008, "step": 3579 }, { "epoch": 0.04882017900732303, "grad_norm": 11.65854549407959, "learning_rate": 1.0207149578579789e-07, "loss": 0.2888, "num_input_tokens_seen": 7507804160, "step": 3580 }, { "epoch": 0.04909140222403038, "grad_norm": 24.025188446044922, "learning_rate": 1.0203298621531923e-07, "loss": 0.4813, "num_input_tokens_seen": 7509901312, "step": 3581 }, { "epoch": 0.049362625440737724, "grad_norm": 22.057125091552734, "learning_rate": 1.0199483715164687e-07, "loss": 0.5313, "num_input_tokens_seen": 7511998464, "step": 3582 }, { "epoch": 0.04963384865744508, "grad_norm": 21.940088272094727, "learning_rate": 1.0195704862548167e-07, "loss": 0.8506, "num_input_tokens_seen": 7514095616, "step": 3583 }, { "epoch": 0.049905071874152426, "grad_norm": 14.703182220458984, "learning_rate": 1.0191962066723448e-07, "loss": 0.4072, "num_input_tokens_seen": 7516192768, "step": 3584 }, { "epoch": 0.05017629509085978, "grad_norm": 8.556319236755371, "learning_rate": 1.0188255330702583e-07, "loss": 0.2465, "num_input_tokens_seen": 7518289920, "step": 3585 }, { "epoch": 0.05044751830756713, "grad_norm": 18.820598602294922, "learning_rate": 1.0184584657468615e-07, "loss": 0.6037, "num_input_tokens_seen": 7520387072, "step": 3586 }, { "epoch": 0.050718741524274476, "grad_norm": 11.99779224395752, "learning_rate": 1.018095004997556e-07, "loss": 0.3087, "num_input_tokens_seen": 7522484224, "step": 3587 }, { "epoch": 0.05098996474098183, "grad_norm": 11.369059562683105, "learning_rate": 1.0177351511148414e-07, "loss": 0.2991, "num_input_tokens_seen": 7524581376, "step": 3588 }, { "epoch": 0.05126118795768918, "grad_norm": 16.702545166015625, "learning_rate": 1.0173789043883147e-07, "loss": 0.476, "num_input_tokens_seen": 7526678528, "step": 3589 }, { "epoch": 0.051532411174396525, "grad_norm": 13.388700485229492, "learning_rate": 1.0170262651046687e-07, "loss": 0.3989, "num_input_tokens_seen": 7528775680, "step": 3590 }, { "epoch": 0.05180363439110388, "grad_norm": 16.760042190551758, "learning_rate": 1.0166772335476951e-07, "loss": 0.476, "num_input_tokens_seen": 7530872832, "step": 3591 }, { "epoch": 0.05207485760781123, "grad_norm": 16.29047203063965, "learning_rate": 1.0163318099982808e-07, "loss": 0.4348, "num_input_tokens_seen": 7532969984, "step": 3592 }, { "epoch": 0.05234608082451858, "grad_norm": 14.072612762451172, "learning_rate": 1.0159899947344094e-07, "loss": 0.4399, "num_input_tokens_seen": 7535067136, "step": 3593 }, { "epoch": 0.05261730404122593, "grad_norm": 12.154025077819824, "learning_rate": 1.0156517880311614e-07, "loss": 0.3253, "num_input_tokens_seen": 7537164288, "step": 3594 }, { "epoch": 0.05288852725793328, "grad_norm": 14.722796440124512, "learning_rate": 1.0153171901607118e-07, "loss": 0.3324, "num_input_tokens_seen": 7539261440, "step": 3595 }, { "epoch": 0.05315975047464063, "grad_norm": 17.35401153564453, "learning_rate": 1.0149862013923329e-07, "loss": 0.6772, "num_input_tokens_seen": 7541358592, "step": 3596 }, { "epoch": 0.05343097369134798, "grad_norm": 13.542460441589355, "learning_rate": 1.0146588219923917e-07, "loss": 0.4358, "num_input_tokens_seen": 7543455744, "step": 3597 }, { "epoch": 0.05370219690805533, "grad_norm": 11.1475248336792, "learning_rate": 1.0143350522243509e-07, "loss": 0.3717, "num_input_tokens_seen": 7545552896, "step": 3598 }, { "epoch": 0.05397342012476268, "grad_norm": 14.331890106201172, "learning_rate": 1.0140148923487675e-07, "loss": 0.4832, "num_input_tokens_seen": 7547650048, "step": 3599 }, { "epoch": 0.05424464334147003, "grad_norm": 13.898726463317871, "learning_rate": 1.0136983426232945e-07, "loss": 0.4623, "num_input_tokens_seen": 7549747200, "step": 3600 }, { "epoch": 0.05451586655817738, "grad_norm": 9.847304344177246, "learning_rate": 1.0133854033026789e-07, "loss": 0.2438, "num_input_tokens_seen": 7551844352, "step": 3601 }, { "epoch": 0.05478708977488473, "grad_norm": 16.682790756225586, "learning_rate": 1.0130760746387622e-07, "loss": 0.2657, "num_input_tokens_seen": 7553941504, "step": 3602 }, { "epoch": 0.05505831299159208, "grad_norm": 12.895301818847656, "learning_rate": 1.0127703568804805e-07, "loss": 0.3056, "num_input_tokens_seen": 7556038656, "step": 3603 }, { "epoch": 0.05532953620829943, "grad_norm": 14.191023826599121, "learning_rate": 1.0124682502738638e-07, "loss": 0.368, "num_input_tokens_seen": 7558135808, "step": 3604 }, { "epoch": 0.05560075942500678, "grad_norm": 18.443326950073242, "learning_rate": 1.0121697550620365e-07, "loss": 0.5653, "num_input_tokens_seen": 7560232960, "step": 3605 }, { "epoch": 0.055871982641714134, "grad_norm": 15.984533309936523, "learning_rate": 1.0118748714852156e-07, "loss": 0.4285, "num_input_tokens_seen": 7562330112, "step": 3606 }, { "epoch": 0.05614320585842148, "grad_norm": 12.376087188720703, "learning_rate": 1.011583599780712e-07, "loss": 0.2091, "num_input_tokens_seen": 7564427264, "step": 3607 }, { "epoch": 0.05641442907512883, "grad_norm": 21.866743087768555, "learning_rate": 1.01129594018293e-07, "loss": 0.6654, "num_input_tokens_seen": 7566524416, "step": 3608 }, { "epoch": 0.05668565229183618, "grad_norm": 18.539138793945312, "learning_rate": 1.0110118929233682e-07, "loss": 0.4428, "num_input_tokens_seen": 7568621568, "step": 3609 }, { "epoch": 0.05695687550854353, "grad_norm": 19.51919174194336, "learning_rate": 1.0107314582306156e-07, "loss": 0.6288, "num_input_tokens_seen": 7570718720, "step": 3610 }, { "epoch": 0.05722809872525088, "grad_norm": 16.458036422729492, "learning_rate": 1.0104546363303566e-07, "loss": 0.5354, "num_input_tokens_seen": 7572815872, "step": 3611 }, { "epoch": 0.05749932194195823, "grad_norm": 18.08392906188965, "learning_rate": 1.0101814274453661e-07, "loss": 0.6547, "num_input_tokens_seen": 7574913024, "step": 3612 }, { "epoch": 0.05777054515866558, "grad_norm": 16.274539947509766, "learning_rate": 1.0099118317955127e-07, "loss": 0.5016, "num_input_tokens_seen": 7577010176, "step": 3613 }, { "epoch": 0.058041768375372935, "grad_norm": 16.817398071289062, "learning_rate": 1.0096458495977564e-07, "loss": 0.5801, "num_input_tokens_seen": 7579107328, "step": 3614 }, { "epoch": 0.05831299159208028, "grad_norm": 12.463071823120117, "learning_rate": 1.0093834810661498e-07, "loss": 0.3728, "num_input_tokens_seen": 7581204480, "step": 3615 }, { "epoch": 0.05858421480878763, "grad_norm": 11.567461967468262, "learning_rate": 1.0091247264118372e-07, "loss": 0.3667, "num_input_tokens_seen": 7583301632, "step": 3616 }, { "epoch": 0.058855438025494984, "grad_norm": 10.735450744628906, "learning_rate": 1.0088695858430539e-07, "loss": 0.2851, "num_input_tokens_seen": 7585398784, "step": 3617 }, { "epoch": 0.05912666124220233, "grad_norm": 13.669662475585938, "learning_rate": 1.0086180595651278e-07, "loss": 0.4644, "num_input_tokens_seen": 7587495936, "step": 3618 }, { "epoch": 0.059397884458909686, "grad_norm": 15.31719970703125, "learning_rate": 1.0083701477804778e-07, "loss": 0.5703, "num_input_tokens_seen": 7589593088, "step": 3619 }, { "epoch": 0.059669107675617034, "grad_norm": 14.730269432067871, "learning_rate": 1.0081258506886134e-07, "loss": 0.5415, "num_input_tokens_seen": 7591690240, "step": 3620 }, { "epoch": 0.05994033089232438, "grad_norm": 12.143908500671387, "learning_rate": 1.0078851684861357e-07, "loss": 0.3662, "num_input_tokens_seen": 7593787392, "step": 3621 }, { "epoch": 0.060211554109031735, "grad_norm": 12.585685729980469, "learning_rate": 1.0076481013667376e-07, "loss": 0.2725, "num_input_tokens_seen": 7595884544, "step": 3622 }, { "epoch": 0.06048277732573908, "grad_norm": 11.40467643737793, "learning_rate": 1.0074146495212001e-07, "loss": 0.307, "num_input_tokens_seen": 7597981696, "step": 3623 }, { "epoch": 0.06075400054244643, "grad_norm": 16.664461135864258, "learning_rate": 1.0071848131373972e-07, "loss": 0.7096, "num_input_tokens_seen": 7600078848, "step": 3624 }, { "epoch": 0.061025223759153785, "grad_norm": 11.929876327514648, "learning_rate": 1.0069585924002924e-07, "loss": 0.3869, "num_input_tokens_seen": 7602176000, "step": 3625 }, { "epoch": 0.06129644697586113, "grad_norm": 20.532058715820312, "learning_rate": 1.0067359874919395e-07, "loss": 0.7844, "num_input_tokens_seen": 7604273152, "step": 3626 }, { "epoch": 0.06156767019256849, "grad_norm": 10.287115097045898, "learning_rate": 1.0065169985914826e-07, "loss": 0.2573, "num_input_tokens_seen": 7606370304, "step": 3627 }, { "epoch": 0.061838893409275834, "grad_norm": 17.162843704223633, "learning_rate": 1.0063016258751553e-07, "loss": 0.6302, "num_input_tokens_seen": 7608467456, "step": 3628 }, { "epoch": 0.06211011662598318, "grad_norm": 14.848957061767578, "learning_rate": 1.0060898695162816e-07, "loss": 0.4309, "num_input_tokens_seen": 7610564608, "step": 3629 }, { "epoch": 0.062381339842690536, "grad_norm": 17.66904640197754, "learning_rate": 1.005881729685275e-07, "loss": 0.4711, "num_input_tokens_seen": 7612661760, "step": 3630 }, { "epoch": 0.06265256305939788, "grad_norm": 20.336261749267578, "learning_rate": 1.0056772065496387e-07, "loss": 0.8298, "num_input_tokens_seen": 7614758912, "step": 3631 }, { "epoch": 0.06292378627610523, "grad_norm": 22.26197052001953, "learning_rate": 1.005476300273965e-07, "loss": 0.6695, "num_input_tokens_seen": 7616856064, "step": 3632 }, { "epoch": 0.06319500949281258, "grad_norm": 14.296065330505371, "learning_rate": 1.0052790110199348e-07, "loss": 0.4288, "num_input_tokens_seen": 7618953216, "step": 3633 }, { "epoch": 0.06346623270951994, "grad_norm": 18.81780433654785, "learning_rate": 1.0050853389463205e-07, "loss": 0.6585, "num_input_tokens_seen": 7621050368, "step": 3634 }, { "epoch": 0.06373745592622729, "grad_norm": 11.14263916015625, "learning_rate": 1.0048952842089805e-07, "loss": 0.3255, "num_input_tokens_seen": 7623147520, "step": 3635 }, { "epoch": 0.06400867914293464, "grad_norm": 12.55125904083252, "learning_rate": 1.0047088469608648e-07, "loss": 0.315, "num_input_tokens_seen": 7625244672, "step": 3636 }, { "epoch": 0.06427990235964198, "grad_norm": 20.127553939819336, "learning_rate": 1.00452602735201e-07, "loss": 0.3163, "num_input_tokens_seen": 7627341824, "step": 3637 }, { "epoch": 0.06455112557634933, "grad_norm": 18.10858726501465, "learning_rate": 1.0043468255295435e-07, "loss": 0.6731, "num_input_tokens_seen": 7629438976, "step": 3638 }, { "epoch": 0.06482234879305669, "grad_norm": 15.758679389953613, "learning_rate": 1.0041712416376795e-07, "loss": 0.5683, "num_input_tokens_seen": 7631536128, "step": 3639 }, { "epoch": 0.06509357200976404, "grad_norm": 12.081664085388184, "learning_rate": 1.0039992758177211e-07, "loss": 0.2851, "num_input_tokens_seen": 7633633280, "step": 3640 }, { "epoch": 0.06536479522647139, "grad_norm": 12.943633079528809, "learning_rate": 1.0038309282080596e-07, "loss": 0.2524, "num_input_tokens_seen": 7635730432, "step": 3641 }, { "epoch": 0.06563601844317873, "grad_norm": 16.313081741333008, "learning_rate": 1.0036661989441755e-07, "loss": 0.5379, "num_input_tokens_seen": 7637827584, "step": 3642 }, { "epoch": 0.06590724165988608, "grad_norm": 14.559349060058594, "learning_rate": 1.0035050881586364e-07, "loss": 0.3989, "num_input_tokens_seen": 7639924736, "step": 3643 }, { "epoch": 0.06617846487659344, "grad_norm": 12.672063827514648, "learning_rate": 1.0033475959810974e-07, "loss": 0.3097, "num_input_tokens_seen": 7642021888, "step": 3644 }, { "epoch": 0.06644968809330079, "grad_norm": 12.448989868164062, "learning_rate": 1.0031937225383036e-07, "loss": 0.4511, "num_input_tokens_seen": 7644119040, "step": 3645 }, { "epoch": 0.06672091131000814, "grad_norm": 12.244915962219238, "learning_rate": 1.0030434679540853e-07, "loss": 0.2887, "num_input_tokens_seen": 7646216192, "step": 3646 }, { "epoch": 0.06699213452671549, "grad_norm": 16.625158309936523, "learning_rate": 1.0028968323493623e-07, "loss": 0.5713, "num_input_tokens_seen": 7648313344, "step": 3647 }, { "epoch": 0.06726335774342283, "grad_norm": 18.604345321655273, "learning_rate": 1.0027538158421413e-07, "loss": 0.6223, "num_input_tokens_seen": 7650410496, "step": 3648 }, { "epoch": 0.06753458096013018, "grad_norm": 17.25090217590332, "learning_rate": 1.002614418547516e-07, "loss": 0.6491, "num_input_tokens_seen": 7652507648, "step": 3649 }, { "epoch": 0.06780580417683754, "grad_norm": 17.5065975189209, "learning_rate": 1.0024786405776686e-07, "loss": 0.5556, "num_input_tokens_seen": 7654604800, "step": 3650 }, { "epoch": 0.06807702739354489, "grad_norm": 13.661022186279297, "learning_rate": 1.0023464820418676e-07, "loss": 0.451, "num_input_tokens_seen": 7656701952, "step": 3651 }, { "epoch": 0.06834825061025224, "grad_norm": 14.327960968017578, "learning_rate": 1.00221794304647e-07, "loss": 0.3287, "num_input_tokens_seen": 7658799104, "step": 3652 }, { "epoch": 0.06861947382695958, "grad_norm": 21.388771057128906, "learning_rate": 1.0020930236949182e-07, "loss": 0.363, "num_input_tokens_seen": 7660896256, "step": 3653 }, { "epoch": 0.06889069704366693, "grad_norm": 9.650550842285156, "learning_rate": 1.0019717240877424e-07, "loss": 0.253, "num_input_tokens_seen": 7662993408, "step": 3654 }, { "epoch": 0.0691619202603743, "grad_norm": 17.260183334350586, "learning_rate": 1.001854044322561e-07, "loss": 0.7714, "num_input_tokens_seen": 7665090560, "step": 3655 }, { "epoch": 0.06943314347708164, "grad_norm": 22.375675201416016, "learning_rate": 1.0017399844940774e-07, "loss": 0.4834, "num_input_tokens_seen": 7667187712, "step": 3656 }, { "epoch": 0.06970436669378899, "grad_norm": 13.567085266113281, "learning_rate": 1.0016295446940827e-07, "loss": 0.3653, "num_input_tokens_seen": 7669284864, "step": 3657 }, { "epoch": 0.06997558991049634, "grad_norm": 17.413497924804688, "learning_rate": 1.001522725011455e-07, "loss": 0.7221, "num_input_tokens_seen": 7671382016, "step": 3658 }, { "epoch": 0.07024681312720368, "grad_norm": 13.802844047546387, "learning_rate": 1.0014195255321583e-07, "loss": 0.4207, "num_input_tokens_seen": 7673479168, "step": 3659 }, { "epoch": 0.07051803634391104, "grad_norm": 13.901693344116211, "learning_rate": 1.0013199463392433e-07, "loss": 0.3311, "num_input_tokens_seen": 7675576320, "step": 3660 }, { "epoch": 0.07078925956061839, "grad_norm": 11.964288711547852, "learning_rate": 1.0012239875128484e-07, "loss": 0.2948, "num_input_tokens_seen": 7677673472, "step": 3661 }, { "epoch": 0.07106048277732574, "grad_norm": 16.391555786132812, "learning_rate": 1.0011316491301973e-07, "loss": 0.4533, "num_input_tokens_seen": 7679770624, "step": 3662 }, { "epoch": 0.07133170599403309, "grad_norm": 15.69762134552002, "learning_rate": 1.0010429312656006e-07, "loss": 0.5279, "num_input_tokens_seen": 7681867776, "step": 3663 }, { "epoch": 0.07160292921074043, "grad_norm": 18.312103271484375, "learning_rate": 1.000957833990454e-07, "loss": 0.6914, "num_input_tokens_seen": 7683964928, "step": 3664 }, { "epoch": 0.0718741524274478, "grad_norm": 11.982022285461426, "learning_rate": 1.0008763573732421e-07, "loss": 0.2825, "num_input_tokens_seen": 7686062080, "step": 3665 }, { "epoch": 0.07214537564415514, "grad_norm": 15.081099510192871, "learning_rate": 1.0007985014795331e-07, "loss": 0.4631, "num_input_tokens_seen": 7688159232, "step": 3666 }, { "epoch": 0.07241659886086249, "grad_norm": 14.211804389953613, "learning_rate": 1.0007242663719824e-07, "loss": 0.4301, "num_input_tokens_seen": 7690256384, "step": 3667 }, { "epoch": 0.07268782207756984, "grad_norm": 16.612119674682617, "learning_rate": 1.0006536521103325e-07, "loss": 0.5014, "num_input_tokens_seen": 7692353536, "step": 3668 }, { "epoch": 0.07295904529427719, "grad_norm": 16.976350784301758, "learning_rate": 1.0005866587514106e-07, "loss": 0.5847, "num_input_tokens_seen": 7694450688, "step": 3669 }, { "epoch": 0.07323026851098453, "grad_norm": 16.93351936340332, "learning_rate": 1.0005232863491297e-07, "loss": 0.5873, "num_input_tokens_seen": 7696547840, "step": 3670 }, { "epoch": 0.0735014917276919, "grad_norm": 19.781625747680664, "learning_rate": 1.0004635349544907e-07, "loss": 0.5883, "num_input_tokens_seen": 7698644992, "step": 3671 }, { "epoch": 0.07377271494439924, "grad_norm": 15.636222839355469, "learning_rate": 1.0004074046155789e-07, "loss": 0.3702, "num_input_tokens_seen": 7700742144, "step": 3672 }, { "epoch": 0.07404393816110659, "grad_norm": 20.640913009643555, "learning_rate": 1.000354895377565e-07, "loss": 0.7943, "num_input_tokens_seen": 7702839296, "step": 3673 }, { "epoch": 0.07431516137781394, "grad_norm": 16.68242073059082, "learning_rate": 1.0003060072827073e-07, "loss": 0.4785, "num_input_tokens_seen": 7704936448, "step": 3674 }, { "epoch": 0.07458638459452128, "grad_norm": 14.670442581176758, "learning_rate": 1.0002607403703492e-07, "loss": 0.3074, "num_input_tokens_seen": 7707033600, "step": 3675 }, { "epoch": 0.07485760781122865, "grad_norm": 19.64070701599121, "learning_rate": 1.000219094676919e-07, "loss": 0.7391, "num_input_tokens_seen": 7709130752, "step": 3676 }, { "epoch": 0.075128831027936, "grad_norm": 12.361337661743164, "learning_rate": 1.0001810702359326e-07, "loss": 0.3917, "num_input_tokens_seen": 7711227904, "step": 3677 }, { "epoch": 0.07540005424464334, "grad_norm": 11.465216636657715, "learning_rate": 1.0001466670779896e-07, "loss": 0.3472, "num_input_tokens_seen": 7713325056, "step": 3678 }, { "epoch": 0.07567127746135069, "grad_norm": 10.229789733886719, "learning_rate": 1.000115885230777e-07, "loss": 0.2105, "num_input_tokens_seen": 7715422208, "step": 3679 }, { "epoch": 0.07594250067805804, "grad_norm": 11.707015991210938, "learning_rate": 1.0000887247190662e-07, "loss": 0.3553, "num_input_tokens_seen": 7717519360, "step": 3680 }, { "epoch": 0.0762137238947654, "grad_norm": 17.815807342529297, "learning_rate": 1.000065185564716e-07, "loss": 0.7441, "num_input_tokens_seen": 7719616512, "step": 3681 }, { "epoch": 0.07648494711147275, "grad_norm": 16.822296142578125, "learning_rate": 1.0000452677866691e-07, "loss": 0.5363, "num_input_tokens_seen": 7721713664, "step": 3682 }, { "epoch": 0.07675617032818009, "grad_norm": 16.27877426147461, "learning_rate": 1.0000289714009542e-07, "loss": 0.5075, "num_input_tokens_seen": 7723810816, "step": 3683 }, { "epoch": 0.07702739354488744, "grad_norm": 14.539108276367188, "learning_rate": 1.000016296420687e-07, "loss": 0.4568, "num_input_tokens_seen": 7725907968, "step": 3684 }, { "epoch": 0.07729861676159479, "grad_norm": 16.22719383239746, "learning_rate": 1.0000072428560674e-07, "loss": 0.494, "num_input_tokens_seen": 7728005120, "step": 3685 }, { "epoch": 0.07756983997830215, "grad_norm": 15.531390190124512, "learning_rate": 1.000001810714381e-07, "loss": 0.4097, "num_input_tokens_seen": 7730102272, "step": 3686 }, { "epoch": 0.0778410631950095, "grad_norm": 10.840725898742676, "learning_rate": 1e-07, "loss": 0.3166, "num_input_tokens_seen": 7732199424, "step": 3687 }, { "epoch": 0.0778410631950095, "num_input_tokens_seen": 7732199424, "step": 3687, "total_flos": 4.352218066950134e+19, "train_loss": 0.036126901625228955, "train_runtime": 9429.4679, "train_samples_per_second": 0.782, "train_steps_per_second": 0.391 } ], "logging_steps": 1.0, "max_steps": 3687, "num_input_tokens_seen": 7732199424, "num_train_epochs": 1, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.352218066950134e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }