|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 15767, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.001268472125325046, |
|
"grad_norm": 0.8813716769218445, |
|
"learning_rate": 4.993657639373375e-05, |
|
"loss": 0.8575, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.002536944250650092, |
|
"grad_norm": 0.941986083984375, |
|
"learning_rate": 4.98731527874675e-05, |
|
"loss": 0.8299, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.003805416375975138, |
|
"grad_norm": 0.8295786380767822, |
|
"learning_rate": 4.980972918120125e-05, |
|
"loss": 0.8491, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.005073888501300184, |
|
"grad_norm": 0.7794184684753418, |
|
"learning_rate": 4.9749476755248304e-05, |
|
"loss": 0.8294, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.00634236062662523, |
|
"grad_norm": 0.8349530100822449, |
|
"learning_rate": 4.9686053148982053e-05, |
|
"loss": 0.8443, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.007610832751950276, |
|
"grad_norm": 0.8651715517044067, |
|
"learning_rate": 4.96226295427158e-05, |
|
"loss": 0.817, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.008879304877275321, |
|
"grad_norm": 0.8325722217559814, |
|
"learning_rate": 4.955920593644955e-05, |
|
"loss": 0.7915, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.010147777002600368, |
|
"grad_norm": 0.8690646886825562, |
|
"learning_rate": 4.94957823301833e-05, |
|
"loss": 0.8673, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.011416249127925413, |
|
"grad_norm": 0.8512411117553711, |
|
"learning_rate": 4.9432358723917043e-05, |
|
"loss": 0.8621, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.01268472125325046, |
|
"grad_norm": 0.8262362480163574, |
|
"learning_rate": 4.936893511765079e-05, |
|
"loss": 0.8387, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.013953193378575505, |
|
"grad_norm": 0.8982943892478943, |
|
"learning_rate": 4.930551151138454e-05, |
|
"loss": 0.8199, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.015221665503900552, |
|
"grad_norm": 1.024151086807251, |
|
"learning_rate": 4.92452590854316e-05, |
|
"loss": 0.8289, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.016490137629225597, |
|
"grad_norm": 0.8547594547271729, |
|
"learning_rate": 4.918183547916535e-05, |
|
"loss": 0.8457, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.017758609754550642, |
|
"grad_norm": 0.8753955364227295, |
|
"learning_rate": 4.9118411872899095e-05, |
|
"loss": 0.8213, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.01902708187987569, |
|
"grad_norm": 0.8638611435890198, |
|
"learning_rate": 4.9054988266632844e-05, |
|
"loss": 0.8542, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.020295554005200736, |
|
"grad_norm": 0.9179531931877136, |
|
"learning_rate": 4.899156466036659e-05, |
|
"loss": 0.8362, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.02156402613052578, |
|
"grad_norm": 0.9708409905433655, |
|
"learning_rate": 4.892814105410034e-05, |
|
"loss": 0.8485, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.022832498255850826, |
|
"grad_norm": 0.9580267667770386, |
|
"learning_rate": 4.8864717447834085e-05, |
|
"loss": 0.8234, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.024100970381175875, |
|
"grad_norm": 0.8894557356834412, |
|
"learning_rate": 4.8801293841567834e-05, |
|
"loss": 0.8227, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.02536944250650092, |
|
"grad_norm": 0.8640332221984863, |
|
"learning_rate": 4.873787023530158e-05, |
|
"loss": 0.8695, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.026637914631825965, |
|
"grad_norm": 0.9173359870910645, |
|
"learning_rate": 4.867444662903533e-05, |
|
"loss": 0.8427, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.02790638675715101, |
|
"grad_norm": 0.8366764187812805, |
|
"learning_rate": 4.8611023022769075e-05, |
|
"loss": 0.8248, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.02917485888247606, |
|
"grad_norm": 0.801325798034668, |
|
"learning_rate": 4.8547599416502824e-05, |
|
"loss": 0.8649, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.030443331007801104, |
|
"grad_norm": 1.0194541215896606, |
|
"learning_rate": 4.848417581023657e-05, |
|
"loss": 0.7695, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.03171180313312615, |
|
"grad_norm": 0.8728939294815063, |
|
"learning_rate": 4.842075220397032e-05, |
|
"loss": 0.8076, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.032980275258451194, |
|
"grad_norm": 0.9340566992759705, |
|
"learning_rate": 4.8357328597704065e-05, |
|
"loss": 0.8078, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.03424874738377624, |
|
"grad_norm": 0.8570923209190369, |
|
"learning_rate": 4.8293904991437814e-05, |
|
"loss": 0.85, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.035517219509101285, |
|
"grad_norm": 0.7447881698608398, |
|
"learning_rate": 4.823048138517156e-05, |
|
"loss": 0.864, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.03678569163442633, |
|
"grad_norm": 0.9067574143409729, |
|
"learning_rate": 4.816705777890531e-05, |
|
"loss": 0.8536, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.03805416375975138, |
|
"grad_norm": 1.009699821472168, |
|
"learning_rate": 4.8103634172639055e-05, |
|
"loss": 0.8023, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.039322635885076423, |
|
"grad_norm": 0.9121712446212769, |
|
"learning_rate": 4.8040210566372804e-05, |
|
"loss": 0.8643, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.04059110801040147, |
|
"grad_norm": 0.8111468553543091, |
|
"learning_rate": 4.797678696010656e-05, |
|
"loss": 0.8364, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.04185958013572652, |
|
"grad_norm": 0.8392479419708252, |
|
"learning_rate": 4.79133633538403e-05, |
|
"loss": 0.8694, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.04312805226105156, |
|
"grad_norm": 1.0781140327453613, |
|
"learning_rate": 4.784993974757405e-05, |
|
"loss": 0.8613, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.04439652438637661, |
|
"grad_norm": 1.0560787916183472, |
|
"learning_rate": 4.7786516141307794e-05, |
|
"loss": 0.8348, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.04566499651170165, |
|
"grad_norm": 0.9052237868309021, |
|
"learning_rate": 4.772309253504155e-05, |
|
"loss": 0.8389, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.0469334686370267, |
|
"grad_norm": 0.698861300945282, |
|
"learning_rate": 4.765966892877529e-05, |
|
"loss": 0.8556, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.04820194076235175, |
|
"grad_norm": 0.8764225840568542, |
|
"learning_rate": 4.759624532250904e-05, |
|
"loss": 0.8294, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.04947041288767679, |
|
"grad_norm": 0.9750523567199707, |
|
"learning_rate": 4.7532821716242784e-05, |
|
"loss": 0.8109, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.05073888501300184, |
|
"grad_norm": 0.931287407875061, |
|
"learning_rate": 4.746939810997654e-05, |
|
"loss": 0.8256, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.05200735713832689, |
|
"grad_norm": 0.8329412341117859, |
|
"learning_rate": 4.740597450371028e-05, |
|
"loss": 0.8394, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.05327582926365193, |
|
"grad_norm": 0.8462470173835754, |
|
"learning_rate": 4.734255089744403e-05, |
|
"loss": 0.8445, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.05454430138897698, |
|
"grad_norm": 0.8757215142250061, |
|
"learning_rate": 4.7279127291177774e-05, |
|
"loss": 0.8467, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.05581277351430202, |
|
"grad_norm": 0.8820921182632446, |
|
"learning_rate": 4.721570368491153e-05, |
|
"loss": 0.8415, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.05708124563962707, |
|
"grad_norm": 0.9306124448776245, |
|
"learning_rate": 4.715228007864527e-05, |
|
"loss": 0.8488, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.05834971776495212, |
|
"grad_norm": 0.9606575965881348, |
|
"learning_rate": 4.708885647237902e-05, |
|
"loss": 0.8495, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.05961818989027716, |
|
"grad_norm": 0.7876096963882446, |
|
"learning_rate": 4.702543286611277e-05, |
|
"loss": 0.8364, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.06088666201560221, |
|
"grad_norm": 1.05940842628479, |
|
"learning_rate": 4.696200925984652e-05, |
|
"loss": 0.856, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.06215513414092725, |
|
"grad_norm": 0.8594347238540649, |
|
"learning_rate": 4.689858565358026e-05, |
|
"loss": 0.8114, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.0634236062662523, |
|
"grad_norm": 0.8395501971244812, |
|
"learning_rate": 4.683516204731401e-05, |
|
"loss": 0.8645, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.06469207839157734, |
|
"grad_norm": 0.9333510994911194, |
|
"learning_rate": 4.677173844104776e-05, |
|
"loss": 0.8196, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.06596055051690239, |
|
"grad_norm": 0.8195106983184814, |
|
"learning_rate": 4.670831483478151e-05, |
|
"loss": 0.8586, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.06722902264222744, |
|
"grad_norm": 0.9601035118103027, |
|
"learning_rate": 4.664489122851525e-05, |
|
"loss": 0.8884, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.06849749476755249, |
|
"grad_norm": 1.006823182106018, |
|
"learning_rate": 4.6581467622249e-05, |
|
"loss": 0.8301, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.06976596689287753, |
|
"grad_norm": 0.9135039448738098, |
|
"learning_rate": 4.651804401598275e-05, |
|
"loss": 0.8267, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.07103443901820257, |
|
"grad_norm": 0.929247260093689, |
|
"learning_rate": 4.64546204097165e-05, |
|
"loss": 0.8569, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.07230291114352762, |
|
"grad_norm": 0.8837388753890991, |
|
"learning_rate": 4.639119680345025e-05, |
|
"loss": 0.8236, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.07357138326885267, |
|
"grad_norm": 0.9246069192886353, |
|
"learning_rate": 4.63309443774973e-05, |
|
"loss": 0.8736, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.07483985539417772, |
|
"grad_norm": 0.803130567073822, |
|
"learning_rate": 4.626752077123105e-05, |
|
"loss": 0.841, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.07610832751950276, |
|
"grad_norm": 0.8640844225883484, |
|
"learning_rate": 4.62040971649648e-05, |
|
"loss": 0.8171, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.0773767996448278, |
|
"grad_norm": 0.7979689240455627, |
|
"learning_rate": 4.614067355869855e-05, |
|
"loss": 0.8282, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.07864527177015285, |
|
"grad_norm": 0.9452027678489685, |
|
"learning_rate": 4.60772499524323e-05, |
|
"loss": 0.8724, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.0799137438954779, |
|
"grad_norm": 0.8859258890151978, |
|
"learning_rate": 4.601382634616604e-05, |
|
"loss": 0.8825, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.08118221602080294, |
|
"grad_norm": 0.7669050693511963, |
|
"learning_rate": 4.595040273989979e-05, |
|
"loss": 0.7727, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.08245068814612799, |
|
"grad_norm": 0.9593196511268616, |
|
"learning_rate": 4.588697913363354e-05, |
|
"loss": 0.8217, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.08371916027145304, |
|
"grad_norm": 0.9165802597999573, |
|
"learning_rate": 4.582355552736729e-05, |
|
"loss": 0.8288, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.08498763239677808, |
|
"grad_norm": 0.8536924123764038, |
|
"learning_rate": 4.576013192110103e-05, |
|
"loss": 0.85, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.08625610452210312, |
|
"grad_norm": 1.2818777561187744, |
|
"learning_rate": 4.569670831483479e-05, |
|
"loss": 0.8595, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.08752457664742817, |
|
"grad_norm": 0.8300599455833435, |
|
"learning_rate": 4.563328470856853e-05, |
|
"loss": 0.8843, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.08879304877275322, |
|
"grad_norm": 0.8837544322013855, |
|
"learning_rate": 4.556986110230228e-05, |
|
"loss": 0.869, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.09006152089807827, |
|
"grad_norm": 0.8470463156700134, |
|
"learning_rate": 4.550643749603602e-05, |
|
"loss": 0.8748, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.0913299930234033, |
|
"grad_norm": 0.9581688046455383, |
|
"learning_rate": 4.544301388976978e-05, |
|
"loss": 0.8545, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.09259846514872835, |
|
"grad_norm": 0.9453362822532654, |
|
"learning_rate": 4.537959028350352e-05, |
|
"loss": 0.8134, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.0938669372740534, |
|
"grad_norm": 1.0060471296310425, |
|
"learning_rate": 4.531616667723727e-05, |
|
"loss": 0.8136, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.09513540939937845, |
|
"grad_norm": 1.0802257061004639, |
|
"learning_rate": 4.525274307097101e-05, |
|
"loss": 0.8548, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.0964038815247035, |
|
"grad_norm": 0.8368780612945557, |
|
"learning_rate": 4.518931946470477e-05, |
|
"loss": 0.8709, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.09767235365002853, |
|
"grad_norm": 0.8320122361183167, |
|
"learning_rate": 4.512589585843851e-05, |
|
"loss": 0.8181, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.09894082577535358, |
|
"grad_norm": 0.8374115824699402, |
|
"learning_rate": 4.506247225217226e-05, |
|
"loss": 0.8549, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.10020929790067863, |
|
"grad_norm": 0.7658076882362366, |
|
"learning_rate": 4.4999048645906e-05, |
|
"loss": 0.8453, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.10147777002600368, |
|
"grad_norm": 0.8223689198493958, |
|
"learning_rate": 4.493562503963976e-05, |
|
"loss": 0.9502, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.10274624215132873, |
|
"grad_norm": 0.8949286341667175, |
|
"learning_rate": 4.487220143337351e-05, |
|
"loss": 0.8565, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.10401471427665378, |
|
"grad_norm": 0.9013976454734802, |
|
"learning_rate": 4.480877782710725e-05, |
|
"loss": 0.8313, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.10528318640197881, |
|
"grad_norm": 0.7802047729492188, |
|
"learning_rate": 4.4745354220841e-05, |
|
"loss": 0.8598, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.10655165852730386, |
|
"grad_norm": 1.0458945035934448, |
|
"learning_rate": 4.468193061457475e-05, |
|
"loss": 0.8358, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.10782013065262891, |
|
"grad_norm": 0.8241732716560364, |
|
"learning_rate": 4.46185070083085e-05, |
|
"loss": 0.8473, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.10908860277795396, |
|
"grad_norm": 0.794696569442749, |
|
"learning_rate": 4.455508340204224e-05, |
|
"loss": 0.8557, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.110357074903279, |
|
"grad_norm": 0.9596499800682068, |
|
"learning_rate": 4.449165979577599e-05, |
|
"loss": 0.8727, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.11162554702860404, |
|
"grad_norm": 0.789816677570343, |
|
"learning_rate": 4.442823618950974e-05, |
|
"loss": 0.8475, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.11289401915392909, |
|
"grad_norm": 0.8604167103767395, |
|
"learning_rate": 4.436481258324349e-05, |
|
"loss": 0.8426, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.11416249127925414, |
|
"grad_norm": 0.7762212157249451, |
|
"learning_rate": 4.430138897697723e-05, |
|
"loss": 0.8532, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.11543096340457919, |
|
"grad_norm": 0.7988696694374084, |
|
"learning_rate": 4.423796537071098e-05, |
|
"loss": 0.8398, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.11669943552990424, |
|
"grad_norm": 0.7536550164222717, |
|
"learning_rate": 4.417454176444473e-05, |
|
"loss": 0.8669, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.11796790765522927, |
|
"grad_norm": 0.8086602687835693, |
|
"learning_rate": 4.411111815817848e-05, |
|
"loss": 0.858, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.11923637978055432, |
|
"grad_norm": 0.8665058612823486, |
|
"learning_rate": 4.404769455191222e-05, |
|
"loss": 0.8445, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.12050485190587937, |
|
"grad_norm": 0.9934934973716736, |
|
"learning_rate": 4.398427094564597e-05, |
|
"loss": 0.8548, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.12177332403120442, |
|
"grad_norm": 1.0823296308517456, |
|
"learning_rate": 4.392084733937972e-05, |
|
"loss": 0.8598, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.12304179615652946, |
|
"grad_norm": 0.767232358455658, |
|
"learning_rate": 4.385742373311347e-05, |
|
"loss": 0.8375, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.1243102682818545, |
|
"grad_norm": 0.9011818170547485, |
|
"learning_rate": 4.379400012684721e-05, |
|
"loss": 0.8322, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.12557874040717956, |
|
"grad_norm": 0.7959470152854919, |
|
"learning_rate": 4.373057652058096e-05, |
|
"loss": 0.8742, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.1268472125325046, |
|
"grad_norm": 0.7924261093139648, |
|
"learning_rate": 4.366715291431471e-05, |
|
"loss": 0.854, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.12811568465782963, |
|
"grad_norm": 0.7666494250297546, |
|
"learning_rate": 4.360372930804846e-05, |
|
"loss": 0.8068, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.12938415678315468, |
|
"grad_norm": 0.7881539463996887, |
|
"learning_rate": 4.354030570178221e-05, |
|
"loss": 0.8483, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.13065262890847973, |
|
"grad_norm": 0.8898656964302063, |
|
"learning_rate": 4.3476882095515956e-05, |
|
"loss": 0.8818, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.13192110103380478, |
|
"grad_norm": 0.8294678330421448, |
|
"learning_rate": 4.3413458489249705e-05, |
|
"loss": 0.8118, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.13318957315912983, |
|
"grad_norm": 0.867928147315979, |
|
"learning_rate": 4.335003488298345e-05, |
|
"loss": 0.8602, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.13445804528445487, |
|
"grad_norm": 0.8367668390274048, |
|
"learning_rate": 4.32866112767172e-05, |
|
"loss": 0.8859, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.13572651740977992, |
|
"grad_norm": 0.8341355919837952, |
|
"learning_rate": 4.3223187670450946e-05, |
|
"loss": 0.8484, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.13699498953510497, |
|
"grad_norm": 0.8836013674736023, |
|
"learning_rate": 4.3159764064184695e-05, |
|
"loss": 0.851, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.13826346166043002, |
|
"grad_norm": 0.7626996040344238, |
|
"learning_rate": 4.309634045791844e-05, |
|
"loss": 0.8308, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.13953193378575507, |
|
"grad_norm": 0.926589846611023, |
|
"learning_rate": 4.303291685165219e-05, |
|
"loss": 0.8699, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.14080040591108012, |
|
"grad_norm": 0.79881751537323, |
|
"learning_rate": 4.2969493245385936e-05, |
|
"loss": 0.8215, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.14206887803640514, |
|
"grad_norm": 0.96977299451828, |
|
"learning_rate": 4.2906069639119685e-05, |
|
"loss": 0.8218, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.1433373501617302, |
|
"grad_norm": 0.786342442035675, |
|
"learning_rate": 4.284264603285343e-05, |
|
"loss": 0.8128, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.14460582228705524, |
|
"grad_norm": 0.8864126205444336, |
|
"learning_rate": 4.277922242658718e-05, |
|
"loss": 0.8931, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.14587429441238028, |
|
"grad_norm": 0.9522060751914978, |
|
"learning_rate": 4.2715798820320926e-05, |
|
"loss": 0.8592, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.14714276653770533, |
|
"grad_norm": 0.7771050333976746, |
|
"learning_rate": 4.2652375214054675e-05, |
|
"loss": 0.8473, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.14841123866303038, |
|
"grad_norm": 0.9344542026519775, |
|
"learning_rate": 4.258895160778842e-05, |
|
"loss": 0.8611, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.14967971078835543, |
|
"grad_norm": 0.8358260989189148, |
|
"learning_rate": 4.252552800152217e-05, |
|
"loss": 0.8689, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.15094818291368048, |
|
"grad_norm": 0.9203112125396729, |
|
"learning_rate": 4.2462104395255916e-05, |
|
"loss": 0.8701, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.15221665503900553, |
|
"grad_norm": 0.87474125623703, |
|
"learning_rate": 4.2398680788989665e-05, |
|
"loss": 0.8441, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.15348512716433058, |
|
"grad_norm": 0.838341474533081, |
|
"learning_rate": 4.233525718272341e-05, |
|
"loss": 0.8402, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.1547535992896556, |
|
"grad_norm": 0.9312511682510376, |
|
"learning_rate": 4.227183357645716e-05, |
|
"loss": 0.8594, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.15602207141498065, |
|
"grad_norm": 0.8382570147514343, |
|
"learning_rate": 4.220840997019091e-05, |
|
"loss": 0.8629, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.1572905435403057, |
|
"grad_norm": 0.7457637190818787, |
|
"learning_rate": 4.2144986363924655e-05, |
|
"loss": 0.8518, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.15855901566563074, |
|
"grad_norm": 0.9622685313224792, |
|
"learning_rate": 4.2081562757658405e-05, |
|
"loss": 0.8555, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.1598274877909558, |
|
"grad_norm": 0.8078732490539551, |
|
"learning_rate": 4.201813915139215e-05, |
|
"loss": 0.8499, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.16109595991628084, |
|
"grad_norm": 0.9720255136489868, |
|
"learning_rate": 4.19547155451259e-05, |
|
"loss": 0.8273, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.1623644320416059, |
|
"grad_norm": 0.86973637342453, |
|
"learning_rate": 4.1891291938859645e-05, |
|
"loss": 0.8212, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.16363290416693094, |
|
"grad_norm": 0.7209818363189697, |
|
"learning_rate": 4.1827868332593395e-05, |
|
"loss": 0.8207, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.16490137629225599, |
|
"grad_norm": 0.8860530853271484, |
|
"learning_rate": 4.176444472632714e-05, |
|
"loss": 0.8513, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.16616984841758103, |
|
"grad_norm": 0.8513688445091248, |
|
"learning_rate": 4.170102112006089e-05, |
|
"loss": 0.8581, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.16743832054290608, |
|
"grad_norm": 0.805892825126648, |
|
"learning_rate": 4.1637597513794635e-05, |
|
"loss": 0.8173, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.1687067926682311, |
|
"grad_norm": 0.7897241115570068, |
|
"learning_rate": 4.1574173907528385e-05, |
|
"loss": 0.8358, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.16997526479355615, |
|
"grad_norm": 0.8170486092567444, |
|
"learning_rate": 4.151075030126213e-05, |
|
"loss": 0.8259, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.1712437369188812, |
|
"grad_norm": 0.993430495262146, |
|
"learning_rate": 4.144732669499588e-05, |
|
"loss": 0.8485, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.17251220904420625, |
|
"grad_norm": 0.8168458342552185, |
|
"learning_rate": 4.1383903088729625e-05, |
|
"loss": 0.8446, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.1737806811695313, |
|
"grad_norm": 0.9205940365791321, |
|
"learning_rate": 4.1320479482463375e-05, |
|
"loss": 0.8119, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.17504915329485635, |
|
"grad_norm": 0.7091718912124634, |
|
"learning_rate": 4.1257055876197124e-05, |
|
"loss": 0.8672, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.1763176254201814, |
|
"grad_norm": 0.7693396210670471, |
|
"learning_rate": 4.119363226993087e-05, |
|
"loss": 0.8285, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.17758609754550644, |
|
"grad_norm": 1.0046180486679077, |
|
"learning_rate": 4.1130208663664615e-05, |
|
"loss": 0.8405, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.1788545696708315, |
|
"grad_norm": 0.8736918568611145, |
|
"learning_rate": 4.1066785057398365e-05, |
|
"loss": 0.838, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.18012304179615654, |
|
"grad_norm": 0.9117953777313232, |
|
"learning_rate": 4.1003361451132114e-05, |
|
"loss": 0.866, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.18139151392148156, |
|
"grad_norm": 0.7822251319885254, |
|
"learning_rate": 4.093993784486586e-05, |
|
"loss": 0.9023, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.1826599860468066, |
|
"grad_norm": 0.8296000361442566, |
|
"learning_rate": 4.087651423859961e-05, |
|
"loss": 0.8215, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.18392845817213166, |
|
"grad_norm": 0.814051628112793, |
|
"learning_rate": 4.0813090632333355e-05, |
|
"loss": 0.8362, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.1851969302974567, |
|
"grad_norm": 0.8576933145523071, |
|
"learning_rate": 4.0749667026067104e-05, |
|
"loss": 0.8584, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.18646540242278176, |
|
"grad_norm": 1.1711878776550293, |
|
"learning_rate": 4.068624341980085e-05, |
|
"loss": 0.8411, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.1877338745481068, |
|
"grad_norm": 0.8704431056976318, |
|
"learning_rate": 4.06228198135346e-05, |
|
"loss": 0.8548, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.18900234667343185, |
|
"grad_norm": 0.8817090392112732, |
|
"learning_rate": 4.0559396207268345e-05, |
|
"loss": 0.8284, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.1902708187987569, |
|
"grad_norm": 0.8044705390930176, |
|
"learning_rate": 4.0495972601002094e-05, |
|
"loss": 0.8242, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.19153929092408195, |
|
"grad_norm": 0.7911235690116882, |
|
"learning_rate": 4.043254899473584e-05, |
|
"loss": 0.8239, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.192807763049407, |
|
"grad_norm": 0.8657433390617371, |
|
"learning_rate": 4.036912538846959e-05, |
|
"loss": 0.8538, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.19407623517473205, |
|
"grad_norm": 0.7675552368164062, |
|
"learning_rate": 4.0305701782203335e-05, |
|
"loss": 0.8036, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.19534470730005707, |
|
"grad_norm": 0.8326247930526733, |
|
"learning_rate": 4.0242278175937084e-05, |
|
"loss": 0.8808, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.19661317942538212, |
|
"grad_norm": 0.9370586276054382, |
|
"learning_rate": 4.017885456967083e-05, |
|
"loss": 0.8826, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.19788165155070717, |
|
"grad_norm": 1.0956268310546875, |
|
"learning_rate": 4.011543096340458e-05, |
|
"loss": 0.8272, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.19915012367603221, |
|
"grad_norm": 0.8846516609191895, |
|
"learning_rate": 4.0052007357138325e-05, |
|
"loss": 0.7996, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.20041859580135726, |
|
"grad_norm": 0.8572809100151062, |
|
"learning_rate": 3.998858375087208e-05, |
|
"loss": 0.8098, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.2016870679266823, |
|
"grad_norm": 0.9393579959869385, |
|
"learning_rate": 3.992516014460582e-05, |
|
"loss": 0.8076, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.20295554005200736, |
|
"grad_norm": 0.7605802416801453, |
|
"learning_rate": 3.986173653833957e-05, |
|
"loss": 0.8284, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.2042240121773324, |
|
"grad_norm": 0.6856004595756531, |
|
"learning_rate": 3.9798312932073315e-05, |
|
"loss": 0.8277, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.20549248430265746, |
|
"grad_norm": 0.889754593372345, |
|
"learning_rate": 3.973488932580707e-05, |
|
"loss": 0.863, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.2067609564279825, |
|
"grad_norm": 0.6881526112556458, |
|
"learning_rate": 3.967146571954081e-05, |
|
"loss": 0.8966, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.20802942855330755, |
|
"grad_norm": 0.9273505806922913, |
|
"learning_rate": 3.960804211327456e-05, |
|
"loss": 0.867, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.20929790067863258, |
|
"grad_norm": 0.9193258285522461, |
|
"learning_rate": 3.954461850700831e-05, |
|
"loss": 0.8131, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.21056637280395762, |
|
"grad_norm": 0.9435883164405823, |
|
"learning_rate": 3.948119490074206e-05, |
|
"loss": 0.8066, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.21183484492928267, |
|
"grad_norm": 0.8440527319908142, |
|
"learning_rate": 3.941777129447581e-05, |
|
"loss": 0.8195, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.21310331705460772, |
|
"grad_norm": 0.7809598445892334, |
|
"learning_rate": 3.935434768820955e-05, |
|
"loss": 0.8503, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.21437178917993277, |
|
"grad_norm": 0.7411904335021973, |
|
"learning_rate": 3.92909240819433e-05, |
|
"loss": 0.8055, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.21564026130525782, |
|
"grad_norm": 0.9117131233215332, |
|
"learning_rate": 3.922750047567705e-05, |
|
"loss": 0.8789, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.21690873343058287, |
|
"grad_norm": 0.816663920879364, |
|
"learning_rate": 3.91640768694108e-05, |
|
"loss": 0.8311, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.21817720555590792, |
|
"grad_norm": 0.7956539988517761, |
|
"learning_rate": 3.910065326314454e-05, |
|
"loss": 0.8475, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.21944567768123296, |
|
"grad_norm": 0.858045756816864, |
|
"learning_rate": 3.903722965687829e-05, |
|
"loss": 0.8234, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.220714149806558, |
|
"grad_norm": 0.8653853535652161, |
|
"learning_rate": 3.897380605061204e-05, |
|
"loss": 0.8363, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.22198262193188303, |
|
"grad_norm": 0.8181082606315613, |
|
"learning_rate": 3.891038244434579e-05, |
|
"loss": 0.8224, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.22325109405720808, |
|
"grad_norm": 0.8543662428855896, |
|
"learning_rate": 3.884695883807953e-05, |
|
"loss": 0.8572, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.22451956618253313, |
|
"grad_norm": 0.8624857068061829, |
|
"learning_rate": 3.878353523181328e-05, |
|
"loss": 0.822, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.22578803830785818, |
|
"grad_norm": 0.8224851489067078, |
|
"learning_rate": 3.872011162554703e-05, |
|
"loss": 0.8651, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.22705651043318323, |
|
"grad_norm": 0.7411419749259949, |
|
"learning_rate": 3.865668801928078e-05, |
|
"loss": 0.8125, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.22832498255850828, |
|
"grad_norm": 0.7959814667701721, |
|
"learning_rate": 3.859326441301452e-05, |
|
"loss": 0.8178, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.22959345468383333, |
|
"grad_norm": 0.8626113533973694, |
|
"learning_rate": 3.852984080674827e-05, |
|
"loss": 0.8471, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.23086192680915837, |
|
"grad_norm": 0.8034262657165527, |
|
"learning_rate": 3.846641720048202e-05, |
|
"loss": 0.8552, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.23213039893448342, |
|
"grad_norm": 0.7086619734764099, |
|
"learning_rate": 3.840299359421577e-05, |
|
"loss": 0.8164, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.23339887105980847, |
|
"grad_norm": 0.8688384890556335, |
|
"learning_rate": 3.833956998794951e-05, |
|
"loss": 0.8591, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.23466734318513352, |
|
"grad_norm": 0.8770636916160583, |
|
"learning_rate": 3.827614638168326e-05, |
|
"loss": 0.835, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.23593581531045854, |
|
"grad_norm": 0.6980867981910706, |
|
"learning_rate": 3.821272277541702e-05, |
|
"loss": 0.8467, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.2372042874357836, |
|
"grad_norm": 0.7267412543296814, |
|
"learning_rate": 3.814929916915076e-05, |
|
"loss": 0.8238, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.23847275956110864, |
|
"grad_norm": 0.7703206539154053, |
|
"learning_rate": 3.808587556288451e-05, |
|
"loss": 0.8233, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.2397412316864337, |
|
"grad_norm": 0.8153693079948425, |
|
"learning_rate": 3.802245195661825e-05, |
|
"loss": 0.8361, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.24100970381175874, |
|
"grad_norm": 0.7741461992263794, |
|
"learning_rate": 3.795902835035201e-05, |
|
"loss": 0.9083, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.24227817593708378, |
|
"grad_norm": 0.8532843589782715, |
|
"learning_rate": 3.789560474408575e-05, |
|
"loss": 0.8692, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.24354664806240883, |
|
"grad_norm": 0.8939579725265503, |
|
"learning_rate": 3.78321811378195e-05, |
|
"loss": 0.8336, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.24481512018773388, |
|
"grad_norm": 0.8242042660713196, |
|
"learning_rate": 3.776875753155325e-05, |
|
"loss": 0.7705, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.24608359231305893, |
|
"grad_norm": 0.9386515021324158, |
|
"learning_rate": 3.7705333925287e-05, |
|
"loss": 0.8021, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.24735206443838398, |
|
"grad_norm": 0.9312074184417725, |
|
"learning_rate": 3.764191031902074e-05, |
|
"loss": 0.8113, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.248620536563709, |
|
"grad_norm": 0.8644290566444397, |
|
"learning_rate": 3.757848671275449e-05, |
|
"loss": 0.8336, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.24988900868903405, |
|
"grad_norm": 0.7512555122375488, |
|
"learning_rate": 3.751506310648824e-05, |
|
"loss": 0.832, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.2511574808143591, |
|
"grad_norm": 0.8139438629150391, |
|
"learning_rate": 3.745163950022199e-05, |
|
"loss": 0.8583, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.2524259529396842, |
|
"grad_norm": 0.800618588924408, |
|
"learning_rate": 3.738821589395573e-05, |
|
"loss": 0.8822, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.2536944250650092, |
|
"grad_norm": 0.8025221228599548, |
|
"learning_rate": 3.732479228768948e-05, |
|
"loss": 0.8326, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.25496289719033427, |
|
"grad_norm": 0.9008484482765198, |
|
"learning_rate": 3.726136868142323e-05, |
|
"loss": 0.8686, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.25623136931565926, |
|
"grad_norm": 0.8681203126907349, |
|
"learning_rate": 3.719794507515698e-05, |
|
"loss": 0.8329, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.2574998414409843, |
|
"grad_norm": 0.7322081327438354, |
|
"learning_rate": 3.713452146889072e-05, |
|
"loss": 0.8128, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.25876831356630936, |
|
"grad_norm": 0.8935046792030334, |
|
"learning_rate": 3.707109786262447e-05, |
|
"loss": 0.8317, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.2600367856916344, |
|
"grad_norm": 0.8813959360122681, |
|
"learning_rate": 3.700767425635822e-05, |
|
"loss": 0.855, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.26130525781695946, |
|
"grad_norm": 0.7657369375228882, |
|
"learning_rate": 3.694425065009197e-05, |
|
"loss": 0.7865, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.2625737299422845, |
|
"grad_norm": 0.7684091329574585, |
|
"learning_rate": 3.688082704382572e-05, |
|
"loss": 0.8039, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.26384220206760955, |
|
"grad_norm": 0.7692010998725891, |
|
"learning_rate": 3.681740343755946e-05, |
|
"loss": 0.841, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.2651106741929346, |
|
"grad_norm": 0.8434869050979614, |
|
"learning_rate": 3.675397983129321e-05, |
|
"loss": 0.831, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.26637914631825965, |
|
"grad_norm": 0.7663161158561707, |
|
"learning_rate": 3.669055622502696e-05, |
|
"loss": 0.7749, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.2676476184435847, |
|
"grad_norm": 0.8157145380973816, |
|
"learning_rate": 3.662713261876071e-05, |
|
"loss": 0.8067, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.26891609056890975, |
|
"grad_norm": 0.8220170140266418, |
|
"learning_rate": 3.656370901249445e-05, |
|
"loss": 0.7872, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.2701845626942348, |
|
"grad_norm": 0.9565273523330688, |
|
"learning_rate": 3.6500285406228205e-05, |
|
"loss": 0.8294, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.27145303481955985, |
|
"grad_norm": 0.775265634059906, |
|
"learning_rate": 3.643686179996195e-05, |
|
"loss": 0.8129, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.2727215069448849, |
|
"grad_norm": 0.9544994235038757, |
|
"learning_rate": 3.63734381936957e-05, |
|
"loss": 0.8212, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.27398997907020994, |
|
"grad_norm": 1.041999340057373, |
|
"learning_rate": 3.631001458742944e-05, |
|
"loss": 0.8178, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.275258451195535, |
|
"grad_norm": 0.8267019987106323, |
|
"learning_rate": 3.6246590981163195e-05, |
|
"loss": 0.8306, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.27652692332086004, |
|
"grad_norm": 0.849429190158844, |
|
"learning_rate": 3.618316737489694e-05, |
|
"loss": 0.8179, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.2777953954461851, |
|
"grad_norm": 0.8038565516471863, |
|
"learning_rate": 3.611974376863069e-05, |
|
"loss": 0.82, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.27906386757151014, |
|
"grad_norm": 0.7065672874450684, |
|
"learning_rate": 3.605632016236443e-05, |
|
"loss": 0.8313, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.2803323396968352, |
|
"grad_norm": 0.8574305772781372, |
|
"learning_rate": 3.5992896556098185e-05, |
|
"loss": 0.8529, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.28160081182216024, |
|
"grad_norm": 0.8780136108398438, |
|
"learning_rate": 3.592947294983193e-05, |
|
"loss": 0.8181, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.28286928394748523, |
|
"grad_norm": 0.8531150817871094, |
|
"learning_rate": 3.586604934356568e-05, |
|
"loss": 0.8044, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.2841377560728103, |
|
"grad_norm": 0.807064414024353, |
|
"learning_rate": 3.580262573729942e-05, |
|
"loss": 0.8388, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.2854062281981353, |
|
"grad_norm": 0.8407668471336365, |
|
"learning_rate": 3.5739202131033175e-05, |
|
"loss": 0.8579, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.2866747003234604, |
|
"grad_norm": 0.9750702977180481, |
|
"learning_rate": 3.567577852476692e-05, |
|
"loss": 0.8606, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.2879431724487854, |
|
"grad_norm": 0.8305932283401489, |
|
"learning_rate": 3.561235491850067e-05, |
|
"loss": 0.8312, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.28921164457411047, |
|
"grad_norm": 0.7954509854316711, |
|
"learning_rate": 3.5548931312234416e-05, |
|
"loss": 0.8504, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.2904801166994355, |
|
"grad_norm": 0.8049436211585999, |
|
"learning_rate": 3.5485507705968165e-05, |
|
"loss": 0.887, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.29174858882476057, |
|
"grad_norm": 0.7284989953041077, |
|
"learning_rate": 3.5422084099701914e-05, |
|
"loss": 0.811, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.2930170609500856, |
|
"grad_norm": 0.8093234300613403, |
|
"learning_rate": 3.535866049343566e-05, |
|
"loss": 0.8339, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.29428553307541067, |
|
"grad_norm": 0.7123083472251892, |
|
"learning_rate": 3.5295236887169406e-05, |
|
"loss": 0.8251, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.2955540052007357, |
|
"grad_norm": 0.8694972395896912, |
|
"learning_rate": 3.5231813280903155e-05, |
|
"loss": 0.8166, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.29682247732606076, |
|
"grad_norm": 0.8641796708106995, |
|
"learning_rate": 3.5168389674636904e-05, |
|
"loss": 0.811, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.2980909494513858, |
|
"grad_norm": 0.8516160249710083, |
|
"learning_rate": 3.510496606837065e-05, |
|
"loss": 0.8965, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.29935942157671086, |
|
"grad_norm": 0.866295337677002, |
|
"learning_rate": 3.5041542462104396e-05, |
|
"loss": 0.8301, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.3006278937020359, |
|
"grad_norm": 0.8344794511795044, |
|
"learning_rate": 3.4978118855838145e-05, |
|
"loss": 0.8512, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.30189636582736096, |
|
"grad_norm": 0.8196772933006287, |
|
"learning_rate": 3.4914695249571894e-05, |
|
"loss": 0.8471, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.303164837952686, |
|
"grad_norm": 0.66361004114151, |
|
"learning_rate": 3.485127164330564e-05, |
|
"loss": 0.8528, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.30443331007801105, |
|
"grad_norm": 0.8254550099372864, |
|
"learning_rate": 3.4787848037039386e-05, |
|
"loss": 0.8133, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.3057017822033361, |
|
"grad_norm": 0.7867299914360046, |
|
"learning_rate": 3.4724424430773135e-05, |
|
"loss": 0.8183, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.30697025432866115, |
|
"grad_norm": 1.0029250383377075, |
|
"learning_rate": 3.4661000824506884e-05, |
|
"loss": 0.8192, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.3082387264539862, |
|
"grad_norm": 0.8651610612869263, |
|
"learning_rate": 3.459757721824063e-05, |
|
"loss": 0.8755, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.3095071985793112, |
|
"grad_norm": 0.8182082772254944, |
|
"learning_rate": 3.4534153611974376e-05, |
|
"loss": 0.8661, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.31077567070463624, |
|
"grad_norm": 1.0455305576324463, |
|
"learning_rate": 3.4470730005708125e-05, |
|
"loss": 0.8134, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.3120441428299613, |
|
"grad_norm": 0.7838363647460938, |
|
"learning_rate": 3.4407306399441874e-05, |
|
"loss": 0.8531, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.31331261495528634, |
|
"grad_norm": 0.712868332862854, |
|
"learning_rate": 3.434388279317562e-05, |
|
"loss": 0.8109, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.3145810870806114, |
|
"grad_norm": 0.8088375329971313, |
|
"learning_rate": 3.428045918690937e-05, |
|
"loss": 0.8717, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.31584955920593644, |
|
"grad_norm": 0.8567003607749939, |
|
"learning_rate": 3.421703558064312e-05, |
|
"loss": 0.7997, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.3171180313312615, |
|
"grad_norm": 0.7826852202415466, |
|
"learning_rate": 3.4153611974376864e-05, |
|
"loss": 0.8071, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.31838650345658653, |
|
"grad_norm": 0.7257752418518066, |
|
"learning_rate": 3.4090188368110613e-05, |
|
"loss": 0.8053, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.3196549755819116, |
|
"grad_norm": 0.9757511019706726, |
|
"learning_rate": 3.402676476184436e-05, |
|
"loss": 0.8195, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.32092344770723663, |
|
"grad_norm": 0.9173797369003296, |
|
"learning_rate": 3.396334115557811e-05, |
|
"loss": 0.8662, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.3221919198325617, |
|
"grad_norm": 0.8483538627624512, |
|
"learning_rate": 3.3899917549311854e-05, |
|
"loss": 0.8339, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.32346039195788673, |
|
"grad_norm": 1.0109443664550781, |
|
"learning_rate": 3.3836493943045604e-05, |
|
"loss": 0.8731, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.3247288640832118, |
|
"grad_norm": 0.8521518707275391, |
|
"learning_rate": 3.377307033677935e-05, |
|
"loss": 0.8367, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.3259973362085368, |
|
"grad_norm": 0.8792763352394104, |
|
"learning_rate": 3.37096467305131e-05, |
|
"loss": 0.857, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.3272658083338619, |
|
"grad_norm": 0.7366037368774414, |
|
"learning_rate": 3.3646223124246844e-05, |
|
"loss": 0.8556, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.3285342804591869, |
|
"grad_norm": 0.9895220994949341, |
|
"learning_rate": 3.3582799517980594e-05, |
|
"loss": 0.885, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.32980275258451197, |
|
"grad_norm": 0.8205326199531555, |
|
"learning_rate": 3.351937591171434e-05, |
|
"loss": 0.7703, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.331071224709837, |
|
"grad_norm": 0.7155152559280396, |
|
"learning_rate": 3.345595230544809e-05, |
|
"loss": 0.8479, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.33233969683516207, |
|
"grad_norm": 0.7578288912773132, |
|
"learning_rate": 3.3392528699181834e-05, |
|
"loss": 0.8019, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.3336081689604871, |
|
"grad_norm": 0.8876450657844543, |
|
"learning_rate": 3.3329105092915584e-05, |
|
"loss": 0.8644, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.33487664108581217, |
|
"grad_norm": 0.8280789256095886, |
|
"learning_rate": 3.326568148664933e-05, |
|
"loss": 0.8358, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.33614511321113716, |
|
"grad_norm": 0.8563920855522156, |
|
"learning_rate": 3.320225788038308e-05, |
|
"loss": 0.8908, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.3374135853364622, |
|
"grad_norm": 0.8536137342453003, |
|
"learning_rate": 3.3138834274116824e-05, |
|
"loss": 0.8188, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.33868205746178726, |
|
"grad_norm": 0.909870982170105, |
|
"learning_rate": 3.3075410667850574e-05, |
|
"loss": 0.8175, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.3399505295871123, |
|
"grad_norm": 0.7365782260894775, |
|
"learning_rate": 3.301198706158432e-05, |
|
"loss": 0.8463, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.34121900171243735, |
|
"grad_norm": 0.7502683997154236, |
|
"learning_rate": 3.294856345531807e-05, |
|
"loss": 0.7959, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.3424874738377624, |
|
"grad_norm": 0.6690531373023987, |
|
"learning_rate": 3.288513984905182e-05, |
|
"loss": 0.8266, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.34375594596308745, |
|
"grad_norm": 0.9006823301315308, |
|
"learning_rate": 3.2821716242785564e-05, |
|
"loss": 0.8397, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.3450244180884125, |
|
"grad_norm": 0.826954185962677, |
|
"learning_rate": 3.275829263651932e-05, |
|
"loss": 0.8022, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.34629289021373755, |
|
"grad_norm": 0.8345193266868591, |
|
"learning_rate": 3.269486903025306e-05, |
|
"loss": 0.8148, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.3475613623390626, |
|
"grad_norm": 0.7929914593696594, |
|
"learning_rate": 3.263144542398681e-05, |
|
"loss": 0.8073, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.34882983446438764, |
|
"grad_norm": 0.9218955636024475, |
|
"learning_rate": 3.2568021817720554e-05, |
|
"loss": 0.8576, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.3500983065897127, |
|
"grad_norm": 0.8164006471633911, |
|
"learning_rate": 3.250459821145431e-05, |
|
"loss": 0.8434, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.35136677871503774, |
|
"grad_norm": 0.7511663436889648, |
|
"learning_rate": 3.244117460518805e-05, |
|
"loss": 0.7878, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.3526352508403628, |
|
"grad_norm": 0.8193197846412659, |
|
"learning_rate": 3.23777509989218e-05, |
|
"loss": 0.8608, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.35390372296568784, |
|
"grad_norm": 0.9112285375595093, |
|
"learning_rate": 3.2314327392655544e-05, |
|
"loss": 0.8404, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.3551721950910129, |
|
"grad_norm": 0.75201416015625, |
|
"learning_rate": 3.22509037863893e-05, |
|
"loss": 0.813, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.35644066721633794, |
|
"grad_norm": 0.8154107332229614, |
|
"learning_rate": 3.218748018012304e-05, |
|
"loss": 0.8083, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.357709139341663, |
|
"grad_norm": 0.7872757315635681, |
|
"learning_rate": 3.212405657385679e-05, |
|
"loss": 0.8358, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.35897761146698803, |
|
"grad_norm": 0.8861322999000549, |
|
"learning_rate": 3.206063296759054e-05, |
|
"loss": 0.8328, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.3602460835923131, |
|
"grad_norm": 0.7487745881080627, |
|
"learning_rate": 3.199720936132429e-05, |
|
"loss": 0.8626, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.36151455571763813, |
|
"grad_norm": 0.9134072065353394, |
|
"learning_rate": 3.193378575505803e-05, |
|
"loss": 0.8187, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.3627830278429631, |
|
"grad_norm": 0.8179683089256287, |
|
"learning_rate": 3.187036214879178e-05, |
|
"loss": 0.8223, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.3640514999682882, |
|
"grad_norm": 0.9501960873603821, |
|
"learning_rate": 3.180693854252553e-05, |
|
"loss": 0.8064, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.3653199720936132, |
|
"grad_norm": 0.8398934006690979, |
|
"learning_rate": 3.174351493625928e-05, |
|
"loss": 0.8279, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.36658844421893827, |
|
"grad_norm": 0.7718421816825867, |
|
"learning_rate": 3.168009132999302e-05, |
|
"loss": 0.8453, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.3678569163442633, |
|
"grad_norm": 0.7935000061988831, |
|
"learning_rate": 3.161666772372677e-05, |
|
"loss": 0.8292, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.36912538846958837, |
|
"grad_norm": 0.8383910655975342, |
|
"learning_rate": 3.155324411746052e-05, |
|
"loss": 0.8116, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.3703938605949134, |
|
"grad_norm": 0.7147135734558105, |
|
"learning_rate": 3.148982051119427e-05, |
|
"loss": 0.8146, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.37166233272023846, |
|
"grad_norm": 0.792220950126648, |
|
"learning_rate": 3.142639690492802e-05, |
|
"loss": 0.8278, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.3729308048455635, |
|
"grad_norm": 0.9010721445083618, |
|
"learning_rate": 3.136297329866176e-05, |
|
"loss": 0.8217, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.37419927697088856, |
|
"grad_norm": 0.7762110233306885, |
|
"learning_rate": 3.129954969239551e-05, |
|
"loss": 0.7813, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.3754677490962136, |
|
"grad_norm": 0.8827633261680603, |
|
"learning_rate": 3.123612608612926e-05, |
|
"loss": 0.8087, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.37673622122153866, |
|
"grad_norm": 0.7577320337295532, |
|
"learning_rate": 3.117270247986301e-05, |
|
"loss": 0.8198, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.3780046933468637, |
|
"grad_norm": 0.8777590990066528, |
|
"learning_rate": 3.110927887359675e-05, |
|
"loss": 0.8144, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.37927316547218876, |
|
"grad_norm": 0.8621464967727661, |
|
"learning_rate": 3.10458552673305e-05, |
|
"loss": 0.837, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.3805416375975138, |
|
"grad_norm": 0.842014491558075, |
|
"learning_rate": 3.098243166106425e-05, |
|
"loss": 0.7945, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.38181010972283885, |
|
"grad_norm": 0.8439661860466003, |
|
"learning_rate": 3.0919008054798e-05, |
|
"loss": 0.7907, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.3830785818481639, |
|
"grad_norm": 0.7277771830558777, |
|
"learning_rate": 3.085558444853174e-05, |
|
"loss": 0.7792, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.38434705397348895, |
|
"grad_norm": 0.8808379173278809, |
|
"learning_rate": 3.07921608422655e-05, |
|
"loss": 0.8356, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.385615526098814, |
|
"grad_norm": 0.7495381832122803, |
|
"learning_rate": 3.072873723599924e-05, |
|
"loss": 0.8252, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.38688399822413905, |
|
"grad_norm": 0.8698447942733765, |
|
"learning_rate": 3.066531362973299e-05, |
|
"loss": 0.8633, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.3881524703494641, |
|
"grad_norm": 0.8586138486862183, |
|
"learning_rate": 3.060189002346673e-05, |
|
"loss": 0.8542, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.3894209424747891, |
|
"grad_norm": 0.8375261425971985, |
|
"learning_rate": 3.053846641720049e-05, |
|
"loss": 0.8353, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.39068941460011414, |
|
"grad_norm": 0.723111093044281, |
|
"learning_rate": 3.047504281093423e-05, |
|
"loss": 0.8011, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.3919578867254392, |
|
"grad_norm": 0.7700281143188477, |
|
"learning_rate": 3.041161920466798e-05, |
|
"loss": 0.8309, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 0.39322635885076423, |
|
"grad_norm": 0.8734796643257141, |
|
"learning_rate": 3.0348195598401725e-05, |
|
"loss": 0.7967, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.3944948309760893, |
|
"grad_norm": 0.8601865172386169, |
|
"learning_rate": 3.0284771992135474e-05, |
|
"loss": 0.819, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 0.39576330310141433, |
|
"grad_norm": 0.8753730058670044, |
|
"learning_rate": 3.022134838586922e-05, |
|
"loss": 0.814, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.3970317752267394, |
|
"grad_norm": 0.8561723232269287, |
|
"learning_rate": 3.015792477960297e-05, |
|
"loss": 0.8073, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 0.39830024735206443, |
|
"grad_norm": 0.8156526684761047, |
|
"learning_rate": 3.009450117333672e-05, |
|
"loss": 0.8706, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 0.3995687194773895, |
|
"grad_norm": 0.7783714532852173, |
|
"learning_rate": 3.0031077567070464e-05, |
|
"loss": 0.7929, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.4008371916027145, |
|
"grad_norm": 0.7870660424232483, |
|
"learning_rate": 2.9967653960804216e-05, |
|
"loss": 0.8169, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.4021056637280396, |
|
"grad_norm": 0.9100777506828308, |
|
"learning_rate": 2.990423035453796e-05, |
|
"loss": 0.8365, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 0.4033741358533646, |
|
"grad_norm": 0.8847957849502563, |
|
"learning_rate": 2.984080674827171e-05, |
|
"loss": 0.8178, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 0.4046426079786897, |
|
"grad_norm": 0.9754513502120972, |
|
"learning_rate": 2.9777383142005454e-05, |
|
"loss": 0.8457, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 0.4059110801040147, |
|
"grad_norm": 0.8685561418533325, |
|
"learning_rate": 2.9713959535739206e-05, |
|
"loss": 0.823, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.40717955222933977, |
|
"grad_norm": 0.8493009209632874, |
|
"learning_rate": 2.965053592947295e-05, |
|
"loss": 0.8221, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 0.4084480243546648, |
|
"grad_norm": 0.954006016254425, |
|
"learning_rate": 2.95871123232067e-05, |
|
"loss": 0.811, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 0.40971649647998987, |
|
"grad_norm": 11.963293075561523, |
|
"learning_rate": 2.9523688716940444e-05, |
|
"loss": 0.8032, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 0.4109849686053149, |
|
"grad_norm": 0.9225347638130188, |
|
"learning_rate": 2.9460265110674196e-05, |
|
"loss": 0.8645, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 0.41225344073063996, |
|
"grad_norm": 0.7878516316413879, |
|
"learning_rate": 2.939684150440794e-05, |
|
"loss": 0.8303, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.413521912855965, |
|
"grad_norm": 0.815546989440918, |
|
"learning_rate": 2.933341789814169e-05, |
|
"loss": 0.7873, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 0.41479038498129006, |
|
"grad_norm": 0.7072951197624207, |
|
"learning_rate": 2.9269994291875437e-05, |
|
"loss": 0.8231, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 0.4160588571066151, |
|
"grad_norm": 0.7275887131690979, |
|
"learning_rate": 2.9206570685609186e-05, |
|
"loss": 0.8445, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 0.4173273292319401, |
|
"grad_norm": 0.7888057827949524, |
|
"learning_rate": 2.9143147079342932e-05, |
|
"loss": 0.7819, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 0.41859580135726515, |
|
"grad_norm": 0.8133323788642883, |
|
"learning_rate": 2.907972347307668e-05, |
|
"loss": 0.7918, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.4198642734825902, |
|
"grad_norm": 0.9344608187675476, |
|
"learning_rate": 2.9016299866810427e-05, |
|
"loss": 0.805, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 0.42113274560791525, |
|
"grad_norm": 0.9130797386169434, |
|
"learning_rate": 2.8952876260544176e-05, |
|
"loss": 0.808, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.4224012177332403, |
|
"grad_norm": 0.8647783398628235, |
|
"learning_rate": 2.8889452654277922e-05, |
|
"loss": 0.8027, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 0.42366968985856535, |
|
"grad_norm": 0.8025421500205994, |
|
"learning_rate": 2.882602904801167e-05, |
|
"loss": 0.811, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 0.4249381619838904, |
|
"grad_norm": 0.7249786853790283, |
|
"learning_rate": 2.876260544174542e-05, |
|
"loss": 0.7597, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.42620663410921544, |
|
"grad_norm": 0.7876623868942261, |
|
"learning_rate": 2.8699181835479166e-05, |
|
"loss": 0.8297, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 0.4274751062345405, |
|
"grad_norm": 0.8284019827842712, |
|
"learning_rate": 2.8635758229212916e-05, |
|
"loss": 0.7985, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 0.42874357835986554, |
|
"grad_norm": 0.7855024337768555, |
|
"learning_rate": 2.857233462294666e-05, |
|
"loss": 0.8414, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 0.4300120504851906, |
|
"grad_norm": 0.789946436882019, |
|
"learning_rate": 2.850891101668041e-05, |
|
"loss": 0.8307, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 0.43128052261051564, |
|
"grad_norm": 0.7722301483154297, |
|
"learning_rate": 2.8445487410414156e-05, |
|
"loss": 0.8101, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.4325489947358407, |
|
"grad_norm": 0.8557891249656677, |
|
"learning_rate": 2.8382063804147906e-05, |
|
"loss": 0.846, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.43381746686116573, |
|
"grad_norm": 0.8216169476509094, |
|
"learning_rate": 2.831864019788165e-05, |
|
"loss": 0.7966, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 0.4350859389864908, |
|
"grad_norm": 0.87419593334198, |
|
"learning_rate": 2.82552165916154e-05, |
|
"loss": 0.7777, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 0.43635441111181583, |
|
"grad_norm": 0.8956803679466248, |
|
"learning_rate": 2.8191792985349146e-05, |
|
"loss": 0.8066, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 0.4376228832371409, |
|
"grad_norm": 0.8264901638031006, |
|
"learning_rate": 2.8128369379082896e-05, |
|
"loss": 0.8489, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.43889135536246593, |
|
"grad_norm": 0.7960401773452759, |
|
"learning_rate": 2.806494577281664e-05, |
|
"loss": 0.8257, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 0.440159827487791, |
|
"grad_norm": 0.7691190838813782, |
|
"learning_rate": 2.8001522166550394e-05, |
|
"loss": 0.8365, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 0.441428299613116, |
|
"grad_norm": 0.7433714270591736, |
|
"learning_rate": 2.7938098560284136e-05, |
|
"loss": 0.7755, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 0.4426967717384411, |
|
"grad_norm": 0.7270233035087585, |
|
"learning_rate": 2.787467495401789e-05, |
|
"loss": 0.808, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 0.44396524386376607, |
|
"grad_norm": 0.7907856106758118, |
|
"learning_rate": 2.781125134775163e-05, |
|
"loss": 0.8038, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.4452337159890911, |
|
"grad_norm": 0.7421363592147827, |
|
"learning_rate": 2.7747827741485384e-05, |
|
"loss": 0.8058, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 0.44650218811441617, |
|
"grad_norm": 0.8635361790657043, |
|
"learning_rate": 2.7684404135219126e-05, |
|
"loss": 0.786, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 0.4477706602397412, |
|
"grad_norm": 0.9545580744743347, |
|
"learning_rate": 2.762098052895288e-05, |
|
"loss": 0.8431, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 0.44903913236506626, |
|
"grad_norm": 0.9529020190238953, |
|
"learning_rate": 2.755755692268662e-05, |
|
"loss": 0.8082, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 0.4503076044903913, |
|
"grad_norm": 0.8344403505325317, |
|
"learning_rate": 2.7494133316420374e-05, |
|
"loss": 0.8031, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.45157607661571636, |
|
"grad_norm": 0.8490266799926758, |
|
"learning_rate": 2.7430709710154123e-05, |
|
"loss": 0.7952, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 0.4528445487410414, |
|
"grad_norm": 0.8201053738594055, |
|
"learning_rate": 2.736728610388787e-05, |
|
"loss": 0.8057, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 0.45411302086636646, |
|
"grad_norm": 0.7206814289093018, |
|
"learning_rate": 2.7303862497621618e-05, |
|
"loss": 0.7907, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 0.4553814929916915, |
|
"grad_norm": 0.7151837944984436, |
|
"learning_rate": 2.7240438891355364e-05, |
|
"loss": 0.7989, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 0.45664996511701655, |
|
"grad_norm": 0.73557049036026, |
|
"learning_rate": 2.7177015285089113e-05, |
|
"loss": 0.7887, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.4579184372423416, |
|
"grad_norm": 0.8012831807136536, |
|
"learning_rate": 2.711359167882286e-05, |
|
"loss": 0.7963, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 0.45918690936766665, |
|
"grad_norm": 0.7544090747833252, |
|
"learning_rate": 2.7050168072556608e-05, |
|
"loss": 0.8021, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 0.4604553814929917, |
|
"grad_norm": 0.8027962446212769, |
|
"learning_rate": 2.6986744466290354e-05, |
|
"loss": 0.8346, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 0.46172385361831675, |
|
"grad_norm": 0.8969400525093079, |
|
"learning_rate": 2.6923320860024103e-05, |
|
"loss": 0.8362, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 0.4629923257436418, |
|
"grad_norm": 0.8505738377571106, |
|
"learning_rate": 2.685989725375785e-05, |
|
"loss": 0.8253, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.46426079786896685, |
|
"grad_norm": 0.8005324602127075, |
|
"learning_rate": 2.6796473647491598e-05, |
|
"loss": 0.8133, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 0.4655292699942919, |
|
"grad_norm": 0.8265887498855591, |
|
"learning_rate": 2.6733050041225344e-05, |
|
"loss": 0.7974, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 0.46679774211961694, |
|
"grad_norm": 0.8310080170631409, |
|
"learning_rate": 2.6669626434959093e-05, |
|
"loss": 0.8311, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 0.468066214244942, |
|
"grad_norm": 0.8837007284164429, |
|
"learning_rate": 2.660620282869284e-05, |
|
"loss": 0.8086, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 0.46933468637026704, |
|
"grad_norm": 0.8574205040931702, |
|
"learning_rate": 2.6542779222426588e-05, |
|
"loss": 0.8045, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.47060315849559203, |
|
"grad_norm": 0.7924466729164124, |
|
"learning_rate": 2.6479355616160334e-05, |
|
"loss": 0.831, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 0.4718716306209171, |
|
"grad_norm": 0.8812252283096313, |
|
"learning_rate": 2.6415932009894083e-05, |
|
"loss": 0.8414, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 0.47314010274624213, |
|
"grad_norm": 0.8879112601280212, |
|
"learning_rate": 2.635250840362783e-05, |
|
"loss": 0.8037, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 0.4744085748715672, |
|
"grad_norm": 0.8532351851463318, |
|
"learning_rate": 2.6289084797361578e-05, |
|
"loss": 0.7961, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 0.4756770469968922, |
|
"grad_norm": 0.7800135016441345, |
|
"learning_rate": 2.6225661191095324e-05, |
|
"loss": 0.8158, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.4769455191222173, |
|
"grad_norm": 0.8264251947402954, |
|
"learning_rate": 2.6162237584829073e-05, |
|
"loss": 0.7902, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 0.4782139912475423, |
|
"grad_norm": 0.9216287136077881, |
|
"learning_rate": 2.6098813978562826e-05, |
|
"loss": 0.8387, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 0.4794824633728674, |
|
"grad_norm": 0.8331848382949829, |
|
"learning_rate": 2.6035390372296568e-05, |
|
"loss": 0.8309, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 0.4807509354981924, |
|
"grad_norm": 0.7791485786437988, |
|
"learning_rate": 2.597196676603032e-05, |
|
"loss": 0.7954, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 0.48201940762351747, |
|
"grad_norm": 0.8223782777786255, |
|
"learning_rate": 2.5908543159764063e-05, |
|
"loss": 0.8012, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.4832878797488425, |
|
"grad_norm": 0.7362112998962402, |
|
"learning_rate": 2.5845119553497816e-05, |
|
"loss": 0.8548, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 0.48455635187416757, |
|
"grad_norm": 0.9084497094154358, |
|
"learning_rate": 2.578169594723156e-05, |
|
"loss": 0.7889, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 0.4858248239994926, |
|
"grad_norm": 0.8268200755119324, |
|
"learning_rate": 2.571827234096531e-05, |
|
"loss": 0.778, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 0.48709329612481767, |
|
"grad_norm": 0.683861494064331, |
|
"learning_rate": 2.5654848734699057e-05, |
|
"loss": 0.8079, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 0.4883617682501427, |
|
"grad_norm": 0.9449877142906189, |
|
"learning_rate": 2.5591425128432806e-05, |
|
"loss": 0.7967, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.48963024037546776, |
|
"grad_norm": 0.8445014357566833, |
|
"learning_rate": 2.552800152216655e-05, |
|
"loss": 0.7774, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 0.4908987125007928, |
|
"grad_norm": 0.811717689037323, |
|
"learning_rate": 2.54645779159003e-05, |
|
"loss": 0.7659, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 0.49216718462611786, |
|
"grad_norm": 0.7867732644081116, |
|
"learning_rate": 2.5401154309634047e-05, |
|
"loss": 0.8539, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 0.4934356567514429, |
|
"grad_norm": 0.7808852791786194, |
|
"learning_rate": 2.5337730703367796e-05, |
|
"loss": 0.8294, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 0.49470412887676796, |
|
"grad_norm": 0.8998913168907166, |
|
"learning_rate": 2.527430709710154e-05, |
|
"loss": 0.8507, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.495972601002093, |
|
"grad_norm": 0.7990160584449768, |
|
"learning_rate": 2.521088349083529e-05, |
|
"loss": 0.7261, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 0.497241073127418, |
|
"grad_norm": 0.7361629605293274, |
|
"learning_rate": 2.5147459884569037e-05, |
|
"loss": 0.8058, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 0.49850954525274305, |
|
"grad_norm": 0.849870502948761, |
|
"learning_rate": 2.5084036278302786e-05, |
|
"loss": 0.7708, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 0.4997780173780681, |
|
"grad_norm": 0.7667344808578491, |
|
"learning_rate": 2.502061267203653e-05, |
|
"loss": 0.818, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 0.5010464895033931, |
|
"grad_norm": 0.800609827041626, |
|
"learning_rate": 2.495718906577028e-05, |
|
"loss": 0.8248, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.5023149616287182, |
|
"grad_norm": 0.7429226040840149, |
|
"learning_rate": 2.489376545950403e-05, |
|
"loss": 0.8395, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 0.5035834337540432, |
|
"grad_norm": 0.7970502972602844, |
|
"learning_rate": 2.4830341853237776e-05, |
|
"loss": 0.8182, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 0.5048519058793683, |
|
"grad_norm": 1.7285821437835693, |
|
"learning_rate": 2.4766918246971525e-05, |
|
"loss": 0.8241, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 0.5061203780046933, |
|
"grad_norm": 0.8314895629882812, |
|
"learning_rate": 2.470349464070527e-05, |
|
"loss": 0.8017, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 0.5073888501300184, |
|
"grad_norm": 0.9516363143920898, |
|
"learning_rate": 2.464007103443902e-05, |
|
"loss": 0.8395, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.5086573222553434, |
|
"grad_norm": 0.8164798617362976, |
|
"learning_rate": 2.4576647428172766e-05, |
|
"loss": 0.8033, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 0.5099257943806685, |
|
"grad_norm": 0.7700650691986084, |
|
"learning_rate": 2.4513223821906515e-05, |
|
"loss": 0.779, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 0.5111942665059935, |
|
"grad_norm": 0.8437737226486206, |
|
"learning_rate": 2.444980021564026e-05, |
|
"loss": 0.8112, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 0.5124627386313185, |
|
"grad_norm": 0.8371322751045227, |
|
"learning_rate": 2.438637660937401e-05, |
|
"loss": 0.8152, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 0.5137312107566436, |
|
"grad_norm": 0.8382763862609863, |
|
"learning_rate": 2.4322953003107756e-05, |
|
"loss": 0.8414, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.5149996828819686, |
|
"grad_norm": 0.9525557160377502, |
|
"learning_rate": 2.4259529396841505e-05, |
|
"loss": 0.8091, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 0.5162681550072937, |
|
"grad_norm": 0.7620564699172974, |
|
"learning_rate": 2.419610579057525e-05, |
|
"loss": 0.7981, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 0.5175366271326187, |
|
"grad_norm": 0.8722305297851562, |
|
"learning_rate": 2.4132682184309003e-05, |
|
"loss": 0.8079, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 0.5188050992579438, |
|
"grad_norm": 0.8774722218513489, |
|
"learning_rate": 2.406925857804275e-05, |
|
"loss": 0.7937, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 0.5200735713832688, |
|
"grad_norm": 0.7515254616737366, |
|
"learning_rate": 2.40058349717765e-05, |
|
"loss": 0.8134, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.5213420435085939, |
|
"grad_norm": 0.8385280966758728, |
|
"learning_rate": 2.3942411365510244e-05, |
|
"loss": 0.8064, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 0.5226105156339189, |
|
"grad_norm": 0.8530700206756592, |
|
"learning_rate": 2.3878987759243994e-05, |
|
"loss": 0.8249, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 0.523878987759244, |
|
"grad_norm": 0.7866977453231812, |
|
"learning_rate": 2.381556415297774e-05, |
|
"loss": 0.803, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 0.525147459884569, |
|
"grad_norm": 0.8509036302566528, |
|
"learning_rate": 2.375214054671149e-05, |
|
"loss": 0.8307, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 0.5264159320098941, |
|
"grad_norm": 0.8268348574638367, |
|
"learning_rate": 2.3688716940445234e-05, |
|
"loss": 0.8205, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.5276844041352191, |
|
"grad_norm": 0.9176819920539856, |
|
"learning_rate": 2.3625293334178984e-05, |
|
"loss": 0.7774, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 0.5289528762605442, |
|
"grad_norm": 0.758176326751709, |
|
"learning_rate": 2.356186972791273e-05, |
|
"loss": 0.8067, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 0.5302213483858692, |
|
"grad_norm": 0.7369076609611511, |
|
"learning_rate": 2.349844612164648e-05, |
|
"loss": 0.7673, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 0.5314898205111943, |
|
"grad_norm": 0.8413040041923523, |
|
"learning_rate": 2.3435022515380224e-05, |
|
"loss": 0.8289, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 0.5327582926365193, |
|
"grad_norm": 0.8975269794464111, |
|
"learning_rate": 2.3371598909113974e-05, |
|
"loss": 0.8097, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.5340267647618444, |
|
"grad_norm": 0.8501763343811035, |
|
"learning_rate": 2.330817530284772e-05, |
|
"loss": 0.7867, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 0.5352952368871694, |
|
"grad_norm": 0.9364180564880371, |
|
"learning_rate": 2.324475169658147e-05, |
|
"loss": 0.7925, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 0.5365637090124945, |
|
"grad_norm": 0.6347882151603699, |
|
"learning_rate": 2.3181328090315214e-05, |
|
"loss": 0.7635, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 0.5378321811378195, |
|
"grad_norm": 0.8539864420890808, |
|
"learning_rate": 2.3117904484048964e-05, |
|
"loss": 0.7993, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 0.5391006532631445, |
|
"grad_norm": 0.8893634080886841, |
|
"learning_rate": 2.305448087778271e-05, |
|
"loss": 0.7809, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.5403691253884696, |
|
"grad_norm": 0.7993662357330322, |
|
"learning_rate": 2.299105727151646e-05, |
|
"loss": 0.8639, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 0.5416375975137946, |
|
"grad_norm": 0.8157054781913757, |
|
"learning_rate": 2.2927633665250208e-05, |
|
"loss": 0.8121, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 0.5429060696391197, |
|
"grad_norm": 0.8141036033630371, |
|
"learning_rate": 2.2864210058983954e-05, |
|
"loss": 0.7748, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 0.5441745417644447, |
|
"grad_norm": 0.8311188220977783, |
|
"learning_rate": 2.2803957633031015e-05, |
|
"loss": 0.8182, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 0.5454430138897698, |
|
"grad_norm": 0.8923128247261047, |
|
"learning_rate": 2.274053402676476e-05, |
|
"loss": 0.7841, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.5467114860150948, |
|
"grad_norm": 0.8246520757675171, |
|
"learning_rate": 2.267711042049851e-05, |
|
"loss": 0.8146, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 0.5479799581404199, |
|
"grad_norm": 0.8469933271408081, |
|
"learning_rate": 2.261368681423226e-05, |
|
"loss": 0.817, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 0.5492484302657449, |
|
"grad_norm": 0.8211717009544373, |
|
"learning_rate": 2.2550263207966005e-05, |
|
"loss": 0.8082, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 0.55051690239107, |
|
"grad_norm": 0.9137957692146301, |
|
"learning_rate": 2.2486839601699754e-05, |
|
"loss": 0.7691, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 0.551785374516395, |
|
"grad_norm": 0.8431654572486877, |
|
"learning_rate": 2.2423415995433503e-05, |
|
"loss": 0.812, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.5530538466417201, |
|
"grad_norm": 0.9029563069343567, |
|
"learning_rate": 2.235999238916725e-05, |
|
"loss": 0.8667, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 0.5543223187670451, |
|
"grad_norm": 0.8180502653121948, |
|
"learning_rate": 2.2296568782900998e-05, |
|
"loss": 0.8171, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 0.5555907908923702, |
|
"grad_norm": 0.844530463218689, |
|
"learning_rate": 2.2233145176634744e-05, |
|
"loss": 0.7784, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 0.5568592630176952, |
|
"grad_norm": 0.7153404355049133, |
|
"learning_rate": 2.2169721570368493e-05, |
|
"loss": 0.7891, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 0.5581277351430203, |
|
"grad_norm": 0.7020410895347595, |
|
"learning_rate": 2.2106297964102242e-05, |
|
"loss": 0.8042, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.5593962072683453, |
|
"grad_norm": 0.7581042647361755, |
|
"learning_rate": 2.2042874357835988e-05, |
|
"loss": 0.8337, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 0.5606646793936704, |
|
"grad_norm": 0.7392009496688843, |
|
"learning_rate": 2.1979450751569737e-05, |
|
"loss": 0.8151, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 0.5619331515189954, |
|
"grad_norm": 0.8381578326225281, |
|
"learning_rate": 2.1916027145303483e-05, |
|
"loss": 0.7923, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 0.5632016236443205, |
|
"grad_norm": 1.0505058765411377, |
|
"learning_rate": 2.1852603539037232e-05, |
|
"loss": 0.8117, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 0.5644700957696455, |
|
"grad_norm": 0.67955082654953, |
|
"learning_rate": 2.1789179932770978e-05, |
|
"loss": 0.7921, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.5657385678949705, |
|
"grad_norm": 0.798687219619751, |
|
"learning_rate": 2.1725756326504727e-05, |
|
"loss": 0.8038, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 0.5670070400202956, |
|
"grad_norm": 0.989431619644165, |
|
"learning_rate": 2.1662332720238473e-05, |
|
"loss": 0.8317, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 0.5682755121456206, |
|
"grad_norm": 0.8161944150924683, |
|
"learning_rate": 2.1598909113972222e-05, |
|
"loss": 0.8222, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 0.5695439842709457, |
|
"grad_norm": 0.8795542120933533, |
|
"learning_rate": 2.1535485507705968e-05, |
|
"loss": 0.7717, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 0.5708124563962707, |
|
"grad_norm": 0.7453576326370239, |
|
"learning_rate": 2.1472061901439717e-05, |
|
"loss": 0.805, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.5720809285215958, |
|
"grad_norm": 0.8081907033920288, |
|
"learning_rate": 2.1408638295173463e-05, |
|
"loss": 0.7757, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 0.5733494006469207, |
|
"grad_norm": 0.7817357778549194, |
|
"learning_rate": 2.1345214688907212e-05, |
|
"loss": 0.7528, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 0.5746178727722459, |
|
"grad_norm": 0.8645827770233154, |
|
"learning_rate": 2.1281791082640958e-05, |
|
"loss": 0.7713, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 0.5758863448975708, |
|
"grad_norm": 0.8567843437194824, |
|
"learning_rate": 2.1218367476374707e-05, |
|
"loss": 0.7447, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 0.577154817022896, |
|
"grad_norm": 0.7494439482688904, |
|
"learning_rate": 2.1154943870108453e-05, |
|
"loss": 0.7661, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.5784232891482209, |
|
"grad_norm": 0.8079215884208679, |
|
"learning_rate": 2.1091520263842206e-05, |
|
"loss": 0.8119, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 0.579691761273546, |
|
"grad_norm": 0.9019980430603027, |
|
"learning_rate": 2.102809665757595e-05, |
|
"loss": 0.8125, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 0.580960233398871, |
|
"grad_norm": 0.7452351450920105, |
|
"learning_rate": 2.09646730513097e-05, |
|
"loss": 0.7691, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 0.5822287055241961, |
|
"grad_norm": 0.7727750539779663, |
|
"learning_rate": 2.0901249445043446e-05, |
|
"loss": 0.7925, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 0.5834971776495211, |
|
"grad_norm": 0.7486307621002197, |
|
"learning_rate": 2.0837825838777196e-05, |
|
"loss": 0.7668, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.5847656497748462, |
|
"grad_norm": 0.8719222545623779, |
|
"learning_rate": 2.077440223251094e-05, |
|
"loss": 0.8196, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 0.5860341219001712, |
|
"grad_norm": 0.7641133069992065, |
|
"learning_rate": 2.071097862624469e-05, |
|
"loss": 0.7559, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 0.5873025940254963, |
|
"grad_norm": 0.8036416172981262, |
|
"learning_rate": 2.0647555019978436e-05, |
|
"loss": 0.8285, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 0.5885710661508213, |
|
"grad_norm": 0.8614276051521301, |
|
"learning_rate": 2.0584131413712186e-05, |
|
"loss": 0.7951, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 0.5898395382761464, |
|
"grad_norm": 0.8406545519828796, |
|
"learning_rate": 2.052070780744593e-05, |
|
"loss": 0.7714, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.5911080104014714, |
|
"grad_norm": 0.9403005838394165, |
|
"learning_rate": 2.045728420117968e-05, |
|
"loss": 0.7999, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 0.5923764825267964, |
|
"grad_norm": 0.8395708799362183, |
|
"learning_rate": 2.0393860594913426e-05, |
|
"loss": 0.8069, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 0.5936449546521215, |
|
"grad_norm": 0.8432602286338806, |
|
"learning_rate": 2.0330436988647176e-05, |
|
"loss": 0.8189, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 0.5949134267774465, |
|
"grad_norm": 0.7362537980079651, |
|
"learning_rate": 2.026701338238092e-05, |
|
"loss": 0.8069, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 0.5961818989027716, |
|
"grad_norm": 0.7601738572120667, |
|
"learning_rate": 2.020358977611467e-05, |
|
"loss": 0.7449, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.5974503710280966, |
|
"grad_norm": 0.8012720346450806, |
|
"learning_rate": 2.0140166169848416e-05, |
|
"loss": 0.7831, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 0.5987188431534217, |
|
"grad_norm": 0.7670310139656067, |
|
"learning_rate": 2.0076742563582166e-05, |
|
"loss": 0.8393, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 0.5999873152787467, |
|
"grad_norm": 0.8244422674179077, |
|
"learning_rate": 2.001331895731591e-05, |
|
"loss": 0.7709, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 0.6012557874040718, |
|
"grad_norm": 0.7943612933158875, |
|
"learning_rate": 1.994989535104966e-05, |
|
"loss": 0.819, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 0.6025242595293968, |
|
"grad_norm": 0.9540635347366333, |
|
"learning_rate": 1.9886471744783406e-05, |
|
"loss": 0.7899, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.6037927316547219, |
|
"grad_norm": 0.9198821783065796, |
|
"learning_rate": 1.9823048138517156e-05, |
|
"loss": 0.8239, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 0.6050612037800469, |
|
"grad_norm": 0.7837796807289124, |
|
"learning_rate": 1.9759624532250905e-05, |
|
"loss": 0.7929, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 0.606329675905372, |
|
"grad_norm": 0.8205187320709229, |
|
"learning_rate": 1.9696200925984654e-05, |
|
"loss": 0.819, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 0.607598148030697, |
|
"grad_norm": 0.8532772064208984, |
|
"learning_rate": 1.96327773197184e-05, |
|
"loss": 0.7455, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 0.6088666201560221, |
|
"grad_norm": 0.8524623513221741, |
|
"learning_rate": 1.956935371345215e-05, |
|
"loss": 0.8089, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.6101350922813471, |
|
"grad_norm": 0.8614479899406433, |
|
"learning_rate": 1.9505930107185895e-05, |
|
"loss": 0.8059, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 0.6114035644066722, |
|
"grad_norm": 0.7598078846931458, |
|
"learning_rate": 1.9442506500919644e-05, |
|
"loss": 0.7832, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 0.6126720365319972, |
|
"grad_norm": 0.809009850025177, |
|
"learning_rate": 1.937908289465339e-05, |
|
"loss": 0.7259, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 0.6139405086573223, |
|
"grad_norm": 0.7381779551506042, |
|
"learning_rate": 1.931565928838714e-05, |
|
"loss": 0.7676, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 0.6152089807826473, |
|
"grad_norm": 0.8887180685997009, |
|
"learning_rate": 1.9252235682120888e-05, |
|
"loss": 0.8126, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.6164774529079724, |
|
"grad_norm": 0.7270573973655701, |
|
"learning_rate": 1.9188812075854634e-05, |
|
"loss": 0.7786, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 0.6177459250332974, |
|
"grad_norm": 0.7978057861328125, |
|
"learning_rate": 1.9125388469588383e-05, |
|
"loss": 0.8281, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 0.6190143971586224, |
|
"grad_norm": 0.8202372789382935, |
|
"learning_rate": 1.906196486332213e-05, |
|
"loss": 0.7656, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 0.6202828692839475, |
|
"grad_norm": 0.9720300436019897, |
|
"learning_rate": 1.8998541257055878e-05, |
|
"loss": 0.7881, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 0.6215513414092725, |
|
"grad_norm": 0.9297833442687988, |
|
"learning_rate": 1.8935117650789624e-05, |
|
"loss": 0.8237, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.6228198135345976, |
|
"grad_norm": 0.7593715190887451, |
|
"learning_rate": 1.8871694044523373e-05, |
|
"loss": 0.7574, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 0.6240882856599226, |
|
"grad_norm": 0.8537524938583374, |
|
"learning_rate": 1.880827043825712e-05, |
|
"loss": 0.7969, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 0.6253567577852477, |
|
"grad_norm": 0.770918607711792, |
|
"learning_rate": 1.8744846831990868e-05, |
|
"loss": 0.7894, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 0.6266252299105727, |
|
"grad_norm": 0.7605695724487305, |
|
"learning_rate": 1.8681423225724614e-05, |
|
"loss": 0.782, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 0.6278937020358978, |
|
"grad_norm": 0.8978208303451538, |
|
"learning_rate": 1.8617999619458363e-05, |
|
"loss": 0.7979, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.6291621741612228, |
|
"grad_norm": 0.7393850088119507, |
|
"learning_rate": 1.855457601319211e-05, |
|
"loss": 0.8139, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 0.6304306462865479, |
|
"grad_norm": 0.7255131602287292, |
|
"learning_rate": 1.8491152406925858e-05, |
|
"loss": 0.7802, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 0.6316991184118729, |
|
"grad_norm": 0.7080028653144836, |
|
"learning_rate": 1.8427728800659607e-05, |
|
"loss": 0.8059, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 0.632967590537198, |
|
"grad_norm": 0.8282076716423035, |
|
"learning_rate": 1.8364305194393357e-05, |
|
"loss": 0.7905, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 0.634236062662523, |
|
"grad_norm": 0.8741589784622192, |
|
"learning_rate": 1.8300881588127102e-05, |
|
"loss": 0.8174, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.6355045347878481, |
|
"grad_norm": 0.7435175776481628, |
|
"learning_rate": 1.823745798186085e-05, |
|
"loss": 0.7678, |
|
"step": 10020 |
|
}, |
|
{ |
|
"epoch": 0.6367730069131731, |
|
"grad_norm": 0.7347603440284729, |
|
"learning_rate": 1.817720555590791e-05, |
|
"loss": 0.8225, |
|
"step": 10040 |
|
}, |
|
{ |
|
"epoch": 0.6380414790384982, |
|
"grad_norm": 0.8974965214729309, |
|
"learning_rate": 1.811378194964166e-05, |
|
"loss": 0.7766, |
|
"step": 10060 |
|
}, |
|
{ |
|
"epoch": 0.6393099511638232, |
|
"grad_norm": 0.7255268692970276, |
|
"learning_rate": 1.8050358343375408e-05, |
|
"loss": 0.8, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 0.6405784232891483, |
|
"grad_norm": 0.7062020897865295, |
|
"learning_rate": 1.7986934737109154e-05, |
|
"loss": 0.7088, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.6418468954144733, |
|
"grad_norm": 0.8076253533363342, |
|
"learning_rate": 1.7923511130842903e-05, |
|
"loss": 0.8152, |
|
"step": 10120 |
|
}, |
|
{ |
|
"epoch": 0.6431153675397984, |
|
"grad_norm": 0.8340699672698975, |
|
"learning_rate": 1.786008752457665e-05, |
|
"loss": 0.7985, |
|
"step": 10140 |
|
}, |
|
{ |
|
"epoch": 0.6443838396651234, |
|
"grad_norm": 0.7522137761116028, |
|
"learning_rate": 1.7796663918310398e-05, |
|
"loss": 0.7704, |
|
"step": 10160 |
|
}, |
|
{ |
|
"epoch": 0.6456523117904484, |
|
"grad_norm": 0.8227932453155518, |
|
"learning_rate": 1.7733240312044144e-05, |
|
"loss": 0.828, |
|
"step": 10180 |
|
}, |
|
{ |
|
"epoch": 0.6469207839157735, |
|
"grad_norm": 0.7742383480072021, |
|
"learning_rate": 1.7669816705777893e-05, |
|
"loss": 0.7893, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.6481892560410984, |
|
"grad_norm": 0.7038094401359558, |
|
"learning_rate": 1.760639309951164e-05, |
|
"loss": 0.7699, |
|
"step": 10220 |
|
}, |
|
{ |
|
"epoch": 0.6494577281664236, |
|
"grad_norm": 0.8382614850997925, |
|
"learning_rate": 1.7542969493245388e-05, |
|
"loss": 0.8254, |
|
"step": 10240 |
|
}, |
|
{ |
|
"epoch": 0.6507262002917485, |
|
"grad_norm": 0.9173989295959473, |
|
"learning_rate": 1.7479545886979134e-05, |
|
"loss": 0.7603, |
|
"step": 10260 |
|
}, |
|
{ |
|
"epoch": 0.6519946724170737, |
|
"grad_norm": 0.7602284550666809, |
|
"learning_rate": 1.7416122280712883e-05, |
|
"loss": 0.8019, |
|
"step": 10280 |
|
}, |
|
{ |
|
"epoch": 0.6532631445423986, |
|
"grad_norm": 0.8012353181838989, |
|
"learning_rate": 1.735269867444663e-05, |
|
"loss": 0.7944, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.6545316166677237, |
|
"grad_norm": 0.8844314217567444, |
|
"learning_rate": 1.7289275068180378e-05, |
|
"loss": 0.7785, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 0.6558000887930487, |
|
"grad_norm": 0.7556779980659485, |
|
"learning_rate": 1.7225851461914124e-05, |
|
"loss": 0.7597, |
|
"step": 10340 |
|
}, |
|
{ |
|
"epoch": 0.6570685609183738, |
|
"grad_norm": 0.8446857929229736, |
|
"learning_rate": 1.7162427855647873e-05, |
|
"loss": 0.7941, |
|
"step": 10360 |
|
}, |
|
{ |
|
"epoch": 0.6583370330436988, |
|
"grad_norm": 0.7313318848609924, |
|
"learning_rate": 1.709900424938162e-05, |
|
"loss": 0.8017, |
|
"step": 10380 |
|
}, |
|
{ |
|
"epoch": 0.6596055051690239, |
|
"grad_norm": 0.8298467397689819, |
|
"learning_rate": 1.7035580643115368e-05, |
|
"loss": 0.7869, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.6608739772943489, |
|
"grad_norm": 0.8003538846969604, |
|
"learning_rate": 1.6972157036849114e-05, |
|
"loss": 0.8002, |
|
"step": 10420 |
|
}, |
|
{ |
|
"epoch": 0.662142449419674, |
|
"grad_norm": 0.7555122971534729, |
|
"learning_rate": 1.6908733430582863e-05, |
|
"loss": 0.7632, |
|
"step": 10440 |
|
}, |
|
{ |
|
"epoch": 0.663410921544999, |
|
"grad_norm": 0.7712675333023071, |
|
"learning_rate": 1.684530982431661e-05, |
|
"loss": 0.7441, |
|
"step": 10460 |
|
}, |
|
{ |
|
"epoch": 0.6646793936703241, |
|
"grad_norm": 0.6845158338546753, |
|
"learning_rate": 1.6781886218050358e-05, |
|
"loss": 0.7384, |
|
"step": 10480 |
|
}, |
|
{ |
|
"epoch": 0.6659478657956491, |
|
"grad_norm": 0.8500059843063354, |
|
"learning_rate": 1.6718462611784107e-05, |
|
"loss": 0.8052, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.6672163379209742, |
|
"grad_norm": 0.759861946105957, |
|
"learning_rate": 1.6655039005517856e-05, |
|
"loss": 0.828, |
|
"step": 10520 |
|
}, |
|
{ |
|
"epoch": 0.6684848100462992, |
|
"grad_norm": 0.7759114503860474, |
|
"learning_rate": 1.6591615399251602e-05, |
|
"loss": 0.7719, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 0.6697532821716243, |
|
"grad_norm": 0.8368454575538635, |
|
"learning_rate": 1.652819179298535e-05, |
|
"loss": 0.8383, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 0.6710217542969493, |
|
"grad_norm": 0.8691524267196655, |
|
"learning_rate": 1.6464768186719097e-05, |
|
"loss": 0.7822, |
|
"step": 10580 |
|
}, |
|
{ |
|
"epoch": 0.6722902264222743, |
|
"grad_norm": 0.8464477062225342, |
|
"learning_rate": 1.6401344580452846e-05, |
|
"loss": 0.7802, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.6735586985475994, |
|
"grad_norm": 0.796231210231781, |
|
"learning_rate": 1.6337920974186592e-05, |
|
"loss": 0.7821, |
|
"step": 10620 |
|
}, |
|
{ |
|
"epoch": 0.6748271706729244, |
|
"grad_norm": 0.7409220933914185, |
|
"learning_rate": 1.627449736792034e-05, |
|
"loss": 0.8203, |
|
"step": 10640 |
|
}, |
|
{ |
|
"epoch": 0.6760956427982495, |
|
"grad_norm": 0.7760050892829895, |
|
"learning_rate": 1.6211073761654087e-05, |
|
"loss": 0.7673, |
|
"step": 10660 |
|
}, |
|
{ |
|
"epoch": 0.6773641149235745, |
|
"grad_norm": 0.7795297503471375, |
|
"learning_rate": 1.6147650155387836e-05, |
|
"loss": 0.7545, |
|
"step": 10680 |
|
}, |
|
{ |
|
"epoch": 0.6786325870488996, |
|
"grad_norm": 0.8562922477722168, |
|
"learning_rate": 1.6084226549121585e-05, |
|
"loss": 0.7744, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.6799010591742246, |
|
"grad_norm": 0.8879472613334656, |
|
"learning_rate": 1.602080294285533e-05, |
|
"loss": 0.7775, |
|
"step": 10720 |
|
}, |
|
{ |
|
"epoch": 0.6811695312995497, |
|
"grad_norm": 0.674929141998291, |
|
"learning_rate": 1.595737933658908e-05, |
|
"loss": 0.8155, |
|
"step": 10740 |
|
}, |
|
{ |
|
"epoch": 0.6824380034248747, |
|
"grad_norm": 0.8436025381088257, |
|
"learning_rate": 1.5893955730322826e-05, |
|
"loss": 0.7855, |
|
"step": 10760 |
|
}, |
|
{ |
|
"epoch": 0.6837064755501998, |
|
"grad_norm": 0.7950330972671509, |
|
"learning_rate": 1.5830532124056575e-05, |
|
"loss": 0.8171, |
|
"step": 10780 |
|
}, |
|
{ |
|
"epoch": 0.6849749476755248, |
|
"grad_norm": 0.7402753233909607, |
|
"learning_rate": 1.576710851779032e-05, |
|
"loss": 0.7543, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.6862434198008499, |
|
"grad_norm": 0.7969671487808228, |
|
"learning_rate": 1.570368491152407e-05, |
|
"loss": 0.811, |
|
"step": 10820 |
|
}, |
|
{ |
|
"epoch": 0.6875118919261749, |
|
"grad_norm": 0.9241589307785034, |
|
"learning_rate": 1.5640261305257816e-05, |
|
"loss": 0.8234, |
|
"step": 10840 |
|
}, |
|
{ |
|
"epoch": 0.6887803640515, |
|
"grad_norm": 0.8808215260505676, |
|
"learning_rate": 1.5576837698991565e-05, |
|
"loss": 0.7646, |
|
"step": 10860 |
|
}, |
|
{ |
|
"epoch": 0.690048836176825, |
|
"grad_norm": 0.7900111675262451, |
|
"learning_rate": 1.551341409272531e-05, |
|
"loss": 0.7461, |
|
"step": 10880 |
|
}, |
|
{ |
|
"epoch": 0.6913173083021501, |
|
"grad_norm": 0.9008402824401855, |
|
"learning_rate": 1.544999048645906e-05, |
|
"loss": 0.7693, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.6925857804274751, |
|
"grad_norm": 0.925081729888916, |
|
"learning_rate": 1.538656688019281e-05, |
|
"loss": 0.8174, |
|
"step": 10920 |
|
}, |
|
{ |
|
"epoch": 0.6938542525528002, |
|
"grad_norm": 0.8141810297966003, |
|
"learning_rate": 1.532314327392656e-05, |
|
"loss": 0.8009, |
|
"step": 10940 |
|
}, |
|
{ |
|
"epoch": 0.6951227246781252, |
|
"grad_norm": 0.8973850011825562, |
|
"learning_rate": 1.5259719667660305e-05, |
|
"loss": 0.7731, |
|
"step": 10960 |
|
}, |
|
{ |
|
"epoch": 0.6963911968034503, |
|
"grad_norm": 0.7652609348297119, |
|
"learning_rate": 1.5196296061394052e-05, |
|
"loss": 0.7751, |
|
"step": 10980 |
|
}, |
|
{ |
|
"epoch": 0.6976596689287753, |
|
"grad_norm": 0.8361225724220276, |
|
"learning_rate": 1.51328724551278e-05, |
|
"loss": 0.7551, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.6989281410541003, |
|
"grad_norm": 0.7935757040977478, |
|
"learning_rate": 1.5069448848861547e-05, |
|
"loss": 0.8139, |
|
"step": 11020 |
|
}, |
|
{ |
|
"epoch": 0.7001966131794254, |
|
"grad_norm": 0.7135019898414612, |
|
"learning_rate": 1.5006025242595295e-05, |
|
"loss": 0.7768, |
|
"step": 11040 |
|
}, |
|
{ |
|
"epoch": 0.7014650853047504, |
|
"grad_norm": 0.811869204044342, |
|
"learning_rate": 1.4942601636329042e-05, |
|
"loss": 0.7697, |
|
"step": 11060 |
|
}, |
|
{ |
|
"epoch": 0.7027335574300755, |
|
"grad_norm": 0.9030170440673828, |
|
"learning_rate": 1.487917803006279e-05, |
|
"loss": 0.8206, |
|
"step": 11080 |
|
}, |
|
{ |
|
"epoch": 0.7040020295554005, |
|
"grad_norm": 0.765082836151123, |
|
"learning_rate": 1.4815754423796537e-05, |
|
"loss": 0.8204, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.7052705016807256, |
|
"grad_norm": 0.7715885639190674, |
|
"learning_rate": 1.4752330817530285e-05, |
|
"loss": 0.7718, |
|
"step": 11120 |
|
}, |
|
{ |
|
"epoch": 0.7065389738060506, |
|
"grad_norm": 0.7729353904724121, |
|
"learning_rate": 1.4688907211264034e-05, |
|
"loss": 0.8165, |
|
"step": 11140 |
|
}, |
|
{ |
|
"epoch": 0.7078074459313757, |
|
"grad_norm": 0.6622787117958069, |
|
"learning_rate": 1.4625483604997781e-05, |
|
"loss": 0.7618, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 0.7090759180567007, |
|
"grad_norm": 0.820572555065155, |
|
"learning_rate": 1.4562059998731529e-05, |
|
"loss": 0.7887, |
|
"step": 11180 |
|
}, |
|
{ |
|
"epoch": 0.7103443901820258, |
|
"grad_norm": 0.7710301876068115, |
|
"learning_rate": 1.4498636392465276e-05, |
|
"loss": 0.7712, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.7116128623073508, |
|
"grad_norm": 0.8138539791107178, |
|
"learning_rate": 1.4435212786199024e-05, |
|
"loss": 0.7989, |
|
"step": 11220 |
|
}, |
|
{ |
|
"epoch": 0.7128813344326759, |
|
"grad_norm": 0.7800792455673218, |
|
"learning_rate": 1.4371789179932771e-05, |
|
"loss": 0.7641, |
|
"step": 11240 |
|
}, |
|
{ |
|
"epoch": 0.7141498065580009, |
|
"grad_norm": 0.809686005115509, |
|
"learning_rate": 1.4308365573666519e-05, |
|
"loss": 0.8053, |
|
"step": 11260 |
|
}, |
|
{ |
|
"epoch": 0.715418278683326, |
|
"grad_norm": 0.8002369403839111, |
|
"learning_rate": 1.4244941967400266e-05, |
|
"loss": 0.8044, |
|
"step": 11280 |
|
}, |
|
{ |
|
"epoch": 0.716686750808651, |
|
"grad_norm": 0.8907930850982666, |
|
"learning_rate": 1.4181518361134014e-05, |
|
"loss": 0.7896, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.7179552229339761, |
|
"grad_norm": 0.8205035328865051, |
|
"learning_rate": 1.4118094754867761e-05, |
|
"loss": 0.7609, |
|
"step": 11320 |
|
}, |
|
{ |
|
"epoch": 0.7192236950593011, |
|
"grad_norm": 0.7667264342308044, |
|
"learning_rate": 1.4054671148601512e-05, |
|
"loss": 0.8022, |
|
"step": 11340 |
|
}, |
|
{ |
|
"epoch": 0.7204921671846262, |
|
"grad_norm": 0.7035322189331055, |
|
"learning_rate": 1.399124754233526e-05, |
|
"loss": 0.7969, |
|
"step": 11360 |
|
}, |
|
{ |
|
"epoch": 0.7217606393099512, |
|
"grad_norm": 0.7853593230247498, |
|
"learning_rate": 1.3927823936069007e-05, |
|
"loss": 0.7839, |
|
"step": 11380 |
|
}, |
|
{ |
|
"epoch": 0.7230291114352763, |
|
"grad_norm": 0.9023504853248596, |
|
"learning_rate": 1.3864400329802755e-05, |
|
"loss": 0.7867, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.7242975835606013, |
|
"grad_norm": 0.8038562536239624, |
|
"learning_rate": 1.3800976723536502e-05, |
|
"loss": 0.76, |
|
"step": 11420 |
|
}, |
|
{ |
|
"epoch": 0.7255660556859262, |
|
"grad_norm": 0.8277421593666077, |
|
"learning_rate": 1.373755311727025e-05, |
|
"loss": 0.8377, |
|
"step": 11440 |
|
}, |
|
{ |
|
"epoch": 0.7268345278112514, |
|
"grad_norm": 0.7307552099227905, |
|
"learning_rate": 1.3674129511003997e-05, |
|
"loss": 0.7962, |
|
"step": 11460 |
|
}, |
|
{ |
|
"epoch": 0.7281029999365763, |
|
"grad_norm": 0.7248812913894653, |
|
"learning_rate": 1.3610705904737745e-05, |
|
"loss": 0.7655, |
|
"step": 11480 |
|
}, |
|
{ |
|
"epoch": 0.7293714720619014, |
|
"grad_norm": 0.8142716288566589, |
|
"learning_rate": 1.3547282298471492e-05, |
|
"loss": 0.7874, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.7306399441872264, |
|
"grad_norm": 0.8528370261192322, |
|
"learning_rate": 1.348385869220524e-05, |
|
"loss": 0.7478, |
|
"step": 11520 |
|
}, |
|
{ |
|
"epoch": 0.7319084163125515, |
|
"grad_norm": 0.7856337428092957, |
|
"learning_rate": 1.3420435085938987e-05, |
|
"loss": 0.7682, |
|
"step": 11540 |
|
}, |
|
{ |
|
"epoch": 0.7331768884378765, |
|
"grad_norm": 0.8709967136383057, |
|
"learning_rate": 1.3357011479672735e-05, |
|
"loss": 0.7951, |
|
"step": 11560 |
|
}, |
|
{ |
|
"epoch": 0.7344453605632016, |
|
"grad_norm": 0.7879327535629272, |
|
"learning_rate": 1.3293587873406482e-05, |
|
"loss": 0.8052, |
|
"step": 11580 |
|
}, |
|
{ |
|
"epoch": 0.7357138326885266, |
|
"grad_norm": 0.7124823331832886, |
|
"learning_rate": 1.323016426714023e-05, |
|
"loss": 0.7847, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.7369823048138517, |
|
"grad_norm": 0.8641963601112366, |
|
"learning_rate": 1.3166740660873977e-05, |
|
"loss": 0.7953, |
|
"step": 11620 |
|
}, |
|
{ |
|
"epoch": 0.7382507769391767, |
|
"grad_norm": 0.777748167514801, |
|
"learning_rate": 1.3103317054607725e-05, |
|
"loss": 0.7786, |
|
"step": 11640 |
|
}, |
|
{ |
|
"epoch": 0.7395192490645018, |
|
"grad_norm": 0.9086549878120422, |
|
"learning_rate": 1.3039893448341472e-05, |
|
"loss": 0.7954, |
|
"step": 11660 |
|
}, |
|
{ |
|
"epoch": 0.7407877211898268, |
|
"grad_norm": 0.7550273537635803, |
|
"learning_rate": 1.297646984207522e-05, |
|
"loss": 0.7679, |
|
"step": 11680 |
|
}, |
|
{ |
|
"epoch": 0.7420561933151519, |
|
"grad_norm": 0.8174465894699097, |
|
"learning_rate": 1.2913046235808967e-05, |
|
"loss": 0.7829, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.7433246654404769, |
|
"grad_norm": 0.8319543600082397, |
|
"learning_rate": 1.2849622629542715e-05, |
|
"loss": 0.7592, |
|
"step": 11720 |
|
}, |
|
{ |
|
"epoch": 0.744593137565802, |
|
"grad_norm": 0.7246963381767273, |
|
"learning_rate": 1.2786199023276462e-05, |
|
"loss": 0.7925, |
|
"step": 11740 |
|
}, |
|
{ |
|
"epoch": 0.745861609691127, |
|
"grad_norm": 0.7811394333839417, |
|
"learning_rate": 1.2722775417010213e-05, |
|
"loss": 0.8063, |
|
"step": 11760 |
|
}, |
|
{ |
|
"epoch": 0.7471300818164521, |
|
"grad_norm": 0.9180453419685364, |
|
"learning_rate": 1.266252299105727e-05, |
|
"loss": 0.8391, |
|
"step": 11780 |
|
}, |
|
{ |
|
"epoch": 0.7483985539417771, |
|
"grad_norm": 0.6986908912658691, |
|
"learning_rate": 1.2599099384791018e-05, |
|
"loss": 0.7773, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.7496670260671022, |
|
"grad_norm": 0.8293908834457397, |
|
"learning_rate": 1.2535675778524766e-05, |
|
"loss": 0.7888, |
|
"step": 11820 |
|
}, |
|
{ |
|
"epoch": 0.7509354981924272, |
|
"grad_norm": 0.9369567036628723, |
|
"learning_rate": 1.2472252172258515e-05, |
|
"loss": 0.772, |
|
"step": 11840 |
|
}, |
|
{ |
|
"epoch": 0.7522039703177522, |
|
"grad_norm": 0.884286105632782, |
|
"learning_rate": 1.2408828565992263e-05, |
|
"loss": 0.7806, |
|
"step": 11860 |
|
}, |
|
{ |
|
"epoch": 0.7534724424430773, |
|
"grad_norm": 0.749497652053833, |
|
"learning_rate": 1.234540495972601e-05, |
|
"loss": 0.7501, |
|
"step": 11880 |
|
}, |
|
{ |
|
"epoch": 0.7547409145684023, |
|
"grad_norm": 0.6741966605186462, |
|
"learning_rate": 1.2281981353459758e-05, |
|
"loss": 0.7672, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.7560093866937274, |
|
"grad_norm": 0.8107251524925232, |
|
"learning_rate": 1.2218557747193505e-05, |
|
"loss": 0.7765, |
|
"step": 11920 |
|
}, |
|
{ |
|
"epoch": 0.7572778588190524, |
|
"grad_norm": 0.9146373867988586, |
|
"learning_rate": 1.2155134140927253e-05, |
|
"loss": 0.7462, |
|
"step": 11940 |
|
}, |
|
{ |
|
"epoch": 0.7585463309443775, |
|
"grad_norm": 0.9027043581008911, |
|
"learning_rate": 1.2091710534661e-05, |
|
"loss": 0.7819, |
|
"step": 11960 |
|
}, |
|
{ |
|
"epoch": 0.7598148030697025, |
|
"grad_norm": 0.7713417410850525, |
|
"learning_rate": 1.202828692839475e-05, |
|
"loss": 0.7684, |
|
"step": 11980 |
|
}, |
|
{ |
|
"epoch": 0.7610832751950276, |
|
"grad_norm": 0.8822270631790161, |
|
"learning_rate": 1.1964863322128497e-05, |
|
"loss": 0.7524, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.7623517473203526, |
|
"grad_norm": 0.8402985334396362, |
|
"learning_rate": 1.1901439715862244e-05, |
|
"loss": 0.795, |
|
"step": 12020 |
|
}, |
|
{ |
|
"epoch": 0.7636202194456777, |
|
"grad_norm": 0.7556558847427368, |
|
"learning_rate": 1.1838016109595992e-05, |
|
"loss": 0.773, |
|
"step": 12040 |
|
}, |
|
{ |
|
"epoch": 0.7648886915710027, |
|
"grad_norm": 0.7098413705825806, |
|
"learning_rate": 1.177459250332974e-05, |
|
"loss": 0.7888, |
|
"step": 12060 |
|
}, |
|
{ |
|
"epoch": 0.7661571636963278, |
|
"grad_norm": 0.6865963935852051, |
|
"learning_rate": 1.1711168897063487e-05, |
|
"loss": 0.7844, |
|
"step": 12080 |
|
}, |
|
{ |
|
"epoch": 0.7674256358216528, |
|
"grad_norm": 0.9354507923126221, |
|
"learning_rate": 1.1647745290797234e-05, |
|
"loss": 0.8064, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 0.7686941079469779, |
|
"grad_norm": 0.8983631134033203, |
|
"learning_rate": 1.1584321684530983e-05, |
|
"loss": 0.759, |
|
"step": 12120 |
|
}, |
|
{ |
|
"epoch": 0.7699625800723029, |
|
"grad_norm": 0.9061852693557739, |
|
"learning_rate": 1.1520898078264731e-05, |
|
"loss": 0.7959, |
|
"step": 12140 |
|
}, |
|
{ |
|
"epoch": 0.771231052197628, |
|
"grad_norm": 0.8606493473052979, |
|
"learning_rate": 1.1457474471998478e-05, |
|
"loss": 0.7926, |
|
"step": 12160 |
|
}, |
|
{ |
|
"epoch": 0.772499524322953, |
|
"grad_norm": 0.9167592525482178, |
|
"learning_rate": 1.1394050865732226e-05, |
|
"loss": 0.7813, |
|
"step": 12180 |
|
}, |
|
{ |
|
"epoch": 0.7737679964482781, |
|
"grad_norm": 0.866223931312561, |
|
"learning_rate": 1.1330627259465975e-05, |
|
"loss": 0.7827, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.7750364685736031, |
|
"grad_norm": 0.720427930355072, |
|
"learning_rate": 1.1267203653199723e-05, |
|
"loss": 0.813, |
|
"step": 12220 |
|
}, |
|
{ |
|
"epoch": 0.7763049406989282, |
|
"grad_norm": 0.8172628283500671, |
|
"learning_rate": 1.120378004693347e-05, |
|
"loss": 0.7733, |
|
"step": 12240 |
|
}, |
|
{ |
|
"epoch": 0.7775734128242532, |
|
"grad_norm": 0.741121768951416, |
|
"learning_rate": 1.1140356440667218e-05, |
|
"loss": 0.7434, |
|
"step": 12260 |
|
}, |
|
{ |
|
"epoch": 0.7788418849495782, |
|
"grad_norm": 0.723564624786377, |
|
"learning_rate": 1.1076932834400965e-05, |
|
"loss": 0.7789, |
|
"step": 12280 |
|
}, |
|
{ |
|
"epoch": 0.7801103570749033, |
|
"grad_norm": 0.9289072155952454, |
|
"learning_rate": 1.1013509228134713e-05, |
|
"loss": 0.7894, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 0.7813788292002283, |
|
"grad_norm": 0.8132310509681702, |
|
"learning_rate": 1.095008562186846e-05, |
|
"loss": 0.8153, |
|
"step": 12320 |
|
}, |
|
{ |
|
"epoch": 0.7826473013255534, |
|
"grad_norm": 0.967943549156189, |
|
"learning_rate": 1.0886662015602208e-05, |
|
"loss": 0.7821, |
|
"step": 12340 |
|
}, |
|
{ |
|
"epoch": 0.7839157734508784, |
|
"grad_norm": 0.7738404273986816, |
|
"learning_rate": 1.0823238409335955e-05, |
|
"loss": 0.7855, |
|
"step": 12360 |
|
}, |
|
{ |
|
"epoch": 0.7851842455762035, |
|
"grad_norm": 0.8411769270896912, |
|
"learning_rate": 1.0759814803069703e-05, |
|
"loss": 0.7988, |
|
"step": 12380 |
|
}, |
|
{ |
|
"epoch": 0.7864527177015285, |
|
"grad_norm": 0.8962435722351074, |
|
"learning_rate": 1.0696391196803452e-05, |
|
"loss": 0.8321, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.7877211898268536, |
|
"grad_norm": 0.7484604716300964, |
|
"learning_rate": 1.06329675905372e-05, |
|
"loss": 0.7686, |
|
"step": 12420 |
|
}, |
|
{ |
|
"epoch": 0.7889896619521786, |
|
"grad_norm": 0.802546501159668, |
|
"learning_rate": 1.0569543984270947e-05, |
|
"loss": 0.7904, |
|
"step": 12440 |
|
}, |
|
{ |
|
"epoch": 0.7902581340775037, |
|
"grad_norm": 0.7103933691978455, |
|
"learning_rate": 1.0506120378004694e-05, |
|
"loss": 0.756, |
|
"step": 12460 |
|
}, |
|
{ |
|
"epoch": 0.7915266062028287, |
|
"grad_norm": 0.6866100430488586, |
|
"learning_rate": 1.0442696771738442e-05, |
|
"loss": 0.7736, |
|
"step": 12480 |
|
}, |
|
{ |
|
"epoch": 0.7927950783281538, |
|
"grad_norm": 0.7697407603263855, |
|
"learning_rate": 1.037927316547219e-05, |
|
"loss": 0.7796, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.7940635504534788, |
|
"grad_norm": 0.8658385276794434, |
|
"learning_rate": 1.0315849559205937e-05, |
|
"loss": 0.787, |
|
"step": 12520 |
|
}, |
|
{ |
|
"epoch": 0.7953320225788039, |
|
"grad_norm": 0.8282449245452881, |
|
"learning_rate": 1.0252425952939684e-05, |
|
"loss": 0.7666, |
|
"step": 12540 |
|
}, |
|
{ |
|
"epoch": 0.7966004947041289, |
|
"grad_norm": 0.8376625776290894, |
|
"learning_rate": 1.0189002346673432e-05, |
|
"loss": 0.7313, |
|
"step": 12560 |
|
}, |
|
{ |
|
"epoch": 0.797868966829454, |
|
"grad_norm": 0.8002750277519226, |
|
"learning_rate": 1.012557874040718e-05, |
|
"loss": 0.8, |
|
"step": 12580 |
|
}, |
|
{ |
|
"epoch": 0.799137438954779, |
|
"grad_norm": 0.7849326729774475, |
|
"learning_rate": 1.0062155134140929e-05, |
|
"loss": 0.7761, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.8004059110801041, |
|
"grad_norm": 0.8501541018486023, |
|
"learning_rate": 9.998731527874676e-06, |
|
"loss": 0.7845, |
|
"step": 12620 |
|
}, |
|
{ |
|
"epoch": 0.801674383205429, |
|
"grad_norm": 0.7818264365196228, |
|
"learning_rate": 9.935307921608424e-06, |
|
"loss": 0.7669, |
|
"step": 12640 |
|
}, |
|
{ |
|
"epoch": 0.8029428553307542, |
|
"grad_norm": 0.9117188453674316, |
|
"learning_rate": 9.871884315342171e-06, |
|
"loss": 0.7578, |
|
"step": 12660 |
|
}, |
|
{ |
|
"epoch": 0.8042113274560792, |
|
"grad_norm": 0.8058929443359375, |
|
"learning_rate": 9.808460709075919e-06, |
|
"loss": 0.7676, |
|
"step": 12680 |
|
}, |
|
{ |
|
"epoch": 0.8054797995814041, |
|
"grad_norm": 0.8033195734024048, |
|
"learning_rate": 9.745037102809666e-06, |
|
"loss": 0.8151, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 0.8067482717067292, |
|
"grad_norm": 0.898897647857666, |
|
"learning_rate": 9.681613496543414e-06, |
|
"loss": 0.8267, |
|
"step": 12720 |
|
}, |
|
{ |
|
"epoch": 0.8080167438320542, |
|
"grad_norm": 0.9970609545707703, |
|
"learning_rate": 9.618189890277161e-06, |
|
"loss": 0.8075, |
|
"step": 12740 |
|
}, |
|
{ |
|
"epoch": 0.8092852159573793, |
|
"grad_norm": 0.904344916343689, |
|
"learning_rate": 9.554766284010909e-06, |
|
"loss": 0.8, |
|
"step": 12760 |
|
}, |
|
{ |
|
"epoch": 0.8105536880827043, |
|
"grad_norm": 0.8318148255348206, |
|
"learning_rate": 9.491342677744656e-06, |
|
"loss": 0.8027, |
|
"step": 12780 |
|
}, |
|
{ |
|
"epoch": 0.8118221602080294, |
|
"grad_norm": 0.8471246957778931, |
|
"learning_rate": 9.427919071478404e-06, |
|
"loss": 0.7587, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.8130906323333544, |
|
"grad_norm": 0.7848266363143921, |
|
"learning_rate": 9.364495465212153e-06, |
|
"loss": 0.7499, |
|
"step": 12820 |
|
}, |
|
{ |
|
"epoch": 0.8143591044586795, |
|
"grad_norm": 0.9037428498268127, |
|
"learning_rate": 9.3010718589459e-06, |
|
"loss": 0.7865, |
|
"step": 12840 |
|
}, |
|
{ |
|
"epoch": 0.8156275765840045, |
|
"grad_norm": 0.7049270868301392, |
|
"learning_rate": 9.237648252679648e-06, |
|
"loss": 0.817, |
|
"step": 12860 |
|
}, |
|
{ |
|
"epoch": 0.8168960487093296, |
|
"grad_norm": 0.7613449096679688, |
|
"learning_rate": 9.174224646413395e-06, |
|
"loss": 0.806, |
|
"step": 12880 |
|
}, |
|
{ |
|
"epoch": 0.8181645208346546, |
|
"grad_norm": 0.7704141139984131, |
|
"learning_rate": 9.110801040147143e-06, |
|
"loss": 0.7307, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 0.8194329929599797, |
|
"grad_norm": 0.707279622554779, |
|
"learning_rate": 9.04737743388089e-06, |
|
"loss": 0.7838, |
|
"step": 12920 |
|
}, |
|
{ |
|
"epoch": 0.8207014650853047, |
|
"grad_norm": 0.7817753553390503, |
|
"learning_rate": 8.983953827614638e-06, |
|
"loss": 0.7994, |
|
"step": 12940 |
|
}, |
|
{ |
|
"epoch": 0.8219699372106298, |
|
"grad_norm": 0.8321487307548523, |
|
"learning_rate": 8.920530221348385e-06, |
|
"loss": 0.7843, |
|
"step": 12960 |
|
}, |
|
{ |
|
"epoch": 0.8232384093359548, |
|
"grad_norm": 0.799281120300293, |
|
"learning_rate": 8.857106615082133e-06, |
|
"loss": 0.771, |
|
"step": 12980 |
|
}, |
|
{ |
|
"epoch": 0.8245068814612799, |
|
"grad_norm": 0.8843486309051514, |
|
"learning_rate": 8.79368300881588e-06, |
|
"loss": 0.8269, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.8257753535866049, |
|
"grad_norm": 0.6699514985084534, |
|
"learning_rate": 8.73025940254963e-06, |
|
"loss": 0.7398, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 0.82704382571193, |
|
"grad_norm": 0.7868858575820923, |
|
"learning_rate": 8.666835796283377e-06, |
|
"loss": 0.7779, |
|
"step": 13040 |
|
}, |
|
{ |
|
"epoch": 0.828312297837255, |
|
"grad_norm": 0.8733574151992798, |
|
"learning_rate": 8.603412190017125e-06, |
|
"loss": 0.7977, |
|
"step": 13060 |
|
}, |
|
{ |
|
"epoch": 0.8295807699625801, |
|
"grad_norm": 0.7439238429069519, |
|
"learning_rate": 8.539988583750872e-06, |
|
"loss": 0.7587, |
|
"step": 13080 |
|
}, |
|
{ |
|
"epoch": 0.8308492420879051, |
|
"grad_norm": 0.8214549422264099, |
|
"learning_rate": 8.476564977484621e-06, |
|
"loss": 0.8181, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 0.8321177142132302, |
|
"grad_norm": 0.8577607870101929, |
|
"learning_rate": 8.413141371218369e-06, |
|
"loss": 0.793, |
|
"step": 13120 |
|
}, |
|
{ |
|
"epoch": 0.8333861863385552, |
|
"grad_norm": 0.6957492828369141, |
|
"learning_rate": 8.349717764952116e-06, |
|
"loss": 0.755, |
|
"step": 13140 |
|
}, |
|
{ |
|
"epoch": 0.8346546584638802, |
|
"grad_norm": 0.981088399887085, |
|
"learning_rate": 8.286294158685864e-06, |
|
"loss": 0.7769, |
|
"step": 13160 |
|
}, |
|
{ |
|
"epoch": 0.8359231305892053, |
|
"grad_norm": 0.7333866357803345, |
|
"learning_rate": 8.222870552419611e-06, |
|
"loss": 0.7536, |
|
"step": 13180 |
|
}, |
|
{ |
|
"epoch": 0.8371916027145303, |
|
"grad_norm": 0.8152589201927185, |
|
"learning_rate": 8.159446946153359e-06, |
|
"loss": 0.7497, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 0.8384600748398554, |
|
"grad_norm": 0.8962567448616028, |
|
"learning_rate": 8.096023339887106e-06, |
|
"loss": 0.7838, |
|
"step": 13220 |
|
}, |
|
{ |
|
"epoch": 0.8397285469651804, |
|
"grad_norm": 0.6861271262168884, |
|
"learning_rate": 8.032599733620855e-06, |
|
"loss": 0.777, |
|
"step": 13240 |
|
}, |
|
{ |
|
"epoch": 0.8409970190905055, |
|
"grad_norm": 0.7273656725883484, |
|
"learning_rate": 7.969176127354603e-06, |
|
"loss": 0.7345, |
|
"step": 13260 |
|
}, |
|
{ |
|
"epoch": 0.8422654912158305, |
|
"grad_norm": 0.7643877267837524, |
|
"learning_rate": 7.90575252108835e-06, |
|
"loss": 0.7837, |
|
"step": 13280 |
|
}, |
|
{ |
|
"epoch": 0.8435339633411556, |
|
"grad_norm": 0.695196270942688, |
|
"learning_rate": 7.842328914822098e-06, |
|
"loss": 0.7361, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 0.8448024354664806, |
|
"grad_norm": 0.6783697009086609, |
|
"learning_rate": 7.778905308555845e-06, |
|
"loss": 0.7315, |
|
"step": 13320 |
|
}, |
|
{ |
|
"epoch": 0.8460709075918057, |
|
"grad_norm": 0.8633202314376831, |
|
"learning_rate": 7.715481702289593e-06, |
|
"loss": 0.7713, |
|
"step": 13340 |
|
}, |
|
{ |
|
"epoch": 0.8473393797171307, |
|
"grad_norm": 0.7902844548225403, |
|
"learning_rate": 7.65205809602334e-06, |
|
"loss": 0.7732, |
|
"step": 13360 |
|
}, |
|
{ |
|
"epoch": 0.8486078518424558, |
|
"grad_norm": 1.1263785362243652, |
|
"learning_rate": 7.588634489757088e-06, |
|
"loss": 0.7549, |
|
"step": 13380 |
|
}, |
|
{ |
|
"epoch": 0.8498763239677808, |
|
"grad_norm": 0.7141507863998413, |
|
"learning_rate": 7.5252108834908354e-06, |
|
"loss": 0.7498, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 0.8511447960931059, |
|
"grad_norm": 0.7436708211898804, |
|
"learning_rate": 7.461787277224583e-06, |
|
"loss": 0.7525, |
|
"step": 13420 |
|
}, |
|
{ |
|
"epoch": 0.8524132682184309, |
|
"grad_norm": 0.7840022444725037, |
|
"learning_rate": 7.3983636709583304e-06, |
|
"loss": 0.7658, |
|
"step": 13440 |
|
}, |
|
{ |
|
"epoch": 0.853681740343756, |
|
"grad_norm": 0.7560069561004639, |
|
"learning_rate": 7.33494006469208e-06, |
|
"loss": 0.7946, |
|
"step": 13460 |
|
}, |
|
{ |
|
"epoch": 0.854950212469081, |
|
"grad_norm": 0.7361060976982117, |
|
"learning_rate": 7.271516458425827e-06, |
|
"loss": 0.7722, |
|
"step": 13480 |
|
}, |
|
{ |
|
"epoch": 0.8562186845944061, |
|
"grad_norm": 0.8141864538192749, |
|
"learning_rate": 7.208092852159575e-06, |
|
"loss": 0.751, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.8574871567197311, |
|
"grad_norm": 0.7860879898071289, |
|
"learning_rate": 7.144669245893322e-06, |
|
"loss": 0.7401, |
|
"step": 13520 |
|
}, |
|
{ |
|
"epoch": 0.8587556288450562, |
|
"grad_norm": 1.1111942529678345, |
|
"learning_rate": 7.08124563962707e-06, |
|
"loss": 0.803, |
|
"step": 13540 |
|
}, |
|
{ |
|
"epoch": 0.8600241009703812, |
|
"grad_norm": 0.7983526587486267, |
|
"learning_rate": 7.017822033360817e-06, |
|
"loss": 0.7963, |
|
"step": 13560 |
|
}, |
|
{ |
|
"epoch": 0.8612925730957062, |
|
"grad_norm": 0.7415090799331665, |
|
"learning_rate": 6.954398427094565e-06, |
|
"loss": 0.7997, |
|
"step": 13580 |
|
}, |
|
{ |
|
"epoch": 0.8625610452210313, |
|
"grad_norm": 0.8375813364982605, |
|
"learning_rate": 6.890974820828312e-06, |
|
"loss": 0.8052, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 0.8638295173463563, |
|
"grad_norm": 0.8622868657112122, |
|
"learning_rate": 6.82755121456206e-06, |
|
"loss": 0.7816, |
|
"step": 13620 |
|
}, |
|
{ |
|
"epoch": 0.8650979894716814, |
|
"grad_norm": 0.9229819774627686, |
|
"learning_rate": 6.764127608295807e-06, |
|
"loss": 0.7558, |
|
"step": 13640 |
|
}, |
|
{ |
|
"epoch": 0.8663664615970064, |
|
"grad_norm": 0.8334788084030151, |
|
"learning_rate": 6.700704002029556e-06, |
|
"loss": 0.7648, |
|
"step": 13660 |
|
}, |
|
{ |
|
"epoch": 0.8676349337223315, |
|
"grad_norm": 0.8499231338500977, |
|
"learning_rate": 6.637280395763304e-06, |
|
"loss": 0.7954, |
|
"step": 13680 |
|
}, |
|
{ |
|
"epoch": 0.8689034058476565, |
|
"grad_norm": 0.8455031514167786, |
|
"learning_rate": 6.573856789497051e-06, |
|
"loss": 0.7973, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 0.8701718779729816, |
|
"grad_norm": 0.7882249355316162, |
|
"learning_rate": 6.510433183230799e-06, |
|
"loss": 0.8031, |
|
"step": 13720 |
|
}, |
|
{ |
|
"epoch": 0.8714403500983066, |
|
"grad_norm": 0.7251647114753723, |
|
"learning_rate": 6.447009576964546e-06, |
|
"loss": 0.7581, |
|
"step": 13740 |
|
}, |
|
{ |
|
"epoch": 0.8727088222236317, |
|
"grad_norm": 0.8899182081222534, |
|
"learning_rate": 6.383585970698294e-06, |
|
"loss": 0.7981, |
|
"step": 13760 |
|
}, |
|
{ |
|
"epoch": 0.8739772943489567, |
|
"grad_norm": 0.7742959260940552, |
|
"learning_rate": 6.320162364432042e-06, |
|
"loss": 0.7606, |
|
"step": 13780 |
|
}, |
|
{ |
|
"epoch": 0.8752457664742818, |
|
"grad_norm": 0.7180908918380737, |
|
"learning_rate": 6.25673875816579e-06, |
|
"loss": 0.8301, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 0.8765142385996068, |
|
"grad_norm": 0.7653104066848755, |
|
"learning_rate": 6.193315151899537e-06, |
|
"loss": 0.8017, |
|
"step": 13820 |
|
}, |
|
{ |
|
"epoch": 0.8777827107249319, |
|
"grad_norm": 0.802506148815155, |
|
"learning_rate": 6.1298915456332856e-06, |
|
"loss": 0.7634, |
|
"step": 13840 |
|
}, |
|
{ |
|
"epoch": 0.8790511828502569, |
|
"grad_norm": 0.882520318031311, |
|
"learning_rate": 6.066467939367033e-06, |
|
"loss": 0.777, |
|
"step": 13860 |
|
}, |
|
{ |
|
"epoch": 0.880319654975582, |
|
"grad_norm": 0.7464948892593384, |
|
"learning_rate": 6.0030443331007806e-06, |
|
"loss": 0.7495, |
|
"step": 13880 |
|
}, |
|
{ |
|
"epoch": 0.881588127100907, |
|
"grad_norm": 0.7769840359687805, |
|
"learning_rate": 5.939620726834528e-06, |
|
"loss": 0.7562, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 0.882856599226232, |
|
"grad_norm": 0.9281843304634094, |
|
"learning_rate": 5.8761971205682756e-06, |
|
"loss": 0.7845, |
|
"step": 13920 |
|
}, |
|
{ |
|
"epoch": 0.884125071351557, |
|
"grad_norm": 0.801986813545227, |
|
"learning_rate": 5.812773514302024e-06, |
|
"loss": 0.7301, |
|
"step": 13940 |
|
}, |
|
{ |
|
"epoch": 0.8853935434768821, |
|
"grad_norm": 0.8647619485855103, |
|
"learning_rate": 5.749349908035771e-06, |
|
"loss": 0.7693, |
|
"step": 13960 |
|
}, |
|
{ |
|
"epoch": 0.8866620156022071, |
|
"grad_norm": 0.8235803246498108, |
|
"learning_rate": 5.685926301769519e-06, |
|
"loss": 0.7972, |
|
"step": 13980 |
|
}, |
|
{ |
|
"epoch": 0.8879304877275321, |
|
"grad_norm": 0.8338538408279419, |
|
"learning_rate": 5.622502695503266e-06, |
|
"loss": 0.7951, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.8891989598528572, |
|
"grad_norm": 0.704741895198822, |
|
"learning_rate": 5.559079089237014e-06, |
|
"loss": 0.7446, |
|
"step": 14020 |
|
}, |
|
{ |
|
"epoch": 0.8904674319781822, |
|
"grad_norm": 0.7731455564498901, |
|
"learning_rate": 5.495655482970762e-06, |
|
"loss": 0.7266, |
|
"step": 14040 |
|
}, |
|
{ |
|
"epoch": 0.8917359041035073, |
|
"grad_norm": 0.779869794845581, |
|
"learning_rate": 5.43223187670451e-06, |
|
"loss": 0.7638, |
|
"step": 14060 |
|
}, |
|
{ |
|
"epoch": 0.8930043762288323, |
|
"grad_norm": 0.7645334005355835, |
|
"learning_rate": 5.368808270438257e-06, |
|
"loss": 0.7868, |
|
"step": 14080 |
|
}, |
|
{ |
|
"epoch": 0.8942728483541574, |
|
"grad_norm": 0.8010347485542297, |
|
"learning_rate": 5.305384664172005e-06, |
|
"loss": 0.8038, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 0.8955413204794824, |
|
"grad_norm": 0.830556333065033, |
|
"learning_rate": 5.241961057905752e-06, |
|
"loss": 0.7774, |
|
"step": 14120 |
|
}, |
|
{ |
|
"epoch": 0.8968097926048075, |
|
"grad_norm": 0.9386767745018005, |
|
"learning_rate": 5.178537451639501e-06, |
|
"loss": 0.7591, |
|
"step": 14140 |
|
}, |
|
{ |
|
"epoch": 0.8980782647301325, |
|
"grad_norm": 0.8357464671134949, |
|
"learning_rate": 5.115113845373248e-06, |
|
"loss": 0.7642, |
|
"step": 14160 |
|
}, |
|
{ |
|
"epoch": 0.8993467368554576, |
|
"grad_norm": 0.7423475384712219, |
|
"learning_rate": 5.051690239106996e-06, |
|
"loss": 0.7586, |
|
"step": 14180 |
|
}, |
|
{ |
|
"epoch": 0.9006152089807826, |
|
"grad_norm": 0.8751846551895142, |
|
"learning_rate": 4.988266632840743e-06, |
|
"loss": 0.7536, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 0.9018836811061077, |
|
"grad_norm": 0.8088937997817993, |
|
"learning_rate": 4.924843026574491e-06, |
|
"loss": 0.7747, |
|
"step": 14220 |
|
}, |
|
{ |
|
"epoch": 0.9031521532314327, |
|
"grad_norm": 0.7831218242645264, |
|
"learning_rate": 4.861419420308239e-06, |
|
"loss": 0.7741, |
|
"step": 14240 |
|
}, |
|
{ |
|
"epoch": 0.9044206253567578, |
|
"grad_norm": 0.8346021175384521, |
|
"learning_rate": 4.7979958140419865e-06, |
|
"loss": 0.8034, |
|
"step": 14260 |
|
}, |
|
{ |
|
"epoch": 0.9056890974820828, |
|
"grad_norm": 0.7575668692588806, |
|
"learning_rate": 4.734572207775735e-06, |
|
"loss": 0.7866, |
|
"step": 14280 |
|
}, |
|
{ |
|
"epoch": 0.9069575696074079, |
|
"grad_norm": 0.8374447822570801, |
|
"learning_rate": 4.671148601509482e-06, |
|
"loss": 0.7808, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 0.9082260417327329, |
|
"grad_norm": 0.7750478982925415, |
|
"learning_rate": 4.60772499524323e-06, |
|
"loss": 0.7628, |
|
"step": 14320 |
|
}, |
|
{ |
|
"epoch": 0.909494513858058, |
|
"grad_norm": 0.8722181916236877, |
|
"learning_rate": 4.54747256929029e-06, |
|
"loss": 0.7744, |
|
"step": 14340 |
|
}, |
|
{ |
|
"epoch": 0.910762985983383, |
|
"grad_norm": 0.8774197101593018, |
|
"learning_rate": 4.484048963024038e-06, |
|
"loss": 0.8039, |
|
"step": 14360 |
|
}, |
|
{ |
|
"epoch": 0.9120314581087081, |
|
"grad_norm": 0.677240788936615, |
|
"learning_rate": 4.420625356757786e-06, |
|
"loss": 0.7648, |
|
"step": 14380 |
|
}, |
|
{ |
|
"epoch": 0.9132999302340331, |
|
"grad_norm": 0.8468155264854431, |
|
"learning_rate": 4.357201750491533e-06, |
|
"loss": 0.7494, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 0.9145684023593581, |
|
"grad_norm": 0.8869547247886658, |
|
"learning_rate": 4.293778144225281e-06, |
|
"loss": 0.7713, |
|
"step": 14420 |
|
}, |
|
{ |
|
"epoch": 0.9158368744846832, |
|
"grad_norm": 0.8307056427001953, |
|
"learning_rate": 4.230354537959028e-06, |
|
"loss": 0.7784, |
|
"step": 14440 |
|
}, |
|
{ |
|
"epoch": 0.9171053466100082, |
|
"grad_norm": 0.8972311019897461, |
|
"learning_rate": 4.166930931692777e-06, |
|
"loss": 0.7653, |
|
"step": 14460 |
|
}, |
|
{ |
|
"epoch": 0.9183738187353333, |
|
"grad_norm": 0.7319260835647583, |
|
"learning_rate": 4.103507325426524e-06, |
|
"loss": 0.7802, |
|
"step": 14480 |
|
}, |
|
{ |
|
"epoch": 0.9196422908606583, |
|
"grad_norm": 0.8915144801139832, |
|
"learning_rate": 4.040083719160272e-06, |
|
"loss": 0.7709, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.9209107629859834, |
|
"grad_norm": 1.0568279027938843, |
|
"learning_rate": 3.976660112894019e-06, |
|
"loss": 0.7961, |
|
"step": 14520 |
|
}, |
|
{ |
|
"epoch": 0.9221792351113084, |
|
"grad_norm": 0.9176843166351318, |
|
"learning_rate": 3.913236506627767e-06, |
|
"loss": 0.7758, |
|
"step": 14540 |
|
}, |
|
{ |
|
"epoch": 0.9234477072366335, |
|
"grad_norm": 0.8266422152519226, |
|
"learning_rate": 3.849812900361515e-06, |
|
"loss": 0.7896, |
|
"step": 14560 |
|
}, |
|
{ |
|
"epoch": 0.9247161793619585, |
|
"grad_norm": 0.8376505970954895, |
|
"learning_rate": 3.7863892940952626e-06, |
|
"loss": 0.7529, |
|
"step": 14580 |
|
}, |
|
{ |
|
"epoch": 0.9259846514872836, |
|
"grad_norm": 0.7567136287689209, |
|
"learning_rate": 3.72296568782901e-06, |
|
"loss": 0.7361, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 0.9272531236126086, |
|
"grad_norm": 0.8554660677909851, |
|
"learning_rate": 3.6595420815627576e-06, |
|
"loss": 0.7808, |
|
"step": 14620 |
|
}, |
|
{ |
|
"epoch": 0.9285215957379337, |
|
"grad_norm": 0.8139016032218933, |
|
"learning_rate": 3.5961184752965055e-06, |
|
"loss": 0.7743, |
|
"step": 14640 |
|
}, |
|
{ |
|
"epoch": 0.9297900678632587, |
|
"grad_norm": 0.8457810282707214, |
|
"learning_rate": 3.532694869030253e-06, |
|
"loss": 0.7431, |
|
"step": 14660 |
|
}, |
|
{ |
|
"epoch": 0.9310585399885838, |
|
"grad_norm": 0.7392221093177795, |
|
"learning_rate": 3.469271262764001e-06, |
|
"loss": 0.7678, |
|
"step": 14680 |
|
}, |
|
{ |
|
"epoch": 0.9323270121139088, |
|
"grad_norm": 0.74676913022995, |
|
"learning_rate": 3.405847656497749e-06, |
|
"loss": 0.7739, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 0.9335954842392339, |
|
"grad_norm": 0.9203459620475769, |
|
"learning_rate": 3.3424240502314964e-06, |
|
"loss": 0.7746, |
|
"step": 14720 |
|
}, |
|
{ |
|
"epoch": 0.9348639563645589, |
|
"grad_norm": 0.8364453911781311, |
|
"learning_rate": 3.279000443965244e-06, |
|
"loss": 0.8001, |
|
"step": 14740 |
|
}, |
|
{ |
|
"epoch": 0.936132428489884, |
|
"grad_norm": 0.8923636078834534, |
|
"learning_rate": 3.2155768376989914e-06, |
|
"loss": 0.7831, |
|
"step": 14760 |
|
}, |
|
{ |
|
"epoch": 0.937400900615209, |
|
"grad_norm": 0.8334746360778809, |
|
"learning_rate": 3.1521532314327397e-06, |
|
"loss": 0.7731, |
|
"step": 14780 |
|
}, |
|
{ |
|
"epoch": 0.9386693727405341, |
|
"grad_norm": 0.7588502764701843, |
|
"learning_rate": 3.0887296251664873e-06, |
|
"loss": 0.7625, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 0.9399378448658591, |
|
"grad_norm": 0.7889774441719055, |
|
"learning_rate": 3.0253060189002348e-06, |
|
"loss": 0.7354, |
|
"step": 14820 |
|
}, |
|
{ |
|
"epoch": 0.9412063169911841, |
|
"grad_norm": 0.7171176671981812, |
|
"learning_rate": 2.9618824126339823e-06, |
|
"loss": 0.7341, |
|
"step": 14840 |
|
}, |
|
{ |
|
"epoch": 0.9424747891165092, |
|
"grad_norm": 0.9590219259262085, |
|
"learning_rate": 2.89845880636773e-06, |
|
"loss": 0.8243, |
|
"step": 14860 |
|
}, |
|
{ |
|
"epoch": 0.9437432612418342, |
|
"grad_norm": 0.8210035562515259, |
|
"learning_rate": 2.8350352001014777e-06, |
|
"loss": 0.7863, |
|
"step": 14880 |
|
}, |
|
{ |
|
"epoch": 0.9450117333671593, |
|
"grad_norm": 0.9011463522911072, |
|
"learning_rate": 2.7716115938352256e-06, |
|
"loss": 0.7325, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 0.9462802054924843, |
|
"grad_norm": 0.8472886681556702, |
|
"learning_rate": 2.7081879875689735e-06, |
|
"loss": 0.7409, |
|
"step": 14920 |
|
}, |
|
{ |
|
"epoch": 0.9475486776178094, |
|
"grad_norm": 0.8105041980743408, |
|
"learning_rate": 2.644764381302721e-06, |
|
"loss": 0.7598, |
|
"step": 14940 |
|
}, |
|
{ |
|
"epoch": 0.9488171497431344, |
|
"grad_norm": 0.9121299386024475, |
|
"learning_rate": 2.581340775036469e-06, |
|
"loss": 0.8026, |
|
"step": 14960 |
|
}, |
|
{ |
|
"epoch": 0.9500856218684595, |
|
"grad_norm": 0.7272994518280029, |
|
"learning_rate": 2.5179171687702165e-06, |
|
"loss": 0.771, |
|
"step": 14980 |
|
}, |
|
{ |
|
"epoch": 0.9513540939937845, |
|
"grad_norm": 0.9553162455558777, |
|
"learning_rate": 2.454493562503964e-06, |
|
"loss": 0.7562, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.9526225661191096, |
|
"grad_norm": 0.775790810585022, |
|
"learning_rate": 2.391069956237712e-06, |
|
"loss": 0.7682, |
|
"step": 15020 |
|
}, |
|
{ |
|
"epoch": 0.9538910382444346, |
|
"grad_norm": 0.7874183654785156, |
|
"learning_rate": 2.3276463499714594e-06, |
|
"loss": 0.7706, |
|
"step": 15040 |
|
}, |
|
{ |
|
"epoch": 0.9551595103697597, |
|
"grad_norm": 0.9175487756729126, |
|
"learning_rate": 2.2642227437052073e-06, |
|
"loss": 0.7904, |
|
"step": 15060 |
|
}, |
|
{ |
|
"epoch": 0.9564279824950846, |
|
"grad_norm": 0.7636487483978271, |
|
"learning_rate": 2.200799137438955e-06, |
|
"loss": 0.7568, |
|
"step": 15080 |
|
}, |
|
{ |
|
"epoch": 0.9576964546204098, |
|
"grad_norm": 0.8234472870826721, |
|
"learning_rate": 2.1373755311727023e-06, |
|
"loss": 0.7902, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 0.9589649267457347, |
|
"grad_norm": 0.7153974175453186, |
|
"learning_rate": 2.0739519249064502e-06, |
|
"loss": 0.7305, |
|
"step": 15120 |
|
}, |
|
{ |
|
"epoch": 0.9602333988710599, |
|
"grad_norm": 0.8177540302276611, |
|
"learning_rate": 2.010528318640198e-06, |
|
"loss": 0.7584, |
|
"step": 15140 |
|
}, |
|
{ |
|
"epoch": 0.9615018709963848, |
|
"grad_norm": 0.7617529630661011, |
|
"learning_rate": 1.9471047123739457e-06, |
|
"loss": 0.74, |
|
"step": 15160 |
|
}, |
|
{ |
|
"epoch": 0.96277034312171, |
|
"grad_norm": 0.8607798218727112, |
|
"learning_rate": 1.8836811061076934e-06, |
|
"loss": 0.7867, |
|
"step": 15180 |
|
}, |
|
{ |
|
"epoch": 0.9640388152470349, |
|
"grad_norm": 0.7388313412666321, |
|
"learning_rate": 1.820257499841441e-06, |
|
"loss": 0.7561, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 0.96530728737236, |
|
"grad_norm": 0.8492052555084229, |
|
"learning_rate": 1.756833893575189e-06, |
|
"loss": 0.7769, |
|
"step": 15220 |
|
}, |
|
{ |
|
"epoch": 0.966575759497685, |
|
"grad_norm": 0.7551003694534302, |
|
"learning_rate": 1.6934102873089365e-06, |
|
"loss": 0.7641, |
|
"step": 15240 |
|
}, |
|
{ |
|
"epoch": 0.96784423162301, |
|
"grad_norm": 0.8290461897850037, |
|
"learning_rate": 1.629986681042684e-06, |
|
"loss": 0.7537, |
|
"step": 15260 |
|
}, |
|
{ |
|
"epoch": 0.9691127037483351, |
|
"grad_norm": 0.8409222960472107, |
|
"learning_rate": 1.566563074776432e-06, |
|
"loss": 0.7591, |
|
"step": 15280 |
|
}, |
|
{ |
|
"epoch": 0.9703811758736601, |
|
"grad_norm": 0.7970009446144104, |
|
"learning_rate": 1.5031394685101795e-06, |
|
"loss": 0.7763, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 0.9716496479989852, |
|
"grad_norm": 0.8285433650016785, |
|
"learning_rate": 1.4397158622439274e-06, |
|
"loss": 0.7357, |
|
"step": 15320 |
|
}, |
|
{ |
|
"epoch": 0.9729181201243102, |
|
"grad_norm": 0.8585550785064697, |
|
"learning_rate": 1.376292255977675e-06, |
|
"loss": 0.7706, |
|
"step": 15340 |
|
}, |
|
{ |
|
"epoch": 0.9741865922496353, |
|
"grad_norm": 0.7604003548622131, |
|
"learning_rate": 1.3128686497114226e-06, |
|
"loss": 0.7455, |
|
"step": 15360 |
|
}, |
|
{ |
|
"epoch": 0.9754550643749603, |
|
"grad_norm": 0.7896905541419983, |
|
"learning_rate": 1.2494450434451703e-06, |
|
"loss": 0.7824, |
|
"step": 15380 |
|
}, |
|
{ |
|
"epoch": 0.9767235365002854, |
|
"grad_norm": 0.8073090314865112, |
|
"learning_rate": 1.186021437178918e-06, |
|
"loss": 0.7836, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 0.9779920086256104, |
|
"grad_norm": 0.8942602276802063, |
|
"learning_rate": 1.1225978309126657e-06, |
|
"loss": 0.7993, |
|
"step": 15420 |
|
}, |
|
{ |
|
"epoch": 0.9792604807509355, |
|
"grad_norm": 0.7417428493499756, |
|
"learning_rate": 1.0591742246464135e-06, |
|
"loss": 0.7828, |
|
"step": 15440 |
|
}, |
|
{ |
|
"epoch": 0.9805289528762605, |
|
"grad_norm": 0.7453452944755554, |
|
"learning_rate": 9.957506183801612e-07, |
|
"loss": 0.7589, |
|
"step": 15460 |
|
}, |
|
{ |
|
"epoch": 0.9817974250015856, |
|
"grad_norm": 0.7909823656082153, |
|
"learning_rate": 9.323270121139089e-07, |
|
"loss": 0.7776, |
|
"step": 15480 |
|
}, |
|
{ |
|
"epoch": 0.9830658971269106, |
|
"grad_norm": 0.8445958495140076, |
|
"learning_rate": 8.689034058476566e-07, |
|
"loss": 0.8027, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.9843343692522357, |
|
"grad_norm": 0.8094016313552856, |
|
"learning_rate": 8.054797995814042e-07, |
|
"loss": 0.7426, |
|
"step": 15520 |
|
}, |
|
{ |
|
"epoch": 0.9856028413775607, |
|
"grad_norm": 0.9366889595985413, |
|
"learning_rate": 7.420561933151519e-07, |
|
"loss": 0.7754, |
|
"step": 15540 |
|
}, |
|
{ |
|
"epoch": 0.9868713135028858, |
|
"grad_norm": 0.8186490535736084, |
|
"learning_rate": 6.786325870488996e-07, |
|
"loss": 0.7699, |
|
"step": 15560 |
|
}, |
|
{ |
|
"epoch": 0.9881397856282108, |
|
"grad_norm": 0.7232887148857117, |
|
"learning_rate": 6.152089807826473e-07, |
|
"loss": 0.7384, |
|
"step": 15580 |
|
}, |
|
{ |
|
"epoch": 0.9894082577535359, |
|
"grad_norm": 0.7465324401855469, |
|
"learning_rate": 5.517853745163951e-07, |
|
"loss": 0.775, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 0.9906767298788609, |
|
"grad_norm": 0.8372469544410706, |
|
"learning_rate": 4.883617682501427e-07, |
|
"loss": 0.7497, |
|
"step": 15620 |
|
}, |
|
{ |
|
"epoch": 0.991945202004186, |
|
"grad_norm": 0.8273568749427795, |
|
"learning_rate": 4.2493816198389043e-07, |
|
"loss": 0.7738, |
|
"step": 15640 |
|
}, |
|
{ |
|
"epoch": 0.993213674129511, |
|
"grad_norm": 0.7157850861549377, |
|
"learning_rate": 3.6151455571763815e-07, |
|
"loss": 0.7398, |
|
"step": 15660 |
|
}, |
|
{ |
|
"epoch": 0.994482146254836, |
|
"grad_norm": 0.8117349147796631, |
|
"learning_rate": 2.980909494513858e-07, |
|
"loss": 0.8053, |
|
"step": 15680 |
|
}, |
|
{ |
|
"epoch": 0.9957506183801611, |
|
"grad_norm": 0.869315505027771, |
|
"learning_rate": 2.3466734318513352e-07, |
|
"loss": 0.7819, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 0.9970190905054861, |
|
"grad_norm": 0.7544413208961487, |
|
"learning_rate": 1.712437369188812e-07, |
|
"loss": 0.805, |
|
"step": 15720 |
|
}, |
|
{ |
|
"epoch": 0.9982875626308112, |
|
"grad_norm": 0.9937105178833008, |
|
"learning_rate": 1.0782013065262892e-07, |
|
"loss": 0.7808, |
|
"step": 15740 |
|
}, |
|
{ |
|
"epoch": 0.9995560347561362, |
|
"grad_norm": 0.7431527972221375, |
|
"learning_rate": 4.439652438637661e-08, |
|
"loss": 0.7574, |
|
"step": 15760 |
|
} |
|
], |
|
"logging_steps": 20, |
|
"max_steps": 15767, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 15767, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.607614351420162e+18, |
|
"train_batch_size": 18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|