prxy5604's picture
Training in progress, step 269, checkpoint
595e9ab verified
{
"best_metric": 1.2524751424789429,
"best_model_checkpoint": "miner_id_24/checkpoint-250",
"epoch": 3.005586592178771,
"eval_steps": 50,
"global_step": 269,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0111731843575419,
"grad_norm": 162.40016174316406,
"learning_rate": 3.3333333333333333e-06,
"loss": 7.5612,
"step": 1
},
{
"epoch": 0.0111731843575419,
"eval_loss": 10.110641479492188,
"eval_runtime": 3.5559,
"eval_samples_per_second": 42.465,
"eval_steps_per_second": 21.373,
"step": 1
},
{
"epoch": 0.0223463687150838,
"grad_norm": 201.2489776611328,
"learning_rate": 6.666666666666667e-06,
"loss": 8.9985,
"step": 2
},
{
"epoch": 0.0335195530726257,
"grad_norm": 181.89503479003906,
"learning_rate": 1e-05,
"loss": 9.1837,
"step": 3
},
{
"epoch": 0.0446927374301676,
"grad_norm": 153.2060546875,
"learning_rate": 1.3333333333333333e-05,
"loss": 8.9975,
"step": 4
},
{
"epoch": 0.055865921787709494,
"grad_norm": 104.02288818359375,
"learning_rate": 1.6666666666666667e-05,
"loss": 7.9499,
"step": 5
},
{
"epoch": 0.0670391061452514,
"grad_norm": 92.24451446533203,
"learning_rate": 2e-05,
"loss": 7.5897,
"step": 6
},
{
"epoch": 0.0782122905027933,
"grad_norm": 65.57331085205078,
"learning_rate": 2.3333333333333336e-05,
"loss": 6.5642,
"step": 7
},
{
"epoch": 0.0893854748603352,
"grad_norm": 28.319116592407227,
"learning_rate": 2.6666666666666667e-05,
"loss": 5.9872,
"step": 8
},
{
"epoch": 0.1005586592178771,
"grad_norm": 84.72834777832031,
"learning_rate": 3e-05,
"loss": 5.6777,
"step": 9
},
{
"epoch": 0.11173184357541899,
"grad_norm": 62.04957962036133,
"learning_rate": 3.3333333333333335e-05,
"loss": 5.452,
"step": 10
},
{
"epoch": 0.12290502793296089,
"grad_norm": 23.43031883239746,
"learning_rate": 3.6666666666666666e-05,
"loss": 5.0966,
"step": 11
},
{
"epoch": 0.1340782122905028,
"grad_norm": 11.170083045959473,
"learning_rate": 4e-05,
"loss": 4.9892,
"step": 12
},
{
"epoch": 0.1452513966480447,
"grad_norm": 9.359170913696289,
"learning_rate": 4.3333333333333334e-05,
"loss": 4.8074,
"step": 13
},
{
"epoch": 0.1564245810055866,
"grad_norm": 7.227917194366455,
"learning_rate": 4.666666666666667e-05,
"loss": 4.5527,
"step": 14
},
{
"epoch": 0.16759776536312848,
"grad_norm": 5.746039867401123,
"learning_rate": 5e-05,
"loss": 4.4023,
"step": 15
},
{
"epoch": 0.1787709497206704,
"grad_norm": 5.6375274658203125,
"learning_rate": 5.333333333333333e-05,
"loss": 4.2542,
"step": 16
},
{
"epoch": 0.18994413407821228,
"grad_norm": 5.626579761505127,
"learning_rate": 5.666666666666667e-05,
"loss": 4.205,
"step": 17
},
{
"epoch": 0.2011173184357542,
"grad_norm": 6.251049041748047,
"learning_rate": 6e-05,
"loss": 4.1443,
"step": 18
},
{
"epoch": 0.2122905027932961,
"grad_norm": 5.178546905517578,
"learning_rate": 6.333333333333333e-05,
"loss": 3.7947,
"step": 19
},
{
"epoch": 0.22346368715083798,
"grad_norm": 5.792481899261475,
"learning_rate": 6.666666666666667e-05,
"loss": 3.6496,
"step": 20
},
{
"epoch": 0.2346368715083799,
"grad_norm": 8.204561233520508,
"learning_rate": 7e-05,
"loss": 3.4046,
"step": 21
},
{
"epoch": 0.24581005586592178,
"grad_norm": 6.862538814544678,
"learning_rate": 7.333333333333333e-05,
"loss": 3.2466,
"step": 22
},
{
"epoch": 0.2569832402234637,
"grad_norm": 6.1007490158081055,
"learning_rate": 7.666666666666667e-05,
"loss": 3.1117,
"step": 23
},
{
"epoch": 0.2681564245810056,
"grad_norm": 4.2520365715026855,
"learning_rate": 8e-05,
"loss": 3.1683,
"step": 24
},
{
"epoch": 0.27932960893854747,
"grad_norm": 16.54338264465332,
"learning_rate": 8.333333333333334e-05,
"loss": 3.0407,
"step": 25
},
{
"epoch": 0.2905027932960894,
"grad_norm": 5.3217034339904785,
"learning_rate": 8.666666666666667e-05,
"loss": 2.8435,
"step": 26
},
{
"epoch": 0.3016759776536313,
"grad_norm": 5.066921234130859,
"learning_rate": 9e-05,
"loss": 2.8639,
"step": 27
},
{
"epoch": 0.3128491620111732,
"grad_norm": 4.023467063903809,
"learning_rate": 9.333333333333334e-05,
"loss": 2.7216,
"step": 28
},
{
"epoch": 0.3240223463687151,
"grad_norm": 5.351987361907959,
"learning_rate": 9.666666666666667e-05,
"loss": 2.503,
"step": 29
},
{
"epoch": 0.33519553072625696,
"grad_norm": 10.897500038146973,
"learning_rate": 0.0001,
"loss": 2.5283,
"step": 30
},
{
"epoch": 0.3463687150837989,
"grad_norm": 3.5467631816864014,
"learning_rate": 9.999568045802217e-05,
"loss": 2.4514,
"step": 31
},
{
"epoch": 0.3575418994413408,
"grad_norm": 4.198688507080078,
"learning_rate": 9.998272257842641e-05,
"loss": 2.6802,
"step": 32
},
{
"epoch": 0.3687150837988827,
"grad_norm": 2.829094409942627,
"learning_rate": 9.996112860009688e-05,
"loss": 2.359,
"step": 33
},
{
"epoch": 0.37988826815642457,
"grad_norm": 2.836956262588501,
"learning_rate": 9.993090225407743e-05,
"loss": 2.1674,
"step": 34
},
{
"epoch": 0.39106145251396646,
"grad_norm": 3.0385210514068604,
"learning_rate": 9.989204876292688e-05,
"loss": 2.1289,
"step": 35
},
{
"epoch": 0.4022346368715084,
"grad_norm": 3.1165988445281982,
"learning_rate": 9.984457483981669e-05,
"loss": 2.0466,
"step": 36
},
{
"epoch": 0.4134078212290503,
"grad_norm": 2.5134077072143555,
"learning_rate": 9.978848868737098e-05,
"loss": 2.0756,
"step": 37
},
{
"epoch": 0.4245810055865922,
"grad_norm": 2.617610454559326,
"learning_rate": 9.972379999624936e-05,
"loss": 1.8898,
"step": 38
},
{
"epoch": 0.43575418994413406,
"grad_norm": 2.49375581741333,
"learning_rate": 9.96505199434725e-05,
"loss": 1.8914,
"step": 39
},
{
"epoch": 0.44692737430167595,
"grad_norm": 2.5499541759490967,
"learning_rate": 9.956866119049095e-05,
"loss": 1.74,
"step": 40
},
{
"epoch": 0.4581005586592179,
"grad_norm": 2.702899217605591,
"learning_rate": 9.947823788099753e-05,
"loss": 1.6642,
"step": 41
},
{
"epoch": 0.4692737430167598,
"grad_norm": 2.9417200088500977,
"learning_rate": 9.937926563848346e-05,
"loss": 1.7848,
"step": 42
},
{
"epoch": 0.48044692737430167,
"grad_norm": 2.8817129135131836,
"learning_rate": 9.927176156353899e-05,
"loss": 1.7686,
"step": 43
},
{
"epoch": 0.49162011173184356,
"grad_norm": 2.706786632537842,
"learning_rate": 9.91557442308987e-05,
"loss": 1.5918,
"step": 44
},
{
"epoch": 0.5027932960893855,
"grad_norm": 4.0628342628479,
"learning_rate": 9.903123368623216e-05,
"loss": 2.0585,
"step": 45
},
{
"epoch": 0.5139664804469274,
"grad_norm": 2.9532899856567383,
"learning_rate": 9.889825144268029e-05,
"loss": 1.9343,
"step": 46
},
{
"epoch": 0.5251396648044693,
"grad_norm": 2.189305305480957,
"learning_rate": 9.875682047713846e-05,
"loss": 1.9791,
"step": 47
},
{
"epoch": 0.5363128491620112,
"grad_norm": 2.71520733833313,
"learning_rate": 9.860696522628639e-05,
"loss": 2.0729,
"step": 48
},
{
"epoch": 0.547486033519553,
"grad_norm": 2.677870988845825,
"learning_rate": 9.844871158236591e-05,
"loss": 1.8758,
"step": 49
},
{
"epoch": 0.5586592178770949,
"grad_norm": 2.1260600090026855,
"learning_rate": 9.828208688870735e-05,
"loss": 1.8366,
"step": 50
},
{
"epoch": 0.5586592178770949,
"eval_loss": 1.7909045219421387,
"eval_runtime": 3.541,
"eval_samples_per_second": 42.643,
"eval_steps_per_second": 21.463,
"step": 50
},
{
"epoch": 0.5698324022346368,
"grad_norm": 1.9885830879211426,
"learning_rate": 9.810711993500507e-05,
"loss": 1.7649,
"step": 51
},
{
"epoch": 0.5810055865921788,
"grad_norm": 1.8383235931396484,
"learning_rate": 9.792384095234313e-05,
"loss": 1.6706,
"step": 52
},
{
"epoch": 0.5921787709497207,
"grad_norm": 2.144374370574951,
"learning_rate": 9.773228160797188e-05,
"loss": 1.5816,
"step": 53
},
{
"epoch": 0.6033519553072626,
"grad_norm": 1.8016228675842285,
"learning_rate": 9.753247499983649e-05,
"loss": 1.7056,
"step": 54
},
{
"epoch": 0.6145251396648045,
"grad_norm": 1.934396505355835,
"learning_rate": 9.732445565085824e-05,
"loss": 1.8101,
"step": 55
},
{
"epoch": 0.6256983240223464,
"grad_norm": 1.7126082181930542,
"learning_rate": 9.71082595029695e-05,
"loss": 1.7412,
"step": 56
},
{
"epoch": 0.6368715083798883,
"grad_norm": 2.053506851196289,
"learning_rate": 9.688392391090373e-05,
"loss": 1.6484,
"step": 57
},
{
"epoch": 0.6480446927374302,
"grad_norm": 1.8453326225280762,
"learning_rate": 9.665148763574123e-05,
"loss": 1.7627,
"step": 58
},
{
"epoch": 0.659217877094972,
"grad_norm": 1.7451444864273071,
"learning_rate": 9.64109908382119e-05,
"loss": 1.4411,
"step": 59
},
{
"epoch": 0.6703910614525139,
"grad_norm": 1.7417210340499878,
"learning_rate": 9.616247507175623e-05,
"loss": 1.546,
"step": 60
},
{
"epoch": 0.6815642458100558,
"grad_norm": 2.221048355102539,
"learning_rate": 9.590598327534564e-05,
"loss": 1.7249,
"step": 61
},
{
"epoch": 0.6927374301675978,
"grad_norm": 2.038527727127075,
"learning_rate": 9.564155976606339e-05,
"loss": 1.8133,
"step": 62
},
{
"epoch": 0.7039106145251397,
"grad_norm": 2.2194125652313232,
"learning_rate": 9.536925023144742e-05,
"loss": 1.7806,
"step": 63
},
{
"epoch": 0.7150837988826816,
"grad_norm": 1.9220144748687744,
"learning_rate": 9.508910172159635e-05,
"loss": 1.5446,
"step": 64
},
{
"epoch": 0.7262569832402235,
"grad_norm": 1.841299057006836,
"learning_rate": 9.480116264104011e-05,
"loss": 1.4487,
"step": 65
},
{
"epoch": 0.7374301675977654,
"grad_norm": 2.3878183364868164,
"learning_rate": 9.450548274037653e-05,
"loss": 1.4077,
"step": 66
},
{
"epoch": 0.7486033519553073,
"grad_norm": 2.5445780754089355,
"learning_rate": 9.420211310767533e-05,
"loss": 1.8375,
"step": 67
},
{
"epoch": 0.7597765363128491,
"grad_norm": 1.9894568920135498,
"learning_rate": 9.389110615965102e-05,
"loss": 1.5472,
"step": 68
},
{
"epoch": 0.770949720670391,
"grad_norm": 1.5543500185012817,
"learning_rate": 9.35725156326063e-05,
"loss": 1.5146,
"step": 69
},
{
"epoch": 0.7821229050279329,
"grad_norm": 1.6577750444412231,
"learning_rate": 9.324639657314742e-05,
"loss": 1.5323,
"step": 70
},
{
"epoch": 0.7932960893854749,
"grad_norm": 1.6345881223678589,
"learning_rate": 9.291280532867302e-05,
"loss": 1.6763,
"step": 71
},
{
"epoch": 0.8044692737430168,
"grad_norm": 1.576483130455017,
"learning_rate": 9.257179953763845e-05,
"loss": 1.6414,
"step": 72
},
{
"epoch": 0.8156424581005587,
"grad_norm": 1.661361813545227,
"learning_rate": 9.222343811959693e-05,
"loss": 1.6203,
"step": 73
},
{
"epoch": 0.8268156424581006,
"grad_norm": 1.7419025897979736,
"learning_rate": 9.186778126501916e-05,
"loss": 1.4892,
"step": 74
},
{
"epoch": 0.8379888268156425,
"grad_norm": 1.5462729930877686,
"learning_rate": 9.150489042489367e-05,
"loss": 1.6116,
"step": 75
},
{
"epoch": 0.8491620111731844,
"grad_norm": 1.5897197723388672,
"learning_rate": 9.113482830010918e-05,
"loss": 1.5627,
"step": 76
},
{
"epoch": 0.8603351955307262,
"grad_norm": 1.6474636793136597,
"learning_rate": 9.075765883062093e-05,
"loss": 1.5695,
"step": 77
},
{
"epoch": 0.8715083798882681,
"grad_norm": 1.5242871046066284,
"learning_rate": 9.037344718440322e-05,
"loss": 1.5686,
"step": 78
},
{
"epoch": 0.88268156424581,
"grad_norm": 1.4273405075073242,
"learning_rate": 8.99822597461894e-05,
"loss": 1.2552,
"step": 79
},
{
"epoch": 0.8938547486033519,
"grad_norm": 1.63188636302948,
"learning_rate": 8.958416410600187e-05,
"loss": 1.379,
"step": 80
},
{
"epoch": 0.9050279329608939,
"grad_norm": 1.5915851593017578,
"learning_rate": 8.917922904747384e-05,
"loss": 1.5448,
"step": 81
},
{
"epoch": 0.9162011173184358,
"grad_norm": 1.4518282413482666,
"learning_rate": 8.876752453596462e-05,
"loss": 1.3964,
"step": 82
},
{
"epoch": 0.9273743016759777,
"grad_norm": 1.5592926740646362,
"learning_rate": 8.834912170647101e-05,
"loss": 1.3952,
"step": 83
},
{
"epoch": 0.9385474860335196,
"grad_norm": 1.836943507194519,
"learning_rate": 8.792409285133642e-05,
"loss": 1.542,
"step": 84
},
{
"epoch": 0.9497206703910615,
"grad_norm": 1.6657090187072754,
"learning_rate": 8.749251140776016e-05,
"loss": 1.4903,
"step": 85
},
{
"epoch": 0.9608938547486033,
"grad_norm": 1.8024177551269531,
"learning_rate": 8.705445194510868e-05,
"loss": 1.4487,
"step": 86
},
{
"epoch": 0.9720670391061452,
"grad_norm": 1.9545961618423462,
"learning_rate": 8.66099901520315e-05,
"loss": 1.5293,
"step": 87
},
{
"epoch": 0.9832402234636871,
"grad_norm": 2.1410531997680664,
"learning_rate": 8.615920282338355e-05,
"loss": 1.4279,
"step": 88
},
{
"epoch": 0.994413407821229,
"grad_norm": 1.8293300867080688,
"learning_rate": 8.570216784695637e-05,
"loss": 1.5349,
"step": 89
},
{
"epoch": 1.005586592178771,
"grad_norm": 2.8246774673461914,
"learning_rate": 8.52389641900206e-05,
"loss": 2.6159,
"step": 90
},
{
"epoch": 1.0167597765363128,
"grad_norm": 1.4292484521865845,
"learning_rate": 8.476967188568188e-05,
"loss": 1.6759,
"step": 91
},
{
"epoch": 1.0279329608938548,
"grad_norm": 1.2621123790740967,
"learning_rate": 8.429437201905254e-05,
"loss": 1.4449,
"step": 92
},
{
"epoch": 1.0391061452513966,
"grad_norm": 1.4738688468933105,
"learning_rate": 8.381314671324159e-05,
"loss": 1.4352,
"step": 93
},
{
"epoch": 1.0502793296089385,
"grad_norm": 1.4630380868911743,
"learning_rate": 8.332607911516545e-05,
"loss": 1.4337,
"step": 94
},
{
"epoch": 1.0614525139664805,
"grad_norm": 1.4351942539215088,
"learning_rate": 8.283325338118153e-05,
"loss": 1.4363,
"step": 95
},
{
"epoch": 1.0726256983240223,
"grad_norm": 1.1830869913101196,
"learning_rate": 8.233475466254765e-05,
"loss": 1.4274,
"step": 96
},
{
"epoch": 1.0837988826815643,
"grad_norm": 1.3410512208938599,
"learning_rate": 8.183066909070947e-05,
"loss": 1.3677,
"step": 97
},
{
"epoch": 1.094972067039106,
"grad_norm": 1.4120603799819946,
"learning_rate": 8.132108376241849e-05,
"loss": 1.3283,
"step": 98
},
{
"epoch": 1.106145251396648,
"grad_norm": 1.2132238149642944,
"learning_rate": 8.08060867246834e-05,
"loss": 1.3961,
"step": 99
},
{
"epoch": 1.1173184357541899,
"grad_norm": 1.466185450553894,
"learning_rate": 8.028576695955711e-05,
"loss": 1.3326,
"step": 100
},
{
"epoch": 1.1173184357541899,
"eval_loss": 1.4376939535140991,
"eval_runtime": 3.541,
"eval_samples_per_second": 42.644,
"eval_steps_per_second": 21.463,
"step": 100
},
{
"epoch": 1.1284916201117319,
"grad_norm": 1.4509429931640625,
"learning_rate": 7.97602143687623e-05,
"loss": 1.3226,
"step": 101
},
{
"epoch": 1.1396648044692737,
"grad_norm": 1.453986406326294,
"learning_rate": 7.922951975815811e-05,
"loss": 1.2928,
"step": 102
},
{
"epoch": 1.1508379888268156,
"grad_norm": 1.3562222719192505,
"learning_rate": 7.869377482205042e-05,
"loss": 1.2992,
"step": 103
},
{
"epoch": 1.1620111731843576,
"grad_norm": 1.3154160976409912,
"learning_rate": 7.815307212734888e-05,
"loss": 1.2855,
"step": 104
},
{
"epoch": 1.1731843575418994,
"grad_norm": 1.438166856765747,
"learning_rate": 7.760750509757298e-05,
"loss": 1.1333,
"step": 105
},
{
"epoch": 1.1843575418994414,
"grad_norm": 1.4575860500335693,
"learning_rate": 7.705716799671019e-05,
"loss": 1.2916,
"step": 106
},
{
"epoch": 1.1955307262569832,
"grad_norm": 1.525395154953003,
"learning_rate": 7.650215591292888e-05,
"loss": 1.2428,
"step": 107
},
{
"epoch": 1.2067039106145252,
"grad_norm": 1.7797048091888428,
"learning_rate": 7.594256474214882e-05,
"loss": 1.3893,
"step": 108
},
{
"epoch": 1.217877094972067,
"grad_norm": 1.5947117805480957,
"learning_rate": 7.537849117147212e-05,
"loss": 1.3861,
"step": 109
},
{
"epoch": 1.229050279329609,
"grad_norm": 1.948651671409607,
"learning_rate": 7.481003266247744e-05,
"loss": 1.3417,
"step": 110
},
{
"epoch": 1.2402234636871508,
"grad_norm": 1.8325766324996948,
"learning_rate": 7.423728743438048e-05,
"loss": 1.361,
"step": 111
},
{
"epoch": 1.2513966480446927,
"grad_norm": 1.443632960319519,
"learning_rate": 7.366035444706347e-05,
"loss": 1.3431,
"step": 112
},
{
"epoch": 1.2625698324022347,
"grad_norm": 1.3943829536437988,
"learning_rate": 7.307933338397667e-05,
"loss": 1.3545,
"step": 113
},
{
"epoch": 1.2737430167597765,
"grad_norm": 1.4311888217926025,
"learning_rate": 7.249432463491498e-05,
"loss": 1.3241,
"step": 114
},
{
"epoch": 1.2849162011173183,
"grad_norm": 1.2513173818588257,
"learning_rate": 7.190542927867234e-05,
"loss": 1.4081,
"step": 115
},
{
"epoch": 1.2960893854748603,
"grad_norm": 1.2229642868041992,
"learning_rate": 7.131274906557725e-05,
"loss": 1.5218,
"step": 116
},
{
"epoch": 1.3072625698324023,
"grad_norm": 1.3140783309936523,
"learning_rate": 7.071638639991207e-05,
"loss": 1.4195,
"step": 117
},
{
"epoch": 1.318435754189944,
"grad_norm": 1.3129864931106567,
"learning_rate": 7.011644432221958e-05,
"loss": 1.4635,
"step": 118
},
{
"epoch": 1.329608938547486,
"grad_norm": 1.3215774297714233,
"learning_rate": 6.95130264914993e-05,
"loss": 1.447,
"step": 119
},
{
"epoch": 1.3407821229050279,
"grad_norm": 1.2216203212738037,
"learning_rate": 6.890623716729724e-05,
"loss": 1.1941,
"step": 120
},
{
"epoch": 1.3519553072625698,
"grad_norm": 1.3087091445922852,
"learning_rate": 6.82961811916917e-05,
"loss": 1.4038,
"step": 121
},
{
"epoch": 1.3631284916201118,
"grad_norm": 1.1932587623596191,
"learning_rate": 6.768296397117848e-05,
"loss": 1.3638,
"step": 122
},
{
"epoch": 1.3743016759776536,
"grad_norm": 1.3358490467071533,
"learning_rate": 6.706669145845863e-05,
"loss": 1.3817,
"step": 123
},
{
"epoch": 1.3854748603351954,
"grad_norm": 1.1948127746582031,
"learning_rate": 6.644747013413168e-05,
"loss": 1.2415,
"step": 124
},
{
"epoch": 1.3966480446927374,
"grad_norm": 1.1789873838424683,
"learning_rate": 6.582540698829781e-05,
"loss": 1.0901,
"step": 125
},
{
"epoch": 1.4078212290502794,
"grad_norm": 1.2974153757095337,
"learning_rate": 6.520060950207185e-05,
"loss": 1.2798,
"step": 126
},
{
"epoch": 1.4189944134078212,
"grad_norm": 1.408875584602356,
"learning_rate": 6.457318562901256e-05,
"loss": 1.3745,
"step": 127
},
{
"epoch": 1.4301675977653632,
"grad_norm": 1.4454888105392456,
"learning_rate": 6.394324377647028e-05,
"loss": 1.3028,
"step": 128
},
{
"epoch": 1.441340782122905,
"grad_norm": 1.549099087715149,
"learning_rate": 6.331089278685599e-05,
"loss": 1.3293,
"step": 129
},
{
"epoch": 1.452513966480447,
"grad_norm": 1.4389153718948364,
"learning_rate": 6.26762419188355e-05,
"loss": 1.1637,
"step": 130
},
{
"epoch": 1.463687150837989,
"grad_norm": 1.408833622932434,
"learning_rate": 6.203940082845144e-05,
"loss": 1.1842,
"step": 131
},
{
"epoch": 1.4748603351955307,
"grad_norm": 1.5742685794830322,
"learning_rate": 6.140047955017671e-05,
"loss": 1.2201,
"step": 132
},
{
"epoch": 1.4860335195530725,
"grad_norm": 1.8778325319290161,
"learning_rate": 6.075958847790262e-05,
"loss": 1.1914,
"step": 133
},
{
"epoch": 1.4972067039106145,
"grad_norm": 1.5008203983306885,
"learning_rate": 6.011683834586473e-05,
"loss": 1.2746,
"step": 134
},
{
"epoch": 1.5083798882681565,
"grad_norm": 1.3712409734725952,
"learning_rate": 5.947234020951015e-05,
"loss": 1.244,
"step": 135
},
{
"epoch": 1.5195530726256983,
"grad_norm": 1.298775315284729,
"learning_rate": 5.882620542630901e-05,
"loss": 1.213,
"step": 136
},
{
"epoch": 1.5307262569832403,
"grad_norm": 1.2305636405944824,
"learning_rate": 5.8178545636514145e-05,
"loss": 1.3607,
"step": 137
},
{
"epoch": 1.541899441340782,
"grad_norm": 1.1469392776489258,
"learning_rate": 5.752947274387147e-05,
"loss": 1.349,
"step": 138
},
{
"epoch": 1.553072625698324,
"grad_norm": 1.1323195695877075,
"learning_rate": 5.687909889628529e-05,
"loss": 1.3132,
"step": 139
},
{
"epoch": 1.564245810055866,
"grad_norm": 1.0949897766113281,
"learning_rate": 5.622753646644102e-05,
"loss": 1.2518,
"step": 140
},
{
"epoch": 1.5754189944134078,
"grad_norm": 1.1683642864227295,
"learning_rate": 5.557489803238933e-05,
"loss": 1.3319,
"step": 141
},
{
"epoch": 1.5865921787709496,
"grad_norm": 1.2218724489212036,
"learning_rate": 5.492129635809473e-05,
"loss": 1.3427,
"step": 142
},
{
"epoch": 1.5977653631284916,
"grad_norm": 1.133186936378479,
"learning_rate": 5.426684437395196e-05,
"loss": 1.2725,
"step": 143
},
{
"epoch": 1.6089385474860336,
"grad_norm": 1.2256780862808228,
"learning_rate": 5.361165515727374e-05,
"loss": 1.2708,
"step": 144
},
{
"epoch": 1.6201117318435754,
"grad_norm": 1.2726577520370483,
"learning_rate": 5.295584191275308e-05,
"loss": 1.3795,
"step": 145
},
{
"epoch": 1.6312849162011172,
"grad_norm": 1.2547286748886108,
"learning_rate": 5.229951795290353e-05,
"loss": 1.3061,
"step": 146
},
{
"epoch": 1.6424581005586592,
"grad_norm": 1.2261368036270142,
"learning_rate": 5.164279667848094e-05,
"loss": 1.245,
"step": 147
},
{
"epoch": 1.6536312849162011,
"grad_norm": 1.2689229249954224,
"learning_rate": 5.0985791558889785e-05,
"loss": 1.3172,
"step": 148
},
{
"epoch": 1.6648044692737431,
"grad_norm": 1.2573802471160889,
"learning_rate": 5.032861611257783e-05,
"loss": 1.2967,
"step": 149
},
{
"epoch": 1.675977653631285,
"grad_norm": 1.4018381834030151,
"learning_rate": 4.967138388742218e-05,
"loss": 1.2735,
"step": 150
},
{
"epoch": 1.675977653631285,
"eval_loss": 1.326945424079895,
"eval_runtime": 3.5376,
"eval_samples_per_second": 42.685,
"eval_steps_per_second": 21.484,
"step": 150
},
{
"epoch": 1.6871508379888267,
"grad_norm": 1.5333133935928345,
"learning_rate": 4.901420844111021e-05,
"loss": 1.3852,
"step": 151
},
{
"epoch": 1.6983240223463687,
"grad_norm": 1.3856674432754517,
"learning_rate": 4.835720332151907e-05,
"loss": 1.1126,
"step": 152
},
{
"epoch": 1.7094972067039107,
"grad_norm": 1.6033830642700195,
"learning_rate": 4.770048204709648e-05,
"loss": 1.3011,
"step": 153
},
{
"epoch": 1.7206703910614525,
"grad_norm": 1.6162670850753784,
"learning_rate": 4.7044158087246926e-05,
"loss": 1.2179,
"step": 154
},
{
"epoch": 1.7318435754189943,
"grad_norm": 1.744236946105957,
"learning_rate": 4.6388344842726264e-05,
"loss": 1.1941,
"step": 155
},
{
"epoch": 1.7430167597765363,
"grad_norm": 1.5789906978607178,
"learning_rate": 4.5733155626048036e-05,
"loss": 1.3045,
"step": 156
},
{
"epoch": 1.7541899441340782,
"grad_norm": 1.226691722869873,
"learning_rate": 4.507870364190527e-05,
"loss": 1.3675,
"step": 157
},
{
"epoch": 1.7653631284916202,
"grad_norm": 1.3142743110656738,
"learning_rate": 4.4425101967610674e-05,
"loss": 1.4387,
"step": 158
},
{
"epoch": 1.776536312849162,
"grad_norm": 1.0727818012237549,
"learning_rate": 4.377246353355899e-05,
"loss": 1.2175,
"step": 159
},
{
"epoch": 1.7877094972067038,
"grad_norm": 1.1169428825378418,
"learning_rate": 4.312090110371473e-05,
"loss": 1.365,
"step": 160
},
{
"epoch": 1.7988826815642458,
"grad_norm": 1.0804024934768677,
"learning_rate": 4.247052725612852e-05,
"loss": 1.3312,
"step": 161
},
{
"epoch": 1.8100558659217878,
"grad_norm": 1.2123122215270996,
"learning_rate": 4.1821454363485866e-05,
"loss": 1.3837,
"step": 162
},
{
"epoch": 1.8212290502793296,
"grad_norm": 1.1932836771011353,
"learning_rate": 4.1173794573690996e-05,
"loss": 1.3426,
"step": 163
},
{
"epoch": 1.8324022346368714,
"grad_norm": 1.0667084455490112,
"learning_rate": 4.052765979048986e-05,
"loss": 1.223,
"step": 164
},
{
"epoch": 1.8435754189944134,
"grad_norm": 1.1312137842178345,
"learning_rate": 3.988316165413528e-05,
"loss": 1.2159,
"step": 165
},
{
"epoch": 1.8547486033519553,
"grad_norm": 1.135401725769043,
"learning_rate": 3.924041152209739e-05,
"loss": 1.1931,
"step": 166
},
{
"epoch": 1.8659217877094973,
"grad_norm": 1.2866289615631104,
"learning_rate": 3.859952044982329e-05,
"loss": 1.3266,
"step": 167
},
{
"epoch": 1.8770949720670391,
"grad_norm": 1.3166481256484985,
"learning_rate": 3.7960599171548574e-05,
"loss": 1.2229,
"step": 168
},
{
"epoch": 1.888268156424581,
"grad_norm": 1.2265543937683105,
"learning_rate": 3.732375808116451e-05,
"loss": 1.2899,
"step": 169
},
{
"epoch": 1.899441340782123,
"grad_norm": 1.2559617757797241,
"learning_rate": 3.668910721314402e-05,
"loss": 1.13,
"step": 170
},
{
"epoch": 1.910614525139665,
"grad_norm": 1.2486090660095215,
"learning_rate": 3.605675622352973e-05,
"loss": 1.2083,
"step": 171
},
{
"epoch": 1.9217877094972067,
"grad_norm": 1.406841516494751,
"learning_rate": 3.542681437098745e-05,
"loss": 1.2352,
"step": 172
},
{
"epoch": 1.9329608938547485,
"grad_norm": 1.4469358921051025,
"learning_rate": 3.479939049792817e-05,
"loss": 1.2662,
"step": 173
},
{
"epoch": 1.9441340782122905,
"grad_norm": 1.4758639335632324,
"learning_rate": 3.417459301170219e-05,
"loss": 1.2432,
"step": 174
},
{
"epoch": 1.9553072625698324,
"grad_norm": 1.3574011325836182,
"learning_rate": 3.355252986586832e-05,
"loss": 1.1044,
"step": 175
},
{
"epoch": 1.9664804469273744,
"grad_norm": 1.8598283529281616,
"learning_rate": 3.293330854154136e-05,
"loss": 1.3518,
"step": 176
},
{
"epoch": 1.9776536312849162,
"grad_norm": 1.7371652126312256,
"learning_rate": 3.2317036028821523e-05,
"loss": 1.2855,
"step": 177
},
{
"epoch": 1.988826815642458,
"grad_norm": 1.4092578887939453,
"learning_rate": 3.1703818808308324e-05,
"loss": 1.3558,
"step": 178
},
{
"epoch": 2.0,
"grad_norm": 2.122119188308716,
"learning_rate": 3.109376283270277e-05,
"loss": 1.6615,
"step": 179
},
{
"epoch": 2.011173184357542,
"grad_norm": 1.219295859336853,
"learning_rate": 3.0486973508500727e-05,
"loss": 1.4017,
"step": 180
},
{
"epoch": 2.022346368715084,
"grad_norm": 1.1538641452789307,
"learning_rate": 2.988355567778043e-05,
"loss": 1.2616,
"step": 181
},
{
"epoch": 2.0335195530726256,
"grad_norm": 1.0700401067733765,
"learning_rate": 2.9283613600087933e-05,
"loss": 1.2648,
"step": 182
},
{
"epoch": 2.0446927374301676,
"grad_norm": 0.9880838394165039,
"learning_rate": 2.8687250934422772e-05,
"loss": 1.1885,
"step": 183
},
{
"epoch": 2.0558659217877095,
"grad_norm": 1.051222801208496,
"learning_rate": 2.8094570721327662e-05,
"loss": 1.1746,
"step": 184
},
{
"epoch": 2.0670391061452515,
"grad_norm": 1.0639920234680176,
"learning_rate": 2.750567536508504e-05,
"loss": 1.2354,
"step": 185
},
{
"epoch": 2.078212290502793,
"grad_norm": 1.0196352005004883,
"learning_rate": 2.6920666616023327e-05,
"loss": 1.2534,
"step": 186
},
{
"epoch": 2.089385474860335,
"grad_norm": 1.0681345462799072,
"learning_rate": 2.6339645552936536e-05,
"loss": 1.1096,
"step": 187
},
{
"epoch": 2.100558659217877,
"grad_norm": 1.196368932723999,
"learning_rate": 2.5762712565619528e-05,
"loss": 1.1021,
"step": 188
},
{
"epoch": 2.111731843575419,
"grad_norm": 1.0771485567092896,
"learning_rate": 2.5189967337522573e-05,
"loss": 1.0798,
"step": 189
},
{
"epoch": 2.122905027932961,
"grad_norm": 1.11617910861969,
"learning_rate": 2.46215088285279e-05,
"loss": 1.1074,
"step": 190
},
{
"epoch": 2.1340782122905027,
"grad_norm": 1.1497374773025513,
"learning_rate": 2.4057435257851175e-05,
"loss": 1.1183,
"step": 191
},
{
"epoch": 2.1452513966480447,
"grad_norm": 1.1393492221832275,
"learning_rate": 2.349784408707112e-05,
"loss": 1.0711,
"step": 192
},
{
"epoch": 2.1564245810055866,
"grad_norm": 1.0959515571594238,
"learning_rate": 2.2942832003289823e-05,
"loss": 1.0437,
"step": 193
},
{
"epoch": 2.1675977653631286,
"grad_norm": 1.155680775642395,
"learning_rate": 2.2392494902427025e-05,
"loss": 1.047,
"step": 194
},
{
"epoch": 2.17877094972067,
"grad_norm": 1.157484531402588,
"learning_rate": 2.1846927872651137e-05,
"loss": 1.0649,
"step": 195
},
{
"epoch": 2.189944134078212,
"grad_norm": 1.2944828271865845,
"learning_rate": 2.1306225177949585e-05,
"loss": 1.1103,
"step": 196
},
{
"epoch": 2.201117318435754,
"grad_norm": 1.5556319952011108,
"learning_rate": 2.07704802418419e-05,
"loss": 1.2251,
"step": 197
},
{
"epoch": 2.212290502793296,
"grad_norm": 1.4562413692474365,
"learning_rate": 2.0239785631237705e-05,
"loss": 1.0627,
"step": 198
},
{
"epoch": 2.223463687150838,
"grad_norm": 1.393690824508667,
"learning_rate": 1.9714233040442915e-05,
"loss": 0.9858,
"step": 199
},
{
"epoch": 2.2346368715083798,
"grad_norm": 1.6887412071228027,
"learning_rate": 1.9193913275316626e-05,
"loss": 1.1138,
"step": 200
},
{
"epoch": 2.2346368715083798,
"eval_loss": 1.2823649644851685,
"eval_runtime": 3.5435,
"eval_samples_per_second": 42.613,
"eval_steps_per_second": 21.448,
"step": 200
},
{
"epoch": 2.2458100558659218,
"grad_norm": 2.0551750659942627,
"learning_rate": 1.8678916237581522e-05,
"loss": 1.0606,
"step": 201
},
{
"epoch": 2.2569832402234637,
"grad_norm": 1.3848985433578491,
"learning_rate": 1.816933090929055e-05,
"loss": 1.2715,
"step": 202
},
{
"epoch": 2.2681564245810057,
"grad_norm": 1.1580418348312378,
"learning_rate": 1.7665245337452368e-05,
"loss": 1.1072,
"step": 203
},
{
"epoch": 2.2793296089385473,
"grad_norm": 1.1287815570831299,
"learning_rate": 1.716674661881848e-05,
"loss": 1.1187,
"step": 204
},
{
"epoch": 2.2905027932960893,
"grad_norm": 1.1857539415359497,
"learning_rate": 1.667392088483456e-05,
"loss": 1.3102,
"step": 205
},
{
"epoch": 2.3016759776536313,
"grad_norm": 1.0973467826843262,
"learning_rate": 1.6186853286758397e-05,
"loss": 1.2696,
"step": 206
},
{
"epoch": 2.3128491620111733,
"grad_norm": 1.0390197038650513,
"learning_rate": 1.570562798094747e-05,
"loss": 1.2218,
"step": 207
},
{
"epoch": 2.3240223463687153,
"grad_norm": 1.2470310926437378,
"learning_rate": 1.5230328114318127e-05,
"loss": 1.3041,
"step": 208
},
{
"epoch": 2.335195530726257,
"grad_norm": 1.1841509342193604,
"learning_rate": 1.4761035809979395e-05,
"loss": 1.2008,
"step": 209
},
{
"epoch": 2.346368715083799,
"grad_norm": 1.027030110359192,
"learning_rate": 1.4297832153043656e-05,
"loss": 1.0652,
"step": 210
},
{
"epoch": 2.357541899441341,
"grad_norm": 1.1887884140014648,
"learning_rate": 1.3840797176616466e-05,
"loss": 1.1927,
"step": 211
},
{
"epoch": 2.368715083798883,
"grad_norm": 1.1091079711914062,
"learning_rate": 1.3390009847968504e-05,
"loss": 1.157,
"step": 212
},
{
"epoch": 2.3798882681564244,
"grad_norm": 1.1894493103027344,
"learning_rate": 1.2945548054891321e-05,
"loss": 1.1287,
"step": 213
},
{
"epoch": 2.3910614525139664,
"grad_norm": 1.2934393882751465,
"learning_rate": 1.2507488592239847e-05,
"loss": 1.127,
"step": 214
},
{
"epoch": 2.4022346368715084,
"grad_norm": 1.1570240259170532,
"learning_rate": 1.2075907148663579e-05,
"loss": 1.0553,
"step": 215
},
{
"epoch": 2.4134078212290504,
"grad_norm": 1.172692060470581,
"learning_rate": 1.1650878293528994e-05,
"loss": 1.0155,
"step": 216
},
{
"epoch": 2.4245810055865924,
"grad_norm": 1.197474718093872,
"learning_rate": 1.1232475464035385e-05,
"loss": 1.135,
"step": 217
},
{
"epoch": 2.435754189944134,
"grad_norm": 1.1875431537628174,
"learning_rate": 1.0820770952526155e-05,
"loss": 1.1129,
"step": 218
},
{
"epoch": 2.446927374301676,
"grad_norm": 1.4438897371292114,
"learning_rate": 1.0415835893998116e-05,
"loss": 1.2491,
"step": 219
},
{
"epoch": 2.458100558659218,
"grad_norm": 1.3826245069503784,
"learning_rate": 1.0017740253810609e-05,
"loss": 1.1198,
"step": 220
},
{
"epoch": 2.46927374301676,
"grad_norm": 1.3730559349060059,
"learning_rate": 9.62655281559679e-06,
"loss": 1.0396,
"step": 221
},
{
"epoch": 2.4804469273743015,
"grad_norm": 1.6272450685501099,
"learning_rate": 9.242341169379076e-06,
"loss": 1.1171,
"step": 222
},
{
"epoch": 2.4916201117318435,
"grad_norm": 1.9734584093093872,
"learning_rate": 8.865171699890834e-06,
"loss": 1.1374,
"step": 223
},
{
"epoch": 2.5027932960893855,
"grad_norm": 1.3264243602752686,
"learning_rate": 8.49510957510633e-06,
"loss": 1.3275,
"step": 224
},
{
"epoch": 2.5139664804469275,
"grad_norm": 1.3713715076446533,
"learning_rate": 8.132218734980852e-06,
"loss": 1.2514,
"step": 225
},
{
"epoch": 2.5251396648044695,
"grad_norm": 1.2266292572021484,
"learning_rate": 7.776561880403072e-06,
"loss": 1.1443,
"step": 226
},
{
"epoch": 2.536312849162011,
"grad_norm": 1.2601666450500488,
"learning_rate": 7.4282004623615396e-06,
"loss": 1.2684,
"step": 227
},
{
"epoch": 2.547486033519553,
"grad_norm": 1.222358226776123,
"learning_rate": 7.0871946713269856e-06,
"loss": 1.1953,
"step": 228
},
{
"epoch": 2.558659217877095,
"grad_norm": 1.1266566514968872,
"learning_rate": 6.753603426852589e-06,
"loss": 1.1542,
"step": 229
},
{
"epoch": 2.5698324022346366,
"grad_norm": 1.035483479499817,
"learning_rate": 6.427484367393699e-06,
"loss": 1.1903,
"step": 230
},
{
"epoch": 2.5810055865921786,
"grad_norm": 1.0014451742172241,
"learning_rate": 6.108893840348995e-06,
"loss": 1.0823,
"step": 231
},
{
"epoch": 2.5921787709497206,
"grad_norm": 1.1314361095428467,
"learning_rate": 5.797886892324694e-06,
"loss": 1.2311,
"step": 232
},
{
"epoch": 2.6033519553072626,
"grad_norm": 1.0532621145248413,
"learning_rate": 5.494517259623477e-06,
"loss": 1.1181,
"step": 233
},
{
"epoch": 2.6145251396648046,
"grad_norm": 1.0458093881607056,
"learning_rate": 5.198837358959901e-06,
"loss": 1.1474,
"step": 234
},
{
"epoch": 2.6256983240223466,
"grad_norm": 1.1469138860702515,
"learning_rate": 4.910898278403669e-06,
"loss": 1.1677,
"step": 235
},
{
"epoch": 2.636871508379888,
"grad_norm": 1.135291337966919,
"learning_rate": 4.630749768552589e-06,
"loss": 1.1351,
"step": 236
},
{
"epoch": 2.64804469273743,
"grad_norm": 1.09344482421875,
"learning_rate": 4.358440233936617e-06,
"loss": 1.1221,
"step": 237
},
{
"epoch": 2.659217877094972,
"grad_norm": 1.190507411956787,
"learning_rate": 4.094016724654359e-06,
"loss": 1.0893,
"step": 238
},
{
"epoch": 2.6703910614525137,
"grad_norm": 1.1962109804153442,
"learning_rate": 3.837524928243774e-06,
"loss": 1.0827,
"step": 239
},
{
"epoch": 2.6815642458100557,
"grad_norm": 1.3094158172607422,
"learning_rate": 3.589009161788104e-06,
"loss": 1.0457,
"step": 240
},
{
"epoch": 2.6927374301675977,
"grad_norm": 1.3989758491516113,
"learning_rate": 3.3485123642587658e-06,
"loss": 1.0812,
"step": 241
},
{
"epoch": 2.7039106145251397,
"grad_norm": 1.4350848197937012,
"learning_rate": 3.116076089096265e-06,
"loss": 1.0437,
"step": 242
},
{
"epoch": 2.7150837988826817,
"grad_norm": 1.4926385879516602,
"learning_rate": 2.8917404970305097e-06,
"loss": 1.0832,
"step": 243
},
{
"epoch": 2.7262569832402237,
"grad_norm": 1.7393081188201904,
"learning_rate": 2.675544349141779e-06,
"loss": 1.0668,
"step": 244
},
{
"epoch": 2.7374301675977653,
"grad_norm": 1.9382210969924927,
"learning_rate": 2.4675250001635232e-06,
"loss": 1.0107,
"step": 245
},
{
"epoch": 2.7486033519553073,
"grad_norm": 0.7568252086639404,
"learning_rate": 2.2677183920281343e-06,
"loss": 1.0928,
"step": 246
},
{
"epoch": 2.7597765363128492,
"grad_norm": 0.9040629267692566,
"learning_rate": 2.076159047656889e-06,
"loss": 1.3023,
"step": 247
},
{
"epoch": 2.770949720670391,
"grad_norm": 0.9287111163139343,
"learning_rate": 1.892880064994934e-06,
"loss": 1.1539,
"step": 248
},
{
"epoch": 2.782122905027933,
"grad_norm": 0.9362537860870361,
"learning_rate": 1.7179131112926627e-06,
"loss": 1.1163,
"step": 249
},
{
"epoch": 2.793296089385475,
"grad_norm": 0.9989909529685974,
"learning_rate": 1.551288417634106e-06,
"loss": 1.1343,
"step": 250
},
{
"epoch": 2.793296089385475,
"eval_loss": 1.2524751424789429,
"eval_runtime": 3.5444,
"eval_samples_per_second": 42.603,
"eval_steps_per_second": 21.442,
"step": 250
},
{
"epoch": 2.804469273743017,
"grad_norm": 0.999933660030365,
"learning_rate": 1.3930347737136196e-06,
"loss": 1.1728,
"step": 251
},
{
"epoch": 2.815642458100559,
"grad_norm": 1.084299921989441,
"learning_rate": 1.2431795228615372e-06,
"loss": 1.1947,
"step": 252
},
{
"epoch": 2.826815642458101,
"grad_norm": 1.0853486061096191,
"learning_rate": 1.101748557319715e-06,
"loss": 1.1514,
"step": 253
},
{
"epoch": 2.8379888268156424,
"grad_norm": 1.0919440984725952,
"learning_rate": 9.687663137678604e-07,
"loss": 1.1536,
"step": 254
},
{
"epoch": 2.8491620111731844,
"grad_norm": 1.0306580066680908,
"learning_rate": 8.442557691013043e-07,
"loss": 1.1416,
"step": 255
},
{
"epoch": 2.8603351955307263,
"grad_norm": 1.1208820343017578,
"learning_rate": 7.282384364610206e-07,
"loss": 1.1873,
"step": 256
},
{
"epoch": 2.871508379888268,
"grad_norm": 1.0364559888839722,
"learning_rate": 6.207343615165561e-07,
"loss": 1.0184,
"step": 257
},
{
"epoch": 2.88268156424581,
"grad_norm": 0.9816321730613708,
"learning_rate": 5.217621190024779e-07,
"loss": 1.017,
"step": 258
},
{
"epoch": 2.893854748603352,
"grad_norm": 1.1331318616867065,
"learning_rate": 4.3133880950905205e-07,
"loss": 1.1131,
"step": 259
},
{
"epoch": 2.905027932960894,
"grad_norm": 1.209434151649475,
"learning_rate": 3.494800565275125e-07,
"loss": 1.18,
"step": 260
},
{
"epoch": 2.916201117318436,
"grad_norm": 1.2905470132827759,
"learning_rate": 2.762000037506485e-07,
"loss": 1.1171,
"step": 261
},
{
"epoch": 2.927374301675978,
"grad_norm": 1.216124176979065,
"learning_rate": 2.115113126290258e-07,
"loss": 1.0804,
"step": 262
},
{
"epoch": 2.9385474860335195,
"grad_norm": 1.300574541091919,
"learning_rate": 1.554251601833201e-07,
"loss": 1.0849,
"step": 263
},
{
"epoch": 2.9497206703910615,
"grad_norm": 1.376492977142334,
"learning_rate": 1.0795123707312283e-07,
"loss": 1.0931,
"step": 264
},
{
"epoch": 2.9608938547486034,
"grad_norm": 1.507287859916687,
"learning_rate": 6.909774592258056e-08,
"loss": 1.1502,
"step": 265
},
{
"epoch": 2.972067039106145,
"grad_norm": 1.6715766191482544,
"learning_rate": 3.8871399903134265e-08,
"loss": 1.0179,
"step": 266
},
{
"epoch": 2.983240223463687,
"grad_norm": 2.2043421268463135,
"learning_rate": 1.7277421573608232e-08,
"loss": 0.9667,
"step": 267
},
{
"epoch": 2.994413407821229,
"grad_norm": 1.0078221559524536,
"learning_rate": 4.319541977831909e-09,
"loss": 1.1912,
"step": 268
},
{
"epoch": 3.005586592178771,
"grad_norm": 2.012565851211548,
"learning_rate": 0.0,
"loss": 1.8972,
"step": 269
}
],
"logging_steps": 1,
"max_steps": 269,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 50,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 5,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.738167723150541e+16,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}