ciloku's picture
Training in progress, step 200, checkpoint
61d839f verified
{
"best_metric": 0.5063894391059875,
"best_model_checkpoint": "miner_id_24/checkpoint-200",
"epoch": 0.07582938388625593,
"eval_steps": 50,
"global_step": 200,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0003791469194312796,
"grad_norm": 2.3607397079467773,
"learning_rate": 6e-06,
"loss": 2.3674,
"step": 1
},
{
"epoch": 0.0003791469194312796,
"eval_loss": 1.6981829404830933,
"eval_runtime": 342.6691,
"eval_samples_per_second": 12.963,
"eval_steps_per_second": 3.242,
"step": 1
},
{
"epoch": 0.0007582938388625592,
"grad_norm": 2.2280433177948,
"learning_rate": 1.2e-05,
"loss": 1.7743,
"step": 2
},
{
"epoch": 0.001137440758293839,
"grad_norm": 2.1979401111602783,
"learning_rate": 1.8e-05,
"loss": 1.7584,
"step": 3
},
{
"epoch": 0.0015165876777251184,
"grad_norm": 2.395792245864868,
"learning_rate": 2.4e-05,
"loss": 1.7522,
"step": 4
},
{
"epoch": 0.0018957345971563982,
"grad_norm": 2.4676456451416016,
"learning_rate": 3e-05,
"loss": 1.6792,
"step": 5
},
{
"epoch": 0.002274881516587678,
"grad_norm": 1.9544316530227661,
"learning_rate": 3.6e-05,
"loss": 1.5587,
"step": 6
},
{
"epoch": 0.002654028436018957,
"grad_norm": 1.9189141988754272,
"learning_rate": 4.2e-05,
"loss": 1.5178,
"step": 7
},
{
"epoch": 0.003033175355450237,
"grad_norm": 1.8870714902877808,
"learning_rate": 4.8e-05,
"loss": 1.4159,
"step": 8
},
{
"epoch": 0.0034123222748815166,
"grad_norm": 2.4351065158843994,
"learning_rate": 5.4000000000000005e-05,
"loss": 1.297,
"step": 9
},
{
"epoch": 0.0037914691943127963,
"grad_norm": 2.2076759338378906,
"learning_rate": 6e-05,
"loss": 1.1759,
"step": 10
},
{
"epoch": 0.004170616113744076,
"grad_norm": 1.7516471147537231,
"learning_rate": 5.999589914977407e-05,
"loss": 1.1303,
"step": 11
},
{
"epoch": 0.004549763033175356,
"grad_norm": 1.5598973035812378,
"learning_rate": 5.998359772022778e-05,
"loss": 0.9447,
"step": 12
},
{
"epoch": 0.0049289099526066355,
"grad_norm": 1.4637537002563477,
"learning_rate": 5.996309907444915e-05,
"loss": 0.9049,
"step": 13
},
{
"epoch": 0.005308056872037914,
"grad_norm": 1.9019067287445068,
"learning_rate": 5.9934408816563236e-05,
"loss": 0.8778,
"step": 14
},
{
"epoch": 0.005687203791469194,
"grad_norm": 1.5424095392227173,
"learning_rate": 5.98975347902001e-05,
"loss": 0.8591,
"step": 15
},
{
"epoch": 0.006066350710900474,
"grad_norm": 1.3776594400405884,
"learning_rate": 5.9852487076350345e-05,
"loss": 0.7328,
"step": 16
},
{
"epoch": 0.0064454976303317535,
"grad_norm": 1.3162641525268555,
"learning_rate": 5.979927799060915e-05,
"loss": 0.7089,
"step": 17
},
{
"epoch": 0.006824644549763033,
"grad_norm": 1.2267054319381714,
"learning_rate": 5.9737922079809257e-05,
"loss": 0.7679,
"step": 18
},
{
"epoch": 0.007203791469194313,
"grad_norm": 1.1532204151153564,
"learning_rate": 5.9668436118044054e-05,
"loss": 0.7228,
"step": 19
},
{
"epoch": 0.007582938388625593,
"grad_norm": 1.3049921989440918,
"learning_rate": 5.959083910208167e-05,
"loss": 0.7155,
"step": 20
},
{
"epoch": 0.007962085308056872,
"grad_norm": 1.3345766067504883,
"learning_rate": 5.9505152246171474e-05,
"loss": 0.7048,
"step": 21
},
{
"epoch": 0.008341232227488152,
"grad_norm": 1.353661060333252,
"learning_rate": 5.941139897624428e-05,
"loss": 0.77,
"step": 22
},
{
"epoch": 0.008720379146919432,
"grad_norm": 1.2893695831298828,
"learning_rate": 5.9309604923507984e-05,
"loss": 0.7045,
"step": 23
},
{
"epoch": 0.009099526066350712,
"grad_norm": 1.139123558998108,
"learning_rate": 5.9199797917440176e-05,
"loss": 0.6403,
"step": 24
},
{
"epoch": 0.009478672985781991,
"grad_norm": 1.1754416227340698,
"learning_rate": 5.908200797817991e-05,
"loss": 0.6316,
"step": 25
},
{
"epoch": 0.009857819905213271,
"grad_norm": 1.3844693899154663,
"learning_rate": 5.895626730832046e-05,
"loss": 0.6302,
"step": 26
},
{
"epoch": 0.010236966824644549,
"grad_norm": 1.0756479501724243,
"learning_rate": 5.882261028410545e-05,
"loss": 0.5767,
"step": 27
},
{
"epoch": 0.010616113744075829,
"grad_norm": 1.1990379095077515,
"learning_rate": 5.8681073446030734e-05,
"loss": 0.5942,
"step": 28
},
{
"epoch": 0.010995260663507108,
"grad_norm": 1.2510521411895752,
"learning_rate": 5.853169548885461e-05,
"loss": 0.6293,
"step": 29
},
{
"epoch": 0.011374407582938388,
"grad_norm": 1.0987434387207031,
"learning_rate": 5.8374517251019035e-05,
"loss": 0.5444,
"step": 30
},
{
"epoch": 0.011753554502369668,
"grad_norm": 0.9962146878242493,
"learning_rate": 5.820958170348484e-05,
"loss": 0.5284,
"step": 31
},
{
"epoch": 0.012132701421800948,
"grad_norm": 1.1842890977859497,
"learning_rate": 5.8036933937983825e-05,
"loss": 0.5764,
"step": 32
},
{
"epoch": 0.012511848341232227,
"grad_norm": 1.1667747497558594,
"learning_rate": 5.7856621154691217e-05,
"loss": 0.6086,
"step": 33
},
{
"epoch": 0.012890995260663507,
"grad_norm": 1.1322537660598755,
"learning_rate": 5.766869264932154e-05,
"loss": 0.5379,
"step": 34
},
{
"epoch": 0.013270142180094787,
"grad_norm": 1.07042396068573,
"learning_rate": 5.747319979965172e-05,
"loss": 0.5195,
"step": 35
},
{
"epoch": 0.013649289099526066,
"grad_norm": 1.023470163345337,
"learning_rate": 5.727019605147488e-05,
"loss": 0.543,
"step": 36
},
{
"epoch": 0.014028436018957346,
"grad_norm": 1.1579279899597168,
"learning_rate": 5.7059736903988775e-05,
"loss": 0.5042,
"step": 37
},
{
"epoch": 0.014407582938388626,
"grad_norm": 1.0734813213348389,
"learning_rate": 5.684187989462291e-05,
"loss": 0.5276,
"step": 38
},
{
"epoch": 0.014786729857819906,
"grad_norm": 1.0260837078094482,
"learning_rate": 5.661668458330836e-05,
"loss": 0.4359,
"step": 39
},
{
"epoch": 0.015165876777251185,
"grad_norm": 0.9480069279670715,
"learning_rate": 5.638421253619467e-05,
"loss": 0.437,
"step": 40
},
{
"epoch": 0.015545023696682465,
"grad_norm": 1.0021203756332397,
"learning_rate": 5.614452730881832e-05,
"loss": 0.487,
"step": 41
},
{
"epoch": 0.015924170616113745,
"grad_norm": 0.9898553490638733,
"learning_rate": 5.589769442872722e-05,
"loss": 0.4045,
"step": 42
},
{
"epoch": 0.016303317535545023,
"grad_norm": 1.1742738485336304,
"learning_rate": 5.5643781377566175e-05,
"loss": 0.4673,
"step": 43
},
{
"epoch": 0.016682464454976304,
"grad_norm": 1.0911322832107544,
"learning_rate": 5.538285757262806e-05,
"loss": 0.4501,
"step": 44
},
{
"epoch": 0.017061611374407582,
"grad_norm": 1.2655229568481445,
"learning_rate": 5.5114994347875856e-05,
"loss": 0.4084,
"step": 45
},
{
"epoch": 0.017440758293838864,
"grad_norm": 1.3887434005737305,
"learning_rate": 5.48402649344406e-05,
"loss": 0.3981,
"step": 46
},
{
"epoch": 0.01781990521327014,
"grad_norm": 1.1302635669708252,
"learning_rate": 5.455874444060078e-05,
"loss": 0.4061,
"step": 47
},
{
"epoch": 0.018199052132701423,
"grad_norm": 0.8546876907348633,
"learning_rate": 5.427050983124843e-05,
"loss": 0.3365,
"step": 48
},
{
"epoch": 0.0185781990521327,
"grad_norm": 0.8889405727386475,
"learning_rate": 5.397563990684774e-05,
"loss": 0.2457,
"step": 49
},
{
"epoch": 0.018957345971563982,
"grad_norm": 0.9020243883132935,
"learning_rate": 5.367421528189181e-05,
"loss": 0.2258,
"step": 50
},
{
"epoch": 0.018957345971563982,
"eval_loss": 0.6159882545471191,
"eval_runtime": 344.9982,
"eval_samples_per_second": 12.875,
"eval_steps_per_second": 3.22,
"step": 50
},
{
"epoch": 0.01933649289099526,
"grad_norm": 2.4133408069610596,
"learning_rate": 5.336631836286338e-05,
"loss": 1.5293,
"step": 51
},
{
"epoch": 0.019715639810426542,
"grad_norm": 1.7940583229064941,
"learning_rate": 5.3052033325705774e-05,
"loss": 0.8329,
"step": 52
},
{
"epoch": 0.02009478672985782,
"grad_norm": 1.4519822597503662,
"learning_rate": 5.2731446092810044e-05,
"loss": 0.7793,
"step": 53
},
{
"epoch": 0.020473933649289098,
"grad_norm": 1.110809326171875,
"learning_rate": 5.240464430952462e-05,
"loss": 0.7678,
"step": 54
},
{
"epoch": 0.02085308056872038,
"grad_norm": 1.0020638704299927,
"learning_rate": 5.207171732019395e-05,
"loss": 0.7016,
"step": 55
},
{
"epoch": 0.021232227488151657,
"grad_norm": 1.0127681493759155,
"learning_rate": 5.1732756143732675e-05,
"loss": 0.6799,
"step": 56
},
{
"epoch": 0.02161137440758294,
"grad_norm": 1.004103422164917,
"learning_rate": 5.1387853448741916e-05,
"loss": 0.7143,
"step": 57
},
{
"epoch": 0.021990521327014217,
"grad_norm": 1.1212753057479858,
"learning_rate": 5.103710352817465e-05,
"loss": 0.6442,
"step": 58
},
{
"epoch": 0.022369668246445498,
"grad_norm": 1.13973069190979,
"learning_rate": 5.068060227355698e-05,
"loss": 0.7217,
"step": 59
},
{
"epoch": 0.022748815165876776,
"grad_norm": 1.1548775434494019,
"learning_rate": 5.0318447148772234e-05,
"loss": 0.708,
"step": 60
},
{
"epoch": 0.023127962085308058,
"grad_norm": 1.039237380027771,
"learning_rate": 4.995073716341545e-05,
"loss": 0.6506,
"step": 61
},
{
"epoch": 0.023507109004739336,
"grad_norm": 1.0963629484176636,
"learning_rate": 4.957757284572506e-05,
"loss": 0.6685,
"step": 62
},
{
"epoch": 0.023886255924170617,
"grad_norm": 1.0616074800491333,
"learning_rate": 4.91990562150995e-05,
"loss": 0.6461,
"step": 63
},
{
"epoch": 0.024265402843601895,
"grad_norm": 1.0263690948486328,
"learning_rate": 4.881529075420611e-05,
"loss": 0.5949,
"step": 64
},
{
"epoch": 0.024644549763033177,
"grad_norm": 1.1550185680389404,
"learning_rate": 4.8426381380690036e-05,
"loss": 0.6779,
"step": 65
},
{
"epoch": 0.025023696682464455,
"grad_norm": 0.9933828115463257,
"learning_rate": 4.8032434418490753e-05,
"loss": 0.6456,
"step": 66
},
{
"epoch": 0.025402843601895736,
"grad_norm": 1.00188410282135,
"learning_rate": 4.7633557568774194e-05,
"loss": 0.6354,
"step": 67
},
{
"epoch": 0.025781990521327014,
"grad_norm": 1.0373916625976562,
"learning_rate": 4.722985988048831e-05,
"loss": 0.6082,
"step": 68
},
{
"epoch": 0.026161137440758295,
"grad_norm": 0.9275899529457092,
"learning_rate": 4.6821451720550184e-05,
"loss": 0.6274,
"step": 69
},
{
"epoch": 0.026540284360189573,
"grad_norm": 0.9165945649147034,
"learning_rate": 4.640844474367282e-05,
"loss": 0.6005,
"step": 70
},
{
"epoch": 0.02691943127962085,
"grad_norm": 0.9279446601867676,
"learning_rate": 4.5990951861839815e-05,
"loss": 0.5644,
"step": 71
},
{
"epoch": 0.027298578199052133,
"grad_norm": 0.856342077255249,
"learning_rate": 4.5569087213436455e-05,
"loss": 0.5951,
"step": 72
},
{
"epoch": 0.02767772511848341,
"grad_norm": 0.95192950963974,
"learning_rate": 4.514296613204532e-05,
"loss": 0.5506,
"step": 73
},
{
"epoch": 0.028056872037914692,
"grad_norm": 1.0360265970230103,
"learning_rate": 4.471270511491525e-05,
"loss": 0.64,
"step": 74
},
{
"epoch": 0.02843601895734597,
"grad_norm": 0.9761951565742493,
"learning_rate": 4.427842179111221e-05,
"loss": 0.5912,
"step": 75
},
{
"epoch": 0.02881516587677725,
"grad_norm": 0.8985703587532043,
"learning_rate": 4.3840234889360634e-05,
"loss": 0.5406,
"step": 76
},
{
"epoch": 0.02919431279620853,
"grad_norm": 0.9850453734397888,
"learning_rate": 4.33982642055842e-05,
"loss": 0.6044,
"step": 77
},
{
"epoch": 0.02957345971563981,
"grad_norm": 0.9527453184127808,
"learning_rate": 4.2952630570154785e-05,
"loss": 0.5498,
"step": 78
},
{
"epoch": 0.02995260663507109,
"grad_norm": 0.8898606300354004,
"learning_rate": 4.250345581485871e-05,
"loss": 0.498,
"step": 79
},
{
"epoch": 0.03033175355450237,
"grad_norm": 1.0251396894454956,
"learning_rate": 4.205086273958909e-05,
"loss": 0.5551,
"step": 80
},
{
"epoch": 0.03071090047393365,
"grad_norm": 0.9887294173240662,
"learning_rate": 4.1594975078773565e-05,
"loss": 0.5348,
"step": 81
},
{
"epoch": 0.03109004739336493,
"grad_norm": 0.9571895599365234,
"learning_rate": 4.113591746754662e-05,
"loss": 0.5147,
"step": 82
},
{
"epoch": 0.03146919431279621,
"grad_norm": 0.8705043792724609,
"learning_rate": 4.06738154076755e-05,
"loss": 0.4836,
"step": 83
},
{
"epoch": 0.03184834123222749,
"grad_norm": 0.9828000068664551,
"learning_rate": 4.020879523324929e-05,
"loss": 0.5008,
"step": 84
},
{
"epoch": 0.03222748815165877,
"grad_norm": 0.9241380095481873,
"learning_rate": 3.974098407614051e-05,
"loss": 0.4675,
"step": 85
},
{
"epoch": 0.032606635071090045,
"grad_norm": 0.9122922420501709,
"learning_rate": 3.927050983124842e-05,
"loss": 0.5142,
"step": 86
},
{
"epoch": 0.03298578199052133,
"grad_norm": 0.8403096795082092,
"learning_rate": 3.8797501121533946e-05,
"loss": 0.4439,
"step": 87
},
{
"epoch": 0.03336492890995261,
"grad_norm": 0.910892128944397,
"learning_rate": 3.832208726285534e-05,
"loss": 0.4519,
"step": 88
},
{
"epoch": 0.033744075829383886,
"grad_norm": 0.9453192353248596,
"learning_rate": 3.784439822861459e-05,
"loss": 0.4743,
"step": 89
},
{
"epoch": 0.034123222748815164,
"grad_norm": 0.9362481832504272,
"learning_rate": 3.7364564614223976e-05,
"loss": 0.4588,
"step": 90
},
{
"epoch": 0.03450236966824644,
"grad_norm": 0.8357151746749878,
"learning_rate": 3.688271760140255e-05,
"loss": 0.3989,
"step": 91
},
{
"epoch": 0.03488151658767773,
"grad_norm": 0.9535480737686157,
"learning_rate": 3.6398988922312406e-05,
"loss": 0.4292,
"step": 92
},
{
"epoch": 0.035260663507109005,
"grad_norm": 0.8818041086196899,
"learning_rate": 3.591351082354441e-05,
"loss": 0.376,
"step": 93
},
{
"epoch": 0.03563981042654028,
"grad_norm": 0.9949309229850769,
"learning_rate": 3.54264160299633e-05,
"loss": 0.4139,
"step": 94
},
{
"epoch": 0.03601895734597156,
"grad_norm": 0.8851547837257385,
"learning_rate": 3.493783770842202e-05,
"loss": 0.3293,
"step": 95
},
{
"epoch": 0.036398104265402846,
"grad_norm": 0.8010683059692383,
"learning_rate": 3.444790943135526e-05,
"loss": 0.3311,
"step": 96
},
{
"epoch": 0.036777251184834124,
"grad_norm": 0.7428768873214722,
"learning_rate": 3.3956765140262074e-05,
"loss": 0.2844,
"step": 97
},
{
"epoch": 0.0371563981042654,
"grad_norm": 0.9548898339271545,
"learning_rate": 3.346453910908759e-05,
"loss": 0.376,
"step": 98
},
{
"epoch": 0.03753554502369668,
"grad_norm": 1.0641423463821411,
"learning_rate": 3.297136590751389e-05,
"loss": 0.3601,
"step": 99
},
{
"epoch": 0.037914691943127965,
"grad_norm": 0.7730882167816162,
"learning_rate": 3.247738036416998e-05,
"loss": 0.2464,
"step": 100
},
{
"epoch": 0.037914691943127965,
"eval_loss": 0.5508739948272705,
"eval_runtime": 345.5738,
"eval_samples_per_second": 12.854,
"eval_steps_per_second": 3.215,
"step": 100
},
{
"epoch": 0.03829383886255924,
"grad_norm": 1.8822741508483887,
"learning_rate": 3.1982717529770985e-05,
"loss": 1.4074,
"step": 101
},
{
"epoch": 0.03867298578199052,
"grad_norm": 1.272026777267456,
"learning_rate": 3.148751264019667e-05,
"loss": 0.7856,
"step": 102
},
{
"epoch": 0.0390521327014218,
"grad_norm": 1.161899209022522,
"learning_rate": 3.099190107951924e-05,
"loss": 0.6956,
"step": 103
},
{
"epoch": 0.039431279620853084,
"grad_norm": 1.0137699842453003,
"learning_rate": 3.049601834299076e-05,
"loss": 0.6313,
"step": 104
},
{
"epoch": 0.03981042654028436,
"grad_norm": 1.0051027536392212,
"learning_rate": 3e-05,
"loss": 0.6899,
"step": 105
},
{
"epoch": 0.04018957345971564,
"grad_norm": 0.9395660758018494,
"learning_rate": 2.9503981657009246e-05,
"loss": 0.6345,
"step": 106
},
{
"epoch": 0.04056872037914692,
"grad_norm": 0.8711168766021729,
"learning_rate": 2.9008098920480752e-05,
"loss": 0.6447,
"step": 107
},
{
"epoch": 0.040947867298578196,
"grad_norm": 1.1823228597640991,
"learning_rate": 2.851248735980333e-05,
"loss": 0.663,
"step": 108
},
{
"epoch": 0.04132701421800948,
"grad_norm": 0.9468877911567688,
"learning_rate": 2.801728247022902e-05,
"loss": 0.6448,
"step": 109
},
{
"epoch": 0.04170616113744076,
"grad_norm": 0.8840280175209045,
"learning_rate": 2.7522619635830034e-05,
"loss": 0.6168,
"step": 110
},
{
"epoch": 0.04208530805687204,
"grad_norm": 0.9660087823867798,
"learning_rate": 2.702863409248612e-05,
"loss": 0.6257,
"step": 111
},
{
"epoch": 0.042464454976303315,
"grad_norm": 0.9464119076728821,
"learning_rate": 2.6535460890912416e-05,
"loss": 0.6864,
"step": 112
},
{
"epoch": 0.0428436018957346,
"grad_norm": 1.0040414333343506,
"learning_rate": 2.604323485973793e-05,
"loss": 0.6249,
"step": 113
},
{
"epoch": 0.04322274881516588,
"grad_norm": 0.9637667536735535,
"learning_rate": 2.555209056864474e-05,
"loss": 0.5605,
"step": 114
},
{
"epoch": 0.043601895734597156,
"grad_norm": 0.9878002405166626,
"learning_rate": 2.5062162291577978e-05,
"loss": 0.536,
"step": 115
},
{
"epoch": 0.043981042654028434,
"grad_norm": 1.0215932130813599,
"learning_rate": 2.4573583970036712e-05,
"loss": 0.6054,
"step": 116
},
{
"epoch": 0.04436018957345972,
"grad_norm": 0.9616793394088745,
"learning_rate": 2.4086489176455595e-05,
"loss": 0.5458,
"step": 117
},
{
"epoch": 0.044739336492890996,
"grad_norm": 0.9715977311134338,
"learning_rate": 2.36010110776876e-05,
"loss": 0.5747,
"step": 118
},
{
"epoch": 0.045118483412322274,
"grad_norm": 0.9184060096740723,
"learning_rate": 2.3117282398597456e-05,
"loss": 0.5391,
"step": 119
},
{
"epoch": 0.04549763033175355,
"grad_norm": 0.9710631370544434,
"learning_rate": 2.263543538577603e-05,
"loss": 0.5382,
"step": 120
},
{
"epoch": 0.04587677725118484,
"grad_norm": 0.8977413773536682,
"learning_rate": 2.215560177138541e-05,
"loss": 0.5466,
"step": 121
},
{
"epoch": 0.046255924170616115,
"grad_norm": 0.9749202728271484,
"learning_rate": 2.167791273714467e-05,
"loss": 0.5813,
"step": 122
},
{
"epoch": 0.04663507109004739,
"grad_norm": 1.0063494443893433,
"learning_rate": 2.1202498878466062e-05,
"loss": 0.5904,
"step": 123
},
{
"epoch": 0.04701421800947867,
"grad_norm": 1.0267138481140137,
"learning_rate": 2.072949016875158e-05,
"loss": 0.5809,
"step": 124
},
{
"epoch": 0.04739336492890995,
"grad_norm": 0.916971743106842,
"learning_rate": 2.0259015923859498e-05,
"loss": 0.5226,
"step": 125
},
{
"epoch": 0.047772511848341234,
"grad_norm": 0.9448188543319702,
"learning_rate": 1.979120476675071e-05,
"loss": 0.5503,
"step": 126
},
{
"epoch": 0.04815165876777251,
"grad_norm": 0.8960833549499512,
"learning_rate": 1.9326184592324503e-05,
"loss": 0.5472,
"step": 127
},
{
"epoch": 0.04853080568720379,
"grad_norm": 0.9310248494148254,
"learning_rate": 1.8864082532453373e-05,
"loss": 0.554,
"step": 128
},
{
"epoch": 0.04890995260663507,
"grad_norm": 0.8872618675231934,
"learning_rate": 1.840502492122644e-05,
"loss": 0.5129,
"step": 129
},
{
"epoch": 0.04928909952606635,
"grad_norm": 0.9954771399497986,
"learning_rate": 1.7949137260410924e-05,
"loss": 0.5348,
"step": 130
},
{
"epoch": 0.04966824644549763,
"grad_norm": 0.8588917255401611,
"learning_rate": 1.7496544185141295e-05,
"loss": 0.4947,
"step": 131
},
{
"epoch": 0.05004739336492891,
"grad_norm": 0.8972111940383911,
"learning_rate": 1.7047369429845216e-05,
"loss": 0.5559,
"step": 132
},
{
"epoch": 0.05042654028436019,
"grad_norm": 0.8581748008728027,
"learning_rate": 1.6601735794415806e-05,
"loss": 0.4764,
"step": 133
},
{
"epoch": 0.05080568720379147,
"grad_norm": 0.8218798041343689,
"learning_rate": 1.615976511063937e-05,
"loss": 0.4472,
"step": 134
},
{
"epoch": 0.05118483412322275,
"grad_norm": 0.8914713859558105,
"learning_rate": 1.5721578208887793e-05,
"loss": 0.4507,
"step": 135
},
{
"epoch": 0.05156398104265403,
"grad_norm": 0.891606330871582,
"learning_rate": 1.5287294885084766e-05,
"loss": 0.4769,
"step": 136
},
{
"epoch": 0.051943127962085306,
"grad_norm": 0.9011698365211487,
"learning_rate": 1.4857033867954697e-05,
"loss": 0.5192,
"step": 137
},
{
"epoch": 0.05232227488151659,
"grad_norm": 0.8463770151138306,
"learning_rate": 1.4430912786563554e-05,
"loss": 0.4164,
"step": 138
},
{
"epoch": 0.05270142180094787,
"grad_norm": 0.9531325101852417,
"learning_rate": 1.4009048138160195e-05,
"loss": 0.4199,
"step": 139
},
{
"epoch": 0.05308056872037915,
"grad_norm": 0.8462571501731873,
"learning_rate": 1.3591555256327199e-05,
"loss": 0.4274,
"step": 140
},
{
"epoch": 0.053459715639810425,
"grad_norm": 0.862342894077301,
"learning_rate": 1.3178548279449822e-05,
"loss": 0.3976,
"step": 141
},
{
"epoch": 0.0538388625592417,
"grad_norm": 0.7988135814666748,
"learning_rate": 1.2770140119511693e-05,
"loss": 0.3224,
"step": 142
},
{
"epoch": 0.05421800947867299,
"grad_norm": 0.8891763091087341,
"learning_rate": 1.2366442431225809e-05,
"loss": 0.4034,
"step": 143
},
{
"epoch": 0.054597156398104266,
"grad_norm": 0.9104804396629333,
"learning_rate": 1.1967565581509248e-05,
"loss": 0.3469,
"step": 144
},
{
"epoch": 0.054976303317535544,
"grad_norm": 0.8460737466812134,
"learning_rate": 1.1573618619309965e-05,
"loss": 0.323,
"step": 145
},
{
"epoch": 0.05535545023696682,
"grad_norm": 0.7780117988586426,
"learning_rate": 1.1184709245793889e-05,
"loss": 0.2783,
"step": 146
},
{
"epoch": 0.05573459715639811,
"grad_norm": 0.8993276357650757,
"learning_rate": 1.0800943784900502e-05,
"loss": 0.3748,
"step": 147
},
{
"epoch": 0.056113744075829385,
"grad_norm": 0.8467982411384583,
"learning_rate": 1.042242715427494e-05,
"loss": 0.2837,
"step": 148
},
{
"epoch": 0.05649289099526066,
"grad_norm": 0.8935648202896118,
"learning_rate": 1.004926283658455e-05,
"loss": 0.3136,
"step": 149
},
{
"epoch": 0.05687203791469194,
"grad_norm": 0.9242781400680542,
"learning_rate": 9.681552851227774e-06,
"loss": 0.2324,
"step": 150
},
{
"epoch": 0.05687203791469194,
"eval_loss": 0.5187870860099792,
"eval_runtime": 345.6495,
"eval_samples_per_second": 12.851,
"eval_steps_per_second": 3.214,
"step": 150
},
{
"epoch": 0.057251184834123225,
"grad_norm": 1.3798567056655884,
"learning_rate": 9.319397726443026e-06,
"loss": 1.0638,
"step": 151
},
{
"epoch": 0.0576303317535545,
"grad_norm": 1.1334023475646973,
"learning_rate": 8.962896471825342e-06,
"loss": 0.6878,
"step": 152
},
{
"epoch": 0.05800947867298578,
"grad_norm": 1.0488735437393188,
"learning_rate": 8.61214655125809e-06,
"loss": 0.7195,
"step": 153
},
{
"epoch": 0.05838862559241706,
"grad_norm": 1.0702718496322632,
"learning_rate": 8.267243856267331e-06,
"loss": 0.5706,
"step": 154
},
{
"epoch": 0.058767772511848344,
"grad_norm": 0.9154753088951111,
"learning_rate": 7.928282679806052e-06,
"loss": 0.5379,
"step": 155
},
{
"epoch": 0.05914691943127962,
"grad_norm": 1.022912859916687,
"learning_rate": 7.595355690475393e-06,
"loss": 0.6094,
"step": 156
},
{
"epoch": 0.0595260663507109,
"grad_norm": 1.1987192630767822,
"learning_rate": 7.268553907189964e-06,
"loss": 0.6575,
"step": 157
},
{
"epoch": 0.05990521327014218,
"grad_norm": 1.0124115943908691,
"learning_rate": 6.947966674294236e-06,
"loss": 0.6735,
"step": 158
},
{
"epoch": 0.060284360189573456,
"grad_norm": 1.0271550416946411,
"learning_rate": 6.6336816371366305e-06,
"loss": 0.5953,
"step": 159
},
{
"epoch": 0.06066350710900474,
"grad_norm": 1.0127638578414917,
"learning_rate": 6.325784718108196e-06,
"loss": 0.6548,
"step": 160
},
{
"epoch": 0.06104265402843602,
"grad_norm": 0.8849831223487854,
"learning_rate": 6.0243600931522595e-06,
"loss": 0.6348,
"step": 161
},
{
"epoch": 0.0614218009478673,
"grad_norm": 0.9687224626541138,
"learning_rate": 5.72949016875158e-06,
"loss": 0.5602,
"step": 162
},
{
"epoch": 0.061800947867298575,
"grad_norm": 0.9322991371154785,
"learning_rate": 5.44125555939923e-06,
"loss": 0.6793,
"step": 163
},
{
"epoch": 0.06218009478672986,
"grad_norm": 0.8064345717430115,
"learning_rate": 5.159735065559399e-06,
"loss": 0.4625,
"step": 164
},
{
"epoch": 0.06255924170616113,
"grad_norm": 0.9643939137458801,
"learning_rate": 4.885005652124144e-06,
"loss": 0.6543,
"step": 165
},
{
"epoch": 0.06293838862559242,
"grad_norm": 0.9320531487464905,
"learning_rate": 4.617142427371934e-06,
"loss": 0.5322,
"step": 166
},
{
"epoch": 0.0633175355450237,
"grad_norm": 0.7888116240501404,
"learning_rate": 4.3562186224338265e-06,
"loss": 0.4682,
"step": 167
},
{
"epoch": 0.06369668246445498,
"grad_norm": 0.9415287375450134,
"learning_rate": 4.102305571272783e-06,
"loss": 0.6159,
"step": 168
},
{
"epoch": 0.06407582938388626,
"grad_norm": 0.8795173168182373,
"learning_rate": 3.855472691181678e-06,
"loss": 0.5751,
"step": 169
},
{
"epoch": 0.06445497630331753,
"grad_norm": 0.9194141030311584,
"learning_rate": 3.615787463805331e-06,
"loss": 0.5376,
"step": 170
},
{
"epoch": 0.06483412322274881,
"grad_norm": 0.9153357744216919,
"learning_rate": 3.383315416691646e-06,
"loss": 0.5864,
"step": 171
},
{
"epoch": 0.06521327014218009,
"grad_norm": 0.8993094563484192,
"learning_rate": 3.158120105377096e-06,
"loss": 0.5267,
"step": 172
},
{
"epoch": 0.06559241706161137,
"grad_norm": 0.8435608744621277,
"learning_rate": 2.940263096011233e-06,
"loss": 0.5413,
"step": 173
},
{
"epoch": 0.06597156398104266,
"grad_norm": 0.9139825105667114,
"learning_rate": 2.729803948525125e-06,
"loss": 0.4994,
"step": 174
},
{
"epoch": 0.06635071090047394,
"grad_norm": 0.8998444080352783,
"learning_rate": 2.526800200348275e-06,
"loss": 0.5262,
"step": 175
},
{
"epoch": 0.06672985781990522,
"grad_norm": 0.8086129426956177,
"learning_rate": 2.3313073506784575e-06,
"loss": 0.4989,
"step": 176
},
{
"epoch": 0.0671090047393365,
"grad_norm": 0.805290937423706,
"learning_rate": 2.143378845308791e-06,
"loss": 0.4473,
"step": 177
},
{
"epoch": 0.06748815165876777,
"grad_norm": 0.8601678609848022,
"learning_rate": 1.9630660620161777e-06,
"loss": 0.4955,
"step": 178
},
{
"epoch": 0.06786729857819905,
"grad_norm": 0.875322699546814,
"learning_rate": 1.790418296515165e-06,
"loss": 0.5145,
"step": 179
},
{
"epoch": 0.06824644549763033,
"grad_norm": 0.8989061117172241,
"learning_rate": 1.625482748980961e-06,
"loss": 0.5265,
"step": 180
},
{
"epoch": 0.0686255924170616,
"grad_norm": 0.8949686884880066,
"learning_rate": 1.4683045111453942e-06,
"loss": 0.4835,
"step": 181
},
{
"epoch": 0.06900473933649288,
"grad_norm": 0.9829505681991577,
"learning_rate": 1.3189265539692707e-06,
"loss": 0.4863,
"step": 182
},
{
"epoch": 0.06938388625592418,
"grad_norm": 0.9040274024009705,
"learning_rate": 1.1773897158945557e-06,
"loss": 0.4871,
"step": 183
},
{
"epoch": 0.06976303317535545,
"grad_norm": 0.8389787077903748,
"learning_rate": 1.0437326916795432e-06,
"loss": 0.4487,
"step": 184
},
{
"epoch": 0.07014218009478673,
"grad_norm": 0.8816047310829163,
"learning_rate": 9.179920218200888e-07,
"loss": 0.4567,
"step": 185
},
{
"epoch": 0.07052132701421801,
"grad_norm": 0.951728343963623,
"learning_rate": 8.002020825598277e-07,
"loss": 0.4308,
"step": 186
},
{
"epoch": 0.07090047393364929,
"grad_norm": 0.9444287419319153,
"learning_rate": 6.90395076492022e-07,
"loss": 0.4475,
"step": 187
},
{
"epoch": 0.07127962085308057,
"grad_norm": 0.7584393620491028,
"learning_rate": 5.886010237557194e-07,
"loss": 0.3609,
"step": 188
},
{
"epoch": 0.07165876777251184,
"grad_norm": 0.8242846727371216,
"learning_rate": 4.94847753828529e-07,
"loss": 0.3971,
"step": 189
},
{
"epoch": 0.07203791469194312,
"grad_norm": 0.8676409125328064,
"learning_rate": 4.091608979183303e-07,
"loss": 0.3851,
"step": 190
},
{
"epoch": 0.07241706161137441,
"grad_norm": 0.8776127696037292,
"learning_rate": 3.315638819559452e-07,
"loss": 0.3837,
"step": 191
},
{
"epoch": 0.07279620853080569,
"grad_norm": 0.9045898914337158,
"learning_rate": 2.6207792019074414e-07,
"loss": 0.4312,
"step": 192
},
{
"epoch": 0.07317535545023697,
"grad_norm": 0.9427354335784912,
"learning_rate": 2.0072200939085573e-07,
"loss": 0.4484,
"step": 193
},
{
"epoch": 0.07355450236966825,
"grad_norm": 0.8181795477867126,
"learning_rate": 1.475129236496575e-07,
"loss": 0.3509,
"step": 194
},
{
"epoch": 0.07393364928909953,
"grad_norm": 0.7674962878227234,
"learning_rate": 1.0246520979990459e-07,
"loss": 0.3238,
"step": 195
},
{
"epoch": 0.0743127962085308,
"grad_norm": 0.8669456839561462,
"learning_rate": 6.559118343676396e-08,
"loss": 0.3703,
"step": 196
},
{
"epoch": 0.07469194312796208,
"grad_norm": 0.8156073093414307,
"learning_rate": 3.690092555085789e-08,
"loss": 0.2708,
"step": 197
},
{
"epoch": 0.07507109004739336,
"grad_norm": 0.8143777847290039,
"learning_rate": 1.640227977221853e-08,
"loss": 0.269,
"step": 198
},
{
"epoch": 0.07545023696682464,
"grad_norm": 0.8005136847496033,
"learning_rate": 4.1008502259298755e-09,
"loss": 0.2552,
"step": 199
},
{
"epoch": 0.07582938388625593,
"grad_norm": 0.886074423789978,
"learning_rate": 0.0,
"loss": 0.2134,
"step": 200
},
{
"epoch": 0.07582938388625593,
"eval_loss": 0.5063894391059875,
"eval_runtime": 344.7064,
"eval_samples_per_second": 12.886,
"eval_steps_per_second": 3.223,
"step": 200
}
],
"logging_steps": 1,
"max_steps": 200,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 50,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 4,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 3.092452836429005e+17,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}