lesso's picture
Training in progress, epoch 1, checkpoint
ceb21cd verified
raw
history blame
35.4 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0012771392081736,
"eval_steps": 500,
"global_step": 196,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.005108556832694764,
"grad_norm": 0.06296706199645996,
"learning_rate": 0.00019998715457999314,
"loss": 10.3788,
"step": 1
},
{
"epoch": 0.010217113665389528,
"grad_norm": 0.06155822426080704,
"learning_rate": 0.0001999486216200688,
"loss": 10.3876,
"step": 2
},
{
"epoch": 0.01532567049808429,
"grad_norm": 0.04243969917297363,
"learning_rate": 0.0001998844110196681,
"loss": 10.383,
"step": 3
},
{
"epoch": 0.020434227330779056,
"grad_norm": 0.08324499428272247,
"learning_rate": 0.00019979453927503364,
"loss": 10.3854,
"step": 4
},
{
"epoch": 0.02554278416347382,
"grad_norm": 0.06429007649421692,
"learning_rate": 0.00019967902947497156,
"loss": 10.3783,
"step": 5
},
{
"epoch": 0.03065134099616858,
"grad_norm": 0.06722702831029892,
"learning_rate": 0.00019953791129491983,
"loss": 10.3771,
"step": 6
},
{
"epoch": 0.035759897828863345,
"grad_norm": 0.04376779869198799,
"learning_rate": 0.00019937122098932428,
"loss": 10.3768,
"step": 7
},
{
"epoch": 0.04086845466155811,
"grad_norm": 0.057710230350494385,
"learning_rate": 0.0001991790013823246,
"loss": 10.3731,
"step": 8
},
{
"epoch": 0.04597701149425287,
"grad_norm": 0.06998121738433838,
"learning_rate": 0.00019896130185675261,
"loss": 10.3783,
"step": 9
},
{
"epoch": 0.05108556832694764,
"grad_norm": 0.06011413037776947,
"learning_rate": 0.00019871817834144504,
"loss": 10.3773,
"step": 10
},
{
"epoch": 0.0561941251596424,
"grad_norm": 0.05903960019350052,
"learning_rate": 0.00019844969329687527,
"loss": 10.379,
"step": 11
},
{
"epoch": 0.06130268199233716,
"grad_norm": 0.06023373827338219,
"learning_rate": 0.00019815591569910654,
"loss": 10.3718,
"step": 12
},
{
"epoch": 0.06641123882503193,
"grad_norm": 0.07594680041074753,
"learning_rate": 0.00019783692102207155,
"loss": 10.3802,
"step": 13
},
{
"epoch": 0.07151979565772669,
"grad_norm": 0.07722143083810806,
"learning_rate": 0.00019749279121818235,
"loss": 10.3835,
"step": 14
},
{
"epoch": 0.07662835249042145,
"grad_norm": 0.06465964019298553,
"learning_rate": 0.0001971236146972764,
"loss": 10.3811,
"step": 15
},
{
"epoch": 0.08173690932311622,
"grad_norm": 0.08145278692245483,
"learning_rate": 0.00019672948630390294,
"loss": 10.3771,
"step": 16
},
{
"epoch": 0.08684546615581099,
"grad_norm": 0.06017107516527176,
"learning_rate": 0.00019631050729295707,
"loss": 10.3816,
"step": 17
},
{
"epoch": 0.09195402298850575,
"grad_norm": 0.07251092046499252,
"learning_rate": 0.00019586678530366606,
"loss": 10.3833,
"step": 18
},
{
"epoch": 0.0970625798212005,
"grad_norm": 0.04419001191854477,
"learning_rate": 0.00019539843433193639,
"loss": 10.3781,
"step": 19
},
{
"epoch": 0.10217113665389528,
"grad_norm": 0.0600612536072731,
"learning_rate": 0.00019490557470106686,
"loss": 10.3828,
"step": 20
},
{
"epoch": 0.10727969348659004,
"grad_norm": 0.07362791895866394,
"learning_rate": 0.00019438833303083678,
"loss": 10.3844,
"step": 21
},
{
"epoch": 0.1123882503192848,
"grad_norm": 0.06240883469581604,
"learning_rate": 0.00019384684220497605,
"loss": 10.3754,
"step": 22
},
{
"epoch": 0.11749680715197956,
"grad_norm": 0.06827075779438019,
"learning_rate": 0.0001932812413370265,
"loss": 10.3727,
"step": 23
},
{
"epoch": 0.12260536398467432,
"grad_norm": 0.06609390676021576,
"learning_rate": 0.0001926916757346022,
"loss": 10.3814,
"step": 24
},
{
"epoch": 0.1277139208173691,
"grad_norm": 0.06842462718486786,
"learning_rate": 0.00019207829686205882,
"loss": 10.3716,
"step": 25
},
{
"epoch": 0.13282247765006386,
"grad_norm": 0.06660367548465729,
"learning_rate": 0.00019144126230158127,
"loss": 10.3826,
"step": 26
},
{
"epoch": 0.13793103448275862,
"grad_norm": 0.08344798535108566,
"learning_rate": 0.00019078073571269922,
"loss": 10.3783,
"step": 27
},
{
"epoch": 0.14303959131545338,
"grad_norm": 0.08380083739757538,
"learning_rate": 0.0001900968867902419,
"loss": 10.3732,
"step": 28
},
{
"epoch": 0.14814814814814814,
"grad_norm": 0.0946107804775238,
"learning_rate": 0.00018938989122074197,
"loss": 10.37,
"step": 29
},
{
"epoch": 0.1532567049808429,
"grad_norm": 0.06197074055671692,
"learning_rate": 0.00018865993063730004,
"loss": 10.3761,
"step": 30
},
{
"epoch": 0.1583652618135377,
"grad_norm": 0.0819283053278923,
"learning_rate": 0.00018790719257292174,
"loss": 10.371,
"step": 31
},
{
"epoch": 0.16347381864623245,
"grad_norm": 0.09213186800479889,
"learning_rate": 0.00018713187041233896,
"loss": 10.3725,
"step": 32
},
{
"epoch": 0.1685823754789272,
"grad_norm": 0.08193743228912354,
"learning_rate": 0.00018633416334232753,
"loss": 10.3725,
"step": 33
},
{
"epoch": 0.17369093231162197,
"grad_norm": 0.07233049720525742,
"learning_rate": 0.00018551427630053463,
"loss": 10.374,
"step": 34
},
{
"epoch": 0.17879948914431673,
"grad_norm": 0.09028832614421844,
"learning_rate": 0.00018467241992282843,
"loss": 10.3732,
"step": 35
},
{
"epoch": 0.1839080459770115,
"grad_norm": 0.10519926995038986,
"learning_rate": 0.00018380881048918405,
"loss": 10.3769,
"step": 36
},
{
"epoch": 0.18901660280970625,
"grad_norm": 0.06255685538053513,
"learning_rate": 0.0001829236698681195,
"loss": 10.378,
"step": 37
},
{
"epoch": 0.194125159642401,
"grad_norm": 0.08233755081892014,
"learning_rate": 0.0001820172254596956,
"loss": 10.3704,
"step": 38
},
{
"epoch": 0.19923371647509577,
"grad_norm": 0.05263242870569229,
"learning_rate": 0.0001810897101370951,
"loss": 10.3834,
"step": 39
},
{
"epoch": 0.20434227330779056,
"grad_norm": 0.1387779265642166,
"learning_rate": 0.00018014136218679567,
"loss": 10.3707,
"step": 40
},
{
"epoch": 0.20945083014048532,
"grad_norm": 0.08574540168046951,
"learning_rate": 0.000179172425247352,
"loss": 10.377,
"step": 41
},
{
"epoch": 0.21455938697318008,
"grad_norm": 0.06211940944194794,
"learning_rate": 0.000178183148246803,
"loss": 10.3647,
"step": 42
},
{
"epoch": 0.21966794380587484,
"grad_norm": 0.09258867800235748,
"learning_rate": 0.00017717378533872017,
"loss": 10.3722,
"step": 43
},
{
"epoch": 0.2247765006385696,
"grad_norm": 0.08623415976762772,
"learning_rate": 0.00017614459583691346,
"loss": 10.3754,
"step": 44
},
{
"epoch": 0.22988505747126436,
"grad_norm": 0.1082397997379303,
"learning_rate": 0.00017509584414881113,
"loss": 10.3736,
"step": 45
},
{
"epoch": 0.23499361430395913,
"grad_norm": 0.0896887555718422,
"learning_rate": 0.00017402779970753155,
"loss": 10.3732,
"step": 46
},
{
"epoch": 0.24010217113665389,
"grad_norm": 0.09336112439632416,
"learning_rate": 0.00017294073690266344,
"loss": 10.373,
"step": 47
},
{
"epoch": 0.24521072796934865,
"grad_norm": 0.08002498000860214,
"learning_rate": 0.00017183493500977278,
"loss": 10.3689,
"step": 48
},
{
"epoch": 0.2503192848020434,
"grad_norm": 0.07161203771829605,
"learning_rate": 0.00017071067811865476,
"loss": 10.3702,
"step": 49
},
{
"epoch": 0.2554278416347382,
"grad_norm": 0.121755450963974,
"learning_rate": 0.00016956825506034867,
"loss": 10.3821,
"step": 50
},
{
"epoch": 0.26053639846743293,
"grad_norm": 0.08657146245241165,
"learning_rate": 0.00016840795933293463,
"loss": 10.3803,
"step": 51
},
{
"epoch": 0.2656449553001277,
"grad_norm": 0.07844258099794388,
"learning_rate": 0.0001672300890261317,
"loss": 10.3719,
"step": 52
},
{
"epoch": 0.2707535121328225,
"grad_norm": 0.08680905401706696,
"learning_rate": 0.00016603494674471593,
"loss": 10.3649,
"step": 53
},
{
"epoch": 0.27586206896551724,
"grad_norm": 0.10259506851434708,
"learning_rate": 0.00016482283953077887,
"loss": 10.3725,
"step": 54
},
{
"epoch": 0.280970625798212,
"grad_norm": 0.08116184920072556,
"learning_rate": 0.00016359407878484552,
"loss": 10.3681,
"step": 55
},
{
"epoch": 0.28607918263090676,
"grad_norm": 0.06668803840875626,
"learning_rate": 0.00016234898018587337,
"loss": 10.3731,
"step": 56
},
{
"epoch": 0.29118773946360155,
"grad_norm": 0.08566579967737198,
"learning_rate": 0.00016108786361015143,
"loss": 10.37,
"step": 57
},
{
"epoch": 0.2962962962962963,
"grad_norm": 0.11228417605161667,
"learning_rate": 0.00015981105304912162,
"loss": 10.3716,
"step": 58
},
{
"epoch": 0.30140485312899107,
"grad_norm": 0.07987914979457855,
"learning_rate": 0.00015851887652614237,
"loss": 10.3712,
"step": 59
},
{
"epoch": 0.3065134099616858,
"grad_norm": 0.1201016902923584,
"learning_rate": 0.00015721166601221698,
"loss": 10.3647,
"step": 60
},
{
"epoch": 0.3116219667943806,
"grad_norm": 0.08329416066408157,
"learning_rate": 0.00015588975734070717,
"loss": 10.3792,
"step": 61
},
{
"epoch": 0.3167305236270754,
"grad_norm": 0.10588457435369492,
"learning_rate": 0.00015455349012105486,
"loss": 10.3688,
"step": 62
},
{
"epoch": 0.3218390804597701,
"grad_norm": 0.12081307172775269,
"learning_rate": 0.00015320320765153367,
"loss": 10.3642,
"step": 63
},
{
"epoch": 0.3269476372924649,
"grad_norm": 0.08195498585700989,
"learning_rate": 0.00015183925683105254,
"loss": 10.3682,
"step": 64
},
{
"epoch": 0.33205619412515963,
"grad_norm": 0.06908663362264633,
"learning_rate": 0.0001504619880700346,
"loss": 10.367,
"step": 65
},
{
"epoch": 0.3371647509578544,
"grad_norm": 0.09832872450351715,
"learning_rate": 0.0001490717552003938,
"loss": 10.377,
"step": 66
},
{
"epoch": 0.34227330779054915,
"grad_norm": 0.0903470367193222,
"learning_rate": 0.00014766891538463254,
"loss": 10.3772,
"step": 67
},
{
"epoch": 0.34738186462324394,
"grad_norm": 0.09079986065626144,
"learning_rate": 0.00014625382902408356,
"loss": 10.366,
"step": 68
},
{
"epoch": 0.3524904214559387,
"grad_norm": 0.05147509649395943,
"learning_rate": 0.0001448268596663197,
"loss": 10.3652,
"step": 69
},
{
"epoch": 0.35759897828863346,
"grad_norm": 0.14060315489768982,
"learning_rate": 0.00014338837391175582,
"loss": 10.3629,
"step": 70
},
{
"epoch": 0.36270753512132825,
"grad_norm": 0.11259511113166809,
"learning_rate": 0.0001419387413194657,
"loss": 10.3691,
"step": 71
},
{
"epoch": 0.367816091954023,
"grad_norm": 0.09399597346782684,
"learning_rate": 0.00014047833431223938,
"loss": 10.3617,
"step": 72
},
{
"epoch": 0.37292464878671777,
"grad_norm": 0.11537044495344162,
"learning_rate": 0.00013900752808090468,
"loss": 10.366,
"step": 73
},
{
"epoch": 0.3780332056194125,
"grad_norm": 0.11738992482423782,
"learning_rate": 0.00013752670048793744,
"loss": 10.3471,
"step": 74
},
{
"epoch": 0.3831417624521073,
"grad_norm": 0.11229626834392548,
"learning_rate": 0.00013603623197038536,
"loss": 10.3645,
"step": 75
},
{
"epoch": 0.388250319284802,
"grad_norm": 0.14596045017242432,
"learning_rate": 0.00013453650544213076,
"loss": 10.3751,
"step": 76
},
{
"epoch": 0.3933588761174968,
"grad_norm": 0.06448373198509216,
"learning_rate": 0.00013302790619551674,
"loss": 10.3572,
"step": 77
},
{
"epoch": 0.39846743295019155,
"grad_norm": 0.11809150874614716,
"learning_rate": 0.0001315108218023621,
"loss": 10.3695,
"step": 78
},
{
"epoch": 0.40357598978288634,
"grad_norm": 0.08517828583717346,
"learning_rate": 0.00012998564201439116,
"loss": 10.3629,
"step": 79
},
{
"epoch": 0.4086845466155811,
"grad_norm": 0.12408934533596039,
"learning_rate": 0.00012845275866310324,
"loss": 10.3607,
"step": 80
},
{
"epoch": 0.41379310344827586,
"grad_norm": 0.12312637269496918,
"learning_rate": 0.00012691256555910768,
"loss": 10.3654,
"step": 81
},
{
"epoch": 0.41890166028097064,
"grad_norm": 0.12126640975475311,
"learning_rate": 0.00012536545839095074,
"loss": 10.3617,
"step": 82
},
{
"epoch": 0.4240102171136654,
"grad_norm": 0.15894931554794312,
"learning_rate": 0.00012381183462345982,
"loss": 10.3564,
"step": 83
},
{
"epoch": 0.42911877394636017,
"grad_norm": 0.07874782383441925,
"learning_rate": 0.00012225209339563145,
"loss": 10.3653,
"step": 84
},
{
"epoch": 0.4342273307790549,
"grad_norm": 0.12640683352947235,
"learning_rate": 0.00012068663541808909,
"loss": 10.3631,
"step": 85
},
{
"epoch": 0.4393358876117497,
"grad_norm": 0.10024367272853851,
"learning_rate": 0.00011911586287013725,
"loss": 10.3672,
"step": 86
},
{
"epoch": 0.4444444444444444,
"grad_norm": 0.12512411177158356,
"learning_rate": 0.00011754017929643817,
"loss": 10.3611,
"step": 87
},
{
"epoch": 0.4495530012771392,
"grad_norm": 0.1329926997423172,
"learning_rate": 0.00011595998950333793,
"loss": 10.3591,
"step": 88
},
{
"epoch": 0.454661558109834,
"grad_norm": 0.11963527649641037,
"learning_rate": 0.00011437569945486819,
"loss": 10.365,
"step": 89
},
{
"epoch": 0.45977011494252873,
"grad_norm": 0.11451319605112076,
"learning_rate": 0.00011278771616845061,
"loss": 10.3549,
"step": 90
},
{
"epoch": 0.4648786717752235,
"grad_norm": 0.09816919267177582,
"learning_rate": 0.00011119644761033078,
"loss": 10.3577,
"step": 91
},
{
"epoch": 0.46998722860791825,
"grad_norm": 0.15458016097545624,
"learning_rate": 0.00010960230259076818,
"loss": 10.3668,
"step": 92
},
{
"epoch": 0.47509578544061304,
"grad_norm": 0.1010480597615242,
"learning_rate": 0.00010800569065900933,
"loss": 10.3592,
"step": 93
},
{
"epoch": 0.48020434227330777,
"grad_norm": 0.11193486303091049,
"learning_rate": 0.0001064070219980713,
"loss": 10.3568,
"step": 94
},
{
"epoch": 0.48531289910600256,
"grad_norm": 0.10354366153478622,
"learning_rate": 0.00010480670731936208,
"loss": 10.3509,
"step": 95
},
{
"epoch": 0.4904214559386973,
"grad_norm": 0.15040072798728943,
"learning_rate": 0.00010320515775716555,
"loss": 10.3566,
"step": 96
},
{
"epoch": 0.4955300127713921,
"grad_norm": 0.10507191717624664,
"learning_rate": 0.0001016027847630174,
"loss": 10.3547,
"step": 97
},
{
"epoch": 0.5006385696040868,
"grad_norm": 0.11989340931177139,
"learning_rate": 0.0001,
"loss": 10.3656,
"step": 98
},
{
"epoch": 0.5057471264367817,
"grad_norm": 0.12597250938415527,
"learning_rate": 9.839721523698264e-05,
"loss": 10.3657,
"step": 99
},
{
"epoch": 0.5108556832694764,
"grad_norm": 0.16077381372451782,
"learning_rate": 9.679484224283449e-05,
"loss": 10.3527,
"step": 100
},
{
"epoch": 0.5159642401021711,
"grad_norm": 0.12048140168190002,
"learning_rate": 9.519329268063795e-05,
"loss": 10.3627,
"step": 101
},
{
"epoch": 0.5210727969348659,
"grad_norm": 0.10793527215719223,
"learning_rate": 9.359297800192872e-05,
"loss": 10.3508,
"step": 102
},
{
"epoch": 0.5261813537675607,
"grad_norm": 0.070986308157444,
"learning_rate": 9.199430934099068e-05,
"loss": 10.3545,
"step": 103
},
{
"epoch": 0.5312899106002554,
"grad_norm": 0.08290573209524155,
"learning_rate": 9.039769740923183e-05,
"loss": 10.3568,
"step": 104
},
{
"epoch": 0.5363984674329502,
"grad_norm": 0.09891117364168167,
"learning_rate": 8.880355238966923e-05,
"loss": 10.3572,
"step": 105
},
{
"epoch": 0.541507024265645,
"grad_norm": 0.10156704485416412,
"learning_rate": 8.721228383154939e-05,
"loss": 10.3588,
"step": 106
},
{
"epoch": 0.5466155810983397,
"grad_norm": 0.10102825611829758,
"learning_rate": 8.562430054513184e-05,
"loss": 10.3583,
"step": 107
},
{
"epoch": 0.5517241379310345,
"grad_norm": 0.10080267488956451,
"learning_rate": 8.404001049666211e-05,
"loss": 10.365,
"step": 108
},
{
"epoch": 0.5568326947637292,
"grad_norm": 0.08475471287965775,
"learning_rate": 8.245982070356185e-05,
"loss": 10.3606,
"step": 109
},
{
"epoch": 0.561941251596424,
"grad_norm": 0.13166114687919617,
"learning_rate": 8.08841371298628e-05,
"loss": 10.3538,
"step": 110
},
{
"epoch": 0.5670498084291188,
"grad_norm": 0.1031697690486908,
"learning_rate": 7.931336458191092e-05,
"loss": 10.3672,
"step": 111
},
{
"epoch": 0.5721583652618135,
"grad_norm": 0.08675719797611237,
"learning_rate": 7.774790660436858e-05,
"loss": 10.3575,
"step": 112
},
{
"epoch": 0.5772669220945083,
"grad_norm": 0.10204310715198517,
"learning_rate": 7.618816537654018e-05,
"loss": 10.3607,
"step": 113
},
{
"epoch": 0.5823754789272031,
"grad_norm": 0.12430866807699203,
"learning_rate": 7.463454160904928e-05,
"loss": 10.3586,
"step": 114
},
{
"epoch": 0.5874840357598978,
"grad_norm": 0.09749015420675278,
"learning_rate": 7.308743444089232e-05,
"loss": 10.3493,
"step": 115
},
{
"epoch": 0.5925925925925926,
"grad_norm": 0.11196263879537582,
"learning_rate": 7.154724133689677e-05,
"loss": 10.3479,
"step": 116
},
{
"epoch": 0.5977011494252874,
"grad_norm": 0.1198868602514267,
"learning_rate": 7.001435798560883e-05,
"loss": 10.349,
"step": 117
},
{
"epoch": 0.6028097062579821,
"grad_norm": 0.10104350000619888,
"learning_rate": 6.848917819763793e-05,
"loss": 10.353,
"step": 118
},
{
"epoch": 0.6079182630906769,
"grad_norm": 0.09838002175092697,
"learning_rate": 6.697209380448333e-05,
"loss": 10.3552,
"step": 119
},
{
"epoch": 0.6130268199233716,
"grad_norm": 0.12079446762800217,
"learning_rate": 6.546349455786926e-05,
"loss": 10.3541,
"step": 120
},
{
"epoch": 0.6181353767560664,
"grad_norm": 0.07775072753429413,
"learning_rate": 6.396376802961468e-05,
"loss": 10.3623,
"step": 121
},
{
"epoch": 0.6232439335887612,
"grad_norm": 0.10357434302568436,
"learning_rate": 6.24732995120626e-05,
"loss": 10.3578,
"step": 122
},
{
"epoch": 0.6283524904214559,
"grad_norm": 0.11605259031057358,
"learning_rate": 6.0992471919095315e-05,
"loss": 10.3513,
"step": 123
},
{
"epoch": 0.6334610472541508,
"grad_norm": 0.0947057381272316,
"learning_rate": 5.952166568776062e-05,
"loss": 10.3484,
"step": 124
},
{
"epoch": 0.6385696040868455,
"grad_norm": 0.09507757425308228,
"learning_rate": 5.806125868053433e-05,
"loss": 10.3521,
"step": 125
},
{
"epoch": 0.6436781609195402,
"grad_norm": 0.09619959443807602,
"learning_rate": 5.6611626088244194e-05,
"loss": 10.3583,
"step": 126
},
{
"epoch": 0.648786717752235,
"grad_norm": 0.0792725682258606,
"learning_rate": 5.5173140333680306e-05,
"loss": 10.354,
"step": 127
},
{
"epoch": 0.6538952745849298,
"grad_norm": 0.105789415538311,
"learning_rate": 5.37461709759165e-05,
"loss": 10.3583,
"step": 128
},
{
"epoch": 0.6590038314176245,
"grad_norm": 0.10440662503242493,
"learning_rate": 5.2331084615367485e-05,
"loss": 10.3523,
"step": 129
},
{
"epoch": 0.6641123882503193,
"grad_norm": 0.10326969623565674,
"learning_rate": 5.092824479960625e-05,
"loss": 10.3574,
"step": 130
},
{
"epoch": 0.669220945083014,
"grad_norm": 0.08976173400878906,
"learning_rate": 4.953801192996543e-05,
"loss": 10.3548,
"step": 131
},
{
"epoch": 0.6743295019157088,
"grad_norm": 0.09830819070339203,
"learning_rate": 4.8160743168947496e-05,
"loss": 10.3497,
"step": 132
},
{
"epoch": 0.6794380587484036,
"grad_norm": 0.12448275089263916,
"learning_rate": 4.6796792348466356e-05,
"loss": 10.3661,
"step": 133
},
{
"epoch": 0.6845466155810983,
"grad_norm": 0.1166510358452797,
"learning_rate": 4.544650987894514e-05,
"loss": 10.3572,
"step": 134
},
{
"epoch": 0.6896551724137931,
"grad_norm": 0.11028123646974564,
"learning_rate": 4.4110242659292836e-05,
"loss": 10.3607,
"step": 135
},
{
"epoch": 0.6947637292464879,
"grad_norm": 0.14801649749279022,
"learning_rate": 4.278833398778306e-05,
"loss": 10.3544,
"step": 136
},
{
"epoch": 0.6998722860791826,
"grad_norm": 0.10162738710641861,
"learning_rate": 4.148112347385762e-05,
"loss": 10.3514,
"step": 137
},
{
"epoch": 0.7049808429118773,
"grad_norm": 0.08851826190948486,
"learning_rate": 4.0188946950878404e-05,
"loss": 10.3545,
"step": 138
},
{
"epoch": 0.7100893997445722,
"grad_norm": 0.0674971416592598,
"learning_rate": 3.8912136389848576e-05,
"loss": 10.3527,
"step": 139
},
{
"epoch": 0.7151979565772669,
"grad_norm": 0.07837618142366409,
"learning_rate": 3.7651019814126654e-05,
"loss": 10.3495,
"step": 140
},
{
"epoch": 0.7203065134099617,
"grad_norm": 0.13337913155555725,
"learning_rate": 3.6405921215154494e-05,
"loss": 10.3601,
"step": 141
},
{
"epoch": 0.7254150702426565,
"grad_norm": 0.1068163514137268,
"learning_rate": 3.517716046922118e-05,
"loss": 10.3553,
"step": 142
},
{
"epoch": 0.7305236270753512,
"grad_norm": 0.09636684507131577,
"learning_rate": 3.3965053255284084e-05,
"loss": 10.3609,
"step": 143
},
{
"epoch": 0.735632183908046,
"grad_norm": 0.10259624570608139,
"learning_rate": 3.276991097386831e-05,
"loss": 10.3514,
"step": 144
},
{
"epoch": 0.7407407407407407,
"grad_norm": 0.08181837946176529,
"learning_rate": 3.159204066706539e-05,
"loss": 10.3508,
"step": 145
},
{
"epoch": 0.7458492975734355,
"grad_norm": 0.10370145738124847,
"learning_rate": 3.0431744939651364e-05,
"loss": 10.3518,
"step": 146
},
{
"epoch": 0.7509578544061303,
"grad_norm": 0.11586267501115799,
"learning_rate": 2.9289321881345254e-05,
"loss": 10.3494,
"step": 147
},
{
"epoch": 0.756066411238825,
"grad_norm": 0.07326210290193558,
"learning_rate": 2.8165064990227252e-05,
"loss": 10.3618,
"step": 148
},
{
"epoch": 0.7611749680715197,
"grad_norm": 0.07508935779333115,
"learning_rate": 2.7059263097336597e-05,
"loss": 10.354,
"step": 149
},
{
"epoch": 0.7662835249042146,
"grad_norm": 0.09258796274662018,
"learning_rate": 2.5972200292468464e-05,
"loss": 10.3528,
"step": 150
},
{
"epoch": 0.7713920817369093,
"grad_norm": 0.09141936898231506,
"learning_rate": 2.4904155851188872e-05,
"loss": 10.3475,
"step": 151
},
{
"epoch": 0.776500638569604,
"grad_norm": 0.06457791477441788,
"learning_rate": 2.3855404163086558e-05,
"loss": 10.3594,
"step": 152
},
{
"epoch": 0.7816091954022989,
"grad_norm": 0.0867636650800705,
"learning_rate": 2.282621466127982e-05,
"loss": 10.357,
"step": 153
},
{
"epoch": 0.7867177522349936,
"grad_norm": 0.07386265695095062,
"learning_rate": 2.181685175319702e-05,
"loss": 10.3601,
"step": 154
},
{
"epoch": 0.7918263090676884,
"grad_norm": 0.09301372617483139,
"learning_rate": 2.0827574752648038e-05,
"loss": 10.3521,
"step": 155
},
{
"epoch": 0.7969348659003831,
"grad_norm": 0.09825734794139862,
"learning_rate": 1.985863781320435e-05,
"loss": 10.3518,
"step": 156
},
{
"epoch": 0.8020434227330779,
"grad_norm": 0.09146848320960999,
"learning_rate": 1.891028986290492e-05,
"loss": 10.3543,
"step": 157
},
{
"epoch": 0.8071519795657727,
"grad_norm": 0.10811572521924973,
"learning_rate": 1.7982774540304403e-05,
"loss": 10.3486,
"step": 158
},
{
"epoch": 0.8122605363984674,
"grad_norm": 0.10702500492334366,
"learning_rate": 1.7076330131880526e-05,
"loss": 10.3506,
"step": 159
},
{
"epoch": 0.8173690932311622,
"grad_norm": 0.09714721143245697,
"learning_rate": 1.619118951081594e-05,
"loss": 10.3542,
"step": 160
},
{
"epoch": 0.822477650063857,
"grad_norm": 0.09883596003055573,
"learning_rate": 1.5327580077171587e-05,
"loss": 10.355,
"step": 161
},
{
"epoch": 0.8275862068965517,
"grad_norm": 0.07649561762809753,
"learning_rate": 1.4485723699465392e-05,
"loss": 10.3488,
"step": 162
},
{
"epoch": 0.8326947637292464,
"grad_norm": 0.0714419037103653,
"learning_rate": 1.3665836657672493e-05,
"loss": 10.3463,
"step": 163
},
{
"epoch": 0.8378033205619413,
"grad_norm": 0.09287029504776001,
"learning_rate": 1.286812958766106e-05,
"loss": 10.3506,
"step": 164
},
{
"epoch": 0.842911877394636,
"grad_norm": 0.0882708728313446,
"learning_rate": 1.2092807427078279e-05,
"loss": 10.3538,
"step": 165
},
{
"epoch": 0.8480204342273308,
"grad_norm": 0.0939875915646553,
"learning_rate": 1.134006936269999e-05,
"loss": 10.3498,
"step": 166
},
{
"epoch": 0.8531289910600255,
"grad_norm": 0.10232677310705185,
"learning_rate": 1.0610108779258044e-05,
"loss": 10.3573,
"step": 167
},
{
"epoch": 0.8582375478927203,
"grad_norm": 0.09528540074825287,
"learning_rate": 9.903113209758096e-06,
"loss": 10.3558,
"step": 168
},
{
"epoch": 0.8633461047254151,
"grad_norm": 0.08980326354503632,
"learning_rate": 9.219264287300799e-06,
"loss": 10.3564,
"step": 169
},
{
"epoch": 0.8684546615581098,
"grad_norm": 0.0988466814160347,
"learning_rate": 8.558737698418761e-06,
"loss": 10.3518,
"step": 170
},
{
"epoch": 0.8735632183908046,
"grad_norm": 0.09410892426967621,
"learning_rate": 7.921703137941173e-06,
"loss": 10.3501,
"step": 171
},
{
"epoch": 0.8786717752234994,
"grad_norm": 0.0989941656589508,
"learning_rate": 7.308324265397836e-06,
"loss": 10.3563,
"step": 172
},
{
"epoch": 0.8837803320561941,
"grad_norm": 0.12267417460680008,
"learning_rate": 6.718758662973523e-06,
"loss": 10.3536,
"step": 173
},
{
"epoch": 0.8888888888888888,
"grad_norm": 0.09049776196479797,
"learning_rate": 6.153157795023956e-06,
"loss": 10.3524,
"step": 174
},
{
"epoch": 0.8939974457215837,
"grad_norm": 0.10820028930902481,
"learning_rate": 5.611666969163243e-06,
"loss": 10.3631,
"step": 175
},
{
"epoch": 0.8991060025542784,
"grad_norm": 0.08953528106212616,
"learning_rate": 5.094425298933136e-06,
"loss": 10.3467,
"step": 176
},
{
"epoch": 0.9042145593869731,
"grad_norm": 0.12113954871892929,
"learning_rate": 4.601565668063623e-06,
"loss": 10.3563,
"step": 177
},
{
"epoch": 0.909323116219668,
"grad_norm": 0.08720055967569351,
"learning_rate": 4.133214696333942e-06,
"loss": 10.356,
"step": 178
},
{
"epoch": 0.9144316730523627,
"grad_norm": 0.09181863814592361,
"learning_rate": 3.689492707042974e-06,
"loss": 10.3552,
"step": 179
},
{
"epoch": 0.9195402298850575,
"grad_norm": 0.09627222269773483,
"learning_rate": 3.270513696097055e-06,
"loss": 10.3541,
"step": 180
},
{
"epoch": 0.9246487867177522,
"grad_norm": 0.11901070922613144,
"learning_rate": 2.876385302723628e-06,
"loss": 10.3597,
"step": 181
},
{
"epoch": 0.929757343550447,
"grad_norm": 0.0824970006942749,
"learning_rate": 2.5072087818176382e-06,
"loss": 10.3537,
"step": 182
},
{
"epoch": 0.9348659003831418,
"grad_norm": 0.10875054448843002,
"learning_rate": 2.1630789779284675e-06,
"loss": 10.3465,
"step": 183
},
{
"epoch": 0.9399744572158365,
"grad_norm": 0.08391505479812622,
"learning_rate": 1.8440843008934561e-06,
"loss": 10.3537,
"step": 184
},
{
"epoch": 0.9450830140485313,
"grad_norm": 0.0997113361954689,
"learning_rate": 1.5503067031247598e-06,
"loss": 10.3528,
"step": 185
},
{
"epoch": 0.9501915708812261,
"grad_norm": 0.1007162481546402,
"learning_rate": 1.2818216585549825e-06,
"loss": 10.3498,
"step": 186
},
{
"epoch": 0.9553001277139208,
"grad_norm": 0.0825323835015297,
"learning_rate": 1.0386981432474074e-06,
"loss": 10.3464,
"step": 187
},
{
"epoch": 0.9604086845466155,
"grad_norm": 0.07373323291540146,
"learning_rate": 8.209986176753948e-07,
"loss": 10.3546,
"step": 188
},
{
"epoch": 0.9655172413793104,
"grad_norm": 0.10504812747240067,
"learning_rate": 6.287790106757396e-07,
"loss": 10.3522,
"step": 189
},
{
"epoch": 0.9706257982120051,
"grad_norm": 0.07949844747781754,
"learning_rate": 4.62088705080177e-07,
"loss": 10.3554,
"step": 190
},
{
"epoch": 0.9757343550446999,
"grad_norm": 0.09091745316982269,
"learning_rate": 3.2097052502843007e-07,
"loss": 10.3477,
"step": 191
},
{
"epoch": 0.9808429118773946,
"grad_norm": 0.11523844301700592,
"learning_rate": 2.054607249663665e-07,
"loss": 10.3526,
"step": 192
},
{
"epoch": 0.9859514687100894,
"grad_norm": 0.09753245860338211,
"learning_rate": 1.1558898033191546e-07,
"loss": 10.3566,
"step": 193
},
{
"epoch": 0.9910600255427842,
"grad_norm": 0.08241667598485947,
"learning_rate": 5.137837993121064e-08,
"loss": 10.3576,
"step": 194
},
{
"epoch": 0.9961685823754789,
"grad_norm": 0.08668815344572067,
"learning_rate": 1.2845420006879494e-08,
"loss": 10.3612,
"step": 195
},
{
"epoch": 0.9961685823754789,
"eval_loss": 10.354520797729492,
"eval_runtime": 0.2879,
"eval_samples_per_second": 145.877,
"eval_steps_per_second": 145.877,
"step": 195
},
{
"epoch": 1.0012771392081736,
"grad_norm": 0.08981695771217346,
"learning_rate": 0.0,
"loss": 12.4365,
"step": 196
},
{
"epoch": 1.0012771392081736,
"eval_loss": 10.354520797729492,
"eval_runtime": 0.2791,
"eval_samples_per_second": 150.484,
"eval_steps_per_second": 150.484,
"step": 196
}
],
"logging_steps": 1,
"max_steps": 196,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2534457999360.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}