|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9996020692399522, |
|
"eval_steps": 1000, |
|
"global_step": 628, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0015917230401910067, |
|
"grad_norm": 3.414436464235707, |
|
"learning_rate": 7.936507936507936e-09, |
|
"logits/chosen": -2.458916187286377, |
|
"logits/rejected": -2.50296688079834, |
|
"logps/chosen": -112.41877746582031, |
|
"logps/rejected": -153.82745361328125, |
|
"loss": 0.6931, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01591723040191007, |
|
"grad_norm": 3.47500960837261, |
|
"learning_rate": 7.936507936507936e-08, |
|
"logits/chosen": -2.5214052200317383, |
|
"logits/rejected": -2.508702039718628, |
|
"logps/chosen": -143.01979064941406, |
|
"logps/rejected": -133.09210205078125, |
|
"loss": 0.6932, |
|
"rewards/accuracies": 0.4166666567325592, |
|
"rewards/chosen": -0.00031594105530530214, |
|
"rewards/margins": -0.00045727533870376647, |
|
"rewards/rejected": 0.00014133438526187092, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03183446080382014, |
|
"grad_norm": 3.343704534649061, |
|
"learning_rate": 1.5873015873015872e-07, |
|
"logits/chosen": -2.4521286487579346, |
|
"logits/rejected": -2.448809862136841, |
|
"logps/chosen": -151.37095642089844, |
|
"logps/rejected": -162.80328369140625, |
|
"loss": 0.6931, |
|
"rewards/accuracies": 0.43437498807907104, |
|
"rewards/chosen": 4.1505660192342475e-05, |
|
"rewards/margins": -0.00036320873186923563, |
|
"rewards/rejected": 0.000404714432079345, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0477516912057302, |
|
"grad_norm": 3.366996896296212, |
|
"learning_rate": 2.3809523809523806e-07, |
|
"logits/chosen": -2.4904391765594482, |
|
"logits/rejected": -2.479889392852783, |
|
"logps/chosen": -147.56060791015625, |
|
"logps/rejected": -150.96810913085938, |
|
"loss": 0.6921, |
|
"rewards/accuracies": 0.550000011920929, |
|
"rewards/chosen": 0.003838698612526059, |
|
"rewards/margins": 0.002006606664508581, |
|
"rewards/rejected": 0.001832091948017478, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.06366892160764027, |
|
"grad_norm": 3.3714607115747497, |
|
"learning_rate": 3.1746031746031743e-07, |
|
"logits/chosen": -2.503981590270996, |
|
"logits/rejected": -2.486074447631836, |
|
"logps/chosen": -153.0246124267578, |
|
"logps/rejected": -145.2753448486328, |
|
"loss": 0.6899, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": 0.007592098321765661, |
|
"rewards/margins": 0.00429057702422142, |
|
"rewards/rejected": 0.0033015217632055283, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07958615200955034, |
|
"grad_norm": 4.06423931986909, |
|
"learning_rate": 3.968253968253968e-07, |
|
"logits/chosen": -2.4848971366882324, |
|
"logits/rejected": -2.4830713272094727, |
|
"logps/chosen": -146.40438842773438, |
|
"logps/rejected": -152.903564453125, |
|
"loss": 0.6849, |
|
"rewards/accuracies": 0.5531250238418579, |
|
"rewards/chosen": -0.01682356372475624, |
|
"rewards/margins": 0.0171933826059103, |
|
"rewards/rejected": -0.03401694819331169, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0955033824114604, |
|
"grad_norm": 3.933215748835345, |
|
"learning_rate": 4.761904761904761e-07, |
|
"logits/chosen": -2.4494073390960693, |
|
"logits/rejected": -2.425264835357666, |
|
"logps/chosen": -164.67039489746094, |
|
"logps/rejected": -159.42379760742188, |
|
"loss": 0.6811, |
|
"rewards/accuracies": 0.528124988079071, |
|
"rewards/chosen": -0.13071545958518982, |
|
"rewards/margins": 0.01678677648305893, |
|
"rewards/rejected": -0.14750224351882935, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.11142061281337047, |
|
"grad_norm": 6.465696654064857, |
|
"learning_rate": 4.998106548810311e-07, |
|
"logits/chosen": -2.4344656467437744, |
|
"logits/rejected": -2.3922276496887207, |
|
"logps/chosen": -171.0955352783203, |
|
"logps/rejected": -158.62640380859375, |
|
"loss": 0.6666, |
|
"rewards/accuracies": 0.596875011920929, |
|
"rewards/chosen": -0.17716214060783386, |
|
"rewards/margins": 0.07567773759365082, |
|
"rewards/rejected": -0.2528398633003235, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.12733784321528055, |
|
"grad_norm": 9.086502750868021, |
|
"learning_rate": 4.988839406031596e-07, |
|
"logits/chosen": -2.384453058242798, |
|
"logits/rejected": -2.3677101135253906, |
|
"logps/chosen": -153.10414123535156, |
|
"logps/rejected": -182.54360961914062, |
|
"loss": 0.6529, |
|
"rewards/accuracies": 0.690625011920929, |
|
"rewards/chosen": -0.20188739895820618, |
|
"rewards/margins": 0.12667986750602722, |
|
"rewards/rejected": -0.3285672962665558, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.14325507361719061, |
|
"grad_norm": 14.480905623348013, |
|
"learning_rate": 4.971879403278432e-07, |
|
"logits/chosen": -2.2432703971862793, |
|
"logits/rejected": -2.2313826084136963, |
|
"logps/chosen": -185.1287384033203, |
|
"logps/rejected": -201.3049774169922, |
|
"loss": 0.6398, |
|
"rewards/accuracies": 0.643750011920929, |
|
"rewards/chosen": -0.39208894968032837, |
|
"rewards/margins": 0.144916832447052, |
|
"rewards/rejected": -0.5370057821273804, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.15917230401910068, |
|
"grad_norm": 12.626265646656822, |
|
"learning_rate": 4.947278962947386e-07, |
|
"logits/chosen": -1.7865173816680908, |
|
"logits/rejected": -1.7619308233261108, |
|
"logps/chosen": -234.49526977539062, |
|
"logps/rejected": -266.22235107421875, |
|
"loss": 0.6186, |
|
"rewards/accuracies": 0.659375011920929, |
|
"rewards/chosen": -0.8486677408218384, |
|
"rewards/margins": 0.2785636782646179, |
|
"rewards/rejected": -1.127231478691101, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.17508953442101075, |
|
"grad_norm": 11.440443186979932, |
|
"learning_rate": 4.915114123589732e-07, |
|
"logits/chosen": -1.477772831916809, |
|
"logits/rejected": -1.394840955734253, |
|
"logps/chosen": -206.17166137695312, |
|
"logps/rejected": -243.156005859375, |
|
"loss": 0.6156, |
|
"rewards/accuracies": 0.684374988079071, |
|
"rewards/chosen": -0.719870388507843, |
|
"rewards/margins": 0.34280186891555786, |
|
"rewards/rejected": -1.0626722574234009, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.1910067648229208, |
|
"grad_norm": 14.957458897670607, |
|
"learning_rate": 4.875484304880629e-07, |
|
"logits/chosen": -0.8948208093643188, |
|
"logits/rejected": -0.773229718208313, |
|
"logps/chosen": -234.4495086669922, |
|
"logps/rejected": -270.4801330566406, |
|
"loss": 0.5931, |
|
"rewards/accuracies": 0.699999988079071, |
|
"rewards/chosen": -0.8048633337020874, |
|
"rewards/margins": 0.34693923592567444, |
|
"rewards/rejected": -1.151802659034729, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.20692399522483088, |
|
"grad_norm": 13.673290760971415, |
|
"learning_rate": 4.828512000318616e-07, |
|
"logits/chosen": -0.20325860381126404, |
|
"logits/rejected": -0.022473735734820366, |
|
"logps/chosen": -233.56399536132812, |
|
"logps/rejected": -278.225830078125, |
|
"loss": 0.5842, |
|
"rewards/accuracies": 0.7093750238418579, |
|
"rewards/chosen": -0.7881739139556885, |
|
"rewards/margins": 0.4328451156616211, |
|
"rewards/rejected": -1.2210190296173096, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.22284122562674094, |
|
"grad_norm": 15.252755388662429, |
|
"learning_rate": 4.774342398605221e-07, |
|
"logits/chosen": -0.4290715754032135, |
|
"logits/rejected": -0.26301848888397217, |
|
"logps/chosen": -245.796142578125, |
|
"logps/rejected": -278.6325988769531, |
|
"loss": 0.5803, |
|
"rewards/accuracies": 0.6781250238418579, |
|
"rewards/chosen": -0.9634370803833008, |
|
"rewards/margins": 0.42050686478614807, |
|
"rewards/rejected": -1.383944034576416, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.238758456028651, |
|
"grad_norm": 15.98257839847235, |
|
"learning_rate": 4.713142934875005e-07, |
|
"logits/chosen": 0.04155537486076355, |
|
"logits/rejected": 0.4105464518070221, |
|
"logps/chosen": -294.55078125, |
|
"logps/rejected": -322.9659118652344, |
|
"loss": 0.5928, |
|
"rewards/accuracies": 0.684374988079071, |
|
"rewards/chosen": -1.3449639081954956, |
|
"rewards/margins": 0.41515955328941345, |
|
"rewards/rejected": -1.7601234912872314, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2546756864305611, |
|
"grad_norm": 16.296020720117806, |
|
"learning_rate": 4.64510277316316e-07, |
|
"logits/chosen": 0.07806523889303207, |
|
"logits/rejected": 0.4625118672847748, |
|
"logps/chosen": -242.23171997070312, |
|
"logps/rejected": -306.7259216308594, |
|
"loss": 0.5619, |
|
"rewards/accuracies": 0.7281249761581421, |
|
"rewards/chosen": -1.0831429958343506, |
|
"rewards/margins": 0.6153541803359985, |
|
"rewards/rejected": -1.6984970569610596, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.27059291683247116, |
|
"grad_norm": 15.29787072223291, |
|
"learning_rate": 4.570432221710314e-07, |
|
"logits/chosen": 0.8908351063728333, |
|
"logits/rejected": 1.1234996318817139, |
|
"logps/chosen": -285.02685546875, |
|
"logps/rejected": -332.22576904296875, |
|
"loss": 0.5968, |
|
"rewards/accuracies": 0.6499999761581421, |
|
"rewards/chosen": -1.369902491569519, |
|
"rewards/margins": 0.42481645941734314, |
|
"rewards/rejected": -1.7947190999984741, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.28651014723438123, |
|
"grad_norm": 12.440638905265972, |
|
"learning_rate": 4.4893620829118124e-07, |
|
"logits/chosen": 0.45191067457199097, |
|
"logits/rejected": 0.8097447156906128, |
|
"logps/chosen": -267.55511474609375, |
|
"logps/rejected": -303.0255432128906, |
|
"loss": 0.5782, |
|
"rewards/accuracies": 0.690625011920929, |
|
"rewards/chosen": -1.130518913269043, |
|
"rewards/margins": 0.47403573989868164, |
|
"rewards/rejected": -1.6045547723770142, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.3024273776362913, |
|
"grad_norm": 14.662664269581565, |
|
"learning_rate": 4.40214293992074e-07, |
|
"logits/chosen": 0.07046165317296982, |
|
"logits/rejected": 0.5015997886657715, |
|
"logps/chosen": -276.8005676269531, |
|
"logps/rejected": -317.6512451171875, |
|
"loss": 0.5684, |
|
"rewards/accuracies": 0.699999988079071, |
|
"rewards/chosen": -1.328792929649353, |
|
"rewards/margins": 0.5068601369857788, |
|
"rewards/rejected": -1.835653305053711, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.31834460803820136, |
|
"grad_norm": 13.757436126080059, |
|
"learning_rate": 4.3090443821097566e-07, |
|
"logits/chosen": 0.2187742292881012, |
|
"logits/rejected": 0.5610963106155396, |
|
"logps/chosen": -268.94757080078125, |
|
"logps/rejected": -333.0284729003906, |
|
"loss": 0.5596, |
|
"rewards/accuracies": 0.7124999761581421, |
|
"rewards/chosen": -1.2326629161834717, |
|
"rewards/margins": 0.5554476976394653, |
|
"rewards/rejected": -1.7881107330322266, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.3342618384401114, |
|
"grad_norm": 12.638546225207325, |
|
"learning_rate": 4.210354171785795e-07, |
|
"logits/chosen": 0.5555538535118103, |
|
"logits/rejected": 0.9611243009567261, |
|
"logps/chosen": -258.8009338378906, |
|
"logps/rejected": -315.17242431640625, |
|
"loss": 0.5627, |
|
"rewards/accuracies": 0.7250000238418579, |
|
"rewards/chosen": -1.1692590713500977, |
|
"rewards/margins": 0.5407182574272156, |
|
"rewards/rejected": -1.709977388381958, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.3501790688420215, |
|
"grad_norm": 12.533508219187622, |
|
"learning_rate": 4.1063773547332584e-07, |
|
"logits/chosen": 0.4853738248348236, |
|
"logits/rejected": 0.7825923562049866, |
|
"logps/chosen": -255.5019073486328, |
|
"logps/rejected": -307.07843017578125, |
|
"loss": 0.5684, |
|
"rewards/accuracies": 0.6656249761581421, |
|
"rewards/chosen": -1.0966681241989136, |
|
"rewards/margins": 0.5020471215248108, |
|
"rewards/rejected": -1.59871506690979, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.36609629924393156, |
|
"grad_norm": 13.14652988888945, |
|
"learning_rate": 3.997435317334988e-07, |
|
"logits/chosen": 1.3910820484161377, |
|
"logits/rejected": 1.7489607334136963, |
|
"logps/chosen": -285.7388916015625, |
|
"logps/rejected": -342.02557373046875, |
|
"loss": 0.5556, |
|
"rewards/accuracies": 0.731249988079071, |
|
"rewards/chosen": -1.430323600769043, |
|
"rewards/margins": 0.5691004991531372, |
|
"rewards/rejected": -1.9994239807128906, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.3820135296458416, |
|
"grad_norm": 16.893312636051345, |
|
"learning_rate": 3.8838647931853684e-07, |
|
"logits/chosen": 0.9681981801986694, |
|
"logits/rejected": 1.3831474781036377, |
|
"logps/chosen": -274.5780029296875, |
|
"logps/rejected": -334.71624755859375, |
|
"loss": 0.565, |
|
"rewards/accuracies": 0.684374988079071, |
|
"rewards/chosen": -1.331866979598999, |
|
"rewards/margins": 0.6004709005355835, |
|
"rewards/rejected": -1.932337999343872, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.3979307600477517, |
|
"grad_norm": 16.93758119869443, |
|
"learning_rate": 3.7660168222660824e-07, |
|
"logits/chosen": -0.0617893747985363, |
|
"logits/rejected": 0.3982541561126709, |
|
"logps/chosen": -291.5325012207031, |
|
"logps/rejected": -349.10357666015625, |
|
"loss": 0.546, |
|
"rewards/accuracies": 0.7281249761581421, |
|
"rewards/chosen": -1.3192743062973022, |
|
"rewards/margins": 0.6018632650375366, |
|
"rewards/rejected": -1.9211375713348389, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.41384799044966175, |
|
"grad_norm": 13.594747823748454, |
|
"learning_rate": 3.6442556659016475e-07, |
|
"logits/chosen": -0.1596679985523224, |
|
"logits/rejected": 0.12459886074066162, |
|
"logps/chosen": -283.2134704589844, |
|
"logps/rejected": -335.0664367675781, |
|
"loss": 0.5677, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.3012964725494385, |
|
"rewards/margins": 0.5175098180770874, |
|
"rewards/rejected": -1.8188062906265259, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.4297652208515718, |
|
"grad_norm": 13.029006886811343, |
|
"learning_rate": 3.5189576808485404e-07, |
|
"logits/chosen": 0.6624835729598999, |
|
"logits/rejected": 0.867287278175354, |
|
"logps/chosen": -290.3371887207031, |
|
"logps/rejected": -335.45745849609375, |
|
"loss": 0.5603, |
|
"rewards/accuracies": 0.7093750238418579, |
|
"rewards/chosen": -1.3652312755584717, |
|
"rewards/margins": 0.44845834374427795, |
|
"rewards/rejected": -1.8136895895004272, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.4456824512534819, |
|
"grad_norm": 14.229518064110062, |
|
"learning_rate": 3.390510155998023e-07, |
|
"logits/chosen": 1.0853196382522583, |
|
"logits/rejected": 1.5426385402679443, |
|
"logps/chosen": -301.244873046875, |
|
"logps/rejected": -352.26214599609375, |
|
"loss": 0.5576, |
|
"rewards/accuracies": 0.721875011920929, |
|
"rewards/chosen": -1.4498573541641235, |
|
"rewards/margins": 0.5846196413040161, |
|
"rewards/rejected": -2.0344767570495605, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.46159968165539195, |
|
"grad_norm": 13.247814891025222, |
|
"learning_rate": 3.2593101152883795e-07, |
|
"logits/chosen": 0.6425064206123352, |
|
"logits/rejected": 1.0470354557037354, |
|
"logps/chosen": -304.13165283203125, |
|
"logps/rejected": -356.2009582519531, |
|
"loss": 0.5699, |
|
"rewards/accuracies": 0.6781250238418579, |
|
"rewards/chosen": -1.4935743808746338, |
|
"rewards/margins": 0.543701708316803, |
|
"rewards/rejected": -2.037276029586792, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.477516912057302, |
|
"grad_norm": 14.136317593633118, |
|
"learning_rate": 3.125763090526674e-07, |
|
"logits/chosen": 1.010498285293579, |
|
"logits/rejected": 1.5458686351776123, |
|
"logps/chosen": -312.71893310546875, |
|
"logps/rejected": -367.11956787109375, |
|
"loss": 0.5526, |
|
"rewards/accuracies": 0.6968749761581421, |
|
"rewards/chosen": -1.607622742652893, |
|
"rewards/margins": 0.5898184180259705, |
|
"rewards/rejected": -2.1974411010742188, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.4934341424592121, |
|
"grad_norm": 13.599580845693339, |
|
"learning_rate": 2.9902818679131775e-07, |
|
"logits/chosen": 1.556601881980896, |
|
"logits/rejected": 2.024049758911133, |
|
"logps/chosen": -317.42938232421875, |
|
"logps/rejected": -371.38623046875, |
|
"loss": 0.5571, |
|
"rewards/accuracies": 0.6937500238418579, |
|
"rewards/chosen": -1.6657898426055908, |
|
"rewards/margins": 0.5584100484848022, |
|
"rewards/rejected": -2.2241997718811035, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.5093513728611222, |
|
"grad_norm": 15.486325517579422, |
|
"learning_rate": 2.8532852121428733e-07, |
|
"logits/chosen": 2.424159526824951, |
|
"logits/rejected": 2.8681328296661377, |
|
"logps/chosen": -326.8099670410156, |
|
"logps/rejected": -373.8744201660156, |
|
"loss": 0.5411, |
|
"rewards/accuracies": 0.6656249761581421, |
|
"rewards/chosen": -1.7966935634613037, |
|
"rewards/margins": 0.5875706076622009, |
|
"rewards/rejected": -2.3842644691467285, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.5252686032630323, |
|
"grad_norm": 13.394275013231457, |
|
"learning_rate": 2.715196572027789e-07, |
|
"logits/chosen": 3.215583324432373, |
|
"logits/rejected": 3.725792646408081, |
|
"logps/chosen": -343.72344970703125, |
|
"logps/rejected": -432.32916259765625, |
|
"loss": 0.5414, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.9976279735565186, |
|
"rewards/margins": 0.806243896484375, |
|
"rewards/rejected": -2.8038716316223145, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.5411858336649423, |
|
"grad_norm": 13.373618178706888, |
|
"learning_rate": 2.5764427716409815e-07, |
|
"logits/chosen": 3.000930070877075, |
|
"logits/rejected": 3.6059279441833496, |
|
"logps/chosen": -347.54022216796875, |
|
"logps/rejected": -404.14654541015625, |
|
"loss": 0.5507, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.9271198511123657, |
|
"rewards/margins": 0.6116258502006531, |
|
"rewards/rejected": -2.538745880126953, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.5571030640668524, |
|
"grad_norm": 14.475581092351593, |
|
"learning_rate": 2.4374526910277886e-07, |
|
"logits/chosen": 2.816096782684326, |
|
"logits/rejected": 3.264390230178833, |
|
"logps/chosen": -353.548828125, |
|
"logps/rejected": -420.4794006347656, |
|
"loss": 0.5564, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -2.0463950634002686, |
|
"rewards/margins": 0.6025586128234863, |
|
"rewards/rejected": -2.648953914642334, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.5730202944687625, |
|
"grad_norm": 14.682353004178683, |
|
"learning_rate": 2.2986559405621886e-07, |
|
"logits/chosen": 2.8994641304016113, |
|
"logits/rejected": 3.608001232147217, |
|
"logps/chosen": -360.53143310546875, |
|
"logps/rejected": -417.3299865722656, |
|
"loss": 0.5525, |
|
"rewards/accuracies": 0.7124999761581421, |
|
"rewards/chosen": -1.981393814086914, |
|
"rewards/margins": 0.6567780375480652, |
|
"rewards/rejected": -2.638171911239624, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.5889375248706725, |
|
"grad_norm": 18.068831076710342, |
|
"learning_rate": 2.160481533045751e-07, |
|
"logits/chosen": 2.5152053833007812, |
|
"logits/rejected": 3.3655197620391846, |
|
"logps/chosen": -339.1783752441406, |
|
"logps/rejected": -407.2680358886719, |
|
"loss": 0.5377, |
|
"rewards/accuracies": 0.7437499761581421, |
|
"rewards/chosen": -1.932498574256897, |
|
"rewards/margins": 0.7526292204856873, |
|
"rewards/rejected": -2.6851277351379395, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.6048547552725826, |
|
"grad_norm": 13.903302227889153, |
|
"learning_rate": 2.0233565576536564e-07, |
|
"logits/chosen": 2.7033469676971436, |
|
"logits/rejected": 3.3517203330993652, |
|
"logps/chosen": -352.88543701171875, |
|
"logps/rejected": -414.46807861328125, |
|
"loss": 0.5464, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -2.0058979988098145, |
|
"rewards/margins": 0.6991748809814453, |
|
"rewards/rejected": -2.7050728797912598, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.6207719856744927, |
|
"grad_norm": 14.437387596932032, |
|
"learning_rate": 1.887704859826528e-07, |
|
"logits/chosen": 2.677304744720459, |
|
"logits/rejected": 3.3016350269317627, |
|
"logps/chosen": -345.10186767578125, |
|
"logps/rejected": -422.6465759277344, |
|
"loss": 0.5441, |
|
"rewards/accuracies": 0.753125011920929, |
|
"rewards/chosen": -1.9631824493408203, |
|
"rewards/margins": 0.7392277717590332, |
|
"rewards/rejected": -2.7024102210998535, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.6366892160764027, |
|
"grad_norm": 15.504689068952228, |
|
"learning_rate": 1.7539457311884675e-07, |
|
"logits/chosen": 2.6285808086395264, |
|
"logits/rejected": 3.402756452560425, |
|
"logps/chosen": -350.1034240722656, |
|
"logps/rejected": -420.3218688964844, |
|
"loss": 0.5387, |
|
"rewards/accuracies": 0.737500011920929, |
|
"rewards/chosen": -1.9105831384658813, |
|
"rewards/margins": 0.7946080565452576, |
|
"rewards/rejected": -2.705191135406494, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.6526064464783128, |
|
"grad_norm": 12.730811962712194, |
|
"learning_rate": 1.6224926135406693e-07, |
|
"logits/chosen": 2.7364511489868164, |
|
"logits/rejected": 3.1644163131713867, |
|
"logps/chosen": -364.25469970703125, |
|
"logps/rejected": -419.5701599121094, |
|
"loss": 0.5618, |
|
"rewards/accuracies": 0.690625011920929, |
|
"rewards/chosen": -2.067072629928589, |
|
"rewards/margins": 0.5998597741127014, |
|
"rewards/rejected": -2.6669325828552246, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.6685236768802229, |
|
"grad_norm": 14.2356920381672, |
|
"learning_rate": 1.4937518209365108e-07, |
|
"logits/chosen": 2.5111005306243896, |
|
"logits/rejected": 3.2514469623565674, |
|
"logps/chosen": -381.5763244628906, |
|
"logps/rejected": -433.4600524902344, |
|
"loss": 0.5494, |
|
"rewards/accuracies": 0.778124988079071, |
|
"rewards/chosen": -2.099307060241699, |
|
"rewards/margins": 0.6568456888198853, |
|
"rewards/rejected": -2.756152629852295, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.6844409072821329, |
|
"grad_norm": 15.52832700397283, |
|
"learning_rate": 1.3681212837880977e-07, |
|
"logits/chosen": 2.709282398223877, |
|
"logits/rejected": 3.0862138271331787, |
|
"logps/chosen": -374.52288818359375, |
|
"logps/rejected": -447.3203125, |
|
"loss": 0.5539, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -2.311601161956787, |
|
"rewards/margins": 0.6582018136978149, |
|
"rewards/rejected": -2.9698030948638916, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.700358137684043, |
|
"grad_norm": 13.300949643909883, |
|
"learning_rate": 1.2459893188861613e-07, |
|
"logits/chosen": 2.896282196044922, |
|
"logits/rejected": 3.553002119064331, |
|
"logps/chosen": -382.75054931640625, |
|
"logps/rejected": -458.756591796875, |
|
"loss": 0.5426, |
|
"rewards/accuracies": 0.765625, |
|
"rewards/chosen": -2.2905333042144775, |
|
"rewards/margins": 0.7688990831375122, |
|
"rewards/rejected": -3.0594325065612793, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.716275368085953, |
|
"grad_norm": 14.064737088004772, |
|
"learning_rate": 1.1277334291351145e-07, |
|
"logits/chosen": 3.348231077194214, |
|
"logits/rejected": 4.010863780975342, |
|
"logps/chosen": -383.77569580078125, |
|
"logps/rejected": -465.36767578125, |
|
"loss": 0.5372, |
|
"rewards/accuracies": 0.7281249761581421, |
|
"rewards/chosen": -2.3260040283203125, |
|
"rewards/margins": 0.8284826278686523, |
|
"rewards/rejected": -3.1544864177703857, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.7321925984878631, |
|
"grad_norm": 13.977439772366749, |
|
"learning_rate": 1.0137191367132078e-07, |
|
"logits/chosen": 3.446108341217041, |
|
"logits/rejected": 3.9590706825256348, |
|
"logps/chosen": -375.79705810546875, |
|
"logps/rejected": -461.6592712402344, |
|
"loss": 0.5282, |
|
"rewards/accuracies": 0.7250000238418579, |
|
"rewards/chosen": -2.228874683380127, |
|
"rewards/margins": 0.7711628675460815, |
|
"rewards/rejected": -3.000037670135498, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.7481098288897732, |
|
"grad_norm": 15.466730145904025, |
|
"learning_rate": 9.042988532644249e-08, |
|
"logits/chosen": 3.792773485183716, |
|
"logits/rejected": 4.424340724945068, |
|
"logps/chosen": -395.2272033691406, |
|
"logps/rejected": -461.55352783203125, |
|
"loss": 0.5425, |
|
"rewards/accuracies": 0.690625011920929, |
|
"rewards/chosen": -2.369027614593506, |
|
"rewards/margins": 0.6728731393814087, |
|
"rewards/rejected": -3.041900157928467, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.7640270592916832, |
|
"grad_norm": 14.83331875533172, |
|
"learning_rate": 7.998107906142839e-08, |
|
"logits/chosen": 3.4053382873535156, |
|
"logits/rejected": 3.8911843299865723, |
|
"logps/chosen": -388.53155517578125, |
|
"logps/rejected": -455.40185546875, |
|
"loss": 0.5425, |
|
"rewards/accuracies": 0.7281249761581421, |
|
"rewards/chosen": -2.44258189201355, |
|
"rewards/margins": 0.6656660437583923, |
|
"rewards/rejected": -3.108247756958008, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.7799442896935933, |
|
"grad_norm": 15.620806793742632, |
|
"learning_rate": 7.005779153764682e-08, |
|
"logits/chosen": 3.4802298545837402, |
|
"logits/rejected": 3.920941114425659, |
|
"logps/chosen": -378.8732604980469, |
|
"logps/rejected": -466.740478515625, |
|
"loss": 0.5309, |
|
"rewards/accuracies": 0.731249988079071, |
|
"rewards/chosen": -2.4023540019989014, |
|
"rewards/margins": 0.7670234441757202, |
|
"rewards/rejected": -3.169377565383911, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.7958615200955034, |
|
"grad_norm": 17.188173872991122, |
|
"learning_rate": 6.069069506815325e-08, |
|
"logits/chosen": 3.2549641132354736, |
|
"logits/rejected": 4.011484146118164, |
|
"logps/chosen": -376.5075988769531, |
|
"logps/rejected": -464.5885314941406, |
|
"loss": 0.5443, |
|
"rewards/accuracies": 0.731249988079071, |
|
"rewards/chosen": -2.2434165477752686, |
|
"rewards/margins": 0.8864636421203613, |
|
"rewards/rejected": -3.129879951477051, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.8117787504974134, |
|
"grad_norm": 15.502718885907624, |
|
"learning_rate": 5.190874281132851e-08, |
|
"logits/chosen": 3.2985496520996094, |
|
"logits/rejected": 3.910979747772217, |
|
"logps/chosen": -390.97381591796875, |
|
"logps/rejected": -462.67816162109375, |
|
"loss": 0.5418, |
|
"rewards/accuracies": 0.7093750238418579, |
|
"rewards/chosen": -2.4295220375061035, |
|
"rewards/margins": 0.7420149445533752, |
|
"rewards/rejected": -3.171536922454834, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.8276959808993235, |
|
"grad_norm": 14.242739712121452, |
|
"learning_rate": 4.373907927832513e-08, |
|
"logits/chosen": 3.5766899585723877, |
|
"logits/rejected": 4.207441806793213, |
|
"logps/chosen": -382.2040710449219, |
|
"logps/rejected": -460.5889587402344, |
|
"loss": 0.5461, |
|
"rewards/accuracies": 0.6968749761581421, |
|
"rewards/chosen": -2.4501075744628906, |
|
"rewards/margins": 0.7706823348999023, |
|
"rewards/rejected": -3.220790147781372, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.8436132113012336, |
|
"grad_norm": 12.998296781925998, |
|
"learning_rate": 3.620695643093924e-08, |
|
"logits/chosen": 3.5845916271209717, |
|
"logits/rejected": 4.0032148361206055, |
|
"logps/chosen": -372.14898681640625, |
|
"logps/rejected": -455.68212890625, |
|
"loss": 0.5414, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -2.2561497688293457, |
|
"rewards/margins": 0.7765234112739563, |
|
"rewards/rejected": -3.032672643661499, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.8595304417031436, |
|
"grad_norm": 17.292649632838344, |
|
"learning_rate": 2.9335655629243645e-08, |
|
"logits/chosen": 3.523092269897461, |
|
"logits/rejected": 3.9805259704589844, |
|
"logps/chosen": -388.76800537109375, |
|
"logps/rejected": -462.15625, |
|
"loss": 0.5454, |
|
"rewards/accuracies": 0.721875011920929, |
|
"rewards/chosen": -2.4221224784851074, |
|
"rewards/margins": 0.6886903643608093, |
|
"rewards/rejected": -3.1108126640319824, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.8754476721050537, |
|
"grad_norm": 13.631324448665103, |
|
"learning_rate": 2.31464156702382e-08, |
|
"logits/chosen": 3.492096424102783, |
|
"logits/rejected": 4.266209125518799, |
|
"logps/chosen": -393.3036804199219, |
|
"logps/rejected": -445.7330627441406, |
|
"loss": 0.5519, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -2.4094600677490234, |
|
"rewards/margins": 0.6608883738517761, |
|
"rewards/rejected": -3.0703485012054443, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.8913649025069638, |
|
"grad_norm": 15.338497216975412, |
|
"learning_rate": 1.7658367139945228e-08, |
|
"logits/chosen": 3.498994827270508, |
|
"logits/rejected": 3.9349846839904785, |
|
"logps/chosen": -387.2663269042969, |
|
"logps/rejected": -449.8990173339844, |
|
"loss": 0.5364, |
|
"rewards/accuracies": 0.659375011920929, |
|
"rewards/chosen": -2.4634361267089844, |
|
"rewards/margins": 0.5876127481460571, |
|
"rewards/rejected": -3.051048755645752, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.9072821329088738, |
|
"grad_norm": 13.324839693428412, |
|
"learning_rate": 1.2888473281864597e-08, |
|
"logits/chosen": 3.275439500808716, |
|
"logits/rejected": 3.773510694503784, |
|
"logps/chosen": -398.3496398925781, |
|
"logps/rejected": -472.8216247558594, |
|
"loss": 0.5281, |
|
"rewards/accuracies": 0.6968749761581421, |
|
"rewards/chosen": -2.447859287261963, |
|
"rewards/margins": 0.6992862820625305, |
|
"rewards/rejected": -3.1471457481384277, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.9231993633107839, |
|
"grad_norm": 16.191483693504896, |
|
"learning_rate": 8.851477564560061e-09, |
|
"logits/chosen": 3.820493221282959, |
|
"logits/rejected": 4.351908206939697, |
|
"logps/chosen": -389.7720947265625, |
|
"logps/rejected": -459.3812561035156, |
|
"loss": 0.55, |
|
"rewards/accuracies": 0.6968749761581421, |
|
"rewards/chosen": -2.404622793197632, |
|
"rewards/margins": 0.6956027746200562, |
|
"rewards/rejected": -3.1002252101898193, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.939116593712694, |
|
"grad_norm": 12.583560577719238, |
|
"learning_rate": 5.559858110443016e-09, |
|
"logits/chosen": 3.1959145069122314, |
|
"logits/rejected": 3.7669143676757812, |
|
"logps/chosen": -405.40618896484375, |
|
"logps/rejected": -479.1631774902344, |
|
"loss": 0.5251, |
|
"rewards/accuracies": 0.7124999761581421, |
|
"rewards/chosen": -2.4700753688812256, |
|
"rewards/margins": 0.7442026138305664, |
|
"rewards/rejected": -3.214277982711792, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.955033824114604, |
|
"grad_norm": 15.78612631007661, |
|
"learning_rate": 3.023789126611137e-09, |
|
"logits/chosen": 3.5097384452819824, |
|
"logits/rejected": 4.086105823516846, |
|
"logps/chosen": -395.22027587890625, |
|
"logps/rejected": -475.81787109375, |
|
"loss": 0.5376, |
|
"rewards/accuracies": 0.706250011920929, |
|
"rewards/chosen": -2.4702706336975098, |
|
"rewards/margins": 0.8460418581962585, |
|
"rewards/rejected": -3.316312313079834, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.9709510545165141, |
|
"grad_norm": 15.913083563945913, |
|
"learning_rate": 1.2511094569571668e-09, |
|
"logits/chosen": 3.64544939994812, |
|
"logits/rejected": 4.29144811630249, |
|
"logps/chosen": -404.07501220703125, |
|
"logps/rejected": -467.01153564453125, |
|
"loss": 0.5567, |
|
"rewards/accuracies": 0.659375011920929, |
|
"rewards/chosen": -2.5605225563049316, |
|
"rewards/margins": 0.6661086082458496, |
|
"rewards/rejected": -3.2266311645507812, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.9868682849184242, |
|
"grad_norm": 15.29692820991952, |
|
"learning_rate": 2.4729835275189016e-10, |
|
"logits/chosen": 3.532379150390625, |
|
"logits/rejected": 3.938063859939575, |
|
"logps/chosen": -392.1960754394531, |
|
"logps/rejected": -481.100830078125, |
|
"loss": 0.5331, |
|
"rewards/accuracies": 0.7093750238418579, |
|
"rewards/chosen": -2.4987053871154785, |
|
"rewards/margins": 0.7970455288887024, |
|
"rewards/rejected": -3.295750379562378, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.9996020692399522, |
|
"step": 628, |
|
"total_flos": 0.0, |
|
"train_loss": 0.5721431234080321, |
|
"train_runtime": 15529.5131, |
|
"train_samples_per_second": 10.354, |
|
"train_steps_per_second": 0.04 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 628, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|