full_review / trainer_state.json
sumuks's picture
End of training
ea862ef verified
raw
history blame
32.2 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 600,
"global_step": 1770,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01694915254237288,
"grad_norm": 0.9151965379714966,
"learning_rate": 2.824858757062147e-06,
"loss": 2.0144,
"step": 10
},
{
"epoch": 0.03389830508474576,
"grad_norm": 0.3579045534133911,
"learning_rate": 5.649717514124294e-06,
"loss": 1.9709,
"step": 20
},
{
"epoch": 0.05084745762711865,
"grad_norm": 0.26751402020454407,
"learning_rate": 8.47457627118644e-06,
"loss": 1.8976,
"step": 30
},
{
"epoch": 0.06779661016949153,
"grad_norm": 0.17199155688285828,
"learning_rate": 1.1299435028248587e-05,
"loss": 1.832,
"step": 40
},
{
"epoch": 0.0847457627118644,
"grad_norm": 0.15522244572639465,
"learning_rate": 1.4124293785310736e-05,
"loss": 1.7838,
"step": 50
},
{
"epoch": 0.1016949152542373,
"grad_norm": 0.1401771456003189,
"learning_rate": 1.694915254237288e-05,
"loss": 1.7742,
"step": 60
},
{
"epoch": 0.11864406779661017,
"grad_norm": 0.13913491368293762,
"learning_rate": 1.977401129943503e-05,
"loss": 1.7601,
"step": 70
},
{
"epoch": 0.13559322033898305,
"grad_norm": 0.13437530398368835,
"learning_rate": 2.2598870056497175e-05,
"loss": 1.7553,
"step": 80
},
{
"epoch": 0.15254237288135594,
"grad_norm": 0.14121650159358978,
"learning_rate": 2.5423728813559322e-05,
"loss": 1.7318,
"step": 90
},
{
"epoch": 0.1694915254237288,
"grad_norm": 0.15103434026241302,
"learning_rate": 2.8248587570621472e-05,
"loss": 1.7234,
"step": 100
},
{
"epoch": 0.1864406779661017,
"grad_norm": 0.1516140103340149,
"learning_rate": 3.107344632768362e-05,
"loss": 1.731,
"step": 110
},
{
"epoch": 0.2033898305084746,
"grad_norm": 0.14474578201770782,
"learning_rate": 3.389830508474576e-05,
"loss": 1.7048,
"step": 120
},
{
"epoch": 0.22033898305084745,
"grad_norm": 0.1441410928964615,
"learning_rate": 3.672316384180791e-05,
"loss": 1.7091,
"step": 130
},
{
"epoch": 0.23728813559322035,
"grad_norm": 0.14374051988124847,
"learning_rate": 3.954802259887006e-05,
"loss": 1.7163,
"step": 140
},
{
"epoch": 0.2542372881355932,
"grad_norm": 0.15048211812973022,
"learning_rate": 4.2372881355932206e-05,
"loss": 1.7085,
"step": 150
},
{
"epoch": 0.2711864406779661,
"grad_norm": 0.13899974524974823,
"learning_rate": 4.519774011299435e-05,
"loss": 1.6966,
"step": 160
},
{
"epoch": 0.288135593220339,
"grad_norm": 0.14798595011234283,
"learning_rate": 4.80225988700565e-05,
"loss": 1.7086,
"step": 170
},
{
"epoch": 0.3050847457627119,
"grad_norm": 0.1401553452014923,
"learning_rate": 4.999956245830044e-05,
"loss": 1.7086,
"step": 180
},
{
"epoch": 0.3220338983050847,
"grad_norm": 0.14466150104999542,
"learning_rate": 4.9991784365248394e-05,
"loss": 1.6885,
"step": 190
},
{
"epoch": 0.3389830508474576,
"grad_norm": 0.1339222490787506,
"learning_rate": 4.997428660526443e-05,
"loss": 1.705,
"step": 200
},
{
"epoch": 0.3559322033898305,
"grad_norm": 0.13571041822433472,
"learning_rate": 4.994707598348085e-05,
"loss": 1.6889,
"step": 210
},
{
"epoch": 0.3728813559322034,
"grad_norm": 0.12768986821174622,
"learning_rate": 4.991016308250258e-05,
"loss": 1.6939,
"step": 220
},
{
"epoch": 0.3898305084745763,
"grad_norm": 0.1278401017189026,
"learning_rate": 4.9863562258291494e-05,
"loss": 1.699,
"step": 230
},
{
"epoch": 0.4067796610169492,
"grad_norm": 0.1307385265827179,
"learning_rate": 4.980729163458312e-05,
"loss": 1.6969,
"step": 240
},
{
"epoch": 0.423728813559322,
"grad_norm": 0.12608782947063446,
"learning_rate": 4.974137309583804e-05,
"loss": 1.6931,
"step": 250
},
{
"epoch": 0.4406779661016949,
"grad_norm": 0.9130762815475464,
"learning_rate": 4.966583227873079e-05,
"loss": 1.7061,
"step": 260
},
{
"epoch": 0.4576271186440678,
"grad_norm": 0.12975232303142548,
"learning_rate": 4.9580698562179297e-05,
"loss": 1.6956,
"step": 270
},
{
"epoch": 0.4745762711864407,
"grad_norm": 0.1291048377752304,
"learning_rate": 4.948600505591905e-05,
"loss": 1.7004,
"step": 280
},
{
"epoch": 0.4915254237288136,
"grad_norm": 0.12629887461662292,
"learning_rate": 4.938178858762622e-05,
"loss": 1.693,
"step": 290
},
{
"epoch": 0.5084745762711864,
"grad_norm": 0.12622225284576416,
"learning_rate": 4.926808968859483e-05,
"loss": 1.6819,
"step": 300
},
{
"epoch": 0.5254237288135594,
"grad_norm": 0.12781433761119843,
"learning_rate": 4.9144952577973596e-05,
"loss": 1.6984,
"step": 310
},
{
"epoch": 0.5423728813559322,
"grad_norm": 0.1239006370306015,
"learning_rate": 4.90124251455684e-05,
"loss": 1.6875,
"step": 320
},
{
"epoch": 0.559322033898305,
"grad_norm": 0.1296393871307373,
"learning_rate": 4.88705589332173e-05,
"loss": 1.6758,
"step": 330
},
{
"epoch": 0.576271186440678,
"grad_norm": 0.12318539619445801,
"learning_rate": 4.871940911474513e-05,
"loss": 1.695,
"step": 340
},
{
"epoch": 0.5932203389830508,
"grad_norm": 0.12802407145500183,
"learning_rate": 4.8559034474505614e-05,
"loss": 1.6862,
"step": 350
},
{
"epoch": 0.6101694915254238,
"grad_norm": 0.12415086477994919,
"learning_rate": 4.838949738451929e-05,
"loss": 1.6652,
"step": 360
},
{
"epoch": 0.6271186440677966,
"grad_norm": 0.12488778680562973,
"learning_rate": 4.821086378021608e-05,
"loss": 1.6898,
"step": 370
},
{
"epoch": 0.6440677966101694,
"grad_norm": 0.12341786175966263,
"learning_rate": 4.802320313479214e-05,
"loss": 1.6766,
"step": 380
},
{
"epoch": 0.6610169491525424,
"grad_norm": 0.12222524732351303,
"learning_rate": 4.7826588432190614e-05,
"loss": 1.6838,
"step": 390
},
{
"epoch": 0.6779661016949152,
"grad_norm": 0.19455482065677643,
"learning_rate": 4.76210961387172e-05,
"loss": 1.6702,
"step": 400
},
{
"epoch": 0.6949152542372882,
"grad_norm": 0.12282032519578934,
"learning_rate": 4.7406806173301285e-05,
"loss": 1.6793,
"step": 410
},
{
"epoch": 0.711864406779661,
"grad_norm": 0.12586097419261932,
"learning_rate": 4.7183801876414294e-05,
"loss": 1.6611,
"step": 420
},
{
"epoch": 0.7288135593220338,
"grad_norm": 0.12690460681915283,
"learning_rate": 4.695216997765737e-05,
"loss": 1.6896,
"step": 430
},
{
"epoch": 0.7457627118644068,
"grad_norm": 0.1262531876564026,
"learning_rate": 4.6712000562031e-05,
"loss": 1.6756,
"step": 440
},
{
"epoch": 0.7627118644067796,
"grad_norm": 0.12429999560117722,
"learning_rate": 4.6463387034899645e-05,
"loss": 1.6658,
"step": 450
},
{
"epoch": 0.7796610169491526,
"grad_norm": 0.12122652679681778,
"learning_rate": 4.6206426085665046e-05,
"loss": 1.6841,
"step": 460
},
{
"epoch": 0.7966101694915254,
"grad_norm": 0.1276836097240448,
"learning_rate": 4.594121765016229e-05,
"loss": 1.6947,
"step": 470
},
{
"epoch": 0.8135593220338984,
"grad_norm": 0.12912869453430176,
"learning_rate": 4.5667864871793345e-05,
"loss": 1.6888,
"step": 480
},
{
"epoch": 0.8305084745762712,
"grad_norm": 0.12462390959262848,
"learning_rate": 4.538647406141308e-05,
"loss": 1.6675,
"step": 490
},
{
"epoch": 0.847457627118644,
"grad_norm": 0.12297125160694122,
"learning_rate": 4.509715465598344e-05,
"loss": 1.6828,
"step": 500
},
{
"epoch": 0.864406779661017,
"grad_norm": 0.12475843727588654,
"learning_rate": 4.480001917601185e-05,
"loss": 1.6713,
"step": 510
},
{
"epoch": 0.8813559322033898,
"grad_norm": 0.1244443878531456,
"learning_rate": 4.449518318179029e-05,
"loss": 1.6544,
"step": 520
},
{
"epoch": 0.8983050847457628,
"grad_norm": 0.12166355550289154,
"learning_rate": 4.418276522845231e-05,
"loss": 1.6679,
"step": 530
},
{
"epoch": 0.9152542372881356,
"grad_norm": 0.1233435645699501,
"learning_rate": 4.386288681986516e-05,
"loss": 1.6704,
"step": 540
},
{
"epoch": 0.9322033898305084,
"grad_norm": 0.12299688160419464,
"learning_rate": 4.353567236137521e-05,
"loss": 1.6759,
"step": 550
},
{
"epoch": 0.9491525423728814,
"grad_norm": 0.12501336634159088,
"learning_rate": 4.3201249111424877e-05,
"loss": 1.6617,
"step": 560
},
{
"epoch": 0.9661016949152542,
"grad_norm": 0.1281379759311676,
"learning_rate": 4.2859747132060006e-05,
"loss": 1.6632,
"step": 570
},
{
"epoch": 0.9830508474576272,
"grad_norm": 0.12356545031070709,
"learning_rate": 4.251129923834685e-05,
"loss": 1.6841,
"step": 580
},
{
"epoch": 1.0,
"grad_norm": 0.12595489621162415,
"learning_rate": 4.215604094671835e-05,
"loss": 1.6448,
"step": 590
},
{
"epoch": 1.0169491525423728,
"grad_norm": 0.1292751580476761,
"learning_rate": 4.1794110422269825e-05,
"loss": 1.6261,
"step": 600
},
{
"epoch": 1.0169491525423728,
"eval_loss": 1.6485204696655273,
"eval_runtime": 4.4104,
"eval_samples_per_second": 86.387,
"eval_steps_per_second": 10.883,
"step": 600
},
{
"epoch": 1.0338983050847457,
"grad_norm": 0.13464362919330597,
"learning_rate": 4.142564842502461e-05,
"loss": 1.6501,
"step": 610
},
{
"epoch": 1.0508474576271187,
"grad_norm": 0.13430601358413696,
"learning_rate": 4.1050798255190405e-05,
"loss": 1.6435,
"step": 620
},
{
"epoch": 1.0677966101694916,
"grad_norm": 0.1346493363380432,
"learning_rate": 4.0669705697427754e-05,
"loss": 1.6567,
"step": 630
},
{
"epoch": 1.0847457627118644,
"grad_norm": 0.13532420992851257,
"learning_rate": 4.0282518964152296e-05,
"loss": 1.6527,
"step": 640
},
{
"epoch": 1.1016949152542372,
"grad_norm": 0.1375868022441864,
"learning_rate": 3.988938863789278e-05,
"loss": 1.6408,
"step": 650
},
{
"epoch": 1.11864406779661,
"grad_norm": 0.13993752002716064,
"learning_rate": 3.949046761272736e-05,
"loss": 1.6441,
"step": 660
},
{
"epoch": 1.1355932203389831,
"grad_norm": 0.1394631415605545,
"learning_rate": 3.908591103482088e-05,
"loss": 1.6454,
"step": 670
},
{
"epoch": 1.152542372881356,
"grad_norm": 0.1412103921175003,
"learning_rate": 3.867587624208627e-05,
"loss": 1.6689,
"step": 680
},
{
"epoch": 1.1694915254237288,
"grad_norm": 0.1373807042837143,
"learning_rate": 3.826052270299356e-05,
"loss": 1.6475,
"step": 690
},
{
"epoch": 1.1864406779661016,
"grad_norm": 0.14468206465244293,
"learning_rate": 3.784001195455027e-05,
"loss": 1.6469,
"step": 700
},
{
"epoch": 1.2033898305084745,
"grad_norm": 0.14381623268127441,
"learning_rate": 3.7414507539477315e-05,
"loss": 1.6335,
"step": 710
},
{
"epoch": 1.2203389830508475,
"grad_norm": 0.1429220587015152,
"learning_rate": 3.698417494260494e-05,
"loss": 1.6289,
"step": 720
},
{
"epoch": 1.2372881355932204,
"grad_norm": 0.14810022711753845,
"learning_rate": 3.654918152651319e-05,
"loss": 1.6428,
"step": 730
},
{
"epoch": 1.2542372881355932,
"grad_norm": 0.14096328616142273,
"learning_rate": 3.610969646644222e-05,
"loss": 1.6592,
"step": 740
},
{
"epoch": 1.271186440677966,
"grad_norm": 0.14148060977458954,
"learning_rate": 3.566589068449761e-05,
"loss": 1.6251,
"step": 750
},
{
"epoch": 1.288135593220339,
"grad_norm": 0.14239969849586487,
"learning_rate": 3.5217936783176216e-05,
"loss": 1.6708,
"step": 760
},
{
"epoch": 1.305084745762712,
"grad_norm": 0.14554665982723236,
"learning_rate": 3.476600897823864e-05,
"loss": 1.6334,
"step": 770
},
{
"epoch": 1.3220338983050848,
"grad_norm": 0.14560729265213013,
"learning_rate": 3.431028303095415e-05,
"loss": 1.6475,
"step": 780
},
{
"epoch": 1.3389830508474576,
"grad_norm": 0.14253520965576172,
"learning_rate": 3.385093617974458e-05,
"loss": 1.6418,
"step": 790
},
{
"epoch": 1.3559322033898304,
"grad_norm": 0.1438617706298828,
"learning_rate": 3.338814707125377e-05,
"loss": 1.6296,
"step": 800
},
{
"epoch": 1.3728813559322033,
"grad_norm": 0.14633050560951233,
"learning_rate": 3.2922095690869224e-05,
"loss": 1.636,
"step": 810
},
{
"epoch": 1.3898305084745763,
"grad_norm": 0.1484278440475464,
"learning_rate": 3.2452963292723305e-05,
"loss": 1.6156,
"step": 820
},
{
"epoch": 1.4067796610169492,
"grad_norm": 0.14433981478214264,
"learning_rate": 3.1980932329200756e-05,
"loss": 1.6474,
"step": 830
},
{
"epoch": 1.423728813559322,
"grad_norm": 0.1485067456960678,
"learning_rate": 3.150618637998041e-05,
"loss": 1.6356,
"step": 840
},
{
"epoch": 1.4406779661016949,
"grad_norm": 0.14338724315166473,
"learning_rate": 3.1028910080638354e-05,
"loss": 1.6396,
"step": 850
},
{
"epoch": 1.457627118644068,
"grad_norm": 0.14848238229751587,
"learning_rate": 3.054928905084049e-05,
"loss": 1.6402,
"step": 860
},
{
"epoch": 1.4745762711864407,
"grad_norm": 0.14883729815483093,
"learning_rate": 3.006750982215234e-05,
"loss": 1.6282,
"step": 870
},
{
"epoch": 1.4915254237288136,
"grad_norm": 0.15096916258335114,
"learning_rate": 2.9583759765494267e-05,
"loss": 1.6402,
"step": 880
},
{
"epoch": 1.5084745762711864,
"grad_norm": 0.1506650447845459,
"learning_rate": 2.9098227018270134e-05,
"loss": 1.6392,
"step": 890
},
{
"epoch": 1.5254237288135593,
"grad_norm": 0.14937371015548706,
"learning_rate": 2.8611100411198037e-05,
"loss": 1.6472,
"step": 900
},
{
"epoch": 1.542372881355932,
"grad_norm": 0.14744701981544495,
"learning_rate": 2.8122569394871225e-05,
"loss": 1.6215,
"step": 910
},
{
"epoch": 1.559322033898305,
"grad_norm": 0.14676077663898468,
"learning_rate": 2.763282396607803e-05,
"loss": 1.6313,
"step": 920
},
{
"epoch": 1.576271186440678,
"grad_norm": 0.14905640482902527,
"learning_rate": 2.7142054593909422e-05,
"loss": 1.6452,
"step": 930
},
{
"epoch": 1.5932203389830508,
"grad_norm": 0.14924922585487366,
"learning_rate": 2.6650452145682762e-05,
"loss": 1.6386,
"step": 940
},
{
"epoch": 1.6101694915254239,
"grad_norm": 0.15311497449874878,
"learning_rate": 2.615820781271075e-05,
"loss": 1.6301,
"step": 950
},
{
"epoch": 1.6271186440677967,
"grad_norm": 0.15127715468406677,
"learning_rate": 2.566551303594437e-05,
"loss": 1.6247,
"step": 960
},
{
"epoch": 1.6440677966101696,
"grad_norm": 0.15096339583396912,
"learning_rate": 2.5172559431518727e-05,
"loss": 1.6387,
"step": 970
},
{
"epoch": 1.6610169491525424,
"grad_norm": 0.15415184199810028,
"learning_rate": 2.467953871623073e-05,
"loss": 1.6289,
"step": 980
},
{
"epoch": 1.6779661016949152,
"grad_norm": 0.14868630468845367,
"learning_rate": 2.4186642632977697e-05,
"loss": 1.6339,
"step": 990
},
{
"epoch": 1.694915254237288,
"grad_norm": 0.1516588181257248,
"learning_rate": 2.3694062876185738e-05,
"loss": 1.6172,
"step": 1000
},
{
"epoch": 1.711864406779661,
"grad_norm": 0.15194956958293915,
"learning_rate": 2.3201991017257007e-05,
"loss": 1.6169,
"step": 1010
},
{
"epoch": 1.7288135593220337,
"grad_norm": 0.15378808975219727,
"learning_rate": 2.2710618430064843e-05,
"loss": 1.6156,
"step": 1020
},
{
"epoch": 1.7457627118644068,
"grad_norm": 0.15048706531524658,
"learning_rate": 2.222013621652565e-05,
"loss": 1.6348,
"step": 1030
},
{
"epoch": 1.7627118644067796,
"grad_norm": 0.1514016091823578,
"learning_rate": 2.173073513227667e-05,
"loss": 1.6162,
"step": 1040
},
{
"epoch": 1.7796610169491527,
"grad_norm": 0.15484623610973358,
"learning_rate": 2.1242605512488248e-05,
"loss": 1.6162,
"step": 1050
},
{
"epoch": 1.7966101694915255,
"grad_norm": 0.1544799506664276,
"learning_rate": 2.0755937197839802e-05,
"loss": 1.6146,
"step": 1060
},
{
"epoch": 1.8135593220338984,
"grad_norm": 0.15184859931468964,
"learning_rate": 2.0270919460688055e-05,
"loss": 1.6178,
"step": 1070
},
{
"epoch": 1.8305084745762712,
"grad_norm": 0.154934361577034,
"learning_rate": 1.9787740931456165e-05,
"loss": 1.6287,
"step": 1080
},
{
"epoch": 1.847457627118644,
"grad_norm": 0.15421293675899506,
"learning_rate": 1.9306589525272756e-05,
"loss": 1.6348,
"step": 1090
},
{
"epoch": 1.8644067796610169,
"grad_norm": 0.15633705258369446,
"learning_rate": 1.8827652368888888e-05,
"loss": 1.6295,
"step": 1100
},
{
"epoch": 1.8813559322033897,
"grad_norm": 0.15599501132965088,
"learning_rate": 1.835111572790183e-05,
"loss": 1.629,
"step": 1110
},
{
"epoch": 1.8983050847457628,
"grad_norm": 0.1540609449148178,
"learning_rate": 1.7877164934313628e-05,
"loss": 1.6127,
"step": 1120
},
{
"epoch": 1.9152542372881356,
"grad_norm": 0.1557774394750595,
"learning_rate": 1.740598431445282e-05,
"loss": 1.6345,
"step": 1130
},
{
"epoch": 1.9322033898305084,
"grad_norm": 0.15810750424861908,
"learning_rate": 1.6937757117287278e-05,
"loss": 1.6291,
"step": 1140
},
{
"epoch": 1.9491525423728815,
"grad_norm": 0.15605640411376953,
"learning_rate": 1.6472665443155983e-05,
"loss": 1.6311,
"step": 1150
},
{
"epoch": 1.9661016949152543,
"grad_norm": 0.15914802253246307,
"learning_rate": 1.6010890172947606e-05,
"loss": 1.6082,
"step": 1160
},
{
"epoch": 1.9830508474576272,
"grad_norm": 0.15677359700202942,
"learning_rate": 1.5552610897753292e-05,
"loss": 1.6266,
"step": 1170
},
{
"epoch": 2.0,
"grad_norm": 0.15751083195209503,
"learning_rate": 1.509800584902108e-05,
"loss": 1.6265,
"step": 1180
},
{
"epoch": 2.016949152542373,
"grad_norm": 0.17009219527244568,
"learning_rate": 1.4647251829239139e-05,
"loss": 1.5776,
"step": 1190
},
{
"epoch": 2.0338983050847457,
"grad_norm": 0.16682426631450653,
"learning_rate": 1.4200524143174677e-05,
"loss": 1.5922,
"step": 1200
},
{
"epoch": 2.0338983050847457,
"eval_loss": 1.6305924654006958,
"eval_runtime": 4.3076,
"eval_samples_per_second": 88.449,
"eval_steps_per_second": 11.143,
"step": 1200
},
{
"epoch": 2.0508474576271185,
"grad_norm": 0.16748355329036713,
"learning_rate": 1.3757996529695411e-05,
"loss": 1.5948,
"step": 1210
},
{
"epoch": 2.0677966101694913,
"grad_norm": 0.17169494926929474,
"learning_rate": 1.331984109420001e-05,
"loss": 1.5884,
"step": 1220
},
{
"epoch": 2.084745762711864,
"grad_norm": 0.17144790291786194,
"learning_rate": 1.2886228241683749e-05,
"loss": 1.6067,
"step": 1230
},
{
"epoch": 2.1016949152542375,
"grad_norm": 0.16989319026470184,
"learning_rate": 1.2457326610465642e-05,
"loss": 1.6103,
"step": 1240
},
{
"epoch": 2.1186440677966103,
"grad_norm": 0.1745283454656601,
"learning_rate": 1.2033303006602444e-05,
"loss": 1.5979,
"step": 1250
},
{
"epoch": 2.135593220338983,
"grad_norm": 0.17248323559761047,
"learning_rate": 1.1614322339015484e-05,
"loss": 1.5786,
"step": 1260
},
{
"epoch": 2.152542372881356,
"grad_norm": 0.16860243678092957,
"learning_rate": 1.1200547555355067e-05,
"loss": 1.5903,
"step": 1270
},
{
"epoch": 2.169491525423729,
"grad_norm": 0.17552419006824493,
"learning_rate": 1.0792139578627865e-05,
"loss": 1.5913,
"step": 1280
},
{
"epoch": 2.1864406779661016,
"grad_norm": 0.17908188700675964,
"learning_rate": 1.0389257244611602e-05,
"loss": 1.592,
"step": 1290
},
{
"epoch": 2.2033898305084745,
"grad_norm": 0.17985224723815918,
"learning_rate": 9.992057240081487e-06,
"loss": 1.5968,
"step": 1300
},
{
"epoch": 2.2203389830508473,
"grad_norm": 0.174930602312088,
"learning_rate": 9.600694041872518e-06,
"loss": 1.5893,
"step": 1310
},
{
"epoch": 2.23728813559322,
"grad_norm": 0.17611059546470642,
"learning_rate": 9.215319856801158e-06,
"loss": 1.5861,
"step": 1320
},
{
"epoch": 2.2542372881355934,
"grad_norm": 0.1744425892829895,
"learning_rate": 8.836084562469921e-06,
"loss": 1.606,
"step": 1330
},
{
"epoch": 2.2711864406779663,
"grad_norm": 0.17969445884227753,
"learning_rate": 8.463135648977773e-06,
"loss": 1.5911,
"step": 1340
},
{
"epoch": 2.288135593220339,
"grad_norm": 0.1790507286787033,
"learning_rate": 8.096618161559142e-06,
"loss": 1.5832,
"step": 1350
},
{
"epoch": 2.305084745762712,
"grad_norm": 0.17973437905311584,
"learning_rate": 7.736674644173695e-06,
"loss": 1.5868,
"step": 1360
},
{
"epoch": 2.3220338983050848,
"grad_norm": 0.17950280010700226,
"learning_rate": 7.383445084068907e-06,
"loss": 1.5788,
"step": 1370
},
{
"epoch": 2.3389830508474576,
"grad_norm": 0.17697365581989288,
"learning_rate": 7.037066857337058e-06,
"loss": 1.5769,
"step": 1380
},
{
"epoch": 2.3559322033898304,
"grad_norm": 0.17717047035694122,
"learning_rate": 6.69767467548762e-06,
"loss": 1.5676,
"step": 1390
},
{
"epoch": 2.3728813559322033,
"grad_norm": 0.17924557626247406,
"learning_rate": 6.365400533056012e-06,
"loss": 1.5896,
"step": 1400
},
{
"epoch": 2.389830508474576,
"grad_norm": 0.17875994741916656,
"learning_rate": 6.040373656269041e-06,
"loss": 1.5698,
"step": 1410
},
{
"epoch": 2.406779661016949,
"grad_norm": 0.18066538870334625,
"learning_rate": 5.722720452786928e-06,
"loss": 1.5881,
"step": 1420
},
{
"epoch": 2.423728813559322,
"grad_norm": 0.18073728680610657,
"learning_rate": 5.412564462541544e-06,
"loss": 1.5864,
"step": 1430
},
{
"epoch": 2.440677966101695,
"grad_norm": 0.18556879460811615,
"learning_rate": 5.110026309689922e-06,
"loss": 1.592,
"step": 1440
},
{
"epoch": 2.457627118644068,
"grad_norm": 0.17982180416584015,
"learning_rate": 4.815223655701812e-06,
"loss": 1.5912,
"step": 1450
},
{
"epoch": 2.4745762711864407,
"grad_norm": 0.1827942430973053,
"learning_rate": 4.5282711535994115e-06,
"loss": 1.5732,
"step": 1460
},
{
"epoch": 2.4915254237288136,
"grad_norm": 0.18152937293052673,
"learning_rate": 4.2492804033671145e-06,
"loss": 1.5994,
"step": 1470
},
{
"epoch": 2.5084745762711864,
"grad_norm": 0.18349309265613556,
"learning_rate": 3.978359908548746e-06,
"loss": 1.5906,
"step": 1480
},
{
"epoch": 2.5254237288135593,
"grad_norm": 0.18495327234268188,
"learning_rate": 3.71561503404885e-06,
"loss": 1.593,
"step": 1490
},
{
"epoch": 2.542372881355932,
"grad_norm": 0.18758471310138702,
"learning_rate": 3.4611479651548457e-06,
"loss": 1.6067,
"step": 1500
},
{
"epoch": 2.559322033898305,
"grad_norm": 0.18525028228759766,
"learning_rate": 3.2150576677956458e-06,
"loss": 1.5926,
"step": 1510
},
{
"epoch": 2.576271186440678,
"grad_norm": 0.17536324262619019,
"learning_rate": 2.977439850052366e-06,
"loss": 1.5849,
"step": 1520
},
{
"epoch": 2.593220338983051,
"grad_norm": 0.18296369910240173,
"learning_rate": 2.7483869249360915e-06,
"loss": 1.5828,
"step": 1530
},
{
"epoch": 2.610169491525424,
"grad_norm": 0.1857452690601349,
"learning_rate": 2.527987974447113e-06,
"loss": 1.5802,
"step": 1540
},
{
"epoch": 2.6271186440677967,
"grad_norm": 0.18499261140823364,
"learning_rate": 2.316328714929633e-06,
"loss": 1.5869,
"step": 1550
},
{
"epoch": 2.6440677966101696,
"grad_norm": 0.18661075830459595,
"learning_rate": 2.113491463735437e-06,
"loss": 1.5813,
"step": 1560
},
{
"epoch": 2.6610169491525424,
"grad_norm": 0.18289625644683838,
"learning_rate": 1.9195551072095086e-06,
"loss": 1.5751,
"step": 1570
},
{
"epoch": 2.6779661016949152,
"grad_norm": 0.1815192848443985,
"learning_rate": 1.7345950700099617e-06,
"loss": 1.5887,
"step": 1580
},
{
"epoch": 2.694915254237288,
"grad_norm": 0.17822951078414917,
"learning_rate": 1.558683285774304e-06,
"loss": 1.5868,
"step": 1590
},
{
"epoch": 2.711864406779661,
"grad_norm": 0.17891579866409302,
"learning_rate": 1.3918881691434128e-06,
"loss": 1.5841,
"step": 1600
},
{
"epoch": 2.7288135593220337,
"grad_norm": 0.18717658519744873,
"learning_rate": 1.2342745891540857e-06,
"loss": 1.5783,
"step": 1610
},
{
"epoch": 2.7457627118644066,
"grad_norm": 0.1843055635690689,
"learning_rate": 1.0859038440105164e-06,
"loss": 1.5761,
"step": 1620
},
{
"epoch": 2.7627118644067794,
"grad_norm": 0.18406164646148682,
"learning_rate": 9.46833637244568e-07,
"loss": 1.5948,
"step": 1630
},
{
"epoch": 2.7796610169491527,
"grad_norm": 0.1829424947500229,
"learning_rate": 8.171180552740048e-07,
"loss": 1.5791,
"step": 1640
},
{
"epoch": 2.7966101694915255,
"grad_norm": 0.18232209980487823,
"learning_rate": 6.9680754636752e-07,
"loss": 1.6034,
"step": 1650
},
{
"epoch": 2.8135593220338984,
"grad_norm": 0.18290942907333374,
"learning_rate": 5.859489010246494e-07,
"loss": 1.5902,
"step": 1660
},
{
"epoch": 2.830508474576271,
"grad_norm": 0.17814739048480988,
"learning_rate": 4.845852337782953e-07,
"loss": 1.5887,
"step": 1670
},
{
"epoch": 2.847457627118644,
"grad_norm": 0.1855529248714447,
"learning_rate": 3.9275596642685543e-07,
"loss": 1.5685,
"step": 1680
},
{
"epoch": 2.864406779661017,
"grad_norm": 0.1889604926109314,
"learning_rate": 3.1049681270249044e-07,
"loss": 1.5738,
"step": 1690
},
{
"epoch": 2.8813559322033897,
"grad_norm": 0.18516628444194794,
"learning_rate": 2.3783976438156442e-07,
"loss": 1.5753,
"step": 1700
},
{
"epoch": 2.898305084745763,
"grad_norm": 0.18402761220932007,
"learning_rate": 1.7481307884256727e-07,
"loss": 1.5773,
"step": 1710
},
{
"epoch": 2.915254237288136,
"grad_norm": 0.18513433635234833,
"learning_rate": 1.2144126807641376e-07,
"loss": 1.5865,
"step": 1720
},
{
"epoch": 2.9322033898305087,
"grad_norm": 0.1811605542898178,
"learning_rate": 7.774508915337209e-08,
"loss": 1.5831,
"step": 1730
},
{
"epoch": 2.9491525423728815,
"grad_norm": 0.17931722104549408,
"learning_rate": 4.374153615033794e-08,
"loss": 1.5919,
"step": 1740
},
{
"epoch": 2.9661016949152543,
"grad_norm": 0.18402303755283356,
"learning_rate": 1.9443833541599178e-08,
"loss": 1.5807,
"step": 1750
},
{
"epoch": 2.983050847457627,
"grad_norm": 0.18316687643527985,
"learning_rate": 4.8614310556388544e-09,
"loss": 1.5941,
"step": 1760
},
{
"epoch": 3.0,
"grad_norm": 0.1795782446861267,
"learning_rate": 0.0,
"loss": 1.5941,
"step": 1770
},
{
"epoch": 3.0,
"step": 1770,
"total_flos": 1.0291750314285466e+19,
"train_loss": 1.6438454094579664,
"train_runtime": 3711.4901,
"train_samples_per_second": 30.48,
"train_steps_per_second": 0.477
}
],
"logging_steps": 10,
"max_steps": 1770,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.0291750314285466e+19,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}