Safetensors
qwen2
reasoning
ptrdvn's picture
Upload folder using huggingface_hub
465d606 verified
raw
history blame
19.7 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 10,
"global_step": 98,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01020408163265306,
"grad_norm": 1.5935372114181519,
"learning_rate": 1e-05,
"loss": 1.1572,
"step": 1
},
{
"epoch": 0.02040816326530612,
"grad_norm": 1.8600682020187378,
"learning_rate": 9.997377845227577e-06,
"loss": 1.3198,
"step": 2
},
{
"epoch": 0.030612244897959183,
"grad_norm": 1.662075400352478,
"learning_rate": 9.98951413118856e-06,
"loss": 1.0636,
"step": 3
},
{
"epoch": 0.04081632653061224,
"grad_norm": 1.6947352886199951,
"learning_rate": 9.97641710583307e-06,
"loss": 1.2853,
"step": 4
},
{
"epoch": 0.05102040816326531,
"grad_norm": 1.231564998626709,
"learning_rate": 9.958100506132127e-06,
"loss": 1.0573,
"step": 5
},
{
"epoch": 0.061224489795918366,
"grad_norm": 1.550899863243103,
"learning_rate": 9.934583543669454e-06,
"loss": 1.5486,
"step": 6
},
{
"epoch": 0.07142857142857142,
"grad_norm": 1.017148733139038,
"learning_rate": 9.905890884491196e-06,
"loss": 0.8672,
"step": 7
},
{
"epoch": 0.08163265306122448,
"grad_norm": 1.2291380167007446,
"learning_rate": 9.872052623234632e-06,
"loss": 1.132,
"step": 8
},
{
"epoch": 0.09183673469387756,
"grad_norm": 1.1650055646896362,
"learning_rate": 9.833104251563058e-06,
"loss": 1.2419,
"step": 9
},
{
"epoch": 0.10204081632653061,
"grad_norm": 0.9753865003585815,
"learning_rate": 9.789086620939936e-06,
"loss": 0.9946,
"step": 10
},
{
"epoch": 0.10204081632653061,
"eval_loss": 0.9807827472686768,
"eval_runtime": 0.2604,
"eval_samples_per_second": 30.717,
"eval_steps_per_second": 3.84,
"step": 10
},
{
"epoch": 0.11224489795918367,
"grad_norm": 1.191615343093872,
"learning_rate": 9.740045899781353e-06,
"loss": 1.3435,
"step": 11
},
{
"epoch": 0.12244897959183673,
"grad_norm": 0.9799769520759583,
"learning_rate": 9.68603352503172e-06,
"loss": 0.7463,
"step": 12
},
{
"epoch": 0.1326530612244898,
"grad_norm": 1.0710457563400269,
"learning_rate": 9.627106148213521e-06,
"loss": 1.329,
"step": 13
},
{
"epoch": 0.14285714285714285,
"grad_norm": 1.071432113647461,
"learning_rate": 9.563325576007702e-06,
"loss": 1.3815,
"step": 14
},
{
"epoch": 0.15306122448979592,
"grad_norm": 0.9442039132118225,
"learning_rate": 9.494758705426978e-06,
"loss": 1.0305,
"step": 15
},
{
"epoch": 0.16326530612244897,
"grad_norm": 1.022035837173462,
"learning_rate": 9.421477453650118e-06,
"loss": 1.3317,
"step": 16
},
{
"epoch": 0.17346938775510204,
"grad_norm": 1.0751738548278809,
"learning_rate": 9.343558682590757e-06,
"loss": 1.323,
"step": 17
},
{
"epoch": 0.1836734693877551,
"grad_norm": 1.0566898584365845,
"learning_rate": 9.261084118279846e-06,
"loss": 1.5154,
"step": 18
},
{
"epoch": 0.19387755102040816,
"grad_norm": 1.0009628534317017,
"learning_rate": 9.174140265146356e-06,
"loss": 1.2371,
"step": 19
},
{
"epoch": 0.20408163265306123,
"grad_norm": 0.9259373545646667,
"learning_rate": 9.082818315286054e-06,
"loss": 1.042,
"step": 20
},
{
"epoch": 0.20408163265306123,
"eval_loss": 0.9510844945907593,
"eval_runtime": 0.2601,
"eval_samples_per_second": 30.76,
"eval_steps_per_second": 3.845,
"step": 20
},
{
"epoch": 0.21428571428571427,
"grad_norm": 0.9078995585441589,
"learning_rate": 8.987214052813605e-06,
"loss": 1.07,
"step": 21
},
{
"epoch": 0.22448979591836735,
"grad_norm": 0.9820942282676697,
"learning_rate": 8.887427753398249e-06,
"loss": 1.2102,
"step": 22
},
{
"epoch": 0.23469387755102042,
"grad_norm": 1.0587691068649292,
"learning_rate": 8.783564079088478e-06,
"loss": 1.0913,
"step": 23
},
{
"epoch": 0.24489795918367346,
"grad_norm": 0.8137860894203186,
"learning_rate": 8.675731968536004e-06,
"loss": 0.8923,
"step": 24
},
{
"epoch": 0.25510204081632654,
"grad_norm": 0.9471170902252197,
"learning_rate": 8.564044522734147e-06,
"loss": 1.1723,
"step": 25
},
{
"epoch": 0.2653061224489796,
"grad_norm": 0.8885685205459595,
"learning_rate": 8.448618886390523e-06,
"loss": 1.1206,
"step": 26
},
{
"epoch": 0.2755102040816326,
"grad_norm": 0.8613781332969666,
"learning_rate": 8.329576125058406e-06,
"loss": 1.1261,
"step": 27
},
{
"epoch": 0.2857142857142857,
"grad_norm": 0.9407626986503601,
"learning_rate": 8.207041098155701e-06,
"loss": 1.2727,
"step": 28
},
{
"epoch": 0.29591836734693877,
"grad_norm": 0.8954651951789856,
"learning_rate": 8.081142328004638e-06,
"loss": 1.2197,
"step": 29
},
{
"epoch": 0.30612244897959184,
"grad_norm": 0.8864608407020569,
"learning_rate": 7.952011865029614e-06,
"loss": 1.1236,
"step": 30
},
{
"epoch": 0.30612244897959184,
"eval_loss": 0.9308969974517822,
"eval_runtime": 0.2599,
"eval_samples_per_second": 30.775,
"eval_steps_per_second": 3.847,
"step": 30
},
{
"epoch": 0.3163265306122449,
"grad_norm": 0.8076571226119995,
"learning_rate": 7.819785149254534e-06,
"loss": 0.9098,
"step": 31
},
{
"epoch": 0.32653061224489793,
"grad_norm": 0.8902921080589294,
"learning_rate": 7.68460086824492e-06,
"loss": 1.0564,
"step": 32
},
{
"epoch": 0.336734693877551,
"grad_norm": 0.9010928273200989,
"learning_rate": 7.546600811643816e-06,
"loss": 1.0174,
"step": 33
},
{
"epoch": 0.3469387755102041,
"grad_norm": 0.8672641515731812,
"learning_rate": 7.405929722454026e-06,
"loss": 1.1875,
"step": 34
},
{
"epoch": 0.35714285714285715,
"grad_norm": 0.7878830432891846,
"learning_rate": 7.262735145222696e-06,
"loss": 0.9453,
"step": 35
},
{
"epoch": 0.3673469387755102,
"grad_norm": 0.85221266746521,
"learning_rate": 7.117167271287453e-06,
"loss": 1.0147,
"step": 36
},
{
"epoch": 0.37755102040816324,
"grad_norm": 0.8591777682304382,
"learning_rate": 6.969378781246436e-06,
"loss": 0.995,
"step": 37
},
{
"epoch": 0.3877551020408163,
"grad_norm": 0.9585978388786316,
"learning_rate": 6.819524684817439e-06,
"loss": 0.9636,
"step": 38
},
{
"epoch": 0.3979591836734694,
"grad_norm": 0.9561256170272827,
"learning_rate": 6.667762158254104e-06,
"loss": 1.1511,
"step": 39
},
{
"epoch": 0.40816326530612246,
"grad_norm": 0.9076462388038635,
"learning_rate": 6.514250379489754e-06,
"loss": 1.238,
"step": 40
},
{
"epoch": 0.40816326530612246,
"eval_loss": 0.9210565090179443,
"eval_runtime": 0.26,
"eval_samples_per_second": 30.764,
"eval_steps_per_second": 3.846,
"step": 40
},
{
"epoch": 0.41836734693877553,
"grad_norm": 0.779888391494751,
"learning_rate": 6.3591503611817155e-06,
"loss": 0.927,
"step": 41
},
{
"epoch": 0.42857142857142855,
"grad_norm": 0.826004147529602,
"learning_rate": 6.202624781831269e-06,
"loss": 1.1018,
"step": 42
},
{
"epoch": 0.4387755102040816,
"grad_norm": 0.8641281127929688,
"learning_rate": 6.044837815156377e-06,
"loss": 0.938,
"step": 43
},
{
"epoch": 0.4489795918367347,
"grad_norm": 0.896023154258728,
"learning_rate": 5.885954957896115e-06,
"loss": 1.0764,
"step": 44
},
{
"epoch": 0.45918367346938777,
"grad_norm": 0.9072222709655762,
"learning_rate": 5.726142856227453e-06,
"loss": 1.2437,
"step": 45
},
{
"epoch": 0.46938775510204084,
"grad_norm": 0.7771997451782227,
"learning_rate": 5.5655691309764225e-06,
"loss": 1.0162,
"step": 46
},
{
"epoch": 0.47959183673469385,
"grad_norm": 0.8864902853965759,
"learning_rate": 5.404402201807022e-06,
"loss": 1.1398,
"step": 47
},
{
"epoch": 0.4897959183673469,
"grad_norm": 0.702143669128418,
"learning_rate": 5.242811110572243e-06,
"loss": 0.8257,
"step": 48
},
{
"epoch": 0.5,
"grad_norm": 0.7524011731147766,
"learning_rate": 5.080965344012509e-06,
"loss": 0.9318,
"step": 49
},
{
"epoch": 0.5102040816326531,
"grad_norm": 0.7233298420906067,
"learning_rate": 4.919034655987493e-06,
"loss": 0.8178,
"step": 50
},
{
"epoch": 0.5102040816326531,
"eval_loss": 0.9119246006011963,
"eval_runtime": 0.2595,
"eval_samples_per_second": 30.823,
"eval_steps_per_second": 3.853,
"step": 50
},
{
"epoch": 0.5204081632653061,
"grad_norm": 0.8432551622390747,
"learning_rate": 4.757188889427761e-06,
"loss": 1.0659,
"step": 51
},
{
"epoch": 0.5306122448979592,
"grad_norm": 0.772146999835968,
"learning_rate": 4.59559779819298e-06,
"loss": 1.0016,
"step": 52
},
{
"epoch": 0.5408163265306123,
"grad_norm": 0.8176573514938354,
"learning_rate": 4.434430869023579e-06,
"loss": 1.0541,
"step": 53
},
{
"epoch": 0.5510204081632653,
"grad_norm": 0.730097770690918,
"learning_rate": 4.27385714377255e-06,
"loss": 0.8034,
"step": 54
},
{
"epoch": 0.5612244897959183,
"grad_norm": 0.7190036177635193,
"learning_rate": 4.1140450421038865e-06,
"loss": 0.8833,
"step": 55
},
{
"epoch": 0.5714285714285714,
"grad_norm": 0.732845664024353,
"learning_rate": 3.955162184843625e-06,
"loss": 0.8837,
"step": 56
},
{
"epoch": 0.5816326530612245,
"grad_norm": 0.770910918712616,
"learning_rate": 3.7973752181687336e-06,
"loss": 1.0044,
"step": 57
},
{
"epoch": 0.5918367346938775,
"grad_norm": 0.7635073065757751,
"learning_rate": 3.6408496388182857e-06,
"loss": 0.9338,
"step": 58
},
{
"epoch": 0.6020408163265306,
"grad_norm": 0.8665493130683899,
"learning_rate": 3.4857496205102475e-06,
"loss": 1.1868,
"step": 59
},
{
"epoch": 0.6122448979591837,
"grad_norm": 0.7718795537948608,
"learning_rate": 3.3322378417458985e-06,
"loss": 0.951,
"step": 60
},
{
"epoch": 0.6122448979591837,
"eval_loss": 0.9065544605255127,
"eval_runtime": 0.2595,
"eval_samples_per_second": 30.831,
"eval_steps_per_second": 3.854,
"step": 60
},
{
"epoch": 0.6224489795918368,
"grad_norm": 0.932878851890564,
"learning_rate": 3.180475315182563e-06,
"loss": 1.2871,
"step": 61
},
{
"epoch": 0.6326530612244898,
"grad_norm": 0.8946093320846558,
"learning_rate": 3.0306212187535653e-06,
"loss": 0.9965,
"step": 62
},
{
"epoch": 0.6428571428571429,
"grad_norm": 0.7468050718307495,
"learning_rate": 2.882832728712551e-06,
"loss": 0.8174,
"step": 63
},
{
"epoch": 0.6530612244897959,
"grad_norm": 0.7200383543968201,
"learning_rate": 2.7372648547773063e-06,
"loss": 0.9641,
"step": 64
},
{
"epoch": 0.6632653061224489,
"grad_norm": 0.7331112623214722,
"learning_rate": 2.594070277545975e-06,
"loss": 0.8579,
"step": 65
},
{
"epoch": 0.673469387755102,
"grad_norm": 0.713594377040863,
"learning_rate": 2.4533991883561868e-06,
"loss": 0.872,
"step": 66
},
{
"epoch": 0.6836734693877551,
"grad_norm": 0.8468301296234131,
"learning_rate": 2.315399131755081e-06,
"loss": 1.1029,
"step": 67
},
{
"epoch": 0.6938775510204082,
"grad_norm": 0.7141689658164978,
"learning_rate": 2.1802148507454675e-06,
"loss": 0.8727,
"step": 68
},
{
"epoch": 0.7040816326530612,
"grad_norm": 0.7567382454872131,
"learning_rate": 2.0479881349703885e-06,
"loss": 0.9211,
"step": 69
},
{
"epoch": 0.7142857142857143,
"grad_norm": 0.7020692825317383,
"learning_rate": 1.9188576719953635e-06,
"loss": 0.7869,
"step": 70
},
{
"epoch": 0.7142857142857143,
"eval_loss": 0.9032285809516907,
"eval_runtime": 0.2594,
"eval_samples_per_second": 30.835,
"eval_steps_per_second": 3.854,
"step": 70
},
{
"epoch": 0.7244897959183674,
"grad_norm": 0.853243350982666,
"learning_rate": 1.7929589018443016e-06,
"loss": 1.1762,
"step": 71
},
{
"epoch": 0.7346938775510204,
"grad_norm": 0.8287662267684937,
"learning_rate": 1.6704238749415958e-06,
"loss": 1.0489,
"step": 72
},
{
"epoch": 0.7448979591836735,
"grad_norm": 0.6973975300788879,
"learning_rate": 1.5513811136094786e-06,
"loss": 0.8249,
"step": 73
},
{
"epoch": 0.7551020408163265,
"grad_norm": 0.8150553703308105,
"learning_rate": 1.4359554772658551e-06,
"loss": 1.0011,
"step": 74
},
{
"epoch": 0.7653061224489796,
"grad_norm": 0.7974684238433838,
"learning_rate": 1.3242680314639995e-06,
"loss": 1.0774,
"step": 75
},
{
"epoch": 0.7755102040816326,
"grad_norm": 0.8233014941215515,
"learning_rate": 1.2164359209115235e-06,
"loss": 1.1492,
"step": 76
},
{
"epoch": 0.7857142857142857,
"grad_norm": 0.8305495977401733,
"learning_rate": 1.1125722466017547e-06,
"loss": 1.1329,
"step": 77
},
{
"epoch": 0.7959183673469388,
"grad_norm": 0.7523290514945984,
"learning_rate": 1.012785947186397e-06,
"loss": 0.9375,
"step": 78
},
{
"epoch": 0.8061224489795918,
"grad_norm": 0.8073377013206482,
"learning_rate": 9.171816847139447e-07,
"loss": 1.0936,
"step": 79
},
{
"epoch": 0.8163265306122449,
"grad_norm": 0.8685086965560913,
"learning_rate": 8.258597348536452e-07,
"loss": 1.2626,
"step": 80
},
{
"epoch": 0.8163265306122449,
"eval_loss": 0.9012266397476196,
"eval_runtime": 0.2595,
"eval_samples_per_second": 30.833,
"eval_steps_per_second": 3.854,
"step": 80
},
{
"epoch": 0.826530612244898,
"grad_norm": 0.877649188041687,
"learning_rate": 7.389158817201541e-07,
"loss": 1.3479,
"step": 81
},
{
"epoch": 0.8367346938775511,
"grad_norm": 0.7652468085289001,
"learning_rate": 6.564413174092443e-07,
"loss": 0.9699,
"step": 82
},
{
"epoch": 0.8469387755102041,
"grad_norm": 0.8436093926429749,
"learning_rate": 5.785225463498828e-07,
"loss": 1.1757,
"step": 83
},
{
"epoch": 0.8571428571428571,
"grad_norm": 0.8732744455337524,
"learning_rate": 5.05241294573024e-07,
"loss": 1.1877,
"step": 84
},
{
"epoch": 0.8673469387755102,
"grad_norm": 0.7433239221572876,
"learning_rate": 4.3667442399229985e-07,
"loss": 0.9878,
"step": 85
},
{
"epoch": 0.8775510204081632,
"grad_norm": 0.8225981593132019,
"learning_rate": 3.728938517864794e-07,
"loss": 1.145,
"step": 86
},
{
"epoch": 0.8877551020408163,
"grad_norm": 0.8591734170913696,
"learning_rate": 3.1396647496828245e-07,
"loss": 1.128,
"step": 87
},
{
"epoch": 0.8979591836734694,
"grad_norm": 0.7502849102020264,
"learning_rate": 2.599541002186479e-07,
"loss": 0.9712,
"step": 88
},
{
"epoch": 0.9081632653061225,
"grad_norm": 0.7429087162017822,
"learning_rate": 2.109133790600648e-07,
"loss": 0.9278,
"step": 89
},
{
"epoch": 0.9183673469387755,
"grad_norm": 0.8890379667282104,
"learning_rate": 1.6689574843694433e-07,
"loss": 1.2385,
"step": 90
},
{
"epoch": 0.9183673469387755,
"eval_loss": 0.9008685350418091,
"eval_runtime": 0.2594,
"eval_samples_per_second": 30.837,
"eval_steps_per_second": 3.855,
"step": 90
},
{
"epoch": 0.9285714285714286,
"grad_norm": 0.9237527251243591,
"learning_rate": 1.2794737676536993e-07,
"loss": 1.0866,
"step": 91
},
{
"epoch": 0.9387755102040817,
"grad_norm": 0.8110918402671814,
"learning_rate": 9.410911550880474e-08,
"loss": 1.0466,
"step": 92
},
{
"epoch": 0.9489795918367347,
"grad_norm": 0.7134807109832764,
"learning_rate": 6.54164563305465e-08,
"loss": 0.8466,
"step": 93
},
{
"epoch": 0.9591836734693877,
"grad_norm": 0.8616638779640198,
"learning_rate": 4.189949386787462e-08,
"loss": 1.1521,
"step": 94
},
{
"epoch": 0.9693877551020408,
"grad_norm": 0.9501765966415405,
"learning_rate": 2.358289416693027e-08,
"loss": 1.1505,
"step": 95
},
{
"epoch": 0.9795918367346939,
"grad_norm": 0.7652553915977478,
"learning_rate": 1.0485868811441757e-08,
"loss": 0.9055,
"step": 96
},
{
"epoch": 0.9897959183673469,
"grad_norm": 0.862601637840271,
"learning_rate": 2.6221547724253337e-09,
"loss": 1.0224,
"step": 97
},
{
"epoch": 1.0,
"grad_norm": 0.712780773639679,
"learning_rate": 0.0,
"loss": 0.9534,
"step": 98
},
{
"epoch": 1.0,
"step": 98,
"total_flos": 3.270639151271117e+16,
"train_loss": 1.0707874030483013,
"train_runtime": 199.5779,
"train_samples_per_second": 3.908,
"train_steps_per_second": 0.491
}
],
"logging_steps": 1,
"max_steps": 98,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 98,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 3.270639151271117e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}