marianna13's picture
Upload folder using huggingface_hub
ef3111a verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9980430528375734,
"eval_steps": 500,
"global_step": 1149,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02609262883235486,
"grad_norm": 2.5580028899176965,
"learning_rate": 5e-06,
"loss": 0.9679,
"step": 10
},
{
"epoch": 0.05218525766470972,
"grad_norm": 1.2382684301314397,
"learning_rate": 5e-06,
"loss": 0.8485,
"step": 20
},
{
"epoch": 0.07827788649706457,
"grad_norm": 1.7244888669737717,
"learning_rate": 5e-06,
"loss": 0.8086,
"step": 30
},
{
"epoch": 0.10437051532941943,
"grad_norm": 0.6684245608414933,
"learning_rate": 5e-06,
"loss": 0.7921,
"step": 40
},
{
"epoch": 0.1304631441617743,
"grad_norm": 1.2452393350160769,
"learning_rate": 5e-06,
"loss": 0.7827,
"step": 50
},
{
"epoch": 0.15655577299412915,
"grad_norm": 0.596876900380745,
"learning_rate": 5e-06,
"loss": 0.7727,
"step": 60
},
{
"epoch": 0.182648401826484,
"grad_norm": 0.6733648796408476,
"learning_rate": 5e-06,
"loss": 0.7657,
"step": 70
},
{
"epoch": 0.20874103065883887,
"grad_norm": 0.6054360604030639,
"learning_rate": 5e-06,
"loss": 0.7592,
"step": 80
},
{
"epoch": 0.23483365949119372,
"grad_norm": 0.6078628726535039,
"learning_rate": 5e-06,
"loss": 0.7575,
"step": 90
},
{
"epoch": 0.2609262883235486,
"grad_norm": 0.667808601099263,
"learning_rate": 5e-06,
"loss": 0.7535,
"step": 100
},
{
"epoch": 0.28701891715590344,
"grad_norm": 0.5998036644971939,
"learning_rate": 5e-06,
"loss": 0.7505,
"step": 110
},
{
"epoch": 0.3131115459882583,
"grad_norm": 0.601234098332995,
"learning_rate": 5e-06,
"loss": 0.7462,
"step": 120
},
{
"epoch": 0.33920417482061316,
"grad_norm": 0.5470502921448889,
"learning_rate": 5e-06,
"loss": 0.7464,
"step": 130
},
{
"epoch": 0.365296803652968,
"grad_norm": 0.5354881588471849,
"learning_rate": 5e-06,
"loss": 0.7448,
"step": 140
},
{
"epoch": 0.3913894324853229,
"grad_norm": 0.638179541327912,
"learning_rate": 5e-06,
"loss": 0.7403,
"step": 150
},
{
"epoch": 0.41748206131767773,
"grad_norm": 0.7539597801432362,
"learning_rate": 5e-06,
"loss": 0.7415,
"step": 160
},
{
"epoch": 0.4435746901500326,
"grad_norm": 0.6349380651529639,
"learning_rate": 5e-06,
"loss": 0.7397,
"step": 170
},
{
"epoch": 0.46966731898238745,
"grad_norm": 0.6486113350477142,
"learning_rate": 5e-06,
"loss": 0.7346,
"step": 180
},
{
"epoch": 0.4957599478147423,
"grad_norm": 0.524423192769518,
"learning_rate": 5e-06,
"loss": 0.7336,
"step": 190
},
{
"epoch": 0.5218525766470972,
"grad_norm": 0.8675137881298488,
"learning_rate": 5e-06,
"loss": 0.7357,
"step": 200
},
{
"epoch": 0.547945205479452,
"grad_norm": 0.6181294717672231,
"learning_rate": 5e-06,
"loss": 0.7324,
"step": 210
},
{
"epoch": 0.5740378343118069,
"grad_norm": 0.5310035054090366,
"learning_rate": 5e-06,
"loss": 0.7311,
"step": 220
},
{
"epoch": 0.6001304631441617,
"grad_norm": 0.5616827688931022,
"learning_rate": 5e-06,
"loss": 0.7286,
"step": 230
},
{
"epoch": 0.6262230919765166,
"grad_norm": 0.6811648368494987,
"learning_rate": 5e-06,
"loss": 0.7291,
"step": 240
},
{
"epoch": 0.6523157208088715,
"grad_norm": 0.7733201970691932,
"learning_rate": 5e-06,
"loss": 0.7276,
"step": 250
},
{
"epoch": 0.6784083496412263,
"grad_norm": 0.6192873191830528,
"learning_rate": 5e-06,
"loss": 0.7272,
"step": 260
},
{
"epoch": 0.7045009784735812,
"grad_norm": 0.8068236017156095,
"learning_rate": 5e-06,
"loss": 0.7249,
"step": 270
},
{
"epoch": 0.730593607305936,
"grad_norm": 0.8150623115239886,
"learning_rate": 5e-06,
"loss": 0.728,
"step": 280
},
{
"epoch": 0.7566862361382909,
"grad_norm": 0.724911564188439,
"learning_rate": 5e-06,
"loss": 0.7264,
"step": 290
},
{
"epoch": 0.7827788649706457,
"grad_norm": 0.6359822287899792,
"learning_rate": 5e-06,
"loss": 0.723,
"step": 300
},
{
"epoch": 0.8088714938030006,
"grad_norm": 0.5527636979800452,
"learning_rate": 5e-06,
"loss": 0.7232,
"step": 310
},
{
"epoch": 0.8349641226353555,
"grad_norm": 0.499616870483793,
"learning_rate": 5e-06,
"loss": 0.7222,
"step": 320
},
{
"epoch": 0.8610567514677103,
"grad_norm": 0.5494308637029189,
"learning_rate": 5e-06,
"loss": 0.7174,
"step": 330
},
{
"epoch": 0.8871493803000652,
"grad_norm": 0.5742968590351128,
"learning_rate": 5e-06,
"loss": 0.719,
"step": 340
},
{
"epoch": 0.91324200913242,
"grad_norm": 0.48431807220252887,
"learning_rate": 5e-06,
"loss": 0.7206,
"step": 350
},
{
"epoch": 0.9393346379647749,
"grad_norm": 0.4853602229330187,
"learning_rate": 5e-06,
"loss": 0.7178,
"step": 360
},
{
"epoch": 0.9654272667971298,
"grad_norm": 0.501484126053991,
"learning_rate": 5e-06,
"loss": 0.7204,
"step": 370
},
{
"epoch": 0.9915198956294846,
"grad_norm": 0.5012192063059191,
"learning_rate": 5e-06,
"loss": 0.7179,
"step": 380
},
{
"epoch": 1.0176125244618395,
"grad_norm": 0.9524671132017011,
"learning_rate": 5e-06,
"loss": 0.7451,
"step": 390
},
{
"epoch": 1.0437051532941943,
"grad_norm": 0.524028424392148,
"learning_rate": 5e-06,
"loss": 0.6803,
"step": 400
},
{
"epoch": 1.0697977821265492,
"grad_norm": 0.5407200294759114,
"learning_rate": 5e-06,
"loss": 0.6793,
"step": 410
},
{
"epoch": 1.095890410958904,
"grad_norm": 0.5857287578156233,
"learning_rate": 5e-06,
"loss": 0.6811,
"step": 420
},
{
"epoch": 1.121983039791259,
"grad_norm": 0.5280070429598902,
"learning_rate": 5e-06,
"loss": 0.6807,
"step": 430
},
{
"epoch": 1.1480756686236138,
"grad_norm": 0.4857639927609277,
"learning_rate": 5e-06,
"loss": 0.6776,
"step": 440
},
{
"epoch": 1.1741682974559686,
"grad_norm": 0.5945511698424941,
"learning_rate": 5e-06,
"loss": 0.68,
"step": 450
},
{
"epoch": 1.2002609262883235,
"grad_norm": 0.589811732149492,
"learning_rate": 5e-06,
"loss": 0.6831,
"step": 460
},
{
"epoch": 1.2263535551206783,
"grad_norm": 0.5534380098945336,
"learning_rate": 5e-06,
"loss": 0.6821,
"step": 470
},
{
"epoch": 1.2524461839530332,
"grad_norm": 0.5126525326289294,
"learning_rate": 5e-06,
"loss": 0.6798,
"step": 480
},
{
"epoch": 1.278538812785388,
"grad_norm": 0.6254986498705398,
"learning_rate": 5e-06,
"loss": 0.6797,
"step": 490
},
{
"epoch": 1.304631441617743,
"grad_norm": 0.49035273013831415,
"learning_rate": 5e-06,
"loss": 0.683,
"step": 500
},
{
"epoch": 1.3307240704500978,
"grad_norm": 0.5403945589242636,
"learning_rate": 5e-06,
"loss": 0.6769,
"step": 510
},
{
"epoch": 1.3568166992824526,
"grad_norm": 0.5501681060254049,
"learning_rate": 5e-06,
"loss": 0.6807,
"step": 520
},
{
"epoch": 1.3829093281148075,
"grad_norm": 0.6772693119372506,
"learning_rate": 5e-06,
"loss": 0.6806,
"step": 530
},
{
"epoch": 1.4090019569471623,
"grad_norm": 0.5937727651699936,
"learning_rate": 5e-06,
"loss": 0.6797,
"step": 540
},
{
"epoch": 1.4350945857795172,
"grad_norm": 0.5931432683082334,
"learning_rate": 5e-06,
"loss": 0.6769,
"step": 550
},
{
"epoch": 1.461187214611872,
"grad_norm": 0.6044257552885106,
"learning_rate": 5e-06,
"loss": 0.6821,
"step": 560
},
{
"epoch": 1.487279843444227,
"grad_norm": 0.5487167986355509,
"learning_rate": 5e-06,
"loss": 0.6779,
"step": 570
},
{
"epoch": 1.5133724722765818,
"grad_norm": 0.5046084173932228,
"learning_rate": 5e-06,
"loss": 0.6797,
"step": 580
},
{
"epoch": 1.5394651011089366,
"grad_norm": 0.588333435552722,
"learning_rate": 5e-06,
"loss": 0.681,
"step": 590
},
{
"epoch": 1.5655577299412915,
"grad_norm": 0.5857191319203284,
"learning_rate": 5e-06,
"loss": 0.6821,
"step": 600
},
{
"epoch": 1.5916503587736464,
"grad_norm": 0.5197393978903909,
"learning_rate": 5e-06,
"loss": 0.6815,
"step": 610
},
{
"epoch": 1.6177429876060012,
"grad_norm": 0.5171612878613464,
"learning_rate": 5e-06,
"loss": 0.6786,
"step": 620
},
{
"epoch": 1.643835616438356,
"grad_norm": 0.566201693304294,
"learning_rate": 5e-06,
"loss": 0.6793,
"step": 630
},
{
"epoch": 1.669928245270711,
"grad_norm": 0.5044017255239589,
"learning_rate": 5e-06,
"loss": 0.6776,
"step": 640
},
{
"epoch": 1.6960208741030658,
"grad_norm": 0.478490790469601,
"learning_rate": 5e-06,
"loss": 0.6754,
"step": 650
},
{
"epoch": 1.7221135029354206,
"grad_norm": 0.4963811798560764,
"learning_rate": 5e-06,
"loss": 0.6798,
"step": 660
},
{
"epoch": 1.7482061317677755,
"grad_norm": 0.5278784589567511,
"learning_rate": 5e-06,
"loss": 0.6768,
"step": 670
},
{
"epoch": 1.7742987606001304,
"grad_norm": 0.5387531965847336,
"learning_rate": 5e-06,
"loss": 0.6795,
"step": 680
},
{
"epoch": 1.8003913894324852,
"grad_norm": 0.5875937914583538,
"learning_rate": 5e-06,
"loss": 0.6781,
"step": 690
},
{
"epoch": 1.82648401826484,
"grad_norm": 0.5355905144164954,
"learning_rate": 5e-06,
"loss": 0.6792,
"step": 700
},
{
"epoch": 1.852576647097195,
"grad_norm": 0.6338025629920409,
"learning_rate": 5e-06,
"loss": 0.6821,
"step": 710
},
{
"epoch": 1.8786692759295498,
"grad_norm": 0.5291801310990664,
"learning_rate": 5e-06,
"loss": 0.677,
"step": 720
},
{
"epoch": 1.9047619047619047,
"grad_norm": 0.590990170736626,
"learning_rate": 5e-06,
"loss": 0.6766,
"step": 730
},
{
"epoch": 1.9308545335942595,
"grad_norm": 0.5482515525172131,
"learning_rate": 5e-06,
"loss": 0.6751,
"step": 740
},
{
"epoch": 1.9569471624266144,
"grad_norm": 0.620356822725272,
"learning_rate": 5e-06,
"loss": 0.6765,
"step": 750
},
{
"epoch": 1.9830397912589692,
"grad_norm": 0.5642132027456009,
"learning_rate": 5e-06,
"loss": 0.6776,
"step": 760
},
{
"epoch": 2.009132420091324,
"grad_norm": 0.7543688135393036,
"learning_rate": 5e-06,
"loss": 0.7153,
"step": 770
},
{
"epoch": 2.035225048923679,
"grad_norm": 0.5581679413180287,
"learning_rate": 5e-06,
"loss": 0.6374,
"step": 780
},
{
"epoch": 2.061317677756034,
"grad_norm": 0.6393564816609882,
"learning_rate": 5e-06,
"loss": 0.6364,
"step": 790
},
{
"epoch": 2.0874103065883887,
"grad_norm": 0.614009057667238,
"learning_rate": 5e-06,
"loss": 0.6402,
"step": 800
},
{
"epoch": 2.1135029354207435,
"grad_norm": 0.5346940357958085,
"learning_rate": 5e-06,
"loss": 0.6407,
"step": 810
},
{
"epoch": 2.1395955642530984,
"grad_norm": 0.559106756693009,
"learning_rate": 5e-06,
"loss": 0.638,
"step": 820
},
{
"epoch": 2.1656881930854532,
"grad_norm": 0.603074822031344,
"learning_rate": 5e-06,
"loss": 0.6363,
"step": 830
},
{
"epoch": 2.191780821917808,
"grad_norm": 0.6669696137807871,
"learning_rate": 5e-06,
"loss": 0.6394,
"step": 840
},
{
"epoch": 2.217873450750163,
"grad_norm": 0.6196755868842616,
"learning_rate": 5e-06,
"loss": 0.6407,
"step": 850
},
{
"epoch": 2.243966079582518,
"grad_norm": 0.5462481539172109,
"learning_rate": 5e-06,
"loss": 0.6407,
"step": 860
},
{
"epoch": 2.2700587084148727,
"grad_norm": 0.5378190251307421,
"learning_rate": 5e-06,
"loss": 0.6401,
"step": 870
},
{
"epoch": 2.2961513372472275,
"grad_norm": 0.5143159135941981,
"learning_rate": 5e-06,
"loss": 0.6403,
"step": 880
},
{
"epoch": 2.3222439660795824,
"grad_norm": 0.514592343488391,
"learning_rate": 5e-06,
"loss": 0.6412,
"step": 890
},
{
"epoch": 2.3483365949119372,
"grad_norm": 0.5516158921785709,
"learning_rate": 5e-06,
"loss": 0.6434,
"step": 900
},
{
"epoch": 2.374429223744292,
"grad_norm": 0.669009619898686,
"learning_rate": 5e-06,
"loss": 0.6386,
"step": 910
},
{
"epoch": 2.400521852576647,
"grad_norm": 0.6127707444182544,
"learning_rate": 5e-06,
"loss": 0.6378,
"step": 920
},
{
"epoch": 2.426614481409002,
"grad_norm": 0.6147599175693709,
"learning_rate": 5e-06,
"loss": 0.6411,
"step": 930
},
{
"epoch": 2.4527071102413567,
"grad_norm": 0.5183387934748579,
"learning_rate": 5e-06,
"loss": 0.6431,
"step": 940
},
{
"epoch": 2.4787997390737115,
"grad_norm": 0.5481797536089282,
"learning_rate": 5e-06,
"loss": 0.6415,
"step": 950
},
{
"epoch": 2.5048923679060664,
"grad_norm": 0.49281844671077113,
"learning_rate": 5e-06,
"loss": 0.6416,
"step": 960
},
{
"epoch": 2.5309849967384213,
"grad_norm": 0.5590165679106437,
"learning_rate": 5e-06,
"loss": 0.6413,
"step": 970
},
{
"epoch": 2.557077625570776,
"grad_norm": 0.4928056280476181,
"learning_rate": 5e-06,
"loss": 0.6398,
"step": 980
},
{
"epoch": 2.583170254403131,
"grad_norm": 0.6836197415170101,
"learning_rate": 5e-06,
"loss": 0.6458,
"step": 990
},
{
"epoch": 2.609262883235486,
"grad_norm": 0.5754142213185701,
"learning_rate": 5e-06,
"loss": 0.6391,
"step": 1000
},
{
"epoch": 2.6353555120678407,
"grad_norm": 0.5860168635077211,
"learning_rate": 5e-06,
"loss": 0.6412,
"step": 1010
},
{
"epoch": 2.6614481409001955,
"grad_norm": 0.7689811537018695,
"learning_rate": 5e-06,
"loss": 0.6419,
"step": 1020
},
{
"epoch": 2.6875407697325504,
"grad_norm": 0.7522119854978059,
"learning_rate": 5e-06,
"loss": 0.643,
"step": 1030
},
{
"epoch": 2.7136333985649053,
"grad_norm": 0.5487355635946237,
"learning_rate": 5e-06,
"loss": 0.6395,
"step": 1040
},
{
"epoch": 2.73972602739726,
"grad_norm": 0.6331210725666454,
"learning_rate": 5e-06,
"loss": 0.6409,
"step": 1050
},
{
"epoch": 2.765818656229615,
"grad_norm": 0.5745683262495273,
"learning_rate": 5e-06,
"loss": 0.6447,
"step": 1060
},
{
"epoch": 2.79191128506197,
"grad_norm": 0.5686919776335425,
"learning_rate": 5e-06,
"loss": 0.6396,
"step": 1070
},
{
"epoch": 2.8180039138943247,
"grad_norm": 0.6015098918868974,
"learning_rate": 5e-06,
"loss": 0.6414,
"step": 1080
},
{
"epoch": 2.8440965427266796,
"grad_norm": 0.4915249083089861,
"learning_rate": 5e-06,
"loss": 0.641,
"step": 1090
},
{
"epoch": 2.8701891715590344,
"grad_norm": 0.5923242648103054,
"learning_rate": 5e-06,
"loss": 0.6418,
"step": 1100
},
{
"epoch": 2.8962818003913893,
"grad_norm": 0.5453451599354968,
"learning_rate": 5e-06,
"loss": 0.6435,
"step": 1110
},
{
"epoch": 2.922374429223744,
"grad_norm": 0.6120521256139003,
"learning_rate": 5e-06,
"loss": 0.6448,
"step": 1120
},
{
"epoch": 2.948467058056099,
"grad_norm": 0.5414568129256186,
"learning_rate": 5e-06,
"loss": 0.6423,
"step": 1130
},
{
"epoch": 2.974559686888454,
"grad_norm": 0.5372353896161117,
"learning_rate": 5e-06,
"loss": 0.6412,
"step": 1140
},
{
"epoch": 2.9980430528375734,
"step": 1149,
"total_flos": 1924199035699200.0,
"train_loss": 0.6906216756689538,
"train_runtime": 17455.8647,
"train_samples_per_second": 33.704,
"train_steps_per_second": 0.066
}
],
"logging_steps": 10,
"max_steps": 1149,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1924199035699200.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}