|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3660409556313993, |
|
"acc_stderr": 0.014077223108470137, |
|
"acc_norm": 0.4300341296928328, |
|
"acc_norm_stderr": 0.014467631559137991 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.41894045010953995, |
|
"acc_stderr": 0.0049237725818484955, |
|
"acc_norm": 0.5630352519418442, |
|
"acc_norm_stderr": 0.0049499693630176535 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.52046783625731, |
|
"acc_stderr": 0.038316105328219316, |
|
"acc_norm": 0.52046783625731, |
|
"acc_norm_stderr": 0.038316105328219316 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.5631067961165048, |
|
"acc_stderr": 0.04911147107365777, |
|
"acc_norm": 0.5631067961165048, |
|
"acc_norm_stderr": 0.04911147107365777 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.5542784163473818, |
|
"acc_stderr": 0.017774297282479506, |
|
"acc_norm": 0.5542784163473818, |
|
"acc_norm_stderr": 0.017774297282479506 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.5037037037037037, |
|
"acc_stderr": 0.043192236258113324, |
|
"acc_norm": 0.5037037037037037, |
|
"acc_norm_stderr": 0.043192236258113324 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.4085106382978723, |
|
"acc_stderr": 0.03213418026701576, |
|
"acc_norm": 0.4085106382978723, |
|
"acc_norm_stderr": 0.03213418026701576 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.4036144578313253, |
|
"acc_stderr": 0.03819486140758397, |
|
"acc_norm": 0.4036144578313253, |
|
"acc_norm_stderr": 0.03819486140758397 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.5176848874598071, |
|
"acc_stderr": 0.028380322849077138, |
|
"acc_norm": 0.5176848874598071, |
|
"acc_norm_stderr": 0.028380322849077138 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.47533632286995514, |
|
"acc_stderr": 0.03351695167652628, |
|
"acc_norm": 0.47533632286995514, |
|
"acc_norm_stderr": 0.03351695167652628 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.5419847328244275, |
|
"acc_stderr": 0.04369802690578757, |
|
"acc_norm": 0.5419847328244275, |
|
"acc_norm_stderr": 0.04369802690578757 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.46, |
|
"acc_stderr": 0.05009082659620332, |
|
"acc_norm": 0.46, |
|
"acc_norm_stderr": 0.05009082659620332 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.6161616161616161, |
|
"acc_stderr": 0.03464881675016337, |
|
"acc_norm": 0.6161616161616161, |
|
"acc_norm_stderr": 0.03464881675016337 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.5448275862068965, |
|
"acc_stderr": 0.04149886942192117, |
|
"acc_norm": 0.5448275862068965, |
|
"acc_norm_stderr": 0.04149886942192117 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.2647058823529412, |
|
"acc_stderr": 0.04389869956808778, |
|
"acc_norm": 0.2647058823529412, |
|
"acc_norm_stderr": 0.04389869956808778 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.032478490123081544, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.032478490123081544 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.4794871794871795, |
|
"acc_stderr": 0.025329663163489943, |
|
"acc_norm": 0.4794871794871795, |
|
"acc_norm_stderr": 0.025329663163489943 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.62, |
|
"acc_stderr": 0.048783173121456316, |
|
"acc_norm": 0.62, |
|
"acc_norm_stderr": 0.048783173121456316 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.36, |
|
"acc_stderr": 0.04824181513244218, |
|
"acc_norm": 0.36, |
|
"acc_norm_stderr": 0.04824181513244218 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.4166666666666667, |
|
"acc_stderr": 0.04766075165356461, |
|
"acc_norm": 0.4166666666666667, |
|
"acc_norm_stderr": 0.04766075165356461 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.4039408866995074, |
|
"acc_stderr": 0.03452453903822039, |
|
"acc_norm": 0.4039408866995074, |
|
"acc_norm_stderr": 0.03452453903822039 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.4806451612903226, |
|
"acc_stderr": 0.0284226874043121, |
|
"acc_norm": 0.4806451612903226, |
|
"acc_norm_stderr": 0.0284226874043121 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.7008547008547008, |
|
"acc_stderr": 0.029996951858349472, |
|
"acc_norm": 0.7008547008547008, |
|
"acc_norm_stderr": 0.029996951858349472 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.47924528301886793, |
|
"acc_stderr": 0.030746349975723456, |
|
"acc_norm": 0.47924528301886793, |
|
"acc_norm_stderr": 0.030746349975723456 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.04789131426105757, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.04789131426105757 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.027940457136228405, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.027940457136228405 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.31788079470198677, |
|
"acc_stderr": 0.03802039760107903, |
|
"acc_norm": 0.31788079470198677, |
|
"acc_norm_stderr": 0.03802039760107903 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.6268656716417911, |
|
"acc_stderr": 0.03419832608176007, |
|
"acc_norm": 0.6268656716417911, |
|
"acc_norm_stderr": 0.03419832608176007 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.37572254335260113, |
|
"acc_stderr": 0.036928207672648664, |
|
"acc_norm": 0.37572254335260113, |
|
"acc_norm_stderr": 0.036928207672648664 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.3492063492063492, |
|
"acc_stderr": 0.02455229220934265, |
|
"acc_norm": 0.3492063492063492, |
|
"acc_norm_stderr": 0.02455229220934265 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.4722222222222222, |
|
"acc_stderr": 0.04174752578923185, |
|
"acc_norm": 0.4722222222222222, |
|
"acc_norm_stderr": 0.04174752578923185 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.64, |
|
"acc_stderr": 0.048241815132442176, |
|
"acc_norm": 0.64, |
|
"acc_norm_stderr": 0.048241815132442176 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.47109826589595377, |
|
"acc_stderr": 0.026874085883518348, |
|
"acc_norm": 0.47109826589595377, |
|
"acc_norm_stderr": 0.026874085883518348 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.4601226993865031, |
|
"acc_stderr": 0.0391585729143697, |
|
"acc_norm": 0.4601226993865031, |
|
"acc_norm_stderr": 0.0391585729143697 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.49691358024691357, |
|
"acc_stderr": 0.027820214158594384, |
|
"acc_norm": 0.49691358024691357, |
|
"acc_norm_stderr": 0.027820214158594384 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.26, |
|
"acc_stderr": 0.04408440022768078, |
|
"acc_norm": 0.26, |
|
"acc_norm_stderr": 0.04408440022768078 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.5284974093264249, |
|
"acc_stderr": 0.036025735712884414, |
|
"acc_norm": 0.5284974093264249, |
|
"acc_norm_stderr": 0.036025735712884414 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.04434600701584925, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.04434600701584925 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.6422018348623854, |
|
"acc_stderr": 0.020552060784827818, |
|
"acc_norm": 0.6422018348623854, |
|
"acc_norm_stderr": 0.020552060784827818 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.24603174603174602, |
|
"acc_stderr": 0.03852273364924315, |
|
"acc_norm": 0.24603174603174602, |
|
"acc_norm_stderr": 0.03852273364924315 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.5620915032679739, |
|
"acc_stderr": 0.02840830202033269, |
|
"acc_norm": 0.5620915032679739, |
|
"acc_norm_stderr": 0.02840830202033269 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.51, |
|
"acc_stderr": 0.05024183937956911, |
|
"acc_norm": 0.51, |
|
"acc_norm_stderr": 0.05024183937956911 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6446280991735537, |
|
"acc_stderr": 0.04369236326573981, |
|
"acc_norm": 0.6446280991735537, |
|
"acc_norm_stderr": 0.04369236326573981 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.5131578947368421, |
|
"acc_stderr": 0.04067533136309173, |
|
"acc_norm": 0.5131578947368421, |
|
"acc_norm_stderr": 0.04067533136309173 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.4395424836601307, |
|
"acc_stderr": 0.020079420408087918, |
|
"acc_norm": 0.4395424836601307, |
|
"acc_norm_stderr": 0.020079420408087918 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.3404255319148936, |
|
"acc_stderr": 0.02826765748265014, |
|
"acc_norm": 0.3404255319148936, |
|
"acc_norm_stderr": 0.02826765748265014 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.2767857142857143, |
|
"acc_stderr": 0.04246624336697624, |
|
"acc_norm": 0.2767857142857143, |
|
"acc_norm_stderr": 0.04246624336697624 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.33796296296296297, |
|
"acc_stderr": 0.032259413526312945, |
|
"acc_norm": 0.33796296296296297, |
|
"acc_norm_stderr": 0.032259413526312945 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.2346368715083799, |
|
"acc_stderr": 0.014173044098303656, |
|
"acc_norm": 0.2346368715083799, |
|
"acc_norm_stderr": 0.014173044098303656 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.046882617226215034, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.046882617226215034 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.57, |
|
"acc_stderr": 0.04975698519562428, |
|
"acc_norm": 0.57, |
|
"acc_norm_stderr": 0.04975698519562428 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.46691176470588236, |
|
"acc_stderr": 0.030306257722468317, |
|
"acc_norm": 0.46691176470588236, |
|
"acc_norm_stderr": 0.030306257722468317 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.47346938775510206, |
|
"acc_stderr": 0.03196412734523272, |
|
"acc_norm": 0.47346938775510206, |
|
"acc_norm_stderr": 0.03196412734523272 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.6075949367088608, |
|
"acc_stderr": 0.0317847187456473, |
|
"acc_norm": 0.6075949367088608, |
|
"acc_norm_stderr": 0.0317847187456473 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.32333767926988266, |
|
"acc_stderr": 0.011946565758447204, |
|
"acc_norm": 0.32333767926988266, |
|
"acc_norm_stderr": 0.011946565758447204 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.49019607843137253, |
|
"acc_stderr": 0.03508637358630573, |
|
"acc_norm": 0.49019607843137253, |
|
"acc_norm_stderr": 0.03508637358630573 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.6181818181818182, |
|
"acc_stderr": 0.03793713171165636, |
|
"acc_norm": 0.6181818181818182, |
|
"acc_norm_stderr": 0.03793713171165636 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.26438188494492043, |
|
"mc1_stderr": 0.015438211119522512, |
|
"mc2": 0.40241254956351097, |
|
"mc2_stderr": 0.015340553744152264 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.5112160566706021, |
|
"acc_stderr": 0.01718602846948929, |
|
"acc_norm": 0.5419126328217237, |
|
"acc_norm_stderr": 0.017129852117911144 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "BM-K/yi-ko-6b-it-v1.0.0", |
|
"model_sha": "1401792e5c974a79e0f6ccb7f060003d0d54e2e5", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |