|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3626279863481229, |
|
"acc_stderr": 0.014049106564955, |
|
"acc_norm": 0.40187713310580203, |
|
"acc_norm_stderr": 0.014327268614578273 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.41276638119896436, |
|
"acc_stderr": 0.004913253031155685, |
|
"acc_norm": 0.5246962756423024, |
|
"acc_norm_stderr": 0.004983691099110914 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.5380116959064327, |
|
"acc_stderr": 0.03823727092882307, |
|
"acc_norm": 0.5380116959064327, |
|
"acc_norm_stderr": 0.03823727092882307 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.5339805825242718, |
|
"acc_stderr": 0.04939291447273481, |
|
"acc_norm": 0.5339805825242718, |
|
"acc_norm_stderr": 0.04939291447273481 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.4482758620689655, |
|
"acc_stderr": 0.01778403453499246, |
|
"acc_norm": 0.4482758620689655, |
|
"acc_norm_stderr": 0.01778403453499246 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.35555555555555557, |
|
"acc_stderr": 0.04135176749720386, |
|
"acc_norm": 0.35555555555555557, |
|
"acc_norm_stderr": 0.04135176749720386 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.4085106382978723, |
|
"acc_stderr": 0.03213418026701576, |
|
"acc_norm": 0.4085106382978723, |
|
"acc_norm_stderr": 0.03213418026701576 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.3674698795180723, |
|
"acc_stderr": 0.03753267402120574, |
|
"acc_norm": 0.3674698795180723, |
|
"acc_norm_stderr": 0.03753267402120574 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.41479099678456594, |
|
"acc_stderr": 0.02798268045975956, |
|
"acc_norm": 0.41479099678456594, |
|
"acc_norm_stderr": 0.02798268045975956 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.3721973094170404, |
|
"acc_stderr": 0.03244305283008731, |
|
"acc_norm": 0.3721973094170404, |
|
"acc_norm_stderr": 0.03244305283008731 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.46564885496183206, |
|
"acc_stderr": 0.043749285605997376, |
|
"acc_norm": 0.46564885496183206, |
|
"acc_norm_stderr": 0.043749285605997376 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.0446196043338474, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.0446196043338474 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.5505050505050505, |
|
"acc_stderr": 0.0354413249194797, |
|
"acc_norm": 0.5505050505050505, |
|
"acc_norm_stderr": 0.0354413249194797 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.38620689655172413, |
|
"acc_stderr": 0.04057324734419035, |
|
"acc_norm": 0.38620689655172413, |
|
"acc_norm_stderr": 0.04057324734419035 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.22549019607843138, |
|
"acc_stderr": 0.04158307533083286, |
|
"acc_norm": 0.22549019607843138, |
|
"acc_norm_stderr": 0.04158307533083286 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.4495798319327731, |
|
"acc_stderr": 0.03231293497137707, |
|
"acc_norm": 0.4495798319327731, |
|
"acc_norm_stderr": 0.03231293497137707 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.4153846153846154, |
|
"acc_stderr": 0.024985354923102315, |
|
"acc_norm": 0.4153846153846154, |
|
"acc_norm_stderr": 0.024985354923102315 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.26, |
|
"acc_stderr": 0.04408440022768078, |
|
"acc_norm": 0.26, |
|
"acc_norm_stderr": 0.04408440022768078 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.4166666666666667, |
|
"acc_stderr": 0.04766075165356461, |
|
"acc_norm": 0.4166666666666667, |
|
"acc_norm_stderr": 0.04766075165356461 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.33497536945812806, |
|
"acc_stderr": 0.033208527423483104, |
|
"acc_norm": 0.33497536945812806, |
|
"acc_norm_stderr": 0.033208527423483104 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.4129032258064516, |
|
"acc_stderr": 0.028009138125400387, |
|
"acc_norm": 0.4129032258064516, |
|
"acc_norm_stderr": 0.028009138125400387 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.6538461538461539, |
|
"acc_stderr": 0.031166957367235897, |
|
"acc_norm": 0.6538461538461539, |
|
"acc_norm_stderr": 0.031166957367235897 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.44528301886792454, |
|
"acc_stderr": 0.030588052974270658, |
|
"acc_norm": 0.44528301886792454, |
|
"acc_norm_stderr": 0.030588052974270658 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.4727272727272727, |
|
"acc_stderr": 0.04782001791380063, |
|
"acc_norm": 0.4727272727272727, |
|
"acc_norm_stderr": 0.04782001791380063 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.27037037037037037, |
|
"acc_stderr": 0.027080372815145654, |
|
"acc_norm": 0.27037037037037037, |
|
"acc_norm_stderr": 0.027080372815145654 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.31125827814569534, |
|
"acc_stderr": 0.03780445850526733, |
|
"acc_norm": 0.31125827814569534, |
|
"acc_norm_stderr": 0.03780445850526733 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.5522388059701493, |
|
"acc_stderr": 0.03516184772952166, |
|
"acc_norm": 0.5522388059701493, |
|
"acc_norm_stderr": 0.03516184772952166 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.37572254335260113, |
|
"acc_stderr": 0.036928207672648664, |
|
"acc_norm": 0.37572254335260113, |
|
"acc_norm_stderr": 0.036928207672648664 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.29894179894179895, |
|
"acc_stderr": 0.02357760479165582, |
|
"acc_norm": 0.29894179894179895, |
|
"acc_norm_stderr": 0.02357760479165582 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.3402777777777778, |
|
"acc_stderr": 0.039621355734862175, |
|
"acc_norm": 0.3402777777777778, |
|
"acc_norm_stderr": 0.039621355734862175 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.4, |
|
"acc_stderr": 0.04923659639173309, |
|
"acc_norm": 0.4, |
|
"acc_norm_stderr": 0.04923659639173309 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.55, |
|
"acc_stderr": 0.05, |
|
"acc_norm": 0.55, |
|
"acc_norm_stderr": 0.05 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.4161849710982659, |
|
"acc_stderr": 0.026538189104705484, |
|
"acc_norm": 0.4161849710982659, |
|
"acc_norm_stderr": 0.026538189104705484 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.44785276073619634, |
|
"acc_stderr": 0.03906947479456601, |
|
"acc_norm": 0.44785276073619634, |
|
"acc_norm_stderr": 0.03906947479456601 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.4074074074074074, |
|
"acc_stderr": 0.027339546640662724, |
|
"acc_norm": 0.4074074074074074, |
|
"acc_norm_stderr": 0.027339546640662724 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.0479372485441102, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.0479372485441102 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.44041450777202074, |
|
"acc_stderr": 0.035827245300360945, |
|
"acc_norm": 0.44041450777202074, |
|
"acc_norm_stderr": 0.035827245300360945 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.21052631578947367, |
|
"acc_stderr": 0.038351539543994194, |
|
"acc_norm": 0.21052631578947367, |
|
"acc_norm_stderr": 0.038351539543994194 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.4954128440366973, |
|
"acc_stderr": 0.021436420955529424, |
|
"acc_norm": 0.4954128440366973, |
|
"acc_norm_stderr": 0.021436420955529424 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.3492063492063492, |
|
"acc_stderr": 0.04263906892795133, |
|
"acc_norm": 0.3492063492063492, |
|
"acc_norm_stderr": 0.04263906892795133 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.47058823529411764, |
|
"acc_stderr": 0.028580341065138286, |
|
"acc_norm": 0.47058823529411764, |
|
"acc_norm_stderr": 0.028580341065138286 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.4, |
|
"acc_stderr": 0.04923659639173309, |
|
"acc_norm": 0.4, |
|
"acc_norm_stderr": 0.04923659639173309 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6528925619834711, |
|
"acc_stderr": 0.043457245702925335, |
|
"acc_norm": 0.6528925619834711, |
|
"acc_norm_stderr": 0.043457245702925335 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.4407894736842105, |
|
"acc_stderr": 0.04040311062490436, |
|
"acc_norm": 0.4407894736842105, |
|
"acc_norm_stderr": 0.04040311062490436 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.29901960784313725, |
|
"acc_stderr": 0.018521756215423027, |
|
"acc_norm": 0.29901960784313725, |
|
"acc_norm_stderr": 0.018521756215423027 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.31560283687943264, |
|
"acc_stderr": 0.027724989449509314, |
|
"acc_norm": 0.31560283687943264, |
|
"acc_norm_stderr": 0.027724989449509314 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.15178571428571427, |
|
"acc_stderr": 0.034057028381856924, |
|
"acc_norm": 0.15178571428571427, |
|
"acc_norm_stderr": 0.034057028381856924 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.39814814814814814, |
|
"acc_stderr": 0.033384734032074016, |
|
"acc_norm": 0.39814814814814814, |
|
"acc_norm_stderr": 0.033384734032074016 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.24692737430167597, |
|
"acc_stderr": 0.014422292204808852, |
|
"acc_norm": 0.24692737430167597, |
|
"acc_norm_stderr": 0.014422292204808852 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.047937248544110196, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.047937248544110196 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.36, |
|
"acc_stderr": 0.04824181513244218, |
|
"acc_norm": 0.36, |
|
"acc_norm_stderr": 0.04824181513244218 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.3235294117647059, |
|
"acc_stderr": 0.02841820861940679, |
|
"acc_norm": 0.3235294117647059, |
|
"acc_norm_stderr": 0.02841820861940679 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.46122448979591835, |
|
"acc_stderr": 0.031912820526692774, |
|
"acc_norm": 0.46122448979591835, |
|
"acc_norm_stderr": 0.031912820526692774 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.5232067510548524, |
|
"acc_stderr": 0.032512152011410174, |
|
"acc_norm": 0.5232067510548524, |
|
"acc_norm_stderr": 0.032512152011410174 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.3070404172099087, |
|
"acc_stderr": 0.011780959114513778, |
|
"acc_norm": 0.3070404172099087, |
|
"acc_norm_stderr": 0.011780959114513778 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.4068627450980392, |
|
"acc_stderr": 0.03447891136353382, |
|
"acc_norm": 0.4068627450980392, |
|
"acc_norm_stderr": 0.03447891136353382 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.46060606060606063, |
|
"acc_stderr": 0.03892207016552013, |
|
"acc_norm": 0.46060606060606063, |
|
"acc_norm_stderr": 0.03892207016552013 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.29498164014687883, |
|
"mc1_stderr": 0.015964400965589674, |
|
"mc2": 0.47296551445370655, |
|
"mc2_stderr": 0.016489115600580966 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.5631641086186541, |
|
"acc_stderr": 0.01705263355985607, |
|
"acc_norm": 0.5855962219598583, |
|
"acc_norm_stderr": 0.016936583383943615 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "jiwoochris/ko-llama2-v2", |
|
"model_sha": "bfe6a2095cc43e82103cbdff36721810ef4057e3", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |