|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.33532423208191126, |
|
"acc_stderr": 0.01379618294778556, |
|
"acc_norm": 0.38139931740614336, |
|
"acc_norm_stderr": 0.01419438908668526 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.3703445528779128, |
|
"acc_stderr": 0.004819100456867818, |
|
"acc_norm": 0.481876120294762, |
|
"acc_norm_stderr": 0.004986502296931182 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.4619883040935672, |
|
"acc_stderr": 0.03823727092882307, |
|
"acc_norm": 0.4619883040935672, |
|
"acc_norm_stderr": 0.03823727092882307 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.5631067961165048, |
|
"acc_stderr": 0.049111471073657764, |
|
"acc_norm": 0.5631067961165048, |
|
"acc_norm_stderr": 0.049111471073657764 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.46871008939974457, |
|
"acc_stderr": 0.017844918090468544, |
|
"acc_norm": 0.46871008939974457, |
|
"acc_norm_stderr": 0.017844918090468544 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.4148148148148148, |
|
"acc_stderr": 0.042561937679014075, |
|
"acc_norm": 0.4148148148148148, |
|
"acc_norm_stderr": 0.042561937679014075 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.40425531914893614, |
|
"acc_stderr": 0.03208115750788684, |
|
"acc_norm": 0.40425531914893614, |
|
"acc_norm_stderr": 0.03208115750788684 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.43373493975903615, |
|
"acc_stderr": 0.03858158940685515, |
|
"acc_norm": 0.43373493975903615, |
|
"acc_norm_stderr": 0.03858158940685515 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.5112540192926045, |
|
"acc_stderr": 0.028390897396863533, |
|
"acc_norm": 0.5112540192926045, |
|
"acc_norm_stderr": 0.028390897396863533 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.4439461883408072, |
|
"acc_stderr": 0.03334625674242728, |
|
"acc_norm": 0.4439461883408072, |
|
"acc_norm_stderr": 0.03334625674242728 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.5038167938931297, |
|
"acc_stderr": 0.043851623256015534, |
|
"acc_norm": 0.5038167938931297, |
|
"acc_norm_stderr": 0.043851623256015534 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.37, |
|
"acc_stderr": 0.04852365870939099, |
|
"acc_norm": 0.37, |
|
"acc_norm_stderr": 0.04852365870939099 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.5505050505050505, |
|
"acc_stderr": 0.035441324919479704, |
|
"acc_norm": 0.5505050505050505, |
|
"acc_norm_stderr": 0.035441324919479704 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.4482758620689655, |
|
"acc_stderr": 0.041443118108781506, |
|
"acc_norm": 0.4482758620689655, |
|
"acc_norm_stderr": 0.041443118108781506 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.24509803921568626, |
|
"acc_stderr": 0.04280105837364395, |
|
"acc_norm": 0.24509803921568626, |
|
"acc_norm_stderr": 0.04280105837364395 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.542016806722689, |
|
"acc_stderr": 0.03236361111951941, |
|
"acc_norm": 0.542016806722689, |
|
"acc_norm_stderr": 0.03236361111951941 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.4794871794871795, |
|
"acc_stderr": 0.025329663163489943, |
|
"acc_norm": 0.4794871794871795, |
|
"acc_norm_stderr": 0.025329663163489943 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.57, |
|
"acc_stderr": 0.04975698519562429, |
|
"acc_norm": 0.57, |
|
"acc_norm_stderr": 0.04975698519562429 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.34, |
|
"acc_stderr": 0.047609522856952344, |
|
"acc_norm": 0.34, |
|
"acc_norm_stderr": 0.047609522856952344 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.5277777777777778, |
|
"acc_stderr": 0.04826217294139894, |
|
"acc_norm": 0.5277777777777778, |
|
"acc_norm_stderr": 0.04826217294139894 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.43842364532019706, |
|
"acc_stderr": 0.03491207857486519, |
|
"acc_norm": 0.43842364532019706, |
|
"acc_norm_stderr": 0.03491207857486519 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.4935483870967742, |
|
"acc_stderr": 0.02844163823354051, |
|
"acc_norm": 0.4935483870967742, |
|
"acc_norm_stderr": 0.02844163823354051 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.7435897435897436, |
|
"acc_stderr": 0.028605953702004243, |
|
"acc_norm": 0.7435897435897436, |
|
"acc_norm_stderr": 0.028605953702004243 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.4339622641509434, |
|
"acc_stderr": 0.030503292013342592, |
|
"acc_norm": 0.4339622641509434, |
|
"acc_norm_stderr": 0.030503292013342592 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.509090909090909, |
|
"acc_stderr": 0.0478833976870286, |
|
"acc_norm": 0.509090909090909, |
|
"acc_norm_stderr": 0.0478833976870286 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.337037037037037, |
|
"acc_stderr": 0.028820884666253252, |
|
"acc_norm": 0.337037037037037, |
|
"acc_norm_stderr": 0.028820884666253252 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.2582781456953642, |
|
"acc_stderr": 0.035737053147634576, |
|
"acc_norm": 0.2582781456953642, |
|
"acc_norm_stderr": 0.035737053147634576 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.5970149253731343, |
|
"acc_stderr": 0.034683432951111266, |
|
"acc_norm": 0.5970149253731343, |
|
"acc_norm_stderr": 0.034683432951111266 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.3815028901734104, |
|
"acc_stderr": 0.03703851193099521, |
|
"acc_norm": 0.3815028901734104, |
|
"acc_norm_stderr": 0.03703851193099521 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.373015873015873, |
|
"acc_stderr": 0.02490699045899257, |
|
"acc_norm": 0.373015873015873, |
|
"acc_norm_stderr": 0.02490699045899257 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.3472222222222222, |
|
"acc_stderr": 0.039812405437178615, |
|
"acc_norm": 0.3472222222222222, |
|
"acc_norm_stderr": 0.039812405437178615 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.41, |
|
"acc_stderr": 0.049431107042371025, |
|
"acc_norm": 0.41, |
|
"acc_norm_stderr": 0.049431107042371025 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.58, |
|
"acc_stderr": 0.04960449637488584, |
|
"acc_norm": 0.58, |
|
"acc_norm_stderr": 0.04960449637488584 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.523121387283237, |
|
"acc_stderr": 0.026890297881303118, |
|
"acc_norm": 0.523121387283237, |
|
"acc_norm_stderr": 0.026890297881303118 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.5153374233128835, |
|
"acc_stderr": 0.039265223787088424, |
|
"acc_norm": 0.5153374233128835, |
|
"acc_norm_stderr": 0.039265223787088424 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.4382716049382716, |
|
"acc_stderr": 0.027607914087400473, |
|
"acc_norm": 0.4382716049382716, |
|
"acc_norm_stderr": 0.027607914087400473 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.0479372485441102, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.0479372485441102 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.5129533678756477, |
|
"acc_stderr": 0.0360722806104775, |
|
"acc_norm": 0.5129533678756477, |
|
"acc_norm_stderr": 0.0360722806104775 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.04434600701584925, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.04434600701584925 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.4917431192660551, |
|
"acc_stderr": 0.021434399918214338, |
|
"acc_norm": 0.4917431192660551, |
|
"acc_norm_stderr": 0.021434399918214338 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.35714285714285715, |
|
"acc_stderr": 0.04285714285714281, |
|
"acc_norm": 0.35714285714285715, |
|
"acc_norm_stderr": 0.04285714285714281 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.45751633986928103, |
|
"acc_stderr": 0.028526383452142628, |
|
"acc_norm": 0.45751633986928103, |
|
"acc_norm_stderr": 0.028526383452142628 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.48, |
|
"acc_stderr": 0.050211673156867795, |
|
"acc_norm": 0.48, |
|
"acc_norm_stderr": 0.050211673156867795 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6942148760330579, |
|
"acc_stderr": 0.04205953933884124, |
|
"acc_norm": 0.6942148760330579, |
|
"acc_norm_stderr": 0.04205953933884124 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.4407894736842105, |
|
"acc_stderr": 0.04040311062490435, |
|
"acc_norm": 0.4407894736842105, |
|
"acc_norm_stderr": 0.04040311062490435 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.39869281045751637, |
|
"acc_stderr": 0.01980828131744984, |
|
"acc_norm": 0.39869281045751637, |
|
"acc_norm_stderr": 0.01980828131744984 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.35815602836879434, |
|
"acc_stderr": 0.028602085862759412, |
|
"acc_norm": 0.35815602836879434, |
|
"acc_norm_stderr": 0.028602085862759412 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.4017857142857143, |
|
"acc_stderr": 0.04653333146973646, |
|
"acc_norm": 0.4017857142857143, |
|
"acc_norm_stderr": 0.04653333146973646 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.4351851851851852, |
|
"acc_stderr": 0.03381200005643525, |
|
"acc_norm": 0.4351851851851852, |
|
"acc_norm_stderr": 0.03381200005643525 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.34413407821229053, |
|
"acc_stderr": 0.015889221313307094, |
|
"acc_norm": 0.34413407821229053, |
|
"acc_norm_stderr": 0.015889221313307094 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.44, |
|
"acc_stderr": 0.04988876515698589, |
|
"acc_norm": 0.44, |
|
"acc_norm_stderr": 0.04988876515698589 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.58, |
|
"acc_stderr": 0.049604496374885836, |
|
"acc_norm": 0.58, |
|
"acc_norm_stderr": 0.049604496374885836 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.41544117647058826, |
|
"acc_stderr": 0.029935342707877743, |
|
"acc_norm": 0.41544117647058826, |
|
"acc_norm_stderr": 0.029935342707877743 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.563265306122449, |
|
"acc_stderr": 0.03175195237583323, |
|
"acc_norm": 0.563265306122449, |
|
"acc_norm_stderr": 0.03175195237583323 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.5738396624472574, |
|
"acc_stderr": 0.03219035703131774, |
|
"acc_norm": 0.5738396624472574, |
|
"acc_norm_stderr": 0.03219035703131774 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.32790091264667537, |
|
"acc_stderr": 0.011989936640666535, |
|
"acc_norm": 0.32790091264667537, |
|
"acc_norm_stderr": 0.011989936640666535 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.44607843137254904, |
|
"acc_stderr": 0.03488845451304974, |
|
"acc_norm": 0.44607843137254904, |
|
"acc_norm_stderr": 0.03488845451304974 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.45454545454545453, |
|
"acc_stderr": 0.03888176921674099, |
|
"acc_norm": 0.45454545454545453, |
|
"acc_norm_stderr": 0.03888176921674099 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2864137086903305, |
|
"mc1_stderr": 0.015826142439502342, |
|
"mc2": 0.4613168911756529, |
|
"mc2_stderr": 0.015417066073991514 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.5076741440377804, |
|
"acc_stderr": 0.017188329219654273, |
|
"acc_norm": 0.5678866587957497, |
|
"acc_norm_stderr": 0.017031170198851742 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "mistralai/Mistral-7B-v0.1", |
|
"model_sha": "5e9c98b96d071dce59368012254c55b0ec6f8658", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |