|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.2619453924914676, |
|
"acc_stderr": 0.012849054826858114, |
|
"acc_norm": 0.3302047781569966, |
|
"acc_norm_stderr": 0.013743085603760427 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.3604859589723163, |
|
"acc_stderr": 0.004791601975612766, |
|
"acc_norm": 0.45429197371041624, |
|
"acc_norm_stderr": 0.004968888130290065 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.1871345029239766, |
|
"acc_stderr": 0.029913127232368025, |
|
"acc_norm": 0.1871345029239766, |
|
"acc_norm_stderr": 0.029913127232368025 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.1941747572815534, |
|
"acc_stderr": 0.03916667762822584, |
|
"acc_norm": 0.1941747572815534, |
|
"acc_norm_stderr": 0.03916667762822584 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.21966794380587484, |
|
"acc_stderr": 0.014805384478371163, |
|
"acc_norm": 0.21966794380587484, |
|
"acc_norm_stderr": 0.014805384478371163 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.2740740740740741, |
|
"acc_stderr": 0.03853254836552003, |
|
"acc_norm": 0.2740740740740741, |
|
"acc_norm_stderr": 0.03853254836552003 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.046882617226215034, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.046882617226215034 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.2127659574468085, |
|
"acc_stderr": 0.026754391348039783, |
|
"acc_norm": 0.2127659574468085, |
|
"acc_norm_stderr": 0.026754391348039783 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.21686746987951808, |
|
"acc_stderr": 0.03208284450356365, |
|
"acc_norm": 0.21686746987951808, |
|
"acc_norm_stderr": 0.03208284450356365 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.28938906752411575, |
|
"acc_stderr": 0.02575586592263294, |
|
"acc_norm": 0.28938906752411575, |
|
"acc_norm_stderr": 0.02575586592263294 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.11659192825112108, |
|
"acc_stderr": 0.02153963981624447, |
|
"acc_norm": 0.11659192825112108, |
|
"acc_norm_stderr": 0.02153963981624447 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.26717557251908397, |
|
"acc_stderr": 0.03880848301082396, |
|
"acc_norm": 0.26717557251908397, |
|
"acc_norm_stderr": 0.03880848301082396 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.29797979797979796, |
|
"acc_stderr": 0.032586303838365555, |
|
"acc_norm": 0.29797979797979796, |
|
"acc_norm_stderr": 0.032586303838365555 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.2413793103448276, |
|
"acc_stderr": 0.03565998174135302, |
|
"acc_norm": 0.2413793103448276, |
|
"acc_norm_stderr": 0.03565998174135302 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.23529411764705882, |
|
"acc_stderr": 0.04220773659171452, |
|
"acc_norm": 0.23529411764705882, |
|
"acc_norm_stderr": 0.04220773659171452 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.31512605042016806, |
|
"acc_stderr": 0.030176808288974337, |
|
"acc_norm": 0.31512605042016806, |
|
"acc_norm_stderr": 0.030176808288974337 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.32564102564102565, |
|
"acc_stderr": 0.02375966576741229, |
|
"acc_norm": 0.32564102564102565, |
|
"acc_norm_stderr": 0.02375966576741229 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.2, |
|
"acc_stderr": 0.04020151261036846, |
|
"acc_norm": 0.2, |
|
"acc_norm_stderr": 0.04020151261036846 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.21296296296296297, |
|
"acc_stderr": 0.03957835471980981, |
|
"acc_norm": 0.21296296296296297, |
|
"acc_norm_stderr": 0.03957835471980981 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.30049261083743845, |
|
"acc_stderr": 0.03225799476233485, |
|
"acc_norm": 0.30049261083743845, |
|
"acc_norm_stderr": 0.03225799476233485 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.026069362295335134, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.026069362295335134 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.19658119658119658, |
|
"acc_stderr": 0.02603538609895129, |
|
"acc_norm": 0.19658119658119658, |
|
"acc_norm_stderr": 0.02603538609895129 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.27547169811320754, |
|
"acc_stderr": 0.02749566368372406, |
|
"acc_norm": 0.27547169811320754, |
|
"acc_norm_stderr": 0.02749566368372406 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.21818181818181817, |
|
"acc_stderr": 0.03955932861795833, |
|
"acc_norm": 0.21818181818181817, |
|
"acc_norm_stderr": 0.03955932861795833 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.22592592592592592, |
|
"acc_stderr": 0.02549753263960954, |
|
"acc_norm": 0.22592592592592592, |
|
"acc_norm_stderr": 0.02549753263960954 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.3509933774834437, |
|
"acc_stderr": 0.03896981964257375, |
|
"acc_norm": 0.3509933774834437, |
|
"acc_norm_stderr": 0.03896981964257375 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.24875621890547264, |
|
"acc_stderr": 0.030567675938916707, |
|
"acc_norm": 0.24875621890547264, |
|
"acc_norm_stderr": 0.030567675938916707 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.2832369942196532, |
|
"acc_stderr": 0.034355680560478746, |
|
"acc_norm": 0.2832369942196532, |
|
"acc_norm_stderr": 0.034355680560478746 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.23809523809523808, |
|
"acc_stderr": 0.021935878081184756, |
|
"acc_norm": 0.23809523809523808, |
|
"acc_norm_stderr": 0.021935878081184756 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.2847222222222222, |
|
"acc_stderr": 0.03773809990686935, |
|
"acc_norm": 0.2847222222222222, |
|
"acc_norm_stderr": 0.03773809990686935 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.29, |
|
"acc_stderr": 0.04560480215720683, |
|
"acc_norm": 0.29, |
|
"acc_norm_stderr": 0.04560480215720683 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.2658959537572254, |
|
"acc_stderr": 0.023786203255508297, |
|
"acc_norm": 0.2658959537572254, |
|
"acc_norm_stderr": 0.023786203255508297 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.2392638036809816, |
|
"acc_stderr": 0.03351953879521272, |
|
"acc_norm": 0.2392638036809816, |
|
"acc_norm_stderr": 0.03351953879521272 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.19753086419753085, |
|
"acc_stderr": 0.02215288992789894, |
|
"acc_norm": 0.19753086419753085, |
|
"acc_norm_stderr": 0.02215288992789894 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.27, |
|
"acc_stderr": 0.044619604333847394, |
|
"acc_norm": 0.27, |
|
"acc_norm_stderr": 0.044619604333847394 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.32124352331606215, |
|
"acc_stderr": 0.033699508685490674, |
|
"acc_norm": 0.32124352331606215, |
|
"acc_norm_stderr": 0.033699508685490674 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2631578947368421, |
|
"acc_stderr": 0.0414243971948936, |
|
"acc_norm": 0.2631578947368421, |
|
"acc_norm_stderr": 0.0414243971948936 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.3669724770642202, |
|
"acc_stderr": 0.02066467565952053, |
|
"acc_norm": 0.3669724770642202, |
|
"acc_norm_stderr": 0.02066467565952053 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.29365079365079366, |
|
"acc_stderr": 0.04073524322147126, |
|
"acc_norm": 0.29365079365079366, |
|
"acc_norm_stderr": 0.04073524322147126 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.2973856209150327, |
|
"acc_stderr": 0.026173908506718576, |
|
"acc_norm": 0.2973856209150327, |
|
"acc_norm_stderr": 0.026173908506718576 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.2, |
|
"acc_stderr": 0.04020151261036844, |
|
"acc_norm": 0.2, |
|
"acc_norm_stderr": 0.04020151261036844 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.1322314049586777, |
|
"acc_stderr": 0.030922788320445812, |
|
"acc_norm": 0.1322314049586777, |
|
"acc_norm_stderr": 0.030922788320445812 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.34210526315789475, |
|
"acc_stderr": 0.03860731599316092, |
|
"acc_norm": 0.34210526315789475, |
|
"acc_norm_stderr": 0.03860731599316092 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.2173202614379085, |
|
"acc_stderr": 0.016684820929148598, |
|
"acc_norm": 0.2173202614379085, |
|
"acc_norm_stderr": 0.016684820929148598 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.21631205673758866, |
|
"acc_stderr": 0.024561720560562793, |
|
"acc_norm": 0.21631205673758866, |
|
"acc_norm_stderr": 0.024561720560562793 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.20535714285714285, |
|
"acc_stderr": 0.03834241021419073, |
|
"acc_norm": 0.20535714285714285, |
|
"acc_norm_stderr": 0.03834241021419073 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.4675925925925926, |
|
"acc_stderr": 0.03402801581358966, |
|
"acc_norm": 0.4675925925925926, |
|
"acc_norm_stderr": 0.03402801581358966 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.27262569832402234, |
|
"acc_stderr": 0.014893391735249608, |
|
"acc_norm": 0.27262569832402234, |
|
"acc_norm_stderr": 0.014893391735249608 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.33, |
|
"acc_stderr": 0.04725815626252604, |
|
"acc_norm": 0.33, |
|
"acc_norm_stderr": 0.04725815626252604 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.23, |
|
"acc_stderr": 0.04229525846816506, |
|
"acc_norm": 0.23, |
|
"acc_norm_stderr": 0.04229525846816506 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.44485294117647056, |
|
"acc_stderr": 0.030187532060329383, |
|
"acc_norm": 0.44485294117647056, |
|
"acc_norm_stderr": 0.030187532060329383 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.39591836734693875, |
|
"acc_stderr": 0.03130802899065685, |
|
"acc_norm": 0.39591836734693875, |
|
"acc_norm_stderr": 0.03130802899065685 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.22362869198312235, |
|
"acc_stderr": 0.027123298205229972, |
|
"acc_norm": 0.22362869198312235, |
|
"acc_norm_stderr": 0.027123298205229972 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.24967405475880053, |
|
"acc_stderr": 0.011054538377832327, |
|
"acc_norm": 0.24967405475880053, |
|
"acc_norm_stderr": 0.011054538377832327 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.2549019607843137, |
|
"acc_stderr": 0.030587591351604246, |
|
"acc_norm": 0.2549019607843137, |
|
"acc_norm_stderr": 0.030587591351604246 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.2787878787878788, |
|
"acc_stderr": 0.03501438706296781, |
|
"acc_norm": 0.2787878787878788, |
|
"acc_norm_stderr": 0.03501438706296781 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.26193390452876375, |
|
"mc1_stderr": 0.015392118805015016, |
|
"mc2": 0.40797537743571977, |
|
"mc2_stderr": 0.014976707161150397 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.2975206611570248, |
|
"acc_stderr": 0.01571774220508993, |
|
"acc_norm": 0.3742621015348288, |
|
"acc_norm_stderr": 0.01663791778979874 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "heegyu/AULM-5.8b-v0804-hf", |
|
"model_sha": "ddcfd46cc8b42d7fb6ad822d97b6c30dfd3c028b", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |