|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.26791808873720135, |
|
"acc_stderr": 0.012942030195136421, |
|
"acc_norm": 0.310580204778157, |
|
"acc_norm_stderr": 0.013522292098053057 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.33389762995419237, |
|
"acc_stderr": 0.004706398252382464, |
|
"acc_norm": 0.4122684724158534, |
|
"acc_norm_stderr": 0.004912370023913011 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.4269005847953216, |
|
"acc_stderr": 0.03793620616529916, |
|
"acc_norm": 0.4269005847953216, |
|
"acc_norm_stderr": 0.03793620616529916 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.2912621359223301, |
|
"acc_stderr": 0.04498676320572922, |
|
"acc_norm": 0.2912621359223301, |
|
"acc_norm_stderr": 0.04498676320572922 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.34738186462324394, |
|
"acc_stderr": 0.01702667174865573, |
|
"acc_norm": 0.34738186462324394, |
|
"acc_norm_stderr": 0.01702667174865573 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.2740740740740741, |
|
"acc_stderr": 0.03853254836552003, |
|
"acc_norm": 0.2740740740740741, |
|
"acc_norm_stderr": 0.03853254836552003 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.28, |
|
"acc_stderr": 0.045126085985421276, |
|
"acc_norm": 0.28, |
|
"acc_norm_stderr": 0.045126085985421276 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.35319148936170214, |
|
"acc_stderr": 0.031245325202761923, |
|
"acc_norm": 0.35319148936170214, |
|
"acc_norm_stderr": 0.031245325202761923 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.2891566265060241, |
|
"acc_stderr": 0.03529486801511115, |
|
"acc_norm": 0.2891566265060241, |
|
"acc_norm_stderr": 0.03529486801511115 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.3440514469453376, |
|
"acc_stderr": 0.026981478043648022, |
|
"acc_norm": 0.3440514469453376, |
|
"acc_norm_stderr": 0.026981478043648022 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.3721973094170404, |
|
"acc_stderr": 0.032443052830087304, |
|
"acc_norm": 0.3721973094170404, |
|
"acc_norm_stderr": 0.032443052830087304 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.2748091603053435, |
|
"acc_stderr": 0.03915345408847837, |
|
"acc_norm": 0.2748091603053435, |
|
"acc_norm_stderr": 0.03915345408847837 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.2676767676767677, |
|
"acc_stderr": 0.031544498882702866, |
|
"acc_norm": 0.2676767676767677, |
|
"acc_norm_stderr": 0.031544498882702866 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.2896551724137931, |
|
"acc_stderr": 0.037800192304380135, |
|
"acc_norm": 0.2896551724137931, |
|
"acc_norm_stderr": 0.037800192304380135 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.23529411764705882, |
|
"acc_stderr": 0.04220773659171453, |
|
"acc_norm": 0.23529411764705882, |
|
"acc_norm_stderr": 0.04220773659171453 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.2773109243697479, |
|
"acc_stderr": 0.029079374539480007, |
|
"acc_norm": 0.2773109243697479, |
|
"acc_norm_stderr": 0.029079374539480007 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.24871794871794872, |
|
"acc_stderr": 0.0219169577092138, |
|
"acc_norm": 0.24871794871794872, |
|
"acc_norm_stderr": 0.0219169577092138 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.39, |
|
"acc_stderr": 0.04902071300001974, |
|
"acc_norm": 0.39, |
|
"acc_norm_stderr": 0.04902071300001974 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.23, |
|
"acc_stderr": 0.04229525846816505, |
|
"acc_norm": 0.23, |
|
"acc_norm_stderr": 0.04229525846816505 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.04557239513497752, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.04557239513497752 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.2413793103448276, |
|
"acc_stderr": 0.030108330718011625, |
|
"acc_norm": 0.2413793103448276, |
|
"acc_norm_stderr": 0.030108330718011625 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.3032258064516129, |
|
"acc_stderr": 0.026148685930671746, |
|
"acc_norm": 0.3032258064516129, |
|
"acc_norm_stderr": 0.026148685930671746 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.41025641025641024, |
|
"acc_stderr": 0.03222414045241107, |
|
"acc_norm": 0.41025641025641024, |
|
"acc_norm_stderr": 0.03222414045241107 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.27547169811320754, |
|
"acc_stderr": 0.027495663683724057, |
|
"acc_norm": 0.27547169811320754, |
|
"acc_norm_stderr": 0.027495663683724057 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.38181818181818183, |
|
"acc_stderr": 0.04653429807913508, |
|
"acc_norm": 0.38181818181818183, |
|
"acc_norm_stderr": 0.04653429807913508 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.22962962962962963, |
|
"acc_stderr": 0.025644108639267624, |
|
"acc_norm": 0.22962962962962963, |
|
"acc_norm_stderr": 0.025644108639267624 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.2119205298013245, |
|
"acc_stderr": 0.03336767086567977, |
|
"acc_norm": 0.2119205298013245, |
|
"acc_norm_stderr": 0.03336767086567977 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.3383084577114428, |
|
"acc_stderr": 0.03345563070339192, |
|
"acc_norm": 0.3383084577114428, |
|
"acc_norm_stderr": 0.03345563070339192 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.2543352601156069, |
|
"acc_stderr": 0.0332055644308557, |
|
"acc_norm": 0.2543352601156069, |
|
"acc_norm_stderr": 0.0332055644308557 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.21693121693121692, |
|
"acc_stderr": 0.021227082449445045, |
|
"acc_norm": 0.21693121693121692, |
|
"acc_norm_stderr": 0.021227082449445045 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.2361111111111111, |
|
"acc_stderr": 0.03551446610810826, |
|
"acc_norm": 0.2361111111111111, |
|
"acc_norm_stderr": 0.03551446610810826 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.22, |
|
"acc_stderr": 0.041633319989322695, |
|
"acc_norm": 0.22, |
|
"acc_norm_stderr": 0.041633319989322695 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.0479372485441102, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.0479372485441102 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.2745664739884393, |
|
"acc_stderr": 0.02402774515526502, |
|
"acc_norm": 0.2745664739884393, |
|
"acc_norm_stderr": 0.02402774515526502 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.26993865030674846, |
|
"acc_stderr": 0.03487825168497892, |
|
"acc_norm": 0.26993865030674846, |
|
"acc_norm_stderr": 0.03487825168497892 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.3117283950617284, |
|
"acc_stderr": 0.02577311116963045, |
|
"acc_norm": 0.3117283950617284, |
|
"acc_norm_stderr": 0.02577311116963045 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.24352331606217617, |
|
"acc_stderr": 0.030975436386845436, |
|
"acc_norm": 0.24352331606217617, |
|
"acc_norm_stderr": 0.030975436386845436 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2543859649122807, |
|
"acc_stderr": 0.040969851398436716, |
|
"acc_norm": 0.2543859649122807, |
|
"acc_norm_stderr": 0.040969851398436716 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.24403669724770644, |
|
"acc_stderr": 0.018415286351416416, |
|
"acc_norm": 0.24403669724770644, |
|
"acc_norm_stderr": 0.018415286351416416 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.20634920634920634, |
|
"acc_stderr": 0.036196045241242494, |
|
"acc_norm": 0.20634920634920634, |
|
"acc_norm_stderr": 0.036196045241242494 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.28431372549019607, |
|
"acc_stderr": 0.025829163272757475, |
|
"acc_norm": 0.28431372549019607, |
|
"acc_norm_stderr": 0.025829163272757475 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.38, |
|
"acc_stderr": 0.04878317312145632, |
|
"acc_norm": 0.38, |
|
"acc_norm_stderr": 0.04878317312145632 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.4462809917355372, |
|
"acc_stderr": 0.04537935177947879, |
|
"acc_norm": 0.4462809917355372, |
|
"acc_norm_stderr": 0.04537935177947879 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.17763157894736842, |
|
"acc_stderr": 0.03110318238312338, |
|
"acc_norm": 0.17763157894736842, |
|
"acc_norm_stderr": 0.03110318238312338 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.2696078431372549, |
|
"acc_stderr": 0.017952449196987862, |
|
"acc_norm": 0.2696078431372549, |
|
"acc_norm_stderr": 0.017952449196987862 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.25177304964539005, |
|
"acc_stderr": 0.025892151156709405, |
|
"acc_norm": 0.25177304964539005, |
|
"acc_norm_stderr": 0.025892151156709405 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.30357142857142855, |
|
"acc_stderr": 0.04364226155841044, |
|
"acc_norm": 0.30357142857142855, |
|
"acc_norm_stderr": 0.04364226155841044 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.24537037037037038, |
|
"acc_stderr": 0.029346665094372937, |
|
"acc_norm": 0.24537037037037038, |
|
"acc_norm_stderr": 0.029346665094372937 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.2424581005586592, |
|
"acc_stderr": 0.01433352205921789, |
|
"acc_norm": 0.2424581005586592, |
|
"acc_norm_stderr": 0.01433352205921789 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621504, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621504 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.28, |
|
"acc_stderr": 0.04512608598542127, |
|
"acc_norm": 0.28, |
|
"acc_norm_stderr": 0.04512608598542127 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.22426470588235295, |
|
"acc_stderr": 0.025336848563332372, |
|
"acc_norm": 0.22426470588235295, |
|
"acc_norm_stderr": 0.025336848563332372 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.35918367346938773, |
|
"acc_stderr": 0.03071356045510849, |
|
"acc_norm": 0.35918367346938773, |
|
"acc_norm_stderr": 0.03071356045510849 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.28270042194092826, |
|
"acc_stderr": 0.029312814153955924, |
|
"acc_norm": 0.28270042194092826, |
|
"acc_norm_stderr": 0.029312814153955924 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.24511082138200782, |
|
"acc_stderr": 0.010986307870045517, |
|
"acc_norm": 0.24511082138200782, |
|
"acc_norm_stderr": 0.010986307870045517 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.25980392156862747, |
|
"acc_stderr": 0.030778554678693254, |
|
"acc_norm": 0.25980392156862747, |
|
"acc_norm_stderr": 0.030778554678693254 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.2545454545454545, |
|
"acc_stderr": 0.0340150671524904, |
|
"acc_norm": 0.2545454545454545, |
|
"acc_norm_stderr": 0.0340150671524904 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2766217870257038, |
|
"mc1_stderr": 0.015659605755326905, |
|
"mc2": 0.4649376014172755, |
|
"mc2_stderr": 0.015443831068166118 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.26564344746162927, |
|
"acc_stderr": 0.015185107107791248, |
|
"acc_norm": 0.35182998819362454, |
|
"acc_norm_stderr": 0.016418206451218054 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "eclipsemint/kollama2-7b-v0", |
|
"model_sha": "e2a3ee343f997cca7ad3e25b5d970376d79c5b4e", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |