|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.34812286689419797, |
|
"acc_stderr": 0.013921008595179344, |
|
"acc_norm": 0.40784982935153585, |
|
"acc_norm_stderr": 0.014361097288449691 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.36875124477195775, |
|
"acc_stderr": 0.0048148030984368085, |
|
"acc_norm": 0.4722166899024099, |
|
"acc_norm_stderr": 0.004982072108448084 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.45614035087719296, |
|
"acc_stderr": 0.03820042586602966, |
|
"acc_norm": 0.45614035087719296, |
|
"acc_norm_stderr": 0.03820042586602966 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.5145631067961165, |
|
"acc_stderr": 0.04948637324026637, |
|
"acc_norm": 0.5145631067961165, |
|
"acc_norm_stderr": 0.04948637324026637 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.4840357598978289, |
|
"acc_stderr": 0.01787084750608173, |
|
"acc_norm": 0.4840357598978289, |
|
"acc_norm_stderr": 0.01787084750608173 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.35555555555555557, |
|
"acc_stderr": 0.04135176749720386, |
|
"acc_norm": 0.35555555555555557, |
|
"acc_norm_stderr": 0.04135176749720386 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.37, |
|
"acc_stderr": 0.048523658709391, |
|
"acc_norm": 0.37, |
|
"acc_norm_stderr": 0.048523658709391 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.3574468085106383, |
|
"acc_stderr": 0.03132941789476425, |
|
"acc_norm": 0.3574468085106383, |
|
"acc_norm_stderr": 0.03132941789476425 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.3795180722891566, |
|
"acc_stderr": 0.03777798822748016, |
|
"acc_norm": 0.3795180722891566, |
|
"acc_norm_stderr": 0.03777798822748016 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.5016077170418006, |
|
"acc_stderr": 0.02839794490780661, |
|
"acc_norm": 0.5016077170418006, |
|
"acc_norm_stderr": 0.02839794490780661 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.38565022421524664, |
|
"acc_stderr": 0.03266842214289201, |
|
"acc_norm": 0.38565022421524664, |
|
"acc_norm_stderr": 0.03266842214289201 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.4351145038167939, |
|
"acc_stderr": 0.04348208051644858, |
|
"acc_norm": 0.4351145038167939, |
|
"acc_norm_stderr": 0.04348208051644858 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.43, |
|
"acc_stderr": 0.049756985195624284, |
|
"acc_norm": 0.43, |
|
"acc_norm_stderr": 0.049756985195624284 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.5808080808080808, |
|
"acc_stderr": 0.035155207286704175, |
|
"acc_norm": 0.5808080808080808, |
|
"acc_norm_stderr": 0.035155207286704175 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.4827586206896552, |
|
"acc_stderr": 0.04164188720169377, |
|
"acc_norm": 0.4827586206896552, |
|
"acc_norm_stderr": 0.04164188720169377 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.29411764705882354, |
|
"acc_stderr": 0.045338381959297736, |
|
"acc_norm": 0.29411764705882354, |
|
"acc_norm_stderr": 0.045338381959297736 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.47478991596638653, |
|
"acc_stderr": 0.0324371805513741, |
|
"acc_norm": 0.47478991596638653, |
|
"acc_norm_stderr": 0.0324371805513741 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.4461538461538462, |
|
"acc_stderr": 0.025203571773028333, |
|
"acc_norm": 0.4461538461538462, |
|
"acc_norm_stderr": 0.025203571773028333 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.050251890762960605, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.050251890762960605 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.32, |
|
"acc_stderr": 0.04688261722621504, |
|
"acc_norm": 0.32, |
|
"acc_norm_stderr": 0.04688261722621504 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.49074074074074076, |
|
"acc_stderr": 0.04832853553437055, |
|
"acc_norm": 0.49074074074074076, |
|
"acc_norm_stderr": 0.04832853553437055 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.37438423645320196, |
|
"acc_stderr": 0.03405155380561952, |
|
"acc_norm": 0.37438423645320196, |
|
"acc_norm_stderr": 0.03405155380561952 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.432258064516129, |
|
"acc_stderr": 0.02818173972001941, |
|
"acc_norm": 0.432258064516129, |
|
"acc_norm_stderr": 0.02818173972001941 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.6666666666666666, |
|
"acc_stderr": 0.030882736974138653, |
|
"acc_norm": 0.6666666666666666, |
|
"acc_norm_stderr": 0.030882736974138653 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.5094339622641509, |
|
"acc_stderr": 0.030767394707808107, |
|
"acc_norm": 0.5094339622641509, |
|
"acc_norm_stderr": 0.030767394707808107 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.4090909090909091, |
|
"acc_stderr": 0.04709306978661896, |
|
"acc_norm": 0.4090909090909091, |
|
"acc_norm_stderr": 0.04709306978661896 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.3074074074074074, |
|
"acc_stderr": 0.028133252578815642, |
|
"acc_norm": 0.3074074074074074, |
|
"acc_norm_stderr": 0.028133252578815642 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.3443708609271523, |
|
"acc_stderr": 0.03879687024073327, |
|
"acc_norm": 0.3443708609271523, |
|
"acc_norm_stderr": 0.03879687024073327 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.6069651741293532, |
|
"acc_stderr": 0.0345368246603156, |
|
"acc_norm": 0.6069651741293532, |
|
"acc_norm_stderr": 0.0345368246603156 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.4393063583815029, |
|
"acc_stderr": 0.037842719328874674, |
|
"acc_norm": 0.4393063583815029, |
|
"acc_norm_stderr": 0.037842719328874674 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.3544973544973545, |
|
"acc_stderr": 0.024636830602841997, |
|
"acc_norm": 0.3544973544973545, |
|
"acc_norm_stderr": 0.024636830602841997 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.3472222222222222, |
|
"acc_stderr": 0.039812405437178615, |
|
"acc_norm": 0.3472222222222222, |
|
"acc_norm_stderr": 0.039812405437178615 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.0479372485441102, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.0479372485441102 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.55, |
|
"acc_stderr": 0.05, |
|
"acc_norm": 0.55, |
|
"acc_norm_stderr": 0.05 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.4884393063583815, |
|
"acc_stderr": 0.026911898686377913, |
|
"acc_norm": 0.4884393063583815, |
|
"acc_norm_stderr": 0.026911898686377913 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.44171779141104295, |
|
"acc_stderr": 0.039015918258361836, |
|
"acc_norm": 0.44171779141104295, |
|
"acc_norm_stderr": 0.039015918258361836 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.43209876543209874, |
|
"acc_stderr": 0.02756301097160668, |
|
"acc_norm": 0.43209876543209874, |
|
"acc_norm_stderr": 0.02756301097160668 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.38, |
|
"acc_stderr": 0.048783173121456316, |
|
"acc_norm": 0.38, |
|
"acc_norm_stderr": 0.048783173121456316 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.47668393782383417, |
|
"acc_stderr": 0.03604513672442205, |
|
"acc_norm": 0.47668393782383417, |
|
"acc_norm_stderr": 0.03604513672442205 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2982456140350877, |
|
"acc_stderr": 0.04303684033537317, |
|
"acc_norm": 0.2982456140350877, |
|
"acc_norm_stderr": 0.04303684033537317 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.48807339449541287, |
|
"acc_stderr": 0.021431223617362223, |
|
"acc_norm": 0.48807339449541287, |
|
"acc_norm_stderr": 0.021431223617362223 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.29365079365079366, |
|
"acc_stderr": 0.04073524322147124, |
|
"acc_norm": 0.29365079365079366, |
|
"acc_norm_stderr": 0.04073524322147124 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.4673202614379085, |
|
"acc_stderr": 0.02856869975222588, |
|
"acc_norm": 0.4673202614379085, |
|
"acc_norm_stderr": 0.02856869975222588 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.43, |
|
"acc_stderr": 0.049756985195624284, |
|
"acc_norm": 0.43, |
|
"acc_norm_stderr": 0.049756985195624284 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.628099173553719, |
|
"acc_stderr": 0.04412015806624504, |
|
"acc_norm": 0.628099173553719, |
|
"acc_norm_stderr": 0.04412015806624504 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.4342105263157895, |
|
"acc_stderr": 0.040335656678483184, |
|
"acc_norm": 0.4342105263157895, |
|
"acc_norm_stderr": 0.040335656678483184 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.3660130718954248, |
|
"acc_stderr": 0.019488025745529658, |
|
"acc_norm": 0.3660130718954248, |
|
"acc_norm_stderr": 0.019488025745529658 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.32269503546099293, |
|
"acc_stderr": 0.027889139300534792, |
|
"acc_norm": 0.32269503546099293, |
|
"acc_norm_stderr": 0.027889139300534792 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.3125, |
|
"acc_stderr": 0.043994650575715215, |
|
"acc_norm": 0.3125, |
|
"acc_norm_stderr": 0.043994650575715215 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.3611111111111111, |
|
"acc_stderr": 0.032757734861009996, |
|
"acc_norm": 0.3611111111111111, |
|
"acc_norm_stderr": 0.032757734861009996 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.23016759776536314, |
|
"acc_stderr": 0.014078339253425807, |
|
"acc_norm": 0.23016759776536314, |
|
"acc_norm_stderr": 0.014078339253425807 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.43, |
|
"acc_stderr": 0.049756985195624284, |
|
"acc_norm": 0.43, |
|
"acc_norm_stderr": 0.049756985195624284 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.45, |
|
"acc_stderr": 0.05, |
|
"acc_norm": 0.45, |
|
"acc_norm_stderr": 0.05 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.3860294117647059, |
|
"acc_stderr": 0.029573269134411124, |
|
"acc_norm": 0.3860294117647059, |
|
"acc_norm_stderr": 0.029573269134411124 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.4775510204081633, |
|
"acc_stderr": 0.03197694118713673, |
|
"acc_norm": 0.4775510204081633, |
|
"acc_norm_stderr": 0.03197694118713673 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.5864978902953587, |
|
"acc_stderr": 0.03205649904851858, |
|
"acc_norm": 0.5864978902953587, |
|
"acc_norm_stderr": 0.03205649904851858 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.3155149934810952, |
|
"acc_stderr": 0.011869184843058643, |
|
"acc_norm": 0.3155149934810952, |
|
"acc_norm_stderr": 0.011869184843058643 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.49019607843137253, |
|
"acc_stderr": 0.03508637358630572, |
|
"acc_norm": 0.49019607843137253, |
|
"acc_norm_stderr": 0.03508637358630572 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.5151515151515151, |
|
"acc_stderr": 0.03902551007374448, |
|
"acc_norm": 0.5151515151515151, |
|
"acc_norm_stderr": 0.03902551007374448 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2766217870257038, |
|
"mc1_stderr": 0.015659605755326912, |
|
"mc2": 0.4484601943910918, |
|
"mc2_stderr": 0.015458891626438749 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.35093896713615025, |
|
"acc_stderr": 0.016360395003030395, |
|
"acc_norm": 0.4612676056338028, |
|
"acc_norm_stderr": 0.017088275735102635 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "maywell/Synatra_TbST11B_EP01", |
|
"model_sha": "54fd9892d9189a077506a024602f48d83af7d383", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |