{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.29266211604095566, "acc_stderr": 0.013295916103619411, "acc_norm": 0.36689419795221845, "acc_norm_stderr": 0.014084133118104303 }, "harness|ko_hellaswag|10": { "acc": 0.2984465245966939, "acc_stderr": 0.004566412808642455, "acc_norm": 0.3603863772156941, "acc_norm_stderr": 0.004791313101877055 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.6666666666666666, "acc_stderr": 0.036155076303109344, "acc_norm": 0.6666666666666666, "acc_norm_stderr": 0.036155076303109344 }, "harness|ko_mmlu_management|5": { "acc": 0.7572815533980582, "acc_stderr": 0.04245022486384493, "acc_norm": 0.7572815533980582, "acc_norm_stderr": 0.04245022486384493 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.6143039591315453, "acc_stderr": 0.01740647661921291, "acc_norm": 0.6143039591315453, "acc_norm_stderr": 0.01740647661921291 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.45925925925925926, "acc_stderr": 0.04304979692464243, "acc_norm": 0.45925925925925926, "acc_norm_stderr": 0.04304979692464243 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.5319148936170213, "acc_stderr": 0.03261936918467381, "acc_norm": 0.5319148936170213, "acc_norm_stderr": 0.03261936918467381 }, "harness|ko_mmlu_virology|5": { "acc": 0.42771084337349397, "acc_stderr": 0.03851597683718533, "acc_norm": 0.42771084337349397, "acc_norm_stderr": 0.03851597683718533 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.5691318327974276, "acc_stderr": 0.028125340983972718, "acc_norm": 0.5691318327974276, "acc_norm_stderr": 0.028125340983972718 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.515695067264574, "acc_stderr": 0.0335412657542081, "acc_norm": 0.515695067264574, "acc_norm_stderr": 0.0335412657542081 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.6106870229007634, "acc_stderr": 0.04276486542814591, "acc_norm": 0.6106870229007634, "acc_norm_stderr": 0.04276486542814591 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.48, "acc_stderr": 0.05021167315686779, "acc_norm": 0.48, "acc_norm_stderr": 0.05021167315686779 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.702020202020202, "acc_stderr": 0.03258630383836556, "acc_norm": 0.702020202020202, "acc_norm_stderr": 0.03258630383836556 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.5586206896551724, "acc_stderr": 0.04137931034482757, "acc_norm": 0.5586206896551724, "acc_norm_stderr": 0.04137931034482757 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.37254901960784315, "acc_stderr": 0.04810840148082636, "acc_norm": 0.37254901960784315, "acc_norm_stderr": 0.04810840148082636 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.6008403361344538, "acc_stderr": 0.03181110032413926, "acc_norm": 0.6008403361344538, "acc_norm_stderr": 0.03181110032413926 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.6051282051282051, "acc_stderr": 0.024784316942156413, "acc_norm": 0.6051282051282051, "acc_norm_stderr": 0.024784316942156413 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.67, "acc_stderr": 0.04725815626252611, "acc_norm": 0.67, "acc_norm_stderr": 0.04725815626252611 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.4, "acc_stderr": 0.04923659639173309, "acc_norm": 0.4, "acc_norm_stderr": 0.04923659639173309 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.6481481481481481, "acc_stderr": 0.04616631111801713, "acc_norm": 0.6481481481481481, "acc_norm_stderr": 0.04616631111801713 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.5615763546798029, "acc_stderr": 0.03491207857486519, "acc_norm": 0.5615763546798029, "acc_norm_stderr": 0.03491207857486519 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.6129032258064516, "acc_stderr": 0.0277093596750325, "acc_norm": 0.6129032258064516, "acc_norm_stderr": 0.0277093596750325 }, "harness|ko_mmlu_marketing|5": { "acc": 0.811965811965812, "acc_stderr": 0.025598193686652244, "acc_norm": 0.811965811965812, "acc_norm_stderr": 0.025598193686652244 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.5509433962264151, "acc_stderr": 0.030612730713641092, "acc_norm": 0.5509433962264151, "acc_norm_stderr": 0.030612730713641092 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.5636363636363636, "acc_stderr": 0.04750185058907297, "acc_norm": 0.5636363636363636, "acc_norm_stderr": 0.04750185058907297 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.40370370370370373, "acc_stderr": 0.029914812342227624, "acc_norm": 0.40370370370370373, "acc_norm_stderr": 0.029914812342227624 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.3708609271523179, "acc_stderr": 0.03943966699183629, "acc_norm": 0.3708609271523179, "acc_norm_stderr": 0.03943966699183629 }, "harness|ko_mmlu_sociology|5": { "acc": 0.6716417910447762, "acc_stderr": 0.033206858897443244, "acc_norm": 0.6716417910447762, "acc_norm_stderr": 0.033206858897443244 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.5433526011560693, "acc_stderr": 0.03798106566014498, "acc_norm": 0.5433526011560693, "acc_norm_stderr": 0.03798106566014498 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.544973544973545, "acc_stderr": 0.02564692836104939, "acc_norm": 0.544973544973545, "acc_norm_stderr": 0.02564692836104939 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.4722222222222222, "acc_stderr": 0.04174752578923183, "acc_norm": 0.4722222222222222, "acc_norm_stderr": 0.04174752578923183 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.39, "acc_stderr": 0.04902071300001975, "acc_norm": 0.39, "acc_norm_stderr": 0.04902071300001975 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.69, "acc_stderr": 0.04648231987117316, "acc_norm": 0.69, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.5635838150289018, "acc_stderr": 0.026700545424943684, "acc_norm": 0.5635838150289018, "acc_norm_stderr": 0.026700545424943684 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.5644171779141104, "acc_stderr": 0.03895632464138938, "acc_norm": 0.5644171779141104, "acc_norm_stderr": 0.03895632464138938 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.5987654320987654, "acc_stderr": 0.027272582849839803, "acc_norm": 0.5987654320987654, "acc_norm_stderr": 0.027272582849839803 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.35, "acc_stderr": 0.047937248544110196, "acc_norm": 0.35, "acc_norm_stderr": 0.047937248544110196 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.6528497409326425, "acc_stderr": 0.03435696168361355, "acc_norm": 0.6528497409326425, "acc_norm_stderr": 0.03435696168361355 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.42105263157894735, "acc_stderr": 0.046446020912223177, "acc_norm": 0.42105263157894735, "acc_norm_stderr": 0.046446020912223177 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.6972477064220184, "acc_stderr": 0.01969871143475635, "acc_norm": 0.6972477064220184, "acc_norm_stderr": 0.01969871143475635 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.46825396825396826, "acc_stderr": 0.044631127206771704, "acc_norm": 0.46825396825396826, "acc_norm_stderr": 0.044631127206771704 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.5784313725490197, "acc_stderr": 0.028275490156791462, "acc_norm": 0.5784313725490197, "acc_norm_stderr": 0.028275490156791462 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.61, "acc_stderr": 0.049020713000019756, "acc_norm": 0.61, "acc_norm_stderr": 0.049020713000019756 }, "harness|ko_mmlu_international_law|5": { "acc": 0.7107438016528925, "acc_stderr": 0.04139112727635463, "acc_norm": 0.7107438016528925, "acc_norm_stderr": 0.04139112727635463 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.6513157894736842, "acc_stderr": 0.03878139888797611, "acc_norm": 0.6513157894736842, "acc_norm_stderr": 0.03878139888797611 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.4934640522875817, "acc_stderr": 0.020226106567657814, "acc_norm": 0.4934640522875817, "acc_norm_stderr": 0.020226106567657814 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.375886524822695, "acc_stderr": 0.028893955412115886, "acc_norm": 0.375886524822695, "acc_norm_stderr": 0.028893955412115886 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.42857142857142855, "acc_stderr": 0.04697113923010213, "acc_norm": 0.42857142857142855, "acc_norm_stderr": 0.04697113923010213 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.5601851851851852, "acc_stderr": 0.03385177976044811, "acc_norm": 0.5601851851851852, "acc_norm_stderr": 0.03385177976044811 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.3787709497206704, "acc_stderr": 0.01622353351036513, "acc_norm": 0.3787709497206704, "acc_norm_stderr": 0.01622353351036513 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.44, "acc_stderr": 0.04988876515698589, "acc_norm": 0.44, "acc_norm_stderr": 0.04988876515698589 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.68, "acc_stderr": 0.04688261722621505, "acc_norm": 0.68, "acc_norm_stderr": 0.04688261722621505 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.45588235294117646, "acc_stderr": 0.03025437257397669, "acc_norm": 0.45588235294117646, "acc_norm_stderr": 0.03025437257397669 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.6612244897959184, "acc_stderr": 0.030299506562154188, "acc_norm": 0.6612244897959184, "acc_norm_stderr": 0.030299506562154188 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.7341772151898734, "acc_stderr": 0.02875679962965833, "acc_norm": 0.7341772151898734, "acc_norm_stderr": 0.02875679962965833 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.38265971316818775, "acc_stderr": 0.012413595882893279, "acc_norm": 0.38265971316818775, "acc_norm_stderr": 0.012413595882893279 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.6274509803921569, "acc_stderr": 0.03393388584958404, "acc_norm": 0.6274509803921569, "acc_norm_stderr": 0.03393388584958404 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.6303030303030303, "acc_stderr": 0.03769430314512569, "acc_norm": 0.6303030303030303, "acc_norm_stderr": 0.03769430314512569 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2582619339045288, "mc1_stderr": 0.0153218216884762, "mc2": 0.49811141402996134, "mc2_stderr": 0.01690674036318429 }, "harness|ko_commongen_v2|2": { "acc": 0.3234946871310508, "acc_stderr": 0.01608362729048367, "acc_norm": 0.4946871310507674, "acc_norm_stderr": 0.01718938362722969 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "Alibaba-NLP/gte-Qwen2-7B-instruct", "model_sha": "e26182b2122f4435e8b3ebecbf363990f409b45b", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }