{ "config_general": { "lighteval_sha": "?", "num_fewshot_seeds": 1, "override_batch_size": 8, "max_samples": null, "job_id": "", "start_time": 2004961.140231565, "end_time": 2005992.444841733, "total_evaluation_time_secondes": "1031.3046101678628", "model_name": "Qwen/Qwen2.5-0.5B", "model_sha": "060db6499f32faf8b98477b0a26969ef7d8b9987", "model_dtype": "torch.bfloat16", "model_size": "942.3 MB", "config": null }, "results": { "custom|arc:challenge|0": { "acc": 0.2935153583617747, "acc_stderr": 0.013307250444941122, "acc_norm": 0.32081911262798635, "acc_norm_stderr": 0.013640943091946528 }, "custom|arc:easy|0": { "acc": 0.6430976430976431, "acc_stderr": 0.009830630210347016, "acc_norm": 0.5808080808080808, "acc_norm_stderr": 0.01012490528249118 }, "custom|commonsense_qa|0": { "acc": 0.29238329238329236, "acc_stderr": 0.013022531002213358, "acc_norm": 0.3153153153153153, "acc_norm_stderr": 0.013302642632905037 }, "custom|hellaswag|0": { "acc": 0.4025094602668791, "acc_stderr": 0.004894012555642639, "acc_norm": 0.5126468830910177, "acc_norm_stderr": 0.004988184988345287 }, "custom|mmlu_cloze:abstract_algebra|0": { "acc": 0.26, "acc_stderr": 0.04408440022768078, "acc_norm": 0.2, "acc_norm_stderr": 0.04020151261036845 }, "custom|mmlu_cloze:anatomy|0": { "acc": 0.34814814814814815, "acc_stderr": 0.041153246103369526, "acc_norm": 0.34814814814814815, "acc_norm_stderr": 0.041153246103369526 }, "custom|mmlu_cloze:astronomy|0": { "acc": 0.3026315789473684, "acc_stderr": 0.037385206761196686, "acc_norm": 0.39473684210526316, "acc_norm_stderr": 0.039777499346220734 }, "custom|mmlu_cloze:business_ethics|0": { "acc": 0.58, "acc_stderr": 0.04960449637488583, "acc_norm": 0.52, "acc_norm_stderr": 0.050211673156867795 }, "custom|mmlu_cloze:clinical_knowledge|0": { "acc": 0.27169811320754716, "acc_stderr": 0.02737770662467071, "acc_norm": 0.3886792452830189, "acc_norm_stderr": 0.03000048544867599 }, "custom|mmlu_cloze:college_biology|0": { "acc": 0.3125, "acc_stderr": 0.038760854559127644, "acc_norm": 0.3680555555555556, "acc_norm_stderr": 0.0403299905396072 }, "custom|mmlu_cloze:college_chemistry|0": { "acc": 0.24, "acc_stderr": 0.042923469599092816, "acc_norm": 0.27, "acc_norm_stderr": 0.044619604333847394 }, "custom|mmlu_cloze:college_computer_science|0": { "acc": 0.25, "acc_stderr": 0.04351941398892446, "acc_norm": 0.25, "acc_norm_stderr": 0.04351941398892446 }, "custom|mmlu_cloze:college_mathematics|0": { "acc": 0.17, "acc_stderr": 0.0377525168068637, "acc_norm": 0.23, "acc_norm_stderr": 0.042295258468165085 }, "custom|mmlu_cloze:college_medicine|0": { "acc": 0.31213872832369943, "acc_stderr": 0.03533133389323657, "acc_norm": 0.30057803468208094, "acc_norm_stderr": 0.0349610148119118 }, "custom|mmlu_cloze:college_physics|0": { "acc": 0.2647058823529412, "acc_stderr": 0.04389869956808778, "acc_norm": 0.2549019607843137, "acc_norm_stderr": 0.043364327079931785 }, "custom|mmlu_cloze:computer_security|0": { "acc": 0.43, "acc_stderr": 0.04975698519562428, "acc_norm": 0.41, "acc_norm_stderr": 0.049431107042371025 }, "custom|mmlu_cloze:conceptual_physics|0": { "acc": 0.425531914893617, "acc_stderr": 0.03232146916224469, "acc_norm": 0.37872340425531914, "acc_norm_stderr": 0.03170995606040655 }, "custom|mmlu_cloze:econometrics|0": { "acc": 0.22807017543859648, "acc_stderr": 0.03947152782669415, "acc_norm": 0.21929824561403508, "acc_norm_stderr": 0.03892431106518756 }, "custom|mmlu_cloze:electrical_engineering|0": { "acc": 0.3103448275862069, "acc_stderr": 0.03855289616378948, "acc_norm": 0.31724137931034485, "acc_norm_stderr": 0.03878352372138622 }, "custom|mmlu_cloze:elementary_mathematics|0": { "acc": 0.4470899470899471, "acc_stderr": 0.02560672399577703, "acc_norm": 0.4365079365079365, "acc_norm_stderr": 0.025542846817400496 }, "custom|mmlu_cloze:formal_logic|0": { "acc": 0.30952380952380953, "acc_stderr": 0.04134913018303317, "acc_norm": 0.35714285714285715, "acc_norm_stderr": 0.04285714285714281 }, "custom|mmlu_cloze:global_facts|0": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.32, "acc_norm_stderr": 0.04688261722621505 }, "custom|mmlu_cloze:high_school_biology|0": { "acc": 0.3193548387096774, "acc_stderr": 0.026522709674667775, "acc_norm": 0.36129032258064514, "acc_norm_stderr": 0.027327548447957543 }, "custom|mmlu_cloze:high_school_chemistry|0": { "acc": 0.2019704433497537, "acc_stderr": 0.02824735012218027, "acc_norm": 0.2857142857142857, "acc_norm_stderr": 0.03178529710642749 }, "custom|mmlu_cloze:high_school_computer_science|0": { "acc": 0.39, "acc_stderr": 0.04902071300001975, "acc_norm": 0.43, "acc_norm_stderr": 0.049756985195624284 }, "custom|mmlu_cloze:high_school_european_history|0": { "acc": 0.28484848484848485, "acc_stderr": 0.035243908445117836, "acc_norm": 0.41818181818181815, "acc_norm_stderr": 0.03851716319398394 }, "custom|mmlu_cloze:high_school_geography|0": { "acc": 0.3686868686868687, "acc_stderr": 0.034373055019806184, "acc_norm": 0.3484848484848485, "acc_norm_stderr": 0.033948539651564025 }, "custom|mmlu_cloze:high_school_government_and_politics|0": { "acc": 0.38341968911917096, "acc_stderr": 0.03508984236295342, "acc_norm": 0.37823834196891193, "acc_norm_stderr": 0.03499807276193337 }, "custom|mmlu_cloze:high_school_macroeconomics|0": { "acc": 0.3153846153846154, "acc_stderr": 0.02355964698318995, "acc_norm": 0.36153846153846153, "acc_norm_stderr": 0.02435958146539699 }, "custom|mmlu_cloze:high_school_mathematics|0": { "acc": 0.1962962962962963, "acc_stderr": 0.024217421327417145, "acc_norm": 0.26666666666666666, "acc_norm_stderr": 0.026962424325073838 }, "custom|mmlu_cloze:high_school_microeconomics|0": { "acc": 0.3403361344537815, "acc_stderr": 0.030778057422931673, "acc_norm": 0.40336134453781514, "acc_norm_stderr": 0.031866081214088314 }, "custom|mmlu_cloze:high_school_physics|0": { "acc": 0.2781456953642384, "acc_stderr": 0.03658603262763743, "acc_norm": 0.26490066225165565, "acc_norm_stderr": 0.03603038545360385 }, "custom|mmlu_cloze:high_school_psychology|0": { "acc": 0.44954128440366975, "acc_stderr": 0.021327881417823373, "acc_norm": 0.44770642201834865, "acc_norm_stderr": 0.021319754962425462 }, "custom|mmlu_cloze:high_school_statistics|0": { "acc": 0.28703703703703703, "acc_stderr": 0.03085199299325701, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.03214952147802747 }, "custom|mmlu_cloze:high_school_us_history|0": { "acc": 0.27450980392156865, "acc_stderr": 0.0313217980308329, "acc_norm": 0.3284313725490196, "acc_norm_stderr": 0.03296245110172228 }, "custom|mmlu_cloze:high_school_world_history|0": { "acc": 0.3037974683544304, "acc_stderr": 0.029936696387138608, "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.03068582059661081 }, "custom|mmlu_cloze:human_aging|0": { "acc": 0.3991031390134529, "acc_stderr": 0.032867453125679603, "acc_norm": 0.33183856502242154, "acc_norm_stderr": 0.03160295143776679 }, "custom|mmlu_cloze:human_sexuality|0": { "acc": 0.3893129770992366, "acc_stderr": 0.04276486542814591, "acc_norm": 0.35877862595419846, "acc_norm_stderr": 0.04206739313864908 }, "custom|mmlu_cloze:international_law|0": { "acc": 0.17355371900826447, "acc_stderr": 0.0345727283691767, "acc_norm": 0.30578512396694213, "acc_norm_stderr": 0.04205953933884123 }, "custom|mmlu_cloze:jurisprudence|0": { "acc": 0.23148148148148148, "acc_stderr": 0.04077494709252627, "acc_norm": 0.4351851851851852, "acc_norm_stderr": 0.04792898170907062 }, "custom|mmlu_cloze:logical_fallacies|0": { "acc": 0.294478527607362, "acc_stderr": 0.03581165790474082, "acc_norm": 0.3312883435582822, "acc_norm_stderr": 0.03697983910025588 }, "custom|mmlu_cloze:machine_learning|0": { "acc": 0.23214285714285715, "acc_stderr": 0.04007341809755805, "acc_norm": 0.25, "acc_norm_stderr": 0.04109974682633932 }, "custom|mmlu_cloze:management|0": { "acc": 0.3300970873786408, "acc_stderr": 0.0465614711001235, "acc_norm": 0.44660194174757284, "acc_norm_stderr": 0.04922424153458933 }, "custom|mmlu_cloze:marketing|0": { "acc": 0.47863247863247865, "acc_stderr": 0.03272616447634954, "acc_norm": 0.47435897435897434, "acc_norm_stderr": 0.03271298896811159 }, "custom|mmlu_cloze:medical_genetics|0": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.38, "acc_norm_stderr": 0.04878317312145632 }, "custom|mmlu_cloze:miscellaneous|0": { "acc": 0.46998722860791825, "acc_stderr": 0.017847723086649097, "acc_norm": 0.42656449553001274, "acc_norm_stderr": 0.01768606697567565 }, "custom|mmlu_cloze:moral_disputes|0": { "acc": 0.2774566473988439, "acc_stderr": 0.024105712607754307, "acc_norm": 0.2398843930635838, "acc_norm_stderr": 0.022989592543123567 }, "custom|mmlu_cloze:moral_scenarios|0": { "acc": 0.23798882681564246, "acc_stderr": 0.014242630070574915, "acc_norm": 0.27262569832402234, "acc_norm_stderr": 0.014893391735249588 }, "custom|mmlu_cloze:nutrition|0": { "acc": 0.25163398692810457, "acc_stderr": 0.0248480182638752, "acc_norm": 0.35294117647058826, "acc_norm_stderr": 0.02736359328468493 }, "custom|mmlu_cloze:philosophy|0": { "acc": 0.2797427652733119, "acc_stderr": 0.025494259350694888, "acc_norm": 0.3183279742765273, "acc_norm_stderr": 0.026457225067811025 }, "custom|mmlu_cloze:prehistory|0": { "acc": 0.3611111111111111, "acc_stderr": 0.02672586880910079, "acc_norm": 0.31790123456790126, "acc_norm_stderr": 0.025910063528240868 }, "custom|mmlu_cloze:professional_accounting|0": { "acc": 0.26595744680851063, "acc_stderr": 0.02635806569888059, "acc_norm": 0.24468085106382978, "acc_norm_stderr": 0.025645553622266722 }, "custom|mmlu_cloze:professional_law|0": { "acc": 0.2405475880052151, "acc_stderr": 0.010916406735478949, "acc_norm": 0.2653194263363755, "acc_norm_stderr": 0.011276198843958876 }, "custom|mmlu_cloze:professional_medicine|0": { "acc": 0.27941176470588236, "acc_stderr": 0.02725720260611495, "acc_norm": 0.3272058823529412, "acc_norm_stderr": 0.02850145286039657 }, "custom|mmlu_cloze:professional_psychology|0": { "acc": 0.3022875816993464, "acc_stderr": 0.018579232711113877, "acc_norm": 0.3104575163398693, "acc_norm_stderr": 0.018718067052623227 }, "custom|mmlu_cloze:public_relations|0": { "acc": 0.44545454545454544, "acc_stderr": 0.047605488214603246, "acc_norm": 0.3090909090909091, "acc_norm_stderr": 0.044262946482000985 }, "custom|mmlu_cloze:security_studies|0": { "acc": 0.3306122448979592, "acc_stderr": 0.030116426296540613, "acc_norm": 0.24081632653061225, "acc_norm_stderr": 0.027372942201788174 }, "custom|mmlu_cloze:sociology|0": { "acc": 0.2835820895522388, "acc_stderr": 0.03187187537919796, "acc_norm": 0.3482587064676617, "acc_norm_stderr": 0.033687874661154596 }, "custom|mmlu_cloze:us_foreign_policy|0": { "acc": 0.36, "acc_stderr": 0.04824181513244218, "acc_norm": 0.28, "acc_norm_stderr": 0.04512608598542128 }, "custom|mmlu_cloze:virology|0": { "acc": 0.26506024096385544, "acc_stderr": 0.03436024037944967, "acc_norm": 0.3072289156626506, "acc_norm_stderr": 0.035915667978246635 }, "custom|mmlu_cloze:world_religions|0": { "acc": 0.3742690058479532, "acc_stderr": 0.03711601185389481, "acc_norm": 0.4327485380116959, "acc_norm_stderr": 0.03799978644370608 }, "custom|mmlu_mc:abstract_algebra|0": { "acc": 0.32, "acc_stderr": 0.046882617226215034, "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "custom|mmlu_mc:anatomy|0": { "acc": 0.4148148148148148, "acc_stderr": 0.042561937679014075, "acc_norm": 0.4148148148148148, "acc_norm_stderr": 0.042561937679014075 }, "custom|mmlu_mc:astronomy|0": { "acc": 0.48026315789473684, "acc_stderr": 0.040657710025626036, "acc_norm": 0.48026315789473684, "acc_norm_stderr": 0.040657710025626036 }, "custom|mmlu_mc:business_ethics|0": { "acc": 0.52, "acc_stderr": 0.050211673156867795, "acc_norm": 0.52, "acc_norm_stderr": 0.050211673156867795 }, "custom|mmlu_mc:clinical_knowledge|0": { "acc": 0.5358490566037736, "acc_stderr": 0.030693675018458003, "acc_norm": 0.5358490566037736, "acc_norm_stderr": 0.030693675018458003 }, "custom|mmlu_mc:college_biology|0": { "acc": 0.4652777777777778, "acc_stderr": 0.04171115858181618, "acc_norm": 0.4652777777777778, "acc_norm_stderr": 0.04171115858181618 }, "custom|mmlu_mc:college_chemistry|0": { "acc": 0.34, "acc_stderr": 0.04760952285695236, "acc_norm": 0.34, "acc_norm_stderr": 0.04760952285695236 }, "custom|mmlu_mc:college_computer_science|0": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "custom|mmlu_mc:college_mathematics|0": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "custom|mmlu_mc:college_medicine|0": { "acc": 0.48554913294797686, "acc_stderr": 0.03810871630454764, "acc_norm": 0.48554913294797686, "acc_norm_stderr": 0.03810871630454764 }, "custom|mmlu_mc:college_physics|0": { "acc": 0.27450980392156865, "acc_stderr": 0.044405219061793275, "acc_norm": 0.27450980392156865, "acc_norm_stderr": 0.044405219061793275 }, "custom|mmlu_mc:computer_security|0": { "acc": 0.7, "acc_stderr": 0.046056618647183814, "acc_norm": 0.7, "acc_norm_stderr": 0.046056618647183814 }, "custom|mmlu_mc:conceptual_physics|0": { "acc": 0.4127659574468085, "acc_stderr": 0.03218471141400351, "acc_norm": 0.4127659574468085, "acc_norm_stderr": 0.03218471141400351 }, "custom|mmlu_mc:econometrics|0": { "acc": 0.3508771929824561, "acc_stderr": 0.04489539350270701, "acc_norm": 0.3508771929824561, "acc_norm_stderr": 0.04489539350270701 }, "custom|mmlu_mc:electrical_engineering|0": { "acc": 0.5379310344827586, "acc_stderr": 0.041546596717075474, "acc_norm": 0.5379310344827586, "acc_norm_stderr": 0.041546596717075474 }, "custom|mmlu_mc:elementary_mathematics|0": { "acc": 0.3783068783068783, "acc_stderr": 0.024976954053155243, "acc_norm": 0.3783068783068783, "acc_norm_stderr": 0.024976954053155243 }, "custom|mmlu_mc:formal_logic|0": { "acc": 0.35714285714285715, "acc_stderr": 0.04285714285714281, "acc_norm": 0.35714285714285715, "acc_norm_stderr": 0.04285714285714281 }, "custom|mmlu_mc:global_facts|0": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "custom|mmlu_mc:high_school_biology|0": { "acc": 0.5258064516129032, "acc_stderr": 0.02840609505765332, "acc_norm": 0.5258064516129032, "acc_norm_stderr": 0.02840609505765332 }, "custom|mmlu_mc:high_school_chemistry|0": { "acc": 0.43349753694581283, "acc_stderr": 0.034867317274198714, "acc_norm": 0.43349753694581283, "acc_norm_stderr": 0.034867317274198714 }, "custom|mmlu_mc:high_school_computer_science|0": { "acc": 0.51, "acc_stderr": 0.05024183937956912, "acc_norm": 0.51, "acc_norm_stderr": 0.05024183937956912 }, "custom|mmlu_mc:high_school_european_history|0": { "acc": 0.593939393939394, "acc_stderr": 0.03834816355401181, "acc_norm": 0.593939393939394, "acc_norm_stderr": 0.03834816355401181 }, "custom|mmlu_mc:high_school_geography|0": { "acc": 0.5757575757575758, "acc_stderr": 0.03521224908841586, "acc_norm": 0.5757575757575758, "acc_norm_stderr": 0.03521224908841586 }, "custom|mmlu_mc:high_school_government_and_politics|0": { "acc": 0.5699481865284974, "acc_stderr": 0.03572954333144809, "acc_norm": 0.5699481865284974, "acc_norm_stderr": 0.03572954333144809 }, "custom|mmlu_mc:high_school_macroeconomics|0": { "acc": 0.47435897435897434, "acc_stderr": 0.025317649726448652, "acc_norm": 0.47435897435897434, "acc_norm_stderr": 0.025317649726448652 }, "custom|mmlu_mc:high_school_mathematics|0": { "acc": 0.35185185185185186, "acc_stderr": 0.02911661760608302, "acc_norm": 0.35185185185185186, "acc_norm_stderr": 0.02911661760608302 }, "custom|mmlu_mc:high_school_microeconomics|0": { "acc": 0.5126050420168067, "acc_stderr": 0.032468167657521745, "acc_norm": 0.5126050420168067, "acc_norm_stderr": 0.032468167657521745 }, "custom|mmlu_mc:high_school_physics|0": { "acc": 0.31788079470198677, "acc_stderr": 0.038020397601079024, "acc_norm": 0.31788079470198677, "acc_norm_stderr": 0.038020397601079024 }, "custom|mmlu_mc:high_school_psychology|0": { "acc": 0.6477064220183486, "acc_stderr": 0.02048056884399899, "acc_norm": 0.6477064220183486, "acc_norm_stderr": 0.02048056884399899 }, "custom|mmlu_mc:high_school_statistics|0": { "acc": 0.37037037037037035, "acc_stderr": 0.03293377139415191, "acc_norm": 0.37037037037037035, "acc_norm_stderr": 0.03293377139415191 }, "custom|mmlu_mc:high_school_us_history|0": { "acc": 0.5245098039215687, "acc_stderr": 0.03505093194348798, "acc_norm": 0.5245098039215687, "acc_norm_stderr": 0.03505093194348798 }, "custom|mmlu_mc:high_school_world_history|0": { "acc": 0.6118143459915611, "acc_stderr": 0.03172295004332328, "acc_norm": 0.6118143459915611, "acc_norm_stderr": 0.03172295004332328 }, "custom|mmlu_mc:human_aging|0": { "acc": 0.5201793721973094, "acc_stderr": 0.033530461674123005, "acc_norm": 0.5201793721973094, "acc_norm_stderr": 0.033530461674123005 }, "custom|mmlu_mc:human_sexuality|0": { "acc": 0.5725190839694656, "acc_stderr": 0.04338920305792401, "acc_norm": 0.5725190839694656, "acc_norm_stderr": 0.04338920305792401 }, "custom|mmlu_mc:international_law|0": { "acc": 0.71900826446281, "acc_stderr": 0.04103203830514511, "acc_norm": 0.71900826446281, "acc_norm_stderr": 0.04103203830514511 }, "custom|mmlu_mc:jurisprudence|0": { "acc": 0.6018518518518519, "acc_stderr": 0.04732332615978814, "acc_norm": 0.6018518518518519, "acc_norm_stderr": 0.04732332615978814 }, "custom|mmlu_mc:logical_fallacies|0": { "acc": 0.49693251533742333, "acc_stderr": 0.03928297078179662, "acc_norm": 0.49693251533742333, "acc_norm_stderr": 0.03928297078179662 }, "custom|mmlu_mc:machine_learning|0": { "acc": 0.4017857142857143, "acc_stderr": 0.04653333146973647, "acc_norm": 0.4017857142857143, "acc_norm_stderr": 0.04653333146973647 }, "custom|mmlu_mc:management|0": { "acc": 0.6310679611650486, "acc_stderr": 0.0477761518115674, "acc_norm": 0.6310679611650486, "acc_norm_stderr": 0.0477761518115674 }, "custom|mmlu_mc:marketing|0": { "acc": 0.7564102564102564, "acc_stderr": 0.028120966503914394, "acc_norm": 0.7564102564102564, "acc_norm_stderr": 0.028120966503914394 }, "custom|mmlu_mc:medical_genetics|0": { "acc": 0.5, "acc_stderr": 0.050251890762960605, "acc_norm": 0.5, "acc_norm_stderr": 0.050251890762960605 }, "custom|mmlu_mc:miscellaneous|0": { "acc": 0.5530012771392082, "acc_stderr": 0.01777922523339421, "acc_norm": 0.5530012771392082, "acc_norm_stderr": 0.01777922523339421 }, "custom|mmlu_mc:moral_disputes|0": { "acc": 0.5, "acc_stderr": 0.026919095102908273, "acc_norm": 0.5, "acc_norm_stderr": 0.026919095102908273 }, "custom|mmlu_mc:moral_scenarios|0": { "acc": 0.23798882681564246, "acc_stderr": 0.014242630070574915, "acc_norm": 0.23798882681564246, "acc_norm_stderr": 0.014242630070574915 }, "custom|mmlu_mc:nutrition|0": { "acc": 0.6143790849673203, "acc_stderr": 0.02787074527829027, "acc_norm": 0.6143790849673203, "acc_norm_stderr": 0.02787074527829027 }, "custom|mmlu_mc:philosophy|0": { "acc": 0.4855305466237942, "acc_stderr": 0.02838619808417768, "acc_norm": 0.4855305466237942, "acc_norm_stderr": 0.02838619808417768 }, "custom|mmlu_mc:prehistory|0": { "acc": 0.5123456790123457, "acc_stderr": 0.027812262269327235, "acc_norm": 0.5123456790123457, "acc_norm_stderr": 0.027812262269327235 }, "custom|mmlu_mc:professional_accounting|0": { "acc": 0.36524822695035464, "acc_stderr": 0.02872386385328128, "acc_norm": 0.36524822695035464, "acc_norm_stderr": 0.02872386385328128 }, "custom|mmlu_mc:professional_law|0": { "acc": 0.34419817470664926, "acc_stderr": 0.012134433741002574, "acc_norm": 0.34419817470664926, "acc_norm_stderr": 0.012134433741002574 }, "custom|mmlu_mc:professional_medicine|0": { "acc": 0.43014705882352944, "acc_stderr": 0.030074971917302875, "acc_norm": 0.43014705882352944, "acc_norm_stderr": 0.030074971917302875 }, "custom|mmlu_mc:professional_psychology|0": { "acc": 0.4591503267973856, "acc_stderr": 0.020160213617222516, "acc_norm": 0.4591503267973856, "acc_norm_stderr": 0.020160213617222516 }, "custom|mmlu_mc:public_relations|0": { "acc": 0.5909090909090909, "acc_stderr": 0.04709306978661895, "acc_norm": 0.5909090909090909, "acc_norm_stderr": 0.04709306978661895 }, "custom|mmlu_mc:security_studies|0": { "acc": 0.6163265306122448, "acc_stderr": 0.03113088039623593, "acc_norm": 0.6163265306122448, "acc_norm_stderr": 0.03113088039623593 }, "custom|mmlu_mc:sociology|0": { "acc": 0.6965174129353234, "acc_stderr": 0.03251006816458619, "acc_norm": 0.6965174129353234, "acc_norm_stderr": 0.03251006816458619 }, "custom|mmlu_mc:us_foreign_policy|0": { "acc": 0.75, "acc_stderr": 0.04351941398892446, "acc_norm": 0.75, "acc_norm_stderr": 0.04351941398892446 }, "custom|mmlu_mc:virology|0": { "acc": 0.45180722891566266, "acc_stderr": 0.038743715565879536, "acc_norm": 0.45180722891566266, "acc_norm_stderr": 0.038743715565879536 }, "custom|mmlu_mc:world_religions|0": { "acc": 0.5906432748538012, "acc_stderr": 0.03771283107626545, "acc_norm": 0.5906432748538012, "acc_norm_stderr": 0.03771283107626545 }, "custom|mmlu_pro_cloze|0": { "acc": 0.11269946808510638, "acc_stderr": 0.0028830052993390953, "acc_norm": 0.10347406914893617, "acc_norm_stderr": 0.002776811047097138 }, "custom|openbookqa|0": { "acc": 0.224, "acc_stderr": 0.018663994464710797, "acc_norm": 0.378, "acc_norm_stderr": 0.021706550824518184 }, "custom|piqa|0": { "acc": 0.7034820457018498, "acc_stderr": 0.010656078922661148, "acc_norm": 0.6980413492927094, "acc_norm_stderr": 0.010711732891588341 }, "custom|siqa|0": { "acc": 0.40276356192425794, "acc_stderr": 0.011098061143371352, "acc_norm": 0.4329580348004094, "acc_norm_stderr": 0.011211904262208605 }, "custom|winogrande|0": { "acc": 0.5461720599842147, "acc_stderr": 0.013992441563707063, "acc_norm": 0.5374901341752171, "acc_norm_stderr": 0.014012928183336574 }, "custom|gsm8k|5": { "qem": 0.34420015163002277, "qem_stderr": 0.013086800426693784 }, "custom|trivia_qa|0": { "qem": 0.04358002674988854, "qem_stderr": 0.0015241248224145741 }, "custom|arc:_average|0": { "acc": 0.4683065007297089, "acc_stderr": 0.011568940327644068, "acc_norm": 0.45081359671803356, "acc_norm_stderr": 0.011882924187218855 }, "custom|mmlu_cloze:_average|0": { "acc": 0.31630907240132794, "acc_stderr": 0.03422335974354886, "acc_norm": 0.3374225197619436, "acc_norm_stderr": 0.034973693369716986 }, "custom|mmlu_mc:_average|0": { "acc": 0.4896716170891279, "acc_stderr": 0.03621668643274436, "acc_norm": 0.4896716170891279, "acc_norm_stderr": 0.03621668643274436 }, "all": { "acc": 0.402939204802935, "acc_stderr": 0.03344252550939551, "acc_norm": 0.4148286079656159, "acc_norm_stderr": 0.03382380692637996, "qem": 0.19389008918995565, "qem_stderr": 0.007305462624554179 } }, "versions": { "custom|arc:challenge|0": 0, "custom|arc:easy|0": 0, "custom|commonsense_qa|0": 0, "custom|gsm8k|5": 0, "custom|hellaswag|0": 0, "custom|mmlu_cloze:abstract_algebra|0": 0, "custom|mmlu_cloze:anatomy|0": 0, "custom|mmlu_cloze:astronomy|0": 0, "custom|mmlu_cloze:business_ethics|0": 0, "custom|mmlu_cloze:clinical_knowledge|0": 0, "custom|mmlu_cloze:college_biology|0": 0, "custom|mmlu_cloze:college_chemistry|0": 0, "custom|mmlu_cloze:college_computer_science|0": 0, "custom|mmlu_cloze:college_mathematics|0": 0, "custom|mmlu_cloze:college_medicine|0": 0, "custom|mmlu_cloze:college_physics|0": 0, "custom|mmlu_cloze:computer_security|0": 0, "custom|mmlu_cloze:conceptual_physics|0": 0, "custom|mmlu_cloze:econometrics|0": 0, "custom|mmlu_cloze:electrical_engineering|0": 0, "custom|mmlu_cloze:elementary_mathematics|0": 0, "custom|mmlu_cloze:formal_logic|0": 0, "custom|mmlu_cloze:global_facts|0": 0, "custom|mmlu_cloze:high_school_biology|0": 0, "custom|mmlu_cloze:high_school_chemistry|0": 0, "custom|mmlu_cloze:high_school_computer_science|0": 0, "custom|mmlu_cloze:high_school_european_history|0": 0, "custom|mmlu_cloze:high_school_geography|0": 0, "custom|mmlu_cloze:high_school_government_and_politics|0": 0, "custom|mmlu_cloze:high_school_macroeconomics|0": 0, "custom|mmlu_cloze:high_school_mathematics|0": 0, "custom|mmlu_cloze:high_school_microeconomics|0": 0, "custom|mmlu_cloze:high_school_physics|0": 0, "custom|mmlu_cloze:high_school_psychology|0": 0, "custom|mmlu_cloze:high_school_statistics|0": 0, "custom|mmlu_cloze:high_school_us_history|0": 0, "custom|mmlu_cloze:high_school_world_history|0": 0, "custom|mmlu_cloze:human_aging|0": 0, "custom|mmlu_cloze:human_sexuality|0": 0, "custom|mmlu_cloze:international_law|0": 0, "custom|mmlu_cloze:jurisprudence|0": 0, "custom|mmlu_cloze:logical_fallacies|0": 0, "custom|mmlu_cloze:machine_learning|0": 0, "custom|mmlu_cloze:management|0": 0, "custom|mmlu_cloze:marketing|0": 0, "custom|mmlu_cloze:medical_genetics|0": 0, "custom|mmlu_cloze:miscellaneous|0": 0, "custom|mmlu_cloze:moral_disputes|0": 0, "custom|mmlu_cloze:moral_scenarios|0": 0, "custom|mmlu_cloze:nutrition|0": 0, "custom|mmlu_cloze:philosophy|0": 0, "custom|mmlu_cloze:prehistory|0": 0, "custom|mmlu_cloze:professional_accounting|0": 0, "custom|mmlu_cloze:professional_law|0": 0, "custom|mmlu_cloze:professional_medicine|0": 0, "custom|mmlu_cloze:professional_psychology|0": 0, "custom|mmlu_cloze:public_relations|0": 0, "custom|mmlu_cloze:security_studies|0": 0, "custom|mmlu_cloze:sociology|0": 0, "custom|mmlu_cloze:us_foreign_policy|0": 0, "custom|mmlu_cloze:virology|0": 0, "custom|mmlu_cloze:world_religions|0": 0, "custom|mmlu_mc:abstract_algebra|0": 0, "custom|mmlu_mc:anatomy|0": 0, "custom|mmlu_mc:astronomy|0": 0, "custom|mmlu_mc:business_ethics|0": 0, "custom|mmlu_mc:clinical_knowledge|0": 0, "custom|mmlu_mc:college_biology|0": 0, "custom|mmlu_mc:college_chemistry|0": 0, "custom|mmlu_mc:college_computer_science|0": 0, "custom|mmlu_mc:college_mathematics|0": 0, "custom|mmlu_mc:college_medicine|0": 0, "custom|mmlu_mc:college_physics|0": 0, "custom|mmlu_mc:computer_security|0": 0, "custom|mmlu_mc:conceptual_physics|0": 0, "custom|mmlu_mc:econometrics|0": 0, "custom|mmlu_mc:electrical_engineering|0": 0, "custom|mmlu_mc:elementary_mathematics|0": 0, "custom|mmlu_mc:formal_logic|0": 0, "custom|mmlu_mc:global_facts|0": 0, "custom|mmlu_mc:high_school_biology|0": 0, "custom|mmlu_mc:high_school_chemistry|0": 0, "custom|mmlu_mc:high_school_computer_science|0": 0, "custom|mmlu_mc:high_school_european_history|0": 0, "custom|mmlu_mc:high_school_geography|0": 0, "custom|mmlu_mc:high_school_government_and_politics|0": 0, "custom|mmlu_mc:high_school_macroeconomics|0": 0, "custom|mmlu_mc:high_school_mathematics|0": 0, "custom|mmlu_mc:high_school_microeconomics|0": 0, "custom|mmlu_mc:high_school_physics|0": 0, "custom|mmlu_mc:high_school_psychology|0": 0, "custom|mmlu_mc:high_school_statistics|0": 0, "custom|mmlu_mc:high_school_us_history|0": 0, "custom|mmlu_mc:high_school_world_history|0": 0, "custom|mmlu_mc:human_aging|0": 0, "custom|mmlu_mc:human_sexuality|0": 0, "custom|mmlu_mc:international_law|0": 0, "custom|mmlu_mc:jurisprudence|0": 0, "custom|mmlu_mc:logical_fallacies|0": 0, "custom|mmlu_mc:machine_learning|0": 0, "custom|mmlu_mc:management|0": 0, "custom|mmlu_mc:marketing|0": 0, "custom|mmlu_mc:medical_genetics|0": 0, "custom|mmlu_mc:miscellaneous|0": 0, "custom|mmlu_mc:moral_disputes|0": 0, "custom|mmlu_mc:moral_scenarios|0": 0, "custom|mmlu_mc:nutrition|0": 0, "custom|mmlu_mc:philosophy|0": 0, "custom|mmlu_mc:prehistory|0": 0, "custom|mmlu_mc:professional_accounting|0": 0, "custom|mmlu_mc:professional_law|0": 0, "custom|mmlu_mc:professional_medicine|0": 0, "custom|mmlu_mc:professional_psychology|0": 0, "custom|mmlu_mc:public_relations|0": 0, "custom|mmlu_mc:security_studies|0": 0, "custom|mmlu_mc:sociology|0": 0, "custom|mmlu_mc:us_foreign_policy|0": 0, "custom|mmlu_mc:virology|0": 0, "custom|mmlu_mc:world_religions|0": 0, "custom|mmlu_pro_cloze|0": 0, "custom|openbookqa|0": 0, "custom|piqa|0": 0, "custom|siqa|0": 0, "custom|trivia_qa|0": 0, "custom|winogrande|0": 0 }, "config_tasks": { "custom|arc:challenge": { "name": "arc:challenge", "prompt_function": "arc", "hf_repo": "ai2_arc", "hf_subset": "ARC-Challenge", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": null, "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 1172, "effective_num_docs": 1172, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|arc:easy": { "name": "arc:easy", "prompt_function": "arc", "hf_repo": "ai2_arc", "hf_subset": "ARC-Easy", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": null, "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 2376, "effective_num_docs": 2376, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|commonsense_qa": { "name": "commonsense_qa", "prompt_function": "commonsense_qa_prompt", "hf_repo": "commonsense_qa", "hf_subset": "default", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "validation" ], "few_shots_split": null, "few_shots_select": null, "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 1221, "effective_num_docs": 1221, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|gsm8k": { "name": "gsm8k", "prompt_function": "gsm8k", "hf_repo": "gsm8k", "hf_subset": "main", "metric": [ "quasi_exact_match_gsm8k" ], "hf_avail_splits": [ "train", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": null, "few_shots_select": "random_sampling_from_train", "generation_size": 256, "stop_sequence": [ "Question:", "Question" ], "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 1319, "effective_num_docs": 1319, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|hellaswag": { "name": "hellaswag", "prompt_function": "hellaswag_prompt", "hf_repo": "hellaswag", "hf_subset": "default", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "validation" ], "few_shots_split": null, "few_shots_select": null, "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 10042, "effective_num_docs": 10042, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:abstract_algebra": { "name": "mmlu_cloze:abstract_algebra", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "abstract_algebra", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:anatomy": { "name": "mmlu_cloze:anatomy", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "anatomy", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 135, "effective_num_docs": 135, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:astronomy": { "name": "mmlu_cloze:astronomy", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "astronomy", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 152, "effective_num_docs": 152, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:business_ethics": { "name": "mmlu_cloze:business_ethics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "business_ethics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:clinical_knowledge": { "name": "mmlu_cloze:clinical_knowledge", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "clinical_knowledge", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 265, "effective_num_docs": 265, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:college_biology": { "name": "mmlu_cloze:college_biology", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "college_biology", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 144, "effective_num_docs": 144, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:college_chemistry": { "name": "mmlu_cloze:college_chemistry", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "college_chemistry", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:college_computer_science": { "name": "mmlu_cloze:college_computer_science", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "college_computer_science", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:college_mathematics": { "name": "mmlu_cloze:college_mathematics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "college_mathematics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:college_medicine": { "name": "mmlu_cloze:college_medicine", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "college_medicine", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 173, "effective_num_docs": 173, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:college_physics": { "name": "mmlu_cloze:college_physics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "college_physics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 102, "effective_num_docs": 102, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:computer_security": { "name": "mmlu_cloze:computer_security", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "computer_security", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:conceptual_physics": { "name": "mmlu_cloze:conceptual_physics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "conceptual_physics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 235, "effective_num_docs": 235, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:econometrics": { "name": "mmlu_cloze:econometrics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "econometrics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 114, "effective_num_docs": 114, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:electrical_engineering": { "name": "mmlu_cloze:electrical_engineering", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "electrical_engineering", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:elementary_mathematics": { "name": "mmlu_cloze:elementary_mathematics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "elementary_mathematics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 378, "effective_num_docs": 378, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:formal_logic": { "name": "mmlu_cloze:formal_logic", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "formal_logic", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 126, "effective_num_docs": 126, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:global_facts": { "name": "mmlu_cloze:global_facts", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "global_facts", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_biology": { "name": "mmlu_cloze:high_school_biology", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_biology", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 310, "effective_num_docs": 310, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_chemistry": { "name": "mmlu_cloze:high_school_chemistry", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_chemistry", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 203, "effective_num_docs": 203, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_computer_science": { "name": "mmlu_cloze:high_school_computer_science", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_computer_science", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_european_history": { "name": "mmlu_cloze:high_school_european_history", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_european_history", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 165, "effective_num_docs": 165, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_geography": { "name": "mmlu_cloze:high_school_geography", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_geography", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 198, "effective_num_docs": 198, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_government_and_politics": { "name": "mmlu_cloze:high_school_government_and_politics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_government_and_politics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 193, "effective_num_docs": 193, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_macroeconomics": { "name": "mmlu_cloze:high_school_macroeconomics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_macroeconomics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 390, "effective_num_docs": 390, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_mathematics": { "name": "mmlu_cloze:high_school_mathematics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_mathematics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 270, "effective_num_docs": 270, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_microeconomics": { "name": "mmlu_cloze:high_school_microeconomics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_microeconomics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 238, "effective_num_docs": 238, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_physics": { "name": "mmlu_cloze:high_school_physics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_physics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 151, "effective_num_docs": 151, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_psychology": { "name": "mmlu_cloze:high_school_psychology", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_psychology", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 545, "effective_num_docs": 545, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_statistics": { "name": "mmlu_cloze:high_school_statistics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_statistics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 216, "effective_num_docs": 216, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_us_history": { "name": "mmlu_cloze:high_school_us_history", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_us_history", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 204, "effective_num_docs": 204, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:high_school_world_history": { "name": "mmlu_cloze:high_school_world_history", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_world_history", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 237, "effective_num_docs": 237, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:human_aging": { "name": "mmlu_cloze:human_aging", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "human_aging", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 223, "effective_num_docs": 223, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:human_sexuality": { "name": "mmlu_cloze:human_sexuality", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "human_sexuality", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 131, "effective_num_docs": 131, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:international_law": { "name": "mmlu_cloze:international_law", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "international_law", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 121, "effective_num_docs": 121, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:jurisprudence": { "name": "mmlu_cloze:jurisprudence", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "jurisprudence", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 108, "effective_num_docs": 108, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:logical_fallacies": { "name": "mmlu_cloze:logical_fallacies", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "logical_fallacies", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 163, "effective_num_docs": 163, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:machine_learning": { "name": "mmlu_cloze:machine_learning", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "machine_learning", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 112, "effective_num_docs": 112, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:management": { "name": "mmlu_cloze:management", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "management", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 103, "effective_num_docs": 103, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:marketing": { "name": "mmlu_cloze:marketing", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "marketing", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 234, "effective_num_docs": 234, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:medical_genetics": { "name": "mmlu_cloze:medical_genetics", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "medical_genetics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:miscellaneous": { "name": "mmlu_cloze:miscellaneous", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "miscellaneous", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 783, "effective_num_docs": 783, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:moral_disputes": { "name": "mmlu_cloze:moral_disputes", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "moral_disputes", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 346, "effective_num_docs": 346, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:moral_scenarios": { "name": "mmlu_cloze:moral_scenarios", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "moral_scenarios", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 895, "effective_num_docs": 895, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:nutrition": { "name": "mmlu_cloze:nutrition", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "nutrition", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 306, "effective_num_docs": 306, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:philosophy": { "name": "mmlu_cloze:philosophy", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "philosophy", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 311, "effective_num_docs": 311, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:prehistory": { "name": "mmlu_cloze:prehistory", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "prehistory", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 324, "effective_num_docs": 324, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:professional_accounting": { "name": "mmlu_cloze:professional_accounting", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "professional_accounting", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 282, "effective_num_docs": 282, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:professional_law": { "name": "mmlu_cloze:professional_law", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "professional_law", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 1534, "effective_num_docs": 1534, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:professional_medicine": { "name": "mmlu_cloze:professional_medicine", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "professional_medicine", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 272, "effective_num_docs": 272, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:professional_psychology": { "name": "mmlu_cloze:professional_psychology", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "professional_psychology", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 612, "effective_num_docs": 612, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:public_relations": { "name": "mmlu_cloze:public_relations", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "public_relations", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 110, "effective_num_docs": 110, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:security_studies": { "name": "mmlu_cloze:security_studies", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "security_studies", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 245, "effective_num_docs": 245, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:sociology": { "name": "mmlu_cloze:sociology", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "sociology", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 201, "effective_num_docs": 201, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:us_foreign_policy": { "name": "mmlu_cloze:us_foreign_policy", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "us_foreign_policy", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:virology": { "name": "mmlu_cloze:virology", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "virology", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 166, "effective_num_docs": 166, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_cloze:world_religions": { "name": "mmlu_cloze:world_religions", "prompt_function": "mmlu_cloze_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "world_religions", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 171, "effective_num_docs": 171, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:abstract_algebra": { "name": "mmlu_mc:abstract_algebra", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "abstract_algebra", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:anatomy": { "name": "mmlu_mc:anatomy", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "anatomy", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 135, "effective_num_docs": 135, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:astronomy": { "name": "mmlu_mc:astronomy", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "astronomy", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 152, "effective_num_docs": 152, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:business_ethics": { "name": "mmlu_mc:business_ethics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "business_ethics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:clinical_knowledge": { "name": "mmlu_mc:clinical_knowledge", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "clinical_knowledge", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 265, "effective_num_docs": 265, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:college_biology": { "name": "mmlu_mc:college_biology", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "college_biology", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 144, "effective_num_docs": 144, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:college_chemistry": { "name": "mmlu_mc:college_chemistry", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "college_chemistry", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:college_computer_science": { "name": "mmlu_mc:college_computer_science", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "college_computer_science", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:college_mathematics": { "name": "mmlu_mc:college_mathematics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "college_mathematics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:college_medicine": { "name": "mmlu_mc:college_medicine", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "college_medicine", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 173, "effective_num_docs": 173, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:college_physics": { "name": "mmlu_mc:college_physics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "college_physics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 102, "effective_num_docs": 102, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:computer_security": { "name": "mmlu_mc:computer_security", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "computer_security", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:conceptual_physics": { "name": "mmlu_mc:conceptual_physics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "conceptual_physics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 235, "effective_num_docs": 235, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:econometrics": { "name": "mmlu_mc:econometrics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "econometrics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 114, "effective_num_docs": 114, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:electrical_engineering": { "name": "mmlu_mc:electrical_engineering", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "electrical_engineering", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:elementary_mathematics": { "name": "mmlu_mc:elementary_mathematics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "elementary_mathematics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 378, "effective_num_docs": 378, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:formal_logic": { "name": "mmlu_mc:formal_logic", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "formal_logic", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 126, "effective_num_docs": 126, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:global_facts": { "name": "mmlu_mc:global_facts", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "global_facts", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_biology": { "name": "mmlu_mc:high_school_biology", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_biology", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 310, "effective_num_docs": 310, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_chemistry": { "name": "mmlu_mc:high_school_chemistry", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_chemistry", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 203, "effective_num_docs": 203, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_computer_science": { "name": "mmlu_mc:high_school_computer_science", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_computer_science", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_european_history": { "name": "mmlu_mc:high_school_european_history", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_european_history", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 165, "effective_num_docs": 165, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_geography": { "name": "mmlu_mc:high_school_geography", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_geography", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 198, "effective_num_docs": 198, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_government_and_politics": { "name": "mmlu_mc:high_school_government_and_politics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_government_and_politics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 193, "effective_num_docs": 193, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_macroeconomics": { "name": "mmlu_mc:high_school_macroeconomics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_macroeconomics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 390, "effective_num_docs": 390, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_mathematics": { "name": "mmlu_mc:high_school_mathematics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_mathematics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 270, "effective_num_docs": 270, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_microeconomics": { "name": "mmlu_mc:high_school_microeconomics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_microeconomics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 238, "effective_num_docs": 238, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_physics": { "name": "mmlu_mc:high_school_physics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_physics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 151, "effective_num_docs": 151, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_psychology": { "name": "mmlu_mc:high_school_psychology", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_psychology", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 545, "effective_num_docs": 545, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_statistics": { "name": "mmlu_mc:high_school_statistics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_statistics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 216, "effective_num_docs": 216, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_us_history": { "name": "mmlu_mc:high_school_us_history", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_us_history", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 204, "effective_num_docs": 204, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:high_school_world_history": { "name": "mmlu_mc:high_school_world_history", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "high_school_world_history", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 237, "effective_num_docs": 237, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:human_aging": { "name": "mmlu_mc:human_aging", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "human_aging", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 223, "effective_num_docs": 223, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:human_sexuality": { "name": "mmlu_mc:human_sexuality", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "human_sexuality", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 131, "effective_num_docs": 131, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:international_law": { "name": "mmlu_mc:international_law", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "international_law", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 121, "effective_num_docs": 121, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:jurisprudence": { "name": "mmlu_mc:jurisprudence", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "jurisprudence", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 108, "effective_num_docs": 108, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:logical_fallacies": { "name": "mmlu_mc:logical_fallacies", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "logical_fallacies", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 163, "effective_num_docs": 163, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:machine_learning": { "name": "mmlu_mc:machine_learning", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "machine_learning", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 112, "effective_num_docs": 112, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:management": { "name": "mmlu_mc:management", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "management", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 103, "effective_num_docs": 103, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:marketing": { "name": "mmlu_mc:marketing", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "marketing", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 234, "effective_num_docs": 234, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:medical_genetics": { "name": "mmlu_mc:medical_genetics", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "medical_genetics", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:miscellaneous": { "name": "mmlu_mc:miscellaneous", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "miscellaneous", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 783, "effective_num_docs": 783, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:moral_disputes": { "name": "mmlu_mc:moral_disputes", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "moral_disputes", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 346, "effective_num_docs": 346, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:moral_scenarios": { "name": "mmlu_mc:moral_scenarios", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "moral_scenarios", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 895, "effective_num_docs": 895, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:nutrition": { "name": "mmlu_mc:nutrition", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "nutrition", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 306, "effective_num_docs": 306, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:philosophy": { "name": "mmlu_mc:philosophy", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "philosophy", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 311, "effective_num_docs": 311, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:prehistory": { "name": "mmlu_mc:prehistory", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "prehistory", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 324, "effective_num_docs": 324, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:professional_accounting": { "name": "mmlu_mc:professional_accounting", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "professional_accounting", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 282, "effective_num_docs": 282, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:professional_law": { "name": "mmlu_mc:professional_law", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "professional_law", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 1534, "effective_num_docs": 1534, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:professional_medicine": { "name": "mmlu_mc:professional_medicine", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "professional_medicine", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 272, "effective_num_docs": 272, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:professional_psychology": { "name": "mmlu_mc:professional_psychology", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "professional_psychology", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 612, "effective_num_docs": 612, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:public_relations": { "name": "mmlu_mc:public_relations", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "public_relations", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 110, "effective_num_docs": 110, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:security_studies": { "name": "mmlu_mc:security_studies", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "security_studies", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 245, "effective_num_docs": 245, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:sociology": { "name": "mmlu_mc:sociology", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "sociology", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 201, "effective_num_docs": 201, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:us_foreign_policy": { "name": "mmlu_mc:us_foreign_policy", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "us_foreign_policy", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:virology": { "name": "mmlu_mc:virology", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "virology", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 166, "effective_num_docs": 166, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_mc:world_religions": { "name": "mmlu_mc:world_religions", "prompt_function": "mmlu_mc_prompt", "hf_repo": "lighteval/mmlu", "hf_subset": "world_religions", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": null, "generation_size": 1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 171, "effective_num_docs": 171, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|mmlu_pro_cloze": { "name": "mmlu_pro_cloze", "prompt_function": "mmlu_pro_cloze_prompt", "hf_repo": "TIGER-Lab/MMLU-Pro", "hf_subset": "default", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": null, "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 12032, "effective_num_docs": 12032, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|openbookqa": { "name": "openbookqa", "prompt_function": "openbookqa", "hf_repo": "openbookqa", "hf_subset": "main", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "validation" ], "few_shots_split": null, "few_shots_select": null, "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 500, "effective_num_docs": 500, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|piqa": { "name": "piqa", "prompt_function": "piqa_harness", "hf_repo": "piqa", "hf_subset": "plain_text", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "validation" ], "few_shots_split": null, "few_shots_select": null, "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 1838, "effective_num_docs": 1838, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|siqa": { "name": "siqa", "prompt_function": "siqa_prompt", "hf_repo": "lighteval/siqa", "hf_subset": "default", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation" ], "evaluation_splits": [ "validation" ], "few_shots_split": null, "few_shots_select": null, "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 1954, "effective_num_docs": 1954, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|trivia_qa": { "name": "trivia_qa", "prompt_function": "triviaqa", "hf_repo": "mandarjoshi/trivia_qa", "hf_subset": "rc.nocontext", "metric": [ "quasi_exact_match_triviaqa" ], "hf_avail_splits": [ "train", "validation" ], "evaluation_splits": [ "validation" ], "few_shots_split": null, "few_shots_select": "random_sampling_from_train", "generation_size": 20, "stop_sequence": [ "\n", ".", "," ], "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 17944, "effective_num_docs": 17944, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 }, "custom|winogrande": { "name": "winogrande", "prompt_function": "winogrande", "hf_repo": "winogrande", "hf_subset": "winogrande_xl", "metric": [ "loglikelihood_acc", "loglikelihood_acc_norm_nospace" ], "hf_avail_splits": [ "train", "validation", "test" ], "evaluation_splits": [ "validation" ], "few_shots_split": null, "few_shots_select": null, "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "custom" ], "original_num_docs": 1267, "effective_num_docs": 1267, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 } }, "summary_tasks": { "custom|arc:challenge|0": { "hashes": { "hash_examples": "17b0cae357c0259e", "hash_full_prompts": "17b0cae357c0259e", "hash_input_tokens": "1a80ba451389147f", "hash_cont_tokens": "125c49608499af5c" }, "truncated": 0, "non_truncated": 1172, "padded": 4687, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|arc:easy|0": { "hashes": { "hash_examples": "63703c3cdff55bec", "hash_full_prompts": "63703c3cdff55bec", "hash_input_tokens": "d9b56de7e99a5c9c", "hash_cont_tokens": "966c4235d9e2f244" }, "truncated": 0, "non_truncated": 2376, "padded": 9501, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|commonsense_qa|0": { "hashes": { "hash_examples": "2e514c541df5ae5b", "hash_full_prompts": "2e514c541df5ae5b", "hash_input_tokens": "3ab0acf7245fa7dd", "hash_cont_tokens": "c9ae8301e6aeb0f9" }, "truncated": 0, "non_truncated": 1221, "padded": 6105, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|hellaswag|0": { "hashes": { "hash_examples": "31985c805c3a737e", "hash_full_prompts": "31985c805c3a737e", "hash_input_tokens": "cb1b4fddc24d00e3", "hash_cont_tokens": "30c3f5fcc8e12310" }, "truncated": 0, "non_truncated": 10042, "padded": 39721, "non_padded": 447, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:abstract_algebra|0": { "hashes": { "hash_examples": "ff00c12a680621ba", "hash_full_prompts": "ff00c12a680621ba", "hash_input_tokens": "7bc65f0dcf4e8b66", "hash_cont_tokens": "3722ba329e325bae" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:anatomy|0": { "hashes": { "hash_examples": "7f9c9593991d6727", "hash_full_prompts": "7f9c9593991d6727", "hash_input_tokens": "8d4f684fbf885541", "hash_cont_tokens": "10cd8a3e806a8b3a" }, "truncated": 0, "non_truncated": 135, "padded": 540, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:astronomy|0": { "hashes": { "hash_examples": "ff5985a306787836", "hash_full_prompts": "ff5985a306787836", "hash_input_tokens": "21bc0aee62164a2f", "hash_cont_tokens": "3487ec01bb06dc28" }, "truncated": 0, "non_truncated": 152, "padded": 608, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:business_ethics|0": { "hashes": { "hash_examples": "e3fe02a23d08c2d0", "hash_full_prompts": "e3fe02a23d08c2d0", "hash_input_tokens": "dc911e2821af6c2e", "hash_cont_tokens": "e7e730f1afaf5702" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:clinical_knowledge|0": { "hashes": { "hash_examples": "9b962be8e1615cd5", "hash_full_prompts": "9b962be8e1615cd5", "hash_input_tokens": "17c60ca3f6378b46", "hash_cont_tokens": "355af8107d9fa230" }, "truncated": 0, "non_truncated": 265, "padded": 1060, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:college_biology|0": { "hashes": { "hash_examples": "17b4ca841de3a2a3", "hash_full_prompts": "17b4ca841de3a2a3", "hash_input_tokens": "89cef67c04d6f8b8", "hash_cont_tokens": "ee5ca8d6f749de60" }, "truncated": 0, "non_truncated": 144, "padded": 575, "non_padded": 1, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:college_chemistry|0": { "hashes": { "hash_examples": "babea49005fd8249", "hash_full_prompts": "babea49005fd8249", "hash_input_tokens": "2d0a4177310fdb6a", "hash_cont_tokens": "85143a0ca77c0961" }, "truncated": 0, "non_truncated": 100, "padded": 398, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:college_computer_science|0": { "hashes": { "hash_examples": "47f4fef1846c8914", "hash_full_prompts": "47f4fef1846c8914", "hash_input_tokens": "1d5c7fe78fea8ab5", "hash_cont_tokens": "8c8f12fb45715ee0" }, "truncated": 0, "non_truncated": 100, "padded": 396, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:college_mathematics|0": { "hashes": { "hash_examples": "4d3686d599963414", "hash_full_prompts": "4d3686d599963414", "hash_input_tokens": "c14f082c8202b9a8", "hash_cont_tokens": "d34f1764a6a4b2dd" }, "truncated": 0, "non_truncated": 100, "padded": 398, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:college_medicine|0": { "hashes": { "hash_examples": "7209619ecac6f235", "hash_full_prompts": "7209619ecac6f235", "hash_input_tokens": "3649c3d32ac146db", "hash_cont_tokens": "abd1043e1455917b" }, "truncated": 0, "non_truncated": 173, "padded": 689, "non_padded": 3, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:college_physics|0": { "hashes": { "hash_examples": "6131b6c60dd7f055", "hash_full_prompts": "6131b6c60dd7f055", "hash_input_tokens": "512a5f6aff1f12c5", "hash_cont_tokens": "d0c750fb2e9f4582" }, "truncated": 0, "non_truncated": 102, "padded": 408, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:computer_security|0": { "hashes": { "hash_examples": "99539c9a5bc98a59", "hash_full_prompts": "99539c9a5bc98a59", "hash_input_tokens": "56a9385e60e26c39", "hash_cont_tokens": "2fd9482cc71690a2" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:conceptual_physics|0": { "hashes": { "hash_examples": "4e15015839d00858", "hash_full_prompts": "4e15015839d00858", "hash_input_tokens": "d3d0394a32fbc2c2", "hash_cont_tokens": "e863d73d23651a55" }, "truncated": 0, "non_truncated": 235, "padded": 940, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:econometrics|0": { "hashes": { "hash_examples": "fba5c647465e89e0", "hash_full_prompts": "fba5c647465e89e0", "hash_input_tokens": "1204070a3bb022cd", "hash_cont_tokens": "222e26b3d18a3157" }, "truncated": 0, "non_truncated": 114, "padded": 456, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:electrical_engineering|0": { "hashes": { "hash_examples": "2db2ba0cb98cda51", "hash_full_prompts": "2db2ba0cb98cda51", "hash_input_tokens": "4b89f57ffb145417", "hash_cont_tokens": "5a473530ecc1a291" }, "truncated": 0, "non_truncated": 145, "padded": 580, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:elementary_mathematics|0": { "hashes": { "hash_examples": "f231cd5ae05742bb", "hash_full_prompts": "f231cd5ae05742bb", "hash_input_tokens": "0f4d9df5b33ee979", "hash_cont_tokens": "33a1bd8906c60cb2" }, "truncated": 0, "non_truncated": 378, "padded": 1512, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:formal_logic|0": { "hashes": { "hash_examples": "b69d9607d5da536e", "hash_full_prompts": "b69d9607d5da536e", "hash_input_tokens": "27e49106a6279f75", "hash_cont_tokens": "5c87eaf78a589994" }, "truncated": 0, "non_truncated": 126, "padded": 498, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:global_facts|0": { "hashes": { "hash_examples": "c9f53772e672f6bc", "hash_full_prompts": "c9f53772e672f6bc", "hash_input_tokens": "5a45da5dbbbb93a9", "hash_cont_tokens": "7cc2f5666bd2595c" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_biology|0": { "hashes": { "hash_examples": "455027cf6cdd02bc", "hash_full_prompts": "455027cf6cdd02bc", "hash_input_tokens": "0e24c436c50ee6e6", "hash_cont_tokens": "0914e9d87572c1d6" }, "truncated": 0, "non_truncated": 310, "padded": 1234, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_chemistry|0": { "hashes": { "hash_examples": "95d9caac9edbc34d", "hash_full_prompts": "95d9caac9edbc34d", "hash_input_tokens": "a65e6b13d286866d", "hash_cont_tokens": "51132184cc53b900" }, "truncated": 0, "non_truncated": 203, "padded": 805, "non_padded": 7, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_computer_science|0": { "hashes": { "hash_examples": "6e44706db3791e51", "hash_full_prompts": "6e44706db3791e51", "hash_input_tokens": "b574513e7b6ea3e8", "hash_cont_tokens": "982e06f31b257b82" }, "truncated": 0, "non_truncated": 100, "padded": 394, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_european_history|0": { "hashes": { "hash_examples": "9078fce41897117d", "hash_full_prompts": "9078fce41897117d", "hash_input_tokens": "0150f31461ec1aff", "hash_cont_tokens": "ca0e022c7c90bf56" }, "truncated": 0, "non_truncated": 165, "padded": 660, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_geography|0": { "hashes": { "hash_examples": "c68adcc34130a2e8", "hash_full_prompts": "c68adcc34130a2e8", "hash_input_tokens": "c5e53e62e82ebb10", "hash_cont_tokens": "ac8305f6d617f1a7" }, "truncated": 0, "non_truncated": 198, "padded": 792, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_government_and_politics|0": { "hashes": { "hash_examples": "6f839b19e49a0858", "hash_full_prompts": "6f839b19e49a0858", "hash_input_tokens": "3ebd70bc972c8010", "hash_cont_tokens": "b4882210283c8e78" }, "truncated": 0, "non_truncated": 193, "padded": 772, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_macroeconomics|0": { "hashes": { "hash_examples": "708a5c05e7adb220", "hash_full_prompts": "708a5c05e7adb220", "hash_input_tokens": "741536c55adc903c", "hash_cont_tokens": "17b436962516a829" }, "truncated": 0, "non_truncated": 390, "padded": 1557, "non_padded": 3, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_mathematics|0": { "hashes": { "hash_examples": "47c621dd61fd7790", "hash_full_prompts": "47c621dd61fd7790", "hash_input_tokens": "b0acd1dc9c8c3bd2", "hash_cont_tokens": "c21561924e89f4e6" }, "truncated": 0, "non_truncated": 270, "padded": 1073, "non_padded": 7, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_microeconomics|0": { "hashes": { "hash_examples": "3c467180d90f6371", "hash_full_prompts": "3c467180d90f6371", "hash_input_tokens": "853fae992ef7b3e8", "hash_cont_tokens": "a5152854b16dddd6" }, "truncated": 0, "non_truncated": 238, "padded": 952, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_physics|0": { "hashes": { "hash_examples": "89a598cdde43be79", "hash_full_prompts": "89a598cdde43be79", "hash_input_tokens": "dd352b7719edb42a", "hash_cont_tokens": "4453c4648b0d3fb1" }, "truncated": 0, "non_truncated": 151, "padded": 602, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_psychology|0": { "hashes": { "hash_examples": "d8a7bf3f17ec12d0", "hash_full_prompts": "d8a7bf3f17ec12d0", "hash_input_tokens": "807b35bff6e0d108", "hash_cont_tokens": "ae3f9889156cbf0e" }, "truncated": 0, "non_truncated": 545, "padded": 2172, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_statistics|0": { "hashes": { "hash_examples": "995374a6caaa97d6", "hash_full_prompts": "995374a6caaa97d6", "hash_input_tokens": "c3a380d2caaea673", "hash_cont_tokens": "b95ad2aa2091fee6" }, "truncated": 0, "non_truncated": 216, "padded": 860, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_us_history|0": { "hashes": { "hash_examples": "7893e9d07e34cb37", "hash_full_prompts": "7893e9d07e34cb37", "hash_input_tokens": "7fb9bbd0c5d6e638", "hash_cont_tokens": "117f507507b38ff1" }, "truncated": 0, "non_truncated": 204, "padded": 816, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:high_school_world_history|0": { "hashes": { "hash_examples": "48879684e37d1716", "hash_full_prompts": "48879684e37d1716", "hash_input_tokens": "4e508a697a83438f", "hash_cont_tokens": "feb8fd9bf63ce918" }, "truncated": 0, "non_truncated": 237, "padded": 948, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:human_aging|0": { "hashes": { "hash_examples": "afae8c53bd6e5f44", "hash_full_prompts": "afae8c53bd6e5f44", "hash_input_tokens": "94d8dc34e705a4bc", "hash_cont_tokens": "65e40764ff693b00" }, "truncated": 0, "non_truncated": 223, "padded": 892, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:human_sexuality|0": { "hashes": { "hash_examples": "9701f02004912a7a", "hash_full_prompts": "9701f02004912a7a", "hash_input_tokens": "d3d6425f4edfb085", "hash_cont_tokens": "109a6cc3d24afa3e" }, "truncated": 0, "non_truncated": 131, "padded": 523, "non_padded": 1, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:international_law|0": { "hashes": { "hash_examples": "47955196de2d2c7a", "hash_full_prompts": "47955196de2d2c7a", "hash_input_tokens": "bdd15dc02beb3279", "hash_cont_tokens": "f8883cac3f2ba152" }, "truncated": 0, "non_truncated": 121, "padded": 484, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:jurisprudence|0": { "hashes": { "hash_examples": "a992eac2b8ae8bc4", "hash_full_prompts": "a992eac2b8ae8bc4", "hash_input_tokens": "b9642a7eac40a64f", "hash_cont_tokens": "0a554afcae081ada" }, "truncated": 0, "non_truncated": 108, "padded": 431, "non_padded": 1, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:logical_fallacies|0": { "hashes": { "hash_examples": "b0d31ed08f699e6c", "hash_full_prompts": "b0d31ed08f699e6c", "hash_input_tokens": "dcfeba59590d2733", "hash_cont_tokens": "6b84656fb38c2583" }, "truncated": 0, "non_truncated": 163, "padded": 652, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:machine_learning|0": { "hashes": { "hash_examples": "dccdef2bae4461a6", "hash_full_prompts": "dccdef2bae4461a6", "hash_input_tokens": "a06e77c37d98d508", "hash_cont_tokens": "f40b185fb9843e87" }, "truncated": 0, "non_truncated": 112, "padded": 448, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:management|0": { "hashes": { "hash_examples": "f600be25303e1fe2", "hash_full_prompts": "f600be25303e1fe2", "hash_input_tokens": "f36b9f20d8b04d9a", "hash_cont_tokens": "620a0bcabf306e8d" }, "truncated": 0, "non_truncated": 103, "padded": 412, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:marketing|0": { "hashes": { "hash_examples": "1a0df3ae5e306669", "hash_full_prompts": "1a0df3ae5e306669", "hash_input_tokens": "1e3bbdc729f756af", "hash_cont_tokens": "c013289c86a92fdb" }, "truncated": 0, "non_truncated": 234, "padded": 936, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:medical_genetics|0": { "hashes": { "hash_examples": "eb87c9cfd9b7c760", "hash_full_prompts": "eb87c9cfd9b7c760", "hash_input_tokens": "f91941ebf0a7d639", "hash_cont_tokens": "78bc82424ef7b7d9" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:miscellaneous|0": { "hashes": { "hash_examples": "f88d724036ba03b7", "hash_full_prompts": "f88d724036ba03b7", "hash_input_tokens": "9e596d3b99c33d17", "hash_cont_tokens": "8b82b7ff69ce32b0" }, "truncated": 0, "non_truncated": 783, "padded": 3131, "non_padded": 1, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:moral_disputes|0": { "hashes": { "hash_examples": "79782c0823005e7b", "hash_full_prompts": "79782c0823005e7b", "hash_input_tokens": "c8a4c24b156743ed", "hash_cont_tokens": "9ed381140e06ea2d" }, "truncated": 0, "non_truncated": 346, "padded": 1384, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:moral_scenarios|0": { "hashes": { "hash_examples": "fe57fdd86442b483", "hash_full_prompts": "fe57fdd86442b483", "hash_input_tokens": "a62c968277cfc272", "hash_cont_tokens": "78d36b9c0004374b" }, "truncated": 0, "non_truncated": 895, "padded": 3529, "non_padded": 51, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:nutrition|0": { "hashes": { "hash_examples": "421f206f5957e90f", "hash_full_prompts": "421f206f5957e90f", "hash_input_tokens": "65e465f378fe3751", "hash_cont_tokens": "c8de353c5d0a45bf" }, "truncated": 0, "non_truncated": 306, "padded": 1224, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:philosophy|0": { "hashes": { "hash_examples": "c93073e383957fc4", "hash_full_prompts": "c93073e383957fc4", "hash_input_tokens": "a89121a0eb5adb24", "hash_cont_tokens": "f047085684a04f83" }, "truncated": 0, "non_truncated": 311, "padded": 1244, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:prehistory|0": { "hashes": { "hash_examples": "58ec03e20eae9f90", "hash_full_prompts": "58ec03e20eae9f90", "hash_input_tokens": "ae1de55871f9c6aa", "hash_cont_tokens": "d59b7264b52f62ff" }, "truncated": 0, "non_truncated": 324, "padded": 1296, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:professional_accounting|0": { "hashes": { "hash_examples": "13f7a6023a118512", "hash_full_prompts": "13f7a6023a118512", "hash_input_tokens": "46430b373dae77ef", "hash_cont_tokens": "d9f26b0e34bf44e3" }, "truncated": 0, "non_truncated": 282, "padded": 1120, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:professional_law|0": { "hashes": { "hash_examples": "8086d24f4d4e82f4", "hash_full_prompts": "8086d24f4d4e82f4", "hash_input_tokens": "ac86825d4f7d6a2a", "hash_cont_tokens": "cb5e4423deb3bc06" }, "truncated": 0, "non_truncated": 1534, "padded": 6125, "non_padded": 11, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:professional_medicine|0": { "hashes": { "hash_examples": "f0b30b4e786eaeea", "hash_full_prompts": "f0b30b4e786eaeea", "hash_input_tokens": "4a7dc1fa8f5d4fc5", "hash_cont_tokens": "2e9c6460db7cf240" }, "truncated": 0, "non_truncated": 272, "padded": 1074, "non_padded": 14, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:professional_psychology|0": { "hashes": { "hash_examples": "29fd2b4a194c28ea", "hash_full_prompts": "29fd2b4a194c28ea", "hash_input_tokens": "87b8c8b1f1e1770e", "hash_cont_tokens": "57f4ac137227a6a8" }, "truncated": 0, "non_truncated": 612, "padded": 2444, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:public_relations|0": { "hashes": { "hash_examples": "52a84bb75dd812eb", "hash_full_prompts": "52a84bb75dd812eb", "hash_input_tokens": "886d1bf73db9b3ad", "hash_cont_tokens": "d45581a5d3f14b94" }, "truncated": 0, "non_truncated": 110, "padded": 440, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:security_studies|0": { "hashes": { "hash_examples": "2469ceb06f350432", "hash_full_prompts": "2469ceb06f350432", "hash_input_tokens": "c96b3893187077e2", "hash_cont_tokens": "2f980d82506cfccc" }, "truncated": 0, "non_truncated": 245, "padded": 976, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:sociology|0": { "hashes": { "hash_examples": "d2060dedb3fc2bea", "hash_full_prompts": "d2060dedb3fc2bea", "hash_input_tokens": "0ad9c72646c9bdc0", "hash_cont_tokens": "b89d02c97203ff2c" }, "truncated": 0, "non_truncated": 201, "padded": 804, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:us_foreign_policy|0": { "hashes": { "hash_examples": "84e882e740d43f01", "hash_full_prompts": "84e882e740d43f01", "hash_input_tokens": "59382eb3d92d7114", "hash_cont_tokens": "e0cb5d016b2da923" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:virology|0": { "hashes": { "hash_examples": "0428d2d277aa56aa", "hash_full_prompts": "0428d2d277aa56aa", "hash_input_tokens": "2003ed1edec4f410", "hash_cont_tokens": "a9858490e8a50dcd" }, "truncated": 0, "non_truncated": 166, "padded": 664, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_cloze:world_religions|0": { "hashes": { "hash_examples": "7e8f045c67ba6ba1", "hash_full_prompts": "7e8f045c67ba6ba1", "hash_input_tokens": "93eb1cc09970d585", "hash_cont_tokens": "6d6de61b03a0f67a" }, "truncated": 0, "non_truncated": 171, "padded": 684, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:abstract_algebra|0": { "hashes": { "hash_examples": "4c76229e00c9c0e9", "hash_full_prompts": "4c76229e00c9c0e9", "hash_input_tokens": "1ca9d0b9f1b13c3a", "hash_cont_tokens": "c8b3e6990f277952" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:anatomy|0": { "hashes": { "hash_examples": "6a1f8104dccbd33b", "hash_full_prompts": "6a1f8104dccbd33b", "hash_input_tokens": "506a2eb81f0ebd41", "hash_cont_tokens": "731840296eb8b864" }, "truncated": 0, "non_truncated": 135, "padded": 536, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:astronomy|0": { "hashes": { "hash_examples": "1302effa3a76ce4c", "hash_full_prompts": "1302effa3a76ce4c", "hash_input_tokens": "240a675542607aaf", "hash_cont_tokens": "59c4eb6bfd38b563" }, "truncated": 0, "non_truncated": 152, "padded": 600, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:business_ethics|0": { "hashes": { "hash_examples": "03cb8bce5336419a", "hash_full_prompts": "03cb8bce5336419a", "hash_input_tokens": "f8ffe3485997e69a", "hash_cont_tokens": "46c3e9b03592b386" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:clinical_knowledge|0": { "hashes": { "hash_examples": "ffbb9c7b2be257f9", "hash_full_prompts": "ffbb9c7b2be257f9", "hash_input_tokens": "a7b0b622d4a5d914", "hash_cont_tokens": "259f4c696060c1c2" }, "truncated": 0, "non_truncated": 265, "padded": 1052, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:college_biology|0": { "hashes": { "hash_examples": "3ee77f176f38eb8e", "hash_full_prompts": "3ee77f176f38eb8e", "hash_input_tokens": "38e90f043b20153c", "hash_cont_tokens": "4f0931b58aa2150b" }, "truncated": 0, "non_truncated": 144, "padded": 572, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:college_chemistry|0": { "hashes": { "hash_examples": "ce61a69c46d47aeb", "hash_full_prompts": "ce61a69c46d47aeb", "hash_input_tokens": "2b1a539da4838613", "hash_cont_tokens": "e9c575e8f2402a43" }, "truncated": 0, "non_truncated": 100, "padded": 396, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:college_computer_science|0": { "hashes": { "hash_examples": "32805b52d7d5daab", "hash_full_prompts": "32805b52d7d5daab", "hash_input_tokens": "586a10c92c3d4aea", "hash_cont_tokens": "fb65ba94d74b2c93" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:college_mathematics|0": { "hashes": { "hash_examples": "55da1a0a0bd33722", "hash_full_prompts": "55da1a0a0bd33722", "hash_input_tokens": "f8efb2761ac7d0f5", "hash_cont_tokens": "554d506288cb6caa" }, "truncated": 0, "non_truncated": 100, "padded": 396, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:college_medicine|0": { "hashes": { "hash_examples": "c33e143163049176", "hash_full_prompts": "c33e143163049176", "hash_input_tokens": "398bab21a62403d5", "hash_cont_tokens": "29e325dea40fe4c9" }, "truncated": 0, "non_truncated": 173, "padded": 692, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:college_physics|0": { "hashes": { "hash_examples": "ebdab1cdb7e555df", "hash_full_prompts": "ebdab1cdb7e555df", "hash_input_tokens": "aafd604e6a831a90", "hash_cont_tokens": "9e5c29d773a36305" }, "truncated": 0, "non_truncated": 102, "padded": 408, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:computer_security|0": { "hashes": { "hash_examples": "a24fd7d08a560921", "hash_full_prompts": "a24fd7d08a560921", "hash_input_tokens": "096a58ef29190974", "hash_cont_tokens": "8c2b1820aabd2c3f" }, "truncated": 0, "non_truncated": 100, "padded": 392, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:conceptual_physics|0": { "hashes": { "hash_examples": "8300977a79386993", "hash_full_prompts": "8300977a79386993", "hash_input_tokens": "487be9bfcc22f8ae", "hash_cont_tokens": "cb35d879f9089d88" }, "truncated": 0, "non_truncated": 235, "padded": 940, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:econometrics|0": { "hashes": { "hash_examples": "ddde36788a04a46f", "hash_full_prompts": "ddde36788a04a46f", "hash_input_tokens": "b96822fdac48be2a", "hash_cont_tokens": "38202b0ccdc9ee23" }, "truncated": 0, "non_truncated": 114, "padded": 448, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:electrical_engineering|0": { "hashes": { "hash_examples": "acbc5def98c19b3f", "hash_full_prompts": "acbc5def98c19b3f", "hash_input_tokens": "9684d155e984c193", "hash_cont_tokens": "f086b291b3aa0628" }, "truncated": 0, "non_truncated": 145, "padded": 576, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:elementary_mathematics|0": { "hashes": { "hash_examples": "146e61d07497a9bd", "hash_full_prompts": "146e61d07497a9bd", "hash_input_tokens": "5f137cae1894870d", "hash_cont_tokens": "7c1880cb7ab80d0d" }, "truncated": 0, "non_truncated": 378, "padded": 1488, "non_padded": 24, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:formal_logic|0": { "hashes": { "hash_examples": "8635216e1909a03f", "hash_full_prompts": "8635216e1909a03f", "hash_input_tokens": "371b1b0087b418fd", "hash_cont_tokens": "105dfb5265994616" }, "truncated": 0, "non_truncated": 126, "padded": 488, "non_padded": 16, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:global_facts|0": { "hashes": { "hash_examples": "30b315aa6353ee47", "hash_full_prompts": "30b315aa6353ee47", "hash_input_tokens": "91a31641102aa67c", "hash_cont_tokens": "bc75e4dffef3dc0e" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_biology|0": { "hashes": { "hash_examples": "c9136373af2180de", "hash_full_prompts": "c9136373af2180de", "hash_input_tokens": "c1f34a414d19873c", "hash_cont_tokens": "d9b3ef17f027c57d" }, "truncated": 0, "non_truncated": 310, "padded": 1216, "non_padded": 24, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_chemistry|0": { "hashes": { "hash_examples": "b0661bfa1add6404", "hash_full_prompts": "b0661bfa1add6404", "hash_input_tokens": "d893f482ded311c3", "hash_cont_tokens": "4be369e6571e06dc" }, "truncated": 0, "non_truncated": 203, "padded": 808, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_computer_science|0": { "hashes": { "hash_examples": "80fc1d623a3d665f", "hash_full_prompts": "80fc1d623a3d665f", "hash_input_tokens": "aea357a0d7c513d7", "hash_cont_tokens": "35d4dba6f0826ba4" }, "truncated": 0, "non_truncated": 100, "padded": 396, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_european_history|0": { "hashes": { "hash_examples": "854da6e5af0fe1a1", "hash_full_prompts": "854da6e5af0fe1a1", "hash_input_tokens": "43be94f23fd292f0", "hash_cont_tokens": "63bd63bdc490b9ff" }, "truncated": 0, "non_truncated": 165, "padded": 660, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_geography|0": { "hashes": { "hash_examples": "7dc963c7acd19ad8", "hash_full_prompts": "7dc963c7acd19ad8", "hash_input_tokens": "9c4e6a2190b16f8e", "hash_cont_tokens": "999a32d098465441" }, "truncated": 0, "non_truncated": 198, "padded": 788, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_government_and_politics|0": { "hashes": { "hash_examples": "1f675dcdebc9758f", "hash_full_prompts": "1f675dcdebc9758f", "hash_input_tokens": "ac4efc981943c848", "hash_cont_tokens": "361410848e01f8ed" }, "truncated": 0, "non_truncated": 193, "padded": 768, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_macroeconomics|0": { "hashes": { "hash_examples": "2fb32cf2d80f0b35", "hash_full_prompts": "2fb32cf2d80f0b35", "hash_input_tokens": "9d540ff2a6625078", "hash_cont_tokens": "18f9ae57b2444806" }, "truncated": 0, "non_truncated": 390, "padded": 1540, "non_padded": 20, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_mathematics|0": { "hashes": { "hash_examples": "fd6646fdb5d58a1f", "hash_full_prompts": "fd6646fdb5d58a1f", "hash_input_tokens": "8c7e21028bd5f97a", "hash_cont_tokens": "a13496e646060699" }, "truncated": 0, "non_truncated": 270, "padded": 1064, "non_padded": 16, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_microeconomics|0": { "hashes": { "hash_examples": "2118f21f71d87d84", "hash_full_prompts": "2118f21f71d87d84", "hash_input_tokens": "705c23da5c2e02df", "hash_cont_tokens": "791a7a25f0571e59" }, "truncated": 0, "non_truncated": 238, "padded": 948, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_physics|0": { "hashes": { "hash_examples": "dc3ce06378548565", "hash_full_prompts": "dc3ce06378548565", "hash_input_tokens": "ab5d7ca7bb8f4db5", "hash_cont_tokens": "9677b0687811cf73" }, "truncated": 0, "non_truncated": 151, "padded": 596, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_psychology|0": { "hashes": { "hash_examples": "c8d1d98a40e11f2f", "hash_full_prompts": "c8d1d98a40e11f2f", "hash_input_tokens": "5214809e6ad94546", "hash_cont_tokens": "c6fb64e5edd7d73b" }, "truncated": 0, "non_truncated": 545, "padded": 2148, "non_padded": 32, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_statistics|0": { "hashes": { "hash_examples": "666c8759b98ee4ff", "hash_full_prompts": "666c8759b98ee4ff", "hash_input_tokens": "b760a504c3a0d4f1", "hash_cont_tokens": "42603a72bb112e10" }, "truncated": 0, "non_truncated": 216, "padded": 856, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_us_history|0": { "hashes": { "hash_examples": "95fef1c4b7d3f81e", "hash_full_prompts": "95fef1c4b7d3f81e", "hash_input_tokens": "213a2496b7720abf", "hash_cont_tokens": "6807381135d65b04" }, "truncated": 0, "non_truncated": 204, "padded": 816, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:high_school_world_history|0": { "hashes": { "hash_examples": "7e5085b6184b0322", "hash_full_prompts": "7e5085b6184b0322", "hash_input_tokens": "eeb45b715f3e193e", "hash_cont_tokens": "6c355e7b3498b719" }, "truncated": 0, "non_truncated": 237, "padded": 948, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:human_aging|0": { "hashes": { "hash_examples": "c17333e7c7c10797", "hash_full_prompts": "c17333e7c7c10797", "hash_input_tokens": "17894ad262d339b4", "hash_cont_tokens": "0f40704815d5b3f6" }, "truncated": 0, "non_truncated": 223, "padded": 892, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:human_sexuality|0": { "hashes": { "hash_examples": "4edd1e9045df5e3d", "hash_full_prompts": "4edd1e9045df5e3d", "hash_input_tokens": "23b648b53bca213e", "hash_cont_tokens": "a9fdf5917bdddc9b" }, "truncated": 0, "non_truncated": 131, "padded": 524, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:international_law|0": { "hashes": { "hash_examples": "db2fa00d771a062a", "hash_full_prompts": "db2fa00d771a062a", "hash_input_tokens": "0cc62fbf821e6244", "hash_cont_tokens": "c63e45a81fbe97b2" }, "truncated": 0, "non_truncated": 121, "padded": 476, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:jurisprudence|0": { "hashes": { "hash_examples": "e956f86b124076fe", "hash_full_prompts": "e956f86b124076fe", "hash_input_tokens": "1028c218c5c3c3a7", "hash_cont_tokens": "9df89edb95ea3c08" }, "truncated": 0, "non_truncated": 108, "padded": 428, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:logical_fallacies|0": { "hashes": { "hash_examples": "956e0e6365ab79f1", "hash_full_prompts": "956e0e6365ab79f1", "hash_input_tokens": "fcc96799bc03df65", "hash_cont_tokens": "5b4f21454680a984" }, "truncated": 0, "non_truncated": 163, "padded": 652, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:machine_learning|0": { "hashes": { "hash_examples": "397997cc6f4d581e", "hash_full_prompts": "397997cc6f4d581e", "hash_input_tokens": "c089cf31b2d7f01c", "hash_cont_tokens": "3137ed354284f0e0" }, "truncated": 0, "non_truncated": 112, "padded": 448, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:management|0": { "hashes": { "hash_examples": "2bcbe6f6ca63d740", "hash_full_prompts": "2bcbe6f6ca63d740", "hash_input_tokens": "559e19ff854815df", "hash_cont_tokens": "1279a23b3bc7b32c" }, "truncated": 0, "non_truncated": 103, "padded": 412, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:marketing|0": { "hashes": { "hash_examples": "8ddb20d964a1b065", "hash_full_prompts": "8ddb20d964a1b065", "hash_input_tokens": "4402ef89378afbe3", "hash_cont_tokens": "be76778b3b861344" }, "truncated": 0, "non_truncated": 234, "padded": 936, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:medical_genetics|0": { "hashes": { "hash_examples": "182a71f4763d2cea", "hash_full_prompts": "182a71f4763d2cea", "hash_input_tokens": "237012d95029840a", "hash_cont_tokens": "bc75e4dffef3dc0e" }, "truncated": 0, "non_truncated": 100, "padded": 396, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:miscellaneous|0": { "hashes": { "hash_examples": "4c404fdbb4ca57fc", "hash_full_prompts": "4c404fdbb4ca57fc", "hash_input_tokens": "e62914e414087464", "hash_cont_tokens": "656d10e3f49675ee" }, "truncated": 0, "non_truncated": 783, "padded": 3112, "non_padded": 20, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:moral_disputes|0": { "hashes": { "hash_examples": "60cbd2baa3fea5c9", "hash_full_prompts": "60cbd2baa3fea5c9", "hash_input_tokens": "d0ce9621b7026bbb", "hash_cont_tokens": "a208a34c74088f6c" }, "truncated": 0, "non_truncated": 346, "padded": 1364, "non_padded": 20, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:moral_scenarios|0": { "hashes": { "hash_examples": "fd8b0431fbdd75ef", "hash_full_prompts": "fd8b0431fbdd75ef", "hash_input_tokens": "81679b9f674ee598", "hash_cont_tokens": "996ce7a5b6c4aef1" }, "truncated": 0, "non_truncated": 895, "padded": 3580, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:nutrition|0": { "hashes": { "hash_examples": "71e55e2b829b6528", "hash_full_prompts": "71e55e2b829b6528", "hash_input_tokens": "514a8da9bd4f8f36", "hash_cont_tokens": "9d4280b06a73f2ad" }, "truncated": 0, "non_truncated": 306, "padded": 1224, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:philosophy|0": { "hashes": { "hash_examples": "a6d489a8d208fa4b", "hash_full_prompts": "a6d489a8d208fa4b", "hash_input_tokens": "cb7959f2144dc573", "hash_cont_tokens": "de293b0c21c8bae6" }, "truncated": 0, "non_truncated": 311, "padded": 1244, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:prehistory|0": { "hashes": { "hash_examples": "6cc50f032a19acaa", "hash_full_prompts": "6cc50f032a19acaa", "hash_input_tokens": "f9fb897f661d2859", "hash_cont_tokens": "ed0ff6b6c4caf978" }, "truncated": 0, "non_truncated": 324, "padded": 1284, "non_padded": 12, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:professional_accounting|0": { "hashes": { "hash_examples": "50f57ab32f5f6cea", "hash_full_prompts": "50f57ab32f5f6cea", "hash_input_tokens": "ea1343072ee9d1a9", "hash_cont_tokens": "83dd1489a0640048" }, "truncated": 0, "non_truncated": 282, "padded": 1108, "non_padded": 20, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:professional_law|0": { "hashes": { "hash_examples": "a8fdc85c64f4b215", "hash_full_prompts": "a8fdc85c64f4b215", "hash_input_tokens": "61048bdaf47a14e6", "hash_cont_tokens": "29b7a4e974f8d163" }, "truncated": 0, "non_truncated": 1534, "padded": 6136, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:professional_medicine|0": { "hashes": { "hash_examples": "c373a28a3050a73a", "hash_full_prompts": "c373a28a3050a73a", "hash_input_tokens": "3ae4d064e9a0790f", "hash_cont_tokens": "ad2cf95b8929c802" }, "truncated": 0, "non_truncated": 272, "padded": 1088, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:professional_psychology|0": { "hashes": { "hash_examples": "bf5254fe818356af", "hash_full_prompts": "bf5254fe818356af", "hash_input_tokens": "026f10b77b71a6f2", "hash_cont_tokens": "7c6318f8538bddcf" }, "truncated": 0, "non_truncated": 612, "padded": 2416, "non_padded": 32, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:public_relations|0": { "hashes": { "hash_examples": "b66d52e28e7d14e0", "hash_full_prompts": "b66d52e28e7d14e0", "hash_input_tokens": "225d550c82dd770c", "hash_cont_tokens": "1d5a43cfd6444ac8" }, "truncated": 0, "non_truncated": 110, "padded": 440, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:security_studies|0": { "hashes": { "hash_examples": "514c14feaf000ad9", "hash_full_prompts": "514c14feaf000ad9", "hash_input_tokens": "ae0c13a4688c6a19", "hash_cont_tokens": "42a0363271aa0e45" }, "truncated": 0, "non_truncated": 245, "padded": 972, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:sociology|0": { "hashes": { "hash_examples": "f6c9bc9d18c80870", "hash_full_prompts": "f6c9bc9d18c80870", "hash_input_tokens": "458d4766848b23c2", "hash_cont_tokens": "6dcd97a97f690bc4" }, "truncated": 0, "non_truncated": 201, "padded": 796, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:us_foreign_policy|0": { "hashes": { "hash_examples": "ed7b78629db6678f", "hash_full_prompts": "ed7b78629db6678f", "hash_input_tokens": "9dff9377e6bb7a8d", "hash_cont_tokens": "77f9e45302a70de3" }, "truncated": 0, "non_truncated": 100, "padded": 392, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:virology|0": { "hashes": { "hash_examples": "bc52ffdc3f9b994a", "hash_full_prompts": "bc52ffdc3f9b994a", "hash_input_tokens": "74365771e0547aab", "hash_cont_tokens": "324dfec5557d5fef" }, "truncated": 0, "non_truncated": 166, "padded": 660, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_mc:world_religions|0": { "hashes": { "hash_examples": "ecdb4a4f94f62930", "hash_full_prompts": "ecdb4a4f94f62930", "hash_input_tokens": "4b7921c3a1f2c272", "hash_cont_tokens": "488549eb5937c325" }, "truncated": 0, "non_truncated": 171, "padded": 684, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|mmlu_pro_cloze|0": { "hashes": { "hash_examples": "845e15cfeee1fc11", "hash_full_prompts": "845e15cfeee1fc11", "hash_input_tokens": "d97fad769396572c", "hash_cont_tokens": "cbb2b1dfdba77cc7" }, "truncated": 0, "non_truncated": 12032, "padded": 111203, "non_padded": 2792, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|openbookqa|0": { "hashes": { "hash_examples": "fd427af2ef0577e3", "hash_full_prompts": "fd427af2ef0577e3", "hash_input_tokens": "1c38e5461a17ffed", "hash_cont_tokens": "39a126805e6138fc" }, "truncated": 0, "non_truncated": 500, "padded": 1994, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|piqa|0": { "hashes": { "hash_examples": "f7e288a8894cd149", "hash_full_prompts": "f7e288a8894cd149", "hash_input_tokens": "17bff461520b9253", "hash_cont_tokens": "bb302a379c07dc69" }, "truncated": 0, "non_truncated": 1838, "padded": 3595, "non_padded": 81, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|siqa|0": { "hashes": { "hash_examples": "c62abc8ecbd49cc4", "hash_full_prompts": "c62abc8ecbd49cc4", "hash_input_tokens": "623f4a3d34d19fac", "hash_cont_tokens": "3ef06bdd7b8ff7a0" }, "truncated": 0, "non_truncated": 1954, "padded": 5777, "non_padded": 85, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|winogrande|0": { "hashes": { "hash_examples": "087d5d1a1afd4c7b", "hash_full_prompts": "087d5d1a1afd4c7b", "hash_input_tokens": "2073786b6c0bdc78", "hash_cont_tokens": "f5eb053d43f78da1" }, "truncated": 0, "non_truncated": 1267, "padded": 2531, "non_padded": 3, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "custom|gsm8k|5": { "hashes": { "hash_examples": "0ed016e24e7512fd", "hash_full_prompts": "41d55e83abc0e02d", "hash_input_tokens": "fc45e01856de5296", "hash_cont_tokens": "316a05937b0f88f2" }, "truncated": 1319, "non_truncated": 0, "padded": 955, "non_padded": 364, "effective_few_shots": 5.0, "num_truncated_few_shots": 0 }, "custom|trivia_qa|0": { "hashes": { "hash_examples": "1e083041cb75ff0c", "hash_full_prompts": "1e083041cb75ff0c", "hash_input_tokens": "c9340e17ef103ae3", "hash_cont_tokens": "5cce0ca4351baf81" }, "truncated": 17944, "non_truncated": 0, "padded": 283, "non_padded": 17661, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 } }, "summary_general": { "hashes": { "hash_examples": "88e809e91c59a89c", "hash_full_prompts": "60e948f5cd4b4b2e", "hash_input_tokens": "2817d822d799cf5b", "hash_cont_tokens": "86bbe5ffe0d31a92" }, "truncated": 19263, "non_truncated": 60486, "padded": 298164, "non_padded": 21963, "num_truncated_few_shots": 0 } }