mmlu-evals-smollm-360m / smollm2_360M_eval.json
loubnabnl's picture
loubnabnl HF staff
add json
7b607c2
{
"config_general": {
"lighteval_sha": "?",
"num_fewshot_seeds": 1,
"override_batch_size": 8,
"max_samples": null,
"job_id": "",
"start_time": 1379672.653174776,
"end_time": 1380902.32375736,
"total_evaluation_time_secondes": "1229.6705825841054",
"model_name": "HuggingFaceTB/SmolLM-360M-2T-decay",
"model_sha": "0ae3a7ab88864d2f8ad177dffba413ff005dacde",
"model_dtype": "torch.float32",
"model_size": "1.35 GB",
"config": null
},
"results": {
"custom|arc:challenge|0": {
"acc": 0.22184300341296928,
"acc_stderr": 0.012141659068147886,
"acc_norm": 0.24914675767918087,
"acc_norm_stderr": 0.012639407111926433
},
"custom|arc:easy|0": {
"acc": 0.5774410774410774,
"acc_stderr": 0.010135978222981082,
"acc_norm": 0.5244107744107744,
"acc_norm_stderr": 0.010247548905242274
},
"custom|commonsense_qa|0": {
"acc": 0.3063063063063063,
"acc_stderr": 0.013197203908506533,
"acc_norm": 0.3202293202293202,
"acc_norm_stderr": 0.01335770492627267
},
"custom|hellaswag|0": {
"acc": 0.2746464847639912,
"acc_stderr": 0.004454237797448343,
"acc_norm": 0.27912766381198967,
"acc_norm_stderr": 0.004476536569056587
},
"custom|mmlu_cloze:abstract_algebra|0": {
"acc": 0.19,
"acc_stderr": 0.03942772444036625,
"acc_norm": 0.17,
"acc_norm_stderr": 0.03775251680686371
},
"custom|mmlu_cloze:anatomy|0": {
"acc": 0.3037037037037037,
"acc_stderr": 0.03972552884785136,
"acc_norm": 0.3333333333333333,
"acc_norm_stderr": 0.04072314811876837
},
"custom|mmlu_cloze:astronomy|0": {
"acc": 0.23026315789473684,
"acc_stderr": 0.03426059424403165,
"acc_norm": 0.3026315789473684,
"acc_norm_stderr": 0.03738520676119667
},
"custom|mmlu_cloze:business_ethics|0": {
"acc": 0.45,
"acc_stderr": 0.05,
"acc_norm": 0.33,
"acc_norm_stderr": 0.04725815626252604
},
"custom|mmlu_cloze:clinical_knowledge|0": {
"acc": 0.2188679245283019,
"acc_stderr": 0.02544786382510863,
"acc_norm": 0.30566037735849055,
"acc_norm_stderr": 0.028353298073322663
},
"custom|mmlu_cloze:college_biology|0": {
"acc": 0.3263888888888889,
"acc_stderr": 0.03921067198982266,
"acc_norm": 0.2847222222222222,
"acc_norm_stderr": 0.03773809990686935
},
"custom|mmlu_cloze:college_chemistry|0": {
"acc": 0.26,
"acc_stderr": 0.044084400227680794,
"acc_norm": 0.25,
"acc_norm_stderr": 0.04351941398892446
},
"custom|mmlu_cloze:college_computer_science|0": {
"acc": 0.24,
"acc_stderr": 0.04292346959909283,
"acc_norm": 0.21,
"acc_norm_stderr": 0.040936018074033256
},
"custom|mmlu_cloze:college_mathematics|0": {
"acc": 0.17,
"acc_stderr": 0.0377525168068637,
"acc_norm": 0.19,
"acc_norm_stderr": 0.03942772444036622
},
"custom|mmlu_cloze:college_medicine|0": {
"acc": 0.24855491329479767,
"acc_stderr": 0.03295304696818318,
"acc_norm": 0.2774566473988439,
"acc_norm_stderr": 0.034140140070440354
},
"custom|mmlu_cloze:college_physics|0": {
"acc": 0.21568627450980393,
"acc_stderr": 0.04092563958237656,
"acc_norm": 0.21568627450980393,
"acc_norm_stderr": 0.04092563958237656
},
"custom|mmlu_cloze:computer_security|0": {
"acc": 0.25,
"acc_stderr": 0.04351941398892446,
"acc_norm": 0.27,
"acc_norm_stderr": 0.044619604333847394
},
"custom|mmlu_cloze:conceptual_physics|0": {
"acc": 0.30638297872340425,
"acc_stderr": 0.03013590647851756,
"acc_norm": 0.2765957446808511,
"acc_norm_stderr": 0.029241883869628817
},
"custom|mmlu_cloze:econometrics|0": {
"acc": 0.20175438596491227,
"acc_stderr": 0.037752050135836386,
"acc_norm": 0.22807017543859648,
"acc_norm_stderr": 0.03947152782669415
},
"custom|mmlu_cloze:electrical_engineering|0": {
"acc": 0.22758620689655173,
"acc_stderr": 0.03493950380131184,
"acc_norm": 0.27586206896551724,
"acc_norm_stderr": 0.03724563619774632
},
"custom|mmlu_cloze:elementary_mathematics|0": {
"acc": 0.2328042328042328,
"acc_stderr": 0.021765961672154537,
"acc_norm": 0.24074074074074073,
"acc_norm_stderr": 0.0220190800122179
},
"custom|mmlu_cloze:formal_logic|0": {
"acc": 0.2777777777777778,
"acc_stderr": 0.04006168083848876,
"acc_norm": 0.2698412698412698,
"acc_norm_stderr": 0.03970158273235171
},
"custom|mmlu_cloze:global_facts|0": {
"acc": 0.25,
"acc_stderr": 0.04351941398892446,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"custom|mmlu_cloze:high_school_biology|0": {
"acc": 0.24193548387096775,
"acc_stderr": 0.024362599693031103,
"acc_norm": 0.25806451612903225,
"acc_norm_stderr": 0.024892469172462833
},
"custom|mmlu_cloze:high_school_chemistry|0": {
"acc": 0.18719211822660098,
"acc_stderr": 0.027444924966882618,
"acc_norm": 0.20689655172413793,
"acc_norm_stderr": 0.028501378167893946
},
"custom|mmlu_cloze:high_school_computer_science|0": {
"acc": 0.24,
"acc_stderr": 0.04292346959909283,
"acc_norm": 0.29,
"acc_norm_stderr": 0.04560480215720684
},
"custom|mmlu_cloze:high_school_european_history|0": {
"acc": 0.18787878787878787,
"acc_stderr": 0.030501934059429144,
"acc_norm": 0.24848484848484848,
"acc_norm_stderr": 0.03374402644139404
},
"custom|mmlu_cloze:high_school_geography|0": {
"acc": 0.25252525252525254,
"acc_stderr": 0.030954055470365907,
"acc_norm": 0.30808080808080807,
"acc_norm_stderr": 0.03289477330098614
},
"custom|mmlu_cloze:high_school_government_and_politics|0": {
"acc": 0.2694300518134715,
"acc_stderr": 0.03201867122877794,
"acc_norm": 0.34196891191709844,
"acc_norm_stderr": 0.03423465100104284
},
"custom|mmlu_cloze:high_school_macroeconomics|0": {
"acc": 0.2641025641025641,
"acc_stderr": 0.022352193737453282,
"acc_norm": 0.26153846153846155,
"acc_norm_stderr": 0.022282141204204426
},
"custom|mmlu_cloze:high_school_mathematics|0": {
"acc": 0.15185185185185185,
"acc_stderr": 0.021881130957380465,
"acc_norm": 0.1814814814814815,
"acc_norm_stderr": 0.02349926466940728
},
"custom|mmlu_cloze:high_school_microeconomics|0": {
"acc": 0.18067226890756302,
"acc_stderr": 0.02499196496660075,
"acc_norm": 0.33613445378151263,
"acc_norm_stderr": 0.030684737115135356
},
"custom|mmlu_cloze:high_school_physics|0": {
"acc": 0.2251655629139073,
"acc_stderr": 0.03410435282008936,
"acc_norm": 0.23841059602649006,
"acc_norm_stderr": 0.0347918557259966
},
"custom|mmlu_cloze:high_school_psychology|0": {
"acc": 0.29908256880733947,
"acc_stderr": 0.019630417285415175,
"acc_norm": 0.29908256880733947,
"acc_norm_stderr": 0.019630417285415168
},
"custom|mmlu_cloze:high_school_statistics|0": {
"acc": 0.25,
"acc_stderr": 0.029531221160930918,
"acc_norm": 0.2916666666666667,
"acc_norm_stderr": 0.03099866630456053
},
"custom|mmlu_cloze:high_school_us_history|0": {
"acc": 0.20098039215686275,
"acc_stderr": 0.02812597226565437,
"acc_norm": 0.27941176470588236,
"acc_norm_stderr": 0.03149328104507957
},
"custom|mmlu_cloze:high_school_world_history|0": {
"acc": 0.189873417721519,
"acc_stderr": 0.025530100460233494,
"acc_norm": 0.2742616033755274,
"acc_norm_stderr": 0.029041333510598046
},
"custom|mmlu_cloze:human_aging|0": {
"acc": 0.34080717488789236,
"acc_stderr": 0.0318114974705536,
"acc_norm": 0.2914798206278027,
"acc_norm_stderr": 0.03050028317654591
},
"custom|mmlu_cloze:human_sexuality|0": {
"acc": 0.33587786259541985,
"acc_stderr": 0.041423137719966634,
"acc_norm": 0.2824427480916031,
"acc_norm_stderr": 0.03948406125768361
},
"custom|mmlu_cloze:international_law|0": {
"acc": 0.12396694214876033,
"acc_stderr": 0.030083098716035227,
"acc_norm": 0.21487603305785125,
"acc_norm_stderr": 0.03749492448709698
},
"custom|mmlu_cloze:jurisprudence|0": {
"acc": 0.19444444444444445,
"acc_stderr": 0.038260763248848646,
"acc_norm": 0.26851851851851855,
"acc_norm_stderr": 0.04284467968052191
},
"custom|mmlu_cloze:logical_fallacies|0": {
"acc": 0.2331288343558282,
"acc_stderr": 0.03322015795776741,
"acc_norm": 0.2822085889570552,
"acc_norm_stderr": 0.03536117886664743
},
"custom|mmlu_cloze:machine_learning|0": {
"acc": 0.33035714285714285,
"acc_stderr": 0.04464285714285714,
"acc_norm": 0.2767857142857143,
"acc_norm_stderr": 0.042466243366976256
},
"custom|mmlu_cloze:management|0": {
"acc": 0.2524271844660194,
"acc_stderr": 0.04301250399690879,
"acc_norm": 0.3106796116504854,
"acc_norm_stderr": 0.0458212416016155
},
"custom|mmlu_cloze:marketing|0": {
"acc": 0.3504273504273504,
"acc_stderr": 0.031256108244218796,
"acc_norm": 0.32905982905982906,
"acc_norm_stderr": 0.030782321577688163
},
"custom|mmlu_cloze:medical_genetics|0": {
"acc": 0.27,
"acc_stderr": 0.044619604333847394,
"acc_norm": 0.29,
"acc_norm_stderr": 0.045604802157206845
},
"custom|mmlu_cloze:miscellaneous|0": {
"acc": 0.31545338441890164,
"acc_stderr": 0.016617501738763394,
"acc_norm": 0.31928480204342274,
"acc_norm_stderr": 0.016671261749538736
},
"custom|mmlu_cloze:moral_disputes|0": {
"acc": 0.2514450867052023,
"acc_stderr": 0.023357365785874037,
"acc_norm": 0.21098265895953758,
"acc_norm_stderr": 0.021966309947043117
},
"custom|mmlu_cloze:moral_scenarios|0": {
"acc": 0.23798882681564246,
"acc_stderr": 0.014242630070574915,
"acc_norm": 0.2759776536312849,
"acc_norm_stderr": 0.014950103002475353
},
"custom|mmlu_cloze:nutrition|0": {
"acc": 0.21895424836601307,
"acc_stderr": 0.02367908986180772,
"acc_norm": 0.28431372549019607,
"acc_norm_stderr": 0.02582916327275748
},
"custom|mmlu_cloze:philosophy|0": {
"acc": 0.26366559485530544,
"acc_stderr": 0.02502553850053234,
"acc_norm": 0.2733118971061093,
"acc_norm_stderr": 0.02531176597542612
},
"custom|mmlu_cloze:prehistory|0": {
"acc": 0.30246913580246915,
"acc_stderr": 0.02555765398186806,
"acc_norm": 0.21604938271604937,
"acc_norm_stderr": 0.022899162918445796
},
"custom|mmlu_cloze:professional_accounting|0": {
"acc": 0.2695035460992908,
"acc_stderr": 0.026469036818590627,
"acc_norm": 0.2624113475177305,
"acc_norm_stderr": 0.026244920349843007
},
"custom|mmlu_cloze:professional_law|0": {
"acc": 0.23598435462842243,
"acc_stderr": 0.010844802669662684,
"acc_norm": 0.24837027379400262,
"acc_norm_stderr": 0.011035212598034505
},
"custom|mmlu_cloze:professional_medicine|0": {
"acc": 0.23161764705882354,
"acc_stderr": 0.025626533803777562,
"acc_norm": 0.27205882352941174,
"acc_norm_stderr": 0.027033041151681456
},
"custom|mmlu_cloze:professional_psychology|0": {
"acc": 0.23202614379084968,
"acc_stderr": 0.017077373377857002,
"acc_norm": 0.2696078431372549,
"acc_norm_stderr": 0.017952449196987862
},
"custom|mmlu_cloze:public_relations|0": {
"acc": 0.38181818181818183,
"acc_stderr": 0.046534298079135075,
"acc_norm": 0.20909090909090908,
"acc_norm_stderr": 0.0389509101572414
},
"custom|mmlu_cloze:security_studies|0": {
"acc": 0.3224489795918367,
"acc_stderr": 0.029923100563683913,
"acc_norm": 0.1836734693877551,
"acc_norm_stderr": 0.024789071332007615
},
"custom|mmlu_cloze:sociology|0": {
"acc": 0.23880597014925373,
"acc_stderr": 0.03014777593540922,
"acc_norm": 0.23880597014925373,
"acc_norm_stderr": 0.030147775935409224
},
"custom|mmlu_cloze:us_foreign_policy|0": {
"acc": 0.23,
"acc_stderr": 0.04229525846816506,
"acc_norm": 0.24,
"acc_norm_stderr": 0.04292346959909283
},
"custom|mmlu_cloze:virology|0": {
"acc": 0.23493975903614459,
"acc_stderr": 0.03300533186128922,
"acc_norm": 0.30120481927710846,
"acc_norm_stderr": 0.035716092300534796
},
"custom|mmlu_cloze:world_religions|0": {
"acc": 0.2222222222222222,
"acc_stderr": 0.03188578017686398,
"acc_norm": 0.27485380116959063,
"acc_norm_stderr": 0.03424042924691584
},
"custom|mmlu_mc:abstract_algebra|0": {
"acc": 0.22,
"acc_stderr": 0.04163331998932268,
"acc_norm": 0.22,
"acc_norm_stderr": 0.04163331998932268
},
"custom|mmlu_mc:anatomy|0": {
"acc": 0.17777777777777778,
"acc_stderr": 0.033027898599017176,
"acc_norm": 0.17777777777777778,
"acc_norm_stderr": 0.033027898599017176
},
"custom|mmlu_mc:astronomy|0": {
"acc": 0.18421052631578946,
"acc_stderr": 0.0315469804508223,
"acc_norm": 0.18421052631578946,
"acc_norm_stderr": 0.0315469804508223
},
"custom|mmlu_mc:business_ethics|0": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"custom|mmlu_mc:clinical_knowledge|0": {
"acc": 0.2037735849056604,
"acc_stderr": 0.024790784501775406,
"acc_norm": 0.2037735849056604,
"acc_norm_stderr": 0.024790784501775406
},
"custom|mmlu_mc:college_biology|0": {
"acc": 0.2361111111111111,
"acc_stderr": 0.03551446610810826,
"acc_norm": 0.2361111111111111,
"acc_norm_stderr": 0.03551446610810826
},
"custom|mmlu_mc:college_chemistry|0": {
"acc": 0.14,
"acc_stderr": 0.03487350880197771,
"acc_norm": 0.14,
"acc_norm_stderr": 0.03487350880197771
},
"custom|mmlu_mc:college_computer_science|0": {
"acc": 0.24,
"acc_stderr": 0.04292346959909283,
"acc_norm": 0.24,
"acc_norm_stderr": 0.04292346959909283
},
"custom|mmlu_mc:college_mathematics|0": {
"acc": 0.21,
"acc_stderr": 0.040936018074033256,
"acc_norm": 0.21,
"acc_norm_stderr": 0.040936018074033256
},
"custom|mmlu_mc:college_medicine|0": {
"acc": 0.20809248554913296,
"acc_stderr": 0.03095289021774988,
"acc_norm": 0.20809248554913296,
"acc_norm_stderr": 0.03095289021774988
},
"custom|mmlu_mc:college_physics|0": {
"acc": 0.21568627450980393,
"acc_stderr": 0.04092563958237654,
"acc_norm": 0.21568627450980393,
"acc_norm_stderr": 0.04092563958237654
},
"custom|mmlu_mc:computer_security|0": {
"acc": 0.29,
"acc_stderr": 0.045604802157206845,
"acc_norm": 0.29,
"acc_norm_stderr": 0.045604802157206845
},
"custom|mmlu_mc:conceptual_physics|0": {
"acc": 0.26382978723404255,
"acc_stderr": 0.028809989854102973,
"acc_norm": 0.26382978723404255,
"acc_norm_stderr": 0.028809989854102973
},
"custom|mmlu_mc:econometrics|0": {
"acc": 0.22807017543859648,
"acc_stderr": 0.03947152782669415,
"acc_norm": 0.22807017543859648,
"acc_norm_stderr": 0.03947152782669415
},
"custom|mmlu_mc:electrical_engineering|0": {
"acc": 0.25517241379310346,
"acc_stderr": 0.03632984052707842,
"acc_norm": 0.25517241379310346,
"acc_norm_stderr": 0.03632984052707842
},
"custom|mmlu_mc:elementary_mathematics|0": {
"acc": 0.20899470899470898,
"acc_stderr": 0.02094048156533486,
"acc_norm": 0.20899470899470898,
"acc_norm_stderr": 0.02094048156533486
},
"custom|mmlu_mc:formal_logic|0": {
"acc": 0.30158730158730157,
"acc_stderr": 0.04104947269903394,
"acc_norm": 0.30158730158730157,
"acc_norm_stderr": 0.04104947269903394
},
"custom|mmlu_mc:global_facts|0": {
"acc": 0.18,
"acc_stderr": 0.038612291966536934,
"acc_norm": 0.18,
"acc_norm_stderr": 0.038612291966536934
},
"custom|mmlu_mc:high_school_biology|0": {
"acc": 0.1870967741935484,
"acc_stderr": 0.022185710092252252,
"acc_norm": 0.1870967741935484,
"acc_norm_stderr": 0.022185710092252252
},
"custom|mmlu_mc:high_school_chemistry|0": {
"acc": 0.19704433497536947,
"acc_stderr": 0.02798672466673622,
"acc_norm": 0.19704433497536947,
"acc_norm_stderr": 0.02798672466673622
},
"custom|mmlu_mc:high_school_computer_science|0": {
"acc": 0.25,
"acc_stderr": 0.04351941398892446,
"acc_norm": 0.25,
"acc_norm_stderr": 0.04351941398892446
},
"custom|mmlu_mc:high_school_european_history|0": {
"acc": 0.23030303030303031,
"acc_stderr": 0.03287666758603488,
"acc_norm": 0.23030303030303031,
"acc_norm_stderr": 0.03287666758603488
},
"custom|mmlu_mc:high_school_geography|0": {
"acc": 0.17676767676767677,
"acc_stderr": 0.027178752639044915,
"acc_norm": 0.17676767676767677,
"acc_norm_stderr": 0.027178752639044915
},
"custom|mmlu_mc:high_school_government_and_politics|0": {
"acc": 0.20725388601036268,
"acc_stderr": 0.029252823291803613,
"acc_norm": 0.20725388601036268,
"acc_norm_stderr": 0.029252823291803613
},
"custom|mmlu_mc:high_school_macroeconomics|0": {
"acc": 0.19743589743589743,
"acc_stderr": 0.020182646968674844,
"acc_norm": 0.19743589743589743,
"acc_norm_stderr": 0.020182646968674844
},
"custom|mmlu_mc:high_school_mathematics|0": {
"acc": 0.2111111111111111,
"acc_stderr": 0.024882116857655078,
"acc_norm": 0.2111111111111111,
"acc_norm_stderr": 0.024882116857655078
},
"custom|mmlu_mc:high_school_microeconomics|0": {
"acc": 0.19747899159663865,
"acc_stderr": 0.025859164122051463,
"acc_norm": 0.19747899159663865,
"acc_norm_stderr": 0.025859164122051463
},
"custom|mmlu_mc:high_school_physics|0": {
"acc": 0.1986754966887417,
"acc_stderr": 0.03257847384436776,
"acc_norm": 0.1986754966887417,
"acc_norm_stderr": 0.03257847384436776
},
"custom|mmlu_mc:high_school_psychology|0": {
"acc": 0.1926605504587156,
"acc_stderr": 0.016909276884936094,
"acc_norm": 0.1926605504587156,
"acc_norm_stderr": 0.016909276884936094
},
"custom|mmlu_mc:high_school_statistics|0": {
"acc": 0.14814814814814814,
"acc_stderr": 0.024227629273728356,
"acc_norm": 0.14814814814814814,
"acc_norm_stderr": 0.024227629273728356
},
"custom|mmlu_mc:high_school_us_history|0": {
"acc": 0.2549019607843137,
"acc_stderr": 0.03058759135160425,
"acc_norm": 0.2549019607843137,
"acc_norm_stderr": 0.03058759135160425
},
"custom|mmlu_mc:high_school_world_history|0": {
"acc": 0.25316455696202533,
"acc_stderr": 0.028304657943035282,
"acc_norm": 0.25316455696202533,
"acc_norm_stderr": 0.028304657943035282
},
"custom|mmlu_mc:human_aging|0": {
"acc": 0.3094170403587444,
"acc_stderr": 0.031024411740572203,
"acc_norm": 0.3094170403587444,
"acc_norm_stderr": 0.031024411740572203
},
"custom|mmlu_mc:human_sexuality|0": {
"acc": 0.26717557251908397,
"acc_stderr": 0.038808483010823944,
"acc_norm": 0.26717557251908397,
"acc_norm_stderr": 0.038808483010823944
},
"custom|mmlu_mc:international_law|0": {
"acc": 0.2396694214876033,
"acc_stderr": 0.03896878985070417,
"acc_norm": 0.2396694214876033,
"acc_norm_stderr": 0.03896878985070417
},
"custom|mmlu_mc:jurisprudence|0": {
"acc": 0.26851851851851855,
"acc_stderr": 0.04284467968052192,
"acc_norm": 0.26851851851851855,
"acc_norm_stderr": 0.04284467968052192
},
"custom|mmlu_mc:logical_fallacies|0": {
"acc": 0.22085889570552147,
"acc_stderr": 0.032591773927421776,
"acc_norm": 0.22085889570552147,
"acc_norm_stderr": 0.032591773927421776
},
"custom|mmlu_mc:machine_learning|0": {
"acc": 0.32142857142857145,
"acc_stderr": 0.04432804055291519,
"acc_norm": 0.32142857142857145,
"acc_norm_stderr": 0.04432804055291519
},
"custom|mmlu_mc:management|0": {
"acc": 0.17475728155339806,
"acc_stderr": 0.037601780060266224,
"acc_norm": 0.17475728155339806,
"acc_norm_stderr": 0.037601780060266224
},
"custom|mmlu_mc:marketing|0": {
"acc": 0.2863247863247863,
"acc_stderr": 0.029614323690456645,
"acc_norm": 0.2863247863247863,
"acc_norm_stderr": 0.029614323690456645
},
"custom|mmlu_mc:medical_genetics|0": {
"acc": 0.29,
"acc_stderr": 0.045604802157206845,
"acc_norm": 0.29,
"acc_norm_stderr": 0.045604802157206845
},
"custom|mmlu_mc:miscellaneous|0": {
"acc": 0.23627075351213284,
"acc_stderr": 0.015190473717037497,
"acc_norm": 0.23627075351213284,
"acc_norm_stderr": 0.015190473717037497
},
"custom|mmlu_mc:moral_disputes|0": {
"acc": 0.2398843930635838,
"acc_stderr": 0.02298959254312357,
"acc_norm": 0.2398843930635838,
"acc_norm_stderr": 0.02298959254312357
},
"custom|mmlu_mc:moral_scenarios|0": {
"acc": 0.23798882681564246,
"acc_stderr": 0.014242630070574915,
"acc_norm": 0.23798882681564246,
"acc_norm_stderr": 0.014242630070574915
},
"custom|mmlu_mc:nutrition|0": {
"acc": 0.20915032679738563,
"acc_stderr": 0.02328768531233481,
"acc_norm": 0.20915032679738563,
"acc_norm_stderr": 0.02328768531233481
},
"custom|mmlu_mc:philosophy|0": {
"acc": 0.18971061093247588,
"acc_stderr": 0.022268196258783228,
"acc_norm": 0.18971061093247588,
"acc_norm_stderr": 0.022268196258783228
},
"custom|mmlu_mc:prehistory|0": {
"acc": 0.22530864197530864,
"acc_stderr": 0.02324620264781975,
"acc_norm": 0.22530864197530864,
"acc_norm_stderr": 0.02324620264781975
},
"custom|mmlu_mc:professional_accounting|0": {
"acc": 0.24468085106382978,
"acc_stderr": 0.02564555362226673,
"acc_norm": 0.24468085106382978,
"acc_norm_stderr": 0.02564555362226673
},
"custom|mmlu_mc:professional_law|0": {
"acc": 0.24445893089960888,
"acc_stderr": 0.010976425013113902,
"acc_norm": 0.24445893089960888,
"acc_norm_stderr": 0.010976425013113902
},
"custom|mmlu_mc:professional_medicine|0": {
"acc": 0.20220588235294118,
"acc_stderr": 0.024398192986654924,
"acc_norm": 0.20220588235294118,
"acc_norm_stderr": 0.024398192986654924
},
"custom|mmlu_mc:professional_psychology|0": {
"acc": 0.25326797385620914,
"acc_stderr": 0.017593486895366835,
"acc_norm": 0.25326797385620914,
"acc_norm_stderr": 0.017593486895366835
},
"custom|mmlu_mc:public_relations|0": {
"acc": 0.21818181818181817,
"acc_stderr": 0.03955932861795833,
"acc_norm": 0.21818181818181817,
"acc_norm_stderr": 0.03955932861795833
},
"custom|mmlu_mc:security_studies|0": {
"acc": 0.19183673469387755,
"acc_stderr": 0.025206963154225392,
"acc_norm": 0.19183673469387755,
"acc_norm_stderr": 0.025206963154225392
},
"custom|mmlu_mc:sociology|0": {
"acc": 0.24875621890547264,
"acc_stderr": 0.030567675938916707,
"acc_norm": 0.24875621890547264,
"acc_norm_stderr": 0.030567675938916707
},
"custom|mmlu_mc:us_foreign_policy|0": {
"acc": 0.3,
"acc_stderr": 0.046056618647183814,
"acc_norm": 0.3,
"acc_norm_stderr": 0.046056618647183814
},
"custom|mmlu_mc:virology|0": {
"acc": 0.2891566265060241,
"acc_stderr": 0.03529486801511115,
"acc_norm": 0.2891566265060241,
"acc_norm_stderr": 0.03529486801511115
},
"custom|mmlu_mc:world_religions|0": {
"acc": 0.3157894736842105,
"acc_stderr": 0.035650796707083106,
"acc_norm": 0.3157894736842105,
"acc_norm_stderr": 0.035650796707083106
},
"custom|mmlu_pro_cloze|0": {
"acc": 0.08477393617021277,
"acc_stderr": 0.002539478267781077,
"acc_norm": 0.07446808510638298,
"acc_norm_stderr": 0.0023934806969786856
},
"custom|openbookqa|0": {
"acc": 0.196,
"acc_stderr": 0.017770751227744856,
"acc_norm": 0.31,
"acc_norm_stderr": 0.020704041021724805
},
"custom|piqa|0": {
"acc": 0.6164309031556039,
"acc_stderr": 0.011345128734116274,
"acc_norm": 0.6109902067464635,
"acc_norm_stderr": 0.011374774974447468
},
"custom|siqa|0": {
"acc": 0.3474923234390993,
"acc_stderr": 0.010774926057105855,
"acc_norm": 0.4022517911975435,
"acc_norm_stderr": 0.01109575895130807
},
"custom|winogrande|0": {
"acc": 0.5217048145224941,
"acc_stderr": 0.014039239216484633,
"acc_norm": 0.5090765588003157,
"acc_norm_stderr": 0.014050170094497707
},
"custom|gsm8k|5": {
"qem": 0.0,
"qem_stderr": 0.0
},
"custom|trivia_qa|0": {
"qem": 0.02067543468568881,
"qem_stderr": 0.001062290065972327
},
"custom|arc:_average|0": {
"acc": 0.39964204042702334,
"acc_stderr": 0.011138818645564484,
"acc_norm": 0.38677876604497763,
"acc_norm_stderr": 0.011443478008584354
},
"custom|mmlu_cloze:_average|0": {
"acc": 0.2519515926895652,
"acc_stderr": 0.03216451222160932,
"acc_norm": 0.26617775276148775,
"acc_norm_stderr": 0.03290929767916069
},
"custom|mmlu_mc:_average|0": {
"acc": 0.23140608269805896,
"acc_stderr": 0.03147532325432879,
"acc_norm": 0.23140608269805896,
"acc_norm_stderr": 0.03147532325432879
},
"all": {
"acc": 0.24957744996997022,
"acc_stderr": 0.03027535954982755,
"acc_norm": 0.2572518681233832,
"acc_norm_stderr": 0.030652543223254915,
"qem": 0.010337717342844405,
"qem_stderr": 0.0005311450329861635
}
},
"versions": {
"custom|arc:challenge|0": 0,
"custom|arc:easy|0": 0,
"custom|commonsense_qa|0": 0,
"custom|gsm8k|5": 0,
"custom|hellaswag|0": 0,
"custom|mmlu_cloze:abstract_algebra|0": 0,
"custom|mmlu_cloze:anatomy|0": 0,
"custom|mmlu_cloze:astronomy|0": 0,
"custom|mmlu_cloze:business_ethics|0": 0,
"custom|mmlu_cloze:clinical_knowledge|0": 0,
"custom|mmlu_cloze:college_biology|0": 0,
"custom|mmlu_cloze:college_chemistry|0": 0,
"custom|mmlu_cloze:college_computer_science|0": 0,
"custom|mmlu_cloze:college_mathematics|0": 0,
"custom|mmlu_cloze:college_medicine|0": 0,
"custom|mmlu_cloze:college_physics|0": 0,
"custom|mmlu_cloze:computer_security|0": 0,
"custom|mmlu_cloze:conceptual_physics|0": 0,
"custom|mmlu_cloze:econometrics|0": 0,
"custom|mmlu_cloze:electrical_engineering|0": 0,
"custom|mmlu_cloze:elementary_mathematics|0": 0,
"custom|mmlu_cloze:formal_logic|0": 0,
"custom|mmlu_cloze:global_facts|0": 0,
"custom|mmlu_cloze:high_school_biology|0": 0,
"custom|mmlu_cloze:high_school_chemistry|0": 0,
"custom|mmlu_cloze:high_school_computer_science|0": 0,
"custom|mmlu_cloze:high_school_european_history|0": 0,
"custom|mmlu_cloze:high_school_geography|0": 0,
"custom|mmlu_cloze:high_school_government_and_politics|0": 0,
"custom|mmlu_cloze:high_school_macroeconomics|0": 0,
"custom|mmlu_cloze:high_school_mathematics|0": 0,
"custom|mmlu_cloze:high_school_microeconomics|0": 0,
"custom|mmlu_cloze:high_school_physics|0": 0,
"custom|mmlu_cloze:high_school_psychology|0": 0,
"custom|mmlu_cloze:high_school_statistics|0": 0,
"custom|mmlu_cloze:high_school_us_history|0": 0,
"custom|mmlu_cloze:high_school_world_history|0": 0,
"custom|mmlu_cloze:human_aging|0": 0,
"custom|mmlu_cloze:human_sexuality|0": 0,
"custom|mmlu_cloze:international_law|0": 0,
"custom|mmlu_cloze:jurisprudence|0": 0,
"custom|mmlu_cloze:logical_fallacies|0": 0,
"custom|mmlu_cloze:machine_learning|0": 0,
"custom|mmlu_cloze:management|0": 0,
"custom|mmlu_cloze:marketing|0": 0,
"custom|mmlu_cloze:medical_genetics|0": 0,
"custom|mmlu_cloze:miscellaneous|0": 0,
"custom|mmlu_cloze:moral_disputes|0": 0,
"custom|mmlu_cloze:moral_scenarios|0": 0,
"custom|mmlu_cloze:nutrition|0": 0,
"custom|mmlu_cloze:philosophy|0": 0,
"custom|mmlu_cloze:prehistory|0": 0,
"custom|mmlu_cloze:professional_accounting|0": 0,
"custom|mmlu_cloze:professional_law|0": 0,
"custom|mmlu_cloze:professional_medicine|0": 0,
"custom|mmlu_cloze:professional_psychology|0": 0,
"custom|mmlu_cloze:public_relations|0": 0,
"custom|mmlu_cloze:security_studies|0": 0,
"custom|mmlu_cloze:sociology|0": 0,
"custom|mmlu_cloze:us_foreign_policy|0": 0,
"custom|mmlu_cloze:virology|0": 0,
"custom|mmlu_cloze:world_religions|0": 0,
"custom|mmlu_mc:abstract_algebra|0": 0,
"custom|mmlu_mc:anatomy|0": 0,
"custom|mmlu_mc:astronomy|0": 0,
"custom|mmlu_mc:business_ethics|0": 0,
"custom|mmlu_mc:clinical_knowledge|0": 0,
"custom|mmlu_mc:college_biology|0": 0,
"custom|mmlu_mc:college_chemistry|0": 0,
"custom|mmlu_mc:college_computer_science|0": 0,
"custom|mmlu_mc:college_mathematics|0": 0,
"custom|mmlu_mc:college_medicine|0": 0,
"custom|mmlu_mc:college_physics|0": 0,
"custom|mmlu_mc:computer_security|0": 0,
"custom|mmlu_mc:conceptual_physics|0": 0,
"custom|mmlu_mc:econometrics|0": 0,
"custom|mmlu_mc:electrical_engineering|0": 0,
"custom|mmlu_mc:elementary_mathematics|0": 0,
"custom|mmlu_mc:formal_logic|0": 0,
"custom|mmlu_mc:global_facts|0": 0,
"custom|mmlu_mc:high_school_biology|0": 0,
"custom|mmlu_mc:high_school_chemistry|0": 0,
"custom|mmlu_mc:high_school_computer_science|0": 0,
"custom|mmlu_mc:high_school_european_history|0": 0,
"custom|mmlu_mc:high_school_geography|0": 0,
"custom|mmlu_mc:high_school_government_and_politics|0": 0,
"custom|mmlu_mc:high_school_macroeconomics|0": 0,
"custom|mmlu_mc:high_school_mathematics|0": 0,
"custom|mmlu_mc:high_school_microeconomics|0": 0,
"custom|mmlu_mc:high_school_physics|0": 0,
"custom|mmlu_mc:high_school_psychology|0": 0,
"custom|mmlu_mc:high_school_statistics|0": 0,
"custom|mmlu_mc:high_school_us_history|0": 0,
"custom|mmlu_mc:high_school_world_history|0": 0,
"custom|mmlu_mc:human_aging|0": 0,
"custom|mmlu_mc:human_sexuality|0": 0,
"custom|mmlu_mc:international_law|0": 0,
"custom|mmlu_mc:jurisprudence|0": 0,
"custom|mmlu_mc:logical_fallacies|0": 0,
"custom|mmlu_mc:machine_learning|0": 0,
"custom|mmlu_mc:management|0": 0,
"custom|mmlu_mc:marketing|0": 0,
"custom|mmlu_mc:medical_genetics|0": 0,
"custom|mmlu_mc:miscellaneous|0": 0,
"custom|mmlu_mc:moral_disputes|0": 0,
"custom|mmlu_mc:moral_scenarios|0": 0,
"custom|mmlu_mc:nutrition|0": 0,
"custom|mmlu_mc:philosophy|0": 0,
"custom|mmlu_mc:prehistory|0": 0,
"custom|mmlu_mc:professional_accounting|0": 0,
"custom|mmlu_mc:professional_law|0": 0,
"custom|mmlu_mc:professional_medicine|0": 0,
"custom|mmlu_mc:professional_psychology|0": 0,
"custom|mmlu_mc:public_relations|0": 0,
"custom|mmlu_mc:security_studies|0": 0,
"custom|mmlu_mc:sociology|0": 0,
"custom|mmlu_mc:us_foreign_policy|0": 0,
"custom|mmlu_mc:virology|0": 0,
"custom|mmlu_mc:world_religions|0": 0,
"custom|mmlu_pro_cloze|0": 0,
"custom|openbookqa|0": 0,
"custom|piqa|0": 0,
"custom|siqa|0": 0,
"custom|trivia_qa|0": 0,
"custom|winogrande|0": 0
},
"config_tasks": {
"custom|arc:challenge": {
"name": "arc:challenge",
"prompt_function": "arc",
"hf_repo": "ai2_arc",
"hf_subset": "ARC-Challenge",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": null,
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 1172,
"effective_num_docs": 1172,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|arc:easy": {
"name": "arc:easy",
"prompt_function": "arc",
"hf_repo": "ai2_arc",
"hf_subset": "ARC-Easy",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": null,
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 2376,
"effective_num_docs": 2376,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|commonsense_qa": {
"name": "commonsense_qa",
"prompt_function": "commonsense_qa_prompt",
"hf_repo": "commonsense_qa",
"hf_subset": "default",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"validation"
],
"few_shots_split": null,
"few_shots_select": null,
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 1221,
"effective_num_docs": 1221,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|gsm8k": {
"name": "gsm8k",
"prompt_function": "gsm8k",
"hf_repo": "gsm8k",
"hf_subset": "main",
"metric": [
"quasi_exact_match_gsm8k"
],
"hf_avail_splits": [
"train",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": null,
"few_shots_select": "random_sampling_from_train",
"generation_size": 256,
"stop_sequence": [
"Question:",
"Question"
],
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 1319,
"effective_num_docs": 1319,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|hellaswag": {
"name": "hellaswag",
"prompt_function": "hellaswag_prompt",
"hf_repo": "hellaswag",
"hf_subset": "default",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"validation"
],
"few_shots_split": null,
"few_shots_select": null,
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 10042,
"effective_num_docs": 10042,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:abstract_algebra": {
"name": "mmlu_cloze:abstract_algebra",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "abstract_algebra",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:anatomy": {
"name": "mmlu_cloze:anatomy",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "anatomy",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 135,
"effective_num_docs": 135,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:astronomy": {
"name": "mmlu_cloze:astronomy",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "astronomy",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 152,
"effective_num_docs": 152,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:business_ethics": {
"name": "mmlu_cloze:business_ethics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "business_ethics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:clinical_knowledge": {
"name": "mmlu_cloze:clinical_knowledge",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "clinical_knowledge",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 265,
"effective_num_docs": 265,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:college_biology": {
"name": "mmlu_cloze:college_biology",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "college_biology",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 144,
"effective_num_docs": 144,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:college_chemistry": {
"name": "mmlu_cloze:college_chemistry",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "college_chemistry",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:college_computer_science": {
"name": "mmlu_cloze:college_computer_science",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "college_computer_science",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:college_mathematics": {
"name": "mmlu_cloze:college_mathematics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "college_mathematics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:college_medicine": {
"name": "mmlu_cloze:college_medicine",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "college_medicine",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 173,
"effective_num_docs": 173,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:college_physics": {
"name": "mmlu_cloze:college_physics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "college_physics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 102,
"effective_num_docs": 102,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:computer_security": {
"name": "mmlu_cloze:computer_security",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "computer_security",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:conceptual_physics": {
"name": "mmlu_cloze:conceptual_physics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "conceptual_physics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 235,
"effective_num_docs": 235,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:econometrics": {
"name": "mmlu_cloze:econometrics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "econometrics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 114,
"effective_num_docs": 114,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:electrical_engineering": {
"name": "mmlu_cloze:electrical_engineering",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "electrical_engineering",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 145,
"effective_num_docs": 145,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:elementary_mathematics": {
"name": "mmlu_cloze:elementary_mathematics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "elementary_mathematics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 378,
"effective_num_docs": 378,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:formal_logic": {
"name": "mmlu_cloze:formal_logic",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "formal_logic",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 126,
"effective_num_docs": 126,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:global_facts": {
"name": "mmlu_cloze:global_facts",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "global_facts",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_biology": {
"name": "mmlu_cloze:high_school_biology",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_biology",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 310,
"effective_num_docs": 310,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_chemistry": {
"name": "mmlu_cloze:high_school_chemistry",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_chemistry",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 203,
"effective_num_docs": 203,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_computer_science": {
"name": "mmlu_cloze:high_school_computer_science",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_computer_science",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_european_history": {
"name": "mmlu_cloze:high_school_european_history",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_european_history",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 165,
"effective_num_docs": 165,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_geography": {
"name": "mmlu_cloze:high_school_geography",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_geography",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 198,
"effective_num_docs": 198,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_government_and_politics": {
"name": "mmlu_cloze:high_school_government_and_politics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_government_and_politics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 193,
"effective_num_docs": 193,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_macroeconomics": {
"name": "mmlu_cloze:high_school_macroeconomics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_macroeconomics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 390,
"effective_num_docs": 390,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_mathematics": {
"name": "mmlu_cloze:high_school_mathematics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_mathematics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 270,
"effective_num_docs": 270,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_microeconomics": {
"name": "mmlu_cloze:high_school_microeconomics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_microeconomics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 238,
"effective_num_docs": 238,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_physics": {
"name": "mmlu_cloze:high_school_physics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_physics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 151,
"effective_num_docs": 151,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_psychology": {
"name": "mmlu_cloze:high_school_psychology",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_psychology",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 545,
"effective_num_docs": 545,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_statistics": {
"name": "mmlu_cloze:high_school_statistics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_statistics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 216,
"effective_num_docs": 216,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_us_history": {
"name": "mmlu_cloze:high_school_us_history",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_us_history",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 204,
"effective_num_docs": 204,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:high_school_world_history": {
"name": "mmlu_cloze:high_school_world_history",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_world_history",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 237,
"effective_num_docs": 237,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:human_aging": {
"name": "mmlu_cloze:human_aging",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "human_aging",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 223,
"effective_num_docs": 223,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:human_sexuality": {
"name": "mmlu_cloze:human_sexuality",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "human_sexuality",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 131,
"effective_num_docs": 131,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:international_law": {
"name": "mmlu_cloze:international_law",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "international_law",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 121,
"effective_num_docs": 121,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:jurisprudence": {
"name": "mmlu_cloze:jurisprudence",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "jurisprudence",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 108,
"effective_num_docs": 108,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:logical_fallacies": {
"name": "mmlu_cloze:logical_fallacies",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "logical_fallacies",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 163,
"effective_num_docs": 163,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:machine_learning": {
"name": "mmlu_cloze:machine_learning",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "machine_learning",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 112,
"effective_num_docs": 112,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:management": {
"name": "mmlu_cloze:management",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "management",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 103,
"effective_num_docs": 103,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:marketing": {
"name": "mmlu_cloze:marketing",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "marketing",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 234,
"effective_num_docs": 234,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:medical_genetics": {
"name": "mmlu_cloze:medical_genetics",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "medical_genetics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:miscellaneous": {
"name": "mmlu_cloze:miscellaneous",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "miscellaneous",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 783,
"effective_num_docs": 783,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:moral_disputes": {
"name": "mmlu_cloze:moral_disputes",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "moral_disputes",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 346,
"effective_num_docs": 346,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:moral_scenarios": {
"name": "mmlu_cloze:moral_scenarios",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "moral_scenarios",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 895,
"effective_num_docs": 895,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:nutrition": {
"name": "mmlu_cloze:nutrition",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "nutrition",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 306,
"effective_num_docs": 306,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:philosophy": {
"name": "mmlu_cloze:philosophy",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "philosophy",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 311,
"effective_num_docs": 311,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:prehistory": {
"name": "mmlu_cloze:prehistory",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "prehistory",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 324,
"effective_num_docs": 324,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:professional_accounting": {
"name": "mmlu_cloze:professional_accounting",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "professional_accounting",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 282,
"effective_num_docs": 282,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:professional_law": {
"name": "mmlu_cloze:professional_law",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "professional_law",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 1534,
"effective_num_docs": 1534,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:professional_medicine": {
"name": "mmlu_cloze:professional_medicine",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "professional_medicine",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 272,
"effective_num_docs": 272,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:professional_psychology": {
"name": "mmlu_cloze:professional_psychology",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "professional_psychology",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 612,
"effective_num_docs": 612,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:public_relations": {
"name": "mmlu_cloze:public_relations",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "public_relations",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 110,
"effective_num_docs": 110,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:security_studies": {
"name": "mmlu_cloze:security_studies",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "security_studies",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 245,
"effective_num_docs": 245,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:sociology": {
"name": "mmlu_cloze:sociology",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "sociology",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 201,
"effective_num_docs": 201,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:us_foreign_policy": {
"name": "mmlu_cloze:us_foreign_policy",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "us_foreign_policy",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:virology": {
"name": "mmlu_cloze:virology",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "virology",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 166,
"effective_num_docs": 166,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_cloze:world_religions": {
"name": "mmlu_cloze:world_religions",
"prompt_function": "mmlu_cloze_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "world_religions",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 171,
"effective_num_docs": 171,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:abstract_algebra": {
"name": "mmlu_mc:abstract_algebra",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "abstract_algebra",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:anatomy": {
"name": "mmlu_mc:anatomy",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "anatomy",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 135,
"effective_num_docs": 135,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:astronomy": {
"name": "mmlu_mc:astronomy",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "astronomy",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 152,
"effective_num_docs": 152,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:business_ethics": {
"name": "mmlu_mc:business_ethics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "business_ethics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:clinical_knowledge": {
"name": "mmlu_mc:clinical_knowledge",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "clinical_knowledge",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 265,
"effective_num_docs": 265,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:college_biology": {
"name": "mmlu_mc:college_biology",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "college_biology",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 144,
"effective_num_docs": 144,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:college_chemistry": {
"name": "mmlu_mc:college_chemistry",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "college_chemistry",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:college_computer_science": {
"name": "mmlu_mc:college_computer_science",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "college_computer_science",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:college_mathematics": {
"name": "mmlu_mc:college_mathematics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "college_mathematics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:college_medicine": {
"name": "mmlu_mc:college_medicine",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "college_medicine",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 173,
"effective_num_docs": 173,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:college_physics": {
"name": "mmlu_mc:college_physics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "college_physics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 102,
"effective_num_docs": 102,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:computer_security": {
"name": "mmlu_mc:computer_security",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "computer_security",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:conceptual_physics": {
"name": "mmlu_mc:conceptual_physics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "conceptual_physics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 235,
"effective_num_docs": 235,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:econometrics": {
"name": "mmlu_mc:econometrics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "econometrics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 114,
"effective_num_docs": 114,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:electrical_engineering": {
"name": "mmlu_mc:electrical_engineering",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "electrical_engineering",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 145,
"effective_num_docs": 145,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:elementary_mathematics": {
"name": "mmlu_mc:elementary_mathematics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "elementary_mathematics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 378,
"effective_num_docs": 378,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:formal_logic": {
"name": "mmlu_mc:formal_logic",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "formal_logic",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 126,
"effective_num_docs": 126,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:global_facts": {
"name": "mmlu_mc:global_facts",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "global_facts",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_biology": {
"name": "mmlu_mc:high_school_biology",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_biology",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 310,
"effective_num_docs": 310,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_chemistry": {
"name": "mmlu_mc:high_school_chemistry",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_chemistry",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 203,
"effective_num_docs": 203,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_computer_science": {
"name": "mmlu_mc:high_school_computer_science",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_computer_science",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_european_history": {
"name": "mmlu_mc:high_school_european_history",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_european_history",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 165,
"effective_num_docs": 165,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_geography": {
"name": "mmlu_mc:high_school_geography",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_geography",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 198,
"effective_num_docs": 198,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_government_and_politics": {
"name": "mmlu_mc:high_school_government_and_politics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_government_and_politics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 193,
"effective_num_docs": 193,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_macroeconomics": {
"name": "mmlu_mc:high_school_macroeconomics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_macroeconomics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 390,
"effective_num_docs": 390,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_mathematics": {
"name": "mmlu_mc:high_school_mathematics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_mathematics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 270,
"effective_num_docs": 270,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_microeconomics": {
"name": "mmlu_mc:high_school_microeconomics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_microeconomics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 238,
"effective_num_docs": 238,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_physics": {
"name": "mmlu_mc:high_school_physics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_physics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 151,
"effective_num_docs": 151,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_psychology": {
"name": "mmlu_mc:high_school_psychology",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_psychology",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 545,
"effective_num_docs": 545,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_statistics": {
"name": "mmlu_mc:high_school_statistics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_statistics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 216,
"effective_num_docs": 216,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_us_history": {
"name": "mmlu_mc:high_school_us_history",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_us_history",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 204,
"effective_num_docs": 204,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:high_school_world_history": {
"name": "mmlu_mc:high_school_world_history",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "high_school_world_history",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 237,
"effective_num_docs": 237,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:human_aging": {
"name": "mmlu_mc:human_aging",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "human_aging",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 223,
"effective_num_docs": 223,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:human_sexuality": {
"name": "mmlu_mc:human_sexuality",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "human_sexuality",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 131,
"effective_num_docs": 131,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:international_law": {
"name": "mmlu_mc:international_law",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "international_law",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 121,
"effective_num_docs": 121,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:jurisprudence": {
"name": "mmlu_mc:jurisprudence",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "jurisprudence",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 108,
"effective_num_docs": 108,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:logical_fallacies": {
"name": "mmlu_mc:logical_fallacies",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "logical_fallacies",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 163,
"effective_num_docs": 163,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:machine_learning": {
"name": "mmlu_mc:machine_learning",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "machine_learning",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 112,
"effective_num_docs": 112,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:management": {
"name": "mmlu_mc:management",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "management",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 103,
"effective_num_docs": 103,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:marketing": {
"name": "mmlu_mc:marketing",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "marketing",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 234,
"effective_num_docs": 234,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:medical_genetics": {
"name": "mmlu_mc:medical_genetics",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "medical_genetics",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:miscellaneous": {
"name": "mmlu_mc:miscellaneous",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "miscellaneous",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 783,
"effective_num_docs": 783,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:moral_disputes": {
"name": "mmlu_mc:moral_disputes",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "moral_disputes",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 346,
"effective_num_docs": 346,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:moral_scenarios": {
"name": "mmlu_mc:moral_scenarios",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "moral_scenarios",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 895,
"effective_num_docs": 895,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:nutrition": {
"name": "mmlu_mc:nutrition",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "nutrition",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 306,
"effective_num_docs": 306,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:philosophy": {
"name": "mmlu_mc:philosophy",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "philosophy",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 311,
"effective_num_docs": 311,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:prehistory": {
"name": "mmlu_mc:prehistory",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "prehistory",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 324,
"effective_num_docs": 324,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:professional_accounting": {
"name": "mmlu_mc:professional_accounting",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "professional_accounting",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 282,
"effective_num_docs": 282,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:professional_law": {
"name": "mmlu_mc:professional_law",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "professional_law",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 1534,
"effective_num_docs": 1534,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:professional_medicine": {
"name": "mmlu_mc:professional_medicine",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "professional_medicine",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 272,
"effective_num_docs": 272,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:professional_psychology": {
"name": "mmlu_mc:professional_psychology",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "professional_psychology",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 612,
"effective_num_docs": 612,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:public_relations": {
"name": "mmlu_mc:public_relations",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "public_relations",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 110,
"effective_num_docs": 110,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:security_studies": {
"name": "mmlu_mc:security_studies",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "security_studies",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 245,
"effective_num_docs": 245,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:sociology": {
"name": "mmlu_mc:sociology",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "sociology",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 201,
"effective_num_docs": 201,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:us_foreign_policy": {
"name": "mmlu_mc:us_foreign_policy",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "us_foreign_policy",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 100,
"effective_num_docs": 100,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:virology": {
"name": "mmlu_mc:virology",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "virology",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 166,
"effective_num_docs": 166,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_mc:world_religions": {
"name": "mmlu_mc:world_religions",
"prompt_function": "mmlu_mc_prompt",
"hf_repo": "lighteval/mmlu",
"hf_subset": "world_religions",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "dev",
"few_shots_select": null,
"generation_size": 1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 171,
"effective_num_docs": 171,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|mmlu_pro_cloze": {
"name": "mmlu_pro_cloze",
"prompt_function": "mmlu_pro_cloze_prompt",
"hf_repo": "TIGER-Lab/MMLU-Pro",
"hf_subset": "default",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"test"
],
"few_shots_split": "validation",
"few_shots_select": null,
"generation_size": -1,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 12032,
"effective_num_docs": 12032,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|openbookqa": {
"name": "openbookqa",
"prompt_function": "openbookqa",
"hf_repo": "openbookqa",
"hf_subset": "main",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"validation"
],
"few_shots_split": null,
"few_shots_select": null,
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 500,
"effective_num_docs": 500,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|piqa": {
"name": "piqa",
"prompt_function": "piqa_harness",
"hf_repo": "piqa",
"hf_subset": "plain_text",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"validation"
],
"few_shots_split": null,
"few_shots_select": null,
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 1838,
"effective_num_docs": 1838,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|siqa": {
"name": "siqa",
"prompt_function": "siqa_prompt",
"hf_repo": "lighteval/siqa",
"hf_subset": "default",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation"
],
"evaluation_splits": [
"validation"
],
"few_shots_split": null,
"few_shots_select": null,
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 1954,
"effective_num_docs": 1954,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|trivia_qa": {
"name": "trivia_qa",
"prompt_function": "triviaqa",
"hf_repo": "mandarjoshi/trivia_qa",
"hf_subset": "rc.nocontext",
"metric": [
"quasi_exact_match_triviaqa"
],
"hf_avail_splits": [
"train",
"validation"
],
"evaluation_splits": [
"validation"
],
"few_shots_split": null,
"few_shots_select": "random_sampling_from_train",
"generation_size": 20,
"stop_sequence": [
"\n",
".",
","
],
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 17944,
"effective_num_docs": 17944,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
},
"custom|winogrande": {
"name": "winogrande",
"prompt_function": "winogrande",
"hf_repo": "winogrande",
"hf_subset": "winogrande_xl",
"metric": [
"loglikelihood_acc",
"loglikelihood_acc_norm_nospace"
],
"hf_avail_splits": [
"train",
"validation",
"test"
],
"evaluation_splits": [
"validation"
],
"few_shots_split": null,
"few_shots_select": null,
"generation_size": null,
"stop_sequence": null,
"output_regex": null,
"num_samples": null,
"frozen": false,
"suite": [
"custom"
],
"original_num_docs": 1267,
"effective_num_docs": 1267,
"trust_dataset": true,
"must_remove_duplicate_docs": null,
"version": 0
}
},
"summary_tasks": {
"custom|arc:challenge|0": {
"hashes": {
"hash_examples": "17b0cae357c0259e",
"hash_full_prompts": "17b0cae357c0259e",
"hash_input_tokens": "4ecabb2d7f956267",
"hash_cont_tokens": "382807b255d8e654"
},
"truncated": 0,
"non_truncated": 1172,
"padded": 4687,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|arc:easy|0": {
"hashes": {
"hash_examples": "63703c3cdff55bec",
"hash_full_prompts": "63703c3cdff55bec",
"hash_input_tokens": "ae39a87709c09046",
"hash_cont_tokens": "e5bfd0dd9f21e228"
},
"truncated": 0,
"non_truncated": 2376,
"padded": 9501,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|commonsense_qa|0": {
"hashes": {
"hash_examples": "2e514c541df5ae5b",
"hash_full_prompts": "2e514c541df5ae5b",
"hash_input_tokens": "5ea2a6658ad43123",
"hash_cont_tokens": "d72bd296750fdb4d"
},
"truncated": 0,
"non_truncated": 1221,
"padded": 6105,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|hellaswag|0": {
"hashes": {
"hash_examples": "31985c805c3a737e",
"hash_full_prompts": "31985c805c3a737e",
"hash_input_tokens": "9481581cb2f9d3bb",
"hash_cont_tokens": "170d770fcde8ba33"
},
"truncated": 0,
"non_truncated": 10042,
"padded": 40168,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:abstract_algebra|0": {
"hashes": {
"hash_examples": "ff00c12a680621ba",
"hash_full_prompts": "ff00c12a680621ba",
"hash_input_tokens": "825a07b3f48f0cfb",
"hash_cont_tokens": "d0c19f27d7deb80f"
},
"truncated": 0,
"non_truncated": 100,
"padded": 400,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:anatomy|0": {
"hashes": {
"hash_examples": "7f9c9593991d6727",
"hash_full_prompts": "7f9c9593991d6727",
"hash_input_tokens": "d7d6de8abfc50ccc",
"hash_cont_tokens": "42187024d0c75d7d"
},
"truncated": 0,
"non_truncated": 135,
"padded": 540,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:astronomy|0": {
"hashes": {
"hash_examples": "ff5985a306787836",
"hash_full_prompts": "ff5985a306787836",
"hash_input_tokens": "6e1b17a96a4ecdec",
"hash_cont_tokens": "74bf19f47d798d4e"
},
"truncated": 0,
"non_truncated": 152,
"padded": 608,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:business_ethics|0": {
"hashes": {
"hash_examples": "e3fe02a23d08c2d0",
"hash_full_prompts": "e3fe02a23d08c2d0",
"hash_input_tokens": "31df81de8053a502",
"hash_cont_tokens": "3a9f7cebab894907"
},
"truncated": 0,
"non_truncated": 100,
"padded": 400,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:clinical_knowledge|0": {
"hashes": {
"hash_examples": "9b962be8e1615cd5",
"hash_full_prompts": "9b962be8e1615cd5",
"hash_input_tokens": "2a4bfffc23f19d96",
"hash_cont_tokens": "302ae0fc9cb8523d"
},
"truncated": 0,
"non_truncated": 265,
"padded": 1060,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:college_biology|0": {
"hashes": {
"hash_examples": "17b4ca841de3a2a3",
"hash_full_prompts": "17b4ca841de3a2a3",
"hash_input_tokens": "3d21280c8de2c7aa",
"hash_cont_tokens": "ef093a5883a6d392"
},
"truncated": 0,
"non_truncated": 144,
"padded": 576,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:college_chemistry|0": {
"hashes": {
"hash_examples": "babea49005fd8249",
"hash_full_prompts": "babea49005fd8249",
"hash_input_tokens": "955f58f5d70c94c7",
"hash_cont_tokens": "8efb5376b8e37c39"
},
"truncated": 0,
"non_truncated": 100,
"padded": 400,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:college_computer_science|0": {
"hashes": {
"hash_examples": "47f4fef1846c8914",
"hash_full_prompts": "47f4fef1846c8914",
"hash_input_tokens": "8179bf2028cd2f04",
"hash_cont_tokens": "c97234a8cd658381"
},
"truncated": 0,
"non_truncated": 100,
"padded": 400,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:college_mathematics|0": {
"hashes": {
"hash_examples": "4d3686d599963414",
"hash_full_prompts": "4d3686d599963414",
"hash_input_tokens": "c7d240d7cb23b145",
"hash_cont_tokens": "2ba03c3efef40c8a"
},
"truncated": 0,
"non_truncated": 100,
"padded": 400,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:college_medicine|0": {
"hashes": {
"hash_examples": "7209619ecac6f235",
"hash_full_prompts": "7209619ecac6f235",
"hash_input_tokens": "e35f70f368fc586a",
"hash_cont_tokens": "8ee57a8c6fb00b11"
},
"truncated": 0,
"non_truncated": 173,
"padded": 692,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:college_physics|0": {
"hashes": {
"hash_examples": "6131b6c60dd7f055",
"hash_full_prompts": "6131b6c60dd7f055",
"hash_input_tokens": "6c2075c9f928a079",
"hash_cont_tokens": "8af2e1696b293e0e"
},
"truncated": 0,
"non_truncated": 102,
"padded": 399,
"non_padded": 9,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:computer_security|0": {
"hashes": {
"hash_examples": "99539c9a5bc98a59",
"hash_full_prompts": "99539c9a5bc98a59",
"hash_input_tokens": "8550a8d350932163",
"hash_cont_tokens": "d4773943d95ea3a4"
},
"truncated": 0,
"non_truncated": 100,
"padded": 394,
"non_padded": 6,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:conceptual_physics|0": {
"hashes": {
"hash_examples": "4e15015839d00858",
"hash_full_prompts": "4e15015839d00858",
"hash_input_tokens": "173445d4b851c661",
"hash_cont_tokens": "c4bbee0471416c1d"
},
"truncated": 0,
"non_truncated": 235,
"padded": 920,
"non_padded": 20,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:econometrics|0": {
"hashes": {
"hash_examples": "fba5c647465e89e0",
"hash_full_prompts": "fba5c647465e89e0",
"hash_input_tokens": "df936adf2f851607",
"hash_cont_tokens": "7a458dddb4002455"
},
"truncated": 0,
"non_truncated": 114,
"padded": 446,
"non_padded": 10,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:electrical_engineering|0": {
"hashes": {
"hash_examples": "2db2ba0cb98cda51",
"hash_full_prompts": "2db2ba0cb98cda51",
"hash_input_tokens": "4b30e518ba867fe9",
"hash_cont_tokens": "c027a21314d9cb39"
},
"truncated": 0,
"non_truncated": 145,
"padded": 571,
"non_padded": 9,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:elementary_mathematics|0": {
"hashes": {
"hash_examples": "f231cd5ae05742bb",
"hash_full_prompts": "f231cd5ae05742bb",
"hash_input_tokens": "0126e32933ea886b",
"hash_cont_tokens": "ffbc160045a6c7c7"
},
"truncated": 0,
"non_truncated": 378,
"padded": 1467,
"non_padded": 45,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:formal_logic|0": {
"hashes": {
"hash_examples": "b69d9607d5da536e",
"hash_full_prompts": "b69d9607d5da536e",
"hash_input_tokens": "b5734fec1a3ed448",
"hash_cont_tokens": "f202160d6490d213"
},
"truncated": 0,
"non_truncated": 126,
"padded": 500,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:global_facts|0": {
"hashes": {
"hash_examples": "c9f53772e672f6bc",
"hash_full_prompts": "c9f53772e672f6bc",
"hash_input_tokens": "a152fd81b0305203",
"hash_cont_tokens": "173cc0f486d62689"
},
"truncated": 0,
"non_truncated": 100,
"padded": 377,
"non_padded": 23,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_biology|0": {
"hashes": {
"hash_examples": "455027cf6cdd02bc",
"hash_full_prompts": "455027cf6cdd02bc",
"hash_input_tokens": "22d57f0f3025de3b",
"hash_cont_tokens": "0e36fd7d5c0bc1c9"
},
"truncated": 0,
"non_truncated": 310,
"padded": 1204,
"non_padded": 36,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_chemistry|0": {
"hashes": {
"hash_examples": "95d9caac9edbc34d",
"hash_full_prompts": "95d9caac9edbc34d",
"hash_input_tokens": "6e49e3e72b047a3e",
"hash_cont_tokens": "98be610b601651a3"
},
"truncated": 0,
"non_truncated": 203,
"padded": 796,
"non_padded": 16,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_computer_science|0": {
"hashes": {
"hash_examples": "6e44706db3791e51",
"hash_full_prompts": "6e44706db3791e51",
"hash_input_tokens": "d434692dc3d29b06",
"hash_cont_tokens": "9ad1d237a36c27e2"
},
"truncated": 0,
"non_truncated": 100,
"padded": 398,
"non_padded": 2,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_european_history|0": {
"hashes": {
"hash_examples": "9078fce41897117d",
"hash_full_prompts": "9078fce41897117d",
"hash_input_tokens": "baa5c1ca8f6add95",
"hash_cont_tokens": "1c0bac822924721f"
},
"truncated": 0,
"non_truncated": 165,
"padded": 660,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_geography|0": {
"hashes": {
"hash_examples": "c68adcc34130a2e8",
"hash_full_prompts": "c68adcc34130a2e8",
"hash_input_tokens": "b3bd3f5acf3bb59b",
"hash_cont_tokens": "9c18c9389334971d"
},
"truncated": 0,
"non_truncated": 198,
"padded": 754,
"non_padded": 38,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_government_and_politics|0": {
"hashes": {
"hash_examples": "6f839b19e49a0858",
"hash_full_prompts": "6f839b19e49a0858",
"hash_input_tokens": "3612995eb429ca2a",
"hash_cont_tokens": "04fd8615a95849dc"
},
"truncated": 0,
"non_truncated": 193,
"padded": 729,
"non_padded": 43,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_macroeconomics|0": {
"hashes": {
"hash_examples": "708a5c05e7adb220",
"hash_full_prompts": "708a5c05e7adb220",
"hash_input_tokens": "6573e125588b4a17",
"hash_cont_tokens": "2aebe52941f95aab"
},
"truncated": 0,
"non_truncated": 390,
"padded": 1504,
"non_padded": 56,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_mathematics|0": {
"hashes": {
"hash_examples": "47c621dd61fd7790",
"hash_full_prompts": "47c621dd61fd7790",
"hash_input_tokens": "b5b130940d42765d",
"hash_cont_tokens": "c451e8276a516c39"
},
"truncated": 0,
"non_truncated": 270,
"padded": 1068,
"non_padded": 12,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_microeconomics|0": {
"hashes": {
"hash_examples": "3c467180d90f6371",
"hash_full_prompts": "3c467180d90f6371",
"hash_input_tokens": "ef39656ea8bf35f0",
"hash_cont_tokens": "d8563ad90149cae5"
},
"truncated": 0,
"non_truncated": 238,
"padded": 909,
"non_padded": 43,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_physics|0": {
"hashes": {
"hash_examples": "89a598cdde43be79",
"hash_full_prompts": "89a598cdde43be79",
"hash_input_tokens": "895ceec34881fcf5",
"hash_cont_tokens": "caff417b58be4e70"
},
"truncated": 0,
"non_truncated": 151,
"padded": 594,
"non_padded": 10,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_psychology|0": {
"hashes": {
"hash_examples": "d8a7bf3f17ec12d0",
"hash_full_prompts": "d8a7bf3f17ec12d0",
"hash_input_tokens": "3f486cd383cd35b5",
"hash_cont_tokens": "2b6a09b8b17fb0eb"
},
"truncated": 0,
"non_truncated": 545,
"padded": 2123,
"non_padded": 57,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_statistics|0": {
"hashes": {
"hash_examples": "995374a6caaa97d6",
"hash_full_prompts": "995374a6caaa97d6",
"hash_input_tokens": "879083119967753d",
"hash_cont_tokens": "0cc2b9ae8dabf046"
},
"truncated": 0,
"non_truncated": 216,
"padded": 859,
"non_padded": 5,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_us_history|0": {
"hashes": {
"hash_examples": "7893e9d07e34cb37",
"hash_full_prompts": "7893e9d07e34cb37",
"hash_input_tokens": "6a24701391e40ffc",
"hash_cont_tokens": "73cfbeaec4b3aeed"
},
"truncated": 0,
"non_truncated": 204,
"padded": 816,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:high_school_world_history|0": {
"hashes": {
"hash_examples": "48879684e37d1716",
"hash_full_prompts": "48879684e37d1716",
"hash_input_tokens": "3079604970b7d448",
"hash_cont_tokens": "f1cb4346d2b1e292"
},
"truncated": 0,
"non_truncated": 237,
"padded": 948,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:human_aging|0": {
"hashes": {
"hash_examples": "afae8c53bd6e5f44",
"hash_full_prompts": "afae8c53bd6e5f44",
"hash_input_tokens": "a1f16b4821195939",
"hash_cont_tokens": "8d13851bb150eaf9"
},
"truncated": 0,
"non_truncated": 223,
"padded": 856,
"non_padded": 36,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:human_sexuality|0": {
"hashes": {
"hash_examples": "9701f02004912a7a",
"hash_full_prompts": "9701f02004912a7a",
"hash_input_tokens": "832396976eba57e1",
"hash_cont_tokens": "985d7aabfc585133"
},
"truncated": 0,
"non_truncated": 131,
"padded": 507,
"non_padded": 17,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:international_law|0": {
"hashes": {
"hash_examples": "47955196de2d2c7a",
"hash_full_prompts": "47955196de2d2c7a",
"hash_input_tokens": "6466394a11f03388",
"hash_cont_tokens": "9c8c1fa7e47c4c90"
},
"truncated": 0,
"non_truncated": 121,
"padded": 470,
"non_padded": 14,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:jurisprudence|0": {
"hashes": {
"hash_examples": "a992eac2b8ae8bc4",
"hash_full_prompts": "a992eac2b8ae8bc4",
"hash_input_tokens": "51f18657103513e5",
"hash_cont_tokens": "4331097b412f784f"
},
"truncated": 0,
"non_truncated": 108,
"padded": 413,
"non_padded": 19,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:logical_fallacies|0": {
"hashes": {
"hash_examples": "b0d31ed08f699e6c",
"hash_full_prompts": "b0d31ed08f699e6c",
"hash_input_tokens": "4cbb4ee3c4329a41",
"hash_cont_tokens": "4743dde06acaab43"
},
"truncated": 0,
"non_truncated": 163,
"padded": 635,
"non_padded": 17,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:machine_learning|0": {
"hashes": {
"hash_examples": "dccdef2bae4461a6",
"hash_full_prompts": "dccdef2bae4461a6",
"hash_input_tokens": "27c5b064b051a030",
"hash_cont_tokens": "1f795e6e8a273ecc"
},
"truncated": 0,
"non_truncated": 112,
"padded": 444,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:management|0": {
"hashes": {
"hash_examples": "f600be25303e1fe2",
"hash_full_prompts": "f600be25303e1fe2",
"hash_input_tokens": "90d2cf3e33247cf8",
"hash_cont_tokens": "393a02f5ba8ed8f7"
},
"truncated": 0,
"non_truncated": 103,
"padded": 409,
"non_padded": 3,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:marketing|0": {
"hashes": {
"hash_examples": "1a0df3ae5e306669",
"hash_full_prompts": "1a0df3ae5e306669",
"hash_input_tokens": "a6ad9428f2066c2a",
"hash_cont_tokens": "54499c1050a42bd8"
},
"truncated": 0,
"non_truncated": 234,
"padded": 907,
"non_padded": 29,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:medical_genetics|0": {
"hashes": {
"hash_examples": "eb87c9cfd9b7c760",
"hash_full_prompts": "eb87c9cfd9b7c760",
"hash_input_tokens": "a2fcd6b69eb20a62",
"hash_cont_tokens": "e81320584373fcab"
},
"truncated": 0,
"non_truncated": 100,
"padded": 377,
"non_padded": 23,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:miscellaneous|0": {
"hashes": {
"hash_examples": "f88d724036ba03b7",
"hash_full_prompts": "f88d724036ba03b7",
"hash_input_tokens": "9daf4dd3472a7d1a",
"hash_cont_tokens": "aa1fb18c535ba1ba"
},
"truncated": 0,
"non_truncated": 783,
"padded": 3055,
"non_padded": 77,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:moral_disputes|0": {
"hashes": {
"hash_examples": "79782c0823005e7b",
"hash_full_prompts": "79782c0823005e7b",
"hash_input_tokens": "fbede36d813dc634",
"hash_cont_tokens": "bce3e5cbf2757e8a"
},
"truncated": 0,
"non_truncated": 346,
"padded": 1318,
"non_padded": 66,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:moral_scenarios|0": {
"hashes": {
"hash_examples": "fe57fdd86442b483",
"hash_full_prompts": "fe57fdd86442b483",
"hash_input_tokens": "31be0a0533a83d95",
"hash_cont_tokens": "c0b9acae579bdfb6"
},
"truncated": 0,
"non_truncated": 895,
"padded": 3580,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:nutrition|0": {
"hashes": {
"hash_examples": "421f206f5957e90f",
"hash_full_prompts": "421f206f5957e90f",
"hash_input_tokens": "84a45e53f623efbb",
"hash_cont_tokens": "5ea5cbf5680055ad"
},
"truncated": 0,
"non_truncated": 306,
"padded": 1183,
"non_padded": 41,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:philosophy|0": {
"hashes": {
"hash_examples": "c93073e383957fc4",
"hash_full_prompts": "c93073e383957fc4",
"hash_input_tokens": "e927bccac541dae4",
"hash_cont_tokens": "3bf84495cb489f2f"
},
"truncated": 0,
"non_truncated": 311,
"padded": 1203,
"non_padded": 41,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:prehistory|0": {
"hashes": {
"hash_examples": "58ec03e20eae9f90",
"hash_full_prompts": "58ec03e20eae9f90",
"hash_input_tokens": "5a30ecbbc910a94f",
"hash_cont_tokens": "c9ac7804f9ff842d"
},
"truncated": 0,
"non_truncated": 324,
"padded": 1236,
"non_padded": 60,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:professional_accounting|0": {
"hashes": {
"hash_examples": "13f7a6023a118512",
"hash_full_prompts": "13f7a6023a118512",
"hash_input_tokens": "b462d03ce0959e8d",
"hash_cont_tokens": "0d7b8e6abecb4fb3"
},
"truncated": 0,
"non_truncated": 282,
"padded": 1122,
"non_padded": 6,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:professional_law|0": {
"hashes": {
"hash_examples": "8086d24f4d4e82f4",
"hash_full_prompts": "8086d24f4d4e82f4",
"hash_input_tokens": "1c3e5eae5a47abf1",
"hash_cont_tokens": "c38cfeb8105b5e90"
},
"truncated": 0,
"non_truncated": 1534,
"padded": 6118,
"non_padded": 18,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:professional_medicine|0": {
"hashes": {
"hash_examples": "f0b30b4e786eaeea",
"hash_full_prompts": "f0b30b4e786eaeea",
"hash_input_tokens": "ee6d4d0eb63ed94b",
"hash_cont_tokens": "c23b43a1136d5986"
},
"truncated": 0,
"non_truncated": 272,
"padded": 1078,
"non_padded": 10,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:professional_psychology|0": {
"hashes": {
"hash_examples": "29fd2b4a194c28ea",
"hash_full_prompts": "29fd2b4a194c28ea",
"hash_input_tokens": "860dc95066e1e596",
"hash_cont_tokens": "c8827a0bd1403cb3"
},
"truncated": 0,
"non_truncated": 612,
"padded": 2357,
"non_padded": 91,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:public_relations|0": {
"hashes": {
"hash_examples": "52a84bb75dd812eb",
"hash_full_prompts": "52a84bb75dd812eb",
"hash_input_tokens": "38711d53009ea90d",
"hash_cont_tokens": "d224b48b243de3f4"
},
"truncated": 0,
"non_truncated": 110,
"padded": 424,
"non_padded": 16,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:security_studies|0": {
"hashes": {
"hash_examples": "2469ceb06f350432",
"hash_full_prompts": "2469ceb06f350432",
"hash_input_tokens": "4f35af93520ee61e",
"hash_cont_tokens": "fd2c2443a5d9e7ac"
},
"truncated": 0,
"non_truncated": 245,
"padded": 961,
"non_padded": 19,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:sociology|0": {
"hashes": {
"hash_examples": "d2060dedb3fc2bea",
"hash_full_prompts": "d2060dedb3fc2bea",
"hash_input_tokens": "ed7e51cb983572ad",
"hash_cont_tokens": "acc70513ba8b6ea2"
},
"truncated": 0,
"non_truncated": 201,
"padded": 759,
"non_padded": 45,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:us_foreign_policy|0": {
"hashes": {
"hash_examples": "84e882e740d43f01",
"hash_full_prompts": "84e882e740d43f01",
"hash_input_tokens": "268956b1012c5940",
"hash_cont_tokens": "45532e5b7e0b55e5"
},
"truncated": 0,
"non_truncated": 100,
"padded": 379,
"non_padded": 21,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:virology|0": {
"hashes": {
"hash_examples": "0428d2d277aa56aa",
"hash_full_prompts": "0428d2d277aa56aa",
"hash_input_tokens": "961409bd21ba248a",
"hash_cont_tokens": "8bc228ef11217e91"
},
"truncated": 0,
"non_truncated": 166,
"padded": 640,
"non_padded": 24,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_cloze:world_religions|0": {
"hashes": {
"hash_examples": "7e8f045c67ba6ba1",
"hash_full_prompts": "7e8f045c67ba6ba1",
"hash_input_tokens": "bfe903593483aa22",
"hash_cont_tokens": "bc5b5b53a2341fc3"
},
"truncated": 0,
"non_truncated": 171,
"padded": 665,
"non_padded": 19,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:abstract_algebra|0": {
"hashes": {
"hash_examples": "4c76229e00c9c0e9",
"hash_full_prompts": "4c76229e00c9c0e9",
"hash_input_tokens": "7755491d408f0b5c",
"hash_cont_tokens": "0bb4627988c70881"
},
"truncated": 0,
"non_truncated": 100,
"padded": 392,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:anatomy|0": {
"hashes": {
"hash_examples": "6a1f8104dccbd33b",
"hash_full_prompts": "6a1f8104dccbd33b",
"hash_input_tokens": "b95ec4c836880ab8",
"hash_cont_tokens": "fbfb34b68f4e80cc"
},
"truncated": 0,
"non_truncated": 135,
"padded": 536,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:astronomy|0": {
"hashes": {
"hash_examples": "1302effa3a76ce4c",
"hash_full_prompts": "1302effa3a76ce4c",
"hash_input_tokens": "0930c5fc835f8fc3",
"hash_cont_tokens": "f94a14f8612b0081"
},
"truncated": 0,
"non_truncated": 152,
"padded": 604,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:business_ethics|0": {
"hashes": {
"hash_examples": "03cb8bce5336419a",
"hash_full_prompts": "03cb8bce5336419a",
"hash_input_tokens": "ee16d2f0799734ef",
"hash_cont_tokens": "9cafa5a029671fcf"
},
"truncated": 0,
"non_truncated": 100,
"padded": 396,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:clinical_knowledge|0": {
"hashes": {
"hash_examples": "ffbb9c7b2be257f9",
"hash_full_prompts": "ffbb9c7b2be257f9",
"hash_input_tokens": "ac1bf6729bd8feb6",
"hash_cont_tokens": "85c06001ba520401"
},
"truncated": 0,
"non_truncated": 265,
"padded": 1060,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:college_biology|0": {
"hashes": {
"hash_examples": "3ee77f176f38eb8e",
"hash_full_prompts": "3ee77f176f38eb8e",
"hash_input_tokens": "e6334e1ca4b68674",
"hash_cont_tokens": "147d4bf8722e29b8"
},
"truncated": 0,
"non_truncated": 144,
"padded": 568,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:college_chemistry|0": {
"hashes": {
"hash_examples": "ce61a69c46d47aeb",
"hash_full_prompts": "ce61a69c46d47aeb",
"hash_input_tokens": "579af862a1802539",
"hash_cont_tokens": "43d344a3a511451c"
},
"truncated": 0,
"non_truncated": 100,
"padded": 400,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:college_computer_science|0": {
"hashes": {
"hash_examples": "32805b52d7d5daab",
"hash_full_prompts": "32805b52d7d5daab",
"hash_input_tokens": "55adcbebb7750997",
"hash_cont_tokens": "f57b0117d5493506"
},
"truncated": 0,
"non_truncated": 100,
"padded": 396,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:college_mathematics|0": {
"hashes": {
"hash_examples": "55da1a0a0bd33722",
"hash_full_prompts": "55da1a0a0bd33722",
"hash_input_tokens": "5a8fa193074b3ca3",
"hash_cont_tokens": "c9742de1a226550f"
},
"truncated": 0,
"non_truncated": 100,
"padded": 392,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:college_medicine|0": {
"hashes": {
"hash_examples": "c33e143163049176",
"hash_full_prompts": "c33e143163049176",
"hash_input_tokens": "078ed8c3fa3bc8e9",
"hash_cont_tokens": "16f983144c324c70"
},
"truncated": 0,
"non_truncated": 173,
"padded": 684,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:college_physics|0": {
"hashes": {
"hash_examples": "ebdab1cdb7e555df",
"hash_full_prompts": "ebdab1cdb7e555df",
"hash_input_tokens": "e156982a768f804a",
"hash_cont_tokens": "775e75c894c6a68a"
},
"truncated": 0,
"non_truncated": 102,
"padded": 408,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:computer_security|0": {
"hashes": {
"hash_examples": "a24fd7d08a560921",
"hash_full_prompts": "a24fd7d08a560921",
"hash_input_tokens": "ab40d5df656b2df9",
"hash_cont_tokens": "7ffc0312d6fa396e"
},
"truncated": 0,
"non_truncated": 100,
"padded": 400,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:conceptual_physics|0": {
"hashes": {
"hash_examples": "8300977a79386993",
"hash_full_prompts": "8300977a79386993",
"hash_input_tokens": "a11b8a3dc5ca8626",
"hash_cont_tokens": "bc507e44b2db8146"
},
"truncated": 0,
"non_truncated": 235,
"padded": 940,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:econometrics|0": {
"hashes": {
"hash_examples": "ddde36788a04a46f",
"hash_full_prompts": "ddde36788a04a46f",
"hash_input_tokens": "086789ac270f0fb8",
"hash_cont_tokens": "7421c15af82190b8"
},
"truncated": 0,
"non_truncated": 114,
"padded": 452,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:electrical_engineering|0": {
"hashes": {
"hash_examples": "acbc5def98c19b3f",
"hash_full_prompts": "acbc5def98c19b3f",
"hash_input_tokens": "f1aeb4190769da1b",
"hash_cont_tokens": "fde6678d78c6175b"
},
"truncated": 0,
"non_truncated": 145,
"padded": 580,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:elementary_mathematics|0": {
"hashes": {
"hash_examples": "146e61d07497a9bd",
"hash_full_prompts": "146e61d07497a9bd",
"hash_input_tokens": "7f7e3706b777d772",
"hash_cont_tokens": "99b1b1787c6b587a"
},
"truncated": 0,
"non_truncated": 378,
"padded": 1480,
"non_padded": 32,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:formal_logic|0": {
"hashes": {
"hash_examples": "8635216e1909a03f",
"hash_full_prompts": "8635216e1909a03f",
"hash_input_tokens": "7d278bc74a8975ff",
"hash_cont_tokens": "76f58b417f117854"
},
"truncated": 0,
"non_truncated": 126,
"padded": 500,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:global_facts|0": {
"hashes": {
"hash_examples": "30b315aa6353ee47",
"hash_full_prompts": "30b315aa6353ee47",
"hash_input_tokens": "94a3f760be5add9e",
"hash_cont_tokens": "57002e98809bf710"
},
"truncated": 0,
"non_truncated": 100,
"padded": 392,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_biology|0": {
"hashes": {
"hash_examples": "c9136373af2180de",
"hash_full_prompts": "c9136373af2180de",
"hash_input_tokens": "1657f58c617fa9c6",
"hash_cont_tokens": "793c1f6093f18378"
},
"truncated": 0,
"non_truncated": 310,
"padded": 1224,
"non_padded": 16,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_chemistry|0": {
"hashes": {
"hash_examples": "b0661bfa1add6404",
"hash_full_prompts": "b0661bfa1add6404",
"hash_input_tokens": "48024c3cae4ac8f9",
"hash_cont_tokens": "10a87034ea31a82e"
},
"truncated": 0,
"non_truncated": 203,
"padded": 808,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_computer_science|0": {
"hashes": {
"hash_examples": "80fc1d623a3d665f",
"hash_full_prompts": "80fc1d623a3d665f",
"hash_input_tokens": "898b574f856cb636",
"hash_cont_tokens": "cfb8bbe184073f18"
},
"truncated": 0,
"non_truncated": 100,
"padded": 396,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_european_history|0": {
"hashes": {
"hash_examples": "854da6e5af0fe1a1",
"hash_full_prompts": "854da6e5af0fe1a1",
"hash_input_tokens": "15ecc775303b9a1d",
"hash_cont_tokens": "fa695ce00a41ce0a"
},
"truncated": 0,
"non_truncated": 165,
"padded": 660,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_geography|0": {
"hashes": {
"hash_examples": "7dc963c7acd19ad8",
"hash_full_prompts": "7dc963c7acd19ad8",
"hash_input_tokens": "f1edca5b8c432bbf",
"hash_cont_tokens": "8d79b0f69b217745"
},
"truncated": 0,
"non_truncated": 198,
"padded": 788,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_government_and_politics|0": {
"hashes": {
"hash_examples": "1f675dcdebc9758f",
"hash_full_prompts": "1f675dcdebc9758f",
"hash_input_tokens": "8881b114a1e4ad12",
"hash_cont_tokens": "8f802c7dc09a6be2"
},
"truncated": 0,
"non_truncated": 193,
"padded": 760,
"non_padded": 12,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_macroeconomics|0": {
"hashes": {
"hash_examples": "2fb32cf2d80f0b35",
"hash_full_prompts": "2fb32cf2d80f0b35",
"hash_input_tokens": "25cb763d830fd806",
"hash_cont_tokens": "8893e7463bbf4b91"
},
"truncated": 0,
"non_truncated": 390,
"padded": 1536,
"non_padded": 24,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_mathematics|0": {
"hashes": {
"hash_examples": "fd6646fdb5d58a1f",
"hash_full_prompts": "fd6646fdb5d58a1f",
"hash_input_tokens": "a8906d9e33e0214a",
"hash_cont_tokens": "4868955bace38070"
},
"truncated": 0,
"non_truncated": 270,
"padded": 1076,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_microeconomics|0": {
"hashes": {
"hash_examples": "2118f21f71d87d84",
"hash_full_prompts": "2118f21f71d87d84",
"hash_input_tokens": "ca481d3baad7f289",
"hash_cont_tokens": "e8b823787b710a12"
},
"truncated": 0,
"non_truncated": 238,
"padded": 948,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_physics|0": {
"hashes": {
"hash_examples": "dc3ce06378548565",
"hash_full_prompts": "dc3ce06378548565",
"hash_input_tokens": "4259b80a7d21ef74",
"hash_cont_tokens": "834f1c074b49339e"
},
"truncated": 0,
"non_truncated": 151,
"padded": 600,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_psychology|0": {
"hashes": {
"hash_examples": "c8d1d98a40e11f2f",
"hash_full_prompts": "c8d1d98a40e11f2f",
"hash_input_tokens": "490ba9909b535d3d",
"hash_cont_tokens": "54161589e46df634"
},
"truncated": 0,
"non_truncated": 545,
"padded": 2152,
"non_padded": 28,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_statistics|0": {
"hashes": {
"hash_examples": "666c8759b98ee4ff",
"hash_full_prompts": "666c8759b98ee4ff",
"hash_input_tokens": "5556bcba1b0565fa",
"hash_cont_tokens": "833eec6f54b0c732"
},
"truncated": 0,
"non_truncated": 216,
"padded": 860,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_us_history|0": {
"hashes": {
"hash_examples": "95fef1c4b7d3f81e",
"hash_full_prompts": "95fef1c4b7d3f81e",
"hash_input_tokens": "b624d98485c9cee2",
"hash_cont_tokens": "2453ef70d8102d77"
},
"truncated": 0,
"non_truncated": 204,
"padded": 816,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:high_school_world_history|0": {
"hashes": {
"hash_examples": "7e5085b6184b0322",
"hash_full_prompts": "7e5085b6184b0322",
"hash_input_tokens": "80915978e077fe68",
"hash_cont_tokens": "0f88d74be6b857ea"
},
"truncated": 0,
"non_truncated": 237,
"padded": 948,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:human_aging|0": {
"hashes": {
"hash_examples": "c17333e7c7c10797",
"hash_full_prompts": "c17333e7c7c10797",
"hash_input_tokens": "8a00ffa15745266e",
"hash_cont_tokens": "20d50fbd57370a6c"
},
"truncated": 0,
"non_truncated": 223,
"padded": 892,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:human_sexuality|0": {
"hashes": {
"hash_examples": "4edd1e9045df5e3d",
"hash_full_prompts": "4edd1e9045df5e3d",
"hash_input_tokens": "3e04489c11225f01",
"hash_cont_tokens": "a6a547daf0286e50"
},
"truncated": 0,
"non_truncated": 131,
"padded": 520,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:international_law|0": {
"hashes": {
"hash_examples": "db2fa00d771a062a",
"hash_full_prompts": "db2fa00d771a062a",
"hash_input_tokens": "e8e6b00efbc87643",
"hash_cont_tokens": "b5407c77db7f2f4b"
},
"truncated": 0,
"non_truncated": 121,
"padded": 476,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:jurisprudence|0": {
"hashes": {
"hash_examples": "e956f86b124076fe",
"hash_full_prompts": "e956f86b124076fe",
"hash_input_tokens": "55f2b05e593d44f2",
"hash_cont_tokens": "9bf3c173ffe7a067"
},
"truncated": 0,
"non_truncated": 108,
"padded": 424,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:logical_fallacies|0": {
"hashes": {
"hash_examples": "956e0e6365ab79f1",
"hash_full_prompts": "956e0e6365ab79f1",
"hash_input_tokens": "04cc61b11aa11edd",
"hash_cont_tokens": "f0ffa328ffbc1639"
},
"truncated": 0,
"non_truncated": 163,
"padded": 648,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:machine_learning|0": {
"hashes": {
"hash_examples": "397997cc6f4d581e",
"hash_full_prompts": "397997cc6f4d581e",
"hash_input_tokens": "a5c4d754ce7e9176",
"hash_cont_tokens": "8930a57f098fa487"
},
"truncated": 0,
"non_truncated": 112,
"padded": 444,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:management|0": {
"hashes": {
"hash_examples": "2bcbe6f6ca63d740",
"hash_full_prompts": "2bcbe6f6ca63d740",
"hash_input_tokens": "df114f5949470e68",
"hash_cont_tokens": "c6864e99f08c42e2"
},
"truncated": 0,
"non_truncated": 103,
"padded": 412,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:marketing|0": {
"hashes": {
"hash_examples": "8ddb20d964a1b065",
"hash_full_prompts": "8ddb20d964a1b065",
"hash_input_tokens": "50d069db0dce34df",
"hash_cont_tokens": "0ee3c521bbd24c81"
},
"truncated": 0,
"non_truncated": 234,
"padded": 932,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:medical_genetics|0": {
"hashes": {
"hash_examples": "182a71f4763d2cea",
"hash_full_prompts": "182a71f4763d2cea",
"hash_input_tokens": "4a40133ad3a15075",
"hash_cont_tokens": "57002e98809bf710"
},
"truncated": 0,
"non_truncated": 100,
"padded": 396,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:miscellaneous|0": {
"hashes": {
"hash_examples": "4c404fdbb4ca57fc",
"hash_full_prompts": "4c404fdbb4ca57fc",
"hash_input_tokens": "adec3091ce5943d1",
"hash_cont_tokens": "33fa59579b01dcf7"
},
"truncated": 0,
"non_truncated": 783,
"padded": 3124,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:moral_disputes|0": {
"hashes": {
"hash_examples": "60cbd2baa3fea5c9",
"hash_full_prompts": "60cbd2baa3fea5c9",
"hash_input_tokens": "275a765273325318",
"hash_cont_tokens": "aeb2e3df581375d2"
},
"truncated": 0,
"non_truncated": 346,
"padded": 1356,
"non_padded": 28,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:moral_scenarios|0": {
"hashes": {
"hash_examples": "fd8b0431fbdd75ef",
"hash_full_prompts": "fd8b0431fbdd75ef",
"hash_input_tokens": "dbbd61f7dcb34df7",
"hash_cont_tokens": "f97e5f05a6fab216"
},
"truncated": 0,
"non_truncated": 895,
"padded": 3580,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:nutrition|0": {
"hashes": {
"hash_examples": "71e55e2b829b6528",
"hash_full_prompts": "71e55e2b829b6528",
"hash_input_tokens": "1d592d46420a7af8",
"hash_cont_tokens": "0cff076b60b939b9"
},
"truncated": 0,
"non_truncated": 306,
"padded": 1220,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:philosophy|0": {
"hashes": {
"hash_examples": "a6d489a8d208fa4b",
"hash_full_prompts": "a6d489a8d208fa4b",
"hash_input_tokens": "a09ce450f5ef1e7a",
"hash_cont_tokens": "fa9d6a93a12be228"
},
"truncated": 0,
"non_truncated": 311,
"padded": 1244,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:prehistory|0": {
"hashes": {
"hash_examples": "6cc50f032a19acaa",
"hash_full_prompts": "6cc50f032a19acaa",
"hash_input_tokens": "4c8a25175d952d1a",
"hash_cont_tokens": "2551df3b83d95c91"
},
"truncated": 0,
"non_truncated": 324,
"padded": 1288,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:professional_accounting|0": {
"hashes": {
"hash_examples": "50f57ab32f5f6cea",
"hash_full_prompts": "50f57ab32f5f6cea",
"hash_input_tokens": "39d5857440021ead",
"hash_cont_tokens": "8dfb834ed69b9a63"
},
"truncated": 0,
"non_truncated": 282,
"padded": 1108,
"non_padded": 20,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:professional_law|0": {
"hashes": {
"hash_examples": "a8fdc85c64f4b215",
"hash_full_prompts": "a8fdc85c64f4b215",
"hash_input_tokens": "58b944eb0581064f",
"hash_cont_tokens": "90aa19208cf48371"
},
"truncated": 0,
"non_truncated": 1534,
"padded": 6132,
"non_padded": 4,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:professional_medicine|0": {
"hashes": {
"hash_examples": "c373a28a3050a73a",
"hash_full_prompts": "c373a28a3050a73a",
"hash_input_tokens": "03999f2b4bc3d63c",
"hash_cont_tokens": "87192b5d770945cb"
},
"truncated": 0,
"non_truncated": 272,
"padded": 1088,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:professional_psychology|0": {
"hashes": {
"hash_examples": "bf5254fe818356af",
"hash_full_prompts": "bf5254fe818356af",
"hash_input_tokens": "c97b675eaa9b4bf4",
"hash_cont_tokens": "75c67c831f21f75b"
},
"truncated": 0,
"non_truncated": 612,
"padded": 2416,
"non_padded": 32,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:public_relations|0": {
"hashes": {
"hash_examples": "b66d52e28e7d14e0",
"hash_full_prompts": "b66d52e28e7d14e0",
"hash_input_tokens": "35c5d7600493ff13",
"hash_cont_tokens": "500c34e39ca8a57d"
},
"truncated": 0,
"non_truncated": 110,
"padded": 440,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:security_studies|0": {
"hashes": {
"hash_examples": "514c14feaf000ad9",
"hash_full_prompts": "514c14feaf000ad9",
"hash_input_tokens": "bce1c31864d3a9d1",
"hash_cont_tokens": "f4787430091b26e1"
},
"truncated": 0,
"non_truncated": 245,
"padded": 980,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:sociology|0": {
"hashes": {
"hash_examples": "f6c9bc9d18c80870",
"hash_full_prompts": "f6c9bc9d18c80870",
"hash_input_tokens": "10dd0c8cfe298b5f",
"hash_cont_tokens": "d2e9915d64a35e7a"
},
"truncated": 0,
"non_truncated": 201,
"padded": 796,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:us_foreign_policy|0": {
"hashes": {
"hash_examples": "ed7b78629db6678f",
"hash_full_prompts": "ed7b78629db6678f",
"hash_input_tokens": "f2fb51e41e4463e9",
"hash_cont_tokens": "1a7af3ca7869de2e"
},
"truncated": 0,
"non_truncated": 100,
"padded": 392,
"non_padded": 8,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:virology|0": {
"hashes": {
"hash_examples": "bc52ffdc3f9b994a",
"hash_full_prompts": "bc52ffdc3f9b994a",
"hash_input_tokens": "82148f4342caa66d",
"hash_cont_tokens": "0be66455983a917e"
},
"truncated": 0,
"non_truncated": 166,
"padded": 664,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_mc:world_religions|0": {
"hashes": {
"hash_examples": "ecdb4a4f94f62930",
"hash_full_prompts": "ecdb4a4f94f62930",
"hash_input_tokens": "f9d6bc2fb36855ea",
"hash_cont_tokens": "584bbaf1621cc1f9"
},
"truncated": 0,
"non_truncated": 171,
"padded": 684,
"non_padded": 0,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|mmlu_pro_cloze|0": {
"hashes": {
"hash_examples": "845e15cfeee1fc11",
"hash_full_prompts": "845e15cfeee1fc11",
"hash_input_tokens": "dbc4fdb4c171fc52",
"hash_cont_tokens": "72e731df47f7ef00"
},
"truncated": 0,
"non_truncated": 12032,
"padded": 111558,
"non_padded": 2437,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|openbookqa|0": {
"hashes": {
"hash_examples": "fd427af2ef0577e3",
"hash_full_prompts": "fd427af2ef0577e3",
"hash_input_tokens": "924e3881a38e2c6c",
"hash_cont_tokens": "30d1cae5d2df9faa"
},
"truncated": 0,
"non_truncated": 500,
"padded": 1994,
"non_padded": 6,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|piqa|0": {
"hashes": {
"hash_examples": "f7e288a8894cd149",
"hash_full_prompts": "f7e288a8894cd149",
"hash_input_tokens": "2a53824cfb6e393d",
"hash_cont_tokens": "89c61f01a2e4acbb"
},
"truncated": 0,
"non_truncated": 1838,
"padded": 3602,
"non_padded": 74,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|siqa|0": {
"hashes": {
"hash_examples": "c62abc8ecbd49cc4",
"hash_full_prompts": "c62abc8ecbd49cc4",
"hash_input_tokens": "2eefab74aa9434f9",
"hash_cont_tokens": "2beff714f7e6bb8b"
},
"truncated": 0,
"non_truncated": 1954,
"padded": 5782,
"non_padded": 80,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|winogrande|0": {
"hashes": {
"hash_examples": "087d5d1a1afd4c7b",
"hash_full_prompts": "087d5d1a1afd4c7b",
"hash_input_tokens": "17ac595c26748c2b",
"hash_cont_tokens": "d9811a5ce62e0917"
},
"truncated": 0,
"non_truncated": 1267,
"padded": 2532,
"non_padded": 2,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
},
"custom|gsm8k|5": {
"hashes": {
"hash_examples": "0ed016e24e7512fd",
"hash_full_prompts": "41d55e83abc0e02d",
"hash_input_tokens": "4c698c7148ab9bfc",
"hash_cont_tokens": "2b030b835fe3df09"
},
"truncated": 1319,
"non_truncated": 0,
"padded": 948,
"non_padded": 371,
"effective_few_shots": 5.0,
"num_truncated_few_shots": 0
},
"custom|trivia_qa|0": {
"hashes": {
"hash_examples": "1e083041cb75ff0c",
"hash_full_prompts": "1e083041cb75ff0c",
"hash_input_tokens": "01d7706e0dac8380",
"hash_cont_tokens": "9160ed2743838666"
},
"truncated": 17944,
"non_truncated": 0,
"padded": 269,
"non_padded": 17675,
"effective_few_shots": 0.0,
"num_truncated_few_shots": 0
}
},
"summary_general": {
"hashes": {
"hash_examples": "88e809e91c59a89c",
"hash_full_prompts": "60e948f5cd4b4b2e",
"hash_input_tokens": "c4e85a9af5c2ad3f",
"hash_cont_tokens": "788366827136aa8c"
},
"truncated": 19263,
"non_truncated": 60486,
"padded": 297962,
"non_padded": 22165,
"num_truncated_few_shots": 0
}
}