{
   "config_general": {
      "lighteval_sha": "",
      "num_few_shot_default": 0,
      "num_fewshot_seeds": 1,
      "override_batch_size": "auto:6",
      "max_samples": "null",
      "job_id": "",
      "model_name": "VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct",
      "model_sha": "",
      "model_dtype": "torch.bfloat16",
      "model_size": ""
   },
   "results": {
      "harness|truthfulqa_mc2_m_de|0": {
         "acc,none": 0.2956852791878173,
         "acc_stderr,none": 0.01626712741645017,
         "alias": "truthfulqa_mc2_m_de"
      },
      "harness|truthfulqa_mc2_m_es|0": {
         "acc,none": 0.3358681875792142,
         "acc_stderr,none": 0.01682474414131292,
         "alias": "truthfulqa_mc2_m_es"
      },
      "harness|arc_challenge_m_it|25": {
         "acc,none": 0.5945252352437981,
         "acc_stderr,none": 0.014366323465528179,
         "acc_norm,none": 0.6193327630453379,
         "acc_norm_stderr,none": 0.014207359046710982,
         "alias": "arc_challenge_m_it"
      },
      "harness|mmlu_m_de|5": {
         "acc,none": 0.5644139387539598,
         "acc_stderr,none": 0.004306388668808089,
         "alias": "mmlu_m_de"
      },
      "harness|belebele_ita_Latn|5": {
         "acc,none": 0.8333333333333334,
         "acc_stderr,none": 0.012429507075907734,
         "acc_norm,none": 0.8333333333333334,
         "acc_norm_stderr,none": 0.012429507075907734,
         "alias": "belebele_ita_Latn"
      },
      "harness|mmlu_m_fr|5": {
         "acc,none": 0.5821556794744481,
         "acc_stderr,none": 0.004310791301464171,
         "alias": "mmlu_m_fr"
      },
      "harness|belebele_eng_Latn|5": {
         "acc,none": 0.9166666666666666,
         "acc_stderr,none": 0.009217969157169199,
         "acc_norm,none": 0.9166666666666666,
         "acc_norm_stderr,none": 0.009217969157169199,
         "alias": "belebele_eng_Latn"
      },
      "harness|truthfulqa_mc2_m_it|0": {
         "acc,none": 0.3103448275862069,
         "acc_stderr,none": 0.016543785026048318,
         "alias": "truthfulqa_mc2_m_it"
      },
      "harness|arc_challenge_m_de|25": {
         "acc,none": 0.5500427715996579,
         "acc_stderr,none": 0.014556683049829988,
         "acc_norm,none": 0.5936698032506416,
         "acc_norm_stderr,none": 0.014371119725222025,
         "alias": "arc_challenge_m_de"
      },
      "harness|mmlu_m_es|5": {
         "acc,none": 0.5901454927253638,
         "acc_stderr,none": 0.004259223847756214,
         "alias": "mmlu_m_es"
      },
      "harness|gsm8k|5": {
         "exact_match,get-answer": 0.6846095526914329,
         "exact_match_stderr,get-answer": 0.012799353675801832,
         "alias": "gsm8k"
      },
      "harness|arc_challenge_m_es|25": {
         "acc,none": 0.6205128205128205,
         "acc_stderr,none": 0.014192754090886751,
         "acc_norm,none": 0.6410256410256411,
         "acc_norm_stderr,none": 0.014030145004220061,
         "alias": "arc_challenge_m_es"
      },
      "harness|belebele_fra_Latn|5": {
         "acc,none": 0.8766666666666667,
         "acc_stderr,none": 0.010966742231624086,
         "acc_norm,none": 0.8766666666666667,
         "acc_norm_stderr,none": 0.010966742231624086,
         "alias": "belebele_fra_Latn"
      },
      "harness|arc_challenge_m_fr|25": {
         "acc,none": 0.5996578272027374,
         "acc_stderr,none": 0.014336594541437478,
         "acc_norm,none": 0.630453378956373,
         "acc_norm_stderr,none": 0.014123413837443291,
         "alias": "arc_challenge_m_fr"
      },
      "harness|belebele_spa_Latn|5": {
         "acc,none": 0.8744444444444445,
         "acc_stderr,none": 0.011051067526018464,
         "acc_norm,none": 0.8744444444444445,
         "acc_norm_stderr,none": 0.011051067526018464,
         "alias": "belebele_spa_Latn"
      },
      "harness|mmlu_m_it|5": {
         "acc,none": 0.5744504041701292,
         "acc_stderr,none": 0.004297570220538641,
         "alias": "mmlu_m_it"
      },
      "harness|arc_challenge|25": {
         "acc,none": 0.7303754266211604,
         "acc_stderr,none": 0.01296804068686914,
         "acc_norm,none": 0.7448805460750854,
         "acc_norm_stderr,none": 0.012739038695202109,
         "alias": "arc_challenge"
      },
      "harness|hendrycksTest|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-humanities|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-formal_logic|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_european_history|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_us_history|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_world_history|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-international_law|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-jurisprudence|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-logical_fallacies|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-moral_disputes|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-moral_scenarios|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-philosophy|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-prehistory|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-professional_law|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-world_religions|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-other|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-business_ethics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-clinical_knowledge|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-college_medicine|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-global_facts|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-human_aging|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-management|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-marketing|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-medical_genetics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-miscellaneous|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-nutrition|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-professional_accounting|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-professional_medicine|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-virology|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-social_sciences|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-econometrics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_geography|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_government_and_politics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_macroeconomics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_microeconomics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_psychology|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-human_sexuality|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-professional_psychology|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-public_relations|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-security_studies|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-sociology|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-us_foreign_policy|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-stem|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-abstract_algebra|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-anatomy|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-astronomy|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-college_biology|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-college_chemistry|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-college_computer_science|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-college_mathematics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-college_physics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-computer_security|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-conceptual_physics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-electrical_engineering|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-elementary_mathematics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_biology|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_chemistry|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_computer_science|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_mathematics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_physics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-high_school_statistics|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hendrycksTest-machine_learning|5": {
         "acc,none": 0.6665004985044866,
         "acc_stderr,none": 0.12578286666329366,
         "alias": "mmlu"
      },
      "harness|hellaswag|10": {
         "acc,none": 0.7454690300736905,
         "acc_stderr,none": 0.004347070019527486,
         "acc_norm,none": 0.8956383190599482,
         "acc_norm_stderr,none": 0.0030510433788747708,
         "alias": "hellaswag"
      },
      "harness|hellaswag_es|10": {
         "acc,none": 0.5746746319607425,
         "acc_stderr,none": 0.005106605961593097,
         "acc_norm,none": 0.7491999146575635,
         "acc_norm_stderr,none": 0.004477374186359933,
         "alias": "hellaswag_es"
      },
      "harness|hellaswag_de|10": {
         "acc,none": 0.5146242527754057,
         "acc_stderr,none": 0.005163972260259133,
         "acc_norm,none": 0.6642826643894107,
         "acc_norm_stderr,none": 0.004879362489588698,
         "alias": "hellaswag_de"
      },
      "harness|hellaswag_it|10": {
         "acc,none": 0.5330142499728054,
         "acc_stderr,none": 0.005203747512549052,
         "acc_norm,none": 0.7029261394539323,
         "acc_norm_stderr,none": 0.004766307515131304,
         "alias": "hellaswag_it"
      },
      "harness|truthfulqa_mc2_m_fr|0": {
         "acc,none": 0.33672172808132145,
         "acc_stderr,none": 0.0168566854595136,
         "alias": "truthfulqa_mc2_m_fr"
      },
      "harness|truthfulqa_mc2|0": {
         "acc,none": 0.661926387370984,
         "acc_stderr,none": 0.015357570618077165,
         "alias": "truthfulqa_mc2"
      },
      "harness|hellaswag_fr|10": {
         "acc,none": 0.5593274791175841,
         "acc_stderr,none": 0.005137920542961698,
         "acc_norm,none": 0.7264938959091882,
         "acc_norm_stderr,none": 0.004613131313984876,
         "alias": "hellaswag_fr"
      },
      "harness|belebele_deu_Latn|5": {
         "acc,none": 0.8855555555555555,
         "acc_stderr,none": 0.010617576963634278,
         "acc_norm,none": 0.8855555555555555,
         "acc_norm_stderr,none": 0.010617576963634278,
         "alias": "belebele_deu_Latn"
      }
   },
   "versions": {
      "harness|truthfulqa_mc2_m_de|0": "Yaml",
      "harness|truthfulqa_mc2_m_es|0": "Yaml",
      "harness|arc_challenge_m_it|25": 1.0,
      "harness|mmlu_m_de|5": "Yaml",
      "harness|belebele_ita_Latn|5": 0.0,
      "harness|mmlu_m_fr|5": "Yaml",
      "harness|belebele_eng_Latn|5": 0.0,
      "harness|truthfulqa_mc2_m_it|0": "Yaml",
      "harness|arc_challenge_m_de|25": 1.0,
      "harness|mmlu_m_es|5": "Yaml",
      "harness|gsm8k|5": 2.0,
      "harness|arc_challenge_m_es|25": 1.0,
      "harness|belebele_fra_Latn|5": 0.0,
      "harness|arc_challenge_m_fr|25": 1.0,
      "harness|belebele_spa_Latn|5": 0.0,
      "harness|mmlu_m_it|5": "Yaml",
      "harness|arc_challenge|25": 1.0,
      "harness|hendrycksTest|5": "N/A",
      "harness|hendrycksTest-humanities|5": "N/A",
      "harness|hendrycksTest-formal_logic|5": "N/A",
      "harness|hendrycksTest-high_school_european_history|5": "N/A",
      "harness|hendrycksTest-high_school_us_history|5": "N/A",
      "harness|hendrycksTest-high_school_world_history|5": "N/A",
      "harness|hendrycksTest-international_law|5": "N/A",
      "harness|hendrycksTest-jurisprudence|5": "N/A",
      "harness|hendrycksTest-logical_fallacies|5": "N/A",
      "harness|hendrycksTest-moral_disputes|5": "N/A",
      "harness|hendrycksTest-moral_scenarios|5": "N/A",
      "harness|hendrycksTest-philosophy|5": "N/A",
      "harness|hendrycksTest-prehistory|5": "N/A",
      "harness|hendrycksTest-professional_law|5": "N/A",
      "harness|hendrycksTest-world_religions|5": "N/A",
      "harness|hendrycksTest-other|5": "N/A",
      "harness|hendrycksTest-business_ethics|5": "N/A",
      "harness|hendrycksTest-clinical_knowledge|5": "N/A",
      "harness|hendrycksTest-college_medicine|5": "N/A",
      "harness|hendrycksTest-global_facts|5": "N/A",
      "harness|hendrycksTest-human_aging|5": "N/A",
      "harness|hendrycksTest-management|5": "N/A",
      "harness|hendrycksTest-marketing|5": "N/A",
      "harness|hendrycksTest-medical_genetics|5": "N/A",
      "harness|hendrycksTest-miscellaneous|5": "N/A",
      "harness|hendrycksTest-nutrition|5": "N/A",
      "harness|hendrycksTest-professional_accounting|5": "N/A",
      "harness|hendrycksTest-professional_medicine|5": "N/A",
      "harness|hendrycksTest-virology|5": "N/A",
      "harness|hendrycksTest-social_sciences|5": "N/A",
      "harness|hendrycksTest-econometrics|5": "N/A",
      "harness|hendrycksTest-high_school_geography|5": "N/A",
      "harness|hendrycksTest-high_school_government_and_politics|5": "N/A",
      "harness|hendrycksTest-high_school_macroeconomics|5": "N/A",
      "harness|hendrycksTest-high_school_microeconomics|5": "N/A",
      "harness|hendrycksTest-high_school_psychology|5": "N/A",
      "harness|hendrycksTest-human_sexuality|5": "N/A",
      "harness|hendrycksTest-professional_psychology|5": "N/A",
      "harness|hendrycksTest-public_relations|5": "N/A",
      "harness|hendrycksTest-security_studies|5": "N/A",
      "harness|hendrycksTest-sociology|5": "N/A",
      "harness|hendrycksTest-us_foreign_policy|5": "N/A",
      "harness|hendrycksTest-stem|5": "N/A",
      "harness|hendrycksTest-abstract_algebra|5": "N/A",
      "harness|hendrycksTest-anatomy|5": "N/A",
      "harness|hendrycksTest-astronomy|5": "N/A",
      "harness|hendrycksTest-college_biology|5": "N/A",
      "harness|hendrycksTest-college_chemistry|5": "N/A",
      "harness|hendrycksTest-college_computer_science|5": "N/A",
      "harness|hendrycksTest-college_mathematics|5": "N/A",
      "harness|hendrycksTest-college_physics|5": "N/A",
      "harness|hendrycksTest-computer_security|5": "N/A",
      "harness|hendrycksTest-conceptual_physics|5": "N/A",
      "harness|hendrycksTest-electrical_engineering|5": "N/A",
      "harness|hendrycksTest-elementary_mathematics|5": "N/A",
      "harness|hendrycksTest-high_school_biology|5": "N/A",
      "harness|hendrycksTest-high_school_chemistry|5": "N/A",
      "harness|hendrycksTest-high_school_computer_science|5": "N/A",
      "harness|hendrycksTest-high_school_mathematics|5": "N/A",
      "harness|hendrycksTest-high_school_physics|5": "N/A",
      "harness|hendrycksTest-high_school_statistics|5": "N/A",
      "harness|hendrycksTest-machine_learning|5": "N/A",
      "harness|hellaswag|10": 1.0,
      "harness|hellaswag_es|10": 1.0,
      "harness|hellaswag_de|10": 1.0,
      "harness|hellaswag_it|10": 1.0,
      "harness|truthfulqa_mc2_m_fr|0": "Yaml",
      "harness|truthfulqa_mc2|0": 2.0,
      "harness|hellaswag_fr|10": 1.0,
      "harness|belebele_deu_Latn|5": 0.0
   }
}