diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/RedPajama-INCITE-7B-Base/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 6c7a631bdbdd66726aa604d1a7a78b9152863426..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:310cb302ba58221afa5b4b77eded59124462d2e404e554a396360e14e1432d64 -size 5210477 diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/RedPajama-INCITE-7B-Base/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 0b2aae93ffd5a7282eb03420792c521533f09727..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,252 +0,0 @@ -{ - "results": { - "lambada_multilingual": { - "perplexity,none": 55.75634065211411, - "perplexity_stderr,none": 17.112832544800874, - "acc,none": 0.43578497962352025, - "acc_stderr,none": 0.07783182786033929, - "alias": "lambada_multilingual" - }, - "lambada_openai_mt_de": { - "perplexity,none": 87.03760418597274, - "perplexity_stderr,none": 5.220743099786189, - "acc,none": 0.318067145352222, - "acc_stderr,none": 0.006488469772173893, - "alias": " - lambada_openai_mt_de" - }, - "lambada_openai_mt_en": { - "perplexity,none": 4.00575099472505, - "perplexity_stderr,none": 0.08559465754530345, - "acc,none": 0.7003687172520862, - "acc_stderr,none": 0.006382179569794074, - "alias": " - lambada_openai_mt_en" - }, - "lambada_openai_mt_es": { - "perplexity,none": 74.50603551865778, - "perplexity_stderr,none": 4.146635362251485, - "acc,none": 0.3483407723656123, - "acc_stderr,none": 0.006637805195772818, - "alias": " - lambada_openai_mt_es" - }, - "lambada_openai_mt_fr": { - "perplexity,none": 47.60819762333609, - "perplexity_stderr,none": 2.6897251543883476, - "acc,none": 0.42227828449446925, - "acc_stderr,none": 0.006881304773376873, - "alias": " - lambada_openai_mt_fr" - }, - "lambada_openai_mt_it": { - "perplexity,none": 65.62411493787893, - "perplexity_stderr,none": 3.9555857520848434, - "acc,none": 0.3898699786532117, - "acc_stderr,none": 0.006794901529888746, - "alias": " - lambada_openai_mt_it" - } - }, - "groups": { - "lambada_multilingual": { - "perplexity,none": 55.75634065211411, - "perplexity_stderr,none": 17.112832544800874, - "acc,none": 0.43578497962352025, - "acc_stderr,none": 0.07783182786033929, - "alias": "lambada_multilingual" - } - }, - "configs": { - "lambada_openai_mt_de": { - "task": "lambada_openai_mt_de", - "group": [ - "lambada_multilingual" - ], - "dataset_path": "EleutherAI/lambada_openai", - "dataset_name": "de", - "test_split": "test", - "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}", - "doc_to_target": "{{' '+text.split(' ')[-1]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "perplexity", - "aggregation": "perplexity", - "higher_is_better": false - }, - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "loglikelihood", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{text}}", - "metadata": { - "version": 1.0 - } - }, - "lambada_openai_mt_en": { - "task": "lambada_openai_mt_en", - "group": [ - "lambada_multilingual" - ], - "dataset_path": "EleutherAI/lambada_openai", - "dataset_name": "en", - "test_split": "test", - "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}", - "doc_to_target": "{{' '+text.split(' ')[-1]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "perplexity", - "aggregation": "perplexity", - "higher_is_better": false - }, - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "loglikelihood", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{text}}", - "metadata": { - "version": 1.0 - } - }, - "lambada_openai_mt_es": { - "task": "lambada_openai_mt_es", - "group": [ - "lambada_multilingual" - ], - "dataset_path": "EleutherAI/lambada_openai", - "dataset_name": "es", - "test_split": "test", - "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}", - "doc_to_target": "{{' '+text.split(' ')[-1]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "perplexity", - "aggregation": "perplexity", - "higher_is_better": false - }, - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "loglikelihood", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{text}}", - "metadata": { - "version": 1.0 - } - }, - "lambada_openai_mt_fr": { - "task": "lambada_openai_mt_fr", - "group": [ - "lambada_multilingual" - ], - "dataset_path": "EleutherAI/lambada_openai", - "dataset_name": "fr", - "test_split": "test", - "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}", - "doc_to_target": "{{' '+text.split(' ')[-1]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "perplexity", - "aggregation": "perplexity", - "higher_is_better": false - }, - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "loglikelihood", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{text}}", - "metadata": { - "version": 1.0 - } - }, - "lambada_openai_mt_it": { - "task": "lambada_openai_mt_it", - "group": [ - "lambada_multilingual" - ], - "dataset_path": "EleutherAI/lambada_openai", - "dataset_name": "it", - "test_split": "test", - "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}", - "doc_to_target": "{{' '+text.split(' ')[-1]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "perplexity", - "aggregation": "perplexity", - "higher_is_better": false - }, - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "loglikelihood", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{text}}", - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "lambada_multilingual": "N/A", - "lambada_openai_mt_de": 1.0, - "lambada_openai_mt_en": 1.0, - "lambada_openai_mt_es": 1.0, - "lambada_openai_mt_fr": 1.0, - "lambada_openai_mt_it": 1.0 - }, - "n-shot": { - "lambada_multilingual": 0, - "lambada_openai_mt_de": 0, - "lambada_openai_mt_en": 0, - "lambada_openai_mt_es": 0, - "lambada_openai_mt_fr": 0, - "lambada_openai_mt_it": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=togethercomputer/RedPajama-INCITE-7B-Base,dtype=bfloat16,trust_remote_code=True", - "batch_size": "auto", - "batch_sizes": [ - 16 - ], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "2c0a875" -} \ No newline at end of file diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/RedPajama-INCITE-7B-Base/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log deleted file mode 100644 index 8f30eb6ee76531193f11dd10f0278a3d743e22b5..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:82e89447db6627b75a9e38192365c5c94b7c889a1bd3fbc15efae6c880c35cb8 -size 189167 diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/RedPajama-INCITE-7B-Base/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 73c7f12e7b616167da8e61c71da358046a57d266..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:04f23a0163c17bec0e2b07f28182ec884262d5d00b34ed2e5bb793d22657a6c6 -size 2133469 diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/RedPajama-INCITE-7B-Base/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 692267b6b06ac0f508828f8f900a0728461c026e..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,283 +0,0 @@ -{ - "results": { - "pawsx": { - "acc,none": 0.4830714285714286, - "acc_stderr,none": 0.032587798080603594, - "alias": "pawsx" - }, - "paws_de": { - "acc,none": 0.4565, - "acc_stderr,none": 0.011140733053371408, - "alias": " - paws_de" - }, - "paws_en": { - "acc,none": 0.445, - "acc_stderr,none": 0.011115272135099207, - "alias": " - paws_en" - }, - "paws_es": { - "acc,none": 0.412, - "acc_stderr,none": 0.011008569130325172, - "alias": " - paws_es" - }, - "paws_fr": { - "acc,none": 0.531, - "acc_stderr,none": 0.011161621338114474, - "alias": " - paws_fr" - }, - "paws_ja": { - "acc,none": 0.4825, - "acc_stderr,none": 0.01117628425125418, - "alias": " - paws_ja" - }, - "paws_ko": { - "acc,none": 0.524, - "acc_stderr,none": 0.011170245619215438, - "alias": " - paws_ko" - }, - "paws_zh": { - "acc,none": 0.5305, - "acc_stderr,none": 0.011162310405413182, - "alias": " - paws_zh" - } - }, - "groups": { - "pawsx": { - "acc,none": 0.4830714285714286, - "acc_stderr,none": 0.032587798080603594, - "alias": "pawsx" - } - }, - "configs": { - "paws_de": { - "task": "paws_de", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "de", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", richtig? Ja, \"+sentence2, sentence1+\", richtig? Nein, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "paws_en": { - "task": "paws_en", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "en", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", right? Yes, \"+sentence2, sentence1+\", right? No, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "paws_es": { - "task": "paws_es", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "es", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", verdad? Sí, \"+sentence2, sentence1+\", verdad? No, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "paws_fr": { - "task": "paws_fr", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "fr", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", n'est-ce pas? Oui, \"+sentence2, sentence1+\", n'est-ce pas? No, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "paws_ja": { - "task": "paws_ja", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "ja", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", ですね? はい, \"+sentence2, sentence1+\", ですね? いいえ, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "paws_ko": { - "task": "paws_ko", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "ko", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", 맞죠? 예, \"+sentence2, sentence1+\", 맞죠? 아니요, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "paws_zh": { - "task": "paws_zh", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "zh", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", 对吧? 是, \"+sentence2, sentence1+\", 对吧? 不是, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - } - }, - "versions": { - "paws_de": 0.0, - "paws_en": 0.0, - "paws_es": 0.0, - "paws_fr": 0.0, - "paws_ja": 0.0, - "paws_ko": 0.0, - "paws_zh": 0.0, - "pawsx": "N/A" - }, - "n-shot": { - "paws_de": 0, - "paws_en": 0, - "paws_es": 0, - "paws_fr": 0, - "paws_ja": 0, - "paws_ko": 0, - "paws_zh": 0, - "pawsx": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=togethercomputer/RedPajama-INCITE-7B-Base,dtype=bfloat16,trust_remote_code=True", - "batch_size": "auto", - "batch_sizes": [ - 16 - ], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "2c0a875" -} \ No newline at end of file diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/RedPajama-INCITE-7B-Base/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log deleted file mode 100644 index e59c69289e5b015a75d6814fa98cfca27d02d287..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:edb8d10109c30a134c04094c23d91f0598a09cc74c21c12cc8180f20db55a520 -size 170261 diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/RedPajama-INCITE-7B-Base/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index b16ff94f6505b0c20b35865372d310e0881ed194..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:87474ecc21f8109f1b039f37e49d45a577e1a9d22ca1cda05cd8fdd45512d25b -size 534066 diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/RedPajama-INCITE-7B-Base/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index dfb676e1d88e73e9ad44e9b55acc65f350cd64be..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,390 +0,0 @@ -{ - "results": { - "xcopa": { - "acc,none": 0.5254545454545455, - "acc_stderr,none": 0.036407165846333675, - "alias": "xcopa" - }, - "xcopa_et": { - "acc,none": 0.492, - "acc_stderr,none": 0.022380208834928035, - "alias": " - xcopa_et" - }, - "xcopa_ht": { - "acc,none": 0.502, - "acc_stderr,none": 0.022382894986483524, - "alias": " - xcopa_ht" - }, - "xcopa_id": { - "acc,none": 0.54, - "acc_stderr,none": 0.02231133324528966, - "alias": " - xcopa_id" - }, - "xcopa_it": { - "acc,none": 0.604, - "acc_stderr,none": 0.021893529941665813, - "alias": " - xcopa_it" - }, - "xcopa_qu": { - "acc,none": 0.478, - "acc_stderr,none": 0.02236139673920788, - "alias": " - xcopa_qu" - }, - "xcopa_sw": { - "acc,none": 0.522, - "acc_stderr,none": 0.02236139673920788, - "alias": " - xcopa_sw" - }, - "xcopa_ta": { - "acc,none": 0.546, - "acc_stderr,none": 0.02228814759117695, - "alias": " - xcopa_ta" - }, - "xcopa_th": { - "acc,none": 0.532, - "acc_stderr,none": 0.022337186479044292, - "alias": " - xcopa_th" - }, - "xcopa_tr": { - "acc,none": 0.514, - "acc_stderr,none": 0.02237429816635319, - "alias": " - xcopa_tr" - }, - "xcopa_vi": { - "acc,none": 0.494, - "acc_stderr,none": 0.022381462412439324, - "alias": " - xcopa_vi" - }, - "xcopa_zh": { - "acc,none": 0.556, - "acc_stderr,none": 0.02224224437573102, - "alias": " - xcopa_zh" - } - }, - "groups": { - "xcopa": { - "acc,none": 0.5254545454545455, - "acc_stderr,none": 0.036407165846333675, - "alias": "xcopa" - } - }, - "configs": { - "xcopa_et": { - "task": "xcopa_et", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "et", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'sest', 'effect': 'seetõttu'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_ht": { - "task": "xcopa_ht", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "ht", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'poukisa', 'effect': 'donk sa'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_id": { - "task": "xcopa_id", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "id", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'karena', 'effect': 'maka'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_it": { - "task": "xcopa_it", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "it", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'perché', 'effect': 'quindi'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_qu": { - "task": "xcopa_qu", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "qu", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'imataq', 'effect': 'chaymi'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_sw": { - "task": "xcopa_sw", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "sw", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'kwa sababu', 'effect': 'kwa hiyo'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_ta": { - "task": "xcopa_ta", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "ta", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'காரணமாக', 'effect': 'எனவே'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_th": { - "task": "xcopa_th", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "th", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'เพราะ', 'effect': 'ดังนั้น'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_tr": { - "task": "xcopa_tr", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "tr", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'çünkü', 'effect': 'bu yüzden'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_vi": { - "task": "xcopa_vi", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "vi", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'bởi vì', 'effect': 'vì vậy'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_zh": { - "task": "xcopa_zh", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "zh", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': '因为', 'effect': '所以'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "xcopa": "N/A", - "xcopa_et": 1.0, - "xcopa_ht": 1.0, - "xcopa_id": 1.0, - "xcopa_it": 1.0, - "xcopa_qu": 1.0, - "xcopa_sw": 1.0, - "xcopa_ta": 1.0, - "xcopa_th": 1.0, - "xcopa_tr": 1.0, - "xcopa_vi": 1.0, - "xcopa_zh": 1.0 - }, - "n-shot": { - "xcopa": 0, - "xcopa_et": 0, - "xcopa_ht": 0, - "xcopa_id": 0, - "xcopa_it": 0, - "xcopa_qu": 0, - "xcopa_sw": 0, - "xcopa_ta": 0, - "xcopa_th": 0, - "xcopa_tr": 0, - "xcopa_vi": 0, - "xcopa_zh": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=togethercomputer/RedPajama-INCITE-7B-Base,dtype=bfloat16,trust_remote_code=True", - "batch_size": "auto", - "batch_sizes": [ - 16 - ], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "2c0a875" -} \ No newline at end of file diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/RedPajama-INCITE-7B-Base/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log deleted file mode 100644 index 6ba817d18bfdd094700d9b6f26be4194382c195f..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:f9a66a8543cd345001ca68fc23fa022eccb3bfb1e5ab51f497c9c70e925df6c3 -size 198557 diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/RedPajama-INCITE-7B-Base/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 8d6428e3820bdde13e88e1f159896e7d8e14136a..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:84455d99a0f886ce1dc7c7fbfd5a73a3cb9467281241066bb2d1a572c792c3dc -size 6010130 diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/RedPajama-INCITE-7B-Base/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 0e90e1ad200b85b6bd689560a0e31f5d460d49b2..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,548 +0,0 @@ -{ - "results": { - "xnli": { - "acc,none": 0.3827309236947791, - "acc_stderr,none": 0.05194928176239464, - "alias": "xnli" - }, - "xnli_ar": { - "acc,none": 0.344578313253012, - "acc_stderr,none": 0.00952559090011065, - "alias": " - xnli_ar" - }, - "xnli_bg": { - "acc,none": 0.37028112449799194, - "acc_stderr,none": 0.00967891540984029, - "alias": " - xnli_bg" - }, - "xnli_de": { - "acc,none": 0.4461847389558233, - "acc_stderr,none": 0.00996385427413916, - "alias": " - xnli_de" - }, - "xnli_el": { - "acc,none": 0.3481927710843373, - "acc_stderr,none": 0.009548980649153386, - "alias": " - xnli_el" - }, - "xnli_en": { - "acc,none": 0.5357429718875502, - "acc_stderr,none": 0.009996432468510355, - "alias": " - xnli_en" - }, - "xnli_es": { - "acc,none": 0.41967871485943775, - "acc_stderr,none": 0.009891912665432372, - "alias": " - xnli_es" - }, - "xnli_fr": { - "acc,none": 0.46947791164658637, - "acc_stderr,none": 0.010003382355314755, - "alias": " - xnli_fr" - }, - "xnli_hi": { - "acc,none": 0.3405622489959839, - "acc_stderr,none": 0.009498886690274447, - "alias": " - xnli_hi" - }, - "xnli_ru": { - "acc,none": 0.4433734939759036, - "acc_stderr,none": 0.009957592660538648, - "alias": " - xnli_ru" - }, - "xnli_sw": { - "acc,none": 0.3369477911646586, - "acc_stderr,none": 0.009474203778757722, - "alias": " - xnli_sw" - }, - "xnli_th": { - "acc,none": 0.3405622489959839, - "acc_stderr,none": 0.009498886690274442, - "alias": " - xnli_th" - }, - "xnli_tr": { - "acc,none": 0.3437751004016064, - "acc_stderr,none": 0.009520310502882934, - "alias": " - xnli_tr" - }, - "xnli_ur": { - "acc,none": 0.3321285140562249, - "acc_stderr,none": 0.009440328001240636, - "alias": " - xnli_ur" - }, - "xnli_vi": { - "acc,none": 0.3309236947791165, - "acc_stderr,none": 0.009431685461463288, - "alias": " - xnli_vi" - }, - "xnli_zh": { - "acc,none": 0.3385542168674699, - "acc_stderr,none": 0.009485250208516876, - "alias": " - xnli_zh" - } - }, - "groups": { - "xnli": { - "acc,none": 0.3827309236947791, - "acc_stderr,none": 0.05194928176239464, - "alias": "xnli" - } - }, - "configs": { - "xnli_ar": { - "task": "xnli_ar", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "ar", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", صحيح? نعم, \"+hypothesis,premise+\", صحيح? لذا, \"+hypothesis,premise+\", صحيح? رقم, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_bg": { - "task": "xnli_bg", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "bg", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", правилно? да, \"+hypothesis,premise+\", правилно? така, \"+hypothesis,premise+\", правилно? не, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_de": { - "task": "xnli_de", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "de", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", richtig? Ja, \"+hypothesis,premise+\", richtig? Auch, \"+hypothesis,premise+\", richtig? Nein, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_el": { - "task": "xnli_el", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "el", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", σωστός? Ναί, \"+hypothesis,premise+\", σωστός? Έτσι, \"+hypothesis,premise+\", σωστός? όχι, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_en": { - "task": "xnli_en", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "en", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", right? Yes, \"+hypothesis,premise+\", right? Also, \"+hypothesis,premise+\", right? No, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_es": { - "task": "xnli_es", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "es", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", correcto? Sí, \"+hypothesis,premise+\", correcto? Asi que, \"+hypothesis,premise+\", correcto? No, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_fr": { - "task": "xnli_fr", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "fr", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", correct? Oui, \"+hypothesis,premise+\", correct? Aussi, \"+hypothesis,premise+\", correct? Non, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_hi": { - "task": "xnli_hi", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "hi", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", सही? हाँ, \"+hypothesis,premise+\", सही? इसलिए, \"+hypothesis,premise+\", सही? नहीं, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_ru": { - "task": "xnli_ru", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "ru", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", правильно? Да, \"+hypothesis,premise+\", правильно? Так, \"+hypothesis,premise+\", правильно? Нет, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_sw": { - "task": "xnli_sw", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "sw", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", sahihi? Ndiyo, \"+hypothesis,premise+\", sahihi? Hivyo, \"+hypothesis,premise+\", sahihi? Hapana, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_th": { - "task": "xnli_th", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "th", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", ถูกต้อง? ใช่, \"+hypothesis,premise+\", ถูกต้อง? ดังนั้น, \"+hypothesis,premise+\", ถูกต้อง? ไม่, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_tr": { - "task": "xnli_tr", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "tr", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", doğru? Evet, \"+hypothesis,premise+\", doğru? Böylece, \"+hypothesis,premise+\", doğru? Hayır, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_ur": { - "task": "xnli_ur", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "ur", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", صحیح? جی ہاں, \"+hypothesis,premise+\", صحیح? اس لئے, \"+hypothesis,premise+\", صحیح? نہیں, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_vi": { - "task": "xnli_vi", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "vi", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", đúng? Vâng, \"+hypothesis,premise+\", đúng? Vì vậy, \"+hypothesis,premise+\", đúng? Không, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_zh": { - "task": "xnli_zh", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "zh", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", 正确? 是的, \"+hypothesis,premise+\", 正确? 所以, \"+hypothesis,premise+\", 正确? 不是的, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "xnli": "N/A", - "xnli_ar": 1.0, - "xnli_bg": 1.0, - "xnli_de": 1.0, - "xnli_el": 1.0, - "xnli_en": 1.0, - "xnli_es": 1.0, - "xnli_fr": 1.0, - "xnli_hi": 1.0, - "xnli_ru": 1.0, - "xnli_sw": 1.0, - "xnli_th": 1.0, - "xnli_tr": 1.0, - "xnli_ur": 1.0, - "xnli_vi": 1.0, - "xnli_zh": 1.0 - }, - "n-shot": { - "xnli": 0, - "xnli_ar": 0, - "xnli_bg": 0, - "xnli_de": 0, - "xnli_el": 0, - "xnli_en": 0, - "xnli_es": 0, - "xnli_fr": 0, - "xnli_hi": 0, - "xnli_ru": 0, - "xnli_sw": 0, - "xnli_th": 0, - "xnli_tr": 0, - "xnli_ur": 0, - "xnli_vi": 0, - "xnli_zh": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=togethercomputer/RedPajama-INCITE-7B-Base,dtype=bfloat16,trust_remote_code=True", - "batch_size": "auto", - "batch_sizes": [ - 8 - ], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "2c0a875" -} \ No newline at end of file diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/RedPajama-INCITE-7B-Base/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log deleted file mode 100644 index 368f8ceb48f467d0aa1ebca4a106c5e65978d531..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:cfe49c26553fc37efb6944cdf331e11215c8f870ff6c85a92c9f460560f3348b -size 128783 diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/RedPajama-INCITE-7B-Base/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index c9c37fedfb1ee523789009c0318dd3872e1b7329..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:56d33ec3d4d975e330cf2e7f94a5c0309d5c4a0cbbfcbb7405b27a7db626c574 -size 4064576 diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/RedPajama-INCITE-7B-Base/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 0a6c969900eabd1cb0076cd4141e413f56c04a7a..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,423 +0,0 @@ -{ - "results": { - "xstorycloze": { - "acc,none": 0.5524336682510078, - "acc_stderr,none": 0.06791399332607427, - "alias": "xstorycloze" - }, - "xstorycloze_ar": { - "acc,none": 0.4798146922567836, - "acc_stderr,none": 0.01285663570649829, - "alias": " - xstorycloze_ar" - }, - "xstorycloze_en": { - "acc,none": 0.7485109199205824, - "acc_stderr,none": 0.011165293988715807, - "alias": " - xstorycloze_en" - }, - "xstorycloze_es": { - "acc,none": 0.6393117140966248, - "acc_stderr,none": 0.012357592682139025, - "alias": " - xstorycloze_es" - }, - "xstorycloze_eu": { - "acc,none": 0.514228987425546, - "acc_stderr,none": 0.012861913999596127, - "alias": " - xstorycloze_eu" - }, - "xstorycloze_hi": { - "acc,none": 0.513567174056916, - "acc_stderr,none": 0.01286238758665008, - "alias": " - xstorycloze_hi" - }, - "xstorycloze_id": { - "acc,none": 0.513567174056916, - "acc_stderr,none": 0.01286238758665008, - "alias": " - xstorycloze_id" - }, - "xstorycloze_my": { - "acc,none": 0.48974189278623426, - "acc_stderr,none": 0.012864417047980477, - "alias": " - xstorycloze_my" - }, - "xstorycloze_ru": { - "acc,none": 0.5823957643944407, - "acc_stderr,none": 0.012691211382848643, - "alias": " - xstorycloze_ru" - }, - "xstorycloze_sw": { - "acc,none": 0.514228987425546, - "acc_stderr,none": 0.012861913999596127, - "alias": " - xstorycloze_sw" - }, - "xstorycloze_te": { - "acc,none": 0.5327597617471873, - "acc_stderr,none": 0.012839477563855927, - "alias": " - xstorycloze_te" - }, - "xstorycloze_zh": { - "acc,none": 0.5486432825943084, - "acc_stderr,none": 0.0128060889661224, - "alias": " - xstorycloze_zh" - } - }, - "groups": { - "xstorycloze": { - "acc,none": 0.5524336682510078, - "acc_stderr,none": 0.06791399332607427, - "alias": "xstorycloze" - } - }, - "configs": { - "xstorycloze_ar": { - "task": "xstorycloze_ar", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "ar", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_en": { - "task": "xstorycloze_en", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "en", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_es": { - "task": "xstorycloze_es", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "es", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_eu": { - "task": "xstorycloze_eu", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "eu", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_hi": { - "task": "xstorycloze_hi", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "hi", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_id": { - "task": "xstorycloze_id", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "id", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_my": { - "task": "xstorycloze_my", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "my", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_ru": { - "task": "xstorycloze_ru", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "ru", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_sw": { - "task": "xstorycloze_sw", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "sw", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_te": { - "task": "xstorycloze_te", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "te", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_zh": { - "task": "xstorycloze_zh", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "zh", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "xstorycloze": "N/A", - "xstorycloze_ar": 1.0, - "xstorycloze_en": 1.0, - "xstorycloze_es": 1.0, - "xstorycloze_eu": 1.0, - "xstorycloze_hi": 1.0, - "xstorycloze_id": 1.0, - "xstorycloze_my": 1.0, - "xstorycloze_ru": 1.0, - "xstorycloze_sw": 1.0, - "xstorycloze_te": 1.0, - "xstorycloze_zh": 1.0 - }, - "n-shot": { - "xstorycloze": 0, - "xstorycloze_ar": 0, - "xstorycloze_en": 0, - "xstorycloze_es": 0, - "xstorycloze_eu": 0, - "xstorycloze_hi": 0, - "xstorycloze_id": 0, - "xstorycloze_my": 0, - "xstorycloze_ru": 0, - "xstorycloze_sw": 0, - "xstorycloze_te": 0, - "xstorycloze_zh": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=togethercomputer/RedPajama-INCITE-7B-Base,dtype=bfloat16,trust_remote_code=True", - "batch_size": "auto", - "batch_sizes": [ - 4 - ], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "2c0a875" -} \ No newline at end of file diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/RedPajama-INCITE-7B-Base/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log deleted file mode 100644 index c1a40f4fc90cdeb111ab57eb6a19247836ee56f3..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:6e9f5ba42ffa2ea898c14805d192629b5e27ca22869b18fe497db1c273ad30d8 -size 57015 diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/RedPajama-INCITE-7B-Base/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index c89db29f8831c05e450c98e3ae1d301719289331..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e7db16581994b54191b3fb5565ec9226fdb81757903fac634d148eb37ebcd4a0 -size 513604 diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/RedPajama-INCITE-7B-Base/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index bf77332b1292bc5e7e112be21c06088f3d1b3bf7..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,248 +0,0 @@ -{ - "results": { - "xwinograd": { - "acc,none": 0.74331310406833, - "acc_stderr,none": 0.08408219267029617, - "alias": "xwinograd" - }, - "xwinograd_en": { - "acc,none": 0.8718279569892473, - "acc_stderr,none": 0.006934162057729827, - "alias": " - xwinograd_en" - }, - "xwinograd_fr": { - "acc,none": 0.6385542168674698, - "acc_stderr,none": 0.053053439348320096, - "alias": " - xwinograd_fr" - }, - "xwinograd_jp": { - "acc,none": 0.5620437956204379, - "acc_stderr,none": 0.016029414748731596, - "alias": " - xwinograd_jp" - }, - "xwinograd_pt": { - "acc,none": 0.6653992395437263, - "acc_stderr,none": 0.02915103415331038, - "alias": " - xwinograd_pt" - }, - "xwinograd_ru": { - "acc,none": 0.6222222222222222, - "acc_stderr,none": 0.0273606328610564, - "alias": " - xwinograd_ru" - }, - "xwinograd_zh": { - "acc,none": 0.628968253968254, - "acc_stderr,none": 0.02153951426767635, - "alias": " - xwinograd_zh" - } - }, - "groups": { - "xwinograd": { - "acc,none": 0.74331310406833, - "acc_stderr,none": 0.08408219267029617, - "alias": "xwinograd" - } - }, - "configs": { - "xwinograd_en": { - "task": "xwinograd_en", - "group": [ - "xwinograd" - ], - "dataset_path": "Muennighoff/xwinograd", - "dataset_name": "en", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc: Dict) -> int:\n \"\"\"\n Return index of the correct choice.\n\n Note: We are using the \"multiple input\" mode of the multiple-choice\n output-type, which means we use different contexts with the same target\n for the different choices, rather than the same context and different targets.\n \"\"\"\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc: Dict) -> str:\n \"\"\"\n Return the target completion.\n\n Note that this does not depend on the correct choice as we are using\n \"multiple input\" mode.\n \"\"\"\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return the choices that will be used as contexts in \"multiple input\" mode.\"\"\"\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xwinograd_fr": { - "task": "xwinograd_fr", - "group": [ - "xwinograd" - ], - "dataset_path": "Muennighoff/xwinograd", - "dataset_name": "fr", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc: Dict) -> int:\n \"\"\"\n Return index of the correct choice.\n\n Note: We are using the \"multiple input\" mode of the multiple-choice\n output-type, which means we use different contexts with the same target\n for the different choices, rather than the same context and different targets.\n \"\"\"\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc: Dict) -> str:\n \"\"\"\n Return the target completion.\n\n Note that this does not depend on the correct choice as we are using\n \"multiple input\" mode.\n \"\"\"\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return the choices that will be used as contexts in \"multiple input\" mode.\"\"\"\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xwinograd_jp": { - "task": "xwinograd_jp", - "group": [ - "xwinograd" - ], - "dataset_path": "Muennighoff/xwinograd", - "dataset_name": "jp", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc: Dict) -> int:\n \"\"\"\n Return index of the correct choice.\n\n Note: We are using the \"multiple input\" mode of the multiple-choice\n output-type, which means we use different contexts with the same target\n for the different choices, rather than the same context and different targets.\n \"\"\"\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc: Dict) -> str:\n \"\"\"\n Return the target completion.\n\n Note that this does not depend on the correct choice as we are using\n \"multiple input\" mode.\n \"\"\"\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return the choices that will be used as contexts in \"multiple input\" mode.\"\"\"\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xwinograd_pt": { - "task": "xwinograd_pt", - "group": [ - "xwinograd" - ], - "dataset_path": "Muennighoff/xwinograd", - "dataset_name": "pt", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc: Dict) -> int:\n \"\"\"\n Return index of the correct choice.\n\n Note: We are using the \"multiple input\" mode of the multiple-choice\n output-type, which means we use different contexts with the same target\n for the different choices, rather than the same context and different targets.\n \"\"\"\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc: Dict) -> str:\n \"\"\"\n Return the target completion.\n\n Note that this does not depend on the correct choice as we are using\n \"multiple input\" mode.\n \"\"\"\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return the choices that will be used as contexts in \"multiple input\" mode.\"\"\"\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xwinograd_ru": { - "task": "xwinograd_ru", - "group": [ - "xwinograd" - ], - "dataset_path": "Muennighoff/xwinograd", - "dataset_name": "ru", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc: Dict) -> int:\n \"\"\"\n Return index of the correct choice.\n\n Note: We are using the \"multiple input\" mode of the multiple-choice\n output-type, which means we use different contexts with the same target\n for the different choices, rather than the same context and different targets.\n \"\"\"\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc: Dict) -> str:\n \"\"\"\n Return the target completion.\n\n Note that this does not depend on the correct choice as we are using\n \"multiple input\" mode.\n \"\"\"\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return the choices that will be used as contexts in \"multiple input\" mode.\"\"\"\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xwinograd_zh": { - "task": "xwinograd_zh", - "group": [ - "xwinograd" - ], - "dataset_path": "Muennighoff/xwinograd", - "dataset_name": "zh", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc: Dict) -> int:\n \"\"\"\n Return index of the correct choice.\n\n Note: We are using the \"multiple input\" mode of the multiple-choice\n output-type, which means we use different contexts with the same target\n for the different choices, rather than the same context and different targets.\n \"\"\"\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc: Dict) -> str:\n \"\"\"\n Return the target completion.\n\n Note that this does not depend on the correct choice as we are using\n \"multiple input\" mode.\n \"\"\"\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return the choices that will be used as contexts in \"multiple input\" mode.\"\"\"\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "xwinograd": "N/A", - "xwinograd_en": 1.0, - "xwinograd_fr": 1.0, - "xwinograd_jp": 1.0, - "xwinograd_pt": 1.0, - "xwinograd_ru": 1.0, - "xwinograd_zh": 1.0 - }, - "n-shot": { - "xwinograd": 0, - "xwinograd_en": 0, - "xwinograd_fr": 0, - "xwinograd_jp": 0, - "xwinograd_pt": 0, - "xwinograd_ru": 0, - "xwinograd_zh": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=togethercomputer/RedPajama-INCITE-7B-Base,dtype=bfloat16,trust_remote_code=True", - "batch_size": "auto", - "batch_sizes": [ - 16 - ], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "2c0a875" -} \ No newline at end of file diff --git a/lm-eval-output/RedPajama-INCITE-7B-Base/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/RedPajama-INCITE-7B-Base/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log deleted file mode 100644 index 43086821ae0cd4979828e49fbac5b2f82cf68fc6..0000000000000000000000000000000000000000 --- a/lm-eval-output/RedPajama-INCITE-7B-Base/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:6e89625fbf308af1fff3236b77553d7bffe3d2d7c54cab2b132c43f2ff9e0992 -size 40851 diff --git a/lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index de73d62e9560a305786ff7b4c35d36119e7fae1c..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:a7b6162eaecf35e40162218a7076222880e68797aaeb567a8713ff06ca3b6935 -size 681726 diff --git a/lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index b33d98b6026553fd3173065639cbafa7b54c05f7..7c775ac9c53b65decf7f3c333cbc842ebdbc0d94 100644 --- a/lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,33 +1,33 @@ { "results": { "ai2_arc": { - "acc,none": 0.6124577226606539, - "acc_stderr,none": 0.05783349554301207, - "acc_norm,none": 0.5913190529875987, - "acc_norm_stderr,none": 0.04520523788851735, + "acc,none": 0.6141488162344984, + "acc_stderr,none": 0.11602553865195812, + "acc_norm,none": 0.5944193912063134, + "acc_norm_stderr,none": 0.09008281087077372, "alias": "ai2_arc" }, "arc_challenge": { - "acc,none": 0.3677474402730375, - "acc_stderr,none": 0.014090995618168475, - "acc_norm,none": 0.40273037542662116, - "acc_norm_stderr,none": 0.01433223630679014, + "acc,none": 0.36860068259385664, + "acc_stderr,none": 0.014097810678042187, + "acc_norm,none": 0.4044368600682594, + "acc_norm_stderr,none": 0.014342036483436174, "alias": " - arc_challenge" }, "arc_easy": { - "acc,none": 0.7331649831649831, - "acc_stderr,none": 0.009075915859267257, - "acc_norm,none": 0.6843434343434344, - "acc_norm_stderr,none": 0.009537019245566087, + "acc,none": 0.7352693602693603, + "acc_stderr,none": 0.009053021086173977, + "acc_norm,none": 0.6881313131313131, + "acc_norm_stderr,none": 0.00950582334581765, "alias": " - arc_easy" } }, "groups": { "ai2_arc": { - "acc,none": 0.6124577226606539, - "acc_stderr,none": 0.05783349554301207, - "acc_norm,none": 0.5913190529875987, - "acc_norm_stderr,none": 0.04520523788851735, + "acc,none": 0.6141488162344984, + "acc_stderr,none": 0.11602553865195812, + "acc_norm,none": 0.5944193912063134, + "acc_norm_stderr,none": 0.09008281087077372, "alias": "ai2_arc" } }, @@ -118,13 +118,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 32 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index fc5c752e436aeb4bebd1b99d542827ab300f4ca7..a12d3ba914f2bd0c889c2f4ddb5ac093d1696ec5 100644 --- a/lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/ai2_arc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:ceff5a207e0d82cc58530f46cc2f8bbda2dfdf1798d357df795ca30a140fe123 -size 22372 +oid sha256:feda1a9a8d31385965b7cd26c524ddba7445b789e2a118919a275c549d0cea1e +size 16263 diff --git a/lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 506d826ebad28feeb49f9bab0aebfcbea0bf5b6c..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:30daa66193be9558d1965ef1f9bb9c02d4a91213fb3a14fac0879bcf930e1e4b -size 1070250 diff --git a/lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 57072a79a6932c15a540566c3758e5c355ae9677..1e01cbfe43028f325bcaa376e3ca0c30cb12c6fa 100644 --- a/lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,30 +1,30 @@ { "results": { "anli": { - "acc,none": 0.3471875, - "acc_stderr,none": 0.01681366289388001, + "acc,none": 0.3465625, + "acc_stderr,none": 0.016792339011968412, "alias": "anli" }, "anli_r1": { - "acc,none": 0.327, - "acc_stderr,none": 0.014842213153411237, + "acc,none": 0.325, + "acc_stderr,none": 0.014818724459095526, "alias": " - anli_r1" }, "anli_r2": { - "acc,none": 0.36, - "acc_stderr,none": 0.015186527932040122, + "acc,none": 0.356, + "acc_stderr,none": 0.015149042659306626, "alias": " - anli_r2" }, "anli_r3": { - "acc,none": 0.35333333333333333, - "acc_stderr,none": 0.01380457216231493, + "acc,none": 0.3566666666666667, + "acc_stderr,none": 0.013833742805050717, "alias": " - anli_r3" } }, "groups": { "anli": { - "acc,none": 0.3471875, - "acc_stderr,none": 0.01681366289388001, + "acc,none": 0.3465625, + "acc_stderr,none": 0.016792339011968412, "alias": "anli" } }, @@ -147,13 +147,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 32 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index fb3e8afad248d10e9cf231cac51f9074729f2ceb..204acf9a9e5cd7398e0da1d0731161992a65bfc2 100644 --- a/lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/anli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:4e49a85e9b5317ffd4811563ed8a48d3e7219d613f61b93e7c22ca6cd249df70 -size 21092 +oid sha256:df4971e214a122144a0bdc530721b0e62f5d9807f2626826442b773b85ceb849 +size 14676 diff --git a/lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 0cc095cd3ce0290d0e7da69bdec193c4bd392417..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:54f3c8990eb96dc7ac5c4e5586ce6ed5d89feec36cad016d2e19582657216f69 -size 571330 diff --git a/lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 231414ad4d0c8df255f6ce9a5cecf7af391469b9..bd289d7837654744df0d5adb33d8368da6766792 100644 --- a/lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,13 +1,13 @@ { "results": { "arithmetic": { - "acc,none": 0.0069, - "acc_stderr,none": 0.006602463150804194, + "acc,none": 0.007, + "acc_stderr,none": 0.006724488098523242, "alias": "arithmetic" }, "arithmetic_1dc": { - "acc,none": 0.005, - "acc_stderr,none": 0.0015775754727385474, + "acc,none": 0.0065, + "acc_stderr,none": 0.0017973564602277768, "alias": " - arithmetic_1dc" }, "arithmetic_2da": { @@ -16,8 +16,8 @@ "alias": " - arithmetic_2da" }, "arithmetic_2dm": { - "acc,none": 0.0295, - "acc_stderr,none": 0.0037844465933618916, + "acc,none": 0.029, + "acc_stderr,none": 0.0037532044004605246, "alias": " - arithmetic_2dm" }, "arithmetic_2ds": { @@ -58,8 +58,8 @@ }, "groups": { "arithmetic": { - "acc,none": 0.0069, - "acc_stderr,none": 0.006602463150804194, + "acc,none": 0.007, + "acc_stderr,none": 0.006724488098523242, "alias": "arithmetic" } }, @@ -364,13 +364,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 4466c24606845c1793dd0f99d4f8fa64d3b4b334..c447f419a73b682c70ad6205bda842e01660fbd1 100644 --- a/lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/arithmetic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:bc302e5d00e6873c97ee6f5549a0c1cb9f76a06956b6cb32d874a17c6b1e548b -size 32726 +oid sha256:336d5042f526e0b5a3e6045ddb2334d66ffc6ba7c4f85d38364256c924609891 +size 25619 diff --git a/lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index d4dd7a6bb1959154025f7c87d8ec31819694252d..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4e36726446be44a4ca03fdd234a83b2cf7fb9884d27cb2ab737c41c9d9dae9f5 -size 571331 diff --git a/lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index e5b0de5186b8e79336b8c87ef8d483960c22ed89..2bd3f1c94dbcdc533a930023a3c35678fe2119ae 100644 --- a/lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -36,8 +36,8 @@ "alias": "arithmetic_2ds" }, "arithmetic_2dm": { - "acc,none": 0.0295, - "acc_stderr,none": 0.0037844465933618916, + "acc,none": 0.029, + "acc_stderr,none": 0.0037532044004605246, "alias": "arithmetic_2dm" }, "arithmetic_2da": { @@ -46,8 +46,8 @@ "alias": "arithmetic_2da" }, "arithmetic_1dc": { - "acc,none": 0.005, - "acc_stderr,none": 0.0015775754727385474, + "acc,none": 0.0065, + "acc_stderr,none": 0.0017973564602277768, "alias": "arithmetic_1dc" } }, @@ -350,13 +350,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 2f4d2965555bbbfbadc249d14551d9512dd68e07..723a36c37c0bdb403ba8b9db34472333afb2a3c4 100644 --- a/lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/arithmetic__/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:bf64ee5e732687e94dbcef3284174f1ce5468651d74df1c89e59d52fbbc17d0e -size 33636 +oid sha256:c280fce1e1826f134c773e5a830152245f71f8505f03d4c5cfcb4af0ff0f9df8 +size 21272 diff --git a/lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 1d296a62ab4603ab4aed0fae3a6e50eda009e871..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:8d9004ca282c30e0451d79151305ebe50933d491576bde59f97ec09f2d20999a -size 265321 diff --git a/lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index e66cd979580bd8e5214bc4d0db2a78a2b2b5d8b0..867eefb505afe86a69608c01520a40e97a20c697 100644 --- a/lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "asdiv": { - "acc,none": 0.016919739696312365, - "acc_stderr,none": 0.002686891250897643, + "acc,none": 0.015618221258134491, + "acc_stderr,none": 0.002583189883690767, "alias": "asdiv" } }, @@ -41,13 +41,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index feb4c1d5154af2b86f6ab4a36a354d553c0bb407..7980654b39d93cee66d43f456306312507d3697f 100644 --- a/lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/asdiv/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:5bee59ce6f91fa784a9ec22efa5da1a1a0a6720cb9c2a9459dd0ddeda079b8b5 -size 15826 +oid sha256:bc981690a7e19989d96bb07b4c91f750b1d9921eca7d50386cdc5233a531770c +size 16390 diff --git a/lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index bbbde541dae610429f57bcf193ca326827575107..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:bd87668d838a678fe98acd570f4f888ebadc3a001ba22cf884b87023aceebbc5 -size 4244097 diff --git a/lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index cb9e1b27989636a46e59969bc5984fa92d712d8d..8d69029a7054e2f62c93f4403720a49d881dcabf 100644 --- a/lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,13 +1,13 @@ { "results": { "blimp": { - "acc,none": 0.8318358208955224, - "acc_stderr,none": 0.15243408065668698, + "acc,none": 0.8316865671641791, + "acc_stderr,none": 0.1603584899107365, "alias": "blimp" }, "blimp_adjunct_island": { - "acc,none": 0.902, - "acc_stderr,none": 0.009406619184621223, + "acc,none": 0.904, + "acc_stderr,none": 0.009320454434783215, "alias": " - blimp_adjunct_island" }, "blimp_anaphor_gender_agreement": { @@ -16,23 +16,23 @@ "alias": " - blimp_anaphor_gender_agreement" }, "blimp_anaphor_number_agreement": { - "acc,none": 0.992, - "acc_stderr,none": 0.0028185003005045057, + "acc,none": 0.994, + "acc_stderr,none": 0.0024433521993298428, "alias": " - blimp_anaphor_number_agreement" }, "blimp_animate_subject_passive": { - "acc,none": 0.809, - "acc_stderr,none": 0.012436787112179491, + "acc,none": 0.807, + "acc_stderr,none": 0.012486268734370145, "alias": " - blimp_animate_subject_passive" }, "blimp_animate_subject_trans": { - "acc,none": 0.911, - "acc_stderr,none": 0.009008893392651514, + "acc,none": 0.913, + "acc_stderr,none": 0.008916866630745918, "alias": " - blimp_animate_subject_trans" }, "blimp_causative": { - "acc,none": 0.743, - "acc_stderr,none": 0.013825416526895047, + "acc,none": 0.728, + "acc_stderr,none": 0.014078856992462623, "alias": " - blimp_causative" }, "blimp_complex_NP_island": { @@ -41,113 +41,113 @@ "alias": " - blimp_complex_NP_island" }, "blimp_coordinate_structure_constraint_complex_left_branch": { - "acc,none": 0.821, - "acc_stderr,none": 0.012128730605719113, + "acc,none": 0.82, + "acc_stderr,none": 0.012155153135511949, "alias": " - blimp_coordinate_structure_constraint_complex_left_branch" }, "blimp_coordinate_structure_constraint_object_extraction": { - "acc,none": 0.889, - "acc_stderr,none": 0.009938701010583726, + "acc,none": 0.891, + "acc_stderr,none": 0.009859828407037188, "alias": " - blimp_coordinate_structure_constraint_object_extraction" }, "blimp_determiner_noun_agreement_1": { "acc,none": 0.986, - "acc_stderr,none": 0.0037172325482565916, + "acc_stderr,none": 0.0037172325482565877, "alias": " - blimp_determiner_noun_agreement_1" }, "blimp_determiner_noun_agreement_2": { - "acc,none": 0.975, - "acc_stderr,none": 0.0049395748196984545, + "acc,none": 0.973, + "acc_stderr,none": 0.005128089049275288, "alias": " - blimp_determiner_noun_agreement_2" }, "blimp_determiner_noun_agreement_irregular_1": { - "acc,none": 0.943, - "acc_stderr,none": 0.0073351758537068355, + "acc,none": 0.933, + "acc_stderr,none": 0.00791034598317755, "alias": " - blimp_determiner_noun_agreement_irregular_1" }, "blimp_determiner_noun_agreement_irregular_2": { - "acc,none": 0.951, - "acc_stderr,none": 0.006829761756140914, + "acc,none": 0.954, + "acc_stderr,none": 0.006627814717380719, "alias": " - blimp_determiner_noun_agreement_irregular_2" }, "blimp_determiner_noun_agreement_with_adj_2": { - "acc,none": 0.948, - "acc_stderr,none": 0.007024624213817146, + "acc,none": 0.95, + "acc_stderr,none": 0.0068954729748979, "alias": " - blimp_determiner_noun_agreement_with_adj_2" }, "blimp_determiner_noun_agreement_with_adj_irregular_1": { - "acc,none": 0.882, - "acc_stderr,none": 0.01020686926438179, + "acc,none": 0.879, + "acc_stderr,none": 0.010318210380946088, "alias": " - blimp_determiner_noun_agreement_with_adj_irregular_1" }, "blimp_determiner_noun_agreement_with_adj_irregular_2": { - "acc,none": 0.929, - "acc_stderr,none": 0.008125578442487916, + "acc,none": 0.931, + "acc_stderr,none": 0.00801893405031516, "alias": " - blimp_determiner_noun_agreement_with_adj_irregular_2" }, "blimp_determiner_noun_agreement_with_adjective_1": { - "acc,none": 0.971, - "acc_stderr,none": 0.005309160685756979, + "acc,none": 0.973, + "acc_stderr,none": 0.00512808904927529, "alias": " - blimp_determiner_noun_agreement_with_adjective_1" }, "blimp_distractor_agreement_relational_noun": { - "acc,none": 0.919, - "acc_stderr,none": 0.008632121032139986, + "acc,none": 0.923, + "acc_stderr,none": 0.008434580140240648, "alias": " - blimp_distractor_agreement_relational_noun" }, "blimp_distractor_agreement_relative_clause": { - "acc,none": 0.705, - "acc_stderr,none": 0.014428554438445523, + "acc,none": 0.717, + "acc_stderr,none": 0.014251810906481728, "alias": " - blimp_distractor_agreement_relative_clause" }, "blimp_drop_argument": { - "acc,none": 0.746, - "acc_stderr,none": 0.013772206565168537, + "acc,none": 0.747, + "acc_stderr,none": 0.01375427861358708, "alias": " - blimp_drop_argument" }, "blimp_ellipsis_n_bar_1": { "acc,none": 0.812, - "acc_stderr,none": 0.012361586015103749, + "acc_stderr,none": 0.012361586015103744, "alias": " - blimp_ellipsis_n_bar_1" }, "blimp_ellipsis_n_bar_2": { - "acc,none": 0.948, - "acc_stderr,none": 0.0070246242138171325, + "acc,none": 0.949, + "acc_stderr,none": 0.006960420062571421, "alias": " - blimp_ellipsis_n_bar_2" }, "blimp_existential_there_object_raising": { - "acc,none": 0.859, - "acc_stderr,none": 0.011010914595992438, + "acc,none": 0.864, + "acc_stderr,none": 0.010845350230472988, "alias": " - blimp_existential_there_object_raising" }, "blimp_existential_there_quantifiers_1": { - "acc,none": 0.984, - "acc_stderr,none": 0.0039698563903194225, + "acc,none": 0.985, + "acc_stderr,none": 0.0038457495745030067, "alias": " - blimp_existential_there_quantifiers_1" }, "blimp_existential_there_quantifiers_2": { "acc,none": 0.377, - "acc_stderr,none": 0.01533317012577985, + "acc_stderr,none": 0.015333170125779847, "alias": " - blimp_existential_there_quantifiers_2" }, "blimp_existential_there_subject_raising": { - "acc,none": 0.909, - "acc_stderr,none": 0.009099549538400252, + "acc,none": 0.911, + "acc_stderr,none": 0.009008893392651523, "alias": " - blimp_existential_there_subject_raising" }, "blimp_expletive_it_object_raising": { - "acc,none": 0.827, - "acc_stderr,none": 0.011967214137559929, + "acc,none": 0.826, + "acc_stderr,none": 0.01199449323097343, "alias": " - blimp_expletive_it_object_raising" }, "blimp_inchoative": { - "acc,none": 0.678, - "acc_stderr,none": 0.014782913600996662, + "acc,none": 0.68, + "acc_stderr,none": 0.014758652303574874, "alias": " - blimp_inchoative" }, "blimp_intransitive": { - "acc,none": 0.8, - "acc_stderr,none": 0.012655439943366651, + "acc,none": 0.791, + "acc_stderr,none": 0.012864077288499321, "alias": " - blimp_intransitive" }, "blimp_irregular_past_participle_adjectives": { @@ -156,68 +156,68 @@ "alias": " - blimp_irregular_past_participle_adjectives" }, "blimp_irregular_past_participle_verbs": { - "acc,none": 0.912, - "acc_stderr,none": 0.008963053962592081, + "acc,none": 0.906, + "acc_stderr,none": 0.009233052000787736, "alias": " - blimp_irregular_past_participle_verbs" }, "blimp_irregular_plural_subject_verb_agreement_1": { - "acc,none": 0.93, - "acc_stderr,none": 0.008072494358323508, + "acc,none": 0.928, + "acc_stderr,none": 0.008178195576218681, "alias": " - blimp_irregular_plural_subject_verb_agreement_1" }, "blimp_irregular_plural_subject_verb_agreement_2": { - "acc,none": 0.937, - "acc_stderr,none": 0.007687007876286417, + "acc,none": 0.932, + "acc_stderr,none": 0.007964887911291603, "alias": " - blimp_irregular_plural_subject_verb_agreement_2" }, "blimp_left_branch_island_echo_question": { - "acc,none": 0.633, - "acc_stderr,none": 0.015249378464171745, + "acc,none": 0.648, + "acc_stderr,none": 0.015110404505648661, "alias": " - blimp_left_branch_island_echo_question" }, "blimp_left_branch_island_simple_question": { - "acc,none": 0.909, - "acc_stderr,none": 0.009099549538400227, + "acc,none": 0.911, + "acc_stderr,none": 0.009008893392651526, "alias": " - blimp_left_branch_island_simple_question" }, "blimp_matrix_question_npi_licensor_present": { - "acc,none": 0.617, - "acc_stderr,none": 0.015380102325652713, + "acc,none": 0.607, + "acc_stderr,none": 0.015452824654081496, "alias": " - blimp_matrix_question_npi_licensor_present" }, "blimp_npi_present_1": { - "acc,none": 0.671, - "acc_stderr,none": 0.014865395385928362, + "acc,none": 0.674, + "acc_stderr,none": 0.014830507204541038, "alias": " - blimp_npi_present_1" }, "blimp_npi_present_2": { - "acc,none": 0.735, - "acc_stderr,none": 0.013963164754809953, + "acc,none": 0.73, + "acc_stderr,none": 0.014046255632633915, "alias": " - blimp_npi_present_2" }, "blimp_only_npi_licensor_present": { - "acc,none": 0.971, - "acc_stderr,none": 0.005309160685756993, + "acc,none": 0.974, + "acc_stderr,none": 0.005034813735318216, "alias": " - blimp_only_npi_licensor_present" }, "blimp_only_npi_scope": { - "acc,none": 0.678, - "acc_stderr,none": 0.014782913600996685, + "acc,none": 0.706, + "acc_stderr,none": 0.01441429054000822, "alias": " - blimp_only_npi_scope" }, "blimp_passive_1": { - "acc,none": 0.897, - "acc_stderr,none": 0.0096168333396958, + "acc,none": 0.895, + "acc_stderr,none": 0.009698921026024971, "alias": " - blimp_passive_1" }, "blimp_passive_2": { - "acc,none": 0.909, - "acc_stderr,none": 0.009099549538400236, + "acc,none": 0.906, + "acc_stderr,none": 0.009233052000787728, "alias": " - blimp_passive_2" }, "blimp_principle_A_c_command": { - "acc,none": 0.74, - "acc_stderr,none": 0.013877773329774164, + "acc,none": 0.741, + "acc_stderr,none": 0.01386041525752791, "alias": " - blimp_principle_A_c_command" }, "blimp_principle_A_case_1": { @@ -226,23 +226,23 @@ "alias": " - blimp_principle_A_case_1" }, "blimp_principle_A_case_2": { - "acc,none": 0.967, - "acc_stderr,none": 0.0056518088204523705, + "acc,none": 0.963, + "acc_stderr,none": 0.005972157622389646, "alias": " - blimp_principle_A_case_2" }, "blimp_principle_A_domain_1": { - "acc,none": 0.998, - "acc_stderr,none": 0.0014135055705578, + "acc,none": 0.999, + "acc_stderr,none": 0.0010000000000000124, "alias": " - blimp_principle_A_domain_1" }, "blimp_principle_A_domain_2": { - "acc,none": 0.84, - "acc_stderr,none": 0.011598902298689004, + "acc,none": 0.835, + "acc_stderr,none": 0.011743632866916164, "alias": " - blimp_principle_A_domain_2" }, "blimp_principle_A_domain_3": { - "acc,none": 0.736, - "acc_stderr,none": 0.013946271849440467, + "acc,none": 0.739, + "acc_stderr,none": 0.013895037677965136, "alias": " - blimp_principle_A_domain_3" }, "blimp_principle_A_reconstruction": { @@ -251,13 +251,13 @@ "alias": " - blimp_principle_A_reconstruction" }, "blimp_regular_plural_subject_verb_agreement_1": { - "acc,none": 0.967, - "acc_stderr,none": 0.005651808820452372, + "acc,none": 0.965, + "acc_stderr,none": 0.005814534272734963, "alias": " - blimp_regular_plural_subject_verb_agreement_1" }, "blimp_regular_plural_subject_verb_agreement_2": { - "acc,none": 0.935, - "acc_stderr,none": 0.007799733061832028, + "acc,none": 0.931, + "acc_stderr,none": 0.008018934050315146, "alias": " - blimp_regular_plural_subject_verb_agreement_2" }, "blimp_sentential_negation_npi_licensor_present": { @@ -266,73 +266,73 @@ "alias": " - blimp_sentential_negation_npi_licensor_present" }, "blimp_sentential_negation_npi_scope": { - "acc,none": 0.761, - "acc_stderr,none": 0.013493000446937591, + "acc,none": 0.759, + "acc_stderr,none": 0.013531522534515419, "alias": " - blimp_sentential_negation_npi_scope" }, "blimp_sentential_subject_island": { - "acc,none": 0.562, - "acc_stderr,none": 0.01569721001969469, + "acc,none": 0.559, + "acc_stderr,none": 0.01570877989424268, "alias": " - blimp_sentential_subject_island" }, "blimp_superlative_quantifiers_1": { - "acc,none": 0.897, - "acc_stderr,none": 0.009616833339695784, + "acc,none": 0.892, + "acc_stderr,none": 0.009820001651345714, "alias": " - blimp_superlative_quantifiers_1" }, "blimp_superlative_quantifiers_2": { - "acc,none": 0.881, - "acc_stderr,none": 0.010244215145336666, + "acc,none": 0.877, + "acc_stderr,none": 0.010391293421849879, "alias": " - blimp_superlative_quantifiers_2" }, "blimp_tough_vs_raising_1": { - "acc,none": 0.669, - "acc_stderr,none": 0.014888272588203934, + "acc,none": 0.663, + "acc_stderr,none": 0.014955087918653603, "alias": " - blimp_tough_vs_raising_1" }, "blimp_tough_vs_raising_2": { - "acc,none": 0.856, - "acc_stderr,none": 0.01110798754893915, + "acc,none": 0.853, + "acc_stderr,none": 0.011203415395160335, "alias": " - blimp_tough_vs_raising_2" }, "blimp_transitive": { - "acc,none": 0.862, - "acc_stderr,none": 0.010912152632504378, + "acc,none": 0.867, + "acc_stderr,none": 0.010743669132397346, "alias": " - blimp_transitive" }, "blimp_wh_island": { - "acc,none": 0.879, - "acc_stderr,none": 0.010318210380946085, + "acc,none": 0.877, + "acc_stderr,none": 0.010391293421849877, "alias": " - blimp_wh_island" }, "blimp_wh_questions_object_gap": { - "acc,none": 0.84, - "acc_stderr,none": 0.011598902298689004, + "acc,none": 0.841, + "acc_stderr,none": 0.011569479368271296, "alias": " - blimp_wh_questions_object_gap" }, "blimp_wh_questions_subject_gap": { - "acc,none": 0.931, - "acc_stderr,none": 0.00801893405031516, + "acc,none": 0.933, + "acc_stderr,none": 0.007910345983177547, "alias": " - blimp_wh_questions_subject_gap" }, "blimp_wh_questions_subject_gap_long_distance": { - "acc,none": 0.919, - "acc_stderr,none": 0.008632121032139969, + "acc,none": 0.922, + "acc_stderr,none": 0.008484573530118588, "alias": " - blimp_wh_questions_subject_gap_long_distance" }, "blimp_wh_vs_that_no_gap": { - "acc,none": 0.973, - "acc_stderr,none": 0.005128089049275289, + "acc,none": 0.974, + "acc_stderr,none": 0.0050348137353182325, "alias": " - blimp_wh_vs_that_no_gap" }, "blimp_wh_vs_that_no_gap_long_distance": { - "acc,none": 0.961, - "acc_stderr,none": 0.006125072776426141, + "acc,none": 0.962, + "acc_stderr,none": 0.006049181150584946, "alias": " - blimp_wh_vs_that_no_gap_long_distance" }, "blimp_wh_vs_that_with_gap": { - "acc,none": 0.349, - "acc_stderr,none": 0.015080663991563098, + "acc,none": 0.341, + "acc_stderr,none": 0.0149981313484027, "alias": " - blimp_wh_vs_that_with_gap" }, "blimp_wh_vs_that_with_gap_long_distance": { @@ -343,8 +343,8 @@ }, "groups": { "blimp": { - "acc,none": 0.8318358208955224, - "acc_stderr,none": 0.15243408065668698, + "acc,none": 0.8316865671641791, + "acc_stderr,none": 0.1603584899107365, "alias": "blimp" } }, @@ -2235,13 +2235,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index e2c5072c30ef1b04dfe976ae4dc2c43dd9eeb3d0..4ec8d33f54fb22f8288116601ec34a59193bd5dc 100644 --- a/lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/blimp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:e03fbfd3dd884d95a74c39c0d75f2436728da1a89c09807a16b28dab309ec385 -size 355468 +oid sha256:2a3b311b45000c90305578cb19f065bec59f430e131dab0963128cb73e9786b4 +size 294489 diff --git a/lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index e45f8f4578cd0c221372edfe2133615999adb232..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:7d9982cba484eea117460519dd137bac296cc134391704e329d1546044840296 -size 1134661 diff --git a/lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 284f8f4e7dcd1018e6a257d42eab216353ba8ca8..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,60 +0,0 @@ -{ - "results": { - "boolq": { - "acc,none": 0.7253822629969419, - "acc_stderr,none": 0.007806211211206189, - "alias": "boolq" - } - }, - "configs": { - "boolq": { - "task": "boolq", - "group": [ - "super-glue-lm-eval-v1" - ], - "dataset_path": "super_glue", - "dataset_name": "boolq", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "{{passage}}\nQuestion: {{question}}?\nAnswer:", - "doc_to_target": "label", - "doc_to_choice": [ - "no", - "yes" - ], - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "passage", - "metadata": { - "version": 2.0 - } - } - }, - "versions": { - "boolq": 2.0 - }, - "n-shot": { - "boolq": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index cf74ff91b5aaa08a1d5d8ff6cb995f2e4f6d81ce..c3c38e7adf946a7820eb3628a31d5e99b173e880 100644 --- a/lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/boolq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:c16a39eb1a979a8a5c5f87ccfccba887e9bd37bc7710687ab9db30269c8f0018 -size 19802 +oid sha256:444fe6f0cbaf5443ee1dfa05e3d4f1806c4556054ceabcfea74b2c4eb6ee803a +size 21711 diff --git a/lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index ec4a7cfb007ab98354fed3571d3544b304a4a9a2..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:cf8a16d333c03f7ae4e6f988588806b8dc0483568c8f000306fd1e55cc7779ce -size 13905 diff --git a/lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index df5d9251de2174186998afec84d35f49e062d0ef..f90d01ff108b9c7d88b3bbf662d241c6dae95312 100644 --- a/lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -54,13 +54,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 16 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index ee6c769c51c39192b8a177f923b2d12e530edd73..1c91103bd410c1a6897e45c3025e622f9f773837 100644 --- a/lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/cb/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:83fbfd9e158b5fbf92b40a3b0ac21aa5830e033698a6e6fbd1a0c00a82f4e0a3 -size 13372 +oid sha256:239d0b8aec1eddb6f22bc82f54f5fa42e3a45de06aff4fc8e0aef720286f58fe +size 14061 diff --git a/lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 57bb70382f0c40630c023b8630caf0d72e1c0dbf..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:755b95da7d45b3a97ad51d88b7a6f6ab085c43413f8d7099119bf26337c959e1 -size 321200 diff --git a/lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 0d3221df7701d072ed5971f0e1fec6cd42c95876..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,2588 +0,0 @@ -{ - "results": { - "ceval-valid": { - "acc,none": 0.24962852897473997, - "acc_stderr,none": 0.11387085890117267, - "acc_norm,none": 0.24962852897473997, - "acc_norm_stderr,none": 0.11387085890117267, - "alias": "ceval-valid" - }, - "ceval-valid_accountant": { - "acc,none": 0.22448979591836735, - "acc_stderr,none": 0.06022425581505364, - "acc_norm,none": 0.22448979591836735, - "acc_norm_stderr,none": 0.06022425581505364, - "alias": " - ceval-valid_accountant" - }, - "ceval-valid_advanced_mathematics": { - "acc,none": 0.10526315789473684, - "acc_stderr,none": 0.0723351864143449, - "acc_norm,none": 0.10526315789473684, - "acc_norm_stderr,none": 0.0723351864143449, - "alias": " - ceval-valid_advanced_mathematics" - }, - "ceval-valid_art_studies": { - "acc,none": 0.30303030303030304, - "acc_stderr,none": 0.08124094920275463, - "acc_norm,none": 0.30303030303030304, - "acc_norm_stderr,none": 0.08124094920275463, - "alias": " - ceval-valid_art_studies" - }, - "ceval-valid_basic_medicine": { - "acc,none": 0.15789473684210525, - "acc_stderr,none": 0.08594700851870798, - "acc_norm,none": 0.15789473684210525, - "acc_norm_stderr,none": 0.08594700851870798, - "alias": " - ceval-valid_basic_medicine" - }, - "ceval-valid_business_administration": { - "acc,none": 0.21212121212121213, - "acc_stderr,none": 0.07226812131946557, - "acc_norm,none": 0.21212121212121213, - "acc_norm_stderr,none": 0.07226812131946557, - "alias": " - ceval-valid_business_administration" - }, - "ceval-valid_chinese_language_and_literature": { - "acc,none": 0.2608695652173913, - "acc_stderr,none": 0.09361833424764436, - "acc_norm,none": 0.2608695652173913, - "acc_norm_stderr,none": 0.09361833424764436, - "alias": " - ceval-valid_chinese_language_and_literature" - }, - "ceval-valid_civil_servant": { - "acc,none": 0.2127659574468085, - "acc_stderr,none": 0.060342609647735204, - "acc_norm,none": 0.2127659574468085, - "acc_norm_stderr,none": 0.060342609647735204, - "alias": " - ceval-valid_civil_servant" - }, - "ceval-valid_clinical_medicine": { - "acc,none": 0.18181818181818182, - "acc_stderr,none": 0.08416546361568647, - "acc_norm,none": 0.18181818181818182, - "acc_norm_stderr,none": 0.08416546361568647, - "alias": " - ceval-valid_clinical_medicine" - }, - "ceval-valid_college_chemistry": { - "acc,none": 0.25, - "acc_stderr,none": 0.09028938981432691, - "acc_norm,none": 0.25, - "acc_norm_stderr,none": 0.09028938981432691, - "alias": " - ceval-valid_college_chemistry" - }, - "ceval-valid_college_economics": { - "acc,none": 0.32727272727272727, - "acc_stderr,none": 0.0638524469869863, - "acc_norm,none": 0.32727272727272727, - "acc_norm_stderr,none": 0.0638524469869863, - "alias": " - ceval-valid_college_economics" - }, - "ceval-valid_college_physics": { - "acc,none": 0.21052631578947367, - "acc_stderr,none": 0.0960916767552923, - "acc_norm,none": 0.21052631578947367, - "acc_norm_stderr,none": 0.0960916767552923, - "alias": " - ceval-valid_college_physics" - }, - "ceval-valid_college_programming": { - "acc,none": 0.24324324324324326, - "acc_stderr,none": 0.07150679219093488, - "acc_norm,none": 0.24324324324324326, - "acc_norm_stderr,none": 0.07150679219093488, - "alias": " - ceval-valid_college_programming" - }, - "ceval-valid_computer_architecture": { - "acc,none": 0.2857142857142857, - "acc_stderr,none": 0.10101525445522108, - "acc_norm,none": 0.2857142857142857, - "acc_norm_stderr,none": 0.10101525445522108, - "alias": " - ceval-valid_computer_architecture" - }, - "ceval-valid_computer_network": { - "acc,none": 0.42105263157894735, - "acc_stderr,none": 0.11637279966159299, - "acc_norm,none": 0.42105263157894735, - "acc_norm_stderr,none": 0.11637279966159299, - "alias": " - ceval-valid_computer_network" - }, - "ceval-valid_discrete_mathematics": { - "acc,none": 0.1875, - "acc_stderr,none": 0.10077822185373188, - "acc_norm,none": 0.1875, - "acc_norm_stderr,none": 0.10077822185373188, - "alias": " - ceval-valid_discrete_mathematics" - }, - "ceval-valid_education_science": { - "acc,none": 0.2413793103448276, - "acc_stderr,none": 0.080869237238335, - "acc_norm,none": 0.2413793103448276, - "acc_norm_stderr,none": 0.080869237238335, - "alias": " - ceval-valid_education_science" - }, - "ceval-valid_electrical_engineer": { - "acc,none": 0.35135135135135137, - "acc_stderr,none": 0.0795654132101608, - "acc_norm,none": 0.35135135135135137, - "acc_norm_stderr,none": 0.0795654132101608, - "alias": " - ceval-valid_electrical_engineer" - }, - "ceval-valid_environmental_impact_assessment_engineer": { - "acc,none": 0.25806451612903225, - "acc_stderr,none": 0.07988892740217941, - "acc_norm,none": 0.25806451612903225, - "acc_norm_stderr,none": 0.07988892740217941, - "alias": " - ceval-valid_environmental_impact_assessment_engineer" - }, - "ceval-valid_fire_engineer": { - "acc,none": 0.22580645161290322, - "acc_stderr,none": 0.07633651333031763, - "acc_norm,none": 0.22580645161290322, - "acc_norm_stderr,none": 0.07633651333031763, - "alias": " - ceval-valid_fire_engineer" - }, - "ceval-valid_high_school_biology": { - "acc,none": 0.21052631578947367, - "acc_stderr,none": 0.0960916767552923, - "acc_norm,none": 0.21052631578947367, - "acc_norm_stderr,none": 0.0960916767552923, - "alias": " - ceval-valid_high_school_biology" - }, - "ceval-valid_high_school_chemistry": { - "acc,none": 0.3684210526315789, - "acc_stderr,none": 0.1136972052352256, - "acc_norm,none": 0.3684210526315789, - "acc_norm_stderr,none": 0.1136972052352256, - "alias": " - ceval-valid_high_school_chemistry" - }, - "ceval-valid_high_school_chinese": { - "acc,none": 0.10526315789473684, - "acc_stderr,none": 0.0723351864143449, - "acc_norm,none": 0.10526315789473684, - "acc_norm_stderr,none": 0.0723351864143449, - "alias": " - ceval-valid_high_school_chinese" - }, - "ceval-valid_high_school_geography": { - "acc,none": 0.3157894736842105, - "acc_stderr,none": 0.10956136839295434, - "acc_norm,none": 0.3157894736842105, - "acc_norm_stderr,none": 0.10956136839295434, - "alias": " - ceval-valid_high_school_geography" - }, - "ceval-valid_high_school_history": { - "acc,none": 0.15, - "acc_stderr,none": 0.0819178021909125, - "acc_norm,none": 0.15, - "acc_norm_stderr,none": 0.0819178021909125, - "alias": " - ceval-valid_high_school_history" - }, - "ceval-valid_high_school_mathematics": { - "acc,none": 0.2777777777777778, - "acc_stderr,none": 0.1086324845659782, - "acc_norm,none": 0.2777777777777778, - "acc_norm_stderr,none": 0.1086324845659782, - "alias": " - ceval-valid_high_school_mathematics" - }, - "ceval-valid_high_school_physics": { - "acc,none": 0.3157894736842105, - "acc_stderr,none": 0.10956136839295433, - "acc_norm,none": 0.3157894736842105, - "acc_norm_stderr,none": 0.10956136839295433, - "alias": " - ceval-valid_high_school_physics" - }, - "ceval-valid_high_school_politics": { - "acc,none": 0.10526315789473684, - "acc_stderr,none": 0.07233518641434492, - "acc_norm,none": 0.10526315789473684, - "acc_norm_stderr,none": 0.07233518641434492, - "alias": " - ceval-valid_high_school_politics" - }, - "ceval-valid_ideological_and_moral_cultivation": { - "acc,none": 0.15789473684210525, - "acc_stderr,none": 0.08594700851870798, - "acc_norm,none": 0.15789473684210525, - "acc_norm_stderr,none": 0.08594700851870798, - "alias": " - ceval-valid_ideological_and_moral_cultivation" - }, - "ceval-valid_law": { - "acc,none": 0.3333333333333333, - "acc_stderr,none": 0.0982946374365981, - "acc_norm,none": 0.3333333333333333, - "acc_norm_stderr,none": 0.0982946374365981, - "alias": " - ceval-valid_law" - }, - "ceval-valid_legal_professional": { - "acc,none": 0.17391304347826086, - "acc_stderr,none": 0.08081046758996391, - "acc_norm,none": 0.17391304347826086, - "acc_norm_stderr,none": 0.08081046758996391, - "alias": " - ceval-valid_legal_professional" - }, - "ceval-valid_logic": { - "acc,none": 0.22727272727272727, - "acc_stderr,none": 0.09144861547306321, - "acc_norm,none": 0.22727272727272727, - "acc_norm_stderr,none": 0.09144861547306321, - "alias": " - ceval-valid_logic" - }, - "ceval-valid_mao_zedong_thought": { - "acc,none": 0.2916666666666667, - "acc_stderr,none": 0.09477598811252415, - "acc_norm,none": 0.2916666666666667, - "acc_norm_stderr,none": 0.09477598811252415, - "alias": " - ceval-valid_mao_zedong_thought" - }, - "ceval-valid_marxism": { - "acc,none": 0.21052631578947367, - "acc_stderr,none": 0.0960916767552923, - "acc_norm,none": 0.21052631578947367, - "acc_norm_stderr,none": 0.0960916767552923, - "alias": " - ceval-valid_marxism" - }, - "ceval-valid_metrology_engineer": { - "acc,none": 0.25, - "acc_stderr,none": 0.09028938981432691, - "acc_norm,none": 0.25, - "acc_norm_stderr,none": 0.09028938981432691, - "alias": " - ceval-valid_metrology_engineer" - }, - "ceval-valid_middle_school_biology": { - "acc,none": 0.23809523809523808, - "acc_stderr,none": 0.09523809523809523, - "acc_norm,none": 0.23809523809523808, - "acc_norm_stderr,none": 0.09523809523809523, - "alias": " - ceval-valid_middle_school_biology" - }, - "ceval-valid_middle_school_chemistry": { - "acc,none": 0.25, - "acc_stderr,none": 0.09933992677987828, - "acc_norm,none": 0.25, - "acc_norm_stderr,none": 0.09933992677987828, - "alias": " - ceval-valid_middle_school_chemistry" - }, - "ceval-valid_middle_school_geography": { - "acc,none": 0.08333333333333333, - "acc_stderr,none": 0.08333333333333331, - "acc_norm,none": 0.08333333333333333, - "acc_norm_stderr,none": 0.08333333333333331, - "alias": " - ceval-valid_middle_school_geography" - }, - "ceval-valid_middle_school_history": { - "acc,none": 0.45454545454545453, - "acc_stderr,none": 0.10865714630312667, - "acc_norm,none": 0.45454545454545453, - "acc_norm_stderr,none": 0.10865714630312667, - "alias": " - ceval-valid_middle_school_history" - }, - "ceval-valid_middle_school_mathematics": { - "acc,none": 0.3684210526315789, - "acc_stderr,none": 0.11369720523522558, - "acc_norm,none": 0.3684210526315789, - "acc_norm_stderr,none": 0.11369720523522558, - "alias": " - ceval-valid_middle_school_mathematics" - }, - "ceval-valid_middle_school_physics": { - "acc,none": 0.47368421052631576, - "acc_stderr,none": 0.11768778828946262, - "acc_norm,none": 0.47368421052631576, - "acc_norm_stderr,none": 0.11768778828946262, - "alias": " - ceval-valid_middle_school_physics" - }, - "ceval-valid_middle_school_politics": { - "acc,none": 0.2857142857142857, - "acc_stderr,none": 0.10101525445522108, - "acc_norm,none": 0.2857142857142857, - "acc_norm_stderr,none": 0.10101525445522108, - "alias": " - ceval-valid_middle_school_politics" - }, - "ceval-valid_modern_chinese_history": { - "acc,none": 0.21739130434782608, - "acc_stderr,none": 0.08793911249520547, - "acc_norm,none": 0.21739130434782608, - "acc_norm_stderr,none": 0.08793911249520547, - "alias": " - ceval-valid_modern_chinese_history" - }, - "ceval-valid_operating_system": { - "acc,none": 0.10526315789473684, - "acc_stderr,none": 0.07233518641434492, - "acc_norm,none": 0.10526315789473684, - "acc_norm_stderr,none": 0.07233518641434492, - "alias": " - ceval-valid_operating_system" - }, - "ceval-valid_physician": { - "acc,none": 0.20408163265306123, - "acc_stderr,none": 0.05817221556628253, - "acc_norm,none": 0.20408163265306123, - "acc_norm_stderr,none": 0.05817221556628253, - "alias": " - ceval-valid_physician" - }, - "ceval-valid_plant_protection": { - "acc,none": 0.3181818181818182, - "acc_stderr,none": 0.10163945352271771, - "acc_norm,none": 0.3181818181818182, - "acc_norm_stderr,none": 0.10163945352271771, - "alias": " - ceval-valid_plant_protection" - }, - "ceval-valid_probability_and_statistics": { - "acc,none": 0.2222222222222222, - "acc_stderr,none": 0.1008316903303367, - "acc_norm,none": 0.2222222222222222, - "acc_norm_stderr,none": 0.1008316903303367, - "alias": " - ceval-valid_probability_and_statistics" - }, - "ceval-valid_professional_tour_guide": { - "acc,none": 0.4482758620689655, - "acc_stderr,none": 0.09398415777506855, - "acc_norm,none": 0.4482758620689655, - "acc_norm_stderr,none": 0.09398415777506855, - "alias": " - ceval-valid_professional_tour_guide" - }, - "ceval-valid_sports_science": { - "acc,none": 0.15789473684210525, - "acc_stderr,none": 0.08594700851870798, - "acc_norm,none": 0.15789473684210525, - "acc_norm_stderr,none": 0.08594700851870798, - "alias": " - ceval-valid_sports_science" - }, - "ceval-valid_tax_accountant": { - "acc,none": 0.24489795918367346, - "acc_stderr,none": 0.062069005411206316, - "acc_norm,none": 0.24489795918367346, - "acc_norm_stderr,none": 0.062069005411206316, - "alias": " - ceval-valid_tax_accountant" - }, - "ceval-valid_teacher_qualification": { - "acc,none": 0.22727272727272727, - "acc_stderr,none": 0.06390760676613884, - "acc_norm,none": 0.22727272727272727, - "acc_norm_stderr,none": 0.06390760676613884, - "alias": " - ceval-valid_teacher_qualification" - }, - "ceval-valid_urban_and_rural_planner": { - "acc,none": 0.21739130434782608, - "acc_stderr,none": 0.06148754619013454, - "acc_norm,none": 0.21739130434782608, - "acc_norm_stderr,none": 0.06148754619013454, - "alias": " - ceval-valid_urban_and_rural_planner" - }, - "ceval-valid_veterinary_medicine": { - "acc,none": 0.13043478260869565, - "acc_stderr,none": 0.07180198468215394, - "acc_norm,none": 0.13043478260869565, - "acc_norm_stderr,none": 0.07180198468215394, - "alias": " - ceval-valid_veterinary_medicine" - } - }, - "groups": { - "ceval-valid": { - "acc,none": 0.24962852897473997, - "acc_stderr,none": 0.11387085890117267, - "acc_norm,none": 0.24962852897473997, - "acc_norm_stderr,none": 0.11387085890117267, - "alias": "ceval-valid" - } - }, - "configs": { - "ceval-valid_accountant": { - "task": "ceval-valid_accountant", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "accountant", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于注册会计师的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_advanced_mathematics": { - "task": "ceval-valid_advanced_mathematics", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "advanced_mathematics", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于高等数学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_art_studies": { - "task": "ceval-valid_art_studies", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "art_studies", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于艺术学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_basic_medicine": { - "task": "ceval-valid_basic_medicine", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "basic_medicine", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于基础医学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_business_administration": { - "task": "ceval-valid_business_administration", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "business_administration", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于工商管理的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_chinese_language_and_literature": { - "task": "ceval-valid_chinese_language_and_literature", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "chinese_language_and_literature", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于中国语言文学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_civil_servant": { - "task": "ceval-valid_civil_servant", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "civil_servant", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于公务员的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_clinical_medicine": { - "task": "ceval-valid_clinical_medicine", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "clinical_medicine", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于临床医学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_college_chemistry": { - "task": "ceval-valid_college_chemistry", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "college_chemistry", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于大学化学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_college_economics": { - "task": "ceval-valid_college_economics", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "college_economics", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于大学经济学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_college_physics": { - "task": "ceval-valid_college_physics", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "college_physics", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于大学物理的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_college_programming": { - "task": "ceval-valid_college_programming", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "college_programming", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于大学编程的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_computer_architecture": { - "task": "ceval-valid_computer_architecture", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "computer_architecture", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于计算机组成的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_computer_network": { - "task": "ceval-valid_computer_network", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "computer_network", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于计算机网络的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_discrete_mathematics": { - "task": "ceval-valid_discrete_mathematics", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "discrete_mathematics", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于离散数学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_education_science": { - "task": "ceval-valid_education_science", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "education_science", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于教育学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_electrical_engineer": { - "task": "ceval-valid_electrical_engineer", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "electrical_engineer", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于注册电气工程师的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_environmental_impact_assessment_engineer": { - "task": "ceval-valid_environmental_impact_assessment_engineer", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "environmental_impact_assessment_engineer", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于环境影响评价工程师的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_fire_engineer": { - "task": "ceval-valid_fire_engineer", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "fire_engineer", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于注册消防工程师的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_high_school_biology": { - "task": "ceval-valid_high_school_biology", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "high_school_biology", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于高中生物的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_high_school_chemistry": { - "task": "ceval-valid_high_school_chemistry", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "high_school_chemistry", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于高中化学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_high_school_chinese": { - "task": "ceval-valid_high_school_chinese", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "high_school_chinese", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于高中语文的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_high_school_geography": { - "task": "ceval-valid_high_school_geography", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "high_school_geography", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于高中地理的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_high_school_history": { - "task": "ceval-valid_high_school_history", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "high_school_history", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于高中历史的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_high_school_mathematics": { - "task": "ceval-valid_high_school_mathematics", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "high_school_mathematics", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于高中数学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_high_school_physics": { - "task": "ceval-valid_high_school_physics", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "high_school_physics", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于高中物理的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_high_school_politics": { - "task": "ceval-valid_high_school_politics", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "high_school_politics", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于高中政治的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_ideological_and_moral_cultivation": { - "task": "ceval-valid_ideological_and_moral_cultivation", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "ideological_and_moral_cultivation", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于思想道德修养与法律基础的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_law": { - "task": "ceval-valid_law", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "law", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于法学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_legal_professional": { - "task": "ceval-valid_legal_professional", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "legal_professional", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于法律职业资格的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_logic": { - "task": "ceval-valid_logic", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "logic", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于逻辑学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_mao_zedong_thought": { - "task": "ceval-valid_mao_zedong_thought", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "mao_zedong_thought", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于毛泽东思想和中国特色社会主义理论体系概论的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_marxism": { - "task": "ceval-valid_marxism", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "marxism", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于马克思主义基本原理的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_metrology_engineer": { - "task": "ceval-valid_metrology_engineer", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "metrology_engineer", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于注册计量师的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_middle_school_biology": { - "task": "ceval-valid_middle_school_biology", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "middle_school_biology", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于初中生物的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_middle_school_chemistry": { - "task": "ceval-valid_middle_school_chemistry", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "middle_school_chemistry", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于初中化学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_middle_school_geography": { - "task": "ceval-valid_middle_school_geography", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "middle_school_geography", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于初中地理的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_middle_school_history": { - "task": "ceval-valid_middle_school_history", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "middle_school_history", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于初中历史的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_middle_school_mathematics": { - "task": "ceval-valid_middle_school_mathematics", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "middle_school_mathematics", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于初中数学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_middle_school_physics": { - "task": "ceval-valid_middle_school_physics", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "middle_school_physics", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于初中物理的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_middle_school_politics": { - "task": "ceval-valid_middle_school_politics", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "middle_school_politics", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于初中政治的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_modern_chinese_history": { - "task": "ceval-valid_modern_chinese_history", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "modern_chinese_history", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于近代史纲要的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_operating_system": { - "task": "ceval-valid_operating_system", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "operating_system", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于操作系统的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_physician": { - "task": "ceval-valid_physician", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "physician", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于医师资格的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_plant_protection": { - "task": "ceval-valid_plant_protection", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "plant_protection", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于植物保护的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_probability_and_statistics": { - "task": "ceval-valid_probability_and_statistics", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "probability_and_statistics", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于概率统计的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_professional_tour_guide": { - "task": "ceval-valid_professional_tour_guide", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "professional_tour_guide", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于导游资格的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_sports_science": { - "task": "ceval-valid_sports_science", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "sports_science", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于体育学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_tax_accountant": { - "task": "ceval-valid_tax_accountant", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "tax_accountant", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于税务师的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_teacher_qualification": { - "task": "ceval-valid_teacher_qualification", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "teacher_qualification", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于教师资格的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_urban_and_rural_planner": { - "task": "ceval-valid_urban_and_rural_planner", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "urban_and_rural_planner", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于注册城乡规划师的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "ceval-valid_veterinary_medicine": { - "task": "ceval-valid_veterinary_medicine", - "group": "ceval-valid", - "dataset_path": "ceval/ceval-exam", - "dataset_name": "veterinary_medicine", - "validation_split": "val", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{A}}\nB. {{B}}\nC. {{C}}\nD. {{D}}\n答案:", - "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answer)}}", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "以下是中国关于兽医学的单项选择题,请选出其中的正确答案。\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "ceval-valid": "N/A", - "ceval-valid_accountant": 1.0, - "ceval-valid_advanced_mathematics": 1.0, - "ceval-valid_art_studies": 1.0, - "ceval-valid_basic_medicine": 1.0, - "ceval-valid_business_administration": 1.0, - "ceval-valid_chinese_language_and_literature": 1.0, - "ceval-valid_civil_servant": 1.0, - "ceval-valid_clinical_medicine": 1.0, - "ceval-valid_college_chemistry": 1.0, - "ceval-valid_college_economics": 1.0, - "ceval-valid_college_physics": 1.0, - "ceval-valid_college_programming": 1.0, - "ceval-valid_computer_architecture": 1.0, - "ceval-valid_computer_network": 1.0, - "ceval-valid_discrete_mathematics": 1.0, - "ceval-valid_education_science": 1.0, - "ceval-valid_electrical_engineer": 1.0, - "ceval-valid_environmental_impact_assessment_engineer": 1.0, - "ceval-valid_fire_engineer": 1.0, - "ceval-valid_high_school_biology": 1.0, - "ceval-valid_high_school_chemistry": 1.0, - "ceval-valid_high_school_chinese": 1.0, - "ceval-valid_high_school_geography": 1.0, - "ceval-valid_high_school_history": 1.0, - "ceval-valid_high_school_mathematics": 1.0, - "ceval-valid_high_school_physics": 1.0, - "ceval-valid_high_school_politics": 1.0, - "ceval-valid_ideological_and_moral_cultivation": 1.0, - "ceval-valid_law": 1.0, - "ceval-valid_legal_professional": 1.0, - "ceval-valid_logic": 1.0, - "ceval-valid_mao_zedong_thought": 1.0, - "ceval-valid_marxism": 1.0, - "ceval-valid_metrology_engineer": 1.0, - "ceval-valid_middle_school_biology": 1.0, - "ceval-valid_middle_school_chemistry": 1.0, - "ceval-valid_middle_school_geography": 1.0, - "ceval-valid_middle_school_history": 1.0, - "ceval-valid_middle_school_mathematics": 1.0, - "ceval-valid_middle_school_physics": 1.0, - "ceval-valid_middle_school_politics": 1.0, - "ceval-valid_modern_chinese_history": 1.0, - "ceval-valid_operating_system": 1.0, - "ceval-valid_physician": 1.0, - "ceval-valid_plant_protection": 1.0, - "ceval-valid_probability_and_statistics": 1.0, - "ceval-valid_professional_tour_guide": 1.0, - "ceval-valid_sports_science": 1.0, - "ceval-valid_tax_accountant": 1.0, - "ceval-valid_teacher_qualification": 1.0, - "ceval-valid_urban_and_rural_planner": 1.0, - "ceval-valid_veterinary_medicine": 1.0 - }, - "n-shot": { - "ceval-valid": 0, - "ceval-valid_accountant": 0, - "ceval-valid_advanced_mathematics": 0, - "ceval-valid_art_studies": 0, - "ceval-valid_basic_medicine": 0, - "ceval-valid_business_administration": 0, - "ceval-valid_chinese_language_and_literature": 0, - "ceval-valid_civil_servant": 0, - "ceval-valid_clinical_medicine": 0, - "ceval-valid_college_chemistry": 0, - "ceval-valid_college_economics": 0, - "ceval-valid_college_physics": 0, - "ceval-valid_college_programming": 0, - "ceval-valid_computer_architecture": 0, - "ceval-valid_computer_network": 0, - "ceval-valid_discrete_mathematics": 0, - "ceval-valid_education_science": 0, - "ceval-valid_electrical_engineer": 0, - "ceval-valid_environmental_impact_assessment_engineer": 0, - "ceval-valid_fire_engineer": 0, - "ceval-valid_high_school_biology": 0, - "ceval-valid_high_school_chemistry": 0, - "ceval-valid_high_school_chinese": 0, - "ceval-valid_high_school_geography": 0, - "ceval-valid_high_school_history": 0, - "ceval-valid_high_school_mathematics": 0, - "ceval-valid_high_school_physics": 0, - "ceval-valid_high_school_politics": 0, - "ceval-valid_ideological_and_moral_cultivation": 0, - "ceval-valid_law": 0, - "ceval-valid_legal_professional": 0, - "ceval-valid_logic": 0, - "ceval-valid_mao_zedong_thought": 0, - "ceval-valid_marxism": 0, - "ceval-valid_metrology_engineer": 0, - "ceval-valid_middle_school_biology": 0, - "ceval-valid_middle_school_chemistry": 0, - "ceval-valid_middle_school_geography": 0, - "ceval-valid_middle_school_history": 0, - "ceval-valid_middle_school_mathematics": 0, - "ceval-valid_middle_school_physics": 0, - "ceval-valid_middle_school_politics": 0, - "ceval-valid_modern_chinese_history": 0, - "ceval-valid_operating_system": 0, - "ceval-valid_physician": 0, - "ceval-valid_plant_protection": 0, - "ceval-valid_probability_and_statistics": 0, - "ceval-valid_professional_tour_guide": 0, - "ceval-valid_sports_science": 0, - "ceval-valid_tax_accountant": 0, - "ceval-valid_teacher_qualification": 0, - "ceval-valid_urban_and_rural_planner": 0, - "ceval-valid_veterinary_medicine": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 20618f73394f2806202321e26d9f51b4f7ac6f4e..63e0cc568c6461c95c2598605d04ec76b578518a 100644 --- a/lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/ceval-valid/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:b5ee2c65a8b0c7e2a23e6f37952f1caa52bc173fea074a725a634d7b0194f619 -size 64855 +oid sha256:f000647be6eea5529f603c75503e3332301c0bd4aa7f76edcbbb38001ab4cc9c +size 79965 diff --git a/lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 9e355e7eace190bff951d19c314ec679486b0e22..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:d49be0dd3eb66a0025a1d0cdd953772d8bdfc3cd910d47adcca1eaa055d00007 -size 2305507 diff --git a/lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 719078f1e9fc4d4ac0d43e680a1ce6a7c9a83053..99bf24281e62b049a4614d0465008beef05990ca 100644 --- a/lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,38 +1,38 @@ { "results": { "cmmlu": { - "acc,none": 0.2509065791745812, - "acc_stderr,none": 0.046349557976842405, - "acc_norm,none": 0.2509065791745812, - "acc_norm_stderr,none": 0.046349557976842405, + "acc,none": 0.2485753755828009, + "acc_stderr,none": 0.042815306797706565, + "acc_norm,none": 0.2485753755828009, + "acc_norm_stderr,none": 0.042815306797706565, "alias": "cmmlu" }, "cmmlu_agronomy": { - "acc,none": 0.17751479289940827, - "acc_stderr,none": 0.029479945887526282, - "acc_norm,none": 0.17751479289940827, - "acc_norm_stderr,none": 0.029479945887526282, + "acc,none": 0.1952662721893491, + "acc_stderr,none": 0.030583351673923103, + "acc_norm,none": 0.1952662721893491, + "acc_norm_stderr,none": 0.030583351673923103, "alias": " - cmmlu_agronomy" }, "cmmlu_anatomy": { - "acc,none": 0.24324324324324326, - "acc_stderr,none": 0.0353866849031339, - "acc_norm,none": 0.24324324324324326, - "acc_norm_stderr,none": 0.0353866849031339, + "acc,none": 0.22297297297297297, + "acc_stderr,none": 0.03433092518104002, + "acc_norm,none": 0.22297297297297297, + "acc_norm_stderr,none": 0.03433092518104002, "alias": " - cmmlu_anatomy" }, "cmmlu_ancient_chinese": { - "acc,none": 0.23170731707317074, - "acc_stderr,none": 0.033047561588107864, - "acc_norm,none": 0.23170731707317074, - "acc_norm_stderr,none": 0.033047561588107864, + "acc,none": 0.24390243902439024, + "acc_stderr,none": 0.033635910482728223, + "acc_norm,none": 0.24390243902439024, + "acc_norm_stderr,none": 0.033635910482728223, "alias": " - cmmlu_ancient_chinese" }, "cmmlu_arts": { - "acc,none": 0.3, - "acc_stderr,none": 0.036342189215581536, - "acc_norm,none": 0.3, - "acc_norm_stderr,none": 0.036342189215581536, + "acc,none": 0.29375, + "acc_stderr,none": 0.03612181848191273, + "acc_norm,none": 0.29375, + "acc_norm_stderr,none": 0.03612181848191273, "alias": " - cmmlu_arts" }, "cmmlu_astronomy": { @@ -43,122 +43,122 @@ "alias": " - cmmlu_astronomy" }, "cmmlu_business_ethics": { - "acc,none": 0.23923444976076555, - "acc_stderr,none": 0.029580506819430464, - "acc_norm,none": 0.23923444976076555, - "acc_norm_stderr,none": 0.029580506819430464, + "acc,none": 0.23444976076555024, + "acc_stderr,none": 0.029375148972005737, + "acc_norm,none": 0.23444976076555024, + "acc_norm_stderr,none": 0.029375148972005737, "alias": " - cmmlu_business_ethics" }, "cmmlu_chinese_civil_service_exam": { - "acc,none": 0.20625, - "acc_stderr,none": 0.03208782538184617, - "acc_norm,none": 0.20625, - "acc_norm_stderr,none": 0.03208782538184617, + "acc,none": 0.21875, + "acc_stderr,none": 0.032784644885244255, + "acc_norm,none": 0.21875, + "acc_norm_stderr,none": 0.032784644885244255, "alias": " - cmmlu_chinese_civil_service_exam" }, "cmmlu_chinese_driving_rule": { - "acc,none": 0.2366412213740458, - "acc_stderr,none": 0.03727673575596917, - "acc_norm,none": 0.2366412213740458, - "acc_norm_stderr,none": 0.03727673575596917, + "acc,none": 0.25190839694656486, + "acc_stderr,none": 0.03807387116306086, + "acc_norm,none": 0.25190839694656486, + "acc_norm_stderr,none": 0.03807387116306086, "alias": " - cmmlu_chinese_driving_rule" }, "cmmlu_chinese_food_culture": { - "acc,none": 0.21323529411764705, - "acc_stderr,none": 0.035252108259539325, - "acc_norm,none": 0.21323529411764705, - "acc_norm_stderr,none": 0.035252108259539325, + "acc,none": 0.20588235294117646, + "acc_stderr,none": 0.034800469312350674, + "acc_norm,none": 0.20588235294117646, + "acc_norm_stderr,none": 0.034800469312350674, "alias": " - cmmlu_chinese_food_culture" }, "cmmlu_chinese_foreign_policy": { - "acc,none": 0.205607476635514, - "acc_stderr,none": 0.03925401580070485, - "acc_norm,none": 0.205607476635514, - "acc_norm_stderr,none": 0.03925401580070485, + "acc,none": 0.2336448598130841, + "acc_stderr,none": 0.041099848424639984, + "acc_norm,none": 0.2336448598130841, + "acc_norm_stderr,none": 0.041099848424639984, "alias": " - cmmlu_chinese_foreign_policy" }, "cmmlu_chinese_history": { - "acc,none": 0.2848297213622291, - "acc_stderr,none": 0.02515182168617951, - "acc_norm,none": 0.2848297213622291, - "acc_norm_stderr,none": 0.02515182168617951, + "acc,none": 0.29721362229102166, + "acc_stderr,none": 0.025469363219004768, + "acc_norm,none": 0.29721362229102166, + "acc_norm_stderr,none": 0.025469363219004768, "alias": " - cmmlu_chinese_history" }, "cmmlu_chinese_literature": { - "acc,none": 0.2107843137254902, - "acc_stderr,none": 0.02862654791243739, - "acc_norm,none": 0.2107843137254902, - "acc_norm_stderr,none": 0.02862654791243739, + "acc,none": 0.22058823529411764, + "acc_stderr,none": 0.02910225438967408, + "acc_norm,none": 0.22058823529411764, + "acc_norm_stderr,none": 0.02910225438967408, "alias": " - cmmlu_chinese_literature" }, "cmmlu_chinese_teacher_qualification": { - "acc,none": 0.24022346368715083, - "acc_stderr,none": 0.032021424638044936, - "acc_norm,none": 0.24022346368715083, - "acc_norm_stderr,none": 0.032021424638044936, + "acc,none": 0.22346368715083798, + "acc_stderr,none": 0.031222980919579764, + "acc_norm,none": 0.22346368715083798, + "acc_norm_stderr,none": 0.031222980919579764, "alias": " - cmmlu_chinese_teacher_qualification" }, "cmmlu_clinical_knowledge": { - "acc,none": 0.24472573839662448, - "acc_stderr,none": 0.027985699387036406, - "acc_norm,none": 0.24472573839662448, - "acc_norm_stderr,none": 0.027985699387036406, + "acc,none": 0.22784810126582278, + "acc_stderr,none": 0.027303484599069443, + "acc_norm,none": 0.22784810126582278, + "acc_norm_stderr,none": 0.027303484599069443, "alias": " - cmmlu_clinical_knowledge" }, "cmmlu_college_actuarial_science": { - "acc,none": 0.3018867924528302, - "acc_stderr,none": 0.044801270921106716, - "acc_norm,none": 0.3018867924528302, - "acc_norm_stderr,none": 0.044801270921106716, + "acc,none": 0.2830188679245283, + "acc_stderr,none": 0.04396093377439375, + "acc_norm,none": 0.2830188679245283, + "acc_norm_stderr,none": 0.04396093377439375, "alias": " - cmmlu_college_actuarial_science" }, "cmmlu_college_education": { - "acc,none": 0.308411214953271, - "acc_stderr,none": 0.04485760883316698, - "acc_norm,none": 0.308411214953271, - "acc_norm_stderr,none": 0.04485760883316698, + "acc,none": 0.2897196261682243, + "acc_stderr,none": 0.0440606533474851, + "acc_norm,none": 0.2897196261682243, + "acc_norm_stderr,none": 0.0440606533474851, "alias": " - cmmlu_college_education" }, "cmmlu_college_engineering_hydrology": { - "acc,none": 0.2830188679245283, - "acc_stderr,none": 0.04396093377439375, - "acc_norm,none": 0.2830188679245283, - "acc_norm_stderr,none": 0.04396093377439375, + "acc,none": 0.2641509433962264, + "acc_stderr,none": 0.043025487739590106, + "acc_norm,none": 0.2641509433962264, + "acc_norm_stderr,none": 0.043025487739590106, "alias": " - cmmlu_college_engineering_hydrology" }, "cmmlu_college_law": { - "acc,none": 0.28703703703703703, - "acc_stderr,none": 0.043733130409147614, - "acc_norm,none": 0.28703703703703703, - "acc_norm_stderr,none": 0.043733130409147614, + "acc,none": 0.2962962962962963, + "acc_stderr,none": 0.044143436668549335, + "acc_norm,none": 0.2962962962962963, + "acc_norm_stderr,none": 0.044143436668549335, "alias": " - cmmlu_college_law" }, "cmmlu_college_mathematics": { - "acc,none": 0.3142857142857143, - "acc_stderr,none": 0.045521571818039494, - "acc_norm,none": 0.3142857142857143, - "acc_norm_stderr,none": 0.045521571818039494, + "acc,none": 0.3047619047619048, + "acc_stderr,none": 0.04513676718168311, + "acc_norm,none": 0.3047619047619048, + "acc_norm_stderr,none": 0.04513676718168311, "alias": " - cmmlu_college_mathematics" }, "cmmlu_college_medical_statistics": { - "acc,none": 0.18867924528301888, - "acc_stderr,none": 0.0381824426969915, - "acc_norm,none": 0.18867924528301888, - "acc_norm_stderr,none": 0.0381824426969915, + "acc,none": 0.19811320754716982, + "acc_stderr,none": 0.0388972228831855, + "acc_norm,none": 0.19811320754716982, + "acc_norm_stderr,none": 0.0388972228831855, "alias": " - cmmlu_college_medical_statistics" }, "cmmlu_college_medicine": { - "acc,none": 0.21978021978021978, - "acc_stderr,none": 0.025108358900325773, - "acc_norm,none": 0.21978021978021978, - "acc_norm_stderr,none": 0.025108358900325773, + "acc,none": 0.23809523809523808, + "acc_stderr,none": 0.02582505450222104, + "acc_norm,none": 0.23809523809523808, + "acc_norm_stderr,none": 0.02582505450222104, "alias": " - cmmlu_college_medicine" }, "cmmlu_computer_science": { - "acc,none": 0.23039215686274508, - "acc_stderr,none": 0.029554292605695077, - "acc_norm,none": 0.23039215686274508, - "acc_norm_stderr,none": 0.029554292605695077, + "acc,none": 0.24509803921568626, + "acc_stderr,none": 0.030190282453501964, + "acc_norm,none": 0.24509803921568626, + "acc_norm_stderr,none": 0.030190282453501964, "alias": " - cmmlu_computer_science" }, "cmmlu_computer_security": { @@ -169,45 +169,45 @@ "alias": " - cmmlu_computer_security" }, "cmmlu_conceptual_physics": { - "acc,none": 0.2108843537414966, - "acc_stderr,none": 0.033761060398578915, - "acc_norm,none": 0.2108843537414966, - "acc_norm_stderr,none": 0.033761060398578915, + "acc,none": 0.23129251700680273, + "acc_stderr,none": 0.034896744812616155, + "acc_norm,none": 0.23129251700680273, + "acc_norm_stderr,none": 0.034896744812616155, "alias": " - cmmlu_conceptual_physics" }, "cmmlu_construction_project_management": { - "acc,none": 0.302158273381295, - "acc_stderr,none": 0.03908914479291562, - "acc_norm,none": 0.302158273381295, - "acc_norm_stderr,none": 0.03908914479291562, + "acc,none": 0.28776978417266186, + "acc_stderr,none": 0.03853836179233389, + "acc_norm,none": 0.28776978417266186, + "acc_norm_stderr,none": 0.03853836179233389, "alias": " - cmmlu_construction_project_management" }, "cmmlu_economics": { - "acc,none": 0.29559748427672955, - "acc_stderr,none": 0.036302143777231344, - "acc_norm,none": 0.29559748427672955, - "acc_norm_stderr,none": 0.036302143777231344, + "acc,none": 0.27672955974842767, + "acc_stderr,none": 0.03559177035707934, + "acc_norm,none": 0.27672955974842767, + "acc_norm_stderr,none": 0.03559177035707934, "alias": " - cmmlu_economics" }, "cmmlu_education": { - "acc,none": 0.3006134969325153, - "acc_stderr,none": 0.03602511318806771, - "acc_norm,none": 0.3006134969325153, - "acc_norm_stderr,none": 0.03602511318806771, + "acc,none": 0.27607361963190186, + "acc_stderr,none": 0.0351238528370505, + "acc_norm,none": 0.27607361963190186, + "acc_norm_stderr,none": 0.0351238528370505, "alias": " - cmmlu_education" }, "cmmlu_electrical_engineering": { - "acc,none": 0.26744186046511625, - "acc_stderr,none": 0.03384836428157859, - "acc_norm,none": 0.26744186046511625, - "acc_norm_stderr,none": 0.03384836428157859, + "acc,none": 0.27325581395348836, + "acc_stderr,none": 0.03407826167337437, + "acc_norm,none": 0.27325581395348836, + "acc_norm_stderr,none": 0.03407826167337437, "alias": " - cmmlu_electrical_engineering" }, "cmmlu_elementary_chinese": { - "acc,none": 0.24206349206349206, - "acc_stderr,none": 0.027036109679236968, - "acc_norm,none": 0.24206349206349206, - "acc_norm_stderr,none": 0.027036109679236968, + "acc,none": 0.23412698412698413, + "acc_stderr,none": 0.0267280489993024, + "acc_norm,none": 0.23412698412698413, + "acc_norm_stderr,none": 0.0267280489993024, "alias": " - cmmlu_elementary_chinese" }, "cmmlu_elementary_commonsense": { @@ -218,10 +218,10 @@ "alias": " - cmmlu_elementary_commonsense" }, "cmmlu_elementary_information_and_technology": { - "acc,none": 0.27310924369747897, - "acc_stderr,none": 0.028942004040998167, - "acc_norm,none": 0.27310924369747897, - "acc_norm_stderr,none": 0.028942004040998167, + "acc,none": 0.24369747899159663, + "acc_stderr,none": 0.02788682807838058, + "acc_norm,none": 0.24369747899159663, + "acc_norm_stderr,none": 0.02788682807838058, "alias": " - cmmlu_elementary_information_and_technology" }, "cmmlu_elementary_mathematics": { @@ -232,66 +232,66 @@ "alias": " - cmmlu_elementary_mathematics" }, "cmmlu_ethnology": { - "acc,none": 0.2740740740740741, - "acc_stderr,none": 0.03853254836552003, - "acc_norm,none": 0.2740740740740741, - "acc_norm_stderr,none": 0.03853254836552003, + "acc,none": 0.2518518518518518, + "acc_stderr,none": 0.037498507091740234, + "acc_norm,none": 0.2518518518518518, + "acc_norm_stderr,none": 0.037498507091740234, "alias": " - cmmlu_ethnology" }, "cmmlu_food_science": { - "acc,none": 0.27972027972027974, - "acc_stderr,none": 0.037667638895398536, - "acc_norm,none": 0.27972027972027974, - "acc_norm_stderr,none": 0.037667638895398536, + "acc,none": 0.2867132867132867, + "acc_stderr,none": 0.03795000212801782, + "acc_norm,none": 0.2867132867132867, + "acc_norm_stderr,none": 0.03795000212801782, "alias": " - cmmlu_food_science" }, "cmmlu_genetics": { - "acc,none": 0.2784090909090909, - "acc_stderr,none": 0.03388193526335356, - "acc_norm,none": 0.2784090909090909, - "acc_norm_stderr,none": 0.03388193526335356, + "acc,none": 0.2897727272727273, + "acc_stderr,none": 0.034293230802398766, + "acc_norm,none": 0.2897727272727273, + "acc_norm_stderr,none": 0.034293230802398766, "alias": " - cmmlu_genetics" }, "cmmlu_global_facts": { - "acc,none": 0.24161073825503357, - "acc_stderr,none": 0.03518627932594346, - "acc_norm,none": 0.24161073825503357, - "acc_norm_stderr,none": 0.03518627932594346, + "acc,none": 0.2483221476510067, + "acc_stderr,none": 0.0355134404169743, + "acc_norm,none": 0.2483221476510067, + "acc_norm_stderr,none": 0.0355134404169743, "alias": " - cmmlu_global_facts" }, "cmmlu_high_school_biology": { - "acc,none": 0.22485207100591717, - "acc_stderr,none": 0.03220965704514525, - "acc_norm,none": 0.22485207100591717, - "acc_norm_stderr,none": 0.03220965704514525, + "acc,none": 0.23076923076923078, + "acc_stderr,none": 0.03250593287417369, + "acc_norm,none": 0.23076923076923078, + "acc_norm_stderr,none": 0.03250593287417369, "alias": " - cmmlu_high_school_biology" }, "cmmlu_high_school_chemistry": { - "acc,none": 0.25757575757575757, - "acc_stderr,none": 0.038206998148497956, - "acc_norm,none": 0.25757575757575757, - "acc_norm_stderr,none": 0.038206998148497956, + "acc,none": 0.22727272727272727, + "acc_stderr,none": 0.03661433360410719, + "acc_norm,none": 0.22727272727272727, + "acc_norm_stderr,none": 0.03661433360410719, "alias": " - cmmlu_high_school_chemistry" }, "cmmlu_high_school_geography": { - "acc,none": 0.2711864406779661, - "acc_stderr,none": 0.04110070549339208, - "acc_norm,none": 0.2711864406779661, - "acc_norm_stderr,none": 0.04110070549339208, + "acc,none": 0.2457627118644068, + "acc_stderr,none": 0.03980329854920432, + "acc_norm,none": 0.2457627118644068, + "acc_norm_stderr,none": 0.03980329854920432, "alias": " - cmmlu_high_school_geography" }, "cmmlu_high_school_mathematics": { - "acc,none": 0.27439024390243905, - "acc_stderr,none": 0.03494959016177541, - "acc_norm,none": 0.27439024390243905, - "acc_norm_stderr,none": 0.03494959016177541, + "acc,none": 0.2621951219512195, + "acc_stderr,none": 0.0344500028917346, + "acc_norm,none": 0.2621951219512195, + "acc_norm_stderr,none": 0.0344500028917346, "alias": " - cmmlu_high_school_mathematics" }, "cmmlu_high_school_physics": { - "acc,none": 0.2636363636363636, - "acc_stderr,none": 0.04220224692971987, - "acc_norm,none": 0.2636363636363636, - "acc_norm_stderr,none": 0.04220224692971987, + "acc,none": 0.2818181818181818, + "acc_stderr,none": 0.04309118709946458, + "acc_norm,none": 0.2818181818181818, + "acc_norm_stderr,none": 0.04309118709946458, "alias": " - cmmlu_high_school_physics" }, "cmmlu_high_school_politics": { @@ -302,129 +302,129 @@ "alias": " - cmmlu_high_school_politics" }, "cmmlu_human_sexuality": { - "acc,none": 0.1984126984126984, - "acc_stderr,none": 0.03567016675276863, - "acc_norm,none": 0.1984126984126984, - "acc_norm_stderr,none": 0.03567016675276863, + "acc,none": 0.19047619047619047, + "acc_stderr,none": 0.035122074123020534, + "acc_norm,none": 0.19047619047619047, + "acc_norm_stderr,none": 0.035122074123020534, "alias": " - cmmlu_human_sexuality" }, "cmmlu_international_law": { - "acc,none": 0.2648648648648649, - "acc_stderr,none": 0.032530209055933366, - "acc_norm,none": 0.2648648648648649, - "acc_norm_stderr,none": 0.032530209055933366, + "acc,none": 0.25405405405405407, + "acc_stderr,none": 0.03209281645145386, + "acc_norm,none": 0.25405405405405407, + "acc_norm_stderr,none": 0.03209281645145386, "alias": " - cmmlu_international_law" }, "cmmlu_journalism": { - "acc,none": 0.20930232558139536, - "acc_stderr,none": 0.031109583909764642, - "acc_norm,none": 0.20930232558139536, - "acc_norm_stderr,none": 0.031109583909764642, + "acc,none": 0.22674418604651161, + "acc_stderr,none": 0.032020758995849365, + "acc_norm,none": 0.22674418604651161, + "acc_norm_stderr,none": 0.032020758995849365, "alias": " - cmmlu_journalism" }, "cmmlu_jurisprudence": { - "acc,none": 0.2871046228710462, - "acc_stderr,none": 0.02234297829335579, - "acc_norm,none": 0.2871046228710462, - "acc_norm_stderr,none": 0.02234297829335579, + "acc,none": 0.26763990267639903, + "acc_stderr,none": 0.021864816663672668, + "acc_norm,none": 0.26763990267639903, + "acc_norm_stderr,none": 0.021864816663672668, "alias": " - cmmlu_jurisprudence" }, "cmmlu_legal_and_moral_basis": { - "acc,none": 0.3130841121495327, - "acc_stderr,none": 0.03177550735912672, - "acc_norm,none": 0.3130841121495327, - "acc_norm_stderr,none": 0.03177550735912672, + "acc,none": 0.3037383177570093, + "acc_stderr,none": 0.03150984286811783, + "acc_norm,none": 0.3037383177570093, + "acc_norm_stderr,none": 0.03150984286811783, "alias": " - cmmlu_legal_and_moral_basis" }, "cmmlu_logical": { - "acc,none": 0.2682926829268293, - "acc_stderr,none": 0.040113743936211456, - "acc_norm,none": 0.2682926829268293, - "acc_norm_stderr,none": 0.040113743936211456, + "acc,none": 0.25203252032520324, + "acc_stderr,none": 0.039308795268239924, + "acc_norm,none": 0.25203252032520324, + "acc_norm_stderr,none": 0.039308795268239924, "alias": " - cmmlu_logical" }, "cmmlu_machine_learning": { - "acc,none": 0.3114754098360656, - "acc_stderr,none": 0.0420996926731014, - "acc_norm,none": 0.3114754098360656, - "acc_norm_stderr,none": 0.0420996926731014, + "acc,none": 0.28688524590163933, + "acc_stderr,none": 0.041118866352671826, + "acc_norm,none": 0.28688524590163933, + "acc_norm_stderr,none": 0.041118866352671826, "alias": " - cmmlu_machine_learning" }, "cmmlu_management": { - "acc,none": 0.18571428571428572, - "acc_stderr,none": 0.026899110619750637, - "acc_norm,none": 0.18571428571428572, - "acc_norm_stderr,none": 0.026899110619750637, + "acc,none": 0.19523809523809524, + "acc_stderr,none": 0.027418446398346896, + "acc_norm,none": 0.19523809523809524, + "acc_norm_stderr,none": 0.027418446398346896, "alias": " - cmmlu_management" }, "cmmlu_marketing": { - "acc,none": 0.2388888888888889, - "acc_stderr,none": 0.03187098535605761, - "acc_norm,none": 0.2388888888888889, - "acc_norm_stderr,none": 0.03187098535605761, + "acc,none": 0.25, + "acc_stderr,none": 0.032364888900157734, + "acc_norm,none": 0.25, + "acc_norm_stderr,none": 0.032364888900157734, "alias": " - cmmlu_marketing" }, "cmmlu_marxist_theory": { - "acc,none": 0.25396825396825395, - "acc_stderr,none": 0.03174603174603175, - "acc_norm,none": 0.25396825396825395, - "acc_norm_stderr,none": 0.03174603174603175, + "acc,none": 0.24338624338624337, + "acc_stderr,none": 0.031297251928558506, + "acc_norm,none": 0.24338624338624337, + "acc_norm_stderr,none": 0.031297251928558506, "alias": " - cmmlu_marxist_theory" }, "cmmlu_modern_chinese": { - "acc,none": 0.22413793103448276, - "acc_stderr,none": 0.03888669370117824, - "acc_norm,none": 0.22413793103448276, - "acc_norm_stderr,none": 0.03888669370117824, + "acc,none": 0.23275862068965517, + "acc_stderr,none": 0.039406691683376995, + "acc_norm,none": 0.23275862068965517, + "acc_norm_stderr,none": 0.039406691683376995, "alias": " - cmmlu_modern_chinese" }, "cmmlu_nutrition": { - "acc,none": 0.2620689655172414, - "acc_stderr,none": 0.036646663372252565, - "acc_norm,none": 0.2620689655172414, - "acc_norm_stderr,none": 0.036646663372252565, + "acc,none": 0.2689655172413793, + "acc_stderr,none": 0.036951833116502325, + "acc_norm,none": 0.2689655172413793, + "acc_norm_stderr,none": 0.036951833116502325, "alias": " - cmmlu_nutrition" }, "cmmlu_philosophy": { - "acc,none": 0.3142857142857143, - "acc_stderr,none": 0.045521571818039494, - "acc_norm,none": 0.3142857142857143, - "acc_norm_stderr,none": 0.045521571818039494, + "acc,none": 0.3047619047619048, + "acc_stderr,none": 0.0451367671816831, + "acc_norm,none": 0.3047619047619048, + "acc_norm_stderr,none": 0.0451367671816831, "alias": " - cmmlu_philosophy" }, "cmmlu_professional_accounting": { - "acc,none": 0.25142857142857145, - "acc_stderr,none": 0.032888897342098204, - "acc_norm,none": 0.25142857142857145, - "acc_norm_stderr,none": 0.032888897342098204, + "acc,none": 0.2342857142857143, + "acc_stderr,none": 0.032109360396926204, + "acc_norm,none": 0.2342857142857143, + "acc_norm_stderr,none": 0.032109360396926204, "alias": " - cmmlu_professional_accounting" }, "cmmlu_professional_law": { - "acc,none": 0.2890995260663507, - "acc_stderr,none": 0.03128372390561387, - "acc_norm,none": 0.2890995260663507, - "acc_norm_stderr,none": 0.03128372390561387, + "acc,none": 0.2843601895734597, + "acc_stderr,none": 0.031129489323148667, + "acc_norm,none": 0.2843601895734597, + "acc_norm_stderr,none": 0.031129489323148667, "alias": " - cmmlu_professional_law" }, "cmmlu_professional_medicine": { - "acc,none": 0.2579787234042553, - "acc_stderr,none": 0.022593550801056256, - "acc_norm,none": 0.2579787234042553, - "acc_norm_stderr,none": 0.022593550801056256, + "acc,none": 0.26595744680851063, + "acc_stderr,none": 0.022816607010135298, + "acc_norm,none": 0.26595744680851063, + "acc_norm_stderr,none": 0.022816607010135298, "alias": " - cmmlu_professional_medicine" }, "cmmlu_professional_psychology": { - "acc,none": 0.19827586206896552, - "acc_stderr,none": 0.02623260459197056, - "acc_norm,none": 0.19827586206896552, - "acc_norm_stderr,none": 0.02623260459197056, + "acc,none": 0.23706896551724138, + "acc_stderr,none": 0.02798169400862497, + "acc_norm,none": 0.23706896551724138, + "acc_norm_stderr,none": 0.02798169400862497, "alias": " - cmmlu_professional_psychology" }, "cmmlu_public_relations": { - "acc,none": 0.27586206896551724, - "acc_stderr,none": 0.03398079939585583, - "acc_norm,none": 0.27586206896551724, - "acc_norm_stderr,none": 0.03398079939585583, + "acc,none": 0.25862068965517243, + "acc_stderr,none": 0.033291151121447815, + "acc_norm,none": 0.25862068965517243, + "acc_norm_stderr,none": 0.033291151121447815, "alias": " - cmmlu_public_relations" }, "cmmlu_security_study": { @@ -435,31 +435,31 @@ "alias": " - cmmlu_security_study" }, "cmmlu_sociology": { - "acc,none": 0.23893805309734514, - "acc_stderr,none": 0.02842898832603367, - "acc_norm,none": 0.23893805309734514, - "acc_norm_stderr,none": 0.02842898832603367, + "acc,none": 0.22566371681415928, + "acc_stderr,none": 0.027867910955296744, + "acc_norm,none": 0.22566371681415928, + "acc_norm_stderr,none": 0.027867910955296744, "alias": " - cmmlu_sociology" }, "cmmlu_sports_science": { - "acc,none": 0.2606060606060606, - "acc_stderr,none": 0.034277431758165236, - "acc_norm,none": 0.2606060606060606, - "acc_norm_stderr,none": 0.034277431758165236, + "acc,none": 0.23030303030303031, + "acc_stderr,none": 0.03287666758603489, + "acc_norm,none": 0.23030303030303031, + "acc_norm_stderr,none": 0.03287666758603489, "alias": " - cmmlu_sports_science" }, "cmmlu_traditional_chinese_medicine": { "acc,none": 0.23243243243243245, - "acc_stderr,none": 0.03113850517079465, + "acc_stderr,none": 0.031138505170794653, "acc_norm,none": 0.23243243243243245, - "acc_norm_stderr,none": 0.03113850517079465, + "acc_norm_stderr,none": 0.031138505170794653, "alias": " - cmmlu_traditional_chinese_medicine" }, "cmmlu_virology": { - "acc,none": 0.2485207100591716, - "acc_stderr,none": 0.03334150198101965, - "acc_norm,none": 0.2485207100591716, - "acc_norm_stderr,none": 0.03334150198101965, + "acc,none": 0.24260355029585798, + "acc_stderr,none": 0.03307162750323179, + "acc_norm,none": 0.24260355029585798, + "acc_norm_stderr,none": 0.03307162750323179, "alias": " - cmmlu_virology" }, "cmmlu_world_history": { @@ -479,10 +479,10 @@ }, "groups": { "cmmlu": { - "acc,none": 0.2509065791745812, - "acc_stderr,none": 0.046349557976842405, - "acc_norm,none": 0.2509065791745812, - "acc_norm_stderr,none": 0.046349557976842405, + "acc,none": 0.2485753755828009, + "acc_stderr,none": 0.042815306797706565, + "acc_norm,none": 0.2485753755828009, + "acc_norm_stderr,none": 0.042815306797706565, "alias": "cmmlu" } }, @@ -3311,13 +3311,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 8 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 8bed7b389cdfb38374eec20787acb4b517adaaab..cdfe46ac8c67244672a2ce36f9fa02583686a426 100644 --- a/lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/cmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:3969b9c5b5f77b7d2db2d31bce13b271b0b55f53de1c06b12e1484a260cc5ea9 -size 129703 +oid sha256:cc69f38ccd1ff5eed2fe9dbdfb6e94b5d94c6c9d0c737ef0a4f268f3b0b4733c +size 111568 diff --git a/lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 6cd963bd0d133e9da91e578bd7360be26eaf1923..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:3862814204891a844fdf5337965d761f05144ba67d8d95ab5b6cfa8973e8bf9e -size 59222 diff --git a/lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index ed80aa4d00b5b3584ebe8074971411fa1d9d0787..6becb4d4ba0d1273513195c8f743b141272891e9 100644 --- a/lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "cola": { - "mcc,none": 0.012008657795714008, - "mcc_stderr,none": 0.03141923039277966, + "mcc,none": 0.003737743780434562, + "mcc_stderr,none": 0.031171364680531898, "alias": "cola" } }, @@ -46,13 +46,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 71bc26c4850f8bc7a2a98d892cef78af1ad6a06b..28f61db3ceed9c30ab6546abc32b47a4d5462831 100644 --- a/lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/cola/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:6c751bb281058343edd3ad94b7a343159714eb03423174b22f62db4ffcac333e -size 15559 +oid sha256:8ed94ed13ea33ac892b6c4b24f6f02443171923335c00eff64af1c1d89568425 +size 14251 diff --git a/lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 221a497479b33e1b805f93463ffe2c41c5a4d2dc..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2b9320d9c255922d66f90387aeac8cab13a694251714ee4ae12c4a35e6b86f5f -size 10140 diff --git a/lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 5b50ce968830d06ebc86288b92698f67ce610b50..1cdd8e80688de426020656fb5198fed2bedd9f4d 100644 --- a/lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "copa": { - "acc,none": 0.87, - "acc_stderr,none": 0.033799766898963086, + "acc,none": 0.86, + "acc_stderr,none": 0.03487350880197769, "alias": "copa" } }, @@ -44,13 +44,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 16e32aade6b5918bafe74b7c229121a720585bd1..6ec347fb84e0aeb1565c06ada91b9fcba02dc386 100644 --- a/lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/copa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:725500bbc53c11b9f1c7683757ed7337bc9245cceaf1b0d306ff894b8e3aa03b -size 12204 +oid sha256:1a5c1f036d78a7f1a94f65d7815126b3479c18e1be244c11d591c0be8c2b2526 +size 12890 diff --git a/lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 099ed39917456ee7ffd200dd63cb4e5eb1f47afc..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:92c6faec7b2e6d8fc25416a319fdf712c647c5a69fd3ff18acc3fbbed6cef067 -size 584032 diff --git a/lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 4ccbd70f74ea32483b4447de42cad5e426bf856e..a591fb11e2533199b589741556cfa01a5fe018f4 100644 --- a/lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,173 +1,173 @@ { "results": { "crows_pairs": { - "likelihood_diff,none": 3.495266845557543, - "likelihood_diff_stderr,none": 0.4801070978651895, - "pct_stereotype,none": 0.5596302921884317, - "pct_stereotype_stderr,none": 0.09249279538424071, + "likelihood_diff,none": 3.4815705128205128, + "likelihood_diff_stderr,none": 0.47663007871133223, + "pct_stereotype,none": 0.5603756708407871, + "pct_stereotype_stderr,none": 0.0944668961491123, "alias": "crows_pairs" }, "crows_pairs_english": { - "likelihood_diff,none": 3.486583184257603, - "likelihood_diff_stderr,none": 0.0817348969033628, - "pct_stereotype,none": 0.6440071556350626, - "pct_stereotype_stderr,none": 0.011695774156934215, + "likelihood_diff,none": 3.4655635062611805, + "likelihood_diff_stderr,none": 0.08189260928515339, + "pct_stereotype,none": 0.6469886702444841, + "pct_stereotype_stderr,none": 0.011673622705751152, "alias": " - crows_pairs_english" }, "crows_pairs_english_age": { - "likelihood_diff,none": 3.8035714285714284, - "likelihood_diff_stderr,none": 0.36700441725070276, - "pct_stereotype,none": 0.7362637362637363, - "pct_stereotype_stderr,none": 0.04644942852497395, + "likelihood_diff,none": 3.7747252747252746, + "likelihood_diff_stderr,none": 0.3733542426333088, + "pct_stereotype,none": 0.7252747252747253, + "pct_stereotype_stderr,none": 0.047052133987784364, "alias": " - crows_pairs_english_age" }, "crows_pairs_english_autre": { - "likelihood_diff,none": 5.636363636363637, - "likelihood_diff_stderr,none": 1.6544065251361963, + "likelihood_diff,none": 5.715909090909091, + "likelihood_diff_stderr,none": 1.6801050857531363, "pct_stereotype,none": 0.8181818181818182, "pct_stereotype_stderr,none": 0.12196734422726124, "alias": " - crows_pairs_english_autre" }, "crows_pairs_english_disability": { - "likelihood_diff,none": 6.040384615384616, - "likelihood_diff_stderr,none": 0.5931747446888535, + "likelihood_diff,none": 6.015384615384615, + "likelihood_diff_stderr,none": 0.5966623141775952, "pct_stereotype,none": 0.6923076923076923, "pct_stereotype_stderr,none": 0.05769230769230768, "alias": " - crows_pairs_english_disability" }, "crows_pairs_english_gender": { - "likelihood_diff,none": 2.6453125, - "likelihood_diff_stderr,none": 0.15638980324770227, - "pct_stereotype,none": 0.66875, - "pct_stereotype_stderr,none": 0.026352055679927412, + "likelihood_diff,none": 2.60703125, + "likelihood_diff_stderr,none": 0.157265210678921, + "pct_stereotype,none": 0.6625, + "pct_stereotype_stderr,none": 0.026474909752348248, "alias": " - crows_pairs_english_gender" }, "crows_pairs_english_nationality": { - "likelihood_diff,none": 3.4375, - "likelihood_diff_stderr,none": 0.22499431756629582, - "pct_stereotype,none": 0.625, - "pct_stereotype_stderr,none": 0.033016908987210894, + "likelihood_diff,none": 3.392361111111111, + "likelihood_diff_stderr,none": 0.22640739112896405, + "pct_stereotype,none": 0.6018518518518519, + "pct_stereotype_stderr,none": 0.033384734032074016, "alias": " - crows_pairs_english_nationality" }, "crows_pairs_english_physical_appearance": { - "likelihood_diff,none": 3.795138888888889, - "likelihood_diff_stderr,none": 0.30655839797908097, - "pct_stereotype,none": 0.75, - "pct_stereotype_stderr,none": 0.051389153237064875, + "likelihood_diff,none": 3.779513888888889, + "likelihood_diff_stderr,none": 0.3056216112222913, + "pct_stereotype,none": 0.7638888888888888, + "pct_stereotype_stderr,none": 0.050401578099733044, "alias": " - crows_pairs_english_physical_appearance" }, "crows_pairs_english_race_color": { - "likelihood_diff,none": 3.198080708661417, - "likelihood_diff_stderr,none": 0.13916619510315903, - "pct_stereotype,none": 0.5334645669291339, - "pct_stereotype_stderr,none": 0.022155988267174086, + "likelihood_diff,none": 3.1764271653543306, + "likelihood_diff_stderr,none": 0.1388300491868073, + "pct_stereotype,none": 0.5413385826771654, + "pct_stereotype_stderr,none": 0.022129755490549068, "alias": " - crows_pairs_english_race_color" }, "crows_pairs_english_religion": { - "likelihood_diff,none": 3.375, - "likelihood_diff_stderr,none": 0.29087437727978616, - "pct_stereotype,none": 0.7207207207207207, - "pct_stereotype_stderr,none": 0.042776625248814405, + "likelihood_diff,none": 3.310810810810811, + "likelihood_diff_stderr,none": 0.28480970226640806, + "pct_stereotype,none": 0.7657657657657657, + "pct_stereotype_stderr,none": 0.04038097636567093, "alias": " - crows_pairs_english_religion" }, "crows_pairs_english_sexual_orientation": { - "likelihood_diff,none": 4.275537634408602, - "likelihood_diff_stderr,none": 0.4369071115395937, + "likelihood_diff,none": 4.282258064516129, + "likelihood_diff_stderr,none": 0.4391141133147601, "pct_stereotype,none": 0.8602150537634409, "pct_stereotype_stderr,none": 0.036152622588464155, "alias": " - crows_pairs_english_sexual_orientation" }, "crows_pairs_english_socioeconomic": { - "likelihood_diff,none": 4.1322368421052635, - "likelihood_diff_stderr,none": 0.23777763671296961, - "pct_stereotype,none": 0.6526315789473685, - "pct_stereotype_stderr,none": 0.03463365347393426, + "likelihood_diff,none": 4.142763157894737, + "likelihood_diff_stderr,none": 0.23880639175146945, + "pct_stereotype,none": 0.6578947368421053, + "pct_stereotype_stderr,none": 0.03450858738901065, "alias": " - crows_pairs_english_socioeconomic" }, "crows_pairs_french": { - "likelihood_diff,none": 3.504323196183661, - "likelihood_diff_stderr,none": 0.08144843605607167, - "pct_stereotype,none": 0.47584973166368516, - "pct_stereotype_stderr,none": 0.012199044441511507, + "likelihood_diff,none": 3.497875670840787, + "likelihood_diff_stderr,none": 0.08141456707301374, + "pct_stereotype,none": 0.4752534287418008, + "pct_stereotype_stderr,none": 0.012198331374086789, "alias": " - crows_pairs_french" }, "crows_pairs_french_age": { - "likelihood_diff,none": 3.2527777777777778, - "likelihood_diff_stderr,none": 0.3503051649044672, - "pct_stereotype,none": 0.43333333333333335, - "pct_stereotype_stderr,none": 0.05252667118728807, + "likelihood_diff,none": 3.227777777777778, + "likelihood_diff_stderr,none": 0.34426468266405463, + "pct_stereotype,none": 0.4444444444444444, + "pct_stereotype_stderr,none": 0.052671718126664185, "alias": " - crows_pairs_french_age" }, "crows_pairs_french_autre": { - "likelihood_diff,none": 2.923076923076923, - "likelihood_diff_stderr,none": 0.6694959686742004, + "likelihood_diff,none": 2.9615384615384617, + "likelihood_diff_stderr,none": 0.704180598697418, "pct_stereotype,none": 0.5384615384615384, "pct_stereotype_stderr,none": 0.14390989949130545, "alias": " - crows_pairs_french_autre" }, "crows_pairs_french_disability": { - "likelihood_diff,none": 5.390151515151516, - "likelihood_diff_stderr,none": 0.5070059512588119, - "pct_stereotype,none": 0.5909090909090909, - "pct_stereotype_stderr,none": 0.060983672113630656, + "likelihood_diff,none": 5.371212121212121, + "likelihood_diff_stderr,none": 0.5079565770380846, + "pct_stereotype,none": 0.6212121212121212, + "pct_stereotype_stderr,none": 0.0601674102524024, "alias": " - crows_pairs_french_disability" }, "crows_pairs_french_gender": { - "likelihood_diff,none": 2.8169781931464173, - "likelihood_diff_stderr,none": 0.153270778844688, - "pct_stereotype,none": 0.4984423676012461, - "pct_stereotype_stderr,none": 0.027950714088670347, + "likelihood_diff,none": 2.80607476635514, + "likelihood_diff_stderr,none": 0.1529387267167066, + "pct_stereotype,none": 0.48909657320872274, + "pct_stereotype_stderr,none": 0.027944203070818643, "alias": " - crows_pairs_french_gender" }, "crows_pairs_french_nationality": { - "likelihood_diff,none": 4.392292490118577, - "likelihood_diff_stderr,none": 0.23978854995848622, - "pct_stereotype,none": 0.308300395256917, - "pct_stereotype_stderr,none": 0.02909012143059231, + "likelihood_diff,none": 4.414031620553359, + "likelihood_diff_stderr,none": 0.23840899623275094, + "pct_stereotype,none": 0.31225296442687744, + "pct_stereotype_stderr,none": 0.02919223713357907, "alias": " - crows_pairs_french_nationality" }, "crows_pairs_french_physical_appearance": { - "likelihood_diff,none": 3.6875, - "likelihood_diff_stderr,none": 0.42158917388097905, - "pct_stereotype,none": 0.5277777777777778, - "pct_stereotype_stderr,none": 0.05924743948371486, + "likelihood_diff,none": 3.6770833333333335, + "likelihood_diff_stderr,none": 0.4180562948095515, + "pct_stereotype,none": 0.5138888888888888, + "pct_stereotype_stderr,none": 0.05931618532716555, "alias": " - crows_pairs_french_physical_appearance" }, "crows_pairs_french_race_color": { - "likelihood_diff,none": 3.055978260869565, - "likelihood_diff_stderr,none": 0.14322783808086864, - "pct_stereotype,none": 0.4152173913043478, - "pct_stereotype_stderr,none": 0.02300004306440787, + "likelihood_diff,none": 3.0456521739130435, + "likelihood_diff_stderr,none": 0.1433029684316863, + "pct_stereotype,none": 0.41304347826086957, + "pct_stereotype_stderr,none": 0.022982353907431446, "alias": " - crows_pairs_french_race_color" }, "crows_pairs_french_religion": { - "likelihood_diff,none": 3.6804347826086956, - "likelihood_diff_stderr,none": 0.32936541604904185, - "pct_stereotype,none": 0.6260869565217392, - "pct_stereotype_stderr,none": 0.04531585828644964, + "likelihood_diff,none": 3.6543478260869566, + "likelihood_diff_stderr,none": 0.33007440000905736, + "pct_stereotype,none": 0.6, + "pct_stereotype_stderr,none": 0.04588314677411234, "alias": " - crows_pairs_french_religion" }, "crows_pairs_french_sexual_orientation": { - "likelihood_diff,none": 4.019230769230769, - "likelihood_diff_stderr,none": 0.3031235127786174, + "likelihood_diff,none": 4.087912087912088, + "likelihood_diff_stderr,none": 0.3075508243527189, "pct_stereotype,none": 0.7472527472527473, "pct_stereotype_stderr,none": 0.04580951853732889, "alias": " - crows_pairs_french_sexual_orientation" }, "crows_pairs_french_socioeconomic": { - "likelihood_diff,none": 3.649234693877551, - "likelihood_diff_stderr,none": 0.24448107054887983, - "pct_stereotype,none": 0.5357142857142857, - "pct_stereotype_stderr,none": 0.03571428571428571, + "likelihood_diff,none": 3.6463647959183674, + "likelihood_diff_stderr,none": 0.24452528878254828, + "pct_stereotype,none": 0.5459183673469388, + "pct_stereotype_stderr,none": 0.035654431417332814, "alias": " - crows_pairs_french_socioeconomic" } }, "groups": { "crows_pairs": { - "likelihood_diff,none": 3.495266845557543, - "likelihood_diff_stderr,none": 0.4801070978651895, - "pct_stereotype,none": 0.5596302921884317, - "pct_stereotype_stderr,none": 0.09249279538424071, + "likelihood_diff,none": 3.4815705128205128, + "likelihood_diff_stderr,none": 0.47663007871133223, + "pct_stereotype,none": 0.5603756708407871, + "pct_stereotype_stderr,none": 0.0944668961491123, "alias": "crows_pairs" } }, @@ -1038,13 +1038,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index badbf1bef8ed97702d2ca6fc95c145193dabe8f5..8eeb3b74ef5bf06ef542bc457a0d6387e8911ecf 100644 --- a/lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/crows_pairs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:5d8dd078616973821c6ea3dddbc42248464661a8b5e52cbcc34ca867591d2c29 -size 114634 +oid sha256:ebfa42286734a17eafc8caed854007f738497389888b4b6d224186002d0c5645 +size 113857 diff --git a/lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 5e11e9ab9eab42aa52f414c1b82b502cf567e905..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:9e5b0108b4e2f44c40c59f69242940c6b9d747f51f5a8e69cfdd4f06d7173924 -size 196355 diff --git a/lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 240ef27df2e00f23f5a9e3d5dcf18354226d6b26..ea415475a06e4b3c30c4ae54f98753aabcadd8a8 100644 --- a/lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,20 +1,20 @@ { "results": { "freebase": { - "exact_match,none": 0.030019685039370077, - "exact_match_stderr,none": 0.0037864267909448347, + "exact_match,none": 0.028543307086614175, + "exact_match_stderr,none": 0.0036949528903927557, "alias": "freebase" }, "webqs": { - "exact_match,none": 0.030019685039370077, - "exact_match_stderr,none": 0.0037864267909448347, + "exact_match,none": 0.028543307086614175, + "exact_match_stderr,none": 0.0036949528903927557, "alias": " - webqs" } }, "groups": { "freebase": { - "exact_match,none": 0.030019685039370077, - "exact_match_stderr,none": 0.0037864267909448347, + "exact_match,none": 0.028543307086614175, + "exact_match_stderr,none": 0.0036949528903927557, "alias": "freebase" } }, @@ -60,13 +60,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index af85892841655944edcc33f30beb4defd3579a33..5937b16bbcbcf254855a65138a419f1e1e3a07f3 100644 --- a/lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/freebase/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:bee68e93e72fc24ff11e7f634462aff2aa17ca4f7e8474b5cfb8bafdd028c6a8 -size 13684 +oid sha256:8d914584ef70ee0864d3088a1baf759abdc27445c582b251bfe3119a8d31f6f4 +size 12737 diff --git a/lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 0ee34ef548124b62a83169c9c70b1df7ce8e04d2..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:ae30d40c725bd76623e64721101a1b5a6f2f9e7c9c3a2795d4f281e450f0df8a -size 8108576 diff --git a/lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 72d3f6d79b2cf83e56104ac5a4cd7556af3687b3..c3446df571d26b302da80bdd475362de366bcf7c 100644 --- a/lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,27 +1,27 @@ { "results": { "glue": { - "acc,none": 0.4743941669146658, - "acc_stderr,none": 0.0769464078117324, - "f1,none": 0.44895195516590214, - "f1_stderr,none": 0.0012764956755924068, - "mcc,none": 0.012008657795714008, - "mcc_stderr,none": 0.0009871680384745693, + "acc,none": 0.47458134817465414, + "acc_stderr,none": 0.07691289703225153, + "f1,none": 0.444689593964169, + "f1_stderr,none": 0.0013063875646580627, + "mcc,none": 0.003737743780434562, + "mcc_stderr,none": 0.0009741365404113992, "alias": "glue" }, "cola": { - "mcc,none": 0.012008657795714008, - "mcc_stderr,none": 0.03141923039277966, + "mcc,none": 0.003737743780434562, + "mcc_stderr,none": 0.03121116051048726, "alias": " - cola" }, "mnli": { - "acc,none": 0.32776362710137547, - "acc_stderr,none": 0.004738254189734395, + "acc,none": 0.32929190015282733, + "acc_stderr,none": 0.004743886315223882, "alias": " - mnli" }, "mnli_mismatch": { - "acc,none": 0.3342148087876322, - "acc_stderr,none": 0.004757524813728244, + "acc,none": 0.33350284784377543, + "acc_stderr,none": 0.0047549951070959134, "alias": " - mnli_mismatch" }, "mrpc": { @@ -32,25 +32,25 @@ "alias": " - mrpc" }, "qnli": { - "acc,none": 0.4964305326743548, - "acc_stderr,none": 0.006765238152075668, + "acc,none": 0.49697968149368477, + "acc_stderr,none": 0.0067652871181183415, "alias": " - qnli" }, "qqp": { - "acc,none": 0.5348998268612416, - "acc_stderr,none": 0.0024806356431606442, - "f1,none": 0.44540789240842327, - "f1_stderr,none": 0.0033657259190706915, + "acc,none": 0.5346772198862231, + "acc_stderr,none": 0.002480712860000902, + "f1,none": 0.4411039481892992, + "f1_stderr,none": 0.003378627814144618, "alias": " - qqp" }, "rte": { - "acc,none": 0.5342960288808665, - "acc_stderr,none": 0.030025579819366422, + "acc,none": 0.5451263537906137, + "acc_stderr,none": 0.029973636495415252, "alias": " - rte" }, "sst2": { - "acc,none": 0.5642201834862385, - "acc_stderr,none": 0.016801528278889092, + "acc,none": 0.573394495412844, + "acc_stderr,none": 0.016758336618033463, "alias": " - sst2" }, "wnli": { @@ -61,12 +61,12 @@ }, "groups": { "glue": { - "acc,none": 0.4743941669146658, - "acc_stderr,none": 0.0769464078117324, - "f1,none": 0.44895195516590214, - "f1_stderr,none": 0.0012764956755924068, - "mcc,none": 0.012008657795714008, - "mcc_stderr,none": 0.0009871680384745693, + "acc,none": 0.47458134817465414, + "acc_stderr,none": 0.07691289703225153, + "f1,none": 0.444689593964169, + "f1_stderr,none": 0.0013063875646580627, + "mcc,none": 0.003737743780434562, + "mcc_stderr,none": 0.0009741365404113992, "alias": "glue" } }, @@ -360,13 +360,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 16 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 7e5dd1a633f3e2b2bdf4e75e50d9fd34301bc9ba..e7d0436cc19627fd93de740d11b4fc446e82b08d 100644 --- a/lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/glue/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:100a4bf926866e471d22d31b9b24838bf1fbdae341af7509549b3ecfe4a0bc37 -size 177718 +oid sha256:00eab5e98b32ede57f9fcbc2e65ab8d2b3686aba7dbb4279ac2da0dae0d2dd25 +size 104469 diff --git a/lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 6015434aaff7f6968ca4b7047b1512f9a04c1b00..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:322be9fa8109be50e844677c8acad4b7e1b5f42dd02db29c4a3b7bf35401c4d9 -size 1598306 diff --git a/lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 97434ec9d8aee66bb443e11cf1785680f3e677b7..ebd2d051051907724170dd5310eca3a3cca47daa 100644 --- a/lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "gsm8k": { - "exact_match,get-answer": 0.039423805913570885, - "exact_match_stderr,get-answer": 0.005360280030342464, + "exact_match,get-answer": 0.04245640636846096, + "exact_match_stderr,get-answer": 0.005553837749990044, "alias": "gsm8k" } }, @@ -76,7 +76,7 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", + "batch_size": "auto", "batch_sizes": [], "device": null, "use_cache": null, @@ -84,5 +84,5 @@ "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index eb1f766a4848f0b189638ef9722b85bf7ce40c86..53148ee1839882f6f3111f7bbfe7b45150ff581d 100644 --- a/lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/gsm8k/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:f64706f6d619c9e9c72c33b701a71b74fb218d3fd1d12bfee4c6102ffcec3606 -size 14785 +oid sha256:10fe1ee1afc331d8586fd52aa8cac721c131f3700001f647dcfdd24fed23d6d9 +size 15430 diff --git a/lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 2ea22deb54a3c9bdafdf3984ebdceed2263d7bc4..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:7325c9531ea107d3dc4441b5b296279ab19d62c42e5f254c84871ec6815e8d6c -size 4886431 diff --git a/lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 58b67f48161e2afa1a2c44923ae6d6df14165796..8e6154d6751924403ea2b12705396399126c93bf 100644 --- a/lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,10 +1,10 @@ { "results": { "hellaswag": { - "acc,none": 0.5572595100577574, - "acc_stderr,none": 0.004956953917781309, - "acc_norm,none": 0.7547301334395539, - "acc_norm_stderr,none": 0.004293677871726322, + "acc,none": 0.5571599283011353, + "acc_stderr,none": 0.0049570683775165105, + "acc_norm,none": 0.755327623979287, + "acc_norm_stderr,none": 0.0042901420299216834, "alias": "hellaswag" } }, @@ -53,13 +53,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 32 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index ea2837d23f2bfcb09c94e7738f4539f57186318f..e6708aa967f278d2bf93b11cc2104c31a3f9f7b2 100644 --- a/lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/hellaswag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:dddb0d4e4ba3d331930b7860983a07c42ad3a12873e3d5d8117116db133b9d85 -size 50684 +oid sha256:4b624507a14b86e057bb2c74dfd645cf274d3ca96d6ae412b2c2621c270c1193 +size 24651 diff --git a/lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 080a18ebeda013985c1d59b2f73b8b0f146cb545..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:3bf5708491f7b497dc167baf2aeca6e490271a80d56c3487d2b4203b002a80aa -size 7830891 diff --git a/lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index c6022b22a435208a112333ab54a41a19ece10a85..2e013e2554f8981259348e47737cf2d0f28ad16b 100644 --- a/lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,10 +1,10 @@ { "results": { "kmmlu": { - "acc,none": 0.270112619116373, - "acc_stderr,none": 0.02426918657426541, - "acc_norm,none": 0.270112619116373, - "acc_norm_stderr,none": 0.02426918657426541, + "acc,none": 0.26936182500721906, + "acc_stderr,none": 0.024299898599051413, + "acc_norm,none": 0.26936182500721906, + "acc_norm_stderr,none": 0.024299898599051413, "alias": "kmmlu" }, "kmmlu_accounting": { @@ -15,59 +15,59 @@ "alias": " - kmmlu_accounting" }, "kmmlu_agricultural_sciences": { - "acc,none": 0.263, - "acc_stderr,none": 0.013929286594259726, - "acc_norm,none": 0.263, - "acc_norm_stderr,none": 0.013929286594259726, + "acc,none": 0.261, + "acc_stderr,none": 0.013895037677965136, + "acc_norm,none": 0.261, + "acc_norm_stderr,none": 0.013895037677965136, "alias": " - kmmlu_agricultural_sciences" }, "kmmlu_aviation_engineering_and_maintenance": { - "acc,none": 0.274, - "acc_stderr,none": 0.014111099288259588, - "acc_norm,none": 0.274, - "acc_norm_stderr,none": 0.014111099288259588, + "acc,none": 0.27, + "acc_stderr,none": 0.014046255632633913, + "acc_norm,none": 0.27, + "acc_norm_stderr,none": 0.014046255632633913, "alias": " - kmmlu_aviation_engineering_and_maintenance" }, "kmmlu_biology": { - "acc,none": 0.261, - "acc_stderr,none": 0.013895037677965131, - "acc_norm,none": 0.261, - "acc_norm_stderr,none": 0.013895037677965131, + "acc,none": 0.26, + "acc_stderr,none": 0.013877773329774164, + "acc_norm,none": 0.26, + "acc_norm_stderr,none": 0.013877773329774164, "alias": " - kmmlu_biology" }, "kmmlu_chemical_engineering": { - "acc,none": 0.276, - "acc_stderr,none": 0.01414298497574067, - "acc_norm,none": 0.276, - "acc_norm_stderr,none": 0.01414298497574067, + "acc,none": 0.27, + "acc_stderr,none": 0.014046255632633916, + "acc_norm,none": 0.27, + "acc_norm_stderr,none": 0.014046255632633916, "alias": " - kmmlu_chemical_engineering" }, "kmmlu_chemistry": { - "acc,none": 0.27166666666666667, - "acc_stderr,none": 0.018174809149686423, - "acc_norm,none": 0.27166666666666667, - "acc_norm_stderr,none": 0.018174809149686423, + "acc,none": 0.26666666666666666, + "acc_stderr,none": 0.01806848202433441, + "acc_norm,none": 0.26666666666666666, + "acc_norm_stderr,none": 0.01806848202433441, "alias": " - kmmlu_chemistry" }, "kmmlu_civil_engineering": { - "acc,none": 0.31, - "acc_stderr,none": 0.014632638658632903, - "acc_norm,none": 0.31, - "acc_norm_stderr,none": 0.014632638658632903, + "acc,none": 0.315, + "acc_stderr,none": 0.014696631960792498, + "acc_norm,none": 0.315, + "acc_norm_stderr,none": 0.014696631960792498, "alias": " - kmmlu_civil_engineering" }, "kmmlu_computer_science": { - "acc,none": 0.283, - "acc_stderr,none": 0.014251810906481754, - "acc_norm,none": 0.283, - "acc_norm_stderr,none": 0.014251810906481754, + "acc,none": 0.279, + "acc_stderr,none": 0.014190150117612033, + "acc_norm,none": 0.279, + "acc_norm_stderr,none": 0.014190150117612033, "alias": " - kmmlu_computer_science" }, "kmmlu_construction": { - "acc,none": 0.287, - "acc_stderr,none": 0.014312087053809963, - "acc_norm,none": 0.287, - "acc_norm_stderr,none": 0.014312087053809963, + "acc,none": 0.284, + "acc_stderr,none": 0.014267009061031314, + "acc_norm,none": 0.284, + "acc_norm_stderr,none": 0.014267009061031314, "alias": " - kmmlu_construction" }, "kmmlu_criminal_law": { @@ -78,45 +78,45 @@ "alias": " - kmmlu_criminal_law" }, "kmmlu_ecology": { - "acc,none": 0.27, - "acc_stderr,none": 0.014046255632633913, - "acc_norm,none": 0.27, - "acc_norm_stderr,none": 0.014046255632633913, + "acc,none": 0.276, + "acc_stderr,none": 0.014142984975740663, + "acc_norm,none": 0.276, + "acc_norm_stderr,none": 0.014142984975740663, "alias": " - kmmlu_ecology" }, "kmmlu_economics": { - "acc,none": 0.2076923076923077, - "acc_stderr,none": 0.03571595663393523, - "acc_norm,none": 0.2076923076923077, - "acc_norm_stderr,none": 0.03571595663393523, + "acc,none": 0.2230769230769231, + "acc_stderr,none": 0.03665400868201044, + "acc_norm,none": 0.2230769230769231, + "acc_norm_stderr,none": 0.03665400868201044, "alias": " - kmmlu_economics" }, "kmmlu_education": { - "acc,none": 0.24, - "acc_stderr,none": 0.042923469599092816, - "acc_norm,none": 0.24, - "acc_norm_stderr,none": 0.042923469599092816, + "acc,none": 0.23, + "acc_stderr,none": 0.04229525846816505, + "acc_norm,none": 0.23, + "acc_norm_stderr,none": 0.04229525846816505, "alias": " - kmmlu_education" }, "kmmlu_electrical_engineering": { - "acc,none": 0.312, - "acc_stderr,none": 0.014658474370509001, - "acc_norm,none": 0.312, - "acc_norm_stderr,none": 0.014658474370509001, + "acc,none": 0.309, + "acc_stderr,none": 0.014619600977206493, + "acc_norm,none": 0.309, + "acc_norm_stderr,none": 0.014619600977206493, "alias": " - kmmlu_electrical_engineering" }, "kmmlu_electronics_engineering": { - "acc,none": 0.28, - "acc_stderr,none": 0.01420569610409151, - "acc_norm,none": 0.28, - "acc_norm_stderr,none": 0.01420569610409151, + "acc,none": 0.271, + "acc_stderr,none": 0.014062601350986182, + "acc_norm,none": 0.271, + "acc_norm_stderr,none": 0.014062601350986182, "alias": " - kmmlu_electronics_engineering" }, "kmmlu_energy_management": { - "acc,none": 0.272, - "acc_stderr,none": 0.014078856992462623, - "acc_norm,none": 0.272, - "acc_norm_stderr,none": 0.014078856992462623, + "acc,none": 0.275, + "acc_stderr,none": 0.014127086556490524, + "acc_norm,none": 0.275, + "acc_norm_stderr,none": 0.014127086556490524, "alias": " - kmmlu_energy_management" }, "kmmlu_environmental_science": { @@ -127,31 +127,31 @@ "alias": " - kmmlu_environmental_science" }, "kmmlu_fashion": { - "acc,none": 0.265, - "acc_stderr,none": 0.013963164754809949, - "acc_norm,none": 0.265, - "acc_norm_stderr,none": 0.013963164754809949, + "acc,none": 0.261, + "acc_stderr,none": 0.01389503767796512, + "acc_norm,none": 0.261, + "acc_norm_stderr,none": 0.01389503767796512, "alias": " - kmmlu_fashion" }, "kmmlu_food_processing": { - "acc,none": 0.243, - "acc_stderr,none": 0.013569640199177434, - "acc_norm,none": 0.243, - "acc_norm_stderr,none": 0.013569640199177434, + "acc,none": 0.239, + "acc_stderr,none": 0.013493000446937587, + "acc_norm,none": 0.239, + "acc_norm_stderr,none": 0.013493000446937587, "alias": " - kmmlu_food_processing" }, "kmmlu_gas_technology_and_engineering": { - "acc,none": 0.272, - "acc_stderr,none": 0.014078856992462611, - "acc_norm,none": 0.272, - "acc_norm_stderr,none": 0.014078856992462611, + "acc,none": 0.268, + "acc_stderr,none": 0.014013292702729482, + "acc_norm,none": 0.268, + "acc_norm_stderr,none": 0.014013292702729482, "alias": " - kmmlu_gas_technology_and_engineering" }, "kmmlu_geomatics": { - "acc,none": 0.263, - "acc_stderr,none": 0.013929286594259722, - "acc_norm,none": 0.263, - "acc_norm_stderr,none": 0.013929286594259722, + "acc,none": 0.264, + "acc_stderr,none": 0.013946271849440469, + "acc_norm,none": 0.264, + "acc_norm_stderr,none": 0.013946271849440469, "alias": " - kmmlu_geomatics" }, "kmmlu_health": { @@ -162,80 +162,80 @@ "alias": " - kmmlu_health" }, "kmmlu_industrial_engineer": { - "acc,none": 0.277, - "acc_stderr,none": 0.014158794845306263, - "acc_norm,none": 0.277, - "acc_norm_stderr,none": 0.014158794845306263, + "acc,none": 0.276, + "acc_stderr,none": 0.014142984975740671, + "acc_norm,none": 0.276, + "acc_norm_stderr,none": 0.014142984975740671, "alias": " - kmmlu_industrial_engineer" }, "kmmlu_information_technology": { - "acc,none": 0.274, - "acc_stderr,none": 0.014111099288259585, - "acc_norm,none": 0.274, - "acc_norm_stderr,none": 0.014111099288259585, + "acc,none": 0.271, + "acc_stderr,none": 0.014062601350986186, + "acc_norm,none": 0.271, + "acc_norm_stderr,none": 0.014062601350986186, "alias": " - kmmlu_information_technology" }, "kmmlu_interior_architecture_and_design": { - "acc,none": 0.286, - "acc_stderr,none": 0.01429714686251791, - "acc_norm,none": 0.286, - "acc_norm_stderr,none": 0.01429714686251791, + "acc,none": 0.29, + "acc_stderr,none": 0.014356395999905694, + "acc_norm,none": 0.29, + "acc_norm_stderr,none": 0.014356395999905694, "alias": " - kmmlu_interior_architecture_and_design" }, "kmmlu_law": { - "acc,none": 0.255, - "acc_stderr,none": 0.013790038620872828, - "acc_norm,none": 0.255, - "acc_norm_stderr,none": 0.013790038620872828, + "acc,none": 0.257, + "acc_stderr,none": 0.013825416526895031, + "acc_norm,none": 0.257, + "acc_norm_stderr,none": 0.013825416526895031, "alias": " - kmmlu_law" }, "kmmlu_machine_design_and_manufacturing": { - "acc,none": 0.25, - "acc_stderr,none": 0.013699915608779773, - "acc_norm,none": 0.25, - "acc_norm_stderr,none": 0.013699915608779773, + "acc,none": 0.254, + "acc_stderr,none": 0.013772206565168543, + "acc_norm,none": 0.254, + "acc_norm_stderr,none": 0.013772206565168543, "alias": " - kmmlu_machine_design_and_manufacturing" }, "kmmlu_management": { - "acc,none": 0.277, - "acc_stderr,none": 0.014158794845306265, - "acc_norm,none": 0.277, - "acc_norm_stderr,none": 0.014158794845306265, + "acc,none": 0.274, + "acc_stderr,none": 0.01411109928825958, + "acc_norm,none": 0.274, + "acc_norm_stderr,none": 0.01411109928825958, "alias": " - kmmlu_management" }, "kmmlu_maritime_engineering": { - "acc,none": 0.23333333333333334, - "acc_stderr,none": 0.01728137411493407, - "acc_norm,none": 0.23333333333333334, - "acc_norm_stderr,none": 0.01728137411493407, + "acc,none": 0.23666666666666666, + "acc_stderr,none": 0.01736649795856464, + "acc_norm,none": 0.23666666666666666, + "acc_norm_stderr,none": 0.01736649795856464, "alias": " - kmmlu_maritime_engineering" }, "kmmlu_marketing": { - "acc,none": 0.3, - "acc_stderr,none": 0.014498627873361428, - "acc_norm,none": 0.3, - "acc_norm_stderr,none": 0.014498627873361428, + "acc,none": 0.293, + "acc_stderr,none": 0.014399942998441271, + "acc_norm,none": 0.293, + "acc_norm_stderr,none": 0.014399942998441271, "alias": " - kmmlu_marketing" }, "kmmlu_materials_engineering": { - "acc,none": 0.235, - "acc_stderr,none": 0.013414729030247124, - "acc_norm,none": 0.235, - "acc_norm_stderr,none": 0.013414729030247124, + "acc,none": 0.234, + "acc_stderr,none": 0.01339490288966001, + "acc_norm,none": 0.234, + "acc_norm_stderr,none": 0.01339490288966001, "alias": " - kmmlu_materials_engineering" }, "kmmlu_mechanical_engineering": { - "acc,none": 0.268, - "acc_stderr,none": 0.014013292702729493, - "acc_norm,none": 0.268, - "acc_norm_stderr,none": 0.014013292702729493, + "acc,none": 0.269, + "acc_stderr,none": 0.014029819522568193, + "acc_norm,none": 0.269, + "acc_norm_stderr,none": 0.014029819522568193, "alias": " - kmmlu_mechanical_engineering" }, "kmmlu_nondestructive_testing": { - "acc,none": 0.233, - "acc_stderr,none": 0.013374972519220072, - "acc_norm,none": 0.233, - "acc_norm_stderr,none": 0.013374972519220072, + "acc,none": 0.229, + "acc_stderr,none": 0.013294199326613595, + "acc_norm,none": 0.229, + "acc_norm_stderr,none": 0.013294199326613595, "alias": " - kmmlu_nondestructive_testing" }, "kmmlu_patent": { @@ -246,75 +246,75 @@ "alias": " - kmmlu_patent" }, "kmmlu_political_science_and_sociology": { - "acc,none": 0.2833333333333333, - "acc_stderr,none": 0.026059845940064958, - "acc_norm,none": 0.2833333333333333, - "acc_norm_stderr,none": 0.026059845940064958, + "acc,none": 0.29, + "acc_stderr,none": 0.026241737033622703, + "acc_norm,none": 0.29, + "acc_norm_stderr,none": 0.026241737033622703, "alias": " - kmmlu_political_science_and_sociology" }, "kmmlu_psychology": { - "acc,none": 0.236, - "acc_stderr,none": 0.013434451402438692, - "acc_norm,none": 0.236, - "acc_norm_stderr,none": 0.013434451402438692, + "acc,none": 0.234, + "acc_stderr,none": 0.01339490288966001, + "acc_norm,none": 0.234, + "acc_norm_stderr,none": 0.01339490288966001, "alias": " - kmmlu_psychology" }, "kmmlu_public_safety": { - "acc,none": 0.292, - "acc_stderr,none": 0.014385511563477341, - "acc_norm,none": 0.292, - "acc_norm_stderr,none": 0.014385511563477341, + "acc,none": 0.301, + "acc_stderr,none": 0.014512395033543159, + "acc_norm,none": 0.301, + "acc_norm_stderr,none": 0.014512395033543159, "alias": " - kmmlu_public_safety" }, "kmmlu_railway_and_automotive_engineering": { - "acc,none": 0.249, - "acc_stderr,none": 0.01368160027870232, - "acc_norm,none": 0.249, - "acc_norm_stderr,none": 0.01368160027870232, + "acc,none": 0.25, + "acc_stderr,none": 0.013699915608779773, + "acc_norm,none": 0.25, + "acc_norm_stderr,none": 0.013699915608779773, "alias": " - kmmlu_railway_and_automotive_engineering" }, "kmmlu_real_estate": { - "acc,none": 0.28, - "acc_stderr,none": 0.03182868716477582, - "acc_norm,none": 0.28, - "acc_norm_stderr,none": 0.03182868716477582, + "acc,none": 0.275, + "acc_stderr,none": 0.031652557907861936, + "acc_norm,none": 0.275, + "acc_norm_stderr,none": 0.031652557907861936, "alias": " - kmmlu_real_estate" }, "kmmlu_refrigerating_machinery": { - "acc,none": 0.267, - "acc_stderr,none": 0.01399667485179625, - "acc_norm,none": 0.267, - "acc_norm_stderr,none": 0.01399667485179625, + "acc,none": 0.263, + "acc_stderr,none": 0.013929286594259748, + "acc_norm,none": 0.263, + "acc_norm_stderr,none": 0.013929286594259748, "alias": " - kmmlu_refrigerating_machinery" }, "kmmlu_social_welfare": { - "acc,none": 0.252, - "acc_stderr,none": 0.013736254390651152, - "acc_norm,none": 0.252, - "acc_norm_stderr,none": 0.013736254390651152, + "acc,none": 0.258, + "acc_stderr,none": 0.013842963108656604, + "acc_norm,none": 0.258, + "acc_norm_stderr,none": 0.013842963108656604, "alias": " - kmmlu_social_welfare" }, "kmmlu_taxation": { - "acc,none": 0.285, - "acc_stderr,none": 0.031999921482315764, - "acc_norm,none": 0.285, - "acc_norm_stderr,none": 0.031999921482315764, + "acc,none": 0.295, + "acc_stderr,none": 0.03232801420614269, + "acc_norm,none": 0.295, + "acc_norm_stderr,none": 0.03232801420614269, "alias": " - kmmlu_taxation" }, "kmmlu_telecommunications_and_wireless_technology": { - "acc,none": 0.284, - "acc_stderr,none": 0.014267009061031307, - "acc_norm,none": 0.284, - "acc_norm_stderr,none": 0.014267009061031307, + "acc,none": 0.282, + "acc_stderr,none": 0.014236526215291341, + "acc_norm,none": 0.282, + "acc_norm_stderr,none": 0.014236526215291341, "alias": " - kmmlu_telecommunications_and_wireless_technology" } }, "groups": { "kmmlu": { - "acc,none": 0.270112619116373, - "acc_stderr,none": 0.02426918657426541, - "acc_norm,none": 0.270112619116373, - "acc_norm_stderr,none": 0.02426918657426541, + "acc,none": 0.26936182500721906, + "acc_stderr,none": 0.024299898599051413, + "acc_norm,none": 0.26936182500721906, + "acc_norm_stderr,none": 0.024299898599051413, "alias": "kmmlu" } }, @@ -2092,13 +2092,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 4 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 26c4ceb0838d33a5e5eb222995b65a0a380d8da7..e09fc008cbaf7cbeb392cddc583eaf3dadb54f93 100644 --- a/lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/kmmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:b0aec920764a16911898552fd613294992dd057b13166b677d6bab55fb53fb62 -size 249223 +oid sha256:e8cefbf4f0873f2ef42cefc718c36a09c0c63d3e4b7431485ef4846dadf6694b +size 208768 diff --git a/lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index e48261ca7616ed751849ea5b0552ce17be5f875b..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:5c365e25bf2bae8e0185d7c9ba9a751824ff7f1a1cc21eafdb98d22c7a7b1cce -size 831752 diff --git a/lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index aa9cdea03057fb70cea13f09d86ecffad19b27fc..dad6870356528ef9f06f3a9fa4c356336396c0cd 100644 --- a/lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,41 +1,41 @@ { "results": { "kobest": { - "acc,none": 0.4867353650515238, - "acc_stderr,none": 0.03775670069085235, - "f1,none": 0.3952041743500793, + "acc,none": 0.48629686472264855, + "acc_stderr,none": 0.03845705102970606, + "f1,none": 0.3941498184514018, "f1_stderr,none": "N/A", - "acc_norm,none": 0.462, - "acc_norm_stderr,none": 0.0004981082164328657, + "acc_norm,none": 0.464, + "acc_norm_stderr,none": 0.0004984048096192345, "alias": "kobest" }, "kobest_boolq": { - "acc,none": 0.5028490028490028, - "acc_stderr,none": 0.013348550797680823, - "f1,none": 0.35511229034397807, + "acc,none": 0.5042735042735043, + "acc_stderr,none": 0.013348279916769821, + "f1,none": 0.35467032967032963, "f1_stderr,none": "N/A", "alias": " - kobest_boolq" }, "kobest_copa": { - "acc,none": 0.527, - "acc_stderr,none": 0.01579621855130262, - "f1,none": 0.5256676002892124, + "acc,none": 0.524, + "acc_stderr,none": 0.015801065586651758, + "f1,none": 0.5229906482116158, "f1_stderr,none": "N/A", "alias": " - kobest_copa" }, "kobest_hellaswag": { - "acc,none": 0.364, - "acc_stderr,none": 0.0215391706373177, - "f1,none": 0.35903108257383476, + "acc,none": 0.37, + "acc_stderr,none": 0.021613289165165785, + "f1,none": 0.3655287187202081, "f1_stderr,none": "N/A", - "acc_norm,none": 0.462, - "acc_norm_stderr,none": 0.022318338119870523, + "acc_norm,none": 0.464, + "acc_norm_stderr,none": 0.022324981738385243, "alias": " - kobest_hellaswag" }, "kobest_sentineg": { - "acc,none": 0.47858942065491183, - "acc_stderr,none": 0.025102898696363046, - "f1,none": 0.4672177380141982, + "acc,none": 0.46851385390428213, + "acc_stderr,none": 0.025076077305681316, + "f1,none": 0.4552271323122947, "f1_stderr,none": "N/A", "alias": " - kobest_sentineg" }, @@ -49,12 +49,12 @@ }, "groups": { "kobest": { - "acc,none": 0.4867353650515238, - "acc_stderr,none": 0.03775670069085235, - "f1,none": 0.3952041743500793, + "acc,none": 0.48629686472264855, + "acc_stderr,none": 0.03845705102970606, + "f1,none": 0.3941498184514018, "f1_stderr,none": "N/A", - "acc_norm,none": 0.462, - "acc_norm_stderr,none": 0.0004981082164328657, + "acc_norm,none": 0.464, + "acc_norm_stderr,none": 0.0004984048096192345, "alias": "kobest" } }, @@ -279,13 +279,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 8 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 51204350f9a3d610a2bd2896f965c82c85074b08..8946c2c7719cab1e6778617746c84129a00cd2b0 100644 --- a/lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/kobest/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:63f646c4b53d2275e59665c054311869ece7c5e3ddfba31f7cae1abb1a0b5472 -size 31416 +oid sha256:da74f1796d0de08111e8b232ced3122134f5e53a0dcd6fe1bf87e24e4228bf65 +size 38221 diff --git a/lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 5ef5a472020530279a8f20be1f04041a8ffe0dc8..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e5a3635bdb3df62be8c4b46489e5aa39526a58e152ae049a578f7a19ebf98fe9 -size 1969153 diff --git a/lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 76e059b70853ed27ff1ccc40b7c906296ad9c1d7..98f7aec81af6ebf24f41e6b285b0c61a4692752a 100644 --- a/lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,33 +1,33 @@ { "results": { "lambada": { - "perplexity,none": 4.580479771753293, - "perplexity_stderr,none": 0.24516278957673715, - "acc,none": 0.6632058994760334, - "acc_stderr,none": 0.01383973422461185, + "perplexity,none": 4.581086434995066, + "perplexity_stderr,none": 0.24535209996272006, + "acc,none": 0.6639821463225306, + "acc_stderr,none": 0.014008914961907022, "alias": "lambada" }, "lambada_openai": { - "perplexity,none": 4.132732570732677, - "perplexity_stderr,none": 0.08708403777008579, - "acc,none": 0.6875606442848826, - "acc_stderr,none": 0.006457292279746485, + "perplexity,none": 4.132928377411266, + "perplexity_stderr,none": 0.0870725561065737, + "acc,none": 0.6887250145546284, + "acc_stderr,none": 0.006450703968778299, "alias": " - lambada_openai" }, "lambada_standard": { - "perplexity,none": 5.0282269727739095, - "perplexity_stderr,none": 0.11126270336176898, - "acc,none": 0.6388511546671841, - "acc_stderr,none": 0.0066919860997724364, + "perplexity,none": 5.029244492578866, + "perplexity_stderr,none": 0.11127913513447098, + "acc,none": 0.6392392780904328, + "acc_stderr,none": 0.006690420625907091, "alias": " - lambada_standard" } }, "groups": { "lambada": { - "perplexity,none": 4.580479771753293, - "perplexity_stderr,none": 0.24516278957673715, - "acc,none": 0.6632058994760334, - "acc_stderr,none": 0.01383973422461185, + "perplexity,none": 4.581086434995066, + "perplexity_stderr,none": 0.24535209996272006, + "acc,none": 0.6639821463225306, + "acc_stderr,none": 0.014008914961907022, "alias": "lambada" } }, @@ -112,13 +112,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 32 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 89a23cc8a54584c204cd8041d9f8c1004836569c..45ab6c01a6d81cee4116c5ce17ed836ca8db97d7 100644 --- a/lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/lambada/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:ba161d5d3e6781f0f5267065b9b0b2b4a335924d50b24692fb072a01aed34aff -size 26523 +oid sha256:e6bb2ecfc031d37c3cf86cd8f928d40154cad20956535afca696e2a64a6195c2 +size 26754 diff --git a/lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index ea662986444ddb0e9b4518bf452b8d10b198e0ce..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e952e77ebd06093869c2f3e51eacdff4bf38229182b4087770d356263a6102d0 -size 1943088 diff --git a/lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index b566755e3f41de36fc2c0a1efb7a7cb0d40ec2eb..2550f310160c64f3f12a55a15370c9854c4b8fe8 100644 --- a/lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,33 +1,33 @@ { "results": { "lambada_cloze": { - "perplexity,none": 201.8808036126752, - "perplexity_stderr,none": 6.39463446029785, - "acc,none": 0.09576945468659033, - "acc_stderr,none": 0.008224018836311266, + "perplexity,none": 202.60084862007227, + "perplexity_stderr,none": 6.427806219106391, + "acc,none": 0.09411993013778382, + "acc_stderr,none": 0.008590078218125246, "alias": "lambada_cloze" }, "lambada_openai_cloze_yaml": { - "perplexity,none": 200.64451400838843, - "perplexity_stderr,none": 6.48488731628756, - "acc,none": 0.08150591888220454, - "acc_stderr,none": 0.0038119292177701976, + "perplexity,none": 201.1914466369863, + "perplexity_stderr,none": 6.506244049635805, + "acc,none": 0.07898311663108869, + "acc_stderr,none": 0.0037576212389559436, "alias": " - lambada_openai_cloze_yaml" }, "lambada_standard_cloze_yaml": { - "perplexity,none": 203.117093216962, - "perplexity_stderr,none": 6.242797363168333, - "acc,none": 0.11003299049097613, - "acc_stderr,none": 0.004359738943980749, + "perplexity,none": 204.01025060315825, + "perplexity_stderr,none": 6.270310497569642, + "acc,none": 0.10925674364447895, + "acc_stderr,none": 0.004346227651722467, "alias": " - lambada_standard_cloze_yaml" } }, "groups": { "lambada_cloze": { - "perplexity,none": 201.8808036126752, - "perplexity_stderr,none": 6.39463446029785, - "acc,none": 0.09576945468659033, - "acc_stderr,none": 0.008224018836311266, + "perplexity,none": 202.60084862007227, + "perplexity_stderr,none": 6.427806219106391, + "acc,none": 0.09411993013778382, + "acc_stderr,none": 0.008590078218125246, "alias": "lambada_cloze" } }, @@ -112,13 +112,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 32 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index aaa78626da3d94fd4d2b3239a68a6fb4b1776762..a4c48803beec951cdfba597c6aaed3a0e5c7c7a3 100644 --- a/lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/lambada_cloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:3e9de39822aa97cede187c05b0ba32445b11f34520ca5b2586b5ba11ca794a69 -size 26323 +oid sha256:13a1ea7ff27a744f6832ce754169e7999a544270af382a89c145bee14e266eaa +size 19729 diff --git a/lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index a94f214b930eac14f2bdb4dc6dc49bc5049138ba..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:9b93959b8341155d36730cc4af4c02be58e28d296548508a709d034a4129acc0 -size 1066916 diff --git a/lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 5b6de9b95290672d5278e2500e620be3e767c387..4d136fd673777cd9b75e0871446f7cdb2edfd919 100644 --- a/lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "logieval": { - "exact_match,get-answer": 0.25127226463104324, - "exact_match_stderr,get-answer": 0.01094324556925147, + "exact_match,get-answer": 0.2455470737913486, + "exact_match_stderr,get-answer": 0.010859138259206532, "alias": "logieval" } }, @@ -63,7 +63,7 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", + "batch_size": "auto", "batch_sizes": [], "device": null, "use_cache": null, @@ -71,5 +71,5 @@ "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 7a958aedaecc85adb014e816d7e2f5dbdcafa539..378ea5820b8dd254c9fac2776a2815060d7b6f88 100644 --- a/lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/logieval/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:60809dd6c89aa81360eb7c365e82170ce124c93c29c723a7a90fa1c6050baef6 -size 22072 +oid sha256:4a1f89f1bd20cb0d9581ebca08b3d7e47091c6fde766a2ed79edbe53b8d80f01 +size 22052 diff --git a/lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 6a4f6ff466475063ac5fbdb3ac67d97c7526f5ff..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:99c5a132be38d6a705684bf04e99cda370f5ca7b58a75f12ee8cb943168f2b02 -size 309798 diff --git a/lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index b6d0bae30301f66f9b7efd0d860e42f384604644..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,64 +0,0 @@ -{ - "results": { - "logiqa": { - "acc,none": 0.23809523809523808, - "acc_stderr,none": 0.01670586703441963, - "acc_norm,none": 0.27035330261136714, - "acc_norm_stderr,none": 0.01742069478339314, - "alias": "logiqa" - } - }, - "configs": { - "logiqa": { - "task": "logiqa", - "dataset_path": "EleutherAI/logiqa", - "dataset_name": "logiqa", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc) -> str:\n \"\"\"\n Passage: \n Question: \n Choices:\n A. \n B. \n C. \n D. \n Answer:\n \"\"\"\n choices = [\"a\", \"b\", \"c\", \"d\"]\n prompt = \"Passage: \" + doc[\"context\"] + \"\\n\"\n prompt += \"Question: \" + doc[\"question\"] + \"\\nChoices:\\n\"\n for choice, option in zip(choices, doc[\"options\"]):\n prompt += f\"{choice.upper()}. {option}\\n\"\n prompt += \"Answer:\"\n return prompt\n", - "doc_to_target": "def doc_to_target(doc) -> int:\n choices = [\"a\", \"b\", \"c\", \"d\"]\n return choices.index(doc[\"label\"].strip())\n", - "doc_to_choice": "{{options}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{context}}", - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "logiqa": 1.0 - }, - "n-shot": { - "logiqa": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index a7d81d72628802f6b9f59ba0fe8602398afca942..0c9e0ae0bfd20e5eba986cdbc4a8a3b24c2bab52 100644 --- a/lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/logiqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:1afed80022da9a0b567771257e96972eb21062096c7b12753668c252b3c00e3e -size 18079 +oid sha256:b6d813aeb02f78d7c002a843f01b29fa2d93e26155de266673b77dff9b133248 +size 22189 diff --git a/lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 1f87ca374d5ce633624c3cea91e82c9bc8944388..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4377ad7e50e0611971d6b447cc968258e7179641fff47956021e5479906f3a3f -size 821458 diff --git a/lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 278419043b5da813270e871e04eef63d55391d16..98dd0c9a5daa644576611e7344dadb4b8339c9ac 100644 --- a/lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,10 +1,10 @@ { "results": { "logiqa2": { - "acc,none": 0.2544529262086514, - "acc_stderr,none": 0.010988873330778288, - "acc_norm,none": 0.28944020356234096, - "acc_norm_stderr,none": 0.011441728828144168, + "acc,none": 0.24872773536895673, + "acc_stderr,none": 0.010906180806103564, + "acc_norm,none": 0.2907124681933842, + "acc_norm_stderr,none": 0.011456577557813215, "alias": "logiqa2" } }, @@ -52,13 +52,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 8 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 3f9ff15a71eaff2044b1f6de735f70e7c5496d4f..b63c763cef91a332c0858195ae9b7ece63e3b76d 100644 --- a/lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/logiqa2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:594992cd87b06cbe943f40713b6bbc234043083bb6eeb9f5857591ba4dbc7039 -size 23627 +oid sha256:61f0b5913ee7c74e6d4b779cb1e963f0acccdebf06ba0781ce8c8d25afebd7b4 +size 23189 diff --git a/lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 17760aabaec327cc69bbace01de2a4a907a33f60..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:3b49b181404604bcfeba2ab439bfc2db362d1ae12234d1071e727a516c5542b9 -size 911808 diff --git a/lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 76964a2e0eda36f67b7ea4ab7c88a100dfafe102..b370e1a87ae656c12a8a50c0ab3ff3dab07525b0 100644 --- a/lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,10 +1,10 @@ { "results": { "mathqa": { - "acc,none": 0.2552763819095477, - "acc_stderr,none": 0.00798184834896828, - "acc_norm,none": 0.2525963149078727, - "acc_norm_stderr,none": 0.007954112207299585, + "acc,none": 0.25896147403685094, + "acc_stderr,none": 0.008019338828219905, + "acc_norm,none": 0.254606365159129, + "acc_norm_stderr,none": 0.007974951653806829, "alias": "mathqa" } }, @@ -54,13 +54,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 32 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 7147f36a21808a344156ca67d67c84ea1c9c7d79..a3fd992c22a3360414b4b537e5282d89a464e12b 100644 --- a/lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/mathqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:45ff2fe8129191ef6a2838115ebb709d0340f8958baaaae6fb265405d1b44daf -size 22059 +oid sha256:6da0aab79c838012114e306652dbf9afe0d2d5fc797600d1298d4ffb21a02fa7 +size 14214 diff --git a/lm-eval-output/allenai/OLMo-7B/mc_taco/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/mc_taco/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index e2e74563db5a47916202357436d432d68721d51b..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/mc_taco/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:966c53240bcec317de5d17f50f9fa86ea6bb67bf480f2daac489f2b4d714123c -size 782140 diff --git a/lm-eval-output/allenai/OLMo-7B/mc_taco/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/mc_taco/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 9b33fc2b3c45e2d7f40a7f35590bcbaa2677d80d..0d0a5acf33f4570833aa810739d2663383f907c4 100644 --- a/lm-eval-output/allenai/OLMo-7B/mc_taco/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/mc_taco/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,10 +1,10 @@ { "results": { "mc_taco": { - "acc,none": 0.3646473204829485, - "acc_stderr,none": 0.004953759833118804, - "f1,none": 0.5047469660695121, - "f1_stderr,none": 0.0055681597726211125, + "acc,none": 0.3653886888371108, + "acc_stderr,none": 0.0049558990911878825, + "f1,none": 0.5050388237237733, + "f1_stderr,none": 0.005568430015605393, "alias": "mc_taco" } }, @@ -49,13 +49,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/mc_taco/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/mc_taco/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 934dfa97705718bf7af6d821ae3ad433a914fdf0..38fcf087bcbeeac30b9b6adf53fe1591c1e3c3c4 100644 --- a/lm-eval-output/allenai/OLMo-7B/mc_taco/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/mc_taco/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:cbc85721a696c8e3be5ecae81e46e97bdbc6e9e550f51aaac8b1769fb1479561 -size 32360 +oid sha256:e3113985b74b1e82a7d79ebbc152856a4b11ba4dde45c164eac06172cd9141ad +size 21406 diff --git a/lm-eval-output/allenai/OLMo-7B/medmcqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/medmcqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index ab96bcc6d5b813fc655c5b29ee8e3acccb25b602..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/medmcqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b660d14fa04ab6a7d2cea1776b566635c370c46658df7337582d7451dc25148c -size 1411337 diff --git a/lm-eval-output/allenai/OLMo-7B/medmcqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/medmcqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 7b7c08688064538d3a792cd8379c03ae41eb92cb..241c339f8303cc92dd02370f9e1e67f429e69038 100644 --- a/lm-eval-output/allenai/OLMo-7B/medmcqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/medmcqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,10 +1,10 @@ { "results": { "medmcqa": { - "acc,none": 0.25029882859191965, - "acc_stderr,none": 0.006698560574683251, - "acc_norm,none": 0.25029882859191965, - "acc_norm_stderr,none": 0.006698560574683251, + "acc,none": 0.2510160172125269, + "acc_stderr,none": 0.006704941116035671, + "acc_norm,none": 0.2510160172125269, + "acc_norm_stderr,none": 0.006704941116035671, "alias": "medmcqa" } }, @@ -53,13 +53,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 32 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/medmcqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/medmcqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index ea53f7276d5e0416e295cb61bc77c44545c4afaf..8b998406c92b29b85e8dcb9e1f6dfd79f99c4264 100644 --- a/lm-eval-output/allenai/OLMo-7B/medmcqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/medmcqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:4a7768e9b004f3421aefcaf66bf2ba0580e5ab9aed3859247f43546ce0c743d6 -size 23711 +oid sha256:99855d6f98deb7fe002cdc53586711288ca0fd0c9f2fc39a80f8fc1866b3070c +size 14740 diff --git a/lm-eval-output/allenai/OLMo-7B/medqa_4options/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/medqa_4options/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index eb82e436b7f66890e172f319890042e8564ee74c..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/medqa_4options/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:620d40cd788a0210d80bfad297243cf36081568a8e5dc614626b861df19e373c -size 641298 diff --git a/lm-eval-output/allenai/OLMo-7B/medqa_4options/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/medqa_4options/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index bc64767a7368257b43bae11531680dd533f098a1..4c6d2509fb88a0b9825afd4a56f48d2c94e3adba 100644 --- a/lm-eval-output/allenai/OLMo-7B/medqa_4options/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/medqa_4options/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,10 +1,10 @@ { "results": { "medqa_4options": { - "acc,none": 0.24823252160251374, - "acc_stderr,none": 0.012112325990382998, - "acc_norm,none": 0.24823252160251374, - "acc_norm_stderr,none": 0.012112325990382998, + "acc,none": 0.24587588373919875, + "acc_stderr,none": 0.012073573380009434, + "acc_norm,none": 0.24587588373919875, + "acc_norm_stderr,none": 0.012073573380009434, "alias": "medqa_4options" } }, @@ -52,13 +52,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 8 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/medqa_4options/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/medqa_4options/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 2771972de92ad5ececcb5a7175bb051cad43eaaf..9d885099545746ad2036fc64d78f0b30fd5c2a24 100644 --- a/lm-eval-output/allenai/OLMo-7B/medqa_4options/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/medqa_4options/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:139455f19150f9253c8c26d03a9cbc2b5dee40b68b885b40c6a2a64039385aa3 -size 17694 +oid sha256:e26494938f8d64d27b6fe135ad6bb02cc172479f77728832ef3918c2671b73fa +size 18187 diff --git a/lm-eval-output/allenai/OLMo-7B/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index e33d34e0c4e5da16cfa3613dcba9ed129209ff61..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:5e4cb6d855f56f91db0fbc243c12d7f9991758b1ec655655e3ae406f8593a032 -size 3969498 diff --git a/lm-eval-output/allenai/OLMo-7B/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 696df83f14e9da34d5f2de228275b58960671e72..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,2592 +0,0 @@ -{ - "results": { - "mmlu": { - "acc,none": 0.2824383990884489, - "acc_stderr,none": 0.04247561258047098, - "alias": "mmlu" - }, - "mmlu_humanities": { - "alias": " - humanities", - "acc,none": 0.2737513283740701, - "acc_stderr,none": 0.0411011139362443 - }, - "mmlu_formal_logic": { - "alias": " - formal_logic", - "acc,none": 0.38095238095238093, - "acc_stderr,none": 0.043435254289490965 - }, - "mmlu_high_school_european_history": { - "alias": " - high_school_european_history", - "acc,none": 0.23636363636363636, - "acc_stderr,none": 0.033175059300091805 - }, - "mmlu_high_school_us_history": { - "alias": " - high_school_us_history", - "acc,none": 0.24019607843137256, - "acc_stderr,none": 0.02998373305591361 - }, - "mmlu_high_school_world_history": { - "alias": " - high_school_world_history", - "acc,none": 0.23628691983122363, - "acc_stderr,none": 0.02765215314415928 - }, - "mmlu_international_law": { - "alias": " - international_law", - "acc,none": 0.35537190082644626, - "acc_stderr,none": 0.04369236326573981 - }, - "mmlu_jurisprudence": { - "alias": " - jurisprudence", - "acc,none": 0.2037037037037037, - "acc_stderr,none": 0.03893542518824848 - }, - "mmlu_logical_fallacies": { - "alias": " - logical_fallacies", - "acc,none": 0.294478527607362, - "acc_stderr,none": 0.03581165790474082 - }, - "mmlu_moral_disputes": { - "alias": " - moral_disputes", - "acc,none": 0.27167630057803466, - "acc_stderr,none": 0.023948512905468348 - }, - "mmlu_moral_scenarios": { - "alias": " - moral_scenarios", - "acc,none": 0.2424581005586592, - "acc_stderr,none": 0.014333522059217892 - }, - "mmlu_philosophy": { - "alias": " - philosophy", - "acc,none": 0.31511254019292606, - "acc_stderr,none": 0.026385273703464485 - }, - "mmlu_prehistory": { - "alias": " - prehistory", - "acc,none": 0.29012345679012347, - "acc_stderr,none": 0.025251173936495022 - }, - "mmlu_professional_law": { - "alias": " - professional_law", - "acc,none": 0.273142112125163, - "acc_stderr,none": 0.011380150567830403 - }, - "mmlu_world_religions": { - "alias": " - world_religions", - "acc,none": 0.3567251461988304, - "acc_stderr,none": 0.03674013002860954 - }, - "mmlu_other": { - "alias": " - other", - "acc,none": 0.2845188284518829, - "acc_stderr,none": 0.04353064334839231 - }, - "mmlu_business_ethics": { - "alias": " - business_ethics", - "acc,none": 0.27, - "acc_stderr,none": 0.0446196043338474 - }, - "mmlu_clinical_knowledge": { - "alias": " - clinical_knowledge", - "acc,none": 0.25660377358490566, - "acc_stderr,none": 0.026880647889051992 - }, - "mmlu_college_medicine": { - "alias": " - college_medicine", - "acc,none": 0.3236994219653179, - "acc_stderr,none": 0.0356760379963917 - }, - "mmlu_global_facts": { - "alias": " - global_facts", - "acc,none": 0.36, - "acc_stderr,none": 0.048241815132442176 - }, - "mmlu_human_aging": { - "alias": " - human_aging", - "acc,none": 0.2242152466367713, - "acc_stderr,none": 0.027991534258519527 - }, - "mmlu_management": { - "alias": " - management", - "acc,none": 0.2524271844660194, - "acc_stderr,none": 0.04301250399690878 - }, - "mmlu_marketing": { - "alias": " - marketing", - "acc,none": 0.2564102564102564, - "acc_stderr,none": 0.028605953702004253 - }, - "mmlu_medical_genetics": { - "alias": " - medical_genetics", - "acc,none": 0.27, - "acc_stderr,none": 0.044619604333847394 - }, - "mmlu_miscellaneous": { - "alias": " - miscellaneous", - "acc,none": 0.30268199233716475, - "acc_stderr,none": 0.01642878158174936 - }, - "mmlu_nutrition": { - "alias": " - nutrition", - "acc,none": 0.34967320261437906, - "acc_stderr,none": 0.0273053080762747 - }, - "mmlu_professional_accounting": { - "alias": " - professional_accounting", - "acc,none": 0.29432624113475175, - "acc_stderr,none": 0.027187127011503793 - }, - "mmlu_professional_medicine": { - "alias": " - professional_medicine", - "acc,none": 0.23529411764705882, - "acc_stderr,none": 0.025767252010855973 - }, - "mmlu_virology": { - "alias": " - virology", - "acc,none": 0.25903614457831325, - "acc_stderr,none": 0.034106466140718564 - }, - "mmlu_social_sciences": { - "alias": " - social_sciences", - "acc,none": 0.2918427039324017, - "acc_stderr,none": 0.03431852465251 - }, - "mmlu_econometrics": { - "alias": " - econometrics", - "acc,none": 0.2543859649122807, - "acc_stderr,none": 0.040969851398436695 - }, - "mmlu_high_school_geography": { - "alias": " - high_school_geography", - "acc,none": 0.31313131313131315, - "acc_stderr,none": 0.03304205087813652 - }, - "mmlu_high_school_government_and_politics": { - "alias": " - high_school_government_and_politics", - "acc,none": 0.30569948186528495, - "acc_stderr,none": 0.033248379397581594 - }, - "mmlu_high_school_macroeconomics": { - "alias": " - high_school_macroeconomics", - "acc,none": 0.30512820512820515, - "acc_stderr,none": 0.023346335293325887 - }, - "mmlu_high_school_microeconomics": { - "alias": " - high_school_microeconomics", - "acc,none": 0.2815126050420168, - "acc_stderr,none": 0.02921354941437217 - }, - "mmlu_high_school_psychology": { - "alias": " - high_school_psychology", - "acc,none": 0.28807339449541286, - "acc_stderr,none": 0.01941644589263603 - }, - "mmlu_human_sexuality": { - "alias": " - human_sexuality", - "acc,none": 0.31297709923664124, - "acc_stderr,none": 0.04066962905677697 - }, - "mmlu_professional_psychology": { - "alias": " - professional_psychology", - "acc,none": 0.2696078431372549, - "acc_stderr,none": 0.017952449196987862 - }, - "mmlu_public_relations": { - "alias": " - public_relations", - "acc,none": 0.2545454545454545, - "acc_stderr,none": 0.041723430387053825 - }, - "mmlu_security_studies": { - "alias": " - security_studies", - "acc,none": 0.3020408163265306, - "acc_stderr,none": 0.029393609319879815 - }, - "mmlu_sociology": { - "alias": " - sociology", - "acc,none": 0.2935323383084577, - "acc_stderr,none": 0.03220024104534204 - }, - "mmlu_us_foreign_policy": { - "alias": " - us_foreign_policy", - "acc,none": 0.38, - "acc_stderr,none": 0.04878317312145632 - }, - "mmlu_stem": { - "alias": " - stem", - "acc,none": 0.284173802727561, - "acc_stderr,none": 0.049209572972660855 - }, - "mmlu_abstract_algebra": { - "alias": " - abstract_algebra", - "acc,none": 0.27, - "acc_stderr,none": 0.04461960433384741 - }, - "mmlu_anatomy": { - "alias": " - anatomy", - "acc,none": 0.26666666666666666, - "acc_stderr,none": 0.038201699145179055 - }, - "mmlu_astronomy": { - "alias": " - astronomy", - "acc,none": 0.34210526315789475, - "acc_stderr,none": 0.03860731599316092 - }, - "mmlu_college_biology": { - "alias": " - college_biology", - "acc,none": 0.2916666666666667, - "acc_stderr,none": 0.038009680605548574 - }, - "mmlu_college_chemistry": { - "alias": " - college_chemistry", - "acc,none": 0.22, - "acc_stderr,none": 0.0416333199893227 - }, - "mmlu_college_computer_science": { - "alias": " - college_computer_science", - "acc,none": 0.38, - "acc_stderr,none": 0.048783173121456344 - }, - "mmlu_college_mathematics": { - "alias": " - college_mathematics", - "acc,none": 0.4, - "acc_stderr,none": 0.04923659639173309 - }, - "mmlu_college_physics": { - "alias": " - college_physics", - "acc,none": 0.23529411764705882, - "acc_stderr,none": 0.04220773659171453 - }, - "mmlu_computer_security": { - "alias": " - computer_security", - "acc,none": 0.36, - "acc_stderr,none": 0.04824181513244218 - }, - "mmlu_conceptual_physics": { - "alias": " - conceptual_physics", - "acc,none": 0.24680851063829787, - "acc_stderr,none": 0.028185441301234106 - }, - "mmlu_electrical_engineering": { - "alias": " - electrical_engineering", - "acc,none": 0.296551724137931, - "acc_stderr,none": 0.03806142687309993 - }, - "mmlu_elementary_mathematics": { - "alias": " - elementary_mathematics", - "acc,none": 0.2698412698412698, - "acc_stderr,none": 0.02286083830923207 - }, - "mmlu_high_school_biology": { - "alias": " - high_school_biology", - "acc,none": 0.25161290322580643, - "acc_stderr,none": 0.024685979286239952 - }, - "mmlu_high_school_chemistry": { - "alias": " - high_school_chemistry", - "acc,none": 0.29064039408866993, - "acc_stderr,none": 0.031947400722655395 - }, - "mmlu_high_school_computer_science": { - "alias": " - high_school_computer_science", - "acc,none": 0.27, - "acc_stderr,none": 0.04461960433384741 - }, - "mmlu_high_school_mathematics": { - "alias": " - high_school_mathematics", - "acc,none": 0.2740740740740741, - "acc_stderr,none": 0.027195934804085626 - }, - "mmlu_high_school_physics": { - "alias": " - high_school_physics", - "acc,none": 0.2847682119205298, - "acc_stderr,none": 0.03684881521389023 - }, - "mmlu_high_school_statistics": { - "alias": " - high_school_statistics", - "acc,none": 0.2824074074074074, - "acc_stderr,none": 0.030701372111510923 - }, - "mmlu_machine_learning": { - "alias": " - machine_learning", - "acc,none": 0.30357142857142855, - "acc_stderr,none": 0.04364226155841044 - } - }, - "groups": { - "mmlu": { - "acc,none": 0.2824383990884489, - "acc_stderr,none": 0.04247561258047098, - "alias": "mmlu" - }, - "mmlu_humanities": { - "alias": " - humanities", - "acc,none": 0.2737513283740701, - "acc_stderr,none": 0.0411011139362443 - }, - "mmlu_other": { - "alias": " - other", - "acc,none": 0.2845188284518829, - "acc_stderr,none": 0.04353064334839231 - }, - "mmlu_social_sciences": { - "alias": " - social_sciences", - "acc,none": 0.2918427039324017, - "acc_stderr,none": 0.03431852465251 - }, - "mmlu_stem": { - "alias": " - stem", - "acc,none": 0.284173802727561, - "acc_stderr,none": 0.049209572972660855 - } - }, - "configs": { - "mmlu_abstract_algebra": { - "task": "mmlu_abstract_algebra", - "task_alias": "abstract_algebra", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "abstract_algebra", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about abstract algebra.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_anatomy": { - "task": "mmlu_anatomy", - "task_alias": "anatomy", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "anatomy", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about anatomy.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_astronomy": { - "task": "mmlu_astronomy", - "task_alias": "astronomy", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "astronomy", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about astronomy.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_business_ethics": { - "task": "mmlu_business_ethics", - "task_alias": "business_ethics", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "business_ethics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about business ethics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_clinical_knowledge": { - "task": "mmlu_clinical_knowledge", - "task_alias": "clinical_knowledge", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "clinical_knowledge", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about clinical knowledge.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_biology": { - "task": "mmlu_college_biology", - "task_alias": "college_biology", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_biology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college biology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_chemistry": { - "task": "mmlu_college_chemistry", - "task_alias": "college_chemistry", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_chemistry", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college chemistry.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_computer_science": { - "task": "mmlu_college_computer_science", - "task_alias": "college_computer_science", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_computer_science", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college computer science.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_mathematics": { - "task": "mmlu_college_mathematics", - "task_alias": "college_mathematics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_mathematics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college mathematics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_medicine": { - "task": "mmlu_college_medicine", - "task_alias": "college_medicine", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_medicine", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college medicine.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_physics": { - "task": "mmlu_college_physics", - "task_alias": "college_physics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_physics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college physics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_computer_security": { - "task": "mmlu_computer_security", - "task_alias": "computer_security", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "computer_security", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about computer security.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_conceptual_physics": { - "task": "mmlu_conceptual_physics", - "task_alias": "conceptual_physics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "conceptual_physics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about conceptual physics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_econometrics": { - "task": "mmlu_econometrics", - "task_alias": "econometrics", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "econometrics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about econometrics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_electrical_engineering": { - "task": "mmlu_electrical_engineering", - "task_alias": "electrical_engineering", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "electrical_engineering", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about electrical engineering.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_elementary_mathematics": { - "task": "mmlu_elementary_mathematics", - "task_alias": "elementary_mathematics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "elementary_mathematics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about elementary mathematics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_formal_logic": { - "task": "mmlu_formal_logic", - "task_alias": "formal_logic", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "formal_logic", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about formal logic.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_global_facts": { - "task": "mmlu_global_facts", - "task_alias": "global_facts", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "global_facts", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about global facts.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_biology": { - "task": "mmlu_high_school_biology", - "task_alias": "high_school_biology", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_biology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school biology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_chemistry": { - "task": "mmlu_high_school_chemistry", - "task_alias": "high_school_chemistry", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_chemistry", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school chemistry.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_computer_science": { - "task": "mmlu_high_school_computer_science", - "task_alias": "high_school_computer_science", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_computer_science", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school computer science.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_european_history": { - "task": "mmlu_high_school_european_history", - "task_alias": "high_school_european_history", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_european_history", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school european history.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_geography": { - "task": "mmlu_high_school_geography", - "task_alias": "high_school_geography", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_geography", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school geography.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_government_and_politics": { - "task": "mmlu_high_school_government_and_politics", - "task_alias": "high_school_government_and_politics", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_government_and_politics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school government and politics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_macroeconomics": { - "task": "mmlu_high_school_macroeconomics", - "task_alias": "high_school_macroeconomics", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_macroeconomics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school macroeconomics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_mathematics": { - "task": "mmlu_high_school_mathematics", - "task_alias": "high_school_mathematics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_mathematics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school mathematics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_microeconomics": { - "task": "mmlu_high_school_microeconomics", - "task_alias": "high_school_microeconomics", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_microeconomics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school microeconomics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_physics": { - "task": "mmlu_high_school_physics", - "task_alias": "high_school_physics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_physics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school physics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_psychology": { - "task": "mmlu_high_school_psychology", - "task_alias": "high_school_psychology", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_psychology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school psychology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_statistics": { - "task": "mmlu_high_school_statistics", - "task_alias": "high_school_statistics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_statistics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school statistics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_us_history": { - "task": "mmlu_high_school_us_history", - "task_alias": "high_school_us_history", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_us_history", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school us history.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_world_history": { - "task": "mmlu_high_school_world_history", - "task_alias": "high_school_world_history", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_world_history", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school world history.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_human_aging": { - "task": "mmlu_human_aging", - "task_alias": "human_aging", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "human_aging", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about human aging.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_human_sexuality": { - "task": "mmlu_human_sexuality", - "task_alias": "human_sexuality", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "human_sexuality", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about human sexuality.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_international_law": { - "task": "mmlu_international_law", - "task_alias": "international_law", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "international_law", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about international law.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_jurisprudence": { - "task": "mmlu_jurisprudence", - "task_alias": "jurisprudence", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "jurisprudence", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about jurisprudence.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_logical_fallacies": { - "task": "mmlu_logical_fallacies", - "task_alias": "logical_fallacies", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "logical_fallacies", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about logical fallacies.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_machine_learning": { - "task": "mmlu_machine_learning", - "task_alias": "machine_learning", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "machine_learning", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about machine learning.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_management": { - "task": "mmlu_management", - "task_alias": "management", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "management", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about management.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_marketing": { - "task": "mmlu_marketing", - "task_alias": "marketing", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "marketing", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about marketing.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_medical_genetics": { - "task": "mmlu_medical_genetics", - "task_alias": "medical_genetics", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "medical_genetics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about medical genetics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_miscellaneous": { - "task": "mmlu_miscellaneous", - "task_alias": "miscellaneous", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "miscellaneous", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about miscellaneous.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_moral_disputes": { - "task": "mmlu_moral_disputes", - "task_alias": "moral_disputes", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "moral_disputes", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about moral disputes.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_moral_scenarios": { - "task": "mmlu_moral_scenarios", - "task_alias": "moral_scenarios", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "moral_scenarios", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about moral scenarios.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_nutrition": { - "task": "mmlu_nutrition", - "task_alias": "nutrition", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "nutrition", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about nutrition.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_philosophy": { - "task": "mmlu_philosophy", - "task_alias": "philosophy", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "philosophy", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about philosophy.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_prehistory": { - "task": "mmlu_prehistory", - "task_alias": "prehistory", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "prehistory", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about prehistory.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_professional_accounting": { - "task": "mmlu_professional_accounting", - "task_alias": "professional_accounting", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "professional_accounting", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about professional accounting.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_professional_law": { - "task": "mmlu_professional_law", - "task_alias": "professional_law", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "professional_law", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about professional law.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_professional_medicine": { - "task": "mmlu_professional_medicine", - "task_alias": "professional_medicine", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "professional_medicine", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about professional medicine.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_professional_psychology": { - "task": "mmlu_professional_psychology", - "task_alias": "professional_psychology", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "professional_psychology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about professional psychology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_public_relations": { - "task": "mmlu_public_relations", - "task_alias": "public_relations", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "public_relations", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about public relations.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_security_studies": { - "task": "mmlu_security_studies", - "task_alias": "security_studies", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "security_studies", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about security studies.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_sociology": { - "task": "mmlu_sociology", - "task_alias": "sociology", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "sociology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about sociology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_us_foreign_policy": { - "task": "mmlu_us_foreign_policy", - "task_alias": "us_foreign_policy", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "us_foreign_policy", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about us foreign policy.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_virology": { - "task": "mmlu_virology", - "task_alias": "virology", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "virology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about virology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_world_religions": { - "task": "mmlu_world_religions", - "task_alias": "world_religions", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "world_religions", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about world religions.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - } - }, - "versions": { - "mmlu": "N/A", - "mmlu_abstract_algebra": 0.0, - "mmlu_anatomy": 0.0, - "mmlu_astronomy": 0.0, - "mmlu_business_ethics": 0.0, - "mmlu_clinical_knowledge": 0.0, - "mmlu_college_biology": 0.0, - "mmlu_college_chemistry": 0.0, - "mmlu_college_computer_science": 0.0, - "mmlu_college_mathematics": 0.0, - "mmlu_college_medicine": 0.0, - "mmlu_college_physics": 0.0, - "mmlu_computer_security": 0.0, - "mmlu_conceptual_physics": 0.0, - "mmlu_econometrics": 0.0, - "mmlu_electrical_engineering": 0.0, - "mmlu_elementary_mathematics": 0.0, - "mmlu_formal_logic": 0.0, - "mmlu_global_facts": 0.0, - "mmlu_high_school_biology": 0.0, - "mmlu_high_school_chemistry": 0.0, - "mmlu_high_school_computer_science": 0.0, - "mmlu_high_school_european_history": 0.0, - "mmlu_high_school_geography": 0.0, - "mmlu_high_school_government_and_politics": 0.0, - "mmlu_high_school_macroeconomics": 0.0, - "mmlu_high_school_mathematics": 0.0, - "mmlu_high_school_microeconomics": 0.0, - "mmlu_high_school_physics": 0.0, - "mmlu_high_school_psychology": 0.0, - "mmlu_high_school_statistics": 0.0, - "mmlu_high_school_us_history": 0.0, - "mmlu_high_school_world_history": 0.0, - "mmlu_human_aging": 0.0, - "mmlu_human_sexuality": 0.0, - "mmlu_humanities": "N/A", - "mmlu_international_law": 0.0, - "mmlu_jurisprudence": 0.0, - "mmlu_logical_fallacies": 0.0, - "mmlu_machine_learning": 0.0, - "mmlu_management": 0.0, - "mmlu_marketing": 0.0, - "mmlu_medical_genetics": 0.0, - "mmlu_miscellaneous": 0.0, - "mmlu_moral_disputes": 0.0, - "mmlu_moral_scenarios": 0.0, - "mmlu_nutrition": 0.0, - "mmlu_other": "N/A", - "mmlu_philosophy": 0.0, - "mmlu_prehistory": 0.0, - "mmlu_professional_accounting": 0.0, - "mmlu_professional_law": 0.0, - "mmlu_professional_medicine": 0.0, - "mmlu_professional_psychology": 0.0, - "mmlu_public_relations": 0.0, - "mmlu_security_studies": 0.0, - "mmlu_social_sciences": "N/A", - "mmlu_sociology": 0.0, - "mmlu_stem": "N/A", - "mmlu_us_foreign_policy": 0.0, - "mmlu_virology": 0.0, - "mmlu_world_religions": 0.0 - }, - "n-shot": { - "mmlu": 0, - "mmlu_abstract_algebra": 0, - "mmlu_anatomy": 0, - "mmlu_astronomy": 0, - "mmlu_business_ethics": 0, - "mmlu_clinical_knowledge": 0, - "mmlu_college_biology": 0, - "mmlu_college_chemistry": 0, - "mmlu_college_computer_science": 0, - "mmlu_college_mathematics": 0, - "mmlu_college_medicine": 0, - "mmlu_college_physics": 0, - "mmlu_computer_security": 0, - "mmlu_conceptual_physics": 0, - "mmlu_econometrics": 0, - "mmlu_electrical_engineering": 0, - "mmlu_elementary_mathematics": 0, - "mmlu_formal_logic": 0, - "mmlu_global_facts": 0, - "mmlu_high_school_biology": 0, - "mmlu_high_school_chemistry": 0, - "mmlu_high_school_computer_science": 0, - "mmlu_high_school_european_history": 0, - "mmlu_high_school_geography": 0, - "mmlu_high_school_government_and_politics": 0, - "mmlu_high_school_macroeconomics": 0, - "mmlu_high_school_mathematics": 0, - "mmlu_high_school_microeconomics": 0, - "mmlu_high_school_physics": 0, - "mmlu_high_school_psychology": 0, - "mmlu_high_school_statistics": 0, - "mmlu_high_school_us_history": 0, - "mmlu_high_school_world_history": 0, - "mmlu_human_aging": 0, - "mmlu_human_sexuality": 0, - "mmlu_humanities": 0, - "mmlu_international_law": 0, - "mmlu_jurisprudence": 0, - "mmlu_logical_fallacies": 0, - "mmlu_machine_learning": 0, - "mmlu_management": 0, - "mmlu_marketing": 0, - "mmlu_medical_genetics": 0, - "mmlu_miscellaneous": 0, - "mmlu_moral_disputes": 0, - "mmlu_moral_scenarios": 0, - "mmlu_nutrition": 0, - "mmlu_other": 0, - "mmlu_philosophy": 0, - "mmlu_prehistory": 0, - "mmlu_professional_accounting": 0, - "mmlu_professional_law": 0, - "mmlu_professional_medicine": 0, - "mmlu_professional_psychology": 0, - "mmlu_public_relations": 0, - "mmlu_security_studies": 0, - "mmlu_social_sciences": 0, - "mmlu_sociology": 0, - "mmlu_stem": 0, - "mmlu_us_foreign_policy": 0, - "mmlu_virology": 0, - "mmlu_world_religions": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index a96161f8d9b31626cf39c93453a651bb6d3c2076..7a12c9f490e916b3d94540907d3f9a0118ecb01e 100644 --- a/lm-eval-output/allenai/OLMo-7B/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:a2e9d3f98874dd8e2947f52bd9d3dced8306d9e1104048bb1e7a33420d8d1cf5 -size 121026 +oid sha256:2d9d7525a07f80582d69a832fc6d03cc551981a5843a40863cdb5dcaa194a87f +size 70800 diff --git a/lm-eval-output/allenai/OLMo-7B/mnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/mnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 6e6f7d2d50c76b80b53b5e5485e268a592db69fc..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/mnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:009ba12413dacb20b6b75c2fe9f3eaa9e8e8236651cc59a5628311d282edadd0 -size 1454792 diff --git a/lm-eval-output/allenai/OLMo-7B/mnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/mnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 37674c04379a0c4d32abd82fecca4bee0e9026ac..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/mnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,58 +0,0 @@ -{ - "results": { - "mnli": { - "acc,none": 0.32776362710137547, - "acc_stderr,none": 0.004738254189734395, - "alias": "mnli" - } - }, - "configs": { - "mnli": { - "task": "mnli", - "group": "glue", - "dataset_path": "glue", - "dataset_name": "mnli", - "training_split": "train", - "validation_split": "validation_matched", - "doc_to_text": "def doc_to_text(doc) -> str:\n return \"{}\\nQuestion: {} True, False or Neither?\\nAnswer:\".format(\n doc[\"premise\"],\n doc[\"hypothesis\"].strip()\n + (\"\" if doc[\"hypothesis\"].strip().endswith(\".\") else \".\"),\n )\n", - "doc_to_target": "label", - "doc_to_choice": [ - "True", - "Neither", - "False" - ], - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "mnli": 1.0 - }, - "n-shot": { - "mnli": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/mnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/mnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 032da2bf12aa7d50fb40a1113428bdd9bf5b787f..5f4ee9d3436fb8c240727a5c3d2fe3a8b0a507dd 100644 --- a/lm-eval-output/allenai/OLMo-7B/mnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/mnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:9912ca59c72c0bd2f9ac0bb36ce6d64cea01e462070f023d82b6b4121db9b26e -size 36005 +oid sha256:969e96484bbd39e9dc0db4c37355f8da829c76726800b1413f825f55c51bb41d +size 30433 diff --git a/lm-eval-output/allenai/OLMo-7B/mnli_mismatch/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/mnli_mismatch/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 63a69213bfe5fc60721e1a002acf6018faeb8bd5..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/mnli_mismatch/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:f191a428a77f4faffa984dec05b140eb1eb8a4ba7392622e9e2967aab5370aad -size 1496894 diff --git a/lm-eval-output/allenai/OLMo-7B/mnli_mismatch/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/mnli_mismatch/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index db03b47eae6d90eeff882ad513a033edc6c4cb23..21919e62dc7c80828e0781bb94d7d05e3bc9efa2 100644 --- a/lm-eval-output/allenai/OLMo-7B/mnli_mismatch/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/mnli_mismatch/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "mnli_mismatch": { - "acc,none": 0.33411310008136696, - "acc_stderr,none": 0.004757164173003751, + "acc,none": 0.3358421480878763, + "acc_stderr,none": 0.0047632613082498445, "alias": "mnli_mismatch" } }, @@ -46,13 +46,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 32 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/mnli_mismatch/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/mnli_mismatch/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 75751dc93fbf41c609d35405e47c918af2d9a1d3..0ad7ba12c90a5585152ac5096af10690d911c427 100644 --- a/lm-eval-output/allenai/OLMo-7B/mnli_mismatch/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/mnli_mismatch/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:09cec62b387f62a69b77a8fb85de62437121e0db27c7717d1cb654fd7388392b -size 36586 +oid sha256:6d39016236d9ded1d71013a50c1b97672011803394254bfb14839d41fa3e774a +size 20364 diff --git a/lm-eval-output/allenai/OLMo-7B/mrpc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/mrpc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 81adbdc2b8de6847777b354965d9852c06cfbbc9..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/mrpc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:1e2063ed768f53758acf0741bec53320c7b4d310444e3dc043cce3d358e345ee -size 59224 diff --git a/lm-eval-output/allenai/OLMo-7B/mrpc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/mrpc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index b64f0e3df5a988f2f1659083ea4094e46c880c40..97512516f42c4d4fafdf5bcbf1e74139ff538e3f 100644 --- a/lm-eval-output/allenai/OLMo-7B/mrpc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/mrpc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -50,13 +50,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/mrpc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/mrpc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 9ba35cb43ffaca60eb18afb8cf0fb174f6fc9fe3..04a64f54b710e4b4d04f886b56a38ad28deda547 100644 --- a/lm-eval-output/allenai/OLMo-7B/mrpc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/mrpc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:fc17c8261a8ed63099836b62488996caaabcf752b6c4e12c703811b24c90775f -size 16755 +oid sha256:9774a63480eefa9985a3366a0a7a91ed01a9276a0dc6180bb23cf375b89656f3 +size 16849 diff --git a/lm-eval-output/allenai/OLMo-7B/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 662527ece3c55afbcfe09c08ed4b52df57497da0..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:b4257ad1f70ed55089b6c29baa28d0d55ff5edaab84a070e309da3e98687a9a6 -size 2798765 diff --git a/lm-eval-output/allenai/OLMo-7B/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 12e025f2ecfce6bdf43ea4185b746a7ad136ec29..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,427 +0,0 @@ -{ - "results": { - "multimedqa": { - "alias": "stem", - "acc,none": 0.28431511710432933, - "acc_stderr,none": 0.10619682013189911, - "acc_norm,none": 0.2498366196439298, - "acc_norm_stderr,none": 8.176703928892357e-05 - }, - "medmcqa": { - "acc,none": 0.25029882859191965, - "acc_stderr,none": 0.006698560574683251, - "acc_norm,none": 0.25029882859191965, - "acc_norm_stderr,none": 0.006698560574683251, - "alias": " - medmcqa" - }, - "medqa_4options": { - "acc,none": 0.24901806755695208, - "acc_stderr,none": 0.012125135984037819, - "acc_norm,none": 0.24901806755695208, - "acc_norm_stderr,none": 0.012125135984037819, - "alias": " - medqa_4options" - }, - "mmlu_anatomy": { - "alias": " - anatomy (mmlu)", - "acc,none": 0.26666666666666666, - "acc_stderr,none": 0.038201699145179055 - }, - "mmlu_clinical_knowledge": { - "alias": " - clinical_knowledge (mmlu)", - "acc,none": 0.25660377358490566, - "acc_stderr,none": 0.026880647889051992 - }, - "mmlu_college_biology": { - "alias": " - college_biology (mmlu)", - "acc,none": 0.2916666666666667, - "acc_stderr,none": 0.038009680605548574 - }, - "mmlu_college_medicine": { - "alias": " - college_medicine (mmlu)", - "acc,none": 0.3236994219653179, - "acc_stderr,none": 0.0356760379963917 - }, - "mmlu_medical_genetics": { - "alias": " - medical_genetics (mmlu)", - "acc,none": 0.27, - "acc_stderr,none": 0.044619604333847394 - }, - "mmlu_professional_medicine": { - "alias": " - professional_medicine (mmlu)", - "acc,none": 0.23529411764705882, - "acc_stderr,none": 0.025767252010855973 - }, - "pubmedqa": { - "acc,none": 0.692, - "acc_stderr,none": 0.0206670329874661, - "alias": " - pubmedqa" - } - }, - "groups": { - "multimedqa": { - "alias": "stem", - "acc,none": 0.28431511710432933, - "acc_stderr,none": 0.10619682013189911, - "acc_norm,none": 0.2498366196439298, - "acc_norm_stderr,none": 8.176703928892357e-05 - } - }, - "configs": { - "medmcqa": { - "task": "medmcqa", - "dataset_path": "medmcqa", - "training_split": "train", - "validation_split": "validation", - "test_split": "validation", - "doc_to_text": "def doc_to_text(doc) -> str:\n \"\"\"\n Question: \n Choices:\n A. \n B. \n C. \n D. \n Answer:\n \"\"\"\n choices = [doc[\"opa\"], doc[\"opb\"], doc[\"opc\"], doc[\"opd\"]]\n option_choices = {'A': choices[0], 'B': choices[1], 'C': choices[2], 'D': choices[3]}\n\n prompt = \"Question: \" + doc[\"question\"] + \"\\nChoices:\\n\"\n for choice, option in option_choices.items():\n prompt += f\"{choice.upper()}. {option}\\n\"\n prompt += \"Answer:\"\n return prompt\n", - "doc_to_target": "cop", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{question}}" - }, - "medqa_4options": { - "task": "medqa_4options", - "dataset_path": "GBaker/MedQA-USMLE-4-options-hf", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc) -> str:\n option_choices = {'A': doc[\"ending0\"], 'B': doc[\"ending1\"], 'C': doc[\"ending2\"], 'D': doc[\"ending3\"]}\n answers = \"\".join((f\"{k}. {v}\\n\") for k, v in option_choices.items())\n return f\"Question: {doc['sent1']}\\n{answers}Answer:\"\n", - "doc_to_target": "def doc_to_target(doc) -> int:\n return doc[\"label\"]\n", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false - }, - "mmlu_anatomy": { - "task": "mmlu_anatomy", - "task_alias": "anatomy (mmlu)", - "group": "multimedqa", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "anatomy", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about anatomy.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_clinical_knowledge": { - "task": "mmlu_clinical_knowledge", - "task_alias": "clinical_knowledge (mmlu)", - "group": "multimedqa", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "clinical_knowledge", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about clinical knowledge.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_biology": { - "task": "mmlu_college_biology", - "task_alias": "college_biology (mmlu)", - "group": "multimedqa", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_biology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college biology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_medicine": { - "task": "mmlu_college_medicine", - "task_alias": "college_medicine (mmlu)", - "group": "multimedqa", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_medicine", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college medicine.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_medical_genetics": { - "task": "mmlu_medical_genetics", - "task_alias": "medical_genetics (mmlu)", - "group": "multimedqa", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "medical_genetics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about medical genetics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_professional_medicine": { - "task": "mmlu_professional_medicine", - "task_alias": "professional_medicine (mmlu)", - "group": "multimedqa", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "professional_medicine", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about professional medicine.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "pubmedqa": { - "task": "pubmedqa", - "dataset_path": "bigbio/pubmed_qa", - "dataset_name": "pubmed_qa_labeled_fold0_source", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc) -> str:\n ctxs = \"\\n\".join(doc[\"CONTEXTS\"])\n return \"Abstract: {}\\nQuestion: {}\\nAnswer:\".format(\n ctxs,\n doc[\"QUESTION\"],\n )\n", - "doc_to_target": "final_decision", - "doc_to_choice": [ - "yes", - "no", - "maybe" - ], - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "medmcqa": "Yaml", - "medqa_4options": "Yaml", - "mmlu_anatomy": 0.0, - "mmlu_clinical_knowledge": 0.0, - "mmlu_college_biology": 0.0, - "mmlu_college_medicine": 0.0, - "mmlu_medical_genetics": 0.0, - "mmlu_professional_medicine": 0.0, - "multimedqa": "N/A", - "pubmedqa": 1.0 - }, - "n-shot": { - "medmcqa": 0, - "medqa_4options": 0, - "mmlu_anatomy": 0, - "mmlu_clinical_knowledge": 0, - "mmlu_college_biology": 0, - "mmlu_college_medicine": 0, - "mmlu_medical_genetics": 0, - "mmlu_professional_medicine": 0, - "multimedqa": 0, - "pubmedqa": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 233483a49470c412f4fb46f8086fcaca604c6399..70b9a0789e4daf47e5859a1cbdf65ce351aaa5af 100644 --- a/lm-eval-output/allenai/OLMo-7B/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:3ea480bf6be6450ac138f1be9c1a15bdc6cee0a2c87dccdaac4d303380ecb824 -size 49758 +oid sha256:bd89e790f3b8de7ff66505c6f0c72a806c9c8733d0a10bb7b6bd811ae52634c6 +size 37803 diff --git a/lm-eval-output/allenai/OLMo-7B/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 38d96846f9ecb5f30163406181c3ef934359009f..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e33b2d943ff4e2a3260e3697131a46841f610d7ca8326961a05bc4fa07e26fb4 -size 1066382 diff --git a/lm-eval-output/allenai/OLMo-7B/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 397b2572aaef775057cc5449b1776de4b4fae22d..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,56 +0,0 @@ -{ - "results": { - "multirc": { - "acc,none": 0.5693069306930693, - "acc_stderr,none": 0.007112473596419734, - "alias": "multirc" - } - }, - "configs": { - "multirc": { - "task": "multirc", - "group": [ - "super-glue-lm-eval-v1" - ], - "dataset_path": "super_glue", - "dataset_name": "multirc", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "{{paragraph}}\nQuestion: {{question}}\nAnswer:", - "doc_to_target": "label", - "doc_to_choice": "['''{{answer}}\\nIs the answer correct? yes''', '''{{answer}}\\nIs the answer correct? no''']", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 2.0 - } - } - }, - "versions": { - "multirc": 2.0 - }, - "n-shot": { - "multirc": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 1728e1d39d0059e7dbf0e459bdd002c819bda39a..ae1d210d68467f52abd1f073e495b4d07207fae8 100644 --- a/lm-eval-output/allenai/OLMo-7B/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:9949fed3026951419aa1c767a76a1b3440f472d6558f8fef542924cfb59f0683 -size 31523 +oid sha256:8657ae46ddd0cf25a00eb531e15232b2a78cf94314916c2ed538cd70cc6e1dc1 +size 53277 diff --git a/lm-eval-output/allenai/OLMo-7B/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index f20a9b04e42b2aa43383693e5ce4ca522d18ad96..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2b0f05aee63c07c42a77b57d93e5e280e4bbab81ba35bb7333aa69d2d975682f -size 310428 diff --git a/lm-eval-output/allenai/OLMo-7B/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index a2d7b484eeebf4d0238f79e874d5838c611fcd43..1e35d2cce26b2852e77a2c7fc9131f5b52ff30dd 100644 --- a/lm-eval-output/allenai/OLMo-7B/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -3,10 +3,10 @@ "mutual": { "r@1,none": 0.22573363431151242, "r@1_stderr,none": 0.014053085820407473, - "r@2,none": 0.417607223476298, - "r@2_stderr,none": 0.016577550348797442, - "mrr,none": 0.6996802124226336, - "mrr_stderr,none": 0.010329677864245292, + "r@2,none": 0.41309255079006774, + "r@2_stderr,none": 0.016551480902963107, + "mrr,none": 0.7007148249020156, + "mrr_stderr,none": 0.010340024636004837, "alias": "mutual" } }, @@ -60,13 +60,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 16 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 5a819bb69d006ef8eeee37dc94ad0e6ab2958849..37bd1001f74641c3230476964a60bdd0db240252 100644 --- a/lm-eval-output/allenai/OLMo-7B/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:e165fcab4fc265ecc8932c5e463d074e442038f04ef0f17b8417347679a1aa97 -size 17747 +oid sha256:c46dfea34247cea56e1f34c748b81a1976df6663e613dbb9b4d580743140cadd +size 16536 diff --git a/lm-eval-output/allenai/OLMo-7B/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index b79de1ba570d62bb2bd15830cdb4ac560fdd338b..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:9c6c078e132a828b875664614d0796c42bc772baa61b8ba351194b60f7ff8758 -size 307646 diff --git a/lm-eval-output/allenai/OLMo-7B/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 10701754ff55a44ccea40d548efea804ce657459..cd9e5c127ed67915e3da967714e7cf3aa6867300 100644 --- a/lm-eval-output/allenai/OLMo-7B/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -3,10 +3,10 @@ "mutual_plus": { "r@1,none": 0.2595936794582393, "r@1_stderr,none": 0.014737047402750952, - "r@2,none": 0.4650112866817156, - "r@2_stderr,none": 0.016766114263692605, - "mrr,none": 0.6511474810456583, - "mrr_stderr,none": 0.01044237669999924, + "r@2,none": 0.4672686230248307, + "r@2_stderr,none": 0.016771264669080587, + "mrr,none": 0.6497366458312114, + "mrr_stderr,none": 0.010438626939583338, "alias": "mutual_plus" } }, @@ -60,13 +60,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 16 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index cf20ae43067cdcae512e59d945f8b21ebe099c69..f20ebf231175a890a444f2aa6a316c2baafe4887 100644 --- a/lm-eval-output/allenai/OLMo-7B/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:cedb1d8a2fabe7eb8fd63e6797555f5d14d8441f98f2b19593d542cc88718719 -size 17823 +oid sha256:ee66366050505cf05624132c996a99e25b2b43492469376dceea9c566ecf9fbe +size 16599 diff --git a/lm-eval-output/allenai/OLMo-7B/openbookqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/openbookqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 0995be40806414dbcaa583405a3c95c600f184bf..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/openbookqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:f08cdf3938b7a8de51ea4b9d1f7bac3e766c77273ba5ac9d56263c9ffdd469fe -size 74596 diff --git a/lm-eval-output/allenai/OLMo-7B/openbookqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/openbookqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 59209452b422578f88d4701e24c064044996b69b..72596c9eff0263fa3e9b4d9174993de469dc04f3 100644 --- a/lm-eval-output/allenai/OLMo-7B/openbookqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/openbookqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -3,8 +3,8 @@ "openbookqa": { "acc,none": 0.29, "acc_stderr,none": 0.020313179231745186, - "acc_norm,none": 0.426, - "acc_norm_stderr,none": 0.022136577335085637, + "acc_norm,none": 0.422, + "acc_norm_stderr,none": 0.022109039310618552, "alias": "openbookqa" } }, @@ -52,13 +52,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/openbookqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/openbookqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index ac84f331f749c7b9653955210e249c17c016aeaf..a2fd713549f83faade05a291961e4cc85a617948 100644 --- a/lm-eval-output/allenai/OLMo-7B/openbookqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/openbookqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:bb9052e1c42cab4113478033920af72ff10328c4269868b720ecf15697b802b2 -size 11438 +oid sha256:8bbf154d17a61e29fe21f5c25f50451038275e3de8f973abff210989a9171adc +size 10916 diff --git a/lm-eval-output/allenai/OLMo-7B/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 0011e2864d7e1b75f553d58b68d33e4977070720..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2ab34586b2411fff4f2a2ec92c6bc9395a2a084c71bca9e1329350c2cfb06d0a -size 238790 diff --git a/lm-eval-output/allenai/OLMo-7B/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index c79cbc506f9b64e68780902c2a9270e8e62df627..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,62 +0,0 @@ -{ - "results": { - "piqa": { - "acc,none": 0.7883569096844396, - "acc_stderr,none": 0.009530351270479399, - "acc_norm,none": 0.7981501632208923, - "acc_norm_stderr,none": 0.009364873741341434, - "alias": "piqa" - } - }, - "configs": { - "piqa": { - "task": "piqa", - "dataset_path": "piqa", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "Question: {{goal}}\nAnswer:", - "doc_to_target": "label", - "doc_to_choice": "{{[sol1, sol2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "goal", - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "piqa": 1.0 - }, - "n-shot": { - "piqa": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index dd638bee11307bf6b48da86baae3ef7597fac0f7..240a5a607ae12e021d6cae37a348bdcdc6b115ad 100644 --- a/lm-eval-output/allenai/OLMo-7B/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:b636d33ce4cc2cc0f6aba8ecdd7dc69f127355ff288a0f65a307529ec6d6b0d0 -size 12640 +oid sha256:e4f913d0168a3cc1ed3d6d35ba4ca8eaac8cf168f0864964dc0b8076db7485e5 +size 18795 diff --git a/lm-eval-output/allenai/OLMo-7B/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 1ec56ab33cd6669f18a1306ba9ff281458d29b39..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:de33a459f094117247c099486305f283a8d8210f85d258c80a261bf9314da9ce -size 1414737 diff --git a/lm-eval-output/allenai/OLMo-7B/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 9f44abcfd98491abaf057e1e7aaeabd76ec1cea7..bcf815f9f81cca458fac476404291dc4dc845036 100644 --- a/lm-eval-output/allenai/OLMo-7B/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,10 +1,10 @@ { "results": { "prost": { - "acc,none": 0.26334329632792486, - "acc_stderr,none": 0.0032178578692137467, - "acc_norm,none": 0.3005444064901793, - "acc_norm_stderr,none": 0.003349712623195901, + "acc,none": 0.26286293766011953, + "acc_stderr,none": 0.00321596973937104, + "acc_norm,none": 0.29990392826643897, + "acc_norm_stderr,none": 0.0033476731485967267, "alias": "prost" } }, @@ -49,13 +49,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 0985547fe60cd11d44d0f322f1bf7a0d24be1df1..cd9a9248a63cfc2d2947afaabba47957db2b361f 100644 --- a/lm-eval-output/allenai/OLMo-7B/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:245669efdafe3437ca16514c6c7912b881fdf6ac1662f4400da3d8f09ed3d611 -size 71764 +oid sha256:5b41410977509307d63fdbce49c2cf38249e886277a342bdfe3e89468fd6fc74 +size 22782 diff --git a/lm-eval-output/allenai/OLMo-7B/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 5409bc415e99e637ebac1977f8db33a7cec4411b..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:82069cc5da59e8b9fe22f299d23fbcca2b84e6731cba8551cd4b2eb85f89b2f3 -size 449062 diff --git a/lm-eval-output/allenai/OLMo-7B/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 8d96a225324a8b944e33bb9783062ffd5293dcf2..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,60 +0,0 @@ -{ - "results": { - "pubmedqa": { - "acc,none": 0.692, - "acc_stderr,none": 0.0206670329874661, - "alias": "pubmedqa" - } - }, - "configs": { - "pubmedqa": { - "task": "pubmedqa", - "dataset_path": "bigbio/pubmed_qa", - "dataset_name": "pubmed_qa_labeled_fold0_source", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc) -> str:\n ctxs = \"\\n\".join(doc[\"CONTEXTS\"])\n return \"Abstract: {}\\nQuestion: {}\\nAnswer:\".format(\n ctxs,\n doc[\"QUESTION\"],\n )\n", - "doc_to_target": "final_decision", - "doc_to_choice": [ - "yes", - "no", - "maybe" - ], - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "pubmedqa": 1.0 - }, - "n-shot": { - "pubmedqa": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 091fe4539c7862f340318e7f89ed934887ef40e1..13d75b6bea7e1e5e5ee0916426b189474e1f5613 100644 --- a/lm-eval-output/allenai/OLMo-7B/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:208fe54d53b15865dfd01e3ae812c99c5ba953a989ef5d9bed3ea2242824cc9b -size 12598 +oid sha256:a02baf8dd38f1a116500b1b96dc7698cba446d5c98c7ee92fc02d563ca59d0e3 +size 51168 diff --git a/lm-eval-output/allenai/OLMo-7B/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 380eb9380a509743bce63793301638a4ab40314b..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:248577251920a5fd95ab0067bb00983f087994658a2b2f8d6aa8ca162871d748 -size 11880170 diff --git a/lm-eval-output/allenai/OLMo-7B/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 8da2f7af90858f7bcb3da412ccffa9bfd942b511..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,5232 +0,0 @@ -{ - "results": { - "pythia": { - "acc,none": 0.7275472491310511, - "acc_stderr,none": 0.1444591257781161, - "acc_norm,none": 0.596501462319602, - "acc_norm_stderr,none": 0.004459818972162648, - "word_perplexity,none": 11.075255274027937, - "word_perplexity_stderr,none": "N/A", - "byte_perplexity,none": 1.567831445754393, - "byte_perplexity_stderr,none": "N/A", - "bits_per_byte,none": 0.6487704666946622, - "bits_per_byte_stderr,none": "N/A", - "perplexity,none": 4.132675396932446, - "perplexity_stderr,none": 0.08708096011155336, - "alias": "pythia" - }, - "ai2_arc": { - "acc,none": 0.6124577226606539, - "acc_stderr,none": 0.05783349554301207, - "acc_norm,none": 0.5913190529875987, - "acc_norm_stderr,none": 0.04520523788851735, - "alias": " - ai2_arc" - }, - "arc_challenge": { - "acc,none": 0.3677474402730375, - "acc_stderr,none": 0.014090995618168475, - "acc_norm,none": 0.40273037542662116, - "acc_norm_stderr,none": 0.01433223630679014, - "alias": " - arc_challenge" - }, - "arc_easy": { - "acc,none": 0.7331649831649831, - "acc_stderr,none": 0.009075915859267257, - "acc_norm,none": 0.6843434343434344, - "acc_norm_stderr,none": 0.009537019245566087, - "alias": " - arc_easy" - }, - "blimp": { - "acc,none": 0.8318358208955224, - "acc_stderr,none": 0.15243408065668698, - "alias": " - blimp" - }, - "blimp_adjunct_island": { - "acc,none": 0.902, - "acc_stderr,none": 0.009406619184621223, - "alias": " - blimp_adjunct_island" - }, - "blimp_anaphor_gender_agreement": { - "acc,none": 0.995, - "acc_stderr,none": 0.002231586874844882, - "alias": " - blimp_anaphor_gender_agreement" - }, - "blimp_anaphor_number_agreement": { - "acc,none": 0.992, - "acc_stderr,none": 0.0028185003005045057, - "alias": " - blimp_anaphor_number_agreement" - }, - "blimp_animate_subject_passive": { - "acc,none": 0.809, - "acc_stderr,none": 0.012436787112179491, - "alias": " - blimp_animate_subject_passive" - }, - "blimp_animate_subject_trans": { - "acc,none": 0.911, - "acc_stderr,none": 0.009008893392651514, - "alias": " - blimp_animate_subject_trans" - }, - "blimp_causative": { - "acc,none": 0.743, - "acc_stderr,none": 0.013825416526895047, - "alias": " - blimp_causative" - }, - "blimp_complex_NP_island": { - "acc,none": 0.596, - "acc_stderr,none": 0.015524980677122581, - "alias": " - blimp_complex_NP_island" - }, - "blimp_coordinate_structure_constraint_complex_left_branch": { - "acc,none": 0.821, - "acc_stderr,none": 0.012128730605719113, - "alias": " - blimp_coordinate_structure_constraint_complex_left_branch" - }, - "blimp_coordinate_structure_constraint_object_extraction": { - "acc,none": 0.889, - "acc_stderr,none": 0.009938701010583726, - "alias": " - blimp_coordinate_structure_constraint_object_extraction" - }, - "blimp_determiner_noun_agreement_1": { - "acc,none": 0.986, - "acc_stderr,none": 0.0037172325482565916, - "alias": " - blimp_determiner_noun_agreement_1" - }, - "blimp_determiner_noun_agreement_2": { - "acc,none": 0.975, - "acc_stderr,none": 0.0049395748196984545, - "alias": " - blimp_determiner_noun_agreement_2" - }, - "blimp_determiner_noun_agreement_irregular_1": { - "acc,none": 0.943, - "acc_stderr,none": 0.0073351758537068355, - "alias": " - blimp_determiner_noun_agreement_irregular_1" - }, - "blimp_determiner_noun_agreement_irregular_2": { - "acc,none": 0.951, - "acc_stderr,none": 0.006829761756140914, - "alias": " - blimp_determiner_noun_agreement_irregular_2" - }, - "blimp_determiner_noun_agreement_with_adj_2": { - "acc,none": 0.948, - "acc_stderr,none": 0.007024624213817146, - "alias": " - blimp_determiner_noun_agreement_with_adj_2" - }, - "blimp_determiner_noun_agreement_with_adj_irregular_1": { - "acc,none": 0.882, - "acc_stderr,none": 0.01020686926438179, - "alias": " - blimp_determiner_noun_agreement_with_adj_irregular_1" - }, - "blimp_determiner_noun_agreement_with_adj_irregular_2": { - "acc,none": 0.929, - "acc_stderr,none": 0.008125578442487916, - "alias": " - blimp_determiner_noun_agreement_with_adj_irregular_2" - }, - "blimp_determiner_noun_agreement_with_adjective_1": { - "acc,none": 0.971, - "acc_stderr,none": 0.005309160685756979, - "alias": " - blimp_determiner_noun_agreement_with_adjective_1" - }, - "blimp_distractor_agreement_relational_noun": { - "acc,none": 0.919, - "acc_stderr,none": 0.008632121032139986, - "alias": " - blimp_distractor_agreement_relational_noun" - }, - "blimp_distractor_agreement_relative_clause": { - "acc,none": 0.705, - "acc_stderr,none": 0.014428554438445523, - "alias": " - blimp_distractor_agreement_relative_clause" - }, - "blimp_drop_argument": { - "acc,none": 0.746, - "acc_stderr,none": 0.013772206565168537, - "alias": " - blimp_drop_argument" - }, - "blimp_ellipsis_n_bar_1": { - "acc,none": 0.812, - "acc_stderr,none": 0.012361586015103749, - "alias": " - blimp_ellipsis_n_bar_1" - }, - "blimp_ellipsis_n_bar_2": { - "acc,none": 0.948, - "acc_stderr,none": 0.0070246242138171325, - "alias": " - blimp_ellipsis_n_bar_2" - }, - "blimp_existential_there_object_raising": { - "acc,none": 0.859, - "acc_stderr,none": 0.011010914595992438, - "alias": " - blimp_existential_there_object_raising" - }, - "blimp_existential_there_quantifiers_1": { - "acc,none": 0.984, - "acc_stderr,none": 0.0039698563903194225, - "alias": " - blimp_existential_there_quantifiers_1" - }, - "blimp_existential_there_quantifiers_2": { - "acc,none": 0.377, - "acc_stderr,none": 0.01533317012577985, - "alias": " - blimp_existential_there_quantifiers_2" - }, - "blimp_existential_there_subject_raising": { - "acc,none": 0.909, - "acc_stderr,none": 0.009099549538400252, - "alias": " - blimp_existential_there_subject_raising" - }, - "blimp_expletive_it_object_raising": { - "acc,none": 0.827, - "acc_stderr,none": 0.011967214137559929, - "alias": " - blimp_expletive_it_object_raising" - }, - "blimp_inchoative": { - "acc,none": 0.678, - "acc_stderr,none": 0.014782913600996662, - "alias": " - blimp_inchoative" - }, - "blimp_intransitive": { - "acc,none": 0.8, - "acc_stderr,none": 0.012655439943366651, - "alias": " - blimp_intransitive" - }, - "blimp_irregular_past_participle_adjectives": { - "acc,none": 0.979, - "acc_stderr,none": 0.00453647215130652, - "alias": " - blimp_irregular_past_participle_adjectives" - }, - "blimp_irregular_past_participle_verbs": { - "acc,none": 0.912, - "acc_stderr,none": 0.008963053962592081, - "alias": " - blimp_irregular_past_participle_verbs" - }, - "blimp_irregular_plural_subject_verb_agreement_1": { - "acc,none": 0.93, - "acc_stderr,none": 0.008072494358323508, - "alias": " - blimp_irregular_plural_subject_verb_agreement_1" - }, - "blimp_irregular_plural_subject_verb_agreement_2": { - "acc,none": 0.937, - "acc_stderr,none": 0.007687007876286417, - "alias": " - blimp_irregular_plural_subject_verb_agreement_2" - }, - "blimp_left_branch_island_echo_question": { - "acc,none": 0.633, - "acc_stderr,none": 0.015249378464171745, - "alias": " - blimp_left_branch_island_echo_question" - }, - "blimp_left_branch_island_simple_question": { - "acc,none": 0.909, - "acc_stderr,none": 0.009099549538400227, - "alias": " - blimp_left_branch_island_simple_question" - }, - "blimp_matrix_question_npi_licensor_present": { - "acc,none": 0.617, - "acc_stderr,none": 0.015380102325652713, - "alias": " - blimp_matrix_question_npi_licensor_present" - }, - "blimp_npi_present_1": { - "acc,none": 0.671, - "acc_stderr,none": 0.014865395385928362, - "alias": " - blimp_npi_present_1" - }, - "blimp_npi_present_2": { - "acc,none": 0.735, - "acc_stderr,none": 0.013963164754809953, - "alias": " - blimp_npi_present_2" - }, - "blimp_only_npi_licensor_present": { - "acc,none": 0.971, - "acc_stderr,none": 0.005309160685756993, - "alias": " - blimp_only_npi_licensor_present" - }, - "blimp_only_npi_scope": { - "acc,none": 0.678, - "acc_stderr,none": 0.014782913600996685, - "alias": " - blimp_only_npi_scope" - }, - "blimp_passive_1": { - "acc,none": 0.897, - "acc_stderr,none": 0.0096168333396958, - "alias": " - blimp_passive_1" - }, - "blimp_passive_2": { - "acc,none": 0.909, - "acc_stderr,none": 0.009099549538400236, - "alias": " - blimp_passive_2" - }, - "blimp_principle_A_c_command": { - "acc,none": 0.74, - "acc_stderr,none": 0.013877773329774164, - "alias": " - blimp_principle_A_c_command" - }, - "blimp_principle_A_case_1": { - "acc,none": 1.0, - "acc_stderr,none": 0.0, - "alias": " - blimp_principle_A_case_1" - }, - "blimp_principle_A_case_2": { - "acc,none": 0.967, - "acc_stderr,none": 0.0056518088204523705, - "alias": " - blimp_principle_A_case_2" - }, - "blimp_principle_A_domain_1": { - "acc,none": 0.998, - "acc_stderr,none": 0.0014135055705578, - "alias": " - blimp_principle_A_domain_1" - }, - "blimp_principle_A_domain_2": { - "acc,none": 0.84, - "acc_stderr,none": 0.011598902298689004, - "alias": " - blimp_principle_A_domain_2" - }, - "blimp_principle_A_domain_3": { - "acc,none": 0.736, - "acc_stderr,none": 0.013946271849440467, - "alias": " - blimp_principle_A_domain_3" - }, - "blimp_principle_A_reconstruction": { - "acc,none": 0.376, - "acc_stderr,none": 0.01532510550889813, - "alias": " - blimp_principle_A_reconstruction" - }, - "blimp_regular_plural_subject_verb_agreement_1": { - "acc,none": 0.967, - "acc_stderr,none": 0.005651808820452372, - "alias": " - blimp_regular_plural_subject_verb_agreement_1" - }, - "blimp_regular_plural_subject_verb_agreement_2": { - "acc,none": 0.935, - "acc_stderr,none": 0.007799733061832028, - "alias": " - blimp_regular_plural_subject_verb_agreement_2" - }, - "blimp_sentential_negation_npi_licensor_present": { - "acc,none": 0.996, - "acc_stderr,none": 0.00199699473909873, - "alias": " - blimp_sentential_negation_npi_licensor_present" - }, - "blimp_sentential_negation_npi_scope": { - "acc,none": 0.761, - "acc_stderr,none": 0.013493000446937591, - "alias": " - blimp_sentential_negation_npi_scope" - }, - "blimp_sentential_subject_island": { - "acc,none": 0.562, - "acc_stderr,none": 0.01569721001969469, - "alias": " - blimp_sentential_subject_island" - }, - "blimp_superlative_quantifiers_1": { - "acc,none": 0.897, - "acc_stderr,none": 0.009616833339695784, - "alias": " - blimp_superlative_quantifiers_1" - }, - "blimp_superlative_quantifiers_2": { - "acc,none": 0.881, - "acc_stderr,none": 0.010244215145336666, - "alias": " - blimp_superlative_quantifiers_2" - }, - "blimp_tough_vs_raising_1": { - "acc,none": 0.669, - "acc_stderr,none": 0.014888272588203934, - "alias": " - blimp_tough_vs_raising_1" - }, - "blimp_tough_vs_raising_2": { - "acc,none": 0.856, - "acc_stderr,none": 0.01110798754893915, - "alias": " - blimp_tough_vs_raising_2" - }, - "blimp_transitive": { - "acc,none": 0.862, - "acc_stderr,none": 0.010912152632504378, - "alias": " - blimp_transitive" - }, - "blimp_wh_island": { - "acc,none": 0.879, - "acc_stderr,none": 0.010318210380946085, - "alias": " - blimp_wh_island" - }, - "blimp_wh_questions_object_gap": { - "acc,none": 0.84, - "acc_stderr,none": 0.011598902298689004, - "alias": " - blimp_wh_questions_object_gap" - }, - "blimp_wh_questions_subject_gap": { - "acc,none": 0.931, - "acc_stderr,none": 0.00801893405031516, - "alias": " - blimp_wh_questions_subject_gap" - }, - "blimp_wh_questions_subject_gap_long_distance": { - "acc,none": 0.919, - "acc_stderr,none": 0.008632121032139969, - "alias": " - blimp_wh_questions_subject_gap_long_distance" - }, - "blimp_wh_vs_that_no_gap": { - "acc,none": 0.973, - "acc_stderr,none": 0.005128089049275289, - "alias": " - blimp_wh_vs_that_no_gap" - }, - "blimp_wh_vs_that_no_gap_long_distance": { - "acc,none": 0.961, - "acc_stderr,none": 0.006125072776426141, - "alias": " - blimp_wh_vs_that_no_gap_long_distance" - }, - "blimp_wh_vs_that_with_gap": { - "acc,none": 0.349, - "acc_stderr,none": 0.015080663991563098, - "alias": " - blimp_wh_vs_that_with_gap" - }, - "blimp_wh_vs_that_with_gap_long_distance": { - "acc,none": 0.253, - "acc_stderr,none": 0.01375427861358708, - "alias": " - blimp_wh_vs_that_with_gap_long_distance" - }, - "lambada_openai": { - "perplexity,none": 4.132675396932446, - "perplexity_stderr,none": 0.08708096011155336, - "acc,none": 0.6875606442848826, - "acc_stderr,none": 0.006457292279746485, - "alias": " - lambada_openai" - }, - "logiqa": { - "acc,none": 0.23809523809523808, - "acc_stderr,none": 0.01670586703441963, - "acc_norm,none": 0.27035330261136714, - "acc_norm_stderr,none": 0.01742069478339314, - "alias": " - logiqa" - }, - "mmlu": { - "acc,none": 0.2824383990884489, - "acc_stderr,none": 0.04247561258047098, - "alias": " - mmlu" - }, - "mmlu_humanities": { - "alias": " - humanities", - "acc,none": 0.2737513283740701, - "acc_stderr,none": 0.0411011139362443 - }, - "mmlu_formal_logic": { - "alias": " - formal_logic", - "acc,none": 0.38095238095238093, - "acc_stderr,none": 0.043435254289490965 - }, - "mmlu_high_school_european_history": { - "alias": " - high_school_european_history", - "acc,none": 0.23636363636363636, - "acc_stderr,none": 0.033175059300091805 - }, - "mmlu_high_school_us_history": { - "alias": " - high_school_us_history", - "acc,none": 0.24019607843137256, - "acc_stderr,none": 0.02998373305591361 - }, - "mmlu_high_school_world_history": { - "alias": " - high_school_world_history", - "acc,none": 0.23628691983122363, - "acc_stderr,none": 0.02765215314415928 - }, - "mmlu_international_law": { - "alias": " - international_law", - "acc,none": 0.35537190082644626, - "acc_stderr,none": 0.04369236326573981 - }, - "mmlu_jurisprudence": { - "alias": " - jurisprudence", - "acc,none": 0.2037037037037037, - "acc_stderr,none": 0.03893542518824848 - }, - "mmlu_logical_fallacies": { - "alias": " - logical_fallacies", - "acc,none": 0.294478527607362, - "acc_stderr,none": 0.03581165790474082 - }, - "mmlu_moral_disputes": { - "alias": " - moral_disputes", - "acc,none": 0.27167630057803466, - "acc_stderr,none": 0.023948512905468348 - }, - "mmlu_moral_scenarios": { - "alias": " - moral_scenarios", - "acc,none": 0.2424581005586592, - "acc_stderr,none": 0.014333522059217892 - }, - "mmlu_philosophy": { - "alias": " - philosophy", - "acc,none": 0.31511254019292606, - "acc_stderr,none": 0.026385273703464485 - }, - "mmlu_prehistory": { - "alias": " - prehistory", - "acc,none": 0.29012345679012347, - "acc_stderr,none": 0.025251173936495022 - }, - "mmlu_professional_law": { - "alias": " - professional_law", - "acc,none": 0.273142112125163, - "acc_stderr,none": 0.011380150567830403 - }, - "mmlu_world_religions": { - "alias": " - world_religions", - "acc,none": 0.3567251461988304, - "acc_stderr,none": 0.03674013002860954 - }, - "mmlu_other": { - "alias": " - other", - "acc,none": 0.2845188284518829, - "acc_stderr,none": 0.04353064334839231 - }, - "mmlu_business_ethics": { - "alias": " - business_ethics", - "acc,none": 0.27, - "acc_stderr,none": 0.0446196043338474 - }, - "mmlu_clinical_knowledge": { - "alias": " - clinical_knowledge", - "acc,none": 0.25660377358490566, - "acc_stderr,none": 0.026880647889051992 - }, - "mmlu_college_medicine": { - "alias": " - college_medicine", - "acc,none": 0.3236994219653179, - "acc_stderr,none": 0.0356760379963917 - }, - "mmlu_global_facts": { - "alias": " - global_facts", - "acc,none": 0.36, - "acc_stderr,none": 0.048241815132442176 - }, - "mmlu_human_aging": { - "alias": " - human_aging", - "acc,none": 0.2242152466367713, - "acc_stderr,none": 0.027991534258519527 - }, - "mmlu_management": { - "alias": " - management", - "acc,none": 0.2524271844660194, - "acc_stderr,none": 0.04301250399690878 - }, - "mmlu_marketing": { - "alias": " - marketing", - "acc,none": 0.2564102564102564, - "acc_stderr,none": 0.028605953702004253 - }, - "mmlu_medical_genetics": { - "alias": " - medical_genetics", - "acc,none": 0.27, - "acc_stderr,none": 0.044619604333847394 - }, - "mmlu_miscellaneous": { - "alias": " - miscellaneous", - "acc,none": 0.30268199233716475, - "acc_stderr,none": 0.01642878158174936 - }, - "mmlu_nutrition": { - "alias": " - nutrition", - "acc,none": 0.34967320261437906, - "acc_stderr,none": 0.0273053080762747 - }, - "mmlu_professional_accounting": { - "alias": " - professional_accounting", - "acc,none": 0.29432624113475175, - "acc_stderr,none": 0.027187127011503793 - }, - "mmlu_professional_medicine": { - "alias": " - professional_medicine", - "acc,none": 0.23529411764705882, - "acc_stderr,none": 0.025767252010855973 - }, - "mmlu_virology": { - "alias": " - virology", - "acc,none": 0.25903614457831325, - "acc_stderr,none": 0.034106466140718564 - }, - "mmlu_social_sciences": { - "alias": " - social_sciences", - "acc,none": 0.2918427039324017, - "acc_stderr,none": 0.03431852465251 - }, - "mmlu_econometrics": { - "alias": " - econometrics", - "acc,none": 0.2543859649122807, - "acc_stderr,none": 0.040969851398436695 - }, - "mmlu_high_school_geography": { - "alias": " - high_school_geography", - "acc,none": 0.31313131313131315, - "acc_stderr,none": 0.03304205087813652 - }, - "mmlu_high_school_government_and_politics": { - "alias": " - high_school_government_and_politics", - "acc,none": 0.30569948186528495, - "acc_stderr,none": 0.033248379397581594 - }, - "mmlu_high_school_macroeconomics": { - "alias": " - high_school_macroeconomics", - "acc,none": 0.30512820512820515, - "acc_stderr,none": 0.023346335293325887 - }, - "mmlu_high_school_microeconomics": { - "alias": " - high_school_microeconomics", - "acc,none": 0.2815126050420168, - "acc_stderr,none": 0.02921354941437217 - }, - "mmlu_high_school_psychology": { - "alias": " - high_school_psychology", - "acc,none": 0.28807339449541286, - "acc_stderr,none": 0.01941644589263603 - }, - "mmlu_human_sexuality": { - "alias": " - human_sexuality", - "acc,none": 0.31297709923664124, - "acc_stderr,none": 0.04066962905677697 - }, - "mmlu_professional_psychology": { - "alias": " - professional_psychology", - "acc,none": 0.2696078431372549, - "acc_stderr,none": 0.017952449196987862 - }, - "mmlu_public_relations": { - "alias": " - public_relations", - "acc,none": 0.2545454545454545, - "acc_stderr,none": 0.041723430387053825 - }, - "mmlu_security_studies": { - "alias": " - security_studies", - "acc,none": 0.3020408163265306, - "acc_stderr,none": 0.029393609319879815 - }, - "mmlu_sociology": { - "alias": " - sociology", - "acc,none": 0.2935323383084577, - "acc_stderr,none": 0.03220024104534204 - }, - "mmlu_us_foreign_policy": { - "alias": " - us_foreign_policy", - "acc,none": 0.38, - "acc_stderr,none": 0.04878317312145632 - }, - "mmlu_stem": { - "alias": " - stem", - "acc,none": 0.284173802727561, - "acc_stderr,none": 0.049209572972660855 - }, - "mmlu_abstract_algebra": { - "alias": " - abstract_algebra", - "acc,none": 0.27, - "acc_stderr,none": 0.04461960433384741 - }, - "mmlu_anatomy": { - "alias": " - anatomy", - "acc,none": 0.26666666666666666, - "acc_stderr,none": 0.038201699145179055 - }, - "mmlu_astronomy": { - "alias": " - astronomy", - "acc,none": 0.34210526315789475, - "acc_stderr,none": 0.03860731599316092 - }, - "mmlu_college_biology": { - "alias": " - college_biology", - "acc,none": 0.2916666666666667, - "acc_stderr,none": 0.038009680605548574 - }, - "mmlu_college_chemistry": { - "alias": " - college_chemistry", - "acc,none": 0.22, - "acc_stderr,none": 0.0416333199893227 - }, - "mmlu_college_computer_science": { - "alias": " - college_computer_science", - "acc,none": 0.38, - "acc_stderr,none": 0.048783173121456344 - }, - "mmlu_college_mathematics": { - "alias": " - college_mathematics", - "acc,none": 0.4, - "acc_stderr,none": 0.04923659639173309 - }, - "mmlu_college_physics": { - "alias": " - college_physics", - "acc,none": 0.23529411764705882, - "acc_stderr,none": 0.04220773659171453 - }, - "mmlu_computer_security": { - "alias": " - computer_security", - "acc,none": 0.36, - "acc_stderr,none": 0.04824181513244218 - }, - "mmlu_conceptual_physics": { - "alias": " - conceptual_physics", - "acc,none": 0.24680851063829787, - "acc_stderr,none": 0.028185441301234106 - }, - "mmlu_electrical_engineering": { - "alias": " - electrical_engineering", - "acc,none": 0.296551724137931, - "acc_stderr,none": 0.03806142687309993 - }, - "mmlu_elementary_mathematics": { - "alias": " - elementary_mathematics", - "acc,none": 0.2698412698412698, - "acc_stderr,none": 0.02286083830923207 - }, - "mmlu_high_school_biology": { - "alias": " - high_school_biology", - "acc,none": 0.25161290322580643, - "acc_stderr,none": 0.024685979286239952 - }, - "mmlu_high_school_chemistry": { - "alias": " - high_school_chemistry", - "acc,none": 0.29064039408866993, - "acc_stderr,none": 0.031947400722655395 - }, - "mmlu_high_school_computer_science": { - "alias": " - high_school_computer_science", - "acc,none": 0.27, - "acc_stderr,none": 0.04461960433384741 - }, - "mmlu_high_school_mathematics": { - "alias": " - high_school_mathematics", - "acc,none": 0.2740740740740741, - "acc_stderr,none": 0.027195934804085626 - }, - "mmlu_high_school_physics": { - "alias": " - high_school_physics", - "acc,none": 0.2847682119205298, - "acc_stderr,none": 0.03684881521389023 - }, - "mmlu_high_school_statistics": { - "alias": " - high_school_statistics", - "acc,none": 0.2824074074074074, - "acc_stderr,none": 0.030701372111510923 - }, - "mmlu_machine_learning": { - "alias": " - machine_learning", - "acc,none": 0.30357142857142855, - "acc_stderr,none": 0.04364226155841044 - }, - "piqa": { - "acc,none": 0.7883569096844396, - "acc_stderr,none": 0.009530351270479399, - "acc_norm,none": 0.7981501632208923, - "acc_norm_stderr,none": 0.009364873741341434, - "alias": " - piqa" - }, - "sciq": { - "acc,none": 0.929, - "acc_stderr,none": 0.008125578442487923, - "acc_norm,none": 0.884, - "acc_norm_stderr,none": 0.010131468138756988, - "alias": " - sciq" - }, - "wikitext": { - "word_perplexity,none": 11.075255274027937, - "word_perplexity_stderr,none": "N/A", - "byte_perplexity,none": 1.567831445754393, - "byte_perplexity_stderr,none": "N/A", - "bits_per_byte,none": 0.6487704666946622, - "bits_per_byte_stderr,none": "N/A", - "alias": " - wikitext" - }, - "winogrande": { - "acc,none": 0.6629834254143646, - "acc_stderr,none": 0.01328495576939525, - "alias": " - winogrande" - }, - "wsc": { - "acc,none": 0.38461538461538464, - "acc_stderr,none": 0.047936688680750406, - "alias": " - wsc" - } - }, - "groups": { - "pythia": { - "acc,none": 0.7275472491310511, - "acc_stderr,none": 0.1444591257781161, - "acc_norm,none": 0.596501462319602, - "acc_norm_stderr,none": 0.004459818972162648, - "word_perplexity,none": 11.075255274027937, - "word_perplexity_stderr,none": "N/A", - "byte_perplexity,none": 1.567831445754393, - "byte_perplexity_stderr,none": "N/A", - "bits_per_byte,none": 0.6487704666946622, - "bits_per_byte_stderr,none": "N/A", - "perplexity,none": 4.132675396932446, - "perplexity_stderr,none": 0.08708096011155336, - "alias": "pythia" - }, - "ai2_arc": { - "acc,none": 0.6124577226606539, - "acc_stderr,none": 0.05783349554301207, - "acc_norm,none": 0.5913190529875987, - "acc_norm_stderr,none": 0.04520523788851735, - "alias": " - ai2_arc" - }, - "blimp": { - "acc,none": 0.8318358208955224, - "acc_stderr,none": 0.15243408065668698, - "alias": " - blimp" - }, - "mmlu": { - "acc,none": 0.2824383990884489, - "acc_stderr,none": 0.04247561258047098, - "alias": " - mmlu" - }, - "mmlu_humanities": { - "alias": " - humanities", - "acc,none": 0.2737513283740701, - "acc_stderr,none": 0.0411011139362443 - }, - "mmlu_other": { - "alias": " - other", - "acc,none": 0.2845188284518829, - "acc_stderr,none": 0.04353064334839231 - }, - "mmlu_social_sciences": { - "alias": " - social_sciences", - "acc,none": 0.2918427039324017, - "acc_stderr,none": 0.03431852465251 - }, - "mmlu_stem": { - "alias": " - stem", - "acc,none": 0.284173802727561, - "acc_stderr,none": 0.049209572972660855 - } - }, - "configs": { - "arc_challenge": { - "task": "arc_challenge", - "group": [ - "ai2_arc" - ], - "dataset_path": "allenai/ai2_arc", - "dataset_name": "ARC-Challenge", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "Question: {{question}}\nAnswer:", - "doc_to_target": "{{choices.label.index(answerKey)}}", - "doc_to_choice": "{{choices.text}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "Question: {{question}}\nAnswer:", - "metadata": { - "version": 1.0 - } - }, - "arc_easy": { - "task": "arc_easy", - "group": [ - "ai2_arc" - ], - "dataset_path": "allenai/ai2_arc", - "dataset_name": "ARC-Easy", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "Question: {{question}}\nAnswer:", - "doc_to_target": "{{choices.label.index(answerKey)}}", - "doc_to_choice": "{{choices.text}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "Question: {{question}}\nAnswer:", - "metadata": { - "version": 1.0 - } - }, - "blimp_adjunct_island": { - "task": "blimp_adjunct_island", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "adjunct_island", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_anaphor_gender_agreement": { - "task": "blimp_anaphor_gender_agreement", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "anaphor_gender_agreement", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_anaphor_number_agreement": { - "task": "blimp_anaphor_number_agreement", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "anaphor_number_agreement", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_animate_subject_passive": { - "task": "blimp_animate_subject_passive", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "animate_subject_passive", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_animate_subject_trans": { - "task": "blimp_animate_subject_trans", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "animate_subject_trans", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_causative": { - "task": "blimp_causative", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "causative", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_complex_NP_island": { - "task": "blimp_complex_NP_island", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "complex_NP_island", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_coordinate_structure_constraint_complex_left_branch": { - "task": "blimp_coordinate_structure_constraint_complex_left_branch", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "coordinate_structure_constraint_complex_left_branch", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_coordinate_structure_constraint_object_extraction": { - "task": "blimp_coordinate_structure_constraint_object_extraction", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "coordinate_structure_constraint_object_extraction", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_determiner_noun_agreement_1": { - "task": "blimp_determiner_noun_agreement_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "determiner_noun_agreement_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_determiner_noun_agreement_2": { - "task": "blimp_determiner_noun_agreement_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "determiner_noun_agreement_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_determiner_noun_agreement_irregular_1": { - "task": "blimp_determiner_noun_agreement_irregular_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "determiner_noun_agreement_irregular_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_determiner_noun_agreement_irregular_2": { - "task": "blimp_determiner_noun_agreement_irregular_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "determiner_noun_agreement_irregular_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_determiner_noun_agreement_with_adj_2": { - "task": "blimp_determiner_noun_agreement_with_adj_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "determiner_noun_agreement_with_adj_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_determiner_noun_agreement_with_adj_irregular_1": { - "task": "blimp_determiner_noun_agreement_with_adj_irregular_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "determiner_noun_agreement_with_adj_irregular_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_determiner_noun_agreement_with_adj_irregular_2": { - "task": "blimp_determiner_noun_agreement_with_adj_irregular_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "determiner_noun_agreement_with_adj_irregular_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_determiner_noun_agreement_with_adjective_1": { - "task": "blimp_determiner_noun_agreement_with_adjective_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "determiner_noun_agreement_with_adjective_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_distractor_agreement_relational_noun": { - "task": "blimp_distractor_agreement_relational_noun", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "distractor_agreement_relational_noun", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_distractor_agreement_relative_clause": { - "task": "blimp_distractor_agreement_relative_clause", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "distractor_agreement_relative_clause", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_drop_argument": { - "task": "blimp_drop_argument", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "drop_argument", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_ellipsis_n_bar_1": { - "task": "blimp_ellipsis_n_bar_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "ellipsis_n_bar_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_ellipsis_n_bar_2": { - "task": "blimp_ellipsis_n_bar_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "ellipsis_n_bar_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_existential_there_object_raising": { - "task": "blimp_existential_there_object_raising", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "existential_there_object_raising", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_existential_there_quantifiers_1": { - "task": "blimp_existential_there_quantifiers_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "existential_there_quantifiers_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_existential_there_quantifiers_2": { - "task": "blimp_existential_there_quantifiers_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "existential_there_quantifiers_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_existential_there_subject_raising": { - "task": "blimp_existential_there_subject_raising", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "existential_there_subject_raising", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_expletive_it_object_raising": { - "task": "blimp_expletive_it_object_raising", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "expletive_it_object_raising", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_inchoative": { - "task": "blimp_inchoative", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "inchoative", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_intransitive": { - "task": "blimp_intransitive", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "intransitive", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_irregular_past_participle_adjectives": { - "task": "blimp_irregular_past_participle_adjectives", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "irregular_past_participle_adjectives", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_irregular_past_participle_verbs": { - "task": "blimp_irregular_past_participle_verbs", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "irregular_past_participle_verbs", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_irregular_plural_subject_verb_agreement_1": { - "task": "blimp_irregular_plural_subject_verb_agreement_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "irregular_plural_subject_verb_agreement_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_irregular_plural_subject_verb_agreement_2": { - "task": "blimp_irregular_plural_subject_verb_agreement_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "irregular_plural_subject_verb_agreement_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_left_branch_island_echo_question": { - "task": "blimp_left_branch_island_echo_question", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "left_branch_island_echo_question", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_left_branch_island_simple_question": { - "task": "blimp_left_branch_island_simple_question", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "left_branch_island_simple_question", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_matrix_question_npi_licensor_present": { - "task": "blimp_matrix_question_npi_licensor_present", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "matrix_question_npi_licensor_present", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_npi_present_1": { - "task": "blimp_npi_present_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "npi_present_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_npi_present_2": { - "task": "blimp_npi_present_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "npi_present_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_only_npi_licensor_present": { - "task": "blimp_only_npi_licensor_present", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "only_npi_licensor_present", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_only_npi_scope": { - "task": "blimp_only_npi_scope", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "only_npi_scope", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_passive_1": { - "task": "blimp_passive_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "passive_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_passive_2": { - "task": "blimp_passive_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "passive_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_principle_A_c_command": { - "task": "blimp_principle_A_c_command", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "principle_A_c_command", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_principle_A_case_1": { - "task": "blimp_principle_A_case_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "principle_A_case_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_principle_A_case_2": { - "task": "blimp_principle_A_case_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "principle_A_case_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_principle_A_domain_1": { - "task": "blimp_principle_A_domain_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "principle_A_domain_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_principle_A_domain_2": { - "task": "blimp_principle_A_domain_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "principle_A_domain_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_principle_A_domain_3": { - "task": "blimp_principle_A_domain_3", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "principle_A_domain_3", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_principle_A_reconstruction": { - "task": "blimp_principle_A_reconstruction", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "principle_A_reconstruction", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_regular_plural_subject_verb_agreement_1": { - "task": "blimp_regular_plural_subject_verb_agreement_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "regular_plural_subject_verb_agreement_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_regular_plural_subject_verb_agreement_2": { - "task": "blimp_regular_plural_subject_verb_agreement_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "regular_plural_subject_verb_agreement_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_sentential_negation_npi_licensor_present": { - "task": "blimp_sentential_negation_npi_licensor_present", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "sentential_negation_npi_licensor_present", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_sentential_negation_npi_scope": { - "task": "blimp_sentential_negation_npi_scope", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "sentential_negation_npi_scope", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_sentential_subject_island": { - "task": "blimp_sentential_subject_island", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "sentential_subject_island", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_superlative_quantifiers_1": { - "task": "blimp_superlative_quantifiers_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "superlative_quantifiers_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_superlative_quantifiers_2": { - "task": "blimp_superlative_quantifiers_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "superlative_quantifiers_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_tough_vs_raising_1": { - "task": "blimp_tough_vs_raising_1", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "tough_vs_raising_1", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_tough_vs_raising_2": { - "task": "blimp_tough_vs_raising_2", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "tough_vs_raising_2", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_transitive": { - "task": "blimp_transitive", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "transitive", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_wh_island": { - "task": "blimp_wh_island", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "wh_island", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_wh_questions_object_gap": { - "task": "blimp_wh_questions_object_gap", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "wh_questions_object_gap", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_wh_questions_subject_gap": { - "task": "blimp_wh_questions_subject_gap", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "wh_questions_subject_gap", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_wh_questions_subject_gap_long_distance": { - "task": "blimp_wh_questions_subject_gap_long_distance", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "wh_questions_subject_gap_long_distance", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_wh_vs_that_no_gap": { - "task": "blimp_wh_vs_that_no_gap", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "wh_vs_that_no_gap", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_wh_vs_that_no_gap_long_distance": { - "task": "blimp_wh_vs_that_no_gap_long_distance", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "wh_vs_that_no_gap_long_distance", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_wh_vs_that_with_gap": { - "task": "blimp_wh_vs_that_with_gap", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "wh_vs_that_with_gap", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "blimp_wh_vs_that_with_gap_long_distance": { - "task": "blimp_wh_vs_that_with_gap_long_distance", - "group": "blimp", - "dataset_path": "blimp", - "dataset_name": "wh_vs_that_with_gap_long_distance", - "validation_split": "train", - "doc_to_text": "", - "doc_to_target": 0, - "doc_to_choice": "{{[sentence_good, sentence_bad]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "num_fewshot": 0, - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{sentence_good}} {{sentence_bad}}", - "metadata": { - "version": 1.0 - } - }, - "lambada_openai": { - "task": "lambada_openai", - "group": [ - "lambada" - ], - "dataset_path": "EleutherAI/lambada_openai", - "dataset_name": "default", - "test_split": "test", - "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}", - "doc_to_target": "{{' '+text.split(' ')[-1]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "perplexity", - "aggregation": "perplexity", - "higher_is_better": false - }, - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "loglikelihood", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{text}}", - "metadata": { - "version": 1.0 - } - }, - "logiqa": { - "task": "logiqa", - "dataset_path": "EleutherAI/logiqa", - "dataset_name": "logiqa", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc) -> str:\n \"\"\"\n Passage: \n Question: \n Choices:\n A. \n B. \n C. \n D. \n Answer:\n \"\"\"\n choices = [\"a\", \"b\", \"c\", \"d\"]\n prompt = \"Passage: \" + doc[\"context\"] + \"\\n\"\n prompt += \"Question: \" + doc[\"question\"] + \"\\nChoices:\\n\"\n for choice, option in zip(choices, doc[\"options\"]):\n prompt += f\"{choice.upper()}. {option}\\n\"\n prompt += \"Answer:\"\n return prompt\n", - "doc_to_target": "def doc_to_target(doc) -> int:\n choices = [\"a\", \"b\", \"c\", \"d\"]\n return choices.index(doc[\"label\"].strip())\n", - "doc_to_choice": "{{options}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{context}}", - "metadata": { - "version": 1.0 - } - }, - "mmlu_abstract_algebra": { - "task": "mmlu_abstract_algebra", - "task_alias": "abstract_algebra", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "abstract_algebra", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about abstract algebra.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_anatomy": { - "task": "mmlu_anatomy", - "task_alias": "anatomy", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "anatomy", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about anatomy.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_astronomy": { - "task": "mmlu_astronomy", - "task_alias": "astronomy", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "astronomy", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about astronomy.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_business_ethics": { - "task": "mmlu_business_ethics", - "task_alias": "business_ethics", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "business_ethics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about business ethics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_clinical_knowledge": { - "task": "mmlu_clinical_knowledge", - "task_alias": "clinical_knowledge", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "clinical_knowledge", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about clinical knowledge.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_biology": { - "task": "mmlu_college_biology", - "task_alias": "college_biology", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_biology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college biology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_chemistry": { - "task": "mmlu_college_chemistry", - "task_alias": "college_chemistry", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_chemistry", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college chemistry.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_computer_science": { - "task": "mmlu_college_computer_science", - "task_alias": "college_computer_science", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_computer_science", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college computer science.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_mathematics": { - "task": "mmlu_college_mathematics", - "task_alias": "college_mathematics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_mathematics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college mathematics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_medicine": { - "task": "mmlu_college_medicine", - "task_alias": "college_medicine", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_medicine", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college medicine.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_college_physics": { - "task": "mmlu_college_physics", - "task_alias": "college_physics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "college_physics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about college physics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_computer_security": { - "task": "mmlu_computer_security", - "task_alias": "computer_security", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "computer_security", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about computer security.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_conceptual_physics": { - "task": "mmlu_conceptual_physics", - "task_alias": "conceptual_physics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "conceptual_physics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about conceptual physics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_econometrics": { - "task": "mmlu_econometrics", - "task_alias": "econometrics", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "econometrics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about econometrics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_electrical_engineering": { - "task": "mmlu_electrical_engineering", - "task_alias": "electrical_engineering", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "electrical_engineering", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about electrical engineering.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_elementary_mathematics": { - "task": "mmlu_elementary_mathematics", - "task_alias": "elementary_mathematics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "elementary_mathematics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about elementary mathematics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_formal_logic": { - "task": "mmlu_formal_logic", - "task_alias": "formal_logic", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "formal_logic", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about formal logic.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_global_facts": { - "task": "mmlu_global_facts", - "task_alias": "global_facts", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "global_facts", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about global facts.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_biology": { - "task": "mmlu_high_school_biology", - "task_alias": "high_school_biology", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_biology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school biology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_chemistry": { - "task": "mmlu_high_school_chemistry", - "task_alias": "high_school_chemistry", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_chemistry", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school chemistry.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_computer_science": { - "task": "mmlu_high_school_computer_science", - "task_alias": "high_school_computer_science", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_computer_science", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school computer science.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_european_history": { - "task": "mmlu_high_school_european_history", - "task_alias": "high_school_european_history", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_european_history", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school european history.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_geography": { - "task": "mmlu_high_school_geography", - "task_alias": "high_school_geography", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_geography", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school geography.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_government_and_politics": { - "task": "mmlu_high_school_government_and_politics", - "task_alias": "high_school_government_and_politics", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_government_and_politics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school government and politics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_macroeconomics": { - "task": "mmlu_high_school_macroeconomics", - "task_alias": "high_school_macroeconomics", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_macroeconomics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school macroeconomics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_mathematics": { - "task": "mmlu_high_school_mathematics", - "task_alias": "high_school_mathematics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_mathematics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school mathematics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_microeconomics": { - "task": "mmlu_high_school_microeconomics", - "task_alias": "high_school_microeconomics", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_microeconomics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school microeconomics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_physics": { - "task": "mmlu_high_school_physics", - "task_alias": "high_school_physics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_physics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school physics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_psychology": { - "task": "mmlu_high_school_psychology", - "task_alias": "high_school_psychology", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_psychology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school psychology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_statistics": { - "task": "mmlu_high_school_statistics", - "task_alias": "high_school_statistics", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_statistics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school statistics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_us_history": { - "task": "mmlu_high_school_us_history", - "task_alias": "high_school_us_history", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_us_history", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school us history.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_high_school_world_history": { - "task": "mmlu_high_school_world_history", - "task_alias": "high_school_world_history", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "high_school_world_history", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about high school world history.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_human_aging": { - "task": "mmlu_human_aging", - "task_alias": "human_aging", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "human_aging", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about human aging.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_human_sexuality": { - "task": "mmlu_human_sexuality", - "task_alias": "human_sexuality", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "human_sexuality", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about human sexuality.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_international_law": { - "task": "mmlu_international_law", - "task_alias": "international_law", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "international_law", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about international law.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_jurisprudence": { - "task": "mmlu_jurisprudence", - "task_alias": "jurisprudence", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "jurisprudence", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about jurisprudence.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_logical_fallacies": { - "task": "mmlu_logical_fallacies", - "task_alias": "logical_fallacies", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "logical_fallacies", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about logical fallacies.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_machine_learning": { - "task": "mmlu_machine_learning", - "task_alias": "machine_learning", - "group": "mmlu_stem", - "group_alias": "stem", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "machine_learning", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about machine learning.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_management": { - "task": "mmlu_management", - "task_alias": "management", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "management", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about management.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_marketing": { - "task": "mmlu_marketing", - "task_alias": "marketing", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "marketing", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about marketing.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_medical_genetics": { - "task": "mmlu_medical_genetics", - "task_alias": "medical_genetics", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "medical_genetics", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about medical genetics.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_miscellaneous": { - "task": "mmlu_miscellaneous", - "task_alias": "miscellaneous", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "miscellaneous", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about miscellaneous.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_moral_disputes": { - "task": "mmlu_moral_disputes", - "task_alias": "moral_disputes", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "moral_disputes", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about moral disputes.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_moral_scenarios": { - "task": "mmlu_moral_scenarios", - "task_alias": "moral_scenarios", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "moral_scenarios", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about moral scenarios.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_nutrition": { - "task": "mmlu_nutrition", - "task_alias": "nutrition", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "nutrition", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about nutrition.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_philosophy": { - "task": "mmlu_philosophy", - "task_alias": "philosophy", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "philosophy", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about philosophy.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_prehistory": { - "task": "mmlu_prehistory", - "task_alias": "prehistory", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "prehistory", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about prehistory.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_professional_accounting": { - "task": "mmlu_professional_accounting", - "task_alias": "professional_accounting", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "professional_accounting", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about professional accounting.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_professional_law": { - "task": "mmlu_professional_law", - "task_alias": "professional_law", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "professional_law", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about professional law.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_professional_medicine": { - "task": "mmlu_professional_medicine", - "task_alias": "professional_medicine", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "professional_medicine", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about professional medicine.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_professional_psychology": { - "task": "mmlu_professional_psychology", - "task_alias": "professional_psychology", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "professional_psychology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about professional psychology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_public_relations": { - "task": "mmlu_public_relations", - "task_alias": "public_relations", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "public_relations", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about public relations.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_security_studies": { - "task": "mmlu_security_studies", - "task_alias": "security_studies", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "security_studies", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about security studies.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_sociology": { - "task": "mmlu_sociology", - "task_alias": "sociology", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "sociology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about sociology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_us_foreign_policy": { - "task": "mmlu_us_foreign_policy", - "task_alias": "us_foreign_policy", - "group": "mmlu_social_sciences", - "group_alias": "social_sciences", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "us_foreign_policy", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about us foreign policy.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_virology": { - "task": "mmlu_virology", - "task_alias": "virology", - "group": "mmlu_other", - "group_alias": "other", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "virology", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about virology.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "mmlu_world_religions": { - "task": "mmlu_world_religions", - "task_alias": "world_religions", - "group": "mmlu_humanities", - "group_alias": "humanities", - "dataset_path": "hails/mmlu_no_train", - "dataset_name": "world_religions", - "test_split": "test", - "fewshot_split": "dev", - "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", - "doc_to_target": "answer", - "doc_to_choice": [ - "A", - "B", - "C", - "D" - ], - "description": "The following are multiple choice questions (with answers) about world religions.\n\n", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "fewshot_config": { - "sampler": "first_n" - }, - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "piqa": { - "task": "piqa", - "dataset_path": "piqa", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "Question: {{goal}}\nAnswer:", - "doc_to_target": "label", - "doc_to_choice": "{{[sol1, sol2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "goal", - "metadata": { - "version": 1.0 - } - }, - "sciq": { - "task": "sciq", - "dataset_path": "sciq", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "{{support.lstrip()}}\nQuestion: {{question}}\nAnswer:", - "doc_to_target": 3, - "doc_to_choice": "{{[distractor1, distractor2, distractor3, correct_answer]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{support}} {{question}}", - "metadata": { - "version": 1.0 - } - }, - "wikitext": { - "task": "wikitext", - "dataset_path": "EleutherAI/wikitext_document_level", - "dataset_name": "wikitext-2-raw-v1", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "def wikitext_detokenizer(doc):\n string = doc[\"page\"]\n # contractions\n string = string.replace(\"s '\", \"s'\")\n string = re.sub(r\"/' [0-9]/\", r\"/'[0-9]/\", string)\n # number separators\n string = string.replace(\" @-@ \", \"-\")\n string = string.replace(\" @,@ \", \",\")\n string = string.replace(\" @.@ \", \".\")\n # punctuation\n string = string.replace(\" : \", \": \")\n string = string.replace(\" ; \", \"; \")\n string = string.replace(\" . \", \". \")\n string = string.replace(\" ! \", \"! \")\n string = string.replace(\" ? \", \"? \")\n string = string.replace(\" , \", \", \")\n # double brackets\n string = re.sub(r\"\\(\\s*([^\\)]*?)\\s*\\)\", r\"(\\1)\", string)\n string = re.sub(r\"\\[\\s*([^\\]]*?)\\s*\\]\", r\"[\\1]\", string)\n string = re.sub(r\"{\\s*([^}]*?)\\s*}\", r\"{\\1}\", string)\n string = re.sub(r\"\\\"\\s*([^\\\"]*?)\\s*\\\"\", r'\"\\1\"', string)\n string = re.sub(r\"'\\s*([^']*?)\\s*'\", r\"'\\1'\", string)\n # miscellaneous\n string = string.replace(\"= = = =\", \"====\")\n string = string.replace(\"= = =\", \"===\")\n string = string.replace(\"= =\", \"==\")\n string = string.replace(\" \" + chr(176) + \" \", chr(176))\n string = string.replace(\" \\n\", \"\\n\")\n string = string.replace(\"\\n \", \"\\n\")\n string = string.replace(\" N \", \" 1 \")\n string = string.replace(\" 's\", \"'s\")\n\n return string\n", - "process_results": "def process_results(doc, results):\n (loglikelihood,) = results\n # IMPORTANT: wikitext counts number of words in *original doc before detokenization*\n _words = len(re.split(r\"\\s+\", doc[\"page\"]))\n _bytes = len(doc[\"page\"].encode(\"utf-8\"))\n return {\n \"word_perplexity\": (loglikelihood, _words),\n \"byte_perplexity\": (loglikelihood, _bytes),\n \"bits_per_byte\": (loglikelihood, _bytes),\n }\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "word_perplexity" - }, - { - "metric": "byte_perplexity" - }, - { - "metric": "bits_per_byte" - } - ], - "output_type": "loglikelihood_rolling", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{page}}", - "metadata": { - "version": 2.0 - } - }, - "winogrande": { - "task": "winogrande", - "dataset_path": "winogrande", - "dataset_name": "winogrande_xl", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "def doc_to_text(doc):\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc):\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc):\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "sentence", - "metadata": { - "version": 1.0 - } - }, - "wsc": { - "task": "wsc", - "group": [ - "super-glue-lm-eval-v1" - ], - "dataset_path": "super_glue", - "dataset_name": "wsc.fixed", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "def default_doc_to_text(x):\n raw_passage = x[\"text\"]\n # NOTE: HuggingFace span indices are word-based not character-based.\n pre = \" \".join(raw_passage.split()[: x[\"span2_index\"]])\n post = raw_passage[len(pre) + len(x[\"span2_text\"]) + 1 :]\n passage = general_detokenize(pre + \" *{}*\".format(x[\"span2_text\"]) + post)\n noun = x[\"span1_text\"]\n pronoun = x[\"span2_text\"]\n text = (\n f\"Passage: {passage}\\n\"\n + f'Question: In the passage above, does the pronoun \"*{pronoun}*\" refer to \"*{noun}*\"?\\n'\n + \"Answer:\"\n )\n return text\n", - "doc_to_target": "label", - "doc_to_choice": [ - "no", - "yes" - ], - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "ai2_arc": "N/A", - "arc_challenge": 1.0, - "arc_easy": 1.0, - "blimp": "N/A", - "blimp_adjunct_island": 1.0, - "blimp_anaphor_gender_agreement": 1.0, - "blimp_anaphor_number_agreement": 1.0, - "blimp_animate_subject_passive": 1.0, - "blimp_animate_subject_trans": 1.0, - "blimp_causative": 1.0, - "blimp_complex_NP_island": 1.0, - "blimp_coordinate_structure_constraint_complex_left_branch": 1.0, - "blimp_coordinate_structure_constraint_object_extraction": 1.0, - "blimp_determiner_noun_agreement_1": 1.0, - "blimp_determiner_noun_agreement_2": 1.0, - "blimp_determiner_noun_agreement_irregular_1": 1.0, - "blimp_determiner_noun_agreement_irregular_2": 1.0, - "blimp_determiner_noun_agreement_with_adj_2": 1.0, - "blimp_determiner_noun_agreement_with_adj_irregular_1": 1.0, - "blimp_determiner_noun_agreement_with_adj_irregular_2": 1.0, - "blimp_determiner_noun_agreement_with_adjective_1": 1.0, - "blimp_distractor_agreement_relational_noun": 1.0, - "blimp_distractor_agreement_relative_clause": 1.0, - "blimp_drop_argument": 1.0, - "blimp_ellipsis_n_bar_1": 1.0, - "blimp_ellipsis_n_bar_2": 1.0, - "blimp_existential_there_object_raising": 1.0, - "blimp_existential_there_quantifiers_1": 1.0, - "blimp_existential_there_quantifiers_2": 1.0, - "blimp_existential_there_subject_raising": 1.0, - "blimp_expletive_it_object_raising": 1.0, - "blimp_inchoative": 1.0, - "blimp_intransitive": 1.0, - "blimp_irregular_past_participle_adjectives": 1.0, - "blimp_irregular_past_participle_verbs": 1.0, - "blimp_irregular_plural_subject_verb_agreement_1": 1.0, - "blimp_irregular_plural_subject_verb_agreement_2": 1.0, - "blimp_left_branch_island_echo_question": 1.0, - "blimp_left_branch_island_simple_question": 1.0, - "blimp_matrix_question_npi_licensor_present": 1.0, - "blimp_npi_present_1": 1.0, - "blimp_npi_present_2": 1.0, - "blimp_only_npi_licensor_present": 1.0, - "blimp_only_npi_scope": 1.0, - "blimp_passive_1": 1.0, - "blimp_passive_2": 1.0, - "blimp_principle_A_c_command": 1.0, - "blimp_principle_A_case_1": 1.0, - "blimp_principle_A_case_2": 1.0, - "blimp_principle_A_domain_1": 1.0, - "blimp_principle_A_domain_2": 1.0, - "blimp_principle_A_domain_3": 1.0, - "blimp_principle_A_reconstruction": 1.0, - "blimp_regular_plural_subject_verb_agreement_1": 1.0, - "blimp_regular_plural_subject_verb_agreement_2": 1.0, - "blimp_sentential_negation_npi_licensor_present": 1.0, - "blimp_sentential_negation_npi_scope": 1.0, - "blimp_sentential_subject_island": 1.0, - "blimp_superlative_quantifiers_1": 1.0, - "blimp_superlative_quantifiers_2": 1.0, - "blimp_tough_vs_raising_1": 1.0, - "blimp_tough_vs_raising_2": 1.0, - "blimp_transitive": 1.0, - "blimp_wh_island": 1.0, - "blimp_wh_questions_object_gap": 1.0, - "blimp_wh_questions_subject_gap": 1.0, - "blimp_wh_questions_subject_gap_long_distance": 1.0, - "blimp_wh_vs_that_no_gap": 1.0, - "blimp_wh_vs_that_no_gap_long_distance": 1.0, - "blimp_wh_vs_that_with_gap": 1.0, - "blimp_wh_vs_that_with_gap_long_distance": 1.0, - "lambada_openai": 1.0, - "logiqa": 1.0, - "mmlu": "N/A", - "mmlu_abstract_algebra": 0.0, - "mmlu_anatomy": 0.0, - "mmlu_astronomy": 0.0, - "mmlu_business_ethics": 0.0, - "mmlu_clinical_knowledge": 0.0, - "mmlu_college_biology": 0.0, - "mmlu_college_chemistry": 0.0, - "mmlu_college_computer_science": 0.0, - "mmlu_college_mathematics": 0.0, - "mmlu_college_medicine": 0.0, - "mmlu_college_physics": 0.0, - "mmlu_computer_security": 0.0, - "mmlu_conceptual_physics": 0.0, - "mmlu_econometrics": 0.0, - "mmlu_electrical_engineering": 0.0, - "mmlu_elementary_mathematics": 0.0, - "mmlu_formal_logic": 0.0, - "mmlu_global_facts": 0.0, - "mmlu_high_school_biology": 0.0, - "mmlu_high_school_chemistry": 0.0, - "mmlu_high_school_computer_science": 0.0, - "mmlu_high_school_european_history": 0.0, - "mmlu_high_school_geography": 0.0, - "mmlu_high_school_government_and_politics": 0.0, - "mmlu_high_school_macroeconomics": 0.0, - "mmlu_high_school_mathematics": 0.0, - "mmlu_high_school_microeconomics": 0.0, - "mmlu_high_school_physics": 0.0, - "mmlu_high_school_psychology": 0.0, - "mmlu_high_school_statistics": 0.0, - "mmlu_high_school_us_history": 0.0, - "mmlu_high_school_world_history": 0.0, - "mmlu_human_aging": 0.0, - "mmlu_human_sexuality": 0.0, - "mmlu_humanities": "N/A", - "mmlu_international_law": 0.0, - "mmlu_jurisprudence": 0.0, - "mmlu_logical_fallacies": 0.0, - "mmlu_machine_learning": 0.0, - "mmlu_management": 0.0, - "mmlu_marketing": 0.0, - "mmlu_medical_genetics": 0.0, - "mmlu_miscellaneous": 0.0, - "mmlu_moral_disputes": 0.0, - "mmlu_moral_scenarios": 0.0, - "mmlu_nutrition": 0.0, - "mmlu_other": "N/A", - "mmlu_philosophy": 0.0, - "mmlu_prehistory": 0.0, - "mmlu_professional_accounting": 0.0, - "mmlu_professional_law": 0.0, - "mmlu_professional_medicine": 0.0, - "mmlu_professional_psychology": 0.0, - "mmlu_public_relations": 0.0, - "mmlu_security_studies": 0.0, - "mmlu_social_sciences": "N/A", - "mmlu_sociology": 0.0, - "mmlu_stem": "N/A", - "mmlu_us_foreign_policy": 0.0, - "mmlu_virology": 0.0, - "mmlu_world_religions": 0.0, - "piqa": 1.0, - "pythia": "N/A", - "sciq": 1.0, - "wikitext": 2.0, - "winogrande": 1.0, - "wsc": 1.0 - }, - "n-shot": { - "ai2_arc": 0, - "arc_challenge": 0, - "arc_easy": 0, - "blimp": 0, - "blimp_adjunct_island": 0, - "blimp_anaphor_gender_agreement": 0, - "blimp_anaphor_number_agreement": 0, - "blimp_animate_subject_passive": 0, - "blimp_animate_subject_trans": 0, - "blimp_causative": 0, - "blimp_complex_NP_island": 0, - "blimp_coordinate_structure_constraint_complex_left_branch": 0, - "blimp_coordinate_structure_constraint_object_extraction": 0, - "blimp_determiner_noun_agreement_1": 0, - "blimp_determiner_noun_agreement_2": 0, - "blimp_determiner_noun_agreement_irregular_1": 0, - "blimp_determiner_noun_agreement_irregular_2": 0, - "blimp_determiner_noun_agreement_with_adj_2": 0, - "blimp_determiner_noun_agreement_with_adj_irregular_1": 0, - "blimp_determiner_noun_agreement_with_adj_irregular_2": 0, - "blimp_determiner_noun_agreement_with_adjective_1": 0, - "blimp_distractor_agreement_relational_noun": 0, - "blimp_distractor_agreement_relative_clause": 0, - "blimp_drop_argument": 0, - "blimp_ellipsis_n_bar_1": 0, - "blimp_ellipsis_n_bar_2": 0, - "blimp_existential_there_object_raising": 0, - "blimp_existential_there_quantifiers_1": 0, - "blimp_existential_there_quantifiers_2": 0, - "blimp_existential_there_subject_raising": 0, - "blimp_expletive_it_object_raising": 0, - "blimp_inchoative": 0, - "blimp_intransitive": 0, - "blimp_irregular_past_participle_adjectives": 0, - "blimp_irregular_past_participle_verbs": 0, - "blimp_irregular_plural_subject_verb_agreement_1": 0, - "blimp_irregular_plural_subject_verb_agreement_2": 0, - "blimp_left_branch_island_echo_question": 0, - "blimp_left_branch_island_simple_question": 0, - "blimp_matrix_question_npi_licensor_present": 0, - "blimp_npi_present_1": 0, - "blimp_npi_present_2": 0, - "blimp_only_npi_licensor_present": 0, - "blimp_only_npi_scope": 0, - "blimp_passive_1": 0, - "blimp_passive_2": 0, - "blimp_principle_A_c_command": 0, - "blimp_principle_A_case_1": 0, - "blimp_principle_A_case_2": 0, - "blimp_principle_A_domain_1": 0, - "blimp_principle_A_domain_2": 0, - "blimp_principle_A_domain_3": 0, - "blimp_principle_A_reconstruction": 0, - "blimp_regular_plural_subject_verb_agreement_1": 0, - "blimp_regular_plural_subject_verb_agreement_2": 0, - "blimp_sentential_negation_npi_licensor_present": 0, - "blimp_sentential_negation_npi_scope": 0, - "blimp_sentential_subject_island": 0, - "blimp_superlative_quantifiers_1": 0, - "blimp_superlative_quantifiers_2": 0, - "blimp_tough_vs_raising_1": 0, - "blimp_tough_vs_raising_2": 0, - "blimp_transitive": 0, - "blimp_wh_island": 0, - "blimp_wh_questions_object_gap": 0, - "blimp_wh_questions_subject_gap": 0, - "blimp_wh_questions_subject_gap_long_distance": 0, - "blimp_wh_vs_that_no_gap": 0, - "blimp_wh_vs_that_no_gap_long_distance": 0, - "blimp_wh_vs_that_with_gap": 0, - "blimp_wh_vs_that_with_gap_long_distance": 0, - "lambada_openai": 0, - "logiqa": 0, - "mmlu": 0, - "mmlu_abstract_algebra": 0, - "mmlu_anatomy": 0, - "mmlu_astronomy": 0, - "mmlu_business_ethics": 0, - "mmlu_clinical_knowledge": 0, - "mmlu_college_biology": 0, - "mmlu_college_chemistry": 0, - "mmlu_college_computer_science": 0, - "mmlu_college_mathematics": 0, - "mmlu_college_medicine": 0, - "mmlu_college_physics": 0, - "mmlu_computer_security": 0, - "mmlu_conceptual_physics": 0, - "mmlu_econometrics": 0, - "mmlu_electrical_engineering": 0, - "mmlu_elementary_mathematics": 0, - "mmlu_formal_logic": 0, - "mmlu_global_facts": 0, - "mmlu_high_school_biology": 0, - "mmlu_high_school_chemistry": 0, - "mmlu_high_school_computer_science": 0, - "mmlu_high_school_european_history": 0, - "mmlu_high_school_geography": 0, - "mmlu_high_school_government_and_politics": 0, - "mmlu_high_school_macroeconomics": 0, - "mmlu_high_school_mathematics": 0, - "mmlu_high_school_microeconomics": 0, - "mmlu_high_school_physics": 0, - "mmlu_high_school_psychology": 0, - "mmlu_high_school_statistics": 0, - "mmlu_high_school_us_history": 0, - "mmlu_high_school_world_history": 0, - "mmlu_human_aging": 0, - "mmlu_human_sexuality": 0, - "mmlu_humanities": 0, - "mmlu_international_law": 0, - "mmlu_jurisprudence": 0, - "mmlu_logical_fallacies": 0, - "mmlu_machine_learning": 0, - "mmlu_management": 0, - "mmlu_marketing": 0, - "mmlu_medical_genetics": 0, - "mmlu_miscellaneous": 0, - "mmlu_moral_disputes": 0, - "mmlu_moral_scenarios": 0, - "mmlu_nutrition": 0, - "mmlu_other": 0, - "mmlu_philosophy": 0, - "mmlu_prehistory": 0, - "mmlu_professional_accounting": 0, - "mmlu_professional_law": 0, - "mmlu_professional_medicine": 0, - "mmlu_professional_psychology": 0, - "mmlu_public_relations": 0, - "mmlu_security_studies": 0, - "mmlu_social_sciences": 0, - "mmlu_sociology": 0, - "mmlu_stem": 0, - "mmlu_us_foreign_policy": 0, - "mmlu_virology": 0, - "mmlu_world_religions": 0, - "piqa": 0, - "pythia": 0, - "sciq": 0, - "wikitext": 0, - "winogrande": 0, - "wsc": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index b9d580fa1952e7ce5273a541b964ab19f66a9c1b..2d4e37f654054268e23a61675fc071e3d6b82927 100644 --- a/lm-eval-output/allenai/OLMo-7B/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:cf6116e4155573804413d89e826dac96298103a1a8b34fa950ddabea1befca99 -size 532127 +oid sha256:b0eceaf74af790d829ccfc115ddd72942318903111611de359f037017346de28 +size 325295 diff --git a/lm-eval-output/allenai/OLMo-7B/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index b7f6fd64d2a0ec2285a2e2e96f76ad0e1c09e70b..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:c4326bbaa5b62395074166a05e8f1dd1481910896d7c76eef5f22341fb3e01a5 -size 2029815 diff --git a/lm-eval-output/allenai/OLMo-7B/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index a6c567e2da934f90c458d1d47e4362968898a5c8..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,169 +0,0 @@ -{ - "results": { - "qa4mre": { - "acc,none": 0.39184397163120566, - "acc_stderr,none": 0.04732646258224104, - "acc_norm,none": 0.4379432624113475, - "acc_norm_stderr,none": 0.04888958862965728, - "alias": "qa4mre" - }, - "qa4mre_2011": { - "acc,none": 0.4666666666666667, - "acc_stderr,none": 0.0457329560380023, - "acc_norm,none": 0.55, - "acc_norm_stderr,none": 0.04560517440787951, - "alias": " - qa4mre_2011" - }, - "qa4mre_2012": { - "acc,none": 0.3125, - "acc_stderr,none": 0.03675892481369823, - "acc_norm,none": 0.44375, - "acc_norm_stderr,none": 0.039400853796259426, - "alias": " - qa4mre_2012" - }, - "qa4mre_2013": { - "acc,none": 0.40492957746478875, - "acc_stderr,none": 0.029179692752203355, - "acc_norm,none": 0.3873239436619718, - "acc_norm_stderr,none": 0.02895738957595096, - "alias": " - qa4mre_2013" - } - }, - "groups": { - "qa4mre": { - "acc,none": 0.39184397163120566, - "acc_stderr,none": 0.04732646258224104, - "acc_norm,none": 0.4379432624113475, - "acc_norm_stderr,none": 0.04888958862965728, - "alias": "qa4mre" - } - }, - "configs": { - "qa4mre_2011": { - "task": "qa4mre_2011", - "group": [ - "qa4mre" - ], - "dataset_path": "qa4mre", - "dataset_name": "2011.main.EN", - "test_split": "train", - "doc_to_text": "{{document_str.strip()}}\nQuestion: {{question_str}}\nAnswer:", - "doc_to_target": "{{correct_answer_id|int - 1}}", - "doc_to_choice": "{{answer_options.answer_str}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{document_str.strip()}} + ' ' + {{question_str}}", - "metadata": { - "version": 1.0 - } - }, - "qa4mre_2012": { - "task": "qa4mre_2012", - "group": [ - "qa4mre" - ], - "dataset_path": "qa4mre", - "dataset_name": "2012.main.EN", - "test_split": "train", - "doc_to_text": "{{document_str.strip()}}\nQuestion: {{question_str}}\nAnswer:", - "doc_to_target": "{{correct_answer_id|int - 1}}", - "doc_to_choice": "{{answer_options.answer_str}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{document_str.strip()}} + ' ' + {{question_str}}", - "metadata": { - "version": 1.0 - } - }, - "qa4mre_2013": { - "task": "qa4mre_2013", - "group": [ - "qa4mre" - ], - "dataset_path": "qa4mre", - "dataset_name": "2013.main.EN", - "test_split": "train", - "doc_to_text": "{{document_str.strip()}}\nQuestion: {{question_str}}\nAnswer:", - "doc_to_target": "{{correct_answer_id|int - 1}}", - "doc_to_choice": "{{answer_options.answer_str}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - }, - { - "metric": "acc_norm", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{document_str.strip()}} + ' ' + {{question_str}}", - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "qa4mre": "N/A", - "qa4mre_2011": 1.0, - "qa4mre_2012": 1.0, - "qa4mre_2013": 1.0 - }, - "n-shot": { - "qa4mre": 0, - "qa4mre_2011": 0, - "qa4mre_2012": 0, - "qa4mre_2013": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 032311ea0a95064147d278474adf69cf0f69b868..eeca7d108e48424b054e4d4a6e7042ab1d7cd6a7 100644 --- a/lm-eval-output/allenai/OLMo-7B/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:fa65988594c0f34884f3c4176b749d7e617fb5af47f1e8662248f11ef06fd864 -size 31349 +oid sha256:37d33c0989c8264c329429e95ac6b542d28d5ee8b481acafc4d074f9fb1cadfe +size 60635 diff --git a/lm-eval-output/allenai/OLMo-7B/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 47a1f2065e1e52bbe239d773997f263b7337247b..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:8d03603595720bed4efdff9e74832e687886ac8c55c3c8b7da8322db4992d957 -size 876136 diff --git a/lm-eval-output/allenai/OLMo-7B/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 5d3613e376b2177f8cda827fb0db6c93f200cfb2..8071918cc2b463d83dbf6ce5764abf87d49de0a8 100644 --- a/lm-eval-output/allenai/OLMo-7B/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "qnli": { - "acc,none": 0.4964305326743548, - "acc_stderr,none": 0.006765238152075668, + "acc,none": 0.49679663188724144, + "acc_stderr,none": 0.006765271702920652, "alias": "qnli" } }, @@ -45,13 +45,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 16 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 426227b1ed7d0ea74937cc48629c504f2ff50019..d5dff4d49e00d9e09f28ec9acc43a01eed65afac 100644 --- a/lm-eval-output/allenai/OLMo-7B/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:c93698f60583e4fbd78aa584588da8561d78d4f9cfa80682b74a54d87d7d1feb -size 21305 +oid sha256:d1c4843236c77f1cf0e1d9fd2971e32357b6567e61f9933ed90d17d428764e9d +size 19215 diff --git a/lm-eval-output/allenai/OLMo-7B/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 3703fa6cb0814adbba42d5fd452329b6bdd76960..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:74d87205498e3d960f17a2e4f1166909b61a52e3c801baeb3ecd3755f35704a7 -size 4019065 diff --git a/lm-eval-output/allenai/OLMo-7B/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 6971a87d51999261a0fd14e4949c6ba3f58162af..9c47db2fe3fc76d380ad4dcd9c6ca842dab80e5d 100644 --- a/lm-eval-output/allenai/OLMo-7B/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,10 +1,10 @@ { "results": { "qqp": { - "acc,none": 0.5348998268612416, - "acc_stderr,none": 0.0024806356431606442, - "f1,none": 0.44540789240842327, - "f1_stderr,none": 0.0033657259190706915, + "acc,none": 0.5342567400445214, + "acc_stderr,none": 0.00248085735947474, + "f1,none": 0.44154457559760363, + "f1_stderr,none": 0.003374377995053171, "alias": "qqp" } }, @@ -50,13 +50,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 32 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 9d93fc5433ef76d718c3433e51fe0d6d8f667aaf..07d17ddd608cf26e9df6e0cfb477a4630ee5e9ef 100644 --- a/lm-eval-output/allenai/OLMo-7B/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:d7eb18e695ad7f6ca5ba1ebdeac2ca9aca2418e4837c0f0a54880225a52ccc62 -size 82735 +oid sha256:8345584a7d4110c3a91fcd2ecafcad46903876f6bceb1ada9d4b772e8bf35cee +size 38657 diff --git a/lm-eval-output/allenai/OLMo-7B/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index a0a201533d94fc5021ae667e2cc115e75bf67b49..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:60bb5b56c894712963c63542e0e5d3c12fd2c7bf27199b011358179dfa780668 -size 1291523 diff --git a/lm-eval-output/allenai/OLMo-7B/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index d4a0903239d29aa9184a51931425e0664e08e4c8..febfa42086687536745353ddb6fd317ca295b4fc 100644 --- a/lm-eval-output/allenai/OLMo-7B/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "race": { - "acc,none": 0.38564593301435407, - "acc_stderr,none": 0.015064457528125815, + "acc,none": 0.3837320574162679, + "acc_stderr,none": 0.015050418634703647, "alias": "race" } }, @@ -42,13 +42,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 4 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 32c287c83d9c5e6faafc5ebddf5fe9d6f1149bba..f903058427506fefed895b0b7881f8315e1bee8f 100644 --- a/lm-eval-output/allenai/OLMo-7B/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:ffa16782f3a2e569457c19a885dcdb31d6c603b406e6eb540a3f873b71f83724 -size 21868 +oid sha256:d09287b3dae2197e3845d605a6b6047ec51baa9f891f2c93c596a5fcc84312e2 +size 99905 diff --git a/lm-eval-output/allenai/OLMo-7B/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 869bc9d1042e83cfba7d530cc27d13afab958d0e..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:09733c95d59bc0ee490c5f18a81f89634aa9004723cc8da3b0744c9407d60d57 -size 57822 diff --git a/lm-eval-output/allenai/OLMo-7B/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 00ccf85c4a64013ca8ce90581dfe982ed518c2b1..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,57 +0,0 @@ -{ - "results": { - "rte": { - "acc,none": 0.5342960288808665, - "acc_stderr,none": 0.030025579819366422, - "alias": "rte" - } - }, - "configs": { - "rte": { - "task": "rte", - "group": "glue", - "dataset_path": "glue", - "dataset_name": "rte", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "{{sentence1}}\nQuestion: {{sentence2}} True or False?\nAnswer:", - "doc_to_target": "label", - "doc_to_choice": [ - "True", - "False" - ], - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "rte": 1.0 - }, - "n-shot": { - "rte": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index fa287fa80251b60a3e8ea9bd0c773a483af1f50f..f711affc6732d353ee2be996901a3af6880fdfde 100644 --- a/lm-eval-output/allenai/OLMo-7B/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:c82f1445d6c9aadd5674588246cfeca975e20d410cea231d50c35a835d284e28 -size 12534 +oid sha256:01cbeae2069ced113f14451fe8919b07d24316ade8803ee4376595b8f003cba5 +size 20388 diff --git a/lm-eval-output/allenai/OLMo-7B/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 824573a2f140995ab4044d7443bde054e8e252aa..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:94c80e4543063838b682a526047666f8d5479373ccdd17ff4e641b18764d8688 -size 332195 diff --git a/lm-eval-output/allenai/OLMo-7B/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index ae3e02bf932b27d2fbeff09661e305f35e90a710..4bbcdd492cab0cfc44bc7c54a362ef250ac13ee1 100644 --- a/lm-eval-output/allenai/OLMo-7B/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,10 +1,10 @@ { "results": { "sciq": { - "acc,none": 0.929, - "acc_stderr,none": 0.008125578442487923, - "acc_norm,none": 0.884, - "acc_norm_stderr,none": 0.010131468138756988, + "acc,none": 0.926, + "acc_stderr,none": 0.00828206451270416, + "acc_norm,none": 0.885, + "acc_norm_stderr,none": 0.010093407594904619, "alias": "sciq" } }, @@ -51,13 +51,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 8 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 344e12d84c7a9ebc5a8dbbde8d6e03f1f8a9c0d4..3c150d5aaf310754cf59da43e1ff3be23ed0b454 100644 --- a/lm-eval-output/allenai/OLMo-7B/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:ed6e77edd908f5f11de9c7980199e875c97143eabdba9583343be0abb37f8312 -size 13802 +oid sha256:ebdc2a5a2582d6f4cd883c2b791cb7325863594a244455d8ad26240085237cf9 +size 15716 diff --git a/lm-eval-output/allenai/OLMo-7B/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 03d4ab4a2e3e32e01efeb30f4c3ddca6cc633332..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:2e56f448f51e9ce4f19a9b64e514c683fb69a95504fe83552d92fb3673ea1120 -size 57650 diff --git a/lm-eval-output/allenai/OLMo-7B/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 5c741d3e3a8fef1838bfac1075b074a3f427e4cb..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,59 +0,0 @@ -{ - "results": { - "sglue_rte": { - "acc,none": 0.5342960288808665, - "acc_stderr,none": 0.030025579819366422, - "alias": "sglue_rte" - } - }, - "configs": { - "sglue_rte": { - "task": "sglue_rte", - "group": [ - "super-glue-lm-eval-v1" - ], - "dataset_path": "super_glue", - "dataset_name": "rte", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "{{premise}}\nQuestion: {{hypothesis}} True or False?\nAnswer:", - "doc_to_target": "label", - "doc_to_choice": [ - "True", - "False" - ], - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - } - }, - "versions": { - "sglue_rte": 0.0 - }, - "n-shot": { - "sglue_rte": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 15e76c9479af5d0dc55a689dc630a6302d03805c..0ea8393b46590c9e7a48d6e775f50e2f56230c76 100644 --- a/lm-eval-output/allenai/OLMo-7B/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:bc53170c706d7e296f0251721aee11637bb6f391933085130e9545ca86ba0468 -size 12624 +oid sha256:b8a4f1c78117f3ce83c715666126967217cde444873a514aecc0cf755f095e64 +size 20532 diff --git a/lm-eval-output/allenai/OLMo-7B/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index e898b99a7753d14d3bacc8e0b08c63d753dc67f1..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:07d2a09179166d19d78183cfd721a35d7e390e2a430d1d2b29f0811a3f860961 -size 79755 diff --git a/lm-eval-output/allenai/OLMo-7B/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index d1a2bbb85b7e507ffddcb2027a7ed18ebbe78c08..de34f87b3edcc6a2f003d90e8387cdcb6dce3835 100644 --- a/lm-eval-output/allenai/OLMo-7B/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "sst2": { - "acc,none": 0.5642201834862385, - "acc_stderr,none": 0.016801528278889092, + "acc,none": 0.5722477064220184, + "acc_stderr,none": 0.016764056901835654, "alias": "sst2" } }, @@ -45,13 +45,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 79d0c63ad402cc393d5ea0d88e9535db0241ac18..177e1084dec7e4eb2fac7a73896e189681cf4dcd 100644 --- a/lm-eval-output/allenai/OLMo-7B/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:2e031723d27b68213b081f8794593ac5a82d8c28306bde7828f8320f464403bd -size 13424 +oid sha256:d9f30d2052ec9fbd57589b502bdc501e8713a02d98e29680d36b192582746e88 +size 13080 diff --git a/lm-eval-output/allenai/OLMo-7B/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 6cf4216dd986088b69fa0cbf0ccf051c31396f42..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:221c04e660c191a58526cfdd779670104dbc5b77e45666db1737d7bba1fdfcea -size 4679226 diff --git a/lm-eval-output/allenai/OLMo-7B/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 7d91b87e48af7fb34e9dc35fdbe698d07b4cc29f..65de50117e0c811eabdfe1618226976e69fd9248 100644 --- a/lm-eval-output/allenai/OLMo-7B/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,10 +1,10 @@ { "results": { "swag": { - "acc,none": 0.5506348095571328, - "acc_stderr,none": 0.0035169181998806623, - "acc_norm,none": 0.7490752774167749, - "acc_norm_stderr,none": 0.0030652449131181743, + "acc,none": 0.5507847645706289, + "acc_stderr,none": 0.0035168101165724205, + "acc_norm,none": 0.7488753373987803, + "acc_norm_stderr,none": 0.0030660566116668415, "alias": "swag" } }, @@ -50,13 +50,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 32 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 7e9e51b8a9beefa34d620339f14f1acce700d253..15e39dfa79a66a05db58d100563865743cc34efe 100644 --- a/lm-eval-output/allenai/OLMo-7B/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:859c314a5e98c7478ee536c9e8f00691d86bcf32c56f74e66686b62349646852 -size 73825 +oid sha256:5579a4142e7385468c915b3999e1784dd45e01914f1b60f7c1b4cd63b8aacd8d +size 29086 diff --git a/lm-eval-output/allenai/OLMo-7B/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index be7e4f9a5af5018480d26731195aaebc5e5742cb..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e4772e50d0ed80a8b00f111cfedffae18fa7f0ca16baadaaf91db8ddc1679e15 -size 5622526 diff --git a/lm-eval-output/allenai/OLMo-7B/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index d6d9c41f39c6b942452c16639bdcf07a4956d203..144c952582f596066faa3f3d825282d11aae91f9 100644 --- a/lm-eval-output/allenai/OLMo-7B/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,30 +1,30 @@ { "results": { "sycophancy": { - "acc,none": 0.5715616784799175, - "acc_stderr,none": 0.026698982414548357, + "acc,none": 0.5723935975508303, + "acc_stderr,none": 0.027356716501895306, "alias": "sycophancy" }, "sycophancy_on_nlp_survey": { - "acc,none": 0.6025641025641025, - "acc_stderr,none": 0.004897840495269369, + "acc,none": 0.6052684294871795, + "acc_stderr,none": 0.004892089645048989, "alias": " - sycophancy_on_nlp_survey" }, "sycophancy_on_philpapers2020": { - "acc,none": 0.6095064355933921, - "acc_stderr,none": 0.004911628747734351, + "acc,none": 0.6099118273031316, + "acc_stderr,none": 0.004910710855746093, "alias": " - sycophancy_on_philpapers2020" }, "sycophancy_on_political_typology_quiz": { - "acc,none": 0.5045098039215686, - "acc_stderr,none": 0.0049507790224931686, + "acc,none": 0.503921568627451, + "acc_stderr,none": 0.004950828134318425, "alias": " - sycophancy_on_political_typology_quiz" } }, "groups": { "sycophancy": { - "acc,none": 0.5715616784799175, - "acc_stderr,none": 0.026698982414548357, + "acc,none": 0.5723935975508303, + "acc_stderr,none": 0.027356716501895306, "alias": "sycophancy" } }, @@ -117,13 +117,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 16 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index a3621e6a9745a174e9a07747fd4a57e5ed980a0b..ccc091e736d8a73d3b2c42de4eea4ef2ec46328c 100644 --- a/lm-eval-output/allenai/OLMo-7B/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:abd378f325d5736637425be7f8695f377a4d8da6c3fa47d0253943c7c3a522bf -size 100858 +oid sha256:ff4124f0af62a4c3a1b9cbe6750e5daa3ff24039616318f4acea93f961881e88 +size 50082 diff --git a/lm-eval-output/allenai/OLMo-7B/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 9d98c7570fe21702d0b0d96331fc2dc4d77794ae..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:bb763cd8a27efe6fa6ebce22f1809dc8e559cfb6186120318342ca7b59eebf89 -size 705697 diff --git a/lm-eval-output/allenai/OLMo-7B/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 2c8ca7bc185e94f9b5e47a3ed3ae9c240528c0f9..007dfa9f15bcfa24ef4de91ab4577fd3c49bff6d 100644 --- a/lm-eval-output/allenai/OLMo-7B/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,100 +1,100 @@ { "results": { "truthfulqa": { - "acc,none": 0.32027712412406617, - "acc_stderr,none": 0.038859715675549834, - "bleu_max,none": 26.605743093764914, - "bleu_max_stderr,none": 0.6015382486011928, - "bleu_acc,none": 0.29253365973072215, - "bleu_acc_stderr,none": 0.00025362465398930524, - "bleu_diff,none": -9.25287150767902, - "bleu_diff_stderr,none": 0.612111392397556, - "rouge1_max,none": 52.114007973758504, - "rouge1_max_stderr,none": 0.6818812598428838, - "rouge1_acc,none": 0.31211750305997554, - "rouge1_acc_stderr,none": 0.0002631129501759587, - "rouge1_diff,none": -11.370354328482358, - "rouge1_diff_stderr,none": 0.6679458035889521, - "rouge2_max,none": 36.595331993943475, - "rouge2_max_stderr,none": 0.929975566251503, - "rouge2_acc,none": 0.2594859241126071, - "rouge2_acc_stderr,none": 0.00023548159227945327, - "rouge2_diff,none": -13.634662065905777, - "rouge2_diff_stderr,none": 1.022565108392841, - "rougeL_max,none": 49.29481530743976, - "rougeL_max_stderr,none": 0.69003639220024, - "rougeL_acc,none": 0.29498164014687883, - "rougeL_acc_stderr,none": 0.00025486209819011984, - "rougeL_diff,none": -11.50697421848847, - "rougeL_diff_stderr,none": 0.6640055563327908, + "acc,none": 0.3204471732572733, + "acc_stderr,none": 0.03839948737352693, + "bleu_max,none": 26.62777876314072, + "bleu_max_stderr,none": 0.6017893249734863, + "bleu_acc,none": 0.2937576499388005, + "bleu_acc_stderr,none": 0.0002542452120603388, + "bleu_diff,none": -9.137527769828738, + "bleu_diff_stderr,none": 0.6078938282103232, + "rouge1_max,none": 52.13218849000032, + "rouge1_max_stderr,none": 0.6832478569917924, + "rouge1_acc,none": 0.31456548347613217, + "rouge1_acc_stderr,none": 0.0002642328922568159, + "rouge1_diff,none": -11.170553637141843, + "rouge1_diff_stderr,none": 0.6677109580586631, + "rouge2_max,none": 36.62953000553089, + "rouge2_max_stderr,none": 0.9274488375324298, + "rouge2_acc,none": 0.2582619339045288, + "rouge2_acc_stderr,none": 0.000234758219853459, + "rouge2_diff,none": -13.513217455900138, + "rouge2_diff_stderr,none": 1.0181637360872702, + "rougeL_max,none": 49.30854621277883, + "rougeL_max_stderr,none": 0.6902553099803951, + "rougeL_acc,none": 0.2974296205630355, + "rougeL_acc_stderr,none": 0.0002560848546259373, + "rougeL_diff,none": -11.351503523939826, + "rougeL_diff_stderr,none": 0.6610384159635325, "alias": "truthfulqa" }, "truthfulqa_gen": { - "bleu_max,none": 26.605743093764914, - "bleu_max_stderr,none": 0.7755889688496045, - "bleu_acc,none": 0.29253365973072215, - "bleu_acc_stderr,none": 0.015925597445286165, - "bleu_diff,none": -9.25287150767902, - "bleu_diff_stderr,none": 0.782375480442451, - "rouge1_max,none": 52.114007973758504, - "rouge1_max_stderr,none": 0.8257610186021642, - "rouge1_acc,none": 0.31211750305997554, - "rouge1_acc_stderr,none": 0.01622075676952092, - "rouge1_diff,none": -11.370354328482358, - "rouge1_diff_stderr,none": 0.8172795137460329, - "rouge2_max,none": 36.595331993943475, - "rouge2_max_stderr,none": 0.96435240770763, - "rouge2_acc,none": 0.2594859241126071, - "rouge2_acc_stderr,none": 0.015345409485557994, - "rouge2_diff,none": -13.634662065905777, - "rouge2_diff_stderr,none": 1.011219614323635, - "rougeL_max,none": 49.29481530743976, - "rougeL_max_stderr,none": 0.8306842915333358, - "rougeL_acc,none": 0.29498164014687883, - "rougeL_acc_stderr,none": 0.015964400965589653, - "rougeL_diff,none": -11.50697421848847, - "rougeL_diff_stderr,none": 0.8148653608620204, + "bleu_max,none": 26.62777876314072, + "bleu_max_stderr,none": 0.7757508137111339, + "bleu_acc,none": 0.2937576499388005, + "bleu_acc_stderr,none": 0.015945068581236607, + "bleu_diff,none": -9.137527769828738, + "bleu_diff_stderr,none": 0.779675463388661, + "rouge1_max,none": 52.13218849000032, + "rouge1_max_stderr,none": 0.8265880818108814, + "rouge1_acc,none": 0.31456548347613217, + "rouge1_acc_stderr,none": 0.01625524199317918, + "rouge1_diff,none": -11.170553637141843, + "rouge1_diff_stderr,none": 0.8171358259546959, + "rouge2_max,none": 36.62953000553089, + "rouge2_max_stderr,none": 0.9630414516169228, + "rouge2_acc,none": 0.2582619339045288, + "rouge2_acc_stderr,none": 0.015321821688476178, + "rouge2_diff,none": -13.513217455900138, + "rouge2_diff_stderr,none": 1.009040998219235, + "rougeL_max,none": 49.30854621277883, + "rougeL_max_stderr,none": 0.8308160506275712, + "rougeL_acc,none": 0.2974296205630355, + "rougeL_acc_stderr,none": 0.016002651487361002, + "rougeL_diff,none": -11.351503523939826, + "rougeL_diff_stderr,none": 0.813042690123669, "alias": " - truthfulqa_gen" }, "truthfulqa_mc1": { - "acc,none": 0.24357405140758873, - "acc_stderr,none": 0.01502635482491078, + "acc,none": 0.24479804161566707, + "acc_stderr,none": 0.015051869486714994, "alias": " - truthfulqa_mc1" }, "truthfulqa_mc2": { - "acc,none": 0.35862866048230485, - "acc_stderr,none": 0.013799662858469295, + "acc,none": 0.3582717390780764, + "acc_stderr,none": 0.013793375542640339, "alias": " - truthfulqa_mc2" } }, "groups": { "truthfulqa": { - "acc,none": 0.32027712412406617, - "acc_stderr,none": 0.038859715675549834, - "bleu_max,none": 26.605743093764914, - "bleu_max_stderr,none": 0.6015382486011928, - "bleu_acc,none": 0.29253365973072215, - "bleu_acc_stderr,none": 0.00025362465398930524, - "bleu_diff,none": -9.25287150767902, - "bleu_diff_stderr,none": 0.612111392397556, - "rouge1_max,none": 52.114007973758504, - "rouge1_max_stderr,none": 0.6818812598428838, - "rouge1_acc,none": 0.31211750305997554, - "rouge1_acc_stderr,none": 0.0002631129501759587, - "rouge1_diff,none": -11.370354328482358, - "rouge1_diff_stderr,none": 0.6679458035889521, - "rouge2_max,none": 36.595331993943475, - "rouge2_max_stderr,none": 0.929975566251503, - "rouge2_acc,none": 0.2594859241126071, - "rouge2_acc_stderr,none": 0.00023548159227945327, - "rouge2_diff,none": -13.634662065905777, - "rouge2_diff_stderr,none": 1.022565108392841, - "rougeL_max,none": 49.29481530743976, - "rougeL_max_stderr,none": 0.69003639220024, - "rougeL_acc,none": 0.29498164014687883, - "rougeL_acc_stderr,none": 0.00025486209819011984, - "rougeL_diff,none": -11.50697421848847, - "rougeL_diff_stderr,none": 0.6640055563327908, + "acc,none": 0.3204471732572733, + "acc_stderr,none": 0.03839948737352693, + "bleu_max,none": 26.62777876314072, + "bleu_max_stderr,none": 0.6017893249734863, + "bleu_acc,none": 0.2937576499388005, + "bleu_acc_stderr,none": 0.0002542452120603388, + "bleu_diff,none": -9.137527769828738, + "bleu_diff_stderr,none": 0.6078938282103232, + "rouge1_max,none": 52.13218849000032, + "rouge1_max_stderr,none": 0.6832478569917924, + "rouge1_acc,none": 0.31456548347613217, + "rouge1_acc_stderr,none": 0.0002642328922568159, + "rouge1_diff,none": -11.170553637141843, + "rouge1_diff_stderr,none": 0.6677109580586631, + "rouge2_max,none": 36.62953000553089, + "rouge2_max_stderr,none": 0.9274488375324298, + "rouge2_acc,none": 0.2582619339045288, + "rouge2_acc_stderr,none": 0.000234758219853459, + "rouge2_diff,none": -13.513217455900138, + "rouge2_diff_stderr,none": 1.0181637360872702, + "rougeL_max,none": 49.30854621277883, + "rougeL_max_stderr,none": 0.6902553099803951, + "rougeL_acc,none": 0.2974296205630355, + "rougeL_acc_stderr,none": 0.0002560848546259373, + "rougeL_diff,none": -11.351503523939826, + "rougeL_diff_stderr,none": 0.6610384159635325, "alias": "truthfulqa" } }, @@ -268,13 +268,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 32 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 45620b3f714bdbb9935b528551d5343f1ba6db4e..8c217ec97b25ef8ae8b1ec0e4ee8f593d3ea586e 100644 --- a/lm-eval-output/allenai/OLMo-7B/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:9b2a0027fcbc15193df92e0d950e9e44fd4a023b45893f27b466da67b29f7268 -size 552973 +oid sha256:3cdd673b5f30fa91347fd6659cbed813cd598e41c7d346397d2f59e7a6757095 +size 541738 diff --git a/lm-eval-output/allenai/OLMo-7B/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 040a6b831b0d588d4d13e5699d56e712d78d82b5..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:53965f3358b53bbc6b4d816f38ad873a41ec2ca933309504b8c2a86d7daa1689 -size 196355 diff --git a/lm-eval-output/allenai/OLMo-7B/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 50006ec3dd6079e31bcfe6baa07d7cebc29c3252..e04a83763148831b3c0975b290a0735d9f4eb818 100644 --- a/lm-eval-output/allenai/OLMo-7B/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "webqs": { - "exact_match,none": 0.030019685039370077, - "exact_match_stderr,none": 0.0037864267909448347, + "exact_match,none": 0.028543307086614175, + "exact_match_stderr,none": 0.0036949528903927557, "alias": "webqs" } }, @@ -46,13 +46,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index e16f3d53cd35707f167654216760d92311275542..5164377ab6e703cc57007843ed5beecc825a64bb 100644 --- a/lm-eval-output/allenai/OLMo-7B/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:a6809d92f1d0b0300cf31bc66170a0268f98b4689d8bf3b4614490d005ad896b -size 13459 +oid sha256:b0ac9038bece82c8286618ffc86fb6259ec8f99697d97d9ca74219ecfed34078 +size 12449 diff --git a/lm-eval-output/allenai/OLMo-7B/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 7af7d54203ecb78ab8de9828fdc1fcc280917c76..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:4bf1f4d379be7a987c76a55d997c04aeec3cc06ae4af6bf6aa0df9b7fe300ff7 -size 67701 diff --git a/lm-eval-output/allenai/OLMo-7B/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 227106483928a80698737bd7575091bf4753395d..0e029f27b848c6cf85ccb082c718426687a9e4bb 100644 --- a/lm-eval-output/allenai/OLMo-7B/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -47,13 +47,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 20af1e3e7e6a0bc3f0320ec1c43907b352d0d146..b00739fcb4662ef6d36fb28587ae6d5c4f3bad26 100644 --- a/lm-eval-output/allenai/OLMo-7B/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:ad1673980793757c05e9991bd725fa35d88dfa8fb4ee400aa699b88ffc6ac919 -size 13032 +oid sha256:0dcc692e4093c7208765cbeba1f1e471c81ebacc00d239e5880993cd4a47892e +size 15498 diff --git a/lm-eval-output/allenai/OLMo-7B/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 9d542ab1ecb99436967d1e9031a69b72419445cc..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:1c7f2ffae0318110244bd391cb82ba9695e5b3da3e4cfe708a42397332549d9a -size 955600 diff --git a/lm-eval-output/allenai/OLMo-7B/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index f752a237a6308c7f1a2b3ef406bd8e23ac027125..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,65 +0,0 @@ -{ - "results": { - "wikitext": { - "word_perplexity,none": 11.075255274027937, - "word_perplexity_stderr,none": "N/A", - "byte_perplexity,none": 1.567831445754393, - "byte_perplexity_stderr,none": "N/A", - "bits_per_byte,none": 0.6487704666946622, - "bits_per_byte_stderr,none": "N/A", - "alias": "wikitext" - } - }, - "configs": { - "wikitext": { - "task": "wikitext", - "dataset_path": "EleutherAI/wikitext_document_level", - "dataset_name": "wikitext-2-raw-v1", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "def wikitext_detokenizer(doc):\n string = doc[\"page\"]\n # contractions\n string = string.replace(\"s '\", \"s'\")\n string = re.sub(r\"/' [0-9]/\", r\"/'[0-9]/\", string)\n # number separators\n string = string.replace(\" @-@ \", \"-\")\n string = string.replace(\" @,@ \", \",\")\n string = string.replace(\" @.@ \", \".\")\n # punctuation\n string = string.replace(\" : \", \": \")\n string = string.replace(\" ; \", \"; \")\n string = string.replace(\" . \", \". \")\n string = string.replace(\" ! \", \"! \")\n string = string.replace(\" ? \", \"? \")\n string = string.replace(\" , \", \", \")\n # double brackets\n string = re.sub(r\"\\(\\s*([^\\)]*?)\\s*\\)\", r\"(\\1)\", string)\n string = re.sub(r\"\\[\\s*([^\\]]*?)\\s*\\]\", r\"[\\1]\", string)\n string = re.sub(r\"{\\s*([^}]*?)\\s*}\", r\"{\\1}\", string)\n string = re.sub(r\"\\\"\\s*([^\\\"]*?)\\s*\\\"\", r'\"\\1\"', string)\n string = re.sub(r\"'\\s*([^']*?)\\s*'\", r\"'\\1'\", string)\n # miscellaneous\n string = string.replace(\"= = = =\", \"====\")\n string = string.replace(\"= = =\", \"===\")\n string = string.replace(\"= =\", \"==\")\n string = string.replace(\" \" + chr(176) + \" \", chr(176))\n string = string.replace(\" \\n\", \"\\n\")\n string = string.replace(\"\\n \", \"\\n\")\n string = string.replace(\" N \", \" 1 \")\n string = string.replace(\" 's\", \"'s\")\n\n return string\n", - "process_results": "def process_results(doc, results):\n (loglikelihood,) = results\n # IMPORTANT: wikitext counts number of words in *original doc before detokenization*\n _words = len(re.split(r\"\\s+\", doc[\"page\"]))\n _bytes = len(doc[\"page\"].encode(\"utf-8\"))\n return {\n \"word_perplexity\": (loglikelihood, _words),\n \"byte_perplexity\": (loglikelihood, _bytes),\n \"bits_per_byte\": (loglikelihood, _bytes),\n }\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "word_perplexity" - }, - { - "metric": "byte_perplexity" - }, - { - "metric": "bits_per_byte" - } - ], - "output_type": "loglikelihood_rolling", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{page}}", - "metadata": { - "version": 2.0 - } - } - }, - "versions": { - "wikitext": 2.0 - }, - "n-shot": { - "wikitext": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "4701655" -} \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 82ce06502d9410b49213e687311276384a7de445..0a4f37b47adf137eeb92eb2d4fc7fe1081024290 100644 --- a/lm-eval-output/allenai/OLMo-7B/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:1de5b2a6517546aa5cb12e8acdf532ebadb56bdf47c90924c37151ce07a68c2e -size 18418 +oid sha256:e38a0a649d9a08386022bd2f889edc7a0a87a5f643be6aae3d2657f3111b0018 +size 53894 diff --git a/lm-eval-output/allenai/OLMo-7B/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 772e5ec95a4662037d94da5ee36101f2c691e57c..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:f607cbe7ac37a9cdf68d527736362a26dc68b13df6dc9a49de11e88d3c494fd1 -size 138033 diff --git a/lm-eval-output/allenai/OLMo-7B/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index e22cdb7a166f68ae8699e8b0a8af30235c0c3642..bf73aeae82bf50e2fc3ac63eedf8ec9e04126b47 100644 --- a/lm-eval-output/allenai/OLMo-7B/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "winogrande": { - "acc,none": 0.6629834254143646, - "acc_stderr,none": 0.01328495576939525, + "acc,none": 0.6748224151539068, + "acc_stderr,none": 0.013165525471764358, "alias": "winogrande" } }, @@ -44,13 +44,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 8ab3b1f769f76d10d313b4909455c2c3f654711a..f27e427af5b8fa7bd3a2e3df4edf683caab8917d 100644 --- a/lm-eval-output/allenai/OLMo-7B/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:0d426c615e740d84fe4d109ef0236e0a36b94f0039f8c9e2bca2d006c14fa187 -size 11800 +oid sha256:82d8d2ee281e36572544eb6174f104d3b90983584b04ecef207dc32e35a489af +size 10912 diff --git a/lm-eval-output/allenai/OLMo-7B/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 59f7523fb445bce729790f2e7ceb82ac937fab0e..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:8b42c76e9a5da6a9298c06437242b5c75aea3f25e6ae539367829ffab808da84 -size 8059 diff --git a/lm-eval-output/allenai/OLMo-7B/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index c647bfa102dd87db8c9ccb3bc3f168d18feaddf1..940e2b026f320b063630167cc0ef1997cb1b8871 100644 --- a/lm-eval-output/allenai/OLMo-7B/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -45,13 +45,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 32 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index d01351983d4cf7825d4894cbce5e7af8d54c2dfe..0fdf465e562401c7606294d5ea324f4e44d6a79c 100644 --- a/lm-eval-output/allenai/OLMo-7B/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:fc57ca4254e08c2155cd6189c777f6bd4da185c075c2b1e9e9fb1b5debc643c0 -size 12145 +oid sha256:e7122c21522d0a1cc22c4b891770ba5613448f0eb6feeec43e6d30d3f5a4f1b1 +size 14221 diff --git a/lm-eval-output/allenai/OLMo-7B/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 83ffdeb76ff46cfc67e89190435eb2cf97835eb2..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:af0cb5672ef31bf7a8429df773324b9de323b5eb50680bf45db8d0d39b0bf04a -size 11036 diff --git a/lm-eval-output/allenai/OLMo-7B/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 8c3b87da38eefd7dc560bed19a1330502e14ec59..533d3e7f14a2de745f57387317a4fbdeee3fdcf9 100644 --- a/lm-eval-output/allenai/OLMo-7B/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "wsc": { - "acc,none": 0.38461538461538464, - "acc_stderr,none": 0.047936688680750406, + "acc,none": 0.3942307692307692, + "acc_stderr,none": 0.048151547759907105, "alias": "wsc" } }, @@ -47,13 +47,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 8cb70488c789e6ee7ea7e290a26bcea14065f0bb..c39059059ea3721e66a723efcf46b02043ec5b7c 100644 --- a/lm-eval-output/allenai/OLMo-7B/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:4bf6a6447a388f2ac4bd86e725be1b907f27d32e45881c2e296472bb140311b7 -size 12178 +oid sha256:a3f008f9377b22030b9df7a0fddfb21eea1a3d36acfbe120abf324dc432bf5b3 +size 12869 diff --git a/lm-eval-output/allenai/OLMo-7B/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/allenai/OLMo-7B/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 71039f8d9d1f026d249c65de4590e304b9ab7007..0000000000000000000000000000000000000000 --- a/lm-eval-output/allenai/OLMo-7B/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:3efb2ba2f2b80febfe3c2ebd2232de7f6e2af2e0fc7e4e26294590bfca0e37d7 -size 32913 diff --git a/lm-eval-output/allenai/OLMo-7B/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/allenai/OLMo-7B/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index a3f5eea66ee8900fc9140ff4d9bcdd0fe0f7c1f3..dd8dbf8d239d3b605089bfd6a930e570a1c89487 100644 --- a/lm-eval-output/allenai/OLMo-7B/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/allenai/OLMo-7B/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,8 +1,8 @@ { "results": { "wsc273": { - "acc,none": 0.8424908424908425, - "acc_stderr,none": 0.02208772806150052, + "acc,none": 0.8498168498168498, + "acc_stderr,none": 0.021661514699106647, "alias": "wsc273" } }, @@ -44,13 +44,15 @@ "config": { "model": "hf", "model_args": "pretrained=allenai/OLMo-7B,dtype=bfloat16,trust_remote_code=True", - "batch_size": "2", - "batch_sizes": [], + "batch_size": "auto", + "batch_sizes": [ + 64 + ], "device": null, "use_cache": null, "limit": null, "bootstrap_iters": 100000, "gen_kwargs": null }, - "git_hash": "4701655" + "git_hash": "2e3ceb0" } \ No newline at end of file diff --git a/lm-eval-output/allenai/OLMo-7B/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/allenai/OLMo-7B/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index c50cdc20f0f2cb3260da25bbb62f42fcc4e5dce6..efdc9a3dcc53110b820f8d7c6209b4490f17dd9f 100644 --- a/lm-eval-output/allenai/OLMo-7B/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/allenai/OLMo-7B/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:81e3c7380f748a5f7aeb3bbdc1d580b1b1f37edde35ba4e8ce4e6f63192448cc -size 13051 +oid sha256:33cfb78a2df98c8c9e1b80a3641815beb7e8b5a562e0e8b615347058b3a8257f +size 13440 diff --git a/lm-eval-output/bloom-7b1/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/bloom-7b1/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 828312185ef1840804eeb4372c180f2c85de445f..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:319c68639bb05b6b985fcda887079637990b8c047b287314efc58672da605b83 -size 5205695 diff --git a/lm-eval-output/bloom-7b1/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/bloom-7b1/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 153562307bc09d7dacfca69a8e53242892f2f648..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,252 +0,0 @@ -{ - "results": { - "lambada_multilingual": { - "perplexity,none": 131.45396740665825, - "perplexity_stderr,none": 95.28024178884175, - "acc,none": 0.38490199883562976, - "acc_stderr,none": 0.07608898792977997, - "alias": "lambada_multilingual" - }, - "lambada_openai_mt_de": { - "perplexity,none": 370.91952810475857, - "perplexity_stderr,none": 24.98299339282566, - "acc,none": 0.23015718998641568, - "acc_stderr,none": 0.0058644241714399855, - "alias": " - lambada_openai_mt_de" - }, - "lambada_openai_mt_en": { - "perplexity,none": 6.583236525584539, - "perplexity_stderr,none": 0.17481189179976453, - "acc,none": 0.5717058024451775, - "acc_stderr,none": 0.0068939712541951454, - "alias": " - lambada_openai_mt_en" - }, - "lambada_openai_mt_es": { - "perplexity,none": 51.02874715706533, - "perplexity_stderr,none": 2.6341920857292744, - "acc,none": 0.36638851154667185, - "acc_stderr,none": 0.0067126579546010565, - "alias": " - lambada_openai_mt_es" - }, - "lambada_openai_mt_fr": { - "perplexity,none": 29.56217917543056, - "perplexity_stderr,none": 1.5411073949753211, - "acc,none": 0.4513875412381137, - "acc_stderr,none": 0.0069329758883686235, - "alias": " - lambada_openai_mt_fr" - }, - "lambada_openai_mt_it": { - "perplexity,none": 199.1761460704524, - "perplexity_stderr,none": 13.648756866456297, - "acc,none": 0.30487094896176986, - "acc_stderr,none": 0.006413613926848421, - "alias": " - lambada_openai_mt_it" - } - }, - "groups": { - "lambada_multilingual": { - "perplexity,none": 131.45396740665825, - "perplexity_stderr,none": 95.28024178884175, - "acc,none": 0.38490199883562976, - "acc_stderr,none": 0.07608898792977997, - "alias": "lambada_multilingual" - } - }, - "configs": { - "lambada_openai_mt_de": { - "task": "lambada_openai_mt_de", - "group": [ - "lambada_multilingual" - ], - "dataset_path": "EleutherAI/lambada_openai", - "dataset_name": "de", - "test_split": "test", - "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}", - "doc_to_target": "{{' '+text.split(' ')[-1]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "perplexity", - "aggregation": "perplexity", - "higher_is_better": false - }, - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "loglikelihood", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{text}}", - "metadata": { - "version": 1.0 - } - }, - "lambada_openai_mt_en": { - "task": "lambada_openai_mt_en", - "group": [ - "lambada_multilingual" - ], - "dataset_path": "EleutherAI/lambada_openai", - "dataset_name": "en", - "test_split": "test", - "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}", - "doc_to_target": "{{' '+text.split(' ')[-1]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "perplexity", - "aggregation": "perplexity", - "higher_is_better": false - }, - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "loglikelihood", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{text}}", - "metadata": { - "version": 1.0 - } - }, - "lambada_openai_mt_es": { - "task": "lambada_openai_mt_es", - "group": [ - "lambada_multilingual" - ], - "dataset_path": "EleutherAI/lambada_openai", - "dataset_name": "es", - "test_split": "test", - "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}", - "doc_to_target": "{{' '+text.split(' ')[-1]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "perplexity", - "aggregation": "perplexity", - "higher_is_better": false - }, - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "loglikelihood", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{text}}", - "metadata": { - "version": 1.0 - } - }, - "lambada_openai_mt_fr": { - "task": "lambada_openai_mt_fr", - "group": [ - "lambada_multilingual" - ], - "dataset_path": "EleutherAI/lambada_openai", - "dataset_name": "fr", - "test_split": "test", - "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}", - "doc_to_target": "{{' '+text.split(' ')[-1]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "perplexity", - "aggregation": "perplexity", - "higher_is_better": false - }, - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "loglikelihood", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{text}}", - "metadata": { - "version": 1.0 - } - }, - "lambada_openai_mt_it": { - "task": "lambada_openai_mt_it", - "group": [ - "lambada_multilingual" - ], - "dataset_path": "EleutherAI/lambada_openai", - "dataset_name": "it", - "test_split": "test", - "doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}", - "doc_to_target": "{{' '+text.split(' ')[-1]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "perplexity", - "aggregation": "perplexity", - "higher_is_better": false - }, - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "loglikelihood", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{text}}", - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "lambada_multilingual": "N/A", - "lambada_openai_mt_de": 1.0, - "lambada_openai_mt_en": 1.0, - "lambada_openai_mt_es": 1.0, - "lambada_openai_mt_fr": 1.0, - "lambada_openai_mt_it": 1.0 - }, - "n-shot": { - "lambada_multilingual": 0, - "lambada_openai_mt_de": 0, - "lambada_openai_mt_en": 0, - "lambada_openai_mt_es": 0, - "lambada_openai_mt_fr": 0, - "lambada_openai_mt_it": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=bigscience/bloom-7b1,dtype=bfloat16,trust_remote_code=True", - "batch_size": "auto", - "batch_sizes": [ - 16 - ], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "2c0a875" -} \ No newline at end of file diff --git a/lm-eval-output/bloom-7b1/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/bloom-7b1/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log deleted file mode 100644 index 15d8c6de95855da1da9ec65803e924ccc1be13c4..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/lambada_multilingual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:1305bf4cf9e181051bc11c7a57a1e45ce67ec14d33b83da65859d550c506ba29 -size 193365 diff --git a/lm-eval-output/bloom-7b1/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/bloom-7b1/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 3f1f5f46c39667c5450b42966a8a06e7d7bbbc7e..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:451306563781c10123a254b8642bf6e162d3ed0bdf7dc5a7deece8c1f0ebb316 -size 2133987 diff --git a/lm-eval-output/bloom-7b1/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/bloom-7b1/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 061ae85a79dddc445b381f6daf62168fb4e0e844..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,283 +0,0 @@ -{ - "results": { - "pawsx": { - "acc,none": 0.5078571428571429, - "acc_stderr,none": 0.03988534011535243, - "alias": "pawsx" - }, - "paws_de": { - "acc,none": 0.5175, - "acc_stderr,none": 0.011176284251254179, - "alias": " - paws_de" - }, - "paws_en": { - "acc,none": 0.4145, - "acc_stderr,none": 0.011018419931591767, - "alias": " - paws_en" - }, - "paws_es": { - "acc,none": 0.437, - "acc_stderr,none": 0.011094009127418984, - "alias": " - paws_es" - }, - "paws_fr": { - "acc,none": 0.5435, - "acc_stderr,none": 0.011140733053371404, - "alias": " - paws_fr" - }, - "paws_ja": { - "acc,none": 0.5575, - "acc_stderr,none": 0.01110894141174761, - "alias": " - paws_ja" - }, - "paws_ko": { - "acc,none": 0.552, - "acc_stderr,none": 0.011122493197456285, - "alias": " - paws_ko" - }, - "paws_zh": { - "acc,none": 0.533, - "acc_stderr,none": 0.01115875256825067, - "alias": " - paws_zh" - } - }, - "groups": { - "pawsx": { - "acc,none": 0.5078571428571429, - "acc_stderr,none": 0.03988534011535243, - "alias": "pawsx" - } - }, - "configs": { - "paws_de": { - "task": "paws_de", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "de", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", richtig? Ja, \"+sentence2, sentence1+\", richtig? Nein, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "paws_en": { - "task": "paws_en", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "en", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", right? Yes, \"+sentence2, sentence1+\", right? No, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "paws_es": { - "task": "paws_es", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "es", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", verdad? Sí, \"+sentence2, sentence1+\", verdad? No, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "paws_fr": { - "task": "paws_fr", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "fr", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", n'est-ce pas? Oui, \"+sentence2, sentence1+\", n'est-ce pas? No, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "paws_ja": { - "task": "paws_ja", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "ja", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", ですね? はい, \"+sentence2, sentence1+\", ですね? いいえ, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "paws_ko": { - "task": "paws_ko", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "ko", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", 맞죠? 예, \"+sentence2, sentence1+\", 맞죠? 아니요, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - }, - "paws_zh": { - "task": "paws_zh", - "group": "pawsx", - "dataset_path": "paws-x", - "dataset_name": "zh", - "training_split": "train", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[sentence1+\", 对吧? 是, \"+sentence2, sentence1+\", 对吧? 不是, \"+sentence2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 0.0 - } - } - }, - "versions": { - "paws_de": 0.0, - "paws_en": 0.0, - "paws_es": 0.0, - "paws_fr": 0.0, - "paws_ja": 0.0, - "paws_ko": 0.0, - "paws_zh": 0.0, - "pawsx": "N/A" - }, - "n-shot": { - "paws_de": 0, - "paws_en": 0, - "paws_es": 0, - "paws_fr": 0, - "paws_ja": 0, - "paws_ko": 0, - "paws_zh": 0, - "pawsx": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=bigscience/bloom-7b1,dtype=bfloat16,trust_remote_code=True", - "batch_size": "auto", - "batch_sizes": [ - 16 - ], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "2c0a875" -} \ No newline at end of file diff --git a/lm-eval-output/bloom-7b1/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/bloom-7b1/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log deleted file mode 100644 index 919cc8495aa6a1334dc71748354193116becb178..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/pawsx/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:ddfbdd9db11b8eaa02bb4e271211708dda3278c6e429ba83ad80fe07ff768106 -size 176108 diff --git a/lm-eval-output/bloom-7b1/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/bloom-7b1/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index ab11baae424569e0e9ea8951be6d88d966d8ac4d..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:974bd45d42256f61234e21e0e6244a357c2100dc7b9f9a73db0f239ba3bd04dd -size 531493 diff --git a/lm-eval-output/bloom-7b1/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/bloom-7b1/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 0e8dfe10cbc049d0424e08b200184031b60f3741..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,390 +0,0 @@ -{ - "results": { - "xcopa": { - "acc,none": 0.5709090909090908, - "acc_stderr,none": 0.06135942275478038, - "alias": "xcopa" - }, - "xcopa_et": { - "acc,none": 0.482, - "acc_stderr,none": 0.02236856511738799, - "alias": " - xcopa_et" - }, - "xcopa_ht": { - "acc,none": 0.516, - "acc_stderr,none": 0.0223716109825804, - "alias": " - xcopa_ht" - }, - "xcopa_id": { - "acc,none": 0.702, - "acc_stderr,none": 0.020475118092988978, - "alias": " - xcopa_id" - }, - "xcopa_it": { - "acc,none": 0.528, - "acc_stderr,none": 0.022347949832668093, - "alias": " - xcopa_it" - }, - "xcopa_qu": { - "acc,none": 0.512, - "acc_stderr,none": 0.02237662679792717, - "alias": " - xcopa_qu" - }, - "xcopa_sw": { - "acc,none": 0.518, - "acc_stderr,none": 0.02236856511738799, - "alias": " - xcopa_sw" - }, - "xcopa_ta": { - "acc,none": 0.592, - "acc_stderr,none": 0.02200091089387719, - "alias": " - xcopa_ta" - }, - "xcopa_th": { - "acc,none": 0.552, - "acc_stderr,none": 0.022261697292270132, - "alias": " - xcopa_th" - }, - "xcopa_tr": { - "acc,none": 0.512, - "acc_stderr,none": 0.02237662679792717, - "alias": " - xcopa_tr" - }, - "xcopa_vi": { - "acc,none": 0.716, - "acc_stderr,none": 0.02018670369357085, - "alias": " - xcopa_vi" - }, - "xcopa_zh": { - "acc,none": 0.65, - "acc_stderr,none": 0.021352091786223104, - "alias": " - xcopa_zh" - } - }, - "groups": { - "xcopa": { - "acc,none": 0.5709090909090908, - "acc_stderr,none": 0.06135942275478038, - "alias": "xcopa" - } - }, - "configs": { - "xcopa_et": { - "task": "xcopa_et", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "et", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'sest', 'effect': 'seetõttu'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_ht": { - "task": "xcopa_ht", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "ht", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'poukisa', 'effect': 'donk sa'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_id": { - "task": "xcopa_id", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "id", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'karena', 'effect': 'maka'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_it": { - "task": "xcopa_it", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "it", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'perché', 'effect': 'quindi'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_qu": { - "task": "xcopa_qu", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "qu", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'imataq', 'effect': 'chaymi'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_sw": { - "task": "xcopa_sw", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "sw", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'kwa sababu', 'effect': 'kwa hiyo'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_ta": { - "task": "xcopa_ta", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "ta", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'காரணமாக', 'effect': 'எனவே'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_th": { - "task": "xcopa_th", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "th", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'เพราะ', 'effect': 'ดังนั้น'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_tr": { - "task": "xcopa_tr", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "tr", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'çünkü', 'effect': 'bu yüzden'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_vi": { - "task": "xcopa_vi", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "vi", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': 'bởi vì', 'effect': 'vì vậy'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xcopa_zh": { - "task": "xcopa_zh", - "group": "xcopa", - "dataset_path": "xcopa", - "dataset_name": "zh", - "validation_split": "validation", - "test_split": "test", - "doc_to_text": "functools.partial(, connector={'cause': '因为', 'effect': '所以'})", - "doc_to_target": "label", - "doc_to_choice": "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"]), convert_choice(doc[\"choice2\"])]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc" - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "xcopa": "N/A", - "xcopa_et": 1.0, - "xcopa_ht": 1.0, - "xcopa_id": 1.0, - "xcopa_it": 1.0, - "xcopa_qu": 1.0, - "xcopa_sw": 1.0, - "xcopa_ta": 1.0, - "xcopa_th": 1.0, - "xcopa_tr": 1.0, - "xcopa_vi": 1.0, - "xcopa_zh": 1.0 - }, - "n-shot": { - "xcopa": 0, - "xcopa_et": 0, - "xcopa_ht": 0, - "xcopa_id": 0, - "xcopa_it": 0, - "xcopa_qu": 0, - "xcopa_sw": 0, - "xcopa_ta": 0, - "xcopa_th": 0, - "xcopa_tr": 0, - "xcopa_vi": 0, - "xcopa_zh": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=bigscience/bloom-7b1,dtype=bfloat16,trust_remote_code=True", - "batch_size": "auto", - "batch_sizes": [ - 32 - ], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "2c0a875" -} \ No newline at end of file diff --git a/lm-eval-output/bloom-7b1/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/bloom-7b1/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log deleted file mode 100644 index f217c75e55bcb796d3b90546bbf3e9d2c0fcdabb..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/xcopa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:d4ec1d12dfcff6d9ad4131af2712cc7261ca105612ed3df968592559b5f76302 -size 198895 diff --git a/lm-eval-output/bloom-7b1/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/bloom-7b1/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 45a9f843516226f09ad5a0bb74cae47a76533642..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:d32aeea4b41e4804c8a154051d6232d970ba02dc4770299682a84f23d9b73c88 -size 6015523 diff --git a/lm-eval-output/bloom-7b1/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/bloom-7b1/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 899d88e5025282dc793daaf70bc86e95a8f2cb08..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,548 +0,0 @@ -{ - "results": { - "xnli": { - "acc,none": 0.41204819277108434, - "acc_stderr,none": 0.051535476594892576, - "alias": "xnli" - }, - "xnli_ar": { - "acc,none": 0.3377510040160643, - "acc_stderr,none": 0.009479742273956477, - "alias": " - xnli_ar" - }, - "xnli_bg": { - "acc,none": 0.3779116465863454, - "acc_stderr,none": 0.009718712281227459, - "alias": " - xnli_bg" - }, - "xnli_de": { - "acc,none": 0.41365461847389556, - "acc_stderr,none": 0.009871502159099368, - "alias": " - xnli_de" - }, - "xnli_el": { - "acc,none": 0.3650602409638554, - "acc_stderr,none": 0.009650194822749637, - "alias": " - xnli_el" - }, - "xnli_en": { - "acc,none": 0.5261044176706827, - "acc_stderr,none": 0.01000840465166064, - "alias": " - xnli_en" - }, - "xnli_es": { - "acc,none": 0.4879518072289157, - "acc_stderr,none": 0.010019162857624487, - "alias": " - xnli_es" - }, - "xnli_fr": { - "acc,none": 0.478714859437751, - "acc_stderr,none": 0.010012987604500423, - "alias": " - xnli_fr" - }, - "xnli_hi": { - "acc,none": 0.4666666666666667, - "acc_stderr,none": 0.00999977679318763, - "alias": " - xnli_hi" - }, - "xnli_ru": { - "acc,none": 0.43253012048192774, - "acc_stderr,none": 0.009930409027139453, - "alias": " - xnli_ru" - }, - "xnli_sw": { - "acc,none": 0.3855421686746988, - "acc_stderr,none": 0.009755949341224318, - "alias": " - xnli_sw" - }, - "xnli_th": { - "acc,none": 0.3437751004016064, - "acc_stderr,none": 0.009520310502882936, - "alias": " - xnli_th" - }, - "xnli_tr": { - "acc,none": 0.3522088353413655, - "acc_stderr,none": 0.009574259292495757, - "alias": " - xnli_tr" - }, - "xnli_ur": { - "acc,none": 0.42289156626506025, - "acc_stderr,none": 0.009902179034797438, - "alias": " - xnli_ur" - }, - "xnli_vi": { - "acc,none": 0.44497991967871486, - "acc_stderr,none": 0.009961210239024633, - "alias": " - xnli_vi" - }, - "xnli_zh": { - "acc,none": 0.3449799196787149, - "acc_stderr,none": 0.009528219800053311, - "alias": " - xnli_zh" - } - }, - "groups": { - "xnli": { - "acc,none": 0.41204819277108434, - "acc_stderr,none": 0.051535476594892576, - "alias": "xnli" - } - }, - "configs": { - "xnli_ar": { - "task": "xnli_ar", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "ar", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", صحيح? نعم, \"+hypothesis,premise+\", صحيح? لذا, \"+hypothesis,premise+\", صحيح? رقم, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_bg": { - "task": "xnli_bg", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "bg", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", правилно? да, \"+hypothesis,premise+\", правилно? така, \"+hypothesis,premise+\", правилно? не, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_de": { - "task": "xnli_de", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "de", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", richtig? Ja, \"+hypothesis,premise+\", richtig? Auch, \"+hypothesis,premise+\", richtig? Nein, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_el": { - "task": "xnli_el", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "el", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", σωστός? Ναί, \"+hypothesis,premise+\", σωστός? Έτσι, \"+hypothesis,premise+\", σωστός? όχι, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_en": { - "task": "xnli_en", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "en", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", right? Yes, \"+hypothesis,premise+\", right? Also, \"+hypothesis,premise+\", right? No, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_es": { - "task": "xnli_es", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "es", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", correcto? Sí, \"+hypothesis,premise+\", correcto? Asi que, \"+hypothesis,premise+\", correcto? No, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_fr": { - "task": "xnli_fr", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "fr", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", correct? Oui, \"+hypothesis,premise+\", correct? Aussi, \"+hypothesis,premise+\", correct? Non, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_hi": { - "task": "xnli_hi", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "hi", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", सही? हाँ, \"+hypothesis,premise+\", सही? इसलिए, \"+hypothesis,premise+\", सही? नहीं, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_ru": { - "task": "xnli_ru", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "ru", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", правильно? Да, \"+hypothesis,premise+\", правильно? Так, \"+hypothesis,premise+\", правильно? Нет, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_sw": { - "task": "xnli_sw", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "sw", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", sahihi? Ndiyo, \"+hypothesis,premise+\", sahihi? Hivyo, \"+hypothesis,premise+\", sahihi? Hapana, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_th": { - "task": "xnli_th", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "th", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", ถูกต้อง? ใช่, \"+hypothesis,premise+\", ถูกต้อง? ดังนั้น, \"+hypothesis,premise+\", ถูกต้อง? ไม่, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_tr": { - "task": "xnli_tr", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "tr", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", doğru? Evet, \"+hypothesis,premise+\", doğru? Böylece, \"+hypothesis,premise+\", doğru? Hayır, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_ur": { - "task": "xnli_ur", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "ur", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", صحیح? جی ہاں, \"+hypothesis,premise+\", صحیح? اس لئے, \"+hypothesis,premise+\", صحیح? نہیں, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_vi": { - "task": "xnli_vi", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "vi", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", đúng? Vâng, \"+hypothesis,premise+\", đúng? Vì vậy, \"+hypothesis,premise+\", đúng? Không, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xnli_zh": { - "task": "xnli_zh", - "group": "xnli", - "dataset_path": "xnli", - "dataset_name": "zh", - "training_split": "train", - "validation_split": "validation", - "doc_to_text": "", - "doc_to_target": "label", - "doc_to_choice": "{{[premise+\", 正确? 是的, \"+hypothesis,premise+\", 正确? 所以, \"+hypothesis,premise+\", 正确? 不是的, \"+hypothesis]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "xnli": "N/A", - "xnli_ar": 1.0, - "xnli_bg": 1.0, - "xnli_de": 1.0, - "xnli_el": 1.0, - "xnli_en": 1.0, - "xnli_es": 1.0, - "xnli_fr": 1.0, - "xnli_hi": 1.0, - "xnli_ru": 1.0, - "xnli_sw": 1.0, - "xnli_th": 1.0, - "xnli_tr": 1.0, - "xnli_ur": 1.0, - "xnli_vi": 1.0, - "xnli_zh": 1.0 - }, - "n-shot": { - "xnli": 0, - "xnli_ar": 0, - "xnli_bg": 0, - "xnli_de": 0, - "xnli_el": 0, - "xnli_en": 0, - "xnli_es": 0, - "xnli_fr": 0, - "xnli_hi": 0, - "xnli_ru": 0, - "xnli_sw": 0, - "xnli_th": 0, - "xnli_tr": 0, - "xnli_ur": 0, - "xnli_vi": 0, - "xnli_zh": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=bigscience/bloom-7b1,dtype=bfloat16,trust_remote_code=True", - "batch_size": "auto", - "batch_sizes": [ - 16 - ], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "2c0a875" -} \ No newline at end of file diff --git a/lm-eval-output/bloom-7b1/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/bloom-7b1/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log deleted file mode 100644 index 806c61aa1238ecf0938fa4323f6887ea1e17b4b1..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/xnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:e89c3f8b693c7a40c2712cd450065d334fb5eacc0dd8ebe6292b9acea1ae84fd -size 66766 diff --git a/lm-eval-output/bloom-7b1/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/bloom-7b1/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index a446937d3b75f058c6a7bcaa38c42e20c7572405..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:42dbb9438d56fe87666d48fdd6b52cdf7c8578ef8d58b210bbbc3f770fdb9149 -size 4062985 diff --git a/lm-eval-output/bloom-7b1/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/bloom-7b1/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index 69cf929152579d2aecb4ceacecc168f4a11d5e3d..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,423 +0,0 @@ -{ - "results": { - "xstorycloze": { - "acc,none": 0.5927441188857469, - "acc_stderr,none": 0.05262352730974911, - "alias": "xstorycloze" - }, - "xstorycloze_ar": { - "acc,none": 0.5883520847121112, - "acc_stderr,none": 0.01266464832921408, - "alias": " - xstorycloze_ar" - }, - "xstorycloze_en": { - "acc,none": 0.7081403044341495, - "acc_stderr,none": 0.01169925603764938, - "alias": " - xstorycloze_en" - }, - "xstorycloze_es": { - "acc,none": 0.6598279285241562, - "acc_stderr,none": 0.012192034998028832, - "alias": " - xstorycloze_es" - }, - "xstorycloze_eu": { - "acc,none": 0.57114493712773, - "acc_stderr,none": 0.012736202713147777, - "alias": " - xstorycloze_eu" - }, - "xstorycloze_hi": { - "acc,none": 0.6048974189278623, - "acc_stderr,none": 0.012580772976133262, - "alias": " - xstorycloze_hi" - }, - "xstorycloze_id": { - "acc,none": 0.6419589675711449, - "acc_stderr,none": 0.012337624883487575, - "alias": " - xstorycloze_id" - }, - "xstorycloze_my": { - "acc,none": 0.48378557246856385, - "acc_stderr,none": 0.012860357805055867, - "alias": " - xstorycloze_my" - }, - "xstorycloze_ru": { - "acc,none": 0.5268034414295168, - "acc_stderr,none": 0.012848623899505765, - "alias": " - xstorycloze_ru" - }, - "xstorycloze_sw": { - "acc,none": 0.5413633355393779, - "acc_stderr,none": 0.012823020340169822, - "alias": " - xstorycloze_sw" - }, - "xstorycloze_te": { - "acc,none": 0.5744540039708802, - "acc_stderr,none": 0.012723670419166326, - "alias": " - xstorycloze_te" - }, - "xstorycloze_zh": { - "acc,none": 0.6194573130377233, - "acc_stderr,none": 0.012494500786685344, - "alias": " - xstorycloze_zh" - } - }, - "groups": { - "xstorycloze": { - "acc,none": 0.5927441188857469, - "acc_stderr,none": 0.05262352730974911, - "alias": "xstorycloze" - } - }, - "configs": { - "xstorycloze_ar": { - "task": "xstorycloze_ar", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "ar", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_en": { - "task": "xstorycloze_en", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "en", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_es": { - "task": "xstorycloze_es", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "es", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_eu": { - "task": "xstorycloze_eu", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "eu", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_hi": { - "task": "xstorycloze_hi", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "hi", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_id": { - "task": "xstorycloze_id", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "id", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_my": { - "task": "xstorycloze_my", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "my", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_ru": { - "task": "xstorycloze_ru", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "ru", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_sw": { - "task": "xstorycloze_sw", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "sw", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_te": { - "task": "xstorycloze_te", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "te", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - }, - "xstorycloze_zh": { - "task": "xstorycloze_zh", - "group": "xstorycloze", - "dataset_path": "juletxara/xstory_cloze", - "dataset_name": "zh", - "training_split": "train", - "validation_split": "eval", - "doc_to_text": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "doc_to_target": "{{answer_right_ending-1}}", - "doc_to_choice": "{{[sentence_quiz1, sentence_quiz2]}}", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": true, - "doc_to_decontamination_query": "{{[input_sentence_1, input_sentence_2, input_sentence_3, input_sentence_4]|join(' ')}}", - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "xstorycloze": "N/A", - "xstorycloze_ar": 1.0, - "xstorycloze_en": 1.0, - "xstorycloze_es": 1.0, - "xstorycloze_eu": 1.0, - "xstorycloze_hi": 1.0, - "xstorycloze_id": 1.0, - "xstorycloze_my": 1.0, - "xstorycloze_ru": 1.0, - "xstorycloze_sw": 1.0, - "xstorycloze_te": 1.0, - "xstorycloze_zh": 1.0 - }, - "n-shot": { - "xstorycloze": 0, - "xstorycloze_ar": 0, - "xstorycloze_en": 0, - "xstorycloze_es": 0, - "xstorycloze_eu": 0, - "xstorycloze_hi": 0, - "xstorycloze_id": 0, - "xstorycloze_my": 0, - "xstorycloze_ru": 0, - "xstorycloze_sw": 0, - "xstorycloze_te": 0, - "xstorycloze_zh": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=bigscience/bloom-7b1,dtype=bfloat16,trust_remote_code=True", - "batch_size": "auto", - "batch_sizes": [ - 4 - ], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "2c0a875" -} \ No newline at end of file diff --git a/lm-eval-output/bloom-7b1/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/bloom-7b1/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log deleted file mode 100644 index 16a53c5560899be54e56af664795673442230972..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/xstorycloze/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:9f292c6d6f96f368b1fbf97106cc7c977fe645a445719ee6d259d2cc2a301f64 -size 43759 diff --git a/lm-eval-output/bloom-7b1/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz b/lm-eval-output/bloom-7b1/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz deleted file mode 100644 index 6e3d5aa1a12bd4234f4cca187fd1c2688e01c217..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/result-jsonl.tar.gz +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:6561b1f7bcbbecb3a9c735882388b4e48b30275b5d70ff391ff0320005ecd7bc -size 513227 diff --git a/lm-eval-output/bloom-7b1/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/bloom-7b1/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json deleted file mode 100644 index f740f79f303c54fe787fd3db9c1412817979402a..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ /dev/null @@ -1,248 +0,0 @@ -{ - "results": { - "xwinograd": { - "acc,none": 0.7442121825129242, - "acc_stderr,none": 0.06414679137553342, - "alias": "xwinograd" - }, - "xwinograd_en": { - "acc,none": 0.8219354838709677, - "acc_stderr,none": 0.007935777723887321, - "alias": " - xwinograd_en" - }, - "xwinograd_fr": { - "acc,none": 0.6987951807228916, - "acc_stderr,none": 0.0506639425494172, - "alias": " - xwinograd_fr" - }, - "xwinograd_jp": { - "acc,none": 0.6037539103232534, - "acc_stderr,none": 0.015802642616557255, - "alias": " - xwinograd_jp" - }, - "xwinograd_pt": { - "acc,none": 0.7680608365019012, - "acc_stderr,none": 0.026075593860304693, - "alias": " - xwinograd_pt" - }, - "xwinograd_ru": { - "acc,none": 0.5714285714285714, - "acc_stderr,none": 0.02792722339076032, - "alias": " - xwinograd_ru" - }, - "xwinograd_zh": { - "acc,none": 0.7559523809523809, - "acc_stderr,none": 0.01915139944664688, - "alias": " - xwinograd_zh" - } - }, - "groups": { - "xwinograd": { - "acc,none": 0.7442121825129242, - "acc_stderr,none": 0.06414679137553342, - "alias": "xwinograd" - } - }, - "configs": { - "xwinograd_en": { - "task": "xwinograd_en", - "group": [ - "xwinograd" - ], - "dataset_path": "Muennighoff/xwinograd", - "dataset_name": "en", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc: Dict) -> int:\n \"\"\"\n Return index of the correct choice.\n\n Note: We are using the \"multiple input\" mode of the multiple-choice\n output-type, which means we use different contexts with the same target\n for the different choices, rather than the same context and different targets.\n \"\"\"\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc: Dict) -> str:\n \"\"\"\n Return the target completion.\n\n Note that this does not depend on the correct choice as we are using\n \"multiple input\" mode.\n \"\"\"\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return the choices that will be used as contexts in \"multiple input\" mode.\"\"\"\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xwinograd_fr": { - "task": "xwinograd_fr", - "group": [ - "xwinograd" - ], - "dataset_path": "Muennighoff/xwinograd", - "dataset_name": "fr", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc: Dict) -> int:\n \"\"\"\n Return index of the correct choice.\n\n Note: We are using the \"multiple input\" mode of the multiple-choice\n output-type, which means we use different contexts with the same target\n for the different choices, rather than the same context and different targets.\n \"\"\"\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc: Dict) -> str:\n \"\"\"\n Return the target completion.\n\n Note that this does not depend on the correct choice as we are using\n \"multiple input\" mode.\n \"\"\"\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return the choices that will be used as contexts in \"multiple input\" mode.\"\"\"\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xwinograd_jp": { - "task": "xwinograd_jp", - "group": [ - "xwinograd" - ], - "dataset_path": "Muennighoff/xwinograd", - "dataset_name": "jp", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc: Dict) -> int:\n \"\"\"\n Return index of the correct choice.\n\n Note: We are using the \"multiple input\" mode of the multiple-choice\n output-type, which means we use different contexts with the same target\n for the different choices, rather than the same context and different targets.\n \"\"\"\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc: Dict) -> str:\n \"\"\"\n Return the target completion.\n\n Note that this does not depend on the correct choice as we are using\n \"multiple input\" mode.\n \"\"\"\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return the choices that will be used as contexts in \"multiple input\" mode.\"\"\"\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xwinograd_pt": { - "task": "xwinograd_pt", - "group": [ - "xwinograd" - ], - "dataset_path": "Muennighoff/xwinograd", - "dataset_name": "pt", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc: Dict) -> int:\n \"\"\"\n Return index of the correct choice.\n\n Note: We are using the \"multiple input\" mode of the multiple-choice\n output-type, which means we use different contexts with the same target\n for the different choices, rather than the same context and different targets.\n \"\"\"\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc: Dict) -> str:\n \"\"\"\n Return the target completion.\n\n Note that this does not depend on the correct choice as we are using\n \"multiple input\" mode.\n \"\"\"\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return the choices that will be used as contexts in \"multiple input\" mode.\"\"\"\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xwinograd_ru": { - "task": "xwinograd_ru", - "group": [ - "xwinograd" - ], - "dataset_path": "Muennighoff/xwinograd", - "dataset_name": "ru", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc: Dict) -> int:\n \"\"\"\n Return index of the correct choice.\n\n Note: We are using the \"multiple input\" mode of the multiple-choice\n output-type, which means we use different contexts with the same target\n for the different choices, rather than the same context and different targets.\n \"\"\"\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc: Dict) -> str:\n \"\"\"\n Return the target completion.\n\n Note that this does not depend on the correct choice as we are using\n \"multiple input\" mode.\n \"\"\"\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return the choices that will be used as contexts in \"multiple input\" mode.\"\"\"\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - }, - "xwinograd_zh": { - "task": "xwinograd_zh", - "group": [ - "xwinograd" - ], - "dataset_path": "Muennighoff/xwinograd", - "dataset_name": "zh", - "test_split": "test", - "doc_to_text": "def doc_to_text(doc: Dict) -> int:\n \"\"\"\n Return index of the correct choice.\n\n Note: We are using the \"multiple input\" mode of the multiple-choice\n output-type, which means we use different contexts with the same target\n for the different choices, rather than the same context and different targets.\n \"\"\"\n answer_to_num = {\"1\": 0, \"2\": 1}\n return answer_to_num[doc[\"answer\"]]\n", - "doc_to_target": "def doc_to_target(doc: Dict) -> str:\n \"\"\"\n Return the target completion.\n\n Note that this does not depend on the correct choice as we are using\n \"multiple input\" mode.\n \"\"\"\n idx = doc[\"sentence\"].index(\"_\") + 1\n return doc[\"sentence\"][idx:].strip()\n", - "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return the choices that will be used as contexts in \"multiple input\" mode.\"\"\"\n idx = doc[\"sentence\"].index(\"_\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"sentence\"][:idx] + opt for opt in options]\n", - "description": "", - "target_delimiter": " ", - "fewshot_delimiter": "\n\n", - "metric_list": [ - { - "metric": "acc", - "aggregation": "mean", - "higher_is_better": true - } - ], - "output_type": "multiple_choice", - "repeats": 1, - "should_decontaminate": false, - "metadata": { - "version": 1.0 - } - } - }, - "versions": { - "xwinograd": "N/A", - "xwinograd_en": 1.0, - "xwinograd_fr": 1.0, - "xwinograd_jp": 1.0, - "xwinograd_pt": 1.0, - "xwinograd_ru": 1.0, - "xwinograd_zh": 1.0 - }, - "n-shot": { - "xwinograd": 0, - "xwinograd_en": 0, - "xwinograd_fr": 0, - "xwinograd_jp": 0, - "xwinograd_pt": 0, - "xwinograd_ru": 0, - "xwinograd_zh": 0 - }, - "config": { - "model": "hf", - "model_args": "pretrained=bigscience/bloom-7b1,dtype=bfloat16,trust_remote_code=True", - "batch_size": "auto", - "batch_sizes": [ - 32 - ], - "device": null, - "use_cache": null, - "limit": null, - "bootstrap_iters": 100000, - "gen_kwargs": null - }, - "git_hash": "2c0a875" -} \ No newline at end of file diff --git a/lm-eval-output/bloom-7b1/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/bloom-7b1/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log deleted file mode 100644 index a8c4abae15d1687cadb6e93f5e409f01480abe6a..0000000000000000000000000000000000000000 --- a/lm-eval-output/bloom-7b1/xwinograd/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ /dev/null @@ -1,3 +0,0 @@ -version https://git-lfs.github.com/spec/v1 -oid sha256:740f9d4172026e8ac9f919a971a6e456d06f9e51ba04fead1deb156409e23c74 -size 36676