odia_llama2_7B_base / README.md
leaderboard-pr-bot's picture
Adding Evaluation Results
0863260 verified
|
raw
history blame
5.04 kB
metadata
language:
  - or
license: cc-by-nc-4.0
model-index:
  - name: odia_llama2_7B_base
    results:
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: AI2 Reasoning Challenge (25-Shot)
          type: ai2_arc
          config: ARC-Challenge
          split: test
          args:
            num_few_shot: 25
        metrics:
          - type: acc_norm
            value: 50.77
            name: normalized accuracy
        source:
          url: >-
            https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=OdiaGenAI/odia_llama2_7B_base
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: HellaSwag (10-Shot)
          type: hellaswag
          split: validation
          args:
            num_few_shot: 10
        metrics:
          - type: acc_norm
            value: 75.94
            name: normalized accuracy
        source:
          url: >-
            https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=OdiaGenAI/odia_llama2_7B_base
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: MMLU (5-Shot)
          type: cais/mmlu
          config: all
          split: test
          args:
            num_few_shot: 5
        metrics:
          - type: acc
            value: 46.1
            name: accuracy
        source:
          url: >-
            https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=OdiaGenAI/odia_llama2_7B_base
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: TruthfulQA (0-shot)
          type: truthful_qa
          config: multiple_choice
          split: validation
          args:
            num_few_shot: 0
        metrics:
          - type: mc2
            value: 37.27
        source:
          url: >-
            https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=OdiaGenAI/odia_llama2_7B_base
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: Winogrande (5-shot)
          type: winogrande
          config: winogrande_xl
          split: validation
          args:
            num_few_shot: 5
        metrics:
          - type: acc
            value: 70.8
            name: accuracy
        source:
          url: >-
            https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=OdiaGenAI/odia_llama2_7B_base
          name: Open LLM Leaderboard
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: GSM8k (5-shot)
          type: gsm8k
          config: main
          split: test
          args:
            num_few_shot: 5
        metrics:
          - type: acc
            value: 14.94
            name: accuracy
        source:
          url: >-
            https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=OdiaGenAI/odia_llama2_7B_base
          name: Open LLM Leaderboard

Model Card for Model ID

License: CC BY-NC-SA 4.0

Model description

odia_llama2_7B_base is based on Llama2-7b and finetuned with 180k Odia instruction set. The instruction set is translated data from open-source resources and a prepared domain knowledge instruction set, resulting in good Odia instruction understanding and response generation capabilities.

The code of Odia data generation and other detailed information can be found in our Github project repository: https://github.com/OdiaGenAI/GenerativeAI_and_LLM_Odia.

Instructions for running it can be found at https://github.com/OdiaGenAI/GenerativeAI_and_LLM_Odia.

Licensing Information

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.

CC BY-NC-SA 4.0

Citation Information

If you find this helpful repository, please consider giving 👏 and citing:

@misc{OdiaGenAI_Odia_Llama2_7b_base,
  author = {Shantipriya Parida and Sambit Sekhar and Guneet Singh Kohli},
  title = {OdiaGenAI_Odia_Llama2_7b_base},
  year = {2023},
  publisher = {Hugging Face},
  journal = {Hugging Face repository},
  howpublished = {\url{https://huggingface.co/OdiaGenAI}},
}

Contributors:

  • Guneet Singh Kohli
  • Sambit Shekhar
  • Shantipriya Parida

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 49.30
AI2 Reasoning Challenge (25-Shot) 50.77
HellaSwag (10-Shot) 75.94
MMLU (5-Shot) 46.10
TruthfulQA (0-shot) 37.27
Winogrande (5-shot) 70.80
GSM8k (5-shot) 14.94