rwmasood commited on
Commit
a385b17
·
verified ·
1 Parent(s): e390311

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -109,7 +109,7 @@ print(round(results["mean_perplexity"], 2))
109
  ### Harness Evaluation
110
 
111
  - The performance evaluation is based on the tasks being evaluated on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
112
- The model is evaluated on three benchmark datasets, which include `ARC-Challenge`, `HellaSwag` and `MMLU`.
113
  The library used is [lm-evaluation-harness repository](https://github.com/EleutherAI/lm-evaluation-harness)
114
 
115
 
 
109
  ### Harness Evaluation
110
 
111
  - The performance evaluation is based on the tasks being evaluated on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
112
+ The model is evaluated on three benchmark datasets, which include `ARC-Challenge`, `HellaSwag`, `MMLU` and `IFEval`.
113
  The library used is [lm-evaluation-harness repository](https://github.com/EleutherAI/lm-evaluation-harness)
114
 
115