Locutusque
commited on
Delete duplicate eval
Browse files
README.md
CHANGED
@@ -175,18 +175,6 @@ pipeline = transformers.pipeline(
|
|
175 |
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
176 |
print(outputs[0]["generated_text"])
|
177 |
```
|
178 |
-
## [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
179 |
-
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Locutusque__OpenHercules-2.5-Mistral-7B)
|
180 |
-
|
181 |
-
| Metric |Value|
|
182 |
-
|---------------------------------|----:|
|
183 |
-
|Avg. |66.55|
|
184 |
-
|AI2 Reasoning Challenge (25-Shot)|64.25|
|
185 |
-
|HellaSwag (10-Shot) |84.84|
|
186 |
-
|MMLU (5-Shot) |64.21|
|
187 |
-
|TruthfulQA (0-shot) |47.84|
|
188 |
-
|Winogrande (5-shot) |78.93|
|
189 |
-
|GSM8k (5-shot) |59.21|
|
190 |
|
191 |
## Quants
|
192 |
|
|
|
175 |
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
176 |
print(outputs[0]["generated_text"])
|
177 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
178 |
|
179 |
## Quants
|
180 |
|