Adding Evaluation Results

#1
Files changed (1) hide show
  1. README.md +16 -2
README.md CHANGED
@@ -1,10 +1,10 @@
1
  ---
2
- library_name: transformers
3
  license: llama3
4
- base_model: meta-llama/Meta-Llama-3-8B
5
  tags:
6
  - axolotl
7
  - generated_from_trainer
 
8
  model-index:
9
  - name: L3-Pneuma-8B
10
  results: []
@@ -147,3 +147,17 @@ The following hyperparameters were used during training:
147
  - Pytorch 2.3.1+cu121
148
  - Datasets 2.21.0
149
  - Tokenizers 0.20.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
2
  license: llama3
3
+ library_name: transformers
4
  tags:
5
  - axolotl
6
  - generated_from_trainer
7
+ base_model: meta-llama/Meta-Llama-3-8B
8
  model-index:
9
  - name: L3-Pneuma-8B
10
  results: []
 
147
  - Pytorch 2.3.1+cu121
148
  - Datasets 2.21.0
149
  - Tokenizers 0.20.1
150
+
151
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
152
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Kquant03__L3-Pneuma-8B)
153
+
154
+ | Metric |Value|
155
+ |-------------------|----:|
156
+ |Avg. |16.57|
157
+ |IFEval (0-Shot) |23.74|
158
+ |BBH (3-Shot) |28.82|
159
+ |MATH Lvl 5 (4-Shot)| 4.76|
160
+ |GPQA (0-shot) | 7.61|
161
+ |MuSR (0-shot) |10.21|
162
+ |MMLU-PRO (5-shot) |24.27|
163
+