Files changed (1) hide show
  1. README.md +106 -0
README.md CHANGED
@@ -110,6 +110,98 @@ model-index:
110
  source:
111
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=shadowml/BeagSake-7B
112
  name: Open LLM Leaderboard
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
113
  ---
114
 
115
  # BeagSake-7B
@@ -176,3 +268,17 @@ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-le
176
  |Winogrande (5-shot) |82.16|
177
  |GSM8k (5-shot) |71.80|
178
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
110
  source:
111
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=shadowml/BeagSake-7B
112
  name: Open LLM Leaderboard
113
+ - task:
114
+ type: text-generation
115
+ name: Text Generation
116
+ dataset:
117
+ name: IFEval (0-Shot)
118
+ type: HuggingFaceH4/ifeval
119
+ args:
120
+ num_few_shot: 0
121
+ metrics:
122
+ - type: inst_level_strict_acc and prompt_level_strict_acc
123
+ value: 52.16
124
+ name: strict accuracy
125
+ source:
126
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=shadowml/BeagSake-7B
127
+ name: Open LLM Leaderboard
128
+ - task:
129
+ type: text-generation
130
+ name: Text Generation
131
+ dataset:
132
+ name: BBH (3-Shot)
133
+ type: BBH
134
+ args:
135
+ num_few_shot: 3
136
+ metrics:
137
+ - type: acc_norm
138
+ value: 25.19
139
+ name: normalized accuracy
140
+ source:
141
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=shadowml/BeagSake-7B
142
+ name: Open LLM Leaderboard
143
+ - task:
144
+ type: text-generation
145
+ name: Text Generation
146
+ dataset:
147
+ name: MATH Lvl 5 (4-Shot)
148
+ type: hendrycks/competition_math
149
+ args:
150
+ num_few_shot: 4
151
+ metrics:
152
+ - type: exact_match
153
+ value: 5.44
154
+ name: exact match
155
+ source:
156
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=shadowml/BeagSake-7B
157
+ name: Open LLM Leaderboard
158
+ - task:
159
+ type: text-generation
160
+ name: Text Generation
161
+ dataset:
162
+ name: GPQA (0-shot)
163
+ type: Idavidrein/gpqa
164
+ args:
165
+ num_few_shot: 0
166
+ metrics:
167
+ - type: acc_norm
168
+ value: 4.14
169
+ name: acc_norm
170
+ source:
171
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=shadowml/BeagSake-7B
172
+ name: Open LLM Leaderboard
173
+ - task:
174
+ type: text-generation
175
+ name: Text Generation
176
+ dataset:
177
+ name: MuSR (0-shot)
178
+ type: TAUR-Lab/MuSR
179
+ args:
180
+ num_few_shot: 0
181
+ metrics:
182
+ - type: acc_norm
183
+ value: 9.84
184
+ name: acc_norm
185
+ source:
186
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=shadowml/BeagSake-7B
187
+ name: Open LLM Leaderboard
188
+ - task:
189
+ type: text-generation
190
+ name: Text Generation
191
+ dataset:
192
+ name: MMLU-PRO (5-shot)
193
+ type: TIGER-Lab/MMLU-Pro
194
+ config: main
195
+ split: test
196
+ args:
197
+ num_few_shot: 5
198
+ metrics:
199
+ - type: acc
200
+ value: 17.61
201
+ name: accuracy
202
+ source:
203
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=shadowml/BeagSake-7B
204
+ name: Open LLM Leaderboard
205
  ---
206
 
207
  # BeagSake-7B
 
268
  |Winogrande (5-shot) |82.16|
269
  |GSM8k (5-shot) |71.80|
270
 
271
+
272
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
273
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_shadowml__BeagSake-7B)
274
+
275
+ | Metric |Value|
276
+ |-------------------|----:|
277
+ |Avg. |19.06|
278
+ |IFEval (0-Shot) |52.16|
279
+ |BBH (3-Shot) |25.19|
280
+ |MATH Lvl 5 (4-Shot)| 5.44|
281
+ |GPQA (0-shot) | 4.14|
282
+ |MuSR (0-shot) | 9.84|
283
+ |MMLU-PRO (5-shot) |17.61|
284
+