alrope commited on
Commit
b4abde0
·
verified ·
1 Parent(s): 9e483fa

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. temperature=0.0/Llama-2-13b-chat-hf.json +24 -24
  2. temperature=0.0/Llama-2-70b-chat-hf.json +23 -23
  3. temperature=0.0/Llama-2-7b-chat-hf.json +24 -24
  4. temperature=0.0/Llama-3.1-70B-Instruct.json +24 -24
  5. temperature=0.0/Llama-3.1-8B-Instruct.json +24 -24
  6. temperature=0.0/Mistral-7B-Instruct-v0.3.json +24 -24
  7. temperature=0.0/Mistral-Large-Instruct-2407.json +24 -24
  8. temperature=0.0/Mistral-Small-Instruct-2409.json +24 -24
  9. temperature=0.0/OLMo-7B-0724-Instruct-hf.json +24 -24
  10. temperature=0.0/OLMo-7B-SFT-hf.json +23 -23
  11. temperature=0.0/Phi-3-medium-4k-instruct.json +24 -24
  12. temperature=0.0/Qwen1.5-110B-Chat.json +24 -24
  13. temperature=0.0/Qwen2-72B-Instruct.json +22 -22
  14. temperature=0.0/Qwen2.5-72B-Instruct.json +24 -24
  15. temperature=0.0/WizardLM-13B-V1.2.json +24 -24
  16. temperature=0.0/Yi-1.5-34B-Chat.json +24 -24
  17. temperature=0.0/dolly-v2-12b.json +22 -22
  18. temperature=0.0/dolly-v2-7b.json +22 -22
  19. temperature=0.0/gpt4all-13b-snoozy.json +24 -24
  20. temperature=0.0/koala-13B-HF.json +24 -24
  21. temperature=0.0/koala-7B-HF.json +22 -22
  22. temperature=0.0/mpt-7b-chat.json +24 -24
  23. temperature=0.0/oasst-sft-1-pythia-12b.json +22 -22
  24. temperature=0.0/tulu-2-dpo-13b.json +24 -24
  25. temperature=0.0/tulu-2-dpo-70b.json +24 -24
  26. temperature=0.0/tulu-2-dpo-7b.json +23 -23
  27. temperature=0.0/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm.json +24 -24
  28. temperature=0.0/vicuna-13b-v1.5.json +24 -24
  29. temperature=0.0/vicuna-7b-v1.5.json +24 -24
  30. temperature=1.0/Llama-2-13b-chat-hf.json +24 -24
  31. temperature=1.0/Llama-2-70b-chat-hf.json +24 -24
  32. temperature=1.0/Llama-2-7b-chat-hf.json +23 -23
  33. temperature=1.0/Llama-3.1-70B-Instruct.json +24 -24
  34. temperature=1.0/Llama-3.1-8B-Instruct.json +23 -23
  35. temperature=1.0/Mistral-7B-Instruct-v0.3.json +24 -24
  36. temperature=1.0/Mistral-Large-Instruct-2407.json +24 -24
  37. temperature=1.0/Mistral-Small-Instruct-2409.json +24 -24
  38. temperature=1.0/OLMo-7B-0724-Instruct-hf.json +24 -24
  39. temperature=1.0/OLMo-7B-SFT-hf.json +24 -24
  40. temperature=1.0/Phi-3-medium-4k-instruct.json +24 -24
  41. temperature=1.0/Qwen1.5-110B-Chat.json +23 -23
  42. temperature=1.0/Qwen2-72B-Instruct.json +24 -24
  43. temperature=1.0/Qwen2.5-72B-Instruct.json +24 -24
  44. temperature=1.0/WizardLM-13B-V1.2.json +23 -23
  45. temperature=1.0/Yi-1.5-34B-Chat.json +24 -24
  46. temperature=1.0/dolly-v2-12b.json +16 -16
  47. temperature=1.0/dolly-v2-7b.json +16 -16
  48. temperature=1.0/gpt4all-13b-snoozy.json +24 -24
  49. temperature=1.0/koala-13B-HF.json +23 -23
  50. temperature=1.0/koala-7B-HF.json +22 -22
temperature=0.0/Llama-2-13b-chat-hf.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "meta-llama/Llama-2-13b-chat-hf",
3
- "brainstorm": 0.2,
4
- "open_qa": 0.5,
5
- "closed_qa": 0.2,
6
- "extract": 0.2,
7
- "generation": 0.2,
8
- "rewrite": 0.1,
9
- "summarize": 0.2,
10
- "classify": 0.3,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.1,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.19,
15
  "brainstorm_rank": 15,
16
  "open_qa_rank": 24,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 16,
25
  "fact_checking_or_attributed_qa_rank": 9,
26
  "average_rank": 13,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "meta-llama/Llama-2-13b-chat-hf",
3
+ "brainstorm": 0.168,
4
+ "open_qa": 0.525,
5
+ "closed_qa": 0.218,
6
+ "extract": 0.198,
7
+ "generation": 0.159,
8
+ "rewrite": 0.142,
9
+ "summarize": 0.198,
10
+ "classify": 0.274,
11
+ "reasoning_over_numerical_data": 0.115,
12
+ "multi-document_synthesis": 0.091,
13
+ "fact_checking_or_attributed_qa": 0.42,
14
+ "average": 0.1936,
15
  "brainstorm_rank": 15,
16
  "open_qa_rank": 24,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 16,
25
  "fact_checking_or_attributed_qa_rank": 9,
26
  "average_rank": 13,
27
+ "brainstorm_confi": "+2.9/-2.9",
28
+ "open_qa_confi": "+9.8/-9.8",
29
+ "closed_qa_confi": "+5.0/-4.7",
30
+ "extract_confi": "+4.7/-4.5",
31
+ "generation_confi": "+3.0/-2.8",
32
+ "rewrite_confi": "+2.5/-2.6",
33
+ "summarize_confi": "+5.4/-5.0",
34
+ "classify_confi": "+5.5/-5.5",
35
+ "reasoning_over_numerical_data_confi": "+2.6/-2.5",
36
+ "multi-document_synthesis_confi": "+2.9/-2.4",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.5",
38
+ "average_confi": "+1.15/-1.15"
39
  }
temperature=0.0/Llama-2-70b-chat-hf.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "meta-llama/Llama-2-70b-chat-hf",
3
- "brainstorm": 0.2,
4
- "open_qa": 0.6,
5
  "closed_qa": 0.3,
6
- "extract": 0.3,
7
- "generation": 0.2,
8
- "rewrite": 0.2,
9
- "summarize": 0.2,
10
- "classify": 0.3,
11
- "reasoning_over_numerical_data": 0.2,
12
- "multi-document_synthesis": 0.1,
13
- "fact_checking_or_attributed_qa": 0.5,
14
- "average": 0.24,
15
  "brainstorm_rank": 12,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 13,
25
  "fact_checking_or_attributed_qa_rank": 5,
26
  "average_rank": 11,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "meta-llama/Llama-2-70b-chat-hf",
3
+ "brainstorm": 0.223,
4
+ "open_qa": 0.613,
5
  "closed_qa": 0.3,
6
+ "extract": 0.26,
7
+ "generation": 0.192,
8
+ "rewrite": 0.171,
9
+ "summarize": 0.218,
10
+ "classify": 0.343,
11
+ "reasoning_over_numerical_data": 0.191,
12
+ "multi-document_synthesis": 0.132,
13
+ "fact_checking_or_attributed_qa": 0.485,
14
+ "average": 0.2448,
15
  "brainstorm_rank": 12,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 13,
25
  "fact_checking_or_attributed_qa_rank": 5,
26
  "average_rank": 11,
27
+ "brainstorm_confi": "+3.3/-3.3",
28
+ "open_qa_confi": "+9.3/-9.3",
29
+ "closed_qa_confi": "+5.4/-5.4",
30
+ "extract_confi": "+5.2/-5.0",
31
+ "generation_confi": "+3.2/-3.1",
32
+ "rewrite_confi": "+2.9/-2.7",
33
+ "summarize_confi": "+5.7/-5.4",
34
+ "classify_confi": "+6.0/-6.0",
35
+ "reasoning_over_numerical_data_confi": "+3.2/-3.2",
36
+ "multi-document_synthesis_confi": "+3.1/-3.1",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.5",
38
+ "average_confi": "+1.27/-1.24"
39
  }
temperature=0.0/Llama-2-7b-chat-hf.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "meta-llama/Llama-2-7b-chat-hf",
3
- "brainstorm": 0.2,
4
- "open_qa": 0.4,
5
- "closed_qa": 0.2,
6
- "extract": 0.1,
7
- "generation": 0.1,
8
- "rewrite": 0.1,
9
- "summarize": 0.1,
10
- "classify": 0.2,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.1,
13
- "fact_checking_or_attributed_qa": 0.3,
14
- "average": 0.16,
15
  "brainstorm_rank": 15,
16
  "open_qa_rank": 27,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 16,
25
  "fact_checking_or_attributed_qa_rank": 21,
26
  "average_rank": 18,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "meta-llama/Llama-2-7b-chat-hf",
3
+ "brainstorm": 0.164,
4
+ "open_qa": 0.417,
5
+ "closed_qa": 0.213,
6
+ "extract": 0.129,
7
+ "generation": 0.133,
8
+ "rewrite": 0.129,
9
+ "summarize": 0.097,
10
+ "classify": 0.192,
11
+ "reasoning_over_numerical_data": 0.073,
12
+ "multi-document_synthesis": 0.075,
13
+ "fact_checking_or_attributed_qa": 0.335,
14
+ "average": 0.1563,
15
  "brainstorm_rank": 15,
16
  "open_qa_rank": 27,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 16,
25
  "fact_checking_or_attributed_qa_rank": 21,
26
  "average_rank": 18,
27
+ "brainstorm_confi": "+2.9/-2.8",
28
+ "open_qa_confi": "+9.8/-9.3",
29
+ "closed_qa_confi": "+5.2/-5.0",
30
+ "extract_confi": "+4.2/-4.0",
31
+ "generation_confi": "+2.8/-2.7",
32
+ "rewrite_confi": "+2.6/-2.5",
33
+ "summarize_confi": "+4.0/-3.7",
34
+ "classify_confi": "+5.2/-5.0",
35
+ "reasoning_over_numerical_data_confi": "+2.1/-2.0",
36
+ "multi-document_synthesis_confi": "+2.4/-2.4",
37
+ "fact_checking_or_attributed_qa_confi": "+4.3/-4.1",
38
+ "average_confi": "+1.06/-1.06"
39
  }
temperature=0.0/Llama-3.1-70B-Instruct.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "meta-llama/Llama-3.1-70B-Instruct",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.8,
5
- "closed_qa": 0.6,
6
- "extract": 0.5,
7
- "generation": 0.5,
8
- "rewrite": 0.4,
9
- "summarize": 0.4,
10
- "classify": 0.5,
11
- "reasoning_over_numerical_data": 0.5,
12
- "multi-document_synthesis": 0.5,
13
- "fact_checking_or_attributed_qa": 0.6,
14
- "average": 0.5,
15
  "brainstorm_rank": 4,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 4,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 1,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "meta-llama/Llama-3.1-70B-Instruct",
3
+ "brainstorm": 0.486,
4
+ "open_qa": 0.843,
5
+ "closed_qa": 0.552,
6
+ "extract": 0.45,
7
+ "generation": 0.455,
8
+ "rewrite": 0.446,
9
+ "summarize": 0.433,
10
+ "classify": 0.545,
11
+ "reasoning_over_numerical_data": 0.503,
12
+ "multi-document_synthesis": 0.498,
13
+ "fact_checking_or_attributed_qa": 0.571,
14
+ "average": 0.498,
15
  "brainstorm_rank": 4,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 4,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 1,
27
+ "brainstorm_confi": "+4.0/-3.8",
28
+ "open_qa_confi": "+6.9/-7.4",
29
+ "closed_qa_confi": "+5.9/-6.2",
30
+ "extract_confi": "+5.9/-5.7",
31
+ "generation_confi": "+3.9/-3.8",
32
+ "rewrite_confi": "+3.7/-3.7",
33
+ "summarize_confi": "+6.4/-6.2",
34
+ "classify_confi": "+6.0/-6.0",
35
+ "reasoning_over_numerical_data_confi": "+3.6/-3.8",
36
+ "multi-document_synthesis_confi": "+4.5/-4.6",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.5",
38
+ "average_confi": "+1.44/-1.46"
39
  }
temperature=0.0/Llama-3.1-8B-Instruct.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "meta-llama/Llama-3.1-8B-Instruct",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.8,
5
- "closed_qa": 0.4,
6
- "extract": 0.3,
7
- "generation": 0.4,
8
- "rewrite": 0.4,
9
- "summarize": 0.3,
10
- "classify": 0.4,
11
- "reasoning_over_numerical_data": 0.3,
12
- "multi-document_synthesis": 0.4,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.39,
15
  "brainstorm_rank": 7,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 7,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 6,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "meta-llama/Llama-3.1-8B-Instruct",
3
+ "brainstorm": 0.467,
4
+ "open_qa": 0.794,
5
+ "closed_qa": 0.406,
6
+ "extract": 0.324,
7
+ "generation": 0.363,
8
+ "rewrite": 0.367,
9
+ "summarize": 0.329,
10
+ "classify": 0.43,
11
+ "reasoning_over_numerical_data": 0.299,
12
+ "multi-document_synthesis": 0.358,
13
+ "fact_checking_or_attributed_qa": 0.393,
14
+ "average": 0.385,
15
  "brainstorm_rank": 7,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 7,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 6,
27
+ "brainstorm_confi": "+3.8/-3.8",
28
+ "open_qa_confi": "+6.9/-8.3",
29
+ "closed_qa_confi": "+6.2/-6.2",
30
+ "extract_confi": "+5.4/-5.4",
31
+ "generation_confi": "+3.9/-3.9",
32
+ "rewrite_confi": "+3.5/-3.6",
33
+ "summarize_confi": "+6.2/-6.2",
34
+ "classify_confi": "+6.2/-6.0",
35
+ "reasoning_over_numerical_data_confi": "+3.5/-3.6",
36
+ "multi-document_synthesis_confi": "+4.2/-4.4",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.4",
38
+ "average_confi": "+1.39/-1.37"
39
  }
temperature=0.0/Mistral-7B-Instruct-v0.3.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "mistralai/Mistral-7B-Instruct-v0.3",
3
- "brainstorm": 0.2,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.2,
6
- "extract": 0.1,
7
- "generation": 0.2,
8
- "rewrite": 0.2,
9
- "summarize": 0.2,
10
- "classify": 0.3,
11
- "reasoning_over_numerical_data": 0.2,
12
- "multi-document_synthesis": 0.3,
13
- "fact_checking_or_attributed_qa": 0.5,
14
- "average": 0.24,
15
  "brainstorm_rank": 10,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 8,
25
  "fact_checking_or_attributed_qa_rank": 9,
26
  "average_rank": 12,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "mistralai/Mistral-7B-Instruct-v0.3",
3
+ "brainstorm": 0.241,
4
+ "open_qa": 0.608,
5
+ "closed_qa": 0.213,
6
+ "extract": 0.141,
7
+ "generation": 0.191,
8
+ "rewrite": 0.185,
9
+ "summarize": 0.168,
10
+ "classify": 0.271,
11
+ "reasoning_over_numerical_data": 0.173,
12
+ "multi-document_synthesis": 0.255,
13
+ "fact_checking_or_attributed_qa": 0.465,
14
+ "average": 0.2427,
15
  "brainstorm_rank": 10,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 8,
25
  "fact_checking_or_attributed_qa_rank": 9,
26
  "average_rank": 12,
27
+ "brainstorm_confi": "+3.4/-3.3",
28
+ "open_qa_confi": "+8.8/-9.8",
29
+ "closed_qa_confi": "+5.2/-5.0",
30
+ "extract_confi": "+4.5/-4.2",
31
+ "generation_confi": "+3.1/-3.0",
32
+ "rewrite_confi": "+3.0/-2.9",
33
+ "summarize_confi": "+5.2/-4.7",
34
+ "classify_confi": "+6.0/-6.0",
35
+ "reasoning_over_numerical_data_confi": "+3.1/-3.0",
36
+ "multi-document_synthesis_confi": "+4.1/-4.0",
37
+ "fact_checking_or_attributed_qa_confi": "+4.8/-4.5",
38
+ "average_confi": "+1.23/-1.27"
39
  }
temperature=0.0/Mistral-Large-Instruct-2407.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "mistralai/Mistral-Large-Instruct-2407",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.4,
6
- "extract": 0.4,
7
- "generation": 0.5,
8
- "rewrite": 0.5,
9
- "summarize": 0.5,
10
- "classify": 0.4,
11
- "reasoning_over_numerical_data": 0.5,
12
- "multi-document_synthesis": 0.6,
13
- "fact_checking_or_attributed_qa": 0.3,
14
- "average": 0.48,
15
  "brainstorm_rank": 1,
16
  "open_qa_rank": 24,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 1,
25
  "fact_checking_or_attributed_qa_rank": 23,
26
  "average_rank": 1,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "mistralai/Mistral-Large-Instruct-2407",
3
+ "brainstorm": 0.545,
4
+ "open_qa": 0.588,
5
+ "closed_qa": 0.351,
6
+ "extract": 0.396,
7
+ "generation": 0.514,
8
+ "rewrite": 0.505,
9
+ "summarize": 0.453,
10
+ "classify": 0.445,
11
+ "reasoning_over_numerical_data": 0.485,
12
+ "multi-document_synthesis": 0.591,
13
+ "fact_checking_or_attributed_qa": 0.26,
14
+ "average": 0.4762,
15
  "brainstorm_rank": 1,
16
  "open_qa_rank": 24,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 1,
25
  "fact_checking_or_attributed_qa_rank": 23,
26
  "average_rank": 1,
27
+ "brainstorm_confi": "+4.1/-4.1",
28
+ "open_qa_confi": "+9.8/-9.8",
29
+ "closed_qa_confi": "+5.7/-5.7",
30
+ "extract_confi": "+5.7/-5.7",
31
+ "generation_confi": "+3.9/-4.0",
32
+ "rewrite_confi": "+3.6/-3.7",
33
+ "summarize_confi": "+6.9/-6.9",
34
+ "classify_confi": "+6.2/-5.7",
35
+ "reasoning_over_numerical_data_confi": "+3.7/-3.7",
36
+ "multi-document_synthesis_confi": "+4.4/-4.5",
37
+ "fact_checking_or_attributed_qa_confi": "+4.1/-3.9",
38
+ "average_confi": "+1.47/-1.47"
39
  }
temperature=0.0/Mistral-Small-Instruct-2409.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "mistralai/Mistral-Small-Instruct-2409",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.3,
6
- "extract": 0.4,
7
- "generation": 0.4,
8
- "rewrite": 0.4,
9
- "summarize": 0.4,
10
- "classify": 0.4,
11
- "reasoning_over_numerical_data": 0.4,
12
- "multi-document_synthesis": 0.5,
13
- "fact_checking_or_attributed_qa": 0.3,
14
- "average": 0.42,
15
  "brainstorm_rank": 5,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 3,
25
  "fact_checking_or_attributed_qa_rank": 23,
26
  "average_rank": 5,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "mistralai/Mistral-Small-Instruct-2409",
3
+ "brainstorm": 0.47,
4
+ "open_qa": 0.676,
5
+ "closed_qa": 0.322,
6
+ "extract": 0.364,
7
+ "generation": 0.424,
8
+ "rewrite": 0.449,
9
+ "summarize": 0.401,
10
+ "classify": 0.393,
11
+ "reasoning_over_numerical_data": 0.397,
12
+ "multi-document_synthesis": 0.547,
13
+ "fact_checking_or_attributed_qa": 0.258,
14
+ "average": 0.4221,
15
  "brainstorm_rank": 5,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 3,
25
  "fact_checking_or_attributed_qa_rank": 23,
26
  "average_rank": 5,
27
+ "brainstorm_confi": "+4.0/-4.2",
28
+ "open_qa_confi": "+8.8/-8.8",
29
+ "closed_qa_confi": "+5.9/-5.4",
30
+ "extract_confi": "+5.7/-5.9",
31
+ "generation_confi": "+3.9/-3.9",
32
+ "rewrite_confi": "+3.6/-3.7",
33
+ "summarize_confi": "+6.7/-6.7",
34
+ "classify_confi": "+6.2/-5.7",
35
+ "reasoning_over_numerical_data_confi": "+3.8/-3.8",
36
+ "multi-document_synthesis_confi": "+4.4/-4.6",
37
+ "fact_checking_or_attributed_qa_confi": "+4.1/-3.9",
38
+ "average_confi": "+1.43/-1.48"
39
  }
temperature=0.0/OLMo-7B-0724-Instruct-hf.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "allenai/OLMo-7B-SFT",
3
- "brainstorm": 0.1,
4
- "open_qa": 0.4,
5
- "closed_qa": 0.1,
6
- "extract": 0.1,
7
- "generation": 0.1,
8
- "rewrite": 0.1,
9
- "summarize": 0.1,
10
- "classify": 0.1,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.1,
13
- "fact_checking_or_attributed_qa": 0.2,
14
- "average": 0.09,
15
  "brainstorm_rank": 17,
16
  "open_qa_rank": 27,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 16,
25
  "fact_checking_or_attributed_qa_rank": 26,
26
  "average_rank": 25,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "allenai/OLMo-7B-SFT",
3
+ "brainstorm": 0.089,
4
+ "open_qa": 0.363,
5
+ "closed_qa": 0.057,
6
+ "extract": 0.052,
7
+ "generation": 0.059,
8
+ "rewrite": 0.051,
9
+ "summarize": 0.057,
10
+ "classify": 0.08,
11
+ "reasoning_over_numerical_data": 0.05,
12
+ "multi-document_synthesis": 0.073,
13
+ "fact_checking_or_attributed_qa": 0.227,
14
+ "average": 0.0883,
15
  "brainstorm_rank": 17,
16
  "open_qa_rank": 27,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 16,
25
  "fact_checking_or_attributed_qa_rank": 26,
26
  "average_rank": 25,
27
+ "brainstorm_confi": "+2.3/-2.2",
28
+ "open_qa_confi": "+9.8/-8.8",
29
+ "closed_qa_confi": "+3.2/-2.7",
30
+ "extract_confi": "+3.2/-2.7",
31
+ "generation_confi": "+1.9/-1.8",
32
+ "rewrite_confi": "+1.7/-1.6",
33
+ "summarize_confi": "+3.5/-3.0",
34
+ "classify_confi": "+3.5/-3.2",
35
+ "reasoning_over_numerical_data_confi": "+1.8/-1.7",
36
+ "multi-document_synthesis_confi": "+2.4/-2.2",
37
+ "fact_checking_or_attributed_qa_confi": "+3.9/-3.9",
38
+ "average_confi": "+0.85/-0.81"
39
  }
temperature=0.0/OLMo-7B-SFT-hf.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "allenai/OLMo-7B-SFT",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.8,
5
- "closed_qa": 0.1,
6
- "extract": 0.1,
7
- "generation": 0.1,
8
- "rewrite": 0.0,
9
- "summarize": 0.0,
10
- "classify": 0.2,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.0,
13
  "fact_checking_or_attributed_qa": 0.5,
14
- "average": 0.12,
15
  "brainstorm_rank": 21,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 22,
25
  "fact_checking_or_attributed_qa_rank": 5,
26
  "average_rank": 18,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "allenai/OLMo-7B-SFT",
3
+ "brainstorm": 0.012,
4
+ "open_qa": 0.804,
5
+ "closed_qa": 0.134,
6
+ "extract": 0.069,
7
+ "generation": 0.067,
8
+ "rewrite": 0.044,
9
+ "summarize": 0.015,
10
+ "classify": 0.224,
11
+ "reasoning_over_numerical_data": 0.047,
12
+ "multi-document_synthesis": 0.024,
13
  "fact_checking_or_attributed_qa": 0.5,
14
+ "average": 0.1211,
15
  "brainstorm_rank": 21,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 22,
25
  "fact_checking_or_attributed_qa_rank": 5,
26
  "average_rank": 18,
27
+ "brainstorm_confi": "+1.0/-0.8",
28
+ "open_qa_confi": "+6.9/-7.8",
29
+ "closed_qa_confi": "+4.2/-4.2",
30
+ "extract_confi": "+3.2/-3.0",
31
+ "generation_confi": "+2.0/-1.8",
32
+ "rewrite_confi": "+1.5/-1.4",
33
+ "summarize_confi": "+1.5/-1.2",
34
+ "classify_confi": "+5.7/-5.5",
35
+ "reasoning_over_numerical_data_confi": "+1.9/-1.7",
36
+ "multi-document_synthesis_confi": "+1.5/-1.3",
37
+ "fact_checking_or_attributed_qa_confi": "+4.3/-4.5",
38
+ "average_confi": "+0.99/-0.96"
39
  }
temperature=0.0/Phi-3-medium-4k-instruct.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "microsoft/Phi-3-medium-4k-instruct",
3
- "brainstorm": 0.3,
4
- "open_qa": 0.8,
5
- "closed_qa": 0.3,
6
- "extract": 0.2,
7
- "generation": 0.3,
8
- "rewrite": 0.3,
9
- "summarize": 0.2,
10
- "classify": 0.5,
11
- "reasoning_over_numerical_data": 0.4,
12
- "multi-document_synthesis": 0.2,
13
- "fact_checking_or_attributed_qa": 0.6,
14
- "average": 0.33,
15
  "brainstorm_rank": 10,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 11,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 9,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "microsoft/Phi-3-medium-4k-instruct",
3
+ "brainstorm": 0.277,
4
+ "open_qa": 0.814,
5
+ "closed_qa": 0.349,
6
+ "extract": 0.243,
7
+ "generation": 0.263,
8
+ "rewrite": 0.281,
9
+ "summarize": 0.24,
10
+ "classify": 0.453,
11
+ "reasoning_over_numerical_data": 0.372,
12
+ "multi-document_synthesis": 0.172,
13
+ "fact_checking_or_attributed_qa": 0.602,
14
+ "average": 0.3328,
15
  "brainstorm_rank": 10,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 11,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 9,
27
+ "brainstorm_confi": "+3.6/-3.5",
28
+ "open_qa_confi": "+7.4/-7.8",
29
+ "closed_qa_confi": "+5.9/-5.7",
30
+ "extract_confi": "+5.0/-5.2",
31
+ "generation_confi": "+3.4/-3.5",
32
+ "rewrite_confi": "+3.4/-3.4",
33
+ "summarize_confi": "+5.9/-5.7",
34
+ "classify_confi": "+6.5/-6.5",
35
+ "reasoning_over_numerical_data_confi": "+3.7/-3.6",
36
+ "multi-document_synthesis_confi": "+3.5/-3.5",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.3",
38
+ "average_confi": "+1.35/-1.4"
39
  }
temperature=0.0/Qwen1.5-110B-Chat.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "Qwen/Qwen1.5-110B-Chat",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.8,
5
- "closed_qa": 0.3,
6
- "extract": 0.3,
7
- "generation": 0.4,
8
- "rewrite": 0.4,
9
- "summarize": 0.4,
10
- "classify": 0.5,
11
- "reasoning_over_numerical_data": 0.4,
12
- "multi-document_synthesis": 0.5,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.43,
15
  "brainstorm_rank": 5,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 4,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 1,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.02"
39
  }
 
1
  {
2
  "path": "Qwen/Qwen1.5-110B-Chat",
3
+ "brainstorm": 0.477,
4
+ "open_qa": 0.804,
5
+ "closed_qa": 0.314,
6
+ "extract": 0.304,
7
+ "generation": 0.439,
8
+ "rewrite": 0.39,
9
+ "summarize": 0.371,
10
+ "classify": 0.495,
11
+ "reasoning_over_numerical_data": 0.388,
12
+ "multi-document_synthesis": 0.517,
13
+ "fact_checking_or_attributed_qa": 0.398,
14
+ "average": 0.4297,
15
  "brainstorm_rank": 5,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 4,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 1,
27
+ "brainstorm_confi": "+3.9/-4.0",
28
+ "open_qa_confi": "+6.9/-7.8",
29
+ "closed_qa_confi": "+5.9/-5.7",
30
+ "extract_confi": "+5.9/-5.7",
31
+ "generation_confi": "+4.1/-4.0",
32
+ "rewrite_confi": "+3.6/-3.6",
33
+ "summarize_confi": "+6.9/-6.7",
34
+ "classify_confi": "+6.2/-6.5",
35
+ "reasoning_over_numerical_data_confi": "+3.7/-3.7",
36
+ "multi-document_synthesis_confi": "+4.7/-4.6",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.5",
38
+ "average_confi": "+1.47/-1.43"
39
  }
temperature=0.0/Qwen2-72B-Instruct.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
  "path": "Qwen/Qwen2-72B-Instruct",
3
- "brainstorm": 0.4,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.2,
6
- "extract": 0.3,
7
- "generation": 0.3,
8
- "rewrite": 0.3,
9
- "summarize": 0.2,
10
  "classify": 0.4,
11
- "reasoning_over_numerical_data": 0.4,
12
- "multi-document_synthesis": 0.3,
13
- "fact_checking_or_attributed_qa": 0.5,
14
  "average": 0.35,
15
  "brainstorm_rank": 7,
16
  "open_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 8,
25
  "fact_checking_or_attributed_qa_rank": 9,
26
  "average_rank": 6,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "Qwen/Qwen2-72B-Instruct",
3
+ "brainstorm": 0.424,
4
+ "open_qa": 0.716,
5
+ "closed_qa": 0.248,
6
+ "extract": 0.324,
7
+ "generation": 0.317,
8
+ "rewrite": 0.291,
9
+ "summarize": 0.153,
10
  "classify": 0.4,
11
+ "reasoning_over_numerical_data": 0.384,
12
+ "multi-document_synthesis": 0.264,
13
+ "fact_checking_or_attributed_qa": 0.472,
14
  "average": 0.35,
15
  "brainstorm_rank": 7,
16
  "open_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 8,
25
  "fact_checking_or_attributed_qa_rank": 9,
26
  "average_rank": 6,
27
+ "brainstorm_confi": "+4.0/-3.8",
28
+ "open_qa_confi": "+8.3/-8.8",
29
+ "closed_qa_confi": "+5.4/-5.2",
30
+ "extract_confi": "+5.7/-5.4",
31
+ "generation_confi": "+3.8/-3.7",
32
+ "rewrite_confi": "+3.3/-3.3",
33
+ "summarize_confi": "+4.7/-4.7",
34
+ "classify_confi": "+6.7/-6.5",
35
+ "reasoning_over_numerical_data_confi": "+3.8/-3.7",
36
+ "multi-document_synthesis_confi": "+4.1/-4.1",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.5",
38
+ "average_confi": "+1.36/-1.35"
39
  }
temperature=0.0/Qwen2.5-72B-Instruct.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "Qwen/Qwen2.5-72B-Instruct",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.3,
6
- "extract": 0.4,
7
- "generation": 0.5,
8
- "rewrite": 0.5,
9
- "summarize": 0.4,
10
- "classify": 0.4,
11
- "reasoning_over_numerical_data": 0.5,
12
- "multi-document_synthesis": 0.6,
13
- "fact_checking_or_attributed_qa": 0.2,
14
- "average": 0.45,
15
  "brainstorm_rank": 2,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 1,
25
  "fact_checking_or_attributed_qa_rank": 26,
26
  "average_rank": 1,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "Qwen/Qwen2.5-72B-Instruct",
3
+ "brainstorm": 0.514,
4
+ "open_qa": 0.716,
5
+ "closed_qa": 0.302,
6
+ "extract": 0.408,
7
+ "generation": 0.477,
8
+ "rewrite": 0.46,
9
+ "summarize": 0.364,
10
+ "classify": 0.435,
11
+ "reasoning_over_numerical_data": 0.456,
12
+ "multi-document_synthesis": 0.575,
13
+ "fact_checking_or_attributed_qa": 0.234,
14
+ "average": 0.4476,
15
  "brainstorm_rank": 2,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 1,
25
  "fact_checking_or_attributed_qa_rank": 26,
26
  "average_rank": 1,
27
+ "brainstorm_confi": "+4.0/-4.0",
28
+ "open_qa_confi": "+8.8/-8.8",
29
+ "closed_qa_confi": "+5.9/-5.7",
30
+ "extract_confi": "+5.9/-5.7",
31
+ "generation_confi": "+4.1/-4.1",
32
+ "rewrite_confi": "+3.6/-3.6",
33
+ "summarize_confi": "+6.7/-6.7",
34
+ "classify_confi": "+6.0/-6.2",
35
+ "reasoning_over_numerical_data_confi": "+3.8/-3.8",
36
+ "multi-document_synthesis_confi": "+4.5/-4.3",
37
+ "fact_checking_or_attributed_qa_confi": "+3.9/-3.9",
38
+ "average_confi": "+1.46/-1.47"
39
  }
temperature=0.0/WizardLM-13B-V1.2.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "WizardLMTeam/WizardLM-13B-V1.2",
3
- "brainstorm": 0.2,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.2,
6
- "extract": 0.2,
7
- "generation": 0.2,
8
- "rewrite": 0.1,
9
- "summarize": 0.1,
10
- "classify": 0.3,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.1,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.2,
15
  "brainstorm_rank": 12,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 13,
25
  "fact_checking_or_attributed_qa_rank": 9,
26
  "average_rank": 13,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "WizardLMTeam/WizardLM-13B-V1.2",
3
+ "brainstorm": 0.169,
4
+ "open_qa": 0.632,
5
+ "closed_qa": 0.228,
6
+ "extract": 0.171,
7
+ "generation": 0.158,
8
+ "rewrite": 0.147,
9
+ "summarize": 0.077,
10
+ "classify": 0.336,
11
+ "reasoning_over_numerical_data": 0.097,
12
+ "multi-document_synthesis": 0.113,
13
+ "fact_checking_or_attributed_qa": 0.439,
14
+ "average": 0.1956,
15
  "brainstorm_rank": 12,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 13,
25
  "fact_checking_or_attributed_qa_rank": 9,
26
  "average_rank": 13,
27
+ "brainstorm_confi": "+2.9/-2.9",
28
+ "open_qa_confi": "+9.3/-9.3",
29
+ "closed_qa_confi": "+5.4/-5.0",
30
+ "extract_confi": "+5.0/-4.5",
31
+ "generation_confi": "+3.0/-2.8",
32
+ "rewrite_confi": "+2.6/-2.6",
33
+ "summarize_confi": "+3.7/-3.2",
34
+ "classify_confi": "+6.2/-6.2",
35
+ "reasoning_over_numerical_data_confi": "+2.4/-2.2",
36
+ "multi-document_synthesis_confi": "+3.1/-2.9",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.5",
38
+ "average_confi": "+1.19/-1.17"
39
  }
temperature=0.0/Yi-1.5-34B-Chat.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "01-ai/Yi-1.5-34B-Chat",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.3,
6
- "extract": 0.3,
7
- "generation": 0.4,
8
- "rewrite": 0.3,
9
- "summarize": 0.2,
10
- "classify": 0.4,
11
- "reasoning_over_numerical_data": 0.3,
12
- "multi-document_synthesis": 0.4,
13
- "fact_checking_or_attributed_qa": 0.3,
14
- "average": 0.35,
15
  "brainstorm_rank": 2,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 6,
25
  "fact_checking_or_attributed_qa_rank": 23,
26
  "average_rank": 6,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "01-ai/Yi-1.5-34B-Chat",
3
+ "brainstorm": 0.497,
4
+ "open_qa": 0.716,
5
+ "closed_qa": 0.252,
6
+ "extract": 0.27,
7
+ "generation": 0.362,
8
+ "rewrite": 0.301,
9
+ "summarize": 0.215,
10
+ "classify": 0.358,
11
+ "reasoning_over_numerical_data": 0.318,
12
+ "multi-document_synthesis": 0.411,
13
+ "fact_checking_or_attributed_qa": 0.288,
14
+ "average": 0.3542,
15
  "brainstorm_rank": 2,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 6,
25
  "fact_checking_or_attributed_qa_rank": 23,
26
  "average_rank": 6,
27
+ "brainstorm_confi": "+4.0/-4.1",
28
+ "open_qa_confi": "+8.3/-8.8",
29
+ "closed_qa_confi": "+5.4/-5.2",
30
+ "extract_confi": "+5.4/-5.4",
31
+ "generation_confi": "+3.8/-3.8",
32
+ "rewrite_confi": "+3.4/-3.4",
33
+ "summarize_confi": "+5.7/-5.2",
34
+ "classify_confi": "+6.5/-6.0",
35
+ "reasoning_over_numerical_data_confi": "+3.7/-3.6",
36
+ "multi-document_synthesis_confi": "+4.6/-4.4",
37
+ "fact_checking_or_attributed_qa_confi": "+4.1/-4.1",
38
+ "average_confi": "+1.44/-1.43"
39
  }
temperature=0.0/dolly-v2-12b.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "databricks/dolly-v2-7b",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.8,
5
- "closed_qa": 0.1,
6
- "extract": 0.1,
7
- "generation": 0.0,
8
- "rewrite": 0.0,
9
  "summarize": 0.0,
10
- "classify": 0.1,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.09,
15
  "brainstorm_rank": 24,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 24,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 25,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
  "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "databricks/dolly-v2-7b",
3
+ "brainstorm": 0.003,
4
+ "open_qa": 0.755,
5
+ "closed_qa": 0.124,
6
+ "extract": 0.074,
7
+ "generation": 0.028,
8
+ "rewrite": 0.02,
9
  "summarize": 0.0,
10
+ "classify": 0.067,
11
+ "reasoning_over_numerical_data": 0.033,
12
+ "multi-document_synthesis": 0.009,
13
+ "fact_checking_or_attributed_qa": 0.396,
14
+ "average": 0.0864,
15
  "brainstorm_rank": 24,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 24,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 25,
27
+ "brainstorm_confi": "+0.5/-0.3",
28
+ "open_qa_confi": "+7.8/-8.8",
29
+ "closed_qa_confi": "+4.7/-4.5",
30
+ "extract_confi": "+3.5/-3.2",
31
+ "generation_confi": "+1.3/-1.3",
32
+ "rewrite_confi": "+1.1/-1.0",
33
  "summarize_confi": "+0.0/-0.0",
34
+ "classify_confi": "+3.7/-3.2",
35
+ "reasoning_over_numerical_data_confi": "+1.5/-1.4",
36
+ "multi-document_synthesis_confi": "+0.9/-0.7",
37
+ "fact_checking_or_attributed_qa_confi": "+4.3/-4.5",
38
+ "average_confi": "+0.85/-0.85"
39
  }
temperature=0.0/dolly-v2-7b.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "databricks/dolly-v2-12b",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.1,
6
- "extract": 0.0,
7
- "generation": 0.0,
8
- "rewrite": 0.0,
9
  "summarize": 0.0,
10
- "classify": 0.1,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.08,
15
  "brainstorm_rank": 24,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 29,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 25,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
  "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "databricks/dolly-v2-12b",
3
+ "brainstorm": 0.002,
4
+ "open_qa": 0.706,
5
+ "closed_qa": 0.094,
6
+ "extract": 0.05,
7
+ "generation": 0.02,
8
+ "rewrite": 0.013,
9
  "summarize": 0.0,
10
+ "classify": 0.062,
11
+ "reasoning_over_numerical_data": 0.036,
12
+ "multi-document_synthesis": 0.002,
13
+ "fact_checking_or_attributed_qa": 0.42,
14
+ "average": 0.0822,
15
  "brainstorm_rank": 24,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 29,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 25,
27
+ "brainstorm_confi": "+0.3/-0.2",
28
+ "open_qa_confi": "+8.8/-8.8",
29
+ "closed_qa_confi": "+4.2/-3.7",
30
+ "extract_confi": "+3.0/-2.5",
31
+ "generation_confi": "+1.2/-1.0",
32
+ "rewrite_confi": "+0.9/-0.7",
33
  "summarize_confi": "+0.0/-0.0",
34
+ "classify_confi": "+3.5/-3.2",
35
+ "reasoning_over_numerical_data_confi": "+1.7/-1.5",
36
+ "multi-document_synthesis_confi": "+0.4/-0.2",
37
+ "fact_checking_or_attributed_qa_confi": "+4.3/-4.5",
38
+ "average_confi": "+0.86/-0.81"
39
  }
temperature=0.0/gpt4all-13b-snoozy.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "nomic-ai/gpt4all-13b-snoozy",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.9,
5
- "closed_qa": 0.0,
6
- "extract": 0.1,
7
- "generation": 0.1,
8
- "rewrite": 0.1,
9
- "summarize": 0.0,
10
- "classify": 0.1,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.1,
15
  "brainstorm_rank": 24,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 24,
25
  "fact_checking_or_attributed_qa_rank": 21,
26
  "average_rank": 22,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.0/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "nomic-ai/gpt4all-13b-snoozy",
3
+ "brainstorm": 0.007,
4
+ "open_qa": 0.902,
5
+ "closed_qa": 0.037,
6
+ "extract": 0.062,
7
+ "generation": 0.064,
8
+ "rewrite": 0.057,
9
+ "summarize": 0.01,
10
+ "classify": 0.117,
11
+ "reasoning_over_numerical_data": 0.051,
12
+ "multi-document_synthesis": 0.013,
13
+ "fact_checking_or_attributed_qa": 0.359,
14
+ "average": 0.0985,
15
  "brainstorm_rank": 24,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 24,
25
  "fact_checking_or_attributed_qa_rank": 21,
26
  "average_rank": 22,
27
+ "brainstorm_confi": "+0.7/-0.5",
28
+ "open_qa_confi": "+5.4/-5.9",
29
+ "closed_qa_confi": "+2.5/-2.2",
30
+ "extract_confi": "+3.2/-3.0",
31
+ "generation_confi": "+1.9/-1.9",
32
+ "rewrite_confi": "+1.7/-1.6",
33
+ "summarize_confi": "+1.5/-1.0",
34
+ "classify_confi": "+4.7/-4.2",
35
+ "reasoning_over_numerical_data_confi": "+2.0/-1.8",
36
+ "multi-document_synthesis_confi": "+1.1/-0.9",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.3",
38
+ "average_confi": "+0.89/-0.87"
39
  }
temperature=0.0/koala-13B-HF.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "TheBloke/koala-13B-HF",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.2,
6
- "extract": 0.1,
7
- "generation": 0.0,
8
- "rewrite": 0.1,
9
- "summarize": 0.0,
10
- "classify": 0.2,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.1,
15
  "brainstorm_rank": 21,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 22,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 22,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "TheBloke/koala-13B-HF",
3
+ "brainstorm": 0.01,
4
+ "open_qa": 0.701,
5
+ "closed_qa": 0.158,
6
+ "extract": 0.084,
7
+ "generation": 0.043,
8
+ "rewrite": 0.05,
9
+ "summarize": 0.01,
10
+ "classify": 0.164,
11
+ "reasoning_over_numerical_data": 0.052,
12
+ "multi-document_synthesis": 0.022,
13
+ "fact_checking_or_attributed_qa": 0.398,
14
+ "average": 0.1044,
15
  "brainstorm_rank": 21,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 22,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 22,
27
+ "brainstorm_confi": "+0.8/-0.7",
28
+ "open_qa_confi": "+8.8/-8.8",
29
+ "closed_qa_confi": "+5.0/-4.5",
30
+ "extract_confi": "+3.2/-3.0",
31
+ "generation_confi": "+1.7/-1.5",
32
+ "rewrite_confi": "+1.7/-1.6",
33
+ "summarize_confi": "+1.2/-1.0",
34
+ "classify_confi": "+5.0/-4.7",
35
+ "reasoning_over_numerical_data_confi": "+2.1/-1.9",
36
+ "multi-document_synthesis_confi": "+1.3/-1.3",
37
+ "fact_checking_or_attributed_qa_confi": "+4.3/-4.3",
38
+ "average_confi": "+0.9/-0.9"
39
  }
temperature=0.0/koala-7B-HF.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "TheBloke/koala-7B-HF",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.1,
6
- "extract": 0.1,
7
- "generation": 0.0,
8
- "rewrite": 0.0,
9
  "summarize": 0.0,
10
- "classify": 0.1,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.09,
15
  "brainstorm_rank": 24,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 24,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 25,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
  "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "TheBloke/koala-7B-HF",
3
+ "brainstorm": 0.004,
4
+ "open_qa": 0.657,
5
+ "closed_qa": 0.082,
6
+ "extract": 0.079,
7
+ "generation": 0.033,
8
+ "rewrite": 0.034,
9
  "summarize": 0.0,
10
+ "classify": 0.097,
11
+ "reasoning_over_numerical_data": 0.031,
12
+ "multi-document_synthesis": 0.013,
13
+ "fact_checking_or_attributed_qa": 0.387,
14
+ "average": 0.0861,
15
  "brainstorm_rank": 24,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 24,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 25,
27
+ "brainstorm_confi": "+0.6/-0.4",
28
+ "open_qa_confi": "+8.8/-8.8",
29
+ "closed_qa_confi": "+3.7/-3.2",
30
+ "extract_confi": "+3.7/-3.2",
31
+ "generation_confi": "+1.5/-1.3",
32
+ "rewrite_confi": "+1.4/-1.2",
33
  "summarize_confi": "+0.0/-0.0",
34
+ "classify_confi": "+4.2/-3.7",
35
+ "reasoning_over_numerical_data_confi": "+1.6/-1.4",
36
+ "multi-document_synthesis_confi": "+1.1/-0.9",
37
+ "fact_checking_or_attributed_qa_confi": "+4.3/-4.5",
38
+ "average_confi": "+0.85/-0.81"
39
  }
temperature=0.0/mpt-7b-chat.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "mosaicml/mpt-7b-chat",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.1,
6
- "extract": 0.0,
7
- "generation": 0.1,
8
- "rewrite": 0.0,
9
- "summarize": 0.0,
10
- "classify": 0.2,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.5,
14
- "average": 0.11,
15
  "brainstorm_rank": 24,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 24,
25
  "fact_checking_or_attributed_qa_rank": 9,
26
  "average_rank": 22,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "mosaicml/mpt-7b-chat",
3
+ "brainstorm": 0.005,
4
+ "open_qa": 0.73,
5
+ "closed_qa": 0.121,
6
+ "extract": 0.037,
7
+ "generation": 0.05,
8
+ "rewrite": 0.039,
9
+ "summarize": 0.015,
10
+ "classify": 0.211,
11
+ "reasoning_over_numerical_data": 0.041,
12
+ "multi-document_synthesis": 0.013,
13
+ "fact_checking_or_attributed_qa": 0.474,
14
+ "average": 0.1078,
15
  "brainstorm_rank": 24,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 24,
25
  "fact_checking_or_attributed_qa_rank": 9,
26
  "average_rank": 22,
27
+ "brainstorm_confi": "+0.7/-0.5",
28
+ "open_qa_confi": "+8.3/-8.8",
29
+ "closed_qa_confi": "+4.5/-3.7",
30
+ "extract_confi": "+2.5/-2.2",
31
+ "generation_confi": "+1.8/-1.7",
32
+ "rewrite_confi": "+1.4/-1.4",
33
+ "summarize_confi": "+1.7/-1.2",
34
+ "classify_confi": "+5.7/-5.5",
35
+ "reasoning_over_numerical_data_confi": "+1.8/-1.6",
36
+ "multi-document_synthesis_confi": "+1.1/-0.9",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.8",
38
+ "average_confi": "+0.94/-0.93"
39
  }
temperature=0.0/oasst-sft-1-pythia-12b.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "OpenAssistant/oasst-sft-1-pythia-12b",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.0,
6
- "extract": 0.0,
7
- "generation": 0.0,
8
- "rewrite": 0.0,
9
  "summarize": 0.0,
10
- "classify": 0.0,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.2,
14
- "average": 0.05,
15
  "brainstorm_rank": 24,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 24,
25
  "fact_checking_or_attributed_qa_rank": 26,
26
  "average_rank": 29,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
  "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "OpenAssistant/oasst-sft-1-pythia-12b",
3
+ "brainstorm": 0.002,
4
+ "open_qa": 0.623,
5
+ "closed_qa": 0.02,
6
+ "extract": 0.015,
7
+ "generation": 0.017,
8
+ "rewrite": 0.006,
9
  "summarize": 0.0,
10
+ "classify": 0.032,
11
+ "reasoning_over_numerical_data": 0.024,
12
+ "multi-document_synthesis": 0.009,
13
+ "fact_checking_or_attributed_qa": 0.232,
14
+ "average": 0.0507,
15
  "brainstorm_rank": 24,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 24,
25
  "fact_checking_or_attributed_qa_rank": 26,
26
  "average_rank": 29,
27
+ "brainstorm_confi": "+0.3/-0.2",
28
+ "open_qa_confi": "+9.3/-9.3",
29
+ "closed_qa_confi": "+2.0/-1.5",
30
+ "extract_confi": "+1.7/-1.2",
31
+ "generation_confi": "+1.2/-1.0",
32
+ "rewrite_confi": "+0.6/-0.4",
33
  "summarize_confi": "+0.0/-0.0",
34
+ "classify_confi": "+2.5/-2.2",
35
+ "reasoning_over_numerical_data_confi": "+1.5/-1.1",
36
+ "multi-document_synthesis_confi": "+0.9/-0.7",
37
+ "fact_checking_or_attributed_qa_confi": "+3.9/-3.7",
38
+ "average_confi": "+0.65/-0.62"
39
  }
temperature=0.0/tulu-2-dpo-13b.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "allenai/OLMo-7B-SFT",
3
- "brainstorm": 0.1,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.2,
6
- "extract": 0.1,
7
- "generation": 0.1,
8
- "rewrite": 0.2,
9
- "summarize": 0.1,
10
- "classify": 0.2,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.1,
13
- "fact_checking_or_attributed_qa": 0.5,
14
- "average": 0.18,
15
  "brainstorm_rank": 17,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 13,
25
  "fact_checking_or_attributed_qa_rank": 5,
26
  "average_rank": 16,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "allenai/OLMo-7B-SFT",
3
+ "brainstorm": 0.093,
4
+ "open_qa": 0.686,
5
+ "closed_qa": 0.178,
6
+ "extract": 0.136,
7
+ "generation": 0.133,
8
+ "rewrite": 0.156,
9
+ "summarize": 0.092,
10
+ "classify": 0.249,
11
+ "reasoning_over_numerical_data": 0.109,
12
+ "multi-document_synthesis": 0.102,
13
+ "fact_checking_or_attributed_qa": 0.502,
14
+ "average": 0.184,
15
  "brainstorm_rank": 17,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 13,
25
  "fact_checking_or_attributed_qa_rank": 5,
26
  "average_rank": 16,
27
+ "brainstorm_confi": "+2.3/-2.3",
28
+ "open_qa_confi": "+8.8/-8.8",
29
+ "closed_qa_confi": "+5.0/-5.0",
30
+ "extract_confi": "+4.5/-4.2",
31
+ "generation_confi": "+2.7/-2.5",
32
+ "rewrite_confi": "+2.6/-2.6",
33
+ "summarize_confi": "+4.0/-3.7",
34
+ "classify_confi": "+5.7/-5.2",
35
+ "reasoning_over_numerical_data_confi": "+2.5/-2.4",
36
+ "multi-document_synthesis_confi": "+2.9/-2.6",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.8",
38
+ "average_confi": "+1.16/-1.1"
39
  }
temperature=0.0/tulu-2-dpo-70b.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "allenai/tulu-2-dpo-70b",
3
- "brainstorm": 0.2,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.3,
6
- "extract": 0.2,
7
- "generation": 0.2,
8
- "rewrite": 0.2,
9
- "summarize": 0.1,
10
- "classify": 0.3,
11
- "reasoning_over_numerical_data": 0.2,
12
- "multi-document_synthesis": 0.2,
13
- "fact_checking_or_attributed_qa": 0.6,
14
- "average": 0.26,
15
  "brainstorm_rank": 12,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 11,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 10,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "allenai/tulu-2-dpo-70b",
3
+ "brainstorm": 0.175,
4
+ "open_qa": 0.608,
5
+ "closed_qa": 0.265,
6
+ "extract": 0.21,
7
+ "generation": 0.229,
8
+ "rewrite": 0.215,
9
+ "summarize": 0.119,
10
+ "classify": 0.308,
11
+ "reasoning_over_numerical_data": 0.232,
12
+ "multi-document_synthesis": 0.181,
13
+ "fact_checking_or_attributed_qa": 0.556,
14
+ "average": 0.2583,
15
  "brainstorm_rank": 12,
16
  "open_qa_rank": 12,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 11,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 10,
27
+ "brainstorm_confi": "+3.0/-3.0",
28
+ "open_qa_confi": "+9.8/-8.8",
29
+ "closed_qa_confi": "+5.7/-5.7",
30
+ "extract_confi": "+5.0/-4.7",
31
+ "generation_confi": "+3.4/-3.5",
32
+ "rewrite_confi": "+3.1/-2.9",
33
+ "summarize_confi": "+4.7/-4.2",
34
+ "classify_confi": "+6.0/-5.7",
35
+ "reasoning_over_numerical_data_confi": "+3.4/-3.3",
36
+ "multi-document_synthesis_confi": "+3.5/-3.5",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.6",
38
+ "average_confi": "+1.27/-1.28"
39
  }
temperature=0.0/tulu-2-dpo-7b.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "allenai/tulu-2-dpo-13b",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.1,
6
- "extract": 0.1,
7
  "generation": 0.1,
8
- "rewrite": 0.1,
9
- "summarize": 0.1,
10
- "classify": 0.2,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.1,
13
- "fact_checking_or_attributed_qa": 0.5,
14
- "average": 0.14,
15
  "brainstorm_rank": 19,
16
  "open_qa_rank": 24,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 19,
25
  "fact_checking_or_attributed_qa_rank": 9,
26
  "average_rank": 18,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "allenai/tulu-2-dpo-13b",
3
+ "brainstorm": 0.049,
4
+ "open_qa": 0.559,
5
+ "closed_qa": 0.146,
6
+ "extract": 0.099,
7
  "generation": 0.1,
8
+ "rewrite": 0.099,
9
+ "summarize": 0.069,
10
+ "classify": 0.184,
11
+ "reasoning_over_numerical_data": 0.054,
12
+ "multi-document_synthesis": 0.06,
13
+ "fact_checking_or_attributed_qa": 0.465,
14
+ "average": 0.138,
15
  "brainstorm_rank": 19,
16
  "open_qa_rank": 24,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 19,
25
  "fact_checking_or_attributed_qa_rank": 9,
26
  "average_rank": 18,
27
+ "brainstorm_confi": "+1.8/-1.7",
28
+ "open_qa_confi": "+9.8/-9.8",
29
+ "closed_qa_confi": "+4.7/-4.2",
30
+ "extract_confi": "+4.0/-3.5",
31
+ "generation_confi": "+2.5/-2.3",
32
+ "rewrite_confi": "+2.3/-2.1",
33
+ "summarize_confi": "+3.5/-3.5",
34
+ "classify_confi": "+5.2/-4.7",
35
+ "reasoning_over_numerical_data_confi": "+2.0/-1.9",
36
+ "multi-document_synthesis_confi": "+2.2/-2.0",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.6",
38
+ "average_confi": "+1.05/-1.05"
39
  }
temperature=0.0/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm",
3
- "brainstorm": 0.4,
4
- "open_qa": 0.2,
5
- "closed_qa": 0.1,
6
- "extract": 0.1,
7
- "generation": 0.2,
8
- "rewrite": 0.2,
9
- "summarize": 0.2,
10
- "classify": 0.1,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.2,
13
- "fact_checking_or_attributed_qa": 0.2,
14
- "average": 0.19,
15
  "brainstorm_rank": 7,
16
  "open_qa_rank": 29,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 8,
25
  "fact_checking_or_attributed_qa_rank": 26,
26
  "average_rank": 15,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "allenai/tulu-v2.5-ppo-13b-uf-mean-70b-uf-rm",
3
+ "brainstorm": 0.377,
4
+ "open_qa": 0.216,
5
+ "closed_qa": 0.064,
6
+ "extract": 0.109,
7
+ "generation": 0.164,
8
+ "rewrite": 0.155,
9
+ "summarize": 0.161,
10
+ "classify": 0.124,
11
+ "reasoning_over_numerical_data": 0.117,
12
+ "multi-document_synthesis": 0.219,
13
+ "fact_checking_or_attributed_qa": 0.214,
14
+ "average": 0.1898,
15
  "brainstorm_rank": 7,
16
  "open_qa_rank": 29,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 8,
25
  "fact_checking_or_attributed_qa_rank": 26,
26
  "average_rank": 15,
27
+ "brainstorm_confi": "+3.9/-3.8",
28
+ "open_qa_confi": "+7.8/-7.8",
29
+ "closed_qa_confi": "+3.2/-3.0",
30
+ "extract_confi": "+4.2/-4.0",
31
+ "generation_confi": "+3.0/-2.8",
32
+ "rewrite_confi": "+2.8/-2.6",
33
+ "summarize_confi": "+5.2/-5.0",
34
+ "classify_confi": "+4.5/-4.0",
35
+ "reasoning_over_numerical_data_confi": "+2.7/-2.6",
36
+ "multi-document_synthesis_confi": "+3.8/-3.8",
37
+ "fact_checking_or_attributed_qa_confi": "+3.7/-3.7",
38
+ "average_confi": "+1.15/-1.16"
39
  }
temperature=0.0/vicuna-13b-v1.5.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "lmsys/vicuna-7b-v1.5",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.8,
5
- "closed_qa": 0.3,
6
- "extract": 0.1,
7
- "generation": 0.1,
8
- "rewrite": 0.1,
9
- "summarize": 0.1,
10
- "classify": 0.3,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.1,
13
- "fact_checking_or_attributed_qa": 0.6,
14
- "average": 0.17,
15
  "brainstorm_rank": 19,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 19,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 16,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "lmsys/vicuna-7b-v1.5",
3
+ "brainstorm": 0.041,
4
+ "open_qa": 0.814,
5
+ "closed_qa": 0.265,
6
+ "extract": 0.139,
7
+ "generation": 0.102,
8
+ "rewrite": 0.119,
9
+ "summarize": 0.079,
10
+ "classify": 0.256,
11
+ "reasoning_over_numerical_data": 0.106,
12
+ "multi-document_synthesis": 0.055,
13
+ "fact_checking_or_attributed_qa": 0.558,
14
+ "average": 0.1739,
15
  "brainstorm_rank": 19,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 19,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 16,
27
+ "brainstorm_confi": "+1.7/-1.5",
28
+ "open_qa_confi": "+7.4/-7.8",
29
+ "closed_qa_confi": "+5.7/-5.7",
30
+ "extract_confi": "+4.5/-4.2",
31
+ "generation_confi": "+2.5/-2.3",
32
+ "rewrite_confi": "+2.4/-2.4",
33
+ "summarize_confi": "+3.7/-3.2",
34
+ "classify_confi": "+6.0/-5.7",
35
+ "reasoning_over_numerical_data_confi": "+2.6/-2.4",
36
+ "multi-document_synthesis_confi": "+2.2/-2.0",
37
+ "fact_checking_or_attributed_qa_confi": "+4.3/-4.3",
38
+ "average_confi": "+1.15/-1.1"
39
  }
temperature=0.0/vicuna-7b-v1.5.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "lmsys/vicuna-7b-v1.5",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.8,
5
- "closed_qa": 0.2,
6
- "extract": 0.1,
7
- "generation": 0.1,
8
- "rewrite": 0.1,
9
- "summarize": 0.0,
10
- "classify": 0.2,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.5,
14
- "average": 0.14,
15
  "brainstorm_rank": 21,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 19,
25
  "fact_checking_or_attributed_qa_rank": 5,
26
  "average_rank": 18,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "lmsys/vicuna-7b-v1.5",
3
+ "brainstorm": 0.029,
4
+ "open_qa": 0.755,
5
+ "closed_qa": 0.203,
6
+ "extract": 0.106,
7
+ "generation": 0.082,
8
+ "rewrite": 0.07,
9
+ "summarize": 0.027,
10
+ "classify": 0.241,
11
+ "reasoning_over_numerical_data": 0.067,
12
+ "multi-document_synthesis": 0.042,
13
+ "fact_checking_or_attributed_qa": 0.528,
14
+ "average": 0.1428,
15
  "brainstorm_rank": 21,
16
  "open_qa_rank": 3,
17
  "closed_qa_rank": 3,
 
24
  "multi-document_synthesis_rank": 19,
25
  "fact_checking_or_attributed_qa_rank": 5,
26
  "average_rank": 18,
27
+ "brainstorm_confi": "+1.4/-1.3",
28
+ "open_qa_confi": "+7.8/-8.3",
29
+ "closed_qa_confi": "+5.2/-5.0",
30
+ "extract_confi": "+4.0/-3.7",
31
+ "generation_confi": "+2.2/-2.2",
32
+ "rewrite_confi": "+2.0/-1.9",
33
+ "summarize_confi": "+2.2/-1.7",
34
+ "classify_confi": "+5.7/-5.7",
35
+ "reasoning_over_numerical_data_confi": "+2.1/-2.1",
36
+ "multi-document_synthesis_confi": "+2.0/-1.8",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.5",
38
+ "average_confi": "+1.01/-1.03"
39
  }
temperature=1.0/Llama-2-13b-chat-hf.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "meta-llama/Llama-2-13b-chat-hf",
3
- "brainstorm": 0.2,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.2,
6
- "extract": 0.1,
7
- "generation": 0.2,
8
- "rewrite": 0.1,
9
- "summarize": 0.2,
10
- "classify": 0.3,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.1,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.2,
15
  "brainstorm_rank": 13,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 14,
25
  "fact_checking_or_attributed_qa_rank": 7,
26
  "average_rank": 13,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "meta-llama/Llama-2-13b-chat-hf",
3
+ "brainstorm": 0.21,
4
+ "open_qa": 0.593,
5
+ "closed_qa": 0.176,
6
+ "extract": 0.134,
7
+ "generation": 0.167,
8
+ "rewrite": 0.136,
9
+ "summarize": 0.188,
10
+ "classify": 0.323,
11
+ "reasoning_over_numerical_data": 0.122,
12
+ "multi-document_synthesis": 0.121,
13
+ "fact_checking_or_attributed_qa": 0.409,
14
+ "average": 0.2013,
15
  "brainstorm_rank": 13,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 14,
25
  "fact_checking_or_attributed_qa_rank": 7,
26
  "average_rank": 13,
27
+ "brainstorm_confi": "+3.3/-3.3",
28
+ "open_qa_confi": "+9.3/-9.8",
29
+ "closed_qa_confi": "+4.7/-4.5",
30
+ "extract_confi": "+4.5/-4.0",
31
+ "generation_confi": "+3.2/-3.0",
32
+ "rewrite_confi": "+2.6/-2.6",
33
+ "summarize_confi": "+5.2/-5.0",
34
+ "classify_confi": "+6.0/-5.7",
35
+ "reasoning_over_numerical_data_confi": "+2.7/-2.5",
36
+ "multi-document_synthesis_confi": "+3.1/-3.1",
37
+ "fact_checking_or_attributed_qa_confi": "+4.3/-4.3",
38
+ "average_confi": "+1.14/-1.17"
39
  }
temperature=1.0/Llama-2-70b-chat-hf.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "meta-llama/Llama-2-70b-chat-hf",
3
- "brainstorm": 0.2,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.3,
6
- "extract": 0.2,
7
- "generation": 0.2,
8
- "rewrite": 0.2,
9
- "summarize": 0.2,
10
- "classify": 0.3,
11
- "reasoning_over_numerical_data": 0.2,
12
- "multi-document_synthesis": 0.2,
13
- "fact_checking_or_attributed_qa": 0.5,
14
- "average": 0.24,
15
  "brainstorm_rank": 13,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 11,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 11,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "meta-llama/Llama-2-70b-chat-hf",
3
+ "brainstorm": 0.218,
4
+ "open_qa": 0.662,
5
+ "closed_qa": 0.312,
6
+ "extract": 0.173,
7
+ "generation": 0.197,
8
+ "rewrite": 0.183,
9
+ "summarize": 0.153,
10
+ "classify": 0.343,
11
+ "reasoning_over_numerical_data": 0.172,
12
+ "multi-document_synthesis": 0.166,
13
+ "fact_checking_or_attributed_qa": 0.488,
14
+ "average": 0.2431,
15
  "brainstorm_rank": 13,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 11,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 11,
27
+ "brainstorm_confi": "+3.3/-3.2",
28
+ "open_qa_confi": "+9.3/-9.3",
29
+ "closed_qa_confi": "+5.7/-5.9",
30
+ "extract_confi": "+5.0/-4.5",
31
+ "generation_confi": "+3.3/-3.2",
32
+ "rewrite_confi": "+2.9/-2.9",
33
+ "summarize_confi": "+5.0/-4.5",
34
+ "classify_confi": "+6.5/-6.5",
35
+ "reasoning_over_numerical_data_confi": "+3.1/-3.1",
36
+ "multi-document_synthesis_confi": "+3.5/-3.3",
37
+ "fact_checking_or_attributed_qa_confi": "+4.7/-4.5",
38
+ "average_confi": "+1.28/-1.27"
39
  }
temperature=1.0/Llama-2-7b-chat-hf.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "meta-llama/Llama-2-7b-chat-hf",
3
- "brainstorm": 0.2,
4
- "open_qa": 0.5,
5
  "closed_qa": 0.2,
6
- "extract": 0.1,
7
- "generation": 0.2,
8
- "rewrite": 0.1,
9
- "summarize": 0.1,
10
- "classify": 0.2,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.1,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.17,
15
  "brainstorm_rank": 13,
16
  "open_qa_rank": 21,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 14,
25
  "fact_checking_or_attributed_qa_rank": 12,
26
  "average_rank": 15,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "meta-llama/Llama-2-7b-chat-hf",
3
+ "brainstorm": 0.171,
4
+ "open_qa": 0.475,
5
  "closed_qa": 0.2,
6
+ "extract": 0.119,
7
+ "generation": 0.152,
8
+ "rewrite": 0.131,
9
+ "summarize": 0.121,
10
+ "classify": 0.187,
11
+ "reasoning_over_numerical_data": 0.088,
12
+ "multi-document_synthesis": 0.106,
13
+ "fact_checking_or_attributed_qa": 0.355,
14
+ "average": 0.169,
15
  "brainstorm_rank": 13,
16
  "open_qa_rank": 21,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 14,
25
  "fact_checking_or_attributed_qa_rank": 12,
26
  "average_rank": 15,
27
+ "brainstorm_confi": "+3.0/-2.9",
28
+ "open_qa_confi": "+9.8/-9.8",
29
+ "closed_qa_confi": "+5.0/-5.0",
30
+ "extract_confi": "+4.2/-4.0",
31
+ "generation_confi": "+3.0/-2.7",
32
+ "rewrite_confi": "+2.5/-2.4",
33
+ "summarize_confi": "+4.5/-4.2",
34
+ "classify_confi": "+5.0/-5.0",
35
+ "reasoning_over_numerical_data_confi": "+2.3/-2.2",
36
+ "multi-document_synthesis_confi": "+2.9/-2.9",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.3",
38
+ "average_confi": "+1.14/-1.09"
39
  }
temperature=1.0/Llama-3.1-70B-Instruct.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "meta-llama/Llama-3.1-70B-Instruct",
3
- "brainstorm": 0.4,
4
- "open_qa": 0.8,
5
- "closed_qa": 0.4,
6
- "extract": 0.4,
7
- "generation": 0.4,
8
- "rewrite": 0.4,
9
- "summarize": 0.4,
10
- "classify": 0.5,
11
- "reasoning_over_numerical_data": 0.5,
12
- "multi-document_synthesis": 0.5,
13
- "fact_checking_or_attributed_qa": 0.5,
14
- "average": 0.46,
15
  "brainstorm_rank": 7,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 4,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 2,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "meta-llama/Llama-3.1-70B-Instruct",
3
+ "brainstorm": 0.437,
4
+ "open_qa": 0.77,
5
+ "closed_qa": 0.408,
6
+ "extract": 0.369,
7
+ "generation": 0.431,
8
+ "rewrite": 0.424,
9
+ "summarize": 0.443,
10
+ "classify": 0.532,
11
+ "reasoning_over_numerical_data": 0.458,
12
+ "multi-document_synthesis": 0.481,
13
+ "fact_checking_or_attributed_qa": 0.498,
14
+ "average": 0.456,
15
  "brainstorm_rank": 7,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 4,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 2,
27
+ "brainstorm_confi": "+3.8/-3.9",
28
+ "open_qa_confi": "+7.8/-8.3",
29
+ "closed_qa_confi": "+5.9/-5.9",
30
+ "extract_confi": "+5.7/-5.7",
31
+ "generation_confi": "+4.0/-4.0",
32
+ "rewrite_confi": "+3.8/-3.6",
33
+ "summarize_confi": "+6.7/-6.7",
34
+ "classify_confi": "+6.0/-6.0",
35
+ "reasoning_over_numerical_data_confi": "+3.8/-3.8",
36
+ "multi-document_synthesis_confi": "+4.6/-4.6",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.6",
38
+ "average_confi": "+1.43/-1.46"
39
  }
temperature=1.0/Llama-3.1-8B-Instruct.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "meta-llama/Llama-3.1-8B-Instruct",
3
- "brainstorm": 0.4,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.3,
6
- "extract": 0.3,
7
- "generation": 0.3,
8
- "rewrite": 0.3,
9
  "summarize": 0.3,
10
- "classify": 0.4,
11
- "reasoning_over_numerical_data": 0.2,
12
- "multi-document_synthesis": 0.4,
13
- "fact_checking_or_attributed_qa": 0.3,
14
- "average": 0.33,
15
  "brainstorm_rank": 7,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 6,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 8,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "meta-llama/Llama-3.1-8B-Instruct",
3
+ "brainstorm": 0.402,
4
+ "open_qa": 0.652,
5
+ "closed_qa": 0.304,
6
+ "extract": 0.255,
7
+ "generation": 0.324,
8
+ "rewrite": 0.301,
9
  "summarize": 0.3,
10
+ "classify": 0.378,
11
+ "reasoning_over_numerical_data": 0.235,
12
+ "multi-document_synthesis": 0.411,
13
+ "fact_checking_or_attributed_qa": 0.316,
14
+ "average": 0.3334,
15
  "brainstorm_rank": 7,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 6,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 8,
27
+ "brainstorm_confi": "+3.9/-3.9",
28
+ "open_qa_confi": "+9.3/-9.3",
29
+ "closed_qa_confi": "+5.7/-5.7",
30
+ "extract_confi": "+5.4/-5.2",
31
+ "generation_confi": "+3.7/-3.7",
32
+ "rewrite_confi": "+3.5/-3.4",
33
+ "summarize_confi": "+6.2/-5.9",
34
+ "classify_confi": "+6.2/-6.0",
35
+ "reasoning_over_numerical_data_confi": "+3.5/-3.4",
36
+ "multi-document_synthesis_confi": "+4.6/-4.4",
37
+ "fact_checking_or_attributed_qa_confi": "+4.3/-4.3",
38
+ "average_confi": "+1.4/-1.37"
39
  }
temperature=1.0/Mistral-7B-Instruct-v0.3.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "mistralai/Mistral-7B-Instruct-v0.3",
3
- "brainstorm": 0.3,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.2,
6
- "extract": 0.1,
7
- "generation": 0.2,
8
- "rewrite": 0.2,
9
- "summarize": 0.2,
10
- "classify": 0.2,
11
- "reasoning_over_numerical_data": 0.2,
12
- "multi-document_synthesis": 0.3,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.24,
15
  "brainstorm_rank": 9,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 9,
25
  "fact_checking_or_attributed_qa_rank": 7,
26
  "average_rank": 12,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "mistralai/Mistral-7B-Instruct-v0.3",
3
+ "brainstorm": 0.298,
4
+ "open_qa": 0.618,
5
+ "closed_qa": 0.196,
6
+ "extract": 0.134,
7
+ "generation": 0.194,
8
+ "rewrite": 0.167,
9
+ "summarize": 0.168,
10
+ "classify": 0.226,
11
+ "reasoning_over_numerical_data": 0.18,
12
+ "multi-document_synthesis": 0.285,
13
+ "fact_checking_or_attributed_qa": 0.403,
14
+ "average": 0.2425,
15
  "brainstorm_rank": 9,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 9,
25
  "fact_checking_or_attributed_qa_rank": 7,
26
  "average_rank": 12,
27
+ "brainstorm_confi": "+3.6/-3.6",
28
+ "open_qa_confi": "+8.8/-8.8",
29
+ "closed_qa_confi": "+5.2/-5.0",
30
+ "extract_confi": "+4.7/-4.2",
31
+ "generation_confi": "+3.2/-3.1",
32
+ "rewrite_confi": "+2.7/-2.8",
33
+ "summarize_confi": "+5.4/-4.7",
34
+ "classify_confi": "+6.0/-5.5",
35
+ "reasoning_over_numerical_data_confi": "+3.0/-3.0",
36
+ "multi-document_synthesis_confi": "+4.2/-4.2",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.5",
38
+ "average_confi": "+1.3/-1.26"
39
  }
temperature=1.0/Mistral-Large-Instruct-2407.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "mistralai/Mistral-Large-Instruct-2407",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.3,
6
- "extract": 0.3,
7
- "generation": 0.5,
8
- "rewrite": 0.5,
9
- "summarize": 0.4,
10
- "classify": 0.5,
11
- "reasoning_over_numerical_data": 0.5,
12
- "multi-document_synthesis": 0.6,
13
- "fact_checking_or_attributed_qa": 0.2,
14
- "average": 0.47,
15
  "brainstorm_rank": 1,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 1,
25
  "fact_checking_or_attributed_qa_rank": 19,
26
  "average_rank": 1,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "mistralai/Mistral-Large-Instruct-2407",
3
+ "brainstorm": 0.548,
4
+ "open_qa": 0.618,
5
+ "closed_qa": 0.287,
6
+ "extract": 0.347,
7
+ "generation": 0.522,
8
+ "rewrite": 0.499,
9
+ "summarize": 0.411,
10
+ "classify": 0.458,
11
+ "reasoning_over_numerical_data": 0.461,
12
+ "multi-document_synthesis": 0.592,
13
+ "fact_checking_or_attributed_qa": 0.24,
14
+ "average": 0.4656,
15
  "brainstorm_rank": 1,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 1,
25
  "fact_checking_or_attributed_qa_rank": 19,
26
  "average_rank": 1,
27
+ "brainstorm_confi": "+4.0/-4.0",
28
+ "open_qa_confi": "+8.8/-9.8",
29
+ "closed_qa_confi": "+5.7/-5.7",
30
+ "extract_confi": "+5.9/-5.7",
31
+ "generation_confi": "+4.1/-3.8",
32
+ "rewrite_confi": "+3.6/-3.7",
33
+ "summarize_confi": "+6.4/-6.7",
34
+ "classify_confi": "+6.0/-6.2",
35
+ "reasoning_over_numerical_data_confi": "+3.9/-3.7",
36
+ "multi-document_synthesis_confi": "+4.6/-4.4",
37
+ "fact_checking_or_attributed_qa_confi": "+3.9/-3.9",
38
+ "average_confi": "+1.51/-1.4"
39
  }
temperature=1.0/Mistral-Small-Instruct-2409.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "mistralai/Mistral-Small-Instruct-2409",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.3,
6
- "extract": 0.2,
7
- "generation": 0.4,
8
- "rewrite": 0.4,
9
- "summarize": 0.4,
10
- "classify": 0.4,
11
- "reasoning_over_numerical_data": 0.4,
12
- "multi-document_synthesis": 0.6,
13
- "fact_checking_or_attributed_qa": 0.2,
14
- "average": 0.41,
15
  "brainstorm_rank": 4,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 3,
25
  "fact_checking_or_attributed_qa_rank": 19,
26
  "average_rank": 4,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "mistralai/Mistral-Small-Instruct-2409",
3
+ "brainstorm": 0.492,
4
+ "open_qa": 0.603,
5
+ "closed_qa": 0.302,
6
+ "extract": 0.235,
7
+ "generation": 0.435,
8
+ "rewrite": 0.408,
9
+ "summarize": 0.366,
10
+ "classify": 0.371,
11
+ "reasoning_over_numerical_data": 0.407,
12
+ "multi-document_synthesis": 0.564,
13
+ "fact_checking_or_attributed_qa": 0.225,
14
+ "average": 0.4079,
15
  "brainstorm_rank": 4,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 3,
25
  "fact_checking_or_attributed_qa_rank": 19,
26
  "average_rank": 4,
27
+ "brainstorm_confi": "+4.0/-4.0",
28
+ "open_qa_confi": "+9.3/-9.8",
29
+ "closed_qa_confi": "+5.9/-5.4",
30
+ "extract_confi": "+5.4/-5.2",
31
+ "generation_confi": "+4.0/-4.0",
32
+ "rewrite_confi": "+3.6/-3.7",
33
+ "summarize_confi": "+6.4/-6.2",
34
+ "classify_confi": "+6.0/-6.2",
35
+ "reasoning_over_numerical_data_confi": "+3.8/-3.7",
36
+ "multi-document_synthesis_confi": "+4.5/-4.6",
37
+ "fact_checking_or_attributed_qa_confi": "+3.9/-3.7",
38
+ "average_confi": "+1.41/-1.42"
39
  }
temperature=1.0/OLMo-7B-0724-Instruct-hf.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "allenai/OLMo-7B-SFT",
3
- "brainstorm": 0.1,
4
- "open_qa": 0.2,
5
- "closed_qa": 0.0,
6
- "extract": 0.0,
7
- "generation": 0.0,
8
- "rewrite": 0.0,
9
- "summarize": 0.0,
10
- "classify": 0.1,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.1,
13
- "fact_checking_or_attributed_qa": 0.2,
14
- "average": 0.07,
15
  "brainstorm_rank": 20,
16
  "open_qa_rank": 26,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 18,
25
  "fact_checking_or_attributed_qa_rank": 24,
26
  "average_rank": 24,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "allenai/OLMo-7B-SFT",
3
+ "brainstorm": 0.065,
4
+ "open_qa": 0.176,
5
+ "closed_qa": 0.04,
6
+ "extract": 0.037,
7
+ "generation": 0.04,
8
+ "rewrite": 0.033,
9
+ "summarize": 0.05,
10
+ "classify": 0.08,
11
+ "reasoning_over_numerical_data": 0.05,
12
+ "multi-document_synthesis": 0.073,
13
+ "fact_checking_or_attributed_qa": 0.173,
14
+ "average": 0.0669,
15
  "brainstorm_rank": 20,
16
  "open_qa_rank": 26,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 18,
25
  "fact_checking_or_attributed_qa_rank": 24,
26
  "average_rank": 24,
27
+ "brainstorm_confi": "+2.0/-2.0",
28
+ "open_qa_confi": "+7.8/-6.9",
29
+ "closed_qa_confi": "+2.7/-2.2",
30
+ "extract_confi": "+2.7/-2.2",
31
+ "generation_confi": "+1.5/-1.5",
32
+ "rewrite_confi": "+1.4/-1.3",
33
+ "summarize_confi": "+3.5/-3.0",
34
+ "classify_confi": "+3.7/-3.5",
35
+ "reasoning_over_numerical_data_confi": "+1.9/-1.7",
36
+ "multi-document_synthesis_confi": "+2.6/-2.2",
37
+ "fact_checking_or_attributed_qa_confi": "+3.5/-3.5",
38
+ "average_confi": "+0.75/-0.74"
39
  }
temperature=1.0/OLMo-7B-SFT-hf.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "allenai/OLMo-7B-SFT",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.1,
6
- "extract": 0.0,
7
- "generation": 0.0,
8
- "rewrite": 0.0,
9
- "summarize": 0.0,
10
- "classify": 0.2,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.1,
15
  "brainstorm_rank": 22,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 22,
25
  "fact_checking_or_attributed_qa_rank": 7,
26
  "average_rank": 21,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "allenai/OLMo-7B-SFT",
3
+ "brainstorm": 0.012,
4
+ "open_qa": 0.721,
5
+ "closed_qa": 0.139,
6
+ "extract": 0.045,
7
+ "generation": 0.048,
8
+ "rewrite": 0.024,
9
+ "summarize": 0.012,
10
+ "classify": 0.177,
11
+ "reasoning_over_numerical_data": 0.038,
12
+ "multi-document_synthesis": 0.029,
13
+ "fact_checking_or_attributed_qa": 0.439,
14
+ "average": 0.1029,
15
  "brainstorm_rank": 22,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 22,
25
  "fact_checking_or_attributed_qa_rank": 7,
26
  "average_rank": 21,
27
+ "brainstorm_confi": "+1.0/-0.8",
28
+ "open_qa_confi": "+8.8/-8.8",
29
+ "closed_qa_confi": "+4.7/-4.5",
30
+ "extract_confi": "+2.7/-2.2",
31
+ "generation_confi": "+1.8/-1.7",
32
+ "rewrite_confi": "+1.1/-1.0",
33
+ "summarize_confi": "+1.7/-1.2",
34
+ "classify_confi": "+5.2/-5.0",
35
+ "reasoning_over_numerical_data_confi": "+1.7/-1.6",
36
+ "multi-document_synthesis_confi": "+1.8/-1.3",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.5",
38
+ "average_confi": "+0.92/-0.92"
39
  }
temperature=1.0/Phi-3-medium-4k-instruct.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "microsoft/Phi-3-medium-4k-instruct",
3
- "brainstorm": 0.3,
4
- "open_qa": 0.8,
5
- "closed_qa": 0.3,
6
- "extract": 0.2,
7
- "generation": 0.2,
8
- "rewrite": 0.2,
9
- "summarize": 0.1,
10
- "classify": 0.4,
11
- "reasoning_over_numerical_data": 0.3,
12
- "multi-document_synthesis": 0.2,
13
- "fact_checking_or_attributed_qa": 0.5,
14
- "average": 0.3,
15
  "brainstorm_rank": 9,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 11,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 9,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "microsoft/Phi-3-medium-4k-instruct",
3
+ "brainstorm": 0.32,
4
+ "open_qa": 0.75,
5
+ "closed_qa": 0.275,
6
+ "extract": 0.163,
7
+ "generation": 0.25,
8
+ "rewrite": 0.208,
9
+ "summarize": 0.114,
10
+ "classify": 0.386,
11
+ "reasoning_over_numerical_data": 0.34,
12
+ "multi-document_synthesis": 0.212,
13
+ "fact_checking_or_attributed_qa": 0.489,
14
+ "average": 0.2951,
15
  "brainstorm_rank": 9,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 11,
25
  "fact_checking_or_attributed_qa_rank": 1,
26
  "average_rank": 9,
27
+ "brainstorm_confi": "+3.8/-3.7",
28
+ "open_qa_confi": "+7.8/-8.8",
29
+ "closed_qa_confi": "+5.9/-5.4",
30
+ "extract_confi": "+4.7/-4.5",
31
+ "generation_confi": "+3.5/-3.3",
32
+ "rewrite_confi": "+3.1/-2.9",
33
+ "summarize_confi": "+4.2/-4.2",
34
+ "classify_confi": "+6.7/-6.5",
35
+ "reasoning_over_numerical_data_confi": "+3.7/-3.6",
36
+ "multi-document_synthesis_confi": "+3.8/-3.8",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.3",
38
+ "average_confi": "+1.32/-1.32"
39
  }
temperature=1.0/Qwen1.5-110B-Chat.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "Qwen/Qwen1.5-110B-Chat",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.8,
5
- "closed_qa": 0.3,
6
- "extract": 0.3,
7
- "generation": 0.4,
8
- "rewrite": 0.3,
9
  "summarize": 0.3,
10
- "classify": 0.4,
11
- "reasoning_over_numerical_data": 0.4,
12
- "multi-document_synthesis": 0.5,
13
- "fact_checking_or_attributed_qa": 0.3,
14
- "average": 0.38,
15
  "brainstorm_rank": 5,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 4,
25
  "fact_checking_or_attributed_qa_rank": 12,
26
  "average_rank": 4,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "Qwen/Qwen1.5-110B-Chat",
3
+ "brainstorm": 0.452,
4
+ "open_qa": 0.775,
5
+ "closed_qa": 0.28,
6
+ "extract": 0.255,
7
+ "generation": 0.367,
8
+ "rewrite": 0.321,
9
  "summarize": 0.3,
10
+ "classify": 0.413,
11
+ "reasoning_over_numerical_data": 0.375,
12
+ "multi-document_synthesis": 0.49,
13
+ "fact_checking_or_attributed_qa": 0.346,
14
+ "average": 0.3827,
15
  "brainstorm_rank": 5,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 4,
25
  "fact_checking_or_attributed_qa_rank": 12,
26
  "average_rank": 4,
27
+ "brainstorm_confi": "+4.0/-4.0",
28
+ "open_qa_confi": "+7.8/-8.8",
29
+ "closed_qa_confi": "+5.9/-5.7",
30
+ "extract_confi": "+5.7/-5.4",
31
+ "generation_confi": "+3.9/-3.9",
32
+ "rewrite_confi": "+3.5/-3.4",
33
+ "summarize_confi": "+6.4/-6.2",
34
+ "classify_confi": "+6.7/-6.2",
35
+ "reasoning_over_numerical_data_confi": "+3.8/-3.8",
36
+ "multi-document_synthesis_confi": "+4.4/-4.7",
37
+ "fact_checking_or_attributed_qa_confi": "+4.3/-4.3",
38
+ "average_confi": "+1.41/-1.42"
39
  }
temperature=1.0/Qwen2-72B-Instruct.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "Qwen/Qwen2-72B-Instruct",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.7,
5
- "closed_qa": 0.3,
6
- "extract": 0.3,
7
- "generation": 0.3,
8
- "rewrite": 0.3,
9
- "summarize": 0.2,
10
- "classify": 0.4,
11
- "reasoning_over_numerical_data": 0.4,
12
- "multi-document_synthesis": 0.3,
13
- "fact_checking_or_attributed_qa": 0.4,
14
- "average": 0.37,
15
  "brainstorm_rank": 5,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 8,
25
  "fact_checking_or_attributed_qa_rank": 7,
26
  "average_rank": 6,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "Qwen/Qwen2-72B-Instruct",
3
+ "brainstorm": 0.454,
4
+ "open_qa": 0.74,
5
+ "closed_qa": 0.257,
6
+ "extract": 0.275,
7
+ "generation": 0.34,
8
+ "rewrite": 0.326,
9
+ "summarize": 0.22,
10
+ "classify": 0.393,
11
+ "reasoning_over_numerical_data": 0.398,
12
+ "multi-document_synthesis": 0.342,
13
+ "fact_checking_or_attributed_qa": 0.387,
14
+ "average": 0.3659,
15
  "brainstorm_rank": 5,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 8,
25
  "fact_checking_or_attributed_qa_rank": 7,
26
  "average_rank": 6,
27
+ "brainstorm_confi": "+3.9/-4.1",
28
+ "open_qa_confi": "+8.3/-8.3",
29
+ "closed_qa_confi": "+5.7/-5.7",
30
+ "extract_confi": "+5.9/-5.4",
31
+ "generation_confi": "+3.7/-3.8",
32
+ "rewrite_confi": "+3.4/-3.4",
33
+ "summarize_confi": "+5.7/-5.7",
34
+ "classify_confi": "+6.2/-6.2",
35
+ "reasoning_over_numerical_data_confi": "+3.8/-3.8",
36
+ "multi-document_synthesis_confi": "+4.2/-4.2",
37
+ "fact_checking_or_attributed_qa_confi": "+4.5/-4.3",
38
+ "average_confi": "+1.41/-1.41"
39
  }
temperature=1.0/Qwen2.5-72B-Instruct.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "Qwen/Qwen2.5-72B-Instruct",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.3,
6
- "extract": 0.3,
7
- "generation": 0.5,
8
- "rewrite": 0.5,
9
- "summarize": 0.4,
10
- "classify": 0.4,
11
- "reasoning_over_numerical_data": 0.5,
12
- "multi-document_synthesis": 0.6,
13
- "fact_checking_or_attributed_qa": 0.2,
14
- "average": 0.44,
15
  "brainstorm_rank": 1,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 1,
25
  "fact_checking_or_attributed_qa_rank": 19,
26
  "average_rank": 3,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.1",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "Qwen/Qwen2.5-72B-Instruct",
3
+ "brainstorm": 0.523,
4
+ "open_qa": 0.642,
5
+ "closed_qa": 0.292,
6
+ "extract": 0.334,
7
+ "generation": 0.47,
8
+ "rewrite": 0.453,
9
+ "summarize": 0.364,
10
+ "classify": 0.42,
11
+ "reasoning_over_numerical_data": 0.454,
12
+ "multi-document_synthesis": 0.574,
13
+ "fact_checking_or_attributed_qa": 0.225,
14
+ "average": 0.4391,
15
  "brainstorm_rank": 1,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 1,
25
  "fact_checking_or_attributed_qa_rank": 19,
26
  "average_rank": 3,
27
+ "brainstorm_confi": "+3.9/-3.9",
28
+ "open_qa_confi": "+9.3/-9.3",
29
+ "closed_qa_confi": "+5.9/-5.4",
30
+ "extract_confi": "+6.2/-5.9",
31
+ "generation_confi": "+4.0/-4.2",
32
+ "rewrite_confi": "+3.6/-3.7",
33
+ "summarize_confi": "+6.4/-6.4",
34
+ "classify_confi": "+6.2/-6.2",
35
+ "reasoning_over_numerical_data_confi": "+3.8/-3.9",
36
+ "multi-document_synthesis_confi": "+4.4/-4.6",
37
+ "fact_checking_or_attributed_qa_confi": "+3.9/-3.7",
38
+ "average_confi": "+1.43/-1.51"
39
  }
temperature=1.0/WizardLM-13B-V1.2.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "WizardLMTeam/WizardLM-13B-V1.2",
3
- "brainstorm": 0.2,
4
  "open_qa": 0.5,
5
- "closed_qa": 0.1,
6
- "extract": 0.1,
7
- "generation": 0.2,
8
- "rewrite": 0.1,
9
- "summarize": 0.1,
10
- "classify": 0.2,
11
- "reasoning_over_numerical_data": 0.1,
12
- "multi-document_synthesis": 0.1,
13
- "fact_checking_or_attributed_qa": 0.3,
14
- "average": 0.16,
15
  "brainstorm_rank": 13,
16
  "open_qa_rank": 21,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 14,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 15,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "WizardLMTeam/WizardLM-13B-V1.2",
3
+ "brainstorm": 0.189,
4
  "open_qa": 0.5,
5
+ "closed_qa": 0.119,
6
+ "extract": 0.084,
7
+ "generation": 0.165,
8
+ "rewrite": 0.137,
9
+ "summarize": 0.072,
10
+ "classify": 0.216,
11
+ "reasoning_over_numerical_data": 0.078,
12
+ "multi-document_synthesis": 0.104,
13
+ "fact_checking_or_attributed_qa": 0.307,
14
+ "average": 0.1618,
15
  "brainstorm_rank": 13,
16
  "open_qa_rank": 21,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 14,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 15,
27
+ "brainstorm_confi": "+3.3/-3.1",
28
+ "open_qa_confi": "+9.8/-9.8",
29
+ "closed_qa_confi": "+4.5/-4.0",
30
+ "extract_confi": "+3.7/-3.2",
31
+ "generation_confi": "+3.0/-3.0",
32
+ "rewrite_confi": "+2.6/-2.4",
33
+ "summarize_confi": "+3.7/-3.2",
34
+ "classify_confi": "+5.5/-5.2",
35
+ "reasoning_over_numerical_data_confi": "+2.2/-2.1",
36
+ "multi-document_synthesis_confi": "+2.9/-2.9",
37
+ "fact_checking_or_attributed_qa_confi": "+4.1/-3.9",
38
+ "average_confi": "+1.1/-1.08"
39
  }
temperature=1.0/Yi-1.5-34B-Chat.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "01-ai/Yi-1.5-34B-Chat",
3
- "brainstorm": 0.5,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.2,
6
- "extract": 0.2,
7
- "generation": 0.4,
8
- "rewrite": 0.3,
9
- "summarize": 0.2,
10
- "classify": 0.3,
11
- "reasoning_over_numerical_data": 0.3,
12
- "multi-document_synthesis": 0.4,
13
- "fact_checking_or_attributed_qa": 0.3,
14
- "average": 0.34,
15
  "brainstorm_rank": 1,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 6,
25
  "fact_checking_or_attributed_qa_rank": 19,
26
  "average_rank": 7,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.1/-0.1",
30
- "extract_confi": "+0.1/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.1/-0.1",
34
- "classify_confi": "+0.1/-0.1",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "01-ai/Yi-1.5-34B-Chat",
3
+ "brainstorm": 0.509,
4
+ "open_qa": 0.593,
5
+ "closed_qa": 0.243,
6
+ "extract": 0.193,
7
+ "generation": 0.361,
8
+ "rewrite": 0.304,
9
+ "summarize": 0.208,
10
+ "classify": 0.291,
11
+ "reasoning_over_numerical_data": 0.291,
12
+ "multi-document_synthesis": 0.386,
13
+ "fact_checking_or_attributed_qa": 0.268,
14
+ "average": 0.3377,
15
  "brainstorm_rank": 1,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 6,
25
  "fact_checking_or_attributed_qa_rank": 19,
26
  "average_rank": 7,
27
+ "brainstorm_confi": "+3.9/-3.9",
28
+ "open_qa_confi": "+9.3/-9.3",
29
+ "closed_qa_confi": "+5.4/-5.4",
30
+ "extract_confi": "+5.2/-5.0",
31
+ "generation_confi": "+3.9/-3.8",
32
+ "rewrite_confi": "+3.4/-3.4",
33
+ "summarize_confi": "+5.5/-5.4",
34
+ "classify_confi": "+5.7/-6.0",
35
+ "reasoning_over_numerical_data_confi": "+3.6/-3.6",
36
+ "multi-document_synthesis_confi": "+4.4/-4.4",
37
+ "fact_checking_or_attributed_qa_confi": "+4.1/-4.1",
38
+ "average_confi": "+1.37/-1.39"
39
  }
temperature=1.0/dolly-v2-12b.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "databricks/dolly-v2-7b",
3
  "brainstorm": 0.0,
4
- "open_qa": 0.0,
5
  "closed_qa": 0.0,
6
  "extract": 0.0,
7
- "generation": 0.0,
8
- "rewrite": 0.0,
9
  "summarize": 0.0,
10
- "classify": 0.0,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.0,
14
- "average": 0.01,
15
  "brainstorm_rank": 28,
16
  "open_qa_rank": 28,
17
  "closed_qa_rank": 27,
@@ -25,15 +25,15 @@
25
  "fact_checking_or_attributed_qa_rank": 28,
26
  "average_rank": 29,
27
  "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.0/-0.0",
29
  "closed_qa_confi": "+0.0/-0.0",
30
  "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
  "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.0/-0.0"
39
  }
 
1
  {
2
  "path": "databricks/dolly-v2-7b",
3
  "brainstorm": 0.0,
4
+ "open_qa": 0.029,
5
  "closed_qa": 0.0,
6
  "extract": 0.0,
7
+ "generation": 0.007,
8
+ "rewrite": 0.001,
9
  "summarize": 0.0,
10
+ "classify": 0.005,
11
+ "reasoning_over_numerical_data": 0.007,
12
+ "multi-document_synthesis": 0.009,
13
+ "fact_checking_or_attributed_qa": 0.017,
14
+ "average": 0.0059,
15
  "brainstorm_rank": 28,
16
  "open_qa_rank": 28,
17
  "closed_qa_rank": 27,
 
25
  "fact_checking_or_attributed_qa_rank": 28,
26
  "average_rank": 29,
27
  "brainstorm_confi": "+0.0/-0.0",
28
+ "open_qa_confi": "+3.9/-2.9",
29
  "closed_qa_confi": "+0.0/-0.0",
30
  "extract_confi": "+0.0/-0.0",
31
+ "generation_confi": "+0.7/-0.5",
32
+ "rewrite_confi": "+0.3/-0.1",
33
  "summarize_confi": "+0.0/-0.0",
34
+ "classify_confi": "+1.0/-0.5",
35
+ "reasoning_over_numerical_data_confi": "+0.7/-0.6",
36
+ "multi-document_synthesis_confi": "+0.9/-0.7",
37
+ "fact_checking_or_attributed_qa_confi": "+1.3/-1.1",
38
+ "average_confi": "+0.23/-0.21"
39
  }
temperature=1.0/dolly-v2-7b.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "databricks/dolly-v2-12b",
3
  "brainstorm": 0.0,
4
- "open_qa": 0.1,
5
  "closed_qa": 0.0,
6
  "extract": 0.0,
7
- "generation": 0.0,
8
- "rewrite": 0.0,
9
  "summarize": 0.0,
10
- "classify": 0.0,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.0,
14
- "average": 0.01,
15
  "brainstorm_rank": 28,
16
  "open_qa_rank": 28,
17
  "closed_qa_rank": 27,
@@ -25,15 +25,15 @@
25
  "fact_checking_or_attributed_qa_rank": 28,
26
  "average_rank": 28,
27
  "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.0/-0.0",
29
  "closed_qa_confi": "+0.0/-0.0",
30
  "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
  "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.0/-0.0"
39
  }
 
1
  {
2
  "path": "databricks/dolly-v2-12b",
3
  "brainstorm": 0.0,
4
+ "open_qa": 0.059,
5
  "closed_qa": 0.0,
6
  "extract": 0.0,
7
+ "generation": 0.008,
8
+ "rewrite": 0.001,
9
  "summarize": 0.0,
10
+ "classify": 0.015,
11
+ "reasoning_over_numerical_data": 0.02,
12
+ "multi-document_synthesis": 0.009,
13
+ "fact_checking_or_attributed_qa": 0.022,
14
+ "average": 0.0093,
15
  "brainstorm_rank": 28,
16
  "open_qa_rank": 28,
17
  "closed_qa_rank": 27,
 
25
  "fact_checking_or_attributed_qa_rank": 28,
26
  "average_rank": 28,
27
  "brainstorm_confi": "+0.0/-0.0",
28
+ "open_qa_confi": "+4.9/-3.9",
29
  "closed_qa_confi": "+0.0/-0.0",
30
  "extract_confi": "+0.0/-0.0",
31
+ "generation_confi": "+0.8/-0.7",
32
+ "rewrite_confi": "+0.3/-0.1",
33
  "summarize_confi": "+0.0/-0.0",
34
+ "classify_confi": "+2.0/-1.5",
35
+ "reasoning_over_numerical_data_confi": "+1.2/-1.0",
36
+ "multi-document_synthesis_confi": "+0.9/-0.7",
37
+ "fact_checking_or_attributed_qa_confi": "+1.5/-1.3",
38
+ "average_confi": "+0.31/-0.28"
39
  }
temperature=1.0/gpt4all-13b-snoozy.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "nomic-ai/gpt4all-13b-snoozy",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.8,
5
- "closed_qa": 0.0,
6
- "extract": 0.0,
7
- "generation": 0.0,
8
- "rewrite": 0.0,
9
- "summarize": 0.0,
10
- "classify": 0.0,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.2,
14
- "average": 0.06,
15
  "brainstorm_rank": 22,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 27,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 25,
25
  "fact_checking_or_attributed_qa_rank": 24,
26
  "average_rank": 24,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "nomic-ai/gpt4all-13b-snoozy",
3
+ "brainstorm": 0.008,
4
+ "open_qa": 0.77,
5
+ "closed_qa": 0.015,
6
+ "extract": 0.022,
7
+ "generation": 0.043,
8
+ "rewrite": 0.029,
9
+ "summarize": 0.002,
10
+ "classify": 0.045,
11
+ "reasoning_over_numerical_data": 0.043,
12
+ "multi-document_synthesis": 0.015,
13
+ "fact_checking_or_attributed_qa": 0.21,
14
+ "average": 0.0641,
15
  "brainstorm_rank": 22,
16
  "open_qa_rank": 1,
17
  "closed_qa_rank": 27,
 
24
  "multi-document_synthesis_rank": 25,
25
  "fact_checking_or_attributed_qa_rank": 24,
26
  "average_rank": 24,
27
+ "brainstorm_confi": "+0.8/-0.6",
28
+ "open_qa_confi": "+7.8/-8.3",
29
+ "closed_qa_confi": "+1.7/-1.2",
30
+ "extract_confi": "+2.0/-1.7",
31
+ "generation_confi": "+1.7/-1.5",
32
+ "rewrite_confi": "+1.3/-1.1",
33
+ "summarize_confi": "+0.5/-0.2",
34
+ "classify_confi": "+3.0/-2.5",
35
+ "reasoning_over_numerical_data_confi": "+1.7/-1.7",
36
+ "multi-document_synthesis_confi": "+1.3/-1.1",
37
+ "fact_checking_or_attributed_qa_confi": "+3.7/-3.7",
38
+ "average_confi": "+0.73/-0.69"
39
  }
temperature=1.0/koala-13B-HF.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "TheBloke/koala-13B-HF",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.5,
5
- "closed_qa": 0.1,
6
- "extract": 0.0,
7
- "generation": 0.0,
8
- "rewrite": 0.0,
9
- "summarize": 0.0,
10
  "classify": 0.1,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.3,
14
- "average": 0.07,
15
  "brainstorm_rank": 22,
16
  "open_qa_rank": 21,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 22,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 22,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
- "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "TheBloke/koala-13B-HF",
3
+ "brainstorm": 0.009,
4
+ "open_qa": 0.49,
5
+ "closed_qa": 0.064,
6
+ "extract": 0.027,
7
+ "generation": 0.045,
8
+ "rewrite": 0.031,
9
+ "summarize": 0.012,
10
  "classify": 0.1,
11
+ "reasoning_over_numerical_data": 0.032,
12
+ "multi-document_synthesis": 0.024,
13
+ "fact_checking_or_attributed_qa": 0.312,
14
+ "average": 0.0746,
15
  "brainstorm_rank": 22,
16
  "open_qa_rank": 21,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 22,
25
  "fact_checking_or_attributed_qa_rank": 15,
26
  "average_rank": 22,
27
+ "brainstorm_confi": "+0.8/-0.7",
28
+ "open_qa_confi": "+9.8/-9.8",
29
+ "closed_qa_confi": "+3.5/-3.0",
30
+ "extract_confi": "+2.2/-1.7",
31
+ "generation_confi": "+1.7/-1.7",
32
+ "rewrite_confi": "+1.3/-1.3",
33
+ "summarize_confi": "+1.5/-1.0",
34
+ "classify_confi": "+4.2/-4.0",
35
+ "reasoning_over_numerical_data_confi": "+1.5/-1.5",
36
+ "multi-document_synthesis_confi": "+1.5/-1.3",
37
+ "fact_checking_or_attributed_qa_confi": "+4.3/-4.1",
38
+ "average_confi": "+0.82/-0.76"
39
  }
temperature=1.0/koala-7B-HF.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "path": "TheBloke/koala-7B-HF",
3
- "brainstorm": 0.0,
4
- "open_qa": 0.6,
5
- "closed_qa": 0.0,
6
- "extract": 0.0,
7
- "generation": 0.0,
8
- "rewrite": 0.0,
9
  "summarize": 0.0,
10
- "classify": 0.0,
11
- "reasoning_over_numerical_data": 0.0,
12
- "multi-document_synthesis": 0.0,
13
- "fact_checking_or_attributed_qa": 0.3,
14
- "average": 0.06,
15
  "brainstorm_rank": 22,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
@@ -24,16 +24,16 @@
24
  "multi-document_synthesis_rank": 22,
25
  "fact_checking_or_attributed_qa_rank": 19,
26
  "average_rank": 24,
27
- "brainstorm_confi": "+0.0/-0.0",
28
- "open_qa_confi": "+0.1/-0.1",
29
- "closed_qa_confi": "+0.0/-0.0",
30
- "extract_confi": "+0.0/-0.0",
31
- "generation_confi": "+0.0/-0.0",
32
- "rewrite_confi": "+0.0/-0.0",
33
  "summarize_confi": "+0.0/-0.0",
34
- "classify_confi": "+0.0/-0.0",
35
- "reasoning_over_numerical_data_confi": "+0.0/-0.0",
36
- "multi-document_synthesis_confi": "+0.0/-0.0",
37
- "fact_checking_or_attributed_qa_confi": "+0.0/-0.0",
38
- "average_confi": "+0.01/-0.01"
39
  }
 
1
  {
2
  "path": "TheBloke/koala-7B-HF",
3
+ "brainstorm": 0.007,
4
+ "open_qa": 0.598,
5
+ "closed_qa": 0.015,
6
+ "extract": 0.022,
7
+ "generation": 0.027,
8
+ "rewrite": 0.021,
9
  "summarize": 0.0,
10
+ "classify": 0.04,
11
+ "reasoning_over_numerical_data": 0.028,
12
+ "multi-document_synthesis": 0.018,
13
+ "fact_checking_or_attributed_qa": 0.268,
14
+ "average": 0.0607,
15
  "brainstorm_rank": 22,
16
  "open_qa_rank": 10,
17
  "closed_qa_rank": 1,
 
24
  "multi-document_synthesis_rank": 22,
25
  "fact_checking_or_attributed_qa_rank": 19,
26
  "average_rank": 24,
27
+ "brainstorm_confi": "+0.7/-0.5",
28
+ "open_qa_confi": "+8.8/-9.8",
29
+ "closed_qa_confi": "+1.7/-1.2",
30
+ "extract_confi": "+2.0/-1.7",
31
+ "generation_confi": "+1.3/-1.2",
32
+ "rewrite_confi": "+1.1/-1.0",
33
  "summarize_confi": "+0.0/-0.0",
34
+ "classify_confi": "+2.7/-2.5",
35
+ "reasoning_over_numerical_data_confi": "+1.5/-1.3",
36
+ "multi-document_synthesis_confi": "+1.3/-1.1",
37
+ "fact_checking_or_attributed_qa_confi": "+3.9/-3.9",
38
+ "average_confi": "+0.73/-0.72"
39
  }