Files changed (1) hide show
  1. README.md +117 -1
README.md CHANGED
@@ -1,6 +1,109 @@
1
  ---
2
  license: apache-2.0
3
  inference: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  ---
5
 
6
  # MistralLite 11B Model
@@ -361,4 +464,17 @@ for output in outputs:
361
  ```
362
 
363
  ## Limitations ##
364
- Before using the MistralLite model, it is important to perform your own independent assessment, and take measures to ensure that your use would comply with your own specific quality control practices and standards, and that your use would comply with the local rules, laws, regulations, licenses and terms that apply to you, and your content.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  inference: false
4
+ model-index:
5
+ - name: MistralLite-11B
6
+ results:
7
+ - task:
8
+ type: text-generation
9
+ name: Text Generation
10
+ dataset:
11
+ name: AI2 Reasoning Challenge (25-Shot)
12
+ type: ai2_arc
13
+ config: ARC-Challenge
14
+ split: test
15
+ args:
16
+ num_few_shot: 25
17
+ metrics:
18
+ - type: acc_norm
19
+ value: 57.68
20
+ name: normalized accuracy
21
+ source:
22
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=NurtureAI/MistralLite-11B
23
+ name: Open LLM Leaderboard
24
+ - task:
25
+ type: text-generation
26
+ name: Text Generation
27
+ dataset:
28
+ name: HellaSwag (10-Shot)
29
+ type: hellaswag
30
+ split: validation
31
+ args:
32
+ num_few_shot: 10
33
+ metrics:
34
+ - type: acc_norm
35
+ value: 79.54
36
+ name: normalized accuracy
37
+ source:
38
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=NurtureAI/MistralLite-11B
39
+ name: Open LLM Leaderboard
40
+ - task:
41
+ type: text-generation
42
+ name: Text Generation
43
+ dataset:
44
+ name: MMLU (5-Shot)
45
+ type: cais/mmlu
46
+ config: all
47
+ split: test
48
+ args:
49
+ num_few_shot: 5
50
+ metrics:
51
+ - type: acc
52
+ value: 50.09
53
+ name: accuracy
54
+ source:
55
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=NurtureAI/MistralLite-11B
56
+ name: Open LLM Leaderboard
57
+ - task:
58
+ type: text-generation
59
+ name: Text Generation
60
+ dataset:
61
+ name: TruthfulQA (0-shot)
62
+ type: truthful_qa
63
+ config: multiple_choice
64
+ split: validation
65
+ args:
66
+ num_few_shot: 0
67
+ metrics:
68
+ - type: mc2
69
+ value: 38.27
70
+ source:
71
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=NurtureAI/MistralLite-11B
72
+ name: Open LLM Leaderboard
73
+ - task:
74
+ type: text-generation
75
+ name: Text Generation
76
+ dataset:
77
+ name: Winogrande (5-shot)
78
+ type: winogrande
79
+ config: winogrande_xl
80
+ split: validation
81
+ args:
82
+ num_few_shot: 5
83
+ metrics:
84
+ - type: acc
85
+ value: 76.64
86
+ name: accuracy
87
+ source:
88
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=NurtureAI/MistralLite-11B
89
+ name: Open LLM Leaderboard
90
+ - task:
91
+ type: text-generation
92
+ name: Text Generation
93
+ dataset:
94
+ name: GSM8k (5-shot)
95
+ type: gsm8k
96
+ config: main
97
+ split: test
98
+ args:
99
+ num_few_shot: 5
100
+ metrics:
101
+ - type: acc
102
+ value: 0.38
103
+ name: accuracy
104
+ source:
105
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=NurtureAI/MistralLite-11B
106
+ name: Open LLM Leaderboard
107
  ---
108
 
109
  # MistralLite 11B Model
 
464
  ```
465
 
466
  ## Limitations ##
467
+ Before using the MistralLite model, it is important to perform your own independent assessment, and take measures to ensure that your use would comply with your own specific quality control practices and standards, and that your use would comply with the local rules, laws, regulations, licenses and terms that apply to you, and your content.
468
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
469
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_NurtureAI__MistralLite-11B)
470
+
471
+ | Metric |Value|
472
+ |---------------------------------|----:|
473
+ |Avg. |50.43|
474
+ |AI2 Reasoning Challenge (25-Shot)|57.68|
475
+ |HellaSwag (10-Shot) |79.54|
476
+ |MMLU (5-Shot) |50.09|
477
+ |TruthfulQA (0-shot) |38.27|
478
+ |Winogrande (5-shot) |76.64|
479
+ |GSM8k (5-shot) | 0.38|
480
+