File size: 6,167 Bytes
de821b5
 
 
 
 
 
 
 
 
 
5ab7214
4c71592
de821b5
 
77974b7
 
b45bad5
 
77974b7
 
 
 
 
9253029
77974b7
 
 
de821b5
46e8d4d
 
 
 
 
 
 
 
 
 
0e496e2
46e8d4d
0b4902f
 
8572d46
 
0e496e2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
de821b5
 
4c71592
de821b5
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
---
language:
- en
license: apache-2.0
tags:
- text-generation-inference
- transformers
- unsloth
- mistral
- trl
- sft
base_model: alpindale/Mistral-7B-v0.2
---

# Mistral-7B-v0.2-OpenHermes

![image/webp](https://cdn-uploads.huggingface.co/production/uploads/6455cc8d679315e4ef16fbec/AbagOgU056oIB7S31XESC.webp)

SFT Training Params:
+ Learning Rate: 2e-4
+ Batch Size: 8
+ Gradient Accumulation steps: 4
+ Dataset: teknium/OpenHermes-2.5 (200k split contains a slight bias towards rp and theory of life)
+ r: 16
+ Lora Alpha: 16

Training Time: 13 hours on A100

Prompt Template: ChatML

```
<|im_start|>system
You are a helpful assistant.<|im_end|>
<|im_start|>user
What's the capital of France?<|im_end|>
<|im_start|>assistant
Paris.
```
## Quantizations

[GGUF](https://huggingface.co/macadeliccc/Mistral-7B-v0.2-OpenHermes-GGUF)

[AWQ](https://huggingface.co/macadeliccc/Mistral-7B-v0.2-OpenHermes-AWQ/)

### Evaluations

Thanks to Maxime Labonne for the evalution:

|                                           Model                                           |AGIEval|GPT4All|TruthfulQA|Bigbench|Average|
|-------------------------------------------------------------------------------------------|------:|------:|---------:|-------:|------:|
|[Mistral-7B-v0.2-OpenHermes](https://huggingface.co/macadeliccc/Mistral-7B-v0.2-OpenHermes)|  35.57|  67.15|     42.06|   36.27|  45.26|

### AGIEval
|             Task             |Version| Metric |Value|   |Stderr|
|------------------------------|------:|--------|----:|---|-----:|
|agieval_aqua_rat              |      0|acc     |24.02|±  |  2.69|
|                              |       |acc_norm|21.65|±  |  2.59|
|agieval_logiqa_en             |      0|acc     |28.11|±  |  1.76|
|                              |       |acc_norm|34.56|±  |  1.87|
|agieval_lsat_ar               |      0|acc     |27.83|±  |  2.96|
|                              |       |acc_norm|23.48|±  |  2.80|
|agieval_lsat_lr               |      0|acc     |33.73|±  |  2.10|
|                              |       |acc_norm|33.14|±  |  2.09|
|agieval_lsat_rc               |      0|acc     |48.70|±  |  3.05|
|                              |       |acc_norm|39.78|±  |  2.99|
|agieval_sat_en                |      0|acc     |67.48|±  |  3.27|
|                              |       |acc_norm|64.56|±  |  3.34|
|agieval_sat_en_without_passage|      0|acc     |38.83|±  |  3.40|
|                              |       |acc_norm|37.38|±  |  3.38|
|agieval_sat_math              |      0|acc     |32.27|±  |  3.16|
|                              |       |acc_norm|30.00|±  |  3.10|

Average: 35.57%

### GPT4All
|    Task     |Version| Metric |Value|   |Stderr|
|-------------|------:|--------|----:|---|-----:|
|arc_challenge|      0|acc     |45.05|±  |  1.45|
|             |       |acc_norm|48.46|±  |  1.46|
|arc_easy     |      0|acc     |77.27|±  |  0.86|
|             |       |acc_norm|73.78|±  |  0.90|
|boolq        |      1|acc     |68.62|±  |  0.81|
|hellaswag    |      0|acc     |59.63|±  |  0.49|
|             |       |acc_norm|79.66|±  |  0.40|
|openbookqa   |      0|acc     |31.40|±  |  2.08|
|             |       |acc_norm|43.40|±  |  2.22|
|piqa         |      0|acc     |80.25|±  |  0.93|
|             |       |acc_norm|82.05|±  |  0.90|
|winogrande   |      0|acc     |74.11|±  |  1.23|

Average: 67.15%

### TruthfulQA
|    Task     |Version|Metric|Value|   |Stderr|
|-------------|------:|------|----:|---|-----:|
|truthfulqa_mc|      1|mc1   |27.54|±  |  1.56|
|             |       |mc2   |42.06|±  |  1.44|

Average: 42.06%

### Bigbench
|                      Task                      |Version|       Metric        |Value|   |Stderr|
|------------------------------------------------|------:|---------------------|----:|---|-----:|
|bigbench_causal_judgement                       |      0|multiple_choice_grade|56.32|±  |  3.61|
|bigbench_date_understanding                     |      0|multiple_choice_grade|66.40|±  |  2.46|
|bigbench_disambiguation_qa                      |      0|multiple_choice_grade|45.74|±  |  3.11|
|bigbench_geometric_shapes                       |      0|multiple_choice_grade|10.58|±  |  1.63|
|                                                |       |exact_str_match      | 0.00|±  |  0.00|
|bigbench_logical_deduction_five_objects         |      0|multiple_choice_grade|25.00|±  |  1.94|
|bigbench_logical_deduction_seven_objects        |      0|multiple_choice_grade|17.71|±  |  1.44|
|bigbench_logical_deduction_three_objects        |      0|multiple_choice_grade|37.33|±  |  2.80|
|bigbench_movie_recommendation                   |      0|multiple_choice_grade|29.40|±  |  2.04|
|bigbench_navigate                               |      0|multiple_choice_grade|50.00|±  |  1.58|
|bigbench_reasoning_about_colored_objects        |      0|multiple_choice_grade|42.50|±  |  1.11|
|bigbench_ruin_names                             |      0|multiple_choice_grade|39.06|±  |  2.31|
|bigbench_salient_translation_error_detection    |      0|multiple_choice_grade|12.93|±  |  1.06|
|bigbench_snarks                                 |      0|multiple_choice_grade|69.06|±  |  3.45|
|bigbench_sports_understanding                   |      0|multiple_choice_grade|49.80|±  |  1.59|
|bigbench_temporal_sequences                     |      0|multiple_choice_grade|26.50|±  |  1.40|
|bigbench_tracking_shuffled_objects_five_objects |      0|multiple_choice_grade|21.20|±  |  1.16|
|bigbench_tracking_shuffled_objects_seven_objects|      0|multiple_choice_grade|16.06|±  |  0.88|
|bigbench_tracking_shuffled_objects_three_objects|      0|multiple_choice_grade|37.33|±  |  2.80|

Average: 36.27%

Average score: 45.26%

Elapsed time: 01:49:22

- **Developed by:** macadeliccc
- **License:** apache-2.0
- **Finetuned from model :** alpindale/Mistral-7B-v0.2

This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.

[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)