File size: 2,961 Bytes
c79f12b
 
 
 
 
4659e53
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
eef576b
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
---
license: apache-2.0
pipeline_tag: text-generation
---

![DALL·E 2023-10-11 16.30.48 - Photo of a dark forest with tall, ancient trees. Their branches and leaves form intricate patterns, revealing a portal of swirling magical energy in t.png](https://cdn-uploads.huggingface.co/production/uploads/643c81209f5d314db2de9743/FE8hhNKnRX8UOE5qXNkZD.png)

# Overview

This model is a finetune of Mistral7b on cleaned data from WizardLM Evol Instruct v2 196k. most instances of RLHF were removed from the dataset, so this should be treated as a unscensored model although it is not fully unscensored. 

# Benchmarks

Wizard Mistral was only finetuned on >200k rows of evol instruct multi turn data, however it achieves competetive results when evaluated. below is wizard mistrals benchmark scores compared to the most popular mistral7b finetunes. 

| Model                                            | Average | ARC   | HellaSwag | MMLU | TruthfulQA |
|--------------------------------------------------|---------|-------|-----------|------|------------|
| unaidedelf87777/wizard-mistral-v0.1              | 64.18   | 61.77 | 83.51     | 63.99| 47.46      |
| Undi95/Mistral-11B-TestBench11                   | **67.21**| **64.42**| 83.93     | 63.82| 56.68      |
| Undi95/Mistral-11B-TestBench9                    | 67.13   | 64.08 | 84.24     | **64** | 56.19      |
| ehartford/dolphin-2.1-mistral-7b                 | 67.06   | 64.42 | 84.92     | 63.32| 55.56      |
| ehartford/dolphin-2.1-mistral-7b (Duplicate?)    | 67      | 63.99 | 85        | 63.44| 55.57      |
| Undi95/Mistral-11B-TestBench10                   | 66.99   | 64.25 | 84.24     | 63.9 | 55.57      |
| teknuim/CollectiveCognition-v1.1-Mistral-7B      | 66.56   | 62.12 | 84.17     | 62.35| **57.62**  |
| Weyaxi/SlimOpenOrca-Mistral-7B                   | 66.54   | 62.97 | 83.49     | 62.3 | 57.39      |
| teknuim/CollectiveCognition-v1-Mistral-7B        | 66.28   | 62.37 | **85.5**  | 62.76| 54.48      |
| ehartford/samantha-1-2-mistral-7b                | 65.87   | 64.08 | 85.08     | 63.91| 50.4       |
| Open-Orca/Mistral-7B-SlimOrca                    | 65.85   | 62.54 | 83.86     | 62.77| 54.23      |
| Open-Orca/Mistral-7B-OpenOrca                    | 65.84   | 64.08 | 83.99     | 62.24| 53.05      |

# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_unaidedelf87777__wizard-mistral-v0.1)

| Metric                | Value                     |
|-----------------------|---------------------------|
| Avg.                  | 51.58   |
| ARC (25-shot)         | 61.77          |
| HellaSwag (10-shot)   | 83.51    |
| MMLU (5-shot)         | 63.99         |
| TruthfulQA (0-shot)   | 47.46   |
| Winogrande (5-shot)   | 78.3   |
| GSM8K (5-shot)        | 19.03        |
| DROP (3-shot)         | 7.01         |