File size: 2,294 Bytes
c79f12b
 
 
 
 
4659e53
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
---
license: apache-2.0
pipeline_tag: text-generation
---

![DALL·E 2023-10-11 16.30.48 - Photo of a dark forest with tall, ancient trees. Their branches and leaves form intricate patterns, revealing a portal of swirling magical energy in t.png](https://cdn-uploads.huggingface.co/production/uploads/643c81209f5d314db2de9743/FE8hhNKnRX8UOE5qXNkZD.png)

# Overview

This model is a finetune of Mistral7b on cleaned data from WizardLM Evol Instruct v2 196k. most instances of RLHF were removed from the dataset, so this should be treated as a unscensored model although it is not fully unscensored. 

# Benchmarks

Wizard Mistral was only finetuned on >200k rows of evol instruct multi turn data, however it achieves competetive results when evaluated. below is wizard mistrals benchmark scores compared to the most popular mistral7b finetunes. 

| Model                                            | Average | ARC   | HellaSwag | MMLU | TruthfulQA |
|--------------------------------------------------|---------|-------|-----------|------|------------|
| unaidedelf87777/wizard-mistral-v0.1              | 64.18   | 61.77 | 83.51     | 63.99| 47.46      |
| Undi95/Mistral-11B-TestBench11                   | **67.21**| **64.42**| 83.93     | 63.82| 56.68      |
| Undi95/Mistral-11B-TestBench9                    | 67.13   | 64.08 | 84.24     | **64** | 56.19      |
| ehartford/dolphin-2.1-mistral-7b                 | 67.06   | 64.42 | 84.92     | 63.32| 55.56      |
| ehartford/dolphin-2.1-mistral-7b (Duplicate?)    | 67      | 63.99 | 85        | 63.44| 55.57      |
| Undi95/Mistral-11B-TestBench10                   | 66.99   | 64.25 | 84.24     | 63.9 | 55.57      |
| teknuim/CollectiveCognition-v1.1-Mistral-7B      | 66.56   | 62.12 | 84.17     | 62.35| **57.62**  |
| Weyaxi/SlimOpenOrca-Mistral-7B                   | 66.54   | 62.97 | 83.49     | 62.3 | 57.39      |
| teknuim/CollectiveCognition-v1-Mistral-7B        | 66.28   | 62.37 | **85.5**  | 62.76| 54.48      |
| ehartford/samantha-1-2-mistral-7b                | 65.87   | 64.08 | 85.08     | 63.91| 50.4       |
| Open-Orca/Mistral-7B-SlimOrca                    | 65.85   | 62.54 | 83.86     | 62.77| 54.23      |
| Open-Orca/Mistral-7B-OpenOrca                    | 65.84   | 64.08 | 83.99     | 62.24| 53.05      |