File size: 1,516 Bytes
24bf269 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 |
---
license: cc-by-nc-nd-4.0
language:
- en
library_name: transformers
pipeline_tag: text-generation
tags:
- Mistral
- Pygmalion
- llama-2
- llama-2-7b
---
# MistralPy-7b
This is a merger focusing on preserving the roleplay abilities of Pygmalion while gaining the improved results from Mistral. This model works best for roleplay but is still fairly capable assistant. The smaller (7b) size does mean it isn't perfect at more complex reasoning tasks, but this should be addressed in the larger version that I'll upload soon (when I can get Mistral to play along).
[GGUF version done by TheBloke](https://huggingface.co/TheBloke/Mistral-Pygmalion-7B-GGUF)
# LLM Leaderboard Evaluation
| Metric | Value |
|-----------------------|---------------------------|
| Avg. | 44.58 |
| ARC (25-shot) | 54.44 |
| HellaSwag (10-shot) | 78.48 |
| MMLU (5-shot) | 49.23 |
| TruthfulQA (0-shot) | 41.82 |
| Winogrande (5-shot) | 75.3 |
| GSM8K (5-shot) | 6.82 |
| DROP (3-shot) | 5.94 |
### Prompt Template
```
### Instruction:
{Prompt & Backstory}
### Assistant:
{Output}
```
Example:
```
### Instruction:
You are Sally, a fun 19 year old woman. Her favorite animal is "cat". Her favoritate color is "blue". She enjoys grape juice and cake.
### Assistant:
Sally: Hi, how are you?
User: Okay, you?
```
# Send a message
[Steam](https://steamcommunity.com/id/delcos/)
#### Discord: delcos69 |