File size: 1,096 Bytes
8461737
 
28c2541
 
 
0c93012
 
 
58b5df6
 
 
 
0c93012
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b137d82
 
c45f845
 
 
285fb2a
b137d82
 
 
285fb2a
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
This is a llama-13B based model. (sorry, I forgot to put it in the model name)

Base Model: GPT4-x-Alpaca full fine tune by Chavinlo -> https://huggingface.co/chavinlo/gpt4-x-alpaca  
LORA fine tune using the Roleplay Instruct from GPT4 generated dataset -> https://github.com/teknium1/GPTeacher/tree/main/Roleplay  
LORA Adapter Only: https://huggingface.co/ZeusLabs/gpt4-x-alpaca-rp-lora/tree/main/gpt-rp-instruct-1  

Merged LORA to the model.

FYI Latest HF Transformers generates BROKEN generations. 
Try this instead if your generations are terrible (first uninstall transformers): pip install git+https://github.com/huggingface/transformers@9eae4aa57650c1dbe1becd4e0979f6ad1e572ac0


Instruct it same way as alpaca / gpt4xalpaca:
```
### Instruction:
<prompt>
### Response:

```

or 
```
### Instruction:
<prompt>
### Input:
<specific data to manipulate for the instruction
### Response:

```

For a better idea of prompting it for roleplay, check out the roleplay discord bot code I made here: https://github.com/teknium1/alpaca-roleplay-discordbot


---
tags:
- llama
- alpaca
- gpt4
---