File size: 1,051 Bytes
97a7038
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3760b8f
 
97a7038
3760b8f
97a7038
 
3760b8f
 
97a7038
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
---
license: openrail
datasets:
- JosephusCheung/GuanacoDataset
language:
- en
- zh
- ja
---

# Guanaco: A Multilingual Instruction-Following Language Model Based on LLaMA 7B

This model is trained with modified [alpaca-lora](https://github.com/tloen/alpaca-lora) with lora + embed_tokens + lm_head be trained.

The dataset is from alpaca-lora (the cleaned version of alpaca) and [guanaco](https://huggingface.co/datasets/JosephusCheung/GuanacoDataset).
With trained embed and head, the model perform better at Chinese and Japanese then original LLaMA, and with instruction based prompt. You can use this model more easily.

Since this model is trained by guanaco dataset, you can also use this as chatbot. just use this format:
```
### Instruction:
User: <Message history>
Assistant: <Message history>

### Input:
System: <System response for next message, optional>
User: <Next message>

### Response:
```

**Tips: I just removed the first line of original prompt to reduce token comsumption, plz consider remove it when you want to use this model**