RichardErkhov commited on
Commit
d2530bf
·
verified ·
1 Parent(s): 19c38a2

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +107 -0
README.md ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ Mistral-7B-Instruct-v0.1 - bnb 8bits
11
+ - Model creator: https://huggingface.co/mistralai/
12
+ - Original model: https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1/
13
+
14
+
15
+
16
+
17
+ Original model description:
18
+ ---
19
+ license: apache-2.0
20
+ pipeline_tag: text-generation
21
+ tags:
22
+ - finetuned
23
+ inference: true
24
+ widget:
25
+ - messages:
26
+ - role: user
27
+ content: What is your favorite condiment?
28
+ ---
29
+
30
+ # Model Card for Mistral-7B-Instruct-v0.1
31
+
32
+ The Mistral-7B-Instruct-v0.1 Large Language Model (LLM) is a instruct fine-tuned version of the [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) generative text model using a variety of publicly available conversation datasets.
33
+
34
+ For full details of this model please read our [paper](https://arxiv.org/abs/2310.06825) and [release blog post](https://mistral.ai/news/announcing-mistral-7b/).
35
+
36
+ ## Instruction format
37
+
38
+ In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens. The very first instruction should begin with a begin of sentence id. The next instructions should not. The assistant generation will be ended by the end-of-sentence token id.
39
+
40
+ E.g.
41
+ ```
42
+ text = "<s>[INST] What is your favourite condiment? [/INST]"
43
+ "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!</s> "
44
+ "[INST] Do you have mayonnaise recipes? [/INST]"
45
+ ```
46
+
47
+ This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method:
48
+
49
+ ```python
50
+ from transformers import AutoModelForCausalLM, AutoTokenizer
51
+
52
+ device = "cuda" # the device to load the model onto
53
+
54
+ model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")
55
+ tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.1")
56
+
57
+ messages = [
58
+ {"role": "user", "content": "What is your favourite condiment?"},
59
+ {"role": "assistant", "content": "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!"},
60
+ {"role": "user", "content": "Do you have mayonnaise recipes?"}
61
+ ]
62
+
63
+ encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
64
+
65
+ model_inputs = encodeds.to(device)
66
+ model.to(device)
67
+
68
+ generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
69
+ decoded = tokenizer.batch_decode(generated_ids)
70
+ print(decoded[0])
71
+ ```
72
+
73
+ ## Model Architecture
74
+ This instruction model is based on Mistral-7B-v0.1, a transformer model with the following architecture choices:
75
+ - Grouped-Query Attention
76
+ - Sliding-Window Attention
77
+ - Byte-fallback BPE tokenizer
78
+
79
+ ## Troubleshooting
80
+ - If you see the following error:
81
+ ```
82
+ Traceback (most recent call last):
83
+ File "", line 1, in
84
+ File "/transformers/models/auto/auto_factory.py", line 482, in from_pretrained
85
+ config, kwargs = AutoConfig.from_pretrained(
86
+ File "/transformers/models/auto/configuration_auto.py", line 1022, in from_pretrained
87
+ config_class = CONFIG_MAPPING[config_dict["model_type"]]
88
+ File "/transformers/models/auto/configuration_auto.py", line 723, in getitem
89
+ raise KeyError(key)
90
+ KeyError: 'mistral'
91
+ ```
92
+
93
+ Installing transformers from source should solve the issue
94
+ pip install git+https://github.com/huggingface/transformers
95
+
96
+ This should not be required after transformers-v4.33.4.
97
+
98
+ ## Limitations
99
+
100
+ The Mistral 7B Instruct model is a quick demonstration that the base model can be easily fine-tuned to achieve compelling performance.
101
+ It does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to
102
+ make the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs.
103
+
104
+ ## The Mistral AI Team
105
+
106
+ Albert Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lélio Renard Lavaud, Lucile Saulnier, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, William El Sayed.
107
+