File size: 1,442 Bytes
90b9193
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
aca8d98
90b9193
 
 
 
 
 
5935cf9
90b9193
aca8d98
90b9193
 
f0c5ac2
 
4e475d3
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
---
language:
- en
- fr
- de
- es
- it
- pt
- zh
- ja
- ru
- ko
license: other
license_name: mrl
base_model: mistralai/Pixtral-Large-Instruct-2411
inference: false
license_link: https://mistral.ai/licenses/MRL-0.1.md
library_name: transformers
pipeline_tag: image-text-to-text
---

# Pixtral-Large-Instruct-2411 🧡

Transformers implementation of [Pixtral-Large-Instruct-2411](https://huggingface.co/mistralai/Pixtral-Large-Instruct-2411).  
 
 
## Tokenizer And Prompt Template
Using conversion of v7m1 tokenizer with 32k vocab size.  
 
Chat template in chat_template.json uses the v7 instruct template:
 
```
<s>[SYSTEM_PROMPT] <system prompt>[/SYSTEM_PROMPT][INST] <user message>[/INST] <assistant response></s>[INST] <user message>[/INST]
```

## Notes
*- tool use hasn't been implemented in the template yet. I'll add this in later.*  
*- I've added extra stop tokens between consecutive user messages. Helps contexts where there'll be multiple speakers etc but your milage may vary.*  
*- If you have a better implementation of the tokenizer let me know and I'm happy to swap it out.*  
*- As always pls respect the model license.* 
 

## Quantizations 
EXL2 quants are available in different sizes [here](https://huggingface.co/models?author=nintwentydo&other=base_model:quantized:mistralai/Pixtral-Large-Instruct-2411). You'll need to use dev branch of [ExLlamaV2](https://github.com/turboderp/exllamav2/tree/dev) for vision input.