File size: 3,870 Bytes
178190e daa54e8 178190e daa54e8 9dfd5fb daa54e8 e1aa1a3 daa54e8 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 |
---
base_model: mistralai/Mistral-7B-v0.1
tags:
- mistral
- instruct
- bggpt
- insait
language:
- bg
- en
library_name: transformers
pipeline_tag: text-generation
license: apache-2.0
---
# INSAIT-Institute/BgGPT-7B-Instruct-v0.2
![image/png](https://cdn-uploads.huggingface.co/production/uploads/637e1f8cf7e01589cc17bf7e/p6d0YFHjWCQ3S12jWqO1m.png)
Meet BgGPT-7B, a Bulgarian language model trained from mistralai/Mistral-7B-v0.1. BgGPT is distributed under Apache 2.0 license.
This model was created by [`INSAIT Institute`](https://identrics.ai/), in the scope of the wasper project.
This is an improved version of the model - v0.2.
## Model description
The model is continously pretrained to gain its Bulgarian language and culture capabilities using multiple datasets, including Bulgarian web crawl data, a range of specialized Bulgarian datasets sourced by INSAIT Institute, and machine translations of popular English datasets.
This Bulgarian data was augmented with English datasets to retain English and logical reasoning skills.
The model's tokenizer has been extended to allow for a more efficient encoding of Bulgarian words written in Cyrillic.
This not only increases throughput of Cyrillic text but also performance.
## Instruction format
In order to leverage instruction fine-tuning, your prompt should be surrounded by `[INST]` and `[/INST]` tokens.
The very first instruction should begin with a begin of sequence token `<s>`. Following instructions should not.
The assistant generation will be ended by the end-of-sequence token.
E.g.
```
text = "<s>[INST] Кога е основан Софийският университет? [/INST]"
"Софийският университет „Св. Климент Охридски“ е създаден на 1 октомври 1888 г.</s> "
"[INST] Кой го е основал? [/INST]"
```
This format is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method:
## Benchmarks
The model comes with a set of Benchmarks that are translations of the corresponding English-benchmarks. These are provided at [`https://github.com/insait-institute/lm-evaluation-harness-bg`](https://github.com/insait-institute/lm-evaluation-harness-bg)
As this is an improved version over version 0.1 of the same model and we include benchmark comparisons.
![image/png](https://cdn-uploads.huggingface.co/production/uploads/637e1f8cf7e01589cc17bf7e/aZAEv5qyLcPn5p4KrHpEw.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/637e1f8cf7e01589cc17bf7e/6PafMC6StfUaPY1N8Xrta.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/637e1f8cf7e01589cc17bf7e/L1bKXq4Xiik1ZbTDuCnxj.png)
## Summary
- **Finetuned from:** [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1)
- **Model type:** Causal decoder-only transformer language model
- **Language:** Bulgarian and English
- **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.html)
- **Contact:** [[email protected]](mailto:[email protected])
## Use in 🤗Transformers
First install direct dependencies:
```
pip install transformers torch accelerate
```
If you want faster inference using flash-attention2, you need to install these dependencies:
```bash
pip install packaging ninja
pip install flash-attn
```
Then load the model in transformers:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
model = AutoModelForCausalLM.from_pretrained(
model="INSAIT-Institute/BgGPT-7B-Instruct-v0.2",
device_map="auto",
torch_dtype=torch.bfloat16,
use_flash_attn_2=True # optional
)
```
## Use with GGML / llama.cpp
The model in GGUF format [INSAIT-Institute/BgGPT-7B-Instruct-v0.2-GGUF](https://huggingface.co/INSAIT-Institute/BgGPT-7B-Instruct-v0.2-GGUF)
|