|
--- |
|
language: |
|
- uk |
|
- en |
|
license: apache-2.0 |
|
library_name: peft |
|
tags: |
|
- translation |
|
- mlx |
|
datasets: |
|
- Helsinki-NLP/opus_paracrawl |
|
- turuta/Multi30k-uk |
|
metrics: |
|
- bleu |
|
pipeline_tag: text-generation |
|
base_model: mistralai/Mistral-7B-v0.1 |
|
inference: false |
|
model-index: |
|
- name: Dragoman |
|
results: |
|
- task: |
|
type: translation |
|
name: English-Ukrainian Translation |
|
dataset: |
|
name: FLORES-101 |
|
type: facebook/flores |
|
config: eng_Latn-ukr_Cyrl |
|
split: devtest |
|
metrics: |
|
- type: bleu |
|
value: 32.34 |
|
name: Test BLEU |
|
--- |
|
|
|
# lang-uk/dragoman-4bit |
|
This model was converted to MLX format from the [`lang-uk/dragoman`](https://huggingface.co/lang-uk/dragoman) adapter fused into the [`mistralai/Mistral-7b-v0.1`](https://huggingface.co/mistralai/Mistral-7B-v0.1) |
|
base model and quantized into 4 bits using mlx-lm version **0.4.0**. |
|
Refer to the [original model card](https://huggingface.co/lang-uk/dragoman) for more details on the model. |
|
## Use with mlx |
|
|
|
```bash |
|
pip install mlx-lm |
|
``` |
|
|
|
```python |
|
from mlx_lm import load, generate |
|
|
|
model, tokenizer = load("lang-uk/dragoman-4bit") |
|
response = generate(model, tokenizer, prompt="[INST] who holds this neighborhood? [/INST]", verbose=True) |
|
``` |
|
|
|
Or use from your shell: |
|
|
|
```console |
|
python -m mlx_lm.generate --model lang-uk/dragoman-4bit --prompt '[INST] who holds this neighborhood? [/INST]' --temp 0 --max-tokens 100 |
|
``` |
|
|