File size: 4,059 Bytes
43ac42a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 |
---
base_model:
- BSC-LT/salamandra-7b-instruct
datasets:
- alinia/EADOP-RAG-out-of-domain
language:
- ca
- es
library_name: transformers
license: apache-2.0
pipeline_tag: text-generation
tags:
- legal
---
# Salamandra 7B aligned EADOP Model Card
Salamandra 7B aligned EADOP is a full-finetuning version of
[BSC Language Technologies Unit](https://huggingface.co/BSC-LT)'s
[Salamandra Instruct 7B](https://huggingface.co/BSC-LT/salamandra-7b-instruct)
model of the Barcelona Supercomputing Center focused on improving
the handling of out-of-domain Questions in a RAG instruction-following setting.
The model has been finetuned on a dataset consisting of 2,000+ human annotated in-
and out-of-domain user messages and assistant responses in the context of a chatbot that can
provide helpful information about the current Catalan legislation.
The dataset [alinia/EADOP-RAG-out-of-domain](https://huggingface.co/datasets/alinia/EADOP-RAG-out-of-domain)
was collected in collaboration with the
[Entitat Autònoma del Diari Oficial i de Publicacions (EADOP)](https://dogc.gencat.cat/ca/sobre-el-dogc/eadop/)
and it consists of user messages and assistant responses in Catalan and Spanish.
> [!WARNING]
> **DISCLAIMER:** This model is a proof-of-concept designed to demonstrate the effects of
finetuning an Instruction model with a small dataset of out-of-domain questions in the model's
capability to politely and informatively refuse to answer questions that are out-of-domain.
> As a proof-of-concept, the model is still prone to generate harmful or inappropriate content.
---
## Model Details
Please refer to the [Salamandra Instruct 7B model details](https://huggingface.co/BSC-LT/salamandra-7b-instruct#model-details)
for the specific details about the model architecture and pretraining.
## Intended Use
This model was developed as a proof-of-concept to demonstrate the effects of finetuning
an Instruction model with a small dataset of in- and out-of-domain questions in the model's
capability to politely and informatively refuse to answer questions that are out-of-domain in
the context of a domain-specific RAG-based chatbot.
## How to use
This model uses the ChatML, the same instruction-following conversation format as the base model.
```python
from datetime import datetime
from transformers import AutoTokenizer, AutoModelForCausalLM
import transformers
import torch
model_id = "projecte-aina/salamandra-7b-aligned-EADOP"
text = "Quina és la finalitat del Servei Meterològic de Catalunya ?"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(
model_id,
device_map="auto",
torch_dtype=torch.bfloat16
)
message = [ { "role": "user", "content": text } ]
prompt = tokenizer.apply_chat_template(
message,
tokenize=False,
add_generation_prompt=True
)
inputs = tokenizer.encode(prompt, add_special_tokens=False, return_tensors="pt")
outputs = model.generate(input_ids=inputs.to(model.device), max_new_tokens=200)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
Using this template, each turn is preceded by a `<|im_start|>` delimiter and the role of the entity
(either `user`, for content supplied by the user, or `assistant` for LLM responses), and finished with the `<|im_end|>` token.
---
## Finetuning Data
Please refer to [alinia/EADOP-RAG-out-of-domain](https://huggingface.co/datasets/alinia/EADOP-RAG-out-of-domain) for the Dataset Card.
### Author
This model has been finetuned by [Alinia AI](https://alinia.ai/).
### Contact
For further information, please email [[email protected]](mailto:[email protected]).
### Copyright
Copyright(c) 2024 by Language Technologies Unit, Barcelona Supercomputing Center.
### License
Apache-2.0
### Funding
This work has been promoted and financed by the Generalitat de Catalunya through the Aina project.
### Acknowledgements
The data collection process was supported by the [Entitat Autònoma del Diari Oficial i de Publicacions (EADOP)](https://dogc.gencat.cat/ca/sobre-el-dogc/eadop/). |