File size: 1,502 Bytes
cd980f4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
74312aa
cd980f4
0294241
1180cc3
cd980f4
 
 
 
74028b5
cd980f4
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
---
language:
- en
- de
- fr
- zh
- pt
- nl
- ru
- ko
- it
- es
license: cc-by-nc-4.0
metrics:
- comet
pipeline_tag: translation
---
# Model Card for TowerInstruct-WMT24-Chat-7B

## Model Details

### Model Description

TowerInstruct-WMT24-Chat-7B is a language model that results from fine-tuning TowerBase on TowerBlocks and the WMT24 Chat MT Shared task training set.

TowerInstruct-WMT24-Chat-7B was the best submission of the shared task, winning on all 10 language pairs according to human evaluation (see the task's findings paper [here](https://aclanthology.org/2024.wmt-1.59.pdf)).

It is specifically tailoured for context-aware translation of customer support chats.

Check out our [paper](https://aclanthology.org/2024.wmt-1.100/) for more details and information on training and data.

Information on model usage, out-of-scope usages, risks, etc... are the same as the model cards of the TowerInstruct models. 

## Citation 

```bibtex
@inproceedings{pombal2024improving,
  title={Improving Context Usage for Translating Bilingual Customer Support Chat with Large Language Models},
  author={Pombal, Jos{\'e} and Agrawal, Sweta and Martins, Andr{\'e} FT},
  booktitle={Proceedings of the Ninth Conference on Machine Translation},
  pages={993--1003},
  year={2024}
}
```

[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)