language: sem
tags:
- translation
license: apache-2.0
sem-eng
source group: Semitic languages
target group: English
OPUS readme: sem-eng
model: transformer
source language(s): acm afb amh apc ara arq ary arz heb mlt tir
target language(s): eng
model: transformer
pre-processing: normalization + SentencePiece (spm32k,spm32k)
download original weights: opus2m-2020-08-01.zip
test set translations: opus2m-2020-08-01.test.txt
test set scores: opus2m-2020-08-01.eval.txt
Benchmarks
testset | BLEU | chr-F |
---|---|---|
Tatoeba-test.amh-eng.amh.eng | 37.5 | 0.565 |
Tatoeba-test.ara-eng.ara.eng | 38.9 | 0.566 |
Tatoeba-test.heb-eng.heb.eng | 44.6 | 0.610 |
Tatoeba-test.mlt-eng.mlt.eng | 53.7 | 0.688 |
Tatoeba-test.multi.eng | 41.7 | 0.588 |
Tatoeba-test.tir-eng.tir.eng | 18.3 | 0.370 |
System Info:
hf_name: sem-eng
source_languages: sem
target_languages: eng
opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/sem-eng/README.md
original_repo: Tatoeba-Challenge
tags: ['translation']
prepro: normalization + SentencePiece (spm32k,spm32k)
url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/sem-eng/opus2m-2020-08-01.zip
url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/sem-eng/opus2m-2020-08-01.test.txt
src_alpha3: sem
tgt_alpha3: eng
short_pair: sem-en
chrF2_score: 0.588
bleu: 41.7
brevity_penalty: 0.987
ref_len: 72950.0
src_name: Semitic languages
tgt_name: English
train_date: 2020-08-01
src_alpha2: sem
tgt_alpha2: en
prefer_old: False
long_pair: sem-eng
helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535
transformers_git_sha: 46e9f53347bbe9e989f0335f98465f30886d8173
port_machine: brutasse
port_time: 2020-08-18-01:48