eng-alv

  • source group: English

  • target group: Atlantic-Congo languages

  • OPUS readme: eng-alv

  • model: transformer

  • source language(s): eng

  • target language(s): ewe fuc fuv ibo kin lin lug nya run sag sna swh toi_Latn tso umb wol xho yor zul

  • model: transformer

  • pre-processing: normalization + SentencePiece (spm32k,spm32k)

  • a sentence initial language token is required in the form of >>id<< (id = valid target language ID)

  • download original weights: opus2m-2020-08-01.zip

  • test set translations: opus2m-2020-08-01.test.txt

  • test set scores: opus2m-2020-08-01.eval.txt

Benchmarks

testset BLEU chr-F
Tatoeba-test.eng-ewe.eng.ewe 4.9 0.212
Tatoeba-test.eng-ful.eng.ful 0.6 0.079
Tatoeba-test.eng-ibo.eng.ibo 3.5 0.255
Tatoeba-test.eng-kin.eng.kin 10.5 0.510
Tatoeba-test.eng-lin.eng.lin 1.1 0.273
Tatoeba-test.eng-lug.eng.lug 5.3 0.340
Tatoeba-test.eng.multi 11.4 0.429
Tatoeba-test.eng-nya.eng.nya 18.1 0.595
Tatoeba-test.eng-run.eng.run 13.9 0.484
Tatoeba-test.eng-sag.eng.sag 5.3 0.194
Tatoeba-test.eng-sna.eng.sna 26.2 0.623
Tatoeba-test.eng-swa.eng.swa 1.0 0.141
Tatoeba-test.eng-toi.eng.toi 7.0 0.224
Tatoeba-test.eng-tso.eng.tso 46.7 0.643
Tatoeba-test.eng-umb.eng.umb 7.8 0.359
Tatoeba-test.eng-wol.eng.wol 6.8 0.191
Tatoeba-test.eng-xho.eng.xho 27.1 0.629
Tatoeba-test.eng-yor.eng.yor 17.4 0.356
Tatoeba-test.eng-zul.eng.zul 34.1 0.729

System Info:

  • hf_name: eng-alv

  • source_languages: eng

  • target_languages: alv

  • opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/eng-alv/README.md

  • original_repo: Tatoeba-Challenge

  • tags: ['translation']

  • languages: ['en', 'sn', 'rw', 'wo', 'ig', 'sg', 'ee', 'zu', 'lg', 'ts', 'ln', 'ny', 'yo', 'rn', 'xh', 'alv']

  • src_constituents: {'eng'}

  • tgt_constituents: {'sna', 'kin', 'wol', 'ibo', 'swh', 'sag', 'ewe', 'zul', 'fuc', 'lug', 'tso', 'lin', 'nya', 'yor', 'run', 'xho', 'fuv', 'toi_Latn', 'umb'}

  • src_multilingual: False

  • tgt_multilingual: True

  • prepro: normalization + SentencePiece (spm32k,spm32k)

  • url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/eng-alv/opus2m-2020-08-01.zip

  • url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/eng-alv/opus2m-2020-08-01.test.txt

  • src_alpha3: eng

  • tgt_alpha3: alv

  • short_pair: en-alv

  • chrF2_score: 0.429

  • bleu: 11.4

  • brevity_penalty: 1.0

  • ref_len: 10603.0

  • src_name: English

  • tgt_name: Atlantic-Congo languages

  • train_date: 2020-08-01

  • src_alpha2: en

  • tgt_alpha2: alv

  • prefer_old: False

  • long_pair: eng-alv

  • helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535

  • transformers_git_sha: 2207e5d8cb224e954a7cba69fa4ac2309e9ff30b

  • port_machine: brutasse

  • port_time: 2020-08-21-14:41

Downloads last month
44
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Spaces using Helsinki-NLP/opus-mt-en-alv 7