Skip to content

Latest commit

 

History

History

art-eng

opus-2020-06-28.zip

  • dataset: opus
  • model: transformer
  • source language(s): afh_Latn avk_Latn bzt_Latn dws_Latn epo ido ido_Latn ile_Latn ina_Latn jbo jbo_Cyrl jbo_Latn ldn_Latn lfn_Cyrl lfn_Latn nov_Latn qya qya_Latn sjn_Latn tlh_Latn tzl tzl_Latn vol_Latn
  • target language(s): eng
  • model: transformer
  • pre-processing: normalization + SentencePiece (spm32k,spm32k)
  • download: opus-2020-06-28.zip
  • test set translations: opus-2020-06-28.test.txt
  • test set scores: opus-2020-06-28.eval.txt

Benchmarks

testset BLEU chr-F
Tatoeba-test.afh-eng.afh.eng 1.1 0.097
Tatoeba-test.avk-eng.avk.eng 0.6 0.108
Tatoeba-test.bzt-eng.bzt.eng 0.8 0.109
Tatoeba-test.dws-eng.dws.eng 0.7 0.039
Tatoeba-test.epo-eng.epo.eng 34.7 0.529
Tatoeba-test.ido-eng.ido.eng 13.8 0.318
Tatoeba-test.ile-eng.ile.eng 5.7 0.234
Tatoeba-test.ina-eng.ina.eng 5.7 0.251
Tatoeba-test.jbo-eng.jbo.eng 0.2 0.113
Tatoeba-test.ldn-eng.ldn.eng 0.3 0.082
Tatoeba-test.lfn-eng.lfn.eng 1.5 0.169
Tatoeba-test.multi.eng 11.9 0.291
Tatoeba-test.nov-eng.nov.eng 3.9 0.209
Tatoeba-test.qya-eng.qya.eng 0.3 0.076
Tatoeba-test.sjn-eng.sjn.eng 1.0 0.081
Tatoeba-test.tlh-eng.tlh.eng 0.2 0.124
Tatoeba-test.tzl-eng.tzl.eng 1.1 0.125
Tatoeba-test.vol-eng.vol.eng 0.6 0.115

opus-2020-07-26.zip

  • dataset: opus
  • model: transformer
  • source language(s): afh_Latn avk_Latn dws_Latn epo ido ido_Latn ile_Latn ina_Latn jbo jbo_Cyrl jbo_Latn ldn_Latn lfn_Cyrl lfn_Latn nov_Latn qya qya_Latn sjn_Latn tlh_Latn tzl tzl_Latn vol_Latn
  • target language(s): eng
  • model: transformer
  • pre-processing: normalization + SentencePiece (spm32k,spm32k)
  • download: opus-2020-07-26.zip
  • test set translations: opus-2020-07-26.test.txt
  • test set scores: opus-2020-07-26.eval.txt

Benchmarks

testset BLEU chr-F
Tatoeba-test.afh-eng.afh.eng 1.5 0.084
Tatoeba-test.avk-eng.avk.eng 0.4 0.104
Tatoeba-test.dws-eng.dws.eng 0.5 0.060
Tatoeba-test.epo-eng.epo.eng 34.8 0.529
Tatoeba-test.ido-eng.ido.eng 13.0 0.310
Tatoeba-test.ile-eng.ile.eng 5.2 0.227
Tatoeba-test.ina-eng.ina.eng 5.5 0.250
Tatoeba-test.jbo-eng.jbo.eng 0.2 0.111
Tatoeba-test.ldn-eng.ldn.eng 0.3 0.075
Tatoeba-test.lfn-eng.lfn.eng 1.8 0.171
Tatoeba-test.multi.eng 11.6 0.284
Tatoeba-test.nov-eng.nov.eng 4.2 0.210
Tatoeba-test.qya-eng.qya.eng 0.3 0.099
Tatoeba-test.sjn-eng.sjn.eng 0.5 0.091
Tatoeba-test.tlh-eng.tlh.eng 0.2 0.124
Tatoeba-test.tzl-eng.tzl.eng 1.0 0.115
Tatoeba-test.vol-eng.vol.eng 0.6 0.118

opus2m-2020-08-12.zip

  • dataset: opus2m
  • model: transformer
  • source language(s): afh_Latn avk_Latn dws_Latn epo ido ido_Latn ile_Latn ina_Latn jbo jbo_Cyrl jbo_Latn ldn_Latn lfn_Cyrl lfn_Latn nov_Latn qya qya_Latn sjn_Latn tlh_Latn tzl tzl_Latn vol_Latn
  • target language(s): eng
  • model: transformer
  • pre-processing: normalization + SentencePiece (spm32k,spm32k)
  • download: opus2m-2020-08-12.zip
  • test set translations: opus2m-2020-08-12.test.txt
  • test set scores: opus2m-2020-08-12.eval.txt

Benchmarks

testset BLEU chr-F
Tatoeba-test.afh-eng.afh.eng 1.2 0.099
Tatoeba-test.avk-eng.avk.eng 0.4 0.105
Tatoeba-test.dws-eng.dws.eng 1.6 0.076
Tatoeba-test.epo-eng.epo.eng 34.6 0.530
Tatoeba-test.ido-eng.ido.eng 12.7 0.310
Tatoeba-test.ile-eng.ile.eng 4.6 0.218
Tatoeba-test.ina-eng.ina.eng 5.8 0.254
Tatoeba-test.jbo-eng.jbo.eng 0.2 0.115
Tatoeba-test.ldn-eng.ldn.eng 0.7 0.083
Tatoeba-test.lfn-eng.lfn.eng 1.8 0.172
Tatoeba-test.multi.eng 11.6 0.287
Tatoeba-test.nov-eng.nov.eng 5.1 0.215
Tatoeba-test.qya-eng.qya.eng 0.7 0.113
Tatoeba-test.sjn-eng.sjn.eng 0.9 0.090
Tatoeba-test.tlh-eng.tlh.eng 0.2 0.124
Tatoeba-test.tzl-eng.tzl.eng 1.4 0.109
Tatoeba-test.vol-eng.vol.eng 0.5 0.115

opus1m+bt-2021-05-01.zip

Benchmarks

testset BLEU chr-F #sent #words BP
Tatoeba-test.afh-eng 1.7 0.090 10 54 0.943
Tatoeba-test.avk-eng 0.4 0.108 167 1238 0.768
Tatoeba-test.bzt-eng 1.7 0.124 62 328 1.000
Tatoeba-test.dws-eng 1.2 0.059 10 52 0.741
Tatoeba-test.epo-eng 42.1 0.605 10000 79703 0.973
Tatoeba-test.ido-eng 32.5 0.512 1968 14120 0.988
Tatoeba-test.ido_Latn-eng 32.5 0.512 1967 14114 0.988
Tatoeba-test.ile-eng 13.1 0.323 1711 10439 1.000
Tatoeba-test.ina-eng 16.6 0.399 5000 42826 1.000
Tatoeba-test.jbo_Cyrl-eng 0.3 0.000 1 13 1.000
Tatoeba-test.jbo-eng 0.1 0.104 5000 34483 1.000
Tatoeba-test.jbo_Latn-eng 0.1 0.104 4996 34454 1.000
Tatoeba-test.ldn-eng 0.8 0.097 101 575 1.000
Tatoeba-test.lfn_Cyrl-eng 0.2 0.054 847 5583 0.823
Tatoeba-test.lfn-eng 7.3 0.227 3297 22209 1.000
Tatoeba-test.lfn_Latn-eng 8.8 0.283 2450 16626 1.000
Tatoeba-test.multi-eng 20.0 0.364 10000 73214 1.000
Tatoeba-test.nov-eng 15.2 0.361 198 1318 1.000
Tatoeba-test.qya-eng 0.9 0.102 116 664 0.745
Tatoeba-test.qya_Latn-eng 0.9 0.102 115 654 0.741
Tatoeba-test.sjn-eng 0.6 0.081 44 235 0.835
Tatoeba-test.tlh-eng 0.3 0.109 5000 33078 1.000
Tatoeba-test.tzl-eng 1.0 0.156 166 725 0.952
Tatoeba-test.tzl_Latn-eng 1.0 0.156 165 723 0.952
Tatoeba-test.vol-eng 0.7 0.122 1549 9888 0.781

opus4m+btTCv20210807-2021-09-30.zip

Benchmarks

testset BLEU chr-F #sent #words BP
Tatoeba-test-v2021-08-07.multi-eng 27.0 0.441 10000 75207 1.000
Tatoeba-test-v2021-08-07.multi-multi 27.0 0.441 10000 75207 1.000