Skip to content

Latest commit

 

History

History
 
 

eng-zls

opus-2020-06-28.zip

  • dataset: opus
  • model: transformer
  • source language(s): eng
  • target language(s): bul bul_Latn mkd slv
  • model: transformer
  • pre-processing: normalization + SentencePiece (spm32k,spm32k)
  • a sentence initial language token is required in the form of >>id<< (id = valid target language ID)
  • download: opus-2020-06-28.zip
  • test set translations: opus-2020-06-28.test.txt
  • test set scores: opus-2020-06-28.eval.txt

Benchmarks

testset BLEU chr-F
Tatoeba-test.eng-bul.eng.bul 46.5 0.648
Tatoeba-test.eng-mkd.eng.mkd 44.1 0.635
Tatoeba-test.eng.multi 41.8 0.612
Tatoeba-test.eng-slv.eng.slv 17.9 0.353

opus-2020-07-06.zip

  • dataset: opus
  • model: transformer
  • source language(s): eng
  • target language(s): bos_Latn bul bul_Latn hrv mkd slv srp_Cyrl srp_Latn
  • model: transformer
  • pre-processing: normalization + SentencePiece (spm32k,spm32k)
  • a sentence initial language token is required in the form of >>id<< (id = valid target language ID)
  • download: opus-2020-07-06.zip
  • test set translations: opus-2020-07-06.test.txt
  • test set scores: opus-2020-07-06.eval.txt

Benchmarks

testset BLEU chr-F
Tatoeba-test.eng-bul.eng.bul 46.2 0.646
Tatoeba-test.eng-hbs.eng.hbs 0.8 0.051
Tatoeba-test.eng-mkd.eng.mkd 43.5 0.629
Tatoeba-test.eng.multi 42.4 0.612
Tatoeba-test.eng-slv.eng.slv 17.6 0.348

opus-2020-07-27.zip

  • dataset: opus
  • model: transformer
  • source language(s): eng
  • target language(s): bos_Latn bul bul_Latn hrv mkd slv srp_Cyrl srp_Latn
  • model: transformer
  • pre-processing: normalization + SentencePiece (spm32k,spm32k)
  • a sentence initial language token is required in the form of >>id<< (id = valid target language ID)
  • download: opus-2020-07-27.zip
  • test set translations: opus-2020-07-27.test.txt
  • test set scores: opus-2020-07-27.eval.txt

Benchmarks

testset BLEU chr-F
Tatoeba-test.eng-bul.eng.bul 46.3 0.648
Tatoeba-test.eng-hbs.eng.hbs 40.3 0.613
Tatoeba-test.eng-mkd.eng.mkd 44.4 0.636
Tatoeba-test.eng.multi 41.9 0.615
Tatoeba-test.eng-slv.eng.slv 18.2 0.351

opus2m-2020-08-02.zip

  • dataset: opus2m
  • model: transformer
  • source language(s): eng
  • target language(s): bos_Latn bul bul_Latn hrv mkd slv srp_Cyrl srp_Latn
  • model: transformer
  • pre-processing: normalization + SentencePiece (spm32k,spm32k)
  • a sentence initial language token is required in the form of >>id<< (id = valid target language ID)
  • download: opus2m-2020-08-02.zip
  • test set translations: opus2m-2020-08-02.test.txt
  • test set scores: opus2m-2020-08-02.eval.txt

Benchmarks

testset BLEU chr-F
Tatoeba-test.eng-bul.eng.bul 47.6 0.657
Tatoeba-test.eng-hbs.eng.hbs 40.7 0.619
Tatoeba-test.eng-mkd.eng.mkd 45.2 0.642
Tatoeba-test.eng.multi 42.7 0.622
Tatoeba-test.eng-slv.eng.slv 17.9 0.351