Skip to content

Latest commit

 

History

History
 
 

itc-eng

opus-2020-07-14.zip

  • dataset: opus
  • model: transformer
  • source language(s): arg ast cat cos egl ext fra frm_Latn gcf_Latn glg hat ind ita lad lad_Latn lat_Latn lij lld_Latn lmo max_Latn mfe min mwl oci pap pms por roh ron scn spa tmw_Latn vec wln zlm_Latn zsm_Latn
  • target language(s): eng
  • model: transformer
  • pre-processing: normalization + SentencePiece (spm32k,spm32k)
  • download: opus-2020-07-14.zip
  • test set translations: opus-2020-07-14.test.txt
  • test set scores: opus-2020-07-14.eval.txt

Benchmarks

testset BLEU chr-F
Tatoeba-test.arg-eng.arg.eng 41.4 0.515
Tatoeba-test.ast-eng.ast.eng 33.2 0.493
Tatoeba-test.cat-eng.cat.eng 53.3 0.692
Tatoeba-test.cos-eng.cos.eng 64.2 0.757
Tatoeba-test.egl-eng.egl.eng 3.2 0.181
Tatoeba-test.ext-eng.ext.eng 41.0 0.574
Tatoeba-test.fra-eng.fra.eng 52.8 0.682
Tatoeba-test.frm-eng.frm.eng 23.4 0.409
Tatoeba-test.gcf-eng.gcf.eng 12.6 0.281
Tatoeba-test.glg-eng.glg.eng 53.9 0.696
Tatoeba-test.hat-eng.hat.eng 39.8 0.575
Tatoeba-test.ita-eng.ita.eng 64.0 0.761
Tatoeba-test.lad-eng.lad.eng 21.3 0.443
Tatoeba-test.lat-eng.lat.eng 19.1 0.383
Tatoeba-test.lij-eng.lij.eng 9.4 0.274
Tatoeba-test.lld-eng.lld.eng 17.4 0.326
Tatoeba-test.lmo-eng.lmo.eng 10.9 0.300
Tatoeba-test.mfe-eng.mfe.eng 66.0 0.820
Tatoeba-test.multi.eng 47.1 0.629
Tatoeba-test.mwl-eng.mwl.eng 28.6 0.609
Tatoeba-test.oci-eng.oci.eng 20.3 0.391
Tatoeba-test.pap-eng.pap.eng 52.3 0.634
Tatoeba-test.pms-eng.pms.eng 10.7 0.322
Tatoeba-test.por-eng.por.eng 58.8 0.731
Tatoeba-test.roh-eng.roh.eng 13.8 0.384
Tatoeba-test.ron-eng.ron.eng 54.1 0.694
Tatoeba-test.scn-eng.scn.eng 45.5 0.469
Tatoeba-test.spa-eng.spa.eng 55.5 0.709
Tatoeba-test.vec-eng.vec.eng 21.3 0.364
Tatoeba-test.wln-eng.wln.eng 13.7 0.294

opus-2020-07-19.zip

  • dataset: opus
  • model: transformer
  • source language(s): arg ast cat cos egl ext fra frm_Latn gcf_Latn glg hat ind ita lad lad_Latn lat_Latn lij lld_Latn lmo max_Latn mfe min mwl oci pap pms por roh ron scn spa tmw_Latn vec wln zlm_Latn zsm_Latn
  • target language(s): eng
  • model: transformer
  • pre-processing: normalization + SentencePiece (spm32k,spm32k)
  • download: opus-2020-07-19.zip
  • test set translations: opus-2020-07-19.test.txt
  • test set scores: opus-2020-07-19.eval.txt

Benchmarks

testset BLEU chr-F
Tatoeba-test.arg-eng.arg.eng 40.5 0.512
Tatoeba-test.ast-eng.ast.eng 33.6 0.490
Tatoeba-test.cat-eng.cat.eng 54.3 0.698
Tatoeba-test.cos-eng.cos.eng 60.3 0.690
Tatoeba-test.egl-eng.egl.eng 4.7 0.190
Tatoeba-test.ext-eng.ext.eng 42.7 0.567
Tatoeba-test.fra-eng.fra.eng 52.9 0.684
Tatoeba-test.frm-eng.frm.eng 25.0 0.409
Tatoeba-test.gcf-eng.gcf.eng 17.8 0.330
Tatoeba-test.glg-eng.glg.eng 54.1 0.697
Tatoeba-test.hat-eng.hat.eng 39.8 0.573
Tatoeba-test.ita-eng.ita.eng 64.1 0.762
Tatoeba-test.lad-eng.lad.eng 19.8 0.433
Tatoeba-test.lat-eng.lat.eng 18.8 0.381
Tatoeba-test.lij-eng.lij.eng 9.0 0.280
Tatoeba-test.lld-eng.lld.eng 18.3 0.332
Tatoeba-test.lmo-eng.lmo.eng 10.8 0.300
Tatoeba-test.mfe-eng.mfe.eng 56.9 0.702
Tatoeba-test.msa-eng.msa.eng 40.1 0.587
Tatoeba-test.multi.eng 47.3 0.630
Tatoeba-test.mwl-eng.mwl.eng 41.3 0.707
Tatoeba-test.oci-eng.oci.eng 20.3 0.396
Tatoeba-test.pap-eng.pap.eng 55.5 0.653
Tatoeba-test.pms-eng.pms.eng 12.2 0.325
Tatoeba-test.por-eng.por.eng 58.8 0.731
Tatoeba-test.roh-eng.roh.eng 12.8 0.383
Tatoeba-test.ron-eng.ron.eng 54.3 0.695
Tatoeba-test.scn-eng.scn.eng 50.9 0.503
Tatoeba-test.spa-eng.spa.eng 55.6 0.710
Tatoeba-test.vec-eng.vec.eng 29.3 0.393
Tatoeba-test.wln-eng.wln.eng 8.5 0.272

opus-2020-07-27.zip

  • dataset: opus
  • model: transformer
  • source language(s): arg ast cat cos egl ext fra frm_Latn gcf_Latn glg hat ind ita lad lad_Latn lat_Latn lij lld_Latn lmo max_Latn mfe min mwl oci pap pms por roh ron scn spa tmw_Latn vec wln zlm_Latn zsm_Latn
  • target language(s): eng
  • model: transformer
  • pre-processing: normalization + SentencePiece (spm32k,spm32k)
  • download: opus-2020-07-27.zip
  • test set translations: opus-2020-07-27.test.txt
  • test set scores: opus-2020-07-27.eval.txt

Benchmarks

testset BLEU chr-F
newsdev2016-enro-roneng.ron.eng 36.0 0.623
newsdiscussdev2015-enfr-fraeng.fra.eng 30.7 0.560
newsdiscusstest2015-enfr-fraeng.fra.eng 35.4 0.587
newssyscomb2009-fraeng.fra.eng 28.5 0.555
newssyscomb2009-itaeng.ita.eng 32.0 0.581
newssyscomb2009-spaeng.spa.eng 29.0 0.562
news-test2008-fraeng.fra.eng 25.0 0.530
news-test2008-spaeng.spa.eng 26.2 0.539
newstest2009-fraeng.fra.eng 27.9 0.553
newstest2009-itaeng.ita.eng 31.4 0.577
newstest2009-spaeng.spa.eng 28.8 0.558
newstest2010-fraeng.fra.eng 29.2 0.568
newstest2010-spaeng.spa.eng 32.9 0.593
newstest2011-fraeng.fra.eng 30.7 0.580
newstest2011-spaeng.spa.eng 31.7 0.581
newstest2012-fraeng.fra.eng 30.7 0.574
newstest2012-spaeng.spa.eng 34.8 0.602
newstest2013-fraeng.fra.eng 31.4 0.571
newstest2013-spaeng.spa.eng 32.3 0.588
newstest2014-fren-fraeng.fra.eng 34.0 0.606
newstest2016-enro-roneng.ron.eng 34.5 0.605
Tatoeba-test.arg-eng.arg.eng 40.9 0.512
Tatoeba-test.ast-eng.ast.eng 34.8 0.503
Tatoeba-test.cat-eng.cat.eng 54.4 0.699
Tatoeba-test.cos-eng.cos.eng 63.2 0.659
Tatoeba-test.egl-eng.egl.eng 2.7 0.185
Tatoeba-test.ext-eng.ext.eng 46.4 0.593
Tatoeba-test.fra-eng.fra.eng 53.0 0.685
Tatoeba-test.frm-eng.frm.eng 23.2 0.423
Tatoeba-test.gcf-eng.gcf.eng 17.1 0.310
Tatoeba-test.glg-eng.glg.eng 54.1 0.697
Tatoeba-test.hat-eng.hat.eng 38.3 0.565
Tatoeba-test.ita-eng.ita.eng 64.5 0.764
Tatoeba-test.lad-eng.lad.eng 14.7 0.430
Tatoeba-test.lat-eng.lat.eng 18.8 0.382
Tatoeba-test.lij-eng.lij.eng 9.5 0.275
Tatoeba-test.lld-eng.lld.eng 16.7 0.330
Tatoeba-test.lmo-eng.lmo.eng 10.7 0.315
Tatoeba-test.mfe-eng.mfe.eng 66.0 0.820
Tatoeba-test.msa-eng.msa.eng 40.2 0.587
Tatoeba-test.multi.eng 47.2 0.631
Tatoeba-test.mwl-eng.mwl.eng 41.3 0.707
Tatoeba-test.oci-eng.oci.eng 19.8 0.395
Tatoeba-test.pap-eng.pap.eng 53.6 0.647
Tatoeba-test.pms-eng.pms.eng 11.3 0.325
Tatoeba-test.por-eng.por.eng 59.1 0.733
Tatoeba-test.roh-eng.roh.eng 14.9 0.404
Tatoeba-test.ron-eng.ron.eng 54.2 0.696
Tatoeba-test.scn-eng.scn.eng 40.3 0.403
Tatoeba-test.spa-eng.spa.eng 55.6 0.711
Tatoeba-test.vec-eng.vec.eng 36.2 0.459
Tatoeba-test.wln-eng.wln.eng 18.5 0.335

opus2m-2020-08-01.zip

  • dataset: opus2m
  • model: transformer
  • source language(s): arg ast cat cos egl ext fra frm_Latn gcf_Latn glg hat ind ita lad lad_Latn lat_Latn lij lld_Latn lmo max_Latn mfe min mwl oci pap pms por roh ron scn spa tmw_Latn vec wln zlm_Latn zsm_Latn
  • target language(s): eng
  • model: transformer
  • pre-processing: normalization + SentencePiece (spm32k,spm32k)
  • download: opus2m-2020-08-01.zip
  • test set translations: opus2m-2020-08-01.test.txt
  • test set scores: opus2m-2020-08-01.eval.txt

Benchmarks

testset BLEU chr-F
newsdev2016-enro-roneng.ron.eng 36.5 0.628
newsdiscussdev2015-enfr-fraeng.fra.eng 30.9 0.561
newsdiscusstest2015-enfr-fraeng.fra.eng 35.5 0.590
newssyscomb2009-fraeng.fra.eng 29.2 0.560
newssyscomb2009-itaeng.ita.eng 32.2 0.583
newssyscomb2009-spaeng.spa.eng 29.3 0.563
news-test2008-fraeng.fra.eng 25.2 0.531
news-test2008-spaeng.spa.eng 26.3 0.539
newstest2009-fraeng.fra.eng 28.5 0.555
newstest2009-itaeng.ita.eng 31.6 0.578
newstest2009-spaeng.spa.eng 28.7 0.558
newstest2010-fraeng.fra.eng 29.7 0.571
newstest2010-spaeng.spa.eng 32.8 0.593
newstest2011-fraeng.fra.eng 30.9 0.580
newstest2011-spaeng.spa.eng 31.8 0.582
newstest2012-fraeng.fra.eng 31.1 0.576
newstest2012-spaeng.spa.eng 35.0 0.604
newstest2013-fraeng.fra.eng 31.7 0.573
newstest2013-spaeng.spa.eng 32.4 0.589
newstest2014-fren-fraeng.fra.eng 34.0 0.606
newstest2016-enro-roneng.ron.eng 34.8 0.608
Tatoeba-test.arg-eng.arg.eng 41.5 0.528
Tatoeba-test.ast-eng.ast.eng 36.0 0.519
Tatoeba-test.cat-eng.cat.eng 53.7 0.696
Tatoeba-test.cos-eng.cos.eng 56.5 0.640
Tatoeba-test.egl-eng.egl.eng 4.6 0.217
Tatoeba-test.ext-eng.ext.eng 39.1 0.547
Tatoeba-test.fra-eng.fra.eng 53.4 0.688
Tatoeba-test.frm-eng.frm.eng 22.3 0.409
Tatoeba-test.gcf-eng.gcf.eng 18.7 0.308
Tatoeba-test.glg-eng.glg.eng 54.8 0.701
Tatoeba-test.hat-eng.hat.eng 42.6 0.583
Tatoeba-test.ita-eng.ita.eng 64.8 0.767
Tatoeba-test.lad-eng.lad.eng 14.4 0.433
Tatoeba-test.lat-eng.lat.eng 19.5 0.390
Tatoeba-test.lij-eng.lij.eng 8.9 0.280
Tatoeba-test.lld-eng.lld.eng 17.4 0.331
Tatoeba-test.lmo-eng.lmo.eng 10.8 0.306
Tatoeba-test.mfe-eng.mfe.eng 66.0 0.820
Tatoeba-test.msa-eng.msa.eng 40.8 0.590
Tatoeba-test.multi.eng 47.6 0.634
Tatoeba-test.mwl-eng.mwl.eng 41.3 0.707
Tatoeba-test.oci-eng.oci.eng 20.3 0.401
Tatoeba-test.pap-eng.pap.eng 53.9 0.642
Tatoeba-test.pms-eng.pms.eng 12.2 0.334
Tatoeba-test.por-eng.por.eng 59.3 0.734
Tatoeba-test.roh-eng.roh.eng 17.7 0.420
Tatoeba-test.ron-eng.ron.eng 54.5 0.697
Tatoeba-test.scn-eng.scn.eng 40.0 0.443
Tatoeba-test.spa-eng.spa.eng 55.9 0.712
Tatoeba-test.vec-eng.vec.eng 11.2 0.304
Tatoeba-test.wln-eng.wln.eng 20.9 0.360

opus4m-2020-08-12.zip

  • dataset: opus4m
  • model: transformer
  • source language(s): arg ast cat cos egl ext fra frm_Latn gcf_Latn glg hat ind ita lad lad_Latn lat_Latn lij lld_Latn lmo max_Latn mfe min mwl oci pap pms por roh ron scn spa tmw_Latn vec wln zlm_Latn zsm_Latn
  • target language(s): eng
  • model: transformer
  • pre-processing: normalization + SentencePiece (spm32k,spm32k)
  • download: opus4m-2020-08-12.zip
  • test set translations: opus4m-2020-08-12.test.txt
  • test set scores: opus4m-2020-08-12.eval.txt

Benchmarks

testset BLEU chr-F
newsdev2016-enro-roneng.ron.eng 36.8 0.628
newsdiscussdev2015-enfr-fraeng.fra.eng 31.2 0.563
newsdiscusstest2015-enfr-fraeng.fra.eng 35.2 0.587
newssyscomb2009-fraeng.fra.eng 29.4 0.562
newssyscomb2009-itaeng.ita.eng 33.0 0.587
newssyscomb2009-spaeng.spa.eng 29.1 0.562
news-test2008-fraeng.fra.eng 25.2 0.531
news-test2008-spaeng.spa.eng 26.3 0.539
newstest2009-fraeng.fra.eng 28.3 0.555
newstest2009-itaeng.ita.eng 31.7 0.579
newstest2009-spaeng.spa.eng 28.9 0.559
newstest2010-fraeng.fra.eng 29.9 0.572
newstest2010-spaeng.spa.eng 33.0 0.594
newstest2011-fraeng.fra.eng 30.9 0.581
newstest2011-spaeng.spa.eng 32.0 0.582
newstest2012-fraeng.fra.eng 31.5 0.578
newstest2012-spaeng.spa.eng 35.0 0.604
newstest2013-fraeng.fra.eng 31.6 0.572
newstest2013-spaeng.spa.eng 32.4 0.589
newstest2014-fren-fraeng.fra.eng 34.0 0.607
newstest2016-enro-roneng.ron.eng 35.0 0.608
Tatoeba-test.arg-eng.arg.eng 43.7 0.539
Tatoeba-test.ast-eng.ast.eng 37.7 0.532
Tatoeba-test.cat-eng.cat.eng 54.2 0.700
Tatoeba-test.cos-eng.cos.eng 62.9 0.632
Tatoeba-test.egl-eng.egl.eng 4.2 0.204
Tatoeba-test.ext-eng.ext.eng 46.8 0.597
Tatoeba-test.fra-eng.fra.eng 53.4 0.688
Tatoeba-test.frm-eng.frm.eng 26.4 0.434
Tatoeba-test.gcf-eng.gcf.eng 17.1 0.294
Tatoeba-test.glg-eng.glg.eng 53.9 0.699
Tatoeba-test.hat-eng.hat.eng 42.0 0.578
Tatoeba-test.ita-eng.ita.eng 64.7 0.767
Tatoeba-test.lad-eng.lad.eng 19.8 0.444
Tatoeba-test.lat-eng.lat.eng 19.8 0.392
Tatoeba-test.lij-eng.lij.eng 8.5 0.289
Tatoeba-test.lld-eng.lld.eng 7.2 0.266
Tatoeba-test.lmo-eng.lmo.eng 11.9 0.324
Tatoeba-test.mfe-eng.mfe.eng 66.0 0.820
Tatoeba-test.msa-eng.msa.eng 40.7 0.590
Tatoeba-test.multi.eng 47.8 0.636
Tatoeba-test.mwl-eng.mwl.eng 41.3 0.707
Tatoeba-test.oci-eng.oci.eng 20.4 0.400
Tatoeba-test.pap-eng.pap.eng 53.6 0.647
Tatoeba-test.pms-eng.pms.eng 12.8 0.330
Tatoeba-test.por-eng.por.eng 59.5 0.737
Tatoeba-test.roh-eng.roh.eng 17.4 0.405
Tatoeba-test.ron-eng.ron.eng 54.3 0.696
Tatoeba-test.scn-eng.scn.eng 38.3 0.422
Tatoeba-test.spa-eng.spa.eng 55.8 0.712
Tatoeba-test.vec-eng.vec.eng 7.8 0.313
Tatoeba-test.wln-eng.wln.eng 17.9 0.339