NICT_LOGO.JPG KYOTO-U_LOGO.JPG

WAT

The Workshop on Asian Translation
Evaluation Results

[EVALUATION RESULTS TOP] | [BLEU] | [RIBES] | [AMFM] | [HUMAN (WAT2022)] | [HUMAN (WAT2021)] | [HUMAN (WAT2020)] | [HUMAN (WAT2019)] | [HUMAN (WAT2018)] | [HUMAN (WAT2017)] | [HUMAN (WAT2016)] | [HUMAN (WAT2015)] | [HUMAN (WAT2014)] | [EVALUATION RESULTS USAGE POLICY]

BLEU


# Team Task Date/Time DataID BLEU
Method
Other
Resources
System
Description
juman kytea mecab moses-
tokenizer
stanford-
segmenter-
ctb
stanford-
segmenter-
pku
indic-
tokenizer
unuse myseg kmseg
1ut-mrtBSDja-en2020/09/18 19:35:103966---23.80------NMTYesTransformer-base ensemble of 2 best models trained on large batches of the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 and WMT 2020 corpora; tuned on the full BSD corpus.
2goku20BSDja-en2020/09/15 19:41:513747---23.15------NMTYesmBART pre-training, doc-level ensembled model, JESC parallel corpus
3DEEPNLPBSDja-en2020/09/19 15:00:044048---22.83------NMTYesAn ensemble of transformer models trained on several publicly available JA-EN datasets such as JESC, KFTT, MTNT, etc, and then finetuned on filtered back-translated data followed by finetuning on BSD.
4ut-mrtBSDja-en2020/09/19 20:54:354075---21.64------NMTYesTransformer-base trained on document aligned news data, the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with one previous context sentence.
5ut-mrtBSDja-en2020/09/19 20:20:284067---18.94------NMTYesTransformer-base trained on data from WMT 2020 + the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0
6adapt-dcuBSDja-en2020/09/17 23:50:553836---18.70------NMTYesTraining corpus is a mix of OpenSubtitles, JESC and BSD. Marian-NMT toolkit used to train transformer and fine-tuned on the same corpus oversampled with BSD.
7adapt-dcuBSDja-en2020/09/18 20:07:073970---18.59------NMTYesMarian-NMT toolkit used to train transformer and fine-tuned on source-original synthetic corpus, as well as BSD training corpus.
8ut-mrtBSDja-en2020/09/19 20:47:474073---18.57------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus. Average of 4 best models.
9adapt-dcuBSDja-en2020/09/18 19:26:013963---18.53------NMTYesTraining corpus is a mix of OpenSubtitles, JESC and BSD. Marian-NMT toolkit used to train transformer and fine-tuned on source-original synthetic corpus.
10ut-mrtBSDja-en2020/09/19 20:46:384072---18.05------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus and one previous context sentence. Average of 4 best models.
11ut-mrtBSDja-en2020/09/17 13:49:463802---17.58------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0
12adapt-dcuBSDja-en2020/09/18 18:35:313940---17.33------NMTYesTraining corpus is a mix of OpenSubtitles, JESC and BSD. Marian-NMT toolkit used to train transformer model (baseline model)
13goku20BSDja-en2020/09/15 19:43:023748---17.02------NMTNomBART pre-training, sentence-level single model
14ut-mrtBSDja-en2020/09/19 20:25:234069---16.99------NMTYesTransformer-base trained on data from WMT 2020 without any BSD
15ut-mrtBSDja-en2020/09/17 13:55:013804---15.83------NMTYesTransformer-base trained on the full BSD corpus (80k)
16ut-mrtBSDja-en2020/09/18 17:57:023936---14.49------NMTYesTransformer-small trained on the full BSD corpus (80k) with one previous context sentence
17DEEPNLPBSDja-en2020/09/19 14:57:574047---10.91------NMTNoAn ensemble of two transformer models trained on BSD corpus (20k)
18DEEPNLPBSDja-en2020/09/17 16:21:533808--- 9.70------NMTNoTransformer-base model trained on BSD corpus (20k).
19ut-mrtBSDja-en2020/09/19 19:06:004065--- 7.67------NMTNoTransformer-small (4layer) trained the BSD corpus from GitHub (20k). Average of four best models.
20DEEPNLPBSDja-en2020/10/16 18:03:124162--- 7.47------NMTNo
21ut-mrtBSDja-en2020/09/19 18:55:024059--- 7.43------NMTNoTransformer-small (4layer) trained the BSD corpus from GitHub (20k) with one previous context sentence. Average of four best models.
22ut-mrtBSDja-en2020/09/18 17:19:263919--- 6.88------SMTNoSMT Baseline trained the BSD corpus from GitHub (20k)
23DEEPNLPBSDja-en2020/09/17 15:58:223806--- 6.27------NMTNoTransformer-base model trained on BSD corpus (20k).

Notice:

Back to top

RIBES


# Team Task Date/Time DataID RIBES
Method
Other
Resources
System
Description
juman kytea mecab moses-
tokenizer
stanford-
segmenter-
ctb
stanford-
segmenter-
pku
indic-
tokenizer
unuse myseg kmseg
1goku20BSDja-en2020/09/15 19:41:513747---0.755099------NMTYesmBART pre-training, doc-level ensembled model, JESC parallel corpus
2DEEPNLPBSDja-en2020/09/19 15:00:044048---0.752619------NMTYesAn ensemble of transformer models trained on several publicly available JA-EN datasets such as JESC, KFTT, MTNT, etc, and then finetuned on filtered back-translated data followed by finetuning on BSD.
3ut-mrtBSDja-en2020/09/18 19:35:103966---0.746855------NMTYesTransformer-base ensemble of 2 best models trained on large batches of the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 and WMT 2020 corpora; tuned on the full BSD corpus.
4ut-mrtBSDja-en2020/09/19 20:54:354075---0.746032------NMTYesTransformer-base trained on document aligned news data, the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with one previous context sentence.
5adapt-dcuBSDja-en2020/09/18 20:07:073970---0.732256------NMTYesMarian-NMT toolkit used to train transformer and fine-tuned on source-original synthetic corpus, as well as BSD training corpus.
6adapt-dcuBSDja-en2020/09/18 19:26:013963---0.731639------NMTYesTraining corpus is a mix of OpenSubtitles, JESC and BSD. Marian-NMT toolkit used to train transformer and fine-tuned on source-original synthetic corpus.
7adapt-dcuBSDja-en2020/09/17 23:50:553836---0.730460------NMTYesTraining corpus is a mix of OpenSubtitles, JESC and BSD. Marian-NMT toolkit used to train transformer and fine-tuned on the same corpus oversampled with BSD.
8ut-mrtBSDja-en2020/09/19 20:46:384072---0.723202------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus and one previous context sentence. Average of 4 best models.
9ut-mrtBSDja-en2020/09/19 20:47:474073---0.720809------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus. Average of 4 best models.
10adapt-dcuBSDja-en2020/09/18 18:35:313940---0.714268------NMTYesTraining corpus is a mix of OpenSubtitles, JESC and BSD. Marian-NMT toolkit used to train transformer model (baseline model)
11ut-mrtBSDja-en2020/09/17 13:49:463802---0.710701------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0
12ut-mrtBSDja-en2020/09/17 13:55:013804---0.699950------NMTYesTransformer-base trained on the full BSD corpus (80k)
13ut-mrtBSDja-en2020/09/19 20:20:284067---0.698102------NMTYesTransformer-base trained on data from WMT 2020 + the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0
14ut-mrtBSDja-en2020/09/18 17:57:023936---0.697037------NMTYesTransformer-small trained on the full BSD corpus (80k) with one previous context sentence
15goku20BSDja-en2020/09/15 19:43:023748---0.688501------NMTNomBART pre-training, sentence-level single model
16ut-mrtBSDja-en2020/09/19 20:25:234069---0.677070------NMTYesTransformer-base trained on data from WMT 2020 without any BSD
17DEEPNLPBSDja-en2020/09/19 14:57:574047---0.615230------NMTNoAn ensemble of two transformer models trained on BSD corpus (20k)
18DEEPNLPBSDja-en2020/09/17 16:21:533808---0.606886------NMTNoTransformer-base model trained on BSD corpus (20k).
19DEEPNLPBSDja-en2020/10/16 18:03:124162---0.587745------NMTNo
20ut-mrtBSDja-en2020/09/19 19:06:004065---0.580769------NMTNoTransformer-small (4layer) trained the BSD corpus from GitHub (20k). Average of four best models.
21ut-mrtBSDja-en2020/09/19 18:55:024059---0.576275------NMTNoTransformer-small (4layer) trained the BSD corpus from GitHub (20k) with one previous context sentence. Average of four best models.
22ut-mrtBSDja-en2020/09/18 17:19:263919---0.575330------SMTNoSMT Baseline trained the BSD corpus from GitHub (20k)
23DEEPNLPBSDja-en2020/09/17 15:58:223806---0.531230------NMTNoTransformer-base model trained on BSD corpus (20k).

Notice:

Back to top

AMFM


# Team Task Date/Time DataID AMFM
Method
Other
Resources
System
Description
unuse unuse unuse unuse unuse unuse unuse unuse unuse unuse
1goku20BSDja-en2020/09/15 19:41:513747---0.000000------NMTYesmBART pre-training, doc-level ensembled model, JESC parallel corpus
2goku20BSDja-en2020/09/15 19:43:023748---0.000000------NMTNomBART pre-training, sentence-level single model
3ut-mrtBSDja-en2020/09/17 13:49:463802---0.000000------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0
4ut-mrtBSDja-en2020/09/17 13:55:013804---0.000000------NMTYesTransformer-base trained on the full BSD corpus (80k)
5DEEPNLPBSDja-en2020/09/17 15:58:223806---0.000000------NMTNoTransformer-base model trained on BSD corpus (20k).
6DEEPNLPBSDja-en2020/09/17 16:21:533808---0.000000------NMTNoTransformer-base model trained on BSD corpus (20k).
7adapt-dcuBSDja-en2020/09/17 23:50:553836---0.000000------NMTYesTraining corpus is a mix of OpenSubtitles, JESC and BSD. Marian-NMT toolkit used to train transformer and fine-tuned on the same corpus oversampled with BSD.
8ut-mrtBSDja-en2020/09/18 17:19:263919---0.000000------SMTNoSMT Baseline trained the BSD corpus from GitHub (20k)
9ut-mrtBSDja-en2020/09/18 17:57:023936---0.000000------NMTYesTransformer-small trained on the full BSD corpus (80k) with one previous context sentence
10adapt-dcuBSDja-en2020/09/18 18:35:313940---0.000000------NMTYesTraining corpus is a mix of OpenSubtitles, JESC and BSD. Marian-NMT toolkit used to train transformer model (baseline model)
11adapt-dcuBSDja-en2020/09/18 19:26:013963---0.000000------NMTYesTraining corpus is a mix of OpenSubtitles, JESC and BSD. Marian-NMT toolkit used to train transformer and fine-tuned on source-original synthetic corpus.
12ut-mrtBSDja-en2020/09/18 19:35:103966---0.000000------NMTYesTransformer-base ensemble of 2 best models trained on large batches of the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 and WMT 2020 corpora; tuned on the full BSD corpus.
13adapt-dcuBSDja-en2020/09/18 20:07:073970---0.000000------NMTYesMarian-NMT toolkit used to train transformer and fine-tuned on source-original synthetic corpus, as well as BSD training corpus.
14DEEPNLPBSDja-en2020/09/19 14:57:574047---0.000000------NMTNoAn ensemble of two transformer models trained on BSD corpus (20k)
15DEEPNLPBSDja-en2020/09/19 15:00:044048---0.000000------NMTYesAn ensemble of transformer models trained on several publicly available JA-EN datasets such as JESC, KFTT, MTNT, etc, and then finetuned on filtered back-translated data followed by finetuning on BSD.
16ut-mrtBSDja-en2020/09/19 18:55:024059---0.000000------NMTNoTransformer-small (4layer) trained the BSD corpus from GitHub (20k) with one previous context sentence. Average of four best models.
17ut-mrtBSDja-en2020/09/19 19:06:004065---0.000000------NMTNoTransformer-small (4layer) trained the BSD corpus from GitHub (20k). Average of four best models.
18ut-mrtBSDja-en2020/09/19 20:20:284067---0.000000------NMTYesTransformer-base trained on data from WMT 2020 + the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0
19ut-mrtBSDja-en2020/09/19 20:25:234069---0.000000------NMTYesTransformer-base trained on data from WMT 2020 without any BSD
20ut-mrtBSDja-en2020/09/19 20:46:384072---0.000000------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus and one previous context sentence. Average of 4 best models.
21ut-mrtBSDja-en2020/09/19 20:47:474073---0.000000------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus. Average of 4 best models.
22ut-mrtBSDja-en2020/09/19 20:54:354075---0.000000------NMTYesTransformer-base trained on document aligned news data, the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with one previous context sentence.
23DEEPNLPBSDja-en2020/10/16 18:03:124162---0.000000------NMTNo

Notice:

Back to top

HUMAN (WAT2022)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2021)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2020)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description
1goku20BSDja-en2020/09/15 19:41:5137474.190NMTYesmBART pre-training, doc-level ensembled model, JESC parallel corpus
2DEEPNLPBSDja-en2020/09/19 15:00:0440484.100NMTYesAn ensemble of transformer models trained on several publicly available JA-EN datasets such as JESC, KFTT, MTNT, etc, and then finetuned on filtered back-translated data followed by finetuning on BSD.
3adapt-dcuBSDja-en2020/09/17 23:50:5538363.930NMTYesTraining corpus is a mix of OpenSubtitles, JESC and BSD. Marian-NMT toolkit used to train transformer and fine-tuned on the same corpus oversampled with BSD.
4ut-mrtBSDja-en2020/09/19 20:47:4740733.620NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus. Average of 4 best models.
5goku20BSDja-en2020/09/15 19:43:0237483.570NMTNomBART pre-training, sentence-level single model
6ut-mrtBSDja-en2020/09/19 20:46:3840723.550NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus and one previous context sentence. Average of 4 best models.
7DEEPNLPBSDja-en2020/09/19 14:57:5740472.400NMTNoAn ensemble of two transformer models trained on BSD corpus (20k)

Notice:
Back to top

HUMAN (WAT2019)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2018)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2017)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2016)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2015)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2014)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

EVALUATION RESULTS USAGE POLICY

When you use the WAT evaluation results for any purpose such as:
- writing technical papers,
- making presentations about your system,
- advertising your MT system to the customers,
you can use the information about translation directions, scores (including both automatic and human evaluations) and ranks of your system among others. You can also use the scores of the other systems, but you MUST anonymize the other system's names. In addition, you can show the links (URLs) to the WAT evaluation result pages.

NICT (National Institute of Information and Communications Technology)
Kyoto University
Last Modified: 2018-08-02