NICT_LOGO.JPG KYOTO-U_LOGO.JPG

WAT

The Workshop on Asian Translation
Evaluation Results

[EVALUATION RESULTS TOP] | [BLEU] | [RIBES] | [AMFM] | [HUMAN (WAT2022)] | [HUMAN (WAT2021)] | [HUMAN (WAT2020)] | [HUMAN (WAT2019)] | [HUMAN (WAT2018)] | [HUMAN (WAT2017)] | [HUMAN (WAT2016)] | [HUMAN (WAT2015)] | [HUMAN (WAT2014)] | [EVALUATION RESULTS USAGE POLICY]

BLEU


# Team Task Date/Time DataID BLEU
Method
Other
Resources
System
Description
juman kytea mecab moses-
tokenizer
stanford-
segmenter-
ctb
stanford-
segmenter-
pku
indic-
tokenizer
unuse myseg kmseg
1ut-mrtBSDen-ja2020/09/18 18:49:05394219.5025.5020.54-------NMTYesTransformer-base ensemble of 2 best models trained on large batches of the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0, WMT 2020, and news corpora; tuned on the full BSD corpus.
2goku20BSDen-ja2020/09/15 20:33:23375619.4326.0420.75-------NMTYesmBART pre-training, doc-level ensembled model, JESC parallel corpus
3DEEPNLPBSDen-ja2020/09/19 15:03:40405019.3926.5920.95-------NMTYesAn ensemble of transformer models trained on several publicly available JA-EN datasets such as JESC, KFTT, MTNT, etc, and then finetuned on filtered back-translated data further finetuning on BSD.
4ut-mrtBSDen-ja2020/09/17 12:04:19379318.8524.8420.04-------NMTYesTransformer-base trained on large batches of the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0, and jParaCrawl; tuned on the full BSD corpus.
5DEEPNLPBSDen-ja2020/09/19 03:23:53402718.7625.8620.19-------NMTYesTraining corpus is a mix of several publicly available JA-EN datasets such as JESC,KFTT,MTNT, etc. Model - Base Transformer trained and then finetuned on BSD.
6ut-mrtBSDen-ja2020/09/19 20:55:04407617.2523.5518.60-------NMTYesTransformer-base trained on document aligned news data, the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with one previous context sentence.
7ut-mrtBSDen-ja2020/09/19 20:21:04406814.8720.5816.08-------NMTYesTransformer-base trained on data from WMT 2020 without any BSD
8ut-mrtBSDen-ja2020/09/19 20:44:13407114.5921.0916.03-------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus and one previous context sentence. Average of 4 best models.
9goku20BSDen-ja2020/09/15 20:27:00375314.4920.8915.75-------NMTNomBART pre-training, doc-level single model
10ut-mrtBSDen-ja2020/09/19 20:49:11407414.1920.3215.49-------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus. Average of 4 best models.
11ut-mrtBSDen-ja2020/09/17 13:50:02380313.7820.0215.26-------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0
12ut-mrtBSDen-ja2020/09/19 20:25:59407012.5517.6013.47-------NMTYesTransformer-base trained on data from WMT 2020 + the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0
13ut-mrtBSDen-ja2020/09/17 13:55:20380512.2018.3713.63-------NMTYesTransformer-base trained on the full BSD corpus (80k)
14ut-mrtBSDen-ja2020/09/18 17:57:22393711.7517.9413.13-------NMTYesTransformer-small trained on the full BSD corpus (80k) with one previous context sentence
15DEEPNLPBSDen-ja2020/09/19 15:01:284049 8.6615.1010.10-------NMTNoAn ensemble of two transformer models trained on BSD corpus (20k)
16DEEPNLPBSDen-ja2020/09/19 03:22:484026 7.9014.03 9.10-------NMTNoTransformer model trained on BSD corpus (20k)
17ut-mrtBSDen-ja2020/09/18 17:19:523920 7.8313.51 9.21-------SMTNoSMT Baseline trained the BSD corpus from GitHub (20k)
18ut-mrtBSDen-ja2020/09/19 18:55:294060 6.7012.13 7.84-------NMTNoTransformer-small (4layer) trained the BSD corpus from GitHub (20k) with one previous context sentence. Average of four best models.
19ut-mrtBSDen-ja2020/09/17 16:51:273810 5.8410.82 6.97-------NMTNoTransformer-small (4layer) trained the BSD corpus from GitHub (20k).

Notice:

Back to top

RIBES


# Team Task Date/Time DataID RIBES
Method
Other
Resources
System
Description
juman kytea mecab moses-
tokenizer
stanford-
segmenter-
ctb
stanford-
segmenter-
pku
indic-
tokenizer
unuse myseg kmseg
1goku20BSDen-ja2020/09/15 20:33:2337560.7454930.7715880.749985-------NMTYesmBART pre-training, doc-level ensembled model, JESC parallel corpus
2DEEPNLPBSDen-ja2020/09/19 15:03:4040500.7442090.7702170.753233-------NMTYesAn ensemble of transformer models trained on several publicly available JA-EN datasets such as JESC, KFTT, MTNT, etc, and then finetuned on filtered back-translated data further finetuning on BSD.
3DEEPNLPBSDen-ja2020/09/19 03:23:5340270.7389470.7672640.749352-------NMTYesTraining corpus is a mix of several publicly available JA-EN datasets such as JESC,KFTT,MTNT, etc. Model - Base Transformer trained and then finetuned on BSD.
4ut-mrtBSDen-ja2020/09/19 20:55:0440760.7346920.7594650.744708-------NMTYesTransformer-base trained on document aligned news data, the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with one previous context sentence.
5ut-mrtBSDen-ja2020/09/18 18:49:0539420.7246240.7587770.740521-------NMTYesTransformer-base ensemble of 2 best models trained on large batches of the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0, WMT 2020, and news corpora; tuned on the full BSD corpus.
6ut-mrtBSDen-ja2020/09/19 20:44:1340710.7208830.7446520.731101-------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus and one previous context sentence. Average of 4 best models.
7ut-mrtBSDen-ja2020/09/17 12:04:1937930.7203650.7551180.736793-------NMTYesTransformer-base trained on large batches of the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0, and jParaCrawl; tuned on the full BSD corpus.
8goku20BSDen-ja2020/09/15 20:27:0037530.7000580.7394680.715971-------NMTNomBART pre-training, doc-level single model
9ut-mrtBSDen-ja2020/09/19 20:49:1140740.6983270.7373320.719519-------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus. Average of 4 best models.
10ut-mrtBSDen-ja2020/09/19 20:21:0440680.6966330.7258030.707683-------NMTYesTransformer-base trained on data from WMT 2020 without any BSD
11ut-mrtBSDen-ja2020/09/18 17:57:2239370.6919090.7232160.705550-------NMTYesTransformer-small trained on the full BSD corpus (80k) with one previous context sentence
12ut-mrtBSDen-ja2020/09/17 13:55:2038050.6874110.7210430.703018-------NMTYesTransformer-base trained on the full BSD corpus (80k)
13ut-mrtBSDen-ja2020/09/17 13:50:0238030.6820910.7221630.701005-------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0
14ut-mrtBSDen-ja2020/09/19 20:25:5940700.6437930.6851480.658451-------NMTYesTransformer-base trained on data from WMT 2020 + the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0
15DEEPNLPBSDen-ja2020/09/19 15:01:2840490.6406690.6862330.658636-------NMTNoAn ensemble of two transformer models trained on BSD corpus (20k)
16DEEPNLPBSDen-ja2020/09/19 03:22:4840260.6322360.6765370.645595-------NMTNoTransformer model trained on BSD corpus (20k)
17ut-mrtBSDen-ja2020/09/18 17:19:5239200.6258460.6667690.641864-------SMTNoSMT Baseline trained the BSD corpus from GitHub (20k)
18ut-mrtBSDen-ja2020/09/19 18:55:2940600.5970370.6483970.613298-------NMTNoTransformer-small (4layer) trained the BSD corpus from GitHub (20k) with one previous context sentence. Average of four best models.
19ut-mrtBSDen-ja2020/09/17 16:51:2738100.5665630.6342740.596934-------NMTNoTransformer-small (4layer) trained the BSD corpus from GitHub (20k).

Notice:

Back to top

AMFM


# Team Task Date/Time DataID AMFM
Method
Other
Resources
System
Description
unuse unuse unuse unuse unuse unuse unuse unuse unuse unuse
1goku20BSDen-ja2020/09/15 20:27:0037530.0000000.0000000.000000-------NMTNomBART pre-training, doc-level single model
2goku20BSDen-ja2020/09/15 20:33:2337560.0000000.0000000.000000-------NMTYesmBART pre-training, doc-level ensembled model, JESC parallel corpus
3ut-mrtBSDen-ja2020/09/17 12:04:1937930.0000000.0000000.000000-------NMTYesTransformer-base trained on large batches of the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0, and jParaCrawl; tuned on the full BSD corpus.
4ut-mrtBSDen-ja2020/09/17 13:50:0238030.0000000.0000000.000000-------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0
5ut-mrtBSDen-ja2020/09/17 13:55:2038050.0000000.0000000.000000-------NMTYesTransformer-base trained on the full BSD corpus (80k)
6ut-mrtBSDen-ja2020/09/17 16:51:2738100.0000000.0000000.000000-------NMTNoTransformer-small (4layer) trained the BSD corpus from GitHub (20k).
7ut-mrtBSDen-ja2020/09/18 17:19:5239200.0000000.0000000.000000-------SMTNoSMT Baseline trained the BSD corpus from GitHub (20k)
8ut-mrtBSDen-ja2020/09/18 17:57:2239370.0000000.0000000.000000-------NMTYesTransformer-small trained on the full BSD corpus (80k) with one previous context sentence
9ut-mrtBSDen-ja2020/09/18 18:49:0539420.0000000.0000000.000000-------NMTYesTransformer-base ensemble of 2 best models trained on large batches of the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0, WMT 2020, and news corpora; tuned on the full BSD corpus.
10DEEPNLPBSDen-ja2020/09/19 03:22:4840260.0000000.0000000.000000-------NMTNoTransformer model trained on BSD corpus (20k)
11DEEPNLPBSDen-ja2020/09/19 03:23:5340270.0000000.0000000.000000-------NMTYesTraining corpus is a mix of several publicly available JA-EN datasets such as JESC,KFTT,MTNT, etc. Model - Base Transformer trained and then finetuned on BSD.
12DEEPNLPBSDen-ja2020/09/19 15:01:2840490.0000000.0000000.000000-------NMTNoAn ensemble of two transformer models trained on BSD corpus (20k)
13DEEPNLPBSDen-ja2020/09/19 15:03:4040500.0000000.0000000.000000-------NMTYesAn ensemble of transformer models trained on several publicly available JA-EN datasets such as JESC, KFTT, MTNT, etc, and then finetuned on filtered back-translated data further finetuning on BSD.
14ut-mrtBSDen-ja2020/09/19 18:55:2940600.0000000.0000000.000000-------NMTNoTransformer-small (4layer) trained the BSD corpus from GitHub (20k) with one previous context sentence. Average of four best models.
15ut-mrtBSDen-ja2020/09/19 20:21:0440680.0000000.0000000.000000-------NMTYesTransformer-base trained on data from WMT 2020 without any BSD
16ut-mrtBSDen-ja2020/09/19 20:25:5940700.0000000.0000000.000000-------NMTYesTransformer-base trained on data from WMT 2020 + the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0
17ut-mrtBSDen-ja2020/09/19 20:44:1340710.0000000.0000000.000000-------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus and one previous context sentence. Average of 4 best models.
18ut-mrtBSDen-ja2020/09/19 20:49:1140740.0000000.0000000.000000-------NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus. Average of 4 best models.
19ut-mrtBSDen-ja2020/09/19 20:55:0440760.0000000.0000000.000000-------NMTYesTransformer-base trained on document aligned news data, the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with one previous context sentence.

Notice:

Back to top

HUMAN (WAT2022)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2021)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2020)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description
1goku20BSDen-ja2020/09/15 20:33:2337564.200NMTYesmBART pre-training, doc-level ensembled model, JESC parallel corpus
2DEEPNLPBSDen-ja2020/09/19 15:03:4040504.130NMTYesAn ensemble of transformer models trained on several publicly available JA-EN datasets such as JESC, KFTT, MTNT, etc, and then finetuned on filtered back-translated data further finetuning on BSD.
3ut-mrtBSDen-ja2020/09/19 20:49:1140743.560NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus. Average of 4 best models.
4goku20BSDen-ja2020/09/15 20:27:0037533.550NMTNomBART pre-training, doc-level single model
5ut-mrtBSDen-ja2020/09/19 20:44:1340713.520NMTYesTransformer-base trained on the full BSD corpus (80k), AMI meeting corpus, Ontonotes 5.0 with domain tags to separate each corpus and one previous context sentence. Average of 4 best models.
6DEEPNLPBSDen-ja2020/09/19 15:01:2840492.600NMTNoAn ensemble of two transformer models trained on BSD corpus (20k)

Notice:
Back to top

HUMAN (WAT2019)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2018)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2017)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2016)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2015)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2014)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

EVALUATION RESULTS USAGE POLICY

When you use the WAT evaluation results for any purpose such as:
- writing technical papers,
- making presentations about your system,
- advertising your MT system to the customers,
you can use the information about translation directions, scores (including both automatic and human evaluations) and ranks of your system among others. You can also use the scores of the other systems, but you MUST anonymize the other system's names. In addition, you can show the links (URLs) to the WAT evaluation result pages.

NICT (National Institute of Information and Communications Technology)
Kyoto University
Last Modified: 2018-08-02