NICT_LOGO.JPG KYOTO-U_LOGO.JPG

WAT

The Workshop on Asian Translation
Evaluation Results

[EVALUATION RESULTS TOP] | [BLEU] | [RIBES] | [AMFM] | [HUMAN (WAT2022)] | [HUMAN (WAT2021)] | [HUMAN (WAT2020)] | [HUMAN (WAT2019)] | [HUMAN (WAT2018)] | [HUMAN (WAT2017)] | [HUMAN (WAT2016)] | [HUMAN (WAT2015)] | [HUMAN (WAT2014)] | [EVALUATION RESULTS USAGE POLICY]

BLEU


# Team Task Date/Time DataID BLEU
Method
Other
Resources
System
Description
juman kytea mecab moses-
tokenizer
stanford-
segmenter-
ctb
stanford-
segmenter-
pku
indic-
tokenizer
unuse myseg kmseg
1SRPOLINDIC21en-te2021/05/04 15:23:246241------16.85---NMTNoEnsemble of one-to-many on all data. Pretrained on BT, finetuned on PMI
2SRPOLINDIC21en-te2021/05/04 16:28:566267------16.82---NMTNoOne-to-many on all data. Pretrained on BT, finetuned on PMI
3IIIT-HINDIC21en-te2021/05/03 18:11:376014------15.61---NMTNoMNMT system (En-XX) trained via exploiting lexical similarity on PMI+CVIT parallel corpus, then improved using back translation on PMI monolingual data followed by fine tuning.
4CFILTINDIC21en-te2021/05/04 01:07:216051------15.52---NMTNoMultilingual(One-to-Many(En-XX)) NMT model based on Transformer with shared encoder and decoder.
5sakuraINDIC21en-te2021/05/04 04:19:166160------15.48---NMTNoPre-training multilingual mBART one2many model with training corpus followed by finetuning on PMI Parallel.
6coastalINDIC21en-te2021/05/04 01:40:506088------12.86---NMTNoseq2seq model trained on all WAT2021 data
7sakuraINDIC21en-te2021/05/01 11:39:525891------11.86---NMTNoFine-tuning of multilingual mBART one2many model with training corpus.
8mcairtINDIC21en-te2021/05/03 17:25:185997------11.17---NMTNomultilingual model(one to many model) trained on all WAT 2021 data by using base transformer.
9SRPOLINDIC21en-te2021/04/21 19:34:105334------10.65---NMTNoBase transformer on all WAT21 data
10IITP-MTINDIC21en-te2021/05/04 18:18:336305------ 6.25---NMTNoOne-to-Many model trained on all training data with base Transformer. All indic language data is romanized. Model fine-tuned on BT PMI monolingual corpus.
11NICT-5INDIC21en-te2021/06/25 11:39:476492------ 5.57---NMTNoUsing PMI and PIB data for fine-tuning on a mbart model trained for over 5 epochs. MNMT model.
12NLPHutINDIC21en-te2021/05/03 00:19:165986------ 4.88---NMTNoTransformer with source and target language tags trained using all languages PMI data. Then fine tuned using all en-te data.
13NICT-5INDIC21en-te2021/04/21 15:46:175291------ 4.59---NMTNoPretrain MBART on IndicCorp and FT on bilingual PMI data. Beam search. Model is bilingual.
14NICT-5INDIC21en-te2021/04/22 11:54:225366------ 4.20---NMTNoMBART+MNMT. Beam 4.
15NLPHutINDIC21en-te2021/03/20 00:20:234618------ 3.42---NMTNoTransformer with target language tag trained using all languages PMI data. Then fine tuned using en-te PMI data.
16ORGANIZERINDIC21en-te2021/04/08 17:26:144806------ 2.80---NMTNoBilingual baseline trained on PMI data. Transformer base. LR=10-3
17gaurvarINDIC21en-te2021/04/25 20:03:275587------ 2.31---NMTNoMulti Task Multi Lingual T5 trained for Multiple Indic Languages
18gaurvarINDIC21en-te2021/05/01 19:34:575935------ 2.31---NMTNoMulti Task Multi Lingual T5 trained for Multiple Indic Languages

Notice:

Back to top

RIBES


# Team Task Date/Time DataID RIBES
Method
Other
Resources
System
Description
juman kytea mecab moses-
tokenizer
stanford-
segmenter-
ctb
stanford-
segmenter-
pku
indic-
tokenizer
unuse myseg kmseg
1SRPOLINDIC21en-te2021/05/04 15:23:246241------0.739835---NMTNoEnsemble of one-to-many on all data. Pretrained on BT, finetuned on PMI
2SRPOLINDIC21en-te2021/05/04 16:28:566267------0.734483---NMTNoOne-to-many on all data. Pretrained on BT, finetuned on PMI
3IIIT-HINDIC21en-te2021/05/03 18:11:376014------0.728432---NMTNoMNMT system (En-XX) trained via exploiting lexical similarity on PMI+CVIT parallel corpus, then improved using back translation on PMI monolingual data followed by fine tuning.
4sakuraINDIC21en-te2021/05/04 04:19:166160------0.725543---NMTNoPre-training multilingual mBART one2many model with training corpus followed by finetuning on PMI Parallel.
5CFILTINDIC21en-te2021/05/04 01:07:216051------0.725496---NMTNoMultilingual(One-to-Many(En-XX)) NMT model based on Transformer with shared encoder and decoder.
6coastalINDIC21en-te2021/05/04 01:40:506088------0.707817---NMTNoseq2seq model trained on all WAT2021 data
7sakuraINDIC21en-te2021/05/01 11:39:525891------0.703612---NMTNoFine-tuning of multilingual mBART one2many model with training corpus.
8mcairtINDIC21en-te2021/05/03 17:25:185997------0.702337---NMTNomultilingual model(one to many model) trained on all WAT 2021 data by using base transformer.
9SRPOLINDIC21en-te2021/04/21 19:34:105334------0.692362---NMTNoBase transformer on all WAT21 data
10NICT-5INDIC21en-te2021/06/25 11:39:476492------0.612627---NMTNoUsing PMI and PIB data for fine-tuning on a mbart model trained for over 5 epochs. MNMT model.
11NICT-5INDIC21en-te2021/04/22 11:54:225366------0.576863---NMTNoMBART+MNMT. Beam 4.
12NLPHutINDIC21en-te2021/05/03 00:19:165986------0.570112---NMTNoTransformer with source and target language tags trained using all languages PMI data. Then fine tuned using all en-te data.
13NICT-5INDIC21en-te2021/04/21 15:46:175291------0.569735---NMTNoPretrain MBART on IndicCorp and FT on bilingual PMI data. Beam search. Model is bilingual.
14NLPHutINDIC21en-te2021/03/20 00:20:234618------0.537365---NMTNoTransformer with target language tag trained using all languages PMI data. Then fine tuned using en-te PMI data.
15IITP-MTINDIC21en-te2021/05/04 18:18:336305------0.530898---NMTNoOne-to-Many model trained on all training data with base Transformer. All indic language data is romanized. Model fine-tuned on BT PMI monolingual corpus.
16ORGANIZERINDIC21en-te2021/04/08 17:26:144806------0.479896---NMTNoBilingual baseline trained on PMI data. Transformer base. LR=10-3
17gaurvarINDIC21en-te2021/04/25 20:03:275587------0.414016---NMTNoMulti Task Multi Lingual T5 trained for Multiple Indic Languages
18gaurvarINDIC21en-te2021/05/01 19:34:575935------0.389727---NMTNoMulti Task Multi Lingual T5 trained for Multiple Indic Languages

Notice:

Back to top

AMFM


# Team Task Date/Time DataID AMFM
Method
Other
Resources
System
Description
unuse unuse unuse unuse unuse unuse unuse unuse unuse unuse
1SRPOLINDIC21en-te2021/05/04 16:28:566267------0.792970---NMTNoOne-to-many on all data. Pretrained on BT, finetuned on PMI
2SRPOLINDIC21en-te2021/05/04 15:23:246241------0.791085---NMTNoEnsemble of one-to-many on all data. Pretrained on BT, finetuned on PMI
3CFILTINDIC21en-te2021/05/04 01:07:216051------0.789820---NMTNoMultilingual(One-to-Many(En-XX)) NMT model based on Transformer with shared encoder and decoder.
4sakuraINDIC21en-te2021/05/04 04:19:166160------0.785055---NMTNoPre-training multilingual mBART one2many model with training corpus followed by finetuning on PMI Parallel.
5mcairtINDIC21en-te2021/05/03 17:25:185997------0.783647---NMTNomultilingual model(one to many model) trained on all WAT 2021 data by using base transformer.
6IIIT-HINDIC21en-te2021/05/03 18:11:376014------0.780218---NMTNoMNMT system (En-XX) trained via exploiting lexical similarity on PMI+CVIT parallel corpus, then improved using back translation on PMI monolingual data followed by fine tuning.
7coastalINDIC21en-te2021/05/04 01:40:506088------0.778251---NMTNoseq2seq model trained on all WAT2021 data
8sakuraINDIC21en-te2021/05/01 11:39:525891------0.772064---NMTNoFine-tuning of multilingual mBART one2many model with training corpus.
9IITP-MTINDIC21en-te2021/05/04 18:18:336305------0.764977---NMTNoOne-to-Many model trained on all training data with base Transformer. All indic language data is romanized. Model fine-tuned on BT PMI monolingual corpus.
10SRPOLINDIC21en-te2021/04/21 19:34:105334------0.763271---NMTNoBase transformer on all WAT21 data
11NICT-5INDIC21en-te2021/04/21 15:46:175291------0.754015---NMTNoPretrain MBART on IndicCorp and FT on bilingual PMI data. Beam search. Model is bilingual.
12NICT-5INDIC21en-te2021/04/22 11:54:225366------0.752068---NMTNoMBART+MNMT. Beam 4.
13NLPHutINDIC21en-te2021/03/20 00:20:234618------0.749881---NMTNoTransformer with target language tag trained using all languages PMI data. Then fine tuned using en-te PMI data.
14NLPHutINDIC21en-te2021/05/03 00:19:165986------0.713960---NMTNoTransformer with source and target language tags trained using all languages PMI data. Then fine tuned using all en-te data.
15ORGANIZERINDIC21en-te2021/04/08 17:26:144806------0.708086---NMTNoBilingual baseline trained on PMI data. Transformer base. LR=10-3
16gaurvarINDIC21en-te2021/05/01 19:34:575935------0.642502---NMTNoMulti Task Multi Lingual T5 trained for Multiple Indic Languages
17gaurvarINDIC21en-te2021/04/25 20:03:275587------0.634376---NMTNoMulti Task Multi Lingual T5 trained for Multiple Indic Languages
18NICT-5INDIC21en-te2021/06/25 11:39:476492------0.000000---NMTNoUsing PMI and PIB data for fine-tuning on a mbart model trained for over 5 epochs. MNMT model.

Notice:

Back to top

HUMAN (WAT2022)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2021)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description
1NICT-5INDIC21en-te2021/04/21 15:46:175291UnderwayNMTNoPretrain MBART on IndicCorp and FT on bilingual PMI data. Beam search. Model is bilingual.
2NICT-5INDIC21en-te2021/04/22 11:54:225366UnderwayNMTNoMBART+MNMT. Beam 4.
3gaurvarINDIC21en-te2021/04/25 20:03:275587UnderwayNMTNoMulti Task Multi Lingual T5 trained for Multiple Indic Languages
4gaurvarINDIC21en-te2021/05/01 19:34:575935UnderwayNMTNoMulti Task Multi Lingual T5 trained for Multiple Indic Languages
5NLPHutINDIC21en-te2021/05/03 00:19:165986UnderwayNMTNoTransformer with source and target language tags trained using all languages PMI data. Then fine tuned using all en-te data.
6mcairtINDIC21en-te2021/05/03 17:25:185997UnderwayNMTNomultilingual model(one to many model) trained on all WAT 2021 data by using base transformer.
7IIIT-HINDIC21en-te2021/05/03 18:11:376014UnderwayNMTNoMNMT system (En-XX) trained via exploiting lexical similarity on PMI+CVIT parallel corpus, then improved using back translation on PMI monolingual data followed by fine tuning.
8CFILTINDIC21en-te2021/05/04 01:07:216051UnderwayNMTNoMultilingual(One-to-Many(En-XX)) NMT model based on Transformer with shared encoder and decoder.
9coastalINDIC21en-te2021/05/04 01:40:506088UnderwayNMTNoseq2seq model trained on all WAT2021 data
10sakuraINDIC21en-te2021/05/04 04:19:166160UnderwayNMTNoPre-training multilingual mBART one2many model with training corpus followed by finetuning on PMI Parallel.
11SRPOLINDIC21en-te2021/05/04 15:23:246241UnderwayNMTNoEnsemble of one-to-many on all data. Pretrained on BT, finetuned on PMI
12SRPOLINDIC21en-te2021/05/04 16:28:566267UnderwayNMTNoOne-to-many on all data. Pretrained on BT, finetuned on PMI
13IITP-MTINDIC21en-te2021/05/04 18:18:336305UnderwayNMTNoOne-to-Many model trained on all training data with base Transformer. All indic language data is romanized. Model fine-tuned on BT PMI monolingual corpus.

Notice:
Back to top

HUMAN (WAT2020)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2019)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2018)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2017)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2016)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2015)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

HUMAN (WAT2014)


# Team Task Date/Time DataID HUMAN
Method
Other
Resources
System
Description

Notice:
Back to top

EVALUATION RESULTS USAGE POLICY

When you use the WAT evaluation results for any purpose such as:
- writing technical papers,
- making presentations about your system,
- advertising your MT system to the customers,
you can use the information about translation directions, scores (including both automatic and human evaluations) and ranks of your system among others. You can also use the scores of the other systems, but you MUST anonymize the other system's names. In addition, you can show the links (URLs) to the WAT evaluation result pages.

NICT (National Institute of Information and Communications Technology)
Kyoto University
Last Modified: 2018-08-02