# |
Team |
Task |
Date/Time |
DataID |
AMFM |
Method
|
Other Resources
|
System Description |
unuse |
unuse |
unuse |
unuse |
unuse |
unuse |
unuse |
unuse |
unuse |
unuse |
|
1 | ORGANIZER | ALT20en-id | 2020/09/01 16:01:54 | 3612 | - | - | - | 0.809985 | - | - | - | - | - | - | NMT | No | Baseline MLNMT En to XX model using ALT, Ubuntu, GNOME and KDE4 data from opus.
Transformer big model. Default settings. |
2 | NICT-5 | ALT20en-id | 2020/09/18 19:15:39 | 3956 | - | - | - | 0.000000 | - | - | - | - | - | - | NMT | No | XX to XX transformer model trained on ALT as well as KDE, GNOME and Ubuntu data from OPUS. Corpora were size balanced. |
3 | NICT-5 | ALT20en-id | 2020/09/18 21:52:04 | 4013 | - | - | - | 0.000000 | - | - | - | - | - | - | NMT | No | XX to XX transformer model trained on ALT as well as KDE, GNOME and Ubuntu data from OPUS. Corpora were size unbalanced. |
4 | sakura | ALT20en-id | 2021/04/29 12:20:02 | 5798 | - | - | - | 0.868025 | - | - | - | - | - | - | NMT | No | Multilingual finetuning of mBART50 finetuned many-to-many model, ensemble of 3 |
5 | NICT-2 | ALT20en-id | 2021/05/01 13:18:16 | 5910 | - | - | - | 0.821204 | - | - | - | - | - | - | NMT | No | Transformer base model, multilingual + mixed domain training with domain fine-tuning. |
6 | NICT-2 | ALT20en-id | 2021/05/01 13:31:17 | 5918 | - | - | - | 0.867678 | - | - | - | - | - | - | NMT | Yes | The extended mBART model, mixed domain training with domain fine-tuning. |
7 | HwTscSU | ALT20en-id | 2022/07/11 12:06:05 | 6737 | - | - | - | 0.000000 | - | - | - | - | - | - | NMT | No | XX to XX transformer model finetune on the baseline trained on IT domain data |