The Workshop on Asian Translation (WAT) is a new open evaluation campaign focusing on Asian languages. We would like to invite a broad range of participants and conduct various forms of machine translation experiments and evaluation. Collecting and sharing our knowledge will allow us to understand the essence of machine translation and the problems to be solved. We are working toward the practical use of machine translation among all Asian countries.
For the 1st WAT, we chose scientific papers as the targeted domain, and selected the languages Japanese, Chinese and English.
What makes WAT unique:
9:30 | Welcome |
9:30 - 10:10 | Invited talk: Hamman Riza |
10:10 - 10:25 | Overview of WAT2014: Toshiaki Nakazawa, Hideya Mino, Isao Goto, Sadao Kurohashi, Eiichiro Sumita |
10:25 - 10:40 | Weblio Pre-reordering Statistical Machine Translation System: Zhongyuan Zhu |
10:40 - 10:55 | Forest-to-String SMT for Asian Language Translation: NAIST at WAT 2014: Graham Neubig |
10:55 - 11:00 | Break |
11:00 - 11:10 | Poster booster session (10 systems) |
11:10 - 11:50 | Poster presentation (all 12 systems) |
11:50 - 12:00 | Closing |
12:00 - 12:05 | Commemorative photo |
Dr. Ir. Hammam Riza, Director, Agency for the Assessment and Application of Technology (BPPT) [Short bio.] Title: Leveraging ASEAN economic communities 2015 through Language Translation [slide] (Click here to more information. ) Time: 9:30am-10:10am |
Crowdsourcing evaluation due | August 31, 2014 | |
System description draft paper due | September 14, 2014 | |
Review feedback | September 21, 2014 | |
System description camera-ready paper due | September 28, 2014 | |
WAT 2014 | 9:30am-12:00pm, October 4, 2014 |
Subtasks:
Dataset:
WAT uses ASPEC for the dataset including training, development, development test and test data.
Participants must get a copy of ASPEC by themselves.
ASPEC consists of approximately 3 million Japanese-English parallel sentences from paper abstracts (ASPEC-JE) and approximately 0.7 million Japanese-Chinese paper excerpts (ASPEC-JC).
Baseline system:
Baseline systems site is now open.
We will evaluate the translation performance of the results submitted through automatic evaluation and human evaluation.
Automatic evaluation:
We will prepare an automatic evaluation server.
You will be able to evaluate the translation results at any time using this server.
Human evaluation:
Human evaluation will be carried out using crowdsourcing.
Organizers will sample 400 sentences from the test data for human evaluation.
Participants can submit translation results for the human evaluation a maximum of twice until the crowdsourcing evaluation is due.
Submission:
Submission site is now open.
(User Name and Password is necessary to access.)
Evaluation results:
Evaluation results site is now open.
Participants who submit results for human evaluation should submit description papers of their translation systems and evaluation results. All submissions and feedback are handled electronically as below.
For questions, comments, etc. please email to "wat-at-nlp.ist.i.kyoto-u.ac.jp".
The Association for Natural Language Processing (Japanese Page)
Asian Scientific Paper Excerpt Corpus (ASPEC)
Japan Science and Technology Agency (JST)
National Institute of Information and Communications Technology (NICT)
2014-10-12: INVITED TALK slide
2014-10-12: PAPERS slides
2014-10-06: PHOTOS
2014-10-01: PAPERS
2014-09-25: TIMETABLE
2014-07-17: IMPORTANT DATES (Extended)
2014-07-14: baseline system site, submission site and evaluation results site open
2014-07-04: registration site open
2014-03-05: site open
JST (Japan Science and Technology Agency)
NICT (National Institute of Information and Communications Technology)
Kyoto University
Last Modified: 2015-02-25