Amrita Saha, Mitesh M. Khapra, et al.
COLING 2016
We propose multi-way, multilingual neural machine translation. The proposed approach enables a single neural translation model to translate between multiple languages, with a number of parameters that grows only linearly with the number of languages. This is made possible by having a single attention mechanism that is shared across all language pairs. We train the proposed multi-way, multilingual model on ten language pairs from WMT′15 simultaneously and observe clear performance improvements over models trained on only one language pair. We empirically evaluate the proposed model on low-resource language translation tasks. In particular, we observe that the proposed multilingual model outperforms strong conventional statistical machine translation systems on Turkish-English and Uzbek-English by incorporating the resources of other language pairs.
Amrita Saha, Mitesh M. Khapra, et al.
COLING 2016
Haitao Mi, Baskaran Sankaran, et al.
EMNLP 2016
Li Yao, Nicolas Ballas, et al.
BMVC 2016
Orhan Firat, Baskaran Sankaran, et al.
EMNLP 2016