Skip to content

Latest commit

 

History

History
24 lines (18 loc) · 1.1 KB

README.md

File metadata and controls

24 lines (18 loc) · 1.1 KB

Toolkit for training neural machine translation systems

This is a Tensorflow reimplementation of Nematus. I used this toolkit for conducting my own research while I was pursuing a PhD at Edinburgh University.

This project was implemented with Tensorflow 1.0 - I have not checked whether it works with later versions.

The neural network model is an attention-based encoder-decoder model for neural machine translation as described in:

Rico Sennrich, Orhan Firat, Kyunghyun Cho, Alexandra Birch, Barry Haddow, Julian Hitschler, Marcin Junczys-Dowmunt, Samuel Läubli, Antonio Valerio Miceli Barone, Jozef Mokry and Maria Nadejde (2017): Nematus: a Toolkit for Neural Machine Translation. In Proceedings of the Software Demonstrations of the 15th Conference of the European Chapter of the Association for Computational Linguistics, Valencia, Spain, pp. 65-68.

and

Dzmitry Bahdanau, Kyunghyun Cho, Yoshua Bengio (2015): Neural Machine Translation by Jointly Learning to Align and Translate, Proceedings of the International Conference on Learning Representations (ICLR).