Informatics and Applications

2019, Volume 13, Issue 3, pp 90-96


  • V. A. Nuriev


The paper describes architecture of a Neural Machine Translation (NMT) system. The subject is brought up since NMT, i. e., translation using artificial neural networks, is now a leading Machine Translation paradigm.
The NMT systems manage to deliver much better quality of output than the machine translators of the previous generation (statistical translation systems) do. Yet, the translation they produce still may contain various errors and it is relatively inaccurate compared with human translations. Therefore, to improve its quality, it is important to see more clearly how an NMT system is built and works. Commonly its architecture consists of two recurrent neural networks, one to get the input text sequence and the other to generate translated output (text sequence). The NMT system often has an attention mechanism helping it cope with long input sequences. As an example, Google's NMT system is taken as the Google Translate service is one of the most highly demanded today, it processes around 143 billion words in more than 100 languages per day. The paper concludes with some perspectives for future research.

[+] References (25)

[+] About this article