Abstract

Neural machine translation is a relatively new approach to statistical machine translation based purely on neural networks.The neural machine translation models often consist of an encoder and a decoder.The encoder extracts a fixed-length representation from a variable-length input sentence, and the decoder generates a correct translation from this representation.In this paper, we focus on analyzing the properties of the neural machine translation using two models; RNN Encoder-Decoder and a newly proposed gated recursive convolutional neural network.We show that the neural machine translation performs relatively well on short sentences without unknown words, but its performance degrades rapidly as the length of the sentence and the number of unknown words increase.Furthermore, we find that the proposed gated recursive convolutional network learns a grammatical structure of a sentence automatically.

Keywords

Computer scienceMachine translationEncoderTranslation (biology)Artificial intelligenceSpeech recognitionOperating system

Affiliated Institutions

Related Publications

Skip-Thought Vectors

We describe an approach for unsupervised learning of a generic, distributed sentence encoder. Using the continuity of text from books, we train an encoder-decoder model that tri...

2015 arXiv (Cornell University) 723 citations

Publication Info

Year
2014
Type
preprint
Pages
103-111
Citations
6358
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

6358
OpenAlex
891
Influential
3978
CrossRef

Cite This

Kyunghyun Cho, Bart van Merriënboer, Dzmitry Bahdanau et al. (2014). On the Properties of Neural Machine Translation: Encoder–Decoder Approaches. Proceedings of SSST-8, Eighth Workshop on Syntax, Semantics and Structure in Statistical Translation , 103-111. https://doi.org/10.3115/v1/w14-4012

Identifiers

DOI
10.3115/v1/w14-4012
arXiv
1409.1259

Data Quality

Data completeness: 84%