Speed Up the Training of Neural Machine Translation

作者:Xinyue Liu, Weixuan Wang, Wenxin Liang, Yuangang Li

摘要

Neural machine translation (NMT) has achieved notable achievements in recent years. Although existing models provide reasonable translation performance, they cost too much training time. Especially, when the corpus is enormous, their computational cost will be extremely high. In this paper, we propose a novel NMT model based on the conventional bidirectional recurrent neural network (bi-RNN). In this model, we apply a tanh activation function, which can learn the future and history context information more sufficiently, to speed up the training process. Experimental results on tasks of German–English and English–French translation demonstrate that the proposed model can save much training time compared with the state-of-the-art models and provide better translation performances.

论文关键词:Neural machine translation (NMT), Bidirectional LSTM, Nolinear activation function

论文评审过程:

论文官网地址:https://doi.org/10.1007/s11063-019-10084-y