With the recent developments in the field of Natural Language Processing, there has been a rise in the use of different architectures for Neural Machine Translation. Transformer architectures are used to achieve state-of-the-art accuracy, but they are very computationally expensive to train. Everyone cannot have such setups consisting of high-end GPUs and other resources. We train our models on low computational resources and investigate the results. As expected, transformers outperformed other architectures, but there were some surprising results. Transformers consisting of more encoders and decoders took more time to train but had fewer BLEU scores. LSTM performed well in the experiment and took comparatively less time to train than transformers, making it suitable to use in situations having time constraints.
翻译:随着最近自然语言处理领域的发展,使用不同结构来进行神经机器翻译的情况有所增加。 变异结构被用于实现最先进的准确性,但是在计算上非常昂贵。 每个人不能拥有由高端GPU和其他资源组成的这种设置。 我们用低计算资源来培训模型并调查结果。 正如所预期的那样,变异器优于其他结构,但有一些出人意料的结果。 由更多的编码器和解码器组成的变异器需要更多的时间来培训,但BLEU的得分却较少。 LSTM在实验中表现良好,比变异器培训的时间要少得多,在有时间限制的情况下适合使用。