Li Rumeng, Wang Xun, Yu Hong
School of Computer Science, University of Massachusetts Amherst, Amherst, MA, United States.
Department of Computer Science, University of Massachusetts Lowell, Lowell, MA, United States.
Proc AAAI Conf Artif Intell. 2020;34(5):8245-8252. doi: 10.1609/aaai.v34i05.6339. Epub 2020 Apr 3.
Neural machine translation (NMT) models have achieved state-of-the-art translation quality with a large quantity of parallel corpora available. However, their performance suffers significantly when it comes to domain-specific translations, in which training data are usually scarce. In this paper, we present a novel NMT model with a new word embedding transition technique for fast domain adaption. We propose to split parameters in the model into two groups: model parameters and meta parameters. The former are used to model the translation while the latter are used to adjust the representational space to generalize the model to different domains. We mimic the domain adaptation of the machine translation model to low-resource domains using multiple translation tasks on different domains. A new training strategy based on meta-learning is developed along with the proposed model to update the model parameters and meta parameters alternately. Experiments on datasets of different domains showed substantial improvements of NMT performances on a limited amount of data.
神经机器翻译(NMT)模型在有大量平行语料库可用的情况下已经实现了最先进的翻译质量。然而,当涉及特定领域的翻译时,它们的性能会显著下降,因为在这些领域中训练数据通常很少。在本文中,我们提出了一种新颖的NMT模型,该模型采用了一种新的词嵌入转换技术来实现快速的领域适应。我们建议将模型中的参数分为两组:模型参数和元参数。前者用于对翻译进行建模,而后者用于调整表示空间,以便将模型推广到不同领域。我们通过在不同领域上进行多个翻译任务来模拟机器翻译模型对低资源领域的适应。随着所提出的模型,还开发了一种基于元学习的新训练策略,以交替更新模型参数和元参数。在不同领域的数据集上进行的实验表明,在有限的数据量上,NMT的性能有了显著提高。