For languages with no annotated resources, transferring knowledge from rich-resource languages is an effective solution for named entity recognition (NER). While all existing methods directly transfer from source-learned model to a target language, in this paper, we propose to fine-tune the learned model with a few similar examples given a test case, which could benefit the prediction by leveraging the structural and semantic information conveyed in such similar examples. To this end, we present a meta-learning algorithm to find a good model parameter initialization that could fast adapt to the given test case and propose to construct multiple pseudo-NER tasks for meta-training by computing sentence similarities. To further improve the model's generalization ability across different languages, we introduce a masking scheme and augment the loss function with an additional maximum term during meta-training. We conduct extensive experiments on cross-lingual named entity recognition with minimal resources over five target languages. The results show that our approach significantly outperforms existing state-of-the-art methods across the board.
翻译:对于没有附加说明资源的语言而言,从丰富资源语言中传授知识是命名实体识别的有效解决办法。 虽然所有现有方法都直接从源学模式向目标语言转移,但在本文件中,我们提议用几个类似实例对学习模式进行微调,并给出一个测试案例,通过利用在类似实例中传递的结构和语义信息,有利于预测。为此,我们提出了一个元学习算法,以找到一个能够快速适应特定测试案例的良好模型参数初始化,并提议通过计算句号相似之处,为元培训构建多种模拟网络任务。为了进一步提高模型在不同语言中的通用能力,我们引入了一种掩码计划,并在元培训期间增加一个最长期限,以扩大损失功能。我们用最起码的资源对五种目标语言的跨语言进行跨语言实体识别的广泛实验。结果显示,我们的方法大大优于目前全局的先进方法。