Biomedical entity linking is the task of linking entity mentions in a biomedical document to referent entities in a knowledge base. Recently, many BERT-based models have been introduced for the task. While these models have achieved competitive results on many datasets, they are computationally expensive and contain about 110M parameters. Little is known about the factors contributing to their impressive performance and whether the over-parameterization is needed. In this work, we shed some light on the inner working mechanisms of these large BERT-based models. Through a set of probing experiments, we have found that the entity linking performance only changes slightly when the input word order is shuffled or when the attention scope is limited to a fixed window size. From these observations, we propose an efficient convolutional neural network with residual connections for biomedical entity linking. Because of the sparse connectivity and weight sharing properties, our model has a small number of parameters and is highly efficient. On five public datasets, our model achieves comparable or even better linking accuracy than the state-of-the-art BERT-based models while having about 60 times fewer parameters.
翻译:生物医学实体的连接是将生物医学文件中提及的实体与知识库中的参考实体联系起来的任务。最近,许多基于生物伦理学的模型已经为这项任务采用了许多基于生物医学的模型。这些模型在许多数据集上取得了竞争性结果,但计算成本很高,包含大约110M参数。对于造成其令人印象深刻的性能的因素以及是否需要超度参数化的情况知之甚少。在这项工作中,我们对这些基于生物医学的大型模型的内部工作机制做了一些说明。通过一系列测试实验,我们发现,只有在输入单词令被打乱或关注范围限于固定窗口大小时,将业绩挂钩的实体才略有变化。我们从这些观察中建议建立一个高效的革命神经网络,与生物医学实体连接的剩余连接连接连接。由于连接和重量共享特性稀少,我们的模型具有少量参数,而且效率很高。在五个公共数据集中,我们的模型比基于生物伦理学的模型的精确度要高出60倍左右。