Lack of repeatability and generalisability are two significant threats to continuing scientific development in Natural Language Processing. Language models and learning methods are so complex that scientific conference papers no longer contain enough space for the technical depth required for replication or reproduction. Taking Target Dependent Sentiment Analysis as a case study, we show how recent work in the field has not consistently released code, or described settings for learning methods in enough detail, and lacks comparability and generalisability in train, test or validation data. To investigate generalisability and to enable state of the art comparative evaluations, we carry out the first reproduction studies of three groups of complementary methods and perform the first large-scale mass evaluation on six different English datasets. Reflecting on our experiences, we recommend that future replication or reproduction experiments should always consider a variety of datasets alongside documenting and releasing their methods and published code in order to minimise the barriers to both repeatability and generalisability. We have released our code with a model zoo on GitHub with Jupyter Notebooks to aid understanding and full documentation, and we recommend that others do the same with their papers at submission time through an anonymised GitHub account.
翻译:语言模式和学习方法非常复杂,科学会议文件不再有足够的空间进行复制或复制所需的技术深度。 以目标依赖感分析为案例研究,我们表明最近实地工作没有始终如一地公布守则,也没有足够详细地描述学习方法的设置,在火车、测试或验证数据方面缺乏可比性和普遍性。 为了调查普遍适用性,并能够进行最新比较评估,我们进行了三组互补方法的首次复制研究,对六种不同的英文数据集进行了第一次大规模大规模评估。我们建议今后的复制或复制实验应始终考虑各种数据集,同时记录和公布其方法和公布编码,以尽量减少重复性和可普及性的障碍。我们发布了在GitHub的模型动物园的代码,配有Jupyter笔记本,以帮助理解和完整文件。我们建议其他人通过匿名的GitHub账户按时提交文件。