Reproducing results in publications by distributing publicly available source code is becoming ever more popular. Given the difficulty of reproducing machine learning (ML) experiments, there have been significant efforts in reducing the variance of these results. As in any science, the ability to consistently reproduce results effectively strengthens the underlying hypothesis of the work, and thus, should be regarded as important as the novel aspect of the research itself. The contribution of this work is a framework that is able to reproduce consistent results and provides a means of easily creating, training, and evaluating natural language processing (NLP) deep learning (DL) models.
翻译:由于复制机器学习(ML)试验的困难,在减少这些结果的差异方面已作出了重大努力,同任何科学一样,不断复制结果的能力有效地加强了工作的基本假设,因此,应当像研究本身的新颖方面一样,被视为重要,这项工作的贡献是一个能够复制一致的结果和提供易于创造、培训和评价自然语言处理(NLP)深层学习模式的手段的框架。