It's hard for neural MWP solvers to deal with tiny local variances. In MWP task, some local changes conserve the original semantic while the others may totally change the underlying logic. Currently, existing datasets for MWP task contain limited samples which are key for neural models to learn to disambiguate different kinds of local variances in questions and solve the questions correctly. In this paper, we propose a set of novel data augmentation approaches to supplement existing datasets with such data that are augmented with different kinds of local variances, and help to improve the generalization ability of current neural models. New samples are generated by knowledge guided entity replacement, and logic guided problem reorganization. The augmentation approaches are ensured to keep the consistency between the new data and their labels. Experimental results have shown the necessity and the effectiveness of our methods.
翻译:神经 MWP 解析器很难解决本地差异很小的问题。 在 MWP 任务中, 一些本地的改变保存了原始语义, 而其他的则可能完全改变基本逻辑。 目前, MWP 任务的现有数据集包含有限的样本, 这些样本是神经模型学会分辨不同类型本地差异并正确解决问题的关键。 在本文中, 我们提出一套新的数据增强方法, 以补充现有数据集, 这些数据增加不同的本地差异, 并帮助提高当前神经模型的普遍化能力。 新的样本是由知识引导的实体替换和逻辑引导的问题重组生成的。 增强方法确保了新数据及其标签的一致性。 实验结果显示了我们方法的必要性和有效性 。