As the numbers of submissions to conferences grow quickly, the task of assessing the quality of academic papers automatically, convincingly, and with high accuracy attracts increasing attention. We argue that studying interpretable dimensions of these submissions could lead to scalable solutions. We extract a collection of writing features, and construct a suite of prediction tasks to assess the usefulness of these features in predicting citation counts and the publication of AI-related papers. Depending on the venues, the writing features can predict the conference vs. workshop appearance with F1 scores up to 60-90, sometimes even outperforming the content-based tf-idf features and RoBERTa. We show that the features describe writing style more than content. To further understand the results, we estimate the causal impact of the most indicative features. Our analysis on writing features provides a perspective to assessing and refining the writing of academic articles at scale.
翻译:随着向会议提交的文件数量迅速增加,自动、令人信服和高度精确地评估学术论文质量的任务日益引起人们的注意。我们争辩说,研究这些论文可解释的方面可能会导致可伸缩的解决办法。我们抽取了一些写作特征,并设计了一套预测任务,以评估这些特点在预测引注数和出版与AI有关的论文方面的效用。根据会议地点,书面特征可以预测会议与F1讲习班的出现相比达到60-90分,有时甚至超过基于内容的tf-idf特征和RoBERTA。我们显示,这些特征描述的写作风格比内容要多。为了进一步理解结果,我们估计了最指示性特征的因果关系。我们对写作特征的分析为评估和改进规模学术文章的写作提供了一个视角。