The recent surge of language models has rapidly expanded NLP research, driving an exponential rise in submissions and acceptances at major conferences. Yet this growth has been shadowed by escalating concerns over conference quality, e.g., plagiarism, reviewer inexperience and collusive bidding. However, existing studies rely largely on qualitative accounts (e.g., expert interviews, social media discussions, etc.), lacking longitudinal empirical evidence. To fill this gap, we conduct a ten year empirical study spanning seven leading conferences. We build a four dimensional bibliometric framework covering conference scale, core citation statistics,impact dispersion, cross venue and journal influence, etc. Notably, we further propose a metric Quality Quantity Elasticity, which measures the elasticity of citation growth relative to acceptance growth. Our findings show that ML venues sustain dominant and stable impact, NLP venues undergo widening stratification with mixed expansion efficiency, and AI venues exhibit structural decline. This study provides the first decade-long, cross-venue empirical evidence on the evolution of major conferences.
翻译:近期语言模型的迅猛发展极大地推动了自然语言处理(NLP)研究,导致主要会议的投稿量与接收量呈指数级增长。然而,这种增长背后伴随着对会议质量的日益担忧,例如剽窃、评审经验不足及合谋投标等问题。现有研究主要依赖于定性描述(如专家访谈、社交媒体讨论等),缺乏纵向实证证据。为填补这一空白,我们开展了一项跨越七个顶尖会议的十年实证研究。我们构建了一个四维文献计量框架,涵盖会议规模、核心引文统计、影响力分散度、跨会议及期刊影响力等维度。值得注意的是,我们进一步提出了“质量-数量弹性”指标,用于衡量引文增长相对于接收量增长的弹性。研究发现:机器学习(ML)类会议保持主导且稳定的影响力;自然语言处理(NLP)类会议呈现日益扩大的分层现象,扩张效率参差不齐;而人工智能(AI)类会议则显示出结构性衰退。本研究首次提供了关于主要会议演变的、跨会议、长达十年的实证证据。