自然语言处理顶会COLING2020最佳论文出炉!

2020 年 12 月 12 日 专知

计算语言学国际会议 COLING 2020(The 28th International Conference on Computational Linguistics)是计算语言学和自然语言处理领域的重要国际会议,由ICCL(International Committee on Computational Linguistics)主办,每两年举办一次,是CCF-B类推荐会议。本届COLING 2020将于2020年12月8日至13日以在线会议的形式举办。COLING 2020共计收到2180篇论文投稿,其中包括2021篇主会投稿、48篇Demo投稿、111篇工业论文投稿,最终有1900余篇论文进入审稿流程。官方Twitter公布了最佳论文。



最佳论文


https://www.aclweb.org/anthology/2020.coling-main.398/


Recent studies have revealed a number of pathologies of neural machine translation (NMT) systems. Hypotheses explaining these mostly suggest there is something fundamentally wrong with NMT as a model or its training algorithm, maximum likelihood estimation (MLE). Most of this evidence was gathered using maximum a posteriori (MAP) decoding, a decision rule aimed at identifying the highest-scoring translation, i.e. the mode. We argue that the evidence corroborates the inadequacy of MAP decoding more than casts doubt on the model and its training algorithm. In this work, we show that translation distributions do reproduce various statistics of the data well, but that beam search strays from such statistics. We show that some of the known pathologies and biases of NMT are due to MAP decoding and not to NMT’s statistical assumptions nor MLE. In particular, we show that the most likely translations under the model accumulate so little probability mass that the mode can be considered essentially arbitrary. We therefore advocate for the use of decision rules that take into account the translation distribution holistically. We show that an approximation to minimum Bayes risk decoding gives competitive results confirming that NMT models do capture important aspects of translation well in expectation.


最佳短论文

https://www.aclweb.org/anthology/2020.coling-main.129/


While recurrent models have been effective in NLP tasks, their performance on context-free languages (CFLs) has been found to be quite weak. Given that CFLs are believed to capture important phenomena such as hierarchical structure in natural languages, this discrepancy in performance calls for an explanation. We study the performance of recurrent models on Dyck-n languages, a particularly important and well-studied class of CFLs. We find that while recurrent models generalize nearly perfectly if the lengths of the training and test strings are from the same range, they perform poorly if the test strings are longer. At the same time, we observe that RNNs are expressive enough to recognize Dyck words of arbitrary lengths in finite precision if their depths are bounded. Hence, we evaluate our models on samples generated from Dyck languages with bounded depth and find that they are indeed able to generalize to much higher lengths. Since natural language datasets have nested dependencies of bounded depth, this may help explain why they perform well in modeling hierarchical dependencies in natural language data despite prior works indicating poor generalization performance on Dyck languages. We perform probing studies to support our results and provide comparisons with Transformers.


专知便捷查看

便捷下载,请关注专知公众号(点击上方蓝色专知关注)

  • 后台回复“COLING” 可以获取《自然语言处理顶会COLING2020最佳论文》专知下载链接索引

专知,专业可信的人工智能知识分发,让认知协作更快更好!欢迎注册登录专知www.zhuanzhi.ai,获取5000+AI主题干货知识资料!
欢迎微信扫一扫加入专知人工智能知识星球群,获取最新AI专业干货知识教程资料和与专家交流咨询
点击“ 阅读原文 ”,了解使用 专知 ,查看获取5000+AI主题知识资源
登录查看更多
3

相关内容

专知会员服务
25+阅读 · 2021年4月2日
【快讯】KDD2020论文出炉,216篇上榜, 你的paper中了吗?
专知会员服务
50+阅读 · 2020年5月16日
ECCV 2020接受论文出炉,1361篇都在这了
专知
23+阅读 · 2020年7月4日
ICLR2019最佳论文出炉
专知
12+阅读 · 2019年5月6日
自然语言处理顶会EMNLP2018接受论文列表!
专知
87+阅读 · 2018年8月26日
自然语言处理顶会 ACL 2018 参会见闻
PaperWeekly
3+阅读 · 2018年7月26日
ACL 2018 计算语言学协会接受论文列表
专知
3+阅读 · 2018年4月27日
Arxiv
21+阅读 · 2019年8月21日
Arxiv
3+阅读 · 2018年3月28日
Arxiv
3+阅读 · 2018年3月2日
Arxiv
6+阅读 · 2018年2月28日
Arxiv
6+阅读 · 2018年2月26日
VIP会员
相关论文
Top
微信扫码咨询专知VIP会员