In the last decade, an increasing number of users have started reporting Adverse Drug Events (ADE) on social media platforms, blogs, and health forums. Given the large volume of reports, pharmacovigilance has focused on ways to use Natural Language Processing (NLP) techniques to rapidly examine these large collections of text, detecting mentions of drug-related adverse reactions to trigger medical investigations. However, despite the growing interest in the task and the advances in NLP, the robustness of these models in face of linguistic phenomena such as negations and speculations is an open research question. Negations and speculations are pervasive phenomena in natural language, and can severely hamper the ability of an automated system to discriminate between factual and nonfactual statements in text. In this paper we take into consideration four state-of-the-art systems for ADE detection on social media texts. We introduce SNAX, a benchmark to test their performance against samples containing negated and speculated ADEs, showing their fragility against these phenomena. We then introduce two possible strategies to increase the robustness of these models, showing that both of them bring significant increases in performance, lowering the number of spurious entities predicted by the models by 60% for negation and 80% for speculations.
翻译:在过去的十年中,越来越多的用户开始在社交媒体平台、博客和卫生论坛上报告反毒毒品事件(ADE),鉴于报告数量庞大,药物警惕侧重于如何利用自然语言处理(NLP)技术快速审查这些大量文本,发现与毒品有关的不良反应引发医学调查;然而,尽管人们对这项任务的兴趣日益浓厚,而且国家禁毒方案也取得了进展,但这些模型在面临诸如否定和投机等语言现象时的稳健性是一个公开的研究问题;错误和投机是自然语言中普遍存在的现象,可能严重妨碍自动化系统在文本中区分事实和非事实声明的能力;在本文件中,我们考虑到四种先进的系统,用于在社会媒体文本上检测与毒品有关的不良反应;我们采用SNAX,这是用来测试其业绩的基准,用含有否定和推测的ADE的样本来测试这些模型的脆弱性;然后我们提出两种可能的战略,以提高这些模型的稳健性,表明这两种模式都带来显著的自然语言现象,并可能严重妨碍自动系统在事实和非事实陈述之间进行区分的能力;在这份文件中,我们考虑到四种最先进的社会媒体文本中,我们通过预测了80 %的投机性实体,降低了80 %。