People are regularly confronted with potentially deceptive statements (e.g., fake news, misleading product reviews, or lies about activities). Only few works on automated text-based deception detection have exploited the potential of deep learning approaches. A critique of deep-learning methods is their lack of interpretability, preventing us from understanding the underlying (linguistic) mechanisms involved in deception. However, recent advancements have made it possible to explain some aspects of such models. This paper proposes and evaluates six deep-learning models, including combinations of BERT (and RoBERTa), MultiHead Attention, co-attentions, and transformers. To understand how the models reach their decisions, we then examine the model's predictions with LIME. We then zoom in on vocabulary uniqueness and the correlation of LIWC categories with the outcome class (truthful vs deceptive). The findings suggest that our transformer-based models can enhance automated deception detection performances (+2.11% in accuracy) and show significant differences pertinent to the usage of LIWC features in truthful and deceptive statements.
翻译:人们经常遇到潜在的欺骗性陈述(例如假新闻、误导性产品审查或对活动进行谎言);只有为数不多的关于自动文本欺骗探测的著作利用了深层学习方法的潜力;对深层学习方法的批评是它们缺乏解释性,使我们无法理解欺骗所涉及的基本(语言)机制;然而,最近的进展使我们有可能解释这些模型的某些方面;本文件提出并评价了六个深层学习模型,包括BERT(和RoBERTA)、多领导人注意、共同关注和变压器的组合;为了解模型是如何达到其决定的,我们然后研究模型与LIME的预测;然后我们放大了LIME的词汇独特性和LIWC类别与结果类的关联性(真实性和欺骗性);研究结果表明,我们基于变压器的模型可以提高自动欺骗性检测性表现(准确度为+2.11%),并显示在真实和欺骗性陈述中使用LICC特征方面存在重大差异。