Due to various and serious adverse impacts of spreading fake news, it is often known that only people with malicious intent would propagate fake news. However, it is not necessarily true based on social science studies. Distinguishing the types of fake news spreaders based on their intent is critical because it will effectively guide how to intervene to mitigate the spread of fake news with different approaches. To this end, we propose an intent classification framework that can best identify the correct intent of fake news. We will leverage deep reinforcement learning (DRL) that can optimize the structural representation of each tweet by removing noisy words from the input sequence when appending an actor to the long short-term memory (LSTM) intent classifier. Policy gradient DRL model (e.g., REINFORCE) can lead the actor to a higher delayed reward. We also devise a new uncertainty-aware immediate reward using a subjective opinion that can explicitly deal with multidimensional uncertainty for effective decision-making. Via 600K training episodes from a fake news tweets dataset with an annotated intent class, we evaluate the performance of uncertainty-aware reward in DRL. Evaluation results demonstrate that our proposed framework efficiently reduces the number of selected words to maintain a high 95\% multi-class accuracy.
翻译:由于传播假新闻产生的各种严重不利影响,人们往往知道,只有恶意的人才会传播假新闻,但根据社会科学研究,这不一定是真实的。根据意图区分假新闻传播者类别至关重要,因为这样做将有效指导如何进行干预,以不同方式减少假新闻的传播。为此,我们提出一个意图分类框架,以最佳方式确定假新闻的正确意图。我们将利用深度强化学习(DRL),使每条推特的结构性代表性得到优化,在将一个行为体附在长期短期记忆(LSTM)意图分类器(LSTM)输入输入序列时,它会从输入序列中删除噪音词。政策梯度 DRL模型(例如REINFORCE)可以引导该行为体获得更高的延迟奖励。我们还利用一种能够明确处理有效决策的多层面不确定性的主观意见,设计新的不确定性即时奖励。我们将利用一个带有附加说明的意向类的假新闻推特数据集制作的600K培训片段,我们评估DRL长期记忆(L)意图分类中的不确定性奖励表现。评估结果显示,我们提议的框架能够有效减少所选的多级词汇的准确性。