Many recent deep learning-based solutions have widely adopted the attention-based mechanism in various tasks of the NLP discipline. However, the inherent characteristics of deep learning models and the flexibility of the attention mechanism increase the models' complexity, thus leading to challenges in model explainability. In this paper, to address this challenge, we propose a novel practical framework by utilizing a two-tier attention architecture to decouple the complexity of explanation and the decision-making process. We apply it in the context of a news article classification task. The experiments on two large-scaled news corpora demonstrate that the proposed model can achieve competitive performance with many state-of-the-art alternatives and illustrate its appropriateness from an explainability perspective.
翻译:最近许多深层次的基于学习的解决方案广泛采用以关注为基础的机制来完成国家采购计划学科的各项任务,然而,深层次学习模式的固有特点和关注机制的灵活性增加了模型的复杂性,从而在模式解释方面造成挑战。为了应对这一挑战,我们在本文件中提出一个新的务实框架,利用一个两级关注结构来区分解释的复杂性和决策过程。我们将其应用于新闻文章分类任务。两个大型新闻公司公司的实验表明,拟议的模式可以实现竞争性业绩,有许多最先进的替代方案,并从解释的角度说明其适当性。