Text classification systems based on contextual embeddings are not viable options for many of the low resource languages. On the other hand, recently introduced capsule networks have shown performance in par with these text classification models. Thus, they could be considered as a viable alternative for text classification for languages that do not have pre-trained contextual embedding models. However, current capsule networks depend upon spatial patterns without considering the sequential features of the text. They are also sub-optimal in capturing the context-level information in longer sequences. This paper presents a novel Dual-State Capsule (DS-Caps) network-based technique for text classification, which is optimized to mitigate these issues. Two varieties of states, namely sentence-level and word-level, are integrated with capsule layers to capture deeper context-level information for language modeling. The dynamic routing process among capsules was also optimized using the context-level information obtained through sentence-level states. The DS-Caps networks outperform the existing capsule network architectures for multiple datasets, particularly for tasks with longer sequences of text. We also demonstrate the superiority of DS-Caps in text classification for a low resource language.
翻译:对许多低资源语言而言,基于背景嵌入的文本分类系统并不是可行的选择。另一方面,最近引进的胶囊网络显示了与这些文本分类模型相同的性能。因此,可以将它们视为对于没有经过事先培训的背景嵌入模型的语文的文本分类的一种可行的替代方法。然而,目前的胶囊网络依赖于空间模式,而没有考虑到文本的顺序特征。在以较长的顺序获取上下文级信息时,它们也是次优的。本文件展示了一种新的基于文本分类的双州胶囊(DS-Caps)网络技术(DS-Caps),这是为缓解这些问题而最优化的。两种类型的国家,即句级和字级,都与胶囊层融合在一起,以获取更深的上下文级语言建模信息。在胶囊间动态定线过程也利用通过判决级状态获得的上下文级信息加以优化。DS-Caps网络超越了现有多数据集的胶囊网络结构,特别是较长的文本序列。我们还展示了DS-Caps的优越性。