Electroencephalography (EEG) and language have been widely explored independently for many downstream tasks (e.g., sentiment analysis, relation detection, etc.). Multimodal approaches that study both domains have not been well explored, even though in recent years, multimodal learning has been seen to be more powerful than its unimodal counterparts. In this study, we want to explore the relationship and dependency between EEG and language, i.e., how one domain reflects and represents the other. To study the relationship at the representation level, we introduced MTAM, a Multimodal Transformer Alignment Model, to observe coordinated representations between the two modalities, and thus employ the transformed representations for downstream applications. We used various relationship alignment-seeking techniques, such as Canonical Correlation Analysis and Wasserstein Distance, as loss functions to transfigure low-level language and EEG features to high-level transformed features. On downstream applications, sentiment analysis, and relation detection, we achieved new state-of-the-art results on two datasets, ZuCo and K-EmoCon. Our method achieved an F1-score improvement of 16.5% on sentiment analysis for K-EmoCon, 26.6% on sentiment analysis of ZuCo, and 31.1% on relation detection of ZuCo. In addition, we provide interpretation of the performance improvement by: (1) visualizing the original feature distribution and the transformed feature distribution, showing the effectiveness of the alignment module for discovering and encoding the relationship between EEG and language; (2) visualizing word-level and sentence-level EEG-language alignment weights, showing the influence of different language semantics as well as EEG frequency features; and (3) visualizing brain topographical maps to provide an intuitive demonstration of the connectivity of EEG and language response in the brain regions.
翻译:许多下游任务(例如情绪分析、关系探测等)独立地对电电子学(EEEG)和语言进行了广泛的独立探讨。研究这两个领域的多模式方法尚未得到很好探讨,尽管近年来,多式学习被认为比单式对口更为强大。在本研究中,我们希望探索EEEG和语言之间的关系和依赖性,即一个领域如何反映和代表另一个领域。为了研究代表性层面的关系,我们引入了MTAM(多式变换器调整模型),以观察两种模式之间协调的表达方式,从而对下游应用程序采用变换的表达方式。我们使用了各种关系协调方法,例如Canonical Colorelation分析 和 Vasserstem距离等,作为将低级别语言和EEEEEG特征转换成高层次。 关于下游应用、情绪分析,我们实现了两个数据集、Z-CUCO的显示E1-直径对E值分析的直观和直径分析的直径分析,提供了EEE-O-Sialalal-realalalalalalalalalalalalalalalisal 。