We propose cross-modal attentive connections, a new dynamic and effective technique for multimodal representation learning from wearable data. Our solution can be integrated into any stage of the pipeline, i.e., after any convolutional layer or block, to create intermediate connections between individual streams responsible for processing each modality. Additionally, our method benefits from two properties. First, it can share information uni-directionally (from one modality to the other) or bi-directionally. Second, it can be integrated into multiple stages at the same time to further allow network gradients to be exchanged in several touch-points. We perform extensive experiments on three public multimodal wearable datasets, WESAD, SWELL-KW, and CASE, and demonstrate that our method can effectively regulate and share information between different modalities to learn better representations. Our experiments further demonstrate that once integrated into simple CNN-based multimodal solutions (2, 3, or 4 modalities), our method can result in superior or competitive performance to state-of-the-art and outperform a variety of baseline uni-modal and classical multimodal methods.
翻译:我们提出跨式关注联系,这是从可磨损的数据中学习多式联运代表的新的动态有效技术,我们的解决办法可以纳入管道的任何阶段,即,在任何卷积层或块块之后,在负责处理每种模式的单个流之间建立中间联系,此外,我们的方法可以有两个属性。首先,我们的方法可以单向地(从一种模式到另一种模式)或双向地分享信息。第二,可以同时将其纳入多个阶段,以便进一步使网络梯度在几个接触点上交换。我们对三种公共多式联运可磨损数据集(WESAD、SWELL-KW和CASE)进行了广泛的实验,并表明我们的方法可以有效地调节不同模式之间的信息并分享,以了解更好的表述方式。我们的实验进一步证明,一旦融入了基于CNN的简单多式联运解决方案(2、3或4种模式),我们的方法可以导致高水平或竞争性的绩效,使各种基线的单式和古型多式联运方法超越。