Automating sleep staging is vital to scale up sleep assessment and diagnosis to serve millions experiencing sleep deprivation and disorders and enable longitudinal sleep monitoring in home environments. Learning from raw polysomnography signals and their derived time-frequency image representations has been prevalent. However, learning from multi-view inputs (e.g., both the raw signals and the time-frequency images) for sleep staging is difficult and not well understood. This work proposes a sequence-to-sequence sleep staging model, XSleepNet, that is capable of learning a joint representation from both raw signals and time-frequency images. Since different views may generalize or overfit at different rates, the proposed network is trained such that the learning pace on each view is adapted based on their generalization/overfitting behavior. In simple terms, the learning on a particular view is speeded up when it is generalizing well and slowed down when it is overfitting. View-specific generalization/overfitting measures are computed on-the-fly during the training course and used to derive weights to blend the gradients from different views. As a result, the network is able to retain the representation power of different views in the joint features which represent the underlying distribution better than those learned by each individual view alone. Furthermore, the XSleepNet architecture is principally designed to gain robustness to the amount of training data and to increase the complementarity between the input views. Experimental results on five databases of different sizes show that XSleepNet consistently outperforms the single-view baselines and the multi-view baseline with a simple fusion strategy. Finally, XSleepNet also outperforms prior sleep staging methods and improves previous state-of-the-art results on the experimental databases.
翻译:自动入睡对于扩大睡眠评估和诊断对于扩大睡眠评估和诊断至关重要,以便帮助数百万人经历睡眠被剥夺和失调,并能够在家庭环境中进行纵向睡眠监测。 学习原始聚光成像信号及其衍生的时间频率图像表现非常普遍。 但是, 学习多视图投入( 如原始信号和时间频率图像)对于睡眠变化来说是困难的, 并且没有很好地理解。 这项工作提议了一个从原始信号和时频图像中学习联合代表的顺序到顺序的睡眠准备模式XSLeptNet。 由于不同的观点可能以不同的速度概括或过度适应, 拟议的网络经过培训, 每一个视图的学习速度都以其一般化/ 超常化行为为基础加以调整。 简单地说, 当某个特定视图在概括化时会加快学习速度, 当它过大时会减缓。 在培训课程期间, 将特定的总体/ 措施在现场计算, 并用来从原始信号和时间频谱图像中得出权重值。 由于结果, 网络能够保持每个视图的宽度或超常变速度, 以其一般化的缩缩缩缩缩缩结构 显示在前的图像中, 。 将每个预测到预测到预测到预测到预测测测的精确度 。 。