Recently, data-driven inertial navigation approaches have demonstrated their capability of using well-trained neural networks to obtain accurate position estimates from inertial measurement units (IMU) measurements. In this paper, we propose a novel robust Contextual Transformer-based network for Inertial Navigation~(CTIN) to accurately predict velocity and trajectory. To this end, we first design a ResNet-based encoder enhanced by local and global multi-head self-attention to capture spatial contextual information from IMU measurements. Then we fuse these spatial representations with temporal knowledge by leveraging multi-head attention in the Transformer decoder. Finally, multi-task learning with uncertainty reduction is leveraged to improve learning efficiency and prediction accuracy of velocity and trajectory. Through extensive experiments over a wide range of inertial datasets~(e.g. RIDI, OxIOD, RoNIN, IDOL, and our own), CTIN is very robust and outperforms state-of-the-art models.
翻译:最近,数据驱动惯性导航方法展示了它们利用训练有素的神经网络从惯性测量单位的测量中获取准确位置估计的能力。在本文中,我们提议为惰性导航~(CTIN)建立一个新的强势环境变异器网络,以准确预测速度和轨迹。为此,我们首先设计一个基于ResNet的编码器,由当地和全球多头自我意识增强,以从IMU测量中获取空间背景信息。然后,我们通过在变换器脱coder中利用多头目的关注,将这些空间表示与时间知识结合起来。最后,利用多任务和减少不确定性的学习来提高学习效率和预测速度和轨迹的准确性。通过对广泛的惯性数据集(例如RIDI、OxIOD、RoNIN、IDOL和我们自己)进行广泛的实验,CTIN非常强大,而且不完善了最新模型。