Conventional methods for human pose estimation either require a high degree of instrumentation, by relying on many inertial measurement units (IMUs), or constraint the recording space, by relying on extrinsic cameras. These deficits are tackled through the approach of human pose estimation from sparse IMU data. We define adjacency adaptive graph convolutional long-short term memory networks (AAGC-LSTM), to tackle human pose estimation based on six IMUs, while incorporating the human body graph structure directly into the network. The AAGC-LSTM combines both spatial and temporal dependency in a single network operation, more memory efficiently than previous approaches. This is made possible by equipping graph convolutions with adjacency adaptivity, which eliminates the problem of information loss in deep or recurrent graph networks, while it also allows for learning unknown dependencies between the human body joints. To further boost accuracy, we propose longitudinal loss weighting to consider natural movement patterns. With our presented approach, we are able to utilize the inherent graph nature of the human body, and thus can outperform the state of the art (SOTA) for human pose estimation from sparse IMU data.
翻译:人类表面估计的常规方法要么需要高度的仪器,依靠许多惯性测量单位(IMUs),要么依靠外部照相机限制记录空间。这些缺陷是通过利用稀薄的IMU数据对人构成进行估计的方法来解决的。我们定义了相近的图形相容性图象长期短期内存网络(AGC-LSTM),以便根据6个IMU(AGC-LSTM)进行人类构成估计,同时将人体的图象结构直接纳入网络。AAGC-LSTM将单一网络操作的空间和时间依赖性结合起来,比以前的方法更高效地结合了记忆。这可以通过使图象变异与相相适应性相结合来解决,从而消除了深度或经常性图形网络中的信息损失问题,同时也有助于学习人体连接之间未知的相互依存关系。为了进一步提高准确性,我们建议纵向减重,以考虑自然运动模式。我们提出的方法是能够利用人体的内在图形性质,从而比以前的方法能够超越从流式数据中得出人类形态的状态。