Autism diagnosis presents a major challenge due to the vast heterogeneity of the condition and the elusive nature of early detection. Atypical gait and gesture patterns are dominant behavioral characteristics of autism and can provide crucial insights for diagnosis. Furthermore, these data can be collected efficiently in a non-intrusive way, facilitating early intervention to optimize positive outcomes. Existing research mainly focuses on associating facial and eye-gaze features with autism. However, very few studies have investigated movement and gesture patterns which can reveal subtle variations and characteristics that are specific to autism. To address this gap, we present an analysis of gesture and gait activity in videos to identify children with autism and quantify the severity of their condition by regressing autism diagnostic observation schedule scores. Our proposed architecture addresses two key factors: (1) an effective feature representation to manifest irregular gesture patterns and (2) a two-stream co-learning framework to enable a comprehensive understanding of its relation to autism from diverse perspectives without explicitly using additional data modality. Experimental results demonstrate the efficacy of utilizing gesture and gait-activity videos for autism analysis.
翻译:自闭症的诊断是一个巨大的挑战,因为该疾病有很大的异质性和早期检测的难度。 非典型的步态和手势模式是自闭症的主要行为特征,可以为诊断提供关键的见解。此外,这些数据可以以非侵入性的方式高效地收集,促进早期干预以优化积极结果。现有的研究主要集中在将面部和眼神特征与自闭症相关联。然而,极少数研究调查了运动和手势模式,这些模式可以揭示特定于自闭症的微妙变化和特征。为了弥补这一差距,我们提出了通过分析视频中的手势和步态活动来识别患有自闭症的儿童并回归自闭症诊断观察表分数来量化其病情严重程度。我们提出的架构解决了两个关键因素:(1)有效的特征表示以展示不规则的手势模式和(2)两个流共同学习框架,实现了全面理解不同视角的手势和步态活动与自闭症之间的关系,无需明确使用其他数据模态。实验结果表明,利用手势和步态活动视频进行自闭症分析的有效性。