Home-cage social behaviour analysis of mice is an invaluable tool to assess therapeutic efficacy of neurodegenerative diseases. Despite tremendous efforts made within the research community, single-camera video recordings are mainly used for such analysis. Because of the potential to create rich descriptions of mouse social behaviors, the use of multi-view video recordings for rodent observations is increasingly receiving much attention. However, identifying social behaviours from various views is still challenging due to the lack of correspondence across data sources. To address this problem, we here propose a novel multiview latent-attention and dynamic discriminative model that jointly learns view-specific and view-shared sub-structures, where the former captures unique dynamics of each view whilst the latter encodes the interaction between the views. Furthermore, a novel multi-view latent-attention variational autoencoder model is introduced in learning the acquired features, enabling us to learn discriminative features in each view. Experimental results on the standard CRMI13 and our multi-view Parkinson's Disease Mouse Behaviour (PDMB) datasets demonstrate that our model outperforms the other state of the arts technologies and effectively deals with the imbalanced data problem.
翻译:尽管研究界作出了巨大努力,但单相机录像主要用于这种分析。由于有可能对老鼠的社会行为作出丰富的描述,使用多视视频记录进行鼠鼠观察越来越受到重视。然而,由于数据源之间缺乏通信,从各种观点中识别社会行为仍具有挑战性。为了解决这一问题,我们在此提出一个新的多视角潜意识和动态歧视模式,共同学习视觉和视觉共享的子结构,前者捕捉每种观点的独特动态,而后者则记录各种观点之间的相互作用。此外,在学习获得的特征时引入了新型的多视角潜意识变异自动编码模型,使我们能够学习每种观点中的歧视性特征。标准CRMI13的实验结果和我们多视角Parkinson的疾病鼠标行为(PDMB)数据集表明,我们的模型超越了艺术技术的另一种状态,有效地处理了不平衡的数据问题。