Embodied Conversational Agents that make use of co-speech gestures can enhance human-machine interactions in many ways. In recent years, data-driven gesture generation approaches for ECAs have attracted considerable research attention, and related methods have continuously improved. Real-time interaction is typically used when researchers evaluate ECA systems that generate rule-based gestures. However, when evaluating the performance of ECAs based on data-driven methods, participants are often required only to watch pre-recorded videos, which cannot provide adequate information about what a person perceives during the interaction. To address this limitation, we explored use of real-time interaction to assess data-driven gesturing ECAs. We provided a testbed framework, and investigated whether gestures could affect human perception of ECAs in the dimensions of human-likeness, animacy, perceived intelligence, and focused attention. Our user study required participants to interact with two ECAs - one with and one without hand gestures. We collected subjective data from the participants' self-report questionnaires and objective data from a gaze tracker. To our knowledge, the current study represents the first attempt to evaluate data-driven gesturing ECAs through real-time interaction and the first experiment using gaze-tracking to examine the effect of ECAs' gestures.
翻译:使用共同声音手势的实时交流器可以在许多方面加强人体-机械的互动。近年来,数据驱动的ECA手势生成方法吸引了相当多的研究关注,相关方法也不断改进。研究人员在评价ECA系统时通常使用实时互动来评价产生有章可循的手势的ECA系统。然而,在根据数据驱动方法评价ECA的绩效时,参与者往往只需要观看预先录制的录像,这些录像无法提供一个人在互动过程中所感知的足够信息。为了应对这一局限性,我们探索了实时互动来评估数据驱动的ECA。我们提供了一个测试框架,并调查了手势是否会影响ECA在人类相似性、敏锐性、感知觉智能和专注方面对人的看法。我们的用户研究要求参与者与两个ECA系统进行互动——一个是用手动手势的,一个是不用手动手势的。我们从参与者的自我报告调查表中收集了主观数据数据,并且从视觉追踪者那里收集了客观数据。根据我们的知识,目前的研究首次尝试通过实际时间对AECA的手势性互动来评价对AECA的图像的实验。