Synthesizing human motion through learning techniques is becoming an increasingly popular approach to alleviating the requirement of new data capture to produce animations. Learning to move naturally from music, i.e., to dance, is one of the more complex motions humans often perform effortlessly. Each dance movement is unique, yet such movements maintain the core characteristics of the dance style. Most approaches addressing this problem with classical convolutional and recursive neural models undergo training and variability issues due to the non-Euclidean geometry of the motion manifold structure.In this paper, we design a novel method based on graph convolutional networks to tackle the problem of automatic dance generation from audio information. Our method uses an adversarial learning scheme conditioned on the input music audios to create natural motions preserving the key movements of different music styles. We evaluate our method with three quantitative metrics of generative methods and a user study. The results suggest that the proposed GCN model outperforms the state-of-the-art dance generation method conditioned on music in different experiments. Moreover, our graph-convolutional approach is simpler, easier to be trained, and capable of generating more realistic motion styles regarding qualitative and different quantitative metrics. It also presented a visual movement perceptual quality comparable to real motion data.
翻译:通过学习技术合成人类运动正在成为一种越来越受欢迎的办法,以减轻对新数据采集的要求,以制作动画。学习自然地从音乐(即舞蹈)到舞蹈,这是人类通常不费力的更复杂的运动之一。每个舞蹈运动都是独特的,但这种运动保持舞蹈风格的核心特点。通过古典变幻和循环神经模型解决这一问题的多数方法都因运动形体结构的非欧立德几何学而接受了培训和变异问题。在本文中,我们设计了一种基于图形革命网络的新方法,以解决从音频信息中自动产生舞蹈的问题。我们的方法使用一种以投入音乐音频为条件的对抗性学习计划来创造自然运动,以保存不同音乐风格的关键运动。我们用三种定量的变异方法来评价我们的方法和用户研究。结果显示,拟议的GCN模型超越了在不同实验中以音乐为条件的状态和艺术舞蹈生成方法。此外,我们用图形革命网络的方法更简单、更方便地、更便于被训练,并且能够以投入的音乐生成更现实的定性和可比较的定量数据流。