论文标题
舞蹈革命:通过课程学习与音乐的长期舞蹈一代
Dance Revolution: Long-Term Dance Generation with Music via Curriculum Learning
论文作者
论文摘要
从远古时代开始跳舞音乐是人类的先天能力之一。但是,在机器学习研究中,从音乐中综合舞蹈运动是一个具有挑战性的问题。最近,研究人员通过复发性神经网络(RNN)等自回归模型综合了人类运动序列。由于预测错误的积累,这种方法通常会产生短序列,这些误差被反馈到神经网络中。在长运动序列的产生中,这个问题变得更加严重。此外,在模特期间,舞蹈和音乐之间的一致性尚未考虑到。在本文中,我们将音乐条件的舞蹈产生形式化为序列学习问题,并设计出一种新颖的Seq2Seq架构,以有效地处理长时间的音乐功能序列,并捕获音乐和舞蹈之间的细粒度对应关系。此外,我们提出了一种新型的课程学习策略,以减轻长运动序列生成中自回归模型的错误积累,该过程将训练过程从训练过程中从使用先前的基础真实运动的完全指导的教师攻击方案中,转变为一种较不引导的自动回归方案,主要是使用生成的运动。广泛的实验表明,我们的方法在自动指标和人类评估方面的现有最新作品明显优于现有的最新面貌。我们还制作了一个演示视频,以在https://www.youtube.com/watch?v=lme20mehez8上展示我们提出的方法的出色性能。
Dancing to music is one of human's innate abilities since ancient times. In machine learning research, however, synthesizing dance movements from music is a challenging problem. Recently, researchers synthesize human motion sequences through autoregressive models like recurrent neural network (RNN). Such an approach often generates short sequences due to an accumulation of prediction errors that are fed back into the neural network. This problem becomes even more severe in the long motion sequence generation. Besides, the consistency between dance and music in terms of style, rhythm and beat is yet to be taken into account during modeling. In this paper, we formalize the music-conditioned dance generation as a sequence-to-sequence learning problem and devise a novel seq2seq architecture to efficiently process long sequences of music features and capture the fine-grained correspondence between music and dance. Furthermore, we propose a novel curriculum learning strategy to alleviate error accumulation of autoregressive models in long motion sequence generation, which gently changes the training process from a fully guided teacher-forcing scheme using the previous ground-truth movements, towards a less guided autoregressive scheme mostly using the generated movements instead. Extensive experiments show that our approach significantly outperforms the existing state-of-the-arts on automatic metrics and human evaluation. We also make a demo video to demonstrate the superior performance of our proposed approach at https://www.youtube.com/watch?v=lmE20MEheZ8.