Generative Adversarial Networks have recently shown promise for video generation, building off of the success of image generation while also addressing a new challenge: time. Although time was analyzed in some early work, the literature has not adequately grown with temporal modeling developments. We study the effects of Neural Differential Equations to model the temporal dynamics of video generation. The paradigm of Neural Differential Equations presents many theoretical strengths including the first continuous representation of time within video generation. In order to address the effects of Neural Differential Equations, we investigate how changes in temporal models affect generated video quality. Our results give support to the usage of Neural Differential Equations as a simple replacement for older temporal generators. While keeping run times similar and decreasing parameter count, we produce a new state-of-the-art model in 64$\times$64 pixel unconditional video generation, with an Inception Score of 15.20.
翻译:创世网络最近展示了对视频生成的希望,在图像生成成功的同时应对新的挑战:时间。虽然在一些早期工作中对时间进行了分析,但文献并没有随着时间模型的发展而充分增长。我们研究了神经差异等量的影响,以模拟视频生成的时间动态。神经差异模型呈现了许多理论优势,包括在视频生成过程中时间的首次连续表达。为了应对神经差异等量的影响,我们研究了时间模型的变化如何影响生成视频质量。我们的结果支持使用神经差异等量作为更老的时间生成器的简单替代。我们虽然保持运行时间相似和减少参数计数,但我们制作了64美元和64美元的新型最新工艺型无条件视频生成,其感知分数为15.20。