Normalizing flow is a class of deep generative models for efficient sampling and density estimation. In practice, the flow often appears as a chain of invertible neural network blocks; to facilitate training, existing works have regularized flow trajectories and designed special network architectures. The current paper develops a neural ODE flow network inspired by the Jordan-Kinderleherer-Otto (JKO) scheme, which allows efficient block-wise training of the residual blocks and avoids inner loops of score matching or variational learning. As the JKO scheme unfolds the dynamic of gradient flow, the proposed model naturally stacks residual network blocks one-by-one, reducing the memory load and difficulty of performing end-to-end training of deep flow networks. We also develop adaptive time reparameterization of the flow network with a progressive refinement of the trajectory in probability space, which improves the model training efficiency and accuracy in practice. Using numerical experiments with synthetic and real data, we show that the proposed JKO-iFlow model achieves similar or better performance in generating new samples compared with existing flow and diffusion models at a significantly reduced computational and memory cost.
翻译:在实践上,流动往往表现为一连串不可逆神经网络块;为了便利培训,现有工程有正常的流程轨迹,并设计了特殊的网络结构。本文开发了一个由Jordan-Kinderleherer-Ottto(JKO)计划启发的神经流流网络,从而可以对残余块进行有效的轮廓培训,避免得分匹配或变异学习的内在循环。随着JKO计划展现梯度流动的动态,拟议的模型自然堆积网络块将逐个地块隔开来,减少了对深流网络进行端到端培训的记忆负荷和难度。我们还开发了流动网络的适应性时间重新计数,逐步完善了概率空间的轨迹,提高了模型培训效率和实践中的准确性。我们用合成数据和真实数据进行数字实验,表明拟议的JKO-ilow模型在生成新样品方面取得了类似或更好的业绩,而现有模型则以大幅降低的计算成本和记忆成本,与现有的流动和传播模型相比,在生成新样品方面实现了类似或更好的业绩。