Deep generative models provide a powerful set of tools to understand real-world data. But as these models improve, they increase in size and complexity, so their computational cost in memory and execution time grows. Using binary weights in neural networks is one method which has shown promise in reducing this cost. However, whether binary neural networks can be used in generative models is an open problem. In this work we show, for the first time, that we can successfully train generative models which utilize binary neural networks. This reduces the computational cost of the models massively. We develop a new class of binary weight normalization, and provide insights for architecture designs of these binarized generative models. We demonstrate that two state-of-the-art deep generative models, the ResNet VAE and Flow++ models, can be binarized effectively using these techniques. We train binary models that achieve loss values close to those of the regular models but are 90%-94% smaller in size, and also allow significant speed-ups in execution time.
翻译:深基因模型提供了一套强大的工具来理解现实世界的数据。 但是随着这些模型的改进,它们的规模和复杂性会增加,因此它们的记忆和执行时间的计算成本会增加。 在神经网络中使用二进制重量是降低这一成本的一个有希望的方法。 但是,二进制神经网络能否在基因模型中使用是一个开放的问题。 在这项工作中,我们第一次显示,我们能够成功地培训使用二进制神经网络的基因模型。这大大降低了模型的计算成本。我们开发了一个新的二进制重量分类,并为这些二进制基因模型的建筑设计提供了见解。我们证明,两种最先进的深层次基因模型,即ResNet VAE 和 Flow++模型,可以使用这些技术有效地实现二进制化。我们培训的二进制模型能够实现接近常规模型的损失值,但规模小于90%至94%,并且可以让执行过程中的快速增长。