The recently proposed Quantum Neuron Born Machine (QNBM) has demonstrated quality initial performance as the first quantum generative machine learning (ML) model proposed with non-linear activations. However, previous investigations have been limited in scope with regards to the model's learnability and simulatability. In this work, we make a considerable leap forward by providing an extensive deep dive into the QNBM's potential as a generative model. We first demonstrate that the QNBM's network representation makes it non-trivial to be classically efficiently simulated. Following this result, we showcase the model's ability to learn (express and train on) a wider set of probability distributions, and benchmark the performance against a classical Restricted Boltzmann Machine (RBM). The QNBM is able to outperform this classical model on all distributions, even for the most optimally trained RBM among our simulations. Specifically, the QNBM outperforms the RBM with an improvement factor of 75.3x, 6.4x, and 3.5x for the discrete Gaussian, cardinality-constrained, and Bars and Stripes distributions respectively. Lastly, we conduct an initial investigation into the model's generalization capabilities and use a KL test to show that the model is able to approximate the ground truth probability distribution more closely than the training distribution when given access to a limited amount of data. Overall, we put forth a stronger case in support of using the QNBM for larger-scale generative tasks.
翻译:最近提议的Qantum 中产中产机械(QNBM) 展示了质量较高的初始性能,作为第一个非线性启动的量子基因化机器学习模型(ML), 显示了质量更高的初步性能, 然而, 先前的调查在模型的可学习性和可模拟性方面范围有限。 在这项工作中, 我们通过对QNBM的潜力进行广泛深入的下潜, 将QNBM的潜力作为一种发型模型, 我们首先表明,QNBM的网络代表制使得它不易被传统地有效地模拟。 之后, 我们展示了模型学习(表达和培训)更广泛的概率分布的(MMLM)能力, 并对照经典限制的Boltzmann机器(RBM) 测试性能。 QNBM能够在所有分布上超越这一经典模式, 即使是在我们模拟中经过最优化培训的成果管理制。 具体地, QNBMMM在改进后, 其总体支持度比75.3x、 6.4x 和3.5x 更强。 在离心性模型中, 模型中, 显示(表达、 基础- 基本数据分布) 将数据分配到我们开始的概率化的概率化的模型 显示, 显示, 我们的分级分级分级分级的分级分级分级的分级的分级的分级的分级的分级的分级到分级到分级的分级, 。