Tensor networks are a powerful modeling framework developed for computational many-body physics, which have only recently been applied within machine learning. In this work we utilize a uniform matrix product state (u-MPS) model for probabilistic modeling of sequence data. We first show that u-MPS enable sequence-level parallelism, with length-n sequences able to be evaluated in depth O(log n). We then introduce a novel generative algorithm giving trained u-MPS the ability to efficiently sample from a wide variety of conditional distributions, each one defined by a regular expression. Special cases of this algorithm correspond to autoregressive and fill-in-the-blank sampling, but more complex regular expressions permit the generation of richly structured data in a manner that has no direct analogue in neural generative models. Experiments on sequence modeling with synthetic and real text data show u-MPS outperforming a variety of baselines and effectively generalizing their predictions in the presence of limited data.
翻译:在这项工作中,我们使用统一的矩阵产品状态模型(u-MPS)模型来对序列数据进行概率建模。我们首先表明,u-MPS能够实现序列级平行,而长度序列可以在O(log n)深度下进行评估。然后我们引入一种新的基因变现算法,使经过培训的u-MPS能够从多种有条件分布中有效取样,每个分布都是定期表达的。这种算法的特殊情况相当于自动递增和填充的抽样,但更复杂的经常表达法允许生成结构丰富的数据,其方式在神经基因化模型中没有直接的类似。用合成和真实文本数据进行序列建模的实验显示,u-MPS在有限的数据存在的情况下超过各种基线,并有效地概括了它们的预测。