This work derives a theoretical value for the entropy of a Linear Additive Markov Process (LAMP), an expressive model able to generate sequences with a given autocorrelation structure. While a first-order Markov Chain model generates new values by conditioning on the current state, the LAMP model takes the transition state from the sequence's history according to some distribution which does not have to be bounded. The LAMP model captures complex relationships and long-range dependencies in data with similar expressibility to a higher-order Markov process. While a higher-order Markov process has a polynomial parameter space, a LAMP model is characterised only by a probability distribution and the transition matrix of an underlying first-order Markov Chain. We prove that the theoretical entropy rate of a LAMP is equivalent to the theoretical entropy rate of the underlying first-order Markov Chain. This surprising result is explained by the randomness introduced by the random process which selects the LAMP transitioning state, and provides a tool to model complex dependencies in data while retaining useful theoretical results. We use the LAMP model to estimate the entropy rate of the LastFM, BrightKite, Wikispeedia and Reuters-21578 datasets. We compare estimates calculated using frequency probability estimates, a first-order Markov model and the LAMP model, and consider two approaches to ensuring the transition matrix is irreducible. In most cases the LAMP entropy rates are lower than those of the alternatives, suggesting that LAMP model is better at accommodating structural dependencies in the processes.
翻译:这项工作为Linear Additive Markov 进程( LAMP) 的精度带来一个理论值, 这是一种能生成序列的表达式模型, 能够以给定的自动调节结构生成序列。 虽然第一个顺序的Markov 链模式通过调整当前状态产生新的值, 但LAMP 模式从序列历史中得出过渡状态, 根据一些不需要约束的分布。 LAMP 模式在数据中捕捉复杂的关系和长距离依赖性, 与更高顺序的Markov 进程相类似。 虽然一个更高顺序的Markov 进程有一个多元参数空间, 但LAMP 结构模型的特征只能通过一个概率分布和对一个基本顺序的Markov 链的过渡矩阵矩阵模型生成新的值。 我们证明, LAMMP 模式的理论性温和率相当于某些基本线的理论性。 我们使用LAMP- 2 模型来计算出一个更精确的RIMFI 模型, 以更精确的RIMI 和最精确的RIMF 模型来测测算。