Inferring models from samples of stochastic processes is challenging, even in the most basic setting in which processes are stationary and ergodic. A principle reason for this, discovered by Blackwell in 1957, is that finite-state generators produce realizations with arbitrarily-long dependencies that require an uncountable infinity of probabilistic features be tracked for optimal prediction. This, in turn, means predictive models are generically infinite-state. Specifically, hidden Markov chains, even if finite, generate stochastic processes that are irreducibly complicated. The consequences are dramatic. For one, no finite expression for their Shannon entropy rate exists. Said simply, one cannot make general statements about how random they are and finite models incur an irreducible excess degree of unpredictability. This was the state of affairs until a recently-introduced method showed how to accurately calculate their entropy rate and, constructively, to determine the minimal set of infinite predictive features. Leveraging this, here we address the complementary challenge of determining how structured hidden Markov processes are by calculating the rate of statistical complexity divergence -- the information dimension of the minimal set of predictive features.
翻译:从随机过程样本中推断模型是具有挑战性的,即使在程序是固定和随机的最基本的环境中也是如此。1957年,布莱克韦尔发现,造成这种情况的一个主要理由是,有限状态发电机具有任意的长期依赖性,需要追踪无法计算的无限概率特征,才能作出最佳预测。这反过来又意味着预测模型一般是无限的。具体地说,隐藏的马尔科夫链,即使有限,也会产生不可复制的复杂过程。后果是巨大的。对于一个问题,其香农诱导率没有有限的表达方式存在。简单地说,不能就它们是如何随机的和有限模型产生不可避免的过度不可预测性程度作一般性陈述。这就是直到最近采用的方法表明如何准确计算其恒温率,以及建设性地确定最起码的无限预测特征组。我们在此探讨如何通过计算统计复杂性的差异来确定结构化的隐形马可夫进程这一补充挑战,即最低预测特征的信息层面。