Recurrent neural networks are used to forecast time series in finance, climate, language, and from many other domains. Reservoir computers are a particularly easily trainable form of recurrent neural network. Recently, a "next-generation" reservoir computer was introduced in which the memory trace involves only a finite number of previous symbols. We explore the inherent limitations of finite-past memory traces in this intriguing proposal. A lower bound from Fano's inequality shows that, on highly non-Markovian processes generated by large probabilistic state machines, next-generation reservoir computers with reasonably long memory traces have an error probability that is at least ~ 60% higher than the minimal attainable error probability in predicting the next observation. More generally, it appears that popular recurrent neural networks fall far short of optimally predicting such complex processes. These results highlight the need for a new generation of optimized recurrent neural network architectures. Alongside this finding, we present concentration-of-measure results for randomly-generated but complex processes. One conclusion is that large probabilistic state machines -- specifically, large $\epsilon$-machines -- are key to generating challenging and structurally-unbiased stimuli for ground-truthing recurrent neural network architectures.
翻译:回归神经网络用于预测金融、气候、语言和许多其他领域的时间序列。蓄水池计算机是一种特别易于训练的递归神经网络。最近,一种“下一代”蓄水池计算机被引入,其中记忆痕迹仅涉及有限数量的先前符号。我们探索有限过去记忆痕迹的固有限制在这个有趣的提议中。来自 Fano 不等式的下界表明,在由大型概率状态机生成的高度非马尔科夫过程上,具有相当长的记忆痕迹的下一代蓄水池计算机的误差概率比预测下一个观测所能达到的最小可达误差概率高至少 ~ 60%。更一般而言,似乎流行的递归神经网络远远不足以最优地预测这种复杂过程。这些结果凸显了一代优化的递归神经网络架构的需求。除此之外,我们还提供关于随机生成的但是具有复杂结构的过程的测量集中结果。一个结论是,大型概率状态机——具体来说是大型 $\epsilon$-machines——是为递归神经网络架构提供挑战性和结构化无偏刺激的关键。