The reliable prediction of the temporal behavior of complex systems is required in numerous scientific fields. This strong interest is however hindered by modeling issues: often, the governing equations describing the physics of the system under consideration are not accessible or, when known, their solution might require a computational time incompatible with the prediction time constraints. Nowadays, approximating complex systems at hand in a generic functional format and informing it ex--nihilo from available observations has become a common practice, as illustrated by the enormous amount of scientific work appeared in the last years. Numerous successful examples based on deep neural networks are already available, although generalizability of the models and margins of guarantee are often overlooked. Here, we consider Long-Short Term Memory neural networks and thoroughly investigate the impact of the training set and its structure on the quality of the long-term prediction. Leveraging insights from ergodic theory, we perform a thorough computational analysis to assess the amount of data sufficient for a priori guaranteeing a faithful model of the physical system. We show how an informed design of the training set, based on invariants of the system and the structure of the underlying attractor, significantly improves the resulting models, opening up avenues for research within the context of active learning. Further, the non-trivial effects of the memory initializations when relying on memory-capable models will be illustrated. Our findings provide evidence-based good-practice on the amount and the choice of data required for an effective data-driven modeling of any complex dynamical system.
翻译:在许多科学领域,需要对复杂系统的时间行为进行可靠的预测,但这种强烈的兴趣却受到模拟问题的影响:通常,描述所审议系统物理的治理方程式无法进入,或者当已知时,其解决方案可能需要一个与预测时间限制不相符的计算时间。如今,以通用功能格式将手头的复杂系统近似复杂系统,并从现有观测结果中为它提供参考,这已成为一种常见做法,过去几年里出现了大量科学工作,说明大量科学工作,许多基于深层神经网络的成功范例已经存在,尽管往往忽视模型的通用性和担保的边缘。在这里,我们考虑长期短期记忆神经网络,并彻底调查培训组及其结构对长期预测质量的影响。我们从原创理论中汲取洞察力,进行彻底的计算分析,评估数据数量足以预先保证物理系统的忠实模型。我们展示了如何根据系统变量和潜在吸引力模型的结构对培训组进行知情设计。我们考虑长期记忆记忆内记忆基础模型的动态分析将大大改进我们最初的模型的动态分析。