Recently there have been many research efforts in developing generative models for self-exciting point processes, partly due to their broad applicability for real-world applications. However, rarely can we quantify how well the generative model captures the nature or ground-truth since it is usually unknown. The challenge typically lies in the fact that the generative models typically provide, at most, good approximations to the ground-truth (e.g., through the rich representative power of neural networks), but they cannot be precisely the ground-truth. We thus cannot use the classic goodness-of-fit (GOF) test framework to evaluate their performance. In this paper, we develop a GOF test for generative models of self-exciting processes by making a new connection to this problem with the classical statistical theory of Quasi-maximum-likelihood estimator (QMLE). We present a non-parametric self-normalizing statistic for the GOF test: the Generalized Score (GS) statistics, and explicitly capture the model misspecification when establishing the asymptotic distribution of the GS statistic. Numerical simulation and real-data experiments validate our theory and demonstrate the proposed GS test's good performance.
翻译:最近,在开发自我探索点过程的基因化模型方面进行了许多研究努力,部分原因是这些模型广泛适用于现实应用,然而,我们很少能够量化基因化模型在自然或地面真实性方面捕捉自然或地面真实性的程度,因为它通常并不为人所知。挑战通常在于,基因化模型通常能向地面真相提供良好的近似(例如,通过神经网络具有丰富代表性的力量),但它们不能精确地反映地面真相。因此,我们不能使用典型的“特惠”测试框架来评价其性能。在本文件中,我们开发了一个GOF测试,通过将这一问题与Qasi-最接近的测算仪(QMLE)的典型统计理论重新联系起来,来测试自我激励过程的基因化模型。我们为GOF测试提供了一种非参数性的自我标准化统计:通用评级(GS)统计数据,在确定拟议的GGS统计的“特征”分布时,我们无法明确捕捉到模型的不精确特性。我们用模型模拟和真实的数据验证,展示了我们拟议的全球统计的模拟和真实性试验。