Likelihood-based inferences have been remarkably successful in wide-spanning application areas. However, even after due diligence in selecting a good model for the data at hand, there is inevitably some amount of model misspecification: outliers, data contamination or inappropriate parametric assumptions such as Gaussianity mean that most models are at best rough approximations of reality. A significant practical concern is that for certain inferences, even small amounts of model misspecification may have a substantial impact; a problem we refer to as brittleness. This article attempts to address the brittleness problem in likelihood-based inferences by choosing the most model friendly data generating process in a discrepancy-based neighbourhood of the empirical measure. This leads to a new Optimistically Weighted Likelihood (OWL), which robustifies the original likelihood by formally accounting for a small amount of model misspecification. Focusing on total variation (TV) neighborhoods, we study theoretical properties, develop inference algorithms and illustrate the methodology in applications to mixture models and regression.
翻译:似然度的推理在广泛的应用领域中取得了显著的成功。然而,即使在为手头数据选择了好的模型之后,仍然不可避免地存在一定程度的模型错误拟合:异常值、数据污染或不合适的参数假设,如高斯性,使得大多数模型最多只是现实的粗略近似。一个重要的实际问题是,在某些推理中,即使存在小量的模型错误拟合,也可能产生显著影响;我们将此问题称为脆性问题。本文试图通过在经验测量的差异性邻域中选择最适合模型的数据生成过程来解决基于似然度推理中的脆性问题。这导致了一种新的“乐观加权似然度” (OWL) 方法,通过正式考虑少量的模型错误拟合来增强原始似然度的鲁棒性。聚焦在总变差 (TV) 邻域,我们研究了理论性质,开发了推理算法,并在混合模型和回归应用中说明了该方法。