Logistic regression models for binomial responses are routinely used in statistical practice. However, the maximum likelihood estimate may not exist due to data separability. We address this issue by considering a conjugate prior penalty which always produces finite estimates. Such a specification has a clear Bayesian interpretation and enjoys several invariance properties, making it an appealing prior choice. We show that the proposed method leads to an accurate approximation of the reduced-bias approach of Firth (1993), resulting in estimators with smaller asymptotic bias than the maximum-likelihood and whose existence is always guaranteed. Moreover, the considered penalized likelihood can be expressed as a genuine likelihood, in which the original data are replaced with a collection of pseudo-counts. Hence, our approach may leverage well established and scalable algorithms for logistic regression. We compare our estimator with alternative reduced-bias methods, vastly improving their computational performance and achieving appealing inferential results.
翻译:统计实践中经常使用二元反应的后勤回归模型,然而,由于数据可分离,最大可能性估计可能不存在。我们通过考虑一致的先前处罚来解决这一问题,这种处罚总是产生有限的估计数。这种规格具有明确的巴伊西亚解释,具有若干差异性,因此它具有先选的吸引力。我们表明,拟议方法导致Firth(1993年)的降低偏差方法的准确近似,从而导致估计者对零食的偏差小于最大相似度,而且其存在总是得到保证。此外,被认为的受处罚的可能性可以作为一种真正的可能性来表达,在这种可能性中,原始数据被以收集的假数字取代。因此,我们的方法可以充分利用固定和可调整的逻辑回归算法。我们用其他降低偏差的方法比较我们的估算法,大大改进它们的计算性,并取得可上诉的推断结果。