Results from randomized experiments (trials) can be severely distorted by outcome misclassification, such as from measurement error or reporting bias in binary outcomes. All existing approaches to outcome misclassification rely on some data-generating (super-population) model and therefore may not be applicable to randomized experiments without additional assumptions. We propose a model-free and finite-population-exact framework for randomized experiments subject to outcome misclassification. A central quantity in our framework is "warning accuracy," defined as the threshold such that the causal conclusion drawn from the measured outcomes may differ from that based on the true outcomes if the outcome measurement accuracy did not surpass that threshold. We show how learning the warning accuracy and related concepts can benefit a randomized experiment subject to outcome misclassification. We show that the warning accuracy can be computed efficiently (even for large datasets) by adaptively reformulating an integer program with respect to the randomization design. Our framework covers both Fisher's sharp null and Neyman's weak null, works for a wide range of randomization designs, and can also be applied to observational studies adopting randomization-based inference. We apply our framework to a large randomized clinical trial for the prevention of prostate cancer.
翻译:随机实验(实验)的结果可能因结果错误分类而严重扭曲,例如测量错误或报告二进制结果偏差。所有现有的结果错误分类方法都依赖于某些数据生成(超人口)模型,因此可能无法在没有额外假设的情况下适用于随机实验。我们为随机随机实验提出了一个不受结果错误分类影响的无模型和有限人口特征框架。我们框架中的一个核心数量是“警报准确性”,其定义是,从测量结果得出的因果结论可能不同于根据真实结果得出的临界值,如果测量结果的准确性没有超过这一临界值。我们表明,在结果错误分类的情况下,对警告准确性和相关概念的学习会给随机实验带来哪些好处。我们表明,通过对随机化设计进行适应性调整性重新组合程序,可以有效地计算警告准确性(即使是大型数据集)。我们的框架涵盖Fishercher的尖锐空空和Neyman的弱无效性,用于广泛的随机化设计,还可以应用于采用基于随机性推断的观察性研究。我们把我们的框架应用于大规模临床癌症预防的大规模随机试验。