A variety of techniques have been proposed to train machine learning classifiers that are independent of a given feature. While this can be an essential technique for enabling background estimation, it may also be useful for reducing uncertainties. We carefully examine theory uncertainties, which typically do not have a statistical origin. We will provide explicit examples of two-point (fragmentation modeling) and continuous (higher-order corrections) uncertainties where decorrelating significantly reduces the apparent uncertainty while the actual uncertainty is much larger. These results suggest that caution should be taken when using decorrelation for these types of uncertainties as long as we do not have a complete decomposition into statistically meaningful components.
翻译:已经提出各种技术来培训独立于某一特性的机器学习分类师。虽然这可能是有利于进行背景估计的必要技术,但也可能有助于减少不确定性。我们仔细研究通常没有统计来源的理论不确定性。我们将提供两点(不成体系建模)和连续(高阶校正)不确定性的明确例子,在这些例子中,与装饰有关的因素大大减少了明显的不确定性,而实际不确定性则大得多。这些结果表明,只要我们没有完全分解成具有统计意义的部分,在对这些类型的不确定性进行设计时,应当谨慎行事。