Countless applications depend on accurate predictions with reliable confidence estimates from modern object detectors. It is well known, however, that neural networks including object detectors produce miscalibrated confidence estimates. Recent work even suggests that detectors' confidence predictions are biased with respect to object size and position, but it is still unclear how this bias relates to the performance of the affected object detectors. We formally prove that the conditional confidence bias is harming the expected performance of object detectors and empirically validate these findings. Specifically, we demonstrate how to modify the histogram binning calibration to not only avoid performance impairment but also improve performance through conditional confidence calibration. We further find that the confidence bias is also present in detections generated on the training data of the detector, which we leverage to perform our de-biasing without using additional data. Moreover, Test Time Augmentation magnifies this bias, which results in even larger performance gains from our calibration method. Finally, we validate our findings on a diverse set of object detection architectures and show improvements of up to 0.6 mAP and 0.8 mAP50 without extra data or training.
翻译:然而,众所周知,包括物体探测器在内的神经网络会得出错误的置信估计。最近的工作甚至表明,探测器的置信预测在物体大小和位置上存在偏差,但这种偏差与受影响物体探测器的性能有何关系,目前还不清楚。我们正式证明,有条件的置信偏差正在损害物体探测器的预期性能,并从经验上证实这些结果。具体地说,我们展示了如何修改直方图的硬化校准,不仅避免了性能受损,而且通过有条件的置信校准提高了性能。我们进一步发现,在探测探测器的培训数据时也存在信任偏差,我们利用这些数据进行分辨偏差,而不使用其他数据。此外,试验时间放大了这种偏差,这导致我们校准方法产生更大的性能收益。最后,我们验证了我们关于各种物体探测结构的调查结果,并显示在没有额外数据或培训的情况下改进了0.6 mAP和0.8 mAP50。