Heteroscedastic regression is the task of supervised learning where each label is subject to noise from a different distribution. This noise can be caused by the labelling process, and impacts negatively the performance of the learning algorithm as it violates the i.i.d. assumptions. In many situations however, the labelling process is able to estimate the variance of such distribution for each label, which can be used as an additional information to mitigate this impact. We adapt an inverse-variance weighted mean square error, based on the Gauss-Markov theorem, for parameter optimization on neural networks. We introduce Batch Inverse-Variance, a loss function which is robust to near-ground truth samples, and allows to control the effective learning rate. Our experimental results show that BIV improves significantly the performance of the networks on two noisy datasets, compared to L2 loss, inverse-variance weighting, as well as a filtering-based baseline.
翻译:电子回归是监督学习的任务,因为每个标签都受到不同分布的噪音的影响。这种噪音可能由标签过程引起,对学习算法的性能产生消极影响,因为它违反i.d.假设。然而,在许多情况下,标签过程能够估计每个标签的这种分布差异,可以用作减轻这种影响的额外信息。我们根据Gaus-Markov理论,为神经网络的参数优化调整了一个逆差加权平均方差。我们引入了批量反差功能,这一损失功能对近地真理样本是强大的,能够控制有效的学习率。我们的实验结果显示,BIV大大改善了两个噪音数据集网络的性能,与L2损失相比,逆差加权和基于过滤的基线。