In supervised machine learning, the choice of loss function implicitly assumes a particular noise distribution over the data. For example, the frequently used mean squared error (MSE) loss assumes a Gaussian noise distribution. The choice of loss function during training and testing affects the performance of artificial neural networks (ANNs). It is known that MSE may yield substandard performance in the presence of outliers. The Cauchy loss function (CLF) assumes a Cauchy noise distribution, and is therefore potentially better suited for data with outliers. This papers aims to determine the extent of robustness and generalisability of the CLF as compared to MSE. CLF and MSE are assessed on a few handcrafted regression problems, and a real-world regression problem with artificially simulated outliers, in the context of ANN training. CLF yielded results that were either comparable to or better than the results yielded by MSE, with a few notable exceptions.
翻译:在受监督的机器学习中,损失选择功能暗含地假定数据有特定的噪音分布。例如,常用的平均平方错误(MSE)损失假定高斯噪音分布。在培训和测试中选择损失功能会影响人工神经网络(ANNS)的性能。众所周知,MSE在外派人员在场的情况下可能会产生低于标准的性能。Cawy损失功能(CLF)假定是粗野的噪音分布,因此可能更适合外派人员的数据。这份文件旨在确定与MSE相比,CLF的稳健性和可概括性的程度。CLF和MSE是在几个手工制造的回归问题中评估的,在ANN培训中,与人工模拟的外科人员一起评估了一个真实世界性回归问题。CLF的结果与MSE的结果相似或优于MSE产生的结果,但有一些显著的例外。