Cross-entropy loss and focal loss are the most common choices when training deep neural networks for classification problems. Generally speaking, however, a good loss function can take on much more flexible forms, and should be tailored for different tasks and datasets. Motivated by how functions can be approximated via Taylor expansion, we propose a simple framework, named PolyLoss, to view and design loss functions as a linear combination of polynomial functions. Our PolyLoss allows the importance of different polynomial bases to be easily adjusted depending on the targeting tasks and datasets, while naturally subsuming the aforementioned cross-entropy loss and focal loss as special cases. Extensive experimental results show that the optimal choice within the PolyLoss is indeed dependent on the task and dataset. Simply by introducing one extra hyperparameter and adding one line of code, our Poly-1 formulation outperforms the cross-entropy loss and focal loss on 2D image classification, instance segmentation, object detection, and 3D object detection tasks, sometimes by a large margin.
翻译:在培训深神经网络处理分类问题时,跨作物损失和焦点损失是最常见的选择。但一般而言,良好的损失功能可以采取更灵活得多的形式,并适合不同的任务和数据集。我们提议一个简单的框架,即称为PolyLos,以将损失功能作为多元函数的线性组合来看待和设计。我们的多边实验允许根据目标任务和数据集来轻而易举地调整不同多球基础的重要性,同时自然地将上述跨作物损失和焦点损失作为特例进行分解。广泛的实验结果显示,在多边试验中,最佳选择的确是取决于任务和数据集。我们采用一个额外的超分量计和增加一行代码,我们的聚一配方体的配方在2D图像分类、例分解、物体探测和3D对象探测任务中,有时以较大幅度来完成跨作物损失和焦点损失。