Convolutional neural networks learned by minimizing the cross-entropy loss are nowadays the standard for image classification. Till now, the statistical theory behind those networks is lacking. We analyze the rate of convergence of the misclassification risk of the estimates towards the optimal misclassification risk. Under suitable assumptions on the smoothness and structure of the aposteriori probability it is shown that these estimates achieve a rate of convergence which is independent of the dimension of the image. The study shed light on the good performance of CNNs learned by cross-entropy loss and partly explains their success in practical applications.
翻译:通过尽量减少跨热带损失而学得的进化神经网络如今已成为图像分类的标准。到目前为止,这些网络背后的统计理论尚缺乏。我们分析了估算错误分类风险与最佳分类错误风险的趋同率。根据关于异种概率平滑和结构的适当假设,可以证明这些估算达到了与图像层面无关的趋同率。这项研究揭示了CNN通过跨热带损失而学得的良好表现,并部分解释了其在实际应用方面的成功。