Fire localization in images and videos is an important step for an autonomous system to combat fire incidents. State-of-art image segmentation methods based on deep neural networks require a large number of pixel-annotated samples to train Convolutional Neural Networks (CNNs) in a fully-supervised manner. In this paper, we consider weakly supervised segmentation of fire in images, in which only image labels are used to train the network. We show that in the case of fire segmentation, which is a binary segmentation problem, the mean value of features in a mid-layer of classification CNN can perform better than conventional Class Activation Mapping (CAM) method. We also propose to further improve the segmentation accuracy by adding a rotation equivariant regularization loss on the features of the last convolutional layer. Our results show noticeable improvements over baseline method for weakly-supervised fire segmentation.
翻译:图像和视频中的火灾定位是应对火灾事件自主系统的一个重要步骤。基于深神经网络的先进图像分解方法需要大量像素加注样本,以在完全监督下培训进化神经网络。本文认为,图像中的火灾分解监管薄弱,其中仅使用图像标签来培训网络。我们表明,在火灾分解(即二元分解问题)的情况下,CNN分类中层的特征平均值比常规的分类活动映射(CAM)方法要好。我们还建议通过在最后一个进化层的特征上增加旋转的等离子调整损失来进一步提高分解的准确性。我们的结果显示,在弱监控消防分解的基线方法方面有明显改进。