Detection and localization of fire in images and videos are important in tackling fire incidents. Although semantic segmentation methods can be used to indicate the location of pixels with fire in the images, their predictions are localized, and they often fail to consider global information of the existence of fire in the image which is implicit in the image labels. We propose a Convolutional Neural Network (CNN) for joint classification and segmentation of fire in images which improves the performance of the fire segmentation. We use a spatial self-attention mechanism to capture long-range dependency between pixels, and a new channel attention module which uses the classification probability as an attention weight. The network is jointly trained for both segmentation and classification, leading to improvement in the performance of the single-task image segmentation methods, and the previous methods proposed for fire segmentation.
翻译:图像和视频中火灾的探测和定位对于处理火灾事件很重要。虽然可以使用语义分解方法来显示图像中火灾的像素位置,但其预测是局部的,而且往往不考虑图像标签中隐含的图像中火灾存在的全球信息。我们提议建立一个动态神经网络,对图像中的火灾进行联合分类和分解,以提高火灾分解的性能。我们使用空间自省机制来捕捉像素之间的远距离依赖性,以及使用分类概率作为关注重量的新频道关注模块。网络在分解和分类方面都得到了联合培训,从而改进了单一任务图像分解方法的性能,以及先前提出的火分解方法。