Cross-modality fusing complementary information of multispectral remote sensing image pairs can improve the perception ability of detection algorithms, making them more robust and reliable for a wider range of applications, such as nighttime detection. Compared with prior methods, we think different features should be processed specifically, the modality-specific features should be retained and enhanced, while the modality-shared features should be cherry-picked from the RGB and thermal IR modalities. Following this idea, a novel and lightweight multispectral feature fusion approach with joint common-modality and differential-modality attentions are proposed, named Cross-Modality Attentive Feature Fusion (CMAFF). Given the intermediate feature maps of RGB and IR images, our module parallel infers attention maps from two separate modalities, common- and differential-modality, then the attention maps are multiplied to the input feature map respectively for adaptive feature enhancement or selection. Extensive experiments demonstrate that our proposed approach can achieve the state-of-the-art performance at a low computation cost.
翻译:多光谱遥感成像配对的交叉模式化补充信息可以提高探测算法的感知能力,使其在夜间探测等更广泛的应用中更加强大和可靠。与以前的方法相比,我们认为应具体处理不同特征,应当保留和加强模式特有特征,而模式共享特征应从RGB和热IR模式中摘取樱桃式的。根据这一想法,建议采用新型和轻度多光谱特征融合法,采用共同模式和差异模式联合关注方式,称为跨模式高度特征聚合法(CMAFF)。鉴于RGB和IR图像的中间特征图,我们的模块平行地从两种不同的模式,即共同模式和差异模式中引出关注地图,然后将关注地图乘以投入特征图,分别用于适应性地貌的增强或选择。广泛的实验表明,我们拟议的方法能够以低计算成本实现最先进的性能。