We propose a new information aggregation method which called Localized Feature Aggregation Module based on the similarity between the feature maps of an encoder and a decoder. The proposed method recovers positional information by emphasizing the similarity between decoder's feature maps with superior semantic information and encoder's feature maps with superior positional information. The proposed method can learn positional information more efficiently than conventional concatenation in the U-net and attention U-net. Additionally, the proposed method also uses localized attention range to reduce the computational cost. Two innovations contributed to improve the segmentation accuracy with lower computational cost. By experiments on the Drosophila cell image dataset and COVID-19 image dataset, we confirmed that our method outperformed conventional methods.
翻译:我们建议一种新的信息汇总方法,即基于编码器和解码器特征图相似性的本地化地物聚合模块。拟议方法恢复定位信息,强调解码器特征图与高级语义信息相似性,而编码器特征图与高级定位信息相似性。拟议方法可以比U-net和注意力U-net更高效地学习定位信息。此外,拟议方法还利用本地关注范围减少计算成本。两项创新有助于以较低的计算成本提高分解准确性。通过对德罗索菲拉细胞图像数据集和COVID-19图像数据集的实验,我们确认我们的方法优于常规方法。