Medical image segmentation has made significant progress in recent years. Deep learning-based methods are recognized as data-hungry techniques, requiring large amounts of data with manual annotations. However, manual annotation is expensive in the field of medical image analysis, which requires domain-specific expertise. To address this challenge, few-shot learning has the potential to learn new classes from only a few examples. In this work, we propose a novel framework for few-shot medical image segmentation, termed CAT-Net, based on cross masked attention Transformer. Our proposed network mines the correlations between the support image and query image, limiting them to focus only on useful foreground information and boosting the representation capacity of both the support prototype and query features. We further design an iterative refinement framework that refines the query image segmentation iteratively and promotes the support feature in turn. We validated the proposed method on three public datasets: Abd-CT, Abd-MRI, and Card-MRI. Experimental results demonstrate the superior performance of our method compared to state-of-the-art methods and the effectiveness of each component. we will release the source codes of our method upon acceptance.
翻译:医学图像分割在近年来取得了重大进展。基于深度学习的方法被认为是数据饥饿的技术,需要大量带有手动标注的数据。然而,医学图像分析领域中的手动注释是一项昂贵的工作,需要特定领域的专业知识。为应对这一挑战,Few Shot 学习有能力从仅有少量实例中学习新类别。本文提出了一种新颖的 Few Shot 医学图像分割框架,称为 CAT-Net,基于交叉掩膜注意力变换器。我们的提议网络挖掘了支持图像和查询图像之间的相关性,将它们限制在仅关注有用的前景信息上,提高了支持原型和查询特征的表示能力。我们进一步设计了一个迭代的改进框架,迭代地细化查询图像分割,反过来促进支持特征。我们在三个公共数据集(Abd-CT,Abd-MRI 和 Card-MRI)上验证了所提出的方法。实验结果表明,与最新的方法相比,我们的方法具有优异的性能,每个组件的有效性也得到了证实。经过接受,我们将发布我们方法的源代码。