Weak supervision learning on classification labels has demonstrated high performance in various tasks. When a few pixel-level fine annotations are also affordable, it is natural to leverage both of the pixel-level (e.g., segmentation) and image level (e.g., classification) annotation to further improve the performance. In computational pathology, however, such weak or mixed supervision learning is still a challenging task, since the high resolution of whole slide images makes it unattainable to perform end-to-end training of classification models. An alternative approach is to analyze such data by patch-base model training, i.e., using self-supervised learning to generate pixel-level pseudo labels for patches. However, such methods usually have model drifting issues, i.e., hard to converge, because the noise accumulates during the self-training process. To handle those problems, we propose a mixed supervision learning framework for super high-resolution images to effectively utilize their various labels (e.g., sufficient image-level coarse annotations and a few pixel-level fine labels). During the patch training stage, this framework can make use of coarse image-level labels to refine self-supervised learning and generate high-quality pixel-level pseudo labels. A comprehensive strategy is proposed to suppress pixel-level false positives and false negatives. Three real-world datasets with very large number of images (i.e., more than 10,000 whole slide images) and various types of labels are used to evaluate the effectiveness of mixed supervision learning. We reduced the false positive rate by around one third compared to state of the art while retaining 100% sensitivity, in the task of image-level classification.
翻译:在分类标签上薄弱的监督学习显示,在各种任务中表现优异。 当一些像素级的微调说明也负担得起时,自然会利用像素级(例如分解)和图像级(例如分类)的注释来进一步改进性能。但在计算病理学中,这种薄弱或混合的监督学习仍是一项艰巨的任务,因为整个幻灯片图像的高分辨率使得无法对分类模型进行端到端的培训。另一种办法是通过补丁基的幻灯片模型培训来分析这些数据,即使用自我监督的学习类型来为补丁级图像生成像素级的假标签(例如分解)和图像级(例如分类),但这种方法通常会模拟漂移问题,即难以趋同,因为在自我培训过程中,噪音会累积。为了处理这些问题,我们建议为超级高分辨率图像制作一个混合的监督学习框架,以便有效地利用它们的各种标签(例如,足够的图像级的分解图和几平级级级的微级的图像级图像级,以及比高级的级级级级级的图像级级的标签) 。在修补制阶段,我们使用这个框架可以对高级的自我学习过程进行更精确的自我级的升级的升级,而要使用。