Multiple Instance Learning (MIL) methods have become increasingly popular for classifying giga-pixel sized Whole-Slide Images (WSIs) in digital pathology. Most MIL methods operate at a single WSI magnification, by processing all the tissue patches. Such a formulation induces high computational requirements, and constrains the contextualization of the WSI-level representation to a single scale. A few MIL methods extend to multiple scales, but are computationally more demanding. In this paper, inspired by the pathological diagnostic process, we propose ZoomMIL, a method that learns to perform multi-level zooming in an end-to-end manner. ZoomMIL builds WSI representations by aggregating tissue-context information from multiple magnifications. The proposed method outperforms the state-of-the-art MIL methods in WSI classification on two large datasets, while significantly reducing the computational demands with regard to Floating-Point Operations (FLOPs) and processing time by up to 40x.
翻译:在数字病理学中对千兆像素大小整流图像进行分类时,多例学习方法越来越受欢迎。大多数MIL方法通过处理所有组织补丁,在单一的WSI放大法中运作。这种配方会产生很高的计算要求,并将WSI层次代表的上下文限制为单一规模。一些MIL方法扩展到多个尺度,但计算上的要求更高。在病理诊断过程的启发下,我们建议ZUMMIL,这是一种学习以端到端方式进行多级缩放的方法。ZUMMIL通过汇总多个放大法的组织文本信息来构建WSI的表示。拟议方法比WSI分类中两个大型数据集的现代MIL方法更符合当前最先进的MIL方法,同时显著地减少了与浮点操作(FLOPs)和处理时间有关的计算要求,最多可达40x。