Recent advances in deep learning and computer vision have reduced many barriers to automated medical image analysis, allowing algorithms to process label-free images and improve performance. However, existing techniques have extreme computational requirements and drop a lot of performance with a reduction in batch size or training epochs. This paper presents Cross Architectural - Self Supervision (CASS), a novel self-supervised learning approach that leverages Transformer and CNN simultaneously. Compared to the existing state of the art self-supervised learning approaches, we empirically show that CASS-trained CNNs and Transformers across four diverse datasets gained an average of 3.8% with 1% labeled data, 5.9% with 10% labeled data, and 10.13% with 100% labeled data while taking 69% less time. We also show that CASS is much more robust to changes in batch size and training epochs. Notably, one of the test datasets comprised histopathology slides of an autoimmune disease, a condition with minimal data that has been underrepresented in medical imaging. The code is open source and is available on GitHub.
翻译:最近深层次学习和计算机愿景的进展减少了自动化医学图像分析的许多障碍,使得算法能够处理无标签图像和改善性能。然而,现有技术具有极端的计算要求,并且随着批量规模或培训规模的减少而大量下降。本文展示了跨建筑-自我监督(CASS),这是一种利用变异器和CNN的新型自我监督学习方法。与现代自监督学习方法的现有状态相比,我们从经验上表明,经过CASS培训的有线电视新闻网和四个不同数据集的变换器获得的平均3.8%,有1%的标签数据,5.9%的标签数据,10.13%的标签数据,减少了69%的时间。我们还表明,CASS对批量规模和培训包的改变具有更大的活力。值得注意的是,其中一个测试数据集是由自动免疫系统疾病的病理学幻灯片构成的,这是医学成象素中代表性极小的一个条件。代码是开放源,可在GitHub获得。