Recent work has investigated the distributions of learned convolution filters through a large-scale study containing hundreds of heterogeneous image models. Surprisingly, on average, the distributions only show minor drifts in comparisons of various studied dimensions including the learned task, image domain, or dataset. However, among the studied image domains, medical imaging models appeared to show significant outliers through "spikey" distributions, and, therefore, learn clusters of highly specific filters different from other domains. Following this observation, we study the collected medical imaging models in more detail. We show that instead of fundamental differences, the outliers are due to specific processing in some architectures. Quite the contrary, for standardized architectures, we find that models trained on medical data do not significantly differ in their filter distributions from similar architectures trained on data from other domains. Our conclusions reinforce previous hypotheses stating that pre-training of imaging models can be done with any kind of diverse image data.
翻译:最近的工作通过一项包含数百个不同图像模型的大规模研究,调查了所学的革命过滤器的分布情况。 令人惊讶的是,平均而言,分布只显示在比较各种研究层面(包括学习的任务、图像域或数据集)方面的细微变化。 但在研究的图像域中,医学成像模型似乎通过“spigey”分布显示显著的外差,因此,通过学习与其它领域不同的高度特定过滤器群落。在观察之后,我们更详细地研究所收集的医疗成像模型。我们发现,与基本差异相比,外差是由于某些结构中的具体处理。相反,对于标准化结构而言,我们发现,在医学数据方面受过培训的模型在过滤分布方面与其他领域数据受过训练的类似结构没有很大区别。我们的结论强化了先前的假设,即对成像模型的训练前训练可以使用任何不同的图像数据完成。