Despite continued advancement in recent years, deep neural networks still rely on large amounts of training data to avoid overfitting. However, labeled training data for real-world applications such as healthcare is limited and difficult to access given longstanding privacy, and strict data sharing policies. By manipulating image datasets in the pixel or feature space, existing data augmentation techniques represent one of the effective ways to improve the quantity and diversity of training data. Here, we look to advance augmentation techniques by building upon the emerging success of text-to-image diffusion probabilistic models in augmenting the training samples of our macroscopic skin disease dataset. We do so by enabling fine-grained control of the image generation process via input text prompts. We demonstrate that this generative data augmentation approach successfully maintains a similar classification accuracy of the visual classifier even when trained on a fully synthetic skin disease dataset. Similar to recent applications of generative models, our study suggests that diffusion models are indeed effective in generating high-quality skin images that do not sacrifice the classifier performance, and can improve the augmentation of training datasets after curation.
翻译:尽管近年来不断进步,但深神经网络仍然依靠大量培训数据来避免过度配置,然而,由于长期隐私和严格的数据共享政策,贴有标签的保健等真实世界应用培训数据有限,难以获取。通过在像素或地貌空间操纵图像数据集,现有数据增强技术是提高培训数据数量和多样性的有效方法之一。在这里,我们期待在文本到图像传播概率模型不断成功的基础上推进增强技术,以增加我们大型皮肤疾病数据集的培训样本。我们这样做是因为能够通过输入文本提示对图像生成过程进行精细的精细控制。我们证明,这种基因化数据增强方法成功地保持了视觉分类的类似分类准确性,即使经过全面合成皮肤疾病数据集的培训。与最近应用的基因化模型类似,我们的研究显示,扩散模型确实有效地生成了不牺牲分类功能的高品质皮肤图像,并能在曲线后改进培训数据集的扩大。