Current state-of-the-art segmentation techniques for ocular images are critically dependent on large-scale annotated datasets, which are labor-intensive to gather and often raise privacy concerns. In this paper, we present a novel framework, called BiOcularGAN, capable of generating synthetic large-scale datasets of photorealistic (visible light and near-infrared) ocular images, together with corresponding segmentation labels to address these issues. At its core, the framework relies on a novel Dual-Branch StyleGAN2 (DB-StyleGAN2) model that facilitates bimodal image generation, and a Semantic Mask Generator (SMG) component that produces semantic annotations by exploiting latent features of the DB-StyleGAN2 model. We evaluate BiOcularGAN through extensive experiments across five diverse ocular datasets and analyze the effects of bimodal data generation on image quality and the produced annotations. Our experimental results show that BiOcularGAN is able to produce high-quality matching bimodal images and annotations (with minimal manual intervention) that can be used to train highly competitive (deep) segmentation models (in a privacy aware-manner) that perform well across multiple real-world datasets. The source code for the BiOcularGAN framework is publicly available at https://github.com/dariant/BiOcularGAN.
翻译:在本文中,我们提出了一个名为BiOOSALGAN(SMG)的新颖框架,它能够产生光现实(可见光和近红外)视觉图像的合成大型数据集,并配有相应的分层标签来解决这些问题。框架的核心是,它依赖于一个新型的双层-Branch StyleGAN2(DB-StyleGAN2)(DB-StyleGAN2)模型,该模型有助于双式图像生成,以及一个用于利用DB-StyleGAN2模型潜在特征生成语义说明的语义面具生成器(SMGMG)元件。我们通过五个不同的视觉数据集的广泛实验来评估双向GAN,并分析双向数据生成对图像质量和制作的注释的影响。我们的实验结果表明,BiOOSLGANAN能够制作高质量的匹配双向图像和说明(与最低限度的手动干预),而SMARMAMS(SMG)组件生成语义说明,通过DB-Stal-G-Stal-BAL Profileal Indeal commissional commission (O-AIal commission) commissional deal supulation the commissional commulational supulational supulational supulations) commilute.我们可以用来进行高竞争性的多种数据。