Training computer-vision related algorithms on medical images for disease diagnosis or image segmentation is difficult in large part due to privacy concerns. For this reason, generative image models are highly sought after to facilitate data sharing. However, 3-D generative models are understudied, and investigation of their privacy leakage is needed. We introduce our 3-D generative model, Transversal GAN (TrGAN), using head & neck PET images which are conditioned on tumour masks as a case study. We define quantitative measures of image fidelity, utility and privacy for our model. These metrics are evaluated in the course of training to identify ideal fidelity, utility and privacy trade-offs and establish the relationships between these parameters. We show that the discriminator of the TrGAN is vulnerable to attack, and that an attacker can identify which samples were used in training with almost perfect accuracy (AUC = 0.99). We also show that an attacker with access to only the generator cannot reliably classify whether a sample had been used for training (AUC = 0.51). This suggests that TrGAN generators, but not discriminators, may be used for sharing synthetic 3-D PET data with minimal privacy risk while maintaining good utility and fidelity.
翻译:对用于疾病诊断或图像分化的医疗图像进行计算机-视觉培训的计算机相关算法培训很困难,这在很大程度上是由于隐私问题。为此原因,为了便利数据共享,人们在大力寻求基因图像模型,以便利数据共享。然而,对3种基因模型的研究不足,需要调查其隐私渗漏。我们引入了我们的3种基因模型,即Transversal GAN(TrGAN),使用以肿瘤遮盖为条件的头部和颈部PET图像作为案例研究。我们定义了图像忠诚、实用性和隐私的定量计量。我们确定了模型的模型。在培训过程中对这些模型进行了评估,以确定理想的忠诚、实用性和隐私的权衡,并确定了这些参数之间的关系。我们表明,TrGAN的区别性模型很容易受到攻击,而且攻击者可以确定在培训中使用了几乎完全准确的样本(AUSC= 0.99),我们还表明,仅接触发电机的攻击者无法可靠地分类是否使用了样本用于培训(AUSC=0.51)。这说明,TrGAN发电机的发电机发电机,但不是歧视性生成者,同时可以与保密性数据共享最低风险。