Training deep neural networks via federated learning allows clients to share the model updated on their data, instead of the original data. In practice, it is shown that a client's private information, unrelated to the main learning task, can be discovered from the shared model, which compromises the promised privacy protection. However, there is still no formal approach for quantifying the leakage of such latent information from the shared model/gradients. As a solution, we introduce and evaluate two mathematically-grounded metrics for better characterizing the amount of information included in the shared gradients computed on the clients' private data. First, using an adaptation of the empirical $\mathcal{V}$-information, we show how to quantify the amount of private latent information captured in gradients that are usable for an attacker. Second, based on a sensitivity analysis}using Jacobian matrices, we show how to measure changes in the gradients with respect to latent information. Further, we show the applicability of our proposed metrics in (i) localizing private latent information in a layer-wise manner, in both settings where we have or we do not have the knowledge of the attackers' capability, and (ii) comparing the capacity of each layer in a neural network in capturing higher-level versus lower-level latent information. Experimental results on three real-world datasets using three benchmark models show the validity of the proposed metrics.
翻译:通过联合会式学习进行深层培训神经网络,使客户能够分享其数据更新的模型,而不是原始数据。在实践中,显示客户的私人信息与主要学习任务无关,可以从共享模型中发现,这有损于所承诺的隐私保护;然而,目前还没有正式的方法来量化共享模型/梯度中此类潜在信息渗漏。作为一种解决办法,我们引入和评估两个数学基础的衡量标准,以便更好地说明根据客户私人数据计算的共同梯度中所含信息的数量。首先,利用经验$mathcal{V}信息,我们展示如何量化用于攻击者的梯度中采集的私人潜在信息数量。第二,根据敏感性分析,我们使用雅各布矩阵,展示如何衡量与隐性信息有关的梯度变化。此外,我们展示了我们所提议的衡量标准的适用性,在(一) 以层方式将私人潜在信息本地化,在两种环境中,我们没有或我们没有使用攻击者网络的三个水平的真实性水平,在三个水平上,在实验性水平上,对比每个水平上,对攻击者基准度能力水平上的真实性水平上的数据。