In conventional split learning, training and testing data often face severe privacy leakage threats. Existing solutions often have to trade learning accuracy for data privacy, or the other way around. We propose a lossless privacy-preserving split learning framework, on the basis of the permutation equivalence properties which are inherent to many neural network modules. We adopt Transformer as the example building block to the framework. It is proved that the Transformer encoder block is permutation equivalent, and thus training/testing could be done equivalently on permuted data. We further introduce shuffling-based privacy guarantee and enhance it by mix-up training. All properties are verified by conducted experiments, which also show strong defence against privacy attacks compared to the state-of-the-art, yet without any accuracy decline.
翻译:在传统的分割学习中,训练和测试数据经常面临严重的隐私泄漏威胁。现有的解决方案常常必须在学习精度和数据隐私之间进行权衡。我们提出了一种无损隐私保护的分割学习框架,基于许多神经网络模块固有的排列等价性质。我们以Transformer为例来构建这个框架。证明了Transformer编码器块是置换等价的,因此能够在置换数据上等效地进行训练/测试。我们进一步引入了基于洗牌的隐私保证,并通过混合训练加以增强。所有属性都通过进行的实验进行了验证,这些实验还表明与现有技术相比,它具有强大的防御隐私攻击的能力,而且没有任何精度下降。