With the increasing applications of language models, it has become crucial to protect these models from leaking private information. Previous work has attempted to tackle this challenge by training RNN-based language models with differential privacy guarantees. However, applying classical differential privacy to language models leads to poor model performance as the underlying privacy notion is over-pessimistic and provides undifferentiated protection for all tokens in the data. Given that the private information in natural language is sparse (for example, the bulk of an email might not carry personally identifiable information), we propose a new privacy notion, selective differential privacy, to provide rigorous privacy guarantees on the sensitive portion of the data to improve model utility. To realize such a new notion, we develop a corresponding privacy mechanism, Selective-DPSGD, for RNN-based language models. Besides language modeling, we also apply the method to a more concrete application--dialog systems. Experiments on both language modeling and dialog system building show that the proposed privacy-preserving mechanism achieves better utilities while remaining safe under various privacy attacks compared to the baselines. The data and code are released at https://github.com/wyshi/lm_privacy to facilitate future research .
翻译:随着语言模型应用的增加,保护这些模型不受私密信息泄漏至关重要。以前的工作试图通过培训基于区域网的基于区域网的语文模型,以不同的隐私保障来应对这一挑战。然而,对语言模型适用古典差异性隐私,导致典型表现不佳,因为基本隐私概念过于悲观,为数据中的所有象征物提供不加区分的保护。鉴于自然语言的私人信息很少(例如,大部分电子邮件可能不包含个人可识别的信息),我们提出一个新的隐私概念,有选择的差别隐私,为数据的敏感部分提供严格的隐私保障,以改善模型的效用。为了实现这样一个新概念,我们为基于区域网的语文模型开发了相应的隐私机制,即Speepive-DPSGD。除了语言模型外,我们还将这一方法应用于更具体的应用程序-对像系统。对语言模型和对话系统建设的实验表明,拟议的隐私保护机制在与基线相比的各种隐私攻击下可以实现更好的公用事业。数据和代码在https://github.com/wyshi/privacyvary上发布,以便利未来研究。