With the increasing adoption of language models in applications involving sensitive data, it has become crucial to protect these models from leaking private information. Previous work has attempted to tackle this challenge by training RNN-based language models with differential privacy guarantees. However, applying classical differential privacy to language models leads to poor model performance as the underlying privacy notion is over-pessimistic and provides undifferentiated protection for all tokens of the data. Given that the private information in natural language is sparse (for example, the bulk of an email might not carry personally identifiable information), we propose a new privacy notion, selective differential privacy, to provide rigorous privacy guarantees on the sensitive portion of the data to improve model utility. To realize such a new notion, we develop a corresponding privacy mechanism, Selective-DPSGD, for RNN-based language models. Besides language modeling, we also apply the method to a more concrete application -- dialog systems. Experiments on both language modeling and dialog system building show that the proposed privacy-preserving mechanism achieves better utilities while remaining safe under various privacy attacks compared to the baselines. The data, code and models are available at https://github.com/wyshi/lm_privacy.
翻译:随着在涉及敏感数据的应用程序中越来越多地采用语言模型,保护这些模型不受泄露私人信息的影响变得至关重要。以前的工作试图通过培训基于区域网的基于区域网的语文模型,提供不同的隐私保障,以克服这一挑战。然而,对语言模型适用古典差异隐私权,导致典型的功能不佳,因为基本隐私概念过于悲观,并为数据的所有象征物提供不加区分的保护。鉴于自然语言的私人信息很少(例如,大部分电子邮件可能不包含个人可识别的信息),我们提出了一个新的隐私概念,有选择的差别隐私,以便为数据敏感部分提供严格的隐私保障,以改善模型的实用性。为了实现这样一个新概念,我们为基于区域网的语文模型开发了相应的隐私机制,即Speepive-DPSGD。除了语言模型外,我们还将这一方法应用于更具体的应用程序 -- -- 对话系统。关于语言模型和对话系统建设的实验表明,拟议的隐私保护机制在各种隐私攻击下比基线更安全,同时保持安全。数据、代码和模型可在 https://githbub.privac.com/wshim/mm.