A key promise of machine learning is the ability to assist users with personal tasks. Because the personal context required to make accurate predictions is often sensitive, we require systems that protect privacy. A gold standard privacy-preserving system will satisfy perfect secrecy, meaning that interactions with the system provably reveal no additional private information to adversaries. This guarantee should hold even as we perform multiple personal tasks over the same underlying data. However, privacy and quality appear to be in tension in existing systems for personal tasks. Neural models typically require lots of training to perform well, while individual users typically hold a limited scale of data, so the systems propose to learn from the aggregate data of multiple users. This violates perfect secrecy and instead, in the last few years, academics have defended these solutions using statistical notions of privacy -- i.e., the probability of learning private information about a user should be reasonably low. Given the vulnerabilities of these solutions, we explore whether the strong perfect secrecy guarantee can be achieved using recent zero-to-few sample adaptation techniques enabled by foundation models. In response, we propose FOCUS, a framework for personal tasks. Evaluating on popular privacy benchmarks, we find the approach, satisfying perfect secrecy, competes with strong collaborative learning baselines on 6 of 7 tasks. We empirically analyze the proposal, highlighting the opportunities and limitations across task types, and model inductive biases and sizes.
翻译:机器学习的关键承诺是帮助用户完成个人任务的能力。由于准确预测所需的个人环境往往很敏感,我们需要保护隐私的系统。金质标准隐私保护系统将满足完全保密,这意味着与系统的互动可能不会向对手透露额外的私人信息。即使我们对同一基本数据执行多重个人任务,这种保证也应保持。然而,隐私和质量似乎在现有个人任务系统中处于紧张状态。神经模型通常需要大量的培训才能很好地运作,而个人用户通常拥有有限的数据规模,因此系统建议从多个用户的总数据中学习。这违反了完美的保密原则,相反,在过去几年里,学术界利用隐私统计概念来捍卫这些解决方案 -- -- 也就是说,学习关于用户的私人信息的可能性应该相当低。鉴于这些解决方案的脆弱性,我们探讨能否利用基础模型所促成的最近的零到零的样本适应技术实现强有力的保密保证。我们提出FOCUS,一个个人任务框架。评估大众隐私基准,我们发现方法,满足完美的保密要求,并用合作性基准衡量任务的规模,我们用合作性基准来竞争。