Deep learning models are increasingly deployed in real-world applications. These models are often deployed on the server-side and receive user data in an information-rich representation to solve a specific task, such as image classification. Since images can contain sensitive information, which users might not be willing to share, privacy protection becomes increasingly important. Adversarial Representation Learning (ARL) is a common approach to train an encoder that runs on the client-side and obfuscates an image. It is assumed, that the obfuscated image can safely be transmitted and used for the task on the server without privacy concerns. However, in this work, we find that training a reconstruction attacker can successfully recover the original image of existing ARL methods. To this end, we introduce a novel ARL method enhanced through low-pass filtering, limiting the available information amount to be encoded in the frequency domain. Our experimental results reveal that our approach withstands reconstruction attacks while outperforming previous state-of-the-art methods regarding the privacy-utility trade-off. We further conduct a user study to qualitatively assess our defense of the reconstruction attack.
翻译:深度学习模型越来越多地被应用到现实世界的应用中。 这些模型往往被安装在服务器上,接收用户数据,在信息丰富的演示中解决特定任务,例如图像分类。 由于图像可能包含敏感信息,用户可能不愿意分享,隐私保护变得日益重要。 反向代表学习(ARL)是培训一个运行在客户端的编码器和模糊图像的常见方法。 假设的是, 模糊的图像可以安全地传输并用于服务器上的任务, 而不考虑隐私问题。 然而, 在这项工作中, 我们发现, 培训一个重建攻击者可以成功恢复现有ARL方法的原始图像。 为此, 我们引入了一种新的ARL方法, 通过低途过滤, 限制现有信息在频率域的编码。 我们的实验结果显示, 我们的方法在超过先前的关于隐私效用交易的状态方法的同时, 能够抵御重建攻击。 我们还进行了用户研究, 以便从质量上评估我们对重建攻击的防御。