Deep neural networks are increasingly deployed for scene analytics, including to evaluate the attention and reaction of people exposed to out-of-home advertisements. However, the features extracted by a deep neural network that was trained to predict a specific, consensual attribute (e.g. emotion) may also encode and thus reveal information about private, protected attributes (e.g. age or gender). In this work, we focus on such leakage of private information at inference time. We consider an adversary with access to the features extracted by the layers of a deployed neural network and use these features to predict private attributes. To prevent the success of such an attack, we modify the training of the network using a confusion loss that encourages the extraction of features that make it difficult for the adversary to accurately predict private attributes. We validate this training approach on image-based tasks using a publicly available dataset. Results show that, compared to the original network, the proposed PrivateNet can reduce the leakage of private information of a state-of-the-art emotion recognition classifier by 2.88% for gender and by 13.06% for age group, with a minimal effect on task accuracy.
翻译:深度神经网络越来越多地用于现场分析,包括评估受家庭外广告影响的人的关注和反应;然而,深神经网络的特征,经过培训,可以预测具体的、双方同意的属性(例如情感),其特征也可能编码,从而披露关于私人受保护属性(例如年龄或性别)的信息;在这项工作中,我们侧重于在推论时间这种私人信息的泄漏;我们认为,可以接触由部署神经网络层提取的特征的对手,并利用这些特征预测私人属性;为防止这种袭击的成功,我们用混乱损失来修改网络的培训,鼓励提取使对手难以准确预测私人属性的特征;我们利用公开的数据集验证这种基于图像的任务的培训方法;结果显示,与原始网络相比,拟议的私人网络可以将最新情感识别分类器的私人信息的泄漏减少2.88%,将性别的私人信息减少13.06%,对任务准确性影响最小。