Recently, there has been an increasing concern about the privacy issue raised by identifiable information in machine learning. However, previous portrait matting methods were all based on identifiable images. To fill the gap, we present P3M-10k, which is the first large-scale anonymized benchmark for Privacy-Preserving Portrait Matting (P3M). P3M-10k consists of 10,421 high resolution face-blurred portrait images along with high-quality alpha mattes, which enables us to systematically evaluate both trimap-free and trimap-based matting methods and obtain some useful findings about model generalization ability under the privacy preserving training (PPT) setting. We also present a unified matting model dubbed P3M-Net that is compatible with both CNN and transformer backbones. To further mitigate the cross-domain performance gap issue under the PPT setting, we devise a simple yet effective Copy and Paste strategy (P3M-CP), which borrows facial information from public celebrity images and directs the network to reacquire the face context at both data and feature level. Extensive experiments on P3M-10k and public benchmarks demonstrate the superiority of P3M-Net over state-of-the-art methods and the effectiveness of P3M-CP in improving the cross-domain generalization ability, implying a great significance of P3M for future research and real-world applications.
翻译:最近,人们对可识别信息在机器学习中引发的隐私问题越来越关注。然而,以前的肖像抠图方法都是基于可识别的图像。为了填补这一空白,我们提出了P3M-10k,这是首个大规模匿名化的隐私保护肖像抠图基准(P3M)。P3M-10k包括10,421个高分辨率面部模糊的肖像图像和高质量的阿尔法遮罩,使我们能够系统地评估无三分图和有三分图的抠图方法,并获得有关在隐私保护训练(PPT)设置下模型泛化能力的一些有用发现。我们还提出了一个统一的抠图模型,称为P3M-Net,它兼容CNN和transformer背骨。为了进一步减轻在PPT设置下跨域性能差距的问题,我们设计了一种简单而有效的Copy and Paste策略(P3M-CP),它从公共名人图像中借用面部信息,并引导网络在数据和特征层面重新获取面部上下文。对P3M-10k和公共基准的广泛实验表明,P3M-Net优于现有的最先进方法,而P3M-CP的有效性在提高跨域泛化能力方面也得到证明,这意味着P3M在未来的研究和实际应用中具有重要意义。