The collection and availability of big data, combined with advances in pre-trained models (e.g. BERT), have revolutionized the predictive performance of natural language processing tasks. This allows corporations to provide machine learning as a service (MLaaS) by encapsulating fine-tuned BERT-based models as APIs. Due to significant commercial interest, there has been a surge of attempts to steal re mote services via model extraction. Although previous works have made progress in defending against model extraction attacks, there has been little discussion on their performance in preventing privacy leakage. This work bridges this gap by launching an attribute inference attack against the extracted BERT model. Our extensive experiments reveal that model extraction can cause severe privacy leakage even when victim models are facilitated with advanced defensive strategies.
翻译:海量数据的收集和提供,加上预先培训的模型(如BERT)的进步,使自然语言处理任务的预测性表现发生了革命性的变化,使公司能够通过将基于BERT的模型精细调整的模型包装成API而提供机器学习服务(MLAAS)。由于商业上的巨大兴趣,有人企图通过模型提取偷取重新移动服务。虽然以前的工作在防范模型提取攻击方面取得了进展,但很少讨论其在防止隐私泄漏方面的表现。这项工作通过对提取出来的BERT模型发起属性推断攻击来弥补这一差距。我们的广泛实验显示,即使受害者模型得到先进的防御战略的便利,模型的提取也会造成严重隐私渗漏。