In this paper, we focus our attention on private Empirical Risk Minimization (ERM), which is one of the most commonly used data analysis method. We take the first step towards solving the above problem by theoretically exploring the effect of epsilon (the parameter of differential privacy that determines the strength of privacy guarantee) on utility of the learning model. We trace the change of utility with modification of epsilon and reveal an established relationship between epsilon and utility. We then formalize this relationship and propose a practical approach for estimating the utility under an arbitrary value of epsilon. Both theoretical analysis and experimental results demonstrate high estimation accuracy and broad applicability of our approach in practical applications. As providing algorithms with strong utility guarantees that also give privacy when possible becomes more and more accepted, our approach would have high practical value and may be likely to be adopted by companies and organizations that would like to preserve privacy but are unwilling to compromise on utility.
翻译:在本文中,我们把注意力集中在私人经验风险最小化(ERM)上,这是最常用的数据分析方法之一。我们从理论上探讨epsilon(差异隐私参数,确定隐私保障的力度)对学习模式的实用性的影响,从而朝着解决上述问题迈出了第一步。我们用对epsilon的修改来追踪功用的变化,并揭示Essilon与实用性之间的既定关系。然后我们正式确定这种关系,提出一种实用的方法,在epsilon的任意价值下评估功用。理论分析和实验结果都表明我们的做法在实际应用中具有很高的估计准确性和广泛适用性。由于为算法提供强有力的效用保证,在可能时也使隐私得到越来越多的接受,我们的方法将具有很高的实际价值,并有可能被愿意保护隐私但不愿在实用性上妥协的公司和组织采用。