We propose a new definition of instance optimality for differentially private estimation algorithms. Our definition requires an optimal algorithm to compete, simultaneously for every dataset $D$, with the best private benchmark algorithm that (a) knows $D$ in advance and (b) is evaluated by its worst-case performance on large subsets of $D$. That is, the benchmark algorithm need not perform well when potentially extreme points are added to $D$; it only has to handle the removal of a small number of real data points that already exist. This makes our benchmark significantly stronger than those proposed in prior work. We nevertheless show, for real-valued datasets, how to construct private algorithms that achieve our notion of instance optimality when estimating a broad class of dataset properties, including means, quantiles, and $\ell_p$-norm minimizers. For means in particular, we provide a detailed analysis and show that our algorithm simultaneously matches or exceeds the asymptotic performance of existing algorithms under a range of distributional assumptions.
翻译:我们提出了对差别性私人估算算法最佳实例的新定义。 我们的定义要求一种最佳算法,在对每套数据集同时对每套美元,与最佳的私人基准算法进行竞争,这种算法:(a) 事先知道美元,和(b) 根据其对大子集美元最坏的性能进行评估。也就是说,当潜在极端点加到美元时,基准算法并不需要很好地发挥作用;它只需要处理删除少量已经存在的真实数据点。这使得我们的基准大大高于先前工作中的建议。然而,对于实际价值的数据集,我们展示了如何在估计广泛的数据集特性类别时,包括手段、四分法和美元/ell_p$-norm 最小化器时,建立能够实现我们实例最佳性概念的私人算法。 具体地说,我们提供了详细的分析,并表明我们的算法在一系列分配假设下同时匹配或超过现有算法的无症状性能。</s>