In this work, we introduce several expected generalization error bounds based on the Wasserstein distance. More precisely, we present full-dataset, single-letter, and random-subset bounds on both the standard setting and the randomized-subsample setting from Steinke and Zakynthinou [2020]. Moreover, we show that, when the loss function is bounded, these bounds recover from below (and thus are tighter than) current bounds based on the relative entropy and, for the standard setting, generate new, non-vacuous bounds also based on the relative entropy. Then, we show how similar bounds featuring the backward channel can be derived with the proposed proof techniques. Finally, we show how various new bounds based on different information measures (e.g., the lautum information or several $f$-divergences) can be derived from the presented bounds.
翻译:在这项工作中,我们引入了基于瓦森斯坦距离的若干预期的概括错误界限。更确切地说,我们在标准设置和Steinke和Zakynthinou [2020] 的随机分抽样设置上都展示了完整的数据集、单字母和随机的子分组界限。此外,我们显示,当损失函数被捆绑时,这些界限从以下(并因此更紧于)基于相对导体的当前界限中恢复过来,而对于标准设置而言,也根据相对导体产生新的、非挥发的界限。然后,我们展示了以拟议验证技术为主的后端通道的类似界限如何产生。最后,我们展示了基于不同信息计量(如Lautum信息或数美元-美元-参数)的不同新界限如何从显示的界限中产生。