In many learning theory problems, a central role is played by a hypothesis class: we might assume that the data is labeled according to a hypothesis in the class (usually referred to as the realizable setting), or we might evaluate the learned model by comparing it with the best hypothesis in the class (the agnostic setting). Taking a step beyond these classic setups that involve only a single hypothesis class, we introduce comparative learning as a combination of the realizable and agnostic settings in PAC learning: given two binary hypothesis classes $S$ and $B$, we assume that the data is labeled according to a hypothesis in the source class $S$ and require the learned model to achieve an accuracy comparable to the best hypothesis in the benchmark class $B$. Even when both $S$ and $B$ have infinite VC dimensions, comparative learning can still have a small sample complexity. We show that the sample complexity of comparative learning is characterized by the mutual VC dimension $\mathsf{VC}(S,B)$ which we define to be the maximum size of a subset shattered by both $S$ and $B$. We also show a similar result in the online setting, where we give a regret characterization in terms of the mutual Littlestone dimension $\mathsf{Ldim}(S,B)$. These results also hold for partial hypotheses. We additionally show that the insights necessary to characterize the sample complexity of comparative learning can be applied to characterize the sample complexity of realizable multiaccuracy and multicalibration using the mutual fat-shattering dimension, an analogue of the mutual VC dimension for real-valued hypotheses. This not only solves an open problem proposed by Hu, Peale, Reingold (2022), but also leads to independently interesting results extending classic ones about regression, boosting, and covering number to our two-hypothesis-class setting.
翻译:在许多学习理论问题中,一个假设类扮演了中心角色:我们可以假设数据是根据类中的一个假设标签(通常称为可实现的设置),或者我们可以通过将数据与类中的最佳假设(不可知的设置)进行比较来评估所学模型。除了这些仅涉及单一假设类的经典设置之外,我们引入比较学习,作为PAC学习中可实现和不可知环境的组合:如果有两个二进制假设类(S$和B$),我们假设数据是根据来源类的假设标签(通常称为可实现的设置),我们可能假设数据是根据该类的假设标签(通常是可实现的设置),或者我们可能需要的多读性模型,或者我们可能实现与基准类中最佳假设相当的精确度。即使美元和美元都具有无限的 VC 层面,我们引入比较学习的样本复杂性的特征是相互的维C $mathfsf{VC} {VC} (S,B) 和 rick 。我们定义的数据的最大尺寸是由 $ 美元 和 美元 美元 的正变的正变。我们用一个相互的 显示的结果, 在网上显示一个真实的 。