Laws of large numbers guarantee that given a large enough sample from some population, the measure of any fixed sub-population is well-estimated by its frequency in the sample. We study laws of large numbers in sampling processes that can affect the environment they are acting upon and interact with it. Specifically, we consider the sequential sampling model proposed by Ben-Eliezer and Yogev (2020), and characterize the classes which admit a uniform law of large numbers in this model: these are exactly the classes that are \emph{online learnable}. Our characterization may be interpreted as an online analogue to the equivalence between learnability and uniform convergence in statistical (PAC) learning. The sample-complexity bounds we obtain are tight for many parameter regimes, and as an application, we determine the optimal regret bounds in online learning, stated in terms of \emph{Littlestone's dimension}, thus resolving the main open question from Ben-David, P\'al, and Shalev-Shwartz (2009), which was also posed by Rakhlin, Sridharan, and Tewari (2015).
翻译:大量数据的法律保证,如果从某些人群中有大量的样本,那么任何固定亚人口量的测量量会根据其抽样的频率而得到准确估计。我们研究在取样过程中大量数量的法规,这些法规会影响它们正在采取行动并与之互动的环境。具体地说,我们考虑Ben-Eliezer和Yogev(202020年)提出的顺序抽样模型,并给这一模型中接受大量数据的统一法规的类别定性:这些类别恰恰是可学习的类别。我们的定性可被解释为在线模拟,相当于统计学(PAC)的可学习性和统一趋同性之间的等同。我们获得的样本兼容性界限对于许多参数系统来说是紧凑的,作为应用,我们决定了在线学习中的最佳遗憾界限,用\emph{Litstone的尺寸表示},从而解决了Ben-David、P\'al和Shalev-Shwartz(2009年)提出的主要问题,这也是Rakhlin、Sridharan和Tewari(2015年)提出的主要问题。