Boosting is one of the most successful ideas in machine learning. The most well-accepted explanations for the low generalization error of boosting algorithms such as AdaBoost stem from margin theory. The study of margins in the context of boosting algorithms was initiated by Schapire, Freund, Bartlett and Lee (1998) and has inspired numerous boosting algorithms and generalization bounds. To date, the strongest known generalization (upper bound) is the $k$th margin bound of Gao and Zhou (2013). Despite the numerous generalization upper bounds that have been proved over the last two decades, nothing is known about the tightness of these bounds. In this paper, we give the first margin-based lower bounds on the generalization error of boosted classifiers. Our lower bounds nearly match the $k$th margin bound and thus almost settle the generalization performance of boosted classifiers in terms of margins.
翻译:推动是机器学习中最成功的理念之一。 AdaBoost 等提振算法的低一般化错误最受人接受的解释来自边际理论。 在提振算法背景下的边际研究由Schapire、Freund、Bartlett和Lee (1998年)发起,激发了许多推力算法和一般化界限。迄今为止,已知最强的概括化(超约束)是加奥和周的边距(2013年)。尽管在过去20年中已经证明有许多普遍化的上限,但这些界限的紧凑性却一无所知。在本文中,我们给出了第一个基于边际的关于提振的分类工普遍化错误的下限。我们的下限几乎与美元边距的边距一致,从而几乎解决了提振的分类师在边距方面的普遍化表现。