Casting machine learning as a type of search, we demonstrate that the proportion of problems that are favorable for a fixed algorithm is strictly bounded, such that no single algorithm can perform well over a large fraction of them. Our results explain why we must either continue to develop new learning methods year after year or move towards highly parameterized models that are both flexible and sensitive to their hyperparameters. We further give an upper bound on the expected performance for a search algorithm as a function of the mutual information between the target and the information resource (e.g., training dataset), proving the importance of certain types of dependence for machine learning. Lastly, we show that the expected per-query probability of success for an algorithm is mathematically equivalent to a single-query probability of success under a distribution (called a search strategy), and prove that the proportion of favorable strategies is also strictly bounded. Thus, whether one holds fixed the search algorithm and considers all possible problems or one fixes the search problem and looks at all possible search strategies, favorable matches are exceedingly rare. The forte (strength) of any algorithm is quantifiably restricted.
翻译:将机器学习作为一种类型的搜索,我们证明,对固定算法有利的问题比例是严格限定的,因此,任何单一算法都无法比大部分算法产生很大效果。我们的结果解释了为什么我们必须年复一年地继续开发新的学习方法,或者转向高度参数化的模型,这些模型既灵活又敏感,对超光度参数也具有灵活性。我们进一步将搜索算法的预期性能作为目标与信息资源(例如培训数据集)之间相互信息的一个函数设定上限,以证明某些类型的依赖对于机器学习的重要性。最后,我们表明,任何算法的预期人均成功概率在数学上等同于在分配(称为搜索战略)下的单一查询成功概率,并证明有利的战略的比例也严格受约束。因此,我们是否将搜索算法固定所有可能的问题,或将所有可能的问题固定在搜索问题上,并且看所有可能的搜索战略(例如,培训数据集),好匹配是极其罕见的。任何算法的直径(strength)都受到限制。