The important recent book by G. Schurz appreciates that the no-free-lunch theorems (NFL) have major implications for the problem of (meta) induction. Here I review the NFL theorems, emphasizing that they do not only concern the case where there is a uniform prior -- they prove that there are "as many priors" (loosely speaking) for which any induction algorithm $A$ out-generalizes some induction algorithm $B$ as vice-versa. Importantly though, in addition to the NFL theorems, there are many {free lunch} theorems. In particular, the NFL theorems can only be used to compare the {marginal} expected performance of an induction algorithm $A$ with the marginal expected performance of an induction algorithm $B$. There is a rich set of free lunches which instead concern the statistical correlations among the generalization errors of induction algorithms. As I describe, the meta-induction algorithms that Schurz advocate as a "solution to Hume's problem" are just an example of such a free lunch based on correlations among the generalization errors of induction algorithms. I end by pointing out that the prior that Schurz advocates, which is uniform over bit frequencies rather than bit patterns, is contradicted by thousands of experiments in statistical physics and by the great success of the maximum entropy procedure in inductive inference.
翻译:G. Schurz最近的重要著作《G. Schurz》认识到,无自由午餐理论(NFL)对(元)上岗问题有重大影响。我在这里回顾NFL理论,强调它们不仅关注有统一的前科的案例,而且强调它们不仅关注“许多前科”(粗略地说),任何上岗算法都“像许多前科”(如许多前科 ) 将一些上岗算法($B$)作为反转法。重要的是,除了NFLL理论外,还有许多非免费午餐理论。特别是,NFL理论只能用来比较上岗算法的预期性能(MUGI),而美元和上岗算法的预期性能(LUB$$)与上岗算法的预期性能差值。 一套丰富的免费午餐,而不是上岗算法的统计性差。Schurz的元演算法,通过普通的上行进算法,在上比普通的进算法的进算法更接近于普通的进算法。