We show that $n$-variable tree-structured Ising models can be learned computationally-efficiently to within total variation distance $\epsilon$ from an optimal $O(n \ln n/\epsilon^2)$ samples, where $O(\cdot)$ hides an absolute constant which, importantly, does not depend on the model being learned - neither its tree nor the magnitude of its edge strengths, on which we place no assumptions. Our guarantees hold, in fact, for the celebrated Chow-Liu [1968] algorithm, using the plug-in estimator for estimating mutual information. While this (or any other) algorithm may fail to identify the structure of the underlying model correctly from a finite sample, we show that it will still learn a tree-structured model that is $\epsilon$-close to the true one in total variation distance, a guarantee called "proper learning." Our guarantees do not follow from known results for the Chow-Liu algorithm and the ensuing literature on learning graphical models, including a recent renaissance of algorithms on this learning challenge, which only yield asymptotic consistency results, or sample-inefficient and/or time-inefficient algorithms, unless further assumptions are placed on the graphical model, such as bounds on the "strengths" of the model's edges/hyperedges. While we establish guarantees for a widely known and simple algorithm, the analysis that this algorithm succeeds and is sample-optimal is quite complex, requiring a hierarchical classification of the edges into layers with different reconstruction guarantees, depending on their strength, combined with delicate uses of the subadditivity of the squared Hellinger distance over graphical models to control the error accumulation.
翻译:我们显示,美元可变树结构的Ising模型可以从一个最佳的 $O(n) n/ n/\\ epsilon2) 样本中以计算效率方式学习到在完全变异性结构的全变数范围内的全变数 。 我们的保证对于庆祝的 Chow-Liu [1968年] 算法, 使用 Expl- in 估测器来估算相互信息, 可以在总变异性积累中, 无论是它的树(\ cdott) 和它的边际力量的大小, 都可以在总变异性结构中, 进行计算。 事实上, 我们的保证对于已经庆祝的Chow- Liu [1968年] 算法和随后的学习图形模型, 包括最近需要学习挑战的变现算法, 而这个算法可能无法从有限的样本中正确识别基本模型的结构结构结构, 我们显示, 它仍然会学习树型模型的精度和直率性分析结果,