Random forests on the one hand, and neural networks on the other hand, have met great success in the machine learning community for their predictive performance. Combinations of both have been proposed in the literature, notably leading to the so-called deep forests (DF) (Zhou \& Feng,2019). In this paper, our aim is not to benchmark DF performances but to investigate instead their underlying mechanisms. Additionally, DF architecture can be generally simplified into more simple and computationally efficient shallow forest networks. Despite some instability, the latter may outperform standard predictive tree-based methods. We exhibit a theoretical framework in which a shallow tree network is shown to enhance the performance of classical decision trees. In such a setting, we provide tight theoretical lower and upper bounds on its excess risk. These theoretical results show the interest of tree-network architectures for well-structured data provided that the first layer, acting as a data encoder, is rich enough.
翻译:一方面是随机森林,另一方面是神经网络,在机器学习社区中因其预测性能而取得了巨大成功。文献中提出了两者的结合,特别是导致所谓的深森林(DF)(Zhou ⁇ feng, 2019年)。在本文中,我们的目的不是为DF的性能设定基准,而是调查其基本机制。此外,DF的架构可以普遍简化为更简单、更计算高效的浅森林网络。尽管存在一些不稳定性,但后者可能优于标准的预测性树方法。我们展示了一个理论框架,在这个框架中,浅树网络展示了提高传统决策树的性能。在这种环境中,我们提供了其过度风险的严格理论下限和上限。这些理论结果表明,树木结构结构结构结构对良好数据的兴趣很大,只要第一层作为数据编码器,足够丰富。