Large language models (LLMs) have proven to be highly effective for solving complex reasoning tasks. Surprisingly, their capabilities can often be improved by iterating on previously generated solutions. In this context, a reasoning plan for generating and combining a set of solutions can be thought of as an algorithm for reasoning using a probabilistic oracle. We introduce a theoretical framework for analyzing such reasoning algorithms. This framework formalizes the principles underlying popular techniques for iterative improvement and answer aggregation, providing a foundation for designing a new generation of more powerful reasoning methods. Unlike approaches for understanding models that rely on architectural specifics, our model is grounded in experimental evidence. As a result, it offers a general perspective that may extend to a wide range of current and future reasoning oracles.
翻译:大型语言模型(LLMs)已被证明在解决复杂推理任务方面具有高效能。令人惊讶的是,通过迭代先前生成的解决方案,其能力往往能得到进一步提升。在此背景下,用于生成和组合一组解决方案的推理计划可被视为一种利用概率性预言机进行推理的算法。我们引入了一个理论框架来分析此类推理算法。该框架形式化了迭代改进与答案聚合等流行技术的基本原理,为设计新一代更强大的推理方法奠定了基础。与依赖架构细节的模型理解方法不同,我们的模型基于实验证据构建。因此,它提供了一个可能适用于当前及未来广泛推理预言机的通用视角。