The generate-filter-refine (iterative paradigm) based on large language models (LLMs) has achieved progress in reasoning, programming, and program discovery in AI+Science. However, the effectiveness of search depends on where to search, namely, how to encode the domain prior into an operationally structured hypothesis space. To this end, this paper proposes a compact formal theory that describes and measures LLM-assisted iterative search guided by domain priors. We represent an agent as a fuzzy relation operator on inputs and outputs to capture feasible transitions; the agent is thereby constrained by a fixed safety envelope. To describe multi-step reasoning/search, we weight all reachable paths by a single continuation parameter and sum them to obtain a coverage generating function; this induces a measure of reachability difficulty; and it provides a geometric interpretation of search on the graph induced by the safety envelope. We further provide the simplest testable inferences and validate them via a majority-vote instantiation. This theory offers a workable language and operational tools to measure agents and their search spaces, proposing a systematic formal description of iterative search constructed by LLMs.
翻译:基于大语言模型(LLMs)的生成-过滤-精炼(迭代范式)在AI+Science领域的推理、编程及程序发现方面取得了进展。然而,搜索的有效性取决于搜索何处,即如何将领域先验编码为可操作的结构化假设空间。为此,本文提出了一种紧凑的形式化理论,用于描述和衡量由领域先验引导的LLM辅助迭代搜索。我们将智能体表示为输入与输出上的模糊关系算子,以捕捉可行的状态转移;智能体因此受限于一个固定的安全包络。为描述多步推理/搜索,我们通过单一延续参数对所有可达路径进行加权并求和,得到一个覆盖生成函数;这导出了一个可达性难度的度量;并提供了在由安全包络诱导的图上进行搜索的几何解释。我们进一步给出了最简单的可检验推论,并通过一个多数投票的实例化进行了验证。该理论提供了一套可用的语言和操作工具来衡量智能体及其搜索空间,为LLMs构建的迭代搜索提出了一个系统化的形式化描述。