The general reinforcement learning agent AIXI may be the only mathematical formalism of artificial general intelligence (AGI) supported by proof that its performance, measured according to a description of intelligence, is optimal. Unfortunately AIXI is incomputable, and its performance is subjective. This paper proposes an alternative, also supported by proof, which overcomes both problems. Integrating research from cognitive science (enactivism), philosophy of language (intension and extension), machine learning and planning (satplan), the notion of an arbitrary task is given mathematical rigour. This serves as an enactive, unified model of learning and reasoning within which a description of intelligence is formalised, along with a computable universal prior we prove grants optimal performance according to that description. This mathematical proof is then further supported with experimental results. The foremost limitation is that intelligence is computationally complex, and must be coupled with domain specific inductive biases to make real-world tasks of practical significance tractable (such as the domain of all tasks an average human might expect to undertake). Finally, by unifying concepts from AI, cognitive science and philosophy in one formalism, we have defined shared language to enable collaborative bridges within and beyond these fields.
翻译:全面强化学习代理AXI可能是人工一般智能的唯一数学形式主义,其依据是证明其根据情报描述衡量的性能是最佳的。不幸的是,AXI是无可争议的,其性能是主观的。本文件提出一种替代方法,也得到证据的支持,克服了这两个问题。将认知科学(活性)、语言哲学(强化和推广)、机器学习和规划(计划)的研究结合起来,对任意性任务的概念给予数学的严格性。这可以作为一种成文的、统一的学习和推理模式,在这种模式中,对情报的描述正式化,以及一个可计算的普遍化之前,我们证明能够根据这一描述实现最佳性性能。这个数学证据随后得到实验性结果的进一步支持。最大的限制是,情报是计算复杂的,必须与具体的领域带带偏见相结合,使现实世界中具有实际意义的任务(例如所有任务领域是人类可能期望承担的通常任务),通过将大赦国际的认知科学和哲学概念统一起来,我们界定了共同语言,以便能够在这些领域内外建立协作桥梁。