The rapid integration of Large Language Models (LLMs) into decentralized physical infrastructure networks (DePIN) is currently bottlenecked by the Verifiability Trilemma, which posits that a decentralized inference system cannot simultaneously achieve high computational integrity, low latency, and low cost. Existing cryptographic solutions, such as Zero-Knowledge Machine Learning (ZKML), suffer from superlinear proving overheads (O(k NlogN)) that render them infeasible for billionparameter models. Conversely, optimistic approaches (opML) impose prohibitive dispute windows, preventing real-time interactivity, while recent "Proof of Quality" (PoQ) paradigms sacrifice cryptographic integrity for subjective semantic evaluation, leaving networks vulnerable to model downgrade attacks and reward hacking. In this paper, we introduce Optimistic TEE-Rollups (OTR), a hybrid verification protocol that harmonizes these constraints. OTR leverages NVIDIA H100 Confidential Computing Trusted Execution Environments (TEEs) to provide sub-second Provisional Finality, underpinned by an optimistic fraud-proof mechanism and stochastic Zero-Knowledge spot-checks to mitigate hardware side-channel risks. We formally define Proof of Efficient Attribution (PoEA), a consensus mechanism that cryptographically binds execution traces to hardware attestations, thereby guaranteeing model authenticity. Extensive simulations demonstrate that OTR achieves 99% of the throughput of centralized baselines with a marginal cost overhead of $0.07 per query, maintaining Byzantine fault tolerance against rational adversaries even in the presence of transient hardware vulnerabilities.
翻译:大型语言模型(LLMs)在去中心化物理基础设施网络(DePIN)中的快速集成,目前正受限于可验证性三难困境——即去中心化推理系统无法同时实现高计算完整性、低延迟与低成本。现有密码学解决方案(如零知识机器学习ZKML)存在超线性证明开销(O(k NlogN)),导致其无法适用于十亿参数级模型。相反,乐观验证方案(opML)因设置过长的争议窗口而阻碍实时交互;而近期提出的“质量证明”(PoQ)范式为追求主观语义评估牺牲了密码学完整性,使网络易受模型降级攻击与奖励操纵威胁。本文提出乐观型TEE-Rollups(OTR),一种融合多方约束的混合验证协议。OTR利用英伟达H100机密计算可信执行环境(TEEs)实现亚秒级临时终局性,并通过乐观欺诈证明机制与随机零知识抽查相结合以缓解硬件侧信道风险。我们形式化定义高效归属证明(PoEA),该共识机制通过密码学方法将执行轨迹与硬件证明绑定,从而确保模型真实性。大量仿真实验表明,OTR在保持拜占庭容错能力(即使面对瞬时硬件漏洞下的理性攻击者)的同时,可实现中心化基线方案99%的吞吐量,且每查询仅产生0.07美元的边际成本开销。