As large language models (LLMs) are used in sensitive fields, accurately verifying their computational provenance without disclosing their training datasets poses a significant challenge, particularly in regulated sectors such as healthcare, which have strict requirements for dataset use. Traditional approaches either incur substantial computational cost to fully verify the entire training process or leak unauthorized information to the verifier. Therefore, we introduce ZKPROV, a novel cryptographic framework allowing users to verify that the LLM's responses to their prompts are trained on datasets certified by the authorities that own them. Additionally, it ensures that the dataset's content is relevant to the users' queries without revealing sensitive information about the datasets or the model parameters. ZKPROV offers a unique balance between privacy and efficiency by binding training datasets, model parameters, and responses, while also attaching zero-knowledge proofs to the responses generated by the LLM to validate these claims. Our experimental results demonstrate sublinear scaling for generating and verifying these proofs, with end-to-end overhead under 3.3 seconds for models up to 8B parameters, presenting a practical solution for real-world applications. We also provide formal security guarantees, proving that our approach preserves dataset confidentiality while ensuring trustworthy dataset provenance.
翻译:随着大型语言模型(LLMs)在敏感领域中的应用日益广泛,如何在不泄露其训练数据集的前提下准确验证其计算溯源成为一个重大挑战,这在医疗等对数据集使用有严格要求的受监管领域尤为突出。传统方法要么需要高昂的计算成本来完整验证整个训练过程,要么会向验证方泄露未授权信息。为此,我们提出了ZKPROV——一种新型密码学框架,允许用户验证LLM对其提示的响应是基于数据所有者权威认证的数据集训练生成的。此外,该框架能确保数据集内容与用户查询相关,同时不泄露数据集或模型参数的敏感信息。ZKPROV通过将训练数据集、模型参数和响应进行绑定,并在LLM生成的响应上附加零知识证明以验证这些声明,在隐私保护与效率之间实现了独特的平衡。实验结果表明,生成和验证这些证明具有亚线性扩展特性,对于参数量高达80亿的模型,端到端开销低于3.3秒,为实际应用提供了可行的解决方案。我们还提供了形式化的安全保证,证明该方法在确保可信数据集溯源的同时能有效保护数据集机密性。