This study investigates the efficacy of Large Language Models (LLMs) in causal discovery. Using newly available open-source LLMs, OLMo and BLOOM, which provide access to their pre-training corpora, we investigate how LLMs address causal discovery through three research questions. We examine: (i) the impact of memorization for accurate causal relation prediction, (ii) the influence of incorrect causal relations in pre-training data, and (iii) the contextual nuances that influence LLMs' understanding of causal relations. Our findings indicate that while LLMs are effective in recognizing causal relations that occur frequently in pre-training data, their ability to generalize to new or rare causal relations is limited. Moreover, the presence of incorrect causal relations significantly undermines the confidence of LLMs in corresponding correct causal relations, and the contextual information critically affects the outcomes of LLMs to discern causal connections between random variables.
翻译:本研究探讨了大型语言模型(LLMs)在因果发现任务中的有效性。利用新近开源且提供预训练语料库访问权限的LLMs(OLMo和BLOOM),我们通过三个研究问题探究LLMs如何处理因果发现。我们考察:(i)记忆效应对准确因果关系预测的影响,(ii)预训练数据中错误因果关系的影响,以及(iii)影响LLMs理解因果关系的上下文细微差别。研究结果表明,虽然LLMs能有效识别预训练数据中频繁出现的因果关系,但其对新出现或罕见因果关系的泛化能力有限。此外,错误因果关系的存在会显著削弱LLMs对相应正确因果关系的置信度,且上下文信息对LLMs辨别随机变量间因果关联的结果具有关键影响。