Hallucinations in large language models pose a critical challenge for applications requiring factual reliability, particularly in high-stakes domains such as finance. This work presents an effective approach for detecting and editing factually incorrect content in model-generated responses based on the provided context. Given a user-defined domain-specific error taxonomy, we construct a synthetic dataset by inserting tagged errors into financial question-answering corpora and then fine-tune four language models, Phi-4, Phi-4-mini, Qwen3-4B, and Qwen3-14B, to detect and edit these factual inaccuracies. Our best-performing model, fine-tuned Phi-4, achieves an 8% improvement in binary F1 score and a 30% gain in overall detection performance compared to OpenAI-o3. Notably, our fine-tuned Phi-4-mini model, despite having only 4 billion parameters, maintains competitive performance with just a 2% drop in binary detection and a 0.1% decline in overall detection compared to OpenAI-o3. Our work provides a practical solution for detecting and editing factual inconsistencies in financial text generation while introducing a generalizable framework that can enhance the trustworthiness and alignment of large language models across diverse applications beyond finance. Our code and data are available at https://github.com/pegasi-ai/shield.
翻译:大型语言模型中的幻觉问题对需要事实可靠性的应用构成了严峻挑战,尤其在金融等高风险领域。本研究提出一种基于给定上下文检测并编辑模型生成回复中事实性错误内容的有效方法。给定用户定义的领域特定错误分类体系,我们通过向金融问答语料库中插入带标签的错误来构建合成数据集,随后对Phi-4、Phi-4-mini、Qwen3-4B和Qwen3-14B四个语言模型进行微调,以检测并编辑这些事实性错误。性能最优的微调Phi-4模型在二元F1分数上较OpenAI-o3提升8%,整体检测性能提升30%。值得注意的是,仅拥有40亿参数的微调Phi-4-mini模型仍保持竞争优势,其二元检测性能仅下降2%,整体检测性能仅降低0.1%。本工作为金融文本生成中的事实不一致性检测与编辑提供了实用解决方案,同时提出了一个可泛化的框架,能够增强大型语言模型在金融及其他多样化应用中的可信度与对齐性。我们的代码与数据公开于https://github.com/pegasi-ai/shield。