Large Language Models (LLMs) have emerged as promising assistants for scientific writing. However, there have been concerns regarding the quality and reliability of the generated text, one of which is the citation accuracy and faithfulness. While most recent work relies on methods such as LLM-as-a-Judge, the reliability of LLM-as-a-Judge alone is also in doubt. In this work, we reframe citation evaluation as a problem of citation attribution alignment, which is assessing whether LLM-generated citations match those a human author would include for the same text. We propose CiteGuard, a retrieval-aware agent framework designed to provide more faithful grounding for citation validation. CiteGuard improves the prior baseline by 12.3%, and achieves up to 65.4% accuracy on the CiteME benchmark, on par with human-level performance (69.7%). It also enables the identification of alternative but valid citations.
翻译:大型语言模型(LLMs)已成为科学写作领域极具潜力的辅助工具。然而,生成文本的质量与可靠性仍存在诸多隐忧,其中引用准确性与忠实度尤为突出。尽管近期研究多依赖LLM-as-a-Judge等方法,但该方法的独立可靠性同样存疑。本研究将引用评估重新定义为引用归因对齐问题,即评估LLM生成的引用是否与人类作者针对相同文本会采用的引用相匹配。我们提出CiteGuard——一个具备检索感知能力的智能体框架,旨在为引用验证提供更可靠的依据。CiteGuard将现有基线性能提升了12.3%,在CiteME基准测试中达到65.4%的准确率,与人类水平表现(69.7%)相当。该框架还能有效识别替代性有效引用。