Large Language Models (LLMs) have emerged as promising assistants for scientific writing. However, there have been concerns regarding the quality and reliability of the generated text, one of which is the citation accuracy and faithfulness. While most recent work relies on methods such as LLM-as-a-Judge, the reliability of LLM-as-a-Judge alone is also in doubt. In this work, we reframe citation evaluation as a problem of citation attribution alignment, which is assessing whether LLM-generated citations match those a human author would include for the same text. We propose CiteGuard, a retrieval-aware agent framework designed to provide more faithful grounding for citation validation. CiteGuard improves the prior baseline by 12.3%, and achieves up to 65.4% accuracy on the CiteME benchmark, on par with human-level performance (69.7%). It also enables the identification of alternative but valid citations.
翻译:大型语言模型(LLMs)已成为科学写作领域颇具前景的辅助工具。然而,人们对其生成文本的质量与可靠性存在担忧,其中引文准确性与忠实度是核心问题之一。尽管近期研究多采用诸如LLM-as-a-Judge等方法,但仅依赖此类方法的可靠性仍存疑。本研究将引文评估重新定义为引文归因对齐问题,即评估LLM生成的引文是否与人类作者针对相同文本会采用的引文相匹配。我们提出CiteGuard——一个具备检索感知能力的智能体框架,旨在为引文验证提供更可靠的依据。CiteGuard将现有基线性能提升了12.3%,在CiteME基准测试中达到65.4%的准确率,与人类水平表现(69.7%)相当。该框架还能有效识别替代性有效引文。