Question answering is a natural language understanding task that involves reasoning over both explicit context and unstated, relevant domain knowledge. Large language models (LLMs), which underpin most contemporary question answering systems, struggle to induce how concepts relate in specialized domains such as medicine. Existing medical LLMs are also costly to train. In this work, we present MEG, a parameter-efficient approach for medical knowledge-augmented LLMs. MEG uses a lightweight mapping network to integrate graph embeddings into the LLM, enabling it to leverage external knowledge in a cost-effective way. We evaluate our method on four popular medical multiple-choice datasets and show that LLMs greatly benefit from the factual grounding provided by knowledge graph embeddings. MEG attains an average of +10.2% accuracy over the Mistral-Instruct baseline, and +6.7% over specialized models like BioMistral. We also show results based on Llama-3. Finally, we show that MEG's performance remains robust to the choice of graph encoder.
翻译:问答是一种自然语言理解任务,需要对显式上下文和未明确陈述的相关领域知识进行推理。支撑当前大多数问答系统的大语言模型(LLMs)难以在医学等专业领域中推导概念间的关联。现有的医学大语言模型训练成本也较高。本文提出MEG,一种参数高效的医学知识增强大语言模型方法。MEG使用轻量级映射网络将图嵌入集成到大语言模型中,使其能够以经济高效的方式利用外部知识。我们在四个主流医学多项选择数据集上评估了该方法,结果表明大语言模型从知识图谱嵌入提供的事实基础中获益显著。MEG在Mistral-Instruct基线上平均准确率提升+10.2%,相较于BioMistral等专业模型提升+6.7%。我们还展示了基于Llama-3的实验结果。最后,我们证明MEG的性能对图编码器的选择具有鲁棒性。