Text-to-SQL is a subtask in semantic parsing that has seen rapid progress with the evolution of Large Language Models (LLMs). However, LLMs face challenges due to hallucination issues and a lack of domain-specific database knowledge(such as table schema and cell values). As a result, they can make errors in generating table names, columns, and matching values to the correct columns in SQL statements. This paper introduces a method of knowledge injection to enhance LLMs' ability to understand schema contents by incorporating prior knowledge. This approach improves their performance in Text-to-SQL tasks. Experimental results show that pre-training LLMs on domain-specific database knowledge and fine-tuning them on downstream Text-to-SQL tasks significantly improves the Execution Match (EX) and Exact Match (EM) metrics across various models. This effectively reduces errors in generating column names and matching values to the columns. Furthermore, the knowledge-injected models can be applied to many downstream Text-to-SQL tasks, demonstrating the generalizability of the approach presented in this paper.
翻译:文本到SQL是语义解析中的一个子任务,随着大语言模型(LLMs)的发展而取得了快速进展。然而,由于幻觉问题以及缺乏特定领域的数据库知识(如表结构和单元格值),LLMs在生成SQL语句中的表名、列名以及将值匹配到正确列时可能出现错误。本文引入了一种知识注入方法,通过融入先验知识来增强LLMs理解模式内容的能力,从而提升其在文本到SQL任务中的性能。实验结果表明,在特定领域数据库知识上对LLMs进行预训练,并在下游文本到SQL任务上进行微调,能够显著提高不同模型在执行匹配(EX)和精确匹配(EM)指标上的表现,有效减少了生成列名和将值匹配到列时的错误。此外,经过知识注入的模型可应用于许多下游文本到SQL任务,这证明了本文所提出方法的泛化能力。