Large Language Models (LLMs) demonstrate remarkable performance in semantic understanding and generation, yet accurately assessing their output reliability remains a significant challenge. While numerous studies have explored calibration techniques, they primarily focus on White-Box LLMs with accessible parameters. Black-Box LLMs, despite their superior performance, pose heightened requirements for calibration techniques due to their API-only interaction constraints. Although recent researches have achieved breakthroughs in black-box LLMs calibration, a systematic survey of these methodologies is still lacking. To bridge this gap, we presents the first comprehensive survey on calibration techniques for black-box LLMs. We first define the Calibration Process of LLMs as comprising two interrelated key steps: Confidence Estimation and Calibration. Second, we conduct a systematic review of applicable methods within black-box settings, and provide insights on the unique challenges and connections in implementing these key steps. Furthermore, we explore typical applications of Calibration Process in black-box LLMs and outline promising future research directions, providing new perspectives for enhancing reliability and human-machine alignment. This is our GitHub link: https://github.com/LiangruXie/Calibration-Process-in-Black-Box-LLMs
翻译:大语言模型在语义理解与生成方面展现出卓越性能,但其输出可靠性的准确评估仍面临重大挑战。尽管已有大量研究探索校准技术,这些研究主要聚焦于参数可访问的白盒大语言模型。黑盒大语言模型虽具备更优性能,但由于其仅能通过API交互的限制,对校准技术提出了更高要求。尽管近期研究已在黑盒大语言模型校准方面取得突破,但针对这些方法的系统性综述仍然缺乏。为填补这一空白,本文首次对黑盒大语言模型的校准技术进行全面综述。首先,我们将大语言模型的校准流程定义为包含两个相互关联的关键步骤:置信度估计与校准。其次,我们系统梳理了黑盒场景下的适用方法,并就实施这些关键步骤时面临的独特挑战与内在联系提供见解。此外,我们探讨了校准流程在黑盒大语言模型中的典型应用场景,并展望了未来有前景的研究方向,为提升模型可靠性与人机对齐提供了新的视角。本项目GitHub链接:https://github.com/LiangruXie/Calibration-Process-in-Black-Box-LLMs