The CIA security triad - Confidentiality, Integrity, and Availability - is a cornerstone of data and cybersecurity. With the emergence of large language model (LLM) applications, a new class of threat, known as prompt injection, was first identified in 2022. Since then, numerous real-world vulnerabilities and exploits have been documented in production LLM systems, including those from leading vendors like OpenAI, Microsoft, Anthropic and Google. This paper compiles real-world exploits and proof-of concept examples, based on the research conducted and publicly documented by the author, demonstrating how prompt injection undermines the CIA triad and poses ongoing risks to cybersecurity and AI systems at large.
翻译:CIA安全三要素——机密性、完整性与可用性——是数据与网络安全体系的基石。随着大语言模型(LLM)应用的出现,一类名为“提示注入”的新型威胁于2022年首次被识别。此后,众多实际漏洞与攻击案例已在生产环境的LLM系统中被记录,包括来自OpenAI、微软、Anthropic和谷歌等领先厂商的系统。本文基于作者开展并公开记录的研究,汇编了实际攻击案例与概念验证示例,系统论证了提示注入如何破坏CIA三要素,并对网络安全及整体人工智能系统构成持续风险。