AI agents have been boosted by large language models. AI agents can function as intelligent assistants and complete tasks on behalf of their users with access to tools and the ability to execute commands in their environments. Through studying and experiencing the workflow of typical AI agents, we have raised several concerns regarding their security. These potential vulnerabilities are not addressed by the frameworks used to build the agents, nor by research aimed at improving the agents. In this paper, we identify and describe these vulnerabilities in detail from a system security perspective, emphasizing their causes and severe effects. Furthermore, we introduce defense mechanisms corresponding to each vulnerability with design and experiments to evaluate their viability. Altogether, this paper contextualizes the security issues in the current development of AI agents and delineates methods to make AI agents safer and more reliable.
翻译:AI Agent 在大语言模型的推动下得到了显著发展。AI Agent 能够作为智能助手,通过访问工具并在其环境中执行命令,代表用户完成任务。通过对典型 AI Agent 工作流程的研究与实践,我们对其安全性提出了若干担忧。这些潜在漏洞既未被构建 Agent 的框架所解决,也未被旨在改进 Agent 的研究所涵盖。本文从系统安全的角度详细识别并描述了这些漏洞,重点分析了其成因与严重后果。此外,我们针对每个漏洞提出了相应的防御机制,并通过设计与实验评估了其可行性。总体而言,本文系统阐述了当前 AI Agent 发展中的安全问题,并提出了使 AI Agent 更安全、更可靠的方法。