Over the past year, artificial intelligence (AI) companies have been increasingly adopting AI safety frameworks. These frameworks outline how companies intend to keep the potential risks associated with developing and deploying frontier AI systems to an acceptable level. Major players like Anthropic, OpenAI, and Google DeepMind have already published their frameworks, while another 13 companies have signaled their intent to release similar frameworks by February 2025. Given their central role in AI companies' efforts to identify and address unacceptable risks from their systems, AI safety frameworks warrant significant scrutiny. To enable governments, academia, and civil society to pass judgment on these frameworks, this paper proposes a grading rubric. The rubric consists of seven evaluation criteria and 21 indicators that concretize the criteria. Each criterion can be graded on a scale from A (gold standard) to F (substandard). The paper also suggests three methods for applying the rubric: surveys, Delphi studies, and audits. The purpose of the grading rubric is to enable nuanced comparisons between frameworks, identify potential areas of improvement, and promote a race to the top in responsible AI development.
翻译:过去一年间,人工智能(AI)公司日益广泛地采用AI安全框架。这些框架阐述了企业如何将开发与部署前沿AI系统相关的潜在风险控制在可接受水平。Anthropic、OpenAI和Google DeepMind等主要参与者已发布其框架,另有13家公司表示将在2025年2月前发布类似框架。鉴于AI安全框架在帮助企业识别和应对系统不可接受风险方面的核心作用,对其进行严格审查至关重要。为助力政府部门、学术界和公民社会对这些框架进行评估,本文提出一套分级评估标准。该标准包含七项评估准则和21项具体化指标,每项准则可按A级(黄金标准)至F级(不合格)进行分级。本文还提出三种应用该标准的方法:问卷调查、德尔菲研究和审计核查。本分级评估标准旨在实现框架间的精细化比较,识别潜在改进领域,并推动负责任AI开发领域的良性竞争。