Advanced AI systems may be developed which exhibit capabilities that present significant risks to public safety or security. They may also exhibit capabilities that may be applied defensively in a wide set of domains, including (but not limited to) developing societal resilience against AI threats. We propose Coordinated Disclosure of Dual-Use Capabilities (CDDC) as a process to guide early information-sharing between advanced AI developers, US government agencies, and other private sector actors about these capabilities. The process centers around an information clearinghouse (the "coordinator") which receives evidence of dual-use capabilities from finders via mandatory and/or voluntary reporting pathways, and passes noteworthy reports to defenders for follow-up (i.e., further analysis and response). This aims to provide the US government, dual-use foundation model developers, and other actors with an overview of AI capabilities that could significantly impact public safety and security, as well as maximal time to respond.
翻译:先进人工智能系统可能展现出对公共安全或安全构成重大风险的能力。它们也可能展现出可在广泛领域(包括但不限于开发针对人工智能威胁的社会韧性)进行防御性应用的能力。我们提出双用途能力协调披露(CDDC)作为一个流程,旨在指导先进人工智能开发者、美国政府机构及其他私营部门参与者之间就这些能力进行早期信息共享。该流程围绕一个信息交换中心(即"协调者")展开,该中心通过强制性和/或自愿性报告途径接收发现者提交的双用途能力证据,并将值得关注的报告传递给防御方进行后续处理(即进一步分析与响应)。此举旨在为美国政府、双用途基础模型开发者及其他相关方提供可能对公共安全与安全产生重大影响的人工智能能力概览,并争取最长的响应时间。