Humans and machines interact more frequently than ever and our societies are becoming increasingly hybrid. A consequence of this hybridisation is the degradation of societal trust due to the prevalence of AI-enabled deception. Yet, despite our understanding of the role of trust in AI in the recent years, we still do not have a computational theory to be able to fully understand and explain the role deception plays in this context. This is a problem because while our ability to explain deception in hybrid societies is delayed, the design of AI agents may keep advancing towards fully autonomous deceptive machines, which would pose new challenges to dealing with deception. In this paper we build a timely and meaningful interdisciplinary perspective on deceptive AI and reinforce a 20 year old socio-cognitive perspective on trust and deception, by proposing the development of DAMAS -- a holistic Multi-Agent Systems (MAS) framework for the socio-cognitive modelling and analysis of deception. In a nutshell this paper covers the topic of modelling and explaining deception using AI approaches from the perspectives of Computer Science, Philosophy, Psychology, Ethics, and Intelligence Analysis.
翻译:人类与机器的交互比以往任何时候都更加频繁,我们的社会正变得日益混合化。这种混合化的一个后果是,由于人工智能驱动的欺骗日益普遍,社会信任度不断下降。然而,尽管近年来我们理解了信任在人工智能中的作用,我们仍然缺乏一种计算理论来充分理解和解释欺骗在此背景下的作用。这是一个问题,因为在我们延迟解释混合社会中欺骗的能力的同时,人工智能体的设计可能持续向完全自主的欺骗机器发展,这将给应对欺骗带来新的挑战。在本文中,我们通过提出开发DAMAS——一个用于欺骗的社会认知建模与分析的整体式多智能体系统框架——构建了一个关于欺骗性人工智能的及时且有意义的跨学科视角,并强化了一个已有20年历史的关于信任与欺骗的社会认知视角。简而言之,本文从计算机科学、哲学、心理学、伦理学和情报分析的视角,涵盖了使用人工智能方法对欺骗进行建模和解释的主题。