Anthropomorphic language describing artificial intelligence (AI) is widespread in media, policy, and everyday discourse; so too are discussions of AI bad behavior, from hallucinations to inappropriate comments. How does humanizing language about AI shape moral judgments when AI behaves badly? Across four experiments (total N = 1,020), we tested whether lexical anthropomorphism (LA) primes shape judgments of AI moral character, behavior morality, and behavioral responsibility. Studies 1-3 tested interactions between anthropomorphic language and humanizing design cues (icons, names, self-referencing) in the context of amoral errors. Study 4 extended this to genuinely immoral AI behavior across seven moral-violation types. Results indicate humanizing language and design cues have little influence on moral judgments of misbehaving AI. Where effects emerged, high-anthropomorphic primes elevated perceptions of an AI's capacity for dishonesty. The type of moral violation observed was the strongest predictor of moral judgments, with harm and degradation violations producing the broadest negative character assessments. Prime drift, horn effects, and egoistic value orientations emerged as potentially important predictors of AI moral judgments.
翻译:描述人工智能(AI)的拟人化用语在媒体、政策和日常话语中广泛存在;同样,从幻觉到不当评论等AI不良行为的讨论也十分普遍。当AI行为不端时,将AI人性化的语言如何影响道德判断?通过四项实验(总样本量N=1,020),我们检验了词汇拟人化(LA)启动是否会影响对AI道德品格、行为道德性及行为责任的判断。研究1-3在非道德错误情境中测试了拟人化语言与人性化设计线索(图标、名称、自我指涉)的交互作用。研究4将此扩展至涵盖七种道德违规类型的真正不道德AI行为。结果表明,人性化语言和设计线索对行为不端AI的道德判断影响甚微。在出现效应的情况下,高拟人化启动提升了人们对AI不诚实能力的感知。观察到的道德违规类型是道德判断的最强预测因子,其中伤害与贬低类违规产生了最广泛的负面品格评价。启动漂移、晕轮效应及利己价值取向成为AI道德判断中潜在的重要预测因子。