Culture fundamentally shapes people's reasoning, behavior, and communication. As people increasingly use generative artificial intelligence (AI) to expedite and automate personal and professional tasks, cultural values embedded in AI models may bias people's authentic expression and contribute to the dominance of certain cultures. We conduct a disaggregated evaluation of cultural bias for five widely used large language models (OpenAI's GPT-4o/4-turbo/4/3.5-turbo/3) by comparing the models' responses to nationally representative survey data. All models exhibit cultural values resembling English-speaking and Protestant European countries. We test cultural prompting as a control strategy to increase cultural alignment for each country/territory. For recent models (GPT-4, 4-turbo, 4o), this improves the cultural alignment of the models' output for 71-81% of countries and territories. We suggest using cultural prompting and ongoing evaluation to reduce cultural bias in the output of generative AI.
翻译:文化从根本上塑造了人们的推理、行为和沟通方式。随着人们越来越多地使用生成式人工智能来加速和自动化个人及专业任务,AI模型中嵌入的文化价值观可能会影响人们的真实表达,并加剧特定文化的主导地位。我们通过比较模型对全国代表性调查数据的响应,对五种广泛使用的大型语言模型(OpenAI的GPT-4o/4-turbo/4/3.5-turbo/3)进行了文化偏见的分解评估。所有模型都表现出与英语国家和新教欧洲国家相似的文化价值观。我们测试了文化提示作为一种控制策略,以提高每个国家/地区的文化对齐度。对于近期模型(GPT-4、4-turbo、4o),该方法改善了71-81%国家和地区输出结果的文化对齐度。我们建议使用文化提示和持续评估来减少生成式AI输出中的文化偏见。