Large Language Models (LLMs) are increasingly integral to information dissemination and decision-making processes. Given their growing societal influence, understanding potential biases, particularly within the political domain, is crucial to prevent undue influence on public opinion and democratic processes. This work investigates political bias and stereotype propagation across eight prominent LLMs using the two-dimensional Political Compass Test (PCT). Initially, the PCT is employed to assess the inherent political leanings of these models. Subsequently, persona prompting with the PCT is used to explore explicit stereotypes across various social dimensions. In a final step, implicit stereotypes are uncovered by evaluating models with multilingual versions of the PCT. Key findings reveal a consistent left-leaning political alignment across all investigated models. Furthermore, while the nature and extent of stereotypes vary considerably between models, implicit stereotypes elicited through language variation are more pronounced than those identified via explicit persona prompting. Interestingly, for most models, implicit and explicit stereotypes show a notable alignment, suggesting a degree of transparency or "awareness" regarding their inherent biases. This study underscores the complex interplay of political bias and stereotypes in LLMs.
翻译:大型语言模型(LLMs)在信息传播与决策过程中日益重要。鉴于其不断增长的社会影响力,理解其潜在偏见——特别是在政治领域——对于防止其对公众舆论和民主进程产生不当影响至关重要。本研究采用二维政治罗盘测试(PCT),对八个主流大型语言模型中的政治偏见与刻板印象传播进行了系统探究。首先,运用PCT评估这些模型固有的政治倾向。随后,通过PCT结合角色提示法,探究模型在不同社会维度上的显性刻板印象。最后,通过多语言版PCT评估模型,揭示其隐性刻板印象。关键发现表明:所有被调查模型均呈现出一致的左倾政治倾向。此外,尽管不同模型间刻板印象的性质与程度存在显著差异,但通过语言变异引发的隐性刻板印象比通过显性角色提示法识别的更为突出。值得注意的是,对于大多数模型,隐性刻板印象与显性刻板印象表现出显著的一致性,这暗示模型对其固有偏见存在某种程度的透明度或“认知”。本研究揭示了大型语言模型中政治偏见与刻板印象之间复杂的相互作用。