As large language models (LLMs) start interacting with each other and generating an increasing amount of text online, it becomes crucial to better understand how information is transformed as it passes from one LLM to the next. While significant research has examined individual LLM behaviors, existing studies have largely overlooked the collective behaviors and information distortions arising from iterated LLM interactions. Small biases, negligible at the single output level, risk being amplified in iterated interactions, potentially leading the content to evolve towards attractor states. In a series of telephone game experiments, we apply a transmission chain design borrowed from the human cultural evolution literature: LLM agents iteratively receive, produce, and transmit texts from the previous to the next agent in the chain. By tracking the evolution of text toxicity, positivity, difficulty, and length across transmission chains, we uncover the existence of biases and attractors, and study their dependence on the initial text, the instructions, language model, and model size. For instance, we find that more open-ended instructions lead to stronger attraction effects compared to more constrained tasks. We also find that different text properties display different sensitivity to attraction effects, with toxicity leading to stronger attractors than length. These findings highlight the importance of accounting for multi-step transmission dynamics and represent a first step towards a more comprehensive understanding of LLM cultural dynamics.
翻译:随着大语言模型(LLMs)开始相互交互并在线上生成日益增多的文本,深入理解信息在LLM间传递时的转化过程变得至关重要。尽管已有大量研究考察了单个LLM的行为,但现有工作大多忽视了迭代LLM交互中产生的集体行为与信息扭曲。在单次输出层面可忽略的微小偏差,在迭代交互中可能被放大,导致内容演化趋向吸引子状态。在一系列传话游戏实验中,我们借鉴人类文化演化研究中的传播链设计:LLM智能体在链式结构中迭代接收、生成并将文本传递给下一个智能体。通过追踪传播链中文本的毒性、积极性、难度和长度的演化,我们揭示了偏差与吸引子的存在,并研究了它们对初始文本、指令、语言模型及模型规模的依赖性。例如,我们发现相较于约束性更强的任务,开放性指令会导致更显著的吸引效应。我们还发现不同文本属性对吸引效应的敏感性存在差异,毒性比长度表现出更强的吸引子特性。这些发现凸显了考虑多步传播动力学的重要性,并为更全面理解LLM文化动态迈出了第一步。