Large Language Models (LLMs) have demonstrated their capabilities across various tasks, from language translation to complex reasoning. Understanding and predicting human behavior and biases are crucial for artificial intelligence (AI) assisted systems to provide useful assistance, yet it remains an open question whether these models can achieve this. This paper addresses this gap by leveraging the reasoning and generative capabilities of the LLMs to predict human behavior in two sequential decision-making tasks. These tasks involve balancing between exploitative and exploratory actions and handling delayed feedback, both essential for simulating real-life decision processes. We compare the performance of LLMs with a cognitive instance-based learning (IBL) model, which imitates human experiential decision-making. Our findings indicate that LLMs excel at rapidly incorporating feedback to enhance prediction accuracy. In contrast, the cognitive IBL model better accounts for human exploratory behaviors and effectively captures loss aversion bias, i.e., the tendency to choose a sub-optimal goal with fewer step-cost penalties rather than exploring to find the optimal choice, even with limited experience. The results highlight the benefits of integrating LLMs with cognitive architectures, suggesting that this synergy could enhance the modeling and understanding of complex human decision-making patterns.
翻译:大语言模型(LLMs)已在从语言翻译到复杂推理的多种任务中展现出其能力。理解并预测人类行为与认知偏差对于人工智能辅助系统提供有效支持至关重要,然而这些模型能否实现这一目标仍是一个开放性问题。本文通过利用大语言模型的推理与生成能力来预测人类在两个序列决策任务中的行为,以填补这一研究空白。这些任务涉及在利用式与探索式行为之间取得平衡,并处理延迟反馈,二者对于模拟现实决策过程均至关重要。我们将大语言模型的性能与一种模拟人类经验式决策的认知实例学习模型进行比较。研究结果表明,大语言模型在快速整合反馈以提高预测准确性方面表现优异。相比之下,认知实例学习模型能更好地解释人类的探索行为,并有效捕捉损失厌恶偏差——即即使在经验有限的情况下,个体也倾向于选择步骤代价惩罚较少但次优的目标,而非通过探索寻找最优选择。这些结果凸显了将大语言模型与认知架构相结合的优势,表明这种协同作用有望提升对复杂人类决策模式的建模与理解能力。