The Rashomon effect -- the existence of multiple, distinct models that achieve nearly equivalent predictive performance -- has emerged as a fundamental phenomenon in modern machine learning and statistics. In this paper, we explore the causes underlying the Rashomon effect, organizing them into three categories: statistical sources arising from finite samples and noise in the data-generating process; structural sources arising from non-convexity of optimization objectives and unobserved variables that create fundamental non-identifiability; and procedural sources arising from limitations of optimization algorithms and deliberate restrictions to suboptimal model classes. We synthesize insights from machine learning, statistics, and optimization literature to provide a unified framework for understanding why the multiplicity of good models arises. A key distinction emerges: statistical multiplicity diminishes with more data, structural multiplicity persists asymptotically and cannot be resolved without different data or additional assumptions, and procedural multiplicity reflects choices made by practitioners. Beyond characterizing causes, we discuss both the challenges and opportunities presented by the Rashomon effect, including implications for inference, interpretability, fairness, and decision-making under uncertainty.
翻译:罗夏效应——即存在多个不同模型却能实现近乎等同预测性能的现象——已成为现代机器学习与统计学中的一个基本现象。本文深入探讨罗夏效应的成因,并将其归纳为三类:源于有限样本和数据生成过程中噪声的统计性成因;源于优化目标非凸性及未观测变量导致根本性不可识别性的结构性成因;以及源于优化算法局限性和人为限制于次优模型类别的程序性成因。我们综合机器学习、统计学与优化理论的研究成果,构建了一个统一框架来解释优质模型多重性产生的机理。关键区别在于:统计多重性随数据量增加而减弱,结构多重性具有渐近持续性且无法通过不同数据或附加假设消除,而程序多重性则反映了实践者的主观选择。除成因分析外,本文进一步探讨罗夏效应带来的挑战与机遇,包括其对统计推断、可解释性、公平性以及不确定性决策等领域的影响。