This position paper argues that the theoretical inconsistency often observed among Responsible AI (RAI) metrics, such as differing fairness definitions or tradeoffs between accuracy and privacy, should be embraced as a valuable feature rather than a flaw to be eliminated. We contend that navigating these inconsistencies, by treating metrics as divergent objectives, yields three key benefits: (1) Normative Pluralism: Maintaining a full suite of potentially contradictory metrics ensures that the diverse moral stances and stakeholder values inherent in RAI are adequately represented. (2) Epistemological Completeness: The use of multiple, sometimes conflicting, metrics allows for a more comprehensive capture of multifaceted ethical concepts, thereby preserving greater informational fidelity about these concepts than any single, simplified definition. (3) Implicit Regularization: Jointly optimizing for theoretically conflicting objectives discourages overfitting to one specific metric, steering models towards solutions with enhanced generalization and robustness under real-world complexities. In contrast, efforts to enforce theoretical consistency by simplifying or pruning metrics risk narrowing this value diversity, losing conceptual depth, and degrading model performance. We therefore advocate for a shift in RAI theory and practice: from getting trapped in inconsistency to characterizing acceptable inconsistency thresholds and elucidating the mechanisms that permit robust, approximated consistency in practice.
翻译:本立场论文主张,在负责任人工智能(RAI)指标中常观察到的理论不一致性——如不同的公平性定义或准确性与隐私之间的权衡——应被视为一种有价值的特性而非需要消除的缺陷。我们认为,通过将指标视为相互冲突的目标来应对这些不一致性,可带来三个关键益处:(1)规范性多元主义:保持一套完整的、可能相互矛盾的指标,能够确保RAI中固有的多样化道德立场和利益相关者价值观得到充分体现。(2)认识论完备性:使用多个有时相互冲突的指标,能够更全面地捕捉多层面的伦理概念,从而比任何单一的简化定义保留更多关于这些概念的信息保真度。(3)隐式正则化:联合优化理论上冲突的目标,可防止模型过度拟合某一特定指标,引导模型在现实世界复杂性下获得具有更强泛化能力和鲁棒性的解决方案。相比之下,通过简化或删减指标来强制实现理论一致性的努力,可能收窄这种价值多样性、丧失概念深度并降低模型性能。因此,我们倡导RAI理论与实践进行转变:从陷入不一致性的困境,转向界定可接受的不一致性阈值,并阐明在实践中允许实现稳健、近似一致性的机制。