We solve the neural network robustness problem by adding Similarity (i.e., correctly predicted depth-matches into training)-awareness and Distance-to-training-distribution-awareness to the existing output Magnitude (i.e., decision-boundary)-awareness of the softmax function. The resulting sdm activation function provides strong signals of the relative epistemic (reducible) predictive uncertainty. We use this novel behavior to further address the complementary HCI problem of mapping the output to human-interpretable summary statistics over relevant partitions of a held-out calibration set. Estimates of prediction-conditional uncertainty are obtained via a parsimonious learned transform over the class-conditional empirical CDFs of the output of a final-layer sdm activation function. For decision-making and as an intrinsic model check, estimates of class-conditional accuracy are obtained by further partitioning the high-probability regions of this calibrated output into class-conditional, region-specific CDFs. The uncertainty estimates from sdm calibration are remarkably robust to test-time distribution shifts and out-of-distribution inputs; incorporate awareness of the effective sample size; provide estimates of uncertainty from the learning and data splitting processes; and are well-suited for selective classification and conditional branching for additional test-time compute based on the predictive uncertainty, as for selective LLM generation, routing, and composition over multiple models and retrieval. Finally, we construct sdm networks, LLMs with uncertainty-aware verification and interpretability-by-exemplar as intrinsic properties. We provide open-source software implementing these results.
翻译:我们通过向现有softmax函数的输出幅度(即决策边界)感知能力中添加相似性(即正确预测的深度匹配)感知和训练分布距离感知,解决了神经网络鲁棒性问题。由此产生的sdm激活函数能够提供关于相对认知(可约简)预测不确定性的强信号。我们利用这种新颖特性进一步解决互补的人机交互问题:将输出映射到预留校准集相关分区上的人类可解释汇总统计量。通过在学习到的简约变换上对最终层sdm激活函数输出的类条件经验累积分布函数进行处理,获得预测条件不确定性的估计值。为支持决策制定并作为内在模型检验,通过将校准输出的高概率区域进一步划分为类条件、区域特定的累积分布函数,获得类条件准确率的估计值。sdm校准产生的不确定性估计具有以下显著特性:对测试时分布偏移和分布外输入表现出卓越的鲁棒性;包含有效样本量的感知能力;提供来自学习过程和数据划分过程的不确定性估计;特别适用于选择性分类和基于预测不确定性的条件分支处理(例如针对选择性大语言模型生成、路由以及多模型与检索的组合场景)。最后,我们构建了sdm网络——将不确定性感知验证和范例可解释性作为内在属性的大语言模型。我们提供了实现这些成果的开源软件。