Predictive models may generate biased predictions when classifying imbalanced datasets. This happens when the model favors the majority class, leading to low performance in accurately predicting the minority class. To address this issue, balancing or resampling methods are critical data-centric AI approaches in the modeling process to improve prediction performance. However, there have been debates and questions about the functionality of these methods in recent years. In particular, many candidate models may exhibit very similar predictive performance, called the Rashomon effect, in model selection, and they may even produce different predictions for the same observations. Selecting one of these models without considering the predictive multiplicity -- which is the case of yielding conflicting models' predictions for any sample -- can result in blind selection. In this paper, the impact of balancing methods on predictive multiplicity is examined using the Rashomon effect. It is crucial because the blind model selection in data-centric AI is risky from a set of approximately equally accurate models. This may lead to severe problems in model selection, validation, and explanation. To tackle this matter, we conducted real dataset experiments to observe the impact of balancing methods on predictive multiplicity through the Rashomon effect by using a newly proposed metric obscurity in addition to the existing ones: ambiguity and discrepancy. Our findings showed that balancing methods inflate the predictive multiplicity and yield varying results. To monitor the trade-off between the prediction performance and predictive multiplicity for conducting the modeling process responsibly, we proposed using the extended version of the performance-gain plot when balancing the training data.
翻译:在对不平衡数据集进行分类时,预测模型可能产生有偏差的预测。当模型倾向于多数类时,会导致对少数类的准确预测性能低下。为解决此问题,平衡或重采样方法作为建模过程中以数据为中心的关键人工智能方法,对于提升预测性能至关重要。然而,近年来关于这些方法的功能一直存在争议和疑问。特别是在模型选择中,许多候选模型可能表现出非常相似的预测性能(称为拉希蒙效应),它们甚至可能对相同的观测样本产生不同的预测。若在选择模型时未考虑预测多重性——即对于任何样本可能产生相互冲突的模型预测的情况——则可能导致盲目选择。本文通过拉希蒙效应研究了平衡方法对预测多重性的影响。这一点至关重要,因为在以数据为中心的人工智能中,从一组近似同等准确的模型中进行盲目选择具有风险,可能导致模型选择、验证和解释方面的严重问题。为解决此问题,我们通过真实数据集实验,除使用现有指标(模糊性和差异性)外,还采用新提出的混淆度指标,观察平衡方法通过拉希蒙效应对预测多重性的影响。我们的研究结果表明,平衡方法会加剧预测多重性并产生不同的结果。为在建模过程中负责任地权衡预测性能与预测多重性,我们建议在平衡训练数据时使用扩展版的性能增益图。