We present FairX, an open-source Python-based benchmarking tool designed for the comprehensive analysis of models under the umbrella of fairness, utility, and eXplainability (XAI). FairX enables users to train benchmarking bias-removal models and evaluate their fairness using a wide array of fairness metrics, data utility metrics, and generate explanations for model predictions, all within a unified framework. Existing benchmarking tools do not have the way to evaluate synthetic data generated from fair generative models, also they do not have the support for training fair generative models either. In FairX, we add fair generative models in the collection of our fair-model library (pre-processing, in-processing, post-processing) and evaluation metrics for evaluating the quality of synthetic fair data. This version of FairX supports both tabular and image datasets. It also allows users to provide their own custom datasets. The open-source FairX benchmarking package is publicly available at https://github.com/fahim-sikder/FairX.
翻译:本文提出FairX,一个基于Python的开源基准测试工具,旨在从公平性、效用与可解释性(XAI)的综合视角对模型进行全面分析。FairX使用户能够在统一框架内训练基准去偏模型,利用多种公平性指标和数据效用指标评估其公平性,并为模型预测生成解释。现有基准测试工具既无法评估公平生成模型产生的合成数据,也不支持训练公平生成模型。在FairX中,我们在公平模型库(预处理、处理中、后处理)中增加了公平生成模型,并提供了用于评估合成公平数据质量的评估指标。当前版本的FairX同时支持表格数据和图像数据集,并允许用户使用自定义数据集。该开源基准测试工具包已在https://github.com/fahim-sikder/FairX 公开提供。