In many real-world scenarios, acquiring all features of a data instance can be expensive or impractical due to monetary cost, latency, or privacy concerns. Active Feature Acquisition (AFA) addresses this challenge by dynamically selecting a subset of informative features for each data instance, trading predictive performance against acquisition cost. While numerous methods have been proposed for AFA, ranging from myopic information-theoretic strategies to non-myopic reinforcement learning approaches, fair and systematic evaluation of these methods has been hindered by a lack of standardized benchmarks. In this paper, we introduce AFABench, the first benchmark framework for AFA. Our benchmark includes a diverse set of synthetic and real-world datasets, supports a wide range of acquisition policies, and provides a modular design that enables easy integration of new methods and tasks. We implement and evaluate representative algorithms from all major categories, including static, myopic, and reinforcement learning-based approaches. To test the lookahead capabilities of AFA policies, we introduce a novel synthetic dataset, CUBE-NM, designed to expose the limitations of myopic selection. Our results highlight key trade-offs between different AFA strategies and provide actionable insights for future research. The benchmark code is available at: https://github.com/Linusaronsson/AFA-Benchmark.
翻译:在许多现实场景中,由于成本、延迟或隐私考虑,获取数据实例的所有特征可能代价高昂或不切实际。主动特征获取通过为每个数据实例动态选择信息丰富的特征子集来解决这一挑战,在预测性能与获取成本之间进行权衡。尽管已提出了从短视的信息论策略到非短视的强化学习方法等多种AFA方法,但由于缺乏标准化基准,这些方法的公平和系统评估一直受到阻碍。本文介绍了首个AFA基准框架AFABench。该基准包含多样化的合成与真实数据集,支持广泛的获取策略,并提供模块化设计以方便新方法和任务的集成。我们实现并评估了所有主要类别的代表性算法,包括静态方法、短视方法和基于强化学习的方法。为测试AFA策略的前瞻能力,我们引入了一个新颖的合成数据集CUBE-NM,其设计旨在揭示短视选择的局限性。实验结果凸显了不同AFA策略间的关键权衡,并为未来研究提供了可行见解。基准代码已发布于:https://github.com/Linusaronsson/AFA-Benchmark。