In the RAG paradigm, the information retrieval module provides context for generators by retrieving and ranking multiple documents to support the aggregation of evidence. However, existing ranking models are primarily optimized for query--document relevance, which often misaligns with generators' preferences for evidence selection and citation, limiting their impact on response quality. Moreover, most approaches do not account for preference differences across generators, resulting in unstable cross-generator performance. We propose \textbf{Rank4Gen}, a generator-aware ranker for RAG that targets the goal of \emph{Ranking for Generators}. Rank4Gen introduces two key preference modeling strategies: (1) \textbf{From Ranking Relevance to Response Quality}, which optimizes ranking with respect to downstream response quality rather than query--document relevance; and (2) \textbf{Generator-Specific Preference Modeling}, which conditions a single ranker on different generators to capture their distinct ranking preferences. To enable such modeling, we construct \textbf{PRISM}, a dataset built from multiple open-source corpora and diverse downstream generators. Experiments on five challenging and recent RAG benchmarks demonstrate that RRank4Gen achieves strong and competitive performance for complex evidence composition in RAG.
翻译:在RAG范式中,信息检索模块通过检索并排序多篇文档来为生成器提供上下文,以支持证据的聚合。然而,现有排序模型主要针对查询-文档相关性进行优化,这常常与生成器在证据选择和引用方面的偏好不一致,从而限制了其对响应质量的提升效果。此外,大多数方法未考虑不同生成器之间的偏好差异,导致跨生成器性能不稳定。我们提出 **Rank4Gen**,一种面向RAG的生成器感知排序器,其目标在于实现**面向生成器的排序**。Rank4Gen引入了两种关键的偏好建模策略:(1)**从排序相关性到响应质量**,即依据下游响应质量而非查询-文档相关性来优化排序;(2)**生成器特定偏好建模**,使单一排序器能够适配不同生成器,以捕捉其独特的排序偏好。为支持此类建模,我们构建了 **PRISM** 数据集,该数据集基于多个开源语料库和多样化的下游生成器构建而成。在五个具有挑战性的近期RAG基准测试上的实验表明,Rank4Gen在RAG中复杂证据组合任务上实现了强劲且具有竞争力的性能。