Conventional recommendation systems (RSs) are typically optimized to enhance performance metrics uniformly across all training samples, inadvertently overlooking the needs of diverse user populations. The performance disparity among various populations can harm the model's robustness to sub-populations due to the varying user properties. While large language models (LLMs) show promise in enhancing RS performance, their practical applicability is hindered by high costs, inference latency, and degraded performance on long user queries. To address these challenges, we propose a hybrid task allocation framework designed to promote social good by equitably serving all user groups. By adopting a two-phase approach, we promote a strategic assignment of tasks for efficient and responsible adaptation of LLMs. Our strategy works by first identifying the weak and inactive users that receive a suboptimal ranking performance by RSs. Next, we use an in-context learning approach for such users, wherein each user interaction history is contextualized as a distinct ranking task. We evaluate our hybrid framework by incorporating eight different recommendation algorithms and three different LLMs -- both open and close-sourced. Our results on three real-world datasets show a significant reduction in weak users and improved robustness to subpopulations without disproportionately escalating costs.
翻译:传统推荐系统通常以优化所有训练样本上的性能指标为目标,这种统一优化方式无意中忽视了多样化用户群体的需求。由于用户属性存在差异,不同群体间的性能差距可能损害模型对于子群体的稳健性。尽管大语言模型在提升推荐系统性能方面展现出潜力,但其实际应用受到高成本、推理延迟以及对长用户查询性能下降等因素的制约。为应对这些挑战,我们提出一种混合任务分配框架,旨在通过公平服务所有用户群体来促进社会效益。该框架采用两阶段方法,通过策略性任务分配实现大语言模型的高效且负责任适应。我们的策略首先识别那些在推荐系统中获得次优排序表现的弱势与非活跃用户,随后对此类用户采用上下文学习方法——将每个用户的交互历史情境化为独立的排序任务进行评估。我们通过整合八种不同推荐算法与三种开源及闭源大语言模型对混合框架进行评估。在三个真实世界数据集上的实验结果表明,该方法能显著减少弱势用户数量,提升对子群体的稳健性,且未造成成本的不成比例增加。