Large Language Models (LLMs) have shown significant potential for improving recommendation systems through their inherent reasoning capabilities and extensive knowledge base. Yet, existing studies predominantly address warm-start scenarios with abundant user-item interaction data, leaving the more challenging cold-start scenarios, where sparse interactions hinder traditional collaborative filtering methods, underexplored. To address this limitation, we propose novel reasoning strategies designed for cold-start item recommendations within the Netflix domain. Our method utilizes the advanced reasoning capabilities of LLMs to effectively infer user preferences, particularly for newly introduced or rarely interacted items. We systematically evaluate supervised fine-tuning, reinforcement learning-based fine-tuning, and hybrid approaches that combine both methods to optimize recommendation performance. Extensive experiments on real-world data demonstrate significant improvements in both methodological efficacy and practical performance in cold-start recommendation contexts. Remarkably, our reasoning-based fine-tuned models outperform Netflix's production ranking model by up to 8% in certain cases.
翻译:大语言模型(LLMs)凭借其固有的推理能力和广泛的知识库,在改进推荐系统方面展现出巨大潜力。然而,现有研究主要针对用户-物品交互数据丰富的热启动场景,而对更具挑战性的冷启动场景——稀疏交互阻碍了传统协同过滤方法——则探索不足。为应对这一局限,我们提出了针对Netflix领域冷启动物品推荐的新型推理策略。该方法利用LLMs的高级推理能力,有效推断用户偏好,特别是针对新引入或交互稀少的物品。我们系统评估了监督微调、基于强化学习的微调以及结合两种方法的混合策略,以优化推荐性能。在真实数据上进行的大量实验表明,该方法在冷启动推荐场景中的方法有效性和实际性能均取得显著提升。值得注意的是,我们基于推理的微调模型在某些情况下优于Netflix的生产排序模型,提升幅度高达8%。