Recent studies on scaling up ranking models have achieved substantial improvement for recommendation systems and search engines. However, most large-scale ranking systems rely on item IDs, where each item is treated as an independent categorical symbol and mapped to a learned embedding. As items rapidly appear and disappear, these embeddings become difficult to train and maintain. This instability impedes effective learning of neural network parameters and limits the scalability of ranking models. In this paper, we show that semantic tokens possess greater scaling potential compared to item IDs. Our proposed framework TRM improves the token generation and application pipeline, leading to 33% reduction in sparse storage while achieving 0.85% AUC increase. Extensive experiments further show that TRM could consistently outperform state-of-the-art models when model capacity scales. Finally, TRM has been successfully deployed on large-scale personalized search engines, yielding 0.26% and 0.75% improvement on user active days and change query ratio respectively through A/B test.
翻译:近期关于扩展排序模型的研究在推荐系统和搜索引擎领域取得了显著进步。然而,大多数大规模排序系统仍依赖于物品ID,即将每个物品视为独立的分类符号并映射到学习得到的嵌入向量。随着物品的快速出现和消失,这些嵌入向量变得难以训练和维护。这种不稳定性阻碍了神经网络参数的有效学习,并限制了排序模型的可扩展性。本文证明,与物品ID相比,语义标记具有更大的扩展潜力。我们提出的TRM框架改进了标记生成与应用流程,在实现AUC提升0.85%的同时,稀疏存储减少了33%。大量实验进一步表明,当模型容量扩展时,TRM能够持续超越最先进的模型。最终,TRM已成功部署于大规模个性化搜索引擎,通过A/B测试在用户活跃天数和查询变更率上分别实现了0.26%和0.75%的提升。