Scalability is a major challenge in modern recommender systems. In sequential recommendations, full Cross-Entropy (CE) loss achieves state-of-the-art recommendation quality but consumes excessive GPU memory with large item catalogs, limiting its practicality. Using a GPU-efficient locality-sensitive hashing-like algorithm for approximating large tensor of logits, this paper introduces a novel RECE (REduced Cross-Entropy) loss. RECE significantly reduces memory consumption while allowing one to enjoy the state-of-the-art performance of full CE loss. Experimental results on various datasets show that RECE cuts training peak memory usage by up to 12 times compared to existing methods while retaining or exceeding performance metrics of CE loss. The approach also opens up new possibilities for large-scale applications in other domains.
翻译:可扩展性是现代推荐系统面临的主要挑战。在序列推荐任务中,全量交叉熵损失虽能实现最优的推荐质量,但在处理大规模商品目录时会导致GPU内存消耗过高,从而限制其实际应用。本文通过采用一种基于GPU高效局部敏感哈希的算法来近似计算大型逻辑张量,提出了一种新颖的RECE(简化交叉熵)损失函数。RECE在显著降低内存消耗的同时,使模型能够保持与全量交叉熵损失相当的最优性能。在不同数据集上的实验结果表明,与现有方法相比,RECE将训练峰值内存使用量最高降低了12倍,同时保持甚至超越了交叉熵损失的性能指标。该方法也为其他领域的大规模应用开辟了新的可能性。