Advancing Loss Functions in Recommender Systems: A Comparative Study with a Rényi Divergence-Based Solution

Authors

  • Shengjia Zhang State Key Laboratory of Blockchain and Data Security, Zhejiang University College of Computer Science, Zhejiang University, China
  • Jiawei Chen State Key Laboratory of Blockchain and Data Security, Zhejiang University College of Computer Science, Zhejiang University, China Hangzhou High-Tech Zone (Binjiang) Institute of Blockchain and Data Security
  • Changdong Li College of Computer Science, Zhejiang University, China
  • Sheng Zhou College of Computer Science, Zhejiang University, China
  • Qihao Shi College of Computer Science, Zhejiang University, China
  • Yan Feng State Key Laboratory of Blockchain and Data Security, Zhejiang University College of Computer Science, Zhejiang University, China
  • Chun Chen State Key Laboratory of Blockchain and Data Security, Zhejiang University College of Computer Science, Zhejiang University, China
  • Can Wang State Key Laboratory of Blockchain and Data Security, Zhejiang University Hangzhou High-Tech Zone (Binjiang) Institute of Blockchain and Data Security

DOI:

https://doi.org/10.1609/aaai.v39i12.33450

Abstract

Loss functions play a pivotal role in optimizing recommendation models. Among various loss functions, Softmax Loss (SL) and Cosine Contrastive Loss (CCL) are particularly effective. Their theoretical connections and differences warrant in-depth exploration. This work conducts comprehensive analyses of these losses, yielding significant insights: 1) Common strengths --- both can be viewed as augmentations of traditional losses with Distributional Robust Optimization (DRO), enhancing robustness to distributional shifts; 2) Respective limitations --- stemming from their use of different distribution distance metrics in DRO optimization, SL exhibits high sensitivity to false negative instances, whereas CCL suffers from low data utilization. To address these limitations, this work proposes a new loss function, DrRL, which generalizes SL and CCL by leveraging Rényi-divergence in DRO optimization. DrRL incorporates the advantageous structures of both SL and CCL, and can be demonstrated to effectively mitigate their limitations. Extensive experiments have been conducted to validate the superiority of DrRL on both recommendation accuracy and robustness.

Downloads

Published

2025-04-11

How to Cite

Zhang, S., Chen, J., Li, C., Zhou, S., Shi, Q., Feng, Y., … Wang, C. (2025). Advancing Loss Functions in Recommender Systems: A Comparative Study with a Rényi Divergence-Based Solution. Proceedings of the AAAI Conference on Artificial Intelligence, 39(12), 13286–13294. https://doi.org/10.1609/aaai.v39i12.33450

Issue

Section

AAAI Technical Track on Data Mining & Knowledge Management II