Ladder Loss for Coherent Visual-Semantic Embedding


  • Mo Zhou Xidian University
  • Zhenxing Niu Alibaba Group
  • Le Wang Xi'an Jiaotong University
  • Zhanning Gao Alibaba Group
  • Qilin Zhang HERE Technologies
  • Gang Hua Wormpex AI Research



For visual-semantic embedding, the existing methods normally treat the relevance between queries and candidates in a bipolar way – relevant or irrelevant, and all “irrelevant” candidates are uniformly pushed away from the query by an equal margin in the embedding space, regardless of their various proximity to the query. This practice disregards relatively discriminative information and could lead to suboptimal ranking in the retrieval results and poorer user experience, especially in the long-tail query scenario where a matching candidate may not necessarily exist. In this paper, we introduce a continuous variable to model the relevance degree between queries and multiple candidates, and propose to learn a coherent embedding space, where candidates with higher relevance degrees are mapped closer to the query than those with lower relevance degrees. In particular, the new ladder loss is proposed by extending the triplet loss inequality to a more general inequality chain, which implements variable push-away margins according to respective relevance degrees. In addition, a proper Coherent Score metric is proposed to better measure the ranking results including those “irrelevant” candidates. Extensive experiments on multiple datasets validate the efficacy of our proposed method, which achieves significant improvement over existing state-of-the-art methods.




How to Cite

Zhou, M., Niu, Z., Wang, L., Gao, Z., Zhang, Q., & Hua, G. (2020). Ladder Loss for Coherent Visual-Semantic Embedding. Proceedings of the AAAI Conference on Artificial Intelligence, 34(07), 13050-13057.



AAAI Technical Track: Vision