A New Formula for Sticker Retrieval: Reply with Stickers in Multi-Modal and Multi-Session Conversation

Authors

  • Bingbing Wang Harbin Institute of Technology, Shenzhen, China Guangdong Provincial Key Laboratory of Novel Security Intelligence Technologies, Shenzhen, China
  • Yiming Du The Chinese University of Hong Kong, Hong Kong, China
  • Bin Liang The Chinese University of Hong Kong, Hong Kong, China
  • Zhixin Bai Harbin Institute of Technology, Shenzhen, China
  • Min Yang Shenzhen Institutes of Advanced Technology Chinese Academy of Sciences, Shenzhen, China
  • Baojun Wang Huawei Noah’s Ark Lab, Shenzhen, China
  • Kam-Fai Wong The Chinese University of Hong Kong, Hong Kong, China
  • Ruifeng Xu Harbin Institute of Technology, Shenzhen, China Guangdong Provincial Key Laboratory of Novel Security Intelligence Technologies, Shenzhen, China Peng Cheng Laboratory, Shenzhen, China

DOI:

https://doi.org/10.1609/aaai.v39i24.34720

Abstract

Stickers are widely used in online chatting, which can vividly express someone's intention, emotion, or attitude. Existing conversation research typically retrieves stickers based on a single session or the previous textual information, which can not adapt to the multi-modal and multi-session nature of the real-world conversation. To this end, we introduce MultiChat, a new dataset for sticker retrieval facing the multi-modal and multi-session conversation, comprising 1,542 sessions, featuring 50,192 utterances and 2,182 stickers. Based on the created dataset, we propose a novel Intent-Guided Sticker Retrieval (IGSR) framework that retrieves stickers for multi-modal and multi-session conversation history drawing support from intent learning. Specifically, we introduce sticker attributes to better leverage the sticker information in multi-modal conversation, which are incorporated with utterances to construct a memory bank. Further, we extract relevant memories for the current conversation from the memory bank to identify the intent of the current conversation, and then retrieve a sticker to respond guided by the intent. Extensive experiments on our MultiChat dataset reveal the robustness and effectiveness of our IGSR approach in multi-session, multi-modal scenarios.

Downloads

Published

2025-04-11

How to Cite

Wang, B., Du, Y., Liang, B., Bai, Z., Yang, M., Wang, B., … Xu, R. (2025). A New Formula for Sticker Retrieval: Reply with Stickers in Multi-Modal and Multi-Session Conversation. Proceedings of the AAAI Conference on Artificial Intelligence, 39(24), 25327–25335. https://doi.org/10.1609/aaai.v39i24.34720

Issue

Section

AAAI Technical Track on Natural Language Processing III