MCL: Multi-Granularity Contrastive Learning Framework for Chinese NER


  • Shan Zhao Hefei University of Technology
  • ChengYu Wang Hefei University of Technology National University of Defense Technology
  • Minghao Hu Information Research Center of Military Science, PLA Academy of Military Science
  • Tianwei Yan National University of Defense Technology
  • Meng Wang Hefei University of Technology



SNLP: Applications, SNLP: Information Extraction


Recently, researchers have applied the word-character lattice framework to integrated word information, which has become very popular for Chinese named entity recognition (NER). However, prior approaches fuse word information by different variants of encoders such as Lattice LSTM or Flat-Lattice Transformer, but are still not data-efficient indeed to fully grasp the depth interaction of cross-granularity and important word information from the lexicon. In this paper, we go beyond the typical lattice structure and propose a novel Multi-Granularity Contrastive Learning framework (MCL), that aims to optimize the inter-granularity distribution distance and emphasize the critical matched words in the lexicon. By carefully combining cross-granularity contrastive learning and bi-granularity contrastive learning, the network can explicitly leverage lexicon information on the initial lattice structure, and further provide more dense interactions of across-granularity, thus significantly improving model performance. Experiments on four Chinese NER datasets show that MCL obtains state-of-the-art results while considering model efficiency. The source code of the proposed method is publicly available at




How to Cite

Zhao, S., Wang, C., Hu, M., Yan, T., & Wang, M. (2023). MCL: Multi-Granularity Contrastive Learning Framework for Chinese NER. Proceedings of the AAAI Conference on Artificial Intelligence, 37(11), 14011-14019.



AAAI Technical Track on Speech & Natural Language Processing