RpBERT: A Text-image Relation Propagation-based BERT Model for Multimodal NER
AbstractRecently multimodal named entity recognition (MNER) has utilized images to improve the accuracy of NER in tweets. However, most of the multimodal methods use attention mechanisms to extract visual clues regardless of whether the text and image are relevant. Practically, the irrelevant text-image pairs account for a large proportion in tweets. The visual clues that are unrelated to the texts will exert uncertain or even negative effects on multimodal model learning. In this paper, we introduce a method of text-image relation propagation into the multimodal BERT model. We integrate soft or hard gates to select visual clues and propose a multitask algorithm to train and validate the effects of relation propagation on the MNER datasets. In the experiments, we deeply analyze the changes in visual attention before and after the use of relation propagation. Our model achieves state-of-the-art performance on the MNER datasets.
How to Cite
Sun, L. ., Wang, J., Zhang, K., Su, Y., & Weng, F. (2021). RpBERT: A Text-image Relation Propagation-based BERT Model for Multimodal NER. Proceedings of the AAAI Conference on Artificial Intelligence, 35(15), 13860-13868. Retrieved from https://ojs.aaai.org/index.php/AAAI/article/view/17633
AAAI Technical Track on Speech and Natural Language Processing II