Explainable Recommendation through Attentive Multi-View Learning


  • Jingyue Gao Peking University
  • Xiting Wang Microsoft Research Asia
  • Yasha Wang Peking University
  • Xing Xie Microsoft Research Asia




Recommender systems have been playing an increasingly important role in our daily life due to the explosive growth of information. Accuracy and explainability are two core aspects when we evaluate a recommendation model and have become one of the fundamental trade-offs in machine learning. In this paper, we propose to alleviate the trade-off between accuracy and explainability by developing an explainable deep model that combines the advantages of deep learning-based models and existing explainable methods. The basic idea is to build an initial network based on an explainable deep hierarchy (e.g., Microsoft Concept Graph) and improve the model accuracy by optimizing key variables in the hierarchy (e.g., node importance and relevance). To ensure accurate rating prediction, we propose an attentive multi-view learning framework. The framework enables us to handle sparse and noisy data by co-regularizing among different feature levels and combining predictions attentively. To mine readable explanations from the hierarchy, we formulate personalized explanation generation as a constrained tree node selection problem and propose a dynamic programming algorithm to solve it. Experimental results show that our model outperforms state-of-the-art methods in terms of both accuracy and explainability.




How to Cite

Gao, J., Wang, X., Wang, Y., & Xie, X. (2019). Explainable Recommendation through Attentive Multi-View Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 33(01), 3622-3629. https://doi.org/10.1609/aaai.v33i01.33013622



AAAI Technical Track: Machine Learning