Unifying Model Explainability and Robustness for Joint Text Classification and Rationale Extraction
Keywords:Speech & Natural Language Processing (SNLP)
AbstractRecent works have shown explainability and robustness are two crucial ingredients of trustworthy and reliable text classification. However, previous works usually address one of two aspects: i) how to extract accurate rationales for explainability while being beneficial to prediction; ii) how to make the predictive model robust to different types of adversarial attacks. Intuitively, a model that produces helpful explanations should be more robust against adversarial attacks, because we cannot trust the model that outputs explanations but changes its prediction under small perturbations. To this end, we propose a joint classification and rationale extraction model named AT-BMC. It includes two key mechanisms: mixed Adversarial Training (AT) is designed to use various perturbations in discrete and embedding space to improve the model’s robustness, and Boundary Match Constraint (BMC) helps to locate rationales more precisely with the guidance of boundary information. Performances on benchmark datasets demonstrate that the proposed AT-BMC outperforms baselines on both classification and rationale extraction by a large margin. Robustness analysis shows that the proposed AT-BMC decreases the attack success rate effectively by up to 69%. The results indicate that there are connections between robust models and better explanations.
How to Cite
Li, D., Hu, B., Chen, Q., Xu, T., Tao, J., & Zhang, Y. (2022). Unifying Model Explainability and Robustness for Joint Text Classification and Rationale Extraction. Proceedings of the AAAI Conference on Artificial Intelligence, 36(10), 10947-10955. https://doi.org/10.1609/aaai.v36i10.21342
AAAI Technical Track on Speech and Natural Language Processing