Towards Learning and Explaining Indirect Causal Effects in Neural Networks

Authors

  • Abbavaram Gowtham Reddy Indian Institute of Technology Hyderabad, India
  • Saketh Bachu Indian Institute of Technology Hyderabad, India
  • Harsharaj Pathak Indian Institute of Technology Hyderabad, India
  • Benin Godfrey L Indian Institute of Technology Hyderabad, India
  • Varshaneya V Honeywell, Bengaluru, India
  • Vineeth N Balasubramanian Indian Institute of Technology Hyderabad, India
  • Satyanarayan Kar Honeywell, Bengaluru, India

DOI:

https://doi.org/10.1609/aaai.v38i13.29399

Keywords:

ML: Deep Learning Algorithms, ML: Causal Learning

Abstract

Recently, there has been a growing interest in learning and explaining causal effects within Neural Network (NN) models. By virtue of NN architectures, previous approaches consider only direct and total causal effects assuming independence among input variables. We view an NN as a structural causal model (SCM) and extend our focus to include indirect causal effects by introducing feedforward connections among input neurons. We propose an ante-hoc method that captures and maintains direct, indirect, and total causal effects during NN model training. We also propose an algorithm for quantifying learned causal effects in an NN model and efficient approximation strategies for quantifying causal effects in high-dimensional data. Extensive experiments conducted on synthetic and real-world datasets demonstrate that the causal effects learned by our ante-hoc method better approximate the ground truth effects compared to existing methods.

Published

2024-03-24

How to Cite

Reddy, A. G., Bachu, S., Pathak, H., Godfrey L, B., V, V., N Balasubramanian, V., & Kar, S. (2024). Towards Learning and Explaining Indirect Causal Effects in Neural Networks. Proceedings of the AAAI Conference on Artificial Intelligence, 38(13), 14802-14810. https://doi.org/10.1609/aaai.v38i13.29399

Issue

Section

AAAI Technical Track on Machine Learning IV