Batch Normalization Is Blind to the First and Second Derivatives of the Loss
DOI:
https://doi.org/10.1609/aaai.v38i18.29978Keywords:
PEAI: Accountability, Interpretability & Explainability, ML: Deep Learning TheoryAbstract
We prove that when we do the Taylor series expansion of the loss function, the BN operation will block the influence of the first-order term and most influence of the second-order term of the loss. We also find that such a problem is caused by the standardization phase of the BN operation. We believe that proving the blocking of certain loss terms provides an analytic perspective for potential detects of a deep model with BN operations, although the blocking problem is not fully equivalent to significant damages in all tasks on benchmark datasets. Experiments show that the BN operation significantly affects feature representations in specific tasks.Downloads
Published
2024-03-24
How to Cite
Zhou, Z., Shen, W., Chen, H., Tang, L., Chen, Y., & Zhang, Q. (2024). Batch Normalization Is Blind to the First and Second Derivatives of the Loss. Proceedings of the AAAI Conference on Artificial Intelligence, 38(18), 20010-20018. https://doi.org/10.1609/aaai.v38i18.29978
Issue
Section
AAAI Technical Track on Philosophy and Ethics of AI