Complexity of Neural Network Training and ETR: Extensions with Effectively Continuous Functions

Authors

  • Teemu Hankala University of Helsinki
  • Miika Hannula University of Helsinki
  • Juha Kontinen University of Helsinki
  • Jonni Virtema University of Sheffield

DOI:

https://doi.org/10.1609/aaai.v38i11.29118

Keywords:

ML: Deep Neural Architectures and Foundation Models, ML: Deep Learning Theory

Abstract

The training problem of neural networks (NNs) is known to be ER-complete with respect to ReLU and linear activation functions. We show that the training problem for NNs equipped with arbitrary activation functions is polynomial-time bireducible to the existential theory of the reals extended with the corresponding activation functions. For effectively continuous activation functions (e.g., the sigmoid function), we obtain an inclusion to low levels of the arithmetical hierarchy. Consequently, the sigmoid activation function leads to the existential theory of the reals with the exponential function, and hence the decidability of training NNs using the sigmoid activation function is equivalent to the decidability of the existential theory of the reals with the exponential function, a long-standing open problem. In contrast, we obtain that the training problem is undecidable if sinusoidal activation functions are considered.

Published

2024-03-24

How to Cite

Hankala, T., Hannula, M., Kontinen, J., & Virtema, J. (2024). Complexity of Neural Network Training and ETR: Extensions with Effectively Continuous Functions. Proceedings of the AAAI Conference on Artificial Intelligence, 38(11), 12278-12285. https://doi.org/10.1609/aaai.v38i11.29118

Issue

Section

AAAI Technical Track on Machine Learning II