Kolmogorov-Arnold Networks Still Catastrophically Forget but Differently from MLP

Authors

  • Anton Lee Victoria University of Wellington
  • Heitor Murilo Gomes Victoria University of Wellington
  • Yaqian Zhang University of Waikato
  • W. Bastiaan Kleijn Victoria University of Wellington

DOI:

https://doi.org/10.1609/aaai.v39i17.33986

Abstract

Catastrophic forgetting is when a neural network loses previously learnt information after learning a new task sequentially. Avoiding catastrophic forgetting could reduce the resources necessary to update neural networks. Recently, Kolmogorov–Arnold Networks (KAN) gained the community's attention as preliminary experiments suggest KAN avoid catastrophic forgetting. KAN replace neural network edges with learnable B-splines and sum incoming edges in nodes. Proponents of KAN argue they avoid forgetting, are more accurate, are interpretable, and use fewer parameters. Our work investigates the claims that KAN avoid catastrophic forgetting, finding that they fail to do so on more complex datasets containing features that overlap between tasks. We give a simple explanation as to why and how KAN catastrophically forget. Motivated by evidence suggesting KAN are superior for symbolic regression, we augment KAN in the same ways as multilayer perceptron (MLP) to perform continual learning tasks, making special accommodations to support KAN. Our experiments found that unmodified KAN often forget more than MLP, but KAN can be better than MLP when combined with continual learning strategies. We aim to highlight some of the current shortcomings and strengths associated with KAN for continual learning.

Published

2025-04-11

How to Cite

Lee, A., Gomes, H. M., Zhang, Y., & Kleijn, W. B. (2025). Kolmogorov-Arnold Networks Still Catastrophically Forget but Differently from MLP. Proceedings of the AAAI Conference on Artificial Intelligence, 39(17), 18053–18061. https://doi.org/10.1609/aaai.v39i17.33986

Issue

Section

AAAI Technical Track on Machine Learning III