FlashKAT: Understanding and Addressing Performance Bottlenecks in the Kolmogorov-Arnold Transformer

Authors

  • Matthew Raffel Oregon State University
  • Lizhong Chen Oregon State University

DOI:

https://doi.org/10.1609/aaai.v40i11.37822

Abstract

The Kolmogorov-Arnold Network (KAN) has been gaining popularity as an alternative to the multilayer perceptron (MLP) due to its greater expressiveness and interpretability. Even so, KAN suffers from training instability and being orders of magnitude slower due to its increased computational cost, limiting its applicability to large-scale tasks. Recently, the Kolmogorov-Arnold Transformer (KAT) has been proposed, achieving FLOPs comparable to traditional Transformer models with MLPs by leveraging Group-Rational KAN (GR-KAN). Unfortunately, despite the comparable FLOPs, our testing shows that KAT remains 123x slower during training, indicating that there are other performance bottlenecks beyond FLOPs. In this paper, we conduct a series of experiments to understand the root cause of the slowdown in KAT. We uncover that the slowdown can be isolated to memory stalls, linked more specifically to inefficient gradient accumulations in the backward pass of GR-KAN. To address this memory bottleneck, we propose FlashKAT, which minimizes accesses to slow memory and the usage of atomic adds through a restructured kernel. Evaluations show that FlashKAT achieves up to an 86.5x training speedup over state-of-the-art KAT while reducing rounding errors in gradient computation.

Published

2026-03-14

How to Cite

Raffel, M., & Chen, L. (2026). FlashKAT: Understanding and Addressing Performance Bottlenecks in the Kolmogorov-Arnold Transformer. Proceedings of the AAAI Conference on Artificial Intelligence, 40(11), 8694–8702. https://doi.org/10.1609/aaai.v40i11.37822

Issue

Section

AAAI Technical Track on Computer Vision VIII