Unlocking the Power of LSTM for Long Term Time Series Forecasting

Authors

  • Yaxuan Kong University of Oxford
  • Zepu Wang The Hong Kong University of Science and Technology (Guangzhou) Duke Kunshan University
  • Yuqi Nie Princeton University
  • Tian Zhou Alibaba Group
  • Stefan Zohren University of Oxford
  • Yuxuan Liang The Hong Kong University of Science and Technology (Guangzhou)
  • Peng Sun Duke Kunshan University
  • Qingsong Wen Squirrel AI

DOI:

https://doi.org/10.1609/aaai.v39i11.33303

Abstract

Traditional recurrent neural network architectures, such as long short-term memory neural networks (LSTM), have historically held a prominent role in time series forecasting (TSF) tasks. While the recently introduced sLSTM for Natural Language Processing (NLP) introduces exponential gating and memory mixing that are beneficial for long term sequential learning, its potential short memory issue is a barrier to applying sLSTM directly in TSF. To address this, we propose a simple yet efficient algorithm named P-sLSTM, which is built upon sLSTM by incorporating patching and channel independence. These modifications substantially enhance sLSTM's performance in TSF, achieving state-of-the-art results. Furthermore, we provide theoretical justifications for our design, and conduct extensive comparative and analytical experiments to fully validate the efficiency and superior performance of our model.

Published

2025-04-11

How to Cite

Kong, Y., Wang, Z., Nie, Y., Zhou, T., Zohren, S., Liang, Y., Sun, P., & Wen, Q. (2025). Unlocking the Power of LSTM for Long Term Time Series Forecasting. Proceedings of the AAAI Conference on Artificial Intelligence, 39(11), 11968-11976. https://doi.org/10.1609/aaai.v39i11.33303

Issue

Section

AAAI Technical Track on Data Mining & Knowledge Management I