RI-Loss: A Learnable Residual-Informed Loss for Time Series Forecasting

Authors

  • Jieting Wang Shanxi University
  • Xiaolei Shang Shanxi University
  • Feijiang Li Shanxi University
  • Furong Peng Shanxi University

DOI:

https://doi.org/10.1609/aaai.v40i31.39832

Abstract

Time series forecasting relies on predicting future values from historical data, yet most state-of-the-art approaches—including transformer and multilayer perceptron-based models—optimize using Mean Squared Error (MSE), which has two fundamental weaknesses: its point-wise error computation fails to capture temporal relationships, and it does not account for inherent noise in the data. To overcome these limitations, we introduce the Residual-Informed Loss (RI-Loss), a novel objective function based on the Hilbert-Schmidt Independence Criterion (HSIC). RI-Loss explicitly models noise structure by enforcing dependence between the residual sequence and a random time series, enabling more robust, noise-aware representations. Theoretically, we derive the first non-asymptotic HSIC bound with explicit double-sample complexity terms, achieving optimal convergence rates through Bernstein-type concentration inequalities and Rademacher complexity analysis. This provides rigorous guarantees for RI-Loss optimization while precisely quantifying kernel space interactions. Empirically, experiments across eight real-world benchmarks and five leading forecasting models demonstrate improvements in predictive performance, validating the effectiveness of our approach.

Published

2026-03-14

How to Cite

Wang, J., Shang, X., Li, F., & Peng, F. (2026). RI-Loss: A Learnable Residual-Informed Loss for Time Series Forecasting. Proceedings of the AAAI Conference on Artificial Intelligence, 40(31), 26277–26284. https://doi.org/10.1609/aaai.v40i31.39832

Issue

Section

AAAI Technical Track on Machine Learning VIII