Enhancing Masked Time-Series Modeling via Dropping Patches

Authors

  • Tianyu Qiu Fudan University
  • Yi Xie Fudan University
  • Hao Niu Fudan University
  • Yun Xiong Fudan University
  • Xiaofeng Gao Shanghai Jiao Tong University

DOI:

https://doi.org/10.1609/aaai.v39i19.34211

Abstract

This paper explores how to enhance existing masked time-series modeling by randomly dropping sub-sequence level patches of time series. On this basis, a simple yet effective method named DropPatch is proposed, which has two remarkable advantages: 1) It improves the pre-training efficiency by a square-level advantage; 2) It provides additional advantages for modeling in scenarios such as in-domain, cross-domain, few-shot learning and cold start. This paper conducts comprehensive experiments to verify the effectiveness of the method and analyze its internal mechanism. Empirically, DropPatch strengthens the attention mechanism, reduces information redundancy and serves as an efficient means of data augmentation. Theoretically, it is proved that DropPatch slows down the rate at which the Transformer representations collapse into the rank-1 linear subspace by randomly dropping patches, thus optimizing the quality of the learned representations.

Downloads

Published

2025-04-11

How to Cite

Qiu, T., Xie, Y., Niu, H., Xiong, Y., & Gao, X. (2025). Enhancing Masked Time-Series Modeling via Dropping Patches. Proceedings of the AAAI Conference on Artificial Intelligence, 39(19), 20077–20085. https://doi.org/10.1609/aaai.v39i19.34211

Issue

Section

AAAI Technical Track on Machine Learning V