Enhancing Masked Time-Series Modeling via Dropping Patches
DOI:
https://doi.org/10.1609/aaai.v39i19.34211Abstract
This paper explores how to enhance existing masked time-series modeling by randomly dropping sub-sequence level patches of time series. On this basis, a simple yet effective method named DropPatch is proposed, which has two remarkable advantages: 1) It improves the pre-training efficiency by a square-level advantage; 2) It provides additional advantages for modeling in scenarios such as in-domain, cross-domain, few-shot learning and cold start. This paper conducts comprehensive experiments to verify the effectiveness of the method and analyze its internal mechanism. Empirically, DropPatch strengthens the attention mechanism, reduces information redundancy and serves as an efficient means of data augmentation. Theoretically, it is proved that DropPatch slows down the rate at which the Transformer representations collapse into the rank-1 linear subspace by randomly dropping patches, thus optimizing the quality of the learned representations.Downloads
Published
2025-04-11
How to Cite
Qiu, T., Xie, Y., Niu, H., Xiong, Y., & Gao, X. (2025). Enhancing Masked Time-Series Modeling via Dropping Patches. Proceedings of the AAAI Conference on Artificial Intelligence, 39(19), 20077–20085. https://doi.org/10.1609/aaai.v39i19.34211
Issue
Section
AAAI Technical Track on Machine Learning V