TS2Vec: Towards Universal Representation of Time Series

Authors

  • Zhihan Yue Peking University Microsoft
  • Yujing Wang Microsoft Peking University
  • Juanyong Duan Microsoft
  • Tianmeng Yang Peking University Microsoft
  • Congrui Huang Microsoft
  • Yunhai Tong Peking University
  • Bixiong Xu Microsoft

DOI:

https://doi.org/10.1609/aaai.v36i8.20881

Keywords:

Machine Learning (ML)

Abstract

This paper presents TS2Vec, a universal framework for learning representations of time series in an arbitrary semantic level. Unlike existing methods, TS2Vec performs contrastive learning in a hierarchical way over augmented context views, which enables a robust contextual representation for each timestamp. Furthermore, to obtain the representation of an arbitrary sub-sequence in the time series, we can apply a simple aggregation over the representations of corresponding timestamps. We conduct extensive experiments on time series classification tasks to evaluate the quality of time series representations. As a result, TS2Vec achieves significant improvement over existing SOTAs of unsupervised time series representation on 125 UCR datasets and 29 UEA datasets. The learned timestamp-level representations also achieve superior results in time series forecasting and anomaly detection tasks. A linear regression trained on top of the learned representations outperforms previous SOTAs of time series forecasting. Furthermore, we present a simple way to apply the learned representations for unsupervised anomaly detection, which establishes SOTA results in the literature. The source code is publicly available at https://github.com/yuezhihan/ts2vec.

Downloads

Published

2022-06-28

How to Cite

Yue, Z., Wang, Y., Duan, J., Yang, T., Huang, C., Tong, Y., & Xu, B. (2022). TS2Vec: Towards Universal Representation of Time Series. Proceedings of the AAAI Conference on Artificial Intelligence, 36(8), 8980-8987. https://doi.org/10.1609/aaai.v36i8.20881

Issue

Section

AAAI Technical Track on Machine Learning III