Learning Transferable Time Series Classifier with Cross-Domain Pre-training from Language Model
arxiv(2024)
摘要
Advancements in self-supervised pre-training (SSL) have significantly
advanced the field of learning transferable time series representations, which
can be very useful in enhancing the downstream task. Despite being effective,
most existing works struggle to achieve cross-domain SSL pre-training, missing
valuable opportunities to integrate patterns and features from different
domains. The main challenge lies in the significant differences in the
characteristics of time-series data across different domains, such as
variations in the number of channels and temporal resolution scales. To address
this challenge, we propose CrossTimeNet, a novel cross-domain SSL learning
framework to learn transferable knowledge from various domains to largely
benefit the target downstream task. One of the key characteristics of
CrossTimeNet is the newly designed time series tokenization module, which could
effectively convert the raw time series into a sequence of discrete tokens
based on a reconstruction optimization process. Besides, we highlight that
predicting a high proportion of corrupted tokens can be very helpful for
extracting informative patterns across different domains during SSL
pre-training, which has been largely overlooked in past years. Furthermore,
unlike previous works, our work treats the pre-training language model (PLM) as
the initialization of the encoder network, investigating the feasibility of
transferring the knowledge learned by the PLM to the time series area. Through
these efforts, the path to cross-domain pre-training of a generic time series
model can be effectively paved. We conduct extensive experiments in a
real-world scenario across various time series classification domains. The
experimental results clearly confirm CrossTimeNet's superior performance.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要