Abstract
Time Series Representation Learning (TSRL) focuses on generating informative representations for various Time Series (TS) modeling tasks. Traditional Self-Supervised Learning (SSL) methods in TSRL fall into four main categories: reconstructive, adversarial, contrastive, and predictive, each with a common challenge of sensitivity to noise and intricate data nuances. Recently, diffusion-based methods have shown advanced generative capabilities. However, they primarily target specific application scenarios like imputation and forecasting, leaving a gap in leveraging diffusion models for generic TSRL. Our work, Time Series Diffusion Embedding (TSDE), bridges this gap as the first diffusion-based SSL TSRL approach. TSDE segments TS data into observed and masked parts using an Imputation-Interpolation-Forecasting (IIF) mask. It applies a trainable embedding function, featuring dual-orthogonal Transformer encoders with a crossover mechanism, to the observed part. We train a reverse diffusion process conditioned on the embeddings, designed to predict noise added to the masked part. Extensive experiments demonstrate TSDE's superiority in imputation, interpolation, forecasting, anomaly detection, classification, and clustering. We also conduct an ablation study, present embedding visualizations, and compare inference speed, further substantiating TSDE's efficiency and validity in learning representations of TS data.
Keywords
Citation
@article{Senane2026SelfSupervised,
title={Self-Supervised Learning of Time Series Representation via Diffusion Process and Imputation-Interpolation-Forecasting Mask},
author={Zineb Senane and Yusuke Tashiro and Mats Nordahl and Lele Cao and Lei You and Ruibo Tu and Valentin Leonhard Buchner and Pawel Andrzej Herman and Vilhelm von Ehrenheim},
year={2026},
url={https://cspaper.org/openprint/20260403.0001v1},
journal={OpenPrint:20260403.0001v1}
}Version History
| Version | Archived Date | Submitter |
|---|---|---|
v1Current | Apr 3, 2026 | Zineb Senane |
