Papers
Topics
Authors
Recent
Search
2000 character limit reached

ROSE: Register Assisted General Time Series Forecasting with Decomposed Frequency Learning

Published 24 May 2024 in cs.LG and stat.ML | (2405.17478v2)

Abstract: With the increasing collection of time series data from various domains, there arises a strong demand for general time series forecasting models pre-trained on a large number of time-series datasets to support a variety of downstream prediction tasks. Enabling general time series forecasting faces two challenges: how to obtain unified representations from multi-domian time series data, and how to capture domain-specific features from time series data across various domains for adaptive transfer in downstream tasks. To address these challenges, we propose a Register Assisted General Time Series Forecasting Model with Decomposed Frequency Learning (ROSE), a novel pre-trained model for time series forecasting. ROSE employs Decomposed Frequency Learning for the pre-training task, which decomposes coupled semantic and periodic information in time series with frequency-based masking and reconstruction to obtain unified representations across domains. We also equip ROSE with a Time Series Register, which learns to generate a register codebook to capture domain-specific representations during pre-training and enhances domain-adaptive transfer by selecting related register tokens on downstream tasks. After pre-training on large-scale time series data, ROSE achieves state-of-the-art forecasting performance on 8 real-world benchmarks. Remarkably, even in few-shot scenarios, it demonstrates competitive or superior performance compared to existing methods trained with full data.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. Tfb: Towards comprehensive and fair benchmarking of time series forecasting methods. arXiv preprint arXiv:2403.20150, 2024.
  2. Timer: Transformers for time series analysis at scale. arXiv preprint arXiv:2402.02368, 2024.
  3. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In International conference on machine learning, pages 27268–27286. PMLR, 2022.
  4. Self-supervised contrastive pre-training for time series via time-frequency consistency. Advances in Neural Information Processing Systems, 35:3988–4003, 2022.
  5. Simmtm: A simple pre-training framework for masked time-series modeling. Advances in Neural Information Processing Systems, 36, 2024.
  6. A time series is worth 64 words: Long-term forecasting with transformers. arXiv preprint arXiv:2211.14730, 2022.
  7. Learning to embed time series patches independently. arXiv preprint arXiv:2312.16427, 2023.
  8. Unified training of universal time series forecasting transformers. arXiv preprint arXiv:2402.02592, 2024.
  9. Knowledge-inspired subdomain adaptation for cross-domain knowledge transfer. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, pages 234–244, 2023.
  10. One fits all: Power general time series analysis by pretrained lm. Advances in neural information processing systems, 36, 2024.
  11. Some recent advances in forecasting and control. Journal of the Royal Statistical Society. Series C (Applied Statistics), 17(2):91–109, 1968.
  12. Graph attention recurrent neural networks for correlated time series forecasting. In MileTS19@KDD, 2019.
  13. A multi-horizon quantile recurrent forecaster. arXiv preprint arXiv:1711.11053, 2017.
  14. Deepar: Probabilistic forecasting with autoregressive recurrent networks. International journal of forecasting, 36(3):1181–1191, 2020.
  15. Moderntcn: A modern pure convolution structure for general time series analysis. In The Twelfth International Conference on Learning Representations, 2024.
  16. Scinet: Time series modeling and forecasting with sample convolution and interaction. Advances in Neural Information Processing Systems, 35:5816–5828, 2022a.
  17. Micn: Multi-scale local and global context modeling for long-term series forecasting. In The Eleventh International Conference on Learning Representations, 2022a.
  18. Timesnet: Temporal 2d-variation modeling for general time series analysis. In The eleventh international conference on learning representations, 2022.
  19. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in neural information processing systems, 34:22419–22430, 2021.
  20. itransformer: Inverted transformers are effective for time series forecasting. arXiv preprint arXiv:2310.06625, 2023.
  21. Pathformer: Multi-scale transformers with adaptive pathways for time series forecasting. 2024.
  22. Ti-mae: Self-supervised masked time series autoencoders. arXiv preprint arXiv:2301.08871, 2023.
  23. A transformer-based framework for multivariate time series representation learning. In Proceedings of the 27th ACM SIGKDD conference on knowledge discovery & data mining, pages 2114–2124, 2021.
  24. Ts2vec: Towards universal representation of time series. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 8980–8987, 2022.
  25. Time-series representation learning via temporal and contextual contrasting. arXiv preprint arXiv:2106.14112, 2021.
  26. Cost: Contrastive learning of disentangled seasonal-trend representations for time series forecasting. arXiv preprint arXiv:2202.01575, 2022.
  27. Learning latent seasonal-trend representations for time series forecasting. Advances in Neural Information Processing Systems, 35:38775–38787, 2022b.
  28. Lag-llama: Towards foundation models for time series forecasting. arXiv preprint arXiv:2310.08278, 2023.
  29. Forecastpfn: Synthetically-trained zero-shot forecasting. Advances in Neural Information Processing Systems, 36, 2024.
  30. Timegpt-1. arXiv preprint arXiv:2310.03589, 2023.
  31. Moment: A family of open time-series foundation models. arXiv preprint arXiv:2402.03885, 2024.
  32. A decoder-only foundation model for time-series forecasting. arXiv preprint arXiv:2310.10688, 2023a.
  33. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  34. E Oran Brigham and RE Morrow. The fast fourier transform. IEEE spectrum, 4(12):63–70, 1967.
  35. Categorical reparameterization with gumbel-softmax. arXiv preprint arXiv:1611.01144, 2016.
  36. Discrete representations strengthen vision transformer robustness. arXiv preprint arXiv:2111.10493, 2021.
  37. Intrinsic dimensionality explains the effectiveness of language model fine-tuning. arXiv preprint arXiv:2012.13255, 2020.
  38. Fits: Modeling time series with 10⁢k10𝑘10k10 italic_k parameters. arXiv preprint arXiv:2307.03756, 2023.
  39. Long-term forecasting with tide: Time-series dense encoder. arXiv preprint arXiv:2304.08424, 2023b.
  40. Are transformers effective for time series forecasting? In Proceedings of the AAAI conference on artificial intelligence, volume 37, pages 11121–11128, 2023.
  41. Monash time series forecasting archive. arXiv preprint arXiv:2105.06643, 2021.
  42. The uea multivariate time series classification archive, 2018. arXiv preprint arXiv:1811.00075, 2018.
  43. The ucr time series archive. IEEE/CAA Journal of Automatica Sinica, 6(6):1293–1305, 2019.
  44. Cautionary tales on air-quality improvement in beijing. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 473(2205):20170457, 2017.
  45. Contrast everything: A hierarchical contrastive framework for medical time-series. Advances in Neural Information Processing Systems, 36, 2024.
  46. Scinet: Time series modeling and forecasting with sample convolution and interaction. Advances in Neural Information Processing Systems, 35:5816–5828, 2022b.
  47. Fred-md: A monthly database for macroeconomic research. Journal of Business & Economic Statistics, 34(4):574–589, 2016.
  48. A review and comparison of strategies for multi-step ahead time series forecasting based on the nn5 forecasting competition. Expert systems with applications, 39(8):7067–7083, 2012.
  49. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019.
  50. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
Citations (3)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 3 tweets with 0 likes about this paper.