Papers
Topics
Authors
Recent
Search
2000 character limit reached

Considering Nonstationary within Multivariate Time Series with Variational Hierarchical Transformer for Forecasting

Published 8 Mar 2024 in cs.LG and cs.AI | (2403.05406v1)

Abstract: The forecasting of Multivariate Time Series (MTS) has long been an important but challenging task. Due to the non-stationary problem across long-distance time steps, previous studies primarily adopt stationarization method to attenuate the non-stationary problem of the original series for better predictability. However, existing methods always adopt the stationarized series, which ignores the inherent non-stationarity, and has difficulty in modeling MTS with complex distributions due to the lack of stochasticity. To tackle these problems, we first develop a powerful hierarchical probabilistic generative module to consider the non-stationarity and stochastic characteristics within MTS, and then combine it with transformer for a well-defined variational generative dynamic model named Hierarchical Time series Variational Transformer (HTV-Trans), which recovers the intrinsic non-stationary information into temporal dependencies. Being a powerful probabilistic model, HTV-Trans is utilized to learn expressive representations of MTS and applied to forecasting tasks. Extensive experiments on diverse datasets show the efficiency of HTV-Trans on MTS forecasting tasks

Definition Search Book Streamline Icon: https://streamlinehq.com
References (29)
  1. Passenger demand forecasting with multi-task convolutional recurrent neural networks. in PAKDD, 29–42.
  2. Box, G. E. P. 1976. Time series analysis, forecasting and control rev. ed. Time series analysis, forecasting and control rev. ed.
  3. Spectral temporal graph neural network for multivariate time-series forecasting. in NeurIPS, 33: 17766–17778.
  4. Autoformer: Searching transformers for visual recognition. in CVPR, 12270–12280.
  5. Infinite Switching Dynamic Probabilistic Network with Bayesian Nonparametric Learning. in TSP.
  6. Switching Poisson gamma dynamical systems. in IJCAI.
  7. Switching Gaussian Mixture Variational RNN for Anomaly Detection of Diverse CDN Websites. in INFOCOM.
  8. SDFVAE: Static and Dynamic Factorized VAE for Anomaly Detection of Multivariate CDN KPIs. in WWW, 3076–3086.
  9. An image is worth 16x16 words: Transformers for image recognition at scale. in ICLR.
  10. Forecasting: principles and practice. OTexts.
  11. Reversible instance normalization for accurate time-series forecasting against distribution shift. in CVPR.
  12. Reformer: The efficient transformer. in ICLR.
  13. Pyraformer: Low-complexity pyramidal attention for long-range time series modeling and forecasting. In ICLR.
  14. Non-stationary Transformers: Exploring the Stationarity in Time Series Forecasting. in NeurIPS.
  15. LSTM-based encoder-decoder for multi-sensor anomaly detection. in ICML.
  16. Adaptive normalization: A novel data normalization approach for non-stationary time series. In The 2010 International Joint Conference on Neural Networks (IJCNN), 1–8. IEEE.
  17. Adaptive Normalization: A novel data normalization approach for non-stationary time series. In International Joint Conference on Neural Networks, IJCNN 2010, Barcelona, Spain, 18-23 July, 2010.
  18. Deep adaptive input normalization for time series forecasting. IEEE transactions on neural networks and learning systems, 31(9): 3760–3765.
  19. Deep Adaptive Input Normalization for Time Series Forecasting. Papers.
  20. DeepAR: Probabilistic forecasting with autoregressive recurrent networks. in International Journal of Forecasting, 36(3): 1181–1191.
  21. Joint modeling of local and global temporal dynamics for multivariate time series forecasting with missing values. in AAAI, 34(04): 5956–5963.
  22. Attention is all you need. in NeurIPS, 30.
  23. Transformers in Time Series: A Survey. in arXiv preprint arXiv:2202.07125.
  24. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in Neural Information Processing Systems, 34: 22419–22430.
  25. Deep multi-view spatial-temporal network for taxi demand prediction. in AAAI, 32(1).
  26. A deep neural network for unsupervised anomaly detection and diagnosis in multivariate time series data. in AAAI, 33(01): 1409–1416.
  27. Crossformer: Transformer Utilizing Cross-Dimension Dependency for Multivariate Time Series Forecasting. In The Eleventh International Conference on Learning Representations.
  28. Informer: Beyond efficient transformer for long sequence time-series forecasting. in AAAI.
  29. FEDformer: Frequency enhanced decomposed transformer for long-term series forecasting. In ICML.
Citations (2)

Summary

  • The paper introduces HTV-Trans, a novel model that integrates a hierarchical probabilistic generative module within a Transformer framework to capture nonstationary dynamics.
  • It incorporates a balancing parameter that effectively manages both stationary and nonstationary patterns, preserving crucial temporal dependencies for improved long-range forecasts.
  • Experimental results on benchmark datasets demonstrate that HTV-Trans outperforms existing Transformer-based models, offering enhanced adaptability to real-world MTS challenges.

Considering Nonstationary within Multivariate Time Series with Variational Hierarchical Transformer for Forecasting

Introduction

The paper "Considering Nonstationary within Multivariate Time Series with Variational Hierarchical Transformer for Forecasting" (2403.05406) addresses the inherent challenges in forecasting multivariate time series (MTS), particularly focusing on the issues of nonstationarity and the stochastic nature of real-world datasets. Historically, methods have leaned towards enforcing stationarity by transforming data to appease models, often at the cost of losing intrinsic temporal dependencies. This work proposes a novel model, the Hierarchical Time series Variational Transformer (HTV-Trans), which integrates a Hierarchical Time Series Probabilistic Generative Module (HTPGM) to effectively model these challenges.

Background and Motivation

Forecasting MTS involves complexities due to correlated temporal dependencies, stochastic elements, and nonstationarity. While recent advances with Transformer models improved forecasting by capturing long-range dependencies, they still suffer from limitations when predicting nonstationary series due to deterministic architectural constraints. Previous solutions often revolved around pre-processing techniques like instance normalization or stationarization, which could lead to over-stationarization, sacrificing the intrinsic properties of the series. The proposed HTV-Trans aims to redress these deficiencies by acknowledging and utilizing the nonstationary and stochastic nature of time series data.

Methodology

HTV-Trans operationalizes its objectives through several key components:

  • Hierarchical Generative Module: This module captures multi-scale nonstationary characteristics, allowing the model to maintain complex data patterns and variability over different time horizons. It structures latent variables hierarchically, representing data distributions that recover the nonstationary and non-deterministic features absent in stationary transformations.
  • Transformer Integration: By embedding HTPGM within the Transformer framework, HTV-Trans leverages the attention mechanism to model intricate temporal dependencies (Figure 1). The addition of a variational inference mechanism enhances the model's representative capacity for MTS, facilitating both reconstruction and prediction tasks.
  • Parameter Balance: The model introduces a parameter α\alpha to balance between stationary and non-stationary information, facilitating the recovery of essential temporal structure (Figure 2). This balance is critical for injecting just the right amount of stochasticity back into the series. Figure 3

    Figure 3: Graphical illustration of different operations of the HTV-Trans: (a) generative process of HTPGM, (b) the inference scheme of HTPGM, (c) the fusion of different scale information and stationarization input series for forecasting.

Experimental Results

The performance evaluation of HTV-Trans was comprehensive, involving multiple datasets such as ETTh1, ETTh2, ETTm1, ETTm2, Weather, and Exchange (Table 1). The results indicated that HTV-Trans consistently outperforms existing Transformer-based models, particularly in scenarios involving long-range forecasting. This superiority is attributed to HTV-Trans's ability to maintain non-stationary dynamics within its latent representations, aiding in better adaptation to complex, real-world data patterns. Figure 4

Figure 4: The effectiveness evaluation of hierarchical architecture on ETTh1 (left), ETTh2 (middle), and ETTm1 (right) datasets.

Implications and Future Directions

The development of HTV-Trans represents a significant step toward more adaptive and realistic MTS forecasting that acknowledges rather than overlooks the nonstationarity and stochasticity inherent in real data. This model not only enhances forecasting accuracy but also sets a precedent for future explorations into hybrid methods that synergize traditional and deep learning paradigms. Potential extensions of this work could explore more sophisticated fusion techniques within the transformer framework to further enhance adaptability, or alternatively, evaluate the HTV-Trans model across more diverse datasets and real-world applications.

Conclusion

HTV-Trans provides a robust framework for addressing nonstationary and stochastic challenges in MTS forecasting by intricately combining probabilistic generative modeling with advanced attention mechanisms. It stands as a versatile model capable of improving forecasting accuracy across varying domains, asserting the importance of incorporating inherent data variabilities into deep learning architectures. This research holds promise not only for improving time series forecasting techniques but also for inspiring innovative approaches to dealing with the dynamic nature of real-world datasets.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Authors (3)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.