Papers
Topics
Authors
Recent
Search
2000 character limit reached

Chain-structured neural architecture search for financial time series forecasting

Published 15 Mar 2024 in q-fin.ST and cs.LG | (2403.14695v2)

Abstract: Neural architecture search (NAS) emerged as a way to automatically optimize neural networks for a specific task and dataset. Despite an abundance of research on NAS for images and natural language applications, similar studies for time series data are lacking. Among NAS search spaces, chain-structured are the simplest and most applicable to small datasets like time series. We compare three popular NAS strategies on chain-structured search spaces: Bayesian optimization (specifically Tree-structured Parzen Estimator), the hyperband method, and reinforcement learning in the context of financial time series forecasting. These strategies were employed to optimize simple well-understood neural architectures like the MLP, 1D CNN, and RNN, with more complex temporal fusion transformers (TFT) and their own optimizers included for comparison. We find Bayesian optimization and the hyperband method performing best among the strategies, and RNN and 1D CNN best among the architectures, but all methods were very close to each other with a high variance due to the difficulty of working with financial datasets. We discuss our approach to overcome the variance and provide implementation recommendations for future users and researchers.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)
  1. Optuna: A next-generation hyperparameter optimization framework. Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2019.
  2. Random search for hyper-parameter optimization. J. Mach. Learn. Res., 13:281–305, 2012.
  3. Algorithms for hyper-parameter optimization. In J. Shawe-Taylor, R. Zemel, P. Bartlett, F. Pereira, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 24. Curran Associates, Inc., 2011.
  4. A tutorial on bayesian optimization of expensive cost functions, with application to active user modeling and hierarchical reinforcement learning. ArXiv, abs/1012.2599, 2010.
  5. Towards an empirical foundation for assessing bayesian optimization of hyperparameters. 2013.
  6. Neural architecture search: A survey. Journal of Machine Learning Research, 20(55):1–21, 2019.
  7. Deep Learning. MIT Press, Cambridge, MA, USA, 2016.
  8. Cnnpred: Cnn-based stock market prediction using a diverse set of variables. Expert Systems with Applications, 129:273–285, 2019. https://doi.org/10.1016/j.eswa.2019.03.029.
  9. Temporal fusion transformers for interpretable multi-horizon time series forecasting. International Journal of Forecasting, 37(4):1748–1764, 2021.
  10. Hyperband: A novel bandit-based approach to hyperparameter optimization. Journal of Machine Learning Research, 18(185):1–52, 2018.
  11. Somshubra Majumdar. neural-architecture-search. https://github.com/titu1994/neural-architecture-search, 2018. Accessed: 2023-02-21.
  12. Kerastuner. https://github.com/keras-team/keras-tuner, 2019.
  13. PyTorch Forecasting Contributors. Pytorch forecasting. https://github.com/jdb78/pytorch-forecasting, 2023.
  14. Carl Edward Rasmussen and Christopher K. I. Williams. Gaussian processes for machine learning. Adaptive computation and machine learning. MIT Press, 2006.
  15. Automated Machine Learning in Action. Manning, 2022.
  16. Practical bayesian optimization of machine learning algorithms. In F. Pereira, C.J. Burges, L. Bottou, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 25. Curran Associates, Inc., 2012.
  17. Taking the human out of the loop: A review of bayesian optimization. Proceedings of the IEEE, 104(1):148–175, 2016.
  18. Neural architecture search: Insights from 1000 papers, Jan 2023.
  19. Neural architecture search with reinforcement learning. In ICLR International Conference on Learning Representations, 2017.
Citations (1)

Summary

  • The paper introduces a chain-structured NAS approach that automates the selection of optimal neural network architectures for financial forecasting.
  • It compares NAS strategies including Bayesian optimization, reinforcement learning, and hyperband across multivariate financial datasets from Japan, Germany, and the US.
  • Results indicate that LSTMs and CNNs optimized with hyperband and Bayesian methods marginally outperform other models despite inherent data variability challenges.

The paper "Chain-structured neural architecture search for financial time series forecasting" by Denis Levchenko and colleagues explores the application of neural architecture search (NAS) strategies to financial time series forecasting. The authors aim to improve model performance by automating the selection of the best neural network architecture using NAS techniques.

Introduction

The paper begins by discussing the advantages of deep neural networks (DNNs), notably their ability to perform automatic feature extraction without extensive manual engineering. Despite these benefits, choosing the optimal DNN architecture often relies on manual selection. To address this, Auto-ML and NAS techniques have emerged, automating the architecture selection process. However, most NAS research focuses on domains like computer vision and natural language processing, with limited exploration in time series data analysis. This work attempts to fill that gap by evaluating NAS strategies on financial time series datasets.

Data and Problem Formulation

The research utilizes real-world financial data provided by Predictive Layer SA, focusing on multivariate daily time series for Japanese, German, and US bonds. The task is a binary classification problem, predicting whether a target feature will increase or decrease in value in the future. This involves dealing with datasets characterized by high dimensionality and limited samples, which is challenging for deep learning methods. To manage this, the authors employ feature reduction techniques, including removing time-derived features and applying principal component analysis (PCA).

Architecture Types and Search Spaces

The authors focus on chain-structured search spaces, which involve simple architectural topologies comprising sequential layers. This approach offers simplicity and robustness, suitable for the small datasets they employ. They consider various architectures, including feedforward neural networks (FFNNs), convolutional neural networks (CNNs), and recurrent neural networks (RNNs), each adapted with hyperparameter search spaces. Additionally, the study examines the Temporal Fusion Transformer (TFT) for comparison, though it performed poorly due to the lack of data.

Challenges and Performance Estimation

Predicting financial markets is intrinsically difficult, resulting in metrics close to random performance. Factors like the stochastic nature of deep learning models contribute to high variance, depending on the training's random seed, creating challenges for NAS methods. To mitigate seed-related performance variations, the authors average multiple trial results.

Search Strategies

The paper evaluates three NAS strategies: Bayesian optimization, reinforcement learning, and the hyperband method. These strategies inform the optimal network configuration through iterative model training and evaluation. Each method navigates hyperparameter configurations differently, with Bayesian optimization leveraging probabilistic models, reinforcement learning using a reward-based system, and hyperband employing successive halving to efficiently allocate resources.

Results

The authors report the best results for the German dataset using an LSTM model optimized by the hyperband method, achieving an AUC score of 0.56 on average. For the Japanese dataset, a Bayesian-optimized CNN performs best. The US dataset posed the greatest challenge, with averages not exceeding random prediction.

Architectures and Search Strategies Compared

LSTMs and CNNs generally outperform FFNNs, with Bayesian optimization and hyperband showing a slight performance edge over reinforcement learning. Implementation difficulties and computational efficiency considerations make Bayesian optimization and hyperband more practical.

Discussion and Future Work

The study notes that hyperparameter optimization showed no significant convergence, indicating the complexity inherent in the task. The authors propose exploring further NAS methods like cell-based search spaces and one-shot NAS techniques. They emphasize examining publicly available datasets to ascertain the generalizability of their findings.

In conclusion, this paper contributes to understanding how NAS can be applied to financial time series forecasting, highlighting both the potential and challenges of employing automated architecture selection in such complex domains.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.