Chain-structured neural architecture search for financial time series forecasting
Abstract: Neural architecture search (NAS) emerged as a way to automatically optimize neural networks for a specific task and dataset. Despite an abundance of research on NAS for images and natural language applications, similar studies for time series data are lacking. Among NAS search spaces, chain-structured are the simplest and most applicable to small datasets like time series. We compare three popular NAS strategies on chain-structured search spaces: Bayesian optimization (specifically Tree-structured Parzen Estimator), the hyperband method, and reinforcement learning in the context of financial time series forecasting. These strategies were employed to optimize simple well-understood neural architectures like the MLP, 1D CNN, and RNN, with more complex temporal fusion transformers (TFT) and their own optimizers included for comparison. We find Bayesian optimization and the hyperband method performing best among the strategies, and RNN and 1D CNN best among the architectures, but all methods were very close to each other with a high variance due to the difficulty of working with financial datasets. We discuss our approach to overcome the variance and provide implementation recommendations for future users and researchers.
- Optuna: A next-generation hyperparameter optimization framework. Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2019.
- Random search for hyper-parameter optimization. J. Mach. Learn. Res., 13:281–305, 2012.
- Algorithms for hyper-parameter optimization. In J. Shawe-Taylor, R. Zemel, P. Bartlett, F. Pereira, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 24. Curran Associates, Inc., 2011.
- A tutorial on bayesian optimization of expensive cost functions, with application to active user modeling and hierarchical reinforcement learning. ArXiv, abs/1012.2599, 2010.
- Towards an empirical foundation for assessing bayesian optimization of hyperparameters. 2013.
- Neural architecture search: A survey. Journal of Machine Learning Research, 20(55):1–21, 2019.
- Deep Learning. MIT Press, Cambridge, MA, USA, 2016.
- Cnnpred: Cnn-based stock market prediction using a diverse set of variables. Expert Systems with Applications, 129:273–285, 2019. https://doi.org/10.1016/j.eswa.2019.03.029.
- Temporal fusion transformers for interpretable multi-horizon time series forecasting. International Journal of Forecasting, 37(4):1748–1764, 2021.
- Hyperband: A novel bandit-based approach to hyperparameter optimization. Journal of Machine Learning Research, 18(185):1–52, 2018.
- Somshubra Majumdar. neural-architecture-search. https://github.com/titu1994/neural-architecture-search, 2018. Accessed: 2023-02-21.
- Kerastuner. https://github.com/keras-team/keras-tuner, 2019.
- PyTorch Forecasting Contributors. Pytorch forecasting. https://github.com/jdb78/pytorch-forecasting, 2023.
- Carl Edward Rasmussen and Christopher K. I. Williams. Gaussian processes for machine learning. Adaptive computation and machine learning. MIT Press, 2006.
- Automated Machine Learning in Action. Manning, 2022.
- Practical bayesian optimization of machine learning algorithms. In F. Pereira, C.J. Burges, L. Bottou, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 25. Curran Associates, Inc., 2012.
- Taking the human out of the loop: A review of bayesian optimization. Proceedings of the IEEE, 104(1):148–175, 2016.
- Neural architecture search: Insights from 1000 papers, Jan 2023.
- Neural architecture search with reinforcement learning. In ICLR International Conference on Learning Representations, 2017.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.